Project Case StudyDevOps / AI // 2025

LogrCtx

An intelligent log reduction and analysis system that turns millions of raw log lines into actionable insights using AI and distributed stream processing.

The Problem

Drowning in Data

Modern microservices generate terabytes of logs daily. In a recent case study with the `dash-cart` service, we were dealing with 900,000+ log lines for a single incident.

Debugging meant scrolling through endless "INFO" messages to find the one "ERROR" that mattered. We needed a way to drain the noise without losing the signal.

Original Volume
900k
Lines of Logs
Reduced Volume
730
Meaningful Patterns
99.92%
Reduction Ratio
The Engine

Intelligent Drainer

Using the Drain3 algorithm, LogrCtx parses logs in real-time, extracting dynamic variables (timestamps, IDs) to identify static templates. This allows us to cluster millions of similar logs into a single representative pattern.

Live Reduction Demo

INTERACTIVE
LOG_REDUCTION_ENGINE_V1
PROCESSING
Raw Stream (1000/s)
Reduced Insight
WAITING FOR PATTERNS...
Select Scenario
AI_AGENT_STATUS
IDLE / READY
retrieval
Scanning 50,000 log lines...
cluster
Identified 3 relevant error clusters.
context
Correlating with 'Service C' database metrics.
insight
Root Cause: Database connection pool exhaustion in Service C caused cascading timeouts in Checkout Service.
The Intelligence

RAG-Powered Insights

Reduction is just the first step. LogrCtx uses Retrieval-Augmented Generation (RAG) to understand the context of the reduced logs.

  • Context-Aware Summaries

    Translates cryptic error codes into human-readable root cause analysis.

  • Comparative Analysis

    Automatically compares logs between Canary and Stable deployments to detect regressions.

Distributed Streaming Architecture

Built to handle massive throughput. The system decouples ingestion from processing using a distributed message queue and pub/sub architecture.

03 // SYSTEM_ARCHITECTURE
● LIVERENDERER: WebGL
FPS: 60
Client Tier
API Gateway
Worker Mesh
NODES: 6WORKERS: 100CONNECTIONS: 6

RabbitMQ Task Queue

Handles backpressure by buffering incoming log streams before distributing them to worker nodes.

Elastic Worker Pool

Stateless worker nodes that scale horizontally to process log chunks in parallel using the Drain3 algorithm.

NATS Real-time Stream

Ultra-low latency pub/sub system that streams processed results back to the client via SSE.

VIEW_MODE: DASHBOARD
Background
Slide Image
VISUAL_FEED
SEQ_01
::

INTERFACE

THE_LENS

VISUALIZING THE
INVISIBLE

> Raw data is useless without insight. The LogrCtx frontend transforms abstract log streams into actionable intelligence using a high-performance React + AntD interface.

01 //

LOGQL_SUPPORT

Native support for complex queries to filter, aggregate, and analyze logs.

02 //

AUTO_RCA

One-click Root Cause Analysis tracing failures to the exact line of code.

03 //

PATTERN_ID

Visualizes log structures and templates for intuitive recognition.

IMPACT

OUTCOME

FROM PROTOTYPE TO
PRODUCTION SCALE

STATUS: DEPLOYED
ORG: SWIGGY
01 // NOISE_REDUCTION
99.9%

Condensed 900k+ daily log lines into < 1000 actionable patterns.

02 // MTTR_IMPACT
-40%

Drastic reduction in recovery time for DevOps teams during incidents.

03 // STACK_FUSION
DevOpsLLMOpsDistributedFull Stack

Engineering convergence of infrastructure, AI, and frontend systems.

From Napkin to Node

Original Handwritten Architecture Diagram

Fig 1.0: The initial handwritten design of the distributed worker system.

API-First Design

POST/api/v1/reduce

Authorization: Bearer <token>

Content-Type: application/json

{ "service": "checkout-api", "time_range": "1h", "reduction_rate": 0.95 }

LogrCtx exposes a full REST API, allowing developers to integrate log reduction directly into their CI/CD pipelines or custom dashboards.