RAG-Based Knowledge Retrieval System
Scalable Retrieval-Augmented Generation system with 100K+ embeddings using LangChain, FastAPI, and Weaviate. Optimized retrieval accuracy with chunking strategies and context window management.
If it's not in prod, it doesn't exist.
Latency is the silent killer of user trust.
RAG is easy. Good RAG is hard.
Context engineering > prompt engineering.
Agents are just while loops with ambition.
Trace everything. Assume nothing.

GenAI Engineer specializing in LLM systems, RAG pipelines, multi-agent orchestration, and production-grade AI infrastructure on AWS and GCP.
Performance metrics from production AI systems
From prototype to production-grade infrastructure
Deep dives into products I've shipped from 0 to 1
Scalable Retrieval-Augmented Generation system with 100K+ embeddings using LangChain, FastAPI, and Weaviate. Optimized retrieval accuracy with chunking strategies and context window management.
Orchestrated multi-agent workflows using LangGraph and LangChain for complex task execution. Integrated LLMs with external systems via APIs, messaging, and automation platforms.
End-to-end document processing system using AWS Lambda, SQS, S3, and OCR engines. Implements pipelines for document ingestion, classification, validation, and structured data extraction.
Building production AI systems from day one
Xperion (Remote)
Texagon
Product leadership capabilities I bring to every team
Tools and platforms I use to ship AI products
Looking to build production AI systems or need a GenAI engineer? Let's talk architecture.
Trusted By
Companies I've helped build and scale products
Wall of Love
Feedback from colleagues and collaborators