I move beyond static models to build resilient, adaptive AI architectures.
Currently focused on GraphRAG, Time-Series Forecasting, and MLOps.
Solving the "Lost in the Middle" problem in long-context retrieval.
- Architecture: Hybrid Retrieval (Neo4j Graph + LanceDB Vectors).
- Key Tech: LangGraph, LangSmith, OpenAI API, Docker.
- Result: Increased context fidelity by 15% over standard RAG baselines.
Deploying 12B-parameter models on consumer hardware (6GB VRAM).
- Architecture: Hybrid Pipeline (WSL2 + Windows).
- Key Tech: 4-bit GGUF Quantization, LoRA Fine-Tuning, WebSocket Bridge.
- Result: Reduced memory footprint by 71% (23.8GB → 6.8GB).