[GsoC 2026 ] Agentic GraphRAG #36 on STaRK-MAG with OpenVINO — Working System + draft Proposal #34801
1Suryansh1
started this conversation in
Google Summer of Code
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi @bharagha @14pankaj,
I'm Suryansh Verma , B,Tech CSE Student from Delhi Technological University. I have a working, benchmarked GraphRAG system running on Intel OpenVINO, and I'd like to share it as my warm-up contribution for Project 36.
My Laptop is intel AI PC with Intel Core Ultra 7 155H with integrated Intel Arc Graphics and NPU — all inference runs on-device via OpenVINO
Results
I evaluated the current system on the STaRK-MAG human-generated split with FULL ALL 84 queries in a fully local, zero-shot setting on 1,872,968 NODES . The current no explicit rerank run achieves Hit@5 = 53.57 with average latency of about 10.2 seconds per query.
human_eval is considered tougher than synthetic split and synthetic has thousands of queries.
the exact numbers here https://arxiv.org/html/2404.13207v3
A few points matter here:
Current System
I built a graph retrieval tool for agents, implemented and benchmarked on STaRK-MAG. The system is designed as a reusable retrieval backend for agentic workflows, not just a benchmark script.
Offline setup
Online retrieval flow
Engineering choices
OpenVINO implementation
Project 36 Direction
My proposal is to package this retrieval core as a GraphQueryTool inside a LangGraph agentic pipeline. The traversal engine remains a deterministic retrieval tool in the loop, while OpenVINO SLMs handle query understanding, reflection, and answer synthesis.
Text-to-Cypher fails silently on schemas — the LLM hallucinates relationship names and returns zero results with no recovery. My Anchor SLM extracts schema-free JSON entities instead. It never generates Cypher. Schema knowledge lives in the retrieval layer, not the LLM.
Project 37 compatibility
( with @naitik-2006 fine tuning and fulfilling its requirement could be done , I have connected with him via linkedin , we can collaborate together towards converging solution with your guidance )
The three SLM layers can be fine tuned properly.
The graph traversal engine is deterministic — no fine-tuning required, keeping the gradient signal clean and targeted.
The same structure is also naturally extensible to multimodal retrieval because the seed-retrieval layer can ingest new embedding types without changing the traversal core.
Conclusion
I'd like to propose this benchmarked implementation as my warm-up contribution for Project 36. It already demonstrates working Neo4j integration, OpenVINO inference, million-node-scale graph traversal, and retrieval evaluation on STaRK-MAG, which I believe maps directly to the goals of the project.
The choice of model for SLM could be further decided and extended to support multimodal nature also.
I'd be happy to open a PR integrating this direction into the edge-ai-libraries ChatQnA pipeline.
Your suggestions and directions would be very helpful before I draft my final proposal
the image shows the PRESENT SYSTEM which would act as deterministic tool

the image below shows the proposed architecture

For Further clarification plz have a look at my architecture images
GitHub: https://github.com/1Suryansh1/graphRAG
I have added the mentors as collaborator and to guide for next steps and corrections ...
Suryansh Verma
B.Tech CSE
Delhi Technological University
Beta Was this translation helpful? Give feedback.
All reactions