[Question]:Cross-chunk relations & multi-hop retrieval design choices in LightRAG #1830
Replies: 3 comments 1 reply
-
Upvote this question. |
Beta Was this translation helpful? Give feedback.
-
Reducing the chunk size and increasing the overlap size may help, but it won't be enough on its own. |
Beta Was this translation helpful? Give feedback.
-
this is a super sharp observation — you're actually brushing up against one of the core edge cases in chunk-based RAG systems:
your mention of "LLM-inferred links across high-frequency entity pairs" is especially important — many systems skip this due to cost or complexity, but without it, you often end up with confident completions that are logically disconnected across sessions. curious if you've tested this on anything like character arcs or evolving definitions (e.g. legal terms across a contract corpus)? you'd be surprised how easily the current design masks drift under the hood. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Do you need to ask a question?
Your Question
Hi LightRAG team 👋,
First, thanks for releasing such a clean and practical RAG framework—its dual-level retrieval has been working great for us!
While reviewing the indexing flow I noticed that
D(·)
become shared nodes, but no new edges are created between entities that never co-occur in the same chunk.I’d love to understand the design rationale and its impact on deeper reasoning:
Why omit explicit cross-chunk edge creation?
Multi-hop coverage
Possible extensions
Would the project welcome an option to add lightweight cross-chunk edges, such as:
Prompting an LLM on a vector‑based top‑k chunk set to infer cross‑chunk relations
Any insights, papers, or code pointers would be greatly appreciated.
Thanks again for the excellent work!
Best regards,
SangwookBaek
Additional Context
No response
Beta Was this translation helpful? Give feedback.
All reactions