Panini proposes a non-parametric continual learning framework for LLMs using Generative Semantic Workspaces (GSW), an entity- and event-aware QA network that consolidates experiences without updating the base model. It outperforms RAG baselines by 5-7% on six QA benchmarks while using 2-30x fewer tokens and reducing unsupported answers. Open-source code is available on GitHub.
Key Points
- 1.Introduces GSW as entity/event-aware QA networks for document representation
- 2.Retrieves inference chains from GSW instead of verbatim chunks for efficiency
- 3.Outperforms baselines by 5-7% on QA benchmarks with 2-30x token savings
- 4.Reduces unsupported answers on unanswerable queries
- 5.Fully open-source with GitHub repo
Impact Analysis
Panini offers a more efficient alternative to RAG for handling evolving data, potentially cutting inference costs and improving reliability in production LLM apps. Researchers and builders can adopt it for continual knowledge integration without retraining.
Technical Details
GSW structures documents into interconnected QA pairs, enabling LLM reasoning over latent knowledge via traversable inference chains. The memory self-consolidates new experiences at write time, optimizing read-time retrieval without redundant compute.