AtlasKV: Augmenting LLMs with Billion-Scale Knowledge Graphs in 20GB VRAM
📰 ArXiv cs.AI
arXiv:2510.17934v2 Announce Type: replace-cross Abstract: Retrieval-augmented generation (RAG) has shown some success in augmenting large language models (LLMs) with external knowledge. However, as a non-parametric knowledge integration paradigm for LLMs, RAG methods heavily rely on external retrieval modules and the retrieved textual context prior. Especially for very large scale knowledge augmentation, they would introduce substantial inference latency due to expensive searches and much longer
DeepCamp AI