HyperMem: Hypergraph Memory for Long-Term Conversations

📰 ArXiv cs.AI

arXiv:2604.08256v2 Announce Type: replace-cross Abstract: Long-term memory is essential for conversational agents to maintain coherence, track persistent tasks, and provide personalized interactions across extended dialogues. However, existing approaches as Retrieval-Augmented Generation (RAG) and graph-based memory mostly rely on pairwise relations, which can hardly capture high-order associations, i.e., joint dependencies among multiple elements, causing fragmented retrieval. To this end, we p

Published 13 Apr 2026
Read full paper → ← Back to Reads