Beyond Code Snippets: Benchmarking LLMs on Repository-Level Question Answering

📰 ArXiv cs.AI

arXiv:2603.26567v1 Announce Type: cross Abstract: Large Language Models (LLMs) have shown impressive capabilities across software engineering tasks, including question answering (QA). However, most studies and benchmarks focus on isolated functions or single-file snippets, overlooking the challenges of real-world program comprehension, which often spans multiple files and system-level dependencies. In this work, we introduce StackRepoQA, the first multi-project, repository-level question answeri

Published 30 Mar 2026
Read full paper → ← Back to News