Beyond Code Snippets: Benchmarking LLMs on Repository-Level Question Answering
📰 ArXiv cs.AI
arXiv:2603.26567v1 Announce Type: cross Abstract: Large Language Models (LLMs) have shown impressive capabilities across software engineering tasks, including question answering (QA). However, most studies and benchmarks focus on isolated functions or single-file snippets, overlooking the challenges of real-world program comprehension, which often spans multiple files and system-level dependencies. In this work, we introduce StackRepoQA, the first multi-project, repository-level question answeri
DeepCamp AI