Concept frustration: Aligning human concepts and machine representations
📰 ArXiv cs.AI
arXiv:2603.29654v1 Announce Type: cross Abstract: Aligning human-interpretable concepts with the internal representations learned by modern machine learning systems remains a central challenge for interpretable AI. We introduce a geometric framework for comparing supervised human concepts with unsupervised intermediate representations extracted from foundation model embeddings. Motivated by the role of conceptual leaps in scientific discovery, we formalise the notion of concept frustration: a co
DeepCamp AI