CropVLM: Learning to Zoom for Fine-Grained Vision-Language Perception
📰 ArXiv cs.AI
arXiv:2511.19820v2 Announce Type: replace-cross Abstract: Vision-Language Models (VLMs) often struggle with tasks that require fine-grained image understanding, such as scene-text recognition or document analysis, due to perception limitations and visual fragmentation. To address these challenges, we introduce CropVLM as an external low-cost method for boosting performance, enabling VLMs to dynamically ''zoom in'' on relevant image regions, enhancing their ability to capture fine details. CropVL
DeepCamp AI