<RETURN_TO_BASE

ByteDance Unveils VGR: Advanced Multimodal LLM with Superior Visual Reasoning

ByteDance introduces VGR, a multimodal large language model that dynamically integrates visual information during reasoning, significantly improving accuracy and efficiency in vision-language benchmarks.

The Importance of Multimodal Reasoning in Vision-Language Tasks

Multimodal reasoning is essential for models to effectively combine visual and textual data, enabling them to interpret charts, answer image-related questions, and understand complex visuals. This capability allows machines to not only see but also comprehend visual information and connect it with language-based reasoning.

Challenges in Visual Reasoning Due to Language Bias

Many existing models rely heavily on linguistic cues even when tasks demand detailed visual analysis. This leads to poor performance in perception-focused applications, where identifying objects or interpreting numerical data from images is crucial. These models tend to guess answers based on language patterns instead of analyzing the visual content, limiting their effectiveness.

Shortcomings of Current Vision-Language Models

Although several techniques have been proposed to enhance model performance, they often fail to capture fine visual details. Some use pre-generated captions or annotated image regions, while others depend on structured prompts for reasoning. However, these approaches suffer from static visual inputs and inflexible processing pipelines, which hinder the integration of vision and reasoning, especially for diverse and open-ended queries.

Introducing VGR: Visual Grounded Reasoning Framework

ByteDance Inc. and the University of Chinese Academy of Sciences developed VGR, a novel model that dynamically interacts with visual elements during reasoning. Unlike traditional models, VGR simultaneously processes image and text streams by identifying key image regions relevant to the question and incorporating them into the answer generation process. Along with VGR, the researchers created the VGR-SFT dataset, enabling learning of visual reasoning with embedded image clues without manual annotations.

Selective Visual Replay for Efficient Image Reasoning

At the heart of VGR lies selective visual replay, a technique allowing the model to retrieve specific image parts as needed. A vision encoder extracts tokens from image regions, storing them in a visual memory pool. When visual information is required during reasoning, the model signals a replay, reintroducing relevant image tokens into the reasoning stream. The AnyRes strategy broadens resolution support while reducing token consumption. Compared to baselines, VGR uses only 144 tokens for image snapshots and 720 tokens for high-resolution regions, cutting token usage by 70%. Training combines supervised learning with an auxiliary loss to improve region selection and interpretation.

Benchmark Performance: Accuracy and Efficiency

Testing VGR against the LLaVA-NeXT-7B baseline yielded impressive results. On the MMStar benchmark, VGR improved accuracy by 4.1 points; on AI2D by 7.1 points; and on ChartQA by 12.9 points, while using only 30% of the visual tokens required by the baseline. In another evaluation, VGR increased MMStar performance by 6.4 points and ChartQA by 14.1 points, demonstrating efficient and accurate multimodal reasoning with fewer resources.

Advancing Beyond Text-Centric Reasoning

This work underscores that integrating visual inputs thoughtfully within reasoning processes overcomes the limitations of text-only deduction. The researchers identified a critical challenge, devised a precise method, and validated its effectiveness with strong empirical results. VGR sets a new standard for combining vision and language in intelligent systems.

For more detailed information, check out the original paper and model release from the researchers.

🇷🇺

Сменить язык

Читать эту статью на русском

Переключить на Русский