<RETURN_TO_BASE

Salesforce Unveils BLIP3-o: Open-Source Multimodal Model Combining CLIP Embeddings and Flow Matching for Image Understanding and Generation

Salesforce has released BLIP3-o, an open-source multimodal model that unifies image understanding and generation using CLIP embeddings and Flow Matching, achieving state-of-the-art results.

Bridging Vision and Language with Unified Multimodal Models

Multimodal modeling aims to develop systems capable of both understanding and generating content across visual and textual formats. These models interpret visual scenes and create new images from natural language prompts, integrating image recognition and generation into a single unified architecture. This approach eliminates the need for separate pipelines and enables more coherent cross-modal interactions.

Challenges in Unified Image Understanding and Generation

A significant challenge is designing architectures that excel at both image understanding and generation without compromising either. Models must comprehend complex visual concepts while producing high-quality images aligned with user prompts. Achieving this requires effective image representations and training methods that support both tasks simultaneously, aligning semantic interpretation with pixel-level synthesis.

Previous Methods and Limitations

Traditional approaches often rely on Variational Autoencoders (VAEs) or CLIP-based encoders for image representation. While VAEs efficiently reconstruct images, they capture lower-level features, resulting in less informative embeddings. CLIP encoders provide rich semantic embeddings learned from large image-text datasets but were not designed for image reconstruction, complicating their use for generation unless paired with diffusion decoders. Mean Squared Error (MSE) loss, commonly used in training, tends to yield deterministic outputs, limiting diversity and quality.

Introducing BLIP3-o: A New Unified Model

Salesforce Research, in collaboration with the University of Maryland and other institutions, introduced BLIP3-o, a family of unified multimodal models. BLIP3-o employs a dual-stage training strategy: first focusing on image understanding, then on image generation. It utilizes CLIP embeddings combined with a diffusion transformer to synthesize new images. The diffusion module is trained independently with a frozen autoregressive backbone to prevent task interference.

Data and Model Architecture

The team developed BLIP3o-60k, a high-quality instruction-tuning dataset generated via GPT-4o, covering a wide range of visual categories such as scenes, objects, gestures, and text. Two BLIP3-o model sizes were created: an 8-billion parameter version trained on both proprietary and public data, and a 4-billion parameter version trained solely on open-source data.

The image generation pipeline is built upon Qwen2.5-VL large language models. Input prompts are transformed into visual features refined by a Flow Matching diffusion transformer based on the Lumina-Next architecture, optimized for speed and quality with 3D rotary position embeddings and grouped-query attention. Images are encoded into 64 fixed-length semantic vectors, regardless of resolution, enabling efficient storage and decoding.

Extensive Training and Evaluation

Training leveraged a massive dataset of 25 million images from sources like CC12M, SA-1B, and JourneyDB, supplemented by 30 million proprietary samples for the 8B model. The instruction-tuning dataset includes 60k samples with challenging prompts, generated by GPT-4o.

BLIP3-o demonstrated outstanding performance across benchmarks. The 8B model scored 0.84 in GenEval for image generation alignment, 0.62 in WISE for reasoning, and achieved high scores in MME-Perception (1682.6), MME-Cognition (647.1), MMMU (50.6), VQAv2 (83.1), and TextVQA (83.1). Human evaluations showed BLIP3-o 8B was preferred over Janus Pro 7B in visual quality (50.4%) and prompt alignment (51.5%) with statistically significant results.

Significance and Open Access

BLIP3-o addresses the dual challenge of image understanding and generation by combining CLIP embeddings, Flow Matching, and a sequential training strategy. It achieves state-of-the-art results while providing an efficient, fully open-source solution for unified multimodal modeling.

Explore the Paper, GitHub, and Hugging Face Model. Follow the research team on Twitter and join the 90k+ ML SubReddit for more updates.

🇷🇺

Сменить язык

Читать эту статью на русском

Переключить на Русский