Dimple: Singapore Researchers Revolutionize Multimodal Text Generation with Discrete Diffusion
Researchers at the National University of Singapore introduce Dimple, a novel discrete diffusion multimodal language model that enhances text generation efficiency and control, outperforming similar autoregressive models.
Emerging Interest in Diffusion Models for Language
Recent months have seen a surge in applying diffusion models, originally crafted for continuous data like images, to natural language processing. This shift has led to the creation of Discrete Diffusion Language Models (DLMs), which interpret text generation as a denoising process. Compared to traditional autoregressive models, DLMs offer benefits such as parallel decoding, enhanced structural control, flexible sequence initialization, explicit output formatting, and improved infilling via bidirectional attention. Moreover, their non-sequential generation allows faster text production. Despite these advantages, most multimodal large language models (MLLMs) like LLaMA, Qwen-VL, and InternVL still depend exclusively on autoregressive methods.
Diverse Approaches in Diffusion-Based Language Models
Work on diffusion language models has explored continuous and discrete diffusion spaces. Continuous methods (e.g., DiffuSeq, SED) utilize embedding or relaxed categorical spaces for smoother generation, while discrete models (e.g., SDDM, RDM) customize diffusion processes tailored to linguistic structures. Training often involves masked language modeling losses or entropy-based score matching. Hybrid models such as AR-Diffusion and SSD-LM combine autoregressive and diffusion techniques to harness the strengths of both. Meanwhile, open-source MLLMs like LLaVA and InternVL have advanced via visual instruction tuning and joint pretraining but remain rooted in autoregressive generation.
Introducing Dimple: The First Discrete Diffusion Multimodal LLM
Researchers at the National University of Singapore have unveiled Dimple, the first Discrete Diffusion Multimodal Large Language Model (DMLLM) integrating a vision encoder with a discrete diffusion-based language model. To address instability and performance challenges in pure diffusion training, they developed a two-phase training strategy: starting with autoregressive alignment, followed by diffusion-based masked language modeling. The Dimple-7B model outperforms LLaVA-NEXT by 3.9% on benchmarks. Innovations include Confident Decoding for dynamic token generation and Structure Priors for precise output control, significantly enhancing inference efficiency, generation flexibility, and structural controllability without compromising performance.
Training and Inference Strategies
Dimple tackles diffusion training inefficiencies like sparse supervision and limited generation coverage by adopting a two-phase training process. Initially, autoregressive training with causal attention aligns vision and language modalities. Subsequently, diffusion training restores generation capabilities. During inference, a dynamic Confident Decoding method adapts token updates based on prediction confidence, accelerating generation. Despite fewer training samples, Dimple achieves competitive benchmark results, outperforming similarly sized autoregressive models though trailing behind larger state-of-the-art systems.
Performance Evaluation and Benefits
Evaluations of Dimple against autoregressive models on instruction-following tasks demonstrate strong performance, often surpassing models trained with comparable data. While it does not match models trained on vastly larger datasets, Dimple benefits from a robust base language model. Ablation studies show that combining autoregressive and diffusion tuning reduces length bias and enhances consistency. Prefilling techniques further boost inference speed with minimal performance loss, making Dimple efficient and competitive for multimodal understanding.
Innovations in Output Control and Efficiency
Dimple’s hybrid training overcomes limitations of pure discrete diffusion, such as instability and length bias, by blending autoregressive and diffusion learning. The confident decoding strategy reduces inference steps notably. Prefilling improves speed with slight performance trade-offs. Additionally, Structure Priors enable fine-grained and controllable outputs, offering formatting and length control that autoregressive models struggle to achieve.
Explore the detailed research paper, model on Hugging Face, and GitHub repository for more information. Follow related discussions on Twitter and join the active ML SubReddit community.
Сменить язык
Читать эту статью на русском