Large Language Models
Apr 14, 2026
New Introspective Diffusion Language Model Matches Quality of Autoregressive Models
Apr 14, 2026
AI Summary
The Introspective Diffusion Language Model (I-DLM) has been developed to address quality gaps in diffusion language models compared to autoregressive models. I-DLM-8B has demonstrated performance on par with its autoregressive counterpart while achieving higher throughput and efficiency.
- Diffusion language models (DLMs) have potential for parallel token generation but often lag in quality behind autoregressive (AR) models.
- The I-DLM introduces introspective strided decoding (ISD) to verify previously generated tokens while generating new ones in a single forward pass.
- I-DLM-8B matches the quality of a same-scale AR model and outperforms LLaDA-2.1-mini (16B) in benchmarks, achieving significant throughput improvements.
- Three bottlenecks in current DLMs are identified: conversion of pretrained AR models, token generation and verification, and integration into existing frameworks.
- I-DLM achieves 2.1-2.5 times throughput over naive baselines and maintains efficiency at high concurrency levels.
- The model utilizes gated LoRA for lossless acceleration and strict causal attention for seamless integration into SGLang.
- I-DLM has been evaluated across 15 benchmarks, demonstrating superior performance compared to previous DLMs.
diffusion modelslanguage modelsai researchnlpmachine learning