Introspective Diffusion Language Models

hackernews

Introspective Diffusion
Language Models

Together AI • UIUC • Princeton • Stanford • UT Austin
* Equal contribution   † Equal advising   ◊ Corresponding author
69.6
AIME-24 (I-DLM-8B)
vs. LLaDA-2.1-mini 43.3
45.7
LCB-v6 (I-DLM-8B)
vs. LLaDA-2.1-mini 30.4
2.9-4.1x
Throughput over
LLaDA-2.1-mini at C=64
Lossless
Bit-for-bit identical
to base AR model

Abstract

Diffusion language models (DLMs) offer a compelling promise: parallel token generation could break the sequential bottleneck of autoregressive (AR) decoding. Yet in practice, DLMs consistently lag behind AR models in quality.

We argue that this gap stems from a fundamental failure of introspective consistency: AR models agree with what they generate, whereas DLMs often do not. We introduce the Introspective Diffusion Language Model (I-DLM), which uses introspective strided decoding (ISD) to verify previously generated tokens while advancing new ones in the same forward pass.

Empirically, I-DLM-8B is the first DLM to match the quality of its same-scale AR counterpart, outperforming LLaDA-2.1-mini (16B) by +26 on AIME-24 and +15 on LiveCodeBench-v6 with half the parameters, while delivering 2.9-4.1x throughput at high concurrency. With gated LoRA, ISD enables bit-for-bit lossless acceleration.


Why Introspective Consistency?

Key Insight: AR training unifies generation and introspection in one forward pass. Existing DLMs miss this — they learn to denoise but not to introspect.

We identify three fundamental bottlenecks in current DLMs:

(1) Low introspective consistency. SDAR: 0.699 vs. I-DLM: 0.984.
(2) Compute inefficiency. TiDAR: ~7.8x overhead vs. I-DLM: ~2.5x.
(3) Infrastructure mismatch. SDAR slope=84 vs. I-DLM: 549.

The I-DLM Method

Introspective-Consistency Training

Convert pretrained AR models via causal attention, logit shift, and an all-masked objective.

Introspective Strided Decoding

Generate N tokens per forward pass while verifying prior tokens via the p/q acceptance criterion.

AR-Compatible Serving

Strict causal attention enables direct integration into SGLang with no custom infrastructure.

Decoding paradigm comparison. I-DLM is a drop-in replacement within AR serving infrastructure.

Results

I-DLM is the first DLM to match same-scale AR quality while surpassing all prior DLMs across 15 benchmarks.

End-to-End Quality

Blue = best non-AR <30B. Bold = best non-AR <100B.

Qwen3
8B
Qwen3
32B
LLaDA-2.1
-mini 16B
LLaDA-2.0
-flash 100B
LLaDA-2.1
-flash 100B
SDAR
8B
SDAR
30B
Mercury
Coder
Gemini
Diffusion
I-DLM
8B
I-DLM
32B
Knowledge & Reasoning
ARC-C95.897.290.2------91.993.2------95.896.8
MMLU83.587.274.5------78.682.8------82.486.8
MMLU-Pro75.180.164.874.876.656.961.5------73.179.7
GPQA-D58.964.146.0------40.236.7------55.662.1
GPQA55.465.053.362.367.3------------54.958.7
Math
GSM8K96.094.789.0------91.791.4------95.094.9
MATH-50095.897.885.0------78.677.8------96.897.6
MathBench93.195.584.2------76.979.3------89.195.6
AIME-2473.176.743.3------10.016.7------69.683.3
AIME-2565.480.043.360.063.310.010.8------60.880.0
Code
HumanEval95.196.386.0------78.787.290.089.693.396.3
MBPP93.495.782.1------72.071.676.676.092.294.6
LCB-v650.358.330.442.545.416.621.7------45.757.1
Instruction Following
IFEval84.784.583.282.683.661.460.6------84.784.7

Throughput

Throughput-latency tradeoff compared with DLMs across batch sizes (1, 4, 16, 64). I-DLM delivers 2.9-4.1x higher throughput than LLaDA-2.1-mini and SDAR at C=64.

Speedup Factor Explorer

In the memory-bound decode regime, TPF closely approximates wall-clock speedup: a TPF of 2.5 represents roughly 2.5x faster decoding than AR. Explore how acceptance rate and stride size affect this below.

0.90
1.12

Gated LoRA adds compute at MASK positions for bit-for-bit lossless output. α=1.12 matches empirical overhead.

--
N=2
--
N=3
--
N=4
--
N=8
Memory-bound: Speedup ≈ TPF = (2+p+...+pN-2) / (2-pN-1)
R-ISD (lossless): Speedup ≈ TPF / α — gated LoRA guarantees bit-for-bit AR output.
How do DLMs perform as they approach compute-bound?

At high concurrency, forward pass latency scales with query count per forward. We can measure compute efficiency as TPF²/query_size — how much useful output each FLOP produces relative to AR (efficiency = 1):

  • SDAR (N=4, p=0.5): TPF ≈ 1.1, processes N=4 queries/forward → compute efficiency = 1.1²/4 ≈ 0.31. Each FLOP produces only 31% as much output as AR. This pushes SDAR into compute-bound early, and its throughput plateaus (batching efficiency slope = 84, see motivation figure).
  • I-DLM (N=4, p=0.9): TPF ≈ 2.9, processes 2N−1=7 queries/forward → compute efficiency = 2.9²/7 ≈ 1.22. Each FLOP produces more useful output than AR — I-DLM stays in the memory-bound regime at concurrency levels where SDAR is already saturated (batching efficiency slope = 549).

Efficiency > 1 means parallel decoding actually saves total compute vs. AR. This is why I-DLM's throughput scales with concurrency while SDAR and LLaDA plateau in the throughput figure above.

Per-Position Acceptance Breakdown

Acceptance compounds geometrically: position k has probability $p^{k-1}$. Position 1 is always accepted (logit shift).


Documentation & Resources

Everything you need to train, serve, and deploy I-DLM. Click any card to expand.

Installation

git clone https://github.com/Introspective-Diffusion/I-DLM.git
cd I-DLM/inference
bash install.sh

See inference/README.md for detailed environment setup.

Quick Start

1. Launch server:

python -m sglang.launch_server \
    --model-path yifanyu/I-DLM-8B \
    --trust-remote-code --tp-size 1 --dtype bfloat16 \
    --mem-fraction-static 0.85 --max-running-requests 32 \
    --attention-backend flashinfer --dllm-algorithm IDLMBlockN \
    --dllm-algorithm-config inference/configs/idlm_blockN4_config.yaml \
    --port 30000

2. Generate:

curl http://localhost:30000/v1/chat/completions \
  -H "Content-Type: application/json" \
  -d '{
    "model": "default",
    "messages": [{"role": "user", "content": "Prove that sqrt(2) is irrational."}],
    "max_tokens": 4096, "temperature": 1.0
  }'

Training

Convert a pretrained AR model into I-DLM via introspective-consistency training:

  • Input: Concatenate fully-masked sequence with clean sequence [x_t | x_0]
  • Attention: Strict causal masking across all positions
  • Loss: Auto-balanced CE on both masked and clean positions
  • Data: 4.5B tokens, 8 H100 GPUs, 2 epochs with stride curriculum (N=2 then N=3)

See training/README.md for scripts and configs.

Inference & ISD

Introspective Strided Decoding (ISD) generates and verifies in a single forward pass:

  • MASK positions: Propose new tokens (distribution q)
  • Clean positions: Verify prior tokens (anchor distribution p)
  • Acceptance: min(1, p(x)/q(x)) guarantees AR-distribution output
  • Stride N=4: TPF=2.96, ~3x wall-clock speedup in memory-bound regime

See inference/README.md for algorithm configs.

Serving (SGLang)

I-DLM uses strict causal attention, enabling direct integration into SGLang with no custom infrastructure:

  • Paged KV cache and continuous batching
  • CUDA graph capture (+42-76% throughput)
  • Stationary-batch decode-loop scheduling (+11-21%)
  • Argmax proposals (+11-15%)
  • Paged-only attention kernel (+10-14%)

Full system achieves 2.1-2.5x throughput over naive baseline.

Lossless R-ISD

Residual ISD (R-ISD) adds a gated LoRA adapter for bit-for-bit lossless acceleration:

  • LoRA active only at MASK positions; verify positions use base-only weights
  • Output is identical to the base AR model by construction
  • LoRA rank=128, overhead factor ~1.12x
  • Model: yifanyu/I-DLM-8B-lora-r128

Model Zoo

ModelBaseDescription
I-DLM-8BQwen3-8BMain model, matches AR quality
I-DLM-32BQwen3-32BLarge scale, outperforms LLaDA-2.1-flash (100B)
I-DLM-8B-LoRAQwen3-8BGated LoRA (rank=128) for lossless R-ISD

All models use trust_remote_code=True (custom SDARForCausalLM architecture).

Benchmarks

We evaluate on 15 benchmarks across 4 categories with thinking mode enabled:

  • Knowledge: ARC-C, MMLU, MMLU-Pro, GPQA-D, GPQA
  • Math: GSM8K, MATH-500, MathBench, AIME-24, AIME-25
  • Code: HumanEval, MBPP, LiveCodeBench-v6
  • Instruction: IFEval

See inference/eval/ for reproduction scripts.


Citation

@article{yu2026introspective,
  title={Introspective Diffusion Language Models},
  author={Yu, Yifan and Jian, Yuqing and Wang, Junxiong and Zhou, Zhongzhu
          and Zhuang, Donglin and Fang, Xinyu and Yanamandra, Sri
          and Wu, Xiaoxia and Wu, Qingyang and Song, Shuaiwen Leon
          and Dao, Tri and Athiwaratkun, Ben and Zou, James
          and Lai, Fan and Xu, Chenfeng},
  journal={arXiv preprint arXiv:7471639},
  year={2026}
}
© 2025 I-DLM Team. Built with care.

Source: hackernews

arrow_back Back to News