A comprehensive list of papers about Large-Language-Diffusion-Models.
Important
Contributions welcome:
-
If you have a relevant paper not included in the library, please contact us! Or, you may also consider submitting 'Pull requests' directly, thank you!
-
If you think your paper is more suitable for another category, please contact us or submit 'Pull requests'.
-
If your paper is accepted, you may consider updating the relevant information.
-
Thank you!
- 🔥🔥🔥 Awesome-LLDM is now open!
- Gemini Diffusion
- Dream-7B
- DreamOn
- What are Diffusion Language Models?
- Generative Modeling by Estimating Gradients of the Data Distribution
Paper Title | Year | Conference/Journal | Remark |
---|---|---|---|
Discrete Diffusion in Large Language and Multimodal Models: A Survey | 2025 | Arxiv | |
Diffusion-based Large Language Models Survey | 2025 | Arxiv | |
A Survey on Parallel Text Generation: From Parallel Decoding to Diffusion Language Models | 2025 | Arxiv |
Paper Title | Year | Conference/Journal | Remark |
---|---|---|---|
David helps Goliath: Inference-Time Collaboration Between Small Specialized and Large General Diffusion LMs | 2023 | NAACL | |
Diffusion Language Models Can Perform Many Tasks with Scaling and Instruction-Finetuning | 2023 | Arxiv | |
TESS 2: A Large-Scale Generalist Diffusion Language Model | 2025 | ACL | Adapted from Mistral-7B-v0.1 |
Scaling Diffusion Language Models via Adaptation from Autoregressive Models | 2025 | ICLR | 127M~7B (GPT2, LLaMA2) |
Large Language Diffusion Models | 2025 | Arxiv | LLaDA-8B |
LLaDA 1.5: Variance-Reduced Preference Optimization for Large Language Diffusion Models | 2025 | Arxiv | |
Large Language Models to Diffusion Finetuning | 2025 | Arxiv | |
LongLLaDA: Unlocking Long Context Capabilities in Diffusion LLMs | 2025 | Arxiv | Long context scaling |
Dream 7B: Diffusion Large Language Models | 2025 | Arxiv |
Paper Title | Year | Conference/Journal | Remark |
---|---|---|---|
Reinforcing the Diffusion Chain of Lateral Thought with Diffusion Language Models | 2025 | Arxiv | |
d1: Scaling Reasoning in Diffusion Large Language Models via Reinforcement Learning | 2025 | Arxiv | |
Diffusion of Thought: Chain-of-Thought Reasoning in Diffusion Language Models | 2024 | NeurIPS | |
wd1: Weighted Policy Optimization for Reasoning in Diffusion Language Models | 2025 | Arxiv | |
Thinking Inside the Mask: In-Place Prompting in Diffusion LLMs | 2025 | Arxiv | |
Review, Remask, Refine (R3): Process-Guided Block Diffusion for Text Generation | 2025 | ICML |
Paper Title | Year | Conference/Journal | Remark |
---|---|---|---|
DINGO: Constrained Inference for Diffusion LLMs | 2025 | Arxiv | Constrained decoding |
DiffuCoder: Understanding and Improving Masked Diffusion Models for Code Generation | 2025 | Arxiv | Coder |
Seed Diffusion: A Large-Scale Diffusion Language Model with High-Speed Inference | 2025 | Arxiv | Coder |
Time Is a Feature: Exploiting Temporal Dynamics in Diffusion Language Models | 2025 | Arxiv | |
The Devil behind the mask: An emergent safety vulnerability of Diffusion LLMs | 2025 | Arxiv | |
Discrete Diffusion VLA: Bringing Discrete Diffusion to Action Decoding in Vision-Language-Action Policies | 2025 | Arxiv | VLA |
LLaDA-VLA: Vision Language Diffusion Action Models | 2025 | Arxiv | VLA |
Paper Title | Year | Conference/Journal | Remark |
---|---|---|---|
Diffuse Everything: Multimodal Diffusion Models on Arbitrary State Spaces | 2025 | ICML | |
MMaDA: Multimodal Large Diffusion Language Models | 2025 | Arxiv | |
LLaDA-V: Large Language Diffusion Models with Visual Instruction Tuning | 2025 | Arxiv | |
Unified Multimodal Discrete Diffusion | 2025 | Arxiv | |
Dimple: Discrete Diffusion Multimodal Large Language Model with Parallel Decoding | 2025 | Arxiv | |
LaViDa: A Large Diffusion Language Model for Multimodal Understanding | 2025 | Arxiv | |
Dual Diffusion for Unified Image Generation and Understanding | 2025 | Arxiv | |
Muddit: Liberating Generation Beyond Text-to-Image with a Unified Discrete Diffusion Model | 2025 | Arxiv | |
Show-o2: Improved Native Unified Multimodal Models | 2025 | Arxiv |
We welcome all researchers to contribute to this repository.
If you have a related paper that was not added to the library, please contact us.
Email: jake630@snu.ac.kr / wjk9904@snu.ac.kr