Fred Zhangzhi Peng, Shuibai Zhang, Alex Tong
Duke University, UW-Madison, Aithyra
🤗 Hugging Face | 💻 Codebase | 📖 Blog
🌐 English Blog | 🇨🇳中文博客
TL;DR: the first/most open release diffusion large language model.
The field of diffusion LLMs is still young, and the answers aren’t clear. Pioneering efforts like Gemini-Diffusion, Seed Diffusion, and Mercury sparked excitement, but they remain closed APIs. You can use them, but you can’t study how they were built.
Open projects like LLaDA and Dream pushed things further by releasing weights and inference code. But they stopped short of what researchers need most: training pipelines, data recipes, and reproducible evaluation.
That’s why we built Open-dLLM: the first full-stack open diffusion language model project.
👉 Our first release is Open-dCoder, focused on code generation. It includes:
With Open-dLLM, you can go from raw data → training → checkpoints → evaluation → inference, all in one repo.
Project | Data | Training Code | Inference | Evaluation | Weights |
---|---|---|---|---|---|
Open-dLLM (ours) | ✅ | ✅ | ✅ | ✅ | ✅ |
LLaDA | ❌ | ❌ | ✅ | ⚠️ limited | ✅ |
Dream | ❌ | ❌ | ✅ | ⚠️ limited | ✅ |
Gemini-Diffusion | ❌ | ❌ | ❌ | ❌ | ❌ (API only) |
Seed Diffusion | ❌ | ❌ | ❌ | ❌ | ❌ (API only) |
Mercury | ❌ | ❌ | ❌ | ❌ | ❌ (API only |
Here’s our Open-dLLM generating a QuickSort algorithm from scratch:
Youtube Video (Please play it, I rly want u to enjoy the music : )
Youtube Video (Please play it, I rly want u to enjoy the music : )