Spaces:
Running
Running
Update README.md
Browse files
README.md
CHANGED
|
@@ -24,13 +24,13 @@ Through discussions with researchers at Safe Superintelligence (SSI) Club, Unive
|
|
| 24 |
|
| 25 |
A similar vision was shared by [**Stefano Ermon**](https://cs.stanford.edu/~ermon/) at ICLR 2025, where he described *Diffusion as a unified paradigm for a multi-modal world model*, a message that echoes and strengthens our belief: that unified generative modeling is the path toward general-purpose superintelligence.
|
| 26 |
|
| 27 |
-
To pursue this vision, we introduced [**Muddit** and **Muddit Plus**](https://github.com/M-E-AGI-Lab/Muddit), unified generative models built upon visual priors (Meissonic), and capable of
|
| 28 |
|
| 29 |
We want to build the world with visual prior, though we sadly agree that the language prior dominates current unified models.
|
| 30 |
Inspired by the success of Mercury by [**Inception Labs**](https://www.inceptionlabs.ai/),
|
| 31 |
-
we developed [**Lumina-DiMOO**](https://arxiv.org/abs/2510.06308).
|
| 32 |
|
| 33 |
-
To further clarify our roadmap, we articulated our perspective in [**From Masks to Worlds: A Hitchhiker’s Guide to World Models**](https://arxiv.org/abs/2510.20668), which traces a five-stage roadmap from early masked modeling to unified generative modeling and the future we are building.
|
| 34 |
|
| 35 |
We look forward to releasing more models and algorithms in this direction. We post related and family papers [here](https://github.com/viiika/Meissonic).
|
| 36 |
We thank our amazing teammates and you, the reader, for your interest in our work.
|
|
|
|
| 24 |
|
| 25 |
A similar vision was shared by [**Stefano Ermon**](https://cs.stanford.edu/~ermon/) at ICLR 2025, where he described *Diffusion as a unified paradigm for a multi-modal world model*, a message that echoes and strengthens our belief: that unified generative modeling is the path toward general-purpose superintelligence.
|
| 26 |
|
| 27 |
+
To pursue this vision, we introduced [**Muddit** and **Muddit Plus**](https://github.com/M-E-AGI-Lab/Muddit), unified generative models built upon visual priors (Meissonic), and capable of generation across text and image within a single architecture and paradigm.
|
| 28 |
|
| 29 |
We want to build the world with visual prior, though we sadly agree that the language prior dominates current unified models.
|
| 30 |
Inspired by the success of Mercury by [**Inception Labs**](https://www.inceptionlabs.ai/),
|
| 31 |
+
we developed [**Lumina-DiMOO**](https://arxiv.org/abs/2510.06308). As a larger scale unified masked diffusion model than Muddit, Lumina-DiMOO achieves state-of-the-art performance among discrete diffusion models to date; and we are still pushing it further! It integrates high-resolution image generation with multimodal capabilities, including text-to-image, image-to-image, and image understanding.
|
| 32 |
|
| 33 |
+
To further clarify our long-term roadmap, we articulated our perspective in [**From Masks to Worlds: A Hitchhiker’s Guide to World Models**](https://arxiv.org/abs/2510.20668), which traces a five-stage roadmap from early masked modeling to unified generative modeling and the future we are building.
|
| 34 |
|
| 35 |
We look forward to releasing more models and algorithms in this direction. We post related and family papers [here](https://github.com/viiika/Meissonic).
|
| 36 |
We thank our amazing teammates and you, the reader, for your interest in our work.
|