Skip to content

[M] Lab — Masked Emergent AGI Lab (2025 - )

[M] Lab (Masked Emergent AGI Lab) (Originally MeissonFlow Research) is a research group exploring the next generation of general-purpose AI systems built on the principle of masking. Our mission is to advance efficient, safe, and unified modeling through masked generative and discriminative learning.

🔬 Research Focus

  • 🧠 Masked modeling in vision, language, and multimodality
  • 🌀 Discrete diffusion and non-autoregressive generation
  • 🧩 Unified architectures for scalable generalization
  • 🧭 Mask-based pathways to AGI-aligned reasoning

We believe masking is a fundamental abstraction toward controllable, efficient, and generalizable intelligence.


MeissonFlow Research (2024 - )

MeissonFlow Research is a non-commercial research group dedicated to advancing generative modeling techniques for structured visual and multimodal content creation. We aim to design models and algorithms that help creators produce high-quality content with greater efficiency and control.

Our journey began with MaskGIT, a pioneering work by Huiwen Chang, which introduced a bidirectional transformer decoder for image synthesis and outperformed traditional raster-scan autoregressive (AR) generation. This paradigm was later extended to text-to-image synthesis in MUSE.

Building upon these foundations, we scaled masked generative modeling with the latest architectural designs and sampling strategies, culminating in Monetico and Meissonic built from scratch, which are on par with leading diffusion models such as SDXL while maintaining greater efficiency.

Having verified the effectiveness of this approach, we began to ask a deeper question, one that reaches beyond performance benchmarks: what foundations are required for general-purpose generative intelligence?
Through discussions with researchers at Safe Superintelligence (SSI) Club, University of Illinois Urbana-Champaign (UIUC) and Riot Video Games, we converged on the vision of a visual-centric world model: a generative and interactive system capable of simulating, interacting with, and reasoning about multimodal environments.

We believe that masking is a fundamental abstraction for building such controllable, efficient, and generalizable intelligence.

A similar vision was shared by Stefano Ermon at ICLR 2025, where he described Diffusion as a unified paradigm for a multi-modal world model, a message that echoes and strengthens our belief: that unified generative modeling is the path toward general-purpose superintelligence.

To pursue this vision, we introduced Muddit and Muddit Plus, unified generative models built upon visual priors (Meissonic), and capable of generation across text and image within a single architecture and paradigm.

We want to build the world with visual prior, though we sadly agree that the language prior dominates current unified models. Inspired by the success of Mercury by Inception Labs, we developed Lumina-DiMOO. As a larger scale unified masked diffusion model than Muddit, Lumina-DiMOO achieves state-of-the-art performance among discrete diffusion models to date; and we are still pushing it further! It integrates high-resolution image generation with multimodal capabilities, including text-to-image, image-to-image, and image understanding.

To further clarify our long-term roadmap, we articulated our perspective in From Masks to Worlds: A Hitchhiker’s Guide to World Models, which traces a five-stage roadmap from early masked modeling to unified generative modeling and the future we are building.

We look forward to releasing more models and algorithms in this direction. We post related and family papers here. We thank our amazing teammates and you, the reader, for your interest in our work.

Special thanks to Style2Paints Research, which helped shape our taste and research direction in the early days.

Pinned Loading

  1. Awesome-World-Models Awesome-World-Models Public

    Official Repo of From Masks to Worlds: A Hitchhiker’s Guide to World Models.

    62

  2. Muddit Muddit Public

    Official Implementation of Muddit [Meissonic II]: Liberating Generation Beyond Text-to-Image with a Unified Discrete Diffusion Model.

    Python 95 1

Repositories

Showing 4 of 4 repositories

Top languages

Loading…

Most used topics

Loading…