MMaDA - Open-Sourced Multimodal Large Diffusion Language Models (dLLMs with block diffusion, mixed-CoT, unified RL)
-
Updated
Feb 14, 2026 - Python
MMaDA - Open-Sourced Multimodal Large Diffusion Language Models (dLLMs with block diffusion, mixed-CoT, unified RL)
Lumina-DiMOO - An Open-Sourced Multi-Modal Large Diffusion Language Model
NextFlow🚀: Unified Sequential Modeling Activates Multimodal Understanding and Generation
Official repository for "Visual Generation Unlocks Human-Like Reasoning through Multimodal World Models", https://arxiv.org/abs/2601.19834
🐧 Unify-Agent: An end-to-end unified multimodal agent for faithful, knowledge-grounded image generation.
UniEval: Unified Holistic Evaluation for Unified Multimodal Understanding and Generation
🌟 Generate and understand multimodal content seamlessly with Lumina-DiMOO, an advanced large language model designed for innovative applications.
Add a description, image, and links to the unified-multimodal-understanding-and-generation topic page so that developers can more easily learn about it.
To associate your repository with the unified-multimodal-understanding-and-generation topic, visit your repo's landing page and select "manage topics."