UniVidX: A Unified Framework for Multimodal Video Generation
- •UniVidX enables versatile video generation by unifying multiple tasks in a single multimodal framework
- •New architecture uses stochastic condition masking and decoupled gated LoRA for flexible synthesis
- •Model demonstrates high performance across diverse domains including RGB, intrinsic maps, and RGBA layers
The landscape of video generation is shifting rapidly, moving away from specialized models for every niche task and toward unified, flexible frameworks. Researchers have unveiled UniVidX, a sophisticated approach designed to consolidate disparate video generation tasks into a single, cohesive system. By leveraging Video Diffusion Models (VDMs), the authors have created a framework capable of handling complex inputs and outputs without the need for training separate, rigid models for each specific goal.
The primary challenge in previous attempts at unified generation was the fixed nature of input-output mappings, which prevented models from effectively learning the underlying correlations between different modalities. UniVidX overcomes this by reformulating pixel-aligned tasks as conditional generation within a shared space. This allows the system to treat disparate data types—such as standard RGB video, intrinsic maps for lighting, or separate RGBA layers—as conditional inputs that inform the final synthesis process.
Three architectural innovations anchor this framework. First, Stochastic Condition Masking (SCM) provides flexibility by dynamically partitioning modalities during training, moving beyond static mappings to enable omni-directional generation. Second, Decoupled Gated LoRA (DGL) applies specific, lightweight adaptations only when a modality serves as the target, effectively preserving the core strengths of the underlying VDM backbone. Finally, Cross-Modal Self-Attention (CMSA) allows the model to exchange critical information across modalities while maintaining their specific characteristics.
These combined techniques allow the model to remain versatile even when trained on smaller datasets, specifically achieving robust performance with fewer than 1,000 videos. By facilitating seamless information exchange, UniVidX demonstrates that it is possible to achieve state-of-the-art results across diverse domains like intrinsic map generation and layered video blending within a single, unified architecture. This represents a significant step toward more efficient, general-purpose generative video systems that can adapt to varying requirements without massive, redundant training pipelines.