MAGI: A New Framework for Autoregressive Video Generation Using Masked Modeling

Top post
Masked Autoregressive Video Generation: A New Approach with MAGI
The generation of videos using Artificial Intelligence (AI) is a rapidly growing field of research with diverse applications, from the entertainment industry to scientific research. A promising approach in this area is autoregressive video generation, where image frames are generated sequentially based on preceding frames. A new framework called MAGI (Masked Autoregressive Generation of Images) presents an innovative hybrid approach that combines masked modeling for intra-frame generation with causal modeling for next-frame generation.
Complete Teacher Forcing: A Key to Improved Video Generation
A central innovation of MAGI is the so-called "Complete Teacher Forcing" (CTF). Traditional approaches often use "Masked Teacher Forcing" (MTF), where masked frames are conditioned on similarly masked frames. CTF, on the other hand, conditions masked frames on fully observable frames. This difference allows for a smoother transition from token-based (patch-based) to frame-based autoregressive generation. The results show a significant improvement in performance: CTF surpasses MTF by an impressive 23% in terms of FVD scores (Fréchet Video Distance) when predicting videos based on the first frame. This advancement underscores the potential of CTF for generating high-quality videos.
Challenges and Solutions
As with many AI-based generation methods, autoregressive video generation also presents challenges, such as "Exposure Bias." This arises from the discrepancy between training and inference: During training, the models are fed with the actual preceding frames, while during inference, they must use their own, potentially erroneous predictions. MAGI addresses this problem through targeted training strategies that aim to increase the model's robustness against such discrepancies.
Impressive Results and Future Prospects
Experiments with MAGI show promising results. The framework is capable of generating long, coherent video sequences of over 100 frames, even when trained with only 16 frames. This ability to generate long sequences based on limited training material opens up new possibilities for scalable and high-quality video generation. The combination of masked and causal modeling, coupled with the innovative CTF method, positions MAGI as a significant contribution to the advancement of AI-powered video generation. Future research could focus on further optimizing training strategies and expanding the application possibilities of MAGI, for example, in the field of video editing or the creation of synthetic training data for other AI models.
Bibliography: Zhou, D., et al. "Taming Teacher Forcing for Masked Autoregressive Video Generation." arXiv preprint arXiv:2501.12389 (2025). https://arxiv.org/abs/2501.12389 https://arxiv.org/html/2501.12389v1 https://neurips.cc/virtual/2024/poster/94115 https://arxiv-sanity-lite.com/ https://openaccess.thecvf.com/content/CVPR2023/papers/Fu_Tell_Me_What_Happened_Unifying_Text-Guided_Video_Completion_via_Multimodal_CVPR_2023_paper.pdf https://www.chatpaper.com/chatpaper/fr?id=4&date=1737475200&page=1 https://github.com/showlab/Awesome-Video-Diffusion https://openreview.net/pdf?id=TFbwV6I0VLg https://www.ecva.net/papers/eccv_2024/papers_ECCV/papers/10270.pdf https://desaixie.github.io/pa-vdm/