Update README.md
Browse files
README.md
CHANGED
@@ -1,5 +1,8 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
|
|
|
|
|
|
3 |
---
|
4 |
|
5 |

|
@@ -34,11 +37,6 @@ This repository contains the code for the MAGI-1 model, pre-trained weights and
|
|
34 |
|
35 |
We present MAGI-1, a world model that generates videos by ***autoregressively*** predicting a sequence of video chunks, defined as fixed-length segments of consecutive frames. Trained to denoise per-chunk noise that increases monotonically over time, MAGI-1 enables causal temporal modeling and naturally supports streaming generation. It achieves strong performance on image-to-video (I2V) tasks conditioned on text instructions, providing high temporal consistency and scalability, which are made possible by several algorithmic innovations and a dedicated infrastructure stack. MAGI-1 further supports controllable generation via chunk-wise prompting, enabling smooth scene transitions, long-horizon synthesis, and fine-grained text-driven control. We believe MAGI-1 offers a promising direction for unifying high-fidelity video generation with flexible instruction control and real-time deployment.
|
36 |
|
37 |
-
<div align="center">
|
38 |
-
<video src="https://github.com/user-attachments/assets/5cfa90e0-f6ed-476b-a194-71f1d309903a
|
39 |
-
" width="70%" poster=""> </video>
|
40 |
-
</div>
|
41 |
-
|
42 |
|
43 |
## 2. Model Summary
|
44 |
|
@@ -220,4 +218,4 @@ If you find our code or model useful in your research, please cite:
|
|
220 |
|
221 |
## 8. Contact
|
222 |
|
223 |
-
If you have any questions, please feel free to raise an issue or contact us at [[email protected]]([email protected]) .
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
pipeline_tag: image-to-video
|
6 |
---
|
7 |
|
8 |

|
|
|
37 |
|
38 |
We present MAGI-1, a world model that generates videos by ***autoregressively*** predicting a sequence of video chunks, defined as fixed-length segments of consecutive frames. Trained to denoise per-chunk noise that increases monotonically over time, MAGI-1 enables causal temporal modeling and naturally supports streaming generation. It achieves strong performance on image-to-video (I2V) tasks conditioned on text instructions, providing high temporal consistency and scalability, which are made possible by several algorithmic innovations and a dedicated infrastructure stack. MAGI-1 further supports controllable generation via chunk-wise prompting, enabling smooth scene transitions, long-horizon synthesis, and fine-grained text-driven control. We believe MAGI-1 offers a promising direction for unifying high-fidelity video generation with flexible instruction control and real-time deployment.
|
39 |
|
|
|
|
|
|
|
|
|
|
|
40 |
|
41 |
## 2. Model Summary
|
42 |
|
|
|
218 |
|
219 |
## 8. Contact
|
220 |
|
221 |
+
If you have any questions, please feel free to raise an issue or contact us at [[email protected]]([email protected]) .
|