Update README.md
Browse files
README.md
CHANGED
|
@@ -1,5 +1,8 @@
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
|
|
|
|
|
|
|
|
|
| 3 |
---
|
| 4 |
|
| 5 |

|
|
@@ -34,11 +37,6 @@ This repository contains the code for the MAGI-1 model, pre-trained weights and
|
|
| 34 |
|
| 35 |
We present MAGI-1, a world model that generates videos by ***autoregressively*** predicting a sequence of video chunks, defined as fixed-length segments of consecutive frames. Trained to denoise per-chunk noise that increases monotonically over time, MAGI-1 enables causal temporal modeling and naturally supports streaming generation. It achieves strong performance on image-to-video (I2V) tasks conditioned on text instructions, providing high temporal consistency and scalability, which are made possible by several algorithmic innovations and a dedicated infrastructure stack. MAGI-1 further supports controllable generation via chunk-wise prompting, enabling smooth scene transitions, long-horizon synthesis, and fine-grained text-driven control. We believe MAGI-1 offers a promising direction for unifying high-fidelity video generation with flexible instruction control and real-time deployment.
|
| 36 |
|
| 37 |
-
<div align="center">
|
| 38 |
-
<video src="https://github.com/user-attachments/assets/5cfa90e0-f6ed-476b-a194-71f1d309903a
|
| 39 |
-
" width="70%" poster=""> </video>
|
| 40 |
-
</div>
|
| 41 |
-
|
| 42 |
|
| 43 |
## 2. Model Summary
|
| 44 |
|
|
@@ -220,4 +218,4 @@ If you find our code or model useful in your research, please cite:
|
|
| 220 |
|
| 221 |
## 8. Contact
|
| 222 |
|
| 223 |
-
If you have any questions, please feel free to raise an issue or contact us at [support@sand.ai](support@sand.ai) .
|
|
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
+
language:
|
| 4 |
+
- en
|
| 5 |
+
pipeline_tag: image-to-video
|
| 6 |
---
|
| 7 |
|
| 8 |

|
|
|
|
| 37 |
|
| 38 |
We present MAGI-1, a world model that generates videos by ***autoregressively*** predicting a sequence of video chunks, defined as fixed-length segments of consecutive frames. Trained to denoise per-chunk noise that increases monotonically over time, MAGI-1 enables causal temporal modeling and naturally supports streaming generation. It achieves strong performance on image-to-video (I2V) tasks conditioned on text instructions, providing high temporal consistency and scalability, which are made possible by several algorithmic innovations and a dedicated infrastructure stack. MAGI-1 further supports controllable generation via chunk-wise prompting, enabling smooth scene transitions, long-horizon synthesis, and fine-grained text-driven control. We believe MAGI-1 offers a promising direction for unifying high-fidelity video generation with flexible instruction control and real-time deployment.
|
| 39 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 40 |
|
| 41 |
## 2. Model Summary
|
| 42 |
|
|
|
|
| 218 |
|
| 219 |
## 8. Contact
|
| 220 |
|
| 221 |
+
If you have any questions, please feel free to raise an issue or contact us at [support@sand.ai](support@sand.ai) .
|