pdw_final_dora / README.md
athul020's picture
Upload folder using huggingface_hub
12cfa0f verified
---
base_model: THUDM/CogVideoX-2b
library_name: peft
tags:
- lora
- dora
- cogvideox
- physics
- video-generation
- warp
---
# PDW — Physics-Corrected CogVideoX-2b World Model (DoRA Adapter)
A **DoRA (Weight-Decomposed Low-Rank Adaptation)** adapter for [CogVideoX-2b](https://huggingface.co/THUDM/CogVideoX-2b), fine-tuned to generate physically accurate videos using **NVIDIA Warp** physics simulation data and **TRD (Temporal Representation Distillation)** with DINOv2-large as teacher.
## Model Details
- **Base model:** THUDM/CogVideoX-2b (1.7B params)
- **Adapter:** DoRA (r=16, lora_alpha=32, use_dora=True)
- **Target modules:** `to_q`, `to_k`, `to_v`, `to_out.0`
- **Trainable params:** 7.6M / 1.7B (0.45%)
- **Physics engine:** NVIDIA Warp (28-scenario 7×4 grid)
- **TRD teacher:** DINOv2-large
- **Hardware:** NVIDIA H100 NVL
- **Training steps:** 400
## Evaluation Results
| Metric | Delta |
|---|---|
| Diffusion MSE | +94.1% |
| Motion score | +1.7% |
| Overall | +47.9% |
## How to Use
```python
from peft import PeftModel
from diffusers import CogVideoXTransformer3DModel
# Load base transformer
base_transformer = CogVideoXTransformer3DModel.from_pretrained(
"THUDM/CogVideoX-2b", subfolder="transformer"
)
# Load DoRA adapter
model = PeftModel.from_pretrained(base_transformer, "athul020/pdw_final_dora")
```
### Framework versions
- PEFT 0.18.1