File size: 1,383 Bytes
12cfa0f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
---
base_model: THUDM/CogVideoX-2b
library_name: peft
tags:
- lora
- dora
- cogvideox
- physics
- video-generation
- warp
---

# PDW — Physics-Corrected CogVideoX-2b World Model (DoRA Adapter)

A **DoRA (Weight-Decomposed Low-Rank Adaptation)** adapter for [CogVideoX-2b](https://huggingface.co/THUDM/CogVideoX-2b), fine-tuned to generate physically accurate videos using **NVIDIA Warp** physics simulation data and **TRD (Temporal Representation Distillation)** with DINOv2-large as teacher.

## Model Details

- **Base model:** THUDM/CogVideoX-2b (1.7B params)
- **Adapter:** DoRA (r=16, lora_alpha=32, use_dora=True)
- **Target modules:** `to_q`, `to_k`, `to_v`, `to_out.0`
- **Trainable params:** 7.6M / 1.7B (0.45%)
- **Physics engine:** NVIDIA Warp (28-scenario 7×4 grid)
- **TRD teacher:** DINOv2-large
- **Hardware:** NVIDIA H100 NVL
- **Training steps:** 400

## Evaluation Results

| Metric | Delta |
|---|---|
| Diffusion MSE | +94.1% |
| Motion score | +1.7% |
| Overall | +47.9% |

## How to Use

```python
from peft import PeftModel
from diffusers import CogVideoXTransformer3DModel

# Load base transformer
base_transformer = CogVideoXTransformer3DModel.from_pretrained(
    "THUDM/CogVideoX-2b", subfolder="transformer"
)

# Load DoRA adapter
model = PeftModel.from_pretrained(base_transformer, "athul020/pdw_final_dora")
```

### Framework versions

- PEFT 0.18.1