| # sfp4_v4_sparse09_hpo_on_ours_p_init2050 checkpoint-700 |
| |
| This upload contains the consolidated WanTransformer3DModel transformer weights |
| from: |
| |
| `checkpoints/sfp4_v4_sparse09_hpo_on_ours_p_init2050_1n_interactive/checkpoint-700` |
|
|
| Contents: |
|
|
| - `transformer/config.json` |
| - `transformer/diffusion_pytorch_model.safetensors` |
| - `backend_snapshot/` |
| - `standalone_inference/` |
|
|
| Training run: |
|
|
| - run name: `sfp4_v4_sparse09_hpo_on_ours_p_init2050_1n_interactive` |
| - source init: `sfp4_v4_sparse06_hpo_on_ours_p_1n_interactive_v2 checkpoint-2050` |
| - attention backend: `SPARSE_FP4_OURS_P_ATTN` |
| - high precision output for backward: enabled |
| - VSA sparsity: `0.9` |
|
|
| This package does not include the distributed optimizer/training-state |
| checkpoint. Use the original `distributed_checkpoint/` directory if exact |
| training resume state is required. |
|
|
| `backend_snapshot/` contains the local FastVideo backend code used by this |
| checkpoint, including `SPARSE_FP4_OURS_P_ATTN`, its Triton forward/backward |
| kernel, FP4 quant helpers, VSA metadata helper, backend wiring, and the exact |
| SFT launch scripts. |
|
|
| It also includes the inference entrypoint snapshot and an example script: |
|
|
| - `backend_snapshot/scripts/inference/run_sfp4_ours_p_checkpoint_700.sh` |
| - `backend_snapshot/training_attention_settings.json` |
| - `standalone_inference/` |
|
|
| Attention setup for this checkpoint: |
|
|
| - self-attention: `SPARSE_FP4_OURS_P_ATTN`, FP4 Q/K/V, sparse 64-token VSA |
| tiles, group-local P quant, dropped-tile mean compensation |
| - cross-attention: dense SDPA fallback, not FP4/sparse |
| - force-dense paths: dense SDPA |
|
|
| `standalone_inference/` is a portable helper for normal inference. It contains |
| an overlay installer, a runner that downloads/loads the checkpoint-700 |
| transformer weights, and the sparse FP4 backend files required by this |
| checkpoint. |
|
|