anytop-checkpoints / README.md
AAAcon's picture
Add dataset card
2bb548b verified
metadata
license: mit
task_categories:
  - other
tags:
  - motion-generation
  - diffusion-models
  - cross-skeleton
  - retargeting
  - anonymous-submission
size_categories:
  - 1K<n<10K

anytop-checkpoints

Pre-trained checkpoints accompanying the anonymous NeurIPS 2026 submission "Why Cross-Skeleton Retargeting Is Non-Identifiable: Structural Limits of Generative Motion Models".

This repository contains a single tarball, anytop-checkpoints.tar.gz (6.5 GB compressed, 7.2 GB extracted), bundling the trained model weights for the 14-method comparison reported in the paper.

Contents

After extraction (tar -xzvf anytop-checkpoints.tar.gz):

anytop-checkpoints/
  README.md                       (per-method usage notes)
  anytop_v5/                      AnyTop transductive (model000175000.pt, 34 MB)
  ace_primary_70/                 ACE-T transductive (ckpt_final.pt, 83 MB)
  ace_inductive_60/               ACE-I inductive (ckpt_final.pt, 83 MB)
  moreflow_primary_70/            MoReFlow-T (ckpt_final.pt, 79 MB)
  moreflow_inductive_60/          MoReFlow-I (ckpt_final.pt, 79 MB)
  anchor_label_flow/              AL-Flow (ckpt_final.pt, 86 MB)
  anchor_label_flow_src/          AL-Flow-Src (ckpt_final.pt, 88 MB)
  anchor_label_flow_src_g/        AL-Flow-Src-G (ckpt_final.pt, 87 MB)
  dpg_sb_v3/                      DPG-SB-v3 (final.pt 62 MB + z_stats.pt)
  moreflow_vqvae/                 71 per-skeleton VQ-VAE encoders (~91 MB each)
                                  Required for ACE / MoReFlow / DPG-SB-v3 inference
  moreflow_caches/                Pre-extracted MoReFlow latents
                                  (only required to retrain DPG-SB-v3 from scratch)

Verification

SHA-256: fc15f536510014a7bb953aa62cd24358de6af11f0d92f0f8360e2ec44b8251e2
sha256sum -c <<< "fc15f536510014a7bb953aa62cd24358de6af11f0d92f0f8360e2ec44b8251e2  anytop-checkpoints.tar.gz"

How to use

The companion code repository contains training scripts, evaluation runners, and the SIF metric implementation. After downloading and extracting this tarball into the code repo's save/ directory, every per-method invocation in the code repo's REPRODUCE.md § 4 will work without retraining.

# From the code repo root:
mkdir -p save
tar -xzvf anytop-checkpoints.tar.gz
mv anytop-checkpoints/* save/
rmdir anytop-checkpoints

# Verify with one method:
python -m eval.baselines.run_anytop_v5 \
    --ckpt save/anytop_v5/model000175000.pt \
    --manifest eval/benchmark_v3/queries_sif/manifest.json \
    --output_dir results/anytop_sif

License

MIT, matching the upstream AnyTop release.

Anonymity

Every args.json file in the package has been swept for personal paths, author names, and W&B entity strings. Checkpoint binaries (*.pt) contain only PyTorch model and optimiser state.