--- datasets: lerobot/pusht library_name: lerobot license: apache-2.0 model_name: act pipeline_tag: robotics tags: - lerobot - robotics - act --- # Model Card for act ACT trained on PushT dataset for 80,000 training steps. Trained for 3.5 hrs on T4 GPU on Colab free. The goal is to use this to compare ACTs performance on PushT against diffusion policy particularly the aspect of action multimodality. Training logs: https://api.wandb.ai/links/ramachandranaadarsh-indian-institute-of-technology-madras/7m9dpcgw [Action Chunking with Transformers (ACT)](https://huggingface.co/papers/2304.13705) is an imitation-learning method that predicts short action chunks instead of single steps. It learns from teleoperated data and often achieves high success rates. This policy has been trained and pushed to the Hub using [LeRobot](https://github.com/huggingface/lerobot). See the full documentation at [LeRobot Docs](https://huggingface.co/docs/lerobot/index). --- ## How to Get Started with the Model For a complete walkthrough, see the [training guide](https://huggingface.co/docs/lerobot/il_robots#train-a-policy). Below is the short version on how to train and run inference/eval: ### Train from scratch ```bash lerobot-train \ --dataset.repo_id=${HF_USER}/ \ --policy.type=act \ --output_dir=outputs/train/ \ --job_name=lerobot_training \ --policy.device=cuda \ --policy.repo_id=${HF_USER}/ --wandb.enable=true ``` _Writes checkpoints to `outputs/train//checkpoints/`._ ### Evaluate the policy/run inference ```bash lerobot-record \ --robot.type=so100_follower \ --dataset.repo_id=/eval_ \ --policy.path=/ \ --episodes=10 ``` Prefix the dataset repo with **eval\_** and supply `--policy.path` pointing to a local or hub checkpoint. --- ## Model Details - **License:** apache-2.0