Improve model card: Add tags, links, abstract, and usage
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,3 +1,69 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: mit
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mit
|
| 3 |
+
pipeline_tag: robotics
|
| 4 |
+
library_name: jax
|
| 5 |
+
---
|
| 6 |
+
|
| 7 |
+
# AimBot: A Simple Auxiliary Visual Cue to Enhance Spatial Awareness of Visuomotor Policies
|
| 8 |
+
|
| 9 |
+
This repository contains model checkpoints and code for **AimBot**, a lightweight visual augmentation technique designed to improve visuomotor policy learning in robotic manipulation.
|
| 10 |
+
|
| 11 |
+
## Paper Abstract
|
| 12 |
+
|
| 13 |
+
In this paper, we propose AimBot, a lightweight visual augmentation technique that provides explicit spatial cues to improve visuomotor policy learning in robotic manipulation. AimBot overlays shooting lines and scope reticles onto multi-view RGB images, offering auxiliary visual guidance that encodes the end-effector's state. The overlays are computed from depth images, camera extrinsics, and the current end-effector pose, explicitly conveying spatial relationships between the gripper and objects in the scene. AimBot incurs minimal computational overhead (less than 1 ms) and requires no changes to model architectures, as it simply replaces original RGB images with augmented counterparts. Despite its simplicity, our results show that AimBot consistently improves the performance of various visuomotor policies in both simulation and real-world settings, highlighting the benefits of spatially grounded visual feedback.
|
| 14 |
+
|
| 15 |
+
## Links
|
| 16 |
+
|
| 17 |
+
* **Paper:** [AimBot: A Simple Auxiliary Visual Cue to Enhance Spatial Awareness of Visuomotor Policies](https://huggingface.co/papers/2508.08113)
|
| 18 |
+
* **Project Page:** [https://aimbot-reticle.github.io/](https://aimbot-reticle.github.io/)
|
| 19 |
+
* **Code Repository:** [https://github.com/aimbot-reticle/AimBot-Pi0](https://github.com/aimbot-reticle/AimBot-Pi0)
|
| 20 |
+
|
| 21 |
+
## Usage
|
| 22 |
+
|
| 23 |
+
You can run inference with the pre-trained AimBot models using the `openpi` package.
|
| 24 |
+
|
| 25 |
+
First, ensure you have the necessary environment setup and dependencies installed as described in the [official GitHub repository's installation guide](https://github.com/aimbot-reticle/AimBot-Pi0#install).
|
| 26 |
+
|
| 27 |
+
Here's an example of how to load a trained policy and run inference:
|
| 28 |
+
|
| 29 |
+
```python
|
| 30 |
+
from openpi.training import config
|
| 31 |
+
from openpi.policies import policy_config
|
| 32 |
+
from openpi.shared import download
|
| 33 |
+
|
| 34 |
+
# Example: Load the pi0_fast_droid model.
|
| 35 |
+
# Adjust 'pi0_fast_droid' and checkpoint_dir to your specific model/checkpoint.
|
| 36 |
+
config = config.get_config("pi0_fast_droid")
|
| 37 |
+
checkpoint_dir = download.maybe_download("s3://openpi-assets/checkpoints/pi0_fast_droid")
|
| 38 |
+
|
| 39 |
+
# Create a trained policy.
|
| 40 |
+
policy = policy_config.create_trained_policy(config, checkpoint_dir)
|
| 41 |
+
|
| 42 |
+
# Prepare your observation data.
|
| 43 |
+
# Replace '...' with actual image tensors (e.g., from PIL.Image or numpy arrays)
|
| 44 |
+
# and other required observation keys (e.g., proprioception, language prompts).
|
| 45 |
+
example = {
|
| 46 |
+
"observation/exterior_image_1_left": ..., # Placeholder for image data
|
| 47 |
+
"observation/wrist_image_left": ..., # Placeholder for image data
|
| 48 |
+
# ... other observation keys as required by the model, e.g., "prompt": "pick up the fork"
|
| 49 |
+
}
|
| 50 |
+
|
| 51 |
+
# Run inference to get action predictions.
|
| 52 |
+
action_chunk = policy.infer(example)["actions"]
|
| 53 |
+
print("Inferred actions:", action_chunk)
|
| 54 |
+
```
|
| 55 |
+
|
| 56 |
+
For more detailed instructions on installation, running simulation/real-world experiments, and fine-tuning, please refer to the [official GitHub repository](https://github.com/aimbot-reticle/AimBot-Pi0).
|
| 57 |
+
|
| 58 |
+
## Citation
|
| 59 |
+
|
| 60 |
+
If you find this work helpful, please consider citing our paper:
|
| 61 |
+
|
| 62 |
+
```bibtex
|
| 63 |
+
@article{aimbot,
|
| 64 |
+
title={AimBot: A Simple Auxiliary Visual Cue to Enhance Spatial Awareness of Visuomotor Policies},
|
| 65 |
+
author={Dai, Yinpei and Lee, Jayjun and et al},
|
| 66 |
+
journal={CoRL},
|
| 67 |
+
year={2025},
|
| 68 |
+
}
|
| 69 |
+
```
|