Improve model card: Add tags, links, abstract, and usage

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +69 -3
README.md CHANGED
@@ -1,3 +1,69 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ pipeline_tag: robotics
4
+ library_name: jax
5
+ ---
6
+
7
+ # AimBot: A Simple Auxiliary Visual Cue to Enhance Spatial Awareness of Visuomotor Policies
8
+
9
+ This repository contains model checkpoints and code for **AimBot**, a lightweight visual augmentation technique designed to improve visuomotor policy learning in robotic manipulation.
10
+
11
+ ## Paper Abstract
12
+
13
+ In this paper, we propose AimBot, a lightweight visual augmentation technique that provides explicit spatial cues to improve visuomotor policy learning in robotic manipulation. AimBot overlays shooting lines and scope reticles onto multi-view RGB images, offering auxiliary visual guidance that encodes the end-effector's state. The overlays are computed from depth images, camera extrinsics, and the current end-effector pose, explicitly conveying spatial relationships between the gripper and objects in the scene. AimBot incurs minimal computational overhead (less than 1 ms) and requires no changes to model architectures, as it simply replaces original RGB images with augmented counterparts. Despite its simplicity, our results show that AimBot consistently improves the performance of various visuomotor policies in both simulation and real-world settings, highlighting the benefits of spatially grounded visual feedback.
14
+
15
+ ## Links
16
+
17
+ * **Paper:** [AimBot: A Simple Auxiliary Visual Cue to Enhance Spatial Awareness of Visuomotor Policies](https://huggingface.co/papers/2508.08113)
18
+ * **Project Page:** [https://aimbot-reticle.github.io/](https://aimbot-reticle.github.io/)
19
+ * **Code Repository:** [https://github.com/aimbot-reticle/AimBot-Pi0](https://github.com/aimbot-reticle/AimBot-Pi0)
20
+
21
+ ## Usage
22
+
23
+ You can run inference with the pre-trained AimBot models using the `openpi` package.
24
+
25
+ First, ensure you have the necessary environment setup and dependencies installed as described in the [official GitHub repository's installation guide](https://github.com/aimbot-reticle/AimBot-Pi0#install).
26
+
27
+ Here's an example of how to load a trained policy and run inference:
28
+
29
+ ```python
30
+ from openpi.training import config
31
+ from openpi.policies import policy_config
32
+ from openpi.shared import download
33
+
34
+ # Example: Load the pi0_fast_droid model.
35
+ # Adjust 'pi0_fast_droid' and checkpoint_dir to your specific model/checkpoint.
36
+ config = config.get_config("pi0_fast_droid")
37
+ checkpoint_dir = download.maybe_download("s3://openpi-assets/checkpoints/pi0_fast_droid")
38
+
39
+ # Create a trained policy.
40
+ policy = policy_config.create_trained_policy(config, checkpoint_dir)
41
+
42
+ # Prepare your observation data.
43
+ # Replace '...' with actual image tensors (e.g., from PIL.Image or numpy arrays)
44
+ # and other required observation keys (e.g., proprioception, language prompts).
45
+ example = {
46
+ "observation/exterior_image_1_left": ..., # Placeholder for image data
47
+ "observation/wrist_image_left": ..., # Placeholder for image data
48
+ # ... other observation keys as required by the model, e.g., "prompt": "pick up the fork"
49
+ }
50
+
51
+ # Run inference to get action predictions.
52
+ action_chunk = policy.infer(example)["actions"]
53
+ print("Inferred actions:", action_chunk)
54
+ ```
55
+
56
+ For more detailed instructions on installation, running simulation/real-world experiments, and fine-tuning, please refer to the [official GitHub repository](https://github.com/aimbot-reticle/AimBot-Pi0).
57
+
58
+ ## Citation
59
+
60
+ If you find this work helpful, please consider citing our paper:
61
+
62
+ ```bibtex
63
+ @article{aimbot,
64
+ title={AimBot: A Simple Auxiliary Visual Cue to Enhance Spatial Awareness of Visuomotor Policies},
65
+ author={Dai, Yinpei and Lee, Jayjun and et al},
66
+ journal={CoRL},
67
+ year={2025},
68
+ }
69
+ ```