Improve model card with metadata, links, and usage details

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +64 -4
README.md CHANGED
@@ -2,9 +2,69 @@
2
  tags:
3
  - model_hub_mixin
4
  - pytorch_model_hub_mixin
 
 
 
5
  ---
6
 
7
- This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
8
- - Code: [More Information Needed]
9
- - Paper: [More Information Needed]
10
- - Docs: [More Information Needed]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  tags:
3
  - model_hub_mixin
4
  - pytorch_model_hub_mixin
5
+ pipeline_tag: robotics
6
+ library_name: lerobot
7
+ license: unknown
8
  ---
9
 
10
+ # Look, Focus, Act: Efficient and Robust Robot Learning via Human Gaze and Foveated Vision Transformers
11
+
12
+ This model is part of the work presented in the paper [**Look, Focus, Act: Efficient and Robust Robot Learning via Human Gaze and Foveated Vision Transformers**](https://huggingface.co/papers/2507.15833).
13
+
14
+ <div align="center">
15
+ <img src="https://ian-chuang.github.io/gaze-av-aloha/media/hero.gif" alt="Look, Focus, Act Hero GIF" width="100%">
16
+ </div>
17
+
18
+ **Project Website:** [https://ian-chuang.github.io/gaze-av-aloha/](https://ian-chuang.github.io/gaze-av-aloha/)
19
+ **Code (GitHub):** [https://github.com/ian-chuang/gaze-av-aloha](https://github.com/ian-chuang/gaze-av-aloha)
20
+
21
+ ## About the Model
22
+
23
+ Human vision is a highly active process driven by gaze, which directs attention and fixation to task-relevant regions and dramatically reduces visual processing. This work explores how incorporating human-like active gaze into robotic policies can enhance both efficiency and performance.
24
+
25
+ This model is a component of an Active Vision robot system that emulates human head movement and eye tracking. It builds on recent advances in foveated image processing and integrates gaze information into Vision Transformers (ViTs) using a foveated patch tokenization scheme. This approach significantly reduces computational overhead (by 94%) and accelerates training (7x) and inference (3x) without sacrificing visual fidelity near regions of interest.
26
+
27
+ The framework supports simultaneously collecting eye-tracking data and robot demonstrations from a human operator, and provides a simulation benchmark and dataset for training robot policies that incorporate human gaze. The findings suggest that human-inspired visual processing offers a useful inductive bias for robotic vision systems.
28
+
29
+ This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration. It can be a pretrained vision encoder (like MAE-pretrained ViT weights) or a task-specific gaze model as detailed in the project's repository.
30
+
31
+ ## Sample Usage
32
+
33
+ To utilize this model as part of the "Look, Focus, Act" framework, you generally need to set up the environment and run training or evaluation scripts provided in the official GitHub repository. The project uses the `LeRobot` library for dataset handling and policy training.
34
+
35
+ For detailed installation instructions, dataset preparation, and policy training, please refer to the [official GitHub repository](https://github.com/ian-chuang/gaze-av-aloha).
36
+
37
+ Here is an example of how a foveated Vision Transformer policy can be trained, incorporating a gaze model (which this repository might represent if it's a specific gaze model instance):
38
+
39
+ ```bash
40
+ # Example from the project's train.py script for Fov-UNet (two-stage with pretrained gaze model)
41
+ python gaze_av_aloha/scripts/train.py \
42
+ policy=foveated_vit_policy \
43
+ task=<task e.g. av_aloha_sim_thread_needle> \
44
+ policy.use_gaze_as_action=false \
45
+ policy.gaze_model_repo_id=iantc104/gaze_model_av_aloha_sim_thread_needle \
46
+ policy.vision_encoder_kwargs.repo_id=iantc104/mae_vitb_foveated_vit \
47
+ policy.optimizer_lr_backbone=1e-5 \
48
+ wandb.enable=true \
49
+ wandb.project=<your_project_name> \
50
+ wandb.entity=<your_wandb_entity> \
51
+ wandb.job_name=fov-unet \
52
+ device=cuda
53
+ ```
54
+ Replace `<your_project_name>` and `<your_wandb_entity>` with your Weights & Biases credentials, and `<task e.g. av_aloha_sim_thread_needle>` with the specific task you are working on.
55
+
56
+ ## Citation
57
+
58
+ If you find this work useful for your research, please consider citing the original paper:
59
+
60
+ ```bibtex
61
+ @misc{chuang2025lookfocusactefficient,
62
+ title={Look, Focus, Act: Efficient and Robust Robot Learning via Human Gaze and Foveated Vision Transformers},
63
+ author={Ian Chuang and Andrew Lee and Dechen Gao and Jinyu Zou and Iman Soltani},
64
+ year={2025},
65
+ eprint={2507.15833},
66
+ archivePrefix={arXiv},
67
+ primaryClass={cs.RO},
68
+ url={https://arxiv.org/abs/2507.15833},
69
+ }
70
+ ```