hardware-pathon-ai commited on
Commit
9fc8ea3
·
verified ·
1 Parent(s): 796cd57

Upload SO-ARM101 reaching policy (999 iterations)

Browse files
Files changed (1) hide show
  1. README.md +126 -0
README.md ADDED
@@ -0,0 +1,126 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: en
3
+ license: mit
4
+ tags:
5
+ - reinforcement-learning
6
+ - robotics
7
+ - isaac-lab
8
+ - manipulation
9
+ - so-arm101
10
+ - inverse-kinematics
11
+ library_name: rsl-rl
12
+ ---
13
+
14
+ # SO-ARM101 Reaching Policy
15
+
16
+ This model is a reinforcement learning policy trained for the **SO-ARM101** robot arm to perform end-effector reaching tasks in Isaac Lab.
17
+
18
+ ## Model Description
19
+
20
+ - **Task**: Move the end-effector to randomly sampled target poses in 3D space
21
+ - **Robot**: SO-ARM101 (6-DOF robotic arm)
22
+ - **Framework**: Isaac Lab 2.3.0 (on Isaac Sim 5.1.0)
23
+ - **Algorithm**: RSL-RL (Robotic Systems Lab - Reinforcement Learning)
24
+ - **Environment**: `Isaac-SO-ARM101-Reach-v0`
25
+ - **Training**: 999 iterations with 4096 parallel environments
26
+
27
+ ## Model Overview
28
+
29
+ This policy learns to control the SO-ARM101 robot arm's joint positions to reach target end-effector poses. The model effectively learns inverse kinematics behavior through reinforcement learning, enabling the robot to accurately position its end-effector at desired 3D locations.
30
+
31
+ ## Training Details
32
+
33
+ ### Environment Configuration
34
+ - **Observation Space**: Joint positions, velocities, and target pose relative to end-effector
35
+ - **Action Space**: Joint position commands (6 DOF)
36
+ - **Reward Function**: Negative distance between end-effector and target pose
37
+ - **Episode Length**: Variable (resets on success or timeout)
38
+
39
+ ### Training Parameters
40
+ - **Parallel Environments**: 4096
41
+ - **Total Iterations**: 999
42
+ - **Training Time**: ~1.5 hours on NVIDIA RTX 4080 Super (16GB VRAM)
43
+ - **Framework**: Isaac Lab with RSL-RL runner
44
+ - **Simulator**: Isaac Sim 5.1.0
45
+
46
+ ### Hardware Used
47
+ - GPU: NVIDIA RTX 4080 Super (16GB VRAM)
48
+ - OS: Ubuntu 24.04 LTS
49
+ - CUDA: 13.0
50
+
51
+ ## Usage
52
+
53
+ ### Prerequisites
54
+
55
+ ```bash
56
+ # Install Isaac Lab (with Docker)
57
+ # See: https://isaac-sim.github.io/IsaacLab/
58
+
59
+ # Clone SO-ARM101 external project
60
+ git clone https://github.com/MuammerBay/isaac_so_arm101.git
61
+ cd isaac_so_arm101
62
+ ```
63
+
64
+ ### Evaluation
65
+
66
+ ```bash
67
+ # Inside Isaac Lab container
68
+ cd /workspace/isaaclab
69
+
70
+ # Run the trained policy
71
+ ./isaaclab.sh -p /workspace/isaac_so_arm101/src/isaac_so_arm101/scripts/rsl_rl/play.py \
72
+ --task Isaac-SO-ARM101-Reach-Play-v0 \
73
+ --checkpoint /path/to/model_999.pt
74
+ ```
75
+
76
+ ### Training From Scratch
77
+
78
+ ```bash
79
+ # Train the policy
80
+ ./isaaclab.sh -p /workspace/isaac_so_arm101/src/isaac_so_arm101/scripts/rsl_rl/train.py \
81
+ --task Isaac-SO-ARM101-Reach-v0 \
82
+ --num_envs 4096 \
83
+ --headless
84
+ ```
85
+
86
+ ## Performance
87
+
88
+ The trained policy demonstrates accurate reaching behavior with the SO-ARM101 robot, successfully moving the end-effector to target positions across the reachable workspace with high precision.
89
+
90
+ ## Use Cases
91
+
92
+ This reaching policy serves as a foundation for:
93
+ - **Inverse Kinematics**: Learned IK controller for end-effector positioning
94
+ - **Manipulation Tasks**: Base controller for pick-and-place, assembly, etc.
95
+ - **Trajectory Following**: Can be extended for path planning applications
96
+ - **Sim-to-Real Transfer**: Ready for deployment on real SO-ARM101 hardware
97
+
98
+ ## Citation
99
+
100
+ If you use this model, please cite:
101
+
102
+ ```bibtex
103
+ @misc{so-arm101-reach-isaaclab,
104
+ title={SO-ARM101 Reaching Policy trained with Isaac Lab},
105
+ author={PathOn AI},
106
+ year={2026},
107
+ howpublished={\url{https://huggingface.co/}},
108
+ }
109
+
110
+ @software{isaaclab,
111
+ author = {Mittal, Mayank and others},
112
+ title = {Isaac Lab: A Unified Framework for Robot Learning},
113
+ url = {https://isaac-sim.github.io/IsaacLab/},
114
+ year = {2024},
115
+ }
116
+ ```
117
+
118
+ ## Related Resources
119
+
120
+ - [Isaac Lab Documentation](https://isaac-sim.github.io/IsaacLab/)
121
+ - [SO-ARM101 Isaac Lab Integration](https://github.com/MuammerBay/isaac_so_arm101)
122
+ - [RSL-RL Library](https://github.com/leggedrobotics/rsl_rl)
123
+
124
+ ## License
125
+
126
+ MIT License