Update README.md
#1
by
mingxinz
- opened
README.md
CHANGED
|
@@ -1,3 +1,172 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
---
|
| 4 |
+
# Index
|
| 5 |
+
|
| 6 |
+
- [Overview](https://docs.google.com/document/d/1qonEBUTEX3zC5PL_pKILSCkBS6WQ5dGzDq4XCJM-hRw/edit#overview)
|
| 7 |
+
- [Bias](https://docs.google.com/document/d/1qonEBUTEX3zC5PL_pKILSCkBS6WQ5dGzDq4XCJM-hRw/edit#bias)
|
| 8 |
+
- [Explainability](https://docs.google.com/document/d/1qonEBUTEX3zC5PL_pKILSCkBS6WQ5dGzDq4XCJM-hRw/edit#explainability)
|
| 9 |
+
- [Privacy](https://docs.google.com/document/d/1qonEBUTEX3zC5PL_pKILSCkBS6WQ5dGzDq4XCJM-hRw/edit#privacy)
|
| 10 |
+
- [Safety & Security](https://docs.google.com/document/d/1qonEBUTEX3zC5PL_pKILSCkBS6WQ5dGzDq4XCJM-hRw/edit#safety--security)
|
| 11 |
+
|
| 12 |
+
# Model Overview
|
| 13 |
+
|
| 14 |
+
### Description:
|
| 15 |
+
|
| 16 |
+
GR00T N1.5 for SO-ARM starter is a vision language action model (VLA) fine-tuned to perform autonomous surgical assistance tasks, specifically surgical instrument management in the Isaac for Healthcare environment. It uses the weights and architecture from the NVIDIA Isaac GR00T N1.5, and is fine-tuned with simulation and real-world data from SO-ARM101 robotic arms.
|
| 17 |
+
|
| 18 |
+
This model is ready for commercial/non-commercial use.
|
| 19 |
+
|
| 20 |
+
### License/Terms of Use
|
| 21 |
+
|
| 22 |
+
NSCL V1 License
|
| 23 |
+
|
| 24 |
+
### Deployment Geography:
|
| 25 |
+
|
| 26 |
+
Global
|
| 27 |
+
|
| 28 |
+
Use Case:
|
| 29 |
+
This model is intended to be used within Isaac for Healthcare as an autonomous SO-ARM starter system that can perform essential surgical assistance duties, including surgical instrument handling and management.
|
| 30 |
+
|
| 31 |
+
|
| 32 |
+
## Reference(s):
|
| 33 |
+
|
| 34 |
+
[Nvidia Isaac GR00T N1.5](https://github.com/NVIDIA/Isaac-GR00T)
|
| 35 |
+
[Isaac For Healthcare](https://github.com/isaac-for-healthcare)
|
| 36 |
+
|
| 37 |
+
## Model Architecture:
|
| 38 |
+
|
| 39 |
+
**Architecture Type:** Vision Language Action model
|
| 40 |
+
|
| 41 |
+
**Network Architecture:** [GR00T N1.5]([https://github.com/NVIDIA/Isaac-GR00T](https://github.com/NVIDIA/Isaac-GR00T))
|
| 42 |
+
|
| 43 |
+
** This model was developed based on GR00T N1.5
|
| 44 |
+
** This model has 3 billion parameters.
|
| 45 |
+
|
| 46 |
+
### **Input**
|
| 47 |
+
|
| 48 |
+
Input Type(s): Vision, State, Language InstructionInput Format:
|
| 49 |
+
|
| 50 |
+
* Vision: Variable number of 224x224 uint8 image frames, coming from cameras
|
| 51 |
+
* State: Floating Point (Robot Proprioception)
|
| 52 |
+
* Language Instruction: String
|
| 53 |
+
|
| 54 |
+
Input Parameters:
|
| 55 |
+
|
| 56 |
+
* Vision: 2D - RGB image, square (224x224)
|
| 57 |
+
* State: 1D - Floating number vector
|
| 58 |
+
* Language Instruction: 1D - String
|
| 59 |
+
|
| 60 |
+
Input Images:
|
| 61 |
+
|
| 62 |
+
* Room Camera: 224x224 uint8 RGB image frames
|
| 63 |
+
* Wrist Camera: 224x224 uint8 RGB image frames
|
| 64 |
+
|
| 65 |
+
Input Prompt:Text String (Language Instruction)
|
| 66 |
+
|
| 67 |
+
**Output**
|
| 68 |
+
|
| 69 |
+
Output Type(s): Actions
|
| 70 |
+
Output Format: Continuous-value vectors
|
| 71 |
+
Output Parameters: Two-Dimensional (2D), 16x6 Tensor
|
| 72 |
+
Other Properties Related to Output: Continuous-value vectors correspond to different motor controls on a robot, which depends on Degrees of Freedom of the robot embodiment.
|
| 73 |
+
|
| 74 |
+
Our AI models are designed and/or optimized to run on NVIDIA GPU-accelerated systems. By leveraging NVIDIA’s hardware (e.g. GPU cores) and software frameworks (e.g., CUDA libraries), the model achieves faster training and inference times compared to CPU-only solutions.
|
| 75 |
+
|
| 76 |
+
## Software Integration:
|
| 77 |
+
|
| 78 |
+
**Runtime Engine(s):**
|
| 79 |
+
|
| 80 |
+
* Pytorch - 2.5.1
|
| 81 |
+
* TensorRT - 10.11.0.33
|
| 82 |
+
|
| 83 |
+
**Supported Hardware Microarchitecture Compatibility:**
|
| 84 |
+
|
| 85 |
+
* NVIDIA Ampere
|
| 86 |
+
|
| 87 |
+
* NVIDIA Blackwell
|
| 88 |
+
|
| 89 |
+
* NVIDIA Hopper
|
| 90 |
+
|
| 91 |
+
**Preferred/Supported Operating System(s):**
|
| 92 |
+
|
| 93 |
+
* Linux (Ubuntu 22.04/24.04 LTS)
|
| 94 |
+
|
| 95 |
+
## Model Version(s):
|
| 96 |
+
|
| 97 |
+
- GR00T N1.5 for SO-ARM starter
|
| 98 |
+
|
| 99 |
+
## Training Datasets:
|
| 100 |
+
|
| 101 |
+
Data Collection Method by Dataset:
|
| 102 |
+
|
| 103 |
+
* Manual teleoperation
|
| 104 |
+
|
| 105 |
+
# Inference:
|
| 106 |
+
|
| 107 |
+
**Engine:** Pytorch / TensorRT
|
| 108 |
+
|
| 109 |
+
**Test Hardware:**
|
| 110 |
+
|
| 111 |
+
* Ada RTX 6000
|
| 112 |
+
|
| 113 |
+
| Inference mode | Average Latency | Memory Usage |
|
| 114 |
+
| :---- | :---- | :---- |
|
| 115 |
+
| PyTorch | 42.16 ± 0.81 ms | 5.7 GB |
|
| 116 |
+
| TensorRT | 26.96 ± 1.86 ms | 6.6 GB |
|
| 117 |
+
|
| 118 |
+
# Limitations:
|
| 119 |
+
|
| 120 |
+
This model was trained on data from the Isaac for Healthcare SO-ARM starter workflow. Therefore, the model will only perform well in that specific surgical assistance environment. This model is not expected to generalize to different robot platforms, surgical instruments, or surgical procedures outside of the trained domain.
|
| 121 |
+
|
| 122 |
+
## Ethical Considerations:
|
| 123 |
+
|
| 124 |
+
NVIDIA believes Trustworthy AI is a shared responsibility, and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse.
|
| 125 |
+
|
| 126 |
+
For more detailed information on ethical considerations for this model, please see the Model Card++ Explainability, Bias, Safety & Security, and Privacy Subcards [Insert Link to Model Card++ subcards here].
|
| 127 |
+
|
| 128 |
+
Please report model quality, risk, security vulnerabilities or NVIDIA AI Concerns [here](https://www.nvidia.com/en-us/support/submit-security-vulnerability/).
|
| 129 |
+
|
| 130 |
+
# Bias
|
| 131 |
+
|
| 132 |
+
| Field | Response |
|
| 133 |
+
| :---- | :---- |
|
| 134 |
+
| Participation considerations from adversely impacted groups, protected class, in model design and testing: | Not Applicable |
|
| 135 |
+
| Measures taken to mitigate against unwanted bias: | Not Applicable |
|
| 136 |
+
|
| 137 |
+
# Explainability
|
| 138 |
+
|
| 139 |
+
| Field | Response |
|
| 140 |
+
| :---- | :---- |
|
| 141 |
+
| Intended Domain: | SO-ARM starter |
|
| 142 |
+
| Model Type: | Robot VLA model |
|
| 143 |
+
| Intended Users: | Isaac For Healthcare users testing the scrub nurse environment. |
|
| 144 |
+
| Output: | Action tensor (outputs the next 16 actions to complete the scissors handling) |
|
| 145 |
+
| Describe how the model works: | Accepts vision, language and robot observations, outputs robot action policy. |
|
| 146 |
+
| Name the adversely impacted groups this has been tested to deliver comparable outcomes regardless of: | Not Applicable |
|
| 147 |
+
| Technical Limitations & Mitigation: | This model was trained on data from the Isaac for Healthcare SO-ARM starter workflow. Therefore, the model will only perform well in that singular environment. This model is not expected to generalize to different robot platforms, surgical instruments, or surgical procedures. |
|
| 148 |
+
| Verified to have met prescribed NVIDIA quality standards: | Yes |
|
| 149 |
+
| Performance Metrics: | Latency, Accuracy |
|
| 150 |
+
| Potential Known Risks: | The model may not perfectly follow surgical protocols or handle unexpected surgical scenarios. This may happen due to: unexpected surgical setups, inconsistent camera positioning, and different deployment environments outside of the Isaac for Healthcare simulation environment. |
|
| 151 |
+
| Licensing: | NSCL V1 License |
|
| 152 |
+
|
| 153 |
+
# Privacy
|
| 154 |
+
|
| 155 |
+
| Field | Response |
|
| 156 |
+
| :---- | :---- |
|
| 157 |
+
| Generatable or reverse engineerable personal data? | None |
|
| 158 |
+
| Personal data used to create this model? | None |
|
| 159 |
+
| How often is dataset reviewed? | Before Release |
|
| 160 |
+
| Is there provenance for all datasets used in training? | Yes |
|
| 161 |
+
| Does data labeling (annotation, metadata) comply with privacy laws? | Yes |
|
| 162 |
+
| Applicable Privacy Policy | [https://www.nvidia.com/en-us/about-nvidia/privacy-policy/](https://www.nvidia.com/en-us/about-nvidia/privacy-policy/) |
|
| 163 |
+
|
| 164 |
+
# Safety & Security
|
| 165 |
+
|
| 166 |
+
| Field | Response |
|
| 167 |
+
| :---- | :---- |
|
| 168 |
+
| Model Application(s): | SO-ARM starter |
|
| 169 |
+
| Model Application Field(s): | Medical Devices, Machinery and Robotics |
|
| 170 |
+
| Describe the life critical impact (if present). | This model could pose significant risks if deployed on a robotic system in real surgical environments without proper validation. This model has been tested with simulation data and limited real-world data using Isaac for Healthcare and may make unexpected movements if attempted to be deployed in a new surgical environment. This model is not expected to generalize to different environments, robot platforms, or surgical procedures. |
|
| 171 |
+
| Use Case Restrictions: | Abide by NSCL V1 License |
|
| 172 |
+
| Model and dataset restrictions: | The Principle of Least Privilege (PoLP) is applied limiting access for dataset generation and model development. Restrictions enforce dataset access during training, and dataset license constraints adhered to. |
|