qaihm-bot commited on
Commit
e3a12f6
·
verified ·
1 Parent(s): 9fe174c

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +9 -3
README.md CHANGED
@@ -34,9 +34,12 @@ More details on model performance across various devices, can be found
34
  - Model size: 21.7 MB
35
 
36
 
 
 
37
  | Device | Chipset | Target Runtime | Inference Time (ms) | Peak Memory Range (MB) | Precision | Primary Compute Unit | Target Model
38
  | ---|---|---|---|---|---|---|---|
39
- | Samsung Galaxy S23 Ultra (Android 13) | Snapdragon® 8 Gen 2 | TFLite | 6.73 ms | 1 - 4 MB | FP16 | NPU | [DDRNet23-Slim.tflite](https://huggingface.co/qualcomm/DDRNet23-Slim/blob/main/DDRNet23-Slim.tflite)
 
40
 
41
 
42
  ## Installation
@@ -93,9 +96,11 @@ device. This script does the following:
93
  python -m qai_hub_models.models.ddrnet23_slim.export
94
  ```
95
 
 
 
96
  ## How does this work?
97
 
98
- This [export script](https://github.com/quic/ai-hub-models/blob/main/qai_hub_models/models/DDRNet23-Slim/export.py)
99
  leverages [Qualcomm® AI Hub](https://aihub.qualcomm.com/) to optimize, validate, and deploy this model
100
  on-device. Lets go through each step below in detail:
101
 
@@ -172,6 +177,7 @@ spot check the output with expected output.
172
  AI Hub. [Sign up for access](https://myaccount.qualcomm.com/signup).
173
 
174
 
 
175
  ## Run demo on a cloud-hosted device
176
 
177
  You can also run the demo on-device.
@@ -208,7 +214,7 @@ Explore all available models on [Qualcomm® AI Hub](https://aihub.qualcomm.com/)
208
  ## License
209
  - The license for the original implementation of DDRNet23-Slim can be found
210
  [here](https://github.com/chenjun2hao/DDRNet.pytorch/blob/main/LICENSE).
211
- - The license for the compiled assets for on-device deployment can be found [here]({deploy_license_url})
212
 
213
  ## References
214
  * [Deep Dual-resolution Networks for Real-time and Accurate Semantic Segmentation of Road Scenes](https://arxiv.org/abs/2101.06085)
 
34
  - Model size: 21.7 MB
35
 
36
 
37
+
38
+
39
  | Device | Chipset | Target Runtime | Inference Time (ms) | Peak Memory Range (MB) | Precision | Primary Compute Unit | Target Model
40
  | ---|---|---|---|---|---|---|---|
41
+ | Samsung Galaxy S23 Ultra (Android 13) | Snapdragon® 8 Gen 2 | TFLite | 6.65 ms | 0 - 26 MB | FP16 | NPU | [DDRNet23-Slim.tflite](https://huggingface.co/qualcomm/DDRNet23-Slim/blob/main/DDRNet23-Slim.tflite)
42
+
43
 
44
 
45
  ## Installation
 
96
  python -m qai_hub_models.models.ddrnet23_slim.export
97
  ```
98
 
99
+
100
+
101
  ## How does this work?
102
 
103
+ This [export script](https://aihub.qualcomm.com/models/ddrnet23_slim/qai_hub_models/models/DDRNet23-Slim/export.py)
104
  leverages [Qualcomm® AI Hub](https://aihub.qualcomm.com/) to optimize, validate, and deploy this model
105
  on-device. Lets go through each step below in detail:
106
 
 
177
  AI Hub. [Sign up for access](https://myaccount.qualcomm.com/signup).
178
 
179
 
180
+
181
  ## Run demo on a cloud-hosted device
182
 
183
  You can also run the demo on-device.
 
214
  ## License
215
  - The license for the original implementation of DDRNet23-Slim can be found
216
  [here](https://github.com/chenjun2hao/DDRNet.pytorch/blob/main/LICENSE).
217
+ - The license for the compiled assets for on-device deployment can be found [here](https://qaihub-public-assets.s3.us-west-2.amazonaws.com/qai-hub-models/Qualcomm+AI+Hub+Proprietary+License.pdf)
218
 
219
  ## References
220
  * [Deep Dual-resolution Networks for Real-time and Accurate Semantic Segmentation of Road Scenes](https://arxiv.org/abs/2101.06085)