qaihm-bot commited on
Commit
da8cdce
·
verified ·
1 Parent(s): eba905d

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +9 -9
README.md CHANGED
@@ -35,8 +35,8 @@ More details on model performance across various devices, can be found
35
 
36
  | Device | Chipset | Target Runtime | Inference Time (ms) | Peak Memory Range (MB) | Precision | Primary Compute Unit | Target Model
37
  | ---|---|---|---|---|---|---|---|
38
- | Samsung Galaxy S23 Ultra (Android 13) | Snapdragon® 8 Gen 2 | TFLite | 10.833 ms | 1 - 3 MB | FP16 | NPU | [FFNet-78S-LowRes.tflite](https://huggingface.co/qualcomm/FFNet-78S-LowRes/blob/main/FFNet-78S-LowRes.tflite)
39
- | Samsung Galaxy S23 Ultra (Android 13) | Snapdragon® 8 Gen 2 | QNN Model Library | 11.41 ms | 1 - 40 MB | FP16 | NPU | [FFNet-78S-LowRes.so](https://huggingface.co/qualcomm/FFNet-78S-LowRes/blob/main/FFNet-78S-LowRes.so)
40
 
41
 
42
  ## Installation
@@ -97,16 +97,16 @@ python -m qai_hub_models.models.ffnet_78s_lowres.export
97
  ```
98
  Profile Job summary of FFNet-78S-LowRes
99
  --------------------------------------------------
100
- Device: Samsung Galaxy S23 Ultra (13)
101
- Estimated Inference Time: 10.83 ms
102
- Estimated Peak Memory Range: 0.64-3.42 MB
103
  Compute Units: NPU (149) | Total (149)
104
 
105
  Profile Job summary of FFNet-78S-LowRes
106
  --------------------------------------------------
107
- Device: Samsung Galaxy S23 Ultra (13)
108
- Estimated Inference Time: 11.41 ms
109
- Estimated Peak Memory Range: 0.54-40.43 MB
110
  Compute Units: NPU (237) | Total (237)
111
 
112
 
@@ -212,7 +212,7 @@ Explore all available models on [Qualcomm® AI Hub](https://aihub.qualcomm.com/)
212
  ## License
213
  - The license for the original implementation of FFNet-78S-LowRes can be found
214
  [here](https://github.com/Qualcomm-AI-research/FFNet/blob/master/LICENSE).
215
- - The license for the compiled assets for on-device deployment can be found [here](https://qaihub-public-assets.s3.us-west-2.amazonaws.com/qai-hub-models/Qualcomm+AI+Hub+Proprietary+License.pdf).
216
 
217
  ## References
218
  * [Simple and Efficient Architectures for Semantic Segmentation](https://arxiv.org/abs/2206.08236)
 
35
 
36
  | Device | Chipset | Target Runtime | Inference Time (ms) | Peak Memory Range (MB) | Precision | Primary Compute Unit | Target Model
37
  | ---|---|---|---|---|---|---|---|
38
+ | Samsung Galaxy S23 Ultra (Android 13) | Snapdragon® 8 Gen 2 | TFLite | 10.81 ms | 0 - 2 MB | FP16 | NPU | [FFNet-78S-LowRes.tflite](https://huggingface.co/qualcomm/FFNet-78S-LowRes/blob/main/FFNet-78S-LowRes.tflite)
39
+ | Samsung Galaxy S23 Ultra (Android 13) | Snapdragon® 8 Gen 2 | QNN Model Library | 11.408 ms | 0 - 50 MB | FP16 | NPU | [FFNet-78S-LowRes.so](https://huggingface.co/qualcomm/FFNet-78S-LowRes/blob/main/FFNet-78S-LowRes.so)
40
 
41
 
42
  ## Installation
 
97
  ```
98
  Profile Job summary of FFNet-78S-LowRes
99
  --------------------------------------------------
100
+ Device: Samsung Galaxy S24 (14)
101
+ Estimated Inference Time: 7.77 ms
102
+ Estimated Peak Memory Range: 0.52-49.82 MB
103
  Compute Units: NPU (149) | Total (149)
104
 
105
  Profile Job summary of FFNet-78S-LowRes
106
  --------------------------------------------------
107
+ Device: Samsung Galaxy S24 (14)
108
+ Estimated Inference Time: 8.08 ms
109
+ Estimated Peak Memory Range: 6.04-69.22 MB
110
  Compute Units: NPU (237) | Total (237)
111
 
112
 
 
212
  ## License
213
  - The license for the original implementation of FFNet-78S-LowRes can be found
214
  [here](https://github.com/Qualcomm-AI-research/FFNet/blob/master/LICENSE).
215
+ - The license for the compiled assets for on-device deployment can be found [here]({deploy_license_url})
216
 
217
  ## References
218
  * [Simple and Efficient Architectures for Semantic Segmentation](https://arxiv.org/abs/2206.08236)