Spaces:
No application file
No application file
alo
Browse files
classification/inference_onnx.py
CHANGED
|
@@ -38,7 +38,7 @@ def load_model(model_path: str):
|
|
| 38 |
if "CUDAExecutionProvider" in available_providers:
|
| 39 |
# "TensorrtExecutionProvider"
|
| 40 |
# "CUDAExecutionProvider"
|
| 41 |
-
providers = ["
|
| 42 |
print("Using CUDA provider")
|
| 43 |
else:
|
| 44 |
providers = ["CPUExecutionProvider"]
|
|
|
|
| 38 |
if "CUDAExecutionProvider" in available_providers:
|
| 39 |
# "TensorrtExecutionProvider"
|
| 40 |
# "CUDAExecutionProvider"
|
| 41 |
+
providers = ["CUDAExecutionProvider", "CPUExecutionProvider"]
|
| 42 |
print("Using CUDA provider")
|
| 43 |
else:
|
| 44 |
providers = ["CPUExecutionProvider"]
|
ultrafast/ultrafastLaneDetector.py
CHANGED
|
@@ -135,7 +135,7 @@ class UltrafastLaneDetector:
|
|
| 135 |
if 'CUDAExecutionProvider' in available_providers:
|
| 136 |
# "TensorrtExecutionProvider"
|
| 137 |
# "CUDAExecutionProvider"
|
| 138 |
-
providers = ['
|
| 139 |
print("Using CUDA provider")
|
| 140 |
else:
|
| 141 |
providers = ['CPUExecutionProvider']
|
|
|
|
| 135 |
if 'CUDAExecutionProvider' in available_providers:
|
| 136 |
# "TensorrtExecutionProvider"
|
| 137 |
# "CUDAExecutionProvider"
|
| 138 |
+
providers = ['CUDAExecutionProvider', 'CPUExecutionProvider']
|
| 139 |
print("Using CUDA provider")
|
| 140 |
else:
|
| 141 |
providers = ['CPUExecutionProvider']
|