Upload cuda_inference_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark_config.json with huggingface_hub
Browse files
cuda_inference_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark_config.json
CHANGED
|
@@ -10,13 +10,14 @@
|
|
| 10 |
"library": "transformers",
|
| 11 |
"model_type": "llama",
|
| 12 |
"device": "cuda",
|
| 13 |
-
"device_ids": "
|
| 14 |
"seed": 42,
|
| 15 |
"inter_op_num_threads": null,
|
| 16 |
"intra_op_num_threads": null,
|
| 17 |
"model_kwargs": {},
|
| 18 |
"processor_kwargs": {},
|
| 19 |
"no_weights": true,
|
|
|
|
| 20 |
"device_map": null,
|
| 21 |
"torch_dtype": null,
|
| 22 |
"eval_mode": true,
|
|
|
|
| 10 |
"library": "transformers",
|
| 11 |
"model_type": "llama",
|
| 12 |
"device": "cuda",
|
| 13 |
+
"device_ids": "4",
|
| 14 |
"seed": 42,
|
| 15 |
"inter_op_num_threads": null,
|
| 16 |
"intra_op_num_threads": null,
|
| 17 |
"model_kwargs": {},
|
| 18 |
"processor_kwargs": {},
|
| 19 |
"no_weights": true,
|
| 20 |
+
"tp_plan": null,
|
| 21 |
"device_map": null,
|
| 22 |
"torch_dtype": null,
|
| 23 |
"eval_mode": true,
|