Upload cpu_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark_config.json with huggingface_hub
Browse files
cpu_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark_config.json
CHANGED
|
@@ -2,7 +2,7 @@
|
|
| 2 |
"name": "cpu_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM",
|
| 3 |
"backend": {
|
| 4 |
"name": "pytorch",
|
| 5 |
-
"version": "2.
|
| 6 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
| 7 |
"model": "hf-internal-testing/tiny-random-LlamaForCausalLM",
|
| 8 |
"processor": "hf-internal-testing/tiny-random-LlamaForCausalLM",
|
|
@@ -85,8 +85,8 @@
|
|
| 85 |
"processor": "x86_64",
|
| 86 |
"python_version": "3.10.16",
|
| 87 |
"optimum_benchmark_version": "0.6.0.dev0",
|
| 88 |
-
"optimum_benchmark_commit": "
|
| 89 |
-
"transformers_version": "4.48.
|
| 90 |
"transformers_commit": null,
|
| 91 |
"accelerate_version": "1.3.0",
|
| 92 |
"accelerate_commit": null,
|
|
|
|
| 2 |
"name": "cpu_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM",
|
| 3 |
"backend": {
|
| 4 |
"name": "pytorch",
|
| 5 |
+
"version": "2.6.0+cpu",
|
| 6 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
| 7 |
"model": "hf-internal-testing/tiny-random-LlamaForCausalLM",
|
| 8 |
"processor": "hf-internal-testing/tiny-random-LlamaForCausalLM",
|
|
|
|
| 85 |
"processor": "x86_64",
|
| 86 |
"python_version": "3.10.16",
|
| 87 |
"optimum_benchmark_version": "0.6.0.dev0",
|
| 88 |
+
"optimum_benchmark_commit": "4eb7a37589fa5efafd23072041135e22808603ce",
|
| 89 |
+
"transformers_version": "4.48.2",
|
| 90 |
"transformers_commit": null,
|
| 91 |
"accelerate_version": "1.3.0",
|
| 92 |
"accelerate_commit": null,
|