IlyasMoutawwakil HF Staff commited on
Commit
bc533eb
·
verified ·
1 Parent(s): 5094d13

Upload cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark_config.json with huggingface_hub

Browse files
cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark_config.json CHANGED
@@ -2,7 +2,7 @@
2
  "name": "cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM",
3
  "backend": {
4
  "name": "pytorch",
5
- "version": "2.7.0+cu128",
6
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
  "model": "hf-internal-testing/tiny-random-LlamaForCausalLM",
8
  "processor": "hf-internal-testing/tiny-random-LlamaForCausalLM",
@@ -79,10 +79,10 @@
79
  "environment": {
80
  "cpu": " AMD EPYC 7R32",
81
  "cpu_count": 16,
82
- "cpu_ram_mb": 66697.252864,
83
  "system": "Linux",
84
  "machine": "x86_64",
85
- "platform": "Linux-5.10.236-228.935.amzn2.x86_64-x86_64-with-glibc2.35",
86
  "processor": "x86_64",
87
  "python_version": "3.10.12",
88
  "gpu": [
@@ -92,15 +92,15 @@
92
  "gpu_vram_mb": 24146608128,
93
  "optimum_benchmark_version": "0.6.0.dev0",
94
  "optimum_benchmark_commit": null,
95
- "transformers_version": "4.52.3",
96
  "transformers_commit": null,
97
- "accelerate_version": "1.7.0",
98
  "accelerate_commit": null,
99
- "diffusers_version": "0.33.1",
100
  "diffusers_commit": null,
101
  "optimum_version": null,
102
  "optimum_commit": null,
103
- "timm_version": "1.0.15",
104
  "timm_commit": null,
105
  "peft_version": null,
106
  "peft_commit": null
 
2
  "name": "cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM",
3
  "backend": {
4
  "name": "pytorch",
5
+ "version": "2.8.0+cu128",
6
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
  "model": "hf-internal-testing/tiny-random-LlamaForCausalLM",
8
  "processor": "hf-internal-testing/tiny-random-LlamaForCausalLM",
 
79
  "environment": {
80
  "cpu": " AMD EPYC 7R32",
81
  "cpu_count": 16,
82
+ "cpu_ram_mb": 66697.248768,
83
  "system": "Linux",
84
  "machine": "x86_64",
85
+ "platform": "Linux-5.10.240-238.959.amzn2.x86_64-x86_64-with-glibc2.35",
86
  "processor": "x86_64",
87
  "python_version": "3.10.12",
88
  "gpu": [
 
92
  "gpu_vram_mb": 24146608128,
93
  "optimum_benchmark_version": "0.6.0.dev0",
94
  "optimum_benchmark_commit": null,
95
+ "transformers_version": "4.55.2",
96
  "transformers_commit": null,
97
+ "accelerate_version": "1.10.0",
98
  "accelerate_commit": null,
99
+ "diffusers_version": "0.34.0",
100
  "diffusers_commit": null,
101
  "optimum_version": null,
102
  "optimum_commit": null,
103
+ "timm_version": "1.0.19",
104
  "timm_commit": null,
105
  "peft_version": null,
106
  "peft_commit": null