IlyasMoutawwakil HF Staff commited on
Commit
a60acce
·
verified ·
1 Parent(s): 488b8e8

Upload cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark_config.json with huggingface_hub

Browse files
cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark_config.json CHANGED
@@ -2,15 +2,15 @@
2
  "name": "cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM",
3
  "backend": {
4
  "name": "pytorch",
5
- "version": "2.3.1+rocm5.7",
6
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
 
 
7
  "task": "text-generation",
8
  "library": "transformers",
9
  "model_type": "llama",
10
- "model": "hf-internal-testing/tiny-random-LlamaForCausalLM",
11
- "processor": "hf-internal-testing/tiny-random-LlamaForCausalLM",
12
  "device": "cuda",
13
- "device_ids": "6",
14
  "seed": 42,
15
  "inter_op_num_threads": null,
16
  "intra_op_num_threads": null,
@@ -24,6 +24,7 @@
24
  "low_cpu_mem_usage": null,
25
  "attn_implementation": null,
26
  "cache_implementation": null,
 
27
  "autocast_enabled": false,
28
  "autocast_dtype": null,
29
  "torch_compile": false,
@@ -50,7 +51,6 @@
50
  "per_device_train_batch_size": 2,
51
  "gradient_accumulation_steps": 1,
52
  "output_dir": "./trainer_output",
53
- "evaluation_strategy": "no",
54
  "eval_strategy": "no",
55
  "save_strategy": "no",
56
  "do_train": true,
@@ -78,10 +78,10 @@
78
  "environment": {
79
  "cpu": " AMD EPYC 7763 64-Core Processor",
80
  "cpu_count": 128,
81
- "cpu_ram_mb": 1082014.490624,
82
  "system": "Linux",
83
  "machine": "x86_64",
84
- "platform": "Linux-5.15.0-122-generic-x86_64-with-glibc2.35",
85
  "processor": "x86_64",
86
  "python_version": "3.10.12",
87
  "gpu": [
@@ -96,19 +96,19 @@
96
  ],
97
  "gpu_count": 8,
98
  "gpu_vram_mb": 549621596160,
99
- "optimum_benchmark_version": "0.5.0.dev0",
100
  "optimum_benchmark_commit": null,
101
- "transformers_version": "4.47.0",
102
  "transformers_commit": null,
103
- "accelerate_version": "1.2.0",
104
  "accelerate_commit": null,
105
- "diffusers_version": "0.31.0",
106
  "diffusers_commit": null,
107
  "optimum_version": null,
108
  "optimum_commit": null,
109
- "timm_version": "1.0.12",
110
  "timm_commit": null,
111
- "peft_version": "0.14.0",
112
  "peft_commit": null
113
  },
114
  "print_report": true,
 
2
  "name": "cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM",
3
  "backend": {
4
  "name": "pytorch",
5
+ "version": "2.7.0+rocm6.3",
6
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "model": "hf-internal-testing/tiny-random-LlamaForCausalLM",
8
+ "processor": "hf-internal-testing/tiny-random-LlamaForCausalLM",
9
  "task": "text-generation",
10
  "library": "transformers",
11
  "model_type": "llama",
 
 
12
  "device": "cuda",
13
+ "device_ids": "5",
14
  "seed": 42,
15
  "inter_op_num_threads": null,
16
  "intra_op_num_threads": null,
 
24
  "low_cpu_mem_usage": null,
25
  "attn_implementation": null,
26
  "cache_implementation": null,
27
+ "allow_tf32": false,
28
  "autocast_enabled": false,
29
  "autocast_dtype": null,
30
  "torch_compile": false,
 
51
  "per_device_train_batch_size": 2,
52
  "gradient_accumulation_steps": 1,
53
  "output_dir": "./trainer_output",
 
54
  "eval_strategy": "no",
55
  "save_strategy": "no",
56
  "do_train": true,
 
78
  "environment": {
79
  "cpu": " AMD EPYC 7763 64-Core Processor",
80
  "cpu_count": 128,
81
+ "cpu_ram_mb": 1082014.486528,
82
  "system": "Linux",
83
  "machine": "x86_64",
84
+ "platform": "Linux-5.15.0-131-generic-x86_64-with-glibc2.35",
85
  "processor": "x86_64",
86
  "python_version": "3.10.12",
87
  "gpu": [
 
96
  ],
97
  "gpu_count": 8,
98
  "gpu_vram_mb": 549621596160,
99
+ "optimum_benchmark_version": "0.6.0.dev0",
100
  "optimum_benchmark_commit": null,
101
+ "transformers_version": "4.51.3",
102
  "transformers_commit": null,
103
+ "accelerate_version": "1.6.0",
104
  "accelerate_commit": null,
105
+ "diffusers_version": "0.33.1",
106
  "diffusers_commit": null,
107
  "optimum_version": null,
108
  "optimum_commit": null,
109
+ "timm_version": "1.0.15",
110
  "timm_commit": null,
111
+ "peft_version": null,
112
  "peft_commit": null
113
  },
114
  "print_report": true,