IlyasMoutawwakil HF Staff commited on
Commit
548eb08
·
verified ·
1 Parent(s): a60acce

Upload cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark.json with huggingface_hub

Browse files
cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark.json CHANGED
@@ -3,15 +3,15 @@
3
  "name": "cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.3.1+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
 
 
8
  "task": "text-generation",
9
  "library": "transformers",
10
  "model_type": "llama",
11
- "model": "hf-internal-testing/tiny-random-LlamaForCausalLM",
12
- "processor": "hf-internal-testing/tiny-random-LlamaForCausalLM",
13
  "device": "cuda",
14
- "device_ids": "6",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
@@ -25,6 +25,7 @@
25
  "low_cpu_mem_usage": null,
26
  "attn_implementation": null,
27
  "cache_implementation": null,
 
28
  "autocast_enabled": false,
29
  "autocast_dtype": null,
30
  "torch_compile": false,
@@ -51,7 +52,6 @@
51
  "per_device_train_batch_size": 2,
52
  "gradient_accumulation_steps": 1,
53
  "output_dir": "./trainer_output",
54
- "evaluation_strategy": "no",
55
  "eval_strategy": "no",
56
  "save_strategy": "no",
57
  "do_train": true,
@@ -79,10 +79,10 @@
79
  "environment": {
80
  "cpu": " AMD EPYC 7763 64-Core Processor",
81
  "cpu_count": 128,
82
- "cpu_ram_mb": 1082014.490624,
83
  "system": "Linux",
84
  "machine": "x86_64",
85
- "platform": "Linux-5.15.0-122-generic-x86_64-with-glibc2.35",
86
  "processor": "x86_64",
87
  "python_version": "3.10.12",
88
  "gpu": [
@@ -97,19 +97,19 @@
97
  ],
98
  "gpu_count": 8,
99
  "gpu_vram_mb": 549621596160,
100
- "optimum_benchmark_version": "0.5.0.dev0",
101
  "optimum_benchmark_commit": null,
102
- "transformers_version": "4.47.0",
103
  "transformers_commit": null,
104
- "accelerate_version": "1.2.0",
105
  "accelerate_commit": null,
106
- "diffusers_version": "0.31.0",
107
  "diffusers_commit": null,
108
  "optimum_version": null,
109
  "optimum_commit": null,
110
- "timm_version": "1.0.12",
111
  "timm_commit": null,
112
- "peft_version": "0.14.0",
113
  "peft_commit": null
114
  },
115
  "print_report": true,
@@ -119,34 +119,34 @@
119
  "overall": {
120
  "memory": {
121
  "unit": "MB",
122
- "max_ram": 1312.206848,
123
  "max_global_vram": 68702.69952,
124
- "max_process_vram": 65339.850752,
125
- "max_reserved": 44.040192,
126
- "max_allocated": 25.06496
127
  },
128
  "latency": {
129
  "unit": "s",
130
  "values": [
131
- 0.5945537719726562,
132
- 0.010716172218322753,
133
- 0.009252650260925293,
134
- 0.009294569969177246,
135
- 0.00920864963531494
136
  ],
137
  "count": 5,
138
- "total": 0.6330258140563965,
139
- "mean": 0.1266051628112793,
140
- "p50": 0.009294569969177246,
141
- "p90": 0.36101873207092294,
142
- "p95": 0.4777862520217894,
143
- "p99": 0.5712002679824829,
144
- "stdev": 0.2339749933892262,
145
- "stdev_": 184.8068342505076
146
  },
147
  "throughput": {
148
  "unit": "samples/s",
149
- "value": 78.9857204710225
150
  },
151
  "energy": null,
152
  "efficiency": null
@@ -154,31 +154,31 @@
154
  "warmup": {
155
  "memory": {
156
  "unit": "MB",
157
- "max_ram": 1312.206848,
158
  "max_global_vram": 68702.69952,
159
- "max_process_vram": 65339.850752,
160
- "max_reserved": 44.040192,
161
- "max_allocated": 25.06496
162
  },
163
  "latency": {
164
  "unit": "s",
165
  "values": [
166
- 0.5945537719726562,
167
- 0.010716172218322753
168
  ],
169
  "count": 2,
170
- "total": 0.605269944190979,
171
- "mean": 0.3026349720954895,
172
- "p50": 0.3026349720954895,
173
- "p90": 0.5361700119972229,
174
- "p95": 0.5653618919849396,
175
- "p99": 0.5887153959751129,
176
- "stdev": 0.29191879987716673,
177
- "stdev_": 96.45904366434507
178
  },
179
  "throughput": {
180
  "unit": "samples/s",
181
- "value": 13.217243110746276
182
  },
183
  "energy": null,
184
  "efficiency": null
@@ -186,32 +186,32 @@
186
  "train": {
187
  "memory": {
188
  "unit": "MB",
189
- "max_ram": 1312.206848,
190
  "max_global_vram": 68702.69952,
191
- "max_process_vram": 65339.850752,
192
- "max_reserved": 44.040192,
193
- "max_allocated": 25.06496
194
  },
195
  "latency": {
196
  "unit": "s",
197
  "values": [
198
- 0.009252650260925293,
199
- 0.009294569969177246,
200
- 0.00920864963531494
201
  ],
202
  "count": 3,
203
- "total": 0.027755869865417476,
204
- "mean": 0.009251956621805826,
205
- "p50": 0.009252650260925293,
206
- "p90": 0.009286186027526856,
207
- "p95": 0.009290377998352051,
208
- "p99": 0.009293731575012208,
209
- "stdev": 3.508025840054541e-05,
210
- "stdev_": 0.37916583307216517
211
  },
212
  "throughput": {
213
  "unit": "samples/s",
214
- "value": 648.5114711691008
215
  },
216
  "energy": null,
217
  "efficiency": null
 
3
  "name": "cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.7.0+rocm6.3",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "model": "hf-internal-testing/tiny-random-LlamaForCausalLM",
9
+ "processor": "hf-internal-testing/tiny-random-LlamaForCausalLM",
10
  "task": "text-generation",
11
  "library": "transformers",
12
  "model_type": "llama",
 
 
13
  "device": "cuda",
14
+ "device_ids": "5",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
 
25
  "low_cpu_mem_usage": null,
26
  "attn_implementation": null,
27
  "cache_implementation": null,
28
+ "allow_tf32": false,
29
  "autocast_enabled": false,
30
  "autocast_dtype": null,
31
  "torch_compile": false,
 
52
  "per_device_train_batch_size": 2,
53
  "gradient_accumulation_steps": 1,
54
  "output_dir": "./trainer_output",
 
55
  "eval_strategy": "no",
56
  "save_strategy": "no",
57
  "do_train": true,
 
79
  "environment": {
80
  "cpu": " AMD EPYC 7763 64-Core Processor",
81
  "cpu_count": 128,
82
+ "cpu_ram_mb": 1082014.486528,
83
  "system": "Linux",
84
  "machine": "x86_64",
85
+ "platform": "Linux-5.15.0-131-generic-x86_64-with-glibc2.35",
86
  "processor": "x86_64",
87
  "python_version": "3.10.12",
88
  "gpu": [
 
97
  ],
98
  "gpu_count": 8,
99
  "gpu_vram_mb": 549621596160,
100
+ "optimum_benchmark_version": "0.6.0.dev0",
101
  "optimum_benchmark_commit": null,
102
+ "transformers_version": "4.51.3",
103
  "transformers_commit": null,
104
+ "accelerate_version": "1.6.0",
105
  "accelerate_commit": null,
106
+ "diffusers_version": "0.33.1",
107
  "diffusers_commit": null,
108
  "optimum_version": null,
109
  "optimum_commit": null,
110
+ "timm_version": "1.0.15",
111
  "timm_commit": null,
112
+ "peft_version": null,
113
  "peft_commit": null
114
  },
115
  "print_report": true,
 
119
  "overall": {
120
  "memory": {
121
  "unit": "MB",
122
+ "max_ram": 2345.480192,
123
  "max_global_vram": 68702.69952,
124
+ "max_process_vram": 0.0,
125
+ "max_reserved": 182.452224,
126
+ "max_allocated": 98.354688
127
  },
128
  "latency": {
129
  "unit": "s",
130
  "values": [
131
+ 1.1069317626953126,
132
+ 0.012313740730285645,
133
+ 0.00990590476989746,
134
+ 0.01005998420715332,
135
+ 0.010109745025634765
136
  ],
137
  "count": 5,
138
+ "total": 1.1493211374282837,
139
+ "mean": 0.22986422748565674,
140
+ "p50": 0.010109745025634765,
141
+ "p90": 0.6690845539093019,
142
+ "p95": 0.888008158302307,
143
+ "p99": 1.0631470418167115,
144
+ "stdev": 0.43853466847278505,
145
+ "stdev_": 190.7798674328954
146
  },
147
  "throughput": {
148
  "unit": "samples/s",
149
+ "value": 43.503941911204905
150
  },
151
  "energy": null,
152
  "efficiency": null
 
154
  "warmup": {
155
  "memory": {
156
  "unit": "MB",
157
+ "max_ram": 2345.480192,
158
  "max_global_vram": 68702.69952,
159
+ "max_process_vram": 0.0,
160
+ "max_reserved": 182.452224,
161
+ "max_allocated": 98.354688
162
  },
163
  "latency": {
164
  "unit": "s",
165
  "values": [
166
+ 1.1069317626953126,
167
+ 0.012313740730285645
168
  ],
169
  "count": 2,
170
+ "total": 1.1192455034255981,
171
+ "mean": 0.5596227517127991,
172
+ "p50": 0.5596227517127991,
173
+ "p90": 0.9974699604988099,
174
+ "p95": 1.0522008615970613,
175
+ "p99": 1.0959855824756624,
176
+ "stdev": 0.5473090109825135,
177
+ "stdev_": 97.79963543429967
178
  },
179
  "throughput": {
180
  "unit": "samples/s",
181
+ "value": 7.147672226973392
182
  },
183
  "energy": null,
184
  "efficiency": null
 
186
  "train": {
187
  "memory": {
188
  "unit": "MB",
189
+ "max_ram": 2345.480192,
190
  "max_global_vram": 68702.69952,
191
+ "max_process_vram": 0.0,
192
+ "max_reserved": 182.452224,
193
+ "max_allocated": 98.354688
194
  },
195
  "latency": {
196
  "unit": "s",
197
  "values": [
198
+ 0.00990590476989746,
199
+ 0.01005998420715332,
200
+ 0.010109745025634765
201
  ],
202
  "count": 3,
203
+ "total": 0.030075634002685547,
204
+ "mean": 0.010025211334228516,
205
+ "p50": 0.01005998420715332,
206
+ "p90": 0.010099792861938475,
207
+ "p95": 0.01010476894378662,
208
+ "p99": 0.010108749809265136,
209
+ "stdev": 8.677394763923839e-05,
210
+ "stdev_": 0.8655572909767097
211
  },
212
  "throughput": {
213
  "unit": "samples/s",
214
+ "value": 598.4911240239433
215
  },
216
  "energy": null,
217
  "efficiency": null