IlyasMoutawwakil HF Staff commited on
Commit
785e22d
·
verified ·
1 Parent(s): 4d4c226

Upload cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark.json with huggingface_hub

Browse files
cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.3.1+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "model": "hf-internal-testing/tiny-random-LlamaForCausalLM",
9
  "processor": "hf-internal-testing/tiny-random-LlamaForCausalLM",
@@ -11,7 +11,7 @@
11
  "library": "transformers",
12
  "model_type": "llama",
13
  "device": "cuda",
14
- "device_ids": "4",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
@@ -110,7 +110,7 @@
110
  "optimum_commit": null,
111
  "timm_version": "1.0.15",
112
  "timm_commit": null,
113
- "peft_version": "0.15.2",
114
  "peft_commit": null
115
  },
116
  "print_report": true,
@@ -120,34 +120,34 @@
120
  "overall": {
121
  "memory": {
122
  "unit": "MB",
123
- "max_ram": 1321.013248,
124
  "max_global_vram": 68702.69952,
125
- "max_process_vram": 60605.60384,
126
- "max_reserved": 44.040192,
127
- "max_allocated": 25.311744
128
  },
129
  "latency": {
130
  "unit": "s",
131
  "values": [
132
- 0.6614071655273438,
133
- 0.009809419631958008,
134
- 0.009661101341247558,
135
- 0.009308781623840332,
136
- 0.009224302291870117
137
  ],
138
  "count": 5,
139
- "total": 0.6994107704162598,
140
- "mean": 0.13988215408325194,
141
- "p50": 0.009661101341247558,
142
- "p90": 0.40076806716918956,
143
- "p95": 0.5310876163482665,
144
- "p99": 0.6353432556915284,
145
- "stdev": 0.26076259556032794,
146
- "stdev_": 186.41591364480493
147
  },
148
  "throughput": {
149
  "unit": "samples/s",
150
- "value": 71.48874754994424
151
  },
152
  "energy": null,
153
  "efficiency": null
@@ -155,31 +155,31 @@
155
  "warmup": {
156
  "memory": {
157
  "unit": "MB",
158
- "max_ram": 1321.013248,
159
  "max_global_vram": 68702.69952,
160
- "max_process_vram": 60605.60384,
161
- "max_reserved": 44.040192,
162
- "max_allocated": 25.311744
163
  },
164
  "latency": {
165
  "unit": "s",
166
  "values": [
167
- 0.6614071655273438,
168
- 0.009809419631958008
169
  ],
170
  "count": 2,
171
- "total": 0.6712165851593018,
172
- "mean": 0.3356082925796509,
173
- "p50": 0.3356082925796509,
174
- "p90": 0.5962473909378052,
175
- "p95": 0.6288272782325744,
176
- "p99": 0.65489118806839,
177
- "stdev": 0.3257988729476929,
178
- "stdev_": 97.07712239272814
179
  },
180
  "throughput": {
181
  "unit": "samples/s",
182
- "value": 11.918656625716922
183
  },
184
  "energy": null,
185
  "efficiency": null
@@ -187,32 +187,32 @@
187
  "train": {
188
  "memory": {
189
  "unit": "MB",
190
- "max_ram": 1321.013248,
191
  "max_global_vram": 68702.69952,
192
- "max_process_vram": 60605.60384,
193
- "max_reserved": 44.040192,
194
- "max_allocated": 25.311744
195
  },
196
  "latency": {
197
  "unit": "s",
198
  "values": [
199
- 0.009661101341247558,
200
- 0.009308781623840332,
201
- 0.009224302291870117
202
  ],
203
  "count": 3,
204
- "total": 0.028194185256958007,
205
- "mean": 0.009398061752319336,
206
- "p50": 0.009308781623840332,
207
- "p90": 0.009590637397766113,
208
- "p95": 0.009625869369506835,
209
- "p99": 0.009654054946899414,
210
- "stdev": 0.00018916757719048932,
211
- "stdev_": 2.0128360738192095
212
  },
213
  "throughput": {
214
  "unit": "samples/s",
215
- "value": 638.4295143112107
216
  },
217
  "energy": null,
218
  "efficiency": null
 
3
  "name": "cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.7.0+rocm6.3",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "model": "hf-internal-testing/tiny-random-LlamaForCausalLM",
9
  "processor": "hf-internal-testing/tiny-random-LlamaForCausalLM",
 
11
  "library": "transformers",
12
  "model_type": "llama",
13
  "device": "cuda",
14
+ "device_ids": "5",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
 
110
  "optimum_commit": null,
111
  "timm_version": "1.0.15",
112
  "timm_commit": null,
113
+ "peft_version": null,
114
  "peft_commit": null
115
  },
116
  "print_report": true,
 
120
  "overall": {
121
  "memory": {
122
  "unit": "MB",
123
+ "max_ram": 2343.985152,
124
  "max_global_vram": 68702.69952,
125
+ "max_process_vram": 0.0,
126
+ "max_reserved": 182.452224,
127
+ "max_allocated": 98.354688
128
  },
129
  "latency": {
130
  "unit": "s",
131
  "values": [
132
+ 1.1652205810546874,
133
+ 0.012251338958740234,
134
+ 0.01122590160369873,
135
+ 0.011280620574951172,
136
+ 0.0112243013381958
137
  ],
138
  "count": 5,
139
+ "total": 1.2112027435302732,
140
+ "mean": 0.24224054870605466,
141
+ "p50": 0.011280620574951172,
142
+ "p90": 0.7040328842163086,
143
+ "p95": 0.9346267326354978,
144
+ "p99": 1.1191018113708495,
145
+ "stdev": 0.46149018165949224,
146
+ "stdev_": 190.5090556162357
147
  },
148
  "throughput": {
149
  "unit": "samples/s",
150
+ "value": 41.281280336490816
151
  },
152
  "energy": null,
153
  "efficiency": null
 
155
  "warmup": {
156
  "memory": {
157
  "unit": "MB",
158
+ "max_ram": 2343.985152,
159
  "max_global_vram": 68702.69952,
160
+ "max_process_vram": 0.0,
161
+ "max_reserved": 182.452224,
162
+ "max_allocated": 98.354688
163
  },
164
  "latency": {
165
  "unit": "s",
166
  "values": [
167
+ 1.1652205810546874,
168
+ 0.012251338958740234
169
  ],
170
  "count": 2,
171
+ "total": 1.1774719200134276,
172
+ "mean": 0.5887359600067138,
173
+ "p50": 0.5887359600067138,
174
+ "p90": 1.0499236568450927,
175
+ "p95": 1.10757211894989,
176
+ "p99": 1.1536908886337278,
177
+ "stdev": 0.5764846210479736,
178
+ "stdev_": 97.91904354566681
179
  },
180
  "throughput": {
181
  "unit": "samples/s",
182
+ "value": 6.794217224227963
183
  },
184
  "energy": null,
185
  "efficiency": null
 
187
  "train": {
188
  "memory": {
189
  "unit": "MB",
190
+ "max_ram": 2343.985152,
191
  "max_global_vram": 68702.69952,
192
+ "max_process_vram": 0.0,
193
+ "max_reserved": 182.452224,
194
+ "max_allocated": 98.354688
195
  },
196
  "latency": {
197
  "unit": "s",
198
  "values": [
199
+ 0.01122590160369873,
200
+ 0.011280620574951172,
201
+ 0.0112243013381958
202
  ],
203
  "count": 3,
204
+ "total": 0.0337308235168457,
205
+ "mean": 0.011243607838948566,
206
+ "p50": 0.01122590160369873,
207
+ "p90": 0.011269676780700684,
208
+ "p95": 0.011275148677825928,
209
+ "p99": 0.011279526195526123,
210
+ "stdev": 2.618010927172042e-05,
211
+ "stdev_": 0.2328443827525794
212
  },
213
  "throughput": {
214
  "unit": "samples/s",
215
+ "value": 533.6365413969369
216
  },
217
  "energy": null,
218
  "efficiency": null