nm-research commited on
Commit
f3c2819
·
verified ·
1 Parent(s): 429104a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +119 -2
README.md CHANGED
@@ -24,7 +24,7 @@ library_name: transformers
24
  - **Model Developers:** Neural Magic
25
 
26
  Quantized version of [ibm-granite/granite-3.1-8b-instruct](https://huggingface.co/ibm-granite/granite-3.1-8b-instruct).
27
- It achieves an average score of xxxx on the [OpenLLM](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard) benchmark (version 1), whereas the unquantized model achieves xxxx.
28
 
29
  ### Model Optimizations
30
 
@@ -178,4 +178,121 @@ evalplus.evaluate \
178
  | HumanEval Pass@1 | 71.00 | 69.90 |
179
 
180
 
181
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  - **Model Developers:** Neural Magic
25
 
26
  Quantized version of [ibm-granite/granite-3.1-8b-instruct](https://huggingface.co/ibm-granite/granite-3.1-8b-instruct).
27
+ It achieves an average score of 70.57 on the [OpenLLM](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard) benchmark (version 1), whereas the unquantized model achieves 70.30.
28
 
29
  ### Model Optimizations
30
 
 
178
  | HumanEval Pass@1 | 71.00 | 69.90 |
179
 
180
 
181
+ ## Inference Performance
182
+
183
+
184
+ This model achieves up to 1.5x speedup in single-stream deployment and up to 1.1x speedup in multi-stream asynchronous deployment on L40 GPUs.
185
+ The following performance benchmarks were conducted with [vLLM](https://docs.vllm.ai/en/latest/) version 0.6.6.post1, and [GuideLLM](https://github.com/neuralmagic/guidellm).
186
+
187
+ ### Single-stream performance (measured with vLLM version 0.6.6.post1)
188
+ <table>
189
+ <tr>
190
+ <td></td>
191
+ <td></td>
192
+ <td></td>
193
+ <th style="text-align: center;" colspan="7" >Latency (s)</th>
194
+ </tr>
195
+ <tr>
196
+ <th>GPU class</th>
197
+ <th>Model</th>
198
+ <th>Speedup</th>
199
+ <th>Code Completion<br>prefill: 256 tokens<br>decode: 1024 tokens</th>
200
+ <th>Docstring Generation<br>prefill: 768 tokens<br>decode: 128 tokens</th>
201
+ <th>Code Fixing<br>prefill: 1024 tokens<br>decode: 1024 tokens</th>
202
+ <th>RAG<br>prefill: 1024 tokens<br>decode: 128 tokens</th>
203
+ <th>Instruction Following<br>prefill: 256 tokens<br>decode: 128 tokens</th>
204
+ <th>Multi-turn Chat<br>prefill: 512 tokens<br>decode: 256 tokens</th>
205
+ <th>Large Summarization<br>prefill: 4096 tokens<br>decode: 512 tokens</th>
206
+ </tr>
207
+ <tr>
208
+ <td style="vertical-align: middle;" rowspan="3" >L40</td>
209
+ <td>granite-3.1-8b-instruct</td>
210
+ <td></td>
211
+ <td>25.1</td>
212
+ <td>3.2</td>
213
+ <td>25.3</td>
214
+ <td>3.2</td>
215
+ <td>3.2</td>
216
+ <td>6.3</td>
217
+ <td>13.4</td>
218
+ </tr>
219
+ <tr>
220
+ <td>granite-3.1-8b-instruct-FP8-dynamic<br>(this model)</td>
221
+ <td>1.47</td>
222
+ <td>16.8</td>
223
+ <td>2.2</td>
224
+ <td>17.1</td>
225
+ <td>2.2</td>
226
+ <td>2.1</td>
227
+ <td>4.2</td>
228
+ <td>9.3</td>
229
+ </tr>
230
+ <tr>
231
+ <td>granite-3.1-8b-instruct-quantized.w4a16</td>
232
+ <td>2.72</td>
233
+ <td>8.9</td>
234
+ <td>1.2</td>
235
+ <td>9.2</td>
236
+ <td>1.2</td>
237
+ <td>1.1</td>
238
+ <td>2.3</td>
239
+ <td>5.3</td>
240
+ </tr>
241
+ </table>
242
+
243
+
244
+ ### Multi-stream asynchronous performance (measured with vLLM version 0.6.6.post1)
245
+ <table>
246
+ <tr>
247
+ <td></td>
248
+ <td></td>
249
+ <td></td>
250
+ <th style="text-align: center;" colspan="7" >Maximum Throughput (Queries per Second)</th>
251
+ </tr>
252
+ <tr>
253
+ <th>GPU class</th>
254
+ <th>Model</th>
255
+ <th>Speedup</th>
256
+ <th>Code Completion<br>prefill: 256 tokens<br>decode: 1024 tokens</th>
257
+ <th>Docstring Generation<br>prefill: 768 tokens<br>decode: 128 tokens</th>
258
+ <th>Code Fixing<br>prefill: 1024 tokens<br>decode: 1024 tokens</th>
259
+ <th>RAG<br>prefill: 1024 tokens<br>decode: 128 tokens</th>
260
+ <th>Instruction Following<br>prefill: 256 tokens<br>decode: 128 tokens</th>
261
+ <th>Multi-turn Chat<br>prefill: 512 tokens<br>decode: 256 tokens</th>
262
+ <th>Large Summarization<br>prefill: 4096 tokens<br>decode: 512 tokens</th>
263
+ </tr>
264
+ <tr>
265
+ <td style="vertical-align: middle;" rowspan="3" >L40</td>
266
+ <td>granite-3.1-8b-instruct</td>
267
+ <td></td>
268
+ <td>1.4</td>
269
+ <td>7.8</td>
270
+ <td>1.1</td>
271
+ <td>6.2</td>
272
+ <td>15.5</td>
273
+ <td>6.0</td>
274
+ <td>0.7</td>
275
+ </tr>
276
+ <tr>
277
+ <td>granite-3.1-8b-instruct-FP8-dynamic<br>(this model)</td>
278
+ <td>1.12</td>
279
+ <td>2.1</td>
280
+ <td>7.4</td>
281
+ <td>1.3</td>
282
+ <td>5.9</td>
283
+ <td>15.3</td>
284
+ <td>6.9</td>
285
+ <td>0.8</td>
286
+ </tr>
287
+ <tr>
288
+ <td>granite-3.1-2b-instruct-quantized.w4a16</td>
289
+ <td>1.29</td>
290
+ <td>2.4</td>
291
+ <td>8.9</td>
292
+ <td>1.4</td>
293
+ <td>7.1</td>
294
+ <td>17.8</td>
295
+ <td>7.8</td>
296
+ <td>1.0</td>
297
+ </tr>
298
+ </table>