File size: 11,500 Bytes
35f0bf2
fd8b0db
 
35f0bf2
cc4e52d
35f0bf2
b93ba77
8520a5f
 
 
 
 
 
 
 
 
 
 
 
 
 
35f0bf2
 
cc4e52d
35f0bf2
 
a3fab31
35f0bf2
ca2d3bb
 
cc4e52d
35f0bf2
cc4e52d
 
 
 
35f0bf2
cc4e52d
35f0bf2
ca2d3bb
 
cc4e52d
35f0bf2
a3fab31
cc4e52d
35f0bf2
ca2d3bb
 
cc4e52d
 
 
 
 
 
35f0bf2
 
cc4e52d
35f0bf2
ca2d3bb
 
cc4e52d
 
 
 
 
 
 
 
 
ca2d3bb
 
cc4e52d
 
 
 
 
 
 
35f0bf2
ca2d3bb
 
a3fab31
35f0bf2
cc4e52d
 
35f0bf2
 
 
 
 
 
 
 
99a17c0
c008ff9
35f0bf2
 
 
 
 
 
 
 
cc4e52d
35f0bf2
 
 
acf0220
35f0bf2
 
 
 
 
4ef69d5
46a89e6
742322e
 
46a89e6
 
742322e
 
46a89e6
 
4ef69d5
 
 
 
 
 
35f0bf2
 
e0a26e9
35f0bf2
 
 
 
 
 
cc4e52d
35f0bf2
 
 
 
 
 
 
 
 
ca2d3bb
 
a3fab31
35f0bf2
 
cc4e52d
 
 
 
 
 
ca2d3bb
 
cc4e52d
 
 
 
 
 
 
 
ca2d3bb
 
a3fab31
cc4e52d
 
 
 
 
 
 
ca2d3bb
 
a3fab31
cc4e52d
 
 
 
 
ca2d3bb
 
a3fab31
cc4e52d
 
 
 
 
 
 
 
ca2d3bb
 
cc4e52d
 
 
 
 
 
 
 
 
 
 
ca2d3bb
 
a3fab31
cc4e52d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9eb8ae0
 
ca2d3bb
 
a3fab31
9eb8ae0
cc4e52d
 
 
 
 
 
 
ca2d3bb
cc4e52d
ca2d3bb
cc4e52d
 
ca2d3bb
cc4e52d
 
 
 
 
 
 
 
 
 
 
 
ca2d3bb
35f0bf2
 
cc4e52d
 
 
 
 
 
 
ca2d3bb
 
cc4e52d
a3fab31
cc4e52d
 
 
 
 
 
 
 
 
 
 
 
 
35f0bf2
cc4e52d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
35f0bf2
 
ca2d3bb
 
a3fab31
35f0bf2
cc4e52d
 
 
ca2d3bb
 
cc4e52d
35f0bf2
cc4e52d
35f0bf2
ca2d3bb
 
cc4e52d
 
 
35f0bf2
cc4e52d
 
 
35f0bf2
cc4e52d
 
 
35f0bf2
cc4e52d
35f0bf2
ca2d3bb
 
cc4e52d
 
 
35f0bf2
 
ca2d3bb
 
a3fab31
35f0bf2
 
cc4e52d
35f0bf2
cc4e52d
 
ca2d3bb
 
cc4e52d
 
 
 
 
 
 
ca2d3bb
 
cc4e52d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ca2d3bb
 
cc4e52d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ca2d3bb
 
cc4e52d
 
 
35f0bf2
 
 
 
ca2d3bb
 
cc4e52d
35f0bf2
cc4e52d
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
---
license: other
license_name: thestageai-elastic
base_model:
- meta-llama/Llama-3.1-8B-Instruct
base_model_relation: quantized
pipeline_tag: text-generation
language:
- zho
- eng
- fra
- spa
- por
- deu
- ita
- rus
- jpn
- kor
- vie
- tha
- ara
---

# Elastic model: Llama-3.1-8B-Instruct


## Overview

---

ElasticModels are the models produced by TheStage AI ANNA: Automated Neural Networks Accelerator. ANNA allows you to control model size, latency and quality with a simple slider movement, routing different compression algorithms to different layers. For each model, we have produced a series of optimized models:

- **XL**: Mathematically equivalent neural network, optimized with our DNN compiler.
- **L**: Near lossless model, with less than 1% degradation obtained on corresponding benchmarks.
- **M**: Faster model, with accuracy degradation less than 1.5%.
- **S**: The fastest model, with accuracy degradation less than 2%.

Models can be accessed via TheStage AI Python SDK: ElasticModels, or deployed as Docker containers with REST API endpoints (see Deploy section).

## Installation

---


### System Requirements

---

| **Property**| **Value** |
 | ---  | ---  |
| **GPU** | L40s, RTX 5090, H100, B200 |
| **Python Version** | 3.10-3.12 |
| **CPU** | Intel/AMD x86_64 |
| **CUDA Version** | 12.9+ |


### TheStage AI Access token setup

---

Install TheStage AI CLI and setup API token:

```bash
pip install thestage
thestage config set --access-token <YOUR_ACCESS_TOKEN>
```

### ElasticModels installation

---

Install TheStage Elastic Models package:

```bash
pip install 'thestage-elastic-models[nvidia,cudnn]' \
    --extra-index-url https://thestage.jfrog.io/artifactory/api/pypi/pypi-thestage-ai-production/simple
pip install --force-reinstall --no-deps nvidia-cudnn-frontend==1.18.0
```

## Usage example

---


Elastic Models provides the same interface as HuggingFace Transformers. Here is an example of how to use the Llama-3.1-8B-Instruct model:

```python
import torch
from transformers import AutoTokenizer
from elastic_models.transformers import AutoModelForCausalLM

# Currently we require to have your HF token
# as we use original weights for part of layers and
# model configuration as well
model_name = "meta-llama/Llama-3.1-8B-Instruct"
hf_token = ''
device = torch.device("cuda")

# Create mode
tokenizer = AutoTokenizer.from_pretrained(
    model_name, token=hf_token
)
model = AutoModelForCausalLM.from_pretrained(
    model_name,
    token=hf_token,
    torch_dtype=torch.bfloat16,
    attn_implementation="sdpa",
    mode='S'
).to(device)
model.generation_config.pad_token_id = tokenizer.eos_token_id

# Inference simple as transformers library
prompt = "Describe basics of DNNs quantization."
messages = [
  {
    "role": "system",
    "content": "You are a search bot, answer on user text queries."
  },
  {
    "role": "user",
    "content": prompt
  }
]

chat_prompt = tokenizer.apply_chat_template(
    messages, add_generation_prompt=True, tokenize=False
)

inputs = tokenizer(chat_prompt, return_tensors="pt")
inputs.to(device)

with torch.inference_mode():
    generate_ids = model.generate(**inputs, max_length=500)

input_len = inputs['input_ids'].shape[1]
generate_ids = generate_ids[:, input_len:]
output = tokenizer.batch_decode(
    generate_ids,
    skip_special_tokens=True,
    clean_up_tokenization_spaces=False
)[0]

# Validate answer
print(f"# Q:\n{prompt}\n")
print(f"# A:\n{output}\n")
```


## Quality Benchmarks

---


We have used the `lm_eval` library to validate the models. For each model size (S, M, L, XL), we have run the following tasks: MMLU, PIQA, Arc Challenge, Winogrande.

![Quality Benchmarking](https://cdn.thestage.ai/production/cms_file_upload/1773422713-7d51617f-e70a-41db-95f9-abd0d9ff338f/Elastic_Llama_3.1_8B_Instruct_MMLU.png)

### Quality Benchmark Results

---

| **Metric/Model Size**| **S**| **M**| **L**| **XL**| **Original**| **W8A8, int8** |
 | ---  | ---  | ---  | ---  | ---  | ---  | ---  |
| **MMLU** | 67.4 | 68.1 | 68.3 | 68.5 | 68.4 | 24.3 |
| **PIQA** | 79.8 | 80.2 | 80.1 | 79.9 | 80.0 | 64.6 |
| **Arc Challenge** | 55.1 | 54.6 | 54.7 | 55.6 | 55.5 | 29.6 |
| **Winogrande** | 73.7 | 73.6 | 73.7 | 74.0 | 74.0 | 62.8 |


## Datasets

---


- **MMLU**: Measures model performance on a diverse set of multiple-choice questions covering various academic subjects, testing general knowledge and reasoning.
- **PIQA**: Evaluates physical commonsense reasoning by asking the model to choose the most plausible solution to everyday physical problems.
- **Arc Challenge**: Assesses scientific and factual reasoning using challenging multiple-choice questions from the AI2 Reasoning Challenge dataset.
- **Winogrande**: Tests commonsense understanding and pronoun resolution through sentences requiring the model to identify the correct referent.

## Metrics

---


- **Accuracy**: Accuracy measures the proportion of model predictions that exactly match the correct answers across evaluation tasks.


## Latency Benchmarks

---


We measured TPS (tokens per second) for each model size using 100 input tokens and 300 output tokens.

![Latency Benchmarking](https://cdn.thestage.ai/production/cms_file_upload/1773422728-414fdb22-5c04-44a6-8686-0602a4293e88/Elastic_Llama_3.1_8B_Instruct_latency.png)

### Latency Benchmark Results

---

Tokens per second for different model sizes on various GPUs.

| **GPU/Model Size**| **S**| **M**| **L**| **XL**| **Original**| **W8A8_int8** |
 | ---  | ---  | ---  | ---  | ---  | ---  | ---  |
| **H100** | 189 | 168 | 156 | 134 | 60 | 191 |
| **L40s** | 72 | 63 | 56 | 45 | 37 | 77 |
| **B200** | 239 | 236 | 207 | 199 | 100 | N/A |
| **GeForce RTX 5090** | 143 | N/A | N/A | N/A | 60 | N/A |
| **GeForce RTX 4090** | 95 | N/A | N/A | N/A | 41 | N/A |


## Benchmarking Methodology

---


The benchmarking was performed on a single GPU with a batch size of 1. Each model was run for 10 iterations, and the average latency was calculated.

> **Algorithm summary:**
> 1. Load the Llama-3.1-8B-Instruct model with the specified size (S, M, L, XL, original).
> 2. Move the model to the GPU.
> 3. Prepare a sample prompt for text generation.
> 4. Run the model for a number of iterations (e.g., 10) and measure the time taken for each iteration. On each iteration:
>    - Synchronize the GPU to flush any previous operations.
>    - Record the start time.
>    - Generate the text using the model.
>    - Synchronize the GPU again.
>    - Record the end time and calculate the TTFT and TPS for that iteration.
> 5. Calculate the average TTFT and TPS over all iterations.


## Serving with Docker Image

---


For serving with Nvidia GPUs, we provide ready-to-go Docker containers with OpenAI-compatible API endpoints.
Using our containers you can set up an inference endpoint on any desired cloud/serverless providers as well as on-premise servers.
You can also use this container to run inference through TheStage AI platform.

### Prebuilt image from ECR

---

Pull docker image and start inference container:

```bash
docker pull public.ecr.aws/i3f7g5s7/thestage/elastic-models:0.2.0-llm-24.09c
```
```bash
docker run --rm -ti \
  --name serving_thestage_model \
  -p 8000:80 \
  -e AUTH_TOKEN=<AUTH_TOKEN> \
  -e MODEL_REPO=meta-llama/Llama-3.1-8B-Instruct \
  -e MODEL_SIZE=<MODEL_SIZE> \
  -e MODEL_BATCH=<MAX_BATCH_SIZE> \
  -e HUGGINGFACE_ACCESS_TOKEN=<HUGGINGFACE_ACCESS_TOKEN> \
  -e THESTAGE_AUTH_TOKEN=<THESTAGE_ACCESS_TOKEN> \
  -v /mnt/hf_cache:/root/.cache/huggingface \
  public.ecr.aws/i3f7g5s7/thestage/elastic-models:0.2.0-llm-24.09c
```

| **Parameter**              | **Description**                                                                                      |
|----------------------------|------------------------------------------------------------------------------------------------------|
| `<MODEL_SIZE>`             | Available: S, M, L, XL.                                                                              |
| `<MAX_BATCH_SIZE>`         | Maximum batch size to process in parallel.                                                           |
| `<HUGGINGFACE_ACCESS_TOKEN>` | Hugging Face access token.                                                                         |
| `<THESTAGE_ACCESS_TOKEN>`  | TheStage token generated on the platform (Profile -> Access tokens).                                 |
| `<AUTH_TOKEN>`             | Token for endpoint authentication. You can set it to any random string; it must match the value used by the client. |

## Invocation

---


You can invoke the endpoint using CURL as follows:

```bash
curl -X POST 'http://127.0.0.1:8000/v1/chat/completions' \
    -H 'Authorization: Bearer 123' \
    -H 'Content-Type: application/json' \
    -H "X-Model-Name: llama-3-1-8b-instruct-<MODEL_SIZE>-bs<MAX_BATCH_SIZE>-paged" \
    -d '{
        "messages":[{"role":"user","content":"Define AI"}]
    }'
```

Or using OpenAI python client:

```python
import os, base64, pathlib, json
from openai import OpenAI

BASE_URL = "http://<your_ip>/v1"
API_KEY  = "123"
MODEL    = "llama-3-1-8b-instruct-<MODEL_SIZE>-bs<MAX_BATCH_SIZE>-paged"

client = OpenAI(
    api_key=API_KEY,
    base_url=BASE_URL,
    default_headers={"X-Model-Name": MODEL}
)

response = client.chat.completions.create(
    model=MODEL,
    messages=[
        {"role": "user", "content": "Define AI"}
    ]
)

print(response.choices[0].message.content)
```

## Endpoint Parameters

---


### Method

---

> **POST** `/v1/chat/completions`

### Header Parameters

---

> `Authorization`: `string`
>
> Bearer token for authentication. Should match the `AUTH_TOKEN` set during container startup.

> `Content-Type`: `string`
>
> Must be set to `application/json`.

> `X-Model-Name`: `string`
>
> Specifies the model to use for generation. Format: `llama-3-1-8b-instruct-<size>-bs<batch_size>`, where `<size>` is one of `S`, `M`, `L`, `XL`, `original` and `<batch_size>` is the maximum batch size configured during container startup.

### Input Body

---

> `messages` : `string`
>
> The input text prompt.


## Deploy on Modal

---


For more details please use the tutorial [Modal deployment](https://docs.thestage.ai/tutorials/source/modal_thestage.html)

### Clone modal serving code

---

```shell
git clone https://github.com/TheStageAI/ElasticModels.git
cd ElasticModels/examples/modal
```

### Configuration of environment variables

---

Set your environment variables in `modal_serving.py`:

```python
# modal_serving.py

ENVS = {
    "MODEL_REPO": "meta-llama/Llama-3.1-8B-Instruct",
    "MODEL_BATCH": "4",
    "THESTAGE_AUTH_TOKEN": "",
    "HUGGINGFACE_ACCESS_TOKEN": "",
    "PORT": "80",
    "PORT_HEALTH": "80",
    "HF_HOME": "/cache/huggingface",
}
```

### Configuration of GPUs

---

Set your desired GPU type and autoscaling variables in `modal_serving.py`:

```python
# modal_serving.py

@app.function(
    image=image,
    gpu="B200",
    min_containers=8,
    max_containers=8,
    timeout=10000,
    ephemeral_disk=600 * 1024,
    volumes={"/opt/project/.cache": HF_CACHE},
    startup_timeout=60*20
)
@modal.web_server(
    80,
    label="meta-llama/Llama-3.1-8B-Instruct-test",
    startup_timeout=60*20
)
def serve():
    pass
```

### Run serving

---

```shell
modal serve modal_serving.py
```


## Links

---

* __Platform__: [app.thestage.ai](https://app.thestage.ai)
* __Subscribe for updates__: [TheStageAI X](https://x.com/TheStageAI)
* __Contact email__: contact@thestage.ai