Update README.md
Browse files
README.md
CHANGED
|
@@ -1,19 +1,3 @@
|
|
| 1 |
-
# Quantization
|
| 2 |
-
Created with [lambda-quant](https://github.com/LambdaLabsML/lambda-quant/tree/f97108fe4a9ee061a7b969b23a9605a6d561863d) on `Python 3.10.12 (main, Nov 6 2024, 20:22:13) [GCC 11.4.0]`
|
| 3 |
-
|
| 4 |
-
Base Model: [meta-llama/Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct)
|
| 5 |
-
|
| 6 |
-
Quantized using [gptqmodel==2.1.0](https://github.com/ModelCloud/GPTQModel)
|
| 7 |
-
|
| 8 |
-
Steps to create:
|
| 9 |
-
1. `git clone https://github.com/LambdaLabsML/lambda-quant`
|
| 10 |
-
2. `git checkout f97108fe4a9ee061a7b969b23a9605a6d561863d`
|
| 11 |
-
3. `python quantize.py -m meta-llama/Llama-3.1-8B-Instruct -q GPTQ-Int4`
|
| 12 |
-
## Evaluation
|
| 13 |
-
TODO
|
| 14 |
-
## Benchmarks
|
| 15 |
-
TODO
|
| 16 |
-
# Base Model README.md
|
| 17 |
---
|
| 18 |
language:
|
| 19 |
- en
|
|
@@ -204,6 +188,26 @@ extra_gated_description: The information you provide will be collected, stored,
|
|
| 204 |
and shared in accordance with the [Meta Privacy Policy](https://www.facebook.com/privacy/policy/).
|
| 205 |
extra_gated_button_content: Submit
|
| 206 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 207 |
|
| 208 |
## Model Information
|
| 209 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
language:
|
| 3 |
- en
|
|
|
|
| 188 |
and shared in accordance with the [Meta Privacy Policy](https://www.facebook.com/privacy/policy/).
|
| 189 |
extra_gated_button_content: Submit
|
| 190 |
---
|
| 191 |
+
# Quantization
|
| 192 |
+
Created with [lambda-quant](https://github.com/LambdaLabsML/lambda-quant/tree/f97108fe4a9ee061a7b969b23a9605a6d561863d) on `Python 3.10.12 (main, Nov 6 2024, 20:22:13) [GCC 11.4.0]`
|
| 193 |
+
|
| 194 |
+
Base Model: [meta-llama/Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct)
|
| 195 |
+
|
| 196 |
+
Quantized using [gptqmodel==2.1.0](https://github.com/ModelCloud/GPTQModel)
|
| 197 |
+
|
| 198 |
+
Steps to create:
|
| 199 |
+
1. `git clone https://github.com/LambdaLabsML/lambda-quant`
|
| 200 |
+
2. `git checkout f97108fe4a9ee061a7b969b23a9605a6d561863d`
|
| 201 |
+
3. `python quantize.py -m meta-llama/Llama-3.1-8B-Instruct -q GPTQ-Int4`
|
| 202 |
+
|
| 203 |
+
## Evaluation
|
| 204 |
+
TODO
|
| 205 |
+
|
| 206 |
+
## Benchmarks
|
| 207 |
+
TODO
|
| 208 |
+
|
| 209 |
+
# Base Model README.md
|
| 210 |
+
|
| 211 |
|
| 212 |
## Model Information
|
| 213 |
|