Update README.md
Browse files
README.md
CHANGED
|
@@ -98,8 +98,26 @@ print(my_list) # Output: ['a', 'b', 'c']
|
|
| 98 |
Note that in Python, lists are mutable, meaning you can add, remove, or modify elements after creating the list.
|
| 99 |
```
|
| 100 |
|
|
|
|
| 101 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 102 |
|
| 103 |
|
| 104 |
|
| 105 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 98 |
Note that in Python, lists are mutable, meaning you can add, remove, or modify elements after creating the list.
|
| 99 |
```
|
| 100 |
|
| 101 |
+
## Size Comparison
|
| 102 |
|
| 103 |
+
The table shows comparison VRAM requirements for loading and training
|
| 104 |
+
of FP16 Base Model and 4bit GPTQ quantized model with PEFT.
|
| 105 |
+
The value for base model referenced from [Model Memory Calculator](https://huggingface.co/docs/accelerate/main/en/usage_guides/model_size_estimator)
|
| 106 |
+
from HuggingFace
|
| 107 |
|
| 108 |
|
| 109 |
|
| 110 |
+
|
| 111 |
+
| Model | Total Size |
|
| 112 |
+
| ------------------------|-------------|
|
| 113 |
+
| Base Model | 28 GB |
|
| 114 |
+
| 4bitQuantized+PEFT | 5.21 GB |
|
| 115 |
+
|
| 116 |
+
|
| 117 |
+
## Acknowledgment
|
| 118 |
+
|
| 119 |
+
Thanks to [@AMerve Noyan](https://huggingface.co/blog/merve/quantization) for precise intro.
|
| 120 |
+
Thanks to [@HuggungFace Team](https://huggingface.co/blog/4bit-transformers-bitsandbytes) for the Blog.
|
| 121 |
+
Thanks to [@Meta](https://huggingface.co/meta-llama) for the Opensourced Model.
|
| 122 |
+
|
| 123 |
+
|