SwastikM commited on
Commit
042e8c3
·
verified ·
1 Parent(s): fddfb41

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -1
README.md CHANGED
@@ -98,8 +98,26 @@ print(my_list) # Output: ['a', 'b', 'c']
98
  Note that in Python, lists are mutable, meaning you can add, remove, or modify elements after creating the list.
99
  ```
100
 
 
101
 
 
 
 
 
102
 
103
 
104
 
105
- COMING SOON
 
 
 
 
 
 
 
 
 
 
 
 
 
 
98
  Note that in Python, lists are mutable, meaning you can add, remove, or modify elements after creating the list.
99
  ```
100
 
101
+ ## Size Comparison
102
 
103
+ The table shows comparison VRAM requirements for loading and training
104
+ of FP16 Base Model and 4bit GPTQ quantized model with PEFT.
105
+ The value for base model referenced from [Model Memory Calculator](https://huggingface.co/docs/accelerate/main/en/usage_guides/model_size_estimator)
106
+ from HuggingFace
107
 
108
 
109
 
110
+
111
+ | Model | Total Size |
112
+ | ------------------------|-------------|
113
+ | Base Model | 28 GB |
114
+ | 4bitQuantized+PEFT | 5.21 GB |
115
+
116
+
117
+ ## Acknowledgment
118
+
119
+ Thanks to [@AMerve Noyan](https://huggingface.co/blog/merve/quantization) for precise intro.
120
+ Thanks to [@HuggungFace Team](https://huggingface.co/blog/4bit-transformers-bitsandbytes) for the Blog.
121
+ Thanks to [@Meta](https://huggingface.co/meta-llama) for the Opensourced Model.
122
+
123
+