RafMuz commited on
Commit
c30aa09
Β·
1 Parent(s): c1c2acd

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -0
README.md ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ tags:
5
+ - code
6
+ ---
7
+ # About
8
+
9
+ Hi, this is the Readme.
10
+
11
+ This Model was created as a study experiment, to re-create alpaca on my end.
12
+ It uses the gururise/AlpacaDataCleaned Dataset ( From April 7 )
13
+
14
+ ---
15
+ # Specifications
16
+
17
+ **Base Model**:
18
+   LLaMA 7B
19
+
20
+ **Training Parameters**:
21
+   Micro_Batch_Size = 8
22
+   Batch_Size = 128
23
+   Gradient_Accumulation_Steps = Batch_Size / Micro_Batch_Size   # ( 0.0625 )
24
+   Epochs = 2
25
+   Learning_Rate = 2e-5
26
+   Cutoff_Len = 256   # This ( 256 ) accounts for about 96% of all data
27
+   Lora_R = 4
28
+   Lora_Alpha = 16
29
+   Lora_Dropout = 0.05
30
+
31
+  
32
+  
33
+ # Files
34
+   adapter_model.binβ€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…*# This is the Fine-tuned Weights that goes over the base LLaMA Model.*
35
+   adapter_config.binβ€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…*# This is Config File for the adapter_model file.*
36
+
37
+   consolidated.00.pthβ€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…*# This File is the Base Model File ( LLaMA 7B ), merged with the fine-tuned weights ( adapter_model.bin ).*
38
+   tokenizer.modelβ€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…*# This is the tokenizer file, it converts the input text ( prompt ) to tokens that the NN can understand.*
39
+   params.jsonβ€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…*# Parameters of the Model.*
40
+
41
+   ggml_model_f16.binβ€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…*# This is the same model ( consolidated.00.pth ), but now it's in 'ggml f16' format. We need this format to quantize it with llama.cpp.*
42
+   **llama-hf-7b**β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…β€…*# This folder contains the same model ( consolidated.00.pth ), but now it's in 'huggingface' format. We need this format to quantize it with GPTQ.*
43
+
44
+   **quantized-model**:
45
+     ggml-model-q4_0.binβ€…β€…β€…*# This is the 4-bit Quantized Model by llama.cpp, I found this to be better than GPTQ.*
46
+     llama7b-4bit-128g.ptβ€…β€…β€…β€…*# This is the Quantized Model by GPTQ. It takes longer to train and gives worse results compared to llama.cpp, but it does have a ( 7.6% ) smaller file size.*
47
+
48
+