Envoid commited on
Commit
235ad2d
·
1 Parent(s): 9edf356

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -0
README.md CHANGED
@@ -1,3 +1,22 @@
1
  ---
2
  license: cc-by-nc-4.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: cc-by-nc-4.0
3
  ---
4
+ This is my first attempt at a qLoRA on Mixtral using the same raw text dataset as the Dendrite model on Mixtral 8x7B Instruct v0.1 with the following notable settings:
5
+ ```
6
+ lora_r 256
7
+ lora_alpha 256
8
+ ..._max_len 256
9
+ learning_rate 0.000001
10
+ num_train_epochs 2
11
+ ```
12
+
13
+ The output is a little less dry but it still maintains the full level of functionality you would expect out of Mixtral instruct. And it still responds to the
14
+
15
+ ```
16
+ [INST]
17
+ Do a thing
18
+ [/INST]
19
+ ```
20
+
21
+ format.
22
+ It's not a massive change in the output but I do plan to run a similar but larger data set with more epochs and a higher learning rate.