updated readme
Browse files
README.md
CHANGED
|
@@ -28,17 +28,25 @@ Evaluating how well the Fine tuned model works as a psychology assistant
|
|
| 28 |
evaluating simply on different fine-tuned models how the same phrase performs on different fine-tuned models
|
| 29 |
|
| 30 |
Models:
|
|
|
|
| 31 |
oskaralf/tiny_merged
|
|
|
|
| 32 |
TinyLlama (4bit)
|
| 33 |
- Fine-tuned on FineTome-100k
|
|
|
|
| 34 |
Llama3.2_1B_4bit
|
|
|
|
| 35 |
Llama 3.2-1B-Instruct (4bit)
|
|
|
|
| 36 |
- Fine-tuned on FineTome-100k
|
|
|
|
| 37 |
Llama1B_psych_v2_merged
|
|
|
|
| 38 |
- Llama 3.2-1B-Instruct (4bit)
|
| 39 |
- Fine-tuned on samhog/psychology10K
|
| 40 |
- Lora r=8
|
| 41 |
Llama1B_psych_v3_merged
|
|
|
|
| 42 |
- Llama 3.2-1B-Instruct (4bit)
|
| 43 |
- Fine-tuned on samhog/psychology10K
|
| 44 |
- Lora r=16
|
|
@@ -47,14 +55,26 @@ Llama1B_psych_v3_merged
|
|
| 47 |
- 4 epochs
|
| 48 |
|
| 49 |
To improve:
|
|
|
|
| 50 |
Model centric approach
|
|
|
|
| 51 |
change r=16 to higher dimension, for more complex LORA matrices, capturing more complex patterns
|
|
|
|
| 52 |
Using bigger model
|
|
|
|
| 53 |
Training more epochs
|
|
|
|
| 54 |
limited due to RAM and time constraint
|
|
|
|
| 55 |
Change learning rate
|
| 56 |
|
| 57 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 58 |
Comparison between different responses on a question:
|
| 59 |
|
| 60 |
“I have anxiety, what can I do?”
|
|
|
|
| 28 |
evaluating simply on different fine-tuned models how the same phrase performs on different fine-tuned models
|
| 29 |
|
| 30 |
Models:
|
| 31 |
+
|
| 32 |
oskaralf/tiny_merged
|
| 33 |
+
|
| 34 |
TinyLlama (4bit)
|
| 35 |
- Fine-tuned on FineTome-100k
|
| 36 |
+
|
| 37 |
Llama3.2_1B_4bit
|
| 38 |
+
|
| 39 |
Llama 3.2-1B-Instruct (4bit)
|
| 40 |
+
|
| 41 |
- Fine-tuned on FineTome-100k
|
| 42 |
+
|
| 43 |
Llama1B_psych_v2_merged
|
| 44 |
+
|
| 45 |
- Llama 3.2-1B-Instruct (4bit)
|
| 46 |
- Fine-tuned on samhog/psychology10K
|
| 47 |
- Lora r=8
|
| 48 |
Llama1B_psych_v3_merged
|
| 49 |
+
|
| 50 |
- Llama 3.2-1B-Instruct (4bit)
|
| 51 |
- Fine-tuned on samhog/psychology10K
|
| 52 |
- Lora r=16
|
|
|
|
| 55 |
- 4 epochs
|
| 56 |
|
| 57 |
To improve:
|
| 58 |
+
|
| 59 |
Model centric approach
|
| 60 |
+
|
| 61 |
change r=16 to higher dimension, for more complex LORA matrices, capturing more complex patterns
|
| 62 |
+
|
| 63 |
Using bigger model
|
| 64 |
+
|
| 65 |
Training more epochs
|
| 66 |
+
|
| 67 |
limited due to RAM and time constraint
|
| 68 |
+
|
| 69 |
Change learning rate
|
| 70 |
|
| 71 |
|
| 72 |
+
|
| 73 |
+
|
| 74 |
+
|
| 75 |
+
|
| 76 |
+
|
| 77 |
+
|
| 78 |
Comparison between different responses on a question:
|
| 79 |
|
| 80 |
“I have anxiety, what can I do?”
|