advy commited on
Commit
f98d555
·
verified ·
1 Parent(s): e10f5b4

Finetune on MentalChat16K - eval_loss: 0.8088

Browse files
Files changed (2) hide show
  1. README.md +1 -1
  2. training_metrics.json +46 -0
README.md CHANGED
@@ -16,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [google/gemma-2b](https://huggingface.co/google/gemma-2b) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 0.8446
20
 
21
  ## Model description
22
 
 
16
 
17
  This model is a fine-tuned version of [google/gemma-2b](https://huggingface.co/google/gemma-2b) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 0.7946
20
 
21
  ## Model description
22
 
training_metrics.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "gemma2b-mental-health",
3
+ "base_model": "google/gemma-2b",
4
+ "dataset": "ShenLab/MentalChat16K",
5
+ "lora_config": {
6
+ "rank": 24,
7
+ "alpha": 48,
8
+ "target_modules": [
9
+ "q_proj",
10
+ "k_proj",
11
+ "v_proj",
12
+ "o_proj",
13
+ "gate_proj",
14
+ "up_proj",
15
+ "down_proj"
16
+ ],
17
+ "dropout": 0.1
18
+ },
19
+ "training": {
20
+ "final_train_loss": 0.7568918577649376,
21
+ "total_steps": 2200,
22
+ "epochs": 4,
23
+ "learning_rate": 0.0002,
24
+ "per_device_batch_size": 3,
25
+ "gradient_accumulation": 2
26
+ },
27
+ "evaluation": {
28
+ "eval_loss": 0.8087502121925354,
29
+ "eval_runtime": 36.1206,
30
+ "eval_samples_per_second": 13.067,
31
+ "eval_steps_per_second": 4.374,
32
+ "epoch": 2.4666292765002806
33
+ },
34
+ "test_eval": {
35
+ "eval_loss": 0.7946024537086487,
36
+ "eval_runtime": 37.1332,
37
+ "eval_samples_per_second": 12.711,
38
+ "eval_steps_per_second": 4.255,
39
+ "epoch": 2.4666292765002806
40
+ },
41
+ "dataset_stats": {
42
+ "train_size": 5347,
43
+ "val_size": 472,
44
+ "test_size": 472
45
+ }
46
+ }