JoshuaFreeman commited on
Commit
10e0d6e
·
verified ·
1 Parent(s): 4b422ad

Upload LoRA model and probe head for run first gemma 2B with last layer

Browse files
value_head_probes/first gemma 2B with last layer/results.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eval_metrics": {"eval_accuracy": 0.5481849825957236, "eval_precision": 0.6286564204263758, "eval_recall": 0.2504938759383643, "eval_f1": 0.35824268964543016, "eval_auc": 0.599054203344184, "eval_lm_loss": 8.58864703611894, "eval_probe_loss": 0.8011899159713225, "eval_sparsity": 0.17750355113636362, "epoch": 2.0}, "train_metrics": {"train_accuracy": 0.9460922712516787, "train_precision": 0.4116284641801773, "train_recall": 0.3297504798464491, "train_f1": 0.36616816454122414, "train_auc": 0.8816274559252752, "train_lm_loss": 8.537573195726443, "train_probe_loss": 0.8766092875064948, "train_sparsity": 0.1735576923076923, "epoch": 2.0}}
value_head_probes/first gemma 2B with last layer/training_config.json ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_name": "google/gemma-2b",
3
+ "data_model_name": "meta-llama/Meta-Llama-3.1-8B-Instruct",
4
+ "layer": 17,
5
+ "adapter_dir": "/workspace/hallucination_detection/value_head_probes/first gemma 2B with last layer",
6
+ "train_split": 0.8,
7
+ "dtype": "torch.bfloat16",
8
+ "overwrite_output_dir": true,
9
+ "max_steps": -1,
10
+ "num_train_epochs": 2,
11
+ "per_device_train_batch_size": 2,
12
+ "per_device_eval_batch_size": 2,
13
+ "lambda_lm": 0.1,
14
+ "logging_steps": 20,
15
+ "eval_steps": 100,
16
+ "overfit_mode": false,
17
+ "lora_layers": [
18
+ 0,
19
+ 1,
20
+ 2,
21
+ 3,
22
+ 4,
23
+ 5,
24
+ 6,
25
+ 7,
26
+ 8,
27
+ 9,
28
+ 10,
29
+ 11,
30
+ 12,
31
+ 13,
32
+ 14,
33
+ 15,
34
+ 16,
35
+ 17,
36
+ 18,
37
+ 19,
38
+ 20,
39
+ 21,
40
+ 22,
41
+ 23,
42
+ 24,
43
+ 25,
44
+ 26,
45
+ 27,
46
+ 28,
47
+ 29,
48
+ 30,
49
+ 31
50
+ ],
51
+ "lora_r": 32,
52
+ "lora_alpha": 64,
53
+ "lora_dropout": 0.05,
54
+ "upload_to_hf": true,
55
+ "load_from_hf": false,
56
+ "probe_threshold": 0.5,
57
+ "hf_repo": "obalcells/labeled-entity-facts",
58
+ "dataset_name": "longfact",
59
+ "max_length": 1200,
60
+ "default_ignore": false,
61
+ "pos_weight": 10.0,
62
+ "neg_weight": 10.0,
63
+ "ignore_window_size": 5,
64
+ "shuffle": true,
65
+ "seed": 42,
66
+ "run_id": "first gemma 2B with last layer"
67
+ }