LisaMegaWatts commited on
Commit
22b16f0
·
verified ·
1 Parent(s): b6013fb

Grammar expert LoRA for SymbioSLM (CoLA acc=0.300)

Browse files
Files changed (2) hide show
  1. experiment_config.json +11 -11
  2. lora_state.pt +2 -2
experiment_config.json CHANGED
@@ -1,26 +1,26 @@
1
  {
2
  "model": "SymbioSLM",
3
- "model_params": 4261650,
4
  "d_model": 256,
5
- "n_layers": 6,
6
  "organelles": [
7
  "causal_conv",
8
  "monarch",
9
  "long_conv"
10
  ],
11
- "n_monarch_heads": 8,
12
- "checkpoint_loaded": false,
13
- "lora_rank": 16,
14
- "lora_alpha": 32.0,
15
  "lora_targets": [
16
  "w1",
17
  "w2"
18
  ],
19
- "lora_trainable_params": 2468116,
20
- "cola_test_accuracy": 0.606,
21
- "base_perplexity": 2045.621205099146,
22
- "lora_perplexity": 2051.0400660540213,
23
  "generations": 25,
24
- "gelation_step": 6,
25
  "date": "20260301"
26
  }
 
1
  {
2
  "model": "SymbioSLM",
3
+ "model_params": 5052696,
4
  "d_model": 256,
5
+ "n_layers": 8,
6
  "organelles": [
7
  "causal_conv",
8
  "monarch",
9
  "long_conv"
10
  ],
11
+ "n_monarch_heads": 1,
12
+ "checkpoint_loaded": true,
13
+ "lora_rank": 4,
14
+ "lora_alpha": 8.0,
15
  "lora_targets": [
16
  "w1",
17
  "w2"
18
  ],
19
+ "lora_trainable_params": 57858,
20
+ "cola_test_accuracy": 0.3,
21
+ "base_perplexity": 133287.60955412162,
22
+ "lora_perplexity": 255971.62295575335,
23
  "generations": 25,
24
+ "gelation_step": null,
25
  "date": "20260301"
26
  }
lora_state.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9745eff983fbdf700e700523409fe0248adf9976744e41e06fe1b3f11362ada6
3
- size 696117
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19f77bf3554c694739e62bfc33fea6182d88aa96a63c009d2306747d5137576d
3
+ size 239605