bapinero commited on
Commit
5fcde2e
·
verified ·
1 Parent(s): 370d29d

Replace with conservative adapters (LR: 1e-4, 500 iters) - 100% test pass rate

Browse files
0000100_adapters.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa6ec1c44520c102de41a223ceca79bbef6528dbddbccebc979f0238bf2f51ca
3
- size 6822661
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b3bc5d89ca37687efc247308185d7d85f1e260cbdf86cc33152f87a25538855
3
+ size 3411329
0000200_adapters.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:974dc971289eba9df04a273689a18a74f580f40f4e4e6bb9245ae3899b7a2d4a
3
- size 6822661
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:679916c91648bf3dc1bbda797cf55653c71aece2429d71cbada57f677c717b8a
3
+ size 3411329
0000300_adapters.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6154b505f4e55979a12eaa089d1d25a5a842115fcf58fa6d97a456fa310ea97e
3
- size 6822661
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2be5932e764c9707d258639bd4297807132c198ce56a1b9a493f7d46ad011cf9
3
+ size 3411329
0000400_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c01ddcc4203753bc387c299883048c6d77c37378dbf0768b3518adad53a3594e
3
+ size 3411329
0000500_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78097e4cccdf6814c66d0c50fd2318772ef682e44d20eb3fcf087204b9018832
3
+ size 3411329
adapter_config.json CHANGED
@@ -1,12 +1,12 @@
1
  {
2
- "adapter_path": "serum_lora_adapters/production_adapters",
3
  "batch_size": 8,
4
  "config": null,
5
- "data": "serum_lora_adapters/training_data",
6
  "fine_tune_type": "lora",
7
  "grad_checkpoint": false,
8
- "iters": 300,
9
- "learning_rate": 0.0003,
10
  "lora_parameters": {
11
  "rank": 8,
12
  "dropout": 0.0,
@@ -16,7 +16,7 @@
16
  "mask_prompt": false,
17
  "max_seq_length": 2048,
18
  "model": "NousResearch/Hermes-2-Pro-Mistral-7B",
19
- "num_layers": 16,
20
  "optimizer": "adam",
21
  "optimizer_config": {
22
  "adam": {},
 
1
  {
2
+ "adapter_path": "serum_lora_adapters/conservative_adapters",
3
  "batch_size": 8,
4
  "config": null,
5
+ "data": "serum_lora_adapters/training_data_v2",
6
  "fine_tune_type": "lora",
7
  "grad_checkpoint": false,
8
+ "iters": 500,
9
+ "learning_rate": 0.0001,
10
  "lora_parameters": {
11
  "rank": 8,
12
  "dropout": 0.0,
 
16
  "mask_prompt": false,
17
  "max_seq_length": 2048,
18
  "model": "NousResearch/Hermes-2-Pro-Mistral-7B",
19
+ "num_layers": 8,
20
  "optimizer": "adam",
21
  "optimizer_config": {
22
  "adam": {},
adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78097e4cccdf6814c66d0c50fd2318772ef682e44d20eb3fcf087204b9018832
3
+ size 3411329