Replace with conservative adapters (LR: 1e-4, 500 iters) - 100% test pass rate
Browse files- 0000100_adapters.safetensors +2 -2
- 0000200_adapters.safetensors +2 -2
- 0000300_adapters.safetensors +2 -2
- 0000400_adapters.safetensors +3 -0
- 0000500_adapters.safetensors +3 -0
- adapter_config.json +5 -5
- adapters.safetensors +3 -0
0000100_adapters.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8b3bc5d89ca37687efc247308185d7d85f1e260cbdf86cc33152f87a25538855
|
| 3 |
+
size 3411329
|
0000200_adapters.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:679916c91648bf3dc1bbda797cf55653c71aece2429d71cbada57f677c717b8a
|
| 3 |
+
size 3411329
|
0000300_adapters.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2be5932e764c9707d258639bd4297807132c198ce56a1b9a493f7d46ad011cf9
|
| 3 |
+
size 3411329
|
0000400_adapters.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c01ddcc4203753bc387c299883048c6d77c37378dbf0768b3518adad53a3594e
|
| 3 |
+
size 3411329
|
0000500_adapters.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:78097e4cccdf6814c66d0c50fd2318772ef682e44d20eb3fcf087204b9018832
|
| 3 |
+
size 3411329
|
adapter_config.json
CHANGED
|
@@ -1,12 +1,12 @@
|
|
| 1 |
{
|
| 2 |
-
"adapter_path": "serum_lora_adapters/
|
| 3 |
"batch_size": 8,
|
| 4 |
"config": null,
|
| 5 |
-
"data": "serum_lora_adapters/
|
| 6 |
"fine_tune_type": "lora",
|
| 7 |
"grad_checkpoint": false,
|
| 8 |
-
"iters":
|
| 9 |
-
"learning_rate": 0.
|
| 10 |
"lora_parameters": {
|
| 11 |
"rank": 8,
|
| 12 |
"dropout": 0.0,
|
|
@@ -16,7 +16,7 @@
|
|
| 16 |
"mask_prompt": false,
|
| 17 |
"max_seq_length": 2048,
|
| 18 |
"model": "NousResearch/Hermes-2-Pro-Mistral-7B",
|
| 19 |
-
"num_layers":
|
| 20 |
"optimizer": "adam",
|
| 21 |
"optimizer_config": {
|
| 22 |
"adam": {},
|
|
|
|
| 1 |
{
|
| 2 |
+
"adapter_path": "serum_lora_adapters/conservative_adapters",
|
| 3 |
"batch_size": 8,
|
| 4 |
"config": null,
|
| 5 |
+
"data": "serum_lora_adapters/training_data_v2",
|
| 6 |
"fine_tune_type": "lora",
|
| 7 |
"grad_checkpoint": false,
|
| 8 |
+
"iters": 500,
|
| 9 |
+
"learning_rate": 0.0001,
|
| 10 |
"lora_parameters": {
|
| 11 |
"rank": 8,
|
| 12 |
"dropout": 0.0,
|
|
|
|
| 16 |
"mask_prompt": false,
|
| 17 |
"max_seq_length": 2048,
|
| 18 |
"model": "NousResearch/Hermes-2-Pro-Mistral-7B",
|
| 19 |
+
"num_layers": 8,
|
| 20 |
"optimizer": "adam",
|
| 21 |
"optimizer_config": {
|
| 22 |
"adam": {},
|
adapters.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:78097e4cccdf6814c66d0c50fd2318772ef682e44d20eb3fcf087204b9018832
|
| 3 |
+
size 3411329
|