EdBerg commited on
Commit
6745b37
·
verified ·
1 Parent(s): 71150d2

EdBerg/outputs3

Browse files
README.md CHANGED
@@ -44,7 +44,7 @@ The following hyperparameters were used during training:
44
  - optimizer: Use paged_adamw_8bit with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
45
  - lr_scheduler_type: linear
46
  - lr_scheduler_warmup_steps: 2
47
- - training_steps: 2000
48
  - mixed_precision_training: Native AMP
49
 
50
  ### Training results
@@ -56,5 +56,5 @@ The following hyperparameters were used during training:
56
  - PEFT 0.13.3.dev0
57
  - Transformers 4.46.1
58
  - Pytorch 2.5.0+cu121
59
- - Datasets 3.0.2
60
  - Tokenizers 0.20.1
 
44
  - optimizer: Use paged_adamw_8bit with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
45
  - lr_scheduler_type: linear
46
  - lr_scheduler_warmup_steps: 2
47
+ - training_steps: 1000
48
  - mixed_precision_training: Native AMP
49
 
50
  ### Training results
 
56
  - PEFT 0.13.3.dev0
57
  - Transformers 4.46.1
58
  - Pytorch 2.5.0+cu121
59
+ - Datasets 3.1.0
60
  - Tokenizers 0.20.1
adapter_config.json CHANGED
@@ -21,13 +21,13 @@
21
  "rank_pattern": {},
22
  "revision": null,
23
  "target_modules": [
24
- "up_proj",
25
  "gate_proj",
 
 
 
26
  "k_proj",
27
  "q_proj",
28
- "down_proj",
29
- "v_proj",
30
- "o_proj"
31
  ],
32
  "task_type": "CAUSAL_LM",
33
  "use_dora": false,
 
21
  "rank_pattern": {},
22
  "revision": null,
23
  "target_modules": [
 
24
  "gate_proj",
25
+ "v_proj",
26
+ "down_proj",
27
+ "o_proj",
28
  "k_proj",
29
  "q_proj",
30
+ "up_proj"
 
 
31
  ],
32
  "task_type": "CAUSAL_LM",
33
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ffd327f40e03c9f45b0997e8ddf7ced4df8fc8ffccf74cf1035f123d642cea98
3
  size 83945296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d62b44a9a00a2bcb6d3d228c555126316aa3217d8f35b6fb5b182c84f06731cb
3
  size 83945296
runs/Oct31_19-16-42_a63a83da7440/events.out.tfevents.1730402205.a63a83da7440.1228.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf3303b33f92a198b76d5c5e3e2048ac57bad367b6525118392ecbd169bc5ad2
3
+ size 216732
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a640119e763e4c9eb7136f248b18a2c38d07abf5f02c39c9e6b4c12a5e65746
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06f278bcce14b20faeed6c7ede331f59e1e7516f78c166354efb11634e6c8d7b
3
  size 5496