Bruno7 commited on
Commit
900a05b
·
verified ·
1 Parent(s): c88c624

Model save

Browse files
README.md CHANGED
@@ -34,21 +34,21 @@ More information needed
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 5.5e-05
37
- - train_batch_size: 32
38
- - eval_batch_size: 8
39
  - seed: 42
40
- - gradient_accumulation_steps: 2
41
  - total_train_batch_size: 64
42
  - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
- - lr_scheduler_warmup_steps: 500
45
- - num_epochs: 3
46
  - mixed_precision_training: Native AMP
47
 
48
  ### Framework versions
49
 
50
  - PEFT 0.15.2
51
- - Transformers 4.52.3
52
  - Pytorch 2.6.0+cu124
53
  - Datasets 3.6.0
54
  - Tokenizers 0.21.1
 
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 5.5e-05
37
+ - train_batch_size: 16
38
+ - eval_batch_size: 16
39
  - seed: 42
40
+ - gradient_accumulation_steps: 4
41
  - total_train_batch_size: 64
42
  - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
+ - lr_scheduler_warmup_steps: 100
45
+ - num_epochs: 5
46
  - mixed_precision_training: Native AMP
47
 
48
  ### Framework versions
49
 
50
  - PEFT 0.15.2
51
+ - Transformers 4.52.4
52
  - Pytorch 2.6.0+cu124
53
  - Datasets 3.6.0
54
  - Tokenizers 0.21.1
adapter_config.json CHANGED
@@ -27,10 +27,10 @@
27
  "rank_pattern": {},
28
  "revision": null,
29
  "target_modules": [
30
- "v_proj",
31
  "q_proj",
32
- "o_proj",
33
- "k_proj"
34
  ],
35
  "task_type": null,
36
  "trainable_token_indices": null,
 
27
  "rank_pattern": {},
28
  "revision": null,
29
  "target_modules": [
30
+ "k_proj",
31
  "q_proj",
32
+ "v_proj",
33
+ "o_proj"
34
  ],
35
  "task_type": null,
36
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:98fb01fc41430119ff64113ddf2d2588d7980066fbc4512c2bc0d45a9745930c
3
  size 78677456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa142720e80dc427de5eb184ff6dbd425df016512dff4ca53db4072bf34bd319
3
  size 78677456
runs/May30_13-34-06_3a133184e7b6/events.out.tfevents.1748612059.3a133184e7b6.76765.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16594e3a785c1086324645b425c86891a719b9c210b726769e242a69008df3a6
3
+ size 5779
runs/May30_13-38-56_3a133184e7b6/events.out.tfevents.1748612344.3a133184e7b6.76765.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22490fcbe77341597da8290f76bdc8cc1b89fbbf2dab1d9e17adbf8e4ab13489
3
+ size 5779
runs/May30_13-41-29_3a133184e7b6/events.out.tfevents.1748612498.3a133184e7b6.76765.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f906d6d24fc314b5bea65c7f67bf67f53c5b414b6680747b56ee496567e129b1
3
+ size 11426
runs/May30_13-41-29_3a133184e7b6/events.out.tfevents.1748612751.3a133184e7b6.76765.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6428b2238d8ad4353af80b8e05e38f4cdef0cb31ae831a3c272331637f425e2d
3
+ size 5757
runs/May30_13-41-29_3a133184e7b6/events.out.tfevents.1748612872.3a133184e7b6.76765.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa6a2b95bdd6c69a9d4350742707bca0e1e032fa24e70fff61a9da4717ae3df6
3
+ size 5964
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:87006930f24df5622768272be5325413bd12ab40ec6d7e94c8b32c9424d80178
3
  size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26cd6ba8f4f81b491eb0aa376a9d16b719450a68abb33db5020fe67822e4b8a5
3
  size 5560
training_log.json CHANGED
@@ -1,9 +1 @@
1
- [
2
- {
3
- "loss": 0.7766,
4
- "grad_norm": 0.9275875687599182,
5
- "learning_rate": 0.0,
6
- "epoch": 0.001190003966679889,
7
- "step": 1
8
- }
9
- ]
 
1
+ []