Mel-Iza0 commited on
Commit
27e61c3
·
verified ·
1 Parent(s): 167d1b8

Upload folder using huggingface_hub

Browse files
adapter_config.json CHANGED
@@ -19,10 +19,10 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
22
  "q_proj",
23
- "v_proj",
24
  "o_proj",
25
- "k_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_rslora": false
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "k_proj",
23
  "q_proj",
 
24
  "o_proj",
25
+ "v_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_rslora": false
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:213e4149e5333fef153ec75df7cfeada3affc4cf1274cbd86e8be5940e882d62
3
  size 27297032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d45a758481e75e92228735e9a83f3fcc6a9330706c1d1f4674c5ddaf8baf3e67
3
  size 27297032
checkpoint-10/adapter_config.json CHANGED
@@ -19,10 +19,10 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
22
  "q_proj",
23
- "v_proj",
24
  "o_proj",
25
- "k_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_rslora": false
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "k_proj",
23
  "q_proj",
 
24
  "o_proj",
25
+ "v_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_rslora": false
checkpoint-10/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:213e4149e5333fef153ec75df7cfeada3affc4cf1274cbd86e8be5940e882d62
3
  size 27297032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d45a758481e75e92228735e9a83f3fcc6a9330706c1d1f4674c5ddaf8baf3e67
3
  size 27297032
checkpoint-10/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:02334521114a0317774a2cd0a9e77f83d2746dc43e5f11f138281e26e55fbfae
3
  size 54678010
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:faec2caafa9315ba8ca661835dbf7217cff01604f343a3ca08b85ed6b0a38dd8
3
  size 54678010
checkpoint-10/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48704b03cfa5f48d380a2c0f70433ae9a9fb0c6fb3cbb0dfa07bd65e22953db4
3
  size 14512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85c8c4aa40cb47f5dfe76e5e3a3d4d92ee05817ba0d04337354a29cc65fc9aa4
3
  size 14512
checkpoint-10/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7c12f4d0596e72b5705a2a240e5d3c50ef73006c8546aac031b8955275186fcc
3
  size 14512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4b81051e9e6293a85fca6be580da01a9e72a63e666fd69eba67be47904ac813
3
  size 14512
checkpoint-10/trainer_state.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "best_metric": 1.268858551979065,
3
  "best_model_checkpoint": "./mistral/29-02-24-Weni-testing_saving_checkpoints-final_Zeroshot-2_max_steps-60_batch_8_2024-02-29_ppid_7/checkpoint-10",
4
  "epoch": 0.006199628022318661,
5
  "eval_steps": 10,
@@ -10,10 +10,10 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.01,
13
- "eval_loss": 1.268858551979065,
14
- "eval_runtime": 210.5046,
15
- "eval_samples_per_second": 13.62,
16
- "eval_steps_per_second": 3.406,
17
  "step": 10
18
  }
19
  ],
@@ -22,7 +22,7 @@
22
  "num_input_tokens_seen": 0,
23
  "num_train_epochs": 1,
24
  "save_steps": 10,
25
- "total_flos": 5578962264653824.0,
26
  "train_batch_size": 8,
27
  "trial_name": null,
28
  "trial_params": null
 
1
  {
2
+ "best_metric": 1.262868046760559,
3
  "best_model_checkpoint": "./mistral/29-02-24-Weni-testing_saving_checkpoints-final_Zeroshot-2_max_steps-60_batch_8_2024-02-29_ppid_7/checkpoint-10",
4
  "epoch": 0.006199628022318661,
5
  "eval_steps": 10,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.01,
13
+ "eval_loss": 1.262868046760559,
14
+ "eval_runtime": 212.6057,
15
+ "eval_samples_per_second": 13.485,
16
+ "eval_steps_per_second": 3.372,
17
  "step": 10
18
  }
19
  ],
 
22
  "num_input_tokens_seen": 0,
23
  "num_train_epochs": 1,
24
  "save_steps": 10,
25
+ "total_flos": 5482620242624512.0,
26
  "train_batch_size": 8,
27
  "trial_name": null,
28
  "trial_params": null
checkpoint-10/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe0fbff02770a0f92e2a7ea855958234a42b39724e3eefaa45f6d3ef5d38af30
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e7d07fb6d0dc64dc7d7208531c4399717ebe171f4b315c4a0abf0bb2ca7ecb8
3
  size 5112
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe0fbff02770a0f92e2a7ea855958234a42b39724e3eefaa45f6d3ef5d38af30
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e7d07fb6d0dc64dc7d7208531c4399717ebe171f4b315c4a0abf0bb2ca7ecb8
3
  size 5112