Mel-Iza0 commited on
Commit
d7e8498
·
verified ·
1 Parent(s): 6b58c31

Upload folder using huggingface_hub

Browse files
checkpoint-10/adapter_config.json CHANGED
@@ -19,10 +19,10 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "o_proj",
23
  "q_proj",
 
24
  "v_proj",
25
- "k_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_rslora": false
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
22
  "q_proj",
23
+ "k_proj",
24
  "v_proj",
25
+ "o_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_rslora": false
checkpoint-10/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1e888ba5b357cca479fdfb83c7a6d1d517139cd69f31a8e5d2b0f19b95f96156
3
  size 27297032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04cb9a8865453bedfca65851459d7c8cb45b8f30dc9f4934adf42aeeabae0444
3
  size 27297032
checkpoint-10/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:38e7804f61425e5bc724e211e81e2632de062c48494ab6889553a89535b9b858
3
  size 54678010
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8228616759eaeec2a7189a3ab7788d7dfcf811a9f0f3e8fead1478ac977e3ac5
3
  size 54678010
checkpoint-10/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:71deba4f5cbb7f14856592e6b45f65cdbc27ef90ce55c1b4360b28f0d3af92eb
3
  size 14512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74b72282c30a7e55d3f15d285f01f3fc9eacba980e5695f003e536ef0086f6fb
3
  size 14512
checkpoint-10/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fd1064c84cc8895e25cf176f0a6304640003de5efcfb37abf2faf7263c4c86b7
3
  size 14512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74be118bc47e43344dc7f8114adf2066553c0dfab0d66a5fea02e6746c03d068
3
  size 14512
checkpoint-10/trainer_state.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "best_metric": 1.2620933055877686,
3
  "best_model_checkpoint": "./mistral/29-02-24-Weni-testing_saving_checkpoints-final_Zeroshot-2_max_steps-60_batch_8_2024-02-29_ppid_7/checkpoint-10",
4
  "epoch": 0.006199628022318661,
5
  "eval_steps": 10,
@@ -10,10 +10,10 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.01,
13
- "eval_loss": 1.2620933055877686,
14
- "eval_runtime": 214.1744,
15
- "eval_samples_per_second": 13.386,
16
- "eval_steps_per_second": 3.348,
17
  "step": 10
18
  }
19
  ],
@@ -22,7 +22,7 @@
22
  "num_input_tokens_seen": 0,
23
  "num_train_epochs": 1,
24
  "save_steps": 10,
25
- "total_flos": 5344598247342080.0,
26
  "train_batch_size": 8,
27
  "trial_name": null,
28
  "trial_params": null
 
1
  {
2
+ "best_metric": 1.2639678716659546,
3
  "best_model_checkpoint": "./mistral/29-02-24-Weni-testing_saving_checkpoints-final_Zeroshot-2_max_steps-60_batch_8_2024-02-29_ppid_7/checkpoint-10",
4
  "epoch": 0.006199628022318661,
5
  "eval_steps": 10,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.01,
13
+ "eval_loss": 1.2639678716659546,
14
+ "eval_runtime": 212.9338,
15
+ "eval_samples_per_second": 13.464,
16
+ "eval_steps_per_second": 3.367,
17
  "step": 10
18
  }
19
  ],
 
22
  "num_input_tokens_seen": 0,
23
  "num_train_epochs": 1,
24
  "save_steps": 10,
25
+ "total_flos": 5207942051659776.0,
26
  "train_batch_size": 8,
27
  "trial_name": null,
28
  "trial_params": null
checkpoint-10/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4a7b951862cc1b6fc3818abc40db6d785393d7d2119a2d8adb59ac9fe86d702f
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f759d9115ca82f09dba4a4485ea9736298ef842bc83818bf3fb831f68f5920ed
3
  size 5112