diagonalge commited on
Commit
e614d70
·
verified ·
1 Parent(s): 38d91d1

Upload task output test1334test1234test1234test34

Browse files
README.md CHANGED
@@ -4,7 +4,7 @@ tags:
4
  - generated_from_trainer
5
  base_model: samoline/b7447218-27e6-491c-b3ab-ea03a5b93541
6
  model-index:
7
- - name: workspace/axolotl/outputs/test1334test1234test1234test12334/texttest
8
  results: []
9
  ---
10
 
@@ -23,7 +23,7 @@ chat_template: llama3
23
  dataset_prepared_path: null
24
  datasets:
25
  - data_files:
26
- - test1334test1234test1234test12334_train_data.json
27
  ds_type: json
28
  format: custom
29
  path: /workspace/axolotl/data
@@ -60,11 +60,11 @@ lora_target_linear: true
60
  lr_scheduler: cosine
61
  max_steps: 1
62
  micro_batch_size: 2
63
- mlflow_experiment_name: /workspace/axolotl/data/test1334test1234test1234test12334_train_data.json
64
  model_type: AutoModelForCausalLM
65
  num_epochs: 1
66
  optimizer: adamw_bnb_8bit
67
- output_dir: /workspace/axolotl/outputs/test1334test1234test1234test12334/texttest
68
  pad_to_sequence_len: true
69
  resume_from_checkpoint: null
70
  s2_attention: null
@@ -79,10 +79,10 @@ trust_remote_code: true
79
  val_set_size: 0.05
80
  wandb_entity: null
81
  wandb_mode: offline
82
- wandb_name: test1334test1234test1234test12334_texttest
83
  wandb_project: Gradients-On-Demand
84
  wandb_run: your_name
85
- wandb_runid: test1334test1234test1234test12334_texttest
86
  warmup_steps: 1
87
  weight_decay: 0.0
88
  xformers_attention: null
@@ -91,7 +91,7 @@ xformers_attention: null
91
 
92
  </details><br>
93
 
94
- # workspace/axolotl/outputs/test1334test1234test1234test12334/texttest
95
 
96
  This model was trained from scratch on the None dataset.
97
 
 
4
  - generated_from_trainer
5
  base_model: samoline/b7447218-27e6-491c-b3ab-ea03a5b93541
6
  model-index:
7
+ - name: workspace/axolotl/outputs/test1334test1234test1234test34/texttest
8
  results: []
9
  ---
10
 
 
23
  dataset_prepared_path: null
24
  datasets:
25
  - data_files:
26
+ - test1334test1234test1234test34_train_data.json
27
  ds_type: json
28
  format: custom
29
  path: /workspace/axolotl/data
 
60
  lr_scheduler: cosine
61
  max_steps: 1
62
  micro_batch_size: 2
63
+ mlflow_experiment_name: /workspace/axolotl/data/test1334test1234test1234test34_train_data.json
64
  model_type: AutoModelForCausalLM
65
  num_epochs: 1
66
  optimizer: adamw_bnb_8bit
67
+ output_dir: /workspace/axolotl/outputs/test1334test1234test1234test34/texttest
68
  pad_to_sequence_len: true
69
  resume_from_checkpoint: null
70
  s2_attention: null
 
79
  val_set_size: 0.05
80
  wandb_entity: null
81
  wandb_mode: offline
82
+ wandb_name: test1334test1234test1234test34_texttest
83
  wandb_project: Gradients-On-Demand
84
  wandb_run: your_name
85
+ wandb_runid: test1334test1234test1234test34_texttest
86
  warmup_steps: 1
87
  weight_decay: 0.0
88
  xformers_attention: null
 
91
 
92
  </details><br>
93
 
94
+ # workspace/axolotl/outputs/test1334test1234test1234test34/texttest
95
 
96
  This model was trained from scratch on the None dataset.
97
 
adapter_config.json CHANGED
@@ -25,10 +25,10 @@
25
  "revision": null,
26
  "target_modules": [
27
  "gate_proj",
28
- "k_proj",
29
- "down_proj",
30
  "q_proj",
 
31
  "up_proj",
 
32
  "v_proj",
33
  "o_proj"
34
  ],
 
25
  "revision": null,
26
  "target_modules": [
27
  "gate_proj",
 
 
28
  "q_proj",
29
+ "k_proj",
30
  "up_proj",
31
+ "down_proj",
32
  "v_proj",
33
  "o_proj"
34
  ],
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a25b065eef1c3d7fc2341c739d39f9e5314db39eee10c9342acee52c4d7aee7
3
  size 22573704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ea6929ca6dde2a3ac9c8a46014547707c434edda5748ced18cd7d5a5201fef2
3
  size 22573704
checkpoint-1/README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: /cache/test1334test1234test1234test12334/models/samoline--b7447218-27e6-491c-b3ab-ea03a5b93541
3
  library_name: peft
4
  ---
5
 
 
1
  ---
2
+ base_model: /cache/test1334test1234test1234test34/models/samoline--b7447218-27e6-491c-b3ab-ea03a5b93541
3
  library_name: peft
4
  ---
5
 
checkpoint-1/adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "/cache/test1334test1234test1234test12334/models/samoline--b7447218-27e6-491c-b3ab-ea03a5b93541",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
@@ -25,10 +25,10 @@
25
  "revision": null,
26
  "target_modules": [
27
  "gate_proj",
28
- "k_proj",
29
- "down_proj",
30
  "q_proj",
 
31
  "up_proj",
 
32
  "v_proj",
33
  "o_proj"
34
  ],
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "/cache/test1334test1234test1234test34/models/samoline--b7447218-27e6-491c-b3ab-ea03a5b93541",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
 
25
  "revision": null,
26
  "target_modules": [
27
  "gate_proj",
 
 
28
  "q_proj",
29
+ "k_proj",
30
  "up_proj",
31
+ "down_proj",
32
  "v_proj",
33
  "o_proj"
34
  ],
checkpoint-1/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a25b065eef1c3d7fc2341c739d39f9e5314db39eee10c9342acee52c4d7aee7
3
  size 22573704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ea6929ca6dde2a3ac9c8a46014547707c434edda5748ced18cd7d5a5201fef2
3
  size 22573704
checkpoint-1/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:10159d605898dbcbf79e301a5f5144c27286a025a489ac232e94d88acb2a6bee
3
  size 11710970
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:984dce112fc24ae4eaba62d2904d0043a7f32ec1a9ca5e5894925adf96d89a5b
3
  size 11710970
checkpoint-1/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8898f6f7795fb6bc31316b2e35945a51a1fa3d28252626c221222713df549b8d
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f97bb552d304801def1800311e9b4f0db4888641a7d6326f60a0099954973d2b
3
  size 14244
checkpoint-1/trainer_state.json CHANGED
@@ -2,7 +2,7 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 0.009216589861751152,
6
  "eval_steps": 500,
7
  "global_step": 1,
8
  "is_hyper_param_search": false,
@@ -10,10 +10,10 @@
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
- "epoch": 0.009216589861751152,
14
- "grad_norm": 0.6244447827339172,
15
  "learning_rate": 0.0,
16
- "loss": 1.0026,
17
  "step": 1
18
  }
19
  ],
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 0.009900990099009901,
6
  "eval_steps": 500,
7
  "global_step": 1,
8
  "is_hyper_param_search": false,
 
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
+ "epoch": 0.009900990099009901,
14
+ "grad_norm": 0.472037672996521,
15
  "learning_rate": 0.0,
16
+ "loss": 0.9592,
17
  "step": 1
18
  }
19
  ],
checkpoint-1/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a1ebb3a101775193404ad81edb57b0ce9d21fb9922f5caea612cd94c8bac42a
3
  size 7224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d896a8d25c643391583d2ed5bb4dced3ad66e83c5f45be84dfcb518f2faa37f
3
  size 7224