astegaras commited on
Commit
9842ec8
·
verified ·
1 Parent(s): 1c718f7

Upload folder using huggingface_hub

Browse files
checkpoint-150/README.md CHANGED
@@ -207,4 +207,4 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
207
  [More Information Needed]
208
  ### Framework versions
209
 
210
- - PEFT 0.18.0
 
207
  [More Information Needed]
208
  ### Framework versions
209
 
210
+ - PEFT 0.16.0
checkpoint-150/adapter_config.json CHANGED
@@ -1,7 +1,5 @@
1
  {
2
- "alora_invocation_tokens": null,
3
  "alpha_pattern": {},
4
- "arrow_config": null,
5
  "auto_mapping": {
6
  "base_model_class": "LlamaForCausalLM",
7
  "parent_library": "transformers.models.llama.modeling_llama",
@@ -10,7 +8,6 @@
10
  "base_model_name_or_path": "unsloth/llama-3.2-3b-instruct-unsloth-bnb-4bit",
11
  "bias": "none",
12
  "corda_config": null,
13
- "ensure_weight_tying": false,
14
  "eva_config": null,
15
  "exclude_modules": null,
16
  "fan_in_fan_out": false,
@@ -27,21 +24,19 @@
27
  "megatron_core": "megatron.core",
28
  "modules_to_save": null,
29
  "peft_type": "LORA",
30
- "peft_version": "0.18.0",
31
  "qalora_group_size": 16,
32
  "r": 16,
33
  "rank_pattern": {},
34
  "revision": null,
35
  "target_modules": [
36
- "v_proj",
37
- "gate_proj",
38
- "o_proj",
39
- "down_proj",
40
  "k_proj",
41
  "up_proj",
 
 
 
 
42
  "q_proj"
43
  ],
44
- "target_parameters": null,
45
  "task_type": "CAUSAL_LM",
46
  "trainable_token_indices": null,
47
  "use_dora": false,
 
1
  {
 
2
  "alpha_pattern": {},
 
3
  "auto_mapping": {
4
  "base_model_class": "LlamaForCausalLM",
5
  "parent_library": "transformers.models.llama.modeling_llama",
 
8
  "base_model_name_or_path": "unsloth/llama-3.2-3b-instruct-unsloth-bnb-4bit",
9
  "bias": "none",
10
  "corda_config": null,
 
11
  "eva_config": null,
12
  "exclude_modules": null,
13
  "fan_in_fan_out": false,
 
24
  "megatron_core": "megatron.core",
25
  "modules_to_save": null,
26
  "peft_type": "LORA",
 
27
  "qalora_group_size": 16,
28
  "r": 16,
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
 
 
 
 
32
  "k_proj",
33
  "up_proj",
34
+ "down_proj",
35
+ "gate_proj",
36
+ "o_proj",
37
+ "v_proj",
38
  "q_proj"
39
  ],
 
40
  "task_type": "CAUSAL_LM",
41
  "trainable_token_indices": null,
42
  "use_dora": false,
checkpoint-150/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:65374b7a5d997154440e34ac4bbd51a566f8d7a4d4ba19c0ffb48634caef35e6
3
  size 97307544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7cee91ee920a50a5870d0d736d3af53b0f3e76d10b4f9ee7d8894b9e3a1765f
3
  size 97307544
checkpoint-150/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2437f673977fef74e8260e1e6d7b3a56d3ba1c164abe4f35ca525e4524b766dc
3
  size 194846331
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a02a7712989a7bbdcd872d31d8a6e212e74ecd866e06a4cc1bc522874d6a270
3
  size 194846331
checkpoint-150/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f1d565802a8e26c4e8a31328752b7a7fdc186d9401aa008e65697d0ad8c22e33
3
  size 14645
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:716d83e48b4ef0ecc739db256420344d8764b568cd2fec53ea128d8653e43804
3
  size 14645
checkpoint-150/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c272906c5c00ea266d4d7b2f058e95352849407edf35e7999971a0f3aead2e82
3
  size 1465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae8ec43895301a2abc48f62a246f249bdb98524006b5b13cfd6c9b55f6c69c2c
3
  size 1465
checkpoint-150/trainer_state.json CHANGED
@@ -2,7 +2,7 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 0.012,
6
  "eval_steps": 500,
7
  "global_step": 150,
8
  "is_hyper_param_search": false,
@@ -10,7 +10,7 @@
10
  "is_world_process_zero": true,
11
  "log_history": [],
12
  "logging_steps": 9999999,
13
- "max_steps": 12500,
14
  "num_input_tokens_seen": 0,
15
  "num_train_epochs": 1,
16
  "save_steps": 150,
@@ -26,8 +26,8 @@
26
  "attributes": {}
27
  }
28
  },
29
- "total_flos": 1.3741129112813568e+16,
30
- "train_batch_size": 2,
31
  "trial_name": null,
32
  "trial_params": null
33
  }
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 0.024,
6
  "eval_steps": 500,
7
  "global_step": 150,
8
  "is_hyper_param_search": false,
 
10
  "is_world_process_zero": true,
11
  "log_history": [],
12
  "logging_steps": 9999999,
13
+ "max_steps": 6250,
14
  "num_input_tokens_seen": 0,
15
  "num_train_epochs": 1,
16
  "save_steps": 150,
 
26
  "attributes": {}
27
  }
28
  },
29
+ "total_flos": 3.38062726877184e+16,
30
+ "train_batch_size": 4,
31
  "trial_name": null,
32
  "trial_params": null
33
  }
checkpoint-150/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cd1e61689e67f92937b424f1859bcd30776e97cdda7fe9a43a40ccef44783eb7
3
- size 6225
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa7d60f39157d6af42ddf961454a0f16be99637b24b2dc9dcfdb9f7339f17b36
3
+ size 6161