abdurafeyf commited on
Commit
dbaca93
·
verified ·
1 Parent(s): 70678f9

Training in progress, step 30

Browse files
adapter_config.json CHANGED
@@ -3,6 +3,7 @@
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "unsloth/llama-3.2-11b-vision-instruct-unsloth-bnb-4bit",
5
  "bias": "none",
 
6
  "eva_config": null,
7
  "exclude_modules": null,
8
  "fan_in_fan_out": false,
@@ -22,8 +23,9 @@
22
  "r": 16,
23
  "rank_pattern": {},
24
  "revision": null,
25
- "target_modules": "(?:.*?(?:language|text).*?(?:self_attn|attention|attn|mlp|feed_forward|ffn|dense).*?(?:q_proj|k_proj|v_proj|o_proj|fc1|fc2|gate_proj|up_proj|down_proj).*?)|(?:\\bmodel\\.layers\\.[\\d]{1,}\\.(?:self_attn|attention|attn|mlp|feed_forward|ffn|dense)\\.(?:(?:q_proj|k_proj|v_proj|o_proj|fc1|fc2|gate_proj|up_proj|down_proj)))",
26
  "task_type": "CAUSAL_LM",
 
27
  "use_dora": false,
28
  "use_rslora": false
29
  }
 
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "unsloth/llama-3.2-11b-vision-instruct-unsloth-bnb-4bit",
5
  "bias": "none",
6
+ "corda_config": null,
7
  "eva_config": null,
8
  "exclude_modules": null,
9
  "fan_in_fan_out": false,
 
23
  "r": 16,
24
  "rank_pattern": {},
25
  "revision": null,
26
+ "target_modules": "(?:.*?(?:vision|image|visual|patch|language|text).*?(?:self_attn|attention|attn|mlp|feed_forward|ffn|dense).*?(?:q_proj|k_proj|v_proj|o_proj|fc1|fc2|gate_proj|up_proj|down_proj).*?)|(?:\\bmodel\\.layers\\.[\\d]{1,}\\.(?:self_attn|attention|attn|mlp|feed_forward|ffn|dense)\\.(?:(?:q_proj|k_proj|v_proj|o_proj|fc1|fc2|gate_proj|up_proj|down_proj)))",
27
  "task_type": "CAUSAL_LM",
28
+ "trainable_token_indices": null,
29
  "use_dora": false,
30
  "use_rslora": false
31
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:411601b34a4c3f9e69aedd3e2b6df23940d97e7511fe326c5a69d8d5185fb8e5
3
- size 209798936
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7222746c399f1e089eed90392babc4b8e7b88f024b17c647c469baf095e1cb2
3
+ size 268855232
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:abd08b46c9c01a903788030f36d9731aad064d0c420e66b88e3e143cda9e4e9a
3
- size 17210268
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d34b8b90d18ab7ffd5282840a84d5d4df705e2c8e6f74c906ee0a1ea5b6bd9dd
3
+ size 17210367
tokenizer_config.json CHANGED
@@ -2070,6 +2070,6 @@
2070
  "pad_token": "<|finetune_right_pad_id|>",
2071
  "padding_side": "right",
2072
  "processor_class": "MllamaProcessor",
2073
- "tokenizer_class": "PreTrainedTokenizerFast",
2074
  "unk_token": null
2075
  }
 
2070
  "pad_token": "<|finetune_right_pad_id|>",
2071
  "padding_side": "right",
2072
  "processor_class": "MllamaProcessor",
2073
+ "tokenizer_class": "PreTrainedTokenizer",
2074
  "unk_token": null
2075
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dfdabf185a3f95a7e2003539099d81cb0096b56fb46259af75394409f3311343
3
- size 5624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4bd08e834fc3df3702a629726c950bb730c10233a6f29ac579e4376b9af1674c
3
+ size 5688