ioky commited on
Commit
f7b6daf
·
verified ·
1 Parent(s): d04b594

Upload all models into one repo

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +16 -0
  2. Qwen2.5-VL-7B-Instruct_1s/lora/sft/README.md +61 -0
  3. Qwen2.5-VL-7B-Instruct_1s/lora/sft/adapter_config.json +120 -0
  4. Qwen2.5-VL-7B-Instruct_1s/lora/sft/adapter_model.safetensors +3 -0
  5. Qwen2.5-VL-7B-Instruct_1s/lora/sft/added_tokens.json +24 -0
  6. Qwen2.5-VL-7B-Instruct_1s/lora/sft/chat_template.json +3 -0
  7. Qwen2.5-VL-7B-Instruct_1s/lora/sft/merges.txt +0 -0
  8. Qwen2.5-VL-7B-Instruct_1s/lora/sft/preprocessor_config.json +29 -0
  9. Qwen2.5-VL-7B-Instruct_1s/lora/sft/special_tokens_map.json +31 -0
  10. Qwen2.5-VL-7B-Instruct_1s/lora/sft/tokenizer.json +3 -0
  11. Qwen2.5-VL-7B-Instruct_1s/lora/sft/tokenizer_config.json +210 -0
  12. Qwen2.5-VL-7B-Instruct_1s/lora/sft/vocab.json +0 -0
  13. Qwen2.5-VL-7B-Instruct_30s/lora/sft/README.md +61 -0
  14. Qwen2.5-VL-7B-Instruct_30s/lora/sft/adapter_config.json +120 -0
  15. Qwen2.5-VL-7B-Instruct_30s/lora/sft/adapter_model.safetensors +3 -0
  16. Qwen2.5-VL-7B-Instruct_30s/lora/sft/added_tokens.json +24 -0
  17. Qwen2.5-VL-7B-Instruct_30s/lora/sft/chat_template.json +3 -0
  18. Qwen2.5-VL-7B-Instruct_30s/lora/sft/merges.txt +0 -0
  19. Qwen2.5-VL-7B-Instruct_30s/lora/sft/preprocessor_config.json +29 -0
  20. Qwen2.5-VL-7B-Instruct_30s/lora/sft/special_tokens_map.json +31 -0
  21. Qwen2.5-VL-7B-Instruct_30s/lora/sft/tokenizer.json +3 -0
  22. Qwen2.5-VL-7B-Instruct_30s/lora/sft/tokenizer_config.json +210 -0
  23. Qwen2.5-VL-7B-Instruct_30s/lora/sft/vocab.json +0 -0
  24. Qwen2.5-VL-7B-Instruct_5s/lora/sft/README.md +61 -0
  25. Qwen2.5-VL-7B-Instruct_5s/lora/sft/adapter_config.json +120 -0
  26. Qwen2.5-VL-7B-Instruct_5s/lora/sft/adapter_model.safetensors +3 -0
  27. Qwen2.5-VL-7B-Instruct_5s/lora/sft/added_tokens.json +24 -0
  28. Qwen2.5-VL-7B-Instruct_5s/lora/sft/chat_template.json +3 -0
  29. Qwen2.5-VL-7B-Instruct_5s/lora/sft/merges.txt +0 -0
  30. Qwen2.5-VL-7B-Instruct_5s/lora/sft/preprocessor_config.json +29 -0
  31. Qwen2.5-VL-7B-Instruct_5s/lora/sft/special_tokens_map.json +31 -0
  32. Qwen2.5-VL-7B-Instruct_5s/lora/sft/tokenizer.json +3 -0
  33. Qwen2.5-VL-7B-Instruct_5s/lora/sft/tokenizer_config.json +210 -0
  34. Qwen2.5-VL-7B-Instruct_5s/lora/sft/vocab.json +0 -0
  35. Qwen2.5-VL-7B-Instruct_60s/lora/sft/README.md +61 -0
  36. Qwen2.5-VL-7B-Instruct_60s/lora/sft/adapter_config.json +120 -0
  37. Qwen2.5-VL-7B-Instruct_60s/lora/sft/adapter_model.safetensors +3 -0
  38. Qwen2.5-VL-7B-Instruct_60s/lora/sft/added_tokens.json +24 -0
  39. Qwen2.5-VL-7B-Instruct_60s/lora/sft/chat_template.json +3 -0
  40. Qwen2.5-VL-7B-Instruct_60s/lora/sft/merges.txt +0 -0
  41. Qwen2.5-VL-7B-Instruct_60s/lora/sft/preprocessor_config.json +29 -0
  42. Qwen2.5-VL-7B-Instruct_60s/lora/sft/special_tokens_map.json +31 -0
  43. Qwen2.5-VL-7B-Instruct_60s/lora/sft/tokenizer.json +3 -0
  44. Qwen2.5-VL-7B-Instruct_60s/lora/sft/tokenizer_config.json +210 -0
  45. Qwen2.5-VL-7B-Instruct_60s/lora/sft/vocab.json +0 -0
  46. Qwen2.5-VL-7B-Instruct_chain1_1s/lora/sft/README.md +61 -0
  47. Qwen2.5-VL-7B-Instruct_chain1_1s/lora/sft/adapter_config.json +120 -0
  48. Qwen2.5-VL-7B-Instruct_chain1_1s/lora/sft/adapter_model.safetensors +3 -0
  49. Qwen2.5-VL-7B-Instruct_chain1_1s/lora/sft/added_tokens.json +24 -0
  50. Qwen2.5-VL-7B-Instruct_chain1_1s/lora/sft/chat_template.json +3 -0
.gitattributes CHANGED
@@ -33,3 +33,19 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Qwen2.5-VL-7B-Instruct_1s/lora/sft/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ Qwen2.5-VL-7B-Instruct_30s/lora/sft/tokenizer.json filter=lfs diff=lfs merge=lfs -text
38
+ Qwen2.5-VL-7B-Instruct_5s/lora/sft/tokenizer.json filter=lfs diff=lfs merge=lfs -text
39
+ Qwen2.5-VL-7B-Instruct_60s/lora/sft/tokenizer.json filter=lfs diff=lfs merge=lfs -text
40
+ Qwen2.5-VL-7B-Instruct_chain1_1s/lora/sft/tokenizer.json filter=lfs diff=lfs merge=lfs -text
41
+ Qwen2.5-VL-7B-Instruct_chain1_30s/lora/sft/tokenizer.json filter=lfs diff=lfs merge=lfs -text
42
+ Qwen2.5-VL-7B-Instruct_chain1_5s/lora/sft/tokenizer.json filter=lfs diff=lfs merge=lfs -text
43
+ Qwen2.5-VL-7B-Instruct_chain1_60s/lora/sft/tokenizer.json filter=lfs diff=lfs merge=lfs -text
44
+ Qwen2.5-VL-7B-Instruct_chain2_1s/lora/sft/tokenizer.json filter=lfs diff=lfs merge=lfs -text
45
+ Qwen2.5-VL-7B-Instruct_chain2_30s/lora/sft/tokenizer.json filter=lfs diff=lfs merge=lfs -text
46
+ Qwen2.5-VL-7B-Instruct_chain2_5s/lora/sft/tokenizer.json filter=lfs diff=lfs merge=lfs -text
47
+ Qwen2.5-VL-7B-Instruct_chain2_60s/lora/sft/tokenizer.json filter=lfs diff=lfs merge=lfs -text
48
+ Qwen2.5-VL-7B-Instruct_chain3_1s/lora/sft/tokenizer.json filter=lfs diff=lfs merge=lfs -text
49
+ Qwen2.5-VL-7B-Instruct_chain3_30s/lora/sft/tokenizer.json filter=lfs diff=lfs merge=lfs -text
50
+ Qwen2.5-VL-7B-Instruct_chain3_5s/lora/sft/tokenizer.json filter=lfs diff=lfs merge=lfs -text
51
+ Qwen2.5-VL-7B-Instruct_chain3_60s/lora/sft/tokenizer.json filter=lfs diff=lfs merge=lfs -text
Qwen2.5-VL-7B-Instruct_1s/lora/sft/README.md ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ license: other
4
+ base_model: /scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct
5
+ tags:
6
+ - llama-factory
7
+ - lora
8
+ - generated_from_trainer
9
+ model-index:
10
+ - name: sft
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # sft
18
+
19
+ This model is a fine-tuned version of [/scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct](https://huggingface.co//scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct) on the GraSP_1s dataset.
20
+
21
+ ## Model description
22
+
23
+ More information needed
24
+
25
+ ## Intended uses & limitations
26
+
27
+ More information needed
28
+
29
+ ## Training and evaluation data
30
+
31
+ More information needed
32
+
33
+ ## Training procedure
34
+
35
+ ### Training hyperparameters
36
+
37
+ The following hyperparameters were used during training:
38
+ - learning_rate: 0.0001
39
+ - train_batch_size: 12
40
+ - eval_batch_size: 8
41
+ - seed: 42
42
+ - distributed_type: multi-GPU
43
+ - num_devices: 2
44
+ - total_train_batch_size: 24
45
+ - total_eval_batch_size: 16
46
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
+ - lr_scheduler_type: cosine
48
+ - lr_scheduler_warmup_ratio: 0.1
49
+ - num_epochs: 1
50
+
51
+ ### Training results
52
+
53
+
54
+
55
+ ### Framework versions
56
+
57
+ - PEFT 0.15.1
58
+ - Transformers 4.51.3
59
+ - Pytorch 2.6.0+cu124
60
+ - Datasets 3.5.0
61
+ - Tokenizers 0.21.1
Qwen2.5-VL-7B-Instruct_1s/lora/sft/adapter_config.json ADDED
@@ -0,0 +1,120 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "/scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 16,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 8,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "layers.15.mlp.gate_proj",
28
+ "layers.13.mlp.down_proj",
29
+ "layers.14.mlp.down_proj",
30
+ "layers.18.mlp.gate_proj",
31
+ "layers.16.mlp.gate_proj",
32
+ "layers.25.mlp.gate_proj",
33
+ "layers.12.mlp.down_proj",
34
+ "layers.12.mlp.gate_proj",
35
+ "layers.20.mlp.down_proj",
36
+ "layers.12.mlp.up_proj",
37
+ "layers.17.mlp.up_proj",
38
+ "layers.17.mlp.gate_proj",
39
+ "layers.7.mlp.up_proj",
40
+ "layers.0.mlp.gate_proj",
41
+ "layers.24.mlp.up_proj",
42
+ "layers.19.mlp.gate_proj",
43
+ "layers.6.mlp.down_proj",
44
+ "layers.20.mlp.up_proj",
45
+ "layers.22.mlp.down_proj",
46
+ "layers.9.mlp.up_proj",
47
+ "layers.7.mlp.gate_proj",
48
+ "layers.2.mlp.up_proj",
49
+ "layers.27.mlp.up_proj",
50
+ "layers.5.mlp.gate_proj",
51
+ "layers.0.mlp.down_proj",
52
+ "k_proj",
53
+ "layers.27.mlp.gate_proj",
54
+ "layers.7.mlp.down_proj",
55
+ "layers.22.mlp.up_proj",
56
+ "v_proj",
57
+ "layers.23.mlp.up_proj",
58
+ "layers.22.mlp.gate_proj",
59
+ "layers.24.mlp.down_proj",
60
+ "layers.3.mlp.down_proj",
61
+ "layers.8.mlp.up_proj",
62
+ "o_proj",
63
+ "layers.4.mlp.gate_proj",
64
+ "layers.1.mlp.up_proj",
65
+ "layers.6.mlp.up_proj",
66
+ "layers.5.mlp.down_proj",
67
+ "layers.5.mlp.up_proj",
68
+ "layers.23.mlp.down_proj",
69
+ "layers.26.mlp.gate_proj",
70
+ "layers.26.mlp.down_proj",
71
+ "layers.9.mlp.gate_proj",
72
+ "layers.10.mlp.up_proj",
73
+ "layers.13.mlp.up_proj",
74
+ "layers.25.mlp.down_proj",
75
+ "layers.18.mlp.down_proj",
76
+ "layers.4.mlp.up_proj",
77
+ "layers.20.mlp.gate_proj",
78
+ "layers.1.mlp.down_proj",
79
+ "layers.17.mlp.down_proj",
80
+ "layers.14.mlp.gate_proj",
81
+ "layers.2.mlp.down_proj",
82
+ "layers.15.mlp.up_proj",
83
+ "layers.11.mlp.gate_proj",
84
+ "layers.4.mlp.down_proj",
85
+ "layers.26.mlp.up_proj",
86
+ "layers.3.mlp.up_proj",
87
+ "layers.6.mlp.gate_proj",
88
+ "layers.10.mlp.down_proj",
89
+ "layers.8.mlp.gate_proj",
90
+ "layers.2.mlp.gate_proj",
91
+ "layers.21.mlp.up_proj",
92
+ "layers.14.mlp.up_proj",
93
+ "layers.10.mlp.gate_proj",
94
+ "layers.11.mlp.down_proj",
95
+ "layers.21.mlp.gate_proj",
96
+ "layers.16.mlp.down_proj",
97
+ "layers.16.mlp.up_proj",
98
+ "layers.0.mlp.up_proj",
99
+ "layers.18.mlp.up_proj",
100
+ "layers.1.mlp.gate_proj",
101
+ "layers.27.mlp.down_proj",
102
+ "layers.24.mlp.gate_proj",
103
+ "layers.25.mlp.up_proj",
104
+ "layers.19.mlp.up_proj",
105
+ "layers.9.mlp.down_proj",
106
+ "layers.19.mlp.down_proj",
107
+ "q_proj",
108
+ "layers.13.mlp.gate_proj",
109
+ "layers.8.mlp.down_proj",
110
+ "layers.11.mlp.up_proj",
111
+ "layers.3.mlp.gate_proj",
112
+ "layers.21.mlp.down_proj",
113
+ "layers.23.mlp.gate_proj",
114
+ "layers.15.mlp.down_proj"
115
+ ],
116
+ "task_type": "CAUSAL_LM",
117
+ "trainable_token_indices": null,
118
+ "use_dora": false,
119
+ "use_rslora": false
120
+ }
Qwen2.5-VL-7B-Instruct_1s/lora/sft/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:250c4f85e1fb3ceae059a574bdae816fad15f4f63de40c3eac8005bb549aa5b7
3
+ size 80792096
Qwen2.5-VL-7B-Instruct_1s/lora/sft/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
Qwen2.5-VL-7B-Instruct_1s/lora/sft/chat_template.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n{% endif %}<|im_start|>{{ message['role'] }}\n{% if message['content'] is string %}{{ message['content'] }}<|im_end|>\n{% else %}{% for content in message['content'] %}{% if content['type'] == 'image' or 'image' in content or 'image_url' in content %}{% set image_count.value = image_count.value + 1 %}{% if add_vision_id %}Picture {{ image_count.value }}: {% endif %}<|vision_start|><|image_pad|><|vision_end|>{% elif content['type'] == 'video' or 'video' in content %}{% set video_count.value = video_count.value + 1 %}{% if add_vision_id %}Video {{ video_count.value }}: {% endif %}<|vision_start|><|video_pad|><|vision_end|>{% elif 'text' in content %}{{ content['text'] }}{% endif %}{% endfor %}<|im_end|>\n{% endif %}{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant\n{% endif %}"
3
+ }
Qwen2.5-VL-7B-Instruct_1s/lora/sft/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
Qwen2.5-VL-7B-Instruct_1s/lora/sft/preprocessor_config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": true,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.48145466,
8
+ 0.4578275,
9
+ 0.40821073
10
+ ],
11
+ "image_processor_type": "Qwen2VLImageProcessor",
12
+ "image_std": [
13
+ 0.26862954,
14
+ 0.26130258,
15
+ 0.27577711
16
+ ],
17
+ "max_pixels": 12845056,
18
+ "merge_size": 2,
19
+ "min_pixels": 3136,
20
+ "patch_size": 14,
21
+ "processor_class": "Qwen2_5_VLProcessor",
22
+ "resample": 3,
23
+ "rescale_factor": 0.00392156862745098,
24
+ "size": {
25
+ "longest_edge": 12845056,
26
+ "shortest_edge": 3136
27
+ },
28
+ "temporal_patch_size": 2
29
+ }
Qwen2.5-VL-7B-Instruct_1s/lora/sft/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
Qwen2.5-VL-7B-Instruct_1s/lora/sft/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
Qwen2.5-VL-7B-Instruct_1s/lora/sft/tokenizer_config.json ADDED
@@ -0,0 +1,210 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n{% endif %}<|im_start|>{{ message['role'] }}\n{% if message['content'] is string %}{{ message['content'] }}<|im_end|>\n{% else %}{% for content in message['content'] %}{% if content['type'] == 'image' or 'image' in content or 'image_url' in content %}{% set image_count.value = image_count.value + 1 %}{% if add_vision_id %}Picture {{ image_count.value }}: {% endif %}<|vision_start|><|image_pad|><|vision_end|>{% elif content['type'] == 'video' or 'video' in content %}{% set video_count.value = video_count.value + 1 %}{% if add_vision_id %}Video {{ video_count.value }}: {% endif %}<|vision_start|><|video_pad|><|vision_end|>{% elif 'text' in content %}{{ content['text'] }}{% endif %}{% endfor %}<|im_end|>\n{% endif %}{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant\n{% endif %}",
199
+ "clean_up_tokenization_spaces": false,
200
+ "eos_token": "<|im_end|>",
201
+ "errors": "replace",
202
+ "extra_special_tokens": {},
203
+ "model_max_length": 131072,
204
+ "pad_token": "<|endoftext|>",
205
+ "padding_side": "right",
206
+ "processor_class": "Qwen2_5_VLProcessor",
207
+ "split_special_tokens": false,
208
+ "tokenizer_class": "Qwen2Tokenizer",
209
+ "unk_token": null
210
+ }
Qwen2.5-VL-7B-Instruct_1s/lora/sft/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
Qwen2.5-VL-7B-Instruct_30s/lora/sft/README.md ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ license: other
4
+ base_model: /scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct
5
+ tags:
6
+ - llama-factory
7
+ - lora
8
+ - generated_from_trainer
9
+ model-index:
10
+ - name: sft
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # sft
18
+
19
+ This model is a fine-tuned version of [/scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct](https://huggingface.co//scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct) on the GraSP_30s dataset.
20
+
21
+ ## Model description
22
+
23
+ More information needed
24
+
25
+ ## Intended uses & limitations
26
+
27
+ More information needed
28
+
29
+ ## Training and evaluation data
30
+
31
+ More information needed
32
+
33
+ ## Training procedure
34
+
35
+ ### Training hyperparameters
36
+
37
+ The following hyperparameters were used during training:
38
+ - learning_rate: 0.0001
39
+ - train_batch_size: 12
40
+ - eval_batch_size: 8
41
+ - seed: 42
42
+ - distributed_type: multi-GPU
43
+ - num_devices: 2
44
+ - total_train_batch_size: 24
45
+ - total_eval_batch_size: 16
46
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
+ - lr_scheduler_type: cosine
48
+ - lr_scheduler_warmup_ratio: 0.1
49
+ - num_epochs: 1
50
+
51
+ ### Training results
52
+
53
+
54
+
55
+ ### Framework versions
56
+
57
+ - PEFT 0.15.1
58
+ - Transformers 4.51.3
59
+ - Pytorch 2.6.0+cu124
60
+ - Datasets 3.5.0
61
+ - Tokenizers 0.21.1
Qwen2.5-VL-7B-Instruct_30s/lora/sft/adapter_config.json ADDED
@@ -0,0 +1,120 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "/scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 16,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 8,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "layers.17.mlp.down_proj",
28
+ "layers.8.mlp.up_proj",
29
+ "layers.20.mlp.up_proj",
30
+ "layers.17.mlp.up_proj",
31
+ "layers.17.mlp.gate_proj",
32
+ "layers.12.mlp.down_proj",
33
+ "layers.8.mlp.down_proj",
34
+ "layers.4.mlp.down_proj",
35
+ "layers.1.mlp.up_proj",
36
+ "layers.12.mlp.gate_proj",
37
+ "layers.16.mlp.down_proj",
38
+ "layers.13.mlp.up_proj",
39
+ "layers.6.mlp.up_proj",
40
+ "layers.10.mlp.up_proj",
41
+ "layers.7.mlp.gate_proj",
42
+ "layers.12.mlp.up_proj",
43
+ "layers.24.mlp.up_proj",
44
+ "layers.4.mlp.up_proj",
45
+ "layers.11.mlp.gate_proj",
46
+ "layers.10.mlp.gate_proj",
47
+ "layers.6.mlp.gate_proj",
48
+ "layers.5.mlp.up_proj",
49
+ "layers.15.mlp.up_proj",
50
+ "layers.9.mlp.gate_proj",
51
+ "layers.25.mlp.gate_proj",
52
+ "layers.2.mlp.up_proj",
53
+ "layers.19.mlp.up_proj",
54
+ "layers.18.mlp.up_proj",
55
+ "layers.8.mlp.gate_proj",
56
+ "layers.26.mlp.gate_proj",
57
+ "layers.22.mlp.gate_proj",
58
+ "layers.6.mlp.down_proj",
59
+ "layers.25.mlp.up_proj",
60
+ "layers.5.mlp.gate_proj",
61
+ "layers.0.mlp.down_proj",
62
+ "layers.21.mlp.down_proj",
63
+ "layers.0.mlp.gate_proj",
64
+ "layers.2.mlp.gate_proj",
65
+ "layers.4.mlp.gate_proj",
66
+ "layers.18.mlp.down_proj",
67
+ "layers.27.mlp.down_proj",
68
+ "v_proj",
69
+ "layers.14.mlp.gate_proj",
70
+ "layers.14.mlp.up_proj",
71
+ "layers.1.mlp.gate_proj",
72
+ "layers.1.mlp.down_proj",
73
+ "layers.26.mlp.down_proj",
74
+ "layers.22.mlp.up_proj",
75
+ "layers.23.mlp.down_proj",
76
+ "layers.9.mlp.up_proj",
77
+ "layers.13.mlp.gate_proj",
78
+ "layers.9.mlp.down_proj",
79
+ "layers.16.mlp.up_proj",
80
+ "layers.23.mlp.gate_proj",
81
+ "layers.19.mlp.down_proj",
82
+ "layers.7.mlp.down_proj",
83
+ "layers.23.mlp.up_proj",
84
+ "layers.15.mlp.gate_proj",
85
+ "layers.3.mlp.gate_proj",
86
+ "layers.21.mlp.up_proj",
87
+ "layers.2.mlp.down_proj",
88
+ "layers.0.mlp.up_proj",
89
+ "layers.15.mlp.down_proj",
90
+ "o_proj",
91
+ "layers.16.mlp.gate_proj",
92
+ "layers.22.mlp.down_proj",
93
+ "layers.27.mlp.up_proj",
94
+ "layers.20.mlp.down_proj",
95
+ "layers.25.mlp.down_proj",
96
+ "layers.3.mlp.up_proj",
97
+ "layers.11.mlp.up_proj",
98
+ "layers.5.mlp.down_proj",
99
+ "layers.27.mlp.gate_proj",
100
+ "layers.24.mlp.down_proj",
101
+ "layers.13.mlp.down_proj",
102
+ "layers.3.mlp.down_proj",
103
+ "layers.21.mlp.gate_proj",
104
+ "layers.26.mlp.up_proj",
105
+ "layers.19.mlp.gate_proj",
106
+ "q_proj",
107
+ "layers.7.mlp.up_proj",
108
+ "layers.18.mlp.gate_proj",
109
+ "layers.20.mlp.gate_proj",
110
+ "layers.11.mlp.down_proj",
111
+ "layers.14.mlp.down_proj",
112
+ "layers.24.mlp.gate_proj",
113
+ "k_proj",
114
+ "layers.10.mlp.down_proj"
115
+ ],
116
+ "task_type": "CAUSAL_LM",
117
+ "trainable_token_indices": null,
118
+ "use_dora": false,
119
+ "use_rslora": false
120
+ }
Qwen2.5-VL-7B-Instruct_30s/lora/sft/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e648a754dc2d55a518307a2e101c68cbb76524d8d11b20c0b33ce2aa521e4f8
3
+ size 80792096
Qwen2.5-VL-7B-Instruct_30s/lora/sft/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
Qwen2.5-VL-7B-Instruct_30s/lora/sft/chat_template.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n{% endif %}<|im_start|>{{ message['role'] }}\n{% if message['content'] is string %}{{ message['content'] }}<|im_end|>\n{% else %}{% for content in message['content'] %}{% if content['type'] == 'image' or 'image' in content or 'image_url' in content %}{% set image_count.value = image_count.value + 1 %}{% if add_vision_id %}Picture {{ image_count.value }}: {% endif %}<|vision_start|><|image_pad|><|vision_end|>{% elif content['type'] == 'video' or 'video' in content %}{% set video_count.value = video_count.value + 1 %}{% if add_vision_id %}Video {{ video_count.value }}: {% endif %}<|vision_start|><|video_pad|><|vision_end|>{% elif 'text' in content %}{{ content['text'] }}{% endif %}{% endfor %}<|im_end|>\n{% endif %}{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant\n{% endif %}"
3
+ }
Qwen2.5-VL-7B-Instruct_30s/lora/sft/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
Qwen2.5-VL-7B-Instruct_30s/lora/sft/preprocessor_config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": true,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.48145466,
8
+ 0.4578275,
9
+ 0.40821073
10
+ ],
11
+ "image_processor_type": "Qwen2VLImageProcessor",
12
+ "image_std": [
13
+ 0.26862954,
14
+ 0.26130258,
15
+ 0.27577711
16
+ ],
17
+ "max_pixels": 12845056,
18
+ "merge_size": 2,
19
+ "min_pixels": 3136,
20
+ "patch_size": 14,
21
+ "processor_class": "Qwen2_5_VLProcessor",
22
+ "resample": 3,
23
+ "rescale_factor": 0.00392156862745098,
24
+ "size": {
25
+ "longest_edge": 12845056,
26
+ "shortest_edge": 3136
27
+ },
28
+ "temporal_patch_size": 2
29
+ }
Qwen2.5-VL-7B-Instruct_30s/lora/sft/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
Qwen2.5-VL-7B-Instruct_30s/lora/sft/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
Qwen2.5-VL-7B-Instruct_30s/lora/sft/tokenizer_config.json ADDED
@@ -0,0 +1,210 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n{% endif %}<|im_start|>{{ message['role'] }}\n{% if message['content'] is string %}{{ message['content'] }}<|im_end|>\n{% else %}{% for content in message['content'] %}{% if content['type'] == 'image' or 'image' in content or 'image_url' in content %}{% set image_count.value = image_count.value + 1 %}{% if add_vision_id %}Picture {{ image_count.value }}: {% endif %}<|vision_start|><|image_pad|><|vision_end|>{% elif content['type'] == 'video' or 'video' in content %}{% set video_count.value = video_count.value + 1 %}{% if add_vision_id %}Video {{ video_count.value }}: {% endif %}<|vision_start|><|video_pad|><|vision_end|>{% elif 'text' in content %}{{ content['text'] }}{% endif %}{% endfor %}<|im_end|>\n{% endif %}{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant\n{% endif %}",
199
+ "clean_up_tokenization_spaces": false,
200
+ "eos_token": "<|im_end|>",
201
+ "errors": "replace",
202
+ "extra_special_tokens": {},
203
+ "model_max_length": 131072,
204
+ "pad_token": "<|endoftext|>",
205
+ "padding_side": "right",
206
+ "processor_class": "Qwen2_5_VLProcessor",
207
+ "split_special_tokens": false,
208
+ "tokenizer_class": "Qwen2Tokenizer",
209
+ "unk_token": null
210
+ }
Qwen2.5-VL-7B-Instruct_30s/lora/sft/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
Qwen2.5-VL-7B-Instruct_5s/lora/sft/README.md ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ license: other
4
+ base_model: /scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct
5
+ tags:
6
+ - llama-factory
7
+ - lora
8
+ - generated_from_trainer
9
+ model-index:
10
+ - name: sft
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # sft
18
+
19
+ This model is a fine-tuned version of [/scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct](https://huggingface.co//scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct) on the GraSP_5s dataset.
20
+
21
+ ## Model description
22
+
23
+ More information needed
24
+
25
+ ## Intended uses & limitations
26
+
27
+ More information needed
28
+
29
+ ## Training and evaluation data
30
+
31
+ More information needed
32
+
33
+ ## Training procedure
34
+
35
+ ### Training hyperparameters
36
+
37
+ The following hyperparameters were used during training:
38
+ - learning_rate: 0.0001
39
+ - train_batch_size: 12
40
+ - eval_batch_size: 8
41
+ - seed: 42
42
+ - distributed_type: multi-GPU
43
+ - num_devices: 2
44
+ - total_train_batch_size: 24
45
+ - total_eval_batch_size: 16
46
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
+ - lr_scheduler_type: cosine
48
+ - lr_scheduler_warmup_ratio: 0.1
49
+ - num_epochs: 1
50
+
51
+ ### Training results
52
+
53
+
54
+
55
+ ### Framework versions
56
+
57
+ - PEFT 0.15.1
58
+ - Transformers 4.51.3
59
+ - Pytorch 2.6.0+cu124
60
+ - Datasets 3.5.0
61
+ - Tokenizers 0.21.1
Qwen2.5-VL-7B-Instruct_5s/lora/sft/adapter_config.json ADDED
@@ -0,0 +1,120 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "/scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 16,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 8,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "layers.20.mlp.gate_proj",
28
+ "layers.16.mlp.down_proj",
29
+ "layers.16.mlp.up_proj",
30
+ "layers.20.mlp.up_proj",
31
+ "layers.1.mlp.down_proj",
32
+ "layers.2.mlp.gate_proj",
33
+ "layers.27.mlp.gate_proj",
34
+ "layers.26.mlp.down_proj",
35
+ "layers.6.mlp.down_proj",
36
+ "layers.2.mlp.down_proj",
37
+ "v_proj",
38
+ "layers.8.mlp.up_proj",
39
+ "layers.13.mlp.up_proj",
40
+ "layers.12.mlp.up_proj",
41
+ "layers.19.mlp.down_proj",
42
+ "layers.16.mlp.gate_proj",
43
+ "o_proj",
44
+ "layers.17.mlp.gate_proj",
45
+ "layers.10.mlp.gate_proj",
46
+ "layers.27.mlp.up_proj",
47
+ "layers.3.mlp.down_proj",
48
+ "layers.9.mlp.up_proj",
49
+ "layers.25.mlp.down_proj",
50
+ "layers.25.mlp.gate_proj",
51
+ "layers.1.mlp.up_proj",
52
+ "layers.26.mlp.gate_proj",
53
+ "layers.9.mlp.gate_proj",
54
+ "layers.7.mlp.gate_proj",
55
+ "layers.21.mlp.up_proj",
56
+ "layers.19.mlp.gate_proj",
57
+ "layers.5.mlp.up_proj",
58
+ "layers.14.mlp.gate_proj",
59
+ "layers.12.mlp.gate_proj",
60
+ "layers.5.mlp.down_proj",
61
+ "layers.15.mlp.up_proj",
62
+ "layers.14.mlp.up_proj",
63
+ "layers.24.mlp.up_proj",
64
+ "layers.23.mlp.down_proj",
65
+ "layers.21.mlp.gate_proj",
66
+ "layers.12.mlp.down_proj",
67
+ "layers.15.mlp.gate_proj",
68
+ "layers.25.mlp.up_proj",
69
+ "layers.21.mlp.down_proj",
70
+ "layers.20.mlp.down_proj",
71
+ "layers.2.mlp.up_proj",
72
+ "layers.4.mlp.up_proj",
73
+ "layers.23.mlp.up_proj",
74
+ "layers.13.mlp.gate_proj",
75
+ "layers.6.mlp.gate_proj",
76
+ "layers.22.mlp.up_proj",
77
+ "layers.19.mlp.up_proj",
78
+ "layers.0.mlp.down_proj",
79
+ "layers.11.mlp.down_proj",
80
+ "layers.13.mlp.down_proj",
81
+ "layers.8.mlp.gate_proj",
82
+ "layers.24.mlp.down_proj",
83
+ "layers.8.mlp.down_proj",
84
+ "layers.3.mlp.up_proj",
85
+ "layers.4.mlp.gate_proj",
86
+ "layers.7.mlp.down_proj",
87
+ "layers.10.mlp.up_proj",
88
+ "layers.22.mlp.gate_proj",
89
+ "layers.17.mlp.up_proj",
90
+ "layers.0.mlp.up_proj",
91
+ "layers.22.mlp.down_proj",
92
+ "layers.18.mlp.up_proj",
93
+ "layers.4.mlp.down_proj",
94
+ "layers.10.mlp.down_proj",
95
+ "layers.3.mlp.gate_proj",
96
+ "layers.18.mlp.down_proj",
97
+ "layers.9.mlp.down_proj",
98
+ "layers.0.mlp.gate_proj",
99
+ "layers.23.mlp.gate_proj",
100
+ "layers.5.mlp.gate_proj",
101
+ "layers.26.mlp.up_proj",
102
+ "layers.1.mlp.gate_proj",
103
+ "layers.6.mlp.up_proj",
104
+ "layers.7.mlp.up_proj",
105
+ "layers.27.mlp.down_proj",
106
+ "layers.15.mlp.down_proj",
107
+ "layers.11.mlp.gate_proj",
108
+ "layers.18.mlp.gate_proj",
109
+ "layers.24.mlp.gate_proj",
110
+ "q_proj",
111
+ "layers.11.mlp.up_proj",
112
+ "layers.14.mlp.down_proj",
113
+ "layers.17.mlp.down_proj",
114
+ "k_proj"
115
+ ],
116
+ "task_type": "CAUSAL_LM",
117
+ "trainable_token_indices": null,
118
+ "use_dora": false,
119
+ "use_rslora": false
120
+ }
Qwen2.5-VL-7B-Instruct_5s/lora/sft/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9470bf1c72518a0285e1e93c3f963b65cd2cf30d06ff18caa80f1aefd965e7c0
3
+ size 80792096
Qwen2.5-VL-7B-Instruct_5s/lora/sft/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
Qwen2.5-VL-7B-Instruct_5s/lora/sft/chat_template.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n{% endif %}<|im_start|>{{ message['role'] }}\n{% if message['content'] is string %}{{ message['content'] }}<|im_end|>\n{% else %}{% for content in message['content'] %}{% if content['type'] == 'image' or 'image' in content or 'image_url' in content %}{% set image_count.value = image_count.value + 1 %}{% if add_vision_id %}Picture {{ image_count.value }}: {% endif %}<|vision_start|><|image_pad|><|vision_end|>{% elif content['type'] == 'video' or 'video' in content %}{% set video_count.value = video_count.value + 1 %}{% if add_vision_id %}Video {{ video_count.value }}: {% endif %}<|vision_start|><|video_pad|><|vision_end|>{% elif 'text' in content %}{{ content['text'] }}{% endif %}{% endfor %}<|im_end|>\n{% endif %}{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant\n{% endif %}"
3
+ }
Qwen2.5-VL-7B-Instruct_5s/lora/sft/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
Qwen2.5-VL-7B-Instruct_5s/lora/sft/preprocessor_config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": true,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.48145466,
8
+ 0.4578275,
9
+ 0.40821073
10
+ ],
11
+ "image_processor_type": "Qwen2VLImageProcessor",
12
+ "image_std": [
13
+ 0.26862954,
14
+ 0.26130258,
15
+ 0.27577711
16
+ ],
17
+ "max_pixels": 12845056,
18
+ "merge_size": 2,
19
+ "min_pixels": 3136,
20
+ "patch_size": 14,
21
+ "processor_class": "Qwen2_5_VLProcessor",
22
+ "resample": 3,
23
+ "rescale_factor": 0.00392156862745098,
24
+ "size": {
25
+ "longest_edge": 12845056,
26
+ "shortest_edge": 3136
27
+ },
28
+ "temporal_patch_size": 2
29
+ }
Qwen2.5-VL-7B-Instruct_5s/lora/sft/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
Qwen2.5-VL-7B-Instruct_5s/lora/sft/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
Qwen2.5-VL-7B-Instruct_5s/lora/sft/tokenizer_config.json ADDED
@@ -0,0 +1,210 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n{% endif %}<|im_start|>{{ message['role'] }}\n{% if message['content'] is string %}{{ message['content'] }}<|im_end|>\n{% else %}{% for content in message['content'] %}{% if content['type'] == 'image' or 'image' in content or 'image_url' in content %}{% set image_count.value = image_count.value + 1 %}{% if add_vision_id %}Picture {{ image_count.value }}: {% endif %}<|vision_start|><|image_pad|><|vision_end|>{% elif content['type'] == 'video' or 'video' in content %}{% set video_count.value = video_count.value + 1 %}{% if add_vision_id %}Video {{ video_count.value }}: {% endif %}<|vision_start|><|video_pad|><|vision_end|>{% elif 'text' in content %}{{ content['text'] }}{% endif %}{% endfor %}<|im_end|>\n{% endif %}{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant\n{% endif %}",
199
+ "clean_up_tokenization_spaces": false,
200
+ "eos_token": "<|im_end|>",
201
+ "errors": "replace",
202
+ "extra_special_tokens": {},
203
+ "model_max_length": 131072,
204
+ "pad_token": "<|endoftext|>",
205
+ "padding_side": "right",
206
+ "processor_class": "Qwen2_5_VLProcessor",
207
+ "split_special_tokens": false,
208
+ "tokenizer_class": "Qwen2Tokenizer",
209
+ "unk_token": null
210
+ }
Qwen2.5-VL-7B-Instruct_5s/lora/sft/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
Qwen2.5-VL-7B-Instruct_60s/lora/sft/README.md ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ license: other
4
+ base_model: /scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct
5
+ tags:
6
+ - llama-factory
7
+ - lora
8
+ - generated_from_trainer
9
+ model-index:
10
+ - name: sft
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # sft
18
+
19
+ This model is a fine-tuned version of [/scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct](https://huggingface.co//scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct) on the GraSP_60s dataset.
20
+
21
+ ## Model description
22
+
23
+ More information needed
24
+
25
+ ## Intended uses & limitations
26
+
27
+ More information needed
28
+
29
+ ## Training and evaluation data
30
+
31
+ More information needed
32
+
33
+ ## Training procedure
34
+
35
+ ### Training hyperparameters
36
+
37
+ The following hyperparameters were used during training:
38
+ - learning_rate: 0.0001
39
+ - train_batch_size: 12
40
+ - eval_batch_size: 8
41
+ - seed: 42
42
+ - distributed_type: multi-GPU
43
+ - num_devices: 2
44
+ - total_train_batch_size: 24
45
+ - total_eval_batch_size: 16
46
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
+ - lr_scheduler_type: cosine
48
+ - lr_scheduler_warmup_ratio: 0.1
49
+ - num_epochs: 1
50
+
51
+ ### Training results
52
+
53
+
54
+
55
+ ### Framework versions
56
+
57
+ - PEFT 0.15.1
58
+ - Transformers 4.51.3
59
+ - Pytorch 2.6.0+cu124
60
+ - Datasets 3.5.0
61
+ - Tokenizers 0.21.1
Qwen2.5-VL-7B-Instruct_60s/lora/sft/adapter_config.json ADDED
@@ -0,0 +1,120 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "/scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 16,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 8,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "layers.4.mlp.up_proj",
28
+ "layers.23.mlp.gate_proj",
29
+ "layers.1.mlp.up_proj",
30
+ "layers.8.mlp.gate_proj",
31
+ "layers.22.mlp.down_proj",
32
+ "layers.27.mlp.gate_proj",
33
+ "layers.22.mlp.up_proj",
34
+ "layers.10.mlp.gate_proj",
35
+ "layers.19.mlp.gate_proj",
36
+ "layers.16.mlp.gate_proj",
37
+ "layers.25.mlp.up_proj",
38
+ "layers.26.mlp.up_proj",
39
+ "layers.26.mlp.gate_proj",
40
+ "layers.8.mlp.up_proj",
41
+ "layers.1.mlp.down_proj",
42
+ "layers.24.mlp.gate_proj",
43
+ "layers.0.mlp.gate_proj",
44
+ "layers.18.mlp.down_proj",
45
+ "layers.17.mlp.up_proj",
46
+ "layers.2.mlp.up_proj",
47
+ "layers.3.mlp.gate_proj",
48
+ "layers.27.mlp.up_proj",
49
+ "layers.11.mlp.up_proj",
50
+ "layers.13.mlp.up_proj",
51
+ "layers.0.mlp.up_proj",
52
+ "layers.4.mlp.down_proj",
53
+ "o_proj",
54
+ "layers.21.mlp.down_proj",
55
+ "layers.24.mlp.up_proj",
56
+ "layers.15.mlp.down_proj",
57
+ "layers.20.mlp.up_proj",
58
+ "k_proj",
59
+ "layers.9.mlp.gate_proj",
60
+ "layers.23.mlp.up_proj",
61
+ "layers.9.mlp.down_proj",
62
+ "layers.3.mlp.down_proj",
63
+ "layers.7.mlp.up_proj",
64
+ "layers.13.mlp.down_proj",
65
+ "layers.12.mlp.gate_proj",
66
+ "layers.15.mlp.up_proj",
67
+ "layers.5.mlp.gate_proj",
68
+ "layers.12.mlp.down_proj",
69
+ "layers.14.mlp.down_proj",
70
+ "layers.12.mlp.up_proj",
71
+ "layers.5.mlp.up_proj",
72
+ "q_proj",
73
+ "layers.16.mlp.down_proj",
74
+ "layers.25.mlp.down_proj",
75
+ "layers.6.mlp.down_proj",
76
+ "layers.20.mlp.gate_proj",
77
+ "layers.24.mlp.down_proj",
78
+ "layers.21.mlp.gate_proj",
79
+ "layers.26.mlp.down_proj",
80
+ "layers.18.mlp.up_proj",
81
+ "layers.27.mlp.down_proj",
82
+ "layers.25.mlp.gate_proj",
83
+ "layers.23.mlp.down_proj",
84
+ "layers.9.mlp.up_proj",
85
+ "layers.10.mlp.down_proj",
86
+ "layers.6.mlp.gate_proj",
87
+ "layers.5.mlp.down_proj",
88
+ "layers.10.mlp.up_proj",
89
+ "layers.8.mlp.down_proj",
90
+ "layers.20.mlp.down_proj",
91
+ "layers.15.mlp.gate_proj",
92
+ "layers.6.mlp.up_proj",
93
+ "layers.21.mlp.up_proj",
94
+ "layers.7.mlp.gate_proj",
95
+ "layers.2.mlp.gate_proj",
96
+ "layers.19.mlp.up_proj",
97
+ "layers.13.mlp.gate_proj",
98
+ "layers.19.mlp.down_proj",
99
+ "layers.4.mlp.gate_proj",
100
+ "layers.18.mlp.gate_proj",
101
+ "layers.17.mlp.gate_proj",
102
+ "layers.17.mlp.down_proj",
103
+ "layers.2.mlp.down_proj",
104
+ "layers.1.mlp.gate_proj",
105
+ "layers.7.mlp.down_proj",
106
+ "layers.0.mlp.down_proj",
107
+ "layers.22.mlp.gate_proj",
108
+ "layers.14.mlp.gate_proj",
109
+ "layers.3.mlp.up_proj",
110
+ "layers.11.mlp.down_proj",
111
+ "v_proj",
112
+ "layers.11.mlp.gate_proj",
113
+ "layers.16.mlp.up_proj",
114
+ "layers.14.mlp.up_proj"
115
+ ],
116
+ "task_type": "CAUSAL_LM",
117
+ "trainable_token_indices": null,
118
+ "use_dora": false,
119
+ "use_rslora": false
120
+ }
Qwen2.5-VL-7B-Instruct_60s/lora/sft/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da9d6af08d0805af81b7281b51a6a136b00cb61c6e2093f13fb4aa085d97f109
3
+ size 80792096
Qwen2.5-VL-7B-Instruct_60s/lora/sft/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
Qwen2.5-VL-7B-Instruct_60s/lora/sft/chat_template.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n{% endif %}<|im_start|>{{ message['role'] }}\n{% if message['content'] is string %}{{ message['content'] }}<|im_end|>\n{% else %}{% for content in message['content'] %}{% if content['type'] == 'image' or 'image' in content or 'image_url' in content %}{% set image_count.value = image_count.value + 1 %}{% if add_vision_id %}Picture {{ image_count.value }}: {% endif %}<|vision_start|><|image_pad|><|vision_end|>{% elif content['type'] == 'video' or 'video' in content %}{% set video_count.value = video_count.value + 1 %}{% if add_vision_id %}Video {{ video_count.value }}: {% endif %}<|vision_start|><|video_pad|><|vision_end|>{% elif 'text' in content %}{{ content['text'] }}{% endif %}{% endfor %}<|im_end|>\n{% endif %}{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant\n{% endif %}"
3
+ }
Qwen2.5-VL-7B-Instruct_60s/lora/sft/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
Qwen2.5-VL-7B-Instruct_60s/lora/sft/preprocessor_config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": true,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.48145466,
8
+ 0.4578275,
9
+ 0.40821073
10
+ ],
11
+ "image_processor_type": "Qwen2VLImageProcessor",
12
+ "image_std": [
13
+ 0.26862954,
14
+ 0.26130258,
15
+ 0.27577711
16
+ ],
17
+ "max_pixels": 12845056,
18
+ "merge_size": 2,
19
+ "min_pixels": 3136,
20
+ "patch_size": 14,
21
+ "processor_class": "Qwen2_5_VLProcessor",
22
+ "resample": 3,
23
+ "rescale_factor": 0.00392156862745098,
24
+ "size": {
25
+ "longest_edge": 12845056,
26
+ "shortest_edge": 3136
27
+ },
28
+ "temporal_patch_size": 2
29
+ }
Qwen2.5-VL-7B-Instruct_60s/lora/sft/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
Qwen2.5-VL-7B-Instruct_60s/lora/sft/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
Qwen2.5-VL-7B-Instruct_60s/lora/sft/tokenizer_config.json ADDED
@@ -0,0 +1,210 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n{% endif %}<|im_start|>{{ message['role'] }}\n{% if message['content'] is string %}{{ message['content'] }}<|im_end|>\n{% else %}{% for content in message['content'] %}{% if content['type'] == 'image' or 'image' in content or 'image_url' in content %}{% set image_count.value = image_count.value + 1 %}{% if add_vision_id %}Picture {{ image_count.value }}: {% endif %}<|vision_start|><|image_pad|><|vision_end|>{% elif content['type'] == 'video' or 'video' in content %}{% set video_count.value = video_count.value + 1 %}{% if add_vision_id %}Video {{ video_count.value }}: {% endif %}<|vision_start|><|video_pad|><|vision_end|>{% elif 'text' in content %}{{ content['text'] }}{% endif %}{% endfor %}<|im_end|>\n{% endif %}{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant\n{% endif %}",
199
+ "clean_up_tokenization_spaces": false,
200
+ "eos_token": "<|im_end|>",
201
+ "errors": "replace",
202
+ "extra_special_tokens": {},
203
+ "model_max_length": 131072,
204
+ "pad_token": "<|endoftext|>",
205
+ "padding_side": "right",
206
+ "processor_class": "Qwen2_5_VLProcessor",
207
+ "split_special_tokens": false,
208
+ "tokenizer_class": "Qwen2Tokenizer",
209
+ "unk_token": null
210
+ }
Qwen2.5-VL-7B-Instruct_60s/lora/sft/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
Qwen2.5-VL-7B-Instruct_chain1_1s/lora/sft/README.md ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ license: other
4
+ base_model: /scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct
5
+ tags:
6
+ - llama-factory
7
+ - lora
8
+ - generated_from_trainer
9
+ model-index:
10
+ - name: sft
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # sft
18
+
19
+ This model is a fine-tuned version of [/scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct](https://huggingface.co//scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct) on the GraSP_chain1_1s dataset.
20
+
21
+ ## Model description
22
+
23
+ More information needed
24
+
25
+ ## Intended uses & limitations
26
+
27
+ More information needed
28
+
29
+ ## Training and evaluation data
30
+
31
+ More information needed
32
+
33
+ ## Training procedure
34
+
35
+ ### Training hyperparameters
36
+
37
+ The following hyperparameters were used during training:
38
+ - learning_rate: 0.0001
39
+ - train_batch_size: 36
40
+ - eval_batch_size: 8
41
+ - seed: 42
42
+ - distributed_type: multi-GPU
43
+ - num_devices: 2
44
+ - total_train_batch_size: 72
45
+ - total_eval_batch_size: 16
46
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
+ - lr_scheduler_type: cosine
48
+ - lr_scheduler_warmup_ratio: 0.1
49
+ - num_epochs: 1
50
+
51
+ ### Training results
52
+
53
+
54
+
55
+ ### Framework versions
56
+
57
+ - PEFT 0.15.1
58
+ - Transformers 4.51.3
59
+ - Pytorch 2.6.0+cu124
60
+ - Datasets 3.5.0
61
+ - Tokenizers 0.21.1
Qwen2.5-VL-7B-Instruct_chain1_1s/lora/sft/adapter_config.json ADDED
@@ -0,0 +1,120 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "/scratch/e1374390/LLaMA-Factory/pretrained/Qwen2.5-VL-7B-Instruct",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 16,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 8,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "layers.10.mlp.up_proj",
28
+ "layers.24.mlp.gate_proj",
29
+ "layers.2.mlp.up_proj",
30
+ "layers.14.mlp.gate_proj",
31
+ "layers.9.mlp.up_proj",
32
+ "layers.26.mlp.up_proj",
33
+ "layers.3.mlp.gate_proj",
34
+ "layers.27.mlp.down_proj",
35
+ "layers.5.mlp.down_proj",
36
+ "layers.11.mlp.gate_proj",
37
+ "k_proj",
38
+ "layers.0.mlp.up_proj",
39
+ "layers.10.mlp.down_proj",
40
+ "layers.16.mlp.down_proj",
41
+ "layers.25.mlp.down_proj",
42
+ "layers.16.mlp.gate_proj",
43
+ "layers.4.mlp.down_proj",
44
+ "layers.18.mlp.gate_proj",
45
+ "layers.4.mlp.up_proj",
46
+ "layers.1.mlp.down_proj",
47
+ "layers.21.mlp.up_proj",
48
+ "layers.8.mlp.gate_proj",
49
+ "layers.19.mlp.down_proj",
50
+ "layers.6.mlp.down_proj",
51
+ "layers.2.mlp.gate_proj",
52
+ "layers.15.mlp.down_proj",
53
+ "layers.8.mlp.down_proj",
54
+ "layers.13.mlp.down_proj",
55
+ "layers.25.mlp.gate_proj",
56
+ "layers.12.mlp.up_proj",
57
+ "layers.27.mlp.up_proj",
58
+ "layers.11.mlp.up_proj",
59
+ "layers.18.mlp.up_proj",
60
+ "layers.24.mlp.down_proj",
61
+ "layers.26.mlp.gate_proj",
62
+ "layers.13.mlp.gate_proj",
63
+ "layers.9.mlp.gate_proj",
64
+ "layers.7.mlp.gate_proj",
65
+ "layers.5.mlp.up_proj",
66
+ "layers.7.mlp.up_proj",
67
+ "layers.23.mlp.up_proj",
68
+ "layers.23.mlp.down_proj",
69
+ "layers.14.mlp.down_proj",
70
+ "layers.21.mlp.down_proj",
71
+ "layers.12.mlp.gate_proj",
72
+ "layers.0.mlp.down_proj",
73
+ "layers.13.mlp.up_proj",
74
+ "layers.23.mlp.gate_proj",
75
+ "layers.22.mlp.down_proj",
76
+ "layers.18.mlp.down_proj",
77
+ "q_proj",
78
+ "layers.11.mlp.down_proj",
79
+ "o_proj",
80
+ "layers.17.mlp.up_proj",
81
+ "layers.26.mlp.down_proj",
82
+ "layers.19.mlp.gate_proj",
83
+ "layers.17.mlp.gate_proj",
84
+ "layers.7.mlp.down_proj",
85
+ "layers.16.mlp.up_proj",
86
+ "layers.0.mlp.gate_proj",
87
+ "layers.5.mlp.gate_proj",
88
+ "layers.22.mlp.up_proj",
89
+ "layers.21.mlp.gate_proj",
90
+ "layers.22.mlp.gate_proj",
91
+ "layers.15.mlp.gate_proj",
92
+ "layers.17.mlp.down_proj",
93
+ "v_proj",
94
+ "layers.19.mlp.up_proj",
95
+ "layers.3.mlp.down_proj",
96
+ "layers.4.mlp.gate_proj",
97
+ "layers.9.mlp.down_proj",
98
+ "layers.10.mlp.gate_proj",
99
+ "layers.1.mlp.gate_proj",
100
+ "layers.14.mlp.up_proj",
101
+ "layers.20.mlp.gate_proj",
102
+ "layers.25.mlp.up_proj",
103
+ "layers.2.mlp.down_proj",
104
+ "layers.20.mlp.down_proj",
105
+ "layers.6.mlp.up_proj",
106
+ "layers.1.mlp.up_proj",
107
+ "layers.20.mlp.up_proj",
108
+ "layers.15.mlp.up_proj",
109
+ "layers.27.mlp.gate_proj",
110
+ "layers.3.mlp.up_proj",
111
+ "layers.8.mlp.up_proj",
112
+ "layers.12.mlp.down_proj",
113
+ "layers.24.mlp.up_proj",
114
+ "layers.6.mlp.gate_proj"
115
+ ],
116
+ "task_type": "CAUSAL_LM",
117
+ "trainable_token_indices": null,
118
+ "use_dora": false,
119
+ "use_rslora": false
120
+ }
Qwen2.5-VL-7B-Instruct_chain1_1s/lora/sft/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65a592d933f81d2743cb62e93a52a4bc90097e4a839ef0d2f2050bac71a44c7c
3
+ size 80792096
Qwen2.5-VL-7B-Instruct_chain1_1s/lora/sft/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
Qwen2.5-VL-7B-Instruct_chain1_1s/lora/sft/chat_template.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n{% endif %}<|im_start|>{{ message['role'] }}\n{% if message['content'] is string %}{{ message['content'] }}<|im_end|>\n{% else %}{% for content in message['content'] %}{% if content['type'] == 'image' or 'image' in content or 'image_url' in content %}{% set image_count.value = image_count.value + 1 %}{% if add_vision_id %}Picture {{ image_count.value }}: {% endif %}<|vision_start|><|image_pad|><|vision_end|>{% elif content['type'] == 'video' or 'video' in content %}{% set video_count.value = video_count.value + 1 %}{% if add_vision_id %}Video {{ video_count.value }}: {% endif %}<|vision_start|><|video_pad|><|vision_end|>{% elif 'text' in content %}{{ content['text'] }}{% endif %}{% endfor %}<|im_end|>\n{% endif %}{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant\n{% endif %}"
3
+ }