Delete 27b-16lora-lr1e5-16lora

#1
by paidaixing - opened
27b-16lora-lr1e5-16lora/Modelfile DELETED
@@ -1,15 +0,0 @@
1
- # ollama modelfile auto-generated by llamafactory
2
-
3
- FROM .
4
-
5
- TEMPLATE """<bos>{{ if .System }}{{ .System }}
6
-
7
- {{ end }}{{ range .Messages }}{{ if eq .Role "user" }}<start_of_turn>user
8
- {{ .Content }}<end_of_turn>
9
- <start_of_turn>model
10
- {{ else if eq .Role "assistant" }}{{ .Content }}<end_of_turn>
11
- {{ end }}{{ end }}"""
12
-
13
- PARAMETER stop "<eos>"
14
- PARAMETER stop "<end_of_turn>"
15
- PARAMETER num_ctx 4096
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
27b-16lora-lr1e5-16lora/added_tokens.json DELETED
@@ -1,3 +0,0 @@
1
- {
2
- "<image_soft_token>": 262144
3
- }
 
 
 
 
27b-16lora-lr1e5-16lora/chat_template.json DELETED
@@ -1,3 +0,0 @@
1
- {
2
- "chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[0]['content'] is string -%}\n {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}\n {%- else -%}\n {%- set first_user_prefix = messages[0]['content'][0]['text'] + '\n\n' -%}\n {%- endif -%}\n {%- set loop_messages = messages[1:] -%}\n{%- else -%}\n {%- set first_user_prefix = \"\" -%}\n {%- set loop_messages = messages -%}\n{%- endif -%}\n{%- for message in loop_messages -%}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) -%}\n {{ raise_exception(\"Conversation roles must alternate user/assistant/user/assistant/...\") }}\n {%- endif -%}\n {%- if (message['role'] == 'assistant') -%}\n {%- set role = \"model\" -%}\n {%- else -%}\n {%- set role = message['role'] -%}\n {%- endif -%}\n {{ '<start_of_turn>' + role + '\n' + (first_user_prefix if loop.first else \"\") }}\n {%- if message['content'] is string -%}\n {{ message['content'] | trim }}\n {%- elif message['content'] is iterable -%}\n {%- for item in message['content'] -%}\n {%- if item['type'] == 'image' -%}\n {{ '<start_of_image>' }}\n {%- elif item['type'] == 'text' -%}\n {{ item['text'] | trim }}\n {%- endif -%}\n {%- endfor -%}\n {%- else -%}\n {{ raise_exception(\"Invalid content type\") }}\n {%- endif -%}\n {{ '<end_of_turn>\n' }}\n{%- endfor -%}\n{%- if add_generation_prompt -%}\n {{'<start_of_turn>model\n'}}\n{%- endif -%}\n"
3
- }
 
 
 
 
27b-16lora-lr1e5-16lora/config.json DELETED
@@ -1,64 +0,0 @@
1
- {
2
- "architectures": [
3
- "Gemma3ForConditionalGeneration"
4
- ],
5
- "boi_token_index": 255999,
6
- "eoi_token_index": 256000,
7
- "eos_token_id": [
8
- 1,
9
- 106
10
- ],
11
- "hidden_size": 5376,
12
- "image_token_index": 262144,
13
- "initializer_range": 0.02,
14
- "mm_tokens_per_image": 256,
15
- "model_type": "gemma3",
16
- "text_config": {
17
- "attention_bias": false,
18
- "attention_dropout": 0.0,
19
- "attn_logit_softcapping": null,
20
- "cache_implementation": "hybrid",
21
- "final_logit_softcapping": null,
22
- "head_dim": 128,
23
- "hidden_activation": "gelu_pytorch_tanh",
24
- "hidden_size": 5376,
25
- "initializer_range": 0.02,
26
- "intermediate_size": 21504,
27
- "max_position_embeddings": 131072,
28
- "model_type": "gemma3_text",
29
- "num_attention_heads": 32,
30
- "num_hidden_layers": 62,
31
- "num_key_value_heads": 16,
32
- "query_pre_attn_scalar": 168,
33
- "rms_norm_eps": 1e-06,
34
- "rope_local_base_freq": 10000.0,
35
- "rope_scaling": {
36
- "factor": 8.0,
37
- "rope_type": "linear"
38
- },
39
- "rope_theta": 1000000.0,
40
- "sliding_window": 1024,
41
- "sliding_window_pattern": 6,
42
- "torch_dtype": "bfloat16",
43
- "use_cache": true,
44
- "vocab_size": 262208
45
- },
46
- "torch_dtype": "bfloat16",
47
- "transformers_version": "4.51.3",
48
- "use_cache": true,
49
- "vision_config": {
50
- "attention_dropout": 0.0,
51
- "hidden_act": "gelu_pytorch_tanh",
52
- "hidden_size": 1152,
53
- "image_size": 896,
54
- "intermediate_size": 4304,
55
- "layer_norm_eps": 1e-06,
56
- "model_type": "siglip_vision_model",
57
- "num_attention_heads": 16,
58
- "num_channels": 3,
59
- "num_hidden_layers": 27,
60
- "patch_size": 14,
61
- "torch_dtype": "bfloat16",
62
- "vision_use_head": false
63
- }
64
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
27b-16lora-lr1e5-16lora/model-00001-of-00012.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f9f4ef6d018b8d6c75fb54b4bcdc03d4c42f9f952ed6ccb440095332c406a4c
3
- size 4854573696
 
 
 
 
27b-16lora-lr1e5-16lora/model-00002-of-00012.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:592d7f349a78eff5c000e7d57a40ff3f91d04f4786803e5c74659c6416f27f76
3
- size 4954792944
 
 
 
 
27b-16lora-lr1e5-16lora/model-00003-of-00012.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6efaf32eba6d6708e67217e6effa6b6f76c516b2eabf293c39d338e7960b18c9
3
- size 4954792976
 
 
 
 
27b-16lora-lr1e5-16lora/model-00005-of-00012.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7d900a281385a4b3ca665524d699a0050a6be0169ed68742c1b59e6830db2b92
3
- size 4954793016
 
 
 
 
27b-16lora-lr1e5-16lora/model-00008-of-00012.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:ad31a01c2eca1bd3611b4ec468770227a2ea0abe70389f38d4a8ce9bbd966e8e
3
- size 4954793016
 
 
 
 
27b-16lora-lr1e5-16lora/model-00009-of-00012.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e821a3ec9282786fddeee5bd2fa9f2331ed1f7c9649ed06361d1d591c08c8248
3
- size 4954793016
 
 
 
 
27b-16lora-lr1e5-16lora/model-00010-of-00012.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:97e26818f94a64f65ac771e7a6ff53eb8efc8bca46440af5a6cb470ef099d1ef
3
- size 4954793016
 
 
 
 
27b-16lora-lr1e5-16lora/model-00011-of-00012.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:84b61af3a081a8fecaaf17d9c4506f67c5d719dee0fa2b970c37334743163dc5
3
- size 4954793016
 
 
 
 
27b-16lora-lr1e5-16lora/model-00012-of-00012.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:cfa9d349de5d8f9924d1066a11106a2861028d26c08bb6298269dbb63cf9e311
3
- size 462476696
 
 
 
 
27b-16lora-lr1e5-16lora/model.safetensors.index.json DELETED
The diff for this file is too large to render. See raw diff
 
27b-16lora-lr1e5-16lora/preprocessor_config.json DELETED
@@ -1,29 +0,0 @@
1
- {
2
- "do_convert_rgb": null,
3
- "do_normalize": true,
4
- "do_pan_and_scan": null,
5
- "do_rescale": true,
6
- "do_resize": true,
7
- "image_mean": [
8
- 0.5,
9
- 0.5,
10
- 0.5
11
- ],
12
- "image_processor_type": "Gemma3ImageProcessor",
13
- "image_seq_length": 256,
14
- "image_std": [
15
- 0.5,
16
- 0.5,
17
- 0.5
18
- ],
19
- "pan_and_scan_max_num_crops": null,
20
- "pan_and_scan_min_crop_size": null,
21
- "pan_and_scan_min_ratio_to_activate": null,
22
- "processor_class": "Gemma3Processor",
23
- "resample": 2,
24
- "rescale_factor": 0.00392156862745098,
25
- "size": {
26
- "height": 896,
27
- "width": 896
28
- }
29
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
27b-16lora-lr1e5-16lora/processor_config.json DELETED
@@ -1,4 +0,0 @@
1
- {
2
- "image_seq_length": 256,
3
- "processor_class": "Gemma3Processor"
4
- }
 
 
 
 
 
27b-16lora-lr1e5-16lora/readme.md DELETED
@@ -1 +0,0 @@
1
- lora=16_lr=1e-5,
 
 
27b-16lora-lr1e5-16lora/special_tokens_map.json DELETED
@@ -1,42 +0,0 @@
1
- {
2
- "additional_special_tokens": [
3
- {
4
- "content": "<end_of_turn>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false
9
- }
10
- ],
11
- "boi_token": "<start_of_image>",
12
- "bos_token": {
13
- "content": "<bos>",
14
- "lstrip": false,
15
- "normalized": false,
16
- "rstrip": false,
17
- "single_word": false
18
- },
19
- "eoi_token": "<end_of_image>",
20
- "eos_token": {
21
- "content": "<eos>",
22
- "lstrip": false,
23
- "normalized": false,
24
- "rstrip": false,
25
- "single_word": false
26
- },
27
- "image_token": "<image_soft_token>",
28
- "pad_token": {
29
- "content": "<pad>",
30
- "lstrip": false,
31
- "normalized": false,
32
- "rstrip": false,
33
- "single_word": false
34
- },
35
- "unk_token": {
36
- "content": "<unk>",
37
- "lstrip": false,
38
- "normalized": false,
39
- "rstrip": false,
40
- "single_word": false
41
- }
42
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
27b-16lora-lr1e5-16lora/tokenizer.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:4667f2089529e8e7657cfb6d1c19910ae71ff5f28aa7ab2ff2763330affad795
3
- size 33384568