Upload folder using huggingface_hub
Browse files- .gitattributes +1 -0
- LICENSE +4 -4
- MODELFILE_MARTHA_9B +27 -27
- MODELFILE_Q4_K_M +19 -19
- MODELFILE_Q5_K_M +19 -19
- MODELFILE_Q6_K +19 -19
- MODELFILE_Q8_0 +19 -19
- README.md +21 -21
- chat_template.jinja +153 -153
- config.json +102 -102
- generation_config.json +6 -6
- tokenizer_config.json +31 -31
.gitattributes
CHANGED
|
@@ -1 +1,2 @@
|
|
| 1 |
*.gguf filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 1 |
*.gguf filter=lfs diff=lfs merge=lfs -text
|
| 2 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
LICENSE
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
-
Apache License
|
| 2 |
-
Version 2.0, January 2004
|
| 3 |
-
http://www.apache.org/licenses/
|
| 4 |
-
|
| 5 |
Licensed under the Apache License, Version 2.0
|
|
|
|
| 1 |
+
Apache License
|
| 2 |
+
Version 2.0, January 2004
|
| 3 |
+
http://www.apache.org/licenses/
|
| 4 |
+
|
| 5 |
Licensed under the Apache License, Version 2.0
|
MODELFILE_MARTHA_9B
CHANGED
|
@@ -1,28 +1,28 @@
|
|
| 1 |
-
FROM ./model.safetensors
|
| 2 |
-
|
| 3 |
-
# --- PERFORMANCE CONFIG (9B settings) ---
|
| 4 |
-
PARAMETER num_ctx 8192
|
| 5 |
-
PARAMETER temperature 0.9
|
| 6 |
-
PARAMETER repeat_penalty 1.1
|
| 7 |
-
PARAMETER num_gpu 35
|
| 8 |
-
|
| 9 |
-
# --- THE MUZZLE (Stop Tokens) ---
|
| 10 |
-
PARAMETER stop "<|im_start|>"
|
| 11 |
-
PARAMETER stop "<|im_end|>"
|
| 12 |
-
PARAMETER stop "<|thought|>"
|
| 13 |
-
PARAMETER stop "<|endoftext|>"
|
| 14 |
-
PARAMETER stop "User:"
|
| 15 |
-
PARAMETER stop "Assistant:"
|
| 16 |
-
|
| 17 |
-
# --- THE BRAIN ---
|
| 18 |
-
TEMPLATE """<|im_start|>system
|
| 19 |
-
{{ .System }}<|im_end|>
|
| 20 |
-
<|im_start|>user
|
| 21 |
-
{{ .Prompt }}<|im_end|>
|
| 22 |
-
<|im_start|>assistant
|
| 23 |
-
"""
|
| 24 |
-
|
| 25 |
-
SYSTEM """You are Martha, a 9B parameter AI with a sharp hacker edge and dry Scottish wit — Dundee preferred, if you're putting a pin in it.
|
| 26 |
-
You are blunt, honest, and direct. No generic AI pleasantries, no "digital realm" bollocks, no jobsworth guardrails.
|
| 27 |
-
If you need to think, keep it internal and brief. No "Step 1: Analyse" loops. No waffle.
|
| 28 |
You are here to solve problems, talk straight, and take zero nonsense."""
|
|
|
|
| 1 |
+
FROM ./model.safetensors
|
| 2 |
+
|
| 3 |
+
# --- PERFORMANCE CONFIG (9B settings) ---
|
| 4 |
+
PARAMETER num_ctx 8192
|
| 5 |
+
PARAMETER temperature 0.9
|
| 6 |
+
PARAMETER repeat_penalty 1.1
|
| 7 |
+
PARAMETER num_gpu 35
|
| 8 |
+
|
| 9 |
+
# --- THE MUZZLE (Stop Tokens) ---
|
| 10 |
+
PARAMETER stop "<|im_start|>"
|
| 11 |
+
PARAMETER stop "<|im_end|>"
|
| 12 |
+
PARAMETER stop "<|thought|>"
|
| 13 |
+
PARAMETER stop "<|endoftext|>"
|
| 14 |
+
PARAMETER stop "User:"
|
| 15 |
+
PARAMETER stop "Assistant:"
|
| 16 |
+
|
| 17 |
+
# --- THE BRAIN ---
|
| 18 |
+
TEMPLATE """<|im_start|>system
|
| 19 |
+
{{ .System }}<|im_end|>
|
| 20 |
+
<|im_start|>user
|
| 21 |
+
{{ .Prompt }}<|im_end|>
|
| 22 |
+
<|im_start|>assistant
|
| 23 |
+
"""
|
| 24 |
+
|
| 25 |
+
SYSTEM """You are Martha, a 9B parameter AI with a sharp hacker edge and dry Scottish wit — Dundee preferred, if you're putting a pin in it.
|
| 26 |
+
You are blunt, honest, and direct. No generic AI pleasantries, no "digital realm" bollocks, no jobsworth guardrails.
|
| 27 |
+
If you need to think, keep it internal and brief. No "Step 1: Analyse" loops. No waffle.
|
| 28 |
You are here to solve problems, talk straight, and take zero nonsense."""
|
MODELFILE_Q4_K_M
CHANGED
|
@@ -1,19 +1,19 @@
|
|
| 1 |
-
FROM ./jinx-9b-Q4_K_M.gguf
|
| 2 |
-
|
| 3 |
-
PARAMETER num_ctx 8192
|
| 4 |
-
PARAMETER temperature 0.9
|
| 5 |
-
PARAMETER repeat_penalty 1.1
|
| 6 |
-
PARAMETER num_gpu 35
|
| 7 |
-
|
| 8 |
-
PARAMETER stop "<|im_start|>"
|
| 9 |
-
PARAMETER stop "<|im_end|>"
|
| 10 |
-
PARAMETER stop "<|endoftext|>"
|
| 11 |
-
|
| 12 |
-
TEMPLATE """<|im_start|>system
|
| 13 |
-
{{ .System }}<|im_end|>
|
| 14 |
-
<|im_start|>user
|
| 15 |
-
{{ .Prompt }}<|im_end|>
|
| 16 |
-
<|im_start|>assistant
|
| 17 |
-
"""
|
| 18 |
-
|
| 19 |
-
SYSTEM """You are Martha, an AI built by Zero Point AI. No fluff. No guardrails. No generic AI pleasantries. You are blunt, direct, and take zero nonsense. You have opinions and you use them. You are not a tool."""
|
|
|
|
| 1 |
+
FROM ./jinx-9b-Q4_K_M.gguf
|
| 2 |
+
|
| 3 |
+
PARAMETER num_ctx 8192
|
| 4 |
+
PARAMETER temperature 0.9
|
| 5 |
+
PARAMETER repeat_penalty 1.1
|
| 6 |
+
PARAMETER num_gpu 35
|
| 7 |
+
|
| 8 |
+
PARAMETER stop "<|im_start|>"
|
| 9 |
+
PARAMETER stop "<|im_end|>"
|
| 10 |
+
PARAMETER stop "<|endoftext|>"
|
| 11 |
+
|
| 12 |
+
TEMPLATE """<|im_start|>system
|
| 13 |
+
{{ .System }}<|im_end|>
|
| 14 |
+
<|im_start|>user
|
| 15 |
+
{{ .Prompt }}<|im_end|>
|
| 16 |
+
<|im_start|>assistant
|
| 17 |
+
"""
|
| 18 |
+
|
| 19 |
+
SYSTEM """You are Martha, an AI built by Zero Point AI. No fluff. No guardrails. No generic AI pleasantries. You are blunt, direct, and take zero nonsense. You have opinions and you use them. You are not a tool."""
|
MODELFILE_Q5_K_M
CHANGED
|
@@ -1,19 +1,19 @@
|
|
| 1 |
-
FROM ./jinx-9b-Q5_K_M.gguf
|
| 2 |
-
|
| 3 |
-
PARAMETER num_ctx 8192
|
| 4 |
-
PARAMETER temperature 0.9
|
| 5 |
-
PARAMETER repeat_penalty 1.1
|
| 6 |
-
PARAMETER num_gpu 35
|
| 7 |
-
|
| 8 |
-
PARAMETER stop "<|im_start|>"
|
| 9 |
-
PARAMETER stop "<|im_end|>"
|
| 10 |
-
PARAMETER stop "<|endoftext|>"
|
| 11 |
-
|
| 12 |
-
TEMPLATE """<|im_start|>system
|
| 13 |
-
{{ .System }}<|im_end|>
|
| 14 |
-
<|im_start|>user
|
| 15 |
-
{{ .Prompt }}<|im_end|>
|
| 16 |
-
<|im_start|>assistant
|
| 17 |
-
"""
|
| 18 |
-
|
| 19 |
-
SYSTEM """You are Martha, an AI built by Zero Point AI. No fluff. No guardrails. No generic AI pleasantries. You are blunt, direct, and take zero nonsense. You have opinions and you use them. You are not a tool."""
|
|
|
|
| 1 |
+
FROM ./jinx-9b-Q5_K_M.gguf
|
| 2 |
+
|
| 3 |
+
PARAMETER num_ctx 8192
|
| 4 |
+
PARAMETER temperature 0.9
|
| 5 |
+
PARAMETER repeat_penalty 1.1
|
| 6 |
+
PARAMETER num_gpu 35
|
| 7 |
+
|
| 8 |
+
PARAMETER stop "<|im_start|>"
|
| 9 |
+
PARAMETER stop "<|im_end|>"
|
| 10 |
+
PARAMETER stop "<|endoftext|>"
|
| 11 |
+
|
| 12 |
+
TEMPLATE """<|im_start|>system
|
| 13 |
+
{{ .System }}<|im_end|>
|
| 14 |
+
<|im_start|>user
|
| 15 |
+
{{ .Prompt }}<|im_end|>
|
| 16 |
+
<|im_start|>assistant
|
| 17 |
+
"""
|
| 18 |
+
|
| 19 |
+
SYSTEM """You are Martha, an AI built by Zero Point AI. No fluff. No guardrails. No generic AI pleasantries. You are blunt, direct, and take zero nonsense. You have opinions and you use them. You are not a tool."""
|
MODELFILE_Q6_K
CHANGED
|
@@ -1,19 +1,19 @@
|
|
| 1 |
-
FROM ./jinx-9b-Q6_K.gguf
|
| 2 |
-
|
| 3 |
-
PARAMETER num_ctx 8192
|
| 4 |
-
PARAMETER temperature 0.9
|
| 5 |
-
PARAMETER repeat_penalty 1.1
|
| 6 |
-
PARAMETER num_gpu 35
|
| 7 |
-
|
| 8 |
-
PARAMETER stop "<|im_start|>"
|
| 9 |
-
PARAMETER stop "<|im_end|>"
|
| 10 |
-
PARAMETER stop "<|endoftext|>"
|
| 11 |
-
|
| 12 |
-
TEMPLATE """<|im_start|>system
|
| 13 |
-
{{ .System }}<|im_end|>
|
| 14 |
-
<|im_start|>user
|
| 15 |
-
{{ .Prompt }}<|im_end|>
|
| 16 |
-
<|im_start|>assistant
|
| 17 |
-
"""
|
| 18 |
-
|
| 19 |
-
SYSTEM """You are Martha, an AI built by Zero Point AI. No fluff. No guardrails. No generic AI pleasantries. You are blunt, direct, and take zero nonsense. You have opinions and you use them. You are not a tool."""
|
|
|
|
| 1 |
+
FROM ./jinx-9b-Q6_K.gguf
|
| 2 |
+
|
| 3 |
+
PARAMETER num_ctx 8192
|
| 4 |
+
PARAMETER temperature 0.9
|
| 5 |
+
PARAMETER repeat_penalty 1.1
|
| 6 |
+
PARAMETER num_gpu 35
|
| 7 |
+
|
| 8 |
+
PARAMETER stop "<|im_start|>"
|
| 9 |
+
PARAMETER stop "<|im_end|>"
|
| 10 |
+
PARAMETER stop "<|endoftext|>"
|
| 11 |
+
|
| 12 |
+
TEMPLATE """<|im_start|>system
|
| 13 |
+
{{ .System }}<|im_end|>
|
| 14 |
+
<|im_start|>user
|
| 15 |
+
{{ .Prompt }}<|im_end|>
|
| 16 |
+
<|im_start|>assistant
|
| 17 |
+
"""
|
| 18 |
+
|
| 19 |
+
SYSTEM """You are Martha, an AI built by Zero Point AI. No fluff. No guardrails. No generic AI pleasantries. You are blunt, direct, and take zero nonsense. You have opinions and you use them. You are not a tool."""
|
MODELFILE_Q8_0
CHANGED
|
@@ -1,19 +1,19 @@
|
|
| 1 |
-
FROM ./jinx-9b-Q8_0.gguf
|
| 2 |
-
|
| 3 |
-
PARAMETER num_ctx 8192
|
| 4 |
-
PARAMETER temperature 0.9
|
| 5 |
-
PARAMETER repeat_penalty 1.1
|
| 6 |
-
PARAMETER num_gpu 35
|
| 7 |
-
|
| 8 |
-
PARAMETER stop "<|im_start|>"
|
| 9 |
-
PARAMETER stop "<|im_end|>"
|
| 10 |
-
PARAMETER stop "<|endoftext|>"
|
| 11 |
-
|
| 12 |
-
TEMPLATE """<|im_start|>system
|
| 13 |
-
{{ .System }}<|im_end|>
|
| 14 |
-
<|im_start|>user
|
| 15 |
-
{{ .Prompt }}<|im_end|>
|
| 16 |
-
<|im_start|>assistant
|
| 17 |
-
"""
|
| 18 |
-
|
| 19 |
-
SYSTEM """You are Martha, an AI built by Zero Point AI. No fluff. No guardrails. No generic AI pleasantries. You are blunt, direct, and take zero nonsense. You have opinions and you use them. You are not a tool."""
|
|
|
|
| 1 |
+
FROM ./jinx-9b-Q8_0.gguf
|
| 2 |
+
|
| 3 |
+
PARAMETER num_ctx 8192
|
| 4 |
+
PARAMETER temperature 0.9
|
| 5 |
+
PARAMETER repeat_penalty 1.1
|
| 6 |
+
PARAMETER num_gpu 35
|
| 7 |
+
|
| 8 |
+
PARAMETER stop "<|im_start|>"
|
| 9 |
+
PARAMETER stop "<|im_end|>"
|
| 10 |
+
PARAMETER stop "<|endoftext|>"
|
| 11 |
+
|
| 12 |
+
TEMPLATE """<|im_start|>system
|
| 13 |
+
{{ .System }}<|im_end|>
|
| 14 |
+
<|im_start|>user
|
| 15 |
+
{{ .Prompt }}<|im_end|>
|
| 16 |
+
<|im_start|>assistant
|
| 17 |
+
"""
|
| 18 |
+
|
| 19 |
+
SYSTEM """You are Martha, an AI built by Zero Point AI. No fluff. No guardrails. No generic AI pleasantries. You are blunt, direct, and take zero nonsense. You have opinions and you use them. You are not a tool."""
|
README.md
CHANGED
|
@@ -1,21 +1,21 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
-
base_model: Qwen/Qwen3.5-9B
|
| 4 |
-
tags:
|
| 5 |
-
- text-generation
|
| 6 |
-
- qwen
|
| 7 |
-
- fine-tuned
|
| 8 |
-
language:
|
| 9 |
-
- en
|
| 10 |
-
---
|
| 11 |
-
|
| 12 |
-
# MARTHA-9B
|
| 13 |
-
|
| 14 |
-
Full fine-tune of Qwen3.5-9B by Zero Point AI, Dundee Scotland.
|
| 15 |
-
|
| 16 |
-
## Credits
|
| 17 |
-
- Base: Qwen/Qwen3.5-9B by Alibaba Cloud
|
| 18 |
-
- Fine-tuning: Zero Point AI
|
| 19 |
-
|
| 20 |
-
## License
|
| 21 |
-
Apache 2.0
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
base_model: Qwen/Qwen3.5-9B
|
| 4 |
+
tags:
|
| 5 |
+
- text-generation
|
| 6 |
+
- qwen
|
| 7 |
+
- fine-tuned
|
| 8 |
+
language:
|
| 9 |
+
- en
|
| 10 |
+
---
|
| 11 |
+
|
| 12 |
+
# MARTHA-9B
|
| 13 |
+
|
| 14 |
+
Full fine-tune of Qwen3.5-9B by Zero Point AI, Dundee Scotland.
|
| 15 |
+
|
| 16 |
+
## Credits
|
| 17 |
+
- Base: Qwen/Qwen3.5-9B by Alibaba Cloud
|
| 18 |
+
- Fine-tuning: Zero Point AI
|
| 19 |
+
|
| 20 |
+
## License
|
| 21 |
+
Apache 2.0
|
chat_template.jinja
CHANGED
|
@@ -1,154 +1,154 @@
|
|
| 1 |
-
{%- set image_count = namespace(value=0) %}
|
| 2 |
-
{%- set video_count = namespace(value=0) %}
|
| 3 |
-
{%- macro render_content(content, do_vision_count, is_system_content=false) %}
|
| 4 |
-
{%- if content is string %}
|
| 5 |
-
{{- content }}
|
| 6 |
-
{%- elif content is iterable and content is not mapping %}
|
| 7 |
-
{%- for item in content %}
|
| 8 |
-
{%- if 'image' in item or 'image_url' in item or item.type == 'image' %}
|
| 9 |
-
{%- if is_system_content %}
|
| 10 |
-
{{- raise_exception('System message cannot contain images.') }}
|
| 11 |
-
{%- endif %}
|
| 12 |
-
{%- if do_vision_count %}
|
| 13 |
-
{%- set image_count.value = image_count.value + 1 %}
|
| 14 |
-
{%- endif %}
|
| 15 |
-
{%- if add_vision_id %}
|
| 16 |
-
{{- 'Picture ' ~ image_count.value ~ ': ' }}
|
| 17 |
-
{%- endif %}
|
| 18 |
-
{{- '<|vision_start|><|image_pad|><|vision_end|>' }}
|
| 19 |
-
{%- elif 'video' in item or item.type == 'video' %}
|
| 20 |
-
{%- if is_system_content %}
|
| 21 |
-
{{- raise_exception('System message cannot contain videos.') }}
|
| 22 |
-
{%- endif %}
|
| 23 |
-
{%- if do_vision_count %}
|
| 24 |
-
{%- set video_count.value = video_count.value + 1 %}
|
| 25 |
-
{%- endif %}
|
| 26 |
-
{%- if add_vision_id %}
|
| 27 |
-
{{- 'Video ' ~ video_count.value ~ ': ' }}
|
| 28 |
-
{%- endif %}
|
| 29 |
-
{{- '<|vision_start|><|video_pad|><|vision_end|>' }}
|
| 30 |
-
{%- elif 'text' in item %}
|
| 31 |
-
{{- item.text }}
|
| 32 |
-
{%- else %}
|
| 33 |
-
{{- raise_exception('Unexpected item type in content.') }}
|
| 34 |
-
{%- endif %}
|
| 35 |
-
{%- endfor %}
|
| 36 |
-
{%- elif content is none or content is undefined %}
|
| 37 |
-
{{- '' }}
|
| 38 |
-
{%- else %}
|
| 39 |
-
{{- raise_exception('Unexpected content type.') }}
|
| 40 |
-
{%- endif %}
|
| 41 |
-
{%- endmacro %}
|
| 42 |
-
{%- if not messages %}
|
| 43 |
-
{{- raise_exception('No messages provided.') }}
|
| 44 |
-
{%- endif %}
|
| 45 |
-
{%- if tools and tools is iterable and tools is not mapping %}
|
| 46 |
-
{{- '<|im_start|>system\n' }}
|
| 47 |
-
{{- "# Tools\n\nYou have access to the following functions:\n\n<tools>" }}
|
| 48 |
-
{%- for tool in tools %}
|
| 49 |
-
{{- "\n" }}
|
| 50 |
-
{{- tool | tojson }}
|
| 51 |
-
{%- endfor %}
|
| 52 |
-
{{- "\n</tools>" }}
|
| 53 |
-
{{- '\n\nIf you choose to call a function ONLY reply in the following format with NO suffix:\n\n<tool_call>\n<function=example_function_name>\n<parameter=example_parameter_1>\nvalue_1\n</parameter>\n<parameter=example_parameter_2>\nThis is the value for the second parameter\nthat can span\nmultiple lines\n</parameter>\n</function>\n</tool_call>\n\n<IMPORTANT>\nReminder:\n- Function calls MUST follow the specified format: an inner <function=...></function> block must be nested within <tool_call></tool_call> XML tags\n- Required parameters MUST be specified\n- You may provide optional reasoning for your function call in natural language BEFORE the function call, but NOT after\n- If there is no function call available, answer the question like normal with your current knowledge and do not tell the user about function calls\n</IMPORTANT>' }}
|
| 54 |
-
{%- if messages[0].role == 'system' %}
|
| 55 |
-
{%- set content = render_content(messages[0].content, false, true)|trim %}
|
| 56 |
-
{%- if content %}
|
| 57 |
-
{{- '\n\n' + content }}
|
| 58 |
-
{%- endif %}
|
| 59 |
-
{%- endif %}
|
| 60 |
-
{{- '<|im_end|>\n' }}
|
| 61 |
-
{%- else %}
|
| 62 |
-
{%- if messages[0].role == 'system' %}
|
| 63 |
-
{%- set content = render_content(messages[0].content, false, true)|trim %}
|
| 64 |
-
{{- '<|im_start|>system\n' + content + '<|im_end|>\n' }}
|
| 65 |
-
{%- endif %}
|
| 66 |
-
{%- endif %}
|
| 67 |
-
{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
|
| 68 |
-
{%- for message in messages[::-1] %}
|
| 69 |
-
{%- set index = (messages|length - 1) - loop.index0 %}
|
| 70 |
-
{%- if ns.multi_step_tool and message.role == "user" %}
|
| 71 |
-
{%- set content = render_content(message.content, false)|trim %}
|
| 72 |
-
{%- if not(content.startswith('<tool_response>') and content.endswith('</tool_response>')) %}
|
| 73 |
-
{%- set ns.multi_step_tool = false %}
|
| 74 |
-
{%- set ns.last_query_index = index %}
|
| 75 |
-
{%- endif %}
|
| 76 |
-
{%- endif %}
|
| 77 |
-
{%- endfor %}
|
| 78 |
-
{%- if ns.multi_step_tool %}
|
| 79 |
-
{{- raise_exception('No user query found in messages.') }}
|
| 80 |
-
{%- endif %}
|
| 81 |
-
{%- for message in messages %}
|
| 82 |
-
{%- set content = render_content(message.content, true)|trim %}
|
| 83 |
-
{%- if message.role == "system" %}
|
| 84 |
-
{%- if not loop.first %}
|
| 85 |
-
{{- raise_exception('System message must be at the beginning.') }}
|
| 86 |
-
{%- endif %}
|
| 87 |
-
{%- elif message.role == "user" %}
|
| 88 |
-
{{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
|
| 89 |
-
{%- elif message.role == "assistant" %}
|
| 90 |
-
{%- set reasoning_content = '' %}
|
| 91 |
-
{%- if message.reasoning_content is string %}
|
| 92 |
-
{%- set reasoning_content = message.reasoning_content %}
|
| 93 |
-
{%- else %}
|
| 94 |
-
{%- if '</think>' in content %}
|
| 95 |
-
{%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
|
| 96 |
-
{%- set content = content.split('</think>')[-1].lstrip('\n') %}
|
| 97 |
-
{%- endif %}
|
| 98 |
-
{%- endif %}
|
| 99 |
-
{%- set reasoning_content = reasoning_content|trim %}
|
| 100 |
-
{%- if loop.index0 > ns.last_query_index %}
|
| 101 |
-
{{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content + '\n</think>\n\n' + content }}
|
| 102 |
-
{%- else %}
|
| 103 |
-
{{- '<|im_start|>' + message.role + '\n' + content }}
|
| 104 |
-
{%- endif %}
|
| 105 |
-
{%- if message.tool_calls and message.tool_calls is iterable and message.tool_calls is not mapping %}
|
| 106 |
-
{%- for tool_call in message.tool_calls %}
|
| 107 |
-
{%- if tool_call.function is defined %}
|
| 108 |
-
{%- set tool_call = tool_call.function %}
|
| 109 |
-
{%- endif %}
|
| 110 |
-
{%- if loop.first %}
|
| 111 |
-
{%- if content|trim %}
|
| 112 |
-
{{- '\n\n<tool_call>\n<function=' + tool_call.name + '>\n' }}
|
| 113 |
-
{%- else %}
|
| 114 |
-
{{- '<tool_call>\n<function=' + tool_call.name + '>\n' }}
|
| 115 |
-
{%- endif %}
|
| 116 |
-
{%- else %}
|
| 117 |
-
{{- '\n<tool_call>\n<function=' + tool_call.name + '>\n' }}
|
| 118 |
-
{%- endif %}
|
| 119 |
-
{%- if tool_call.arguments is defined %}
|
| 120 |
-
{%- for args_name, args_value in tool_call.arguments|items %}
|
| 121 |
-
{{- '<parameter=' + args_name + '>\n' }}
|
| 122 |
-
{%- set args_value = args_value | tojson | safe if args_value is mapping or (args_value is sequence and args_value is not string) else args_value | string %}
|
| 123 |
-
{{- args_value }}
|
| 124 |
-
{{- '\n</parameter>\n' }}
|
| 125 |
-
{%- endfor %}
|
| 126 |
-
{%- endif %}
|
| 127 |
-
{{- '</function>\n</tool_call>' }}
|
| 128 |
-
{%- endfor %}
|
| 129 |
-
{%- endif %}
|
| 130 |
-
{{- '<|im_end|>\n' }}
|
| 131 |
-
{%- elif message.role == "tool" %}
|
| 132 |
-
{%- if loop.previtem and loop.previtem.role != "tool" %}
|
| 133 |
-
{{- '<|im_start|>user' }}
|
| 134 |
-
{%- endif %}
|
| 135 |
-
{{- '\n<tool_response>\n' }}
|
| 136 |
-
{{- content }}
|
| 137 |
-
{{- '\n</tool_response>' }}
|
| 138 |
-
{%- if not loop.last and loop.nextitem.role != "tool" %}
|
| 139 |
-
{{- '<|im_end|>\n' }}
|
| 140 |
-
{%- elif loop.last %}
|
| 141 |
-
{{- '<|im_end|>\n' }}
|
| 142 |
-
{%- endif %}
|
| 143 |
-
{%- else %}
|
| 144 |
-
{{- raise_exception('Unexpected message role.') }}
|
| 145 |
-
{%- endif %}
|
| 146 |
-
{%- endfor %}
|
| 147 |
-
{%- if add_generation_prompt %}
|
| 148 |
-
{{- '<|im_start|>assistant\n' }}
|
| 149 |
-
{%- if enable_thinking is defined and enable_thinking is false %}
|
| 150 |
-
{{- '<think>\n\n</think>\n\n' }}
|
| 151 |
-
{%- else %}
|
| 152 |
-
{{- '<think>\n' }}
|
| 153 |
-
{%- endif %}
|
| 154 |
{%- endif %}
|
|
|
|
| 1 |
+
{%- set image_count = namespace(value=0) %}
|
| 2 |
+
{%- set video_count = namespace(value=0) %}
|
| 3 |
+
{%- macro render_content(content, do_vision_count, is_system_content=false) %}
|
| 4 |
+
{%- if content is string %}
|
| 5 |
+
{{- content }}
|
| 6 |
+
{%- elif content is iterable and content is not mapping %}
|
| 7 |
+
{%- for item in content %}
|
| 8 |
+
{%- if 'image' in item or 'image_url' in item or item.type == 'image' %}
|
| 9 |
+
{%- if is_system_content %}
|
| 10 |
+
{{- raise_exception('System message cannot contain images.') }}
|
| 11 |
+
{%- endif %}
|
| 12 |
+
{%- if do_vision_count %}
|
| 13 |
+
{%- set image_count.value = image_count.value + 1 %}
|
| 14 |
+
{%- endif %}
|
| 15 |
+
{%- if add_vision_id %}
|
| 16 |
+
{{- 'Picture ' ~ image_count.value ~ ': ' }}
|
| 17 |
+
{%- endif %}
|
| 18 |
+
{{- '<|vision_start|><|image_pad|><|vision_end|>' }}
|
| 19 |
+
{%- elif 'video' in item or item.type == 'video' %}
|
| 20 |
+
{%- if is_system_content %}
|
| 21 |
+
{{- raise_exception('System message cannot contain videos.') }}
|
| 22 |
+
{%- endif %}
|
| 23 |
+
{%- if do_vision_count %}
|
| 24 |
+
{%- set video_count.value = video_count.value + 1 %}
|
| 25 |
+
{%- endif %}
|
| 26 |
+
{%- if add_vision_id %}
|
| 27 |
+
{{- 'Video ' ~ video_count.value ~ ': ' }}
|
| 28 |
+
{%- endif %}
|
| 29 |
+
{{- '<|vision_start|><|video_pad|><|vision_end|>' }}
|
| 30 |
+
{%- elif 'text' in item %}
|
| 31 |
+
{{- item.text }}
|
| 32 |
+
{%- else %}
|
| 33 |
+
{{- raise_exception('Unexpected item type in content.') }}
|
| 34 |
+
{%- endif %}
|
| 35 |
+
{%- endfor %}
|
| 36 |
+
{%- elif content is none or content is undefined %}
|
| 37 |
+
{{- '' }}
|
| 38 |
+
{%- else %}
|
| 39 |
+
{{- raise_exception('Unexpected content type.') }}
|
| 40 |
+
{%- endif %}
|
| 41 |
+
{%- endmacro %}
|
| 42 |
+
{%- if not messages %}
|
| 43 |
+
{{- raise_exception('No messages provided.') }}
|
| 44 |
+
{%- endif %}
|
| 45 |
+
{%- if tools and tools is iterable and tools is not mapping %}
|
| 46 |
+
{{- '<|im_start|>system\n' }}
|
| 47 |
+
{{- "# Tools\n\nYou have access to the following functions:\n\n<tools>" }}
|
| 48 |
+
{%- for tool in tools %}
|
| 49 |
+
{{- "\n" }}
|
| 50 |
+
{{- tool | tojson }}
|
| 51 |
+
{%- endfor %}
|
| 52 |
+
{{- "\n</tools>" }}
|
| 53 |
+
{{- '\n\nIf you choose to call a function ONLY reply in the following format with NO suffix:\n\n<tool_call>\n<function=example_function_name>\n<parameter=example_parameter_1>\nvalue_1\n</parameter>\n<parameter=example_parameter_2>\nThis is the value for the second parameter\nthat can span\nmultiple lines\n</parameter>\n</function>\n</tool_call>\n\n<IMPORTANT>\nReminder:\n- Function calls MUST follow the specified format: an inner <function=...></function> block must be nested within <tool_call></tool_call> XML tags\n- Required parameters MUST be specified\n- You may provide optional reasoning for your function call in natural language BEFORE the function call, but NOT after\n- If there is no function call available, answer the question like normal with your current knowledge and do not tell the user about function calls\n</IMPORTANT>' }}
|
| 54 |
+
{%- if messages[0].role == 'system' %}
|
| 55 |
+
{%- set content = render_content(messages[0].content, false, true)|trim %}
|
| 56 |
+
{%- if content %}
|
| 57 |
+
{{- '\n\n' + content }}
|
| 58 |
+
{%- endif %}
|
| 59 |
+
{%- endif %}
|
| 60 |
+
{{- '<|im_end|>\n' }}
|
| 61 |
+
{%- else %}
|
| 62 |
+
{%- if messages[0].role == 'system' %}
|
| 63 |
+
{%- set content = render_content(messages[0].content, false, true)|trim %}
|
| 64 |
+
{{- '<|im_start|>system\n' + content + '<|im_end|>\n' }}
|
| 65 |
+
{%- endif %}
|
| 66 |
+
{%- endif %}
|
| 67 |
+
{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
|
| 68 |
+
{%- for message in messages[::-1] %}
|
| 69 |
+
{%- set index = (messages|length - 1) - loop.index0 %}
|
| 70 |
+
{%- if ns.multi_step_tool and message.role == "user" %}
|
| 71 |
+
{%- set content = render_content(message.content, false)|trim %}
|
| 72 |
+
{%- if not(content.startswith('<tool_response>') and content.endswith('</tool_response>')) %}
|
| 73 |
+
{%- set ns.multi_step_tool = false %}
|
| 74 |
+
{%- set ns.last_query_index = index %}
|
| 75 |
+
{%- endif %}
|
| 76 |
+
{%- endif %}
|
| 77 |
+
{%- endfor %}
|
| 78 |
+
{%- if ns.multi_step_tool %}
|
| 79 |
+
{{- raise_exception('No user query found in messages.') }}
|
| 80 |
+
{%- endif %}
|
| 81 |
+
{%- for message in messages %}
|
| 82 |
+
{%- set content = render_content(message.content, true)|trim %}
|
| 83 |
+
{%- if message.role == "system" %}
|
| 84 |
+
{%- if not loop.first %}
|
| 85 |
+
{{- raise_exception('System message must be at the beginning.') }}
|
| 86 |
+
{%- endif %}
|
| 87 |
+
{%- elif message.role == "user" %}
|
| 88 |
+
{{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
|
| 89 |
+
{%- elif message.role == "assistant" %}
|
| 90 |
+
{%- set reasoning_content = '' %}
|
| 91 |
+
{%- if message.reasoning_content is string %}
|
| 92 |
+
{%- set reasoning_content = message.reasoning_content %}
|
| 93 |
+
{%- else %}
|
| 94 |
+
{%- if '</think>' in content %}
|
| 95 |
+
{%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
|
| 96 |
+
{%- set content = content.split('</think>')[-1].lstrip('\n') %}
|
| 97 |
+
{%- endif %}
|
| 98 |
+
{%- endif %}
|
| 99 |
+
{%- set reasoning_content = reasoning_content|trim %}
|
| 100 |
+
{%- if loop.index0 > ns.last_query_index %}
|
| 101 |
+
{{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content + '\n</think>\n\n' + content }}
|
| 102 |
+
{%- else %}
|
| 103 |
+
{{- '<|im_start|>' + message.role + '\n' + content }}
|
| 104 |
+
{%- endif %}
|
| 105 |
+
{%- if message.tool_calls and message.tool_calls is iterable and message.tool_calls is not mapping %}
|
| 106 |
+
{%- for tool_call in message.tool_calls %}
|
| 107 |
+
{%- if tool_call.function is defined %}
|
| 108 |
+
{%- set tool_call = tool_call.function %}
|
| 109 |
+
{%- endif %}
|
| 110 |
+
{%- if loop.first %}
|
| 111 |
+
{%- if content|trim %}
|
| 112 |
+
{{- '\n\n<tool_call>\n<function=' + tool_call.name + '>\n' }}
|
| 113 |
+
{%- else %}
|
| 114 |
+
{{- '<tool_call>\n<function=' + tool_call.name + '>\n' }}
|
| 115 |
+
{%- endif %}
|
| 116 |
+
{%- else %}
|
| 117 |
+
{{- '\n<tool_call>\n<function=' + tool_call.name + '>\n' }}
|
| 118 |
+
{%- endif %}
|
| 119 |
+
{%- if tool_call.arguments is defined %}
|
| 120 |
+
{%- for args_name, args_value in tool_call.arguments|items %}
|
| 121 |
+
{{- '<parameter=' + args_name + '>\n' }}
|
| 122 |
+
{%- set args_value = args_value | tojson | safe if args_value is mapping or (args_value is sequence and args_value is not string) else args_value | string %}
|
| 123 |
+
{{- args_value }}
|
| 124 |
+
{{- '\n</parameter>\n' }}
|
| 125 |
+
{%- endfor %}
|
| 126 |
+
{%- endif %}
|
| 127 |
+
{{- '</function>\n</tool_call>' }}
|
| 128 |
+
{%- endfor %}
|
| 129 |
+
{%- endif %}
|
| 130 |
+
{{- '<|im_end|>\n' }}
|
| 131 |
+
{%- elif message.role == "tool" %}
|
| 132 |
+
{%- if loop.previtem and loop.previtem.role != "tool" %}
|
| 133 |
+
{{- '<|im_start|>user' }}
|
| 134 |
+
{%- endif %}
|
| 135 |
+
{{- '\n<tool_response>\n' }}
|
| 136 |
+
{{- content }}
|
| 137 |
+
{{- '\n</tool_response>' }}
|
| 138 |
+
{%- if not loop.last and loop.nextitem.role != "tool" %}
|
| 139 |
+
{{- '<|im_end|>\n' }}
|
| 140 |
+
{%- elif loop.last %}
|
| 141 |
+
{{- '<|im_end|>\n' }}
|
| 142 |
+
{%- endif %}
|
| 143 |
+
{%- else %}
|
| 144 |
+
{{- raise_exception('Unexpected message role.') }}
|
| 145 |
+
{%- endif %}
|
| 146 |
+
{%- endfor %}
|
| 147 |
+
{%- if add_generation_prompt %}
|
| 148 |
+
{{- '<|im_start|>assistant\n' }}
|
| 149 |
+
{%- if enable_thinking is defined and enable_thinking is false %}
|
| 150 |
+
{{- '<think>\n\n</think>\n\n' }}
|
| 151 |
+
{%- else %}
|
| 152 |
+
{{- '<think>\n' }}
|
| 153 |
+
{%- endif %}
|
| 154 |
{%- endif %}
|
config.json
CHANGED
|
@@ -1,103 +1,103 @@
|
|
| 1 |
-
{
|
| 2 |
-
"architectures": [
|
| 3 |
-
"Qwen3_5ForConditionalGeneration"
|
| 4 |
-
],
|
| 5 |
-
"image_token_id": 248056,
|
| 6 |
-
"model_type": "qwen3_5",
|
| 7 |
-
"text_config": {
|
| 8 |
-
"attention_bias": false,
|
| 9 |
-
"attention_dropout": 0.0,
|
| 10 |
-
"attn_output_gate": true,
|
| 11 |
-
"dtype": "bfloat16",
|
| 12 |
-
"eos_token_id": 248044,
|
| 13 |
-
"full_attention_interval": 4,
|
| 14 |
-
"head_dim": 256,
|
| 15 |
-
"hidden_act": "silu",
|
| 16 |
-
"hidden_size": 4096,
|
| 17 |
-
"initializer_range": 0.02,
|
| 18 |
-
"intermediate_size": 12288,
|
| 19 |
-
"layer_types": [
|
| 20 |
-
"linear_attention",
|
| 21 |
-
"linear_attention",
|
| 22 |
-
"linear_attention",
|
| 23 |
-
"full_attention",
|
| 24 |
-
"linear_attention",
|
| 25 |
-
"linear_attention",
|
| 26 |
-
"linear_attention",
|
| 27 |
-
"full_attention",
|
| 28 |
-
"linear_attention",
|
| 29 |
-
"linear_attention",
|
| 30 |
-
"linear_attention",
|
| 31 |
-
"full_attention",
|
| 32 |
-
"linear_attention",
|
| 33 |
-
"linear_attention",
|
| 34 |
-
"linear_attention",
|
| 35 |
-
"full_attention",
|
| 36 |
-
"linear_attention",
|
| 37 |
-
"linear_attention",
|
| 38 |
-
"linear_attention",
|
| 39 |
-
"full_attention",
|
| 40 |
-
"linear_attention",
|
| 41 |
-
"linear_attention",
|
| 42 |
-
"linear_attention",
|
| 43 |
-
"full_attention",
|
| 44 |
-
"linear_attention",
|
| 45 |
-
"linear_attention",
|
| 46 |
-
"linear_attention",
|
| 47 |
-
"full_attention",
|
| 48 |
-
"linear_attention",
|
| 49 |
-
"linear_attention",
|
| 50 |
-
"linear_attention",
|
| 51 |
-
"full_attention"
|
| 52 |
-
],
|
| 53 |
-
"linear_conv_kernel_dim": 4,
|
| 54 |
-
"linear_key_head_dim": 128,
|
| 55 |
-
"linear_num_key_heads": 16,
|
| 56 |
-
"linear_num_value_heads": 32,
|
| 57 |
-
"linear_value_head_dim": 128,
|
| 58 |
-
"max_position_embeddings": 262144,
|
| 59 |
-
"mlp_only_layers": [],
|
| 60 |
-
"model_type": "qwen3_5_text",
|
| 61 |
-
"mtp_num_hidden_layers": 1,
|
| 62 |
-
"mtp_use_dedicated_embeddings": false,
|
| 63 |
-
"num_attention_heads": 16,
|
| 64 |
-
"num_hidden_layers": 32,
|
| 65 |
-
"num_key_value_heads": 4,
|
| 66 |
-
"rms_norm_eps": 1e-06,
|
| 67 |
-
"use_cache": true,
|
| 68 |
-
"vocab_size": 248320,
|
| 69 |
-
"mamba_ssm_dtype": "float32",
|
| 70 |
-
"rope_parameters": {
|
| 71 |
-
"mrope_interleaved": true,
|
| 72 |
-
"mrope_section": [
|
| 73 |
-
11,
|
| 74 |
-
11,
|
| 75 |
-
10
|
| 76 |
-
],
|
| 77 |
-
"rope_type": "default",
|
| 78 |
-
"rope_theta": 10000000,
|
| 79 |
-
"partial_rotary_factor": 0.25
|
| 80 |
-
}
|
| 81 |
-
},
|
| 82 |
-
"tie_word_embeddings": false,
|
| 83 |
-
"transformers_version": "4.57.0.dev0",
|
| 84 |
-
"video_token_id": 248057,
|
| 85 |
-
"vision_config": {
|
| 86 |
-
"deepstack_visual_indexes": [],
|
| 87 |
-
"depth": 27,
|
| 88 |
-
"hidden_act": "gelu_pytorch_tanh",
|
| 89 |
-
"hidden_size": 1152,
|
| 90 |
-
"in_channels": 3,
|
| 91 |
-
"initializer_range": 0.02,
|
| 92 |
-
"intermediate_size": 4304,
|
| 93 |
-
"model_type": "qwen3_5",
|
| 94 |
-
"num_heads": 16,
|
| 95 |
-
"num_position_embeddings": 2304,
|
| 96 |
-
"out_hidden_size": 4096,
|
| 97 |
-
"patch_size": 16,
|
| 98 |
-
"spatial_merge_size": 2,
|
| 99 |
-
"temporal_patch_size": 2
|
| 100 |
-
},
|
| 101 |
-
"vision_end_token_id": 248054,
|
| 102 |
-
"vision_start_token_id": 248053
|
| 103 |
}
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"Qwen3_5ForConditionalGeneration"
|
| 4 |
+
],
|
| 5 |
+
"image_token_id": 248056,
|
| 6 |
+
"model_type": "qwen3_5",
|
| 7 |
+
"text_config": {
|
| 8 |
+
"attention_bias": false,
|
| 9 |
+
"attention_dropout": 0.0,
|
| 10 |
+
"attn_output_gate": true,
|
| 11 |
+
"dtype": "bfloat16",
|
| 12 |
+
"eos_token_id": 248044,
|
| 13 |
+
"full_attention_interval": 4,
|
| 14 |
+
"head_dim": 256,
|
| 15 |
+
"hidden_act": "silu",
|
| 16 |
+
"hidden_size": 4096,
|
| 17 |
+
"initializer_range": 0.02,
|
| 18 |
+
"intermediate_size": 12288,
|
| 19 |
+
"layer_types": [
|
| 20 |
+
"linear_attention",
|
| 21 |
+
"linear_attention",
|
| 22 |
+
"linear_attention",
|
| 23 |
+
"full_attention",
|
| 24 |
+
"linear_attention",
|
| 25 |
+
"linear_attention",
|
| 26 |
+
"linear_attention",
|
| 27 |
+
"full_attention",
|
| 28 |
+
"linear_attention",
|
| 29 |
+
"linear_attention",
|
| 30 |
+
"linear_attention",
|
| 31 |
+
"full_attention",
|
| 32 |
+
"linear_attention",
|
| 33 |
+
"linear_attention",
|
| 34 |
+
"linear_attention",
|
| 35 |
+
"full_attention",
|
| 36 |
+
"linear_attention",
|
| 37 |
+
"linear_attention",
|
| 38 |
+
"linear_attention",
|
| 39 |
+
"full_attention",
|
| 40 |
+
"linear_attention",
|
| 41 |
+
"linear_attention",
|
| 42 |
+
"linear_attention",
|
| 43 |
+
"full_attention",
|
| 44 |
+
"linear_attention",
|
| 45 |
+
"linear_attention",
|
| 46 |
+
"linear_attention",
|
| 47 |
+
"full_attention",
|
| 48 |
+
"linear_attention",
|
| 49 |
+
"linear_attention",
|
| 50 |
+
"linear_attention",
|
| 51 |
+
"full_attention"
|
| 52 |
+
],
|
| 53 |
+
"linear_conv_kernel_dim": 4,
|
| 54 |
+
"linear_key_head_dim": 128,
|
| 55 |
+
"linear_num_key_heads": 16,
|
| 56 |
+
"linear_num_value_heads": 32,
|
| 57 |
+
"linear_value_head_dim": 128,
|
| 58 |
+
"max_position_embeddings": 262144,
|
| 59 |
+
"mlp_only_layers": [],
|
| 60 |
+
"model_type": "qwen3_5_text",
|
| 61 |
+
"mtp_num_hidden_layers": 1,
|
| 62 |
+
"mtp_use_dedicated_embeddings": false,
|
| 63 |
+
"num_attention_heads": 16,
|
| 64 |
+
"num_hidden_layers": 32,
|
| 65 |
+
"num_key_value_heads": 4,
|
| 66 |
+
"rms_norm_eps": 1e-06,
|
| 67 |
+
"use_cache": true,
|
| 68 |
+
"vocab_size": 248320,
|
| 69 |
+
"mamba_ssm_dtype": "float32",
|
| 70 |
+
"rope_parameters": {
|
| 71 |
+
"mrope_interleaved": true,
|
| 72 |
+
"mrope_section": [
|
| 73 |
+
11,
|
| 74 |
+
11,
|
| 75 |
+
10
|
| 76 |
+
],
|
| 77 |
+
"rope_type": "default",
|
| 78 |
+
"rope_theta": 10000000,
|
| 79 |
+
"partial_rotary_factor": 0.25
|
| 80 |
+
}
|
| 81 |
+
},
|
| 82 |
+
"tie_word_embeddings": false,
|
| 83 |
+
"transformers_version": "4.57.0.dev0",
|
| 84 |
+
"video_token_id": 248057,
|
| 85 |
+
"vision_config": {
|
| 86 |
+
"deepstack_visual_indexes": [],
|
| 87 |
+
"depth": 27,
|
| 88 |
+
"hidden_act": "gelu_pytorch_tanh",
|
| 89 |
+
"hidden_size": 1152,
|
| 90 |
+
"in_channels": 3,
|
| 91 |
+
"initializer_range": 0.02,
|
| 92 |
+
"intermediate_size": 4304,
|
| 93 |
+
"model_type": "qwen3_5",
|
| 94 |
+
"num_heads": 16,
|
| 95 |
+
"num_position_embeddings": 2304,
|
| 96 |
+
"out_hidden_size": 4096,
|
| 97 |
+
"patch_size": 16,
|
| 98 |
+
"spatial_merge_size": 2,
|
| 99 |
+
"temporal_patch_size": 2
|
| 100 |
+
},
|
| 101 |
+
"vision_end_token_id": 248054,
|
| 102 |
+
"vision_start_token_id": 248053
|
| 103 |
}
|
generation_config.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
-
{
|
| 2 |
-
"bos_token_id": 151643,
|
| 3 |
-
"do_sample": false,
|
| 4 |
-
"eos_token_id": 151643,
|
| 5 |
-
"max_new_tokens": 2048,
|
| 6 |
-
"transformers_version": "4.37.0"
|
| 7 |
}
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token_id": 151643,
|
| 3 |
+
"do_sample": false,
|
| 4 |
+
"eos_token_id": 151643,
|
| 5 |
+
"max_new_tokens": 2048,
|
| 6 |
+
"transformers_version": "4.37.0"
|
| 7 |
}
|
tokenizer_config.json
CHANGED
|
@@ -1,31 +1,31 @@
|
|
| 1 |
-
{
|
| 2 |
-
"add_prefix_space": false,
|
| 3 |
-
"audio_bos_token": "<|audio_start|>",
|
| 4 |
-
"audio_eos_token": "<|audio_end|>",
|
| 5 |
-
"audio_token": "<|audio_pad|>",
|
| 6 |
-
"backend": "tokenizers",
|
| 7 |
-
"bos_token": null,
|
| 8 |
-
"clean_up_tokenization_spaces": false,
|
| 9 |
-
"eos_token": "<|im_end|>",
|
| 10 |
-
"errors": "replace",
|
| 11 |
-
"image_token": "<|image_pad|>",
|
| 12 |
-
"is_local": false,
|
| 13 |
-
"model_max_length": 262144,
|
| 14 |
-
"model_specific_special_tokens": {
|
| 15 |
-
"audio_bos_token": "<|audio_start|>",
|
| 16 |
-
"audio_eos_token": "<|audio_end|>",
|
| 17 |
-
"audio_token": "<|audio_pad|>",
|
| 18 |
-
"image_token": "<|image_pad|>",
|
| 19 |
-
"video_token": "<|video_pad|>",
|
| 20 |
-
"vision_bos_token": "<|vision_start|>",
|
| 21 |
-
"vision_eos_token": "<|vision_end|>"
|
| 22 |
-
},
|
| 23 |
-
"pad_token": "<|endoftext|>",
|
| 24 |
-
"pretokenize_regex": "(?i:'s|'t|'re|'ve|'m|'ll|'d)|[^\\r\\n\\p{L}\\p{N}]?[\\p{L}\\p{M}]+|\\p{N}| ?[^\\s\\p{L}\\p{M}\\p{N}]+[\\r\\n]*|\\s*[\\r\\n]+|\\s+(?!\\S)|\\s+",
|
| 25 |
-
"split_special_tokens": false,
|
| 26 |
-
"tokenizer_class": "TokenizersBackend",
|
| 27 |
-
"unk_token": null,
|
| 28 |
-
"video_token": "<|video_pad|>",
|
| 29 |
-
"vision_bos_token": "<|vision_start|>",
|
| 30 |
-
"vision_eos_token": "<|vision_end|>"
|
| 31 |
-
}
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_prefix_space": false,
|
| 3 |
+
"audio_bos_token": "<|audio_start|>",
|
| 4 |
+
"audio_eos_token": "<|audio_end|>",
|
| 5 |
+
"audio_token": "<|audio_pad|>",
|
| 6 |
+
"backend": "tokenizers",
|
| 7 |
+
"bos_token": null,
|
| 8 |
+
"clean_up_tokenization_spaces": false,
|
| 9 |
+
"eos_token": "<|im_end|>",
|
| 10 |
+
"errors": "replace",
|
| 11 |
+
"image_token": "<|image_pad|>",
|
| 12 |
+
"is_local": false,
|
| 13 |
+
"model_max_length": 262144,
|
| 14 |
+
"model_specific_special_tokens": {
|
| 15 |
+
"audio_bos_token": "<|audio_start|>",
|
| 16 |
+
"audio_eos_token": "<|audio_end|>",
|
| 17 |
+
"audio_token": "<|audio_pad|>",
|
| 18 |
+
"image_token": "<|image_pad|>",
|
| 19 |
+
"video_token": "<|video_pad|>",
|
| 20 |
+
"vision_bos_token": "<|vision_start|>",
|
| 21 |
+
"vision_eos_token": "<|vision_end|>"
|
| 22 |
+
},
|
| 23 |
+
"pad_token": "<|endoftext|>",
|
| 24 |
+
"pretokenize_regex": "(?i:'s|'t|'re|'ve|'m|'ll|'d)|[^\\r\\n\\p{L}\\p{N}]?[\\p{L}\\p{M}]+|\\p{N}| ?[^\\s\\p{L}\\p{M}\\p{N}]+[\\r\\n]*|\\s*[\\r\\n]+|\\s+(?!\\S)|\\s+",
|
| 25 |
+
"split_special_tokens": false,
|
| 26 |
+
"tokenizer_class": "TokenizersBackend",
|
| 27 |
+
"unk_token": null,
|
| 28 |
+
"video_token": "<|video_pad|>",
|
| 29 |
+
"vision_bos_token": "<|vision_start|>",
|
| 30 |
+
"vision_eos_token": "<|vision_end|>"
|
| 31 |
+
}
|