aphoticshaman commited on
Commit
664ed48
·
verified ·
1 Parent(s): d6fbb14

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,69 +1,127 @@
1
  ---
2
  license: apache-2.0
3
- base_model: Qwen/Qwen2.5-72B-Instruct
4
- tags:
5
- - math
6
- - reasoning
7
- - qwen2
8
- - merged
9
- - aimo3
10
  library_name: transformers
 
 
 
 
 
 
 
 
 
 
 
11
  pipeline_tag: text-generation
12
  model-index:
13
- - name: elle-72b-ultimate
14
  results: []
15
  ---
16
 
17
  # Elle-72B-Ultimate
18
 
19
- ## Model Description
20
 
21
- Elle-72B-Ultimate is a fine-tuned version of [Qwen/Qwen2.5-72B-Instruct](https://huggingface.co/Qwen/Qwen2.5-72B-Instruct) optimized for mathematical reasoning and problem-solving, specifically designed for the AI Mathematical Olympiad Progress Prize 3 (AIMO3) competition.
22
-
23
- This is a **merged full model** (LoRA adapter merged into base weights).
 
 
24
 
25
  ## Model Details
26
 
27
- - **Base Model**: Qwen/Qwen2.5-72B-Instruct
28
- - **Parameters**: 72B
29
- - **Precision**: BF16
30
- - **Format**: Safetensors (31 shards)
31
- - **Training Method**: LoRA (r=64, α=128)
 
 
 
32
 
33
  ## Training Data
34
 
35
- Fine-tuned on mathematical reasoning datasets including:
36
- - NuminaMath-CoT
37
- - Custom mathematical reasoning examples
 
 
 
 
 
 
38
 
39
  ## Intended Use
40
 
41
- - Mathematical problem solving
42
- - Olympiad-style competition problems
43
- - Code generation for computational solutions
44
- - Chain-of-thought reasoning
 
 
 
45
 
46
  ## Limitations
47
 
48
- - **Size**: ~144GB in BF16 - requires significant VRAM
49
- - **Quantization Recommended**: For inference on consumer hardware, use AWQ or GPTQ quantized versions
 
 
 
 
 
 
 
 
50
 
51
- ## Usage
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
52
 
53
  ```python
54
  from transformers import AutoModelForCausalLM, AutoTokenizer
55
 
56
  model = AutoModelForCausalLM.from_pretrained(
57
- "aphoticshaman/elle-72b-ultimate",
58
- torch_dtype="auto",
59
  device_map="auto",
60
- trust_remote_code=True
61
  )
62
- tokenizer = AutoTokenizer.from_pretrained("aphoticshaman/elle-72b-ultimate")
63
 
64
  messages = [
65
- {"role": "system", "content": "You are an expert mathematical problem solver."},
66
- {"role": "user", "content": "Find all positive integers n such that n^2 + 1 divides n^3 + 1."}
67
  ]
68
 
69
  text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
@@ -72,14 +130,50 @@ outputs = model.generate(**inputs, max_new_tokens=2048)
72
  print(tokenizer.decode(outputs[0], skip_special_tokens=True))
73
  ```
74
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
75
  ## Citation
76
 
77
  ```bibtex
78
  @misc{elle-72b-ultimate,
79
- author = {aphoticshaman},
80
- title = {Elle-72B-Ultimate: Mathematical Reasoning Model},
81
  year = {2024},
82
  publisher = {HuggingFace},
83
- url = {https://huggingface.co/aphoticshaman/elle-72b-ultimate}
84
  }
85
  ```
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ language:
4
+ - en
 
 
 
 
 
5
  library_name: transformers
6
+ tags:
7
+ - geopolitical-analysis
8
+ - risk-assessment
9
+ - intelligence
10
+ - fine-tuned
11
+ - lora
12
+ - qwen2.5
13
+ - fp16
14
+ base_model: Qwen/Qwen2.5-72B-Instruct
15
+ datasets:
16
+ - custom
17
  pipeline_tag: text-generation
18
  model-index:
19
+ - name: Elle-72B-Ultimate
20
  results: []
21
  ---
22
 
23
  # Elle-72B-Ultimate
24
 
25
+ **Elle** is a fine-tuned geopolitical intelligence model built on Qwen2.5-72B-Instruct-AWQ, specialized for:
26
 
27
+ - Real-time geopolitical risk assessment
28
+ - Multi-source intelligence synthesis
29
+ - Causal chain analysis for global events
30
+ - Regime stability detection
31
+ - Cascade risk prediction
32
 
33
  ## Model Details
34
 
35
+ | Attribute | Value |
36
+ |-----------|-------|
37
+ | Base Model | Qwen/Qwen2.5-72B-Instruct |
38
+ | Fine-tuning Method | LoRA (r=64, alpha=128) |
39
+ | Training Framework | Unsloth + PEFT |
40
+ | Precision | FP16 (full precision merged) |
41
+ | Context Length | 32,768 tokens |
42
+ | Final Training Loss | 0.2544 |
43
 
44
  ## Training Data
45
 
46
+ Elle was trained on curated geopolitical intelligence data including:
47
+
48
+ - **GDELT Event Data**: Global event monitoring and conflict detection
49
+ - **World Bank Indicators**: Economic stability metrics
50
+ - **USGS Seismic Data**: Natural disaster risk factors
51
+ - **Curated Intel Briefings**: Expert-verified geopolitical analysis
52
+ - **Cascade Analysis**: Historical event chain patterns
53
+
54
+ Training used interleaved conversation format with system prompts, user queries, and assistant responses.
55
 
56
  ## Intended Use
57
 
58
+ Elle is designed for:
59
+
60
+ - Enterprise geopolitical risk dashboards
61
+ - Intelligence briefing generation
62
+ - Supply chain risk assessment
63
+ - Investment risk analysis
64
+ - Policy impact modeling
65
 
66
  ## Limitations
67
 
68
+ - Knowledge cutoff aligned with training data (Dec 2024)
69
+ - Requires external data feeds for real-time analysis
70
+ - Should be used as analytical support, not sole decision-maker
71
+ - May reflect biases present in training data sources
72
+
73
+ ## Hardware Requirements
74
+
75
+ - **Inference**: 4x H100/H200 80GB (vLLM recommended)
76
+ - **Memory**: ~280GB VRAM for FP16 model (4x H200 = 320GB)
77
+ - Consider quantizing to AWQ/GPTQ for smaller deployments
78
 
79
+ ## Usage with vLLM
80
+
81
+ ```python
82
+ from vllm import LLM, SamplingParams
83
+
84
+ llm = LLM(
85
+ model="aphoticshaman/Elle-72B-Ultimate",
86
+ tensor_parallel_size=4,
87
+ trust_remote_code=True,
88
+ max_model_len=32768,
89
+ )
90
+
91
+ sampling_params = SamplingParams(
92
+ temperature=0.7,
93
+ top_p=0.9,
94
+ max_tokens=4096,
95
+ )
96
+
97
+ prompt = """<|im_start|>system
98
+ You are Elle, an expert geopolitical intelligence analyst.
99
+ <|im_end|>
100
+ <|im_start|>user
101
+ Analyze the current risk factors affecting semiconductor supply chains.
102
+ <|im_end|>
103
+ <|im_start|>assistant
104
+ """
105
+
106
+ outputs = llm.generate([prompt], sampling_params)
107
+ print(outputs[0].outputs[0].text)
108
+ ```
109
+
110
+ ## Usage with Transformers
111
 
112
  ```python
113
  from transformers import AutoModelForCausalLM, AutoTokenizer
114
 
115
  model = AutoModelForCausalLM.from_pretrained(
116
+ "aphoticshaman/Elle-72B-Ultimate",
 
117
  device_map="auto",
118
+ trust_remote_code=True,
119
  )
120
+ tokenizer = AutoTokenizer.from_pretrained("aphoticshaman/Elle-72B-Ultimate")
121
 
122
  messages = [
123
+ {"role": "system", "content": "You are Elle, an expert geopolitical intelligence analyst."},
124
+ {"role": "user", "content": "What are the key risk indicators for the South China Sea region?"}
125
  ]
126
 
127
  text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
 
130
  print(tokenizer.decode(outputs[0], skip_special_tokens=True))
131
  ```
132
 
133
+ ## Training Configuration
134
+
135
+ ```yaml
136
+ # LoRA Configuration
137
+ lora_r: 64
138
+ lora_alpha: 128
139
+ lora_dropout: 0.05
140
+ target_modules:
141
+ - q_proj
142
+ - k_proj
143
+ - v_proj
144
+ - o_proj
145
+ - gate_proj
146
+ - up_proj
147
+ - down_proj
148
+
149
+ # Training Hyperparameters
150
+ learning_rate: 2e-5
151
+ batch_size: 2
152
+ gradient_accumulation_steps: 8
153
+ epochs: 3
154
+ warmup_ratio: 0.03
155
+ lr_scheduler: cosine
156
+ optimizer: adamw_8bit
157
+ max_seq_length: 8192
158
+ ```
159
+
160
  ## Citation
161
 
162
  ```bibtex
163
  @misc{elle-72b-ultimate,
164
+ author = {LatticeForge},
165
+ title = {Elle-72B-Ultimate: Fine-tuned Geopolitical Intelligence Model},
166
  year = {2024},
167
  publisher = {HuggingFace},
168
+ url = {https://huggingface.co/aphoticshaman/Elle-72B-Ultimate}
169
  }
170
  ```
171
+
172
+ ## License
173
+
174
+ Apache 2.0 - See LICENSE file for details.
175
+
176
+ ## Contact
177
+
178
+ - **Website**: [latticeforge.ai](https://latticeforge.ai)
179
+ - **Issues**: Report issues via HuggingFace discussions
chat_template.jinja CHANGED
@@ -1,4 +1,54 @@
1
- {% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '
2
- ' + message['content'] + '<|im_end|>' + '
3
- '}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
4
- ' }}{% endif %}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0]['role'] == 'system' %}
4
+ {{- messages[0]['content'] }}
5
+ {%- else %}
6
+ {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}
7
+ {%- endif %}
8
+ {{- "\n\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
9
+ {%- for tool in tools %}
10
+ {{- "\n" }}
11
+ {{- tool | tojson }}
12
+ {%- endfor %}
13
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
14
+ {%- else %}
15
+ {%- if messages[0]['role'] == 'system' %}
16
+ {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
17
+ {%- else %}
18
+ {{- '<|im_start|>system\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\n' }}
19
+ {%- endif %}
20
+ {%- endif %}
21
+ {%- for message in messages %}
22
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) or (message.role == "assistant" and not message.tool_calls) %}
23
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
24
+ {%- elif message.role == "assistant" %}
25
+ {{- '<|im_start|>' + message.role }}
26
+ {%- if message.content %}
27
+ {{- '\n' + message.content }}
28
+ {%- endif %}
29
+ {%- for tool_call in message.tool_calls %}
30
+ {%- if tool_call.function is defined %}
31
+ {%- set tool_call = tool_call.function %}
32
+ {%- endif %}
33
+ {{- '\n<tool_call>\n{"name": "' }}
34
+ {{- tool_call.name }}
35
+ {{- '", "arguments": ' }}
36
+ {{- tool_call.arguments | tojson }}
37
+ {{- '}\n</tool_call>' }}
38
+ {%- endfor %}
39
+ {{- '<|im_end|>\n' }}
40
+ {%- elif message.role == "tool" %}
41
+ {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != "tool") %}
42
+ {{- '<|im_start|>user' }}
43
+ {%- endif %}
44
+ {{- '\n<tool_response>\n' }}
45
+ {{- message.content }}
46
+ {{- '\n</tool_response>' }}
47
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
48
+ {{- '<|im_end|>\n' }}
49
+ {%- endif %}
50
+ {%- endif %}
51
+ {%- endfor %}
52
+ {%- if add_generation_prompt %}
53
+ {{- '<|im_start|>assistant\n' }}
54
+ {%- endif %}
config.json CHANGED
@@ -102,7 +102,7 @@
102
  "rope_theta": 1000000.0,
103
  "sliding_window": null,
104
  "tie_word_embeddings": false,
105
- "torch_dtype": "bfloat16",
106
  "transformers_version": "4.55.2",
107
  "use_cache": true,
108
  "use_sliding_window": false,
 
102
  "rope_theta": 1000000.0,
103
  "sliding_window": null,
104
  "tie_word_embeddings": false,
105
+ "torch_dtype": "float16",
106
  "transformers_version": "4.55.2",
107
  "use_cache": true,
108
  "use_sliding_window": false,
model-00001-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:33e443682c5ec4dfcaf00e56c33d71487e52a2629e152f17314ac4cc6421a366
3
- size 4548798728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02e94cb29bf7fca9d9c4001232433066a31f5cb19264f58f7eac5319a78a4a77
3
+ size 4548798704
model-00002-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b42e3a369f47a829a465e9d26455d5b95674bd0712e6b7199fef3b0588455a54
3
- size 4964101384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74a659f56e027be779bc64bd5a0b13940b081d6410f78037c590ea42d0823aba
3
+ size 4964101360
model-00003-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:77050720124af4ed47599f5981cd7e9a396dbd55e1e2f644a1c419dd286fd796
3
- size 4781637328
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18597d4f9887d7bd2c178f8074faff08396fac5837e0db16aa91090c68bdbb26
3
+ size 4781637296
model-00004-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a5dad23fab593e7def1950b1a5875b865ae1c6b601f30db305292b3ffc2a00a
3
- size 4781670320
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f41467836aa2c89c586affb952b0c34a2969bf5389cb146472c885bf679514fa
3
+ size 4781670288
model-00005-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5950bb4b2b17c0c81d1ae52911335a97790a998ee308d0a14e2158f4d70ecc43
3
- size 4781670360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b64a0a445ec1b6e99a5eee676add2ecef3700427a9d1e2be2ed64b6c61634a02
3
+ size 4781670320
model-00006-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3b31272ce62f538dee1e9e9f8a3b4dee5b968ec15bbbfba3eb854a4b309680d1
3
- size 4964101416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5a9fdac270a2902b0668a5656800de418f53c7236213c27e5044cee9ab9847a
3
+ size 4964101384
model-00007-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8c0a5ebb8d05b5d8d32a843de1b2e175466bbf5631779605d03ee77f21163f96
3
- size 4781637360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f906c8ae2f0c9645e86cee0246b651e2f2975873761960c4d358de845e7060f9
3
+ size 4781637328
model-00008-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9dc45f7058d49305ddfeb8864a62dc4ff4838abf653ae0df5e22600b27a66aea
3
- size 4781670360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9fd352fabe610c9b343b18242edf797e7907135ae34ea6102e430063ba55aff
3
+ size 4781670320
model-00009-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e25f3b49f8aa5fc90409e6a4c174f3828663933269b66eb2416e27043b832f55
3
- size 4781670360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e360779c6c4d062d2ec189eeaca7471b2ec4b42ac5b140630eeece5f167be409
3
+ size 4781670320
model-00010-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f07eb9a77a9be793d6e996873ca3a56d1257afa7b986db990c6aecdb914671c5
3
- size 4964101416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4240a89649334f5618fbed1c088d6b43f6175c6ba35b3e6e314f28cc471878d4
3
+ size 4964101384
model-00011-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b22f20ceaac33d46ef5456498acdba8eadc1bebbfb8ab85dbf6566a98199f28c
3
- size 4781637360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54c09f55ca7bbd2cfb06edfd887a955e03c581abb0e72b169de78935e58bc282
3
+ size 4781637328
model-00012-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f8c497f55219ebcb607d24b9bae1dc3564ecb6c799cad3c1bf50bed6bd4115f0
3
- size 4781670360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c651322a75281b6adba0859dd99a9cea8bbd7996bccf9d79b7f68298c04bc1e3
3
+ size 4781670320
model-00013-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48b96ac2655eac43b011371dd1286fb11bb1eaef2cb28ded3dbef94edabcb0ed
3
- size 4781670360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9636ce7e9e81888a89d66689df777fd2b5a91c6b1c3a2423d28b664e95511eb5
3
+ size 4781670320
model-00014-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:23fb165457ef152707cc1c553fa42a2ccf3ac2275e2ce7744df8ce52f74984d4
3
- size 4964101416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3dde93dfeeef63fb6da85ae69fe6658791605a159c0f878ae6504fbeb5d20c20
3
+ size 4964101384
model-00015-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:674a1d53e15dbb069695e67a7bb1efbcea81ef441f3ff024dca75b4cd904c2f1
3
- size 4781637360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9585189efe02c72c1f1f75f76e32dd1848ce523c91ce4bab5fbee43f0f46331b
3
+ size 4781637328
model-00016-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8d580f3ec47ed48fc62f63072632c5a8de7c7bd71c4237a23e05f12656c1b693
3
- size 4781670360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d3244d6bf13386eb2a005c301df84414823294ebc422d7d31e7fa3c6f162138
3
+ size 4781670320
model-00017-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:93d62ebcdb1f3d8fdf5323d0cf3a9cb0905c1634108eb1245e4b3d0bad54fea1
3
- size 4781670360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb35b83cc9c83358ea049c01f4c87a708b0f2bbd1b425d2b0df7e1a8c2246dab
3
+ size 4781670320
model-00018-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2e853d9a60d03dd35ff4d189dd20f89b13efef69e06333054439d635bcff23bd
3
- size 4964101416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8865c1a0237577735e5e4f092b2e5ed1f4788a245415862a2840a9db0dfa085
3
+ size 4964101384
model-00019-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fc98d1520cef145bbb7c68ccd07309fd5adb6d5c601f8e936434acd3611d0ca0
3
- size 4781637360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92f2abf44ca6daac5b04e3dc608b411d636d69087b34fccf12dacc2fa67c9f47
3
+ size 4781637328
model-00020-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e32d886fdb74eddcca2f0b00e2d43cc440574e567a6945cd4d120af93c93ae87
3
- size 4781670360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a9c489e410b3ebdfd12351345a446003c315a05f1c23b915d63acbea941bc18
3
+ size 4781670320
model-00021-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e10fff18374b4a05184448ad96701a94993b19a4a01cd4c5baf6f04c40cb34c2
3
- size 4781670360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c10454ff29c93bdafad6489ed2c557ab182d50c102298e092013df97bc0a7e2
3
+ size 4781670320
model-00022-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ef0df8d6e6cd1c2a99778e2e7b57a98aa70a73ac755bc4c4c926091c3fd3a132
3
- size 4964101416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3d241bb7dc0467865109e6f722c81b69250838a1d306833e54be8b573b96ddb
3
+ size 4964101384
model-00023-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:34b66fb79ce0171484252ce49f736bc97fcd84d5015e3da68930c43c5b761fc5
3
- size 4781637360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35f37ce84ed2480d15420dff581c77822656e4ce30819ecd51c1f9c65b096aed
3
+ size 4781637328
model-00024-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:45dfc26410d63261d59706b42d72a550814973028aa1607bf8461e6708dcad6f
3
- size 4781670360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a03aeca9fba8c3ad87c79ecdf343dc1a96ab775d0587d430b3105d7119b8122a
3
+ size 4781670320
model-00025-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c54113b9e220bf6fa13cda0d41f6a1d4607b2f925d23c9209dc4a5087788a151
3
- size 4781670360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a8ed90dadbd485eadf13dc5b679afdbcf6c435cddbf6499fda30299911546d8
3
+ size 4781670320
model-00026-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:25b6ba219c5874de068b5385df7ca11f0ad4f359fd9751faf7238e4bb6336c34
3
- size 4964101416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ee73ed2b9d7b0572627563778c1e6157172022cb577fea1e2518d77130678aa
3
+ size 4964101384
model-00027-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f788e1cfef952c051f389788fa373e918317c78db2b5389d4ee1e56ca748ecf0
3
- size 4781637360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2b7cc967007bfc1726aac6bead9032204d51907b2fe5ef6c6eabcd58ef260ac
3
+ size 4781637328
model-00028-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3c70b902f54179a7979cf136f5206d1d75f6f826cb47d342622a5b4fac7a53e3
3
- size 4781670360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99da19d5cb225994dfe1e6ca5e5826fe1abc7e879c972b0362bc73235dae2bb8
3
+ size 4781670320
model-00029-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8c072f2a9780870bb82ad9fc91c82a26ef30e3cb133c3bf652cd6ec13026f89e
3
- size 4781670360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6107b88dfcebe349cdc17d13f04b040005a173d7a32dd7ffa4f738f10827861b
3
+ size 4781670320
model-00030-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f5f68a6ea6d2e13bf3a4f15c8bd9f9e3bf61dff46e94f85dd60a434ca1946e3d
3
- size 3208747032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e5082379f1e37f8fb5de2576d746035abe5f19f94496c65d4d09aed4350d706
3
+ size 3208747016
model-00031-of-00031.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cdbc8f91c574681655bd34786cb3923093027809999083ab1843ea595d8b9a49
3
  size 2491416704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:220efb428e7b47b14df2fd0e7afb1da052f6d9a1f0b5ea7beeb440dc8247f949
3
  size 2491416704
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
- size 11421896
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8506e7111b80c6d8635951a02eab0f4e1a8e4e5772da83846579e97b16f61bf
3
+ size 7031673
tokenizer_config.json CHANGED
@@ -195,10 +195,10 @@
195
  "<|video_pad|>"
196
  ],
197
  "bos_token": null,
 
198
  "clean_up_tokenization_spaces": false,
199
  "eos_token": "<|im_end|>",
200
  "errors": "replace",
201
- "extra_special_tokens": {},
202
  "model_max_length": 131072,
203
  "pad_token": "<|endoftext|>",
204
  "split_special_tokens": false,
 
195
  "<|video_pad|>"
196
  ],
197
  "bos_token": null,
198
+ "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
199
  "clean_up_tokenization_spaces": false,
200
  "eos_token": "<|im_end|>",
201
  "errors": "replace",
 
202
  "model_max_length": 131072,
203
  "pad_token": "<|endoftext|>",
204
  "split_special_tokens": false,