mazesmazes commited on
Commit
7e654c9
·
verified ·
1 Parent(s): c9c127a

Push checkpoint-20000 with tokenizer files and diarization.py

Browse files
.gitattributes CHANGED
@@ -1,3 +1,4 @@
1
  *.safetensors filter=lfs diff=lfs merge=lfs -text
2
  *.bin filter=lfs diff=lfs merge=lfs -text
3
  tokenizer_config.json -filter -diff -merge text
 
 
1
  *.safetensors filter=lfs diff=lfs merge=lfs -text
2
  *.bin filter=lfs diff=lfs merge=lfs -text
3
  tokenizer_config.json -filter -diff -merge text
4
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
chat_template.jinja ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0].role == 'system' %}
4
+ {{- messages[0].content + '\n\n' }}
5
+ {%- endif %}
6
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
+ {%- for tool in tools %}
8
+ {{- "\n" }}
9
+ {{- tool | tojson }}
10
+ {%- endfor %}
11
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
+ {%- else %}
13
+ {%- if messages[0].role == 'system' %}
14
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
+ {%- endif %}
16
+ {%- endif %}
17
+ {%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
18
+ {%- for message in messages[::-1] %}
19
+ {%- set index = (messages|length - 1) - loop.index0 %}
20
+ {%- if ns.multi_step_tool and message.role == "user" and message.content is string and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}
21
+ {%- set ns.multi_step_tool = false %}
22
+ {%- set ns.last_query_index = index %}
23
+ {%- endif %}
24
+ {%- endfor %}
25
+ {%- for message in messages %}
26
+ {%- if message.content is string %}
27
+ {%- set content = message.content %}
28
+ {%- else %}
29
+ {%- set content = '' %}
30
+ {%- endif %}
31
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
32
+ {{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
33
+ {%- elif message.role == "assistant" %}
34
+ {%- set reasoning_content = '' %}
35
+ {%- if message.reasoning_content is string %}
36
+ {%- set reasoning_content = message.reasoning_content %}
37
+ {%- else %}
38
+ {%- if '</think>' in content %}
39
+ {%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
40
+ {%- set content = content.split('</think>')[-1].lstrip('\n') %}
41
+ {%- endif %}
42
+ {%- endif %}
43
+ {%- if loop.index0 > ns.last_query_index %}
44
+ {%- if loop.last or (not loop.last and reasoning_content) %}
45
+ {{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content.strip('\n') + '\n</think>\n\n' + content.lstrip('\n') }}
46
+ {%- else %}
47
+ {{- '<|im_start|>' + message.role + '\n' + content }}
48
+ {%- endif %}
49
+ {%- else %}
50
+ {{- '<|im_start|>' + message.role + '\n' + content }}
51
+ {%- endif %}
52
+ {%- if message.tool_calls %}
53
+ {%- for tool_call in message.tool_calls %}
54
+ {%- if (loop.first and content) or (not loop.first) %}
55
+ {{- '\n' }}
56
+ {%- endif %}
57
+ {%- if tool_call.function %}
58
+ {%- set tool_call = tool_call.function %}
59
+ {%- endif %}
60
+ {{- '<tool_call>\n{"name": "' }}
61
+ {{- tool_call.name }}
62
+ {{- '", "arguments": ' }}
63
+ {%- if tool_call.arguments is string %}
64
+ {{- tool_call.arguments }}
65
+ {%- else %}
66
+ {{- tool_call.arguments | tojson }}
67
+ {%- endif %}
68
+ {{- '}\n</tool_call>' }}
69
+ {%- endfor %}
70
+ {%- endif %}
71
+ {{- '<|im_end|>\n' }}
72
+ {%- elif message.role == "tool" %}
73
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
74
+ {{- '<|im_start|>user' }}
75
+ {%- endif %}
76
+ {{- '\n<tool_response>\n' }}
77
+ {{- content }}
78
+ {{- '\n</tool_response>' }}
79
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
80
+ {{- '<|im_end|>\n' }}
81
+ {%- endif %}
82
+ {%- endif %}
83
+ {%- endfor %}
84
+ {%- if add_generation_prompt %}
85
+ {{- '<|im_start|>assistant\n' }}
86
+ {%- if true %}
87
+ {{- '<think>\n\n</think>\n\n' }}
88
+ {%- endif %}
89
+ {%- endif %}
preprocessor_config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "chunk_length": 30,
3
+ "dither": 0.0,
4
+ "feature_extractor_type": "WhisperFeatureExtractor",
5
+ "feature_size": 128,
6
+ "hop_length": 160,
7
+ "n_fft": 400,
8
+ "n_samples": 480000,
9
+ "nb_max_frames": 3000,
10
+ "padding": false,
11
+ "padding_side": "right",
12
+ "padding_value": 0.0,
13
+ "return_attention_mask": false,
14
+ "sampling_rate": 16000,
15
+ "processor_class": "ASRProcessor",
16
+ "auto_map": {
17
+ "AutoProcessor": "asr_processing.ASRProcessor"
18
+ }
19
+ }
projectors.py CHANGED
@@ -85,7 +85,6 @@ class SimpleAdapter(nn.Module):
85
  def forward(self, x: torch.Tensor) -> torch.Tensor:
86
  return self.fc2(self.act(self.fc1(x)))
87
 
88
-
89
  class MOSAProjector(nn.Module):
90
  """MOSA-Base projector: simple 2-layer ReLU router with 4 simple adapters.
91
 
@@ -127,10 +126,7 @@ class MOSAProjector(nn.Module):
127
  # --- 3. Experts (Simple 2-layer GELU adapters) ---
128
  # Each expert: llm_dim -> hidden -> llm_dim (much smaller than frame-stacking)
129
  self.experts = nn.ModuleList(
130
- [
131
- SimpleAdapter(self.llm_dim, adapter_hidden, self.llm_dim)
132
- for _ in range(self.num_experts)
133
- ]
134
  )
135
 
136
  def forward(self, x: torch.Tensor) -> torch.Tensor:
@@ -153,15 +149,18 @@ class MOSAProjector(nn.Module):
153
  routing_weights = F.softmax(self.router(x), dim=-1) # (B, out_len, num_experts)
154
 
155
  # --- 3. Expert Mixture (Dense Execution) ---
156
- expert_outputs = torch.stack([expert(x) for expert in self.experts]) # (E, B, out_len, D)
 
 
157
  return torch.einsum("ebsd, bse -> bsd", expert_outputs, routing_weights)
158
 
159
  def get_output_length(self, input_length: int) -> int:
160
  """Calculate output sequence length after Conv1d downsampling (4x reduction)."""
161
  # Conv1d with stride 2, kernel 3, padding 1: out = (in + 2*1 - 3) // 2 + 1 = (in - 1) // 2 + 1
162
  # Applied twice for 4x total reduction
163
- after_conv1 = (input_length + 2 * 1 - 3) // 2 + 1
164
- return (after_conv1 + 2 * 1 - 3) // 2 + 1
 
165
 
166
 
167
  # =============================================================================
 
85
  def forward(self, x: torch.Tensor) -> torch.Tensor:
86
  return self.fc2(self.act(self.fc1(x)))
87
 
 
88
  class MOSAProjector(nn.Module):
89
  """MOSA-Base projector: simple 2-layer ReLU router with 4 simple adapters.
90
 
 
126
  # --- 3. Experts (Simple 2-layer GELU adapters) ---
127
  # Each expert: llm_dim -> hidden -> llm_dim (much smaller than frame-stacking)
128
  self.experts = nn.ModuleList(
129
+ [SimpleAdapter(self.llm_dim, adapter_hidden, self.llm_dim) for _ in range(self.num_experts)]
 
 
 
130
  )
131
 
132
  def forward(self, x: torch.Tensor) -> torch.Tensor:
 
149
  routing_weights = F.softmax(self.router(x), dim=-1) # (B, out_len, num_experts)
150
 
151
  # --- 3. Expert Mixture (Dense Execution) ---
152
+ expert_outputs = torch.stack(
153
+ [expert(x) for expert in self.experts]
154
+ ) # (E, B, out_len, D)
155
  return torch.einsum("ebsd, bse -> bsd", expert_outputs, routing_weights)
156
 
157
  def get_output_length(self, input_length: int) -> int:
158
  """Calculate output sequence length after Conv1d downsampling (4x reduction)."""
159
  # Conv1d with stride 2, kernel 3, padding 1: out = (in + 2*1 - 3) // 2 + 1 = (in - 1) // 2 + 1
160
  # Applied twice for 4x total reduction
161
+ length = (input_length + 2 * 1 - 3) // 2 + 1 # First conv
162
+ length = (length + 2 * 1 - 3) // 2 + 1 # Second conv
163
+ return length
164
 
165
 
166
  # =============================================================================
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33b674fb8444e2553eae8f1b261093371920a28ef75b5c18f4deb3f9217ed0ba
3
+ size 11422834
tokenizer_config.json ADDED
Binary file (396 Bytes). View file