nehcgs commited on
Commit
8736e6d
·
verified ·
1 Parent(s): 0d2a03d

Upload folder using huggingface_hub

Browse files
Arch-Function-3B-Q2_K.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:69a466966913fd05f92a439290d52cabb7b80f9fd97aa9ab965b5454e176a59d
3
  size 1274755488
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac9a5a3d3ec8188b711c4c2601458cb11dd40ac47f435ddf2ac1adad06163dc2
3
  size 1274755488
Arch-Function-3B-Q3_K_L.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1123fa859a481a61a4a7b1cd8e2b773e2220de5b421ac0d3007e39dc5de96873
3
  size 1707391392
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3ebb8209e432f57e9d7513d3bb60cfb523414d5ec68d316afb5c765a97c9b92
3
  size 1707391392
Arch-Function-3B-Q3_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8ddff89ff7e2026117b40da195efecee218636f9ddad5a232a91377391c4dfb2
3
  size 1590475168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:badb6a19390fe7abcc0325b435e6ddf7ed0cd30d46967987a896bfaefe3ee6e8
3
  size 1590475168
Arch-Function-3B-Q3_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:957cb13016c32c233ed6af62fc4ca1dd2836351fadebb780d2230b08c79a0921
3
  size 1454356896
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c71b8fdd436c7a2d1f00ed4322a28b977c292b3dd6a798c0400e40d0ed14c03
3
  size 1454356896
Arch-Function-3B-Q4_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6d817fe6eb7db7f8aadcd04bd877cc06cac46490bd57c51c73a0cd9cae1482fe
3
  size 1929902496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7172964e86c7ffa28c7931f1bfb846e4c9d304b74439f14c63de0e918d711f0b
3
  size 1929902496
Arch-Function-3B-Q4_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5b527f5af25182063d3a6c78178c5028f889cb6b2ec64b435795ae4c389127b9
3
  size 1834383776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e80ad78953761dd86db7a7c2256ddd528b0c23f95a1e31be1c8bc93e3bf41fb
3
  size 1834383776
Arch-Function-3B-Q5_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4c9693b603547350e6055be08fdc4afa92e13497f10295eba29d2469e9bd8b55
3
  size 2224814496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0dd4e6a39763d2175c0b3a3c64f96b2c391f9b5c3ebc01c1fa26493131173095
3
  size 2224814496
Arch-Function-3B-Q5_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2e743d7ed116ada8bc3f4371b0621aeffa0ed7c3823c28b8ca0edeb52f9a1e12
3
  size 2169665952
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:294d4570d2cedc163c4e31d1385d1f8208bd8441b2fb281cdc58bc03c3c1f199
3
  size 2169665952
Arch-Function-3B-Q6_K.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd71a76bd294fd15b3b5958d872fcda16a9b5b023ab705ee2aef5a56a294b822
3
  size 2538158496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b4ae2627be1de16a42e01103fbd399028394fc3e28bc2cecaf97eb84c8dded3
3
  size 2538158496
Arch-Function-3B.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b7c9da8f7ec415891a2f575945d5e7886d76614fc2fa82b629bb50f629b88529
3
  size 6178316704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74f797d8c23e83f887de4d5cfd13db3d3de4bdb05bd5345b674e014f51f53c0b
3
  size 6178316704
README.md CHANGED
@@ -2,16 +2,16 @@
2
  license: other
3
  license_name: katanemo-research
4
  license_link: >-
5
- https://huggingface.co/katanemo/Arch-Function-3B.gguf/blob/main/LICENSE
6
  base_model:
7
- - Qwen/Qwen2.5-3B-Instruct
8
  language:
9
  - en
10
  pipeline_tag: text-generation
11
  library_name: transformers
12
  ---
13
 
14
- # katanemo/Arch-Function-3B
15
 
16
  ## Overview
17
  The Katanemo Arch-Function collection of large language models (LLMs) is a collection state-of-the-art (SOTA) LLMs specifically designed for **function calling** tasks. The models are designed to understand complex function signatures, identify required parameters, and produce accurate function call outputs based on natural language prompts. Achieving performance on par with GPT-4, these models set a new benchmark in the domain of function-oriented tasks, making them suitable for scenarios where automated API interaction and function execution is crucial.
@@ -54,7 +54,7 @@ Katanemo Arch-Function collection is built on top of the [Qwen 2.5](https://hugg
54
 
55
 
56
  ## Performance Benchmarks
57
- We evaluate Katanemo Arch-Function series on the [Berkeley Function-Calling Leaderboard (BFCL)](https://gorilla.cs.berkeley.edu/leaderboard.html#leaderboard). For each model family, we select the one with the highest rank. The results are shwon below:
58
 
59
  <table>
60
  <tr style="text-align: center; vertical-align: middle; font-weight: bold;">
@@ -75,109 +75,186 @@ We evaluate Katanemo Arch-Function series on the [Berkeley Function-Calling Lead
75
  </tr>
76
  <tr style="text-align: center; vertical-align: middle;">
77
  <td>1</td>
78
- <td>GPT-4-turbo-2024-04-09</td>
79
- <td>59.49%</td>
80
- <td>82.65%</td>
81
- <td>83.80%</td>
82
- <td>73.39%</td>
83
- <td>21.62%</td>
 
 
 
 
 
 
 
 
 
 
 
84
  <td>70.73%</td>
85
- <td>79.79%</td>
86
  </tr>
87
  <tr style="text-align: center; vertical-align: middle;">
88
- <td>3</td>
89
- <td>xLAM-8x22b-r</td>
90
- <td>59.13%</td>
91
- <td>89.75%</td>
92
- <td>89.32%</td>
93
- <td>72.81%</td>
94
- <td>15.62%</td>
95
- <td>97.56%</td>
96
- <td>75.23%</td>
 
 
 
 
 
 
 
 
 
 
 
97
  </tr>
98
  <tr style="text-align: center; vertical-align: middle; font-weight: bold;">
99
  <td> </td>
100
  <td>Arch-Function-7B</td>
101
- <td>57.48%</td>
102
- <td>87.50%</td>
103
- <td>86.80%</td>
104
- <td>72.19%</td>
105
- <td>13.75%</td>
106
- <td>82.93%</td>
107
- <td>79.54%</td>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
108
  </tr>
109
  <tr style="text-align: center; vertical-align: middle; font-weight: bold;">
110
  <td> </td>
111
  <td>Arch-Function-3B</td>
112
- <td>56.23%</td>
113
- <td>85.10%</td>
114
- <td>89.16%</td>
115
- <td>70.72%</td>
116
- <td>12.28%</td>
117
- <td>90.24%</td>
118
- <td>73.98%</td>
119
- </tr>
120
- <tr style="text-align: center; vertical-align: middle;">
121
- <td>7</td>
122
- <td>mistral-large-2407</td>
123
- <td>55.82%</td>
124
- <td>84.12%</td>
125
- <td>83.09%</td>
126
- <td>67.17%</td>
127
- <td>20.50%</td>
128
- <td>78.05%</td>
129
- <td>48.93%</td>
130
- </tr>
131
- <tr style="text-align: center; vertical-align: middle;">
132
- <td>9</td>
133
- <td>Claude-3.5-Sonnet-20240620</td>
134
- <td>54.83%</td>
135
- <td>70.35%</td>
136
- <td>66.34%</td>
137
- <td>71.39%</td>
138
- <td>23.5%</td>
139
- <td>63.41%</td>
140
- <td>75.91%</td>
141
  </tr>
142
  </tr>
143
  <tr style="text-align: center; vertical-align: middle; font-weight: bold;">
144
  <td> </td>
145
  <td>Arch-Function-1.5B</td>
146
- <td>53.61%</td>
147
- <td>82.60%</td>
148
- <td>87.36%</td>
149
- <td>68.19%</td>
150
- <td>8.62%</td>
151
- <td>87.80%</td>
152
- <td>75.90%</td>
153
  </tr>
154
- <tr style="text-align: center; vertical-align: middle;">
155
- <td>11</td>
156
- <td>o1-mini-2024-09-12</td>
157
- <td>53.43%</td>
158
- <td>75.48%</td>
159
- <td>76.86%</td>
160
- <td>71.17%</td>
161
- <td>11.00%</td>
162
- <td>46.34%</td>
163
- <td>88.07%</td>
164
  </tr>
165
- <tr style="text-align: center; vertical-align: middle;">
166
- <td>12</td>
167
- <td>Gemini-1.5-Flash-Preview-0514</td>
168
- <td>53.01%</td>
169
- <td>77.10%</td>
170
- <td>71.23%</td>
171
- <td>71.17%</td>
172
- <td>13.12%</td>
173
- <td>60.98%</td>
174
- <td>76.15%</td>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
175
  </tr>
176
  </table>
177
 
178
 
179
  # Requirements
180
- The code of Arch-Function-3B has been in the Hugging Face `transformers` library and we advise you to install latest version:
181
  ```bash
182
  pip install transformers>=4.37.0
183
  ```
@@ -193,7 +270,7 @@ import json
193
  from typing import Any, Dict, List
194
  from transformers import AutoModelForCausalLM, AutoTokenizer
195
 
196
- model_name = "katanemo/Arch-Function-3B"
197
  model = AutoModelForCausalLM.from_pretrained(
198
  model_name, device_map="auto", torch_dtype="auto", trust_remote_code=True
199
  )
@@ -332,4 +409,4 @@ The current temperature in Seattle is 62 degrees in Fahrenheit.
332
 
333
 
334
  # License
335
- Katanemo Arch-Function collection is distributed under the [Katanemo license](https://huggingface.co/katanemo/Arch-Function-3B.gguf/blob/main/LICENSE).
 
2
  license: other
3
  license_name: katanemo-research
4
  license_link: >-
5
+ https://huggingface.co/katanemolabs/Arch-Function-1.5B/blob/main/LICENSE
6
  base_model:
7
+ - Qwen/Qwen2.5-1.5B-Instruct
8
  language:
9
  - en
10
  pipeline_tag: text-generation
11
  library_name: transformers
12
  ---
13
 
14
+ # katanemo/Arch-Function-1.5B
15
 
16
  ## Overview
17
  The Katanemo Arch-Function collection of large language models (LLMs) is a collection state-of-the-art (SOTA) LLMs specifically designed for **function calling** tasks. The models are designed to understand complex function signatures, identify required parameters, and produce accurate function call outputs based on natural language prompts. Achieving performance on par with GPT-4, these models set a new benchmark in the domain of function-oriented tasks, making them suitable for scenarios where automated API interaction and function execution is crucial.
 
54
 
55
 
56
  ## Performance Benchmarks
57
+ We evaluate Katanemo Arch-Function series on the [Berkeley Function-Calling Leaderboard (BFCL)](https://gorilla.cs.berkeley.edu/leaderboard.html#leaderboard). For each model family, we select the one with the highest rank. The results (as of Oct 21st, 2024) are shwon below:
58
 
59
  <table>
60
  <tr style="text-align: center; vertical-align: middle; font-weight: bold;">
 
75
  </tr>
76
  <tr style="text-align: center; vertical-align: middle;">
77
  <td>1</td>
78
+ <td>GPT-4o-2024-08-06 (FC)</td>
79
+ <td>62.19%</td>
80
+ <td>85.90%</td>
81
+ <td>85.64%</td>
82
+ <td>75.43%</td>
83
+ <td>25.00%</td>
84
+ <td>63.41%</td>
85
+ <td>82.93%</td>
86
+ </tr>
87
+ <tr style="text-align: center; vertical-align: middle;">
88
+ <td>2</td>
89
+ <td>Functionary-Medium-v3.1 (FC)</td>
90
+ <td>62.02%</td>
91
+ <td>89.52%</td>
92
+ <td>89.77%</td>
93
+ <td>73.48%</td>
94
+ <td>23.50%</td>
95
  <td>70.73%</td>
96
+ <td>73.32%</td>
97
  </tr>
98
  <tr style="text-align: center; vertical-align: middle;">
99
+ <td>5</td>
100
+ <td>ToolACE-8B (FC)</td>
101
+ <td>60.44%</td>
102
+ <td>87.06%</td>
103
+ <td>89.52%</td>
104
+ <td>74.99%</td>
105
+ <td>17.38%</td>
106
+ <td>80.49%</td>
107
+ <td>85.71%</td>
108
+ </tr>
109
+ <tr style="text-align: center; vertical-align: middle;">
110
+ <td>6</td>
111
+ <td>o1-preview-2024-09-12 (Prompt)</td>
112
+ <td>59.27%</td>
113
+ <td>86.42%</td>
114
+ <td>88.88%</td>
115
+ <td>73.08%</td>
116
+ <td>17.62%</td>
117
+ <td>73.17%</td>
118
+ <td>74.60%</td>
119
  </tr>
120
  <tr style="text-align: center; vertical-align: middle; font-weight: bold;">
121
  <td> </td>
122
  <td>Arch-Function-7B</td>
123
+ <td>58.44%</td>
124
+ <td>85.58%</td>
125
+ <td>88.14%</td>
126
+ <td>69.08%</td>
127
+ <td>20.50%</td>
128
+ <td>92.68%</td>
129
+ <td>74.05%</td>
130
+ </tr>
131
+ <tr style="text-align: center; vertical-align: middle; ">
132
+ <td>8</td>
133
+ <td>xLAM-8x22b-r (FC)</td>
134
+ <td>57.99%</td>
135
+ <td>88.15%</td>
136
+ <td>90.11%</td>
137
+ <td>71.97%</td>
138
+ <td>14.50%</td>
139
+ <td>85.37%</td>
140
+ <td>67.29%</td>
141
+ </tr>
142
+ <tr style="text-align: center; vertical-align: middle; ">
143
+ <td>9</td>
144
+ <td>Gemini-1.5-Flash-002 (Prompt)</td>
145
+ <td>57.92%</td>
146
+ <td>86.58%</td>
147
+ <td>89.48%</td>
148
+ <td>76.28%</td>
149
+ <td>9.88%</td>
150
+ <td>85.37%</td>
151
+ <td>78.54%</td>
152
+ </tr>
153
+ <tr style="text-align: center; vertical-align: middle; ">
154
+ <td>10</td>
155
+ <td>Hammer2.0-7b (FC)</td>
156
+ <td>57.69%</td>
157
+ <td>90.27%</td>
158
+ <td>89.25%</td>
159
+ <td>69.79%</td>
160
+ <td>14.75%</td>
161
+ <td>95.12%</td>
162
+ <td>68.46%</td>
163
+ </tr>
164
+ <tr style="text-align: center; vertical-align: middle; ">
165
+ <td>12</td>
166
+ <td>Claude-3.5-Sonnet-20240620 (FC)</td>
167
+ <td>57.42%</td>
168
+ <td>70.04%</td>
169
+ <td>66.27%</td>
170
+ <td>74.68%</td>
171
+ <td>28.38%</td>
172
+ <td>68.29%</td>
173
+ <td>74.58%</td>
174
+ </tr>
175
+ <tr style="text-align: center; vertical-align: middle; ">
176
+ <td>13</td>
177
+ <td>mistral-large-2407 (FC)</td>
178
+ <td>56.80%</td>
179
+ <td>86.62%</td>
180
+ <td>84.57%</td>
181
+ <td>68.37%</td>
182
+ <td>20.62%</td>
183
+ <td>75.61%</td>
184
+ <td>49.44%</td>
185
  </tr>
186
  <tr style="text-align: center; vertical-align: middle; font-weight: bold;">
187
  <td> </td>
188
  <td>Arch-Function-3B</td>
189
+ <td>56.57%</td>
190
+ <td>83.62%</td>
191
+ <td>85.36%</td>
192
+ <td>66.90%</td>
193
+ <td>19.50%</td>
194
+ <td>97.56%</td>
195
+ <td>70.99%</td>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
196
  </tr>
197
  </tr>
198
  <tr style="text-align: center; vertical-align: middle; font-weight: bold;">
199
  <td> </td>
200
  <td>Arch-Function-1.5B</td>
201
+ <td>54.52%</td>
202
+ <td>80.31%</td>
203
+ <td>82.04%</td>
204
+ <td>66.19%</td>
205
+ <td>17.25%</td>
206
+ <td>97.56%</td>
207
+ <td>69.95%</td>
208
  </tr>
209
+ <tr style="text-align: center; vertical-align: middle; ">
210
+ <td>19</td>
211
+ <td>xLAM-7b-r (FC)</td>
212
+ <td>54.41%</td>
213
+ <td>81.40%</td>
214
+ <td>83.46%</td>
215
+ <td>67.88%</td>
216
+ <td>14.50%</td>
217
+ <td>97.56%</td>
218
+ <td>64.05%</td>
219
  </tr>
220
+ <tr style="text-align: center; vertical-align: middle; ">
221
+ <td>20</td>
222
+ <td>Qwen2.5-7B-Instruct (Prompt)</td>
223
+ <td>54.27%</td>
224
+ <td>85.79%</td>
225
+ <td>88.13%</td>
226
+ <td>65.97%</td>
227
+ <td>11.25%</td>
228
+ <td>92.68%</td>
229
+ <td>64.95%</td>
230
+ </tr>
231
+ <tr style="text-align: center; vertical-align: middle; ">
232
+ <td>21</td>
233
+ <td>Llama-3.1-70B-Instruct (Prompt)</td>
234
+ <td>53.67%</td>
235
+ <td>88.90%</td>
236
+ <td>89.34%</td>
237
+ <td>61.13%</td>
238
+ <td>12.38%</td>
239
+ <td>92.68%</td>
240
+ <td>58.38%</td>
241
+ </tr>
242
+ <tr style="text-align: center; vertical-align: middle; ">
243
+ <td>22</td>
244
+ <td>Gemma-2-27b-it (Prompt)</td>
245
+ <td>53.66%</td>
246
+ <td>88.52%</td>
247
+ <td>87.89%</td>
248
+ <td>69.48%</td>
249
+ <td>4.12%</td>
250
+ <td>87.8%</td>
251
+ <td>68.76%</td>
252
  </tr>
253
  </table>
254
 
255
 
256
  # Requirements
257
+ The code of Arch-Function-1.5B has been in the Hugging Face `transformers` library and we advise you to install latest version:
258
  ```bash
259
  pip install transformers>=4.37.0
260
  ```
 
270
  from typing import Any, Dict, List
271
  from transformers import AutoModelForCausalLM, AutoTokenizer
272
 
273
+ model_name = "katanemo/Arch-Function-1.5B"
274
  model = AutoModelForCausalLM.from_pretrained(
275
  model_name, device_map="auto", torch_dtype="auto", trust_remote_code=True
276
  )
 
409
 
410
 
411
  # License
412
+ Katanemo Arch-Function collection is distributed under the [Katanemo license](https://huggingface.co/katanemolabs/Arch-Function-1.5B/blob/main/LICENSE).