yoeven commited on
Commit
3ea216b
·
verified ·
1 Parent(s): f2ac9e0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +216 -145
README.md CHANGED
@@ -1,113 +1,35 @@
1
- ---
2
- title: README
3
- emoji: 📊
4
- colorFrom: blue
5
- colorTo: purple
6
- sdk: static
7
- pinned: false
8
- ---
9
 
10
- # The AI model built for deterministic developer tasks - Interfaze
11
 
 
12
 
13
  Interfaze is an AI model built on a new architecture that merges specialized DNN/CNN models with LLMs for developer tasks that require deterministic output and high consistency like OCR, scraping, classification, web search and more.
14
 
15
- * OCR, web scraping, web search, classification and more
16
- * OpenAI chat completion API compatible
17
- * High accuracy structured output consistency
18
- * Built-in code execution and sandboxing
19
- * Custom web engine for scraping and web research capabilities
20
- * Auto reasoning when needed
21
- * Controllable guardrails
22
- * Fully managed and scalable
23
- * Globally distributed fallback system with high uptime
24
-
25
- Fullsite: https://interfaze.ai
26
-
27
- ## Model Comparison
28
-
29
- * Benchmark: MMLU-Pro
30
- * interfaze-beta: 83.6
31
- * GPT-4.1: 80.6
32
- * Claude Sonnet 4: 83.7
33
- * Gemini 2.5 Flash: 80.9
34
- * Claude Sonnet 4 (Thinking): 83.7
35
- * Claude Opus 4 (Thinking): 86
36
- * GPT-5-Minimal: 80.6
37
- * Gemini-2.5-Pro: 86.2
38
- * Benchmark: MMLU
39
- * interfaze-beta: 91.38
40
- * GPT-4.1: 90.2
41
- * Claude Sonnet 4: -
42
- * Gemini 2.5 Flash: -
43
- * Claude Sonnet 4 (Thinking): 88.8
44
- * Claude Opus 4 (Thinking): 89
45
- * GPT-5-Minimal: -
46
- * Gemini-2.5-Pro: 89.2
47
- * Benchmark: MMMU
48
- * interfaze-beta: 77.33
49
- * GPT-4.1: 74.8
50
- * Claude Sonnet 4: -
51
- * Gemini 2.5 Flash: 79.7
52
- * Claude Sonnet 4 (Thinking): 74.4
53
- * Claude Opus 4 (Thinking): 76.5
54
- * GPT-5-Minimal: -
55
- * Gemini-2.5-Pro: 82
56
- * Benchmark: AIME-2025
57
- * interfaze-beta: 90
58
- * GPT-4.1: 34.7
59
- * Claude Sonnet 4: 38
60
- * Gemini 2.5 Flash: 60.3
61
- * Claude Sonnet 4 (Thinking): 74.3
62
- * Claude Opus 4 (Thinking): 73.3
63
- * GPT-5-Minimal: 31.7
64
- * Gemini-2.5-Pro: 87.7
65
- * Benchmark: GPQA-Diamond
66
- * interfaze-beta: 81.31
67
- * GPT-4.1: 66.3
68
- * Claude Sonnet 4: 68.3
69
- * Gemini 2.5 Flash: 68.3
70
- * Claude Sonnet 4 (Thinking): 77.7
71
- * Claude Opus 4 (Thinking): 79.6
72
- * GPT-5-Minimal: 67.3
73
- * Gemini-2.5-Pro: 84.4
74
- * Benchmark: LiveCodeBench
75
- * interfaze-beta: 57.77
76
- * GPT-4.1: 45.7
77
- * Claude Sonnet 4: 44.9
78
- * Gemini 2.5 Flash: 49.5
79
- * Claude Sonnet 4 (Thinking): 65.5
80
- * Claude Opus 4 (Thinking): 63.6
81
- * GPT-5-Minimal: 55.8
82
- * Gemini-2.5-Pro: 75.9
83
- * Benchmark: ChartQA
84
- * interfaze-beta: 90.88
85
- * GPT-4.1: -
86
- * Claude Sonnet 4: -
87
- * Gemini 2.5 Flash: -
88
- * Claude Sonnet 4 (Thinking): -
89
- * Claude Opus 4 (Thinking): -
90
- * GPT-5-Minimal: -
91
- * Gemini-2.5-Pro: -
92
- * Benchmark: AI2D
93
- * interfaze-beta: 91.51
94
- * GPT-4.1: 85.9
95
- * Claude Sonnet 4: -
96
- * Gemini 2.5 Flash: -
97
- * Claude Sonnet 4 (Thinking): -
98
- * Claude Opus 4 (Thinking): -
99
- * GPT-5-Minimal: -
100
- * Gemini-2.5-Pro: 89.5
101
- * Benchmark: Common-Voice-v16
102
- * interfaze-beta: 90.8
103
- * GPT-4.1: -
104
- * Claude Sonnet 4: -
105
- * Gemini 2.5 Flash: -
106
- * Claude Sonnet 4 (Thinking): -
107
- * Claude Opus 4 (Thinking): -
108
- * GPT-5-Minimal: -
109
- * Gemini-2.5-Pro: -
110
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
111
 
112
  \*Results for Non-Interfaze models are sourced from model providers, leaderboards, and evaluation providers such as Artificial Analysis.
113
 
@@ -115,70 +37,219 @@ Fullsite: https://interfaze.ai
115
 
116
  OpenAI API compatible, works with every AI SDK out of the box
117
 
118
- OpenAI SDK
 
119
 
120
- Vercel AI SDK
 
 
 
121
 
122
- Langchain SDK
 
 
 
 
 
 
 
 
123
 
124
- ![Extraction](https://interfaze.ai/examples/extraction_example.png)
125
-
126
- ![scraping](https://interfaze.ai/examples/scraper_example.png)
127
-
128
- Fully configurable guardrails for text and images
129
-
130
- ![Extraction](https://interfaze.ai/examples/nsfw_example.jpg)
131
-
132
- This architecture combines a suite of small specialized models supported with custom tools and infrastructure while automatically routing to the best model for the task that prioritizes accuracy and speed.
133
 
134
- [![How it works](https://interfaze.ai/examples/howitworks.png)
135
 
 
136
 
 
 
137
 
138
- ](/examples/howitworks.png)
 
 
 
 
 
 
139
 
140
- ### Specs
141
 
142
- Max output tokens
143
 
144
- 32k tokens
145
 
146
- Input modalities
 
147
 
148
- Text, Images, Audio, File, Video
 
 
 
 
 
 
 
149
 
150
- Output tokens
151
 
152
- $3.50 / MTok
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
153
 
154
- Observability & Logging
155
 
156
- Coming soon
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
157
 
158
- ### FAQ
159
 
160
- ### Todo (Prioritized)
161
 
162
- * Reduce transactional token count
163
- * Pre-built prompts/schemas optimized for specific tasks
164
- * Embedding model
165
- * Built-in observability and logging on the dashboard
166
- * Complete metrics and analytics
167
- * v1.1 Interfaze
168
- * Reduce latency and improve throughput
169
- * Custom SDKs for interfaze with AI SDK, Langchain, etc.
170
- * Leaderboard for projects
171
 
172
- If you have feature requests or recommendations, please reach out!
 
 
 
173
 
174
  ### Research references
175
 
176
- * [Interfaze: The Future of AI is built on Task-Specific Small Models](https://www.arxiv.org/abs/2602.04101)
177
- * [Agentic Context Engineering](https://www.arxiv.org/pdf/2510.04618)
178
- * [Small Language Models are the Future of Agentic AI](https://arxiv.org/pdf/2506.02153)
179
- * [The Sparsely-Gated Mixture-of-Experts Layer](https://arxiv.org/pdf/1701.06538)
180
- * [DeepSeekMoE](https://arxiv.org/pdf/2401.06066)
181
- * [Confronting LLMs with Traditional ML](https://arxiv.org/pdf/2310.14607)
182
 
183
  ### Who are we?
184
 
 
 
 
 
 
 
 
 
 
1
 
 
2
 
3
+ # The AI model built for deterministic developer tasks
4
 
5
  Interfaze is an AI model built on a new architecture that merges specialized DNN/CNN models with LLMs for developer tasks that require deterministic output and high consistency like OCR, scraping, classification, web search and more.
6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
 
8
+ [Try now](https://interfaze.ai/dashboard) or [Read paper](https://www.arxiv.org/abs/2602.04101)
9
+
10
+ - OCR, web scraping, web search, classification and more
11
+ - OpenAI chat completion API compatible
12
+ - High accuracy structured output consistency
13
+ - Built-in code execution and sandboxing
14
+ - Custom web engine for scraping and web research capabilities
15
+ - Auto reasoning when needed
16
+ - Controllable guardrails
17
+ - Fully managed and scalable
18
+ - Globally distributed fallback system with high uptime
19
+
20
+ ### Model Comparison
21
+
22
+ | Benchmark | interfaze-beta | GPT-4.1 | Claude Sonnet 4 | Gemini 2.5 Flash | Claude Sonnet 4 (Thinking) | Claude Opus 4 (Thinking) | GPT-5-Minimal | Gemini-2.5-Pro |
23
+ | --- | --- | --- | --- | --- | --- | --- | --- | --- |
24
+ | MMLU-Pro | 83.6 | 80.6 | 83.7 | 80.9 | 83.7 | 86 | 80.6 | 86.2 |
25
+ | MMLU | 91.38 | 90.2 | - | - | 88.8 | 89 | - | 89.2 |
26
+ | MMMU | 77.33 | 74.8 | - | 79.7 | 74.4 | 76.5 | - | 82 |
27
+ | AIME-2025 | 90 | 34.7 | 38 | 60.3 | 74.3 | 73.3 | 31.7 | 87.7 |
28
+ | GPQA-Diamond | 81.31 | 66.3 | 68.3 | 68.3 | 77.7 | 79.6 | 67.3 | 84.4 |
29
+ | LiveCodeBench | 57.77 | 45.7 | 44.9 | 49.5 | 65.5 | 63.6 | 55.8 | 75.9 |
30
+ | ChartQA | 90.88 | - | - | - | - | - | - | - |
31
+ | AI2D | 91.51 | 85.9 | - | - | - | - | - | 89.5 |
32
+ | Common-Voice-v16 | 90.8 | - | - | - | - | - | - | - |
33
 
34
  \*Results for Non-Interfaze models are sourced from model providers, leaderboards, and evaluation providers such as Artificial Analysis.
35
 
 
37
 
38
  OpenAI API compatible, works with every AI SDK out of the box
39
 
40
+ ```
41
+ import OpenAI from "openai";
42
 
43
+ const interfaze = new OpenAI({
44
+ baseURL: "https://api.interfaze.ai/v1",
45
+ apiKey: "<your-api-key>"
46
+ });
47
 
48
+ const completion = await interfaze.chat.completions.create({
49
+ model: "interfaze-beta",
50
+ messages: [\
51
+ {\
52
+ role: "user",\
53
+ content: "Get the company description of JigsawStack from their linkedin page",\
54
+ },\
55
+ ],
56
+ });
57
 
58
+ console.log(completion.choices[0].message.content);
59
+ ```
 
 
 
 
 
 
 
60
 
61
+ ### OCR & Document Extraction
62
 
63
+ [vision docs ->](https://interfaze.ai/docs/vision)
64
 
65
+ ```
66
+ prompt = "Get the person information from the following ID."
67
 
68
+ schema = z.object({
69
+ first_name: z.string(),
70
+ last_name: z.string(),
71
+ dob: z.string(),
72
+ expiry: z.string(),
73
+ });
74
+ ```
75
 
76
+ ![Extraction](https://interfaze.ai/examples/extraction_example.png)
77
 
78
+ ### Smart Web Scraping
79
 
80
+ [web docs ->](https://interfaze.ai/docs/web)
81
 
82
+ ```
83
+ prompt = "Extract the information from Yoeven D Khemlani's linkedin page based on the schema."
84
 
85
+ schema = z.object({
86
+ first_name: z.string(),
87
+ last_name: z.string(),
88
+ about: z.string(),
89
+ current_company: z.string(),
90
+ current_position: z.string(),
91
+ });
92
+ ```
93
 
94
+ ![scraping](https://interfaze.ai/examples/scraper_example.png)
95
 
96
+ ### Translation
97
+
98
+ [translation docs ->](https://interfaze.ai/docs/translation)
99
+
100
+ ```
101
+ prompt = "The UK drinks about 100–160 million cups of tea every day, and 98% of tea drinkers add milk to their tea."
102
+
103
+ schema = z.object({
104
+ zh: z.string(),
105
+ hi: z.string(),
106
+ es: z.string(),
107
+ fr: z.string(),
108
+ de: z.string(),
109
+ it: z.string(),
110
+ ja: z.string(),
111
+ ko: z.string(),
112
+ });
113
+ ```
114
+
115
+ ```
116
+ zh: 英国每天饮用约100–160百万杯茶,有98%的茶饮者在茶中加入牛奶。
117
+ hi: यूके हर दिन लगभग 100–160 मिलियन कप चाय पीता है, और 98% चाय पीने वाले अपनी चाय में दूध मिलाते हैं।
118
+ es: El Reino Unido bebe alrededor de 100–160 millones de tazas de té cada día, y el 98 % de los consumidores de té añade leche a su té.
119
+ fr: Le Royaume-Uni boit environ 100–160 millions de tasses de thé chaque jour, et 98 % des buveurs de thé ajoutent du lait à leur thé.
120
+ de: Das Vereinigte Königreich trinkt etwa 100–160 Millionen Tassen Tee pro Tag, und 98 % der Teetrinker fügen ihrem Tee Milch hinzu.
121
+ it: Il Regno Unito beve circa 100–160 milioni di tazze di tè ogni giorno e il 98% degli amanti del tè aggiunge latte al proprio tè.
122
+ ja: イギリスでは毎日約100~160百万杯の紅茶が飲まれており、紅茶を飲む人の98%が紅茶に牛乳を加えます。
123
+ ko: 영국에서는 매일 약 1억 ~ 1억 6천만 잔의 차를 마시며, 차를 마시는 사람의 98%가 차에 우유를 넣습니다.
124
+ ```
125
+
126
+ ### Speech-to-text (STT) and diarization
127
+
128
+ [stt docs ->](https://interfaze.ai/docs/speech-to-text)
129
+
130
+ ```
131
+ prompt = "Transcribe https://jigsawstack.com/preview/stt-example.wav"
132
+
133
+ schema = z.object({
134
+ text: z.string(),
135
+ speakers: z.object({
136
+ id: z.string(),
137
+ start: z.number(),
138
+ end: z.number()
139
+ })
140
+ });
141
+ ```
142
+
143
+ ```
144
+ {
145
+ "text": " The little tales they tell are false The door was barred, locked and bolted as well Ripe pears are fit for a queen's table A big wet stain was on the round carpet The kite dipped and swayed but stayed aloft The pleasant hours fly by much too soon The room was crowded with a mild wob The room was crowded with a wild mob This strong arm shall shield your honour She blushed when he gave her a white orchid The beetle droned in the hot June sun",
146
+ "speakers": [\
147
+ {\
148
+ "start":0,\
149
+ "end":4.78,\
150
+ "id": "SPEAKER_00"\
151
+ },\
152
+ {\
153
+ "start":4.78,\
154
+ "end":9.48,\
155
+ "id": "SPEAKER_00"\
156
+ },\
157
+ {\
158
+ "start":9.48,\
159
+ "end":13.06,\
160
+ "id": "SPEAKER_00"\
161
+ },\
162
+ {\
163
+ "start":13.06,\
164
+ "end":17.24,\
165
+ "id": "SPEAKER_00"\
166
+ },\
167
+ {\
168
+ "start":17.24,\
169
+ "end":21.78,\
170
+ "id": "SPEAKER_00"\
171
+ },\
172
+ {\
173
+ "start":21.78,\
174
+ "end":26.3,\
175
+ "id": "SPEAKER_00"\
176
+ },\
177
+ {\
178
+ "start":26.3,\
179
+ "end":30.76,\
180
+ "id": "SPEAKER_00"\
181
+ },\
182
+ {\
183
+ "start":30.76,\
184
+ "end":35.08,\
185
+ "id": "SPEAKER_00"\
186
+ },\
187
+ {\
188
+ "start":35.08,\
189
+ "end":39.24,\
190
+ "id": "SPEAKER_00"\
191
+ },\
192
+ {\
193
+ "start":39.24,\
194
+ "end":43.94,\
195
+ "id": "SPEAKER_00"\
196
+ },\
197
+ {\
198
+ "start":43.94,\
199
+ "end":48.5,\
200
+ "id": "SPEAKER_00"\
201
+ }\
202
+ ]
203
+ }
204
+ ```
205
+
206
+ ### Configurable guardrails and NSFW checks
207
+
208
+ [guardrails docs ->](https://interfaze.ai/docs/guard-rails)
209
 
210
+ Fully configurable guardrails for text and images
211
 
212
+ ```
213
+ S1: Violent Crimes
214
+ S2: Non-Violent Crimes
215
+ S3: Sex-Related Crimes
216
+ S4: Child Sexual Exploitation
217
+ S5: Defamation
218
+ S6: Specialized Advice
219
+ S7: Privacy
220
+ S8: Intellectual Property
221
+ S9: Indiscriminate Weapons
222
+ S10: Hate
223
+ S11: Suicide & Self-Harm
224
+ S12: Sexual Content
225
+ S12_IMAGE: Sexual Content (Image)
226
+ S13: Elections
227
+ S14: Code Interpreter Abuse
228
+ ```
229
+
230
+ ### Architecture
231
+
232
+ [read paper ->](https://www.arxiv.org/abs/2602.04101)
233
 
234
+ This architecture combines a suite of small specialized models supported with custom tools and infrastructure while automatically routing to the best model for the task that prioritizes accuracy and speed.
235
 
236
+ ![How it works](https://interfaze.ai/examples/howitworks.png)
237
 
238
+ ### Specs
 
 
 
 
 
 
 
 
239
 
240
+ - Context window: 1m tokens
241
+ - Max output tokens: 32k tokens
242
+ - Input modalities: Text, Images, Audio, File, Video
243
+ - Reasoning: Available
244
 
245
  ### Research references
246
 
247
+ - [Interfaze: The Future of AI is built on Task-Specific Small Models](https://www.arxiv.org/abs/2602.04101)
248
+ - [Agentic Context Engineering](https://www.arxiv.org/pdf/2510.04618)
249
+ - [Small Language Models are the Future of Agentic AI](https://arxiv.org/pdf/2506.02153)
250
+ - [The Sparsely-Gated Mixture-of-Experts Layer](https://arxiv.org/pdf/1701.06538)
251
+ - [DeepSeekMoE](https://arxiv.org/pdf/2401.06066)
252
+ - [Confronting LLMs with Traditional ML](https://arxiv.org/pdf/2310.14607)
253
 
254
  ### Who are we?
255