File size: 8,235 Bytes
864c30b
 
 
 
 
 
703ae5d
abc01c2
fbb364d
3ea216b
fbb364d
 
 
 
3ea216b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fbb364d
 
 
 
 
 
 
3ea216b
 
fbb364d
3ea216b
 
 
 
fbb364d
3ea216b
 
 
 
 
 
 
 
 
fbb364d
3ea216b
 
fbb364d
3ea216b
fbb364d
3ea216b
fbb364d
3ea216b
 
fbb364d
3ea216b
 
 
 
 
 
 
fbb364d
3ea216b
fbb364d
3ea216b
fbb364d
3ea216b
fbb364d
3ea216b
 
fbb364d
3ea216b
 
 
 
 
 
 
 
fbb364d
3ea216b
fbb364d
3ea216b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fbb364d
3ea216b
fbb364d
3ea216b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fbb364d
3ea216b
fbb364d
3ea216b
fbb364d
3ea216b
fbb364d
3ea216b
 
 
 
fbb364d
 
 
3ea216b
 
 
 
 
 
fbb364d
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
---
title: Interfaze
thumbnail: >-
  https://cdn-uploads.huggingface.co/production/uploads/62fb53b572a7ab50b4b06fca/II4KdeJkepE_NOg9HNnq_.jpeg
short_description: The AI model built for deterministic developer tasks
---

![Extraction](https://interfaze.ai/banner.png)

# The AI model built for deterministic developer tasks

Interfaze is an AI model built on a new architecture that merges specialized DNN/CNN models with LLMs for developer tasks that require deterministic output and high consistency like OCR, scraping, classification, web search and more.


[Try now](https://interfaze.ai/dashboard) or [Read paper](https://www.arxiv.org/abs/2602.04101)

- OCR, web scraping, web search, classification and more
- OpenAI chat completion API compatible
- High accuracy structured output consistency
- Built-in code execution and sandboxing
- Custom web engine for scraping and web research capabilities
- Auto reasoning when needed
- Controllable guardrails
- Fully managed and scalable
- Globally distributed fallback system with high uptime

### Model Comparison

| Benchmark | interfaze-beta | GPT-4.1 | Claude Sonnet 4 | Gemini 2.5 Flash | Claude Sonnet 4 (Thinking) | Claude Opus 4 (Thinking) | GPT-5-Minimal | Gemini-2.5-Pro |
| --- | --- | --- | --- | --- | --- | --- | --- | --- |
| MMLU-Pro | 83.6 | 80.6 | 83.7 | 80.9 | 83.7 | 86 | 80.6 | 86.2 |
| MMLU | 91.38 | 90.2 | - | - | 88.8 | 89 | - | 89.2 |
| MMMU | 77.33 | 74.8 | - | 79.7 | 74.4 | 76.5 | - | 82 |
| AIME-2025 | 90 | 34.7 | 38 | 60.3 | 74.3 | 73.3 | 31.7 | 87.7 |
| GPQA-Diamond | 81.31 | 66.3 | 68.3 | 68.3 | 77.7 | 79.6 | 67.3 | 84.4 |
| LiveCodeBench | 57.77 | 45.7 | 44.9 | 49.5 | 65.5 | 63.6 | 55.8 | 75.9 |
| ChartQA | 90.88 | - | - | - | - | - | - | - |
| AI2D | 91.51 | 85.9 | - | - | - | - | - | 89.5 |
| Common-Voice-v16 | 90.8 | - | - | - | - | - | - | - |

\*Results for Non-Interfaze models are sourced from model providers, leaderboards, and evaluation providers such as Artificial Analysis.

### Works like any other LLM

OpenAI API compatible, works with every AI SDK out of the box

```
import OpenAI from "openai";

const interfaze = new OpenAI({
    baseURL: "https://api.interfaze.ai/v1",
    apiKey: "<your-api-key>"
});

const completion = await interfaze.chat.completions.create({
    model: "interfaze-beta",
    messages: [\
        {\
            role: "user",\
            content: "Get the company description of JigsawStack from their linkedin page",\
        },\
    ],
});

console.log(completion.choices[0].message.content);
```

### OCR & Document Extraction

[vision docs ->](https://interfaze.ai/docs/vision)

```
prompt = "Get the person information from the following ID."

schema = z.object({
  first_name: z.string(),
  last_name: z.string(),
  dob: z.string(),
  expiry: z.string(),
});
```

![Extraction](https://interfaze.ai/examples/extraction_example.png)

### Smart Web Scraping

[web docs ->](https://interfaze.ai/docs/web)

```
prompt = "Extract the information from Yoeven D Khemlani's linkedin page based on the schema."

schema = z.object({
  first_name: z.string(),
  last_name: z.string(),
  about: z.string(),
  current_company: z.string(),
  current_position: z.string(),
});
```

![scraping](https://interfaze.ai/examples/scraper_example.png)

### Translation

[translation docs ->](https://interfaze.ai/docs/translation)

```
prompt = "The UK drinks about 100–160 million cups of tea every day, and 98% of tea drinkers add milk to their tea."

schema = z.object({
  zh: z.string(),
  hi: z.string(),
  es: z.string(),
  fr: z.string(),
  de: z.string(),
  it: z.string(),
  ja: z.string(),
  ko: z.string(),
});
```

```
zh: 英国每天饮用约100–160百万杯茶,有98%的茶饮者在茶中加入牛奶。
hi: यूके हर दिन लगभग 100–160 मिलियन कप चाय पीता है, और 98% चाय पीने वाले अपनी चाय में दूध मिलाते हैं।
es: El Reino Unido bebe alrededor de 100–160 millones de tazas de té cada día, y el 98 % de los consumidores de té añade leche a su té.
fr: Le Royaume-Uni boit environ 100–160 millions de tasses de thé chaque jour, et 98 % des buveurs de thé ajoutent du lait à leur thé.
de: Das Vereinigte Königreich trinkt etwa 100–160 Millionen Tassen Tee pro Tag, und 98 % der Teetrinker fügen ihrem Tee Milch hinzu.
it: Il Regno Unito beve circa 100–160 milioni di tazze di tè ogni giorno e il 98% degli amanti del tè aggiunge latte al proprio tè.
ja: イギリスでは毎日約100~160百万杯の紅茶が飲まれており、紅茶を飲む人の98%が紅茶に牛乳を加えます。
ko: 영국에서는 매일 약 1억 ~ 1억 6천만 잔의 차를 마시며, 차를 마시는 사람의 98%가 차에 우유를 넣습니다.
```

### Speech-to-text (STT) and diarization

[stt docs ->](https://interfaze.ai/docs/speech-to-text)

```
prompt = "Transcribe https://jigsawstack.com/preview/stt-example.wav"

schema = z.object({
  text: z.string(),
  speakers: z.object({
    id: z.string(),
    start: z.number(),
    end: z.number()
  })
});
```

```
{
  "text": " The little tales they tell are false The door was barred, locked and bolted as well Ripe pears are fit for a queen's table A big wet stain was on the round carpet The kite dipped and swayed but stayed aloft The pleasant hours fly by much too soon The room was crowded with a mild wob The room was crowded with a wild mob This strong arm shall shield your honour She blushed when he gave her a white orchid The beetle droned in the hot June sun",
  "speakers": [\
    {\
      "start":0,\
      "end":4.78,\
      "id": "SPEAKER_00"\
    },\
    {\
      "start":4.78,\
      "end":9.48,\
      "id": "SPEAKER_00"\
    },\
    {\
      "start":9.48,\
      "end":13.06,\
      "id": "SPEAKER_00"\
    },\
    {\
      "start":13.06,\
      "end":17.24,\
      "id": "SPEAKER_00"\
    },\
    {\
      "start":17.24,\
      "end":21.78,\
      "id": "SPEAKER_00"\
    },\
    {\
      "start":21.78,\
      "end":26.3,\
      "id": "SPEAKER_00"\
    },\
    {\
      "start":26.3,\
      "end":30.76,\
      "id": "SPEAKER_00"\
    },\
    {\
      "start":30.76,\
      "end":35.08,\
      "id": "SPEAKER_00"\
    },\
    {\
      "start":35.08,\
      "end":39.24,\
      "id": "SPEAKER_00"\
    },\
    {\
      "start":39.24,\
      "end":43.94,\
      "id": "SPEAKER_00"\
    },\
    {\
      "start":43.94,\
      "end":48.5,\
      "id": "SPEAKER_00"\
    }\
  ]
}
```

### Configurable guardrails and NSFW checks

[guardrails docs ->](https://interfaze.ai/docs/guard-rails)

Fully configurable guardrails for text and images

```
S1: Violent Crimes
S2: Non-Violent Crimes
S3: Sex-Related Crimes
S4: Child Sexual Exploitation
S5: Defamation
S6: Specialized Advice
S7: Privacy
S8: Intellectual Property
S9: Indiscriminate Weapons
S10: Hate
S11: Suicide & Self-Harm
S12: Sexual Content
S12_IMAGE: Sexual Content (Image)
S13: Elections
S14: Code Interpreter Abuse
```

### Architecture

[read paper ->](https://www.arxiv.org/abs/2602.04101)

This architecture combines a suite of small specialized models supported with custom tools and infrastructure while automatically routing to the best model for the task that prioritizes accuracy and speed.

![How it works](https://interfaze.ai/examples/howitworks.png)

### Specs

- Context window: 1m tokens
- Max output tokens: 32k tokens
- Input modalities: Text, Images, Audio, File, Video
- Reasoning: Available

### Research references

- [Interfaze: The Future of AI is built on Task-Specific Small Models](https://www.arxiv.org/abs/2602.04101)
- [Agentic Context Engineering](https://www.arxiv.org/pdf/2510.04618)
- [Small Language Models are the Future of Agentic AI](https://arxiv.org/pdf/2506.02153)
- [The Sparsely-Gated Mixture-of-Experts Layer](https://arxiv.org/pdf/1701.06538)
- [DeepSeekMoE](https://arxiv.org/pdf/2401.06066)
- [Confronting LLMs with Traditional ML](https://arxiv.org/pdf/2310.14607)

### Who are we?

We are a small team of ML, Software and Infrastructure engineers engrossed in the fact that a small model can do a lot more when specialized. Allowing us to make AI available in every dev workflow.