danielrosehill commited on
Commit
81bdde5
·
1 Parent(s): 82697cc
analysis/{context_windows_distribution.txt → context-windows/context_windows_distribution.txt} RENAMED
File without changes
analysis/{context_windows_stats.json → context-windows/context_windows_stats.json} RENAMED
File without changes
analysis/{context_windows_stratification.txt → context-windows/context_windows_stratification.txt} RENAMED
File without changes
analysis/{input_pricing_distribution.txt → pricing/input-tokens/input_pricing_distribution.txt} RENAMED
File without changes
analysis/{input_pricing_stats.json → pricing/input-tokens/input_pricing_stats.json} RENAMED
File without changes
analysis/{input_pricing_stratification.txt → pricing/input-tokens/input_pricing_stratification.txt} RENAMED
File without changes
analysis/{output_pricing_distribution.txt → pricing/output-tokens/output_pricing_distribution.txt} RENAMED
File without changes
analysis/{output_pricing_stats.json → pricing/output-tokens/output_pricing_stats.json} RENAMED
File without changes
analysis/{output_pricing_stratification.txt → pricing/output-tokens/output_pricing_stratification.txt} RENAMED
File without changes
analysis/{output_input_ratio_distribution.txt → pricing/token-ratios/output_input_ratio_distribution.txt} RENAMED
File without changes
analysis/{output_input_ratio_stats.json → pricing/token-ratios/output_input_ratio_stats.json} RENAMED
File without changes
analysis/{output_input_ratio_stratification.txt → pricing/token-ratios/output_input_ratio_stratification.txt} RENAMED
File without changes
analysis/pricing_analysis.py DELETED
@@ -1,259 +0,0 @@
1
- #!/usr/bin/env python3
2
- """
3
- OpenRouter API Pricing Analysis
4
- Analyzes context windows, token pricing, and output/input ratios
5
- """
6
-
7
- import json
8
- import numpy as np
9
- from pathlib import Path
10
- from typing import Dict, List, Any
11
- import statistics
12
-
13
- def load_data(file_path: str) -> List[Dict[str, Any]]:
14
- """Load the models data from JSON file"""
15
- with open(file_path, 'r') as f:
16
- return json.load(f)
17
-
18
- def calculate_distribution_stats(values: List[float], name: str) -> Dict[str, Any]:
19
- """Calculate comprehensive distribution statistics"""
20
- if not values:
21
- return {"error": "No data"}
22
-
23
- values = sorted(values)
24
- n = len(values)
25
-
26
- stats = {
27
- "name": name,
28
- "count": n,
29
- "min": min(values),
30
- "max": max(values),
31
- "mean": statistics.mean(values),
32
- "median": statistics.median(values),
33
- "std_dev": statistics.stdev(values) if n > 1 else 0,
34
- }
35
-
36
- # Quartiles
37
- stats["q1"] = np.percentile(values, 25)
38
- stats["q3"] = np.percentile(values, 75)
39
- stats["iqr"] = stats["q3"] - stats["q1"]
40
-
41
- # Additional percentiles
42
- stats["p10"] = np.percentile(values, 10)
43
- stats["p90"] = np.percentile(values, 90)
44
- stats["p95"] = np.percentile(values, 95)
45
- stats["p99"] = np.percentile(values, 99)
46
-
47
- return stats
48
-
49
- def calculate_stratification(values: List[float], name: str) -> Dict[str, Any]:
50
- """Calculate stratification by creating bins/ranges"""
51
- if not values:
52
- return {"error": "No data"}
53
-
54
- values_array = np.array(values)
55
-
56
- # Determine appropriate bins based on data range
57
- min_val = values_array.min()
58
- max_val = values_array.max()
59
-
60
- # Create stratification bins
61
- if max_val - min_val < 1:
62
- # For very small ranges (like some pricing)
63
- bins = np.linspace(min_val, max_val, 11)
64
- else:
65
- # Use log scale for wide ranges
66
- if min_val > 0:
67
- bins = np.logspace(np.log10(max(min_val, 0.0001)), np.log10(max_val), 11)
68
- else:
69
- bins = np.linspace(min_val, max_val, 11)
70
-
71
- hist, bin_edges = np.histogram(values_array, bins=bins)
72
-
73
- stratification = {
74
- "name": name,
75
- "bins": []
76
- }
77
-
78
- for i in range(len(hist)):
79
- stratification["bins"].append({
80
- "range": f"{bin_edges[i]:.6f} - {bin_edges[i+1]:.6f}",
81
- "count": int(hist[i]),
82
- "percentage": float(hist[i] / len(values) * 100)
83
- })
84
-
85
- return stratification
86
-
87
- def analyze_context_windows(models: List[Dict[str, Any]]) -> tuple:
88
- """Analyze context window distribution"""
89
- context_lengths = [
90
- m['context_length']
91
- for m in models
92
- if m.get('context_length') and m['context_length'] > 0
93
- ]
94
-
95
- distribution = calculate_distribution_stats(context_lengths, "Context Windows")
96
- stratification = calculate_stratification(context_lengths, "Context Windows")
97
-
98
- return distribution, stratification
99
-
100
- def analyze_input_pricing(models: List[Dict[str, Any]]) -> tuple:
101
- """Analyze input token pricing distribution"""
102
- input_prices = [
103
- m['prompt_price_per_1m_tokens']
104
- for m in models
105
- if m.get('prompt_price_per_1m_tokens') is not None and m['prompt_price_per_1m_tokens'] > 0
106
- ]
107
-
108
- distribution = calculate_distribution_stats(input_prices, "Input Token Pricing (per 1M tokens)")
109
- stratification = calculate_stratification(input_prices, "Input Token Pricing")
110
-
111
- return distribution, stratification
112
-
113
- def analyze_output_pricing(models: List[Dict[str, Any]]) -> tuple:
114
- """Analyze output token pricing distribution"""
115
- output_prices = [
116
- m['completion_price_per_1m_tokens']
117
- for m in models
118
- if m.get('completion_price_per_1m_tokens') is not None and m['completion_price_per_1m_tokens'] > 0
119
- ]
120
-
121
- distribution = calculate_distribution_stats(output_prices, "Output Token Pricing (per 1M tokens)")
122
- stratification = calculate_stratification(output_prices, "Output Token Pricing")
123
-
124
- return distribution, stratification
125
-
126
- def analyze_output_input_ratio(models: List[Dict[str, Any]]) -> tuple:
127
- """Analyze output/input token price ratio"""
128
- ratios = []
129
-
130
- for m in models:
131
- input_price = m.get('prompt_price_per_1m_tokens', 0)
132
- output_price = m.get('completion_price_per_1m_tokens', 0)
133
-
134
- # Calculate ratio only for non-free models
135
- if input_price > 0 and output_price > 0:
136
- ratio = output_price / input_price
137
- ratios.append(ratio)
138
-
139
- distribution = calculate_distribution_stats(ratios, "Output/Input Price Ratio")
140
- stratification = calculate_stratification(ratios, "Output/Input Price Ratio")
141
-
142
- return distribution, stratification
143
-
144
- def format_distribution_report(dist: Dict[str, Any]) -> str:
145
- """Format distribution statistics as readable text"""
146
- if "error" in dist:
147
- return f"Error: {dist['error']}"
148
-
149
- report = f"""
150
- {dist['name']}
151
- {'=' * len(dist['name'])}
152
-
153
- Sample Size: {dist['count']:,}
154
-
155
- Central Tendency:
156
- Mean: {dist['mean']:.6f}
157
- Median: {dist['median']:.6f}
158
-
159
- Spread:
160
- Min: {dist['min']:.6f}
161
- Max: {dist['max']:.6f}
162
- Std Dev: {dist['std_dev']:.6f}
163
- IQR: {dist['iqr']:.6f}
164
-
165
- Quartiles:
166
- Q1 (25th percentile): {dist['q1']:.6f}
167
- Q2 (50th percentile): {dist['median']:.6f}
168
- Q3 (75th percentile): {dist['q3']:.6f}
169
-
170
- Additional Percentiles:
171
- 10th: {dist['p10']:.6f}
172
- 90th: {dist['p90']:.6f}
173
- 95th: {dist['p95']:.6f}
174
- 99th: {dist['p99']:.6f}
175
- """
176
- return report
177
-
178
- def format_stratification_report(strat: Dict[str, Any]) -> str:
179
- """Format stratification as readable text"""
180
- if "error" in strat:
181
- return f"Error: {strat['error']}"
182
-
183
- report = f"""
184
- {strat['name']} - Stratification
185
- {'=' * (len(strat['name']) + 18)}
186
-
187
- Range Distribution:
188
- """
189
-
190
- for bin_data in strat['bins']:
191
- report += f" {bin_data['range']:>40s}: {bin_data['count']:>6,} models ({bin_data['percentage']:>5.1f}%)\n"
192
-
193
- return report
194
-
195
- def main():
196
- # Load data
197
- data_file = Path(__file__).parent.parent / 'raw' / 'models.json'
198
- models = load_data(data_file)
199
-
200
- print(f"Loaded {len(models)} models\n")
201
-
202
- # Create output directory
203
- output_dir = Path(__file__).parent
204
- output_dir.mkdir(exist_ok=True)
205
-
206
- # Analyze each metric
207
- analyses = {
208
- 'context_windows': analyze_context_windows(models),
209
- 'input_pricing': analyze_input_pricing(models),
210
- 'output_pricing': analyze_output_pricing(models),
211
- 'output_input_ratio': analyze_output_input_ratio(models)
212
- }
213
-
214
- # Generate reports for each analysis
215
- for metric_name, (distribution, stratification) in analyses.items():
216
- # Create individual report files
217
- dist_report = format_distribution_report(distribution)
218
- strat_report = format_stratification_report(stratification)
219
-
220
- # Write distribution report
221
- dist_file = output_dir / f"{metric_name}_distribution.txt"
222
- with open(dist_file, 'w') as f:
223
- f.write(dist_report)
224
- print(f"Created: {dist_file}")
225
-
226
- # Write stratification report
227
- strat_file = output_dir / f"{metric_name}_stratification.txt"
228
- with open(strat_file, 'w') as f:
229
- f.write(strat_report)
230
- print(f"Created: {strat_file}")
231
-
232
- # Also save as JSON for programmatic access
233
- json_file = output_dir / f"{metric_name}_stats.json"
234
- with open(json_file, 'w') as f:
235
- json.dump({
236
- 'distribution': distribution,
237
- 'stratification': stratification
238
- }, f, indent=2)
239
- print(f"Created: {json_file}")
240
-
241
- # Create comprehensive summary report
242
- summary_file = output_dir / "summary_report.txt"
243
- with open(summary_file, 'w') as f:
244
- f.write("OpenRouter API Pricing Analysis - Summary Report\n")
245
- f.write("=" * 60 + "\n\n")
246
- f.write(f"Dataset: {len(models)} models analyzed\n\n")
247
-
248
- for metric_name, (distribution, stratification) in analyses.items():
249
- f.write("\n" + "=" * 60 + "\n")
250
- f.write(format_distribution_report(distribution))
251
- f.write("\n")
252
- f.write(format_stratification_report(stratification))
253
- f.write("\n")
254
-
255
- print(f"\nCreated: {summary_file}")
256
- print("\nAnalysis complete!")
257
-
258
- if __name__ == "__main__":
259
- main()
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
analysis/vendors/vendor_detailed.txt ADDED
@@ -0,0 +1,1934 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ OpenRouter API - Detailed Vendor Analysis
2
+ ======================================================================
3
+
4
+
5
+ ======================================================================
6
+ Vendor: OpenAI
7
+ ======================================================================
8
+
9
+ Total Models: 47
10
+ Free Models: 1
11
+ Paid Models: 46
12
+
13
+ Input Pricing (per 1M tokens):
14
+ Models with pricing: 46
15
+ Min: $0.030000
16
+ Max: $150.000000
17
+ Mean: $7.915109
18
+ Median: $2.000000
19
+
20
+ Output Pricing (per 1M tokens):
21
+ Models with pricing: 46
22
+ Min: $0.140000
23
+ Max: $600.000000
24
+ Mean: $28.573478
25
+ Median: $8.000000
26
+
27
+ Context Lengths:
28
+ Models with context data: 47
29
+ Min: 4,095 tokens
30
+ Max: 1,047,576 tokens
31
+ Mean: 229,264 tokens
32
+ Median: 131,072 tokens
33
+
34
+ Models (showing first 10 of 47):
35
+ - OpenAI: gpt-oss-safeguard-20b
36
+ - OpenAI: GPT-5 Image Mini
37
+ - OpenAI: GPT-5 Image
38
+ - OpenAI: o3 Deep Research
39
+ - OpenAI: o4 Mini Deep Research
40
+ - OpenAI: GPT-5 Pro
41
+ - OpenAI: GPT-5 Codex
42
+ - OpenAI: GPT-4o Audio
43
+ - OpenAI: GPT-5 Chat
44
+ - OpenAI: GPT-5
45
+ ... and 37 more
46
+
47
+
48
+ ======================================================================
49
+ Vendor: Qwen
50
+ ======================================================================
51
+
52
+ Total Models: 46
53
+ Free Models: 8
54
+ Paid Models: 38
55
+
56
+ Input Pricing (per 1M tokens):
57
+ Models with pricing: 38
58
+ Min: $0.030000
59
+ Max: $1.600000
60
+ Mean: $0.257763
61
+ Median: $0.150000
62
+
63
+ Output Pricing (per 1M tokens):
64
+ Models with pricing: 38
65
+ Min: $0.090000
66
+ Max: $6.400000
67
+ Mean: $1.189421
68
+ Median: $0.575000
69
+
70
+ Context Lengths:
71
+ Models with context data: 46
72
+ Min: 7,500 tokens
73
+ Max: 1,000,000 tokens
74
+ Mean: 185,297 tokens
75
+ Median: 128,000 tokens
76
+
77
+ Models (showing first 10 of 46):
78
+ - Qwen: Qwen3 VL 8B Thinking
79
+ - Qwen: Qwen3 VL 8B Instruct
80
+ - Qwen: Qwen3 VL 30B A3B Thinking
81
+ - Qwen: Qwen3 VL 30B A3B Instruct
82
+ - Qwen: Qwen3 VL 235B A22B Thinking
83
+ - Qwen: Qwen3 VL 235B A22B Instruct
84
+ - Qwen: Qwen3 Max
85
+ - Qwen: Qwen3 Coder Plus
86
+ - Qwen: Qwen3 Coder Flash
87
+ - Qwen: Qwen3 Next 80B A3B Thinking
88
+ ... and 36 more
89
+
90
+
91
+ ======================================================================
92
+ Vendor: Mistral AI
93
+ ======================================================================
94
+
95
+ Total Models: 36
96
+ Free Models: 5
97
+ Paid Models: 31
98
+
99
+ Input Pricing (per 1M tokens):
100
+ Models with pricing: 31
101
+ Min: $0.020000
102
+ Max: $2.000000
103
+ Mean: $0.596968
104
+ Median: $0.200000
105
+
106
+ Output Pricing (per 1M tokens):
107
+ Models with pricing: 31
108
+ Min: $0.040000
109
+ Max: $6.000000
110
+ Mean: $1.717677
111
+ Median: $0.540000
112
+
113
+ Context Lengths:
114
+ Models with context data: 36
115
+ Min: 2,824 tokens
116
+ Max: 262,144 tokens
117
+ Mean: 91,245 tokens
118
+ Median: 112,000 tokens
119
+
120
+ Models (showing first 10 of 36):
121
+ - Mistral: Voxtral Small 24B 2507
122
+ - Mistral: Mistral Medium 3.1
123
+ - Mistral: Codestral 2508
124
+ - Mistral: Devstral Medium
125
+ - Mistral: Devstral Small 1.1
126
+ - Mistral: Mistral Small 3.2 24B (free)
127
+ - Mistral: Mistral Small 3.2 24B
128
+ - Mistral: Magistral Small 2506
129
+ - Mistral: Magistral Medium 2506 (thinking)
130
+ - Mistral: Magistral Medium 2506
131
+ ... and 26 more
132
+
133
+
134
+ ======================================================================
135
+ Vendor: Google
136
+ ======================================================================
137
+
138
+ Total Models: 24
139
+ Free Models: 6
140
+ Paid Models: 18
141
+
142
+ Input Pricing (per 1M tokens):
143
+ Models with pricing: 18
144
+ Min: $0.017030
145
+ Max: $1.250000
146
+ Mean: $0.347891
147
+ Median: $0.100000
148
+
149
+ Output Pricing (per 1M tokens):
150
+ Models with pricing: 18
151
+ Min: $0.040000
152
+ Max: $10.000000
153
+ Mean: $2.389342
154
+ Median: $0.400000
155
+
156
+ Context Lengths:
157
+ Models with context data: 24
158
+ Min: 8,192 tokens
159
+ Max: 1,048,576 tokens
160
+ Mean: 509,173 tokens
161
+ Median: 131,072 tokens
162
+
163
+ Models (showing first 10 of 24):
164
+ - Google: Gemini 2.5 Flash Image (Nano Banana)
165
+ - Google: Gemini 2.5 Flash Preview 09-2025
166
+ - Google: Gemini 2.5 Flash Lite Preview 09-2025
167
+ - Google: Gemini 2.5 Flash Image Preview (Nano Banana)
168
+ - Google: Gemini 2.5 Flash Lite
169
+ - Google: Gemma 3n 2B (free)
170
+ - Google: Gemini 2.5 Flash Lite Preview 06-17
171
+ - Google: Gemini 2.5 Flash
172
+ - Google: Gemini 2.5 Pro
173
+ - Google: Gemini 2.5 Pro Preview 06-05
174
+ ... and 14 more
175
+
176
+
177
+ ======================================================================
178
+ Vendor: Meta
179
+ ======================================================================
180
+
181
+ Total Models: 21
182
+ Free Models: 5
183
+ Paid Models: 16
184
+
185
+ Input Pricing (per 1M tokens):
186
+ Models with pricing: 16
187
+ Min: $0.005000
188
+ Max: $4.000000
189
+ Mean: $0.420875
190
+ Median: $0.140000
191
+
192
+ Output Pricing (per 1M tokens):
193
+ Models with pricing: 16
194
+ Min: $0.010000
195
+ Max: $4.000000
196
+ Mean: $0.493062
197
+ Median: $0.250000
198
+
199
+ Context Lengths:
200
+ Models with context data: 21
201
+ Min: 8,192 tokens
202
+ Max: 1,048,576 tokens
203
+ Mean: 153,649 tokens
204
+ Median: 131,072 tokens
205
+
206
+ Models (showing first 10 of 21):
207
+ - Meta: Llama 3.3 8B Instruct (free)
208
+ - Meta: Llama Guard 4 12B
209
+ - Meta: Llama 4 Maverick (free)
210
+ - Meta: Llama 4 Maverick
211
+ - Meta: Llama 4 Scout (free)
212
+ - Meta: Llama 4 Scout
213
+ - Llama Guard 3 8B
214
+ - Meta: Llama 3.3 70B Instruct (free)
215
+ - Meta: Llama 3.3 70B Instruct
216
+ - Meta: Llama 3.2 1B Instruct
217
+ ... and 11 more
218
+
219
+
220
+ ======================================================================
221
+ Vendor: DeepSeek
222
+ ======================================================================
223
+
224
+ Total Models: 19
225
+ Free Models: 6
226
+ Paid Models: 13
227
+
228
+ Input Pricing (per 1M tokens):
229
+ Models with pricing: 13
230
+ Min: $0.020000
231
+ Max: $0.500000
232
+ Mean: $0.244615
233
+ Median: $0.270000
234
+
235
+ Output Pricing (per 1M tokens):
236
+ Models with pricing: 13
237
+ Min: $0.100000
238
+ Max: $2.180000
239
+ Mean: $0.840000
240
+ Median: $0.840000
241
+
242
+ Context Lengths:
243
+ Models with context data: 19
244
+ Min: 8,192 tokens
245
+ Max: 163,840 tokens
246
+ Mean: 134,950 tokens
247
+ Median: 163,840 tokens
248
+
249
+ Models (showing first 10 of 19):
250
+ - DeepSeek: DeepSeek V3.2 Exp
251
+ - DeepSeek: DeepSeek V3.1 Terminus
252
+ - DeepSeek: DeepSeek V3.1 Terminus (exacto)
253
+ - DeepSeek: DeepSeek V3.1 (free)
254
+ - DeepSeek: DeepSeek V3.1
255
+ - DeepSeek: DeepSeek R1 0528 Qwen3 8B (free)
256
+ - DeepSeek: DeepSeek R1 0528 Qwen3 8B
257
+ - DeepSeek: R1 0528 (free)
258
+ - DeepSeek: R1 0528
259
+ - DeepSeek: DeepSeek Prover V2
260
+ ... and 9 more
261
+
262
+
263
+ ======================================================================
264
+ Vendor: Anthropic
265
+ ======================================================================
266
+
267
+ Total Models: 13
268
+ Free Models: 0
269
+ Paid Models: 13
270
+
271
+ Input Pricing (per 1M tokens):
272
+ Models with pricing: 13
273
+ Min: $0.250000
274
+ Max: $15.000000
275
+ Mean: $5.065385
276
+ Median: $3.000000
277
+
278
+ Output Pricing (per 1M tokens):
279
+ Models with pricing: 13
280
+ Min: $1.250000
281
+ Max: $75.000000
282
+ Mean: $25.326923
283
+ Median: $15.000000
284
+
285
+ Context Lengths:
286
+ Models with context data: 13
287
+ Min: 200,000 tokens
288
+ Max: 1,000,000 tokens
289
+ Mean: 323,077 tokens
290
+ Median: 200,000 tokens
291
+
292
+ Models (showing first 10 of 13):
293
+ - Anthropic: Claude Haiku 4.5
294
+ - Anthropic: Claude Sonnet 4.5
295
+ - Anthropic: Claude Opus 4.1
296
+ - Anthropic: Claude Opus 4
297
+ - Anthropic: Claude Sonnet 4
298
+ - Anthropic: Claude 3.7 Sonnet (thinking)
299
+ - Anthropic: Claude 3.7 Sonnet
300
+ - Anthropic: Claude 3.5 Haiku
301
+ - Anthropic: Claude 3.5 Haiku (2024-10-22)
302
+ - Anthropic: Claude 3.5 Sonnet
303
+ ... and 3 more
304
+
305
+
306
+ ======================================================================
307
+ Vendor: Microsoft
308
+ ======================================================================
309
+
310
+ Total Models: 9
311
+ Free Models: 1
312
+ Paid Models: 8
313
+
314
+ Input Pricing (per 1M tokens):
315
+ Models with pricing: 8
316
+ Min: $0.050000
317
+ Max: $1.000000
318
+ Mean: $0.270000
319
+ Median: $0.100000
320
+
321
+ Output Pricing (per 1M tokens):
322
+ Models with pricing: 8
323
+ Min: $0.100000
324
+ Max: $1.200000
325
+ Mean: $0.433750
326
+ Median: $0.245000
327
+
328
+ Context Lengths:
329
+ Models with context data: 9
330
+ Min: 16,384 tokens
331
+ Max: 163,840 tokens
332
+ Mean: 106,382 tokens
333
+ Median: 128,000 tokens
334
+
335
+ Models:
336
+ - Microsoft: Phi 4 Reasoning Plus
337
+ - Microsoft: MAI DS R1 (free)
338
+ - Microsoft: MAI DS R1
339
+ - Microsoft: Phi 4 Multimodal Instruct
340
+ - Microsoft: Phi 4
341
+ - Microsoft: Phi-3.5 Mini 128K Instruct
342
+ - Microsoft: Phi-3 Mini 128K Instruct
343
+ - Microsoft: Phi-3 Medium 128K Instruct
344
+ - WizardLM-2 8x22B
345
+
346
+
347
+ ======================================================================
348
+ Vendor: Moonshot AI
349
+ ======================================================================
350
+
351
+ Total Models: 7
352
+ Free Models: 1
353
+ Paid Models: 6
354
+
355
+ Input Pricing (per 1M tokens):
356
+ Models with pricing: 6
357
+ Min: $0.290000
358
+ Max: $0.600000
359
+ Mean: $0.446667
360
+ Median: $0.445000
361
+
362
+ Output Pricing (per 1M tokens):
363
+ Models with pricing: 6
364
+ Min: $0.600000
365
+ Max: $2.500000
366
+ Mean: $1.841667
367
+ Median: $2.150000
368
+
369
+ Context Lengths:
370
+ Models with context data: 7
371
+ Min: 32,768 tokens
372
+ Max: 1,048,576 tokens
373
+ Mean: 304,274 tokens
374
+ Median: 262,144 tokens
375
+
376
+ Models:
377
+ - MoonshotAI: Kimi Linear 48B A3B Instruct
378
+ - MoonshotAI: Kimi K2 Thinking
379
+ - MoonshotAI: Kimi K2 0905
380
+ - MoonshotAI: Kimi K2 0905 (exacto)
381
+ - MoonshotAI: Kimi K2 0711 (free)
382
+ - MoonshotAI: Kimi K2 0711
383
+ - MoonshotAI: Kimi Dev 72B
384
+
385
+
386
+ ======================================================================
387
+ Vendor: NVIDIA
388
+ ======================================================================
389
+
390
+ Total Models: 7
391
+ Free Models: 2
392
+ Paid Models: 5
393
+
394
+ Input Pricing (per 1M tokens):
395
+ Models with pricing: 5
396
+ Min: $0.040000
397
+ Max: $0.600000
398
+ Mean: $0.308000
399
+ Median: $0.200000
400
+
401
+ Output Pricing (per 1M tokens):
402
+ Models with pricing: 5
403
+ Min: $0.160000
404
+ Max: $1.800000
405
+ Mean: $0.712000
406
+ Median: $0.600000
407
+
408
+ Context Lengths:
409
+ Models with context data: 7
410
+ Min: 128,000 tokens
411
+ Max: 131,072 tokens
412
+ Mean: 130,194 tokens
413
+ Median: 131,072 tokens
414
+
415
+ Models:
416
+ - NVIDIA: Nemotron Nano 12B 2 VL (free)
417
+ - NVIDIA: Nemotron Nano 12B 2 VL
418
+ - NVIDIA: Llama 3.3 Nemotron Super 49B V1.5
419
+ - NVIDIA: Nemotron Nano 9B V2 (free)
420
+ - NVIDIA: Nemotron Nano 9B V2
421
+ - NVIDIA: Llama 3.1 Nemotron Ultra 253B v1
422
+ - NVIDIA: Llama 3.1 Nemotron 70B Instruct
423
+
424
+
425
+ ======================================================================
426
+ Vendor: Z-AI
427
+ ======================================================================
428
+
429
+ Total Models: 7
430
+ Free Models: 1
431
+ Paid Models: 6
432
+
433
+ Input Pricing (per 1M tokens):
434
+ Models with pricing: 6
435
+ Min: $0.100000
436
+ Max: $0.600000
437
+ Mean: $0.338333
438
+ Median: $0.375000
439
+
440
+ Output Pricing (per 1M tokens):
441
+ Models with pricing: 6
442
+ Min: $0.100000
443
+ Max: $1.900000
444
+ Mean: $1.325000
445
+ Median: $1.650000
446
+
447
+ Context Lengths:
448
+ Models with context data: 7
449
+ Min: 65,536 tokens
450
+ Max: 202,752 tokens
451
+ Mean: 141,751 tokens
452
+ Median: 131,072 tokens
453
+
454
+ Models:
455
+ - Z.AI: GLM 4.6
456
+ - Z.AI: GLM 4.6 (exacto)
457
+ - Z.AI: GLM 4.5V
458
+ - Z.AI: GLM 4.5
459
+ - Z.AI: GLM 4.5 Air (free)
460
+ - Z.AI: GLM 4.5 Air
461
+ - Z.AI: GLM 4 32B
462
+
463
+
464
+ ======================================================================
465
+ Vendor: xAI
466
+ ======================================================================
467
+
468
+ Total Models: 7
469
+ Free Models: 0
470
+ Paid Models: 7
471
+
472
+ Input Pricing (per 1M tokens):
473
+ Models with pricing: 7
474
+ Min: $0.200000
475
+ Max: $3.000000
476
+ Mean: $1.428571
477
+ Median: $0.300000
478
+
479
+ Output Pricing (per 1M tokens):
480
+ Models with pricing: 7
481
+ Min: $0.500000
482
+ Max: $15.000000
483
+ Mean: $6.857143
484
+ Median: $1.500000
485
+
486
+ Context Lengths:
487
+ Models with context data: 7
488
+ Min: 131,072 tokens
489
+ Max: 2,000,000 tokens
490
+ Mean: 433,755 tokens
491
+ Median: 131,072 tokens
492
+
493
+ Models:
494
+ - xAI: Grok 4 Fast
495
+ - xAI: Grok Code Fast 1
496
+ - xAI: Grok 4
497
+ - xAI: Grok 3 Mini
498
+ - xAI: Grok 3
499
+ - xAI: Grok 3 Mini Beta
500
+ - xAI: Grok 3 Beta
501
+
502
+
503
+ ======================================================================
504
+ Vendor: Nous Research
505
+ ======================================================================
506
+
507
+ Total Models: 7
508
+ Free Models: 1
509
+ Paid Models: 6
510
+
511
+ Input Pricing (per 1M tokens):
512
+ Models with pricing: 6
513
+ Min: $0.025000
514
+ Max: $1.000000
515
+ Mean: $0.314167
516
+ Median: $0.225000
517
+
518
+ Output Pricing (per 1M tokens):
519
+ Models with pricing: 6
520
+ Min: $0.080000
521
+ Max: $1.200000
522
+ Mean: $0.591667
523
+ Median: $0.485000
524
+
525
+ Context Lengths:
526
+ Models with context data: 7
527
+ Min: 32,768 tokens
528
+ Max: 131,072 tokens
529
+ Mean: 93,623 tokens
530
+ Median: 131,072 tokens
531
+
532
+ Models:
533
+ - Nous: Hermes 4 70B
534
+ - Nous: Hermes 4 405B
535
+ - Nous: DeepHermes 3 Mistral 24B Preview
536
+ - Nous: Hermes 3 70B Instruct
537
+ - Nous: Hermes 3 405B Instruct (free)
538
+ - Nous: Hermes 3 405B Instruct
539
+ - NousResearch: Hermes 2 Pro - Llama-3 8B
540
+
541
+
542
+ ======================================================================
543
+ Vendor: Perplexity
544
+ ======================================================================
545
+
546
+ Total Models: 6
547
+ Free Models: 0
548
+ Paid Models: 6
549
+
550
+ Input Pricing (per 1M tokens):
551
+ Models with pricing: 6
552
+ Min: $1.000000
553
+ Max: $3.000000
554
+ Mean: $2.000000
555
+ Median: $2.000000
556
+
557
+ Output Pricing (per 1M tokens):
558
+ Models with pricing: 6
559
+ Min: $1.000000
560
+ Max: $15.000000
561
+ Mean: $8.666667
562
+ Median: $8.000000
563
+
564
+ Context Lengths:
565
+ Models with context data: 6
566
+ Min: 127,000 tokens
567
+ Max: 200,000 tokens
568
+ Mean: 151,679 tokens
569
+ Median: 128,000 tokens
570
+
571
+ Models:
572
+ - Perplexity: Sonar Pro Search
573
+ - Perplexity: Sonar Reasoning Pro
574
+ - Perplexity: Sonar Pro
575
+ - Perplexity: Sonar Deep Research
576
+ - Perplexity: Sonar Reasoning
577
+ - Perplexity: Sonar
578
+
579
+
580
+ ======================================================================
581
+ Vendor: Baidu
582
+ ======================================================================
583
+
584
+ Total Models: 5
585
+ Free Models: 0
586
+ Paid Models: 5
587
+
588
+ Input Pricing (per 1M tokens):
589
+ Models with pricing: 5
590
+ Min: $0.070000
591
+ Max: $0.420000
592
+ Mean: $0.196000
593
+ Median: $0.140000
594
+
595
+ Output Pricing (per 1M tokens):
596
+ Models with pricing: 5
597
+ Min: $0.280000
598
+ Max: $1.250000
599
+ Mean: $0.694000
600
+ Median: $0.560000
601
+
602
+ Context Lengths:
603
+ Models with context data: 5
604
+ Min: 30,000 tokens
605
+ Max: 131,072 tokens
606
+ Mean: 105,414 tokens
607
+ Median: 123,000 tokens
608
+
609
+ Models:
610
+ - Baidu: ERNIE 4.5 21B A3B Thinking
611
+ - Baidu: ERNIE 4.5 21B A3B
612
+ - Baidu: ERNIE 4.5 VL 28B A3B
613
+ - Baidu: ERNIE 4.5 VL 424B A47B
614
+ - Baidu: ERNIE 4.5 300B A47B
615
+
616
+
617
+ ======================================================================
618
+ Vendor: TheDrummer
619
+ ======================================================================
620
+
621
+ Total Models: 5
622
+ Free Models: 0
623
+ Paid Models: 5
624
+
625
+ Input Pricing (per 1M tokens):
626
+ Models with pricing: 5
627
+ Min: $0.170000
628
+ Max: $0.650000
629
+ Mean: $0.404000
630
+ Median: $0.400000
631
+
632
+ Output Pricing (per 1M tokens):
633
+ Models with pricing: 5
634
+ Min: $0.400000
635
+ Max: $1.000000
636
+ Mean: $0.626000
637
+ Median: $0.500000
638
+
639
+ Context Lengths:
640
+ Models with context data: 5
641
+ Min: 32,768 tokens
642
+ Max: 131,072 tokens
643
+ Mean: 72,090 tokens
644
+ Median: 32,768 tokens
645
+
646
+ Models:
647
+ - TheDrummer: Cydonia 24B V4.1
648
+ - TheDrummer: Anubis 70B V1.1
649
+ - TheDrummer: Skyfall 36B V2
650
+ - TheDrummer: UnslopNemo 12B
651
+ - TheDrummer: Rocinante 12B
652
+
653
+
654
+ ======================================================================
655
+ Vendor: Arcee AI
656
+ ======================================================================
657
+
658
+ Total Models: 5
659
+ Free Models: 0
660
+ Paid Models: 5
661
+
662
+ Input Pricing (per 1M tokens):
663
+ Models with pricing: 5
664
+ Min: $0.048000
665
+ Max: $0.900000
666
+ Mean: $0.475600
667
+ Median: $0.500000
668
+
669
+ Output Pricing (per 1M tokens):
670
+ Models with pricing: 5
671
+ Min: $0.150000
672
+ Max: $3.300000
673
+ Mean: $1.126000
674
+ Median: $0.800000
675
+
676
+ Context Lengths:
677
+ Models with context data: 5
678
+ Min: 32,768 tokens
679
+ Max: 131,072 tokens
680
+ Mean: 98,304 tokens
681
+ Median: 131,072 tokens
682
+
683
+ Models:
684
+ - Arcee AI: AFM 4.5B
685
+ - Arcee AI: Spotlight
686
+ - Arcee AI: Maestro Reasoning
687
+ - Arcee AI: Virtuoso Large
688
+ - Arcee AI: Coder Large
689
+
690
+
691
+ ======================================================================
692
+ Vendor: Sao10k
693
+ ======================================================================
694
+
695
+ Total Models: 5
696
+ Free Models: 0
697
+ Paid Models: 5
698
+
699
+ Input Pricing (per 1M tokens):
700
+ Models with pricing: 5
701
+ Min: $0.040000
702
+ Max: $3.000000
703
+ Mean: $1.164000
704
+ Median: $0.650000
705
+
706
+ Output Pricing (per 1M tokens):
707
+ Models with pricing: 5
708
+ Min: $0.050000
709
+ Max: $3.000000
710
+ Mean: $1.206000
711
+ Median: $0.750000
712
+
713
+ Context Lengths:
714
+ Models with context data: 5
715
+ Min: 8,192 tokens
716
+ Max: 131,072 tokens
717
+ Mean: 39,245 tokens
718
+ Median: 16,000 tokens
719
+
720
+ Models:
721
+ - Sao10K: Llama 3.1 70B Hanami x1
722
+ - Sao10K: Llama 3.3 Euryale 70B
723
+ - Sao10K: Llama 3.1 Euryale 70B v2.2
724
+ - Sao10K: Llama 3 8B Lunaris
725
+ - Sao10k: Llama 3 Euryale 70B v2.1
726
+
727
+
728
+ ======================================================================
729
+ Vendor: Amazon
730
+ ======================================================================
731
+
732
+ Total Models: 4
733
+ Free Models: 0
734
+ Paid Models: 4
735
+
736
+ Input Pricing (per 1M tokens):
737
+ Models with pricing: 4
738
+ Min: $0.035000
739
+ Max: $2.500000
740
+ Mean: $0.848750
741
+ Median: $0.430000
742
+
743
+ Output Pricing (per 1M tokens):
744
+ Models with pricing: 4
745
+ Min: $0.140000
746
+ Max: $12.500000
747
+ Mean: $4.020000
748
+ Median: $1.720000
749
+
750
+ Context Lengths:
751
+ Models with context data: 4
752
+ Min: 128,000 tokens
753
+ Max: 1,000,000 tokens
754
+ Mean: 432,000 tokens
755
+ Median: 300,000 tokens
756
+
757
+ Models:
758
+ - Amazon: Nova Premier 1.0
759
+ - Amazon: Nova Lite 1.0
760
+ - Amazon: Nova Micro 1.0
761
+ - Amazon: Nova Pro 1.0
762
+
763
+
764
+ ======================================================================
765
+ Vendor: Minimax
766
+ ======================================================================
767
+
768
+ Total Models: 4
769
+ Free Models: 1
770
+ Paid Models: 3
771
+
772
+ Input Pricing (per 1M tokens):
773
+ Models with pricing: 3
774
+ Min: $0.200000
775
+ Max: $0.400000
776
+ Mean: $0.285000
777
+ Median: $0.255000
778
+
779
+ Output Pricing (per 1M tokens):
780
+ Models with pricing: 3
781
+ Min: $1.020000
782
+ Max: $2.200000
783
+ Mean: $1.440000
784
+ Median: $1.100000
785
+
786
+ Context Lengths:
787
+ Models with context data: 4
788
+ Min: 204,800 tokens
789
+ Max: 1,000,192 tokens
790
+ Mean: 602,448 tokens
791
+ Median: 602,400 tokens
792
+
793
+ Models:
794
+ - MiniMax: MiniMax M2 (free)
795
+ - MiniMax: MiniMax M2
796
+ - MiniMax: MiniMax M1
797
+ - MiniMax: MiniMax-01
798
+
799
+
800
+ ======================================================================
801
+ Vendor: DeepCogito
802
+ ======================================================================
803
+
804
+ Total Models: 4
805
+ Free Models: 0
806
+ Paid Models: 4
807
+
808
+ Input Pricing (per 1M tokens):
809
+ Models with pricing: 4
810
+ Min: $0.180000
811
+ Max: $3.500000
812
+ Mean: $1.452500
813
+ Median: $1.065000
814
+
815
+ Output Pricing (per 1M tokens):
816
+ Models with pricing: 4
817
+ Min: $0.590000
818
+ Max: $3.500000
819
+ Mean: $1.555000
820
+ Median: $1.065000
821
+
822
+ Context Lengths:
823
+ Models with context data: 4
824
+ Min: 32,767 tokens
825
+ Max: 163,840 tokens
826
+ Mean: 65,536 tokens
827
+ Median: 32,768 tokens
828
+
829
+ Models:
830
+ - Deep Cogito: Cogito V2 Preview Llama 405B
831
+ - Deep Cogito: Cogito V2 Preview Llama 70B
832
+ - Cogito V2 Preview Llama 109B
833
+ - Deep Cogito: Cogito V2 Preview Deepseek 671B
834
+
835
+
836
+ ======================================================================
837
+ Vendor: TNG Technology
838
+ ======================================================================
839
+
840
+ Total Models: 4
841
+ Free Models: 2
842
+ Paid Models: 2
843
+
844
+ Input Pricing (per 1M tokens):
845
+ Models with pricing: 2
846
+ Min: $0.300000
847
+ Max: $0.300000
848
+ Mean: $0.300000
849
+ Median: $0.300000
850
+
851
+ Output Pricing (per 1M tokens):
852
+ Models with pricing: 2
853
+ Min: $1.200000
854
+ Max: $1.200000
855
+ Mean: $1.200000
856
+ Median: $1.200000
857
+
858
+ Context Lengths:
859
+ Models with context data: 4
860
+ Min: 163,840 tokens
861
+ Max: 163,840 tokens
862
+ Mean: 163,840 tokens
863
+ Median: 163,840 tokens
864
+
865
+ Models:
866
+ - TNG: DeepSeek R1T2 Chimera (free)
867
+ - TNG: DeepSeek R1T2 Chimera
868
+ - TNG: DeepSeek R1T Chimera (free)
869
+ - TNG: DeepSeek R1T Chimera
870
+
871
+
872
+ ======================================================================
873
+ Vendor: Cohere
874
+ ======================================================================
875
+
876
+ Total Models: 4
877
+ Free Models: 0
878
+ Paid Models: 4
879
+
880
+ Input Pricing (per 1M tokens):
881
+ Models with pricing: 4
882
+ Min: $0.037500
883
+ Max: $2.500000
884
+ Mean: $1.296875
885
+ Median: $1.325000
886
+
887
+ Output Pricing (per 1M tokens):
888
+ Models with pricing: 4
889
+ Min: $0.150000
890
+ Max: $10.000000
891
+ Mean: $5.187500
892
+ Median: $5.300000
893
+
894
+ Context Lengths:
895
+ Models with context data: 4
896
+ Min: 128,000 tokens
897
+ Max: 256,000 tokens
898
+ Mean: 160,000 tokens
899
+ Median: 128,000 tokens
900
+
901
+ Models:
902
+ - Cohere: Command A
903
+ - Cohere: Command R7B (12-2024)
904
+ - Cohere: Command R+ (08-2024)
905
+ - Cohere: Command R (08-2024)
906
+
907
+
908
+ ======================================================================
909
+ Vendor: Aion Labs
910
+ ======================================================================
911
+
912
+ Total Models: 3
913
+ Free Models: 0
914
+ Paid Models: 3
915
+
916
+ Input Pricing (per 1M tokens):
917
+ Models with pricing: 3
918
+ Min: $0.200000
919
+ Max: $4.000000
920
+ Mean: $1.633333
921
+ Median: $0.700000
922
+
923
+ Output Pricing (per 1M tokens):
924
+ Models with pricing: 3
925
+ Min: $0.200000
926
+ Max: $8.000000
927
+ Mean: $3.200000
928
+ Median: $1.400000
929
+
930
+ Context Lengths:
931
+ Models with context data: 3
932
+ Min: 32,768 tokens
933
+ Max: 131,072 tokens
934
+ Mean: 98,304 tokens
935
+ Median: 131,072 tokens
936
+
937
+ Models:
938
+ - AionLabs: Aion-1.0
939
+ - AionLabs: Aion-1.0-Mini
940
+ - AionLabs: Aion-RP 1.0 (8B)
941
+
942
+
943
+ ======================================================================
944
+ Vendor: OpenRouter
945
+ ======================================================================
946
+
947
+ Total Models: 2
948
+ Free Models: 1
949
+ Paid Models: 1
950
+
951
+ Context Lengths:
952
+ Models with context data: 2
953
+ Min: 256,000 tokens
954
+ Max: 2,000,000 tokens
955
+ Mean: 1,128,000 tokens
956
+ Median: 1,128,000 tokens
957
+
958
+ Models:
959
+ - Polaris Alpha
960
+ - Auto Router
961
+
962
+
963
+ ======================================================================
964
+ Vendor: Liquid AI
965
+ ======================================================================
966
+
967
+ Total Models: 2
968
+ Free Models: 0
969
+ Paid Models: 2
970
+
971
+ Input Pricing (per 1M tokens):
972
+ Models with pricing: 2
973
+ Min: $0.050000
974
+ Max: $0.050000
975
+ Mean: $0.050000
976
+ Median: $0.050000
977
+
978
+ Output Pricing (per 1M tokens):
979
+ Models with pricing: 2
980
+ Min: $0.100000
981
+ Max: $0.100000
982
+ Mean: $0.100000
983
+ Median: $0.100000
984
+
985
+ Context Lengths:
986
+ Models with context data: 2
987
+ Min: 32,768 tokens
988
+ Max: 32,768 tokens
989
+ Mean: 32,768 tokens
990
+ Median: 32,768 tokens
991
+
992
+ Models:
993
+ - LiquidAI/LFM2-8B-A1B
994
+ - LiquidAI/LFM2-2.6B
995
+
996
+
997
+ ======================================================================
998
+ Vendor: Inclusion AI
999
+ ======================================================================
1000
+
1001
+ Total Models: 2
1002
+ Free Models: 0
1003
+ Paid Models: 2
1004
+
1005
+ Input Pricing (per 1M tokens):
1006
+ Models with pricing: 2
1007
+ Min: $0.570000
1008
+ Max: $0.570000
1009
+ Mean: $0.570000
1010
+ Median: $0.570000
1011
+
1012
+ Output Pricing (per 1M tokens):
1013
+ Models with pricing: 2
1014
+ Min: $2.280000
1015
+ Max: $2.280000
1016
+ Mean: $2.280000
1017
+ Median: $2.280000
1018
+
1019
+ Context Lengths:
1020
+ Models with context data: 2
1021
+ Min: 131,072 tokens
1022
+ Max: 131,072 tokens
1023
+ Mean: 131,072 tokens
1024
+ Median: 131,072 tokens
1025
+
1026
+ Models:
1027
+ - inclusionAI: Ring 1T
1028
+ - inclusionAI: Ling-1T
1029
+
1030
+
1031
+ ======================================================================
1032
+ Vendor: Alibaba
1033
+ ======================================================================
1034
+
1035
+ Total Models: 2
1036
+ Free Models: 1
1037
+ Paid Models: 1
1038
+
1039
+ Input Pricing (per 1M tokens):
1040
+ Models with pricing: 1
1041
+ Min: $0.090000
1042
+ Max: $0.090000
1043
+ Mean: $0.090000
1044
+ Median: $0.090000
1045
+
1046
+ Output Pricing (per 1M tokens):
1047
+ Models with pricing: 1
1048
+ Min: $0.400000
1049
+ Max: $0.400000
1050
+ Mean: $0.400000
1051
+ Median: $0.400000
1052
+
1053
+ Context Lengths:
1054
+ Models with context data: 2
1055
+ Min: 131,072 tokens
1056
+ Max: 131,072 tokens
1057
+ Mean: 131,072 tokens
1058
+ Median: 131,072 tokens
1059
+
1060
+ Models:
1061
+ - Tongyi DeepResearch 30B A3B (free)
1062
+ - Tongyi DeepResearch 30B A3B
1063
+
1064
+
1065
+ ======================================================================
1066
+ Vendor: Meituan
1067
+ ======================================================================
1068
+
1069
+ Total Models: 2
1070
+ Free Models: 1
1071
+ Paid Models: 1
1072
+
1073
+ Input Pricing (per 1M tokens):
1074
+ Models with pricing: 1
1075
+ Min: $0.150000
1076
+ Max: $0.150000
1077
+ Mean: $0.150000
1078
+ Median: $0.150000
1079
+
1080
+ Output Pricing (per 1M tokens):
1081
+ Models with pricing: 1
1082
+ Min: $0.750000
1083
+ Max: $0.750000
1084
+ Mean: $0.750000
1085
+ Median: $0.750000
1086
+
1087
+ Context Lengths:
1088
+ Models with context data: 2
1089
+ Min: 131,072 tokens
1090
+ Max: 131,072 tokens
1091
+ Mean: 131,072 tokens
1092
+ Median: 131,072 tokens
1093
+
1094
+ Models:
1095
+ - Meituan: LongCat Flash Chat (free)
1096
+ - Meituan: LongCat Flash Chat
1097
+
1098
+
1099
+ ======================================================================
1100
+ Vendor: AI21 Labs
1101
+ ======================================================================
1102
+
1103
+ Total Models: 2
1104
+ Free Models: 0
1105
+ Paid Models: 2
1106
+
1107
+ Input Pricing (per 1M tokens):
1108
+ Models with pricing: 2
1109
+ Min: $0.200000
1110
+ Max: $2.000000
1111
+ Mean: $1.100000
1112
+ Median: $1.100000
1113
+
1114
+ Output Pricing (per 1M tokens):
1115
+ Models with pricing: 2
1116
+ Min: $0.400000
1117
+ Max: $8.000000
1118
+ Mean: $4.200000
1119
+ Median: $4.200000
1120
+
1121
+ Context Lengths:
1122
+ Models with context data: 2
1123
+ Min: 256,000 tokens
1124
+ Max: 256,000 tokens
1125
+ Mean: 256,000 tokens
1126
+ Median: 256,000 tokens
1127
+
1128
+ Models:
1129
+ - AI21: Jamba Mini 1.7
1130
+ - AI21: Jamba Large 1.7
1131
+
1132
+
1133
+ ======================================================================
1134
+ Vendor: Morph
1135
+ ======================================================================
1136
+
1137
+ Total Models: 2
1138
+ Free Models: 0
1139
+ Paid Models: 2
1140
+
1141
+ Input Pricing (per 1M tokens):
1142
+ Models with pricing: 2
1143
+ Min: $0.800000
1144
+ Max: $0.900000
1145
+ Mean: $0.850000
1146
+ Median: $0.850000
1147
+
1148
+ Output Pricing (per 1M tokens):
1149
+ Models with pricing: 2
1150
+ Min: $1.200000
1151
+ Max: $1.900000
1152
+ Mean: $1.550000
1153
+ Median: $1.550000
1154
+
1155
+ Context Lengths:
1156
+ Models with context data: 2
1157
+ Min: 81,920 tokens
1158
+ Max: 262,144 tokens
1159
+ Mean: 172,032 tokens
1160
+ Median: 172,032 tokens
1161
+
1162
+ Models:
1163
+ - Morph: Morph V3 Large
1164
+ - Morph: Morph V3 Fast
1165
+
1166
+
1167
+ ======================================================================
1168
+ Vendor: Inception
1169
+ ======================================================================
1170
+
1171
+ Total Models: 2
1172
+ Free Models: 0
1173
+ Paid Models: 2
1174
+
1175
+ Input Pricing (per 1M tokens):
1176
+ Models with pricing: 2
1177
+ Min: $0.250000
1178
+ Max: $0.250000
1179
+ Mean: $0.250000
1180
+ Median: $0.250000
1181
+
1182
+ Output Pricing (per 1M tokens):
1183
+ Models with pricing: 2
1184
+ Min: $1.000000
1185
+ Max: $1.000000
1186
+ Mean: $1.000000
1187
+ Median: $1.000000
1188
+
1189
+ Context Lengths:
1190
+ Models with context data: 2
1191
+ Min: 128,000 tokens
1192
+ Max: 128,000 tokens
1193
+ Mean: 128,000 tokens
1194
+ Median: 128,000 tokens
1195
+
1196
+ Models:
1197
+ - Inception: Mercury
1198
+ - Inception: Mercury Coder
1199
+
1200
+
1201
+ ======================================================================
1202
+ Vendor: Arli AI
1203
+ ======================================================================
1204
+
1205
+ Total Models: 2
1206
+ Free Models: 1
1207
+ Paid Models: 1
1208
+
1209
+ Input Pricing (per 1M tokens):
1210
+ Models with pricing: 1
1211
+ Min: $0.030000
1212
+ Max: $0.030000
1213
+ Mean: $0.030000
1214
+ Median: $0.030000
1215
+
1216
+ Output Pricing (per 1M tokens):
1217
+ Models with pricing: 1
1218
+ Min: $0.110000
1219
+ Max: $0.110000
1220
+ Mean: $0.110000
1221
+ Median: $0.110000
1222
+
1223
+ Context Lengths:
1224
+ Models with context data: 2
1225
+ Min: 32,768 tokens
1226
+ Max: 32,768 tokens
1227
+ Mean: 32,768 tokens
1228
+ Median: 32,768 tokens
1229
+
1230
+ Models:
1231
+ - ArliAI: QwQ 32B RpR v1 (free)
1232
+ - ArliAI: QwQ 32B RpR v1
1233
+
1234
+
1235
+ ======================================================================
1236
+ Vendor: Agentica
1237
+ ======================================================================
1238
+
1239
+ Total Models: 2
1240
+ Free Models: 1
1241
+ Paid Models: 1
1242
+
1243
+ Input Pricing (per 1M tokens):
1244
+ Models with pricing: 1
1245
+ Min: $0.015000
1246
+ Max: $0.015000
1247
+ Mean: $0.015000
1248
+ Median: $0.015000
1249
+
1250
+ Output Pricing (per 1M tokens):
1251
+ Models with pricing: 1
1252
+ Min: $0.015000
1253
+ Max: $0.015000
1254
+ Mean: $0.015000
1255
+ Median: $0.015000
1256
+
1257
+ Context Lengths:
1258
+ Models with context data: 2
1259
+ Min: 96,000 tokens
1260
+ Max: 96,000 tokens
1261
+ Mean: 96,000 tokens
1262
+ Median: 96,000 tokens
1263
+
1264
+ Models:
1265
+ - Agentica: Deepcoder 14B Preview (free)
1266
+ - Agentica: Deepcoder 14B Preview
1267
+
1268
+
1269
+ ======================================================================
1270
+ Vendor: Inflection AI
1271
+ ======================================================================
1272
+
1273
+ Total Models: 2
1274
+ Free Models: 0
1275
+ Paid Models: 2
1276
+
1277
+ Input Pricing (per 1M tokens):
1278
+ Models with pricing: 2
1279
+ Min: $2.500000
1280
+ Max: $2.500000
1281
+ Mean: $2.500000
1282
+ Median: $2.500000
1283
+
1284
+ Output Pricing (per 1M tokens):
1285
+ Models with pricing: 2
1286
+ Min: $10.000000
1287
+ Max: $10.000000
1288
+ Mean: $10.000000
1289
+ Median: $10.000000
1290
+
1291
+ Context Lengths:
1292
+ Models with context data: 2
1293
+ Min: 8,000 tokens
1294
+ Max: 8,000 tokens
1295
+ Mean: 8,000 tokens
1296
+ Median: 8,000 tokens
1297
+
1298
+ Models:
1299
+ - Inflection: Inflection 3 Productivity
1300
+ - Inflection: Inflection 3 Pi
1301
+
1302
+
1303
+ ======================================================================
1304
+ Vendor: Neversleep
1305
+ ======================================================================
1306
+
1307
+ Total Models: 2
1308
+ Free Models: 0
1309
+ Paid Models: 2
1310
+
1311
+ Input Pricing (per 1M tokens):
1312
+ Models with pricing: 2
1313
+ Min: $0.090000
1314
+ Max: $1.000000
1315
+ Mean: $0.545000
1316
+ Median: $0.545000
1317
+
1318
+ Output Pricing (per 1M tokens):
1319
+ Models with pricing: 2
1320
+ Min: $0.600000
1321
+ Max: $1.750000
1322
+ Mean: $1.175000
1323
+ Median: $1.175000
1324
+
1325
+ Context Lengths:
1326
+ Models with context data: 2
1327
+ Min: 4,096 tokens
1328
+ Max: 32,768 tokens
1329
+ Mean: 18,432 tokens
1330
+ Median: 18,432 tokens
1331
+
1332
+ Models:
1333
+ - NeverSleep: Lumimaid v0.2 8B
1334
+ - Noromaid 20B
1335
+
1336
+
1337
+ ======================================================================
1338
+ Vendor: Kwai
1339
+ ======================================================================
1340
+
1341
+ Total Models: 1
1342
+ Free Models: 1
1343
+ Paid Models: 0
1344
+
1345
+ Context Lengths:
1346
+ Models with context data: 1
1347
+ Min: 256,000 tokens
1348
+ Max: 256,000 tokens
1349
+ Mean: 256,000 tokens
1350
+ Median: 256,000 tokens
1351
+
1352
+ Models:
1353
+ - Kwaipilot: Kat Coder (free)
1354
+
1355
+
1356
+ ======================================================================
1357
+ Vendor: IBM
1358
+ ======================================================================
1359
+
1360
+ Total Models: 1
1361
+ Free Models: 0
1362
+ Paid Models: 1
1363
+
1364
+ Input Pricing (per 1M tokens):
1365
+ Models with pricing: 1
1366
+ Min: $0.017000
1367
+ Max: $0.017000
1368
+ Mean: $0.017000
1369
+ Median: $0.017000
1370
+
1371
+ Output Pricing (per 1M tokens):
1372
+ Models with pricing: 1
1373
+ Min: $0.110000
1374
+ Max: $0.110000
1375
+ Mean: $0.110000
1376
+ Median: $0.110000
1377
+
1378
+ Context Lengths:
1379
+ Models with context data: 1
1380
+ Min: 131,000 tokens
1381
+ Max: 131,000 tokens
1382
+ Mean: 131,000 tokens
1383
+ Median: 131,000 tokens
1384
+
1385
+ Models:
1386
+ - IBM: Granite 4.0 Micro
1387
+
1388
+
1389
+ ======================================================================
1390
+ Vendor: Relace
1391
+ ======================================================================
1392
+
1393
+ Total Models: 1
1394
+ Free Models: 0
1395
+ Paid Models: 1
1396
+
1397
+ Input Pricing (per 1M tokens):
1398
+ Models with pricing: 1
1399
+ Min: $0.850000
1400
+ Max: $0.850000
1401
+ Mean: $0.850000
1402
+ Median: $0.850000
1403
+
1404
+ Output Pricing (per 1M tokens):
1405
+ Models with pricing: 1
1406
+ Min: $1.250000
1407
+ Max: $1.250000
1408
+ Mean: $1.250000
1409
+ Median: $1.250000
1410
+
1411
+ Context Lengths:
1412
+ Models with context data: 1
1413
+ Min: 256,000 tokens
1414
+ Max: 256,000 tokens
1415
+ Mean: 256,000 tokens
1416
+ Median: 256,000 tokens
1417
+
1418
+ Models:
1419
+ - Relace: Relace Apply 3
1420
+
1421
+
1422
+ ======================================================================
1423
+ Vendor: OpenGVLab
1424
+ ======================================================================
1425
+
1426
+ Total Models: 1
1427
+ Free Models: 0
1428
+ Paid Models: 1
1429
+
1430
+ Input Pricing (per 1M tokens):
1431
+ Models with pricing: 1
1432
+ Min: $0.070000
1433
+ Max: $0.070000
1434
+ Mean: $0.070000
1435
+ Median: $0.070000
1436
+
1437
+ Output Pricing (per 1M tokens):
1438
+ Models with pricing: 1
1439
+ Min: $0.260000
1440
+ Max: $0.260000
1441
+ Mean: $0.260000
1442
+ Median: $0.260000
1443
+
1444
+ Context Lengths:
1445
+ Models with context data: 1
1446
+ Min: 32,768 tokens
1447
+ Max: 32,768 tokens
1448
+ Mean: 32,768 tokens
1449
+ Median: 32,768 tokens
1450
+
1451
+ Models:
1452
+ - OpenGVLab: InternVL3 78B
1453
+
1454
+
1455
+ ======================================================================
1456
+ Vendor: Stepfun AI
1457
+ ======================================================================
1458
+
1459
+ Total Models: 1
1460
+ Free Models: 0
1461
+ Paid Models: 1
1462
+
1463
+ Input Pricing (per 1M tokens):
1464
+ Models with pricing: 1
1465
+ Min: $0.570000
1466
+ Max: $0.570000
1467
+ Mean: $0.570000
1468
+ Median: $0.570000
1469
+
1470
+ Output Pricing (per 1M tokens):
1471
+ Models with pricing: 1
1472
+ Min: $1.420000
1473
+ Max: $1.420000
1474
+ Mean: $1.420000
1475
+ Median: $1.420000
1476
+
1477
+ Context Lengths:
1478
+ Models with context data: 1
1479
+ Min: 65,536 tokens
1480
+ Max: 65,536 tokens
1481
+ Mean: 65,536 tokens
1482
+ Median: 65,536 tokens
1483
+
1484
+ Models:
1485
+ - StepFun: Step3
1486
+
1487
+
1488
+ ======================================================================
1489
+ Vendor: ByteDance
1490
+ ======================================================================
1491
+
1492
+ Total Models: 1
1493
+ Free Models: 0
1494
+ Paid Models: 1
1495
+
1496
+ Input Pricing (per 1M tokens):
1497
+ Models with pricing: 1
1498
+ Min: $0.100000
1499
+ Max: $0.100000
1500
+ Mean: $0.100000
1501
+ Median: $0.100000
1502
+
1503
+ Output Pricing (per 1M tokens):
1504
+ Models with pricing: 1
1505
+ Min: $0.200000
1506
+ Max: $0.200000
1507
+ Mean: $0.200000
1508
+ Median: $0.200000
1509
+
1510
+ Context Lengths:
1511
+ Models with context data: 1
1512
+ Min: 128,000 tokens
1513
+ Max: 128,000 tokens
1514
+ Mean: 128,000 tokens
1515
+ Median: 128,000 tokens
1516
+
1517
+ Models:
1518
+ - ByteDance: UI-TARS 7B
1519
+
1520
+
1521
+ ======================================================================
1522
+ Vendor: Switchpoint
1523
+ ======================================================================
1524
+
1525
+ Total Models: 1
1526
+ Free Models: 0
1527
+ Paid Models: 1
1528
+
1529
+ Input Pricing (per 1M tokens):
1530
+ Models with pricing: 1
1531
+ Min: $0.850000
1532
+ Max: $0.850000
1533
+ Mean: $0.850000
1534
+ Median: $0.850000
1535
+
1536
+ Output Pricing (per 1M tokens):
1537
+ Models with pricing: 1
1538
+ Min: $3.400000
1539
+ Max: $3.400000
1540
+ Mean: $3.400000
1541
+ Median: $3.400000
1542
+
1543
+ Context Lengths:
1544
+ Models with context data: 1
1545
+ Min: 131,072 tokens
1546
+ Max: 131,072 tokens
1547
+ Mean: 131,072 tokens
1548
+ Median: 131,072 tokens
1549
+
1550
+ Models:
1551
+ - Switchpoint Router
1552
+
1553
+
1554
+ ======================================================================
1555
+ Vendor: THUDM
1556
+ ======================================================================
1557
+
1558
+ Total Models: 1
1559
+ Free Models: 0
1560
+ Paid Models: 1
1561
+
1562
+ Input Pricing (per 1M tokens):
1563
+ Models with pricing: 1
1564
+ Min: $0.035000
1565
+ Max: $0.035000
1566
+ Mean: $0.035000
1567
+ Median: $0.035000
1568
+
1569
+ Output Pricing (per 1M tokens):
1570
+ Models with pricing: 1
1571
+ Min: $0.138000
1572
+ Max: $0.138000
1573
+ Mean: $0.138000
1574
+ Median: $0.138000
1575
+
1576
+ Context Lengths:
1577
+ Models with context data: 1
1578
+ Min: 65,536 tokens
1579
+ Max: 65,536 tokens
1580
+ Mean: 65,536 tokens
1581
+ Median: 65,536 tokens
1582
+
1583
+ Models:
1584
+ - THUDM: GLM 4.1V 9B Thinking
1585
+
1586
+
1587
+ ======================================================================
1588
+ Vendor: Cognitive Computations
1589
+ ======================================================================
1590
+
1591
+ Total Models: 1
1592
+ Free Models: 1
1593
+ Paid Models: 0
1594
+
1595
+ Context Lengths:
1596
+ Models with context data: 1
1597
+ Min: 32,768 tokens
1598
+ Max: 32,768 tokens
1599
+ Mean: 32,768 tokens
1600
+ Median: 32,768 tokens
1601
+
1602
+ Models:
1603
+ - Venice: Uncensored (free)
1604
+
1605
+
1606
+ ======================================================================
1607
+ Vendor: Tencent
1608
+ ======================================================================
1609
+
1610
+ Total Models: 1
1611
+ Free Models: 0
1612
+ Paid Models: 1
1613
+
1614
+ Input Pricing (per 1M tokens):
1615
+ Models with pricing: 1
1616
+ Min: $0.140000
1617
+ Max: $0.140000
1618
+ Mean: $0.140000
1619
+ Median: $0.140000
1620
+
1621
+ Output Pricing (per 1M tokens):
1622
+ Models with pricing: 1
1623
+ Min: $0.570000
1624
+ Max: $0.570000
1625
+ Mean: $0.570000
1626
+ Median: $0.570000
1627
+
1628
+ Context Lengths:
1629
+ Models with context data: 1
1630
+ Min: 131,072 tokens
1631
+ Max: 131,072 tokens
1632
+ Mean: 131,072 tokens
1633
+ Median: 131,072 tokens
1634
+
1635
+ Models:
1636
+ - Tencent: Hunyuan A13B Instruct
1637
+
1638
+
1639
+ ======================================================================
1640
+ Vendor: EleutherAI
1641
+ ======================================================================
1642
+
1643
+ Total Models: 1
1644
+ Free Models: 0
1645
+ Paid Models: 1
1646
+
1647
+ Input Pricing (per 1M tokens):
1648
+ Models with pricing: 1
1649
+ Min: $0.800000
1650
+ Max: $0.800000
1651
+ Mean: $0.800000
1652
+ Median: $0.800000
1653
+
1654
+ Output Pricing (per 1M tokens):
1655
+ Models with pricing: 1
1656
+ Min: $1.200000
1657
+ Max: $1.200000
1658
+ Mean: $1.200000
1659
+ Median: $1.200000
1660
+
1661
+ Context Lengths:
1662
+ Models with context data: 1
1663
+ Min: 4,096 tokens
1664
+ Max: 4,096 tokens
1665
+ Mean: 4,096 tokens
1666
+ Median: 4,096 tokens
1667
+
1668
+ Models:
1669
+ - EleutherAI: Llemma 7b
1670
+
1671
+
1672
+ ======================================================================
1673
+ Vendor: AlfredPros
1674
+ ======================================================================
1675
+
1676
+ Total Models: 1
1677
+ Free Models: 0
1678
+ Paid Models: 1
1679
+
1680
+ Input Pricing (per 1M tokens):
1681
+ Models with pricing: 1
1682
+ Min: $0.800000
1683
+ Max: $0.800000
1684
+ Mean: $0.800000
1685
+ Median: $0.800000
1686
+
1687
+ Output Pricing (per 1M tokens):
1688
+ Models with pricing: 1
1689
+ Min: $1.200000
1690
+ Max: $1.200000
1691
+ Mean: $1.200000
1692
+ Median: $1.200000
1693
+
1694
+ Context Lengths:
1695
+ Models with context data: 1
1696
+ Min: 4,096 tokens
1697
+ Max: 4,096 tokens
1698
+ Mean: 4,096 tokens
1699
+ Median: 4,096 tokens
1700
+
1701
+ Models:
1702
+ - AlfredPros: CodeLLaMa 7B Instruct Solidity
1703
+
1704
+
1705
+ ======================================================================
1706
+ Vendor: Allen Institute for AI
1707
+ ======================================================================
1708
+
1709
+ Total Models: 1
1710
+ Free Models: 0
1711
+ Paid Models: 1
1712
+
1713
+ Input Pricing (per 1M tokens):
1714
+ Models with pricing: 1
1715
+ Min: $0.200000
1716
+ Max: $0.200000
1717
+ Mean: $0.200000
1718
+ Median: $0.200000
1719
+
1720
+ Output Pricing (per 1M tokens):
1721
+ Models with pricing: 1
1722
+ Min: $0.350000
1723
+ Max: $0.350000
1724
+ Mean: $0.350000
1725
+ Median: $0.350000
1726
+
1727
+ Context Lengths:
1728
+ Models with context data: 1
1729
+ Min: 4,096 tokens
1730
+ Max: 4,096 tokens
1731
+ Mean: 4,096 tokens
1732
+ Median: 4,096 tokens
1733
+
1734
+ Models:
1735
+ - AllenAI: Olmo 2 32B Instruct
1736
+
1737
+
1738
+ ======================================================================
1739
+ Vendor: Raifle
1740
+ ======================================================================
1741
+
1742
+ Total Models: 1
1743
+ Free Models: 0
1744
+ Paid Models: 1
1745
+
1746
+ Input Pricing (per 1M tokens):
1747
+ Models with pricing: 1
1748
+ Min: $4.500000
1749
+ Max: $4.500000
1750
+ Mean: $4.500000
1751
+ Median: $4.500000
1752
+
1753
+ Output Pricing (per 1M tokens):
1754
+ Models with pricing: 1
1755
+ Min: $4.500000
1756
+ Max: $4.500000
1757
+ Mean: $4.500000
1758
+ Median: $4.500000
1759
+
1760
+ Context Lengths:
1761
+ Models with context data: 1
1762
+ Min: 16,000 tokens
1763
+ Max: 16,000 tokens
1764
+ Mean: 16,000 tokens
1765
+ Median: 16,000 tokens
1766
+
1767
+ Models:
1768
+ - SorcererLM 8x22B
1769
+
1770
+
1771
+ ======================================================================
1772
+ Vendor: Anthracite
1773
+ ======================================================================
1774
+
1775
+ Total Models: 1
1776
+ Free Models: 0
1777
+ Paid Models: 1
1778
+
1779
+ Input Pricing (per 1M tokens):
1780
+ Models with pricing: 1
1781
+ Min: $3.000000
1782
+ Max: $3.000000
1783
+ Mean: $3.000000
1784
+ Median: $3.000000
1785
+
1786
+ Output Pricing (per 1M tokens):
1787
+ Models with pricing: 1
1788
+ Min: $5.000000
1789
+ Max: $5.000000
1790
+ Mean: $5.000000
1791
+ Median: $5.000000
1792
+
1793
+ Context Lengths:
1794
+ Models with context data: 1
1795
+ Min: 16,384 tokens
1796
+ Max: 16,384 tokens
1797
+ Mean: 16,384 tokens
1798
+ Median: 16,384 tokens
1799
+
1800
+ Models:
1801
+ - Magnum v4 72B
1802
+
1803
+
1804
+ ======================================================================
1805
+ Vendor: Alpindale
1806
+ ======================================================================
1807
+
1808
+ Total Models: 1
1809
+ Free Models: 0
1810
+ Paid Models: 1
1811
+
1812
+ Input Pricing (per 1M tokens):
1813
+ Models with pricing: 1
1814
+ Min: $4.000000
1815
+ Max: $4.000000
1816
+ Mean: $4.000000
1817
+ Median: $4.000000
1818
+
1819
+ Output Pricing (per 1M tokens):
1820
+ Models with pricing: 1
1821
+ Min: $5.500000
1822
+ Max: $5.500000
1823
+ Mean: $5.500000
1824
+ Median: $5.500000
1825
+
1826
+ Context Lengths:
1827
+ Models with context data: 1
1828
+ Min: 6,144 tokens
1829
+ Max: 6,144 tokens
1830
+ Mean: 6,144 tokens
1831
+ Median: 6,144 tokens
1832
+
1833
+ Models:
1834
+ - Goliath 120B
1835
+
1836
+
1837
+ ======================================================================
1838
+ Vendor: Mancer
1839
+ ======================================================================
1840
+
1841
+ Total Models: 1
1842
+ Free Models: 0
1843
+ Paid Models: 1
1844
+
1845
+ Input Pricing (per 1M tokens):
1846
+ Models with pricing: 1
1847
+ Min: $1.125000
1848
+ Max: $1.125000
1849
+ Mean: $1.125000
1850
+ Median: $1.125000
1851
+
1852
+ Output Pricing (per 1M tokens):
1853
+ Models with pricing: 1
1854
+ Min: $1.125000
1855
+ Max: $1.125000
1856
+ Mean: $1.125000
1857
+ Median: $1.125000
1858
+
1859
+ Context Lengths:
1860
+ Models with context data: 1
1861
+ Min: 8,000 tokens
1862
+ Max: 8,000 tokens
1863
+ Mean: 8,000 tokens
1864
+ Median: 8,000 tokens
1865
+
1866
+ Models:
1867
+ - Mancer: Weaver (alpha)
1868
+
1869
+
1870
+ ======================================================================
1871
+ Vendor: Undi95
1872
+ ======================================================================
1873
+
1874
+ Total Models: 1
1875
+ Free Models: 0
1876
+ Paid Models: 1
1877
+
1878
+ Input Pricing (per 1M tokens):
1879
+ Models with pricing: 1
1880
+ Min: $0.450000
1881
+ Max: $0.450000
1882
+ Mean: $0.450000
1883
+ Median: $0.450000
1884
+
1885
+ Output Pricing (per 1M tokens):
1886
+ Models with pricing: 1
1887
+ Min: $0.650000
1888
+ Max: $0.650000
1889
+ Mean: $0.650000
1890
+ Median: $0.650000
1891
+
1892
+ Context Lengths:
1893
+ Models with context data: 1
1894
+ Min: 6,144 tokens
1895
+ Max: 6,144 tokens
1896
+ Mean: 6,144 tokens
1897
+ Median: 6,144 tokens
1898
+
1899
+ Models:
1900
+ - ReMM SLERP 13B
1901
+
1902
+
1903
+ ======================================================================
1904
+ Vendor: Gryphe
1905
+ ======================================================================
1906
+
1907
+ Total Models: 1
1908
+ Free Models: 0
1909
+ Paid Models: 1
1910
+
1911
+ Input Pricing (per 1M tokens):
1912
+ Models with pricing: 1
1913
+ Min: $0.060000
1914
+ Max: $0.060000
1915
+ Mean: $0.060000
1916
+ Median: $0.060000
1917
+
1918
+ Output Pricing (per 1M tokens):
1919
+ Models with pricing: 1
1920
+ Min: $0.060000
1921
+ Max: $0.060000
1922
+ Mean: $0.060000
1923
+ Median: $0.060000
1924
+
1925
+ Context Lengths:
1926
+ Models with context data: 1
1927
+ Min: 4,096 tokens
1928
+ Max: 4,096 tokens
1929
+ Mean: 4,096 tokens
1930
+ Median: 4,096 tokens
1931
+
1932
+ Models:
1933
+ - MythoMax 13B
1934
+
analysis/vendors/vendor_distribution.txt ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ OpenRouter API - Vendor Distribution
2
+ ======================================================================
3
+
4
+ Total Models: 342
5
+ Total Vendors: 55
6
+
7
+ Market Share by Model Count:
8
+ ----------------------------------------------------------------------
9
+ Vendor Models Share
10
+ ----------------------------------------------------------------------
11
+ OpenAI 47 13.74%
12
+ Qwen 46 13.45%
13
+ Mistral AI 36 10.53%
14
+ Google 24 7.02%
15
+ Meta 21 6.14%
16
+ DeepSeek 19 5.56%
17
+ Anthropic 13 3.80%
18
+ Microsoft 9 2.63%
19
+ Moonshot AI 7 2.05%
20
+ NVIDIA 7 2.05%
21
+ Z-AI 7 2.05%
22
+ xAI 7 2.05%
23
+ Nous Research 7 2.05%
24
+ Perplexity 6 1.75%
25
+ Baidu 5 1.46%
26
+ TheDrummer 5 1.46%
27
+ Arcee AI 5 1.46%
28
+ Sao10k 5 1.46%
29
+ Amazon 4 1.17%
30
+ Minimax 4 1.17%
31
+ DeepCogito 4 1.17%
32
+ TNG Technology 4 1.17%
33
+ Cohere 4 1.17%
34
+ Aion Labs 3 0.88%
35
+ OpenRouter 2 0.58%
36
+ Liquid AI 2 0.58%
37
+ Inclusion AI 2 0.58%
38
+ Alibaba 2 0.58%
39
+ Meituan 2 0.58%
40
+ AI21 Labs 2 0.58%
41
+ Morph 2 0.58%
42
+ Inception 2 0.58%
43
+ Arli AI 2 0.58%
44
+ Agentica 2 0.58%
45
+ Inflection AI 2 0.58%
46
+ Neversleep 2 0.58%
47
+ Kwai 1 0.29%
48
+ IBM 1 0.29%
49
+ Relace 1 0.29%
50
+ OpenGVLab 1 0.29%
51
+ Stepfun AI 1 0.29%
52
+ ByteDance 1 0.29%
53
+ Switchpoint 1 0.29%
54
+ THUDM 1 0.29%
55
+ Cognitive Computations 1 0.29%
56
+ Tencent 1 0.29%
57
+ EleutherAI 1 0.29%
58
+ AlfredPros 1 0.29%
59
+ Allen Institute for AI 1 0.29%
60
+ Raifle 1 0.29%
61
+ Anthracite 1 0.29%
62
+ Alpindale 1 0.29%
63
+ Mancer 1 0.29%
64
+ Undi95 1 0.29%
65
+ Gryphe 1 0.29%
66
+
analysis/vendors/vendor_stats.json ADDED
@@ -0,0 +1,1903 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "Kwai": {
3
+ "vendor": "Kwai",
4
+ "total_models": 1,
5
+ "free_models": 1,
6
+ "paid_models": 0,
7
+ "model_names": [
8
+ "Kwaipilot: Kat Coder (free)"
9
+ ],
10
+ "input_pricing": null,
11
+ "output_pricing": null,
12
+ "context_lengths": {
13
+ "count": 1,
14
+ "min": 256000,
15
+ "max": 256000,
16
+ "mean": 256000,
17
+ "median": 256000
18
+ }
19
+ },
20
+ "Moonshot AI": {
21
+ "vendor": "Moonshot AI",
22
+ "total_models": 7,
23
+ "free_models": 1,
24
+ "paid_models": 6,
25
+ "model_names": [
26
+ "MoonshotAI: Kimi Linear 48B A3B Instruct",
27
+ "MoonshotAI: Kimi K2 Thinking",
28
+ "MoonshotAI: Kimi K2 0905",
29
+ "MoonshotAI: Kimi K2 0905 (exacto)",
30
+ "MoonshotAI: Kimi K2 0711 (free)",
31
+ "MoonshotAI: Kimi K2 0711",
32
+ "MoonshotAI: Kimi Dev 72B"
33
+ ],
34
+ "input_pricing": {
35
+ "count": 6,
36
+ "min": 0.29,
37
+ "max": 0.6,
38
+ "mean": 0.44666666666666666,
39
+ "median": 0.445
40
+ },
41
+ "output_pricing": {
42
+ "count": 6,
43
+ "min": 0.6,
44
+ "max": 2.5,
45
+ "mean": 1.8416666666666666,
46
+ "median": 2.15
47
+ },
48
+ "context_lengths": {
49
+ "count": 7,
50
+ "min": 32768,
51
+ "max": 1048576,
52
+ "mean": 304274.28571428574,
53
+ "median": 262144
54
+ }
55
+ },
56
+ "OpenRouter": {
57
+ "vendor": "OpenRouter",
58
+ "total_models": 2,
59
+ "free_models": 1,
60
+ "paid_models": 1,
61
+ "model_names": [
62
+ "Polaris Alpha",
63
+ "Auto Router"
64
+ ],
65
+ "input_pricing": null,
66
+ "output_pricing": null,
67
+ "context_lengths": {
68
+ "count": 2,
69
+ "min": 256000,
70
+ "max": 2000000,
71
+ "mean": 1128000,
72
+ "median": 1128000.0
73
+ }
74
+ },
75
+ "Amazon": {
76
+ "vendor": "Amazon",
77
+ "total_models": 4,
78
+ "free_models": 0,
79
+ "paid_models": 4,
80
+ "model_names": [
81
+ "Amazon: Nova Premier 1.0",
82
+ "Amazon: Nova Lite 1.0",
83
+ "Amazon: Nova Micro 1.0",
84
+ "Amazon: Nova Pro 1.0"
85
+ ],
86
+ "input_pricing": {
87
+ "count": 4,
88
+ "min": 0.035,
89
+ "max": 2.5,
90
+ "mean": 0.84875,
91
+ "median": 0.42999999999999994
92
+ },
93
+ "output_pricing": {
94
+ "count": 4,
95
+ "min": 0.14,
96
+ "max": 12.5,
97
+ "mean": 4.02,
98
+ "median": 1.7199999999999998
99
+ },
100
+ "context_lengths": {
101
+ "count": 4,
102
+ "min": 128000,
103
+ "max": 1000000,
104
+ "mean": 432000,
105
+ "median": 300000.0
106
+ }
107
+ },
108
+ "Perplexity": {
109
+ "vendor": "Perplexity",
110
+ "total_models": 6,
111
+ "free_models": 0,
112
+ "paid_models": 6,
113
+ "model_names": [
114
+ "Perplexity: Sonar Pro Search",
115
+ "Perplexity: Sonar Reasoning Pro",
116
+ "Perplexity: Sonar Pro",
117
+ "Perplexity: Sonar Deep Research",
118
+ "Perplexity: Sonar Reasoning",
119
+ "Perplexity: Sonar"
120
+ ],
121
+ "input_pricing": {
122
+ "count": 6,
123
+ "min": 1.0,
124
+ "max": 3.0,
125
+ "mean": 2.0,
126
+ "median": 2.0
127
+ },
128
+ "output_pricing": {
129
+ "count": 6,
130
+ "min": 1.0,
131
+ "max": 15.0,
132
+ "mean": 8.666666666666666,
133
+ "median": 8.0
134
+ },
135
+ "context_lengths": {
136
+ "count": 6,
137
+ "min": 127000,
138
+ "max": 200000,
139
+ "mean": 151678.66666666666,
140
+ "median": 128000.0
141
+ }
142
+ },
143
+ "Mistral AI": {
144
+ "vendor": "Mistral AI",
145
+ "total_models": 36,
146
+ "free_models": 5,
147
+ "paid_models": 31,
148
+ "model_names": [
149
+ "Mistral: Voxtral Small 24B 2507",
150
+ "Mistral: Mistral Medium 3.1",
151
+ "Mistral: Codestral 2508",
152
+ "Mistral: Devstral Medium",
153
+ "Mistral: Devstral Small 1.1",
154
+ "Mistral: Mistral Small 3.2 24B (free)",
155
+ "Mistral: Mistral Small 3.2 24B",
156
+ "Mistral: Magistral Small 2506",
157
+ "Mistral: Magistral Medium 2506 (thinking)",
158
+ "Mistral: Magistral Medium 2506",
159
+ "Mistral: Devstral Small 2505",
160
+ "Mistral: Mistral Medium 3",
161
+ "Mistral: Mistral Small 3.1 24B (free)",
162
+ "Mistral: Mistral Small 3.1 24B",
163
+ "Mistral: Saba",
164
+ "Mistral: Mistral Small 3 (free)",
165
+ "Mistral: Mistral Small 3",
166
+ "Mistral: Codestral 2501",
167
+ "Mistral Large 2411",
168
+ "Mistral Large 2407",
169
+ "Mistral: Pixtral Large 2411",
170
+ "Mistral: Ministral 3B",
171
+ "Mistral: Ministral 8B",
172
+ "Mistral: Pixtral 12B",
173
+ "Mistral: Mistral Nemo (free)",
174
+ "Mistral: Mistral Nemo",
175
+ "Mistral: Mistral 7B Instruct (free)",
176
+ "Mistral: Mistral 7B Instruct",
177
+ "Mistral: Mistral 7B Instruct v0.3",
178
+ "Mistral: Mixtral 8x22B Instruct",
179
+ "Mistral Large",
180
+ "Mistral Tiny",
181
+ "Mistral Small",
182
+ "Mistral: Mistral 7B Instruct v0.2",
183
+ "Mistral: Mixtral 8x7B Instruct",
184
+ "Mistral: Mistral 7B Instruct v0.1"
185
+ ],
186
+ "input_pricing": {
187
+ "count": 31,
188
+ "min": 0.02,
189
+ "max": 2.0,
190
+ "mean": 0.5969677419354839,
191
+ "median": 0.19999999999999998
192
+ },
193
+ "output_pricing": {
194
+ "count": 31,
195
+ "min": 0.04,
196
+ "max": 6.0,
197
+ "mean": 1.7176774193548388,
198
+ "median": 0.54
199
+ },
200
+ "context_lengths": {
201
+ "count": 36,
202
+ "min": 2824,
203
+ "max": 262144,
204
+ "mean": 91244.66666666667,
205
+ "median": 112000.0
206
+ }
207
+ },
208
+ "OpenAI": {
209
+ "vendor": "OpenAI",
210
+ "total_models": 47,
211
+ "free_models": 1,
212
+ "paid_models": 46,
213
+ "model_names": [
214
+ "OpenAI: gpt-oss-safeguard-20b",
215
+ "OpenAI: GPT-5 Image Mini",
216
+ "OpenAI: GPT-5 Image",
217
+ "OpenAI: o3 Deep Research",
218
+ "OpenAI: o4 Mini Deep Research",
219
+ "OpenAI: GPT-5 Pro",
220
+ "OpenAI: GPT-5 Codex",
221
+ "OpenAI: GPT-4o Audio",
222
+ "OpenAI: GPT-5 Chat",
223
+ "OpenAI: GPT-5",
224
+ "OpenAI: GPT-5 Mini",
225
+ "OpenAI: GPT-5 Nano",
226
+ "OpenAI: gpt-oss-120b",
227
+ "OpenAI: gpt-oss-120b (exacto)",
228
+ "OpenAI: gpt-oss-20b (free)",
229
+ "OpenAI: gpt-oss-20b",
230
+ "OpenAI: o3 Pro",
231
+ "OpenAI: Codex Mini",
232
+ "OpenAI: o4 Mini High",
233
+ "OpenAI: o3",
234
+ "OpenAI: o4 Mini",
235
+ "OpenAI: GPT-4.1",
236
+ "OpenAI: GPT-4.1 Mini",
237
+ "OpenAI: GPT-4.1 Nano",
238
+ "OpenAI: o1-pro",
239
+ "OpenAI: GPT-4o-mini Search Preview",
240
+ "OpenAI: GPT-4o Search Preview",
241
+ "OpenAI: o3 Mini High",
242
+ "OpenAI: o3 Mini",
243
+ "OpenAI: o1",
244
+ "OpenAI: GPT-4o (2024-11-20)",
245
+ "OpenAI: ChatGPT-4o",
246
+ "OpenAI: GPT-4o (2024-08-06)",
247
+ "OpenAI: GPT-4o-mini",
248
+ "OpenAI: GPT-4o-mini (2024-07-18)",
249
+ "OpenAI: GPT-4o (2024-05-13)",
250
+ "OpenAI: GPT-4o",
251
+ "OpenAI: GPT-4o (extended)",
252
+ "OpenAI: GPT-4 Turbo",
253
+ "OpenAI: GPT-4 Turbo Preview",
254
+ "OpenAI: GPT-3.5 Turbo (older v0613)",
255
+ "OpenAI: GPT-4 Turbo (older v1106)",
256
+ "OpenAI: GPT-3.5 Turbo Instruct",
257
+ "OpenAI: GPT-3.5 Turbo 16k",
258
+ "OpenAI: GPT-4 (older v0314)",
259
+ "OpenAI: GPT-4",
260
+ "OpenAI: GPT-3.5 Turbo"
261
+ ],
262
+ "input_pricing": {
263
+ "count": 46,
264
+ "min": 0.03,
265
+ "max": 150.0,
266
+ "mean": 7.915108695652174,
267
+ "median": 2.0
268
+ },
269
+ "output_pricing": {
270
+ "count": 46,
271
+ "min": 0.14,
272
+ "max": 600.0,
273
+ "mean": 28.573478260869564,
274
+ "median": 8.0
275
+ },
276
+ "context_lengths": {
277
+ "count": 47,
278
+ "min": 4095,
279
+ "max": 1047576,
280
+ "mean": 229264.46808510637,
281
+ "median": 131072
282
+ }
283
+ },
284
+ "NVIDIA": {
285
+ "vendor": "NVIDIA",
286
+ "total_models": 7,
287
+ "free_models": 2,
288
+ "paid_models": 5,
289
+ "model_names": [
290
+ "NVIDIA: Nemotron Nano 12B 2 VL (free)",
291
+ "NVIDIA: Nemotron Nano 12B 2 VL",
292
+ "NVIDIA: Llama 3.3 Nemotron Super 49B V1.5",
293
+ "NVIDIA: Nemotron Nano 9B V2 (free)",
294
+ "NVIDIA: Nemotron Nano 9B V2",
295
+ "NVIDIA: Llama 3.1 Nemotron Ultra 253B v1",
296
+ "NVIDIA: Llama 3.1 Nemotron 70B Instruct"
297
+ ],
298
+ "input_pricing": {
299
+ "count": 5,
300
+ "min": 0.04,
301
+ "max": 0.6,
302
+ "mean": 0.308,
303
+ "median": 0.19999999999999998
304
+ },
305
+ "output_pricing": {
306
+ "count": 5,
307
+ "min": 0.16,
308
+ "max": 1.7999999999999998,
309
+ "mean": 0.712,
310
+ "median": 0.6
311
+ },
312
+ "context_lengths": {
313
+ "count": 7,
314
+ "min": 128000,
315
+ "max": 131072,
316
+ "mean": 130194.28571428571,
317
+ "median": 131072
318
+ }
319
+ },
320
+ "Minimax": {
321
+ "vendor": "Minimax",
322
+ "total_models": 4,
323
+ "free_models": 1,
324
+ "paid_models": 3,
325
+ "model_names": [
326
+ "MiniMax: MiniMax M2 (free)",
327
+ "MiniMax: MiniMax M2",
328
+ "MiniMax: MiniMax M1",
329
+ "MiniMax: MiniMax-01"
330
+ ],
331
+ "input_pricing": {
332
+ "count": 3,
333
+ "min": 0.19999999999999998,
334
+ "max": 0.39999999999999997,
335
+ "mean": 0.285,
336
+ "median": 0.255
337
+ },
338
+ "output_pricing": {
339
+ "count": 3,
340
+ "min": 1.02,
341
+ "max": 2.2,
342
+ "mean": 1.4400000000000002,
343
+ "median": 1.1
344
+ },
345
+ "context_lengths": {
346
+ "count": 4,
347
+ "min": 204800,
348
+ "max": 1000192,
349
+ "mean": 602448,
350
+ "median": 602400.0
351
+ }
352
+ },
353
+ "Liquid AI": {
354
+ "vendor": "Liquid AI",
355
+ "total_models": 2,
356
+ "free_models": 0,
357
+ "paid_models": 2,
358
+ "model_names": [
359
+ "LiquidAI/LFM2-8B-A1B",
360
+ "LiquidAI/LFM2-2.6B"
361
+ ],
362
+ "input_pricing": {
363
+ "count": 2,
364
+ "min": 0.049999999999999996,
365
+ "max": 0.049999999999999996,
366
+ "mean": 0.049999999999999996,
367
+ "median": 0.049999999999999996
368
+ },
369
+ "output_pricing": {
370
+ "count": 2,
371
+ "min": 0.09999999999999999,
372
+ "max": 0.09999999999999999,
373
+ "mean": 0.09999999999999999,
374
+ "median": 0.09999999999999999
375
+ },
376
+ "context_lengths": {
377
+ "count": 2,
378
+ "min": 32768,
379
+ "max": 32768,
380
+ "mean": 32768,
381
+ "median": 32768.0
382
+ }
383
+ },
384
+ "IBM": {
385
+ "vendor": "IBM",
386
+ "total_models": 1,
387
+ "free_models": 0,
388
+ "paid_models": 1,
389
+ "model_names": [
390
+ "IBM: Granite 4.0 Micro"
391
+ ],
392
+ "input_pricing": {
393
+ "count": 1,
394
+ "min": 0.017,
395
+ "max": 0.017,
396
+ "mean": 0.017,
397
+ "median": 0.017
398
+ },
399
+ "output_pricing": {
400
+ "count": 1,
401
+ "min": 0.11,
402
+ "max": 0.11,
403
+ "mean": 0.11,
404
+ "median": 0.11
405
+ },
406
+ "context_lengths": {
407
+ "count": 1,
408
+ "min": 131000,
409
+ "max": 131000,
410
+ "mean": 131000,
411
+ "median": 131000
412
+ }
413
+ },
414
+ "DeepCogito": {
415
+ "vendor": "DeepCogito",
416
+ "total_models": 4,
417
+ "free_models": 0,
418
+ "paid_models": 4,
419
+ "model_names": [
420
+ "Deep Cogito: Cogito V2 Preview Llama 405B",
421
+ "Deep Cogito: Cogito V2 Preview Llama 70B",
422
+ "Cogito V2 Preview Llama 109B",
423
+ "Deep Cogito: Cogito V2 Preview Deepseek 671B"
424
+ ],
425
+ "input_pricing": {
426
+ "count": 4,
427
+ "min": 0.18,
428
+ "max": 3.5,
429
+ "mean": 1.4525,
430
+ "median": 1.065
431
+ },
432
+ "output_pricing": {
433
+ "count": 4,
434
+ "min": 0.59,
435
+ "max": 3.5,
436
+ "mean": 1.555,
437
+ "median": 1.065
438
+ },
439
+ "context_lengths": {
440
+ "count": 4,
441
+ "min": 32767,
442
+ "max": 163840,
443
+ "mean": 65535.75,
444
+ "median": 32768.0
445
+ }
446
+ },
447
+ "Anthropic": {
448
+ "vendor": "Anthropic",
449
+ "total_models": 13,
450
+ "free_models": 0,
451
+ "paid_models": 13,
452
+ "model_names": [
453
+ "Anthropic: Claude Haiku 4.5",
454
+ "Anthropic: Claude Sonnet 4.5",
455
+ "Anthropic: Claude Opus 4.1",
456
+ "Anthropic: Claude Opus 4",
457
+ "Anthropic: Claude Sonnet 4",
458
+ "Anthropic: Claude 3.7 Sonnet (thinking)",
459
+ "Anthropic: Claude 3.7 Sonnet",
460
+ "Anthropic: Claude 3.5 Haiku",
461
+ "Anthropic: Claude 3.5 Haiku (2024-10-22)",
462
+ "Anthropic: Claude 3.5 Sonnet",
463
+ "Anthropic: Claude 3.5 Sonnet (2024-06-20)",
464
+ "Anthropic: Claude 3 Haiku",
465
+ "Anthropic: Claude 3 Opus"
466
+ ],
467
+ "input_pricing": {
468
+ "count": 13,
469
+ "min": 0.25,
470
+ "max": 15.0,
471
+ "mean": 5.065384615384615,
472
+ "median": 3.0
473
+ },
474
+ "output_pricing": {
475
+ "count": 13,
476
+ "min": 1.25,
477
+ "max": 75.0,
478
+ "mean": 25.326923076923077,
479
+ "median": 15.0
480
+ },
481
+ "context_lengths": {
482
+ "count": 13,
483
+ "min": 200000,
484
+ "max": 1000000,
485
+ "mean": 323076.92307692306,
486
+ "median": 200000
487
+ }
488
+ },
489
+ "Qwen": {
490
+ "vendor": "Qwen",
491
+ "total_models": 46,
492
+ "free_models": 8,
493
+ "paid_models": 38,
494
+ "model_names": [
495
+ "Qwen: Qwen3 VL 8B Thinking",
496
+ "Qwen: Qwen3 VL 8B Instruct",
497
+ "Qwen: Qwen3 VL 30B A3B Thinking",
498
+ "Qwen: Qwen3 VL 30B A3B Instruct",
499
+ "Qwen: Qwen3 VL 235B A22B Thinking",
500
+ "Qwen: Qwen3 VL 235B A22B Instruct",
501
+ "Qwen: Qwen3 Max",
502
+ "Qwen: Qwen3 Coder Plus",
503
+ "Qwen: Qwen3 Coder Flash",
504
+ "Qwen: Qwen3 Next 80B A3B Thinking",
505
+ "Qwen: Qwen3 Next 80B A3B Instruct",
506
+ "Qwen: Qwen Plus 0728",
507
+ "Qwen: Qwen Plus 0728 (thinking)",
508
+ "Qwen: Qwen3 30B A3B Thinking 2507",
509
+ "Qwen: Qwen3 Coder 30B A3B Instruct",
510
+ "Qwen: Qwen3 30B A3B Instruct 2507",
511
+ "Qwen: Qwen3 235B A22B Thinking 2507",
512
+ "Qwen: Qwen3 Coder 480B A35B (free)",
513
+ "Qwen: Qwen3 Coder 480B A35B",
514
+ "Qwen: Qwen3 Coder 480B A35B (exacto)",
515
+ "Qwen: Qwen3 235B A22B Instruct 2507",
516
+ "Qwen: Qwen3 4B (free)",
517
+ "Qwen: Qwen3 30B A3B (free)",
518
+ "Qwen: Qwen3 30B A3B",
519
+ "Qwen: Qwen3 8B",
520
+ "Qwen: Qwen3 14B (free)",
521
+ "Qwen: Qwen3 14B",
522
+ "Qwen: Qwen3 32B",
523
+ "Qwen: Qwen3 235B A22B (free)",
524
+ "Qwen: Qwen3 235B A22B",
525
+ "Qwen: Qwen2.5 Coder 7B Instruct",
526
+ "Qwen: Qwen2.5 VL 32B Instruct (free)",
527
+ "Qwen: Qwen2.5 VL 32B Instruct",
528
+ "Qwen: QwQ 32B",
529
+ "Qwen: Qwen VL Plus",
530
+ "Qwen: Qwen VL Max",
531
+ "Qwen: Qwen-Turbo",
532
+ "Qwen: Qwen2.5 VL 72B Instruct",
533
+ "Qwen: Qwen-Plus",
534
+ "Qwen: Qwen-Max ",
535
+ "Qwen2.5 Coder 32B Instruct (free)",
536
+ "Qwen2.5 Coder 32B Instruct",
537
+ "Qwen: Qwen2.5 7B Instruct",
538
+ "Qwen2.5 72B Instruct (free)",
539
+ "Qwen2.5 72B Instruct",
540
+ "Qwen: Qwen2.5-VL 7B Instruct"
541
+ ],
542
+ "input_pricing": {
543
+ "count": 38,
544
+ "min": 0.03,
545
+ "max": 1.5999999999999999,
546
+ "mean": 0.2577631578947368,
547
+ "median": 0.15
548
+ },
549
+ "output_pricing": {
550
+ "count": 38,
551
+ "min": 0.09,
552
+ "max": 6.3999999999999995,
553
+ "mean": 1.189421052631579,
554
+ "median": 0.575
555
+ },
556
+ "context_lengths": {
557
+ "count": 46,
558
+ "min": 7500,
559
+ "max": 1000000,
560
+ "mean": 185296.60869565216,
561
+ "median": 128000.0
562
+ }
563
+ },
564
+ "Inclusion AI": {
565
+ "vendor": "Inclusion AI",
566
+ "total_models": 2,
567
+ "free_models": 0,
568
+ "paid_models": 2,
569
+ "model_names": [
570
+ "inclusionAI: Ring 1T",
571
+ "inclusionAI: Ling-1T"
572
+ ],
573
+ "input_pricing": {
574
+ "count": 2,
575
+ "min": 0.5700000000000001,
576
+ "max": 0.5700000000000001,
577
+ "mean": 0.5700000000000001,
578
+ "median": 0.5700000000000001
579
+ },
580
+ "output_pricing": {
581
+ "count": 2,
582
+ "min": 2.2800000000000002,
583
+ "max": 2.2800000000000002,
584
+ "mean": 2.2800000000000002,
585
+ "median": 2.2800000000000002
586
+ },
587
+ "context_lengths": {
588
+ "count": 2,
589
+ "min": 131072,
590
+ "max": 131072,
591
+ "mean": 131072,
592
+ "median": 131072.0
593
+ }
594
+ },
595
+ "Baidu": {
596
+ "vendor": "Baidu",
597
+ "total_models": 5,
598
+ "free_models": 0,
599
+ "paid_models": 5,
600
+ "model_names": [
601
+ "Baidu: ERNIE 4.5 21B A3B Thinking",
602
+ "Baidu: ERNIE 4.5 21B A3B",
603
+ "Baidu: ERNIE 4.5 VL 28B A3B",
604
+ "Baidu: ERNIE 4.5 VL 424B A47B ",
605
+ "Baidu: ERNIE 4.5 300B A47B "
606
+ ],
607
+ "input_pricing": {
608
+ "count": 5,
609
+ "min": 0.07,
610
+ "max": 0.42,
611
+ "mean": 0.196,
612
+ "median": 0.14
613
+ },
614
+ "output_pricing": {
615
+ "count": 5,
616
+ "min": 0.28,
617
+ "max": 1.25,
618
+ "mean": 0.6940000000000001,
619
+ "median": 0.56
620
+ },
621
+ "context_lengths": {
622
+ "count": 5,
623
+ "min": 30000,
624
+ "max": 131072,
625
+ "mean": 105414.4,
626
+ "median": 123000
627
+ }
628
+ },
629
+ "Google": {
630
+ "vendor": "Google",
631
+ "total_models": 24,
632
+ "free_models": 6,
633
+ "paid_models": 18,
634
+ "model_names": [
635
+ "Google: Gemini 2.5 Flash Image (Nano Banana)",
636
+ "Google: Gemini 2.5 Flash Preview 09-2025",
637
+ "Google: Gemini 2.5 Flash Lite Preview 09-2025",
638
+ "Google: Gemini 2.5 Flash Image Preview (Nano Banana)",
639
+ "Google: Gemini 2.5 Flash Lite",
640
+ "Google: Gemma 3n 2B (free)",
641
+ "Google: Gemini 2.5 Flash Lite Preview 06-17",
642
+ "Google: Gemini 2.5 Flash",
643
+ "Google: Gemini 2.5 Pro",
644
+ "Google: Gemini 2.5 Pro Preview 06-05",
645
+ "Google: Gemma 3n 4B (free)",
646
+ "Google: Gemma 3n 4B",
647
+ "Google: Gemini 2.5 Pro Preview 05-06",
648
+ "Google: Gemma 3 4B (free)",
649
+ "Google: Gemma 3 4B",
650
+ "Google: Gemma 3 12B (free)",
651
+ "Google: Gemma 3 12B",
652
+ "Google: Gemma 3 27B (free)",
653
+ "Google: Gemma 3 27B",
654
+ "Google: Gemini 2.0 Flash Lite",
655
+ "Google: Gemini 2.0 Flash",
656
+ "Google: Gemini 2.0 Flash Experimental (free)",
657
+ "Google: Gemma 2 27B",
658
+ "Google: Gemma 2 9B"
659
+ ],
660
+ "input_pricing": {
661
+ "count": 18,
662
+ "min": 0.01703012,
663
+ "max": 1.25,
664
+ "mean": 0.3478905622222222,
665
+ "median": 0.09999999999999999
666
+ },
667
+ "output_pricing": {
668
+ "count": 18,
669
+ "min": 0.04,
670
+ "max": 10.0,
671
+ "mean": 2.389341866666667,
672
+ "median": 0.39999999999999997
673
+ },
674
+ "context_lengths": {
675
+ "count": 24,
676
+ "min": 8192,
677
+ "max": 1048576,
678
+ "mean": 509173.3333333333,
679
+ "median": 131072.0
680
+ }
681
+ },
682
+ "Z-AI": {
683
+ "vendor": "Z-AI",
684
+ "total_models": 7,
685
+ "free_models": 1,
686
+ "paid_models": 6,
687
+ "model_names": [
688
+ "Z.AI: GLM 4.6",
689
+ "Z.AI: GLM 4.6 (exacto)",
690
+ "Z.AI: GLM 4.5V",
691
+ "Z.AI: GLM 4.5",
692
+ "Z.AI: GLM 4.5 Air (free)",
693
+ "Z.AI: GLM 4.5 Air",
694
+ "Z.AI: GLM 4 32B "
695
+ ],
696
+ "input_pricing": {
697
+ "count": 6,
698
+ "min": 0.09999999999999999,
699
+ "max": 0.6,
700
+ "mean": 0.3383333333333333,
701
+ "median": 0.375
702
+ },
703
+ "output_pricing": {
704
+ "count": 6,
705
+ "min": 0.09999999999999999,
706
+ "max": 1.9,
707
+ "mean": 1.325,
708
+ "median": 1.65
709
+ },
710
+ "context_lengths": {
711
+ "count": 7,
712
+ "min": 65536,
713
+ "max": 202752,
714
+ "mean": 141750.85714285713,
715
+ "median": 131072
716
+ }
717
+ },
718
+ "DeepSeek": {
719
+ "vendor": "DeepSeek",
720
+ "total_models": 19,
721
+ "free_models": 6,
722
+ "paid_models": 13,
723
+ "model_names": [
724
+ "DeepSeek: DeepSeek V3.2 Exp",
725
+ "DeepSeek: DeepSeek V3.1 Terminus",
726
+ "DeepSeek: DeepSeek V3.1 Terminus (exacto)",
727
+ "DeepSeek: DeepSeek V3.1 (free)",
728
+ "DeepSeek: DeepSeek V3.1",
729
+ "DeepSeek: DeepSeek R1 0528 Qwen3 8B (free)",
730
+ "DeepSeek: DeepSeek R1 0528 Qwen3 8B",
731
+ "DeepSeek: R1 0528 (free)",
732
+ "DeepSeek: R1 0528",
733
+ "DeepSeek: DeepSeek Prover V2",
734
+ "DeepSeek: DeepSeek V3 0324 (free)",
735
+ "DeepSeek: DeepSeek V3 0324",
736
+ "DeepSeek: R1 Distill Qwen 32B",
737
+ "DeepSeek: R1 Distill Qwen 14B",
738
+ "DeepSeek: R1 Distill Llama 70B (free)",
739
+ "DeepSeek: R1 Distill Llama 70B",
740
+ "DeepSeek: R1 (free)",
741
+ "DeepSeek: R1",
742
+ "DeepSeek: DeepSeek V3"
743
+ ],
744
+ "input_pricing": {
745
+ "count": 13,
746
+ "min": 0.02,
747
+ "max": 0.5,
748
+ "mean": 0.24461538461538462,
749
+ "median": 0.27
750
+ },
751
+ "output_pricing": {
752
+ "count": 13,
753
+ "min": 0.09999999999999999,
754
+ "max": 2.1799999999999997,
755
+ "mean": 0.84,
756
+ "median": 0.84
757
+ },
758
+ "context_lengths": {
759
+ "count": 19,
760
+ "min": 8192,
761
+ "max": 163840,
762
+ "mean": 134950.31578947368,
763
+ "median": 163840
764
+ }
765
+ },
766
+ "TheDrummer": {
767
+ "vendor": "TheDrummer",
768
+ "total_models": 5,
769
+ "free_models": 0,
770
+ "paid_models": 5,
771
+ "model_names": [
772
+ "TheDrummer: Cydonia 24B V4.1",
773
+ "TheDrummer: Anubis 70B V1.1",
774
+ "TheDrummer: Skyfall 36B V2",
775
+ "TheDrummer: UnslopNemo 12B",
776
+ "TheDrummer: Rocinante 12B"
777
+ ],
778
+ "input_pricing": {
779
+ "count": 5,
780
+ "min": 0.16999999999999998,
781
+ "max": 0.65,
782
+ "mean": 0.40399999999999997,
783
+ "median": 0.39999999999999997
784
+ },
785
+ "output_pricing": {
786
+ "count": 5,
787
+ "min": 0.39999999999999997,
788
+ "max": 1.0,
789
+ "mean": 0.626,
790
+ "median": 0.5
791
+ },
792
+ "context_lengths": {
793
+ "count": 5,
794
+ "min": 32768,
795
+ "max": 131072,
796
+ "mean": 72089.6,
797
+ "median": 32768
798
+ }
799
+ },
800
+ "Relace": {
801
+ "vendor": "Relace",
802
+ "total_models": 1,
803
+ "free_models": 0,
804
+ "paid_models": 1,
805
+ "model_names": [
806
+ "Relace: Relace Apply 3"
807
+ ],
808
+ "input_pricing": {
809
+ "count": 1,
810
+ "min": 0.85,
811
+ "max": 0.85,
812
+ "mean": 0.85,
813
+ "median": 0.85
814
+ },
815
+ "output_pricing": {
816
+ "count": 1,
817
+ "min": 1.25,
818
+ "max": 1.25,
819
+ "mean": 1.25,
820
+ "median": 1.25
821
+ },
822
+ "context_lengths": {
823
+ "count": 1,
824
+ "min": 256000,
825
+ "max": 256000,
826
+ "mean": 256000,
827
+ "median": 256000
828
+ }
829
+ },
830
+ "xAI": {
831
+ "vendor": "xAI",
832
+ "total_models": 7,
833
+ "free_models": 0,
834
+ "paid_models": 7,
835
+ "model_names": [
836
+ "xAI: Grok 4 Fast",
837
+ "xAI: Grok Code Fast 1",
838
+ "xAI: Grok 4",
839
+ "xAI: Grok 3 Mini",
840
+ "xAI: Grok 3",
841
+ "xAI: Grok 3 Mini Beta",
842
+ "xAI: Grok 3 Beta"
843
+ ],
844
+ "input_pricing": {
845
+ "count": 7,
846
+ "min": 0.19999999999999998,
847
+ "max": 3.0,
848
+ "mean": 1.4285714285714286,
849
+ "median": 0.3
850
+ },
851
+ "output_pricing": {
852
+ "count": 7,
853
+ "min": 0.5,
854
+ "max": 15.0,
855
+ "mean": 6.857142857142857,
856
+ "median": 1.5
857
+ },
858
+ "context_lengths": {
859
+ "count": 7,
860
+ "min": 131072,
861
+ "max": 2000000,
862
+ "mean": 433755.4285714286,
863
+ "median": 131072
864
+ }
865
+ },
866
+ "Alibaba": {
867
+ "vendor": "Alibaba",
868
+ "total_models": 2,
869
+ "free_models": 1,
870
+ "paid_models": 1,
871
+ "model_names": [
872
+ "Tongyi DeepResearch 30B A3B (free)",
873
+ "Tongyi DeepResearch 30B A3B"
874
+ ],
875
+ "input_pricing": {
876
+ "count": 1,
877
+ "min": 0.09,
878
+ "max": 0.09,
879
+ "mean": 0.09,
880
+ "median": 0.09
881
+ },
882
+ "output_pricing": {
883
+ "count": 1,
884
+ "min": 0.39999999999999997,
885
+ "max": 0.39999999999999997,
886
+ "mean": 0.39999999999999997,
887
+ "median": 0.39999999999999997
888
+ },
889
+ "context_lengths": {
890
+ "count": 2,
891
+ "min": 131072,
892
+ "max": 131072,
893
+ "mean": 131072,
894
+ "median": 131072.0
895
+ }
896
+ },
897
+ "Arcee AI": {
898
+ "vendor": "Arcee AI",
899
+ "total_models": 5,
900
+ "free_models": 0,
901
+ "paid_models": 5,
902
+ "model_names": [
903
+ "Arcee AI: AFM 4.5B",
904
+ "Arcee AI: Spotlight",
905
+ "Arcee AI: Maestro Reasoning",
906
+ "Arcee AI: Virtuoso Large",
907
+ "Arcee AI: Coder Large"
908
+ ],
909
+ "input_pricing": {
910
+ "count": 5,
911
+ "min": 0.048,
912
+ "max": 0.8999999999999999,
913
+ "mean": 0.47559999999999997,
914
+ "median": 0.5
915
+ },
916
+ "output_pricing": {
917
+ "count": 5,
918
+ "min": 0.15,
919
+ "max": 3.3000000000000003,
920
+ "mean": 1.1260000000000001,
921
+ "median": 0.7999999999999999
922
+ },
923
+ "context_lengths": {
924
+ "count": 5,
925
+ "min": 32768,
926
+ "max": 131072,
927
+ "mean": 98304,
928
+ "median": 131072
929
+ }
930
+ },
931
+ "OpenGVLab": {
932
+ "vendor": "OpenGVLab",
933
+ "total_models": 1,
934
+ "free_models": 0,
935
+ "paid_models": 1,
936
+ "model_names": [
937
+ "OpenGVLab: InternVL3 78B"
938
+ ],
939
+ "input_pricing": {
940
+ "count": 1,
941
+ "min": 0.07,
942
+ "max": 0.07,
943
+ "mean": 0.07,
944
+ "median": 0.07
945
+ },
946
+ "output_pricing": {
947
+ "count": 1,
948
+ "min": 0.26,
949
+ "max": 0.26,
950
+ "mean": 0.26,
951
+ "median": 0.26
952
+ },
953
+ "context_lengths": {
954
+ "count": 1,
955
+ "min": 32768,
956
+ "max": 32768,
957
+ "mean": 32768,
958
+ "median": 32768
959
+ }
960
+ },
961
+ "Meituan": {
962
+ "vendor": "Meituan",
963
+ "total_models": 2,
964
+ "free_models": 1,
965
+ "paid_models": 1,
966
+ "model_names": [
967
+ "Meituan: LongCat Flash Chat (free)",
968
+ "Meituan: LongCat Flash Chat"
969
+ ],
970
+ "input_pricing": {
971
+ "count": 1,
972
+ "min": 0.15,
973
+ "max": 0.15,
974
+ "mean": 0.15,
975
+ "median": 0.15
976
+ },
977
+ "output_pricing": {
978
+ "count": 1,
979
+ "min": 0.75,
980
+ "max": 0.75,
981
+ "mean": 0.75,
982
+ "median": 0.75
983
+ },
984
+ "context_lengths": {
985
+ "count": 2,
986
+ "min": 131072,
987
+ "max": 131072,
988
+ "mean": 131072,
989
+ "median": 131072.0
990
+ }
991
+ },
992
+ "Stepfun AI": {
993
+ "vendor": "Stepfun AI",
994
+ "total_models": 1,
995
+ "free_models": 0,
996
+ "paid_models": 1,
997
+ "model_names": [
998
+ "StepFun: Step3"
999
+ ],
1000
+ "input_pricing": {
1001
+ "count": 1,
1002
+ "min": 0.5700000000000001,
1003
+ "max": 0.5700000000000001,
1004
+ "mean": 0.5700000000000001,
1005
+ "median": 0.5700000000000001
1006
+ },
1007
+ "output_pricing": {
1008
+ "count": 1,
1009
+ "min": 1.42,
1010
+ "max": 1.42,
1011
+ "mean": 1.42,
1012
+ "median": 1.42
1013
+ },
1014
+ "context_lengths": {
1015
+ "count": 1,
1016
+ "min": 65536,
1017
+ "max": 65536,
1018
+ "mean": 65536,
1019
+ "median": 65536
1020
+ }
1021
+ },
1022
+ "Nous Research": {
1023
+ "vendor": "Nous Research",
1024
+ "total_models": 7,
1025
+ "free_models": 1,
1026
+ "paid_models": 6,
1027
+ "model_names": [
1028
+ "Nous: Hermes 4 70B",
1029
+ "Nous: Hermes 4 405B",
1030
+ "Nous: DeepHermes 3 Mistral 24B Preview",
1031
+ "Nous: Hermes 3 70B Instruct",
1032
+ "Nous: Hermes 3 405B Instruct (free)",
1033
+ "Nous: Hermes 3 405B Instruct",
1034
+ "NousResearch: Hermes 2 Pro - Llama-3 8B"
1035
+ ],
1036
+ "input_pricing": {
1037
+ "count": 6,
1038
+ "min": 0.024999999999999998,
1039
+ "max": 1.0,
1040
+ "mean": 0.31416666666666665,
1041
+ "median": 0.22499999999999998
1042
+ },
1043
+ "output_pricing": {
1044
+ "count": 6,
1045
+ "min": 0.08,
1046
+ "max": 1.2,
1047
+ "mean": 0.5916666666666667,
1048
+ "median": 0.485
1049
+ },
1050
+ "context_lengths": {
1051
+ "count": 7,
1052
+ "min": 32768,
1053
+ "max": 131072,
1054
+ "mean": 93622.85714285714,
1055
+ "median": 131072
1056
+ }
1057
+ },
1058
+ "AI21 Labs": {
1059
+ "vendor": "AI21 Labs",
1060
+ "total_models": 2,
1061
+ "free_models": 0,
1062
+ "paid_models": 2,
1063
+ "model_names": [
1064
+ "AI21: Jamba Mini 1.7",
1065
+ "AI21: Jamba Large 1.7"
1066
+ ],
1067
+ "input_pricing": {
1068
+ "count": 2,
1069
+ "min": 0.19999999999999998,
1070
+ "max": 2.0,
1071
+ "mean": 1.1,
1072
+ "median": 1.1
1073
+ },
1074
+ "output_pricing": {
1075
+ "count": 2,
1076
+ "min": 0.39999999999999997,
1077
+ "max": 8.0,
1078
+ "mean": 4.2,
1079
+ "median": 4.2
1080
+ },
1081
+ "context_lengths": {
1082
+ "count": 2,
1083
+ "min": 256000,
1084
+ "max": 256000,
1085
+ "mean": 256000,
1086
+ "median": 256000.0
1087
+ }
1088
+ },
1089
+ "ByteDance": {
1090
+ "vendor": "ByteDance",
1091
+ "total_models": 1,
1092
+ "free_models": 0,
1093
+ "paid_models": 1,
1094
+ "model_names": [
1095
+ "ByteDance: UI-TARS 7B "
1096
+ ],
1097
+ "input_pricing": {
1098
+ "count": 1,
1099
+ "min": 0.09999999999999999,
1100
+ "max": 0.09999999999999999,
1101
+ "mean": 0.09999999999999999,
1102
+ "median": 0.09999999999999999
1103
+ },
1104
+ "output_pricing": {
1105
+ "count": 1,
1106
+ "min": 0.19999999999999998,
1107
+ "max": 0.19999999999999998,
1108
+ "mean": 0.19999999999999998,
1109
+ "median": 0.19999999999999998
1110
+ },
1111
+ "context_lengths": {
1112
+ "count": 1,
1113
+ "min": 128000,
1114
+ "max": 128000,
1115
+ "mean": 128000,
1116
+ "median": 128000
1117
+ }
1118
+ },
1119
+ "Switchpoint": {
1120
+ "vendor": "Switchpoint",
1121
+ "total_models": 1,
1122
+ "free_models": 0,
1123
+ "paid_models": 1,
1124
+ "model_names": [
1125
+ "Switchpoint Router"
1126
+ ],
1127
+ "input_pricing": {
1128
+ "count": 1,
1129
+ "min": 0.85,
1130
+ "max": 0.85,
1131
+ "mean": 0.85,
1132
+ "median": 0.85
1133
+ },
1134
+ "output_pricing": {
1135
+ "count": 1,
1136
+ "min": 3.4,
1137
+ "max": 3.4,
1138
+ "mean": 3.4,
1139
+ "median": 3.4
1140
+ },
1141
+ "context_lengths": {
1142
+ "count": 1,
1143
+ "min": 131072,
1144
+ "max": 131072,
1145
+ "mean": 131072,
1146
+ "median": 131072
1147
+ }
1148
+ },
1149
+ "THUDM": {
1150
+ "vendor": "THUDM",
1151
+ "total_models": 1,
1152
+ "free_models": 0,
1153
+ "paid_models": 1,
1154
+ "model_names": [
1155
+ "THUDM: GLM 4.1V 9B Thinking"
1156
+ ],
1157
+ "input_pricing": {
1158
+ "count": 1,
1159
+ "min": 0.035,
1160
+ "max": 0.035,
1161
+ "mean": 0.035,
1162
+ "median": 0.035
1163
+ },
1164
+ "output_pricing": {
1165
+ "count": 1,
1166
+ "min": 0.13799999999999998,
1167
+ "max": 0.13799999999999998,
1168
+ "mean": 0.13799999999999998,
1169
+ "median": 0.13799999999999998
1170
+ },
1171
+ "context_lengths": {
1172
+ "count": 1,
1173
+ "min": 65536,
1174
+ "max": 65536,
1175
+ "mean": 65536,
1176
+ "median": 65536
1177
+ }
1178
+ },
1179
+ "Cognitive Computations": {
1180
+ "vendor": "Cognitive Computations",
1181
+ "total_models": 1,
1182
+ "free_models": 1,
1183
+ "paid_models": 0,
1184
+ "model_names": [
1185
+ "Venice: Uncensored (free)"
1186
+ ],
1187
+ "input_pricing": null,
1188
+ "output_pricing": null,
1189
+ "context_lengths": {
1190
+ "count": 1,
1191
+ "min": 32768,
1192
+ "max": 32768,
1193
+ "mean": 32768,
1194
+ "median": 32768
1195
+ }
1196
+ },
1197
+ "Tencent": {
1198
+ "vendor": "Tencent",
1199
+ "total_models": 1,
1200
+ "free_models": 0,
1201
+ "paid_models": 1,
1202
+ "model_names": [
1203
+ "Tencent: Hunyuan A13B Instruct"
1204
+ ],
1205
+ "input_pricing": {
1206
+ "count": 1,
1207
+ "min": 0.14,
1208
+ "max": 0.14,
1209
+ "mean": 0.14,
1210
+ "median": 0.14
1211
+ },
1212
+ "output_pricing": {
1213
+ "count": 1,
1214
+ "min": 0.5700000000000001,
1215
+ "max": 0.5700000000000001,
1216
+ "mean": 0.5700000000000001,
1217
+ "median": 0.5700000000000001
1218
+ },
1219
+ "context_lengths": {
1220
+ "count": 1,
1221
+ "min": 131072,
1222
+ "max": 131072,
1223
+ "mean": 131072,
1224
+ "median": 131072
1225
+ }
1226
+ },
1227
+ "TNG Technology": {
1228
+ "vendor": "TNG Technology",
1229
+ "total_models": 4,
1230
+ "free_models": 2,
1231
+ "paid_models": 2,
1232
+ "model_names": [
1233
+ "TNG: DeepSeek R1T2 Chimera (free)",
1234
+ "TNG: DeepSeek R1T2 Chimera",
1235
+ "TNG: DeepSeek R1T Chimera (free)",
1236
+ "TNG: DeepSeek R1T Chimera"
1237
+ ],
1238
+ "input_pricing": {
1239
+ "count": 2,
1240
+ "min": 0.3,
1241
+ "max": 0.3,
1242
+ "mean": 0.3,
1243
+ "median": 0.3
1244
+ },
1245
+ "output_pricing": {
1246
+ "count": 2,
1247
+ "min": 1.2,
1248
+ "max": 1.2,
1249
+ "mean": 1.2,
1250
+ "median": 1.2
1251
+ },
1252
+ "context_lengths": {
1253
+ "count": 4,
1254
+ "min": 163840,
1255
+ "max": 163840,
1256
+ "mean": 163840,
1257
+ "median": 163840.0
1258
+ }
1259
+ },
1260
+ "Morph": {
1261
+ "vendor": "Morph",
1262
+ "total_models": 2,
1263
+ "free_models": 0,
1264
+ "paid_models": 2,
1265
+ "model_names": [
1266
+ "Morph: Morph V3 Large",
1267
+ "Morph: Morph V3 Fast"
1268
+ ],
1269
+ "input_pricing": {
1270
+ "count": 2,
1271
+ "min": 0.7999999999999999,
1272
+ "max": 0.8999999999999999,
1273
+ "mean": 0.8499999999999999,
1274
+ "median": 0.8499999999999999
1275
+ },
1276
+ "output_pricing": {
1277
+ "count": 2,
1278
+ "min": 1.2,
1279
+ "max": 1.9,
1280
+ "mean": 1.5499999999999998,
1281
+ "median": 1.5499999999999998
1282
+ },
1283
+ "context_lengths": {
1284
+ "count": 2,
1285
+ "min": 81920,
1286
+ "max": 262144,
1287
+ "mean": 172032,
1288
+ "median": 172032.0
1289
+ }
1290
+ },
1291
+ "Inception": {
1292
+ "vendor": "Inception",
1293
+ "total_models": 2,
1294
+ "free_models": 0,
1295
+ "paid_models": 2,
1296
+ "model_names": [
1297
+ "Inception: Mercury",
1298
+ "Inception: Mercury Coder"
1299
+ ],
1300
+ "input_pricing": {
1301
+ "count": 2,
1302
+ "min": 0.25,
1303
+ "max": 0.25,
1304
+ "mean": 0.25,
1305
+ "median": 0.25
1306
+ },
1307
+ "output_pricing": {
1308
+ "count": 2,
1309
+ "min": 1.0,
1310
+ "max": 1.0,
1311
+ "mean": 1.0,
1312
+ "median": 1.0
1313
+ },
1314
+ "context_lengths": {
1315
+ "count": 2,
1316
+ "min": 128000,
1317
+ "max": 128000,
1318
+ "mean": 128000,
1319
+ "median": 128000.0
1320
+ }
1321
+ },
1322
+ "Meta": {
1323
+ "vendor": "Meta",
1324
+ "total_models": 21,
1325
+ "free_models": 5,
1326
+ "paid_models": 16,
1327
+ "model_names": [
1328
+ "Meta: Llama 3.3 8B Instruct (free)",
1329
+ "Meta: Llama Guard 4 12B",
1330
+ "Meta: Llama 4 Maverick (free)",
1331
+ "Meta: Llama 4 Maverick",
1332
+ "Meta: Llama 4 Scout (free)",
1333
+ "Meta: Llama 4 Scout",
1334
+ "Llama Guard 3 8B",
1335
+ "Meta: Llama 3.3 70B Instruct (free)",
1336
+ "Meta: Llama 3.3 70B Instruct",
1337
+ "Meta: Llama 3.2 1B Instruct",
1338
+ "Meta: Llama 3.2 3B Instruct (free)",
1339
+ "Meta: Llama 3.2 3B Instruct",
1340
+ "Meta: Llama 3.2 11B Vision Instruct",
1341
+ "Meta: Llama 3.2 90B Vision Instruct",
1342
+ "Meta: Llama 3.1 405B (base)",
1343
+ "Meta: Llama 3.1 8B Instruct",
1344
+ "Meta: Llama 3.1 405B Instruct",
1345
+ "Meta: Llama 3.1 70B Instruct",
1346
+ "Meta: LlamaGuard 2 8B",
1347
+ "Meta: Llama 3 70B Instruct",
1348
+ "Meta: Llama 3 8B Instruct"
1349
+ ],
1350
+ "input_pricing": {
1351
+ "count": 16,
1352
+ "min": 0.005,
1353
+ "max": 4.0,
1354
+ "mean": 0.420875,
1355
+ "median": 0.14
1356
+ },
1357
+ "output_pricing": {
1358
+ "count": 16,
1359
+ "min": 0.01,
1360
+ "max": 4.0,
1361
+ "mean": 0.49306249999999996,
1362
+ "median": 0.25
1363
+ },
1364
+ "context_lengths": {
1365
+ "count": 21,
1366
+ "min": 8192,
1367
+ "max": 1048576,
1368
+ "mean": 153648.7619047619,
1369
+ "median": 131072
1370
+ }
1371
+ },
1372
+ "Microsoft": {
1373
+ "vendor": "Microsoft",
1374
+ "total_models": 9,
1375
+ "free_models": 1,
1376
+ "paid_models": 8,
1377
+ "model_names": [
1378
+ "Microsoft: Phi 4 Reasoning Plus",
1379
+ "Microsoft: MAI DS R1 (free)",
1380
+ "Microsoft: MAI DS R1",
1381
+ "Microsoft: Phi 4 Multimodal Instruct",
1382
+ "Microsoft: Phi 4",
1383
+ "Microsoft: Phi-3.5 Mini 128K Instruct",
1384
+ "Microsoft: Phi-3 Mini 128K Instruct",
1385
+ "Microsoft: Phi-3 Medium 128K Instruct",
1386
+ "WizardLM-2 8x22B"
1387
+ ],
1388
+ "input_pricing": {
1389
+ "count": 8,
1390
+ "min": 0.049999999999999996,
1391
+ "max": 1.0,
1392
+ "mean": 0.27,
1393
+ "median": 0.09999999999999999
1394
+ },
1395
+ "output_pricing": {
1396
+ "count": 8,
1397
+ "min": 0.09999999999999999,
1398
+ "max": 1.2,
1399
+ "mean": 0.43374999999999997,
1400
+ "median": 0.245
1401
+ },
1402
+ "context_lengths": {
1403
+ "count": 9,
1404
+ "min": 16384,
1405
+ "max": 163840,
1406
+ "mean": 106382.22222222222,
1407
+ "median": 128000
1408
+ }
1409
+ },
1410
+ "EleutherAI": {
1411
+ "vendor": "EleutherAI",
1412
+ "total_models": 1,
1413
+ "free_models": 0,
1414
+ "paid_models": 1,
1415
+ "model_names": [
1416
+ "EleutherAI: Llemma 7b"
1417
+ ],
1418
+ "input_pricing": {
1419
+ "count": 1,
1420
+ "min": 0.7999999999999999,
1421
+ "max": 0.7999999999999999,
1422
+ "mean": 0.7999999999999999,
1423
+ "median": 0.7999999999999999
1424
+ },
1425
+ "output_pricing": {
1426
+ "count": 1,
1427
+ "min": 1.2,
1428
+ "max": 1.2,
1429
+ "mean": 1.2,
1430
+ "median": 1.2
1431
+ },
1432
+ "context_lengths": {
1433
+ "count": 1,
1434
+ "min": 4096,
1435
+ "max": 4096,
1436
+ "mean": 4096,
1437
+ "median": 4096
1438
+ }
1439
+ },
1440
+ "AlfredPros": {
1441
+ "vendor": "AlfredPros",
1442
+ "total_models": 1,
1443
+ "free_models": 0,
1444
+ "paid_models": 1,
1445
+ "model_names": [
1446
+ "AlfredPros: CodeLLaMa 7B Instruct Solidity"
1447
+ ],
1448
+ "input_pricing": {
1449
+ "count": 1,
1450
+ "min": 0.7999999999999999,
1451
+ "max": 0.7999999999999999,
1452
+ "mean": 0.7999999999999999,
1453
+ "median": 0.7999999999999999
1454
+ },
1455
+ "output_pricing": {
1456
+ "count": 1,
1457
+ "min": 1.2,
1458
+ "max": 1.2,
1459
+ "mean": 1.2,
1460
+ "median": 1.2
1461
+ },
1462
+ "context_lengths": {
1463
+ "count": 1,
1464
+ "min": 4096,
1465
+ "max": 4096,
1466
+ "mean": 4096,
1467
+ "median": 4096
1468
+ }
1469
+ },
1470
+ "Arli AI": {
1471
+ "vendor": "Arli AI",
1472
+ "total_models": 2,
1473
+ "free_models": 1,
1474
+ "paid_models": 1,
1475
+ "model_names": [
1476
+ "ArliAI: QwQ 32B RpR v1 (free)",
1477
+ "ArliAI: QwQ 32B RpR v1"
1478
+ ],
1479
+ "input_pricing": {
1480
+ "count": 1,
1481
+ "min": 0.03,
1482
+ "max": 0.03,
1483
+ "mean": 0.03,
1484
+ "median": 0.03
1485
+ },
1486
+ "output_pricing": {
1487
+ "count": 1,
1488
+ "min": 0.11,
1489
+ "max": 0.11,
1490
+ "mean": 0.11,
1491
+ "median": 0.11
1492
+ },
1493
+ "context_lengths": {
1494
+ "count": 2,
1495
+ "min": 32768,
1496
+ "max": 32768,
1497
+ "mean": 32768,
1498
+ "median": 32768.0
1499
+ }
1500
+ },
1501
+ "Agentica": {
1502
+ "vendor": "Agentica",
1503
+ "total_models": 2,
1504
+ "free_models": 1,
1505
+ "paid_models": 1,
1506
+ "model_names": [
1507
+ "Agentica: Deepcoder 14B Preview (free)",
1508
+ "Agentica: Deepcoder 14B Preview"
1509
+ ],
1510
+ "input_pricing": {
1511
+ "count": 1,
1512
+ "min": 0.015,
1513
+ "max": 0.015,
1514
+ "mean": 0.015,
1515
+ "median": 0.015
1516
+ },
1517
+ "output_pricing": {
1518
+ "count": 1,
1519
+ "min": 0.015,
1520
+ "max": 0.015,
1521
+ "mean": 0.015,
1522
+ "median": 0.015
1523
+ },
1524
+ "context_lengths": {
1525
+ "count": 2,
1526
+ "min": 96000,
1527
+ "max": 96000,
1528
+ "mean": 96000,
1529
+ "median": 96000.0
1530
+ }
1531
+ },
1532
+ "Allen Institute for AI": {
1533
+ "vendor": "Allen Institute for AI",
1534
+ "total_models": 1,
1535
+ "free_models": 0,
1536
+ "paid_models": 1,
1537
+ "model_names": [
1538
+ "AllenAI: Olmo 2 32B Instruct"
1539
+ ],
1540
+ "input_pricing": {
1541
+ "count": 1,
1542
+ "min": 0.19999999999999998,
1543
+ "max": 0.19999999999999998,
1544
+ "mean": 0.19999999999999998,
1545
+ "median": 0.19999999999999998
1546
+ },
1547
+ "output_pricing": {
1548
+ "count": 1,
1549
+ "min": 0.35,
1550
+ "max": 0.35,
1551
+ "mean": 0.35,
1552
+ "median": 0.35
1553
+ },
1554
+ "context_lengths": {
1555
+ "count": 1,
1556
+ "min": 4096,
1557
+ "max": 4096,
1558
+ "mean": 4096,
1559
+ "median": 4096
1560
+ }
1561
+ },
1562
+ "Cohere": {
1563
+ "vendor": "Cohere",
1564
+ "total_models": 4,
1565
+ "free_models": 0,
1566
+ "paid_models": 4,
1567
+ "model_names": [
1568
+ "Cohere: Command A",
1569
+ "Cohere: Command R7B (12-2024)",
1570
+ "Cohere: Command R+ (08-2024)",
1571
+ "Cohere: Command R (08-2024)"
1572
+ ],
1573
+ "input_pricing": {
1574
+ "count": 4,
1575
+ "min": 0.0375,
1576
+ "max": 2.5,
1577
+ "mean": 1.296875,
1578
+ "median": 1.325
1579
+ },
1580
+ "output_pricing": {
1581
+ "count": 4,
1582
+ "min": 0.15,
1583
+ "max": 10.0,
1584
+ "mean": 5.1875,
1585
+ "median": 5.3
1586
+ },
1587
+ "context_lengths": {
1588
+ "count": 4,
1589
+ "min": 128000,
1590
+ "max": 256000,
1591
+ "mean": 160000,
1592
+ "median": 128000.0
1593
+ }
1594
+ },
1595
+ "Aion Labs": {
1596
+ "vendor": "Aion Labs",
1597
+ "total_models": 3,
1598
+ "free_models": 0,
1599
+ "paid_models": 3,
1600
+ "model_names": [
1601
+ "AionLabs: Aion-1.0",
1602
+ "AionLabs: Aion-1.0-Mini",
1603
+ "AionLabs: Aion-RP 1.0 (8B)"
1604
+ ],
1605
+ "input_pricing": {
1606
+ "count": 3,
1607
+ "min": 0.19999999999999998,
1608
+ "max": 4.0,
1609
+ "mean": 1.6333333333333333,
1610
+ "median": 0.7
1611
+ },
1612
+ "output_pricing": {
1613
+ "count": 3,
1614
+ "min": 0.19999999999999998,
1615
+ "max": 8.0,
1616
+ "mean": 3.2,
1617
+ "median": 1.4
1618
+ },
1619
+ "context_lengths": {
1620
+ "count": 3,
1621
+ "min": 32768,
1622
+ "max": 131072,
1623
+ "mean": 98304,
1624
+ "median": 131072
1625
+ }
1626
+ },
1627
+ "Sao10k": {
1628
+ "vendor": "Sao10k",
1629
+ "total_models": 5,
1630
+ "free_models": 0,
1631
+ "paid_models": 5,
1632
+ "model_names": [
1633
+ "Sao10K: Llama 3.1 70B Hanami x1",
1634
+ "Sao10K: Llama 3.3 Euryale 70B",
1635
+ "Sao10K: Llama 3.1 Euryale 70B v2.2",
1636
+ "Sao10K: Llama 3 8B Lunaris",
1637
+ "Sao10k: Llama 3 Euryale 70B v2.1"
1638
+ ],
1639
+ "input_pricing": {
1640
+ "count": 5,
1641
+ "min": 0.04,
1642
+ "max": 3.0,
1643
+ "mean": 1.164,
1644
+ "median": 0.65
1645
+ },
1646
+ "output_pricing": {
1647
+ "count": 5,
1648
+ "min": 0.049999999999999996,
1649
+ "max": 3.0,
1650
+ "mean": 1.206,
1651
+ "median": 0.75
1652
+ },
1653
+ "context_lengths": {
1654
+ "count": 5,
1655
+ "min": 8192,
1656
+ "max": 131072,
1657
+ "mean": 39244.8,
1658
+ "median": 16000
1659
+ }
1660
+ },
1661
+ "Raifle": {
1662
+ "vendor": "Raifle",
1663
+ "total_models": 1,
1664
+ "free_models": 0,
1665
+ "paid_models": 1,
1666
+ "model_names": [
1667
+ "SorcererLM 8x22B"
1668
+ ],
1669
+ "input_pricing": {
1670
+ "count": 1,
1671
+ "min": 4.5,
1672
+ "max": 4.5,
1673
+ "mean": 4.5,
1674
+ "median": 4.5
1675
+ },
1676
+ "output_pricing": {
1677
+ "count": 1,
1678
+ "min": 4.5,
1679
+ "max": 4.5,
1680
+ "mean": 4.5,
1681
+ "median": 4.5
1682
+ },
1683
+ "context_lengths": {
1684
+ "count": 1,
1685
+ "min": 16000,
1686
+ "max": 16000,
1687
+ "mean": 16000,
1688
+ "median": 16000
1689
+ }
1690
+ },
1691
+ "Anthracite": {
1692
+ "vendor": "Anthracite",
1693
+ "total_models": 1,
1694
+ "free_models": 0,
1695
+ "paid_models": 1,
1696
+ "model_names": [
1697
+ "Magnum v4 72B"
1698
+ ],
1699
+ "input_pricing": {
1700
+ "count": 1,
1701
+ "min": 3.0,
1702
+ "max": 3.0,
1703
+ "mean": 3.0,
1704
+ "median": 3.0
1705
+ },
1706
+ "output_pricing": {
1707
+ "count": 1,
1708
+ "min": 5.0,
1709
+ "max": 5.0,
1710
+ "mean": 5.0,
1711
+ "median": 5.0
1712
+ },
1713
+ "context_lengths": {
1714
+ "count": 1,
1715
+ "min": 16384,
1716
+ "max": 16384,
1717
+ "mean": 16384,
1718
+ "median": 16384
1719
+ }
1720
+ },
1721
+ "Inflection AI": {
1722
+ "vendor": "Inflection AI",
1723
+ "total_models": 2,
1724
+ "free_models": 0,
1725
+ "paid_models": 2,
1726
+ "model_names": [
1727
+ "Inflection: Inflection 3 Productivity",
1728
+ "Inflection: Inflection 3 Pi"
1729
+ ],
1730
+ "input_pricing": {
1731
+ "count": 2,
1732
+ "min": 2.5,
1733
+ "max": 2.5,
1734
+ "mean": 2.5,
1735
+ "median": 2.5
1736
+ },
1737
+ "output_pricing": {
1738
+ "count": 2,
1739
+ "min": 10.0,
1740
+ "max": 10.0,
1741
+ "mean": 10.0,
1742
+ "median": 10.0
1743
+ },
1744
+ "context_lengths": {
1745
+ "count": 2,
1746
+ "min": 8000,
1747
+ "max": 8000,
1748
+ "mean": 8000,
1749
+ "median": 8000.0
1750
+ }
1751
+ },
1752
+ "Neversleep": {
1753
+ "vendor": "Neversleep",
1754
+ "total_models": 2,
1755
+ "free_models": 0,
1756
+ "paid_models": 2,
1757
+ "model_names": [
1758
+ "NeverSleep: Lumimaid v0.2 8B",
1759
+ "Noromaid 20B"
1760
+ ],
1761
+ "input_pricing": {
1762
+ "count": 2,
1763
+ "min": 0.09,
1764
+ "max": 1.0,
1765
+ "mean": 0.545,
1766
+ "median": 0.545
1767
+ },
1768
+ "output_pricing": {
1769
+ "count": 2,
1770
+ "min": 0.6,
1771
+ "max": 1.75,
1772
+ "mean": 1.175,
1773
+ "median": 1.175
1774
+ },
1775
+ "context_lengths": {
1776
+ "count": 2,
1777
+ "min": 4096,
1778
+ "max": 32768,
1779
+ "mean": 18432,
1780
+ "median": 18432.0
1781
+ }
1782
+ },
1783
+ "Alpindale": {
1784
+ "vendor": "Alpindale",
1785
+ "total_models": 1,
1786
+ "free_models": 0,
1787
+ "paid_models": 1,
1788
+ "model_names": [
1789
+ "Goliath 120B"
1790
+ ],
1791
+ "input_pricing": {
1792
+ "count": 1,
1793
+ "min": 4.0,
1794
+ "max": 4.0,
1795
+ "mean": 4.0,
1796
+ "median": 4.0
1797
+ },
1798
+ "output_pricing": {
1799
+ "count": 1,
1800
+ "min": 5.5,
1801
+ "max": 5.5,
1802
+ "mean": 5.5,
1803
+ "median": 5.5
1804
+ },
1805
+ "context_lengths": {
1806
+ "count": 1,
1807
+ "min": 6144,
1808
+ "max": 6144,
1809
+ "mean": 6144,
1810
+ "median": 6144
1811
+ }
1812
+ },
1813
+ "Mancer": {
1814
+ "vendor": "Mancer",
1815
+ "total_models": 1,
1816
+ "free_models": 0,
1817
+ "paid_models": 1,
1818
+ "model_names": [
1819
+ "Mancer: Weaver (alpha)"
1820
+ ],
1821
+ "input_pricing": {
1822
+ "count": 1,
1823
+ "min": 1.125,
1824
+ "max": 1.125,
1825
+ "mean": 1.125,
1826
+ "median": 1.125
1827
+ },
1828
+ "output_pricing": {
1829
+ "count": 1,
1830
+ "min": 1.125,
1831
+ "max": 1.125,
1832
+ "mean": 1.125,
1833
+ "median": 1.125
1834
+ },
1835
+ "context_lengths": {
1836
+ "count": 1,
1837
+ "min": 8000,
1838
+ "max": 8000,
1839
+ "mean": 8000,
1840
+ "median": 8000
1841
+ }
1842
+ },
1843
+ "Undi95": {
1844
+ "vendor": "Undi95",
1845
+ "total_models": 1,
1846
+ "free_models": 0,
1847
+ "paid_models": 1,
1848
+ "model_names": [
1849
+ "ReMM SLERP 13B"
1850
+ ],
1851
+ "input_pricing": {
1852
+ "count": 1,
1853
+ "min": 0.44999999999999996,
1854
+ "max": 0.44999999999999996,
1855
+ "mean": 0.44999999999999996,
1856
+ "median": 0.44999999999999996
1857
+ },
1858
+ "output_pricing": {
1859
+ "count": 1,
1860
+ "min": 0.65,
1861
+ "max": 0.65,
1862
+ "mean": 0.65,
1863
+ "median": 0.65
1864
+ },
1865
+ "context_lengths": {
1866
+ "count": 1,
1867
+ "min": 6144,
1868
+ "max": 6144,
1869
+ "mean": 6144,
1870
+ "median": 6144
1871
+ }
1872
+ },
1873
+ "Gryphe": {
1874
+ "vendor": "Gryphe",
1875
+ "total_models": 1,
1876
+ "free_models": 0,
1877
+ "paid_models": 1,
1878
+ "model_names": [
1879
+ "MythoMax 13B"
1880
+ ],
1881
+ "input_pricing": {
1882
+ "count": 1,
1883
+ "min": 0.06,
1884
+ "max": 0.06,
1885
+ "mean": 0.06,
1886
+ "median": 0.06
1887
+ },
1888
+ "output_pricing": {
1889
+ "count": 1,
1890
+ "min": 0.06,
1891
+ "max": 0.06,
1892
+ "mean": 0.06,
1893
+ "median": 0.06
1894
+ },
1895
+ "context_lengths": {
1896
+ "count": 1,
1897
+ "min": 4096,
1898
+ "max": 4096,
1899
+ "mean": 4096,
1900
+ "median": 4096
1901
+ }
1902
+ }
1903
+ }
analysis/vendors/vendor_summary.txt ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ OpenRouter API - Models by Vendor Summary
2
+ ======================================================================
3
+
4
+ Total Vendors: 55
5
+ Total Models: 342
6
+
7
+ Top Vendors by Model Count:
8
+ ----------------------------------------------------------------------
9
+ Vendor Models Free Paid
10
+ ----------------------------------------------------------------------
11
+ OpenAI 47 1 46
12
+ Qwen 46 8 38
13
+ Mistral AI 36 5 31
14
+ Google 24 6 18
15
+ Meta 21 5 16
16
+ DeepSeek 19 6 13
17
+ Anthropic 13 0 13
18
+ Microsoft 9 1 8
19
+ Moonshot AI 7 1 6
20
+ NVIDIA 7 2 5
21
+ Z-AI 7 1 6
22
+ xAI 7 0 7
23
+ Nous Research 7 1 6
24
+ Perplexity 6 0 6
25
+ Baidu 5 0 5
26
+ TheDrummer 5 0 5
27
+ Arcee AI 5 0 5
28
+ Sao10k 5 0 5
29
+ Amazon 4 0 4
30
+ Minimax 4 1 3
31
+