File size: 2,176 Bytes
f0b8a78
 
 
 
 
 
 
 
 
b50ffc0
0ca1f40
510a014
 
60b7962
510a014
 
 
25c3208
510a014
de5151f
510a014
 
 
 
c9499eb
510a014
 
c9499eb
 
 
 
510a014
c9499eb
18a2cd4
 
b50ffc0
5985c5e
 
71742a8
 
 
 
 
b50ffc0
510a014
44b1dfd
 
 
510a014
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
---
title: README
emoji: 📈
colorFrom: yellow
colorTo: green
sdk: static
pinned: false
---


<div class="grid lg:grid-cols-2 gap-x-4 gap-y-7">
	<p class="lg:col-span-3">
	  Welcome to CARROT-LLM-Routing! For a given desired trade off between performance and cost, 
      CARROT makes it easy to pick the best model among a set of 13 LLMs for any query. Below you may read the CARROT paper, replicate the training process of CARROT, or see how to utilize CARROT out of the box for routing.
	</p>
	<a href="https://arxiv.org/" class="block overflow-hidden group">
		<div
			class="w-40 h-39 object-cover mb-2 rounded-lg flex items-center justify-center bg-[#ECFAFF]"
		>
			<img alt="" src="fmselect_gpt4o_comparison.png" class="w-40" />
		</div>
		<div class="underline">Read the paper</div>
	</a>
	<a
		href="https://github.com/somerstep"
		class="block overflow-hidden"
	>
		<div
			class="w-40 h-39 object-cover mb-2 rounded-lg flex items-center justify-center bg-[#ECFAFF]"
		>
			<img alt="" src="logo.png" class="w-40" />
		</div>
		<div class="underline">Train CARROT</div>
      </a>
      
<p class="lg:col-span-3">
As is, CARROT supports routing to the following collection of large language models.

|                      | claude-3-5-sonnet-v1 | titan-text-premier-v1 | openai-gpt-4o | openai-gpt-4o-mini | granite-3-2b-instruct | granite-3-8b-instruct | llama-3-1-70b-instruct | llama-3-1-8b-instruct | llama-3-2-1b-instruct | llama-3-2-3b-instruct | llama-3-3-70b-instruct | mixtral-8x7b-instruct | llama-3-405b-instruct |
|----------------------|---------------------|----------------------|---------------|--------------------|----------------------|----------------------|----------------------|----------------------|----------------------|----------------------|----------------------|----------------------|----------------------|
| **Input Token Cost ($ per 1M tokens)**  | 3   | 0.5  | 2.5  | 0.15  | 0.1  | 0.2  | 0.9  | 0.2  | 0.06  | 0.06  | 0.9  | 0.6  | 3.5  |
| **Output Token Cost ($ per 1M tokens)** | 15  | 1.5  | 10   | 0.6   | 0.1  | 0.2  | 0.9  | 0.2  | 0.06  | 0.06  | 0.9  | 0.6  | 3.5  |

</p>

```python 
your_code = do_some_stuff
```