Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -14,22 +14,22 @@ if not os.path.exists(AI_PFP) or not os.path.exists(USER_PFP):
|
|
| 14 |
|
| 15 |
model_info = {
|
| 16 |
"command-a-03-2025": {
|
| 17 |
-
"description": "Command A is our most performant model to date, excelling at tool use, agents, retrieval augmented generation (RAG), and multilingual use cases.",
|
| 18 |
"context": "256K",
|
| 19 |
"output": "8K"
|
| 20 |
},
|
| 21 |
"command-r7b-12-2024": {
|
| 22 |
-
"description": "
|
| 23 |
"context": "128K",
|
| 24 |
"output": "4K"
|
| 25 |
},
|
| 26 |
"command-r-plus-04-2024": {
|
| 27 |
-
"description": "
|
| 28 |
"context": "128K",
|
| 29 |
"output": "4K"
|
| 30 |
},
|
| 31 |
"command-r-plus": {
|
| 32 |
-
"description": "
|
| 33 |
"context": "128K",
|
| 34 |
"output": "4K"
|
| 35 |
},
|
|
|
|
| 14 |
|
| 15 |
model_info = {
|
| 16 |
"command-a-03-2025": {
|
| 17 |
+
"description": "Command A is our most performant model to date, excelling at tool use, agents, retrieval augmented generation (RAG), and multilingual use cases. Command A has a context length of 256K, only requires two GPUs to run, and has 150% higher throughput compared to Command R+ 08-2024.",
|
| 18 |
"context": "256K",
|
| 19 |
"output": "8K"
|
| 20 |
},
|
| 21 |
"command-r7b-12-2024": {
|
| 22 |
+
"description": "command-r7b-12-2024 is a small, fast update delivered in December 2024. It excels at RAG, tool use, agents, and similar tasks requiring complex reasoning and multiple steps.",
|
| 23 |
"context": "128K",
|
| 24 |
"output": "4K"
|
| 25 |
},
|
| 26 |
"command-r-plus-04-2024": {
|
| 27 |
+
"description": "Command R+ is an instruction-following conversational model that performs language tasks at a higher quality, more reliably, and with a longer context than previous models. It is best suited for complex RAG workflows and multi-step tool use.",
|
| 28 |
"context": "128K",
|
| 29 |
"output": "4K"
|
| 30 |
},
|
| 31 |
"command-r-plus": {
|
| 32 |
+
"description": "command-r-plus is an alias for command-r-plus-04-2024, so if you use command-r-plus in the API, that’s the model you’re pointing to.",
|
| 33 |
"context": "128K",
|
| 34 |
"output": "4K"
|
| 35 |
},
|