| | --- |
| | license: apache-2.0 |
| | --- |
| | |
| | ## Overview |
| |
|
| | The Mixtral-7x8B Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts. The Mistral-7x8Boutperforms Llama 2 70B on most benchmarks we tested. |
| |
|
| | ## Variants |
| |
|
| | | No | Variant | Cortex CLI command | |
| | | --- | --- | --- | |
| | | 1 | [7x8b-gguf](https://huggingface.co/cortexhub/mixtral/tree/7x8b-gguf) | `cortex run mixtral:7x8b-gguf` | |
| |
|
| | ## Use it with Jan (UI) |
| |
|
| | 1. Install **Jan** using [Quickstart](https://jan.ai/docs/quickstart) |
| | 2. Use in Jan model Hub: |
| | ``` |
| | cortexhub/mixtral |
| | ``` |
| | |
| | ## Use it with Cortex (CLI) |
| |
|
| | 1. Install **Cortex** using [Quickstart](https://cortex.jan.ai/docs/quickstart) |
| | 2. Run the model with command: |
| | ``` |
| | cortex run mixtral |
| | ``` |
| | |
| | ## Credits |
| |
|
| | - **Author:** Mistralai |
| | - **Converter:** [Homebrew](https://www.homebrew.ltd/) |