# MiniMax-M2 Model Repository This is the official MiniMax-M2 model repository containing a 230B parameter MoE model with 10B active parameters, optimized for coding and agentic workflows. ## Model Information - **Model Type**: Mixture of Experts (MoE) - **Total Parameters**: 230B - **Active Parameters**: 10B - **Architecture**: Transformer-based MoE - **License**: Modified MIT - **Pipeline Tag**: text-generation ## Usage This model can be used with various inference frameworks: ### Transformers ```python from transformers import AutoModelForCausalLM, AutoTokenizer model = AutoModelForCausalLM.from_pretrained("your-username/MiniMax-M2") tokenizer = AutoTokenizer.from_pretrained("your-username/MiniMax-M2") ``` ### vLLM ```python from vllm import LLM, SamplingParams llm = LLM(model="your-username/MiniMax-M2") ``` ### SGLang ```python from sglang import function, system, user, assistant, gen, select @function def multi_turn_question(s, question): s += system("You are a helpful assistant.") s += user(question) s += assistant(gen("answer", max_tokens=256)) return s["answer"] ``` ## Model Details - **Context Length**: 128K tokens - **Thinking Format**: Uses `...` tags for reasoning - **Recommended Parameters**: - Temperature: 1.0 - Top-p: 0.95 - Top-k: 40 ## Deployment Guides See the `docs/` directory for detailed deployment guides: - [Transformers Guide](docs/transformers_deploy_guide.md) - [vLLM Guide](docs/vllm_deploy_guide.md) - [SGLang Guide](docs/sglang_deploy_guide.md) - [MLX Guide](docs/mlx_deploy_guide.md) ## License This model is released under the Modified MIT License. See the [license file](https://github.com/MiniMax-AI/MiniMax-M2/blob/main/LICENSE) for details.