MiniMax-M2 / model_card.md
luhkronn's picture
Upload folder using huggingface_hub
ffb99ea verified
metadata
tags:
  - text-generation
  - conversational
  - coding
  - agent
  - moe
  - large-language-model
license: other
license_name: modified-mit
license_link: https://github.com/MiniMax-AI/MiniMax-M2/blob/main/LICENSE
library_name: transformers
pipeline_tag: text-generation

MiniMax-M2

MiniMax-M2 is a Mini model built for Max coding & agentic workflows. It's a compact, fast, and cost-effective MoE model (230 billion total parameters with 10 billion active parameters) built for elite performance in coding and agentic tasks.