| --- |
| AIGC: |
| ContentProducer: Minimax Agent AI |
| ContentPropagator: Minimax Agent AI |
| Label: AIGC |
| ProduceID: f3e961de220519135b7936401f9c497b |
| PropagateID: f3e961de220519135b7936401f9c497b |
| ReservedCode1: 30450221008b926720cc537a337609a6396807cefd6f2465e1a733f88cb72655e7ed3b5a1e0220073082e844d423175f71300fa33a443d56620f52022574850f68f6c58be981c9 |
| ReservedCode2: 3045022100cee9a5ea6ceee0d1355538f5b52d08108adca91f6b0bd514a775e3cd43616f5e02200b1208fe8656e20f91c6bf8f9d6f4e07d3780abe35035a516e3fe4ffb4de7e6a |
| --- |
| |
| # shenwen-coderV2-Instruct |
|
|
|  |
|
|
| [](https://huggingface.co/shenwenAI/shenwen-coderV2-Instruct)[](https://huggingface.co/shenwenAI/shenwen-coderV2-Instruct)[](https://huggingface.co/shenwenAI/shenwen-coderV2-Instruct) |
|
|
| ## Model Overview |
|
|
| **shenwen-coderV2-Instruct** is an instruction-tuned code generation model based on Qwen2.5-Coder-0.5B-Instruct, optimized for various code generation tasks. |
|
|
| ## Model Details |
|
|
| - **Base Model**: Qwen2.5-Coder-0.5B-Instruct |
| - **Tensor Type**: BF16 |
| - **Parameters**: 0.5B |
| - **Architecture**: qwen2 |
|
|
| ## Usage |
|
|
| ### Using Transformers |
|
|
| ```python |
| from transformers import AutoModelForCausalLM, AutoTokenizer |
| |
| model_name = "shenwenAI/shenwen-coderV2-Instruct" |
| tokenizer = AutoTokenizer.from_pretrained(model_name) |
| model = AutoModelForCausalLM.from_pretrained(model_name) |
| |
| prompt = "Write a Python function to calculate factorial:" |
| inputs = tokenizer(prompt, return_tensors="pt") |
| outputs = model.generate(**inputs, max_new_tokens=512) |
| print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
| ``` |
|
|
| ### Using vLLM |
|
|
| ```python |
| from vllm import LLM, SamplingParams |
| |
| llm = LLM(model="shenwenAI/shenwen-coderV2-Instruct") |
| sampling_params = SamplingParams(temperature=0.8, top_p=0.95, max_tokens=512) |
| |
| prompts = ["Write a Python function to calculate factorial:"] |
| outputs = llm.generate(prompts, sampling_params) |
| print(outputs[0].outputs[0].text) |
| ``` |
|
|
| ## Usage with swllm.cpp (Optimized Code Generation) |
|
|
| For optimized code generation, we recommend using our custom **swllm.cpp** tool: |
|
|
| ```bash |
| # Clone swllm.cpp |
| git clone https://github.com/shenwenAI/swllm.cpp |
| cd swllm.cpp |
| |
| # Build with this model |
| # Convert model to GGUF format first if needed |
| |
| # Run inference |
| ./build/bin/swllm-cli -m path/to/model.gguf -n 512 -p "Write a Python function to calculate factorial:" |
| ``` |
|
|
| **swllm.cpp** provides optimized code generation capabilities for enhanced performance and quality. |
|
|
| ## Quantization |
|
|
| For quantized versions, please visit: [shenwenAI/shenwen-coderV2-GGUF](https://huggingface.co/shenwenAI/shenwen-coderV2-GGUF) |
|
|
| | Quantization | Size | |
| | --- | --- | |
| | Q2_K | 339 MB | |
| | Q4_K_M | 398 MB | |
| | Q5_K_M | 420 MB | |
| | Q8_0 | 531 MB | |
| | F16 | 994 MB | |
|
|
| ## License |
|
|
| Apache 2.0 - See [LICENSE](https://huggingface.co/shenwenAI/shenwen-coderV2-Instruct/blob/main/LICENSE) |
|
|
| ## Acknowledgments |
|
|
| - [Qwen Team](https://github.com/QwenLM/Qwen) for Qwen2.5-Coder |
| - [shenwenAI](https://huggingface.co/shenwenAI) for model training and optimization |
|
|
| ## Connect With Us |
|
|
| - **GitHub**: [https://github.com/shenwenAI](https://github.com/shenwenAI) |
| - **HuggingFace**: [https://huggingface.co/shenwenAI](https://huggingface.co/shenwenAI) |
| - **Twitter/X**: [https://x.com/shenwenai](https://x.com/shenwenai) |
|
|
| --- |
|
|
| *If this model is helpful, please consider giving us a star on GitHub and following us on social media!* |