shenwen-coderV2-Instruct

Hugging Face

ModelFormatLicense

Model Overview

shenwen-coderV2-Instruct is an instruction-tuned code generation model based on Qwen2.5-Coder-0.5B-Instruct, optimized for various code generation tasks.

Model Details

  • Base Model: Qwen2.5-Coder-0.5B-Instruct
  • Tensor Type: BF16
  • Parameters: 0.5B
  • Architecture: qwen2

Usage

Using Transformers

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "shenwenAI/shenwen-coderV2-Instruct"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

prompt = "Write a Python function to calculate factorial:"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Using vLLM

from vllm import LLM, SamplingParams

llm = LLM(model="shenwenAI/shenwen-coderV2-Instruct")
sampling_params = SamplingParams(temperature=0.8, top_p=0.95, max_tokens=512)

prompts = ["Write a Python function to calculate factorial:"]
outputs = llm.generate(prompts, sampling_params)
print(outputs[0].outputs[0].text)

Usage with swllm.cpp (Optimized Code Generation)

For optimized code generation, we recommend using our custom swllm.cpp tool:

# Clone swllm.cpp
git clone https://github.com/shenwenAI/swllm.cpp
cd swllm.cpp

# Build with this model
# Convert model to GGUF format first if needed

# Run inference
./build/bin/swllm-cli -m path/to/model.gguf -n 512 -p "Write a Python function to calculate factorial:"

swllm.cpp provides optimized code generation capabilities for enhanced performance and quality.

Quantization

For quantized versions, please visit: shenwenAI/shenwen-coderV2-GGUF

Quantization Size
Q2_K 339 MB
Q4_K_M 398 MB
Q5_K_M 420 MB
Q8_0 531 MB
F16 994 MB

License

Apache 2.0 - See LICENSE

Acknowledgments

Connect With Us


If this model is helpful, please consider giving us a star on GitHub and following us on social media!

Downloads last month
83
Safetensors
Model size
0.5B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for shenwenAI/shenwen-coderV2-Instruct

Quantizations
1 model