How to use from
SGLang
Install from pip and serve model
# Install SGLang from pip:
pip install sglang
# Start the SGLang server:
python3 -m sglang.launch_server \
    --model-path "HelpingAI/Cipher-20B" \
    --host 0.0.0.0 \
    --port 30000
# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:30000/v1/chat/completions" \
	-H "Content-Type: application/json" \
	--data '{
		"model": "HelpingAI/Cipher-20B",
		"messages": [
			{
				"role": "user",
				"content": "What is the capital of France?"
			}
		]
	}'
Use Docker images
docker run --gpus all \
    --shm-size 32g \
    -p 30000:30000 \
    -v ~/.cache/huggingface:/root/.cache/huggingface \
    --env "HF_TOKEN=<secret>" \
    --ipc=host \
    lmsysorg/sglang:latest \
    python3 -m sglang.launch_server \
        --model-path "HelpingAI/Cipher-20B" \
        --host 0.0.0.0 \
        --port 30000
# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:30000/v1/chat/completions" \
	-H "Content-Type: application/json" \
	--data '{
		"model": "HelpingAI/Cipher-20B",
		"messages": [
			{
				"role": "user",
				"content": "What is the capital of France?"
			}
		]
	}'
Quick Links
💻 Cipher-20B
GitHub Organization Hugging Face Model License Join Community Discussion
[📜 License](https://helpingai.co/license) | [🌐 Website](https://helpingai.co)
Model Size Task Deployment Speed

🌟 About Cipher-20B

Cipher-20B is a 20 billion parameter causal language model designed for code generation.

💻 Implementation

Using Transformers

from transformers import AutoModelForCausalLM, AutoTokenizer

# Load Cipher-20B
model = AutoModelForCausalLM.from_pretrained("HelpingAI/Cipher-20B")
tokenizer = AutoTokenizer.from_pretrained("HelpingAI/Cipher-20B")

# Example usage
code_task = [
    {"role": "system", "content": "You are Cipher"},
    {"role": "user", "content": "Write a Python function to calculate the Fibonacci sequence."}
]

inputs = tokenizer.apply_chat_template(
    code_task,
    add_generation_prompt=True,
    return_tensors="pt"
)

outputs = model.generate(
    inputs,
    max_new_tokens=256,
    temperature=0.7,
    top_p=0.9,
)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))

⚙️ Training Details

Training Data

  • Trained on a large dataset of code, programming tasks, and technical documentation.
  • Fine-tuned for multiple programming languages like Python, JavaScript, and C++.

Capabilities

  • Generates code in multiple languages.
  • Detects and corrects common coding errors.
  • Provides clear explanations of code.

⚠️ Limitations

  • May generate verbose code depending on the input.
  • Long code generation may exceed token limits.
  • Ambiguous instructions can lead to incomplete or incorrect code.
  • Prioritizes efficiency in code generation.

Safety

  • Avoids generating harmful or malicious code.
  • Will not assist with illegal or unethical activities.

📚 Citation

@misc{cipher2024,
  author = {Abhay Koul},
  title = {Cipher-20B: Your Ultimate Code Buddy},
  year = {2024},
  publisher = {HelpingAI},
  journal = {HuggingFace},
  howpublished = {\url{https://huggingface.co/HelpingAI/Cipher-20B}}
}

Built with dedication, precision, and passion by HelpingAI

WebsiteGitHubDiscordHuggingFace

Downloads last month
16
Safetensors
Model size
21B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for HelpingAI/Cipher-20B

Quantizations
2 models