How to use from the
Use from the
Transformers library
# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("text-generation", model="deca-ai/2-pro-coder")
messages = [
    {"role": "user", "content": "Who are you?"},
]
pipe(messages)
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("deca-ai/2-pro-coder")
model = AutoModelForCausalLM.from_pretrained("deca-ai/2-pro-coder")
messages = [
    {"role": "user", "content": "Who are you?"},
]
inputs = tokenizer.apply_chat_template(
	messages,
	add_generation_prompt=True,
	tokenize=True,
	return_dict=True,
	return_tensors="pt",
).to(model.device)

outputs = model.generate(**inputs, max_new_tokens=40)
print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:]))
Quick Links

Deca 2 Banner The Deca 2 PRO model, currently in BETA, is built on cutting-edge architectures like Perplexity's R1 1776, LLaMA 3, and Qwen 2, delivering extraordinary performance. With a focus on insane speed and high efficiency, Deca 2 PRO is revolutionizing text generation and setting new standards in the industry.

As more capabilities are added, Deca 2 PRO will evolve into a more powerful, any-to-any model in the future. While it’s focused on text generation for now, its foundation is designed to scale, bringing even more advanced functionalities to come.

This model is trained on code-related tasks.

Downloads last month
7
Safetensors
Model size
69B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for deca-ai/2-pro-coder

Finetuned
(1)
this model