Text Generation
Transformers
Safetensors
English
mistral
conversational
custom_code
text-generation-inference
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("adalbertojunior/DUSMistral", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("adalbertojunior/DUSMistral", trust_remote_code=True)
messages = [
{"role": "user", "content": "Who are you?"},
]
inputs = tokenizer.apply_chat_template(
messages,
add_generation_prompt=True,
tokenize=True,
return_dict=True,
return_tensors="pt",
).to(model.device)
outputs = model.generate(**inputs, max_new_tokens=40)
print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:]))Quick Links
This model draws inspiration from SOLAR, but introduces a novel approach to increasing the model's depth without the traditional method of duplicating layers. By rearranging the order of layers during inference, it maintains the advantages of depth upscaling while preserving the original parameter count. Furthermore, it undergoes additional fine-tuning using the Dolphin dataset. The foundational architecture for this experiment is based on Dolphin.
Use
# pip install transformers
from transformers import AutoTokenizer, AutoModelForCausalLM
model_id = "adalbertojunior/DUSMistral"
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(model_id, trust_remote_code=True)
# Format message with the CHATML chat template
messages = [{"role": "user", "content": "Hello, how are you?"}]
input_ids = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt")
gen_tokens = model.generate(
input_ids,
max_new_tokens=100,
do_sample=True,
temperature=0.3,
)
gen_text = tokenizer.decode(gen_tokens[0])
print(gen_text)
- Downloads last month
- 84
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="adalbertojunior/DUSMistral", trust_remote_code=True) messages = [ {"role": "user", "content": "Who are you?"}, ] pipe(messages)