DeepSeek Model

This is a converted version of the DeepSeek model.

Model Description

  • Model Type: Causal Language Model
  • Language: English
  • Base Architecture: LLaMA
  • Context Length: 2048 tokens
  • Parameters: Custom implementation

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("ashwinij2/deepseek-llama-converted")
tokenizer = AutoTokenizer.from_pretrained("ashwinij2/deepseek-llama-converted")

text = "Hello, how are you?"
inputs = tokenizer(text, return_tensors="pt")
outputs = model.generate(**inputs, max_length=50)
print(tokenizer.decode(outputs[0]))
Downloads last month
5
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support