--- language: en tags: - deepseek - llama - transformers license: apache-2.0 --- # DeepSeek Model This is a converted version of the DeepSeek model. ## Model Description - **Model Type:** Causal Language Model - **Language:** English - **Base Architecture:** LLaMA - **Context Length:** 2048 tokens - **Parameters:** Custom implementation ## Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer model = AutoModelForCausalLM.from_pretrained("ashwinij2/deepseek-llama-converted") tokenizer = AutoTokenizer.from_pretrained("ashwinij2/deepseek-llama-converted") text = "Hello, how are you?" inputs = tokenizer(text, return_tensors="pt") outputs = model.generate(**inputs, max_length=50) print(tokenizer.decode(outputs[0])) ```