--- base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B library_name: peft pipeline_tag: text-generation language: en tags: - deepseek - text-generation - conversational --- # DeepSeek Chatbot This is a fine-tuned version of DeepSeek-R1-Distill-Qwen-1.5B, optimized for conversational AI applications. The model maintains the base model's capabilities while being tuned for improved dialogue interactions. ## Model Details ### Model Description - **Developed by:** Trinoid - **Model type:** Conversational Language Model - **Language(s):** English - **License:** Same as base model (DeepSeek-R1-Distill-Qwen-1.5B) - **Finetuned from model:** deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B ## Uses ### Direct Use This model can be used for: - General conversation - Text generation - Question answering - Chat-based applications Example usage: ```python from huggingface_hub import InferenceClient client = InferenceClient("Trinoid/Deepseek_Chatbot") messages = [ {"role": "system", "content": "You are a helpful assistant."}, {"role": "user", "content": "Hello, how are you?"} ] response = client.chat_completion( messages, max_tokens=512, temperature=0.7, top_p=0.95 ) ``` ### Out-of-Scope Use This model should not be used for: - Generation of harmful or malicious content - Spreading misinformation - Production of illegal content - Making critical decisions without human oversight ## Training Details ### Training Procedure #### Training Hyperparameters - **Training regime:** fp16 mixed precision - **Framework:** PEFT (Parameter-Efficient Fine-Tuning) - **PEFT Method:** LoRA - **Version:** PEFT 0.14.0 ## Technical Specifications ### Model Architecture and Objective - Base architecture: DeepSeek-R1-Distill-Qwen-1.5B - Fine-tuning method: PEFT/LoRA - Primary objective: Conversational AI ### Compute Infrastructure #### Software - PEFT 0.14.0 - Transformers - Python 3.x ## Model Card Contact For questions or issues about this model, please open an issue in the model repository.