--- language: en tags: - conversational - roleplay - dippy - dialogpt - bittensor license: mit --- # Dippy DialoGPT Optimized This is a fine-tuned version of microsoft/DialoGPT-medium optimized for conversational AI with Dippy personality. ## Model Details - Base model: microsoft/DialoGPT-medium - Fine-tuned for: Conversational AI, roleplay, helpful assistant interactions - Optimized for: Bittensor SN11 Dippy subnet ## Usage ```python from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("Gatescrispy/dippy-dialogpt-optimized") model = AutoModelForCausalLM.from_pretrained("Gatescrispy/dippy-dialogpt-optimized") # Generate response inputs = tokenizer.encode("Hello! How are you today?", return_tensors="pt") outputs = model.generate(inputs, max_length=50, pad_token_id=tokenizer.eos_token_id) response = tokenizer.decode(outputs[0], skip_special_tokens=True) print(response) ``` ## Training - Dataset: Custom Dippy personality conversations - Training: 1 epoch with learning rate scheduling - Hardware: NVIDIA RTX 3090 ## Bittensor Integration This model is designed for Bittensor SN11 Dippy subnet integration.