File size: 1,176 Bytes
786d288 266decc 786d288 266decc 786d288 266decc 786d288 266decc 786d288 266decc 786d288 266decc 786d288 266decc 786d288 266decc 786d288 266decc 786d288 266decc |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 |
---
language: en
tags:
- conversational
- roleplay
- dippy
- dialogpt
- bittensor
license: mit
---
# Dippy DialoGPT Optimized
This is a fine-tuned version of microsoft/DialoGPT-medium optimized for conversational AI with Dippy personality.
## Model Details
- Base model: microsoft/DialoGPT-medium
- Fine-tuned for: Conversational AI, roleplay, helpful assistant interactions
- Optimized for: Bittensor SN11 Dippy subnet
## Usage
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("Gatescrispy/dippy-dialogpt-optimized")
model = AutoModelForCausalLM.from_pretrained("Gatescrispy/dippy-dialogpt-optimized")
# Generate response
inputs = tokenizer.encode("Hello! How are you today?", return_tensors="pt")
outputs = model.generate(inputs, max_length=50, pad_token_id=tokenizer.eos_token_id)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
```
## Training
- Dataset: Custom Dippy personality conversations
- Training: 1 epoch with learning rate scheduling
- Hardware: NVIDIA RTX 3090
## Bittensor Integration
This model is designed for Bittensor SN11 Dippy subnet integration.
|