| LoRA for OPT-1.3b, finetuned for chatting. Note that this model is meant to be conversation tuned, NOT instruct tuned. | |
| Trained on a selection of data from https://www.kaggle.com/datasets/jef1056/discord-data | |
| Prompting format: | |
| ``` | |
| [Person 1]: [message] | |
| [Person 2]: [message] | |
| [Person 1]: | |
| ``` | |
| Example chat: | |
| ``` | |
| human: hello | |
| bot: hi how are you? | |
| human: good, you? | |
| bot: I'm doing alright. How about yourself? | |
| human: what are you doing now? | |
| bot: just watching netflix and playing games. What about you? | |
| human: nothing much | |
| bot: same here | |
| human: write me a poem | |
| bot: ok i will do it when im not on my phone | |
| ``` | |
| Training params: | |
| ```py | |
| transformers.TrainingArguments( | |
| per_device_train_batch_size=4, | |
| fp16=True, | |
| warmup_steps=30, | |
| max_steps=200, | |
| learning_rate=1.7e-4, | |
| logging_steps=1, | |
| output_dir='outputs', | |
| report_to="none" | |
| ) | |
| ``` |