YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

LoRA for OPT-1.3b, finetuned for chatting. Note that this model is meant to be conversation tuned, NOT instruct tuned.

Trained on a selection of data from https://www.kaggle.com/datasets/jef1056/discord-data

Prompting format:

[Person 1]: [message]
[Person 2]: [message]
[Person 1]:

Example chat:

human: hello
bot: hi how are you?
human: good, you?
bot: I'm doing alright. How about yourself?
human: what are you doing now?
bot: just watching netflix and playing games. What about you?
human: nothing much
bot: same here
human: write me a poem
bot: ok i will do it when im not on my phone

Training params:

transformers.TrainingArguments(
    per_device_train_batch_size=4, 
    fp16=True,
    warmup_steps=30,
    max_steps=200,
    learning_rate=1.7e-4,
    logging_steps=1, 
    output_dir='outputs',
    report_to="none"
)
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support