| base_model: mistralai/Mistral-7B-Instruct-v0.3 | |
| datasets: | |
| - nroggendorff/mayo | |
| language: | |
| - en | |
| license: mit | |
| tags: | |
| - trl | |
| - sft | |
| - sgd | |
| model-index: | |
| - name: mayo | |
| results: [] | |
| # Mayonnaise LLM | |
| Mayo is a language model fine-tuned on the [Mayo dataset](https://huggingface.co/datasets/nroggendorff/mayo) using Supervised Fine-Tuning (SFT) and Teacher Reinforced Learning (TRL) techniques. It is based on the [Mistral 7b Model](mistralai/Mistral-7B-Instruct-v0.3) | |
| ## Features | |
| - Utilizes SFT and TRL techniques for improved performance | |
| - Supports English language | |
| ## Usage | |
| To use the Mayo LLM, you can load the model using the Hugging Face Transformers library: | |
| ```python | |
| from transformers import pipeline | |
| pipe = pipeline("text-generation", model="nroggendorff/mayo") | |
| question = "What color is the sky?" | |
| conv = [{"role": "user", "content": question}] | |
| response = pipe(conv, max_new_tokens=32)[0]['generated_text'][-1]['content'] | |
| print(response) | |
| ``` | |
| ## License | |
| This project is licensed under the MIT License. |