Spaces:
Configuration error
Configuration error
| --- | |
| # Alpaca Model Card | |
| ## Model details | |
| **Organization developing the model** | |
| Stanford Hashimoto Group | |
| **Model date** | |
| Alpaca was trained in March 2023 | |
| **Model version** | |
| This is version 1 of the model. | |
| **Model type** | |
| Alpaca models are instruction-following models finetuned from LLaMA models. | |
| **More information** | |
| Please see our blog post at `link` for more information. | |
| **Citations details** | |
| Please cite the [github repo](https://github.com/tatsu-lab/stanford_alpaca) if you use the data or code in this repo. | |
| **License** | |
| Code and data are licensed under the Apache 2.0 license. | |
| **Where to send questions or comments about the model** | |
| Questions and comments about LLaMA can be sent via the [GitHub repository](https://github.com/tatsu-lab/stanford_alpaca) of the project, by opening an issue. | |
| ## Intended use | |
| **Primary intended uses** | |
| The primary use of Alpaca is research on instruction following large language models. | |
| **Primary intended users** | |
| The primary intended users of the model are researchers in natural language processing, machine learning and artificial intelligence. | |
| **Out-of-scope use cases** | |
| Alpaca models are not finetuned with human feedback and are not intended for use in production systems. | |
| Alpaca models are trained from data generated using the OpenAI API and thus any usage must not be competing with the OpenAI API. | |
| ## Metrics | |
| **Model performance measures** | |
| the Alpaca 7B model has been evaluated using blinded pairwise comparison with OpenAI's text-davinci-003 on the self-instruct evaluation set. | |
| Our student authors have judged the Alpaca 7B model to be on par with text-davinci-003, with a win rate around 50%. | |
| **Approaches to uncertainty and variability** | |
| We have only finetuned a single Alpaca model at each model size, and thus we do not have a good sense of the variability of the model. | |
| ## Evaluation datasets | |
| The model was evaluated on the self-instruct evaluation set. | |
| ## Training dataset | |
| The model was trained on 52K instruction following data, which is release in the [Github repository](https://github.com/tatsu-lab/stanford_alpaca). |