Update README.md
Browse files
README.md
CHANGED
|
@@ -127,7 +127,7 @@ metrics:
|
|
| 127 |
|
| 128 |
- **Developed by:** Cohere For AI
|
| 129 |
- **Model type:** a Transformer style autoregressive massively multilingual language model.
|
| 130 |
-
- **Paper**: [Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model](arxiv.
|
| 131 |
- **Point of Contact**: Cohere For AI: [cohere.for.ai](cohere.for.ai)
|
| 132 |
- **Languages**: Refer to the list of languages in the `language` section of this model card.
|
| 133 |
- **License**: Apache-2.0
|
|
@@ -180,16 +180,16 @@ The Aya model is trained on the following datasets:
|
|
| 180 |
- [DataProvenance collection](https://huggingface.co/datasets/DataProvenanceInitiative/Commercially-Verified-Licenses)
|
| 181 |
- ShareGPT-Command
|
| 182 |
|
| 183 |
-
All datasets are subset to the 101 languages supported by [mT5]. See the [paper](arxiv.
|
| 184 |
|
| 185 |
## Evaluation
|
| 186 |
|
| 187 |
-
We refer to Section 5 from our paper for multilingual eval across 99 languages – including discriminative
|
| 188 |
|
| 189 |
## Bias, Risks, and Limitations
|
| 190 |
|
| 191 |
|
| 192 |
-
For a detailed overview of our effort at safety mitigation and benchmarking toxicity and bias across multiple languages, we refer Sections 6 and 7 of our paper: [Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model](arxiv.
|
| 193 |
|
| 194 |
We hope that the release of the Aya model will make community-based redteaming efforts possible, by exposing an open-source massively-multilingual model for community research.
|
| 195 |
|
|
|
|
| 127 |
|
| 128 |
- **Developed by:** Cohere For AI
|
| 129 |
- **Model type:** a Transformer style autoregressive massively multilingual language model.
|
| 130 |
+
- **Paper**: [Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model](https://arxiv.org/abs/2402.07827)
|
| 131 |
- **Point of Contact**: Cohere For AI: [cohere.for.ai](cohere.for.ai)
|
| 132 |
- **Languages**: Refer to the list of languages in the `language` section of this model card.
|
| 133 |
- **License**: Apache-2.0
|
|
|
|
| 180 |
- [DataProvenance collection](https://huggingface.co/datasets/DataProvenanceInitiative/Commercially-Verified-Licenses)
|
| 181 |
- ShareGPT-Command
|
| 182 |
|
| 183 |
+
All datasets are subset to the 101 languages supported by [mT5]. See the [paper](https://arxiv.org/abs/2402.07827) for details about filtering and pruning.
|
| 184 |
|
| 185 |
## Evaluation
|
| 186 |
|
| 187 |
+
We refer to Section 5 from our paper for multilingual eval across 99 languages – including discriminative and generative tasks, human evaluation, and simulated win rates that cover both held-out tasks and in-distribution performance.
|
| 188 |
|
| 189 |
## Bias, Risks, and Limitations
|
| 190 |
|
| 191 |
|
| 192 |
+
For a detailed overview of our effort at safety mitigation and benchmarking toxicity and bias across multiple languages, we refer to Sections 6 and 7 of our paper: [Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model](https://arxiv.org/abs/2402.07827).
|
| 193 |
|
| 194 |
We hope that the release of the Aya model will make community-based redteaming efforts possible, by exposing an open-source massively-multilingual model for community research.
|
| 195 |
|