Commit ·
8712ba2
1
Parent(s): 0a82d0d
Update README.md
Browse files
README.md
CHANGED
|
@@ -23,14 +23,14 @@ We open-source Orca 2 to encourage further research on the development, evaluati
|
|
| 23 |
|
| 24 |
## How was Orca 2 evaluated?
|
| 25 |
|
| 26 |
-
+ Orca 2 has been evaluated on a large number of tasks ranging from reasoning to safety. Please refer to Section 6 and Appendix in the paper for details on evaluations.
|
| 27 |
|
| 28 |
## Model Details
|
| 29 |
|
| 30 |
-
Orca 2 is a finetuned version of LLAMA-2. Orca 2’s training data is a synthetic dataset that was created to enhance the small model’s reasoning abilities. All synthetic training data was
|
| 31 |
More details about the model can be found at: LINK to Tech Report
|
| 32 |
|
| 33 |
-
|
| 34 |
|
| 35 |
## License
|
| 36 |
|
|
@@ -41,7 +41,7 @@ Llama 2 is licensed under the [LLAMA 2 Community License](https://ai.meta.com/ll
|
|
| 41 |
## Bias, Risks, and Limitations
|
| 42 |
|
| 43 |
Orca 2, built upon the LLaMA 2 model family, retains many of its limitations, as well as the
|
| 44 |
-
common limitations of other large language models or limitation
|
| 45 |
process, including:
|
| 46 |
|
| 47 |
**Data Biases**: Large language models, trained on extensive data, can inadvertently carry
|
|
@@ -148,9 +148,9 @@ print(second_turn_answer)
|
|
| 148 |
**Safe inference with Azure AI Content Safety**
|
| 149 |
|
| 150 |
The usage of [Azure AI Content Safety](https://azure.microsoft.com/en-us/products/ai-services/ai-content-safety/) on top of model prediction is strongly encouraged
|
| 151 |
-
and can help
|
| 152 |
-
that uses AI to
|
| 153 |
-
|
| 154 |
self-harm with multiple severity levels and multi-lingual detection.
|
| 155 |
|
| 156 |
```python
|
|
|
|
| 23 |
|
| 24 |
## How was Orca 2 evaluated?
|
| 25 |
|
| 26 |
+
+ Orca 2 has been evaluated on a large number of tasks ranging from reasoning to grounding and safety. Please refer to Section 6 and Appendix in the paper for details on evaluations.
|
| 27 |
|
| 28 |
## Model Details
|
| 29 |
|
| 30 |
+
Orca 2 is a finetuned version of LLAMA-2. Orca 2’s training data is a synthetic dataset that was created to enhance the small model’s reasoning abilities. All synthetic training data was moderated using the Microsoft Azure content filters.
|
| 31 |
More details about the model can be found at: LINK to Tech Report
|
| 32 |
|
| 33 |
+
Please refer to LLaMA-2 technical report for details on the model architecture.
|
| 34 |
|
| 35 |
## License
|
| 36 |
|
|
|
|
| 41 |
## Bias, Risks, and Limitations
|
| 42 |
|
| 43 |
Orca 2, built upon the LLaMA 2 model family, retains many of its limitations, as well as the
|
| 44 |
+
common limitations of other large language models or limitation caused by its training
|
| 45 |
process, including:
|
| 46 |
|
| 47 |
**Data Biases**: Large language models, trained on extensive data, can inadvertently carry
|
|
|
|
| 148 |
**Safe inference with Azure AI Content Safety**
|
| 149 |
|
| 150 |
The usage of [Azure AI Content Safety](https://azure.microsoft.com/en-us/products/ai-services/ai-content-safety/) on top of model prediction is strongly encouraged
|
| 151 |
+
and can help preventing some of content harms. Azure AI Content Safety is a content moderation platform
|
| 152 |
+
that uses AI to moderate content. By having Azure AI Content Safety on the output of Orca 2,
|
| 153 |
+
the model output can be moderated by scanning it for different harm categories including sexual content, violence, hate, and
|
| 154 |
self-harm with multiple severity levels and multi-lingual detection.
|
| 155 |
|
| 156 |
```python
|