Update README.md
Browse files
README.md
CHANGED
|
@@ -10,6 +10,8 @@ language:
|
|
| 10 |
# About
|
| 11 |
This model is Lightblue's QLoRA finetune of OpenOrca's [Open-Orca/OpenOrcaxOpenChat-Preview2-13B](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B) model on Japanese fine-tuning datasets.
|
| 12 |
|
|
|
|
|
|
|
| 13 |
We trained on equal samples of the following three datasets:
|
| 14 |
* [SNOW](https://huggingface.co/datasets/snow_simplified_japanese_corpus)
|
| 15 |
* [TyDiQA (Ja)](https://huggingface.co/datasets/khalidalt/tydiqa-goldp)
|
|
@@ -20,6 +22,18 @@ which resulted in a dataset of 13,167 samples total.
|
|
| 20 |
These three datasets were chosen as they represent three distinct fine-tuning tasks (Text simplification, question answering, and text summarization, respectively) which we hypothesize can help to improve the language models suitability for dealing with Japanese data.
|
| 21 |
These three datasets make up the model name: STX.
|
| 22 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 23 |
# How to use
|
| 24 |
|
| 25 |
```python
|
|
|
|
| 10 |
# About
|
| 11 |
This model is Lightblue's QLoRA finetune of OpenOrca's [Open-Orca/OpenOrcaxOpenChat-Preview2-13B](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B) model on Japanese fine-tuning datasets.
|
| 12 |
|
| 13 |
+
This model specialises on answering **Closed Question Answering** in Japanese. Input a piece of reference text, ask a question, and see the model answer based on the reference text.
|
| 14 |
+
|
| 15 |
We trained on equal samples of the following three datasets:
|
| 16 |
* [SNOW](https://huggingface.co/datasets/snow_simplified_japanese_corpus)
|
| 17 |
* [TyDiQA (Ja)](https://huggingface.co/datasets/khalidalt/tydiqa-goldp)
|
|
|
|
| 22 |
These three datasets were chosen as they represent three distinct fine-tuning tasks (Text simplification, question answering, and text summarization, respectively) which we hypothesize can help to improve the language models suitability for dealing with Japanese data.
|
| 23 |
These three datasets make up the model name: STX.
|
| 24 |
|
| 25 |
+
With these datasets, we achieve the following scores on the JGLUE benchmark:
|
| 26 |
+
|
| 27 |
+
| Model Name | Open-Orca/OpenOrcaxOpenChat-Preview2-13B | lightblue/openorca_stx |
|
| 28 |
+
|------------------------|------------------------------------------|------------------------|
|
| 29 |
+
| jsquad-1.1-0.3 | 0.692 | 0.836 |
|
| 30 |
+
| jcommonsenseqa-1.1-0.3 | 0.831 | 0.782 |
|
| 31 |
+
| jnli-1.1-0.3 | 0.504 | 0.48 |
|
| 32 |
+
| marc_ja-1.1-0.3 | 0.936 | 0.959 |
|
| 33 |
+
|
| 34 |
+
Our model achieves much better results on the question answering benchmark (JSQuAD) than the base checkpoint without monstrous degradation of performance on multi-choice question benchmarks (JCommonSense, JNLI, MARC-Ja) purely through QLoRA training.
|
| 35 |
+
This shows the potential for applying strong language models such as [Open-Orca/OpenOrcaxOpenChat-Preview2-13B](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B) to minimal QLoRA fine-tuning using Japanese fine-tuning datasets to achieve better results at narrow NLP tasks.
|
| 36 |
+
|
| 37 |
# How to use
|
| 38 |
|
| 39 |
```python
|