Update README.md
Browse files
README.md
CHANGED
|
@@ -18,11 +18,11 @@ widget:
|
|
| 18 |
This model is based on the [facebook/mbart-large-50](https://huggingface.co/facebook/mbart-large-50) architecture, renowned for its effectiveness in translation and text generation tasks across numerous languages.
|
| 19 |
|
| 20 |
## Training Data
|
| 21 |
-
The model was fine-tuned on a subset of
|
| 22 |
Dataset [skypro1111/ubertext-2-news-verbalized](https://huggingface.co/datasets/skypro1111/ubertext-2-news-verbalized)
|
| 23 |
|
| 24 |
## Training Procedure
|
| 25 |
-
The model underwent
|
| 26 |
|
| 27 |
```python
|
| 28 |
from transformers import MBartForConditionalGeneration, AutoTokenizer, Trainer, TrainingArguments
|
|
|
|
| 18 |
This model is based on the [facebook/mbart-large-50](https://huggingface.co/facebook/mbart-large-50) architecture, renowned for its effectiveness in translation and text generation tasks across numerous languages.
|
| 19 |
|
| 20 |
## Training Data
|
| 21 |
+
The model was fine-tuned on a subset of 457,610 sentences from the Ubertext dataset, focusing on news content. The verbalized equivalents were created using Google Gemini Pro, providing a rich basis for learning text transformation tasks.
|
| 22 |
Dataset [skypro1111/ubertext-2-news-verbalized](https://huggingface.co/datasets/skypro1111/ubertext-2-news-verbalized)
|
| 23 |
|
| 24 |
## Training Procedure
|
| 25 |
+
The model underwent 410,000 training steps.
|
| 26 |
|
| 27 |
```python
|
| 28 |
from transformers import MBartForConditionalGeneration, AutoTokenizer, Trainer, TrainingArguments
|