Commit
·
8600bb6
1
Parent(s):
00d361e
Update README.md
Browse files
README.md
CHANGED
|
@@ -3,13 +3,13 @@ datasets:
|
|
| 3 |
- Siddharth63/biological_dataset
|
| 4 |
license: artistic-2.0
|
| 5 |
---
|
| 6 |
-
Bioul2-tiny-nl6
|
| 7 |
|
| 8 |
-
Pretrained T5 model on
|
| 9 |
|
| 10 |
Note: The Hugging Face inference widget is deactivated because this model needs a text-to-text fine-tuning on a specific downstream task to be useful in practice.
|
| 11 |
|
| 12 |
-
Model description
|
| 13 |
T5 is an encoder-decoder model and treats all NLP problems in a text-to-text format.
|
| 14 |
|
| 15 |
Finnish T5 is a transformers model pretrained on a very large corpus of Finnish data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and outputs from those texts.
|
|
@@ -28,7 +28,7 @@ In total, this model has 31 million parameters.
|
|
| 28 |
|
| 29 |
|
| 30 |
|
| 31 |
-
UL2 pretraining objective
|
| 32 |
This model was pretrained with the UL2's Mixture-of-Denoisers (MoD) objective, that combines diverse pre-training paradigms together. UL2 frames different objective functions for training language models as denoising tasks, where the model has to recover missing sub-sequences of a given input. During pre-training it uses a novel mixture-of-denoisers that samples from a varied set of such objectives, each with different configurations. UL2 is trained using a mixture of three denoising tasks: (1) R-denoising (or regular span corruption), which emulates the standard T5 span corruption objective; (2) X-denoising (or extreme span corruption); and (3) S-denoising (or sequential PrefixLM). During pre-training, we sample from the available denoising tasks based on user-specified ratios.
|
| 33 |
|
| 34 |
UL2 introduces a notion of mode switching, wherein downstream fine-tuning is associated with specific pre-training denoising task. During the pretraining, a paradigm token is inserted to the input ([NLU] for R-denoising, [NLG] for X-denoising, or [S2S] for S-denoising) indicating the denoising task at hand. Then, during fine-tuning the same input token should be inserted to get the best performance for different downstream fine-tuning tasks.
|
|
@@ -38,6 +38,6 @@ This model was only pretrained in a self-supervised way excluding any supervised
|
|
| 38 |
|
| 39 |
Note: For fine-tuning, most likely you can get better results if you insert a prefix token of [NLU], [NLG], or [S2S] to your input texts. For general language understanding fine-tuning tasks, you could use the [NLU] token. For GPT-style causal language generation, you could use the [S2S] token. The token [NLG] of the X-denoising pretrain task is somewhat mix between the language understanding and causal language generation so the token [NLG] could maybe be used for language generation fine-tuning too.
|
| 40 |
|
| 41 |
-
Acknowledgements
|
| 42 |
-
This project would not have been possible without compute generously provided by Google through the TPU Research Cloud. Thanks to the Finnish-NLP authors for releasing their code for the UL2 objective and associated task definitions as well as their guidance. Thanks to Yeb Havinga for helping me get started with the t5x framework.
|
| 43 |
|
|
|
|
| 3 |
- Siddharth63/biological_dataset
|
| 4 |
license: artistic-2.0
|
| 5 |
---
|
| 6 |
+
# Bioul2-tiny-nl6
|
| 7 |
|
| 8 |
+
Pretrained T5 model on Biological dataset using a UL2 (Mixture-of-Denoisers) objective. T5 model was introduced in this paper and first released at this page. The UL2 objective was introduced in this paper and first released at this page.
|
| 9 |
|
| 10 |
Note: The Hugging Face inference widget is deactivated because this model needs a text-to-text fine-tuning on a specific downstream task to be useful in practice.
|
| 11 |
|
| 12 |
+
## Model description
|
| 13 |
T5 is an encoder-decoder model and treats all NLP problems in a text-to-text format.
|
| 14 |
|
| 15 |
Finnish T5 is a transformers model pretrained on a very large corpus of Finnish data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and outputs from those texts.
|
|
|
|
| 28 |
|
| 29 |
|
| 30 |
|
| 31 |
+
## UL2 pretraining objective
|
| 32 |
This model was pretrained with the UL2's Mixture-of-Denoisers (MoD) objective, that combines diverse pre-training paradigms together. UL2 frames different objective functions for training language models as denoising tasks, where the model has to recover missing sub-sequences of a given input. During pre-training it uses a novel mixture-of-denoisers that samples from a varied set of such objectives, each with different configurations. UL2 is trained using a mixture of three denoising tasks: (1) R-denoising (or regular span corruption), which emulates the standard T5 span corruption objective; (2) X-denoising (or extreme span corruption); and (3) S-denoising (or sequential PrefixLM). During pre-training, we sample from the available denoising tasks based on user-specified ratios.
|
| 33 |
|
| 34 |
UL2 introduces a notion of mode switching, wherein downstream fine-tuning is associated with specific pre-training denoising task. During the pretraining, a paradigm token is inserted to the input ([NLU] for R-denoising, [NLG] for X-denoising, or [S2S] for S-denoising) indicating the denoising task at hand. Then, during fine-tuning the same input token should be inserted to get the best performance for different downstream fine-tuning tasks.
|
|
|
|
| 38 |
|
| 39 |
Note: For fine-tuning, most likely you can get better results if you insert a prefix token of [NLU], [NLG], or [S2S] to your input texts. For general language understanding fine-tuning tasks, you could use the [NLU] token. For GPT-style causal language generation, you could use the [S2S] token. The token [NLG] of the X-denoising pretrain task is somewhat mix between the language understanding and causal language generation so the token [NLG] could maybe be used for language generation fine-tuning too.
|
| 40 |
|
| 41 |
+
## Acknowledgements
|
| 42 |
+
This project would not have been possible without compute generously provided by Google through the [Google TPU Research Cloud](https://sites.research.google/trc/about/). Thanks to the Finnish-NLP authors for releasing their code for the UL2 objective and associated task definitions as well as their guidance. Thanks to Yeb Havinga for helping me get started with the t5x framework.
|
| 43 |
|