wall-e-zz commited on
Commit
6cfb019
·
1 Parent(s): 6571a73

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -23
README.md CHANGED
@@ -9,30 +9,12 @@ app_file: app.py
9
  pinned: false
10
  ---
11
 
12
- # Configuration
13
 
14
- `title`: _string_
15
- Display title for the Space
16
 
17
- `emoji`: _string_
18
- Space emoji (emoji-only character allowed)
19
 
20
- `colorFrom`: _string_
21
- Color for Thumbnail gradient (red, yellow, green, blue, indigo, purple, pink, gray)
22
 
23
- `colorTo`: _string_
24
- Color for Thumbnail gradient (red, yellow, green, blue, indigo, purple, pink, gray)
25
-
26
- `sdk`: _string_
27
- Can be either `gradio` or `streamlit`
28
-
29
- `sdk_version` : _string_
30
- Only applicable for `streamlit` SDK.
31
- See [doc](https://hf.co/docs/hub/spaces) for more info on supported versions.
32
-
33
- `app_file`: _string_
34
- Path to your main application file (which contains either `gradio` or `streamlit` Python code).
35
- Path is relative to the root of the repository.
36
-
37
- `pinned`: _boolean_
38
- Whether the Space stays on top of your list.
 
9
  pinned: false
10
  ---
11
 
12
+ # mBART-50
13
 
14
+ mBART-50 is a multilingual Sequence-to-Sequence model pre-trained using the "Multilingual Denoising Pretraining" objective. It was introduced in [Multilingual Translation with Extensible Multilingual Pretraining and Finetuning](https://arxiv.org/abs/2008.00401) paper.
 
15
 
16
+ ## Model description
 
17
 
18
+ mBART-50 is a multilingual Sequence-to-Sequence model. It was introduced to show that multilingual translation models can be created through multilingual fine-tuning.
19
+ Instead of fine-tuning on one direction, a pre-trained model is fine-tuned on many directions simultaneously. mBART-50 is created using the original mBART model and extended to add extra 25 languages to support multilingual machine translation models of 50 languages. The pre-training objective is explained below.
20