Update README.md
Browse files
README.md
CHANGED
|
@@ -25,18 +25,60 @@ It achieves the following results on the evaluation set:
|
|
| 25 |
|
| 26 |
## Model description
|
| 27 |
|
| 28 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 29 |
|
| 30 |
## Intended uses & limitations
|
| 31 |
|
| 32 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
|
| 34 |
## Training and evaluation data
|
| 35 |
|
| 36 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 37 |
|
| 38 |
## Training procedure
|
| 39 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 40 |
### Training hyperparameters
|
| 41 |
|
| 42 |
The following hyperparameters were used during training:
|
|
|
|
| 25 |
|
| 26 |
## Model description
|
| 27 |
|
| 28 |
+
This model is a fine-tuned version of facebook/mbart-large-50, a multilingual sequence-to-sequence Transformer model, adapted for the task of Spanish gender neutralization.
|
| 29 |
+
The goal of the model is to transform gender-marked Spanish sentences into gender-neutral reformulations, preserving meaning while reducing grammatical gender marking. This task can be framed as a monolingual translation problem (Spanish → neutral Spanish).
|
| 30 |
+
The model was trained using the Hugging Face Transformers library and follows a standard encoder–decoder architecture with transfer learning from the pretrained mBART model.
|
| 31 |
+
The resulting system performs controlled rewriting rather than translation between languages, making it suitable for experiments in:
|
| 32 |
+
|
| 33 |
+
- inclusive language generation
|
| 34 |
+
- stylistic rewriting
|
| 35 |
+
- bias reduction in text
|
| 36 |
+
- controlled text transformation
|
| 37 |
|
| 38 |
## Intended uses & limitations
|
| 39 |
|
| 40 |
+
This model is intended for:
|
| 41 |
+
- Research experiments in NLP and inclusive language
|
| 42 |
+
- Educational purposes in courses on Machine Translation or Text Generation
|
| 43 |
+
- Demonstrations of transfer learning using multilingual seq2seq models
|
| 44 |
+
- Automatic rewriting of short Spanish sentences into gender-neutral forms
|
| 45 |
|
| 46 |
## Training and evaluation data
|
| 47 |
|
| 48 |
+
The model was trained on the Spanish Gender Neutralization dataset available on Hugging Face:
|
| 49 |
+
👉 hackathon-pln-es/neutral-es
|
| 50 |
+
This dataset contains pairs of aligned sentences:
|
| 51 |
+
- gendered: original sentence with grammatical gender marking
|
| 52 |
+
- neutral: reformulated gender-neutral version
|
| 53 |
+
|
| 54 |
+
The dataset already includes a predefined split:
|
| 55 |
+
- Training set
|
| 56 |
+
- Test set
|
| 57 |
+
|
| 58 |
+
The dataset is relatively small and designed mainly for educational and experimental purposes, not for large-scale production systems.
|
| 59 |
+
Before training, the data was:
|
| 60 |
+
- tokenized using the mBART tokenizer
|
| 61 |
+
- truncated/padded to model limits
|
| 62 |
+
- converted into input/label format for seq2seq training
|
| 63 |
+
|
| 64 |
+
Evaluation was performed using the BLEU score (sacrebleu), a standard metric in machine translation.
|
| 65 |
|
| 66 |
## Training procedure
|
| 67 |
|
| 68 |
+
The model was trained using the Hugging Face Trainer API for sequence-to-sequence learning.
|
| 69 |
+
Training steps:
|
| 70 |
+
1. The pretrained model facebook/mbart-large-50 was loaded.
|
| 71 |
+
2. The dataset was tokenized using the corresponding mBART tokenizer.
|
| 72 |
+
3. Inputs were formatted as:
|
| 73 |
+
- source: gendered sentence
|
| 74 |
+
- target: neutral sentence
|
| 75 |
+
4. The model was fine-tuned using transfer learning.
|
| 76 |
+
5. Training was performed on GPU in Google Colab.
|
| 77 |
+
6. Evaluation during training used the sacrebleu metric.
|
| 78 |
+
7. The final model was uploaded to the Hugging Face Hub.
|
| 79 |
+
|
| 80 |
+
The model therefore learns to perform monolingual rewriting via multilingual translation architecture.
|
| 81 |
+
|
| 82 |
### Training hyperparameters
|
| 83 |
|
| 84 |
The following hyperparameters were used during training:
|