CarolGuga commited on
Commit
bd04132
·
verified ·
1 Parent(s): 045351d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +45 -3
README.md CHANGED
@@ -25,18 +25,60 @@ It achieves the following results on the evaluation set:
25
 
26
  ## Model description
27
 
28
- More information needed
 
 
 
 
 
 
 
 
29
 
30
  ## Intended uses & limitations
31
 
32
- More information needed
 
 
 
 
33
 
34
  ## Training and evaluation data
35
 
36
- More information needed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
37
 
38
  ## Training procedure
39
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
  ### Training hyperparameters
41
 
42
  The following hyperparameters were used during training:
 
25
 
26
  ## Model description
27
 
28
+ This model is a fine-tuned version of facebook/mbart-large-50, a multilingual sequence-to-sequence Transformer model, adapted for the task of Spanish gender neutralization.
29
+ The goal of the model is to transform gender-marked Spanish sentences into gender-neutral reformulations, preserving meaning while reducing grammatical gender marking. This task can be framed as a monolingual translation problem (Spanish → neutral Spanish).
30
+ The model was trained using the Hugging Face Transformers library and follows a standard encoder–decoder architecture with transfer learning from the pretrained mBART model.
31
+ The resulting system performs controlled rewriting rather than translation between languages, making it suitable for experiments in:
32
+
33
+ - inclusive language generation
34
+ - stylistic rewriting
35
+ - bias reduction in text
36
+ - controlled text transformation
37
 
38
  ## Intended uses & limitations
39
 
40
+ This model is intended for:
41
+ - Research experiments in NLP and inclusive language
42
+ - Educational purposes in courses on Machine Translation or Text Generation
43
+ - Demonstrations of transfer learning using multilingual seq2seq models
44
+ - Automatic rewriting of short Spanish sentences into gender-neutral forms
45
 
46
  ## Training and evaluation data
47
 
48
+ The model was trained on the Spanish Gender Neutralization dataset available on Hugging Face:
49
+ 👉 hackathon-pln-es/neutral-es
50
+ This dataset contains pairs of aligned sentences:
51
+ - gendered: original sentence with grammatical gender marking
52
+ - neutral: reformulated gender-neutral version
53
+
54
+ The dataset already includes a predefined split:
55
+ - Training set
56
+ - Test set
57
+
58
+ The dataset is relatively small and designed mainly for educational and experimental purposes, not for large-scale production systems.
59
+ Before training, the data was:
60
+ - tokenized using the mBART tokenizer
61
+ - truncated/padded to model limits
62
+ - converted into input/label format for seq2seq training
63
+
64
+ Evaluation was performed using the BLEU score (sacrebleu), a standard metric in machine translation.
65
 
66
  ## Training procedure
67
 
68
+ The model was trained using the Hugging Face Trainer API for sequence-to-sequence learning.
69
+ Training steps:
70
+ 1. The pretrained model facebook/mbart-large-50 was loaded.
71
+ 2. The dataset was tokenized using the corresponding mBART tokenizer.
72
+ 3. Inputs were formatted as:
73
+ - source: gendered sentence
74
+ - target: neutral sentence
75
+ 4. The model was fine-tuned using transfer learning.
76
+ 5. Training was performed on GPU in Google Colab.
77
+ 6. Evaluation during training used the sacrebleu metric.
78
+ 7. The final model was uploaded to the Hugging Face Hub.
79
+
80
+ The model therefore learns to perform monolingual rewriting via multilingual translation architecture.
81
+
82
  ### Training hyperparameters
83
 
84
  The following hyperparameters were used during training: