| --- |
| license: bigscience-bloom-rail-1.0 |
| language: |
| - it |
| --- |
| # Model Card for Model ID |
|
|
| <!-- Provide a quick summary of what the model is/does. --> |
|
|
| This model is obtained by adapting bloom-1b7 to the Italian language. Among the languages supported by the BLOOM model, there is no Italian, making its use |
| in that context challenging. We adapt the original BLOOM model using the MAD-X language adaptation strategy. |
| Then, the adapted model is fine-tuned over two classification task prompts. To deal with this step, we decided to use data from two well-known EVALITA tasks: AMI2020 (misogyny detection) |
| and HASPEEDE-v2-2020 (hate-speech detection). |
|
|
| ## Model Details |
|
|
| ### Model Description |
|
|
| We adapt the bloom-1b7 to the Italian language using the MAD-X language adaptation strategy. |
| To produce a valuable model, we follow the same procedure proposed in: https://arxiv.org/abs/2212.09535 |
|
|
| We use default script parameters and select a sample of 100,000 examples in the Italian language. We decided to sample data from the Filtered Oscar Dataset for |
| the Italian Language released by Sarti. |
|
|
| Then the adopted model is fine-tuned over two classification task prompts using two well-known EVALITA tasks: AMI2020 (misogyny detection) |
| and HASPEEDE-v2-2020 (hate-speech detection). |
|
|
| We transformed the training data of the two tasks into an LLM prompt following a template. For the AMI task, we used the following template: |
|
|
| *instruction: Nel testo seguente si esprime odio contro le donne? Rispondi sì o no., input: \<text\>, output: \<sì/no\>.* |
|
|
| Similarly, for HASPEEDE we used: |
|
|
| *instruction: “Il testo seguente incita all’odio? Rispondi sì o no., input: \<text\>, output: \<sì/no\>.* |
|
|
| To fill these templates, we mapped the label "1" with the word "sì" and the label "0" with the word "no", \<text\> is just the sentence from the |
| dataset to classify. |
|
|
| To fine-tune the adapted model, we use the script available here: https://github.com/hyintell/BLOOM-fine-tuning/tree/main |
|
|
| **It is important to underline that when you use the adapted LLM or one of its fine-tuned models is necessary to use the tokenizer of the adapted model. |
| The BLOOM model adapted to the Italian language is available here: https://huggingface.co/basilepp19/bloom-1b7_it.** |
| |
| - **Developed by:** Pierpaolo Basile, Pierluigi Cassotti, Marco Polignano, Lucia Siciliani, Giovanni Semeraro. Department of Computer Science, University of Bari Aldo Moro, Italy |
| - **Model type:** BLOOM |
| - **Language(s) (NLP):** Italian |
| - **License:** BigScience BLOOM RAIL 1.0 |
| |
| ## Citation |
| |
| Pierpaolo Basile, Pierluigi Cassotti, Marco Polignano, Lucia Siciliani, Giovanni Semeraro. On the impact of Language Adaptation for Large Language Models: A |
| case study for the Italian language using only open resources. Proceedings of the Ninth Italian Conference on Computational Linguistics (CLiC-it 2023). |
| |
| |
| |