SenaSoft commited on
Commit
239a9c0
·
verified ·
1 Parent(s): 425059f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +139 -134
README.md CHANGED
@@ -1,199 +1,204 @@
1
  ---
2
- library_name: transformers
3
- tags: []
 
 
 
 
 
 
 
 
 
 
 
 
4
  ---
5
 
6
- # Model Card for Model ID
7
-
8
- <!-- Provide a quick summary of what the model is/does. -->
9
 
 
 
10
 
 
11
 
12
  ## Model Details
13
 
14
  ### Model Description
 
 
 
 
 
 
 
 
 
 
 
15
 
16
- <!-- Provide a longer summary of what this model is. -->
17
-
18
- This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
19
-
20
- - **Developed by:** [More Information Needed]
21
- - **Funded by [optional]:** [More Information Needed]
22
- - **Shared by [optional]:** [More Information Needed]
23
- - **Model type:** [More Information Needed]
24
- - **Language(s) (NLP):** [More Information Needed]
25
- - **License:** [More Information Needed]
26
- - **Finetuned from model [optional]:** [More Information Needed]
27
-
28
- ### Model Sources [optional]
29
-
30
- <!-- Provide the basic links for the model. -->
31
-
32
- - **Repository:** [More Information Needed]
33
- - **Paper [optional]:** [More Information Needed]
34
- - **Demo [optional]:** [More Information Needed]
35
 
36
  ## Uses
37
 
38
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
-
40
  ### Direct Use
 
41
 
42
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
 
 
43
 
44
- [More Information Needed]
45
 
46
- ### Downstream Use [optional]
 
47
 
48
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
 
49
 
50
- [More Information Needed]
 
 
51
 
52
- ### Out-of-Scope Use
 
 
 
 
53
 
54
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
 
 
 
55
 
56
- [More Information Needed]
57
 
58
  ## Bias, Risks, and Limitations
59
-
60
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
-
62
- [More Information Needed]
63
 
64
  ### Recommendations
 
 
65
 
66
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
-
68
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
 
70
  ## How to Get Started with the Model
 
 
71
 
72
- Use the code below to get started with the model.
 
 
73
 
74
- [More Information Needed]
75
 
76
  ## Training Details
77
 
78
  ### Training Data
79
-
80
- <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
-
82
- [More Information Needed]
 
83
 
84
  ### Training Procedure
85
 
86
- <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
-
88
- #### Preprocessing [optional]
89
-
90
- [More Information Needed]
91
-
92
 
93
  #### Training Hyperparameters
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
94
 
95
- - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
-
97
- #### Speeds, Sizes, Times [optional]
98
-
99
- <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
-
101
- [More Information Needed]
102
 
103
  ## Evaluation
104
 
105
- <!-- This section describes the evaluation protocols and provides the results. -->
106
-
107
- ### Testing Data, Factors & Metrics
108
-
109
- #### Testing Data
110
-
111
- <!-- This should link to a Dataset Card if possible. -->
112
-
113
- [More Information Needed]
114
-
115
- #### Factors
116
 
117
- <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
-
119
- [More Information Needed]
120
-
121
- #### Metrics
122
-
123
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
-
125
- [More Information Needed]
126
 
127
  ### Results
 
 
128
 
129
- [More Information Needed]
130
-
131
- #### Summary
132
-
133
-
134
-
135
- ## Model Examination [optional]
136
-
137
- <!-- Relevant interpretability work for the model goes here -->
138
-
139
- [More Information Needed]
140
 
141
  ## Environmental Impact
 
 
 
 
142
 
143
- <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
-
145
- Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
-
147
- - **Hardware Type:** [More Information Needed]
148
- - **Hours used:** [More Information Needed]
149
- - **Cloud Provider:** [More Information Needed]
150
- - **Compute Region:** [More Information Needed]
151
- - **Carbon Emitted:** [More Information Needed]
152
 
153
- ## Technical Specifications [optional]
154
 
155
  ### Model Architecture and Objective
156
-
157
- [More Information Needed]
 
158
 
159
  ### Compute Infrastructure
 
 
 
 
 
 
160
 
161
- [More Information Needed]
162
-
163
- #### Hardware
164
-
165
- [More Information Needed]
166
-
167
- #### Software
168
-
169
- [More Information Needed]
170
-
171
- ## Citation [optional]
172
-
173
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
-
175
- **BibTeX:**
176
-
177
- [More Information Needed]
178
-
179
- **APA:**
180
-
181
- [More Information Needed]
182
-
183
- ## Glossary [optional]
184
-
185
- <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
 
187
- [More Information Needed]
188
 
189
- ## More Information [optional]
 
 
 
 
 
 
 
 
 
190
 
191
- [More Information Needed]
 
192
 
193
- ## Model Card Authors [optional]
194
 
195
- [More Information Needed]
 
196
 
197
  ## Model Card Contact
 
 
 
198
 
199
- [More Information Needed]
 
 
 
 
1
  ---
2
+ base_model: nlptown/bert-base-multilingual-uncased-sentiment
3
+ library_name: peft
4
+ tags:
5
+ - base_model:adapter:nlptown/bert-base-multilingual-uncased-sentiment
6
+ - lora
7
+ - transformers
8
+ - sentiment-analysis
9
+ - spanish
10
+ - social-media
11
+ - text-classification
12
+ datasets:
13
+ - pyupeu/social-media-peruvian-sentiment
14
+ metrics:
15
+ - accuracy
16
  ---
17
 
18
+ # SenaSoft/chdv-sentiment-analysis
 
 
19
 
20
+ Este modelo es una adaptación de `nlptown/bert-base-multilingual-uncased-sentiment` entrenado con **LoRA** para la tarea de **análisis de sentimiento** en textos en español.
21
+ Fue ajustado específicamente en un dataset de publicaciones en redes sociales peruanas con 3 etiquetas: **negativo, neutral y positivo**.
22
 
23
+ ---
24
 
25
  ## Model Details
26
 
27
  ### Model Description
28
+ - **Autores:** Christopher Aponte y David Navarro
29
+ - **Modelo base:** [nlptown/bert-base-multilingual-uncased-sentiment](https://huggingface.co/nlptown/bert-base-multilingual-uncased-sentiment)
30
+ - **Tipo de modelo:** Clasificación de secuencias (sentiment analysis)
31
+ - **Idiomas:** Multilingual
32
+ - **Tarea:** Análisis de sentimiento en textos cortos (ej. redes sociales)
33
+ - **License:** Apache 2.0
34
+ - **Finetuned from:** `nlptown/bert-base-multilingual-uncased-sentiment`
35
+
36
+ ### Model Sources
37
+ - **Repository:** [Hugging Face Hub](https://huggingface.co/SenaSoft/chdv-sentiment-analysis)
38
+ - **Dataset:** [pyupeu/social-media-peruvian-sentiment](https://huggingface.co/datasets/pyupeu/social-media-peruvian-sentiment)
39
 
40
+ ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
41
 
42
  ## Uses
43
 
 
 
44
  ### Direct Use
45
+ Este modelo puede usarse directamente para clasificar textos en español en **positivo, negativo o neutral**.
46
 
47
+ Ejemplos de uso:
48
+ ```python
49
+ from transformers import pipeline
50
 
51
+ clasificador = pipeline("text-classification", model="SenaSoft/chdv-sentiment-analysis")
52
 
53
+ print(clasificador("Me encanta este producto"))
54
+ # [{'label': 'positivo', 'score': 0.92}]
55
 
56
+ print(clasificador("No me gustó para nada"))
57
+ # [{'label': 'negativo', 'score': 0.88}]
58
 
59
+ print(clasificador("Está bien, pero podría mejorar"))
60
+ # [{'label': 'neutral', 'score': 0.74}]
61
+ ```
62
 
63
+ ### Downstream Use
64
+ Puede integrarse en:
65
+ - Plataformas de monitoreo de redes sociales.
66
+ - Sistemas de atención al cliente.
67
+ - Análisis de feedback de usuarios.
68
 
69
+ ### Out-of-Scope Use
70
+ - No está entrenado para otros idiomas distintos al español.
71
+ - No es un detector de sarcasmo ni de sentimientos complejos como miedo o sorpresa.
72
+ - No debe usarse en contextos médicos, legales o financieros sin verificación adicional.
73
 
74
+ ---
75
 
76
  ## Bias, Risks, and Limitations
77
+ - Entrenado con un dataset específico (Perú, redes sociales). Puede no generalizar igual en otros países o contextos.
78
+ - Riesgo de sesgo cultural y lingüístico.
79
+ - Puede fallar en casos de ironía, sarcasmo o ambigüedad semántica.
 
80
 
81
  ### Recommendations
82
+ - Se recomienda evaluar antes de usar en producción fuera de su dominio.
83
+ - Puede complementarse con técnicas de **ensembles** o datasets más diversos.
84
 
85
+ ---
 
 
86
 
87
  ## How to Get Started with the Model
88
+ ```python
89
+ from transformers import pipeline
90
 
91
+ clasificador = pipeline("text-classification", model="SenaSoft/chdv-sentiment-analysis")
92
+ clasificador("El servicio estuvo excelente, volveré pronto!")
93
+ ```
94
 
95
+ ---
96
 
97
  ## Training Details
98
 
99
  ### Training Data
100
+ Dataset: [pyupeu/social-media-peruvian-sentiment](https://huggingface.co/datasets/pyupeu/social-media-peruvian-sentiment)
101
+ Etiquetas:
102
+ - `0`: negativo
103
+ - `1`: neutral
104
+ - `2`: positivo
105
 
106
  ### Training Procedure
107
 
108
+ #### Preprocessing
109
+ - Textos truncados y/o rellenados a `max_length=128`.
110
+ - Tokenización con `AutoTokenizer` de `bert-base-multilingual-uncased-sentiment`.
 
 
 
111
 
112
  #### Training Hyperparameters
113
+ - **LoRA Config:**
114
+ - `r=8`
115
+ - `lora_alpha=16`
116
+ - `lora_dropout=0.7`
117
+ - **TrainingArguments:**
118
+ - `batch_size=16`
119
+ - `num_train_epochs=20`
120
+ - `learning_rate=5e-5`
121
+ - `weight_decay=0.01`
122
+ - `eval_strategy="epoch"`
123
+ - `save_strategy="epoch"`
124
+ - `load_best_model_at_end=True`
125
+ - `metric_for_best_model="accuracy"`
126
+
127
+ #### Speeds, Sizes, Times
128
+ - Número de épocas: 20
129
+ - Mejor resultado en **epoch 15**
130
+ - Checkpoint tamaño aproximado: ~420 MB
131
 
132
+ ---
 
 
 
 
 
 
133
 
134
  ## Evaluation
135
 
136
+ ### Testing Data
137
+ Dataset de prueba del mismo `pyupeu/social-media-peruvian-sentiment`.
 
 
 
 
 
 
 
 
 
138
 
139
+ ### Metrics
140
+ - **Accuracy**: 0.66
141
+ - **Training Loss (epoch 15):** 0.722
142
+ - **Validation Loss (epoch 15):** 0.785
 
 
 
 
 
143
 
144
  ### Results
145
+ - Mejor rendimiento alcanzado en epoch 15.
146
+ - Buen ajuste considerando dataset pequeño y dominio específico.
147
 
148
+ ---
 
 
 
 
 
 
 
 
 
 
149
 
150
  ## Environmental Impact
151
+ - **Hardware:** NVIDIA GPU (única)
152
+ - **Tiempo de entrenamiento:** ~2 horas
153
+ - **Cloud Provider:** Local / Dockerizado
154
+ - **Carbon Emitted:** No estimado
155
 
156
+ ---
 
 
 
 
 
 
 
 
157
 
158
+ ## Technical Specifications
159
 
160
  ### Model Architecture and Objective
161
+ - Modelo base: BERT Multilingüe (`uncased`)
162
+ - Objetivo: clasificación en 3 clases
163
+ - Técnica usada: Fine-tuning con LoRA
164
 
165
  ### Compute Infrastructure
166
+ - **Hardware:** GPU local
167
+ - **Software:**
168
+ - Python 3.10+
169
+ - Transformers 4.56.1
170
+ - PEFT 0.17.1
171
+ - Datasets 3.0.2
172
 
173
+ ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
174
 
175
+ ## Citation
176
 
177
+ **BibTeX**
178
+ ```bibtex
179
+ @misc{aponte2025chdv,
180
+ title={SenaSoft CHDV Sentiment Analysis},
181
+ author={Aponte, Christopher and Navarro, David},
182
+ year={2025},
183
+ howpublished={Hugging Face},
184
+ url={https://huggingface.co/SenaSoft/chdv-sentiment-analysis}
185
+ }
186
+ ```
187
 
188
+ **APA**
189
+ Aponte, C., & Navarro, D. (2025). *SenaSoft CHDV Sentiment Analysis* [Model]. Hugging Face. https://huggingface.co/SenaSoft/chdv-sentiment-analysis
190
 
191
+ ---
192
 
193
+ ## Model Card Authors
194
+ Christopher Aponte y David Navarro
195
 
196
  ## Model Card Contact
197
+ Para consultas: [Christopher Aponte & David Navarro, SenaSoft 2025]
198
+
199
+ ---
200
 
201
+ ### Framework versions
202
+ - **PEFT:** 0.17.1
203
+ - **Transformers:** 4.56.1
204
+ - **Datasets:** 3.0.2