Sathwik3 commited on
Commit
6415243
·
verified ·
1 Parent(s): aa30061

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -51
README.md CHANGED
@@ -145,10 +145,10 @@ print(result)
145
  ### Training Data
146
 
147
  The model was fine-tuned on an emotion classification dataset. Specific dataset details:
148
- - **Dataset:** [Dataset name and link - placeholder for specific information]
149
- - **Size:** [Number of training examples - placeholder]
150
- - **Emotion categories:** [List of emotion labels - placeholder]
151
- - **Data split:** [Train/validation/test split information - placeholder]
152
 
153
  ### Training Procedure
154
 
@@ -157,24 +157,16 @@ The model was fine-tuned on an emotion classification dataset. Specific dataset
157
  - Text tokenization using DistilBERT tokenizer
158
  - Maximum sequence length: 512 tokens
159
  - Truncation and padding applied as needed
160
- - Text normalization: [specific preprocessing steps - placeholder]
161
 
162
  #### Training Hyperparameters
163
 
164
- - **Training regime:** Mixed precision (fp16) [placeholder - adjust if different]
165
  - **Optimizer:** AdamW
166
- - **Learning rate:** [e.g., 2e-5 - placeholder]
167
- - **Batch size:** [e.g., 16 or 32 - placeholder]
168
- - **Number of epochs:** [e.g., 3-5 - placeholder]
169
- - **Weight decay:** [e.g., 0.01 - placeholder]
170
- - **Warmup steps:** [placeholder]
171
- - **Scheduler:** [e.g., Linear with warmup - placeholder]
172
 
173
- #### Training Infrastructure
174
-
175
- - **Hardware:** [GPU type, e.g., NVIDIA Tesla V100 - placeholder]
176
- - **Training time:** [Approximate duration - placeholder]
177
- - **Framework:** PyTorch with Hugging Face Transformers
178
 
179
  ## Evaluation
180
 
@@ -201,35 +193,10 @@ The model's performance is evaluated using:
201
 
202
  | Metric | Value |
203
  |--------|-------|
204
- | Accuracy | [e.g., 0.XX - placeholder] |
205
- | Macro F1 | [e.g., 0.XX - placeholder] |
206
- | Weighted F1 | [e.g., 0.XX - placeholder] |
207
- | Macro Precision | [e.g., 0.XX - placeholder] |
208
- | Macro Recall | [e.g., 0.XX - placeholder] |
209
-
210
- #### Per-Class Performance
211
-
212
- [Placeholder for per-class metrics table]
213
-
214
- | Emotion | Precision | Recall | F1-Score | Support |
215
- |---------|-----------|--------|----------|----------|
216
- | [Class 1] | [0.XX] | [0.XX] | [0.XX] | [N] |
217
- | [Class 2] | [0.XX] | [0.XX] | [0.XX] | [N] |
218
- | ... | ... | ... | ... | ... |
219
-
220
- ### Summary
221
 
222
- The model demonstrates strong performance on emotion classification tasks, with particular strengths in [specific aspects - placeholder]. Areas for potential improvement include [specific areas - placeholder].
223
 
224
- ## Environmental Impact
225
-
226
- Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
227
-
228
- - **Hardware Type:** [e.g., NVIDIA Tesla V100 - placeholder]
229
- - **Hours used:** [placeholder]
230
- - **Cloud Provider:** [e.g., AWS, GCP, Azure, or on-premises - placeholder]
231
- - **Compute Region:** [e.g., us-east-1 - placeholder]
232
- - **Carbon Emitted:** [e.g., XX kg CO2eq - placeholder]
233
 
234
  ## Technical Specifications
235
 
@@ -244,22 +211,16 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
244
  - **Max Sequence Length:** 512 tokens
245
  - **Vocabulary Size:** 30,522 tokens
246
 
247
- ### Compute Infrastructure
248
-
249
- #### Hardware
250
-
251
- [Placeholder for specific hardware information - e.g., GPU type, CPU, memory]
252
 
253
  #### Software
254
 
255
  - **Framework:** PyTorch
256
  - **Library:** Hugging Face Transformers
257
- - **Python Version:** [e.g., 3.8+ - placeholder]
258
  - **Key Dependencies:**
259
  - transformers
260
  - torch
261
  - tokenizers
262
- - datasets (if applicable)
263
 
264
  ## Citation
265
 
 
145
  ### Training Data
146
 
147
  The model was fine-tuned on an emotion classification dataset. Specific dataset details:
148
+ - **Dataset:** Emotion dataset
149
+ - **Size:** 16000
150
+ - **Emotion categories:** ['sadness', 'joy', 'love', 'anger', 'fear', 'surprise']
151
+ - **Data split:** Train,Validation,Test
152
 
153
  ### Training Procedure
154
 
 
157
  - Text tokenization using DistilBERT tokenizer
158
  - Maximum sequence length: 512 tokens
159
  - Truncation and padding applied as needed
 
160
 
161
  #### Training Hyperparameters
162
 
163
+ - **Training regime:** Mixed precision (fp16)
164
  - **Optimizer:** AdamW
165
+ - **Learning rate:** 2e-5
166
+ - **Batch size:** 64
167
+ - **Number of epochs:** 2
168
+ - **Weight decay:** 0.01
 
 
169
 
 
 
 
 
 
170
 
171
  ## Evaluation
172
 
 
193
 
194
  | Metric | Value |
195
  |--------|-------|
196
+ | Accuracy | 0.9295 |
197
+ | Weighted F1 | 0.9292 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
198
 
 
199
 
 
 
 
 
 
 
 
 
 
200
 
201
  ## Technical Specifications
202
 
 
211
  - **Max Sequence Length:** 512 tokens
212
  - **Vocabulary Size:** 30,522 tokens
213
 
 
 
 
 
 
214
 
215
  #### Software
216
 
217
  - **Framework:** PyTorch
218
  - **Library:** Hugging Face Transformers
219
+ - **Python Version:** 3.10
220
  - **Key Dependencies:**
221
  - transformers
222
  - torch
223
  - tokenizers
 
224
 
225
  ## Citation
226