Sinanmz commited on
Commit
dad8859
·
verified ·
1 Parent(s): 04823ae

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +54 -84
README.md CHANGED
@@ -64,12 +64,6 @@ This BERT-based multilabel multiclass classifier is designed to predict the genr
64
 
65
  ## How to Get Started with the Model
66
 
67
- Here's the "How to Get Started with the Model" section for your model card:
68
-
69
- ---
70
-
71
- ## How to Get Started with the Model
72
-
73
  Use the code below to get started with the model.
74
 
75
  ```python
@@ -108,100 +102,76 @@ print(f"Predicted genres: {predicted_genres}")
108
 
109
 
110
 
111
- ## Evaluation
112
-
113
- <!-- This section describes the evaluation protocols and provides the results. -->
114
-
115
- ### Testing Data, Factors & Metrics
116
-
117
- #### Testing Data
118
-
119
- <!-- This should link to a Dataset Card if possible. -->
120
-
121
- [More Information Needed]
122
-
123
- #### Factors
124
 
125
- <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
126
 
127
- [More Information Needed]
128
 
129
  #### Metrics
130
 
131
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
132
-
133
- [More Information Needed]
134
 
135
  ### Results
136
 
137
- [More Information Needed]
138
-
139
  #### Summary
140
 
 
141
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
142
 
143
- ## Model Examination [optional]
144
-
145
- <!-- Relevant interpretability work for the model goes here -->
146
-
147
- [More Information Needed]
148
-
149
- ## Environmental Impact
150
-
151
- <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
152
-
153
- Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
154
-
155
- - **Hardware Type:** [More Information Needed]
156
- - **Hours used:** [More Information Needed]
157
- - **Cloud Provider:** [More Information Needed]
158
- - **Compute Region:** [More Information Needed]
159
- - **Carbon Emitted:** [More Information Needed]
160
-
161
- ## Technical Specifications [optional]
162
-
163
- ### Model Architecture and Objective
164
-
165
- [More Information Needed]
166
-
167
- ### Compute Infrastructure
168
-
169
- [More Information Needed]
170
-
171
- #### Hardware
172
-
173
- [More Information Needed]
174
-
175
- #### Software
176
-
177
- [More Information Needed]
178
-
179
- ## Citation [optional]
180
-
181
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
182
-
183
- **BibTeX:**
184
-
185
- [More Information Needed]
186
-
187
- **APA:**
188
-
189
- [More Information Needed]
190
-
191
- ## Glossary [optional]
192
-
193
- <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
194
-
195
- [More Information Needed]
196
 
197
- ## More Information [optional]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
198
 
199
- [More Information Needed]
200
 
201
- ## Model Card Authors [optional]
202
 
203
- [More Information Needed]
204
 
205
  ## Model Card Contact
206
 
207
- [More Information Needed]
 
 
64
 
65
  ## How to Get Started with the Model
66
 
 
 
 
 
 
 
67
  Use the code below to get started with the model.
68
 
69
  ```python
 
102
 
103
 
104
 
105
+ Here's the "Evaluation" section for your model card:
 
 
 
 
 
 
 
 
 
 
 
 
106
 
 
107
 
108
+ ## Evaluation
109
 
110
  #### Metrics
111
 
112
+ The evaluation metrics used for this model include precision, recall, and F1-score. These metrics were chosen because they provide a comprehensive view of the model's performance, particularly in a multilabel classification setting where it is important to understand not only how many correct predictions were made but also the balance between precision (accuracy of the positive predictions) and recall (the ability to find all positive instances).
 
 
113
 
114
  ### Results
115
 
 
 
116
  #### Summary
117
 
118
+ Below are the classification reports for the train, validation, and test splits of the dataset.
119
 
120
+ **Classification Report for Train Split:**
121
+ ```
122
+ precision recall f1-score support
123
+
124
+ Action 1.00 1.00 1.00 1655
125
+ Drama 1.00 1.00 1.00 4109
126
+ Comedy 1.00 1.00 1.00 2094
127
+ Animation 1.00 1.00 1.00 669
128
+ Crime 1.00 1.00 1.00 1284
129
+
130
+ micro avg 1.00 1.00 1.00 9811
131
+ macro avg 1.00 1.00 1.00 9811
132
+ weighted avg 1.00 1.00 1.00 9811
133
+ samples avg 1.00 1.00 1.00 9811
134
+ ```
135
 
136
+ **Classification Report for Val Split:**
137
+ ```
138
+ precision recall f1-score support
139
+
140
+ Action 0.70 0.73 0.71 220
141
+ Drama 0.77 0.84 0.80 507
142
+ Comedy 0.69 0.54 0.61 260
143
+ Animation 0.59 0.44 0.50 80
144
+ Crime 0.72 0.66 0.69 165
145
+
146
+ micro avg 0.73 0.71 0.72 1232
147
+ macro avg 0.70 0.64 0.66 1232
148
+ weighted avg 0.72 0.71 0.71 1232
149
+ samples avg 0.75 0.74 0.71 1232
150
+ ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
151
 
152
+ **Classification Report for Test Split:**
153
+ ```
154
+ precision recall f1-score support
155
+
156
+ Action 0.62 0.66 0.64 191
157
+ Drama 0.80 0.85 0.82 520
158
+ Comedy 0.69 0.58 0.63 260
159
+ Animation 0.60 0.49 0.54 78
160
+ Crime 0.65 0.67 0.66 154
161
+
162
+ micro avg 0.72 0.71 0.72 1203
163
+ macro avg 0.67 0.65 0.66 1203
164
+ weighted avg 0.72 0.71 0.71 1203
165
+ samples avg 0.75 0.75 0.72 1203
166
+ ```
167
 
168
+ The results indicate that the model performs well on the training data with high precision, recall, and F1-scores across all genres. However, there is a drop in performance on the validation and test splits, highlighting areas where the model could be further improved to generalize better to unseen data.
169
 
170
+ ## Model Card Authors
171
 
172
+ Sina Namazi
173
 
174
  ## Model Card Contact
175
 
176
+ - **Github:** github.com/Sinanmz
177
+ - **Hugging Face:** huggingface.co/Sinanmz