Datasets:

Modalities:
Text
Formats:
parquet
Languages:
Persian
ArXiv:
Libraries:
Datasets
pandas
License:
Samoed commited on
Commit
7c7de0b
·
verified ·
1 Parent(s): e7ae011

Add dataset card

Browse files
Files changed (1) hide show
  1. README.md +7 -146
README.md CHANGED
@@ -6,7 +6,7 @@ language:
6
  license: unknown
7
  multilinguality: monolingual
8
  source_datasets:
9
- - PNLPhub/DigiMag
10
  task_categories:
11
  - text-classification
12
  task_ids: []
@@ -58,7 +58,7 @@ A total of 8,515 articles scraped from Digikala Online Magazine. This dataset in
58
  | Reference | https://hooshvare.github.io/docs/datasets/tc |
59
 
60
  Source datasets:
61
- - [PNLPhub/DigiMag](https://huggingface.co/datasets/PNLPhub/DigiMag)
62
 
63
 
64
  ## How to evaluate on this task
@@ -68,15 +68,15 @@ You can evaluate an embedding model on this dataset using the following code:
68
  ```python
69
  import mteb
70
 
71
- task = mteb.get_tasks(["DigikalamagClassification"])
72
- evaluator = mteb.MTEB(task)
73
 
74
  model = mteb.get_model(YOUR_MODEL)
75
  evaluator.run(model)
76
  ```
77
 
78
  <!-- Datasets want link to arxiv in readme to autolink dataset with paper -->
79
- To learn more about how to run models on `mteb` task check out the [GitHub repitory](https://github.com/embeddings-benchmark/mteb).
80
 
81
  ## Citation
82
 
@@ -96,7 +96,7 @@ If you use this dataset, please cite the dataset as well as [mteb](https://githu
96
  }
97
 
98
  @article{muennighoff2022mteb,
99
- author = {Muennighoff, Niklas and Tazi, Nouamane and Magne, Lo{\"\i}c and Reimers, Nils},
100
  title = {MTEB: Massive Text Embedding Benchmark},
101
  publisher = {arXiv},
102
  journal={arXiv preprint arXiv:2210.07316},
@@ -121,146 +121,7 @@ desc_stats = task.metadata.descriptive_stats
121
  ```
122
 
123
  ```json
124
- {
125
- "test": {
126
- "num_samples": 852,
127
- "number_of_characters": 2935825,
128
- "number_texts_intersect_with_train": 0,
129
- "min_text_length": 40,
130
- "average_text_length": 3445.8039906103286,
131
- "max_text_length": 36911,
132
- "unique_texts": 852,
133
- "min_labels_per_text": 5,
134
- "average_label_per_text": 12.762910798122066,
135
- "max_labels_per_text": 14,
136
- "unique_labels": 18,
137
- "labels": {
138
- "\u0633": {
139
- "count": 328
140
- },
141
- "\u0644": {
142
- "count": 715
143
- },
144
- "\u0627": {
145
- "count": 787
146
- },
147
- "\u0645": {
148
- "count": 642
149
- },
150
- "\u062a": {
151
- "count": 488
152
- },
153
- " ": {
154
- "count": 1470
155
- },
156
- "\u0648": {
157
- "count": 1590
158
- },
159
- "\u0632": {
160
- "count": 358
161
- },
162
- "\u06cc": {
163
- "count": 1975
164
- },
165
- "\u0628": {
166
- "count": 408
167
- },
168
- "\u0639": {
169
- "count": 289
170
- },
171
- "\u06a9": {
172
- "count": 302
173
- },
174
- "\u0646": {
175
- "count": 624
176
- },
177
- "\u0698": {
178
- "count": 277
179
- },
180
- "\u0647": {
181
- "count": 180
182
- },
183
- "\u0631": {
184
- "count": 193
185
- },
186
- "\u062f": {
187
- "count": 235
188
- },
189
- "\u062e": {
190
- "count": 13
191
- }
192
- }
193
- },
194
- "train": {
195
- "num_samples": 6896,
196
- "number_of_characters": 23218475,
197
- "number_texts_intersect_with_train": null,
198
- "min_text_length": 114,
199
- "average_text_length": 3366.948230858469,
200
- "max_text_length": 53321,
201
- "unique_texts": 6896,
202
- "min_labels_per_text": 5,
203
- "average_label_per_text": 12.764936194895592,
204
- "max_labels_per_text": 14,
205
- "unique_labels": 18,
206
- "labels": {
207
- "\u0639": {
208
- "count": 2342
209
- },
210
- "\u0644": {
211
- "count": 5794
212
- },
213
- "\u0645": {
214
- "count": 5194
215
- },
216
- " ": {
217
- "count": 11904
218
- },
219
- "\u0648": {
220
- "count": 12878
221
- },
222
- "\u062a": {
223
- "count": 3961
224
- },
225
- "\u06a9": {
226
- "count": 2451
227
- },
228
- "\u0646": {
229
- "count": 5046
230
- },
231
- "\u0698": {
232
- "count": 2245
233
- },
234
- "\u06cc": {
235
- "count": 15977
236
- },
237
- "\u0647": {
238
- "count": 1451
239
- },
240
- "\u0631": {
241
- "count": 1552
242
- },
243
- "\u0633": {
244
- "count": 2654
245
- },
246
- "\u0627": {
247
- "count": 6371
248
- },
249
- "\u0628": {
250
- "count": 3309
251
- },
252
- "\u0632": {
253
- "count": 2897
254
- },
255
- "\u062f": {
256
- "count": 1900
257
- },
258
- "\u062e": {
259
- "count": 101
260
- }
261
- }
262
- }
263
- }
264
  ```
265
 
266
  </details>
 
6
  license: unknown
7
  multilinguality: monolingual
8
  source_datasets:
9
+ - mteb/DigikalamagClassification
10
  task_categories:
11
  - text-classification
12
  task_ids: []
 
58
  | Reference | https://hooshvare.github.io/docs/datasets/tc |
59
 
60
  Source datasets:
61
+ - [mteb/DigikalamagClassification](https://huggingface.co/datasets/mteb/DigikalamagClassification)
62
 
63
 
64
  ## How to evaluate on this task
 
68
  ```python
69
  import mteb
70
 
71
+ task = mteb.get_task("DigikalamagClassification")
72
+ evaluator = mteb.MTEB([task])
73
 
74
  model = mteb.get_model(YOUR_MODEL)
75
  evaluator.run(model)
76
  ```
77
 
78
  <!-- Datasets want link to arxiv in readme to autolink dataset with paper -->
79
+ To learn more about how to run models on `mteb` task check out the [GitHub repository](https://github.com/embeddings-benchmark/mteb).
80
 
81
  ## Citation
82
 
 
96
  }
97
 
98
  @article{muennighoff2022mteb,
99
+ author = {Muennighoff, Niklas and Tazi, Nouamane and Magne, Loïc and Reimers, Nils},
100
  title = {MTEB: Massive Text Embedding Benchmark},
101
  publisher = {arXiv},
102
  journal={arXiv preprint arXiv:2210.07316},
 
121
  ```
122
 
123
  ```json
124
+ {}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
125
  ```
126
 
127
  </details>