Datasets:

Modalities:
Text
Formats:
parquet
ArXiv:
Libraries:
Datasets
pandas
Samoed commited on
Commit
a34edc2
·
verified ·
1 Parent(s): fa463a1

Add dataset card

Browse files
Files changed (1) hide show
  1. README.md +9 -38
README.md CHANGED
@@ -2,6 +2,8 @@
2
  language:
3
  - cmn
4
  multilinguality: monolingual
 
 
5
  task_categories:
6
  - sentence-similarity
7
  task_ids:
@@ -56,6 +58,8 @@ A Chinese dataset for textual relatedness
56
  | Reference | https://aclanthology.org/2021.emnlp-main.357 |
57
 
58
 
 
 
59
  ## How to evaluate on this task
60
 
61
  You can evaluate an embedding model on this dataset using the following code:
@@ -63,15 +67,15 @@ You can evaluate an embedding model on this dataset using the following code:
63
  ```python
64
  import mteb
65
 
66
- task = mteb.get_tasks(["ATEC"])
67
- evaluator = mteb.MTEB(task)
68
 
69
  model = mteb.get_model(YOUR_MODEL)
70
  evaluator.run(model)
71
  ```
72
 
73
  <!-- Datasets want link to arxiv in readme to autolink dataset with paper -->
74
- To learn more about how to run models on `mteb` task check out the [GitHub repitory](https://github.com/embeddings-benchmark/mteb).
75
 
76
  ## Citation
77
 
@@ -112,7 +116,7 @@ Yih, Scott Wen-tau},
112
  }
113
 
114
  @article{muennighoff2022mteb,
115
- author = {Muennighoff, Niklas and Tazi, Nouamane and Magne, Lo{\"\i}c and Reimers, Nils},
116
  title = {MTEB: Massive Text Embedding Benchmark},
117
  publisher = {arXiv},
118
  journal={arXiv preprint arXiv:2210.07316},
@@ -137,40 +141,7 @@ desc_stats = task.metadata.descriptive_stats
137
  ```
138
 
139
  ```json
140
- {
141
- "validation": {
142
- "num_samples": 20000,
143
- "number_of_characters": 536573,
144
- "unique_pairs": 20000,
145
- "min_sentence1_length": 5,
146
- "average_sentence1_len": 13.4172,
147
- "max_sentence1_length": 84,
148
- "unique_sentence1": 19909,
149
- "min_sentence2_length": 5,
150
- "average_sentence2_len": 13.41145,
151
- "max_sentence2_length": 82,
152
- "unique_sentence2": 19882,
153
- "min_score": 0,
154
- "avg_score": 0.1844,
155
- "max_score": 1
156
- },
157
- "test": {
158
- "num_samples": 20000,
159
- "number_of_characters": 536531,
160
- "unique_pairs": 20000,
161
- "min_sentence1_length": 5,
162
- "average_sentence1_len": 13.40835,
163
- "max_sentence1_length": 97,
164
- "unique_sentence1": 19911,
165
- "min_sentence2_length": 5,
166
- "average_sentence2_len": 13.4182,
167
- "max_sentence2_length": 88,
168
- "unique_sentence2": 19907,
169
- "min_score": 0,
170
- "avg_score": 0.1805,
171
- "max_score": 1
172
- }
173
- }
174
  ```
175
 
176
  </details>
 
2
  language:
3
  - cmn
4
  multilinguality: monolingual
5
+ source_datasets:
6
+ - C-MTEB/ATEC
7
  task_categories:
8
  - sentence-similarity
9
  task_ids:
 
58
  | Reference | https://aclanthology.org/2021.emnlp-main.357 |
59
 
60
 
61
+
62
+
63
  ## How to evaluate on this task
64
 
65
  You can evaluate an embedding model on this dataset using the following code:
 
67
  ```python
68
  import mteb
69
 
70
+ task = mteb.get_task("ATEC")
71
+ evaluator = mteb.MTEB([task])
72
 
73
  model = mteb.get_model(YOUR_MODEL)
74
  evaluator.run(model)
75
  ```
76
 
77
  <!-- Datasets want link to arxiv in readme to autolink dataset with paper -->
78
+ To learn more about how to run models on `mteb` task check out the [GitHub repository](https://github.com/embeddings-benchmark/mteb).
79
 
80
  ## Citation
81
 
 
116
  }
117
 
118
  @article{muennighoff2022mteb,
119
+ author = {Muennighoff, Niklas and Tazi, Nouamane and Magne, Loïc and Reimers, Nils},
120
  title = {MTEB: Massive Text Embedding Benchmark},
121
  publisher = {arXiv},
122
  journal={arXiv preprint arXiv:2210.07316},
 
141
  ```
142
 
143
  ```json
144
+ {}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
145
  ```
146
 
147
  </details>