Update README.md
Browse files
README.md
CHANGED
|
@@ -76,6 +76,9 @@ Descriptions of the benchmark datasets used for evaluation are as follows:
|
|
| 76 |
- **XPQARetrieval**
|
| 77 |
A real-world dataset constructed from user queries and relevant product documents in a Korean e-commerce platform.
|
| 78 |
|
|
|
|
|
|
|
|
|
|
| 79 |
#### Sparse Embedding
|
| 80 |
| Model Name | # params | Avg. NDCG | NDCG@1 | NDCG@3 | NDCG@5 | NDCG@10 |
|
| 81 |
|------|:---:|:---:|:---:|:---:|:---:|:---:|
|
|
|
|
| 76 |
- **XPQARetrieval**
|
| 77 |
A real-world dataset constructed from user queries and relevant product documents in a Korean e-commerce platform.
|
| 78 |
|
| 79 |
+
> **Tip:**
|
| 80 |
+
> While many benchmark datasets are available for evaluation, in this project we chose to use only those that contain clean positive documents for each query. Keep in mind that a benchmark dataset is just that—a benchmark. For real-world applications, it is best to construct an evaluation dataset tailored to your specific domain and evaluate embedding models, such as PIXIE, in that environment to determine the most suitable one.
|
| 81 |
+
|
| 82 |
#### Sparse Embedding
|
| 83 |
| Model Name | # params | Avg. NDCG | NDCG@1 | NDCG@3 | NDCG@5 | NDCG@10 |
|
| 84 |
|------|:---:|:---:|:---:|:---:|:---:|:---:|
|