liuqi6777 commited on
Commit
317c4d0
Β·
verified Β·
1 Parent(s): d6d653f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +47 -4
README.md CHANGED
@@ -47,9 +47,10 @@ pipeline_tag: feature-extraction
47
  </div>
48
 
49
 
50
- <a href="https://Alibaba-NLP.github.io/E2Rank/">[πŸ€– Website]</a> |
51
- <a href="https://arxiv.org/pdf/2510.00000">[πŸ“„ Arxiv Paper]</a> |
52
- <a href="https://huggingface.co/collections/Alibaba-NLP/e2rank">[πŸ€— Huggingface Collection]</a> |
 
53
 
54
  </div>
55
 
@@ -367,7 +368,49 @@ print(scores.tolist())
367
 
368
  Since E2Rank extends a single text embedding model to perform both high-quality retrieval and listwise reranking, you can directly use it to build an end-to-end search system. By reusing the embeddings computed during the retrieval stage, E2Rank only need to compute the pseudo query's embedding and can efficiently rerank the retrieved documents with minimal additional computational overhead.
369
 
370
- Example code is comming soon.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
371
 
372
 
373
  # 🚩 Citation
 
47
  </div>
48
 
49
 
50
+ <a href="https://Alibaba-NLP.github.io/E2Rank/">πŸ€– Website</a> |
51
+ <a href="https://arxiv.org/pdf/2510.00000">πŸ“„ Arxiv Paper</a> |
52
+ <a href="https://huggingface.co/collections/Alibaba-NLP/e2rank">πŸ€— Huggingface Collection</a> |
53
+ <a href="#🚩-citation">🚩 Citation</a>
54
 
55
  </div>
56
 
 
368
 
369
  Since E2Rank extends a single text embedding model to perform both high-quality retrieval and listwise reranking, you can directly use it to build an end-to-end search system. By reusing the embeddings computed during the retrieval stage, E2Rank only need to compute the pseudo query's embedding and can efficiently rerank the retrieved documents with minimal additional computational overhead.
370
 
371
+ Example code is coming soon.
372
+
373
+ # πŸ“Š Evaluation
374
+
375
+ ## Reranking Benchmark
376
+
377
+ ### BEIR
378
+
379
+ | | Covid | NFCorpus | Touche | DBPedia | SciFact | Signal | News | Robust | Avg. |
380
+ |------------------------------------------------------------|:-----:|:--------:|:------:|:-------:|:-------:|:------:|:-----:|:------:|:---------:|
381
+ | BM25 | 59.47 | 30.75 | 44.22 | 31.80 | 67.89 | 33.05 | 39.52 | 40.70 | 43.43 |
382
+ | *Zero-shot Listwise Reranker* | | | | | | | | | |
383
+ | RankGPT-4o | 83.41 | 39.67 | 32.26 | 45.56 | 77.41 | 34.20 | 51.92 | 60.25 | 53.09 |
384
+ | RankGPT-4o-mini | 80.03 | 38.73 | 30.91 | 44.54 | 73.14 | 33.64 | 50.91 | 57.41 | 51.16 |
385
+ | RankQwen3-14B | 84.45 | 38.94 | 38.30 | 44.52 | 78.64 | 33.58 | 51.24 | 59.66 | 53.67 |
386
+ | RankQwen3-32B | 83.48 | 39.22 | 37.13 | 45.00 | 78.22 | 32.12 | 51.08 | 60.74 | 53.37 |
387
+ | *Fine-tuned Listwise Reranker based on Qwen3* | | | | | | | | | |
388
+ | RankQwen3-0.6B | 78.35 | 36.41 | 37.54 | 39.19 | 71.01 | 30.96 | 44.43 | 46.31 | 48.03 |
389
+ | RankQwen3-4B | 83.91 | 39.88 | 32.66 | 43.91 | 76.37 | 32.15 | 50.81 | 59.36 | 52.38 |
390
+ | RankQwen3-8B | 85.37 | 40.05 | 31.73 | 45.44 | 78.96 | 32.48 | 52.36 | 60.72 | 53.39 |
391
+ | *Ours* | | | | | | | | | |
392
+ | **E2Rank-0.6B** | 79.17 | 38.60 | 41.91 | 41.96 | 73.43 | 35.26 | 52.75 | 53.67 | 52.09 |
393
+ | **E2Rank-4B** | 83.30 | 39.20 | 43.16 | 42.95 | 77.19 | 34.48 | 52.71 | 60.16 | 54.14 |
394
+ | **E2Rank-8B** | 84.09 | 39.08 | 42.06 | 43.44 | 77.49 | 34.01 | 54.25 | 60.34 | **54.35** |
395
+
396
+ ## Embedding Benchmark
397
+
398
+ ### MTEB (Eng, v1)
399
+
400
+ | Models | Retr. | Rerank. | Clust. | PairClass. | Class. | STS | Summ. | Avg. |
401
+ |------------------------------------|:-----:|:-------:|:------:|:----------:|:------:|:-----:|:-----:|:---------:|
402
+ | Instructor-xl | 49.26 | 57.29 | 44.74 | 86.62 | 73.12 | 83.06 | 32.32 | 61.79 |
403
+ | BGE-large-en-v1.5 | 54.29 | 60.03 | 46.08 | 87.12 | 75.97 | 83.11 | 31.61 | 64.23 |
404
+ | GritLM-7B | 53.10 | 61.30 | 48.90 | 86.90 | 77.00 | 82.80 | 29.40 | 64.70 |
405
+ | E5-Mistral-7b-v1 | 52.78 | 60.38 | 47.78 | 88.47 | 76.80 | 83.77 | 31.90 | 64.56 |
406
+ | Echo-Mistral-7b-v1 | 55.52 | 58.14 | 46.32 | 87.34 | 77.43 | 82.56 | 30.73 | 64.68 |
407
+ | LLM2Vec-Mistral-7B | 55.99 | 58.42 | 45.54 | 87.99 | 76.63 | 84.09 | 29.96 | 64.80 |
408
+ | LLM2Vec-Meta-LLaMA-3-8B | 56.63 | 59.68 | 46.45 | 87.80 | 75.92 | 83.58 | 30.94 | 65.01 |
409
+ | **E2Rank-0.6B** | 51.74 | 55.97 | 40.85 | 83.93 | 73.66 | 81.41 | 30.90 | 61.25 |
410
+ | **E2Rank-4B** | 55.33 | 59.10 | 44.27 | 87.14 | 77.08 | 84.03 | 30.06 | 64.47 |
411
+ | **E2Rank-8B** | 56.89 | 59.58 | 44.75 | 86.96 | 76.81 | 84.52 | 30.23 | **65.03** |
412
+
413
+ > Note: For baselines, we only compared with models that are trained using public datasets.
414
 
415
 
416
  # 🚩 Citation