File size: 1,567 Bytes
a30de77 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 | ---
pipeline_tag: sentence-similarity
tags:
- sentence-transformers
- feature-extraction
- sentence-similarity
- lore
- logic-oriented-retrieval
license: mit
---
For more details please refer to our github repo: https://github.com/FlagOpen/FlagEmbedding
# Lore-Bge3: Logic-ORiented Retriever Enhancement for BGE-M3
This model is a fine-tuned version of [BAAI/bge-m3](https://arxiv.org/pdf/2402.03216.pdf) using the LORE (Logic-ORiented Retriever Enhancement) method. It significantly improves retrieval performance for complex logical expressions and queries.
## LORE Method Overview
LORE is a novel embedding enhancement method that improves retrieval performance through fine-grained contrastive learning:
- **Three-tier Contrastive Learning**: Fine-grained sample classification with P (Positive), N1 (Distractor), and N2 (Negative) samples
- **Dual Encoder Architecture**: Frozen document encoder M_d and trainable query encoder M_q
- **InfoNCE-based Loss**: Differentiated weights for hierarchical separation P ≻ N1 ≻ N2
- **Query Rewriting**: LLM-assisted dataset construction with discourse relations from Rhetorical Structure Theory (RST)
- **No External Dependencies**: Requires no external supervision, resources, or pre-retrieval analysis
## Key Improvements
- **Enhanced Logical Reasoning**: Improved ability to handle complex logical expressions in queries
- **Fine-grained Discrimination**: Better distinction between relevant content and distractors
- **Maintained Efficiency**: Preserves the computational efficiency of the original model
|