|
|
--- |
|
|
datasets: |
|
|
- Jasaxion/LexSemBridge_eval |
|
|
language: |
|
|
- en |
|
|
library_name: sentence-transformers |
|
|
license: apache-2.0 |
|
|
pipeline_tag: feature-extraction |
|
|
tags: |
|
|
- sentence-transformers |
|
|
- sentence-similarity |
|
|
- feature-extraction |
|
|
- lexsembridge |
|
|
widget: [] |
|
|
--- |
|
|
|
|
|
# LexSemBridge: Fine-Grained Dense Representation Enhancement through Token-Aware Embedding Augmentation |
|
|
|
|
|
This model implements **LexSemBridge**, a unified framework that enhances dense query representations through fine-grained, input-aware vector modulation. LexSemBridge constructs latent enhancement vectors from input tokens using statistical, learned, and contextual paradigms, integrating them with dense embeddings via element-wise interaction. It naturally extends to both text and vision modalities with an appropriate tokenization, aiming to improve performance on fine-grained retrieval tasks where precise keyword alignment and span-level localization are crucial. |
|
|
|
|
|
The model is based on the paper [LexSemBridge: Fine-Grained Dense Representation Enhancement through Token-Aware Embedding Augmentation](https://huggingface.co/papers/2508.17858). |
|
|
|
|
|
For the official code and further details, please refer to the [GitHub repository](https://github.com/Jasaxion/LexSemBridge/). |
|
|
|
|
|
## Model Details |
|
|
|
|
|
### Model Description |
|
|
- **Model Type:** Sentence Transformer |
|
|
- **Maximum Sequence Length:** 512 tokens |
|
|
- **Output Dimensionality:** 1024 tokens |
|
|
- **Similarity Function:** Cosine Similarity |
|
|
|
|
|
### Model Sources |
|
|
|
|
|
- **Paper:** [LexSemBridge: Fine-Grained Dense Representation Enhancement through Token-Aware Embedding Augmentation](https://huggingface.co/papers/2508.17858) |
|
|
- **Code/GitHub Repository:** https://github.com/Jasaxion/LexSemBridge/ |
|
|
- **Documentation:** [Sentence Transformers Documentation](https://sbert.net) |
|
|
- **Repository (Sentence Transformers Library):** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers) |
|
|
- **Hugging Face (Sentence Transformers Models):** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers) |
|
|
|
|
|
### Full Model Architecture |
|
|
|
|
|
``` |
|
|
SentenceTransformer( |
|
|
(0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel |
|
|
(1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True}) |
|
|
(2): Normalize() |
|
|
) |
|
|
``` |
|
|
|
|
|
## Usage |
|
|
|
|
|
### Direct Usage (Sentence Transformers) |
|
|
|
|
|
First install the Sentence Transformers library: |
|
|
|
|
|
```bash |
|
|
pip install -U sentence-transformers |
|
|
``` |
|
|
|
|
|
Then you can load this model and run inference. |
|
|
```python |
|
|
from sentence_transformers import SentenceTransformer |
|
|
|
|
|
# Download from the 🤗 Hub |
|
|
model = SentenceTransformer("Jasaxion/LexSemBridge_CLR_snowflake") # Example: LexSemBridge-CLR-snowflake model |
|
|
# Run inference |
|
|
sentences = [ |
|
|
'The weather is lovely today.', |
|
|
"It's so sunny outside!", |
|
|
'He drove to the stadium.', |
|
|
] |
|
|
embeddings = model.encode(sentences) |
|
|
print(embeddings.shape) |
|
|
# [3, 1024] |
|
|
|
|
|
# Get the similarity scores for the embeddings |
|
|
similarities = model.similarity(embeddings, embeddings) |
|
|
print(similarities.shape) |
|
|
# [3, 3] |
|
|
``` |
|
|
|
|
|
## Training Details |
|
|
|
|
|
### Framework Versions |
|
|
- Python: 3.10.14 |
|
|
- Sentence Transformers: 3.1.0.dev0 |
|
|
- Transformers: 4.44.2 |
|
|
- PyTorch: 2.4.1+cu121 |
|
|
- Accelerate: 0.34.2 |
|
|
- Datasets: 2.21.0 |
|
|
- Tokenizers: 0.19.1 |
|
|
|
|
|
## Citation |
|
|
|
|
|
### BibTeX |
|
|
```bibtex |
|
|
@article{zhan2025lexsembridge, |
|
|
title={LexSemBridge: Fine-Grained Dense Representation Enhancement through Token-Aware Embedding Augmentation}, |
|
|
author={Zhan, Shaoxiong and Lin, Hai and Tan, Hongming and Cai, Xiaodong and Zheng, Hai-Tao and Su, Xin and Shan, Zifei and Liu, Ruitong and Kim, Hong-Gee}, |
|
|
journal={arXiv preprint arXiv:2508.17858}, |
|
|
year={2025} |
|
|
} |
|
|
``` |