Improve model card for LexSemBridge: Add paper/code links, update pipeline tag, license, and datasets

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +26 -66
README.md CHANGED
@@ -1,36 +1,42 @@
1
  ---
2
- datasets: []
3
- language: []
 
 
4
  library_name: sentence-transformers
5
- pipeline_tag: sentence-similarity
 
6
  tags:
7
  - sentence-transformers
8
  - sentence-similarity
9
  - feature-extraction
 
10
  widget: []
11
  ---
12
 
13
- # SentenceTransformer
14
 
15
- This is a [sentence-transformers](https://www.SBERT.net) model trained. It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
 
 
 
 
16
 
17
  ## Model Details
18
 
19
  ### Model Description
20
  - **Model Type:** Sentence Transformer
21
- <!-- - **Base model:** [Unknown](https://huggingface.co/unknown) -->
22
  - **Maximum Sequence Length:** 512 tokens
23
  - **Output Dimensionality:** 1024 tokens
24
  - **Similarity Function:** Cosine Similarity
25
- <!-- - **Training Dataset:** Unknown -->
26
- <!-- - **Language:** Unknown -->
27
- <!-- - **License:** Unknown -->
28
 
29
  ### Model Sources
30
 
 
 
31
  - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
32
- - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
33
- - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
34
 
35
  ### Full Model Architecture
36
 
@@ -57,7 +63,7 @@ Then you can load this model and run inference.
57
  from sentence_transformers import SentenceTransformer
58
 
59
  # Download from the 🤗 Hub
60
- model = SentenceTransformer("sentence_transformers_model_id")
61
  # Run inference
62
  sentences = [
63
  'The weather is lovely today.',
@@ -74,42 +80,6 @@ print(similarities.shape)
74
  # [3, 3]
75
  ```
76
 
77
- <!--
78
- ### Direct Usage (Transformers)
79
-
80
- <details><summary>Click to see the direct usage in Transformers</summary>
81
-
82
- </details>
83
- -->
84
-
85
- <!--
86
- ### Downstream Usage (Sentence Transformers)
87
-
88
- You can finetune this model on your own dataset.
89
-
90
- <details><summary>Click to expand</summary>
91
-
92
- </details>
93
- -->
94
-
95
- <!--
96
- ### Out-of-Scope Use
97
-
98
- *List how the model may foreseeably be misused and address what users ought not to do with the model.*
99
- -->
100
-
101
- <!--
102
- ## Bias, Risks and Limitations
103
-
104
- *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
105
- -->
106
-
107
- <!--
108
- ### Recommendations
109
-
110
- *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
111
- -->
112
-
113
  ## Training Details
114
 
115
  ### Framework Versions
@@ -124,21 +94,11 @@ You can finetune this model on your own dataset.
124
  ## Citation
125
 
126
  ### BibTeX
127
-
128
- <!--
129
- ## Glossary
130
-
131
- *Clearly define terms in order to be accessible across audiences.*
132
- -->
133
-
134
- <!--
135
- ## Model Card Authors
136
-
137
- *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
138
- -->
139
-
140
- <!--
141
- ## Model Card Contact
142
-
143
- *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
144
- -->
 
1
  ---
2
+ datasets:
3
+ - Jasaxion/LexSemBridge_eval
4
+ language:
5
+ - en
6
  library_name: sentence-transformers
7
+ license: apache-2.0
8
+ pipeline_tag: feature-extraction
9
  tags:
10
  - sentence-transformers
11
  - sentence-similarity
12
  - feature-extraction
13
+ - lexsembridge
14
  widget: []
15
  ---
16
 
17
+ # LexSemBridge: Fine-Grained Dense Representation Enhancement through Token-Aware Embedding Augmentation
18
 
19
+ This model implements **LexSemBridge**, a unified framework that enhances dense query representations through fine-grained, input-aware vector modulation. LexSemBridge constructs latent enhancement vectors from input tokens using statistical, learned, and contextual paradigms, integrating them with dense embeddings via element-wise interaction. It operates as a plug-in without modifying the backbone encoder and naturally extends to both text and vision modalities, aiming to improve performance on fine-grained retrieval tasks where precise keyword alignment and span-level localization are crucial.
20
+
21
+ The model is based on the paper [LexSemBridge: Fine-Grained Dense Representation Enhancement through Token-Aware Embedding Augmentation](https://huggingface.co/papers/2508.17858).
22
+
23
+ For the official code and further details, please refer to the [GitHub repository](https://github.com/Jasaxion/LexSemBridge/).
24
 
25
  ## Model Details
26
 
27
  ### Model Description
28
  - **Model Type:** Sentence Transformer
 
29
  - **Maximum Sequence Length:** 512 tokens
30
  - **Output Dimensionality:** 1024 tokens
31
  - **Similarity Function:** Cosine Similarity
 
 
 
32
 
33
  ### Model Sources
34
 
35
+ - **Paper:** [LexSemBridge: Fine-Grained Dense Representation Enhancement through Token-Aware Embedding Augmentation](https://huggingface.co/papers/2508.17858)
36
+ - **Code/GitHub Repository:** https://github.com/Jasaxion/LexSemBridge/
37
  - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
38
+ - **Repository (Sentence Transformers Library):** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
39
+ - **Hugging Face (Sentence Transformers Models):** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
40
 
41
  ### Full Model Architecture
42
 
 
63
  from sentence_transformers import SentenceTransformer
64
 
65
  # Download from the 🤗 Hub
66
+ model = SentenceTransformer("Jasaxion/LexSemBridge_CLR_snowflake") # Example: LexSemBridge-CLR-snowflake model
67
  # Run inference
68
  sentences = [
69
  'The weather is lovely today.',
 
80
  # [3, 3]
81
  ```
82
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
83
  ## Training Details
84
 
85
  ### Framework Versions
 
94
  ## Citation
95
 
96
  ### BibTeX
97
+ ```bibtex
98
+ @article{wu2025lexsembridge,
99
+ title={LexSemBridge: Fine-Grained Dense Representation Enhancement through Token-Aware Embedding Augmentation},
100
+ author={Wu, Zhiyong and Wu, Zhenyu and Xu, Fangzhi and Wang, Yian and Sun, Qiushi and Jia, Chengyou and Cheng, Kanzhi and Ding, Zichen and Chen, Liheng and Liang, Paul Pu and others},
101
+ journal={arXiv preprint arXiv:2508.17858},
102
+ year={2025}
103
+ }
104
+ ```