radoslavralev commited on
Commit
667f199
·
verified ·
1 Parent(s): 7dd7e00

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -15
README.md CHANGED
@@ -38,14 +38,17 @@ model-index:
38
  value: 0.92
39
  name: Cosine Ap
40
  ---
 
41
 
42
- # Redis semantic caching embedding model based on Alibaba-NLP/gte-modernbert-base
 
 
43
 
44
  This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [Alibaba-NLP/gte-modernbert-base](https://huggingface.co/Alibaba-NLP/gte-modernbert-base) on the [Quora](https://www.kaggle.com/datasets/quora/question-pairs-dataset) dataset. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity for the purpose of semantic caching.
45
 
46
- ## Model Details
47
 
48
- ### Model Description
49
  - **Model Type:** Sentence Transformer
50
  - **Base model:** [Alibaba-NLP/gte-modernbert-base](https://huggingface.co/Alibaba-NLP/gte-modernbert-base) <!-- at revision bc02f0a92d1b6dd82108036f6cb4b7b423fb7434 -->
51
  - **Maximum Sequence Length:** 8192 tokens
@@ -56,13 +59,13 @@ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [A
56
  <!-- - **Language:** Unknown -->
57
  <!-- - **License:** Unknown -->
58
 
59
- ### Model Sources
60
 
61
  - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
62
  - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
63
  - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
64
 
65
- ### Full Model Architecture
66
 
67
  ```
68
  SentenceTransformer(
@@ -71,7 +74,7 @@ SentenceTransformer(
71
  )
72
  ```
73
 
74
- ## Usage
75
 
76
  First install the Sentence Transformers library:
77
 
@@ -101,7 +104,7 @@ print(similarities.shape)
101
 
102
  ```
103
 
104
- #### Binary Classification
105
 
106
 
107
  | Metric | Value |
@@ -113,27 +116,27 @@ print(similarities.shape)
113
  | **cosine_ap** | 0.92 |
114
 
115
 
116
- ### Training Dataset
117
 
118
- #### Quora
119
 
120
  * Dataset: [Quora](https://www.kaggle.com/datasets/quora/question-pairs-dataset)
121
  * Size: 323491 training samples
122
  * Columns: <code>question_1</code>, <code>question_2</code>, and <code>label</code>
123
 
124
- ### Evaluation Dataset
125
 
126
- #### Quora
127
 
128
  * Dataset: [Quora](https://www.kaggle.com/datasets/quora/question-pairs-dataset)
129
  * Size: 53486 evaluation samples
130
  * Columns: <code>question_1</code>, <code>question_2</code>, and <code>label</code>
131
 
132
- ## Citation
133
 
134
- ### BibTeX
135
 
136
- #### Redis Langcache-embed Models
137
  ```bibtex
138
  @inproceedings{langcache-embed-v1,
139
  title = "Advancing Semantic Caching for LLMs with Domain-Specific Embeddings and Synthetic Data",
@@ -144,7 +147,7 @@ print(similarities.shape)
144
  }
145
  ```
146
 
147
- #### Sentence Transformers
148
  ```bibtex
149
  @inproceedings{reimers-2019-sentence-bert,
150
  title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
 
38
  value: 0.92
39
  name: Cosine Ap
40
  ---
41
+ # WARNING: This is an outdated model.
42
 
43
+ # 🚀 Check out [our new v3-small model](https://huggingface.co/redis/langcache-embed-v3-small), trained for improved inference speed, lighter footprint, and better semantic matching for caching.
44
+ ---
45
+ ## Redis semantic caching embedding model based on Alibaba-NLP/gte-modernbert-base
46
 
47
  This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [Alibaba-NLP/gte-modernbert-base](https://huggingface.co/Alibaba-NLP/gte-modernbert-base) on the [Quora](https://www.kaggle.com/datasets/quora/question-pairs-dataset) dataset. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity for the purpose of semantic caching.
48
 
49
+ ### Model Details
50
 
51
+ #### Model Description
52
  - **Model Type:** Sentence Transformer
53
  - **Base model:** [Alibaba-NLP/gte-modernbert-base](https://huggingface.co/Alibaba-NLP/gte-modernbert-base) <!-- at revision bc02f0a92d1b6dd82108036f6cb4b7b423fb7434 -->
54
  - **Maximum Sequence Length:** 8192 tokens
 
59
  <!-- - **Language:** Unknown -->
60
  <!-- - **License:** Unknown -->
61
 
62
+ #### Model Sources
63
 
64
  - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
65
  - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
66
  - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
67
 
68
+ #### Full Model Architecture
69
 
70
  ```
71
  SentenceTransformer(
 
74
  )
75
  ```
76
 
77
+ ### Usage
78
 
79
  First install the Sentence Transformers library:
80
 
 
104
 
105
  ```
106
 
107
+ ##### Binary Classification
108
 
109
 
110
  | Metric | Value |
 
116
  | **cosine_ap** | 0.92 |
117
 
118
 
119
+ #### Training Dataset
120
 
121
+ ##### Quora
122
 
123
  * Dataset: [Quora](https://www.kaggle.com/datasets/quora/question-pairs-dataset)
124
  * Size: 323491 training samples
125
  * Columns: <code>question_1</code>, <code>question_2</code>, and <code>label</code>
126
 
127
+ #### Evaluation Dataset
128
 
129
+ ##### Quora
130
 
131
  * Dataset: [Quora](https://www.kaggle.com/datasets/quora/question-pairs-dataset)
132
  * Size: 53486 evaluation samples
133
  * Columns: <code>question_1</code>, <code>question_2</code>, and <code>label</code>
134
 
135
+ ### Citation
136
 
137
+ #### BibTeX
138
 
139
+ ##### Redis Langcache-embed Models
140
  ```bibtex
141
  @inproceedings{langcache-embed-v1,
142
  title = "Advancing Semantic Caching for LLMs with Domain-Specific Embeddings and Synthetic Data",
 
147
  }
148
  ```
149
 
150
+ ##### Sentence Transformers
151
  ```bibtex
152
  @inproceedings{reimers-2019-sentence-bert,
153
  title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",