Update README.md
Browse files
README.md
CHANGED
|
@@ -38,14 +38,17 @@ model-index:
|
|
| 38 |
value: 0.92
|
| 39 |
name: Cosine Ap
|
| 40 |
---
|
|
|
|
| 41 |
|
| 42 |
-
#
|
|
|
|
|
|
|
| 43 |
|
| 44 |
This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [Alibaba-NLP/gte-modernbert-base](https://huggingface.co/Alibaba-NLP/gte-modernbert-base) on the [Quora](https://www.kaggle.com/datasets/quora/question-pairs-dataset) dataset. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity for the purpose of semantic caching.
|
| 45 |
|
| 46 |
-
|
| 47 |
|
| 48 |
-
|
| 49 |
- **Model Type:** Sentence Transformer
|
| 50 |
- **Base model:** [Alibaba-NLP/gte-modernbert-base](https://huggingface.co/Alibaba-NLP/gte-modernbert-base) <!-- at revision bc02f0a92d1b6dd82108036f6cb4b7b423fb7434 -->
|
| 51 |
- **Maximum Sequence Length:** 8192 tokens
|
|
@@ -56,13 +59,13 @@ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [A
|
|
| 56 |
<!-- - **Language:** Unknown -->
|
| 57 |
<!-- - **License:** Unknown -->
|
| 58 |
|
| 59 |
-
|
| 60 |
|
| 61 |
- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
|
| 62 |
- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
|
| 63 |
- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
|
| 64 |
|
| 65 |
-
|
| 66 |
|
| 67 |
```
|
| 68 |
SentenceTransformer(
|
|
@@ -71,7 +74,7 @@ SentenceTransformer(
|
|
| 71 |
)
|
| 72 |
```
|
| 73 |
|
| 74 |
-
|
| 75 |
|
| 76 |
First install the Sentence Transformers library:
|
| 77 |
|
|
@@ -101,7 +104,7 @@ print(similarities.shape)
|
|
| 101 |
|
| 102 |
```
|
| 103 |
|
| 104 |
-
|
| 105 |
|
| 106 |
|
| 107 |
| Metric | Value |
|
|
@@ -113,27 +116,27 @@ print(similarities.shape)
|
|
| 113 |
| **cosine_ap** | 0.92 |
|
| 114 |
|
| 115 |
|
| 116 |
-
|
| 117 |
|
| 118 |
-
|
| 119 |
|
| 120 |
* Dataset: [Quora](https://www.kaggle.com/datasets/quora/question-pairs-dataset)
|
| 121 |
* Size: 323491 training samples
|
| 122 |
* Columns: <code>question_1</code>, <code>question_2</code>, and <code>label</code>
|
| 123 |
|
| 124 |
-
|
| 125 |
|
| 126 |
-
|
| 127 |
|
| 128 |
* Dataset: [Quora](https://www.kaggle.com/datasets/quora/question-pairs-dataset)
|
| 129 |
* Size: 53486 evaluation samples
|
| 130 |
* Columns: <code>question_1</code>, <code>question_2</code>, and <code>label</code>
|
| 131 |
|
| 132 |
-
|
| 133 |
|
| 134 |
-
|
| 135 |
|
| 136 |
-
|
| 137 |
```bibtex
|
| 138 |
@inproceedings{langcache-embed-v1,
|
| 139 |
title = "Advancing Semantic Caching for LLMs with Domain-Specific Embeddings and Synthetic Data",
|
|
@@ -144,7 +147,7 @@ print(similarities.shape)
|
|
| 144 |
}
|
| 145 |
```
|
| 146 |
|
| 147 |
-
|
| 148 |
```bibtex
|
| 149 |
@inproceedings{reimers-2019-sentence-bert,
|
| 150 |
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
|
|
|
|
| 38 |
value: 0.92
|
| 39 |
name: Cosine Ap
|
| 40 |
---
|
| 41 |
+
# WARNING: This is an outdated model.
|
| 42 |
|
| 43 |
+
# 🚀 Check out [our new v3-small model](https://huggingface.co/redis/langcache-embed-v3-small), trained for improved inference speed, lighter footprint, and better semantic matching for caching.
|
| 44 |
+
---
|
| 45 |
+
## Redis semantic caching embedding model based on Alibaba-NLP/gte-modernbert-base
|
| 46 |
|
| 47 |
This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [Alibaba-NLP/gte-modernbert-base](https://huggingface.co/Alibaba-NLP/gte-modernbert-base) on the [Quora](https://www.kaggle.com/datasets/quora/question-pairs-dataset) dataset. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity for the purpose of semantic caching.
|
| 48 |
|
| 49 |
+
### Model Details
|
| 50 |
|
| 51 |
+
#### Model Description
|
| 52 |
- **Model Type:** Sentence Transformer
|
| 53 |
- **Base model:** [Alibaba-NLP/gte-modernbert-base](https://huggingface.co/Alibaba-NLP/gte-modernbert-base) <!-- at revision bc02f0a92d1b6dd82108036f6cb4b7b423fb7434 -->
|
| 54 |
- **Maximum Sequence Length:** 8192 tokens
|
|
|
|
| 59 |
<!-- - **Language:** Unknown -->
|
| 60 |
<!-- - **License:** Unknown -->
|
| 61 |
|
| 62 |
+
#### Model Sources
|
| 63 |
|
| 64 |
- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
|
| 65 |
- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
|
| 66 |
- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
|
| 67 |
|
| 68 |
+
#### Full Model Architecture
|
| 69 |
|
| 70 |
```
|
| 71 |
SentenceTransformer(
|
|
|
|
| 74 |
)
|
| 75 |
```
|
| 76 |
|
| 77 |
+
### Usage
|
| 78 |
|
| 79 |
First install the Sentence Transformers library:
|
| 80 |
|
|
|
|
| 104 |
|
| 105 |
```
|
| 106 |
|
| 107 |
+
##### Binary Classification
|
| 108 |
|
| 109 |
|
| 110 |
| Metric | Value |
|
|
|
|
| 116 |
| **cosine_ap** | 0.92 |
|
| 117 |
|
| 118 |
|
| 119 |
+
#### Training Dataset
|
| 120 |
|
| 121 |
+
##### Quora
|
| 122 |
|
| 123 |
* Dataset: [Quora](https://www.kaggle.com/datasets/quora/question-pairs-dataset)
|
| 124 |
* Size: 323491 training samples
|
| 125 |
* Columns: <code>question_1</code>, <code>question_2</code>, and <code>label</code>
|
| 126 |
|
| 127 |
+
#### Evaluation Dataset
|
| 128 |
|
| 129 |
+
##### Quora
|
| 130 |
|
| 131 |
* Dataset: [Quora](https://www.kaggle.com/datasets/quora/question-pairs-dataset)
|
| 132 |
* Size: 53486 evaluation samples
|
| 133 |
* Columns: <code>question_1</code>, <code>question_2</code>, and <code>label</code>
|
| 134 |
|
| 135 |
+
### Citation
|
| 136 |
|
| 137 |
+
#### BibTeX
|
| 138 |
|
| 139 |
+
##### Redis Langcache-embed Models
|
| 140 |
```bibtex
|
| 141 |
@inproceedings{langcache-embed-v1,
|
| 142 |
title = "Advancing Semantic Caching for LLMs with Domain-Specific Embeddings and Synthetic Data",
|
|
|
|
| 147 |
}
|
| 148 |
```
|
| 149 |
|
| 150 |
+
##### Sentence Transformers
|
| 151 |
```bibtex
|
| 152 |
@inproceedings{reimers-2019-sentence-bert,
|
| 153 |
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
|