Sentence Similarity
sentence-transformers
Safetensors
bert
feature-extraction
Generated from Trainer
dataset_size:81041718
loss:CoSENTLoss
text-embeddings-inference
Instructions to use samheym/GerBi-Encoder with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- sentence-transformers
How to use samheym/GerBi-Encoder with sentence-transformers:
from sentence_transformers import SentenceTransformer model = SentenceTransformer("samheym/GerBi-Encoder") sentences = [ "Welcher Whisky wird für blaues Blut verwendet?", "Versuchen Sie auch: 1 Blue Bloods Single Malt Scotch. 2 Scotch betrunken auf blauem Blut. 3 Finnerty's Scotch Whisky Blue Bloods. 4 trinkt Tom Selleck Scotch. 5 blaue Blut Scotch.", "...Menge... Ein Mol Saccharose und ein Mol Glucose beziehen sich auf die gleiche Menge an Molekülen. Denken Sie daran, dass der Maulwurf die Zähleinheit des Chemikers ist. Ein Mol von etwas ist 6.022137x10^23 Partikel einer Substanz.", "Pilgerfahrt nach Mekka. Mansa MÃ…«sÄÂ , entweder der Enkel oder der Großneffe von Sundiata, dem Begründer seiner Dynastie, bestieg 1307 den Thron. Im 17. Jahr seiner Herrschaft (1324) machte er sich auf den Weg seine berühmte Pilgerfahrt nach Mekka. Es war diese Pilgerfahrt, die die Welt für den erstaunlichen Reichtum Malis erweckte." ] embeddings = model.encode(sentences) similarities = model.similarity(embeddings, embeddings) print(similarities.shape) # [4, 4] - Notebooks
- Google Colab
- Kaggle
samheym commited on
Commit ·
33a8452
1
Parent(s): f093dd4
update readme
Browse files
README.md
CHANGED
|
@@ -92,7 +92,17 @@ library_name: sentence-transformers
|
|
| 92 |
|
| 93 |
# SentenceTransformer based on deepset/gbert-base
|
| 94 |
|
| 95 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 96 |
|
| 97 |
## Model Details
|
| 98 |
|
|
|
|
| 92 |
|
| 93 |
# SentenceTransformer based on deepset/gbert-base
|
| 94 |
|
| 95 |
+
## Model Description
|
| 96 |
+
|
| 97 |
+
This model was trained as part of a Bachelor's thesis evaluating and comparing different retrieval architectures, specifically ColBERT, Cross-Encoders, and Bi-Encoders. The primary focus of this work was to assess retrieval effectiveness in a German-language information retrieval setting.
|
| 98 |
+
Intended Use
|
| 99 |
+
|
| 100 |
+
The model is intended for research and experimentation in the field of Information Retrieval (IR). It can be used to analyze retrieval quality in contrast to other architectures or as a reference for future work on ColBERT in German.
|
| 101 |
+
Limitations & Warnings
|
| 102 |
+
|
| 103 |
+
## ⚠ Not for Production Use
|
| 104 |
+
This model is experimental and has not been optimized for production use. Performance, robustness, and scalability considerations have not been fully addressed, and real-world deployment is strongly discouraged.
|
| 105 |
+
|
| 106 |
|
| 107 |
## Model Details
|
| 108 |
|