Instructions to use ml6team/cross-encoder-mmarco-german-distilbert-base with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- sentence-transformers
How to use ml6team/cross-encoder-mmarco-german-distilbert-base with sentence-transformers:
from sentence_transformers import CrossEncoder model = CrossEncoder("ml6team/cross-encoder-mmarco-german-distilbert-base") query = "Which planet is known as the Red Planet?" passages = [ "Venus is often called Earth's twin because of its similar size and proximity.", "Mars, known for its reddish appearance, is often referred to as the Red Planet.", "Jupiter, the largest planet in our solar system, has a prominent red spot.", "Saturn, famous for its rings, is sometimes mistaken for the Red Planet." ] scores = model.predict([(query, passage) for passage in passages]) print(scores) - Notebooks
- Google Colab
- Kaggle
cross-encoder-mmarco-german-distilbert-base
Model description:
This model is a fine-tuned cross-encoder on the MMARCO dataset which is the machine translated version of the MS MARCO dataset. As base model for the fine-tuning we use distilbert-base-multilingual-cased
Model input samples are tuples of the following format, either
<query, positive_paragraph> assigned to 1 or <query, negative_paragraph> assigned to 0.
The model was trained for 1 epoch.
Model usage
The cross-encoder model can be used like this:
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name')
scores = model.predict([('Query 1', 'Paragraph 1'), ('Query 2', 'Paragraph 2')])
The model will predict scores for the pairs ('Query 1', 'Paragraph 1') and ('Query 2', 'Paragraph 2').
For more details on the usage of the cross-encoder models have a look into the Sentence-Transformers
Model Performance:
Model evaluation was done on 2000 evaluation paragraphs of the dataset.
| Accuracy | F1-Score | Precision | Recall |
|---|---|---|---|
| 89.70 | 86.82 | 86.82 | 93.50 |
- Downloads last month
- 8,216