Text Ranking
sentence-transformers
Safetensors
multilingual
new
cross-encoder
reranker
Generated from Trainer
dataset_size:16862
loss:BinaryCrossEntropyLoss
custom_code
Eval Results (legacy)
text-embeddings-inference
Instructions to use cometadata/gte-multilingual-reranker-affiliations with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- sentence-transformers
How to use cometadata/gte-multilingual-reranker-affiliations with sentence-transformers:
from sentence_transformers import CrossEncoder model = CrossEncoder("cometadata/gte-multilingual-reranker-affiliations", trust_remote_code=True) query = "Which planet is known as the Red Planet?" passages = [ "Venus is often called Earth's twin because of its similar size and proximity.", "Mars, known for its reddish appearance, is often referred to as the Red Planet.", "Jupiter, the largest planet in our solar system, has a prominent red spot.", "Saturn, famous for its rings, is sometimes mistaken for the Red Planet." ] scores = model.predict([(query, passage) for passage in passages]) print(scores) - Notebooks
- Google Colab
- Kaggle
Welcome to the community
The community tab is the place to discuss and collaborate with the HF community!