Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks
Paper
• 1908.10084 • Published
• 12
This is a sentence-transformers model finetuned from BAAI/bge-m3. It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
SentenceTransformer(
(0): Transformer({'max_seq_length': 8192, 'do_lower_case': False}) with Transformer model: XLMRobertaModel
(1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
(2): Normalize()
)
First install the Sentence Transformers library:
pip install -U sentence-transformers
Then you can load this model and run inference.
from sentence_transformers import SentenceTransformer
# Download from the 🤗 Hub
model = SentenceTransformer("collaborativeearth/bge-m3_wri")
# Run inference
sentences = [
'what is the wri meat initiative?',
'Toward "Better" Meat? Aligning Meat Sourcing Strategies with Corporate Climate and Sustainability Goals Toward “Better” Meat? Aligning meat sourcing strategies with corporate climate and sustainability goals\n\nWOR L D WOR L D R E S O U R C E S R E S O U R C E S I NS T I T U T E I NS T I T U T E\n\nRICHARD WAITE is the Acting Director for Agriculture Initiatives at WRI.\n\nis a doctoral student with Oxford University’s Environmental Change Institute and a former Research Analyst for WRI’s Food and Climate Programs.\n\nCLARA CHO is the Data Analyst for the Coolfood initiative at WRI. Contact: clara.cho@wri.org.\n\nWe are pleased to acknowledge our institutional strategic partners that provide core funding to WRI: the Netherlands Ministry of Foreign Affairs, Royal Danish Ministry of Foreign Affairs, and Swedish International Development Cooperation Agency.\n\nThe authors acknowledge the following individuals for their valuable guidance and critical reviews:',
'Pilot analysis of global ecosystems: Grassland ecosystems Although GLASOD was by necessity a somewhat subjective assessment it was extremely carefully prepared by leading experts in the field. It remains the only global database on the status of human-induced soil degradation, and no other data set comes as close to defining the extent of desertification at the global scale (UNEP 1997: V).',
]
embeddings = model.encode(sentences)
print(embeddings.shape)
# [3, 1024]
# Get the similarity scores for the embeddings
similarities = model.similarity(embeddings, embeddings)
print(similarities.shape)
# [3, 3]
ir-evalInformationRetrievalEvaluator| Metric | Value |
|---|---|
| cosine_accuracy@1 | 0.3403 |
| cosine_accuracy@3 | 0.5389 |
| cosine_accuracy@5 | 0.6212 |
| cosine_accuracy@10 | 0.7122 |
| cosine_precision@1 | 0.3403 |
| cosine_precision@3 | 0.1796 |
| cosine_precision@5 | 0.1242 |
| cosine_precision@10 | 0.0712 |
| cosine_recall@1 | 0.3403 |
| cosine_recall@3 | 0.5389 |
| cosine_recall@5 | 0.6212 |
| cosine_recall@10 | 0.7122 |
| cosine_ndcg@10 | 0.5191 |
| cosine_mrr@10 | 0.458 |
| cosine_map@100 | 0.4673 |
question and answer| question | answer | |
|---|---|---|
| type | string | string |
| details |
|
|
| question | answer |
|---|---|
what is the economic case of restoration |
The Economic Case for Landscape Restoration in Latin America The Economic Case for Landscape Restoration in Latin America |
economic case of landscape restoration in latin america |
The Economic Case for Landscape Restoration in Latin America The Economic Case for Landscape Restoration in Latin America |
what is lata-american landscape |
The Economic Case for Landscape Restoration in Latin America Agriculture and forestry exports from Latin America represent about 13 percent of the global trade of food, feed, and fiber and account for a majority of employment outside large urban areas—numbers only expected to grow as Latin America is called upon to meet an increasing global demand for food. Yet, since the turn of the century, about 37 million hectares of natural forests, savannas and wetlands have been transformed to expand agriculture. Cumulative, unsustainable land-use practices have led to the degradation of about 300 million hectares, resulting in a reduction in yields and quality of production, and in losses in biomass content, soil quality, surface water hydrology, and biodiversity. Deforestation, land-use change, and unsustainable agricultural activities are also currently the largest drivers of climate change in the region, accounting for 56 percent of all greenhouse gas emissions. Today, while some progress ha... |
MultipleNegativesRankingLoss with these parameters:{
"scale": 20.0,
"similarity_fct": "cos_sim"
}
eval_strategy: stepsper_device_train_batch_size: 32learning_rate: 1e-06num_train_epochs: 2warmup_ratio: 0.1fp16: Truegradient_checkpointing: Truebatch_sampler: no_duplicatesoverwrite_output_dir: Falsedo_predict: Falseeval_strategy: stepsprediction_loss_only: Trueper_device_train_batch_size: 32per_device_eval_batch_size: 8per_gpu_train_batch_size: Noneper_gpu_eval_batch_size: Nonegradient_accumulation_steps: 1eval_accumulation_steps: Nonetorch_empty_cache_steps: Nonelearning_rate: 1e-06weight_decay: 0.0adam_beta1: 0.9adam_beta2: 0.999adam_epsilon: 1e-08max_grad_norm: 1.0num_train_epochs: 2max_steps: -1lr_scheduler_type: linearlr_scheduler_kwargs: {}warmup_ratio: 0.1warmup_steps: 0log_level: passivelog_level_replica: warninglog_on_each_node: Truelogging_nan_inf_filter: Truesave_safetensors: Truesave_on_each_node: Falsesave_only_model: Falserestore_callback_states_from_checkpoint: Falseno_cuda: Falseuse_cpu: Falseuse_mps_device: Falseseed: 42data_seed: Nonejit_mode_eval: Falseuse_ipex: Falsebf16: Falsefp16: Truefp16_opt_level: O1half_precision_backend: autobf16_full_eval: Falsefp16_full_eval: Falsetf32: Nonelocal_rank: 0ddp_backend: Nonetpu_num_cores: Nonetpu_metrics_debug: Falsedebug: []dataloader_drop_last: Falsedataloader_num_workers: 0dataloader_prefetch_factor: Nonepast_index: -1disable_tqdm: Falseremove_unused_columns: Truelabel_names: Noneload_best_model_at_end: Falseignore_data_skip: Falsefsdp: []fsdp_min_num_params: 0fsdp_config: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}tp_size: 0fsdp_transformer_layer_cls_to_wrap: Noneaccelerator_config: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}deepspeed: Nonelabel_smoothing_factor: 0.0optim: adamw_torchoptim_args: Noneadafactor: Falsegroup_by_length: Falselength_column_name: lengthddp_find_unused_parameters: Noneddp_bucket_cap_mb: Noneddp_broadcast_buffers: Falsedataloader_pin_memory: Truedataloader_persistent_workers: Falseskip_memory_metrics: Trueuse_legacy_prediction_loop: Falsepush_to_hub: Falseresume_from_checkpoint: Nonehub_model_id: Nonehub_strategy: every_savehub_private_repo: Nonehub_always_push: Falsegradient_checkpointing: Truegradient_checkpointing_kwargs: Noneinclude_inputs_for_metrics: Falseinclude_for_metrics: []eval_do_concat_batches: Truefp16_backend: autopush_to_hub_model_id: Nonepush_to_hub_organization: Nonemp_parameters: auto_find_batch_size: Falsefull_determinism: Falsetorchdynamo: Noneray_scope: lastddp_timeout: 1800torch_compile: Falsetorch_compile_backend: Nonetorch_compile_mode: Noneinclude_tokens_per_second: Falseinclude_num_input_tokens_seen: Falseneftune_noise_alpha: Noneoptim_target_modules: Nonebatch_eval_metrics: Falseeval_on_start: Falseuse_liger_kernel: Falseeval_use_gather_object: Falseaverage_tokens_across_devices: Falseprompts: Nonebatch_sampler: no_duplicatesmulti_dataset_batch_sampler: proportional| Epoch | Step | Training Loss | ir-eval_cosine_ndcg@10 |
|---|---|---|---|
| -1 | -1 | - | 0.4718 |
| 0.0389 | 100 | 0.7439 | - |
| 0.0779 | 200 | 0.6208 | - |
| 0.1168 | 300 | 0.4568 | - |
| 0.1558 | 400 | 0.3713 | - |
| 0.1947 | 500 | 0.3263 | 0.5004 |
| 0.2336 | 600 | 0.2722 | - |
| 0.2726 | 700 | 0.2521 | - |
| 0.3115 | 800 | 0.2541 | - |
| 0.3505 | 900 | 0.2348 | - |
| 0.3894 | 1000 | 0.2321 | 0.5090 |
| 0.4283 | 1100 | 0.2313 | - |
| 0.4673 | 1200 | 0.2195 | - |
| 0.5062 | 1300 | 0.2286 | - |
| 0.5452 | 1400 | 0.2188 | - |
| 0.5841 | 1500 | 0.2166 | 0.5115 |
| 0.6231 | 1600 | 0.2194 | - |
| 0.6620 | 1700 | 0.2006 | - |
| 0.7009 | 1800 | 0.1954 | - |
| 0.7399 | 1900 | 0.2157 | - |
| 0.7788 | 2000 | 0.2059 | 0.5154 |
| 0.8178 | 2100 | 0.203 | - |
| 0.8567 | 2200 | 0.1949 | - |
| 0.8956 | 2300 | 0.1943 | - |
| 0.9346 | 2400 | 0.206 | - |
| 0.9735 | 2500 | 0.2015 | 0.5175 |
| 1.0125 | 2600 | 0.1801 | - |
| 1.0514 | 2700 | 0.1867 | - |
| 1.0903 | 2800 | 0.1914 | - |
| 1.1293 | 2900 | 0.1827 | - |
| 1.1682 | 3000 | 0.1899 | 0.5165 |
| 1.2072 | 3100 | 0.1707 | - |
| 1.2461 | 3200 | 0.1872 | - |
| 1.2850 | 3300 | 0.1943 | - |
| 1.3240 | 3400 | 0.1854 | - |
| 1.3629 | 3500 | 0.1747 | 0.5182 |
| 1.4019 | 3600 | 0.1764 | - |
| 1.4408 | 3700 | 0.1866 | - |
| 1.4798 | 3800 | 0.1855 | - |
| 1.5187 | 3900 | 0.1782 | - |
| 1.5576 | 4000 | 0.1744 | 0.5181 |
| 1.5966 | 4100 | 0.1793 | - |
| 1.6355 | 4200 | 0.187 | - |
| 1.6745 | 4300 | 0.1907 | - |
| 1.7134 | 4400 | 0.1781 | - |
| 1.7523 | 4500 | 0.1825 | 0.5185 |
| 1.7913 | 4600 | 0.1981 | - |
| 1.8302 | 4700 | 0.1751 | - |
| 1.8692 | 4800 | 0.1824 | - |
| 1.9081 | 4900 | 0.1866 | - |
| 1.9470 | 5000 | 0.188 | 0.5191 |
| 1.9860 | 5100 | 0.1838 | - |
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "https://arxiv.org/abs/1908.10084",
}
@misc{henderson2017efficient,
title={Efficient Natural Language Response Suggestion for Smart Reply},
author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
year={2017},
eprint={1705.00652},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
Base model
BAAI/bge-m3