Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks
Paper
•
1908.10084
•
Published
•
9
This is a sentence-transformers model finetuned from sentence-transformers/all-MiniLM-L6-v2. It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
SentenceTransformer(
(0): Transformer({'max_seq_length': 256, 'do_lower_case': False, 'architecture': 'BertModel'})
(1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
(2): Normalize()
)
First install the Sentence Transformers library:
pip install -U sentence-transformers
Then you can load this model and run inference.
from sentence_transformers import SentenceTransformer
# Download from the 🤗 Hub
model = SentenceTransformer("its-zion-18/projfinetuned")
# Run inference
sentences = [
'Manager, Social Sales & Newsletters',
'Manager, Social Sales & Newsletters Collaborate with the sales team to drive revenue and sales for social advertising business, ensuring revenue goals are hit quarterly and annually.\nManage day-to-day sales requests providing targeting information and inventory requests to sales planners & reps to assist in their pre-sale initiatives\nCollaborate with the sales team to drive revenue and sales for newsletter partnerships, ensuring revenue goals are hit quarterly and annually.\nAttend live events and partner with stakeholders to ensure contracted social elements for clients are operating smoothly when necessary\nManage communications and organization from pre-sale to post-sale including managing tracker of programs, working on brainstorms in pre-sale, joining kick off calls in post-sale, looping in social activation and newsletter teams and providing campaign details\nStrong understanding of social media best practices, trends and advertising on social platforms\nWe are holding ourselves accountable and ensuring that all decisions are made with equity top of mind. Bachelor’s Degree in Business, Marketing, Communications or a related field\n3-5 years of relevant experience\nCreative storyteller\nCreator economy experience a must\nStrong proficiency in Excel, Google Drive\nAbility to multi-task, efficiently manage time and prioritize deliverables\nForbes has estimated the compensation range set forth above in good faith.\xa0 The compensation range is what we believe we will offer, and ultimately pay, a successful candidate.\xa0 In determining this range, we consider the experience, level of education (if applicable to the role), knowledge, skills, and abilities required to be had by a successful candidate as well as the budget and the company’s pay rates, generally. This said, we may have to make changes to our compensation estimates and job descriptions from time to time and we expressly reserve the right to do so.\xa0 Should we make any such changes, this advertisement will be revised to reflect such revisions.\xa0 We encourage you to occasionally re-visit this advertisement to ensure that you are abreast of any changes.\xa0 Thank you for your interest in joining Forbes!\nForbes aims to offer employees the flexibility they need in order to be successful. Some positions may require candidates to be based in a specific location for consideration while some roles may be fully remote (within the U.S.) if it aligns with the needs of the position. This position is only open to candidates residing in California, Colorado, Connecticut, District of Columbia, Florida, Georgia, Maine, Maryland, Massachusetts, New Jersey, New York, North Carolina, Pennsylvania, South Carolina, Tennessee, Texas, & Washington. Due to business operations and compliance requirements, we are unable to consider applicants based outside these states at this time.\nCreate a Job Alert\nInterested in building your career at Forbes? Get future opportunities sent straight to your email.\nAccepted file types: pdf, doc, docx, txt, rtf Forbes is an iconic global media brand that has symbolized success for over a century. Fueled by journalism that informs and inspires, Forbes spotlights the doers and doings shaping industries, achieving success and making an impact on the world. Forbes connects and convenes the most influential communities ranging from billionaires, business leaders and rising entrepreneurs to creators and innovators. The Forbes brand reaches more than 140 million people monthly worldwide through its trusted journalism, signature ForbesLive events and 49 licensed local editions in 81 countries.\nForbes Media is hiring a Manager, Social Sales & Newsletters to join our AI & Strategic Platforms team. This role will focus on driving revenue, strategic business development and representing the social and newsletter products at Forbes. This includes collaborating with key stakeholders at Forbes to drive sales, develop creative strategy for social activations, create new products and work with clients on bringing social advertising to life. This role will report into the Senior Director, Client Activation & Paid Social.\nOur office operates as a remote and hybrid workforce, with in-person collaboration expected for certain roles. Based on the responsibilities of this position, a hybrid work schedule with a minimum of one in-person collaboration day per week at our Jersey City, NJ headquarters is expected.\nParticipating in client-facing conversations as a social media and newsletter expert on behalf of Forbes when needed\nBrainstorm and ideate social & newsletter packages attached to Forbes’ products including editorial, live partnerships, content and social only\nBrainstorm and bring to market new social activations and opportunities for Forbes Live Event advertisers\nLiaison with key Forbes stakeholders including sales reps, account managers, sales planners, editorial social, live programming, integrated marketing, video producers, etc.\nPartner with external partners including LinkedIn, Meta, TikTok, influencer talent agencies, etc.\nForbes is committed to diversity, equity and inclusion and uses our various platforms to help the world build back with equity at its core. As the leading business media brand, we have a responsibility to ensure our content and experiences reflect the diverse audiences we serve. We have the unique ability to do good, leveraging our considerable editorial voice to highlight inequities and challenge systems.',
'Firmware Engineer Evaluate, benchmark, and select SoC/MCU platforms suitable for vision workloads and potential cloud integration.\nPrototype and deploy vision pipelines: capture raw image frames, perform preprocessing (ISP, color conversions), and manage AI inference models either locally or via cloud services.\nCollaborate with hardware teams to select optimal image sensors, lenses, and camera modules for object detection and classification tasks.\nImplement low-level drivers (I²C/SPI, GPIO, PWM) and incorporate them into device trees and board-support packages.\nWork with other team members to define and implement inter-processor communication protocols (SPI/UART/I²C) for sensor data exchange and system telemetry.\nEnsure coexistence of camera subsystems with other embedded peripherals (touchscreen, motors, sensors, connectivity modules) without performance degradation.\nExperience with OTA firmware updates for MCUs (bootloaders) and SoCs (partition management). Bachelor’s or Master’s degree in Computer Science, Electrical Engineering, or related discipline, or equivalent experience.\n5+ years experience in embedded firmware engineering, including at least 3 years focused on camera or vision applications.\nStrong proficiency in C/C++ (Linux kernel drivers or RTOS firmware), with knowledge of memory management, DMA, interrupts, and real-time system constraints.\nHands-on experience with camera interfaces: MIPI-CSI-2, USB Video Class (UVC), or parallel RGB; comfortable working with image sensor datasheets, register programming, and sensor configuration.\nFamiliarity with vision frameworks (OpenCV, TensorRT, TensorFlow Lite, Edge TPU) or cloud-based AI processing platforms (AWS, Azure, Google Cloud); capable of prototyping object detection or segmentation pipelines, profiling performance, and suggesting optimizations.\nSolid understanding of I²C, SPI, UART, and GPIO interfaces.\nProven prototyping ability—quickly translating high-level performance targets into functional prototypes and effectively evaluating hardware-software tradeoffs.\nSkilled in using logic analyzers, oscilloscopes, and debugging tools (JTAG, GDB) to troubleshoot camera issues.\nExcellent communication skills, capable of producing clear documentation, architecture diagrams, and engaging in effective cross-team collaboration.\nExperience with camera calibration techniques, including lens distortion correction and color-space conversions.\nExperience with integrating touchscreen panels (LVDS/DSI/HDMI) and touch controllers (I²C/SPI), and developing responsive user interfaces.\nDevelop and optimize embedded UI frameworks (e.g., Qt Embedded, LVGL).\nFamiliarity with Linux system architectures, cross-compilation, build systems, and kernel module integration.\nPrevious experience in startups or rapid prototyping environments.\nFamiliarity with Agile methodologies, version control (GitHub), CI/CD pipelines, and project management tools (JIRA).\nInterested in building your career at Mill? Get future opportunities sent straight to your email.\nAccepted file types: pdf, doc, docx, txt, rtf Mill is all about answering a simple question: how can we prevent waste? Less waste can save time, money, energy, maybe even our planet. And there’s no better place to start than food. Food waste is one of the most solvable climate problems facing us today. Plus, our trash really stinks. It’s gross, heavy, and our least favorite chore. At Mill we are striving to build a better environment for all, as we take on climate and kitchen change.\nWe are seeking an extraordinary Firmware Engineer to lead the vision system development for our next-generation product. You will evaluate SoC and MCU platforms, and develop camera-based AI applications for object detection (either on-device or cloud-based). Collaborating closely with hardware designers, connectivity experts, and system engineers, your work will be crucial in turning innovative concepts into functional prototypes and products.',
]
embeddings = model.encode(sentences)
print(embeddings.shape)
# [3, 384]
# Get the similarity scores for the embeddings
similarities = model.similarity(embeddings, embeddings)
print(similarities)
# tensor([[ 1.0000, 0.7013, -0.0001],
# [ 0.7013, 1.0000, 0.1803],
# [ -0.0001, 0.1803, 1.0000]])
sentence_0 and sentence_1| sentence_0 | sentence_1 | |
|---|---|---|
| type | string | string |
| details |
|
|
| sentence_0 | sentence_1 |
|---|---|
Electronics Engineer |
Electronics Engineer National Telecommunications and Information Administration Perform basic or applied research or engineering in such areas as propagation theory, scattering and diffraction, antenna design and optimization, radiated electromagnetic interference, and microwave systems. |
Procurement Agent |
Procurement Agent Boeing Communicate performance expectations and metrics to the supplier |
Digital Design Developer, Opinion |
Digital Design Developer, Opinion The New York Times Company Office: New York, NY Department: Editing |
MultipleNegativesRankingLoss with these parameters:{
"scale": 20.0,
"similarity_fct": "cos_sim",
"gather_across_devices": false
}
per_device_train_batch_size: 16per_device_eval_batch_size: 16num_train_epochs: 1multi_dataset_batch_sampler: round_robinoverwrite_output_dir: Falsedo_predict: Falseeval_strategy: noprediction_loss_only: Trueper_device_train_batch_size: 16per_device_eval_batch_size: 16per_gpu_train_batch_size: Noneper_gpu_eval_batch_size: Nonegradient_accumulation_steps: 1eval_accumulation_steps: Nonetorch_empty_cache_steps: Nonelearning_rate: 5e-05weight_decay: 0.0adam_beta1: 0.9adam_beta2: 0.999adam_epsilon: 1e-08max_grad_norm: 1num_train_epochs: 1max_steps: -1lr_scheduler_type: linearlr_scheduler_kwargs: {}warmup_ratio: 0.0warmup_steps: 0log_level: passivelog_level_replica: warninglog_on_each_node: Truelogging_nan_inf_filter: Truesave_safetensors: Truesave_on_each_node: Falsesave_only_model: Falserestore_callback_states_from_checkpoint: Falseno_cuda: Falseuse_cpu: Falseuse_mps_device: Falseseed: 42data_seed: Nonejit_mode_eval: Falseuse_ipex: Falsebf16: Falsefp16: Falsefp16_opt_level: O1half_precision_backend: autobf16_full_eval: Falsefp16_full_eval: Falsetf32: Nonelocal_rank: 0ddp_backend: Nonetpu_num_cores: Nonetpu_metrics_debug: Falsedebug: []dataloader_drop_last: Falsedataloader_num_workers: 0dataloader_prefetch_factor: Nonepast_index: -1disable_tqdm: Falseremove_unused_columns: Truelabel_names: Noneload_best_model_at_end: Falseignore_data_skip: Falsefsdp: []fsdp_min_num_params: 0fsdp_config: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}fsdp_transformer_layer_cls_to_wrap: Noneaccelerator_config: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}parallelism_config: Nonedeepspeed: Nonelabel_smoothing_factor: 0.0optim: adamw_torch_fusedoptim_args: Noneadafactor: Falsegroup_by_length: Falselength_column_name: lengthddp_find_unused_parameters: Noneddp_bucket_cap_mb: Noneddp_broadcast_buffers: Falsedataloader_pin_memory: Truedataloader_persistent_workers: Falseskip_memory_metrics: Trueuse_legacy_prediction_loop: Falsepush_to_hub: Falseresume_from_checkpoint: Nonehub_model_id: Nonehub_strategy: every_savehub_private_repo: Nonehub_always_push: Falsehub_revision: Nonegradient_checkpointing: Falsegradient_checkpointing_kwargs: Noneinclude_inputs_for_metrics: Falseinclude_for_metrics: []eval_do_concat_batches: Truefp16_backend: autopush_to_hub_model_id: Nonepush_to_hub_organization: Nonemp_parameters: auto_find_batch_size: Falsefull_determinism: Falsetorchdynamo: Noneray_scope: lastddp_timeout: 1800torch_compile: Falsetorch_compile_backend: Nonetorch_compile_mode: Noneinclude_tokens_per_second: Falseinclude_num_input_tokens_seen: Falseneftune_noise_alpha: Noneoptim_target_modules: Nonebatch_eval_metrics: Falseeval_on_start: Falseuse_liger_kernel: Falseliger_kernel_config: Noneeval_use_gather_object: Falseaverage_tokens_across_devices: Falseprompts: Nonebatch_sampler: batch_samplermulti_dataset_batch_sampler: round_robinrouter_mapping: {}learning_rate_mapping: {}@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "https://arxiv.org/abs/1908.10084",
}
@misc{henderson2017efficient,
title={Efficient Natural Language Response Suggestion for Smart Reply},
author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
year={2017},
eprint={1705.00652},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
Base model
sentence-transformers/all-MiniLM-L6-v2