metadata
tags:
- sentence-transformers
- sentence-similarity
- feature-extraction
- generated_from_trainer
- dataset_size:5000
- loss:CosineSimilarityLoss
base_model: sentence-transformers/all-MiniLM-L6-v2
widget:
- source_sentence: >-
looking Product Manager expertise AWS Cybersecurity JavaScript Cloud
Architecture candidate responsible designing implementing maintaining
solutions using modern technologies
sentences:
- >-
Emily Barry professional skilled JavaScript Machine Learning Kubernetes
Computer Vision Experienced working multiple projects involving cloud
technologies modern software development practices
- >-
Stephen Baker professional skilled React AWS Node.js NLP Experienced
working multiple projects involving cloud technologies modern software
development practices
- >-
James Jackson professional skilled Node.js Cybersecurity Kubernetes
Docker Experienced working multiple projects involving cloud
technologies modern software development practices
- source_sentence: >-
looking Software Engineer expertise AWS TensorFlow NLP Node.js candidate
responsible designing implementing maintaining solutions using modern
technologies
sentences:
- >-
Jennifer Thompson professional skilled JavaScript TensorFlow Computer
Vision Django Experienced working multiple projects involving cloud
technologies modern software development practices
- >-
Lisa Bell professional skilled Python TensorFlow Computer Vision Machine
Learning Experienced working multiple projects involving cloud
technologies modern software development practices
- >-
Susan Rogers professional skilled Docker Cybersecurity Machine Learning
Python Experienced working multiple projects involving cloud
technologies modern software development practices
- source_sentence: >-
looking DevOps Engineer expertise Cybersecurity Machine Learning SQL
TensorFlow candidate responsible designing implementing maintaining
solutions using modern technologies
sentences:
- >-
Kenneth Jones professional skilled NLP Node.js Cybersecurity Cloud
Architecture Experienced working multiple projects involving cloud
technologies modern software development practices
- >-
Matthew Mcintyre professional skilled NoSQL Kubernetes React Docker
Experienced working multiple projects involving cloud technologies
modern software development practices
- >-
William Wilson professional skilled SQL Kubernetes CI/CD Security
Analysis Experienced working multiple projects involving cloud
technologies modern software development practices
- source_sentence: >-
looking Software Engineer expertise Cybersecurity NLP SQL Django candidate
responsible designing implementing maintaining solutions using modern
technologies
sentences:
- >-
Daniel Stewart professional skilled JavaScript Python Cybersecurity
TensorFlow Experienced working multiple projects involving cloud
technologies modern software development practices
- >-
Kristy Massey MD professional skilled Django Security Analysis
JavaScript Cybersecurity Experienced working multiple projects involving
cloud technologies modern software development practices
- >-
Melanie Sutton professional skilled Django CI/CD JavaScript SQL
Experienced working multiple projects involving cloud technologies
modern software development practices
- source_sentence: >-
looking AI Researcher expertise CI/CD Docker TensorFlow JavaScript
candidate responsible designing implementing maintaining solutions using
modern technologies
sentences:
- >-
Dr. William Ramirez professional skilled NoSQL React CI/CD Cloud
Architecture Experienced working multiple projects involving cloud
technologies modern software development practices
- >-
Rebecca Wiley professional skilled Python Kubernetes Node.js JavaScript
Experienced working multiple projects involving cloud technologies
modern software development practices
- >-
Roberta Graham professional skilled Flask Machine Learning Node.js
Docker Experienced working multiple projects involving cloud
technologies modern software development practices
pipeline_tag: sentence-similarity
library_name: sentence-transformers
SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
This is a sentence-transformers model finetuned from sentence-transformers/all-MiniLM-L6-v2. It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
Model Details
Model Description
- Model Type: Sentence Transformer
- Base model: sentence-transformers/all-MiniLM-L6-v2
- Maximum Sequence Length: 256 tokens
- Output Dimensionality: 384 dimensions
- Similarity Function: Cosine Similarity
Model Sources
- Documentation: Sentence Transformers Documentation
- Repository: Sentence Transformers on GitHub
- Hugging Face: Sentence Transformers on Hugging Face
Full Model Architecture
SentenceTransformer(
(0): Transformer({'max_seq_length': 256, 'do_lower_case': False}) with Transformer model: BertModel
(1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
(2): Normalize()
)
Usage
Direct Usage (Sentence Transformers)
First install the Sentence Transformers library:
pip install -U sentence-transformers
Then you can load this model and run inference.
from sentence_transformers import SentenceTransformer
# Download from the 🤗 Hub
model = SentenceTransformer("sentence_transformers_model_id")
# Run inference
sentences = [
'looking AI Researcher expertise CI/CD Docker TensorFlow JavaScript candidate responsible designing implementing maintaining solutions using modern technologies',
'Roberta Graham professional skilled Flask Machine Learning Node.js Docker Experienced working multiple projects involving cloud technologies modern software development practices',
'Rebecca Wiley professional skilled Python Kubernetes Node.js JavaScript Experienced working multiple projects involving cloud technologies modern software development practices',
]
embeddings = model.encode(sentences)
print(embeddings.shape)
# [3, 384]
# Get the similarity scores for the embeddings
similarities = model.similarity(embeddings, embeddings)
print(similarities.shape)
# [3, 3]
Training Details
Training Dataset
Unnamed Dataset
- Size: 5,000 training samples
- Columns:
sentence_0,sentence_1, andlabel - Approximate statistics based on the first 1000 samples:
sentence_0 sentence_1 label type string string float details - min: 20 tokens
- mean: 24.72 tokens
- max: 32 tokens
- min: 22 tokens
- mean: 26.26 tokens
- max: 34 tokens
- min: 0.4
- mean: 0.71
- max: 1.0
- Samples:
sentence_0 sentence_1 label looking AI Researcher expertise CI/CD Python Computer Vision Flask candidate responsible designing implementing maintaining solutions using modern technologiesDeanna Gibson professional skilled Security Analysis Node.js Machine Learning Kubernetes Experienced working multiple projects involving cloud technologies modern software development practices0.481looking Machine Learning Engineer expertise AWS Kubernetes Python Django candidate responsible designing implementing maintaining solutions using modern technologiesAmanda Johnson professional skilled AWS NLP Node.js Security Analysis Experienced working multiple projects involving cloud technologies modern software development practices0.982looking Cybersecurity Analyst expertise JavaScript Python Node.js NoSQL candidate responsible designing implementing maintaining solutions using modern technologiesAlicia Patton professional skilled Node.js TensorFlow SQL NoSQL Experienced working multiple projects involving cloud technologies modern software development practices0.597 - Loss:
CosineSimilarityLosswith these parameters:{ "loss_fct": "torch.nn.modules.loss.MSELoss" }
Training Hyperparameters
Non-Default Hyperparameters
per_device_train_batch_size: 16per_device_eval_batch_size: 16num_train_epochs: 30multi_dataset_batch_sampler: round_robin
All Hyperparameters
Click to expand
overwrite_output_dir: Falsedo_predict: Falseeval_strategy: noprediction_loss_only: Trueper_device_train_batch_size: 16per_device_eval_batch_size: 16per_gpu_train_batch_size: Noneper_gpu_eval_batch_size: Nonegradient_accumulation_steps: 1eval_accumulation_steps: Nonetorch_empty_cache_steps: Nonelearning_rate: 5e-05weight_decay: 0.0adam_beta1: 0.9adam_beta2: 0.999adam_epsilon: 1e-08max_grad_norm: 1num_train_epochs: 30max_steps: -1lr_scheduler_type: linearlr_scheduler_kwargs: {}warmup_ratio: 0.0warmup_steps: 0log_level: passivelog_level_replica: warninglog_on_each_node: Truelogging_nan_inf_filter: Truesave_safetensors: Truesave_on_each_node: Falsesave_only_model: Falserestore_callback_states_from_checkpoint: Falseno_cuda: Falseuse_cpu: Falseuse_mps_device: Falseseed: 42data_seed: Nonejit_mode_eval: Falseuse_ipex: Falsebf16: Falsefp16: Falsefp16_opt_level: O1half_precision_backend: autobf16_full_eval: Falsefp16_full_eval: Falsetf32: Nonelocal_rank: 0ddp_backend: Nonetpu_num_cores: Nonetpu_metrics_debug: Falsedebug: []dataloader_drop_last: Falsedataloader_num_workers: 0dataloader_prefetch_factor: Nonepast_index: -1disable_tqdm: Falseremove_unused_columns: Truelabel_names: Noneload_best_model_at_end: Falseignore_data_skip: Falsefsdp: []fsdp_min_num_params: 0fsdp_config: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}tp_size: 0fsdp_transformer_layer_cls_to_wrap: Noneaccelerator_config: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}deepspeed: Nonelabel_smoothing_factor: 0.0optim: adamw_torchoptim_args: Noneadafactor: Falsegroup_by_length: Falselength_column_name: lengthddp_find_unused_parameters: Noneddp_bucket_cap_mb: Noneddp_broadcast_buffers: Falsedataloader_pin_memory: Truedataloader_persistent_workers: Falseskip_memory_metrics: Trueuse_legacy_prediction_loop: Falsepush_to_hub: Falseresume_from_checkpoint: Nonehub_model_id: Nonehub_strategy: every_savehub_private_repo: Nonehub_always_push: Falsegradient_checkpointing: Falsegradient_checkpointing_kwargs: Noneinclude_inputs_for_metrics: Falseinclude_for_metrics: []eval_do_concat_batches: Truefp16_backend: autopush_to_hub_model_id: Nonepush_to_hub_organization: Nonemp_parameters:auto_find_batch_size: Falsefull_determinism: Falsetorchdynamo: Noneray_scope: lastddp_timeout: 1800torch_compile: Falsetorch_compile_backend: Nonetorch_compile_mode: Noneinclude_tokens_per_second: Falseinclude_num_input_tokens_seen: Falseneftune_noise_alpha: Noneoptim_target_modules: Nonebatch_eval_metrics: Falseeval_on_start: Falseuse_liger_kernel: Falseeval_use_gather_object: Falseaverage_tokens_across_devices: Falseprompts: Nonebatch_sampler: batch_samplermulti_dataset_batch_sampler: round_robin
Training Logs
| Epoch | Step | Training Loss |
|---|---|---|
| 1.5974 | 500 | 0.0324 |
| 3.1949 | 1000 | 0.0298 |
| 4.7923 | 1500 | 0.028 |
| 6.3898 | 2000 | 0.025 |
| 7.9872 | 2500 | 0.0229 |
| 9.5847 | 3000 | 0.0198 |
| 11.1821 | 3500 | 0.0179 |
| 12.7796 | 4000 | 0.0156 |
| 14.3770 | 4500 | 0.014 |
| 15.9744 | 5000 | 0.0127 |
| 17.5719 | 5500 | 0.0115 |
| 19.1693 | 6000 | 0.0104 |
| 20.7668 | 6500 | 0.0098 |
| 22.3642 | 7000 | 0.009 |
| 23.9617 | 7500 | 0.0086 |
| 25.5591 | 8000 | 0.0082 |
| 27.1565 | 8500 | 0.0078 |
| 28.7540 | 9000 | 0.0076 |
Framework Versions
- Python: 3.11.12
- Sentence Transformers: 3.4.1
- Transformers: 4.51.3
- PyTorch: 2.6.0+cu124
- Accelerate: 1.5.2
- Datasets: 3.5.0
- Tokenizers: 0.21.1
Citation
BibTeX
Sentence Transformers
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "https://arxiv.org/abs/1908.10084",
}