SentenceTransformer based on sentence-transformers/all-distilroberta-v1
This is a sentence-transformers model finetuned from sentence-transformers/all-distilroberta-v1 on the ai-job-embedding-finetuning dataset. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
Model Details
Model Description
- Model Type: Sentence Transformer
- Base model: sentence-transformers/all-distilroberta-v1
- Maximum Sequence Length: 512 tokens
- Output Dimensionality: 768 dimensions
- Similarity Function: Cosine Similarity
- Training Dataset:
Model Sources
Full Model Architecture
SentenceTransformer(
(0): Transformer({'max_seq_length': 512, 'do_lower_case': False, 'architecture': 'RobertaModel'})
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
(2): Normalize()
)
Usage
Direct Usage (Sentence Transformers)
First install the Sentence Transformers library:
pip install -U sentence-transformers
Then you can load this model and run inference.
from sentence_transformers import SentenceTransformer
model = SentenceTransformer("Fe2x/distilroberta-ai-job-embeddings")
queries = [
"Deep learning research, large-scale driving data, road scene understanding",
]
documents = [
"experience where customer success continues to motivate what is next.\n\nNetradyne is committed to building a world-class team of technologists and industry experts to deliver products that improve safety, increase productivity, and optimize collaboration within organizations. With growth exceeding 4x year over year, our solution is quickly being recognized as a significant disruptive technology – that has put ‘legacy’ providers in a “spin” cycle trying to catch up. Our team is growing, and we need forward-thinking, uncompromising, competitive team members to continue to facilitate our growth.\n\nAI Engineer - Deep Learning\n\nWe are looking for a highly independent and self-driven Senior Research Engineer who is passionate about pushing the boundaries of deep learning research, to join our fast-growing technology team. This person should be able to work autonomously, think creatively, and explore new ideas and approaches to tackle complex problems in the field. You will have an opportunity to work with very large-scale real-world driving data. Netradyne analyzes over 100 million miles of driving data every month, covering over 1.25 million miles of US roads. This role provides a unique opportunity to work with cutting-edge technology and tackle complex problems in the field of deep learning using vast real-world datasets. The Deep Learning Research Engineer will have the chance to make a significant impact on road safety and advance the field of deep learning research. If you are driven by curiosity and have a passion for innovation, we encourage you to apply.\n\nResponsibilities\n\nDevelop and implement deep learning algorithms to extract valuable insights from large-scale real-world vision data.Design and commercialize algorithms characterizing driving behavior.Innovate and develop proof-of-concept solutions showcasing novel capabilities.\n\n\nRequirements\n\nPh.D. in Computer Science, Electrical Engineering, or a related field with publications in top conferences (CVPR/NeurIPs/ICML/ICLR).Strong background in deep learning, machine learning, and computer vision.Excellent programming skills – Python.Proficiency in PyTorch or TensorFlow.Experience with training large models with huge datasets.Ability to take abstract product concepts and turn them into reality.Location: San Diego, CA - Hybrid\n\n\nDesired Skills\n\nExperience with image, video, and time-series data.Experience with road scene understanding (objects, lanes, interactions, signs, etc.).Experience with person/driver scene understanding (pose, distracted, eye status etc.).Experience with Predictive analytics.\n\n\nOther Essential Abilities and Skills: \n\nStrong analytical and problem-solving skills.Excellent verbal and written communication skills.Energetic or passionate about AI.Ability to work independently and as part of a team.\n\n\nEconomic Package Includes:\n\nSalary $145,000- $180,000Company Paid Health Care, Dental, and Vision CoverageIncluding Coverage for your partner and dependentsThree Health Care Plan OptionsFSA and HSA OptionsGenerous PTO and Sick Leave401(K) Disability and Life Insurance Benefits$50 phone stipend per pay period\n\nSan Diego Pay Range\n\n$145,000—$180,000 USD\n\nWe are committed to an inclusive and diverse team. Netradyne is an equal-opportunity employer. We do not discriminate based on race, color, ethnicity, ancestry, national origin, religion, sex, gender, gender identity, gender expression, sexual orientation, age, disability, veteran status, genetic information, marital status, or any legally protected status.\n\nIf there is a match between your experiences/skills and the Company's needs, we will contact you directly.\n\nNetradyne is an equal-opportunity employer.\n\nApplicants only - Recruiting agencies do not contact.\n\nCalifornia Consumer Privacy Act Notice\n\nThis notice applies if you are a resident of California (“California Consumer”) and have provided Personal Information to Netradyne that is subject to the California Consumer Privacy Act (“CCPA”). We typically collect Personal Information in the capacity of a service provider to our clients, who are responsible for providing notice to their employees and contractors and complying with CCPA requirements.\n\nDuring the past 12 months, we have collected the following categories of Personal Information: (a) identifiers; (b) biometric information (see our Biometric Data Privacy Policy for more information); (c) Internet or other electronic network activity information; (d) geolocation data; (e) Audio, electronic, visual, thermal, olfactory, or similar information; (f) professional or employment-related information (from job applicants and from clients regarding their employees and contractors); and (g) education information (from job applicants). We will not discriminate against any person that exercises any rights under the CCPA.\n\nWe have collected this Personal Information for the business purposes and commercial purposes described in this Policy, including to provide the Services to our clients, process job applications, and for marketing and promotion.\n\nThe sources of such Personal Information are you, our clients and our service providers. We have shared this information this only with our clients (if you are an employee or contractor of them) or our service providers.\n\nIf you are a California Consumer, you have the following rights under the CCPA:\n\nYou have the right to request:The categories and specific pieces of your Personal Information that we’ve collected;The categories of sources from which we collected your Personal Information;The business or commercial purposes for which we collected or sold your Personal Information; andThe categories of third parties with which we shared your Personal Information.You can submit a request to us for the following additional information:The categories of third parties to whom we’ve sold Personal Information, and the category or categories of Personal Information sold to each; andThe categories of third parties to whom we’ve disclosed Personal Information, and the category or categories of Personal Information disclosed to each.You can request that we delete the Personal Information we have collected about you, except for situations when that information is necessary for us to: provide you with a product or service that you requested; perform a contract we entered into with you; maintain the functionality or security of our systems; comply with or exercise rights provided by the law; or use the information internally in ways that are compatible with the context in which you provided the information to us, or that are reasonably aligned with your expectations based on your relationship with us.You have the right to request that we not sell your Personal Information. However, we do not offer this opt-out as we do not sell your Personal Information as that term is defined under the CCPA.\n\nYou can make a request under the CCPA by e-mailing us at privacy@netradyne.com We may request additional information from you to verify your identify. You may also designate an authorized agent to submit a request on your behalf. To do so, we will require either (1) a valid power of attorney, or (2) signed written permission from you. In the event your authorized agent is relying on signed written permission, we may also need to verify your identity and/or contact you directly to confirm permission to proceed with the request.\n\nAs noted above, if your request concerns Personal Information collected in our capacity as a service provider to a client, we are not responsible for responding to the request and may send the request to the client for a response.\n\nGoverning law\n\nThis Services are provided in the United States, and are located and targeted to persons in the United States and our policies are directed at compliance with those laws. If you are uncertain whether this Policy conflicts with the applicable local privacy laws where you are located, you should not submit your Personal Information to Netradyne.",
'QUALIFICATIONSMust-Have:Bachelor’s Degree in Computer Science, Information Systems, or related field.A minimum of 3-5 years of experience as a data engineer or in a similar role (SQL, Python, etc.)Experience working in cloud environments (AWS, Azure, etc.)Solid understanding of data governance principles and practices.Knowledge of a Data Catalog, Data Lineage, and Data Quality frameworksPrior experience with Data governance tools such as Atlan, Collibra, Alation, Manta, etc. is highly desired.Strong analytical and technical problem-solving skills.Excellent interpersonal and communication skills.Takes ownership and pride in end-to-end delivery of projects and initiatives.Comfort with a data-intensive and high transaction volume environmentDeadline-driven mindsetNice-to-have:Prior experience in Finance and Asset management domain is a plus.Prior experience with Snowflake and DBT is a plus',
'Qualifications\n\nYour Experience\n\nM.S. or Ph.D degree in Computer Science, Mathematics, Electrical Engineering or related field or equivalent military experience required8+ years industry experience in Machine Learning techniques and data analytics8+ experience in design, algorithms and data structures - Expertise with one or more of the following languages is must - Java, C++, Python, RustExperience with NLP, Recommender Systems, and LLM is strongly preferredExperience with Formal Methods toolchain (z3, cvc5, TLA+) will be a plusExcellent communication skills with the ability to influence at all levels of the organizationA self driven individual contributor and an excellent team player\n\nAdditional Information\n\nThe Team\n\nDrawing on the near real-time data collected through PAN-OS device telemetry, our industry-leading next generation insights product (AIOps for NGFW) gives large cybersecurity operators a force multiplier that provides visibility into the health of their next-generation-firewall (NGFW) devices. It enables early detection of issues at various levels of the stack via advanced time-series forecasting and anomaly detection using novel deep learning techniques. Our goal is to be able to prevent service-impacting issues in critical security infrastructure that operates 24/7/365 with zero false positives and zero false negatives.You will be working on the best large language model in the cyber security industry.\n\nOur Commitment\n\nWe’re trailblazers that dream big, take risks, and challenge cybersecurity’s status quo. It’s simple: we can’t accomplish our mission without diverse teams innovating, together.\n\nWe are committed to providing reasonable accommodations for all qualified individuals with a disability. If you require assistance or accommodation due to a disability or special need, please contact us at accommodations@paloaltonetworks.com.\n\nPalo Alto Networks is \n\nAll your information will be kept confidential according to \n\nThe compensation offered for this position will depend on qualifications, experience, and work location. For candidates who receive an offer at the posted level, the starting base salary (for non-sales roles) or base salary + commission target (for sales/commissioned roles) is expected to be between $140,100/yr to $220,600/yr. The offered compensation may also include restricted stock units and a bonus. A description of our employee benefits may be found here.\n\nIs role eligible for Immigration Sponsorship?: Yes',
]
query_embeddings = model.encode_query(queries)
document_embeddings = model.encode_document(documents)
print(query_embeddings.shape, document_embeddings.shape)
similarities = model.similarity(query_embeddings, document_embeddings)
print(similarities)
Evaluation
Metrics
Triplet
| Metric |
ai-job-validation |
ai-job-test |
| cosine_accuracy |
0.9894 |
1.0 |
Training Details
Training Dataset
ai-job-embedding-finetuning
Evaluation Dataset
ai-job-embedding-finetuning
Training Hyperparameters
Non-Default Hyperparameters
eval_strategy: steps
per_device_train_batch_size: 16
per_device_eval_batch_size: 16
learning_rate: 2e-05
num_train_epochs: 1
warmup_ratio: 0.1
batch_sampler: no_duplicates
All Hyperparameters
Click to expand
overwrite_output_dir: False
do_predict: False
eval_strategy: steps
prediction_loss_only: True
per_device_train_batch_size: 16
per_device_eval_batch_size: 16
per_gpu_train_batch_size: None
per_gpu_eval_batch_size: None
gradient_accumulation_steps: 1
eval_accumulation_steps: None
torch_empty_cache_steps: None
learning_rate: 2e-05
weight_decay: 0.0
adam_beta1: 0.9
adam_beta2: 0.999
adam_epsilon: 1e-08
max_grad_norm: 1.0
num_train_epochs: 1
max_steps: -1
lr_scheduler_type: linear
lr_scheduler_kwargs: {}
warmup_ratio: 0.1
warmup_steps: 0
log_level: passive
log_level_replica: warning
log_on_each_node: True
logging_nan_inf_filter: True
save_safetensors: True
save_on_each_node: False
save_only_model: False
restore_callback_states_from_checkpoint: False
no_cuda: False
use_cpu: False
use_mps_device: False
seed: 42
data_seed: None
jit_mode_eval: False
use_ipex: False
bf16: False
fp16: False
fp16_opt_level: O1
half_precision_backend: auto
bf16_full_eval: False
fp16_full_eval: False
tf32: None
local_rank: 0
ddp_backend: None
tpu_num_cores: None
tpu_metrics_debug: False
debug: []
dataloader_drop_last: False
dataloader_num_workers: 0
dataloader_prefetch_factor: None
past_index: -1
disable_tqdm: False
remove_unused_columns: True
label_names: None
load_best_model_at_end: False
ignore_data_skip: False
fsdp: []
fsdp_min_num_params: 0
fsdp_config: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
fsdp_transformer_layer_cls_to_wrap: None
accelerator_config: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
deepspeed: None
label_smoothing_factor: 0.0
optim: adamw_torch_fused
optim_args: None
adafactor: False
group_by_length: False
length_column_name: length
ddp_find_unused_parameters: None
ddp_bucket_cap_mb: None
ddp_broadcast_buffers: False
dataloader_pin_memory: True
dataloader_persistent_workers: False
skip_memory_metrics: True
use_legacy_prediction_loop: False
push_to_hub: False
resume_from_checkpoint: None
hub_model_id: None
hub_strategy: every_save
hub_private_repo: None
hub_always_push: False
hub_revision: None
gradient_checkpointing: False
gradient_checkpointing_kwargs: None
include_inputs_for_metrics: False
include_for_metrics: []
eval_do_concat_batches: True
fp16_backend: auto
push_to_hub_model_id: None
push_to_hub_organization: None
mp_parameters:
auto_find_batch_size: False
full_determinism: False
torchdynamo: None
ray_scope: last
ddp_timeout: 1800
torch_compile: False
torch_compile_backend: None
torch_compile_mode: None
include_tokens_per_second: False
include_num_input_tokens_seen: False
neftune_noise_alpha: None
optim_target_modules: None
batch_eval_metrics: False
eval_on_start: False
use_liger_kernel: False
liger_kernel_config: None
eval_use_gather_object: False
average_tokens_across_devices: False
prompts: None
batch_sampler: no_duplicates
multi_dataset_batch_sampler: proportional
router_mapping: {}
learning_rate_mapping: {}
Training Logs
| Epoch |
Step |
ai-job-validation_cosine_accuracy |
ai-job-test_cosine_accuracy |
| -1 |
-1 |
0.9894 |
1.0 |
Framework Versions
- Python: 3.12.11
- Sentence Transformers: 5.1.0
- Transformers: 4.55.2
- PyTorch: 2.8.0+cu126
- Accelerate: 1.10.0
- Datasets: 4.0.0
- Tokenizers: 0.21.4
Citation
BibTeX
Sentence Transformers
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "https://arxiv.org/abs/1908.10084",
}
MultipleNegativesRankingLoss
@misc{henderson2017efficient,
title={Efficient Natural Language Response Suggestion for Smart Reply},
author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
year={2017},
eprint={1705.00652},
archivePrefix={arXiv},
primaryClass={cs.CL}
}