SentenceTransformer based on intfloat/multilingual-e5-large
This is a sentence-transformers model finetuned from intfloat/multilingual-e5-large on the q2q_data and q2p_data datasets. It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
Model Details
Model Description
- Model Type: Sentence Transformer
- Base model: intfloat/multilingual-e5-large
- Maximum Sequence Length: 512 tokens
- Output Dimensionality: 1024 dimensions
- Similarity Function: Cosine Similarity
- Training Datasets:
Model Sources
Full Model Architecture
SentenceTransformer(
(0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: XLMRobertaModel
(1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
(2): Normalize()
)
Usage
Direct Usage (Sentence Transformers)
First install the Sentence Transformers library:
pip install -U sentence-transformers
Then you can load this model and run inference.
from sentence_transformers import SentenceTransformer
model = SentenceTransformer("George2002/sledopyt_embedder")
sentences = [
'query: Как должен выглядеть ярлык на Транспортной единице (ТЕ)?',
'query: Где подается заявление на выплату вкладов Чеченского банка?',
'query: Какие данные обязательно должны быть на ярлыке для Транспортной единицы?',
]
embeddings = model.encode(sentences)
print(embeddings.shape)
similarities = model.similarity(embeddings, embeddings)
print(similarities.shape)
Training Details
Training Datasets
q2q_data
q2p_data
Evaluation Datasets
q2q_data
q2p_data
Training Hyperparameters
Non-Default Hyperparameters
eval_strategy: steps
per_device_train_batch_size: 32
learning_rate: 1e-05
weight_decay: 0.01
warmup_ratio: 0.1
load_best_model_at_end: True
push_to_hub: True
hub_model_id: George2002/sledopyt_embedder
hub_strategy: end
All Hyperparameters
Click to expand
overwrite_output_dir: False
do_predict: False
eval_strategy: steps
prediction_loss_only: True
per_device_train_batch_size: 32
per_device_eval_batch_size: 8
per_gpu_train_batch_size: None
per_gpu_eval_batch_size: None
gradient_accumulation_steps: 1
eval_accumulation_steps: None
torch_empty_cache_steps: None
learning_rate: 1e-05
weight_decay: 0.01
adam_beta1: 0.9
adam_beta2: 0.999
adam_epsilon: 1e-08
max_grad_norm: 1.0
num_train_epochs: 3
max_steps: -1
lr_scheduler_type: linear
lr_scheduler_kwargs: {}
warmup_ratio: 0.1
warmup_steps: 0
log_level: passive
log_level_replica: warning
log_on_each_node: True
logging_nan_inf_filter: True
save_safetensors: True
save_on_each_node: False
save_only_model: False
restore_callback_states_from_checkpoint: False
no_cuda: False
use_cpu: False
use_mps_device: False
seed: 42
data_seed: None
jit_mode_eval: False
use_ipex: False
bf16: False
fp16: False
fp16_opt_level: O1
half_precision_backend: auto
bf16_full_eval: False
fp16_full_eval: False
tf32: None
local_rank: 0
ddp_backend: None
tpu_num_cores: None
tpu_metrics_debug: False
debug: []
dataloader_drop_last: True
dataloader_num_workers: 0
dataloader_prefetch_factor: None
past_index: -1
disable_tqdm: False
remove_unused_columns: True
label_names: None
load_best_model_at_end: True
ignore_data_skip: False
fsdp: []
fsdp_min_num_params: 0
fsdp_config: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
tp_size: 0
fsdp_transformer_layer_cls_to_wrap: None
accelerator_config: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
deepspeed: None
label_smoothing_factor: 0.0
optim: adamw_torch
optim_args: None
adafactor: False
group_by_length: False
length_column_name: length
ddp_find_unused_parameters: None
ddp_bucket_cap_mb: None
ddp_broadcast_buffers: False
dataloader_pin_memory: True
dataloader_persistent_workers: False
skip_memory_metrics: True
use_legacy_prediction_loop: False
push_to_hub: True
resume_from_checkpoint: None
hub_model_id: George2002/sledopyt_embedder
hub_strategy: end
hub_private_repo: None
hub_always_push: False
gradient_checkpointing: False
gradient_checkpointing_kwargs: None
include_inputs_for_metrics: False
include_for_metrics: []
eval_do_concat_batches: True
fp16_backend: auto
push_to_hub_model_id: None
push_to_hub_organization: None
mp_parameters:
auto_find_batch_size: False
full_determinism: False
torchdynamo: None
ray_scope: last
ddp_timeout: 1800
torch_compile: False
torch_compile_backend: None
torch_compile_mode: None
include_tokens_per_second: False
include_num_input_tokens_seen: False
neftune_noise_alpha: None
optim_target_modules: None
batch_eval_metrics: False
eval_on_start: False
use_liger_kernel: False
eval_use_gather_object: False
average_tokens_across_devices: False
prompts: None
batch_sampler: batch_sampler
multi_dataset_batch_sampler: proportional
Training Logs
Click to expand
| Epoch |
Step |
Training Loss |
q2q data loss |
q2p data loss |
| 0.0096 |
10 |
5.0879 |
- |
- |
| 0.0192 |
20 |
5.1046 |
- |
- |
| 0.0288 |
30 |
5.0837 |
- |
- |
| 0.0384 |
40 |
5.0778 |
- |
- |
| 0.0479 |
50 |
5.0643 |
5.0714 |
5.0469 |
| 0.0575 |
60 |
5.0424 |
- |
- |
| 0.0671 |
70 |
5.0255 |
- |
- |
| 0.0767 |
80 |
5.0099 |
- |
- |
| 0.0863 |
90 |
5.0063 |
- |
- |
| 0.0959 |
100 |
5.0033 |
5.0147 |
5.0005 |
| 0.1055 |
110 |
5.003 |
- |
- |
| 0.1151 |
120 |
4.9967 |
- |
- |
| 0.1246 |
130 |
4.998 |
- |
- |
| 0.1342 |
140 |
5.0012 |
- |
- |
| 0.1438 |
150 |
4.9989 |
5.0095 |
4.9424 |
| 0.1534 |
160 |
4.9908 |
- |
- |
| 0.1630 |
170 |
4.9735 |
- |
- |
| 0.1726 |
180 |
4.9965 |
- |
- |
| 0.1822 |
190 |
4.9825 |
- |
- |
| 0.1918 |
200 |
4.9514 |
5.0074 |
4.8744 |
| 0.2013 |
210 |
4.9521 |
- |
- |
| 0.2109 |
220 |
4.968 |
- |
- |
| 0.2205 |
230 |
4.96 |
- |
- |
| 0.2301 |
240 |
4.9758 |
- |
- |
| 0.2397 |
250 |
4.9834 |
5.0065 |
4.8428 |
| 0.2493 |
260 |
4.9273 |
- |
- |
| 0.2589 |
270 |
4.9796 |
- |
- |
| 0.2685 |
280 |
4.9517 |
- |
- |
| 0.2780 |
290 |
4.9763 |
- |
- |
| 0.2876 |
300 |
4.9372 |
5.0056 |
4.8253 |
| 0.2972 |
310 |
4.9325 |
- |
- |
| 0.3068 |
320 |
4.9477 |
- |
- |
| 0.3164 |
330 |
4.9455 |
- |
- |
| 0.3260 |
340 |
4.9258 |
- |
- |
| 0.3356 |
350 |
4.9799 |
5.0045 |
4.8434 |
| 0.3452 |
360 |
4.9791 |
- |
- |
| 0.3547 |
370 |
4.9437 |
- |
- |
| 0.3643 |
380 |
4.9873 |
- |
- |
| 0.3739 |
390 |
4.9425 |
- |
- |
| 0.3835 |
400 |
4.9837 |
5.0043 |
4.8419 |
| 0.3931 |
410 |
5.0006 |
- |
- |
| 0.4027 |
420 |
4.9831 |
- |
- |
| 0.4123 |
430 |
4.9531 |
- |
- |
| 0.4219 |
440 |
4.9856 |
- |
- |
| 0.4314 |
450 |
4.8996 |
5.0056 |
4.8652 |
| 0.4410 |
460 |
4.9467 |
- |
- |
| 0.4506 |
470 |
4.9724 |
- |
- |
| 0.4602 |
480 |
4.9797 |
- |
- |
| 0.4698 |
490 |
4.9735 |
- |
- |
| 0.4794 |
500 |
4.8765 |
5.0036 |
4.8457 |
| 0.4890 |
510 |
4.9136 |
- |
- |
| 0.4986 |
520 |
4.9688 |
- |
- |
| 0.5081 |
530 |
4.9436 |
- |
- |
| 0.5177 |
540 |
5.0017 |
- |
- |
| 0.5273 |
550 |
4.9867 |
5.0043 |
4.8609 |
| 0.5369 |
560 |
4.9716 |
- |
- |
| 0.5465 |
570 |
4.9338 |
- |
- |
| 0.5561 |
580 |
4.9975 |
- |
- |
| 0.5657 |
590 |
4.9485 |
- |
- |
| 0.5753 |
600 |
4.8959 |
5.0026 |
4.7809 |
| 0.5849 |
610 |
4.9769 |
- |
- |
| 0.5944 |
620 |
4.9407 |
- |
- |
| 0.6040 |
630 |
4.9941 |
- |
- |
| 0.6136 |
640 |
4.976 |
- |
- |
| 0.6232 |
650 |
4.986 |
5.0030 |
4.8102 |
| 0.6328 |
660 |
4.94 |
- |
- |
| 0.6424 |
670 |
4.9917 |
- |
- |
| 0.6520 |
680 |
4.9938 |
- |
- |
| 0.6616 |
690 |
4.9373 |
- |
- |
| 0.6711 |
700 |
5.0235 |
5.0321 |
4.8794 |
| 0.6807 |
710 |
4.939 |
- |
- |
| 0.6903 |
720 |
4.9682 |
- |
- |
| 0.6999 |
730 |
4.9813 |
- |
- |
| 0.7095 |
740 |
4.9442 |
- |
- |
| 0.7191 |
750 |
4.9354 |
5.0024 |
4.8053 |
| 0.7287 |
760 |
4.9105 |
- |
- |
| 0.7383 |
770 |
4.9271 |
- |
- |
| 0.7478 |
780 |
4.9476 |
- |
- |
| 0.7574 |
790 |
4.8887 |
- |
- |
| 0.7670 |
800 |
4.9576 |
5.0020 |
4.7397 |
| 0.7766 |
810 |
4.9577 |
- |
- |
| 0.7862 |
820 |
4.8974 |
- |
- |
| 0.7958 |
830 |
5.0033 |
- |
- |
| 0.8054 |
840 |
4.976 |
- |
- |
| 0.8150 |
850 |
4.9915 |
5.0422 |
4.9367 |
| 0.8245 |
860 |
4.9856 |
- |
- |
| 0.8341 |
870 |
4.9566 |
- |
- |
| 0.8437 |
880 |
4.8738 |
- |
- |
| 0.8533 |
890 |
4.9258 |
- |
- |
| 0.8629 |
900 |
4.9118 |
5.0033 |
4.7945 |
| 0.8725 |
910 |
4.9782 |
- |
- |
| 0.8821 |
920 |
4.8659 |
- |
- |
| 0.8917 |
930 |
4.9197 |
- |
- |
| 0.9012 |
940 |
4.9281 |
- |
- |
| 0.9108 |
950 |
4.9427 |
5.0021 |
4.7857 |
| 0.9204 |
960 |
4.8966 |
- |
- |
| 0.9300 |
970 |
4.9657 |
- |
- |
| 0.9396 |
980 |
4.9597 |
- |
- |
| 0.9492 |
990 |
4.9627 |
- |
- |
| 0.9588 |
1000 |
4.8603 |
5.0017 |
4.8015 |
| 0.9684 |
1010 |
4.9817 |
- |
- |
| 0.9779 |
1020 |
4.813 |
- |
- |
| 0.9875 |
1030 |
4.9688 |
- |
- |
| 0.9971 |
1040 |
4.9802 |
- |
- |
| 1.0067 |
1050 |
4.8651 |
5.0015 |
4.7961 |
| 1.0163 |
1060 |
4.9286 |
- |
- |
| 1.0259 |
1070 |
4.9124 |
- |
- |
| 1.0355 |
1080 |
4.9927 |
- |
- |
| 1.0451 |
1090 |
4.9769 |
- |
- |
| 1.0547 |
1100 |
5.0158 |
5.0046 |
5.1203 |
| 1.0642 |
1110 |
5.0234 |
- |
- |
| 1.0738 |
1120 |
4.9903 |
- |
- |
| 1.0834 |
1130 |
5.008 |
- |
- |
| 1.0930 |
1140 |
4.9987 |
- |
- |
| 1.1026 |
1150 |
5.0091 |
5.0155 |
5.1031 |
| 1.1122 |
1160 |
5.0562 |
- |
- |
| 1.1218 |
1170 |
5.0129 |
- |
- |
| 1.1314 |
1180 |
5.0033 |
- |
- |
| 1.1409 |
1190 |
4.9961 |
- |
- |
| 1.1505 |
1200 |
4.988 |
5.0005 |
5.0025 |
| 1.1601 |
1210 |
4.9687 |
- |
- |
| 1.1697 |
1220 |
4.9824 |
- |
- |
| 1.1793 |
1230 |
4.9955 |
- |
- |
| 1.1889 |
1240 |
4.9943 |
- |
- |
| 1.1985 |
1250 |
5.0552 |
5.0002 |
5.0005 |
| 1.2081 |
1260 |
5.0073 |
- |
- |
| 1.2176 |
1270 |
4.9928 |
- |
- |
| 1.2272 |
1280 |
5.0183 |
- |
- |
| 1.2368 |
1290 |
4.9932 |
- |
- |
| 1.2464 |
1300 |
4.9737 |
5.0003 |
5.0018 |
| 1.2560 |
1310 |
5.012 |
- |
- |
| 1.2656 |
1320 |
5.0138 |
- |
- |
| 1.2752 |
1330 |
5.0107 |
- |
- |
| 1.2848 |
1340 |
5.0226 |
- |
- |
| 1.2943 |
1350 |
4.9827 |
5.0009 |
5.0021 |
| 1.3039 |
1360 |
5.0089 |
- |
- |
| 1.3135 |
1370 |
4.9869 |
- |
- |
| 1.3231 |
1380 |
5.0178 |
- |
- |
| 1.3327 |
1390 |
5.0038 |
- |
- |
| 1.3423 |
1400 |
4.9761 |
5.0003 |
5.0008 |
| 1.3519 |
1410 |
5.0025 |
- |
- |
| 1.3615 |
1420 |
4.9973 |
- |
- |
| 1.3710 |
1430 |
4.9911 |
- |
- |
| 1.3806 |
1440 |
5.0088 |
- |
- |
| 1.3902 |
1450 |
4.986 |
5.0001 |
5.0005 |
| 1.3998 |
1460 |
5.0025 |
- |
- |
| 1.4094 |
1470 |
4.9907 |
- |
- |
| 1.4190 |
1480 |
5.0012 |
- |
- |
| 1.4286 |
1490 |
4.9928 |
- |
- |
| 1.4382 |
1500 |
4.9955 |
5.0002 |
5.0006 |
| 1.4477 |
1510 |
4.9952 |
- |
- |
| 1.4573 |
1520 |
4.9875 |
- |
- |
| 1.4669 |
1530 |
5.0027 |
- |
- |
| 1.4765 |
1540 |
4.963 |
- |
- |
| 1.4861 |
1550 |
4.9662 |
5.0016 |
5.0061 |
| 1.4957 |
1560 |
4.9284 |
- |
- |
| 1.5053 |
1570 |
4.9773 |
- |
- |
| 1.5149 |
1580 |
5.0154 |
- |
- |
| 1.5244 |
1590 |
4.9956 |
- |
- |
| 1.5340 |
1600 |
5.0147 |
5.0152 |
5.0622 |
| 1.5436 |
1610 |
5.0216 |
- |
- |
| 1.5532 |
1620 |
5.0276 |
- |
- |
| 1.5628 |
1630 |
5.0115 |
- |
- |
| 1.5724 |
1640 |
4.9836 |
- |
- |
| 1.5820 |
1650 |
5.0171 |
5.0001 |
5.0022 |
| 1.5916 |
1660 |
5.0266 |
- |
- |
| 1.6012 |
1670 |
4.9617 |
- |
- |
| 1.6107 |
1680 |
4.9691 |
- |
- |
| 1.6203 |
1690 |
5.0004 |
- |
- |
| 1.6299 |
1700 |
5.0173 |
5.0399 |
5.0408 |
| 1.6395 |
1710 |
4.9824 |
- |
- |
| 1.6491 |
1720 |
5.0099 |
- |
- |
| 1.6587 |
1730 |
5.003 |
- |
- |
| 1.6683 |
1740 |
5.0039 |
- |
- |
| 1.6779 |
1750 |
5.0031 |
5.0021 |
5.0106 |
| 1.6874 |
1760 |
4.9992 |
- |
- |
| 1.6970 |
1770 |
4.9997 |
- |
- |
| 1.7066 |
1780 |
4.9999 |
- |
- |
| 1.7162 |
1790 |
5.0022 |
- |
- |
| 1.7258 |
1800 |
4.9988 |
5.0017 |
5.0100 |
| 1.7354 |
1810 |
4.9987 |
- |
- |
| 1.7450 |
1820 |
5.0029 |
- |
- |
| 1.7546 |
1830 |
4.9949 |
- |
- |
| 1.7641 |
1840 |
4.999 |
- |
- |
| 1.7737 |
1850 |
5.0017 |
5.0033 |
4.9873 |
| 1.7833 |
1860 |
4.9866 |
- |
- |
| 1.7929 |
1870 |
4.9976 |
- |
- |
| 1.8025 |
1880 |
4.9784 |
- |
- |
| 1.8121 |
1890 |
4.9824 |
- |
- |
| 1.8217 |
1900 |
4.9945 |
5.0142 |
5.1557 |
| 1.8313 |
1910 |
4.9445 |
- |
- |
| 1.8408 |
1920 |
4.9477 |
- |
- |
| 1.8504 |
1930 |
5.0047 |
- |
- |
| 1.8600 |
1940 |
4.9307 |
- |
- |
| 1.8696 |
1950 |
4.9634 |
5.0023 |
4.8261 |
| 1.8792 |
1960 |
4.9791 |
- |
- |
| 1.8888 |
1970 |
4.9792 |
- |
- |
| 1.8984 |
1980 |
4.9041 |
- |
- |
| 1.9080 |
1990 |
4.9349 |
- |
- |
| 1.9175 |
2000 |
4.8942 |
5.0013 |
4.7501 |
| 1.9271 |
2010 |
4.9871 |
- |
- |
| 1.9367 |
2020 |
4.9631 |
- |
- |
| 1.9463 |
2030 |
4.9604 |
- |
- |
| 1.9559 |
2040 |
4.9346 |
- |
- |
| 1.9655 |
2050 |
4.9398 |
5.0012 |
4.7617 |
| 1.9751 |
2060 |
4.9262 |
- |
- |
| 1.9847 |
2070 |
4.9505 |
- |
- |
| 1.9942 |
2080 |
4.9667 |
- |
- |
| 2.0038 |
2090 |
4.8763 |
- |
- |
| 2.0134 |
2100 |
4.9116 |
5.0010 |
4.7720 |
| 2.0230 |
2110 |
4.9367 |
- |
- |
| 2.0326 |
2120 |
4.9546 |
- |
- |
| 2.0422 |
2130 |
4.9743 |
- |
- |
| 2.0518 |
2140 |
4.9795 |
- |
- |
| 2.0614 |
2150 |
4.9647 |
5.0010 |
4.7732 |
| 2.0709 |
2160 |
4.9856 |
- |
- |
| 2.0805 |
2170 |
4.9553 |
- |
- |
| 2.0901 |
2180 |
4.9479 |
- |
- |
| 2.0997 |
2190 |
4.9672 |
- |
- |
| 2.1093 |
2200 |
4.9645 |
5.0011 |
4.7576 |
| 2.1189 |
2210 |
4.9299 |
- |
- |
| 2.1285 |
2220 |
4.9777 |
- |
- |
| 2.1381 |
2230 |
4.94 |
- |
- |
| 2.1477 |
2240 |
4.978 |
- |
- |
| 2.1572 |
2250 |
4.968 |
4.9994 |
4.7557 |
| 2.1668 |
2260 |
4.9512 |
- |
- |
| 2.1764 |
2270 |
4.9539 |
- |
- |
| 2.1860 |
2280 |
4.9508 |
- |
- |
| 2.1956 |
2290 |
4.8871 |
- |
- |
| 2.2052 |
2300 |
4.909 |
5.0010 |
4.7444 |
| 2.2148 |
2310 |
4.9587 |
- |
- |
| 2.2244 |
2320 |
4.8956 |
- |
- |
| 2.2339 |
2330 |
4.9891 |
- |
- |
| 2.2435 |
2340 |
4.8795 |
- |
- |
| 2.2531 |
2350 |
4.887 |
5.0010 |
4.7448 |
| 2.2627 |
2360 |
4.9723 |
- |
- |
| 2.2723 |
2370 |
4.8967 |
- |
- |
| 2.2819 |
2380 |
4.8975 |
- |
- |
| 2.2915 |
2390 |
4.9177 |
- |
- |
| 2.3011 |
2400 |
4.9272 |
5.0010 |
4.7287 |
| 2.3106 |
2410 |
4.9283 |
- |
- |
| 2.3202 |
2420 |
4.9061 |
- |
- |
| 2.3298 |
2430 |
4.9279 |
- |
- |
| 2.3394 |
2440 |
4.9856 |
- |
- |
| 2.3490 |
2450 |
4.8988 |
5.0010 |
4.7268 |
| 2.3586 |
2460 |
4.9269 |
- |
- |
| 2.3682 |
2470 |
4.9318 |
- |
- |
| 2.3778 |
2480 |
4.8814 |
- |
- |
| 2.3873 |
2490 |
4.9912 |
- |
- |
| 2.3969 |
2500 |
4.9226 |
5.0009 |
4.7197 |
| 2.4065 |
2510 |
4.9437 |
- |
- |
| 2.4161 |
2520 |
4.9553 |
- |
- |
| 2.4257 |
2530 |
4.9355 |
- |
- |
| 2.4353 |
2540 |
4.9063 |
- |
- |
| 2.4449 |
2550 |
4.9095 |
4.9973 |
4.7481 |
| 2.4545 |
2560 |
4.9624 |
- |
- |
| 2.4640 |
2570 |
4.9731 |
- |
- |
| 2.4736 |
2580 |
4.9156 |
- |
- |
| 2.4832 |
2590 |
4.8714 |
- |
- |
| 2.4928 |
2600 |
4.9532 |
5.0009 |
4.7327 |
| 2.5024 |
2610 |
4.9196 |
- |
- |
| 2.5120 |
2620 |
4.9477 |
- |
- |
| 2.5216 |
2630 |
4.9725 |
- |
- |
| 2.5312 |
2640 |
4.9483 |
- |
- |
| 2.5407 |
2650 |
4.9124 |
5.0008 |
4.7752 |
| 2.5503 |
2660 |
4.9056 |
- |
- |
| 2.5599 |
2670 |
4.9396 |
- |
- |
| 2.5695 |
2680 |
4.9472 |
- |
- |
| 2.5791 |
2690 |
4.9322 |
- |
- |
| 2.5887 |
2700 |
4.9147 |
5.0008 |
4.7360 |
| 2.5983 |
2710 |
4.9511 |
- |
- |
| 2.6079 |
2720 |
4.9229 |
- |
- |
| 2.6174 |
2730 |
4.9207 |
- |
- |
| 2.6270 |
2740 |
4.9695 |
- |
- |
| 2.6366 |
2750 |
4.8886 |
5.0008 |
4.7616 |
| 2.6462 |
2760 |
4.9878 |
- |
- |
| 2.6558 |
2770 |
4.9647 |
- |
- |
| 2.6654 |
2780 |
4.9552 |
- |
- |
| 2.6750 |
2790 |
5.0171 |
- |
- |
| 2.6846 |
2800 |
4.9379 |
5.0008 |
4.7563 |
| 2.6942 |
2810 |
4.9727 |
- |
- |
| 2.7037 |
2820 |
4.9798 |
- |
- |
| 2.7133 |
2830 |
4.9726 |
- |
- |
| 2.7229 |
2840 |
4.956 |
- |
- |
| 2.7325 |
2850 |
4.9512 |
5.0007 |
4.7669 |
| 2.7421 |
2860 |
4.9705 |
- |
- |
| 2.7517 |
2870 |
4.8603 |
- |
- |
| 2.7613 |
2880 |
4.9764 |
- |
- |
| 2.7709 |
2890 |
4.9187 |
- |
- |
| 2.7804 |
2900 |
4.8941 |
5.0008 |
4.7311 |
| 2.7900 |
2910 |
4.9592 |
- |
- |
| 2.7996 |
2920 |
4.9141 |
- |
- |
| 2.8092 |
2930 |
4.9198 |
- |
- |
| 2.8188 |
2940 |
5.0112 |
- |
- |
| 2.8284 |
2950 |
4.9778 |
5.0007 |
4.7333 |
| 2.8380 |
2960 |
4.8999 |
- |
- |
| 2.8476 |
2970 |
4.9223 |
- |
- |
| 2.8571 |
2980 |
4.9369 |
- |
- |
| 2.8667 |
2990 |
4.8722 |
- |
- |
| 2.8763 |
3000 |
4.9299 |
5.0008 |
4.7280 |
| 2.8859 |
3010 |
4.8457 |
- |
- |
| 2.8955 |
3020 |
4.8864 |
- |
- |
| 2.9051 |
3030 |
4.882 |
- |
- |
| 2.9147 |
3040 |
4.8897 |
- |
- |
| 2.9243 |
3050 |
4.9663 |
5.0007 |
4.7238 |
| 2.9338 |
3060 |
4.946 |
- |
- |
| 2.9434 |
3070 |
4.9555 |
- |
- |
| 2.9530 |
3080 |
4.9005 |
- |
- |
| 2.9626 |
3090 |
4.9097 |
- |
- |
| 2.9722 |
3100 |
4.924 |
5.0007 |
4.7231 |
| 2.9818 |
3110 |
4.8929 |
- |
- |
| 2.9914 |
3120 |
4.93 |
- |
- |
Framework Versions
- Python: 3.10.12
- Sentence Transformers: 4.1.0
- Transformers: 4.51.3
- PyTorch: 2.6.0+cu124
- Accelerate: 1.6.0
- Datasets: 3.5.0
- Tokenizers: 0.21.1
Citation
BibTeX
Sentence Transformers
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "https://arxiv.org/abs/1908.10084",
}
TripletLoss
@misc{hermans2017defense,
title={In Defense of the Triplet Loss for Person Re-Identification},
author={Alexander Hermans and Lucas Beyer and Bastian Leibe},
year={2017},
eprint={1703.07737},
archivePrefix={arXiv},
primaryClass={cs.CV}
}