metadata
base_model: buddhist-nlp/buddhist-sentence-similarity
library_name: sentence-transformers
pipeline_tag: sentence-similarity
tags:
- sentence-transformers
- sentence-similarity
- feature-extraction
- generated_from_trainer
- dataset_size:3779
- loss:MultipleNegativesRankingLoss
widget:
- source_sentence: oM ma Ni pad+me hUM|
sentences:
- 'xamuq bayidal tögüsüqsen bayidal ügei:'
- >-
burxan zarliq bolboi: subudi öün-dü you sedkikü: γurban mingγani yeke
mingγan yertünčüyin orondu γazariyin tōsun kedüi bui töüni olon kemēn
sedkikü buyu:
- 'oṁ ma ni pad me huṁ:'
- source_sentence: >-
thogs pa med pa nyon cig| ngas skye bo kun dang yon bdag tu gyur pa rnams
la phan pa dang bde ba′i phyir| gnam sa snang brgyad ces bya ba′i chos ′di
brjod pas snang srid kyi sa bdag lha klu kun zhi ba dang bde legs su bcos
so|
sentences:
- >-
bi öüni sonosun meded: bi dēdü-yi ayilidxan üyiledbei: tere imaqtadu eši
üzüülüqseni tula: ögüülen durdaxui-gi bi üyiledbei::
- >-
biligiyin činadu kürüqsen Dorǰi Zodboyin ači tusayin tayilbur-ēce: toyin
ǰiluba nirvān bolxu-yi caqtu yeke γaixamšiqtai belgeši üzüülüqsen: bölöq
inü arban dabtaγar bui:: ::
- >-
türbel ügei čingna: bi xamuq törölkitön kigēd öqligöyin ezen
boluqsan-noγoud tusa bolun buyan aγui yeke bolxu-yin tula: yeke nom
oqtorγui γazariyin nayiman gegēn öüni ögüüleqsen-yēr: gegēn sangsar
γazariyin ezen xamuq doqšin kluši amurliulun sayin amuγuulang-tu bolγon
üyiledümüi::
- source_sentence: bdag gis sdig pa gang bgyis dang||
sentences:
- 'mini kilince ali üyiledüqsen::'
- >-
arban küčüni dēdü altan gerel-yēr: suduriyin ayimagiyin erketü xān maši
manglai öüni: yertünčü-yi tedküqči ta bügüde: maši batu kičēnggüi-bēr
tedkü::
- >-
tegēd bodhi-sadv-nariyin manglai xutuqtu nidübēr üzüqči erketü: amidābha
terigüüten arban nigen ǰeva burxadiyin gegēn-dü amitani tusa üyiledküi
sedkil öüskebei:
- source_sentence: >-
lha dang mi dang ngan song gsum dag tu|| ji ltar byas pa'i las kyis srid
par skye||
sentences:
- >-
izuurtani köböün tedeni öbörö törlökitün busu-bēr uxan üyiled xarin ülü
xariqči bodhi-sadv maha-sadv-nar-luγā adali üzün üyiledküi: tedeni
üküküi caqtu arban xoyor togünčilen boluqsad ȫdö bolǰi: izourtani köböün
yeke külgüni suduriyin xān öüni sonosoqsan xarin orčilong-du ülü törökü
bolyu: xarin törkü ötülkü kigēd ebedekü ükükü ülü üzekü bolyu: busun
tālaxoi-luγā ülü xaγacan: xorson ülü tālaxoi-yi-luγā učirxu ügei bui:
- >-
tenggeri kümün kigēd γurban mou zayātan-du: yamaru üyiledüqsen zayāni
sangsartu törökü
- 'subudi ayiladxabai: ilaγun tögüsüqsen tere beye yeke mün:'
- source_sentence: >-
ji ltar chos smra ba de'i lus mi ngal bar 'gyur ba dang| lus kyi dbang po
bde bar 'gyur ba dang| rab tu dga' ba skye bar 'gyur ba dang| gang gi slad
du sangs rgyas brgya stong dag la dge ba'i rtsa ba bskrun pa'i sems can
rnams kyi don gyi slad du| gser 'od dam pa mdo sde'i dbang po'i rgyal po
'di 'dzam bu'i gling du yun ring du gnas shing myur du nub par mi 'gyur ba
dang| sems can rnams kyang gser 'od dam pa mdo sde'i dbang po'i rgyal po
'di nyan par 'gyur ba dang| ye shes kyi phung po bsam gyis mi khyab pa
thob par 'gyur ba dang| shes rab dang ldan par 'gyur ba dang| bsod nams
kyi phung po rab tu 'dzin par 'gyur ba dang| ma 'ongs pa'i dus na bskal pa
bye ba khrag khrig brgya stong phrag du mar lha dang mi'i bde ba bsam gyis
mi khyab pa myong bar 'gyur ba dang| de bzhin gshegs pa dang 'grogs par
'gyur ba dang| ma 'ongs pa'i dus na bla na med pa yang dag par rdzogs pa'i
byang chub mngon par rdzogs par 'tshang rgya bar 'gyur ba dang| sems can
dmyal ba dang| dud 'gro'i skye gnas dang| gshin rje'i 'jig rten gyi sdug
bsngal thams cad shin tu rgyun chad par 'gyur bar de'i spu'i bu ga rnams
su mdangs stsal bar bgyi'o||
sentences:
- 'dusul xōsun zoun üzüq:'
- >-
ödügē yeke rahā-noγoudi zasamui: naran saran kigēd ulān nidütü: ülemǰi
kigēd γadasun bui: basang saben-ba kigēd bariqči Rahula: urtu söültü
utān kigēd: yeke raha tedeni čü öbör kigēd: öqligöyin ezen-lügē
zokilduulun zasamui: amurlin sayin amuγuulang boltuγai::
- >-
yamaru nom ӧgüüleqči dgeslong töüni beye ülü alzoulun üyiledün: beyeyin
erketü-yi amuγuulang bolγon: sayitur bayasxan üyiledkü kigēd: keni
tulada zoun mingγan burxan-noγoudtu buyani ündüsü öüskeqsen
amitan-noγoudiyin tusayin tulada: suduriyin aimagiyin erketü xān dēdü
altan gerel öüni: ‘zambutib-tu önidö orošiulun ötör ülü ecüdken üyiledkü
kigēd: amitan-noγoudčü suduriyin ayimagiyin erketü xān dēdü altan gerel
öüni sonosun üyiledkü kigēd: belge biligiyin coqco sedkiši ügei olun
üyiledkü: biliq-lügē tögüsün üyiledkü kigēd: buyani coqco sayitur barin
üyiledkü: irē ödüi caqtu olon zoun mingγan kraq kriq ǯeva γalab-tu:
kümün tenggeriyin amuγuulang sedkiši ügei edlen üiledkü kigēd:
tögünčilen boluqsan-luγā nӧkücün üyiledkü: irē ödüi caqtu dēre ügei
sayitur dousuqsan bodhi-du ilerkei dousun burxan bolxu kigēd tamuyin
amitan kigēd adousuni tӧrӧkui oron erligiyin yertünčüyin zobolong xamugi
tasulun üyiledküye: tӧüni šara üsün-noγoud- tu önggü ogün üyiledümüi:
SentenceTransformer based on buddhist-nlp/buddhist-sentence-similarity
This is a sentence-transformers model finetuned from buddhist-nlp/buddhist-sentence-similarity. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
Model Details
Model Description
- Model Type: Sentence Transformer
- Base model: buddhist-nlp/buddhist-sentence-similarity
- Maximum Sequence Length: 512 tokens
- Output Dimensionality: 768 tokens
- Similarity Function: Cosine Similarity
Model Sources
- Documentation: Sentence Transformers Documentation
- Repository: Sentence Transformers on GitHub
- Hugging Face: Sentence Transformers on Hugging Face
Full Model Architecture
SentenceTransformer(
(0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
)
Usage
Direct Usage (Sentence Transformers)
First install the Sentence Transformers library:
pip install -U sentence-transformers
Then you can load this model and run inference.
from sentence_transformers import SentenceTransformer
# Download from the 🤗 Hub
model = SentenceTransformer("sentence_transformers_model_id")
# Run inference
sentences = [
"ji ltar chos smra ba de'i lus mi ngal bar 'gyur ba dang| lus kyi dbang po bde bar 'gyur ba dang| rab tu dga' ba skye bar 'gyur ba dang| gang gi slad du sangs rgyas brgya stong dag la dge ba'i rtsa ba bskrun pa'i sems can rnams kyi don gyi slad du| gser 'od dam pa mdo sde'i dbang po'i rgyal po 'di 'dzam bu'i gling du yun ring du gnas shing myur du nub par mi 'gyur ba dang| sems can rnams kyang gser 'od dam pa mdo sde'i dbang po'i rgyal po 'di nyan par 'gyur ba dang| ye shes kyi phung po bsam gyis mi khyab pa thob par 'gyur ba dang| shes rab dang ldan par 'gyur ba dang| bsod nams kyi phung po rab tu 'dzin par 'gyur ba dang| ma 'ongs pa'i dus na bskal pa bye ba khrag khrig brgya stong phrag du mar lha dang mi'i bde ba bsam gyis mi khyab pa myong bar 'gyur ba dang| de bzhin gshegs pa dang 'grogs par 'gyur ba dang| ma 'ongs pa'i dus na bla na med pa yang dag par rdzogs pa'i byang chub mngon par rdzogs par 'tshang rgya bar 'gyur ba dang| sems can dmyal ba dang| dud 'gro'i skye gnas dang| gshin rje'i 'jig rten gyi sdug bsngal thams cad shin tu rgyun chad par 'gyur bar de'i spu'i bu ga rnams su mdangs stsal bar bgyi'o||",
'yamaru nom ӧgüüleqči dgeslong töüni beye ülü alzoulun üyiledün: beyeyin erketü-yi amuγuulang bolγon: sayitur bayasxan üyiledkü kigēd: keni tulada zoun mingγan burxan-noγoudtu buyani ündüsü öüskeqsen amitan-noγoudiyin tusayin tulada: suduriyin aimagiyin erketü xān dēdü altan gerel öüni: ‘zambutib-tu önidö orošiulun ötör ülü ecüdken üyiledkü kigēd: amitan-noγoudčü suduriyin ayimagiyin erketü xān dēdü altan gerel öüni sonosun üyiledkü kigēd: belge biligiyin coqco sedkiši ügei olun üyiledkü: biliq-lügē tögüsün üyiledkü kigēd: buyani coqco sayitur barin üyiledkü: irē ödüi caqtu olon zoun mingγan kraq kriq ǯeva γalab-tu: kümün tenggeriyin amuγuulang sedkiši ügei edlen üiledkü kigēd: tögünčilen boluqsan-luγā nӧkücün üyiledkü: irē ödüi caqtu dēre ügei sayitur dousuqsan bodhi-du ilerkei dousun burxan bolxu kigēd tamuyin amitan kigēd adousuni tӧrӧkui oron erligiyin yertünčüyin zobolong xamugi tasulun üyiledküye: tӧüni šara üsün-noγoud- tu önggü ogün üyiledümüi:',
'dusul xōsun zoun üzüq:',
]
embeddings = model.encode(sentences)
print(embeddings.shape)
# [3, 768]
# Get the similarity scores for the embeddings
similarities = model.similarity(embeddings, embeddings)
print(similarities.shape)
# [3, 3]
Training Details
Training Dataset
Unnamed Dataset
- Size: 3,779 training samples
- Columns:
sentence_0andsentence_1 - Approximate statistics based on the first 1000 samples:
sentence_0 sentence_1 type string string details - min: 3 tokens
- mean: 56.43 tokens
- max: 426 tokens
- min: 7 tokens
- mean: 57.5 tokens
- max: 397 tokens
- Samples:
sentence_0 sentence_1 bdag ni khyod kyi srog bcod du 'ongs pas gsung rab mdo sde’i sgra thos pa tsam gyi bdag gi mthu stobs kyang rab tu nyamszhe sdang rnams kyang rab tu zhi zhing nyams las gang gi phyir ni 'byung ba mi 'byung basgyu ma smig rgyu lta bu chags med pa - Loss:
MultipleNegativesRankingLosswith these parameters:{ "scale": 20.0, "similarity_fct": "cos_sim" }
Training Hyperparameters
Non-Default Hyperparameters
eval_strategy: stepsper_device_train_batch_size: 16per_device_eval_batch_size: 16num_train_epochs: 80fp16: Truemulti_dataset_batch_sampler: round_robin
All Hyperparameters
Click to expand
overwrite_output_dir: Falsedo_predict: Falseeval_strategy: stepsprediction_loss_only: Trueper_device_train_batch_size: 16per_device_eval_batch_size: 16per_gpu_train_batch_size: Noneper_gpu_eval_batch_size: Nonegradient_accumulation_steps: 1eval_accumulation_steps: Nonetorch_empty_cache_steps: Nonelearning_rate: 5e-05weight_decay: 0.0adam_beta1: 0.9adam_beta2: 0.999adam_epsilon: 1e-08max_grad_norm: 1num_train_epochs: 80max_steps: -1lr_scheduler_type: linearlr_scheduler_kwargs: {}warmup_ratio: 0.0warmup_steps: 0log_level: passivelog_level_replica: warninglog_on_each_node: Truelogging_nan_inf_filter: Truesave_safetensors: Truesave_on_each_node: Falsesave_only_model: Falserestore_callback_states_from_checkpoint: Falseno_cuda: Falseuse_cpu: Falseuse_mps_device: Falseseed: 42data_seed: Nonejit_mode_eval: Falseuse_ipex: Falsebf16: Falsefp16: Truefp16_opt_level: O1half_precision_backend: autobf16_full_eval: Falsefp16_full_eval: Falsetf32: Nonelocal_rank: 0ddp_backend: Nonetpu_num_cores: Nonetpu_metrics_debug: Falsedebug: []dataloader_drop_last: Falsedataloader_num_workers: 0dataloader_prefetch_factor: Nonepast_index: -1disable_tqdm: Falseremove_unused_columns: Truelabel_names: Noneload_best_model_at_end: Falseignore_data_skip: Falsefsdp: []fsdp_min_num_params: 0fsdp_config: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}fsdp_transformer_layer_cls_to_wrap: Noneaccelerator_config: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}deepspeed: Nonelabel_smoothing_factor: 0.0optim: adamw_torchoptim_args: Noneadafactor: Falsegroup_by_length: Falselength_column_name: lengthddp_find_unused_parameters: Noneddp_bucket_cap_mb: Noneddp_broadcast_buffers: Falsedataloader_pin_memory: Truedataloader_persistent_workers: Falseskip_memory_metrics: Trueuse_legacy_prediction_loop: Falsepush_to_hub: Falseresume_from_checkpoint: Nonehub_model_id: Nonehub_strategy: every_savehub_private_repo: Falsehub_always_push: Falsegradient_checkpointing: Falsegradient_checkpointing_kwargs: Noneinclude_inputs_for_metrics: Falseeval_do_concat_batches: Truefp16_backend: autopush_to_hub_model_id: Nonepush_to_hub_organization: Nonemp_parameters:auto_find_batch_size: Falsefull_determinism: Falsetorchdynamo: Noneray_scope: lastddp_timeout: 1800torch_compile: Falsetorch_compile_backend: Nonetorch_compile_mode: Nonedispatch_batches: Nonesplit_batches: Noneinclude_tokens_per_second: Falseinclude_num_input_tokens_seen: Falseneftune_noise_alpha: Noneoptim_target_modules: Nonebatch_eval_metrics: Falseeval_on_start: Falseeval_use_gather_object: Falsebatch_sampler: batch_samplermulti_dataset_batch_sampler: round_robin
Training Logs
Click to expand
| Epoch | Step | Training Loss |
|---|---|---|
| 1.0 | 237 | - |
| 1.2658 | 300 | - |
| 2.0 | 474 | - |
| 2.1097 | 500 | 1.3285 |
| 2.5316 | 600 | - |
| 3.0 | 711 | - |
| 3.7975 | 900 | - |
| 4.0 | 948 | - |
| 4.2194 | 1000 | 0.4782 |
| 5.0 | 1185 | - |
| 5.0633 | 1200 | - |
| 6.0 | 1422 | - |
| 6.3291 | 1500 | 0.2195 |
| 7.0 | 1659 | - |
| 7.5949 | 1800 | - |
| 8.0 | 1896 | - |
| 8.4388 | 2000 | 0.1024 |
| 8.8608 | 2100 | - |
| 9.0 | 2133 | - |
| 10.0 | 2370 | - |
| 10.1266 | 2400 | - |
| 10.5485 | 2500 | 0.054 |
| 11.0 | 2607 | - |
| 11.3924 | 2700 | - |
| 12.0 | 2844 | - |
| 12.6582 | 3000 | 0.0277 |
| 13.0 | 3081 | - |
| 13.9241 | 3300 | - |
| 14.0 | 3318 | - |
| 14.7679 | 3500 | 0.0205 |
| 15.0 | 3555 | - |
| 15.1899 | 3600 | - |
| 16.0 | 3792 | - |
| 16.4557 | 3900 | - |
| 16.8776 | 4000 | 0.0173 |
| 17.0 | 4029 | - |
| 17.7215 | 4200 | - |
| 18.0 | 4266 | - |
| 18.9873 | 4500 | 0.0177 |
| 19.0 | 4503 | - |
| 20.0 | 4740 | - |
| 20.2532 | 4800 | - |
| 21.0 | 4977 | - |
| 21.0970 | 5000 | 0.0114 |
| 21.5190 | 5100 | - |
| 22.0 | 5214 | - |
| 22.7848 | 5400 | - |
| 23.0 | 5451 | - |
| 23.2068 | 5500 | 0.0115 |
| 24.0 | 5688 | - |
| 24.0506 | 5700 | - |
| 25.0 | 5925 | - |
| 25.3165 | 6000 | 0.0095 |
| 26.0 | 6162 | - |
| 26.5823 | 6300 | - |
| 27.0 | 6399 | - |
| 27.4262 | 6500 | 0.0123 |
| 27.8481 | 6600 | - |
| 28.0 | 6636 | - |
| 29.0 | 6873 | - |
| 29.1139 | 6900 | - |
| 29.5359 | 7000 | 0.0087 |
| 30.0 | 7110 | - |
| 30.3797 | 7200 | - |
| 31.0 | 7347 | - |
| 31.6456 | 7500 | 0.0074 |
| 32.0 | 7584 | - |
| 32.9114 | 7800 | - |
| 33.0 | 7821 | - |
| 33.7553 | 8000 | 0.0108 |
| 34.0 | 8058 | - |
| 34.1772 | 8100 | - |
| 35.0 | 8295 | - |
| 35.4430 | 8400 | - |
| 35.8650 | 8500 | 0.0074 |
| 36.0 | 8532 | - |
| 36.7089 | 8700 | - |
| 37.0 | 8769 | - |
| 37.9747 | 9000 | 0.0068 |
| 38.0 | 9006 | - |
| 39.0 | 9243 | - |
| 39.2405 | 9300 | - |
| 40.0 | 9480 | - |
| 40.0844 | 9500 | 0.0053 |
| 40.5063 | 9600 | - |
| 41.0 | 9717 | - |
| 41.7722 | 9900 | - |
| 42.0 | 9954 | - |
| 42.1941 | 10000 | 0.0066 |
| 43.0 | 10191 | - |
| 43.0380 | 10200 | - |
| 44.0 | 10428 | - |
| 44.3038 | 10500 | 0.0073 |
| 45.0 | 10665 | - |
| 45.5696 | 10800 | - |
| 46.0 | 10902 | - |
| 46.4135 | 11000 | 0.0067 |
| 46.8354 | 11100 | - |
| 47.0 | 11139 | - |
| 48.0 | 11376 | - |
| 48.1013 | 11400 | - |
| 48.5232 | 11500 | 0.0061 |
| 49.0 | 11613 | - |
| 49.3671 | 11700 | - |
| 50.0 | 11850 | - |
| 50.6329 | 12000 | 0.0062 |
| 51.0 | 12087 | - |
| 51.8987 | 12300 | - |
| 52.0 | 12324 | - |
| 52.7426 | 12500 | 0.0051 |
| 53.0 | 12561 | - |
| 53.1646 | 12600 | - |
| 54.0 | 12798 | - |
| 54.4304 | 12900 | - |
| 54.8523 | 13000 | 0.0052 |
| 55.0 | 13035 | - |
| 55.6962 | 13200 | - |
| 56.0 | 13272 | - |
| 56.9620 | 13500 | 0.004 |
| 57.0 | 13509 | - |
| 58.0 | 13746 | - |
| 58.2278 | 13800 | - |
| 59.0 | 13983 | - |
Framework Versions
- Python: 3.12.1
- Sentence Transformers: 3.1.1
- Transformers: 4.44.2
- PyTorch: 2.6.0.dev20240923+cu121
- Accelerate: 1.12.0
- Datasets: 4.4.2
- Tokenizers: 0.19.1
Citation
BibTeX
Sentence Transformers
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "https://arxiv.org/abs/1908.10084",
}
MultipleNegativesRankingLoss
@misc{henderson2017efficient,
title={Efficient Natural Language Response Suggestion for Smart Reply},
author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
year={2017},
eprint={1705.00652},
archivePrefix={arXiv},
primaryClass={cs.CL}
}