avsolatorio commited on
Commit
abf5e76
·
1 Parent(s): a095fd0

Training for 0.0 epochs, 500 steps, 5.3311 loss, 2.7874564459930317e-08 learning rate.

Browse files
Files changed (49) hide show
  1. 1_Pooling/config.json +7 -0
  2. 3_MixtureEmbeddingsModel/MixSentenceTransformer_config.json +15 -0
  3. 3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/1_Pooling/config.json +7 -0
  4. 3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/README.md +94 -0
  5. 3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/config.json +32 -0
  6. 3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/config_sentence_transformers.json +7 -0
  7. 3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/model.safetensors +3 -0
  8. 3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/modules.json +14 -0
  9. 3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/sentence_bert_config.json +4 -0
  10. 3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/special_tokens_map.json +37 -0
  11. 3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/tokenizer.json +0 -0
  12. 3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/tokenizer_config.json +57 -0
  13. 3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/vocab.txt +0 -0
  14. 3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/1_Pooling/config.json +7 -0
  15. 3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/README.md +2702 -0
  16. 3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/config.json +26 -0
  17. 3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/config_sentence_transformers.json +7 -0
  18. 3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/model.safetensors +3 -0
  19. 3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/modules.json +20 -0
  20. 3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/sentence_bert_config.json +4 -0
  21. 3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/special_tokens_map.json +37 -0
  22. 3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/tokenizer.json +0 -0
  23. 3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/tokenizer_config.json +62 -0
  24. 3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/vocab.txt +0 -0
  25. 3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/1_Pooling/config.json +7 -0
  26. 3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/2_Dense/config.json +1 -0
  27. 3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/2_Dense/pytorch_model.bin +3 -0
  28. 3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/README.md +51 -0
  29. 3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/config.json +61 -0
  30. 3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/config_sentence_transformers.json +7 -0
  31. 3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/model.safetensors +3 -0
  32. 3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/modules.json +26 -0
  33. 3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/sentence_bert_config.json +4 -0
  34. 3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/special_tokens_map.json +125 -0
  35. 3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/spiece.model +3 -0
  36. 3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/tokenizer.json +0 -0
  37. 3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/tokenizer_config.json +941 -0
  38. 3_MixtureEmbeddingsModel/gate.bin +3 -0
  39. README.md +176 -0
  40. config.json +26 -0
  41. config_sentence_transformers.json +7 -0
  42. freeze_encoder.txt +1 -0
  43. model.safetensors +3 -0
  44. modules.json +26 -0
  45. sentence_bert_config.json +4 -0
  46. special_tokens_map.json +37 -0
  47. tokenizer.json +0 -0
  48. tokenizer_config.json +64 -0
  49. vocab.txt +0 -0
1_Pooling/config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 384,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false
7
+ }
3_MixtureEmbeddingsModel/MixSentenceTransformer_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "expert_model_names": [
3
+ "infgrad/stella-base-en-v2",
4
+ "thenlper/gte-base",
5
+ "sentence-transformers/gtr-t5-base"
6
+ ],
7
+ "encoder_dim": 384,
8
+ "topk": 2,
9
+ "freeze_experts": false,
10
+ "normalize_experts": false,
11
+ "has_blender": false,
12
+ "has_noise": false,
13
+ "use_encoder_expert": false,
14
+ "use_gate_norm_last": false
15
+ }
3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/1_Pooling/config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false
7
+ }
3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/README.md ADDED
@@ -0,0 +1,94 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: sentence-similarity
3
+ tags:
4
+ - sentence-transformers
5
+ - feature-extraction
6
+ - sentence-similarity
7
+ - transformers
8
+
9
+ ---
10
+
11
+ # {MODEL_NAME}
12
+
13
+ This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
14
+
15
+ <!--- Describe your model here -->
16
+
17
+ ## Usage (Sentence-Transformers)
18
+
19
+ Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
20
+
21
+ ```
22
+ pip install -U sentence-transformers
23
+ ```
24
+
25
+ Then you can use the model like this:
26
+
27
+ ```python
28
+ from sentence_transformers import SentenceTransformer
29
+ sentences = ["This is an example sentence", "Each sentence is converted"]
30
+
31
+ model = SentenceTransformer('{MODEL_NAME}')
32
+ embeddings = model.encode(sentences)
33
+ print(embeddings)
34
+ ```
35
+
36
+
37
+
38
+ ## Usage (HuggingFace Transformers)
39
+ Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
40
+
41
+ ```python
42
+ from transformers import AutoTokenizer, AutoModel
43
+ import torch
44
+
45
+
46
+ #Mean Pooling - Take attention mask into account for correct averaging
47
+ def mean_pooling(model_output, attention_mask):
48
+ token_embeddings = model_output[0] #First element of model_output contains all token embeddings
49
+ input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
50
+ return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9)
51
+
52
+
53
+ # Sentences we want sentence embeddings for
54
+ sentences = ['This is an example sentence', 'Each sentence is converted']
55
+
56
+ # Load model from HuggingFace Hub
57
+ tokenizer = AutoTokenizer.from_pretrained('{MODEL_NAME}')
58
+ model = AutoModel.from_pretrained('{MODEL_NAME}')
59
+
60
+ # Tokenize sentences
61
+ encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')
62
+
63
+ # Compute token embeddings
64
+ with torch.no_grad():
65
+ model_output = model(**encoded_input)
66
+
67
+ # Perform pooling. In this case, mean pooling.
68
+ sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask'])
69
+
70
+ print("Sentence embeddings:")
71
+ print(sentence_embeddings)
72
+ ```
73
+
74
+
75
+
76
+ ## Evaluation Results
77
+
78
+ <!--- Describe how your model was evaluated -->
79
+
80
+ For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
81
+
82
+
83
+
84
+ ## Full Model Architecture
85
+ ```
86
+ EncoderSentenceTransformer(
87
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel
88
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
89
+ )
90
+ ```
91
+
92
+ ## Citing & Authors
93
+
94
+ <!--- Describe where people can find more information -->
3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/root/.cache/torch/sentence_transformers/infgrad_stella-base-en-v2",
3
+ "architectures": [
4
+ "BertModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "LABEL_0"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 3072,
17
+ "label2id": {
18
+ "LABEL_0": 0
19
+ },
20
+ "layer_norm_eps": 1e-12,
21
+ "max_position_embeddings": 512,
22
+ "model_type": "bert",
23
+ "num_attention_heads": 12,
24
+ "num_hidden_layers": 12,
25
+ "pad_token_id": 0,
26
+ "position_embedding_type": "absolute",
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.36.2",
29
+ "type_vocab_size": 2,
30
+ "use_cache": true,
31
+ "vocab_size": 30522
32
+ }
3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.2.2",
4
+ "transformers": "4.36.2",
5
+ "pytorch": "2.1.2+cu121"
6
+ }
7
+ }
3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:caaac1823f5ad9ab016e335d7e5b7aaa2c8ffbf169581465f141a188155f848b
3
+ size 437951328
3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/tokenizer_config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 512,
50
+ "never_split": null,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "strip_accents": null,
54
+ "tokenize_chinese_chars": true,
55
+ "tokenizer_class": "BertTokenizer",
56
+ "unk_token": "[UNK]"
57
+ }
3_MixtureEmbeddingsModel/expert_00_infgrad_stella-base-en-v2/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/1_Pooling/config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false
7
+ }
3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/README.md ADDED
@@ -0,0 +1,2702 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - mteb
4
+ - sentence-similarity
5
+ - sentence-transformers
6
+ - Sentence Transformers
7
+ model-index:
8
+ - name: gte-base
9
+ results:
10
+ - task:
11
+ type: Classification
12
+ dataset:
13
+ type: mteb/amazon_counterfactual
14
+ name: MTEB AmazonCounterfactualClassification (en)
15
+ config: en
16
+ split: test
17
+ revision: e8379541af4e31359cca9fbcf4b00f2671dba205
18
+ metrics:
19
+ - type: accuracy
20
+ value: 74.17910447761193
21
+ - type: ap
22
+ value: 36.827146398068926
23
+ - type: f1
24
+ value: 68.11292888046363
25
+ - task:
26
+ type: Classification
27
+ dataset:
28
+ type: mteb/amazon_polarity
29
+ name: MTEB AmazonPolarityClassification
30
+ config: default
31
+ split: test
32
+ revision: e2d317d38cd51312af73b3d32a06d1a08b442046
33
+ metrics:
34
+ - type: accuracy
35
+ value: 91.77345000000001
36
+ - type: ap
37
+ value: 88.33530426691347
38
+ - type: f1
39
+ value: 91.76549906404642
40
+ - task:
41
+ type: Classification
42
+ dataset:
43
+ type: mteb/amazon_reviews_multi
44
+ name: MTEB AmazonReviewsClassification (en)
45
+ config: en
46
+ split: test
47
+ revision: 1399c76144fd37290681b995c656ef9b2e06e26d
48
+ metrics:
49
+ - type: accuracy
50
+ value: 48.964
51
+ - type: f1
52
+ value: 48.22995586184998
53
+ - task:
54
+ type: Retrieval
55
+ dataset:
56
+ type: arguana
57
+ name: MTEB ArguAna
58
+ config: default
59
+ split: test
60
+ revision: None
61
+ metrics:
62
+ - type: map_at_1
63
+ value: 32.147999999999996
64
+ - type: map_at_10
65
+ value: 48.253
66
+ - type: map_at_100
67
+ value: 49.038
68
+ - type: map_at_1000
69
+ value: 49.042
70
+ - type: map_at_3
71
+ value: 43.433
72
+ - type: map_at_5
73
+ value: 46.182
74
+ - type: mrr_at_1
75
+ value: 32.717
76
+ - type: mrr_at_10
77
+ value: 48.467
78
+ - type: mrr_at_100
79
+ value: 49.252
80
+ - type: mrr_at_1000
81
+ value: 49.254999999999995
82
+ - type: mrr_at_3
83
+ value: 43.599
84
+ - type: mrr_at_5
85
+ value: 46.408
86
+ - type: ndcg_at_1
87
+ value: 32.147999999999996
88
+ - type: ndcg_at_10
89
+ value: 57.12199999999999
90
+ - type: ndcg_at_100
91
+ value: 60.316
92
+ - type: ndcg_at_1000
93
+ value: 60.402
94
+ - type: ndcg_at_3
95
+ value: 47.178
96
+ - type: ndcg_at_5
97
+ value: 52.146
98
+ - type: precision_at_1
99
+ value: 32.147999999999996
100
+ - type: precision_at_10
101
+ value: 8.542
102
+ - type: precision_at_100
103
+ value: 0.9900000000000001
104
+ - type: precision_at_1000
105
+ value: 0.1
106
+ - type: precision_at_3
107
+ value: 19.346
108
+ - type: precision_at_5
109
+ value: 14.026
110
+ - type: recall_at_1
111
+ value: 32.147999999999996
112
+ - type: recall_at_10
113
+ value: 85.42
114
+ - type: recall_at_100
115
+ value: 99.004
116
+ - type: recall_at_1000
117
+ value: 99.644
118
+ - type: recall_at_3
119
+ value: 58.037000000000006
120
+ - type: recall_at_5
121
+ value: 70.128
122
+ - task:
123
+ type: Clustering
124
+ dataset:
125
+ type: mteb/arxiv-clustering-p2p
126
+ name: MTEB ArxivClusteringP2P
127
+ config: default
128
+ split: test
129
+ revision: a122ad7f3f0291bf49cc6f4d32aa80929df69d5d
130
+ metrics:
131
+ - type: v_measure
132
+ value: 48.59706013699614
133
+ - task:
134
+ type: Clustering
135
+ dataset:
136
+ type: mteb/arxiv-clustering-s2s
137
+ name: MTEB ArxivClusteringS2S
138
+ config: default
139
+ split: test
140
+ revision: f910caf1a6075f7329cdf8c1a6135696f37dbd53
141
+ metrics:
142
+ - type: v_measure
143
+ value: 43.01463593002057
144
+ - task:
145
+ type: Reranking
146
+ dataset:
147
+ type: mteb/askubuntudupquestions-reranking
148
+ name: MTEB AskUbuntuDupQuestions
149
+ config: default
150
+ split: test
151
+ revision: 2000358ca161889fa9c082cb41daa8dcfb161a54
152
+ metrics:
153
+ - type: map
154
+ value: 61.80250355752458
155
+ - type: mrr
156
+ value: 74.79455216989844
157
+ - task:
158
+ type: STS
159
+ dataset:
160
+ type: mteb/biosses-sts
161
+ name: MTEB BIOSSES
162
+ config: default
163
+ split: test
164
+ revision: d3fb88f8f02e40887cd149695127462bbcf29b4a
165
+ metrics:
166
+ - type: cos_sim_pearson
167
+ value: 89.87448576082345
168
+ - type: cos_sim_spearman
169
+ value: 87.64235843637468
170
+ - type: euclidean_pearson
171
+ value: 88.4901825511062
172
+ - type: euclidean_spearman
173
+ value: 87.74537283182033
174
+ - type: manhattan_pearson
175
+ value: 88.39040638362911
176
+ - type: manhattan_spearman
177
+ value: 87.62669542888003
178
+ - task:
179
+ type: Classification
180
+ dataset:
181
+ type: mteb/banking77
182
+ name: MTEB Banking77Classification
183
+ config: default
184
+ split: test
185
+ revision: 0fd18e25b25c072e09e0d92ab615fda904d66300
186
+ metrics:
187
+ - type: accuracy
188
+ value: 85.06818181818183
189
+ - type: f1
190
+ value: 85.02524460098233
191
+ - task:
192
+ type: Clustering
193
+ dataset:
194
+ type: mteb/biorxiv-clustering-p2p
195
+ name: MTEB BiorxivClusteringP2P
196
+ config: default
197
+ split: test
198
+ revision: 65b79d1d13f80053f67aca9498d9402c2d9f1f40
199
+ metrics:
200
+ - type: v_measure
201
+ value: 38.20471092679967
202
+ - task:
203
+ type: Clustering
204
+ dataset:
205
+ type: mteb/biorxiv-clustering-s2s
206
+ name: MTEB BiorxivClusteringS2S
207
+ config: default
208
+ split: test
209
+ revision: 258694dd0231531bc1fd9de6ceb52a0853c6d908
210
+ metrics:
211
+ - type: v_measure
212
+ value: 36.58967592147641
213
+ - task:
214
+ type: Retrieval
215
+ dataset:
216
+ type: BeIR/cqadupstack
217
+ name: MTEB CQADupstackAndroidRetrieval
218
+ config: default
219
+ split: test
220
+ revision: None
221
+ metrics:
222
+ - type: map_at_1
223
+ value: 32.411
224
+ - type: map_at_10
225
+ value: 45.162
226
+ - type: map_at_100
227
+ value: 46.717
228
+ - type: map_at_1000
229
+ value: 46.836
230
+ - type: map_at_3
231
+ value: 41.428
232
+ - type: map_at_5
233
+ value: 43.54
234
+ - type: mrr_at_1
235
+ value: 39.914
236
+ - type: mrr_at_10
237
+ value: 51.534
238
+ - type: mrr_at_100
239
+ value: 52.185
240
+ - type: mrr_at_1000
241
+ value: 52.22
242
+ - type: mrr_at_3
243
+ value: 49.046
244
+ - type: mrr_at_5
245
+ value: 50.548
246
+ - type: ndcg_at_1
247
+ value: 39.914
248
+ - type: ndcg_at_10
249
+ value: 52.235
250
+ - type: ndcg_at_100
251
+ value: 57.4
252
+ - type: ndcg_at_1000
253
+ value: 58.982
254
+ - type: ndcg_at_3
255
+ value: 47.332
256
+ - type: ndcg_at_5
257
+ value: 49.62
258
+ - type: precision_at_1
259
+ value: 39.914
260
+ - type: precision_at_10
261
+ value: 10.258000000000001
262
+ - type: precision_at_100
263
+ value: 1.6219999999999999
264
+ - type: precision_at_1000
265
+ value: 0.20500000000000002
266
+ - type: precision_at_3
267
+ value: 23.462
268
+ - type: precision_at_5
269
+ value: 16.71
270
+ - type: recall_at_1
271
+ value: 32.411
272
+ - type: recall_at_10
273
+ value: 65.408
274
+ - type: recall_at_100
275
+ value: 87.248
276
+ - type: recall_at_1000
277
+ value: 96.951
278
+ - type: recall_at_3
279
+ value: 50.349999999999994
280
+ - type: recall_at_5
281
+ value: 57.431
282
+ - task:
283
+ type: Retrieval
284
+ dataset:
285
+ type: BeIR/cqadupstack
286
+ name: MTEB CQADupstackEnglishRetrieval
287
+ config: default
288
+ split: test
289
+ revision: None
290
+ metrics:
291
+ - type: map_at_1
292
+ value: 31.911
293
+ - type: map_at_10
294
+ value: 42.608000000000004
295
+ - type: map_at_100
296
+ value: 43.948
297
+ - type: map_at_1000
298
+ value: 44.089
299
+ - type: map_at_3
300
+ value: 39.652
301
+ - type: map_at_5
302
+ value: 41.236
303
+ - type: mrr_at_1
304
+ value: 40.064
305
+ - type: mrr_at_10
306
+ value: 48.916
307
+ - type: mrr_at_100
308
+ value: 49.539
309
+ - type: mrr_at_1000
310
+ value: 49.583
311
+ - type: mrr_at_3
312
+ value: 46.741
313
+ - type: mrr_at_5
314
+ value: 48.037
315
+ - type: ndcg_at_1
316
+ value: 40.064
317
+ - type: ndcg_at_10
318
+ value: 48.442
319
+ - type: ndcg_at_100
320
+ value: 52.798
321
+ - type: ndcg_at_1000
322
+ value: 54.871
323
+ - type: ndcg_at_3
324
+ value: 44.528
325
+ - type: ndcg_at_5
326
+ value: 46.211
327
+ - type: precision_at_1
328
+ value: 40.064
329
+ - type: precision_at_10
330
+ value: 9.178
331
+ - type: precision_at_100
332
+ value: 1.452
333
+ - type: precision_at_1000
334
+ value: 0.193
335
+ - type: precision_at_3
336
+ value: 21.614
337
+ - type: precision_at_5
338
+ value: 15.185
339
+ - type: recall_at_1
340
+ value: 31.911
341
+ - type: recall_at_10
342
+ value: 58.155
343
+ - type: recall_at_100
344
+ value: 76.46300000000001
345
+ - type: recall_at_1000
346
+ value: 89.622
347
+ - type: recall_at_3
348
+ value: 46.195
349
+ - type: recall_at_5
350
+ value: 51.288999999999994
351
+ - task:
352
+ type: Retrieval
353
+ dataset:
354
+ type: BeIR/cqadupstack
355
+ name: MTEB CQADupstackGamingRetrieval
356
+ config: default
357
+ split: test
358
+ revision: None
359
+ metrics:
360
+ - type: map_at_1
361
+ value: 40.597
362
+ - type: map_at_10
363
+ value: 54.290000000000006
364
+ - type: map_at_100
365
+ value: 55.340999999999994
366
+ - type: map_at_1000
367
+ value: 55.388999999999996
368
+ - type: map_at_3
369
+ value: 50.931000000000004
370
+ - type: map_at_5
371
+ value: 52.839999999999996
372
+ - type: mrr_at_1
373
+ value: 46.646
374
+ - type: mrr_at_10
375
+ value: 57.524
376
+ - type: mrr_at_100
377
+ value: 58.225
378
+ - type: mrr_at_1000
379
+ value: 58.245999999999995
380
+ - type: mrr_at_3
381
+ value: 55.235
382
+ - type: mrr_at_5
383
+ value: 56.589
384
+ - type: ndcg_at_1
385
+ value: 46.646
386
+ - type: ndcg_at_10
387
+ value: 60.324999999999996
388
+ - type: ndcg_at_100
389
+ value: 64.30900000000001
390
+ - type: ndcg_at_1000
391
+ value: 65.19
392
+ - type: ndcg_at_3
393
+ value: 54.983000000000004
394
+ - type: ndcg_at_5
395
+ value: 57.621
396
+ - type: precision_at_1
397
+ value: 46.646
398
+ - type: precision_at_10
399
+ value: 9.774
400
+ - type: precision_at_100
401
+ value: 1.265
402
+ - type: precision_at_1000
403
+ value: 0.13799999999999998
404
+ - type: precision_at_3
405
+ value: 24.911
406
+ - type: precision_at_5
407
+ value: 16.977999999999998
408
+ - type: recall_at_1
409
+ value: 40.597
410
+ - type: recall_at_10
411
+ value: 74.773
412
+ - type: recall_at_100
413
+ value: 91.61200000000001
414
+ - type: recall_at_1000
415
+ value: 97.726
416
+ - type: recall_at_3
417
+ value: 60.458
418
+ - type: recall_at_5
419
+ value: 66.956
420
+ - task:
421
+ type: Retrieval
422
+ dataset:
423
+ type: BeIR/cqadupstack
424
+ name: MTEB CQADupstackGisRetrieval
425
+ config: default
426
+ split: test
427
+ revision: None
428
+ metrics:
429
+ - type: map_at_1
430
+ value: 27.122
431
+ - type: map_at_10
432
+ value: 36.711
433
+ - type: map_at_100
434
+ value: 37.775
435
+ - type: map_at_1000
436
+ value: 37.842999999999996
437
+ - type: map_at_3
438
+ value: 33.693
439
+ - type: map_at_5
440
+ value: 35.607
441
+ - type: mrr_at_1
442
+ value: 29.153000000000002
443
+ - type: mrr_at_10
444
+ value: 38.873999999999995
445
+ - type: mrr_at_100
446
+ value: 39.739000000000004
447
+ - type: mrr_at_1000
448
+ value: 39.794000000000004
449
+ - type: mrr_at_3
450
+ value: 36.102000000000004
451
+ - type: mrr_at_5
452
+ value: 37.876
453
+ - type: ndcg_at_1
454
+ value: 29.153000000000002
455
+ - type: ndcg_at_10
456
+ value: 42.048
457
+ - type: ndcg_at_100
458
+ value: 47.144999999999996
459
+ - type: ndcg_at_1000
460
+ value: 48.901
461
+ - type: ndcg_at_3
462
+ value: 36.402
463
+ - type: ndcg_at_5
464
+ value: 39.562999999999995
465
+ - type: precision_at_1
466
+ value: 29.153000000000002
467
+ - type: precision_at_10
468
+ value: 6.4750000000000005
469
+ - type: precision_at_100
470
+ value: 0.951
471
+ - type: precision_at_1000
472
+ value: 0.11299999999999999
473
+ - type: precision_at_3
474
+ value: 15.479999999999999
475
+ - type: precision_at_5
476
+ value: 11.028
477
+ - type: recall_at_1
478
+ value: 27.122
479
+ - type: recall_at_10
480
+ value: 56.279999999999994
481
+ - type: recall_at_100
482
+ value: 79.597
483
+ - type: recall_at_1000
484
+ value: 92.804
485
+ - type: recall_at_3
486
+ value: 41.437000000000005
487
+ - type: recall_at_5
488
+ value: 49.019
489
+ - task:
490
+ type: Retrieval
491
+ dataset:
492
+ type: BeIR/cqadupstack
493
+ name: MTEB CQADupstackMathematicaRetrieval
494
+ config: default
495
+ split: test
496
+ revision: None
497
+ metrics:
498
+ - type: map_at_1
499
+ value: 17.757
500
+ - type: map_at_10
501
+ value: 26.739
502
+ - type: map_at_100
503
+ value: 28.015
504
+ - type: map_at_1000
505
+ value: 28.127999999999997
506
+ - type: map_at_3
507
+ value: 23.986
508
+ - type: map_at_5
509
+ value: 25.514
510
+ - type: mrr_at_1
511
+ value: 22.015
512
+ - type: mrr_at_10
513
+ value: 31.325999999999997
514
+ - type: mrr_at_100
515
+ value: 32.368
516
+ - type: mrr_at_1000
517
+ value: 32.426
518
+ - type: mrr_at_3
519
+ value: 28.897000000000002
520
+ - type: mrr_at_5
521
+ value: 30.147000000000002
522
+ - type: ndcg_at_1
523
+ value: 22.015
524
+ - type: ndcg_at_10
525
+ value: 32.225
526
+ - type: ndcg_at_100
527
+ value: 38.405
528
+ - type: ndcg_at_1000
529
+ value: 40.932
530
+ - type: ndcg_at_3
531
+ value: 27.403
532
+ - type: ndcg_at_5
533
+ value: 29.587000000000003
534
+ - type: precision_at_1
535
+ value: 22.015
536
+ - type: precision_at_10
537
+ value: 5.9830000000000005
538
+ - type: precision_at_100
539
+ value: 1.051
540
+ - type: precision_at_1000
541
+ value: 0.13899999999999998
542
+ - type: precision_at_3
543
+ value: 13.391
544
+ - type: precision_at_5
545
+ value: 9.602
546
+ - type: recall_at_1
547
+ value: 17.757
548
+ - type: recall_at_10
549
+ value: 44.467
550
+ - type: recall_at_100
551
+ value: 71.53699999999999
552
+ - type: recall_at_1000
553
+ value: 89.281
554
+ - type: recall_at_3
555
+ value: 31.095
556
+ - type: recall_at_5
557
+ value: 36.818
558
+ - task:
559
+ type: Retrieval
560
+ dataset:
561
+ type: BeIR/cqadupstack
562
+ name: MTEB CQADupstackPhysicsRetrieval
563
+ config: default
564
+ split: test
565
+ revision: None
566
+ metrics:
567
+ - type: map_at_1
568
+ value: 30.354
569
+ - type: map_at_10
570
+ value: 42.134
571
+ - type: map_at_100
572
+ value: 43.429
573
+ - type: map_at_1000
574
+ value: 43.532
575
+ - type: map_at_3
576
+ value: 38.491
577
+ - type: map_at_5
578
+ value: 40.736
579
+ - type: mrr_at_1
580
+ value: 37.247
581
+ - type: mrr_at_10
582
+ value: 47.775
583
+ - type: mrr_at_100
584
+ value: 48.522999999999996
585
+ - type: mrr_at_1000
586
+ value: 48.567
587
+ - type: mrr_at_3
588
+ value: 45.059
589
+ - type: mrr_at_5
590
+ value: 46.811
591
+ - type: ndcg_at_1
592
+ value: 37.247
593
+ - type: ndcg_at_10
594
+ value: 48.609
595
+ - type: ndcg_at_100
596
+ value: 53.782
597
+ - type: ndcg_at_1000
598
+ value: 55.666000000000004
599
+ - type: ndcg_at_3
600
+ value: 42.866
601
+ - type: ndcg_at_5
602
+ value: 46.001
603
+ - type: precision_at_1
604
+ value: 37.247
605
+ - type: precision_at_10
606
+ value: 8.892999999999999
607
+ - type: precision_at_100
608
+ value: 1.341
609
+ - type: precision_at_1000
610
+ value: 0.168
611
+ - type: precision_at_3
612
+ value: 20.5
613
+ - type: precision_at_5
614
+ value: 14.976
615
+ - type: recall_at_1
616
+ value: 30.354
617
+ - type: recall_at_10
618
+ value: 62.273
619
+ - type: recall_at_100
620
+ value: 83.65599999999999
621
+ - type: recall_at_1000
622
+ value: 95.82000000000001
623
+ - type: recall_at_3
624
+ value: 46.464
625
+ - type: recall_at_5
626
+ value: 54.225
627
+ - task:
628
+ type: Retrieval
629
+ dataset:
630
+ type: BeIR/cqadupstack
631
+ name: MTEB CQADupstackProgrammersRetrieval
632
+ config: default
633
+ split: test
634
+ revision: None
635
+ metrics:
636
+ - type: map_at_1
637
+ value: 26.949
638
+ - type: map_at_10
639
+ value: 37.230000000000004
640
+ - type: map_at_100
641
+ value: 38.644
642
+ - type: map_at_1000
643
+ value: 38.751999999999995
644
+ - type: map_at_3
645
+ value: 33.816
646
+ - type: map_at_5
647
+ value: 35.817
648
+ - type: mrr_at_1
649
+ value: 33.446999999999996
650
+ - type: mrr_at_10
651
+ value: 42.970000000000006
652
+ - type: mrr_at_100
653
+ value: 43.873
654
+ - type: mrr_at_1000
655
+ value: 43.922
656
+ - type: mrr_at_3
657
+ value: 40.467999999999996
658
+ - type: mrr_at_5
659
+ value: 41.861
660
+ - type: ndcg_at_1
661
+ value: 33.446999999999996
662
+ - type: ndcg_at_10
663
+ value: 43.403000000000006
664
+ - type: ndcg_at_100
665
+ value: 49.247
666
+ - type: ndcg_at_1000
667
+ value: 51.361999999999995
668
+ - type: ndcg_at_3
669
+ value: 38.155
670
+ - type: ndcg_at_5
671
+ value: 40.643
672
+ - type: precision_at_1
673
+ value: 33.446999999999996
674
+ - type: precision_at_10
675
+ value: 8.128
676
+ - type: precision_at_100
677
+ value: 1.274
678
+ - type: precision_at_1000
679
+ value: 0.163
680
+ - type: precision_at_3
681
+ value: 18.493000000000002
682
+ - type: precision_at_5
683
+ value: 13.333
684
+ - type: recall_at_1
685
+ value: 26.949
686
+ - type: recall_at_10
687
+ value: 56.006
688
+ - type: recall_at_100
689
+ value: 80.99199999999999
690
+ - type: recall_at_1000
691
+ value: 95.074
692
+ - type: recall_at_3
693
+ value: 40.809
694
+ - type: recall_at_5
695
+ value: 47.57
696
+ - task:
697
+ type: Retrieval
698
+ dataset:
699
+ type: BeIR/cqadupstack
700
+ name: MTEB CQADupstackRetrieval
701
+ config: default
702
+ split: test
703
+ revision: None
704
+ metrics:
705
+ - type: map_at_1
706
+ value: 27.243583333333333
707
+ - type: map_at_10
708
+ value: 37.193250000000006
709
+ - type: map_at_100
710
+ value: 38.44833333333334
711
+ - type: map_at_1000
712
+ value: 38.56083333333333
713
+ - type: map_at_3
714
+ value: 34.06633333333333
715
+ - type: map_at_5
716
+ value: 35.87858333333334
717
+ - type: mrr_at_1
718
+ value: 32.291583333333335
719
+ - type: mrr_at_10
720
+ value: 41.482749999999996
721
+ - type: mrr_at_100
722
+ value: 42.33583333333333
723
+ - type: mrr_at_1000
724
+ value: 42.38683333333333
725
+ - type: mrr_at_3
726
+ value: 38.952999999999996
727
+ - type: mrr_at_5
728
+ value: 40.45333333333333
729
+ - type: ndcg_at_1
730
+ value: 32.291583333333335
731
+ - type: ndcg_at_10
732
+ value: 42.90533333333334
733
+ - type: ndcg_at_100
734
+ value: 48.138666666666666
735
+ - type: ndcg_at_1000
736
+ value: 50.229083333333335
737
+ - type: ndcg_at_3
738
+ value: 37.76133333333334
739
+ - type: ndcg_at_5
740
+ value: 40.31033333333334
741
+ - type: precision_at_1
742
+ value: 32.291583333333335
743
+ - type: precision_at_10
744
+ value: 7.585583333333333
745
+ - type: precision_at_100
746
+ value: 1.2045000000000001
747
+ - type: precision_at_1000
748
+ value: 0.15733333333333335
749
+ - type: precision_at_3
750
+ value: 17.485416666666666
751
+ - type: precision_at_5
752
+ value: 12.5145
753
+ - type: recall_at_1
754
+ value: 27.243583333333333
755
+ - type: recall_at_10
756
+ value: 55.45108333333334
757
+ - type: recall_at_100
758
+ value: 78.25858333333335
759
+ - type: recall_at_1000
760
+ value: 92.61716666666665
761
+ - type: recall_at_3
762
+ value: 41.130583333333334
763
+ - type: recall_at_5
764
+ value: 47.73133333333334
765
+ - task:
766
+ type: Retrieval
767
+ dataset:
768
+ type: BeIR/cqadupstack
769
+ name: MTEB CQADupstackStatsRetrieval
770
+ config: default
771
+ split: test
772
+ revision: None
773
+ metrics:
774
+ - type: map_at_1
775
+ value: 26.325
776
+ - type: map_at_10
777
+ value: 32.795
778
+ - type: map_at_100
779
+ value: 33.96
780
+ - type: map_at_1000
781
+ value: 34.054
782
+ - type: map_at_3
783
+ value: 30.64
784
+ - type: map_at_5
785
+ value: 31.771
786
+ - type: mrr_at_1
787
+ value: 29.908
788
+ - type: mrr_at_10
789
+ value: 35.83
790
+ - type: mrr_at_100
791
+ value: 36.868
792
+ - type: mrr_at_1000
793
+ value: 36.928
794
+ - type: mrr_at_3
795
+ value: 33.896
796
+ - type: mrr_at_5
797
+ value: 34.893
798
+ - type: ndcg_at_1
799
+ value: 29.908
800
+ - type: ndcg_at_10
801
+ value: 36.746
802
+ - type: ndcg_at_100
803
+ value: 42.225
804
+ - type: ndcg_at_1000
805
+ value: 44.523
806
+ - type: ndcg_at_3
807
+ value: 32.82
808
+ - type: ndcg_at_5
809
+ value: 34.583000000000006
810
+ - type: precision_at_1
811
+ value: 29.908
812
+ - type: precision_at_10
813
+ value: 5.6129999999999995
814
+ - type: precision_at_100
815
+ value: 0.9079999999999999
816
+ - type: precision_at_1000
817
+ value: 0.11800000000000001
818
+ - type: precision_at_3
819
+ value: 13.753000000000002
820
+ - type: precision_at_5
821
+ value: 9.417
822
+ - type: recall_at_1
823
+ value: 26.325
824
+ - type: recall_at_10
825
+ value: 45.975
826
+ - type: recall_at_100
827
+ value: 70.393
828
+ - type: recall_at_1000
829
+ value: 87.217
830
+ - type: recall_at_3
831
+ value: 35.195
832
+ - type: recall_at_5
833
+ value: 39.69
834
+ - task:
835
+ type: Retrieval
836
+ dataset:
837
+ type: BeIR/cqadupstack
838
+ name: MTEB CQADupstackTexRetrieval
839
+ config: default
840
+ split: test
841
+ revision: None
842
+ metrics:
843
+ - type: map_at_1
844
+ value: 17.828
845
+ - type: map_at_10
846
+ value: 25.759
847
+ - type: map_at_100
848
+ value: 26.961000000000002
849
+ - type: map_at_1000
850
+ value: 27.094
851
+ - type: map_at_3
852
+ value: 23.166999999999998
853
+ - type: map_at_5
854
+ value: 24.610000000000003
855
+ - type: mrr_at_1
856
+ value: 21.61
857
+ - type: mrr_at_10
858
+ value: 29.605999999999998
859
+ - type: mrr_at_100
860
+ value: 30.586000000000002
861
+ - type: mrr_at_1000
862
+ value: 30.664
863
+ - type: mrr_at_3
864
+ value: 27.214
865
+ - type: mrr_at_5
866
+ value: 28.571
867
+ - type: ndcg_at_1
868
+ value: 21.61
869
+ - type: ndcg_at_10
870
+ value: 30.740000000000002
871
+ - type: ndcg_at_100
872
+ value: 36.332
873
+ - type: ndcg_at_1000
874
+ value: 39.296
875
+ - type: ndcg_at_3
876
+ value: 26.11
877
+ - type: ndcg_at_5
878
+ value: 28.297
879
+ - type: precision_at_1
880
+ value: 21.61
881
+ - type: precision_at_10
882
+ value: 5.643
883
+ - type: precision_at_100
884
+ value: 1.0
885
+ - type: precision_at_1000
886
+ value: 0.14400000000000002
887
+ - type: precision_at_3
888
+ value: 12.4
889
+ - type: precision_at_5
890
+ value: 9.119
891
+ - type: recall_at_1
892
+ value: 17.828
893
+ - type: recall_at_10
894
+ value: 41.876000000000005
895
+ - type: recall_at_100
896
+ value: 66.648
897
+ - type: recall_at_1000
898
+ value: 87.763
899
+ - type: recall_at_3
900
+ value: 28.957
901
+ - type: recall_at_5
902
+ value: 34.494
903
+ - task:
904
+ type: Retrieval
905
+ dataset:
906
+ type: BeIR/cqadupstack
907
+ name: MTEB CQADupstackUnixRetrieval
908
+ config: default
909
+ split: test
910
+ revision: None
911
+ metrics:
912
+ - type: map_at_1
913
+ value: 27.921000000000003
914
+ - type: map_at_10
915
+ value: 37.156
916
+ - type: map_at_100
917
+ value: 38.399
918
+ - type: map_at_1000
919
+ value: 38.498
920
+ - type: map_at_3
921
+ value: 34.134
922
+ - type: map_at_5
923
+ value: 35.936
924
+ - type: mrr_at_1
925
+ value: 32.649
926
+ - type: mrr_at_10
927
+ value: 41.19
928
+ - type: mrr_at_100
929
+ value: 42.102000000000004
930
+ - type: mrr_at_1000
931
+ value: 42.157
932
+ - type: mrr_at_3
933
+ value: 38.464
934
+ - type: mrr_at_5
935
+ value: 40.148
936
+ - type: ndcg_at_1
937
+ value: 32.649
938
+ - type: ndcg_at_10
939
+ value: 42.679
940
+ - type: ndcg_at_100
941
+ value: 48.27
942
+ - type: ndcg_at_1000
943
+ value: 50.312
944
+ - type: ndcg_at_3
945
+ value: 37.269000000000005
946
+ - type: ndcg_at_5
947
+ value: 40.055
948
+ - type: precision_at_1
949
+ value: 32.649
950
+ - type: precision_at_10
951
+ value: 7.155
952
+ - type: precision_at_100
953
+ value: 1.124
954
+ - type: precision_at_1000
955
+ value: 0.14100000000000001
956
+ - type: precision_at_3
957
+ value: 16.791
958
+ - type: precision_at_5
959
+ value: 12.015
960
+ - type: recall_at_1
961
+ value: 27.921000000000003
962
+ - type: recall_at_10
963
+ value: 55.357
964
+ - type: recall_at_100
965
+ value: 79.476
966
+ - type: recall_at_1000
967
+ value: 93.314
968
+ - type: recall_at_3
969
+ value: 40.891
970
+ - type: recall_at_5
971
+ value: 47.851
972
+ - task:
973
+ type: Retrieval
974
+ dataset:
975
+ type: BeIR/cqadupstack
976
+ name: MTEB CQADupstackWebmastersRetrieval
977
+ config: default
978
+ split: test
979
+ revision: None
980
+ metrics:
981
+ - type: map_at_1
982
+ value: 25.524
983
+ - type: map_at_10
984
+ value: 35.135
985
+ - type: map_at_100
986
+ value: 36.665
987
+ - type: map_at_1000
988
+ value: 36.886
989
+ - type: map_at_3
990
+ value: 31.367
991
+ - type: map_at_5
992
+ value: 33.724
993
+ - type: mrr_at_1
994
+ value: 30.631999999999998
995
+ - type: mrr_at_10
996
+ value: 39.616
997
+ - type: mrr_at_100
998
+ value: 40.54
999
+ - type: mrr_at_1000
1000
+ value: 40.585
1001
+ - type: mrr_at_3
1002
+ value: 36.462
1003
+ - type: mrr_at_5
1004
+ value: 38.507999999999996
1005
+ - type: ndcg_at_1
1006
+ value: 30.631999999999998
1007
+ - type: ndcg_at_10
1008
+ value: 41.61
1009
+ - type: ndcg_at_100
1010
+ value: 47.249
1011
+ - type: ndcg_at_1000
1012
+ value: 49.662
1013
+ - type: ndcg_at_3
1014
+ value: 35.421
1015
+ - type: ndcg_at_5
1016
+ value: 38.811
1017
+ - type: precision_at_1
1018
+ value: 30.631999999999998
1019
+ - type: precision_at_10
1020
+ value: 8.123
1021
+ - type: precision_at_100
1022
+ value: 1.5810000000000002
1023
+ - type: precision_at_1000
1024
+ value: 0.245
1025
+ - type: precision_at_3
1026
+ value: 16.337
1027
+ - type: precision_at_5
1028
+ value: 12.568999999999999
1029
+ - type: recall_at_1
1030
+ value: 25.524
1031
+ - type: recall_at_10
1032
+ value: 54.994
1033
+ - type: recall_at_100
1034
+ value: 80.03099999999999
1035
+ - type: recall_at_1000
1036
+ value: 95.25099999999999
1037
+ - type: recall_at_3
1038
+ value: 37.563
1039
+ - type: recall_at_5
1040
+ value: 46.428999999999995
1041
+ - task:
1042
+ type: Retrieval
1043
+ dataset:
1044
+ type: BeIR/cqadupstack
1045
+ name: MTEB CQADupstackWordpressRetrieval
1046
+ config: default
1047
+ split: test
1048
+ revision: None
1049
+ metrics:
1050
+ - type: map_at_1
1051
+ value: 22.224
1052
+ - type: map_at_10
1053
+ value: 30.599999999999998
1054
+ - type: map_at_100
1055
+ value: 31.526
1056
+ - type: map_at_1000
1057
+ value: 31.629
1058
+ - type: map_at_3
1059
+ value: 27.491
1060
+ - type: map_at_5
1061
+ value: 29.212
1062
+ - type: mrr_at_1
1063
+ value: 24.214
1064
+ - type: mrr_at_10
1065
+ value: 32.632
1066
+ - type: mrr_at_100
1067
+ value: 33.482
1068
+ - type: mrr_at_1000
1069
+ value: 33.550000000000004
1070
+ - type: mrr_at_3
1071
+ value: 29.852
1072
+ - type: mrr_at_5
1073
+ value: 31.451
1074
+ - type: ndcg_at_1
1075
+ value: 24.214
1076
+ - type: ndcg_at_10
1077
+ value: 35.802
1078
+ - type: ndcg_at_100
1079
+ value: 40.502
1080
+ - type: ndcg_at_1000
1081
+ value: 43.052
1082
+ - type: ndcg_at_3
1083
+ value: 29.847
1084
+ - type: ndcg_at_5
1085
+ value: 32.732
1086
+ - type: precision_at_1
1087
+ value: 24.214
1088
+ - type: precision_at_10
1089
+ value: 5.804
1090
+ - type: precision_at_100
1091
+ value: 0.885
1092
+ - type: precision_at_1000
1093
+ value: 0.121
1094
+ - type: precision_at_3
1095
+ value: 12.692999999999998
1096
+ - type: precision_at_5
1097
+ value: 9.242
1098
+ - type: recall_at_1
1099
+ value: 22.224
1100
+ - type: recall_at_10
1101
+ value: 49.849
1102
+ - type: recall_at_100
1103
+ value: 71.45
1104
+ - type: recall_at_1000
1105
+ value: 90.583
1106
+ - type: recall_at_3
1107
+ value: 34.153
1108
+ - type: recall_at_5
1109
+ value: 41.004000000000005
1110
+ - task:
1111
+ type: Retrieval
1112
+ dataset:
1113
+ type: climate-fever
1114
+ name: MTEB ClimateFEVER
1115
+ config: default
1116
+ split: test
1117
+ revision: None
1118
+ metrics:
1119
+ - type: map_at_1
1120
+ value: 12.386999999999999
1121
+ - type: map_at_10
1122
+ value: 20.182
1123
+ - type: map_at_100
1124
+ value: 21.86
1125
+ - type: map_at_1000
1126
+ value: 22.054000000000002
1127
+ - type: map_at_3
1128
+ value: 17.165
1129
+ - type: map_at_5
1130
+ value: 18.643
1131
+ - type: mrr_at_1
1132
+ value: 26.906000000000002
1133
+ - type: mrr_at_10
1134
+ value: 37.907999999999994
1135
+ - type: mrr_at_100
1136
+ value: 38.868
1137
+ - type: mrr_at_1000
1138
+ value: 38.913
1139
+ - type: mrr_at_3
1140
+ value: 34.853
1141
+ - type: mrr_at_5
1142
+ value: 36.567
1143
+ - type: ndcg_at_1
1144
+ value: 26.906000000000002
1145
+ - type: ndcg_at_10
1146
+ value: 28.103
1147
+ - type: ndcg_at_100
1148
+ value: 35.073
1149
+ - type: ndcg_at_1000
1150
+ value: 38.653
1151
+ - type: ndcg_at_3
1152
+ value: 23.345
1153
+ - type: ndcg_at_5
1154
+ value: 24.828
1155
+ - type: precision_at_1
1156
+ value: 26.906000000000002
1157
+ - type: precision_at_10
1158
+ value: 8.547
1159
+ - type: precision_at_100
1160
+ value: 1.617
1161
+ - type: precision_at_1000
1162
+ value: 0.22799999999999998
1163
+ - type: precision_at_3
1164
+ value: 17.025000000000002
1165
+ - type: precision_at_5
1166
+ value: 12.834000000000001
1167
+ - type: recall_at_1
1168
+ value: 12.386999999999999
1169
+ - type: recall_at_10
1170
+ value: 33.306999999999995
1171
+ - type: recall_at_100
1172
+ value: 57.516
1173
+ - type: recall_at_1000
1174
+ value: 77.74799999999999
1175
+ - type: recall_at_3
1176
+ value: 21.433
1177
+ - type: recall_at_5
1178
+ value: 25.915
1179
+ - task:
1180
+ type: Retrieval
1181
+ dataset:
1182
+ type: dbpedia-entity
1183
+ name: MTEB DBPedia
1184
+ config: default
1185
+ split: test
1186
+ revision: None
1187
+ metrics:
1188
+ - type: map_at_1
1189
+ value: 9.322
1190
+ - type: map_at_10
1191
+ value: 20.469
1192
+ - type: map_at_100
1193
+ value: 28.638
1194
+ - type: map_at_1000
1195
+ value: 30.433
1196
+ - type: map_at_3
1197
+ value: 14.802000000000001
1198
+ - type: map_at_5
1199
+ value: 17.297
1200
+ - type: mrr_at_1
1201
+ value: 68.75
1202
+ - type: mrr_at_10
1203
+ value: 76.29599999999999
1204
+ - type: mrr_at_100
1205
+ value: 76.62400000000001
1206
+ - type: mrr_at_1000
1207
+ value: 76.633
1208
+ - type: mrr_at_3
1209
+ value: 75.083
1210
+ - type: mrr_at_5
1211
+ value: 75.771
1212
+ - type: ndcg_at_1
1213
+ value: 54.87499999999999
1214
+ - type: ndcg_at_10
1215
+ value: 41.185
1216
+ - type: ndcg_at_100
1217
+ value: 46.400000000000006
1218
+ - type: ndcg_at_1000
1219
+ value: 54.223
1220
+ - type: ndcg_at_3
1221
+ value: 45.489000000000004
1222
+ - type: ndcg_at_5
1223
+ value: 43.161
1224
+ - type: precision_at_1
1225
+ value: 68.75
1226
+ - type: precision_at_10
1227
+ value: 32.300000000000004
1228
+ - type: precision_at_100
1229
+ value: 10.607999999999999
1230
+ - type: precision_at_1000
1231
+ value: 2.237
1232
+ - type: precision_at_3
1233
+ value: 49.083
1234
+ - type: precision_at_5
1235
+ value: 41.6
1236
+ - type: recall_at_1
1237
+ value: 9.322
1238
+ - type: recall_at_10
1239
+ value: 25.696
1240
+ - type: recall_at_100
1241
+ value: 52.898
1242
+ - type: recall_at_1000
1243
+ value: 77.281
1244
+ - type: recall_at_3
1245
+ value: 15.943
1246
+ - type: recall_at_5
1247
+ value: 19.836000000000002
1248
+ - task:
1249
+ type: Classification
1250
+ dataset:
1251
+ type: mteb/emotion
1252
+ name: MTEB EmotionClassification
1253
+ config: default
1254
+ split: test
1255
+ revision: 4f58c6b202a23cf9a4da393831edf4f9183cad37
1256
+ metrics:
1257
+ - type: accuracy
1258
+ value: 48.650000000000006
1259
+ - type: f1
1260
+ value: 43.528467245539396
1261
+ - task:
1262
+ type: Retrieval
1263
+ dataset:
1264
+ type: fever
1265
+ name: MTEB FEVER
1266
+ config: default
1267
+ split: test
1268
+ revision: None
1269
+ metrics:
1270
+ - type: map_at_1
1271
+ value: 66.56
1272
+ - type: map_at_10
1273
+ value: 76.767
1274
+ - type: map_at_100
1275
+ value: 77.054
1276
+ - type: map_at_1000
1277
+ value: 77.068
1278
+ - type: map_at_3
1279
+ value: 75.29299999999999
1280
+ - type: map_at_5
1281
+ value: 76.24
1282
+ - type: mrr_at_1
1283
+ value: 71.842
1284
+ - type: mrr_at_10
1285
+ value: 81.459
1286
+ - type: mrr_at_100
1287
+ value: 81.58800000000001
1288
+ - type: mrr_at_1000
1289
+ value: 81.59100000000001
1290
+ - type: mrr_at_3
1291
+ value: 80.188
1292
+ - type: mrr_at_5
1293
+ value: 81.038
1294
+ - type: ndcg_at_1
1295
+ value: 71.842
1296
+ - type: ndcg_at_10
1297
+ value: 81.51899999999999
1298
+ - type: ndcg_at_100
1299
+ value: 82.544
1300
+ - type: ndcg_at_1000
1301
+ value: 82.829
1302
+ - type: ndcg_at_3
1303
+ value: 78.92
1304
+ - type: ndcg_at_5
1305
+ value: 80.406
1306
+ - type: precision_at_1
1307
+ value: 71.842
1308
+ - type: precision_at_10
1309
+ value: 10.066
1310
+ - type: precision_at_100
1311
+ value: 1.076
1312
+ - type: precision_at_1000
1313
+ value: 0.11199999999999999
1314
+ - type: precision_at_3
1315
+ value: 30.703000000000003
1316
+ - type: precision_at_5
1317
+ value: 19.301
1318
+ - type: recall_at_1
1319
+ value: 66.56
1320
+ - type: recall_at_10
1321
+ value: 91.55
1322
+ - type: recall_at_100
1323
+ value: 95.67099999999999
1324
+ - type: recall_at_1000
1325
+ value: 97.539
1326
+ - type: recall_at_3
1327
+ value: 84.46900000000001
1328
+ - type: recall_at_5
1329
+ value: 88.201
1330
+ - task:
1331
+ type: Retrieval
1332
+ dataset:
1333
+ type: fiqa
1334
+ name: MTEB FiQA2018
1335
+ config: default
1336
+ split: test
1337
+ revision: None
1338
+ metrics:
1339
+ - type: map_at_1
1340
+ value: 20.087
1341
+ - type: map_at_10
1342
+ value: 32.830999999999996
1343
+ - type: map_at_100
1344
+ value: 34.814
1345
+ - type: map_at_1000
1346
+ value: 34.999
1347
+ - type: map_at_3
1348
+ value: 28.198
1349
+ - type: map_at_5
1350
+ value: 30.779
1351
+ - type: mrr_at_1
1352
+ value: 38.889
1353
+ - type: mrr_at_10
1354
+ value: 48.415
1355
+ - type: mrr_at_100
1356
+ value: 49.187
1357
+ - type: mrr_at_1000
1358
+ value: 49.226
1359
+ - type: mrr_at_3
1360
+ value: 45.705
1361
+ - type: mrr_at_5
1362
+ value: 47.225
1363
+ - type: ndcg_at_1
1364
+ value: 38.889
1365
+ - type: ndcg_at_10
1366
+ value: 40.758
1367
+ - type: ndcg_at_100
1368
+ value: 47.671
1369
+ - type: ndcg_at_1000
1370
+ value: 50.744
1371
+ - type: ndcg_at_3
1372
+ value: 36.296
1373
+ - type: ndcg_at_5
1374
+ value: 37.852999999999994
1375
+ - type: precision_at_1
1376
+ value: 38.889
1377
+ - type: precision_at_10
1378
+ value: 11.466
1379
+ - type: precision_at_100
1380
+ value: 1.8499999999999999
1381
+ - type: precision_at_1000
1382
+ value: 0.24
1383
+ - type: precision_at_3
1384
+ value: 24.126
1385
+ - type: precision_at_5
1386
+ value: 18.21
1387
+ - type: recall_at_1
1388
+ value: 20.087
1389
+ - type: recall_at_10
1390
+ value: 48.042
1391
+ - type: recall_at_100
1392
+ value: 73.493
1393
+ - type: recall_at_1000
1394
+ value: 91.851
1395
+ - type: recall_at_3
1396
+ value: 32.694
1397
+ - type: recall_at_5
1398
+ value: 39.099000000000004
1399
+ - task:
1400
+ type: Retrieval
1401
+ dataset:
1402
+ type: hotpotqa
1403
+ name: MTEB HotpotQA
1404
+ config: default
1405
+ split: test
1406
+ revision: None
1407
+ metrics:
1408
+ - type: map_at_1
1409
+ value: 38.096000000000004
1410
+ - type: map_at_10
1411
+ value: 56.99999999999999
1412
+ - type: map_at_100
1413
+ value: 57.914
1414
+ - type: map_at_1000
1415
+ value: 57.984
1416
+ - type: map_at_3
1417
+ value: 53.900999999999996
1418
+ - type: map_at_5
1419
+ value: 55.827000000000005
1420
+ - type: mrr_at_1
1421
+ value: 76.19200000000001
1422
+ - type: mrr_at_10
1423
+ value: 81.955
1424
+ - type: mrr_at_100
1425
+ value: 82.164
1426
+ - type: mrr_at_1000
1427
+ value: 82.173
1428
+ - type: mrr_at_3
1429
+ value: 80.963
1430
+ - type: mrr_at_5
1431
+ value: 81.574
1432
+ - type: ndcg_at_1
1433
+ value: 76.19200000000001
1434
+ - type: ndcg_at_10
1435
+ value: 65.75
1436
+ - type: ndcg_at_100
1437
+ value: 68.949
1438
+ - type: ndcg_at_1000
1439
+ value: 70.342
1440
+ - type: ndcg_at_3
1441
+ value: 61.29
1442
+ - type: ndcg_at_5
1443
+ value: 63.747
1444
+ - type: precision_at_1
1445
+ value: 76.19200000000001
1446
+ - type: precision_at_10
1447
+ value: 13.571
1448
+ - type: precision_at_100
1449
+ value: 1.6070000000000002
1450
+ - type: precision_at_1000
1451
+ value: 0.179
1452
+ - type: precision_at_3
1453
+ value: 38.663
1454
+ - type: precision_at_5
1455
+ value: 25.136999999999997
1456
+ - type: recall_at_1
1457
+ value: 38.096000000000004
1458
+ - type: recall_at_10
1459
+ value: 67.853
1460
+ - type: recall_at_100
1461
+ value: 80.365
1462
+ - type: recall_at_1000
1463
+ value: 89.629
1464
+ - type: recall_at_3
1465
+ value: 57.995
1466
+ - type: recall_at_5
1467
+ value: 62.843
1468
+ - task:
1469
+ type: Classification
1470
+ dataset:
1471
+ type: mteb/imdb
1472
+ name: MTEB ImdbClassification
1473
+ config: default
1474
+ split: test
1475
+ revision: 3d86128a09e091d6018b6d26cad27f2739fc2db7
1476
+ metrics:
1477
+ - type: accuracy
1478
+ value: 85.95200000000001
1479
+ - type: ap
1480
+ value: 80.73847277002109
1481
+ - type: f1
1482
+ value: 85.92406135678594
1483
+ - task:
1484
+ type: Retrieval
1485
+ dataset:
1486
+ type: msmarco
1487
+ name: MTEB MSMARCO
1488
+ config: default
1489
+ split: dev
1490
+ revision: None
1491
+ metrics:
1492
+ - type: map_at_1
1493
+ value: 20.916999999999998
1494
+ - type: map_at_10
1495
+ value: 33.23
1496
+ - type: map_at_100
1497
+ value: 34.427
1498
+ - type: map_at_1000
1499
+ value: 34.477000000000004
1500
+ - type: map_at_3
1501
+ value: 29.292
1502
+ - type: map_at_5
1503
+ value: 31.6
1504
+ - type: mrr_at_1
1505
+ value: 21.547
1506
+ - type: mrr_at_10
1507
+ value: 33.839999999999996
1508
+ - type: mrr_at_100
1509
+ value: 34.979
1510
+ - type: mrr_at_1000
1511
+ value: 35.022999999999996
1512
+ - type: mrr_at_3
1513
+ value: 29.988
1514
+ - type: mrr_at_5
1515
+ value: 32.259
1516
+ - type: ndcg_at_1
1517
+ value: 21.519
1518
+ - type: ndcg_at_10
1519
+ value: 40.209
1520
+ - type: ndcg_at_100
1521
+ value: 45.954
1522
+ - type: ndcg_at_1000
1523
+ value: 47.187
1524
+ - type: ndcg_at_3
1525
+ value: 32.227
1526
+ - type: ndcg_at_5
1527
+ value: 36.347
1528
+ - type: precision_at_1
1529
+ value: 21.519
1530
+ - type: precision_at_10
1531
+ value: 6.447
1532
+ - type: precision_at_100
1533
+ value: 0.932
1534
+ - type: precision_at_1000
1535
+ value: 0.104
1536
+ - type: precision_at_3
1537
+ value: 13.877999999999998
1538
+ - type: precision_at_5
1539
+ value: 10.404
1540
+ - type: recall_at_1
1541
+ value: 20.916999999999998
1542
+ - type: recall_at_10
1543
+ value: 61.7
1544
+ - type: recall_at_100
1545
+ value: 88.202
1546
+ - type: recall_at_1000
1547
+ value: 97.588
1548
+ - type: recall_at_3
1549
+ value: 40.044999999999995
1550
+ - type: recall_at_5
1551
+ value: 49.964999999999996
1552
+ - task:
1553
+ type: Classification
1554
+ dataset:
1555
+ type: mteb/mtop_domain
1556
+ name: MTEB MTOPDomainClassification (en)
1557
+ config: en
1558
+ split: test
1559
+ revision: d80d48c1eb48d3562165c59d59d0034df9fff0bf
1560
+ metrics:
1561
+ - type: accuracy
1562
+ value: 93.02781577747379
1563
+ - type: f1
1564
+ value: 92.83653922768306
1565
+ - task:
1566
+ type: Classification
1567
+ dataset:
1568
+ type: mteb/mtop_intent
1569
+ name: MTEB MTOPIntentClassification (en)
1570
+ config: en
1571
+ split: test
1572
+ revision: ae001d0e6b1228650b7bd1c2c65fb50ad11a8aba
1573
+ metrics:
1574
+ - type: accuracy
1575
+ value: 72.04286365709075
1576
+ - type: f1
1577
+ value: 53.43867658525793
1578
+ - task:
1579
+ type: Classification
1580
+ dataset:
1581
+ type: mteb/amazon_massive_intent
1582
+ name: MTEB MassiveIntentClassification (en)
1583
+ config: en
1584
+ split: test
1585
+ revision: 31efe3c427b0bae9c22cbb560b8f15491cc6bed7
1586
+ metrics:
1587
+ - type: accuracy
1588
+ value: 71.47276395427035
1589
+ - type: f1
1590
+ value: 69.77017399597342
1591
+ - task:
1592
+ type: Classification
1593
+ dataset:
1594
+ type: mteb/amazon_massive_scenario
1595
+ name: MTEB MassiveScenarioClassification (en)
1596
+ config: en
1597
+ split: test
1598
+ revision: 7d571f92784cd94a019292a1f45445077d0ef634
1599
+ metrics:
1600
+ - type: accuracy
1601
+ value: 76.3819771351715
1602
+ - type: f1
1603
+ value: 76.8484533435409
1604
+ - task:
1605
+ type: Clustering
1606
+ dataset:
1607
+ type: mteb/medrxiv-clustering-p2p
1608
+ name: MTEB MedrxivClusteringP2P
1609
+ config: default
1610
+ split: test
1611
+ revision: e7a26af6f3ae46b30dde8737f02c07b1505bcc73
1612
+ metrics:
1613
+ - type: v_measure
1614
+ value: 33.16515993299593
1615
+ - task:
1616
+ type: Clustering
1617
+ dataset:
1618
+ type: mteb/medrxiv-clustering-s2s
1619
+ name: MTEB MedrxivClusteringS2S
1620
+ config: default
1621
+ split: test
1622
+ revision: 35191c8c0dca72d8ff3efcd72aa802307d469663
1623
+ metrics:
1624
+ - type: v_measure
1625
+ value: 31.77145323314774
1626
+ - task:
1627
+ type: Reranking
1628
+ dataset:
1629
+ type: mteb/mind_small
1630
+ name: MTEB MindSmallReranking
1631
+ config: default
1632
+ split: test
1633
+ revision: 3bdac13927fdc888b903db93b2ffdbd90b295a69
1634
+ metrics:
1635
+ - type: map
1636
+ value: 32.53637706586391
1637
+ - type: mrr
1638
+ value: 33.7312926288863
1639
+ - task:
1640
+ type: Retrieval
1641
+ dataset:
1642
+ type: nfcorpus
1643
+ name: MTEB NFCorpus
1644
+ config: default
1645
+ split: test
1646
+ revision: None
1647
+ metrics:
1648
+ - type: map_at_1
1649
+ value: 7.063999999999999
1650
+ - type: map_at_10
1651
+ value: 15.046999999999999
1652
+ - type: map_at_100
1653
+ value: 19.116
1654
+ - type: map_at_1000
1655
+ value: 20.702
1656
+ - type: map_at_3
1657
+ value: 10.932
1658
+ - type: map_at_5
1659
+ value: 12.751999999999999
1660
+ - type: mrr_at_1
1661
+ value: 50.464
1662
+ - type: mrr_at_10
1663
+ value: 58.189
1664
+ - type: mrr_at_100
1665
+ value: 58.733999999999995
1666
+ - type: mrr_at_1000
1667
+ value: 58.769000000000005
1668
+ - type: mrr_at_3
1669
+ value: 56.24400000000001
1670
+ - type: mrr_at_5
1671
+ value: 57.68299999999999
1672
+ - type: ndcg_at_1
1673
+ value: 48.142
1674
+ - type: ndcg_at_10
1675
+ value: 37.897
1676
+ - type: ndcg_at_100
1677
+ value: 35.264
1678
+ - type: ndcg_at_1000
1679
+ value: 44.033
1680
+ - type: ndcg_at_3
1681
+ value: 42.967
1682
+ - type: ndcg_at_5
1683
+ value: 40.815
1684
+ - type: precision_at_1
1685
+ value: 50.15500000000001
1686
+ - type: precision_at_10
1687
+ value: 28.235
1688
+ - type: precision_at_100
1689
+ value: 8.994
1690
+ - type: precision_at_1000
1691
+ value: 2.218
1692
+ - type: precision_at_3
1693
+ value: 40.041
1694
+ - type: precision_at_5
1695
+ value: 35.046
1696
+ - type: recall_at_1
1697
+ value: 7.063999999999999
1698
+ - type: recall_at_10
1699
+ value: 18.598
1700
+ - type: recall_at_100
1701
+ value: 35.577999999999996
1702
+ - type: recall_at_1000
1703
+ value: 67.43
1704
+ - type: recall_at_3
1705
+ value: 11.562999999999999
1706
+ - type: recall_at_5
1707
+ value: 14.771
1708
+ - task:
1709
+ type: Retrieval
1710
+ dataset:
1711
+ type: nq
1712
+ name: MTEB NQ
1713
+ config: default
1714
+ split: test
1715
+ revision: None
1716
+ metrics:
1717
+ - type: map_at_1
1718
+ value: 29.046
1719
+ - type: map_at_10
1720
+ value: 44.808
1721
+ - type: map_at_100
1722
+ value: 45.898
1723
+ - type: map_at_1000
1724
+ value: 45.927
1725
+ - type: map_at_3
1726
+ value: 40.19
1727
+ - type: map_at_5
1728
+ value: 42.897
1729
+ - type: mrr_at_1
1730
+ value: 32.706
1731
+ - type: mrr_at_10
1732
+ value: 47.275
1733
+ - type: mrr_at_100
1734
+ value: 48.075
1735
+ - type: mrr_at_1000
1736
+ value: 48.095
1737
+ - type: mrr_at_3
1738
+ value: 43.463
1739
+ - type: mrr_at_5
1740
+ value: 45.741
1741
+ - type: ndcg_at_1
1742
+ value: 32.706
1743
+ - type: ndcg_at_10
1744
+ value: 52.835
1745
+ - type: ndcg_at_100
1746
+ value: 57.345
1747
+ - type: ndcg_at_1000
1748
+ value: 57.985
1749
+ - type: ndcg_at_3
1750
+ value: 44.171
1751
+ - type: ndcg_at_5
1752
+ value: 48.661
1753
+ - type: precision_at_1
1754
+ value: 32.706
1755
+ - type: precision_at_10
1756
+ value: 8.895999999999999
1757
+ - type: precision_at_100
1758
+ value: 1.143
1759
+ - type: precision_at_1000
1760
+ value: 0.12
1761
+ - type: precision_at_3
1762
+ value: 20.238999999999997
1763
+ - type: precision_at_5
1764
+ value: 14.728
1765
+ - type: recall_at_1
1766
+ value: 29.046
1767
+ - type: recall_at_10
1768
+ value: 74.831
1769
+ - type: recall_at_100
1770
+ value: 94.192
1771
+ - type: recall_at_1000
1772
+ value: 98.897
1773
+ - type: recall_at_3
1774
+ value: 52.37500000000001
1775
+ - type: recall_at_5
1776
+ value: 62.732
1777
+ - task:
1778
+ type: Retrieval
1779
+ dataset:
1780
+ type: quora
1781
+ name: MTEB QuoraRetrieval
1782
+ config: default
1783
+ split: test
1784
+ revision: None
1785
+ metrics:
1786
+ - type: map_at_1
1787
+ value: 70.38799999999999
1788
+ - type: map_at_10
1789
+ value: 84.315
1790
+ - type: map_at_100
1791
+ value: 84.955
1792
+ - type: map_at_1000
1793
+ value: 84.971
1794
+ - type: map_at_3
1795
+ value: 81.33399999999999
1796
+ - type: map_at_5
1797
+ value: 83.21300000000001
1798
+ - type: mrr_at_1
1799
+ value: 81.03
1800
+ - type: mrr_at_10
1801
+ value: 87.395
1802
+ - type: mrr_at_100
1803
+ value: 87.488
1804
+ - type: mrr_at_1000
1805
+ value: 87.48899999999999
1806
+ - type: mrr_at_3
1807
+ value: 86.41499999999999
1808
+ - type: mrr_at_5
1809
+ value: 87.074
1810
+ - type: ndcg_at_1
1811
+ value: 81.04
1812
+ - type: ndcg_at_10
1813
+ value: 88.151
1814
+ - type: ndcg_at_100
1815
+ value: 89.38199999999999
1816
+ - type: ndcg_at_1000
1817
+ value: 89.479
1818
+ - type: ndcg_at_3
1819
+ value: 85.24000000000001
1820
+ - type: ndcg_at_5
1821
+ value: 86.856
1822
+ - type: precision_at_1
1823
+ value: 81.04
1824
+ - type: precision_at_10
1825
+ value: 13.372
1826
+ - type: precision_at_100
1827
+ value: 1.526
1828
+ - type: precision_at_1000
1829
+ value: 0.157
1830
+ - type: precision_at_3
1831
+ value: 37.217
1832
+ - type: precision_at_5
1833
+ value: 24.502
1834
+ - type: recall_at_1
1835
+ value: 70.38799999999999
1836
+ - type: recall_at_10
1837
+ value: 95.452
1838
+ - type: recall_at_100
1839
+ value: 99.59700000000001
1840
+ - type: recall_at_1000
1841
+ value: 99.988
1842
+ - type: recall_at_3
1843
+ value: 87.11
1844
+ - type: recall_at_5
1845
+ value: 91.662
1846
+ - task:
1847
+ type: Clustering
1848
+ dataset:
1849
+ type: mteb/reddit-clustering
1850
+ name: MTEB RedditClustering
1851
+ config: default
1852
+ split: test
1853
+ revision: 24640382cdbf8abc73003fb0fa6d111a705499eb
1854
+ metrics:
1855
+ - type: v_measure
1856
+ value: 59.334991029213235
1857
+ - task:
1858
+ type: Clustering
1859
+ dataset:
1860
+ type: mteb/reddit-clustering-p2p
1861
+ name: MTEB RedditClusteringP2P
1862
+ config: default
1863
+ split: test
1864
+ revision: 282350215ef01743dc01b456c7f5241fa8937f16
1865
+ metrics:
1866
+ - type: v_measure
1867
+ value: 62.586500854616666
1868
+ - task:
1869
+ type: Retrieval
1870
+ dataset:
1871
+ type: scidocs
1872
+ name: MTEB SCIDOCS
1873
+ config: default
1874
+ split: test
1875
+ revision: None
1876
+ metrics:
1877
+ - type: map_at_1
1878
+ value: 5.153
1879
+ - type: map_at_10
1880
+ value: 14.277000000000001
1881
+ - type: map_at_100
1882
+ value: 16.922
1883
+ - type: map_at_1000
1884
+ value: 17.302999999999997
1885
+ - type: map_at_3
1886
+ value: 9.961
1887
+ - type: map_at_5
1888
+ value: 12.257
1889
+ - type: mrr_at_1
1890
+ value: 25.4
1891
+ - type: mrr_at_10
1892
+ value: 37.458000000000006
1893
+ - type: mrr_at_100
1894
+ value: 38.681
1895
+ - type: mrr_at_1000
1896
+ value: 38.722
1897
+ - type: mrr_at_3
1898
+ value: 34.1
1899
+ - type: mrr_at_5
1900
+ value: 36.17
1901
+ - type: ndcg_at_1
1902
+ value: 25.4
1903
+ - type: ndcg_at_10
1904
+ value: 23.132
1905
+ - type: ndcg_at_100
1906
+ value: 32.908
1907
+ - type: ndcg_at_1000
1908
+ value: 38.754
1909
+ - type: ndcg_at_3
1910
+ value: 21.82
1911
+ - type: ndcg_at_5
1912
+ value: 19.353
1913
+ - type: precision_at_1
1914
+ value: 25.4
1915
+ - type: precision_at_10
1916
+ value: 12.1
1917
+ - type: precision_at_100
1918
+ value: 2.628
1919
+ - type: precision_at_1000
1920
+ value: 0.402
1921
+ - type: precision_at_3
1922
+ value: 20.732999999999997
1923
+ - type: precision_at_5
1924
+ value: 17.34
1925
+ - type: recall_at_1
1926
+ value: 5.153
1927
+ - type: recall_at_10
1928
+ value: 24.54
1929
+ - type: recall_at_100
1930
+ value: 53.293
1931
+ - type: recall_at_1000
1932
+ value: 81.57
1933
+ - type: recall_at_3
1934
+ value: 12.613
1935
+ - type: recall_at_5
1936
+ value: 17.577
1937
+ - task:
1938
+ type: STS
1939
+ dataset:
1940
+ type: mteb/sickr-sts
1941
+ name: MTEB SICK-R
1942
+ config: default
1943
+ split: test
1944
+ revision: a6ea5a8cab320b040a23452cc28066d9beae2cee
1945
+ metrics:
1946
+ - type: cos_sim_pearson
1947
+ value: 84.86284404925333
1948
+ - type: cos_sim_spearman
1949
+ value: 78.85870555294795
1950
+ - type: euclidean_pearson
1951
+ value: 82.20105295276093
1952
+ - type: euclidean_spearman
1953
+ value: 78.92125617009592
1954
+ - type: manhattan_pearson
1955
+ value: 82.15840025289069
1956
+ - type: manhattan_spearman
1957
+ value: 78.85955732900803
1958
+ - task:
1959
+ type: STS
1960
+ dataset:
1961
+ type: mteb/sts12-sts
1962
+ name: MTEB STS12
1963
+ config: default
1964
+ split: test
1965
+ revision: a0d554a64d88156834ff5ae9920b964011b16384
1966
+ metrics:
1967
+ - type: cos_sim_pearson
1968
+ value: 84.98747423389027
1969
+ - type: cos_sim_spearman
1970
+ value: 75.71298531799367
1971
+ - type: euclidean_pearson
1972
+ value: 81.59709559192291
1973
+ - type: euclidean_spearman
1974
+ value: 75.40622749225653
1975
+ - type: manhattan_pearson
1976
+ value: 81.55553547608804
1977
+ - type: manhattan_spearman
1978
+ value: 75.39380235424899
1979
+ - task:
1980
+ type: STS
1981
+ dataset:
1982
+ type: mteb/sts13-sts
1983
+ name: MTEB STS13
1984
+ config: default
1985
+ split: test
1986
+ revision: 7e90230a92c190f1bf69ae9002b8cea547a64cca
1987
+ metrics:
1988
+ - type: cos_sim_pearson
1989
+ value: 83.76861330695503
1990
+ - type: cos_sim_spearman
1991
+ value: 85.72991921531624
1992
+ - type: euclidean_pearson
1993
+ value: 84.84504307397536
1994
+ - type: euclidean_spearman
1995
+ value: 86.02679162824732
1996
+ - type: manhattan_pearson
1997
+ value: 84.79969439220142
1998
+ - type: manhattan_spearman
1999
+ value: 85.99238837291625
2000
+ - task:
2001
+ type: STS
2002
+ dataset:
2003
+ type: mteb/sts14-sts
2004
+ name: MTEB STS14
2005
+ config: default
2006
+ split: test
2007
+ revision: 6031580fec1f6af667f0bd2da0a551cf4f0b2375
2008
+ metrics:
2009
+ - type: cos_sim_pearson
2010
+ value: 83.31929747511796
2011
+ - type: cos_sim_spearman
2012
+ value: 81.50806522502528
2013
+ - type: euclidean_pearson
2014
+ value: 82.93936686512777
2015
+ - type: euclidean_spearman
2016
+ value: 81.54403447993224
2017
+ - type: manhattan_pearson
2018
+ value: 82.89696981900828
2019
+ - type: manhattan_spearman
2020
+ value: 81.52817825470865
2021
+ - task:
2022
+ type: STS
2023
+ dataset:
2024
+ type: mteb/sts15-sts
2025
+ name: MTEB STS15
2026
+ config: default
2027
+ split: test
2028
+ revision: ae752c7c21bf194d8b67fd573edf7ae58183cbe3
2029
+ metrics:
2030
+ - type: cos_sim_pearson
2031
+ value: 87.14413295332908
2032
+ - type: cos_sim_spearman
2033
+ value: 88.81032027008195
2034
+ - type: euclidean_pearson
2035
+ value: 88.19205563407645
2036
+ - type: euclidean_spearman
2037
+ value: 88.89738339479216
2038
+ - type: manhattan_pearson
2039
+ value: 88.11075942004189
2040
+ - type: manhattan_spearman
2041
+ value: 88.8297061675564
2042
+ - task:
2043
+ type: STS
2044
+ dataset:
2045
+ type: mteb/sts16-sts
2046
+ name: MTEB STS16
2047
+ config: default
2048
+ split: test
2049
+ revision: 4d8694f8f0e0100860b497b999b3dbed754a0513
2050
+ metrics:
2051
+ - type: cos_sim_pearson
2052
+ value: 82.15980075557017
2053
+ - type: cos_sim_spearman
2054
+ value: 83.81896308594801
2055
+ - type: euclidean_pearson
2056
+ value: 83.11195254311338
2057
+ - type: euclidean_spearman
2058
+ value: 84.10479481755407
2059
+ - type: manhattan_pearson
2060
+ value: 83.13915225100556
2061
+ - type: manhattan_spearman
2062
+ value: 84.09895591027859
2063
+ - task:
2064
+ type: STS
2065
+ dataset:
2066
+ type: mteb/sts17-crosslingual-sts
2067
+ name: MTEB STS17 (en-en)
2068
+ config: en-en
2069
+ split: test
2070
+ revision: af5e6fb845001ecf41f4c1e033ce921939a2a68d
2071
+ metrics:
2072
+ - type: cos_sim_pearson
2073
+ value: 87.93669480147919
2074
+ - type: cos_sim_spearman
2075
+ value: 87.89861394614361
2076
+ - type: euclidean_pearson
2077
+ value: 88.37316413202339
2078
+ - type: euclidean_spearman
2079
+ value: 88.18033817842569
2080
+ - type: manhattan_pearson
2081
+ value: 88.39427578879469
2082
+ - type: manhattan_spearman
2083
+ value: 88.09185009236847
2084
+ - task:
2085
+ type: STS
2086
+ dataset:
2087
+ type: mteb/sts22-crosslingual-sts
2088
+ name: MTEB STS22 (en)
2089
+ config: en
2090
+ split: test
2091
+ revision: 6d1ba47164174a496b7fa5d3569dae26a6813b80
2092
+ metrics:
2093
+ - type: cos_sim_pearson
2094
+ value: 66.62215083348255
2095
+ - type: cos_sim_spearman
2096
+ value: 67.33243665716736
2097
+ - type: euclidean_pearson
2098
+ value: 67.60871701996284
2099
+ - type: euclidean_spearman
2100
+ value: 66.75929225238659
2101
+ - type: manhattan_pearson
2102
+ value: 67.63907838970992
2103
+ - type: manhattan_spearman
2104
+ value: 66.79313656754846
2105
+ - task:
2106
+ type: STS
2107
+ dataset:
2108
+ type: mteb/stsbenchmark-sts
2109
+ name: MTEB STSBenchmark
2110
+ config: default
2111
+ split: test
2112
+ revision: b0fddb56ed78048fa8b90373c8a3cfc37b684831
2113
+ metrics:
2114
+ - type: cos_sim_pearson
2115
+ value: 84.65549191934764
2116
+ - type: cos_sim_spearman
2117
+ value: 85.73266847750143
2118
+ - type: euclidean_pearson
2119
+ value: 85.75609932254318
2120
+ - type: euclidean_spearman
2121
+ value: 85.9452287759371
2122
+ - type: manhattan_pearson
2123
+ value: 85.69717413063573
2124
+ - type: manhattan_spearman
2125
+ value: 85.86546318377046
2126
+ - task:
2127
+ type: Reranking
2128
+ dataset:
2129
+ type: mteb/scidocs-reranking
2130
+ name: MTEB SciDocsRR
2131
+ config: default
2132
+ split: test
2133
+ revision: d3c5e1fc0b855ab6097bf1cda04dd73947d7caab
2134
+ metrics:
2135
+ - type: map
2136
+ value: 87.08164129085783
2137
+ - type: mrr
2138
+ value: 96.2877273416489
2139
+ - task:
2140
+ type: Retrieval
2141
+ dataset:
2142
+ type: scifact
2143
+ name: MTEB SciFact
2144
+ config: default
2145
+ split: test
2146
+ revision: None
2147
+ metrics:
2148
+ - type: map_at_1
2149
+ value: 62.09400000000001
2150
+ - type: map_at_10
2151
+ value: 71.712
2152
+ - type: map_at_100
2153
+ value: 72.128
2154
+ - type: map_at_1000
2155
+ value: 72.14399999999999
2156
+ - type: map_at_3
2157
+ value: 68.93
2158
+ - type: map_at_5
2159
+ value: 70.694
2160
+ - type: mrr_at_1
2161
+ value: 65.0
2162
+ - type: mrr_at_10
2163
+ value: 72.572
2164
+ - type: mrr_at_100
2165
+ value: 72.842
2166
+ - type: mrr_at_1000
2167
+ value: 72.856
2168
+ - type: mrr_at_3
2169
+ value: 70.44399999999999
2170
+ - type: mrr_at_5
2171
+ value: 71.744
2172
+ - type: ndcg_at_1
2173
+ value: 65.0
2174
+ - type: ndcg_at_10
2175
+ value: 76.178
2176
+ - type: ndcg_at_100
2177
+ value: 77.887
2178
+ - type: ndcg_at_1000
2179
+ value: 78.227
2180
+ - type: ndcg_at_3
2181
+ value: 71.367
2182
+ - type: ndcg_at_5
2183
+ value: 73.938
2184
+ - type: precision_at_1
2185
+ value: 65.0
2186
+ - type: precision_at_10
2187
+ value: 10.033
2188
+ - type: precision_at_100
2189
+ value: 1.097
2190
+ - type: precision_at_1000
2191
+ value: 0.11199999999999999
2192
+ - type: precision_at_3
2193
+ value: 27.667
2194
+ - type: precision_at_5
2195
+ value: 18.4
2196
+ - type: recall_at_1
2197
+ value: 62.09400000000001
2198
+ - type: recall_at_10
2199
+ value: 89.022
2200
+ - type: recall_at_100
2201
+ value: 96.833
2202
+ - type: recall_at_1000
2203
+ value: 99.333
2204
+ - type: recall_at_3
2205
+ value: 75.922
2206
+ - type: recall_at_5
2207
+ value: 82.428
2208
+ - task:
2209
+ type: PairClassification
2210
+ dataset:
2211
+ type: mteb/sprintduplicatequestions-pairclassification
2212
+ name: MTEB SprintDuplicateQuestions
2213
+ config: default
2214
+ split: test
2215
+ revision: d66bd1f72af766a5cc4b0ca5e00c162f89e8cc46
2216
+ metrics:
2217
+ - type: cos_sim_accuracy
2218
+ value: 99.82178217821782
2219
+ - type: cos_sim_ap
2220
+ value: 95.71282508220798
2221
+ - type: cos_sim_f1
2222
+ value: 90.73120494335737
2223
+ - type: cos_sim_precision
2224
+ value: 93.52441613588111
2225
+ - type: cos_sim_recall
2226
+ value: 88.1
2227
+ - type: dot_accuracy
2228
+ value: 99.73960396039604
2229
+ - type: dot_ap
2230
+ value: 92.98534606529098
2231
+ - type: dot_f1
2232
+ value: 86.83024536805209
2233
+ - type: dot_precision
2234
+ value: 86.96088264794383
2235
+ - type: dot_recall
2236
+ value: 86.7
2237
+ - type: euclidean_accuracy
2238
+ value: 99.82475247524752
2239
+ - type: euclidean_ap
2240
+ value: 95.72927039014849
2241
+ - type: euclidean_f1
2242
+ value: 90.89974293059126
2243
+ - type: euclidean_precision
2244
+ value: 93.54497354497354
2245
+ - type: euclidean_recall
2246
+ value: 88.4
2247
+ - type: manhattan_accuracy
2248
+ value: 99.82574257425742
2249
+ - type: manhattan_ap
2250
+ value: 95.72142177390405
2251
+ - type: manhattan_f1
2252
+ value: 91.00152516522625
2253
+ - type: manhattan_precision
2254
+ value: 92.55429162357808
2255
+ - type: manhattan_recall
2256
+ value: 89.5
2257
+ - type: max_accuracy
2258
+ value: 99.82574257425742
2259
+ - type: max_ap
2260
+ value: 95.72927039014849
2261
+ - type: max_f1
2262
+ value: 91.00152516522625
2263
+ - task:
2264
+ type: Clustering
2265
+ dataset:
2266
+ type: mteb/stackexchange-clustering
2267
+ name: MTEB StackExchangeClustering
2268
+ config: default
2269
+ split: test
2270
+ revision: 6cbc1f7b2bc0622f2e39d2c77fa502909748c259
2271
+ metrics:
2272
+ - type: v_measure
2273
+ value: 66.63957663468679
2274
+ - task:
2275
+ type: Clustering
2276
+ dataset:
2277
+ type: mteb/stackexchange-clustering-p2p
2278
+ name: MTEB StackExchangeClusteringP2P
2279
+ config: default
2280
+ split: test
2281
+ revision: 815ca46b2622cec33ccafc3735d572c266efdb44
2282
+ metrics:
2283
+ - type: v_measure
2284
+ value: 36.003307257923964
2285
+ - task:
2286
+ type: Reranking
2287
+ dataset:
2288
+ type: mteb/stackoverflowdupquestions-reranking
2289
+ name: MTEB StackOverflowDupQuestions
2290
+ config: default
2291
+ split: test
2292
+ revision: e185fbe320c72810689fc5848eb6114e1ef5ec69
2293
+ metrics:
2294
+ - type: map
2295
+ value: 53.005825525863905
2296
+ - type: mrr
2297
+ value: 53.854683919022165
2298
+ - task:
2299
+ type: Summarization
2300
+ dataset:
2301
+ type: mteb/summeval
2302
+ name: MTEB SummEval
2303
+ config: default
2304
+ split: test
2305
+ revision: cda12ad7615edc362dbf25a00fdd61d3b1eaf93c
2306
+ metrics:
2307
+ - type: cos_sim_pearson
2308
+ value: 30.503611569974098
2309
+ - type: cos_sim_spearman
2310
+ value: 31.17155564248449
2311
+ - type: dot_pearson
2312
+ value: 26.740428413981306
2313
+ - type: dot_spearman
2314
+ value: 26.55727635469746
2315
+ - task:
2316
+ type: Retrieval
2317
+ dataset:
2318
+ type: trec-covid
2319
+ name: MTEB TRECCOVID
2320
+ config: default
2321
+ split: test
2322
+ revision: None
2323
+ metrics:
2324
+ - type: map_at_1
2325
+ value: 0.23600000000000002
2326
+ - type: map_at_10
2327
+ value: 1.7670000000000001
2328
+ - type: map_at_100
2329
+ value: 10.208
2330
+ - type: map_at_1000
2331
+ value: 25.997999999999998
2332
+ - type: map_at_3
2333
+ value: 0.605
2334
+ - type: map_at_5
2335
+ value: 0.9560000000000001
2336
+ - type: mrr_at_1
2337
+ value: 84.0
2338
+ - type: mrr_at_10
2339
+ value: 90.167
2340
+ - type: mrr_at_100
2341
+ value: 90.167
2342
+ - type: mrr_at_1000
2343
+ value: 90.167
2344
+ - type: mrr_at_3
2345
+ value: 89.667
2346
+ - type: mrr_at_5
2347
+ value: 90.167
2348
+ - type: ndcg_at_1
2349
+ value: 77.0
2350
+ - type: ndcg_at_10
2351
+ value: 68.783
2352
+ - type: ndcg_at_100
2353
+ value: 54.196
2354
+ - type: ndcg_at_1000
2355
+ value: 52.077
2356
+ - type: ndcg_at_3
2357
+ value: 71.642
2358
+ - type: ndcg_at_5
2359
+ value: 70.45700000000001
2360
+ - type: precision_at_1
2361
+ value: 84.0
2362
+ - type: precision_at_10
2363
+ value: 73.0
2364
+ - type: precision_at_100
2365
+ value: 55.48
2366
+ - type: precision_at_1000
2367
+ value: 23.102
2368
+ - type: precision_at_3
2369
+ value: 76.0
2370
+ - type: precision_at_5
2371
+ value: 74.8
2372
+ - type: recall_at_1
2373
+ value: 0.23600000000000002
2374
+ - type: recall_at_10
2375
+ value: 1.9869999999999999
2376
+ - type: recall_at_100
2377
+ value: 13.749
2378
+ - type: recall_at_1000
2379
+ value: 50.157
2380
+ - type: recall_at_3
2381
+ value: 0.633
2382
+ - type: recall_at_5
2383
+ value: 1.0290000000000001
2384
+ - task:
2385
+ type: Retrieval
2386
+ dataset:
2387
+ type: webis-touche2020
2388
+ name: MTEB Touche2020
2389
+ config: default
2390
+ split: test
2391
+ revision: None
2392
+ metrics:
2393
+ - type: map_at_1
2394
+ value: 1.437
2395
+ - type: map_at_10
2396
+ value: 8.791
2397
+ - type: map_at_100
2398
+ value: 15.001999999999999
2399
+ - type: map_at_1000
2400
+ value: 16.549
2401
+ - type: map_at_3
2402
+ value: 3.8080000000000003
2403
+ - type: map_at_5
2404
+ value: 5.632000000000001
2405
+ - type: mrr_at_1
2406
+ value: 20.408
2407
+ - type: mrr_at_10
2408
+ value: 36.96
2409
+ - type: mrr_at_100
2410
+ value: 37.912
2411
+ - type: mrr_at_1000
2412
+ value: 37.912
2413
+ - type: mrr_at_3
2414
+ value: 29.592000000000002
2415
+ - type: mrr_at_5
2416
+ value: 34.489999999999995
2417
+ - type: ndcg_at_1
2418
+ value: 19.387999999999998
2419
+ - type: ndcg_at_10
2420
+ value: 22.554
2421
+ - type: ndcg_at_100
2422
+ value: 35.197
2423
+ - type: ndcg_at_1000
2424
+ value: 46.58
2425
+ - type: ndcg_at_3
2426
+ value: 20.285
2427
+ - type: ndcg_at_5
2428
+ value: 21.924
2429
+ - type: precision_at_1
2430
+ value: 20.408
2431
+ - type: precision_at_10
2432
+ value: 21.837
2433
+ - type: precision_at_100
2434
+ value: 7.754999999999999
2435
+ - type: precision_at_1000
2436
+ value: 1.537
2437
+ - type: precision_at_3
2438
+ value: 21.769
2439
+ - type: precision_at_5
2440
+ value: 23.673
2441
+ - type: recall_at_1
2442
+ value: 1.437
2443
+ - type: recall_at_10
2444
+ value: 16.314999999999998
2445
+ - type: recall_at_100
2446
+ value: 47.635
2447
+ - type: recall_at_1000
2448
+ value: 82.963
2449
+ - type: recall_at_3
2450
+ value: 4.955
2451
+ - type: recall_at_5
2452
+ value: 8.805
2453
+ - task:
2454
+ type: Classification
2455
+ dataset:
2456
+ type: mteb/toxic_conversations_50k
2457
+ name: MTEB ToxicConversationsClassification
2458
+ config: default
2459
+ split: test
2460
+ revision: d7c0de2777da35d6aae2200a62c6e0e5af397c4c
2461
+ metrics:
2462
+ - type: accuracy
2463
+ value: 71.6128
2464
+ - type: ap
2465
+ value: 14.279639861175664
2466
+ - type: f1
2467
+ value: 54.922292491204274
2468
+ - task:
2469
+ type: Classification
2470
+ dataset:
2471
+ type: mteb/tweet_sentiment_extraction
2472
+ name: MTEB TweetSentimentExtractionClassification
2473
+ config: default
2474
+ split: test
2475
+ revision: d604517c81ca91fe16a244d1248fc021f9ecee7a
2476
+ metrics:
2477
+ - type: accuracy
2478
+ value: 57.01188455008489
2479
+ - type: f1
2480
+ value: 57.377953019225515
2481
+ - task:
2482
+ type: Clustering
2483
+ dataset:
2484
+ type: mteb/twentynewsgroups-clustering
2485
+ name: MTEB TwentyNewsgroupsClustering
2486
+ config: default
2487
+ split: test
2488
+ revision: 6125ec4e24fa026cec8a478383ee943acfbd5449
2489
+ metrics:
2490
+ - type: v_measure
2491
+ value: 52.306769136544254
2492
+ - task:
2493
+ type: PairClassification
2494
+ dataset:
2495
+ type: mteb/twittersemeval2015-pairclassification
2496
+ name: MTEB TwitterSemEval2015
2497
+ config: default
2498
+ split: test
2499
+ revision: 70970daeab8776df92f5ea462b6173c0b46fd2d1
2500
+ metrics:
2501
+ - type: cos_sim_accuracy
2502
+ value: 85.64701674912082
2503
+ - type: cos_sim_ap
2504
+ value: 72.46600945328552
2505
+ - type: cos_sim_f1
2506
+ value: 67.96572367648784
2507
+ - type: cos_sim_precision
2508
+ value: 61.21801649397336
2509
+ - type: cos_sim_recall
2510
+ value: 76.38522427440633
2511
+ - type: dot_accuracy
2512
+ value: 82.33295583238957
2513
+ - type: dot_ap
2514
+ value: 62.54843443071716
2515
+ - type: dot_f1
2516
+ value: 60.38378562507096
2517
+ - type: dot_precision
2518
+ value: 52.99980067769583
2519
+ - type: dot_recall
2520
+ value: 70.15831134564644
2521
+ - type: euclidean_accuracy
2522
+ value: 85.7423854085951
2523
+ - type: euclidean_ap
2524
+ value: 72.76873850945174
2525
+ - type: euclidean_f1
2526
+ value: 68.23556960543262
2527
+ - type: euclidean_precision
2528
+ value: 61.3344559040202
2529
+ - type: euclidean_recall
2530
+ value: 76.88654353562005
2531
+ - type: manhattan_accuracy
2532
+ value: 85.74834594981225
2533
+ - type: manhattan_ap
2534
+ value: 72.66825372446462
2535
+ - type: manhattan_f1
2536
+ value: 68.21539194662853
2537
+ - type: manhattan_precision
2538
+ value: 62.185056472632496
2539
+ - type: manhattan_recall
2540
+ value: 75.54089709762533
2541
+ - type: max_accuracy
2542
+ value: 85.74834594981225
2543
+ - type: max_ap
2544
+ value: 72.76873850945174
2545
+ - type: max_f1
2546
+ value: 68.23556960543262
2547
+ - task:
2548
+ type: PairClassification
2549
+ dataset:
2550
+ type: mteb/twitterurlcorpus-pairclassification
2551
+ name: MTEB TwitterURLCorpus
2552
+ config: default
2553
+ split: test
2554
+ revision: 8b6510b0b1fa4e4c4f879467980e9be563ec1cdf
2555
+ metrics:
2556
+ - type: cos_sim_accuracy
2557
+ value: 88.73171110334924
2558
+ - type: cos_sim_ap
2559
+ value: 85.51855542063649
2560
+ - type: cos_sim_f1
2561
+ value: 77.95706775700934
2562
+ - type: cos_sim_precision
2563
+ value: 74.12524298805887
2564
+ - type: cos_sim_recall
2565
+ value: 82.20665229442562
2566
+ - type: dot_accuracy
2567
+ value: 86.94842240074514
2568
+ - type: dot_ap
2569
+ value: 80.90995345771762
2570
+ - type: dot_f1
2571
+ value: 74.20765027322403
2572
+ - type: dot_precision
2573
+ value: 70.42594385285575
2574
+ - type: dot_recall
2575
+ value: 78.41854019094548
2576
+ - type: euclidean_accuracy
2577
+ value: 88.73753250281368
2578
+ - type: euclidean_ap
2579
+ value: 85.54712254033734
2580
+ - type: euclidean_f1
2581
+ value: 78.07565728654365
2582
+ - type: euclidean_precision
2583
+ value: 75.1120597652081
2584
+ - type: euclidean_recall
2585
+ value: 81.282722513089
2586
+ - type: manhattan_accuracy
2587
+ value: 88.72588970388482
2588
+ - type: manhattan_ap
2589
+ value: 85.52118291594071
2590
+ - type: manhattan_f1
2591
+ value: 78.04428724070593
2592
+ - type: manhattan_precision
2593
+ value: 74.83219105490002
2594
+ - type: manhattan_recall
2595
+ value: 81.54450261780106
2596
+ - type: max_accuracy
2597
+ value: 88.73753250281368
2598
+ - type: max_ap
2599
+ value: 85.54712254033734
2600
+ - type: max_f1
2601
+ value: 78.07565728654365
2602
+ language:
2603
+ - en
2604
+ license: mit
2605
+ ---
2606
+
2607
+ # gte-base
2608
+
2609
+ General Text Embeddings (GTE) model. [Towards General Text Embeddings with Multi-stage Contrastive Learning](https://arxiv.org/abs/2308.03281)
2610
+
2611
+ The GTE models are trained by Alibaba DAMO Academy. They are mainly based on the BERT framework and currently offer three different sizes of models, including [GTE-large](https://huggingface.co/thenlper/gte-large), [GTE-base](https://huggingface.co/thenlper/gte-base), and [GTE-small](https://huggingface.co/thenlper/gte-small). The GTE models are trained on a large-scale corpus of relevance text pairs, covering a wide range of domains and scenarios. This enables the GTE models to be applied to various downstream tasks of text embeddings, including **information retrieval**, **semantic textual similarity**, **text reranking**, etc.
2612
+
2613
+ ## Metrics
2614
+
2615
+ We compared the performance of the GTE models with other popular text embedding models on the MTEB benchmark. For more detailed comparison results, please refer to the [MTEB leaderboard](https://huggingface.co/spaces/mteb/leaderboard).
2616
+
2617
+
2618
+
2619
+ | Model Name | Model Size (GB) | Dimension | Sequence Length | Average (56) | Clustering (11) | Pair Classification (3) | Reranking (4) | Retrieval (15) | STS (10) | Summarization (1) | Classification (12) |
2620
+ |:----:|:---:|:---:|:---:|:---:|:---:|:---:|:---:|:---:|:---:|:---:|:---:|
2621
+ | [**gte-large**](https://huggingface.co/thenlper/gte-large) | 0.67 | 1024 | 512 | **63.13** | 46.84 | 85.00 | 59.13 | 52.22 | 83.35 | 31.66 | 73.33 |
2622
+ | [**gte-base**](https://huggingface.co/thenlper/gte-base) | 0.22 | 768 | 512 | **62.39** | 46.2 | 84.57 | 58.61 | 51.14 | 82.3 | 31.17 | 73.01 |
2623
+ | [e5-large-v2](https://huggingface.co/intfloat/e5-large-v2) | 1.34 | 1024| 512 | 62.25 | 44.49 | 86.03 | 56.61 | 50.56 | 82.05 | 30.19 | 75.24 |
2624
+ | [e5-base-v2](https://huggingface.co/intfloat/e5-base-v2) | 0.44 | 768 | 512 | 61.5 | 43.80 | 85.73 | 55.91 | 50.29 | 81.05 | 30.28 | 73.84 |
2625
+ | [**gte-small**](https://huggingface.co/thenlper/gte-small) | 0.07 | 384 | 512 | **61.36** | 44.89 | 83.54 | 57.7 | 49.46 | 82.07 | 30.42 | 72.31 |
2626
+ | [text-embedding-ada-002](https://platform.openai.com/docs/guides/embeddings) | - | 1536 | 8192 | 60.99 | 45.9 | 84.89 | 56.32 | 49.25 | 80.97 | 30.8 | 70.93 |
2627
+ | [e5-small-v2](https://huggingface.co/intfloat/e5-base-v2) | 0.13 | 384 | 512 | 59.93 | 39.92 | 84.67 | 54.32 | 49.04 | 80.39 | 31.16 | 72.94 |
2628
+ | [sentence-t5-xxl](https://huggingface.co/sentence-transformers/sentence-t5-xxl) | 9.73 | 768 | 512 | 59.51 | 43.72 | 85.06 | 56.42 | 42.24 | 82.63 | 30.08 | 73.42 |
2629
+ | [all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2) | 0.44 | 768 | 514 | 57.78 | 43.69 | 83.04 | 59.36 | 43.81 | 80.28 | 27.49 | 65.07 |
2630
+ | [sgpt-bloom-7b1-msmarco](https://huggingface.co/bigscience/sgpt-bloom-7b1-msmarco) | 28.27 | 4096 | 2048 | 57.59 | 38.93 | 81.9 | 55.65 | 48.22 | 77.74 | 33.6 | 66.19 |
2631
+ | [all-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L12-v2) | 0.13 | 384 | 512 | 56.53 | 41.81 | 82.41 | 58.44 | 42.69 | 79.8 | 27.9 | 63.21 |
2632
+ | [all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) | 0.09 | 384 | 512 | 56.26 | 42.35 | 82.37 | 58.04 | 41.95 | 78.9 | 30.81 | 63.05 |
2633
+ | [contriever-base-msmarco](https://huggingface.co/nthakur/contriever-base-msmarco) | 0.44 | 768 | 512 | 56.00 | 41.1 | 82.54 | 53.14 | 41.88 | 76.51 | 30.36 | 66.68 |
2634
+ | [sentence-t5-base](https://huggingface.co/sentence-transformers/sentence-t5-base) | 0.22 | 768 | 512 | 55.27 | 40.21 | 85.18 | 53.09 | 33.63 | 81.14 | 31.39 | 69.81 |
2635
+
2636
+
2637
+ ## Usage
2638
+
2639
+ Code example
2640
+
2641
+ ```python
2642
+ import torch.nn.functional as F
2643
+ from torch import Tensor
2644
+ from transformers import AutoTokenizer, AutoModel
2645
+
2646
+ def average_pool(last_hidden_states: Tensor,
2647
+ attention_mask: Tensor) -> Tensor:
2648
+ last_hidden = last_hidden_states.masked_fill(~attention_mask[..., None].bool(), 0.0)
2649
+ return last_hidden.sum(dim=1) / attention_mask.sum(dim=1)[..., None]
2650
+
2651
+ input_texts = [
2652
+ "what is the capital of China?",
2653
+ "how to implement quick sort in python?",
2654
+ "Beijing",
2655
+ "sorting algorithms"
2656
+ ]
2657
+
2658
+ tokenizer = AutoTokenizer.from_pretrained("thenlper/gte-base")
2659
+ model = AutoModel.from_pretrained("thenlper/gte-base")
2660
+
2661
+ # Tokenize the input texts
2662
+ batch_dict = tokenizer(input_texts, max_length=512, padding=True, truncation=True, return_tensors='pt')
2663
+
2664
+ outputs = model(**batch_dict)
2665
+ embeddings = average_pool(outputs.last_hidden_state, batch_dict['attention_mask'])
2666
+
2667
+ # (Optionally) normalize embeddings
2668
+ embeddings = F.normalize(embeddings, p=2, dim=1)
2669
+ scores = (embeddings[:1] @ embeddings[1:].T) * 100
2670
+ print(scores.tolist())
2671
+ ```
2672
+
2673
+ Use with sentence-transformers:
2674
+ ```python
2675
+ from sentence_transformers import SentenceTransformer
2676
+ from sentence_transformers.util import cos_sim
2677
+
2678
+ sentences = ['That is a happy person', 'That is a very happy person']
2679
+
2680
+ model = SentenceTransformer('thenlper/gte-base')
2681
+ embeddings = model.encode(sentences)
2682
+ print(cos_sim(embeddings[0], embeddings[1]))
2683
+ ```
2684
+
2685
+ ### Limitation
2686
+
2687
+ This model exclusively caters to English texts, and any lengthy texts will be truncated to a maximum of 512 tokens.
2688
+
2689
+ ### Citation
2690
+
2691
+ If you find our paper or models helpful, please consider citing them as follows:
2692
+
2693
+ ```
2694
+ @misc{li2023general,
2695
+ title={Towards General Text Embeddings with Multi-stage Contrastive Learning},
2696
+ author={Zehan Li and Xin Zhang and Yanzhao Zhang and Dingkun Long and Pengjun Xie and Meishan Zhang},
2697
+ year={2023},
2698
+ eprint={2308.03281},
2699
+ archivePrefix={arXiv},
2700
+ primaryClass={cs.CL}
2701
+ }
2702
+ ```
3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/root/.cache/torch/sentence_transformers/thenlper_gte-base/",
3
+ "architectures": [
4
+ "BertModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.36.2",
23
+ "type_vocab_size": 2,
24
+ "use_cache": true,
25
+ "vocab_size": 30522
26
+ }
3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.2.2",
4
+ "transformers": "4.36.2",
5
+ "pytorch": "2.1.2+cu121"
6
+ }
7
+ }
3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aba45b56e1121519c7e01fd78752dbb3ed583195732cd8cfc2564a8ac75f01f4
3
+ size 437951328
3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
+ }
20
+ ]
3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/tokenizer_config.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "mask_token": "[MASK]",
48
+ "max_length": 128,
49
+ "model_max_length": 512,
50
+ "pad_to_multiple_of": null,
51
+ "pad_token": "[PAD]",
52
+ "pad_token_type_id": 0,
53
+ "padding_side": "right",
54
+ "sep_token": "[SEP]",
55
+ "stride": 0,
56
+ "strip_accents": null,
57
+ "tokenize_chinese_chars": true,
58
+ "tokenizer_class": "BertTokenizer",
59
+ "truncation_side": "right",
60
+ "truncation_strategy": "longest_first",
61
+ "unk_token": "[UNK]"
62
+ }
3_MixtureEmbeddingsModel/expert_01_thenlper_gte-base/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/1_Pooling/config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false
7
+ }
3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/2_Dense/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"in_features": 768, "out_features": 768, "bias": false, "activation_function": "torch.nn.modules.linear.Identity"}
3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/2_Dense/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:747e3b94d4dfb73a2bd470f72915632e71aeb0ae4f7311c520e205b334f9630f
3
+ size 2360634
3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/README.md ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: sentence-similarity
3
+ language: en
4
+ license: apache-2.0
5
+ tags:
6
+ - sentence-transformers
7
+ - feature-extraction
8
+ - sentence-similarity
9
+ - transformers
10
+ ---
11
+
12
+ # sentence-transformers/gtr-t5-base
13
+
14
+ This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space. The model was specifically trained for the task of sematic search.
15
+
16
+ This model was converted from the Tensorflow model [gtr-base-1](https://tfhub.dev/google/gtr/gtr-base/1) to PyTorch. When using this model, have a look at the publication: [Large Dual Encoders Are Generalizable Retrievers](https://arxiv.org/abs/2112.07899). The tfhub model and this PyTorch model can produce slightly different embeddings, however, when run on the same benchmarks, they produce identical results.
17
+
18
+ The model uses only the encoder from a T5-base model. The weights are stored in FP16.
19
+
20
+
21
+ ## Usage (Sentence-Transformers)
22
+
23
+ Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
24
+
25
+ ```
26
+ pip install -U sentence-transformers
27
+ ```
28
+
29
+ Then you can use the model like this:
30
+
31
+ ```python
32
+ from sentence_transformers import SentenceTransformer
33
+ sentences = ["This is an example sentence", "Each sentence is converted"]
34
+
35
+ model = SentenceTransformer('sentence-transformers/gtr-t5-base')
36
+ embeddings = model.encode(sentences)
37
+ print(embeddings)
38
+ ```
39
+
40
+ The model requires sentence-transformers version 2.2.0 or newer.
41
+
42
+ ## Evaluation Results
43
+
44
+ For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name=sentence-transformers/gtr-t5-base)
45
+
46
+
47
+
48
+ ## Citing & Authors
49
+
50
+ If you find this model helpful, please cite the respective publication:
51
+ [Large Dual Encoders Are Generalizable Retrievers](https://arxiv.org/abs/2112.07899)
3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/config.json ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/root/.cache/torch/sentence_transformers/sentence-transformers_gtr-t5-base/",
3
+ "architectures": [
4
+ "T5EncoderModel"
5
+ ],
6
+ "classifier_dropout": 0.0,
7
+ "d_ff": 3072,
8
+ "d_kv": 64,
9
+ "d_model": 768,
10
+ "decoder_start_token_id": 0,
11
+ "dense_act_fn": "relu",
12
+ "dropout_rate": 0.1,
13
+ "eos_token_id": 1,
14
+ "feed_forward_proj": "relu",
15
+ "initializer_factor": 1.0,
16
+ "is_encoder_decoder": true,
17
+ "is_gated_act": false,
18
+ "layer_norm_epsilon": 1e-06,
19
+ "model_type": "t5",
20
+ "n_positions": 512,
21
+ "num_decoder_layers": 12,
22
+ "num_heads": 12,
23
+ "num_layers": 12,
24
+ "output_past": true,
25
+ "pad_token_id": 0,
26
+ "relative_attention_max_distance": 128,
27
+ "relative_attention_num_buckets": 32,
28
+ "task_specific_params": {
29
+ "summarization": {
30
+ "early_stopping": true,
31
+ "length_penalty": 2.0,
32
+ "max_length": 200,
33
+ "min_length": 30,
34
+ "no_repeat_ngram_size": 3,
35
+ "num_beams": 4,
36
+ "prefix": "summarize: "
37
+ },
38
+ "translation_en_to_de": {
39
+ "early_stopping": true,
40
+ "max_length": 300,
41
+ "num_beams": 4,
42
+ "prefix": "translate English to German: "
43
+ },
44
+ "translation_en_to_fr": {
45
+ "early_stopping": true,
46
+ "max_length": 300,
47
+ "num_beams": 4,
48
+ "prefix": "translate English to French: "
49
+ },
50
+ "translation_en_to_ro": {
51
+ "early_stopping": true,
52
+ "max_length": 300,
53
+ "num_beams": 4,
54
+ "prefix": "translate English to Romanian: "
55
+ }
56
+ },
57
+ "torch_dtype": "float32",
58
+ "transformers_version": "4.36.2",
59
+ "use_cache": true,
60
+ "vocab_size": 32128
61
+ }
3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.2.0",
4
+ "transformers": "4.7.0",
5
+ "pytorch": "1.9.0+cu102"
6
+ }
7
+ }
3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:325f2adc00dfdbcd78117693e3f88818c0cd7411818e1650acd73a0369a151d0
3
+ size 438525864
3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/modules.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Dense",
18
+ "type": "sentence_transformers.models.Dense"
19
+ },
20
+ {
21
+ "idx": 3,
22
+ "name": "3",
23
+ "path": "3_Normalize",
24
+ "type": "sentence_transformers.models.Normalize"
25
+ }
26
+ ]
3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/special_tokens_map.json ADDED
@@ -0,0 +1,125 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<extra_id_0>",
4
+ "<extra_id_1>",
5
+ "<extra_id_2>",
6
+ "<extra_id_3>",
7
+ "<extra_id_4>",
8
+ "<extra_id_5>",
9
+ "<extra_id_6>",
10
+ "<extra_id_7>",
11
+ "<extra_id_8>",
12
+ "<extra_id_9>",
13
+ "<extra_id_10>",
14
+ "<extra_id_11>",
15
+ "<extra_id_12>",
16
+ "<extra_id_13>",
17
+ "<extra_id_14>",
18
+ "<extra_id_15>",
19
+ "<extra_id_16>",
20
+ "<extra_id_17>",
21
+ "<extra_id_18>",
22
+ "<extra_id_19>",
23
+ "<extra_id_20>",
24
+ "<extra_id_21>",
25
+ "<extra_id_22>",
26
+ "<extra_id_23>",
27
+ "<extra_id_24>",
28
+ "<extra_id_25>",
29
+ "<extra_id_26>",
30
+ "<extra_id_27>",
31
+ "<extra_id_28>",
32
+ "<extra_id_29>",
33
+ "<extra_id_30>",
34
+ "<extra_id_31>",
35
+ "<extra_id_32>",
36
+ "<extra_id_33>",
37
+ "<extra_id_34>",
38
+ "<extra_id_35>",
39
+ "<extra_id_36>",
40
+ "<extra_id_37>",
41
+ "<extra_id_38>",
42
+ "<extra_id_39>",
43
+ "<extra_id_40>",
44
+ "<extra_id_41>",
45
+ "<extra_id_42>",
46
+ "<extra_id_43>",
47
+ "<extra_id_44>",
48
+ "<extra_id_45>",
49
+ "<extra_id_46>",
50
+ "<extra_id_47>",
51
+ "<extra_id_48>",
52
+ "<extra_id_49>",
53
+ "<extra_id_50>",
54
+ "<extra_id_51>",
55
+ "<extra_id_52>",
56
+ "<extra_id_53>",
57
+ "<extra_id_54>",
58
+ "<extra_id_55>",
59
+ "<extra_id_56>",
60
+ "<extra_id_57>",
61
+ "<extra_id_58>",
62
+ "<extra_id_59>",
63
+ "<extra_id_60>",
64
+ "<extra_id_61>",
65
+ "<extra_id_62>",
66
+ "<extra_id_63>",
67
+ "<extra_id_64>",
68
+ "<extra_id_65>",
69
+ "<extra_id_66>",
70
+ "<extra_id_67>",
71
+ "<extra_id_68>",
72
+ "<extra_id_69>",
73
+ "<extra_id_70>",
74
+ "<extra_id_71>",
75
+ "<extra_id_72>",
76
+ "<extra_id_73>",
77
+ "<extra_id_74>",
78
+ "<extra_id_75>",
79
+ "<extra_id_76>",
80
+ "<extra_id_77>",
81
+ "<extra_id_78>",
82
+ "<extra_id_79>",
83
+ "<extra_id_80>",
84
+ "<extra_id_81>",
85
+ "<extra_id_82>",
86
+ "<extra_id_83>",
87
+ "<extra_id_84>",
88
+ "<extra_id_85>",
89
+ "<extra_id_86>",
90
+ "<extra_id_87>",
91
+ "<extra_id_88>",
92
+ "<extra_id_89>",
93
+ "<extra_id_90>",
94
+ "<extra_id_91>",
95
+ "<extra_id_92>",
96
+ "<extra_id_93>",
97
+ "<extra_id_94>",
98
+ "<extra_id_95>",
99
+ "<extra_id_96>",
100
+ "<extra_id_97>",
101
+ "<extra_id_98>",
102
+ "<extra_id_99>"
103
+ ],
104
+ "eos_token": {
105
+ "content": "</s>",
106
+ "lstrip": false,
107
+ "normalized": false,
108
+ "rstrip": false,
109
+ "single_word": false
110
+ },
111
+ "pad_token": {
112
+ "content": "<pad>",
113
+ "lstrip": false,
114
+ "normalized": false,
115
+ "rstrip": false,
116
+ "single_word": false
117
+ },
118
+ "unk_token": {
119
+ "content": "<unk>",
120
+ "lstrip": false,
121
+ "normalized": false,
122
+ "rstrip": false,
123
+ "single_word": false
124
+ }
125
+ }
3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86
3
+ size 791656
3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
3_MixtureEmbeddingsModel/expert_02_sentence-transformers_gtr-t5-base/tokenizer_config.json ADDED
@@ -0,0 +1,941 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<pad>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "</s>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "<unk>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "32000": {
28
+ "content": "<extra_id_99>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "32001": {
36
+ "content": "<extra_id_98>",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "32002": {
44
+ "content": "<extra_id_97>",
45
+ "lstrip": false,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ },
51
+ "32003": {
52
+ "content": "<extra_id_96>",
53
+ "lstrip": false,
54
+ "normalized": false,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": true
58
+ },
59
+ "32004": {
60
+ "content": "<extra_id_95>",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": true
66
+ },
67
+ "32005": {
68
+ "content": "<extra_id_94>",
69
+ "lstrip": false,
70
+ "normalized": false,
71
+ "rstrip": false,
72
+ "single_word": false,
73
+ "special": true
74
+ },
75
+ "32006": {
76
+ "content": "<extra_id_93>",
77
+ "lstrip": false,
78
+ "normalized": false,
79
+ "rstrip": false,
80
+ "single_word": false,
81
+ "special": true
82
+ },
83
+ "32007": {
84
+ "content": "<extra_id_92>",
85
+ "lstrip": false,
86
+ "normalized": false,
87
+ "rstrip": false,
88
+ "single_word": false,
89
+ "special": true
90
+ },
91
+ "32008": {
92
+ "content": "<extra_id_91>",
93
+ "lstrip": false,
94
+ "normalized": false,
95
+ "rstrip": false,
96
+ "single_word": false,
97
+ "special": true
98
+ },
99
+ "32009": {
100
+ "content": "<extra_id_90>",
101
+ "lstrip": false,
102
+ "normalized": false,
103
+ "rstrip": false,
104
+ "single_word": false,
105
+ "special": true
106
+ },
107
+ "32010": {
108
+ "content": "<extra_id_89>",
109
+ "lstrip": false,
110
+ "normalized": false,
111
+ "rstrip": false,
112
+ "single_word": false,
113
+ "special": true
114
+ },
115
+ "32011": {
116
+ "content": "<extra_id_88>",
117
+ "lstrip": false,
118
+ "normalized": false,
119
+ "rstrip": false,
120
+ "single_word": false,
121
+ "special": true
122
+ },
123
+ "32012": {
124
+ "content": "<extra_id_87>",
125
+ "lstrip": false,
126
+ "normalized": false,
127
+ "rstrip": false,
128
+ "single_word": false,
129
+ "special": true
130
+ },
131
+ "32013": {
132
+ "content": "<extra_id_86>",
133
+ "lstrip": false,
134
+ "normalized": false,
135
+ "rstrip": false,
136
+ "single_word": false,
137
+ "special": true
138
+ },
139
+ "32014": {
140
+ "content": "<extra_id_85>",
141
+ "lstrip": false,
142
+ "normalized": false,
143
+ "rstrip": false,
144
+ "single_word": false,
145
+ "special": true
146
+ },
147
+ "32015": {
148
+ "content": "<extra_id_84>",
149
+ "lstrip": false,
150
+ "normalized": false,
151
+ "rstrip": false,
152
+ "single_word": false,
153
+ "special": true
154
+ },
155
+ "32016": {
156
+ "content": "<extra_id_83>",
157
+ "lstrip": false,
158
+ "normalized": false,
159
+ "rstrip": false,
160
+ "single_word": false,
161
+ "special": true
162
+ },
163
+ "32017": {
164
+ "content": "<extra_id_82>",
165
+ "lstrip": false,
166
+ "normalized": false,
167
+ "rstrip": false,
168
+ "single_word": false,
169
+ "special": true
170
+ },
171
+ "32018": {
172
+ "content": "<extra_id_81>",
173
+ "lstrip": false,
174
+ "normalized": false,
175
+ "rstrip": false,
176
+ "single_word": false,
177
+ "special": true
178
+ },
179
+ "32019": {
180
+ "content": "<extra_id_80>",
181
+ "lstrip": false,
182
+ "normalized": false,
183
+ "rstrip": false,
184
+ "single_word": false,
185
+ "special": true
186
+ },
187
+ "32020": {
188
+ "content": "<extra_id_79>",
189
+ "lstrip": false,
190
+ "normalized": false,
191
+ "rstrip": false,
192
+ "single_word": false,
193
+ "special": true
194
+ },
195
+ "32021": {
196
+ "content": "<extra_id_78>",
197
+ "lstrip": false,
198
+ "normalized": false,
199
+ "rstrip": false,
200
+ "single_word": false,
201
+ "special": true
202
+ },
203
+ "32022": {
204
+ "content": "<extra_id_77>",
205
+ "lstrip": false,
206
+ "normalized": false,
207
+ "rstrip": false,
208
+ "single_word": false,
209
+ "special": true
210
+ },
211
+ "32023": {
212
+ "content": "<extra_id_76>",
213
+ "lstrip": false,
214
+ "normalized": false,
215
+ "rstrip": false,
216
+ "single_word": false,
217
+ "special": true
218
+ },
219
+ "32024": {
220
+ "content": "<extra_id_75>",
221
+ "lstrip": false,
222
+ "normalized": false,
223
+ "rstrip": false,
224
+ "single_word": false,
225
+ "special": true
226
+ },
227
+ "32025": {
228
+ "content": "<extra_id_74>",
229
+ "lstrip": false,
230
+ "normalized": false,
231
+ "rstrip": false,
232
+ "single_word": false,
233
+ "special": true
234
+ },
235
+ "32026": {
236
+ "content": "<extra_id_73>",
237
+ "lstrip": false,
238
+ "normalized": false,
239
+ "rstrip": false,
240
+ "single_word": false,
241
+ "special": true
242
+ },
243
+ "32027": {
244
+ "content": "<extra_id_72>",
245
+ "lstrip": false,
246
+ "normalized": false,
247
+ "rstrip": false,
248
+ "single_word": false,
249
+ "special": true
250
+ },
251
+ "32028": {
252
+ "content": "<extra_id_71>",
253
+ "lstrip": false,
254
+ "normalized": false,
255
+ "rstrip": false,
256
+ "single_word": false,
257
+ "special": true
258
+ },
259
+ "32029": {
260
+ "content": "<extra_id_70>",
261
+ "lstrip": false,
262
+ "normalized": false,
263
+ "rstrip": false,
264
+ "single_word": false,
265
+ "special": true
266
+ },
267
+ "32030": {
268
+ "content": "<extra_id_69>",
269
+ "lstrip": false,
270
+ "normalized": false,
271
+ "rstrip": false,
272
+ "single_word": false,
273
+ "special": true
274
+ },
275
+ "32031": {
276
+ "content": "<extra_id_68>",
277
+ "lstrip": false,
278
+ "normalized": false,
279
+ "rstrip": false,
280
+ "single_word": false,
281
+ "special": true
282
+ },
283
+ "32032": {
284
+ "content": "<extra_id_67>",
285
+ "lstrip": false,
286
+ "normalized": false,
287
+ "rstrip": false,
288
+ "single_word": false,
289
+ "special": true
290
+ },
291
+ "32033": {
292
+ "content": "<extra_id_66>",
293
+ "lstrip": false,
294
+ "normalized": false,
295
+ "rstrip": false,
296
+ "single_word": false,
297
+ "special": true
298
+ },
299
+ "32034": {
300
+ "content": "<extra_id_65>",
301
+ "lstrip": false,
302
+ "normalized": false,
303
+ "rstrip": false,
304
+ "single_word": false,
305
+ "special": true
306
+ },
307
+ "32035": {
308
+ "content": "<extra_id_64>",
309
+ "lstrip": false,
310
+ "normalized": false,
311
+ "rstrip": false,
312
+ "single_word": false,
313
+ "special": true
314
+ },
315
+ "32036": {
316
+ "content": "<extra_id_63>",
317
+ "lstrip": false,
318
+ "normalized": false,
319
+ "rstrip": false,
320
+ "single_word": false,
321
+ "special": true
322
+ },
323
+ "32037": {
324
+ "content": "<extra_id_62>",
325
+ "lstrip": false,
326
+ "normalized": false,
327
+ "rstrip": false,
328
+ "single_word": false,
329
+ "special": true
330
+ },
331
+ "32038": {
332
+ "content": "<extra_id_61>",
333
+ "lstrip": false,
334
+ "normalized": false,
335
+ "rstrip": false,
336
+ "single_word": false,
337
+ "special": true
338
+ },
339
+ "32039": {
340
+ "content": "<extra_id_60>",
341
+ "lstrip": false,
342
+ "normalized": false,
343
+ "rstrip": false,
344
+ "single_word": false,
345
+ "special": true
346
+ },
347
+ "32040": {
348
+ "content": "<extra_id_59>",
349
+ "lstrip": false,
350
+ "normalized": false,
351
+ "rstrip": false,
352
+ "single_word": false,
353
+ "special": true
354
+ },
355
+ "32041": {
356
+ "content": "<extra_id_58>",
357
+ "lstrip": false,
358
+ "normalized": false,
359
+ "rstrip": false,
360
+ "single_word": false,
361
+ "special": true
362
+ },
363
+ "32042": {
364
+ "content": "<extra_id_57>",
365
+ "lstrip": false,
366
+ "normalized": false,
367
+ "rstrip": false,
368
+ "single_word": false,
369
+ "special": true
370
+ },
371
+ "32043": {
372
+ "content": "<extra_id_56>",
373
+ "lstrip": false,
374
+ "normalized": false,
375
+ "rstrip": false,
376
+ "single_word": false,
377
+ "special": true
378
+ },
379
+ "32044": {
380
+ "content": "<extra_id_55>",
381
+ "lstrip": false,
382
+ "normalized": false,
383
+ "rstrip": false,
384
+ "single_word": false,
385
+ "special": true
386
+ },
387
+ "32045": {
388
+ "content": "<extra_id_54>",
389
+ "lstrip": false,
390
+ "normalized": false,
391
+ "rstrip": false,
392
+ "single_word": false,
393
+ "special": true
394
+ },
395
+ "32046": {
396
+ "content": "<extra_id_53>",
397
+ "lstrip": false,
398
+ "normalized": false,
399
+ "rstrip": false,
400
+ "single_word": false,
401
+ "special": true
402
+ },
403
+ "32047": {
404
+ "content": "<extra_id_52>",
405
+ "lstrip": false,
406
+ "normalized": false,
407
+ "rstrip": false,
408
+ "single_word": false,
409
+ "special": true
410
+ },
411
+ "32048": {
412
+ "content": "<extra_id_51>",
413
+ "lstrip": false,
414
+ "normalized": false,
415
+ "rstrip": false,
416
+ "single_word": false,
417
+ "special": true
418
+ },
419
+ "32049": {
420
+ "content": "<extra_id_50>",
421
+ "lstrip": false,
422
+ "normalized": false,
423
+ "rstrip": false,
424
+ "single_word": false,
425
+ "special": true
426
+ },
427
+ "32050": {
428
+ "content": "<extra_id_49>",
429
+ "lstrip": false,
430
+ "normalized": false,
431
+ "rstrip": false,
432
+ "single_word": false,
433
+ "special": true
434
+ },
435
+ "32051": {
436
+ "content": "<extra_id_48>",
437
+ "lstrip": false,
438
+ "normalized": false,
439
+ "rstrip": false,
440
+ "single_word": false,
441
+ "special": true
442
+ },
443
+ "32052": {
444
+ "content": "<extra_id_47>",
445
+ "lstrip": false,
446
+ "normalized": false,
447
+ "rstrip": false,
448
+ "single_word": false,
449
+ "special": true
450
+ },
451
+ "32053": {
452
+ "content": "<extra_id_46>",
453
+ "lstrip": false,
454
+ "normalized": false,
455
+ "rstrip": false,
456
+ "single_word": false,
457
+ "special": true
458
+ },
459
+ "32054": {
460
+ "content": "<extra_id_45>",
461
+ "lstrip": false,
462
+ "normalized": false,
463
+ "rstrip": false,
464
+ "single_word": false,
465
+ "special": true
466
+ },
467
+ "32055": {
468
+ "content": "<extra_id_44>",
469
+ "lstrip": false,
470
+ "normalized": false,
471
+ "rstrip": false,
472
+ "single_word": false,
473
+ "special": true
474
+ },
475
+ "32056": {
476
+ "content": "<extra_id_43>",
477
+ "lstrip": false,
478
+ "normalized": false,
479
+ "rstrip": false,
480
+ "single_word": false,
481
+ "special": true
482
+ },
483
+ "32057": {
484
+ "content": "<extra_id_42>",
485
+ "lstrip": false,
486
+ "normalized": false,
487
+ "rstrip": false,
488
+ "single_word": false,
489
+ "special": true
490
+ },
491
+ "32058": {
492
+ "content": "<extra_id_41>",
493
+ "lstrip": false,
494
+ "normalized": false,
495
+ "rstrip": false,
496
+ "single_word": false,
497
+ "special": true
498
+ },
499
+ "32059": {
500
+ "content": "<extra_id_40>",
501
+ "lstrip": false,
502
+ "normalized": false,
503
+ "rstrip": false,
504
+ "single_word": false,
505
+ "special": true
506
+ },
507
+ "32060": {
508
+ "content": "<extra_id_39>",
509
+ "lstrip": false,
510
+ "normalized": false,
511
+ "rstrip": false,
512
+ "single_word": false,
513
+ "special": true
514
+ },
515
+ "32061": {
516
+ "content": "<extra_id_38>",
517
+ "lstrip": false,
518
+ "normalized": false,
519
+ "rstrip": false,
520
+ "single_word": false,
521
+ "special": true
522
+ },
523
+ "32062": {
524
+ "content": "<extra_id_37>",
525
+ "lstrip": false,
526
+ "normalized": false,
527
+ "rstrip": false,
528
+ "single_word": false,
529
+ "special": true
530
+ },
531
+ "32063": {
532
+ "content": "<extra_id_36>",
533
+ "lstrip": false,
534
+ "normalized": false,
535
+ "rstrip": false,
536
+ "single_word": false,
537
+ "special": true
538
+ },
539
+ "32064": {
540
+ "content": "<extra_id_35>",
541
+ "lstrip": false,
542
+ "normalized": false,
543
+ "rstrip": false,
544
+ "single_word": false,
545
+ "special": true
546
+ },
547
+ "32065": {
548
+ "content": "<extra_id_34>",
549
+ "lstrip": false,
550
+ "normalized": false,
551
+ "rstrip": false,
552
+ "single_word": false,
553
+ "special": true
554
+ },
555
+ "32066": {
556
+ "content": "<extra_id_33>",
557
+ "lstrip": false,
558
+ "normalized": false,
559
+ "rstrip": false,
560
+ "single_word": false,
561
+ "special": true
562
+ },
563
+ "32067": {
564
+ "content": "<extra_id_32>",
565
+ "lstrip": false,
566
+ "normalized": false,
567
+ "rstrip": false,
568
+ "single_word": false,
569
+ "special": true
570
+ },
571
+ "32068": {
572
+ "content": "<extra_id_31>",
573
+ "lstrip": false,
574
+ "normalized": false,
575
+ "rstrip": false,
576
+ "single_word": false,
577
+ "special": true
578
+ },
579
+ "32069": {
580
+ "content": "<extra_id_30>",
581
+ "lstrip": false,
582
+ "normalized": false,
583
+ "rstrip": false,
584
+ "single_word": false,
585
+ "special": true
586
+ },
587
+ "32070": {
588
+ "content": "<extra_id_29>",
589
+ "lstrip": false,
590
+ "normalized": false,
591
+ "rstrip": false,
592
+ "single_word": false,
593
+ "special": true
594
+ },
595
+ "32071": {
596
+ "content": "<extra_id_28>",
597
+ "lstrip": false,
598
+ "normalized": false,
599
+ "rstrip": false,
600
+ "single_word": false,
601
+ "special": true
602
+ },
603
+ "32072": {
604
+ "content": "<extra_id_27>",
605
+ "lstrip": false,
606
+ "normalized": false,
607
+ "rstrip": false,
608
+ "single_word": false,
609
+ "special": true
610
+ },
611
+ "32073": {
612
+ "content": "<extra_id_26>",
613
+ "lstrip": false,
614
+ "normalized": false,
615
+ "rstrip": false,
616
+ "single_word": false,
617
+ "special": true
618
+ },
619
+ "32074": {
620
+ "content": "<extra_id_25>",
621
+ "lstrip": false,
622
+ "normalized": false,
623
+ "rstrip": false,
624
+ "single_word": false,
625
+ "special": true
626
+ },
627
+ "32075": {
628
+ "content": "<extra_id_24>",
629
+ "lstrip": false,
630
+ "normalized": false,
631
+ "rstrip": false,
632
+ "single_word": false,
633
+ "special": true
634
+ },
635
+ "32076": {
636
+ "content": "<extra_id_23>",
637
+ "lstrip": false,
638
+ "normalized": false,
639
+ "rstrip": false,
640
+ "single_word": false,
641
+ "special": true
642
+ },
643
+ "32077": {
644
+ "content": "<extra_id_22>",
645
+ "lstrip": false,
646
+ "normalized": false,
647
+ "rstrip": false,
648
+ "single_word": false,
649
+ "special": true
650
+ },
651
+ "32078": {
652
+ "content": "<extra_id_21>",
653
+ "lstrip": false,
654
+ "normalized": false,
655
+ "rstrip": false,
656
+ "single_word": false,
657
+ "special": true
658
+ },
659
+ "32079": {
660
+ "content": "<extra_id_20>",
661
+ "lstrip": false,
662
+ "normalized": false,
663
+ "rstrip": false,
664
+ "single_word": false,
665
+ "special": true
666
+ },
667
+ "32080": {
668
+ "content": "<extra_id_19>",
669
+ "lstrip": false,
670
+ "normalized": false,
671
+ "rstrip": false,
672
+ "single_word": false,
673
+ "special": true
674
+ },
675
+ "32081": {
676
+ "content": "<extra_id_18>",
677
+ "lstrip": false,
678
+ "normalized": false,
679
+ "rstrip": false,
680
+ "single_word": false,
681
+ "special": true
682
+ },
683
+ "32082": {
684
+ "content": "<extra_id_17>",
685
+ "lstrip": false,
686
+ "normalized": false,
687
+ "rstrip": false,
688
+ "single_word": false,
689
+ "special": true
690
+ },
691
+ "32083": {
692
+ "content": "<extra_id_16>",
693
+ "lstrip": false,
694
+ "normalized": false,
695
+ "rstrip": false,
696
+ "single_word": false,
697
+ "special": true
698
+ },
699
+ "32084": {
700
+ "content": "<extra_id_15>",
701
+ "lstrip": false,
702
+ "normalized": false,
703
+ "rstrip": false,
704
+ "single_word": false,
705
+ "special": true
706
+ },
707
+ "32085": {
708
+ "content": "<extra_id_14>",
709
+ "lstrip": false,
710
+ "normalized": false,
711
+ "rstrip": false,
712
+ "single_word": false,
713
+ "special": true
714
+ },
715
+ "32086": {
716
+ "content": "<extra_id_13>",
717
+ "lstrip": false,
718
+ "normalized": false,
719
+ "rstrip": false,
720
+ "single_word": false,
721
+ "special": true
722
+ },
723
+ "32087": {
724
+ "content": "<extra_id_12>",
725
+ "lstrip": false,
726
+ "normalized": false,
727
+ "rstrip": false,
728
+ "single_word": false,
729
+ "special": true
730
+ },
731
+ "32088": {
732
+ "content": "<extra_id_11>",
733
+ "lstrip": false,
734
+ "normalized": false,
735
+ "rstrip": false,
736
+ "single_word": false,
737
+ "special": true
738
+ },
739
+ "32089": {
740
+ "content": "<extra_id_10>",
741
+ "lstrip": false,
742
+ "normalized": false,
743
+ "rstrip": false,
744
+ "single_word": false,
745
+ "special": true
746
+ },
747
+ "32090": {
748
+ "content": "<extra_id_9>",
749
+ "lstrip": false,
750
+ "normalized": false,
751
+ "rstrip": false,
752
+ "single_word": false,
753
+ "special": true
754
+ },
755
+ "32091": {
756
+ "content": "<extra_id_8>",
757
+ "lstrip": false,
758
+ "normalized": false,
759
+ "rstrip": false,
760
+ "single_word": false,
761
+ "special": true
762
+ },
763
+ "32092": {
764
+ "content": "<extra_id_7>",
765
+ "lstrip": false,
766
+ "normalized": false,
767
+ "rstrip": false,
768
+ "single_word": false,
769
+ "special": true
770
+ },
771
+ "32093": {
772
+ "content": "<extra_id_6>",
773
+ "lstrip": false,
774
+ "normalized": false,
775
+ "rstrip": false,
776
+ "single_word": false,
777
+ "special": true
778
+ },
779
+ "32094": {
780
+ "content": "<extra_id_5>",
781
+ "lstrip": false,
782
+ "normalized": false,
783
+ "rstrip": false,
784
+ "single_word": false,
785
+ "special": true
786
+ },
787
+ "32095": {
788
+ "content": "<extra_id_4>",
789
+ "lstrip": false,
790
+ "normalized": false,
791
+ "rstrip": false,
792
+ "single_word": false,
793
+ "special": true
794
+ },
795
+ "32096": {
796
+ "content": "<extra_id_3>",
797
+ "lstrip": false,
798
+ "normalized": false,
799
+ "rstrip": false,
800
+ "single_word": false,
801
+ "special": true
802
+ },
803
+ "32097": {
804
+ "content": "<extra_id_2>",
805
+ "lstrip": false,
806
+ "normalized": false,
807
+ "rstrip": false,
808
+ "single_word": false,
809
+ "special": true
810
+ },
811
+ "32098": {
812
+ "content": "<extra_id_1>",
813
+ "lstrip": false,
814
+ "normalized": false,
815
+ "rstrip": false,
816
+ "single_word": false,
817
+ "special": true
818
+ },
819
+ "32099": {
820
+ "content": "<extra_id_0>",
821
+ "lstrip": false,
822
+ "normalized": false,
823
+ "rstrip": false,
824
+ "single_word": false,
825
+ "special": true
826
+ }
827
+ },
828
+ "additional_special_tokens": [
829
+ "<extra_id_0>",
830
+ "<extra_id_1>",
831
+ "<extra_id_2>",
832
+ "<extra_id_3>",
833
+ "<extra_id_4>",
834
+ "<extra_id_5>",
835
+ "<extra_id_6>",
836
+ "<extra_id_7>",
837
+ "<extra_id_8>",
838
+ "<extra_id_9>",
839
+ "<extra_id_10>",
840
+ "<extra_id_11>",
841
+ "<extra_id_12>",
842
+ "<extra_id_13>",
843
+ "<extra_id_14>",
844
+ "<extra_id_15>",
845
+ "<extra_id_16>",
846
+ "<extra_id_17>",
847
+ "<extra_id_18>",
848
+ "<extra_id_19>",
849
+ "<extra_id_20>",
850
+ "<extra_id_21>",
851
+ "<extra_id_22>",
852
+ "<extra_id_23>",
853
+ "<extra_id_24>",
854
+ "<extra_id_25>",
855
+ "<extra_id_26>",
856
+ "<extra_id_27>",
857
+ "<extra_id_28>",
858
+ "<extra_id_29>",
859
+ "<extra_id_30>",
860
+ "<extra_id_31>",
861
+ "<extra_id_32>",
862
+ "<extra_id_33>",
863
+ "<extra_id_34>",
864
+ "<extra_id_35>",
865
+ "<extra_id_36>",
866
+ "<extra_id_37>",
867
+ "<extra_id_38>",
868
+ "<extra_id_39>",
869
+ "<extra_id_40>",
870
+ "<extra_id_41>",
871
+ "<extra_id_42>",
872
+ "<extra_id_43>",
873
+ "<extra_id_44>",
874
+ "<extra_id_45>",
875
+ "<extra_id_46>",
876
+ "<extra_id_47>",
877
+ "<extra_id_48>",
878
+ "<extra_id_49>",
879
+ "<extra_id_50>",
880
+ "<extra_id_51>",
881
+ "<extra_id_52>",
882
+ "<extra_id_53>",
883
+ "<extra_id_54>",
884
+ "<extra_id_55>",
885
+ "<extra_id_56>",
886
+ "<extra_id_57>",
887
+ "<extra_id_58>",
888
+ "<extra_id_59>",
889
+ "<extra_id_60>",
890
+ "<extra_id_61>",
891
+ "<extra_id_62>",
892
+ "<extra_id_63>",
893
+ "<extra_id_64>",
894
+ "<extra_id_65>",
895
+ "<extra_id_66>",
896
+ "<extra_id_67>",
897
+ "<extra_id_68>",
898
+ "<extra_id_69>",
899
+ "<extra_id_70>",
900
+ "<extra_id_71>",
901
+ "<extra_id_72>",
902
+ "<extra_id_73>",
903
+ "<extra_id_74>",
904
+ "<extra_id_75>",
905
+ "<extra_id_76>",
906
+ "<extra_id_77>",
907
+ "<extra_id_78>",
908
+ "<extra_id_79>",
909
+ "<extra_id_80>",
910
+ "<extra_id_81>",
911
+ "<extra_id_82>",
912
+ "<extra_id_83>",
913
+ "<extra_id_84>",
914
+ "<extra_id_85>",
915
+ "<extra_id_86>",
916
+ "<extra_id_87>",
917
+ "<extra_id_88>",
918
+ "<extra_id_89>",
919
+ "<extra_id_90>",
920
+ "<extra_id_91>",
921
+ "<extra_id_92>",
922
+ "<extra_id_93>",
923
+ "<extra_id_94>",
924
+ "<extra_id_95>",
925
+ "<extra_id_96>",
926
+ "<extra_id_97>",
927
+ "<extra_id_98>",
928
+ "<extra_id_99>"
929
+ ],
930
+ "clean_up_tokenization_spaces": true,
931
+ "eos_token": "</s>",
932
+ "extra_ids": 100,
933
+ "max_length": null,
934
+ "model_max_length": 512,
935
+ "pad_to_multiple_of": null,
936
+ "pad_token": "<pad>",
937
+ "pad_token_type_id": 0,
938
+ "padding_side": "right",
939
+ "tokenizer_class": "T5Tokenizer",
940
+ "unk_token": "<unk>"
941
+ }
3_MixtureEmbeddingsModel/gate.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f6521fccca52e1517bc6adaee335c408ca0e551eba44e6e0d9b8779d315bff7
3
+ size 6022
README.md ADDED
@@ -0,0 +1,176 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: sentence-similarity
3
+ tags:
4
+ - sentence-transformers
5
+ - feature-extraction
6
+ - sentence-similarity
7
+ language: en
8
+ license: apache-2.0
9
+ datasets:
10
+ - s2orc
11
+ - flax-sentence-embeddings/stackexchange_xml
12
+ - ms_marco
13
+ - gooaq
14
+ - yahoo_answers_topics
15
+ - code_search_net
16
+ - search_qa
17
+ - eli5
18
+ - snli
19
+ - multi_nli
20
+ - wikihow
21
+ - natural_questions
22
+ - trivia_qa
23
+ - embedding-data/sentence-compression
24
+ - embedding-data/flickr30k-captions
25
+ - embedding-data/altlex
26
+ - embedding-data/simple-wiki
27
+ - embedding-data/QQP
28
+ - embedding-data/SPECTER
29
+ - embedding-data/PAQ_pairs
30
+ - embedding-data/WikiAnswers
31
+
32
+ ---
33
+
34
+
35
+ # all-MiniLM-L6-v2
36
+ This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search.
37
+
38
+ ## Usage (Sentence-Transformers)
39
+ Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
40
+
41
+ ```
42
+ pip install -U sentence-transformers
43
+ ```
44
+
45
+ Then you can use the model like this:
46
+ ```python
47
+ from sentence_transformers import SentenceTransformer
48
+ sentences = ["This is an example sentence", "Each sentence is converted"]
49
+
50
+ model = SentenceTransformer('sentence-transformers/all-MiniLM-L6-v2')
51
+ embeddings = model.encode(sentences)
52
+ print(embeddings)
53
+ ```
54
+
55
+ ## Usage (HuggingFace Transformers)
56
+ Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
57
+
58
+ ```python
59
+ from transformers import AutoTokenizer, AutoModel
60
+ import torch
61
+ import torch.nn.functional as F
62
+
63
+ #Mean Pooling - Take attention mask into account for correct averaging
64
+ def mean_pooling(model_output, attention_mask):
65
+ token_embeddings = model_output[0] #First element of model_output contains all token embeddings
66
+ input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
67
+ return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9)
68
+
69
+
70
+ # Sentences we want sentence embeddings for
71
+ sentences = ['This is an example sentence', 'Each sentence is converted']
72
+
73
+ # Load model from HuggingFace Hub
74
+ tokenizer = AutoTokenizer.from_pretrained('sentence-transformers/all-MiniLM-L6-v2')
75
+ model = AutoModel.from_pretrained('sentence-transformers/all-MiniLM-L6-v2')
76
+
77
+ # Tokenize sentences
78
+ encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')
79
+
80
+ # Compute token embeddings
81
+ with torch.no_grad():
82
+ model_output = model(**encoded_input)
83
+
84
+ # Perform pooling
85
+ sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask'])
86
+
87
+ # Normalize embeddings
88
+ sentence_embeddings = F.normalize(sentence_embeddings, p=2, dim=1)
89
+
90
+ print("Sentence embeddings:")
91
+ print(sentence_embeddings)
92
+ ```
93
+
94
+ ## Evaluation Results
95
+
96
+ For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name=sentence-transformers/all-MiniLM-L6-v2)
97
+
98
+ ------
99
+
100
+ ## Background
101
+
102
+ The project aims to train sentence embedding models on very large sentence level datasets using a self-supervised
103
+ contrastive learning objective. We used the pretrained [`nreimers/MiniLM-L6-H384-uncased`](https://huggingface.co/nreimers/MiniLM-L6-H384-uncased) model and fine-tuned in on a
104
+ 1B sentence pairs dataset. We use a contrastive learning objective: given a sentence from the pair, the model should predict which out of a set of randomly sampled other sentences, was actually paired with it in our dataset.
105
+
106
+ We developped this model during the
107
+ [Community week using JAX/Flax for NLP & CV](https://discuss.huggingface.co/t/open-to-the-community-community-week-using-jax-flax-for-nlp-cv/7104),
108
+ organized by Hugging Face. We developped this model as part of the project:
109
+ [Train the Best Sentence Embedding Model Ever with 1B Training Pairs](https://discuss.huggingface.co/t/train-the-best-sentence-embedding-model-ever-with-1b-training-pairs/7354). We benefited from efficient hardware infrastructure to run the project: 7 TPUs v3-8, as well as intervention from Googles Flax, JAX, and Cloud team member about efficient deep learning frameworks.
110
+
111
+ ## Intended uses
112
+
113
+ Our model is intented to be used as a sentence and short paragraph encoder. Given an input text, it ouptuts a vector which captures
114
+ the semantic information. The sentence vector may be used for information retrieval, clustering or sentence similarity tasks.
115
+
116
+ By default, input text longer than 256 word pieces is truncated.
117
+
118
+
119
+ ## Training procedure
120
+
121
+ ### Pre-training
122
+
123
+ We use the pretrained [`nreimers/MiniLM-L6-H384-uncased`](https://huggingface.co/nreimers/MiniLM-L6-H384-uncased) model. Please refer to the model card for more detailed information about the pre-training procedure.
124
+
125
+ ### Fine-tuning
126
+
127
+ We fine-tune the model using a contrastive objective. Formally, we compute the cosine similarity from each possible sentence pairs from the batch.
128
+ We then apply the cross entropy loss by comparing with true pairs.
129
+
130
+ #### Hyper parameters
131
+
132
+ We trained ou model on a TPU v3-8. We train the model during 100k steps using a batch size of 1024 (128 per TPU core).
133
+ We use a learning rate warm up of 500. The sequence length was limited to 128 tokens. We used the AdamW optimizer with
134
+ a 2e-5 learning rate. The full training script is accessible in this current repository: `train_script.py`.
135
+
136
+ #### Training data
137
+
138
+ We use the concatenation from multiple datasets to fine-tune our model. The total number of sentence pairs is above 1 billion sentences.
139
+ We sampled each dataset given a weighted probability which configuration is detailed in the `data_config.json` file.
140
+
141
+
142
+ | Dataset | Paper | Number of training tuples |
143
+ |--------------------------------------------------------|:----------------------------------------:|:--------------------------:|
144
+ | [Reddit comments (2015-2018)](https://github.com/PolyAI-LDN/conversational-datasets/tree/master/reddit) | [paper](https://arxiv.org/abs/1904.06472) | 726,484,430 |
145
+ | [S2ORC](https://github.com/allenai/s2orc) Citation pairs (Abstracts) | [paper](https://aclanthology.org/2020.acl-main.447/) | 116,288,806 |
146
+ | [WikiAnswers](https://github.com/afader/oqa#wikianswers-corpus) Duplicate question pairs | [paper](https://doi.org/10.1145/2623330.2623677) | 77,427,422 |
147
+ | [PAQ](https://github.com/facebookresearch/PAQ) (Question, Answer) pairs | [paper](https://arxiv.org/abs/2102.07033) | 64,371,441 |
148
+ | [S2ORC](https://github.com/allenai/s2orc) Citation pairs (Titles) | [paper](https://aclanthology.org/2020.acl-main.447/) | 52,603,982 |
149
+ | [S2ORC](https://github.com/allenai/s2orc) (Title, Abstract) | [paper](https://aclanthology.org/2020.acl-main.447/) | 41,769,185 |
150
+ | [Stack Exchange](https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_xml) (Title, Body) pairs | - | 25,316,456 |
151
+ | [Stack Exchange](https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_xml) (Title+Body, Answer) pairs | - | 21,396,559 |
152
+ | [Stack Exchange](https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_xml) (Title, Answer) pairs | - | 21,396,559 |
153
+ | [MS MARCO](https://microsoft.github.io/msmarco/) triplets | [paper](https://doi.org/10.1145/3404835.3462804) | 9,144,553 |
154
+ | [GOOAQ: Open Question Answering with Diverse Answer Types](https://github.com/allenai/gooaq) | [paper](https://arxiv.org/pdf/2104.08727.pdf) | 3,012,496 |
155
+ | [Yahoo Answers](https://www.kaggle.com/soumikrakshit/yahoo-answers-dataset) (Title, Answer) | [paper](https://proceedings.neurips.cc/paper/2015/hash/250cf8b51c773f3f8dc8b4be867a9a02-Abstract.html) | 1,198,260 |
156
+ | [Code Search](https://huggingface.co/datasets/code_search_net) | - | 1,151,414 |
157
+ | [COCO](https://cocodataset.org/#home) Image captions | [paper](https://link.springer.com/chapter/10.1007%2F978-3-319-10602-1_48) | 828,395|
158
+ | [SPECTER](https://github.com/allenai/specter) citation triplets | [paper](https://doi.org/10.18653/v1/2020.acl-main.207) | 684,100 |
159
+ | [Yahoo Answers](https://www.kaggle.com/soumikrakshit/yahoo-answers-dataset) (Question, Answer) | [paper](https://proceedings.neurips.cc/paper/2015/hash/250cf8b51c773f3f8dc8b4be867a9a02-Abstract.html) | 681,164 |
160
+ | [Yahoo Answers](https://www.kaggle.com/soumikrakshit/yahoo-answers-dataset) (Title, Question) | [paper](https://proceedings.neurips.cc/paper/2015/hash/250cf8b51c773f3f8dc8b4be867a9a02-Abstract.html) | 659,896 |
161
+ | [SearchQA](https://huggingface.co/datasets/search_qa) | [paper](https://arxiv.org/abs/1704.05179) | 582,261 |
162
+ | [Eli5](https://huggingface.co/datasets/eli5) | [paper](https://doi.org/10.18653/v1/p19-1346) | 325,475 |
163
+ | [Flickr 30k](https://shannon.cs.illinois.edu/DenotationGraph/) | [paper](https://transacl.org/ojs/index.php/tacl/article/view/229/33) | 317,695 |
164
+ | [Stack Exchange](https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_xml) Duplicate questions (titles) | | 304,525 |
165
+ | AllNLI ([SNLI](https://nlp.stanford.edu/projects/snli/) and [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) | [paper SNLI](https://doi.org/10.18653/v1/d15-1075), [paper MultiNLI](https://doi.org/10.18653/v1/n18-1101) | 277,230 |
166
+ | [Stack Exchange](https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_xml) Duplicate questions (bodies) | | 250,519 |
167
+ | [Stack Exchange](https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_xml) Duplicate questions (titles+bodies) | | 250,460 |
168
+ | [Sentence Compression](https://github.com/google-research-datasets/sentence-compression) | [paper](https://www.aclweb.org/anthology/D13-1155/) | 180,000 |
169
+ | [Wikihow](https://github.com/pvl/wikihow_pairs_dataset) | [paper](https://arxiv.org/abs/1810.09305) | 128,542 |
170
+ | [Altlex](https://github.com/chridey/altlex/) | [paper](https://aclanthology.org/P16-1135.pdf) | 112,696 |
171
+ | [Quora Question Triplets](https://quoradata.quora.com/First-Quora-Dataset-Release-Question-Pairs) | - | 103,663 |
172
+ | [Simple Wikipedia](https://cs.pomona.edu/~dkauchak/simplification/) | [paper](https://www.aclweb.org/anthology/P11-2117/) | 102,225 |
173
+ | [Natural Questions (NQ)](https://ai.google.com/research/NaturalQuestions) | [paper](https://transacl.org/ojs/index.php/tacl/article/view/1455) | 100,231 |
174
+ | [SQuAD2.0](https://rajpurkar.github.io/SQuAD-explorer/) | [paper](https://aclanthology.org/P18-2124.pdf) | 87,599 |
175
+ | [TriviaQA](https://huggingface.co/datasets/trivia_qa) | - | 73,346 |
176
+ | **Total** | | **1,170,060,424** |
config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "medi-data/sentence-transformers_all-MiniLM-L6-v2/",
3
+ "architectures": [
4
+ "BertModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 384,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 1536,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 6,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.36.2",
23
+ "type_vocab_size": 2,
24
+ "use_cache": true,
25
+ "vocab_size": 30522
26
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.0.0",
4
+ "transformers": "4.6.1",
5
+ "pytorch": "1.8.1"
6
+ }
7
+ }
freeze_encoder.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ False
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2860a570142753e17e2071260859cbd7ea6d7d4b0cf1b01531a5a6161e966bc3
3
+ size 90864192
modules.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
+ },
20
+ {
21
+ "idx": 3,
22
+ "name": "3",
23
+ "path": "3_MixtureEmbeddingsModel",
24
+ "type": "mixemb.models.MixtureEmbeddingsModel"
25
+ }
26
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 256,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "mask_token": "[MASK]",
49
+ "max_length": 128,
50
+ "model_max_length": 512,
51
+ "never_split": null,
52
+ "pad_to_multiple_of": null,
53
+ "pad_token": "[PAD]",
54
+ "pad_token_type_id": 0,
55
+ "padding_side": "right",
56
+ "sep_token": "[SEP]",
57
+ "stride": 0,
58
+ "strip_accents": null,
59
+ "tokenize_chinese_chars": true,
60
+ "tokenizer_class": "BertTokenizer",
61
+ "truncation_side": "right",
62
+ "truncation_strategy": "longest_first",
63
+ "unk_token": "[UNK]"
64
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff