radoslavralev commited on
Commit
2173df4
·
verified ·
1 Parent(s): dcfbd74

Training in progress, step 5000

Browse files
1_Pooling/config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "word_embedding_dimension": 384,
3
- "pooling_mode_cls_token": false,
4
- "pooling_mode_mean_tokens": true,
5
  "pooling_mode_max_tokens": false,
6
  "pooling_mode_mean_sqrt_len_tokens": false,
7
  "pooling_mode_weightedmean_tokens": false,
 
1
  {
2
+ "word_embedding_dimension": 512,
3
+ "pooling_mode_cls_token": true,
4
+ "pooling_mode_mean_tokens": false,
5
  "pooling_mode_max_tokens": false,
6
  "pooling_mode_mean_sqrt_len_tokens": false,
7
  "pooling_mode_weightedmean_tokens": false,
Information-Retrieval_evaluation_BeIR-touche2020-subset-test_results.csv CHANGED
@@ -1,2 +1,3 @@
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.7551020408163265,0.8979591836734694,0.9387755102040817,1.0,0.7551020408163265,0.01681883497544576,0.6802721088435373,0.045273391320759554,0.6489795918367346,0.07198855467813525,0.563265306122449,0.12417651370073833,0.8328474246841594,0.6023501966395867,0.2434385293084787
 
 
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.7551020408163265,0.8979591836734694,0.9387755102040817,1.0,0.7551020408163265,0.01681883497544576,0.6802721088435373,0.045273391320759554,0.6489795918367346,0.07198855467813525,0.563265306122449,0.12417651370073833,0.8328474246841594,0.6023501966395867,0.2434385293084787
3
+ -1,-1,0.7346938775510204,0.9387755102040817,0.9591836734693877,0.9795918367346939,0.7346938775510204,0.01628348794626864,0.6802721088435373,0.045158349206937544,0.6612244897959185,0.07322241057348641,0.5857142857142857,0.12952247914108056,0.8306122448979592,0.6211658380172899,0.24407616368208723
Information-Retrieval_evaluation_NanoArguAna_results.csv CHANGED
@@ -1,2 +1,3 @@
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.18,0.5,0.66,0.74,0.18,0.18,0.16666666666666663,0.5,0.13200000000000003,0.66,0.07400000000000001,0.74,0.3599682539682539,0.45218312003145433,0.3658170202780539
 
 
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.18,0.5,0.66,0.74,0.18,0.18,0.16666666666666663,0.5,0.13200000000000003,0.66,0.07400000000000001,0.74,0.3599682539682539,0.45218312003145433,0.3658170202780539
3
+ -1,-1,0.14,0.38,0.5,0.66,0.14,0.14,0.12666666666666665,0.38,0.1,0.5,0.06600000000000002,0.66,0.300095238095238,0.38615266678375515,0.3088425567963239
Information-Retrieval_evaluation_NanoClimateFEVER_results.csv CHANGED
@@ -1,2 +1,3 @@
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.1,0.32,0.44,0.6,0.1,0.04333333333333333,0.11333333333333333,0.154,0.092,0.214,0.066,0.2723333333333333,0.23579365079365078,0.18832347198247595,0.13278630044723194
 
 
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.1,0.32,0.44,0.6,0.1,0.04333333333333333,0.11333333333333333,0.154,0.092,0.214,0.066,0.2723333333333333,0.23579365079365078,0.18832347198247595,0.13278630044723194
3
+ -1,-1,0.1,0.26,0.34,0.6,0.1,0.04666666666666666,0.08666666666666666,0.12399999999999999,0.07200000000000001,0.16899999999999998,0.066,0.2973333333333333,0.21938095238095237,0.1853895398720514,0.1267197025068282
Information-Retrieval_evaluation_NanoDBPedia_results.csv CHANGED
@@ -1,2 +1,3 @@
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.58,0.82,0.86,0.88,0.58,0.05774032197349101,0.5266666666666666,0.13525154291594316,0.444,0.17429307751101658,0.36199999999999993,0.24557535689700863,0.7,0.46037876007701023,0.32047705412103555
 
 
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.58,0.82,0.86,0.88,0.58,0.05774032197349101,0.5266666666666666,0.13525154291594316,0.444,0.17429307751101658,0.36199999999999993,0.24557535689700863,0.7,0.46037876007701023,0.32047705412103555
3
+ -1,-1,0.68,0.8,0.86,0.88,0.68,0.08435515343632806,0.5066666666666666,0.13854094008435847,0.44800000000000006,0.17555319487559032,0.376,0.25077742902228617,0.7498333333333334,0.4897177201468777,0.34021634749539026
Information-Retrieval_evaluation_NanoFEVER_results.csv CHANGED
@@ -1,2 +1,3 @@
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.56,0.8,0.88,0.88,0.56,0.5266666666666666,0.28,0.7533333333333333,0.184,0.8333333333333333,0.092,0.8333333333333333,0.6789999999999999,0.7016221865098926,0.6483203115492273
 
 
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.56,0.8,0.88,0.88,0.56,0.5266666666666666,0.28,0.7533333333333333,0.184,0.8333333333333333,0.092,0.8333333333333333,0.6789999999999999,0.7016221865098926,0.6483203115492273
3
+ -1,-1,0.6,0.76,0.86,0.88,0.6,0.5466666666666666,0.2533333333333333,0.7066666666666666,0.17199999999999996,0.7966666666666665,0.088,0.8166666666666665,0.6975555555555556,0.6976146491512496,0.6474697943858455
Information-Retrieval_evaluation_NanoFiQA2018_results.csv CHANGED
@@ -1,2 +1,3 @@
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.24,0.38,0.42,0.52,0.24,0.13307936507936508,0.15333333333333332,0.21543650793650795,0.11600000000000002,0.2601031746031746,0.07600000000000001,0.3241587301587302,0.3178809523809524,0.2635447070179544,0.22615568818713977
 
 
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.24,0.38,0.42,0.52,0.24,0.13307936507936508,0.15333333333333332,0.21543650793650795,0.11600000000000002,0.2601031746031746,0.07600000000000001,0.3241587301587302,0.3178809523809524,0.2635447070179544,0.22615568818713977
3
+ -1,-1,0.26,0.42,0.46,0.58,0.26,0.11974603174603175,0.18,0.25293650793650796,0.124,0.28343650793650793,0.08399999999999999,0.37834920634920627,0.34807936507936504,0.289405411189281,0.2371668409805281
Information-Retrieval_evaluation_NanoHotpotQA_results.csv CHANGED
@@ -1,2 +1,3 @@
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.5,0.58,0.62,0.68,0.5,0.25,0.2333333333333333,0.35,0.16399999999999998,0.41,0.094,0.47,0.5529999999999999,0.43627646197603637,0.37828612061221206
 
 
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.5,0.58,0.62,0.68,0.5,0.25,0.2333333333333333,0.35,0.16399999999999998,0.41,0.094,0.47,0.5529999999999999,0.43627646197603637,0.37828612061221206
3
+ -1,-1,0.54,0.64,0.7,0.74,0.54,0.27,0.26666666666666666,0.4,0.184,0.46,0.106,0.53,0.606,0.48356059599535955,0.4167638398079239
Information-Retrieval_evaluation_NanoMSMARCO_results.csv CHANGED
@@ -1,2 +1,3 @@
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.26,0.52,0.6,0.62,0.26,0.26,0.1733333333333333,0.52,0.12,0.6,0.062,0.62,0.40519047619047627,0.45904886208148177,0.4260102142025637
 
 
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.26,0.52,0.6,0.62,0.26,0.26,0.1733333333333333,0.52,0.12,0.6,0.062,0.62,0.40519047619047627,0.45904886208148177,0.4260102142025637
3
+ -1,-1,0.28,0.54,0.62,0.8,0.28,0.28,0.17999999999999997,0.54,0.124,0.62,0.08,0.8,0.43837301587301575,0.5241911345526384,0.4480618800320956
Information-Retrieval_evaluation_NanoNFCorpus_results.csv CHANGED
@@ -1,2 +1,3 @@
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.32,0.5,0.52,0.64,0.32,0.01260685895239504,0.31333333333333335,0.03463231741620742,0.27599999999999997,0.06850002262294216,0.24399999999999994,0.10851262864104039,0.4212380952380952,0.2703690747449406,0.10029196368651581
 
 
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.32,0.5,0.52,0.64,0.32,0.01260685895239504,0.31333333333333335,0.03463231741620742,0.27599999999999997,0.06850002262294216,0.24399999999999994,0.10851262864104039,0.4212380952380952,0.2703690747449406,0.10029196368651581
3
+ -1,-1,0.34,0.52,0.58,0.68,0.34,0.012133063569139098,0.33333333333333326,0.05465564949455657,0.29600000000000004,0.07192795043792813,0.25,0.10822085751351866,0.44152380952380954,0.2847553576589848,0.108499059841261
Information-Retrieval_evaluation_NanoNQ_results.csv CHANGED
@@ -1,2 +1,3 @@
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.32,0.5,0.6,0.62,0.32,0.3,0.1733333333333333,0.47,0.128,0.58,0.066,0.6,0.4272222222222222,0.4619884812398348,0.42411983365963474
 
 
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.32,0.5,0.6,0.62,0.32,0.3,0.1733333333333333,0.47,0.128,0.58,0.066,0.6,0.4272222222222222,0.4619884812398348,0.42411983365963474
3
+ -1,-1,0.36,0.56,0.6,0.62,0.36,0.35,0.19333333333333333,0.53,0.128,0.58,0.066,0.6,0.4625,0.490897686812855,0.46206363135240186
Information-Retrieval_evaluation_NanoQuoraRetrieval_results.csv CHANGED
@@ -1,2 +1,3 @@
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.94,1.0,1.0,1.0,0.94,0.8273333333333334,0.4133333333333333,0.9653333333333333,0.25999999999999995,0.9793333333333334,0.13799999999999998,0.9966666666666666,0.9666666666666667,0.9697624312418531,0.9551233766233765
 
 
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.94,1.0,1.0,1.0,0.94,0.8273333333333334,0.4133333333333333,0.9653333333333333,0.25999999999999995,0.9793333333333334,0.13799999999999998,0.9966666666666666,0.9666666666666667,0.9697624312418531,0.9551233766233765
3
+ -1,-1,0.92,1.0,1.0,1.0,0.92,0.8073333333333332,0.4133333333333333,0.9653333333333333,0.264,0.986,0.13399999999999998,0.99,0.96,0.960129267031932,0.945489898989899
Information-Retrieval_evaluation_NanoSCIDOCS_results.csv CHANGED
@@ -1,2 +1,3 @@
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.42,0.6,0.72,0.82,0.42,0.08866666666666667,0.33333333333333326,0.20866666666666664,0.272,0.2806666666666667,0.17999999999999997,0.3696666666666666,0.540047619047619,0.36082794471047336,0.2862806075456748
 
 
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.42,0.6,0.72,0.82,0.42,0.08866666666666667,0.33333333333333326,0.20866666666666664,0.272,0.2806666666666667,0.17999999999999997,0.3696666666666666,0.540047619047619,0.36082794471047336,0.2862806075456748
3
+ -1,-1,0.48,0.72,0.76,0.8,0.48,0.10166666666666666,0.3666666666666666,0.22766666666666666,0.28400000000000003,0.29266666666666663,0.19199999999999995,0.3946666666666666,0.5951904761904762,0.3895503827770311,0.3082897117657901
Information-Retrieval_evaluation_NanoSciFact_results.csv CHANGED
@@ -1,2 +1,3 @@
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.38,0.46,0.5,0.58,0.38,0.345,0.15999999999999998,0.43,0.10800000000000003,0.475,0.068,0.58,0.44026984126984126,0.46384622999765257,0.43257979600699104
 
 
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.38,0.46,0.5,0.58,0.38,0.345,0.15999999999999998,0.43,0.10800000000000003,0.475,0.068,0.58,0.44026984126984126,0.46384622999765257,0.43257979600699104
3
+ -1,-1,0.38,0.54,0.58,0.7,0.38,0.345,0.19999999999999996,0.525,0.12800000000000003,0.565,0.08,0.7,0.4820793650793651,0.5292195947118973,0.47730440170572996
Information-Retrieval_evaluation_NanoTouche2020_results.csv CHANGED
@@ -1,2 +1,3 @@
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.46938775510204084,0.8367346938775511,0.9387755102040817,1.0,0.46938775510204084,0.032657982947973084,0.44897959183673464,0.09621881460341672,0.42040816326530606,0.1425551052100505,0.3346938775510204,0.22061476067159091,0.6573129251700679,0.3807140713282222,0.2698119698398041
 
 
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.46938775510204084,0.8367346938775511,0.9387755102040817,1.0,0.46938775510204084,0.032657982947973084,0.44897959183673464,0.09621881460341672,0.42040816326530606,0.1425551052100505,0.3346938775510204,0.22061476067159091,0.6573129251700679,0.3807140713282222,0.2698119698398041
3
+ -1,-1,0.5102040816326531,0.8163265306122449,0.8571428571428571,0.9591836734693877,0.5102040816326531,0.04030730530317779,0.45578231292517,0.10027039527564566,0.4040816326530612,0.14754618693234572,0.3428571428571428,0.2268233238254859,0.6711613216715256,0.3942611497955867,0.28013001290517386
NanoBEIR_evaluation_mean_results.csv CHANGED
@@ -1,2 +1,3 @@
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.4053375196232339,0.601287284144427,0.6737519623233909,0.7369230769230769,0.4053375196232339,0.23516034838101724,0.26838304552590264,0.37175942432349296,0.20895447409733128,0.43675267025234743,0.1428226059654631,0.4908354981821823,0.5156608233036803,0.45145275407225244,0.3820046351353431
 
 
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.4053375196232339,0.601287284144427,0.6737519623233909,0.7369230769230769,0.4053375196232339,0.23516034838101724,0.26838304552590264,0.37175942432349296,0.20895447409733128,0.43675267025234743,0.1428226059654631,0.4908354981821823,0.5156608233036803,0.45145275407225244,0.3820046351353431
3
+ -1,-1,0.4300156985871272,0.6120251177394034,0.6705494505494506,0.7614756671899527,0.4300156985871272,0.24183652979907766,0.2740345368916797,0.3803900122659796,0.20985243328100472,0.4344459364242849,0.14852747252747253,0.5194490371828588,0.536290187137126,0.46960347357534615,0.39284751373578397
README.md CHANGED
@@ -5,231 +5,51 @@ tags:
5
  - feature-extraction
6
  - dense
7
  - generated_from_trainer
8
- - dataset_size:713743
9
  - loss:MultipleNegativesRankingLoss
10
- base_model: sentence-transformers/all-MiniLM-L12-v2
11
  widget:
12
- - source_sentence: 'Abraham Lincoln: Why is the Gettysburg Address so memorable?'
13
  sentences:
14
- - 'Abraham Lincoln: Why is the Gettysburg Address so memorable?'
15
- - What does the Gettysburg Address really mean?
16
- - What is eatalo.com?
17
- - source_sentence: Has the influence of Ancient Carthage in science, math, and society
18
- been underestimated?
19
  sentences:
20
- - How does one earn money online without an investment from home?
21
- - Has the influence of Ancient Carthage in science, math, and society been underestimated?
22
- - Has the influence of the Ancient Etruscans in science and math been underestimated?
23
- - source_sentence: Is there any app that shares charging to others like share it how
24
- we transfer files?
25
  sentences:
26
- - How do you think of Chinese claims that the present Private Arbitration is illegal,
27
- its verdict violates the UNCLOS and is illegal?
28
- - Is there any app that shares charging to others like share it how we transfer
29
- files?
30
- - Are there any platforms that provides end-to-end encryption for file transfer/
31
- sharing?
32
- - source_sentence: Why AAP’s MLA Dinesh Mohaniya has been arrested?
33
  sentences:
34
- - What are your views on the latest sex scandal by AAP MLA Sandeep Kumar?
35
- - What is a dc current? What are some examples?
36
- - Why AAP’s MLA Dinesh Mohaniya has been arrested?
37
- - source_sentence: What is the difference between economic growth and economic development?
38
  sentences:
39
- - How cold can the Gobi Desert get, and how do its average temperatures compare
40
- to the ones in the Simpson Desert?
41
- - the difference between economic growth and economic development is What?
42
- - What is the difference between economic growth and economic development?
43
  pipeline_tag: sentence-similarity
44
  library_name: sentence-transformers
45
- metrics:
46
- - cosine_accuracy@1
47
- - cosine_accuracy@3
48
- - cosine_accuracy@5
49
- - cosine_accuracy@10
50
- - cosine_precision@1
51
- - cosine_precision@3
52
- - cosine_precision@5
53
- - cosine_precision@10
54
- - cosine_recall@1
55
- - cosine_recall@3
56
- - cosine_recall@5
57
- - cosine_recall@10
58
- - cosine_ndcg@10
59
- - cosine_mrr@10
60
- - cosine_map@100
61
- model-index:
62
- - name: SentenceTransformer based on sentence-transformers/all-MiniLM-L12-v2
63
- results:
64
- - task:
65
- type: information-retrieval
66
- name: Information Retrieval
67
- dataset:
68
- name: NanoMSMARCO
69
- type: NanoMSMARCO
70
- metrics:
71
- - type: cosine_accuracy@1
72
- value: 0.28
73
- name: Cosine Accuracy@1
74
- - type: cosine_accuracy@3
75
- value: 0.54
76
- name: Cosine Accuracy@3
77
- - type: cosine_accuracy@5
78
- value: 0.62
79
- name: Cosine Accuracy@5
80
- - type: cosine_accuracy@10
81
- value: 0.8
82
- name: Cosine Accuracy@10
83
- - type: cosine_precision@1
84
- value: 0.28
85
- name: Cosine Precision@1
86
- - type: cosine_precision@3
87
- value: 0.17999999999999997
88
- name: Cosine Precision@3
89
- - type: cosine_precision@5
90
- value: 0.124
91
- name: Cosine Precision@5
92
- - type: cosine_precision@10
93
- value: 0.08
94
- name: Cosine Precision@10
95
- - type: cosine_recall@1
96
- value: 0.28
97
- name: Cosine Recall@1
98
- - type: cosine_recall@3
99
- value: 0.54
100
- name: Cosine Recall@3
101
- - type: cosine_recall@5
102
- value: 0.62
103
- name: Cosine Recall@5
104
- - type: cosine_recall@10
105
- value: 0.8
106
- name: Cosine Recall@10
107
- - type: cosine_ndcg@10
108
- value: 0.5241911345526384
109
- name: Cosine Ndcg@10
110
- - type: cosine_mrr@10
111
- value: 0.43837301587301575
112
- name: Cosine Mrr@10
113
- - type: cosine_map@100
114
- value: 0.4480711307258977
115
- name: Cosine Map@100
116
- - task:
117
- type: information-retrieval
118
- name: Information Retrieval
119
- dataset:
120
- name: NanoNQ
121
- type: NanoNQ
122
- metrics:
123
- - type: cosine_accuracy@1
124
- value: 0.36
125
- name: Cosine Accuracy@1
126
- - type: cosine_accuracy@3
127
- value: 0.56
128
- name: Cosine Accuracy@3
129
- - type: cosine_accuracy@5
130
- value: 0.6
131
- name: Cosine Accuracy@5
132
- - type: cosine_accuracy@10
133
- value: 0.62
134
- name: Cosine Accuracy@10
135
- - type: cosine_precision@1
136
- value: 0.36
137
- name: Cosine Precision@1
138
- - type: cosine_precision@3
139
- value: 0.19333333333333333
140
- name: Cosine Precision@3
141
- - type: cosine_precision@5
142
- value: 0.128
143
- name: Cosine Precision@5
144
- - type: cosine_precision@10
145
- value: 0.066
146
- name: Cosine Precision@10
147
- - type: cosine_recall@1
148
- value: 0.35
149
- name: Cosine Recall@1
150
- - type: cosine_recall@3
151
- value: 0.53
152
- name: Cosine Recall@3
153
- - type: cosine_recall@5
154
- value: 0.58
155
- name: Cosine Recall@5
156
- - type: cosine_recall@10
157
- value: 0.6
158
- name: Cosine Recall@10
159
- - type: cosine_ndcg@10
160
- value: 0.490897686812855
161
- name: Cosine Ndcg@10
162
- - type: cosine_mrr@10
163
- value: 0.4625
164
- name: Cosine Mrr@10
165
- - type: cosine_map@100
166
- value: 0.46206363135240186
167
- name: Cosine Map@100
168
- - task:
169
- type: nano-beir
170
- name: Nano BEIR
171
- dataset:
172
- name: NanoBEIR mean
173
- type: NanoBEIR_mean
174
- metrics:
175
- - type: cosine_accuracy@1
176
- value: 0.32
177
- name: Cosine Accuracy@1
178
- - type: cosine_accuracy@3
179
- value: 0.55
180
- name: Cosine Accuracy@3
181
- - type: cosine_accuracy@5
182
- value: 0.61
183
- name: Cosine Accuracy@5
184
- - type: cosine_accuracy@10
185
- value: 0.71
186
- name: Cosine Accuracy@10
187
- - type: cosine_precision@1
188
- value: 0.32
189
- name: Cosine Precision@1
190
- - type: cosine_precision@3
191
- value: 0.18666666666666665
192
- name: Cosine Precision@3
193
- - type: cosine_precision@5
194
- value: 0.126
195
- name: Cosine Precision@5
196
- - type: cosine_precision@10
197
- value: 0.07300000000000001
198
- name: Cosine Precision@10
199
- - type: cosine_recall@1
200
- value: 0.315
201
- name: Cosine Recall@1
202
- - type: cosine_recall@3
203
- value: 0.535
204
- name: Cosine Recall@3
205
- - type: cosine_recall@5
206
- value: 0.6
207
- name: Cosine Recall@5
208
- - type: cosine_recall@10
209
- value: 0.7
210
- name: Cosine Recall@10
211
- - type: cosine_ndcg@10
212
- value: 0.5075444106827467
213
- name: Cosine Ndcg@10
214
- - type: cosine_mrr@10
215
- value: 0.4504365079365079
216
- name: Cosine Mrr@10
217
- - type: cosine_map@100
218
- value: 0.4550673810391498
219
- name: Cosine Map@100
220
  ---
221
 
222
- # SentenceTransformer based on sentence-transformers/all-MiniLM-L12-v2
223
 
224
- This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L12-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
225
 
226
  ## Model Details
227
 
228
  ### Model Description
229
  - **Model Type:** Sentence Transformer
230
- - **Base model:** [sentence-transformers/all-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L12-v2) <!-- at revision 936af83a2ecce5fe87a09109ff5cbcefe073173a -->
231
  - **Maximum Sequence Length:** 128 tokens
232
- - **Output Dimensionality:** 384 dimensions
233
  - **Similarity Function:** Cosine Similarity
234
  <!-- - **Training Dataset:** Unknown -->
235
  <!-- - **Language:** Unknown -->
@@ -246,8 +66,7 @@ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [s
246
  ```
247
  SentenceTransformer(
248
  (0): Transformer({'max_seq_length': 128, 'do_lower_case': False, 'architecture': 'BertModel'})
249
- (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
250
- (2): Normalize()
251
  )
252
  ```
253
 
@@ -266,23 +85,23 @@ Then you can load this model and run inference.
266
  from sentence_transformers import SentenceTransformer
267
 
268
  # Download from the 🤗 Hub
269
- model = SentenceTransformer("redis/model-b-structured")
270
  # Run inference
271
  sentences = [
272
- 'What is the difference between economic growth and economic development?',
273
- 'What is the difference between economic growth and economic development?',
274
- 'the difference between economic growth and economic development is What?',
275
  ]
276
  embeddings = model.encode(sentences)
277
  print(embeddings.shape)
278
- # [3, 384]
279
 
280
  # Get the similarity scores for the embeddings
281
  similarities = model.similarity(embeddings, embeddings)
282
  print(similarities)
283
- # tensor([[ 1.0000, 1.0000, -0.0663],
284
- # [ 1.0000, 1.0000, -0.0663],
285
- # [-0.0663, -0.0663, 1.0001]])
286
  ```
287
 
288
  <!--
@@ -309,65 +128,6 @@ You can finetune this model on your own dataset.
309
  *List how the model may foreseeably be misused and address what users ought not to do with the model.*
310
  -->
311
 
312
- ## Evaluation
313
-
314
- ### Metrics
315
-
316
- #### Information Retrieval
317
-
318
- * Datasets: `NanoMSMARCO` and `NanoNQ`
319
- * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
320
-
321
- | Metric | NanoMSMARCO | NanoNQ |
322
- |:--------------------|:------------|:-----------|
323
- | cosine_accuracy@1 | 0.28 | 0.36 |
324
- | cosine_accuracy@3 | 0.54 | 0.56 |
325
- | cosine_accuracy@5 | 0.62 | 0.6 |
326
- | cosine_accuracy@10 | 0.8 | 0.62 |
327
- | cosine_precision@1 | 0.28 | 0.36 |
328
- | cosine_precision@3 | 0.18 | 0.1933 |
329
- | cosine_precision@5 | 0.124 | 0.128 |
330
- | cosine_precision@10 | 0.08 | 0.066 |
331
- | cosine_recall@1 | 0.28 | 0.35 |
332
- | cosine_recall@3 | 0.54 | 0.53 |
333
- | cosine_recall@5 | 0.62 | 0.58 |
334
- | cosine_recall@10 | 0.8 | 0.6 |
335
- | **cosine_ndcg@10** | **0.5242** | **0.4909** |
336
- | cosine_mrr@10 | 0.4384 | 0.4625 |
337
- | cosine_map@100 | 0.4481 | 0.4621 |
338
-
339
- #### Nano BEIR
340
-
341
- * Dataset: `NanoBEIR_mean`
342
- * Evaluated with [<code>NanoBEIREvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.NanoBEIREvaluator) with these parameters:
343
- ```json
344
- {
345
- "dataset_names": [
346
- "msmarco",
347
- "nq"
348
- ],
349
- "dataset_id": "lightonai/NanoBEIR-en"
350
- }
351
- ```
352
-
353
- | Metric | Value |
354
- |:--------------------|:-----------|
355
- | cosine_accuracy@1 | 0.32 |
356
- | cosine_accuracy@3 | 0.55 |
357
- | cosine_accuracy@5 | 0.61 |
358
- | cosine_accuracy@10 | 0.71 |
359
- | cosine_precision@1 | 0.32 |
360
- | cosine_precision@3 | 0.1867 |
361
- | cosine_precision@5 | 0.126 |
362
- | cosine_precision@10 | 0.073 |
363
- | cosine_recall@1 | 0.315 |
364
- | cosine_recall@3 | 0.535 |
365
- | cosine_recall@5 | 0.6 |
366
- | cosine_recall@10 | 0.7 |
367
- | **cosine_ndcg@10** | **0.5075** |
368
- | cosine_mrr@10 | 0.4504 |
369
- | cosine_map@100 | 0.4551 |
370
-
371
  <!--
372
  ## Bias, Risks and Limitations
373
 
@@ -386,49 +146,23 @@ You can finetune this model on your own dataset.
386
 
387
  #### Unnamed Dataset
388
 
389
- * Size: 713,743 training samples
390
- * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
391
- * Approximate statistics based on the first 1000 samples:
392
- | | anchor | positive | negative |
393
- |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
394
- | type | string | string | string |
395
- | details | <ul><li>min: 6 tokens</li><li>mean: 16.07 tokens</li><li>max: 53 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.03 tokens</li><li>max: 53 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.81 tokens</li><li>max: 58 tokens</li></ul> |
396
- * Samples:
397
- | anchor | positive | negative |
398
- |:-------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------|
399
- | <code>Which one is better Linux OS? Ubuntu or Mint?</code> | <code>Why do you use Linux Mint?</code> | <code>Which one is not better Linux OS ? Ubuntu or Mint ?</code> |
400
- | <code>What is flow?</code> | <code>What is flow?</code> | <code>What are flow lines?</code> |
401
- | <code>How is Trump planning to get Mexico to pay for his supposed wall?</code> | <code>How is it possible for Donald Trump to force Mexico to pay for the wall?</code> | <code>Why do we connect the positive terminal before the negative terminal to ground in a vehicle battery?</code> |
402
- * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
403
- ```json
404
- {
405
- "scale": 7.0,
406
- "similarity_fct": "cos_sim",
407
- "gather_across_devices": false
408
- }
409
- ```
410
-
411
- ### Evaluation Dataset
412
-
413
- #### Unnamed Dataset
414
-
415
- * Size: 40,000 evaluation samples
416
- * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
417
  * Approximate statistics based on the first 1000 samples:
418
- | | anchor | positive | negative |
419
  |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
420
  | type | string | string | string |
421
- | details | <ul><li>min: 6 tokens</li><li>mean: 15.52 tokens</li><li>max: 74 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 15.51 tokens</li><li>max: 74 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.79 tokens</li><li>max: 69 tokens</li></ul> |
422
  * Samples:
423
- | anchor | positive | negative |
424
- |:-------------------------------------------------------------------------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------|
425
- | <code>Why are all my questions on Quora marked needing improvement?</code> | <code>Why are all my questions immediately being marked as needing improvement?</code> | <code>For a post-graduate student in IIT, is it allowed to take an external scholarship as a top-up to his/her MHRD assistantship?</code> |
426
- | <code>Can blue butter fly needle with vaccum tube be reused? Is it HIV risk? . Heard the needle is too small to be reused . Had blood draw at clinic?</code> | <code>Can blue butter fly needle with vaccum tube be reused? Is it HIV risk? . Heard the needle is too small to be reused . Had blood draw at clinic?</code> | <code>Can blue butter fly needle with vaccum tube be reused not ? Is it HIV risk ? . Heard the needle is too small to be reused . Had blood draw at clinic ?</code> |
427
- | <code>Why do people still believe the world is flat?</code> | <code>Why are there still people who believe the world is flat?</code> | <code>I'm not able to buy Udemy course .it is not accepting mine and my friends debit card.my card can be used for Flipkart .how to purchase now?</code> |
428
  * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
429
  ```json
430
  {
431
- "scale": 7.0,
432
  "similarity_fct": "cos_sim",
433
  "gather_across_devices": false
434
  }
@@ -437,49 +171,36 @@ You can finetune this model on your own dataset.
437
  ### Training Hyperparameters
438
  #### Non-Default Hyperparameters
439
 
440
- - `eval_strategy`: steps
441
- - `per_device_train_batch_size`: 128
442
- - `per_device_eval_batch_size`: 128
443
- - `learning_rate`: 2e-05
444
- - `weight_decay`: 0.0001
445
- - `max_steps`: 5000
446
- - `warmup_ratio`: 0.1
447
  - `fp16`: True
448
- - `dataloader_drop_last`: True
449
- - `dataloader_num_workers`: 1
450
- - `dataloader_prefetch_factor`: 1
451
- - `load_best_model_at_end`: True
452
- - `optim`: adamw_torch
453
- - `ddp_find_unused_parameters`: False
454
- - `push_to_hub`: True
455
- - `hub_model_id`: redis/model-b-structured
456
- - `eval_on_start`: True
457
 
458
  #### All Hyperparameters
459
  <details><summary>Click to expand</summary>
460
 
461
  - `overwrite_output_dir`: False
462
  - `do_predict`: False
463
- - `eval_strategy`: steps
464
  - `prediction_loss_only`: True
465
- - `per_device_train_batch_size`: 128
466
- - `per_device_eval_batch_size`: 128
467
  - `per_gpu_train_batch_size`: None
468
  - `per_gpu_eval_batch_size`: None
469
  - `gradient_accumulation_steps`: 1
470
  - `eval_accumulation_steps`: None
471
  - `torch_empty_cache_steps`: None
472
- - `learning_rate`: 2e-05
473
- - `weight_decay`: 0.0001
474
  - `adam_beta1`: 0.9
475
  - `adam_beta2`: 0.999
476
  - `adam_epsilon`: 1e-08
477
- - `max_grad_norm`: 1.0
478
- - `num_train_epochs`: 3.0
479
- - `max_steps`: 5000
480
  - `lr_scheduler_type`: linear
481
  - `lr_scheduler_kwargs`: {}
482
- - `warmup_ratio`: 0.1
483
  - `warmup_steps`: 0
484
  - `log_level`: passive
485
  - `log_level_replica`: warning
@@ -507,14 +228,14 @@ You can finetune this model on your own dataset.
507
  - `tpu_num_cores`: None
508
  - `tpu_metrics_debug`: False
509
  - `debug`: []
510
- - `dataloader_drop_last`: True
511
- - `dataloader_num_workers`: 1
512
- - `dataloader_prefetch_factor`: 1
513
  - `past_index`: -1
514
  - `disable_tqdm`: False
515
  - `remove_unused_columns`: True
516
  - `label_names`: None
517
- - `load_best_model_at_end`: True
518
  - `ignore_data_skip`: False
519
  - `fsdp`: []
520
  - `fsdp_min_num_params`: 0
@@ -524,23 +245,23 @@ You can finetune this model on your own dataset.
524
  - `parallelism_config`: None
525
  - `deepspeed`: None
526
  - `label_smoothing_factor`: 0.0
527
- - `optim`: adamw_torch
528
  - `optim_args`: None
529
  - `adafactor`: False
530
  - `group_by_length`: False
531
  - `length_column_name`: length
532
  - `project`: huggingface
533
  - `trackio_space_id`: trackio
534
- - `ddp_find_unused_parameters`: False
535
  - `ddp_bucket_cap_mb`: None
536
  - `ddp_broadcast_buffers`: False
537
  - `dataloader_pin_memory`: True
538
  - `dataloader_persistent_workers`: False
539
  - `skip_memory_metrics`: True
540
  - `use_legacy_prediction_loop`: False
541
- - `push_to_hub`: True
542
  - `resume_from_checkpoint`: None
543
- - `hub_model_id`: redis/model-b-structured
544
  - `hub_strategy`: every_save
545
  - `hub_private_repo`: None
546
  - `hub_always_push`: False
@@ -567,43 +288,31 @@ You can finetune this model on your own dataset.
567
  - `neftune_noise_alpha`: None
568
  - `optim_target_modules`: None
569
  - `batch_eval_metrics`: False
570
- - `eval_on_start`: True
571
  - `use_liger_kernel`: False
572
  - `liger_kernel_config`: None
573
  - `eval_use_gather_object`: False
574
  - `average_tokens_across_devices`: True
575
  - `prompts`: None
576
  - `batch_sampler`: batch_sampler
577
- - `multi_dataset_batch_sampler`: proportional
578
  - `router_mapping`: {}
579
  - `learning_rate_mapping`: {}
580
 
581
  </details>
582
 
583
  ### Training Logs
584
- | Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
585
- |:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------:|:----------------------------:|
586
- | 0 | 0 | - | 0.8240 | 0.5887 | 0.5786 | 0.5836 |
587
- | 0.0448 | 250 | 0.7214 | 0.4428 | 0.5512 | 0.5408 | 0.5460 |
588
- | 0.0897 | 500 | 0.5356 | 0.4086 | 0.5509 | 0.5218 | 0.5363 |
589
- | 0.1345 | 750 | 0.5038 | 0.3926 | 0.5310 | 0.5203 | 0.5257 |
590
- | 0.1793 | 1000 | 0.4887 | 0.3845 | 0.5389 | 0.5128 | 0.5259 |
591
- | 0.2242 | 1250 | 0.4756 | 0.3780 | 0.5481 | 0.4881 | 0.5181 |
592
- | 0.2690 | 1500 | 0.4636 | 0.3742 | 0.5361 | 0.5060 | 0.5210 |
593
- | 0.3138 | 1750 | 0.4572 | 0.3698 | 0.5311 | 0.5125 | 0.5218 |
594
- | 0.3587 | 2000 | 0.4529 | 0.3666 | 0.5290 | 0.5177 | 0.5233 |
595
- | 0.4035 | 2250 | 0.446 | 0.3638 | 0.5260 | 0.5117 | 0.5189 |
596
- | 0.4484 | 2500 | 0.4434 | 0.3620 | 0.5121 | 0.5105 | 0.5113 |
597
- | 0.4932 | 2750 | 0.4374 | 0.3597 | 0.5333 | 0.5026 | 0.5179 |
598
- | 0.5380 | 3000 | 0.4351 | 0.3577 | 0.5156 | 0.4949 | 0.5052 |
599
- | 0.5829 | 3250 | 0.433 | 0.3561 | 0.5153 | 0.4934 | 0.5043 |
600
- | 0.6277 | 3500 | 0.4339 | 0.3551 | 0.5148 | 0.5109 | 0.5129 |
601
- | 0.6725 | 3750 | 0.4299 | 0.3535 | 0.5061 | 0.4921 | 0.4991 |
602
- | 0.7174 | 4000 | 0.4283 | 0.3535 | 0.5239 | 0.4886 | 0.5063 |
603
- | 0.7622 | 4250 | 0.4276 | 0.3522 | 0.5208 | 0.4890 | 0.5049 |
604
- | 0.8070 | 4500 | 0.4257 | 0.3523 | 0.5239 | 0.4864 | 0.5051 |
605
- | 0.8519 | 4750 | 0.4284 | 0.3520 | 0.5246 | 0.4883 | 0.5065 |
606
- | 0.8967 | 5000 | 0.4262 | 0.3516 | 0.5242 | 0.4909 | 0.5075 |
607
 
608
 
609
  ### Framework Versions
@@ -612,7 +321,7 @@ You can finetune this model on your own dataset.
612
  - Transformers: 4.57.3
613
  - PyTorch: 2.9.1+cu128
614
  - Accelerate: 1.12.0
615
- - Datasets: 2.21.0
616
  - Tokenizers: 0.22.1
617
 
618
  ## Citation
 
5
  - feature-extraction
6
  - dense
7
  - generated_from_trainer
8
+ - dataset_size:100000
9
  - loss:MultipleNegativesRankingLoss
10
+ base_model: prajjwal1/bert-small
11
  widget:
12
+ - source_sentence: How do I calculate IQ?
13
  sentences:
14
+ - What is the easiest way to know my IQ?
15
+ - How do I calculate not IQ ?
16
+ - What are some creative and innovative business ideas with less investment in India?
17
+ - source_sentence: How can I learn martial arts in my home?
 
18
  sentences:
19
+ - How can I learn martial arts by myself?
20
+ - What are the advantages and disadvantages of investing in gold?
21
+ - Can people see that I have looked at their pictures on instagram if I am not following
22
+ them?
23
+ - source_sentence: When Enterprise picks you up do you have to take them back?
24
  sentences:
25
+ - Are there any software Training institute in Tuticorin?
26
+ - When Enterprise picks you up do you have to take them back?
27
+ - When Enterprise picks you up do them have to take youback?
28
+ - source_sentence: What are some non-capital goods?
 
 
 
29
  sentences:
30
+ - What are capital goods?
31
+ - How is the value of [math]\pi[/math] calculated?
32
+ - What are some non-capital goods?
33
+ - source_sentence: What is the QuickBooks technical support phone number in New York?
34
  sentences:
35
+ - What caused the Great Depression?
36
+ - Can I apply for PR in Canada?
37
+ - Which is the best QuickBooks Hosting Support Number in New York?
 
38
  pipeline_tag: sentence-similarity
39
  library_name: sentence-transformers
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
  ---
41
 
42
+ # SentenceTransformer based on prajjwal1/bert-small
43
 
44
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [prajjwal1/bert-small](https://huggingface.co/prajjwal1/bert-small). It maps sentences & paragraphs to a 512-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
45
 
46
  ## Model Details
47
 
48
  ### Model Description
49
  - **Model Type:** Sentence Transformer
50
+ - **Base model:** [prajjwal1/bert-small](https://huggingface.co/prajjwal1/bert-small) <!-- at revision 0ec5f86f27c1a77d704439db5e01c307ea11b9d4 -->
51
  - **Maximum Sequence Length:** 128 tokens
52
+ - **Output Dimensionality:** 512 dimensions
53
  - **Similarity Function:** Cosine Similarity
54
  <!-- - **Training Dataset:** Unknown -->
55
  <!-- - **Language:** Unknown -->
 
66
  ```
67
  SentenceTransformer(
68
  (0): Transformer({'max_seq_length': 128, 'do_lower_case': False, 'architecture': 'BertModel'})
69
+ (1): Pooling({'word_embedding_dimension': 512, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
 
70
  )
71
  ```
72
 
 
85
  from sentence_transformers import SentenceTransformer
86
 
87
  # Download from the 🤗 Hub
88
+ model = SentenceTransformer("sentence_transformers_model_id")
89
  # Run inference
90
  sentences = [
91
+ 'What is the QuickBooks technical support phone number in New York?',
92
+ 'Which is the best QuickBooks Hosting Support Number in New York?',
93
+ 'Can I apply for PR in Canada?',
94
  ]
95
  embeddings = model.encode(sentences)
96
  print(embeddings.shape)
97
+ # [3, 512]
98
 
99
  # Get the similarity scores for the embeddings
100
  similarities = model.similarity(embeddings, embeddings)
101
  print(similarities)
102
+ # tensor([[1.0000, 0.8563, 0.0594],
103
+ # [0.8563, 1.0000, 0.1245],
104
+ # [0.0594, 0.1245, 1.0000]])
105
  ```
106
 
107
  <!--
 
128
  *List how the model may foreseeably be misused and address what users ought not to do with the model.*
129
  -->
130
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
131
  <!--
132
  ## Bias, Risks and Limitations
133
 
 
146
 
147
  #### Unnamed Dataset
148
 
149
+ * Size: 100,000 training samples
150
+ * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>sentence_2</code>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
151
  * Approximate statistics based on the first 1000 samples:
152
+ | | sentence_0 | sentence_1 | sentence_2 |
153
  |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
154
  | type | string | string | string |
155
+ | details | <ul><li>min: 6 tokens</li><li>mean: 15.79 tokens</li><li>max: 66 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 15.68 tokens</li><li>max: 66 tokens</li></ul> | <ul><li>min: 7 tokens</li><li>mean: 16.37 tokens</li><li>max: 67 tokens</li></ul> |
156
  * Samples:
157
+ | sentence_0 | sentence_1 | sentence_2 |
158
+ |:-----------------------------------------------------------------|:-----------------------------------------------------------------|:----------------------------------------------------------------------------------|
159
+ | <code>Is masturbating bad for boys?</code> | <code>Is masturbating bad for boys?</code> | <code>How harmful or unhealthy is masturbation?</code> |
160
+ | <code>Does a train engine move in reverse?</code> | <code>Does a train engine move in reverse?</code> | <code>Time moves forward, not in reverse. Doesn't that make time a vector?</code> |
161
+ | <code>What is the most badass thing anyone has ever done?</code> | <code>What is the most badass thing anyone has ever done?</code> | <code>anyone is the most badass thing Whathas ever done?</code> |
162
  * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
163
  ```json
164
  {
165
+ "scale": 20.0,
166
  "similarity_fct": "cos_sim",
167
  "gather_across_devices": false
168
  }
 
171
  ### Training Hyperparameters
172
  #### Non-Default Hyperparameters
173
 
174
+ - `per_device_train_batch_size`: 64
175
+ - `per_device_eval_batch_size`: 64
 
 
 
 
 
176
  - `fp16`: True
177
+ - `multi_dataset_batch_sampler`: round_robin
 
 
 
 
 
 
 
 
178
 
179
  #### All Hyperparameters
180
  <details><summary>Click to expand</summary>
181
 
182
  - `overwrite_output_dir`: False
183
  - `do_predict`: False
184
+ - `eval_strategy`: no
185
  - `prediction_loss_only`: True
186
+ - `per_device_train_batch_size`: 64
187
+ - `per_device_eval_batch_size`: 64
188
  - `per_gpu_train_batch_size`: None
189
  - `per_gpu_eval_batch_size`: None
190
  - `gradient_accumulation_steps`: 1
191
  - `eval_accumulation_steps`: None
192
  - `torch_empty_cache_steps`: None
193
+ - `learning_rate`: 5e-05
194
+ - `weight_decay`: 0.0
195
  - `adam_beta1`: 0.9
196
  - `adam_beta2`: 0.999
197
  - `adam_epsilon`: 1e-08
198
+ - `max_grad_norm`: 1
199
+ - `num_train_epochs`: 3
200
+ - `max_steps`: -1
201
  - `lr_scheduler_type`: linear
202
  - `lr_scheduler_kwargs`: {}
203
+ - `warmup_ratio`: 0.0
204
  - `warmup_steps`: 0
205
  - `log_level`: passive
206
  - `log_level_replica`: warning
 
228
  - `tpu_num_cores`: None
229
  - `tpu_metrics_debug`: False
230
  - `debug`: []
231
+ - `dataloader_drop_last`: False
232
+ - `dataloader_num_workers`: 0
233
+ - `dataloader_prefetch_factor`: None
234
  - `past_index`: -1
235
  - `disable_tqdm`: False
236
  - `remove_unused_columns`: True
237
  - `label_names`: None
238
+ - `load_best_model_at_end`: False
239
  - `ignore_data_skip`: False
240
  - `fsdp`: []
241
  - `fsdp_min_num_params`: 0
 
245
  - `parallelism_config`: None
246
  - `deepspeed`: None
247
  - `label_smoothing_factor`: 0.0
248
+ - `optim`: adamw_torch_fused
249
  - `optim_args`: None
250
  - `adafactor`: False
251
  - `group_by_length`: False
252
  - `length_column_name`: length
253
  - `project`: huggingface
254
  - `trackio_space_id`: trackio
255
+ - `ddp_find_unused_parameters`: None
256
  - `ddp_bucket_cap_mb`: None
257
  - `ddp_broadcast_buffers`: False
258
  - `dataloader_pin_memory`: True
259
  - `dataloader_persistent_workers`: False
260
  - `skip_memory_metrics`: True
261
  - `use_legacy_prediction_loop`: False
262
+ - `push_to_hub`: False
263
  - `resume_from_checkpoint`: None
264
+ - `hub_model_id`: None
265
  - `hub_strategy`: every_save
266
  - `hub_private_repo`: None
267
  - `hub_always_push`: False
 
288
  - `neftune_noise_alpha`: None
289
  - `optim_target_modules`: None
290
  - `batch_eval_metrics`: False
291
+ - `eval_on_start`: False
292
  - `use_liger_kernel`: False
293
  - `liger_kernel_config`: None
294
  - `eval_use_gather_object`: False
295
  - `average_tokens_across_devices`: True
296
  - `prompts`: None
297
  - `batch_sampler`: batch_sampler
298
+ - `multi_dataset_batch_sampler`: round_robin
299
  - `router_mapping`: {}
300
  - `learning_rate_mapping`: {}
301
 
302
  </details>
303
 
304
  ### Training Logs
305
+ | Epoch | Step | Training Loss |
306
+ |:------:|:----:|:-------------:|
307
+ | 0.3199 | 500 | 0.4294 |
308
+ | 0.6398 | 1000 | 0.1268 |
309
+ | 0.9597 | 1500 | 0.1 |
310
+ | 1.2796 | 2000 | 0.0792 |
311
+ | 1.5995 | 2500 | 0.0706 |
312
+ | 1.9194 | 3000 | 0.0687 |
313
+ | 2.2393 | 3500 | 0.0584 |
314
+ | 2.5592 | 4000 | 0.057 |
315
+ | 2.8791 | 4500 | 0.0581 |
 
 
 
 
 
 
 
 
 
 
 
 
316
 
317
 
318
  ### Framework Versions
 
321
  - Transformers: 4.57.3
322
  - PyTorch: 2.9.1+cu128
323
  - Accelerate: 1.12.0
324
+ - Datasets: 4.4.2
325
  - Tokenizers: 0.22.1
326
 
327
  ## Citation
config.json CHANGED
@@ -5,7 +5,6 @@
5
  "attention_probs_dropout_prob": 0.1,
6
  "classifier_dropout": null,
7
  "dtype": "float32",
8
- "gradient_checkpointing": false,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 384,
 
5
  "attention_probs_dropout_prob": 0.1,
6
  "classifier_dropout": null,
7
  "dtype": "float32",
 
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.1,
10
  "hidden_size": 384,
config_sentence_transformers.json CHANGED
@@ -1,10 +1,10 @@
1
  {
 
2
  "__version__": {
3
  "sentence_transformers": "5.2.0",
4
  "transformers": "4.57.3",
5
  "pytorch": "2.9.1+cu128"
6
  },
7
- "model_type": "SentenceTransformer",
8
  "prompts": {
9
  "query": "",
10
  "document": ""
 
1
  {
2
+ "model_type": "SentenceTransformer",
3
  "__version__": {
4
  "sentence_transformers": "5.2.0",
5
  "transformers": "4.57.3",
6
  "pytorch": "2.9.1+cu128"
7
  },
 
8
  "prompts": {
9
  "query": "",
10
  "document": ""
eval/Information-Retrieval_evaluation_NanoMSMARCO_results.csv CHANGED
@@ -41,3 +41,24 @@ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accurac
41
  0.8070301291248206,4500,0.28,0.54,0.62,0.8,0.28,0.28,0.17999999999999997,0.54,0.124,0.62,0.08,0.8,0.43813492063492065,0.5239338733662281,0.44783147531841366
42
  0.8518651362984218,4750,0.28,0.54,0.62,0.8,0.28,0.28,0.17999999999999997,0.54,0.124,0.62,0.08,0.8,0.43884920634920627,0.5246486116281323,0.4486126806792123
43
  0.896700143472023,5000,0.28,0.54,0.62,0.8,0.28,0.28,0.17999999999999997,0.54,0.124,0.62,0.08,0.8,0.43837301587301575,0.5241911345526384,0.4480711307258977
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
41
  0.8070301291248206,4500,0.28,0.54,0.62,0.8,0.28,0.28,0.17999999999999997,0.54,0.124,0.62,0.08,0.8,0.43813492063492065,0.5239338733662281,0.44783147531841366
42
  0.8518651362984218,4750,0.28,0.54,0.62,0.8,0.28,0.28,0.17999999999999997,0.54,0.124,0.62,0.08,0.8,0.43884920634920627,0.5246486116281323,0.4486126806792123
43
  0.896700143472023,5000,0.28,0.54,0.62,0.8,0.28,0.28,0.17999999999999997,0.54,0.124,0.62,0.08,0.8,0.43837301587301575,0.5241911345526384,0.4480711307258977
44
+ 0,0,0.44,0.66,0.72,0.8,0.44,0.44,0.22,0.66,0.14400000000000002,0.72,0.08,0.8,0.5702460317460318,0.6259279298239366,0.582094782035277
45
+ 0.04483500717360115,250,0.38,0.56,0.7,0.78,0.38,0.38,0.18666666666666668,0.56,0.14,0.7,0.07800000000000001,0.78,0.5063571428571428,0.5723432352498731,0.5154920420839786
46
+ 0.0896700143472023,500,0.44,0.68,0.72,0.76,0.44,0.44,0.22666666666666668,0.68,0.14400000000000002,0.72,0.07600000000000001,0.76,0.5658333333333334,0.6142278592436732,0.5758522013148719
47
+ 0.13450502152080343,750,0.38,0.64,0.7,0.76,0.38,0.38,0.21333333333333332,0.64,0.14,0.7,0.07600000000000001,0.76,0.5191904761904762,0.5785699532446995,0.5305262079655163
48
+ 0.1793400286944046,1000,0.34,0.64,0.7,0.76,0.34,0.34,0.21333333333333332,0.64,0.14,0.7,0.07600000000000001,0.76,0.4943888888888889,0.5597276353562957,0.5057309981684981
49
+ 0.22417503586800575,1250,0.3,0.64,0.66,0.78,0.3,0.3,0.21333333333333332,0.64,0.132,0.66,0.07800000000000001,0.78,0.47455555555555556,0.5489922962110942,0.484680435019931
50
+ 0.26901004304160686,1500,0.3,0.6,0.68,0.74,0.3,0.3,0.19999999999999996,0.6,0.136,0.68,0.07400000000000001,0.74,0.45966666666666667,0.5286298641401985,0.4732919814942032
51
+ 0.31384505021520803,1750,0.32,0.62,0.68,0.74,0.32,0.32,0.20666666666666667,0.62,0.136,0.68,0.07400000000000001,0.74,0.4728333333333333,0.5386226869314987,0.4861326654307037
52
+ 0.3586800573888092,2000,0.28,0.58,0.68,0.74,0.28,0.28,0.19333333333333333,0.58,0.136,0.68,0.07400000000000001,0.74,0.44511111111111107,0.5175424750735735,0.45847456493883976
53
+ 0.4035150645624103,2250,0.28,0.58,0.62,0.74,0.28,0.28,0.19333333333333333,0.58,0.124,0.62,0.07400000000000001,0.74,0.4369444444444444,0.5105173465545532,0.4502006020766632
54
+ 0.4483500717360115,2500,0.3,0.58,0.66,0.74,0.3,0.3,0.19333333333333333,0.58,0.132,0.66,0.07400000000000001,0.74,0.45011111111111113,0.5208158975114078,0.46310611993959677
55
+ 0.4931850789096126,2750,0.3,0.58,0.64,0.74,0.3,0.3,0.19333333333333333,0.58,0.128,0.64,0.07400000000000001,0.74,0.44733333333333336,0.5182229662632195,0.4604496242018111
56
+ 0.5380200860832137,3000,0.28,0.58,0.64,0.74,0.28,0.28,0.19333333333333333,0.58,0.128,0.64,0.07400000000000001,0.74,0.4392460317460317,0.5126528007272476,0.45193414032521817
57
+ 0.582855093256815,3250,0.28,0.58,0.64,0.74,0.28,0.28,0.19333333333333333,0.58,0.128,0.64,0.07400000000000001,0.74,0.43777777777777777,0.5111918877674148,0.45068354413480427
58
+ 0.6276901004304161,3500,0.28,0.58,0.64,0.76,0.28,0.28,0.19333333333333333,0.58,0.128,0.64,0.07600000000000001,0.76,0.43977777777777777,0.5169731842937726,0.45163005170491566
59
+ 0.6725251076040172,3750,0.28,0.58,0.64,0.74,0.28,0.28,0.19333333333333333,0.58,0.128,0.64,0.07400000000000001,0.74,0.4398888888888889,0.5126947044351451,0.45337430630173947
60
+ 0.7173601147776184,4000,0.28,0.58,0.64,0.74,0.28,0.28,0.19333333333333333,0.58,0.128,0.64,0.07400000000000001,0.74,0.44038888888888894,0.5130431784425653,0.4537472527472527
61
+ 0.7621951219512195,4250,0.28,0.58,0.64,0.72,0.28,0.28,0.19333333333333333,0.58,0.128,0.64,0.07200000000000001,0.72,0.4388888888888889,0.5077898829255643,0.453482045393225
62
+ 0.8070301291248206,4500,0.28,0.58,0.64,0.72,0.28,0.28,0.19333333333333333,0.58,0.128,0.64,0.07200000000000001,0.72,0.4383888888888889,0.5072618819162075,0.45318728603379055
63
+ 0.8518651362984218,4750,0.28,0.58,0.64,0.72,0.28,0.28,0.19333333333333333,0.58,0.128,0.64,0.07200000000000001,0.72,0.4353333333333333,0.5049319844672133,0.4499328092290969
64
+ 0.896700143472023,5000,0.28,0.58,0.64,0.72,0.28,0.28,0.19333333333333333,0.58,0.128,0.64,0.07200000000000001,0.72,0.4386111111111111,0.5075011853031293,0.4533366047009664
eval/Information-Retrieval_evaluation_NanoNQ_results.csv CHANGED
@@ -41,3 +41,24 @@ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accurac
41
  0.8070301291248206,4500,0.36,0.54,0.6,0.62,0.36,0.35,0.18666666666666665,0.51,0.128,0.58,0.066,0.6,0.457,0.48636462189353685,0.4563546379037804
42
  0.8518651362984218,4750,0.36,0.56,0.6,0.62,0.36,0.35,0.19333333333333333,0.53,0.128,0.58,0.066,0.6,0.4591666666666667,0.4882790917414258,0.4588111953162766
43
  0.896700143472023,5000,0.36,0.56,0.6,0.62,0.36,0.35,0.19333333333333333,0.53,0.128,0.58,0.066,0.6,0.4625,0.490897686812855,0.46206363135240186
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
41
  0.8070301291248206,4500,0.36,0.54,0.6,0.62,0.36,0.35,0.18666666666666665,0.51,0.128,0.58,0.066,0.6,0.457,0.48636462189353685,0.4563546379037804
42
  0.8518651362984218,4750,0.36,0.56,0.6,0.62,0.36,0.35,0.19333333333333333,0.53,0.128,0.58,0.066,0.6,0.4591666666666667,0.4882790917414258,0.4588111953162766
43
  0.896700143472023,5000,0.36,0.56,0.6,0.62,0.36,0.35,0.19333333333333333,0.53,0.128,0.58,0.066,0.6,0.4625,0.490897686812855,0.46206363135240186
44
+ 0,0,0.58,0.68,0.74,0.78,0.58,0.55,0.22666666666666666,0.64,0.15600000000000003,0.7,0.086,0.75,0.6453888888888889,0.6582829486426273,0.6272793246081896
45
+ 0.04483500717360115,250,0.36,0.6,0.7,0.74,0.36,0.33,0.2,0.56,0.14400000000000002,0.67,0.07600000000000001,0.7,0.4950238095238095,0.5298490669233439,0.4739194711924051
46
+ 0.0896700143472023,500,0.38,0.6,0.66,0.72,0.38,0.35,0.20666666666666664,0.58,0.136,0.63,0.07400000000000001,0.68,0.4970238095238095,0.5301322778895181,0.4849673712311871
47
+ 0.13450502152080343,750,0.4,0.56,0.62,0.72,0.4,0.37,0.19333333333333333,0.54,0.128,0.59,0.07600000000000001,0.7,0.49674603174603177,0.5354593371355858,0.4859822715906391
48
+ 0.1793400286944046,1000,0.38,0.54,0.66,0.72,0.38,0.35,0.18666666666666665,0.51,0.14,0.64,0.07600000000000001,0.7,0.48783333333333334,0.5290925388478536,0.4760392045402088
49
+ 0.22417503586800575,1250,0.38,0.54,0.64,0.7,0.38,0.35,0.18666666666666665,0.51,0.136,0.62,0.07400000000000001,0.68,0.48983333333333334,0.5254672343498678,0.4772819462279719
50
+ 0.26901004304160686,1500,0.4,0.56,0.62,0.7,0.4,0.37,0.19333333333333333,0.53,0.132,0.6,0.07400000000000001,0.68,0.4997222222222223,0.5330560600816254,0.4884345453661391
51
+ 0.31384505021520803,1750,0.38,0.56,0.6,0.72,0.38,0.36,0.19333333333333333,0.53,0.128,0.58,0.07600000000000001,0.69,0.4893015873015873,0.5297468208024404,0.48082567230874385
52
+ 0.3586800573888092,2000,0.36,0.54,0.62,0.7,0.36,0.33,0.18666666666666665,0.51,0.132,0.6,0.07400000000000001,0.68,0.4766666666666666,0.5155145399346358,0.4650632990555934
53
+ 0.4035150645624103,2250,0.4,0.54,0.62,0.7,0.4,0.37,0.18666666666666665,0.51,0.132,0.6,0.07400000000000001,0.68,0.4954444444444445,0.5291615713880786,0.48395073767717256
54
+ 0.4483500717360115,2500,0.34,0.54,0.6,0.68,0.34,0.32,0.18,0.5,0.128,0.58,0.07200000000000001,0.66,0.45960317460317457,0.4985929658046555,0.44934741543505036
55
+ 0.4931850789096126,2750,0.34,0.54,0.6,0.68,0.34,0.32,0.18666666666666665,0.51,0.128,0.58,0.07200000000000001,0.66,0.4625555555555556,0.5016337772874723,0.4528968956200227
56
+ 0.5380200860832137,3000,0.36,0.54,0.6,0.66,0.36,0.33,0.18666666666666665,0.51,0.128,0.58,0.07,0.64,0.4662222222222222,0.49688329838352674,0.4538908182490129
57
+ 0.582855093256815,3250,0.34,0.54,0.6,0.66,0.34,0.31,0.18,0.5,0.128,0.58,0.07,0.64,0.45388888888888884,0.48690966392410095,0.4390928696092712
58
+ 0.6276901004304161,3500,0.36,0.56,0.62,0.66,0.36,0.33,0.19333333333333333,0.53,0.132,0.6,0.07,0.64,0.4698333333333333,0.5000478522638012,0.4567747023729727
59
+ 0.6725251076040172,3750,0.34,0.54,0.6,0.66,0.34,0.31,0.18666666666666665,0.51,0.128,0.58,0.07,0.64,0.45666666666666667,0.489852219887722,0.4434927668385896
60
+ 0.7173601147776184,4000,0.34,0.54,0.6,0.66,0.34,0.31,0.18666666666666665,0.51,0.128,0.58,0.07,0.64,0.45305555555555554,0.48694492719385785,0.4402135952187105
61
+ 0.7621951219512195,4250,0.34,0.54,0.6,0.66,0.34,0.31,0.18666666666666665,0.51,0.128,0.58,0.07,0.64,0.4553888888888889,0.48868704724851,0.44262350670371914
62
+ 0.8070301291248206,4500,0.34,0.54,0.6,0.66,0.34,0.31,0.18666666666666665,0.51,0.128,0.58,0.07,0.64,0.4526904761904762,0.4867145189304679,0.4396658900478081
63
+ 0.8518651362984218,4750,0.32,0.54,0.62,0.66,0.32,0.3,0.18666666666666665,0.51,0.132,0.6,0.07,0.64,0.4471666666666667,0.48748319998633916,0.44239180468060907
64
+ 0.896700143472023,5000,0.32,0.54,0.6,0.66,0.32,0.3,0.18666666666666665,0.51,0.128,0.58,0.07,0.64,0.4465,0.48687028758380874,0.44172587957864395
eval/NanoBEIR_evaluation_mean_results.csv CHANGED
@@ -41,3 +41,24 @@ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accurac
41
  0.8070301291248206,4500,0.32,0.54,0.61,0.71,0.32,0.315,0.1833333333333333,0.525,0.126,0.6,0.07300000000000001,0.7,0.4475674603174603,0.5051492476298824,0.452093056611097
42
  0.8518651362984218,4750,0.32,0.55,0.61,0.71,0.32,0.315,0.18666666666666665,0.535,0.126,0.6,0.07300000000000001,0.7,0.44900793650793647,0.506463851684779,0.45371193799774445
43
  0.896700143472023,5000,0.32,0.55,0.61,0.71,0.32,0.315,0.18666666666666665,0.535,0.126,0.6,0.07300000000000001,0.7,0.4504365079365079,0.5075444106827467,0.4550673810391498
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
41
  0.8070301291248206,4500,0.32,0.54,0.61,0.71,0.32,0.315,0.1833333333333333,0.525,0.126,0.6,0.07300000000000001,0.7,0.4475674603174603,0.5051492476298824,0.452093056611097
42
  0.8518651362984218,4750,0.32,0.55,0.61,0.71,0.32,0.315,0.18666666666666665,0.535,0.126,0.6,0.07300000000000001,0.7,0.44900793650793647,0.506463851684779,0.45371193799774445
43
  0.896700143472023,5000,0.32,0.55,0.61,0.71,0.32,0.315,0.18666666666666665,0.535,0.126,0.6,0.07300000000000001,0.7,0.4504365079365079,0.5075444106827467,0.4550673810391498
44
+ 0,0,0.51,0.67,0.73,0.79,0.51,0.495,0.22333333333333333,0.65,0.15000000000000002,0.71,0.08299999999999999,0.775,0.6078174603174604,0.6421054392332819,0.6046870533217332
45
+ 0.04483500717360115,250,0.37,0.5800000000000001,0.7,0.76,0.37,0.355,0.19333333333333336,0.56,0.14200000000000002,0.685,0.07700000000000001,0.74,0.5006904761904761,0.5510961510866085,0.49470575663819183
46
+ 0.0896700143472023,500,0.41000000000000003,0.64,0.69,0.74,0.41000000000000003,0.395,0.21666666666666667,0.63,0.14,0.675,0.07500000000000001,0.72,0.5314285714285715,0.5721800685665956,0.5304097862730295
47
+ 0.13450502152080343,750,0.39,0.6000000000000001,0.6599999999999999,0.74,0.39,0.375,0.2033333333333333,0.5900000000000001,0.134,0.645,0.07600000000000001,0.73,0.507968253968254,0.5570146451901427,0.5082542397780777
48
+ 0.1793400286944046,1000,0.36,0.5900000000000001,0.6799999999999999,0.74,0.36,0.345,0.19999999999999998,0.575,0.14,0.6699999999999999,0.07600000000000001,0.73,0.4911111111111111,0.5444100871020746,0.4908851013543535
49
+ 0.22417503586800575,1250,0.33999999999999997,0.5900000000000001,0.65,0.74,0.33999999999999997,0.32499999999999996,0.19999999999999998,0.575,0.134,0.64,0.07600000000000001,0.73,0.4821944444444445,0.5372297652804809,0.48098119062395145
50
+ 0.26901004304160686,1500,0.35,0.5800000000000001,0.65,0.72,0.35,0.33499999999999996,0.19666666666666666,0.565,0.134,0.64,0.07400000000000001,0.71,0.4796944444444445,0.530842962110912,0.48086326343017116
51
+ 0.31384505021520803,1750,0.35,0.5900000000000001,0.64,0.73,0.35,0.33999999999999997,0.2,0.575,0.132,0.63,0.07500000000000001,0.715,0.4810674603174603,0.5341847538669695,0.48347916886972375
52
+ 0.3586800573888092,2000,0.32,0.56,0.65,0.72,0.32,0.30500000000000005,0.19,0.5449999999999999,0.134,0.64,0.07400000000000001,0.71,0.46088888888888885,0.5165285075041046,0.4617689319972166
53
+ 0.4035150645624103,2250,0.34,0.56,0.62,0.72,0.34,0.325,0.19,0.5449999999999999,0.128,0.61,0.07400000000000001,0.71,0.46619444444444447,0.5198394589713159,0.46707566987691784
54
+ 0.4483500717360115,2500,0.32,0.56,0.63,0.71,0.32,0.31,0.18666666666666665,0.54,0.13,0.62,0.07300000000000001,0.7,0.45485714285714285,0.5097044316580317,0.45622676768732356
55
+ 0.4931850789096126,2750,0.32,0.56,0.62,0.71,0.32,0.31,0.19,0.5449999999999999,0.128,0.61,0.07300000000000001,0.7,0.4549444444444445,0.5099283717753459,0.4566732599109169
56
+ 0.5380200860832137,3000,0.32,0.56,0.62,0.7,0.32,0.30500000000000005,0.19,0.5449999999999999,0.128,0.61,0.07200000000000001,0.69,0.45273412698412696,0.5047680495553872,0.45291247928711553
57
+ 0.582855093256815,3250,0.31000000000000005,0.56,0.62,0.7,0.31000000000000005,0.29500000000000004,0.18666666666666665,0.54,0.128,0.61,0.07200000000000001,0.69,0.4458333333333333,0.49905077584575785,0.44488820687203773
58
+ 0.6276901004304161,3500,0.32,0.5700000000000001,0.63,0.71,0.32,0.30500000000000005,0.19333333333333333,0.5549999999999999,0.13,0.62,0.07300000000000001,0.7,0.4548055555555556,0.5085105182787869,0.45420237703894417
59
+ 0.6725251076040172,3750,0.31000000000000005,0.56,0.62,0.7,0.31000000000000005,0.29500000000000004,0.19,0.5449999999999999,0.128,0.61,0.07200000000000001,0.69,0.4482777777777778,0.5012734621614335,0.44843353657016455
60
+ 0.7173601147776184,4000,0.31000000000000005,0.56,0.62,0.7,0.31000000000000005,0.29500000000000004,0.19,0.5449999999999999,0.128,0.61,0.07200000000000001,0.69,0.44672222222222224,0.4999940528182116,0.4469804239829816
61
+ 0.7621951219512195,4250,0.31000000000000005,0.56,0.62,0.69,0.31000000000000005,0.29500000000000004,0.19,0.5449999999999999,0.128,0.61,0.07100000000000001,0.6799999999999999,0.44713888888888886,0.4982384650870371,0.44805277604847205
62
+ 0.8070301291248206,4500,0.31000000000000005,0.56,0.62,0.69,0.31000000000000005,0.29500000000000004,0.19,0.5449999999999999,0.128,0.61,0.07100000000000001,0.6799999999999999,0.44553968253968257,0.4969882004233377,0.4464265880407993
63
+ 0.8518651362984218,4750,0.30000000000000004,0.56,0.63,0.69,0.30000000000000004,0.29000000000000004,0.19,0.5449999999999999,0.13,0.62,0.07100000000000001,0.6799999999999999,0.44125000000000003,0.49620759222677624,0.446162306954853
64
+ 0.896700143472023,5000,0.30000000000000004,0.56,0.62,0.69,0.30000000000000004,0.29000000000000004,0.19,0.5449999999999999,0.128,0.61,0.07100000000000001,0.6799999999999999,0.4425555555555556,0.497185736443469,0.44753124213980516
final_metrics.json CHANGED
@@ -1,16 +1,231 @@
1
  {
2
- "val_cosine_accuracy@1": 0.83295,
3
- "val_cosine_accuracy@3": 0.9048,
4
- "val_cosine_accuracy@5": 0.930625,
5
- "val_cosine_precision@1": 0.83295,
6
- "val_cosine_precision@3": 0.3016,
7
- "val_cosine_precision@5": 0.186125,
8
- "val_cosine_recall@1": 0.83295,
9
- "val_cosine_recall@3": 0.9048,
10
- "val_cosine_recall@5": 0.930625,
11
- "val_cosine_ndcg@10": 0.8953809850299808,
12
- "val_cosine_mrr@1": 0.83295,
13
- "val_cosine_mrr@5": 0.8708829166666628,
14
- "val_cosine_mrr@10": 0.8748452281746,
15
- "val_cosine_map@100": 0.8769581868758156
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  }
 
1
  {
2
+ "nano_beir": {
3
+ "NanoClimateFEVER_cosine_accuracy@1": 0.1,
4
+ "NanoClimateFEVER_cosine_accuracy@3": 0.26,
5
+ "NanoClimateFEVER_cosine_accuracy@5": 0.34,
6
+ "NanoClimateFEVER_cosine_accuracy@10": 0.6,
7
+ "NanoClimateFEVER_cosine_precision@1": 0.1,
8
+ "NanoClimateFEVER_cosine_precision@3": 0.08666666666666666,
9
+ "NanoClimateFEVER_cosine_precision@5": 0.07200000000000001,
10
+ "NanoClimateFEVER_cosine_precision@10": 0.066,
11
+ "NanoClimateFEVER_cosine_recall@1": 0.04666666666666666,
12
+ "NanoClimateFEVER_cosine_recall@3": 0.12399999999999999,
13
+ "NanoClimateFEVER_cosine_recall@5": 0.16899999999999998,
14
+ "NanoClimateFEVER_cosine_recall@10": 0.2973333333333333,
15
+ "NanoClimateFEVER_cosine_ndcg@10": 0.1853895398720514,
16
+ "NanoClimateFEVER_cosine_mrr@10": 0.21938095238095237,
17
+ "NanoClimateFEVER_cosine_map@100": 0.1267197025068282,
18
+ "NanoDBPedia_cosine_accuracy@1": 0.68,
19
+ "NanoDBPedia_cosine_accuracy@3": 0.8,
20
+ "NanoDBPedia_cosine_accuracy@5": 0.86,
21
+ "NanoDBPedia_cosine_accuracy@10": 0.88,
22
+ "NanoDBPedia_cosine_precision@1": 0.68,
23
+ "NanoDBPedia_cosine_precision@3": 0.5066666666666666,
24
+ "NanoDBPedia_cosine_precision@5": 0.44800000000000006,
25
+ "NanoDBPedia_cosine_precision@10": 0.376,
26
+ "NanoDBPedia_cosine_recall@1": 0.08435515343632806,
27
+ "NanoDBPedia_cosine_recall@3": 0.13854094008435847,
28
+ "NanoDBPedia_cosine_recall@5": 0.17555319487559032,
29
+ "NanoDBPedia_cosine_recall@10": 0.25077742902228617,
30
+ "NanoDBPedia_cosine_ndcg@10": 0.4897177201468777,
31
+ "NanoDBPedia_cosine_mrr@10": 0.7498333333333334,
32
+ "NanoDBPedia_cosine_map@100": 0.34021634749539026,
33
+ "NanoFEVER_cosine_accuracy@1": 0.6,
34
+ "NanoFEVER_cosine_accuracy@3": 0.76,
35
+ "NanoFEVER_cosine_accuracy@5": 0.86,
36
+ "NanoFEVER_cosine_accuracy@10": 0.88,
37
+ "NanoFEVER_cosine_precision@1": 0.6,
38
+ "NanoFEVER_cosine_precision@3": 0.2533333333333333,
39
+ "NanoFEVER_cosine_precision@5": 0.17199999999999996,
40
+ "NanoFEVER_cosine_precision@10": 0.088,
41
+ "NanoFEVER_cosine_recall@1": 0.5466666666666666,
42
+ "NanoFEVER_cosine_recall@3": 0.7066666666666666,
43
+ "NanoFEVER_cosine_recall@5": 0.7966666666666665,
44
+ "NanoFEVER_cosine_recall@10": 0.8166666666666665,
45
+ "NanoFEVER_cosine_ndcg@10": 0.6976146491512496,
46
+ "NanoFEVER_cosine_mrr@10": 0.6975555555555556,
47
+ "NanoFEVER_cosine_map@100": 0.6474697943858455,
48
+ "NanoFiQA2018_cosine_accuracy@1": 0.26,
49
+ "NanoFiQA2018_cosine_accuracy@3": 0.42,
50
+ "NanoFiQA2018_cosine_accuracy@5": 0.46,
51
+ "NanoFiQA2018_cosine_accuracy@10": 0.58,
52
+ "NanoFiQA2018_cosine_precision@1": 0.26,
53
+ "NanoFiQA2018_cosine_precision@3": 0.18,
54
+ "NanoFiQA2018_cosine_precision@5": 0.124,
55
+ "NanoFiQA2018_cosine_precision@10": 0.08399999999999999,
56
+ "NanoFiQA2018_cosine_recall@1": 0.11974603174603175,
57
+ "NanoFiQA2018_cosine_recall@3": 0.25293650793650796,
58
+ "NanoFiQA2018_cosine_recall@5": 0.28343650793650793,
59
+ "NanoFiQA2018_cosine_recall@10": 0.37834920634920627,
60
+ "NanoFiQA2018_cosine_ndcg@10": 0.289405411189281,
61
+ "NanoFiQA2018_cosine_mrr@10": 0.34807936507936504,
62
+ "NanoFiQA2018_cosine_map@100": 0.2371668409805281,
63
+ "NanoHotpotQA_cosine_accuracy@1": 0.54,
64
+ "NanoHotpotQA_cosine_accuracy@3": 0.64,
65
+ "NanoHotpotQA_cosine_accuracy@5": 0.7,
66
+ "NanoHotpotQA_cosine_accuracy@10": 0.74,
67
+ "NanoHotpotQA_cosine_precision@1": 0.54,
68
+ "NanoHotpotQA_cosine_precision@3": 0.26666666666666666,
69
+ "NanoHotpotQA_cosine_precision@5": 0.184,
70
+ "NanoHotpotQA_cosine_precision@10": 0.106,
71
+ "NanoHotpotQA_cosine_recall@1": 0.27,
72
+ "NanoHotpotQA_cosine_recall@3": 0.4,
73
+ "NanoHotpotQA_cosine_recall@5": 0.46,
74
+ "NanoHotpotQA_cosine_recall@10": 0.53,
75
+ "NanoHotpotQA_cosine_ndcg@10": 0.48356059599535955,
76
+ "NanoHotpotQA_cosine_mrr@10": 0.606,
77
+ "NanoHotpotQA_cosine_map@100": 0.4167638398079239,
78
+ "NanoMSMARCO_cosine_accuracy@1": 0.28,
79
+ "NanoMSMARCO_cosine_accuracy@3": 0.54,
80
+ "NanoMSMARCO_cosine_accuracy@5": 0.62,
81
+ "NanoMSMARCO_cosine_accuracy@10": 0.8,
82
+ "NanoMSMARCO_cosine_precision@1": 0.28,
83
+ "NanoMSMARCO_cosine_precision@3": 0.17999999999999997,
84
+ "NanoMSMARCO_cosine_precision@5": 0.124,
85
+ "NanoMSMARCO_cosine_precision@10": 0.08,
86
+ "NanoMSMARCO_cosine_recall@1": 0.28,
87
+ "NanoMSMARCO_cosine_recall@3": 0.54,
88
+ "NanoMSMARCO_cosine_recall@5": 0.62,
89
+ "NanoMSMARCO_cosine_recall@10": 0.8,
90
+ "NanoMSMARCO_cosine_ndcg@10": 0.5241911345526384,
91
+ "NanoMSMARCO_cosine_mrr@10": 0.43837301587301575,
92
+ "NanoMSMARCO_cosine_map@100": 0.4480618800320956,
93
+ "NanoNFCorpus_cosine_accuracy@1": 0.34,
94
+ "NanoNFCorpus_cosine_accuracy@3": 0.52,
95
+ "NanoNFCorpus_cosine_accuracy@5": 0.58,
96
+ "NanoNFCorpus_cosine_accuracy@10": 0.68,
97
+ "NanoNFCorpus_cosine_precision@1": 0.34,
98
+ "NanoNFCorpus_cosine_precision@3": 0.33333333333333326,
99
+ "NanoNFCorpus_cosine_precision@5": 0.29600000000000004,
100
+ "NanoNFCorpus_cosine_precision@10": 0.25,
101
+ "NanoNFCorpus_cosine_recall@1": 0.012133063569139098,
102
+ "NanoNFCorpus_cosine_recall@3": 0.05465564949455657,
103
+ "NanoNFCorpus_cosine_recall@5": 0.07192795043792813,
104
+ "NanoNFCorpus_cosine_recall@10": 0.10822085751351866,
105
+ "NanoNFCorpus_cosine_ndcg@10": 0.2847553576589848,
106
+ "NanoNFCorpus_cosine_mrr@10": 0.44152380952380954,
107
+ "NanoNFCorpus_cosine_map@100": 0.108499059841261,
108
+ "NanoNQ_cosine_accuracy@1": 0.36,
109
+ "NanoNQ_cosine_accuracy@3": 0.56,
110
+ "NanoNQ_cosine_accuracy@5": 0.6,
111
+ "NanoNQ_cosine_accuracy@10": 0.62,
112
+ "NanoNQ_cosine_precision@1": 0.36,
113
+ "NanoNQ_cosine_precision@3": 0.19333333333333333,
114
+ "NanoNQ_cosine_precision@5": 0.128,
115
+ "NanoNQ_cosine_precision@10": 0.066,
116
+ "NanoNQ_cosine_recall@1": 0.35,
117
+ "NanoNQ_cosine_recall@3": 0.53,
118
+ "NanoNQ_cosine_recall@5": 0.58,
119
+ "NanoNQ_cosine_recall@10": 0.6,
120
+ "NanoNQ_cosine_ndcg@10": 0.490897686812855,
121
+ "NanoNQ_cosine_mrr@10": 0.4625,
122
+ "NanoNQ_cosine_map@100": 0.46206363135240186,
123
+ "NanoQuoraRetrieval_cosine_accuracy@1": 0.92,
124
+ "NanoQuoraRetrieval_cosine_accuracy@3": 1.0,
125
+ "NanoQuoraRetrieval_cosine_accuracy@5": 1.0,
126
+ "NanoQuoraRetrieval_cosine_accuracy@10": 1.0,
127
+ "NanoQuoraRetrieval_cosine_precision@1": 0.92,
128
+ "NanoQuoraRetrieval_cosine_precision@3": 0.4133333333333333,
129
+ "NanoQuoraRetrieval_cosine_precision@5": 0.264,
130
+ "NanoQuoraRetrieval_cosine_precision@10": 0.13399999999999998,
131
+ "NanoQuoraRetrieval_cosine_recall@1": 0.8073333333333332,
132
+ "NanoQuoraRetrieval_cosine_recall@3": 0.9653333333333333,
133
+ "NanoQuoraRetrieval_cosine_recall@5": 0.986,
134
+ "NanoQuoraRetrieval_cosine_recall@10": 0.99,
135
+ "NanoQuoraRetrieval_cosine_ndcg@10": 0.960129267031932,
136
+ "NanoQuoraRetrieval_cosine_mrr@10": 0.96,
137
+ "NanoQuoraRetrieval_cosine_map@100": 0.945489898989899,
138
+ "NanoSCIDOCS_cosine_accuracy@1": 0.48,
139
+ "NanoSCIDOCS_cosine_accuracy@3": 0.72,
140
+ "NanoSCIDOCS_cosine_accuracy@5": 0.76,
141
+ "NanoSCIDOCS_cosine_accuracy@10": 0.8,
142
+ "NanoSCIDOCS_cosine_precision@1": 0.48,
143
+ "NanoSCIDOCS_cosine_precision@3": 0.3666666666666666,
144
+ "NanoSCIDOCS_cosine_precision@5": 0.28400000000000003,
145
+ "NanoSCIDOCS_cosine_precision@10": 0.19199999999999995,
146
+ "NanoSCIDOCS_cosine_recall@1": 0.10166666666666666,
147
+ "NanoSCIDOCS_cosine_recall@3": 0.22766666666666666,
148
+ "NanoSCIDOCS_cosine_recall@5": 0.29266666666666663,
149
+ "NanoSCIDOCS_cosine_recall@10": 0.3946666666666666,
150
+ "NanoSCIDOCS_cosine_ndcg@10": 0.3895503827770311,
151
+ "NanoSCIDOCS_cosine_mrr@10": 0.5951904761904762,
152
+ "NanoSCIDOCS_cosine_map@100": 0.3082897117657901,
153
+ "NanoArguAna_cosine_accuracy@1": 0.14,
154
+ "NanoArguAna_cosine_accuracy@3": 0.38,
155
+ "NanoArguAna_cosine_accuracy@5": 0.5,
156
+ "NanoArguAna_cosine_accuracy@10": 0.66,
157
+ "NanoArguAna_cosine_precision@1": 0.14,
158
+ "NanoArguAna_cosine_precision@3": 0.12666666666666665,
159
+ "NanoArguAna_cosine_precision@5": 0.1,
160
+ "NanoArguAna_cosine_precision@10": 0.06600000000000002,
161
+ "NanoArguAna_cosine_recall@1": 0.14,
162
+ "NanoArguAna_cosine_recall@3": 0.38,
163
+ "NanoArguAna_cosine_recall@5": 0.5,
164
+ "NanoArguAna_cosine_recall@10": 0.66,
165
+ "NanoArguAna_cosine_ndcg@10": 0.38615266678375515,
166
+ "NanoArguAna_cosine_mrr@10": 0.300095238095238,
167
+ "NanoArguAna_cosine_map@100": 0.3088425567963239,
168
+ "NanoSciFact_cosine_accuracy@1": 0.38,
169
+ "NanoSciFact_cosine_accuracy@3": 0.54,
170
+ "NanoSciFact_cosine_accuracy@5": 0.58,
171
+ "NanoSciFact_cosine_accuracy@10": 0.7,
172
+ "NanoSciFact_cosine_precision@1": 0.38,
173
+ "NanoSciFact_cosine_precision@3": 0.19999999999999996,
174
+ "NanoSciFact_cosine_precision@5": 0.12800000000000003,
175
+ "NanoSciFact_cosine_precision@10": 0.08,
176
+ "NanoSciFact_cosine_recall@1": 0.345,
177
+ "NanoSciFact_cosine_recall@3": 0.525,
178
+ "NanoSciFact_cosine_recall@5": 0.565,
179
+ "NanoSciFact_cosine_recall@10": 0.7,
180
+ "NanoSciFact_cosine_ndcg@10": 0.5292195947118973,
181
+ "NanoSciFact_cosine_mrr@10": 0.4820793650793651,
182
+ "NanoSciFact_cosine_map@100": 0.47730440170572996,
183
+ "NanoTouche2020_cosine_accuracy@1": 0.5102040816326531,
184
+ "NanoTouche2020_cosine_accuracy@3": 0.8163265306122449,
185
+ "NanoTouche2020_cosine_accuracy@5": 0.8571428571428571,
186
+ "NanoTouche2020_cosine_accuracy@10": 0.9591836734693877,
187
+ "NanoTouche2020_cosine_precision@1": 0.5102040816326531,
188
+ "NanoTouche2020_cosine_precision@3": 0.45578231292517,
189
+ "NanoTouche2020_cosine_precision@5": 0.4040816326530612,
190
+ "NanoTouche2020_cosine_precision@10": 0.3428571428571428,
191
+ "NanoTouche2020_cosine_recall@1": 0.04030730530317779,
192
+ "NanoTouche2020_cosine_recall@3": 0.10027039527564566,
193
+ "NanoTouche2020_cosine_recall@5": 0.14754618693234572,
194
+ "NanoTouche2020_cosine_recall@10": 0.2268233238254859,
195
+ "NanoTouche2020_cosine_ndcg@10": 0.3942611497955867,
196
+ "NanoTouche2020_cosine_mrr@10": 0.6711613216715256,
197
+ "NanoTouche2020_cosine_map@100": 0.28013001290517386,
198
+ "NanoBEIR_mean_cosine_accuracy@1": 0.4300156985871272,
199
+ "NanoBEIR_mean_cosine_accuracy@3": 0.6120251177394034,
200
+ "NanoBEIR_mean_cosine_accuracy@5": 0.6705494505494506,
201
+ "NanoBEIR_mean_cosine_accuracy@10": 0.7614756671899527,
202
+ "NanoBEIR_mean_cosine_precision@1": 0.4300156985871272,
203
+ "NanoBEIR_mean_cosine_precision@3": 0.2740345368916797,
204
+ "NanoBEIR_mean_cosine_precision@5": 0.20985243328100472,
205
+ "NanoBEIR_mean_cosine_precision@10": 0.14852747252747253,
206
+ "NanoBEIR_mean_cosine_recall@1": 0.24183652979907766,
207
+ "NanoBEIR_mean_cosine_recall@3": 0.3803900122659796,
208
+ "NanoBEIR_mean_cosine_recall@5": 0.4344459364242849,
209
+ "NanoBEIR_mean_cosine_recall@10": 0.5194490371828588,
210
+ "NanoBEIR_mean_cosine_ndcg@10": 0.46960347357534615,
211
+ "NanoBEIR_mean_cosine_mrr@10": 0.536290187137126,
212
+ "NanoBEIR_mean_cosine_map@100": 0.39284751373578397
213
+ },
214
+ "beir_touche2020": {
215
+ "BeIR-touche2020-subset-test_cosine_accuracy@1": 0.7346938775510204,
216
+ "BeIR-touche2020-subset-test_cosine_accuracy@3": 0.9387755102040817,
217
+ "BeIR-touche2020-subset-test_cosine_accuracy@5": 0.9591836734693877,
218
+ "BeIR-touche2020-subset-test_cosine_accuracy@10": 0.9795918367346939,
219
+ "BeIR-touche2020-subset-test_cosine_precision@1": 0.7346938775510204,
220
+ "BeIR-touche2020-subset-test_cosine_precision@3": 0.6802721088435373,
221
+ "BeIR-touche2020-subset-test_cosine_precision@5": 0.6612244897959185,
222
+ "BeIR-touche2020-subset-test_cosine_precision@10": 0.5857142857142857,
223
+ "BeIR-touche2020-subset-test_cosine_recall@1": 0.01628348794626864,
224
+ "BeIR-touche2020-subset-test_cosine_recall@3": 0.045158349206937544,
225
+ "BeIR-touche2020-subset-test_cosine_recall@5": 0.07322241057348641,
226
+ "BeIR-touche2020-subset-test_cosine_recall@10": 0.12952247914108056,
227
+ "BeIR-touche2020-subset-test_cosine_ndcg@10": 0.6211658380172899,
228
+ "BeIR-touche2020-subset-test_cosine_mrr@10": 0.8306122448979592,
229
+ "BeIR-touche2020-subset-test_cosine_map@100": 0.24407616368208723
230
+ }
231
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:db59c43bf96bbee3c0ba2dacf6445ee23cd7102f00450058446032338eff15a7
3
  size 133462128
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:325ea67b4c6fb434aa13d441e188dab3f657079050a07c5c0072112cfb0cb217
3
  size 133462128
modules.json CHANGED
@@ -10,11 +10,5 @@
10
  "name": "1",
11
  "path": "1_Pooling",
12
  "type": "sentence_transformers.models.Pooling"
13
- },
14
- {
15
- "idx": 2,
16
- "name": "2",
17
- "path": "2_Normalize",
18
- "type": "sentence_transformers.models.Normalize"
19
  }
20
  ]
 
10
  "name": "1",
11
  "path": "1_Pooling",
12
  "type": "sentence_transformers.models.Pooling"
 
 
 
 
 
 
13
  }
14
  ]
tokenizer_config.json CHANGED
@@ -41,14 +41,14 @@
41
  "special": true
42
  }
43
  },
44
- "clean_up_tokenization_spaces": false,
45
  "cls_token": "[CLS]",
46
  "do_basic_tokenize": true,
47
  "do_lower_case": true,
48
  "extra_special_tokens": {},
49
  "mask_token": "[MASK]",
50
  "max_length": 128,
51
- "model_max_length": 128,
52
  "never_split": null,
53
  "pad_to_multiple_of": null,
54
  "pad_token": "[PAD]",
 
41
  "special": true
42
  }
43
  },
44
+ "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
  "do_basic_tokenize": true,
47
  "do_lower_case": true,
48
  "extra_special_tokens": {},
49
  "mask_token": "[MASK]",
50
  "max_length": 128,
51
+ "model_max_length": 512,
52
  "never_split": null,
53
  "pad_to_multiple_of": null,
54
  "pad_token": "[PAD]",