Commit ·
49f48ec
1
Parent(s): f9e3310
Update bert-base-uncased_results.csv
Browse files- bert-base-uncased_results.csv +57 -57
bert-base-uncased_results.csv
CHANGED
|
@@ -1,57 +1,57 @@
|
|
| 1 |
-
dataset,metric,value
|
| 2 |
-
AmazonCounterfactualClassification,accuracy,0.7425373134328358
|
| 3 |
-
AmazonPolarityClassification,accuracy,0.7132945
|
| 4 |
-
AmazonReviewsClassification,accuracy,0.33564
|
| 5 |
-
Banking77Classification,accuracy,0.6340584415584415
|
| 6 |
-
EmotionClassification,accuracy,0.3528
|
| 7 |
-
ImdbClassification,accuracy,0.653456
|
| 8 |
-
MassiveIntentClassification,accuracy,0.5988231338264962
|
| 9 |
-
MassiveScenarioClassification,accuracy,0.6427706792199059
|
| 10 |
-
MTOPDomainClassification,accuracy,0.8262653898768809
|
| 11 |
-
MTOPIntentClassification,accuracy,0.6813725490196079
|
| 12 |
-
ToxicConversationsClassification,accuracy,0.6999679999999999
|
| 13 |
-
TweetSentimentExtractionClassification,accuracy,0.5180814940577249
|
| 14 |
-
ArxivClusteringP2P,v_measure,0.3518932830729758
|
| 15 |
-
ArxivClusteringS2S,v_measure,0.2750822951507033
|
| 16 |
-
BiorxivClusteringP2P,v_measure,0.3012278640553615
|
| 17 |
-
BiorxivClusteringS2S,v_measure,0.2476601053047804
|
| 18 |
-
MedrxivClusteringP2P,v_measure,0.26087688306606044
|
| 19 |
-
MedrxivClusteringS2S,v_measure,0.23604914608225602
|
| 20 |
-
RedditClustering,v_measure,0.27241718557705
|
| 21 |
-
RedditClusteringP2P,v_measure,0.43323854460464056
|
| 22 |
-
StackExchangeClustering,v_measure,0.4358261870547655
|
| 23 |
-
StackExchangeClusteringP2P,v_measure,0.2654851675795123
|
| 24 |
-
TwentyNewsgroupsClustering,v_measure,0.23354321403189832
|
| 25 |
-
SprintDuplicateQuestions,ap,0.36808703728970593
|
| 26 |
-
TwitterSemEval2015,ap,0.5589788752976392
|
| 27 |
-
TwitterURLCorpus,ap,0.7628732783437441
|
| 28 |
-
AskUbuntuDupQuestions,map,0.4584088706528255
|
| 29 |
-
MindSmallReranking,map,0.28366637355845425
|
| 30 |
-
SciDocsRR,map,0.6493728987830145
|
| 31 |
-
StackOverflowDupQuestions,map,0.34615478798860166
|
| 32 |
-
ArguAna,ndcg_at_10,0.28294
|
| 33 |
-
ClimateFEVER,ndcg_at_10,0.0541
|
| 34 |
-
CQADupstackRetrieval,ndcg_at_10,0.05506583333333333
|
| 35 |
-
DBPedia,ndcg_at_10,0.04132
|
| 36 |
-
FEVER,ndcg_at_10,0.033
|
| 37 |
-
FiQA2018,ndcg_at_10,0.02191
|
| 38 |
-
HotpotQA,ndcg_at_10,0.0826
|
| 39 |
-
MSMARCO,ndcg_at_10,0.06176
|
| 40 |
-
NFCorpus,ndcg_at_10,0.04304
|
| 41 |
-
NQ,ndcg_at_10,0.02615
|
| 42 |
-
QuoraRetrieval,ndcg_at_10,0.61029
|
| 43 |
-
SCIDOCS,ndcg_at_10,0.02815
|
| 44 |
-
SciFact,ndcg_at_10,0.13339
|
| 45 |
-
Touche2020,ndcg_at_10,0.00967
|
| 46 |
-
TRECCOVID,ndcg_at_10,0.14745
|
| 47 |
-
BIOSSES,cosine_spearman,0.5469823428818151
|
| 48 |
-
SICK-R,cosine_spearman,0.5864506948179484
|
| 49 |
-
STS12,cosine_spearman,0.30871788407575457
|
| 50 |
-
STS13,cosine_spearman,0.5989485045425808
|
| 51 |
-
STS14,cosine_spearman,0.4772791352844941
|
| 52 |
-
STS15,cosine_spearman,0.6028567383446486
|
| 53 |
-
STS16,cosine_spearman,0.6373272693604403
|
| 54 |
-
STS17,cosine_spearman,0.6410023907260639
|
| 55 |
-
STS22,cosine_spearman,0.563667996535454
|
| 56 |
-
STSBenchmark,cosine_spearman,0.4729108172174081
|
| 57 |
-
SummEval,cosine_spearman,0.2981716355664815
|
|
|
|
| 1 |
+
model,dataset,metric,value
|
| 2 |
+
bert-base-uncased,AmazonCounterfactualClassification,accuracy,0.7425373134328358
|
| 3 |
+
bert-base-uncased,AmazonPolarityClassification,accuracy,0.7132945
|
| 4 |
+
bert-base-uncased,AmazonReviewsClassification,accuracy,0.33564
|
| 5 |
+
bert-base-uncased,Banking77Classification,accuracy,0.6340584415584415
|
| 6 |
+
bert-base-uncased,EmotionClassification,accuracy,0.3528
|
| 7 |
+
bert-base-uncased,ImdbClassification,accuracy,0.653456
|
| 8 |
+
bert-base-uncased,MassiveIntentClassification,accuracy,0.5988231338264962
|
| 9 |
+
bert-base-uncased,MassiveScenarioClassification,accuracy,0.6427706792199059
|
| 10 |
+
bert-base-uncased,MTOPDomainClassification,accuracy,0.8262653898768809
|
| 11 |
+
bert-base-uncased,MTOPIntentClassification,accuracy,0.6813725490196079
|
| 12 |
+
bert-base-uncased,ToxicConversationsClassification,accuracy,0.6999679999999999
|
| 13 |
+
bert-base-uncased,TweetSentimentExtractionClassification,accuracy,0.5180814940577249
|
| 14 |
+
bert-base-uncased,ArxivClusteringP2P,v_measure,0.3518932830729758
|
| 15 |
+
bert-base-uncased,ArxivClusteringS2S,v_measure,0.2750822951507033
|
| 16 |
+
bert-base-uncased,BiorxivClusteringP2P,v_measure,0.3012278640553615
|
| 17 |
+
bert-base-uncased,BiorxivClusteringS2S,v_measure,0.2476601053047804
|
| 18 |
+
bert-base-uncased,MedrxivClusteringP2P,v_measure,0.26087688306606044
|
| 19 |
+
bert-base-uncased,MedrxivClusteringS2S,v_measure,0.23604914608225602
|
| 20 |
+
bert-base-uncased,RedditClustering,v_measure,0.27241718557705
|
| 21 |
+
bert-base-uncased,RedditClusteringP2P,v_measure,0.43323854460464056
|
| 22 |
+
bert-base-uncased,StackExchangeClustering,v_measure,0.4358261870547655
|
| 23 |
+
bert-base-uncased,StackExchangeClusteringP2P,v_measure,0.2654851675795123
|
| 24 |
+
bert-base-uncased,TwentyNewsgroupsClustering,v_measure,0.23354321403189832
|
| 25 |
+
bert-base-uncased,SprintDuplicateQuestions,ap,0.36808703728970593
|
| 26 |
+
bert-base-uncased,TwitterSemEval2015,ap,0.5589788752976392
|
| 27 |
+
bert-base-uncased,TwitterURLCorpus,ap,0.7628732783437441
|
| 28 |
+
bert-base-uncased,AskUbuntuDupQuestions,map,0.4584088706528255
|
| 29 |
+
bert-base-uncased,MindSmallReranking,map,0.28366637355845425
|
| 30 |
+
bert-base-uncased,SciDocsRR,map,0.6493728987830145
|
| 31 |
+
bert-base-uncased,StackOverflowDupQuestions,map,0.34615478798860166
|
| 32 |
+
bert-base-uncased,ArguAna,ndcg_at_10,0.28294
|
| 33 |
+
bert-base-uncased,ClimateFEVER,ndcg_at_10,0.0541
|
| 34 |
+
bert-base-uncased,CQADupstackRetrieval,ndcg_at_10,0.05506583333333333
|
| 35 |
+
bert-base-uncased,DBPedia,ndcg_at_10,0.04132
|
| 36 |
+
bert-base-uncased,FEVER,ndcg_at_10,0.033
|
| 37 |
+
bert-base-uncased,FiQA2018,ndcg_at_10,0.02191
|
| 38 |
+
bert-base-uncased,HotpotQA,ndcg_at_10,0.0826
|
| 39 |
+
bert-base-uncased,MSMARCO,ndcg_at_10,0.06176
|
| 40 |
+
bert-base-uncased,NFCorpus,ndcg_at_10,0.04304
|
| 41 |
+
bert-base-uncased,NQ,ndcg_at_10,0.02615
|
| 42 |
+
bert-base-uncased,QuoraRetrieval,ndcg_at_10,0.61029
|
| 43 |
+
bert-base-uncased,SCIDOCS,ndcg_at_10,0.02815
|
| 44 |
+
bert-base-uncased,SciFact,ndcg_at_10,0.13339
|
| 45 |
+
bert-base-uncased,Touche2020,ndcg_at_10,0.00967
|
| 46 |
+
bert-base-uncased,TRECCOVID,ndcg_at_10,0.14745
|
| 47 |
+
bert-base-uncased,BIOSSES,cosine_spearman,0.5469823428818151
|
| 48 |
+
bert-base-uncased,SICK-R,cosine_spearman,0.5864506948179484
|
| 49 |
+
bert-base-uncased,STS12,cosine_spearman,0.30871788407575457
|
| 50 |
+
bert-base-uncased,STS13,cosine_spearman,0.5989485045425808
|
| 51 |
+
bert-base-uncased,STS14,cosine_spearman,0.4772791352844941
|
| 52 |
+
bert-base-uncased,STS15,cosine_spearman,0.6028567383446486
|
| 53 |
+
bert-base-uncased,STS16,cosine_spearman,0.6373272693604403
|
| 54 |
+
bert-base-uncased,STS17,cosine_spearman,0.6410023907260639
|
| 55 |
+
bert-base-uncased,STS22,cosine_spearman,0.563667996535454
|
| 56 |
+
bert-base-uncased,STSBenchmark,cosine_spearman,0.4729108172174081
|
| 57 |
+
bert-base-uncased,SummEval,cosine_spearman,0.2981716355664815
|