Dataset Viewer
Auto-converted to Parquet Duplicate
model name
stringlengths
5
56
dataset
stringclasses
3 values
method
stringclasses
4 values
file name
stringclasses
1 value
submitter
stringclasses
2 values
MICRO precision
float64
0.03
0.96
MICRO recall
float64
0
0.94
MICRO f1
float64
0
0.94
MACRO precision
float64
0.13
0.94
MACRO recall
float64
0
0.93
MACRO f1
float64
0
0.93
detail result
stringlengths
1.16k
1.48k
01-ai@Yi-1.5-6B-Chat
Setting1
icl
answer.txt
zhaorui
0.673741
0.336458
0.448793
0.44781
0.248924
0.319981
{ "ID": { "precision": 0.7566633761105627, "recall": 0.34253156071947266, "f1": 0.4715834807352149, "support": 8951 }, "CONTACT": { "precision": 0.0, "recall": 0.0, "f1": 0.0, "support": 10 }, "DATE": { "precision": 0.7972571428...
01-ai@Yi-1.5-6B-Chat
Setting2
icl
answer.txt
zhaorui
0.396261
0.1807
0.248212
0.317656
0.175866
0.226393
{ "ID": { "precision": 0.5772517159266004, "recall": 0.6081759149940968, "f1": 0.5923104563420769, "support": 6776 }, "CONTACT": { "precision": 0.23333333333333334, "recall": 0.006329113924050633, "f1": 0.01232394366197183, "support": 1106 ...
01-ai@Yi-1.5-6B-Chat
Setting3
icl
answer.txt
zhaorui
0.833257
0.600661
0.698094
0.801466
0.576268
0.670462
{ "ID": { "precision": 0.8891784391715813, "recall": 0.8323276415457697, "f1": 0.859814323607427, "support": 7789 }, "CONTACT": { "precision": 0.8794117647058823, "recall": 0.5652173913043478, "f1": 0.6881472957422324, "support": 529 }, ...
CRF++
Setting1
crf
answer.txt
zhoarui
0.657063
0.266374
0.379072
0.448857
0.279593
0.344559
{ "LOCATION": { "precision": 0.16030534351145037, "recall": 0.006390748630553865, "f1": 0.012291483757682178, "support": 9858 }, "ID": { "precision": 0.9885844748858448, "recall": 0.19349793319182215, "f1": 0.3236475754461366, "support": 8951 ...
CRF++
Setting2
crf
answer.txt
zhoarui
0.303494
0.088684
0.13726
0.368191
0.092017
0.147237
{ "LOCATION": { "precision": 0.03461935928349983, "recall": 0.02296355535245059, "f1": 0.027611786523799718, "support": 17506 }, "ID": { "precision": 0.434659514230368, "recall": 0.4304899645808737, "f1": 0.4325646919255579, "support": 6776 ...
CRF++
Setting3
crf
answer.txt
zhoarui
0.856147
0.525169
0.651005
0.851477
0.49451
0.625658
{ "LOCATION": { "precision": 0.6475649630012046, "recall": 0.2723850886717336, "f1": 0.3834709059410985, "support": 13815 }, "ID": { "precision": 0.9730736663844199, "recall": 0.7377070227243548, "f1": 0.8391996494815248, "support": 7789 },...
EleutherAI@pythia-1b
Setting1
icl
answer.txt
zhaorui
0.538551
0.250135
0.341608
0.36385
0.153981
0.216387
{ "ID": { "precision": 0.45235707121364094, "recall": 0.20154172718132052, "f1": 0.27884689697812814, "support": 8951 }, "CONTACT": { "precision": 0.0, "recall": 0.0, "f1": 0.0, "support": 10 }, "DATE": { "precision": 0.74636723...
EleutherAI@pythia-1b
Setting2
icl
answer.txt
zhaorui
0.229472
0.151139
0.182245
0.246173
0.147191
0.184229
{ "ID": { "precision": 0.4749574105621806, "recall": 0.6171782762691853, "f1": 0.5368076503433669, "support": 6776 }, "CONTACT": { "precision": 0.11428571428571428, "recall": 0.003616636528028933, "f1": 0.007011393514460999, "support": 1106 ...
EleutherAI@pythia-1b
Setting3
icl
answer.txt
zhaorui
0.861082
0.552898
0.673405
0.814266
0.491727
0.613168
{ "ID": { "precision": 0.87248322147651, "recall": 0.7844395942996534, "f1": 0.8261222282314765, "support": 7789 }, "CONTACT": { "precision": 0.8852459016393442, "recall": 0.5103969754253308, "f1": 0.6474820143884891, "support": 529 }, ...
EleutherAI@pythia-2.8b
Setting1
icl
answer.txt
zhaorui
0.597884
0.235008
0.337396
0.391379
0.148751
0.215571
{ "ID": { "precision": 0.5385248149163696, "recall": 0.21941682493576137, "f1": 0.31179552309890457, "support": 8951 }, "CONTACT": { "precision": 0.0, "recall": 0.0, "f1": 0.0, "support": 10 }, "DATE": { "precision": 0.628522920...
EleutherAI@pythia-2.8b
Setting2
icl
answer.txt
zhaorui
0.258988
0.136246
0.178558
0.249814
0.141902
0.180994
{ "ID": { "precision": 0.5102875706894477, "recall": 0.6258854781582054, "f1": 0.5622058726055544, "support": 6776 }, "CONTACT": { "precision": 0.017543859649122806, "recall": 0.0009041591320072332, "f1": 0.0017196904557179706, "support": 1106 ...
EleutherAI@pythia-2.8b
Setting3
icl
answer.txt
zhaorui
0.883802
0.55635
0.682849
0.818117
0.511118
0.629165
{ "ID": { "precision": 0.8936535162950258, "recall": 0.8026704326614457, "f1": 0.845722015556307, "support": 7789 }, "CONTACT": { "precision": 0.7795698924731183, "recall": 0.5482041587901701, "f1": 0.6437291897891232, "support": 529 }, ...
NCRF++
Setting1
NCRF++
answer.txt
zhaorui
0.850321
0.410928
0.554087
0.623446
0.29958
0.404695
{ "NAME": { "precision": 0.8325492689129053, "recall": 0.7720787642966631, "f1": 0.8011745992903462, "support": 8481 }, "ID": { "precision": 0.9909584086799277, "recall": 0.18366662942687967, "f1": 0.3098963242224317, "support": 8951 }, ...
NCRF++
Setting2
NCRF++
answer.txt
zhaorui
0.451742
0.210843
0.2875
0.440042
0.162872
0.237748
{ "NAME": { "precision": 0.7314139941690962, "recall": 0.21825893099885812, "f1": 0.33619498303949075, "support": 18391 }, "ID": { "precision": 0.7721874325199741, "recall": 0.5277449822904369, "f1": 0.62698343122644, "support": 6776 }, ...
NCRF++
Setting3
NCRF++
answer.txt
zhaorui
0.808136
0.566309
0.665948
0.822376
0.522093
0.638701
{ "NAME": { "precision": 0.8729248152976966, "recall": 0.7312509101499928, "f1": 0.7958318475375412, "support": 13734 }, "ID": { "precision": 0.9626706133859563, "recall": 0.7515727307741688, "f1": 0.8441240086517664, "support": 7789 }, ...
Qwen@Qwen1.5-7B-Chat
Setting1
icl
answer.txt
zhaorui
0.027778
0.000114
0.000227
0.126225
0.00008
0.00016
{ "ID": { "precision": 0.0, "recall": 0.0, "f1": 0.0, "support": 8951 }, "CONTACT": { "precision": 0.0, "recall": 0.0, "f1": 0.0, "support": 10 }, "DATE": { "precision": 0.5, "recall": 0.0002610625244746117, "f1"...
Qwen@Qwen1.5-7B-Chat
Setting2
icl
answer.txt
zhaorui
0.476744
0.001945
0.003874
0.273566
0.001119
0.00223
{ "ID": { "precision": 0.0, "recall": 0.0, "f1": 0.0, "support": 6776 }, "CONTACT": { "precision": 0.0, "recall": 0.0, "f1": 0.0, "support": 1106 }, "DATE": { "precision": 0.5463917525773195, "recall": 0.0035748010252259...
Qwen@Qwen1.5-7B-Chat
Setting3
icl
answer.txt
zhaorui
0.873993
0.025087
0.048773
0.871603
0.018961
0.037114
{ "ID": { "precision": 1.0, "recall": 0.00115547567081782, "f1": 0.002308284175429597, "support": 7789 }, "CONTACT": { "precision": 1.0, "recall": 0.001890359168241966, "f1": 0.0037735849056603774, "support": 529 }, "DATE": { "p...
deepseek-ai@DeepSeek-R1-Distill-Llama-8B
Setting1
icl
answer.txt
zhaorui
0.67086
0.205094
0.314147
0.476263
0.124079
0.196869
{ "NAME": { "precision": 0.5066832452595585, "recall": 0.19219431670793538, "f1": 0.2786801162591896, "support": 8481 }, "ID": { "precision": 0.7588126159554731, "recall": 0.22846609317394703, "f1": 0.3511935428473295, "support": 8951 }, ...
deepseek-ai@DeepSeek-R1-Distill-Llama-8B
Setting2
icl
answer.txt
zhaorui
0.326696
0.140455
0.196451
0.348844
0.138422
0.198198
{ "NAME": { "precision": 0.3600294804078123, "recall": 0.15937143167853843, "f1": 0.22094075079149705, "support": 18391 }, "ID": { "precision": 0.6116489446119664, "recall": 0.5687721369539551, "f1": 0.5894318268716067, "support": 6776 }, ...
deepseek-ai@DeepSeek-R1-Distill-Llama-8B
Setting3
icl
answer.txt
zhaorui
0.907324
0.467927
0.617431
0.877819
0.446832
0.592213
{ "ID": { "precision": 0.9341042293679709, "recall": 0.757093336756965, "f1": 0.836335271592682, "support": 7789 }, "DATE": { "precision": 0.9165489514288819, "recall": 0.4398498200969912, "f1": 0.5944326990838619, "support": 19177 }, "...
meta-llama@Llama-2-7b-hf
Setting1
icl
answer.txt
zhaorui
0.519124
0.282271
0.365697
0.335145
0.172249
0.227549
{ "ID": { "precision": 0.4086677367576244, "recall": 0.14221874650876998, "f1": 0.21100613293552128, "support": 8951 }, "CONTACT": { "precision": 0.0, "recall": 0.0, "f1": 0.0, "support": 10 }, "DATE": { "precision": 0.628837393...
meta-llama@Llama-2-7b-hf
Setting2
icl
answer.txt
zhaorui
0.229651
0.18019
0.201936
0.282236
0.168357
0.210906
{ "ID": { "precision": 0.5350502512562814, "recall": 0.6285419126328218, "f1": 0.5780401737242128, "support": 6776 }, "CONTACT": { "precision": 0.1206896551724138, "recall": 0.006329113924050633, "f1": 0.012027491408934709, "support": 1106 ...
meta-llama@Llama-2-7b-hf
Setting3
icl
answer.txt
zhaorui
0.852069
0.608621
0.710058
0.799693
0.566826
0.663419
{ "ID": { "precision": 0.896279905960448, "recall": 0.8320708691744768, "f1": 0.8629826897470039, "support": 7789 }, "CONTACT": { "precision": 0.7916666666666666, "recall": 0.610586011342155, "f1": 0.6894343649946638, "support": 529 }, ...
meta-llama@Meta-Llama-3-8B-Instruct
Setting1
icl
answer.txt
zhaorui
0.507072
0.258397
0.342342
0.368275
0.162881
0.225865
{ "ID": { "precision": 0.4260834345889024, "recall": 0.11752876773544856, "f1": 0.1842381786339755, "support": 8951 }, "CONTACT": { "precision": 0.0, "recall": 0.0, "f1": 0.0, "support": 10 }, "DATE": { "precision": 0.4078395624...
meta-llama@Meta-Llama-3-8B-Instruct
Setting2
icl
answer.txt
zhaorui
0.283601
0.223033
0.249696
0.28171
0.1646
0.20779
{ "ID": { "precision": 0.362962962962963, "recall": 0.27479338842975204, "f1": 0.312783470519066, "support": 6776 }, "CONTACT": { "precision": 0.1320754716981132, "recall": 0.012658227848101266, "f1": 0.023102310231023097, "support": 1106 }...
meta-llama@Meta-Llama-3-8B-Instruct
Setting3
icl
answer.txt
zhaorui
0.837661
0.611693
0.707061
0.815803
0.581267
0.678848
{ "ID": { "precision": 0.9124541540424175, "recall": 0.7346257542688407, "f1": 0.8139402560455191, "support": 7789 }, "CONTACT": { "precision": 0.7980997624703088, "recall": 0.6351606805293005, "f1": 0.7073684210526315, "support": 529 }, ...
meta-llama@Meta-Llama-3-8B
Setting1
icl
answer.txt
zhaorui
0.540875
0.081052
0.140978
0.292196
0.050096
0.085529
{ "ID": { "precision": 0.22379032258064516, "recall": 0.012400849067143336, "f1": 0.023499523658304225, "support": 8951 }, "CONTACT": { "precision": 0.0, "recall": 0.0, "f1": 0.0, "support": 10 }, "DATE": { "precision": 0.706453...
meta-llama@Meta-Llama-3-8B
Setting2
icl
answer.txt
zhaorui
0.299782
0.075154
0.12018
0.296362
0.051534
0.087801
{ "ID": { "precision": 0.38205499276411, "recall": 0.07792207792207792, "f1": 0.12944349105172837, "support": 6776 }, "CONTACT": { "precision": 0.1935483870967742, "recall": 0.0054249547920434, "f1": 0.010554089709762533, "support": 1106 },...
meta-llama@Meta-Llama-3-8B
Setting3
icl
answer.txt
zhaorui
0.676322
0.07057
0.127804
0.683701
0.073847
0.133297
{ "ID": { "precision": 0.6742857142857143, "recall": 0.015149569906278084, "f1": 0.029633350075339022, "support": 7789 }, "CONTACT": { "precision": 0.7878787878787878, "recall": 0.04914933837429111, "f1": 0.09252669039145905, "support": 529 ...
microsoft@Phi-3-mini-4k-instruct
Setting1
icl
answer.txt
zhaorui
0.618346
0.257144
0.363234
0.435172
0.184519
0.259153
{ "ID": { "precision": 0.4675850891410049, "recall": 0.12892414255390458, "f1": 0.20211927489272266, "support": 8951 }, "CONTACT": { "precision": 0.0, "recall": 0.0, "f1": 0.0, "support": 10 }, "DATE": { "precision": 0.786998276...
microsoft@Phi-3-mini-4k-instruct
Setting2
icl
answer.txt
zhaorui
0.504577
0.207202
0.29377
0.400923
0.189641
0.257487
{ "ID": { "precision": 0.6379661016949153, "recall": 0.5554899645808736, "f1": 0.5938781950142, "support": 6776 }, "CONTACT": { "precision": 0.11538461538461539, "recall": 0.0081374321880651, "f1": 0.015202702702702702, "support": 1106 }, ...
microsoft@Phi-3-mini-4k-instruct
Setting3
icl
answer.txt
zhaorui
0.866033
0.590933
0.702511
0.837363
0.554243
0.667003
{ "ID": { "precision": 0.8974398519432449, "recall": 0.7470792142765439, "f1": 0.8153856932670077, "support": 7789 }, "CONTACT": { "precision": 0.8721590909090909, "recall": 0.5803402646502835, "f1": 0.6969353007945518, "support": 529 }, ...
mistralai@Mistral-7B-Instruct-v0.3
Setting1
icl
answer.txt
zhaorui
0.539066
0.340247
0.417179
0.414726
0.258951
0.318829
{ "ID": { "precision": 0.40844444444444444, "recall": 0.2053401854541392, "f1": 0.2732882313582633, "support": 8951 }, "CONTACT": { "precision": 0.0, "recall": 0.0, "f1": 0.0, "support": 10 }, "DATE": { "precision": 0.7598395108...
mistralai@Mistral-7B-Instruct-v0.3
Setting2
icl
answer.txt
zhaorui
0.391887
0.290064
0.333374
0.342535
0.250306
0.289246
{ "ID": { "precision": 0.4313184003272988, "recall": 0.6223435655253837, "f1": 0.5095148915604423, "support": 6776 }, "CONTACT": { "precision": 0.1592920353982301, "recall": 0.0162748643761302, "f1": 0.029532403609515995, "support": 1106 },...
mistralai@Mistral-7B-Instruct-v0.3
Setting3
icl
answer.txt
zhaorui
0.840219
0.657688
0.737832
0.80437
0.634122
0.709172
{ "ID": { "precision": 0.8709163346613545, "recall": 0.8419566054692516, "f1": 0.8561916574188915, "support": 7789 }, "CONTACT": { "precision": 0.8289156626506025, "recall": 0.6502835538752363, "f1": 0.728813559322034, "support": 529 }, ...
zhaorui-nb@Llama-2-7b-hf._.lora_ft._.Setting1
Setting1
ft
answer.txt
zhaorui
0.80789
0.725791
0.764643
0.642001
0.642574
0.642287
{ "ID": { "precision": 0.89810681853163, "recall": 0.6518824712322645, "f1": 0.7554375970999482, "support": 8951 }, "CONTACT": { "precision": 0.3333333333333333, "recall": 0.6, "f1": 0.42857142857142855, "support": 10 }, "DATE": { ...
zhaorui-nb@Llama-2-7b-hf._.lora_ft._.Setting2
Setting2
ft
answer.txt
zhaorui
0.639199
0.461741
0.536168
0.595992
0.413827
0.488479
{ "ID": { "precision": 0.5254995836802664, "recall": 0.7451298701298701, "f1": 0.6163330078124999, "support": 6776 }, "CONTACT": { "precision": 0.9284064665127021, "recall": 0.36347197106690776, "f1": 0.5224171539961013, "support": 1106 }, ...
zhaorui-nb@Llama-2-7b-hf._.lora_ft._.Setting3
Setting3
ft
answer.txt
zhaorui
0.939451
0.91787
0.928535
0.927537
0.89448
0.910709
{ "ID": { "precision": 0.9340963400830085, "recall": 0.9535242007959943, "f1": 0.943710292249047, "support": 7789 }, "CONTACT": { "precision": 0.9544468546637744, "recall": 0.831758034026465, "f1": 0.888888888888889, "support": 529 }, "...
zhaorui-nb@Meta-Llama-3-8B-Instruct._.lora_ft._.Setting1
Setting1
ft
answer.txt
zhaorui
0.782024
0.747358
0.764298
0.587606
0.663966
0.623456
{ "ID": { "precision": 0.9190300623936472, "recall": 0.7240531784158195, "f1": 0.8099731300381179, "support": 8951 }, "CONTACT": { "precision": 0.12280701754385964, "recall": 0.7, "f1": 0.208955223880597, "support": 10 }, "DATE": { ...
zhaorui-nb@Meta-Llama-3-8B-Instruct._.lora_ft._.Setting2
Setting2
ft
answer.txt
zhaorui
0.604263
0.433603
0.504902
0.5665
0.415873
0.479639
{ "ID": { "precision": 0.4935290918977705, "recall": 0.6697166469893743, "f1": 0.5682800075136184, "support": 6776 }, "CONTACT": { "precision": 0.8682926829268293, "recall": 0.48282097649186256, "f1": 0.620569436374201, "support": 1106 }, ...
zhaorui-nb@Meta-Llama-3-8B-Instruct._.lora_ft._.Setting3
Setting3
ft
answer.txt
zhaorui
0.950231
0.939356
0.944762
0.938533
0.926599
0.932528
{ "ID": { "precision": 0.936260446551079, "recall": 0.9636667094620619, "f1": 0.9497659116791092, "support": 7789 }, "CONTACT": { "precision": 0.932806324110672, "recall": 0.8922495274102079, "f1": 0.9120772946859904, "support": 529 }, ...
zhaorui-nb@Meta-Llama-3-8B._.lora_ft._.Setting1
Setting1
ft
answer.txt
zhaorui
0.731445
0.72015
0.725754
0.603949
0.653613
0.6278
{ "ID": { "precision": 0.7390877192982456, "recall": 0.5883141548430343, "f1": 0.6551380940532471, "support": 8951 }, "CONTACT": { "precision": 0.5, "recall": 0.7, "f1": 0.5833333333333334, "support": 10 }, "DATE": { "precision"...
zhaorui-nb@Meta-Llama-3-8B._.lora_ft._.Setting2
Setting2
ft
answer.txt
zhaorui
0.604846
0.476516
0.533066
0.567618
0.450128
0.502091
{ "ID": { "precision": 0.5055170921678926, "recall": 0.6896399055489965, "f1": 0.5833957553058677, "support": 6776 }, "CONTACT": { "precision": 0.8285302593659942, "recall": 0.5198915009041591, "f1": 0.638888888888889, "support": 1106 }, ...
zhaorui-nb@Meta-Llama-3-8B._.lora_ft._.Setting3
Setting3
ft
answer.txt
zhaorui
0.949879
0.938348
0.944078
0.941295
0.928518
0.934863
{ "ID": { "precision": 0.9289779757485771, "recall": 0.9639234818333547, "f1": 0.9461281582761011, "support": 7789 }, "CONTACT": { "precision": 0.944, "recall": 0.8922495274102079, "f1": 0.9173955296404275, "support": 529 }, "DATE": { ...
zhaorui-nb@Mistral-7B-Instruct-v0.3._.lora_ft._.Setting1
Setting1
ft
answer.txt
zhaorui
0.856992
0.786872
0.820437
0.670888
0.647933
0.659211
{ "ID": { "precision": 0.9347982191584088, "recall": 0.7271813205228466, "f1": 0.8180218675380168, "support": 8951 }, "CONTACT": { "precision": 0.3076923076923077, "recall": 0.4, "f1": 0.34782608695652173, "support": 10 }, "DATE": { ...
zhaorui-nb@Mistral-7B-Instruct-v0.3._.lora_ft._.Setting2
Setting2
ft
answer.txt
zhaorui
0.572191
0.362503
0.443827
0.533084
0.358032
0.428364
{ "ID": { "precision": 0.4414388489208633, "recall": 0.6791617473435655, "f1": 0.5350851694668914, "support": 6776 }, "CONTACT": { "precision": 0.7743362831858407, "recall": 0.15822784810126583, "f1": 0.2627627627627628, "support": 1106 }, ...
zhaorui-nb@Mistral-7B-Instruct-v0.3._.lora_ft._.Setting3
Setting3
ft
answer.txt
zhaorui
0.954996
0.934682
0.94473
0.943861
0.922748
0.933185
{ "ID": { "precision": 0.9509366636931311, "recall": 0.9580177172936192, "f1": 0.9544640573036582, "support": 7789 }, "CONTACT": { "precision": 0.9387755102040817, "recall": 0.8695652173913043, "f1": 0.9028459273797841, "support": 529 }, ...
zhaorui-nb@Phi-3-mini-4k-instruct._.lora_ft._.Setting1
Setting1
ft
answer.txt
zhaorui
0.797989
0.723512
0.758928
0.592968
0.653247
0.62165
{ "ID": { "precision": 0.8848518725544997, "recall": 0.7074069936319964, "f1": 0.7862420065809896, "support": 8951 }, "CONTACT": { "precision": 0.28, "recall": 0.7, "f1": 0.4, "support": 10 }, "DATE": { "precision": 0.8538339631...
zhaorui-nb@Phi-3-mini-4k-instruct._.lora_ft._.Setting2
Setting2
ft
answer.txt
zhaorui
0.602849
0.583662
0.5931
0.561752
0.501562
0.529954
{ "ID": { "precision": 0.5282856528285653, "recall": 0.7249114521841794, "f1": 0.6111733233793704, "support": 6776 }, "CONTACT": { "precision": 0.851764705882353, "recall": 0.32730560578661844, "f1": 0.47289353363814507, "support": 1106 }, ...
zhaorui-nb@Phi-3-mini-4k-instruct._.lora_ft._.Setting3
Setting3
ft
answer.txt
zhaorui
0.927589
0.910768
0.919101
0.916656
0.891351
0.903827
{ "ID": { "precision": 0.9325942915392457, "recall": 0.9396584927461805, "f1": 0.9361130651659525, "support": 7789 }, "CONTACT": { "precision": 0.9388185654008439, "recall": 0.8412098298676749, "f1": 0.8873379860418744, "support": 529 }, ...
zhaorui-nb@Qwen1.5-7B-Chat._.lora_ft._.Setting1
Setting1
ft
answer.txt
zhaorui
0.715421
0.71055
0.712977
0.58507
0.647171
0.614556
{ "ID": { "precision": 0.862237080373311, "recall": 0.6915428443749302, "f1": 0.7675139491630503, "support": 8951 }, "CONTACT": { "precision": 0.3888888888888889, "recall": 0.7, "f1": 0.5, "support": 10 }, "DATE": { "precision":...
zhaorui-nb@Qwen1.5-7B-Chat._.lora_ft._.Setting2
Setting2
ft
answer.txt
zhaorui
0.563489
0.460958
0.507093
0.557387
0.374568
0.448046
{ "ID": { "precision": 0.3249097472924188, "recall": 0.6375442739079102, "f1": 0.43045037863690716, "support": 6776 }, "CONTACT": { "precision": 1.0, "recall": 0.0009041591320072332, "f1": 0.001806684733514002, "support": 1106 }, "DATE"...
zhaorui-nb@Qwen1.5-7B-Chat._.lora_ft._.Setting3
Setting3
ft
answer.txt
zhaorui
0.946103
0.932634
0.93932
0.930536
0.914584
0.922491
{ "ID": { "precision": 0.9327973872629066, "recall": 0.953395814610348, "f1": 0.942984126984127, "support": 7789 }, "CONTACT": { "precision": 0.9458333333333333, "recall": 0.8582230623818525, "f1": 0.8999008919722498, "support": 529 }, ...
zhaorui-nb@Yi-1.5-6B-Chat._.lora_ft._.Setting1
Setting1
ft
answer.txt
zhaorui
0.814126
0.779266
0.796314
0.59846
0.636761
0.617016
{ "ID": { "precision": 0.9259150374834632, "recall": 0.703720254720143, "f1": 0.7996699250983876, "support": 8951 }, "CONTACT": { "precision": 0.0547945205479452, "recall": 0.4, "f1": 0.09638554216867469, "support": 10 }, "DATE": { ...
zhaorui-nb@Yi-1.5-6B-Chat._.lora_ft._.Setting2
Setting2
ft
answer.txt
zhaorui
0.604842
0.466614
0.526812
0.56391
0.383105
0.456248
{ "ID": { "precision": 0.4680517853414074, "recall": 0.6615997638724912, "f1": 0.5482450776568424, "support": 6776 }, "CONTACT": { "precision": 0.8809523809523809, "recall": 0.06690777576853527, "f1": 0.12436974789915968, "support": 1106 },...
zhaorui-nb@Yi-1.5-6B-Chat._.lora_ft._.Setting3
Setting3
ft
answer.txt
zhaorui
0.93232
0.915706
0.923938
0.925996
0.905732
0.915752
{ "ID": { "precision": 0.9366515837104072, "recall": 0.956733855437155, "f1": 0.9465862178469355, "support": 7789 }, "CONTACT": { "precision": 0.9475890985324947, "recall": 0.8544423440453687, "f1": 0.8986083499005963, "support": 529 }, ...
zhaorui-nb@pythia-1b._.lora_ft._.Setting1
Setting1
ft
answer.txt
zhaorui
0.680503
0.588473
0.631151
0.445592
0.492792
0.468005
{ "ID": { "precision": 0.5646608797999118, "recall": 0.42877890738464974, "f1": 0.4874269748539497, "support": 8951 }, "CONTACT": { "precision": 0.017857142857142856, "recall": 0.1, "f1": 0.030303030303030304, "support": 10 }, "DATE": {...
zhaorui-nb@pythia-1b._.lora_ft._.Setting2
Setting2
ft
answer.txt
zhaorui
0.49663
0.344266
0.406645
0.379872
0.257412
0.306876
{ "ID": { "precision": 0.2760147141342976, "recall": 0.6533353010625738, "f1": 0.3880780188472496, "support": 6776 }, "CONTACT": { "precision": 0.0, "recall": 0.0, "f1": 0.0, "support": 1106 }, "DATE": { "precision": 0.603295589...
zhaorui-nb@pythia-1b._.lora_ft._.Setting3
Setting3
ft
answer.txt
zhaorui
0.887324
0.834319
0.860005
0.86639
0.796946
0.830219
{ "ID": { "precision": 0.8819513406156901, "recall": 0.9121838490178457, "f1": 0.8968128747238876, "support": 7789 }, "CONTACT": { "precision": 0.8789808917197452, "recall": 0.782608695652174, "f1": 0.828, "support": 529 }, "DATE": { ...
zhaorui-nb@pythia-2.8b._.lora_ft._.Setting1
Setting1
ft
answer.txt
zhaorui
0.67938
0.608985
0.642259
0.557119
0.515451
0.535476
{ "ID": { "precision": 0.5445768592547896, "recall": 0.40330689308457157, "f1": 0.4634146341463415, "support": 8951 }, "CONTACT": { "precision": 0.26666666666666666, "recall": 0.4, "f1": 0.32, "support": 10 }, "DATE": { "precisi...
zhaorui-nb@pythia-2.8b._.lora_ft._.Setting2
Setting2
ft
answer.txt
zhaorui
0.468348
0.399867
0.431407
0.317785
0.303954
0.310716
{ "ID": { "precision": 0.3203775680177679, "recall": 0.6812278630460449, "f1": 0.43580060422960726, "support": 6776 }, "CONTACT": { "precision": 0.0, "recall": 0.0, "f1": 0.0, "support": 1106 }, "DATE": { "precision": 0.60855189...
zhaorui-nb@pythia-2.8b._.lora_ft._.Setting3
Setting3
ft
answer.txt
zhaorui
0.910426
0.873212
0.891431
0.907407
0.84512
0.875156
{ "ID": { "precision": 0.9455119768360095, "recall": 0.9223263576839132, "f1": 0.9337752648339507, "support": 7789 }, "CONTACT": { "precision": 0.9186295503211992, "recall": 0.8109640831758034, "f1": 0.8614457831325301, "support": 529 }, ...

No dataset card yet

Downloads last month
3

Space using zhaorui-nb/leaderboard-score 1