craffel HF Staff commited on
Commit
1e71b5d
·
verified ·
1 Parent(s): 907c0a1

Upload nemotron_synthetic_1T_exp_7B/metrics.eval.jsonl with huggingface_hub

Browse files
nemotron_synthetic_1T_exp_7B/metrics.eval.jsonl ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ {"created_at": "2025-11-13T00:20:13.953184", "global_step": 30000, "commonsense_qa": {"alias": "commonsense_qa", "acc,none": 0.20147420147420148, "acc_stderr,none": 0.011483500195202905}, "hellaswag": {"alias": "hellaswag", "acc,none": 0.48137821151165106, "acc_stderr,none": 0.004986319587524967, "acc_norm,none": 0.6365265883290181, "acc_norm_stderr,none": 0.004800164434233253}, "mmlu": {"acc,none": 0.24177467597208374, "acc_stderr,none": 0.0036040777826654743, "alias": "mmlu"}, "mmlu_humanities": {"acc,none": 0.25249734325185974, "acc_stderr,none": 0.006328666341290716, "alias": " - humanities"}, "mmlu_formal_logic": {"alias": " - formal_logic", "acc,none": 0.2777777777777778, "acc_stderr,none": 0.04006168083848878}, "mmlu_high_school_european_history": {"alias": " - high_school_european_history", "acc,none": 0.24242424242424243, "acc_stderr,none": 0.03346409881055953}, "mmlu_high_school_us_history": {"alias": " - high_school_us_history", "acc,none": 0.25980392156862747, "acc_stderr,none": 0.03077855467869326}, "mmlu_high_school_world_history": {"alias": " - high_school_world_history", "acc,none": 0.29535864978902954, "acc_stderr,none": 0.029696338713422882}, "mmlu_international_law": {"alias": " - international_law", "acc,none": 0.2727272727272727, "acc_stderr,none": 0.04065578140908705}, "mmlu_jurisprudence": {"alias": " - jurisprudence", "acc,none": 0.26851851851851855, "acc_stderr,none": 0.04284467968052191}, "mmlu_logical_fallacies": {"alias": " - logical_fallacies", "acc,none": 0.24539877300613497, "acc_stderr,none": 0.03380939813943354}, "mmlu_moral_disputes": {"alias": " - moral_disputes", "acc,none": 0.2774566473988439, "acc_stderr,none": 0.024105712607754307}, "mmlu_moral_scenarios": {"alias": " - moral_scenarios", "acc,none": 0.23798882681564246, "acc_stderr,none": 0.014242630070574885}, "mmlu_philosophy": {"alias": " - philosophy", "acc,none": 0.19292604501607716, "acc_stderr,none": 0.022411516780911363}, "mmlu_prehistory": {"alias": " - prehistory", "acc,none": 0.25, "acc_stderr,none": 0.02409347123262133}, "mmlu_professional_law": {"alias": " - professional_law", "acc,none": 0.24641460234680573, "acc_stderr,none": 0.011005971399927239}, "mmlu_world_religions": {"alias": " - world_religions", "acc,none": 0.3508771929824561, "acc_stderr,none": 0.036602988340491624}, "mmlu_other": {"acc,none": 0.2603797875764403, "acc_stderr,none": 0.007844141249228068, "alias": " - other"}, "mmlu_business_ethics": {"alias": " - business_ethics", "acc,none": 0.35, "acc_stderr,none": 0.0479372485441102}, "mmlu_clinical_knowledge": {"alias": " - clinical_knowledge", "acc,none": 0.22264150943396227, "acc_stderr,none": 0.025604233470899095}, "mmlu_college_medicine": {"alias": " - college_medicine", "acc,none": 0.21965317919075145, "acc_stderr,none": 0.031568093627031744}, "mmlu_global_facts": {"alias": " - global_facts", "acc,none": 0.17, "acc_stderr,none": 0.0377525168068637}, "mmlu_human_aging": {"alias": " - human_aging", "acc,none": 0.34080717488789236, "acc_stderr,none": 0.031811497470553604}, "mmlu_management": {"alias": " - management", "acc,none": 0.24271844660194175, "acc_stderr,none": 0.04245022486384493}, "mmlu_marketing": {"alias": " - marketing", "acc,none": 0.32051282051282054, "acc_stderr,none": 0.03057281131029961}, "mmlu_medical_genetics": {"alias": " - medical_genetics", "acc,none": 0.3, "acc_stderr,none": 0.046056618647183814}, "mmlu_miscellaneous": {"alias": " - miscellaneous", "acc,none": 0.2656449553001277, "acc_stderr,none": 0.015794302487888726}, "mmlu_nutrition": {"alias": " - nutrition", "acc,none": 0.24836601307189543, "acc_stderr,none": 0.02473998135511359}, "mmlu_professional_accounting": {"alias": " - professional_accounting", "acc,none": 0.24468085106382978, "acc_stderr,none": 0.02564555362226673}, "mmlu_professional_medicine": {"alias": " - professional_medicine", "acc,none": 0.18382352941176472, "acc_stderr,none": 0.023529242185193106}, "mmlu_virology": {"alias": " - virology", "acc,none": 0.3072289156626506, "acc_stderr,none": 0.03591566797824662}, "mmlu_social_sciences": {"acc,none": 0.22716932076698082, "acc_stderr,none": 0.007542251453874756, "alias": " - social sciences"}, "mmlu_econometrics": {"alias": " - econometrics", "acc,none": 0.23684210526315788, "acc_stderr,none": 0.03999423879281335}, "mmlu_high_school_geography": {"alias": " - high_school_geography", "acc,none": 0.20202020202020202, "acc_stderr,none": 0.02860620428922989}, "mmlu_high_school_government_and_politics": {"alias": " - high_school_government_and_politics", "acc,none": 0.20725388601036268, "acc_stderr,none": 0.02925282329180362}, "mmlu_high_school_macroeconomics": {"alias": " - high_school_macroeconomics", "acc,none": 0.2076923076923077, "acc_stderr,none": 0.02056753956724679}, "mmlu_high_school_microeconomics": {"alias": " - high_school_microeconomics", "acc,none": 0.21008403361344538, "acc_stderr,none": 0.026461398717471874}, "mmlu_high_school_psychology": {"alias": " - high_school_psychology", "acc,none": 0.1926605504587156, "acc_stderr,none": 0.016909276884936094}, "mmlu_human_sexuality": {"alias": " - human_sexuality", "acc,none": 0.2824427480916031, "acc_stderr,none": 0.03948406125768361}, "mmlu_professional_psychology": {"alias": " - professional_psychology", "acc,none": 0.26143790849673204, "acc_stderr,none": 0.017776947157528044}, "mmlu_public_relations": {"alias": " - public_relations", "acc,none": 0.23636363636363636, "acc_stderr,none": 0.04069306319721375}, "mmlu_security_studies": {"alias": " - security_studies", "acc,none": 0.19183673469387755, "acc_stderr,none": 0.025206963154225423}, "mmlu_sociology": {"alias": " - sociology", "acc,none": 0.26865671641791045, "acc_stderr,none": 0.03134328358208954}, "mmlu_us_foreign_policy": {"alias": " - us_foreign_policy", "acc,none": 0.32, "acc_stderr,none": 0.04688261722621504}, "mmlu_stem": {"acc,none": 0.22169362511893434, "acc_stderr,none": 0.00738316747225605, "alias": " - stem"}, "mmlu_abstract_algebra": {"alias": " - abstract_algebra", "acc,none": 0.22, "acc_stderr,none": 0.04163331998932268}, "mmlu_anatomy": {"alias": " - anatomy", "acc,none": 0.2074074074074074, "acc_stderr,none": 0.03502553170678317}, "mmlu_astronomy": {"alias": " - astronomy", "acc,none": 0.17763157894736842, "acc_stderr,none": 0.031103182383123398}, "mmlu_college_biology": {"alias": " - college_biology", "acc,none": 0.2638888888888889, "acc_stderr,none": 0.03685651095897532}, "mmlu_college_chemistry": {"alias": " - college_chemistry", "acc,none": 0.19, "acc_stderr,none": 0.039427724440366234}, "mmlu_college_computer_science": {"alias": " - college_computer_science", "acc,none": 0.28, "acc_stderr,none": 0.04512608598542128}, "mmlu_college_mathematics": {"alias": " - college_mathematics", "acc,none": 0.21, "acc_stderr,none": 0.040936018074033256}, "mmlu_college_physics": {"alias": " - college_physics", "acc,none": 0.23529411764705882, "acc_stderr,none": 0.04220773659171453}, "mmlu_computer_security": {"alias": " - computer_security", "acc,none": 0.28, "acc_stderr,none": 0.04512608598542128}, "mmlu_conceptual_physics": {"alias": " - conceptual_physics", "acc,none": 0.2851063829787234, "acc_stderr,none": 0.029513196625539355}, "mmlu_electrical_engineering": {"alias": " - electrical_engineering", "acc,none": 0.25517241379310346, "acc_stderr,none": 0.03632984052707842}, "mmlu_elementary_mathematics": {"alias": " - elementary_mathematics", "acc,none": 0.21164021164021163, "acc_stderr,none": 0.021037331505262893}, "mmlu_high_school_biology": {"alias": " - high_school_biology", "acc,none": 0.2129032258064516, "acc_stderr,none": 0.02328766512726853}, "mmlu_high_school_chemistry": {"alias": " - high_school_chemistry", "acc,none": 0.15763546798029557, "acc_stderr,none": 0.025639014131172404}, "mmlu_high_school_computer_science": {"alias": " - high_school_computer_science", "acc,none": 0.25, "acc_stderr,none": 0.04351941398892446}, "mmlu_high_school_mathematics": {"alias": " - high_school_mathematics", "acc,none": 0.21481481481481482, "acc_stderr,none": 0.02504044387700069}, "mmlu_high_school_physics": {"alias": " - high_school_physics", "acc,none": 0.1986754966887417, "acc_stderr,none": 0.032578473844367774}, "mmlu_high_school_statistics": {"alias": " - high_school_statistics", "acc,none": 0.1574074074074074, "acc_stderr,none": 0.02483717351824239}, "mmlu_machine_learning": {"alias": " - machine_learning", "acc,none": 0.3125, "acc_stderr,none": 0.043994650575715215}, "sciq": {"alias": "sciq", "acc,none": 0.931, "acc_stderr,none": 0.008018934050315167, "acc_norm,none": 0.906, "acc_norm_stderr,none": 0.009233052000787736}}
2
+ {"created_at": "2025-11-17T17:38:01.343222", "global_step": 150000, "commonsense_qa": {"alias": "commonsense_qa", "acc,none": 0.579033579033579, "acc_stderr,none": 0.014134995908456715}, "hellaswag": {"alias": "hellaswag", "acc,none": 0.5781716789484167, "acc_stderr,none": 0.004928420903026554, "acc_norm,none": 0.7637920732921729, "acc_norm_stderr,none": 0.0042388331660634816}, "mmlu": {"acc,none": 0.5566158666856573, "acc_stderr,none": 0.0039265686575764005, "alias": "mmlu"}, "mmlu_humanities": {"acc,none": 0.5032943676939426, "acc_stderr,none": 0.0067606516264965165, "alias": " - humanities"}, "mmlu_formal_logic": {"alias": " - formal_logic", "acc,none": 0.30158730158730157, "acc_stderr,none": 0.04104947269903394}, "mmlu_high_school_european_history": {"alias": " - high_school_european_history", "acc,none": 0.696969696969697, "acc_stderr,none": 0.035886248000917075}, "mmlu_high_school_us_history": {"alias": " - high_school_us_history", "acc,none": 0.7156862745098039, "acc_stderr,none": 0.031660096793998116}, "mmlu_high_school_world_history": {"alias": " - high_school_world_history", "acc,none": 0.7383966244725738, "acc_stderr,none": 0.028609516716994934}, "mmlu_international_law": {"alias": " - international_law", "acc,none": 0.7024793388429752, "acc_stderr,none": 0.041733491480835}, "mmlu_jurisprudence": {"alias": " - jurisprudence", "acc,none": 0.6666666666666666, "acc_stderr,none": 0.04557239513497751}, "mmlu_logical_fallacies": {"alias": " - logical_fallacies", "acc,none": 0.6932515337423313, "acc_stderr,none": 0.03623089915724147}, "mmlu_moral_disputes": {"alias": " - moral_disputes", "acc,none": 0.653179190751445, "acc_stderr,none": 0.025624723994030457}, "mmlu_moral_scenarios": {"alias": " - moral_scenarios", "acc,none": 0.2424581005586592, "acc_stderr,none": 0.014333522059217892}, "mmlu_philosophy": {"alias": " - philosophy", "acc,none": 0.684887459807074, "acc_stderr,none": 0.02638527370346449}, "mmlu_prehistory": {"alias": " - prehistory", "acc,none": 0.6944444444444444, "acc_stderr,none": 0.02563082497562135}, "mmlu_professional_law": {"alias": " - professional_law", "acc,none": 0.3983050847457627, "acc_stderr,none": 0.012503310565166254}, "mmlu_world_religions": {"alias": " - world_religions", "acc,none": 0.7719298245614035, "acc_stderr,none": 0.032180937956023566}, "mmlu_other": {"acc,none": 0.6337302864499518, "acc_stderr,none": 0.008312173846022779, "alias": " - other"}, "mmlu_business_ethics": {"alias": " - business_ethics", "acc,none": 0.57, "acc_stderr,none": 0.04975698519562428}, "mmlu_clinical_knowledge": {"alias": " - clinical_knowledge", "acc,none": 0.6188679245283019, "acc_stderr,none": 0.029890609686286648}, "mmlu_college_medicine": {"alias": " - college_medicine", "acc,none": 0.5202312138728323, "acc_stderr,none": 0.03809342081273957}, "mmlu_global_facts": {"alias": " - global_facts", "acc,none": 0.39, "acc_stderr,none": 0.04902071300001975}, "mmlu_human_aging": {"alias": " - human_aging", "acc,none": 0.6636771300448431, "acc_stderr,none": 0.031708824268455}, "mmlu_management": {"alias": " - management", "acc,none": 0.7572815533980582, "acc_stderr,none": 0.04245022486384495}, "mmlu_marketing": {"alias": " - marketing", "acc,none": 0.8205128205128205, "acc_stderr,none": 0.02514093595033546}, "mmlu_medical_genetics": {"alias": " - medical_genetics", "acc,none": 0.68, "acc_stderr,none": 0.04688261722621505}, "mmlu_miscellaneous": {"alias": " - miscellaneous", "acc,none": 0.7726692209450831, "acc_stderr,none": 0.014987270640946012}, "mmlu_nutrition": {"alias": " - nutrition", "acc,none": 0.6372549019607843, "acc_stderr,none": 0.027530078447110314}, "mmlu_professional_accounting": {"alias": " - professional_accounting", "acc,none": 0.3971631205673759, "acc_stderr,none": 0.0291898056735871}, "mmlu_professional_medicine": {"alias": " - professional_medicine", "acc,none": 0.5073529411764706, "acc_stderr,none": 0.030369552523902173}, "mmlu_virology": {"alias": " - virology", "acc,none": 0.5, "acc_stderr,none": 0.03892494720807614}, "mmlu_social_sciences": {"acc,none": 0.6603834904127397, "acc_stderr,none": 0.008260740874851987, "alias": " - social sciences"}, "mmlu_econometrics": {"alias": " - econometrics", "acc,none": 0.30701754385964913, "acc_stderr,none": 0.0433913832257986}, "mmlu_high_school_geography": {"alias": " - high_school_geography", "acc,none": 0.7272727272727273, "acc_stderr,none": 0.03173071239071724}, "mmlu_high_school_government_and_politics": {"alias": " - high_school_government_and_politics", "acc,none": 0.7979274611398963, "acc_stderr,none": 0.02897908979429673}, "mmlu_high_school_macroeconomics": {"alias": " - high_school_macroeconomics", "acc,none": 0.5435897435897435, "acc_stderr,none": 0.025254485424799602}, "mmlu_high_school_microeconomics": {"alias": " - high_school_microeconomics", "acc,none": 0.6176470588235294, "acc_stderr,none": 0.03156663099215416}, "mmlu_high_school_psychology": {"alias": " - high_school_psychology", "acc,none": 0.7944954128440367, "acc_stderr,none": 0.017324352325016}, "mmlu_human_sexuality": {"alias": " - human_sexuality", "acc,none": 0.6946564885496184, "acc_stderr,none": 0.040393149787245605}, "mmlu_professional_psychology": {"alias": " - professional_psychology", "acc,none": 0.5571895424836601, "acc_stderr,none": 0.020095083154577347}, "mmlu_public_relations": {"alias": " - public_relations", "acc,none": 0.6272727272727273, "acc_stderr,none": 0.04631381319425464}, "mmlu_security_studies": {"alias": " - security_studies", "acc,none": 0.6857142857142857, "acc_stderr,none": 0.02971932942241748}, "mmlu_sociology": {"alias": " - sociology", "acc,none": 0.8059701492537313, "acc_stderr,none": 0.02796267760476892}, "mmlu_us_foreign_policy": {"alias": " - us_foreign_policy", "acc,none": 0.76, "acc_stderr,none": 0.04292346959909283}, "mmlu_stem": {"acc,none": 0.4589280050745322, "acc_stderr,none": 0.008482301640737499, "alias": " - stem"}, "mmlu_abstract_algebra": {"alias": " - abstract_algebra", "acc,none": 0.22, "acc_stderr,none": 0.04163331998932269}, "mmlu_anatomy": {"alias": " - anatomy", "acc,none": 0.5481481481481482, "acc_stderr,none": 0.04299268905480863}, "mmlu_astronomy": {"alias": " - astronomy", "acc,none": 0.631578947368421, "acc_stderr,none": 0.039255233810529325}, "mmlu_college_biology": {"alias": " - college_biology", "acc,none": 0.6458333333333334, "acc_stderr,none": 0.039994111357535424}, "mmlu_college_chemistry": {"alias": " - college_chemistry", "acc,none": 0.32, "acc_stderr,none": 0.046882617226215034}, "mmlu_college_computer_science": {"alias": " - college_computer_science", "acc,none": 0.48, "acc_stderr,none": 0.050211673156867795}, "mmlu_college_mathematics": {"alias": " - college_mathematics", "acc,none": 0.36, "acc_stderr,none": 0.04824181513244218}, "mmlu_college_physics": {"alias": " - college_physics", "acc,none": 0.2549019607843137, "acc_stderr,none": 0.04336432707993177}, "mmlu_computer_security": {"alias": " - computer_security", "acc,none": 0.74, "acc_stderr,none": 0.044084400227680794}, "mmlu_conceptual_physics": {"alias": " - conceptual_physics", "acc,none": 0.502127659574468, "acc_stderr,none": 0.03268572658667492}, "mmlu_electrical_engineering": {"alias": " - electrical_engineering", "acc,none": 0.5379310344827586, "acc_stderr,none": 0.04154659671707548}, "mmlu_elementary_mathematics": {"alias": " - elementary_mathematics", "acc,none": 0.31216931216931215, "acc_stderr,none": 0.023865206836972602}, "mmlu_high_school_biology": {"alias": " - high_school_biology", "acc,none": 0.7032258064516129, "acc_stderr,none": 0.02598850079241189}, "mmlu_high_school_chemistry": {"alias": " - high_school_chemistry", "acc,none": 0.46798029556650245, "acc_stderr,none": 0.035107665979592154}, "mmlu_high_school_computer_science": {"alias": " - high_school_computer_science", "acc,none": 0.55, "acc_stderr,none": 0.05}, "mmlu_high_school_mathematics": {"alias": " - high_school_mathematics", "acc,none": 0.2740740740740741, "acc_stderr,none": 0.027195934804085622}, "mmlu_high_school_physics": {"alias": " - high_school_physics", "acc,none": 0.3443708609271523, "acc_stderr,none": 0.03879687024073327}, "mmlu_high_school_statistics": {"alias": " - high_school_statistics", "acc,none": 0.4212962962962963, "acc_stderr,none": 0.03367462138896078}, "mmlu_machine_learning": {"alias": " - machine_learning", "acc,none": 0.41964285714285715, "acc_stderr,none": 0.04684099321077106}, "sciq": {"alias": "sciq", "acc,none": 0.956, "acc_stderr,none": 0.006488921798427414, "acc_norm,none": 0.938, "acc_norm_stderr,none": 0.007629823996280304}}