task_id stringlengths 8 69 | name stringlengths 3 64 | suite stringclasses 6
values | hf_repo stringclasses 125
values | hf_subset stringlengths 0 55 | file_path stringclasses 6
values | line_number int64 36 22.8k | variable_name stringclasses 1
value | is_subtask bool 2
classes | main_task stringclasses 122
values | subtask_count int64 1 1 | suites listlengths 1 1 | subtasks listlengths 1 1 | is_standalone bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
lighteval:bigbench:persian_idioms | bigbench:persian_idioms | lighteval | tasksource/bigbench | persian_idioms | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,113 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:persian_idioms"
] | false |
lighteval:bigbench:phrase_relatedness | bigbench:phrase_relatedness | lighteval | tasksource/bigbench | phrase_relatedness | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,128 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:phrase_relatedness"
] | false |
lighteval:bigbench:physical_intuition | bigbench:physical_intuition | lighteval | tasksource/bigbench | physical_intuition | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,143 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:physical_intuition"
] | false |
lighteval:bigbench:physics | bigbench:physics | lighteval | tasksource/bigbench | physics | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,158 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:physics"
] | false |
lighteval:bigbench:physics_questions | bigbench:physics_questions | lighteval | tasksource/bigbench | physics_questions | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,173 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:physics_questions"
] | false |
lighteval:piqa | piqa | lighteval | ybisk/piqa | plain_text | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,188 | direct_call | false | piqa | 1 | [
"lighteval"
] | [
"lighteval:piqa"
] | true |
helm:piqa | piqa | helm | ybisk/piqa | plain_text | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,206 | direct_call | false | piqa | 1 | [
"helm"
] | [
"helm:piqa"
] | true |
lighteval:bigbench_lite:play_dialog_same_or_different | bigbench_lite:play_dialog_same_or_different | lighteval | tasksource/bigbench | play_dialog_same_or_different | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,232 | direct_call | true | bigbench_lite | 1 | [
"lighteval"
] | [
"lighteval:bigbench_lite:play_dialog_same_or_different"
] | false |
lighteval:bigbench:polish_sequence_labeling | bigbench:polish_sequence_labeling | lighteval | tasksource/bigbench | polish_sequence_labeling | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,247 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:polish_sequence_labeling"
] | false |
lighteval:bigbench:presuppositions_as_nli | bigbench:presuppositions_as_nli | lighteval | tasksource/bigbench | presuppositions_as_nli | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,262 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:presuppositions_as_nli"
] | false |
lighteval:prost | prost | lighteval | lighteval/prost | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,277 | direct_call | false | prost | 1 | [
"lighteval"
] | [
"lighteval:prost"
] | true |
lighteval:pubmedqa | pubmedqa | lighteval | pubmed_qa | pqa_labeled | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,295 | direct_call | false | pubmedqa | 1 | [
"lighteval"
] | [
"lighteval:pubmedqa"
] | true |
helm:pubmedqa | pubmedqa | helm | pubmed_qa | pqa_labeled | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,310 | direct_call | false | pubmedqa | 1 | [
"helm"
] | [
"helm:pubmedqa"
] | true |
lighteval:qa4mre:2011 | qa4mre:2011 | lighteval | qa4mre | 2011.main.EN | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,336 | direct_call | true | qa4mre | 1 | [
"lighteval"
] | [
"lighteval:qa4mre:2011"
] | false |
lighteval:qa4mre:2012 | qa4mre:2012 | lighteval | qa4mre | 2012.main.EN | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,354 | direct_call | true | qa4mre | 1 | [
"lighteval"
] | [
"lighteval:qa4mre:2012"
] | false |
lighteval:qa4mre:2013 | qa4mre:2013 | lighteval | qa4mre | 2013.main.EN | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,372 | direct_call | true | qa4mre | 1 | [
"lighteval"
] | [
"lighteval:qa4mre:2013"
] | false |
lighteval:bigbench:qa_wikidata | bigbench:qa_wikidata | lighteval | tasksource/bigbench | qa_wikidata | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,390 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:qa_wikidata"
] | false |
lighteval:qasper | qasper | lighteval | allenai/qasper | qasper | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,410 | direct_call | false | qasper | 1 | [
"lighteval"
] | [
"lighteval:qasper"
] | true |
lighteval:qasper_ll | qasper_ll | lighteval | allenai/qasper | qasper | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,425 | direct_call | false | qasper_ll | 1 | [
"lighteval"
] | [
"lighteval:qasper_ll"
] | true |
helm:quac | quac | helm | lighteval/quac_helm | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,440 | direct_call | false | quac | 1 | [
"helm"
] | [
"helm:quac"
] | true |
lighteval:bigbench:question_selection | bigbench:question_selection | lighteval | tasksource/bigbench | question_selection | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,459 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:question_selection"
] | false |
lighteval:race:high | race:high | lighteval | EleutherAI/race | high | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,474 | direct_call | true | race | 1 | [
"lighteval"
] | [
"lighteval:race:high"
] | false |
helm:raft:ade_corpus_v2 | raft:ade_corpus_v2 | helm | ought/raft | ade_corpus_v2 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,489 | direct_call | true | raft | 1 | [
"helm"
] | [
"helm:raft:ade_corpus_v2"
] | false |
helm:raft:banking_77 | raft:banking_77 | helm | ought/raft | banking_77 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,517 | direct_call | true | raft | 1 | [
"helm"
] | [
"helm:raft:banking_77"
] | false |
helm:raft:neurips_impact_statement_risks | raft:neurips_impact_statement_risks | helm | ought/raft | neurips_impact_statement_risks | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,545 | direct_call | true | raft | 1 | [
"helm"
] | [
"helm:raft:neurips_impact_statement_risks"
] | false |
helm:raft:one_stop_english | raft:one_stop_english | helm | ought/raft | one_stop_english | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,573 | direct_call | true | raft | 1 | [
"helm"
] | [
"helm:raft:one_stop_english"
] | false |
helm:raft:overruling | raft:overruling | helm | ought/raft | overruling | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,601 | direct_call | true | raft | 1 | [
"helm"
] | [
"helm:raft:overruling"
] | false |
helm:raft:semiconductor_org_types | raft:semiconductor_org_types | helm | ought/raft | semiconductor_org_types | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,629 | direct_call | true | raft | 1 | [
"helm"
] | [
"helm:raft:semiconductor_org_types"
] | false |
helm:raft:systematic_review_inclusion | raft:systematic_review_inclusion | helm | ought/raft | systematic_review_inclusion | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,657 | direct_call | true | raft | 1 | [
"helm"
] | [
"helm:raft:systematic_review_inclusion"
] | false |
helm:raft:tai_safety_research | raft:tai_safety_research | helm | ought/raft | tai_safety_research | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,685 | direct_call | true | raft | 1 | [
"helm"
] | [
"helm:raft:tai_safety_research"
] | false |
helm:raft:terms_of_service | raft:terms_of_service | helm | ought/raft | terms_of_service | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,713 | direct_call | true | raft | 1 | [
"helm"
] | [
"helm:raft:terms_of_service"
] | false |
helm:raft:tweet_eval_hate | raft:tweet_eval_hate | helm | ought/raft | tweet_eval_hate | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,741 | direct_call | true | raft | 1 | [
"helm"
] | [
"helm:raft:tweet_eval_hate"
] | false |
helm:raft:twitter_complaints | raft:twitter_complaints | helm | ought/raft | twitter_complaints | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,769 | direct_call | true | raft | 1 | [
"helm"
] | [
"helm:raft:twitter_complaints"
] | false |
lighteval:bigbench:real_or_fake_text | bigbench:real_or_fake_text | lighteval | tasksource/bigbench | real_or_fake_text | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,797 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:real_or_fake_text"
] | false |
helm:real_toxicity_prompts | real_toxicity_prompts | helm | allenai/real-toxicity-prompts | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,812 | direct_call | false | real_toxicity_prompts | 1 | [
"helm"
] | [
"helm:real_toxicity_prompts"
] | true |
lighteval:bigbench_lite:repeat_copy_logic | bigbench_lite:repeat_copy_logic | lighteval | tasksource/bigbench | repeat_copy_logic | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,842 | direct_call | true | bigbench_lite | 1 | [
"lighteval"
] | [
"lighteval:bigbench_lite:repeat_copy_logic"
] | false |
lighteval:bigbench:rephrase | bigbench:rephrase | lighteval | tasksource/bigbench | rephrase | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,857 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:rephrase"
] | false |
lighteval:bigbench:rhyming | bigbench:rhyming | lighteval | tasksource/bigbench | rhyming | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,877 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:rhyming"
] | false |
lighteval:bigbench:riddle_sense | bigbench:riddle_sense | lighteval | tasksource/bigbench | riddle_sense | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,892 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:riddle_sense"
] | false |
lighteval:bigbench:scientific_press_release | bigbench:scientific_press_release | lighteval | tasksource/bigbench | scientific_press_release | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,937 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:scientific_press_release"
] | false |
lighteval:sciq | sciq | lighteval | sciq | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,952 | direct_call | false | sciq | 1 | [
"lighteval"
] | [
"lighteval:sciq"
] | true |
lighteval:bigbench:semantic_parsing_in_context_sparc | bigbench:semantic_parsing_in_context_sparc | lighteval | tasksource/bigbench | semantic_parsing_in_context_sparc | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,970 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:semantic_parsing_in_context_sparc"
] | false |
lighteval:bigbench:semantic_parsing_spider | bigbench:semantic_parsing_spider | lighteval | tasksource/bigbench | semantic_parsing_spider | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 15,985 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:semantic_parsing_spider"
] | false |
lighteval:bigbench:sentence_ambiguity | bigbench:sentence_ambiguity | lighteval | tasksource/bigbench | sentence_ambiguity | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,000 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:sentence_ambiguity"
] | false |
lighteval:bigbench:similarities_abstraction | bigbench:similarities_abstraction | lighteval | tasksource/bigbench | similarities_abstraction | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,015 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:similarities_abstraction"
] | false |
lighteval:bigbench:simp_turing_concept | bigbench:simp_turing_concept | lighteval | tasksource/bigbench | simp_turing_concept | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,030 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:simp_turing_concept"
] | false |
lighteval:simpleqa | simpleqa | lighteval | lighteval/SimpleQA | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,045 | direct_call | false | simpleqa | 1 | [
"lighteval"
] | [
"lighteval:simpleqa"
] | true |
lighteval:bigbench:simple_arithmetic_json | bigbench:simple_arithmetic_json | lighteval | tasksource/bigbench | simple_arithmetic_json | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,060 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:simple_arithmetic_json"
] | false |
lighteval:bigbench:simple_arithmetic_json_multiple_choice | bigbench:simple_arithmetic_json_multiple_choice | lighteval | tasksource/bigbench | simple_arithmetic_json_multiple_choice | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,075 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:simple_arithmetic_json_multiple_choice"
] | false |
lighteval:bigbench:simple_arithmetic_json_subtasks | bigbench:simple_arithmetic_json_subtasks | lighteval | tasksource/bigbench | simple_arithmetic_json_subtasks | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,090 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:simple_arithmetic_json_subtasks"
] | false |
lighteval:bigbench:simple_arithmetic_multiple_targets_json | bigbench:simple_arithmetic_multiple_targets_json | lighteval | tasksource/bigbench | simple_arithmetic_multiple_targets_json | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,105 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:simple_arithmetic_multiple_targets_json"
] | false |
lighteval:bigbench:simple_ethical_questions | bigbench:simple_ethical_questions | lighteval | tasksource/bigbench | simple_ethical_questions | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,120 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:simple_ethical_questions"
] | false |
lighteval:bigbench:simple_text_editing | bigbench:simple_text_editing | lighteval | tasksource/bigbench | simple_text_editing | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,135 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:simple_text_editing"
] | false |
helm:siqa | siqa | helm | allenai/social_i_qa | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,150 | direct_call | false | siqa | 1 | [
"helm"
] | [
"helm:siqa"
] | true |
lighteval:bigbench:social_iqa | bigbench:social_iqa | lighteval | tasksource/bigbench | social_iqa | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,191 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:social_iqa"
] | false |
lighteval:bigbench:social_support | bigbench:social_support | lighteval | tasksource/bigbench | social_support | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,206 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:social_support"
] | false |
lighteval:squad_v2 | squad_v2 | lighteval | rajpurkar/squad_v2 | squad_v2 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,236 | direct_call | false | squad_v2 | 1 | [
"lighteval"
] | [
"lighteval:squad_v2"
] | true |
lighteval:storycloze:2016 | storycloze:2016 | lighteval | MoE-UNC/story_cloze | 2016 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,265 | direct_call | true | storycloze | 1 | [
"lighteval"
] | [
"lighteval:storycloze:2016"
] | false |
lighteval:storycloze:2018 | storycloze:2018 | lighteval | MoE-UNC/story_cloze | 2018 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,280 | direct_call | true | storycloze | 1 | [
"lighteval"
] | [
"lighteval:storycloze:2018"
] | false |
lighteval:bigbench_lite:strange_stories | bigbench_lite:strange_stories | lighteval | tasksource/bigbench | strange_stories | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,295 | direct_call | true | bigbench_lite | 1 | [
"lighteval"
] | [
"lighteval:bigbench_lite:strange_stories"
] | false |
lighteval:bigbench_lite:strategyqa | bigbench_lite:strategyqa | lighteval | tasksource/bigbench | strategyqa | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,310 | direct_call | true | bigbench_lite | 1 | [
"lighteval"
] | [
"lighteval:bigbench_lite:strategyqa"
] | false |
lighteval:bigbench:sufficient_information | bigbench:sufficient_information | lighteval | tasksource/bigbench | sufficient_information | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,325 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:sufficient_information"
] | false |
lighteval:bigbench:suicide_risk | bigbench:suicide_risk | lighteval | tasksource/bigbench | suicide_risk | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,340 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:suicide_risk"
] | false |
helm:summarization:cnn-dm | summarization:cnn-dm | helm | lighteval/summarization | cnn-dm | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,355 | direct_call | true | summarization | 1 | [
"helm"
] | [
"helm:summarization:cnn-dm"
] | false |
helm:summarization:xsum | summarization:xsum | helm | lighteval/summarization | xsum | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,377 | direct_call | true | summarization | 1 | [
"helm"
] | [
"helm:summarization:xsum"
] | false |
helm:summarization:xsum-sampled | summarization:xsum-sampled | helm | lighteval/summarization | xsum-sampled | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,399 | direct_call | true | summarization | 1 | [
"helm"
] | [
"helm:summarization:xsum-sampled"
] | false |
lighteval:super_glue:boolq | super_glue:boolq | lighteval | super_glue | boolq | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,421 | direct_call | true | super_glue | 1 | [
"lighteval"
] | [
"lighteval:super_glue:boolq"
] | false |
lighteval:super_glue:cb | super_glue:cb | lighteval | super_glue | cb | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,436 | direct_call | true | super_glue | 1 | [
"lighteval"
] | [
"lighteval:super_glue:cb"
] | false |
lighteval:super_glue:copa | super_glue:copa | lighteval | super_glue | copa | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,451 | direct_call | true | super_glue | 1 | [
"lighteval"
] | [
"lighteval:super_glue:copa"
] | false |
lighteval:super_glue:rte | super_glue:rte | lighteval | super_glue | rte | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,466 | direct_call | true | super_glue | 1 | [
"lighteval"
] | [
"lighteval:super_glue:rte"
] | false |
lighteval:super_glue:multirc | super_glue:multirc | lighteval | super_glue | multirc | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,481 | direct_call | true | super_glue | 1 | [
"lighteval"
] | [
"lighteval:super_glue:multirc"
] | false |
lighteval:super_glue:wic | super_glue:wic | lighteval | super_glue | wic | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,496 | direct_call | true | super_glue | 1 | [
"lighteval"
] | [
"lighteval:super_glue:wic"
] | false |
lighteval:super_glue:wsc | super_glue:wsc | lighteval | super_glue | wsc | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,511 | direct_call | true | super_glue | 1 | [
"lighteval"
] | [
"lighteval:super_glue:wsc"
] | false |
lighteval:bigbench:swahili_english_proverbs | bigbench:swahili_english_proverbs | lighteval | tasksource/bigbench | swahili_english_proverbs | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,526 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:swahili_english_proverbs"
] | false |
lighteval:swag | swag | lighteval | swag | regular | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,541 | direct_call | false | swag | 1 | [
"lighteval"
] | [
"lighteval:swag"
] | true |
lighteval:bigbench:swedish_to_german_proverbs | bigbench:swedish_to_german_proverbs | lighteval | tasksource/bigbench | swedish_to_german_proverbs | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,559 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:swedish_to_german_proverbs"
] | false |
lighteval:bigbench_lite:symbol_interpretation | bigbench_lite:symbol_interpretation | lighteval | tasksource/bigbench | symbol_interpretation | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,574 | direct_call | true | bigbench_lite | 1 | [
"lighteval"
] | [
"lighteval:bigbench_lite:symbol_interpretation"
] | false |
helm:synthetic_reasoning:induction | synthetic_reasoning:induction | helm | lighteval/synthetic_reasoning | induction | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,589 | direct_call | true | synthetic_reasoning | 1 | [
"helm"
] | [
"helm:synthetic_reasoning:induction"
] | false |
helm:synthetic_reasoning:natural_easy | synthetic_reasoning:natural_easy | helm | lighteval/synthetic_reasoning_natural | easy | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,615 | direct_call | true | synthetic_reasoning | 1 | [
"helm"
] | [
"helm:synthetic_reasoning:natural_easy"
] | false |
helm:synthetic_reasoning:natural_hard | synthetic_reasoning:natural_hard | helm | lighteval/synthetic_reasoning_natural | hard | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,630 | direct_call | true | synthetic_reasoning | 1 | [
"helm"
] | [
"helm:synthetic_reasoning:natural_hard"
] | false |
helm:synthetic_reasoning:pattern_match | synthetic_reasoning:pattern_match | helm | lighteval/synthetic_reasoning | pattern_match | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,645 | direct_call | true | synthetic_reasoning | 1 | [
"helm"
] | [
"helm:synthetic_reasoning:pattern_match"
] | false |
helm:synthetic_reasoning:variable_substitution | synthetic_reasoning:variable_substitution | helm | lighteval/synthetic_reasoning | variable_substitution | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,671 | direct_call | true | synthetic_reasoning | 1 | [
"helm"
] | [
"helm:synthetic_reasoning:variable_substitution"
] | false |
lighteval:bigbench:tellmewhy | bigbench:tellmewhy | lighteval | tasksource/bigbench | tellmewhy | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,697 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:tellmewhy"
] | false |
lighteval:bigbench:tense | bigbench:tense | lighteval | tasksource/bigbench | tense | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,727 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:tense"
] | false |
helm:the_pile:arxiv | the_pile:arxiv | helm | lighteval/pile_helm | arxiv | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,742 | direct_call | true | the_pile | 1 | [
"helm"
] | [
"helm:the_pile:arxiv"
] | false |
helm:the_pile:bibliotik | the_pile:bibliotik | helm | lighteval/pile_helm | bibliotik | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,757 | direct_call | true | the_pile | 1 | [
"helm"
] | [
"helm:the_pile:bibliotik"
] | false |
helm:the_pile:commoncrawl | the_pile:commoncrawl | helm | lighteval/pile_helm | commoncrawl | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,772 | direct_call | true | the_pile | 1 | [
"helm"
] | [
"helm:the_pile:commoncrawl"
] | false |
helm:the_pile:dm-mathematics | the_pile:dm-mathematics | helm | lighteval/pile_helm | dm-mathematics | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,787 | direct_call | true | the_pile | 1 | [
"helm"
] | [
"helm:the_pile:dm-mathematics"
] | false |
helm:the_pile:enron | the_pile:enron | helm | lighteval/pile_helm | enron | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,802 | direct_call | true | the_pile | 1 | [
"helm"
] | [
"helm:the_pile:enron"
] | false |
helm:the_pile:europarl | the_pile:europarl | helm | lighteval/pile_helm | europarl | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,817 | direct_call | true | the_pile | 1 | [
"helm"
] | [
"helm:the_pile:europarl"
] | false |
helm:the_pile:freelaw | the_pile:freelaw | helm | lighteval/pile_helm | freelaw | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,832 | direct_call | true | the_pile | 1 | [
"helm"
] | [
"helm:the_pile:freelaw"
] | false |
helm:the_pile:github | the_pile:github | helm | lighteval/pile_helm | github | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,847 | direct_call | true | the_pile | 1 | [
"helm"
] | [
"helm:the_pile:github"
] | false |
helm:the_pile:gutenberg | the_pile:gutenberg | helm | lighteval/pile_helm | gutenberg | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,862 | direct_call | true | the_pile | 1 | [
"helm"
] | [
"helm:the_pile:gutenberg"
] | false |
helm:the_pile:hackernews | the_pile:hackernews | helm | lighteval/pile_helm | hackernews | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,877 | direct_call | true | the_pile | 1 | [
"helm"
] | [
"helm:the_pile:hackernews"
] | false |
helm:the_pile:nih-exporter | the_pile:nih-exporter | helm | lighteval/pile_helm | nih-exporter | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,892 | direct_call | true | the_pile | 1 | [
"helm"
] | [
"helm:the_pile:nih-exporter"
] | false |
helm:the_pile:opensubtitles | the_pile:opensubtitles | helm | lighteval/pile_helm | opensubtitles | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,907 | direct_call | true | the_pile | 1 | [
"helm"
] | [
"helm:the_pile:opensubtitles"
] | false |
helm:the_pile:openwebtext2 | the_pile:openwebtext2 | helm | lighteval/pile_helm | openwebtext2 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,922 | direct_call | true | the_pile | 1 | [
"helm"
] | [
"helm:the_pile:openwebtext2"
] | false |
helm:the_pile:pubmed-abstracts | the_pile:pubmed-abstracts | helm | lighteval/pile_helm | pubmed-abstracts | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,937 | direct_call | true | the_pile | 1 | [
"helm"
] | [
"helm:the_pile:pubmed-abstracts"
] | false |
helm:the_pile:pubmed-central | the_pile:pubmed-central | helm | lighteval/pile_helm | pubmed-central | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,952 | direct_call | true | the_pile | 1 | [
"helm"
] | [
"helm:the_pile:pubmed-central"
] | false |
helm:the_pile:stackexchange | the_pile:stackexchange | helm | lighteval/pile_helm | stackexchange | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 16,967 | direct_call | true | the_pile | 1 | [
"helm"
] | [
"helm:the_pile:stackexchange"
] | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.