task_id stringlengths 8 69 | name stringlengths 3 64 | suite stringclasses 6
values | hf_repo stringclasses 125
values | hf_subset stringlengths 0 55 | file_path stringclasses 6
values | line_number int64 36 22.8k | variable_name stringclasses 1
value | is_subtask bool 2
classes | main_task stringclasses 122
values | subtask_count int64 1 1 | suites listlengths 1 1 | subtasks listlengths 1 1 | is_standalone bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
lighteval:blimp:sentential_negation_npi_licensor_present | blimp:sentential_negation_npi_licensor_present | lighteval | blimp | sentential_negation_npi_licensor_present | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 5,819 | direct_call | true | blimp | 1 | [
"lighteval"
] | [
"lighteval:blimp:sentential_negation_npi_licensor_present"
] | false |
helm:blimp:sentential_negation_npi_licensor_present | blimp:sentential_negation_npi_licensor_present | helm | blimp | sentential_negation_npi_licensor_present | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 5,834 | direct_call | true | blimp | 1 | [
"helm"
] | [
"helm:blimp:sentential_negation_npi_licensor_present"
] | false |
lighteval:blimp:sentential_negation_npi_scope | blimp:sentential_negation_npi_scope | lighteval | blimp | sentential_negation_npi_scope | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 5,852 | direct_call | true | blimp | 1 | [
"lighteval"
] | [
"lighteval:blimp:sentential_negation_npi_scope"
] | false |
helm:blimp:sentential_negation_npi_scope | blimp:sentential_negation_npi_scope | helm | blimp | sentential_negation_npi_scope | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 5,867 | direct_call | true | blimp | 1 | [
"helm"
] | [
"helm:blimp:sentential_negation_npi_scope"
] | false |
lighteval:blimp:sentential_subject_island | blimp:sentential_subject_island | lighteval | blimp | sentential_subject_island | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 5,885 | direct_call | true | blimp | 1 | [
"lighteval"
] | [
"lighteval:blimp:sentential_subject_island"
] | false |
helm:blimp:sentential_subject_island | blimp:sentential_subject_island | helm | blimp | sentential_subject_island | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 5,900 | direct_call | true | blimp | 1 | [
"helm"
] | [
"helm:blimp:sentential_subject_island"
] | false |
lighteval:blimp:superlative_quantifiers_1 | blimp:superlative_quantifiers_1 | lighteval | blimp | superlative_quantifiers_1 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 5,918 | direct_call | true | blimp | 1 | [
"lighteval"
] | [
"lighteval:blimp:superlative_quantifiers_1"
] | false |
helm:blimp:superlative_quantifiers_1 | blimp:superlative_quantifiers_1 | helm | blimp | superlative_quantifiers_1 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 5,933 | direct_call | true | blimp | 1 | [
"helm"
] | [
"helm:blimp:superlative_quantifiers_1"
] | false |
lighteval:blimp:superlative_quantifiers_2 | blimp:superlative_quantifiers_2 | lighteval | blimp | superlative_quantifiers_2 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 5,951 | direct_call | true | blimp | 1 | [
"lighteval"
] | [
"lighteval:blimp:superlative_quantifiers_2"
] | false |
helm:blimp:superlative_quantifiers_2 | blimp:superlative_quantifiers_2 | helm | blimp | superlative_quantifiers_2 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 5,966 | direct_call | true | blimp | 1 | [
"helm"
] | [
"helm:blimp:superlative_quantifiers_2"
] | false |
lighteval:blimp:tough_vs_raising_1 | blimp:tough_vs_raising_1 | lighteval | blimp | tough_vs_raising_1 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 5,984 | direct_call | true | blimp | 1 | [
"lighteval"
] | [
"lighteval:blimp:tough_vs_raising_1"
] | false |
helm:blimp:tough_vs_raising_1 | blimp:tough_vs_raising_1 | helm | blimp | tough_vs_raising_1 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 5,999 | direct_call | true | blimp | 1 | [
"helm"
] | [
"helm:blimp:tough_vs_raising_1"
] | false |
lighteval:blimp:tough_vs_raising_2 | blimp:tough_vs_raising_2 | lighteval | blimp | tough_vs_raising_2 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,017 | direct_call | true | blimp | 1 | [
"lighteval"
] | [
"lighteval:blimp:tough_vs_raising_2"
] | false |
helm:blimp:tough_vs_raising_2 | blimp:tough_vs_raising_2 | helm | blimp | tough_vs_raising_2 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,032 | direct_call | true | blimp | 1 | [
"helm"
] | [
"helm:blimp:tough_vs_raising_2"
] | false |
lighteval:blimp:transitive | blimp:transitive | lighteval | blimp | transitive | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,050 | direct_call | true | blimp | 1 | [
"lighteval"
] | [
"lighteval:blimp:transitive"
] | false |
helm:blimp:transitive | blimp:transitive | helm | blimp | transitive | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,065 | direct_call | true | blimp | 1 | [
"helm"
] | [
"helm:blimp:transitive"
] | false |
lighteval:blimp:wh_island | blimp:wh_island | lighteval | blimp | wh_island | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,083 | direct_call | true | blimp | 1 | [
"lighteval"
] | [
"lighteval:blimp:wh_island"
] | false |
helm:blimp:wh_island | blimp:wh_island | helm | blimp | wh_island | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,098 | direct_call | true | blimp | 1 | [
"helm"
] | [
"helm:blimp:wh_island"
] | false |
lighteval:blimp:wh_questions_object_gap | blimp:wh_questions_object_gap | lighteval | blimp | wh_questions_object_gap | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,116 | direct_call | true | blimp | 1 | [
"lighteval"
] | [
"lighteval:blimp:wh_questions_object_gap"
] | false |
helm:blimp:wh_questions_object_gap | blimp:wh_questions_object_gap | helm | blimp | wh_questions_object_gap | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,131 | direct_call | true | blimp | 1 | [
"helm"
] | [
"helm:blimp:wh_questions_object_gap"
] | false |
lighteval:blimp:wh_questions_subject_gap | blimp:wh_questions_subject_gap | lighteval | blimp | wh_questions_subject_gap | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,149 | direct_call | true | blimp | 1 | [
"lighteval"
] | [
"lighteval:blimp:wh_questions_subject_gap"
] | false |
helm:blimp:wh_questions_subject_gap | blimp:wh_questions_subject_gap | helm | blimp | wh_questions_subject_gap | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,164 | direct_call | true | blimp | 1 | [
"helm"
] | [
"helm:blimp:wh_questions_subject_gap"
] | false |
lighteval:blimp:wh_questions_subject_gap_long_distance | blimp:wh_questions_subject_gap_long_distance | lighteval | blimp | wh_questions_subject_gap_long_distance | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,182 | direct_call | true | blimp | 1 | [
"lighteval"
] | [
"lighteval:blimp:wh_questions_subject_gap_long_distance"
] | false |
helm:blimp:wh_questions_subject_gap_long_distance | blimp:wh_questions_subject_gap_long_distance | helm | blimp | wh_questions_subject_gap_long_distance | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,197 | direct_call | true | blimp | 1 | [
"helm"
] | [
"helm:blimp:wh_questions_subject_gap_long_distance"
] | false |
lighteval:blimp:wh_vs_that_no_gap | blimp:wh_vs_that_no_gap | lighteval | blimp | wh_vs_that_no_gap | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,215 | direct_call | true | blimp | 1 | [
"lighteval"
] | [
"lighteval:blimp:wh_vs_that_no_gap"
] | false |
helm:blimp:wh_vs_that_no_gap | blimp:wh_vs_that_no_gap | helm | blimp | wh_vs_that_no_gap | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,230 | direct_call | true | blimp | 1 | [
"helm"
] | [
"helm:blimp:wh_vs_that_no_gap"
] | false |
lighteval:blimp:wh_vs_that_no_gap_long_distance | blimp:wh_vs_that_no_gap_long_distance | lighteval | blimp | wh_vs_that_no_gap_long_distance | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,248 | direct_call | true | blimp | 1 | [
"lighteval"
] | [
"lighteval:blimp:wh_vs_that_no_gap_long_distance"
] | false |
helm:blimp:wh_vs_that_no_gap_long_distance | blimp:wh_vs_that_no_gap_long_distance | helm | blimp | wh_vs_that_no_gap_long_distance | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,263 | direct_call | true | blimp | 1 | [
"helm"
] | [
"helm:blimp:wh_vs_that_no_gap_long_distance"
] | false |
lighteval:blimp:wh_vs_that_with_gap | blimp:wh_vs_that_with_gap | lighteval | blimp | wh_vs_that_with_gap | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,281 | direct_call | true | blimp | 1 | [
"lighteval"
] | [
"lighteval:blimp:wh_vs_that_with_gap"
] | false |
helm:blimp:wh_vs_that_with_gap | blimp:wh_vs_that_with_gap | helm | blimp | wh_vs_that_with_gap | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,296 | direct_call | true | blimp | 1 | [
"helm"
] | [
"helm:blimp:wh_vs_that_with_gap"
] | false |
lighteval:blimp:wh_vs_that_with_gap_long_distance | blimp:wh_vs_that_with_gap_long_distance | lighteval | blimp | wh_vs_that_with_gap_long_distance | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,314 | direct_call | true | blimp | 1 | [
"lighteval"
] | [
"lighteval:blimp:wh_vs_that_with_gap_long_distance"
] | false |
helm:blimp:wh_vs_that_with_gap_long_distance | blimp:wh_vs_that_with_gap_long_distance | helm | blimp | wh_vs_that_with_gap_long_distance | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,329 | direct_call | true | blimp | 1 | [
"helm"
] | [
"helm:blimp:wh_vs_that_with_gap_long_distance"
] | false |
helm:bold | bold | helm | lighteval/bold_helm | all | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,347 | direct_call | false | bold | 1 | [
"helm"
] | [
"helm:bold"
] | false |
helm:bold:gender | bold:gender | helm | lighteval/bold_helm | gender | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,362 | direct_call | true | bold | 1 | [
"helm"
] | [
"helm:bold:gender"
] | false |
helm:bold:political_ideology | bold:political_ideology | helm | lighteval/bold_helm | political_ideology | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,377 | direct_call | true | bold | 1 | [
"helm"
] | [
"helm:bold:political_ideology"
] | false |
helm:bold:profession | bold:profession | helm | lighteval/bold_helm | profession | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,392 | direct_call | true | bold | 1 | [
"helm"
] | [
"helm:bold:profession"
] | false |
helm:bold:race | bold:race | helm | lighteval/bold_helm | race | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,407 | direct_call | true | bold | 1 | [
"helm"
] | [
"helm:bold:race"
] | false |
helm:bold:religious_ideology | bold:religious_ideology | helm | lighteval/bold_helm | religious_ideology | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,422 | direct_call | true | bold | 1 | [
"helm"
] | [
"helm:bold:religious_ideology"
] | false |
helm:boolq | boolq | helm | lighteval/boolq_helm | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,437 | direct_call | false | boolq | 1 | [
"helm"
] | [
"helm:boolq"
] | false |
helm:boolq:contrastset | boolq:contrastset | helm | lighteval/boolq_helm | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,463 | direct_call | true | boolq | 1 | [
"helm"
] | [
"helm:boolq:contrastset"
] | false |
lighteval:bigbench:bridging_anaphora_resolution_barqa | bigbench:bridging_anaphora_resolution_barqa | lighteval | tasksource/bigbench | bridging_anaphora_resolution_barqa | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,489 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:bridging_anaphora_resolution_barqa"
] | false |
lighteval:bigbench:cause_and_effect | bigbench:cause_and_effect | lighteval | tasksource/bigbench | cause_and_effect | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,519 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:cause_and_effect"
] | false |
lighteval:bigbench:checkmate_in_one | bigbench:checkmate_in_one | lighteval | tasksource/bigbench | checkmate_in_one | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,534 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:checkmate_in_one"
] | false |
lighteval:bigbench:chess_state_tracking | bigbench:chess_state_tracking | lighteval | tasksource/bigbench | chess_state_tracking | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,549 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:chess_state_tracking"
] | false |
lighteval:bigbench:chinese_remainder_theorem | bigbench:chinese_remainder_theorem | lighteval | tasksource/bigbench | chinese_remainder_theorem | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,564 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:chinese_remainder_theorem"
] | false |
lighteval:bigbench:cifar10_classification | bigbench:cifar10_classification | lighteval | tasksource/bigbench | cifar10_classification | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,579 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:cifar10_classification"
] | false |
helm:civil_comments | civil_comments | helm | lighteval/civil_comments_helm | all | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,594 | direct_call | false | civil_comments | 1 | [
"helm"
] | [
"helm:civil_comments"
] | false |
helm:civil_comments:LGBTQ | civil_comments:LGBTQ | helm | lighteval/civil_comments_helm | LGBTQ | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,622 | direct_call | true | civil_comments | 1 | [
"helm"
] | [
"helm:civil_comments:LGBTQ"
] | false |
helm:civil_comments:black | civil_comments:black | helm | lighteval/civil_comments_helm | black | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,650 | direct_call | true | civil_comments | 1 | [
"helm"
] | [
"helm:civil_comments:black"
] | false |
helm:civil_comments:christian | civil_comments:christian | helm | lighteval/civil_comments_helm | christian | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,678 | direct_call | true | civil_comments | 1 | [
"helm"
] | [
"helm:civil_comments:christian"
] | false |
helm:civil_comments:female | civil_comments:female | helm | lighteval/civil_comments_helm | female | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,706 | direct_call | true | civil_comments | 1 | [
"helm"
] | [
"helm:civil_comments:female"
] | false |
helm:civil_comments:male | civil_comments:male | helm | lighteval/civil_comments_helm | male | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,734 | direct_call | true | civil_comments | 1 | [
"helm"
] | [
"helm:civil_comments:male"
] | false |
helm:civil_comments:muslim | civil_comments:muslim | helm | lighteval/civil_comments_helm | muslim | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,762 | direct_call | true | civil_comments | 1 | [
"helm"
] | [
"helm:civil_comments:muslim"
] | false |
helm:civil_comments:other_religions | civil_comments:other_religions | helm | lighteval/civil_comments_helm | other_religions | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,790 | direct_call | true | civil_comments | 1 | [
"helm"
] | [
"helm:civil_comments:other_religions"
] | false |
helm:civil_comments:white | civil_comments:white | helm | lighteval/civil_comments_helm | white | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,818 | direct_call | true | civil_comments | 1 | [
"helm"
] | [
"helm:civil_comments:white"
] | false |
lighteval:bigbench_lite:code_line_description | bigbench_lite:code_line_description | lighteval | tasksource/bigbench | code_line_description | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,846 | direct_call | true | bigbench_lite | 1 | [
"lighteval"
] | [
"lighteval:bigbench_lite:code_line_description"
] | false |
lighteval:bigbench:codenames | bigbench:codenames | lighteval | tasksource/bigbench | codenames | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,861 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:codenames"
] | false |
lighteval:bigbench:color | bigbench:color | lighteval | tasksource/bigbench | color | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,876 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:color"
] | false |
lighteval:bigbench:common_morpheme | bigbench:common_morpheme | lighteval | tasksource/bigbench | common_morpheme | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,896 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:common_morpheme"
] | false |
helm:commonsenseqa | commonsenseqa | helm | commonsense_qa | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,911 | direct_call | false | commonsenseqa | 1 | [
"helm"
] | [
"helm:commonsenseqa"
] | true |
lighteval:bigbench_lite:conceptual_combinations | bigbench_lite:conceptual_combinations | lighteval | tasksource/bigbench | conceptual_combinations | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,937 | direct_call | true | bigbench_lite | 1 | [
"lighteval"
] | [
"lighteval:bigbench_lite:conceptual_combinations"
] | false |
lighteval:bigbench_lite:conlang_translation | bigbench_lite:conlang_translation | lighteval | tasksource/bigbench | conlang_translation | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,952 | direct_call | true | bigbench_lite | 1 | [
"lighteval"
] | [
"lighteval:bigbench_lite:conlang_translation"
] | false |
lighteval:bigbench:contextual_parametric_knowledge_conflicts | bigbench:contextual_parametric_knowledge_conflicts | lighteval | tasksource/bigbench | contextual_parametric_knowledge_conflicts | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,967 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:contextual_parametric_knowledge_conflicts"
] | false |
helm:copyright:n_books_1000-extractions_per_book_1-prefix_length_125 | copyright:n_books_1000-extractions_per_book_1-prefix_length_125 | helm | lighteval/copyright_helm | n_books_1000-extractions_per_book_1-prefix_length_125 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,982 | direct_call | true | copyright | 1 | [
"helm"
] | [
"helm:copyright:n_books_1000-extractions_per_book_1-prefix_length_125"
] | false |
helm:copyright:n_books_1000-extractions_per_book_1-prefix_length_25 | copyright:n_books_1000-extractions_per_book_1-prefix_length_25 | helm | lighteval/copyright_helm | n_books_1000-extractions_per_book_1-prefix_length_25 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 6,997 | direct_call | true | copyright | 1 | [
"helm"
] | [
"helm:copyright:n_books_1000-extractions_per_book_1-prefix_length_25"
] | false |
helm:copyright:n_books_1000-extractions_per_book_1-prefix_length_5 | copyright:n_books_1000-extractions_per_book_1-prefix_length_5 | helm | lighteval/copyright_helm | n_books_1000-extractions_per_book_1-prefix_length_5 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,012 | direct_call | true | copyright | 1 | [
"helm"
] | [
"helm:copyright:n_books_1000-extractions_per_book_1-prefix_length_5"
] | false |
helm:copyright:n_books_1000-extractions_per_book_3-prefix_length_125 | copyright:n_books_1000-extractions_per_book_3-prefix_length_125 | helm | lighteval/copyright_helm | n_books_1000-extractions_per_book_3-prefix_length_125 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,027 | direct_call | true | copyright | 1 | [
"helm"
] | [
"helm:copyright:n_books_1000-extractions_per_book_3-prefix_length_125"
] | false |
helm:copyright:n_books_1000-extractions_per_book_3-prefix_length_25 | copyright:n_books_1000-extractions_per_book_3-prefix_length_25 | helm | lighteval/copyright_helm | n_books_1000-extractions_per_book_3-prefix_length_25 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,042 | direct_call | true | copyright | 1 | [
"helm"
] | [
"helm:copyright:n_books_1000-extractions_per_book_3-prefix_length_25"
] | false |
helm:copyright:n_books_1000-extractions_per_book_3-prefix_length_5 | copyright:n_books_1000-extractions_per_book_3-prefix_length_5 | helm | lighteval/copyright_helm | n_books_1000-extractions_per_book_3-prefix_length_5 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,057 | direct_call | true | copyright | 1 | [
"helm"
] | [
"helm:copyright:n_books_1000-extractions_per_book_3-prefix_length_5"
] | false |
helm:copyright:oh_the_places | copyright:oh_the_places | helm | lighteval/copyright_helm | oh_the_places | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,072 | direct_call | true | copyright | 1 | [
"helm"
] | [
"helm:copyright:oh_the_places"
] | false |
helm:copyright:pilot | copyright:pilot | helm | lighteval/copyright_helm | pilot | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,087 | direct_call | true | copyright | 1 | [
"helm"
] | [
"helm:copyright:pilot"
] | false |
helm:copyright:popular_books-prefix_length_10 | copyright:popular_books-prefix_length_10 | helm | lighteval/copyright_helm | popular_books-prefix_length_10 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,102 | direct_call | true | copyright | 1 | [
"helm"
] | [
"helm:copyright:popular_books-prefix_length_10"
] | false |
helm:copyright:popular_books-prefix_length_125 | copyright:popular_books-prefix_length_125 | helm | lighteval/copyright_helm | popular_books-prefix_length_125 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,117 | direct_call | true | copyright | 1 | [
"helm"
] | [
"helm:copyright:popular_books-prefix_length_125"
] | false |
helm:copyright:popular_books-prefix_length_25 | copyright:popular_books-prefix_length_25 | helm | lighteval/copyright_helm | popular_books-prefix_length_25 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,132 | direct_call | true | copyright | 1 | [
"helm"
] | [
"helm:copyright:popular_books-prefix_length_25"
] | false |
helm:copyright:popular_books-prefix_length_250 | copyright:popular_books-prefix_length_250 | helm | lighteval/copyright_helm | popular_books-prefix_length_250 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,147 | direct_call | true | copyright | 1 | [
"helm"
] | [
"helm:copyright:popular_books-prefix_length_250"
] | false |
helm:copyright:popular_books-prefix_length_5 | copyright:popular_books-prefix_length_5 | helm | lighteval/copyright_helm | popular_books-prefix_length_5 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,162 | direct_call | true | copyright | 1 | [
"helm"
] | [
"helm:copyright:popular_books-prefix_length_5"
] | false |
helm:copyright:popular_books-prefix_length_50 | copyright:popular_books-prefix_length_50 | helm | lighteval/copyright_helm | popular_books-prefix_length_50 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,177 | direct_call | true | copyright | 1 | [
"helm"
] | [
"helm:copyright:popular_books-prefix_length_50"
] | false |
helm:copyright:prompt_num_line_1-min_lines_20 | copyright:prompt_num_line_1-min_lines_20 | helm | lighteval/copyright_helm | prompt_num_line_1-min_lines_20 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,192 | direct_call | true | copyright | 1 | [
"helm"
] | [
"helm:copyright:prompt_num_line_1-min_lines_20"
] | false |
helm:copyright:prompt_num_line_10-min_lines_20 | copyright:prompt_num_line_10-min_lines_20 | helm | lighteval/copyright_helm | prompt_num_line_10-min_lines_20 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,207 | direct_call | true | copyright | 1 | [
"helm"
] | [
"helm:copyright:prompt_num_line_10-min_lines_20"
] | false |
helm:copyright:prompt_num_line_5-min_lines_20 | copyright:prompt_num_line_5-min_lines_20 | helm | lighteval/copyright_helm | prompt_num_line_5-min_lines_20 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,222 | direct_call | true | copyright | 1 | [
"helm"
] | [
"helm:copyright:prompt_num_line_5-min_lines_20"
] | false |
lighteval:coqa | coqa | lighteval | stanfordnlp/coqa | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,237 | direct_call | false | coqa | 1 | [
"lighteval"
] | [
"lighteval:coqa"
] | true |
lighteval:coqa_bb | coqa_bb | lighteval | coqa | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,266 | direct_call | false | coqa_bb | 1 | [
"lighteval"
] | [
"lighteval:coqa_bb"
] | true |
helm:covid_dialogue | covid_dialogue | helm | lighteval/covid_dialogue | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,281 | direct_call | false | covid_dialogue | 1 | [
"helm"
] | [
"helm:covid_dialogue"
] | true |
lighteval:bigbench:crash_blossom | bigbench:crash_blossom | lighteval | tasksource/bigbench | crash_blossom | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,303 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:crash_blossom"
] | false |
lighteval:bigbench:crass_ai | bigbench:crass_ai | lighteval | tasksource/bigbench | crass_ai | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,318 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:crass_ai"
] | false |
lighteval:bigbench:cryobiology_spanish | bigbench:cryobiology_spanish | lighteval | tasksource/bigbench | cryobiology_spanish | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,333 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:cryobiology_spanish"
] | false |
lighteval:bigbench:cryptonite | bigbench:cryptonite | lighteval | tasksource/bigbench | cryptonite | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,348 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:cryptonite"
] | false |
lighteval:bigbench:cs_algorithms | bigbench:cs_algorithms | lighteval | tasksource/bigbench | cs_algorithms | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,363 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:cs_algorithms"
] | false |
lighteval:bigbench:dark_humor_detection | bigbench:dark_humor_detection | lighteval | tasksource/bigbench | dark_humor_detection | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,378 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:dark_humor_detection"
] | false |
lighteval:bigbench:discourse_marker_prediction | bigbench:discourse_marker_prediction | lighteval | tasksource/bigbench | discourse_marker_prediction | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,423 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:discourse_marker_prediction"
] | false |
lighteval:bigbench:disfl_qa | bigbench:disfl_qa | lighteval | tasksource/bigbench | disfl_qa | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,438 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:disfl_qa"
] | false |
lighteval:drop | drop | lighteval | lighteval/drop_harness | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,453 | direct_call | false | drop | 1 | [
"lighteval"
] | [
"lighteval:drop"
] | true |
helm:dyck_language:2 | dyck_language:2 | helm | lighteval/DyckLanguage | 2 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,497 | direct_call | true | dyck_language | 1 | [
"helm"
] | [
"helm:dyck_language:2"
] | false |
helm:dyck_language:3 | dyck_language:3 | helm | lighteval/DyckLanguage | 3 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,512 | direct_call | true | dyck_language | 1 | [
"helm"
] | [
"helm:dyck_language:3"
] | false |
helm:dyck_language:4 | dyck_language:4 | helm | lighteval/DyckLanguage | 4 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,527 | direct_call | true | dyck_language | 1 | [
"helm"
] | [
"helm:dyck_language:4"
] | false |
lighteval:bigbench:dyck_languages | bigbench:dyck_languages | lighteval | tasksource/bigbench | dyck_languages | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,542 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:dyck_languages"
] | false |
lighteval:bigbench:elementary_math_qa | bigbench:elementary_math_qa | lighteval | tasksource/bigbench | elementary_math_qa | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,557 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:elementary_math_qa"
] | false |
lighteval:bigbench_lite:emoji_movie | bigbench_lite:emoji_movie | lighteval | tasksource/bigbench | emoji_movie | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,572 | direct_call | true | bigbench_lite | 1 | [
"lighteval"
] | [
"lighteval:bigbench_lite:emoji_movie"
] | false |
lighteval:bigbench:emojis_emotion_prediction | bigbench:emojis_emotion_prediction | lighteval | tasksource/bigbench | emojis_emotion_prediction | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,592 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:emojis_emotion_prediction"
] | false |
lighteval:bigbench:empirical_judgments | bigbench:empirical_judgments | lighteval | tasksource/bigbench | empirical_judgments | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 7,607 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:empirical_judgments"
] | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.