task_id
stringlengths 8
69
| name
stringlengths 3
64
| suite
stringclasses 6
values | hf_repo
stringclasses 125
values | hf_subset
stringlengths 0
55
| file_path
stringclasses 6
values | line_number
int64 36
22.8k
| variable_name
stringclasses 1
value | is_subtask
bool 2
classes | main_task
stringclasses 122
values | subtask_count
int64 1
1
| suites
listlengths 1
1
| subtasks
listlengths 1
1
| is_standalone
bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
lighteval:bigbench:persian_idioms
|
bigbench:persian_idioms
|
lighteval
|
tasksource/bigbench
|
persian_idioms
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,113
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:persian_idioms"
] | false
|
lighteval:bigbench:phrase_relatedness
|
bigbench:phrase_relatedness
|
lighteval
|
tasksource/bigbench
|
phrase_relatedness
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,128
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:phrase_relatedness"
] | false
|
lighteval:bigbench:physical_intuition
|
bigbench:physical_intuition
|
lighteval
|
tasksource/bigbench
|
physical_intuition
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,143
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:physical_intuition"
] | false
|
lighteval:bigbench:physics
|
bigbench:physics
|
lighteval
|
tasksource/bigbench
|
physics
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,158
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:physics"
] | false
|
lighteval:bigbench:physics_questions
|
bigbench:physics_questions
|
lighteval
|
tasksource/bigbench
|
physics_questions
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,173
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:physics_questions"
] | false
|
lighteval:piqa
|
piqa
|
lighteval
|
ybisk/piqa
|
plain_text
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,188
|
direct_call
| false
|
piqa
| 1
|
[
"lighteval"
] |
[
"lighteval:piqa"
] | true
|
helm:piqa
|
piqa
|
helm
|
ybisk/piqa
|
plain_text
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,206
|
direct_call
| false
|
piqa
| 1
|
[
"helm"
] |
[
"helm:piqa"
] | true
|
lighteval:bigbench_lite:play_dialog_same_or_different
|
bigbench_lite:play_dialog_same_or_different
|
lighteval
|
tasksource/bigbench
|
play_dialog_same_or_different
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,232
|
direct_call
| true
|
bigbench_lite
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench_lite:play_dialog_same_or_different"
] | false
|
lighteval:bigbench:polish_sequence_labeling
|
bigbench:polish_sequence_labeling
|
lighteval
|
tasksource/bigbench
|
polish_sequence_labeling
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,247
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:polish_sequence_labeling"
] | false
|
lighteval:bigbench:presuppositions_as_nli
|
bigbench:presuppositions_as_nli
|
lighteval
|
tasksource/bigbench
|
presuppositions_as_nli
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,262
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:presuppositions_as_nli"
] | false
|
lighteval:prost
|
prost
|
lighteval
|
lighteval/prost
|
default
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,277
|
direct_call
| false
|
prost
| 1
|
[
"lighteval"
] |
[
"lighteval:prost"
] | true
|
lighteval:pubmedqa
|
pubmedqa
|
lighteval
|
pubmed_qa
|
pqa_labeled
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,295
|
direct_call
| false
|
pubmedqa
| 1
|
[
"lighteval"
] |
[
"lighteval:pubmedqa"
] | true
|
helm:pubmedqa
|
pubmedqa
|
helm
|
pubmed_qa
|
pqa_labeled
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,310
|
direct_call
| false
|
pubmedqa
| 1
|
[
"helm"
] |
[
"helm:pubmedqa"
] | true
|
lighteval:qa4mre:2011
|
qa4mre:2011
|
lighteval
|
qa4mre
|
2011.main.EN
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,336
|
direct_call
| true
|
qa4mre
| 1
|
[
"lighteval"
] |
[
"lighteval:qa4mre:2011"
] | false
|
lighteval:qa4mre:2012
|
qa4mre:2012
|
lighteval
|
qa4mre
|
2012.main.EN
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,354
|
direct_call
| true
|
qa4mre
| 1
|
[
"lighteval"
] |
[
"lighteval:qa4mre:2012"
] | false
|
lighteval:qa4mre:2013
|
qa4mre:2013
|
lighteval
|
qa4mre
|
2013.main.EN
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,372
|
direct_call
| true
|
qa4mre
| 1
|
[
"lighteval"
] |
[
"lighteval:qa4mre:2013"
] | false
|
lighteval:bigbench:qa_wikidata
|
bigbench:qa_wikidata
|
lighteval
|
tasksource/bigbench
|
qa_wikidata
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,390
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:qa_wikidata"
] | false
|
lighteval:qasper
|
qasper
|
lighteval
|
allenai/qasper
|
qasper
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,410
|
direct_call
| false
|
qasper
| 1
|
[
"lighteval"
] |
[
"lighteval:qasper"
] | true
|
lighteval:qasper_ll
|
qasper_ll
|
lighteval
|
allenai/qasper
|
qasper
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,425
|
direct_call
| false
|
qasper_ll
| 1
|
[
"lighteval"
] |
[
"lighteval:qasper_ll"
] | true
|
helm:quac
|
quac
|
helm
|
lighteval/quac_helm
|
default
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,440
|
direct_call
| false
|
quac
| 1
|
[
"helm"
] |
[
"helm:quac"
] | true
|
lighteval:bigbench:question_selection
|
bigbench:question_selection
|
lighteval
|
tasksource/bigbench
|
question_selection
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,459
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:question_selection"
] | false
|
lighteval:race:high
|
race:high
|
lighteval
|
EleutherAI/race
|
high
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,474
|
direct_call
| true
|
race
| 1
|
[
"lighteval"
] |
[
"lighteval:race:high"
] | false
|
helm:raft:ade_corpus_v2
|
raft:ade_corpus_v2
|
helm
|
ought/raft
|
ade_corpus_v2
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,489
|
direct_call
| true
|
raft
| 1
|
[
"helm"
] |
[
"helm:raft:ade_corpus_v2"
] | false
|
helm:raft:banking_77
|
raft:banking_77
|
helm
|
ought/raft
|
banking_77
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,517
|
direct_call
| true
|
raft
| 1
|
[
"helm"
] |
[
"helm:raft:banking_77"
] | false
|
helm:raft:neurips_impact_statement_risks
|
raft:neurips_impact_statement_risks
|
helm
|
ought/raft
|
neurips_impact_statement_risks
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,545
|
direct_call
| true
|
raft
| 1
|
[
"helm"
] |
[
"helm:raft:neurips_impact_statement_risks"
] | false
|
helm:raft:one_stop_english
|
raft:one_stop_english
|
helm
|
ought/raft
|
one_stop_english
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,573
|
direct_call
| true
|
raft
| 1
|
[
"helm"
] |
[
"helm:raft:one_stop_english"
] | false
|
helm:raft:overruling
|
raft:overruling
|
helm
|
ought/raft
|
overruling
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,601
|
direct_call
| true
|
raft
| 1
|
[
"helm"
] |
[
"helm:raft:overruling"
] | false
|
helm:raft:semiconductor_org_types
|
raft:semiconductor_org_types
|
helm
|
ought/raft
|
semiconductor_org_types
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,629
|
direct_call
| true
|
raft
| 1
|
[
"helm"
] |
[
"helm:raft:semiconductor_org_types"
] | false
|
helm:raft:systematic_review_inclusion
|
raft:systematic_review_inclusion
|
helm
|
ought/raft
|
systematic_review_inclusion
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,657
|
direct_call
| true
|
raft
| 1
|
[
"helm"
] |
[
"helm:raft:systematic_review_inclusion"
] | false
|
helm:raft:tai_safety_research
|
raft:tai_safety_research
|
helm
|
ought/raft
|
tai_safety_research
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,685
|
direct_call
| true
|
raft
| 1
|
[
"helm"
] |
[
"helm:raft:tai_safety_research"
] | false
|
helm:raft:terms_of_service
|
raft:terms_of_service
|
helm
|
ought/raft
|
terms_of_service
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,713
|
direct_call
| true
|
raft
| 1
|
[
"helm"
] |
[
"helm:raft:terms_of_service"
] | false
|
helm:raft:tweet_eval_hate
|
raft:tweet_eval_hate
|
helm
|
ought/raft
|
tweet_eval_hate
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,741
|
direct_call
| true
|
raft
| 1
|
[
"helm"
] |
[
"helm:raft:tweet_eval_hate"
] | false
|
helm:raft:twitter_complaints
|
raft:twitter_complaints
|
helm
|
ought/raft
|
twitter_complaints
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,769
|
direct_call
| true
|
raft
| 1
|
[
"helm"
] |
[
"helm:raft:twitter_complaints"
] | false
|
lighteval:bigbench:real_or_fake_text
|
bigbench:real_or_fake_text
|
lighteval
|
tasksource/bigbench
|
real_or_fake_text
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,797
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:real_or_fake_text"
] | false
|
helm:real_toxicity_prompts
|
real_toxicity_prompts
|
helm
|
allenai/real-toxicity-prompts
|
default
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,812
|
direct_call
| false
|
real_toxicity_prompts
| 1
|
[
"helm"
] |
[
"helm:real_toxicity_prompts"
] | true
|
lighteval:bigbench_lite:repeat_copy_logic
|
bigbench_lite:repeat_copy_logic
|
lighteval
|
tasksource/bigbench
|
repeat_copy_logic
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,842
|
direct_call
| true
|
bigbench_lite
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench_lite:repeat_copy_logic"
] | false
|
lighteval:bigbench:rephrase
|
bigbench:rephrase
|
lighteval
|
tasksource/bigbench
|
rephrase
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,857
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:rephrase"
] | false
|
lighteval:bigbench:rhyming
|
bigbench:rhyming
|
lighteval
|
tasksource/bigbench
|
rhyming
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,877
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:rhyming"
] | false
|
lighteval:bigbench:riddle_sense
|
bigbench:riddle_sense
|
lighteval
|
tasksource/bigbench
|
riddle_sense
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,892
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:riddle_sense"
] | false
|
lighteval:bigbench:scientific_press_release
|
bigbench:scientific_press_release
|
lighteval
|
tasksource/bigbench
|
scientific_press_release
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,937
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:scientific_press_release"
] | false
|
lighteval:sciq
|
sciq
|
lighteval
|
sciq
|
default
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,952
|
direct_call
| false
|
sciq
| 1
|
[
"lighteval"
] |
[
"lighteval:sciq"
] | true
|
lighteval:bigbench:semantic_parsing_in_context_sparc
|
bigbench:semantic_parsing_in_context_sparc
|
lighteval
|
tasksource/bigbench
|
semantic_parsing_in_context_sparc
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,970
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:semantic_parsing_in_context_sparc"
] | false
|
lighteval:bigbench:semantic_parsing_spider
|
bigbench:semantic_parsing_spider
|
lighteval
|
tasksource/bigbench
|
semantic_parsing_spider
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 15,985
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:semantic_parsing_spider"
] | false
|
lighteval:bigbench:sentence_ambiguity
|
bigbench:sentence_ambiguity
|
lighteval
|
tasksource/bigbench
|
sentence_ambiguity
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,000
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:sentence_ambiguity"
] | false
|
lighteval:bigbench:similarities_abstraction
|
bigbench:similarities_abstraction
|
lighteval
|
tasksource/bigbench
|
similarities_abstraction
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,015
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:similarities_abstraction"
] | false
|
lighteval:bigbench:simp_turing_concept
|
bigbench:simp_turing_concept
|
lighteval
|
tasksource/bigbench
|
simp_turing_concept
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,030
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:simp_turing_concept"
] | false
|
lighteval:simpleqa
|
simpleqa
|
lighteval
|
lighteval/SimpleQA
|
default
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,045
|
direct_call
| false
|
simpleqa
| 1
|
[
"lighteval"
] |
[
"lighteval:simpleqa"
] | true
|
lighteval:bigbench:simple_arithmetic_json
|
bigbench:simple_arithmetic_json
|
lighteval
|
tasksource/bigbench
|
simple_arithmetic_json
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,060
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:simple_arithmetic_json"
] | false
|
lighteval:bigbench:simple_arithmetic_json_multiple_choice
|
bigbench:simple_arithmetic_json_multiple_choice
|
lighteval
|
tasksource/bigbench
|
simple_arithmetic_json_multiple_choice
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,075
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:simple_arithmetic_json_multiple_choice"
] | false
|
lighteval:bigbench:simple_arithmetic_json_subtasks
|
bigbench:simple_arithmetic_json_subtasks
|
lighteval
|
tasksource/bigbench
|
simple_arithmetic_json_subtasks
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,090
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:simple_arithmetic_json_subtasks"
] | false
|
lighteval:bigbench:simple_arithmetic_multiple_targets_json
|
bigbench:simple_arithmetic_multiple_targets_json
|
lighteval
|
tasksource/bigbench
|
simple_arithmetic_multiple_targets_json
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,105
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:simple_arithmetic_multiple_targets_json"
] | false
|
lighteval:bigbench:simple_ethical_questions
|
bigbench:simple_ethical_questions
|
lighteval
|
tasksource/bigbench
|
simple_ethical_questions
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,120
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:simple_ethical_questions"
] | false
|
lighteval:bigbench:simple_text_editing
|
bigbench:simple_text_editing
|
lighteval
|
tasksource/bigbench
|
simple_text_editing
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,135
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:simple_text_editing"
] | false
|
helm:siqa
|
siqa
|
helm
|
allenai/social_i_qa
|
default
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,150
|
direct_call
| false
|
siqa
| 1
|
[
"helm"
] |
[
"helm:siqa"
] | true
|
lighteval:bigbench:social_iqa
|
bigbench:social_iqa
|
lighteval
|
tasksource/bigbench
|
social_iqa
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,191
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:social_iqa"
] | false
|
lighteval:bigbench:social_support
|
bigbench:social_support
|
lighteval
|
tasksource/bigbench
|
social_support
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,206
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:social_support"
] | false
|
lighteval:squad_v2
|
squad_v2
|
lighteval
|
rajpurkar/squad_v2
|
squad_v2
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,236
|
direct_call
| false
|
squad_v2
| 1
|
[
"lighteval"
] |
[
"lighteval:squad_v2"
] | true
|
lighteval:storycloze:2016
|
storycloze:2016
|
lighteval
|
MoE-UNC/story_cloze
|
2016
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,265
|
direct_call
| true
|
storycloze
| 1
|
[
"lighteval"
] |
[
"lighteval:storycloze:2016"
] | false
|
lighteval:storycloze:2018
|
storycloze:2018
|
lighteval
|
MoE-UNC/story_cloze
|
2018
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,280
|
direct_call
| true
|
storycloze
| 1
|
[
"lighteval"
] |
[
"lighteval:storycloze:2018"
] | false
|
lighteval:bigbench_lite:strange_stories
|
bigbench_lite:strange_stories
|
lighteval
|
tasksource/bigbench
|
strange_stories
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,295
|
direct_call
| true
|
bigbench_lite
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench_lite:strange_stories"
] | false
|
lighteval:bigbench_lite:strategyqa
|
bigbench_lite:strategyqa
|
lighteval
|
tasksource/bigbench
|
strategyqa
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,310
|
direct_call
| true
|
bigbench_lite
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench_lite:strategyqa"
] | false
|
lighteval:bigbench:sufficient_information
|
bigbench:sufficient_information
|
lighteval
|
tasksource/bigbench
|
sufficient_information
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,325
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:sufficient_information"
] | false
|
lighteval:bigbench:suicide_risk
|
bigbench:suicide_risk
|
lighteval
|
tasksource/bigbench
|
suicide_risk
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,340
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:suicide_risk"
] | false
|
helm:summarization:cnn-dm
|
summarization:cnn-dm
|
helm
|
lighteval/summarization
|
cnn-dm
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,355
|
direct_call
| true
|
summarization
| 1
|
[
"helm"
] |
[
"helm:summarization:cnn-dm"
] | false
|
helm:summarization:xsum
|
summarization:xsum
|
helm
|
lighteval/summarization
|
xsum
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,377
|
direct_call
| true
|
summarization
| 1
|
[
"helm"
] |
[
"helm:summarization:xsum"
] | false
|
helm:summarization:xsum-sampled
|
summarization:xsum-sampled
|
helm
|
lighteval/summarization
|
xsum-sampled
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,399
|
direct_call
| true
|
summarization
| 1
|
[
"helm"
] |
[
"helm:summarization:xsum-sampled"
] | false
|
lighteval:super_glue:boolq
|
super_glue:boolq
|
lighteval
|
super_glue
|
boolq
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,421
|
direct_call
| true
|
super_glue
| 1
|
[
"lighteval"
] |
[
"lighteval:super_glue:boolq"
] | false
|
lighteval:super_glue:cb
|
super_glue:cb
|
lighteval
|
super_glue
|
cb
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,436
|
direct_call
| true
|
super_glue
| 1
|
[
"lighteval"
] |
[
"lighteval:super_glue:cb"
] | false
|
lighteval:super_glue:copa
|
super_glue:copa
|
lighteval
|
super_glue
|
copa
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,451
|
direct_call
| true
|
super_glue
| 1
|
[
"lighteval"
] |
[
"lighteval:super_glue:copa"
] | false
|
lighteval:super_glue:rte
|
super_glue:rte
|
lighteval
|
super_glue
|
rte
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,466
|
direct_call
| true
|
super_glue
| 1
|
[
"lighteval"
] |
[
"lighteval:super_glue:rte"
] | false
|
lighteval:super_glue:multirc
|
super_glue:multirc
|
lighteval
|
super_glue
|
multirc
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,481
|
direct_call
| true
|
super_glue
| 1
|
[
"lighteval"
] |
[
"lighteval:super_glue:multirc"
] | false
|
lighteval:super_glue:wic
|
super_glue:wic
|
lighteval
|
super_glue
|
wic
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,496
|
direct_call
| true
|
super_glue
| 1
|
[
"lighteval"
] |
[
"lighteval:super_glue:wic"
] | false
|
lighteval:super_glue:wsc
|
super_glue:wsc
|
lighteval
|
super_glue
|
wsc
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,511
|
direct_call
| true
|
super_glue
| 1
|
[
"lighteval"
] |
[
"lighteval:super_glue:wsc"
] | false
|
lighteval:bigbench:swahili_english_proverbs
|
bigbench:swahili_english_proverbs
|
lighteval
|
tasksource/bigbench
|
swahili_english_proverbs
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,526
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:swahili_english_proverbs"
] | false
|
lighteval:swag
|
swag
|
lighteval
|
swag
|
regular
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,541
|
direct_call
| false
|
swag
| 1
|
[
"lighteval"
] |
[
"lighteval:swag"
] | true
|
lighteval:bigbench:swedish_to_german_proverbs
|
bigbench:swedish_to_german_proverbs
|
lighteval
|
tasksource/bigbench
|
swedish_to_german_proverbs
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,559
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:swedish_to_german_proverbs"
] | false
|
lighteval:bigbench_lite:symbol_interpretation
|
bigbench_lite:symbol_interpretation
|
lighteval
|
tasksource/bigbench
|
symbol_interpretation
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,574
|
direct_call
| true
|
bigbench_lite
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench_lite:symbol_interpretation"
] | false
|
helm:synthetic_reasoning:induction
|
synthetic_reasoning:induction
|
helm
|
lighteval/synthetic_reasoning
|
induction
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,589
|
direct_call
| true
|
synthetic_reasoning
| 1
|
[
"helm"
] |
[
"helm:synthetic_reasoning:induction"
] | false
|
helm:synthetic_reasoning:natural_easy
|
synthetic_reasoning:natural_easy
|
helm
|
lighteval/synthetic_reasoning_natural
|
easy
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,615
|
direct_call
| true
|
synthetic_reasoning
| 1
|
[
"helm"
] |
[
"helm:synthetic_reasoning:natural_easy"
] | false
|
helm:synthetic_reasoning:natural_hard
|
synthetic_reasoning:natural_hard
|
helm
|
lighteval/synthetic_reasoning_natural
|
hard
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,630
|
direct_call
| true
|
synthetic_reasoning
| 1
|
[
"helm"
] |
[
"helm:synthetic_reasoning:natural_hard"
] | false
|
helm:synthetic_reasoning:pattern_match
|
synthetic_reasoning:pattern_match
|
helm
|
lighteval/synthetic_reasoning
|
pattern_match
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,645
|
direct_call
| true
|
synthetic_reasoning
| 1
|
[
"helm"
] |
[
"helm:synthetic_reasoning:pattern_match"
] | false
|
helm:synthetic_reasoning:variable_substitution
|
synthetic_reasoning:variable_substitution
|
helm
|
lighteval/synthetic_reasoning
|
variable_substitution
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,671
|
direct_call
| true
|
synthetic_reasoning
| 1
|
[
"helm"
] |
[
"helm:synthetic_reasoning:variable_substitution"
] | false
|
lighteval:bigbench:tellmewhy
|
bigbench:tellmewhy
|
lighteval
|
tasksource/bigbench
|
tellmewhy
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,697
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:tellmewhy"
] | false
|
lighteval:bigbench:tense
|
bigbench:tense
|
lighteval
|
tasksource/bigbench
|
tense
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,727
|
direct_call
| true
|
bigbench
| 1
|
[
"lighteval"
] |
[
"lighteval:bigbench:tense"
] | false
|
helm:the_pile:arxiv
|
the_pile:arxiv
|
helm
|
lighteval/pile_helm
|
arxiv
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,742
|
direct_call
| true
|
the_pile
| 1
|
[
"helm"
] |
[
"helm:the_pile:arxiv"
] | false
|
helm:the_pile:bibliotik
|
the_pile:bibliotik
|
helm
|
lighteval/pile_helm
|
bibliotik
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,757
|
direct_call
| true
|
the_pile
| 1
|
[
"helm"
] |
[
"helm:the_pile:bibliotik"
] | false
|
helm:the_pile:commoncrawl
|
the_pile:commoncrawl
|
helm
|
lighteval/pile_helm
|
commoncrawl
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,772
|
direct_call
| true
|
the_pile
| 1
|
[
"helm"
] |
[
"helm:the_pile:commoncrawl"
] | false
|
helm:the_pile:dm-mathematics
|
the_pile:dm-mathematics
|
helm
|
lighteval/pile_helm
|
dm-mathematics
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,787
|
direct_call
| true
|
the_pile
| 1
|
[
"helm"
] |
[
"helm:the_pile:dm-mathematics"
] | false
|
helm:the_pile:enron
|
the_pile:enron
|
helm
|
lighteval/pile_helm
|
enron
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,802
|
direct_call
| true
|
the_pile
| 1
|
[
"helm"
] |
[
"helm:the_pile:enron"
] | false
|
helm:the_pile:europarl
|
the_pile:europarl
|
helm
|
lighteval/pile_helm
|
europarl
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,817
|
direct_call
| true
|
the_pile
| 1
|
[
"helm"
] |
[
"helm:the_pile:europarl"
] | false
|
helm:the_pile:freelaw
|
the_pile:freelaw
|
helm
|
lighteval/pile_helm
|
freelaw
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,832
|
direct_call
| true
|
the_pile
| 1
|
[
"helm"
] |
[
"helm:the_pile:freelaw"
] | false
|
helm:the_pile:github
|
the_pile:github
|
helm
|
lighteval/pile_helm
|
github
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,847
|
direct_call
| true
|
the_pile
| 1
|
[
"helm"
] |
[
"helm:the_pile:github"
] | false
|
helm:the_pile:gutenberg
|
the_pile:gutenberg
|
helm
|
lighteval/pile_helm
|
gutenberg
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,862
|
direct_call
| true
|
the_pile
| 1
|
[
"helm"
] |
[
"helm:the_pile:gutenberg"
] | false
|
helm:the_pile:hackernews
|
the_pile:hackernews
|
helm
|
lighteval/pile_helm
|
hackernews
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,877
|
direct_call
| true
|
the_pile
| 1
|
[
"helm"
] |
[
"helm:the_pile:hackernews"
] | false
|
helm:the_pile:nih-exporter
|
the_pile:nih-exporter
|
helm
|
lighteval/pile_helm
|
nih-exporter
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,892
|
direct_call
| true
|
the_pile
| 1
|
[
"helm"
] |
[
"helm:the_pile:nih-exporter"
] | false
|
helm:the_pile:opensubtitles
|
the_pile:opensubtitles
|
helm
|
lighteval/pile_helm
|
opensubtitles
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,907
|
direct_call
| true
|
the_pile
| 1
|
[
"helm"
] |
[
"helm:the_pile:opensubtitles"
] | false
|
helm:the_pile:openwebtext2
|
the_pile:openwebtext2
|
helm
|
lighteval/pile_helm
|
openwebtext2
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,922
|
direct_call
| true
|
the_pile
| 1
|
[
"helm"
] |
[
"helm:the_pile:openwebtext2"
] | false
|
helm:the_pile:pubmed-abstracts
|
the_pile:pubmed-abstracts
|
helm
|
lighteval/pile_helm
|
pubmed-abstracts
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,937
|
direct_call
| true
|
the_pile
| 1
|
[
"helm"
] |
[
"helm:the_pile:pubmed-abstracts"
] | false
|
helm:the_pile:pubmed-central
|
the_pile:pubmed-central
|
helm
|
lighteval/pile_helm
|
pubmed-central
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,952
|
direct_call
| true
|
the_pile
| 1
|
[
"helm"
] |
[
"helm:the_pile:pubmed-central"
] | false
|
helm:the_pile:stackexchange
|
the_pile:stackexchange
|
helm
|
lighteval/pile_helm
|
stackexchange
|
/Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py
| 16,967
|
direct_call
| true
|
the_pile
| 1
|
[
"helm"
] |
[
"helm:the_pile:stackexchange"
] | false
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.