inputs stringlengths 38 313k | targets stringlengths 0 4.86k | _template_idx int64 0 9 | _task_source stringclasses 1
value | _task_name stringlengths 19 85 | _template_type stringclasses 2
values | embedding listlengths 1.02k 1.02k |
|---|---|---|---|---|---|---|
instruction:
Generate a topic word for the given sentence. A topic word is usually the name of a person, place, animal, concept or object the sentence is focusing on.
question:
child development refers to biological, psychological and emotional changes that happen in humans between birth and the end of adolescence, at ... | shepreth
| 9 | NIv2 | task645_summarization | fs_opt | [
0.20023919641971588,
0.17994913458824158,
-0.4165772795677185,
-0.16807398200035095,
-0.6082602739334106,
-0.7590060234069824,
0.0715254470705986,
0.3373364806175232,
0.15590544044971466,
-0.850281298160553,
-0.11715102940797806,
0.23711642622947693,
-1.085972785949707,
0.25884002447128296... |
You will be given a definition of a task first, then some input of the task.
A text is given in Panjabi. Translate it from the Panjabi language to the Tamil language. The translation must not omit or add information to the original sentence.
திருப்பூர் அருகே உள்ள பெருமாநல்லூர் கிராமத்தில் நடைபெற்ற விழாவில், பல்வேறு வள... | ਪ੍ਰਧਾਨ ਮੰਤਰੀ ਨੇ ਤਿਰੂਪੁਰ ਦੇ ਪੇਰੂਮੰਨਲੂਰ ਪਿੰਡ ਵਿੱਚ ਕਈ ਵਿਕਾਸ ਪ੍ਰੋਜੈਕਟ ਦਾ ਉਦਘਾਟਨ ਕੀਤਾ। | 1 | NIv2 | task1016_pib_translation_tamil_punjabi | zs_opt | [
0.3806263208389282,
0.6967498064041138,
-0.15627725422382355,
0.2558242082595825,
0.001430792035534978,
-0.7105085849761963,
0.8284541368484497,
0.5786981582641602,
-0.09274283796548843,
0.5753283500671387,
-0.5080763101577759,
0.08409659564495087,
-0.831657350063324,
0.38432079553604126,
... |
You will be given a definition of a task first, then some input of the task.
Categorize the comment on the basis of severe toxicity. If the comment is severely toxic output Yes, otherwise output No.
Let him hang
Output: | Yes | 1 | NIv2 | task1725_civil_comments_severtoxicity_classification | zs_opt | [
-0.21693235635757446,
0.3710702061653137,
0.5126121044158936,
0.0933082103729248,
-0.10306807607412338,
-0.6766558289527893,
0.65589439868927,
0.5992007255554199,
0.16494405269622803,
0.4843577742576599,
-0.44445204734802246,
0.022917479276657104,
-0.11464079469442368,
-0.4408804178237915,... |
Given the task definition and input, reply with output. You are given a paragraph, with numbered sentences, and a question. Write the numbers of the sentences needed to answer the given question. Separate multiple sentence numbers with a comma (", "). E.g. "1, 2" would be the correct response if Sent 1 and Sent 2 are n... | 1, 3. | 5 | NIv2 | task049_multirc_questions_needed_to_answer | zs_opt | [
0.4523957073688507,
0.3404579162597656,
-0.29354143142700195,
-0.0013036224991083145,
0.25509703159332275,
-0.35602858662605286,
0.7803512811660767,
0.5059555172920227,
-0.5864365696907043,
0.05710972473025322,
-0.502057671546936,
0.20237502455711365,
-0.8238778114318848,
0.088139921426773... |
Detailed Instructions: In this task, you will be shown an extract from a movie plot and a question. You need to provide the correct answer for it. Short answers containing words that are present in the passage are preferred.
Q: This article's plot summary may be too long or excessively detailed. Please help improve it ... | Answer: A Sphere. | 9 | NIv2 | task194_duorc_answer_generation | zs_opt | [
0.5512210726737976,
0.3048609793186188,
-0.2907380163669586,
-0.7411141395568848,
-0.0015208381228148937,
-0.05229805409908295,
0.046246666461229324,
0.6931378841400146,
-0.5310620665550232,
0.1334407925605774,
-0.2140740305185318,
0.685116171836853,
-0.18649162352085114,
0.050214562565088... |
In this task, you are given two facts, and a multiple-choice question. Based on the given facts, answer the question with index of the correct option (e.g, "A").
Example Input: Fact1: Secondary meristem is located within and around the vascular tissues., Fact2: Plant tissues are produced from meristems., Question: Wha... | H
| 3 | NIv2 | task1297_qasc_question_answering | fs_opt | [
0.5180855989456177,
0.17297950387001038,
-0.7071354985237122,
0.7973112463951111,
-0.2674269378185272,
-0.9640873670578003,
0.05907832831144333,
0.9365221261978149,
0.044507671147584915,
-0.13249757885932922,
-0.5603652000427246,
0.439873605966568,
-0.08941638469696045,
0.18809214234352112... |
Given the task definition, example input & output, solve the new input case.
You are given a sentence in Spanish. Your job is to translate the Spanish sentence into Hebrew.
Example: Pero comenzamos con la idea de un niño por vez.
Output: אבל התחלנו לממש את הרעיון של ללמד כל ילד בנפרד.
The Spanish sentence is correctly ... | קניתי לה בגדים משומשים וקניתי לה את הבובה הראשונה שלה. | 1 | NIv2 | task1229_ted_translation_es_he | fs_opt | [
-0.7045596837997437,
0.606110692024231,
-0.061723493039608,
0.17041254043579102,
-0.13319523632526398,
0.46471506357192993,
0.6323596835136414,
0.781709611415863,
0.3671860694885254,
-0.44490718841552734,
-0.32392072677612305,
0.22659778594970703,
-0.5422555208206177,
-0.14686158299446106,... |
Given the task definition, example input & output, solve the new input case.
Given an Amazon customer review, write a title for the review. The preferred titles are under fifteen words.
Example: I was very surprised at the high quality of the stitching, the sturdiness of the handles and the padding for my laptop. The p... | Simple, easy, use it right away | 1 | NIv2 | task1342_amazon_us_reviews_title | fs_opt | [
-0.35294461250305176,
0.2630242109298706,
-0.2703113853931427,
0.18067815899848938,
0.9236404895782471,
0.0813906267285347,
0.4109225869178772,
0.7555415630340576,
0.22535723447799683,
0.7605326771736145,
-0.07480591535568237,
0.28349602222442627,
-0.056524187326431274,
-0.4417437911033630... |
Detailed Instructions: You are given a sentence in Japanese. Your job is to translate the Japanese sentence into Galician.
Problem:この時点で、子供は家へ帰れるのでしょうか ?
Solution: | Nesa altura, ¿pode regresar o neno a casa? | 8 | NIv2 | task1095_ted_translation_ja_gl | zs_opt | [
0.45041483640670776,
0.9932952523231506,
-0.024781249463558197,
-0.29069992899894714,
-0.21192197501659393,
-0.67303067445755,
0.5453006029129028,
0.8873908519744873,
-0.5969940423965454,
-0.34764885902404785,
-0.14667972922325134,
1.3713243007659912,
-0.47948795557022095,
0.18490564823150... |
In this task, you're given a four sentences of story written in natural language. Your job is to complete end part of the story by predicting appropriate last sentence which is coherent with the given sentences.
One example is below.
Q: Sentence1: Rick grew up in a troubled household. Sentence2: He never found good sup... | Amazingly, she won the part. | 9 | NIv2 | task105_story_cloze-rocstories_sentence_generation | fs_opt | [
-0.20091399550437927,
1.4504787921905518,
-0.01278490200638771,
-0.7642013430595398,
-0.22715425491333008,
-0.0410672165453434,
0.63984215259552,
0.38304466009140015,
0.39027053117752075,
0.15216699242591858,
-0.2777929902076721,
-0.06083740293979645,
-0.5415946245193481,
0.300659149885177... |
Given the task definition, example input & output, solve the new input case.
In this task, you are given an input list A. You need to find all the elements of the list that are numbers and calculate their sum.
Example: ['238', 'h', '92', 'U', '2799']
Output: 3129
Here, the numbers in the input list are '238', '92' and ... | 33412 | 1 | NIv2 | task499_extract_and_add_all_numbers_from_list | fs_opt | [
-0.7460570931434631,
0.33516910672187805,
-0.35610124468803406,
-0.4509347081184387,
0.35914573073387146,
-0.08974483609199524,
-0.3879179358482361,
0.050509434193372726,
-0.7487916946411133,
-0.2767994999885559,
-0.7553877830505371,
-0.0796123519539833,
0.5653066635131836,
-0.448255151510... |
Detailed Instructions: Given a post that is a real-life anecdote of a complex ethical situation and an associated claim about its type, verify if the claim is true or not. The claim asks if the posts are historical or hypothetical. The posts are "HISTORICAL" when the author has already done something and they are "HYPO... | no | 8 | NIv2 | task501_scruples_anecdotes_post_type_verification | zs_opt | [
-0.7898770570755005,
0.2613539397716522,
0.6964956521987915,
0.36108776926994324,
-0.3316381275653839,
-0.7614154815673828,
0.6747276782989502,
0.7271745204925537,
-0.3502199053764343,
0.09343749284744263,
-0.12541711330413818,
0.37479168176651,
-0.5319184064865112,
-0.05803503841161728,
... |
Detailed Instructions: In this task, you're given an ambiguous question (which can be answered in more than one way). Your task is to provide one question which clarifies the input question and it has one unique answer, and also provide an answer to the generated question. Generated question and answer should be separa... | What was last the world cup england won?
1966 FIFA World Cup | 9 | NIv2 | task671_ambigqa_text_generation | zs_opt | [
-0.29838210344314575,
0.9175251722335815,
0.39128774404525757,
-0.8843308091163635,
-0.05061184614896774,
-0.7284571528434753,
0.14138753712177277,
0.04049713909626007,
-0.6072194576263428,
-0.5378113389015198,
-0.6754703521728516,
0.43913671374320984,
-0.2072899341583252,
0.04308597743511... |
Given a sentence in the Japanese and Filipino language. Your task is check if the Filipino sentence is translation of Japanese. if the translation is correct than generate label "Yes", otherwise generate label "No".
One example is below.
Q: Japanese: 詳細は昨日UTC17時30分、英国議会でイギリスのルス・ケリー運輸大臣によって伝えられた。
Filipino: Ang mga det... | No | 9 | NIv2 | task1120_alt_ja_fil_answer_generation | fs_opt | [
-0.8669588565826416,
-0.6780481338500977,
0.1395692229270935,
0.09868813306093216,
0.14723441004753113,
-0.7521005272865295,
0.45630943775177,
0.3273126780986786,
-0.2615508437156677,
-0.38676631450653076,
0.04086752235889435,
0.9320008158683777,
-0.07711745798587799,
0.915936291217804,
... |
Given the task definition and input, reply with output. Given a real-life anecdote of a complex ethical situation, identify who is wrong here. Classify each post into the following categories : AUTHOR (if the author of the post is wrong in the situation), OTHER (if any of the others mentioned in the paragraph is wrong)... | OTHER | 5 | NIv2 | task498_scruples_anecdotes_whoiswrong_classification | zs_opt | [
0.16951601207256317,
0.26080137491226196,
-0.058874621987342834,
-0.6567946076393127,
0.08754023909568787,
-0.40425294637680054,
0.4323292374610901,
0.6069880723953247,
0.207830548286438,
-0.37527793645858765,
-0.2750256061553955,
0.31568413972854614,
-0.5540894269943237,
-0.31500253081321... |
You will be given a definition of a task first, then some input of the task.
In this task, you need to Translate Czech text to English.
Rešení se týká zpusobu prumyslové výroby vetšího množství amorfní formy hemivápenaté soli (3R,5R) 7-[3-fenyl-4-fenylkarbamoyl-2-(4-fluorfenyl)-5-isopropylpyrrol-1-yl]-3,5-dihydroxyhep... | The present invention relates to a process for the industrial production of more amorphous form of (3R, 5R) 7- [3-phenyl-4-phenylcarbamoyl-2- (4-fluorophenyl) -5-isopropylpyrrol-1-yl] -3,5-dihydroxyheptanoic hemi-calcium salt. low surface area (atorvastatin of formula I) based on controlled precipitation and its use in... | 1 | NIv2 | task842_para_pdt_cs_en_translation | zs_opt | [
0.0493728443980217,
0.2739468812942505,
0.11235970258712769,
-0.13569006323814392,
0.053295571357011795,
0.5011024475097656,
0.27081894874572754,
0.9880713224411011,
0.48757272958755493,
-0.2916027903556824,
-0.36111217737197876,
0.6448112726211548,
-0.031141089275479317,
0.272569686174392... |
In this task, you need to remove all words of a given length in the sentence. The number of letters in a word determine its length, for example, the length of the word "apple" is 5.
One example is below.
Q: Sentence: 'breakfast and a glass of milk is on the table'. Remove all words of length '3' in the given sentence.
... | couple of women are under an umbrella | 9 | NIv2 | task377_remove_words_of_given_length | fs_opt | [
0.3466586470603943,
0.23257820308208466,
-0.21358993649482727,
-0.6324073672294617,
0.12828266620635986,
-0.7909702658653259,
0.07200220227241516,
0.5719027519226074,
-0.008043646812438965,
-0.3521999716758728,
-1.1868515014648438,
-0.710607647895813,
-0.6324774622917175,
-0.41617867350578... |
You are given a sentence in Hebrew. Your job is to translate the Hebrew sentence into Galician.
Q: אתה לוחץ התחל, אף אחד לא נוגע בדבר, והיא מסוגלת להגיע ליעד שלה בצורה אוטונומית לחלוטין.
A: | Púlsase o botón de inicio, e sen tocar nada, pódese chegar ao destino de forma completamente autónoma. | 4 | NIv2 | task1110_ted_translation_he_gl | zs_opt | [
-0.1847292184829712,
1.230409026145935,
-0.0008802349912002683,
-0.40391090512275696,
-0.44262176752090454,
0.03101259469985962,
1.1458783149719238,
0.7346686124801636,
0.3251684606075287,
-0.4965611696243286,
0.08020003885030746,
0.49267828464508057,
-0.6117438077926636,
-0.12658953666687... |
A text is given in Panjabi. Translate it from the Panjabi language to the Urdu language. The translation must not omit or add information to the original sentence.
Example: درخواست واپس لینے کا عمل کامیابی کے ساتھ مکمل ہونے پر امیدوار کو درخواست واپس لینے کی تصدیق کے لئے ای میل اور ایس ایم ایس بھیجاجائے گا۔درخواست جب ا... | Solution: ਇਸ ਸਮਾਗਮ ਦੌਰਾਨ ਭਾਰਤ ਤੇ ਬਰਤਾਨੀਆ ਦੋਹਾਂ ਨੇ ਖਾਣ ਪੀਣ ਦੀਆਂ ਵਸਤਾਂ , ਸਿਹਤ ਸੇਵਾਵਾਂ ਅਤੇ ਡੇਟਾ ਸੇਵਾਵਾਂ ਦੇ ਵਪਾਰ ਵਿੱਚ ਆਉਣ ਵਾਲੇ ਅੜਿੱਕਿਆਂ ਨੂੰ ਦੂਰ ਕਰਨ ਲਈ ਤਿੰਨ ਨਵੇਂ ਦੁਵੱਲੇ ਕਾਰਜ ਸਮੂਹ ਗਠਿਤ ਕਰਨ ‘ਤੇ ਸਹਿਮਤੀ ਪ੍ਰਗਟਾਈ । | 5 | NIv2 | task1052_pib_translation_urdu_punjabi | fs_opt | [
-0.2959510385990143,
1.0660477876663208,
0.23296692967414856,
-0.06024348735809326,
-0.41662275791168213,
-0.8815134763717651,
0.7423447370529175,
0.36764729022979736,
-0.2997192442417145,
-0.1327478587627411,
-0.23812749981880188,
-0.1237402856349945,
-0.08751115202903748,
0.1107454597949... |
You will be given a sentence containing a pronoun/person name and an emotion. From these implicit parameters, the main goal is to find the gender of the person (male / female).
Alan told us all about the recent hilarious events. | male | 0 | NIv2 | task1336_peixian_equity_evaluation_corpus_gender_classifier | zs_opt | [
-0.33221501111984253,
0.5168811082839966,
-0.22560067474842072,
-0.684755802154541,
-0.48092690110206604,
-0.34835293889045715,
0.26270508766174316,
0.11320802569389343,
0.657558262348175,
0.20873631536960602,
-1.06270170211792,
-0.05172542855143547,
-0.7172412872314453,
0.3096334338188171... |
Part 1. Definition
In this task, you will be presented with a text, a pronoun from the text, and two candidate names. You should determine what the pronoun refers to and classify the answers into A, B, or Neither. A and B here are referring to option A and option B. Position of the pronoun in the text is showed within ... | B | 7 | NIv2 | task329_gap_classification | fs_opt | [
-0.019210226833820343,
0.2062709629535675,
0.2338564097881317,
-0.2218143790960312,
-0.21946030855178833,
-0.09229540824890137,
0.8072242736816406,
1.0266832113265991,
0.5730076432228088,
-0.024669870734214783,
-0.6537794470787048,
0.5611398220062256,
-0.3687724173069,
-0.44507351517677307... |
In this task you are given a tweet and you must identify whether the tweet contains any offense or any form of (untargeted) profanity. Label the post as NOT if the post does not contain offense or profanity. Non-offensive posts do not include any form of offense or profanity. label the post as OFFENSIVE if the post co... | NOT
| 0 | NIv2 | task286_olid_offense_judgment | fs_opt | [
-1.1601454019546509,
0.5504993200302124,
0.4923418462276459,
0.16031725704669952,
-0.3220842480659485,
-0.538352370262146,
0.48784685134887695,
0.3701023459434509,
0.7401720285415649,
0.17150726914405823,
0.03840680420398712,
0.11538827419281006,
-0.6631280183792114,
-0.47264838218688965,
... |
In the following task, you are given a yes/no question, its answer, and additional information which includes a description of a key term in the question and several facts. Your task is to "decompose the question"i.e. write the steps required to construct the given answer where each step is a question that can be answe... | Step1: Which element is the pigment verdigris derived from?, Step2: Is #1 copper and verdigris also used as a fungicide?, Wikipedia page for step 1: Verdigris, Wikipedia page for step 2: Verdigris, not_definitive_answer: False, incorrect_answer: False | 0 | NIv2 | task168_strategyqa_question_decomposition | zs_opt | [
0.969504714012146,
-0.06773364543914795,
-0.5680842995643616,
-0.0241162721067667,
-0.3285045623779297,
-0.9380969405174255,
1.066127061843872,
0.9837414026260376,
-0.00316671677865088,
0.26547980308532715,
-0.6290615200996399,
0.22066166996955872,
-0.16657350957393646,
0.19895018637180328... |
Part 1. Definition
You are given a sentence in English. Your job is to translate the English sentence into Japanese.
Part 2. Example
And it was primarily because Kiribati realized that this was in their own self-interest to do this.
Answer: 第一にキリバス共和国が彼ら自身の利益になると理解したからです
Explanation: The English sentence is correctly t... | まったく文化的なものだと | 7 | NIv2 | task1218_ted_translation_en_ja | fs_opt | [
0.2832602262496948,
0.48404937982559204,
0.005511992610991001,
-0.2315722107887268,
-0.11631709337234497,
-0.5460371375083923,
0.9252772331237793,
-0.6707518100738525,
-0.006809836253523827,
-0.3518761992454529,
-0.674534261226654,
-0.05149085074663162,
-0.3811926245689392,
0.5496394634246... |
Teacher:Generate an appropriate title for the given text. The generated title must be short and include the main topic of the text. The preferred titles are under fifteen words.
Teacher: Now, understand the problem? Solve this instance: The first minister signed a memorandum of understanding (MOU) with SinoFortone and ... | Row over apparent collapse of Scots-China investment pact | 6 | NIv2 | task1356_xlsum_title_generation | zs_opt | [
0.11561243236064911,
0.38834497332572937,
-0.3011798858642578,
-0.3164310157299042,
0.7512979507446289,
-0.20652595162391663,
0.3977081775665283,
0.9202693700790405,
-0.15669003129005432,
0.2180432826280594,
0.7795113325119019,
0.9940393567085266,
-0.04977024719119072,
0.2332448959350586,
... |
In this task, you will be shown a short story with a beginning, two potential middles, and an ending. Your job is to choose the middle statement that makes the story incoherent / implausible by indicating 1 or 2 in the output. If both sentences are plausible, pick the one that makes less sense.
[EX Q]: Beginning: Chuc... | 1
| 6 | NIv2 | task070_abductivenli_incorrect_classification | fs_opt | [
-0.12961162626743317,
0.09613270312547684,
-0.3243660628795624,
0.168828547000885,
-0.9074883460998535,
-0.3080228567123413,
0.39676544070243835,
0.7235896587371826,
0.3679448664188385,
-0.3622332811355591,
-0.7743173241615295,
-0.27936506271362305,
-0.18563348054885864,
-0.486795097589492... |
In this task you will be given a list of integers. You should only return an integer if the first digit is the same as the last digit in the number. If an integer has a single digit, then it should be returned. If there are no integers that start and end with the same digit then an empty list ("[]") should be returned.... | [] | 3 | NIv2 | task372_synthetic_palindrome_numbers | fs_opt | [
-0.6178319454193115,
0.4642873704433441,
-0.19833160936832428,
-0.7881254553794861,
-0.24654492735862732,
0.1370515376329422,
1.107214331626892,
0.07258340716362,
-0.2930668592453003,
0.30437028408050537,
-0.3133126199245453,
0.21223264932632446,
0.04786312207579613,
-0.2149335891008377,
... |
Detailed Instructions: In this task, a passage will be given and the goal is to generate a question about temporal relations based on that passage. A temporal relation describes the relation between two things with respect to time e.g., something happens/starts/terminates/... immediately/shortly/usually/... before/afte... | What event has already happened? | 8 | NIv2 | task389_torque_generate_temporal_question | zs_opt | [
-0.5973731279373169,
0.8117358684539795,
0.22941668331623077,
0.09988798201084137,
-0.12848974764347076,
-0.4463695287704468,
0.15967941284179688,
0.6126202940940857,
0.5687974691390991,
0.41735678911209106,
-0.28664958477020264,
-0.1692105233669281,
-1.073594570159912,
0.461819052696228,
... |
This task is about generating an incorrect answer to a question given the question and a true statement related to the question. The answer must be related to the question, but incorrectly answer it given the context.
Fact: a stove generates heat for cooking usually. Question: which one of these can help a person cook... | a stack of clothes | 0 | NIv2 | task1400_obqa_incorrect_answer_generation | zs_opt | [
1.0850822925567627,
0.8816654682159424,
-0.5370121002197266,
0.702709972858429,
-0.19106480479240417,
-0.41254717111587524,
0.6867940425872803,
0.6124420762062073,
-0.04899617284536362,
-0.30195996165275574,
0.009823477827012539,
0.024452317506074905,
-0.6877803206443787,
0.146775275468826... |
In this task, you are given an input list A. You need to convert all the alphabets in the list with a number representing their position in the English alphabet. E.g., replace A by 1, B by 2, a by 1, b by 2, and so on.
['Q', '2971', '5709', '4005', '4209', 'A', '159', '4273'] | 17, 2971, 5709, 4005, 4209, 1, 159, 4273 | 0 | NIv2 | task622_replace_alphabets_in_a_list_by_their_position_in_english_alphabet | zs_opt | [
-0.5823842883110046,
0.5331799387931824,
0.5348348021507263,
-0.8463524580001831,
-0.05160202831029892,
-0.589706301689148,
0.07421259582042694,
-0.08701006323099136,
0.23112091422080994,
-0.28166085481643677,
0.1690024435520172,
-0.0541863776743412,
0.1547655463218689,
-0.3245422244071960... |
You are given a sentence in Arabic. Your job is to translate the Arabic sentence into Polish.
Example Input: فى نخب نجاحاتكم المتواصلة
Example Output: Tak więc, życzę Wam nieustających sukcesów.
Example Input: و أساسا مشكلتنا هي وجود أحدهم في وسط هذه المرحلة الطويلة, يتوقفون فقط للحظة, و يتركون شيئا في الأنزيم لثانية... | Bywa też lepszy niż nam to wmówiono, ale musimy zobaczyć go takim, jaki jest, ze wszystkimi swoimi problemami. Dopiero widząc wszystkie problemy, możemy je rozwiązać i żyć w świecie, gdzie wszyscy żyją długo i szczęśliwie. (Śmieje się)
| 3 | NIv2 | task1107_ted_translation_ar_pl | fs_opt | [
-0.4066479802131653,
0.65691739320755,
-0.21512731909751892,
-0.4332033693790436,
-0.602714478969574,
-0.20632806420326233,
0.8801098465919495,
0.5110476016998291,
-0.012315062806010246,
0.014701145701110363,
-0.8208432197570801,
0.8843475580215454,
-0.6851083040237427,
0.46212947368621826... |
You will be given a definition of a task first, then some input of the task.
In this task, you're given a review from Amazon and your task is to generate the name of the category of the product based on the review given by the user. The categories are: kitchen, office product, watch, wireless, other, toy, digital video... | electronics | 1 | NIv2 | task617_amazonreview_category_text_generation | zs_opt | [
-0.5373115539550781,
-0.13602551817893982,
-0.4332618713378906,
0.39556002616882324,
0.5593624114990234,
0.4847244322299957,
0.274236798286438,
0.4956774115562439,
-0.3598018288612366,
0.4681205749511719,
-0.000519562978297472,
-0.10581757873296738,
-0.4750567078590393,
-0.4918506741523742... |
Detailed Instructions: In this task, you are given a passage which has a question and the context. You have to generate an answer to the question based on the information present in the context.
See one example below:
Problem: Context: Chronic rhinosinusitis (CRS) is a heterogeneous disease with an uncertain pathogenes... | Our results indicated that CSTMP could induce apoptosis and mitochondrial dysfunction in RPMI8226 cells via CHOP-dependent ER stress. | 4 | NIv2 | task849_pubmedqa_answer_generation | fs_opt | [
0.205876886844635,
0.5719090700149536,
-0.3884480893611908,
-0.004083888605237007,
0.35761186480522156,
-0.451141357421875,
0.2282816469669342,
0.7233549356460571,
0.019577395170927048,
0.36764147877693176,
-0.3961016535758972,
-0.11061593890190125,
-0.661065936088562,
0.2720632553100586,
... |
Given the task definition and input, reply with output. Given a premise, an initial context, an original ending, and a counterfactual context, the task is to generate a new story ending aligned with the counterfactual context and as close to the original ending as possible. Each instance consists of a five-sentence sto... | Normally she was the center of attention, and now had gotten her way. She told the photographer she was open to anything as he took the pictures. Her sisters were abhorred and it created tension between them. | 5 | NIv2 | task269_csrg_counterfactual_story_generation | zs_opt | [
0.222197487950325,
-0.16068649291992188,
0.17186713218688965,
-0.06764382123947144,
-0.05385969579219818,
-0.5208808183670044,
-0.31104257702827454,
1.06227707862854,
-0.05463334172964096,
0.013776984065771103,
-0.9339959621429443,
0.19093501567840576,
-0.31549468636512756,
0.2443615496158... |
Given the task definition, example input & output, solve the new input case.
In this task, you are given a set of context paragraphs, some supporting facts and an answer of a question. Your task is to generate question for given answer based on set of context paragraphs, supporting facts and an answer.
Example: Context... | Which female film directors career spanned more than one film era, Dorothy Arzner or Gurinder Chadha? | 1 | NIv2 | task191_hotpotqa_question_generation | fs_opt | [
0.7686780691146851,
0.331909716129303,
-0.8883602619171143,
0.9677903056144714,
0.6835002899169922,
-0.01891596056520939,
0.6804456114768982,
0.5217878818511963,
-0.09768174588680267,
0.4193882942199707,
-0.06787365674972534,
0.5116052627563477,
-0.8114304542541504,
-0.09846308827400208,
... |
Q: This task is about using the specified sentence and converting the sentence to Resource Description Framework (RDF) triplets of the form (subject, predicate object). The RDF triplets generated must be such that the triplets accurately capture the structure and semantics of the input sentence. The input is a sentence... | [['Azerbaijan', 'CAPITAL', 'Baku'], ["Baku Turkish Martyrs' Memorial", 'MATERIAL', '"Red granite and white marble"'], ['Azerbaijan', 'LEADER_TITLE', 'Prime Minister of Azerbaijan'], ["Baku Turkish Martyrs' Memorial", 'DEDICATED_TO', '"Ottoman Army soldiers killed in the Battle of Baku"'], ["Baku Turkish Martyrs' Memori... | 7 | NIv2 | task1410_dart_relationship_extraction | zs_opt | [
0.11741147935390472,
0.38278627395629883,
-0.4543219208717346,
-0.4737851917743683,
-0.19611051678657532,
-0.06886740028858185,
1.692000150680542,
0.13657639920711517,
0.2357691526412964,
0.42860931158065796,
-0.611611545085907,
0.5580452680587769,
0.18117383122444153,
0.5304913520812988,
... |
In this task you will be given a list of integers. You should remove all of the integers that are divisible by 3 from the list. If every integer in the input list is divisible by 3 then an empty list should be returned. Zero is divisible by 3.
One example is below.
Q: [2, 5, 9, 6, 11]
A: [2, 5, 11]
Rationale: 6 and 9 a... | [64, 47, 100] | 9 | NIv2 | task370_synthetic_remove_divisible_by_3 | fs_opt | [
0.03625943511724472,
-0.041057731956243515,
-0.2233906239271164,
-0.967121958732605,
0.39488309621810913,
-0.440155565738678,
0.9896541833877563,
0.09754704684019089,
-0.18537238240242004,
0.3298410177230835,
-0.7431906461715698,
0.03788759186863899,
-0.1679539829492569,
-0.451558589935302... |
In this task, you are given a context paragraph of the tweet and question. Your task is to generate right answer of given question based on given context tweet paragraph.
Q: Context: American Airlines #AA280 is diverting to Narita with 15 passengers injured in turbulence Flightradar24 (@flightradar24) December 16, 2014... | 15 | 4 | NIv2 | task239_tweetqa_answer_generation | zs_opt | [
0.4955459237098694,
0.050610363483428955,
-0.048356495797634125,
0.06749792397022247,
-0.5274760723114014,
0.8768383264541626,
0.34745216369628906,
0.6238605976104736,
-0.4990389049053192,
-0.10664836317300797,
-0.3156377673149109,
0.18131133913993835,
-0.013855421915650368,
-0.19785791635... |
You will be given a definition of a task first, then some input of the task.
In this task, you will be presented with a passage and have to answer a question based on your understanding of the events inferred from the passage. Among the entities, try to find the best entity that is most likely to fill in "_" and classi... | (A) | 1 | NIv2 | task302_record_classification | zs_opt | [
-0.15081334114074707,
0.7319259643554688,
-0.3312355875968933,
0.019709374755620956,
0.6066769361495972,
0.3589938282966614,
0.5845553874969482,
0.5899676084518433,
0.475829154253006,
0.1156342476606369,
-0.1688866913318634,
0.6406189799308777,
-0.7590868473052979,
0.1504165530204773,
0.... |
This task is about using the specified sentence and converting the sentence to Resource Description Framework (RDF) triplets of the form (subject, predicate object). The RDF triplets generated must be such that the triplets accurately capture the structure and semantics of the input sentence. The input is a sentence an... | [['Fitzbillies', 'eatType', 'restaurant'], ['Fitzbillies', 'food', 'Italian'], ['Fitzbillies', 'priceRange', 'more than £30'], ['Fitzbillies', 'near', 'The Six Bells']] | 3 | NIv2 | task1410_dart_relationship_extraction | fs_opt | [
-0.1996445655822754,
0.35638517141342163,
-0.5420775413513184,
0.1687895804643631,
-0.47082915902137756,
-0.1937393695116043,
0.05678939074277878,
0.2580071687698364,
0.010649897158145905,
-0.29559314250946045,
-0.5135879516601562,
0.16631144285202026,
0.004741494078189135,
0.605224609375,... |
Given the task definition and input, reply with output. You will be given a piece of text either about an everyday event, or a general statement. If the event seems a plausible event to you, or the general statement makes sense matches your commonsense, output 'True', otherwise output 'False'.
When heating some tea in... | False | 5 | NIv2 | task116_com2sense_commonsense_reasoning | zs_opt | [
-0.526970386505127,
0.5407053828239441,
-0.2339487373828888,
0.5830659866333008,
-0.4343402683734894,
-0.3947829008102417,
0.25449150800704956,
0.13771869242191315,
-0.03746010735630989,
-0.11864468455314636,
-0.31675079464912415,
0.06475168466567993,
0.22323238849639893,
-0.00304894126020... |
In this task, you're given an ambiguous question (which can be answered in more than one way). Your task is to provide one question which clarifies the input question and it has one unique answer, and also provide an answer to the generated question. Generated question and answer should be separated with a new line.
[... | When, based on evidence, did we begin calling ourselves the united states of america?
January 2, 1776
| 6 | NIv2 | task671_ambigqa_text_generation | fs_opt | [
-1.1226592063903809,
0.31669774651527405,
0.0713861882686615,
0.37117302417755127,
-0.8889521956443787,
0.14987275004386902,
0.24050140380859375,
0.32954317331314087,
0.12229949235916138,
-0.45923805236816406,
-0.4252699613571167,
-0.43263769149780273,
-0.5192110538482666,
-0.2525906562805... |
In this task you are given a sentence. You must judge whether there exist two consecutive words within the sentence with flipped orders, that is, whether the sentence will make sense and be correct if the order of two consecutive words changes. Label the instances as "Inversion" or "Original" based on your judgment.
I... | Output: Original
| 2 | NIv2 | task428_senteval_inversion | fs_opt | [
-0.8894059062004089,
0.7765215635299683,
0.10388584434986115,
0.03580181673169136,
-0.23713107407093048,
0.0052552605047822,
0.3861781358718872,
0.30812111496925354,
0.3530966639518738,
-0.12086385488510132,
-0.5921527147293091,
-0.3065202236175537,
-0.4888449013233185,
-0.0821124464273452... |
In this task, you are given two phrases: Head and Tail, separated with <sep>. The Head and the Tail events are short phrases possibly involving participants. The names of specific people have been replaced by generic words (e.g., PersonX, PersonY, PersonZ). PersonX is always the subject of the event. You have to determ... | Yes
| 7 | NIv2 | task1205_atomic_classification_isafter | fs_opt | [
0.651993453502655,
0.38138920068740845,
0.40362977981567383,
-0.35122716426849365,
-0.5429041385650635,
-0.9282739162445068,
1.3855565786361694,
0.3154583275318146,
-0.4547167420387268,
-0.7344315648078918,
-0.42059653997421265,
-0.5416325330734253,
-0.5426772832870483,
0.5162909626960754,... |
Detailed Instructions: You are given a sentence in Arabic. Your job is to translate the Arabic sentence into Farsi.
See one example below:
Problem: وقد كتبن اللغة وغالبا ما كانت لغة آلة وأحيانا شفرة ثنائية ترسل بعدها بالبريد لمركز البيانات لتنقل على شريط أو بطاقة ورقية مثقبة ثم تعاد الكرة مرة أخرى للتأكد.
Solution: و آ... | اون دماغمه در راست ، و حالا می خواهیم بریم اینجا ، درست اونجا. | 4 | NIv2 | task1108_ted_translation_ar_fa | fs_opt | [
-0.41226547956466675,
0.1859067678451538,
-0.34461015462875366,
-0.6145222783088684,
-0.5066582560539246,
-0.2498522251844406,
1.1586298942565918,
0.1977289617061615,
0.2578699588775635,
-0.16887667775154114,
-1.1118206977844238,
0.47268569469451904,
-0.08751437813043594,
0.444306910037994... |
You are given a statement written in Kannada. Choose the most logical word from the given 4 options which can be used to replace the <MASK> token in the statement. Output the word from the correct option .
Let me give you an example: Statement: ಕೆಂಪು ಸಮುದ್ರವು ಹಿಂದೂ ಮಹಾಸಾಗರದ ಕಡಲಾಚೆಯ <MASK> ಮತ್ತು ಏಷ್ಯಾ ಖಂಡದ ನಡುವೆ ಇರುವ ... | ರಷ್ಯಾ | 8 | NIv2 | task948_wiki_cloze_kn_multiple_choice_question_answering | fs_opt | [
0.1677732765674591,
0.21673239767551422,
-0.5169717073440552,
-0.21566300094127655,
-0.24253812432289124,
0.014566587284207344,
-0.6799794435501099,
0.34530800580978394,
-0.5949777364730835,
0.4363666772842407,
0.10838712751865387,
-0.5627550482749939,
0.46613165736198425,
-0.3927235901355... |
Part 1. Definition
You are given a sentence in Hebrew. Your job is to translate the Hebrew sentence into Portugese.
Part 2. Example
אחרי הכל, הם החליטו, והם עשו,
Answer: Afinal de contas elas é que decidem, e decidiram mesmo.
Explanation: The Hebrew sentence is correctly translated into Portugese, because the meaning i... | Eu estou centrada na minha essência, e o meu eu está suspenso. | 7 | NIv2 | task1114_ted_translation_he_pt | fs_opt | [
0.2558891177177429,
0.7478198409080505,
0.3830541670322418,
-0.03886901214718819,
-0.18326276540756226,
-0.21589088439941406,
1.0446386337280273,
0.2686944901943207,
0.5389734506607056,
-0.4758867621421814,
-0.8331347703933716,
0.3454332947731018,
-0.8259609937667847,
-0.14166495203971863,... |
In this task, you are given a question and answer for that. The answer will fill in the blank or will complete the unfinished question. Your task is to provide an explanation based on the given question or fill-in-the-blank statement, and answer.
Example input: Question: The government of Canada is considering impleme... | Nations that have more industry produce more hazardous waste. | 3 | NIv2 | task223_quartz_explanation_generation | fs_opt | [
-0.38074204325675964,
0.3210655152797699,
-0.03490392118692398,
-0.9102954864501953,
-0.1141124963760376,
-0.3542380928993225,
0.5176899433135986,
0.7524294853210449,
-0.4359443485736847,
-0.05229046195745468,
-0.20143255591392517,
-0.4180375337600708,
0.04745432361960411,
0.25235962867736... |
Q: You are given a statement written in Kannada. Choose the most logical word from the given 4 options which can be used to replace the <MASK> token in the statement. Output the word from the correct option .
Statement: IMFನಲ್ಲಿ ಸದಸ್ಯರ ಮೀಸಲು ಪ್ರಮಾಣವು, ಅದರ ಚಂದಾದಾರತ್ವ, ಮತಗಳ ಗಣನೆ, IMFನಿಂದ ಸಾಲಪಡೆಯಬಲ್ಲ ಅವಕಾಶ, ಹಾಗೂ ವಿಶೇಷ ಹಿಂ... | ಕೆನಡಾ | 7 | NIv2 | task948_wiki_cloze_kn_multiple_choice_question_answering | zs_opt | [
0.8518600463867188,
0.2745143473148346,
-0.12698127329349518,
0.1063004583120346,
0.08778393268585205,
-0.20723405480384827,
0.8226367235183716,
0.6800979375839233,
-0.25176453590393066,
0.19465412199497223,
-0.5737589597702026,
0.08108082413673401,
-0.9654386043548584,
0.10480017215013504... |
Given the task definition and input, reply with output. In this task, you will be presented with a text and a pronoun. You should write an implausible answer to the question of what is the pronoun's reference. Even though there exist multiple wrong answers, we only need a single wrong answer. Position of the pronoun in... | Len Righi | 5 | NIv2 | task331_gap_incorrect_answer_generation | zs_opt | [
-0.796242356300354,
1.107586145401001,
0.024164481088519096,
-0.03783462941646576,
-0.09348222613334656,
-0.20126809179782867,
1.0358237028121948,
0.19149789214134216,
0.21222971379756927,
-0.5164811611175537,
-0.8783957958221436,
0.5454471111297607,
-0.28417110443115234,
-0.22811222076416... |
In this task, you are given two sentences in Persian separated with <sep>, and you have to determine whether the sentences are paraphrases or not. Classify your answers into "paraphrase" and "not-paraphrase".
Ex Input:
چرا پوست ما چرب یا مرطوب می شود؟ <sep> تحقیق درباره چرا پوست ما چرب می شود؟
Ex Output:
paraphrase
... | not-paraphrase
| 1 | NIv2 | task465_parsinlu_qqp_classification | fs_opt | [
-1.1002016067504883,
0.7728413343429565,
-0.06328052282333374,
-0.42530930042266846,
-0.7205729484558105,
-0.08309910446405411,
1.0010868310928345,
0.49458736181259155,
0.3827865719795227,
0.22259068489074707,
-0.8018620610237122,
-0.1957491785287857,
-0.4324946403503418,
-0.10206767916679... |
You will be given a definition of a task first, then some input of the task.
You are given a sentence in Arabic. Your job is to translate the Arabic sentence into English.
أو يقومون بإحراق البلاستيك للوصول إلى المعادن في منازل حرق كما ترون هنا
Output: | Or they burn the plastics to get to the metals in burn houses like you see here. | 1 | NIv2 | task1230_ted_translation_ar_en | zs_opt | [
-0.1913742572069168,
0.8017199635505676,
-0.09323370456695557,
-0.47738099098205566,
-0.5139994621276855,
0.4245700240135193,
0.8860379457473755,
-0.02271563932299614,
0.5446145534515381,
-0.1395488679409027,
-0.5663977861404419,
0.5169286131858826,
-0.24762442708015442,
0.0341790020465850... |
Detailed Instructions: In this task, you're given the title of a five-sentence story, the first four sentences, and two options for the fifth sentence as a and b. Your job is to pick the sentence option that does not connect with the rest of the story, indicating your choice as 'a' or 'b'. If both sentences are plausib... | b | 8 | NIv2 | task214_rocstories_incorrect_ending_classification | zs_opt | [
0.0506485179066658,
0.48519423604011536,
0.15493150055408478,
-0.16915611922740936,
-0.23835331201553345,
-0.08787517994642258,
0.4555366039276123,
1.021140694618225,
0.17523078620433807,
0.27985504269599915,
-0.750442624092102,
-0.7378694415092468,
-0.09901735186576843,
-0.051754705607891... |
You are given a statement in Croatian, a question word and four choices in Croation. If the question word is "cause", you should choose the option that is most likely to be the cause of the statement. If the question word is "effect", you should pick the choice that is most likely to be a consequence of the statement. ... | Izašla je iz reda.
| 5 | NIv2 | task1628_copa_hr_question_answering | fs_opt | [
0.8272838592529297,
0.6170474886894226,
-0.259623259305954,
0.09372229129076004,
-0.019171271473169327,
-1.3987047672271729,
0.805732786655426,
0.9604240655899048,
-0.5390172600746155,
-0.30678242444992065,
0.1245613545179367,
0.6244313716888428,
-0.4031484127044678,
-0.44555994868278503,
... |
In this task you will be given a list of integers. You should remove any integer that is not prime. A prime integer is an integer that is only divisible by '1' and itself. The output should be the list of prime numbers in the input list. If there are no primes in the input list an empty list ("[]") should be returned.
... | Solution: [563, 307, 769, 449, 401, 673, 499] | 5 | NIv2 | task366_synthetic_return_primes | fs_opt | [
-0.47507452964782715,
0.17178970575332642,
-0.3888348937034607,
-0.38354361057281494,
0.18223154544830322,
-0.24334120750427246,
1.0621867179870605,
0.6085165739059448,
-0.14708736538887024,
0.3894997239112854,
-1.1419180631637573,
-0.642937958240509,
-0.2775406539440155,
-0.37545484304428... |
Q: In this task, you are given a set of context paragraph and some supporting facts to answer a question. Your task is to generate answer for given question based on set of context paragraphs and supporting facts.
Context_1 : Magnus Tideman (born 9 April 1963), is a former professional tennis player from Sweden. He enj... | Manuel Orantes | 7 | NIv2 | task170_hotpotqa_answer_generation | zs_opt | [
0.20985248684883118,
0.503041684627533,
-1.2107778787612915,
0.17559808492660522,
0.9241883754730225,
0.06237971782684326,
0.7403704524040222,
0.7112278938293457,
-0.10453743487596512,
0.4021297097206116,
-0.20595312118530273,
1.1627784967422485,
-0.5802471041679382,
0.41611605882644653,
... |
instruction:
Given an entity, a before event, an after event, and an attribute related to the entity, generate a sentence as output. Your sentence should show the changes in the attribute of the entity.
question:
entity: drink
before: ungarnished
after: garnished
attr: taste
answer:
taste of drink was ungarnished be... | step of the download was missing for the internet channel before and installed for the internet channel afterwards
| 9 | NIv2 | task1631_openpi_answer_generation | fs_opt | [
-0.42446279525756836,
0.7195297479629517,
0.42313867807388306,
-0.08019042015075684,
-0.6355662941932678,
-0.9019905924797058,
-0.10637757182121277,
0.5713594555854797,
-0.3396005630493164,
-0.4500521123409271,
0.044653668999671936,
-0.37151557207107544,
-1.166460633277893,
-0.330657780170... |
Classify the given comment into 'Hope Speech' or 'Not Hope Speech'. A hope speech is defined as content that is encouraging, positive or supportive contents and talks about equality, diversity or inclusion
Ex Input:
“god gave us a choice. my choice is to love” wait i’m literally crying i love him so much
Ex Output:
H... | Hope Speech
| 1 | NIv2 | task679_hope_edi_english_text_classification | fs_opt | [
0.08752399682998657,
-0.10383739322423935,
0.2082114815711975,
0.39368852972984314,
-0.4964061379432678,
-1.2271673679351807,
0.2074781060218811,
0.719559907913208,
0.040773965418338776,
0.43946021795272827,
-0.16060343384742737,
-0.1220368891954422,
-0.10306796431541443,
-0.14106011390686... |
Detailed Instructions: Given an Amazon review, indicate whether it is a 'Positive Review' or 'Negative Review'.
Problem:Wish I had this along time ago. Comes with 2 programs. Installs very easy. Its like having your own cloud with out the risk of someone getting your information & making all those payments. Back ups ca... | Positive Review | 8 | NIv2 | task1343_amazon_us_reviews_rating | zs_opt | [
-0.2122645080089569,
0.2056088149547577,
0.15921352803707123,
-0.702741265296936,
0.03635788708925247,
-0.0925382748246193,
0.8308558464050293,
0.1282581388950348,
0.36448293924331665,
0.3059222102165222,
0.20695748925209045,
-0.7667979598045349,
-0.49279963970184326,
0.15944957733154297,
... |
Teacher:In this task, you will be shown an extract from a movie plot. You need to read the extract and create questions that can be answered from the extract. Once you create a question you also need to provide the correct answer for it. Write your answer like "Question: your question, Answer: your answer". The questio... | Question: What happens to some of the castaways?, Answer: They die | 6 | NIv2 | task182_duorc_question_generation | zs_opt | [
0.7440078854560852,
0.10823614150285721,
-0.2701059579849243,
-0.6627163290977478,
-0.4511512517929077,
-0.8948228359222412,
0.8082073926925659,
0.4300386905670166,
-0.3806763291358948,
-0.08513612300157547,
0.06718416512012482,
-0.37604695558547974,
-1.1414227485656738,
0.2324443459510803... |
Instructions: In this task, you're given a statement, and three sentences as choices. Your job is to determine which sentence clearly disagrees with the statement. Indicate your answer as '1', '2', or '3' corresponding to the choice number of the selected sentence.
Input: Statement: But skeptics suggest a darker that ... | 2 | 3 | NIv2 | task202_mnli_contradiction_classification | zs_opt | [
-0.47374236583709717,
-0.06740251183509827,
0.059723805636167526,
-0.4070274233818054,
-0.006169954314827919,
-0.5799590349197388,
0.47816041111946106,
0.5450254678726196,
0.08096557855606079,
-0.08502434939146042,
-0.08479864150285721,
-0.0561966747045517,
-0.3986426591873169,
-0.64936190... |
Given the task definition, example input & output, solve the new input case.
In this task, you're given a text and question. The question is about the topic of the document. You are expected to generate an incorrect answer. The incorrect answer should be among one of the following categories: 1)Company, 2)Educational I... | Written work | 1 | NIv2 | task631_dbpedia_14_incorrect_answer_generation | fs_opt | [
-0.5830345153808594,
0.010228323750197887,
0.24587740004062653,
-0.5461642742156982,
0.9421853423118591,
-0.7225300073623657,
-0.8321143388748169,
0.7070088386535645,
0.22912392020225525,
-0.33187299966812134,
-0.23938895761966705,
-0.3394853174686432,
-0.38453081250190735,
-0.293160796165... |
You are given a sentence in Polish. Your job is to translate the Polish sentence into Arabic.
Ex Input:
Ale właśnie wtedy firma przechodzi przez ciemny okres do etapu demokratycznego.
Ex Output:
ولكن بعد ذلك تمر الشركة خلال هذه المرحلة القاتمة لهذا الحزب الديمقراطي.
Ex Input:
To samo dotyczy świadomości.
Ex Output... | وهنا صحيفة صدرت صباح اليوم ، ولدينا بعض أخبار ويمبلدون ، وهذا جيد.
| 1 | NIv2 | task1259_ted_translation_pl_ar | fs_opt | [
-0.9817095994949341,
-0.14698728919029236,
-0.2882927656173706,
-0.6689993143081665,
-1.0874576568603516,
-0.06436443328857422,
1.5289337635040283,
0.01622636243700981,
0.8677916526794434,
-0.5312406420707703,
-0.2686103582382202,
0.21981719136238098,
-0.28651857376098633,
0.17522539198398... |
Detailed Instructions: In this task, you need to count the number of words in a sentence that contain the given letter
Q: Sentence: 'a stack of books with a clock on top of them'. How many words contain the letter 'h' in the sentence.
A: | 2 | 9 | NIv2 | task161_count_words_containing_letter | zs_opt | [
0.34584492444992065,
0.514787495136261,
-0.006124386563897133,
-0.5549102425575256,
0.10840961337089539,
-0.09165018051862717,
0.34085357189178467,
-0.4477006793022156,
0.15016299486160278,
-1.2555298805236816,
-0.37269333004951477,
-0.6450681090354919,
-0.42253339290618896,
0.359350919723... |
Detailed Instructions: You are given a conversation between two people. 'Person1:' and 'Person2:' are used to separate their respective dialogues. You are required to assign a label 'formal' if there is an absence of emotion and a presence of questions anywhere within the conversation. If such pattern is not found assi... | informal | 9 | NIv2 | task1533_daily_dialog_formal_classification | zs_opt | [
-0.42891383171081543,
0.579512357711792,
0.3872032165527344,
-0.4400428831577301,
-0.17459598183631897,
-0.5339593887329102,
1.0345003604888916,
-0.06271837651729584,
0.4157671630382538,
0.04770730808377266,
-0.330559641122818,
-0.410125195980072,
-0.15596839785575867,
0.28422296047210693,... |
Detailed Instructions: Given a review text from amazon and its polarity (either positive or negative). Generate answer "True" if given sentence and its polarity match, otherwise generate answer "False". A review is considered positive if the reviewer is satisfied with the product. Otherwise, it is considered negative.
... | False | 4 | NIv2 | task494_review_polarity_answer_generation | fs_opt | [
-0.2900112569332123,
-0.7641597390174866,
-0.01991838589310646,
-0.02172144502401352,
0.10561272501945496,
-0.890038013458252,
1.1915748119354248,
-0.16129502654075623,
0.1878998577594757,
-0.27382051944732666,
-0.11972180753946304,
0.3838699162006378,
-0.8604415655136108,
-0.6136275529861... |
TASK DEFINITION: In this task, you are given a summary for US Congressional and California state bill, your task is to generate a Title for this bill. The preferred titles are under forty words and mention the purpose of the bill.
PROBLEM: Directs the Secretary of Health and Human Services to: (1) establish the Child ... | Federal Resource Efficient Building Materials Act of 1993
| 8 | NIv2 | task1659_title_generation | fs_opt | [
0.31938618421554565,
0.7492862939834595,
-0.8922796249389648,
-0.164922297000885,
0.6818488240242004,
-0.09841227531433105,
0.8424636721611023,
0.7992259860038757,
-0.7610121369361877,
0.4488618075847626,
-0.12646269798278809,
0.5445953607559204,
-0.4236353039741516,
0.09513431787490845,
... |
Detailed Instructions: In this task, you're given two sentences. Indicate if the first sentence clearly entails the second sentence (i.e., one can conclude the 2nd sentence by reading the 1st one). Indicate your answer with '1' if the first sentence entails the second sentence, otherwise answer with '0'.
Problem:Senten... | 0 | 8 | NIv2 | task1344_glue_entailment_classification | zs_opt | [
-0.9920933246612549,
0.37371212244033813,
0.010447089560329914,
-0.3846282958984375,
-0.020742453634738922,
0.06626978516578674,
0.2094675898551941,
0.49335646629333496,
0.5065991878509521,
0.7623468637466431,
-0.49291521310806274,
-0.4210868179798126,
-0.323932945728302,
-0.63921922445297... |
In this task you will be given some text dialogue and you need to infer the underlying emotion of the text. The possible emotions are happy, sad, angry, or other.
Example: i don't work i could take your shift i am a student
Example solution: other
Example explanation: In this dialogue someone is informing another pers... | Solution: sad | 5 | NIv2 | task517_emo_classify_emotion_of_dialogue | fs_opt | [
-1.0912059545516968,
0.46615922451019287,
0.6063237190246582,
-1.0503137111663818,
0.18208636343479156,
-0.5970829725265503,
0.7569952011108398,
0.5435917973518372,
0.7869125008583069,
-0.11427166312932968,
-0.021684860810637474,
-0.6023781299591064,
-0.4948650300502777,
-0.259917736053466... |
Given a premise, an initial context, an original ending, and a counterfactual context, the task is to generate a new story ending aligned with the counterfactual context and as close to the original ending as possible. Each instance consists of a five-sentence story. The premise is the first sentence of a story, and th... | She filled her cart, thinking she would choose a favorite later. Then, she realized that she had checked out with all of the items. Most of the items she didn't even like!
| 5 | NIv2 | task269_csrg_counterfactual_story_generation | fs_opt | [
0.6839343309402466,
0.3771721124649048,
-0.012850400060415268,
-0.00684606796130538,
0.46220967173576355,
-1.1245038509368896,
0.5199998617172241,
1.104992151260376,
-0.6114416122436523,
-0.013500083237886429,
-0.6716808080673218,
-0.015236305072903633,
-0.19652487337589264,
0.098017595708... |
Definition: Generate an explanation for the given claim using the provided supporting material from the paragraph. Please consider the following points while generating an output. 1) The claim will always have supporting proof in the paragraph, and the paragraph will have a clear point of view supporting the claim. 2) ... | Maryland’s attorney general has announced three people have been indicted for allegedly operating unlicensed assisted living facilities in the Baltimore area. | 2 | NIv2 | task1369_healthfact_sentence_generation | zs_opt | [
0.510344922542572,
0.23473623394966125,
-0.2826961874961853,
0.3062332272529602,
0.1167217418551445,
-0.7948091626167297,
-0.44433265924453735,
1.2581274509429932,
-0.22555652260780334,
0.677420973777771,
-0.2831800580024719,
0.30548766255378723,
-0.06317378580570221,
-0.33380326628685,
... |
Given the task definition and input, reply with output. In this task, you are given a sentence and a profession that is mentioned in the sentence. The gender of the person with the given profession can be identified using the gendered pronouns that refer to them in the text. You are expected to identify the gender. Sel... | Male | 5 | NIv2 | task340_winomt_classification_gender_pro | zs_opt | [
-1.4048278331756592,
0.874452531337738,
-0.36749038100242615,
-0.5259125232696533,
0.10250957310199738,
0.010898984968662262,
-0.3900071680545807,
0.2667842507362366,
1.2407467365264893,
0.27256453037261963,
-0.9349895715713501,
0.21712036430835724,
-0.4804437756538391,
-0.3760862648487091... |
Teacher:In this task you will be given a claim and a perspective. You should determine whether that perspective supports or undermines the claim. If the perspective could possibly convince someone with different view, it is supporting, otherwise it is undermining.
Teacher: Now, understand the problem? Solve this instan... | support | 6 | NIv2 | task738_perspectrum_classification | zs_opt | [
-0.889731228351593,
0.44607001543045044,
0.29672056436538696,
-0.7069902420043945,
-0.0940907821059227,
-0.5599938035011292,
1.0612989664077759,
1.0731425285339355,
0.7819675207138062,
0.12176910042762756,
-0.38141775131225586,
-0.30158546566963196,
-0.2347095012664795,
-0.2435261309146881... |
Part 1. Definition
You will be given a context, a subject and a relation. Your task is to generate a question based on the subject and relation. The generated question should include the given subject. Try to use a minimum number of words that are not present in either context, subject or relation while generating ques... | Which award did Dragonwings get? | 7 | NIv2 | task1325_qa_zre_question_generation_on_subject_relation | fs_opt | [
0.3946867883205414,
0.3490620255470276,
-0.4480375349521637,
-0.5995762944221497,
-0.3079095780849457,
-0.09884814918041229,
1.0762319564819336,
0.5604746341705322,
0.19474834203720093,
-0.17752346396446228,
-0.03110450878739357,
0.71104496717453,
-1.5214033126831055,
0.1286938488483429,
... |
You will be given a definition of a task first, then some input of the task.
In this task, you will be presented with a question in Dutch language, and you have to write the named entities from the question if present. B denotes the first item of a phrase and an I any non-initial word. Here is the list of terms used: p... | Pena: B-PER, Col: B-LOC | 1 | NIv2 | task1544_conll2002_named_entity_recognition_answer_generation | zs_opt | [
-0.40681958198547363,
0.2824179530143738,
0.14423039555549622,
0.009152858518064022,
-0.2670374810695648,
-0.09666115045547485,
1.0750126838684082,
0.03113635443150997,
0.42908909916877747,
-0.23182496428489685,
-0.684561014175415,
0.1842014193534851,
0.007746447343379259,
-0.2197883725166... |
In this task, you are given a sentence in the Hindi language and your task is to convert it into the English language. In translation, keep numbers as it is and make it sentence case (capitalize only the first word of each sentence and noun).
Example input: 2007 में फ़्रांस, पेरिस के पार्क डेस प्रिंसेस में हुए रग्बी व... | Minutes after the flow of hot mud was visible from Lake Coatepeque, a huge cloud of ash appeared, which reached an altitude of 50,000 feet, then covering the northwestern part of the country which contains the largest concentration of coffee farms (main export). | 3 | NIv2 | task433_alt_hi_en_translation | fs_opt | [
-0.096027672290802,
0.8088387250900269,
-0.15773048996925354,
0.1281396448612213,
0.17583568394184113,
-0.8004413843154907,
-0.9070528745651245,
0.9594066143035889,
0.003620031289756298,
-0.3364405333995819,
-0.14628472924232483,
-0.014701265841722488,
-0.1802486777305603,
0.36858552694320... |
You are given a sentence in Spanish. Your job is to translate the Spanish sentence into Italian.
Example input: Aquí esta la medusa ala-x de la muerte.
Example output: Qui c'è la medusa mortale con le ali ad X.
Example explanation: The Spanish sentence is correctly translated into Italian, because the meaning is prese... | Uno dei problemi quando si cerca di capire quanto sia sana la terra, è che non abbiamo dati esatti su com'era il tempo 60 anni fa, ed ancora meno dati sullo strato d'ozono. | 3 | NIv2 | task1101_ted_translation_es_it | fs_opt | [
-0.5360714197158813,
0.8271958827972412,
-0.5982657670974731,
-0.287483811378479,
-0.8009528517723083,
-1.1130152940750122,
0.27727746963500977,
1.082446575164795,
-0.027691196650266647,
-0.1623959243297577,
-0.28129416704177856,
0.38101813197135925,
-0.24494409561157227,
0.881733715534210... |
Detailed Instructions: Given a sentence in the Japanese and Central Khmer language. Your task is check if the Khamer sentence is translation of Japanese. if the translation is correct than generate label "Yes", otherwise generate label "No".
Q: Japanese: ベーティは既婚者に対し、彼らのパートナーについての知識をテストし、どのカップルが互いを良く知っているかを調べるテレビ番組に500回... | Yes | 9 | NIv2 | task1123_alt_ja_khm_answer_generation | zs_opt | [
-0.04562224820256233,
-0.1879386007785797,
-0.24325300753116608,
0.12214392423629761,
0.5635826587677002,
-0.6994838118553162,
0.7208303809165955,
0.4377993941307068,
-0.24695245921611786,
-0.4470711946487427,
-0.6278030872344971,
0.3041096329689026,
-1.0350481271743774,
0.5934343934059143... |
In this task, you're given a pair of sentences, sentence 1 and sentence 2, that agree with each other. Your job is to alter sentence 2 so that the pair neither agree nor contradict each other. Generated sentences must be short, with less than 15 words. New information can be introduced. Avoid using pronouns to confuse ... | The man is going to a birthday party. | 4 | NIv2 | task184_snli_entailment_to_neutral_text_modification | zs_opt | [
-0.38193854689598083,
1.0344877243041992,
0.8532674908638,
-0.6646091938018799,
0.19950640201568604,
-0.17415523529052734,
-0.11181944608688354,
0.4650362730026245,
0.5138604044914246,
-0.2739039659500122,
-0.8995589017868042,
-0.38021761178970337,
-0.8993855714797974,
-0.34898191690444946... |
Teacher: Given a post that is a real-life anecdote of a complex ethical situation and an associated claim about its type, verify if the claim is true or not. The claim asks if the posts are historical or hypothetical. The posts are "HISTORICAL" when the author has already done something and they are "HYPOTHETICAL" when... | yes | 2 | NIv2 | task501_scruples_anecdotes_post_type_verification | fs_opt | [
-0.017069756984710693,
0.18343684077262878,
0.00699990801513195,
0.24828116595745087,
0.17927603423595428,
-0.2142685055732727,
0.43137356638908386,
1.3113452196121216,
-0.36984437704086304,
-0.40838974714279175,
-0.03626769408583641,
-0.14345508813858032,
-0.0005058166570961475,
-0.247656... |
Q: You are given a short poem which is written by a kid. Your task is to predict if the kid is in elementary school or high school. Anyone who is from grade 1st to 6th-grade will be considered as in elementary school, and 7th to 12th-grade kids will be considered as in high school. There are only two possible outputs, ... | elementary | 7 | NIv2 | task1712_poki_classification | zs_opt | [
0.08682163059711456,
0.9013949632644653,
0.6309442520141602,
-0.10681755840778351,
-0.20537368953227997,
-0.9382491707801819,
0.1927175223827362,
0.6569569706916809,
0.20375584065914154,
-0.8862200975418091,
-0.25398656725883484,
0.5001198053359985,
0.13499923050403595,
-0.0223970487713813... |
Q: In this task, you are given dvd product reviews in Japanese language. The goal is to classify the review as "POS" if the overall sentiment of the review is positive or as "NEG" if the overall sentiment of the review is negative.
本編は非常に期待している。でもこれは...。 . こちらは、デュナンのフィギアが付いているが、それは別に如何でもいい。それより、気になるのは、ライナーノートすら付いていない事だ... | NEG | 7 | NIv2 | task486_cls_japanese_dvd_classification | zs_opt | [
-0.5777980089187622,
-0.7219439148902893,
-0.02819114178419113,
-0.5082377195358276,
-0.11824776232242584,
-0.1843893676996231,
0.8426495790481567,
-0.08187654614448547,
0.19906210899353027,
0.4624518156051636,
-0.009221253916621208,
-0.6740036606788635,
0.35328128933906555,
-0.08184786140... |
Detailed Instructions: Given a text passage, come up with an appropriate title for it. The title should accurately represent what the main focus of the passage is. The title of a passage may be present in the first few sentences of the passage and/or be referred to multiple times in the passage. The title should be 1-5... | Mount Shasta | 9 | NIv2 | task769_qed_summarization | zs_opt | [
0.038767654448747635,
0.4355684518814087,
-0.2812076807022095,
0.18272997438907623,
-0.32781702280044556,
-0.935685396194458,
0.49102872610092163,
0.33886533975601196,
-0.0014191583031788468,
0.04011142998933792,
0.1537407636642456,
0.3209538459777832,
-0.8673768639564514,
0.17965191602706... |
You will be given a definition of a task first, then some input of the task.
In this task, you are given two sentences. Your task is to classify the given sentences as "Yes" if they have same meaning; otherwise, classify them as "No".
Sentence-1: I'd rather do something relaxing.<sep>Sentence-2: I would like to go to... | No | 1 | NIv2 | task566_circa_classification | zs_opt | [
-0.5681252479553223,
0.4764217138290405,
-0.24571602046489716,
0.29352977871894836,
0.25839561223983765,
-0.19888430833816528,
0.662316083908081,
0.45966196060180664,
0.5255485773086548,
0.14668889343738556,
0.24960783123970032,
-0.8020855188369751,
-0.8310378789901733,
-0.754361093044281,... |
Instructions: Given a review text from amazon and its polarity (either positive or negative). Generate answer "True" if given sentence and its polarity match, otherwise generate answer "False". A review is considered positive if the reviewer is satisfied with the product. Otherwise, it is considered negative.
Input: Re... | False | 3 | NIv2 | task494_review_polarity_answer_generation | zs_opt | [
-0.7546682357788086,
0.00046470650704577565,
0.6459747552871704,
-0.35323241353034973,
0.7880609631538391,
-0.72159343957901,
1.1684505939483643,
0.29307806491851807,
0.5602631568908691,
0.3827006220817566,
-0.4373849034309387,
-0.4072711169719696,
-0.6020014882087708,
-0.34278926253318787... |
Detailed Instructions: Given reviews from Amazon, classify those review based on their content into two classes: Negative or Positive.
Q: ...really quite staggeringly, breathtakingly bad - unless you're particularly fond of big explosions, that is, in which case you'll absolutely love it. If you don't, prepare for a to... | Negative | 9 | NIv2 | task493_review_polarity_classification | zs_opt | [
0.024775497615337372,
0.22304342687129974,
0.3841556906700134,
-0.40017080307006836,
0.3396903872489929,
-0.3946462571620941,
0.39956846833229065,
0.5223530530929565,
0.12150750309228897,
0.8736668825149536,
-0.051074057817459106,
-0.04721969738602638,
-0.8349798917770386,
-0.2855702340602... |
Definition: In this task, you're given a review from Amazon and category of the product based on the review given by the user and your task is classify whether the given category match the review. Generate "True" if given review and its category match, otherwise generate "False".
Input: Reviews: This product came in a ... | True | 2 | NIv2 | task1308_amazonreview_category_classification | zs_opt | [
-0.06980695575475693,
0.0669872984290123,
-0.12123171240091324,
-0.5346983671188354,
-0.2560197412967682,
0.6527101993560791,
0.4352293610572815,
0.5976847410202026,
0.05343691259622574,
0.3688896596431732,
-0.7616876363754272,
0.34494298696517944,
-0.7588688731193542,
0.056930359452962875... |
You will be given a definition of a task first, then some input of the task.
In this task, you are given a sentence in either Spanish or English. Your task is to determine the language of the input sentence. Input sentences can only be in Spanish or English, and they cannot be in two languages at the same time.
En c... | Spanish | 1 | NIv2 | task533_europarl_es-en_language_identification | zs_opt | [
-0.1908409595489502,
0.6925686597824097,
0.45453810691833496,
-0.2546861171722412,
0.4189455509185791,
-0.288337379693985,
0.4694545269012451,
0.6536086797714233,
0.02158396691083908,
-0.29664984345436096,
0.11939571797847748,
0.07403214275836945,
0.2704128623008728,
-0.28779923915863037,
... |
Teacher:Given a passage, construct a question on the basis of the information present in the passage. Construct the question in such a way that (i) it is unambiguous, (ii) it is answerable from the passage, (iii) its answer is unique (iv) it answer uses text phrases from the passage. Avoid creating questions that (i) c... | Does reference range for induced sputum eosinophil counts in Korean adult population? | 6 | NIv2 | task847_pubmedqa_question_generation | zs_opt | [
0.50118088722229,
0.3020339012145996,
-0.427486389875412,
-0.655427098274231,
0.16450057923793793,
-0.5163005590438843,
0.43029820919036865,
0.3556740880012512,
0.2826302945613861,
0.18072409927845,
-0.06629937887191772,
0.7762894630432129,
-0.5035128593444824,
-0.0065144565887749195,
0.... |
In this task, you will be given a set of steps that are required to answer a specific question. Your job is to generate that question. Each given step refers to either an entity (known or unknown), a propery of an entity or a query operation (count, group, union, etc.) Knowing those operations and how they appear in th... | If one dog photo is taken outside in a grassy area, while the other is taken inside in a private home setting.
****
| 4 | NIv2 | task184_break_generate_question | fs_opt | [
0.6637201309204102,
-0.11649814993143082,
-0.8446066379547119,
0.7699766159057617,
0.5097723007202148,
-0.5728235244750977,
1.0530200004577637,
0.5437079071998596,
-0.6581841111183167,
0.5388755798339844,
-0.5552747845649719,
0.40792298316955566,
0.056511081755161285,
0.31975823640823364,
... |
instruction:
Given a set of five words, generate the word from the set that does not belong (i.e. is the least relevant) with the other words. Words are separated by commas.
question:
cookies, replace, fudge, banned, lie
answer:
cookies
question:
hunting, fish, kayak, goat, raft
answer:
goat
question:
harsh, cheese... | cheese
| 9 | NIv2 | task142_odd-man-out_classification_no_category | fs_opt | [
0.2395343780517578,
0.5707640051841736,
0.28987517952919006,
-0.46967482566833496,
0.027589986100792885,
-0.3153347373008728,
-0.16952212154865265,
0.6387671232223511,
-0.15799911320209503,
-0.2586495876312256,
-0.5179309248924255,
-0.6188175082206726,
-0.7424267530441284,
-0.1250706911087... |
TASK DEFINITION: In this task, you are given a sentence in the Gujarati language and your task is to convert Gujarati sentence into the English language.
PROBLEM: બે ઘેટાં, એક કેમેરા તરફ જોઈ રહ્યા છે, જ્યારે અન્ય જુએ છે તે અરણ્યમાં છે
SOLUTION: Two sheep, one looking at the camera, while the other looks away are in th... | A blue and yellow Alaska railroad train driving through a grassy and woody area.
| 8 | NIv2 | task439_eng_guj_parallel_corpus_gu_en_translation | fs_opt | [
-0.9105151891708374,
0.2036648690700531,
0.17189106345176697,
-0.42190980911254883,
-0.10670127719640732,
-0.007733954582363367,
0.2230636477470398,
0.43047165870666504,
-0.048281148076057434,
-0.542777419090271,
-0.8222371339797974,
0.11783868074417114,
-0.2991308271884918,
0.128327697515... |
Instructions: In this task, you need to reverse the order of words in the given sentence.
Input: Sentence: two women sit on a boat with a stuffed panda
Output: | panda stuffed a with boat a on sit women two | 3 | NIv2 | task376_reverse_order_of_words | zs_opt | [
0.11518692970275879,
0.4099622368812561,
-0.7631891965866089,
-0.1348683089017868,
-0.13966944813728333,
-0.541662871837616,
0.46684786677360535,
0.20230184495449066,
0.22990281879901886,
0.18583422899246216,
-0.535464882850647,
-0.43579643964767456,
-0.23044806718826294,
-0.35751885175704... |
Definition: In this task, you will be given a set of steps that are required to answer a specific question. Your job is to generate that question. Each given step refers to either an entity (known or unknown), a propery of an entity or a query operation (count, group, union, etc.) Knowing those operations and how they ... | Is there a object the same color as the cylinder to the left of the large green sphere? | 2 | NIv2 | task184_break_generate_question | zs_opt | [
0.7613474726676941,
-0.1300843060016632,
-0.7560949325561523,
0.466827392578125,
0.3044534921646118,
-0.2861185669898987,
0.7648060321807861,
0.4469287395477295,
-0.2728431820869446,
0.6967788934707642,
-0.4079313576221466,
0.4369301199913025,
0.01991335302591324,
0.09287559986114502,
0.... |
Q: Generate the answer for a given question. The question requires common knowledge of English. Make sure that the answer is in detail, long, consistent with the context, and requires general knowledge of English. To be less repetitive, please vary your language for each question. Give a proper explanation to answer, a... | The city names you quote are all derived from Spanish, where 'San' (or 'Santo', as @tchrist clarifies below) is the male inflection and 'Santa' the female one. However, Santa Claus isn't derived from Spanish, but from Dutch, where it was originally rendered as Sante Klaas, and was modified to Santa when it was adopted ... | 7 | NIv2 | task225_english_language_answer_generation | zs_opt | [
-0.17714472115039825,
-0.040220506489276886,
-0.39932358264923096,
-0.4288991093635559,
0.3661254048347473,
-1.1824133396148682,
0.1962684541940689,
0.18752828240394592,
0.1846887320280075,
0.41727566719055176,
0.6621904373168945,
0.5856801867485046,
-1.046900749206543,
-0.3658174872398376... |
You are given a sentence in Spanish. Your job is to translate the Spanish sentence into Portugese.
Q: La razón por la que no estamos absolutamente seguros estriba en la naturaleza de la preservación de pruebas en nuestro sistema judicial, pero eso sería para toda una charla TEDx en otro momento. (Risas) Así que en el ... | Eu não sabia falar inglês, e estava atrasada em termos de ler e escrever.
****
| 4 | NIv2 | task1104_ted_translation_es_pt | fs_opt | [
-0.36270463466644287,
0.9102091193199158,
0.35661184787750244,
-0.4331313371658325,
-0.2478211522102356,
-1.131061315536499,
-0.38990581035614014,
1.2469216585159302,
0.5008090734481812,
-0.4090835750102997,
-0.00560466293245554,
0.5547162294387817,
-1.2255401611328125,
0.09282681345939636... |
The input is a tweet which can be Hate Speech or Offensive. Given such a tweet, output a phrase from the tweet that makes it hate speech or offensive. The output should contain only one such phrase. The output has to be from within the tweet itself. Do not generate words or phrases not present in the tweet.
Example in... | faggots | 3 | NIv2 | task1504_hatexplain_answer_generation | fs_opt | [
-0.8232287764549255,
0.7027309536933899,
0.5002581477165222,
0.7628888487815857,
-0.4756791293621063,
-1.3226381540298462,
-0.10182907432317734,
0.39060062170028687,
0.6355755925178528,
0.25878268480300903,
-0.35057201981544495,
0.28017422556877136,
-0.5243439674377441,
-0.9630041718482971... |
You are given a math word problem and you are supposed to apply a single mathematical operators like addition, subtraction, multiplication or division on the numbers embedded in the text to answer the following question and then only report final the numerical answer.
Ex Input:
Rose has 9 apples and 12 erasers . If sh... | 24
| 1 | NIv2 | task868_mawps_singleop_question_answering | fs_opt | [
0.17447331547737122,
0.7076525688171387,
-0.8924700021743774,
0.11022672057151794,
-0.5194705724716187,
0.04846634715795517,
0.38413727283477783,
0.522014856338501,
-0.692829966545105,
-0.37245145440101624,
-0.04712952300906181,
0.414072722196579,
-0.26461389660835266,
-0.15809160470962524... |
Given a sentence, fill out the missing word with a 'no' or a number (between zero and ten). You should write the numbers with english alphabet, like: four instead of 4.
Q: Cords begin to form between the ages of six and ____ months.
A: | nine | 4 | NIv2 | task1359_numer_sense_answer_generation | zs_opt | [
0.2132159322500229,
1.2179880142211914,
-0.501697301864624,
-0.5538680553436279,
-0.6542981863021851,
-0.13937753438949585,
0.09756551682949066,
0.9300966262817383,
-0.5340483784675598,
-0.2609190046787262,
-0.025301586836576462,
-0.4210997223854065,
-1.1683266162872314,
-0.161953821778297... |
Given the task definition and input, reply with output. In this task, you are given a review of product in Polish language and a question whether this review has positive sentiment. You are expected to generate the answer. The output should be "Yes" or "No". Don't generate anything apart from "Yes", "No".
Text: Wiem, ... | No | 5 | NIv2 | task635_allegro_reviews_answer_generation | zs_opt | [
-0.8239787817001343,
0.5857346057891846,
-0.4125974774360657,
0.07816453278064728,
-0.00725115742534399,
-0.3397764563560486,
1.2664200067520142,
-0.0076880864799022675,
-0.25464677810668945,
0.20641712844371796,
0.4089067578315735,
0.8270793557167053,
0.0754595398902893,
0.215231165289878... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.