Question-ID stringlengths 18 22 | RawQuestion stringlengths 18 933 | ProcessedQuestion stringlengths 17 932 | Parses dict | template_name stringclasses 5
values | template stringclasses 5
values | rendered_input stringlengths 50 1.11k | rendered_output stringlengths 1 58 |
|---|---|---|---|---|---|---|---|
FreebaseQA-train-200 | What city will host the 2020 Summer Olympics? | what city will host the 2020 summer olympics | {
"Parse-Id": [
"FreebaseQA-train-200.P0"
],
"PotentialTopicEntityMention": [
"2020 summer olympics"
],
"TopicEntityName": [
"2020 summer olympics"
],
"TopicEntityMid": [
"m.05gj2h"
],
"InferentialChain": [
"time.event.locations"
],
"Answers": [
{
"AnswersMid": [
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: What city will host the 2020 Summer Olympics? | Locations |
FreebaseQA-train-201 | Which city was awarded the 2020 Summer Olympics last month, beating Istanbul and Madrid ? | which city was awarded the 2020 summer olympics last month, beating istanbul and madrid | {
"Parse-Id": [
"FreebaseQA-train-201.P0"
],
"PotentialTopicEntityMention": [
"2020 summer olympics"
],
"TopicEntityName": [
"2020 summer olympics"
],
"TopicEntityMid": [
"m.05gj2h"
],
"InferentialChain": [
"time.event.locations"
],
"Answers": [
{
"AnswersMid": [
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which city was awarded the 2020 Summer Olympics last month, beating Istanbul and Madrid ? | Locations |
FreebaseQA-train-202 | What 1965 film was credited with saving 20th Century Fox from bankruptcy? | what 1965 film was credited with saving 20th century fox from bankruptcy | {
"Parse-Id": [
"FreebaseQA-train-202.P0",
"FreebaseQA-train-202.P1"
],
"PotentialTopicEntityMention": [
"20th century fox",
"20th century fox"
],
"TopicEntityName": [
"20th century fox",
"20th century fox"
],
"TopicEntityMid": [
"m.016tt2",
"m.016tt2"
],
"InferentialCh... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: What 1965 film was credited with saving 20th Century Fox from bankruptcy? | Film |
FreebaseQA-train-203 | 'Dookie' and '21st. Century Breakdown' are albums by which American rock band? | 'dookie' and '21st. century breakdown' are albums by which american rock band | {
"Parse-Id": [
"FreebaseQA-train-203.P0",
"FreebaseQA-train-203.P1",
"FreebaseQA-train-203.P2",
"FreebaseQA-train-203.P3",
"FreebaseQA-train-203.P4"
],
"PotentialTopicEntityMention": [
"21st. century breakdown",
"21st. century breakdown",
"21st. century breakdown",
"dookie",
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: 'Dookie' and '21st. Century Breakdown' are albums by which American rock band? | Producers |
FreebaseQA-train-204 | Who released an album in May 2009 called 21st Century Breakdown? | who released an album in may 2009 called 21st century breakdown | {
"Parse-Id": [
"FreebaseQA-train-204.P0",
"FreebaseQA-train-204.P1",
"FreebaseQA-train-204.P2",
"FreebaseQA-train-204.P3"
],
"PotentialTopicEntityMention": [
"21st century breakdown",
"21st century breakdown",
"21st century breakdown",
"21st century breakdown"
],
"TopicEntityN... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who released an album in May 2009 called 21st Century Breakdown? | Producers |
FreebaseQA-train-205 | Who stepped down as chief of 21st Century Fox in 2015? | who stepped down as chief of 21st century fox in 2015 | {
"Parse-Id": [
"FreebaseQA-train-205.P0"
],
"PotentialTopicEntityMention": [
"21st century fox"
],
"TopicEntityName": [
"21st century fox"
],
"TopicEntityMid": [
"m.0n_hqj6"
],
"InferentialChain": [
"organization.organization.board_members..organization.organization_board_membersh... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who stepped down as chief of 21st Century Fox in 2015? | Member |
FreebaseQA-train-206 | In which country is the spectacular 1650 stone-built 23-arch Khaju Bridge? | in which country is the spectacular 1650 stone-built 23-arch khaju bridge | {
"Parse-Id": [
"FreebaseQA-train-206.P0"
],
"PotentialTopicEntityMention": [
"23-arch khaju bridge"
],
"TopicEntityName": [
"khaju bridge"
],
"TopicEntityMid": [
"m.02877xv"
],
"InferentialChain": [
"location.location.containedby"
],
"Answers": [
{
"AnswersMid": [
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: In which country is the spectacular 1650 stone-built 23-arch Khaju Bridge? | Containedby |
FreebaseQA-train-207 | Who played 'Tony Wilson' in the 202 film '24 Hour Party People'? | who played 'tony wilson' in the 202 film '24 hour party people' | {
"Parse-Id": [
"FreebaseQA-train-207.P0"
],
"PotentialTopicEntityMention": [
"24 hour party people"
],
"TopicEntityName": [
"24 hour party people"
],
"TopicEntityMid": [
"m.01_bnz"
],
"InferentialChain": [
"film.film.starring..film.performance.actor"
],
"Answers": [
{
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who played 'Tony Wilson' in the 202 film '24 Hour Party People'? | Actor |
FreebaseQA-train-208 | Which singer who sang '24 Hours From Tulsa' died April 2006 at the age of 65? | which singer who sang '24 hours from tulsa' died april 2006 at the age of 65 | {
"Parse-Id": [
"FreebaseQA-train-208.P0"
],
"PotentialTopicEntityMention": [
"24 hours from tulsa"
],
"TopicEntityName": [
"twenty four hours from tulsa"
],
"TopicEntityMid": [
"m.011mzj6"
],
"InferentialChain": [
"music.recording.artist"
],
"Answers": [
{
"AnswersMi... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which singer who sang '24 Hours From Tulsa' died April 2006 at the age of 65? | Artist |
FreebaseQA-train-209 | What was the 2007 segue! to the film 28 Days Later? | what was the 2007 segue! to the film 28 days later | {
"Parse-Id": [
"FreebaseQA-train-209.P0"
],
"PotentialTopicEntityMention": [
"28 days later"
],
"TopicEntityName": [
"28 days later"
],
"TopicEntityMid": [
"m.012kyx"
],
"InferentialChain": [
"film.film.sequel"
],
"Answers": [
{
"AnswersMid": [
"m.0fhd04"
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: What was the 2007 segue! to the film 28 Days Later? | Sequel |
FreebaseQA-train-210 | '2 Become 1', 'Say You'll Be There' and 'Viva Forever' were all number one hit singles by which British group? | '2 become 1', 'say you'll be there' and 'viva forever' were all number one hit singles by which british group | {
"Parse-Id": [
"FreebaseQA-train-210.P0",
"FreebaseQA-train-210.P1",
"FreebaseQA-train-210.P2",
"FreebaseQA-train-210.P3",
"FreebaseQA-train-210.P4",
"FreebaseQA-train-210.P5",
"FreebaseQA-train-210.P6",
"FreebaseQA-train-210.P7",
"FreebaseQA-train-210.P8"
],
"PotentialTopicEn... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: '2 Become 1', 'Say You'll Be There' and 'Viva Forever' were all number one hit singles by which British group? | Artist |
FreebaseQA-train-211 | '2 Become 1', 'Say You'll Be There' and 'Viva Forever' were all number one hit singles by which British group? | '2 become 1', 'say you'll be there' and 'viva forever' were all number one hit singles by which british group | {
"Parse-Id": [
"FreebaseQA-train-211.P0",
"FreebaseQA-train-211.P1",
"FreebaseQA-train-211.P2",
"FreebaseQA-train-211.P3",
"FreebaseQA-train-211.P4",
"FreebaseQA-train-211.P5",
"FreebaseQA-train-211.P6",
"FreebaseQA-train-211.P7",
"FreebaseQA-train-211.P8",
"FreebaseQA-train-2... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: '2 Become 1', 'Say You'll Be There' and 'Viva Forever' were all number one hit singles by which British group? | Artist |
FreebaseQA-train-212 | Kenneth Moore portrayed which famous 2nd World War fighter pilot in the 1956 film Reach for the Sky? | kenneth moore portrayed which famous 2nd world war fighter pilot in the 1956 film reach for the sky | {
"Parse-Id": [
"FreebaseQA-train-212.P0",
"FreebaseQA-train-212.P1"
],
"PotentialTopicEntityMention": [
"2nd world war",
"reach for the sky"
],
"TopicEntityName": [
"world war ii",
"reach for the sky"
],
"TopicEntityMid": [
"m.081pw",
"m.04cgkz"
],
"InferentialChain": ... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Kenneth Moore portrayed which famous 2nd World War fighter pilot in the 1956 film Reach for the Sky? | Works |
FreebaseQA-train-213 | Which 'kind of love' did 2 Pac have a hit with in 1996? | which 'kind of love' did 2 pac have a hit with in 1996 | {
"Parse-Id": [
"FreebaseQA-train-213.P0"
],
"PotentialTopicEntityMention": [
"2 pac"
],
"TopicEntityName": [
"tupac shakur"
],
"TopicEntityMid": [
"m.07pzc"
],
"InferentialChain": [
"music.artist.track"
],
"Answers": [
{
"AnswersMid": [
"m.0112h_g"
],
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which 'kind of love' did 2 Pac have a hit with in 1996? | Track |
FreebaseQA-train-214 | At which narrow pass did Leonidas and 300 Spartans fight to the death, in order to slow down the Persian army? | at which narrow pass did leonidas and 300 spartans fight to the death, in order to slow down the persian army | {
"Parse-Id": [
"FreebaseQA-train-214.P0"
],
"PotentialTopicEntityMention": [
"300 spartans"
],
"TopicEntityName": [
"battle of thermopylae"
],
"TopicEntityMid": [
"m.02pyp8p"
],
"InferentialChain": [
"time.event.locations"
],
"Answers": [
{
"AnswersMid": [
"m... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: At which narrow pass did Leonidas and 300 Spartans fight to the death, in order to slow down the Persian army? | Locations |
FreebaseQA-train-215 | In 480 BC, Leonidas led his 300 Spartans against the Persian army, led by Xerxes, in which land battle? | in 480 bc, leonidas led his 300 spartans against the persian army, led by xerxes, in which land battle | {
"Parse-Id": [
"FreebaseQA-train-215.P0"
],
"PotentialTopicEntityMention": [
"300 spartans"
],
"TopicEntityName": [
"battle of thermopylae"
],
"TopicEntityMid": [
"m.0hzngz9"
],
"InferentialChain": [
"time.event.locations"
],
"Answers": [
{
"AnswersMid": [
"m... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: In 480 BC, Leonidas led his 300 Spartans against the Persian army, led by Xerxes, in which land battle? | Locations |
FreebaseQA-train-216 | Which famous actor is the lead singer with the rock group 30 odd foot of grunts? | which famous actor is the lead singer with the rock group 30 odd foot of grunts | {
"Parse-Id": [
"FreebaseQA-train-216.P0"
],
"PotentialTopicEntityMention": [
"30 odd foot of grunts"
],
"TopicEntityName": [
"30 odd foot of grunts"
],
"TopicEntityMid": [
"m.01mh7_2"
],
"InferentialChain": [
"music.musical_group.member..music.group_membership.member"
],
"Answ... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which famous actor is the lead singer with the rock group 30 odd foot of grunts? | Member |
FreebaseQA-train-217 | Who was the author of 'The 39 steps'? | who was the author of 'the 39 steps' | {
"Parse-Id": [
"FreebaseQA-train-217.P0",
"FreebaseQA-train-217.P1"
],
"PotentialTopicEntityMention": [
"39 steps",
"39 steps"
],
"TopicEntityName": [
"the 39 steps",
"the 39 steps"
],
"TopicEntityMid": [
"m.09xcn4",
"m.0llgk"
],
"InferentialChain": [
"film.film.st... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who was the author of 'The 39 steps'? | Story by |
FreebaseQA-train-218 | Who wrote the novel `The 39 Steps`? | who wrote the novel `the 39 steps` | {
"Parse-Id": [
"FreebaseQA-train-218.P0",
"FreebaseQA-train-218.P1"
],
"PotentialTopicEntityMention": [
"39 steps",
"39 steps"
],
"TopicEntityName": [
"the 39 steps",
"the 39 steps"
],
"TopicEntityMid": [
"m.09xcn4",
"m.0llgk"
],
"InferentialChain": [
"film.film.st... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who wrote the novel `The 39 Steps`? | Story by |
FreebaseQA-train-219 | Who wrote the novel The 39 Steps? | who wrote the novel the 39 steps | {
"Parse-Id": [
"FreebaseQA-train-219.P0"
],
"PotentialTopicEntityMention": [
"39 steps"
],
"TopicEntityName": [
"the 39 steps"
],
"TopicEntityMid": [
"m.0llgk"
],
"InferentialChain": [
"book.written_work.author"
],
"Answers": [
{
"AnswersMid": [
"m.0llf9"
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who wrote the novel The 39 Steps? | Author |
FreebaseQA-train-220 | Which British singer released a 2009 album entitled '3 Words'? | which british singer released a 2009 album entitled '3 words' | {
"Parse-Id": [
"FreebaseQA-train-220.P0",
"FreebaseQA-train-220.P1"
],
"PotentialTopicEntityMention": [
"3 words",
"3 words"
],
"TopicEntityName": [
"3 words",
"3 words"
],
"TopicEntityMid": [
"m.01123htn",
"m.07kfy9x"
],
"InferentialChain": [
"music.recording.arti... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which British singer released a 2009 album entitled '3 Words'? | Artist |
FreebaseQA-train-221 | Which US actor, who appeared in the film The 40 year old Virgin, also played the lead role in the US version of The Office? | which us actor, who appeared in the film the 40 year old virgin, also played the lead role in the us version of the office | {
"Parse-Id": [
"FreebaseQA-train-221.P0",
"FreebaseQA-train-221.P1",
"FreebaseQA-train-221.P2"
],
"PotentialTopicEntityMention": [
"40 year old virgin",
"us version of \"\"the office",
"us version of \"\"the office"
],
"TopicEntityName": [
"the 40-year-old virgin",
"the office... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which US actor, who appeared in the film The 40 year old Virgin, also played the lead role in the US version of The Office? | Actor |
FreebaseQA-train-222 | Which Emeritus Professor at Edinburgh University has written several series of novels including '2 Pillars Of Wisdom' and '44 Scotland Street'? | which emeritus professor at edinburgh university has written several series of novels including '2 pillars of wisdom' and '44 scotland street' | {
"Parse-Id": [
"FreebaseQA-train-222.P0",
"FreebaseQA-train-222.P1",
"FreebaseQA-train-222.P2"
],
"PotentialTopicEntityMention": [
"44 scotland street",
"44 scotland street",
"44 scotland street"
],
"TopicEntityName": [
"44 scotland street",
"44 scotland street",
"44 scotl... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which Emeritus Professor at Edinburgh University has written several series of novels including '2 Pillars Of Wisdom' and '44 Scotland Street'? | Author editor |
FreebaseQA-train-223 | What Rhodesian-born Scottish writer, who did a reading at SPL on Friday, is the creative forced behind the No. 1 Ladies' Detective Agency, 44 Scotland Street, Sunday Philosophy Club, and the Portuguese Irregular Verbs book series? | what rhodesian-born scottish writer, who did a reading at spl on friday, is the creative forced behind the no. 1 ladies' detective agency, 44 scotland street, sunday philosophy club, and the portuguese irregular verbs book series | {
"Parse-Id": [
"FreebaseQA-train-223.P0",
"FreebaseQA-train-223.P1",
"FreebaseQA-train-223.P2",
"FreebaseQA-train-223.P3",
"FreebaseQA-train-223.P4",
"FreebaseQA-train-223.P5",
"FreebaseQA-train-223.P6",
"FreebaseQA-train-223.P7"
],
"PotentialTopicEntityMention": [
"44 scotlan... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: What Rhodesian-born Scottish writer, who did a reading at SPL on Friday, is the creative forced behind the No. 1 Ladies' Detective Agency, 44 Scotland Street, Sunday Philosophy Club, and the Portuguese Irregular Verbs book series? | Author |
FreebaseQA-train-224 | What kind of shop did 50 Cent rap about in 2005? | what kind of shop did 50 cent rap about in 2005 | {
"Parse-Id": [
"FreebaseQA-train-224.P0"
],
"PotentialTopicEntityMention": [
"50 cent"
],
"TopicEntityName": [
"50 cent"
],
"TopicEntityMid": [
"m.01vvyc_"
],
"InferentialChain": [
"music.artist.track"
],
"Answers": [
{
"AnswersMid": [
"m.010qghs"
],
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: What kind of shop did 50 Cent rap about in 2005? | Track |
FreebaseQA-train-225 | Who hosted the '$64,000 Question' on British television? | who hosted the '$64,000 question' on british television | {
"Parse-Id": [
"FreebaseQA-train-225.P0",
"FreebaseQA-train-225.P1"
],
"PotentialTopicEntityMention": [
"64,000 question",
"64,000 question"
],
"TopicEntityName": [
"the $64,000 question",
"the $64,000 question"
],
"TopicEntityMid": [
"m.09k70gp",
"m.09k70gp"
],
"Infer... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who hosted the '$64,000 Question' on British television? | Actor |
FreebaseQA-train-226 | On television who live at 742 Evergreen Terrace, Springfield? | on television who live at 742 evergreen terrace, springfield | {
"Parse-Id": [
"FreebaseQA-train-226.P0",
"FreebaseQA-train-226.P1"
],
"PotentialTopicEntityMention": [
"742 evergreen terrace",
"springfield"
],
"TopicEntityName": [
"the simpsons house",
"springfield"
],
"TopicEntityMid": [
"m.08mftp",
"m.0m3n5"
],
"InferentialChain"... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: On television who live at 742 Evergreen Terrace, Springfield? | Universe |
FreebaseQA-train-227 | Which family live at 742 Evergreen Terrace? | which family live at 742 evergreen terrace | {
"Parse-Id": [
"FreebaseQA-train-227.P0"
],
"PotentialTopicEntityMention": [
"742 evergreen terrace"
],
"TopicEntityName": [
"the simpsons house"
],
"TopicEntityMid": [
"m.08mftp"
],
"InferentialChain": [
"fictional_universe.fictional_setting.universe"
],
"Answers": [
{
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which family live at 742 Evergreen Terrace? | Universe |
FreebaseQA-train-228 | Which fictional family live at 742 Evergreen Terrace, Springfield? | which fictional family live at 742 evergreen terrace, springfield | {
"Parse-Id": [
"FreebaseQA-train-228.P0",
"FreebaseQA-train-228.P1"
],
"PotentialTopicEntityMention": [
"742 evergreen terrace",
"springfield"
],
"TopicEntityName": [
"the simpsons house",
"springfield"
],
"TopicEntityMid": [
"m.08mftp",
"m.0m3n5"
],
"InferentialChain"... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which fictional family live at 742 Evergreen Terrace, Springfield? | Universe |
FreebaseQA-train-229 | Which TV family live at 742 Evergreen Terrace? | which tv family live at 742 evergreen terrace | {
"Parse-Id": [
"FreebaseQA-train-229.P0"
],
"PotentialTopicEntityMention": [
"742 evergreen terrace"
],
"TopicEntityName": [
"the simpsons house"
],
"TopicEntityMid": [
"m.08mftp"
],
"InferentialChain": [
"fictional_universe.fictional_setting.universe"
],
"Answers": [
{
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which TV family live at 742 Evergreen Terrace? | Universe |
FreebaseQA-train-230 | Which composers 7th Symphony is known as the Leningrad symphony? | which composers 7th symphony is known as the leningrad symphony | {
"Parse-Id": [
"FreebaseQA-train-230.P0"
],
"PotentialTopicEntityMention": [
"7th symphony"
],
"TopicEntityName": [
"symphony no. 7"
],
"TopicEntityMid": [
"m.049rr2j"
],
"InferentialChain": [
"music.album.artist"
],
"Answers": [
{
"AnswersMid": [
"m.02ck1"
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which composers 7th Symphony is known as the Leningrad symphony? | Artist |
FreebaseQA-train-231 | 84 Charing Cross Road is a book based on 20 years of correspondence between which New York writer and Frank Doel, an antiquarian bookseller in London? | 84 charing cross road is a book based on 20 years of correspondence between which new york writer and frank doel, an antiquarian bookseller in london | {
"Parse-Id": [
"FreebaseQA-train-231.P0"
],
"PotentialTopicEntityMention": [
"84 charing cross road"
],
"TopicEntityName": [
"84, charing cross road"
],
"TopicEntityMid": [
"m.03lzhc"
],
"InferentialChain": [
"book.written_work.author"
],
"Answers": [
{
"AnswersMid":... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: 84 Charing Cross Road is a book based on 20 years of correspondence between which New York writer and Frank Doel, an antiquarian bookseller in London? | Author |
FreebaseQA-train-232 | '8 Mile' was a 2002 film starring rapper Eminem, in which US city was it set? | '8 mile' was a 2002 film starring rapper eminem, in which us city was it set | {
"Parse-Id": [
"FreebaseQA-train-232.P0",
"FreebaseQA-train-232.P1"
],
"PotentialTopicEntityMention": [
"8 mile",
"eminem"
],
"TopicEntityName": [
"8 mile",
"eminem"
],
"TopicEntityMid": [
"m.017d93",
"m.01vsgrn"
],
"InferentialChain": [
"film.film.featured_film_lo... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: '8 Mile' was a 2002 film starring rapper Eminem, in which US city was it set? | Featured film locations |
FreebaseQA-train-233 | Who was the star of 8 Mile (2002) a hip-hop drama film which was a fictionalized account of his growing up in Detroit and early career in rap? | who was the star of 8 mile (2002) a hip-hop drama film which was a fictionalized account of his growing up in detroit and early career in rap | {
"Parse-Id": [
"FreebaseQA-train-233.P0"
],
"PotentialTopicEntityMention": [
"8 mile"
],
"TopicEntityName": [
"8 mile"
],
"TopicEntityMid": [
"m.017d93"
],
"InferentialChain": [
"film.film.starring..film.performance.actor"
],
"Answers": [
{
"AnswersMid": [
"m... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who was the star of 8 Mile (2002) a hip-hop drama film which was a fictionalized account of his growing up in Detroit and early career in rap? | Actor |
FreebaseQA-train-234 | Which stand-up comedian presents 8 Out Of 10 Cats and Distraction? | which stand-up comedian presents 8 out of 10 cats and distraction | {
"Parse-Id": [
"FreebaseQA-train-234.P0",
"FreebaseQA-train-234.P1"
],
"PotentialTopicEntityMention": [
"8 out of 10 cats",
"distraction"
],
"TopicEntityName": [
"8 out of 10 cats",
"distraction"
],
"TopicEntityMid": [
"m.06km6x",
"m.098n5g"
],
"InferentialChain": [
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which stand-up comedian presents 8 Out Of 10 Cats and Distraction? | Person |
FreebaseQA-train-235 | Who had a hit in 1984 with '99 Red Balloons'? | who had a hit in 1984 with '99 red balloons' | {
"Parse-Id": [
"FreebaseQA-train-235.P0",
"FreebaseQA-train-235.P1",
"FreebaseQA-train-235.P2",
"FreebaseQA-train-235.P3",
"FreebaseQA-train-235.P4"
],
"PotentialTopicEntityMention": [
"99 red balloons",
"99 red balloons",
"99 red balloons",
"99 red balloons",
"99 red ball... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who had a hit in 1984 with '99 Red Balloons'? | Artist |
FreebaseQA-train-236 | Who had a hit in 1984 with `99 Red Balloons`? | who had a hit in 1984 with `99 red balloons` | {
"Parse-Id": [
"FreebaseQA-train-236.P0",
"FreebaseQA-train-236.P1",
"FreebaseQA-train-236.P2",
"FreebaseQA-train-236.P3"
],
"PotentialTopicEntityMention": [
"99 red balloons",
"99 red balloons",
"99 red balloons",
"99 red balloons"
],
"TopicEntityName": [
"99 luftballons"... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who had a hit in 1984 with `99 Red Balloons`? | Lyricist |
FreebaseQA-train-237 | What name is given to the Fourth (4th) movement in Beethoven's / Ninth (9th) Symphony? | what name is given to the fourth (4th) movement in beethoven's / ninth (9th) symphony | {
"Parse-Id": [
"FreebaseQA-train-237.P0",
"FreebaseQA-train-237.P1"
],
"PotentialTopicEntityMention": [
"9th",
"beethoven"
],
"TopicEntityName": [
"symphony no. 9",
"ludwig van beethoven"
],
"TopicEntityMid": [
"m.01hrpy8",
"m.04k15"
],
"InferentialChain": [
"music... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: What name is given to the Fourth (4th) movement in Beethoven's / Ninth (9th) Symphony? | Compositions |
FreebaseQA-train-238 | Which singer/songwriter wrote the musical '9 to 5'? | which singer/songwriter wrote the musical '9 to 5' | {
"Parse-Id": [
"FreebaseQA-train-238.P0",
"FreebaseQA-train-238.P1",
"FreebaseQA-train-238.P2",
"FreebaseQA-train-238.P3",
"FreebaseQA-train-238.P4",
"FreebaseQA-train-238.P5"
],
"PotentialTopicEntityMention": [
"9 to 5",
"9 to 5",
"9 to 5",
"9 to 5",
"9 to 5",
"9 ... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which singer/songwriter wrote the musical '9 to 5'? | Composer |
FreebaseQA-train-239 | Who sang the theme song in 9 to 5? | who sang the theme song in 9 to 5 | {
"Parse-Id": [
"FreebaseQA-train-239.P0",
"FreebaseQA-train-239.P1",
"FreebaseQA-train-239.P2",
"FreebaseQA-train-239.P3"
],
"PotentialTopicEntityMention": [
"9 to 5",
"9 to 5",
"9 to 5",
"9 to 5"
],
"TopicEntityName": [
"9 to 5",
"9 to 5",
"9 to 5",
"9 to 5"
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who sang the theme song in 9 to 5? | Composer |
FreebaseQA-train-240 | Who had a UK top ten hit in July, 1980 with `9 to 5`? | who had a uk top ten hit in july, 1980 with `9 to 5` | {
"Parse-Id": [
"FreebaseQA-train-240.P0"
],
"PotentialTopicEntityMention": [
"9 to 5"
],
"TopicEntityName": [
"9 to 5"
],
"TopicEntityMid": [
"m.0_vb6g"
],
"InferentialChain": [
"music.recording.artist"
],
"Answers": [
{
"AnswersMid": [
"m.01vzxld"
],
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who had a UK top ten hit in July, 1980 with `9 to 5`? | Artist |
FreebaseQA-train-241 | A.A. Milne is most famous for creating which Bear? | a.a. milne is most famous for creating which bear | {
"Parse-Id": [
"FreebaseQA-train-241.P0",
"FreebaseQA-train-241.P1"
],
"PotentialTopicEntityMention": [
"a.a. milne",
"a.a. milne"
],
"TopicEntityName": [
"a. a. milne",
"a. a. milne"
],
"TopicEntityMid": [
"m.0k_s",
"m.0k_s"
],
"InferentialChain": [
"book.author.b... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: A.A. Milne is most famous for creating which Bear? | Book editions published |
FreebaseQA-train-242 | Odense and Aarhus are major cities in which European country? | odense and aarhus are major cities in which european country | {
"Parse-Id": [
"FreebaseQA-train-242.P0",
"FreebaseQA-train-242.P1",
"FreebaseQA-train-242.P2"
],
"PotentialTopicEntityMention": [
"aarhus",
"aarhus",
"odense"
],
"TopicEntityName": [
"aarhus",
"aarhus",
"odense"
],
"TopicEntityMid": [
"m.0lzp",
"m.0lzp",
"... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Odense and Aarhus are major cities in which European country? | Country |
FreebaseQA-train-243 | In the Bible who was the elder sister of Moses and Aaron? | in the bible who was the elder sister of moses and aaron | {
"Parse-Id": [
"FreebaseQA-train-243.P0",
"FreebaseQA-train-243.P1"
],
"PotentialTopicEntityMention": [
"aaron",
"moses"
],
"TopicEntityName": [
"aaron",
"moses"
],
"TopicEntityMid": [
"m.0llm",
"m.04yfh"
],
"InferentialChain": [
"people.person.sibling_s..people.si... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: In the Bible who was the elder sister of Moses and Aaron? | Sibling |
FreebaseQA-train-244 | In the Bible, who was the sister of Moses and Aaron? | in the bible, who was the sister of moses and aaron | {
"Parse-Id": [
"FreebaseQA-train-244.P0",
"FreebaseQA-train-244.P1"
],
"PotentialTopicEntityMention": [
"aaron",
"moses"
],
"TopicEntityName": [
"aaron",
"moses"
],
"TopicEntityMid": [
"m.0llm",
"m.04yfh"
],
"InferentialChain": [
"people.person.sibling_s..people.si... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: In the Bible, who was the sister of Moses and Aaron? | Sibling |
FreebaseQA-train-245 | In the Old Testament, who is the sister of Moses and Aaron? | in the old testament, who is the sister of moses and aaron | {
"Parse-Id": [
"FreebaseQA-train-245.P0",
"FreebaseQA-train-245.P1"
],
"PotentialTopicEntityMention": [
"aaron",
"moses"
],
"TopicEntityName": [
"aaron",
"moses"
],
"TopicEntityMid": [
"m.0llm",
"m.04yfh"
],
"InferentialChain": [
"people.person.sibling_s..people.si... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: In the Old Testament, who is the sister of Moses and Aaron? | Sibling |
FreebaseQA-train-246 | What is the name of Moses and Aaron's sister? | what is the name of moses and aaron's sister | {
"Parse-Id": [
"FreebaseQA-train-246.P0",
"FreebaseQA-train-246.P1"
],
"PotentialTopicEntityMention": [
"aaron",
"moses"
],
"TopicEntityName": [
"aaron",
"moses"
],
"TopicEntityMid": [
"m.0llm",
"m.04yfh"
],
"InferentialChain": [
"people.person.sibling_s..people.si... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: What is the name of Moses and Aaron's sister? | Sibling |
FreebaseQA-train-247 | Aaron Copeland wrote a ballet about which American folk hero? | aaron copeland wrote a ballet about which american folk hero | {
"Parse-Id": [
"FreebaseQA-train-247.P0"
],
"PotentialTopicEntityMention": [
"aaron copeland"
],
"TopicEntityName": [
"aaron copeland"
],
"TopicEntityMid": [
"m.0djg9"
],
"InferentialChain": [
"base.ballet.ballet_composer.ballets_composed"
],
"Answers": [
{
"AnswersM... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Aaron Copeland wrote a ballet about which American folk hero? | Ballets composed |
FreebaseQA-train-248 | Which ballet by Aaron Copeland, first performed in 1944, contains a mountain range and a season in its name. It tells the story of a celebration of the building of a Pennsylvanian farmhouse? | which ballet by aaron copeland, first performed in 1944, contains a mountain range and a season in its name. it tells the story of a celebration of the building of a pennsylvanian farmhouse | {
"Parse-Id": [
"FreebaseQA-train-248.P0"
],
"PotentialTopicEntityMention": [
"aaron copeland"
],
"TopicEntityName": [
"aaron copeland"
],
"TopicEntityMid": [
"m.0djg9"
],
"InferentialChain": [
"music.composer.compositions"
],
"Answers": [
{
"AnswersMid": [
"m... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which ballet by Aaron Copeland, first performed in 1944, contains a mountain range and a season in its name. It tells the story of a celebration of the building of a Pennsylvanian farmhouse? | Compositions |
FreebaseQA-train-249 | In October 2009 it was announced that director Sam Taylor-Wood was to marry Aaron Johnson, 23 years her junior, who were director and star respecitively of Nowhere Boy, a film about whom? | in october 2009 it was announced that director sam taylor-wood was to marry aaron johnson, 23 years her junior, who were director and star respecitively of nowhere boy, a film about whom | {
"Parse-Id": [
"FreebaseQA-train-249.P0",
"FreebaseQA-train-249.P1",
"FreebaseQA-train-249.P2"
],
"PotentialTopicEntityMention": [
"aaron johnson",
"nowhere boy",
"nowhere boy"
],
"TopicEntityName": [
"aaron taylor-johnson",
"nowhere boy",
"nowhere boy"
],
"TopicEntity... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: In October 2009 it was announced that director Sam Taylor-Wood was to marry Aaron Johnson, 23 years her junior, who were director and star respecitively of Nowhere Boy, a film about whom? | Character |
FreebaseQA-train-250 | Alex Band and Aaron Kamin make up which band? | alex band and aaron kamin make up which band | {
"Parse-Id": [
"FreebaseQA-train-250.P0",
"FreebaseQA-train-250.P1"
],
"PotentialTopicEntityMention": [
"aaron kamin",
"alex band"
],
"TopicEntityName": [
"aaron kamin",
"alex band"
],
"TopicEntityMid": [
"m.0768fk",
"m.01vcslm"
],
"InferentialChain": [
"music.grou... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Alex Band and Aaron Kamin make up which band? | Group |
FreebaseQA-train-251 | Played by Aaron Paul, who was Walter White's co-protagonist in 'Breaking Bad'? | played by aaron paul, who was walter white's co-protagonist in 'breaking bad' | {
"Parse-Id": [
"FreebaseQA-train-251.P0",
"FreebaseQA-train-251.P1",
"FreebaseQA-train-251.P2"
],
"PotentialTopicEntityMention": [
"aaron paul",
"aaron paul",
"breaking bad"
],
"TopicEntityName": [
"aaron paul",
"aaron paul",
"breaking bad"
],
"TopicEntityMid": [
"... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Played by Aaron Paul, who was Walter White's co-protagonist in 'Breaking Bad'? | Character |
FreebaseQA-train-252 | With which song did Swedish pop group Abba win the 1974 Eurovision Song Contest? | with which song did swedish pop group abba win the 1974 eurovision song contest | {
"Parse-Id": [
"FreebaseQA-train-252.P0",
"FreebaseQA-train-252.P1"
],
"PotentialTopicEntityMention": [
"abba",
"abba"
],
"TopicEntityName": [
"abba",
"abba"
],
"TopicEntityMid": [
"m.01hnmty",
"m.0kjh"
],
"InferentialChain": [
"music.release.track",
"music.art... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: With which song did Swedish pop group Abba win the 1974 Eurovision Song Contest? | Track |
FreebaseQA-train-253 | What is the title of Abba's last UK number one hit single? | what is the title of abba's last uk number one hit single | {
"Parse-Id": [
"FreebaseQA-train-253.P0",
"FreebaseQA-train-253.P1",
"FreebaseQA-train-253.P2"
],
"PotentialTopicEntityMention": [
"abba",
"abba",
"abba"
],
"TopicEntityName": [
"abba",
"abba",
"abba"
],
"TopicEntityMid": [
"m.038t528",
"m.0kjh",
"m.0vyvpd8... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: What is the title of Abba's last UK number one hit single? | Track |
FreebaseQA-train-254 | What is the title of the last UK number one hit for Swedish group Abba? | what is the title of the last uk number one hit for swedish group abba | {
"Parse-Id": [
"FreebaseQA-train-254.P0",
"FreebaseQA-train-254.P1",
"FreebaseQA-train-254.P2"
],
"PotentialTopicEntityMention": [
"abba",
"abba",
"abba"
],
"TopicEntityName": [
"abba",
"abba",
"abba"
],
"TopicEntityMid": [
"m.038t528",
"m.0kjh",
"m.0vyvpd8... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: What is the title of the last UK number one hit for Swedish group Abba? | Track |
FreebaseQA-train-255 | What was the only single by ABBA to reach number 1 on the US Billboard charts? | what was the only single by abba to reach number 1 on the us billboard charts | {
"Parse-Id": [
"FreebaseQA-train-255.P0",
"FreebaseQA-train-255.P1"
],
"PotentialTopicEntityMention": [
"abba",
"abba"
],
"TopicEntityName": [
"abba",
"abba"
],
"TopicEntityMid": [
"m.038t528",
"m.0kjh"
],
"InferentialChain": [
"music.release.track",
"music.com... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: What was the only single by ABBA to reach number 1 on the US Billboard charts? | Track |
FreebaseQA-train-256 | What was Abba`s first British hit single? | what was abba`s first british hit single | {
"Parse-Id": [
"FreebaseQA-train-256.P0",
"FreebaseQA-train-256.P1",
"FreebaseQA-train-256.P2"
],
"PotentialTopicEntityMention": [
"abba",
"abba",
"abba"
],
"TopicEntityName": [
"abba",
"abba",
"abba"
],
"TopicEntityMid": [
"m.038t528",
"m.0kjh",
"m.0vyvpd8... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: What was Abba`s first British hit single? | Track |
FreebaseQA-train-257 | No more carefree laughter, Silence ever after are the opening lines of which Abba Number One hit ? | no more carefree laughter, silence ever after are the opening lines of which abba number one hit | {
"Parse-Id": [
"FreebaseQA-train-257.P0",
"FreebaseQA-train-257.P1"
],
"PotentialTopicEntityMention": [
"abba",
"abba"
],
"TopicEntityName": [
"abba",
"abba"
],
"TopicEntityMid": [
"m.038t528",
"m.0kjh"
],
"InferentialChain": [
"music.release.track",
"music.art... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: No more carefree laughter, Silence ever after are the opening lines of which Abba Number One hit ? | Track |
FreebaseQA-train-258 | What was Abba's last Number One single in the UK? | what was abba's last number one single in the uk | {
"Parse-Id": [
"FreebaseQA-train-258.P0"
],
"PotentialTopicEntityMention": [
"abba"
],
"TopicEntityName": [
"abba"
],
"TopicEntityMid": [
"m.0kjh"
],
"InferentialChain": [
"music.artist.album"
],
"Answers": [
{
"AnswersMid": [
"m.01hnm5g"
],
"Answ... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: What was Abba's last Number One single in the UK? | Album |
FreebaseQA-train-259 | The group 'Abba' was formed in which European capital city in 1971? | the group 'abba' was formed in which european capital city in 1971 | {
"Parse-Id": [
"FreebaseQA-train-259.P0"
],
"PotentialTopicEntityMention": [
"abba"
],
"TopicEntityName": [
"abba"
],
"TopicEntityMid": [
"m.0kjh"
],
"InferentialChain": [
"music.artist.origin"
],
"Answers": [
{
"AnswersMid": [
"m.06mxs"
],
"Answe... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: The group 'Abba' was formed in which European capital city in 1971? | Origin |
FreebaseQA-train-260 | Where, in the novel The Count of Monte Christo, by Alexandre Dumas, does the Abbe Faria befriend Edmond Dantes during his time in prison? | where, in the novel the count of monte christo, by alexandre dumas, does the abbe faria befriend edmond dantes during his time in prison | {
"Parse-Id": [
"FreebaseQA-train-260.P0",
"FreebaseQA-train-260.P1",
"FreebaseQA-train-260.P2"
],
"PotentialTopicEntityMention": [
"abbe faria",
"count of monte christo",
"monte christo"
],
"TopicEntityName": [
"abbe faria",
"the count of monte cristo",
"monte cristo"
],... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Where, in the novel The Count of Monte Christo, by Alexandre Dumas, does the Abbe Faria befriend Edmond Dantes during his time in prison? | Prison |
FreebaseQA-train-261 | Abbotsford House, on the banks of the Tweed, was the home of which author? | abbotsford house, on the banks of the tweed, was the home of which author | {
"Parse-Id": [
"FreebaseQA-train-261.P0"
],
"PotentialTopicEntityMention": [
"abbotsford house"
],
"TopicEntityName": [
"abbotsford house"
],
"TopicEntityMid": [
"m.0pvy"
],
"InferentialChain": [
"architecture.structure.owner..architecture.ownership.owner"
],
"Answers": [
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Abbotsford House, on the banks of the Tweed, was the home of which author? | Owner |
FreebaseQA-train-262 | Which author lived at Abbotsford House near Galashiels in the Scottish borders? | which author lived at abbotsford house near galashiels in the scottish borders | {
"Parse-Id": [
"FreebaseQA-train-262.P0"
],
"PotentialTopicEntityMention": [
"abbotsford house"
],
"TopicEntityName": [
"abbotsford house"
],
"TopicEntityMid": [
"m.0pvy"
],
"InferentialChain": [
"architecture.structure.owner..architecture.ownership.owner"
],
"Answers": [
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which author lived at Abbotsford House near Galashiels in the Scottish borders? | Owner |
FreebaseQA-train-263 | Which British city is served by 'Abbotsinch Airport'? | which british city is served by 'abbotsinch airport' | {
"Parse-Id": [
"FreebaseQA-train-263.P0"
],
"PotentialTopicEntityMention": [
"abbotsinch airport"
],
"TopicEntityName": [
"glasgow airport"
],
"TopicEntityMid": [
"m.01pnl_"
],
"InferentialChain": [
"aviation.airport.serves"
],
"Answers": [
{
"AnswersMid": [
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which British city is served by 'Abbotsinch Airport'? | Serves |
FreebaseQA-train-264 | Which 1970's American TV detective series, was created by film writer Abby Mann? | which 1970's american tv detective series, was created by film writer abby mann | {
"Parse-Id": [
"FreebaseQA-train-264.P0",
"FreebaseQA-train-264.P1"
],
"PotentialTopicEntityMention": [
"abby mann",
"abby mann"
],
"TopicEntityName": [
"abby mann",
"abby mann"
],
"TopicEntityMid": [
"m.0bvjjq",
"m.0bvjjq"
],
"InferentialChain": [
"tv.tv_program_c... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which 1970's American TV detective series, was created by film writer Abby Mann? | Programs created |
FreebaseQA-train-265 | In music, Pulp, Def Leppard and ABC all come from which English city? | in music, pulp, def leppard and abc all come from which english city | {
"Parse-Id": [
"FreebaseQA-train-265.P0",
"FreebaseQA-train-265.P1",
"FreebaseQA-train-265.P2",
"FreebaseQA-train-265.P3",
"FreebaseQA-train-265.P4",
"FreebaseQA-train-265.P5"
],
"PotentialTopicEntityMention": [
"abc",
"abc",
"def leppard",
"def leppard",
"pulp",
"... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: In music, Pulp, Def Leppard and ABC all come from which English city? | Origin |
FreebaseQA-train-266 | ABC and NBC are TV networks in which country ? | abc and nbc are tv networks in which country | {
"Parse-Id": [
"FreebaseQA-train-266.P0",
"FreebaseQA-train-266.P1",
"FreebaseQA-train-266.P2"
],
"PotentialTopicEntityMention": [
"abc",
"abc",
"abc"
],
"TopicEntityName": [
"abc",
"abc",
"american broadcasting company"
],
"TopicEntityMid": [
"m.0bqs52l",
"m.0... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: ABC and NBC are TV networks in which country ? | Country |
FreebaseQA-train-267 | In which country were elections planned for 1991 were stopped by the military (leading to Civil War), were held in 1995, and, in 1999, after a series of short-term military leaders, Abdelaziz Bouteflika was elected president? | in which country were elections planned for 1991 were stopped by the military (leading to civil war), were held in 1995, and, in 1999, after a series of short-term military leaders, abdelaziz bouteflika was elected president | {
"Parse-Id": [
"FreebaseQA-train-267.P0"
],
"PotentialTopicEntityMention": [
"abdelaziz bouteflika"
],
"TopicEntityName": [
"abdelaziz bouteflika"
],
"TopicEntityMid": [
"m.0224kz"
],
"InferentialChain": [
"government.politician.government_positions_held..government.government_pos... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: In which country were elections planned for 1991 were stopped by the military (leading to Civil War), were held in 1995, and, in 1999, after a series of short-term military leaders, Abdelaziz Bouteflika was elected president? | Jurisdiction of office |
FreebaseQA-train-268 | What was Abdel Kechiche's 2008 film about France's North African community? | what was abdel kechiche's 2008 film about france's north african community | {
"Parse-Id": [
"FreebaseQA-train-268.P0",
"FreebaseQA-train-268.P1"
],
"PotentialTopicEntityMention": [
"abdel kechiche",
"abdel kechiche"
],
"TopicEntityName": [
"abdellatif kechiche",
"abdellatif kechiche"
],
"TopicEntityMid": [
"m.0gpvk3",
"m.0gpvk3"
],
"Inferential... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: What was Abdel Kechiche's 2008 film about France's North African community? | Film |
FreebaseQA-train-269 | Who was Minister-Chairman of the Russian Provisional Government and Prime Minister of Russia from 21 July 1917 (following the abdication of Tsar Nicholas II) to 8 November 1917 (the Bolshevik revolution), who is buried in Putney Vale Cemetery, London? | who was minister-chairman of the russian provisional government and prime minister of russia from 21 july 1917 (following the abdication of tsar nicholas ii) to 8 november 1917 (the bolshevik revolution), who is buried in putney vale cemetery, london | {
"Parse-Id": [
"FreebaseQA-train-269.P0"
],
"PotentialTopicEntityMention": [
"abdication of tsar"
],
"TopicEntityName": [
"russian revolution"
],
"TopicEntityMid": [
"m.06fby"
],
"InferentialChain": [
"base.culturalevent.event.entity_involved"
],
"Answers": [
{
"Answ... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who was Minister-Chairman of the Russian Provisional Government and Prime Minister of Russia from 21 July 1917 (following the abdication of Tsar Nicholas II) to 8 November 1917 (the Bolshevik revolution), who is buried in Putney Vale Cemetery... | Entity involved |
FreebaseQA-train-270 | Abdul Aziz Ibn Saud was the first king of which Asian state? | abdul aziz ibn saud was the first king of which asian state | {
"Parse-Id": [
"FreebaseQA-train-270.P0"
],
"PotentialTopicEntityMention": [
"abdul aziz ibn saud"
],
"TopicEntityName": [
"ibn saud"
],
"TopicEntityMid": [
"m.0195b9"
],
"InferentialChain": [
"people.person.nationality"
],
"Answers": [
{
"AnswersMid": [
"m.0... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Abdul Aziz Ibn Saud was the first king of which Asian state? | Nationality |
FreebaseQA-train-271 | What modern-day country was ruled between 1902 and 1964 by Abdul Aziz Ibn Saud and his son, Emir Saud? | what modern-day country was ruled between 1902 and 1964 by abdul aziz ibn saud and his son, emir saud | {
"Parse-Id": [
"FreebaseQA-train-271.P0",
"FreebaseQA-train-271.P1"
],
"PotentialTopicEntityMention": [
"abdul aziz ibn saud",
"saud"
],
"TopicEntityName": [
"ibn saud",
"house of saud"
],
"TopicEntityMid": [
"m.0195b9",
"m.0240lv"
],
"InferentialChain": [
"people.... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: What modern-day country was ruled between 1902 and 1964 by Abdul Aziz Ibn Saud and his son, Emir Saud? | Nationality |
FreebaseQA-train-272 | Abebe Bikila was the only athlete to win the Olympic marathon twice and in successive Olympic games. What country did he represent? | abebe bikila was the only athlete to win the olympic marathon twice and in successive olympic games. what country did he represent | {
"Parse-Id": [
"FreebaseQA-train-272.P0",
"FreebaseQA-train-272.P1",
"FreebaseQA-train-272.P2"
],
"PotentialTopicEntityMention": [
"abebe bikila",
"abebe bikila",
"abebe bikila"
],
"TopicEntityName": [
"abebe bikila",
"abebe bikila",
"abebe bikila"
],
"TopicEntityMid":... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Abebe Bikila was the only athlete to win the Olympic marathon twice and in successive Olympic games. What country did he represent? | Country |
FreebaseQA-train-273 | Which country does marathon man Abel Anton come form? | which country does marathon man abel anton come form | {
"Parse-Id": [
"FreebaseQA-train-273.P0",
"FreebaseQA-train-273.P1"
],
"PotentialTopicEntityMention": [
"abel anton",
"abel anton"
],
"TopicEntityName": [
"abel anton",
"abel anton"
],
"TopicEntityMid": [
"m.09l2n_",
"m.09l2n_"
],
"InferentialChain": [
"people.pers... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which country does marathon man Abel Anton come form? | Nationality |
FreebaseQA-train-274 | The Aberdare mountains are in which African country? | the aberdare mountains are in which african country | {
"Parse-Id": [
"FreebaseQA-train-274.P0"
],
"PotentialTopicEntityMention": [
"aberdare mountains"
],
"TopicEntityName": [
"aberdare range"
],
"TopicEntityMid": [
"m.06kdp6"
],
"InferentialChain": [
"location.location.containedby"
],
"Answers": [
{
"AnswersMid": [
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: The Aberdare mountains are in which African country? | Containedby |
FreebaseQA-train-275 | In which town or city is the University of Abertay? | in which town or city is the university of abertay | {
"Parse-Id": [
"FreebaseQA-train-275.P0"
],
"PotentialTopicEntityMention": [
"abertay"
],
"TopicEntityName": [
"abertay university"
],
"TopicEntityMid": [
"m.03p81c"
],
"InferentialChain": [
"location.location.containedby"
],
"Answers": [
{
"AnswersMid": [
"m... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: In which town or city is the University of Abertay? | Containedby |
FreebaseQA-train-276 | In 2010, Veera Musikapong, Abhisit Vejjajiva, Khattiya Sawasdiphol, Suthep Thaugsuban, Prompong Nopparit, Colonel Sansern Kaewkamnerd, Natthawut Saikua and Buranaj Smutharaks were prominent in the politics of which country? | in 2010, veera musikapong, abhisit vejjajiva, khattiya sawasdiphol, suthep thaugsuban, prompong nopparit, colonel sansern kaewkamnerd, natthawut saikua and buranaj smutharaks were prominent in the politics of which country | {
"Parse-Id": [
"FreebaseQA-train-276.P0"
],
"PotentialTopicEntityMention": [
"abhisit vejjajiva"
],
"TopicEntityName": [
"abhisit vejjajiva"
],
"TopicEntityMid": [
"m.087cw4"
],
"InferentialChain": [
"government.politician.government_positions_held..government.government_position_... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: In 2010, Veera Musikapong, Abhisit Vejjajiva, Khattiya Sawasdiphol, Suthep Thaugsuban, Prompong Nopparit, Colonel Sansern Kaewkamnerd, Natthawut Saikua and Buranaj Smutharaks were prominent in the politics of which country? | Jurisdiction of office |
FreebaseQA-train-277 | Abhorson is an executioner in which Shakespeare play? | abhorson is an executioner in which shakespeare play | {
"Parse-Id": [
"FreebaseQA-train-277.P0",
"FreebaseQA-train-277.P1"
],
"PotentialTopicEntityMention": [
"abhorson",
"abhorson"
],
"TopicEntityName": [
"abhorson",
"abhorson"
],
"TopicEntityMid": [
"m.0zzbjb6",
"m.0zzbjb6"
],
"InferentialChain": [
"film.film_charact... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Abhorson is an executioner in which Shakespeare play? | Film |
FreebaseQA-train-278 | Abidjan is the former capital and a major battleground of what civil-war-torn nation in 2011? | abidjan is the former capital and a major battleground of what civil-war-torn nation in 2011 | {
"Parse-Id": [
"FreebaseQA-train-278.P0"
],
"PotentialTopicEntityMention": [
"abidjan"
],
"TopicEntityName": [
"abidjan"
],
"TopicEntityMid": [
"m.0fmyd"
],
"InferentialChain": [
"base.biblioness.bibs_location.country"
],
"Answers": [
{
"AnswersMid": [
"m.0fv... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Abidjan is the former capital and a major battleground of what civil-war-torn nation in 2011? | Country |
FreebaseQA-train-279 | Who wrote the plays Nuts in May and Abigail's Party? | who wrote the plays nuts in may and abigail's party | {
"Parse-Id": [
"FreebaseQA-train-279.P0",
"FreebaseQA-train-279.P1",
"FreebaseQA-train-279.P2",
"FreebaseQA-train-279.P3"
],
"PotentialTopicEntityMention": [
"abigail 's party",
"nuts in may",
"nuts in may",
"nuts in may"
],
"TopicEntityName": [
"abigail's party",
"nut... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who wrote the plays Nuts in May and Abigail's Party? | Author |
FreebaseQA-train-280 | Abkhazia, recognised as independent by Russia, is a breakaway region from which former Soviet republic? | abkhazia, recognised as independent by russia, is a breakaway region from which former soviet republic | {
"Parse-Id": [
"FreebaseQA-train-280.P0",
"FreebaseQA-train-280.P1"
],
"PotentialTopicEntityMention": [
"abkhazia",
"abkhazia"
],
"TopicEntityName": [
"abkhazia",
"abkhazia"
],
"TopicEntityMid": [
"m.011jw",
"m.011jw"
],
"InferentialChain": [
"location.administrati... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Abkhazia, recognised as independent by Russia, is a breakaway region from which former Soviet republic? | Country |
FreebaseQA-train-281 | The Abominable Snowman, a mythological ape-like cryptid said to inhabit the Himalayan region of Nepal and Tibet, is also known as what? | the abominable snowman, a mythological ape-like cryptid said to inhabit the himalayan region of nepal and tibet, is also known as what | {
"Parse-Id": [
"FreebaseQA-train-281.P0",
"FreebaseQA-train-281.P1",
"FreebaseQA-train-281.P2",
"FreebaseQA-train-281.P3",
"FreebaseQA-train-281.P4"
],
"PotentialTopicEntityMention": [
"abominable snowman",
"himalayan region",
"nepal",
"tibet",
"tibet"
],
"TopicEntityN... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: The Abominable Snowman, a mythological ape-like cryptid said to inhabit the Himalayan region of Nepal and Tibet, is also known as what? | Cryptid classification |
FreebaseQA-train-282 | The Abominable Snowman is said to inhabit which mountain range? | the abominable snowman is said to inhabit which mountain range | {
"Parse-Id": [
"FreebaseQA-train-282.P0",
"FreebaseQA-train-282.P1"
],
"PotentialTopicEntityMention": [
"abominable snowman",
"abominable snowman"
],
"TopicEntityName": [
"yeti",
"yeti"
],
"TopicEntityMid": [
"m.0hfj1",
"m.0hfj1"
],
"InferentialChain": [
"base.myst... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: The Abominable Snowman is said to inhabit which mountain range? | Area of occurrence |
FreebaseQA-train-283 | With which mountain range would you associate the Abonimable Snowman? | with which mountain range would you associate the abonimable snowman | {
"Parse-Id": [
"FreebaseQA-train-283.P0",
"FreebaseQA-train-283.P1"
],
"PotentialTopicEntityMention": [
"abonimable snowman",
"abonimable snowman"
],
"TopicEntityName": [
"yeti",
"yeti"
],
"TopicEntityMid": [
"m.0hfj1",
"m.0hfj1"
],
"InferentialChain": [
"base.myst... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: With which mountain range would you associate the Abonimable Snowman? | Area of occurrence |
FreebaseQA-train-284 | What links the US, 2 May 2011 and Abottabad in Pakistan? | what links the us, 2 may 2011 and abottabad in pakistan | {
"Parse-Id": [
"FreebaseQA-train-284.P0"
],
"PotentialTopicEntityMention": [
"abottabad"
],
"TopicEntityName": [
"abbottabad"
],
"TopicEntityMid": [
"m.011hs1"
],
"InferentialChain": [
"location.location.events"
],
"Answers": [
{
"AnswersMid": [
"m.0bwk4lh"
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: What links the US, 2 May 2011 and Abottabad in Pakistan? | Events |
FreebaseQA-train-285 | By what name is the Battle of Aboukir Bay (1798) more commonly known? | by what name is the battle of aboukir bay (1798) more commonly known | {
"Parse-Id": [
"FreebaseQA-train-285.P0"
],
"PotentialTopicEntityMention": [
"aboukir bay"
],
"TopicEntityName": [
"abu qir bay"
],
"TopicEntityMid": [
"m.06v4lf"
],
"InferentialChain": [
"location.location.events"
],
"Answers": [
{
"AnswersMid": [
"m.01gq_"
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: By what name is the Battle of Aboukir Bay (1798) more commonly known? | Events |
FreebaseQA-train-286 | Who wrote the 1998 novel About A Boy? | who wrote the 1998 novel about a boy | {
"Parse-Id": [
"FreebaseQA-train-286.P0",
"FreebaseQA-train-286.P1"
],
"PotentialTopicEntityMention": [
"about a boy",
"about a boy"
],
"TopicEntityName": [
"about a boy",
"about a boy"
],
"TopicEntityMid": [
"m.04tvp02",
"m.0jv0d"
],
"InferentialChain": [
"book.bo... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who wrote the 1998 novel About A Boy? | Author editor |
FreebaseQA-train-287 | Who wrote the book `About A Boy` which was later turned into a successful film starring Hugh Grant? | who wrote the book `about a boy` which was later turned into a successful film starring hugh grant | {
"Parse-Id": [
"FreebaseQA-train-287.P0",
"FreebaseQA-train-287.P1",
"FreebaseQA-train-287.P2"
],
"PotentialTopicEntityMention": [
"about a boy",
"about a boy",
"about a boy"
],
"TopicEntityName": [
"about a boy",
"about a boy",
"about a boy"
],
"TopicEntityMid": [
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who wrote the book `About A Boy` which was later turned into a successful film starring Hugh Grant? | Author editor |
FreebaseQA-train-288 | Who wrote the novels About A Boy, How To Be Good and High Fidelity? | who wrote the novels about a boy, how to be good and high fidelity | {
"Parse-Id": [
"FreebaseQA-train-288.P0",
"FreebaseQA-train-288.P1",
"FreebaseQA-train-288.P2",
"FreebaseQA-train-288.P3",
"FreebaseQA-train-288.P4",
"FreebaseQA-train-288.P5",
"FreebaseQA-train-288.P6"
],
"PotentialTopicEntityMention": [
"about a boy",
"about a boy",
"abo... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who wrote the novels About A Boy, How To Be Good and High Fidelity? | Author editor |
FreebaseQA-train-289 | Who had parts in 'St. Elmo's Fire', 'About Last Night ' and 'Wayne's World'? | who had parts in 'st. elmo's fire', 'about last night ' and 'wayne's world' | {
"Parse-Id": [
"FreebaseQA-train-289.P0",
"FreebaseQA-train-289.P1"
],
"PotentialTopicEntityMention": [
"about last night",
"wayne?s world"
],
"TopicEntityName": [
"about last night",
"wayne's world"
],
"TopicEntityMid": [
"m.03rl2l",
"m.0kbwb"
],
"InferentialChain": [... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who had parts in 'St. Elmo's Fire', 'About Last Night ' and 'Wayne's World'? | Actor |
FreebaseQA-train-290 | Abracadabra (1982) and The Joker (1990) were big hits for which group? | abracadabra (1982) and the joker (1990) were big hits for which group | {
"Parse-Id": [
"FreebaseQA-train-290.P0",
"FreebaseQA-train-290.P1",
"FreebaseQA-train-290.P2"
],
"PotentialTopicEntityMention": [
"abracadabra",
"joker",
"joker"
],
"TopicEntityName": [
"abracadabra",
"the joker",
"the joker"
],
"TopicEntityMid": [
"m.0f794zy",
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Abracadabra (1982) and The Joker (1990) were big hits for which group? | Artist |
FreebaseQA-train-291 | In the Bible who was the son of Abraham and Hagar? | in the bible who was the son of abraham and hagar | {
"Parse-Id": [
"FreebaseQA-train-291.P0",
"FreebaseQA-train-291.P1"
],
"PotentialTopicEntityMention": [
"abraham",
"hagar"
],
"TopicEntityName": [
"abraham",
"hagar"
],
"TopicEntityMid": [
"m.0pwb",
"m.01sq9c"
],
"InferentialChain": [
"people.person.children",
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: In the Bible who was the son of Abraham and Hagar? | Children |
FreebaseQA-train-292 | In the Bible, who was the son of Abraham and Hagar? | in the bible, who was the son of abraham and hagar | {
"Parse-Id": [
"FreebaseQA-train-292.P0",
"FreebaseQA-train-292.P1"
],
"PotentialTopicEntityMention": [
"abraham",
"hagar"
],
"TopicEntityName": [
"abraham",
"hagar"
],
"TopicEntityMid": [
"m.0pwb",
"m.01sq9c"
],
"InferentialChain": [
"people.person.children",
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: In the Bible, who was the son of Abraham and Hagar? | Children |
FreebaseQA-train-293 | In the Bible who was the son of Abraham and Sarah? | in the bible who was the son of abraham and sarah | {
"Parse-Id": [
"FreebaseQA-train-293.P0",
"FreebaseQA-train-293.P1"
],
"PotentialTopicEntityMention": [
"abraham",
"sarah"
],
"TopicEntityName": [
"abraham",
"sarah"
],
"TopicEntityMid": [
"m.0pwb",
"m.01b5wx"
],
"InferentialChain": [
"people.person.children",
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: In the Bible who was the son of Abraham and Sarah? | Children |
FreebaseQA-train-294 | In the Bible who was the father of Abraham? | in the bible who was the father of abraham | {
"Parse-Id": [
"FreebaseQA-train-294.P0"
],
"PotentialTopicEntityMention": [
"abraham"
],
"TopicEntityName": [
"abraham"
],
"TopicEntityMid": [
"m.0pwb"
],
"InferentialChain": [
"people.person.parents"
],
"Answers": [
{
"AnswersMid": [
"m.0j0hf"
],
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: In the Bible who was the father of Abraham? | Parents |
FreebaseQA-train-295 | In the Bible, who was the wife of Abraham? | in the bible, who was the wife of abraham | {
"Parse-Id": [
"FreebaseQA-train-295.P0"
],
"PotentialTopicEntityMention": [
"abraham"
],
"TopicEntityName": [
"abraham"
],
"TopicEntityMid": [
"m.0pwb"
],
"InferentialChain": [
"people.person.spouse_s..people.marriage.spouse"
],
"Answers": [
{
"AnswersMid": [
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: In the Bible, who was the wife of Abraham? | Spouse |
FreebaseQA-train-296 | In the Bible, who bore Abraham a son called Ishmael? | in the bible, who bore abraham a son called ishmael | {
"Parse-Id": [
"FreebaseQA-train-296.P0"
],
"PotentialTopicEntityMention": [
"abraham"
],
"TopicEntityName": [
"abraham"
],
"TopicEntityMid": [
"m.0pwb"
],
"InferentialChain": [
"people.person.spouse_s..people.marriage.spouse"
],
"Answers": [
{
"AnswersMid": [
... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: In the Bible, who bore Abraham a son called Ishmael? | Spouse |
FreebaseQA-train-297 | In the Bible, who was the servant of Sarah who became the concubine of Abraham and mother of Ishmael? | in the bible, who was the servant of sarah who became the concubine of abraham and mother of ishmael | {
"Parse-Id": [
"FreebaseQA-train-297.P0",
"FreebaseQA-train-297.P1"
],
"PotentialTopicEntityMention": [
"abraham",
"ishmael"
],
"TopicEntityName": [
"abraham",
"ishmael"
],
"TopicEntityMid": [
"m.0pwb",
"m.01ghr4"
],
"InferentialChain": [
"people.person.spouse_s..p... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: In the Bible, who was the servant of Sarah who became the concubine of Abraham and mother of Ishmael? | Spouse |
FreebaseQA-train-298 | Which alliteratively-named politician served as Abraham Lincoln's Vice-President between 1861 and 1865? | which alliteratively-named politician served as abraham lincoln's vice-president between 1861 and 1865 | {
"Parse-Id": [
"FreebaseQA-train-298.P0"
],
"PotentialTopicEntityMention": [
"abraham lincoln"
],
"TopicEntityName": [
"abraham lincoln"
],
"TopicEntityMid": [
"m.0gzh"
],
"InferentialChain": [
"government.us_president.vice_president"
],
"Answers": [
{
"AnswersMid": ... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Which alliteratively-named politician served as Abraham Lincoln's Vice-President between 1861 and 1865? | Vice president |
FreebaseQA-train-299 | Who immediately succeeded Abraham Lincoln as President? | who immediately succeeded abraham lincoln as president | {
"Parse-Id": [
"FreebaseQA-train-299.P0"
],
"PotentialTopicEntityMention": [
"abraham lincoln"
],
"TopicEntityName": [
"abraham lincoln"
],
"TopicEntityMid": [
"m.0gzh"
],
"InferentialChain": [
"government.us_president.vice_president"
],
"Answers": [
{
"AnswersMid": ... | inference_chain_prompt | What category best describes the answer to the following question?
Question: {{RawQuestion}}
|||
{% set answer = Parses.InferentialChain | first %}
{{ answer.split(".") | last | capitalize | replace("_", " ")}}
| What category best describes the answer to the following question?
Question: Who immediately succeeded Abraham Lincoln as President? | Vice president |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.