Question-ID
stringlengths
18
22
RawQuestion
stringlengths
18
933
ProcessedQuestion
stringlengths
17
932
Parses
dict
template_name
stringclasses
5 values
template
stringclasses
5 values
rendered_input
stringlengths
50
1.11k
rendered_output
stringlengths
1
58
FreebaseQA-train-200
What city will host the 2020 Summer Olympics?
what city will host the 2020 summer olympics
{ "Parse-Id": [ "FreebaseQA-train-200.P0" ], "PotentialTopicEntityMention": [ "2020 summer olympics" ], "TopicEntityName": [ "2020 summer olympics" ], "TopicEntityMid": [ "m.05gj2h" ], "InferentialChain": [ "time.event.locations" ], "Answers": [ { "AnswersMid": [ ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: What city will host the 2020 Summer Olympics?
Locations
FreebaseQA-train-201
Which city was awarded the 2020 Summer Olympics last month, beating Istanbul and Madrid ?
which city was awarded the 2020 summer olympics last month, beating istanbul and madrid
{ "Parse-Id": [ "FreebaseQA-train-201.P0" ], "PotentialTopicEntityMention": [ "2020 summer olympics" ], "TopicEntityName": [ "2020 summer olympics" ], "TopicEntityMid": [ "m.05gj2h" ], "InferentialChain": [ "time.event.locations" ], "Answers": [ { "AnswersMid": [ ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which city was awarded the 2020 Summer Olympics last month, beating Istanbul and Madrid ?
Locations
FreebaseQA-train-202
What 1965 film was credited with saving 20th Century Fox from bankruptcy?
what 1965 film was credited with saving 20th century fox from bankruptcy
{ "Parse-Id": [ "FreebaseQA-train-202.P0", "FreebaseQA-train-202.P1" ], "PotentialTopicEntityMention": [ "20th century fox", "20th century fox" ], "TopicEntityName": [ "20th century fox", "20th century fox" ], "TopicEntityMid": [ "m.016tt2", "m.016tt2" ], "InferentialCh...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: What 1965 film was credited with saving 20th Century Fox from bankruptcy?
Film
FreebaseQA-train-203
'Dookie' and '21st. Century Breakdown' are albums by which American rock band?
'dookie' and '21st. century breakdown' are albums by which american rock band
{ "Parse-Id": [ "FreebaseQA-train-203.P0", "FreebaseQA-train-203.P1", "FreebaseQA-train-203.P2", "FreebaseQA-train-203.P3", "FreebaseQA-train-203.P4" ], "PotentialTopicEntityMention": [ "21st. century breakdown", "21st. century breakdown", "21st. century breakdown", "dookie", ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: 'Dookie' and '21st. Century Breakdown' are albums by which American rock band?
Producers
FreebaseQA-train-204
Who released an album in May 2009 called 21st Century Breakdown?
who released an album in may 2009 called 21st century breakdown
{ "Parse-Id": [ "FreebaseQA-train-204.P0", "FreebaseQA-train-204.P1", "FreebaseQA-train-204.P2", "FreebaseQA-train-204.P3" ], "PotentialTopicEntityMention": [ "21st century breakdown", "21st century breakdown", "21st century breakdown", "21st century breakdown" ], "TopicEntityN...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who released an album in May 2009 called 21st Century Breakdown?
Producers
FreebaseQA-train-205
Who stepped down as chief of 21st Century Fox in 2015?
who stepped down as chief of 21st century fox in 2015
{ "Parse-Id": [ "FreebaseQA-train-205.P0" ], "PotentialTopicEntityMention": [ "21st century fox" ], "TopicEntityName": [ "21st century fox" ], "TopicEntityMid": [ "m.0n_hqj6" ], "InferentialChain": [ "organization.organization.board_members..organization.organization_board_membersh...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who stepped down as chief of 21st Century Fox in 2015?
Member
FreebaseQA-train-206
In which country is the spectacular 1650 stone-built 23-arch Khaju Bridge?
in which country is the spectacular 1650 stone-built 23-arch khaju bridge
{ "Parse-Id": [ "FreebaseQA-train-206.P0" ], "PotentialTopicEntityMention": [ "23-arch khaju bridge" ], "TopicEntityName": [ "khaju bridge" ], "TopicEntityMid": [ "m.02877xv" ], "InferentialChain": [ "location.location.containedby" ], "Answers": [ { "AnswersMid": [ ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: In which country is the spectacular 1650 stone-built 23-arch Khaju Bridge?
Containedby
FreebaseQA-train-207
Who played 'Tony Wilson' in the 202 film '24 Hour Party People'?
who played 'tony wilson' in the 202 film '24 hour party people'
{ "Parse-Id": [ "FreebaseQA-train-207.P0" ], "PotentialTopicEntityMention": [ "24 hour party people" ], "TopicEntityName": [ "24 hour party people" ], "TopicEntityMid": [ "m.01_bnz" ], "InferentialChain": [ "film.film.starring..film.performance.actor" ], "Answers": [ { ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who played 'Tony Wilson' in the 202 film '24 Hour Party People'?
Actor
FreebaseQA-train-208
Which singer who sang '24 Hours From Tulsa' died April 2006 at the age of 65?
which singer who sang '24 hours from tulsa' died april 2006 at the age of 65
{ "Parse-Id": [ "FreebaseQA-train-208.P0" ], "PotentialTopicEntityMention": [ "24 hours from tulsa" ], "TopicEntityName": [ "twenty four hours from tulsa" ], "TopicEntityMid": [ "m.011mzj6" ], "InferentialChain": [ "music.recording.artist" ], "Answers": [ { "AnswersMi...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which singer who sang '24 Hours From Tulsa' died April 2006 at the age of 65?
Artist
FreebaseQA-train-209
What was the 2007 segue! to the film 28 Days Later?
what was the 2007 segue! to the film 28 days later
{ "Parse-Id": [ "FreebaseQA-train-209.P0" ], "PotentialTopicEntityMention": [ "28 days later" ], "TopicEntityName": [ "28 days later" ], "TopicEntityMid": [ "m.012kyx" ], "InferentialChain": [ "film.film.sequel" ], "Answers": [ { "AnswersMid": [ "m.0fhd04" ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: What was the 2007 segue! to the film 28 Days Later?
Sequel
FreebaseQA-train-210
'2 Become 1', 'Say You'll Be There' and 'Viva Forever' were all number one hit singles by which British group?
'2 become 1', 'say you'll be there' and 'viva forever' were all number one hit singles by which british group
{ "Parse-Id": [ "FreebaseQA-train-210.P0", "FreebaseQA-train-210.P1", "FreebaseQA-train-210.P2", "FreebaseQA-train-210.P3", "FreebaseQA-train-210.P4", "FreebaseQA-train-210.P5", "FreebaseQA-train-210.P6", "FreebaseQA-train-210.P7", "FreebaseQA-train-210.P8" ], "PotentialTopicEn...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: '2 Become 1', 'Say You'll Be There' and 'Viva Forever' were all number one hit singles by which British group?
Artist
FreebaseQA-train-211
'2 Become 1', 'Say You'll Be There' and 'Viva Forever' were all number one hit singles by which British group?
'2 become 1', 'say you'll be there' and 'viva forever' were all number one hit singles by which british group
{ "Parse-Id": [ "FreebaseQA-train-211.P0", "FreebaseQA-train-211.P1", "FreebaseQA-train-211.P2", "FreebaseQA-train-211.P3", "FreebaseQA-train-211.P4", "FreebaseQA-train-211.P5", "FreebaseQA-train-211.P6", "FreebaseQA-train-211.P7", "FreebaseQA-train-211.P8", "FreebaseQA-train-2...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: '2 Become 1', 'Say You'll Be There' and 'Viva Forever' were all number one hit singles by which British group?
Artist
FreebaseQA-train-212
Kenneth Moore portrayed which famous 2nd World War fighter pilot in the 1956 film Reach for the Sky?
kenneth moore portrayed which famous 2nd world war fighter pilot in the 1956 film reach for the sky
{ "Parse-Id": [ "FreebaseQA-train-212.P0", "FreebaseQA-train-212.P1" ], "PotentialTopicEntityMention": [ "2nd world war", "reach for the sky" ], "TopicEntityName": [ "world war ii", "reach for the sky" ], "TopicEntityMid": [ "m.081pw", "m.04cgkz" ], "InferentialChain": ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Kenneth Moore portrayed which famous 2nd World War fighter pilot in the 1956 film Reach for the Sky?
Works
FreebaseQA-train-213
Which 'kind of love' did 2 Pac have a hit with in 1996?
which 'kind of love' did 2 pac have a hit with in 1996
{ "Parse-Id": [ "FreebaseQA-train-213.P0" ], "PotentialTopicEntityMention": [ "2 pac" ], "TopicEntityName": [ "tupac shakur" ], "TopicEntityMid": [ "m.07pzc" ], "InferentialChain": [ "music.artist.track" ], "Answers": [ { "AnswersMid": [ "m.0112h_g" ], ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which 'kind of love' did 2 Pac have a hit with in 1996?
Track
FreebaseQA-train-214
At which narrow pass did Leonidas and 300 Spartans fight to the death, in order to slow down the Persian army?
at which narrow pass did leonidas and 300 spartans fight to the death, in order to slow down the persian army
{ "Parse-Id": [ "FreebaseQA-train-214.P0" ], "PotentialTopicEntityMention": [ "300 spartans" ], "TopicEntityName": [ "battle of thermopylae" ], "TopicEntityMid": [ "m.02pyp8p" ], "InferentialChain": [ "time.event.locations" ], "Answers": [ { "AnswersMid": [ "m...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: At which narrow pass did Leonidas and 300 Spartans fight to the death, in order to slow down the Persian army?
Locations
FreebaseQA-train-215
In 480 BC, Leonidas led his 300 Spartans against the Persian army, led by Xerxes, in which land battle?
in 480 bc, leonidas led his 300 spartans against the persian army, led by xerxes, in which land battle
{ "Parse-Id": [ "FreebaseQA-train-215.P0" ], "PotentialTopicEntityMention": [ "300 spartans" ], "TopicEntityName": [ "battle of thermopylae" ], "TopicEntityMid": [ "m.0hzngz9" ], "InferentialChain": [ "time.event.locations" ], "Answers": [ { "AnswersMid": [ "m...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: In 480 BC, Leonidas led his 300 Spartans against the Persian army, led by Xerxes, in which land battle?
Locations
FreebaseQA-train-216
Which famous actor is the lead singer with the rock group 30 odd foot of grunts?
which famous actor is the lead singer with the rock group 30 odd foot of grunts
{ "Parse-Id": [ "FreebaseQA-train-216.P0" ], "PotentialTopicEntityMention": [ "30 odd foot of grunts" ], "TopicEntityName": [ "30 odd foot of grunts" ], "TopicEntityMid": [ "m.01mh7_2" ], "InferentialChain": [ "music.musical_group.member..music.group_membership.member" ], "Answ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which famous actor is the lead singer with the rock group 30 odd foot of grunts?
Member
FreebaseQA-train-217
Who was the author of 'The 39 steps'?
who was the author of 'the 39 steps'
{ "Parse-Id": [ "FreebaseQA-train-217.P0", "FreebaseQA-train-217.P1" ], "PotentialTopicEntityMention": [ "39 steps", "39 steps" ], "TopicEntityName": [ "the 39 steps", "the 39 steps" ], "TopicEntityMid": [ "m.09xcn4", "m.0llgk" ], "InferentialChain": [ "film.film.st...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who was the author of 'The 39 steps'?
Story by
FreebaseQA-train-218
Who wrote the novel `The 39 Steps`?
who wrote the novel `the 39 steps`
{ "Parse-Id": [ "FreebaseQA-train-218.P0", "FreebaseQA-train-218.P1" ], "PotentialTopicEntityMention": [ "39 steps", "39 steps" ], "TopicEntityName": [ "the 39 steps", "the 39 steps" ], "TopicEntityMid": [ "m.09xcn4", "m.0llgk" ], "InferentialChain": [ "film.film.st...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who wrote the novel `The 39 Steps`?
Story by
FreebaseQA-train-219
Who wrote the novel The 39 Steps?
who wrote the novel the 39 steps
{ "Parse-Id": [ "FreebaseQA-train-219.P0" ], "PotentialTopicEntityMention": [ "39 steps" ], "TopicEntityName": [ "the 39 steps" ], "TopicEntityMid": [ "m.0llgk" ], "InferentialChain": [ "book.written_work.author" ], "Answers": [ { "AnswersMid": [ "m.0llf9" ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who wrote the novel The 39 Steps?
Author
FreebaseQA-train-220
Which British singer released a 2009 album entitled '3 Words'?
which british singer released a 2009 album entitled '3 words'
{ "Parse-Id": [ "FreebaseQA-train-220.P0", "FreebaseQA-train-220.P1" ], "PotentialTopicEntityMention": [ "3 words", "3 words" ], "TopicEntityName": [ "3 words", "3 words" ], "TopicEntityMid": [ "m.01123htn", "m.07kfy9x" ], "InferentialChain": [ "music.recording.arti...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which British singer released a 2009 album entitled '3 Words'?
Artist
FreebaseQA-train-221
Which US actor, who appeared in the film The 40 year old Virgin, also played the lead role in the US version of The Office?
which us actor, who appeared in the film the 40 year old virgin, also played the lead role in the us version of the office
{ "Parse-Id": [ "FreebaseQA-train-221.P0", "FreebaseQA-train-221.P1", "FreebaseQA-train-221.P2" ], "PotentialTopicEntityMention": [ "40 year old virgin", "us version of \"\"the office", "us version of \"\"the office" ], "TopicEntityName": [ "the 40-year-old virgin", "the office...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which US actor, who appeared in the film The 40 year old Virgin, also played the lead role in the US version of The Office?
Actor
FreebaseQA-train-222
Which Emeritus Professor at Edinburgh University has written several series of novels including '2 Pillars Of Wisdom' and '44 Scotland Street'?
which emeritus professor at edinburgh university has written several series of novels including '2 pillars of wisdom' and '44 scotland street'
{ "Parse-Id": [ "FreebaseQA-train-222.P0", "FreebaseQA-train-222.P1", "FreebaseQA-train-222.P2" ], "PotentialTopicEntityMention": [ "44 scotland street", "44 scotland street", "44 scotland street" ], "TopicEntityName": [ "44 scotland street", "44 scotland street", "44 scotl...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which Emeritus Professor at Edinburgh University has written several series of novels including '2 Pillars Of Wisdom' and '44 Scotland Street'?
Author editor
FreebaseQA-train-223
What Rhodesian-born Scottish writer, who did a reading at SPL on Friday, is the creative forced behind the No. 1 Ladies' Detective Agency, 44 Scotland Street, Sunday Philosophy Club, and the Portuguese Irregular Verbs book series?
what rhodesian-born scottish writer, who did a reading at spl on friday, is the creative forced behind the no. 1 ladies' detective agency, 44 scotland street, sunday philosophy club, and the portuguese irregular verbs book series
{ "Parse-Id": [ "FreebaseQA-train-223.P0", "FreebaseQA-train-223.P1", "FreebaseQA-train-223.P2", "FreebaseQA-train-223.P3", "FreebaseQA-train-223.P4", "FreebaseQA-train-223.P5", "FreebaseQA-train-223.P6", "FreebaseQA-train-223.P7" ], "PotentialTopicEntityMention": [ "44 scotlan...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: What Rhodesian-born Scottish writer, who did a reading at SPL on Friday, is the creative forced behind the No. 1 Ladies' Detective Agency, 44 Scotland Street, Sunday Philosophy Club, and the Portuguese Irregular Verbs book series?
Author
FreebaseQA-train-224
What kind of shop did 50 Cent rap about in 2005?
what kind of shop did 50 cent rap about in 2005
{ "Parse-Id": [ "FreebaseQA-train-224.P0" ], "PotentialTopicEntityMention": [ "50 cent" ], "TopicEntityName": [ "50 cent" ], "TopicEntityMid": [ "m.01vvyc_" ], "InferentialChain": [ "music.artist.track" ], "Answers": [ { "AnswersMid": [ "m.010qghs" ], ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: What kind of shop did 50 Cent rap about in 2005?
Track
FreebaseQA-train-225
Who hosted the '$64,000 Question' on British television?
who hosted the '$64,000 question' on british television
{ "Parse-Id": [ "FreebaseQA-train-225.P0", "FreebaseQA-train-225.P1" ], "PotentialTopicEntityMention": [ "64,000 question", "64,000 question" ], "TopicEntityName": [ "the $64,000 question", "the $64,000 question" ], "TopicEntityMid": [ "m.09k70gp", "m.09k70gp" ], "Infer...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who hosted the '$64,000 Question' on British television?
Actor
FreebaseQA-train-226
On television who live at 742 Evergreen Terrace, Springfield?
on television who live at 742 evergreen terrace, springfield
{ "Parse-Id": [ "FreebaseQA-train-226.P0", "FreebaseQA-train-226.P1" ], "PotentialTopicEntityMention": [ "742 evergreen terrace", "springfield" ], "TopicEntityName": [ "the simpsons house", "springfield" ], "TopicEntityMid": [ "m.08mftp", "m.0m3n5" ], "InferentialChain"...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: On television who live at 742 Evergreen Terrace, Springfield?
Universe
FreebaseQA-train-227
Which family live at 742 Evergreen Terrace?
which family live at 742 evergreen terrace
{ "Parse-Id": [ "FreebaseQA-train-227.P0" ], "PotentialTopicEntityMention": [ "742 evergreen terrace" ], "TopicEntityName": [ "the simpsons house" ], "TopicEntityMid": [ "m.08mftp" ], "InferentialChain": [ "fictional_universe.fictional_setting.universe" ], "Answers": [ { ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which family live at 742 Evergreen Terrace?
Universe
FreebaseQA-train-228
Which fictional family live at 742 Evergreen Terrace, Springfield?
which fictional family live at 742 evergreen terrace, springfield
{ "Parse-Id": [ "FreebaseQA-train-228.P0", "FreebaseQA-train-228.P1" ], "PotentialTopicEntityMention": [ "742 evergreen terrace", "springfield" ], "TopicEntityName": [ "the simpsons house", "springfield" ], "TopicEntityMid": [ "m.08mftp", "m.0m3n5" ], "InferentialChain"...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which fictional family live at 742 Evergreen Terrace, Springfield?
Universe
FreebaseQA-train-229
Which TV family live at 742 Evergreen Terrace?
which tv family live at 742 evergreen terrace
{ "Parse-Id": [ "FreebaseQA-train-229.P0" ], "PotentialTopicEntityMention": [ "742 evergreen terrace" ], "TopicEntityName": [ "the simpsons house" ], "TopicEntityMid": [ "m.08mftp" ], "InferentialChain": [ "fictional_universe.fictional_setting.universe" ], "Answers": [ { ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which TV family live at 742 Evergreen Terrace?
Universe
FreebaseQA-train-230
Which composers 7th Symphony is known as the Leningrad symphony?
which composers 7th symphony is known as the leningrad symphony
{ "Parse-Id": [ "FreebaseQA-train-230.P0" ], "PotentialTopicEntityMention": [ "7th symphony" ], "TopicEntityName": [ "symphony no. 7" ], "TopicEntityMid": [ "m.049rr2j" ], "InferentialChain": [ "music.album.artist" ], "Answers": [ { "AnswersMid": [ "m.02ck1" ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which composers 7th Symphony is known as the Leningrad symphony?
Artist
FreebaseQA-train-231
84 Charing Cross Road is a book based on 20 years of correspondence between which New York writer and Frank Doel, an antiquarian bookseller in London?
84 charing cross road is a book based on 20 years of correspondence between which new york writer and frank doel, an antiquarian bookseller in london
{ "Parse-Id": [ "FreebaseQA-train-231.P0" ], "PotentialTopicEntityMention": [ "84 charing cross road" ], "TopicEntityName": [ "84, charing cross road" ], "TopicEntityMid": [ "m.03lzhc" ], "InferentialChain": [ "book.written_work.author" ], "Answers": [ { "AnswersMid":...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: 84 Charing Cross Road is a book based on 20 years of correspondence between which New York writer and Frank Doel, an antiquarian bookseller in London?
Author
FreebaseQA-train-232
'8 Mile' was a 2002 film starring rapper Eminem, in which US city was it set?
'8 mile' was a 2002 film starring rapper eminem, in which us city was it set
{ "Parse-Id": [ "FreebaseQA-train-232.P0", "FreebaseQA-train-232.P1" ], "PotentialTopicEntityMention": [ "8 mile", "eminem" ], "TopicEntityName": [ "8 mile", "eminem" ], "TopicEntityMid": [ "m.017d93", "m.01vsgrn" ], "InferentialChain": [ "film.film.featured_film_lo...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: '8 Mile' was a 2002 film starring rapper Eminem, in which US city was it set?
Featured film locations
FreebaseQA-train-233
Who was the star of 8 Mile (2002) a hip-hop drama film which was a fictionalized account of his growing up in Detroit and early career in rap?
who was the star of 8 mile (2002) a hip-hop drama film which was a fictionalized account of his growing up in detroit and early career in rap
{ "Parse-Id": [ "FreebaseQA-train-233.P0" ], "PotentialTopicEntityMention": [ "8 mile" ], "TopicEntityName": [ "8 mile" ], "TopicEntityMid": [ "m.017d93" ], "InferentialChain": [ "film.film.starring..film.performance.actor" ], "Answers": [ { "AnswersMid": [ "m...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who was the star of 8 Mile (2002) a hip-hop drama film which was a fictionalized account of his growing up in Detroit and early career in rap?
Actor
FreebaseQA-train-234
Which stand-up comedian presents 8 Out Of 10 Cats and Distraction?
which stand-up comedian presents 8 out of 10 cats and distraction
{ "Parse-Id": [ "FreebaseQA-train-234.P0", "FreebaseQA-train-234.P1" ], "PotentialTopicEntityMention": [ "8 out of 10 cats", "distraction" ], "TopicEntityName": [ "8 out of 10 cats", "distraction" ], "TopicEntityMid": [ "m.06km6x", "m.098n5g" ], "InferentialChain": [ ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which stand-up comedian presents 8 Out Of 10 Cats and Distraction?
Person
FreebaseQA-train-235
Who had a hit in 1984 with '99 Red Balloons'?
who had a hit in 1984 with '99 red balloons'
{ "Parse-Id": [ "FreebaseQA-train-235.P0", "FreebaseQA-train-235.P1", "FreebaseQA-train-235.P2", "FreebaseQA-train-235.P3", "FreebaseQA-train-235.P4" ], "PotentialTopicEntityMention": [ "99 red balloons", "99 red balloons", "99 red balloons", "99 red balloons", "99 red ball...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who had a hit in 1984 with '99 Red Balloons'?
Artist
FreebaseQA-train-236
Who had a hit in 1984 with `99 Red Balloons`?
who had a hit in 1984 with `99 red balloons`
{ "Parse-Id": [ "FreebaseQA-train-236.P0", "FreebaseQA-train-236.P1", "FreebaseQA-train-236.P2", "FreebaseQA-train-236.P3" ], "PotentialTopicEntityMention": [ "99 red balloons", "99 red balloons", "99 red balloons", "99 red balloons" ], "TopicEntityName": [ "99 luftballons"...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who had a hit in 1984 with `99 Red Balloons`?
Lyricist
FreebaseQA-train-237
What name is given to the Fourth (4th) movement in Beethoven's / Ninth (9th) Symphony?
what name is given to the fourth (4th) movement in beethoven's / ninth (9th) symphony
{ "Parse-Id": [ "FreebaseQA-train-237.P0", "FreebaseQA-train-237.P1" ], "PotentialTopicEntityMention": [ "9th", "beethoven" ], "TopicEntityName": [ "symphony no. 9", "ludwig van beethoven" ], "TopicEntityMid": [ "m.01hrpy8", "m.04k15" ], "InferentialChain": [ "music...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: What name is given to the Fourth (4th) movement in Beethoven's / Ninth (9th) Symphony?
Compositions
FreebaseQA-train-238
Which singer/songwriter wrote the musical '9 to 5'?
which singer/songwriter wrote the musical '9 to 5'
{ "Parse-Id": [ "FreebaseQA-train-238.P0", "FreebaseQA-train-238.P1", "FreebaseQA-train-238.P2", "FreebaseQA-train-238.P3", "FreebaseQA-train-238.P4", "FreebaseQA-train-238.P5" ], "PotentialTopicEntityMention": [ "9 to 5", "9 to 5", "9 to 5", "9 to 5", "9 to 5", "9 ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which singer/songwriter wrote the musical '9 to 5'?
Composer
FreebaseQA-train-239
Who sang the theme song in 9 to 5?
who sang the theme song in 9 to 5
{ "Parse-Id": [ "FreebaseQA-train-239.P0", "FreebaseQA-train-239.P1", "FreebaseQA-train-239.P2", "FreebaseQA-train-239.P3" ], "PotentialTopicEntityMention": [ "9 to 5", "9 to 5", "9 to 5", "9 to 5" ], "TopicEntityName": [ "9 to 5", "9 to 5", "9 to 5", "9 to 5" ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who sang the theme song in 9 to 5?
Composer
FreebaseQA-train-240
Who had a UK top ten hit in July, 1980 with `9 to 5`?
who had a uk top ten hit in july, 1980 with `9 to 5`
{ "Parse-Id": [ "FreebaseQA-train-240.P0" ], "PotentialTopicEntityMention": [ "9 to 5" ], "TopicEntityName": [ "9 to 5" ], "TopicEntityMid": [ "m.0_vb6g" ], "InferentialChain": [ "music.recording.artist" ], "Answers": [ { "AnswersMid": [ "m.01vzxld" ], ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who had a UK top ten hit in July, 1980 with `9 to 5`?
Artist
FreebaseQA-train-241
A.A. Milne is most famous for creating which Bear?
a.a. milne is most famous for creating which bear
{ "Parse-Id": [ "FreebaseQA-train-241.P0", "FreebaseQA-train-241.P1" ], "PotentialTopicEntityMention": [ "a.a. milne", "a.a. milne" ], "TopicEntityName": [ "a. a. milne", "a. a. milne" ], "TopicEntityMid": [ "m.0k_s", "m.0k_s" ], "InferentialChain": [ "book.author.b...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: A.A. Milne is most famous for creating which Bear?
Book editions published
FreebaseQA-train-242
Odense and Aarhus are major cities in which European country?
odense and aarhus are major cities in which european country
{ "Parse-Id": [ "FreebaseQA-train-242.P0", "FreebaseQA-train-242.P1", "FreebaseQA-train-242.P2" ], "PotentialTopicEntityMention": [ "aarhus", "aarhus", "odense" ], "TopicEntityName": [ "aarhus", "aarhus", "odense" ], "TopicEntityMid": [ "m.0lzp", "m.0lzp", "...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Odense and Aarhus are major cities in which European country?
Country
FreebaseQA-train-243
In the Bible who was the elder sister of Moses and Aaron?
in the bible who was the elder sister of moses and aaron
{ "Parse-Id": [ "FreebaseQA-train-243.P0", "FreebaseQA-train-243.P1" ], "PotentialTopicEntityMention": [ "aaron", "moses" ], "TopicEntityName": [ "aaron", "moses" ], "TopicEntityMid": [ "m.0llm", "m.04yfh" ], "InferentialChain": [ "people.person.sibling_s..people.si...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: In the Bible who was the elder sister of Moses and Aaron?
Sibling
FreebaseQA-train-244
In the Bible, who was the sister of Moses and Aaron?
in the bible, who was the sister of moses and aaron
{ "Parse-Id": [ "FreebaseQA-train-244.P0", "FreebaseQA-train-244.P1" ], "PotentialTopicEntityMention": [ "aaron", "moses" ], "TopicEntityName": [ "aaron", "moses" ], "TopicEntityMid": [ "m.0llm", "m.04yfh" ], "InferentialChain": [ "people.person.sibling_s..people.si...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: In the Bible, who was the sister of Moses and Aaron?
Sibling
FreebaseQA-train-245
In the Old Testament, who is the sister of Moses and Aaron?
in the old testament, who is the sister of moses and aaron
{ "Parse-Id": [ "FreebaseQA-train-245.P0", "FreebaseQA-train-245.P1" ], "PotentialTopicEntityMention": [ "aaron", "moses" ], "TopicEntityName": [ "aaron", "moses" ], "TopicEntityMid": [ "m.0llm", "m.04yfh" ], "InferentialChain": [ "people.person.sibling_s..people.si...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: In the Old Testament, who is the sister of Moses and Aaron?
Sibling
FreebaseQA-train-246
What is the name of Moses and Aaron's sister?
what is the name of moses and aaron's sister
{ "Parse-Id": [ "FreebaseQA-train-246.P0", "FreebaseQA-train-246.P1" ], "PotentialTopicEntityMention": [ "aaron", "moses" ], "TopicEntityName": [ "aaron", "moses" ], "TopicEntityMid": [ "m.0llm", "m.04yfh" ], "InferentialChain": [ "people.person.sibling_s..people.si...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: What is the name of Moses and Aaron's sister?
Sibling
FreebaseQA-train-247
Aaron Copeland wrote a ballet about which American folk hero?
aaron copeland wrote a ballet about which american folk hero
{ "Parse-Id": [ "FreebaseQA-train-247.P0" ], "PotentialTopicEntityMention": [ "aaron copeland" ], "TopicEntityName": [ "aaron copeland" ], "TopicEntityMid": [ "m.0djg9" ], "InferentialChain": [ "base.ballet.ballet_composer.ballets_composed" ], "Answers": [ { "AnswersM...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Aaron Copeland wrote a ballet about which American folk hero?
Ballets composed
FreebaseQA-train-248
Which ballet by Aaron Copeland, first performed in 1944, contains a mountain range and a season in its name. It tells the story of a celebration of the building of a Pennsylvanian farmhouse?
which ballet by aaron copeland, first performed in 1944, contains a mountain range and a season in its name. it tells the story of a celebration of the building of a pennsylvanian farmhouse
{ "Parse-Id": [ "FreebaseQA-train-248.P0" ], "PotentialTopicEntityMention": [ "aaron copeland" ], "TopicEntityName": [ "aaron copeland" ], "TopicEntityMid": [ "m.0djg9" ], "InferentialChain": [ "music.composer.compositions" ], "Answers": [ { "AnswersMid": [ "m...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which ballet by Aaron Copeland, first performed in 1944, contains a mountain range and a season in its name. It tells the story of a celebration of the building of a Pennsylvanian farmhouse?
Compositions
FreebaseQA-train-249
In October 2009 it was announced that director Sam Taylor-Wood was to marry Aaron Johnson, 23 years her junior, who were director and star respecitively of Nowhere Boy, a film about whom?
in october 2009 it was announced that director sam taylor-wood was to marry aaron johnson, 23 years her junior, who were director and star respecitively of nowhere boy, a film about whom
{ "Parse-Id": [ "FreebaseQA-train-249.P0", "FreebaseQA-train-249.P1", "FreebaseQA-train-249.P2" ], "PotentialTopicEntityMention": [ "aaron johnson", "nowhere boy", "nowhere boy" ], "TopicEntityName": [ "aaron taylor-johnson", "nowhere boy", "nowhere boy" ], "TopicEntity...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: In October 2009 it was announced that director Sam Taylor-Wood was to marry Aaron Johnson, 23 years her junior, who were director and star respecitively of Nowhere Boy, a film about whom?
Character
FreebaseQA-train-250
Alex Band and Aaron Kamin make up which band?
alex band and aaron kamin make up which band
{ "Parse-Id": [ "FreebaseQA-train-250.P0", "FreebaseQA-train-250.P1" ], "PotentialTopicEntityMention": [ "aaron kamin", "alex band" ], "TopicEntityName": [ "aaron kamin", "alex band" ], "TopicEntityMid": [ "m.0768fk", "m.01vcslm" ], "InferentialChain": [ "music.grou...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Alex Band and Aaron Kamin make up which band?
Group
FreebaseQA-train-251
Played by Aaron Paul, who was Walter White's co-protagonist in 'Breaking Bad'?
played by aaron paul, who was walter white's co-protagonist in 'breaking bad'
{ "Parse-Id": [ "FreebaseQA-train-251.P0", "FreebaseQA-train-251.P1", "FreebaseQA-train-251.P2" ], "PotentialTopicEntityMention": [ "aaron paul", "aaron paul", "breaking bad" ], "TopicEntityName": [ "aaron paul", "aaron paul", "breaking bad" ], "TopicEntityMid": [ "...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Played by Aaron Paul, who was Walter White's co-protagonist in 'Breaking Bad'?
Character
FreebaseQA-train-252
With which song did Swedish pop group Abba win the 1974 Eurovision Song Contest?
with which song did swedish pop group abba win the 1974 eurovision song contest
{ "Parse-Id": [ "FreebaseQA-train-252.P0", "FreebaseQA-train-252.P1" ], "PotentialTopicEntityMention": [ "abba", "abba" ], "TopicEntityName": [ "abba", "abba" ], "TopicEntityMid": [ "m.01hnmty", "m.0kjh" ], "InferentialChain": [ "music.release.track", "music.art...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: With which song did Swedish pop group Abba win the 1974 Eurovision Song Contest?
Track
FreebaseQA-train-253
What is the title of Abba's last UK number one hit single?
what is the title of abba's last uk number one hit single
{ "Parse-Id": [ "FreebaseQA-train-253.P0", "FreebaseQA-train-253.P1", "FreebaseQA-train-253.P2" ], "PotentialTopicEntityMention": [ "abba", "abba", "abba" ], "TopicEntityName": [ "abba", "abba", "abba" ], "TopicEntityMid": [ "m.038t528", "m.0kjh", "m.0vyvpd8...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: What is the title of Abba's last UK number one hit single?
Track
FreebaseQA-train-254
What is the title of the last UK number one hit for Swedish group Abba?
what is the title of the last uk number one hit for swedish group abba
{ "Parse-Id": [ "FreebaseQA-train-254.P0", "FreebaseQA-train-254.P1", "FreebaseQA-train-254.P2" ], "PotentialTopicEntityMention": [ "abba", "abba", "abba" ], "TopicEntityName": [ "abba", "abba", "abba" ], "TopicEntityMid": [ "m.038t528", "m.0kjh", "m.0vyvpd8...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: What is the title of the last UK number one hit for Swedish group Abba?
Track
FreebaseQA-train-255
What was the only single by ABBA to reach number 1 on the US Billboard charts?
what was the only single by abba to reach number 1 on the us billboard charts
{ "Parse-Id": [ "FreebaseQA-train-255.P0", "FreebaseQA-train-255.P1" ], "PotentialTopicEntityMention": [ "abba", "abba" ], "TopicEntityName": [ "abba", "abba" ], "TopicEntityMid": [ "m.038t528", "m.0kjh" ], "InferentialChain": [ "music.release.track", "music.com...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: What was the only single by ABBA to reach number 1 on the US Billboard charts?
Track
FreebaseQA-train-256
What was Abba`s first British hit single?
what was abba`s first british hit single
{ "Parse-Id": [ "FreebaseQA-train-256.P0", "FreebaseQA-train-256.P1", "FreebaseQA-train-256.P2" ], "PotentialTopicEntityMention": [ "abba", "abba", "abba" ], "TopicEntityName": [ "abba", "abba", "abba" ], "TopicEntityMid": [ "m.038t528", "m.0kjh", "m.0vyvpd8...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: What was Abba`s first British hit single?
Track
FreebaseQA-train-257
No more carefree laughter, Silence ever after are the opening lines of which Abba Number One hit ?
no more carefree laughter, silence ever after are the opening lines of which abba number one hit
{ "Parse-Id": [ "FreebaseQA-train-257.P0", "FreebaseQA-train-257.P1" ], "PotentialTopicEntityMention": [ "abba", "abba" ], "TopicEntityName": [ "abba", "abba" ], "TopicEntityMid": [ "m.038t528", "m.0kjh" ], "InferentialChain": [ "music.release.track", "music.art...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: No more carefree laughter, Silence ever after are the opening lines of which Abba Number One hit ?
Track
FreebaseQA-train-258
What was Abba's last Number One single in the UK?
what was abba's last number one single in the uk
{ "Parse-Id": [ "FreebaseQA-train-258.P0" ], "PotentialTopicEntityMention": [ "abba" ], "TopicEntityName": [ "abba" ], "TopicEntityMid": [ "m.0kjh" ], "InferentialChain": [ "music.artist.album" ], "Answers": [ { "AnswersMid": [ "m.01hnm5g" ], "Answ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: What was Abba's last Number One single in the UK?
Album
FreebaseQA-train-259
The group 'Abba' was formed in which European capital city in 1971?
the group 'abba' was formed in which european capital city in 1971
{ "Parse-Id": [ "FreebaseQA-train-259.P0" ], "PotentialTopicEntityMention": [ "abba" ], "TopicEntityName": [ "abba" ], "TopicEntityMid": [ "m.0kjh" ], "InferentialChain": [ "music.artist.origin" ], "Answers": [ { "AnswersMid": [ "m.06mxs" ], "Answe...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: The group 'Abba' was formed in which European capital city in 1971?
Origin
FreebaseQA-train-260
Where, in the novel The Count of Monte Christo, by Alexandre Dumas, does the Abbe Faria befriend Edmond Dantes during his time in prison?
where, in the novel the count of monte christo, by alexandre dumas, does the abbe faria befriend edmond dantes during his time in prison
{ "Parse-Id": [ "FreebaseQA-train-260.P0", "FreebaseQA-train-260.P1", "FreebaseQA-train-260.P2" ], "PotentialTopicEntityMention": [ "abbe faria", "count of monte christo", "monte christo" ], "TopicEntityName": [ "abbe faria", "the count of monte cristo", "monte cristo" ],...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Where, in the novel The Count of Monte Christo, by Alexandre Dumas, does the Abbe Faria befriend Edmond Dantes during his time in prison?
Prison
FreebaseQA-train-261
Abbotsford House, on the banks of the Tweed, was the home of which author?
abbotsford house, on the banks of the tweed, was the home of which author
{ "Parse-Id": [ "FreebaseQA-train-261.P0" ], "PotentialTopicEntityMention": [ "abbotsford house" ], "TopicEntityName": [ "abbotsford house" ], "TopicEntityMid": [ "m.0pvy" ], "InferentialChain": [ "architecture.structure.owner..architecture.ownership.owner" ], "Answers": [ ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Abbotsford House, on the banks of the Tweed, was the home of which author?
Owner
FreebaseQA-train-262
Which author lived at Abbotsford House near Galashiels in the Scottish borders?
which author lived at abbotsford house near galashiels in the scottish borders
{ "Parse-Id": [ "FreebaseQA-train-262.P0" ], "PotentialTopicEntityMention": [ "abbotsford house" ], "TopicEntityName": [ "abbotsford house" ], "TopicEntityMid": [ "m.0pvy" ], "InferentialChain": [ "architecture.structure.owner..architecture.ownership.owner" ], "Answers": [ ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which author lived at Abbotsford House near Galashiels in the Scottish borders?
Owner
FreebaseQA-train-263
Which British city is served by 'Abbotsinch Airport'?
which british city is served by 'abbotsinch airport'
{ "Parse-Id": [ "FreebaseQA-train-263.P0" ], "PotentialTopicEntityMention": [ "abbotsinch airport" ], "TopicEntityName": [ "glasgow airport" ], "TopicEntityMid": [ "m.01pnl_" ], "InferentialChain": [ "aviation.airport.serves" ], "Answers": [ { "AnswersMid": [ ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which British city is served by 'Abbotsinch Airport'?
Serves
FreebaseQA-train-264
Which 1970's American TV detective series, was created by film writer Abby Mann?
which 1970's american tv detective series, was created by film writer abby mann
{ "Parse-Id": [ "FreebaseQA-train-264.P0", "FreebaseQA-train-264.P1" ], "PotentialTopicEntityMention": [ "abby mann", "abby mann" ], "TopicEntityName": [ "abby mann", "abby mann" ], "TopicEntityMid": [ "m.0bvjjq", "m.0bvjjq" ], "InferentialChain": [ "tv.tv_program_c...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which 1970's American TV detective series, was created by film writer Abby Mann?
Programs created
FreebaseQA-train-265
In music, Pulp, Def Leppard and ABC all come from which English city?
in music, pulp, def leppard and abc all come from which english city
{ "Parse-Id": [ "FreebaseQA-train-265.P0", "FreebaseQA-train-265.P1", "FreebaseQA-train-265.P2", "FreebaseQA-train-265.P3", "FreebaseQA-train-265.P4", "FreebaseQA-train-265.P5" ], "PotentialTopicEntityMention": [ "abc", "abc", "def leppard", "def leppard", "pulp", "...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: In music, Pulp, Def Leppard and ABC all come from which English city?
Origin
FreebaseQA-train-266
ABC and NBC are TV networks in which country ?
abc and nbc are tv networks in which country
{ "Parse-Id": [ "FreebaseQA-train-266.P0", "FreebaseQA-train-266.P1", "FreebaseQA-train-266.P2" ], "PotentialTopicEntityMention": [ "abc", "abc", "abc" ], "TopicEntityName": [ "abc", "abc", "american broadcasting company" ], "TopicEntityMid": [ "m.0bqs52l", "m.0...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: ABC and NBC are TV networks in which country ?
Country
FreebaseQA-train-267
In which country were elections planned for 1991 were stopped by the military (leading to Civil War), were held in 1995, and, in 1999, after a series of short-term military leaders, Abdelaziz Bouteflika was elected president?
in which country were elections planned for 1991 were stopped by the military (leading to civil war), were held in 1995, and, in 1999, after a series of short-term military leaders, abdelaziz bouteflika was elected president
{ "Parse-Id": [ "FreebaseQA-train-267.P0" ], "PotentialTopicEntityMention": [ "abdelaziz bouteflika" ], "TopicEntityName": [ "abdelaziz bouteflika" ], "TopicEntityMid": [ "m.0224kz" ], "InferentialChain": [ "government.politician.government_positions_held..government.government_pos...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: In which country were elections planned for 1991 were stopped by the military (leading to Civil War), were held in 1995, and, in 1999, after a series of short-term military leaders, Abdelaziz Bouteflika was elected president?
Jurisdiction of office
FreebaseQA-train-268
What was Abdel Kechiche's 2008 film about France's North African community?
what was abdel kechiche's 2008 film about france's north african community
{ "Parse-Id": [ "FreebaseQA-train-268.P0", "FreebaseQA-train-268.P1" ], "PotentialTopicEntityMention": [ "abdel kechiche", "abdel kechiche" ], "TopicEntityName": [ "abdellatif kechiche", "abdellatif kechiche" ], "TopicEntityMid": [ "m.0gpvk3", "m.0gpvk3" ], "Inferential...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: What was Abdel Kechiche's 2008 film about France's North African community?
Film
FreebaseQA-train-269
Who was Minister-Chairman of the Russian Provisional Government and Prime Minister of Russia from 21 July 1917 (following the abdication of Tsar Nicholas II) to 8 November 1917 (the Bolshevik revolution), who is buried in Putney Vale Cemetery, London?
who was minister-chairman of the russian provisional government and prime minister of russia from 21 july 1917 (following the abdication of tsar nicholas ii) to 8 november 1917 (the bolshevik revolution), who is buried in putney vale cemetery, london
{ "Parse-Id": [ "FreebaseQA-train-269.P0" ], "PotentialTopicEntityMention": [ "abdication of tsar" ], "TopicEntityName": [ "russian revolution" ], "TopicEntityMid": [ "m.06fby" ], "InferentialChain": [ "base.culturalevent.event.entity_involved" ], "Answers": [ { "Answ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who was Minister-Chairman of the Russian Provisional Government and Prime Minister of Russia from 21 July 1917 (following the abdication of Tsar Nicholas II) to 8 November 1917 (the Bolshevik revolution), who is buried in Putney Vale Cemetery...
Entity involved
FreebaseQA-train-270
Abdul Aziz Ibn Saud was the first king of which Asian state?
abdul aziz ibn saud was the first king of which asian state
{ "Parse-Id": [ "FreebaseQA-train-270.P0" ], "PotentialTopicEntityMention": [ "abdul aziz ibn saud" ], "TopicEntityName": [ "ibn saud" ], "TopicEntityMid": [ "m.0195b9" ], "InferentialChain": [ "people.person.nationality" ], "Answers": [ { "AnswersMid": [ "m.0...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Abdul Aziz Ibn Saud was the first king of which Asian state?
Nationality
FreebaseQA-train-271
What modern-day country was ruled between 1902 and 1964 by Abdul Aziz Ibn Saud and his son, Emir Saud?
what modern-day country was ruled between 1902 and 1964 by abdul aziz ibn saud and his son, emir saud
{ "Parse-Id": [ "FreebaseQA-train-271.P0", "FreebaseQA-train-271.P1" ], "PotentialTopicEntityMention": [ "abdul aziz ibn saud", "saud" ], "TopicEntityName": [ "ibn saud", "house of saud" ], "TopicEntityMid": [ "m.0195b9", "m.0240lv" ], "InferentialChain": [ "people....
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: What modern-day country was ruled between 1902 and 1964 by Abdul Aziz Ibn Saud and his son, Emir Saud?
Nationality
FreebaseQA-train-272
Abebe Bikila was the only athlete to win the Olympic marathon twice and in successive Olympic games. What country did he represent?
abebe bikila was the only athlete to win the olympic marathon twice and in successive olympic games. what country did he represent
{ "Parse-Id": [ "FreebaseQA-train-272.P0", "FreebaseQA-train-272.P1", "FreebaseQA-train-272.P2" ], "PotentialTopicEntityMention": [ "abebe bikila", "abebe bikila", "abebe bikila" ], "TopicEntityName": [ "abebe bikila", "abebe bikila", "abebe bikila" ], "TopicEntityMid":...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Abebe Bikila was the only athlete to win the Olympic marathon twice and in successive Olympic games. What country did he represent?
Country
FreebaseQA-train-273
Which country does marathon man Abel Anton come form?
which country does marathon man abel anton come form
{ "Parse-Id": [ "FreebaseQA-train-273.P0", "FreebaseQA-train-273.P1" ], "PotentialTopicEntityMention": [ "abel anton", "abel anton" ], "TopicEntityName": [ "abel anton", "abel anton" ], "TopicEntityMid": [ "m.09l2n_", "m.09l2n_" ], "InferentialChain": [ "people.pers...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which country does marathon man Abel Anton come form?
Nationality
FreebaseQA-train-274
The Aberdare mountains are in which African country?
the aberdare mountains are in which african country
{ "Parse-Id": [ "FreebaseQA-train-274.P0" ], "PotentialTopicEntityMention": [ "aberdare mountains" ], "TopicEntityName": [ "aberdare range" ], "TopicEntityMid": [ "m.06kdp6" ], "InferentialChain": [ "location.location.containedby" ], "Answers": [ { "AnswersMid": [ ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: The Aberdare mountains are in which African country?
Containedby
FreebaseQA-train-275
In which town or city is the University of Abertay?
in which town or city is the university of abertay
{ "Parse-Id": [ "FreebaseQA-train-275.P0" ], "PotentialTopicEntityMention": [ "abertay" ], "TopicEntityName": [ "abertay university" ], "TopicEntityMid": [ "m.03p81c" ], "InferentialChain": [ "location.location.containedby" ], "Answers": [ { "AnswersMid": [ "m...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: In which town or city is the University of Abertay?
Containedby
FreebaseQA-train-276
In 2010, Veera Musikapong, Abhisit Vejjajiva, Khattiya Sawasdiphol, Suthep Thaugsuban, Prompong Nopparit, Colonel Sansern Kaewkamnerd, Natthawut Saikua and Buranaj Smutharaks were prominent in the politics of which country?
in 2010, veera musikapong, abhisit vejjajiva, khattiya sawasdiphol, suthep thaugsuban, prompong nopparit, colonel sansern kaewkamnerd, natthawut saikua and buranaj smutharaks were prominent in the politics of which country
{ "Parse-Id": [ "FreebaseQA-train-276.P0" ], "PotentialTopicEntityMention": [ "abhisit vejjajiva" ], "TopicEntityName": [ "abhisit vejjajiva" ], "TopicEntityMid": [ "m.087cw4" ], "InferentialChain": [ "government.politician.government_positions_held..government.government_position_...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: In 2010, Veera Musikapong, Abhisit Vejjajiva, Khattiya Sawasdiphol, Suthep Thaugsuban, Prompong Nopparit, Colonel Sansern Kaewkamnerd, Natthawut Saikua and Buranaj Smutharaks were prominent in the politics of which country?
Jurisdiction of office
FreebaseQA-train-277
Abhorson is an executioner in which Shakespeare play?
abhorson is an executioner in which shakespeare play
{ "Parse-Id": [ "FreebaseQA-train-277.P0", "FreebaseQA-train-277.P1" ], "PotentialTopicEntityMention": [ "abhorson", "abhorson" ], "TopicEntityName": [ "abhorson", "abhorson" ], "TopicEntityMid": [ "m.0zzbjb6", "m.0zzbjb6" ], "InferentialChain": [ "film.film_charact...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Abhorson is an executioner in which Shakespeare play?
Film
FreebaseQA-train-278
Abidjan is the former capital and a major battleground of what civil-war-torn nation in 2011?
abidjan is the former capital and a major battleground of what civil-war-torn nation in 2011
{ "Parse-Id": [ "FreebaseQA-train-278.P0" ], "PotentialTopicEntityMention": [ "abidjan" ], "TopicEntityName": [ "abidjan" ], "TopicEntityMid": [ "m.0fmyd" ], "InferentialChain": [ "base.biblioness.bibs_location.country" ], "Answers": [ { "AnswersMid": [ "m.0fv...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Abidjan is the former capital and a major battleground of what civil-war-torn nation in 2011?
Country
FreebaseQA-train-279
Who wrote the plays Nuts in May and Abigail's Party?
who wrote the plays nuts in may and abigail's party
{ "Parse-Id": [ "FreebaseQA-train-279.P0", "FreebaseQA-train-279.P1", "FreebaseQA-train-279.P2", "FreebaseQA-train-279.P3" ], "PotentialTopicEntityMention": [ "abigail 's party", "nuts in may", "nuts in may", "nuts in may" ], "TopicEntityName": [ "abigail's party", "nut...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who wrote the plays Nuts in May and Abigail's Party?
Author
FreebaseQA-train-280
Abkhazia, recognised as independent by Russia, is a breakaway region from which former Soviet republic?
abkhazia, recognised as independent by russia, is a breakaway region from which former soviet republic
{ "Parse-Id": [ "FreebaseQA-train-280.P0", "FreebaseQA-train-280.P1" ], "PotentialTopicEntityMention": [ "abkhazia", "abkhazia" ], "TopicEntityName": [ "abkhazia", "abkhazia" ], "TopicEntityMid": [ "m.011jw", "m.011jw" ], "InferentialChain": [ "location.administrati...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Abkhazia, recognised as independent by Russia, is a breakaway region from which former Soviet republic?
Country
FreebaseQA-train-281
The Abominable Snowman, a mythological ape-like cryptid said to inhabit the Himalayan region of Nepal and Tibet, is also known as what?
the abominable snowman, a mythological ape-like cryptid said to inhabit the himalayan region of nepal and tibet, is also known as what
{ "Parse-Id": [ "FreebaseQA-train-281.P0", "FreebaseQA-train-281.P1", "FreebaseQA-train-281.P2", "FreebaseQA-train-281.P3", "FreebaseQA-train-281.P4" ], "PotentialTopicEntityMention": [ "abominable snowman", "himalayan region", "nepal", "tibet", "tibet" ], "TopicEntityN...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: The Abominable Snowman, a mythological ape-like cryptid said to inhabit the Himalayan region of Nepal and Tibet, is also known as what?
Cryptid classification
FreebaseQA-train-282
The Abominable Snowman is said to inhabit which mountain range?
the abominable snowman is said to inhabit which mountain range
{ "Parse-Id": [ "FreebaseQA-train-282.P0", "FreebaseQA-train-282.P1" ], "PotentialTopicEntityMention": [ "abominable snowman", "abominable snowman" ], "TopicEntityName": [ "yeti", "yeti" ], "TopicEntityMid": [ "m.0hfj1", "m.0hfj1" ], "InferentialChain": [ "base.myst...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: The Abominable Snowman is said to inhabit which mountain range?
Area of occurrence
FreebaseQA-train-283
With which mountain range would you associate the Abonimable Snowman?
with which mountain range would you associate the abonimable snowman
{ "Parse-Id": [ "FreebaseQA-train-283.P0", "FreebaseQA-train-283.P1" ], "PotentialTopicEntityMention": [ "abonimable snowman", "abonimable snowman" ], "TopicEntityName": [ "yeti", "yeti" ], "TopicEntityMid": [ "m.0hfj1", "m.0hfj1" ], "InferentialChain": [ "base.myst...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: With which mountain range would you associate the Abonimable Snowman?
Area of occurrence
FreebaseQA-train-284
What links the US, 2 May 2011 and Abottabad in Pakistan?
what links the us, 2 may 2011 and abottabad in pakistan
{ "Parse-Id": [ "FreebaseQA-train-284.P0" ], "PotentialTopicEntityMention": [ "abottabad" ], "TopicEntityName": [ "abbottabad" ], "TopicEntityMid": [ "m.011hs1" ], "InferentialChain": [ "location.location.events" ], "Answers": [ { "AnswersMid": [ "m.0bwk4lh" ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: What links the US, 2 May 2011 and Abottabad in Pakistan?
Events
FreebaseQA-train-285
By what name is the Battle of Aboukir Bay (1798) more commonly known?
by what name is the battle of aboukir bay (1798) more commonly known
{ "Parse-Id": [ "FreebaseQA-train-285.P0" ], "PotentialTopicEntityMention": [ "aboukir bay" ], "TopicEntityName": [ "abu qir bay" ], "TopicEntityMid": [ "m.06v4lf" ], "InferentialChain": [ "location.location.events" ], "Answers": [ { "AnswersMid": [ "m.01gq_" ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: By what name is the Battle of Aboukir Bay (1798) more commonly known?
Events
FreebaseQA-train-286
Who wrote the 1998 novel About A Boy?
who wrote the 1998 novel about a boy
{ "Parse-Id": [ "FreebaseQA-train-286.P0", "FreebaseQA-train-286.P1" ], "PotentialTopicEntityMention": [ "about a boy", "about a boy" ], "TopicEntityName": [ "about a boy", "about a boy" ], "TopicEntityMid": [ "m.04tvp02", "m.0jv0d" ], "InferentialChain": [ "book.bo...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who wrote the 1998 novel About A Boy?
Author editor
FreebaseQA-train-287
Who wrote the book `About A Boy` which was later turned into a successful film starring Hugh Grant?
who wrote the book `about a boy` which was later turned into a successful film starring hugh grant
{ "Parse-Id": [ "FreebaseQA-train-287.P0", "FreebaseQA-train-287.P1", "FreebaseQA-train-287.P2" ], "PotentialTopicEntityMention": [ "about a boy", "about a boy", "about a boy" ], "TopicEntityName": [ "about a boy", "about a boy", "about a boy" ], "TopicEntityMid": [ ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who wrote the book `About A Boy` which was later turned into a successful film starring Hugh Grant?
Author editor
FreebaseQA-train-288
Who wrote the novels About A Boy, How To Be Good and High Fidelity?
who wrote the novels about a boy, how to be good and high fidelity
{ "Parse-Id": [ "FreebaseQA-train-288.P0", "FreebaseQA-train-288.P1", "FreebaseQA-train-288.P2", "FreebaseQA-train-288.P3", "FreebaseQA-train-288.P4", "FreebaseQA-train-288.P5", "FreebaseQA-train-288.P6" ], "PotentialTopicEntityMention": [ "about a boy", "about a boy", "abo...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who wrote the novels About A Boy, How To Be Good and High Fidelity?
Author editor
FreebaseQA-train-289
Who had parts in 'St. Elmo's Fire', 'About Last Night ' and 'Wayne's World'?
who had parts in 'st. elmo's fire', 'about last night ' and 'wayne's world'
{ "Parse-Id": [ "FreebaseQA-train-289.P0", "FreebaseQA-train-289.P1" ], "PotentialTopicEntityMention": [ "about last night", "wayne?s world" ], "TopicEntityName": [ "about last night", "wayne's world" ], "TopicEntityMid": [ "m.03rl2l", "m.0kbwb" ], "InferentialChain": [...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who had parts in 'St. Elmo's Fire', 'About Last Night ' and 'Wayne's World'?
Actor
FreebaseQA-train-290
Abracadabra (1982) and The Joker (1990) were big hits for which group?
abracadabra (1982) and the joker (1990) were big hits for which group
{ "Parse-Id": [ "FreebaseQA-train-290.P0", "FreebaseQA-train-290.P1", "FreebaseQA-train-290.P2" ], "PotentialTopicEntityMention": [ "abracadabra", "joker", "joker" ], "TopicEntityName": [ "abracadabra", "the joker", "the joker" ], "TopicEntityMid": [ "m.0f794zy", ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Abracadabra (1982) and The Joker (1990) were big hits for which group?
Artist
FreebaseQA-train-291
In the Bible who was the son of Abraham and Hagar?
in the bible who was the son of abraham and hagar
{ "Parse-Id": [ "FreebaseQA-train-291.P0", "FreebaseQA-train-291.P1" ], "PotentialTopicEntityMention": [ "abraham", "hagar" ], "TopicEntityName": [ "abraham", "hagar" ], "TopicEntityMid": [ "m.0pwb", "m.01sq9c" ], "InferentialChain": [ "people.person.children", ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: In the Bible who was the son of Abraham and Hagar?
Children
FreebaseQA-train-292
In the Bible, who was the son of Abraham and Hagar?
in the bible, who was the son of abraham and hagar
{ "Parse-Id": [ "FreebaseQA-train-292.P0", "FreebaseQA-train-292.P1" ], "PotentialTopicEntityMention": [ "abraham", "hagar" ], "TopicEntityName": [ "abraham", "hagar" ], "TopicEntityMid": [ "m.0pwb", "m.01sq9c" ], "InferentialChain": [ "people.person.children", ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: In the Bible, who was the son of Abraham and Hagar?
Children
FreebaseQA-train-293
In the Bible who was the son of Abraham and Sarah?
in the bible who was the son of abraham and sarah
{ "Parse-Id": [ "FreebaseQA-train-293.P0", "FreebaseQA-train-293.P1" ], "PotentialTopicEntityMention": [ "abraham", "sarah" ], "TopicEntityName": [ "abraham", "sarah" ], "TopicEntityMid": [ "m.0pwb", "m.01b5wx" ], "InferentialChain": [ "people.person.children", ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: In the Bible who was the son of Abraham and Sarah?
Children
FreebaseQA-train-294
In the Bible who was the father of Abraham?
in the bible who was the father of abraham
{ "Parse-Id": [ "FreebaseQA-train-294.P0" ], "PotentialTopicEntityMention": [ "abraham" ], "TopicEntityName": [ "abraham" ], "TopicEntityMid": [ "m.0pwb" ], "InferentialChain": [ "people.person.parents" ], "Answers": [ { "AnswersMid": [ "m.0j0hf" ], ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: In the Bible who was the father of Abraham?
Parents
FreebaseQA-train-295
In the Bible, who was the wife of Abraham?
in the bible, who was the wife of abraham
{ "Parse-Id": [ "FreebaseQA-train-295.P0" ], "PotentialTopicEntityMention": [ "abraham" ], "TopicEntityName": [ "abraham" ], "TopicEntityMid": [ "m.0pwb" ], "InferentialChain": [ "people.person.spouse_s..people.marriage.spouse" ], "Answers": [ { "AnswersMid": [ ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: In the Bible, who was the wife of Abraham?
Spouse
FreebaseQA-train-296
In the Bible, who bore Abraham a son called Ishmael?
in the bible, who bore abraham a son called ishmael
{ "Parse-Id": [ "FreebaseQA-train-296.P0" ], "PotentialTopicEntityMention": [ "abraham" ], "TopicEntityName": [ "abraham" ], "TopicEntityMid": [ "m.0pwb" ], "InferentialChain": [ "people.person.spouse_s..people.marriage.spouse" ], "Answers": [ { "AnswersMid": [ ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: In the Bible, who bore Abraham a son called Ishmael?
Spouse
FreebaseQA-train-297
In the Bible, who was the servant of Sarah who became the concubine of Abraham and mother of Ishmael?
in the bible, who was the servant of sarah who became the concubine of abraham and mother of ishmael
{ "Parse-Id": [ "FreebaseQA-train-297.P0", "FreebaseQA-train-297.P1" ], "PotentialTopicEntityMention": [ "abraham", "ishmael" ], "TopicEntityName": [ "abraham", "ishmael" ], "TopicEntityMid": [ "m.0pwb", "m.01ghr4" ], "InferentialChain": [ "people.person.spouse_s..p...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: In the Bible, who was the servant of Sarah who became the concubine of Abraham and mother of Ishmael?
Spouse
FreebaseQA-train-298
Which alliteratively-named politician served as Abraham Lincoln's Vice-President between 1861 and 1865?
which alliteratively-named politician served as abraham lincoln's vice-president between 1861 and 1865
{ "Parse-Id": [ "FreebaseQA-train-298.P0" ], "PotentialTopicEntityMention": [ "abraham lincoln" ], "TopicEntityName": [ "abraham lincoln" ], "TopicEntityMid": [ "m.0gzh" ], "InferentialChain": [ "government.us_president.vice_president" ], "Answers": [ { "AnswersMid": ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Which alliteratively-named politician served as Abraham Lincoln's Vice-President between 1861 and 1865?
Vice president
FreebaseQA-train-299
Who immediately succeeded Abraham Lincoln as President?
who immediately succeeded abraham lincoln as president
{ "Parse-Id": [ "FreebaseQA-train-299.P0" ], "PotentialTopicEntityMention": [ "abraham lincoln" ], "TopicEntityName": [ "abraham lincoln" ], "TopicEntityMid": [ "m.0gzh" ], "InferentialChain": [ "government.us_president.vice_president" ], "Answers": [ { "AnswersMid": ...
inference_chain_prompt
What category best describes the answer to the following question? Question: {{RawQuestion}} ||| {% set answer = Parses.InferentialChain | first %} {{ answer.split(".") | last | capitalize | replace("_", " ")}}
What category best describes the answer to the following question? Question: Who immediately succeeded Abraham Lincoln as President?
Vice president