Dataset Viewer
Auto-converted to Parquet Duplicate
tokens
sequencelengths
7
500
labels
sequencelengths
9
502
mentions
listlengths
0
60
input_ids
sequencelengths
9
502
attention_mask
sequencelengths
9
502
[ "artificial", "intelligence", "(", "ai", ")", ",", "in", "its", "broad", "##est", "sense", ",", "is", "intelligence", "exhibited", "by", "machines", ",", "particularly", "computer", "systems", ".", "it", "is", "a", "field", "of", "research", "in", "computer"...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 1, 0, 0, 1, 2, 0, 0, 0, 0, 1, 2, 2, 0, 1, 2, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 1, 2, 2, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 13, "end_token_idx": 14, "identifier": "/wiki/Intelligence", "tokens": [ "intelligence" ] }, { "start_token_idx": 16, "end_token_idx": 17, "identifier": "/wiki/Machine", "tokens": [ "machines" ] }, { "start_token_idx": 19, "end...
[ 101, 7976, 4454, 1006, 9932, 1007, 1010, 1999, 2049, 5041, 4355, 3168, 1010, 2003, 4454, 8176, 2011, 6681, 1010, 3391, 3274, 3001, 1012, 2009, 2003, 1037, 2492, 1997, 2470, 1999, 3274, 2671, 2008, 11791, 1998, 2913, 4725, 1998, 4007, 2008...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "least", "equal", "level", "—", "is", "among", "the", "field", "'", "s", "long", "-", "term", "goals", ".", "to", "reach", "these", "goals", ",", "ai", "researchers", "have", "adapted", "and", "integrated", "a", "wide", "range", "of", "techniques", ",", ...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 1, 2, 0, 1, 2, 0, 1, 2, 2, 0, 0, 0, 0, 0, 1, 0, 1, 2, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 33, "end_token_idx": 34, "identifier": "/wiki/State_space_search", "tokens": [ "search" ] }, { "start_token_idx": 35, "end_token_idx": 37, "identifier": "/wiki/Mathematical_optimization", "tokens": [ "mathematical", "optimization" ] ...
[ 101, 2560, 5020, 2504, 1517, 2003, 2426, 1996, 2492, 1005, 1055, 2146, 1011, 2744, 3289, 1012, 2000, 3362, 2122, 3289, 1010, 9932, 6950, 2031, 5967, 1998, 6377, 1037, 2898, 2846, 1997, 5461, 1010, 2164, 3945, 1998, 8045, 20600, 1010, 5337...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "facts", "\"", "or", "\"", "statements", "\"", "that", "they", "could", "express", "verbal", "##ly", ")", ".", "there", "is", "also", "the", "difficulty", "of", "knowledge", "acquisition", ",", "the", "problem", "of", "obtaining", "knowledge", "for", "ai", ...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 20, "end_token_idx": 22, "identifier": "/wiki/Knowledge_acquisition", "tokens": [ "knowledge", "acquisition" ] }, { "start_token_idx": 52, "end_token_idx": 54, "identifier": "/wiki/Rational_agent", "tokens": [ "rational", "agent" ...
[ 101, 8866, 1000, 2030, 1000, 8635, 1000, 2008, 2027, 2071, 4671, 12064, 2135, 1007, 1012, 2045, 2003, 2036, 1996, 7669, 1997, 3716, 7654, 1010, 1996, 3291, 1997, 11381, 3716, 2005, 9932, 5097, 1012, 1031, 1039, 1033, 2019, 1000, 4005, 100...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "programs", "that", "can", "improve", "their", "performance", "on", "a", "given", "task", "automatically", ".", "it", "has", "been", "a", "part", "of", "ai", "from", "the", "beginning", ".", "[", "e", "]", "there", "are", "several", "kinds", "of", "machi...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 2, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 0, 0, 0, 0...
[ { "start_token_idx": 34, "end_token_idx": 40, "identifier": "/wiki/Unsupervised_learning", "tokens": [ "un", "##su", "##per", "##vis", "##ed", "learning" ] }, { "start_token_idx": 57, "end_token_idx": 59, "identifier": "/wiki/Supervised_learnin...
[ 101, 3454, 2008, 2064, 5335, 2037, 2836, 2006, 1037, 2445, 4708, 8073, 1012, 2009, 2038, 2042, 1037, 2112, 1997, 9932, 2013, 1996, 2927, 1012, 1031, 1041, 1033, 2045, 2024, 2195, 7957, 1997, 3698, 4083, 1012, 4895, 6342, 4842, 11365, 2098...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "input", ".", "the", "field", "includes", "speech", "recognition", ",", "image", "classification", ",", "facial", "recognition", ",", "object", "recognition", ",", "object", "tracking", ",", "and", "robotic", "perception", ".", "affect", "##ive", "computing", "i...
[ 0, 0, 0, 0, 0, 0, 1, 2, 0, 1, 2, 0, 1, 2, 0, 1, 2, 0, 1, 2, 0, 0, 1, 2, 0, 1, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 2, 2, 2, 0, 0, 0, 0, 0, 1, 2, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 5, "end_token_idx": 7, "identifier": "/wiki/Speech_recognition", "tokens": [ "speech", "recognition" ] }, { "start_token_idx": 8, "end_token_idx": 10, "identifier": "/wiki/Image_classification", "tokens": [ "image", "classification...
[ 101, 7953, 1012, 1996, 2492, 2950, 4613, 5038, 1010, 3746, 5579, 1010, 13268, 5038, 1010, 4874, 5038, 1010, 4874, 9651, 1010, 1998, 20478, 10617, 1012, 7461, 3512, 9798, 2003, 2019, 18593, 12977, 2008, 8681, 3001, 2008, 6807, 1010, 17841, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "survive", "each", "generation", ".", "distributed", "search", "processes", "can", "coordinate", "via", "swarm", "intelligence", "algorithms", ".", "two", "popular", "swarm", "algorithms", "used", "in", "search", "are", "particle", "swarm", "optimization", "(", "i...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 0, 0, 0, 0, 1, 2, 0, 0, 1, 2, 2, 0, 0, 0, 1, 2, 0, 0, 0, 1, 0, 0, 0, 1, 0, 1, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2...
[ { "start_token_idx": 10, "end_token_idx": 12, "identifier": "/wiki/Swarm_intelligence", "tokens": [ "swarm", "intelligence" ] }, { "start_token_idx": 22, "end_token_idx": 25, "identifier": "/wiki/Particle_swarm_optimization", "tokens": [ "particle", "s...
[ 101, 5788, 2169, 4245, 1012, 5500, 3945, 6194, 2064, 13530, 3081, 21708, 4454, 13792, 1012, 2048, 2759, 21708, 13792, 2109, 1999, 3945, 2024, 10811, 21708, 20600, 1006, 4427, 2011, 4743, 19311, 2075, 1007, 1998, 14405, 5701, 20600, 1006, 44...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "have", "been", "developed", "that", "analyze", "how", "an", "agent", "can", "make", "choices", "and", "plan", ",", "using", "decision", "theory", ",", "decision", "analysis", ",", "and", "information", "value", "theory", ".", "these", "tools", "include", "m...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 0, 1, 2, 0, 0, 1, 2, 2, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 2, 0, 0, 1, 2, 0, 1, 2, 0, 1, 2, 0, 1, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1...
[ { "start_token_idx": 15, "end_token_idx": 17, "identifier": "/wiki/Decision_theory", "tokens": [ "decision", "theory" ] }, { "start_token_idx": 18, "end_token_idx": 20, "identifier": "/wiki/Decision_analysis", "tokens": [ "decision", "analysis" ] ...
[ 101, 2031, 2042, 2764, 2008, 17908, 2129, 2019, 4005, 2064, 2191, 9804, 1998, 2933, 1010, 2478, 3247, 3399, 1010, 3247, 4106, 1010, 1998, 2592, 3643, 3399, 1012, 2122, 5906, 2421, 4275, 2107, 2004, 28003, 2615, 3247, 6194, 1010, 8790, 324...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "weights", "that", "will", "get", "the", "right", "output", "for", "each", "input", "during", "training", ".", "the", "most", "common", "training", "technique", "is", "the", "back", "##pro", "##pa", "##gation", "algorithm", ".", "neural", "networks", "learn", ...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 2, 2, 0, 0, 0...
[ { "start_token_idx": 20, "end_token_idx": 24, "identifier": "/wiki/Backpropagation", "tokens": [ "back", "##pro", "##pa", "##gation" ] }, { "start_token_idx": 38, "end_token_idx": 40, "identifier": "/wiki/Pattern_recognition", "tokens": [ "find",...
[ 101, 15871, 2008, 2097, 2131, 1996, 2157, 6434, 2005, 2169, 7953, 2076, 2731, 1012, 1996, 2087, 2691, 2731, 6028, 2003, 1996, 2067, 21572, 4502, 12540, 9896, 1012, 15756, 6125, 4553, 2000, 2944, 3375, 6550, 2090, 20407, 1998, 27852, 1998, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "like", "text", "by", "repeatedly", "predicting", "the", "next", "token", ".", "typically", ",", "a", "subsequent", "training", "phase", "makes", "the", "model", "more", "truth", "##ful", ",", "useful", "and", "harmless", ",", "usually", "with", "a", "techni...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 0, 0, 0...
[ { "start_token_idx": 31, "end_token_idx": 36, "identifier": "/wiki/Reinforcement_learning_from_human_feedback", "tokens": [ "reinforcement", "learning", "from", "human", "feedback" ] }, { "start_token_idx": 57, "end_token_idx": 60, "identifier": "/wi...
[ 101, 2066, 3793, 2011, 8385, 29458, 1996, 2279, 19204, 1012, 4050, 1010, 1037, 4745, 2731, 4403, 3084, 1996, 2944, 2062, 3606, 3993, 1010, 6179, 1998, 19741, 1010, 2788, 2007, 1037, 6028, 2170, 23895, 4083, 2013, 2529, 12247, 1006, 1054, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ ".", "for", "example", ",", "alpha", "##fold", "2", "(", "2021", ")", "demonstrated", "the", "ability", "to", "approximate", ",", "in", "hours", "rather", "than", "months", ",", "the", "3d", "structure", "of", "a", "protein", ".", "in", "202", "##3", "...
[ 0, 0, 0, 0, 0, 1, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 4, "end_token_idx": 7, "identifier": "/wiki/AlphaFold_2", "tokens": [ "alpha", "##fold", "2" ] }, { "start_token_idx": 24, "end_token_idx": 28, "identifier": "/wiki/Protein_structure", "tokens": [ "structure", "of", "a"...
[ 101, 1012, 2005, 2742, 1010, 6541, 10371, 1016, 1006, 25682, 1007, 7645, 1996, 3754, 2000, 15796, 1010, 1999, 2847, 2738, 2084, 2706, 1010, 1996, 7605, 3252, 1997, 1037, 5250, 1012, 1999, 16798, 2509, 1010, 2009, 2001, 2988, 2008, 9932, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ ",", "financial", "planning", ",", "and", "pension", "advice", "in", "the", "process", ",", "but", "i", "’", "m", "not", "sure", "it", "will", "un", "##lea", "##sh", "a", "new", "wave", "of", "[", "e", ".", "g", ".", ",", "sophisticated", "]", "pen...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 51, "end_token_idx": 54, "identifier": "/wiki/Command_and_control", "tokens": [ "command", "and", "control" ] }, { "start_token_idx": 88, "end_token_idx": 90, "identifier": "/wiki/Vehicular_automation", "tokens": [ "autonomous", ...
[ 101, 1010, 3361, 4041, 1010, 1998, 11550, 6040, 1999, 1996, 2832, 1010, 2021, 1045, 1521, 1049, 2025, 2469, 2009, 2097, 4895, 19738, 4095, 1037, 2047, 4400, 1997, 1031, 1041, 1012, 1043, 1012, 1010, 12138, 1033, 11550, 8144, 1012, 1000, 2...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "data", "and", "applications", ",", "mainly", "for", "\"", "classification", ",", "regression", ",", "cluster", "##ing", ",", "forecast", "##ing", ",", "generation", ",", "discovery", ",", "and", "the", "development", "of", "new", "scientific", "insights", "\"...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 116, "end_token_idx": 121, "identifier": "/wiki/Demis_Hassabis", "tokens": [ "demi", "##s", "has", "##sa", "##bis" ] }, { "start_token_idx": 122, "end_token_idx": 124, "identifier": "/wiki/DeepMind", "tokens": [ "deep",...
[ 101, 2951, 1998, 5097, 1010, 3701, 2005, 1000, 5579, 1010, 26237, 1010, 9324, 2075, 1010, 19939, 2075, 1010, 4245, 1010, 5456, 1010, 1998, 1996, 2458, 1997, 2047, 4045, 20062, 1000, 2005, 2742, 2005, 13648, 4654, 7361, 20644, 3215, 1010, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "copyright", "##ed", "work", "\"", ".", "website", "owners", "who", "do", "not", "wish", "to", "have", "their", "content", "scraped", "can", "indicate", "it", "in", "a", "\"", "robots", ".", "tx", "##t", "\"", "file", ".", "in", "202", "##3", ",", "l...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 2, 0, 1, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 22, "end_token_idx": 26, "identifier": "/wiki/Robots.txt", "tokens": [ "robots", ".", "tx", "##t" ] }, { "start_token_idx": 37, "end_token_idx": 41, "identifier": "/wiki/John_Grisham", "tokens": [ "john", "gr", "#...
[ 101, 9385, 2098, 2147, 1000, 1012, 4037, 5608, 2040, 2079, 2025, 4299, 2000, 2031, 2037, 4180, 20378, 2064, 5769, 2009, 1999, 1037, 1000, 13507, 1012, 19067, 2102, 1000, 5371, 1012, 1999, 16798, 2509, 1010, 2877, 6048, 1006, 2164, 2198, 2...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "labeling", "feature", "mistakenly", "identified", "jack", "##y", "al", "##cine", "and", "a", "friend", "as", "\"", "gorilla", "##s", "\"", "because", "they", "were", "black", ".", "the", "system", "was", "trained", "on", "a", "data", "##set", "that", "con...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 100, "end_token_idx": 102, "identifier": "/wiki/COMPAS_(software)", "tokens": [ "com", "##pas" ] }, { "start_token_idx": 109, "end_token_idx": 114, "identifier": "/wiki/U.S._court", "tokens": [ "u", ".", "s", ".", ...
[ 101, 28847, 3444, 20706, 4453, 2990, 2100, 2632, 16567, 1998, 1037, 2767, 2004, 1000, 23526, 2015, 1000, 2138, 2027, 2020, 2304, 1012, 1996, 2291, 2001, 4738, 2006, 1037, 2951, 13462, 2008, 4838, 2200, 2261, 4871, 1997, 2304, 2111, 1010, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "##ete", "##cted", "because", "the", "developers", "are", "overwhelmingly", "white", "and", "male", ":", "among", "ai", "engineers", ",", "about", "4", "%", "are", "black", "and", "20", "%", "are", "women", ".", "at", "its", "202", "##2", "conference", "...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 2, 2, 2, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 30, "end_token_idx": 38, "identifier": "/wiki/ACM_Conference_on_Fairness,_Accountability,_and_Transparency", "tokens": [ "conference", "on", "fairness", ",", "accountability", ",", "and", "transparency" ] }, { "start_to...
[ 101, 12870, 10985, 2138, 1996, 9797, 2024, 24783, 2317, 1998, 3287, 1024, 2426, 9932, 6145, 1010, 2055, 1018, 1003, 2024, 2304, 1998, 2322, 1003, 2024, 2308, 1012, 2012, 2049, 16798, 2475, 3034, 2006, 26935, 1010, 17842, 1010, 1998, 16987, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "lime", "can", "locally", "approximate", "a", "model", "with", "a", "simpler", ",", "interpret", "##able", "model", ".", "multi", "##tas", "##k", "learning", "provides", "a", "large", "number", "of", "outputs", "in", "addition", "to", "the", "target", "class...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 2, 2, 0, 1, 2, 2, 0, 0, 1, 2, 0, 0, 0, 0...
[ { "start_token_idx": 14, "end_token_idx": 18, "identifier": "/wiki/Multitask_learning", "tokens": [ "multi", "##tas", "##k", "learning" ] }, { "start_token_idx": 46, "end_token_idx": 51, "identifier": "/wiki/Deconvolution", "tokens": [ "deco", ...
[ 101, 14123, 2064, 7246, 15796, 1037, 2944, 2007, 1037, 16325, 1010, 17841, 3085, 2944, 1012, 4800, 10230, 2243, 4083, 3640, 1037, 2312, 2193, 1997, 27852, 1999, 2804, 2000, 1996, 4539, 5579, 1012, 2122, 2060, 27852, 2064, 2393, 9797, 2139, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "in", "the", "past", ",", "technology", "has", "tended", "to", "increase", "rather", "than", "reduce", "total", "employment", ",", "but", "economists", "acknowledge", "that", "\"", "we", "'", "re", "in", "un", "##cha", "##rted", "territory", "\"", "with", ...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 56, "end_token_idx": 57, "identifier": "/wiki/Unemployment", "tokens": [ "unemployment" ] }, { "start_token_idx": 70, "end_token_idx": 71, "identifier": "/wiki/Productivity", "tokens": [ "productivity" ] }, { "start_token_idx": 73,...
[ 101, 1999, 1996, 2627, 1010, 2974, 2038, 11121, 2000, 3623, 2738, 2084, 5547, 2561, 6107, 1010, 2021, 22171, 13399, 2008, 1000, 2057, 1005, 2128, 1999, 4895, 7507, 17724, 3700, 1000, 2007, 9932, 1012, 1037, 5002, 1997, 22171, 3662, 18185, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
End of preview. Expand in Data Studio

Dataset Card for "mention_detection_bert_dataset"

More Information needed

Downloads last month
4