Dataset Viewer
Auto-converted to Parquet Duplicate
tokens
sequencelengths
17
502
labels
sequencelengths
17
502
mentions
listlengths
0
59
input_ids
sequencelengths
17
502
attention_mask
sequencelengths
17
502
[ "[CLS]", "Art", "##ific", "##ial", "intelligence", "(", "AI", ")", ",", "in", "its", "broad", "##est", "sense", ",", "is", "intelligence", "exhibited", "by", "machines", ",", "particularly", "computer", "systems", ".", "It", "is", "a", "field", "of", "rese...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 1, 0, 0, 1, 2, 0, 0, 0, 0, 1, 2, 2, 0, 1, 2, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 1, 2, 2, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 16, "end_token_idx": 17, "identifier": "/wiki/Intelligence", "tokens": [ "intelligence" ] }, { "start_token_idx": 19, "end_token_idx": 20, "identifier": "/wiki/Machine", "tokens": [ "machines" ] }, { "start_token_idx": 22, "end...
[ 101, 2051, 19814, 2916, 4810, 113, 19016, 114, 117, 1107, 1157, 4728, 2556, 2305, 117, 1110, 4810, 7799, 1118, 6555, 117, 2521, 2775, 2344, 119, 1135, 1110, 170, 1768, 1104, 1844, 1107, 2775, 2598, 1115, 11926, 1105, 2527, 4069, 1105, 3...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "[CLS]", "General", "intelligence", "—", "the", "ability", "to", "complete", "any", "task", "perform", "##able", "by", "a", "human", "on", "an", "at", "least", "equal", "level", "—", "is", "among", "the", "field", "'", "s", "long", "-", "term", "goals", ...
[ 0, 1, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 1, 2, 0, 1, 2, 0, 1, 2, 2, 0, 0...
[ { "start_token_idx": 1, "end_token_idx": 3, "identifier": "/wiki/Artificial_general_intelligence", "tokens": [ "General", "intelligence" ] }, { "start_token_idx": 51, "end_token_idx": 52, "identifier": "/wiki/State_space_search", "tokens": [ "search" ] ...
[ 101, 1615, 4810, 783, 1103, 2912, 1106, 2335, 1251, 4579, 3870, 1895, 1118, 170, 1769, 1113, 1126, 1120, 1655, 4463, 1634, 783, 1110, 1621, 1103, 1768, 112, 188, 1263, 118, 1858, 2513, 119, 1706, 2519, 1292, 2513, 117, 19016, 6962, 1138...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "[CLS]", "of", "atomic", "facts", "that", "the", "average", "person", "knows", "is", "enormous", ")", ";", "and", "the", "sub", "-", "symbolic", "form", "of", "most", "common", "##sen", "##se", "knowledge", "(", "much", "of", "what", "people", "know", "i...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 56, "end_token_idx": 58, "identifier": "/wiki/Knowledge_acquisition", "tokens": [ "knowledge", "acquisition" ] }, { "start_token_idx": 88, "end_token_idx": 90, "identifier": "/wiki/Rational_agent", "tokens": [ "rational", "agent" ...
[ 101, 1104, 12861, 9193, 1115, 1103, 1903, 1825, 3520, 1110, 7883, 114, 132, 1105, 1103, 4841, 118, 13516, 1532, 1104, 1211, 1887, 3792, 2217, 3044, 113, 1277, 1104, 1184, 1234, 1221, 1110, 1136, 2533, 1112, 107, 9193, 107, 1137, 107, 84...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "[CLS]", ",", "or", "it", "can", "be", "learned", ".", "Game", "theory", "describes", "the", "rational", "behavior", "of", "multiple", "interacting", "agents", "and", "is", "used", "in", "AI", "programs", "that", "make", "decisions", "that", "involve", "othe...
[ 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 8, "end_token_idx": 10, "identifier": "/wiki/Game_theory", "tokens": [ "Game", "theory" ] }, { "start_token_idx": 32, "end_token_idx": 34, "identifier": "/wiki/Machine_learning", "tokens": [ "Machine", "learning" ] }, { ...
[ 101, 117, 1137, 1122, 1169, 1129, 3560, 119, 3497, 2749, 4856, 1103, 12478, 4658, 1104, 2967, 24775, 5789, 1105, 1110, 1215, 1107, 19016, 2648, 1115, 1294, 6134, 1115, 8803, 1168, 5789, 119, 7792, 3776, 1110, 1103, 2025, 1104, 2648, 1115,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "[CLS]", "perception", "is", "the", "ability", "to", "use", "input", "from", "sensors", "(", "such", "as", "cameras", ",", "microphone", "##s", ",", "wireless", "signals", ",", "active", "lid", "##ar", ",", "son", "##ar", ",", "radar", ",", "and", "ta", ...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 0, 1, 2, 0...
[ { "start_token_idx": 22, "end_token_idx": 24, "identifier": "/wiki/Lidar", "tokens": [ "lid", "##ar" ] }, { "start_token_idx": 31, "end_token_idx": 35, "identifier": "/wiki/Tactile_sensor", "tokens": [ "ta", "##ct", "##ile", "sensors" ]...
[ 101, 11170, 1110, 1103, 2912, 1106, 1329, 7758, 1121, 15145, 113, 1216, 1112, 8457, 117, 16976, 1116, 117, 12784, 7981, 117, 2327, 14753, 1813, 117, 1488, 1813, 117, 7746, 117, 1105, 27629, 5822, 4759, 15145, 114, 1106, 1260, 7641, 2093, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "[CLS]", "##ly", "adjusting", "them", "to", "minimize", "a", "loss", "function", ".", "V", "##arian", "##ts", "of", "gradient", "descent", "are", "commonly", "used", "to", "train", "neural", "networks", ".", "Another", "type", "of", "local", "search", "is", ...
[ 0, 0, 0, 0, 0, 0, 0, 1, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0...
[ { "start_token_idx": 7, "end_token_idx": 9, "identifier": "/wiki/Loss_function", "tokens": [ "loss", "function" ] }, { "start_token_idx": 30, "end_token_idx": 33, "identifier": "/wiki/Evolutionary_computation", "tokens": [ "evolutionary", "com", ...
[ 101, 1193, 21763, 1172, 1106, 20220, 170, 2445, 3053, 119, 159, 7968, 2145, 1104, 19848, 6585, 1132, 3337, 1215, 1106, 2669, 18250, 6379, 119, 2543, 2076, 1104, 1469, 3403, 1110, 14745, 3254, 19675, 117, 1134, 8469, 1106, 1122, 21126, 119...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "[CLS]", "true", ".", "Non", "-", "mon", "##oto", "##nic", "logic", "##s", ",", "including", "logic", "programming", "with", "ne", "##gation", "as", "failure", ",", "are", "designed", "to", "handle", "default", "reasoning", ".", "Other", "specialized", "vers...
[ 0, 0, 0, 1, 2, 2, 2, 2, 2, 2, 0, 0, 0, 0, 0, 1, 2, 2, 2, 0, 0, 0, 0, 0, 1, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 3, "end_token_idx": 10, "identifier": "/wiki/Non-monotonic_logic", "tokens": [ "Non", "-", "mon", "##oto", "##nic", "logic", "##s" ] }, { "start_token_idx": 15, "end_token_idx": 19, "identifier": "/wiki/Negation_as_fa...
[ 101, 2276, 119, 7922, 118, 19863, 12355, 7770, 8738, 1116, 117, 1259, 8738, 4159, 1114, 24928, 10716, 1112, 4290, 117, 1132, 2011, 1106, 4282, 12973, 14417, 119, 2189, 7623, 3827, 1104, 8738, 1138, 1151, 1872, 1106, 5594, 1242, 2703, 1377...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "[CLS]", "based", "on", "a", "collection", "of", "nodes", "also", "known", "as", "artificial", "neurons", ",", "which", "loosely", "model", "the", "neurons", "in", "a", "biological", "brain", ".", "It", "is", "trained", "to", "recognise", "patterns", ";", ...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 10, "end_token_idx": 12, "identifier": "/wiki/Artificial_neurons", "tokens": [ "artificial", "neurons" ] }, { "start_token_idx": 17, "end_token_idx": 18, "identifier": "/wiki/Neurons", "tokens": [ "neurons" ] }, { "start_toke...
[ 101, 1359, 1113, 170, 2436, 1104, 15029, 1145, 1227, 1112, 8246, 16993, 117, 1134, 12158, 2235, 1103, 16993, 1107, 170, 7269, 3575, 119, 1135, 1110, 3972, 1106, 18986, 6692, 132, 1517, 3972, 117, 1122, 1169, 18986, 1343, 6692, 1107, 4489,...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "[CLS]", "used", "for", "bench", "##mark", "testing", ",", "such", "as", "Image", "##Net", ".", "[", "j", "]", "Gene", "##rative", "pre", "-", "trained", "transform", "##ers", "(", "GP", "##T", ")", "are", "large", "language", "models", "that", "are", ...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 0, 0, 0, 0, 1, 2, 2, 2, 2, 2, 2, 0, 0, 0, 0, 0, 1, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 9, "end_token_idx": 11, "identifier": "/wiki/ImageNet", "tokens": [ "Image", "##Net" ] }, { "start_token_idx": 15, "end_token_idx": 22, "identifier": "/wiki/Generative_pre-trained_transformer", "tokens": [ "Gene", "##rative", ...
[ 101, 1215, 1111, 6757, 8519, 5193, 117, 1216, 1112, 15065, 25264, 119, 164, 179, 166, 9066, 15306, 3073, 118, 3972, 11303, 1468, 113, 15175, 1942, 114, 1132, 1415, 1846, 3584, 1115, 1132, 1359, 1113, 1103, 27921, 6085, 1206, 1734, 1107, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "[CLS]", "Facebook", ",", "Apple", "'", "s", "i", "##P", "##hot", "##o", "and", "T", "##ik", "##T", "##ok", ")", ".", "The", "application", "of", "AI", "in", "medicine", "and", "medical", "research", "has", "the", "potential", "to", "increase", "patient"...
[ 0, 1, 0, 0, 0, 0, 1, 2, 2, 2, 0, 1, 2, 2, 2, 0, 0, 0, 0, 0, 0, 0, 1, 0, 1, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 1, "end_token_idx": 2, "identifier": "/wiki/Facebook", "tokens": [ "Facebook" ] }, { "start_token_idx": 6, "end_token_idx": 10, "identifier": "/wiki/IPhoto", "tokens": [ "i", "##P", "##hot", "##o" ] }, { "start_to...
[ 101, 9078, 117, 7302, 112, 188, 178, 2101, 12217, 1186, 1105, 157, 4847, 1942, 5926, 114, 119, 1109, 4048, 1104, 19016, 1107, 5182, 1105, 2657, 1844, 1144, 1103, 3209, 1106, 2773, 5351, 1920, 1105, 3068, 1104, 1297, 119, 4737, 1103, 110...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "[CLS]", ",", "or", "Atari", "games", ".", "In", "2019", ",", "Deep", "##M", "##ind", "'", "s", "Alpha", "##S", "##tar", "achieved", "grand", "##master", "level", "in", "Star", "##C", "##raft", "II", ",", "a", "particularly", "challenging", "real", "-", ...
[ 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 3, "end_token_idx": 4, "identifier": "/wiki/Atari", "tokens": [ "Atari" ] }, { "start_token_idx": 22, "end_token_idx": 26, "identifier": "/wiki/StarCraft_II", "tokens": [ "Star", "##C", "##raft", "II" ] }, { "star...
[ 101, 117, 1137, 20694, 1638, 119, 1130, 10351, 117, 7786, 2107, 24704, 112, 188, 8461, 1708, 6817, 3890, 5372, 6532, 1634, 1107, 2537, 1658, 15371, 1563, 117, 170, 2521, 10467, 1842, 118, 1159, 5564, 1342, 1115, 6808, 13975, 3044, 1104, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "[CLS]", "a", "ho", "##ax", "of", "an", "attack", "on", "the", "Pentagon", ",", "as", "well", "as", "the", "usage", "in", "professional", "creative", "arts", ".", "There", "are", "also", "thousands", "of", "successful", "AI", "applications", "used", "to", ...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 9, "end_token_idx": 10, "identifier": "/wiki/The_Pentagon", "tokens": [ "Pentagon" ] }, { "start_token_idx": 65, "end_token_idx": 67, "identifier": "/wiki/Energy_storage", "tokens": [ "energy", "storage" ] }, { "start_token_i...
[ 101, 170, 16358, 7897, 1104, 1126, 2035, 1113, 1103, 25122, 117, 1112, 1218, 1112, 1103, 7991, 1107, 1848, 6228, 3959, 119, 1247, 1132, 1145, 4674, 1104, 2265, 19016, 4683, 1215, 1106, 9474, 2747, 2645, 1111, 2747, 7519, 1137, 4300, 119, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "[CLS]", "clearly", "une", "##thical", "and", "a", "violation", "of", "the", "right", "to", "privacy", ".", "AI", "developers", "argue", "that", "this", "is", "the", "only", "way", "to", "deliver", "valuable", "applications", ".", "and", "have", "developed",...
[ 0, 0, 1, 2, 0, 0, 0, 0, 0, 1, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 2, 0, 1, 2, 2, 0, 1, 2, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 2, "end_token_idx": 4, "identifier": "/wiki/Unethical", "tokens": [ "une", "##thical" ] }, { "start_token_idx": 9, "end_token_idx": 12, "identifier": "/wiki/Right_to_privacy", "tokens": [ "right", "to", "privacy" ] }, ...
[ 101, 3817, 25731, 26610, 1105, 170, 11574, 1104, 1103, 1268, 1106, 9909, 119, 19016, 10300, 6982, 1115, 1142, 1110, 1103, 1178, 1236, 1106, 7852, 7468, 4683, 119, 1105, 1138, 1872, 1317, 4884, 1115, 2661, 1106, 8333, 9909, 1229, 1253, 116...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "[CLS]", "real", "photographs", ",", "recordings", ",", "films", ",", "or", "human", "writing", ".", "It", "is", "possible", "for", "bad", "actors", "to", "use", "this", "technology", "to", "create", "massive", "amounts", "of", "mi", "##sin", "##formation", ...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 35, "end_token_idx": 38, "identifier": "/wiki/Geoffrey_Hinton", "tokens": [ "Geoffrey", "Hi", "##nton" ] }, { "start_token_idx": 96, "end_token_idx": 98, "identifier": "/wiki/Training_data", "tokens": [ "training", "data" ...
[ 101, 1842, 6810, 117, 5982, 117, 2441, 117, 1137, 1769, 2269, 119, 1135, 1110, 1936, 1111, 2213, 5681, 1106, 1329, 1142, 2815, 1106, 2561, 4672, 7919, 1104, 1940, 10606, 24152, 1137, 11516, 119, 19016, 8578, 10985, 8790, 13124, 4448, 4517...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "[CLS]", "does", "not", "explicitly", "mention", "a", "problematic", "feature", "(", "such", "as", "\"", "race", "\"", "or", "\"", "gender", "\"", ")", ".", "The", "feature", "will", "co", "##rrel", "##ate", "with", "other", "features", "(", "like", "\""...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 264, "end_token_idx": 278, "identifier": "/wiki/ACM_Conference_on_Fairness,_Accountability,_and_Transparency", "tokens": [ "Conference", "on", "Fair", "##ness", ",", "A", "##cco", "##unta", "##bility", ",", "and",...
[ 101, 1674, 1136, 12252, 4734, 170, 20405, 2672, 113, 1216, 1112, 107, 1886, 107, 1137, 107, 5772, 107, 114, 119, 1109, 2672, 1209, 1884, 18337, 2193, 1114, 1168, 1956, 113, 1176, 107, 4134, 107, 117, 107, 6001, 1607, 107, 1137, 107, 1...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "[CLS]", "to", "show", "the", "scale", ".", "Another", "machine", "learning", "system", "designed", "to", "help", "effectively", "all", "##oc", "##ate", "medical", "resources", "was", "found", "to", "class", "##ify", "patients", "with", "as", "##th", "##ma", ...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 150, "end_token_idx": 154, "identifier": "/wiki/General_Data_Protection_Regulation", "tokens": [ "General", "Data", "Protection", "Regulation" ] }, { "start_token_idx": 210, "end_token_idx": 213, "identifier": "/wiki/DARPA", "token...
[ 101, 1106, 1437, 1103, 3418, 119, 2543, 3395, 3776, 1449, 2011, 1106, 1494, 5877, 1155, 13335, 2193, 2657, 3979, 1108, 1276, 1106, 1705, 6120, 4420, 1114, 1112, 1582, 1918, 1112, 1217, 1120, 107, 1822, 3187, 107, 1104, 5694, 1121, 20673, ...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
[ "[CLS]", "United", "States", "and", "others", "disagreed", ".", "By", "2015", ",", "over", "fifty", "countries", "were", "reported", "to", "be", "researching", "battlefield", "robots", ".", "AI", "tools", "make", "it", "easier", "for", "author", "##itarian", ...
[ 0, 1, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 2, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 1, 2, 0, 0, 1, 0, 1, 2, 0, 0, 0, 0, 0, 0, 1, 2, 0, 0, 0, 0, 0, 0, 0...
[ { "start_token_idx": 1, "end_token_idx": 3, "identifier": "/wiki/United_States", "tokens": [ "United", "States" ] }, { "start_token_idx": 27, "end_token_idx": 30, "identifier": "/wiki/Authoritarian", "tokens": [ "author", "##itarian", "government...
[ 101, 1244, 1311, 1105, 1639, 19786, 119, 1650, 1410, 117, 1166, 5547, 2182, 1127, 2103, 1106, 1129, 22513, 13777, 16013, 119, 19016, 5537, 1294, 1122, 5477, 1111, 2351, 12403, 6670, 1106, 19723, 1654, 1147, 4037, 1107, 1317, 3242, 119, 10...
[ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1...
End of preview. Expand in Data Studio

Dataset Card for "mention-detection-bert-large-cased"

More Information needed

Downloads last month
2