tokens sequencelengths 7 500 | labels sequencelengths 9 502 | mentions listlengths 0 60 | input_ids sequencelengths 9 502 | attention_mask sequencelengths 9 502 |
|---|---|---|---|---|
[
"artificial",
"intelligence",
"(",
"ai",
")",
",",
"in",
"its",
"broad",
"##est",
"sense",
",",
"is",
"intelligence",
"exhibited",
"by",
"machines",
",",
"particularly",
"computer",
"systems",
".",
"it",
"is",
"a",
"field",
"of",
"research",
"in",
"computer"... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
1,
0,
0,
1,
2,
0,
0,
0,
0,
1,
2,
2,
0,
1,
2,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
1,
2,
2,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
{
"start_token_idx": 13,
"end_token_idx": 14,
"identifier": "/wiki/Intelligence",
"tokens": [
"intelligence"
]
},
{
"start_token_idx": 16,
"end_token_idx": 17,
"identifier": "/wiki/Machine",
"tokens": [
"machines"
]
},
{
"start_token_idx": 19,
"end... | [
101,
7976,
4454,
1006,
9932,
1007,
1010,
1999,
2049,
5041,
4355,
3168,
1010,
2003,
4454,
8176,
2011,
6681,
1010,
3391,
3274,
3001,
1012,
2009,
2003,
1037,
2492,
1997,
2470,
1999,
3274,
2671,
2008,
11791,
1998,
2913,
4725,
1998,
4007,
2008... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... |
[
"least",
"equal",
"level",
"—",
"is",
"among",
"the",
"field",
"'",
"s",
"long",
"-",
"term",
"goals",
".",
"to",
"reach",
"these",
"goals",
",",
"ai",
"researchers",
"have",
"adapted",
"and",
"integrated",
"a",
"wide",
"range",
"of",
"techniques",
",",
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
1,
2,
0,
1,
2,
0,
1,
2,
2,
0,
0,
0,
0,
0,
1,
0,
1,
2,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0... | [
{
"start_token_idx": 33,
"end_token_idx": 34,
"identifier": "/wiki/State_space_search",
"tokens": [
"search"
]
},
{
"start_token_idx": 35,
"end_token_idx": 37,
"identifier": "/wiki/Mathematical_optimization",
"tokens": [
"mathematical",
"optimization"
]
... | [
101,
2560,
5020,
2504,
1517,
2003,
2426,
1996,
2492,
1005,
1055,
2146,
1011,
2744,
3289,
1012,
2000,
3362,
2122,
3289,
1010,
9932,
6950,
2031,
5967,
1998,
6377,
1037,
2898,
2846,
1997,
5461,
1010,
2164,
3945,
1998,
8045,
20600,
1010,
5337... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... |
[
"facts",
"\"",
"or",
"\"",
"statements",
"\"",
"that",
"they",
"could",
"express",
"verbal",
"##ly",
")",
".",
"there",
"is",
"also",
"the",
"difficulty",
"of",
"knowledge",
"acquisition",
",",
"the",
"problem",
"of",
"obtaining",
"knowledge",
"for",
"ai",
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
{
"start_token_idx": 20,
"end_token_idx": 22,
"identifier": "/wiki/Knowledge_acquisition",
"tokens": [
"knowledge",
"acquisition"
]
},
{
"start_token_idx": 52,
"end_token_idx": 54,
"identifier": "/wiki/Rational_agent",
"tokens": [
"rational",
"agent"
... | [
101,
8866,
1000,
2030,
1000,
8635,
1000,
2008,
2027,
2071,
4671,
12064,
2135,
1007,
1012,
2045,
2003,
2036,
1996,
7669,
1997,
3716,
7654,
1010,
1996,
3291,
1997,
11381,
3716,
2005,
9932,
5097,
1012,
1031,
1039,
1033,
2019,
1000,
4005,
100... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... |
[
"programs",
"that",
"can",
"improve",
"their",
"performance",
"on",
"a",
"given",
"task",
"automatically",
".",
"it",
"has",
"been",
"a",
"part",
"of",
"ai",
"from",
"the",
"beginning",
".",
"[",
"e",
"]",
"there",
"are",
"several",
"kinds",
"of",
"machi... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
2,
2,
2,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
0,
0,
0,
0... | [
{
"start_token_idx": 34,
"end_token_idx": 40,
"identifier": "/wiki/Unsupervised_learning",
"tokens": [
"un",
"##su",
"##per",
"##vis",
"##ed",
"learning"
]
},
{
"start_token_idx": 57,
"end_token_idx": 59,
"identifier": "/wiki/Supervised_learnin... | [
101,
3454,
2008,
2064,
5335,
2037,
2836,
2006,
1037,
2445,
4708,
8073,
1012,
2009,
2038,
2042,
1037,
2112,
1997,
9932,
2013,
1996,
2927,
1012,
1031,
1041,
1033,
2045,
2024,
2195,
7957,
1997,
3698,
4083,
1012,
4895,
6342,
4842,
11365,
2098... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... |
[
"input",
".",
"the",
"field",
"includes",
"speech",
"recognition",
",",
"image",
"classification",
",",
"facial",
"recognition",
",",
"object",
"recognition",
",",
"object",
"tracking",
",",
"and",
"robotic",
"perception",
".",
"affect",
"##ive",
"computing",
"i... | [
0,
0,
0,
0,
0,
0,
1,
2,
0,
1,
2,
0,
1,
2,
0,
1,
2,
0,
1,
2,
0,
0,
1,
2,
0,
1,
2,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
2,
2,
2,
2,
0,
0,
0,
0,
0,
1,
2,
0,
0,
0,
0,
0,
0... | [
{
"start_token_idx": 5,
"end_token_idx": 7,
"identifier": "/wiki/Speech_recognition",
"tokens": [
"speech",
"recognition"
]
},
{
"start_token_idx": 8,
"end_token_idx": 10,
"identifier": "/wiki/Image_classification",
"tokens": [
"image",
"classification... | [
101,
7953,
1012,
1996,
2492,
2950,
4613,
5038,
1010,
3746,
5579,
1010,
13268,
5038,
1010,
4874,
5038,
1010,
4874,
9651,
1010,
1998,
20478,
10617,
1012,
7461,
3512,
9798,
2003,
2019,
18593,
12977,
2008,
8681,
3001,
2008,
6807,
1010,
17841,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... |
[
"survive",
"each",
"generation",
".",
"distributed",
"search",
"processes",
"can",
"coordinate",
"via",
"swarm",
"intelligence",
"algorithms",
".",
"two",
"popular",
"swarm",
"algorithms",
"used",
"in",
"search",
"are",
"particle",
"swarm",
"optimization",
"(",
"i... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
2,
0,
0,
0,
0,
1,
2,
0,
0,
1,
2,
2,
0,
0,
0,
1,
2,
0,
0,
0,
1,
0,
0,
0,
1,
0,
1,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2... | [
{
"start_token_idx": 10,
"end_token_idx": 12,
"identifier": "/wiki/Swarm_intelligence",
"tokens": [
"swarm",
"intelligence"
]
},
{
"start_token_idx": 22,
"end_token_idx": 25,
"identifier": "/wiki/Particle_swarm_optimization",
"tokens": [
"particle",
"s... | [
101,
5788,
2169,
4245,
1012,
5500,
3945,
6194,
2064,
13530,
3081,
21708,
4454,
13792,
1012,
2048,
2759,
21708,
13792,
2109,
1999,
3945,
2024,
10811,
21708,
20600,
1006,
4427,
2011,
4743,
19311,
2075,
1007,
1998,
14405,
5701,
20600,
1006,
44... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... |
[
"have",
"been",
"developed",
"that",
"analyze",
"how",
"an",
"agent",
"can",
"make",
"choices",
"and",
"plan",
",",
"using",
"decision",
"theory",
",",
"decision",
"analysis",
",",
"and",
"information",
"value",
"theory",
".",
"these",
"tools",
"include",
"m... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
0,
1,
2,
0,
0,
1,
2,
2,
0,
0,
0,
0,
0,
0,
0,
1,
2,
2,
2,
0,
0,
1,
2,
0,
1,
2,
0,
1,
2,
0,
1,
2,
2,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
1... | [
{
"start_token_idx": 15,
"end_token_idx": 17,
"identifier": "/wiki/Decision_theory",
"tokens": [
"decision",
"theory"
]
},
{
"start_token_idx": 18,
"end_token_idx": 20,
"identifier": "/wiki/Decision_analysis",
"tokens": [
"decision",
"analysis"
]
... | [
101,
2031,
2042,
2764,
2008,
17908,
2129,
2019,
4005,
2064,
2191,
9804,
1998,
2933,
1010,
2478,
3247,
3399,
1010,
3247,
4106,
1010,
1998,
2592,
3643,
3399,
1012,
2122,
5906,
2421,
4275,
2107,
2004,
28003,
2615,
3247,
6194,
1010,
8790,
324... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... |
[
"weights",
"that",
"will",
"get",
"the",
"right",
"output",
"for",
"each",
"input",
"during",
"training",
".",
"the",
"most",
"common",
"training",
"technique",
"is",
"the",
"back",
"##pro",
"##pa",
"##gation",
"algorithm",
".",
"neural",
"networks",
"learn",
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
2,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
2,
2,
2,
0,
0,
0... | [
{
"start_token_idx": 20,
"end_token_idx": 24,
"identifier": "/wiki/Backpropagation",
"tokens": [
"back",
"##pro",
"##pa",
"##gation"
]
},
{
"start_token_idx": 38,
"end_token_idx": 40,
"identifier": "/wiki/Pattern_recognition",
"tokens": [
"find",... | [
101,
15871,
2008,
2097,
2131,
1996,
2157,
6434,
2005,
2169,
7953,
2076,
2731,
1012,
1996,
2087,
2691,
2731,
6028,
2003,
1996,
2067,
21572,
4502,
12540,
9896,
1012,
15756,
6125,
4553,
2000,
2944,
3375,
6550,
2090,
20407,
1998,
27852,
1998,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... |
[
"like",
"text",
"by",
"repeatedly",
"predicting",
"the",
"next",
"token",
".",
"typically",
",",
"a",
"subsequent",
"training",
"phase",
"makes",
"the",
"model",
"more",
"truth",
"##ful",
",",
"useful",
"and",
"harmless",
",",
"usually",
"with",
"a",
"techni... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
2,
2,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
2,
0,
0,
0... | [
{
"start_token_idx": 31,
"end_token_idx": 36,
"identifier": "/wiki/Reinforcement_learning_from_human_feedback",
"tokens": [
"reinforcement",
"learning",
"from",
"human",
"feedback"
]
},
{
"start_token_idx": 57,
"end_token_idx": 60,
"identifier": "/wi... | [
101,
2066,
3793,
2011,
8385,
29458,
1996,
2279,
19204,
1012,
4050,
1010,
1037,
4745,
2731,
4403,
3084,
1996,
2944,
2062,
3606,
3993,
1010,
6179,
1998,
19741,
1010,
2788,
2007,
1037,
6028,
2170,
23895,
4083,
2013,
2529,
12247,
1006,
1054,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... |
[
".",
"for",
"example",
",",
"alpha",
"##fold",
"2",
"(",
"2021",
")",
"demonstrated",
"the",
"ability",
"to",
"approximate",
",",
"in",
"hours",
"rather",
"than",
"months",
",",
"the",
"3d",
"structure",
"of",
"a",
"protein",
".",
"in",
"202",
"##3",
"... | [
0,
0,
0,
0,
0,
1,
2,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
2,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
{
"start_token_idx": 4,
"end_token_idx": 7,
"identifier": "/wiki/AlphaFold_2",
"tokens": [
"alpha",
"##fold",
"2"
]
},
{
"start_token_idx": 24,
"end_token_idx": 28,
"identifier": "/wiki/Protein_structure",
"tokens": [
"structure",
"of",
"a"... | [
101,
1012,
2005,
2742,
1010,
6541,
10371,
1016,
1006,
25682,
1007,
7645,
1996,
3754,
2000,
15796,
1010,
1999,
2847,
2738,
2084,
2706,
1010,
1996,
7605,
3252,
1997,
1037,
5250,
1012,
1999,
16798,
2509,
1010,
2009,
2001,
2988,
2008,
9932,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... |
[
",",
"financial",
"planning",
",",
"and",
"pension",
"advice",
"in",
"the",
"process",
",",
"but",
"i",
"’",
"m",
"not",
"sure",
"it",
"will",
"un",
"##lea",
"##sh",
"a",
"new",
"wave",
"of",
"[",
"e",
".",
"g",
".",
",",
"sophisticated",
"]",
"pen... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
{
"start_token_idx": 51,
"end_token_idx": 54,
"identifier": "/wiki/Command_and_control",
"tokens": [
"command",
"and",
"control"
]
},
{
"start_token_idx": 88,
"end_token_idx": 90,
"identifier": "/wiki/Vehicular_automation",
"tokens": [
"autonomous",
... | [
101,
1010,
3361,
4041,
1010,
1998,
11550,
6040,
1999,
1996,
2832,
1010,
2021,
1045,
1521,
1049,
2025,
2469,
2009,
2097,
4895,
19738,
4095,
1037,
2047,
4400,
1997,
1031,
1041,
1012,
1043,
1012,
1010,
12138,
1033,
11550,
8144,
1012,
1000,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... |
[
"data",
"and",
"applications",
",",
"mainly",
"for",
"\"",
"classification",
",",
"regression",
",",
"cluster",
"##ing",
",",
"forecast",
"##ing",
",",
"generation",
",",
"discovery",
",",
"and",
"the",
"development",
"of",
"new",
"scientific",
"insights",
"\"... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
{
"start_token_idx": 116,
"end_token_idx": 121,
"identifier": "/wiki/Demis_Hassabis",
"tokens": [
"demi",
"##s",
"has",
"##sa",
"##bis"
]
},
{
"start_token_idx": 122,
"end_token_idx": 124,
"identifier": "/wiki/DeepMind",
"tokens": [
"deep",... | [
101,
2951,
1998,
5097,
1010,
3701,
2005,
1000,
5579,
1010,
26237,
1010,
9324,
2075,
1010,
19939,
2075,
1010,
4245,
1010,
5456,
1010,
1998,
1996,
2458,
1997,
2047,
4045,
20062,
1000,
2005,
2742,
2005,
13648,
4654,
7361,
20644,
3215,
1010,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... |
[
"copyright",
"##ed",
"work",
"\"",
".",
"website",
"owners",
"who",
"do",
"not",
"wish",
"to",
"have",
"their",
"content",
"scraped",
"can",
"indicate",
"it",
"in",
"a",
"\"",
"robots",
".",
"tx",
"##t",
"\"",
"file",
".",
"in",
"202",
"##3",
",",
"l... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
2,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
2,
2,
0,
1,
2,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
{
"start_token_idx": 22,
"end_token_idx": 26,
"identifier": "/wiki/Robots.txt",
"tokens": [
"robots",
".",
"tx",
"##t"
]
},
{
"start_token_idx": 37,
"end_token_idx": 41,
"identifier": "/wiki/John_Grisham",
"tokens": [
"john",
"gr",
"#... | [
101,
9385,
2098,
2147,
1000,
1012,
4037,
5608,
2040,
2079,
2025,
4299,
2000,
2031,
2037,
4180,
20378,
2064,
5769,
2009,
1999,
1037,
1000,
13507,
1012,
19067,
2102,
1000,
5371,
1012,
1999,
16798,
2509,
1010,
2877,
6048,
1006,
2164,
2198,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... |
[
"labeling",
"feature",
"mistakenly",
"identified",
"jack",
"##y",
"al",
"##cine",
"and",
"a",
"friend",
"as",
"\"",
"gorilla",
"##s",
"\"",
"because",
"they",
"were",
"black",
".",
"the",
"system",
"was",
"trained",
"on",
"a",
"data",
"##set",
"that",
"con... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
{
"start_token_idx": 100,
"end_token_idx": 102,
"identifier": "/wiki/COMPAS_(software)",
"tokens": [
"com",
"##pas"
]
},
{
"start_token_idx": 109,
"end_token_idx": 114,
"identifier": "/wiki/U.S._court",
"tokens": [
"u",
".",
"s",
".",
... | [
101,
28847,
3444,
20706,
4453,
2990,
2100,
2632,
16567,
1998,
1037,
2767,
2004,
1000,
23526,
2015,
1000,
2138,
2027,
2020,
2304,
1012,
1996,
2291,
2001,
4738,
2006,
1037,
2951,
13462,
2008,
4838,
2200,
2261,
4871,
1997,
2304,
2111,
1010,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... |
[
"##ete",
"##cted",
"because",
"the",
"developers",
"are",
"overwhelmingly",
"white",
"and",
"male",
":",
"among",
"ai",
"engineers",
",",
"about",
"4",
"%",
"are",
"black",
"and",
"20",
"%",
"are",
"women",
".",
"at",
"its",
"202",
"##2",
"conference",
"... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
2,
2,
2,
2,
2,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
2,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
{
"start_token_idx": 30,
"end_token_idx": 38,
"identifier": "/wiki/ACM_Conference_on_Fairness,_Accountability,_and_Transparency",
"tokens": [
"conference",
"on",
"fairness",
",",
"accountability",
",",
"and",
"transparency"
]
},
{
"start_to... | [
101,
12870,
10985,
2138,
1996,
9797,
2024,
24783,
2317,
1998,
3287,
1024,
2426,
9932,
6145,
1010,
2055,
1018,
1003,
2024,
2304,
1998,
2322,
1003,
2024,
2308,
1012,
2012,
2049,
16798,
2475,
3034,
2006,
26935,
1010,
17842,
1010,
1998,
16987,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... |
[
"lime",
"can",
"locally",
"approximate",
"a",
"model",
"with",
"a",
"simpler",
",",
"interpret",
"##able",
"model",
".",
"multi",
"##tas",
"##k",
"learning",
"provides",
"a",
"large",
"number",
"of",
"outputs",
"in",
"addition",
"to",
"the",
"target",
"class... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
2,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
2,
2,
2,
0,
1,
2,
2,
0,
0,
1,
2,
0,
0,
0,
0... | [
{
"start_token_idx": 14,
"end_token_idx": 18,
"identifier": "/wiki/Multitask_learning",
"tokens": [
"multi",
"##tas",
"##k",
"learning"
]
},
{
"start_token_idx": 46,
"end_token_idx": 51,
"identifier": "/wiki/Deconvolution",
"tokens": [
"deco",
... | [
101,
14123,
2064,
7246,
15796,
1037,
2944,
2007,
1037,
16325,
1010,
17841,
3085,
2944,
1012,
4800,
10230,
2243,
4083,
3640,
1037,
2312,
2193,
1997,
27852,
1999,
2804,
2000,
1996,
4539,
5579,
1012,
2122,
2060,
27852,
2064,
2393,
9797,
2139,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... |
[
"in",
"the",
"past",
",",
"technology",
"has",
"tended",
"to",
"increase",
"rather",
"than",
"reduce",
"total",
"employment",
",",
"but",
"economists",
"acknowledge",
"that",
"\"",
"we",
"'",
"re",
"in",
"un",
"##cha",
"##rted",
"territory",
"\"",
"with",
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0... | [
{
"start_token_idx": 56,
"end_token_idx": 57,
"identifier": "/wiki/Unemployment",
"tokens": [
"unemployment"
]
},
{
"start_token_idx": 70,
"end_token_idx": 71,
"identifier": "/wiki/Productivity",
"tokens": [
"productivity"
]
},
{
"start_token_idx": 73,... | [
101,
1999,
1996,
2627,
1010,
2974,
2038,
11121,
2000,
3623,
2738,
2084,
5547,
2561,
6107,
1010,
2021,
22171,
13399,
2008,
1000,
2057,
1005,
2128,
1999,
4895,
7507,
17724,
3700,
1000,
2007,
9932,
1012,
1037,
5002,
1997,
22171,
3662,
18185,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... |
End of preview. Expand in Data Studio
- Downloads last month
- 4