{"id": 147459, "image": "COCO_train2014_000000147459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man riding on a skateboard\"."}], "task": "refering", "answer_template": "The \"a man riding on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 104, 105, 106, 107, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 288, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316, 335, 336, 338, 339, 340, 357, 358, 359, 361, 362, 363, 380, 383, 384, 385, 386, 405, 406, 407, 408, 409, 428, 429], "bbox": [0.47015624999999994, 0.18217726396917147, 0.8564999999999999, 0.9619460500963392], "iscrowd": 0, "area": 41208.142700000004, "rle": {"size": [519, 640], "counts": "`jh49e?901O000O2O00001O0O2O000O01N100O1O2O01O01O0000001O0O10001N1001O10O01O100XEBf5>YJHb59\\JM_54^J3^5L`FFS3a0[6J`F0k26e6K_F:a2Lo6J`Fd0X2BW7K_Fn0P2XO`7K_FW1h1mNi7M]FZ1h1iNj7M^F]1e1gNl7M^F`1c1bNo7O\\Fc1b1_NQ8N]Ff1`1\\NZ4XO_Mh0fLg1^1ZNV4AbM?hLk1]1TNZ4E]M=lLl1[1RN[4IZM9QMn1X1PN\\4MWM6TMP2W1lM_40SM5WMR2S1jMb41RM3YMU2Q1gMc44oL2]MU2o0eMe46kL1bMV2l0bMg49iLOdMZ2i0^Mi4bLLmM]2c0ZMm4?jKlNlNo0K^2`0WMo4c0fKROhNg03`2;VMT5b0bKa0oNj14XMZ5?`K`0SOk1MYMa5=[Ka0WOl1F[Mf5;ZK?[OV3[5[LWKa0_OT3Z5\\LSKb0CR3Z5^LPKa0GQ3Y5_LmJa0JQ3X5`LkJa0Mn2Y5bLgJa01m2X5bLdJc04l2Y5aL`Jd08k2X5bL]Jd0lJSM_O[2f5`0lJUMCU2b5e0kJXMFo1`5f0lJ[MHj1]5j0lJ\\MKe1[5l0lJ_MN`1X5n0kJbM1[1X5o0hJfM3W1X5Q1eJhM5T1Y5S1aJiM9Q1X5V1]JiM=n0X5Y1YJjMa0j0X5\\1VJjMe0g0V5VNcIe2a0^Nh0d0V5QNiIn27]Nl0a0k5Q1WI_NQ1=i5T1UI_NT1:g5Z1SI\\NX17f5^1QI[N[14e5b1nH[N`10a5g1nHYNc1M`5l1kHWNg1J_5P2iHVNk1G[5U2hHUNo1CZ5Y2fHTNR2@HQO\\3]3hJRNW2\\OE[OV3Y3mJPNZ2ZOBDQ3S3QKPN^2VO@Ml2c4fM]K]O4j2`4kMZKZO7k2_4lMXKYO9l2_4mMUKWO=k2_4oMRKVO`0k2^4PNQKTOb0l2]4RNnJROe0l2`4QNiJSOh0l2a4QNeJQOl0n2a4PNaJROo0m2c4M]K4e4I\\K7f4F[K:g4CYK>j4^OWKb0k4\\OUKd0m4YOTKh0m4VORKk0P5SOPKm0R5POoJP1S5nNmJR1V5jNjJW1Y5fNgJ[1[5bNeJ^1]5_NdJa1_5\\N`Je1c5WN^Ji1e5TN[Jl1h5QNWJQ2j5lMWJT2l5iMTJW2o5eMRJ[2Q6`MPJa2S6ZMoIf2S6VMoIk2S6PMoIP3T6kLmIV3V6fLkIZ3X6aLjI_3X6]LiId3Z6WLhIj3b7iJ`HW5X90O2O00001N1^FeJX8[5hGgJV8Z5iGhJU8X5kGjJR8X5mGjJQ8V5oGlJo7T5QHmJm7U5RHmJS2F\\2]5aKoJP2H\\2Y5eKPKk1K^2V5fKQKi1L^2T5iKRKe1O_2P5kKSKb11a2l4nKTK^14a2h4RLTKZ18a2f4ULSKV1c2^4`K`J9e0`1a0d2\\4`K`J=d0[1e0e2W4`KcJa0b0W1g0f2U4^KfJe0?S1k0g2b4WLdJo0m0g2`4[LdJj0Q1h2[4_LeJe0T1j2X4aLeJb0V1k2U4dL\\NX3e1iL\\NU3d1lL^NQ3c1oL^No2c1QM_Nk2c1VM]Nh2e1WM\\Ng2e1YM]Nc2e1^M[N`2g1_M[N^2f1cMZN[2h1dMZNX2h1hMYNV2h1kMXNS2j1nMVNo1k1RNcIPOl3k2c2VN[ITOT4c2c2ZNRIYO\\4Z2c2^NkH]Oc4R2d2SO_Mj0c2SO`Mk0a2TObMi0`2TOdMh0^2VOeMh0\\2WOgMf0[2WOhMg0Y2_OcM=_2MXM_NVKV1d7`0]M^Od2a0^M]Od2`0^M^Od2`0_M^Ob2a0_M^Oc2?`M^Ob2`0`M_Ob2?`M_Oa2?bM_O_2`0cM]Oa2?bM_O`2?bM_O`2>cM@_2>cM_O`2>cM@_2=g6O2M2O1N3N1N3N0O2O1N1O2O0O2O1N101N2O0Ob\\^1"}, "label": "a man riding on a skateboard"}]} {"id": 147459, "image": "COCO_train2014_000000147459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skateboarder with black shoes\"."}], "task": "refering", "answer_template": "The \"skateboarder with black shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 104, 105, 106, 107, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 288, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316, 335, 336, 338, 339, 340, 357, 358, 359, 361, 362, 363, 380, 383, 384, 385, 386, 405, 406, 407, 408, 409, 428, 429], "bbox": [0.47015624999999994, 0.18217726396917147, 0.8564999999999999, 0.9619460500963392], "iscrowd": 0, "area": 41208.142700000004, "rle": {"size": [519, 640], "counts": "`jh49e?901O000O2O00001O0O2O000O01N100O1O2O01O01O0000001O0O10001N1001O10O01O100XEBf5>YJHb59\\JM_54^J3^5L`FFS3a0[6J`F0k26e6K_F:a2Lo6J`Fd0X2BW7K_Fn0P2XO`7K_FW1h1mNi7M]FZ1h1iNj7M^F]1e1gNl7M^F`1c1bNo7O\\Fc1b1_NQ8N]Ff1`1\\NZ4XO_Mh0fLg1^1ZNV4AbM?hLk1]1TNZ4E]M=lLl1[1RN[4IZM9QMn1X1PN\\4MWM6TMP2W1lM_40SM5WMR2S1jMb41RM3YMU2Q1gMc44oL2]MU2o0eMe46kL1bMV2l0bMg49iLOdMZ2i0^Mi4bLLmM]2c0ZMm4?jKlNlNo0K^2`0WMo4c0fKROhNg03`2;VMT5b0bKa0oNj14XMZ5?`K`0SOk1MYMa5=[Ka0WOl1F[Mf5;ZK?[OV3[5[LWKa0_OT3Z5\\LSKb0CR3Z5^LPKa0GQ3Y5_LmJa0JQ3X5`LkJa0Mn2Y5bLgJa01m2X5bLdJc04l2Y5aL`Jd08k2X5bL]Jd0lJSM_O[2f5`0lJUMCU2b5e0kJXMFo1`5f0lJ[MHj1]5j0lJ\\MKe1[5l0lJ_MN`1X5n0kJbM1[1X5o0hJfM3W1X5Q1eJhM5T1Y5S1aJiM9Q1X5V1]JiM=n0X5Y1YJjMa0j0X5\\1VJjMe0g0V5VNcIe2a0^Nh0d0V5QNiIn27]Nl0a0k5Q1WI_NQ1=i5T1UI_NT1:g5Z1SI\\NX17f5^1QI[N[14e5b1nH[N`10a5g1nHYNc1M`5l1kHWNg1J_5P2iHVNk1G[5U2hHUNo1CZ5Y2fHTNR2@HQO\\3]3hJRNW2\\OE[OV3Y3mJPNZ2ZOBDQ3S3QKPN^2VO@Ml2c4fM]K]O4j2`4kMZKZO7k2_4lMXKYO9l2_4mMUKWO=k2_4oMRKVO`0k2^4PNQKTOb0l2]4RNnJROe0l2`4QNiJSOh0l2a4QNeJQOl0n2a4PNaJROo0m2c4M]K4e4I\\K7f4F[K:g4CYK>j4^OWKb0k4\\OUKd0m4YOTKh0m4VORKk0P5SOPKm0R5POoJP1S5nNmJR1V5jNjJW1Y5fNgJ[1[5bNeJ^1]5_NdJa1_5\\N`Je1c5WN^Ji1e5TN[Jl1h5QNWJQ2j5lMWJT2l5iMTJW2o5eMRJ[2Q6`MPJa2S6ZMoIf2S6VMoIk2S6PMoIP3T6kLmIV3V6fLkIZ3X6aLjI_3X6]LiId3Z6WLhIj3b7iJ`HW5X90O2O00001N1^FeJX8[5hGgJV8Z5iGhJU8X5kGjJR8X5mGjJQ8V5oGlJo7T5QHmJm7U5RHmJS2F\\2]5aKoJP2H\\2Y5eKPKk1K^2V5fKQKi1L^2T5iKRKe1O_2P5kKSKb11a2l4nKTK^14a2h4RLTKZ18a2f4ULSKV1c2^4`K`J9e0`1a0d2\\4`K`J=d0[1e0e2W4`KcJa0b0W1g0f2U4^KfJe0?S1k0g2b4WLdJo0m0g2`4[LdJj0Q1h2[4_LeJe0T1j2X4aLeJb0V1k2U4dL\\NX3e1iL\\NU3d1lL^NQ3c1oL^No2c1QM_Nk2c1VM]Nh2e1WM\\Ng2e1YM]Nc2e1^M[N`2g1_M[N^2f1cMZN[2h1dMZNX2h1hMYNV2h1kMXNS2j1nMVNo1k1RNcIPOl3k2c2VN[ITOT4c2c2ZNRIYO\\4Z2c2^NkH]Oc4R2d2SO_Mj0c2SO`Mk0a2TObMi0`2TOdMh0^2VOeMh0\\2WOgMf0[2WOhMg0Y2_OcM=_2MXM_NVKV1d7`0]M^Od2a0^M]Od2`0^M^Od2`0_M^Ob2a0_M^Oc2?`M^Ob2`0`M_Ob2?`M_Oa2?bM_O_2`0cM]Oa2?bM_O`2?bM_O`2>cM@_2>cM_O`2>cM@_2=g6O2M2O1N3N1N3N0O2O1N1O2O0O2O1N101N2O0Ob\\^1"}, "label": "skateboarder with black shoes"}]} {"id": 466949, "image": "COCO_train2014_000000466949.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a van standing behind the yellow vehicle\"."}], "task": "refering", "answer_template": "The \"a van standing behind the yellow vehicle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234], "bbox": [0.0, 0.44503529411764703, 0.20392187499999997, 0.7096705882352942], "iscrowd": 0, "area": 9957.663549999996, "rle": {"size": [425, 640], "counts": "o5S3V:00O1O1L4O11O00000000000O100000001O000000000O1002N000001O000000000000000000001O000000000000001O3M3M1O1O001O1O1O00001O000000001O00000000000000O1000000O10000O100N3M2N2N2N2000000000000000000ROWF\\Ni9Q1jFoNV9c0XG]OZ;00000001O0000000001O000000000000O101O000000000O100000000000000O2O0Hhob6"}, "label": "a van standing behind the yellow vehicle"}]} {"id": 466949, "image": "COCO_train2014_000000466949.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a van parked behind a yellow school bus\"."}], "task": "refering", "answer_template": "The \"a van parked behind a yellow school bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234], "bbox": [0.0, 0.44503529411764703, 0.20392187499999997, 0.7096705882352942], "iscrowd": 0, "area": 9957.663549999996, "rle": {"size": [425, 640], "counts": "o5S3V:00O1O1L4O11O00000000000O100000001O000000000O1002N000001O000000000000000000001O000000000000001O3M3M1O1O001O1O1O00001O000000001O00000000000000O1000000O10000O100N3M2N2N2N2000000000000000000ROWF\\Ni9Q1jFoNV9c0XG]OZ;00000001O0000000001O000000000000O101O000000000O100000000000000O2O0Hhob6"}, "label": "a van parked behind a yellow school bus"}]} {"id": 466949, "image": "COCO_train2014_000000466949.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver colored sedan\"."}], "task": "refering", "answer_template": "The \"a silver colored sedan\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 204, 205, 206, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275, 296, 297, 298], "bbox": [0.8154375, 0.5268705882352941, 0.998234375, 0.8666588235294117], "iscrowd": 0, "area": 11050.414449999997, "rle": {"size": [425, 640], "counts": "Rnh6`0a<9O1N1O2N2EROYDn0b;YO[Di0`;]O]De0`;^O_Dc0];f0N3M3L4M2N2N3O0000001O0000001O0000001O0000001O01O0001O0000001O00001O0000001O000000003M6K2M2N2N3M2N2N2N2N1O1O00001O00001O00001N100O2O0O101N1O101N100O2N100O2O0O1O010O01O00001O010O001O00010O001O000010O01O000010O01mNR1lM^8"}, "label": "a silver colored sedan"}]} {"id": 466949, "image": "COCO_train2014_000000466949.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a car parked in front of school bus\"."}], "task": "refering", "answer_template": "The \"a car parked in front of school bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 204, 205, 206, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275, 296, 297, 298], "bbox": [0.8154375, 0.5268705882352941, 0.998234375, 0.8666588235294117], "iscrowd": 0, "area": 11050.414449999997, "rle": {"size": [425, 640], "counts": "Rnh6`0a<9O1N1O2N2EROYDn0b;YO[Di0`;]O]De0`;^O_Dc0];f0N3M3L4M2N2N3O0000001O0000001O0000001O0000001O01O0001O0000001O00001O0000001O000000003M6K2M2N2N3M2N2N2N2N1O1O00001O00001O00001N100O2O0O101N1O101N100O2N100O2O0O1O010O01O00001O010O001O00010O001O000010O01O000010O01mNR1lM^8"}, "label": "a car parked in front of school bus"}]} {"id": 376838, "image": "COCO_train2014_000000376838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with no hat , with his back toward the camera\"."}], "task": "refering", "answer_template": "The \"a man with no hat , with his back toward the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 99, 100, 101, 117, 118, 119, 120, 135, 136, 137, 138, 153, 154, 155, 171, 172, 173, 189, 190, 191, 208, 209, 226, 227], "bbox": [0.5189400000000001, 0.3408753315649867, 0.67742, 0.9576657824933686], "iscrowd": 0, "area": 11662.9175, "rle": {"size": [377, 500], "counts": "bbo22`;g0XOg0ZO=C8H9G9F9K7L4L4L4L3M4M3L4L3M2N2N2N2N2N2SOTLjIQ4P6RLnIZ4d5jKZJc4Z5^KdJn4:[JV4S700000000000000000O10000000000000001O1O1^I\\Km5f4oI^Ko5f4jI^KT6V5K5K4L3N3L3M4L3D in this image.", "objects": [{"patches": [82, 83, 99, 100, 101, 117, 118, 119, 120, 135, 136, 137, 138, 153, 154, 155, 171, 172, 173, 189, 190, 191, 208, 209, 226, 227], "bbox": [0.5189400000000001, 0.3408753315649867, 0.67742, 0.9576657824933686], "iscrowd": 0, "area": 11662.9175, "rle": {"size": [377, 500], "counts": "bbo22`;g0XOg0ZO=C8H9G9F9K7L4L4L4L3M4M3L4L3M2N2N2N2N2N2SOTLjIQ4P6RLnIZ4d5jKZJc4Z5^KdJn4:[JV4S700000000000000000O10000000000000001O1O1^I\\Km5f4oI^Ko5f4jI^KT6V5K5K4L3N3L3M4L3D in this image.", "objects": [{"patches": [80, 81, 97, 98, 99, 115, 116, 117, 133, 134, 135, 151, 152, 153, 169, 170, 171, 187, 188, 189, 205, 206, 207], "bbox": [0.41888, 0.28872679045092836, 0.5300199999999999, 0.9213793103448277], "iscrowd": 0, "area": 8998.710450000004, "rle": {"size": [377, 500], "counts": "^S]22f;h0XOg0YOh0XOg0ZOg0XOg0YOh0XOa0_O1O1N3N1N2O1O2M2YOmIgK\\5AUKg4BdKW5KVK`4DcKR53YKY4GaKm4=ZKQ4L_Kg4f0\\Kj3R5]LlJb3Q5cLnJ]3m4jLRKT3k4mLoJ_N4d4h4gL_KgNGa4f4kKeLEdN`4h6aKWI]4j6eKTI[4l68100O1O100O101N100O10B\\KgIc4l0]Ko36TK]4j0aKl38YKV4i0gKg39_KS4d0hKg3;dKU4:dKl3=jKU5o3QKPLP5g1oI0W1XNl4U1lJ0>jNg4c0kK0D\\Oc40iL0kNOh9O1O1O10nUf2"}, "label": "soldier without coat"}]} {"id": 376838, "image": "COCO_train2014_000000376838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the middle wearing black\"."}], "task": "refering", "answer_template": "The \"the man in the middle wearing black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 97, 98, 99, 115, 116, 117, 133, 134, 135, 151, 152, 153, 169, 170, 171, 187, 188, 189, 205, 206, 207], "bbox": [0.41888, 0.28872679045092836, 0.5300199999999999, 0.9213793103448277], "iscrowd": 0, "area": 8998.710450000004, "rle": {"size": [377, 500], "counts": "^S]22f;h0XOg0YOh0XOg0ZOg0XOg0YOh0XOa0_O1O1N3N1N2O1O2M2YOmIgK\\5AUKg4BdKW5KVK`4DcKR53YKY4GaKm4=ZKQ4L_Kg4f0\\Kj3R5]LlJb3Q5cLnJ]3m4jLRKT3k4mLoJ_N4d4h4gL_KgNGa4f4kKeLEdN`4h6aKWI]4j6eKTI[4l68100O1O100O101N100O10B\\KgIc4l0]Ko36TK]4j0aKl38YKV4i0gKg39_KS4d0hKg3;dKU4:dKl3=jKU5o3QKPLP5g1oI0W1XNl4U1lJ0>jNg4c0kK0D\\Oc40iL0kNOh9O1O1O10nUf2"}, "label": "the man in the middle wearing black"}]} {"id": 376838, "image": "COCO_train2014_000000376838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a hat with a strap\"."}], "task": "refering", "answer_template": "The \"a man in a hat with a strap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 95, 96, 97, 113, 114, 115, 131, 132, 133, 149, 150, 151, 167, 168, 169, 185, 186, 203, 204, 221, 222], "bbox": [0.29144, 0.3078779840848806, 0.43038, 0.9505570291777189], "iscrowd": 0, "area": 10308.403400000003, "rle": {"size": [377, 500], "counts": "doe17Z;:F9G9bNROgGX1U8SOZGV1e2[Nj2g2dLjMZ3X2SLZNk3j1fKcNY4_1dKcN[4i1WKZNg4k1TKWNk4l1QKVNn4R2hJQNV5Z4O1O1N2O1O1N2O1O1O1O1O1O100O10000O10FSKhIX4NULQ6LQJo33nKf5 in this image.", "objects": [{"patches": [78, 79, 95, 96, 97, 113, 114, 115, 131, 132, 133, 149, 150, 151, 167, 168, 169, 185, 186, 203, 204, 221, 222], "bbox": [0.29144, 0.3078779840848806, 0.43038, 0.9505570291777189], "iscrowd": 0, "area": 10308.403400000003, "rle": {"size": [377, 500], "counts": "doe17Z;:F9G9bNROgGX1U8SOZGV1e2[Nj2g2dLjMZ3X2SLZNk3j1fKcNY4_1dKcN[4i1WKZNg4k1TKWNk4l1QKVNn4R2hJQNV5Z4O1O1N2O1O1N2O1O1O1O1O1O100O10000O10FSKhIX4NULQ6LQJo33nKf5 in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 237, 238, 239, 240, 241, 242, 243, 244, 245, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.2483125, 0.3051875, 0.74325, 0.9718541666666667], "iscrowd": 0, "area": 85246.46705000002, "rle": {"size": [480, 640], "counts": "ToZ22j>6J5K6K4K6J5K6J5L3L4L4L4M3L4L4L4L4M3L4L4L4M3L4L4L4L4M3dIcLQ1a3oNkLa0Y3_OTM0P30\\MAg2?fMPO^2P1nM`NV2`1WNoMm1Q2`N^Md1b2hNoL[1Q3RO^LR1b3ZOQLg0o3FdK:\\42XKNh49QKGo49QKGo49QKGo4U5000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000eJZK in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 237, 238, 239, 240, 241, 242, 243, 244, 245, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.2483125, 0.3051875, 0.74325, 0.9718541666666667], "iscrowd": 0, "area": 85246.46705000002, "rle": {"size": [480, 640], "counts": "ToZ22j>6J5K6K4K6J5K6J5L3L4L4L4M3L4L4L4L4M3L4L4L4M3L4L4L4L4M3dIcLQ1a3oNkLa0Y3_OTM0P30\\MAg2?fMPO^2P1nM`NV2`1WNoMm1Q2`N^Md1b2hNoL[1Q3RO^LR1b3ZOQLg0o3FdK:\\42XKNh49QKGo49QKGo49QKGo4U5000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000eJZK in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 26, 27, 28, 29, 30, 31, 32, 33, 34, 50, 51, 52, 53, 54, 55, 56, 57, 73, 74, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190], "bbox": [0.108359375, 0.0175, 0.4994375, 0.48702083333333335], "iscrowd": 0, "area": 36246.51975, "rle": {"size": [480, 640], "counts": "TQQ11m>2N3N1N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N2O1N2O1N2O1N2N2O1N3cJaNVM`1j2fNnL\\1R3jNfLX1Z3nN_LS1a3SOWLo0i3WOPLj0P4\\OhKf0X4@aKa0_4EYK=g4IQK9o4MjJ4V52bJ0^56[JKe5;SJGm5?kICU6c0dI^O\\6g0]I[Oc6k0VIVOj6P1nHROR7T1gHmNY7Y1_HiNa7]1WHeNi7a1PH`NP8X4000001O0000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000ZMTHYNk7g1^HPNb7P2gHgMY7Y2PI^MP7b2YIUMg6k2bIlL^6T3kIcLU6]3TJZLl5f3]JQLc5o3fJhKZ5X4g2000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000004LP1POQ1oNQ1oNQ1oNake4"}, "label": "the laptop in the back to the left"}]} {"id": 344073, "image": "COCO_train2014_000000344073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the computer on the back left\"."}], "task": "refering", "answer_template": "The \"the computer on the back left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 26, 27, 28, 29, 30, 31, 32, 33, 34, 50, 51, 52, 53, 54, 55, 56, 57, 73, 74, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190], "bbox": [0.108359375, 0.0175, 0.4994375, 0.48702083333333335], "iscrowd": 0, "area": 36246.51975, "rle": {"size": [480, 640], "counts": "TQQ11m>2N3N1N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N2O1N2O1N2O1N2N2O1N3cJaNVM`1j2fNnL\\1R3jNfLX1Z3nN_LS1a3SOWLo0i3WOPLj0P4\\OhKf0X4@aKa0_4EYK=g4IQK9o4MjJ4V52bJ0^56[JKe5;SJGm5?kICU6c0dI^O\\6g0]I[Oc6k0VIVOj6P1nHROR7T1gHmNY7Y1_HiNa7]1WHeNi7a1PH`NP8X4000001O0000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000ZMTHYNk7g1^HPNb7P2gHgMY7Y2PI^MP7b2YIUMg6k2bIlL^6T3kIcLU6]3TJZLl5f3]JQLc5o3fJhKZ5X4g2000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000004LP1POQ1oNQ1oNQ1oNake4"}, "label": "the computer on the back left"}]} {"id": 344073, "image": "COCO_train2014_000000344073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top and rightmost of three open laptops\"."}], "task": "refering", "answer_template": "The \"the top and rightmost of three open laptops\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 18, 19, 34, 35, 36, 37, 38, 39, 40, 41, 42, 57, 58, 59, 60, 61, 62, 63, 64, 65, 80, 81, 82, 83, 84, 85, 86, 87, 88, 103, 104, 105, 106, 107, 108, 109, 110, 129, 130, 131, 132, 133, 134, 154, 155, 156, 157, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204], "bbox": [0.50234375, 0.010687499999999999, 0.891609375, 0.5028333333333334], "iscrowd": 0, "area": 37726.562750000005, "rle": {"size": [480, 640], "counts": "Unf4g0Y>g0YOg0YOf0ZOg0YOd0\\O000000000001O01O000000000000000000000000000001O000000000000000001O000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O00000000000000000000000VGbLR6_3QI_Mm6c2VH[Ng7a4N2N000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000O10000QM\\HbNd7V1eHiN[7o0nHPOR7h0XIVOh6c0`I\\O`6 in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32], "bbox": [0.0016875000000000002, 0.0050625, 0.57471875, 0.0994375], "iscrowd": 0, "area": 11223.06035, "rle": {"size": [480, 640], "counts": "R?7i>>B:F0000000000001O000000000000001O000000000000001O0000000000001O000000000000001O000000000000001O0000000000001O000000000000001O000000000000001O0000000000001O000000000000001O000000000000001O0000000000001O000000000000001O000000000000001O000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000000000O1M3M3N200001O0000001O00001O0000001O00001O00001O0000001O00O1O1N2O1N2N200O1O1O1O1O100001O00000000001O00000000001O00000000001O0000000000001O00000000001O0000O10000000000000000000000000000000000000000000000000000O1000000O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100000000000000000000000000000000O1000000000000001O000000000000001O00000000000000001O000000000000001OO1N2N2O1N2N2N2O1NP`o3"}, "label": "the yellow table shade above the man in the foreground"}]} {"id": 516106, "image": "COCO_train2014_000000516106.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white umbrella on the left\"."}], "task": "refering", "answer_template": "The \"white umbrella on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32], "bbox": [0.0016875000000000002, 0.0050625, 0.57471875, 0.0994375], "iscrowd": 0, "area": 11223.06035, "rle": {"size": [480, 640], "counts": "R?7i>>B:F0000000000001O000000000000001O000000000000001O0000000000001O000000000000001O000000000000001O0000000000001O000000000000001O000000000000001O0000000000001O000000000000001O000000000000001O0000000000001O000000000000001O000000000000001O000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000000000O1M3M3N200001O0000001O00001O0000001O00001O00001O0000001O00O1O1N2O1N2N200O1O1O1O1O100001O00000000001O00000000001O00000000001O0000000000001O00000000001O0000O10000000000000000000000000000000000000000000000000000O1000000O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100000000000000000000000000000000O1000000000000001O000000000000001O00000000000000001O000000000000001OO1N2N2O1N2N2N2O1NP`o3"}, "label": "white umbrella on the left"}]} {"id": 516106, "image": "COCO_train2014_000000516106.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bowl on soup on a brown table and black lunch tray\"."}], "task": "refering", "answer_template": "The \"bowl on soup on a brown table and black lunch tray\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.131453125, 0.6741666666666667, 0.9977499999999999, 0.99775], "iscrowd": 0, "area": 63019.14394999999, "rle": {"size": [480, 640], "counts": "_jW15j> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 65, 88, 111, 134, 156, 157, 179, 180], "bbox": [0.559125, 0.0033750000000000004, 1.0, 0.4763541666666667], "iscrowd": 0, "area": 12117.137750000002, "rle": {"size": [480, 640], "counts": "hjW52l>2N2N3L3N3M2N2N3M2N3O00001O001O00001O00001O001O0000001O000000001O000000001O00000000001O000000001O0000000000001O0000000000001O0000000000001O0000000000001O0O100000001O0000000000001O0000000000001O000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000lGoNY2Q1ZM]Oe2c0oLIQ37bL6^3JVLb0j3]OjKQ1U4oN_K]1a4cNRKj1n4VNfJV2Z5jMZJc2e5]MnIP3R6PMbI\\3^6dLUIi3k6VLjHP4\\7PLWHg3_8g0f0ZOf0ZOf0ZOf0ZOf0ZOf0ZOf0I7000000000000000000000000000000000000000000000000000000000000O100000000O1000000O100000000O100000000O1000000O100000000O100000000001O00001O00001O0000001O00001O00001O00001O00001O0000001O0000]Oa0"}, "label": "pole on a green umbrella"}]} {"id": 516106, "image": "COCO_train2014_000000516106.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green umbrella and its pole\"."}], "task": "refering", "answer_template": "The \"green umbrella and its pole\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 65, 88, 111, 134, 156, 157, 179, 180], "bbox": [0.559125, 0.0033750000000000004, 1.0, 0.4763541666666667], "iscrowd": 0, "area": 12117.137750000002, "rle": {"size": [480, 640], "counts": "hjW52l>2N2N3L3N3M2N2N3M2N3O00001O001O00001O00001O001O0000001O000000001O000000001O00000000001O000000001O0000000000001O0000000000001O0000000000001O0000000000001O0O100000001O0000000000001O0000000000001O000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000000000000O100000000000000O1000lGoNY2Q1ZM]Oe2c0oLIQ37bL6^3JVLb0j3]OjKQ1U4oN_K]1a4cNRKj1n4VNfJV2Z5jMZJc2e5]MnIP3R6PMbI\\3^6dLUIi3k6VLjHP4\\7PLWHg3_8g0f0ZOf0ZOf0ZOf0ZOf0ZOf0ZOf0I7000000000000000000000000000000000000000000000000000000000000O100000000O1000000O100000000O100000000O1000000O100000000O100000000001O00001O00001O0000001O00001O00001O00001O00001O0000001O0000]Oa0"}, "label": "green umbrella and its pole"}]} {"id": 213005, "image": "COCO_train2014_000000213005.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small lamb lying closest to the adult\"."}], "task": "refering", "answer_template": "The \"a small lamb lying closest to the adult\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 155, 156, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225], "bbox": [0.49959375, 0.46452777777777776, 0.7947656249999999, 0.7634444444444444], "iscrowd": 0, "area": 13722.29645, "rle": {"size": [360, 640], "counts": "ig`37o:4L2O2M3N1O2N2N1O2N101O1N101N101N1O2N2O0O2N2N100O2O0O101N1O100O100O1O100O100O100O100O100O10000O2O0O10000O10000O2O000O10000O100000000000000O1000000000000000000000000O100000000O10000O10000O10000QOSNbHn1]7VN`Hj1`7XN^Hh1b7o0O100O100O100O100O1000000000000000000000000000000O10000000000000000000000001O00000O100000000000000O101O000000000O2O1O001O001O0O2O1O001O001O0O2O1hMjG\\1V8dNlGZ1U8eNmGY1T8eNoGY1Q8gNQHW1P8hNRHV1n7jNUHR1m7lNXHP1i7oNZHn0f7RO^Hi0d7UO`Hh0a7WObHe0_7[OdHb0S9N3L4E;FUj]1"}, "label": "a small lamb lying closest to the adult"}]} {"id": 213005, "image": "COCO_train2014_000000213005.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lamb looking up at a sheep\"."}], "task": "refering", "answer_template": "The \"a lamb looking up at a sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 155, 156, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225], "bbox": [0.49959375, 0.46452777777777776, 0.7947656249999999, 0.7634444444444444], "iscrowd": 0, "area": 13722.29645, "rle": {"size": [360, 640], "counts": "ig`37o:4L2O2M3N1O2N2N1O2N101O1N101N101N1O2N2O0O2N2N100O2O0O101N1O100O100O1O100O100O100O100O100O10000O2O0O10000O10000O2O000O10000O100000000000000O1000000000000000000000000O100000000O10000O10000O10000QOSNbHn1]7VN`Hj1`7XN^Hh1b7o0O100O100O100O100O1000000000000000000000000000000O10000000000000000000000001O00000O100000000000000O101O000000000O2O1O001O001O0O2O1O001O001O0O2O1hMjG\\1V8dNlGZ1U8eNmGY1T8eNoGY1Q8gNQHW1P8hNRHV1n7jNUHR1m7lNXHP1i7oNZHn0f7RO^Hi0d7UO`Hh0a7WObHe0_7[OdHb0S9N3L4E;FUj]1"}, "label": "a lamb looking up at a sheep"}]} {"id": 213005, "image": "COCO_train2014_000000213005.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sheep laying in grass\"."}], "task": "refering", "answer_template": "The \"sheep laying in grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 61, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 149, 150, 151, 152, 155, 156, 157, 158, 172, 173, 174, 175, 178, 179, 180, 181, 182, 195, 196, 202, 203, 204, 205], "bbox": [0.454515625, 0.17277777777777778, 0.9274999999999999, 0.6754722222222223], "iscrowd": 0, "area": 28495.89385, "rle": {"size": [360, 640], "counts": "h]V3:l:3M4L3M4M2N3M3M2N2O1N2N2N2N3N0O001O10O01O1O001O1hHgN`4Y1`KjN]4W1aKkN^4V1aKkN^4W1`KiNT4f1hK\\NU4i1hKXNW4k1gKUNX4o1eKPN[4U2`KlM_4X2^KhMb4\\2ZKdMe4a2VK_Mi4e2UK[Mi4i2UKVMj4n2TKQMk4T3WJ_L?=Y5X3SJ_La0;[5b3`JaLa5`3ZJcLg5^3TJeLn5Z40O1O2N100O1O100O00100O1O010O100O00100O00100O1O010O100O00100O1000000000O10000000000001O0010OO101N2N101N100O1O100O1O100O10001O001O00000000000O10000000003M5K0000OnNS1M3O1dMeGj1\\8TNhGj1X8UNkGi1U8VNnGh1Q8WNSHg1m7XNVHe1k7ZNXHd1h7ZNZHf1f7YN[Hg1e7WN]Hh1b800000001N1000000O100O10000O10000000010O0001O00001O00001O000010O0001O00010O000010O01O1O010O010O0010O01O010O10O01O010O010O1O100O0WGRNY8n1fGSNY8n1fGSNY8n1fGTNY8l1fGUNY8k1fGWNZ8i1eGYNY8h1fG\\NV8e1iG`NS8`1lGjNj7W1UHoNe7R1ZHSOb7m0\\HXO`7i0_HXOa7h0^HXOb7W2001N101O00001O001O00001N101O001O001O00001N101O001O001O1O001O1N101O001O1O001O1O0O2O1O001O001O1O00100O001O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O001O2N2N3L3N3M2M4M2M4M2N6I7J7Hd0]OWn?"}, "label": "sheep laying in grass"}]} {"id": 213005, "image": "COCO_train2014_000000213005.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mother sheep with three of her kids\"."}], "task": "refering", "answer_template": "The \"a mother sheep with three of her kids\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 61, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 149, 150, 151, 152, 155, 156, 157, 158, 172, 173, 174, 175, 178, 179, 180, 181, 182, 195, 196, 202, 203, 204, 205], "bbox": [0.454515625, 0.17277777777777778, 0.9274999999999999, 0.6754722222222223], "iscrowd": 0, "area": 28495.89385, "rle": {"size": [360, 640], "counts": "h]V3:l:3M4L3M4M2N3M3M2N2O1N2N2N2N3N0O001O10O01O1O001O1hHgN`4Y1`KjN]4W1aKkN^4V1aKkN^4W1`KiNT4f1hK\\NU4i1hKXNW4k1gKUNX4o1eKPN[4U2`KlM_4X2^KhMb4\\2ZKdMe4a2VK_Mi4e2UK[Mi4i2UKVMj4n2TKQMk4T3WJ_L?=Y5X3SJ_La0;[5b3`JaLa5`3ZJcLg5^3TJeLn5Z40O1O2N100O1O100O00100O1O010O100O00100O00100O1O010O100O00100O1000000000O10000000000001O0010OO101N2N101N100O1O100O1O100O10001O001O00000000000O10000000003M5K0000OnNS1M3O1dMeGj1\\8TNhGj1X8UNkGi1U8VNnGh1Q8WNSHg1m7XNVHe1k7ZNXHd1h7ZNZHf1f7YN[Hg1e7WN]Hh1b800000001N1000000O100O10000O10000000010O0001O00001O00001O000010O0001O00010O000010O01O1O010O010O0010O01O010O10O01O010O010O1O100O0WGRNY8n1fGSNY8n1fGSNY8n1fGTNY8l1fGUNY8k1fGWNZ8i1eGYNY8h1fG\\NV8e1iG`NS8`1lGjNj7W1UHoNe7R1ZHSOb7m0\\HXO`7i0_HXOa7h0^HXOb7W2001N101O00001O001O00001N101O001O001O00001N101O001O001O1O001O1N101O001O1O001O1O0O2O1O001O001O1O00100O001O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O001O2N2N3L3N3M2M4M2M4M2N6I7J7Hd0]OWn?"}, "label": "a mother sheep with three of her kids"}]} {"id": 213005, "image": "COCO_train2014_000000213005.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"one of three lambs sitting near their mom . it has two visible light teal spots on its back\"."}], "task": "refering", "answer_template": "The \"one of three lambs sitting near their mom . it has two visible light teal spots on its back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 214, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 243], "bbox": [0.184015625, 0.4523611111111111, 0.609109375, 0.8260000000000001], "iscrowd": 0, "area": 16976.500300000003, "rle": {"size": [360, 640], "counts": "^gY1141k:8M2O1N3M2N3N1N2O1O100O101N100O1O100O100O100O1O100O100O100O2O0O100O100O100O100O100O101N100O100O10000O2O0O100O100O10001N100O100O2O01O000001O01O0000010O00000010O000000010O000001O0001O00000000001O000001O0001O00000000001O0000000O1O2N1O1O100iNPNUIQ2h6SNVIo1g6UNVIl1i6VNUIk1k6VNSIk1m6WNPIj1P7WNnHj1Q7YNlHh1T7YNjHh1V7ZNgHh1W7[NeHg1[7ZNcHg1]7[N`Hf1`7Q1O10001N10000000010O00000000010O0000000000010O00000000000O101O0000000O1000001O000O100000000O2O00000000000001O01O0001O000_NfHTO^7d0iHYO]7=iHA[7:hHDY7:kHDT7;oHCR7;RIBo6 in this image.", "objects": [{"patches": [125, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 214, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 243], "bbox": [0.184015625, 0.4523611111111111, 0.609109375, 0.8260000000000001], "iscrowd": 0, "area": 16976.500300000003, "rle": {"size": [360, 640], "counts": "^gY1141k:8M2O1N3M2N3N1N2O1O100O101N100O1O100O100O100O1O100O100O100O2O0O100O100O100O100O100O101N100O100O10000O2O0O100O100O10001N100O100O2O01O000001O01O0000010O00000010O000000010O000001O0001O00000000001O000001O0001O00000000001O0000000O1O2N1O1O100iNPNUIQ2h6SNVIo1g6UNVIl1i6VNUIk1k6VNSIk1m6WNPIj1P7WNnHj1Q7YNlHh1T7YNjHh1V7ZNgHh1W7[NeHg1[7ZNcHg1]7[N`Hf1`7Q1O10001N10000000010O00000000010O0000000000010O00000000000O101O0000000O1000001O000O100000000O2O00000000000001O01O0001O000_NfHTO^7d0iHYO]7=iHA[7:hHDY7:kHDT7;oHCR7;RIBo6 in this image.", "objects": [{"patches": [168, 169, 170, 171, 190, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 260, 261, 262, 263], "bbox": [0.27168749999999997, 0.46421545667447306, 0.489265625, 0.7918735362997658], "iscrowd": 0, "area": 15392.5828, "rle": {"size": [427, 640], "counts": "fiX25S=7H7K5K6J5K5L5K4K4M2N3M3M2M4M3M2N3M3M2N3M3M3M2ZFWMg8l2UGVMj8m2RGUMl8o2PGSMo8R3kFPMj8_3PGcLP9l3O1000000O1000000O1000000000O10000000000000000000000000O1000000000000000000000000000000000O100000000000000000001O00000000000000000000000000000000001O0O10000000001O0SLmFb3^901O0AaFRM_9l2eFRM\\9k2gFTMY9j2kFTMU9j2nFUMS9g2QGXMo8f2UGXMk8f2XGYMi8d2[GZMf8c2S1L4M3M3L4M3M3L5L4K4M4K5L3L5L7HoSX4"}, "label": "a cream colored urn with burgundy floral designs"}]} {"id": 368659, "image": "COCO_train2014_000000368659.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and red vas on the wood crate\"."}], "task": "refering", "answer_template": "The \"a white and red vas on the wood crate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 170, 171, 190, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 260, 261, 262, 263], "bbox": [0.27168749999999997, 0.46421545667447306, 0.489265625, 0.7918735362997658], "iscrowd": 0, "area": 15392.5828, "rle": {"size": [427, 640], "counts": "fiX25S=7H7K5K6J5K5L5K4K4M2N3M3M2M4M3M2N3M3M2N3M3M3M2ZFWMg8l2UGVMj8m2RGUMl8o2PGSMo8R3kFPMj8_3PGcLP9l3O1000000O1000000O1000000000O10000000000000000000000000O1000000000000000000000000000000000O100000000000000000001O00000000000000000000000000000000001O0O10000000001O0SLmFb3^901O0AaFRM_9l2eFRM\\9k2gFTMY9j2kFTMU9j2nFUMS9g2QGXMo8f2UGXMk8f2XGYMi8d2[GZMf8c2S1L4M3M3L4M3M3L5L4K4M4K5L3L5L7HoSX4"}, "label": "a white and red vas on the wood crate"}]} {"id": 278549, "image": "COCO_train2014_000000278549.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child is holding a piece of bread\"."}], "task": "refering", "answer_template": "The \"a child is holding a piece of bread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 59, 78, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 194, 195, 196, 197, 198, 212, 213, 214, 215, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 249, 250, 251, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 272, 273, 274, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 294, 295, 296, 297, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.215125, 0.1454225352112676, 0.9540156249999999, 0.9838262910798122], "iscrowd": 0, "area": 80443.13884999997, "rle": {"size": [426, 640], "counts": "ghi1:o<1N2O2M2O1O2M2O2N1N2O2N1N2O2M2O1O2M2O1O2M2O2jMjNUHW1j7TOkGm0T8WOhGk0V8XOgGi0W8\\OdGf0[8]ObGe0\\8^OaGc0^8@_Ga0_8C^G?a8D[G=d8FYG;f8HWG:h80lF2S9S2O1O100O2N100O100O100O2O0O100O100O100O101N1O100O100O100O101N100O100O100O100O2O0M3N2N2M3N2N2N2M4M2N2M3N2N2N2M3N3M2N2N2M3L4M3L4L4N3M2M3N2N2M3N2N2M3N3L3M3M3M3M3N2N2M3N3M2M3N2N2M3M3M3M3M4L3M3L4N2M3M3M3N3N100O100O100O2N100O2O0O2O0O1O2O0O2O000O2O0O2O0O101O0O101N10000O2O000O10000O10000O101O0O10000O100000001O000000000000000010O000000000000000001O000001O001O00001O001O001O001O001O001O001O001O001O001O1O001O001O001O1O001O0^KZM`Ng2^1aM[N`2b1iMWNX2g1nMTNR2k1PNTNQ2i1QNWNP2g1RNXNn1g1TNXNm1e1VNZNk1d1WN[Ni1d1YN[Nh1b1[N]Nf1a1\\N^Nd1`1^N`Nc1[1bNdN_1X1dNhN\\1T1hNlNY1P1jNPOW1k0mNUOS1h0QOWOP1d0TO\\Om0?WOAi0:\\OFf00B0`0FH::\\OOc03SO5m0MkN9U1IdN<\\1G[N?e1CTNc0k1AlMd0T2_OdMf0\\2^O[Mg0e2\\ORMj0n2i401O000000001O2N2N2N1O2N2N2N2M2O2N3M2N2N3M2N2N3M2N2N2N3M2N2N3M2N2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O000000000000000O10O100000000000O10000000000000O010000000O1O10O0100O1O10O01O100O00100O10O01O100O001O100O001O1O1O001O1O10O01O1O1O001O1O1O100O1O1O001O1O1O1O100N2M3N2N2M2O2NL\\EfMa:\\2cEaM[:a251N3N100010O010O011N101N101N1O101N100O10O100O10000O10000O100O01000O100O02O0O2N1O2O0O2N1O2N101N1O2N1O2O0O2N1O2O1N2POUE^Om:;U1J5K6J[h;"}, "label": "a child is holding a piece of bread"}]} {"id": 278549, "image": "COCO_train2014_000000278549.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baby in green\"."}], "task": "refering", "answer_template": "The \"the baby in green\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 59, 78, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 190, 191, 192, 193, 194, 195, 196, 197, 198, 212, 213, 214, 215, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 249, 250, 251, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 272, 273, 274, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 294, 295, 296, 297, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.215125, 0.1454225352112676, 0.9540156249999999, 0.9838262910798122], "iscrowd": 0, "area": 80443.13884999997, "rle": {"size": [426, 640], "counts": "ghi1:o<1N2O2M2O1O2M2O2N1N2O2N1N2O2M2O1O2M2O1O2M2O2jMjNUHW1j7TOkGm0T8WOhGk0V8XOgGi0W8\\OdGf0[8]ObGe0\\8^OaGc0^8@_Ga0_8C^G?a8D[G=d8FYG;f8HWG:h80lF2S9S2O1O100O2N100O100O100O2O0O100O100O100O101N1O100O100O100O101N100O100O100O100O2O0M3N2N2M3N2N2N2M4M2N2M3N2N2N2M3N3M2N2N2M3L4M3L4L4N3M2M3N2N2M3N2N2M3N3L3M3M3M3M3N2N2M3N3M2M3N2N2M3M3M3M3M4L3M3L4N2M3M3M3N3N100O100O100O2N100O2O0O2O0O1O2O0O2O000O2O0O2O0O101O0O101N10000O2O000O10000O10000O101O0O10000O100000001O000000000000000010O000000000000000001O000001O001O00001O001O001O001O001O001O001O001O001O001O1O001O001O001O1O001O0^KZM`Ng2^1aM[N`2b1iMWNX2g1nMTNR2k1PNTNQ2i1QNWNP2g1RNXNn1g1TNXNm1e1VNZNk1d1WN[Ni1d1YN[Nh1b1[N]Nf1a1\\N^Nd1`1^N`Nc1[1bNdN_1X1dNhN\\1T1hNlNY1P1jNPOW1k0mNUOS1h0QOWOP1d0TO\\Om0?WOAi0:\\OFf00B0`0FH::\\OOc03SO5m0MkN9U1IdN<\\1G[N?e1CTNc0k1AlMd0T2_OdMf0\\2^O[Mg0e2\\ORMj0n2i401O000000001O2N2N2N1O2N2N2N2M2O2N3M2N2N3M2N2N3M2N2N2N3M2N2N3M2N2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O000000000000000O10O100000000000O10000000000000O010000000O1O10O0100O1O10O01O100O00100O10O01O100O001O100O001O1O1O001O1O10O01O1O1O001O1O1O100O1O1O001O1O1O1O100N2M3N2N2M2O2NL\\EfMa:\\2cEaM[:a251N3N100010O010O011N101N101N1O101N100O10O100O10000O10000O100O01000O100O02O0O2N1O2O0O2N1O2N101N1O2N1O2O0O2N1O2O1N2POUE^Om:;U1J5K6J[h;"}, "label": "the baby in green"}]} {"id": 278549, "image": "COCO_train2014_000000278549.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little boy putting finger into the mouth\"."}], "task": "refering", "answer_template": "The \"a little boy putting finger into the mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 69, 70, 92, 93, 115, 116, 117, 138, 139, 140, 141, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 301, 302, 303], "bbox": [0.0, 0.16342723004694837, 0.272578125, 0.9003755868544601], "iscrowd": 0, "area": 33004.3616, "rle": {"size": [426, 640], "counts": "V2o8[401O00100O001O010O001O00101N1O1O1O101N1O001O100O1O3M3M3N2M3M3M1O100O2N1O1O100O1O2N1O2O1N2N3M2N3N1N1O2N1O101N1O2N2N101N3M4L4Le0\\OO000000001O01O0000010O001O00010O001O00010O001O0001O00000010O00002N100O1O2N1O100O1O1O00100O1O1O1O10O01O001O01O0O100O1O2O000000001O0000000000O2N1N3N1N3N2N2N2O1N3N1N3N2M3N2L4L4M3K4M4K5J6I7K5L4M3M3M3XN_Eb0e:XOaEd0c:VOcEf0b:SOdEi0`:ROeEh0a:SOdEh0a;M4L4L3M3M3M5KmaQ6"}, "label": "a little boy putting finger into the mouth"}]} {"id": 278549, "image": "COCO_train2014_000000278549.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy licking his fingers\"."}], "task": "refering", "answer_template": "The \"a boy licking his fingers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 69, 70, 92, 93, 115, 116, 117, 138, 139, 140, 141, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 301, 302, 303], "bbox": [0.0, 0.16342723004694837, 0.272578125, 0.9003755868544601], "iscrowd": 0, "area": 33004.3616, "rle": {"size": [426, 640], "counts": "V2o8[401O00100O001O010O001O00101N1O1O1O101N1O001O100O1O3M3M3N2M3M3M1O100O2N1O1O100O1O2N1O2O1N2N3M2N3N1N1O2N1O101N1O2N2N101N3M4L4Le0\\OO000000001O01O0000010O001O00010O001O00010O001O0001O00000010O00002N100O1O2N1O100O1O1O00100O1O1O1O10O01O001O01O0O100O1O2O000000001O0000000000O2N1N3N1N3N2N2N2O1N3N1N3N2M3N2L4L4M3K4M4K5J6I7K5L4M3M3M3XN_Eb0e:XOaEd0c:VOcEf0b:SOdEi0`:ROeEh0a:SOdEh0a;M4L4L3M3M3M5KmaQ6"}, "label": "a boy licking his fingers"}]} {"id": 32801, "image": "COCO_train2014_000000032801.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a worker in a black ballcap reaches over the machinery\"."}], "task": "refering", "answer_template": "The \"a worker in a black ballcap reaches over the machinery\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 41, 42, 43, 57, 58, 59, 60, 61, 77, 78, 79, 97, 98], "bbox": [0.19047999999999998, 0.12901333333333334, 0.4521, 0.43448000000000003], "iscrowd": 0, "area": 4984.3031, "rle": {"size": [375, 500], "counts": "[lR14b;4L3M2OO01O0010O10000000000O100000000O10000000O01000O10O0100O10O010000O0100000O10O1000O100000001N101O00001O0O10001O0000001]O\\OPFe0n9^OQFc0k9@UF`0k9@UFa0i9AVF?i9BWFg0_9\\O_Fn0W9SOgFW1o8i0O1O100O1O100O001000000001O00001O00001O00001O003N3L3M:F3M001O0010O010O1O010O010O1O1000TOZFFf9=eFUO[9m0h01N2N3M2O1iEPOb9d1O12M8H:F6I7I7I8H:ERVT3"}, "label": "a worker in a black ballcap reaches over the machinery"}]} {"id": 32801, "image": "COCO_train2014_000000032801.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black hat\"."}], "task": "refering", "answer_template": "The \"a man in a black hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 41, 42, 43, 57, 58, 59, 60, 61, 77, 78, 79, 97, 98], "bbox": [0.19047999999999998, 0.12901333333333334, 0.4521, 0.43448000000000003], "iscrowd": 0, "area": 4984.3031, "rle": {"size": [375, 500], "counts": "[lR14b;4L3M2OO01O0010O10000000000O100000000O10000000O01000O10O0100O10O010000O0100000O10O1000O100000001N101O00001O0O10001O0000001]O\\OPFe0n9^OQFc0k9@UF`0k9@UFa0i9AVF?i9BWFg0_9\\O_Fn0W9SOgFW1o8i0O1O100O1O100O001000000001O00001O00001O00001O003N3L3M:F3M001O0010O010O1O010O010O1O1000TOZFFf9=eFUO[9m0h01N2N3M2O1iEPOb9d1O12M8H:F6I7I7I8H:ERVT3"}, "label": "a man in a black hat"}]} {"id": 581670, "image": "COCO_train2014_000000581670.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a sandwich near a table\"."}], "task": "refering", "answer_template": "The \"a man holding a sandwich near a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 36, 37, 54, 55, 72, 73, 75, 76, 90, 91, 92, 93, 108, 109, 110], "bbox": [0.0, 0.06493333333333334, 0.23766, 0.5454666666666667], "iscrowd": 0, "area": 7876.174999999998, "rle": {"size": [375, 500], "counts": "i0Z1^:e2[MZ1gNN1O100O1O2N1O100O1O2N1O100O2N1O1O100O2N3UMnHk0V7oNmHQ1U7iNPIU1T7eNoHZ1T7`NPI_1S7\\NPIc1T7WNPIh1R7RNRIm1R7mMQIR2R7hMRIW2Y8N100O1O2N1O100O1O2N010N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O00O1O100O1O1O1O100O1O1O1O1000000001O00001O00001O0000001O00001O0[OXE in this image.", "objects": [{"patches": [18, 36, 37, 54, 55, 72, 73, 75, 76, 90, 91, 92, 93, 108, 109, 110], "bbox": [0.0, 0.06493333333333334, 0.23766, 0.5454666666666667], "iscrowd": 0, "area": 7876.174999999998, "rle": {"size": [375, 500], "counts": "i0Z1^:e2[MZ1gNN1O100O1O2N1O100O1O2N1O100O2N1O1O100O2N3UMnHk0V7oNmHQ1U7iNPIU1T7eNoHZ1T7`NPI_1S7\\NPIc1T7WNPIh1R7RNRIm1R7mMQIR2R7hMRIW2Y8N100O1O2N1O100O1O2N010N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O00O1O100O1O1O1O100O1O1O1O1000000001O00001O00001O0000001O00001O0[OXE in this image.", "objects": [{"patches": [0, 5, 6, 7, 18, 19, 23, 24, 37, 38, 39, 40, 54, 55, 56, 57, 58, 73, 74, 75, 92], "bbox": [0.00684, 0.006853333333333333, 0.44177999999999995, 0.39498666666666665], "iscrowd": 0, "area": 7525.168199999999, "rle": {"size": [375, 500], "counts": "cS11d;kGEU8=fGF[8:`GJ`88[GLe8c10010O01N1O2O0O2N101N1O2O0O2N1O2O0O2N101N1O2O0O2N1O2O0O2N101N1O2ON10001O0O10N2O1N1O2N2N2N101N2N2O001O1O00100O1O001O1O10O01O1O1O00100O1O00100O100O010O100O01000O101N2O1N2O1N2O1N3N1N2O1N2O1N2O1N2N2O1N2O0O2ObZ4O_eK000O1O2N1O101N1O1O1O2N1O1O101N1O101O00001O001O000010O01O00001O001O00N2L4L5L3000000001O0000000010O000001O000001O01O01O00001O0O1O2N1O1O1O2Nb`V3"}, "label": "a hand holding a hamburg"}]} {"id": 581670, "image": "COCO_train2014_000000581670.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man ' s hand grabbing food\"."}], "task": "refering", "answer_template": "The \"a man ' s hand grabbing food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 5, 6, 7, 18, 19, 23, 24, 37, 38, 39, 40, 54, 55, 56, 57, 58, 73, 74, 75, 92], "bbox": [0.00684, 0.006853333333333333, 0.44177999999999995, 0.39498666666666665], "iscrowd": 0, "area": 7525.168199999999, "rle": {"size": [375, 500], "counts": "cS11d;kGEU8=fGF[8:`GJ`88[GLe8c10010O01N1O2O0O2N101N1O2O0O2N1O2O0O2N101N1O2O0O2N1O2O0O2N101N1O2ON10001O0O10N2O1N1O2N2N2N101N2N2O001O1O00100O1O001O1O10O01O1O1O00100O1O00100O100O010O100O01000O101N2O1N2O1N2O1N3N1N2O1N2O1N2O1N2N2O1N2O0O2ObZ4O_eK000O1O2N1O101N1O1O1O2N1O1O101N1O101O00001O001O000010O01O00001O001O00N2L4L5L3000000001O0000000010O000001O000001O01O01O00001O0O1O2N1O1O1O2Nb`V3"}, "label": "a man ' s hand grabbing food"}]} {"id": 581670, "image": "COCO_train2014_000000581670.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the large sandwich that is on the plate next to the knife\"."}], "task": "refering", "answer_template": "The \"the large sandwich that is on the plate next to the knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 80, 81, 82, 83, 84, 85, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 147, 148, 149, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 204], "bbox": [0.19506, 0.2689333333333333, 0.7853199999999999, 0.8676], "iscrowd": 0, "area": 36414.773550000005, "rle": {"size": [375, 500], "counts": "lTT13a;4L3M3M4L3O1N3N1N2O2N1N2O2M2OO0O2N1OaNYOSHe0o7GdG8^8L^G3c81XG2f82VGOi84SGOk85QGMn86nFKQ9_1N2N1O2O1O001N2O001N2O2N1N3N1O1O2M2O2N1N3N1O2N1N3N1O1N3N1O2N1N3N1O2M2O2N1O1N3N1O2M2O2N1O1N101O1N2O1O001N2O000O0100O10O0100O010O1000O0100O010O100O010O10O0100O010O010O010O01O10O010O010O01O010O010O010O0010O010O010O01O10O010O010O01O010O010O010O0100O010O10O01O10O010O01O10O0100O00100O010O1O010O10O01O010O10O01O10O0100O00100O00100O010O00100O010O1O010O10O01O10O0100O0010O0100O00100O010O1O010O1O010O1O001O00100O001O1O010O1O1O1O2O0O1O2N1O101N1O1O2N100O2N1O1O101N1N2O2N2N2N3M2N2N2N2N2N2N3gNTGBn8=RGBP9=QGAQ9=PGAS9>nF@T9?mF_OW9=jFBZ9;gFC]9:cFEb96`FHc96^FHd97\\FHe97]FGe98[FGf99[FEg9:ZFDg9;ZFDg9XF@j9?WF_Ok9`0UF_Om9?f001O1N2O0L5H]QW1"}, "label": "the large sandwich that is on the plate next to the knife"}]} {"id": 581670, "image": "COCO_train2014_000000581670.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich on a plate next to a cutting knife\"."}], "task": "refering", "answer_template": "The \"a sandwich on a plate next to a cutting knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 80, 81, 82, 83, 84, 85, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 147, 148, 149, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 204], "bbox": [0.19506, 0.2689333333333333, 0.7853199999999999, 0.8676], "iscrowd": 0, "area": 36414.773550000005, "rle": {"size": [375, 500], "counts": "lTT13a;4L3M3M4L3O1N3N1N2O2N1N2O2M2OO0O2N1OaNYOSHe0o7GdG8^8L^G3c81XG2f82VGOi84SGOk85QGMn86nFKQ9_1N2N1O2O1O001N2O001N2O2N1N3N1O1O2M2O2N1N3N1O2N1N3N1O1N3N1O2N1N3N1O2M2O2N1O1N3N1O2M2O2N1O1N101O1N2O1O001N2O000O0100O10O0100O010O1000O0100O010O100O010O10O0100O010O010O010O01O10O010O010O01O010O010O010O0010O010O010O01O10O010O010O01O010O010O010O0100O010O10O01O10O010O01O10O0100O00100O010O1O010O10O01O010O10O01O10O0100O00100O00100O010O00100O010O1O010O10O01O10O0100O0010O0100O00100O010O1O010O1O010O1O001O00100O001O1O010O1O1O1O2O0O1O2N1O101N1O1O2N100O2N1O1O101N1N2O2N2N2N3M2N2N2N2N2N2N3gNTGBn8=RGBP9=QGAQ9=PGAS9>nF@T9?mF_OW9=jFBZ9;gFC]9:cFEb96`FHc96^FHd97\\FHe97]FGe98[FGf99[FEg9:ZFDg9;ZFDg9XF@j9?WF_Ok9`0UF_Om9?f001O1N2O0L5H]QW1"}, "label": "a sandwich on a plate next to a cutting knife"}]} {"id": 229415, "image": "COCO_train2014_000000229415.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black airplane\"."}], "task": "refering", "answer_template": "The \"a black airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 152, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 220, 221, 244], "bbox": [0.321625, 0.42601873536299767, 0.68065625, 0.6860889929742389], "iscrowd": 0, "area": 10219.48615, "rle": {"size": [427, 640], "counts": "TTf22V=3M4M2M3M4M2M3M4M2M4L3N2KkNQDY1m;301N1O1O2N100010O100O1O100O1O100O10000;EM3M3M3N2M3N2O1O1O1O1O1O1O1O1O1O1O1O000001N1000000O100000000O1000000O100000000O10001O0O1000000O100000000O1000000O100000000O101O000O100000000O100000000O1000000000000000001O0000000O100000000000000000000000000O10000000001O000000000000000O100000000000000000000000000O10001O00002N2N2N2N2oDTN_:n1[EVNg:j1TEZNl:S2001N10001O0O10oNQ13M3L4M2L5K5J6K3M2N1O1O1A?A?000000O1000001N10000O1000000O1000000O2O000O1000000O1000000Ok0VOajd2"}, "label": "a black airplane"}]} {"id": 229415, "image": "COCO_train2014_000000229415.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a toy plane flying\"."}], "task": "refering", "answer_template": "The \"a toy plane flying\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 152, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 220, 221, 244], "bbox": [0.321625, 0.42601873536299767, 0.68065625, 0.6860889929742389], "iscrowd": 0, "area": 10219.48615, "rle": {"size": [427, 640], "counts": "TTf22V=3M4M2M3M4M2M3M4M2M4L3N2KkNQDY1m;301N1O1O2N100010O100O1O100O1O100O10000;EM3M3M3N2M3N2O1O1O1O1O1O1O1O1O1O1O1O000001N1000000O100000000O1000000O100000000O10001O0O1000000O100000000O1000000O100000000O101O000O100000000O100000000O1000000000000000001O0000000O100000000000000000000000000O10000000001O000000000000000O100000000000000000000000000O10001O00002N2N2N2N2oDTN_:n1[EVNg:j1TEZNl:S2001N10001O0O10oNQ13M3L4M2L5K5J6K3M2N1O1O1A?A?000000O1000001N10000O1000000O1000000O2O000O1000000O1000000Ok0VOajd2"}, "label": "a toy plane flying"}]} {"id": 385066, "image": "COCO_train2014_000000385066.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vehicle in front of the red car\"."}], "task": "refering", "answer_template": "The \"the vehicle in front of the red car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 100, 101, 119, 122, 136, 137, 138, 139, 140, 155], "bbox": [0.47122, 0.43434628975265016, 0.78924, 0.8177385159010602], "iscrowd": 0, "area": 2467.8676, "rle": {"size": [283, 500], "counts": "Q[Q21j82N2N2M2O2N2N2N1OM3M3001O1O001O1O0000O100O100O100000O1000O1000000001O0000001O00001O1O001O1O1O1O001O2N2N2NO10000O101O0O10000O101O001OjIDm3 in this image.", "objects": [{"patches": [80, 81, 82, 83, 100, 101, 119, 122, 136, 137, 138, 139, 140, 155], "bbox": [0.47122, 0.43434628975265016, 0.78924, 0.8177385159010602], "iscrowd": 0, "area": 2467.8676, "rle": {"size": [283, 500], "counts": "Q[Q21j82N2N2M2O2N2N2N1OM3M3001O1O001O1O0000O100O100O100000O1000O1000000001O0000001O00001O1O001O1O1O1O001O2N2N2NO10000O101O0O10000O101O001OjIDm3 in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 23, 36, 37, 38, 39, 40, 41, 42, 43, 54, 55, 56, 57, 58, 59, 60, 61, 72, 73, 74, 75, 76, 77, 78, 79, 80, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 128, 129, 130, 131, 132, 133, 134, 135, 146, 147, 148, 149, 152, 153], "bbox": [0.0, 0.10752650176678445, 0.54044, 0.901060070671378], "iscrowd": 0, "area": 42904.25749999999, "rle": {"size": [283, 500], "counts": "b1T1g7V2iMf0[O0O010O2O000O100O100O10000O100O10000O1000000O2O000O10000O1000000O100O10000O100O10001N10000O1000000000000O1000O1000000000000OdLbK]2^4]MkK`2U4ZMTLc2l3XM\\Le2d3XMbLe2^3XMhLe2X3YMlLe2S3YMQMf2d4N1O2N1O2N1O1O2N1O1O1O1O1O1O1O1O001O1O1O00001O00001O000O2O000001O010O001O010O001O001O001O001O001O001O001OO1O1O100O1O100O1O010O1O001001N2O0O2O1N2O0O2O1N2O0O2O1N2O0O2N2N2N1O2N2N2O0N3K5K5M2O2O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O001O2N1O2N2N101N2N2N1O2N2NmJ\\LP5c3PK^LQ5a3nJ`LS5^3mJdLS5Z3mJgLS5a31O10O01N2N101N1O2N1O2O0O2N2N1O2N101N1O2M2O2M2O1N2N2O1N2dMQJj1b6M4M3K3M4L3M4L7I6J4L4K5Lh[o1"}, "label": "red color car"}]} {"id": 385066, "image": "COCO_train2014_000000385066.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red color car is on the road\"."}], "task": "refering", "answer_template": "The \"a red color car is on the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 23, 36, 37, 38, 39, 40, 41, 42, 43, 54, 55, 56, 57, 58, 59, 60, 61, 72, 73, 74, 75, 76, 77, 78, 79, 80, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 128, 129, 130, 131, 132, 133, 134, 135, 146, 147, 148, 149, 152, 153], "bbox": [0.0, 0.10752650176678445, 0.54044, 0.901060070671378], "iscrowd": 0, "area": 42904.25749999999, "rle": {"size": [283, 500], "counts": "b1T1g7V2iMf0[O0O010O2O000O100O100O10000O100O10000O1000000O2O000O10000O1000000O100O10000O100O10001N10000O1000000000000O1000O1000000000000OdLbK]2^4]MkK`2U4ZMTLc2l3XM\\Le2d3XMbLe2^3XMhLe2X3YMlLe2S3YMQMf2d4N1O2N1O2N1O1O2N1O1O1O1O1O1O1O1O001O1O1O00001O00001O000O2O000001O010O001O010O001O001O001O001O001O001O001OO1O1O100O1O100O1O010O1O001001N2O0O2O1N2O0O2O1N2O0O2O1N2O0O2N2N2N1O2N2N2O0N3K5K5M2O2O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O001O2N1O2N2N101N2N2N1O2N2NmJ\\LP5c3PK^LQ5a3nJ`LS5^3mJdLS5Z3mJgLS5a31O10O01N2N101N1O2N1O2O0O2N2N1O2N101N1O2M2O2M2O1N2N2O1N2dMQJj1b6M4M3K3M4L3M4L7I6J4L4K5Lh[o1"}, "label": "a red color car is on the road"}]} {"id": 401455, "image": "COCO_train2014_000000401455.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older man dressed in blue and denim speaking into a small microphone\"."}], "task": "refering", "answer_template": "The \"an older man dressed in blue and denim speaking into a small microphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 26, 27, 28, 29, 49, 50, 51, 52, 72, 73, 74, 75, 76, 96, 97, 98, 119, 120, 121, 142, 143, 144, 165, 166, 167, 188, 189, 190, 211], "bbox": [0.144625, 0.0080625, 0.31109375000000006, 0.5686041666666667], "iscrowd": 0, "area": 16001.334499999995, "rle": {"size": [480, 640], "counts": "Ye[1c0Z>5L3M4K4M4L2N1O1O1O1O1O100001O0000000013L9G8oNUN^DW2R;b1XOh0\\Ol1[G]IU5g7]IcHc6n7O1O1O1O1O00O101N100O10000O100O2O1L5K4L4L5nNQIlIS7i5X1F;D;E;F;D;1O2N3M2N2N2N2N2N3M2N2M3N2N3M2N2M3N2N3M2N2N2M3N2N3M2N2M3NO1O1O1N101O1N2O1ULaGS1`8eNdH=`7[OaIIa62^JROd5i0^4K4K6J5K6K4K6J5K6J_d^6"}, "label": "an older man dressed in blue and denim speaking into a small microphone"}]} {"id": 401455, "image": "COCO_train2014_000000401455.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a husky old man wearing a blue shirt and jeans\"."}], "task": "refering", "answer_template": "The \"a husky old man wearing a blue shirt and jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 26, 27, 28, 29, 49, 50, 51, 52, 72, 73, 74, 75, 76, 96, 97, 98, 119, 120, 121, 142, 143, 144, 165, 166, 167, 188, 189, 190, 211], "bbox": [0.144625, 0.0080625, 0.31109375000000006, 0.5686041666666667], "iscrowd": 0, "area": 16001.334499999995, "rle": {"size": [480, 640], "counts": "Ye[1c0Z>5L3M4K4M4L2N1O1O1O1O1O100001O0000000013L9G8oNUN^DW2R;b1XOh0\\Ol1[G]IU5g7]IcHc6n7O1O1O1O1O00O101N100O10000O100O2O1L5K4L4L5nNQIlIS7i5X1F;D;E;F;D;1O2N3M2N2N2N2N2N3M2N2M3N2N3M2N2M3N2N3M2N2N2M3N2N3M2N2M3NO1O1O1N101O1N2O1ULaGS1`8eNdH=`7[OaIIa62^JROd5i0^4K4K6J5K6K4K6J5K6J_d^6"}, "label": "a husky old man wearing a blue shirt and jeans"}]} {"id": 401455, "image": "COCO_train2014_000000401455.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a red and black coat sitting and watching people work on a project\"."}], "task": "refering", "answer_template": "The \"a woman wearing a red and black coat sitting and watching people work on a project\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 202, 223, 224, 225, 246, 247, 269, 292, 293, 315, 316, 337, 338, 359, 360, 361, 382, 383], "bbox": [0.60403125, 0.494625, 0.8254999999999999, 0.9890208333333333], "iscrowd": 0, "area": 6216.4582500000015, "rle": {"size": [480, 640], "counts": "hke51n>1O1O1O1O1O1O1O1O1O1O1O1O1O100O1N2O1N3N1O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1O1O2M2O1O1O1O1N2ON2M3L3N2N2N2O12N2N2N2N2NFXOP1L7O10hNgCJZ<5lCEU<;PD_OP in this image.", "objects": [{"patches": [200, 201, 202, 223, 224, 225, 246, 247, 269, 292, 293, 315, 316, 337, 338, 359, 360, 361, 382, 383], "bbox": [0.60403125, 0.494625, 0.8254999999999999, 0.9890208333333333], "iscrowd": 0, "area": 6216.4582500000015, "rle": {"size": [480, 640], "counts": "hke51n>1O1O1O1O1O1O1O1O1O1O1O1O1O100O1N2O1N3N1O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1O1O2M2O1O1O1O1N2ON2M3L3N2N2N2O12N2N2N2N2NFXOP1L7O10hNgCJZ<5lCEU<;PD_OP in this image.", "objects": [{"patches": [148, 171, 172, 173, 174, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269, 288, 289, 290, 291, 292, 311, 312], "bbox": [0.4305, 0.40358333333333335, 0.744953125, 0.787], "iscrowd": 0, "area": 18190.66855, "rle": {"size": [480, 640], "counts": "XbQ47h>4L3N3L4L9H>iCgNj9h1WEjNg:h200000001O001O001O001O00O101O001N101O0O2\\O\\DjMd;U2^DjMc;S2_DmMa;S2`DkMb;T2_DjMc;U2^DjMd;S2^DkMe;S2\\DkMg;S2YDmMi;Q2XDlMj;T2a0O2O00001O000O2O0000001O01O010O0010O010O010O010O01000O0100O1aElMP8T2jGSNU8n1dGYN[8h1]G`Nb8a1WGfNh8[1QGkNo8V1jFQOU9R1aFVO^9n0XFYOg9k0oE\\OP:j2000001O0000001O0000000001O0000001O001ZOlE^LT:b3mE\\LU:b3lE]LV:a3lE^LU:`3lE_LU:a3lE^LU:`3lE^LW:`3kE\\LY:b3hEZL]:d3`001N2N2N2O2M3N1O2M2O2N2N6I4M3M3L3N000O2O0000001O00001O0000001O00001O00000000O100000000000000O2O2N1O1O1O2N100O1O2N1O100O1O2N2N2N01O01O00000001O0000002N1N2O2eNlCFU<7nCHS<4RDJP<2SDLo;1UDMl;1WDMk;OXD0i;M[D1f;L^D2d;J`D3e;F`D7X=L2N2N3M2NPa\\2"}, "label": "the bench closest to the palm tree and on a concrete pedestal"}]} {"id": 32818, "image": "COCO_train2014_000000032818.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bench in the shore near another bench\"."}], "task": "refering", "answer_template": "The \"a bench in the shore near another bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 171, 172, 173, 174, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269, 288, 289, 290, 291, 292, 311, 312], "bbox": [0.4305, 0.40358333333333335, 0.744953125, 0.787], "iscrowd": 0, "area": 18190.66855, "rle": {"size": [480, 640], "counts": "XbQ47h>4L3N3L4L9H>iCgNj9h1WEjNg:h200000001O001O001O001O00O101O001N101O0O2\\O\\DjMd;U2^DjMc;S2_DmMa;S2`DkMb;T2_DjMc;U2^DjMd;S2^DkMe;S2\\DkMg;S2YDmMi;Q2XDlMj;T2a0O2O00001O000O2O0000001O01O010O0010O010O010O010O01000O0100O1aElMP8T2jGSNU8n1dGYN[8h1]G`Nb8a1WGfNh8[1QGkNo8V1jFQOU9R1aFVO^9n0XFYOg9k0oE\\OP:j2000001O0000001O0000000001O0000001O001ZOlE^LT:b3mE\\LU:b3lE]LV:a3lE^LU:`3lE_LU:a3lE^LU:`3lE^LW:`3kE\\LY:b3hEZL]:d3`001N2N2N2O2M3N1O2M2O2N2N6I4M3M3L3N000O2O0000001O00001O0000001O00001O00000000O100000000000000O2O2N1O1O1O2N100O1O2N1O100O1O2N2N2N01O01O00000001O0000002N1N2O2eNlCFU<7nCHS<4RDJP<2SDLo;1UDMl;1WDMk;OXD0i;M[D1f;L^D2d;J`D3e;F`D7X=L2N2N3M2NPa\\2"}, "label": "a bench in the shore near another bench"}]} {"id": 8300, "image": "COCO_train2014_000000008300.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the two giraffe necks togther eating\"."}], "task": "refering", "answer_template": "The \"the two giraffe necks togther eating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 169, 170, 171, 172, 173, 191, 192, 193, 194, 213, 214, 215, 216, 235, 236, 237, 238, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328], "bbox": [0.011984375, 0.4192941176470588, 0.539640625, 0.9858823529411764], "iscrowd": 0, "area": 24430.354250000008, "rle": {"size": [425, 640], "counts": "Yg32T=3N2M3N2N2M3N2M3N2M3N2N2M3N2M3N2N2M3N2M3N2O1O100O1O100O100O1O100O100O1O100O100O1O100O100O100O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O100O100O1O100O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1DhLfFY3Z9;O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O10N1M3N2N2M4_NZG\\Nh8d1^GTNd8k1bGmMa8S2eGeM]8[2iG]MY8b2X1O0O100O101O0O100O101O0O100O10001N100O100O2O000O100O101O0O100O101N10000O101N10000O100O2O000O100O2O0O10000O101N10000O101N100O10001N100O10000O2O0O10000O2O0O10CiNgDX1[;:1O1O010O10000O10O1001N2O1O1N2O1O1O1N2O2N1N2O1O1O1N2O1O1N101O000O10000O10001O0O10000O1000000O2O000O10000000001N100000000O1O4LdWj3"}, "label": "the two giraffe necks togther eating"}]} {"id": 8300, "image": "COCO_train2014_000000008300.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tallest giraffe on the left\"."}], "task": "refering", "answer_template": "The \"the tallest giraffe on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 101, 102, 103, 123, 124, 125, 126, 127, 145, 146, 147, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 276, 277, 278, 299, 322], "bbox": [0.0, 0.2369176470588235, 0.546578125, 1.0], "iscrowd": 0, "area": 22976.368800000004, "rle": {"size": [425, 640], "counts": "R:?i in this image.", "objects": [{"patches": [268, 269, 292, 293, 294, 315, 316, 317, 318, 339, 340, 341, 342, 343], "bbox": [0.678921875, 0.7315529411764706, 0.9626718750000001, 1.0], "iscrowd": 0, "area": 6255.54325, "rle": {"size": [425, 640], "counts": "`kd53T=2N3M4L3L4M4M201O001O001O001O001O001O001O00001N101O001O000000000000000000000000000000000000000O10000O10000O100O100_DnNj:R1QEUOl:g1NO1O001O1O001O001O1O001O001O0000001O00001O0000001O00001O00001O001O1O001O1O2N2N2N2N2N2N1O001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O00001O001O001O1O001O001O001O1O001O001O001O1O4L3M1O1O1O00`a9"}, "label": "a giraffe standing behind three other giraffes , facing the opposite direction"}]} {"id": 8300, "image": "COCO_train2014_000000008300.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the third giraffe from the left\"."}], "task": "refering", "answer_template": "The \"the third giraffe from the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 154, 155, 156, 157, 176, 177, 178, 179, 180, 199, 200, 201, 221, 222, 223, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 285, 286, 287, 288, 289, 290, 291, 308, 309, 310, 311, 312, 313, 314, 330, 331, 332, 333, 334, 335, 336], "bbox": [0.38625, 0.3559294117647059, 0.8705624999999999, 0.9868], "iscrowd": 0, "area": 25043.94614999999, "rle": {"size": [425, 640], "counts": "mlV3e0a<3M3M3M3M3M3M3M3M3M3O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1O100O1O100O100N2O1O1N2O1O1O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O100O1O100O100O1O100O1O1O1O10O01O1O1O1O100O1O1O1O1O100O1O1O1O1O1OL5L2VNfGdN^8[1gG]N\\8d1hGUN[8j1jGnMY8R2lGfMX8Y2mG`MU8a2oGWMU8h2U1010O01O0010O01O010O0010O01O0010O01O010O001O010O0010O01O0010O01O010O1O00100O0010OO2N2N1O2N2N1O2M2O2N2N1O2N3M3M3M3M3M3M3N2O100O0010E:O2N1011N2N3N1N2O1N2O1O1O0000000001O01O0000002N2N3M1O1O10O01O1O010O1O010O1O00100O00100O001O010O001O10O01O001O010O001O1N101N104L3L5LPXR1"}, "label": "the third giraffe from the left"}]} {"id": 8300, "image": "COCO_train2014_000000008300.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"second giraffe from the right\"."}], "task": "refering", "answer_template": "The \"second giraffe from the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 154, 155, 156, 157, 176, 177, 178, 179, 180, 199, 200, 201, 221, 222, 223, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 285, 286, 287, 288, 289, 290, 291, 308, 309, 310, 311, 312, 313, 314, 330, 331, 332, 333, 334, 335, 336], "bbox": [0.38625, 0.3559294117647059, 0.8705624999999999, 0.9868], "iscrowd": 0, "area": 25043.94614999999, "rle": {"size": [425, 640], "counts": "mlV3e0a<3M3M3M3M3M3M3M3M3M3O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1O100O1O100O100N2O1O1N2O1O1O1O1O1O100O1O1O1O100O1O1O100O1O1O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O100O1O100O100O1O100O1O1O1O10O01O1O1O1O100O1O1O1O1O100O1O1O1O1O1OL5L2VNfGdN^8[1gG]N\\8d1hGUN[8j1jGnMY8R2lGfMX8Y2mG`MU8a2oGWMU8h2U1010O01O0010O01O010O0010O01O0010O01O010O001O010O0010O01O0010O01O010O1O00100O0010OO2N2N1O2N2N1O2M2O2N2N1O2N3M3M3M3M3M3M3N2O100O0010E:O2N1011N2N3N1N2O1N2O1O1O0000000001O01O0000002N2N3M1O1O10O01O1O010O1O010O1O00100O00100O001O010O001O10O01O001O010O001O1N101N104L3L5LPXR1"}, "label": "second giraffe from the right"}]} {"id": 376941, "image": "COCO_train2014_000000376941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a black leather coat surrounded by luggage\"."}], "task": "refering", "answer_template": "The \"a person in a black leather coat surrounded by luggage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 30, 31, 32, 33, 34, 47, 48, 49, 50, 51, 52, 53, 65, 66, 67, 68, 69, 70, 71, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 156, 157, 158, 159, 174, 175, 176, 177, 192, 193, 194, 195, 212, 213], "bbox": [0.59662, 0.0, 0.9691000000000001, 0.9236000000000001], "iscrowd": 0, "area": 33959.059149999994, "rle": {"size": [375, 500], "counts": "hc]35^;4N3M2O2N10001O001O0O2L3N3L3M4L3N3L3M3M4M2M4L3M4M2M4L3N3M2O001N2O1O001O1O001N2O1O001O1O1O001O1IVMaGP3X8QMgGe3c7]L\\HX4n6iKQIm4c0TKk40aJi5W1cJ[2f5bM^J\\2d5_MaJ_2a5\\McJc2_5YMeJe2X7N101N2N1O2N2O0O2N2N1O2O1N2N1O2N2O3L4K4M1N2O2M2O1O1O101O000000001N101O00001O001O001O001N10001O001O001O001O2N1N2O1O1O1O1O1O1O1O1O1O1N2O1O1O1O0001dESOl9n0QFUOn9l0oEWOP:j0nEXOQ:j0kEYOQ:Y1SOm0^Ob0J6I7oG`Ld7f3ZH\\L`7j3^HXL[7o3dHRLV7[4I7J5K6WKoJd2X5[MoJY2W5gMPKl1V5SNQK`1V5_NRKT1T5lNRKg0U5XORK>R5BSK7o4HWK0l4OZKJh46]KCd4=aK]Oa4c0dKUO_4j0gKoN[4P1kKiNW4W1U3O1O100O102M8H2N1O1N101O1O1O001N2M2N3M3N1N3M3M3M7JS]5"}, "label": "a person in a black leather coat surrounded by luggage"}]} {"id": 376941, "image": "COCO_train2014_000000376941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man wearing a black leather jacket\"."}], "task": "refering", "answer_template": "The \"the man wearing a black leather jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 30, 31, 32, 33, 34, 47, 48, 49, 50, 51, 52, 53, 65, 66, 67, 68, 69, 70, 71, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 156, 157, 158, 159, 174, 175, 176, 177, 192, 193, 194, 195, 212, 213], "bbox": [0.59662, 0.0, 0.9691000000000001, 0.9236000000000001], "iscrowd": 0, "area": 33959.059149999994, "rle": {"size": [375, 500], "counts": "hc]35^;4N3M2O2N10001O001O0O2L3N3L3M4L3N3L3M3M4M2M4L3M4M2M4L3N3M2O001N2O1O001O1O001N2O1O001O1O1O001O1IVMaGP3X8QMgGe3c7]L\\HX4n6iKQIm4c0TKk40aJi5W1cJ[2f5bM^J\\2d5_MaJ_2a5\\McJc2_5YMeJe2X7N101N2N1O2N2O0O2N2N1O2O1N2N1O2N2O3L4K4M1N2O2M2O1O1O101O000000001N101O00001O001O001O001N10001O001O001O001O2N1N2O1O1O1O1O1O1O1O1O1O1N2O1O1O1O0001dESOl9n0QFUOn9l0oEWOP:j0nEXOQ:j0kEYOQ:Y1SOm0^Ob0J6I7oG`Ld7f3ZH\\L`7j3^HXL[7o3dHRLV7[4I7J5K6WKoJd2X5[MoJY2W5gMPKl1V5SNQK`1V5_NRKT1T5lNRKg0U5XORK>R5BSK7o4HWK0l4OZKJh46]KCd4=aK]Oa4c0dKUO_4j0gKoN[4P1kKiNW4W1U3O1O100O102M8H2N1O1N101O1O1O001N2M2N3M3N1N3M3M3M7JS]5"}, "label": "the man wearing a black leather jacket"}]} {"id": 376941, "image": "COCO_train2014_000000376941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the arm of a person ' s blue jacket\"."}], "task": "refering", "answer_template": "The \"the arm of a person ' s blue jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 36, 54, 55, 72, 73, 90, 91, 108, 109, 126, 127, 144, 145, 162, 163, 180, 181, 182, 198, 199, 200, 201, 216, 217, 218, 219], "bbox": [0.0, 0.11010666666666666, 0.2191, 0.9887733333333334], "iscrowd": 0, "area": 13964.725650000002, "rle": {"size": [375, 500], "counts": "Z12g;5Ll4SKm4SKN2N2N2N2N3M2N2N2N2N2N3M2N2N2N2N2N3M2N2O1N2N2N3M2N2N2N5K5K5K5eLmKhNX4T1hKkN]4P1eKnN`4n0`KQOe4k0[KTOj4g0XKWOl4f0TKYOQ57[KHj4lN_LR1f3cMaM\\2n5O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2O0O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O`j^4"}, "label": "the arm of a person ' s blue jacket"}]} {"id": 24689, "image": "COCO_train2014_000000024689.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl with a blue dress is sitting at a table between an adult and a little boy\"."}], "task": "refering", "answer_template": "The \"a little girl with a blue dress is sitting at a table between an adult and a little boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 127, 128, 129, 150, 151, 152, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 220, 221, 222, 243], "bbox": [0.503796875, 0.28072599531615927, 0.687203125, 0.6734192037470726], "iscrowd": 0, "area": 11420.371800000003, "rle": {"size": [427, 640], "counts": "hmV43U=3L4L4J6M2N3N2M3N2M2O2M2O101O100O1O100O1O100]NhNRGY1k8TOjFm0R9AdF?X9KaF6[90cF0[94bFM\\97bFI\\9;bFE[9a0bF_O\\9e0bF[O]9h0aFXO_9Z2000000O100000000O100000002N4L4L4L=B>C001O0000000000000000O2O1O001O1N2O001O1O0O2O1O1O001O1N3N4L3M3M3L4M3M4VMeF\\1c9ZN_Ff1c:O2O001O001N10001O001O0O101O001J5K6I8H8J6M4M2N2^OnCGT<7d0K5L3M_Uc2"}, "label": "a little girl with a blue dress is sitting at a table between an adult and a little boy"}]} {"id": 24689, "image": "COCO_train2014_000000024689.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little girl with blond hair wearing a blue dress looking down\"."}], "task": "refering", "answer_template": "The \"little girl with blond hair wearing a blue dress looking down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 127, 128, 129, 150, 151, 152, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 220, 221, 222, 243], "bbox": [0.503796875, 0.28072599531615927, 0.687203125, 0.6734192037470726], "iscrowd": 0, "area": 11420.371800000003, "rle": {"size": [427, 640], "counts": "hmV43U=3L4L4J6M2N3N2M3N2M2O2M2O101O100O1O100O1O100]NhNRGY1k8TOjFm0R9AdF?X9KaF6[90cF0[94bFM\\97bFI\\9;bFE[9a0bF_O\\9e0bF[O]9h0aFXO_9Z2000000O100000000O100000002N4L4L4L=B>C001O0000000000000000O2O1O001O1N2O001O1O0O2O1O1O001O1N3N4L3M3M3L4M3M4VMeF\\1c9ZN_Ff1c:O2O001O001N10001O001O0O101O001J5K6I8H8J6M4M2N2^OnCGT<7d0K5L3M_Uc2"}, "label": "little girl with blond hair wearing a blue dress looking down"}]} {"id": 24689, "image": "COCO_train2014_000000024689.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black shirt holding a knife\"."}], "task": "refering", "answer_template": "The \"a woman in a black shirt holding a knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 48, 49, 50, 51, 71, 72, 73, 74, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 301, 302, 303, 304, 305, 306, 307, 308, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.026984374999999998, 0.0853864168618267, 0.5022656249999999, 0.9730444964871193], "iscrowd": 0, "area": 44344.8626, "rle": {"size": [427, 640], "counts": "a[75j<>B>B>B>F9L5J6K5K5J6K5K5J6K5K5J5L5K5L4M3M3M3M3M2N3M3M3M3M3M3M3M3M5K5K5K5K4L5K5K5K4L5K5K5K5K4L4L1O1N2O1O2N1O1O1O1N2L4I7J6J6QOo0M3M3M3L4M3M3O2N1N2O1O1O1N2O1O1O1O1N2O1OWNiM`HV2`7kM`HS2a7nM_HR2`7PN_Ho1a7RN`Hm1_7TNaHl1_7UN`Hj1`7WN`Hi1_7YN`Hg1_7ZNaHe1_7\\NaHd1^7^NaHa1_7`NaH`1_7`NaH`1^7bNaH]1`7cN`H]1_7dNaH[1`7fN_HZ1a7fN_HZ1`7gN`HX1a7iN^HW1b7iN_HU1b7lN]HU1a7lN_HT1a7lN_HU1`7lN_HU1_7lNaHT1_7lNaHU1^7lNaHU1]7lNcHT1]7lNcHU1\\7lNcHU1\\7kNdHU1\\7lNcHU1\\7kNdHV1[7jNeHV1[7kNdHV1[7jNeHW1Z7iNgHV1Y7kNfHV1X7kNhHV1W7jNiHV1W7kNhHV1W7jNiHW1V7jNiHW1V7iNjHW1V7iNjHX1U7iNjHX1U7hNkHX1U7hNkHY1T7hNkHY1T7gNlHY1T7gNlHZ1R7hNmHY1R7gNnHY1R7gNoHY1P7hNoHY1P7gNPIY1P7hNoHY1P7gNPIZ1o6fNQIZ1o6gNPIZ1o6fNQI\\1m6dNSIb1g6_NXIg1b6YN^Il1\\6UNdIQ2V6PNiIV2o5lMQJZ3h4hLWKb3]4`LcKj3R4WLnKS4g3oKYLV4g1ZJgNb1BX4^1`JjNY1H[4U1fJlNQ1O]4l0lJoNi04`4c0QKSO`0:c4:WKUO8a0k7YOWHf0n7SOTHm0U:01O0000001O000010O0001O0000001O00001O00001O00000010O0001O00001O0000001O00001O00001O01O0001O00001O00001O0000001O00001O01O0001O00001O00001O0000001O00001O01O01O0000001O00001O00001O00000010O0001O00001O0000001O00001O00001O01O0001O00001O00001O0000001O00001O01O01O0000001O00001O0000hkU4"}, "label": "a woman in a black shirt holding a knife"}]} {"id": 24689, "image": "COCO_train2014_000000024689.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black sweater preparing to eat\"."}], "task": "refering", "answer_template": "The \"a woman in a black sweater preparing to eat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 48, 49, 50, 51, 71, 72, 73, 74, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 301, 302, 303, 304, 305, 306, 307, 308, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.026984374999999998, 0.0853864168618267, 0.5022656249999999, 0.9730444964871193], "iscrowd": 0, "area": 44344.8626, "rle": {"size": [427, 640], "counts": "a[75j<>B>B>B>F9L5J6K5K5J6K5K5J6K5K5J5L5K5L4M3M3M3M3M2N3M3M3M3M3M3M3M3M5K5K5K5K4L5K5K5K4L5K5K5K5K4L4L1O1N2O1O2N1O1O1O1N2L4I7J6J6QOo0M3M3M3L4M3M3O2N1N2O1O1O1N2O1O1O1O1N2O1OWNiM`HV2`7kM`HS2a7nM_HR2`7PN_Ho1a7RN`Hm1_7TNaHl1_7UN`Hj1`7WN`Hi1_7YN`Hg1_7ZNaHe1_7\\NaHd1^7^NaHa1_7`NaH`1_7`NaH`1^7bNaH]1`7cN`H]1_7dNaH[1`7fN_HZ1a7fN_HZ1`7gN`HX1a7iN^HW1b7iN_HU1b7lN]HU1a7lN_HT1a7lN_HU1`7lN_HU1_7lNaHT1_7lNaHU1^7lNaHU1]7lNcHT1]7lNcHU1\\7lNcHU1\\7kNdHU1\\7lNcHU1\\7kNdHV1[7jNeHV1[7kNdHV1[7jNeHW1Z7iNgHV1Y7kNfHV1X7kNhHV1W7jNiHV1W7kNhHV1W7jNiHW1V7jNiHW1V7iNjHW1V7iNjHX1U7iNjHX1U7hNkHX1U7hNkHY1T7hNkHY1T7gNlHY1T7gNlHZ1R7hNmHY1R7gNnHY1R7gNoHY1P7hNoHY1P7gNPIY1P7hNoHY1P7gNPIZ1o6fNQIZ1o6gNPIZ1o6fNQI\\1m6dNSIb1g6_NXIg1b6YN^Il1\\6UNdIQ2V6PNiIV2o5lMQJZ3h4hLWKb3]4`LcKj3R4WLnKS4g3oKYLV4g1ZJgNb1BX4^1`JjNY1H[4U1fJlNQ1O]4l0lJoNi04`4c0QKSO`0:c4:WKUO8a0k7YOWHf0n7SOTHm0U:01O0000001O000010O0001O0000001O00001O00001O00000010O0001O00001O0000001O00001O00001O01O0001O00001O00001O0000001O00001O01O0001O00001O00001O0000001O00001O01O01O0000001O00001O00001O00000010O0001O00001O0000001O00001O00001O01O0001O00001O00001O0000001O00001O01O01O0000001O00001O0000hkU4"}, "label": "a woman in a black sweater preparing to eat"}]} {"id": 24689, "image": "COCO_train2014_000000024689.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small boy wearing a plaid cap eating pizza\"."}], "task": "refering", "answer_template": "The \"a small boy wearing a plaid cap eating pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 177, 178, 179, 180, 200, 201, 202, 203, 223, 224, 225, 226, 227, 248, 249, 250], "bbox": [0.704796875, 0.40327868852459015, 0.902859375, 0.7007962529274004], "iscrowd": 0, "area": 8978.015399999993, "rle": {"size": [427, 640], "counts": "QXl58R=2N2N2O2M101N101cD]Oo9d0QF\\Oo9e0oE]Oo9d0PF]OP:e0nE[OQ:j0jEXOU:l0fEWOW:o0dEWOV:m0dEJf9b0nE_OR:n1N100O10000O10000O10000O10000O01000O10000O01000O1000O1000000000000O1000O1000000001O0017H0001O001O00001O010O0bNTF]Om9>WFBi9>WFBj9 in this image.", "objects": [{"patches": [154, 155, 156, 157, 177, 178, 179, 180, 200, 201, 202, 203, 223, 224, 225, 226, 227, 248, 249, 250], "bbox": [0.704796875, 0.40327868852459015, 0.902859375, 0.7007962529274004], "iscrowd": 0, "area": 8978.015399999993, "rle": {"size": [427, 640], "counts": "QXl58R=2N2N2O2M101N101cD]Oo9d0QF\\Oo9e0oE]Oo9d0PF]OP:e0nE[OQ:j0jEXOU:l0fEWOW:o0dEWOV:m0dEJf9b0nE_OR:n1N100O10000O10000O10000O10000O01000O10000O01000O1000O1000000000000O1000O1000000001O0017H0001O001O00001O010O0bNTF]Om9>WFBi9>WFBj9 in this image.", "objects": [{"patches": [167, 168, 169, 184, 190, 191, 192, 207, 230, 253, 254, 276, 277, 299, 300, 322, 323, 324], "bbox": [0.0017343750000000002, 0.5210304449648712, 0.38371875, 0.9899297423887587], "iscrowd": 0, "area": 7943.527749999999, "rle": {"size": [427, 640], "counts": "fi0i0m9e2dM\\201O0000000000000000000000000000000000hKVIo1T9O1O001O1O1O1O10O01O1O1O1O1O3M4L4L4L4L4L4L4L4L4L4L4L5K4L4L4L4L4L4M3L4L4L4L4L3M3M^Pa1i0in^N1O0O100000O10000O100O10000O10000O100O10000N2N2O00100000000000000000000O10000000000000000000O100001O00000O1000001O0000000O1K5Lm_T5"}, "label": "the seat in which the blonde woman is sitting on"}]} {"id": 360570, "image": "COCO_train2014_000000360570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with pink and black hair walking a dog\"."}], "task": "refering", "answer_template": "The \"a woman with pink and black hair walking a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 37, 38, 39, 60, 61, 62, 83, 84, 85, 86, 106, 107, 108, 109, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 243, 244, 245, 246, 266, 267, 268, 269, 289, 290, 291, 292, 312, 313, 314, 315, 335, 336, 337, 338], "bbox": [0.535953125, 0.020235849056603774, 0.7652187500000001, 0.9842688679245283], "iscrowd": 0, "area": 31797.931350000003, "rle": {"size": [424, 640], "counts": "kV^42U=1O2M3N2N1O2M3N2N2M2O2N2M3N1O2M3N2N2M2O2N2]JeNUO]1j0dNUO\\1k0fNSOY1m0iNQOX1o0iNPOW1P1jNoNV1P1lNnNU1R1lNmNS1T1nNkNR1U1oNjNQ1`MKIVOe2P1[LV1k0kMi2o0ZL\\1i0fMl2n0YLc1e0aMP3l0ZLj1a0[MT3k0YLP2?VMW3j0XLW2;QM[3i0XL\\29lL^3h0WLc26fLb3g0VLi23bLe3e0WLP3O\\Li3d0VLW3LVLm3c0TL^3JQLP4b0TLd3GkKT4a0SLj3EfKW4`0RLQ4AaK[4>SLY5W3\\KgLg4a2PL]MS4`2oK^MS4c2lK[MW4e2gKZM[4g2dKWM^4i2bKTMb4l2l21O010O1nLUMgKj2W4_MbKb2[4hM]KP1dM_Ol6JZKc0QN@c65UK9^N^O[6b0PKMlN^OQ6n0mJAXO^Oh5Z1iJUOF]O_5g1eJhN2_OV5R2aJ`N<[OQ5]2\\JYNf0VOl4j2XJoMP1TOe4V3TJgMZ1PO_4b3PJ_Md1kNZ4V5j1ROSHdLk7]3WHcLh7\\3ZHdLi7W3YHhLk7S3WHmLl7n2VHRMn7h2THXMo7c2SH\\MQ8_2QHaMR8Z2PHfMT8T2nGkMV8P2mGoMV8l1lGTN[8c1gG]Na8Y1aGfNh8P1ZGPOn8f0TGZOT9Bb0_Oa0eNoom1"}, "label": "a woman with pink and black hair walking a dog"}]} {"id": 360570, "image": "COCO_train2014_000000360570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"s woman walking her dog on sidewalk\"."}], "task": "refering", "answer_template": "The \"s woman walking her dog on sidewalk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 37, 38, 39, 60, 61, 62, 83, 84, 85, 86, 106, 107, 108, 109, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 243, 244, 245, 246, 266, 267, 268, 269, 289, 290, 291, 292, 312, 313, 314, 315, 335, 336, 337, 338], "bbox": [0.535953125, 0.020235849056603774, 0.7652187500000001, 0.9842688679245283], "iscrowd": 0, "area": 31797.931350000003, "rle": {"size": [424, 640], "counts": "kV^42U=1O2M3N2N1O2M3N2N2M2O2N2M3N1O2M3N2N2M2O2N2]JeNUO]1j0dNUO\\1k0fNSOY1m0iNQOX1o0iNPOW1P1jNoNV1P1lNnNU1R1lNmNS1T1nNkNR1U1oNjNQ1`MKIVOe2P1[LV1k0kMi2o0ZL\\1i0fMl2n0YLc1e0aMP3l0ZLj1a0[MT3k0YLP2?VMW3j0XLW2;QM[3i0XL\\29lL^3h0WLc26fLb3g0VLi23bLe3e0WLP3O\\Li3d0VLW3LVLm3c0TL^3JQLP4b0TLd3GkKT4a0SLj3EfKW4`0RLQ4AaK[4>SLY5W3\\KgLg4a2PL]MS4`2oK^MS4c2lK[MW4e2gKZM[4g2dKWM^4i2bKTMb4l2l21O010O1nLUMgKj2W4_MbKb2[4hM]KP1dM_Ol6JZKc0QN@c65UK9^N^O[6b0PKMlN^OQ6n0mJAXO^Oh5Z1iJUOF]O_5g1eJhN2_OV5R2aJ`N<[OQ5]2\\JYNf0VOl4j2XJoMP1TOe4V3TJgMZ1PO_4b3PJ_Md1kNZ4V5j1ROSHdLk7]3WHcLh7\\3ZHdLi7W3YHhLk7S3WHmLl7n2VHRMn7h2THXMo7c2SH\\MQ8_2QHaMR8Z2PHfMT8T2nGkMV8P2mGoMV8l1lGTN[8c1gG]Na8Y1aGfNh8P1ZGPOn8f0TGZOT9Bb0_Oa0eNoom1"}, "label": "s woman walking her dog on sidewalk"}]} {"id": 360570, "image": "COCO_train2014_000000360570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in black baseball cap , tight jeans , black shirt and purple ugg boots\"."}], "task": "refering", "answer_template": "The \"a woman in black baseball cap , tight jeans , black shirt and purple ugg boots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 69, 70, 92, 93, 94, 115, 116, 138, 139, 161, 162, 163, 184, 185, 186, 207, 208, 230, 231, 232, 253, 254], "bbox": [0.0, 0.1318632075471698, 0.103359375, 0.749504716981132], "iscrowd": 0, "area": 12135.469150000003, "rle": {"size": [424, 640], "counts": "W3h0^XJZ1CUNV6`0ZJS1IYNo5a0\\JQ1G^Nn5?_Jn0EbNn5=bJl0C[NX62mJ^1mN`Nn8Z1TGfNo8U1SGkNQ9o0SGoNS9i0RGSOT9f0n1H7G[^]7"}, "label": "a woman in black baseball cap , tight jeans , black shirt and purple ugg boots"}]} {"id": 360570, "image": "COCO_train2014_000000360570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in tight jeans leaning against a storefront\"."}], "task": "refering", "answer_template": "The \"a woman in tight jeans leaning against a storefront\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 69, 70, 92, 93, 94, 115, 116, 138, 139, 161, 162, 163, 184, 185, 186, 207, 208, 230, 231, 232, 253, 254], "bbox": [0.0, 0.1318632075471698, 0.103359375, 0.749504716981132], "iscrowd": 0, "area": 12135.469150000003, "rle": {"size": [424, 640], "counts": "W3h0^XJZ1CUNV6`0ZJS1IYNo5a0\\JQ1G^Nn5?_Jn0EbNn5=bJl0C[NX62mJ^1mN`Nn8Z1TGfNo8U1SGkNQ9o0SGoNS9i0RGSOT9f0n1H7G[^]7"}, "label": "a woman in tight jeans leaning against a storefront"}]} {"id": 163985, "image": "COCO_train2014_000000163985.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy that is skiing\"."}], "task": "refering", "answer_template": "The \"the boy that is skiing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 80, 81, 103, 104, 105, 126, 127, 128, 149, 150, 172, 173, 174, 195, 196, 197, 220], "bbox": [0.474703125, 0.15727777777777777, 0.591109375, 0.7331111111111112], "iscrowd": 0, "area": 7427.1709, "rle": {"size": [360, 640], "counts": "aP[36k:=C>D7F;D;K5K6K4K5L4L3PIfMo4]2jJnMo4U2lJoMQ5U2jJnMU5U2fJnMY5k0PJYOa0O]5h0VJYO82b5e0UJ]O30i5c0TJQ1l5POTJn0n5RORJj0R6VOnIg0U6YOkId0W6]OhIa0[6_OeI=m5jMXJi1K9P6TNPJc106R6[NjI`141T6d0lIZOU6g0kIWOW6i0iIQN06X6j1hIQN5MU6R2fISN6FU6X2eISN8@U6]2cITN:TOZ6h2]IUNS7k1mHVNQ7k1oHVNP7i1PIYNo6g1QI[Nn6c1SI^Nm6a1SI[MJS1R7_1ZIaNe6\\1`IdN_6Y1eIfNZ6W1jIiNV6T1nImNo5Q1UJTOe5j0]JGQ57SKLi4:QKJk4`N[Jm0LcNh0S2o4RNeJ_1=a0a5A_Jb0\\5_OdJd0Y5\\OhJXNUOW2\\6VOfJi0Y5WOiJh0W5VOlJi0V5jNaIYO[1m1U5_NiID`1`1e5^NiJ4XN8o6BfK<[4AiK>a7O2O000O2M2O1N3N1N^kk2"}, "label": "the boy that is skiing"}]} {"id": 163985, "image": "COCO_train2014_000000163985.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"young boy on water skis\"."}], "task": "refering", "answer_template": "The \"young boy on water skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 80, 81, 103, 104, 105, 126, 127, 128, 149, 150, 172, 173, 174, 195, 196, 197, 220], "bbox": [0.474703125, 0.15727777777777777, 0.591109375, 0.7331111111111112], "iscrowd": 0, "area": 7427.1709, "rle": {"size": [360, 640], "counts": "aP[36k:=C>D7F;D;K5K6K4K5L4L3PIfMo4]2jJnMo4U2lJoMQ5U2jJnMU5U2fJnMY5k0PJYOa0O]5h0VJYO82b5e0UJ]O30i5c0TJQ1l5POTJn0n5RORJj0R6VOnIg0U6YOkId0W6]OhIa0[6_OeI=m5jMXJi1K9P6TNPJc106R6[NjI`141T6d0lIZOU6g0kIWOW6i0iIQN06X6j1hIQN5MU6R2fISN6FU6X2eISN8@U6]2cITN:TOZ6h2]IUNS7k1mHVNQ7k1oHVNP7i1PIYNo6g1QI[Nn6c1SI^Nm6a1SI[MJS1R7_1ZIaNe6\\1`IdN_6Y1eIfNZ6W1jIiNV6T1nImNo5Q1UJTOe5j0]JGQ57SKLi4:QKJk4`N[Jm0LcNh0S2o4RNeJ_1=a0a5A_Jb0\\5_OdJd0Y5\\OhJXNUOW2\\6VOfJi0Y5WOiJh0W5VOlJi0V5jNaIYO[1m1U5_NiID`1`1e5^NiJ4XN8o6BfK<[4AiK>a7O2O000O2M2O1N3N1N^kk2"}, "label": "young boy on water skis"}]} {"id": 163985, "image": "COCO_train2014_000000163985.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy with white shirt pulling boy on skis from boat\"."}], "task": "refering", "answer_template": "The \"boy with white shirt pulling boy on skis from boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 41, 42, 43, 44, 63, 64, 65, 66, 67, 86, 87, 88, 89, 90, 109, 110, 111, 112, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250, 251, 258, 266, 267, 268, 269, 270, 271, 272, 273, 274, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296], "bbox": [0.22728125000000002, 0.0615, 0.9446875, 0.9870277777777777], "iscrowd": 0, "area": 47161.021499999995, "rle": {"size": [360, 640], "counts": "mYc1UMGe2a0UMBf2e0TM_Oh2h0QM\\Oj2l0oLYOd2U1VMoN`2^1XMgNd2_1VMeNe2S5K4K5G9G:FkS<"}, "label": "boy with white shirt pulling boy on skis from boat"}]} {"id": 163985, "image": "COCO_train2014_000000163985.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in white shirt and white pants\"."}], "task": "refering", "answer_template": "The \"man in white shirt and white pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 41, 42, 43, 44, 63, 64, 65, 66, 67, 86, 87, 88, 89, 90, 109, 110, 111, 112, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250, 251, 258, 266, 267, 268, 269, 270, 271, 272, 273, 274, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296], "bbox": [0.22728125000000002, 0.0615, 0.9446875, 0.9870277777777777], "iscrowd": 0, "area": 47161.021499999995, "rle": {"size": [360, 640], "counts": "mYc1UMGe2a0UMBf2e0TM_Oh2h0QM\\Oj2l0oLYOd2U1VMoN`2^1XMgNd2_1VMeNe2S5K4K5G9G:FkS<"}, "label": "man in white shirt and white pants"}]} {"id": 106652, "image": "COCO_train2014_000000106652.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a brown dress with a floral scarf standing next to a man\"."}], "task": "refering", "answer_template": "The \"a woman in a brown dress with a floral scarf standing next to a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 57, 58, 59, 80, 81, 82, 83, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 286, 287, 288, 289, 290, 309, 310, 311, 312, 313, 332, 333, 334, 335, 336, 355, 356, 357, 358, 379, 380, 381], "bbox": [0.4375, 0.09345833333333334, 0.6908749999999999, 0.9853541666666668], "iscrowd": 0, "area": 47835.9264, "rle": {"size": [480, 640], "counts": "XdS4c0T>;D=D;E;E;E;E in this image.", "objects": [{"patches": [35, 57, 58, 59, 80, 81, 82, 83, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 286, 287, 288, 289, 290, 309, 310, 311, 312, 313, 332, 333, 334, 335, 336, 355, 356, 357, 358, 379, 380, 381], "bbox": [0.4375, 0.09345833333333334, 0.6908749999999999, 0.9853541666666668], "iscrowd": 0, "area": 47835.9264, "rle": {"size": [480, 640], "counts": "XdS4c0T>;D=D;E;E;E;E in this image.", "objects": [{"patches": [31, 32, 33, 53, 54, 55, 56, 76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 259, 260, 261, 262, 263, 282, 283, 284, 285, 286, 305, 306, 307, 308, 309, 328, 329, 330, 331, 332, 351, 352, 353, 354, 374, 375, 376, 377], "bbox": [0.245515625, 0.066375, 0.4809374999999999, 0.9878958333333333], "iscrowd": 0, "area": 48424.816050000016, "rle": {"size": [480, 640], "counts": "WiY23j>9G9G9G9G9G9G9^DXN]9Q2ZF[NZ9n1]F^NW9k1`F`NU9i1bFcNR9f1eFfNo8P4D;E in this image.", "objects": [{"patches": [31, 32, 33, 53, 54, 55, 56, 76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 259, 260, 261, 262, 263, 282, 283, 284, 285, 286, 305, 306, 307, 308, 309, 328, 329, 330, 331, 332, 351, 352, 353, 354, 374, 375, 376, 377], "bbox": [0.245515625, 0.066375, 0.4809374999999999, 0.9878958333333333], "iscrowd": 0, "area": 48424.816050000016, "rle": {"size": [480, 640], "counts": "WiY23j>9G9G9G9G9G9G9^DXN]9Q2ZF[NZ9n1]F^NW9k1`F`NU9i1bFcNR9f1eFfNo8P4D;E in this image.", "objects": [{"patches": [153, 170, 171, 187, 188, 189, 204, 205, 206, 207, 221, 222, 223, 224, 225, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 272, 273, 274, 275, 289, 290, 291, 292, 306, 307, 308, 309, 323, 324, 325], "bbox": [0.0, 0.4019375, 0.28635416666666663, 0.8426562499999999], "iscrowd": 0, "area": 25703.68890000001, "rle": {"size": [640, 480], "counts": "Q8Q8P<3M3L4M2NO010O1O010O1O010O1O00100O00100O00100O001O10O01O001O1O001O1O001O1O3M3M2N3M3M2O2M3N0O100O1O1O1O1N2O001N2O1O1N2O1N2N101N2N2N2O1N2N1O2O1N2O00001O0O101O000O3N1O1O1N3N1O1O2N1N2O1O2N1N2O1N3M2N1lM[AlNg>S1[AjNg>T1]AhNd>W1_AfNc>Y1_AdNc>Z1_AdNb>\\1_AbNc>\\1_AbNc>\\1`AaNa>_1aA^Na>`1cAYNa>g1dAPNa>n1gAgM]>X2kA^MY>a2c1O2O0O2N102M3N2M2N3N2M3N2M2N3L4L4M3L3M4L4M3L3M4Leoe6"}, "label": "a brown pot with a plant growing out of it"}]} {"id": 106660, "image": "COCO_train2014_000000106660.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue colour flower pot holding up a yellow rose\"."}], "task": "refering", "answer_template": "The \"a blue colour flower pot holding up a yellow rose\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 192, 193, 194, 209, 210, 211, 225, 226, 227, 228, 242, 243, 244, 245, 259, 260, 261, 262, 276, 277, 278, 279, 293, 294, 295, 296, 310, 311, 312, 313, 327, 328, 329, 330], "bbox": [0.26547916666666665, 0.47225, 0.44443750000000004, 0.8592812500000001], "iscrowd": 0, "area": 16344.702499999998, "rle": {"size": [640, 480], "counts": "gj_21kc08H9G8H9G8Z^OmNS`0\\1\\_OQO_`0X1P_OTOl`0R2TNl1L5K4M3L5K4L4QOP1M2O1O2N1N2O2N1N2O2N1O1O2O0O1O2N1O101N1OaHdC\\7[ in this image.", "objects": [{"patches": [176, 192, 193, 194, 209, 210, 211, 225, 226, 227, 228, 242, 243, 244, 245, 259, 260, 261, 262, 276, 277, 278, 279, 293, 294, 295, 296, 310, 311, 312, 313, 327, 328, 329, 330], "bbox": [0.26547916666666665, 0.47225, 0.44443750000000004, 0.8592812500000001], "iscrowd": 0, "area": 16344.702499999998, "rle": {"size": [640, 480], "counts": "gj_21kc08H9G8H9G8Z^OmNS`0\\1\\_OQO_`0X1P_OTOl`0R2TNl1L5K4M3L5K4L4QOP1M2O1O2N1N2O2N1N2O2N1O1O2O0O1O2N1O101N1OaHdC\\7[ in this image.", "objects": [{"patches": [76, 77, 78, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 115, 126, 127, 128], "bbox": [0.0, 0.34573573573573574, 0.441, 0.6002402402402403], "iscrowd": 0, "area": 9578.0025, "rle": {"size": [333, 500], "counts": "Y5o0^90000000000O10000000000000000000000000000000000O1000000000000000000000000000000000O10O1000000000000000000000000000000000000O1000000000000000000000000000000000000O1000000000000000000000000000L4H8K5N101O1N2O1N2O1O1N2O1O1O1N2O1N2N2N2O1000O2O0000001O0000000000O10000N2N2N2N2M30O1000001O0000000000000000000000001O000O100000000000000000001O0000000000000O10`0@1O001O1O1O001O1O1O001O00000000000O4M:D=@Tm2BYiL1k94QFN=MR9;jFJU96gFOX92cF3\\9`0O0010O000001OO;Ed0\\OQbj2"}, "label": "bed on the other side of the lamp"}]} {"id": 393394, "image": "COCO_train2014_000000393394.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed that is farthest away from the flower vase\"."}], "task": "refering", "answer_template": "The \"the bed that is farthest away from the flower vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 115, 126, 127, 128], "bbox": [0.0, 0.34573573573573574, 0.441, 0.6002402402402403], "iscrowd": 0, "area": 9578.0025, "rle": {"size": [333, 500], "counts": "Y5o0^90000000000O10000000000000000000000000000000000O1000000000000000000000000000000000O10O1000000000000000000000000000000000000O1000000000000000000000000000000000000O1000000000000000000000000000L4H8K5N101O1N2O1N2O1O1N2O1O1O1N2O1N2N2N2O1000O2O0000001O0000000000O10000N2N2N2N2M30O1000001O0000000000000000000000001O000O100000000000000000001O0000000000000O10`0@1O001O1O1O001O1O1O001O00000000000O4M:D=@Tm2BYiL1k94QFN=MR9;jFJU96gFOX92cF3\\9`0O0010O000001OO;Ed0\\OQbj2"}, "label": "the bed that is farthest away from the flower vase"}]} {"id": 377017, "image": "COCO_train2014_000000377017.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the brick in his hand\"."}], "task": "refering", "answer_template": "The \"the man with the brick in his hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 38, 39, 40, 41, 42, 43, 44, 45, 46, 57, 58, 59, 60, 61, 62, 63, 64, 65, 76, 77, 78, 79, 81, 82, 83, 94, 95, 96], "bbox": [0.1278, 0.004746666666666666, 0.6339600000000001, 0.4509333333333333], "iscrowd": 0, "area": 25582.98345, "rle": {"size": [375, 500], "counts": "S^g01f;4L4L4L4L3M4L4L4L4L4L3M4L4L4L4L3M4L4L3M1O1O1O1O001O1O1O1O1O1N2O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O5K?A?A4L3M3M2N3M3M1O0000001O0000000000000000000000000000O1000000O100O100O100O10000O1O100O100O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1N2O1N2O1N2O1O1N2dMYHg0h7VO^He0d7XObHd0_7ZOfH`0]7]OiH>Y7_OnH:U7CQI8Q7UOeIe0^6[OeI?^6AeI:]6FfI4]6LfIO\\61hIH[68b20000000000O1000000000000000000000000000bFE`7;_HI^77aHN[72eH1X7OgH4W7LhH6W7JhH8W7HdH>[7B_He0`7ZO\\Hl0c7TOWHS1h7mNRHZ1m7gNmG_1R8bNgGe1X8\\NbGj1]8WN\\GP2c8c0O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O001O001O001O1O001O001O001O1O001O001O001O1O0000O100O100O100lMUI[Ol6?[I@e69cIE^65iIJW6OPJ0Q6JVJ5j5D^J:c5@dJ?\\5ZOlJc0V5WOSKd0S8L4L[mR2"}, "label": "the man with the brick in his hand"}]} {"id": 377017, "image": "COCO_train2014_000000377017.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man kneeling\"."}], "task": "refering", "answer_template": "The \"a man kneeling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 38, 39, 40, 41, 42, 43, 44, 45, 46, 57, 58, 59, 60, 61, 62, 63, 64, 65, 76, 77, 78, 79, 81, 82, 83, 94, 95, 96], "bbox": [0.1278, 0.004746666666666666, 0.6339600000000001, 0.4509333333333333], "iscrowd": 0, "area": 25582.98345, "rle": {"size": [375, 500], "counts": "S^g01f;4L4L4L4L3M4L4L4L4L4L3M4L4L4L4L3M4L4L3M1O1O1O1O001O1O1O1O1O1N2O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O5K?A?A4L3M3M2N3M3M1O0000001O0000000000000000000000000000O1000000O100O100O100O10000O1O100O100O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1N2O1N2O1N2O1O1N2dMYHg0h7VO^He0d7XObHd0_7ZOfH`0]7]OiH>Y7_OnH:U7CQI8Q7UOeIe0^6[OeI?^6AeI:]6FfI4]6LfIO\\61hIH[68b20000000000O1000000000000000000000000000bFE`7;_HI^77aHN[72eH1X7OgH4W7LhH6W7JhH8W7HdH>[7B_He0`7ZO\\Hl0c7TOWHS1h7mNRHZ1m7gNmG_1R8bNgGe1X8\\NbGj1]8WN\\GP2c8c0O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O001O001O001O1O001O001O001O1O001O001O001O1O0000O100O100O100lMUI[Ol6?[I@e69cIE^65iIJW6OPJ0Q6JVJ5j5D^J:c5@dJ?\\5ZOlJc0V5WOSKd0S8L4L[mR2"}, "label": "a man kneeling"}]} {"id": 377017, "image": "COCO_train2014_000000377017.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black shoes with white socks\"."}], "task": "refering", "answer_template": "The \"black shoes with white socks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 29, 30, 31, 32, 33, 34, 51, 69], "bbox": [0.64628, 0.0, 0.9653599999999999, 0.27237333333333336], "iscrowd": 0, "area": 4517.968899999998, "rle": {"size": [375, 500], "counts": "XZf32e;5J6J1O2O0O1O2N100O1O2O0O1O2O0O1O100O1O010O100O010O100O010O100O010O100O010O1O1O001O1O1O1O001O1O1O01000000O1000000O1000000O101O000O10000N2K5L4L4K5K5L4KcQ70_nH4K4M4L4M2N3M2N3M2N3M2N3M2N3M2N2N2N2N2M_OcFnN[9Q1d0O10L4WOi03L4M3O1N2O1N2O1N2O1O1N2O0O2O1O1O100O1O100O100O100O1O1O1M3M3N1N3M3M3N2M3M3N2MQc6"}, "label": "black shoes with white socks"}]} {"id": 377017, "image": "COCO_train2014_000000377017.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the guy with white socks in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the guy with white socks in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 29, 30, 31, 32, 33, 34, 51, 69], "bbox": [0.64628, 0.0, 0.9653599999999999, 0.27237333333333336], "iscrowd": 0, "area": 4517.968899999998, "rle": {"size": [375, 500], "counts": "XZf32e;5J6J1O2O0O1O2N100O1O2O0O1O2O0O1O100O1O010O100O010O100O010O100O010O100O010O1O1O001O1O1O1O001O1O1O01000000O1000000O1000000O101O000O10000N2K5L4L4K5K5L4KcQ70_nH4K4M4L4M2N3M2N3M2N3M2N3M2N3M2N2N2N2N2M_OcFnN[9Q1d0O10L4WOi03L4M3O1N2O1N2O1N2O1O1N2O0O2O1O1O100O1O100O100O100O1O1O1M3M3N1N3M3M3N2M3M3N2MQc6"}, "label": "the guy with white socks in the right hand picture"}]} {"id": 377019, "image": "COCO_train2014_000000377019.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a patterned - upholstery couch to the left of two young people playing a nintendo wii game\"."}], "task": "refering", "answer_template": "The \"a patterned - upholstery couch to the left of two young people playing a nintendo wii game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 183, 184, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220, 221], "bbox": [0.00644, 0.79184, 0.28166, 0.98712], "iscrowd": 0, "area": 8545.663649999999, "rle": {"size": [375, 500], "counts": "m]1i0c9[12N2N3M2N2N000000O100000000O10001O000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O1000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000002N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O2N7Ie0[OeSS4"}, "label": "a patterned - upholstery couch to the left of two young people playing a nintendo wii game"}]} {"id": 377019, "image": "COCO_train2014_000000377019.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white flower sofa\"."}], "task": "refering", "answer_template": "The \"a white flower sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 183, 184, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220, 221], "bbox": [0.00644, 0.79184, 0.28166, 0.98712], "iscrowd": 0, "area": 8545.663649999999, "rle": {"size": [375, 500], "counts": "m]1i0c9[12N2N3M2N2N000000O100000000O10001O000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O1000000000000O1000000000000000000O100000000000000000000O1000000000000000000O100000000002N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O2N7Ie0[OeSS4"}, "label": "a white flower sofa"}]} {"id": 377019, "image": "COCO_train2014_000000377019.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy playing video game with his friends\"."}], "task": "refering", "answer_template": "The \"a boy playing video game with his friends\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 83, 84, 85, 101, 102, 103, 120, 121, 138, 139, 156, 157, 192, 193, 210, 211, 229], "bbox": [0.64932, 0.3, 0.78524, 0.98456], "iscrowd": 0, "area": 5480.433650000001, "rle": {"size": [375, 500], "counts": "ZUg34_;6K5L3M4L4L4J5M4N2M3M2N3M3M3M6K8H7I6J2N3nI[Mk2h2SM[Mj2g2TM\\Mj2e2TM^Mi2e2TM^Mj2c2UM_Mh2c2VM_Mi2c2TM`Mi2b2TMbMi2`2UMcMg2`2VMdMe2`2XMdMd2_2ZMdMa2a2[McMa2`2]McM^2nNdKV3f1UNa2gNlKS3^1\\Na2bNTLR3X1_Nb2_NXLR3S1cNb2[N^LU2JSNR1a1c2WNcLl13ZNd0g1c2SNiLc1;`N6l1e2QNlLZ1e0fNGP2h2oMTMn0h0POZOS2j2oM[M`0l0\\OlNV2m2mMaM8o0B`NZ2P3kMfM4P1EeNn1e2XNkM0P1HcNk1b2]NnMLY2d1I_NRNIY2e1EaN[2\\1fMcN]2Z1cMeNa2X1R4M3M3M4L3M3M3M3M3M3M3M3M4L3MVVW1"}, "label": "a boy playing video game with his friends"}]} {"id": 377019, "image": "COCO_train2014_000000377019.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy holding up a wii remote and facing the tv\"."}], "task": "refering", "answer_template": "The \"a young boy holding up a wii remote and facing the tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 83, 84, 85, 101, 102, 103, 120, 121, 138, 139, 156, 157, 192, 193, 210, 211, 229], "bbox": [0.64932, 0.3, 0.78524, 0.98456], "iscrowd": 0, "area": 5480.433650000001, "rle": {"size": [375, 500], "counts": "ZUg34_;6K5L3M4L4L4J5M4N2M3M2N3M3M3M6K8H7I6J2N3nI[Mk2h2SM[Mj2g2TM\\Mj2e2TM^Mi2e2TM^Mj2c2UM_Mh2c2VM_Mi2c2TM`Mi2b2TMbMi2`2UMcMg2`2VMdMe2`2XMdMd2_2ZMdMa2a2[McMa2`2]McM^2nNdKV3f1UNa2gNlKS3^1\\Na2bNTLR3X1_Nb2_NXLR3S1cNb2[N^LU2JSNR1a1c2WNcLl13ZNd0g1c2SNiLc1;`N6l1e2QNlLZ1e0fNGP2h2oMTMn0h0POZOS2j2oM[M`0l0\\OlNV2m2mMaM8o0B`NZ2P3kMfM4P1EeNn1e2XNkM0P1HcNk1b2]NnMLY2d1I_NRNIY2e1EaN[2\\1fMcN]2Z1cMeNa2X1R4M3M3M4L3M3M3M3M3M3M3M3M4L3MVVW1"}, "label": "a young boy holding up a wii remote and facing the tv"}]} {"id": 377019, "image": "COCO_train2014_000000377019.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a brown shirt watching the tv\"."}], "task": "refering", "answer_template": "The \"a woman wearing a brown shirt watching the tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 118, 119, 120, 135, 136, 137, 138, 153, 154, 155, 156, 157, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 225, 226, 227, 228], "bbox": [0.48702, 0.3917866666666666, 0.74806, 0.9865866666666667], "iscrowd": 0, "area": 19316.13135, "rle": {"size": [375, 500], "counts": "cdi2b0d:b0L3N3L3M4M3L3M4M3L3VGfNT7^1[HROc7b2M3M3N2M3N2O1O1O1N2O1O1O1O1O1O1N2O1O1O1M3L4M3L4M3WOi0N2N2N2N2M3N2N2N2N2N2N2N2O1O10000O10000O10000O10000O100001O001O001O001O001O001O001O1O1O2N1O1O1O2N1O1O1O2N2N4L4L4L4L4L4L2N2N3M2SNgIXN\\6e1mIRNU6k1UJlMm5Q2^JeMe5X2gJ\\M[5a2QKTMR5i2P2M3N3L3N2L5L3M4K4M3M4L3L4M4L3L5L4L4K6K4L4L4K6Klk]1"}, "label": "a woman wearing a brown shirt watching the tv"}]} {"id": 524476, "image": "COCO_train2014_000000524476.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding the dog with both hands\"."}], "task": "refering", "answer_template": "The \"a woman holding the dog with both hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 72, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 286, 299, 300, 301, 302, 303, 304, 305, 306, 309, 322, 323, 324, 325, 326, 327, 328, 329, 331, 332, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.0033750000000000004, 0.18389583333333331, 0.482015625, 0.9883958333333334], "iscrowd": 0, "area": 71773.62465, "rle": {"size": [480, 640], "counts": "ZQ1\\;b32O1N2O1N2O1N20000O10000000000O2O000000000O10000000000O10000000000O10000000000O10000000000O10000001O00000000000000001O000000000000010O00000000000000001O000000000000001O00000000000000001O000000000000001O00001O1O2N2N1O2N1O2N2N1O2N2N1O2O0O2N2N1O2N1O2N2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N4L4L4L5K4L4L4L5K4L4M4K4L4L4L5K4L4L4L5K4L4L4L5K4L4L4Lh7@]H=b7CaH;^7EeH[OX8e0Q301O000000001O000000010O000000XFTOd:1O1O2N1M3N2N2M3N2M3N2M3N2M3N3L3N2M3N2M3N2M3N2M3N2M4M2M3N2M3N2M3N2L4_Oa0_Oa0OO2M3M3M4M2M3M3M3N3L3M3M3N2M3H9F9H8K5K5L5K4K5L5K6J7HhXk4"}, "label": "a woman holding the dog with both hands"}]} {"id": 524476, "image": "COCO_train2014_000000524476.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with long hair holding a dog\"."}], "task": "refering", "answer_template": "The \"a woman with long hair holding a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 72, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 286, 299, 300, 301, 302, 303, 304, 305, 306, 309, 322, 323, 324, 325, 326, 327, 328, 329, 331, 332, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.0033750000000000004, 0.18389583333333331, 0.482015625, 0.9883958333333334], "iscrowd": 0, "area": 71773.62465, "rle": {"size": [480, 640], "counts": "ZQ1\\;b32O1N2O1N2O1N20000O10000000000O2O000000000O10000000000O10000000000O10000000000O10000000000O10000001O00000000000000001O000000000000010O00000000000000001O000000000000001O00000000000000001O000000000000001O00001O1O2N2N1O2N1O2N2N1O2N2N1O2O0O2N2N1O2N1O2N2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N4L4L4L5K4L4L4L5K4L4M4K4L4L4L5K4L4L4L5K4L4L4L5K4L4L4Lh7@]H=b7CaH;^7EeH[OX8e0Q301O000000001O000000010O000000XFTOd:1O1O2N1M3N2N2M3N2M3N2M3N2M3N3L3N2M3N2M3N2M3N2M3N2M4M2M3N2M3N2M3N2L4_Oa0_Oa0OO2M3M3M4M2M3M3M3N3L3M3M3N2M3H9F9H8K5K5L5K4K5L5K6J7HhXk4"}, "label": "a woman with long hair holding a dog"}]} {"id": 524476, "image": "COCO_train2014_000000524476.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman on the right holding the blow dryer\"."}], "task": "refering", "answer_template": "The \"the woman on the right holding the blow dryer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 215, 216, 217, 218, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.366921875, 0.0032291666666666666, 1.0, 0.9902916666666667], "iscrowd": 0, "area": 89205.47605, "rle": {"size": [480, 640], "counts": "U]^31n>2N3N1O2N1O1O1O1O1O1O1O1O0010000O1000O10O100000O010008G;E=DO01O100O010O1O100O0010O01O10O0100O0001O01O01O00010O001O01O01O00010O0010O00010O000_O[NhCe1TV1BoN9Q1GPO8P1HPO7Q1IPO6P1JPO6P1JQO5o0KQO5o0KRO4n0LRO4n0LSO2n0NRO2n0NSO1m0ORO2n0NRO1o0OQO1o0OQO1o0OQO1o0OPO1Q1OoN1Q1OoN1Q1OoN1Q1OnN2R1NnN1S1OmN1S1OmN1S1OmN1S1OlN1U1OkN1U1OkN1U1OkN1U1OkN0U11jN0V10jN0V10jN0V10jN0V10iN0X10hN0X10hN0X10gN1Y1OgN0Z10eN1[1OeN1[1OeNL`14f600000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000001O0000000000000000000000M"}, "label": "the woman on the right holding the blow dryer"}]} {"id": 524476, "image": "COCO_train2014_000000524476.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding a blow dryer\"."}], "task": "refering", "answer_template": "The \"a woman holding a blow dryer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 215, 216, 217, 218, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.366921875, 0.0032291666666666666, 1.0, 0.9902916666666667], "iscrowd": 0, "area": 89205.47605, "rle": {"size": [480, 640], "counts": "U]^31n>2N3N1O2N1O1O1O1O1O1O1O1O0010000O1000O10O100000O010008G;E=DO01O100O010O1O100O0010O01O10O0100O0001O01O01O00010O001O01O01O00010O0010O00010O000_O[NhCe1TV1BoN9Q1GPO8P1HPO7Q1IPO6P1JPO6P1JQO5o0KQO5o0KRO4n0LRO4n0LSO2n0NRO2n0NSO1m0ORO2n0NRO1o0OQO1o0OQO1o0OQO1o0OPO1Q1OoN1Q1OoN1Q1OoN1Q1OnN2R1NnN1S1OmN1S1OmN1S1OmN1S1OlN1U1OkN1U1OkN1U1OkN1U1OkN0U11jN0V10jN0V10jN0V10jN0V10iN0X10hN0X10hN0X10gN1Y1OgN0Z10eN1[1OeN1[1OeNL`14f600000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000001O0000000000000000000000M"}, "label": "a woman holding a blow dryer"}]} {"id": 90310, "image": "COCO_train2014_000000090310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right most black bear\"."}], "task": "refering", "answer_template": "The \"the right most black bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 57, 58, 59, 60, 64, 65, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 204, 205, 206, 221, 222, 223], "bbox": [0.419859375, 0.06863291139240506, 1.0, 0.7021772151898734], "iscrowd": 0, "area": 44733.48350000001, "rle": {"size": [395, 640], "counts": "mkW32Y<1N3M2O1N3M2O2M2N3N1N1O100O1O2O0O1O2N2N2N100O1O1O1O1O2O0O1O1O1O100O1O1O2N1O001O1O100O1O1O1O1O1O1O100O2N2O1N1O100O001O010O1O010O1O100O1O2^OiMmFX2S9nMgFR2X9TNbFm1^9`0O2N1N2O1N2O2M2N2O10010O000001O010O1O2N2N101O001N101O001N2O001N101O001N2O000O10O1000O01000O10O4M7I7H8I4L4K5L3M4K6K5K5J6K5K5J6K5K1N101N100O1O2O0O100O2N100O101N1O10O001O00001O000000001O000000000000000000000000001O0000000O1N2O1RO]IRLe6m3nI_KT6a4k0O100O2O0O100O100O010O100O10O01O100O1O00100O1O1O010O1O100O001O100O1O010O1O1O10O01O1O1O001N2O001N2O010001O0O101O0O2O00001O00001O00001O00001N10000000000000000000000000000O1001O00001O00000010O0001O0000001O0000lGdLZ7]4I5K0001O0O101O0000001O00001O0000001O000O2O0000001O00001O0000001O00001N2O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1N2M3N2N2N2N2N2N2N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2M7Ha0\\NTFJS5"}, "label": "the right most black bear"}]} {"id": 90310, "image": "COCO_train2014_000000090310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bear behind another bear\"."}], "task": "refering", "answer_template": "The \"bear behind another bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 57, 58, 59, 60, 64, 65, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 204, 205, 206, 221, 222, 223], "bbox": [0.419859375, 0.06863291139240506, 1.0, 0.7021772151898734], "iscrowd": 0, "area": 44733.48350000001, "rle": {"size": [395, 640], "counts": "mkW32Y<1N3M2O1N3M2O2M2N3N1N1O100O1O2O0O1O2N2N2N100O1O1O1O1O2O0O1O1O1O100O1O1O2N1O001O1O100O1O1O1O1O1O1O100O2N2O1N1O100O001O010O1O010O1O100O1O2^OiMmFX2S9nMgFR2X9TNbFm1^9`0O2N1N2O1N2O2M2N2O10010O000001O010O1O2N2N101O001N101O001N2O001N101O001N2O000O10O1000O01000O10O4M7I7H8I4L4K5L3M4K6K5K5J6K5K5J6K5K1N101N100O1O2O0O100O2N100O101N1O10O001O00001O000000001O000000000000000000000000001O0000000O1N2O1RO]IRLe6m3nI_KT6a4k0O100O2O0O100O100O010O100O10O01O100O1O00100O1O1O010O1O100O001O100O1O010O1O1O10O01O1O1O001N2O001N2O010001O0O101O0O2O00001O00001O00001O00001N10000000000000000000000000000O1001O00001O00000010O0001O0000001O0000lGdLZ7]4I5K0001O0O101O0000001O00001O0000001O000O2O0000001O00001O0000001O00001N2O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1N2M3N2N2N2N2N2N2N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2M7Ha0\\NTFJS5"}, "label": "bear behind another bear"}]} {"id": 90310, "image": "COCO_train2014_000000090310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small black bear standing in front of another bear in the grass\"."}], "task": "refering", "answer_template": "The \"a small black bear standing in front of another bear in the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 95, 96, 97, 98, 99, 100, 101, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 262], "bbox": [0.077671875, 0.22921518987341774, 0.48959374999999994, 0.8516962025316456], "iscrowd": 0, "area": 37877.862949999995, "rle": {"size": [395, 640], "counts": "d^c04W<5J5L1N2O0O2N2O1N101N2O0O2O0O2N1O2M2O2N1O2M2O2N1O2M2O1N2O1N2O1N3N1O1O2N1O1O2N100O1O1O1O1N2kNnMPHS2o7QNmGP2S8TNhGm1W8WNdGk1\\8ZN^Gg1b8]NYGd1g8P1001O0000001O001O2N3M2OaJTM_1j2_NXMc1f2]NZMf1c2ZN]Mg1b2YN_Mg1`2XNaMh1_2XNaMi1^2WNbMj1]2VNcMj1]2VNcMk1\\2UNeMj1[2UNfMl1Y2TNgMm1X2SNhMm1X2SNhMn1W2QNjMo1V2PNkMQ2T2SL[M4c0j3j3RLWLm3m3oKTLo3R4lKoKR4V4iKlKT4Z4hKgKV4^4gKbKX4c4cK^K\\4f4aKZK]4k4`KUK_4P5]KPKb4T5ZKmJf4V5WKjJi4Z5SKfJm4\\60000000000000O100000000O100000000O2O1O1O1N3N2N2M3N2N2M3N2N2M3fITJQ6S60000O10000O10000O100O010O1O100O100O100O100O100O1O100O100O010O100O100O100O1O100O100O10O0100O100O1O100O100O100O1N2L4M3L4M4K4M31O3M4L3N2M3M4L3M3M3N2M2N3M2NO101N10000O101N100000000O2O0000000O101O001O1O1N3N1O1O2N1O1N3N1O1O1O2N1N2O2N1N2M4L3M3M3M4J5K5K6J5K6J6J5K6J6I6J7I7I6J7H8Ia0_OVnm3"}, "label": "a small black bear standing in front of another bear in the grass"}]} {"id": 90310, "image": "COCO_train2014_000000090310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bear cub on the left\"."}], "task": "refering", "answer_template": "The \"the bear cub on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 95, 96, 97, 98, 99, 100, 101, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 262], "bbox": [0.077671875, 0.22921518987341774, 0.48959374999999994, 0.8516962025316456], "iscrowd": 0, "area": 37877.862949999995, "rle": {"size": [395, 640], "counts": "d^c04W<5J5L1N2O0O2N2O1N101N2O0O2O0O2N1O2M2O2N1O2M2O2N1O2M2O1N2O1N2O1N3N1O1O2N1O1O2N100O1O1O1O1N2kNnMPHS2o7QNmGP2S8TNhGm1W8WNdGk1\\8ZN^Gg1b8]NYGd1g8P1001O0000001O001O2N3M2OaJTM_1j2_NXMc1f2]NZMf1c2ZN]Mg1b2YN_Mg1`2XNaMh1_2XNaMi1^2WNbMj1]2VNcMj1]2VNcMk1\\2UNeMj1[2UNfMl1Y2TNgMm1X2SNhMm1X2SNhMn1W2QNjMo1V2PNkMQ2T2SL[M4c0j3j3RLWLm3m3oKTLo3R4lKoKR4V4iKlKT4Z4hKgKV4^4gKbKX4c4cK^K\\4f4aKZK]4k4`KUK_4P5]KPKb4T5ZKmJf4V5WKjJi4Z5SKfJm4\\60000000000000O100000000O100000000O2O1O1O1N3N2N2M3N2N2M3N2N2M3fITJQ6S60000O10000O10000O100O010O1O100O100O100O100O100O1O100O100O010O100O100O100O1O100O100O10O0100O100O1O100O100O100O1N2L4M3L4M4K4M31O3M4L3N2M3M4L3M3M3N2M2N3M2NO101N10000O101N100000000O2O0000000O101O001O1O1N3N1O1O2N1O1N3N1O1O1O2N1N2O2N1N2M4L3M3M3M4J5K5K6J5K6J6J5K6J6I6J7I7I6J7H8Ia0_OVnm3"}, "label": "the bear cub on the left"}]} {"id": 205000, "image": "COCO_train2014_000000205000.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the reflection of a young woman in a green sweater in a mirror\"."}], "task": "refering", "answer_template": "The \"the reflection of a young woman in a green sweater in a mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 188, 189, 190, 203, 204, 205, 217, 218, 219, 232, 233, 234, 235, 247, 248, 249, 250, 262, 263, 264, 265, 277, 278, 279, 280, 281, 292, 293, 294, 295, 296, 307, 308, 309, 310, 311, 322, 323, 324, 325, 337, 338, 339, 340], "bbox": [0.4797658079625293, 0.51125, 0.7442154566744731, 0.997265625], "iscrowd": 0, "area": 26238.724049999997, "rle": {"size": [640, 427], "counts": "YaP49ac0=CC;E;E;E;D:G3M3M3M4K4M3M3M3M3L4M4L3M0O1UNjIlEV6Q:nImES6R:oImEQ6S:PJkEQ6T:QJkEo5T:SJkEm5U:UJhEl5W:VJhEj5X:WJfEj5Y:XJfEh5Y:ZJfEf5Z:P2O1O010000000000000000000O1000000000000000000000O10O1000000001mHgE]4Z:`KiE_4W:_KkEa4V:[KnEd4S:YKPFf4P:XKRFh4o9TKUFk4l9RKWFm4j9oJYFQ5g9mJ\\FR5e9kJ^FT5c9iJ_FW5b9fJaFY5_9eJdFZ5]9cJeF]5\\9`JgF_5Z9^JiFa5W9[JmFe5T9TJSGk5b;O001O1O001O1O1O1O1eN]BWLd=g3RCdKo in this image.", "objects": [{"patches": [173, 174, 188, 189, 190, 203, 204, 205, 217, 218, 219, 232, 233, 234, 235, 247, 248, 249, 250, 262, 263, 264, 265, 277, 278, 279, 280, 281, 292, 293, 294, 295, 296, 307, 308, 309, 310, 311, 322, 323, 324, 325, 337, 338, 339, 340], "bbox": [0.4797658079625293, 0.51125, 0.7442154566744731, 0.997265625], "iscrowd": 0, "area": 26238.724049999997, "rle": {"size": [640, 427], "counts": "YaP49ac0=CC;E;E;E;D:G3M3M3M4K4M3M3M3M3L4M4L3M0O1UNjIlEV6Q:nImES6R:oImEQ6S:PJkEQ6T:QJkEo5T:SJkEm5U:UJhEl5W:VJhEj5X:WJfEj5Y:XJfEh5Y:ZJfEf5Z:P2O1O010000000000000000000O1000000000000000000000O10O1000000001mHgE]4Z:`KiE_4W:_KkEa4V:[KnEd4S:YKPFf4P:XKRFh4o9TKUFk4l9RKWFm4j9oJYFQ5g9mJ\\FR5e9kJ^FT5c9iJ_FW5b9fJaFY5_9eJdFZ5]9cJeF]5\\9`JgF_5Z9^JiFa5W9[JmFe5T9TJSGk5b;O001O1O001O1O1O1O1eN]BWLd=g3RCdKo in this image.", "objects": [{"patches": [33, 46, 47, 48, 49, 60, 61, 62, 63, 64, 65, 75, 76, 77, 78, 79, 80, 90, 91, 92, 93, 94, 95, 96, 105, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 150, 151, 152, 153, 154, 165, 166, 167, 168, 169, 170, 180, 181, 182, 183, 184, 185, 186, 195, 196, 197, 198, 199, 200, 201, 202, 210, 211, 212, 213, 214, 215, 216, 217, 225, 226, 227, 228, 229, 230, 231, 232, 240, 241, 242, 243, 244, 245, 246, 247, 255, 256, 257, 258, 259, 260, 261, 262, 270, 271, 272, 273, 274, 275, 276, 277, 285, 286, 287, 288, 289, 290, 291, 292, 300, 301, 302, 303, 304, 305, 306, 307, 315, 316, 317, 318, 319, 320, 321, 322, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.0, 0.124859375, 0.5323419203747073, 0.98734375], "iscrowd": 0, "area": 94667.01249999998, "rle": {"size": [640, 427], "counts": "`?U4o9l5O1O10000O100iKmEaNS:\\1TF_Nm9^1ZF^Nf9`1`F\\N`9a1gFZNZ9c1mFYNS9e1SGWNm8f1ZGUNg8i1_GSNa8j1fGRNZ8k1lGRNT8l1PHQNQ8l1THRNl7k1ZHRNf7l1^HQNc7l1bHRN^7l1fHRN[7j1jHSNW7k1nHRNR7m1QIQNo6n1TIPNl6o1WInMj6Q2YImMg6R2]IkMc6T2`IiMa6V2bIhM^6W2eIgM[6X2iIdMX6[2kIaMW6^2lI]MW6b2lIYMW6f2lIUMW6j2mIPMV6o2mIlLV6S3mIiLU6V3nIeLU6Y3oIbLT6]3X5O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O10000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O2O000001O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O0`EdL\\4]3aKiL[4X3cKlLZ4U3cKQMY4P3dKVMX4j2fK[MW4f2fK`MV4a2gKeMU4\\2hKiMU4X2hKnMT4R2jKTNR4m1lKUNS4l1jKWNU4j1hKYNW4h1fKZNZ4g1cK\\N\\4f1`K]N_4d1^K_Na4c1[K_Ne4b1XKaNg4a1VKaNi4a1SKaNm4`1PKcNo4_1mJdNR5^1jJdNV5]1gJfNX5\\1dJgN[5[1aJgN_5Z1_JhN`5Z1\\JiNc5X1ZJjNf5X1VJkNi5W1SJlNl5U1QJmNo5f1\\I]Nc6_3_GdL`8_7O001O1O1O1O1O1O1O1O1O1O001O100O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1Og1YN9G:F:F:F9G:F:F9G:F:F9G:F:F:F9G:F:F9G:Fa\\l3"}, "label": "back image of a woman in a green shirt"}]} {"id": 205000, "image": "COCO_train2014_000000205000.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green colour dressed young lady is standing infront of the mirror\"."}], "task": "refering", "answer_template": "The \"a green colour dressed young lady is standing infront of the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 46, 47, 48, 49, 60, 61, 62, 63, 64, 65, 75, 76, 77, 78, 79, 80, 90, 91, 92, 93, 94, 95, 96, 105, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 150, 151, 152, 153, 154, 165, 166, 167, 168, 169, 170, 180, 181, 182, 183, 184, 185, 186, 195, 196, 197, 198, 199, 200, 201, 202, 210, 211, 212, 213, 214, 215, 216, 217, 225, 226, 227, 228, 229, 230, 231, 232, 240, 241, 242, 243, 244, 245, 246, 247, 255, 256, 257, 258, 259, 260, 261, 262, 270, 271, 272, 273, 274, 275, 276, 277, 285, 286, 287, 288, 289, 290, 291, 292, 300, 301, 302, 303, 304, 305, 306, 307, 315, 316, 317, 318, 319, 320, 321, 322, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.0, 0.124859375, 0.5323419203747073, 0.98734375], "iscrowd": 0, "area": 94667.01249999998, "rle": {"size": [640, 427], "counts": "`?U4o9l5O1O10000O100iKmEaNS:\\1TF_Nm9^1ZF^Nf9`1`F\\N`9a1gFZNZ9c1mFYNS9e1SGWNm8f1ZGUNg8i1_GSNa8j1fGRNZ8k1lGRNT8l1PHQNQ8l1THRNl7k1ZHRNf7l1^HQNc7l1bHRN^7l1fHRN[7j1jHSNW7k1nHRNR7m1QIQNo6n1TIPNl6o1WInMj6Q2YImMg6R2]IkMc6T2`IiMa6V2bIhM^6W2eIgM[6X2iIdMX6[2kIaMW6^2lI]MW6b2lIYMW6f2lIUMW6j2mIPMV6o2mIlLV6S3mIiLU6V3nIeLU6Y3oIbLT6]3X5O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O10000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O2O000001O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O0`EdL\\4]3aKiL[4X3cKlLZ4U3cKQMY4P3dKVMX4j2fK[MW4f2fK`MV4a2gKeMU4\\2hKiMU4X2hKnMT4R2jKTNR4m1lKUNS4l1jKWNU4j1hKYNW4h1fKZNZ4g1cK\\N\\4f1`K]N_4d1^K_Na4c1[K_Ne4b1XKaNg4a1VKaNi4a1SKaNm4`1PKcNo4_1mJdNR5^1jJdNV5]1gJfNX5\\1dJgN[5[1aJgN_5Z1_JhN`5Z1\\JiNc5X1ZJjNf5X1VJkNi5W1SJlNl5U1QJmNo5f1\\I]Nc6_3_GdL`8_7O001O1O1O1O1O1O1O1O1O1O001O100O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1Og1YN9G:F:F:F9G:F:F9G:F:F9G:F:F:F9G:F:F9G:Fa\\l3"}, "label": "a green colour dressed young lady is standing infront of the mirror"}]} {"id": 131277, "image": "COCO_train2014_000000131277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a blue jacket\"."}], "task": "refering", "answer_template": "The \"a man wearing a blue jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 62, 63, 64, 85, 86, 87, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 318, 336, 337, 338, 339, 340, 341, 359, 360, 361, 362, 363, 364, 382, 383, 385, 386], "bbox": [0.6331875, 0.07166666666666667, 0.897296875, 0.9633125], "iscrowd": 0, "area": 45963.097749999986, "rle": {"size": [480, 640], "counts": "XYn52f>8I7H8I8G8I7H8I7H8I7mLhMVI`2b6iMUI`2b6jMUI]2b6nMTIZ2d6oMTIX2c6SNSIU2e6UNRIR2e6XNRIP2f6ZNQIn1e6]NQIk1g6[3G9H8G9H8G9H9F9I7O1O1O1O1OO1O1N3NPNiEZOU:f0RFUOl9k0ZFQOe9m0bFoN\\9Q1kFjNS9V1SGfNl8Y1ZGcNd8\\1cGaN[8`1jG[NT8i1nGSNP8P2SHlMl7V2XHeMf7^2]H^Mb7d2aHXM]7k2fHQMX7R3Y23L3M3N2M3N2M3M4M2M3M3N2M3N2M4L3N2M3M3N2M3N3LRN]FeNb9n0lFROT9`0ZGAe80jG0V8BXH?h7ROfHo0Y7cNUI]1k6TNdIm1g91O1O0eDQNe9P2YFQNg9P2XFPNh9P2WFRNh9o1WFQNi9P2UFRNj9n1VFRNj9o1TFRNl9o1SFRNl9o1RFRNn9n1RFRNn9o1PFSNo9n1PFRNP:`3O00001O001O00001O00001O001N101O1O001O1O1O1O1O1O1O1O1K5J6I7I7J5J7I7J6I7I7J6I7J6I7I7oNQ1UNk1eN[1O0O2N2N2O1N2N2N2N2O1N2N2N2O1N2N1O2O1N2N9G in this image.", "objects": [{"patches": [39, 40, 62, 63, 64, 85, 86, 87, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 318, 336, 337, 338, 339, 340, 341, 359, 360, 361, 362, 363, 364, 382, 383, 385, 386], "bbox": [0.6331875, 0.07166666666666667, 0.897296875, 0.9633125], "iscrowd": 0, "area": 45963.097749999986, "rle": {"size": [480, 640], "counts": "XYn52f>8I7H8I8G8I7H8I7H8I7mLhMVI`2b6iMUI`2b6jMUI]2b6nMTIZ2d6oMTIX2c6SNSIU2e6UNRIR2e6XNRIP2f6ZNQIn1e6]NQIk1g6[3G9H8G9H8G9H9F9I7O1O1O1O1OO1O1N3NPNiEZOU:f0RFUOl9k0ZFQOe9m0bFoN\\9Q1kFjNS9V1SGfNl8Y1ZGcNd8\\1cGaN[8`1jG[NT8i1nGSNP8P2SHlMl7V2XHeMf7^2]H^Mb7d2aHXM]7k2fHQMX7R3Y23L3M3N2M3N2M3M4M2M3M3N2M3N2M4L3N2M3M3N2M3N3LRN]FeNb9n0lFROT9`0ZGAe80jG0V8BXH?h7ROfHo0Y7cNUI]1k6TNdIm1g91O1O0eDQNe9P2YFQNg9P2XFPNh9P2WFRNh9o1WFQNi9P2UFRNj9n1VFRNj9o1TFRNl9o1SFRNl9o1RFRNn9n1RFRNn9o1PFSNo9n1PFRNP:`3O00001O001O00001O00001O001N101O1O001O1O1O1O1O1O1O1O1K5J6I7I7J5J7I7J6I7I7J6I7J6I7I7oNQ1UNk1eN[1O0O2N2N2O1N2N2N2N2O1N2N2N2O1N2N1O2O1N2N9G in this image.", "objects": [{"patches": [98, 99, 120, 121, 122, 123, 143, 144, 145, 146, 167, 168, 169, 170, 190, 191, 192, 193, 213, 214, 215, 216, 236, 237, 238, 239, 259, 260, 261, 262, 281, 282, 283, 284, 285, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 353, 354, 355, 356, 372, 373, 374, 377, 378], "bbox": [0.17865625000000002, 0.2494375, 0.49381249999999993, 0.9775208333333334], "iscrowd": 0, "area": 33885.68519999999, "rle": {"size": [480, 640], "counts": "a[f14h>4L4M3L4L4L4M3L4L4L4M3L4L4L4M3L4L4L4M3L4L4L4M3L4L4L4L4M3O1O1O1O1O1O1oI_Lc0a3\\OaLb0`3\\OdLa0]3^OfL?[3@gL>Z3AiLm1HmM9R2LiM5V21dM0Z26bMJ]2;^MEb2?ZMBe2c0VM]Ok2g0PMZOQ3h0kLXOW3k0dLVO\\3m0`LROc3Q1XLnNj3U1RLiNR4Z1iKeNZ4^1aKaNa4b1[KWNn4_5>B?D;N3N2M2N3M2O2M3M2N3N1N3M2N3N2M2N3N13N1O2M3N2M3N2M2O001N101N2O0O21N101N2O0O2O1N101N101N2O00O010O010O010O010O010O00100O01N1O2N101N1O2cJjHl2W7hLUIW3l6]LaIa3a6PLnIo3R6`K`J^4R8O2N2N2N101N2N1O2N2N2\\L`Ee2`:UMiEg2Y:RMoEk2S:nLUFo2Q;M3M3N2M3M3M4L3M3M3M3M3M3M3L5L3M3M3M3L4M3M3M3M2M2O2N2N2N2N1N3N2N2N2N1N3N2Nm]g4"}, "label": "the woman with her hand on her butt"}]} {"id": 131277, "image": "COCO_train2014_000000131277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a female wearing a black shirt\"."}], "task": "refering", "answer_template": "The \"a female wearing a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 120, 121, 122, 123, 143, 144, 145, 146, 167, 168, 169, 170, 190, 191, 192, 193, 213, 214, 215, 216, 236, 237, 238, 239, 259, 260, 261, 262, 281, 282, 283, 284, 285, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 353, 354, 355, 356, 372, 373, 374, 377, 378], "bbox": [0.17865625000000002, 0.2494375, 0.49381249999999993, 0.9775208333333334], "iscrowd": 0, "area": 33885.68519999999, "rle": {"size": [480, 640], "counts": "a[f14h>4L4M3L4L4L4M3L4L4L4M3L4L4L4M3L4L4L4M3L4L4L4M3L4L4L4L4M3O1O1O1O1O1O1oI_Lc0a3\\OaLb0`3\\OdLa0]3^OfL?[3@gL>Z3AiLm1HmM9R2LiM5V21dM0Z26bMJ]2;^MEb2?ZMBe2c0VM]Ok2g0PMZOQ3h0kLXOW3k0dLVO\\3m0`LROc3Q1XLnNj3U1RLiNR4Z1iKeNZ4^1aKaNa4b1[KWNn4_5>B?D;N3N2M2N3M2O2M3M2N3N1N3M2N3N2M2N3N13N1O2M3N2M3N2M2O001N101N2O0O21N101N2O0O2O1N101N101N2O00O010O010O010O010O010O00100O01N1O2N101N1O2cJjHl2W7hLUIW3l6]LaIa3a6PLnIo3R6`K`J^4R8O2N2N2N101N2N1O2N2N2\\L`Ee2`:UMiEg2Y:RMoEk2S:nLUFo2Q;M3M3N2M3M3M4L3M3M3M3M3M3M3L5L3M3M3M3L4M3M3M3M2M2O2N2N2N2N1N3N2N2N2N1N3N2Nm]g4"}, "label": "a female wearing a black shirt"}]} {"id": 254176, "image": "COCO_train2014_000000254176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorcycle covered in no signs\"."}], "task": "refering", "answer_template": "The \"a motorcycle covered in no signs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 118, 119, 120, 121, 122, 135, 136, 137, 138, 139, 140, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 174, 175, 178, 179, 185, 186, 187, 188, 189, 190, 191, 195, 196, 197, 203, 204, 205, 206, 207, 208, 209, 213, 214, 215], "bbox": [0.2803, 0.4903582089552239, 1.0, 1.0], "iscrowd": 0, "area": 31404.22309999999, "rle": {"size": [335, 500], "counts": "]Q^14[:=C>B=C>B=C:F00000000001O00000000000000001O00000000000000001O000000000000001O000000000000000000000000000000000000000000000000000000000000000000000O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O100O100O100N2N2N2O1N2N2N2N2N2O1N2O1N2O1N2O0O2O1N2O1N2O1N2O1N2O1N2O1O1O1O1O1O1O1O1O1O1O1O100O001O100O1O1O100O1O1O1000000000000000000000[NXKcMh4Q2eKnMZ4S2gKlMY4T2hKkMX4U2iKiMX4W2jKgMV4Y2kKfMU4Z2lKaMX4_2iKXM_4i2aKnLg4R3[100000000000000000O10000000001O00000000000000000000000O100000000001O001O00100O0010O01O10O01O010O010001N101N10O1M2N3N1N3N2L3M4M3L4L5K4L4L4M3K6^Oa0_Oa0DUX:f0[gE;D6J60000O1000000O1000000O1000000O1000001N10000O1O1O1O1N2O1O1O1N2O1O1O1N2O001O1N2000000000000000O100000000000000000000O100000000000000000000OQH"}, "label": "a motorcycle covered in no signs"}]} {"id": 254176, "image": "COCO_train2014_000000254176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a side car with signs on it\"."}], "task": "refering", "answer_template": "The \"a side car with signs on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 118, 119, 120, 121, 122, 135, 136, 137, 138, 139, 140, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 174, 175, 178, 179, 185, 186, 187, 188, 189, 190, 191, 195, 196, 197, 203, 204, 205, 206, 207, 208, 209, 213, 214, 215], "bbox": [0.2803, 0.4903582089552239, 1.0, 1.0], "iscrowd": 0, "area": 31404.22309999999, "rle": {"size": [335, 500], "counts": "]Q^14[:=C>B=C>B=C:F00000000001O00000000000000001O00000000000000001O000000000000001O000000000000000000000000000000000000000000000000000000000000000000000O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O100O100O100N2N2N2O1N2N2N2N2N2O1N2O1N2O1N2O0O2O1N2O1N2O1N2O1N2O1N2O1O1O1O1O1O1O1O1O1O1O1O100O001O100O1O1O100O1O1O1000000000000000000000[NXKcMh4Q2eKnMZ4S2gKlMY4T2hKkMX4U2iKiMX4W2jKgMV4Y2kKfMU4Z2lKaMX4_2iKXM_4i2aKnLg4R3[100000000000000000O10000000001O00000000000000000000000O100000000001O001O00100O0010O01O10O01O010O010001N101N10O1M2N3N1N3N2L3M4M3L4L5K4L4L4M3K6^Oa0_Oa0DUX:f0[gE;D6J60000O1000000O1000000O1000000O1000001N10000O1O1O1O1N2O1O1O1N2O1O1O1N2O001O1N2000000000000000O100000000000000000000O100000000000000000000OQH"}, "label": "a side car with signs on it"}]} {"id": 254176, "image": "COCO_train2014_000000254176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stuffed bear riding sidecar on a motorcycle\"."}], "task": "refering", "answer_template": "The \"a stuffed bear riding sidecar on a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 85, 86, 87, 101, 102, 103, 104, 105, 121, 122], "bbox": [0.6065, 0.34253731343283583, 0.8642799999999999, 0.581820895522388], "iscrowd": 0, "area": 6357.441800000003, "rle": {"size": [335, 500], "counts": "PXS3=Q:1O2O001N101N1^F^OW9c0bFMS9h0O1O1O100O1000O1000O10000000000O100000000000000O1000001O000000000000001O001O001O00001O001O001O001O001O00001O0000001O0000000010O0001O1O100O1O1O1O010O1O001O10O01O1O002N100O100O0010O0O2N100O2O1O1N2O0O101O0M3J7]Ob0N3M2N3N1O2N101N1O1O2O0O2M3N1N3MeSf0"}, "label": "a stuffed bear riding sidecar on a motorcycle"}]} {"id": 254176, "image": "COCO_train2014_000000254176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a teddy bear doll wearing glasses\"."}], "task": "refering", "answer_template": "The \"a teddy bear doll wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 85, 86, 87, 101, 102, 103, 104, 105, 121, 122], "bbox": [0.6065, 0.34253731343283583, 0.8642799999999999, 0.581820895522388], "iscrowd": 0, "area": 6357.441800000003, "rle": {"size": [335, 500], "counts": "PXS3=Q:1O2O001N101N1^F^OW9c0bFMS9h0O1O1O100O1000O1000O10000000000O100000000000000O1000001O000000000000001O001O001O00001O001O001O001O001O00001O0000001O0000000010O0001O1O100O1O1O1O010O1O001O10O01O1O002N100O100O0010O0O2N100O2O1O1N2O0O101O0M3J7]Ob0N3M2N3N1O2N101N1O1O2O0O2M3N1N3MeSf0"}, "label": "a teddy bear doll wearing glasses"}]} {"id": 254176, "image": "COCO_train2014_000000254176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small teddy bear next to teddy bear in goggles\"."}], "task": "refering", "answer_template": "The \"small teddy bear next to teddy bear in goggles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 140, 141, 142, 158, 159, 160, 175, 176, 177, 178, 193, 194, 195], "bbox": [0.7464, 0.5569850746268656, 0.93316, 0.8950149253731343], "iscrowd": 0, "area": 6359.630450000001, "rle": {"size": [335, 500], "counts": "dYj31^:1N3M2O2M2N9H0O1O2O0O100O1O1M3G9F:gNjNWIX1c6SOVIo0d6\\OVIe0f6DSI>m6GmH;R7a1O100O10001N10O001O1O0010O01O10000000000001O00000000000000K5D=O000000000000001O1O0RO[HQOg7m0]HPOc7o0aHnN`7Q1dHkN]7S1cHnN]7Q1aHRO`7m0aHRO`7l0aHTO_7k0bHTO`7k0`HUO`7j0bHUO_7i0bHWO_7h0aHXO_7g0cHXO^7f0cHYO^7g0bHYO_7e0cHZO^7d0cH\\O]7d0cH\\O^7b0cH^Oa7=aHAe7:[HFj74WHLn7NTH1U900O2N1O2N4LUa:"}, "label": "small teddy bear next to teddy bear in goggles"}]} {"id": 254176, "image": "COCO_train2014_000000254176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"legs of brown teddy bear hanging outside black car\"."}], "task": "refering", "answer_template": "The \"legs of brown teddy bear hanging outside black car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 140, 141, 142, 158, 159, 160, 175, 176, 177, 178, 193, 194, 195], "bbox": [0.7464, 0.5569850746268656, 0.93316, 0.8950149253731343], "iscrowd": 0, "area": 6359.630450000001, "rle": {"size": [335, 500], "counts": "dYj31^:1N3M2O2M2N9H0O1O2O0O100O1O1M3G9F:gNjNWIX1c6SOVIo0d6\\OVIe0f6DSI>m6GmH;R7a1O100O10001N10O001O1O0010O01O10000000000001O00000000000000K5D=O000000000000001O1O0RO[HQOg7m0]HPOc7o0aHnN`7Q1dHkN]7S1cHnN]7Q1aHRO`7m0aHRO`7l0aHTO_7k0bHTO`7k0`HUO`7j0bHUO_7i0bHWO_7h0aHXO_7g0cHXO^7f0cHYO^7g0bHYO_7e0cHZO^7d0cH\\O]7d0cH\\O^7b0cH^Oa7=aHAe7:[HFj74WHLn7NTH1U900O2N1O2N4LUa:"}, "label": "legs of brown teddy bear hanging outside black car"}]} {"id": 254176, "image": "COCO_train2014_000000254176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"double headlight on motorcycle over the right side of teddy bear\"."}], "task": "refering", "answer_template": "The \"double headlight on motorcycle over the right side of teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 106, 107, 123, 124, 125, 142, 143, 161], "bbox": [0.8719600000000001, 0.31946268656716414, 0.99882, 0.744955223880597], "iscrowd": 0, "area": 4245.391899999997, "rle": {"size": [335, 500], "counts": "Xj^41]:2N2N1O2N2L4E;L4N2N2O1N2N2O1N2O1N2N3L4L4L4F:K4M4L4M3TObMaIa2Z6fMbI\\2\\6hM_I[2`6k0OfL\\Ii2f6TM^Ij2c6RMdIi2U7M9FO2N1O2N101N1O2N101N1O2O00QNdHk0\\7SOhHl0X7ROkHl0U7SOkHo0U7nNmHR1U7kNkHV1W7gNjHZ1V7cNkH^1W7_NjHb1W7[NiHg1W7WNiHm1V7oMlHT2R7jMnHY2Q7eMPI]2_7VNWIMQ<"}, "label": "double headlight on motorcycle over the right side of teddy bear"}]} {"id": 254176, "image": "COCO_train2014_000000254176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black motor bike with two head lights\"."}], "task": "refering", "answer_template": "The \"a black motor bike with two head lights\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 106, 107, 123, 124, 125, 142, 143, 161], "bbox": [0.8719600000000001, 0.31946268656716414, 0.99882, 0.744955223880597], "iscrowd": 0, "area": 4245.391899999997, "rle": {"size": [335, 500], "counts": "Xj^41]:2N2N1O2N2L4E;L4N2N2O1N2N2O1N2O1N2N3L4L4L4F:K4M4L4M3TObMaIa2Z6fMbI\\2\\6hM_I[2`6k0OfL\\Ii2f6TM^Ij2c6RMdIi2U7M9FO2N1O2N101N1O2N101N1O2O00QNdHk0\\7SOhHl0X7ROkHl0U7SOkHo0U7nNmHR1U7kNkHV1W7gNjHZ1V7cNkH^1W7_NjHb1W7[NiHg1W7WNiHm1V7oMlHT2R7jMnHY2Q7eMPI]2_7VNWIMQ<"}, "label": "a black motor bike with two head lights"}]} {"id": 254176, "image": "COCO_train2014_000000254176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorbike occupied by two men dressed like teddy bear\"."}], "task": "refering", "answer_template": "The \"a motorbike occupied by two men dressed like teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 73, 75, 78, 79, 91, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 180, 181, 182, 183, 184, 198, 199, 200, 201, 202], "bbox": [0.0, 0.29776119402985074, 0.44666, 0.9756119402985076], "iscrowd": 0, "area": 23166.458949999993, "rle": {"size": [335, 500], "counts": "n97T8JaId0_6f10000000000000000000000000000000000aNiMfJW2T5TNgJl1V5[NfJe1X5_NfJa1Y5cNdJ]1Z5gNdJY1Z5kNdJU1[5oNbJQ1]5ROaJn0_5TO_Jl0`5UO`Jk0_5WO`Ji0`5XO^Ji0a5XO]Jj0c5WO[Jj0d5WOZJk0f5VOXJk0g5WOUJl0j5m1000000O1bNgKdL[4Z3hKdLY4Z3kKcLV4\\3lKbLU4_3lK`LS4a3mK^LS4c3mK\\LS4e3nKYLR4h3mKXLS4l3jKTLU4n3jKQLV4P4iKPLW4Q4iKnKW4S4iKmKV4U4hKkKX4V4hKiKX4X4hKhKW4X4jKgKV4Z4iKfKW4[4iKdKW4\\4iKeKV4\\4iKdKW4]4iKbKW4_4hKaKX4_4hKbKW4_4iK`KW4`4iK_KX4b4gK[K\\4f4cKXK_4h4bKUK`4l4_KSKb4m4^KQKd4P5[KoJf4Q5[KmJf4T5YKkJh4U591O001O1mJoJf4_5N2N001O001O001O2N2N2NROeKjKY4j3XLSLf3j3bLRL]3k3kLQLT3o3lLRLS3m3nLSLR3m3nLTLQ3k3PMULP3k3PMVLo2i3RMXLm2h3SMXLm2h3SMXLm2h3SMWLn2j3QMVLo2j3QMULP3l3oLTLQ3l3oLSLR3n3mLRLS3n3mLQLU3n3kLRLU3o3jLPLW3P4iLPLW3R4gLmKZ3T4eLlK[3V4cLiK^3X4aLhK^OAf3n4eL`KEFc3b4TLgK=1JId3_4VLfK<1JKd3^4VLgK;0JMd3\\4XLgK:NKOc3\\4XLgK:NKOd3[4XLhK8ML2c3Y4YLhK7NM1JIa3`4bLhK5MO3HHb3`4bLkK1K32HI`3`4dLkK0J42HI[2JhNf4QOjK1J3P2CaN\\4\\OgK0K4?o1E_N[4]OeK2L2`0o1GUNEFc42eK2K3:GCT1LKc0WO`4OXK1K3;HCR10Hc0UO_45UK1K3:HEQ1n6SOWI1L3:IFn0n6UOVI1L39JGl0o6VOUI1L29LGk0P7VOSI1L39LHj0P7VOSI1L37NJh0P7WORI0M36OKg0P7XOQION350Lg0o6WORION340Ng0n6XOQINN53N1g0m6YOPIMO51N5g0k6YOoHN040N7g0j6ZOnHM15NM:g0i6ZOmHN24ML=i0f6ZOmHM34KMa0g0d6[OlHN4==:c6\\OkHL6<>]6?dIA\\6?eI@Z6`0gI@Y6`0gI@Y6`0gI@Y6?hIAX6?iI@W6`0iIAV6LVI^Oe0e0U6JZI@a0f0U6I[IAa0e0T6H^IB>f0T6F`ID=e0S6FaIEm5MoIE38S6GWJ2JKU63RJ2V7NlH0T70mH0R70PIOo61RIOm61UINk62UINj62WINh62ZIMf62[IOc61_IN`63`IM`62bINS6MYH6e1LP6?PJAm5a0UJ_Og5d0YJ\\Of5d0\\J[Oc5f0^JZO`5f0aJZO]5h0dJWO\\5i0dJXO[5h0gJVOY5j0hJVOW5k0iJTOW5l0jJTOV5k0W21O0000O101O000O101O2M2O1N3N1N2M6I[`j2"}, "label": "a motorbike occupied by two men dressed like teddy bear"}]} {"id": 254176, "image": "COCO_train2014_000000254176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"motorcyle under humans with costumes\"."}], "task": "refering", "answer_template": "The \"motorcyle under humans with costumes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 73, 75, 78, 79, 91, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 180, 181, 182, 183, 184, 198, 199, 200, 201, 202], "bbox": [0.0, 0.29776119402985074, 0.44666, 0.9756119402985076], "iscrowd": 0, "area": 23166.458949999993, "rle": {"size": [335, 500], "counts": "n97T8JaId0_6f10000000000000000000000000000000000aNiMfJW2T5TNgJl1V5[NfJe1X5_NfJa1Y5cNdJ]1Z5gNdJY1Z5kNdJU1[5oNbJQ1]5ROaJn0_5TO_Jl0`5UO`Jk0_5WO`Ji0`5XO^Ji0a5XO]Jj0c5WO[Jj0d5WOZJk0f5VOXJk0g5WOUJl0j5m1000000O1bNgKdL[4Z3hKdLY4Z3kKcLV4\\3lKbLU4_3lK`LS4a3mK^LS4c3mK\\LS4e3nKYLR4h3mKXLS4l3jKTLU4n3jKQLV4P4iKPLW4Q4iKnKW4S4iKmKV4U4hKkKX4V4hKiKX4X4hKhKW4X4jKgKV4Z4iKfKW4[4iKdKW4\\4iKeKV4\\4iKdKW4]4iKbKW4_4hKaKX4_4hKbKW4_4iK`KW4`4iK_KX4b4gK[K\\4f4cKXK_4h4bKUK`4l4_KSKb4m4^KQKd4P5[KoJf4Q5[KmJf4T5YKkJh4U591O001O1mJoJf4_5N2N001O001O001O2N2N2NROeKjKY4j3XLSLf3j3bLRL]3k3kLQLT3o3lLRLS3m3nLSLR3m3nLTLQ3k3PMULP3k3PMVLo2i3RMXLm2h3SMXLm2h3SMXLm2h3SMWLn2j3QMVLo2j3QMULP3l3oLTLQ3l3oLSLR3n3mLRLS3n3mLQLU3n3kLRLU3o3jLPLW3P4iLPLW3R4gLmKZ3T4eLlK[3V4cLiK^3X4aLhK^OAf3n4eL`KEFc3b4TLgK=1JId3_4VLfK<1JKd3^4VLgK;0JMd3\\4XLgK:NKOc3\\4XLgK:NKOd3[4XLhK8ML2c3Y4YLhK7NM1JIa3`4bLhK5MO3HHb3`4bLkK1K32HI`3`4dLkK0J42HI[2JhNf4QOjK1J3P2CaN\\4\\OgK0K4?o1E_N[4]OeK2L2`0o1GUNEFc42eK2K3:GCT1LKc0WO`4OXK1K3;HCR10Hc0UO_45UK1K3:HEQ1n6SOWI1L3:IFn0n6UOVI1L39JGl0o6VOUI1L29LGk0P7VOSI1L39LHj0P7VOSI1L37NJh0P7WORI0M36OKg0P7XOQION350Lg0o6WORION340Ng0n6XOQINN53N1g0m6YOPIMO51N5g0k6YOoHN040N7g0j6ZOnHM15NM:g0i6ZOmHN24ML=i0f6ZOmHM34KMa0g0d6[OlHN4==:c6\\OkHL6<>]6?dIA\\6?eI@Z6`0gI@Y6`0gI@Y6`0gI@Y6?hIAX6?iI@W6`0iIAV6LVI^Oe0e0U6JZI@a0f0U6I[IAa0e0T6H^IB>f0T6F`ID=e0S6FaIEm5MoIE38S6GWJ2JKU63RJ2V7NlH0T70mH0R70PIOo61RIOm61UINk62UINj62WINh62ZIMf62[IOc61_IN`63`IM`62bINS6MYH6e1LP6?PJAm5a0UJ_Og5d0YJ\\Of5d0\\J[Oc5f0^JZO`5f0aJZO]5h0dJWO\\5i0dJXO[5h0gJVOY5j0hJVOW5k0iJTOW5l0jJTOV5k0W21O0000O101O000O101O2M2O1N3N1N2M6I[`j2"}, "label": "motorcyle under humans with costumes"}]} {"id": 532711, "image": "COCO_train2014_000000532711.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair behind the guy wearing the stripes\"."}], "task": "refering", "answer_template": "The \"the chair behind the guy wearing the stripes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 174, 175, 176, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 241, 242, 243, 244, 265], "bbox": [0.46935937499999997, 0.3569791666666667, 0.7, 0.6688124999999999], "iscrowd": 0, "area": 9688.921199999999, "rle": {"size": [480, 640], "counts": "hZ]46i>6K5K6J1O0O2O001O00002M2O2N1O2N2M2O2N1O2N2M2O2N1O2M2O2N2N1N3N1O2N2N1N3N1O2N2M2O2M01O01O01O00001O010O00001O00010O001O000100O00100O0010O0100O0010O01O10O01O010O1O010O10O01O0O2N2O0O2O1N1TOTMdEn2\\:UM`Ek2b:WMYEk2h:WMTEj2l:YMoDh2S;YMiDi2W;`01N101N102N2M3N3L3N2M3M3N2M3M3M3M3M3N2M3M3M4M2M3N2N2N2N2N2M3N2N2N2N2N2N2M4M2M3N2M3N2M3N2M3N2N2M3Negi2"}, "label": "the chair behind the guy wearing the stripes"}]} {"id": 532711, "image": "COCO_train2014_000000532711.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown chair to the top left of pizza\"."}], "task": "refering", "answer_template": "The \"a brown chair to the top left of pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 174, 175, 176, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 241, 242, 243, 244, 265], "bbox": [0.46935937499999997, 0.3569791666666667, 0.7, 0.6688124999999999], "iscrowd": 0, "area": 9688.921199999999, "rle": {"size": [480, 640], "counts": "hZ]46i>6K5K6J1O0O2O001O00002M2O2N1O2N2M2O2N1O2N2M2O2N1O2M2O2N2N1N3N1O2N2N1N3N1O2N2M2O2M01O01O01O00001O010O00001O00010O001O000100O00100O0010O0100O0010O01O10O01O010O1O010O10O01O0O2N2O0O2O1N1TOTMdEn2\\:UM`Ek2b:WMYEk2h:WMTEj2l:YMoDh2S;YMiDi2W;`01N101N102N2M3N3L3N2M3M3N2M3M3M3M3M3N2M3M3M4M2M3N2N2N2N2N2M3N2N2N2N2N2N2M4M2M3N2M3N2M3N2M3N2N2M3Negi2"}, "label": "a brown chair to the top left of pizza"}]} {"id": 532711, "image": "COCO_train2014_000000532711.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person off the screen in a white shirt next to a pizza\"."}], "task": "refering", "answer_template": "The \"a person off the screen in a white shirt next to a pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [322, 326, 327, 328, 329, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.010859375000000001, 0.8055, 0.40542187500000004, 0.9854375], "iscrowd": 0, "area": 11218.30375, "rle": {"size": [480, 640], "counts": "Vg3b0[=S1POP12N2N3M2N2N2N3M2N2N2N2N2N2N2N1O2N2N2N2N1O2N2N2N2N2N1O001O00001O001O001O00001O001O00001O001O001O00001O001O00001OO100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O2O0O100O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000000000O10000000000O1000000000000O1000000000000000000000000000000000000001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O00001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O000O2M2M4M2N3M2M4M2N3M2NSUb5"}, "label": "a person off the screen in a white shirt next to a pizza"}]} {"id": 532711, "image": "COCO_train2014_000000532711.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right shoulder of someone wearing a white shirt and a vest to the left of a pizza\"."}], "task": "refering", "answer_template": "The \"the right shoulder of someone wearing a white shirt and a vest to the left of a pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [322, 326, 327, 328, 329, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.010859375000000001, 0.8055, 0.40542187500000004, 0.9854375], "iscrowd": 0, "area": 11218.30375, "rle": {"size": [480, 640], "counts": "Vg3b0[=S1POP12N2N3M2N2N2N3M2N2N2N2N2N2N2N1O2N2N2N2N1O2N2N2N2N2N1O001O00001O001O001O00001O001O00001O001O001O00001O001O00001OO100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O2O0O100O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000000000O10000000000O1000000000000O1000000000000000000000000000000000000001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O00001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O000O2M2M4M2N3M2M4M2N3M2NSUb5"}, "label": "the right shoulder of someone wearing a white shirt and a vest to the left of a pizza"}]} {"id": 532711, "image": "COCO_train2014_000000532711.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"table nearest woman in white\"."}], "task": "refering", "answer_template": "The \"table nearest woman in white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 214, 215, 216, 217], "bbox": [0.300015625, 0.32997916666666666, 0.5972500000000001, 0.56575], "iscrowd": 0, "area": 8864.072650000007, "rle": {"size": [480, 640], "counts": "\\Xj21m>3N1N3M2N3O0O1O1O1O100O1O1O00100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O1_OhNYCY1f in this image.", "objects": [{"patches": [102, 103, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 194, 195, 196], "bbox": [0.379359375, 0.3211971830985916, 0.6117812499999999, 0.5891314553990611], "iscrowd": 0, "area": 10217.403250000001, "rle": {"size": [426, 640], "counts": "hXU33V<4`D0_;9VDKi;k001O1N101O0O2O001O1N101O001N101O1N101O001N101O1N101O001N101O1O0O2O1O1N3N1O1O1N2O1O1N2O1O1O2M10000O1000O10oE^MY9c2cFaM]9_2^FgMa9Y2ZFlMf9S2WFRNh9h2001JUFlLj9l2^FTMb9e2fF[MY9e2gF[MZ9d2gF\\MX9d2hF\\MX9d2hF[MY9e2eF\\M]9d2aF\\M`9d2^F]Mc9c2\\F]Me9c2YF^Mi9a2WF^Mj9c2YFXMh9h2=00010O000001O01O0001O00100O001O010O001O00100O001O0010O01O1O010O001N1N3N1N3N2M2O2N2O1O1O001O1N2O1O1O001N2O1O1O001O1N2O1O1O001N2O1O1O1O9G8G9H7I5KdnV3"}, "label": "umbrella white orange and black"}]} {"id": 24808, "image": "COCO_train2014_000000024808.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umbrella of the woman in front\"."}], "task": "refering", "answer_template": "The \"the umbrella of the woman in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 194, 195, 196], "bbox": [0.379359375, 0.3211971830985916, 0.6117812499999999, 0.5891314553990611], "iscrowd": 0, "area": 10217.403250000001, "rle": {"size": [426, 640], "counts": "hXU33V<4`D0_;9VDKi;k001O1N101O0O2O001O1N101O001N101O1N101O001N101O1N101O001N101O1O0O2O1O1N3N1O1O1N2O1O1N2O1O1O2M10000O1000O10oE^MY9c2cFaM]9_2^FgMa9Y2ZFlMf9S2WFRNh9h2001JUFlLj9l2^FTMb9e2fF[MY9e2gF[MZ9d2gF\\MX9d2hF\\MX9d2hF[MY9e2eF\\M]9d2aF\\M`9d2^F]Mc9c2\\F]Me9c2YF^Mi9a2WF^Mj9c2YFXMh9h2=00010O000001O01O0001O00100O001O010O001O00100O001O0010O01O1O010O001N1N3N1N3N2M2O2N2O1O1O001O1N2O1O1O001N2O1O1O001O1N2O1O1O001N2O1O1O1O9G8G9H7I5KdnV3"}, "label": "the umbrella of the woman in front"}]} {"id": 33017, "image": "COCO_train2014_000000033017.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"ford suv\"."}], "task": "refering", "answer_template": "The \"ford suv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 136, 137, 138, 139, 140, 141, 142, 143, 145, 146, 147, 148, 153, 154, 163, 164], "bbox": [0.0, 0.0, 0.9067872340425532, 0.41839062499999996], "iscrowd": 0, "area": 95152.77429999999, "rle": {"size": [640, 470], "counts": "3h0Uc0^1eN[1eN\\1dN[1eN[1eN=C0000000000000000O100000000000000000000000000O100000000000000000000000000O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1000000000000001O0000000000000000000000001O000000000000000000000000001O000000000000000000000000001O0000000000000000000000001O000000000000000000000000001O000000000000O1000000000000000000O10000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O100000000O1N2N2O100O100O100O10000O100O100O100O10000O100O100O100O100O1O100O100O1O100O1000000000000000000000000000000000000000000000000000000000000001O1O1O1O1O2N1O1O1O1O2N1O1O2N1O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N2N000000001O0000000000000000001O00000000000000001O0000N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2O1M3N2M3M3N2M3N2O100O1000000O100000000O1000000O100000000cNkBnKV=R4RCfKnRLWAm3i>RLZAk3g>ULZAj3f>VL\\Ah3d>XL^Af3b>YL`Af3`>ZLbAd3^>\\LcAc3]>]LeAa3[>^LhA`3X>`LiA_3W>aLkA]3U>cLlAZ3V>fLlBT2X=kMmBm1W=SNnBd1V=\\NnB\\1V=dNoBS1T=mNQCk0S=UORCb0R=^ORC:R=FRC2R=MTCJP=6]30000O1000o_k0"}, "label": "ford suv"}]} {"id": 33017, "image": "COCO_train2014_000000033017.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a car on the road\"."}], "task": "refering", "answer_template": "The \"a car on the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 136, 137, 138, 139, 140, 141, 142, 143, 145, 146, 147, 148, 153, 154, 163, 164], "bbox": [0.0, 0.0, 0.9067872340425532, 0.41839062499999996], "iscrowd": 0, "area": 95152.77429999999, "rle": {"size": [640, 470], "counts": "3h0Uc0^1eN[1eN\\1dN[1eN[1eN=C0000000000000000O100000000000000000000000000O100000000000000000000000000O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1000000000000001O0000000000000000000000001O000000000000000000000000001O000000000000000000000000001O0000000000000000000000001O000000000000000000000000001O000000000000O1000000000000000000O10000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O100000000O1N2N2O100O100O100O10000O100O100O100O10000O100O100O100O100O1O100O100O1O100O1000000000000000000000000000000000000000000000000000000000000001O1O1O1O1O2N1O1O1O1O2N1O1O2N1O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N2N000000001O0000000000000000001O00000000000000001O0000N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2O1M3N2M3M3N2M3N2O100O1000000O100000000O1000000O100000000cNkBnKV=R4RCfKnRLWAm3i>RLZAk3g>ULZAj3f>VL\\Ah3d>XL^Af3b>YL`Af3`>ZLbAd3^>\\LcAc3]>]LeAa3[>^LhA`3X>`LiA_3W>aLkA]3U>cLlAZ3V>fLlBT2X=kMmBm1W=SNnBd1V=\\NnB\\1V=dNoBS1T=mNQCk0S=UORCb0R=^ORC:R=FRC2R=MTCJP=6]30000O1000o_k0"}, "label": "a car on the road"}]} {"id": 213241, "image": "COCO_train2014_000000213241.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"rectangular decorated cake\"."}], "task": "refering", "answer_template": "The \"rectangular decorated cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 97, 98, 99, 100, 101, 102, 108, 109, 110, 111, 119, 120, 121, 122, 123, 124, 125, 131, 132, 133, 134, 141, 142, 143, 144, 145, 146, 147, 148, 154, 155, 156, 157, 158, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 338, 339, 340, 341, 342, 365], "bbox": [0.06879687500000001, 0.21720833333333334, 0.906046875, 0.8883541666666666], "iscrowd": 0, "area": 120970.55344999999, "rle": {"size": [480, 640], "counts": "jld0`0j=i0K6J5Ko0QO9G9G9G3M2N2N2N3M2M3N2N2N2N3M2N2N2N2N2O2N100O1O100O2N100O1O100O2N100O1O1O100O2N100O1O100O2N100N2O1O1O2N1O1O1O1N2O2N1O1O1O1O1N3N1O1O1O1O1O2N1N2O1O1O1O2N1O1N2O1O1O2N1O1O1O1N2O1O1O2N1O1N2O1O1O1O1O1N3N1O1O1O1O1N2O1O2N1O1O1N2O1O1O1O2N1N2O1O1O1O1O1N3N1O1O1O1O1000001O01O00000000000001O000000000000001O0001O0000000001O000000000000001O00000001O000001O000000000000001O000000000001O0001O000000000000001O000000000000010O000000000000001O000000000000001O01O00000000001O01O000001O000000010O0000001jGgIc7Y6[HkIc7U6[HnId7R6YHSJf7l5XHXJf7i5WH[Jb0[O^6Z6nH^Jb0@Y6o6dIYIU6i6hI_IQ6b6mIeIm5]6PJkIi5W6TJQJf5o5XJTJf5n5WJSJi5b700000000001O000001O00000001O00000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O00000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000001O0000000001O0001O0000000001O000001OO10000O2O00000O10000O101O000O1000000O101O0O1000000O10000O2O00000O100N2UN^IVKd6g4`IVKc6f4bIVK`6g4eIUK]6i4fITK\\6j4hIRKZ6n4hImJ[6S5fIjJ\\6V5fIeJ]6[5dIbJ_6]5cI^J`6b5bIYJa6h5_ITJd6l5^IoIe6Q6U1000001O000001O00000001O00000001O000001O0000000001O0001O000000000001O01O000000000000010O0000000000000010O00000000000001O01O000000000001O04L5K4L5K4L4L5L3L5K4L5K4L4L5K4M4K4L5K4L5K4L4L5L3L5K4L5K4L4L5lNS1jNW1iNV1jNYlk0"}, "label": "rectangular decorated cake"}]} {"id": 213241, "image": "COCO_train2014_000000213241.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white and green color cake\"."}], "task": "refering", "answer_template": "The \"white and green color cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 97, 98, 99, 100, 101, 102, 108, 109, 110, 111, 119, 120, 121, 122, 123, 124, 125, 131, 132, 133, 134, 141, 142, 143, 144, 145, 146, 147, 148, 154, 155, 156, 157, 158, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 338, 339, 340, 341, 342, 365], "bbox": [0.06879687500000001, 0.21720833333333334, 0.906046875, 0.8883541666666666], "iscrowd": 0, "area": 120970.55344999999, "rle": {"size": [480, 640], "counts": "jld0`0j=i0K6J5Ko0QO9G9G9G3M2N2N2N3M2M3N2N2N2N3M2N2N2N2N2O2N100O1O100O2N100O1O100O2N100O1O1O100O2N100O1O100O2N100N2O1O1O2N1O1O1O1N2O2N1O1O1O1O1N3N1O1O1O1O1O2N1N2O1O1O1O2N1O1N2O1O1O2N1O1O1O1N2O1O1O2N1O1N2O1O1O1O1O1N3N1O1O1O1O1N2O1O2N1O1O1N2O1O1O1O2N1N2O1O1O1O1O1N3N1O1O1O1O1000001O01O00000000000001O000000000000001O0001O0000000001O000000000000001O00000001O000001O000000000000001O000000000001O0001O000000000000001O000000000000010O000000000000001O000000000000001O01O00000000001O01O000001O000000010O0000001jGgIc7Y6[HkIc7U6[HnId7R6YHSJf7l5XHXJf7i5WH[Jb0[O^6Z6nH^Jb0@Y6o6dIYIU6i6hI_IQ6b6mIeIm5]6PJkIi5W6TJQJf5o5XJTJf5n5WJSJi5b700000000001O000001O00000001O00000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O00000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O000001O0000000001O0001O0000000001O000001OO10000O2O00000O10000O101O000O1000000O101O0O1000000O10000O2O00000O100N2UN^IVKd6g4`IVKc6f4bIVK`6g4eIUK]6i4fITK\\6j4hIRKZ6n4hImJ[6S5fIjJ\\6V5fIeJ]6[5dIbJ_6]5cI^J`6b5bIYJa6h5_ITJd6l5^IoIe6Q6U1000001O000001O00000001O00000001O000001O0000000001O0001O000000000001O01O000000000000010O0000000000000010O00000000000001O01O000000000001O04L5K4L5K4L4L5L3L5K4L5K4L4L5K4M4K4L5K4L5K4L4L5L3L5K4L5K4L4L5lNS1jNW1iNV1jNYlk0"}, "label": "white and green color cake"}]} {"id": 147710, "image": "COCO_train2014_000000147710.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red suitcase\"."}], "task": "refering", "answer_template": "The \"a red suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 133, 134, 135, 152, 153, 171, 186, 187, 189, 190, 204, 205, 207], "bbox": [0.34628, 0.4786133333333333, 0.55912, 0.9132800000000001], "iscrowd": 0, "area": 5031.6416500000005, "rle": {"size": [375, 500], "counts": "Zdo1b0U;5K4L5K2N1O2O001N101N101N2O06KN101N10O001IdEPOe:h06O001O1O1O001O0BVE0i:LnS20dlM1O2N2N1O2N2N1J7N2N2N1O2N2\\ETOZ:m0dEVO\\:j0bEWO_:Q1000O010000TOjNVGU1j8nNTGR1k8POSGP1m8SOQGm0n8SOSGm0l8TOTGk0l8VOSGk0l8UOUGj0j8XOUGi0l8VOSGl0m8ROSGQ1m8kNUGU1l8gNVG[1b91O2O0O2M2M4L4K6E;F9G:K5M3L3N3M3O0010O10O01O001f0ZOK7I>C=B>@`0@9G:F:F:F_Y`2"}, "label": "a red suitcase"}]} {"id": 147710, "image": "COCO_train2014_000000147710.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red suitcase that the little boy is holding on to\"."}], "task": "refering", "answer_template": "The \"the red suitcase that the little boy is holding on to\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 133, 134, 135, 152, 153, 171, 186, 187, 189, 190, 204, 205, 207], "bbox": [0.34628, 0.4786133333333333, 0.55912, 0.9132800000000001], "iscrowd": 0, "area": 5031.6416500000005, "rle": {"size": [375, 500], "counts": "Zdo1b0U;5K4L5K2N1O2O001N101N101N2O06KN101N10O001IdEPOe:h06O001O1O1O001O0BVE0i:LnS20dlM1O2N2N1O2N2N1J7N2N2N1O2N2\\ETOZ:m0dEVO\\:j0bEWO_:Q1000O010000TOjNVGU1j8nNTGR1k8POSGP1m8SOQGm0n8SOSGm0l8TOTGk0l8VOSGk0l8UOUGj0j8XOUGi0l8VOSGl0m8ROSGQ1m8kNUGU1l8gNVG[1b91O2O0O2M2M4L4K6E;F9G:K5M3L3N3M3O0010O10O01O001f0ZOK7I>C=B>@`0@9G:F:F:F_Y`2"}, "label": "the red suitcase that the little boy is holding on to"}]} {"id": 499966, "image": "COCO_train2014_000000499966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue and white striped couch\"."}], "task": "refering", "answer_template": "The \"a blue and white striped couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 210, 211, 212, 213, 214, 215, 221, 222, 223, 224, 225, 226, 227, 228, 233, 234, 235, 244, 245, 246, 247, 248, 249, 250, 251, 256, 257, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 279, 292, 293, 294, 295, 296, 297, 316, 317, 318, 319, 320, 339, 340, 341, 342, 343], "bbox": [0.130375, 0.4473125, 0.9580625000000002, 0.8860208333333334], "iscrowd": 0, "area": 42648.31225, "rle": {"size": [480, 640], "counts": "aUW12e>=F9L4L3M3M2N2M4M2N2O1N2O1N2O2N2M4MY1gN000O1000O1O00100O1O001O1O100O001O1O1O00100O1O001O1O100O1O1O001O1O100O1M3N2O1N20001N100O100O10000O1KWCRNj\\IhNe6V1d3M3N2M3M2O2M2O2N101N1O2N2N1O2N2O1N3M3M2N3M3M2Kh\\<"}, "label": "a blue and white striped couch"}]} {"id": 499966, "image": "COCO_train2014_000000499966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and blue lined three seater sofa\"."}], "task": "refering", "answer_template": "The \"a white and blue lined three seater sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 210, 211, 212, 213, 214, 215, 221, 222, 223, 224, 225, 226, 227, 228, 233, 234, 235, 244, 245, 246, 247, 248, 249, 250, 251, 256, 257, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 279, 292, 293, 294, 295, 296, 297, 316, 317, 318, 319, 320, 339, 340, 341, 342, 343], "bbox": [0.130375, 0.4473125, 0.9580625000000002, 0.8860208333333334], "iscrowd": 0, "area": 42648.31225, "rle": {"size": [480, 640], "counts": "aUW12e>=F9L4L3M3M2N2M4M2N2O1N2O1N2O2N2M4MY1gN000O1000O1O00100O1O001O1O100O001O1O1O00100O1O001O1O100O1O1O001O1O100O1M3N2O1N20001N100O100O10000O1KWCRNj\\IhNe6V1d3M3N2M3M2O2M2O2N101N1O2N2N1O2N2O1N3M3M2N3M3M2Kh\\<"}, "label": "a white and blue lined three seater sofa"}]} {"id": 499966, "image": "COCO_train2014_000000499966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a middle aged man sitting on a sofa with black dog\"."}], "task": "refering", "answer_template": "The \"a middle aged man sitting on a sofa with black dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 172, 173, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 288, 289], "bbox": [0.350078125, 0.3798958333333333, 0.6232656249999999, 0.7114166666666667], "iscrowd": 0, "area": 13651.964249999997, "rle": {"size": [480, 640], "counts": "WYY32i>6K5M3M3M3M2K5L4N2N2N2O00100O100O10O1N2D;O2N2N1100000O100000O01000O101N10000O1cB_O[bMb9m0oEb1k:`NTE`1l:aNRE`1m:bNQE^1o:eNmD]1R;Y100O1O010O100O02O001N101O1O0O2O001O1N2O2NWEjLi9T3VFQMi9m2TFVMm9i2QFYMQ:9bEZ1<_NU:1cE_16bNW;]1gDeNZ;Z1dDhN^;U1_DoNa;Q1[DSOe;S210000000000000000000000000001O1O001O1O2N1O2N00000000000ZOSDYNm;f1WDWNi;g1\\DVNd;i1_DUNa;h1dDVN\\;h1hDVNX;g1S1M3L4M3M3L5K4M3L4L4M3LRg`3"}, "label": "a middle aged man sitting on a sofa with black dog"}]} {"id": 499966, "image": "COCO_train2014_000000499966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man seated on a sofa\"."}], "task": "refering", "answer_template": "The \"a man seated on a sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 172, 173, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 288, 289], "bbox": [0.350078125, 0.3798958333333333, 0.6232656249999999, 0.7114166666666667], "iscrowd": 0, "area": 13651.964249999997, "rle": {"size": [480, 640], "counts": "WYY32i>6K5M3M3M3M2K5L4N2N2N2O00100O100O10O1N2D;O2N2N1100000O100000O01000O101N10000O1cB_O[bMb9m0oEb1k:`NTE`1l:aNRE`1m:bNQE^1o:eNmD]1R;Y100O1O010O100O02O001N101O1O0O2O001O1N2O2NWEjLi9T3VFQMi9m2TFVMm9i2QFYMQ:9bEZ1<_NU:1cE_16bNW;]1gDeNZ;Z1dDhN^;U1_DoNa;Q1[DSOe;S210000000000000000000000000001O1O001O1O2N1O2N00000000000ZOSDYNm;f1WDWNi;g1\\DVNd;i1_DUNa;h1dDVN\\;h1hDVNX;g1S1M3L4M3M3L5K4M3L4L4M3LRg`3"}, "label": "a man seated on a sofa"}]} {"id": 499966, "image": "COCO_train2014_000000499966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red pillow\"."}], "task": "refering", "answer_template": "The \"a red pillow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [257, 258, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.0, 0.6545416666666667, 0.5486718749999999, 1.0], "iscrowd": 0, "area": 45690.51789999999, "rle": {"size": [480, 640], "counts": "i:o0Q>m1SN[1eN00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1001O000000001O0000001O2N3M3M3M3M3M3M3M3M3M1O0000O1000000000000000000000000000000000000O100000000000000000000000000000000000000O1000000000000000000000000000000000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000001O00001O00001O00001O00001O00001O00001O001O001O1O1O0O2O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001N2O1O001O1O1O001O1O1M2O2M3N2M2O2N2M3N3K4K6I6J7J5J6J7J5J7I6K6I^RW4"}, "label": "a red pillow"}]} {"id": 499966, "image": "COCO_train2014_000000499966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"couch and pillow in the left bottom corner\"."}], "task": "refering", "answer_template": "The \"couch and pillow in the left bottom corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [257, 258, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.0, 0.6545416666666667, 0.5486718749999999, 1.0], "iscrowd": 0, "area": 45690.51789999999, "rle": {"size": [480, 640], "counts": "i:o0Q>m1SN[1eN00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1001O000000001O0000001O2N3M3M3M3M3M3M3M3M3M1O0000O1000000000000000000000000000000000000O100000000000000000000000000000000000000O1000000000000000000000000000000000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000001O00001O00001O00001O00001O00001O00001O001O001O1O1O0O2O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001N2O1O001O1O1O001O1O1M2O2M3N2M2O2N2M3N3K4K6I6J7J5J6J7J5J7I6K6I^RW4"}, "label": "couch and pillow in the left bottom corner"}]} {"id": 532744, "image": "COCO_train2014_000000532744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red nissan pickup truck\"."}], "task": "refering", "answer_template": "The \"the red nissan pickup truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [266, 267, 268, 287, 288, 289, 290, 291, 292, 309, 310, 311, 312, 313, 314, 315, 316, 332, 333, 334, 335, 336, 337, 338, 339, 355, 356, 357, 358, 359, 360, 361, 362, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.438125, 0.6972291666666667, 0.769734375, 0.9990833333333333], "iscrowd": 0, "area": 24284.19765, "rle": {"size": [480, 640], "counts": "oST45j>4L3L5@CVBa0c=C\\B`0`=B^Bb0`=_O^Bd0`=\\O`Bf0^=[O`Bh0^=XObBj0\\=?N2N2N2O1O100L4K5L4L3L4M1N3N1O2M201O0100O02O000O2O000O2O000O2O000O102N2M3L4L4K6K4L3L3N2N200O01000O100O10000O10000O10000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000O1000000000000O100O100O101N10000O100000000000000O101O00000000000O100001O1O0010O01O1O001O001O000000000001O0000000O100000000000001O2N2N1N3N2N1O2FnDbLT;[3:O2M3M3M2O2N2N1O2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N:F>B=C>B>Ba^T2"}, "label": "the red nissan pickup truck"}]} {"id": 532744, "image": "COCO_train2014_000000532744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small orange nissan pick up truck in the traffic\"."}], "task": "refering", "answer_template": "The \"a small orange nissan pick up truck in the traffic\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [266, 267, 268, 287, 288, 289, 290, 291, 292, 309, 310, 311, 312, 313, 314, 315, 316, 332, 333, 334, 335, 336, 337, 338, 339, 355, 356, 357, 358, 359, 360, 361, 362, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.438125, 0.6972291666666667, 0.769734375, 0.9990833333333333], "iscrowd": 0, "area": 24284.19765, "rle": {"size": [480, 640], "counts": "oST45j>4L3L5@CVBa0c=C\\B`0`=B^Bb0`=_O^Bd0`=\\O`Bf0^=[O`Bh0^=XObBj0\\=?N2N2N2O1O100L4K5L4L3L4M1N3N1O2M201O0100O02O000O2O000O2O000O2O000O102N2M3L4L4K6K4L3L3N2N200O01000O100O10000O10000O10000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000O1000000000000O100O100O101N10000O100000000000000O101O00000000000O100001O1O0010O01O1O001O001O000000000001O0000000O100000000000001O2N2N1N3N2N1O2FnDbLT;[3:O2M3M3M2O2N2N1O2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N:F>B=C>B>Ba^T2"}, "label": "a small orange nissan pick up truck in the traffic"}]} {"id": 41233, "image": "COCO_train2014_000000041233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark gray truck\"."}], "task": "refering", "answer_template": "The \"a dark gray truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 130, 150, 151, 152, 153, 154, 174, 175, 176, 177, 197, 198, 199, 200, 220, 221, 222, 223, 244, 245, 246, 267, 268, 269], "bbox": [0.5460625, 0.3016875, 0.7382031250000001, 0.6747291666666667], "iscrowd": 0, "area": 14891.4906, "rle": {"size": [480, 640], "counts": "aVT52n>2N3M2N2N2N2N3M3M8H8H8H7I8H8H8H7I8H8H8H7I5L0O2N1O1O2N1O2O0O1O2N1O2N100O2N1O1O2N101N1O1O3M3M4M2M4L3M3M4L3M01O00000O10001O00000000001O000000000O101O00000000001O1O2N2N2N1O2N2M2O2N2N2N1O2N2N2N2N3M3M3M3L4M3L4M3L3N3M3L4M3L4M3L4M2M4M3M3L4M3L4M3L4F:WOc`^2"}, "label": "a dark gray truck"}]} {"id": 41233, "image": "COCO_train2014_000000041233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"front end of a grey truck without the trailor\"."}], "task": "refering", "answer_template": "The \"front end of a grey truck without the trailor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 130, 150, 151, 152, 153, 154, 174, 175, 176, 177, 197, 198, 199, 200, 220, 221, 222, 223, 244, 245, 246, 267, 268, 269], "bbox": [0.5460625, 0.3016875, 0.7382031250000001, 0.6747291666666667], "iscrowd": 0, "area": 14891.4906, "rle": {"size": [480, 640], "counts": "aVT52n>2N3M2N2N2N2N3M3M8H8H8H7I8H8H8H7I8H8H8H7I5L0O2N1O1O2N1O2O0O1O2N1O2N100O2N1O1O2N101N1O1O3M3M4M2M4L3M3M4L3M01O00000O10001O00000000001O000000000O101O00000000001O1O2N2N2N1O2N2M2O2N2N2N1O2N2N2N2N3M3M3M3L4M3L4M3L3N3M3L4M3L4M3L4M2M4M3M3L4M3L4M3L4F:WOc`^2"}, "label": "front end of a grey truck without the trailor"}]} {"id": 41233, "image": "COCO_train2014_000000041233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange dump truck with its back raised up\"."}], "task": "refering", "answer_template": "The \"an orange dump truck with its back raised up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 136, 137, 158, 159, 160, 180, 181, 182, 183, 202, 203, 204, 205, 206, 228, 229, 251, 252], "bbox": [0.819546875, 0.26160416666666664, 1.0, 0.6046875], "iscrowd": 0, "area": 10255.148400000002, "rle": {"size": [480, 640], "counts": "ZZf73l>4L4K5L4L00O1O1O00100O1O1O03N1O2N2N1O2M3N1O2N2N1N3N9G00O10O10000000000O2O001O001N101O1O001N101O001O1Oa0^O5L1O2N2M3M3M3M3N2M3M3M3M3M3M3N2M4L3M4L3M4L3N3L3M4L3M4N2N1O2N1O1O000001O000000000O100000001O000000000hMSF]Om9`0YF]Og9a0]F]Oc9`0cF]O]9a0gF\\O[9`0jF^OV9`0nF^OR9?SG_Om8?WG_Oi8>\\G@d8=aGA_84nGUOg8IH"}, "label": "an orange dump truck with its back raised up"}]} {"id": 41233, "image": "COCO_train2014_000000041233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange dump truck\"."}], "task": "refering", "answer_template": "The \"orange dump truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 136, 137, 158, 159, 160, 180, 181, 182, 183, 202, 203, 204, 205, 206, 228, 229, 251, 252], "bbox": [0.819546875, 0.26160416666666664, 1.0, 0.6046875], "iscrowd": 0, "area": 10255.148400000002, "rle": {"size": [480, 640], "counts": "ZZf73l>4L4K5L4L00O1O1O00100O1O1O03N1O2N2N1O2M3N1O2N2N1N3N9G00O10O10000000000O2O001O001N101O1O001N101O001O1Oa0^O5L1O2N2M3M3M3M3N2M3M3M3M3M3M3N2M4L3M4L3M4L3N3L3M4L3M4N2N1O2N1O1O000001O000000000O100000001O000000000hMSF]Om9`0YF]Og9a0]F]Oc9`0cF]O]9a0gF\\O[9`0jF^OV9`0nF^OR9?SG_Om8?WG_Oi8>\\G@d8=aGA_84nGUOg8IH"}, "label": "orange dump truck"}]} {"id": 41233, "image": "COCO_train2014_000000041233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white tractor big rig with a blue rope around the bumper\"."}], "task": "refering", "answer_template": "The \"a white tractor big rig with a blue rope around the bumper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 102, 103, 104, 120, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 306, 307, 308, 309], "bbox": [0.067265625, 0.21077083333333332, 0.59865625, 0.7937291666666667], "iscrowd": 0, "area": 61606.10525, "rle": {"size": [480, 640], "counts": "i^d07[>g0hAXOd=U1L4L4M3N1N2O00000000000O101O000000000000001O00000000000000001O0000O1O1O1O1O1O1O1O1O1001O000000001O00000000001O000000001N100O100O2N1000000O101O00000O101O00000O10001O000O10001O000O10001O0O1000000I8CB=C=C=J60010O0000000010O00000001O0001O000001O0001O0001O00000001O01O000000010O1O001O1O100O1O1O1O1O1O100O1O001O1O100O1O00000000001O0000000000O1000000O1000000O10000O1000000M3L4M3O1O1001O00001O001O00001O00001O00001O001O000010O0010O00010O0010O00010O00010N10001O0O101O0O2O000O2O000O2O000O2O000O2O000O2O00001N101O0O101O0O101O0O101O0O101O0O101O0O101O000O2O001N10001N10001N10001N10001N10001N10001O0O2O001N5L4L3L5L4L4K5L3M4K5L4L3L5L4L4K5L3M4L5J7J6J6I7J6J6H8H8I7H9H7H8H8ZOZVh3"}, "label": "a white tractor big rig with a blue rope around the bumper"}]} {"id": 41233, "image": "COCO_train2014_000000041233.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white semi truck is parked by a black one , with a blue strap attached to the front\"."}], "task": "refering", "answer_template": "The \"a white semi truck is parked by a black one , with a blue strap attached to the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 102, 103, 104, 120, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 306, 307, 308, 309], "bbox": [0.067265625, 0.21077083333333332, 0.59865625, 0.7937291666666667], "iscrowd": 0, "area": 61606.10525, "rle": {"size": [480, 640], "counts": "i^d07[>g0hAXOd=U1L4L4M3N1N2O00000000000O101O000000000000001O00000000000000001O0000O1O1O1O1O1O1O1O1O1001O000000001O00000000001O000000001N100O100O2N1000000O101O00000O101O00000O10001O000O10001O000O10001O0O1000000I8CB=C=C=J60010O0000000010O00000001O0001O000001O0001O0001O00000001O01O000000010O1O001O1O100O1O1O1O1O1O100O1O001O1O100O1O00000000001O0000000000O1000000O1000000O10000O1000000M3L4M3O1O1001O00001O001O00001O00001O00001O001O000010O0010O00010O0010O00010O00010N10001O0O101O0O2O000O2O000O2O000O2O000O2O000O2O00001N101O0O101O0O101O0O101O0O101O0O101O0O101O000O2O001N10001N10001N10001N10001N10001N10001O0O2O001N5L4L3L5L4L4K5L3M4K5L4L3L5L4L4K5L3M4L5J7J6J6I7J6J6H8H8I7H9H7H8H8ZOZVh3"}, "label": "a white semi truck is parked by a black one , with a blue strap attached to the front"}]} {"id": 442641, "image": "COCO_train2014_000000442641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing the back side position holding the hand in kite\"."}], "task": "refering", "answer_template": "The \"a man standing the back side position holding the hand in kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 173, 174, 175, 195, 196, 197, 198, 219, 220, 221, 242, 243, 244, 265, 266, 267, 288, 289, 290, 312, 334, 335], "bbox": [0.507484375, 0.3657916666666667, 0.6223750000000001, 0.8392083333333334], "iscrowd": 0, "area": 9662.194200000004, "rle": {"size": [480, 640], "counts": "Tbh44k>3M1O1O1O1O1O2N1O1O1O2N1O01O00VCAT;?jDEoN8V;3hEn0Q:SOiEU1U:kNcE_1Z:cN\\En1\\:^1M1TERLd:^4G0lF_Kh7b4VHaKi7_4WHbKh7^4XHbKh7^4XHcKf7^4ZHbKf7^4ZHaKg7_4YHaKf7h4RHXKn7j4PHVKo7n4mGRKT8Q5hGPKX8R5eGoJd72VHS52lJg75RHT54gJi7R6SHoIl7U6PHlIo7c6O1O1O100000000000000000O11OmIRHU5m7jJVHT5k7iJXHV5h7hJZHX5g7fJ[HY5h7aJ\\H^5g7[J]HQ5HnJR8IYHQ51PK^8i4Q1H8RMkEl0Y:lNPFfNDR2^:POaFm0a9kNgFWOfNb1g:VOeFVOiN\\1f:]OPG4[9KR3KmVa3"}, "label": "a man standing the back side position holding the hand in kite"}]} {"id": 442641, "image": "COCO_train2014_000000442641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy in a brown shirt\"."}], "task": "refering", "answer_template": "The \"a guy in a brown shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 173, 174, 175, 195, 196, 197, 198, 219, 220, 221, 242, 243, 244, 265, 266, 267, 288, 289, 290, 312, 334, 335], "bbox": [0.507484375, 0.3657916666666667, 0.6223750000000001, 0.8392083333333334], "iscrowd": 0, "area": 9662.194200000004, "rle": {"size": [480, 640], "counts": "Tbh44k>3M1O1O1O1O1O2N1O1O1O2N1O01O00VCAT;?jDEoN8V;3hEn0Q:SOiEU1U:kNcE_1Z:cN\\En1\\:^1M1TERLd:^4G0lF_Kh7b4VHaKi7_4WHbKh7^4XHbKh7^4XHcKf7^4ZHbKf7^4ZHaKg7_4YHaKf7h4RHXKn7j4PHVKo7n4mGRKT8Q5hGPKX8R5eGoJd72VHS52lJg75RHT54gJi7R6SHoIl7U6PHlIo7c6O1O1O100000000000000000O11OmIRHU5m7jJVHT5k7iJXHV5h7hJZHX5g7fJ[HY5h7aJ\\H^5g7[J]HQ5HnJR8IYHQ51PK^8i4Q1H8RMkEl0Y:lNPFfNDR2^:POaFm0a9kNgFWOfNb1g:VOeFVOiN\\1f:]OPG4[9KR3KmVa3"}, "label": "a guy in a brown shirt"}]} {"id": 237853, "image": "COCO_train2014_000000237853.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white computer keyboard placed in front of a computer desk\"."}], "task": "refering", "answer_template": "The \"a white computer keyboard placed in front of a computer desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [265, 266, 267, 268, 269, 270, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.23484375000000002, 0.7730281690140846, 0.8181875, 0.9865258215962441], "iscrowd": 0, "area": 24773.012949999997, "rle": {"size": [426, 640], "counts": "VWo14Q=5L4L4L4L4L4L4L4L4L4N2O1O1O1O1O1O1O1O1O1O1000000O100000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O1000000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O100000000000000000000O10000000000000000000000000000O100000000000000000O10000000O10O100000000O10000000000O1000O100000O100000000O10000000000O010000000O10000000000O100000O10O10000000000O10000000000O10O100000O10000000000O100000000O01000000000O100000000O100000O1001N2O1O1O2N1O1N2O1O1O2N1N2O1O1O1O1O2M2O1O1O1O1N3N1O1O1O1O1N3N1O1O1O1N2O2N1O1O1O1N2O2N1O1O1O1N2O2N1O1O1N2N2NSl_1"}, "label": "a white computer keyboard placed in front of a computer desk"}]} {"id": 237853, "image": "COCO_train2014_000000237853.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white computer keyboard in front of a stuffed monkey\"."}], "task": "refering", "answer_template": "The \"a white computer keyboard in front of a stuffed monkey\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [265, 266, 267, 268, 269, 270, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.23484375000000002, 0.7730281690140846, 0.8181875, 0.9865258215962441], "iscrowd": 0, "area": 24773.012949999997, "rle": {"size": [426, 640], "counts": "VWo14Q=5L4L4L4L4L4L4L4L4L4N2O1O1O1O1O1O1O1O1O1O1000000O100000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O1000000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O100000000000000000000O10000000000000000000000000000O100000000000000000O10000000O10O100000000O10000000000O1000O100000O100000000O10000000000O010000000O10000000000O100000O10O10000000000O10000000000O10O100000O10000000000O100000000O01000000000O100000000O100000O1001N2O1O1O2N1O1N2O1O1O2N1N2O1O1O1O1O2M2O1O1O1O1N3N1O1O1O1O1N3N1O1O1O1N2O2N1O1O1O1N2O2N1O1O1O1N2O2N1O1O1N2N2NSl_1"}, "label": "a white computer keyboard in front of a stuffed monkey"}]} {"id": 229678, "image": "COCO_train2014_000000229678.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown wooden chair in the back\"."}], "task": "refering", "answer_template": "The \"the brown wooden chair in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 26, 27, 28, 39, 40, 41, 52, 53, 54, 65, 66, 67, 68, 78, 79, 80, 81, 82, 91, 92, 93, 94, 95, 106, 107, 119, 120, 132, 133], "bbox": [0.005733333333333333, 0.10538, 0.4042933333333334, 0.6215], "iscrowd": 0, "area": 13302.658350000003, "rle": {"size": [500, 375], "counts": "ZQ1Q1b>6K6J5K5K5J6K5K6J5K5K5J6K5mCkL];Z3[DhLf;c30O10O10O1000mDjLh9U3YFkLg9U3XFlLg9U3XFlLh9S3YFmLg9S3XFmLi9R3XFnLh9R3WFoLh9R3XFnLh9Q3XFPMh9P3XFoLi9Q3VFPMj9o2WFQMi9o2VFRMi9o2WFPMj9o2VFRMj9n2VFRMj9m2VFTMj9l2VFSMk9m2TFTMk9l2UFUMk9k2UFUMk9k2TFUMm9j2TFVMl9j2SFWMl9j2TFVMl9i2TFWMm9i2SFWMm9i2RFXMn9h2RFXMn9h2QFXMo9h2RFXMn9h2QFUMS:k2mEnLZ:R3eEiLa:V3`EdLf:\\3h01000O1000O10O10000007H;F;E;E:F in this image.", "objects": [{"patches": [15, 26, 27, 28, 39, 40, 41, 52, 53, 54, 65, 66, 67, 68, 78, 79, 80, 81, 82, 91, 92, 93, 94, 95, 106, 107, 119, 120, 132, 133], "bbox": [0.005733333333333333, 0.10538, 0.4042933333333334, 0.6215], "iscrowd": 0, "area": 13302.658350000003, "rle": {"size": [500, 375], "counts": "ZQ1Q1b>6K6J5K5K5J6K5K6J5K5K5J6K5mCkL];Z3[DhLf;c30O10O10O1000mDjLh9U3YFkLg9U3XFlLg9U3XFlLh9S3YFmLg9S3XFmLi9R3XFnLh9R3WFoLh9R3XFnLh9Q3XFPMh9P3XFoLi9Q3VFPMj9o2WFQMi9o2VFRMi9o2WFPMj9o2VFRMj9n2VFRMj9m2VFTMj9l2VFSMk9m2TFTMk9l2UFUMk9k2UFUMk9k2TFUMm9j2TFVMl9j2SFWMl9j2TFVMl9i2TFWMm9i2SFWMm9i2RFXMn9h2RFXMn9h2QFXMo9h2RFXMn9h2QFUMS:k2mEnLZ:R3eEiLa:V3`EdLf:\\3h01000O1000O10O10000007H;F;E;E:F in this image.", "objects": [{"patches": [78, 79, 80, 91, 92, 93, 104, 105, 106, 117, 119, 132, 144, 145, 146, 156, 157, 158, 159, 160, 161, 169, 170, 171, 172, 173, 174, 175, 182, 183, 184, 185, 186, 187, 188, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 211, 212, 213, 214, 215, 221, 222, 223, 224, 225, 227, 228], "bbox": [0.0, 0.33196, 0.5505333333333333, 1.0], "iscrowd": 0, "area": 33462.09725, "rle": {"size": [500, 375], "counts": "T7a0W3Q3g3kLQJd0h1b2V4\\M`I2Y2d2U4[MaI1[2e2S4YMcI2Y2h2R4VMeI1Z2j2P4TMfI3Y2k2o3RMiI2Y2m2m3QMjI2X2o2m3nLkI3Y2P3j3nLmI2X2R3j3kLoI2X2T3h3jLoI3X2U3h3gLQJ3X2X3d3eLUJ3V2Z3d3cLVJ2W2\\3b3aLWJ3W2^3a3_LXJ3W2_3_3^L[J2V2b3^3\\L[J3W2b3]3ZL]J3V2e3[3YL_J2V2f3Z3WL`J3V2i3X3TLbJ3V2j3W3RLdJ3U2m3U3QLeJ3V2m3T3oKgJ3U2P4S3mKhJ3U2Q4Q3lKkJ2T2T4P3jKkJ3U2T4o2hKmJ3T2X4m2eKoJ3T2Y4k2dKQK3T2[4j2bKRK3T2\\4i2`KTK3S2_4g2_KUK3T2^4g2_KUK2T2`4g2]KVK2T2a4f2]KUK3T2a4f2\\KWK2T2b4e2\\KWK2S2c4f2ZKWK3T2c4d2[KXK2S2d4e2YKYK2S2e4d2YKYK2R2f4e2WKYK3S2e4d2YKYK2R2f4e2WKZK2R2g4d2WKYK3R2g4e2UKZK3R2h4c2VK[K2Q2i4d2TK[K3R2i4c2TK[K3Q2j4c2SK]K2Q2k4b2SK\\K3Q2k4c2RK\\K2Q2m4c2PK]K3P2m4b2QK]K2Q2n4b2YKTKIZ2n4b2[L^Mf3a2ZL`Mf3`2ZL_Mg3a2XL`Mh3`2XL_Mi3`2WLaMi3_2fNQK[1W5]48H9F:G8H9G8H9G9G8H9G9G8H9G8H9G9G3M001O001O001O0O2O001O00001O0O2O0O2N1O2N1O2N1O2O0O2N1O2N1O2N1O2O0O2N1O2N1O2N1O101N1O2N1O2N1O2N101N1O2N1O2N1O2N101N1O2N1O2N1O2N101N1O2N1O1O2N1O2O0O2N1O2N1O2N1O2O0O2N1O2N1O2N1O2N101N1O2N1O2N1O2N100O2N1O27H=RDbMm9j2eEXMZ:T3YEoLf:]3kDfLT;m31O001N101O001O001O001O0O in this image.", "objects": [{"patches": [78, 79, 80, 91, 92, 93, 104, 105, 106, 117, 119, 132, 144, 145, 146, 156, 157, 158, 159, 160, 161, 169, 170, 171, 172, 173, 174, 175, 182, 183, 184, 185, 186, 187, 188, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 211, 212, 213, 214, 215, 221, 222, 223, 224, 225, 227, 228], "bbox": [0.0, 0.33196, 0.5505333333333333, 1.0], "iscrowd": 0, "area": 33462.09725, "rle": {"size": [500, 375], "counts": "T7a0W3Q3g3kLQJd0h1b2V4\\M`I2Y2d2U4[MaI1[2e2S4YMcI2Y2h2R4VMeI1Z2j2P4TMfI3Y2k2o3RMiI2Y2m2m3QMjI2X2o2m3nLkI3Y2P3j3nLmI2X2R3j3kLoI2X2T3h3jLoI3X2U3h3gLQJ3X2X3d3eLUJ3V2Z3d3cLVJ2W2\\3b3aLWJ3W2^3a3_LXJ3W2_3_3^L[J2V2b3^3\\L[J3W2b3]3ZL]J3V2e3[3YL_J2V2f3Z3WL`J3V2i3X3TLbJ3V2j3W3RLdJ3U2m3U3QLeJ3V2m3T3oKgJ3U2P4S3mKhJ3U2Q4Q3lKkJ2T2T4P3jKkJ3U2T4o2hKmJ3T2X4m2eKoJ3T2Y4k2dKQK3T2[4j2bKRK3T2\\4i2`KTK3S2_4g2_KUK3T2^4g2_KUK2T2`4g2]KVK2T2a4f2]KUK3T2a4f2\\KWK2T2b4e2\\KWK2S2c4f2ZKWK3T2c4d2[KXK2S2d4e2YKYK2S2e4d2YKYK2R2f4e2WKYK3S2e4d2YKYK2R2f4e2WKZK2R2g4d2WKYK3R2g4e2UKZK3R2h4c2VK[K2Q2i4d2TK[K3R2i4c2TK[K3Q2j4c2SK]K2Q2k4b2SK\\K3Q2k4c2RK\\K2Q2m4c2PK]K3P2m4b2QK]K2Q2n4b2YKTKIZ2n4b2[L^Mf3a2ZL`Mf3`2ZL_Mg3a2XL`Mh3`2XL_Mi3`2WLaMi3_2fNQK[1W5]48H9F:G8H9G8H9G9G8H9G9G8H9G8H9G9G3M001O001O001O0O2O001O00001O0O2O0O2N1O2N1O2N1O2O0O2N1O2N1O2N1O2O0O2N1O2N1O2N1O101N1O2N1O2N1O2N101N1O2N1O2N1O2N101N1O2N1O2N1O2N101N1O2N1O1O2N1O2O0O2N1O2N1O2N1O2O0O2N1O2N1O2N1O2N101N1O2N1O2N1O2N100O2N1O27H=RDbMm9j2eEXMZ:T3YEoLf:]3kDfLT;m31O001N101O001O001O001O0O in this image.", "objects": [{"patches": [9, 31, 32, 33, 54, 55, 56, 76, 77, 78, 79, 81, 82, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 278, 279, 280, 281, 282, 283, 287, 302, 303, 304, 305], "bbox": [0.084546875, 0.05812646370023419, 0.85765625, 0.9023185011709602], "iscrowd": 0, "area": 87094.08145, "rle": {"size": [427, 640], "counts": "ajf02S<0mD7n:LlD in this image.", "objects": [{"patches": [9, 31, 32, 33, 54, 55, 56, 76, 77, 78, 79, 81, 82, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 278, 279, 280, 281, 282, 283, 287, 302, 303, 304, 305], "bbox": [0.084546875, 0.05812646370023419, 0.85765625, 0.9023185011709602], "iscrowd": 0, "area": 87094.08145, "rle": {"size": [427, 640], "counts": "ajf02S<0mD7n:LlD in this image.", "objects": [{"patches": [171, 172, 173, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 234, 235, 236, 237, 255, 256, 257, 258, 276, 277, 278, 279, 297, 298, 299, 300, 318, 319, 320, 321, 340, 341, 342, 362], "bbox": [0.12183333333333332, 0.3922166666666667, 0.3242, 0.82225], "iscrowd": 0, "area": 21044.769899999996, "rle": {"size": [600, 600], "counts": "eQ[12cb06K5J7I6K6I6K6K4L4M2M2N3N1N3M2O2M2N3N1N3M2O2M2N2O2M2N in this image.", "objects": [{"patches": [171, 172, 173, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 234, 235, 236, 237, 255, 256, 257, 258, 276, 277, 278, 279, 297, 298, 299, 300, 318, 319, 320, 321, 340, 341, 342, 362], "bbox": [0.12183333333333332, 0.3922166666666667, 0.3242, 0.82225], "iscrowd": 0, "area": 21044.769899999996, "rle": {"size": [600, 600], "counts": "eQ[12cb06K5J7I6K6I6K6K4L4M2M2N3N1N3M2O2M2N3N1N3M2O2M2N2O2M2N in this image.", "objects": [{"patches": [153, 154, 155, 156, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 216, 217, 218, 219, 237, 238, 239, 240, 241, 258, 259, 260, 261, 279, 280, 281, 282, 300, 301, 302], "bbox": [0.2553, 0.35541666666666666, 0.4795833333333333, 0.7104499999999999], "iscrowd": 0, "area": 17341.1693, "rle": {"size": [600, 600], "counts": "Uli2;]b01O1O1N2O2N1N2O1O1O1N3N1O1O1N2O1O2N1N2K5L4O1N2O2M2O4K6K5K3L4M4K4M3L5\\BXMP;l2fD^MU;g2bDbMZ;a2^DhM^;^2VDmMe;Z2lCRNo;V2bCUN[O1O1N2O1N2N3N1N2N2O4KR1nNb_f5"}, "label": "the torso of a brown furry teddy bear wearing a red ribbon"}]} {"id": 309, "image": "COCO_train2014_000000000309.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"stuffed teddy bear with a red ribbon\"."}], "task": "refering", "answer_template": "The \"stuffed teddy bear with a red ribbon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 155, 156, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 216, 217, 218, 219, 237, 238, 239, 240, 241, 258, 259, 260, 261, 279, 280, 281, 282, 300, 301, 302], "bbox": [0.2553, 0.35541666666666666, 0.4795833333333333, 0.7104499999999999], "iscrowd": 0, "area": 17341.1693, "rle": {"size": [600, 600], "counts": "Uli2;]b01O1O1N2O2N1N2O1O1O1N3N1O1O1N2O1O2N1N2K5L4O1N2O2M2O4K6K5K3L4M4K4M3L5\\BXMP;l2fD^MU;g2bDbMZ;a2^DhM^;^2VDmMe;Z2lCRNo;V2bCUN[O1O1N2O1N2N3N1N2N2O4KR1nNb_f5"}, "label": "stuffed teddy bear with a red ribbon"}]} {"id": 459082, "image": "COCO_train2014_000000459082.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red double decker bus\"."}], "task": "refering", "answer_template": "The \"the red double decker bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 37, 38, 39, 55, 56, 57, 58, 59, 60, 61, 62, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 308, 309, 310, 311, 312, 313, 314, 315], "bbox": [0.299703125, 0.08295081967213115, 0.7695937500000001, 0.9159250585480094], "iscrowd": 0, "area": 90248.56885000003, "rle": {"size": [427, 640], "counts": "h[`2?g9OcHg0d6MfHn0b6FeHV1d6]OdH`1h6i2C;F:F;M2M4M2O101N1O2N10O01O1O010O1O00100O001O010O1O00100O001O10O01O001O100O2N1O100O1O100O2N1O100O1O100O1O2O0O1O1O100O3M3N2M3M3N2M2N3M3N2M3M3N2M3M3NO0O1O1O2O0O1O2N1O1010O1O1O100O2N100O1O100O1O1O2O0O1O100O1O101N1O1000000O10001O000O100000000O2O00000O10O100000O100000000O1000000000000000000000000O100000000000000000000000000O10000000000000000000000000O0100001O0000000000000000001O0000000O100000001O0000000000000000001O00000O1000000000001O0000000000001O0000001N1000001O0000001O0000001O0000001O00000O2O0000001O00002N1O2N2N1O2N1O2N1O2N1N3N1O2N2N2N8H8H8H8H8H9G8H8H8H8H8H8H9F=D`0@`0@`0@`0@`0@`0@a0_O`0@`0@`0@`0@Pnl1"}, "label": "the red double decker bus"}]} {"id": 459082, "image": "COCO_train2014_000000459082.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red bus\"."}], "task": "refering", "answer_template": "The \"a red bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 37, 38, 39, 55, 56, 57, 58, 59, 60, 61, 62, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 308, 309, 310, 311, 312, 313, 314, 315], "bbox": [0.299703125, 0.08295081967213115, 0.7695937500000001, 0.9159250585480094], "iscrowd": 0, "area": 90248.56885000003, "rle": {"size": [427, 640], "counts": "h[`2?g9OcHg0d6MfHn0b6FeHV1d6]OdH`1h6i2C;F:F;M2M4M2O101N1O2N10O01O1O010O1O00100O001O010O1O00100O001O10O01O001O100O2N1O100O1O100O2N1O100O1O100O1O2O0O1O1O100O3M3N2M3M3N2M2N3M3N2M3M3N2M3M3NO0O1O1O2O0O1O2N1O1010O1O1O100O2N100O1O100O1O1O2O0O1O100O1O101N1O1000000O10001O000O100000000O2O00000O10O100000O100000000O1000000000000000000000000O100000000000000000000000000O10000000000000000000000000O0100001O0000000000000000001O0000000O100000001O0000000000000000001O00000O1000000000001O0000000000001O0000001N1000001O0000001O0000001O0000001O00000O2O0000001O00002N1O2N2N1O2N1O2N1O2N1N3N1O2N2N2N8H8H8H8H8H9G8H8H8H8H8H8H9F=D`0@`0@`0@`0@`0@`0@a0_O`0@`0@`0@`0@Pnl1"}, "label": "a red bus"}]} {"id": 459082, "image": "COCO_train2014_000000459082.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey minivan driving in the opposite direction of a red bus\"."}], "task": "refering", "answer_template": "The \"a grey minivan driving in the opposite direction of a red bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 295, 296, 297, 298], "bbox": [0.761859375, 0.5461826697892271, 1.0, 0.8367213114754097], "iscrowd": 0, "area": 16087.65965, "rle": {"size": [427, 640], "counts": "]h[6l0^ in this image.", "objects": [{"patches": [202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 295, 296, 297, 298], "bbox": [0.761859375, 0.5461826697892271, 1.0, 0.8367213114754097], "iscrowd": 0, "area": 16087.65965, "rle": {"size": [427, 640], "counts": "]h[6l0^ in this image.", "objects": [{"patches": [63, 64, 65, 86, 87, 88, 108, 109, 110, 131, 132, 133, 134, 153, 154, 155, 156, 157, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 246, 247, 248, 249, 269, 270, 271, 272, 292, 293, 294, 295, 314, 315, 316, 317, 318, 337, 338, 339, 340, 341], "bbox": [0.6880625, 0.15504694835680752, 0.9049531250000001, 0.9887558685446011], "iscrowd": 0, "area": 27525.510949999993, "rle": {"size": [426, 640], "counts": "d^g51W60?:]ON4=HLI=4L]O>?KRO?j0IgNb0V1F[Nd0a1EPNe0l1CeMg0X2@[Mj0a2]ORMm0j2ZOiLP1W3SO[LX1f3jNmK_1T4dN_Kf1b4\\NPKo1Q5T31O1N2O1O1O001O1N2O1O1O1O1N10mN\\HbLc7]3fH\\LY7d3nHVLQ7i3TITLk6l3VITLh6m3ZIRLe6m3]ISLb6m3_IRLa6m3bIRL]6n3dIRLZ6n3hIRLW6n3kIQLT6o3mIoKT6P4oInKQ6R4PJmKo5T4RJjKo5V4SJhKm5W4UJgKl5Y4b106Jd0[Of0[O:F2N1N20000001O0001O000001O00000\\GdM[6[2dIgM[6Y2cIiM]6X2aIjM^6V2aIlM^6T2`IoM_6Q2`IQN_6o1`ISN_6m1_IUNb6k1\\IWNc6i1[IZNd6f1[I\\Nd6d1[I^Nd6b1ZI`Nf6`1YIbNf6_1XIcNg6]1WIfNi6Y1VIiNi6W1VIkNi6U1UImNk6T1RIoNm6Q1RIQOl6P1RISO_6[1_IhNR6f1mI[Na0dNV3_3WLPN@Dj3k2cLAP3l0oLVOb2^1VMdN\\2V2VMlM\\2i6A>C>A>B?K410O010O010O0010O001N1O2N1O2N1O2N1O1O2N1O20O2M2N3L3N3L3N3L3N3M2M4M2MdTi0"}, "label": "a woman wearing a grey hat"}]} {"id": 205131, "image": "COCO_train2014_000000205131.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman holding yellow streamers\"."}], "task": "refering", "answer_template": "The \"woman holding yellow streamers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 86, 87, 88, 108, 109, 110, 131, 132, 133, 134, 153, 154, 155, 156, 157, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 246, 247, 248, 249, 269, 270, 271, 272, 292, 293, 294, 295, 314, 315, 316, 317, 318, 337, 338, 339, 340, 341], "bbox": [0.6880625, 0.15504694835680752, 0.9049531250000001, 0.9887558685446011], "iscrowd": 0, "area": 27525.510949999993, "rle": {"size": [426, 640], "counts": "d^g51W60?:]ON4=HLI=4L]O>?KRO?j0IgNb0V1F[Nd0a1EPNe0l1CeMg0X2@[Mj0a2]ORMm0j2ZOiLP1W3SO[LX1f3jNmK_1T4dN_Kf1b4\\NPKo1Q5T31O1N2O1O1O001O1N2O1O1O1O1N10mN\\HbLc7]3fH\\LY7d3nHVLQ7i3TITLk6l3VITLh6m3ZIRLe6m3]ISLb6m3_IRLa6m3bIRL]6n3dIRLZ6n3hIRLW6n3kIQLT6o3mIoKT6P4oInKQ6R4PJmKo5T4RJjKo5V4SJhKm5W4UJgKl5Y4b106Jd0[Of0[O:F2N1N20000001O0001O000001O00000\\GdM[6[2dIgM[6Y2cIiM]6X2aIjM^6V2aIlM^6T2`IoM_6Q2`IQN_6o1`ISN_6m1_IUNb6k1\\IWNc6i1[IZNd6f1[I\\Nd6d1[I^Nd6b1ZI`Nf6`1YIbNf6_1XIcNg6]1WIfNi6Y1VIiNi6W1VIkNi6U1UImNk6T1RIoNm6Q1RIQOl6P1RISO_6[1_IhNR6f1mI[Na0dNV3_3WLPN@Dj3k2cLAP3l0oLVOb2^1VMdN\\2V2VMlM\\2i6A>C>A>B?K410O010O010O0010O001N1O2N1O2N1O2N1O1O2N1O20O2M2N3L3N3L3N3L3N3M2M4M2MdTi0"}, "label": "woman holding yellow streamers"}]} {"id": 205131, "image": "COCO_train2014_000000205131.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a blue shirt flying a kite\"."}], "task": "refering", "answer_template": "The \"a man wearing a blue shirt flying a kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 187, 188, 189, 209, 210, 211, 212, 232, 233, 234, 235, 256, 257, 258, 279, 280, 281, 302, 303, 304, 325, 326, 327], "bbox": [0.12342187499999999, 0.48225352112676056, 0.255984375, 1.0], "iscrowd": 0, "area": 11039.49665000001, "rle": {"size": [426, 640], "counts": "_TQ11W=3RC1\\<2bC1Z<2dC2W<2fC1V in this image.", "objects": [{"patches": [164, 165, 187, 188, 189, 209, 210, 211, 212, 232, 233, 234, 235, 256, 257, 258, 279, 280, 281, 302, 303, 304, 325, 326, 327], "bbox": [0.12342187499999999, 0.48225352112676056, 0.255984375, 1.0], "iscrowd": 0, "area": 11039.49665000001, "rle": {"size": [426, 640], "counts": "_TQ11W=3RC1\\<2bC1Z<2dC2W<2fC1V in this image.", "objects": [{"patches": [120, 121, 122, 142, 143, 144, 145, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 233, 234, 235, 256, 257, 258, 279, 280, 281, 303, 304, 326, 327, 349, 350, 351, 372, 373, 374, 375], "bbox": [0.135171875, 0.32308333333333333, 0.39065625, 1.0], "iscrowd": 0, "area": 22650.810449999997, "rle": {"size": [480, 640], "counts": "oRY1b0\\>5L2N3L3N2N2M2O1N3N1O1N101O1N1N3M3M2N3M2O2M3M2N3M2N3N2O001O1O0RFlMQ7U2ZHkNZ7U1]HCT7>cHb0e6_ORIj0X5fLYK`2\\On0S5UMUKn1EP1n4`MVK`1IT1n4fMQKW1MW1P5gMoJS1NY1P5jMnJm0O]1P5lMmJh0O`1R5lMlJd00c1R5nMjJR5U5QKhJP5W5]2N2O1O10000O100O100O10000O100O100O10000O100O10000O10oIWLk0h3VOcL?]3@oL5Q3KZMJf25gM_OY2a0RNTOn1k0^NjNb1V1iN_NW1a1UOSNk0l1AiM?W2EeM;Z2GeM9[2HdM8\\2JbM6]2LbM4^2MaM3^2OaM1_21_MO`23_MN`23_MMa24^MLa27]MIc28\\MHc2:\\MFd2;[MEe2nBBU=4QCMS=HTC8j=0O100O01000000000000O101O0O101N10000O2O1O1N2Omaf5"}, "label": "one girl with blonde hair and brown long - sleeve shirt eating a lemon"}]} {"id": 500057, "image": "COCO_train2014_000000500057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small child wearing a stripped dress eating\"."}], "task": "refering", "answer_template": "The \"a small child wearing a stripped dress eating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 142, 143, 144, 145, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 233, 234, 235, 256, 257, 258, 279, 280, 281, 303, 304, 326, 327, 349, 350, 351, 372, 373, 374, 375], "bbox": [0.135171875, 0.32308333333333333, 0.39065625, 1.0], "iscrowd": 0, "area": 22650.810449999997, "rle": {"size": [480, 640], "counts": "oRY1b0\\>5L2N3L3N2N2M2O1N3N1O1N101O1N1N3M3M2N3M2O2M3M2N3M2N3N2O001O1O0RFlMQ7U2ZHkNZ7U1]HCT7>cHb0e6_ORIj0X5fLYK`2\\On0S5UMUKn1EP1n4`MVK`1IT1n4fMQKW1MW1P5gMoJS1NY1P5jMnJm0O]1P5lMmJh0O`1R5lMlJd00c1R5nMjJR5U5QKhJP5W5]2N2O1O10000O100O100O10000O100O100O10000O100O10000O10oIWLk0h3VOcL?]3@oL5Q3KZMJf25gM_OY2a0RNTOn1k0^NjNb1V1iN_NW1a1UOSNk0l1AiM?W2EeM;Z2GeM9[2HdM8\\2JbM6]2LbM4^2MaM3^2OaM1_21_MO`23_MN`23_MMa24^MLa27]MIc28\\MHc2:\\MFd2;[MEe2nBBU=4QCMS=HTC8j=0O100O01000000000000O101O0O101N10000O2O1O1N2Omaf5"}, "label": "a small child wearing a stripped dress eating"}]} {"id": 500057, "image": "COCO_train2014_000000500057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl in black dress\"."}], "task": "refering", "answer_template": "The \"the girl in black dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 32, 33, 34, 35, 36, 55, 56, 57, 58, 59, 77, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195], "bbox": [0.37246875, 0.011395833333333332, 0.606734375, 0.5372291666666666], "iscrowd": 0, "area": 27152.414199999992, "rle": {"size": [480, 640], "counts": "Ue_36j>i1VN7J7H8I7H8I6fEjL]8^3XGeLh8`3QGcLo8_3lFdLS9^3iFeLW9]3dFeL\\9]3`FfL`9\\3[FgLd9[3XFhLh9Y3TFjLk9U410O100O00100O100O010O1_NRKPIn4m6YKnHg4P7aKkH_4R7hKiHY4T7mKiHS4S7SLjHn3S7VLkHk3R7ZLlHf3Q7_LlHb3Q7cLmH]3Q7gLlHY3U7iLiHW3W7kLfHV3Z7X200000000000O10000000000000000O10000000000000000O1000000O1000000O10000O10000O10000O10000O10000N2O1O1N2O1N2O1N2dKPHg1Q8XNPHf1R8XNPHg1Q8XNPHf1R8XNQHe1Q8YNQHd1R8[NoGc1S8[NoGc1S8[NPHb1R8]NoGa1S8]NoGa1S8^NnG`1T8^NoG_1S8_NoG_1S8`NnG^1T8`NnG^1T8aNnG\\1T8bNnG^1R8bNnG_1P8aNQH`1n7`NSH_1m7aNSH`1l7`NTHa1k7_NUHb1l7\\NUHc1n7ZNRHg1o7WNQHj1P8TNQHl1Q8QNoGo1S8oMmGR2T8lMlGU2V8hMkGX2V8fMjGZ2X8dMhG]2Z8`MfGa2[8]MfGc2[8ZMfGf2]8WMcGj2^8TMbGm2_8QMbGo2`8nL`GR3b8lL^Gm2iNmLj94]Gi2SOPMa95`Gb2XOVMW98fGX2U9gMoFP2W9oMn3O1O1O1O1OUaf3"}, "label": "the girl in black dress"}]} {"id": 500057, "image": "COCO_train2014_000000500057.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young girl eating a yellow vegetable and wearing a black sleeveless top\"."}], "task": "refering", "answer_template": "The \"a young girl eating a yellow vegetable and wearing a black sleeveless top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 32, 33, 34, 35, 36, 55, 56, 57, 58, 59, 77, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195], "bbox": [0.37246875, 0.011395833333333332, 0.606734375, 0.5372291666666666], "iscrowd": 0, "area": 27152.414199999992, "rle": {"size": [480, 640], "counts": "Ue_36j>i1VN7J7H8I7H8I6fEjL]8^3XGeLh8`3QGcLo8_3lFdLS9^3iFeLW9]3dFeL\\9]3`FfL`9\\3[FgLd9[3XFhLh9Y3TFjLk9U410O100O00100O100O010O1_NRKPIn4m6YKnHg4P7aKkH_4R7hKiHY4T7mKiHS4S7SLjHn3S7VLkHk3R7ZLlHf3Q7_LlHb3Q7cLmH]3Q7gLlHY3U7iLiHW3W7kLfHV3Z7X200000000000O10000000000000000O10000000000000000O1000000O1000000O10000O10000O10000O10000O10000N2O1O1N2O1N2O1N2dKPHg1Q8XNPHf1R8XNPHg1Q8XNPHf1R8XNQHe1Q8YNQHd1R8[NoGc1S8[NoGc1S8[NPHb1R8]NoGa1S8]NoGa1S8^NnG`1T8^NoG_1S8_NoG_1S8`NnG^1T8`NnG^1T8aNnG\\1T8bNnG^1R8bNnG_1P8aNQH`1n7`NSH_1m7aNSH`1l7`NTHa1k7_NUHb1l7\\NUHc1n7ZNRHg1o7WNQHj1P8TNQHl1Q8QNoGo1S8oMmGR2T8lMlGU2V8hMkGX2V8fMjGZ2X8dMhG]2Z8`MfGa2[8]MfGc2[8ZMfGf2]8WMcGj2^8TMbGm2_8QMbGo2`8nL`GR3b8lL^Gm2iNmLj94]Gi2SOPMa95`Gb2XOVMW98fGX2U9gMoFP2W9oMn3O1O1O1O1OUaf3"}, "label": "a young girl eating a yellow vegetable and wearing a black sleeveless top"}]} {"id": 450914, "image": "COCO_train2014_000000450914.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"stray berry was on a cup\"."}], "task": "refering", "answer_template": "The \"stray berry was on a cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 52, 70, 71, 72, 73, 74, 75, 76, 93, 94, 95, 96, 97, 98, 99, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 187, 188, 189], "bbox": [0.038546875, 0.14675, 0.340640625, 0.49314583333333334], "iscrowd": 0, "area": 22341.084499999997, "rle": {"size": [480, 640], "counts": "^k;5i>5J6K5K5J7J5J6K5J6K5K5J6K4M2N2M3N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2O2M2O1N2O1N2O1N3N1N2O1N2O1N3N1N2O0O100O2O0O100O2O0O100O101O0000001O0O1000001O00000O101O0000001O0O1000001O0000001N100000001O000O101O00000000000O1000000000000O11O000O101O0O101O0O101N10000O2O000O2O000O2O000O101N10001N10000O2O000O2O0O101O0O10001N1dN\\1J7I6J6K6M2N2O1N3M2N3M3M3N2M2N3M3M3N2M3N2N2N2N1O2O1N2N2N2N2N2O1N1O2N2N2N2O1NbfU6"}, "label": "stray berry was on a cup"}]} {"id": 450914, "image": "COCO_train2014_000000450914.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small white bowl with a few small strawberries in it\"."}], "task": "refering", "answer_template": "The \"a small white bowl with a few small strawberries in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 52, 70, 71, 72, 73, 74, 75, 76, 93, 94, 95, 96, 97, 98, 99, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 187, 188, 189], "bbox": [0.038546875, 0.14675, 0.340640625, 0.49314583333333334], "iscrowd": 0, "area": 22341.084499999997, "rle": {"size": [480, 640], "counts": "^k;5i>5J6K5K5J7J5J6K5J6K5K5J6K4M2N2M3N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2O2M2O1N2O1N2O1N3N1N2O1N2O1N3N1N2O0O100O2O0O100O2O0O100O101O0000001O0O1000001O00000O101O0000001O0O1000001O0000001N100000001O000O101O00000000000O1000000000000O11O000O101O0O101O0O101N10000O2O000O2O000O2O000O101N10001N10000O2O000O2O0O101O0O10001N1dN\\1J7I6J6K6M2N2O1N3M2N3M3M3N2M2N3M3M3N2M3N2N2N2N1O2O1N2N2N2N2N2O1N1O2N2N2N2O1NbfU6"}, "label": "a small white bowl with a few small strawberries in it"}]} {"id": 450914, "image": "COCO_train2014_000000450914.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"one of two bananas is placed immediately to the right of some oranges\"."}], "task": "refering", "answer_template": "The \"one of two bananas is placed immediately to the right of some oranges\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 124, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 221, 222, 223, 224, 225, 246, 247], "bbox": [0.42353125, 0.24235416666666668, 0.788828125, 0.6211666666666668], "iscrowd": 0, "area": 17704.816600000002, "rle": {"size": [480, 640], "counts": "UUo3?`>2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N2O1N2O1N2O001N2O1N101N2O0O200O1O00100O1O010O1O0010O01O010O1O0010O01O010O001O010O0010O01O1O010O001O010O0010O01O00100O1O010O1O10000O10O0100O100O1000O0100O100O100O01000O100O100O100O10000O100O100O10000O100O100O1O100O100O100O1O100O10O0100O1O100O100O100O00100O100O1O010O100O100O10O100000000O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O2N1O2M3N2N2N1N3N2N2N2M;WOWQo1"}, "label": "one of two bananas is placed immediately to the right of some oranges"}]} {"id": 450914, "image": "COCO_train2014_000000450914.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the banana that is to the outmost left hand side . it is more towards the oranges rather than the apples\"."}], "task": "refering", "answer_template": "The \"the banana that is to the outmost left hand side . it is more towards the oranges rather than the apples\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 124, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 221, 222, 223, 224, 225, 246, 247], "bbox": [0.42353125, 0.24235416666666668, 0.788828125, 0.6211666666666668], "iscrowd": 0, "area": 17704.816600000002, "rle": {"size": [480, 640], "counts": "UUo3?`>2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N2O1N2O1N2O001N2O1N101N2O0O200O1O00100O1O010O1O0010O01O010O1O0010O01O010O001O010O0010O01O1O010O001O010O0010O01O00100O1O010O1O10000O10O0100O100O1000O0100O100O100O01000O100O100O100O10000O100O100O10000O100O100O1O100O100O100O1O100O10O0100O1O100O100O100O00100O100O1O010O100O100O10O100000000O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O2N1O2M3N2N2N1N3N2N2N2M;WOWQo1"}, "label": "the banana that is to the outmost left hand side . it is more towards the oranges rather than the apples"}]} {"id": 450914, "image": "COCO_train2014_000000450914.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bowl of bananas and oranges\"."}], "task": "refering", "answer_template": "The \"bowl of bananas and oranges\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 60, 61, 62, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 309, 310, 311, 312, 313, 314, 315, 316, 334, 335, 336, 337], "bbox": [0.28040625, 0.116, 0.9138437500000001, 0.8434583333333333], "iscrowd": 0, "area": 105788.46470000007, "rle": {"size": [480, 640], "counts": "ZSd23U>l0F:I7H7J7H8I7I7H7J5J6K4K6M3N2M3M3M3M3M2N3M3M3M3M5K5K in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 60, 61, 62, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 309, 310, 311, 312, 313, 314, 315, 316, 334, 335, 336, 337], "bbox": [0.28040625, 0.116, 0.9138437500000001, 0.8434583333333333], "iscrowd": 0, "area": 105788.46470000007, "rle": {"size": [480, 640], "counts": "ZSd23U>l0F:I7H7J7H8I7I7H7J5J6K4K6M3N2M3M3M3M3M2N3M3M3M3M5K5K in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 60, 61, 62, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 309, 310, 311, 312, 313, 314, 315, 316, 334, 335, 336, 337], "bbox": [0.28040625, 0.116, 0.9138437500000001, 0.8434583333333333], "iscrowd": 0, "area": 105788.46470000007, "rle": {"size": [480, 640], "counts": "ZSd23U>l0F:I7H7J7H8I7I7H7J5J6K4K6M3N2M3M3M3M3M2N3M3M3M3M5K5K in this image.", "objects": [{"patches": [80, 81, 82, 83, 103, 104, 105, 106, 107, 108, 128, 129, 130, 131, 132, 153, 154, 155, 156, 177, 178, 179, 201, 202, 225], "bbox": [0.48706250000000006, 0.1836875, 0.8311718750000001, 0.536625], "iscrowd": 0, "area": 11907.474949999996, "rle": {"size": [480, 640], "counts": "[[b4:a>5N3N1O2M2O2N101O0O2O001O001O001N101O001O001O001N101O001O001O001N101O001O001O001N101O001O001O0O2O0010O01O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O0010O010O01O010O010O0010O010O01O010O010O0010O01O010O010O0010O010O01O010O0100O10O0100O100O010O100O010O11O000O10001O000O10001O000O10001O0O1000001O0O1000001O0O1000001N100001O01O0001O0001O01O0000010O00000010O000001O01O000001O01O00O2N1O1O1O2N1O1O1O2N1O1O3M2N3K5K4L5K5K4L5J5L5K5Kg\\b1"}, "label": "the bannana in the back"}]} {"id": 565608, "image": "COCO_train2014_000000565608.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tan colored leather bag\"."}], "task": "refering", "answer_template": "The \"the tan colored leather bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [179, 194, 195, 196, 197, 211, 212, 213, 214, 227, 228, 229, 230, 231, 232, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 277, 278, 279, 295, 296], "bbox": [0.3439166666666667, 0.46281249999999996, 0.6581875, 0.7711875], "iscrowd": 0, "area": 16090.071450000001, "rle": {"size": [640, 480], "counts": "\\aW38hc0:G9G9G:F9F3N00000I6mNS1O1N2N2O1N2N2O0O2O1N2N2O1N1O2O1N2O1N2N2O0O2N2O1N2O1N1O2O1N2N2O1N101N2N2O1N2N2O0O200OO1K5J7I6F:M4N1O11O100O1O00100O1O100O001O100O00O2N11O0010O01O010O001O010O010O10O001N1O2O0O2O001O011N2O0O2O1N2O1N2O1N3N1N2O1N2O2M2O1N3N1N3N1N3N2M2O2M3N1N3N2M2O2M2O2M3N1N3N2M2O4K4M3L5L3L5L3L4M4K4M4KoSV3"}, "label": "the tan colored leather bag"}]} {"id": 565608, "image": "COCO_train2014_000000565608.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tan bag\"."}], "task": "refering", "answer_template": "The \"a tan bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [179, 194, 195, 196, 197, 211, 212, 213, 214, 227, 228, 229, 230, 231, 232, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 277, 278, 279, 295, 296], "bbox": [0.3439166666666667, 0.46281249999999996, 0.6581875, 0.7711875], "iscrowd": 0, "area": 16090.071450000001, "rle": {"size": [640, 480], "counts": "\\aW38hc0:G9G9G:F9F3N00000I6mNS1O1N2N2O1N2N2O0O2O1N2N2O1N1O2O1N2O1N2N2O0O2N2O1N2O1N1O2O1N2N2O1N101N2N2O1N2N2O0O200OO1K5J7I6F:M4N1O11O100O1O00100O1O100O001O100O00O2N11O0010O01O010O001O010O010O10O001N1O2O0O2O001O011N2O0O2O1N2O1N2O1N3N1N2O1N2O2M2O1N3N1N3N1N3N2M2O2M3N1N3N2M2O2M2O2M3N1N3N2M2O4K4M3L5L3L5L3L4M4K4M4KoSV3"}, "label": "a tan bag"}]} {"id": 123247, "image": "COCO_train2014_000000123247.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player in jersey number 43\"."}], "task": "refering", "answer_template": "The \"a baseball player in jersey number 43\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 103, 122, 123, 124, 125, 126, 145, 146, 147, 148, 170, 171, 191, 192, 193, 194, 214, 215, 216, 217, 218, 219, 236, 237, 241, 242, 259, 260], "bbox": [0.26187499999999997, 0.3226578947368421, 0.550890625, 0.8181578947368421], "iscrowd": 0, "area": 11442.606500000002, "rle": {"size": [380, 640], "counts": "Xdn11i;3N1N3O00001O001O00001O0O101O001O000O2N1O2N100O2N1O2N1O1O2M2N3hLoNaKS1]4oN`KT1_4lN`KW1]4kNaKW1]4kNaKW1]4kN`KX1_4iN_KX1`4jN^KU1c4lN\\KS1e4oNXKQ1j4POTKn0n4TOPKk0Q5WOmJi0T5WOjJj0V5VOjJi0W5XOhJh0X5XOhJh0Y5XOdJj0\\5VOeJi0[5XOfJg0[5WOfJh0[5YOdJg0]5YOcJe0_5\\O_Jd0b5\\O^Jb0c5_O\\Ja0e5_OZJ`0h5AWJ=j5DUJgL2eLAd6:jL?X3^OkLa0V3[OmLe0S3XOPMh0Q3UORMj0P70001O001O0O2N2N2N2N2O1N2N2N2N2N1O2N2N^cZ3"}, "label": "a baseball player in jersey number 43"}]} {"id": 123247, "image": "COCO_train2014_000000123247.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the ball player throwing the ball\"."}], "task": "refering", "answer_template": "The \"the ball player throwing the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 103, 122, 123, 124, 125, 126, 145, 146, 147, 148, 170, 171, 191, 192, 193, 194, 214, 215, 216, 217, 218, 219, 236, 237, 241, 242, 259, 260], "bbox": [0.26187499999999997, 0.3226578947368421, 0.550890625, 0.8181578947368421], "iscrowd": 0, "area": 11442.606500000002, "rle": {"size": [380, 640], "counts": "Xdn11i;3N1N3O00001O001O00001O0O101O001O000O2N1O2N100O2N1O2N1O1O2M2N3hLoNaKS1]4oN`KT1_4lN`KW1]4kNaKW1]4kNaKW1]4kN`KX1_4iN_KX1`4jN^KU1c4lN\\KS1e4oNXKQ1j4POTKn0n4TOPKk0Q5WOmJi0T5WOjJj0V5VOjJi0W5XOhJh0X5XOhJh0Y5XOdJj0\\5VOeJi0[5XOfJg0[5WOfJh0[5YOdJg0]5YOcJe0_5\\O_Jd0b5\\O^Jb0c5_O\\Ja0e5_OZJ`0h5AWJ=j5DUJgL2eLAd6:jL?X3^OkLa0V3[OmLe0S3XOPMh0Q3UORMj0P70001O001O0O2N2N2N2N2O1N2N2N2N2N1O2N2N^cZ3"}, "label": "the ball player throwing the ball"}]} {"id": 131449, "image": "COCO_train2014_000000131449.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bus on the right\"."}], "task": "refering", "answer_template": "The \"the bus on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249], "bbox": [0.62375, 0.3550469483568075, 0.8855000000000001, 0.6876291079812207], "iscrowd": 0, "area": 21445.500699999997, "rle": {"size": [426, 640], "counts": "nWV59]V7A[INg60[3O0O2N2N1O2OaVn0"}, "label": "the bus on the right"}]} {"id": 131449, "image": "COCO_train2014_000000131449.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bus no is 1232 and it is on the left side\"."}], "task": "refering", "answer_template": "The \"the bus no is 1232 and it is on the left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 244, 245, 246, 247, 248, 249], "bbox": [0.62375, 0.3550469483568075, 0.8855000000000001, 0.6876291079812207], "iscrowd": 0, "area": 21445.500699999997, "rle": {"size": [426, 640], "counts": "nWV59]V7A[INg60[3O0O2N2N1O2OaVn0"}, "label": "the bus no is 1232 and it is on the left side"}]} {"id": 131449, "image": "COCO_train2014_000000131449.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bus , to the left of another bus\"."}], "task": "refering", "answer_template": "The \"a bus , to the left of another bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221], "bbox": [0.09828125, 0.36039906103286384, 0.638828125, 0.6601643192488263], "iscrowd": 0, "area": 35082.56695000001, "rle": {"size": [426, 640], "counts": "Y_j01X=1N2O1O1N2O1O1O1N2N2J6J6J6I7J6J6I7J6J6I7J6J6I7J6J6I7J60000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000001O0001O0000000000001O000000000000001O00000000000010O00000000000001O0000000000001O00000L4H8L4N2N3M2N2N2N2N2N2N3M21O001O001O010O001O001O00 in this image.", "objects": [{"patches": [118, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221], "bbox": [0.09828125, 0.36039906103286384, 0.638828125, 0.6601643192488263], "iscrowd": 0, "area": 35082.56695000001, "rle": {"size": [426, 640], "counts": "Y_j01X=1N2O1O1N2O1O1O1N2N2J6J6J6I7J6J6I7J6J6I7J6J6I7J6J6I7J60000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000001O0001O0000000000001O000000000000001O00000000000010O00000000000001O0000000000001O00000L4H8L4N2N3M2N2N2N2N2N2N3M21O001O001O010O001O001O00 in this image.", "objects": [{"patches": [78, 79, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 127, 146, 147, 148, 149, 169, 170, 171, 172, 192, 193, 194, 195, 214, 215, 216, 217, 218, 237, 238, 239, 240, 260, 261, 262, 263, 282, 283, 285, 286, 308, 309], "bbox": [0.295265625, 0.20393023255813952, 0.54834375, 0.9169302325581394], "iscrowd": 0, "area": 21169.936799999992, "rle": {"size": [430, 640], "counts": "af_27T=?B8H2N2M3N2N2M3N2N2N2M3N2N2M3N1O2N2M3N2M3bKoM[MT2S2_NjMa1W2aNfM_1\\2bNaM]1a2eN[M\\1f2fNWMZ1j2hNSMX1o2iNnLW1S3kNjLT1Y3mNcLT1_3nN]LR1f3nNWLR1k3oNRLQ1Q4oNlKP1W4QOeKP1^4PO_KP1c4QOZKo0i4QOTKn0o4SOmJn0V5ROgJn0[5SObJn0_5SO^Jo0c5QOZJP1h5POTJR1o5mNnIU1S6kNiIW1Z6hN`I\\1c6dNUIb1m6]NmHg1V7XNdHl1_7SNZHR2i7mMQHX2P8hMjG\\2Y8Z13N2M3M100O1O1O100O1ON3M2N2M4M2N3M>E;F:E;E:F;F:ZKeHe2e7gLeHZ3d7RLfHn3]8001O001O00001O001O0010O01O3N2N3M2N3M2N3M2N2ONK6F:E;F:F:E:H8I7H8I7H7J7H8I7H8I7I7H7PNcFJd95k100O1O100O10O01O2O1N2O2M2O1N2O2M2O1N3N1N2O1N3N1N2O2M2O1N2O2M2O1N3N1N2O1N3N1NZVi3"}, "label": "a skier with the number 377 on their chest"}]} {"id": 41357, "image": "COCO_train2014_000000041357.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person downhill skiing wearing a vest that says 377 on it and a blue jacket\"."}], "task": "refering", "answer_template": "The \"a person downhill skiing wearing a vest that says 377 on it and a blue jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 127, 146, 147, 148, 149, 169, 170, 171, 172, 192, 193, 194, 195, 214, 215, 216, 217, 218, 237, 238, 239, 240, 260, 261, 262, 263, 282, 283, 285, 286, 308, 309], "bbox": [0.295265625, 0.20393023255813952, 0.54834375, 0.9169302325581394], "iscrowd": 0, "area": 21169.936799999992, "rle": {"size": [430, 640], "counts": "af_27T=?B8H2N2M3N2N2M3N2N2N2M3N2N2M3N1O2N2M3N2M3bKoM[MT2S2_NjMa1W2aNfM_1\\2bNaM]1a2eN[M\\1f2fNWMZ1j2hNSMX1o2iNnLW1S3kNjLT1Y3mNcLT1_3nN]LR1f3nNWLR1k3oNRLQ1Q4oNlKP1W4QOeKP1^4PO_KP1c4QOZKo0i4QOTKn0o4SOmJn0V5ROgJn0[5SObJn0_5SO^Jo0c5QOZJP1h5POTJR1o5mNnIU1S6kNiIW1Z6hN`I\\1c6dNUIb1m6]NmHg1V7XNdHl1_7SNZHR2i7mMQHX2P8hMjG\\2Y8Z13N2M3M100O1O1O100O1ON3M2N2M4M2N3M>E;F:E;E:F;F:ZKeHe2e7gLeHZ3d7RLfHn3]8001O001O00001O001O0010O01O3N2N3M2N3M2N3M2N2ONK6F:E;F:F:E:H8I7H8I7H7J7H8I7H8I7I7H7PNcFJd95k100O1O100O10O01O2O1N2O2M2O1N2O2M2O1N3N1N2O1N3N1N2O2M2O1N2O2M2O1N3N1N2O1N3N1NZVi3"}, "label": "a person downhill skiing wearing a vest that says 377 on it and a blue jacket"}]} {"id": 205202, "image": "COCO_train2014_000000205202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the kid on the left\"."}], "task": "refering", "answer_template": "The \"the kid on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 73, 74, 75, 76, 95, 96, 97, 98, 99, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 207, 208, 209, 232, 233, 254, 255, 256, 277, 278, 279, 300, 301, 302, 322, 323, 324, 345, 346], "bbox": [0.015296874999999998, 0.14302277432712215, 0.31785937499999994, 0.9020289855072464], "iscrowd": 0, "area": 22617.22915, "rle": {"size": [483, 640], "counts": "hn48g>;oN_OnBi0e3VOn45VGi0l3UOj47RGh0S4TOi4\\1WKfNf4[1ZKgNc4Z1\\KiNb4W1^KlN^4U1bKnNZ4S1eKPOX4Q1hKROU4n0kKUOQ4l0nKWOn3k0RLVOl3k0TLUOk3l0ULSOj3o0ULQOk3P1ULPOj3Q1VLnNj3S1VLmNi3T1WLkNj3U1VLjNj3V1ZLgNf3Y1]LcNc3^1`L_N`3a1dLZN]3e1gLWNY3j1jLSNV3m1mLoMS3R2mIlMU9S2kFRNP9o1oFVNm8j1RG[Ni8f1VG_Ne8a1[GdN_8^1`GgNZ8[1hGgNR8[1QHfNi7[1[HfN_7\\1eHcNV7_1mHbNn6_1UIbNe6_1_IbN\\6_1hIaNS6`1hH\\MIU1Y7`1oH_MGR1U7`1SIbMIn0o6a1WIfMIj0j6b1]IhMHg0f6a1bImMHa0b6c1fIPNG<_6f1iIRNH6]6h1kIVNIN[6m1lIXNIHZ6Q2lI[NJAY6U2mI]NJZOY6Y2mIaNKROX6]2lIeNLkNW6a2mIfNMeNW6d2QJfNIbNU6i2XJbNCbNT6m2^J^N_OaNT6Q3bJ\\Nd6d1bIZN[6f1kIVNR6k1SJSNi5m1]JQN`5o1eJnMX5S2nJjMo4U2ZITMW6Y2`I_M68W6Y2cIcM54U6Y2fIgM4MU6\\2hIjM3FV6`2fIoM3^OW6c2fISN1WOZ6f2eIVN1QOY6i2gIZNOiN[6m2fI]NObN\\6Q3eIaNc7`1]HbNa7^1eH^NY7a1nH[NP7e1UIYNg6h1_ITN_6l1gIPNW6o1PJmMl5U2ZJhM`5[2PKWMl4k2c32O2N2O1O1O1O1O1O2N1N2M3H8G:G8O1N2N3M2N01O00000O2000000000O100O100O101O01O1O1O1O1O1O1O1O1O1O100O1O1N2N3M2M3O1O1N2O2M2O1O1N2_OoDoLR;Q3TEgLo:X3>0000O10001N1O100O1O2N2O0O2N2O1M3M3M3N2M3M3N2M3BXC]Nk<`1[CZNhM3M3M3M4K4M4K4L5K4K5HPQ^6"}, "label": "the kid on the left"}]} {"id": 205202, "image": "COCO_train2014_000000205202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little boy standing behind chair\"."}], "task": "refering", "answer_template": "The \"little boy standing behind chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 73, 74, 75, 76, 95, 96, 97, 98, 99, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 207, 208, 209, 232, 233, 254, 255, 256, 277, 278, 279, 300, 301, 302, 322, 323, 324, 345, 346], "bbox": [0.015296874999999998, 0.14302277432712215, 0.31785937499999994, 0.9020289855072464], "iscrowd": 0, "area": 22617.22915, "rle": {"size": [483, 640], "counts": "hn48g>;oN_OnBi0e3VOn45VGi0l3UOj47RGh0S4TOi4\\1WKfNf4[1ZKgNc4Z1\\KiNb4W1^KlN^4U1bKnNZ4S1eKPOX4Q1hKROU4n0kKUOQ4l0nKWOn3k0RLVOl3k0TLUOk3l0ULSOj3o0ULQOk3P1ULPOj3Q1VLnNj3S1VLmNi3T1WLkNj3U1VLjNj3V1ZLgNf3Y1]LcNc3^1`L_N`3a1dLZN]3e1gLWNY3j1jLSNV3m1mLoMS3R2mIlMU9S2kFRNP9o1oFVNm8j1RG[Ni8f1VG_Ne8a1[GdN_8^1`GgNZ8[1hGgNR8[1QHfNi7[1[HfN_7\\1eHcNV7_1mHbNn6_1UIbNe6_1_IbN\\6_1hIaNS6`1hH\\MIU1Y7`1oH_MGR1U7`1SIbMIn0o6a1WIfMIj0j6b1]IhMHg0f6a1bImMHa0b6c1fIPNG<_6f1iIRNH6]6h1kIVNIN[6m1lIXNIHZ6Q2lI[NJAY6U2mI]NJZOY6Y2mIaNKROX6]2lIeNLkNW6a2mIfNMeNW6d2QJfNIbNU6i2XJbNCbNT6m2^J^N_OaNT6Q3bJ\\Nd6d1bIZN[6f1kIVNR6k1SJSNi5m1]JQN`5o1eJnMX5S2nJjMo4U2ZITMW6Y2`I_M68W6Y2cIcM54U6Y2fIgM4MU6\\2hIjM3FV6`2fIoM3^OW6c2fISN1WOZ6f2eIVN1QOY6i2gIZNOiN[6m2fI]NObN\\6Q3eIaNc7`1]HbNa7^1eH^NY7a1nH[NP7e1UIYNg6h1_ITN_6l1gIPNW6o1PJmMl5U2ZJhM`5[2PKWMl4k2c32O2N2O1O1O1O1O1O2N1N2M3H8G:G8O1N2N3M2N01O00000O2000000000O100O100O101O01O1O1O1O1O1O1O1O1O1O100O1O1N2N3M2M3O1O1N2O2M2O1O1N2_OoDoLR;Q3TEgLo:X3>0000O10001N1O100O1O2N2O0O2N2O1M3M3M3N2M3M3N2M3BXC]Nk<`1[CZNhM3M3M3M4K4M4K4L5K4K5HPQ^6"}, "label": "little boy standing behind chair"}]} {"id": 205202, "image": "COCO_train2014_000000205202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl in gray hoodie leaning on table\"."}], "task": "refering", "answer_template": "The \"girl in gray hoodie leaning on table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 201, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271], "bbox": [0.42459375, 0.24138716356107662, 0.8592343750000001, 0.6985714285714285], "iscrowd": 0, "area": 32508.034349999984, "rle": {"size": [483, 640], "counts": "PaP43l>6K5N1N3M3M2O2M2N3M3M2O2M3M2N3N2M2N3M3M2O2M3M2O2N2M2O2N1N3N2N1O2M3N1N3M3N1N3M3N1N3M3N1N3M2O0O2O0O1O2O0O2N1O1O2N1O2O0O1O2M2N2N3N1N3M2N2O2O001O00001O0D^KVFc4j9bKQF_4n9:010O1O001O10O01O1O010O1O00100O1O0102L3N2N2M3N2M3N2N2M4M2M3N1O2M2N2O1N3N1O1O1O2O0O1O1O2O000O101N100O1L301N2O1O1O010O1000O1000001N2O1O001O1N1000000O100000000O2O1O1O1N2O1O1O1N2O1O1O1O100O2N100O1O106I7I7UKXFm3n9cK^F[4V:O100O1O1O1O1O1O1O1O2N1O1O10001O2N2N2O1M3M3M3L5L3M3M3L4M3M3M3L3N3M2N2M4M2N2N2nNmBOW=LjB5X=HhB9Z=ChB=n=010O1O010O1O10O0100O1O010O1O1O00100O1O001O100O001O1O10O01O01O0O2O1O1O1O1O1N101O1O1O1N2O1O1O001O1N2O1O1O1O1N4M6J7I_UZ1"}, "label": "girl in gray hoodie leaning on table"}]} {"id": 205202, "image": "COCO_train2014_000000205202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl looking at laptop screen\"."}], "task": "refering", "answer_template": "The \"girl looking at laptop screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 201, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271], "bbox": [0.42459375, 0.24138716356107662, 0.8592343750000001, 0.6985714285714285], "iscrowd": 0, "area": 32508.034349999984, "rle": {"size": [483, 640], "counts": "PaP43l>6K5N1N3M3M2O2M2N3M3M2O2M3M2N3N2M2N3M3M2O2M3M2O2N2M2O2N1N3N2N1O2M3N1N3M3N1N3M3N1N3M3N1N3M2O0O2O0O1O2O0O2N1O1O2N1O2O0O1O2M2N2N3N1N3M2N2O2O001O00001O0D^KVFc4j9bKQF_4n9:010O1O001O10O01O1O010O1O00100O1O0102L3N2N2M3N2M3N2N2M4M2M3N1O2M2N2O1N3N1O1O1O2O0O1O1O2O000O101N100O1L301N2O1O1O010O1000O1000001N2O1O001O1N1000000O100000000O2O1O1O1N2O1O1O1N2O1O1O1O100O2N100O1O106I7I7UKXFm3n9cK^F[4V:O100O1O1O1O1O1O1O1O2N1O1O10001O2N2N2O1M3M3M3L5L3M3M3L4M3M3M3L3N3M2N2M4M2N2N2nNmBOW=LjB5X=HhB9Z=ChB=n=010O1O010O1O10O0100O1O010O1O1O00100O1O001O100O001O1O10O01O01O0O2O1O1O1O1O1N101O1O1O1N2O1O1O001O1N2O1O1O1O1N4M6J7I_UZ1"}, "label": "girl looking at laptop screen"}]} {"id": 205202, "image": "COCO_train2014_000000205202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man types on a laptop while two youngsters watch\"."}], "task": "refering", "answer_template": "The \"a young man types on a laptop while two youngsters watch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 99, 100, 101, 102, 122, 123, 124, 125, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 302, 303, 308, 309, 310, 311, 312, 313, 314, 331, 332, 333, 334, 335, 336, 337, 354, 355, 377], "bbox": [0.154671875, 0.21057971014492752, 0.742796875, 0.9740993788819875], "iscrowd": 0, "area": 50423.32979999998, "rle": {"size": [483, 640], "counts": "RR_1;R>e0[Of0@?L5K5K4M4K5K4L5L3L5K5L3O2N2O0O2N1O2N2N1O2N1O2N2O0O2O1O0O2O001N2O001N2O010O00100O1O010O1O00100O00100O00100O10O10O10O10O10O10O1000O01000O01000O01000O10O1000O01O1O001O1N101O1O001O1O1N101O1K4J7XOh0ZOe0O2N2N1O2N2N2N1O2N2O00100O00100O00100O1O00100O010000000O01000000O1000000O10000000O01000000O100000000OSKfHS2Z7lMnHm1R7SNVIf1i6ZN\\Ia1c6_N`I_1`6aNbI^1]6bNeI\\1\\6cNfI[1Z6dNiI[1W6dNkIZ1V6eNlIZ1S6fNnIY1S6fNoIX1Q6gNQJY1o5fNQJZ1P6eNPJ[1P6eNPJ\\1P6bNPJ_1Q6]NRJd1m5YNVJg1o5QNTJo1R6hMQJY2c6bLlI`3k81gHULW3l3eL[LW3g3fL^LV3d3gLaLU3a3hLdLT3^3bLnLZ3T3eLoLW3S3hLoLU3R3kLPMR3R3lLQMR3P3mLSMo2o2PMSMm2o2RMSMk2o2SMTMj2m2TMWMi2k2UMYMc1oMhLj4b1[Mc1lMjLk4T1iMP2]MjLl4S1kMQ2YMkLn4Q1nMP2VMmLn4o0QNR2QMnLQ5k0SNU2mLoLR5HmJ7Y3o2iLQMS5FRK2W3U3eLQMU5BYO\\3bKQMX5\\O\\Oa3^KQMW:n2iEQMY:n2gEQM[:m2gEQM\\:m2dERM^:m2cEQM_:m2bERM`:l2bERM`:l2aETM`:k2aESMa:k2aESMa:k2aESMa:l2`ERMc:k2_ESMc:k2l0N2O1N2N1O2O1N101O1O001O010O1O001O1O0001O0000010O00001O00010O00001O000001O000000000O101O000000000O101O0000000001O01O0000000000001O0001O1O001O10O01OeNZC:eo;BSD in this image.", "objects": [{"patches": [76, 77, 78, 99, 100, 101, 102, 122, 123, 124, 125, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 302, 303, 308, 309, 310, 311, 312, 313, 314, 331, 332, 333, 334, 335, 336, 337, 354, 355, 377], "bbox": [0.154671875, 0.21057971014492752, 0.742796875, 0.9740993788819875], "iscrowd": 0, "area": 50423.32979999998, "rle": {"size": [483, 640], "counts": "RR_1;R>e0[Of0@?L5K5K4M4K5K4L5L3L5K5L3O2N2O0O2N1O2N2N1O2N1O2N2O0O2O1O0O2O001N2O001N2O010O00100O1O010O1O00100O00100O00100O10O10O10O10O10O10O1000O01000O01000O01000O10O1000O01O1O001O1N101O1O001O1O1N101O1K4J7XOh0ZOe0O2N2N1O2N2N2N1O2N2O00100O00100O00100O1O00100O010000000O01000000O1000000O10000000O01000000O100000000OSKfHS2Z7lMnHm1R7SNVIf1i6ZN\\Ia1c6_N`I_1`6aNbI^1]6bNeI\\1\\6cNfI[1Z6dNiI[1W6dNkIZ1V6eNlIZ1S6fNnIY1S6fNoIX1Q6gNQJY1o5fNQJZ1P6eNPJ[1P6eNPJ\\1P6bNPJ_1Q6]NRJd1m5YNVJg1o5QNTJo1R6hMQJY2c6bLlI`3k81gHULW3l3eL[LW3g3fL^LV3d3gLaLU3a3hLdLT3^3bLnLZ3T3eLoLW3S3hLoLU3R3kLPMR3R3lLQMR3P3mLSMo2o2PMSMm2o2RMSMk2o2SMTMj2m2TMWMi2k2UMYMc1oMhLj4b1[Mc1lMjLk4T1iMP2]MjLl4S1kMQ2YMkLn4Q1nMP2VMmLn4o0QNR2QMnLQ5k0SNU2mLoLR5HmJ7Y3o2iLQMS5FRK2W3U3eLQMU5BYO\\3bKQMX5\\O\\Oa3^KQMW:n2iEQMY:n2gEQM[:m2gEQM\\:m2dERM^:m2cEQM_:m2bERM`:l2bERM`:l2aETM`:k2aESMa:k2aESMa:k2aESMa:l2`ERMc:k2_ESMc:k2l0N2O1N2N1O2O1N101O1O001O010O1O001O1O0001O0000010O00001O00010O00001O000001O000000000O101O000000000O101O0000000001O01O0000000000001O0001O1O001O10O01OeNZC:eo;BSD in this image.", "objects": [{"patches": [281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 331, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0033750000000000004, 0.7071428571428572, 0.42546875, 0.9890269151138718], "iscrowd": 0, "area": 21143.829699999995, "rle": {"size": [483, 640], "counts": "e\\11Q?001O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O2N1N2O1O1O1O100O1O100O100O1O2O0O1O100O100O1O100O1O100O101N1O100O100O1O100O1O100O101N1O100O1O100O101N2N2O1N2N101N101N1O100O100O1O101N1O100O100O1O100O2N100O100O1O100O101N1O100O100O10ElCgMSO010O010O10O0100O0100O0100O010O10O0100O010O010O10O4M4L3L4M3L5L3M2M10000O100O10000OPO_D`Na;`1cD\\N];c1hDYNW;h1nDSNR;m1REoMm:Q2XEkMh:U2]EfMc:Z2R101000O010O10O10O10O10O100O01000O010000:E=D in this image.", "objects": [{"patches": [281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 308, 323, 324, 325, 326, 327, 328, 329, 330, 331, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0033750000000000004, 0.7071428571428572, 0.42546875, 0.9890269151138718], "iscrowd": 0, "area": 21143.829699999995, "rle": {"size": [483, 640], "counts": "e\\11Q?001O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O2N1N2O1O1O1O100O1O100O100O1O2O0O1O100O100O1O100O1O100O101N1O100O100O1O100O1O100O101N1O100O1O100O101N2N2O1N2N101N101N1O100O100O1O101N1O100O100O1O100O2N100O100O1O100O101N1O100O100O10ElCgMSO010O010O10O0100O0100O0100O010O10O0100O010O010O10O4M4L3L4M3L5L3M2M10000O100O10000OPO_D`Na;`1cD\\N];c1hDYNW;h1nDSNR;m1REoMm:Q2XEkMh:U2]EfMc:Z2R101000O010O10O10O10O10O100O01000O010000:E=D in this image.", "objects": [{"patches": [187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 300, 301, 302, 323, 324], "bbox": [0.020265625000000002, 0.5214699792960663, 0.18234375, 0.8794202898550725], "iscrowd": 0, "area": 11966.7806, "rle": {"size": [483, 640], "counts": "Z^66c>=C>D;K5J6J6J7K4K5L4K6K4K5L4L4K6K4M3L4M3M4K4M3M3L5L3M3M3L4MN2O10O01O10O01O100O00100O1O010O1O10O01O100O00100O00100O1O010O1O10O01O100O00100O1O10O01O100O1O100O00100O1O100O00100O1O100O1]M\\Ef0e:PORFc0n9XOeF9\\9DTGNl80eGB[8;X3M3M3M3M3M3N2MX^f7"}, "label": "a back rest of a wooden chair which was occupied by a man"}]} {"id": 205202, "image": "COCO_train2014_000000205202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair with a boy sitting in it and a boy leaning on it\"."}], "task": "refering", "answer_template": "The \"a chair with a boy sitting in it and a boy leaning on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 300, 301, 302, 323, 324], "bbox": [0.020265625000000002, 0.5214699792960663, 0.18234375, 0.8794202898550725], "iscrowd": 0, "area": 11966.7806, "rle": {"size": [483, 640], "counts": "Z^66c>=C>D;K5J6J6J7K4K5L4K6K4K5L4L4K6K4M3L4M3M4K4M3M3L5L3M3M3L4MN2O10O01O10O01O100O00100O1O010O1O10O01O100O00100O00100O1O010O1O10O01O100O00100O1O10O01O100O1O100O00100O1O100O00100O1O100O1]M\\Ef0e:PORFc0n9XOeF9\\9DTGNl80eGB[8;X3M3M3M3M3M3N2MX^f7"}, "label": "a chair with a boy sitting in it and a boy leaning on it"}]} {"id": 237976, "image": "COCO_train2014_000000237976.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl with the camera on her hand to take selfie\"."}], "task": "refering", "answer_template": "The \"a girl with the camera on her hand to take selfie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 101, 117, 118, 135, 152, 169, 186, 304, 305, 319, 320, 321, 322, 334, 335, 336, 337, 338, 339, 350, 351, 352, 353, 354, 355, 356, 367, 368, 369, 370, 371, 372, 373, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.5441041666666667, 0.18534375, 0.9955208333333334, 0.99028125], "iscrowd": 0, "area": 23397.276800000007, "rle": {"size": [640, 480], "counts": "g[T52lc02O1O1N2O1O2M2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1O1O1O1O100O1O1O1O1O001O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O100O1000000O10000O1000000O1000000O10000O1000000O10000O1000000O1000000O100O10000O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O10000O100O100O100O10000O100O100O10000O100O10000O1O1O1O1O1O1O1O1O1O100O100O100^FlKW3U4YL^Ld3b3kKRMQ4o2fK]MW4c2hKaMVMTNn4\\4jMeMSMSNQ5X4kMiMoLRNT5V4lMlMkLRNW5R4mMPNfLSNZ5n3oMRNcLSN]5k3nMWN`LRN_5h3PNZN\\LRNb5d3QNbNSLoMi5`3SNhNlKkMP6]3SNPOdKgMV6Y3VNf1DYN?c1C\\N`0`1B_N`0^1BaNa0\\1@bNd0Z1^OeNe0X1\\OgNf0W1[OhNh0U1YOjNj0T1VOkNgNX3Z1R8Ed3\\LZe0"}, "label": "a girl with the camera on her hand to take selfie"}]} {"id": 237976, "image": "COCO_train2014_000000237976.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl getting ready to take a shelfi in the bathroom\"."}], "task": "refering", "answer_template": "The \"girl getting ready to take a shelfi in the bathroom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 101, 117, 118, 135, 152, 169, 186, 304, 305, 319, 320, 321, 322, 334, 335, 336, 337, 338, 339, 350, 351, 352, 353, 354, 355, 356, 367, 368, 369, 370, 371, 372, 373, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.5441041666666667, 0.18534375, 0.9955208333333334, 0.99028125], "iscrowd": 0, "area": 23397.276800000007, "rle": {"size": [640, 480], "counts": "g[T52lc02O1O1N2O1O2M2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1O1O1O1O100O1O1O1O1O001O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O100O1000000O10000O1000000O1000000O10000O1000000O10000O1000000O1000000O100O10000O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O10000O100O100O100O10000O100O100O10000O100O10000O1O1O1O1O1O1O1O1O1O100O100O100^FlKW3U4YL^Ld3b3kKRMQ4o2fK]MW4c2hKaMVMTNn4\\4jMeMSMSNQ5X4kMiMoLRNT5V4lMlMkLRNW5R4mMPNfLSNZ5n3oMRNcLSN]5k3nMWN`LRN_5h3PNZN\\LRNb5d3QNbNSLoMi5`3SNhNlKkMP6]3SNPOdKgMV6Y3VNf1DYN?c1C\\N`0`1B_N`0^1BaNa0\\1@bNd0Z1^OeNe0X1\\OgNf0W1[OhNh0U1YOjNj0T1VOkNgNX3Z1R8Ed3\\LZe0"}, "label": "girl getting ready to take a shelfi in the bathroom"}]} {"id": 311706, "image": "COCO_train2014_000000311706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table that the laptop is sitting on\"."}], "task": "refering", "answer_template": "The \"the table that the laptop is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 210, 212, 213, 226, 227, 228], "bbox": [0.5447000000000001, 0.7856533333333333, 0.9635600000000001, 1.0], "iscrowd": 0, "area": 1400.58915, "rle": {"size": [375, 500], "counts": "lZT32c;2O1N2O1N2N2O1N20000O10O1O001N2O0O2O1N101O1O01000000O10001N10000O10000O101O0O10000O10000O2O000O10000O1nD_On:=PEE1Oo:5^PA2M4M3M3L4MO1N1O2O1N1O2O1K4L5O1O001O_e6"}, "label": "the table that the laptop is sitting on"}]} {"id": 311706, "image": "COCO_train2014_000000311706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table with computers on it\"."}], "task": "refering", "answer_template": "The \"the table with computers on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 210, 212, 213, 226, 227, 228], "bbox": [0.5447000000000001, 0.7856533333333333, 0.9635600000000001, 1.0], "iscrowd": 0, "area": 1400.58915, "rle": {"size": [375, 500], "counts": "lZT32c;2O1N2O1N2N2O1N20000O10O1O001N2O0O2O1N101O1O01000000O10001N10000O10000O101O0O10000O10000O2O000O10000O1nD_On:=PEE1Oo:5^PA2M4M3M3L4MO1N1O2O1N1O2O1K4L5O1O001O_e6"}, "label": "the table with computers on it"}]} {"id": 311706, "image": "COCO_train2014_000000311706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hand holding a mobile phone near by the two man walking near\"."}], "task": "refering", "answer_template": "The \"a hand holding a mobile phone near by the two man walking near\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 124, 125, 143, 161, 179, 197, 214, 215, 233], "bbox": [0.8646, 0.3887733333333333, 1.0, 0.9797866666666666], "iscrowd": 0, "area": 4429.677050000001, "rle": {"size": [375, 500], "counts": "ifn44d;3L4M3L3N000O100O100O100O1L4L4O1O1O100O0010O01O0010OoIROY1o0eNRO[1n0cNUO]1k0aNWO^1i0`NZO_1g0_N[O`1e0_N\\Ob1d0[N_Oe1a0YNAg1>WNEi1;UNGl18QNKo15oMLR24kMOU21hM2X2NeM5[2KaM9_2G^M;c2D[M?f2@WMc0i2]OSMg0m2YOPMj0P3VOmLl0T3TOhLP1X3POeLS1\\3mN`LV1`3lN[LW1e3kNULY1k3hNQLZ1P4hNjK\\1W4eNcK_1]4cN]Ka1c4`NXKd1h4^NRKe1o4]NjJh1W5YNcJk1]5VN^Jn1b5TNXJP2h5RNRJQ2o5QNjIT2W6f1PJ"}, "label": "a hand holding a mobile phone near by the two man walking near"}]} {"id": 311706, "image": "COCO_train2014_000000311706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hand holding a cellphone\"."}], "task": "refering", "answer_template": "The \"a hand holding a cellphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 124, 125, 143, 161, 179, 197, 214, 215, 233], "bbox": [0.8646, 0.3887733333333333, 1.0, 0.9797866666666666], "iscrowd": 0, "area": 4429.677050000001, "rle": {"size": [375, 500], "counts": "ifn44d;3L4M3L3N000O100O100O100O1L4L4O1O1O100O0010O01O0010OoIROY1o0eNRO[1n0cNUO]1k0aNWO^1i0`NZO_1g0_N[O`1e0_N\\Ob1d0[N_Oe1a0YNAg1>WNEi1;UNGl18QNKo15oMLR24kMOU21hM2X2NeM5[2KaM9_2G^M;c2D[M?f2@WMc0i2]OSMg0m2YOPMj0P3VOmLl0T3TOhLP1X3POeLS1\\3mN`LV1`3lN[LW1e3kNULY1k3hNQLZ1P4hNjK\\1W4eNcK_1]4cN]Ka1c4`NXKd1h4^NRKe1o4]NjJh1W5YNcJk1]5VN^Jn1b5TNXJP2h5RNRJQ2o5QNjIT2W6f1PJ"}, "label": "a hand holding a cellphone"}]} {"id": 311706, "image": "COCO_train2014_000000311706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in green long sleeved shirt with light blue denim pants\"."}], "task": "refering", "answer_template": "The \"man in green long sleeved shirt with light blue denim pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 99, 100, 101, 117, 118, 119, 120, 135, 136, 137, 138, 153, 154, 155, 171, 172, 173, 190, 191, 208, 209, 210, 226, 227], "bbox": [0.49607999999999997, 0.3094133333333333, 0.6945, 0.9641333333333334], "iscrowd": 0, "area": 13941.4504, "rle": {"size": [375, 500], "counts": "TRk21d;2M4M2N2M3N2N3M2mN]OoFf0o8]OnFe0P9^OmFe0o8@mFb0Q9AkFb0R9CjF?T9DiF>U9FgF=U9GgF in this image.", "objects": [{"patches": [82, 83, 99, 100, 101, 117, 118, 119, 120, 135, 136, 137, 138, 153, 154, 155, 171, 172, 173, 190, 191, 208, 209, 210, 226, 227], "bbox": [0.49607999999999997, 0.3094133333333333, 0.6945, 0.9641333333333334], "iscrowd": 0, "area": 13941.4504, "rle": {"size": [375, 500], "counts": "TRk21d;2M4M2N2M3N2N3M2mN]OoFf0o8]OnFe0P9^OmFe0o8@mFb0Q9AkFb0R9CjF?T9DiF>U9FgF=U9GgF in this image.", "objects": [{"patches": [127, 128, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 149], "bbox": [0.001, 0.5573333333333333, 0.437, 0.676], "iscrowd": 0, "area": 3641.25, "rle": {"size": [375, 500], "counts": "Qc0a0U;3N00000000000000O10O100000000000O100000O100000O100000000000O10O1000000000000O1000O1000000000O10000000O100000O100000000000O0100000000000000O10O100000000000O100000O100000O10000000000000O0100000000000000O10O1000000000O10000000O100000O100000000000O10O1000000000000O1000O1000000000O10000000O100000O100000000000O0100000000000000O10O1000000000O1000000000O1000O100000000000:F2M10O100@;nDER;`0000O100000O10000000O100000000000000000O100000000000J6EifV3"}, "label": "the part of the table in front of the garbage can"}]} {"id": 311706, "image": "COCO_train2014_000000311706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table next to the men with a beer bottle on it\"."}], "task": "refering", "answer_template": "The \"the table next to the men with a beer bottle on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 149], "bbox": [0.001, 0.5573333333333333, 0.437, 0.676], "iscrowd": 0, "area": 3641.25, "rle": {"size": [375, 500], "counts": "Qc0a0U;3N00000000000000O10O100000000000O100000O100000O100000000000O10O1000000000000O1000O1000000000O10000000O100000O100000000000O0100000000000000O10O100000000000O100000O100000O10000000000000O0100000000000000O10O1000000000O10000000O100000O100000000000O10O1000000000000O1000O1000000000O10000000O100000O100000000000O0100000000000000O10O1000000000O1000000000O1000O100000000000:F2M10O100@;nDER;`0000O100000O10000000O100000000000000000O100000000000J6EifV3"}, "label": "the table next to the men with a beer bottle on it"}]} {"id": 311706, "image": "COCO_train2014_000000311706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man walking with long hair\"."}], "task": "refering", "answer_template": "The \"a man walking with long hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 97, 98, 99, 114, 115, 116, 117, 132, 133, 134, 135, 151, 152, 153, 169, 170, 187, 188, 189, 205, 206, 207, 223], "bbox": [0.3668, 0.3353333333333333, 0.53284, 0.9366933333333333], "iscrowd": 0, "area": 8340.451200000001, "rle": {"size": [375, 500], "counts": "eVS26a;1O1O1O1O1O001N1000001O0000001O0000000O10OQIGa2:]MHc27]MJb27\\MKd24\\MMc23]MNa24\\MOb22]M0a22\\M1X2:dJSOW2f0S38_JLh10g36\\J`0W1_O[42[JT1f0mNo4NXJk13[Nd5KSJb3m5T1O1OL5M201O0010O1000O010OVO_J_Kb5`4PKoJP5P5TKmJk4S5m0O00N2N3M3O010O1000000O2O001ZNPKcLP5R3bKgL_4T3hKiLX4S3oKjLU4m2SLPMP44VJh1V2RNe33XJe1Z2VN_32ZJc1]2VN]34YJa1`2UN[37YJX1h9A?E;N2N2M3N2Nm`e2"}, "label": "a man walking with long hair"}]} {"id": 311706, "image": "COCO_train2014_000000311706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with longer hair past his ears , standing to the right of another man\"."}], "task": "refering", "answer_template": "The \"a man with longer hair past his ears , standing to the right of another man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 97, 98, 99, 114, 115, 116, 117, 132, 133, 134, 135, 151, 152, 153, 169, 170, 187, 188, 189, 205, 206, 207, 223], "bbox": [0.3668, 0.3353333333333333, 0.53284, 0.9366933333333333], "iscrowd": 0, "area": 8340.451200000001, "rle": {"size": [375, 500], "counts": "eVS26a;1O1O1O1O1O001N1000001O0000001O0000000O10OQIGa2:]MHc27]MJb27\\MKd24\\MMc23]MNa24\\MOb22]M0a22\\M1X2:dJSOW2f0S38_JLh10g36\\J`0W1_O[42[JT1f0mNo4NXJk13[Nd5KSJb3m5T1O1OL5M201O0010O1000O010OVO_J_Kb5`4PKoJP5P5TKmJk4S5m0O00N2N3M3O010O1000000O2O001ZNPKcLP5R3bKgL_4T3hKiLX4S3oKjLU4m2SLPMP44VJh1V2RNe33XJe1Z2VN_32ZJc1]2VN]34YJa1`2UN[37YJX1h9A?E;N2N2M3N2Nm`e2"}, "label": "a man with longer hair past his ears , standing to the right of another man"}]} {"id": 491936, "image": "COCO_train2014_000000491936.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man sitting with the blue and green jacket\"."}], "task": "refering", "answer_template": "The \"the man sitting with the blue and green jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349, 350, 351, 352, 353, 368, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0, 0.3141875, 0.40203125, 0.989875], "iscrowd": 0, "area": 52864.7087, "rle": {"size": [480, 640], "counts": "e5T1[6UOl0S3lL_Mb2T4kK^Ld3e3YLgLZ3\\3cLeL\\3^3bLbL^3`3_LaL`3b3]L_Lc3c3[L]Ld3g3XLZLg3i3WLWLi3k3TLVLk3m3RLTLm3o3QLQLo3Q4nKPLQ4T4lKlKT4`7O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1O1000000000000000000O100000000000000000000000000000000000000000000000000000000001O0000000000000000000000001O0000000000001O001O001O001O001O1O001O001O001O001N101O001O001O001O001O001TMdJRL]5f3jJZLV5`3PK`LQ5X3VKhLj4R3\\KnLe4j2cKUM]4e2jKZMW4b2mK]MS4a2PL^MQ4_2RL`Mn3]2ULcMl3Z2WLeMi3Y2ZLfMg3V2]LiMe3T2]LkMe3S2[LmMf3R2[LmMg3Q2ZLnMg3Q2ZLnMh3P2YLoMi3o1XLPNi3P2VLPNl3n1ULQNm3m1TLRNm3m1TLRNn3l1SLSNo3k1RLTNo3i1SLWNo3b1WL]Nj3]1\\LbNf3X1_LgNc3R1cLmN^33]IfNY3V1\\3]OQJYOf2Z1^5cNfJ\\1Z5bNiJ]1W5`NmJ_1S5^NQKa1o4\\NTKd1l4YNXKf1h4WN\\Kh1o8OO10000O10000O10000O10000O10000O11O000000001O0000001O000000001O000O101N10001N100O2O000O2O0O2O000O2O0O101O0O101N101O0O101N3N1O1N3N1N2O1O2M2O1N3N1O1N2O2M2O1O2M2O1N2OlRc5"}, "label": "the man sitting with the blue and green jacket"}]} {"id": 491936, "image": "COCO_train2014_000000491936.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on a double decker bus tour\"."}], "task": "refering", "answer_template": "The \"man on a double decker bus tour\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349, 350, 351, 352, 353, 368, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.0, 0.3141875, 0.40203125, 0.989875], "iscrowd": 0, "area": 52864.7087, "rle": {"size": [480, 640], "counts": "e5T1[6UOl0S3lL_Mb2T4kK^Ld3e3YLgLZ3\\3cLeL\\3^3bLbL^3`3_LaL`3b3]L_Lc3c3[L]Ld3g3XLZLg3i3WLWLi3k3TLVLk3m3RLTLm3o3QLQLo3Q4nKPLQ4T4lKlKT4`7O1O100O1O100O100O1O100O1O100O1O100O100O1O100O1O100O100O1O1000000000000000000O100000000000000000000000000000000000000000000000000000000001O0000000000000000000000001O0000000000001O001O001O001O001O1O001O001O001O001N101O001O001O001O001O001TMdJRL]5f3jJZLV5`3PK`LQ5X3VKhLj4R3\\KnLe4j2cKUM]4e2jKZMW4b2mK]MS4a2PL^MQ4_2RL`Mn3]2ULcMl3Z2WLeMi3Y2ZLfMg3V2]LiMe3T2]LkMe3S2[LmMf3R2[LmMg3Q2ZLnMg3Q2ZLnMh3P2YLoMi3o1XLPNi3P2VLPNl3n1ULQNm3m1TLRNm3m1TLRNn3l1SLSNo3k1RLTNo3i1SLWNo3b1WL]Nj3]1\\LbNf3X1_LgNc3R1cLmN^33]IfNY3V1\\3]OQJYOf2Z1^5cNfJ\\1Z5bNiJ]1W5`NmJ_1S5^NQKa1o4\\NTKd1l4YNXKf1h4WN\\Kh1o8OO10000O10000O10000O10000O10000O11O000000001O0000001O000000001O000O101N10001N100O2O000O2O0O2O000O2O0O101O0O101N101O0O101N3N1O1N3N1N2O1O2M2O1N3N1O1N2O2M2O1O2M2O1N2OlRc5"}, "label": "man on a double decker bus tour"}]} {"id": 565664, "image": "COCO_train2014_000000565664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby zebra nestled under an adult zebra\"."}], "task": "refering", "answer_template": "The \"a baby zebra nestled under an adult zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 220, 221, 222, 223, 224, 239, 240, 241, 242, 243, 244, 245, 246, 262, 263, 265, 266, 268, 269, 285, 288, 289, 291, 292, 311, 312, 314, 315, 333, 334, 335, 337, 338], "bbox": [0.40615625, 0.46208333333333335, 0.7442656249999999, 0.88025], "iscrowd": 0, "area": 21496.78654999999, "rle": {"size": [480, 640], "counts": "]Vj32l>3oNMPC7j<3nB1l<9lBKQ=;hBHW=R1M3M3N2M2N3N2M2N10O0100O010000O010O10O10O1000O0100O01000O101O1N2M3L3M400000O01000000O01000O10O1000O1000O01000000O01000O10O1WF]NU6c1jIcNQ6^1lIdNT6\\1kIfNT6[1jIfNU6\\1hIgNW6Z1gIhNX6Y1fIgN[6Y1cIiN\\6Y1bIhN^6Y1`IhN`6Y1]IiNc6X1XIkNi6V1QHoMAo0]8T1nGTN]Ol0e8[1`GPNCi0m8f2lF^MT9V4001N2O1N2O0O2O1N2O1N1101O1O001O1N10000000000000WO`FRL_9f3lFXLT9a3UG]Lk8a3ZG\\Lf8a3_G]La8a3cG]L]8a3gG]LY8`3kG_LU8_3oG_LP8`3SH_Lm7^3XH`Lh7^3\\H`Ld7^3P2M3N2N2N1N3N2N2M3N2N2M3N1O2M3N2O1O100000O1000O1000000000O012N1O1O2N1O1O2M2O2N1O1O2N1O2N1N2O2N1O1O2N1O2N1N2O2N1VG^LT6d3gI_LY6b3cI`L^6c3ZIbLg6d3kHcLV7c3[HdLg7b3kGeLV8a3\\GeLf8j41N2O1N2O1K5J6K5SO`FVLc9e3iFQLZ9i3QGoKR9l3T1J6K5N210\\NdDBh1cNd7[1UG3Q1eNk7i0eGa0NVO^81nGh0SOEQ9BnGi0_N5\\9XOYHU2f7lM]HQ2c7oM`Hn1_7SNeHi1[7WNjHd1U7]NPI^1P7aN[IU1d6lNjId0X6\\OUJMT64ZJSOQ6m0P41ed\\2"}, "label": "a baby zebra nestled under an adult zebra"}]} {"id": 565664, "image": "COCO_train2014_000000565664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra calf leaning into mother\"."}], "task": "refering", "answer_template": "The \"a zebra calf leaning into mother\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 220, 221, 222, 223, 224, 239, 240, 241, 242, 243, 244, 245, 246, 262, 263, 265, 266, 268, 269, 285, 288, 289, 291, 292, 311, 312, 314, 315, 333, 334, 335, 337, 338], "bbox": [0.40615625, 0.46208333333333335, 0.7442656249999999, 0.88025], "iscrowd": 0, "area": 21496.78654999999, "rle": {"size": [480, 640], "counts": "]Vj32l>3oNMPC7j<3nB1l<9lBKQ=;hBHW=R1M3M3N2M2N3N2M2N10O0100O010000O010O10O10O1000O0100O01000O101O1N2M3L3M400000O01000000O01000O10O1000O1000O01000000O01000O10O1WF]NU6c1jIcNQ6^1lIdNT6\\1kIfNT6[1jIfNU6\\1hIgNW6Z1gIhNX6Y1fIgN[6Y1cIiN\\6Y1bIhN^6Y1`IhN`6Y1]IiNc6X1XIkNi6V1QHoMAo0]8T1nGTN]Ol0e8[1`GPNCi0m8f2lF^MT9V4001N2O1N2O0O2O1N2O1N1101O1O001O1N10000000000000WO`FRL_9f3lFXLT9a3UG]Lk8a3ZG\\Lf8a3_G]La8a3cG]L]8a3gG]LY8`3kG_LU8_3oG_LP8`3SH_Lm7^3XH`Lh7^3\\H`Ld7^3P2M3N2N2N1N3N2N2M3N2N2M3N1O2M3N2O1O100000O1000O1000000000O012N1O1O2N1O1O2M2O2N1O1O2N1O2N1N2O2N1O1O2N1O2N1N2O2N1VG^LT6d3gI_LY6b3cI`L^6c3ZIbLg6d3kHcLV7c3[HdLg7b3kGeLV8a3\\GeLf8j41N2O1N2O1K5J6K5SO`FVLc9e3iFQLZ9i3QGoKR9l3T1J6K5N210\\NdDBh1cNd7[1UG3Q1eNk7i0eGa0NVO^81nGh0SOEQ9BnGi0_N5\\9XOYHU2f7lM]HQ2c7oM`Hn1_7SNeHi1[7WNjHd1U7]NPI^1P7aN[IU1d6lNjId0X6\\OUJMT64ZJSOQ6m0P41ed\\2"}, "label": "a zebra calf leaning into mother"}]} {"id": 565664, "image": "COCO_train2014_000000565664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"larger of the two zebras\"."}], "task": "refering", "answer_template": "The \"larger of the two zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 107, 108, 109, 110, 111, 112, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 189, 190, 191, 192, 193, 194, 195, 196, 197, 199, 200, 204, 213, 214, 215, 216, 217, 236, 237, 238, 239, 259, 260, 261, 262, 282, 283, 284, 285, 305, 306, 308, 309, 328, 329, 331, 332], "bbox": [0.236484375, 0.21733333333333332, 0.93075, 0.8614791666666667], "iscrowd": 0, "area": 46953.81530000001, "rle": {"size": [480, 640], "counts": "ooV21k>8I7I8G8I8J5MO0O1N3N1N3N1N3N1N3N1O2M2O2M2O0O01N101O0O2O0O2O001N5L6J5mFTOi3S1cK_OY4f0TKLg4;eJ7W52UJ=g54`I8^66jH6S7n3M4M2M100O1O101N10hMhHhLW7V3VI`Li6_3cIWL]6g3nIPLQ6n3\\JhKd5V4gJaKX5^4kJaKT5]4PKbKP5\\4SKcKl4\\4WKcKi4Z4[KeKd4Z4_KeK`4Z4cKeK]4X4hKfKW4Y4lKeKU4Y4nKfKQ4X4SLgKm3W4VLhKi3W4ZLhKe3V4_LiKa3U4bLjK]3U4h3N1N201O00]FPLX8o3eGVLY8j3dG\\LY8e3cG`L\\8_3aGfL]8Z3`GkL_8T3^GQM`8W3UGnLi8Z3lFlLR9b4O1O1O2N2N2N2N3M2N2N2N3M2N2N2N2N3M2M3N2NjNYHVKd7`4gHaKW7V4TIjKj6T4YImKe6b3cH`Kl0o0_6_3iH^Kk0U1Y6[3QI[Kj0Z1S6Y3^JhL`5W3bJjL\\5T3hJlLV5R3mJPMP5n2SKSMk4k2XKVMf4i2]KWMa4g2bKZM\\4d2gK[MY4c2jK]MU4a2oK]MQ4b2QL]Mn3b2UL^Mj3`2YL_Mg3_2]L_Mc3_2`LaM_3^2cLaM]3]2gLaMY3]2jLcMU3[2nLdMR3[2PMdMP3[2SMdMl2[2WMcMj2[2YMcMg2\\2]MaMc2]2aMbM^2]2fM`MZ2_2iM_MW2`2lM_MS2`2QN]Mo1b2TN\\Ml1b2XNZMj1e2ZNRMl1n2S5O100000001O0O10000000000O10000000000O10001O00000O10000000000O1000000000001N10000000000O10000000000O1000001O000O10000000000O10000000000O101O0000000O1000000000000O100000001O0O10000000000O10000000000O10001O0000000O10000000000O10000000001N100000O1L4L4L3M4L4L4O2O1N101O1O001O1O0O2O1O001O001O1N101O1O001O1O0O2O1O001O1O000O0100O1O1O100O1O00100O100O1O10O01O100O1O100O010O1O100O1O10O01O100O1O100O010O1O100O1O10O01O100O100O1O10O01O100O1O100O00100O100O1O10000001O1O1O001O1O1O001O1O1O001O1O10O01O1O1O001O1O1O001O1O1O001N2O0O2O1N2OaMgDW1X;gNmDX1R;dNTE[1j:bN\\E]1c:_NbEa1\\:]NiEa1W:[NoEd1o9YNWFf1h9WN]Fh1a9UNdFk1[9QNkFn1S9PNRGo1Q;O1N101O1O1O1N1N3M3N2M3M3M3M2O2M3M3M3M3N2M3M2N3M3N2M3MUld0"}, "label": "larger of the two zebras"}]} {"id": 565664, "image": "COCO_train2014_000000565664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mother and young baby zebra bonding\"."}], "task": "refering", "answer_template": "The \"a mother and young baby zebra bonding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 107, 108, 109, 110, 111, 112, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 189, 190, 191, 192, 193, 194, 195, 196, 197, 199, 200, 204, 213, 214, 215, 216, 217, 236, 237, 238, 239, 259, 260, 261, 262, 282, 283, 284, 285, 305, 306, 308, 309, 328, 329, 331, 332], "bbox": [0.236484375, 0.21733333333333332, 0.93075, 0.8614791666666667], "iscrowd": 0, "area": 46953.81530000001, "rle": {"size": [480, 640], "counts": "ooV21k>8I7I8G8I8J5MO0O1N3N1N3N1N3N1N3N1O2M2O2M2O0O01N101O0O2O0O2O001N5L6J5mFTOi3S1cK_OY4f0TKLg4;eJ7W52UJ=g54`I8^66jH6S7n3M4M2M100O1O101N10hMhHhLW7V3VI`Li6_3cIWL]6g3nIPLQ6n3\\JhKd5V4gJaKX5^4kJaKT5]4PKbKP5\\4SKcKl4\\4WKcKi4Z4[KeKd4Z4_KeK`4Z4cKeK]4X4hKfKW4Y4lKeKU4Y4nKfKQ4X4SLgKm3W4VLhKi3W4ZLhKe3V4_LiKa3U4bLjK]3U4h3N1N201O00]FPLX8o3eGVLY8j3dG\\LY8e3cG`L\\8_3aGfL]8Z3`GkL_8T3^GQM`8W3UGnLi8Z3lFlLR9b4O1O1O2N2N2N2N3M2N2N2N3M2N2N2N2N3M2M3N2NjNYHVKd7`4gHaKW7V4TIjKj6T4YImKe6b3cH`Kl0o0_6_3iH^Kk0U1Y6[3QI[Kj0Z1S6Y3^JhL`5W3bJjL\\5T3hJlLV5R3mJPMP5n2SKSMk4k2XKVMf4i2]KWMa4g2bKZM\\4d2gK[MY4c2jK]MU4a2oK]MQ4b2QL]Mn3b2UL^Mj3`2YL_Mg3_2]L_Mc3_2`LaM_3^2cLaM]3]2gLaMY3]2jLcMU3[2nLdMR3[2PMdMP3[2SMdMl2[2WMcMj2[2YMcMg2\\2]MaMc2]2aMbM^2]2fM`MZ2_2iM_MW2`2lM_MS2`2QN]Mo1b2TN\\Ml1b2XNZMj1e2ZNRMl1n2S5O100000001O0O10000000000O10000000000O10001O00000O10000000000O1000000000001N10000000000O10000000000O1000001O000O10000000000O10000000000O101O0000000O1000000000000O100000001O0O10000000000O10000000000O10001O0000000O10000000000O10000000001N100000O1L4L4L3M4L4L4O2O1N101O1O001O1O0O2O1O001O001O1N101O1O001O1O0O2O1O001O1O000O0100O1O1O100O1O00100O100O1O10O01O100O1O100O010O1O100O1O10O01O100O1O100O010O1O100O1O10O01O100O100O1O10O01O100O1O100O00100O100O1O10000001O1O1O001O1O1O001O1O1O001O1O10O01O1O1O001O1O1O001O1O1O001N2O0O2O1N2OaMgDW1X;gNmDX1R;dNTE[1j:bN\\E]1c:_NbEa1\\:]NiEa1W:[NoEd1o9YNWFf1h9WN]Fh1a9UNdFk1[9QNkFn1S9PNRGo1Q;O1N101O1O1O1N1N3M3N2M3M3M3M2O2M3M3M3M3N2M3M2N3M3N2M3MUld0"}, "label": "a mother and young baby zebra bonding"}]} {"id": 287140, "image": "COCO_train2014_000000287140.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the racing bike , having white and black colors\"."}], "task": "refering", "answer_template": "The \"the racing bike , having white and black colors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 96, 115, 116, 117, 118, 138, 139, 140, 141, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 278, 279, 281, 282, 283, 301, 304, 305, 306], "bbox": [0.0003125, 0.2819529411764706, 0.32492187499999997, 0.903270588235294], "iscrowd": 0, "area": 30426.04619999999, "rle": {"size": [425, 640], "counts": "l53h<=001O01O000001OcCEm;;RDIk;8TDJj;7UDJk;6UDJj;7VDIi;8WDGi;9XDGj;7VDHl;6TDJb0J\\:60V8:VGYO=?53n7?\\GoN`0a046k7g0nGVO56j7R2THRNh7P2VHTNf7n1WHVNf7l1XHXNd7j1ZHYNc7i1[H[Na7g1]H\\N`7f1^H^N^7d1`H^N_7h3N1N2O2N1O1N2O100O1O1O1O1O100O1O1O1O001O100O1TJcIk4]6SKdIm4\\6RKeIn4\\6oJfIR5Y6mJhIS5Y6kJhIU5X6iJjIW5W6gJjIY5V6fJkIZ5V6cJlI\\5h0VJW4=RK]5f0XJX49SK_55WJI3n45PKJGf5b0[JANN2X54PKg59WJ^O7X5JQKi58hJe4^OTKj57iJe4\\OTKk56lJe4XOTKn57kJd4UOVKP67kJc4TOVKR66lJc4QOWKS67lJn4T5RKnJk4T5UKmJi4S5XKmJg4T5XKnJf4S5ZKnJd4S5[KnJd4S5\\KnJa4T5_KmJ_4T5`KmJ_4T5aKmJ]4T5bKmJ]4T5cKmJZ4U5eKlJZ4V5eKkJY4V5fKkJY4W5dKkJ[4U5eKmJX4U5hKkJV4V5jKkJT4W5lKjJQ4W5oKjJo3W5RLiJl3Y5SLiJj3X5WLhJg3Z5XLgJf3Z5[LgJc3Z5\\LgJd3X5ZLmHJl1k3V5[LnJd3S5\\LmJd3R5\\LTK_3l4`LXK\\3h4eLYK[3f4dL[K]3d4cL\\K]3e4aL\\K`3d4^L]Kc3d4[L\\Kf3S7001O1O1O1O0O2N2O1N101N2O1N3M2O1N3N1N3M2O2N1O2M2O2N1O2M2O2N1O2M2O2N1O2M3N1O1O1O1O1O1O1O1O1O1O10O01O1O1O00000000iJfH_4Z7_KlH^4T7aKkH`4V7^KkHb4W7\\KiHd4Y7ZKfHg4_7SKbHm4m7000O2OcLoGb1Q8^NPHa1Q8]NRH`1o7`NRH_1n7aNTH]1n7`NUH^1o7^NRH`1R8]NPHa1Q8]NRHa1n7_NSH_1o7`NSH^1n7`NUH]1m7bNUH[1l7eNVHX1l7fNXHW1i7bN`HZ1c7cN`HZ1b7dNbHX1a7dNdHW1_7gNeHT1_7iNdHT1^7hNiHR1[7jNoHl0U7oNUIe0R:@Pob5"}, "label": "the racing bike , having white and black colors"}]} {"id": 287140, "image": "COCO_train2014_000000287140.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white colour motor cycle\"."}], "task": "refering", "answer_template": "The \"white colour motor cycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 96, 115, 116, 117, 118, 138, 139, 140, 141, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 278, 279, 281, 282, 283, 301, 304, 305, 306], "bbox": [0.0003125, 0.2819529411764706, 0.32492187499999997, 0.903270588235294], "iscrowd": 0, "area": 30426.04619999999, "rle": {"size": [425, 640], "counts": "l53h<=001O01O000001OcCEm;;RDIk;8TDJj;7UDJk;6UDJj;7VDIi;8WDGi;9XDGj;7VDHl;6TDJb0J\\:60V8:VGYO=?53n7?\\GoN`0a046k7g0nGVO56j7R2THRNh7P2VHTNf7n1WHVNf7l1XHXNd7j1ZHYNc7i1[H[Na7g1]H\\N`7f1^H^N^7d1`H^N_7h3N1N2O2N1O1N2O100O1O1O1O1O100O1O1O1O001O100O1TJcIk4]6SKdIm4\\6RKeIn4\\6oJfIR5Y6mJhIS5Y6kJhIU5X6iJjIW5W6gJjIY5V6fJkIZ5V6cJlI\\5h0VJW4=RK]5f0XJX49SK_55WJI3n45PKJGf5b0[JANN2X54PKg59WJ^O7X5JQKi58hJe4^OTKj57iJe4\\OTKk56lJe4XOTKn57kJd4UOVKP67kJc4TOVKR66lJc4QOWKS67lJn4T5RKnJk4T5UKmJi4S5XKmJg4T5XKnJf4S5ZKnJd4S5[KnJd4S5\\KnJa4T5_KmJ_4T5`KmJ_4T5aKmJ]4T5bKmJ]4T5cKmJZ4U5eKlJZ4V5eKkJY4V5fKkJY4W5dKkJ[4U5eKmJX4U5hKkJV4V5jKkJT4W5lKjJQ4W5oKjJo3W5RLiJl3Y5SLiJj3X5WLhJg3Z5XLgJf3Z5[LgJc3Z5\\LgJd3X5ZLmHJl1k3V5[LnJd3S5\\LmJd3R5\\LTK_3l4`LXK\\3h4eLYK[3f4dL[K]3d4cL\\K]3e4aL\\K`3d4^L]Kc3d4[L\\Kf3S7001O1O1O1O0O2N2O1N101N2O1N3M2O1N3N1N3M2O2N1O2M2O2N1O2M2O2N1O2M2O2N1O2M3N1O1O1O1O1O1O1O1O1O1O10O01O1O1O00000000iJfH_4Z7_KlH^4T7aKkH`4V7^KkHb4W7\\KiHd4Y7ZKfHg4_7SKbHm4m7000O2OcLoGb1Q8^NPHa1Q8]NRH`1o7`NRH_1n7aNTH]1n7`NUH^1o7^NRH`1R8]NPHa1Q8]NRHa1n7_NSH_1o7`NSH^1n7`NUH]1m7bNUH[1l7eNVHX1l7fNXHW1i7bN`HZ1c7cN`HZ1b7dNbHX1a7dNdHW1_7gNeHT1_7iNdHT1^7hNiHR1[7jNoHl0U7oNUIe0R:@Pob5"}, "label": "white colour motor cycle"}]} {"id": 287140, "image": "COCO_train2014_000000287140.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red dirt bike\"."}], "task": "refering", "answer_template": "The \"a red dirt bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 122, 123, 124, 125, 126, 129, 130, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 200, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 261, 262, 263, 264, 265, 267, 268, 269, 270, 287, 291, 292], "bbox": [0.33801562500000004, 0.2939294117647059, 0.756796875, 0.8344705882352941], "iscrowd": 0, "area": 33963.15665000001, "rle": {"size": [425, 640], "counts": "Sii24U=0YC3S[6^OiIb0W6POXJo0h5ROXJm0h5TOWJk0j5VOVJi0j5XOVJg0i5[OWJd0i5]OVJc0j5]OXJa0h5@YJ in this image.", "objects": [{"patches": [101, 102, 103, 104, 122, 123, 124, 125, 126, 129, 130, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 200, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 261, 262, 263, 264, 265, 267, 268, 269, 270, 287, 291, 292], "bbox": [0.33801562500000004, 0.2939294117647059, 0.756796875, 0.8344705882352941], "iscrowd": 0, "area": 33963.15665000001, "rle": {"size": [425, 640], "counts": "Sii24U=0YC3S[6^OiIb0W6POXJo0h5ROXJm0h5TOWJk0j5VOVJi0j5XOVJg0i5[OWJd0i5]OVJc0j5]OXJa0h5@YJ in this image.", "objects": [{"patches": [77, 78, 79, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 145, 146, 147, 169, 170], "bbox": [0.32553125, 0.19905882352941176, 0.50721875, 0.5122117647058824], "iscrowd": 0, "area": 6114.69895, "rle": {"size": [425, 640], "counts": "P_f2:8IZ< in this image.", "objects": [{"patches": [77, 78, 79, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 145, 146, 147, 169, 170], "bbox": [0.32553125, 0.19905882352941176, 0.50721875, 0.5122117647058824], "iscrowd": 0, "area": 6114.69895, "rle": {"size": [425, 640], "counts": "P_f2:8IZ< in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183], "bbox": [0.747109375, 0.003949579831932773, 1.0, 0.9791176470588235], "iscrowd": 0, "area": 35535.5768, "rle": {"size": [238, 640], "counts": "`S_36W7f0ZOf0ZOf0ZOf0ZOe0ZOg0ZOf0ZOf0ZOf0ZOf0[O6J0000000000000000000000000001O000000000000000000000000000O100000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000O10000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O=B>C=C=C=C>B\\M"}, "label": "a white shirt with gray tie and suit coat highlighted with a white flower in the lapel"}]} {"id": 205223, "image": "COCO_train2014_000000205223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the suit on the far right\"."}], "task": "refering", "answer_template": "The \"the suit on the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183], "bbox": [0.747109375, 0.003949579831932773, 1.0, 0.9791176470588235], "iscrowd": 0, "area": 35535.5768, "rle": {"size": [238, 640], "counts": "`S_36W7f0ZOf0ZOf0ZOf0ZOe0ZOg0ZOf0ZOf0ZOf0ZOf0[O6J0000000000000000000000000001O000000000000000000000000000O100000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000O10000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O=B>C=C=C=C>B\\M"}, "label": "the suit on the far right"}]} {"id": 205223, "image": "COCO_train2014_000000205223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"second tie from right\"."}], "task": "refering", "answer_template": "The \"second tie from right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 58, 59, 81, 82, 104, 105, 127, 128, 129, 150, 151, 173, 174], "bbox": [0.53065625, 0.13932773109243696, 0.615890625, 0.9865126050420168], "iscrowd": 0, "area": 7076.0271999999995, "rle": {"size": [238, 640], "counts": "TR_22m2f0i0\\OeN_1o0eNgNl1o0VNlNW2i0lMRO]2h0eMROc2j0_MQOh2j0[MQOk2l0WMoNP3o0QMkNU3S1oLfNW3X1kLbNZ3\\1jL]N\\3`1\\1M2N3M2N3M2N3M0001O001O001O1O0WJ^Of4c0XK@f4a0WKCg4>VKFh4:VKKg4a1M4L4L3M4L4L3M4H8G8H9G9G8H9G9G8H9G8H9G9F9HjPi1"}, "label": "second tie from right"}]} {"id": 205223, "image": "COCO_train2014_000000205223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black tie with a white floral pinned to his suit\"."}], "task": "refering", "answer_template": "The \"black tie with a white floral pinned to his suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166], "bbox": [0.0, 0.002226890756302521, 0.24653125, 0.9865126050420168], "iscrowd": 0, "area": 34454.1651, "rle": {"size": [238, 640], "counts": "P1V3U4X3hL3M3N2\\O^I9e6GaI0b60>00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000007I1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O000RM[R`3"}, "label": "black tie with a white floral pinned to his suit"}]} {"id": 205223, "image": "COCO_train2014_000000205223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tux on the left with the white flower\"."}], "task": "refering", "answer_template": "The \"tux on the left with the white flower\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166], "bbox": [0.0, 0.002226890756302521, 0.24653125, 0.9865126050420168], "iscrowd": 0, "area": 34454.1651, "rle": {"size": [238, 640], "counts": "P1V3U4X3hL3M3N2\\O^I9e6GaI0b60>00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000007I1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O000RM[R`3"}, "label": "tux on the left with the white flower"}]} {"id": 205223, "image": "COCO_train2014_000000205223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second suit\"."}], "task": "refering", "answer_template": "The \"the second suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 28, 29, 30, 31, 32, 33, 34, 51, 52, 53, 54, 55, 56, 57, 74, 75, 76, 77, 78, 79, 80, 97, 98, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172], "bbox": [0.25017187500000004, 0.002605042016806723, 0.495515625, 0.9818067226890757], "iscrowd": 0, "area": 35454.47194999999, "rle": {"size": [238, 640], "counts": "_\\U1k0i3j2VMj2G9O1@XI8j6HWI6j6JWI3k6MWIOk61VIMk638000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O1O1O1O001O1O1O1O1O001O1O1O1O001O001O001O00001O001O00001O001O001O00001O001O000gLkT[2"}, "label": "the second suit"}]} {"id": 205223, "image": "COCO_train2014_000000205223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white silver tie is posed to the left of the logo\"."}], "task": "refering", "answer_template": "The \"a white silver tie is posed to the left of the logo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 34, 35, 36, 37, 38, 39, 40, 57, 58, 59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178], "bbox": [0.49743750000000003, 0.0018067226890756303, 0.7485781250000001, 0.9836134453781512], "iscrowd": 0, "area": 36458.50964999999, "rle": {"size": [238, 640], "counts": "TmY25Y7^1bN_1aN^1bN_1aNZ1fN0000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000001O001O001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O001O00001O00m\\U1"}, "label": "a white silver tie is posed to the left of the logo"}]} {"id": 205223, "image": "COCO_train2014_000000205223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a tuxedo standing 2nd from the right of four people\"."}], "task": "refering", "answer_template": "The \"a man wearing a tuxedo standing 2nd from the right of four people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 34, 35, 36, 37, 38, 39, 40, 57, 58, 59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178], "bbox": [0.49743750000000003, 0.0018067226890756303, 0.7485781250000001, 0.9836134453781512], "iscrowd": 0, "area": 36458.50964999999, "rle": {"size": [238, 640], "counts": "TmY25Y7^1bN_1aN^1bN_1aNZ1fN0000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000001O001O001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O001O00001O00m\\U1"}, "label": "a man wearing a tuxedo standing 2nd from the right of four people"}]} {"id": 205223, "image": "COCO_train2014_000000205223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the fourth image of a grey tie in a series of four\"."}], "task": "refering", "answer_template": "The \"the fourth image of a grey tie in a series of four\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 41, 42, 64, 65, 87, 88, 89, 110, 111, 112, 133, 134, 135, 156, 157, 158, 179, 180, 181], "bbox": [0.779484375, 0.06063025210084033, 0.8948281250000001, 1.0], "iscrowd": 0, "area": 9442.751550000003, "rle": {"size": [238, 640], "counts": "WPd32Z72O2M3M3N2M3O001O1O1O001SML^N5e0o0SORO?]1_OcN5k1IVNKW23jMH\\26dMFb28_MCg2;ZM@l2>UM]OQ3a0PMZOV3d0kLWO[3g0gLSO^3l0cLoNc3o0^LmNg3Q1[LiNk3U1Y1M3N1O1O2N1O1O2N1O1O2N1O1N3L20lIOk41iJ>T5CgJc0W5^OgJe0W5[OhJi0V5WOhJl0V5UOhJo0U5i0L3M3M4L3M3M4L3M5K5K5K5K5K5K5K6J5K5K5K7I7I8\\Oc0YOh0WOl\\?"}, "label": "the fourth image of a grey tie in a series of four"}]} {"id": 205223, "image": "COCO_train2014_000000205223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the mans gray tie on the right side of the picture\"."}], "task": "refering", "answer_template": "The \"the mans gray tie on the right side of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 41, 42, 64, 65, 87, 88, 89, 110, 111, 112, 133, 134, 135, 156, 157, 158, 179, 180, 181], "bbox": [0.779484375, 0.06063025210084033, 0.8948281250000001, 1.0], "iscrowd": 0, "area": 9442.751550000003, "rle": {"size": [238, 640], "counts": "WPd32Z72O2M3M3N2M3O001O1O1O001SML^N5e0o0SORO?]1_OcN5k1IVNKW23jMH\\26dMFb28_MCg2;ZM@l2>UM]OQ3a0PMZOV3d0kLWO[3g0gLSO^3l0cLoNc3o0^LmNg3Q1[LiNk3U1Y1M3N1O1O2N1O1O2N1O1O2N1O1N3L20lIOk41iJ>T5CgJc0W5^OgJe0W5[OhJi0V5WOhJl0V5UOhJo0U5i0L3M3M4L3M3M4L3M5K5K5K5K5K5K5K6J5K5K5K7I7I8\\Oc0YOh0WOl\\?"}, "label": "the mans gray tie on the right side of the picture"}]} {"id": 205223, "image": "COCO_train2014_000000205223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark grey tie to the left of three other ties\"."}], "task": "refering", "answer_template": "The \"a dark grey tie to the left of three other ties\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 47, 70, 71, 92, 93, 94, 115, 116, 117, 139, 140, 162], "bbox": [0.032546874999999996, 0.18752100840336136, 0.105671875, 1.0], "iscrowd": 0, "area": 5760.914249999999, "rle": {"size": [238, 640], "counts": "ZP5g0b6;F:E;F:F7H7K6K5K6I7J6gNgLRNa3e1hLTN^3f1iLRN^3j1gLmMa3n1U1K4L4K4M2N0O10001N10000O1000hJdN^4\\1aKjNX4X1fKPOn3W1PLmNh3Y1ULlNd3X1ZLlN_3Z1_LiN[3\\1bLiNW3[1hLhNQ3^1lLfNn2_1PMeNg1a2WNcMa1Z4H8G:G;E in this image.", "objects": [{"patches": [24, 25, 47, 70, 71, 92, 93, 94, 115, 116, 117, 139, 140, 162], "bbox": [0.032546874999999996, 0.18752100840336136, 0.105671875, 1.0], "iscrowd": 0, "area": 5760.914249999999, "rle": {"size": [238, 640], "counts": "ZP5g0b6;F:E;F:F7H7K6K5K6I7J6gNgLRNa3e1hLTN^3f1iLRN^3j1gLmMa3n1U1K4L4K4M2N0O10001N10000O1000hJdN^4\\1aKjNX4X1fKPOn3W1PLmNh3Y1ULlNd3X1ZLlN_3Z1_LiN[3\\1bLiNW3[1hLhNQ3^1lLfNn2_1PMeNg1a2WNcMa1Z4H8G:G;E in this image.", "objects": [{"patches": [30, 31, 53, 75, 76, 77, 98, 99, 100, 121, 122, 123, 144, 145, 167, 168], "bbox": [0.272265625, 0.17327731092436977, 0.358234375, 0.9962184873949581], "iscrowd": 0, "area": 7283.865349999999, "rle": {"size": [238, 640], "counts": "_cX17e6f0ZOg0XOg0I8I6J6J7I6J6J7I6J7I6J4L2N2N2O1lNbKVO`4g0lKnNV4R1SLcNo3\\1R10000O10000O10000O1000O01000O0L6VK`N_3f1_L]NZ3h1eL[NT3k1iLYNQ3l1mLWNl2o1RMTNg2Q2WMSNb2S2\\MPN]2V2`MnMY2X2eMkMU2Y2jMjMo1\\2nMhMk1^2SNeMR1U3kNh2O2N2MV_o2"}, "label": "the tie at the second from the left"}]} {"id": 205223, "image": "COCO_train2014_000000205223.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second tie from the left in the collage\"."}], "task": "refering", "answer_template": "The \"the second tie from the left in the collage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 53, 75, 76, 77, 98, 99, 100, 121, 122, 123, 144, 145, 167, 168], "bbox": [0.272265625, 0.17327731092436977, 0.358234375, 0.9962184873949581], "iscrowd": 0, "area": 7283.865349999999, "rle": {"size": [238, 640], "counts": "_cX17e6f0ZOg0XOg0I8I6J6J7I6J6J7I6J7I6J4L2N2N2O1lNbKVO`4g0lKnNV4R1SLcNo3\\1R10000O10000O10000O1000O01000O0L6VK`N_3f1_L]NZ3h1eL[NT3k1iLYNQ3l1mLWNl2o1RMTNg2Q2WMSNb2S2\\MPN]2V2`MnMY2X2eMkMU2Y2jMjMo1\\2nMhMk1^2SNeMR1U3kNh2O2N2MV_o2"}, "label": "the second tie from the left in the collage"}]} {"id": 213419, "image": "COCO_train2014_000000213419.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boat with says marleni - n on it\"."}], "task": "refering", "answer_template": "The \"the boat with says marleni - n on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 302, 303], "bbox": [0.00596875, 0.39170960187353626, 0.44479687500000004, 0.8973067915690865], "iscrowd": 0, "area": 44259.75765000001, "rle": {"size": [427, 640], "counts": "dn1e0V in this image.", "objects": [{"patches": [118, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 302, 303], "bbox": [0.00596875, 0.39170960187353626, 0.44479687500000004, 0.8973067915690865], "iscrowd": 0, "area": 44259.75765000001, "rle": {"size": [427, 640], "counts": "dn1e0V in this image.", "objects": [{"patches": [155, 156, 157, 158, 159, 174, 175, 176, 177, 178, 179, 180, 181, 182, 196, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249], "bbox": [0.545484375, 0.42463700234192037, 0.9236093750000001, 0.7274707259953161], "iscrowd": 0, "area": 19081.130500000003, "rle": {"size": [427, 640], "counts": "Xha41Z=3M3M3L4M2N3M3M3L4M2N3M3M3L3N3M3M3M2M4M2N2M3N2N1O0O2O000O2O00001N1000001N10001N10001O0O10001N10001N10000O1000000O10000O10000O10000O1000000O100000000O100000001N100000000O100000000O10000000000O100000000O2O000O1000000O10000O10001O0O10000O1000000O10001N1000000O10000O10001O0O100000000O1000000O2O0000000O100000000O2O00O10O100000O10O1000O1000O1000mM\\F>e9B\\F=c9D]F^G_Ob8`0bG]O^8c0eGYO[8h0gGVOY8i0jGUOV8k0mGROS8n0oGoNR8Q1PHmNo7S1THkNl7U1VHiNj7W1U2O10O10000000O010000000000O10O100000O100000O1000O11O00000O100O2O0O100O1H9N1O1O101N3L4L4M3K5K5KiWd0"}, "label": "a boat with a person in it"}]} {"id": 213419, "image": "COCO_train2014_000000213419.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boat on the right with the person in it\"."}], "task": "refering", "answer_template": "The \"the boat on the right with the person in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 158, 159, 174, 175, 176, 177, 178, 179, 180, 181, 182, 196, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249], "bbox": [0.545484375, 0.42463700234192037, 0.9236093750000001, 0.7274707259953161], "iscrowd": 0, "area": 19081.130500000003, "rle": {"size": [427, 640], "counts": "Xha41Z=3M3M3L4M2N3M3M3L4M2N3M3M3L3N3M3M3M2M4M2N2M3N2N1O0O2O000O2O00001N1000001N10001N10001O0O10001N10001N10000O1000000O10000O10000O10000O1000000O100000000O100000001N100000000O100000000O10000000000O100000000O2O000O1000000O10000O10001O0O10000O1000000O10001N1000000O10000O10001O0O100000000O1000000O2O0000000O100000000O2O00O10O100000O10O1000O1000O1000mM\\F>e9B\\F=c9D]F^G_Ob8`0bG]O^8c0eGYO[8h0gGVOY8i0jGUOV8k0mGROS8n0oGoNR8Q1PHmNo7S1THkNl7U1VHiNj7W1U2O10O10000000O010000000000O10O100000O100000O1000O11O00000O100O2O0O100O1H9N1O1O101N3L4L4M3K5K5KiWd0"}, "label": "the boat on the right with the person in it"}]} {"id": 74156, "image": "COCO_train2014_000000074156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man wearing a black shirt and jeans\"."}], "task": "refering", "answer_template": "The \"the man wearing a black shirt and jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 145, 146, 147, 167, 168, 169, 172, 190, 191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 219, 220, 237, 238, 239, 240, 261], "bbox": [0.28532812500000004, 0.3784633569739953, 0.5905, 0.7455791962174941], "iscrowd": 0, "area": 12926.978250000002, "rle": {"size": [423, 640], "counts": "nj[26n<7H7I7I7XOPOoDW1m:d0L4M3L3N3L4M3N2M2O2M3N1N101N101N101N101N101M2N3M2N3M2N3N1N3N1N3M2O2M3FVLVGi3i8 in this image.", "objects": [{"patches": [123, 145, 146, 147, 167, 168, 169, 172, 190, 191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 219, 220, 237, 238, 239, 240, 261], "bbox": [0.28532812500000004, 0.3784633569739953, 0.5905, 0.7455791962174941], "iscrowd": 0, "area": 12926.978250000002, "rle": {"size": [423, 640], "counts": "nj[26n<7H7I7I7XOPOoDW1m:d0L4M3L3N3L4M3N2M2O2M3N1N101N101N101N101N101M2N3M2N3M2N3N1N3N1N3M2O2M3FVLVGi3i8 in this image.", "objects": [{"patches": [151, 152, 174, 175, 176, 197, 198, 199, 200, 220, 221, 222, 223, 239, 240, 241, 242, 243, 244, 245, 246, 262, 263, 264, 265, 266, 268, 269, 286, 287, 291, 292, 314, 315], "bbox": [0.40284375, 0.4160283687943262, 0.738671875, 0.8831914893617022], "iscrowd": 0, "area": 14534.710300000004, "rle": {"size": [423, 640], "counts": "jkZ32S=4L3M4L3N3N1O1O2N1O2M2O2O0O101N101O0O010O0100O010O010O010001N100O1O10O010O01O10O01O010O10O00001O00000000000O010O100O00100O10O0100O10O0100O0100O01000O10O0100O10O10O100O010O1000O0100O10O0100O10O1OG9001O001O10O01O1O1O1O1O1O1O1O1O1O1O01JPOUDo0j;TOUDl0j;VOUDi0b:b0]E^OZ9WOeG]2[8nMYGT2f8oMVGR2j8RNRGo1m8TNoFm1Q9VNkFl1T9S101N2O10000000000000000001O001O1O2N2N2N2N2N2N2N2N2gLYFn2P:0000000000000000000000O10000O2O1N2O1N2O1N2M3M4M2100kEhMX9Y2ZFZN9ZOd8V4J6K0OO2O2M3N2M4M1M101O1O1O001N2O1O001O1O1N1M5I7J6L4K4M4K5J7oMeEl0a;H8H8G9BQfT2"}, "label": "a man with a multicolored shirt and brown boots is sitting on the ground with something white behind him"}]} {"id": 74156, "image": "COCO_train2014_000000074156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with two other men in the back of a truck with hiking gear\"."}], "task": "refering", "answer_template": "The \"a man with two other men in the back of a truck with hiking gear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 174, 175, 176, 197, 198, 199, 200, 220, 221, 222, 223, 239, 240, 241, 242, 243, 244, 245, 246, 262, 263, 264, 265, 266, 268, 269, 286, 287, 291, 292, 314, 315], "bbox": [0.40284375, 0.4160283687943262, 0.738671875, 0.8831914893617022], "iscrowd": 0, "area": 14534.710300000004, "rle": {"size": [423, 640], "counts": "jkZ32S=4L3M4L3N3N1O1O2N1O2M2O2O0O101N101O0O010O0100O010O010O010001N100O1O10O010O01O10O01O010O10O00001O00000000000O010O100O00100O10O0100O10O0100O0100O01000O10O0100O10O10O100O010O1000O0100O10O0100O10O1OG9001O001O10O01O1O1O1O1O1O1O1O1O1O1O01JPOUDo0j;TOUDl0j;VOUDi0b:b0]E^OZ9WOeG]2[8nMYGT2f8oMVGR2j8RNRGo1m8TNoFm1Q9VNkFl1T9S101N2O10000000000000000001O001O1O2N2N2N2N2N2N2N2N2gLYFn2P:0000000000000000000000O10000O2O1N2O1N2O1N2M3M4M2100kEhMX9Y2ZFZN9ZOd8V4J6K0OO2O2M3N2M4M1M101O1O1O001N2O1O001O1O1N1M5I7J6L4K4M4K5J7oMeEl0a;H8H8G9BQfT2"}, "label": "a man with two other men in the back of a truck with hiking gear"}]} {"id": 74156, "image": "COCO_train2014_000000074156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"three men with camping gear in the back of a truck\"."}], "task": "refering", "answer_template": "The \"three men with camping gear in the back of a truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 96, 108, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.26952718676122933, 1.0, 0.9856737588652482], "iscrowd": 0, "area": 159801.36035, "rle": {"size": [423, 640], "counts": "b3_9h301O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O00000000J6G900000000000000O100001O3M3M3M3M3M1O1O001O1O001O001O:F:F;E:F;E5K1O000000000000000000001O000000000000000000001O000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000_Oa0jNV1000000O1000000O1000000O1000000O1K5H8M30000000000001O1O1O1O1O2N1O1O1O1O2N000000O100000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O1l2TMS4mKPE"}, "label": "three men with camping gear in the back of a truck"}]} {"id": 74156, "image": "COCO_train2014_000000074156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back side of the truck\"."}], "task": "refering", "answer_template": "The \"the back side of the truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 96, 108, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.26952718676122933, 1.0, 0.9856737588652482], "iscrowd": 0, "area": 159801.36035, "rle": {"size": [423, 640], "counts": "b3_9h301O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O00000000J6G900000000000000O100001O3M3M3M3M3M1O1O001O1O001O001O:F:F;E:F;E5K1O000000000000000000001O000000000000000000001O000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000_Oa0jNV1000000O1000000O1000000O1000000O1K5H8M30000000000001O1O1O1O1O2N1O1O1O1O2N000000O100000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O1l2TMS4mKPE"}, "label": "the back side of the truck"}]} {"id": 483766, "image": "COCO_train2014_000000483766.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy with a beige shirt\"."}], "task": "refering", "answer_template": "The \"a boy with a beige shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 112, 113, 129, 130, 146, 147, 148, 149, 163, 164, 165, 166, 167, 180, 181, 182, 183, 184, 197, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 247, 248, 249, 250, 251, 265, 266, 267, 268, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 317, 318, 319, 334, 335, 336, 337, 351, 353, 354], "bbox": [0.5756875, 0.25525000000000003, 0.9157291666666666, 0.90178125], "iscrowd": 0, "area": 36927.18970000001, "rle": {"size": [640, 480], "counts": "Rl\\5b0Rc03^4`9ZKWFe0JT4n9\\KSFi6l9YIRFg6m9\\IPFf6o9\\IoEe6P:]InEd6Q:^ImEb6S:f1O1O001O000^HWFm4n9nJUFQ5o9iJSFW5Q:bJSF]5Q:]JRFa5S:XJPFh5U:QJnEn5V:kInET6V:fIlEZ6n;O1O1O001OO1O1O002N3M2N3L4M2N3M3M2N4I7REjGi9[8TFgGg9`8TFdGg9c8SF`Gh9i8RFZGi9n8PFUGk9c9aMVFTI24Y:i6m11N2N2N2O1N2N2N2O1N1O2N2O0O2N2N2N3N1N2N2N2O1N2\\EaGc9a8hETHW:l8N2N2N2O1N2N2N2O1N2O1O1ORHZFb5d9\\JfF\\5[9bJnFV5S9hJWGP5i8nJ`Gj4a8SKjGd4W8]KoG\\4Q8kKoGm3R8SLVHl0kNhNP9;]He0RObNa8g0gH>kNhN_8i0nH7gNlN\\8k0UI3bNoNY8m0\\IM_NROW8o0`II\\NUOU8P1gIDXNYOQ8R1mI_OUN\\OP8T1nI\\ORN@Q8U1PJVOPNEQ8T1SJSOkMIS8T1VJnNgMNT8T1XJjNeM2T8T1ZJeNbM7U8U1\\J_N_M in this image.", "objects": [{"patches": [96, 112, 113, 129, 130, 146, 147, 148, 149, 163, 164, 165, 166, 167, 180, 181, 182, 183, 184, 197, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 247, 248, 249, 250, 251, 265, 266, 267, 268, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 317, 318, 319, 334, 335, 336, 337, 351, 353, 354], "bbox": [0.5756875, 0.25525000000000003, 0.9157291666666666, 0.90178125], "iscrowd": 0, "area": 36927.18970000001, "rle": {"size": [640, 480], "counts": "Rl\\5b0Rc03^4`9ZKWFe0JT4n9\\KSFi6l9YIRFg6m9\\IPFf6o9\\IoEe6P:]InEd6Q:^ImEb6S:f1O1O001O000^HWFm4n9nJUFQ5o9iJSFW5Q:bJSF]5Q:]JRFa5S:XJPFh5U:QJnEn5V:kInET6V:fIlEZ6n;O1O1O001OO1O1O002N3M2N3L4M2N3M3M2N4I7REjGi9[8TFgGg9`8TFdGg9c8SF`Gh9i8RFZGi9n8PFUGk9c9aMVFTI24Y:i6m11N2N2N2O1N2N2N2O1N1O2N2O0O2N2N2N3N1N2N2N2O1N2\\EaGc9a8hETHW:l8N2N2N2O1N2N2N2O1N2O1O1ORHZFb5d9\\JfF\\5[9bJnFV5S9hJWGP5i8nJ`Gj4a8SKjGd4W8]KoG\\4Q8kKoGm3R8SLVHl0kNhNP9;]He0RObNa8g0gH>kNhN_8i0nH7gNlN\\8k0UI3bNoNY8m0\\IM_NROW8o0`II\\NUOU8P1gIDXNYOQ8R1mI_OUN\\OP8T1nI\\ORN@Q8U1PJVOPNEQ8T1SJSOkMIS8T1VJnNgMNT8T1XJjNeM2T8T1ZJeNbM7U8U1\\J_N_M in this image.", "objects": [{"patches": [58, 59, 75, 76, 92, 93, 102, 103, 104, 105, 106, 107, 108, 109, 110, 121, 122, 123, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 146, 158, 159, 160, 161, 162, 163, 176, 177, 178, 179, 180, 193, 194, 195, 196, 197, 210, 211, 212, 213, 227, 228, 229, 230, 244, 245, 246, 247, 261, 262, 263, 264, 278, 279, 280, 281, 296, 297, 313, 314, 315, 330, 331, 332, 348], "bbox": [0.03795833333333333, 0.16575, 0.6177291666666667, 0.8801562500000001], "iscrowd": 0, "area": 41084.68400000001, "rle": {"size": [640, 480], "counts": "d];2nc01O00001O001O00001O00001^O2P]ONPc05m\\OLRc07k\\OIVc0`001O01O101N100O2N100O1OO2O001N101O00001O001O00001O010O001O00001O0010O0001O001O001O01N1O100O1O1001O2O1N1O2N1O1O00000010O000001O00001O0000001N10001O0000001O0000001O00001O01O0001O0000001O00001O0000001O000010O000001O0000001O000O2O000O101O0O101O0O10001N10000O2O003L3N3M3V^OlNa`0]2M4L3L5L3M3M4M2M3M1O1O010O1O1O1O100O0XAfKj=[4RBWL_=j3]BeLX=[3dBSMQ=n2kBaMi<`2SCnMb in this image.", "objects": [{"patches": [58, 59, 75, 76, 92, 93, 102, 103, 104, 105, 106, 107, 108, 109, 110, 121, 122, 123, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 146, 158, 159, 160, 161, 162, 163, 176, 177, 178, 179, 180, 193, 194, 195, 196, 197, 210, 211, 212, 213, 227, 228, 229, 230, 244, 245, 246, 247, 261, 262, 263, 264, 278, 279, 280, 281, 296, 297, 313, 314, 315, 330, 331, 332, 348], "bbox": [0.03795833333333333, 0.16575, 0.6177291666666667, 0.8801562500000001], "iscrowd": 0, "area": 41084.68400000001, "rle": {"size": [640, 480], "counts": "d];2nc01O00001O001O00001O00001^O2P]ONPc05m\\OLRc07k\\OIVc0`001O01O101N100O2N100O1OO2O001N101O00001O001O00001O010O001O00001O0010O0001O001O001O01N1O100O1O1001O2O1N1O2N1O1O00000010O000001O00001O0000001N10001O0000001O0000001O00001O01O0001O0000001O00001O0000001O000010O000001O0000001O000O2O000O101O0O101O0O10001N10000O2O003L3N3M3V^OlNa`0]2M4L3L5L3M3M4M2M3M1O1O010O1O1O1O100O0XAfKj=[4RBWL_=j3]BeLX=[3dBSMQ=n2kBaMi<`2SCnMb in this image.", "objects": [{"patches": [91, 113, 114, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 275, 298, 321, 343, 344, 366, 367, 389, 390], "bbox": [0.886515625, 0.15954166666666666, 1.0, 0.9887708333333333], "iscrowd": 0, "area": 12341.702849999998, "rle": {"size": [480, 640], "counts": "hoY82l>>B>B`M@b2=aMAa2;cMD^28fMF\\27hMGY25kMIX23kMLV20nMNT2OoM0R2LRN2P2KSN3o1IUN6l1FXN8j1EYN:h1B\\Nd1_O_N`0b1^O`N`0b1]OaNb0`1\\ObNb0`1[OcNd0^1ZOdNd0^1YOeNf0\\1XOgNe0[1XOhNg0Z1UObNP1`1kNgLR3nMT2:]M"}, "label": "person in black standing off to the side"}]} {"id": 229825, "image": "COCO_train2014_000000229825.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person ' s arm wearing a black shirt\"."}], "task": "refering", "answer_template": "The \"a person ' s arm wearing a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [91, 113, 114, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 275, 298, 321, 343, 344, 366, 367, 389, 390], "bbox": [0.886515625, 0.15954166666666666, 1.0, 0.9887708333333333], "iscrowd": 0, "area": 12341.702849999998, "rle": {"size": [480, 640], "counts": "hoY82l>>B>B`M@b2=aMAa2;cMD^28fMF\\27hMGY25kMIX23kMLV20nMNT2OoM0R2LRN2P2KSN3o1IUN6l1FXN8j1EYN:h1B\\Nd1_O_N`0b1^O`N`0b1]OaNb0`1\\ObNb0`1[OcNd0^1ZOdNd0^1YOeNf0\\1XOgNe0[1XOhNg0Z1UObNP1`1kNgLR3nMT2:]M"}, "label": "a person ' s arm wearing a black shirt"}]} {"id": 229825, "image": "COCO_train2014_000000229825.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a blue button up shirt\"."}], "task": "refering", "answer_template": "The \"a woman in a blue button up shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 32, 33, 34, 35, 36, 37, 38, 39, 55, 56, 57, 58, 59, 60, 61, 62, 78, 79, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 198], "bbox": [0.41121875, 0.0008541666666666666, 0.73671875, 0.5042083333333334], "iscrowd": 0, "area": 35041.70725, "rle": {"size": [480, 640], "counts": "oYk3^1b=3M3L4M3M3L3N3L4M3M3L3N2M10001N1iEdMj7\\2QHcNT7^1gHhNX7X1cHmN]7S1^HROa7P1YHVOf7j0UH[Ok7e0[G6c8T300O2O001N101N10001N101N10001N101O0I7H90O0000001O0000001O0000000000000000000000000000000001O000000000000000000000000000000000000O100O100O100O10000O100O100O100O11O00001O00001O00001O00001O00001O0O101O0000001O00001O1O5K5J5L0000000000O1000000000OO2J6J6I7J6J6J6K5L4L4L4L4ZOlF\\KW9b4f0K4L5K5K4C>B>B>C in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 32, 33, 34, 35, 36, 37, 38, 39, 55, 56, 57, 58, 59, 60, 61, 62, 78, 79, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 198], "bbox": [0.41121875, 0.0008541666666666666, 0.73671875, 0.5042083333333334], "iscrowd": 0, "area": 35041.70725, "rle": {"size": [480, 640], "counts": "oYk3^1b=3M3L4M3M3L3N3L4M3M3L3N2M10001N1iEdMj7\\2QHcNT7^1gHhNX7X1cHmN]7S1^HROa7P1YHVOf7j0UH[Ok7e0[G6c8T300O2O001N101N10001N101N10001N101O0I7H90O0000001O0000001O0000000000000000000000000000000001O000000000000000000000000000000000000O100O100O100O10000O100O100O100O11O00001O00001O00001O00001O00001O0O101O0000001O00001O1O5K5J5L0000000000O1000000000OO2J6J6I7J6J6J6K5L4L4L4L4ZOlF\\KW9b4f0K4L5K5K4C>B>B>C in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 26, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 257, 276, 278], "bbox": [0.0033750000000000004, 0.0045000000000000005, 0.22921874999999997, 0.7820208333333334], "iscrowd": 0, "area": 36055.204549999995, "rle": {"size": [480, 640], "counts": "Rn0d;\\3M3M3M3M3L4M3M3M4L3L4N200O100O1O100O100O1O100O1O10000000000000000O10000001O010O1O001O1O1O001O1O1OYFaKn8^4RGcKn8\\4RGdKn8\\4RGeKn8Z4RGgKm8Y4SGhKm8W4SGjKl8V4TGjKm8U4SGkKm8U4SGlKl8T4TGlKk8U4UGkKk8U4UGlKi8U4WGkKi8U4WGlKg8U4YGkKg8U4YGkKf8W4WGlKh8U4TGnKl8S4oFQLP9Q4lFSLm8T4nFPLa8`5J7I6K6K4K5nMWIgKP7X4^IUKg6j4h101N101N1002O0O101N1O101N100Of0ZO3N3L3N2M3M2O2N1O2N1O1O1O1O1O1O10hKaHm0^7ROhHi0X7WOlHf0S7ZORIb0P7ZOVIb0k6\\OZI`0h6\\O[Ib0g6\\O[Ib0h6ZOZIe0h6YOZIf0h6VO[Ih0g6VO[Ih0h6TOZIk0i6PO[In0i6mNZIQ1j6gN\\IX1g6^N`Ia1e6TNcIj1a6kMhIS2f9M3M3M3M4K3M2N3J5I7I8H7H9G8H8K6JXjV7"}, "label": "a woman wearing a blue and white shirt"}]} {"id": 229825, "image": "COCO_train2014_000000229825.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl blue color t shirt and blue color jeans\"."}], "task": "refering", "answer_template": "The \"a girl blue color t shirt and blue color jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 26, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 257, 276, 278], "bbox": [0.0033750000000000004, 0.0045000000000000005, 0.22921874999999997, 0.7820208333333334], "iscrowd": 0, "area": 36055.204549999995, "rle": {"size": [480, 640], "counts": "Rn0d;\\3M3M3M3M3L4M3M3M4L3L4N200O100O1O100O100O1O100O1O10000000000000000O10000001O010O1O001O1O1O001O1O1OYFaKn8^4RGcKn8\\4RGdKn8\\4RGeKn8Z4RGgKm8Y4SGhKm8W4SGjKl8V4TGjKm8U4SGkKm8U4SGlKl8T4TGlKk8U4UGkKk8U4UGlKi8U4WGkKi8U4WGlKg8U4YGkKg8U4YGkKf8W4WGlKh8U4TGnKl8S4oFQLP9Q4lFSLm8T4nFPLa8`5J7I6K6K4K5nMWIgKP7X4^IUKg6j4h101N101N1002O0O101N1O101N100Of0ZO3N3L3N2M3M2O2N1O2N1O1O1O1O1O1O10hKaHm0^7ROhHi0X7WOlHf0S7ZORIb0P7ZOVIb0k6\\OZI`0h6\\O[Ib0g6\\O[Ib0h6ZOZIe0h6YOZIf0h6VO[Ih0g6VO[Ih0h6TOZIk0i6PO[In0i6mNZIQ1j6gN\\IX1g6^N`Ia1e6TNcIj1a6kMhIS2f9M3M3M3M4K3M2N3J5I7I8H7H9G8H8K6JXjV7"}, "label": "a girl blue color t shirt and blue color jeans"}]} {"id": 459201, "image": "COCO_train2014_000000459201.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two zebras standing side by side with another behind them\"."}], "task": "refering", "answer_template": "The \"two zebras standing side by side with another behind them\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 76, 77, 78, 79, 80, 81, 82, 83, 93, 94, 95, 96, 97, 98, 99, 100, 112, 113, 114, 115, 116, 117, 118, 119, 130, 131, 132, 133, 134, 135, 148, 149, 151, 152, 166, 167, 169, 170, 171, 184, 185, 188, 189], "bbox": [0.21552000000000002, 0.2914655172413793, 0.6401800000000001, 0.9181896551724137], "iscrowd": 0, "area": 22087.974300000005, "rle": {"size": [348, 500], "counts": "]kT1>\\:5K6J5K6J5K5K6J5K6iIRNV3S2fLoMX3U2dLmMZ3X2aLkM\\3Y2`LiM^3[2^LhM_3m2kKVMS4X3^KlL_4b3RK`Lm4n3dJSL]5e4000010O0001O00001O001O11N4M4LaNaKfLY4W3nKkLm3R3XLlLh3Q3^LmLa3o2fLnLY3R3kLlLT3T3nLjLR3U3PMjLP3U3SMjLm2U3TMjLl2U3VMjLj2U3XMiLi2V3YMhLh2W3ZMhLg2V3[MhLf2V3]MhLd2W3^MgLc2W3c2N1N2O0O2O1N2O1O2N1001O0010O01O0010O0001O0010O01O0010O000000001O0000001O000000010O0000001O000000001O01O8H9H7H7I5K4L4M3L6J9G9G8H00_MeJA[5d0fKUN[4P2hKhMX4]2jK\\MV4i2kKQMU4T3lKfLT4_3mK[LS4j3]15K5K;E in this image.", "objects": [{"patches": [63, 64, 65, 76, 77, 78, 79, 80, 81, 82, 83, 93, 94, 95, 96, 97, 98, 99, 100, 112, 113, 114, 115, 116, 117, 118, 119, 130, 131, 132, 133, 134, 135, 148, 149, 151, 152, 166, 167, 169, 170, 171, 184, 185, 188, 189], "bbox": [0.21552000000000002, 0.2914655172413793, 0.6401800000000001, 0.9181896551724137], "iscrowd": 0, "area": 22087.974300000005, "rle": {"size": [348, 500], "counts": "]kT1>\\:5K6J5K6J5K5K6J5K6iIRNV3S2fLoMX3U2dLmMZ3X2aLkM\\3Y2`LiM^3[2^LhM_3m2kKVMS4X3^KlL_4b3RK`Lm4n3dJSL]5e4000010O0001O00001O001O11N4M4LaNaKfLY4W3nKkLm3R3XLlLh3Q3^LmLa3o2fLnLY3R3kLlLT3T3nLjLR3U3PMjLP3U3SMjLm2U3TMjLl2U3VMjLj2U3XMiLi2V3YMhLh2W3ZMhLg2V3[MhLf2V3]MhLd2W3^MgLc2W3c2N1N2O0O2O1N2O1O2N1001O0010O01O0010O0001O0010O01O0010O000000001O0000001O000000010O0000001O000000001O01O8H9H7H7I5K4L4M3L6J9G9G8H00_MeJA[5d0fKUN[4P2hKhMX4]2jK\\MV4i2kKQMU4T3lKfLT4_3mK[LS4j3]15K5K;E in this image.", "objects": [{"patches": [63, 64, 65, 66, 79, 80, 81, 82, 83, 84, 101, 102, 103, 117, 118, 119, 120, 121, 135, 136, 137, 153, 154, 155, 171, 172, 189, 190, 191], "bbox": [0.40362, 0.28304597701149425, 0.7419600000000001, 0.8831896551724139], "iscrowd": 0, "area": 9845.592200000005, "rle": {"size": [348, 500], "counts": "mhT21j:1O1O100O1O100O100O1O1O100O1O1O1O1O1O1000000O101O0000000O010000O10000O101N10O10O10O00001O0O101O0O101O0kJGI84LL43NL23OM120NO23MM24MM16NJ17OH19OG091GN93GL95FVND`Nf0[3FQNL]N?d3EkM4XN;n3AgM>PN4Z4]OcMk0gMJh4[O]MZ2f2fMWM[2k2eMQM]2Q3bMdLDYNk2V5aM`LFVNj2\\5`M]Le2e3[MYLd2i3]MYL_2i3aMYL[2i3eMXLW2k3iMWLk1S4UNoK^1\\4bNfKT1b4lN`Km0e4SO]Kf0i4YOXK`0o4_OSK;Q5EQK:o4ESK:m4EUK:k4EWK>d4B]Kh0PNAU5CmL[2h2gMXM_2a2aM_Ma2_2`M_Mc2^2_MaMd2\\2\\McMg2[2ZMdMi2Y2XMeMj2Y2XMfMi2Y2WM[KMn1l2g2hMUMY2k2mMoLS2P3UNhLd0kNmN]4i0^LOHQOi3Z1TLZOg0TOT3m1kKdN]1^Oh2W2bKXNj1@d2Y3aMeL_2Y3dMfL[2X3jMgLU2W3nMhLR2U3SNiLm1T3WNkLh1T3_NgLa1V3S3N3J6I6J7J6N1010O01O001O100O11O1O1O1O100O1O1O1O1O1O1O2O0O1O2N10O01O00100O01N1mMoG[1R8aNSH\\1n7aNXH\\1h7aN]H]1c7`NbH]1`7_NeH_1_8L3lNeFc0]9UOlFg0i9L3L5L3M3L5LUe[1"}, "label": "the second largest zebra with its head and front legs visible"}]} {"id": 459201, "image": "COCO_train2014_000000459201.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra behind other two zebras\"."}], "task": "refering", "answer_template": "The \"zebra behind other two zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 74, 75, 76, 77, 91, 92, 93, 109, 110, 111, 112, 128, 129, 130, 146, 147, 148, 164, 165, 166], "bbox": [0.08928, 0.24382183908045976, 0.3403, 0.835603448275862], "iscrowd": 0, "area": 9340.819099999999, "rle": {"size": [348, 500], "counts": "_^?6c:6J5K6J5K5K4L5L3M4L3M4K4M3M4M2N2N4M2M4L4L4M2M4L4RJSMg3P3_KgM`4\\2YKjMe4Z2RKkMo4X2bJSN^5e31O001O1O1O1O1\\OaJjK`5P4hJmKY5S4jJiKW5Y4e03M2N2N5K5KnNRJ[Mi5]2jJ\\MQ5^2YKeMa4Y2dKgMY4W2kKhMT4W2PLfMP4X2ULeMk3Y2YLfMf3Z2]LbMd3^2_L\\Md3d2^LXMd3h2_LSMd3m2^LnLd3R3T2O02O2M3NgMbHU1[7gNoHV1o6fNZIW1c6eNfIY1X6cNQJ[1k5aN_J]1[4aNiJNT1`1cNgN\\4NkKJ[1]2j2jMjKHa1Z2d2oMjKGg1V2_2TNhKFm1S2[2XNgKDQ2R2W2\\NfKBV2P2T2^NfKAY2o1P2bNeK^O`2k1m1gNcK]OQ3Z1^1Hk401O001N2O1O1O2NO1N2N1O2N2N2O11O0O2O001O001O003L6K00O10O10O0100O010OO2OK5O2N1O101N1O2N1L5KVR`3"}, "label": "zebra behind other two zebras"}]} {"id": 459201, "image": "COCO_train2014_000000459201.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra with its head obscured between two other zebras\"."}], "task": "refering", "answer_template": "The \"a zebra with its head obscured between two other zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 74, 75, 76, 77, 91, 92, 93, 109, 110, 111, 112, 128, 129, 130, 146, 147, 148, 164, 165, 166], "bbox": [0.08928, 0.24382183908045976, 0.3403, 0.835603448275862], "iscrowd": 0, "area": 9340.819099999999, "rle": {"size": [348, 500], "counts": "_^?6c:6J5K6J5K5K4L5L3M4L3M4K4M3M4M2N2N4M2M4L4L4M2M4L4RJSMg3P3_KgM`4\\2YKjMe4Z2RKkMo4X2bJSN^5e31O001O1O1O1O1\\OaJjK`5P4hJmKY5S4jJiKW5Y4e03M2N2N5K5KnNRJ[Mi5]2jJ\\MQ5^2YKeMa4Y2dKgMY4W2kKhMT4W2PLfMP4X2ULeMk3Y2YLfMf3Z2]LbMd3^2_L\\Md3d2^LXMd3h2_LSMd3m2^LnLd3R3T2O02O2M3NgMbHU1[7gNoHV1o6fNZIW1c6eNfIY1X6cNQJ[1k5aN_J]1[4aNiJNT1`1cNgN\\4NkKJ[1]2j2jMjKHa1Z2d2oMjKGg1V2_2TNhKFm1S2[2XNgKDQ2R2W2\\NfKBV2P2T2^NfKAY2o1P2bNeK^O`2k1m1gNcK]OQ3Z1^1Hk401O001N2O1O1O2NO1N2N1O2N2N2O11O0O2O001O001O003L6K00O10O10O0100O010OO2OK5O2N1O101N1O2N1L5KVR`3"}, "label": "a zebra with its head obscured between two other zebras"}]} {"id": 16836, "image": "COCO_train2014_000000016836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a square doughnut lying partially under a round doughnut\"."}], "task": "refering", "answer_template": "The \"a square doughnut lying partially under a round doughnut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 63, 64, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 127, 128, 129, 130, 131, 132, 133, 146, 147, 148, 149, 150, 164, 165, 166, 182, 183, 199], "bbox": [0.29552301255230123, 0.12725, 0.8956276150627616, 0.516890625], "iscrowd": 0, "area": 33376.333150000006, "rle": {"size": [640, 478], "counts": "bmh22mc01O1O1O2N1O1N2O1O2N1O1O1O2N1O1O1O1O2N1O1O1O1O2M2O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1N2O2N1O1O1O2N1O1O1O1O2N1O1O1O1O2N1N2O1O2N1O1O1O1O2N1O1O1O2N1O1O1O1N3N1O1O1O2N1O1O2N1O2N1O2N1O2N1O1O2M2O20O01O001O0010O01O001O010O001O0010O01O001O00010O001O0010O01O001O010O001O001O010O001O0010O0001O0010O01O001O0010O01O001O010O001O0010O01O00001O010O001O0010O01O001O010O001O001O011N2N3M2N3N1N3M2N3M2O2M2N3M3M2N3N1N3M2N3M2O2M2N3M2N3N1N3M3M2N3M2O1N1O1O1O10O0N2N2M3N3L3N2M3N2M3N3L3N2M3N2N3L3N2M3N2M4M2M3N2M3N2M4M2M3N2N2M4M2M3N2M3N3L3N2M3N2M3N3M2M3N2M3N3L3N2M3N2M4M2M4M3M3L4M3L4M3L4M3L4M3L4M7H:G8G:GZSo0"}, "label": "a square doughnut lying partially under a round doughnut"}]} {"id": 16836, "image": "COCO_train2014_000000016836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a square donut with nuts and glaze\"."}], "task": "refering", "answer_template": "The \"this is a square donut with nuts and glaze\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 63, 64, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 127, 128, 129, 130, 131, 132, 133, 146, 147, 148, 149, 150, 164, 165, 166, 182, 183, 199], "bbox": [0.29552301255230123, 0.12725, 0.8956276150627616, 0.516890625], "iscrowd": 0, "area": 33376.333150000006, "rle": {"size": [640, 478], "counts": "bmh22mc01O1O1O2N1O1N2O1O2N1O1O1O2N1O1O1O1O2N1O1O1O1O2M2O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1N2O2N1O1O1O2N1O1O1O1O2N1O1O1O1O2N1N2O1O2N1O1O1O1O2N1O1O1O2N1O1O1O1N3N1O1O1O2N1O1O2N1O2N1O2N1O2N1O1O2M2O20O01O001O0010O01O001O010O001O0010O01O001O00010O001O0010O01O001O010O001O001O010O001O0010O0001O0010O01O001O0010O01O001O010O001O0010O01O00001O010O001O0010O01O001O010O001O001O011N2N3M2N3N1N3M2N3M2O2M2N3M3M2N3N1N3M2N3M2O2M2N3M2N3N1N3M3M2N3M2O1N1O1O1O10O0N2N2M3N3L3N2M3N2M3N3L3N2M3N2N3L3N2M3N2M4M2M3N2M3N2M4M2M3N2N2M4M2M3N2M3N3L3N2M3N2M3N3M2M3N2M3N3L3N2M3N2M4M2M4M3M3L4M3L4M3L4M3L4M3L4M7H:G8G:GZSo0"}, "label": "this is a square donut with nuts and glaze"}]} {"id": 16836, "image": "COCO_train2014_000000016836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a round donut\"."}], "task": "refering", "answer_template": "The \"a round donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 145, 146, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 308, 309, 310, 311, 312, 313, 314, 315, 316, 327, 328, 329, 330, 331, 332, 346], "bbox": [0.02110878661087866, 0.3153125, 0.7207112970711297, 0.8783750000000001], "iscrowd": 0, "area": 93840.91865000002, "rle": {"size": [640, 478], "counts": "mb64ic08H8H8I7H8H8H8I7H8H8H8H8I7H8H8H9H6I5K4L5L3L4M4L3L5L3M4L3L5L3M4L3L4M4M2N3M2N3N1N3M2N2N3M2N3M2N3N1N3M2N3M2N2N3M2N3N1N3M2N2N1O2N2N1O2O1N2N1O2N2N1O2N2M2N3L4M3L3N3L4M2M4M3O001O1O1O001O1O001O1O1O001O1O010O1O001O1O1O001O1O001O1O001O1O1O001O1O00100O001O1O1O001O1O00001O001O00001O001O00001O010O00001O001O001O00001O001O00001O001O000010O01O0000000000O10000000000O10000000000O101O000000000O10000000000O10000000000O100000001O0O1000000000000O10000000000O10000000000O2O00000O1O1O1O100O1O1O1O1O2O0O1O2N1O1O2O0O1O2N1O1O2O0O1O2N1O1O2O0O1O2N1O1O2N1O1O1O2N1O2N1O2M3N1O2N1O2N2N1O2N2M2O2N1O2N2N2N3M2N2N2M3N3M2N2N2N3M2N2N2M3N3M2M3M3M3M4L3L4M3M4L3M3M3L4M4L3M3M3M4K7J5K5K5K5K5J7J5K5K5K5J6K6^MR@h0S`0cNhA2]Xc2"}, "label": "a round donut"}]} {"id": 16836, "image": "COCO_train2014_000000016836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the round donut\"."}], "task": "refering", "answer_template": "The \"the round donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 145, 146, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 308, 309, 310, 311, 312, 313, 314, 315, 316, 327, 328, 329, 330, 331, 332, 346], "bbox": [0.02110878661087866, 0.3153125, 0.7207112970711297, 0.8783750000000001], "iscrowd": 0, "area": 93840.91865000002, "rle": {"size": [640, 478], "counts": "mb64ic08H8H8I7H8H8H8I7H8H8H8H8I7H8H8H9H6I5K4L5L3L4M4L3L5L3M4L3L5L3M4L3L4M4M2N3M2N3N1N3M2N2N3M2N3M2N3N1N3M2N3M2N2N3M2N3N1N3M2N2N1O2N2N1O2O1N2N1O2N2N1O2N2M2N3L4M3L3N3L4M2M4M3O001O1O1O001O1O001O1O1O001O1O010O1O001O1O1O001O1O001O1O001O1O1O001O1O00100O001O1O1O001O1O00001O001O00001O001O00001O010O00001O001O001O00001O001O00001O001O000010O01O0000000000O10000000000O10000000000O101O000000000O10000000000O10000000000O100000001O0O1000000000000O10000000000O10000000000O2O00000O1O1O1O100O1O1O1O1O2O0O1O2N1O1O2O0O1O2N1O1O2O0O1O2N1O1O2O0O1O2N1O1O2N1O1O1O2N1O2N1O2M3N1O2N1O2N2N1O2N2M2O2N1O2N2N2N3M2N2N2M3N3M2N2N2N3M2N2N2M3N3M2M3M3M3M4L3L4M3M4L3M3M3L4M4L3M3M3M4K7J5K5K5K5K5J7J5K5K5K5J6K6^MR@h0S`0cNhA2]Xc2"}, "label": "the round donut"}]} {"id": 90569, "image": "COCO_train2014_000000090569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wine glass on the left\"."}], "task": "refering", "answer_template": "The \"wine glass on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 74, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 190, 191, 192, 193, 194, 214, 215, 237, 238, 260, 261, 283, 284, 304, 305, 306, 307, 308, 309, 328, 329, 330, 331, 332], "bbox": [0.22853125, 0.17044496487119437, 0.489046875, 0.9650351288056207], "iscrowd": 0, "area": 30002.11875, "rle": {"size": [427, 640], "counts": "UQm14T= in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 74, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 190, 191, 192, 193, 194, 214, 215, 237, 238, 260, 261, 283, 284, 304, 305, 306, 307, 308, 309, 328, 329, 330, 331, 332], "bbox": [0.22853125, 0.17044496487119437, 0.489046875, 0.9650351288056207], "iscrowd": 0, "area": 30002.11875, "rle": {"size": [427, 640], "counts": "UQm14T= in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 198, 221, 266, 267, 288, 289, 290, 291, 292], "bbox": [0.513390625, 0.16271662763466044, 0.7478281250000001, 0.8516861826697892], "iscrowd": 0, "area": 22388.281949999997, "rle": {"size": [427, 640], "counts": "mZY46R=6J7I6J7I6J7I6J7J5_IRN\\1T2aNlM]1W2aNkM[1Y2dNhMY1[2fNfMW1]2gNeMU1_2jNaMT1b2kN_MQ1e2mN]MP1f2oN[Mn0h2QOXMl0l2SOUMj0n2TOTMh0o2XOQMf0R3YOnLe0U3YOlLd0X3[OiLb0Z3]OfL`0^3_ObL?a3@_L>d3A\\LQ4BoK>Q4APL?Q4@PL?P4APL?Q4@oK`0R4_OnKa0T4]OlKc0U4[OlKe0U4ZOkKf0V4YOjKg0W4XOiKh0X4WOhKi0Y4VOgKj0Z4TOgKl0Z4SOfKm0\\4QOdKP1a4jN_KV1i4`NYK`1o4VNSKj1o4RNSKn1o4nMSKR2n4lMSKT2o4kMPKU2R5iMnJW2S5hMmJX2U5gMjJY2X5eMhJ[2W5fMiJZ2V5hMiJX2V5iMjJW2U5jMkJV2T5lMkJT2j4WNVKi1f4\\NYKd1d4_N\\Ka1`4cN`K]1\\4fNeKZ1W4jNiKV1T4lNmKT1R4WOdKi0[4XOeKh0[4XOeKi0Y4YOfKg0Z4YOfKg0Z4YOfKg0Y4[OfKe0Z4[OfKe0Y4\\OgKd0Y4\\OgKd0X4^OgKb0Y4^OgKb0X4_OhKa0X4_OhKa0W4AhK?X4AhK?W4BiK>W4ChK=W4DiKB`0@a0_Oa0_O[oR2"}, "label": "glass if wine on the riht"}]} {"id": 90569, "image": "COCO_train2014_000000090569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wine glass to the right of another wine glass\"."}], "task": "refering", "answer_template": "The \"a wine glass to the right of another wine glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 198, 221, 266, 267, 288, 289, 290, 291, 292], "bbox": [0.513390625, 0.16271662763466044, 0.7478281250000001, 0.8516861826697892], "iscrowd": 0, "area": 22388.281949999997, "rle": {"size": [427, 640], "counts": "mZY46R=6J7I6J7I6J7I6J7J5_IRN\\1T2aNlM]1W2aNkM[1Y2dNhMY1[2fNfMW1]2gNeMU1_2jNaMT1b2kN_MQ1e2mN]MP1f2oN[Mn0h2QOXMl0l2SOUMj0n2TOTMh0o2XOQMf0R3YOnLe0U3YOlLd0X3[OiLb0Z3]OfL`0^3_ObL?a3@_L>d3A\\LQ4BoK>Q4APL?Q4@PL?P4APL?Q4@oK`0R4_OnKa0T4]OlKc0U4[OlKe0U4ZOkKf0V4YOjKg0W4XOiKh0X4WOhKi0Y4VOgKj0Z4TOgKl0Z4SOfKm0\\4QOdKP1a4jN_KV1i4`NYK`1o4VNSKj1o4RNSKn1o4nMSKR2n4lMSKT2o4kMPKU2R5iMnJW2S5hMmJX2U5gMjJY2X5eMhJ[2W5fMiJZ2V5hMiJX2V5iMjJW2U5jMkJV2T5lMkJT2j4WNVKi1f4\\NYKd1d4_N\\Ka1`4cN`K]1\\4fNeKZ1W4jNiKV1T4lNmKT1R4WOdKi0[4XOeKh0[4XOeKi0Y4YOfKg0Z4YOfKg0Z4YOfKg0Y4[OfKe0Z4[OfKe0Y4\\OgKd0Y4\\OgKd0X4^OgKb0Y4^OgKb0X4_OhKa0X4_OhKa0W4AhK?X4AhK?W4BiK>W4ChK=W4DiKB`0@a0_Oa0_O[oR2"}, "label": "a wine glass to the right of another wine glass"}]} {"id": 90569, "image": "COCO_train2014_000000090569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the arm of the person on the left that is wearing a bracelet on their wrist\"."}], "task": "refering", "answer_template": "The \"the arm of the person on the left that is wearing a bracelet on their wrist\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 184, 185, 186, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 322], "bbox": [0.0015, 0.5069789227166276, 0.505625, 0.9666276346604216], "iscrowd": 0, "area": 34197.649249999995, "rle": {"size": [427, 640], "counts": "ah0f1X8]3POo02N2N2N2O1N2N1N3N2M3N2M3N2M3ZOoGoKS8P4e0O100O1O001O100O1O1O1O100O1O1O1O10O01O1O1O1O00000O101O000O1000000O101O000O10000O1000001N1000000O1000001N1000000O1000000O2O000O10000O10001N10000O100O1000001O0O100000000O10001O00000O100000001O00000000000000001O000000000000001O00000000000000001O000000000000001O000000000000001O0000000000001O000000000000001O0000000000000O10000000000000000000000000000000000000000000000000000000O10000000O1000000000000000000I7WOh0YOh0XOUU2>kjM`0A`0@;D;F;D:N2O1O1O1O1O1O1O100O100O100O10000O100O100O1000O0100O100O10000O0100000O01000O10O10O10O10O11N100O10000O101N10000O10000O2O0O10000O101N1O2N2N1O2N1O2N2N1O2O0N3K5K4M4K5K5L4K5K5K5K:E;FZ`S4"}, "label": "the arm of the person on the left that is wearing a bracelet on their wrist"}]} {"id": 90569, "image": "COCO_train2014_000000090569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hand on the left holding the wine glass\"."}], "task": "refering", "answer_template": "The \"the hand on the left holding the wine glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 184, 185, 186, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 322], "bbox": [0.0015, 0.5069789227166276, 0.505625, 0.9666276346604216], "iscrowd": 0, "area": 34197.649249999995, "rle": {"size": [427, 640], "counts": "ah0f1X8]3POo02N2N2N2O1N2N1N3N2M3N2M3N2M3ZOoGoKS8P4e0O100O1O001O100O1O1O1O100O1O1O1O10O01O1O1O1O00000O101O000O1000000O101O000O10000O1000001N1000000O1000001N1000000O1000000O2O000O10000O10001N10000O100O1000001O0O100000000O10001O00000O100000001O00000000000000001O000000000000001O00000000000000001O000000000000001O000000000000001O0000000000001O000000000000001O0000000000000O10000000000000000000000000000000000000000000000000000000O10000000O1000000000000000000I7WOh0YOh0XOUU2>kjM`0A`0@;D;F;D:N2O1O1O1O1O1O1O100O100O100O10000O100O100O1000O0100O100O10000O0100000O01000O10O10O10O10O11N100O10000O101N10000O10000O2O0O10000O101N1O2N2N1O2N1O2N2N1O2O0N3K5K4M4K5K5L4K5K5K5K:E;FZ`S4"}, "label": "the hand on the left holding the wine glass"}]} {"id": 90569, "image": "COCO_train2014_000000090569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman on the right\"."}], "task": "refering", "answer_template": "The \"the woman on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 43, 44, 45, 65, 66, 67, 68, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 183, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 265, 266, 267, 268], "bbox": [0.5277499999999999, 0.0, 1.0, 0.797751756440281], "iscrowd": 0, "area": 36322.73699999999, "rle": {"size": [427, 640], "counts": "nW]45Q=7H7I7M3N3M2N2M3N2N3M2N2M3N3M2N2M3N1O1O1O1N2O2N1O10000O1000000O2O000O10000O1000001N0100O010O100O10000O100O100O100O101N100O100000000001O00000O1O1O1O1O1O1O2M2O1O1O1O1O001O1O001O1N101O001O1O00001O00000O10000000000000001O000O1000000000000000001O0O1O1O1N2O1N2O1N2O1N2O1N3N1N2O1O1O100O100O100O101N100O100O100O100O100O101N100O1O100O100O100O100O101N10000000000O10000000001O000000000000000O1000001O000000000000O10000O010O100O100O100O100O100O100O010O100O100O100O100O100O1000O010jNZLhHe3U7RMXHn2d7]1101N2O1N2O1N2O1N2O1N2O1N2O1N21O5L4K5K5K5K2N1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O010O1O1O1O001O1O1O1O001O1O1O1O00eLY3"}, "label": "the woman on the right"}]} {"id": 90569, "image": "COCO_train2014_000000090569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman holding a glass\"."}], "task": "refering", "answer_template": "The \"woman holding a glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 43, 44, 45, 65, 66, 67, 68, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 183, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 265, 266, 267, 268], "bbox": [0.5277499999999999, 0.0, 1.0, 0.797751756440281], "iscrowd": 0, "area": 36322.73699999999, "rle": {"size": [427, 640], "counts": "nW]45Q=7H7I7M3N3M2N2M3N2N3M2N2M3N3M2N2M3N1O1O1O1N2O2N1O10000O1000000O2O000O10000O1000001N0100O010O100O10000O100O100O100O101N100O100000000001O00000O1O1O1O1O1O1O2M2O1O1O1O1O001O1O001O1N101O001O1O00001O00000O10000000000000001O000O1000000000000000001O0O1O1O1N2O1N2O1N2O1N2O1N3N1N2O1O1O100O100O100O101N100O100O100O100O100O101N100O1O100O100O100O100O101N10000000000O10000000001O000000000000000O1000001O000000000000O10000O010O100O100O100O100O100O100O010O100O100O100O100O100O1000O010jNZLhHe3U7RMXHn2d7]1101N2O1N2O1N2O1N2O1N2O1N2O1N21O5L4K5K5K5K2N1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O010O1O1O1O001O1O1O1O001O1O1O1O00eLY3"}, "label": "woman holding a glass"}]} {"id": 295370, "image": "COCO_train2014_000000295370.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown bear lying down on a concrete ground\"."}], "task": "refering", "answer_template": "The \"a brown bear lying down on a concrete ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 303, 304], "bbox": [0.17078125, 0.342140625, 0.5460625, 0.5915781250000001], "iscrowd": 0, "area": 21551.857300000007, "rle": {"size": [640, 640], "counts": "d\\T21nc05K2M3N1O2N1O1N3N1O1O2M2O1O2N1N3N1O1O2N1N2O2N1SOhNQ_OY1Pa0hNl^O[1g1dNm<4XAY1i1fNn<3VAX1k1hNm<2UAX1k1jNoo0hATOX>l0eAWO[>j0`AZO`>g0\\AWOi>i0SAWOR?i0h@WO]?i0[@[Oi?]211O000001O00000000010O0000001O000000010O0000001O0000001N100O1O101N1O100O2N100O1O2O0O100O1O2O0O1O101N1O100O2N100O1O101N11O3M2O2M3M2N3M3N1N1O0001O000000001O0000001O0000001O000000001O0000001N100O100O2O0O10001N100O101N100O2O000O2O0O101N100O101N10001N100O2O0O101N10001N100O2O0O2O1N2O1O0O2O1N2O1N2O1N2O1N101O1N2O1N2O1N2O0O2O1O1N2O1N2O1N2O0O2O1N2O1O1N2O1N2O2M3N2M3N1O2M3N2M3N2M3N2M3NRbe5"}, "label": "a brown bear lying down on a concrete ground"}]} {"id": 295370, "image": "COCO_train2014_000000295370.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown bear laying on some concrete\"."}], "task": "refering", "answer_template": "The \"brown bear laying on some concrete\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 303, 304], "bbox": [0.17078125, 0.342140625, 0.5460625, 0.5915781250000001], "iscrowd": 0, "area": 21551.857300000007, "rle": {"size": [640, 640], "counts": "d\\T21nc05K2M3N1O2N1O1N3N1O1O2M2O1O2N1N3N1O1O2N1N2O2N1SOhNQ_OY1Pa0hNl^O[1g1dNm<4XAY1i1fNn<3VAX1k1hNm<2UAX1k1jNoo0hATOX>l0eAWO[>j0`AZO`>g0\\AWOi>i0SAWOR?i0h@WO]?i0[@[Oi?]211O000001O00000000010O0000001O000000010O0000001O0000001N100O1O101N1O100O2N100O1O2O0O100O1O2O0O1O101N1O100O2N100O1O101N11O3M2O2M3M2N3M3N1N1O0001O000000001O0000001O0000001O000000001O0000001N100O100O2O0O10001N100O101N100O2O000O2O0O101N100O101N10001N100O2O0O101N10001N100O2O0O2O1N2O1O0O2O1N2O1N2O1N2O1N101O1N2O1N2O1N2O0O2O1O1N2O1N2O1N2O0O2O1N2O1O1N2O1N2O2M3N2M3N1O2M3N2M3N2M3N2M3NRbe5"}, "label": "brown bear laying on some concrete"}]} {"id": 90573, "image": "COCO_train2014_000000090573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl wearing a yellow shirt\"."}], "task": "refering", "answer_template": "The \"the girl wearing a yellow shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 84, 85, 86, 107, 108, 109, 130, 131, 132, 153, 154, 155, 175, 176, 177, 178, 179, 199, 200, 201, 202, 222, 223, 224, 225, 245, 246, 247, 248, 249, 268, 269, 271, 272, 273, 291, 292, 293, 296], "bbox": [0.619140625, 0.14157205240174672, 0.919859375, 0.8044978165938865], "iscrowd": 0, "area": 19525.01040000001, "rle": {"size": [458, 640], "counts": "iha52X>0O101N1O101N100O2N100O2O0O1O2O0O101N1O101fC_Oh:b0VEBh:>VEGf::XEK_1_O_6e0SH0Z1]O\\2c0cNOfM6W1YO`2X1lMYO\\N;T1UOd2X1kMWO\\Na0Q1QOh2W1jMVO]Ng0m0mNl2W1hMUO^Nk0k0jNo2W1gMSO^NQ1h0fNS3V1fMSO_NU1d0cNW3V1eMPO`N\\1`0_NZ3V1eMoNaN`1<\\N^3V1cMnNcNd18YNb3U1cMmNbNj15UNf3U1_MmNgNm10RNj3T1dLGBV1MPNm3T1kK=:d0KkMP4T1gK`0>e0GhMT4T1bKa0d0g0BdMY4T1]Kd0g0i0_O`M]4T1XKf0l0j0[O]Ma4S1TKi0P1k0WOZMe4b3RLYOUOUMj4c3oK[OTOSMm4c3mK1S4OkKXMYOX1m4a1hKXM]OU1k4d1fKXMAR1i4f1dKZMEm0h4j1aKZMIj0f4m1_KZMMg0d4o1]K\\M1c0b4R2[K\\M6>`4W2XK\\M:;^4Y2WK^M<7]4\\2TK_Ma02\\4`2QK_Me0OZ4c2oJ_Mi0LX4f4jKXKV4h4mKTKT4l4nKRKR4o4oKoJQ4Q5RLkJo3U5SLiJm3W5ULgJk3Y5WLdJj3\\5YLaJh3^5ZL`Jf3a5[L]Je3c5^LYJc3g5_LWJa3i5aLUJ_3k5cLRJ^3n5eLoI[3V1iJR3n1fK^3S1fJV3m1dKb3R1dJW3m1dKc3Q1bJZ3l1cKg3n0`J]3j1cKj3l0^J`3i1aKo3i0\\Jd3P8XLRHg3Q8ULRHi3R8RLPHm3S8oKPHo3T8lKnGS4U8hKoGV4T9O2O0O2N1O3M4L4K4M4L4L3M3L2O1O1O1O1O1N2O1O1O1kN_D^Ob;?`D@b;=_DDb;:_DFb;7`DIa;4aDL`;1bDO_;NcD1_;MbD3_;JcD6^;GcD:^;CdD=];@eD`0\\;^OeDa0];\\OeDd0\\;YOfDg0[;VOgDj0Z;TOgDl0Z;QOgDo0[<0000000000O10O100000O10000000O10O100000000O10O10000000O100000O1000002N2N2N2N1N2OO1N2N2O1N2O1N101N1O101N100O2L3M3M4K4M3M3Lb_f0"}, "label": "the girl wearing a yellow shirt"}]} {"id": 90573, "image": "COCO_train2014_000000090573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the player in a yellow shirt\"."}], "task": "refering", "answer_template": "The \"the player in a yellow shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 84, 85, 86, 107, 108, 109, 130, 131, 132, 153, 154, 155, 175, 176, 177, 178, 179, 199, 200, 201, 202, 222, 223, 224, 225, 245, 246, 247, 248, 249, 268, 269, 271, 272, 273, 291, 292, 293, 296], "bbox": [0.619140625, 0.14157205240174672, 0.919859375, 0.8044978165938865], "iscrowd": 0, "area": 19525.01040000001, "rle": {"size": [458, 640], "counts": "iha52X>0O101N1O101N100O2N100O2O0O1O2O0O101N1O101fC_Oh:b0VEBh:>VEGf::XEK_1_O_6e0SH0Z1]O\\2c0cNOfM6W1YO`2X1lMYO\\N;T1UOd2X1kMWO\\Na0Q1QOh2W1jMVO]Ng0m0mNl2W1hMUO^Nk0k0jNo2W1gMSO^NQ1h0fNS3V1fMSO_NU1d0cNW3V1eMPO`N\\1`0_NZ3V1eMoNaN`1<\\N^3V1cMnNcNd18YNb3U1cMmNbNj15UNf3U1_MmNgNm10RNj3T1dLGBV1MPNm3T1kK=:d0KkMP4T1gK`0>e0GhMT4T1bKa0d0g0BdMY4T1]Kd0g0i0_O`M]4T1XKf0l0j0[O]Ma4S1TKi0P1k0WOZMe4b3RLYOUOUMj4c3oK[OTOSMm4c3mK1S4OkKXMYOX1m4a1hKXM]OU1k4d1fKXMAR1i4f1dKZMEm0h4j1aKZMIj0f4m1_KZMMg0d4o1]K\\M1c0b4R2[K\\M6>`4W2XK\\M:;^4Y2WK^M<7]4\\2TK_Ma02\\4`2QK_Me0OZ4c2oJ_Mi0LX4f4jKXKV4h4mKTKT4l4nKRKR4o4oKoJQ4Q5RLkJo3U5SLiJm3W5ULgJk3Y5WLdJj3\\5YLaJh3^5ZL`Jf3a5[L]Je3c5^LYJc3g5_LWJa3i5aLUJ_3k5cLRJ^3n5eLoI[3V1iJR3n1fK^3S1fJV3m1dKb3R1dJW3m1dKc3Q1bJZ3l1cKg3n0`J]3j1cKj3l0^J`3i1aKo3i0\\Jd3P8XLRHg3Q8ULRHi3R8RLPHm3S8oKPHo3T8lKnGS4U8hKoGV4T9O2O0O2N1O3M4L4K4M4L4L3M3L2O1O1O1O1O1N2O1O1O1kN_D^Ob;?`D@b;=_DDb;:_DFb;7`DIa;4aDL`;1bDO_;NcD1_;MbD3_;JcD6^;GcD:^;CdD=];@eD`0\\;^OeDa0];\\OeDd0\\;YOfDg0[;VOgDj0Z;TOgDl0Z;QOgDo0[<0000000000O10O100000O10000000O10O100000000O10O10000000O100000O1000002N2N2N2N1N2OO1N2N2O1N2O1N101N1O101N100O2L3M3M4K4M3M3Lb_f0"}, "label": "the player in a yellow shirt"}]} {"id": 90573, "image": "COCO_train2014_000000090573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"group of three children on a soccer field\"."}], "task": "refering", "answer_template": "The \"group of three children on a soccer field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 109, 110, 132, 133, 155, 156, 157, 179, 180, 202, 203, 225, 226, 227, 247, 248, 249, 250, 251, 270, 271, 274, 293, 294], "bbox": [0.7381249999999999, 0.21698689956331876, 0.9556249999999998, 0.7626200873362445], "iscrowd": 0, "area": 8014.312449999999, "rle": {"size": [458, 640], "counts": "eec65S>5K4K6K5gH[Oe0j0UOYOi0k0SOWOk0n0oNUOP1o0jNTOU1P1eNSOY1Q1bNRO]1R1^NPOa1R1bMK[29cMG\\2;dMD]2=aMC_2?`M@`2a0_M_Oa2c0^M\\Ob2e0^MZOb2h0\\MYOc2i0\\MVOe2Q1SMoNm2Y1lLfNT3b1eL]N[3e1cL[N]3g1bLXN^3j1`LVN`3m1^LSNb3n1]LQNc3Q2[LoMf3R2YLmMj3S2SLmMP4R2oKmMT4R2jKmMZ4R2eKlMa4R2`KiMf4T2\\KfMj4X2XKcMn4Z2TKaMQ5^2QK[MV5b2lJYMY5e2iJVMX5l2S3c0]O1N101O001N2OO1N2N8H:F9G:F9G:F9G9G7I8H7Gg`4Bi_K2O1N3M2O1N2N2O1N2N101N2N101N2N2O00100O01000O10O010000O010O10O10O100O010O10O10O100O01000O010O10O10O010000O10000O100O10000O100OSW<"}, "label": "group of three children on a soccer field"}]} {"id": 90573, "image": "COCO_train2014_000000090573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy in a blue uniform is kicking a soccer ball\"."}], "task": "refering", "answer_template": "The \"a young boy in a blue uniform is kicking a soccer ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 96, 97, 98, 119, 120, 121, 122, 142, 143, 144, 145, 164, 165, 166, 167, 168, 188, 189, 190, 211, 212, 213, 234, 235, 236, 258, 259, 281, 282], "bbox": [0.1633125, 0.1943013100436681, 0.32996875000000003, 0.7671834061135371], "iscrowd": 0, "area": 14404.429900000001, "rle": {"size": [458, 640], "counts": "QU_13W>3L4L100cMLcF5Y93cFLZ9GX7U1XHUO;MZ7P1YHSO<0X7P1ZHQO=2V7P1[HmN>7T7o0[HjN`0:S7o0ZHgNc0=P7P1XHdNf0a0o6o0RHeNm0a0n6]2aHjM]7^2VHkMg7\\2lGlMQ8Y4N2M3O2N1O1N2O1O2N1O1]IYI`5k6XJXIh5o6oISIR6T7cIPI]6^72O0O2O0O2N1N3M2N3M2O2M2N3N5K4L3M000000100O1O100O1nLoHiNR7S1TIkNk6T1YIiNh6V1\\IfNe6X1_IeNb6Z1bIcN^6[1gIaNZ6^1jI^NW6`1oI\\NQ6c1SJYNm5f1XJVNi5i1\\JRNe5n1]JoMd5Q2]JmMd5h1hJVNY5h1jJUNX5i1kJUNV5i1QKQNo4o1XKjMi4T2[KhMg4V2\\KgMf4W2h3O1N2N2N1N3M3M3M3N2M3M3M4L3M3M3E;J6Jjfo5"}, "label": "a young boy in a blue uniform is kicking a soccer ball"}]} {"id": 90573, "image": "COCO_train2014_000000090573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in blue kicking a soccer ball\"."}], "task": "refering", "answer_template": "The \"a boy in blue kicking a soccer ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 96, 97, 98, 119, 120, 121, 122, 142, 143, 144, 145, 164, 165, 166, 167, 168, 188, 189, 190, 211, 212, 213, 234, 235, 236, 258, 259, 281, 282], "bbox": [0.1633125, 0.1943013100436681, 0.32996875000000003, 0.7671834061135371], "iscrowd": 0, "area": 14404.429900000001, "rle": {"size": [458, 640], "counts": "QU_13W>3L4L100cMLcF5Y93cFLZ9GX7U1XHUO;MZ7P1YHSO<0X7P1ZHQO=2V7P1[HmN>7T7o0[HjN`0:S7o0ZHgNc0=P7P1XHdNf0a0o6o0RHeNm0a0n6]2aHjM]7^2VHkMg7\\2lGlMQ8Y4N2M3O2N1O1N2O1O2N1O1]IYI`5k6XJXIh5o6oISIR6T7cIPI]6^72O0O2O0O2N1N3M2N3M2O2M2N3N5K4L3M000000100O1O100O1nLoHiNR7S1TIkNk6T1YIiNh6V1\\IfNe6X1_IeNb6Z1bIcN^6[1gIaNZ6^1jI^NW6`1oI\\NQ6c1SJYNm5f1XJVNi5i1\\JRNe5n1]JoMd5Q2]JmMd5h1hJVNY5h1jJUNX5i1kJUNV5i1QKQNo4o1XKjMi4T2[KhMg4V2\\KgMf4W2h3O1N2N2N1N3M3M3M3N2M3M3M4L3M3M3E;J6Jjfo5"}, "label": "a boy in blue kicking a soccer ball"}]} {"id": 74201, "image": "COCO_train2014_000000074201.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glass bowl right top corner\"."}], "task": "refering", "answer_template": "The \"glass bowl right top corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 17, 18, 35, 36, 37, 38, 39, 40, 60, 61, 62, 63], "bbox": [0.46557812500000006, 0.0, 0.816234375, 0.15325], "iscrowd": 0, "area": 10135.319, "rle": {"size": [480, 640], "counts": "Pf[41o>3M3M3M3M2N1O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001OO10000000000000000O1000000000000000000O100000000N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1N2N2N2M3N2N2N2NRZg1"}, "label": "glass bowl right top corner"}]} {"id": 74201, "image": "COCO_train2014_000000074201.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the empty container at the top of the other containers\"."}], "task": "refering", "answer_template": "The \"the empty container at the top of the other containers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 17, 18, 35, 36, 37, 38, 39, 40, 60, 61, 62, 63], "bbox": [0.46557812500000006, 0.0, 0.816234375, 0.15325], "iscrowd": 0, "area": 10135.319, "rle": {"size": [480, 640], "counts": "Pf[41o>3M3M3M3M2N1O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001OO10000000000000000O1000000000000000000O100000000N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1N2N2N2M3N2N2N2NRZg1"}, "label": "the empty container at the top of the other containers"}]} {"id": 74201, "image": "COCO_train2014_000000074201.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue cupcakes sitting in a clear container\"."}], "task": "refering", "answer_template": "The \"blue cupcakes sitting in a clear container\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 136, 157, 158, 159, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298], "bbox": [0.71959375, 0.32206250000000003, 0.9949375, 0.7342291666666667], "iscrowd": 0, "area": 18823.295850000002, "rle": {"size": [480, 640], "counts": "\\Zh64k>5K5K5K4L4L4L5K3M3IhNaBZ1^=5N2RD`NZ:b1bEaN\\:a1`EcN_:^1^EeN`:^1\\EeNc:\\1ZEgNd:[1YEhNf:Y1WEjNg:X1VEkNi:U1WElNh:T1WEnNg:S1XEoNg:Q1YEoNf:R1YEPOf:P1YEROg:n0XESOg:m0XEUOg:k0XEWOg:i0YEXOg:g0XE[Og:e0XE]Og:c0YE]Og:c0XE_Oh:`0WEAi:?WEBh:>WECi:=VEEi:;WEEj::UEGk:9TEIk:7TEJl:6TEKl:3TENl:2SE0l:0TE0l:OTE3l:LSE5m:KSE5m:JSE8l:HUE7l:HTE9k:FVE:j:FWE9i:GWE9i:GWE:i:DYE;g:EYE;g:EZE:f:E[E;f:D[E;e:D\\EXOg0b8nNWH8YOi0_8ROWH4ZOj0]8VOXHN]Ok0Z8YOZHG@o0U8]OZHWON[1g7APJ>P6CPJ;Q6FnI9S6HmI7R6KnI3S6NmI1S60mINT63lIKU66jIJV67jIGW6:iIEV6=jIAW6`0iI^OX6c0hI\\OX6e0\\GYNh1o0n6i0WG\\Ni1j0Q7k0RGaNk1a0U7o0mFeNo17T7U1jFiNS2MU7Z1fFnNV2BV7a1aFQOZ2ZOV7g2kHSMW7o2\\23M3L3N3oEeLW8]3^GVMX8k2^GfMZ8Z2eGhMZ8Y2dGiM[8W2cGlM[8e0kF]Oi0O\\8d0lF]Of01]8a0mF_Od03^8>nF@b03`86a87RGC<8a85RGD<8b83SGF8:d80SGG9:c8NUGI79d8NTGJ89c8LVGL68d8KVGN67e8JUG056f8IUG246h8GTG355h8FSG743i8FSG743j8CUG:12k8BUG<03j8@XG=N2k8_OXG?M2l8\\OZGa0K2k8\\O[Ga0K3k8ZO[Gc0J3k8YO]Gc0I3k8WO^Gf0G3P9QO[Gk0F3V9jNUGR1F4\\9bNPGY1E4c9YNkFb1B5c:K]E5d:I^E6c:I]E7c:I^E6c:I]E7d:G^E8b:H^E8c:G^E8c:G]E9c:F_E9b:F^E:c:E^E:b:F^E:c:D_E:c:D^Ec:@_E?b:_O`E`0a:^O`Ea0b:]O`Eb0a:\\OaEb0`:]OaEc0c:XO_Ef0h:SOZEl0l:mNUER1^ in this image.", "objects": [{"patches": [102, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244], "bbox": [0.3960625, 0.2539375, 0.694375, 0.6269583333333333], "iscrowd": 0, "area": 26955.227399999992, "rle": {"size": [480, 640], "counts": "Zif31k>6G9eAFe=c0UBDf=P1K4YO`NgCd1U<\\NkCi1P in this image.", "objects": [{"patches": [102, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244], "bbox": [0.3960625, 0.2539375, 0.694375, 0.6269583333333333], "iscrowd": 0, "area": 26955.227399999992, "rle": {"size": [480, 640], "counts": "Zif31k>6G9eAFe=c0UBDf=P1K4YO`NgCd1U<\\NkCi1P in this image.", "objects": [{"patches": [58, 60, 81, 82, 83, 84, 104, 105, 106, 107, 128, 129, 130, 152], "bbox": [0.5270937499999999, 0.17170798898071624, 0.6661874999999999, 0.5437190082644628], "iscrowd": 0, "area": 4035.842300000002, "rle": {"size": [363, 640], "counts": "Zbg34o9X1H83M2N3M001O000M4M2M4L10L33N2O11O1O05L3M0O101O000O11O2N2N2N2N2N10O001O001O1O1O001O1O14K5L[OSFIn9I_F6\\:O00I7J6J6I7J6I7J6J6L310RHUN[6k1eIWNZ6i1eIXNZ6h1fIZNY6e1gI\\NX6e1gI]NX6b1gI`NX6a1gI`NY6`1fIaNY6`1fI`N[6_1eIbN[6n1TISNm6b2\\H_Md7S3101O0L5J5K5J7H7J8H9F;F9F:G:G8I8GZh[2"}, "label": "the horse that is looking towards the motorcyles"}]} {"id": 74217, "image": "COCO_train2014_000000074217.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white horse right of another\"."}], "task": "refering", "answer_template": "The \"a white horse right of another\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 60, 81, 82, 83, 84, 104, 105, 106, 107, 128, 129, 130, 152], "bbox": [0.5270937499999999, 0.17170798898071624, 0.6661874999999999, 0.5437190082644628], "iscrowd": 0, "area": 4035.842300000002, "rle": {"size": [363, 640], "counts": "Zbg34o9X1H83M2N3M001O000M4M2M4L10L33N2O11O1O05L3M0O101O000O11O2N2N2N2N2N10O001O001O1O1O001O1O14K5L[OSFIn9I_F6\\:O00I7J6J6I7J6I7J6J6L310RHUN[6k1eIWNZ6i1eIXNZ6h1fIZNY6e1gI\\NX6e1gI]NX6b1gI`NX6a1gI`NY6`1fIaNY6`1fI`N[6_1eIbN[6n1TISNm6b2\\H_Md7S3101O0L5J5K5J7H7J8H9F;F9F:G:G8I8GZh[2"}, "label": "a white horse right of another"}]} {"id": 74217, "image": "COCO_train2014_000000074217.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"police biker 4th from left\"."}], "task": "refering", "answer_template": "The \"police biker 4th from left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 122, 123, 124, 125, 145, 146, 147, 168, 169, 170, 191, 192, 193, 215], "bbox": [0.308203125, 0.3305509641873278, 0.46134375, 0.7074104683195593], "iscrowd": 0, "area": 7722.849300000001, "rle": {"size": [363, 640], "counts": "ooU21V;5K5K4O2N2N1O2ZF_O_8c0^GF[8;cGHZ89eGJX88eGMW85gGMW84hGMX84fGNX84fGNY83eGNZ83eGOZ82dG0Z83cGN]83aGO]83_G1`8]1N3M3M2M4M3M4K6K4M3N3L2O1O1N2O1N2O1O1O1OO1O1O1O1O001O1O1O10000O101O0O10000O2N1O100O3M5K5K5K2O1VNnG7U8GTH0n7MUH1m7LUH3m7KUH3m7JVH4l7JVH4l7KTH4m7KVH2l7MVHOm70UHMl73VHJl74WHFm7:UH_Om7c0]1O2O1O1N2O1O1N2O1O1N2O1O3L4MPUj3"}, "label": "police biker 4th from left"}]} {"id": 74217, "image": "COCO_train2014_000000074217.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle police motorcycle in a line of them on the street\"."}], "task": "refering", "answer_template": "The \"the middle police motorcycle in a line of them on the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 122, 123, 124, 125, 145, 146, 147, 168, 169, 170, 191, 192, 193, 215], "bbox": [0.308203125, 0.3305509641873278, 0.46134375, 0.7074104683195593], "iscrowd": 0, "area": 7722.849300000001, "rle": {"size": [363, 640], "counts": "ooU21V;5K5K4O2N2N1O2ZF_O_8c0^GF[8;cGHZ89eGJX88eGMW85gGMW84hGMX84fGNX84fGNY83eGNZ83eGOZ82dG0Z83cGN]83aGO]83_G1`8]1N3M3M2M4M3M4K6K4M3N3L2O1O1N2O1N2O1O1O1OO1O1O1O1O001O1O1O10000O101O0O10000O2N1O100O3M5K5K5K2O1VNnG7U8GTH0n7MUH1m7LUH3m7KUH3m7JVH4l7JVH4l7KTH4m7KVH2l7MVHOm70UHMl73VHJl74WHFm7:UH_Om7c0]1O2O1O1N2O1O1N2O1O1N2O1O3L4MPUj3"}, "label": "the middle police motorcycle in a line of them on the street"}]} {"id": 377329, "image": "COCO_train2014_000000377329.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black cow with a white splattered stripe on its back\"."}], "task": "refering", "answer_template": "The \"a black cow with a white splattered stripe on its back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 152, 153, 154, 155, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 225, 240, 241, 242, 243, 244, 245, 246, 247, 263, 264, 265, 266, 267, 268, 269, 270, 286, 287, 288, 289, 290], "bbox": [0.460046875, 0.3762060889929742, 0.7943125, 0.8453864168618267], "iscrowd": 0, "area": 28034.17655, "rle": {"size": [427, 640], "counts": "Ymj31\\ in this image.", "objects": [{"patches": [130, 131, 152, 153, 154, 155, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 225, 240, 241, 242, 243, 244, 245, 246, 247, 263, 264, 265, 266, 267, 268, 269, 270, 286, 287, 288, 289, 290], "bbox": [0.460046875, 0.3762060889929742, 0.7943125, 0.8453864168618267], "iscrowd": 0, "area": 28034.17655, "rle": {"size": [427, 640], "counts": "Ymj31\\ in this image.", "objects": [{"patches": [84, 85, 86, 87, 88, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 365, 366, 367], "bbox": [0.538890625, 0.20493750000000002, 1.0, 0.9432083333333333], "iscrowd": 0, "area": 77298.84050000002, "rle": {"size": [480, 640], "counts": "SPR52j>6K4L4L4L4L4L4L4L4hNjNaDZ1^;hN]D]1S9iNSHMeN^1j8SO`HAbN_1d8YOjHYObN^1c8ZOjHZOaN]1d8ZOjH[O`N\\1d8[OkH[O_N[1e8[OjH]O_NY1f8ZOkH_O^NW1e8\\OlH^O^NW1e8\\OlH_O]NW1d8\\OnH_O\\NV1e8\\OnH@[NU1f8[OnHC[NR1e8]OoHCZNQ1f8]OoHf1P7[NoHf1o6\\NPIe1o6\\NPIe1o6[NPIg1n6[NPIh1n6YNPIi1o6XNXH[N0_3f7XNXH\\NO^3h7VNXH`NK]3l7TNWHbNJ\\3n7SNVHeNHZ3Q8QNWHgND[3T8nMWHh2g7YMXHi2g7WMXHl2f7TMUHR3j7oLQHV3n7o1O1O1N10001O00001O0O10001O0000001N10001O0000001N10001O0000001N10001O0000001N10001O0000001N1000001O00001O0000001O00001O0000001O00001O3M2N2N3M2N2N2N3M2N00001O0000001O000000001O00000O2O000O101O0O10001N1000001N10000O2O000O2O001N101O1N101O0O2O001N20O01O010O1O010O010O1O010O00100O010O0010O01O1O0010O01O1O010O001O10O01O001O10O01O0010O01O1O010O001O1O010O001O10O01O00100O001O0010O01O1O010O001O10O01O001O10O01O00100O001O010O001O1O010O000001N1O1O101N1O1O100O2N1O100O2N1O100O2N1O100O1O2N100O1O2N100O2N2N3N1N2N3M2OfK"}, "label": "half a sandwich sitting on the right side of a plate"}]} {"id": 66034, "image": "COCO_train2014_000000066034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sandwich half to right\"."}], "task": "refering", "answer_template": "The \"sandwich half to right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 87, 88, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 365, 366, 367], "bbox": [0.538890625, 0.20493750000000002, 1.0, 0.9432083333333333], "iscrowd": 0, "area": 77298.84050000002, "rle": {"size": [480, 640], "counts": "SPR52j>6K4L4L4L4L4L4L4L4hNjNaDZ1^;hN]D]1S9iNSHMeN^1j8SO`HAbN_1d8YOjHYObN^1c8ZOjHZOaN]1d8ZOjH[O`N\\1d8[OkH[O_N[1e8[OjH]O_NY1f8ZOkH_O^NW1e8\\OlH^O^NW1e8\\OlH_O]NW1d8\\OnH_O\\NV1e8\\OnH@[NU1f8[OnHC[NR1e8]OoHCZNQ1f8]OoHf1P7[NoHf1o6\\NPIe1o6\\NPIe1o6[NPIg1n6[NPIh1n6YNPIi1o6XNXH[N0_3f7XNXH\\NO^3h7VNXH`NK]3l7TNWHbNJ\\3n7SNVHeNHZ3Q8QNWHgND[3T8nMWHh2g7YMXHi2g7WMXHl2f7TMUHR3j7oLQHV3n7o1O1O1N10001O00001O0O10001O0000001N10001O0000001N10001O0000001N10001O0000001N10001O0000001N1000001O00001O0000001O00001O0000001O00001O3M2N2N3M2N2N2N3M2N00001O0000001O000000001O00000O2O000O101O0O10001N1000001N10000O2O000O2O001N101O1N101O0O2O001N20O01O010O1O010O010O1O010O00100O010O0010O01O1O0010O01O1O010O001O10O01O001O10O01O0010O01O1O010O001O1O010O001O10O01O00100O001O0010O01O1O010O001O10O01O001O10O01O00100O001O010O001O1O010O000001N1O1O101N1O1O100O2N1O100O2N1O100O2N1O100O1O2N100O1O2N100O2N2N3N1N2N3M2OfK"}, "label": "sandwich half to right"}]} {"id": 66034, "image": "COCO_train2014_000000066034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwich half on the left\"."}], "task": "refering", "answer_template": "The \"the sandwich half on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 323, 324, 325, 326, 327, 328, 329, 330, 331, 348, 349, 350, 351, 352, 353, 373, 374, 375], "bbox": [0.0, 0.22716666666666668, 0.6055625, 0.9802500000000001], "iscrowd": 0, "area": 107086.0293, "rle": {"size": [480, 640], "counts": "b8_3f6l4N101O001O001O001O001O001N101O001O001O001O001O001O0O2O001O00001O001O001O001N101O001O001O001O001O001N101O001O001O001O001O001N101O001O00001O001O001O0O2O001O001O001O001O001O0O2O001O001O001O00001O001N101O001O001O001O00001O0O2O001O001O001O001O00001O001O001O001O001O0010O0001O001O001O001O001O001O00001O001O001O001O001O001O000010O01O001O001O001O001O00001O001O001O001O000001O0000000000000000000000000000000000000N2N2O2M2N2N2O1N2N2N3N1N2N2O2M2N2O100O2O000O10001N10000O2O000O10001N10000O101O0O10001N10000O101O0O1N2N3M2N2N2O1N3M2N2N2N2N3N1N2N2N3N1O100O1O2O0O100O2N100O1O101N1O100O1O2O0O1O101N1O1N2N2O2M2N2O1N2N3N1N2O1N3N100O100O2O0O1O101N100O1O1O2N1O1O100O2N1O1O1O2N100O1O1O2N1O1O100O2N1O1O1O2N100O1O1O2N1O1fNoFdLR9Z3RGbLR9Y3RGcLS9Y3QGbLT9Y3PGdLT9W3QGdLU9W3oFdLU9X3oFeLS9X3RGcLR9R2iEROc;l0cDlNa;S1cDgN_;Y1fD_N^;_1gDZN];e1hDTN[;j1n001O1N2O001O1O001O1O1O001N2VObBNa=0eBH^=7gBC\\= in this image.", "objects": [{"patches": [72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 323, 324, 325, 326, 327, 328, 329, 330, 331, 348, 349, 350, 351, 352, 353, 373, 374, 375], "bbox": [0.0, 0.22716666666666668, 0.6055625, 0.9802500000000001], "iscrowd": 0, "area": 107086.0293, "rle": {"size": [480, 640], "counts": "b8_3f6l4N101O001O001O001O001O001N101O001O001O001O001O001O0O2O001O00001O001O001O001N101O001O001O001O001O001N101O001O001O001O001O001N101O001O00001O001O001O0O2O001O001O001O001O001O0O2O001O001O001O00001O001N101O001O001O001O00001O0O2O001O001O001O001O00001O001O001O001O001O0010O0001O001O001O001O001O001O00001O001O001O001O001O001O000010O01O001O001O001O001O00001O001O001O001O000001O0000000000000000000000000000000000000N2N2O2M2N2N2O1N2N2N3N1N2N2O2M2N2O100O2O000O10001N10000O2O000O10001N10000O101O0O10001N10000O101O0O1N2N3M2N2N2O1N3M2N2N2N2N3N1N2N2N3N1O100O1O2O0O100O2N100O1O101N1O100O1O2O0O1O101N1O1N2N2O2M2N2O1N2N3N1N2O1N3N100O100O2O0O1O101N100O1O1O2N1O1O100O2N1O1O1O2N100O1O1O2N1O1O100O2N1O1O1O2N100O1O1O2N1O1fNoFdLR9Z3RGbLR9Y3RGcLS9Y3QGbLT9Y3PGdLT9W3QGdLU9W3oFdLU9X3oFeLS9X3RGcLR9R2iEROc;l0cDlNa;S1cDgN_;Y1fD_N^;_1gDZN];e1hDTN[;j1n001O1N2O001O1O001O1O1O001N2VObBNa=0eBH^=7gBC\\= in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 135, 137, 138, 139, 140, 141, 142, 143, 144, 157, 158, 161, 162, 163, 164, 165, 166, 167, 168, 180, 181, 184, 185, 186, 187, 188, 189, 190, 191, 203, 204, 207, 208, 209, 210, 212, 213, 214, 226, 227, 230, 231, 232, 233, 235, 236, 237, 238, 239, 249, 250, 253, 254, 255, 256, 258, 259, 260, 261, 262, 272, 273, 275, 276, 277, 278, 279, 281, 282, 283, 284, 295, 296, 298, 299, 300, 301, 321], "bbox": [0.0028125, 0.0011250000000000001, 1.0, 0.9493249999999999], "iscrowd": 0, "area": 96782.54055, "rle": {"size": [400, 640], "counts": "al0S8n0_300000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O0000000000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O100iM[ObHf0o6LnH4c6>YICX6P1dIPOm5b1oI_Ne5P2WJQNc5Y2YJgMa5b2[J_M_5j2]JWM]5R3`JnLZ5\\3aJeLY5d3cJ]LW5l3fJTLT5V4gJkKS5^4jJbKP5R6J6K5J6J6J6J6J6J6J6J6006J5K6J5K6J6J5K6J6J5K6J6J5K6J5K6J6J5K6J6J5K6J6J5K6J5K4L3M2N2N2N3M2N2fNdG\\N^8\\1hGdN[8S1lGlNV8l0PHTOR8d0UH[Om7=YHCj74\\HLg7K_H5n7XOXHh0h90000000000000000000O100000000000000000000000000000000000000O1000000O1000000]I8=IdNX1Z1hNcN]1[1cNbNb1[1_NaNg1]1YN`Nl1^1TN^NR2_1oM^NV2`1jM]N[2a1eM[Na2c1_MZNf2c1[MZNo2_1QM]N\\3Z1dLcNh3S1YLlNQ4k0oKUOZ4b0fK]Od49]KGl40TKOR5LnJ4R5KoJ4S5KmJ4U5KkJ5V5JjJ5X5IiJ7W5IiJ6Y5IgJ7Z5GgJ8[5GeJ9\\5FdJ9^5EcJ;]5EcJ:_5EaJ;`5CaJd5A]J>e5A[J?f5@ZJ?h5_OYJa0h5^OXJa0i5_OWJa0j5]OWJb0k5]OUJc0l5\\OTJb0n5]OSJ`0P6@PJ=S6CmI;U6EkI8X6GiI6Z6JfI3]6McI0`6OaIOa61_ILd64\\IIg65[IHh66ZIGi66YIIi65YIHj65YIHj66XIGk67WIFl67WIHj66XIJh63[INd60^I0b6MaI4^6JdI6\\6GgI9Y6EiId0Eh1l0dNnN\\O43?h1`0jNiN]1T1D4e28WMIa2c0ZM^O]2o0^MROY2[1YMoN^2^1oLQOi2Z1fLTOQ3Y1\\LVO[3W1SLWOi3j4L4K5L4K5L4L4K5L4O11O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1N3N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1kN]K[Je4a5^K^Jc4_5`K`Ja4\\5bKdJ_4Y5dKfJ]4V5gKiJ[4S5hKlJY4P5kKoJV4n4mKQKT4k4oKUKR4h4QLWKP4f4SLYKo3b4ULiL`2T3XMSKoN^4l0c0G_6BXF"}, "label": "the body of the adult elephant"}]} {"id": 500214, "image": "COCO_train2014_000000500214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby elephant with it ' s mouth open\"."}], "task": "refering", "answer_template": "The \"a baby elephant with it ' s mouth open\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 37, 38, 39, 40, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 237, 238, 239, 240, 241, 242, 243, 245, 246, 248, 249, 262, 263, 265, 266, 268, 269, 270, 271, 272, 284, 285, 286, 288, 289, 290, 291, 292, 293, 294, 295, 311, 312, 314, 315, 316, 317, 318], "bbox": [0.301203125, 0.094025, 0.8769218750000001, 0.956075], "iscrowd": 0, "area": 92770.17550000004, "rle": {"size": [400, 640], "counts": "b`[26W<8I8G8H9G8I8G8H9H7J7H7J7I6J7I6Jh0XOe0[O9G9G9G8G4M1O1O2N1O2N1O2N1O1O2M2O2N1O2N1O2N10001N101O0aLkHb1V7]NlHb1S7^NnHb1R7]NPIa1Q7^NPIb1P7]NRIb1n6]NTIa1m6^NTIb1l6]NVIb1i6^NXIb1h6\\NZIc1g6\\NZId1f6[N[Ie1e6ZN\\Ie1e6ZN\\If1d6YN]Ig1b6YN_If1b6YN_Ig1a6XN`Ih1b6UN_Ij1e6RN\\In1g6mM[IS2h6iMYIV2j6fMXIZ2k6aMWI_2l6]MUIc2n6XMTIg2P7UMQIk2Z800O10000O10000O10000O10000O1000000O10000O10000O10O10O100000O0RObMlG^2Q8VN^Gj1`8S1N3M3N2M2O2M3N1N3N2M2N3N2M3N1N3N2M2O2M3M3N1N3N2N13N3M2N3L4M2N3L3N3M2M4M2N3M3L3N3aMeGc0]8YOfGe0^8VOfGh0\\8TOhGj0[8QOhGn0Z8nNjGo0Z8lNiGS1Z8hNiGW1Y8eNiGZ1[8bNfG^1\\8^NfGb1]8ZNdGf1^8VNeGi1a90000000001O00000000O100000000O100000000O100000000O1jK_N]Ma1a2nNRMR1k2^OhLb0V3M]L3a3=PLDo3d0jK\\OV4h0fKXOY4m0cKSO\\4R1`KnN_4X1[KiNf4Z1VKfNj4^1RKbNn4c1mJ]NS5g1iJYNW5k1dJVN]5n1^JRNd5P2WJQNj5R2PJPNR6V42N2N3M2N2N2N2N2N2N2N2N1O2O1N2N2N2N1O2O1N2N3M3kMaHSOb7e0cH[O`7>dHC^75gHK\\7NhH2[7FjH:Y7_OkHa0X7WOmHj0U7oNoHQ1T7gNQIY1R7`NRI`1Q7XNTIh1o6QNUIo1f81O00000000001O00000000001O00000000001O000000001O00000000001O00000000001O0000000YM\\NbJd1k4SOQKn0m4XOnJh0Q5]OkJc0S5DhJY2@kM?S2_OQNa0n1\\OVNd0h1ZO]Ne0b1XObNi0[1UOiNk0V1ROnNn0P1QOSOP1k0mNYOS1f0jN^OW1?gNEY1:dNJ]13aN1_1N^N6c1G\\N in this image.", "objects": [{"patches": [33, 34, 35, 36, 37, 38, 39, 40, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 237, 238, 239, 240, 241, 242, 243, 245, 246, 248, 249, 262, 263, 265, 266, 268, 269, 270, 271, 272, 284, 285, 286, 288, 289, 290, 291, 292, 293, 294, 295, 311, 312, 314, 315, 316, 317, 318], "bbox": [0.301203125, 0.094025, 0.8769218750000001, 0.956075], "iscrowd": 0, "area": 92770.17550000004, "rle": {"size": [400, 640], "counts": "b`[26W<8I8G8H9G8I8G8H9H7J7H7J7I6J7I6Jh0XOe0[O9G9G9G8G4M1O1O2N1O2N1O2N1O1O2M2O2N1O2N1O2N10001N101O0aLkHb1V7]NlHb1S7^NnHb1R7]NPIa1Q7^NPIb1P7]NRIb1n6]NTIa1m6^NTIb1l6]NVIb1i6^NXIb1h6\\NZIc1g6\\NZId1f6[N[Ie1e6ZN\\Ie1e6ZN\\If1d6YN]Ig1b6YN_If1b6YN_Ig1a6XN`Ih1b6UN_Ij1e6RN\\In1g6mM[IS2h6iMYIV2j6fMXIZ2k6aMWI_2l6]MUIc2n6XMTIg2P7UMQIk2Z800O10000O10000O10000O10000O1000000O10000O10000O10O10O100000O0RObMlG^2Q8VN^Gj1`8S1N3M3N2M2O2M3N1N3N2M2N3N2M3N1N3N2M2O2M3M3N1N3N2N13N3M2N3L4M2N3L3N3M2M4M2N3M3L3N3aMeGc0]8YOfGe0^8VOfGh0\\8TOhGj0[8QOhGn0Z8nNjGo0Z8lNiGS1Z8hNiGW1Y8eNiGZ1[8bNfG^1\\8^NfGb1]8ZNdGf1^8VNeGi1a90000000001O00000000O100000000O100000000O100000000O1jK_N]Ma1a2nNRMR1k2^OhLb0V3M]L3a3=PLDo3d0jK\\OV4h0fKXOY4m0cKSO\\4R1`KnN_4X1[KiNf4Z1VKfNj4^1RKbNn4c1mJ]NS5g1iJYNW5k1dJVN]5n1^JRNd5P2WJQNj5R2PJPNR6V42N2N3M2N2N2N2N2N2N2N2N1O2O1N2N2N2N1O2O1N2N3M3kMaHSOb7e0cH[O`7>dHC^75gHK\\7NhH2[7FjH:Y7_OkHa0X7WOmHj0U7oNoHQ1T7gNQIY1R7`NRI`1Q7XNTIh1o6QNUIo1f81O00000000001O00000000001O00000000001O000000001O00000000001O00000000001O0000000YM\\NbJd1k4SOQKn0m4XOnJh0Q5]OkJc0S5DhJY2@kM?S2_OQNa0n1\\OVNd0h1ZO]Ne0b1XObNi0[1UOiNk0V1ROnNn0P1QOSOP1k0mNYOS1f0jN^OW1?gNEY1:dNJ]13aN1_1N^N6c1G\\N in this image.", "objects": [{"patches": [169, 191, 192, 193, 214, 215, 216, 217, 238, 239, 240, 241, 261, 262, 263, 285, 286, 287], "bbox": [0.33815625, 0.4477453027139875, 0.503234375, 0.7560960334029226], "iscrowd": 0, "area": 8683.939649999998, "rle": {"size": [479, 640], "counts": "XYU32m=V1H7J5K5J4M4L3L5N1N4L4M4K4M3L5K4M2M4M1N1O2O0O101O0O100O100O010O10000O10O001O001O1O001N101O001O1N2O1O1O1O1N2O1N2O2M2O1N2O1N2O001N2O1O1O001OoC_Me;`2YDaMi;j21O001O1N2O1O001N2O1^OdC_N] in this image.", "objects": [{"patches": [169, 191, 192, 193, 214, 215, 216, 217, 238, 239, 240, 241, 261, 262, 263, 285, 286, 287], "bbox": [0.33815625, 0.4477453027139875, 0.503234375, 0.7560960334029226], "iscrowd": 0, "area": 8683.939649999998, "rle": {"size": [479, 640], "counts": "XYU32m=V1H7J5K5J4M4L3L5N1N4L4M4K4M3L5K4M2M4M1N1O2O0O101O0O100O100O010O10000O10O001O001O1O001N101O001O1N2O1O1O1O1N2O1N2O2M2O1N2O1N2O001N2O1O1O001OoC_Me;`2YDaMi;j21O001O1N2O1O001N2O1^OdC_N] in this image.", "objects": [{"patches": [103, 104, 105, 125, 126, 127, 128, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 219, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 261, 262, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 288, 304, 305, 306, 307, 308, 309, 310, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.139125, 0.2984777517564403, 0.58640625, 0.98903981264637], "iscrowd": 0, "area": 43642.95235000001, "rle": {"size": [427, 640], "counts": "P]U11Z=1N1O2O1N2O1N101N2N101N101N1O2O1N101N101N101O00001O0O10001O0000001O000O10001O0000001hFPOS6P1kIROT6o0kIROU6n0iITOW6m0fIUOZ6k0dIWO[6j0dIWO\\6j0aIXO_6h0_IZO`6h0^IYOb6g0\\I[Od6f0YI\\Of6e0XI]Oh6d0VI]Oj6d0SI^Ol6e0PI]OP7d0mH^OS7d0jH^OT7e0hH]OX7e0dH]O\\7g0^H[Oa7k0YHVOg7n0SHTOm7Q1lGQOS8T3N2N2N2M3N2N2N2N2O100O100O100O100O10000O100O100O100O100O10000O100O100O100O100O10000O100O100O100O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1N2O1O1N2O1O1OKmJTHQ5l7TKQHk4n79O1O001O1O001O1O1O001O101N101N100010O01O00001O010O000N3G8I8H7H8G:D;E in this image.", "objects": [{"patches": [103, 104, 105, 125, 126, 127, 128, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 219, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 261, 262, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 288, 304, 305, 306, 307, 308, 309, 310, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.139125, 0.2984777517564403, 0.58640625, 0.98903981264637], "iscrowd": 0, "area": 43642.95235000001, "rle": {"size": [427, 640], "counts": "P]U11Z=1N1O2O1N2O1N101N2N101N101N1O2O1N101N101N101O00001O0O10001O0000001O000O10001O0000001hFPOS6P1kIROT6o0kIROU6n0iITOW6m0fIUOZ6k0dIWO[6j0dIWO\\6j0aIXO_6h0_IZO`6h0^IYOb6g0\\I[Od6f0YI\\Of6e0XI]Oh6d0VI]Oj6d0SI^Ol6e0PI]OP7d0mH^OS7d0jH^OT7e0hH]OX7e0dH]O\\7g0^H[Oa7k0YHVOg7n0SHTOm7Q1lGQOS8T3N2N2N2M3N2N2N2N2O100O100O100O100O10000O100O100O100O100O10000O100O100O100O100O10000O100O100O100O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1N2O1O1N2O1O1OKmJTHQ5l7TKQHk4n79O1O001O1O001O1O1O001O101N101N100010O01O00001O010O000N3G8I8H7H8G:D;E in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 138, 139, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 161, 162, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 184, 185, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 211, 212, 213, 214, 215, 216, 217, 230, 231, 234, 235, 236, 237, 253, 254, 257, 258, 259, 260, 276, 277, 279, 280, 281, 282, 283, 299, 300, 302, 303, 304, 305, 306, 322, 323, 325, 326, 327, 328, 329, 346, 348, 349, 350, 351, 352, 369, 371, 372], "bbox": [0.0, 0.0022500000000000003, 0.691015625, 0.9865208333333333], "iscrowd": 0, "area": 115802.61475, "rle": {"size": [480, 640], "counts": "_7n1[T;VOPEj0W;jNnDV1^<0000O10O1000O100000000O1000O1000O100000O10000O1000000O1000000O10000O100mKiN]JW1b5QOWJo0g5ZOQJg0n5AkI?T6IeI7Y63_IM`6;XIFf6c0SI]Ol6k0mHUOR7S1gHmNW7\\1aHeN^7d1ZH\\Nd7m1UHSNj7U2oGkMP8Y2mGgMQ8]2lGdMS8`2jG`MU8c2iG]MV8f2hGZMX8i2dGWM\\8l2bGTM]8P3`GPM`8R3^GnLa8i400O1O1O100O1O100O1O1O100O1O100O1O1O100O1O1O100000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000O1O1O1N2O1O1O1N2O1O100O010O100O10000O100O100O100O100O100O10000O100O100O2O0O101N1O101N100O2N100O1O2O0O1O2O0O101N10001N100O2O001N101O0O2O1O0O2O0O2O1N101N101O0O2O1N101N101N2O0O2O001N2O0O2O0O2O1N101N101N2O0O2O0O2O1N101N101N2O0O2O0O2O1N2O2M2O2M3N2M3M3N1N3N2M3N2M2O2M3N2M3N1N3N2M3N2M3N1I8G9F:G9Ff0ZOlfl2"}, "label": "a big elephant walking with the small elephant"}]} {"id": 475651, "image": "COCO_train2014_000000475651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the adult elephant\"."}], "task": "refering", "answer_template": "The \"the adult elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 138, 139, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 161, 162, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 184, 185, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 211, 212, 213, 214, 215, 216, 217, 230, 231, 234, 235, 236, 237, 253, 254, 257, 258, 259, 260, 276, 277, 279, 280, 281, 282, 283, 299, 300, 302, 303, 304, 305, 306, 322, 323, 325, 326, 327, 328, 329, 346, 348, 349, 350, 351, 352, 369, 371, 372], "bbox": [0.0, 0.0022500000000000003, 0.691015625, 0.9865208333333333], "iscrowd": 0, "area": 115802.61475, "rle": {"size": [480, 640], "counts": "_7n1[T;VOPEj0W;jNnDV1^<0000O10O1000O100000000O1000O1000O100000O10000O1000000O1000000O10000O100mKiN]JW1b5QOWJo0g5ZOQJg0n5AkI?T6IeI7Y63_IM`6;XIFf6c0SI]Ol6k0mHUOR7S1gHmNW7\\1aHeN^7d1ZH\\Nd7m1UHSNj7U2oGkMP8Y2mGgMQ8]2lGdMS8`2jG`MU8c2iG]MV8f2hGZMX8i2dGWM\\8l2bGTM]8P3`GPM`8R3^GnLa8i400O1O1O100O1O100O1O1O100O1O100O1O1O100O1O1O100000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000O1O1O1N2O1O1O1N2O1O100O010O100O10000O100O100O100O100O100O10000O100O100O2O0O101N1O101N100O2N100O1O2O0O1O2O0O101N10001N100O2O001N101O0O2O1O0O2O0O2O1N101N101O0O2O1N101N101N2O0O2O001N2O0O2O0O2O1N101N101N2O0O2O0O2O1N101N101N2O0O2O0O2O1N2O2M2O2M3N2M3M3N1N3N2M3N2M2O2M3N2M3N1N3N2M3N2M3N1I8G9F:G9Ff0ZOlfl2"}, "label": "the adult elephant"}]} {"id": 492040, "image": "COCO_train2014_000000492040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady sitting on a stool wrapped in a red blanket\"."}], "task": "refering", "answer_template": "The \"a lady sitting on a stool wrapped in a red blanket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 346, 347, 348, 349, 350, 354, 355, 369, 370, 371, 372], "bbox": [0.013453125, 0.28950000000000004, 0.450671875, 0.9778333333333333], "iscrowd": 0, "area": 49320.62265000001, "rle": {"size": [480, 640], "counts": "Sa48d>:G9F:G9F:G9iLeNYHe1]7`N^Hj1W7[NfHn1o6XNlHR2l6PNQIY2g6iMUIa2c6aMZIh2^6ZM^Io2[6SMbIV3V6kLgI_3Q6cLkIg3l5[LRJn3f5TLVJV4b5kK\\J^4\\5dK`Jf4X5\\KeJm4R5UKkJS5o4oJmJU5S5jJkJY5T5hJhJ\\5W5dJgJ_5X5U2O1O100O2N1O1O1O1O1O1O1O1O1O100O100O100O100O100O10000000000O10001O00000000000000000000000000000000000000000000000000000010O00000000001O000000001O0000001O0000001O001O1O001O0O2M2N3M3N3L3M3M3N5J5[HXJj5m5nIYJo5n5hIVJW6o5`IWJ]6P6YIUJe6R6RISJk6R7O2M2N3M3M2O2M201O0O2O8G;F6J5J6K5J6K5K4K101N2O001N101N101O0O2O1N101O0O2O0O2O0O2O001O1N101O1O1O1O1O1O1O001O1N2O8H in this image.", "objects": [{"patches": [94, 95, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 346, 347, 348, 349, 350, 354, 355, 369, 370, 371, 372], "bbox": [0.013453125, 0.28950000000000004, 0.450671875, 0.9778333333333333], "iscrowd": 0, "area": 49320.62265000001, "rle": {"size": [480, 640], "counts": "Sa48d>:G9F:G9F:G9iLeNYHe1]7`N^Hj1W7[NfHn1o6XNlHR2l6PNQIY2g6iMUIa2c6aMZIh2^6ZM^Io2[6SMbIV3V6kLgI_3Q6cLkIg3l5[LRJn3f5TLVJV4b5kK\\J^4\\5dK`Jf4X5\\KeJm4R5UKkJS5o4oJmJU5S5jJkJY5T5hJhJ\\5W5dJgJ_5X5U2O1O100O2N1O1O1O1O1O1O1O1O1O100O100O100O100O100O10000000000O10001O00000000000000000000000000000000000000000000000000000010O00000000001O000000001O0000001O0000001O001O1O001O0O2M2N3M3N3L3M3M3N5J5[HXJj5m5nIYJo5n5hIVJW6o5`IWJ]6P6YIUJe6R6RISJk6R7O2M2N3M3M2O2M201O0O2O8G;F6J5J6K5J6K5K4K101N2O001N101N101O0O2O1N101O0O2O0O2O0O2O001O1N101O1O1O1O1O1O1O001O1N2O8H in this image.", "objects": [{"patches": [130, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 179, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 338, 339], "bbox": [0.611796875, 0.28989583333333335, 0.87303125, 0.8629166666666668], "iscrowd": 0, "area": 28241.3483, "rle": {"size": [480, 640], "counts": "nSh54g>5L4K5L4K6K4L4M3M3M3M3N2M3M3M3M3M3N2M3M3M3M4G8ZOf0\\Od0N2M3M3N2M3M3N2M3oM\\KcIg4[6`K^Ib4a6dKWI_4g6iKPIZ4n6mKlHU4R7QLhHQ4U7QLjHP4U7QLjHQ4S7QLlHP4S7RLkHP4R7RLmHP4Q7QLnHP4P7SLnHo3o6SLPIo3n6RLQIo3m6TLQIn3m6SLRIn3l6TLSIn3k6SLTIo3i6TLUIm3k6SLTIo3k6QLTIP4l6QLRIQ4m6R2O001O1N101O001O1O1O1O1N2O1O1O2N000000000O100001O001O001O001O001O001O0nNQJSIP6k6RJTIn5l6RJTIo5j6SJUIm5k6TJTIm5j6TJVIl5j6UJUIl5i6UJWIk5i6VJVIk5i6VJUIk5j6VJVIk5i6VJVIj5i6WJWIj5h6WJWIi5h6YJWIh5h6XJXIh5h6YJWIh5g6ZJXIf5h6ZJXIg5P1WI_4T1`Je5o0[I_4P1bJf5l0_I^4l7`KYH]4h7`K]H\\4d7cK`HZ4a7cKdHZ4\\7dKfH[4\\7bKgH]4Y7aKjH]4X7`KkH_4U7_KnH_4T7^KoHa4Q7]KQIc4g8N2O1N2O1O001N2O1O1O1O001O1O1O1O1O1O001O1O1O1O1N2O8H9G8H8H8H8H9G8H7I6J6J6J6J6J6J6I7JicU1"}, "label": "woman in grey plaid shirt"}]} {"id": 492040, "image": "COCO_train2014_000000492040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing checked shirt\"."}], "task": "refering", "answer_template": "The \"a woman wearing checked shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 179, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 338, 339], "bbox": [0.611796875, 0.28989583333333335, 0.87303125, 0.8629166666666668], "iscrowd": 0, "area": 28241.3483, "rle": {"size": [480, 640], "counts": "nSh54g>5L4K5L4K6K4L4M3M3M3M3N2M3M3M3M3M3N2M3M3M3M4G8ZOf0\\Od0N2M3M3N2M3M3N2M3oM\\KcIg4[6`K^Ib4a6dKWI_4g6iKPIZ4n6mKlHU4R7QLhHQ4U7QLjHP4U7QLjHQ4S7QLlHP4S7RLkHP4R7RLmHP4Q7QLnHP4P7SLnHo3o6SLPIo3n6RLQIo3m6TLQIn3m6SLRIn3l6TLSIn3k6SLTIo3i6TLUIm3k6SLTIo3k6QLTIP4l6QLRIQ4m6R2O001O1N101O001O1O1O1O1N2O1O1O2N000000000O100001O001O001O001O001O001O0nNQJSIP6k6RJTIn5l6RJTIo5j6SJUIm5k6TJTIm5j6TJVIl5j6UJUIl5i6UJWIk5i6VJVIk5i6VJUIk5j6VJVIk5i6VJVIj5i6WJWIj5h6WJWIi5h6YJWIh5h6XJXIh5h6YJWIh5g6ZJXIf5h6ZJXIg5P1WI_4T1`Je5o0[I_4P1bJf5l0_I^4l7`KYH]4h7`K]H\\4d7cK`HZ4a7cKdHZ4\\7dKfH[4\\7bKgH]4Y7aKjH]4X7`KkH_4U7_KnH_4T7^KoHa4Q7]KQIc4g8N2O1N2O1O001N2O1O1O1O001O1O1O1O1O1O001O1O1O1O1N2O8H9G8H8H8H8H9G8H7I6J6J6J6J6J6J6I7JicU1"}, "label": "a woman wearing checked shirt"}]} {"id": 492040, "image": "COCO_train2014_000000492040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in grey sweater\"."}], "task": "refering", "answer_template": "The \"man in grey sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 125, 126, 127, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 217, 219], "bbox": [0.38851562500000003, 0.23422916666666668, 0.584453125, 0.5946041666666666], "iscrowd": 0, "area": 11401.744350000004, "rle": {"size": [480, 640], "counts": "jmd35f>6J6I6M4N1O2N2N1O2N1O2N1N2O1O1O2N1O1O1O1O1O1O1N2O1O1O1O1O2N2N3M2N8GS1nN3M1O1O1O1000001O0O100000O1N2N2N2N2B=0100O1O10UOj0N3L3N3M210O010O10000O100O10000O10000O100O11O001O00001O0010O1OiLZEU2f:gM^EX2c:eM`E[2`:bMcE^2]:^MgEb2Y:[MjEe2V:XMmEh2S:UMPF\\3_9`LeFf3U9WLnFk3S:O1O1O1O1UOiDkMX;Q2WEcMi:[2dEYM^:c2R1L4N2N1O2N2N1O2N2M3N2N5K5K5K6J5K5K5Kd_l3"}, "label": "man in grey sweater"}]} {"id": 492040, "image": "COCO_train2014_000000492040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man stirring a pot of food\"."}], "task": "refering", "answer_template": "The \"a man stirring a pot of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 125, 126, 127, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 193, 194, 195, 196, 197, 217, 219], "bbox": [0.38851562500000003, 0.23422916666666668, 0.584453125, 0.5946041666666666], "iscrowd": 0, "area": 11401.744350000004, "rle": {"size": [480, 640], "counts": "jmd35f>6J6I6M4N1O2N2N1O2N1O2N1N2O1O1O2N1O1O1O1O1O1O1N2O1O1O1O1O2N2N3M2N8GS1nN3M1O1O1O1000001O0O100000O1N2N2N2N2B=0100O1O10UOj0N3L3N3M210O010O10000O100O10000O10000O100O11O001O00001O0010O1OiLZEU2f:gM^EX2c:eM`E[2`:bMcE^2]:^MgEb2Y:[MjEe2V:XMmEh2S:UMPF\\3_9`LeFf3U9WLnFk3S:O1O1O1O1UOiDkMX;Q2WEcMi:[2dEYM^:c2R1L4N2N1O2N2N1O2N2M3N2N5K5K5K6J5K5K5Kd_l3"}, "label": "a man stirring a pot of food"}]} {"id": 156170, "image": "COCO_train2014_000000156170.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on right wearing white shirt and orange shorts\"."}], "task": "refering", "answer_template": "The \"man on right wearing white shirt and orange shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 158, 159, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 271, 272, 273, 274, 294, 295, 296, 297, 317, 318, 319, 320, 340, 341, 342, 343], "bbox": [0.7360781249999999, 0.4031455399061033, 0.9984375, 0.9887323943661973], "iscrowd": 0, "area": 20832.232, "rle": {"size": [426, 640], "counts": "RUT62X=2M4M3L3N3L2O1N100O2O0O101N101N100O2O0O101N1N3M2N2O2M2N2N3N1N2N3M2O2M2N201O00010O001hHPNT3Q1kHLd3D]3=RIOU34e3KWI3g2b0n3YO]I5X2S1W4eNdI8i1c1_4SNjI;^1n1d4dMQJ>X1S2f4\\MTJb0R1U2j4VMVJf0n0V2l4RMXJh0i0Y2o4lL[Jo0?X2V5gL]JU16W2]5aL`J]1KU2e5\\LbJd1AS2l6SNlHP2T7S200000000O101O0000WOlH\\KT7d4QIWKn6j4XIPKh6Q5\\IjJd6U5i0O1N2O0O2O1NXOi0O1N101O0O2O0011O12O1N2N2N2O1N2N2N2O1N2N2N2O1N5K6J6K6I1O0001O00010O0000010O00000010O0000aIUJ_5l5\\JYJd5f5XJ_Jg5a5UJcJk5]5QJhJn5Y5mIkJS6U5iIPKW6R6O001O10OgL]JkNd5a0nK`NS4m0^401O1O0O2O1O001O1N101O001O1O0O2O1O001O1N101O1O001N2O001O001N2H7@X6"}, "label": "man on right wearing white shirt and orange shorts"}]} {"id": 156170, "image": "COCO_train2014_000000156170.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fat person with white tshirt throwing an yellow disk\"."}], "task": "refering", "answer_template": "The \"a fat person with white tshirt throwing an yellow disk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 158, 159, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 271, 272, 273, 274, 294, 295, 296, 297, 317, 318, 319, 320, 340, 341, 342, 343], "bbox": [0.7360781249999999, 0.4031455399061033, 0.9984375, 0.9887323943661973], "iscrowd": 0, "area": 20832.232, "rle": {"size": [426, 640], "counts": "RUT62X=2M4M3L3N3L2O1N100O2O0O101N101N100O2O0O101N1N3M2N2O2M2N2N3N1N2N3M2O2M2N201O00010O001hHPNT3Q1kHLd3D]3=RIOU34e3KWI3g2b0n3YO]I5X2S1W4eNdI8i1c1_4SNjI;^1n1d4dMQJ>X1S2f4\\MTJb0R1U2j4VMVJf0n0V2l4RMXJh0i0Y2o4lL[Jo0?X2V5gL]JU16W2]5aL`J]1KU2e5\\LbJd1AS2l6SNlHP2T7S200000000O101O0000WOlH\\KT7d4QIWKn6j4XIPKh6Q5\\IjJd6U5i0O1N2O0O2O1NXOi0O1N101O0O2O0011O12O1N2N2N2O1N2N2N2O1N2N2N2O1N5K6J6K6I1O0001O00010O0000010O00000010O0000aIUJ_5l5\\JYJd5f5XJ_Jg5a5UJcJk5]5QJhJn5Y5mIkJS6U5iIPKW6R6O001O10OgL]JkNd5a0nK`NS4m0^401O1O0O2O1O001O1N101O001O1O0O2O1O001O1N101O1O001N2O001O001N2H7@X6"}, "label": "a fat person with white tshirt throwing an yellow disk"}]} {"id": 401962, "image": "COCO_train2014_000000401962.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue chair with a boy sitting in it\"."}], "task": "refering", "answer_template": "The \"the blue chair with a boy sitting in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 18, 19, 20, 36, 37, 38, 54, 55, 56, 72, 73, 74, 75, 90, 91, 92, 93, 94, 108, 109, 110, 111, 112, 126, 127, 128, 129, 130, 145, 146, 147, 163, 164, 165, 182, 183, 201, 202, 219, 220], "bbox": [0.0, 0.0, 0.25504, 0.9623999999999999], "iscrowd": 0, "area": 25480.599400000003, "rle": {"size": [375, 500], "counts": "<[5[63N1N3N2N2M3N1N3N2N2M2O2M3N2M2O2N2M3N1N3N2N2M2O2M3N2M2O2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N1O2N1O2O0O2N1O2N1O2N1O2N1O2N1OWI`M]4^2YKPNd4o1YKXNe4f1YK_Ne4`1XKgNf4W1WKPOf4o0XKVOg4h0WK]Og4b0WKBi4 in this image.", "objects": [{"patches": [0, 1, 2, 3, 18, 19, 20, 36, 37, 38, 54, 55, 56, 72, 73, 74, 75, 90, 91, 92, 93, 94, 108, 109, 110, 111, 112, 126, 127, 128, 129, 130, 145, 146, 147, 163, 164, 165, 182, 183, 201, 202, 219, 220], "bbox": [0.0, 0.0, 0.25504, 0.9623999999999999], "iscrowd": 0, "area": 25480.599400000003, "rle": {"size": [375, 500], "counts": "<[5[63N1N3N2N2M3N1N3N2N2M2O2M3N2M2O2N2M3N1N3N2N2M2O2M3N2M2O2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N1O2N1O2O0O2N1O2N1O2N1O2N1O2N1OWI`M]4^2YKPNd4o1YKXNe4f1YK_Ne4`1XKgNf4W1WKPOf4o0XKVOg4h0WK]Og4b0WKBi4 in this image.", "objects": [{"patches": [20, 21, 22, 23, 37, 38, 39, 40, 41, 42, 43, 44, 45, 56, 57, 58, 59, 60, 61, 62, 63, 75, 76, 77, 78, 92, 93, 94, 95, 96, 112, 113, 114, 129, 130, 131, 132, 148, 149, 150, 165, 166, 167, 168, 183, 184, 185, 186], "bbox": [0.09822, 0.1161271676300578, 0.5238400000000001, 0.8752601156069365], "iscrowd": 0, "area": 21098.32469999999, "rle": {"size": [346, 500], "counts": "Sd`01e:4M3N2O2N100O1O2N1O1O1O2N1O1O1O2M2O1O1O2O0O2O1O2N1O1N3YHhNh5Y1VJlNf5U1XJPOd5R1YJSOc5n0[JWOa5j0]J^O[5d0bJ^O^5b0aJ_O_5b0_J_Oa5a0^JAa5?^JBb5?\\JBd5>[JCd5>\\JAe5`0YJAg5?XJAi5?VJAk5?TJAm5`0RJ_Oo5a0PJ_OP6b0oI^OR6c0lI]OU6c0jI]OW6c0gI^OZ6c0cI^O^6b0aI^O_6c0_I^Ob6\\21OQM_LgN`3X1bLhN^3V1dLjN[3gNcLe04d0Y3fNfLb04f0V3gNkL>1k0T3eNoL;0P1P3eNTM6OU1l2cNZM3LZ1j2bN]M0K^1g2aNcMLIb1]2gNnMROoNIg0`2^2aNM_16\\NKe18VNIl1:kMKU2;aMH`2>UMFl2T41N1O1O2N1N2O1O2N1O1O2N1O1N2G:H7K6J6J5L5M2N3M3M2N3N1N3M3M2NULmJT2P5iMYKS2f4jMbKS2_4fMiKX2\\4^MkK_2[4VMmKh2X4nLoKo2o5000N101N2O1N2O1N101mN_LjJc3T5cLgJ^3W5iLbJY3\\5VMVJj2j5S1O1O2N1^KRKW3o4iLVKQ3l4mL[Kl2f4SM`Kf2b4ZMXLk1i3TNYLi1j3VNXLf1j3YNXLe1i3[NXLb1j3]NXL`1k3_NXL\\1j3cN[LV1h3jN[LQ1h3mN]Ll0f3TO^Le0e3ZO`L`0b3@bL4g3Kd30001O0000001N1000001O00001O000O101O0000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000001O00000001N2N2M3N6JbZ`2"}, "label": "a orange cat standing on it hind legs playing with another cat"}]} {"id": 82484, "image": "COCO_train2014_000000082484.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is an orange cat standing up\"."}], "task": "refering", "answer_template": "The \"this is an orange cat standing up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 23, 37, 38, 39, 40, 41, 42, 43, 44, 45, 56, 57, 58, 59, 60, 61, 62, 63, 75, 76, 77, 78, 92, 93, 94, 95, 96, 112, 113, 114, 129, 130, 131, 132, 148, 149, 150, 165, 166, 167, 168, 183, 184, 185, 186], "bbox": [0.09822, 0.1161271676300578, 0.5238400000000001, 0.8752601156069365], "iscrowd": 0, "area": 21098.32469999999, "rle": {"size": [346, 500], "counts": "Sd`01e:4M3N2O2N100O1O2N1O1O1O2N1O1O1O2M2O1O1O2O0O2O1O2N1O1N3YHhNh5Y1VJlNf5U1XJPOd5R1YJSOc5n0[JWOa5j0]J^O[5d0bJ^O^5b0aJ_O_5b0_J_Oa5a0^JAa5?^JBb5?\\JBd5>[JCd5>\\JAe5`0YJAg5?XJAi5?VJAk5?TJAm5`0RJ_Oo5a0PJ_OP6b0oI^OR6c0lI]OU6c0jI]OW6c0gI^OZ6c0cI^O^6b0aI^O_6c0_I^Ob6\\21OQM_LgN`3X1bLhN^3V1dLjN[3gNcLe04d0Y3fNfLb04f0V3gNkL>1k0T3eNoL;0P1P3eNTM6OU1l2cNZM3LZ1j2bN]M0K^1g2aNcMLIb1]2gNnMROoNIg0`2^2aNM_16\\NKe18VNIl1:kMKU2;aMH`2>UMFl2T41N1O1O2N1N2O1O2N1O1O2N1O1N2G:H7K6J6J5L5M2N3M3M2N3N1N3M3M2NULmJT2P5iMYKS2f4jMbKS2_4fMiKX2\\4^MkK_2[4VMmKh2X4nLoKo2o5000N101N2O1N2O1N101mN_LjJc3T5cLgJ^3W5iLbJY3\\5VMVJj2j5S1O1O2N1^KRKW3o4iLVKQ3l4mL[Kl2f4SM`Kf2b4ZMXLk1i3TNYLi1j3VNXLf1j3YNXLe1i3[NXLb1j3]NXL`1k3_NXL\\1j3cN[LV1h3jN[LQ1h3mN]Ll0f3TO^Le0e3ZO`L`0b3@bL4g3Kd30001O0000001N1000001O00001O000O101O0000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000001O00000001N2N2M3N6JbZ`2"}, "label": "this is an orange cat standing up"}]} {"id": 82484, "image": "COCO_train2014_000000082484.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white cat with grey patterns\"."}], "task": "refering", "answer_template": "The \"a white cat with grey patterns\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 123, 124, 133, 134, 135, 136, 137, 138, 140, 141, 152, 153, 154, 155, 156, 157, 158, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 205, 206, 209], "bbox": [0.41372000000000003, 0.5161271676300578, 0.9018400000000001, 0.9462427745664739], "iscrowd": 0, "area": 13330.539800000002, "rle": {"size": [346, 500], "counts": "RUV21h:2N2O1N2N1O2\\HFf4;XKFg4 in this image.", "objects": [{"patches": [116, 117, 123, 124, 133, 134, 135, 136, 137, 138, 140, 141, 152, 153, 154, 155, 156, 157, 158, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 205, 206, 209], "bbox": [0.41372000000000003, 0.5161271676300578, 0.9018400000000001, 0.9462427745664739], "iscrowd": 0, "area": 13330.539800000002, "rle": {"size": [346, 500], "counts": "RUV21h:2N2O1N2N1O2\\HFf4;XKFg4 in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 56, 57, 61, 62, 63, 73, 74, 75, 76, 77, 78, 79, 80, 82, 83, 84, 85, 86, 87, 96, 97, 98, 99, 100, 101, 102, 103, 105, 106, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 170, 171, 172, 173, 174, 176, 177, 178, 179, 193, 194, 195, 196, 197, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 225, 226, 238, 239, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 266, 283, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 331, 332, 333, 334, 335, 357, 358], "bbox": [0.16665625, 0.1514908256880734, 0.863984375, 1.0], "iscrowd": 0, "area": 59795.65955000002, "rle": {"size": [436, 640], "counts": "UR^11b=101eB5k`CC_<=aCE\\<m2@TM\\KJj4R3IUM[KLk4P3ITMZKOl4m2IUMYK1m4k2HcM6]2JeM5[2KeM5[2JgM5Y2KgM5Y2JiM5W2KiM5W2JkM5U2KkM4V2KkM5T2LmM3S2LnM4R2LnM4R2KoM5Q2KoM5Q2JQN5o1KQN5o1JRN6n1JRN6m1JTN6l1JTN6l1IVN5k1KUN5k1KUN5k1JWN5i1KWN5i1KXN4h1KYN5f1L[N3e1M[N3e1L]N3c1M]N3c1K`N4`1KbN4^1KcN5]1JeN4\\1KeN5[1JgN5X1JjN6V1IlN6T1ImN7S1HnN8R1GoN9Q1FPO:P1DSO;l0EUO;k0DVOf0@\\O`0c0@^O`0b0_O_Oa0a0^OAa0a0[OAe0a0UOCk0>oNGQ1;iNJV17eNM[1b0QNCo1g0cM]O]2e0]M_Oc2d0VMAi2b0oLCQ3?^L0b3Q501O0RMaFg1`9UNcFk1]9QNfFP2Z9nMgFS2Z9jMhFV2X9hMiFZ2W9cMjF^2V9`MlF`2U9TMUGm2i9000O1000000O100000000AnL]FQ3a9QM_Fo2_9SMaFm2]9UMcFk2\\9WMdFh2Z9ZMfFg2W9\\MhFe2U9]MkFc2T9^MlFc2S9^MmFb2R9^MnFc2Q9^MnFc2Q9]MoFd2o8]MQGc2o8^MoFd2P9\\MoFf2P9[MmFh2R9XMmFj2R9VMnFk2Q9VMnFj2R9VMnFk2Q9VMnFk2Q9UMoFl2j3aLS1c0SKm2e3eLW1?RKn2]3nL_14TKn2P3\\Mk1GTKn2d2hMV2[OVKn2X2SNa2oNWKo2o1[Ni2b2oLfMP3[2gLoMW3R2`LWN_3i1YL`Nf3a1QLhNn3Y1iKPOU4X501O1O1N2O1O001NO2J6I7I6K7H9G9G9K4VJmIm3f5`JhJ^1Eo3e5fJfJV1HP4e5mJcJl0MT4b5SK`Jf00T4b5XK_J`02S4b5`K\\J94T4b5fKYJ19U4b5lKUJG?\\4\\5PLUJYOh0f4S5TL\\Kj3f4XLYKe3h4^LVK`3l4bLRK]3n4fLPKX3Q5jLnJT3T5nLjJP3W5SMgJk2[5WMdJf2]5]MaJb2`5^M`J`2P4\\LlKV13\\2k3WM`K?d0X2m3[M^K>d0U2n3`M\\K=e0R2c2[L_MZ1WO:f0o1b2hLXMR1^O9g0k1`2WMRMg0E9h0g1c1aNfMCM6i0d1d1h0aMeMk0b1c1l0aMcMk0_1d1U3[NiLe1Z3ZNdLe1`3ZN^Lf1e3YNYLg1j3XNULg1n3XNPLh1S4UNmKm1S4QNmKP2U4mMkKT2W4iMjKX2W4eMiK\\2Y4cMeK^2]4bM`K`2a4_M_K`2c4_M]K`2e4_MZKa2i4]MWKc2j4]MTKc2n4\\MRKc2P5\\MPKd2o7010O01000O101N2O001N2O1N2O001N2O1N101O1N2O1O001O1N4M2N2N2N2N3L3N2N2N3MkLaF[2Y9dMiFa2Q9_MoFb2o8^MSGa2m8^MUGa2k8_MVG`2j8_MXG`2h8`MXG`2h8_MZGa2d8`M]G_2c8`M_G_2a8aM_G_2a8`MaG_2_8aMbG_2]8aMdG^2\\8bMdG^2\\8bMfG\\2Z8dMgG[2Y8eMhGZ2X8fMiGZ2V8fMkGY2U8gMmGW2S8iMnGV2R8jMoGU2R8jMoGV2P8jMQHU2o7kMSHS2m7mMTHR2l7nMUHQ2k7oMVHP2j7PNWHP2h7PNZHn1f7RN[Hm1e7SN\\Hl1e7SN\\Hl1d7TN]Hl1b7TN`Hj1`7VNaHi1_7XNaHg1_7YNbHg1]7ZNcHe1]7\\NdHc1[7]NfHc1Y7^NgHa1Z7^NgHb1X7_NhH`1X7aNhH_1W7aNkH]1U7dNkH\\1T7dNmH\\1R7eNnHZ1R7gNnHY1R7fNoHY1R7gNoHX1P7hNQIX1o6hNQIW1o6jNQIV1o6iNRIV1n6kNRIU1n6kNSIS1m6nNSIR1m6oNRIQ1m6POSIo0n6QORIo0m6ROSIJXN4f82SIHZN4c85RIG]N1a89RIF]N1a89RIE^N1`8;RID^N1`8 in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 56, 57, 61, 62, 63, 73, 74, 75, 76, 77, 78, 79, 80, 82, 83, 84, 85, 86, 87, 96, 97, 98, 99, 100, 101, 102, 103, 105, 106, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 170, 171, 172, 173, 174, 176, 177, 178, 179, 193, 194, 195, 196, 197, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 225, 226, 238, 239, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 266, 283, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 331, 332, 333, 334, 335, 357, 358], "bbox": [0.16665625, 0.1514908256880734, 0.863984375, 1.0], "iscrowd": 0, "area": 59795.65955000002, "rle": {"size": [436, 640], "counts": "UR^11b=101eB5k`CC_<=aCE\\<m2@TM\\KJj4R3IUM[KLk4P3ITMZKOl4m2IUMYK1m4k2HcM6]2JeM5[2KeM5[2JgM5Y2KgM5Y2JiM5W2KiM5W2JkM5U2KkM4V2KkM5T2LmM3S2LnM4R2LnM4R2KoM5Q2KoM5Q2JQN5o1KQN5o1JRN6n1JRN6m1JTN6l1JTN6l1IVN5k1KUN5k1KUN5k1JWN5i1KWN5i1KXN4h1KYN5f1L[N3e1M[N3e1L]N3c1M]N3c1K`N4`1KbN4^1KcN5]1JeN4\\1KeN5[1JgN5X1JjN6V1IlN6T1ImN7S1HnN8R1GoN9Q1FPO:P1DSO;l0EUO;k0DVOf0@\\O`0c0@^O`0b0_O_Oa0a0^OAa0a0[OAe0a0UOCk0>oNGQ1;iNJV17eNM[1b0QNCo1g0cM]O]2e0]M_Oc2d0VMAi2b0oLCQ3?^L0b3Q501O0RMaFg1`9UNcFk1]9QNfFP2Z9nMgFS2Z9jMhFV2X9hMiFZ2W9cMjF^2V9`MlF`2U9TMUGm2i9000O1000000O100000000AnL]FQ3a9QM_Fo2_9SMaFm2]9UMcFk2\\9WMdFh2Z9ZMfFg2W9\\MhFe2U9]MkFc2T9^MlFc2S9^MmFb2R9^MnFc2Q9^MnFc2Q9]MoFd2o8]MQGc2o8^MoFd2P9\\MoFf2P9[MmFh2R9XMmFj2R9VMnFk2Q9VMnFj2R9VMnFk2Q9VMnFk2Q9UMoFl2j3aLS1c0SKm2e3eLW1?RKn2]3nL_14TKn2P3\\Mk1GTKn2d2hMV2[OVKn2X2SNa2oNWKo2o1[Ni2b2oLfMP3[2gLoMW3R2`LWN_3i1YL`Nf3a1QLhNn3Y1iKPOU4X501O1O1N2O1O001NO2J6I7I6K7H9G9G9K4VJmIm3f5`JhJ^1Eo3e5fJfJV1HP4e5mJcJl0MT4b5SK`Jf00T4b5XK_J`02S4b5`K\\J94T4b5fKYJ19U4b5lKUJG?\\4\\5PLUJYOh0f4S5TL\\Kj3f4XLYKe3h4^LVK`3l4bLRK]3n4fLPKX3Q5jLnJT3T5nLjJP3W5SMgJk2[5WMdJf2]5]MaJb2`5^M`J`2P4\\LlKV13\\2k3WM`K?d0X2m3[M^K>d0U2n3`M\\K=e0R2c2[L_MZ1WO:f0o1b2hLXMR1^O9g0k1`2WMRMg0E9h0g1c1aNfMCM6i0d1d1h0aMeMk0b1c1l0aMcMk0_1d1U3[NiLe1Z3ZNdLe1`3ZN^Lf1e3YNYLg1j3XNULg1n3XNPLh1S4UNmKm1S4QNmKP2U4mMkKT2W4iMjKX2W4eMiK\\2Y4cMeK^2]4bM`K`2a4_M_K`2c4_M]K`2e4_MZKa2i4]MWKc2j4]MTKc2n4\\MRKc2P5\\MPKd2o7010O01000O101N2O001N2O1N2O001N2O1N101O1N2O1O001O1N4M2N2N2N2N3L3N2N2N3MkLaF[2Y9dMiFa2Q9_MoFb2o8^MSGa2m8^MUGa2k8_MVG`2j8_MXG`2h8`MXG`2h8_MZGa2d8`M]G_2c8`M_G_2a8aM_G_2a8`MaG_2_8aMbG_2]8aMdG^2\\8bMdG^2\\8bMfG\\2Z8dMgG[2Y8eMhGZ2X8fMiGZ2V8fMkGY2U8gMmGW2S8iMnGV2R8jMoGU2R8jMoGV2P8jMQHU2o7kMSHS2m7mMTHR2l7nMUHQ2k7oMVHP2j7PNWHP2h7PNZHn1f7RN[Hm1e7SN\\Hl1e7SN\\Hl1d7TN]Hl1b7TN`Hj1`7VNaHi1_7XNaHg1_7YNbHg1]7ZNcHe1]7\\NdHc1[7]NfHc1Y7^NgHa1Z7^NgHb1X7_NhH`1X7aNhH_1W7aNkH]1U7dNkH\\1T7dNmH\\1R7eNnHZ1R7gNnHY1R7fNoHY1R7gNoHX1P7hNQIX1o6hNQIW1o6jNQIV1o6iNRIV1n6kNRIU1n6kNSIS1m6nNSIR1m6oNRIQ1m6POSIo0n6QORIo0m6ROSIJXN4f82SIHZN4c85RIG]N1a89RIF]N1a89RIE^N1`8;RID^N1`8 in this image.", "objects": [{"patches": [115, 116, 138, 139, 140, 146, 147, 161, 162, 163, 164, 167, 168, 169, 170, 186, 187, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 300, 301, 302, 303, 324, 325], "bbox": [0.00146875, 0.3397935779816514, 0.40875000000000006, 0.9419266055045871], "iscrowd": 0, "area": 21762.643650000005, "rle": {"size": [436, 640], "counts": "^c07Q=?A?@?M20O01O001O010O001O0010O01[G]NW5c1jJ`NR5a1mJbNP5^1PKfNm4Y1SKjNj4W1UKlNh4T1YKmNe4T1ZKoNc4Q1]KROa4n0^KUO_4k0bKUO]4l0bKVO\\4j0dKXO[4h0dKYO[4g0eK[OY4f0gK[OW4e0iK\\OV4d0kK[OV4d0kK\\OT4d0mK\\OR4d0oK[OR4d0oK\\OP4d0QL\\Oo3c0RL\\On3d0SL\\Om3c0TL]Ok3c0WL[Oj3d0WL\\Oh3d0YL[Oh3d0YL\\Of3d0\\L[Od3d0]L[Od3d0]L\\Ob3d0_L\\Oa3c0aL\\O^3d0cL\\O]3c0dL\\O\\3d0eL\\O[3c0gL\\OX3g0fLYOZ3m0`LSO_3T1[LlNe3Z1TLfNoLIn5h1lL^OU3h0cLYO]3n0[LQOh3U1PLkNQ4X1kKgNX4Z1eKeN]4]1`KcNb4^1XKbNm4_1lJcNW5_1cJbNa5_1XJcNl5^1QJ`NS6a1jI^NZ6b1cI]N`6d1]I\\Nf6d1VI^Nk6c1RI]NP7d1mH]NU7f1eH[N\\7j1]HVNe7l32M2O2M2O200e0ZO6K3MN1O2N2O0O2N2N8H9G9ZKVHP3T8_LSH^3S9N2M3M3M3M3L4J6K5N20000000000000000O100SOZElNf:P1^EPOb:k0dETO\\:k0eEUO[:i0gEWOX:h0jEXOV:f0mEYOS:e0oE[OQ:d0PF\\Om9e0UF[Oh9f0[FYOb9i0_FWO^9k0hFPOT9S1QGiNl8Y1VGfNg8\\1\\GbNa8`1aG_N\\8c0XFO_1\\OU8f0_FM_1[Oo7g0fFM]1[Oj7h0lFL]1ZOd7k0QGK]1XO_7P1TGHf:;[EEc:<^EDa:=_ECa:=_EC`:>`EB`:=aEC_:=aEC^:>bEA_:?aEA^:`0bE@^:`0bE@^:?cEA\\:`0dE_O]:a0cE_O\\:b0dE]O]:c0dE\\O\\:d0eEZO[:f0fEZOZ:f0gEXOY:i0hEUOY:k0hETOX:l0iEROW:o0jEPOV:o0kEPOU:Q1lEnNT:R1mElNT:T1mEkNR:V1oEhNR:X1PFfNo9Z1V1000O100O10O01O0010O01O010O001O010O0010O01O1O010O001O010O0100O010O10O01O010O10O0100O010O10O010O01O010O01O000001O0000010O0000001O01O1O00100O1O1O10O01O1O100O001O1001N2N3M2N2O1N2N3M2iNhDOmSQ5"}, "label": "the second bird from the front"}]} {"id": 533050, "image": "COCO_train2014_000000533050.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bird flying back\"."}], "task": "refering", "answer_template": "The \"bird flying back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 138, 139, 140, 146, 147, 161, 162, 163, 164, 167, 168, 169, 170, 186, 187, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 300, 301, 302, 303, 324, 325], "bbox": [0.00146875, 0.3397935779816514, 0.40875000000000006, 0.9419266055045871], "iscrowd": 0, "area": 21762.643650000005, "rle": {"size": [436, 640], "counts": "^c07Q=?A?@?M20O01O001O010O001O0010O01[G]NW5c1jJ`NR5a1mJbNP5^1PKfNm4Y1SKjNj4W1UKlNh4T1YKmNe4T1ZKoNc4Q1]KROa4n0^KUO_4k0bKUO]4l0bKVO\\4j0dKXO[4h0dKYO[4g0eK[OY4f0gK[OW4e0iK\\OV4d0kK[OV4d0kK\\OT4d0mK\\OR4d0oK[OR4d0oK\\OP4d0QL\\Oo3c0RL\\On3d0SL\\Om3c0TL]Ok3c0WL[Oj3d0WL\\Oh3d0YL[Oh3d0YL\\Of3d0\\L[Od3d0]L[Od3d0]L\\Ob3d0_L\\Oa3c0aL\\O^3d0cL\\O]3c0dL\\O\\3d0eL\\O[3c0gL\\OX3g0fLYOZ3m0`LSO_3T1[LlNe3Z1TLfNoLIn5h1lL^OU3h0cLYO]3n0[LQOh3U1PLkNQ4X1kKgNX4Z1eKeN]4]1`KcNb4^1XKbNm4_1lJcNW5_1cJbNa5_1XJcNl5^1QJ`NS6a1jI^NZ6b1cI]N`6d1]I\\Nf6d1VI^Nk6c1RI]NP7d1mH]NU7f1eH[N\\7j1]HVNe7l32M2O2M2O200e0ZO6K3MN1O2N2O0O2N2N8H9G9ZKVHP3T8_LSH^3S9N2M3M3M3M3L4J6K5N20000000000000000O100SOZElNf:P1^EPOb:k0dETO\\:k0eEUO[:i0gEWOX:h0jEXOV:f0mEYOS:e0oE[OQ:d0PF\\Om9e0UF[Oh9f0[FYOb9i0_FWO^9k0hFPOT9S1QGiNl8Y1VGfNg8\\1\\GbNa8`1aG_N\\8c0XFO_1\\OU8f0_FM_1[Oo7g0fFM]1[Oj7h0lFL]1ZOd7k0QGK]1XO_7P1TGHf:;[EEc:<^EDa:=_ECa:=_EC`:>`EB`:=aEC_:=aEC^:>bEA_:?aEA^:`0bE@^:`0bE@^:?cEA\\:`0dE_O]:a0cE_O\\:b0dE]O]:c0dE\\O\\:d0eEZO[:f0fEZOZ:f0gEXOY:i0hEUOY:k0hETOX:l0iEROW:o0jEPOV:o0kEPOU:Q1lEnNT:R1mElNT:T1mEkNR:V1oEhNR:X1PFfNo9Z1V1000O100O10O01O0010O01O010O001O010O0010O01O1O010O001O010O0100O010O10O01O010O10O0100O010O10O010O01O010O01O000001O0000010O0000001O01O1O00100O1O1O10O01O1O100O001O1001N2N3M2N2O1N2N3M2iNhDOmSQ5"}, "label": "bird flying back"}]} {"id": 401982, "image": "COCO_train2014_000000401982.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with his hand touching his face\"."}], "task": "refering", "answer_template": "The \"the man with his hand touching his face\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 44, 65, 66, 67, 88, 89, 90, 111, 112, 113, 134, 135, 136, 157, 158, 159, 181, 182, 204, 205, 227, 228, 249, 250, 251], "bbox": [0.8455625, 0.08315028901734103, 0.9561093749999999, 0.9033815028901734], "iscrowd": 0, "area": 13506.817500000003, "rle": {"size": [346, 640], "counts": "Skf51h:1hF2c70YH:_7H^Ha0^OAm6OcIh0WO@T4EWN5ZNk1Y3QN\\N5YNP2V3lM`N6WNT2T3gMdN6VN[2P3_MhN9TN`2n2XMmN9QNh2l2PMkLHk1b0VNS3n2cLPMIk1i4o0_KVMHj1o4j0ZK[MHi1V5e0SKaMHi1[5a0mJeMIi1`5 in this image.", "objects": [{"patches": [42, 43, 44, 65, 66, 67, 88, 89, 90, 111, 112, 113, 134, 135, 136, 157, 158, 159, 181, 182, 204, 205, 227, 228, 249, 250, 251], "bbox": [0.8455625, 0.08315028901734103, 0.9561093749999999, 0.9033815028901734], "iscrowd": 0, "area": 13506.817500000003, "rle": {"size": [346, 640], "counts": "Skf51h:1hF2c70YH:_7H^Ha0^OAm6OcIh0WO@T4EWN5ZNk1Y3QN\\N5YNP2V3lM`N6WNT2T3gMdN6VN[2P3_MhN9TN`2n2XMmN9QNh2l2PMkLHk1b0VNS3n2cLPMIk1i4o0_KVMHj1o4j0ZK[MHi1V5e0SKaMHi1[5a0mJeMIi1`5 in this image.", "objects": [{"patches": [25, 48, 49, 50, 70, 71, 72, 73, 95, 96, 97, 118, 119, 120, 142, 143, 166, 167], "bbox": [0.065359375, 0.1545086705202312, 0.26890625, 0.6381213872832371], "iscrowd": 0, "area": 6150.783100000001, "rle": {"size": [346, 640], "counts": "`Y>2g:20000HMhE2X:NhE2X:OgE0Z:1cE0^:0`E2a:21O100000000001O00000O100POI]G8a8I^G8b8H^G8a8I_G7Z8IkF1k06X89aGH^88bGH]89cGG\\8=aGC_8?_GA`8b0]G_Ob8b0^G]Oc8d0\\G\\Od8d0\\G\\Od8d0[G\\Oe8e0[G[Oe8e0ZG\\Of8d0ZG\\Of8d0YG\\Oh8d0XG\\Oh8d0RGCl8`0QGAP9U10O010O1N1O2N207J5KO1N1O2N2N2O1QHaN_6a1YIfNf6\\1RIjNn6W1kHPOT7Q1eHUO\\7k0aHXO^7i0_HYOa7h0]HZOb7g0[H[Of7e0XH]Oh7j1010O100O100O010O100O100O100O1000000O2O1O1O1O1O001O1O1O1UOlGmNU8m0oGTOU8c0PHGg72^H1`7^OQIb0o6\\ORId0n6[OTIc0V7ROlHm0b8O\\FTOZ9k0cFYO^9Q110O01O1O001O001O1O0WOZF>g9_O\\F?f9_O\\F`0d9^O_Fa0n9O1N2O1O1O1N2O1O5Kcmm4"}, "label": "a baseball player in a batting helmet about to run"}]} {"id": 401982, "image": "COCO_train2014_000000401982.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ball player running\"."}], "task": "refering", "answer_template": "The \"a ball player running\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 48, 49, 50, 70, 71, 72, 73, 95, 96, 97, 118, 119, 120, 142, 143, 166, 167], "bbox": [0.065359375, 0.1545086705202312, 0.26890625, 0.6381213872832371], "iscrowd": 0, "area": 6150.783100000001, "rle": {"size": [346, 640], "counts": "`Y>2g:20000HMhE2X:NhE2X:OgE0Z:1cE0^:0`E2a:21O100000000001O00000O100POI]G8a8I^G8b8H^G8a8I_G7Z8IkF1k06X89aGH^88bGH]89cGG\\8=aGC_8?_GA`8b0]G_Ob8b0^G]Oc8d0\\G\\Od8d0\\G\\Od8d0[G\\Oe8e0[G[Oe8e0ZG\\Of8d0ZG\\Of8d0YG\\Oh8d0XG\\Oh8d0RGCl8`0QGAP9U10O010O1N1O2N207J5KO1N1O2N2N2O1QHaN_6a1YIfNf6\\1RIjNn6W1kHPOT7Q1eHUO\\7k0aHXO^7i0_HYOa7h0]HZOb7g0[H[Of7e0XH]Oh7j1010O100O100O010O100O100O100O1000000O2O1O1O1O1O001O1O1O1UOlGmNU8m0oGTOU8c0PHGg72^H1`7^OQIb0o6\\ORId0n6[OTIc0V7ROlHm0b8O\\FTOZ9k0cFYO^9Q110O01O1O001O001O1O0WOZF>g9_O\\F?f9_O\\F`0d9^O_Fa0n9O1N2O1O1O1N2O1O5Kcmm4"}, "label": "a ball player running"}]} {"id": 401982, "image": "COCO_train2014_000000401982.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wicket keeper wearing green t - shirt and green cap standing in the middle of other two playes\"."}], "task": "refering", "answer_template": "The \"a wicket keeper wearing green t - shirt and green cap standing in the middle of other two playes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 53, 54, 75, 76, 77, 98, 99, 100, 122, 123, 144, 145, 146, 167, 168, 169, 190, 191, 192], "bbox": [0.269109375, 0.09734104046242774, 0.37901562499999997, 0.7409537572254336], "iscrowd": 0, "area": 8319.162649999998, "rle": {"size": [346, 640], "counts": "a[j11i:001N101O1N1nM6TIKj6>mHCS7d0cH_O]7i0YHYOg7n0oGUOQ8R1eGPOW4_OHX27jMHV28jMoMN^OX2b2kMjM8_Om1g2lMcM`0Cc1j2PN[Md0J\\1j2UNVMa00Z1i2XNSMa0POPOg0W2T3]NoL`0SOnNj0U2S3`NlL?UOnNl0R2S3dNgL`0WOlNo0P2R3gNdLi0:`0Q3iNaLk0= in this image.", "objects": [{"patches": [29, 30, 53, 54, 75, 76, 77, 98, 99, 100, 122, 123, 144, 145, 146, 167, 168, 169, 190, 191, 192], "bbox": [0.269109375, 0.09734104046242774, 0.37901562499999997, 0.7409537572254336], "iscrowd": 0, "area": 8319.162649999998, "rle": {"size": [346, 640], "counts": "a[j11i:001N101O1N1nM6TIKj6>mHCS7d0cH_O]7i0YHYOg7n0oGUOQ8R1eGPOW4_OHX27jMHV28jMoMN^OX2b2kMjM8_Om1g2lMcM`0Cc1j2PN[Md0J\\1j2UNVMa00Z1i2XNSMa0POPOg0W2T3]NoL`0SOnNj0U2S3`NlL?UOnNl0R2S3dNgL`0WOlNo0P2R3gNdLi0:`0Q3iNaLk0= in this image.", "objects": [{"patches": [19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321], "bbox": [0.581734375, 0.0013333333333333335, 1.0, 0.9901234567901235], "iscrowd": 0, "area": 74995.843, "rle": {"size": [405, 640], "counts": "Q^c42Y< in this image.", "objects": [{"patches": [19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321], "bbox": [0.581734375, 0.0013333333333333335, 1.0, 0.9901234567901235], "iscrowd": 0, "area": 74995.843, "rle": {"size": [405, 640], "counts": "Q^c42Y< in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 59, 76, 77, 78, 79, 80, 81, 82, 99, 100, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241], "bbox": [0.33703125, 0.17079012345679012, 0.620015625, 0.7550617283950618], "iscrowd": 0, "area": 33838.93280000001, "rle": {"size": [405, 640], "counts": "^`e27^<:F:F;E:F:F:E;F:F:F:F:F:F:F:F:E;F:F;E:F:F:F1O00O0100000000000O10O1000000000O100000O100000O10000000O1000O100000000000O010000000000O10O1000000000O100000O100000O1000000000O10O100000000000O01000000000000O10O1000000000O1000O10000000O10000000O1000O1000000000O10O1000000000000O0100000000000O1000O1001O4L7H7J7I6J6J7I6J7H7J7I6J7I6J7I6I8I6J6J7I6J7I6I8I6J7I6J7I6J7H7J6J7Ifjo2"}, "label": "the cell phone on top"}]} {"id": 524866, "image": "COCO_train2014_000000524866.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the phone in pink\"."}], "task": "refering", "answer_template": "The \"the phone in pink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 59, 76, 77, 78, 79, 80, 81, 82, 99, 100, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 221, 238, 239, 240, 241], "bbox": [0.33703125, 0.17079012345679012, 0.620015625, 0.7550617283950618], "iscrowd": 0, "area": 33838.93280000001, "rle": {"size": [405, 640], "counts": "^`e27^<:F:F;E:F:F:E;F:F:F:F:F:F:F:F:E;F:F;E:F:F:F1O00O0100000000000O10O1000000000O100000O100000O10000000O1000O100000000000O010000000000O10O1000000000O100000O100000O1000000000O10O100000000000O01000000000000O10O1000000000O1000O10000000O10000000O1000O1000000000O10O1000000000000O0100000000000O1000O1001O4L7H7J7I6J6J7I6J7H7J7I6J7I6J7I6I8I6J6J7I6J7I6I8I6J7I6J7I6J7H7J6J7Ifjo2"}, "label": "the phone in pink"}]} {"id": 524866, "image": "COCO_train2014_000000524866.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red flip phone under the smartphone\"."}], "task": "refering", "answer_template": "The \"the red flip phone under the smartphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 31, 32, 33, 34, 35, 36, 52, 53, 54, 55, 56, 57, 58, 59, 60, 75, 76, 82, 83], "bbox": [0.288671875, 0.0494320987654321, 0.62996875, 0.2943703703703704], "iscrowd": 0, "area": 6632.84, "rle": {"size": [405, 640], "counts": "mWY2`0U<2M3M3M3M3N2M3M3M2O1N2N1O2N2O0ODoDWOP;f0[ESOc:k0e0M3M2N3M3N110O10O10000O01000O100000O010000O1000O010000O10O10O10000O01000O1000O010000O10O10O10000O01000O10000O01000O1000O010000O10O10O10000O01000O1000O010000O10O1000O10000O10000O01000O10000O10000O10O10O10000O10000O10000O01000O100O100O100O010O010O010O0010O010O010O010O2O1O010O1O001O1O010O001O10O0100O00100O010O00100O10O2O1O1O2N1N2O1O1O2M2O1O1O1N3NO1O1O001O1N2O001O1O1O1N101N2N2O1N2O1N2O1N101N2N2O2M;FQem2"}, "label": "the red flip phone under the smartphone"}]} {"id": 524866, "image": "COCO_train2014_000000524866.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an unknown electronic device behind a mobile phone on an unknown surface\"."}], "task": "refering", "answer_template": "The \"an unknown electronic device behind a mobile phone on an unknown surface\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 31, 32, 33, 34, 35, 36, 52, 53, 54, 55, 56, 57, 58, 59, 60, 75, 76, 82, 83], "bbox": [0.288671875, 0.0494320987654321, 0.62996875, 0.2943703703703704], "iscrowd": 0, "area": 6632.84, "rle": {"size": [405, 640], "counts": "mWY2`0U<2M3M3M3M3N2M3M3M2O1N2N1O2N2O0ODoDWOP;f0[ESOc:k0e0M3M2N3M3N110O10O10000O01000O100000O010000O1000O010000O10O10O10000O01000O1000O010000O10O10O10000O01000O10000O01000O1000O010000O10O10O10000O01000O1000O010000O10O1000O10000O10000O01000O10000O10000O10O10O10000O10000O10000O01000O100O100O100O010O010O010O0010O010O010O010O2O1O010O1O001O1O010O001O10O0100O00100O010O00100O10O2O1O1O2N1N2O1O1O2M2O1O1O1N3NO1O1O001O1N2O001O1O1O1N101N2N2O1N2O1N2O1N101N2N2O2M;FQem2"}, "label": "an unknown electronic device behind a mobile phone on an unknown surface"}]} {"id": 287302, "image": "COCO_train2014_000000287302.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra on right side , jumping over log\"."}], "task": "refering", "answer_template": "The \"zebra on right side , jumping over log\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 35, 55, 56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 102, 103, 104, 105, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 294, 311, 312, 313, 314, 315, 316, 317, 318, 336, 337, 338, 339, 340, 341, 360, 361, 362, 363, 364], "bbox": [0.420484375, 0.05393177737881508, 0.858578125, 0.7977558348294435], "iscrowd": 0, "area": 52573.06965, "rle": {"size": [557, 640], "counts": "Q\\b4c0c`08O1O010O1O2O0O1O1O100O2N1O101N1O1O2O0O1O2N1O1O010k@QOj=P1SBTOk=l0SBYOi=i0TB\\Oi=e0SBg0c<\\OYCh0e0eMm:f1ZDi0d0PNe:Z1cDj0d0PNg:P4UEULg:n3UEWLg:l3UEXLi:j3REYLn:T5O001N1O2M2N3N1N3N1nEXI^9j6RFoIc9X7E7I2N2N2N2N2N2O1O1O1O1O1O1O1O1O1O001O1O1O1O1O2N1O1O1O2N1O1O2O_G^HV7a7hHbHX7]7fHeH\\7Y7aHkH_7T7`HmHa7S7\\HoHf7o6XHSIi7l6VHUIl7i6RHYIP8e6oG\\IS8n7j0VO1N3N1O2N1O00XHVHm5j7lI^HS6b7eIgHZ6Y7^IPIa6P7ZIVIe6i6ZIZIe6f6ZI]Id6c6ZI`Ie6`6ZIbIe6^6ZIdIe6\\6[IeIc6\\6]IeIb6[6^IfIa6Y6`IhI_6X6bIhI]6X6cIiI\\6W6dIiI\\6W6dIjI[6U6fIlIY6T6hIlIV6U6jIlIU6T6kIlIU6S6kIoIU6P6jIRJU6n5jITJV6k5iIWJW6h5hIYJX6g5gI[JY6d5eI_JZ6a5eIaJ[6^5dIcJ]6\\5bIfJ]6Z5bIhJ^6W5aIkJ`6S5_IoJa6P5]IRKd6m4[IUKe6j4ZIXKf6k4UIWKl6d7100000001N10001O000O2O00000O1000O10O100000O01000000O1000000O1000000O1000000O100O10O1O001O1O001O1O0RN]FRLd9l3bFPL^9m3iFnKW9Q4lFmKU9Q4nFmKS9Q4oFnKQ9Q4XF]KMa0l9Q4UFhKI5S:R4RFmKJ0U:R4oERLKKW:R4lEVLNFW:T4hEZL0AZ:S4dE_L3\\OZ:T4aEcL4YO\\:`5dE_J]:`5cE`J^:^5dEaJ]:^5dEaJ^:]5bEcJ_:\\5bEcJ_:\\5bEcJ_:[5bEfJ^:X5dEgJ]:X5cEhJ_:U5cEjJ^:T5dElJ\\:S5eElJ\\:R5fEmJ[:Q5hEmJY:R5hEmJZ:P5iEnJX:Q5iEnJX:P5jEoJW:[4oDPLm0BV:\\4`FcKa9\\4j1M2O2O1O1O1O101N1O100O1001O0010O000O100O2O1N101N2O1N101O1N101N2O0O2N2N2N1O5K6J6J8Hm0TO8G2N2N1N3M2N3M3M2N3M3M2N3M2N3N2M2N3MURa1"}, "label": "zebra on right side , jumping over log"}]} {"id": 287302, "image": "COCO_train2014_000000287302.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra on the left\"."}], "task": "refering", "answer_template": "The \"the zebra on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 97, 98, 99, 100, 101, 121, 122, 123, 144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 282, 283, 285, 286, 287, 305, 306, 308, 309, 310, 311, 328, 329, 331, 332, 333, 334], "bbox": [0.22157812500000001, 0.16100538599640934, 0.564171875, 0.7534470377019749], "iscrowd": 0, "area": 32962.509249999996, "rle": {"size": [557, 640], "counts": "R`]24Ua06K6I7I6K6I6K6I7J5J7J5VLkMTH\\2g7iMSH\\2j7hMoG_2l7fMnG`2n7cMlGc2P8bMjGd2R8`MgGf2U8_MdGh2X8\\MbGi2Z8\\M`Gj2\\8ZM_Gk2\\8YM`Gk2]8YM^Gl2_8fKWFX1U1V3e8SMVGR3k8mLPGX3P9hLkF\\3U9fLeF_3X9fLaF_3]9dL]F`3b9X2N2M2O2N2N1N3N1O2N2M2O2N2N1N2=BO2M3M3M2O2M2N1O00100O001O1O00cMUHZKj7g4YIVJf6k5]21O1O1O100O2O0O100O100O2N100O100O2O0O11O2TK^Ei2d:TM_Ek2b:SM`Ek2c:QM`En2a:oLcEn2_:oLdEP3]:mLfEQ3\\:kLgEU3X:jLhEW3X:hLiEY3U:gLlEY3S:fLoE[3o9eLQF\\3n9dLSF]3k9bLWF_3h9_LYFb3f9]L\\Fd3b9[L`Fe3_9ZLbFh3\\9WLfFk3Y9RLiFP4V9mKlFT4T9iKmFZ4S9bKoF_4Q9^KQGd4o8XKRGl4l8QKWGP5h8mJ[GT5e8hJ^GY5[:1N2O1O1O1O2N1N3N1O2NhM]DfNb;Y1`DhN^;V1fDiNX;W1jDjNT;U1nDkNQ;S1REnNl:Q1WEnNh:P1[EQOc:R1[EnNc:W1ZEhNf:Y1[EeNe:\\1[EbNf:_1ZE`Nf:a1ZE^Nf:c1[EZNf:g1ZEXNe:j1[ETNf:m1ZERNf:o1[EnMf:S2ZEkMg:W2XEgMi:Z2WEdMj:]2WE`Mj:a2VE]Mj:e2VEYMk:h2UEVMl:ZOkDd2:oMl:TOSEn21lMm:mN\\EW3GjMl;U2VDiMl;V2VDgMk;Y2VDeMl;Z2UDdMl;\\2UDbMm;]2UD`Ml;`2UD^Mm;a2TD]Mm;b2UD\\Mm;c2UDZMl;f2UDXMm;g2TDVMn;j2TDSMn;l2SDRMn;n2Y101O1N9I6J7I6Jh0XO3M4K4M1O00O1O1O001O1O1O1O1O1O001O100O1O1O1O001O1OO1O1O1O1O1O2N1O1O1O100N2L4K5L4L4K5L4LT\\g4"}, "label": "the zebra on the left"}]} {"id": 541255, "image": "COCO_train2014_000000541255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich with green stuff on it that is clear and not blurry\"."}], "task": "refering", "answer_template": "The \"a sandwich with green stuff on it that is clear and not blurry\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 17, 18, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 66, 67, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 88, 89, 90, 91, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 330, 331, 332, 337, 338, 339], "bbox": [0.176921875, 0.001873536299765808, 1.0, 0.9501873536299766], "iscrowd": 0, "area": 162031.38025000007, "rle": {"size": [427, 640], "counts": "VY_11W=4K5L5K4SF@j6d0SI^Oj6g0RI[Ok6i0RIZOk6i0RIYOm6j0PIWOn6k0PIWOo6k0nHXOP7i0nHYOP7i0nHYOQ7i0lHZOR7g0lH[OR7g0lH[OS7g0jH\\OT7e0iH^OU7e0hH^OV7c0hH_OW7b0gHAV7b0fHAY7`0eHCY7>eHDY7?dHC[7>bHF\\7;bHG\\7 in this image.", "objects": [{"patches": [64, 65, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 180, 203], "bbox": [0.803296875, 0.15041666666666667, 1.0, 0.5002916666666667], "iscrowd": 0, "area": 13186.9767, "rle": {"size": [480, 640], "counts": "URa7b0U>j0VO;E in this image.", "objects": [{"patches": [64, 65, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 180, 203], "bbox": [0.803296875, 0.15041666666666667, 1.0, 0.5002916666666667], "iscrowd": 0, "area": 13186.9767, "rle": {"size": [480, 640], "counts": "URa7b0U>j0VO;E in this image.", "objects": [{"patches": [230, 231, 241, 242, 243, 244, 245, 246, 247, 248, 258, 259, 260, 261, 262, 263, 264, 265, 277, 278, 279], "bbox": [0.19718750000000002, 0.5680468750000001, 0.6376458333333334, 0.718265625], "iscrowd": 0, "area": 8164.709849999999, "rle": {"size": [640, 480], "counts": "ghk16gc04N2N2N1O2N2N2O0100000O01000O0100000O0100000O0100000O0100000O010000O010000O0100000O0100000O0100000O010001Oa0^Oif83RYGj0WO6I8I8H1N2O1O010001O0O10000O10001O1N101O1O0O2O001N2O001O0O10O10O1000O10O10O10O1000O10O1000O01000O10O1000O01000O10O1000O01000O10O1000O0100000O01000O0100000O01000O10O0100O1O00100O1O00100O1_OcNZ^Ob1ea0aNT^Oc1ka0:O1O1O01000000O100000000O1000000O1000000O1000000O101O1O1O0O2O1O1O0O2O1O1O0O2O1O001O1I7_O`0@f\\\\3"}, "label": "the bottom bunk with a blue blanket"}]} {"id": 229990, "image": "COCO_train2014_000000229990.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bottom bed of bunkbed set\"."}], "task": "refering", "answer_template": "The \"bottom bed of bunkbed set\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 241, 242, 243, 244, 245, 246, 247, 248, 258, 259, 260, 261, 262, 263, 264, 265, 277, 278, 279], "bbox": [0.19718750000000002, 0.5680468750000001, 0.6376458333333334, 0.718265625], "iscrowd": 0, "area": 8164.709849999999, "rle": {"size": [640, 480], "counts": "ghk16gc04N2N2N1O2N2N2O0100000O01000O0100000O0100000O0100000O0100000O010000O010000O0100000O0100000O0100000O010001Oa0^Oif83RYGj0WO6I8I8H1N2O1O010001O0O10000O10001O1N101O1O0O2O001N2O001O0O10O10O1000O10O10O10O1000O10O1000O01000O10O1000O01000O10O1000O01000O10O1000O0100000O01000O0100000O01000O10O0100O1O00100O1O00100O1_OcNZ^Ob1ea0aNT^Oc1ka0:O1O1O01000000O100000000O1000000O1000000O1000000O101O1O1O0O2O1O1O0O2O1O1O0O2O1O001O1I7_O`0@f\\\\3"}, "label": "bottom bed of bunkbed set"}]} {"id": 238187, "image": "COCO_train2014_000000238187.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top of the banana on the apple\"."}], "task": "refering", "answer_template": "The \"the top of the banana on the apple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 30, 31, 32, 33, 43, 44, 45, 46, 47, 56, 57, 58, 59, 60, 69, 70, 71, 72, 73, 74, 82, 83, 84, 85, 86, 87, 96, 97, 98, 99, 100, 109, 110, 111, 112, 113, 122, 123, 124, 125, 126, 135, 136, 137, 138, 139, 140, 148, 149, 150, 151, 152, 153], "bbox": [0.34978666666666663, 0.055299999999999995, 0.7862399999999999, 0.6539599999999999], "iscrowd": 0, "area": 34016.586500000005, "rle": {"size": [500, 375], "counts": "RRP24Y??B>A?B>A`0A>E;G9G9G8H7H7J6J7I6J7H7J6J7I6J7H7J7I6J6J7H7J7I6J6J7H7J7I3N1O10O01O100O00010O01O00010O00001O010O000010O0001O010O00001O01O01O001O01O01O00001O010O00001O2N2O1N1O2N2N1O2N2N1O2N2N1O2N2N1O2M3N1O2N2N1O2N2N2N1O2N2N1O2N2N3M3M2N3M3M3L4M2N3M3M3L4M3M2N3M3L4M3M2N3M3L4M3M3M2N3L4M3M3M2N3L4M3M3M2N5J8I7I7I7I7H8I7I6J7I7H8I7I7I7H8H7IZhV1"}, "label": "the top of the banana on the apple"}]} {"id": 238187, "image": "COCO_train2014_000000238187.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half cut banana on top of the apple\"."}], "task": "refering", "answer_template": "The \"half cut banana on top of the apple\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 30, 31, 32, 33, 43, 44, 45, 46, 47, 56, 57, 58, 59, 60, 69, 70, 71, 72, 73, 74, 82, 83, 84, 85, 86, 87, 96, 97, 98, 99, 100, 109, 110, 111, 112, 113, 122, 123, 124, 125, 126, 135, 136, 137, 138, 139, 140, 148, 149, 150, 151, 152, 153], "bbox": [0.34978666666666663, 0.055299999999999995, 0.7862399999999999, 0.6539599999999999], "iscrowd": 0, "area": 34016.586500000005, "rle": {"size": [500, 375], "counts": "RRP24Y??B>A?B>A`0A>E;G9G9G8H7H7J6J7I6J7H7J6J7I6J7H7J7I6J6J7H7J7I6J6J7H7J7I3N1O10O01O100O00010O01O00010O00001O010O000010O0001O010O00001O01O01O001O01O01O00001O010O00001O2N2O1N1O2N2N1O2N2N1O2N2N1O2N2N1O2M3N1O2N2N1O2N2N2N1O2N2N1O2N2N3M3M2N3M3M3L4M2N3M3M3L4M3M2N3M3L4M3M2N3M3L4M3M3M2N3L4M3M3M2N3L4M3M3M2N5J8I7I7I7I7H8I7I6J7I7H8I7I7I7H8H7IZhV1"}, "label": "half cut banana on top of the apple"}]} {"id": 238187, "image": "COCO_train2014_000000238187.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a banana slice to the back right of an apple slice\"."}], "task": "refering", "answer_template": "The \"a banana slice to the back right of an apple slice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 115, 126, 127, 128, 129, 140, 141, 142, 154, 155, 168, 180, 181, 193, 194], "bbox": [0.75864, 0.46908, 0.9773066666666667, 0.7954], "iscrowd": 0, "area": 5379.405499999999, "rle": {"size": [500, 375], "counts": "klZ41c?=C in this image.", "objects": [{"patches": [106, 107, 117, 118, 119, 120, 121, 130, 131, 132, 133, 134, 143, 144, 145, 146, 157, 158, 170, 171, 183, 184, 185, 196, 197, 198], "bbox": [0.014186666666666667, 0.47814, 0.34042666666666666, 0.8558], "iscrowd": 0, "area": 11653.24085, "rle": {"size": [500, 375], "counts": "mf21Z?;L4L4M4L3M3L5L3M3M3M4M2M4M3M3M3L4M2N3M3L4M3M3M3L4M3M3M3L3N5K6I6K5K6J5J6K6J5K5J7J5K2N1N2O2N000O1000000OQMRF:n9BYF;g9E[F9e9F_F6a9JcF3]9LfF2Z9NiFOW90mFMR93QGKo85SGIm86VGHj87XGHg89[GEe8:^GDb8;aGC_8=bGB]8>fG@Z8?iG_OW8`0kG_OU8a0mG]OS8b0PH\\OP8c0SH[Ol7f0UHYOk7f0XHWOi7h0ZHVOf7j0[HUOe7j0^HTOb7l0_HSO`7m0cHQO]7o0eHoN[7Q1fHnNZ7Q1iHmNW7S1jHlNV7T1lHjNS7V1oHiNQ7W1QIgNo6Y1SIeNm6Z1UIeNk6[1WIcNi6]1XIbNg6^1\\I`Nd6`1^I^Nb6b1_I\\Nb6c1aI[N_6e1bIZN^6f1dIXN[6h1hIVNX6j1iIUNW6k1i3O2O2N1O2M3N3M3L4M3M3L3N3M3L5L3M3L4M3M3L5Lkih3"}, "label": "a cut piece of bananaplaced on the left corner of the apple"}]} {"id": 238187, "image": "COCO_train2014_000000238187.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small banana piece on table towards left side of the image\"."}], "task": "refering", "answer_template": "The \"a small banana piece on table towards left side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 117, 118, 119, 120, 121, 130, 131, 132, 133, 134, 143, 144, 145, 146, 157, 158, 170, 171, 183, 184, 185, 196, 197, 198], "bbox": [0.014186666666666667, 0.47814, 0.34042666666666666, 0.8558], "iscrowd": 0, "area": 11653.24085, "rle": {"size": [500, 375], "counts": "mf21Z?;L4L4M4L3M3L5L3M3M3M4M2M4M3M3M3L4M2N3M3L4M3M3M3L4M3M3M3L3N5K6I6K5K6J5J6K6J5K5J7J5K2N1N2O2N000O1000000OQMRF:n9BYF;g9E[F9e9F_F6a9JcF3]9LfF2Z9NiFOW90mFMR93QGKo85SGIm86VGHj87XGHg89[GEe8:^GDb8;aGC_8=bGB]8>fG@Z8?iG_OW8`0kG_OU8a0mG]OS8b0PH\\OP8c0SH[Ol7f0UHYOk7f0XHWOi7h0ZHVOf7j0[HUOe7j0^HTOb7l0_HSO`7m0cHQO]7o0eHoN[7Q1fHnNZ7Q1iHmNW7S1jHlNV7T1lHjNS7V1oHiNQ7W1QIgNo6Y1SIeNm6Z1UIeNk6[1WIcNi6]1XIbNg6^1\\I`Nd6`1^I^Nb6b1_I\\Nb6c1aI[N_6e1bIZN^6f1dIXN[6h1hIVNX6j1iIUNW6k1i3O2O2N1O2M3N3M3L4M3M3L3N3M3L5L3M3L4M3M3L5Lkih3"}, "label": "a small banana piece on table towards left side of the image"}]} {"id": 74356, "image": "COCO_train2014_000000074356.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady in a pink shirt preparing food\"."}], "task": "refering", "answer_template": "The \"a lady in a pink shirt preparing food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 121, 122, 123, 135, 136, 137, 138, 150, 151, 152, 166, 167, 181, 182, 183, 184, 198, 199], "bbox": [0.053887587822014055, 0.32043750000000004, 0.3160421545667447, 0.590328125], "iscrowd": 0, "area": 7996.0400500000005, "rle": {"size": [640, 427], "counts": "mc>7hc0>C=B>C9F3N2M3N2M3N2M2O2Dg0YAUOg>j0^AROc>m0_AQOa>n0aAQO_>n0dAPO\\>P1eAoN\\>o0fAoN[>Q1fAnNZ>Q1iAmNW>R1kAmNV>R1kAmNU>R1nAlNR>S1PBlNQ>R1QBmNo=Q1TBnNm=P1VBnNk=P1WBoNj=o0XBPOh=o0[BoNf=P1[BoNf=o0\\BPOe=n0\\BROe=l0\\BTOe=j0\\BVOe=h0\\BYOe=d0\\B^Oc=a0]B@d==]BDd=8^BJb=1aB0`=KcB6]=EhB in this image.", "objects": [{"patches": [106, 107, 108, 121, 122, 123, 135, 136, 137, 138, 150, 151, 152, 166, 167, 181, 182, 183, 184, 198, 199], "bbox": [0.053887587822014055, 0.32043750000000004, 0.3160421545667447, 0.590328125], "iscrowd": 0, "area": 7996.0400500000005, "rle": {"size": [640, 427], "counts": "mc>7hc0>C=B>C9F3N2M3N2M3N2M2O2Dg0YAUOg>j0^AROc>m0_AQOa>n0aAQO_>n0dAPO\\>P1eAoN\\>o0fAoN[>Q1fAnNZ>Q1iAmNW>R1kAmNV>R1kAmNU>R1nAlNR>S1PBlNQ>R1QBmNo=Q1TBnNm=P1VBnNk=P1WBoNj=o0XBPOh=o0[BoNf=P1[BoNf=o0\\BPOe=n0\\BROe=l0\\BTOe=j0\\BVOe=h0\\BYOe=d0\\B^Oc=a0]B@d==]BDd=8^BJb=1aB0`=KcB6]=EhB in this image.", "objects": [{"patches": [180, 181, 202, 203, 204, 217, 218, 219, 220, 221, 222, 223, 225, 226, 227, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 285, 290, 291, 292], "bbox": [0.22284375, 0.5236470588235295, 0.9536562500000001, 0.8413176470588235], "iscrowd": 0, "area": 21406.527600000005, "rle": {"size": [425, 640], "counts": "^_l11W=1000001O0O10001O000O2O0000001N10001O000O100000001N100000000O1000001O0O100000000O101O000jC_Od;a0ZDBd;?\\DBc;>TD_O05k;dEC\\:=bEE]:g0WEZOi:c10000000XO[EhNh:X1e0O1O001O1O0O2O001O1O001O10O10000O100O10000O10000O10000O100O1000000O010000000O10000000000O100000000001O0000dDcNl:]1nDjNQ;h1O1O0000001O000000M3N2O2O000000000000000006J1O00000001O0O100000000HoDXNQ;c1UE[Nl:_1ZEaNf:^1\\EaNe:]1\\EcNd:]1d00001O00001O00000O100000001O00000000000O1000001O00000001O1O2N1O1O1O100O1cDfNi:\\1QEmNj:k1N1O1O1O0000O101O00000000000O1M3TORE[OP;e0TEQOR;o0REeNW;[1;00000000000000000001O0000000000000000REgNl9Y1QFjNP:U1nElNS:T1kEnNU:R1hEQOX:o0fESOZ:m0dEUO\\:l0`EWO`:l0[EVOe:m0UEVOk:d100001O00000000000000000_OYE`Ng:[1aEbN`:W1iEfNW:U1Q1000000000000000001N1000000000000000000O0100O100O1O100O100O0O2O1N2O1O1N2O1N2O1O1N101N2O1O1N2O1N2O1O1N101N2O1O1N2O1N2O1O1N101N2O1O1N2O1N2O10000000O10000000000000000000000000000000003M4L5K5K4L5J5L5K5K4L5K5K4L5K3M001O1O1O1O001O1O1O1O0000000O101O0000000000001O0000000000000NUU<"}, "label": "the plane that is in plain view"}]} {"id": 549499, "image": "COCO_train2014_000000549499.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey jet in front of another jet\"."}], "task": "refering", "answer_template": "The \"a grey jet in front of another jet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 202, 203, 204, 217, 218, 219, 220, 221, 222, 223, 225, 226, 227, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 285, 290, 291, 292], "bbox": [0.22284375, 0.5236470588235295, 0.9536562500000001, 0.8413176470588235], "iscrowd": 0, "area": 21406.527600000005, "rle": {"size": [425, 640], "counts": "^_l11W=1000001O0O10001O000O2O0000001N10001O000O100000001N100000000O1000001O0O100000000O101O000jC_Od;a0ZDBd;?\\DBc;>TD_O05k;dEC\\:=bEE]:g0WEZOi:c10000000XO[EhNh:X1e0O1O001O1O0O2O001O1O001O10O10000O100O10000O10000O10000O100O1000000O010000000O10000000000O100000000001O0000dDcNl:]1nDjNQ;h1O1O0000001O000000M3N2O2O000000000000000006J1O00000001O0O100000000HoDXNQ;c1UE[Nl:_1ZEaNf:^1\\EaNe:]1\\EcNd:]1d00001O00001O00000O100000001O00000000000O1000001O00000001O1O2N1O1O1O100O1cDfNi:\\1QEmNj:k1N1O1O1O0000O101O00000000000O1M3TORE[OP;e0TEQOR;o0REeNW;[1;00000000000000000001O0000000000000000REgNl9Y1QFjNP:U1nElNS:T1kEnNU:R1hEQOX:o0fESOZ:m0dEUO\\:l0`EWO`:l0[EVOe:m0UEVOk:d100001O00000000000000000_OYE`Ng:[1aEbN`:W1iEfNW:U1Q1000000000000000001N1000000000000000000O0100O100O1O100O100O0O2O1N2O1O1N2O1N2O1O1N101N2O1O1N2O1N2O1O1N101N2O1O1N2O1N2O1O1N101N2O1O1N2O1N2O10000000O10000000000000000000000000000000003M4L5K5K4L5J5L5K5K4L5K5K4L5K3M001O1O1O1O001O1O1O1O0000000O101O0000000000001O0000000000000NUU<"}, "label": "a grey jet in front of another jet"}]} {"id": 524925, "image": "COCO_train2014_000000524925.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small silver plane with n6594q printed on the side in black\"."}], "task": "refering", "answer_template": "The \"a small silver plane with n6594q printed on the side in black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 161, 163, 164, 174, 175, 176, 177, 178, 179, 184, 185, 186, 187, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.001546875, 0.4202267573696145, 0.9972031250000001, 0.8898866213151927], "iscrowd": 0, "area": 75004.17700000001, "rle": {"size": [441, 640], "counts": "Qf0i1W:j11N2N2O0O2N2O1M3M3M4K4M4L3L5L5K6I7H7H9G5K4O11N1000000O101O0O1BjNYDV1g;kNXDU1h;kNWDV1h;kNXDU1h;lNWDU1g;lNXDU1h;lNWDT1h;mNXDS1h;nNWDR1h;oNWDR1i;oNVDR1h;oNXDQ1h;oNWDR1i;oNVDQ1i;POWDP1i;QOUDP1j;QOVDo0j;a0O0001O0000O2A>D in this image.", "objects": [{"patches": [140, 141, 161, 163, 164, 174, 175, 176, 177, 178, 179, 184, 185, 186, 187, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.001546875, 0.4202267573696145, 0.9972031250000001, 0.8898866213151927], "iscrowd": 0, "area": 75004.17700000001, "rle": {"size": [441, 640], "counts": "Qf0i1W:j11N2N2O0O2N2O1M3M3M4K4M4L3L5L5K6I7H7H9G5K4O11N1000000O101O0O1BjNYDV1g;kNXDU1h;kNWDV1h;kNXDU1h;lNWDU1g;lNXDU1h;lNWDT1h;mNXDS1h;nNWDR1h;oNWDR1i;oNVDR1h;oNXDQ1h;oNWDR1i;oNVDQ1i;POWDP1i;QOUDP1j;QOVDo0j;a0O0001O0000O2A>D in this image.", "objects": [{"patches": [98, 99, 121, 122, 123, 127, 128, 129, 130, 132, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178], "bbox": [0.27864062500000003, 0.2936281179138322, 0.786234375, 0.4690249433106576], "iscrowd": 0, "area": 12067.86595, "rle": {"size": [441, 640], "counts": "nV]22g=3M2N2N3M2N2N3M2M4M2N2N3[DQOZ:Q1cEoN^:S1`EmN`:V1\\EkNd:W1YEjNg:X1WEhNi:[1SEfNm:\\1PEeNP;]1nDcNR;`1jDaNV;a1gD`NY;b1eD^N[;e1aD\\N_;o1000000O100001N2O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1N2O13MN2N2O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O01O000000000O10000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000O10000O10001N10000O10000O10000O100N2N2N2N2N2O10000000000000000O2O0000000000000O10000000000000000O10000000000000000O2O0000000000000O10O100000000O1000O10O1000000O1000O1000O2O1O1N101N2O1O000000000O2O0000000000000mNbCk0^ in this image.", "objects": [{"patches": [98, 99, 121, 122, 123, 127, 128, 129, 130, 132, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178], "bbox": [0.27864062500000003, 0.2936281179138322, 0.786234375, 0.4690249433106576], "iscrowd": 0, "area": 12067.86595, "rle": {"size": [441, 640], "counts": "nV]22g=3M2N2N3M2N2N3M2M4M2N2N3[DQOZ:Q1cEoN^:S1`EmN`:V1\\EkNd:W1YEjNg:X1WEhNi:[1SEfNm:\\1PEeNP;]1nDcNR;`1jDaNV;a1gD`NY;b1eD^N[;e1aD\\N_;o1000000O100001N2O1O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1N2O13MN2N2O2N1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O01O000000000O10000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000O10000O10001N10000O10000O10000O100N2N2N2N2N2O10000000000000000O2O0000000000000O10000000000000000O10000000000000000O2O0000000000000O10O100000000O1000O10O1000000O1000O1000O2O1O1N101N2O1O000000000O2O0000000000000mNbCk0^ in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 299], "bbox": [0.0, 0.0, 0.6711406249999999, 0.7809791666666667], "iscrowd": 0, "area": 117147.90389999996, "rle": {"size": [480, 640], "counts": "0S1m=U2kMU2kMV2jMU2kMo1QNO1000000O1000000O1000000O1000000O1000000O1000000O10000O10000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O10000O10000O10000O1000001N10000O100O10000O100O100O100O100O10000O100O100O100O100O10000O100O100O100O10000O100O100O100O100O10000O100O100O100O100O10000O100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O1O101N2N101N2O1N2N2O1N2N2O1N2N2O1N2N101N101N1O2O0O2N101N1O2O1N1O2O0O2N101N101N1O2O0O2N101N1O2O0O2N101N101N1O2O0O2N101N1O2O0O2N101N1O2O0O2O0O2N2O1N2N2O1M3L5L3L4M3L4L4M3L4M3L5K4M3L4^OPDoMT in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 299], "bbox": [0.0, 0.0, 0.6711406249999999, 0.7809791666666667], "iscrowd": 0, "area": 117147.90389999996, "rle": {"size": [480, 640], "counts": "0S1m=U2kMU2kMV2jMU2kMo1QNO1000000O1000000O1000000O1000000O1000000O1000000O10000O10000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O10000O10000O10000O1000001N10000O100O10000O100O100O100O100O10000O100O100O100O100O10000O100O100O100O10000O100O100O100O100O10000O100O100O100O100O10000O100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O1O101N2N101N2O1N2N2O1N2N2O1N2N2O1N2N101N101N1O2O0O2N101N1O2O1N1O2O0O2N101N101N1O2O0O2N101N1O2O0O2N101N101N1O2O0O2N101N1O2O0O2N101N1O2O0O2O0O2N2O1N2N2O1M3L5L3L4M3L4L4M3L4M3L5K4M3L4^OPDoMT in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 60, 61, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 250, 251, 252, 274, 275], "bbox": [0.62190625, 0.0022500000000000003, 1.0, 0.6674166666666665], "iscrowd": 0, "area": 55855.31145000002, "rle": {"size": [480, 640], "counts": "^Rk52m>2N2M4M2N3M2N3N1N3N1bDXOb8j0[GYOb8i0]GXOb8j0[GYOc8h0YG]Od8f0WG^Oh8c0TGBj8`0QGEm8`9C]F`0b9AZFd0c9_OXFe0g9\\OUFi0i9XOTFl0i9VOSFn0l9U2N2N1N2O1O0O2O1O1N2O1O1O1N2O1O0O2O1O1O1N2O1O1N2O1O0O2O1O1O1N2O1O1N2O1O0O2O1O1O1N2O1O1N2O001N2O1O0O2O1O1N101O1O1N101O1N101O1N2O001N2O1O001N2O1O0O2O1O1N101O1N2O001N2O001O1N2O001N2O1O001O1O1O001O001N2O001O001O001O1O001O001O0O2O1O001O001O1O001O001O001O1O001O001O1O010O001O001O1O001O001O1O001O001O001O1O001O001O1O001O010O001O1O001O001O1O001O001O001O1O001O001O1O00000010O000000000000000001O00000000000000001O0000000 in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 60, 61, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 250, 251, 252, 274, 275], "bbox": [0.62190625, 0.0022500000000000003, 1.0, 0.6674166666666665], "iscrowd": 0, "area": 55855.31145000002, "rle": {"size": [480, 640], "counts": "^Rk52m>2N2M4M2N3M2N3N1N3N1bDXOb8j0[GYOb8i0]GXOb8j0[GYOc8h0YG]Od8f0WG^Oh8c0TGBj8`0QGEm8`9C]F`0b9AZFd0c9_OXFe0g9\\OUFi0i9XOTFl0i9VOSFn0l9U2N2N1N2O1O0O2O1O1N2O1O1O1N2O1O0O2O1O1O1N2O1O1N2O1O0O2O1O1O1N2O1O1N2O1O0O2O1O1O1N2O1O1N2O001N2O1O0O2O1O1N101O1O1N101O1N101O1N2O001N2O1O001N2O1O0O2O1O1N101O1N2O001N2O001O1N2O001N2O1O001O1O1O001O001N2O001O001O001O1O001O001O0O2O1O001O001O1O001O001O001O1O001O001O1O010O001O001O1O001O001O1O001O001O001O1O001O001O1O001O010O001O1O001O001O1O001O001O001O1O001O001O1O00000010O000000000000000001O00000000000000001O0000000 in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 33, 34, 35, 36, 37, 38, 39, 56, 57, 59, 60, 61, 62, 63, 79, 80, 82, 83, 84, 85, 86, 102, 103, 104, 105, 106, 107, 108, 125, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 294, 310, 311, 312, 313, 314, 315, 316, 317, 333, 334, 335, 336, 337, 338, 339, 357, 358, 359, 360, 361, 380, 381, 382, 383], "bbox": [0.443265625, 0.0005416666666666666, 0.805625, 0.9865208333333333], "iscrowd": 0, "area": 60505.2079, "rle": {"size": [480, 640], "counts": "nWU4W1f=4L3M3M4L3L5L3M3M4L3M3M4K4M4L3M3XOTMXEQ3e:UMQEQ3m:b0EUL_En3^:SLaEo3\\:SLgEl3S:XLoEg3l6RLjK8\\Me3a6\\LSLN^Mf3U6fL\\LD`Me3P6kL`LA`Mc3n5nL]LEeM\\3l5RM^LBfM]3k5QM^LCfM]3k5RM]LBgM]3k5RM^L@gM_3j5SM]L]OjMa3i5RM]L]OhMd3i5QM^L[OhMe3i5QMZLTOeML8P4i5QMXLTOXNQ2[O_OT6]OYLSOWNo1FZOi5EYLROWNm12VO\\5L[LQOVNj1=SOR51\\LQOVNh1`0UOm43]LnNWNg1c0XOh44^LjNZNg1c0[Oe44_LhNYNf1g0_O_44ZM8]NDX44\\MOeNMn35\\M\\OYO?[35YLeNmNd0b1b0W36ZLlNdN;m1d0S35]LkN\\Na0W2?P35]LkNZNa0[2?m26^LiNZN`0^2a0i27`LfNYNa0`2b0f28aLeNXN?d2d0b28VMSO:e0^22aJjNf2=>g0Y23gJfNa2?a0h0U24cMRO;k0o13gMQOh0hN[Og0l23oLa0i0POROi1CoNP1n0WOhNHZOi0m1YOQOQ1l0[OiNn1V1fMSOT1HD^1KkM0R1`0FE^1GlM3R1`0EFn3G`Lc0BFm3EdLd0@Gc78^HHb77_HIa77_HIa76`HJ`75aHK_75cHI]77gHEY7;kHAU7?f300000000000000000001O0O100000O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O1O2N2N2O1N3M2N3ZFiM_6X2`IiM^6Z2_IhM_6Z2_IhM_6[2^IgMa6[2]IfMa6^2[IdMc6_2ZIbMe6b2WI`Mh6c2TI_Mj6d2TI]Mj6g2RI[Ml6i2PIYMo6i2nHZMo6j2nHXMo6j2oHYMn6h2QIZMn6g2nH^Mo6c2nHbMo6`2mHeMP7\\2mHiMQ7X2kHmMR7T2kHPNS7R2hHTNU7P2dHVNZ7m1_HYN^7l1ZH_N^7h4I8I6I7I8H7J6I8H7I8I6I6J6H9G8I6I7J6]MiDR1^;hNeDT1b;gNaDU1e;fN]DV1i;fNZDV1l;dNWDX1P in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 33, 34, 35, 36, 37, 38, 39, 56, 57, 59, 60, 61, 62, 63, 79, 80, 82, 83, 84, 85, 86, 102, 103, 104, 105, 106, 107, 108, 125, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 294, 310, 311, 312, 313, 314, 315, 316, 317, 333, 334, 335, 336, 337, 338, 339, 357, 358, 359, 360, 361, 380, 381, 382, 383], "bbox": [0.443265625, 0.0005416666666666666, 0.805625, 0.9865208333333333], "iscrowd": 0, "area": 60505.2079, "rle": {"size": [480, 640], "counts": "nWU4W1f=4L3M3M4L3L5L3M3M4L3M3M4K4M4L3M3XOTMXEQ3e:UMQEQ3m:b0EUL_En3^:SLaEo3\\:SLgEl3S:XLoEg3l6RLjK8\\Me3a6\\LSLN^Mf3U6fL\\LD`Me3P6kL`LA`Mc3n5nL]LEeM\\3l5RM^LBfM]3k5QM^LCfM]3k5RM]LBgM]3k5RM^L@gM_3j5SM]L]OjMa3i5RM]L]OhMd3i5QM^L[OhMe3i5QMZLTOeML8P4i5QMXLTOXNQ2[O_OT6]OYLSOWNo1FZOi5EYLROWNm12VO\\5L[LQOVNj1=SOR51\\LQOVNh1`0UOm43]LnNWNg1c0XOh44^LjNZNg1c0[Oe44_LhNYNf1g0_O_44ZM8]NDX44\\MOeNMn35\\M\\OYO?[35YLeNmNd0b1b0W36ZLlNdN;m1d0S35]LkN\\Na0W2?P35]LkNZNa0[2?m26^LiNZN`0^2a0i27`LfNYNa0`2b0f28aLeNXN?d2d0b28VMSO:e0^22aJjNf2=>g0Y23gJfNa2?a0h0U24cMRO;k0o13gMQOh0hN[Og0l23oLa0i0POROi1CoNP1n0WOhNHZOi0m1YOQOQ1l0[OiNn1V1fMSOT1HD^1KkM0R1`0FE^1GlM3R1`0EFn3G`Lc0BFm3EdLd0@Gc78^HHb77_HIa77_HIa76`HJ`75aHK_75cHI]77gHEY7;kHAU7?f300000000000000000001O0O100000O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O1O2N2N2O1N3M2N3ZFiM_6X2`IiM^6Z2_IhM_6Z2_IhM_6[2^IgMa6[2]IfMa6^2[IdMc6_2ZIbMe6b2WI`Mh6c2TI_Mj6d2TI]Mj6g2RI[Ml6i2PIYMo6i2nHZMo6j2nHXMo6j2oHYMn6h2QIZMn6g2nH^Mo6c2nHbMo6`2mHeMP7\\2mHiMQ7X2kHmMR7T2kHPNS7R2hHTNU7P2dHVNZ7m1_HYN^7l1ZH_N^7h4I8I6I7I8H7J6I8H7I8I6I6J6H9G8I6I7J6]MiDR1^;hNeDT1b;gNaDU1e;fN]DV1i;fNZDV1l;dNWDX1P in this image.", "objects": [{"patches": [86, 108, 109, 110, 132, 133, 134, 155, 156, 157, 178, 179, 220], "bbox": [0.57059375, 0.15347916666666667, 0.8565625000000001, 0.5951875], "iscrowd": 0, "area": 4452.1779000000015, "rle": {"size": [480, 640], "counts": "P[[5k0T>01O1O0O2O001N101O1O0O4O20001N101O0N3K4KURR10lmmN2O1N2O2M2N2O1N2O1N3M2O2M2O2M2O2L3L5J5L5J5L5J5L5J5L5J43L5K^DjM_:Q2bEUN]:f1dE_N[:\\1eEgN\\:U1dEkN_:R1aEoN`:n0aEROa:m0^ESOd:l0[EUOf:j0YEVOi:i0WEWOk:g0TEYOn:f0QE[OP;d0oD]OR;c0lD]OV;b0iD_OX;`0hD_O[;?dDB];=bDC`;<_DDc;<[DDg; in this image.", "objects": [{"patches": [184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 302, 303], "bbox": [0.0, 0.570677570093458, 0.29921875, 0.8771728971962617], "iscrowd": 0, "area": 18475.2486, "rle": {"size": [428, 640], "counts": "Y85U=8H9H8G5K4L4L4L4N2M3N3M2M3N2M3N2N2M3N2N2N1O2N1O2N2N1O2N1O1N2O1O1O1O1O1O1O1O1O001O001O1O0O2O001O001O1O001O1O001O1O010O1O1O001O1O001O1O001O1O001O001O0000001O000000001O000001O01O000000001O000000001O00000000000000000001O0O10000000001O000O1000000O2O000O1000000O2VM]Fi1c9UN_Fk1a9SNaFl1`9SNaFm1_9QNcFo1]9PNdFP2\\9PNdFo1]9QNcFo1^9oMcFQ2]9oMcFP2^9PNcFo1]9QNbFP2^9PNbFo1_9QNaFn1`9QNaFn1a9QN_Fn1b9RN^Fm1c9SN]Fl1d9TN[Fl1f9TNZFk1h9SNYFl1h9TNXFk1j9TNVFk1k9UNUFk1l9TNTFk1m9UNRFk1P:TNPFk1Q:TNPFk1R:TNnEl1R:TNnEk1T:TNlEj1V:VNiEj1Y:UNgEj1Z:UNgEi1\\:VNdEi1]:WNcEh1_:WNaEg1a:YN_Ef1c:YN\\Eg1e:XN\\Eg1f:XNZEg1g:YNYEf1i:YNWEe1k:[NUEd1m:[NREe1o:ZNREe1[;O1O1O1O1N101N2M3L4N1O00105J6J6K5JRoj5"}, "label": "the white bowl on the left"}]} {"id": 541338, "image": "COCO_train2014_000000541338.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange soup next to a glass of beer\"."}], "task": "refering", "answer_template": "The \"an orange soup next to a glass of beer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 302, 303], "bbox": [0.0, 0.570677570093458, 0.29921875, 0.8771728971962617], "iscrowd": 0, "area": 18475.2486, "rle": {"size": [428, 640], "counts": "Y85U=8H9H8G5K4L4L4L4N2M3N3M2M3N2M3N2N2M3N2N2N1O2N1O2N2N1O2N1O1N2O1O1O1O1O1O1O1O1O001O001O1O0O2O001O001O1O001O1O001O1O010O1O1O001O1O001O1O001O1O001O001O0000001O000000001O000001O01O000000001O000000001O00000000000000000001O0O10000000001O000O1000000O2O000O1000000O2VM]Fi1c9UN_Fk1a9SNaFl1`9SNaFm1_9QNcFo1]9PNdFP2\\9PNdFo1]9QNcFo1^9oMcFQ2]9oMcFP2^9PNcFo1]9QNbFP2^9PNbFo1_9QNaFn1`9QNaFn1a9QN_Fn1b9RN^Fm1c9SN]Fl1d9TN[Fl1f9TNZFk1h9SNYFl1h9TNXFk1j9TNVFk1k9UNUFk1l9TNTFk1m9UNRFk1P:TNPFk1Q:TNPFk1R:TNnEl1R:TNnEk1T:TNlEj1V:VNiEj1Y:UNgEj1Z:UNgEi1\\:VNdEi1]:WNcEh1_:WNaEg1a:YN_Ef1c:YN\\Eg1e:XN\\Eg1f:XNZEg1g:YNYEf1i:YNWEe1k:[NUEd1m:[NREe1o:ZNREe1[;O1O1O1O1N101N2M3L4N1O00105J6J6K5JRoj5"}, "label": "an orange soup next to a glass of beer"}]} {"id": 541338, "image": "COCO_train2014_000000541338.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a bowl of vegetables and seasoning with two utensils placed in the food\"."}], "task": "refering", "answer_template": "The \"this is a bowl of vegetables and seasoning with two utensils placed in the food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 224, 240, 241, 242, 243, 244, 245, 246, 247, 248, 263, 264, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315], "bbox": [0.45228124999999997, 0.5645560747663552, 0.8203125, 0.9402803738317757], "iscrowd": 0, "area": 26196.587750000006, "rle": {"size": [428, 640], "counts": "e_i37S=6I6K4L3L4M4M2N2N001O001ZDmNR;T1mDnNQ;R1PEoNn:R1REnNm:R1SEPOk:Q1UEPOi:P1XEQOf:P1YEROe:n0\\EROd:n0\\ESOb:n0^ESOa:l0_EVO`:j0_EXO_:h0aEYO_:g0aEZO]:f0cE]O[:c0dE_OZ:b0fE@X:?hEDV:`9B_F`0`9A^Fa0a9@^Fb0`9^O_Fd0_9^O`Fc0_9^O_Fe0_9\\O`Fe0_9[O`Fh0^9d1O1O2N1N2O2N1O1O2N001O0O2O001O001O001N101O001O001O0O2O001O001O001N1000000000001O0000000000001O000000000000001O000000000000001O000000000000001O000000000000001O0000000001O1O0O2O0O010O101O0O2O0O10O0100O2O1N2O1N2O1N2O0O2O0O100O101N100O1O2N1O101N1O2N1O1O2N1O1O2N1O2O0O2N1O1O2N1]MPFi1R:UNPFi1Q:VNQFh1Q:VNPFi1Q:VNQFh1Q:VNQFi1o9WNRFg1P:XNPFg1Q:YNoEf1R:ZNnEe1T:ZNkEf1V:ZNjEe1X:ZNgEf1Z:YNfEg1\\:XNdEg1]:YNcEf1_:YNaEg1_:YNaEf1a:YN_Ef1b:ZN^Ee1c:[N]Ed1e:ZN\\Ee1e:[N[Ed1g:[NYEd1i:[NWEd1l:ZNTEe1o:YNQEg1Z;O1O1O1O1O1O1O100O2N1O1O1O1O101N3M4L3M4G9G8Hhh_1"}, "label": "this is a bowl of vegetables and seasoning with two utensils placed in the food"}]} {"id": 541338, "image": "COCO_train2014_000000541338.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of food with two utensils in the bowl\"."}], "task": "refering", "answer_template": "The \"a bowl of food with two utensils in the bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 224, 240, 241, 242, 243, 244, 245, 246, 247, 248, 263, 264, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315], "bbox": [0.45228124999999997, 0.5645560747663552, 0.8203125, 0.9402803738317757], "iscrowd": 0, "area": 26196.587750000006, "rle": {"size": [428, 640], "counts": "e_i37S=6I6K4L3L4M4M2N2N001O001ZDmNR;T1mDnNQ;R1PEoNn:R1REnNm:R1SEPOk:Q1UEPOi:P1XEQOf:P1YEROe:n0\\EROd:n0\\ESOb:n0^ESOa:l0_EVO`:j0_EXO_:h0aEYO_:g0aEZO]:f0cE]O[:c0dE_OZ:b0fE@X:?hEDV:`9B_F`0`9A^Fa0a9@^Fb0`9^O_Fd0_9^O`Fc0_9^O_Fe0_9\\O`Fe0_9[O`Fh0^9d1O1O2N1N2O2N1O1O2N001O0O2O001O001O001N101O001O001O0O2O001O001O001N1000000000001O0000000000001O000000000000001O000000000000001O000000000000001O000000000000001O0000000001O1O0O2O0O010O101O0O2O0O10O0100O2O1N2O1N2O1N2O0O2O0O100O101N100O1O2N1O101N1O2N1O1O2N1O1O2N1O2O0O2N1O1O2N1]MPFi1R:UNPFi1Q:VNQFh1Q:VNPFi1Q:VNQFh1Q:VNQFi1o9WNRFg1P:XNPFg1Q:YNoEf1R:ZNnEe1T:ZNkEf1V:ZNjEe1X:ZNgEf1Z:YNfEg1\\:XNdEg1]:YNcEf1_:YNaEg1_:YNaEf1a:YN_Ef1b:ZN^Ee1c:[N]Ed1e:ZN\\Ee1e:[N[Ed1g:[NYEd1i:[NWEd1l:ZNTEe1o:YNQEg1Z;O1O1O1O1O1O1O100O2N1O1O1O1O101N3M4L3M4G9G8Hhh_1"}, "label": "a bowl of food with two utensils in the bowl"}]} {"id": 541338, "image": "COCO_train2014_000000541338.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small bowl of soup\"."}], "task": "refering", "answer_template": "The \"small bowl of soup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 224, 225, 226, 227, 228, 248, 249, 250], "bbox": [0.727359375, 0.46516355140186916, 0.9437656249999999, 0.7168457943925234], "iscrowd": 0, "area": 8781.07385, "rle": {"size": [428, 640], "counts": "n\\S62Z=4KT1mN0O2O0O2O0O2O001N101N101N101N101N101N101N101O001N101O0O2O001N101O001N10000O101O0O10000O1000000O2O0[EbM\\:^2dEcM[:]2dEfMZ:Z2fEhMX:X2hEjMV:U2kEnMR:R2nEPNP:P2PFRNn9n1RFTNl9k1UFXNh9h1XFYNg9g1YFYNg9g1YFYNh9e1YF[Ng9e1YF[Ng9e1YF[Ng9e1YF[Ng9e1YF[Ng9d1ZF[Ng9e1YF[Ng9e1YF[Ng9e1YF[Ng9d1ZF\\Nf9d1ZF\\Nf9d1ZF\\Nf9d1ZF\\Nf9c1[F]Ne9c1[F]Nf9b1ZF^Nf92oEg0;WOg90PFg0:YOf90QFf09ZOf9OSFf06\\Oh9LSFh05\\Oh9KUFg04^Og9JVFh03^Oh9IUFi03^Oh9HUFk03\\Oj9GSFm04\\Oi9FSFo04ZOj9GQFP14ZOl9DQFR13YOm9DPFT13XOn9BoEV14WOn9CmEW15UOo9CmEX14UOP:AlE[14SOQ:AkE\\15SOP:@kE^14RO^:n0bERO]:o0cEPO]:P1dEoN\\:R1dEnN[:S1eElN\\:S1eEmNZ:T1fEkN[:T1fElNZ:T1eElN]:S1cElN^:S1cEmN^:R1bEmN`:Q1aEoN_:Q1aEnNa:Q1_EoNa:P1`EnNc:Q1\\EoNe:Q1[EmNh:Q1YEnNi:Q1WEmNk:R1VElNm:S1f001N3N2M3N2N3L3N2M3N2N2M4M2N2Mei>"}, "label": "small bowl of soup"}]} {"id": 541338, "image": "COCO_train2014_000000541338.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of soup behind a bowl of pasta\"."}], "task": "refering", "answer_template": "The \"a bowl of soup behind a bowl of pasta\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 224, 225, 226, 227, 228, 248, 249, 250], "bbox": [0.727359375, 0.46516355140186916, 0.9437656249999999, 0.7168457943925234], "iscrowd": 0, "area": 8781.07385, "rle": {"size": [428, 640], "counts": "n\\S62Z=4KT1mN0O2O0O2O0O2O001N101N101N101N101N101N101N101O001N101O0O2O001N101O001N10000O101O0O10000O1000000O2O0[EbM\\:^2dEcM[:]2dEfMZ:Z2fEhMX:X2hEjMV:U2kEnMR:R2nEPNP:P2PFRNn9n1RFTNl9k1UFXNh9h1XFYNg9g1YFYNg9g1YFYNh9e1YF[Ng9e1YF[Ng9e1YF[Ng9e1YF[Ng9e1YF[Ng9d1ZF[Ng9e1YF[Ng9e1YF[Ng9e1YF[Ng9d1ZF\\Nf9d1ZF\\Nf9d1ZF\\Nf9d1ZF\\Nf9c1[F]Ne9c1[F]Nf9b1ZF^Nf92oEg0;WOg90PFg0:YOf90QFf09ZOf9OSFf06\\Oh9LSFh05\\Oh9KUFg04^Og9JVFh03^Oh9IUFi03^Oh9HUFk03\\Oj9GSFm04\\Oi9FSFo04ZOj9GQFP14ZOl9DQFR13YOm9DPFT13XOn9BoEV14WOn9CmEW15UOo9CmEX14UOP:AlE[14SOQ:AkE\\15SOP:@kE^14RO^:n0bERO]:o0cEPO]:P1dEoN\\:R1dEnN[:S1eElN\\:S1eEmNZ:T1fEkN[:T1fElNZ:T1eElN]:S1cElN^:S1cEmN^:R1bEmN`:Q1aEoN_:Q1aEnNa:Q1_EoNa:P1`EnNc:Q1\\EoNe:Q1[EmNh:Q1YEnNi:Q1WEmNk:R1VElNm:S1f001N3N2M3N2N3L3N2M3N2N2M4M2N2Mei>"}, "label": "a bowl of soup behind a bowl of pasta"}]} {"id": 524966, "image": "COCO_train2014_000000524966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"it is a young black male wearing black rimmed glasses with his right foot on a soccer ball\"."}], "task": "refering", "answer_template": "The \"it is a young black male wearing black rimmed glasses with his right foot on a soccer ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 38, 39, 40, 61, 62, 63, 64, 65, 83, 84, 85, 86, 87, 88, 89, 105, 106, 107, 108, 109, 110, 111, 112, 128, 129, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 199, 200, 202, 203, 222, 223, 225, 226, 245, 246, 248, 249, 271, 272, 295], "bbox": [0.572734375, 0.054074941451990634, 0.890578125, 0.8136299765807962], "iscrowd": 0, "area": 25970.204149999998, "rle": {"size": [427, 640], "counts": "\\Vi45Q=5K5K5O1O1N2O1O1N2O1N2O2N1N2O1O1N100O10000O2O000O100O101O000001O0000010O0000001O0001O01O0000001O01O0001O000001QFgNi01V4W1PKPOb0K]4U1QKXO:Dd4T1QK@3_Oj4Q1SKGKZOR5n0RK0DTOY5l0RK8^OmN_5l0RK?VOhNf5i0SKg0oNbNn5f0SKo0gN]NU6d0SKW1aNVN[6c0SK_1ZNQNa6`0UK\\2j4dMUK^2k4aMUK`2j4`MUKc2i4]MWKd2h4\\MWKf2j3SL^KV1g0i2i3\\LVKk0Q1k2g3eLmJ`0[1m2g3SNYLn1e3SNZLo1e3QN[LQ2c3oM\\LS2b3nM]LT2b3lM^LV2`3kM^LW2`3jM`LW2_3lM]LV2b3RNULQ2h3WNQLPOWOj0g4>iKjNDd0b4i0cKcN0`0[4V1\\K[N==U4[3PLaLo3`3TLVLQ4k3TLhKT4X4QLZKY4e4kKoJ^4Q5fKbJb4^5f1O1O10O01O1O10O01O1O2O1N2N2N2O1O010001O0000000O10000000000003N9F6J0000O100O100O1O2O2M2O1N3N1N3N1N2O2YJfJW3\\5`LlJ`3V5VLRKi3P5nKXKR4i4fK^KY4e4]KcKc4^4TKjKk4Y4kJoKU5T6O100O100O1O100O10mJSKg1l4WN\\Kd1d4ZNcK`1]4^NkK]1U4aNSLY1l3fN[LT1f3iNbLQ1^3nNiLl0X3QOPMi0P3VOWMc0k2ZO^M?c2_OdM;]2BkM8V2FQN@PM`NP5m1XN]NiMDQ4m17RNJk18UNIi19VNHg1;XNFf1]NC`1`0_NA_1`0aNA\\1Y6M3M3M3M3M3M3M3M3M3N2M3M3M3M3MbRm0"}, "label": "it is a young black male wearing black rimmed glasses with his right foot on a soccer ball"}]} {"id": 524966, "image": "COCO_train2014_000000524966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy wearing glasses and white cleats with his right foot on a soccer ball\"."}], "task": "refering", "answer_template": "The \"a boy wearing glasses and white cleats with his right foot on a soccer ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 38, 39, 40, 61, 62, 63, 64, 65, 83, 84, 85, 86, 87, 88, 89, 105, 106, 107, 108, 109, 110, 111, 112, 128, 129, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 199, 200, 202, 203, 222, 223, 225, 226, 245, 246, 248, 249, 271, 272, 295], "bbox": [0.572734375, 0.054074941451990634, 0.890578125, 0.8136299765807962], "iscrowd": 0, "area": 25970.204149999998, "rle": {"size": [427, 640], "counts": "\\Vi45Q=5K5K5O1O1N2O1O1N2O1N2O2N1N2O1O1N100O10000O2O000O100O101O000001O0000010O0000001O0001O01O0000001O01O0001O000001QFgNi01V4W1PKPOb0K]4U1QKXO:Dd4T1QK@3_Oj4Q1SKGKZOR5n0RK0DTOY5l0RK8^OmN_5l0RK?VOhNf5i0SKg0oNbNn5f0SKo0gN]NU6d0SKW1aNVN[6c0SK_1ZNQNa6`0UK\\2j4dMUK^2k4aMUK`2j4`MUKc2i4]MWKd2h4\\MWKf2j3SL^KV1g0i2i3\\LVKk0Q1k2g3eLmJ`0[1m2g3SNYLn1e3SNZLo1e3QN[LQ2c3oM\\LS2b3nM]LT2b3lM^LV2`3kM^LW2`3jM`LW2_3lM]LV2b3RNULQ2h3WNQLPOWOj0g4>iKjNDd0b4i0cKcN0`0[4V1\\K[N==U4[3PLaLo3`3TLVLQ4k3TLhKT4X4QLZKY4e4kKoJ^4Q5fKbJb4^5f1O1O10O01O1O10O01O1O2O1N2N2N2O1O010001O0000000O10000000000003N9F6J0000O100O100O1O2O2M2O1N3N1N3N1N2O2YJfJW3\\5`LlJ`3V5VLRKi3P5nKXKR4i4fK^KY4e4]KcKc4^4TKjKk4Y4kJoKU5T6O100O100O1O100O10mJSKg1l4WN\\Kd1d4ZNcK`1]4^NkK]1U4aNSLY1l3fN[LT1f3iNbLQ1^3nNiLl0X3QOPMi0P3VOWMc0k2ZO^M?c2_OdM;]2BkM8V2FQN@PM`NP5m1XN]NiMDQ4m17RNJk18UNIi19VNHg1;XNFf1]NC`1`0_NA_1`0aNA\\1Y6M3M3M3M3M3M3M3M3M3N2M3M3M3M3MbRm0"}, "label": "a boy wearing glasses and white cleats with his right foot on a soccer ball"}]} {"id": 524966, "image": "COCO_train2014_000000524966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy wearing yellow shoes\"."}], "task": "refering", "answer_template": "The \"a young boy wearing yellow shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 33, 34, 56, 57, 58, 79, 80, 81, 102, 103, 104, 125, 126, 127, 148, 149, 150, 171, 172, 173, 174, 194, 195, 196, 197], "bbox": [0.434265625, 0.03362997658079625, 0.584546875, 0.6079625292740047], "iscrowd": 0, "area": 12506.419750000001, "rle": {"size": [427, 640], "counts": "joc34U=8H8I0ON25K8H8I7H9G8H8I7G8F9G:F:TI`L^3i3TLbLg3g3lKaLS4e3eK_LZ4e3^K_La4g3WK]Lh4h3PK\\Lo4h3eJ`LZ5f3UJgLj5]5O2O0O2O1N101N101O0O2O0ON3M3L6H9G9F:F:D<@?G8010O10000O010O100O10O0100O10O3N2N30O00VLjG[2W8ZMSHg2l7PM]HP3c7fLgH[3e801L4L4M3L5K4L4M4K4L5K6K4K3M3N3L3M3M4L3L4L5J6K6J7H7Jke^3"}, "label": "a young boy wearing yellow shoes"}]} {"id": 524966, "image": "COCO_train2014_000000524966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy in the middle with yellow shoes\"."}], "task": "refering", "answer_template": "The \"the boy in the middle with yellow shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 33, 34, 56, 57, 58, 79, 80, 81, 102, 103, 104, 125, 126, 127, 148, 149, 150, 171, 172, 173, 174, 194, 195, 196, 197], "bbox": [0.434265625, 0.03362997658079625, 0.584546875, 0.6079625292740047], "iscrowd": 0, "area": 12506.419750000001, "rle": {"size": [427, 640], "counts": "joc34U=8H8I0ON25K8H8I7H9G8H8I7G8F9G:F:TI`L^3i3TLbLg3g3lKaLS4e3eK_LZ4e3^K_La4g3WK]Lh4h3PK\\Lo4h3eJ`LZ5f3UJgLj5]5O2O0O2O1N101N101O0O2O0ON3M3L6H9G9F:F:D<@?G8010O10000O010O100O10O0100O10O3N2N30O00VLjG[2W8ZMSHg2l7PM]HP3c7fLgH[3e801L4L4M3L5K4L4M4K4L5K6K4K3M3N3L3M3M4L3L4L5J6K6J7H7Jke^3"}, "label": "the boy in the middle with yellow shoes"}]} {"id": 524966, "image": "COCO_train2014_000000524966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a youth male soccer player with a black and white uniform and blue cleats on\"."}], "task": "refering", "answer_template": "The \"a youth male soccer player with a black and white uniform and blue cleats on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 35, 36, 37, 58, 59, 60, 81, 82, 83, 104, 105, 106, 127, 129, 152], "bbox": [0.5173281249999999, 0.0, 0.64096875, 0.42365339578454336], "iscrowd": 0, "area": 6448.8368500000015, "rle": {"size": [427, 640], "counts": "_QZ41Y=3N1N3M3M2O0O2N1O1O2O0O2N101N101N2iFYOg5i0RJDg5=RJ1e52RJ in this image.", "objects": [{"patches": [12, 13, 35, 36, 37, 58, 59, 60, 81, 82, 83, 104, 105, 106, 127, 129, 152], "bbox": [0.5173281249999999, 0.0, 0.64096875, 0.42365339578454336], "iscrowd": 0, "area": 6448.8368500000015, "rle": {"size": [427, 640], "counts": "_QZ41Y=3N1N3M3M2O0O2N1O1O2O0O2N101N101N2iFYOg5i0RJDg5=RJ1e52RJ in this image.", "objects": [{"patches": [23, 46, 115, 116, 117, 138, 139, 161, 184, 207, 230, 231, 253, 254, 276, 299, 322], "bbox": [0.0, 0.10866510538641685, 0.097046875, 0.9839812646370023], "iscrowd": 0, "area": 6477.210399999999, "rle": {"size": [427, 640], "counts": "_14X=9QFY2g3QN`Kn0bMb1U5hMPMg0QNT3V3RLjNl0VN_4]1bJ?P1ZNW4X1fJ`0T1^No3V1hJ=Y1dNh3T1hJ:a1dNd3W1dJ7i1dN`3h3aLZL]3f3cL]LY3d3gL^LW3b3iLaLT3_3mLbLQ3^3QLSKMa1P4[3TLUKKc1n3X3WLUKKe1k3W3YLVKKf1h3U3]LVKKf1f3T3^LWKLh1b3R3bLWKJj1b3n2eLXKIm1a2i3eM\\JHn1Y2n3oMTJGQ2P2S4YNmIET2h1V4bNhIET2_1\\4lN`IDX2U1`4VOZICY2n0c4@TIA\\2e0h4JmH_O^2=m43fH]Ob27o4<`HYOf21R5f0XHVOY;i0d010O10O10O010O0100O010O10O10000000003M3M3L3K4N2M3N2M3Mkk`7"}, "label": "a man playing soccer in a red shirt"}]} {"id": 524966, "image": "COCO_train2014_000000524966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"partiall visible soccer player in red sleaved uniform\"."}], "task": "refering", "answer_template": "The \"partiall visible soccer player in red sleaved uniform\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 46, 115, 116, 117, 138, 139, 161, 184, 207, 230, 231, 253, 254, 276, 299, 322], "bbox": [0.0, 0.10866510538641685, 0.097046875, 0.9839812646370023], "iscrowd": 0, "area": 6477.210399999999, "rle": {"size": [427, 640], "counts": "_14X=9QFY2g3QN`Kn0bMb1U5hMPMg0QNT3V3RLjNl0VN_4]1bJ?P1ZNW4X1fJ`0T1^No3V1hJ=Y1dNh3T1hJ:a1dNd3W1dJ7i1dN`3h3aLZL]3f3cL]LY3d3gL^LW3b3iLaLT3_3mLbLQ3^3QLSKMa1P4[3TLUKKc1n3X3WLUKKe1k3W3YLVKKf1h3U3]LVKKf1f3T3^LWKLh1b3R3bLWKJj1b3n2eLXKIm1a2i3eM\\JHn1Y2n3oMTJGQ2P2S4YNmIET2h1V4bNhIET2_1\\4lN`IDX2U1`4VOZICY2n0c4@TIA\\2e0h4JmH_O^2=m43fH]Ob27o4<`HYOf21R5f0XHVOY;i0d010O10O10O010O0100O010O10O10000000003M3M3L3K4N2M3N2M3Mkk`7"}, "label": "partiall visible soccer player in red sleaved uniform"}]} {"id": 385704, "image": "COCO_train2014_000000385704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the buttoned jacket\"."}], "task": "refering", "answer_template": "The \"the man in the buttoned jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 63, 64, 65, 78, 79, 80, 93, 94, 95, 108, 109, 110, 111, 122, 123, 124, 125, 126, 137, 138, 139, 140, 141, 142, 152, 153, 154, 155, 156, 157, 166, 167, 168, 169, 170, 171, 172, 181, 182, 183, 184, 185, 186, 187, 197, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 227, 228, 229, 230, 231, 232, 233, 242, 243, 244, 245, 246, 247, 248, 257, 258, 259, 260, 261, 262, 272, 273, 274, 275, 276, 277, 288, 289, 290, 291, 292, 303, 304, 305, 306, 307, 308, 319, 320, 321, 322, 335, 336, 337], "bbox": [0.12586854460093896, 0.15546875, 0.5677934272300469, 0.983671875], "iscrowd": 0, "area": 60829.2603, "rle": {"size": [640, 426], "counts": "VRR1b0Vc0c0\\Od0]Oc0]O`0_O:fNkL_A]3`>kLQA]3o>jLa@`3^?a0O101eCfKm8[4iFPLV9P4cFWL\\9k3`FYL_9g3^F]La9d3[F`Lc9a3[FbLd9^3YFfLe9\\3WFhLh9X3UFlLi9U3TFoLk9R3QFRMn9n2oEVMo9k2oEXMP:i2lE[MR:f2kE^MT:c2hEaMV:`2gEdMX:\\2eEhMY:Z2cEiM]:W2`EmM^:T2`EoM^:S2^EQNa:o1\\EUNb:l1[EXNc:j1YEZNe:g1XE]Nf:d1WE`Ng:^5N2mLXGkJj8R5RHSJP8j5SHTJo7i5UHTJm7i5VHUJl7h5XHUJj7g5[HVJf7h5]HVJe7g5_HVJe7e5^HYJg7a5]H[Ji7^5[H`Ji7\\5ZH`Jl7\\5VH`JP8\\5RH`JT8\\5nG`JX8[5kGbJZ8Z5hGbJ^8Z5X3L3M4L4L4L4L1O00000000001O00000000001O00000000001O00000000001O000101O10O01O1O10O01O1O100O001O100O001O103L4L4L4fMgJeE\\5eNQKX:ERG^5WNcK[:QO]Gm6`8VI_Gk6_8WIaGi6]8YIaGj6\\8XIaGk6\\8XIbGj6\\8XIbGj6\\8XIbGj6[8YIcGj6[8WIcGk6]8UIaGm6_8RI`Gn6c8QI[GP7f8PIXGQ7h8PIVGP7j8RITGo6k8RITGo6l8QISGo6m8RIRGo6n8QIPGQ7o8PIPGP7Q9PInFQ7Q9PInFQ7Q9PInFP7S9PIlFQ7S9PIlFQ7T9oHkFQ7U9PIjFQ7V9oHiFR7V9oHiFQ7W9PIhFQ7X9oHgF[7o8fHPGd7g8\\HWGo7_8RH`GY8V8gGiGc8m7^GRHm8k6nE[IU1IW9^6iEhIQ1Ia9Q6aEVJo0Hj9e5[EcJk0Gm9_5\\EjJh0Fn9[5^EoJd0EP:V5`EUKb0CQ:R5aE[K>BS:n4cE_K;BT:i4eEeK8AU:e4gEjK4@W:a4hEoK2^OY:]4jEULM]O[:Y4lEYLJ]O]:S4mEaLF[O_:o3oEeLc;V3`DjLb;P3cDPM^;k2fDTM\\;e2iD\\MY;W2REhMR;j1XEUNn:]1\\EbNh:P1cEnNc:c0gE]O]:6mEHY:IRF5n>O1N1O2O1N2O1NWRc3"}, "label": "the man in the buttoned jacket"}]} {"id": 385704, "image": "COCO_train2014_000000385704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man dressed in a texido holding some clothing standing in a parking lot with another man\"."}], "task": "refering", "answer_template": "The \"a young man dressed in a texido holding some clothing standing in a parking lot with another man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 63, 64, 65, 78, 79, 80, 93, 94, 95, 108, 109, 110, 111, 122, 123, 124, 125, 126, 137, 138, 139, 140, 141, 142, 152, 153, 154, 155, 156, 157, 166, 167, 168, 169, 170, 171, 172, 181, 182, 183, 184, 185, 186, 187, 197, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 227, 228, 229, 230, 231, 232, 233, 242, 243, 244, 245, 246, 247, 248, 257, 258, 259, 260, 261, 262, 272, 273, 274, 275, 276, 277, 288, 289, 290, 291, 292, 303, 304, 305, 306, 307, 308, 319, 320, 321, 322, 335, 336, 337], "bbox": [0.12586854460093896, 0.15546875, 0.5677934272300469, 0.983671875], "iscrowd": 0, "area": 60829.2603, "rle": {"size": [640, 426], "counts": "VRR1b0Vc0c0\\Od0]Oc0]O`0_O:fNkL_A]3`>kLQA]3o>jLa@`3^?a0O101eCfKm8[4iFPLV9P4cFWL\\9k3`FYL_9g3^F]La9d3[F`Lc9a3[FbLd9^3YFfLe9\\3WFhLh9X3UFlLi9U3TFoLk9R3QFRMn9n2oEVMo9k2oEXMP:i2lE[MR:f2kE^MT:c2hEaMV:`2gEdMX:\\2eEhMY:Z2cEiM]:W2`EmM^:T2`EoM^:S2^EQNa:o1\\EUNb:l1[EXNc:j1YEZNe:g1XE]Nf:d1WE`Ng:^5N2mLXGkJj8R5RHSJP8j5SHTJo7i5UHTJm7i5VHUJl7h5XHUJj7g5[HVJf7h5]HVJe7g5_HVJe7e5^HYJg7a5]H[Ji7^5[H`Ji7\\5ZH`Jl7\\5VH`JP8\\5RH`JT8\\5nG`JX8[5kGbJZ8Z5hGbJ^8Z5X3L3M4L4L4L4L1O00000000001O00000000001O00000000001O00000000001O000101O10O01O1O10O01O1O100O001O100O001O103L4L4L4fMgJeE\\5eNQKX:ERG^5WNcK[:QO]Gm6`8VI_Gk6_8WIaGi6]8YIaGj6\\8XIaGk6\\8XIbGj6\\8XIbGj6\\8XIbGj6[8YIcGj6[8WIcGk6]8UIaGm6_8RI`Gn6c8QI[GP7f8PIXGQ7h8PIVGP7j8RITGo6k8RITGo6l8QISGo6m8RIRGo6n8QIPGQ7o8PIPGP7Q9PInFQ7Q9PInFQ7Q9PInFP7S9PIlFQ7S9PIlFQ7T9oHkFQ7U9PIjFQ7V9oHiFR7V9oHiFQ7W9PIhFQ7X9oHgF[7o8fHPGd7g8\\HWGo7_8RH`GY8V8gGiGc8m7^GRHm8k6nE[IU1IW9^6iEhIQ1Ia9Q6aEVJo0Hj9e5[EcJk0Gm9_5\\EjJh0Fn9[5^EoJd0EP:V5`EUKb0CQ:R5aE[K>BS:n4cE_K;BT:i4eEeK8AU:e4gEjK4@W:a4hEoK2^OY:]4jEULM]O[:Y4lEYLJ]O]:S4mEaLF[O_:o3oEeLc;V3`DjLb;P3cDPM^;k2fDTM\\;e2iD\\MY;W2REhMR;j1XEUNn:]1\\EbNh:P1cEnNc:c0gE]O]:6mEHY:IRF5n>O1N1O2O1N2O1NWRc3"}, "label": "a young man dressed in a texido holding some clothing standing in a parking lot with another man"}]} {"id": 385704, "image": "COCO_train2014_000000385704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the purple tie\"."}], "task": "refering", "answer_template": "The \"the man with the purple tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 100, 101, 113, 115, 116, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 158, 159, 160, 161, 162, 173, 174, 175, 176, 177, 188, 189, 190, 191, 192, 203, 204, 205, 206, 207, 219, 220, 221, 222, 234, 235, 236, 249, 250, 251, 264, 265, 266, 279], "bbox": [0.5244600938967136, 0.23760937499999998, 0.8594600938967136, 0.832203125], "iscrowd": 0, "area": 26562.086649999997, "rle": {"size": [640, 426], "counts": "ec[41lc04K4L5L4K5K5K5L3N3O100O1O1O00100k@gNa;Z1XDmNg;T1RDSOm;n0lCXOUZO`Am0_>XOVAo0i>U2O0dDRKf7o4jGcKT8]4\\GULa8l3UG_Li8a3SGfLj8[3QGkLn8U3mFRMP9o2kFWMS9i2jF]MS9V3VFPMh9g6N1N0000O2O0O100O101O0O1hNlEdHT:X7]FZHe9b7jFRHV9j7b1L4M3L3N0O1O2OQOgHZEX7i:PImDo6U;[I`Dc6d;h0^ESHZ9h7cFaH\\9[7_FPI_9k6bFZI\\9c6gF`IW9[6nFfIQ9V6RGmIl8n5YGSJe8j5^GYJ`8h5_GYJ`8j5\\GXJc8k5YGXJd8l5YGUJf8o5UGTJi8o5TGRJk8Q6QGRJl8R6QGoIn8U6mFnIQ9U6kFmIT9V6iFlIU9W6gFkIW9Y6fFiIZ9Y6aFiI`9[8100O100O100O1ZO[FPGe9m8^FRGc9l8`FSG`9j8cFUG^9X1^F`58WIZ9T1cFb56YIX9P1hFc54[IU9m0mFd51_IR9h0SGe5OaIo8e0WGh5LcIm8`0]Gi5JeIj8=bGj5GiIg89fGj5HkIc86kGi5GQJ^81QHg5HVJX8NVHf5G\\J[:_5kE_JW:[5nEeJR:V5TFhJn9R5WFmJj9m4i2VMi2^Ob0O2N101N1O1O2O0O2N1O101N1O2N100O2N6J5L5J5K6JZVU1"}, "label": "the man with the purple tie"}]} {"id": 385704, "image": "COCO_train2014_000000385704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older man\"."}], "task": "refering", "answer_template": "The \"an older man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 100, 101, 113, 115, 116, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 158, 159, 160, 161, 162, 173, 174, 175, 176, 177, 188, 189, 190, 191, 192, 203, 204, 205, 206, 207, 219, 220, 221, 222, 234, 235, 236, 249, 250, 251, 264, 265, 266, 279], "bbox": [0.5244600938967136, 0.23760937499999998, 0.8594600938967136, 0.832203125], "iscrowd": 0, "area": 26562.086649999997, "rle": {"size": [640, 426], "counts": "ec[41lc04K4L5L4K5K5K5L3N3O100O1O1O00100k@gNa;Z1XDmNg;T1RDSOm;n0lCXOUZO`Am0_>XOVAo0i>U2O0dDRKf7o4jGcKT8]4\\GULa8l3UG_Li8a3SGfLj8[3QGkLn8U3mFRMP9o2kFWMS9i2jF]MS9V3VFPMh9g6N1N0000O2O0O100O101O0O1hNlEdHT:X7]FZHe9b7jFRHV9j7b1L4M3L3N0O1O2OQOgHZEX7i:PImDo6U;[I`Dc6d;h0^ESHZ9h7cFaH\\9[7_FPI_9k6bFZI\\9c6gF`IW9[6nFfIQ9V6RGmIl8n5YGSJe8j5^GYJ`8h5_GYJ`8j5\\GXJc8k5YGXJd8l5YGUJf8o5UGTJi8o5TGRJk8Q6QGRJl8R6QGoIn8U6mFnIQ9U6kFmIT9V6iFlIU9W6gFkIW9Y6fFiIZ9Y6aFiI`9[8100O100O100O1ZO[FPGe9m8^FRGc9l8`FSG`9j8cFUG^9X1^F`58WIZ9T1cFb56YIX9P1hFc54[IU9m0mFd51_IR9h0SGe5OaIo8e0WGh5LcIm8`0]Gi5JeIj8=bGj5GiIg89fGj5HkIc86kGi5GQJ^81QHg5HVJX8NVHf5G\\J[:_5kE_JW:[5nEeJR:V5TFhJn9R5WFmJj9m4i2VMi2^Ob0O2N101N1O1O2O0O2N1O101N1O2N100O2N6J5L5J5K6JZVU1"}, "label": "an older man"}]} {"id": 197289, "image": "COCO_train2014_000000197289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwich on the right side of the plate has tomato on it\"."}], "task": "refering", "answer_template": "The \"the sandwich on the right side of the plate has tomato on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 58, 59, 60, 61, 62, 63, 64, 65, 66, 81, 82, 83, 84, 85, 86, 87, 88, 104, 105, 106, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 174, 175, 176, 177, 198, 199], "bbox": [0.5337500000000001, 0.0018266978922716628, 0.939265625, 0.5836299765807963], "iscrowd": 0, "area": 39923.09145, "rle": {"size": [427, 640], "counts": "Ve^46Q=7I7I6J7I6J7I6J7XFXNc7n1[HSN`7T2\\HnM_7Y2]HiM`7[2]HgMb7\\2[HeMd7]2YHeMf7]2WHfMg7[2WHgMh7Z2WHgMh7Z2VHhMi7Y2VHhMi7Y2VHhMi7Y2VHiMh7X2WHiMh7X2WHiMh7X2WHiMh7X2WHiMh7X2VHjMi7W2VHjMi7W2VHkMh7V2WHkMh7Y2THiMj7\\2QHgMl7]2PHfMm7^2nGeMP8n3TIjJ]5Y5VJRKi5P5jI\\KU6n5O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O00O100O100O1O100O100O100O100O1O100O100O100O10000O1000000O100000O0100000000O1000000O1O1O1O100O1O1O1O1O1O100O1WObIdJ_6W5hIfJY6U5nIhJS6S5TJjJm5Q5ZJmJf5n4aJoJ`5P5cJmJ^5R5eJkJ\\5T5^100O1O100O1O1O100O100000O010000000000000000000000O10000000000000000000000O100000000O10000O1000000O10000O1000000O01000O10000O100O100O100O100O1O100O1jNWGjMi8U2YGiMh8V2ZGiMf8V2\\GhMe8W2]GhMc8W2_GgMb8X2aGeM`8Z2bGeM^8Z2dGcM^8[2eGbM]8]2eG_M^8`2dG]M^8b2eGYM^8f2dGWM^8h2S1O1O1O1O1O1O1O1O1O1O1O0O2O1O1O1O1O1O1O1O1O1O1O1E;O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O107H9G9G:F]W`0"}, "label": "the sandwich on the right side of the plate has tomato on it"}]} {"id": 197289, "image": "COCO_train2014_000000197289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green broccoli next to other assorted foods\"."}], "task": "refering", "answer_template": "The \"green broccoli next to other assorted foods\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 342, 343, 344], "bbox": [0.582015625, 0.2950351288056206, 1.0, 0.9454800936768151], "iscrowd": 0, "area": 56365.0654, "rle": {"size": [427, 640], "counts": "a[k42Y=a0_Oa0^Oc0^O?A1N3N1O1N2O1O1N3N1N2O1O1N2O2N1N2O1O1N2O2N1N2O1O1N3N1O1N2O1N2O2N1N2O1O1N2O2N1N2O1O1N2O2N1N10001N10000O2O0O10001N10000O101O0O10001N10000O2O000O101O0O100O101O0O10001N10000`NlKhIV4V6`1O1O101N1O1O1O101N1O1O1O2N100O1O2N1O1O101N1O1O1O101N1O1O1O2N100O1O2N1O100O2N1O1O1O101O0000001O0O10001O00000O2O000000001O0O10001O0000001N1000001O000O10001O0000001N1000001O01O0010O01O000010O01O00010O001O010O000010O01O000010O01O0010O0001O010O00001O001O001O00001O001O00001O001O001O00001O001O00001OO100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1RL"}, "label": "green broccoli next to other assorted foods"}]} {"id": 197289, "image": "COCO_train2014_000000197289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"it is the broccoli next to the beans and the slider with a tomato and a slice of the cheese on it\"."}], "task": "refering", "answer_template": "The \"it is the broccoli next to the beans and the slider with a tomato and a slice of the cheese on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 342, 343, 344], "bbox": [0.582015625, 0.2950351288056206, 1.0, 0.9454800936768151], "iscrowd": 0, "area": 56365.0654, "rle": {"size": [427, 640], "counts": "a[k42Y=a0_Oa0^Oc0^O?A1N3N1O1N2O1O1N3N1N2O1O1N2O2N1N2O1O1N2O2N1N2O1O1N3N1O1N2O1N2O2N1N2O1O1N2O2N1N2O1O1N2O2N1N10001N10000O2O0O10001N10000O101O0O10001N10000O2O000O101O0O100O101O0O10001N10000`NlKhIV4V6`1O1O101N1O1O1O101N1O1O1O2N100O1O2N1O1O101N1O1O1O101N1O1O1O2N100O1O2N1O100O2N1O1O1O101O0000001O0O10001O00000O2O000000001O0O10001O0000001N1000001O000O10001O0000001N1000001O01O0010O01O000010O01O00010O001O010O000010O01O000010O01O0010O0001O010O00001O001O001O00001O001O00001O001O001O00001O001O00001OO100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1RL"}, "label": "it is the broccoli next to the beans and the slider with a tomato and a slice of the cheese on it"}]} {"id": 197289, "image": "COCO_train2014_000000197289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwich with no bread on top\"."}], "task": "refering", "answer_template": "The \"the sandwich with no bread on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 281, 283, 284, 287, 288, 310, 311, 312], "bbox": [0.142109375, 0.2801639344262295, 0.6267968749999999, 0.9214051522248243], "iscrowd": 0, "area": 55315.621799999986, "rle": {"size": [427, 640], "counts": "VXV14T=6I8I6J7I7H4M3O01O00010O00010O000010O01O01O01O01bLnNbJR1^5]OTJb0l5DnIVICj6>TIBl6_30O010O010O10O010O10O10O010O010O01000O2O0O1000001O00000O2O000000001O000O101O0000001O000O10001O0000001O00O0100000000000000O100000000000000O10O10000000000000O100000000000000O1000O1000000000O100000000000000O0100000000000000O100000001O000O1000001O00000O101O000000000O2O000000001O0O1000000100O1O1O2N100O1O1O1O2O0cJSIZ4n6cKTI^4l6_KVIa4k6]KWIc4i6ZKYIg4h6UKZIk4g6RK[Io4e6nJ]IR5d6kJ_IU5a6iJ`IW5a6fJaI[5W700100O001O010O001O001O010O1O001O010O001O001O010O1aN[JVKe5h4^JWKc5f4`JYK`5e4dJYK]5d4fJ\\KY5b4jJ]KW5`4mJ^KS5`4PK_KQ5^4RKaKn4]4VKaKk4\\4XKdKh4Y4\\KeKd4Z4^KeKc4X4`KgKa4V4cKhK`4S4cKkK`4Q4cKmKa4n3cKoK`4m3cKPLb4k3aKSLc4h3aKTLc4h3`KVLd4e3_KYLd4c3`KYLe4h0mHP2g9jM_FT2d9dMdFY2_9_MiF^2T:O1N2N100O1O100O2N1O100O1O101N1O100O2N8H8I7H8H8I7HS\\S3"}, "label": "the sandwich with no bread on top"}]} {"id": 197289, "image": "COCO_train2014_000000197289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the burger that only has a tomato and cheese on top of it\"."}], "task": "refering", "answer_template": "The \"the burger that only has a tomato and cheese on top of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 281, 283, 284, 287, 288, 310, 311, 312], "bbox": [0.142109375, 0.2801639344262295, 0.6267968749999999, 0.9214051522248243], "iscrowd": 0, "area": 55315.621799999986, "rle": {"size": [427, 640], "counts": "VXV14T=6I8I6J7I7H4M3O01O00010O00010O000010O01O01O01O01bLnNbJR1^5]OTJb0l5DnIVICj6>TIBl6_30O010O010O10O010O10O10O010O010O01000O2O0O1000001O00000O2O000000001O000O101O0000001O000O10001O0000001O00O0100000000000000O100000000000000O10O10000000000000O100000000000000O1000O1000000000O100000000000000O0100000000000000O100000001O000O1000001O00000O101O000000000O2O000000001O0O1000000100O1O1O2N100O1O1O1O2O0cJSIZ4n6cKTI^4l6_KVIa4k6]KWIc4i6ZKYIg4h6UKZIk4g6RK[Io4e6nJ]IR5d6kJ_IU5a6iJ`IW5a6fJaI[5W700100O001O010O001O001O010O1O001O010O001O001O010O1aN[JVKe5h4^JWKc5f4`JYK`5e4dJYK]5d4fJ\\KY5b4jJ]KW5`4mJ^KS5`4PK_KQ5^4RKaKn4]4VKaKk4\\4XKdKh4Y4\\KeKd4Z4^KeKc4X4`KgKa4V4cKhK`4S4cKkK`4Q4cKmKa4n3cKoK`4m3cKPLb4k3aKSLc4h3aKTLc4h3`KVLd4e3_KYLd4c3`KYLe4h0mHP2g9jM_FT2d9dMdFY2_9_MiF^2T:O1N2N100O1O100O2N1O100O1O101N1O100O2N8H8I7H8H8I7HS\\S3"}, "label": "the burger that only has a tomato and cheese on top of it"}]} {"id": 197289, "image": "COCO_train2014_000000197289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of bread on top of some kind of meat which is half cut off from the picture\"."}], "task": "refering", "answer_template": "The \"a piece of bread on top of some kind of meat which is half cut off from the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 49, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 253, 254, 255, 256], "bbox": [0.0, 0.15730679156908667, 0.19940625, 0.7865105386416863], "iscrowd": 0, "area": 28817.828649999996, "rle": {"size": [427, 640], "counts": "X2_1k;o2QMo2QMn0RO010000O1000000O10000O1000000O10000O1000000O10O10O10000O1000000000000O101O000000000O100000001O00000O1000000000001N1000000000000O1000010O0000000000000000001O01O00000001O001O1O1O001N2L3N3L4M2M4L4M3L3N3L4M2M4M3N1N3M5K5L4K5^M`HQOd7i0RIdNR7T1fIWN_6b1WJjMn5n1m2I6J7H8I7H8I6J7H8I7Ii[e6"}, "label": "a piece of bread on top of some kind of meat which is half cut off from the picture"}]} {"id": 197289, "image": "COCO_train2014_000000197289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a burger slider that is to the left of all the other burgers\"."}], "task": "refering", "answer_template": "The \"a burger slider that is to the left of all the other burgers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 49, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 253, 254, 255, 256], "bbox": [0.0, 0.15730679156908667, 0.19940625, 0.7865105386416863], "iscrowd": 0, "area": 28817.828649999996, "rle": {"size": [427, 640], "counts": "X2_1k;o2QMo2QMn0RO010000O1000000O10000O1000000O10000O1000000O10O10O10000O1000000000000O101O000000000O100000001O00000O1000000000001N1000000000000O1000010O0000000000000000001O01O00000001O001O1O1O001N2L3N3L4M2M4L4M3L3N3L4M2M4M3N1N3M5K5L4K5^M`HQOd7i0RIdNR7T1fIWN_6b1WJjMn5n1m2I6J7H8I7H8I6J7H8I7Ii[e6"}, "label": "a burger slider that is to the left of all the other burgers"}]} {"id": 221880, "image": "COCO_train2014_000000221880.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black dress with a red shawl standing next to a man\"."}], "task": "refering", "answer_template": "The \"a woman in a black dress with a red shawl standing next to a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 291, 292, 293, 294, 310, 311, 312, 313, 314, 315, 316, 317, 333, 334, 335, 336, 337, 338, 339, 340, 356, 357, 358, 359, 360, 361, 362, 363, 379, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.483703125, 0.2, 0.8157343749999999, 0.99775], "iscrowd": 0, "area": 58450.65345, "rle": {"size": [480, 640], "counts": "`^a4=Y>?G:H7PGTO_3T1PLCc3d0jK5i30jKe0i3@nKQ1d3UORL]1a3iNVLg1]3^N_Ln1U3VNgLV2m2lMQM`2c2cMZMi2Y2ZMeMQ3P2RMmMZ3g1hLWN]3d1fLYN[3f1gLXNZ3h1hLUNY3j1iLTNX3l1iLRNX3n1jLoMW3P2kLoMU3Q2mLlMT3S2nLkMS3U2oLhMR3W2QMfMP3Z2RMcMo2\\2SMbMn2^2TM_Mm2a2UM\\Ml2c2WMZMj2f2XMWMi2i2XMUMi2k2YMRMh2n2ZMoLg2Q3d4000000000000O1000000000000000000000000000000001O00000000000000000000010O00000000000000001O1O001O1O001O1O1O001O1O1O001O1O1O001O1O1O2N2N2N2N2N2N2N4fFnL^6V3VIUMg6o2nHXMR7m2[HbMd7m401O1O001O1O001O001O1O001O001O001O00001O001O00001O001O001O00001O001O001O00001O010O001O001O001O1O001O1O001O1O001O1O001O1O1O1O2N1O2N2N2N2N2N2N2N2N9G4L4L4L4L4L5K4L3L4M2M4L4XNfEcN]:Z1hE`N]:\\1iE^N\\:]1jE\\N\\:^1kEYN\\:c1e1K5J6Km0RO7J5J6K5J`mf1"}, "label": "a woman in a black dress with a red shawl standing next to a man"}]} {"id": 221880, "image": "COCO_train2014_000000221880.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady in black top and red shawl\"."}], "task": "refering", "answer_template": "The \"lady in black top and red shawl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 291, 292, 293, 294, 310, 311, 312, 313, 314, 315, 316, 317, 333, 334, 335, 336, 337, 338, 339, 340, 356, 357, 358, 359, 360, 361, 362, 363, 379, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.483703125, 0.2, 0.8157343749999999, 0.99775], "iscrowd": 0, "area": 58450.65345, "rle": {"size": [480, 640], "counts": "`^a4=Y>?G:H7PGTO_3T1PLCc3d0jK5i30jKe0i3@nKQ1d3UORL]1a3iNVLg1]3^N_Ln1U3VNgLV2m2lMQM`2c2cMZMi2Y2ZMeMQ3P2RMmMZ3g1hLWN]3d1fLYN[3f1gLXNZ3h1hLUNY3j1iLTNX3l1iLRNX3n1jLoMW3P2kLoMU3Q2mLlMT3S2nLkMS3U2oLhMR3W2QMfMP3Z2RMcMo2\\2SMbMn2^2TM_Mm2a2UM\\Ml2c2WMZMj2f2XMWMi2i2XMUMi2k2YMRMh2n2ZMoLg2Q3d4000000000000O1000000000000000000000000000000001O00000000000000000000010O00000000000000001O1O001O1O001O1O1O001O1O1O001O1O1O001O1O1O2N2N2N2N2N2N2N4fFnL^6V3VIUMg6o2nHXMR7m2[HbMd7m401O1O001O1O001O001O1O001O001O001O00001O001O00001O001O001O00001O001O001O00001O010O001O001O001O1O001O1O001O1O001O1O001O1O1O1O2N1O2N2N2N2N2N2N2N2N9G4L4L4L4L4L5K4L3L4M2M4L4XNfEcN]:Z1hE`N]:\\1iE^N\\:]1jE\\N\\:^1kEYN\\:c1e1K5J6Km0RO7J5J6K5J`mf1"}, "label": "lady in black top and red shawl"}]} {"id": 221880, "image": "COCO_train2014_000000221880.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man on the left\"."}], "task": "refering", "answer_template": "The \"the man on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 123, 124, 125, 126, 145, 146, 147, 148, 149, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 233, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 310, 325, 326, 327, 328, 329, 330, 331, 332, 333, 348, 349, 350, 351, 352, 353, 354, 355, 356, 371, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.15337499999999998, 0.2518333333333333, 0.515734375, 0.9866666666666667], "iscrowd": 0, "area": 57591.171449999994, "rle": {"size": [480, 640], "counts": "lX^15c>b0_Oa0^Ob0^Ob0^Ob0E;En1SN0O100O1O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O010O100O1O100O1O100O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1fNhHkJY7T5nHeJS7Y5PIeJQ7Z5RIdJn6Z5UIdJl6[5WIbJj6\\5YIbJh6]5[I`Jf6\\5_IbJb6U5i1G9G9O1O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O10000000000000000000000000000000000000000000000000000000000001O001O001O001O001O001O1O001O001O001O3M2N3nH_K_3c4TLiKk3Z4gKRLX4Q4ZK[Le4g3TK^Ll4f3kJ`LT5f3bJ_LMNd2h3VM^L41a2f3QM_L;2_2d3mL_La05\\2a3iL`Lh06Z2k5bM]JY2d5dMdJV2]5gMjJT2V5jMQKQ2P5lMWKo1j4nM^Kl1e4oMmKa1V4YN_LU1W9_O9G4L5KZk`4"}, "label": "the man on the left"}]} {"id": 221880, "image": "COCO_train2014_000000221880.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man that is wearing glasses\"."}], "task": "refering", "answer_template": "The \"a man that is wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 123, 124, 125, 126, 145, 146, 147, 148, 149, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 233, 234, 235, 236, 237, 238, 239, 240, 241, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 310, 325, 326, 327, 328, 329, 330, 331, 332, 333, 348, 349, 350, 351, 352, 353, 354, 355, 356, 371, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.15337499999999998, 0.2518333333333333, 0.515734375, 0.9866666666666667], "iscrowd": 0, "area": 57591.171449999994, "rle": {"size": [480, 640], "counts": "lX^15c>b0_Oa0^Ob0^Ob0^Ob0E;En1SN0O100O1O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O010O100O1O100O1O100O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1fNhHkJY7T5nHeJS7Y5PIeJQ7Z5RIdJn6Z5UIdJl6[5WIbJj6\\5YIbJh6]5[I`Jf6\\5_IbJb6U5i1G9G9O1O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O10000000000000000000000000000000000000000000000000000000000001O001O001O001O001O001O1O001O001O001O3M2N3nH_K_3c4TLiKk3Z4gKRLX4Q4ZK[Le4g3TK^Ll4f3kJ`LT5f3bJ_LMNd2h3VM^L41a2f3QM_L;2_2d3mL_La05\\2a3iL`Lh06Z2k5bM]JY2d5dMdJV2]5gMjJT2V5jMQKQ2P5lMWKo1j4nM^Kl1e4oMmKa1V4YN_LU1W9_O9G4L5KZk`4"}, "label": "a man that is wearing glasses"}]} {"id": 295613, "image": "COCO_train2014_000000295613.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"adult arm reaching down towards a baby\"."}], "task": "refering", "answer_template": "The \"adult arm reaching down towards a baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 26, 47, 48, 49, 71, 72, 73, 95, 96, 97, 98, 119, 120, 121, 143, 144], "bbox": [0.0, 0.0022466960352422907, 0.2933125, 0.42920704845814983], "iscrowd": 0, "area": 10435.156299999999, "rle": {"size": [454, 640], "counts": "\\1>k000O100000000O100000000O10000000000O100000000O10000001O1O001O001O001O1O001O001O001O1O001O001O001O1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1OeMTE]1k:[N^Ed1b:UNgEj1X:oMPFQ2S;01O001O001N10001O001O001N10001O001O001O0O101O001O001O000O2O001O001O000O2O0000000000000O10000000O1000000000001O000000001O000000001O0000001O000000001O00000O101O000O10001O0O0100O10ORO`C6a in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 26, 47, 48, 49, 71, 72, 73, 95, 96, 97, 98, 119, 120, 121, 143, 144], "bbox": [0.0, 0.0022466960352422907, 0.2933125, 0.42920704845814983], "iscrowd": 0, "area": 10435.156299999999, "rle": {"size": [454, 640], "counts": "\\1>k000O100000000O100000000O10000000000O100000000O10000001O1O001O001O001O1O001O001O001O1O001O001O001O1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1OeMTE]1k:[N^Ed1b:UNgEj1X:oMPFQ2S;01O001O001N10001O001O001N10001O001O001O0O101O001O001O000O2O001O001O000O2O0000000000000O10000000O1000000000001O000000001O000000001O0000001O000000001O00000O101O000O10001O0O0100O10ORO`C6a in this image.", "objects": [{"patches": [172, 173, 191, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 240, 241, 242, 243, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 309, 310, 311, 312, 333], "bbox": [0.33903125, 0.44416299559471367, 0.584078125, 0.9064977973568282], "iscrowd": 0, "area": 16756.349449999998, "rle": {"size": [454, 640], "counts": "m]P32S>4L4N2O1NO1O1N2O001O1NELgB3Y=OfB0[=1dB0[=1cB0]=1aB0_=1_B0`=2]B0c=:0001O0O100000000O2O00000O1000001O0O0TEZOY8f0eG]OZ8c0dGAZ8>eGEZ8;dGHZ88eGKZ85dGM\\83cGN\\82cG0\\81cG0]8ObG3]8NbG3]8NbG4\\8LcG6\\8JdG7[8IeG8YNWO_9a0WH:XN\\O[9:]H;VN[O^9:\\HnE@T: in this image.", "objects": [{"patches": [172, 173, 191, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 240, 241, 242, 243, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 309, 310, 311, 312, 333], "bbox": [0.33903125, 0.44416299559471367, 0.584078125, 0.9064977973568282], "iscrowd": 0, "area": 16756.349449999998, "rle": {"size": [454, 640], "counts": "m]P32S>4L4N2O1NO1O1N2O001O1NELgB3Y=OfB0[=1dB0[=1cB0]=1aB0_=1_B0`=2]B0c=:0001O0O100000000O2O00000O1000001O0O0TEZOY8f0eG]OZ8c0dGAZ8>eGEZ8;dGHZ88eGKZ85dGM\\83cGN\\82cG0\\81cG0]8ObG3]8NbG3]8NbG4\\8LcG6\\8JdG7[8IeG8YNWO_9a0WH:XN\\O[9:]H;VN[O^9:\\HnE@T: in this image.", "objects": [{"patches": [19, 20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 110, 111, 112, 113, 114, 132, 133, 134, 136, 137, 159, 160], "bbox": [0.765296875, 0.0, 1.0, 0.3930176211453745], "iscrowd": 0, "area": 13538.935849999996, "rle": {"size": [454, 640], "counts": "n\\i62R>2N2N2M3N2N200O1O101N1O100O1O100O1O100O1O100O1O100O1O1O100O1O100O1O01O000QOmBl0P in this image.", "objects": [{"patches": [19, 20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 110, 111, 112, 113, 114, 132, 133, 134, 136, 137, 159, 160], "bbox": [0.765296875, 0.0, 1.0, 0.3930176211453745], "iscrowd": 0, "area": 13538.935849999996, "rle": {"size": [454, 640], "counts": "n\\i62R>2N2N2M3N2N200O1O101N1O100O1O100O1O100O1O100O1O100O1O1O100O1O100O1O01O000QOmBl0P in this image.", "objects": [{"patches": [29, 30, 31, 52, 53, 54, 55, 75, 76, 77, 78, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 191, 192, 193, 194, 195, 210, 211, 212, 213, 214, 215, 216, 217, 218, 233, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 287, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 330, 331, 332], "bbox": [0.1534375, 0.09843091334894614, 0.517203125, 0.9821545667447307], "iscrowd": 0, "area": 54031.55375000001, "rle": {"size": [427, 640], "counts": "YTY15h<`0L4L4L5K4C=_Oa0_Ob0^Oa0C=bG_Ll6g3mHfLf6a3TIkL`6Z3[IjLa6\\3YIhLb6_3XIeLd6`3WIdLe6b3UIbLf6U5L4L4L4L4M3M3M3M3M3N2M3M3M3mJSHc4P8WKTHh4X800O1O1O100O10000O10000O10000O1000[H[Kd6d4[I^Ke6b4ZI^Kg6b4WI`Ki6_4VIcKj6]4TIdKm6`4nHaKR7c4hH^KY7f4aH\\K_7h4[HYKf7V5000O100O100O1000nIZK]N8j3]4gMkLT2U3iMTMQ2k2lMZMS2f2jM]MV2c2gM`MY24]LXOX1g0[2N`L[OT1h0\\2KbL\\OS1i0[2IdL^OP1j0\\2EfLAn0j0]2CfLEl0h0^2POgLRN0W2j0h0_2mNZM<5h0`2lN[M=5g0`2kN[M?4g0a2iN\\Ma01f0d2hN[Mc01e0d2gN[Mf0Oa0i2hNXMh0N?l2hNWMi0M=T2QMfMh190SOnNh0X2X2UMaMd1`0LSOROb0h1DbMi26\\Ma1g0FSOXOc3W1\\OZNTM=`3Y1[OZNWM<^3Y1\\OZNYM;[3[1\\OYN\\M:X3]1\\OXN^M:V3^1[OXNbM8S3`1[OWNeM7P3b1[OWNgM5n2d1ZOWNjM4l2e1ZOVNmM3i2g1ZOUNPN2f2i1YOVNSN7\\2c1AUNUN7Z2c1AWNWN3Y2f1@VNZN1W2i1^OWN]NMV2l1]OVN_NMT2m1\\OWNaNJT2:jKl0a3oNcNKQ27nKP1\\3nNfNKP25PLR1Z3nNgNKP21QLW1V3mNkNJP2MRL]1R3lNmNIQ2JRLa1o2lNPOIP2ETLg1k2kNQOIa2\\1]NlNSOH_2]1^NjNTOH^2_1]NiNVOH\\2_1^NiNVOH\\2`1]NhNXOG[2b1\\NgNZOGY2b1]NgN[OFX2d1\\NfN]OFV2d1]NfN]OFV2e1\\NeN_OEU2g1[NdNAES2g1\\NdNBDR2i1[NcNGAm1m1\\NaNc4^1_K`Na4a1`K]N`4c1bK[N^4g1bKWN^4j1bKRNa4o1`KlMc4V2X31O1O2N1O1O1O2N1O1O2mF^MY7d2cH`M[7d2_H_M`7d2\\H^Mc7e2XH^Mg7e2UH]Mj7h2oG[MP8k2hGYMV8m2cGUM\\8Q3\\GRMc8l3O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2M4K4L4RL\\J0g5MhJC]5:nJYOV5d0PKUOT5i0QKQOR5l0TKmNP5Q1TKiNQ5S1UKgNn4V1XK[NT5c1QKlM]5Q2iJ_Me5^2f2M4M3L4M3L4L3N3L4M3L4L4K5K4M4K5K4L5K5L3L5KYcP4"}, "label": "a statue on a bench is being sat next to by a man in shorts"}]} {"id": 418500, "image": "COCO_train2014_000000418500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a statue of a man holding a dog\"."}], "task": "refering", "answer_template": "The \"a statue of a man holding a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 52, 53, 54, 55, 75, 76, 77, 78, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 191, 192, 193, 194, 195, 210, 211, 212, 213, 214, 215, 216, 217, 218, 233, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 287, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 330, 331, 332], "bbox": [0.1534375, 0.09843091334894614, 0.517203125, 0.9821545667447307], "iscrowd": 0, "area": 54031.55375000001, "rle": {"size": [427, 640], "counts": "YTY15h<`0L4L4L5K4C=_Oa0_Ob0^Oa0C=bG_Ll6g3mHfLf6a3TIkL`6Z3[IjLa6\\3YIhLb6_3XIeLd6`3WIdLe6b3UIbLf6U5L4L4L4L4M3M3M3M3M3N2M3M3M3mJSHc4P8WKTHh4X800O1O1O100O10000O10000O10000O1000[H[Kd6d4[I^Ke6b4ZI^Kg6b4WI`Ki6_4VIcKj6]4TIdKm6`4nHaKR7c4hH^KY7f4aH\\K_7h4[HYKf7V5000O100O100O1000nIZK]N8j3]4gMkLT2U3iMTMQ2k2lMZMS2f2jM]MV2c2gM`MY24]LXOX1g0[2N`L[OT1h0\\2KbL\\OS1i0[2IdL^OP1j0\\2EfLAn0j0]2CfLEl0h0^2POgLRN0W2j0h0_2mNZM<5h0`2lN[M=5g0`2kN[M?4g0a2iN\\Ma01f0d2hN[Mc01e0d2gN[Mf0Oa0i2hNXMh0N?l2hNWMi0M=T2QMfMh190SOnNh0X2X2UMaMd1`0LSOROb0h1DbMi26\\Ma1g0FSOXOc3W1\\OZNTM=`3Y1[OZNWM<^3Y1\\OZNYM;[3[1\\OYN\\M:X3]1\\OXN^M:V3^1[OXNbM8S3`1[OWNeM7P3b1[OWNgM5n2d1ZOWNjM4l2e1ZOVNmM3i2g1ZOUNPN2f2i1YOVNSN7\\2c1AUNUN7Z2c1AWNWN3Y2f1@VNZN1W2i1^OWN]NMV2l1]OVN_NMT2m1\\OWNaNJT2:jKl0a3oNcNKQ27nKP1\\3nNfNKP25PLR1Z3nNgNKP21QLW1V3mNkNJP2MRL]1R3lNmNIQ2JRLa1o2lNPOIP2ETLg1k2kNQOIa2\\1]NlNSOH_2]1^NjNTOH^2_1]NiNVOH\\2_1^NiNVOH\\2`1]NhNXOG[2b1\\NgNZOGY2b1]NgN[OFX2d1\\NfN]OFV2d1]NfN]OFV2e1\\NeN_OEU2g1[NdNAES2g1\\NdNBDR2i1[NcNGAm1m1\\NaNc4^1_K`Na4a1`K]N`4c1bK[N^4g1bKWN^4j1bKRNa4o1`KlMc4V2X31O1O2N1O1O1O2N1O1O2mF^MY7d2cH`M[7d2_H_M`7d2\\H^Mc7e2XH^Mg7e2UH]Mj7h2oG[MP8k2hGYMV8m2cGUM\\8Q3\\GRMc8l3O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2M4K4L4RL\\J0g5MhJC]5:nJYOV5d0PKUOT5i0QKQOR5l0TKmNP5Q1TKiNQ5S1UKgNn4V1XK[NT5c1QKlM]5Q2iJ_Me5^2f2M4M3L4M3L4L3N3L4M3L4L4K5K4M4K5K4L5K5L3L5KYcP4"}, "label": "a statue of a man holding a dog"}]} {"id": 418500, "image": "COCO_train2014_000000418500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with no shirt\"."}], "task": "refering", "answer_template": "The \"the man with no shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 35, 36, 37, 57, 58, 59, 60, 81, 82, 83, 84, 96, 97, 101, 102, 103, 104, 105, 106, 107, 108, 119, 125, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 293, 294, 310, 311, 312, 313, 314, 315, 316, 317, 333, 334, 335, 337, 338, 339, 340], "bbox": [0.184296875, 0.034894613583138175, 0.828828125, 1.0], "iscrowd": 0, "area": 58838.7651, "rle": {"size": [427, 640], "counts": "gZa1;n<7K2M3NOM30100O1O1O10N1N2O20O00M4O100O1O1001O1O000001O001C]C0R=LQ\\`11lc_N301O0000001O00001N1000001O00000O2O0O2O1N2O001N2O0O2N2O0O2N2O0O2N2O1N4L4M0O1O100O100`K[OYLf0`3A`L?Z3GeL:U3LkL4o22PMOj27VMIf2;YMGc2<]ME_2>aMDZ2?fMBV2a0jM@Q2d0oM^Om1d0SN]Oj1e0VN]Of1e0ZN\\Oc1f0]N\\O_1f0aN[O\\1g0dNZOY1h0gNYOV1i0jNXOS1j0mNWOP1k0POVOg0R1cMjLGU2d2R1dMmLER2e2R1dMPMEo1e2R1ZM^MkNZOg0W2S3Q1XMQNAR1Q3o0\\MTN@P1P3m0]MXNAo0m2j0`MZNBP1i2g0bM^NCn0g2e0dM`NDo0b2c0gMbNFn0_2a0iMeNFn0\\2>nMeNEQ1X2;RNfN]OgMO[3]2:VNe2d1]M\\Ng2_1ZM`Nj2[1XMeNl2V1UMiNQ3P1QMPOU3i0lLVOZ3c0hL\\O_3B3L3N2N2N001O001O00001O00001O01O2O1N3M2O1N3M2O1N3N1N00010O01O00001O010O001O01O10O0100O10O01O100O10O0100O1O010O01O001O00001O002N1O2N2N1O2N2N2N1O2N2N3M2N3M3M3M3M3M2O2M3Ma0_O6J6J5K6J5K6nLVIWOo6=[I@g67bIGa6OiINY6IPJ5S6AVJ=U6QOUJl0_6YNjIe1P9N2N3M2N2N3M2N2N2N3M2N2N2N2N2N3M2N2N2N2N2N4L3M4L3M4Lb^]1"}, "label": "the man with no shirt"}]} {"id": 418500, "image": "COCO_train2014_000000418500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"middle aged white tourist sitting on a bench next to a statue of a black man who is holding a black and white puppy\"."}], "task": "refering", "answer_template": "The \"middle aged white tourist sitting on a bench next to a statue of a black man who is holding a black and white puppy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 35, 36, 37, 57, 58, 59, 60, 81, 82, 83, 84, 96, 97, 101, 102, 103, 104, 105, 106, 107, 108, 119, 125, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 293, 294, 310, 311, 312, 313, 314, 315, 316, 317, 333, 334, 335, 337, 338, 339, 340], "bbox": [0.184296875, 0.034894613583138175, 0.828828125, 1.0], "iscrowd": 0, "area": 58838.7651, "rle": {"size": [427, 640], "counts": "gZa1;n<7K2M3NOM30100O1O1O10N1N2O20O00M4O100O1O1001O1O000001O001C]C0R=LQ\\`11lc_N301O0000001O00001N1000001O00000O2O0O2O1N2O001N2O0O2N2O0O2N2O0O2N2O1N4L4M0O1O100O100`K[OYLf0`3A`L?Z3GeL:U3LkL4o22PMOj27VMIf2;YMGc2<]ME_2>aMDZ2?fMBV2a0jM@Q2d0oM^Om1d0SN]Oj1e0VN]Of1e0ZN\\Oc1f0]N\\O_1f0aN[O\\1g0dNZOY1h0gNYOV1i0jNXOS1j0mNWOP1k0POVOg0R1cMjLGU2d2R1dMmLER2e2R1dMPMEo1e2R1ZM^MkNZOg0W2S3Q1XMQNAR1Q3o0\\MTN@P1P3m0]MXNAo0m2j0`MZNBP1i2g0bM^NCn0g2e0dM`NDo0b2c0gMbNFn0_2a0iMeNFn0\\2>nMeNEQ1X2;RNfN]OgMO[3]2:VNe2d1]M\\Ng2_1ZM`Nj2[1XMeNl2V1UMiNQ3P1QMPOU3i0lLVOZ3c0hL\\O_3B3L3N2N2N001O001O00001O00001O01O2O1N3M2O1N3M2O1N3N1N00010O01O00001O010O001O01O10O0100O10O01O100O10O0100O1O010O01O001O00001O002N1O2N2N1O2N2N2N1O2N2N3M2N3M3M3M3M3M2O2M3Ma0_O6J6J5K6J5K6nLVIWOo6=[I@g67bIGa6OiINY6IPJ5S6AVJ=U6QOUJl0_6YNjIe1P9N2N3M2N2N3M2N2N2N3M2N2N2N2N2N3M2N2N2N2N2N4L3M4L3M4Lb^]1"}, "label": "middle aged white tourist sitting on a bench next to a statue of a black man who is holding a black and white puppy"}]} {"id": 426705, "image": "COCO_train2014_000000426705.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman sitting in front of a laptop\"."}], "task": "refering", "answer_template": "The \"a woman sitting in front of a laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 97, 98, 99, 115, 116, 117, 118, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 191, 192, 204, 205, 206, 207, 208, 209, 210, 224, 225], "bbox": [0.34978, 0.23308, 0.7018, 0.67702], "iscrowd": 0, "area": 24430.89085, "rle": {"size": [500, 500], "counts": "oee25]?7H7I7I8H7J6I8H7I7I8I[1dN3M210O10O10O100O010kNULaFk3T9oL\\FR3X9jMXFV2e9n1M3N2M3M3M3N2M3M4M2000001O00000010O00000001O0000010O0000001O0000010O000000001O0000010OO10001O001O1O1O1O001O1O1O0O2O1O001O1O1O001O1O001O1N2O001O1O001O1O001O6kK[Fo1Q:]MXFb2S:jLWFU3U;N101O1O1O1O001O1N2O1O001O1O1N2O001O1O1O1N101O1O1O001N101O1O1O2N1N3N1O2N1O2N2M2O2N1O1O1O1N1O1N2O1O1N3N1O1N2O1N2M3L4M3M3M3N2M4L3M3N2M3M3K5L4KS`X2"}, "label": "a woman sitting in front of a laptop"}]} {"id": 426705, "image": "COCO_train2014_000000426705.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman with glasses\"."}], "task": "refering", "answer_template": "The \"the woman with glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 97, 98, 99, 115, 116, 117, 118, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 191, 192, 204, 205, 206, 207, 208, 209, 210, 224, 225], "bbox": [0.34978, 0.23308, 0.7018, 0.67702], "iscrowd": 0, "area": 24430.89085, "rle": {"size": [500, 500], "counts": "oee25]?7H7I7I8H7J6I8H7I7I8I[1dN3M210O10O10O100O010kNULaFk3T9oL\\FR3X9jMXFV2e9n1M3N2M3M3M3N2M3M4M2000001O00000010O00000001O0000010O0000001O0000010O000000001O0000010OO10001O001O1O1O1O001O1O1O0O2O1O001O1O1O001O1O001O1N2O001O1O001O1O001O6kK[Fo1Q:]MXFb2S:jLWFU3U;N101O1O1O1O001O1N2O1O001O1O1N2O001O1O1O1N101O1O1O001N101O1O1O2N1N3N1O2N1O2N2M2O2N1O1O1O1N1O1N2O1O1N3N1O1N2O1N2M3L4M3M3M3N2M4L3M3N2M3M3K5L4KS`X2"}, "label": "the woman with glasses"}]} {"id": 426705, "image": "COCO_train2014_000000426705.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt sitting to the left of a woman wearing glasses\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt sitting to the left of a woman wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 75, 76, 77, 78, 79, 93, 94, 95, 96, 97, 111, 112, 113, 114, 115, 128, 129, 130, 131, 132, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 181, 182, 183, 184, 185, 186, 199, 200, 201, 202, 203, 204, 216, 217, 218, 219, 221, 222, 234, 235, 252, 253, 254, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 311, 312, 313], "bbox": [0.047299999999999995, 0.19256, 0.44594, 1.0], "iscrowd": 0, "area": 41053.31079999999, "rle": {"size": [500, 500], "counts": "\\Uk0mLT3]Ob0_Ob0^OlLjFj0f8]OaGe0Z8WOnGh0P8SOWHm0f7UO\\Hj0a7YO`Hf0]7\\OdHd0Y7_OhH`0U7ClHP5EoJ:P5HPK7o4LPK3Q5NnJ2R5NnJ1S50lJOU51kJNV53iJLX55gJJZ56fJJZ57eJH\\59cJF^5:bJE_5<`JD`5=_JCa5=_JBb5?]JAc5`0\\J@d5`0\\J@d5a0[J_Oe5a0[J_Oe5b0ZJ^Of5c0YJ]Og5c0YJ]Og5d0XJ\\Oh5e0WJ[Oi5e0WJ[Oi5f0VJZOj5g0UJYOk5g0UJYOk5h0TJXOl5i0SJVOn5j0RJVOn5k0QJUOo5l0PJTOP6l0PJTOQ6l0nITOT6j0lIVOU6j0jIVOX6i0gIWOZ6h0fIXO\\6g0cIYO_6f0`IZOa6e0_I[Oc6d0\\I\\Od6e0[I[Oe6e0[I[Oe6f0ZIYOg6h0XIXOh6h0XIXOg6j0XIVOh6k0WIUOi6k0WIVOh6k0WIUOi6l0VITOj6l0VIUOQ3RMEj3YMTOR3VM@g3^MSOQ3[M]Oc3aMSOQ3^MZO`3dMROQ3cMVO\\3iMROP3gMnN\\3QNmNP3lMfN]3YNgNQ3PN^N]3aNdNo2TNWN^3iN^NP3XNnM`3QOXNP3^NfM^3ZOUNo2aN^M`3BoMo2fNVMa3JjMo2iNoLa32fMn2nNgLb3:`Mo2SO^Lb3b0\\Mk2S5UMmJf2Y5YMgJb2_5]MbJ^2b5bM^JY2h5fMYJV2j5jMVJV2k5iMUJW2l5hMUJW2k5iMUJW2l5hMTJW2n5hMTJ\\1eLPOW9DVJY1fLQOU9EWJV1hLSOR9FXJT1iLTOo8HZJP1kLVOk8J^Jk0jLYOi8KoJ8\\LJf8NXLOj30XLMi33YLJh36l5O001O00001O001O00001O001O001O0000UXW4"}, "label": "a man in a black shirt sitting to the left of a woman wearing glasses"}]} {"id": 426705, "image": "COCO_train2014_000000426705.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man sitting at the table\"."}], "task": "refering", "answer_template": "The \"the man sitting at the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 75, 76, 77, 78, 79, 93, 94, 95, 96, 97, 111, 112, 113, 114, 115, 128, 129, 130, 131, 132, 145, 146, 147, 148, 149, 150, 163, 164, 165, 166, 167, 168, 181, 182, 183, 184, 185, 186, 199, 200, 201, 202, 203, 204, 216, 217, 218, 219, 221, 222, 234, 235, 252, 253, 254, 270, 271, 272, 273, 274, 288, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 311, 312, 313], "bbox": [0.047299999999999995, 0.19256, 0.44594, 1.0], "iscrowd": 0, "area": 41053.31079999999, "rle": {"size": [500, 500], "counts": "\\Uk0mLT3]Ob0_Ob0^OlLjFj0f8]OaGe0Z8WOnGh0P8SOWHm0f7UO\\Hj0a7YO`Hf0]7\\OdHd0Y7_OhH`0U7ClHP5EoJ:P5HPK7o4LPK3Q5NnJ2R5NnJ1S50lJOU51kJNV53iJLX55gJJZ56fJJZ57eJH\\59cJF^5:bJE_5<`JD`5=_JCa5=_JBb5?]JAc5`0\\J@d5`0\\J@d5a0[J_Oe5a0[J_Oe5b0ZJ^Of5c0YJ]Og5c0YJ]Og5d0XJ\\Oh5e0WJ[Oi5e0WJ[Oi5f0VJZOj5g0UJYOk5g0UJYOk5h0TJXOl5i0SJVOn5j0RJVOn5k0QJUOo5l0PJTOP6l0PJTOQ6l0nITOT6j0lIVOU6j0jIVOX6i0gIWOZ6h0fIXO\\6g0cIYO_6f0`IZOa6e0_I[Oc6d0\\I\\Od6e0[I[Oe6e0[I[Oe6f0ZIYOg6h0XIXOh6h0XIXOg6j0XIVOh6k0WIUOi6k0WIVOh6k0WIUOi6l0VITOj6l0VIUOQ3RMEj3YMTOR3VM@g3^MSOQ3[M]Oc3aMSOQ3^MZO`3dMROQ3cMVO\\3iMROP3gMnN\\3QNmNP3lMfN]3YNgNQ3PN^N]3aNdNo2TNWN^3iN^NP3XNnM`3QOXNP3^NfM^3ZOUNo2aN^M`3BoMo2fNVMa3JjMo2iNoLa32fMn2nNgLb3:`Mo2SO^Lb3b0\\Mk2S5UMmJf2Y5YMgJb2_5]MbJ^2b5bM^JY2h5fMYJV2j5jMVJV2k5iMUJW2l5hMUJW2k5iMUJW2l5hMTJW2n5hMTJ\\1eLPOW9DVJY1fLQOU9EWJV1hLSOR9FXJT1iLTOo8HZJP1kLVOk8J^Jk0jLYOi8KoJ8\\LJf8NXLOj30XLMi33YLJh36l5O001O00001O001O00001O001O001O0000UXW4"}, "label": "the man sitting at the table"}]} {"id": 451283, "image": "COCO_train2014_000000451283.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red bag that is hanging on the arm of a woman\"."}], "task": "refering", "answer_template": "The \"a red bag that is hanging on the arm of a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [255, 276, 278, 279, 299, 300, 301, 302, 303, 323, 324, 325, 326, 346, 347, 348, 349, 369, 370, 371, 372], "bbox": [0.017203125, 0.6799583333333333, 0.223390625, 0.9837708333333334], "iscrowd": 0, "area": 9255.87935, "rle": {"size": [480, 640], "counts": "S`52m>5J5L4K5L4K5L4L5K4M3M3L4O12N3M:G9F5KN2O1L4K5L4K6K4K5L4K5K5L4K6K4K5K5M2OO1O1O1O1O1Oh0XO9G101N2N1O2N00O1O100O1O2O0O1[MZNoGg1R8[NjGf1X8[NeGe1\\8]N`Gd1a8^N[Gc1f8_NVGb1k8`NPGb1Q9aNjFa1V9aNfF`1[9bN`F`1a9bN[F_1f9cNUF_1m9i11O1O1O1O100O1O1O2N1O1O1O1O1O100O1O1O2N1O1O1O1O1O100O1O2N1O1O1O1N2M3N2N2M3N3L3N2M3N2N2M3N2M3N3M2M3N2M3N2M3N2N2M6K5J6K4L4K4M4JkaX7"}, "label": "a red bag that is hanging on the arm of a woman"}]} {"id": 500440, "image": "COCO_train2014_000000500440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a white shirt behind a table\"."}], "task": "refering", "answer_template": "The \"a man with a white shirt behind a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 27, 28, 29, 30, 31, 50, 51, 52, 53, 54, 73, 75, 76, 77, 99, 100, 101, 123, 124, 147], "bbox": [0.170203125, 0.0, 0.43032812499999995, 0.4151875], "iscrowd": 0, "area": 12706.60925, "rle": {"size": [480, 640], "counts": "]Vc13g>5J6K5J7M2O1N2O2N1O1N2O2N1N2O1O1O2M2N2N2O2M2N2N2K5J7H7J6N2100O1O10O01O100O1O1O100O1O100O1O100O1O100O1001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1N2O2N1O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O2N1O1O2N\\L[En2d:nLcEP3X;O2O1N8H8I7H8H8H4M0O2N1O100O2N100O1OO101N10001N100O101O0011O001O001O0O2N1O2M3N1O2M2O2N1N3N1O2XO[BOg=M^B1d=JaB3b=HcB6V>M3M3NjlZ5"}, "label": "a man with a white shirt behind a table"}]} {"id": 500440, "image": "COCO_train2014_000000500440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white shirt\"."}], "task": "refering", "answer_template": "The \"a man in a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 27, 28, 29, 30, 31, 50, 51, 52, 53, 54, 73, 75, 76, 77, 99, 100, 101, 123, 124, 147], "bbox": [0.170203125, 0.0, 0.43032812499999995, 0.4151875], "iscrowd": 0, "area": 12706.60925, "rle": {"size": [480, 640], "counts": "]Vc13g>5J6K5J7M2O1N2O2N1O1N2O2N1N2O1O1O2M2N2N2O2M2N2N2K5J7H7J6N2100O1O10O01O100O1O1O100O1O100O1O100O1O100O1001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1N2O2N1O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O2N1O1O2N\\L[En2d:nLcEP3X;O2O1N8H8I7H8H8H4M0O2N1O100O2N100O1OO101N10001N100O101O0011O001O001O0O2N1O2M3N1O2M2O2N1N3N1O2XO[BOg=M^B1d=JaB3b=HcB6V>M3M3NjlZ5"}, "label": "a man in a white shirt"}]} {"id": 500440, "image": "COCO_train2014_000000500440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an open laptop sits to the right of a glass of beer\"."}], "task": "refering", "answer_template": "The \"an open laptop sits to the right of a glass of beer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.347796875, 0.005479166666666666, 1.0, 0.9821666666666666], "iscrowd": 0, "area": 163810.141, "rle": {"size": [480, 640], "counts": "daX3`0`>a0_Oa0^Oa0@1O1O1O2N1N2O1O1O1O1O1N3N1O1O1O1O1N2O2N1O3M2N3L4M2N3M3M2N3M3L3N3M3M2N3M3L3N3M3M2N3M3L3]FWKk8l4oFZKn8i4lF\\KR9\\5M2N2M4M2N2N3M2N2N3L3N2N3M2N3M2M3N3M2N2N3M2M3N3M2N3M2N2M4M2N2N3M2N2M4M2N3M2N2]OWGkJl8o4WGQKk8j4XGUKk8f4XGZKh8c4[G]Ke8`4^G`Ka8^4bGbK^8[4eGeK[8W4iGhKX8T4lGlKS8P4RHPLn7l3VHTLj7h3ZHXLf7d3^H\\La7a3cH^L^7^3fHbLZ7Z3jHfLV7V3nHjLQ7S3SImLm6o2WIQMi6k2[IUMe6g2_IYM`6i0UGa0_2fN\\6a0^Gd0Z2kNX69gGg0U2POT64mGg0S2UOo51SHe0R2ZOk5LYHd0Q2@f5H^Hc0P2ET76PIJQ71SIOn6LVI4k6GYI9g:00000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000003M2N3M3M2N3M2N3M2N3M2N3M2N3M2NaN"}, "label": "an open laptop sits to the right of a glass of beer"}]} {"id": 500440, "image": "COCO_train2014_000000500440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white laptop with green keyboard open and turned on with email page showing\"."}], "task": "refering", "answer_template": "The \"white laptop with green keyboard open and turned on with email page showing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.347796875, 0.005479166666666666, 1.0, 0.9821666666666666], "iscrowd": 0, "area": 163810.141, "rle": {"size": [480, 640], "counts": "daX3`0`>a0_Oa0^Oa0@1O1O1O2N1N2O1O1O1O1O1N3N1O1O1O1O1N2O2N1O3M2N3L4M2N3M3M2N3M3L3N3M3M2N3M3L3N3M3M2N3M3L3]FWKk8l4oFZKn8i4lF\\KR9\\5M2N2M4M2N2N3M2N2N3L3N2N3M2N3M2M3N3M2N2N3M2M3N3M2N3M2N2M4M2N2N3M2N2M4M2N3M2N2]OWGkJl8o4WGQKk8j4XGUKk8f4XGZKh8c4[G]Ke8`4^G`Ka8^4bGbK^8[4eGeK[8W4iGhKX8T4lGlKS8P4RHPLn7l3VHTLj7h3ZHXLf7d3^H\\La7a3cH^L^7^3fHbLZ7Z3jHfLV7V3nHjLQ7S3SImLm6o2WIQMi6k2[IUMe6g2_IYM`6i0UGa0_2fN\\6a0^Gd0Z2kNX69gGg0U2POT64mGg0S2UOo51SHe0R2ZOk5LYHd0Q2@f5H^Hc0P2ET76PIJQ71SIOn6LVI4k6GYI9g:00000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000003M2N3M3M2N3M2N3M2N3M2N3M2N3M2NaN"}, "label": "white laptop with green keyboard open and turned on with email page showing"}]} {"id": 549599, "image": "COCO_train2014_000000549599.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman carrying a gray bag\"."}], "task": "refering", "answer_template": "The \"a woman carrying a gray bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 36, 37, 38, 59, 60, 61, 82, 83, 84, 85, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 265, 266, 267, 268, 288, 291, 292, 311, 314, 315], "bbox": [0.52240625, 0.04740046838407494, 0.746296875, 0.9355503512880563], "iscrowd": 0, "area": 30080.921649999982, "rle": {"size": [427, 640], "counts": "nc[42X1OX:g0UEF\\:o0XEWOY:U2@`0A?gMY2L4M3L3N3L4M3N2O1O1O1O10OO0O00010O0O1M[Od0K6M31O1011N1O2O010O01O0102M2O2M3NiN\\JlJb5S5dJjJ]5R5iJkJX5P5oJmJS5n4RKPKo4m4VKPKl4n4WKoJj4P5YKmJh4S5ZKjJh4T5[KiJf4W5\\KfJf4X5]KeJd4[5^KbJc4]5_KaJc4^5_K`Ja4_5j1K6K4K5L5J5L4K4M4K4M3N3L3M4L2N2N100O2N1O2N1O101N1O2N1O1O2N101N1O2N1O2N\\MYMeJg2X5^MgJ`2V5fMhJY2V5mMhJR2U5SNiJm1S5YNlJf1Q5_NmJb1o4dNoJ[1m4kNQKU1l4POSKP1i4UOUKk0ZO\\MQ4P2cLc0ZO_MP4S2dL?YO`MP4V2fL9XOcMn3Z2hL3XOeMm3]2jLNVOgMm3`2kLIVOiMk3d2nLFQOhMk3j2RMLcN\\MV4o2VM in this image.", "objects": [{"patches": [14, 15, 36, 37, 38, 59, 60, 61, 82, 83, 84, 85, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 265, 266, 267, 268, 288, 291, 292, 311, 314, 315], "bbox": [0.52240625, 0.04740046838407494, 0.746296875, 0.9355503512880563], "iscrowd": 0, "area": 30080.921649999982, "rle": {"size": [427, 640], "counts": "nc[42X1OX:g0UEF\\:o0XEWOY:U2@`0A?gMY2L4M3L3N3L4M3N2O1O1O1O10OO0O00010O0O1M[Od0K6M31O1011N1O2O010O01O0102M2O2M3NiN\\JlJb5S5dJjJ]5R5iJkJX5P5oJmJS5n4RKPKo4m4VKPKl4n4WKoJj4P5YKmJh4S5ZKjJh4T5[KiJf4W5\\KfJf4X5]KeJd4[5^KbJc4]5_KaJc4^5_K`Ja4_5j1K6K4K5L5J5L4K4M4K4M3N3L3M4L2N2N100O2N1O2N1O101N1O2N1O1O2N101N1O2N1O2N\\MYMeJg2X5^MgJ`2V5fMhJY2V5mMhJR2U5SNiJm1S5YNlJf1Q5_NmJb1o4dNoJ[1m4kNQKU1l4POSKP1i4UOUKk0ZO\\MQ4P2cLc0ZO_MP4S2dL?YO`MP4V2fL9XOcMn3Z2hL3XOeMm3]2jLNVOgMm3`2kLIVOiMk3d2nLFQOhMk3j2RMLcN\\MV4o2VM in this image.", "objects": [{"patches": [8, 30, 31, 32, 53, 54, 55, 75, 76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 238, 239, 240, 241, 261, 262, 263, 264, 284, 285, 286, 287, 307, 308, 309, 310, 330, 331, 332, 333], "bbox": [0.281875, 0.05843091334894614, 0.5082656249999999, 0.9887587822014051], "iscrowd": 0, "area": 34165.073, "rle": {"size": [427, 640], "counts": "RV[2n0a;o0F:L3N3M3L3N3M3M3L4M4L3L5L4L4K4M3M3L4M3M3L4mNeKPI^4h6QLnHS4i6\\LlHh3l6]LTIg3e6\\LZIi3`6YL_Ik3\\6XLcIl3W6VLhIn3S6TLlIQ4n5QLQJR4k5QLSJR4j5oKVJR4h5oKWJS4h5i1O2N1O1N2O1O2N1N]1dN5K4L5KS2lM7J6J4L1O0000000000000000000001O00001O001O001O00001O001O000YEmNV9U1gFmNT8:VGl0c0kNS8?WGi0b0jNU6O]Je0iNe0a0iNT67YJb0nNa0a0gNT6`0TJ>TO=a0fNT6h0mI<[O8a0fNR6W3XIWN`0fNT6U5cIRKZ6S5^ITKb6j54L3N3L4L3N3L3M4M2M4M2M4MU1jN01O001eHeKd5\\4\\JhK_5Z4`JjK[5X4dJmKV5U4iJoKR5S4mJoKP5S4nJnKQ5T4nJlKQ5V4nJjKU5T4jJmKX5Q4gJoK\\5P4bJPLa5[OnIk3>kL]6X3`IiL^6Z3`IfL_6]3_IfL]6^3`IXMi5k2UJVMi5m2SJUMl5n2oIVMo5m2kIWMT6l2fIYMX6n4O2`KkIf1Y6SNUJb1n5XN]J`1f5ZNaJb1c5WNdJf1^5TNjJg1Y5SNnJ`0kMFZ7DRKb0gMG[7@UKe0bMI\\7\\OYKh0]MIh:3ZEKj:1YELi:1ZEMh:0\\EMe:0^1M3MeoR4"}, "label": "woman in blue outfit with glasses , holding her phone"}]} {"id": 549599, "image": "COCO_train2014_000000549599.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with glasses and blue tennis shoes sits on a bench with other women\"."}], "task": "refering", "answer_template": "The \"a woman with glasses and blue tennis shoes sits on a bench with other women\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 30, 31, 32, 53, 54, 55, 75, 76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 214, 215, 216, 217, 218, 238, 239, 240, 241, 261, 262, 263, 264, 284, 285, 286, 287, 307, 308, 309, 310, 330, 331, 332, 333], "bbox": [0.281875, 0.05843091334894614, 0.5082656249999999, 0.9887587822014051], "iscrowd": 0, "area": 34165.073, "rle": {"size": [427, 640], "counts": "RV[2n0a;o0F:L3N3M3L3N3M3M3L4M4L3L5L4L4K4M3M3L4M3M3L4mNeKPI^4h6QLnHS4i6\\LlHh3l6]LTIg3e6\\LZIi3`6YL_Ik3\\6XLcIl3W6VLhIn3S6TLlIQ4n5QLQJR4k5QLSJR4j5oKVJR4h5oKWJS4h5i1O2N1O1N2O1O2N1N]1dN5K4L5KS2lM7J6J4L1O0000000000000000000001O00001O001O001O00001O001O000YEmNV9U1gFmNT8:VGl0c0kNS8?WGi0b0jNU6O]Je0iNe0a0iNT67YJb0nNa0a0gNT6`0TJ>TO=a0fNT6h0mI<[O8a0fNR6W3XIWN`0fNT6U5cIRKZ6S5^ITKb6j54L3N3L4L3N3L3M4M2M4M2M4MU1jN01O001eHeKd5\\4\\JhK_5Z4`JjK[5X4dJmKV5U4iJoKR5S4mJoKP5S4nJnKQ5T4nJlKQ5V4nJjKU5T4jJmKX5Q4gJoK\\5P4bJPLa5[OnIk3>kL]6X3`IiL^6Z3`IfL_6]3_IfL]6^3`IXMi5k2UJVMi5m2SJUMl5n2oIVMo5m2kIWMT6l2fIYMX6n4O2`KkIf1Y6SNUJb1n5XN]J`1f5ZNaJb1c5WNdJf1^5TNjJg1Y5SNnJ`0kMFZ7DRKb0gMG[7@UKe0bMI\\7\\OYKh0]MIh:3ZEKj:1YELi:1ZEMh:0\\EMe:0^1M3MeoR4"}, "label": "a woman with glasses and blue tennis shoes sits on a bench with other women"}]} {"id": 549599, "image": "COCO_train2014_000000549599.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady on a bench holding an orange bag\"."}], "task": "refering", "answer_template": "The \"a lady on a bench holding an orange bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 48, 49, 50, 51, 71, 72, 73, 74, 93, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236, 237, 256, 257, 258, 259, 260], "bbox": [0.052468749999999995, 0.06981264637002342, 0.31934375, 0.7844262295081967], "iscrowd": 0, "area": 35183.07169999999, "rle": {"size": [427, 640], "counts": "hY>d0f in this image.", "objects": [{"patches": [26, 27, 48, 49, 50, 51, 71, 72, 73, 74, 93, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236, 237, 256, 257, 258, 259, 260], "bbox": [0.052468749999999995, 0.06981264637002342, 0.31934375, 0.7844262295081967], "iscrowd": 0, "area": 35183.07169999999, "rle": {"size": [427, 640], "counts": "hY>d0f in this image.", "objects": [{"patches": [11, 33, 34, 35, 36, 37, 38, 56, 57, 58, 59, 60, 61, 79, 80, 81, 82, 83, 102, 103, 104, 105, 106, 128], "bbox": [0.445375, 0.054543325526932085, 0.66371875, 0.34285714285714286], "iscrowd": 0, "area": 13657.727050000001, "rle": {"size": [427, 640], "counts": "elf3e0`<8M4K5L4K5L4L4L4L5K4SETNT:d2M2O2M3M2N3M2N2N1O2N2O0O2O1O1O0O2O1O000O10O1O1O00100O2N1O1O1O100O2N1O1000010O000001O01O0001O0000010O00001O010O00001O00010O001O00010O001O0000001O000000000O101O000000000000O100O1O1O1O010O1O1O1O100O1O1O1O1O2M2O1N3M2N2O1N3M2N2O1N3M2N2O2M3M3K5K5K5K4L5K5K5K5L:E;EXci2"}, "label": "a bush of plant behind middle woman"}]} {"id": 533220, "image": "COCO_train2014_000000533220.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green plant behind a table visible behind a lady ' s head\"."}], "task": "refering", "answer_template": "The \"green plant behind a table visible behind a lady ' s head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 33, 34, 35, 36, 37, 38, 56, 57, 58, 59, 60, 61, 79, 80, 81, 82, 83, 102, 103, 104, 105, 106, 128], "bbox": [0.445375, 0.054543325526932085, 0.66371875, 0.34285714285714286], "iscrowd": 0, "area": 13657.727050000001, "rle": {"size": [427, 640], "counts": "elf3e0`<8M4K5L4K5L4L4L4L5K4SETNT:d2M2O2M3M2N3M2N2N1O2N2O0O2O1O1O0O2O1O000O10O1O1O00100O2N1O1O1O100O2N1O1000010O000001O01O0001O0000010O00001O010O00001O00010O001O00010O001O0000001O000000000O101O000000000000O100O1O1O1O010O1O1O1O100O1O1O1O1O2M2O1N3M2N2O1N3M2N2O1N3M2N2O2M3M3K5K5K5K4L5K5K5K5L:E;EXci2"}, "label": "green plant behind a table visible behind a lady ' s head"}]} {"id": 533220, "image": "COCO_train2014_000000533220.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in the middle\"."}], "task": "refering", "answer_template": "The \"woman in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 126, 127, 128, 149, 150, 151, 171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 309], "bbox": [0.429375, 0.30517564402810304, 0.6562812499999999, 0.906791569086651], "iscrowd": 0, "area": 22098.742650000004, "rle": {"size": [427, 640], "counts": "lob3\\1X;S1mNf0ZOg0E:F:G:E:O20O0010O010O01O010O010O0010O010O01O01O010O0A?_Ob0]Ob0N201O001O000O2O001O00001N101O0000PIiIn6Z6O10O01O100O100O010O100000O10000000000O10O100000O11O1O1O1O1O1O1N2O2N1O1O1O1O1O1O1WKRIS3o6bL[I]3g6XLcIh3^6iKPJV4c7O100O1O2O0O1aMfGL[84_HROa7n0XIXNi6h1]201O000000000000001O000001O000001O1O2N1O2N1O2NmEaN[8^1[GjNh8U1nFSOV9k0`F]Oc9b0SFGP:7fE1]:\\14M2M4L3M9H9F9G:G8G:F:F9Hf^k2"}, "label": "woman in the middle"}]} {"id": 533220, "image": "COCO_train2014_000000533220.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a black shirt and jeans drinking a beer\"."}], "task": "refering", "answer_template": "The \"a woman with a black shirt and jeans drinking a beer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 126, 127, 128, 149, 150, 151, 171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 309], "bbox": [0.429375, 0.30517564402810304, 0.6562812499999999, 0.906791569086651], "iscrowd": 0, "area": 22098.742650000004, "rle": {"size": [427, 640], "counts": "lob3\\1X;S1mNf0ZOg0E:F:G:E:O20O0010O010O01O010O010O0010O010O01O01O010O0A?_Ob0]Ob0N201O001O000O2O001O00001N101O0000PIiIn6Z6O10O01O100O100O010O100000O10000000000O10O100000O11O1O1O1O1O1O1N2O2N1O1O1O1O1O1O1WKRIS3o6bL[I]3g6XLcIh3^6iKPJV4c7O100O1O2O0O1aMfGL[84_HROa7n0XIXNi6h1]201O000000000000001O000001O000001O1O2N1O2N1O2NmEaN[8^1[GjNh8U1nFSOV9k0`F]Oc9b0SFGP:7fE1]:\\14M2M4L3M9H9F9G:G8G:F:F9Hf^k2"}, "label": "a woman with a black shirt and jeans drinking a beer"}]} {"id": 533220, "image": "COCO_train2014_000000533220.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green plant on the left back side of a patio with tables\"."}], "task": "refering", "answer_template": "The \"a green plant on the left back side of a patio with tables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 117, 118, 119, 120], "bbox": [0.0129375, 0.07353629976580796, 0.24126562499999998, 0.3718735362997658], "iscrowd": 0, "area": 12943.389249999997, "rle": {"size": [427, 640], "counts": "`^33n<;H7N2N3M2N3M2N2N3M2N2N3N1N0VDdNg;a1N0[D^Nc;d1^D\\N_;g1`DYN^;i102N3M2N2N3M2N2N3M2N2O2N100O2O0O2O0O101N100O2N100O2O0O2O2M3N2M3N2M3N3L3M10O0100O010O1000O10000O2O000O101O000O101O0O101O0000001O0000001O00001O0000001O0000001O01O01O0001O01O01O0001O01O01O01O00010O000010O01O010O0003N3L4M3L4L4L3M4K4K5K6I6K5J7J5J6K6I6K5J7J5J\\aZ6"}, "label": "a green plant on the left back side of a patio with tables"}]} {"id": 533220, "image": "COCO_train2014_000000533220.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green bush to the left of two other bushes\"."}], "task": "refering", "answer_template": "The \"a green bush to the left of two other bushes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 117, 118, 119, 120], "bbox": [0.0129375, 0.07353629976580796, 0.24126562499999998, 0.3718735362997658], "iscrowd": 0, "area": 12943.389249999997, "rle": {"size": [427, 640], "counts": "`^33n<;H7N2N3M2N3M2N2N3M2N2N3N1N0VDdNg;a1N0[D^Nc;d1^D\\N_;g1`DYN^;i102N3M2N2N3M2N2N3M2N2O2N100O2O0O2O0O101N100O2N100O2O0O2O2M3N2M3N2M3N3L3M10O0100O010O1000O10000O2O000O101O000O101O0O101O0000001O0000001O00001O0000001O0000001O01O01O0001O01O01O0001O01O01O01O00010O000010O01O010O0003N3L4M3L4L4L3M4K4K5K6I6K5J7J5J6K6I6K5J7J5J\\aZ6"}, "label": "a green bush to the left of two other bushes"}]} {"id": 533220, "image": "COCO_train2014_000000533220.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man laughs at a table\"."}], "task": "refering", "answer_template": "The \"a man laughs at a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 120, 121, 122, 144, 145, 146, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259, 260, 261, 282, 283], "bbox": [0.144578125, 0.28611241217798594, 0.42771875000000004, 0.8617330210772833], "iscrowd": 0, "area": 22279.28875, "rle": {"size": [427, 640], "counts": "^RW1 in this image.", "objects": [{"patches": [97, 98, 99, 120, 121, 122, 144, 145, 146, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 256, 257, 258, 259, 260, 261, 282, 283], "bbox": [0.144578125, 0.28611241217798594, 0.42771875000000004, 0.8617330210772833], "iscrowd": 0, "area": 22279.28875, "rle": {"size": [427, 640], "counts": "^RW1 in this image.", "objects": [{"patches": [110, 111, 112, 133, 134, 135, 156, 157, 158, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 318, 319, 320, 321, 343], "bbox": [0.72109375, 0.30817330210772836, 0.9987031249999999, 0.9897423887587822], "iscrowd": 0, "area": 29073.495099999993, "rle": {"size": [427, 640], "counts": "flP63V=3M2M3N3M2N3L3N2N3M2N3N1N2O2XFoNS7Z1bHlNZ7`1WHgNe7f1kGaNQ8i1cG^NX8j1]G[Na8m1TGWNk8U3O1O2N100O100O2O000O100O2O000O100O2O0O10000O2O0O1O101N100O1jNVKkIj4Q6^KjIc4R6eKjI[4R6kKkIV4Q6oKmIR4o5SLoIm3n5WLQJj3k5[LSJf3l5\\LRJe3n5\\LPJe3P6\\LnIe3R6[LmIf3S6[LkIe3V6h100O1000000000O10O10001O1O1O001O1N2O1O1O001O1O1O1O1O1O001O1O1O1O10O01O1O1O001O001O1O0010O01O001O001O001CX1gN010O01\\ITJg5l5UJWJm5i5nIZJS6g5hI]JX6c5cIaJ_6_5\\IdJe6P6101N1O1O2O0O1O2PMRI@o6`0dImN\\6S1jIfNX6Z1hIdNY6\\1gIcNZ6^1eIbN\\6]1eIaN]6_1bI`Na6^1`I`Nb6`1]I`Ne6^1[IaNg6^1ZIaNh6^1WIaNk6^1UIbNm6]1SIaNo6^1QIbNQ7\\1PIbNR7]1nHbNU7\\1kHdNV7[1kHcNX7[1jHcNW7\\1kHaNY7\\1iHbN^7W1cHgNe7R1[HnNk7l0VHROR8g0oGWOX88THEU8]OcHa0U:M2O2N2M3N2M3N1N3N2M3N2N2Mi1"}, "label": "a man with a purple shirt"}]} {"id": 533220, "image": "COCO_train2014_000000533220.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smiley man with a purple shirt and khaki pants sitting down with his hands together\"."}], "task": "refering", "answer_template": "The \"a smiley man with a purple shirt and khaki pants sitting down with his hands together\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 133, 134, 135, 156, 157, 158, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 318, 319, 320, 321, 343], "bbox": [0.72109375, 0.30817330210772836, 0.9987031249999999, 0.9897423887587822], "iscrowd": 0, "area": 29073.495099999993, "rle": {"size": [427, 640], "counts": "flP63V=3M2M3N3M2N3L3N2N3M2N3N1N2O2XFoNS7Z1bHlNZ7`1WHgNe7f1kGaNQ8i1cG^NX8j1]G[Na8m1TGWNk8U3O1O2N100O100O2O000O100O2O000O100O2O0O10000O2O0O1O101N100O1jNVKkIj4Q6^KjIc4R6eKjI[4R6kKkIV4Q6oKmIR4o5SLoIm3n5WLQJj3k5[LSJf3l5\\LRJe3n5\\LPJe3P6\\LnIe3R6[LmIf3S6[LkIe3V6h100O1000000000O10O10001O1O1O001O1N2O1O1O001O1O1O1O1O1O001O1O1O1O10O01O1O1O001O001O1O0010O01O001O001O001CX1gN010O01\\ITJg5l5UJWJm5i5nIZJS6g5hI]JX6c5cIaJ_6_5\\IdJe6P6101N1O1O2O0O1O2PMRI@o6`0dImN\\6S1jIfNX6Z1hIdNY6\\1gIcNZ6^1eIbN\\6]1eIaN]6_1bI`Na6^1`I`Nb6`1]I`Ne6^1[IaNg6^1ZIaNh6^1WIaNk6^1UIbNm6]1SIaNo6^1QIbNQ7\\1PIbNR7]1nHbNU7\\1kHdNV7[1kHcNX7[1jHcNW7\\1kHaNY7\\1iHbN^7W1cHgNe7R1[HnNk7l0VHROR8g0oGWOX88THEU8]OcHa0U:M2O2N2M3N2M3N1N3N2M3N2N2Mi1"}, "label": "a smiley man with a purple shirt and khaki pants sitting down with his hands together"}]} {"id": 533220, "image": "COCO_train2014_000000533220.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with long dark hair in denim shirt , far left of picture\"."}], "task": "refering", "answer_template": "The \"woman with long dark hair in denim shirt , far left of picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 138, 139, 161, 162, 163, 184, 185, 186, 207, 208, 209, 230, 231, 232, 253, 254, 255], "bbox": [0.0045156250000000005, 0.3335128805620609, 0.121984375, 0.7940281030444963], "iscrowd": 0, "area": 10492.830750000001, "rle": {"size": [427, 640], "counts": "`\\14X=5WCU4g7W1@10000000O0100000000O1000O1000O01O1N2O1O001O1O1N3N1O2N1O1O2N1N3N1O2N1O1O2M2O2N1O2N1O1N3N1O2N1O2N1N2O3M4L4L3M4K5L4L4L4L3M4K5L4L4L3M4L4K5K5I7I7I7I\\RZ7"}, "label": "woman with long dark hair in denim shirt , far left of picture"}]} {"id": 533220, "image": "COCO_train2014_000000533220.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman off to the left with a demin jacket on\"."}], "task": "refering", "answer_template": "The \"a woman off to the left with a demin jacket on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 138, 139, 161, 162, 163, 184, 185, 186, 207, 208, 209, 230, 231, 232, 253, 254, 255], "bbox": [0.0045156250000000005, 0.3335128805620609, 0.121984375, 0.7940281030444963], "iscrowd": 0, "area": 10492.830750000001, "rle": {"size": [427, 640], "counts": "`\\14X=5WCU4g7W1@10000000O0100000000O1000O1000O01O1N2O1O001O1O1N3N1O2N1O1O2N1N3N1O2N1O1O2M2O2N1O2N1O1N3N1O2N1O2N1N2O3M4L4L3M4K5L4L4L4L3M4K5L4L4L3M4L4K5K5I7I7I7I\\RZ7"}, "label": "a woman off to the left with a demin jacket on"}]} {"id": 221927, "image": "COCO_train2014_000000221927.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant in front of a smaller elephant\"."}], "task": "refering", "answer_template": "The \"an elephant in front of a smaller elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 219, 235, 237, 238, 239, 240, 241, 242, 258, 260, 261, 263, 264, 283, 284, 286, 287], "bbox": [0.21742187500000001, 0.3955, 0.5646093750000001, 0.7505624999999999], "iscrowd": 0, "area": 23283.076850000005, "rle": {"size": [480, 640], "counts": "n^Q23g>7I7H9H7H8I8H7H8I8G8I7I8G8I7H9I5OK5K5K6J5K5K5K6J5K5K5J7J5M31O100O2N1O100O1O1O100O1O1O100O1O2O0000000000000000000000001O00000000000O10000000000UDmMi:T2hDLf9g200000001O000000001O000000001O000000001O0O1000001O00SN_F[Na9e1iFQNW9o1RGhMn8X2\\G_Mc8a2fGVMZ8j2PHlLP8S3o10000000000O100000000O1000000000O010000000000O10000000000O10000000000O100004L5K4L5K5K4L5J6K4L5K5K4L1O001O001O001O001O001O001O0010O100O100O100O100O100O100O100O100O100I7F:G9F:G9H8N2N2N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2N2N2N2M3N2J6H8H8I7H8HeaR4"}, "label": "an elephant in front of a smaller elephant"}]} {"id": 221927, "image": "COCO_train2014_000000221927.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant that is front and center\"."}], "task": "refering", "answer_template": "The \"elephant that is front and center\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 219, 235, 237, 238, 239, 240, 241, 242, 258, 260, 261, 263, 264, 283, 284, 286, 287], "bbox": [0.21742187500000001, 0.3955, 0.5646093750000001, 0.7505624999999999], "iscrowd": 0, "area": 23283.076850000005, "rle": {"size": [480, 640], "counts": "n^Q23g>7I7H9H7H8I8H7H8I8G8I7I8G8I7H9I5OK5K5K6J5K5K5K6J5K5K5J7J5M31O100O2N1O100O1O1O100O1O1O100O1O2O0000000000000000000000001O00000000000O10000000000UDmMi:T2hDLf9g200000001O000000001O000000001O000000001O0O1000001O00SN_F[Na9e1iFQNW9o1RGhMn8X2\\G_Mc8a2fGVMZ8j2PHlLP8S3o10000000000O100000000O1000000000O010000000000O10000000000O10000000000O100004L5K4L5K5K4L5J6K4L5K5K4L1O001O001O001O001O001O001O0010O100O100O100O100O100O100O100O100O100I7F:G9F:G9H8N2N2N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2N2N2N2M3N2J6H8H8I7H8HeaR4"}, "label": "elephant that is front and center"}]} {"id": 525039, "image": "COCO_train2014_000000525039.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in the desert throwing a frisbee\"."}], "task": "refering", "answer_template": "The \"a man in the desert throwing a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 150, 151, 172, 173, 174, 195, 196, 197, 218, 219, 220, 241, 242, 243, 264, 265, 266, 287, 288, 289, 310, 311, 312, 313, 333, 334, 355, 356, 357], "bbox": [0.46104687499999997, 0.34645833333333337, 0.62384375, 0.9347083333333334], "iscrowd": 0, "area": 12619.311550000002, "rle": {"size": [480, 640], "counts": "gfZ41o>1N2O1N2N2O1N2O1N1cKETJ^5?bHUO5;W7d0^HWO96V7f0\\H[O\\GVOf0S1f0ZOU7>dGRO<]1h0SOW7=iGQO3g1k0kNV7>`HU1KcNc77_He1POhN_8D^HZ3`7gL[H_3b7eLXH`3e7eLXH[3e7kLXHV3c7QMYHQ3a7[MXHe2g7[2O1N101O1O0SOmHoIS7o5RImIP7Q6]IbId6\\6aI`Ia6^6cI_I]6_6R10YMkIeLU6Q3oJULQ5g3^KnKb4P4dKlK\\4R4jKiKW4U4PLfKP4\\4RL`Kn3c4SLXKn3j4TLRKk3R5VLiJk3Y5WLcJi3_5VL_Jk3d5SL[Jn3g5d24L3]GSJn7o5oGQJQ8R6kGoIU8S6hGoIX8e0fG\\42oJX8a0jG`4MQKX8=mGc4IQK[88PHf4DSK\\85RHW3@kM2jN[82UHX3AkMMlN^8MXHZ3@XNV9h1lFVNT9i1oFUNR9i1QGUNQ9h1RGWNn8g1VGVNk8h1YGUNk8f1YGWNk8L]F in this image.", "objects": [{"patches": [127, 150, 151, 172, 173, 174, 195, 196, 197, 218, 219, 220, 241, 242, 243, 264, 265, 266, 287, 288, 289, 310, 311, 312, 313, 333, 334, 355, 356, 357], "bbox": [0.46104687499999997, 0.34645833333333337, 0.62384375, 0.9347083333333334], "iscrowd": 0, "area": 12619.311550000002, "rle": {"size": [480, 640], "counts": "gfZ41o>1N2O1N2N2O1N2O1N1cKETJ^5?bHUO5;W7d0^HWO96V7f0\\H[O\\GVOf0S1f0ZOU7>dGRO<]1h0SOW7=iGQO3g1k0kNV7>`HU1KcNc77_He1POhN_8D^HZ3`7gL[H_3b7eLXH`3e7eLXH[3e7kLXHV3c7QMYHQ3a7[MXHe2g7[2O1N101O1O0SOmHoIS7o5RImIP7Q6]IbId6\\6aI`Ia6^6cI_I]6_6R10YMkIeLU6Q3oJULQ5g3^KnKb4P4dKlK\\4R4jKiKW4U4PLfKP4\\4RL`Kn3c4SLXKn3j4TLRKk3R5VLiJk3Y5WLcJi3_5VL_Jk3d5SL[Jn3g5d24L3]GSJn7o5oGQJQ8R6kGoIU8S6hGoIX8e0fG\\42oJX8a0jG`4MQKX8=mGc4IQK[88PHf4DSK\\85RHW3@kM2jN[82UHX3AkMMlN^8MXHZ3@XNV9h1lFVNT9i1oFUNR9i1QGUNQ9h1RGWNn8g1VGVNk8h1YGUNk8f1YGWNk8L]F in this image.", "objects": [{"patches": [139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 167, 187, 188, 189, 190, 210, 211, 212, 213, 214, 235, 236, 237, 238], "bbox": [0.07078124999999999, 0.3550625, 0.37584375000000003, 0.6269583333333333], "iscrowd": 0, "area": 10535.367950000003, "rle": {"size": [480, 640], "counts": "The01n>2O1N2N2O1N2O1N2N2O1N2O0O2N2O1N2O1N1O2O0O2O0O2O0O2N101N101N1O2O0O2O0O2O001O001O1O3N1N3M3M3M2N3M3N1N3M2N001O0O2O1O0O2O001N101O1N101O0O2O001O1O0O2O001O1O001O001O001O1O001O001O001O001O1O001O001N101O0100O2O0O100O100O1O101N1O100O1O1O102M4L4M2M4L4M2M4L4L3N3L4L3N2M2N2O0O2N2N2O1N2N2O1N2N2O1N2N1O010O1O10O01O1O001O1O001O1O010O1O001O1O001O1O00100O1O001O1O00001O001O01O01O00001O00001O00010O001O3MgVk5"}, "label": "piece pizzza"}]} {"id": 443136, "image": "COCO_train2014_000000443136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a very large pizza the size of a table\"."}], "task": "refering", "answer_template": "The \"a very large pizza the size of a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 210, 211, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 231, 232, 233, 234, 235, 236, 237, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.4815833333333333, 1.0, 1.0], "iscrowd": 0, "area": 131737.21579999998, "rle": {"size": [480, 640], "counts": "]:[4d:1O1O010O1O100O1O010O1O1O100O010O1O101N1O100O100O1O100O1O2O0O100O1O100O1O2O0O100O2N101N2N2N2O1N2N2O0O1O100O1O1O100O1O1O1O100O1O1O1O1O100O1O10000O10000O10000O10000O10000O10000O10000O1000000O10000O1000000O10000O1000000O1000000O10000O100000000001O001O1O001O1O001O1O001O001O001O001O001O001O001O001O001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O001O00001O001O001O001O00001O001O00dMPKiJo4W5SKgJm4X5VKfJj4Z5WKfJh4Z5ZKdJf4[5]KdJb4\\5`KbJ`4^5aKbJ^4^5dK`J\\4`5fK^JZ4a5hK_JW4a5kK]JU4c5mK\\JR4d5PLZJP4f5QLZJn3f5TLXJl3h5VLWJi3i5YLUJg3j5[LUJe3k5]LTJb3l5`LRJ`3n5bLQJ]3o5dLPJ\\3P6fLoIY3Q6iLmIW3S6kLkIU3U6lLkIS3T6PMjIP3V6QMjIn2V6TMhIl2X6UMhIj2X6XMfIh2Z6YMeIg2[6ZMeIe2[6]McIc2]6^McIa2]6`MbI`2^6bMaI]2^6eMaI[2_6gM`IX2`6iM_IW2a6jM^IV2b6lM]IS2c6nM\\IR2d6oM\\IP2d6RNZIn1f6SNZIl1f6VNXIj1h6WNWIi1h6YNWIg1i6[NUIe1k6\\NTId1l6]NSIc1m6_NPIb1P7_NnHb1R7`NkHa1U7`NiHa1W7`NhH`1X7bNgH]1Y7dNgH[1X7gNhHX1X7jNgHU1Y7Y3000000000000000000O100000000000000000000000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000001O00000000000000000000001O00000000000000000000001O0000000000000000001O000000000000001O00000000000000001O000000000000001O0000000000001O000000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O001O1O001O1O001O1O001O1O001O001O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O2N1O1N3K4M`E"}, "label": "a very large pizza the size of a table"}]} {"id": 443136, "image": "COCO_train2014_000000443136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a huge pepperoni and cheese pizza\"."}], "task": "refering", "answer_template": "The \"a huge pepperoni and cheese pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 210, 211, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 231, 232, 233, 234, 235, 236, 237, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.4815833333333333, 1.0, 1.0], "iscrowd": 0, "area": 131737.21579999998, "rle": {"size": [480, 640], "counts": "]:[4d:1O1O010O1O100O1O010O1O1O100O010O1O101N1O100O100O1O100O1O2O0O100O1O100O1O2O0O100O2N101N2N2N2O1N2N2O0O1O100O1O1O100O1O1O1O100O1O1O1O1O100O1O10000O10000O10000O10000O10000O10000O10000O1000000O10000O1000000O10000O1000000O1000000O10000O100000000001O001O1O001O1O001O1O001O001O001O001O001O001O001O001O001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O001O00001O001O001O001O00001O001O00dMPKiJo4W5SKgJm4X5VKfJj4Z5WKfJh4Z5ZKdJf4[5]KdJb4\\5`KbJ`4^5aKbJ^4^5dK`J\\4`5fK^JZ4a5hK_JW4a5kK]JU4c5mK\\JR4d5PLZJP4f5QLZJn3f5TLXJl3h5VLWJi3i5YLUJg3j5[LUJe3k5]LTJb3l5`LRJ`3n5bLQJ]3o5dLPJ\\3P6fLoIY3Q6iLmIW3S6kLkIU3U6lLkIS3T6PMjIP3V6QMjIn2V6TMhIl2X6UMhIj2X6XMfIh2Z6YMeIg2[6ZMeIe2[6]McIc2]6^McIa2]6`MbI`2^6bMaI]2^6eMaI[2_6gM`IX2`6iM_IW2a6jM^IV2b6lM]IS2c6nM\\IR2d6oM\\IP2d6RNZIn1f6SNZIl1f6VNXIj1h6WNWIi1h6YNWIg1i6[NUIe1k6\\NTId1l6]NSIc1m6_NPIb1P7_NnHb1R7`NkHa1U7`NiHa1W7`NhH`1X7bNgH]1Y7dNgH[1X7gNhHX1X7jNgHU1Y7Y3000000000000000000O100000000000000000000000000O1000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000001O00000000000000000000001O00000000000000000000001O0000000000000000001O000000000000001O00000000000000001O000000000000001O0000000000001O000000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O001O1O001O1O001O1O001O1O001O001O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O2N1O1N3K4M`E"}, "label": "a huge pepperoni and cheese pizza"}]} {"id": 41730, "image": "COCO_train2014_000000041730.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bento box with rice in it\"."}], "task": "refering", "answer_template": "The \"the bento box with rice in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 201, 202, 203], "bbox": [0.35134375, 0.10247916666666666, 0.952703125, 0.4808541666666667], "iscrowd": 0, "area": 57340.47565, "rle": {"size": [480, 640], "counts": "QcY3k0k=>A>E;J5L5J6K5J6J5L5J6J6K4K6K5M3L3M1O100O1O100O1O1O100O1O1O100O1O100O1000000O100000000O1000000O100000000O100000000O1000000O1000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000001O00001O00001O00001O0000001O00001O00001O01O01O00001O00001O00001O00001O001O1O1O1O001O1O1O1O1O001O1O1O1O1O001O00000000000000000000000000000000000001O0000000000000000001O00000000001O00000000001O000O10001O000O1000001O000O1000001O000O101O001O0O101N1O2N1O2N1O1O2N1O2N101N1O1O2N2N2N2N3M2N2M3N2N2N2M4M2N2N2M3N2N2N3L3N2N2N2M6K6J6J6J6I6K6J6J6I7G9H8G9GWm="}, "label": "the bento box with rice in it"}]} {"id": 41730, "image": "COCO_train2014_000000041730.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of rice and seaweed next to another bowl of broccolli and sweet potato\"."}], "task": "refering", "answer_template": "The \"a bowl of rice and seaweed next to another bowl of broccolli and sweet potato\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 201, 202, 203], "bbox": [0.35134375, 0.10247916666666666, 0.952703125, 0.4808541666666667], "iscrowd": 0, "area": 57340.47565, "rle": {"size": [480, 640], "counts": "QcY3k0k=>A>E;J5L5J6K5J6J5L5J6J6K4K6K5M3L3M1O100O1O100O1O1O100O1O1O100O1O100O1000000O100000000O1000000O100000000O100000000O1000000O1000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000001O00001O00001O00001O0000001O00001O00001O01O01O00001O00001O00001O00001O001O1O1O1O001O1O1O1O1O001O1O1O1O1O001O00000000000000000000000000000000000001O0000000000000000001O00000000001O00000000001O000O10001O000O1000001O000O1000001O000O101O001O0O101N1O2N1O2N1O1O2N1O2N101N1O1O2N2N2N2N3M2N2M3N2N2N2M4M2N2N2M3N2N2N3L3N2N2N2M6K6J6J6J6I6K6J6J6I7G9H8G9GWm="}, "label": "a bowl of rice and seaweed next to another bowl of broccolli and sweet potato"}]} {"id": 41730, "image": "COCO_train2014_000000041730.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a container of various foods sitting on a table next to other foods\"."}], "task": "refering", "answer_template": "The \"a container of various foods sitting on a table next to other foods\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361], "bbox": [0.058984375, 0.42852083333333335, 0.8039375, 0.9251458333333333], "iscrowd": 0, "area": 102965.78349999998, "rle": {"size": [480, 640], "counts": "`Tb0>Y>f0[O`0@;E;D in this image.", "objects": [{"patches": [164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361], "bbox": [0.058984375, 0.42852083333333335, 0.8039375, 0.9251458333333333], "iscrowd": 0, "area": 102965.78349999998, "rle": {"size": [480, 640], "counts": "`Tb0>Y>f0[O`0@;E;D in this image.", "objects": [{"patches": [271, 272, 288, 289, 290, 293, 294, 295, 305, 306, 307, 308, 309, 311, 312, 313, 316, 317, 318, 328, 329, 330, 331, 332, 333, 334, 335, 336, 339, 340, 341], "bbox": [0.292609375, 0.7760705882352941, 0.8621718749999999, 1.0], "iscrowd": 0, "area": 14042.947199999997, "rle": {"size": [425, 640], "counts": "[P^23V<2^D6a;i001O1O1O:F0000000001O0000000000000000001O0000000000001O0000000000001O00O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000002N2N7Ia0_O1O00O100O10000O1O1N2N2N200O1000000O100000000O1000000O12N1UOSD8o;C\\D5c;Ko02M2N2N2N2N2N2N2O1N1O2N2N2N2N2N0001O000000000000001O0000000000000000000000000000000000UOCjD=m:6iDJW;T1000000000000000000000000000000000000000000000000000000000000000000000000O1000000O10000O1O1O1O1O1O100OfNZE3e:\\O`F1moS13\\[lN`0A?@;EO1O1N2O1O1N2O1O1N2O1O11O1O001O1O1A>B?O5L8G1O0000000000000000000000000000000000000001O00000000000000000000O100O10O0100OdMPFa1P:RNZGQ1o:@`0@eST1"}, "label": "the fold up chairs in front"}]} {"id": 254726, "image": "COCO_train2014_000000254726.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a row of collapsible lawn chairs in multiple colors\"."}], "task": "refering", "answer_template": "The \"a row of collapsible lawn chairs in multiple colors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [271, 272, 288, 289, 290, 293, 294, 295, 305, 306, 307, 308, 309, 311, 312, 313, 316, 317, 318, 328, 329, 330, 331, 332, 333, 334, 335, 336, 339, 340, 341], "bbox": [0.292609375, 0.7760705882352941, 0.8621718749999999, 1.0], "iscrowd": 0, "area": 14042.947199999997, "rle": {"size": [425, 640], "counts": "[P^23V<2^D6a;i001O1O1O:F0000000001O0000000000000000001O0000000000001O0000000000001O00O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000002N2N7Ia0_O1O00O100O10000O1O1N2N2N200O1000000O100000000O1000000O12N1UOSD8o;C\\D5c;Ko02M2N2N2N2N2N2N2O1N1O2N2N2N2N2N0001O000000000000001O0000000000000000000000000000000000UOCjD=m:6iDJW;T1000000000000000000000000000000000000000000000000000000000000000000000000O1000000O10000O1O1O1O1O1O100OfNZE3e:\\O`F1moS13\\[lN`0A?@;EO1O1N2O1O1N2O1O1N2O1O11O1O001O1O1A>B?O5L8G1O0000000000000000000000000000000000000001O00000000000000000000O100O10O0100OdMPFa1P:RNZGQ1o:@`0@eST1"}, "label": "a row of collapsible lawn chairs in multiple colors"}]} {"id": 418569, "image": "COCO_train2014_000000418569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man playing in a water\"."}], "task": "refering", "answer_template": "The \"the man playing in a water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 170, 171, 172, 193, 194, 195, 216, 217, 218, 238, 239, 240, 241, 261, 262, 264, 265], "bbox": [0.340484375, 0.36189583333333336, 0.533453125, 0.7122083333333333], "iscrowd": 0, "area": 7910.863750000003, "rle": {"size": [480, 640], "counts": "\\oV31m>2O2N1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O100O100O0010O0010O01O00ZO^OoBc0i2N1O21N2O0O2O1N2O001O1O1O001O1O10^NgD\\OX;d0jDZOW;e0kDZOT;f0oDXOS;e0oDZOV;`0lD^Oc;3`DKe;O]D0d;N]D2d;M\\D3V in this image.", "objects": [{"patches": [148, 170, 171, 172, 193, 194, 195, 216, 217, 218, 238, 239, 240, 241, 261, 262, 264, 265], "bbox": [0.340484375, 0.36189583333333336, 0.533453125, 0.7122083333333333], "iscrowd": 0, "area": 7910.863750000003, "rle": {"size": [480, 640], "counts": "\\oV31m>2O2N1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O100O100O0010O0010O01O00ZO^OoBc0i2N1O21N2O0O2O1N2O001O1O1O001O1O10^NgD\\OX;d0jDZOW;e0kDZOT;f0oDXOS;e0oDZOV;`0lD^Oc;3`DKe;O]D0d;N]D2d;M\\D3V in this image.", "objects": [{"patches": [71, 72, 94, 95, 96, 117, 118, 119, 120, 121, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 168, 187, 188, 189, 210, 211, 212, 233, 234, 235, 256, 257, 279, 280, 303, 304, 327], "bbox": [0.11103125, 0.2051875, 0.34090625, 0.8337708333333335], "iscrowd": 0, "area": 15151.872599999999, "rle": {"size": [480, 640], "counts": "U^Q14P=MjD:o:7_D1[;1bD4Y;NdD7Y;JeD:W;GgD?U;BiDb0S;@jDe0U3bNn3j0kHh0S3_NQ4l0iHi0R3]NT4k0gHn0P3YNX4j0eHR1o2VNZ4j0eHR1o2VN[4j0bHS1Q3UN[4j0bHS1Q3TN]4j0_HU1Q3TN_4h0^HU1R3UN_4V3_KlL`4U3^KmLa4V3[KlLd4Y3VKhLj4]3PKeLo4_3lJdLR5a3hJaLW5d3cJ^L\\5g3^J[La5j3YJXLf5m3TJbKgNJT7i4oI]KoNJR7o4gIWKYOJP7V5^IPKEIl6_5VIhJ0Ij6`5TIfJ5Ig6a5RIgJ8Hf6a5PIgJ=Gc6b5oHfJ`0I`6a5nHfJd0I`6U6cIiI`6S6bIlI`6Q6cImI`6o5bIPJa6\\6RIbIQ7]6nHbIU7]6lH`IZ7\\6d03N1N2N2N3N1N3M3M3M3N1N3M3M7JmMiH_LS7`3XIXLf6f3gIQLW6n3TJiKl5T4YJkKe5T4^JlK`5S4cJlK]5CPIg3g1eLW5BVIf3f1fLT5BZIf3d1gLR5B\\Id3e1hLo4D_I`3e1kLl4E`I]3g1lLi4GbIZ3h1nLe4HeIW3i1oLc4JgIS3h1RM`4KmI5YOm1\\2RN_4KQJO]Ol1U2YN]4MPJKIf1m1aNZ4MPJI:Y1_1POW4NPJGk0i0R1BR4OPJD]19e03o3OoIDo4:S12oIAQ5:S15NI37NG48NE3X7O0100O0010O010O0100O0010O010O010O01O10O010O01O100O1000O101O00001O001O010O0010O00000O2O0O1O10O000010O0001O011N2J6K4LWdU6"}, "label": "a man in air throwing frisbee weraing a light yellow tea shirt"}]} {"id": 418569, "image": "COCO_train2014_000000418569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing an orange shirt , jumping\"."}], "task": "refering", "answer_template": "The \"a man wearing an orange shirt , jumping\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 94, 95, 96, 117, 118, 119, 120, 121, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 168, 187, 188, 189, 210, 211, 212, 233, 234, 235, 256, 257, 279, 280, 303, 304, 327], "bbox": [0.11103125, 0.2051875, 0.34090625, 0.8337708333333335], "iscrowd": 0, "area": 15151.872599999999, "rle": {"size": [480, 640], "counts": "U^Q14P=MjD:o:7_D1[;1bD4Y;NdD7Y;JeD:W;GgD?U;BiDb0S;@jDe0U3bNn3j0kHh0S3_NQ4l0iHi0R3]NT4k0gHn0P3YNX4j0eHR1o2VNZ4j0eHR1o2VN[4j0bHS1Q3UN[4j0bHS1Q3TN]4j0_HU1Q3TN_4h0^HU1R3UN_4V3_KlL`4U3^KmLa4V3[KlLd4Y3VKhLj4]3PKeLo4_3lJdLR5a3hJaLW5d3cJ^L\\5g3^J[La5j3YJXLf5m3TJbKgNJT7i4oI]KoNJR7o4gIWKYOJP7V5^IPKEIl6_5VIhJ0Ij6`5TIfJ5Ig6a5RIgJ8Hf6a5PIgJ=Gc6b5oHfJ`0I`6a5nHfJd0I`6U6cIiI`6S6bIlI`6Q6cImI`6o5bIPJa6\\6RIbIQ7]6nHbIU7]6lH`IZ7\\6d03N1N2N2N3N1N3M3M3M3N1N3M3M7JmMiH_LS7`3XIXLf6f3gIQLW6n3TJiKl5T4YJkKe5T4^JlK`5S4cJlK]5CPIg3g1eLW5BVIf3f1fLT5BZIf3d1gLR5B\\Id3e1hLo4D_I`3e1kLl4E`I]3g1lLi4GbIZ3h1nLe4HeIW3i1oLc4JgIS3h1RM`4KmI5YOm1\\2RN_4KQJO]Ol1U2YN]4MPJKIf1m1aNZ4MPJI:Y1_1POW4NPJGk0i0R1BR4OPJD]19e03o3OoIDo4:S12oIAQ5:S15NI37NG48NE3X7O0100O0010O010O0100O0010O010O010O01O10O010O01O100O1000O101O00001O001O010O0010O00000O2O0O1O10O000010O0001O011N2J6K4LWdU6"}, "label": "a man wearing an orange shirt , jumping"}]} {"id": 181009, "image": "COCO_train2014_000000181009.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light blue backpack on the lap of the passenger next to the one using his cell phone\"."}], "task": "refering", "answer_template": "The \"a light blue backpack on the lap of the passenger next to the one using his cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [160, 181, 182, 183, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275, 297, 298], "bbox": [0.75234375, 0.3575, 1.0, 0.7575000000000001], "iscrowd": 0, "area": 14973.571650000002, "rle": {"size": [480, 640], "counts": "ZVR71n>2O1N2O1N2N101N2O1N2O1N2O1O1O0O2O1O1O1O1O1O1O001N2O1O1O1O1O1O1O0O2O1O1N2N2O1N2O1N1O2O1N2O0O2N101N1O2O0O2O0O2N101N101N1O2O0O2O0O2N101N1O2O0O2O0O2N101N101N1O2O0O2N101N2O0O2N101N101N1O2O0O2O0O2N101N1O2O0O2O0O2N101N101N1O2O0O2O0O2N101N1O2O0O2N1O2N1O2O0O2N1O2N1O2O1N1O2N1O2N101N1O2N1O2N101N1O2N1O2N1O2O0O2N1O2N1O2O0O2N1dJ"}, "label": "a light blue backpack on the lap of the passenger next to the one using his cell phone"}]} {"id": 181009, "image": "COCO_train2014_000000181009.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"strap on other person ' s shoulder\"."}], "task": "refering", "answer_template": "The \"strap on other person ' s shoulder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 66, 67, 68, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 132, 133, 134, 155, 156], "bbox": [0.7449375, 0.06741666666666667, 0.9977499999999999, 0.3910208333333333], "iscrowd": 0, "area": 7663.129899999998, "rle": {"size": [480, 640], "counts": "lgo62m>3M4L3M4M2M3M4L3M4L3N2M4L1O00000O10001O0000000000000000001N1000000000000000001O00000000000001O01O00000001O00000000010O00000000001O0001O000001O00000001O10O01O1O001O10O01O1O00100O010O00100O010O1O010O10O01O10O0100O0010O0010O01O00010O0010O0001O010O00010O001O01O01O010O000010O01O01O01O0010O0001O010O0010O0001O010O00010O001O01O0Q>"}, "label": "strap on other person ' s shoulder"}]} {"id": 181009, "image": "COCO_train2014_000000181009.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black backpack strap draped over a shoulder\"."}], "task": "refering", "answer_template": "The \"a black backpack strap draped over a shoulder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 66, 67, 68, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 132, 133, 134, 155, 156], "bbox": [0.7449375, 0.06741666666666667, 0.9977499999999999, 0.3910208333333333], "iscrowd": 0, "area": 7663.129899999998, "rle": {"size": [480, 640], "counts": "lgo62m>3M4L3M4M2M3M4L3M4L3N2M4L1O00000O10001O0000000000000000001N1000000000000000001O00000000000001O01O00000001O00000000010O00000000001O0001O000001O00000001O10O01O1O001O10O01O1O00100O010O00100O010O1O010O10O01O10O0100O0010O0010O01O00010O0010O0001O010O00010O001O01O01O010O000010O01O01O01O0010O0001O010O0010O0001O010O00010O001O01O0Q>"}, "label": "a black backpack strap draped over a shoulder"}]} {"id": 516889, "image": "COCO_train2014_000000516889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant about to charge\"."}], "task": "refering", "answer_template": "The \"elephant about to charge\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 13, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 261, 262, 263, 264, 265, 287, 288], "bbox": [0.219234375, 0.022058823529411766, 0.655921875, 0.7381302521008404], "iscrowd": 0, "area": 50746.34854999998, "rle": {"size": [476, 640], "counts": "ncQ25d>6K4K5K6J5L4K5K6K4K5M2N2N2M2O2N2N1O2N2N2N1O2M3N1O2N2N2N1O2N2N1O2M2O1O2N1O101N10001O001O00001O000O2O00001O001O000SEeLY:\\3_EoL]:Q3]EXM_:f3M3M3M3M4L4L`0@a0_O?A1N3N1O1O2N2N3M2N2N2N4K5L4L4L4L4L4L4L5K4L4KA3N2M3N1N3N2M3N0O2O0O100O100O1000O10000000001^NeK`H\\4o6XLnHh3j6cLSI]3h6jLVIV3e6RMXIn2c6ZMZIf2a6bM[I`2_6gM[I]2e6fMUI]2k6j200000001O0O1N2N2N2N2O2M2N2N2N27J2M4M2M3N2M4M2M4L4M3L3N3L4M3L4M2M4jJjG_3Y8SL[Hc3i7PLaHP4b7bKhH]4l8000iIbKQ2]4oMjKj1U4WNjKj1V4UNjKl1V4TNiKm1W4SNhKn1X4RNgKo1X4RNgKRMNc3[4ZOgKSM9Y3P4DfKTMd0n2f3NeKUMo0c2\\38dKVMX1Z2S3`0eKWM`1Q2k2h0dKXMh1i1d2o0cKYMQ2`1\\2W1cKWM[2Z1R2^1dKVMd2T1g1g1fKRMm2o0]1n1hKoLU3l0S1T2jKmL]3g0i0[2kKkLg3b0>c2lKaLW4d0Lk2oK\\Lh4:YOY3QLYLj4=UOY3RLWLn4=QO[3d1dL]N[3e1cL[N]3f1aL\\N^3f1`LZN_3i1_LXN`3b601O001N1O2O0O2O0O2N101N101N1O2O0O2O001N100O2O001N101O1N101O1N2O1N2O001N2O1O1N101N2O1O1N2O001N2O1L4M2WOQCCR=9SCCQ=8TCDQ=6SCGR=3SCJQ=0TCLQ=NTCOXcV3"}, "label": "elephant about to charge"}]} {"id": 516889, "image": "COCO_train2014_000000516889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is a large elephant standing outside\"."}], "task": "refering", "answer_template": "The \"there is a large elephant standing outside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 13, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 261, 262, 263, 264, 265, 287, 288], "bbox": [0.219234375, 0.022058823529411766, 0.655921875, 0.7381302521008404], "iscrowd": 0, "area": 50746.34854999998, "rle": {"size": [476, 640], "counts": "ncQ25d>6K4K5K6J5L4K5K6K4K5M2N2N2M2O2N2N1O2N2N2N1O2M3N1O2N2N2N1O2N2N1O2M2O1O2N1O101N10001O001O00001O000O2O00001O001O000SEeLY:\\3_EoL]:Q3]EXM_:f3M3M3M3M4L4L`0@a0_O?A1N3N1O1O2N2N3M2N2N2N4K5L4L4L4L4L4L4L5K4L4KA3N2M3N1N3N2M3N0O2O0O100O100O1000O10000000001^NeK`H\\4o6XLnHh3j6cLSI]3h6jLVIV3e6RMXIn2c6ZMZIf2a6bM[I`2_6gM[I]2e6fMUI]2k6j200000001O0O1N2N2N2N2O2M2N2N2N27J2M4M2M3N2M4M2M4L4M3L3N3L4M3L4M2M4jJjG_3Y8SL[Hc3i7PLaHP4b7bKhH]4l8000iIbKQ2]4oMjKj1U4WNjKj1V4UNjKl1V4TNiKm1W4SNhKn1X4RNgKo1X4RNgKRMNc3[4ZOgKSM9Y3P4DfKTMd0n2f3NeKUMo0c2\\38dKVMX1Z2S3`0eKWM`1Q2k2h0dKXMh1i1d2o0cKYMQ2`1\\2W1cKWM[2Z1R2^1dKVMd2T1g1g1fKRMm2o0]1n1hKoLU3l0S1T2jKmL]3g0i0[2kKkLg3b0>c2lKaLW4d0Lk2oK\\Lh4:YOY3QLYLj4=UOY3RLWLn4=QO[3d1dL]N[3e1cL[N]3f1aL\\N^3f1`LZN_3i1_LXN`3b601O001N1O2O0O2O0O2N101N101N1O2O0O2O001N100O2O001N101O1N101O1N2O1N2O001N2O1O1N101N2O1O1N2O001N2O1L4M2WOQCCR=9SCCQ=8TCDQ=6SCGR=3SCJQ=0TCLQ=NTCOXcV3"}, "label": "there is a large elephant standing outside"}]} {"id": 795, "image": "COCO_train2014_000000000795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"train\"."}], "task": "refering", "answer_template": "The \"train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 93, 94, 95, 96, 97, 98, 99, 110, 111, 112, 113, 114, 115, 116, 117, 128, 129, 130, 131, 132, 147, 148, 149, 150, 165, 166, 167], "bbox": [0.15668, 0.36855421686746986, 0.54314, 0.8022590361445784], "iscrowd": 0, "area": 16278.01105, "rle": {"size": [332, 500], "counts": "d_i02S:;EM4O00001O010O00001O010O00001O0O2O0O1O20O2O0O1hFTOn8V101N101N100O2N101N10N1M3M4L3M4M2M3N3M2N3O00001O001O0GYFMh92ZFLf93:01N101O00010O0018G0001O001N1O1O2M2O2NfRZ2"}, "label": "train"}]} {"id": 795, "image": "COCO_train2014_000000000795.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow train is in the station with people walking away from it\"."}], "task": "refering", "answer_template": "The \"a yellow train is in the station with people walking away from it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 93, 94, 95, 96, 97, 98, 99, 110, 111, 112, 113, 114, 115, 116, 117, 128, 129, 130, 131, 132, 147, 148, 149, 150, 165, 166, 167], "bbox": [0.15668, 0.36855421686746986, 0.54314, 0.8022590361445784], "iscrowd": 0, "area": 16278.01105, "rle": {"size": [332, 500], "counts": "d_i02S:;EM4O00001O010O00001O010O00001O0O2O0O1O20O2O0O1hFTOn8V101N101N100O2N101N10N1M3M4L3M4M2M3N3M2N3O00001O001O0GYFMh92ZFLf93:01N101O00010O0018G0001O001N1O1O2M2O2NfRZ2"}, "label": "a yellow train is in the station with people walking away from it"}]} {"id": 33581, "image": "COCO_train2014_000000033581.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the rightmost spoon\"."}], "task": "refering", "answer_template": "The \"the rightmost spoon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 102, 103, 116, 117, 131, 144, 145, 158, 159], "bbox": [0.5236533957845433, 0.22742187500000002, 1.0, 0.44212500000000005], "iscrowd": 0, "area": 1627.1872499999974, "rle": {"size": [640, 427], "counts": "el\\41oc01O1O10O01O1O000001O00000000001O00O10000O10000O10000O01000O10000O10000O01000O10000O0100000O1000O010000O10O10O010O010O0100O010O010O010O1O010O010O01O0010O01O0001N10000O2O00000O2OO1HLe\\O3[c00c\\O0]c01b\\OO^c03`\\OM_c091000O01000O0100O001O10O01O100000000000000000000O100000O1001O0000001O000010O0001O0001O01O00010O0000010O00010O00010O0001O01O01O01O010O01O01O01O01O010O00010O01O01O010O01O010O00010O0010O010O00010O01O010O01O01O010O0010IeK"}, "label": "the rightmost spoon"}]} {"id": 33581, "image": "COCO_train2014_000000033581.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the spoon to the left of the cup\"."}], "task": "refering", "answer_template": "The \"the spoon to the left of the cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 105, 106, 121, 122, 137, 138, 139], "bbox": [0.0007962529274004685, 0.269390625, 0.3300702576112412, 0.42631250000000004], "iscrowd": 0, "area": 1565.8198000000004, "rle": {"size": [640, 427], "counts": "\\59hc000000000O1000O100000O1000000000O2O001O000010O01O01O01O0010O0001O01O01O010O000010O01O01O1O1O010O1O1O1O10N100O1O2O0O2O0O10O010O01O010O01O010O01O010O010O01O010O01O010O01O010O0010O00010O01O010O000010O000010O0001O01O01O00010O00O101O000O101O000O101O000O101O00000O2O00000O10001O0O100000g_b5"}, "label": "the spoon to the left of the cup"}]} {"id": 467760, "image": "COCO_train2014_000000467760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra with its head behind the other zebra\"."}], "task": "refering", "answer_template": "The \"zebra with its head behind the other zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 167, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 239, 240], "bbox": [0.264125, 0.4180281690140845, 0.5558125, 0.6880046948356808], "iscrowd": 0, "area": 7326.071000000002, "rle": {"size": [426, 640], "counts": "ibV21U=3M3M2O2M3M3M3YOKXD6h;f0O1N2O1_DcNU;Y1nDjNQ;P1SESOl:h0WE[Oh:?]ECb:7bEK^:6_EM`:;UEIj:?kDET;Y1O1O1O100000000O10000000O1M3L3N3M3L4N2O1N2O1N2O1N2O2N11O001O1O010O001O1O0010O01O1O0010O01O1O001O010O000000000000000000000000001O000000O3N3M1OO010000O10000O10hDhN]:Y1\\EnNd:R1VETOj:k0QE[Oo:a1O0O10001O000O10001O000ON3L4L4L4J8G9L4M4M2N1N3N2M2O2N1M4M2M4M5JRf3LTZL1N3N2N2M3G9M3N101O0000010N1O1O001N2O1N1O2NVOmC`0W in this image.", "objects": [{"patches": [149, 150, 167, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 239, 240], "bbox": [0.264125, 0.4180281690140845, 0.5558125, 0.6880046948356808], "iscrowd": 0, "area": 7326.071000000002, "rle": {"size": [426, 640], "counts": "ibV21U=3M3M2O2M3M3M3YOKXD6h;f0O1N2O1_DcNU;Y1nDjNQ;P1SESOl:h0WE[Oh:?]ECb:7bEK^:6_EM`:;UEIj:?kDET;Y1O1O1O100000000O10000000O1M3L3N3M3L4N2O1N2O1N2O1N2O2N11O001O1O010O001O1O0010O01O1O0010O01O1O001O010O000000000000000000000000001O000000O3N3M1OO010000O10000O10hDhN]:Y1\\EnNd:R1VETOj:k0QE[Oo:a1O0O10001O000O10001O000ON3L4L4L4J8G9L4M4M2N1N3N2M2O2N1M4M2M4M5JRf3LTZL1N3N2N2M3G9M3N101O0000010N1O1O001N2O1N1O2NVOmC`0W in this image.", "objects": [{"patches": [147, 148, 149, 170, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 244], "bbox": [0.400875, 0.4157276995305164, 0.691046875, 0.6921361502347418], "iscrowd": 0, "area": 9862.727850000005, "rle": {"size": [426, 640], "counts": "TT[33U=3L3N3L4N2N2N1O2N2M101O1O1O001O1N101O1O1O001L4L4M210000O1000000O11O2N3M2N2N2N2N2O2N1N2O001O00001O00000O2O00001XEXNk9d20O1O2N101N1O2N101N1O2N2O0O2N001O010O001O010O001RO[EUOe:k0lEeNT:[1m001O001O010O001O0001N1000000O101O00000O101O000O10001O00000O0100O100O010O10000O010O100O10O010000O10O0100O12N3M2N3M2N2N2N2N1O2N2N2N1O2N2N1O1OO100O100O100O100QOVE]Ol:`0XE^Oj:=[E@g:<]ECe:9_EDc:8bEF_:6eEH^:4fEJ[:l0PEQOS;o0lDPOU;P1lDmNW;S1iDkNX;U1`00O010O010OL5J5K5L5J5L5JXTb2"}, "label": "a full grown zebra facing the left"}]} {"id": 467760, "image": "COCO_train2014_000000467760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra with the shorter of tails\"."}], "task": "refering", "answer_template": "The \"zebra with the shorter of tails\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 149, 170, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 244], "bbox": [0.400875, 0.4157276995305164, 0.691046875, 0.6921361502347418], "iscrowd": 0, "area": 9862.727850000005, "rle": {"size": [426, 640], "counts": "TT[33U=3L3N3L4N2N2N1O2N2M101O1O1O001O1N101O1O1O001L4L4M210000O1000000O11O2N3M2N2N2N2N2O2N1N2O001O00001O00000O2O00001XEXNk9d20O1O2N101N1O2N101N1O2N2O0O2N001O010O001O010O001RO[EUOe:k0lEeNT:[1m001O001O010O001O0001N1000000O101O00000O101O000O10001O00000O0100O100O010O10000O010O100O10O010000O10O0100O12N3M2N3M2N2N2N2N1O2N2N2N1O2N2N1O1OO100O100O100O100QOVE]Ol:`0XE^Oj:=[E@g:<]ECe:9_EDc:8bEF_:6eEH^:4fEJ[:l0PEQOS;o0lDPOU;P1lDmNW;S1iDkNX;U1`00O010O010OL5J5K5L5J5L5JXTb2"}, "label": "zebra with the shorter of tails"}]} {"id": 115505, "image": "COCO_train2014_000000115505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young woman who is standing by a building and talking on a cellphone\"."}], "task": "refering", "answer_template": "The \"a young woman who is standing by a building and talking on a cellphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 164, 165, 166, 167, 187, 188, 189, 190, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 277, 278, 279, 280, 300, 301, 302, 323, 324, 325, 346, 347, 348, 349, 350, 351, 371, 372, 373], "bbox": [0.047296875, 0.3524791666666667, 0.31081250000000005, 0.989875], "iscrowd": 0, "area": 22141.394549999997, "rle": {"size": [480, 640], "counts": "_^>1l>6K4K6K4K6K4L5J5L5J5L4K3N2N3M2N2N3N1N2N3N1N3MM4jNgC]Oa in this image.", "objects": [{"patches": [141, 142, 143, 164, 165, 166, 167, 187, 188, 189, 190, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 277, 278, 279, 280, 300, 301, 302, 323, 324, 325, 346, 347, 348, 349, 350, 351, 371, 372, 373], "bbox": [0.047296875, 0.3524791666666667, 0.31081250000000005, 0.989875], "iscrowd": 0, "area": 22141.394549999997, "rle": {"size": [480, 640], "counts": "_^>1l>6K4K6K4K6K4L5J5L5J5L4K3N2N3M2N2N3N1N2N3N1N3MM4jNgC]Oa in this image.", "objects": [{"patches": [147, 148, 170, 171, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 241, 261, 262, 263, 264, 284, 285, 286, 287, 307, 308, 309, 310, 331, 332, 333, 354, 355, 356, 377, 378, 379], "bbox": [0.37584375, 0.3662916666666666, 0.505625, 0.9707916666666667], "iscrowd": 0, "area": 17539.717950000006, "rle": {"size": [480, 640], "counts": "UXa3l0\\=P1@?I8G9G9G8H9G9G9G9G9G9G9G8H9G9G9G9G5L0O1O1O1gNQI`JP7^5UI^Jl6`5XI]Ji6`5]I\\Jd6b5aIZJ`6c5eIZJ\\6d5iIXJX6f5mIVJT6g5QJVJP6h5UJTJl5i5ZJSJg5l5g100001O000000001O0000001O0000001O0000001O0000001O000PHPJP7P6lHVJR7j5iH^JT7c5fHdJX7\\5dHkJY7U5bHRK[7\\6H9G8H9G8H9G8Gh1YN?A?A?A`0@?A?A?@a0@hhc4"}, "label": "wrapped flowers with woman in pink top directly behind"}]} {"id": 115505, "image": "COCO_train2014_000000115505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a pink shirt and black pants\"."}], "task": "refering", "answer_template": "The \"a woman in a pink shirt and black pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 170, 171, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 241, 261, 262, 263, 264, 284, 285, 286, 287, 307, 308, 309, 310, 331, 332, 333, 354, 355, 356, 377, 378, 379], "bbox": [0.37584375, 0.3662916666666666, 0.505625, 0.9707916666666667], "iscrowd": 0, "area": 17539.717950000006, "rle": {"size": [480, 640], "counts": "UXa3l0\\=P1@?I8G9G9G8H9G9G9G9G9G9G9G8H9G9G9G9G5L0O1O1O1gNQI`JP7^5UI^Jl6`5XI]Ji6`5]I\\Jd6b5aIZJ`6c5eIZJ\\6d5iIXJX6f5mIVJT6g5QJVJP6h5UJTJl5i5ZJSJg5l5g100001O000000001O0000001O0000001O0000001O0000001O000PHPJP7P6lHVJR7j5iH^JT7c5fHdJX7\\5dHkJY7U5bHRK[7\\6H9G8H9G8H9G8Gh1YN?A?A?A`0@?A?A?@a0@hhc4"}, "label": "a woman in a pink shirt and black pants"}]} {"id": 115505, "image": "COCO_train2014_000000115505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in blue shirt and pants holding umbrella and flowers\"."}], "task": "refering", "answer_template": "The \"man in blue shirt and pants holding umbrella and flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 104, 105, 106, 127, 128, 129, 150, 151, 152, 153, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 293, 310, 311, 312, 313, 314, 315, 332, 333, 334, 335, 336, 337, 356, 357, 358, 359, 360, 361, 379, 380, 381, 382, 383, 384], "bbox": [0.435171875, 0.21864583333333334, 0.77703125, 0.9990416666666666], "iscrowd": 0, "area": 45064.47545000002, "rle": {"size": [480, 640], "counts": "\\US4165\\>NaA4\\>:03M5J2O1ON1O2N2L6BcA3^>McA2]>MdA2\\>NeA2[>NdA2\\>N in this image.", "objects": [{"patches": [81, 82, 104, 105, 106, 127, 128, 129, 150, 151, 152, 153, 173, 174, 175, 176, 177, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 293, 310, 311, 312, 313, 314, 315, 332, 333, 334, 335, 336, 337, 356, 357, 358, 359, 360, 361, 379, 380, 381, 382, 383, 384], "bbox": [0.435171875, 0.21864583333333334, 0.77703125, 0.9990416666666666], "iscrowd": 0, "area": 45064.47545000002, "rle": {"size": [480, 640], "counts": "\\US4165\\>NaA4\\>:03M5J2O1ON1O2N2L6BcA3^>McA2]>MdA2\\>NeA2[>NdA2\\>N in this image.", "objects": [{"patches": [22, 23, 24, 25, 26, 27, 28, 38, 39, 40, 41, 42, 43, 44, 45, 54, 55, 56, 57, 58, 59, 60, 61, 62, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 103, 104, 105, 106, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 170, 171, 172, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 195, 205, 206, 207, 208, 209, 210, 211, 223, 224, 225, 226, 227, 228, 240, 241, 242, 243, 244, 257, 258, 259, 260, 261, 274, 275, 276, 277, 293], "bbox": [0.0479375, 0.04284375, 0.6831458333333332, 0.7507187500000001], "iscrowd": 0, "area": 83748.57825, "rle": {"size": [640, 480], "counts": "Zc><_c0g0XOh0YO=B9G9H8G9G9H8G9G8I8G9G9H8J6L4M3L4L3M3M3M4L3M3M3M4L3M3M3M4L2N3M2N2N3M2N3N1N2N3M2N3M2N2N3M2N3M2N2N3M2N3gEZGS9h8bFfGY9\\8\\FSH^9\\9K6J5K4L101N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2O0O2O0O1O2O0O2O0O2N100O2O0O2N101N100O2N101N101N1O101N101N1O2O0O1O2O0O2O0O2N100OO2N1O101N1O2N1O1O2N1O2O0O1O2N1O2N1O1O2O0O2N1O1O2N1O2N100O2N1O2N1O1O2N101N1O1O2N1O2N1O1O2O0O2N1O1O2N1O2N100O2N1O1O2N1O3M2O1N2N2N3M2N2N2N2O1N3M2N2N2N2N3M2O1N2N2N3M2N2N2N2N2O2M2N2N2N2N3M2N2O1N2N3M2N2N2N2N2O2M2N2N2N2O2N1O1O1O1O2N1O1O1O1O2N100O1O1O1O2N1O1O1O1O2N1VNnBeLS=U3SCjLnC>A?B>Blln2"}, "label": "the rye sandwich half that doesn ' t have a bite taken out of it"}]} {"id": 74549, "image": "COCO_train2014_000000074549.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"left portion of sandwich closest to pickle\"."}], "task": "refering", "answer_template": "The \"left portion of sandwich closest to pickle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 23, 24, 25, 26, 27, 28, 38, 39, 40, 41, 42, 43, 44, 45, 54, 55, 56, 57, 58, 59, 60, 61, 62, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 103, 104, 105, 106, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 170, 171, 172, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 195, 205, 206, 207, 208, 209, 210, 211, 223, 224, 225, 226, 227, 228, 240, 241, 242, 243, 244, 257, 258, 259, 260, 261, 274, 275, 276, 277, 293], "bbox": [0.0479375, 0.04284375, 0.6831458333333332, 0.7507187500000001], "iscrowd": 0, "area": 83748.57825, "rle": {"size": [640, 480], "counts": "Zc><_c0g0XOh0YO=B9G9H8G9G9H8G9G8I8G9G9H8J6L4M3L4L3M3M3M4L3M3M3M4L3M3M3M4L2N3M2N2N3M2N3N1N2N3M2N3M2N2N3M2N3M2N2N3M2N3gEZGS9h8bFfGY9\\8\\FSH^9\\9K6J5K4L101N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2O0O2O0O1O2O0O2O0O2N100O2O0O2N101N100O2N101N101N1O101N101N1O2O0O1O2O0O2O0O2N100OO2N1O101N1O2N1O1O2N1O2O0O1O2N1O2N1O1O2O0O2N1O1O2N1O2N100O2N1O2N1O1O2N101N1O1O2N1O2N1O1O2O0O2N1O1O2N1O2N100O2N1O1O2N1O3M2O1N2N2N3M2N2N2N2O1N3M2N2N2N2N3M2O1N2N2N3M2N2N2N2N2O2M2N2N2N2N3M2N2O1N2N3M2N2N2N2N2O2M2N2N2N2O2N1O1O1O1O2N1O1O1O1O2N100O1O1O1O2N1O1O1O1O2N1VNnBeLS=U3SCjLnC>A?B>Blln2"}, "label": "left portion of sandwich closest to pickle"}]} {"id": 74549, "image": "COCO_train2014_000000074549.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a half of a sandwhich with a bite taken out\"."}], "task": "refering", "answer_template": "The \"a half of a sandwhich with a bite taken out\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 95, 96, 97, 98, 99, 100, 111, 112, 113, 114, 115, 116, 117, 118, 128, 129, 130, 131, 132, 133, 134, 135, 146, 147, 148, 149, 150, 151, 152, 162, 163, 164, 165, 166, 167, 168, 169, 178, 179, 180, 181, 182, 183, 184, 185, 186, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 250, 251, 252, 253, 254, 269, 270, 271], "bbox": [0.50225, 0.183859375, 1.0, 0.6950625], "iscrowd": 0, "area": 53110.776099999995, "rle": {"size": [640, 480], "counts": "olf43ic08I7J6L4L001N2O1O001N2O1O1j]OmN]a0S1c^OoN[a0R1c^OQOZa0Q1d^OROZa0o0d^OTOZa0l0f^OVOXa0k0f^OWOXa0k0f^OWOYa0j0f^OVOZa0j0e^OXOZa0i0d^OYO[a0h0c^OZO[a0h0d^OYO[a0h0c^OZO\\a0f0c^O\\O\\a0e0c^O\\O[a0f1O0O2N2O1N2\\MlMZCU2c in this image.", "objects": [{"patches": [79, 80, 81, 95, 96, 97, 98, 99, 100, 111, 112, 113, 114, 115, 116, 117, 118, 128, 129, 130, 131, 132, 133, 134, 135, 146, 147, 148, 149, 150, 151, 152, 162, 163, 164, 165, 166, 167, 168, 169, 178, 179, 180, 181, 182, 183, 184, 185, 186, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 250, 251, 252, 253, 254, 269, 270, 271], "bbox": [0.50225, 0.183859375, 1.0, 0.6950625], "iscrowd": 0, "area": 53110.776099999995, "rle": {"size": [640, 480], "counts": "olf43ic08I7J6L4L001N2O1O001N2O1O1j]OmN]a0S1c^OoN[a0R1c^OQOZa0Q1d^OROZa0o0d^OTOZa0l0f^OVOXa0k0f^OWOXa0k0f^OWOYa0j0f^OVOZa0j0e^OXOZa0i0d^OYO[a0h0c^OZO[a0h0d^OYO[a0h0c^OZO\\a0f0c^O\\O\\a0e0c^O\\O[a0f1O0O2N2O1N2\\MlMZCU2c in this image.", "objects": [{"patches": [5, 6, 7, 21, 22, 23, 37, 38, 86, 102, 103, 119], "bbox": [0.005958333333333333, 0.002453125, 0.4653333333333334, 0.329078125], "iscrowd": 0, "area": 3627.8589, "rle": {"size": [640, 480], "counts": "[R21mc04M3M2N2N1O000O10001O0000001N100000001O000O101O000000001N1000001O00000O01N2N2N2N2O1N2N1O2O1N2N2NdQQ1O_nnN1O1N101O1O1O1O1N101O1O1O1N2O001O001O0O2O001O001N101O001O0O20O01O010O001O010O0010O01O010O0010O01O010O0010O01O0010O01O010O0010O01O010O0010O01O1O010O1O10O10000O1000000O10000O1000000O1000000O100N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1N2O1OocP5"}, "label": "a brown and yellow wrapper"}]} {"id": 74549, "image": "COCO_train2014_000000074549.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the langer wrapper in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the langer wrapper in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 21, 22, 23, 37, 38, 86, 102, 103, 119], "bbox": [0.005958333333333333, 0.002453125, 0.4653333333333334, 0.329078125], "iscrowd": 0, "area": 3627.8589, "rle": {"size": [640, 480], "counts": "[R21mc04M3M2N2N1O000O10001O0000001N100000001O000O101O000000001N1000001O00000O01N2N2N2N2O1N2N1O2O1N2N2NdQQ1O_nnN1O1N101O1O1O1O1N101O1O1O1N2O001O001O0O2O001O001N101O001O0O20O01O010O001O010O0010O01O010O0010O01O010O0010O01O0010O01O010O0010O01O010O0010O01O1O010O1O10O10000O1000000O10000O1000000O1000000O100N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1N2O1OocP5"}, "label": "the langer wrapper in the right hand picture"}]} {"id": 66376, "image": "COCO_train2014_000000066376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a hood and mask has his hands in front of his face on a laptop video\"."}], "task": "refering", "answer_template": "The \"a man in a hood and mask has his hands in front of his face on a laptop video\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 200, 201, 202, 223, 224, 225, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272], "bbox": [0.649109375, 0.45316666666666666, 0.857625, 0.7087916666666667], "iscrowd": 0, "area": 8112.880099999999, "rle": {"size": [480, 640], "counts": "ZkR61n>1N2O1N2O1N2O1N2O1N2N2O1N3N1N2O1N2O1N2O1N2O1N2O10000O2O00000000000O1000000O1N2dNeNQE]1n:hNmDZ1R;iNiDY1V;kNeDW1Z;mNbDT1];QO]DQ1c;ROXDP1g;U1O1O1O1O1O1N2N2O2M2O1N2O1N200O10000000000O101O00000O100000000O100000001O01O2N2N2N2N2N1RN^Dd;^O_Da0c;[O`Dd0b;XOaDg0`;VOcDi0^;UOeDj0\\;ROgDm0Z;POiDo0X;nNkDQ1V;lNmDS1U;jNmDU1T;hNPEV1Q;gNQEY1P;dNSE[1Y<01O1O001O010O001O001O1O00?A7I1O1O1O2N100O1O1O2N1O1O1O2N]iZ1"}, "label": "a man in a hood and mask has his hands in front of his face on a laptop video"}]} {"id": 66376, "image": "COCO_train2014_000000066376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a jacket on a screen\"."}], "task": "refering", "answer_template": "The \"a person in a jacket on a screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 200, 201, 202, 223, 224, 225, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272], "bbox": [0.649109375, 0.45316666666666666, 0.857625, 0.7087916666666667], "iscrowd": 0, "area": 8112.880099999999, "rle": {"size": [480, 640], "counts": "ZkR61n>1N2O1N2O1N2O1N2O1N2N2O1N3N1N2O1N2O1N2O1N2O1N2O10000O2O00000000000O1000000O1N2dNeNQE]1n:hNmDZ1R;iNiDY1V;kNeDW1Z;mNbDT1];QO]DQ1c;ROXDP1g;U1O1O1O1O1O1N2N2O2M2O1N2O1N200O10000000000O101O00000O100000000O100000001O01O2N2N2N2N2N1RN^Dd;^O_Da0c;[O`Dd0b;XOaDg0`;VOcDi0^;UOeDj0\\;ROgDm0Z;POiDo0X;nNkDQ1V;lNmDS1U;jNmDU1T;hNPEV1Q;gNQEY1P;dNSE[1Y<01O1O001O010O001O001O1O00?A7I1O1O1O2N100O1O1O2N1O1O1O2N]iZ1"}, "label": "a person in a jacket on a screen"}]} {"id": 66376, "image": "COCO_train2014_000000066376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with a white hoodie looking at a computer screen\"."}], "task": "refering", "answer_template": "The \"a person with a white hoodie looking at a computer screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 53, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 234, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 327, 328, 329, 330, 331, 332, 351, 352, 353, 354, 355, 356, 374, 375, 376, 377, 378, 379], "bbox": [0.0016875000000000002, 0.024770833333333336, 0.71284375, 0.9864791666666667], "iscrowd": 0, "area": 112961.65915, "rle": {"size": [480, 640], "counts": "no01P?0O1iFb0k2i5001O001O001N101O0cK\\G^2d8^MeG]2\\8^MmG]2S8_MVH\\2k7_M^H\\2b7aMfHZ2[7aMjH^2V7^MoH`2S7[MRId2n6XMWIg2j6UMZIj2f6RM_Im2b6nLcIQ3]6kLhIT3S9O1O1N01F:F:F:E;jNV1N2N2N2O0O2N2N2N2O1N2N2O10O01O1O100O1O100O1O10O01O1O100O1O100O1O1O010O1O100O1O1N2O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1N2O010O10000O10000O10O10O2O001N101O000O2O001N101O0O3N2N1N3N1O2M2O2N1O2N2N101N1O2N101N4L5K4M3L4L5K4M3L4L4M4L3N2M3M4L3M3M3M4L3M3M3M3M4L3M3M3M4L3M3M3M4L3M3M3M3M4L3M3M3M4L3M3M00O2N1O1O1O1O2N1O1O2N1O1N3N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O00000000O1000000O1000000O1000000O1000000O1000000001O0000000000000000001O0000000000000000001O000000000000000000001O000000000TLUMcKk2]4WMaKi2^4ZM`Kf2`4\\M^Kd2a4_M]Kb2b4`M\\K`2c4cM[K]2d4fMZKZ2f4hMXKX2g4kMWKU2i4nMTKR2k4QNSKo1l4TNRKl1n4VNPKj1o4YNoJg1Q5[NmJf1Q5]NmJc1R5`NlJ`1T5bNjJ^1U5eNiJ\\1V5gNgJY1X5kNeJU1Z5oNcJQ1]5RO`Jo0^5UO_Jk0a5XO\\Jh0c5\\OZJe0d5@XJ`0h5CUJ=j5GSJ9m5JPJ7n5MoI3P61mIOS62lINS63mINR62nINQ63oIMP64PJMo53QJMn54RJLn54RJLm55SJLl54TJLk55UJKj56VJJj55WJLg55YJKg56XJJg58XJIf59YJGg5:XJFg5WJDg5>XJBg5a0WJ_Oi5b0VJ^Oi5d0VJ]Oh5e0WJ[Oh5g0`IXL7Q3X6Z1fIgNY6[1eIeNZ6^1dIbN[6`1dI`N[6c1cI^N[6e1cI[N\\6U500O2N1O1O101N1O1O2N1O100O1O1O2N100O1O1O1O2O0O1O1O1O1O2O0O1O1O1O1O2M2O1O1N2O1O2M2O1O1O1N2O2N1N2O1O1N2O2N1N2O1O1N2O1O2N1N2O1O1N2O2N1N3N1O2M2O2N1N2O2N1O2M2O2N1O2M2O2N1O2M2O2N1O2M2O2N1O1N3N1O2N1N3N1O2N1N3N1O2N1N3N1O2M3L4L5K4L4K5L4L4L4L4K5LaSf2"}, "label": "a person with a white hoodie looking at a computer screen"}]} {"id": 66376, "image": "COCO_train2014_000000066376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man leaning back with hoodie on\"."}], "task": "refering", "answer_template": "The \"a man leaning back with hoodie on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 53, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 234, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 327, 328, 329, 330, 331, 332, 351, 352, 353, 354, 355, 356, 374, 375, 376, 377, 378, 379], "bbox": [0.0016875000000000002, 0.024770833333333336, 0.71284375, 0.9864791666666667], "iscrowd": 0, "area": 112961.65915, "rle": {"size": [480, 640], "counts": "no01P?0O1iFb0k2i5001O001O001N101O0cK\\G^2d8^MeG]2\\8^MmG]2S8_MVH\\2k7_M^H\\2b7aMfHZ2[7aMjH^2V7^MoH`2S7[MRId2n6XMWIg2j6UMZIj2f6RM_Im2b6nLcIQ3]6kLhIT3S9O1O1N01F:F:F:E;jNV1N2N2N2O0O2N2N2N2O1N2N2O10O01O1O100O1O100O1O10O01O1O100O1O100O1O1O010O1O100O1O1N2O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1N2O010O10000O10000O10O10O2O001N101O000O2O001N101O0O3N2N1N3N1O2M2O2N1O2N2N101N1O2N101N4L5K4M3L4L5K4M3L4L4M4L3N2M3M4L3M3M3M4L3M3M3M3M4L3M3M3M4L3M3M3M4L3M3M3M3M4L3M3M3M4L3M3M00O2N1O1O1O1O2N1O1O2N1O1N3N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O00000000O1000000O1000000O1000000O1000000O1000000001O0000000000000000001O0000000000000000001O000000000000000000001O000000000TLUMcKk2]4WMaKi2^4ZM`Kf2`4\\M^Kd2a4_M]Kb2b4`M\\K`2c4cM[K]2d4fMZKZ2f4hMXKX2g4kMWKU2i4nMTKR2k4QNSKo1l4TNRKl1n4VNPKj1o4YNoJg1Q5[NmJf1Q5]NmJc1R5`NlJ`1T5bNjJ^1U5eNiJ\\1V5gNgJY1X5kNeJU1Z5oNcJQ1]5RO`Jo0^5UO_Jk0a5XO\\Jh0c5\\OZJe0d5@XJ`0h5CUJ=j5GSJ9m5JPJ7n5MoI3P61mIOS62lINS63mINR62nINQ63oIMP64PJMo53QJMn54RJLn54RJLm55SJLl54TJLk55UJKj56VJJj55WJLg55YJKg56XJJg58XJIf59YJGg5:XJFg5WJDg5>XJBg5a0WJ_Oi5b0VJ^Oi5d0VJ]Oh5e0WJ[Oh5g0`IXL7Q3X6Z1fIgNY6[1eIeNZ6^1dIbN[6`1dI`N[6c1cI^N[6e1cI[N\\6U500O2N1O1O101N1O1O2N1O100O1O1O2N100O1O1O1O2O0O1O1O1O1O2O0O1O1O1O1O2M2O1O1N2O1O2M2O1O1O1N2O2N1N2O1O1N2O2N1N2O1O1N2O1O2N1N2O1O1N2O2N1N3N1O2M2O2N1N2O2N1O2M2O2N1O2M2O2N1O2M2O2N1O2M2O2N1O1N3N1O2N1N3N1O2N1N3N1O2N1N3N1O2M3L4L5K4L4K5L4L4L4L4K5LaSf2"}, "label": "a man leaning back with hoodie on"}]} {"id": 475980, "image": "COCO_train2014_000000475980.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"overstuffed leather tan sofa against the wall with the american flag on it\"."}], "task": "refering", "answer_template": "The \"overstuffed leather tan sofa against the wall with the american flag on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 87, 88, 89, 90, 100, 101, 102, 103, 113, 114, 115, 116, 125, 126, 127, 128, 129, 138, 139, 140, 141, 142, 152, 153, 154, 155, 167, 168, 181], "bbox": [0.6795733333333334, 0.32688, 1.0, 0.73764], "iscrowd": 0, "area": 18048.73585, "rle": {"size": [500, 375], "counts": "_hl3g0m>3L5L3nLPOZGS1c8SOWGP1f8VOSGk0m8UOSGl0l8SOUGm0l8ROTGo0k8QOTGP1l8oNUGR1j8nNUGS1k8lNUGV1j8kNSGW1m8jNQGX1n8iNPGX1P9iNlF[1S9fNgF_1Y9bNaFd1^9]N\\Fh1d9UO[EP1d:PO\\EP1d:PO\\EQ1c:PO\\EP1d:PO\\EQ1c:PO\\EP1d:PO]EP1b:QO]Eo0c:QO]EP1b:QO]Eo0c:QO]EP1b:RO\\En0d:SO[Em0e:TOZEm0e:TOZEl0f:UOYEl0f:UOYEk0g:VOYEj0f:WOYEi0g:XOXEi0g:XOXEh0h:YOWEh0h:YOWEg0i:ZOVEf0j:S2O001O001O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00O1000000000001O000O1000O1000000000nNR1`L[N"}, "label": "overstuffed leather tan sofa against the wall with the american flag on it"}]} {"id": 475980, "image": "COCO_train2014_000000475980.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch on the wall in front of the american flag\"."}], "task": "refering", "answer_template": "The \"the couch on the wall in front of the american flag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 87, 88, 89, 90, 100, 101, 102, 103, 113, 114, 115, 116, 125, 126, 127, 128, 129, 138, 139, 140, 141, 142, 152, 153, 154, 155, 167, 168, 181], "bbox": [0.6795733333333334, 0.32688, 1.0, 0.73764], "iscrowd": 0, "area": 18048.73585, "rle": {"size": [500, 375], "counts": "_hl3g0m>3L5L3nLPOZGS1c8SOWGP1f8VOSGk0m8UOSGl0l8SOUGm0l8ROTGo0k8QOTGP1l8oNUGR1j8nNUGS1k8lNUGV1j8kNSGW1m8jNQGX1n8iNPGX1P9iNlF[1S9fNgF_1Y9bNaFd1^9]N\\Fh1d9UO[EP1d:PO\\EP1d:PO\\EQ1c:PO\\EP1d:PO\\EQ1c:PO\\EP1d:PO]EP1b:QO]Eo0c:QO]EP1b:QO]Eo0c:QO]EP1b:RO\\En0d:SO[Em0e:TOZEm0e:TOZEl0f:UOYEl0f:UOYEk0g:VOYEj0f:WOYEi0g:XOXEi0g:XOXEh0h:YOWEh0h:YOWEg0i:ZOVEf0j:S2O001O001O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00O1000000000001O000O1000O1000000000nNR1`L[N"}, "label": "the couch on the wall in front of the american flag"}]} {"id": 475980, "image": "COCO_train2014_000000475980.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"burlap fabric to the left of cat\"."}], "task": "refering", "answer_template": "The \"burlap fabric to the left of cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 117, 118, 119, 120, 122, 123, 124, 130, 131, 132, 133, 134, 135, 136, 137, 143, 144, 145, 146, 147, 148, 149, 150, 156, 157, 158, 159, 160, 161, 162, 163, 169, 170, 171, 172, 173, 174, 175, 176, 177, 182, 183, 184, 185, 186, 187, 188, 189, 190, 195, 196, 197, 198, 199, 200, 201, 202, 203, 208, 209, 210, 211, 212, 213, 214, 215, 216, 221, 222, 223, 224, 225, 226, 227, 228, 229], "bbox": [0.0, 0.45376, 0.69176, 0.9984], "iscrowd": 0, "area": 58059.8041, "rle": {"size": [500, 375], "counts": "a8Q7W7\\1000000000000001O000000000000001O00000000000000001O00000000000000001O000000000000000000000000O100000000O10000000000O10000000000O100000000O10000000000O10000000000001O0000001O0000001O00000SIjH\\5W7bJjH^5V7aJkH_5U7`JlH`5T7^JoHa5R7]JoHc5Q7\\JPId5P7[JRId5n6[JSIe5n6XJTIh5l6WJUIi5k6VJWIi5i6VJXIj5i6TJXIl5h6SJZIl5f6RJ\\In5e6PJ\\IP6d6oI]IQ6c6nI_IQ6a6nI`IR6a6kIaIU6_6jIcIU6]6jIdIV6\\6iIeIW6\\6gIeIY6[6eIhIZ6X6eIiI[6W6dIjI\\6W6bIjI^6V6aIkI_6U6_ImIa6S6^ImIc6b7000000001O0000000O1000000000000000000O1000000000000O1O100O1O1O100O1O10000O10000O10000O1000000O10000O1000000O10000000000O10000000000O1000000000000O10000000000O10000000000O1000000001O6J5K5K6J5K6J5K6J5K5K6J5K6J5K6J5K5K6J5K6J5K6J5K5K6J5K6J5K6J5K5K6J5K6J5K6J5K5K6J5K6J5KRUh1"}, "label": "burlap fabric to the left of cat"}]} {"id": 475980, "image": "COCO_train2014_000000475980.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown sofa holds a cat\"."}], "task": "refering", "answer_template": "The \"a brown sofa holds a cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 117, 118, 119, 120, 122, 123, 124, 130, 131, 132, 133, 134, 135, 136, 137, 143, 144, 145, 146, 147, 148, 149, 150, 156, 157, 158, 159, 160, 161, 162, 163, 169, 170, 171, 172, 173, 174, 175, 176, 177, 182, 183, 184, 185, 186, 187, 188, 189, 190, 195, 196, 197, 198, 199, 200, 201, 202, 203, 208, 209, 210, 211, 212, 213, 214, 215, 216, 221, 222, 223, 224, 225, 226, 227, 228, 229], "bbox": [0.0, 0.45376, 0.69176, 0.9984], "iscrowd": 0, "area": 58059.8041, "rle": {"size": [500, 375], "counts": "a8Q7W7\\1000000000000001O000000000000001O00000000000000001O00000000000000001O000000000000000000000000O100000000O10000000000O10000000000O100000000O10000000000O10000000000001O0000001O0000001O00000SIjH\\5W7bJjH^5V7aJkH_5U7`JlH`5T7^JoHa5R7]JoHc5Q7\\JPId5P7[JRId5n6[JSIe5n6XJTIh5l6WJUIi5k6VJWIi5i6VJXIj5i6TJXIl5h6SJZIl5f6RJ\\In5e6PJ\\IP6d6oI]IQ6c6nI_IQ6a6nI`IR6a6kIaIU6_6jIcIU6]6jIdIV6\\6iIeIW6\\6gIeIY6[6eIhIZ6X6eIiI[6W6dIjI\\6W6bIjI^6V6aIkI_6U6_ImIa6S6^ImIc6b7000000001O0000000O1000000000000000000O1000000000000O1O100O1O1O100O1O10000O10000O10000O1000000O10000O1000000O10000000000O10000000000O1000000000000O10000000000O10000000000O1000000001O6J5K5K6J5K6J5K6J5K5K6J5K6J5K6J5K5K6J5K6J5K6J5K5K6J5K6J5K6J5K5K6J5K6J5K6J5K5K6J5K6J5KRUh1"}, "label": "a brown sofa holds a cat"}]} {"id": 74577, "image": "COCO_train2014_000000074577.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"donuts\"."}], "task": "refering", "answer_template": "The \"donuts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [291, 293, 294, 295, 304, 305, 306, 307, 308, 309, 310, 319, 320, 321, 322, 323, 324, 325, 334, 336, 337, 338, 339, 340], "bbox": [0.26333333333333336, 0.831453125, 0.7258450704225353, 0.9887656249999999], "iscrowd": 0, "area": 10047.838349999995, "rle": {"size": [640, 426], "counts": "obV23dc09G9G:G80001N100O2O000O101O0O101N10000O2O000O2O0O0100O010O010O10O0100O010O010O10O0100O010O010O1N1N3M3M2N3M2N3O10O1001O000000010O01O001O010O1O001O010O0`]OSOma0n0m]O^Ola0b0o]OEoa0X1O1N1O2N2N2O1N0001N2N101N2N2N2O0O2N2O1N1O2O1N2N2N101N2N2O1N1O2N2G9F9H9GZc02j\\Oc0^O3L3N2M3N2M3N2N2M3N2N2O3M6J6J6J6J1O1O1O1O1O1O1O1O1O1O1O1O1O1O00000000000000iNn^O@Sa0=P_OAQa0Ua0^Ol^Ob0Wa0ZOi^Of0[a0UOg^Ok0]a0POc^OP1aa0lN`^OR1Wb0O2N100O2N1O2N102M3M3M3M3N2M3M3MkaX2"}, "label": "donuts"}]} {"id": 435029, "image": "COCO_train2014_000000435029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a green shirt is holding a frisbee\"."}], "task": "refering", "answer_template": "The \"a woman in a green shirt is holding a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 80, 81, 82, 103, 104, 105, 106, 126, 127, 128, 129, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 240, 241, 243, 244, 245, 263, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 333, 334, 335, 336, 337], "bbox": [0.46453125, 0.1891355140186916, 0.7331875, 0.9955841121495328], "iscrowd": 0, "area": 33517.069399999986, "rle": {"size": [428, 640], "counts": "X^l31V=8I7H7LI8O02N2N2mL_OYI:WMH]94eGH:i21fJF=L_1S1]3@aJE?IT1b1k3TO]JF>F\\1_1i3:gJXNa1^1g3>bJVNh1\\1e3a0_JTNm1[1c1oMZN^34cN_1aNmMl2e0cN\\1gNkMe2j0dNX1mNjM_2P1cNT1TOhMX2U1dNP1ZOgMR2Z1dNm0_OfMl1^1eNi0EeMf1c1eNf0JdM`1g1fNb0NeM]1j1dN?2fMZ1k1dN=4gMX1m1dN<5fMW1n1dN;7eMV1P2cN:8eMV1Q2bN:9dMU1R2bN9;cMT1U2`N7`0aMP1X2`N7c0]Mn0\\2_N6g0[Mj0_2_N5X4LhK4X4KiK4X4KiK5W4JjK5W4JjK6V4IkK7U4HlK9S4GmK9S4FnK:Q4FPL:Q4DPL=o3BRL>n3ASL?m3@TL`0m3^OTLb0l3]OULd0j3[OWLe0i3ZOXLf0i3XOXLh0h3WOYLj0f3UO[Lk0f3SO[Lm0e3RO\\Ln0d3TNjI0c2l1c3QNPJO^2Q2b3mMUJNZ2U2a3jM[JMU2Y2`3hM]JNT2Z2_3eMaJOQ2]2^3]MgJ5l1^2b4aM_K_2b4_M_Ka2a4^M`Kc2`4[MaKh2\\4WMeKk2Y4TMhKn2W4PMjKR3W72N2N3M2N2N2N2N2N3M2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N3M4L3M3TH[Kk6h4oH\\Ko6i4jHZKV7i4dH[K[7h4_H\\K`7X501O1O001O1O1O001O1O0lKRJj0o5UOTJh0m5WOVJe0k5ZOYJa0j5^OYJ>i5AZJ:h5E]J6e5I^J3d5L_JOd5O`JMa53dJVOo5i0XJYN`6e1gI]MQ7c2m1N2N3N2M2O2M3M2O2M3N1N3M3N2M2O2M3M2O2M3N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3N2M3MmVW2"}, "label": "a woman in a green shirt is holding a frisbee"}]} {"id": 435029, "image": "COCO_train2014_000000435029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a green t - shirt holding a white frisbee\"."}], "task": "refering", "answer_template": "The \"a woman in a green t - shirt holding a white frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 80, 81, 82, 103, 104, 105, 106, 126, 127, 128, 129, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 240, 241, 243, 244, 245, 263, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 333, 334, 335, 336, 337], "bbox": [0.46453125, 0.1891355140186916, 0.7331875, 0.9955841121495328], "iscrowd": 0, "area": 33517.069399999986, "rle": {"size": [428, 640], "counts": "X^l31V=8I7H7LI8O02N2N2mL_OYI:WMH]94eGH:i21fJF=L_1S1]3@aJE?IT1b1k3TO]JF>F\\1_1i3:gJXNa1^1g3>bJVNh1\\1e3a0_JTNm1[1c1oMZN^34cN_1aNmMl2e0cN\\1gNkMe2j0dNX1mNjM_2P1cNT1TOhMX2U1dNP1ZOgMR2Z1dNm0_OfMl1^1eNi0EeMf1c1eNf0JdM`1g1fNb0NeM]1j1dN?2fMZ1k1dN=4gMX1m1dN<5fMW1n1dN;7eMV1P2cN:8eMV1Q2bN:9dMU1R2bN9;cMT1U2`N7`0aMP1X2`N7c0]Mn0\\2_N6g0[Mj0_2_N5X4LhK4X4KiK4X4KiK5W4JjK5W4JjK6V4IkK7U4HlK9S4GmK9S4FnK:Q4FPL:Q4DPL=o3BRL>n3ASL?m3@TL`0m3^OTLb0l3]OULd0j3[OWLe0i3ZOXLf0i3XOXLh0h3WOYLj0f3UO[Lk0f3SO[Lm0e3RO\\Ln0d3TNjI0c2l1c3QNPJO^2Q2b3mMUJNZ2U2a3jM[JMU2Y2`3hM]JNT2Z2_3eMaJOQ2]2^3]MgJ5l1^2b4aM_K_2b4_M_Ka2a4^M`Kc2`4[MaKh2\\4WMeKk2Y4TMhKn2W4PMjKR3W72N2N3M2N2N2N2N2N3M2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N3M4L3M3TH[Kk6h4oH\\Ko6i4jHZKV7i4dH[K[7h4_H\\K`7X501O1O001O1O1O001O1O0lKRJj0o5UOTJh0m5WOVJe0k5ZOYJa0j5^OYJ>i5AZJ:h5E]J6e5I^J3d5L_JOd5O`JMa53dJVOo5i0XJYN`6e1gI]MQ7c2m1N2N3N2M2O2M3M2O2M3N1N3M3N2M2O2M3M2O2M3N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3N2M3MmVW2"}, "label": "a woman in a green t - shirt holding a white frisbee"}]} {"id": 435029, "image": "COCO_train2014_000000435029.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with black shorts\"."}], "task": "refering", "answer_template": "The \"man with black shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 75, 76, 77, 78, 99, 100, 101, 121, 122, 123, 124, 143, 144, 145, 146, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 280, 281, 282, 283, 284, 303, 304, 305, 306, 307, 327, 328, 329, 330], "bbox": [0.17054687500000001, 0.19261682242990655, 0.47575, 0.9867990654205608], "iscrowd": 0, "area": 37129.78305, "rle": {"size": [428, 640], "counts": "Rl]19] in this image.", "objects": [{"patches": [54, 75, 76, 77, 78, 99, 100, 101, 121, 122, 123, 124, 143, 144, 145, 146, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 280, 281, 282, 283, 284, 303, 304, 305, 306, 307, 327, 328, 329, 330], "bbox": [0.17054687500000001, 0.19261682242990655, 0.47575, 0.9867990654205608], "iscrowd": 0, "area": 37129.78305, "rle": {"size": [428, 640], "counts": "Rl]19] in this image.", "objects": [{"patches": [79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 137, 157, 158, 159, 160], "bbox": [0.44662499999999994, 0.18822709163346613, 0.9763124999999999, 0.36235059760956173], "iscrowd": 0, "area": 16203.19225, "rle": {"size": [502, 640], "counts": "dY\\48^?3M2N3M2N3M2N2N001O001O001O001O001O001O1O2N1O1O1O2N1O1O1O1O001O0000000000000000001O00O100000000000000O101O000000000O100000000000000O100000000000000O1000000000O100000000000000000000000000000000000000000000000000000000O1000000000000000000000000000001O0000001O00001O0000001O00001N10000010O0001O0000001O00001O000001O01O00000000000000001O000001O0000000001O00000000000001O01O00000000000000001O0001O000001O001O00001O00010O00001O010O00001O00010O00001O01O01O001O00010O00001O01O01O001O01O01O00001O01O01O001O01O01O000010O0001O001O01O01O000010O0001O00010O001O000010O0001O00010O001O000001O001O001O001O1O001O001O001O001O001O001O1O1O001O1O1O001O1O0O2O1N2O0O2O1N2O0O2N2O1N101N2O0O2O1M3L4K8I7HcV7"}, "label": "the longer of the two carrots"}]} {"id": 426838, "image": "COCO_train2014_000000426838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a carrot\"."}], "task": "refering", "answer_template": "The \"a carrot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 137, 157, 158, 159, 160], "bbox": [0.44662499999999994, 0.18822709163346613, 0.9763124999999999, 0.36235059760956173], "iscrowd": 0, "area": 16203.19225, "rle": {"size": [502, 640], "counts": "dY\\48^?3M2N3M2N3M2N2N001O001O001O001O001O001O1O2N1O1O1O2N1O1O1O1O001O0000000000000000001O00O100000000000000O101O000000000O100000000000000O100000000000000O1000000000O100000000000000000000000000000000000000000000000000000000O1000000000000000000000000000001O0000001O00001O0000001O00001N10000010O0001O0000001O00001O000001O01O00000000000000001O000001O0000000001O00000000000001O01O00000000000000001O0001O000001O001O00001O00010O00001O010O00001O00010O00001O01O01O001O00010O00001O01O01O001O01O01O00001O01O01O001O01O01O000010O0001O001O01O01O000010O0001O00010O001O000010O0001O00010O001O000001O001O001O001O1O001O001O001O001O001O001O1O1O001O1O1O001O1O0O2O1N2O0O2O1N2O0O2N2O1N101N2O0O2O1M3L4K8I7HcV7"}, "label": "a carrot"}]} {"id": 426838, "image": "COCO_train2014_000000426838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"carrot touching potatoes\"."}], "task": "refering", "answer_template": "The \"carrot touching potatoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 126, 127, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 227, 228], "bbox": [0.49000000000000005, 0.2608764940239044, 0.972640625, 0.5193625498007969], "iscrowd": 0, "area": 18343.27945, "rle": {"size": [502, 640], "counts": "]Rj4:[?2O1O1O001N2O1O1O001N2O1O1O1N101O1O1O1N2O001O0N201N101O0O101N101N100O2O0O101N101O00001O001O000010O0001O001O00001O001O00001O00001O010O00001O001O00001O001O00001O00010O0010O00010O01O01O01O01O01O010O00010O0010O00010O00010O00010O01O01O01O01O01O01O01O01O01O01O01O010O00010O00010O00010O00010O0010O00010O00010O00010O00010O01O00010O00010O00010O00010O00001O010O000010O0001O00010O00001O010O000010O0001O00010O000010O0001O0010O0000010O0001O00010O00001O01O01O00010O00000010O0001O01O01O000010O0001O01O01O0000010O000010O0001O00010O00001O01O01O0001O01O00010O01O0O2O001O0O2O001O001N101O001N101O00001N1O2L3N3L3N3L3N3L3N3L3N3L3N3J5J7IWd8"}, "label": "carrot touching potatoes"}]} {"id": 426838, "image": "COCO_train2014_000000426838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the carrot which is very next to potatoe\"."}], "task": "refering", "answer_template": "The \"the carrot which is very next to potatoe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 126, 127, 128, 129, 130, 131, 132, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 227, 228], "bbox": [0.49000000000000005, 0.2608764940239044, 0.972640625, 0.5193625498007969], "iscrowd": 0, "area": 18343.27945, "rle": {"size": [502, 640], "counts": "]Rj4:[?2O1O1O001N2O1O1O001N2O1O1O1N101O1O1O1N2O001O0N201N101O0O101N101N100O2O0O101N101O00001O001O000010O0001O001O00001O001O00001O00001O010O00001O001O00001O001O00001O00010O0010O00010O01O01O01O01O01O010O00010O0010O00010O00010O00010O01O01O01O01O01O01O01O01O01O01O01O010O00010O00010O00010O00010O0010O00010O00010O00010O00010O01O00010O00010O00010O00010O00001O010O000010O0001O00010O00001O010O000010O0001O00010O000010O0001O0010O0000010O0001O00010O00001O01O01O00010O00000010O0001O01O01O000010O0001O01O01O0000010O000010O0001O00010O00001O01O01O0001O01O00010O01O0O2O001O0O2O001O001N101O001N101O00001N1O2L3N3L3N3L3N3L3N3L3N3L3N3J5J7IWd8"}, "label": "the carrot which is very next to potatoe"}]} {"id": 402264, "image": "COCO_train2014_000000402264.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a balding man with a black shirt looking down\"."}], "task": "refering", "answer_template": "The \"a balding man with a black shirt looking down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 49, 50, 51, 72, 73, 74, 95, 96, 97, 98, 99, 116, 117, 118, 119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 207, 208, 209, 210], "bbox": [0.0196875, 0.0898876404494382, 0.39, 0.9258426966292135], "iscrowd": 0, "area": 24222.95999999999, "rle": {"size": [267, 640], "counts": "Rd33T84L4M3L4M3L4L4M3M4L3M3M3M3N2M3M3M3M3M3M3M3N2M3M3M3M3M3O1O100O1O100O1O100O00100O1O100O1O100O1O100O1O100O1O100O10000O10O0100O100O10000O100O100O100O10000O100O100O1aMRMAm2;^M_Ob2`0bM\\O_2c0dM[O\\2c0iMYOX2f0lMWOT2h0oMUOQ2j0TNSOi1o0[NmNi0XOoMn1[1hNd0CbMk1o1_N>T2FhM;W2HgM7Y2McM4]2NaM1_22^MOa25\\MJe28XMIg2m20100O1O10O101O00001O00001O0O101O00001O00001O001O00001O001O001O00001O002N1O2O0O2N3M2N3M3M2N3M3N2M2N00O11_LXL]2j3]McLZ2b3]MiL\\2c4001O01O00001O0000001O00001O0000001O0000001O001O001O000010O01O001O00001O001O001O00001O001O0010O0001O001O00001O001O002N2N3M2N2N2N2N2N2N2N2N2N3M2N2N2N2M3N2M3N2M3N2N3L3NR`U3"}, "label": "a balding man with a black shirt looking down"}]} {"id": 402264, "image": "COCO_train2014_000000402264.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black t - shirt looking down\"."}], "task": "refering", "answer_template": "The \"a man in a black t - shirt looking down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 49, 50, 51, 72, 73, 74, 95, 96, 97, 98, 99, 116, 117, 118, 119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 207, 208, 209, 210], "bbox": [0.0196875, 0.0898876404494382, 0.39, 0.9258426966292135], "iscrowd": 0, "area": 24222.95999999999, "rle": {"size": [267, 640], "counts": "Rd33T84L4M3L4M3L4L4M3M4L3M3M3M3N2M3M3M3M3M3M3M3N2M3M3M3M3M3O1O100O1O100O1O100O00100O1O100O1O100O1O100O1O100O1O100O10000O10O0100O100O10000O100O100O100O10000O100O100O1aMRMAm2;^M_Ob2`0bM\\O_2c0dM[O\\2c0iMYOX2f0lMWOT2h0oMUOQ2j0TNSOi1o0[NmNi0XOoMn1[1hNd0CbMk1o1_N>T2FhM;W2HgM7Y2McM4]2NaM1_22^MOa25\\MJe28XMIg2m20100O1O10O101O00001O00001O0O101O00001O00001O001O00001O001O001O00001O002N1O2O0O2N3M2N3M3M2N3M3N2M2N00O11_LXL]2j3]McLZ2b3]MiL\\2c4001O01O00001O0000001O00001O0000001O0000001O001O001O000010O01O001O00001O001O001O00001O001O0010O0001O001O00001O001O002N2N3M2N2N2N2N2N2N2N2N2N3M2N2N2N2M3N2M3N2M3N2N3L3NR`U3"}, "label": "a man in a black t - shirt looking down"}]} {"id": 402264, "image": "COCO_train2014_000000402264.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a buttoned up shirt\"."}], "task": "refering", "answer_template": "The \"a man in a buttoned up shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 34, 35, 36, 37, 38, 57, 58, 59, 60, 80, 81, 82, 83, 102, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224], "bbox": [0.381203125, 0.011423220973782771, 0.768703125, 0.9381273408239701], "iscrowd": 0, "area": 30500.37264999999, "rle": {"size": [267, 640], "counts": "Wjo11Y81N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N2N2O1O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O101N1O1O1O1N2O1O1TNXMSNj2i1]MSNd2j1cMPN_2l1hMPNY2l1oMnMS2o1SNmMn1o1ZNlMg1Q2_NjMc1R2eNiM\\1T2jNgMX1U2oNgMR1W2b2O100O100O10000O100O100O10000O100O100O10000O1000000001O00000000000000001O000000000000001O00000002VOhI]OY6`0iI@Y6 in this image.", "objects": [{"patches": [11, 12, 13, 14, 34, 35, 36, 37, 38, 57, 58, 59, 60, 80, 81, 82, 83, 102, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224], "bbox": [0.381203125, 0.011423220973782771, 0.768703125, 0.9381273408239701], "iscrowd": 0, "area": 30500.37264999999, "rle": {"size": [267, 640], "counts": "Wjo11Y81N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N2N2O1O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O101N1O1O1O1N2O1O1TNXMSNj2i1]MSNd2j1cMPN_2l1hMPNY2l1oMnMS2o1SNmMn1o1ZNlMg1Q2_NjMc1R2eNiM\\1T2jNgMX1U2oNgMR1W2b2O100O100O10000O100O100O10000O100O100O10000O1000000001O00000000000000001O000000000000001O00000002VOhI]OY6`0iI@Y6 in this image.", "objects": [{"patches": [65, 86, 87, 88, 89, 109, 110, 111, 112, 133, 134, 135, 136, 157, 158, 159, 179, 180, 181, 182, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229], "bbox": [0.745109375, 0.2905243445692884, 0.9696718750000001, 0.9504494382022471], "iscrowd": 0, "area": 13783.017850000004, "rle": {"size": [267, 640], "counts": "hgl31Y8101O1O001O0O2O001O001N2O001N1M4nKGM<3DJ`04BIa06_OId05\\OJg04YOKk02VOLm02SONo00QOOW1JiN5Y1JhN5Y1JgN5\\1IdN7]1GdN8^1EeN9]1EdN;^1AdN>^1_OdNa0^1[OeNc0]1ZOeNf0]1VOeNi0]1TOeNk0^1POeNP1\\1jNiNU1Z1fNiNZ1W1cNlN\\1U1`NoN_1R1^NPOc1P1ZNSOe1o0WNTOi1l0TNWOl1i0PN[OP2e0mM^OS2b0jMAV2`0eMD[2B5K4L4L5K4L2N1O1O1N2O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N3N3M3M3M3M2N3M3M3M3M3M3M2N3M3M3M3M3M8H7I8H8Hhf4"}, "label": "the man in the orange hat is on the right of the other two men"}]} {"id": 402264, "image": "COCO_train2014_000000402264.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in a black tshirt and red cap walking in the background\"."}], "task": "refering", "answer_template": "The \"man in a black tshirt and red cap walking in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 86, 87, 88, 89, 109, 110, 111, 112, 133, 134, 135, 136, 157, 158, 159, 179, 180, 181, 182, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229], "bbox": [0.745109375, 0.2905243445692884, 0.9696718750000001, 0.9504494382022471], "iscrowd": 0, "area": 13783.017850000004, "rle": {"size": [267, 640], "counts": "hgl31Y8101O1O001O0O2O001O001N2O001N1M4nKGM<3DJ`04BIa06_OId05\\OJg04YOKk02VOLm02SONo00QOOW1JiN5Y1JhN5Y1JgN5\\1IdN7]1GdN8^1EeN9]1EdN;^1AdN>^1_OdNa0^1[OeNc0]1ZOeNf0]1VOeNi0]1TOeNk0^1POeNP1\\1jNiNU1Z1fNiNZ1W1cNlN\\1U1`NoN_1R1^NPOc1P1ZNSOe1o0WNTOi1l0TNWOl1i0PN[OP2e0mM^OS2b0jMAV2`0eMD[2B5K4L4L5K4L2N1O1O1N2O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N3N3M3M3M3M2N3M3M3M3M3M3M2N3M3M3M3M3M8H7I8H8Hhf4"}, "label": "man in a black tshirt and red cap walking in the background"}]} {"id": 41818, "image": "COCO_train2014_000000041818.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holding a glass of orange juice and smiling at the camera with her hand on her hip\"."}], "task": "refering", "answer_template": "The \"a woman holding a glass of orange juice and smiling at the camera with her hand on her hip\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.5587187499999999, 0.23735416666666667, 1.0, 0.9867916666666667], "iscrowd": 0, "area": 67933.45425000002, "rle": {"size": [480, 640], "counts": "cTX5i0V>4L4K5L4L3M4K5L4K4L3N1N3M3N2M2O2M3L3L4oN`M^Ed2o9SNiEQ2V:TNbEP2]:Z1O1N3N1O100O10000O10cKlEn3T:PLQFm3n9SLTFl3l9SLUFm3k9QLWFo3h9QLZFn3f9QL[Fo3e9oK]FQ4b9oK`FP4X:000001O00001O0O[GULW6k3dIZL\\6e3aI_L`6`3[IeLe6Z3WIkLi6U3SIoLm6P3oHUMR7j2fH^MY7b2aHeM_7[2ZHlMe7T2THTNl7l1PHXNo7h1nG\\NP8f1lG^NS8b1gGeNX8R4O1N2O1O1N2H8G9O1O1O1O100O1[ObHgI_7k4dHiK2YO[7g4kHmKNYOX7b4SIRLHZOV7_4XITLb7k3`HRLb7m3`HQLa7n3aHQL_7n3cHPL^7o3dHPL\\7o3fHoK[7P4gHoKY7P4iHnKX7Q4jHnKV7Q4lHmKU7R4nHlKR7S4PIkKQ7T4RIjKn6U4UIhKl6W4WIgKi6X4P2O1O100O100O100O1O100O100O1O100O100O100O100O100O100O10000O100O10000O100001O0000001O0000001O0000001O0000001O00001O001O001O1O001O1O001O003M3M1O1O2N1O2N1bEcKW:h4TGWKW7o4aHVK\\7]4aGkKl0La7Q5XHSKe7Y5mGlJP8T6N3M1O2N2N2N2N2N2N0000001O0000001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O00001O001O00001O00001O001O00001O001O001O001O1O001O001O001O1O001O001UOPHdJP8[5THbJm7[5XHbJi7\\5\\H`Jd7_5`H^Ja7a5bH\\J_7b5eH[J[7e5Q1O1N2O1N2O1N2O1O1N102M3N2N2M3K5J[G"}, "label": "a woman holding a glass of orange juice and smiling at the camera with her hand on her hip"}]} {"id": 41818, "image": "COCO_train2014_000000041818.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady in black over coat holding a glass of juice and posing\"."}], "task": "refering", "answer_template": "The \"a lady in black over coat holding a glass of juice and posing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.5587187499999999, 0.23735416666666667, 1.0, 0.9867916666666667], "iscrowd": 0, "area": 67933.45425000002, "rle": {"size": [480, 640], "counts": "cTX5i0V>4L4K5L4L3M4K5L4K4L3N1N3M3N2M2O2M3L3L4oN`M^Ed2o9SNiEQ2V:TNbEP2]:Z1O1N3N1O100O10000O10cKlEn3T:PLQFm3n9SLTFl3l9SLUFm3k9QLWFo3h9QLZFn3f9QL[Fo3e9oK]FQ4b9oK`FP4X:000001O00001O0O[GULW6k3dIZL\\6e3aI_L`6`3[IeLe6Z3WIkLi6U3SIoLm6P3oHUMR7j2fH^MY7b2aHeM_7[2ZHlMe7T2THTNl7l1PHXNo7h1nG\\NP8f1lG^NS8b1gGeNX8R4O1N2O1O1N2H8G9O1O1O1O100O1[ObHgI_7k4dHiK2YO[7g4kHmKNYOX7b4SIRLHZOV7_4XITLb7k3`HRLb7m3`HQLa7n3aHQL_7n3cHPL^7o3dHPL\\7o3fHoK[7P4gHoKY7P4iHnKX7Q4jHnKV7Q4lHmKU7R4nHlKR7S4PIkKQ7T4RIjKn6U4UIhKl6W4WIgKi6X4P2O1O100O100O100O1O100O100O1O100O100O100O100O100O100O10000O100O10000O100001O0000001O0000001O0000001O0000001O00001O001O001O1O001O1O001O003M3M1O1O2N1O2N1bEcKW:h4TGWKW7o4aHVK\\7]4aGkKl0La7Q5XHSKe7Y5mGlJP8T6N3M1O2N2N2N2N2N2N0000001O0000001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O00001O001O00001O00001O001O00001O001O001O001O1O001O001O001O1O001O001UOPHdJP8[5THbJm7[5XHbJi7\\5\\H`Jd7_5`H^Ja7a5bH\\J_7b5eH[J[7e5Q1O1N2O1N2O1N2O1O1N102M3N2N2M3K5J[G"}, "label": "a lady in black over coat holding a glass of juice and posing"}]} {"id": 41818, "image": "COCO_train2014_000000041818.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a t - shirt with faces on it\"."}], "task": "refering", "answer_template": "The \"a man in a t - shirt with faces on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 73, 74, 75, 76, 96, 97, 98, 99, 120, 121, 122, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 285, 303, 304, 305, 306, 307, 308, 309, 325, 326, 327, 328, 329, 330, 331, 332, 348, 349, 350, 351, 352, 353, 354, 355, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.09628125, 0.11372916666666667, 0.46453125, 0.989875], "iscrowd": 0, "area": 61710.900349999996, "rle": {"size": [480, 640], "counts": "W[m06e>6J7J6I7J6cBnNc in this image.", "objects": [{"patches": [51, 52, 53, 73, 74, 75, 76, 96, 97, 98, 99, 120, 121, 122, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 284, 285, 303, 304, 305, 306, 307, 308, 309, 325, 326, 327, 328, 329, 330, 331, 332, 348, 349, 350, 351, 352, 353, 354, 355, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.09628125, 0.11372916666666667, 0.46453125, 0.989875], "iscrowd": 0, "area": 61710.900349999996, "rle": {"size": [480, 640], "counts": "W[m06e>6J7J6I7J6cBnNc in this image.", "objects": [{"patches": [183, 184, 197, 198, 199, 212, 213, 214, 226, 227, 228, 229, 241, 242, 243, 244, 256, 257, 258, 259, 271, 272, 273, 274, 286, 287, 288, 289, 302, 303], "bbox": [0.11030588235294118, 0.5455625000000001, 0.31839999999999996, 0.8974062500000001], "iscrowd": 0, "area": 15017.459949999999, "rle": {"size": [640, 425], "counts": "ajm0k0ob0X1iN>UOdMZ_Od2a`0a0K5K5K5K5K5K5L4K5K5K4L5K4L5L3M4L3M4L3L5L4L2N2N2N2N2N2N2N2O1O00100O1O100O00010O0001O010O000010O0001O010O2N1O100O2N1O100O2L3M3M3M3M4L3M3M3M4L3M6I8H8H8H8H8H8H8H8Da0]Oc0]Oc0]Oc0]OQkd5"}, "label": "a big clock face is shown from the side angle"}]} {"id": 369509, "image": "COCO_train2014_000000369509.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"clock face not facing camera\"."}], "task": "refering", "answer_template": "The \"clock face not facing camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [183, 184, 197, 198, 199, 212, 213, 214, 226, 227, 228, 229, 241, 242, 243, 244, 256, 257, 258, 259, 271, 272, 273, 274, 286, 287, 288, 289, 302, 303], "bbox": [0.11030588235294118, 0.5455625000000001, 0.31839999999999996, 0.8974062500000001], "iscrowd": 0, "area": 15017.459949999999, "rle": {"size": [640, 425], "counts": "ajm0k0ob0X1iN>UOdMZ_Od2a`0a0K5K5K5K5K5K5L4K5K5K4L5K4L5L3M4L3M4L3L5L4L2N2N2N2N2N2N2N2O1O00100O1O100O00010O0001O010O000010O0001O010O2N1O100O2N1O100O2L3M3M3M3M4L3M3M3M4L3M6I8H8H8H8H8H8H8H8Da0]Oc0]Oc0]Oc0]OQkd5"}, "label": "clock face not facing camera"}]} {"id": 369509, "image": "COCO_train2014_000000369509.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a clock that has the time 12 : 10 on it\"."}], "task": "refering", "answer_template": "The \"a clock that has the time 12 : 10 on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 190, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 231, 232, 233, 234, 235, 236, 237, 246, 247, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 277, 278, 279, 280, 281, 282, 293, 294, 295], "bbox": [0.4017882352941176, 0.540359375, 0.8407294117647058, 0.84978125], "iscrowd": 0, "area": 28481.187599999994, "rle": {"size": [640, 425], "counts": "WY[38cc0:E;E;G9I7I7J5J4L5K4L5K4L5K4M4K4L5L3M4K4M4L3L5L3M1O2N2N1O2O1N1O2O1N1O2O1N2N101N2N101N2N101N2N101N2N101N2O0O2O1O001O1O0O2O1O001O1O0O2O1O004L2N1O00000000001O000000O100000001O000O100000001O00000O10001O000000000O101O000000001N1000001O0000001N10001O0000001N10001O001O001N2O1N2O0O2O1N2O0O2O1N2O1N101N2O1O1N2O2M3N2M2O2M3N1N3N2M3N1N3N2M3M2M4M3L4M2M4M3L4M2M4M3L4M2M5L5J6K6E:B>B>CYRZ1"}, "label": "a clock that has the time 12 : 10 on it"}]} {"id": 369509, "image": "COCO_train2014_000000369509.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a clock on the watch tower showing 12 : 10 pm\"."}], "task": "refering", "answer_template": "The \"a clock on the watch tower showing 12 : 10 pm\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 190, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 231, 232, 233, 234, 235, 236, 237, 246, 247, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 277, 278, 279, 280, 281, 282, 293, 294, 295], "bbox": [0.4017882352941176, 0.540359375, 0.8407294117647058, 0.84978125], "iscrowd": 0, "area": 28481.187599999994, "rle": {"size": [640, 425], "counts": "WY[38cc0:E;E;G9I7I7J5J4L5K4L5K4L5K4M4K4L5L3M4K4M4L3L5L3M1O2N2N1O2O1N1O2O1N1O2O1N2N101N2N101N2N101N2N101N2N101N2O0O2O1O001O1O0O2O1O001O1O0O2O1O004L2N1O00000000001O000000O100000001O000O100000001O00000O10001O000000000O101O000000001N1000001O0000001N10001O0000001N10001O001O001N2O1N2O0O2O1N2O0O2O1N2O1N101N2O1O1N2O2M3N2M2O2M3N1N3N2M3N1N3N2M3M2M4M3L4M2M4M3L4M2M4M3L4M2M5L5J6K6E:B>B>CYRZ1"}, "label": "a clock on the watch tower showing 12 : 10 pm"}]} {"id": 508780, "image": "COCO_train2014_000000508780.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in the air with his arm out , the arm has tattoos , with a corner of a skateboard in it\"."}], "task": "refering", "answer_template": "The \"a person in the air with his arm out , the arm has tattoos , with a corner of a skateboard in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 24, 25, 26, 33, 34, 35, 36, 38, 39, 40, 41, 47, 48, 49, 50, 51, 52, 53, 54, 62, 63, 64, 65, 66, 67, 68, 69, 78, 79, 80, 81, 82, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 122, 123, 124, 125, 126, 127, 136, 137, 138, 139, 140, 141, 142, 151, 152, 153, 154, 155, 156, 157, 158, 166, 168, 169, 170, 171, 172, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202], "bbox": [0.11216470588235294, 0.0, 0.7920235294117648, 0.5931093749999999], "iscrowd": 0, "area": 44205.948749999996, "rle": {"size": [640, 425], "counts": "QYn0j0Rc04L3L5M200O2N1O101N1O2O0O1O2O00O010O001O10O100O1000O1000hKeNkE\\1T:kNfEU1X:SObEn0[:ZO_Ej0\\:_O\\Ed0_:@_Ed0[:@cEd0Y:^OdEg0Y:ZOdEk0Y:WOdEl0Y:VOdEo0Y:ROdEd1g9^NVFd1j9]NTFe1j9]NSFf1l9[NQFh1m9ZNPFh1P:YNnEi1P:YNmEj1R:WNkEk1U:WNgEl1W:VNfEm1Y:TNdEo1Z:SNcEo1]:RN[EV2c:lMiDh2V;YMgDi2X;YMfDi2Y;XMeDj2Y;YMeDh2Z;YMdDh2\\;YMbDi2\\;YMcDh2\\;YMbDh2a0fJP:d2]Eh29mJZ:Q8\\EYHb:c800O1000001O000000001O000000001O0O1000000000000000000001O000000001O0000001O000000001O00000000010O000000001O0001O0001O0000000010O0000000001O2N2mGYEj6h:SIYEm6j:PIVEP7l:mHUES7m:kHSEU7o:hHREX7P;fHPEZ7R;cHnD^7U;_HkDa7W;\\HjDc7Y;[HgDd7[;ZHfDe7[;[HeDd7\\;[HeDd7\\;[HeDd7\\;\\HdDc7];\\HdDc7];]HcDb7^;]HbDP7KYId;GaDm69PIV;3bDj6V in this image.", "objects": [{"patches": [10, 11, 24, 25, 26, 33, 34, 35, 36, 38, 39, 40, 41, 47, 48, 49, 50, 51, 52, 53, 54, 62, 63, 64, 65, 66, 67, 68, 69, 78, 79, 80, 81, 82, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 122, 123, 124, 125, 126, 127, 136, 137, 138, 139, 140, 141, 142, 151, 152, 153, 154, 155, 156, 157, 158, 166, 168, 169, 170, 171, 172, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202], "bbox": [0.11216470588235294, 0.0, 0.7920235294117648, 0.5931093749999999], "iscrowd": 0, "area": 44205.948749999996, "rle": {"size": [640, 425], "counts": "QYn0j0Rc04L3L5M200O2N1O101N1O2O0O1O2O00O010O001O10O100O1000O1000hKeNkE\\1T:kNfEU1X:SObEn0[:ZO_Ej0\\:_O\\Ed0_:@_Ed0[:@cEd0Y:^OdEg0Y:ZOdEk0Y:WOdEl0Y:VOdEo0Y:ROdEd1g9^NVFd1j9]NTFe1j9]NSFf1l9[NQFh1m9ZNPFh1P:YNnEi1P:YNmEj1R:WNkEk1U:WNgEl1W:VNfEm1Y:TNdEo1Z:SNcEo1]:RN[EV2c:lMiDh2V;YMgDi2X;YMfDi2Y;XMeDj2Y;YMeDh2Z;YMdDh2\\;YMbDi2\\;YMcDh2\\;YMbDh2a0fJP:d2]Eh29mJZ:Q8\\EYHb:c800O1000001O000000001O000000001O0O1000000000000000000001O000000001O0000001O000000001O00000000010O000000001O0001O0001O0000000010O0000000001O2N2mGYEj6h:SIYEm6j:PIVEP7l:mHUES7m:kHSEU7o:hHREX7P;fHPEZ7R;cHnD^7U;_HkDa7W;\\HjDc7Y;[HgDd7[;ZHfDe7[;[HeDd7\\;[HeDd7\\;[HeDd7\\;\\HdDc7];\\HdDc7];]HcDb7^;]HbDP7KYId;GaDm69PIV;3bDj6V in this image.", "objects": [{"patches": [208, 209, 210, 211, 232, 233, 234, 235, 257, 258, 259, 279, 280, 281, 282, 305, 306], "bbox": [0.07678125, 0.5956674473067916, 0.330671875, 0.9188056206088993], "iscrowd": 0, "area": 5107.596399999998, "rle": {"size": [427, 640], "counts": "Wfd01W=6J6M2N3N2M2O2M100O100O10O0100001O1N1C^C0cQDBo;?oCBP in this image.", "objects": [{"patches": [208, 209, 210, 211, 232, 233, 234, 235, 257, 258, 259, 279, 280, 281, 282, 305, 306], "bbox": [0.07678125, 0.5956674473067916, 0.330671875, 0.9188056206088993], "iscrowd": 0, "area": 5107.596399999998, "rle": {"size": [427, 640], "counts": "Wfd01W=6J6M2N3N2M2O2M100O100O10O0100001O1N1C^C0cQDBo;?oCBP in this image.", "objects": [{"patches": [144, 145, 167, 168, 190, 191, 192, 213, 214, 215, 236, 237, 238, 258, 259], "bbox": [0.242765625, 0.39775175644028105, 0.377484375, 0.7846370023419202], "iscrowd": 0, "area": 5800.971000000002, "rle": {"size": [427, 640], "counts": "d[Q23W=2N2N3M2N2N2N2N1O1O2N1O1O2N1O1O2N1O1ODjCFR95]I5P71kHOU77eHIZ7>_HCa7c0YH]Of7j0THVOl7W1gGiNY8\\1bGdN_8_1\\GbNd8c1VG^Nk8P31O0010O01N101N1O2OD;2N101NfF_Lm8`3TGaLl8^3TGcLk8\\3VGeLj8Y3VGiLi8W3WGjLi8T3XGmLh8Q3YGPMg8o2YGRMg8l2ZGUMf8i2[GXMe8g2[GZMe8;PGY1<^Nc87VGY15aNe84[GY1NdNi8O]G[1HhNn8H]G`1CiNS9B\\Gf1_OiNX9\\O\\Gj1[OkN\\9VO\\Go1VOlN[:R21O4kMcET1b:dNbE[1b:^NbEa1b:XNcE]1@gNc;W1`DfNb;X1`DgNb;V1;N2N3M2M3M4M2M3M4M2M3N3L3M3NZfU5"}, "label": "the batter with the blue helmet"}]} {"id": 25455, "image": "COCO_train2014_000000025455.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player dodging an incoming ball\"."}], "task": "refering", "answer_template": "The \"a baseball player dodging an incoming ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 167, 168, 190, 191, 192, 213, 214, 215, 236, 237, 238, 258, 259], "bbox": [0.242765625, 0.39775175644028105, 0.377484375, 0.7846370023419202], "iscrowd": 0, "area": 5800.971000000002, "rle": {"size": [427, 640], "counts": "d[Q23W=2N2N3M2N2N2N2N1O1O2N1O1O2N1O1O2N1O1ODjCFR95]I5P71kHOU77eHIZ7>_HCa7c0YH]Of7j0THVOl7W1gGiNY8\\1bGdN_8_1\\GbNd8c1VG^Nk8P31O0010O01N101N1O2OD;2N101NfF_Lm8`3TGaLl8^3TGcLk8\\3VGeLj8Y3VGiLi8W3WGjLi8T3XGmLh8Q3YGPMg8o2YGRMg8l2ZGUMf8i2[GXMe8g2[GZMe8;PGY1<^Nc87VGY15aNe84[GY1NdNi8O]G[1HhNn8H]G`1CiNS9B\\Gf1_OiNX9\\O\\Gj1[OkN\\9VO\\Go1VOlN[:R21O4kMcET1b:dNbE[1b:^NbEa1b:XNcE]1@gNc;W1`DfNb;X1`DgNb;V1;N2N3M2M3M4M2M3M4M2M3N3L3M3NZfU5"}, "label": "a baseball player dodging an incoming ball"}]} {"id": 246641, "image": "COCO_train2014_000000246641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an open part of the table between a can and two sandwich baskets\"."}], "task": "refering", "answer_template": "The \"an open part of the table between a can and two sandwich baskets\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 69, 70, 71, 72, 92, 93, 94, 95, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279], "bbox": [0.00146875, 0.19219444444444445, 0.244, 0.9844166666666666], "iscrowd": 0, "area": 30666.176850000003, "rle": {"size": [360, 640], "counts": "\\>n7Z30O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10UN]MQKc2n4`MPK_2P5eMmJ[2R5hMlJX2S5kMkJU2T5oMiJP2X5RNfJn1Y5UNeJk1Z5YNcJg1\\5\\NbJc1^5`N`J`1_5dN^J\\1a5gN]JY1b5jN\\JU1d5oNYJQ1f5ROXJn0g5UOWJk0i5XOTJg0l5\\ORJd0c0cMU4l1VKa0a0iMV4j1VK in this image.", "objects": [{"patches": [29, 30, 31, 50, 51, 52, 53, 54, 55, 73, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 120, 121, 122, 123, 124, 144, 145, 146], "bbox": [0.2046875, 0.11347222222222222, 0.4140625, 0.49680555555555556], "iscrowd": 0, "area": 12811.75, "rle": {"size": [360, 640], "counts": "WT^13S;4L4mEIo8;kFKR99dF5U9M`Fj0Q9i0O0O101O0100OO2O001O0O10000O10O10O10O001O1O001000O01000L4M3M3L3M4M2N3M2N2O1N2N3M2O1N2N2N3M2N2N3M2N2N3M2O10000O1000001O00000010O001O1O010O1O001O001O000000000O101O000000001O0O10000O2O0O100O1O2O1N101N2^N^H[Oc7d0^H\\Oc7c0]H\\Oe7c0\\H[Oe7e0[HYOh7e0YHZOi7e0XHXOk7g0UHWOm7i0SHUOP8j0QHSOR8l0nGROV8l0jGROZ8l0fGSO]8k0dGSO_8k0aGTOc8i0]GVOf8h0ZGWOi8g0WGTOQ9i0PGPOY9m0l0F0O2N3M3M3M3M2N4JWgS4"}, "label": "a food on tabule"}]} {"id": 246641, "image": "COCO_train2014_000000246641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a meatball sandwich\"."}], "task": "refering", "answer_template": "The \"a meatball sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 50, 51, 52, 53, 54, 55, 73, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 120, 121, 122, 123, 124, 144, 145, 146], "bbox": [0.2046875, 0.11347222222222222, 0.4140625, 0.49680555555555556], "iscrowd": 0, "area": 12811.75, "rle": {"size": [360, 640], "counts": "WT^13S;4L4mEIo8;kFKR99dF5U9M`Fj0Q9i0O0O101O0100OO2O001O0O10000O10O10O10O001O1O001000O01000L4M3M3L3M4M2N3M2N2O1N2N3M2O1N2N2N3M2N2N3M2N2N3M2O10000O1000001O00000010O001O1O010O1O001O001O000000000O101O000000001O0O10000O2O0O100O1O2O1N101N2^N^H[Oc7d0^H\\Oc7c0]H\\Oe7c0\\H[Oe7e0[HYOh7e0YHZOi7e0XHXOk7g0UHWOm7i0SHUOP8j0QHSOR8l0nGROV8l0jGROZ8l0fGSO]8k0dGSO_8k0aGTOc8i0]GVOf8h0ZGWOi8g0WGTOQ9i0PGPOY9m0l0F0O2N3M3M3M3M2N4JWgS4"}, "label": "a meatball sandwich"}]} {"id": 246641, "image": "COCO_train2014_000000246641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwhich on the right in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the sandwhich on the right in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 129, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 157, 176, 177, 178, 179], "bbox": [0.605484375, 0.23369444444444443, 0.86334375, 0.59775], "iscrowd": 0, "area": 15976.108099999998, "rle": {"size": [360, 640], "counts": "m`X45o:7J6I7J6I7J6K5L4L3M4L4L3M2N2N2N3M2N2N2N2N3M2N2M301N10000O100O1O2N1O1N2O1O1O1O2N1N2O1O1O1O1O1O2M2O1O1O10O3N3L3N2M3N2M01O1N101O000O1000O10O1000O10O100000O01000000O0100000O0100000O100000O010000O1000O10O1000O10O010000O010O100O010O100O010O100O01000O10O010000O010O100O10O0100O100O1O100O100O100O1O100O6K7H7J6I8H5L1N101N1O2O0N2M4M2M?Ak0VOi]n0"}, "label": "the sandwhich on the right in the right hand picture"}]} {"id": 246641, "image": "COCO_train2014_000000246641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"right half of the table top undersandwhiches that have multiple children pictures under the glass of the top\"."}], "task": "refering", "answer_template": "The \"right half of the table top undersandwhiches that have multiple children pictures under the glass of the top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 55, 56, 57, 58, 59, 62, 63, 64, 79, 80, 81, 82, 88, 89, 103, 104, 111, 112, 113, 126, 127, 135, 136, 137, 149, 150, 158, 159, 160, 172, 173, 181, 182, 183, 204, 205, 206, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.38053125, 0.0, 1.0, 0.9985277777777779], "iscrowd": 0, "area": 43747.30765, "rle": {"size": [360, 640], "counts": "Rje26R;;D7J1O1O2M2O1O1O1N2O1O2M2O1O001N101O001N2O001O0O2O001N101O1O0O2O001O0O2O001O1N101O001N101O0O2O1O001N101O001N101O1O0O2O1O0O2O1O1O1N2O1O1O1N3N4L4L3L5L5K7H9H8ZJhKl3_4lKbKT4e4dK]KZ4i4_KYKa4l4YKUKg4P5RKSKm4R5mJoJR5d501O000O2O000O2OK5H8H7J7J6I7K4O2N2N2N1O2N2N2N1O2N2L3N3M3M3N1N3N2M3N1N3N2M3N1N3N2M2O2M3N2M3N2M3M3M3N2M3M3N2M3M3M3O1O1000000O100000000O1000000O100000000O1000000O100000000O1001O0PNgNQJZ1m5AYI?g6<^HDb7>\\HAd7`0\\H@d7`0\\H_Oe7a0[H^Oe7c0[H]Oe7c0[H\\Oe7e0[H\\Od7d0\\H]Oc7c0]H^Oa7c0_H^O`7b0`H_O_7a0aH@]7a0cH_O]7a0cH@\\7`0dHAZ7`0fHAY7?hHAV7`0jHAU7?lHAS7?nHAP7`0QI_Oo6a0RI^On6b0SI\\On6d0SI[Om6e0SI[Om6e0TIZOl6f0UIYOk6g0VIXOk6g0VIXOj6h0WIWOi6i0WIVOj6j0WIUOi6k0XITOh6l0YISOg6m0ZIROf6n0[IQOe6o0\\IPOd6P1\\IPOd6P1]IoNc6Q1^ImNc6S1^IlNb6T1_IkNa6U1`IjN`6V1aIiN_6W1aIiN_6W1bIhN^6X1cIgN]6Y1dIfN\\6Z1eIeN[6[1fIdN[6[1fIdNZ6\\1fIcN[6]1fIbNZ6^1gIaNY6_1hI`NX6`1iI_NW6a1jI^NV6b1kI]NU6c1kI]NU6c1lI\\NT6d1mI[NS6e1nIZNR6f1oIYNQ6g1PJXNP6h1QJWNo5i1QJWNo5i1RJVNn5j1SJVNl5j1UJUNk5k1VJUNi5k1XJUNg5k1ZJTNg5k1YJVNf5j1[JVNc5k1^JUN_5m1bJRN]5o1dJQNZ5P2gJPNV5R2kJnMS5S2mJmMR5T2oJlMo4U2RKkMl4V2UKiMj4X2WKhMg4Y2ZKgMe4Y2\\KgMb4Z2^KfMa4[2`KeM^4\\2cKdM\\4\\2eKdMZ4\\2gKdMX4\\2iKdMU4]2lKbMT4^2lKcMS4]2nKcMQ4]2PLcMo3]2RLcMl3^2ULbMj3^2WLaMi3_2XLaMg3_2YLbMf3^2[LbMd3^2]LbMa3_2`LaM_3_2bL`M^3`2cL`M\\3`2eL`MZ3`2fL`MY3a2hL^MX3b2iL\\MX3d2iL[MW3e2jLZMV3f2kLZMS3g2nLYMQ3g2oLYMR3f2oLZMS3c2nL]MT3`2mL`MU3]2lLbMW3[2jLjMR3T2oLRNm2k1SMYNl2d1UM\\Nm2a1TM_Nn2^1SMbNn2\\1SMeNn2X1SMhNo2U1RMkNP3R1PMPOP3n0QMROQ3k0PMUOR3h0oLYOR3d0oL\\O_2U1bMlN[2U1fMkNW2W1iMkNT2V1mMjNP2X1QNiNl1X1UNhNh1Z1YNgNd1Z1]NgN_1[1bNfNZ1\\1fNfNW1[1jNfNR1\\1oNeNn0\\1SOeNi0]1XOdNe0]1\\OdN`0^1AcN<^1DeN8\\1IeN3]1NeNjNiNbNc2e2gNbNjNgN`2h2hN[NlNjN^2l2hNSNoNoNZ2o2jNkMPOSOX2R3kNeMQOWOU2U3kN_MSO[OS2W3kNXMWO^OP2[3jNRMZOAm1^3jNlL]ODj1a3jNZLL2\\1e3h0WL[Oi3f0RL^Oo3m31O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O^J"}, "label": "right half of the table top undersandwhiches that have multiple children pictures under the glass of the top"}]} {"id": 246641, "image": "COCO_train2014_000000246641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a table with pictures of people on it\"."}], "task": "refering", "answer_template": "The \"this is a table with pictures of people on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 55, 56, 57, 58, 59, 62, 63, 64, 79, 80, 81, 82, 88, 89, 103, 104, 111, 112, 113, 126, 127, 135, 136, 137, 149, 150, 158, 159, 160, 172, 173, 181, 182, 183, 204, 205, 206, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.38053125, 0.0, 1.0, 0.9985277777777779], "iscrowd": 0, "area": 43747.30765, "rle": {"size": [360, 640], "counts": "Rje26R;;D7J1O1O2M2O1O1O1N2O1O2M2O1O001N101O001N2O001O0O2O001N101O1O0O2O001O0O2O001O1N101O001N101O0O2O1O001N101O001N101O1O0O2O1O0O2O1O1O1N2O1O1O1N3N4L4L3L5L5K7H9H8ZJhKl3_4lKbKT4e4dK]KZ4i4_KYKa4l4YKUKg4P5RKSKm4R5mJoJR5d501O000O2O000O2OK5H8H7J7J6I7K4O2N2N2N1O2N2N2N1O2N2L3N3M3M3N1N3N2M3N1N3N2M3N1N3N2M2O2M3N2M3N2M3M3M3N2M3M3N2M3M3M3O1O1000000O100000000O1000000O100000000O1000000O100000000O1001O0PNgNQJZ1m5AYI?g6<^HDb7>\\HAd7`0\\H@d7`0\\H_Oe7a0[H^Oe7c0[H]Oe7c0[H\\Oe7e0[H\\Od7d0\\H]Oc7c0]H^Oa7c0_H^O`7b0`H_O_7a0aH@]7a0cH_O]7a0cH@\\7`0dHAZ7`0fHAY7?hHAV7`0jHAU7?lHAS7?nHAP7`0QI_Oo6a0RI^On6b0SI\\On6d0SI[Om6e0SI[Om6e0TIZOl6f0UIYOk6g0VIXOk6g0VIXOj6h0WIWOi6i0WIVOj6j0WIUOi6k0XITOh6l0YISOg6m0ZIROf6n0[IQOe6o0\\IPOd6P1\\IPOd6P1]IoNc6Q1^ImNc6S1^IlNb6T1_IkNa6U1`IjN`6V1aIiN_6W1aIiN_6W1bIhN^6X1cIgN]6Y1dIfN\\6Z1eIeN[6[1fIdN[6[1fIdNZ6\\1fIcN[6]1fIbNZ6^1gIaNY6_1hI`NX6`1iI_NW6a1jI^NV6b1kI]NU6c1kI]NU6c1lI\\NT6d1mI[NS6e1nIZNR6f1oIYNQ6g1PJXNP6h1QJWNo5i1QJWNo5i1RJVNn5j1SJVNl5j1UJUNk5k1VJUNi5k1XJUNg5k1ZJTNg5k1YJVNf5j1[JVNc5k1^JUN_5m1bJRN]5o1dJQNZ5P2gJPNV5R2kJnMS5S2mJmMR5T2oJlMo4U2RKkMl4V2UKiMj4X2WKhMg4Y2ZKgMe4Y2\\KgMb4Z2^KfMa4[2`KeM^4\\2cKdM\\4\\2eKdMZ4\\2gKdMX4\\2iKdMU4]2lKbMT4^2lKcMS4]2nKcMQ4]2PLcMo3]2RLcMl3^2ULbMj3^2WLaMi3_2XLaMg3_2YLbMf3^2[LbMd3^2]LbMa3_2`LaM_3_2bL`M^3`2cL`M\\3`2eL`MZ3`2fL`MY3a2hL^MX3b2iL\\MX3d2iL[MW3e2jLZMV3f2kLZMS3g2nLYMQ3g2oLYMR3f2oLZMS3c2nL]MT3`2mL`MU3]2lLbMW3[2jLjMR3T2oLRNm2k1SMYNl2d1UM\\Nm2a1TM_Nn2^1SMbNn2\\1SMeNn2X1SMhNo2U1RMkNP3R1PMPOP3n0QMROQ3k0PMUOR3h0oLYOR3d0oL\\O_2U1bMlN[2U1fMkNW2W1iMkNT2V1mMjNP2X1QNiNl1X1UNhNh1Z1YNgNd1Z1]NgN_1[1bNfNZ1\\1fNfNW1[1jNfNR1\\1oNeNn0\\1SOeNi0]1XOdNe0]1\\OdN`0^1AcN<^1DeN8\\1IeN3]1NeNjNiNbNc2e2gNbNjNgN`2h2hN[NlNjN^2l2hNSNoNoNZ2o2jNkMPOSOX2R3kNeMQOWOU2U3kN_MSO[OS2W3kNXMWO^OP2[3jNRMZOAm1^3jNlL]ODj1a3jNZLL2\\1e3h0WL[Oi3f0RL^Oo3m31O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O^J"}, "label": "this is a table with pictures of people on it"}]} {"id": 246641, "image": "COCO_train2014_000000246641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a red and white shirt\"."}], "task": "refering", "answer_template": "The \"a person in a red and white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 69, 70], "bbox": [0.00146875, 0.003916666666666666, 0.26075, 0.2721388888888889], "iscrowd": 0, "area": 12145.47305, "rle": {"size": [360, 640], "counts": "Y;Q3W800O1000000O1000000O1000000O101O000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O1000000O10000O10000O3N2N3L3N2N2N3Lf0[O4L4K6KdWV5"}, "label": "a person in a red and white shirt"}]} {"id": 246641, "image": "COCO_train2014_000000246641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and white blouse\"."}], "task": "refering", "answer_template": "The \"a red and white blouse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 69, 70], "bbox": [0.00146875, 0.003916666666666666, 0.26075, 0.2721388888888889], "iscrowd": 0, "area": 12145.47305, "rle": {"size": [360, 640], "counts": "Y;Q3W800O1000000O1000000O1000000O101O000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O1000000O10000O10000O3N2N3L3N2N2N3Lf0[O4L4K6KdWV5"}, "label": "a red and white blouse"}]} {"id": 525180, "image": "COCO_train2014_000000525180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the blue - green towel tucked into his apron\"."}], "task": "refering", "answer_template": "The \"the man with the blue - green towel tucked into his apron\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 308, 309, 310, 311, 331, 332, 333, 334], "bbox": [0.39834375, 0.4608430913348946, 0.6270625, 0.9989929742388759], "iscrowd": 0, "area": 23521.508, "rle": {"size": [427, 640], "counts": "bfZ3;d in this image.", "objects": [{"patches": [171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 308, 309, 310, 311, 331, 332, 333, 334], "bbox": [0.39834375, 0.4608430913348946, 0.6270625, 0.9989929742388759], "iscrowd": 0, "area": 23521.508, "rle": {"size": [427, 640], "counts": "bfZ3;d in this image.", "objects": [{"patches": [156, 157, 179, 180, 181, 202, 203, 204, 225, 226, 227, 248, 249, 250, 271, 272, 273, 294, 295, 296, 317, 318, 319], "bbox": [0.7921875, 0.4184309133489461, 0.915625, 0.8985245901639344], "iscrowd": 0, "area": 9438.000000000002, "rle": {"size": [427, 640], "counts": "ldc6;b<>M3M2O2M3M2N3N2ZDYOh:W2]O;F4UOhLZG\\3b8fL]G]3?jLd6KjH`3:SMa6_OTIa32WMh6ZOTId3H[MS7SOSIP5l6RKRIP5l6RKSIo4k6TKRIn4m6SKRIn4l6TKRIm4n6UKPIl4o6i0OM3I7I7I7I7J6I7I7I7I7N32100O00100O1OO2L3N2N2M4M2N2M4M2M3N3M2M2ON101O1_LmGh1S8VNTHe1m7XNYHd1h7YN_Ha1b7^NdH]1]7`NjH[1V7cNPILQOOP82VIXO^Od0]7JkJ4U5[O_Ka0h8N2N3M2N3M2M3N3M2N2NkTf0"}, "label": "a man wearing a black apron and a grey shirt standing up , with his body facing the door"}]} {"id": 525180, "image": "COCO_train2014_000000525180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black man in restaurant kitchen with his back turned , wearing grey shirt and black pants\"."}], "task": "refering", "answer_template": "The \"black man in restaurant kitchen with his back turned , wearing grey shirt and black pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 179, 180, 181, 202, 203, 204, 225, 226, 227, 248, 249, 250, 271, 272, 273, 294, 295, 296, 317, 318, 319], "bbox": [0.7921875, 0.4184309133489461, 0.915625, 0.8985245901639344], "iscrowd": 0, "area": 9438.000000000002, "rle": {"size": [427, 640], "counts": "ldc6;b<>M3M2O2M3M2N3N2ZDYOh:W2]O;F4UOhLZG\\3b8fL]G]3?jLd6KjH`3:SMa6_OTIa32WMh6ZOTId3H[MS7SOSIP5l6RKRIP5l6RKSIo4k6TKRIn4m6SKRIn4l6TKRIm4n6UKPIl4o6i0OM3I7I7I7I7J6I7I7I7I7N32100O00100O1OO2L3N2N2M4M2N2M4M2M3N3M2M2ON101O1_LmGh1S8VNTHe1m7XNYHd1h7YN_Ha1b7^NdH]1]7`NjH[1V7cNPILQOOP82VIXO^Od0]7JkJ4U5[O_Ka0h8N2N3M2N3M2M3N3M2N2NkTf0"}, "label": "black man in restaurant kitchen with his back turned , wearing grey shirt and black pants"}]} {"id": 426877, "image": "COCO_train2014_000000426877.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the white shirt\"."}], "task": "refering", "answer_template": "The \"the man in the white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 54, 55, 70, 71, 86, 87, 102, 103, 115, 116, 117, 118, 119, 131, 132, 133, 134, 147, 148, 149, 150, 163, 164, 165, 166, 167, 179, 180, 181, 182, 183, 196, 197, 198, 199, 212, 213, 214, 215, 227, 228, 229, 230, 231, 243, 244, 245, 246, 259, 260, 261, 262, 274, 275, 276, 277, 278, 290, 291, 292, 293, 306, 307, 308, 309, 321, 322, 323, 324, 325, 337, 338, 339, 340, 354, 356], "bbox": [0.09734234234234233, 0.122078125, 0.460518018018018, 0.97403125], "iscrowd": 0, "area": 38160.088200000006, "rle": {"size": [640, 444], "counts": "f^k01lc04M3M3M3M3M3M3L4M4K4H8H8H8H8H8H7I8H7N3M1O1O100OK5L4L4N3O0O10000O2O000O100O101O0O10]JQNWIo1W6fNfIY1Z6kNdIT1\\6oN`IR1`6QO[IP1\\5eMTG_1[3n0b5jMkF[1^3m0g5PNbFW1b3j0n5VNXFS1e3i0S6\\NPFn0h3h0Y6`0bIA_6b0]I_Od6c0WI^Ok6d0QI]Oo6f0lH\\OU7f0fH[O[7h0aHYO`7i0[HYOf7i0VHXOj7k0QHWOP8k0lGVO]OQNi5l2eJUOASNh5j2cJUOBVNi5h2_JTOGXNg5g2^JROJ[Ng5e2YJRON^Nh5b2TJRO3aNg5^2RJSO5cNh5\\2mISO:eNh5Z2iIRO=iNi5W2dIROb0kNi5T2aISOd0nNi5Q2]ISOi0POi5o1YIROl0TOj5l5UJXJj5g5TJ]Jk5a5TJdJk5Y5SJlJm5Q5RJRKo5k4PJXKQ6f4lI^KU6_4jIcKX6Z4gIiKZ6T4eIoK\\6o3aITLa6i3^IZLd6b3[IaLf6\\3YIgLh6V3VImLl6Q3RIRMo6k2PI[Ln7b3QH^LS8_3kGbLY8\\3eGcL`8Z3_GfLe8W3YGjLk8R74L4L2N3M3M2N3M2N3M2N3M3M2N1O001O001O001O00001O001O001O001O0O2O000010nM`FeI`9[6hF[IZ9d6oFRIS9m6VGiHl8W7\\G`He8`7cGVH_8n7eGhG]8[8a14M3L3oIZGe0j8UO^Ge0e8WObGc0b8WOeGd0^8WOjGc0Z8XOmGb0W8XOQHc0R8WOYH`0m7YO_H in this image.", "objects": [{"patches": [38, 54, 55, 70, 71, 86, 87, 102, 103, 115, 116, 117, 118, 119, 131, 132, 133, 134, 147, 148, 149, 150, 163, 164, 165, 166, 167, 179, 180, 181, 182, 183, 196, 197, 198, 199, 212, 213, 214, 215, 227, 228, 229, 230, 231, 243, 244, 245, 246, 259, 260, 261, 262, 274, 275, 276, 277, 278, 290, 291, 292, 293, 306, 307, 308, 309, 321, 322, 323, 324, 325, 337, 338, 339, 340, 354, 356], "bbox": [0.09734234234234233, 0.122078125, 0.460518018018018, 0.97403125], "iscrowd": 0, "area": 38160.088200000006, "rle": {"size": [640, 444], "counts": "f^k01lc04M3M3M3M3M3M3L4M4K4H8H8H8H8H8H7I8H7N3M1O1O100OK5L4L4N3O0O10000O2O000O100O101O0O10]JQNWIo1W6fNfIY1Z6kNdIT1\\6oN`IR1`6QO[IP1\\5eMTG_1[3n0b5jMkF[1^3m0g5PNbFW1b3j0n5VNXFS1e3i0S6\\NPFn0h3h0Y6`0bIA_6b0]I_Od6c0WI^Ok6d0QI]Oo6f0lH\\OU7f0fH[O[7h0aHYO`7i0[HYOf7i0VHXOj7k0QHWOP8k0lGVO]OQNi5l2eJUOASNh5j2cJUOBVNi5h2_JTOGXNg5g2^JROJ[Ng5e2YJRON^Nh5b2TJRO3aNg5^2RJSO5cNh5\\2mISO:eNh5Z2iIRO=iNi5W2dIROb0kNi5T2aISOd0nNi5Q2]ISOi0POi5o1YIROl0TOj5l5UJXJj5g5TJ]Jk5a5TJdJk5Y5SJlJm5Q5RJRKo5k4PJXKQ6f4lI^KU6_4jIcKX6Z4gIiKZ6T4eIoK\\6o3aITLa6i3^IZLd6b3[IaLf6\\3YIgLh6V3VImLl6Q3RIRMo6k2PI[Ln7b3QH^LS8_3kGbLY8\\3eGcL`8Z3_GfLe8W3YGjLk8R74L4L2N3M3M2N3M2N3M2N3M3M2N1O001O001O001O00001O001O001O001O0O2O000010nM`FeI`9[6hF[IZ9d6oFRIS9m6VGiHl8W7\\G`He8`7cGVH_8n7eGhG]8[8a14M3L3oIZGe0j8UO^Ge0e8WObGc0b8WOeGd0^8WOjGc0Z8XOmGb0W8XOQHc0R8WOYH`0m7YO_H in this image.", "objects": [{"patches": [100, 101, 102, 116, 117, 135, 151, 167, 183, 199, 215, 216, 231, 232, 247, 248, 263, 264, 279, 280, 295, 296, 311, 312, 327, 328], "bbox": [0.2610810810810811, 0.266625, 0.5564639639639639, 0.8907343749999999], "iscrowd": 0, "area": 11721.990050000002, "rle": {"size": [640, 444], "counts": "UZY21nc01N2O1O1O1O1O1O1N2O2N2M3N2N9H3M4L3L3N000O2O001N10001O0O0100O001O1O1O001O1O1O1O001O1N2O1N2O0O2O1O1O1N2O001N200O1O1O1O1O1O1O1O101N2N2N2N1O2N2N2Ng\\32`bLo0QOP1POP1POQ1oNS1mNd1\\NP1POQ1oNo0TOf06J6J7I6J6J6J7I6J6J6J7H7J6J6J7I6J6I7I6I5L5K4L4L3L3N3M2N3TLkBg0W=SOTCe0o@bcj3"}, "label": "the boy dressed in red"}]} {"id": 426877, "image": "COCO_train2014_000000426877.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy in a red outfit reaching for a frisbee\"."}], "task": "refering", "answer_template": "The \"a guy in a red outfit reaching for a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 116, 117, 135, 151, 167, 183, 199, 215, 216, 231, 232, 247, 248, 263, 264, 279, 280, 295, 296, 311, 312, 327, 328], "bbox": [0.2610810810810811, 0.266625, 0.5564639639639639, 0.8907343749999999], "iscrowd": 0, "area": 11721.990050000002, "rle": {"size": [640, 444], "counts": "UZY21nc01N2O1O1O1O1O1O1N2O2N2M3N2N9H3M4L3L3N000O2O001N10001O0O0100O001O1O1O001O1O1O1O001O1N2O1N2O0O2O1O1O1N2O001N200O1O1O1O1O1O1O1O101N2N2N2N1O2N2N2Ng\\32`bLo0QOP1POP1POQ1oNS1mNd1\\NP1POQ1oNo0TOf06J6J7I6J6J6J7I6J6J6J7H7J6J6J7I6J6I7I6I5L5K4L4L3L3N3M2N3TLkBg0W=SOTCe0o@bcj3"}, "label": "a guy in a red outfit reaching for a frisbee"}]} {"id": 189330, "image": "COCO_train2014_000000189330.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second red bench back , starting from the closest to the camera\"."}], "task": "refering", "answer_template": "The \"the second red bench back , starting from the closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316], "bbox": [0.18990625, 0.7665661252900232, 0.7810781250000001, 0.8838515081206496], "iscrowd": 0, "area": 17416.822500000006, "rle": {"size": [431, 640], "counts": "bec1\\1S<00001O0000000000000000000000001N10000000000000000009Gn0SOWf6Z1^XI00000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000O10000000000000000000000000000000000000000YOncj1"}, "label": "the second red bench back , starting from the closest to the camera"}]} {"id": 189330, "image": "COCO_train2014_000000189330.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red park benches\"."}], "task": "refering", "answer_template": "The \"red park benches\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316], "bbox": [0.18990625, 0.7665661252900232, 0.7810781250000001, 0.8838515081206496], "iscrowd": 0, "area": 17416.822500000006, "rle": {"size": [431, 640], "counts": "bec1\\1S<00001O0000000000000000000000001N10000000000000000009Gn0SOWf6Z1^XI00000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000O10000000000000000000000000000000000000000YOncj1"}, "label": "red park benches"}]} {"id": 189330, "image": "COCO_train2014_000000189330.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"nearest bench\"."}], "task": "refering", "answer_template": "The \"nearest bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.126265625, 0.8736890951276102, 0.925984375, 0.9882830626450115], "iscrowd": 0, "area": 20548.486399999987, "rle": {"size": [431, 640], "counts": "P`R17n<:F:G9G90000000000000000001O000000000001O00000000000001O0000000000000000000000001O0000000000000001O0000000001O00000000000000000O1N2O6Ja0^OQm5ImRJ;D in this image.", "objects": [{"patches": [29, 30, 31, 32, 33, 34, 35, 36, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 103], "bbox": [0.257453125, 0.11074941451990632, 0.6907343750000001, 0.2895784543325527], "iscrowd": 0, "area": 12881.23955, "rle": {"size": [427, 640], "counts": "VkT2e1f;1O1O0000000010O0000000001O000000\\OiDTOW;l0QEmNn:S1e00000000000000000001O2N1O2N1O010O00001O001O001O000000001O00000000001O00000000001O0001O000001O000000000000000000000000000000000000000000000001O00000000000000000001O0000O1000000O1000000O1000000O101O00000O100O100O100O10000O100O100000000000000000WD@g:b100001O000000000000000000000000000000000000000nNQEHo:8WEBi:>\\E]Od:c0bEWO^:i0gEROY:n0mElNS:T1S1O000000000000000001O1O1O001O1O1O001O0000001O0000001O01O0001O000000000000000000000000000000000000000000000001O000000000001O00000000O100000000000000000000000000000001N100000000O1000000O1000000O1000000O1000000XOh0J6K^`b2"}, "label": "an empty table with seven chairs"}]} {"id": 197525, "image": "COCO_train2014_000000197525.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table partially cover with a red cloth , a white plate , and other various items\"."}], "task": "refering", "answer_template": "The \"a table partially cover with a red cloth , a white plate , and other various items\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 33, 34, 35, 36, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 103], "bbox": [0.257453125, 0.11074941451990632, 0.6907343750000001, 0.2895784543325527], "iscrowd": 0, "area": 12881.23955, "rle": {"size": [427, 640], "counts": "VkT2e1f;1O1O0000000010O0000000001O000000\\OiDTOW;l0QEmNn:S1e00000000000000000001O2N1O2N1O010O00001O001O001O000000001O00000000001O00000000001O0001O000001O000000000000000000000000000000000000000000000001O00000000000000000001O0000O1000000O1000000O1000000O101O00000O100O100O100O10000O100O100000000000000000WD@g:b100001O000000000000000000000000000000000000000nNQEHo:8WEBi:>\\E]Od:c0bEWO^:i0gEROY:n0mElNS:T1S1O000000000000000001O1O1O001O1O1O001O0000001O0000001O01O0001O000000000000000000000000000000000000000000000001O000000000001O00000000O100000000000000000000000000000001N100000000O1000000O1000000O1000000O1000000XOh0J6K^`b2"}, "label": "a table partially cover with a red cloth , a white plate , and other various items"}]} {"id": 197525, "image": "COCO_train2014_000000197525.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large pizza on a table with a woman ' s hand above it\"."}], "task": "refering", "answer_template": "The \"a large pizza on a table with a woman ' s hand above it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 119, 120, 121, 122, 123, 124, 128, 129, 139, 140, 141, 142, 143, 144, 145, 149, 150, 151, 152, 153, 161, 162, 163, 164, 165, 167, 169, 170, 171, 172, 173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.3605854800936768, 0.99528125, 1.0], "iscrowd": 0, "area": 123773.48935000002, "rle": {"size": [427, 640], "counts": "` in this image.", "objects": [{"patches": [116, 117, 118, 119, 120, 121, 122, 123, 124, 128, 129, 139, 140, 141, 142, 143, 144, 145, 149, 150, 151, 152, 153, 161, 162, 163, 164, 165, 167, 169, 170, 171, 172, 173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.3605854800936768, 0.99528125, 1.0], "iscrowd": 0, "area": 123773.48935000002, "rle": {"size": [427, 640], "counts": "` in this image.", "objects": [{"patches": [285, 286, 287, 302, 303, 304, 313, 314, 320, 321, 322, 330, 331, 338], "bbox": [0.4385, 0.70303125, 0.9516875, 0.850546875], "iscrowd": 0, "area": 4630.994399999998, "rle": {"size": [640, 480], "counts": "fhS41ic06L4L4L4L400O100O2O0O010O10000O01O001O1O1O1O1O3L3Mi\\OCPc0:P]OHRc05m\\OMTc00l\\O2Uc0Kk\\O6^c02O11O1O001O1N2O2N2NT^]21`abM=L4L4L4M101O001N101O001O001O0O2O1O1O001O100O1O001O1O100O1O1O2O0O101N100O2O0O101N3N1O2N2M3N2N3M3M2M01O100O100O1O101N100O2N1O2N1N3N1N3N1O2M2O1N3N1O2M2O3L3N3K5K4L5K:Fik="}, "label": "cycle fell down in a floor"}]} {"id": 107425, "image": "COCO_train2014_000000107425.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue bicycle is laying on the ground next to the white bike that is standing up\"."}], "task": "refering", "answer_template": "The \"the blue bicycle is laying on the ground next to the white bike that is standing up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [285, 286, 287, 302, 303, 304, 313, 314, 320, 321, 322, 330, 331, 338], "bbox": [0.4385, 0.70303125, 0.9516875, 0.850546875], "iscrowd": 0, "area": 4630.994399999998, "rle": {"size": [640, 480], "counts": "fhS41ic06L4L4L4L400O100O2O0O010O10000O01O001O1O1O1O1O3L3Mi\\OCPc0:P]OHRc05m\\OMTc00l\\O2Uc0Kk\\O6^c02O11O1O001O1N2O2N2NT^]21`abM=L4L4L4M101O001N101O001O001O0O2O1O1O001O100O1O001O1O100O1O1O2O0O101N100O2O0O101N3N1O2N2M3N2N3M3M2M01O100O100O1O101N100O2N1O2N1N3N1N3N1O2M2O1N3N1O2M2O3L3N3K5K4L5K:Fik="}, "label": "the blue bicycle is laying on the ground next to the white bike that is standing up"}]} {"id": 107425, "image": "COCO_train2014_000000107425.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bicycle upright and leaning against a metal sculpture with a pink water bottle attached to it\"."}], "task": "refering", "answer_template": "The \"a bicycle upright and leaning against a metal sculpture with a pink water bottle attached to it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 260, 261, 262, 263, 264, 265, 266, 277, 278, 279, 280, 281, 282, 283, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 327, 328, 329, 330, 332, 333, 334, 335, 336, 337, 345, 346, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371], "bbox": [0.27295833333333336, 0.616296875, 0.8774583333333335, 0.951109375], "iscrowd": 0, "area": 32712.794650000003, "rle": {"size": [640, 480], "counts": "d[b25hc09G9G9G9G9lM`NYAi1^>YN`AP2X>RNeAW2R>kMlAZ2o=iMnA[2n=gMPB^2k=QNhAR2V>nMiAU2T>mMkAT2T>mMjAV2S>VMTA8g0e2S>SMZA5a0j2S>gLVA9:3eLZA7<26U3Q>bL`A5:43X3Q>_LbA5:51Y3Q>]LeA5:5O[3X>`LjA5L^3o=XLkA5;5I_3Z>\\LnA5F`3R>VLmA5<3Fb3[>\\LoAOGg3n=TLPB6ZLRBJHl3V>ZLcBe3]=[LcBf3\\=ZLdBf3\\=ZLdBf3\\=ZLdBf3\\=ZLdBf3\\=ZLdBg3[=YLeBg3[=YLeBg3[=YLeBg3[=YLeBg3[=YLeBh3Z=XLfBg3[=YLeBg3[=YLeBf3\\=ZLdBf3\\=ZLdBe3]=[LcBe3]=[LcBd3^=\\LoAKLi3U>\\LhA32`3W>]LaA:7Y3X>]L\\Aa0:Q3[>ZMdAf2\\>[M_Ah2b>ZMXAj2h>V1100M3N2hLTAY1o>eNXAT1j>jN\\Ao0g>oN`Aj0b>UOdAc0_>^OdAXNm@]1R13T>8oAAR>Bh@LY1;R>Fh@0Z13o=Li@2\\1Jn=1i@7[1Ao=5i@;\\1YOl=;j@=]1QOl=?j@a0Ua0\\On^Of0o`0ZOR_Og0m`0ZOR_Og0m`0YOS_Oh0k`0ZOU_Og0i`0YOW_Oh0g`0ZOX_Og0Ta0lNl^OV1Ta0iNk^OX1Ua0gNk^OZ1Ta0gNk^OZ1Ua0eNk^O]1Ta0cNk^O^1Ta0bNl^O_1Ua0`Nj^Oa1Ya0[Ng^Og1ia00O1O1O2N1O1O1O2N1O1O101N1O1O1O2N1O1O1O2V@aMo=`2oAdMn=]2QBfMl=[2RBiMl=X2RBkMk=V2TBkMk=V2SBjMn=W2QBhMP>Z2nAeMS>\\2kAdMV>]2iAbMX>`2dAaM]>`2_AbMb>`2ZAZMn>i2l@XMW?j2c@WM_?l2\\@TMf?o2T@RMn?^32N2N2N2N2N21O1O100O1O1O1^OZLh@f3V?]Li@c3V?^Lj@b3U?`Lj@`3U?bLj@^3V?cLi@]3W?cLi@]3W?dLh@\\3X?eLh@Y3Y?hLf@X3Z?hLf@X3Z?iLe@V3\\?kLc@U3]?kLc@U3^?kLa@U3_?lL`@S3a?nL^@P3d?PM\\@n2f?SMY@j2j?WMV@g2k?ZMT@i2i?WMW@l2f?UMY@n2d?SM[@P3b?k0L3M3M100O2NPLk@V3T?jLm@V3R?jLPAU3o>kLRAU3m>kLUAT3j>lLWAT3h>lLZAT3d>kL^AU3a>kLaAT3^>lLcAT3\\>lLfAS3Y>mLhAS3W>lLlAS3T>lLmAU3Q>jLRBU3m=jLUBV3j=jLXBU3g=jL[BV3d=jL\\BW3c=fL`B[3_=_LgBb3X=XLnBh3e>00001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O01O010O0010O00O2O001O000O2O001O0O2O00001N101O00001N101O001N10001O0O2O002N1O1N3N1O1O2M2O1O2N1N2M4L3N2M4M2M5K5L3L5L4K5K4M4K5K5L4K4MR[T1"}, "label": "a bicycle upright and leaning against a metal sculpture with a pink water bottle attached to it"}]} {"id": 107425, "image": "COCO_train2014_000000107425.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white road bike\"."}], "task": "refering", "answer_template": "The \"a white road bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 260, 261, 262, 263, 264, 265, 266, 277, 278, 279, 280, 281, 282, 283, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 327, 328, 329, 330, 332, 333, 334, 335, 336, 337, 345, 346, 348, 349, 350, 351, 352, 353, 354, 368, 369, 370, 371], "bbox": [0.27295833333333336, 0.616296875, 0.8774583333333335, 0.951109375], "iscrowd": 0, "area": 32712.794650000003, "rle": {"size": [640, 480], "counts": "d[b25hc09G9G9G9G9lM`NYAi1^>YN`AP2X>RNeAW2R>kMlAZ2o=iMnA[2n=gMPB^2k=QNhAR2V>nMiAU2T>mMkAT2T>mMjAV2S>VMTA8g0e2S>SMZA5a0j2S>gLVA9:3eLZA7<26U3Q>bL`A5:43X3Q>_LbA5:51Y3Q>]LeA5:5O[3X>`LjA5L^3o=XLkA5;5I_3Z>\\LnA5F`3R>VLmA5<3Fb3[>\\LoAOGg3n=TLPB6ZLRBJHl3V>ZLcBe3]=[LcBf3\\=ZLdBf3\\=ZLdBf3\\=ZLdBf3\\=ZLdBf3\\=ZLdBg3[=YLeBg3[=YLeBg3[=YLeBg3[=YLeBg3[=YLeBh3Z=XLfBg3[=YLeBg3[=YLeBf3\\=ZLdBf3\\=ZLdBe3]=[LcBe3]=[LcBd3^=\\LoAKLi3U>\\LhA32`3W>]LaA:7Y3X>]L\\Aa0:Q3[>ZMdAf2\\>[M_Ah2b>ZMXAj2h>V1100M3N2hLTAY1o>eNXAT1j>jN\\Ao0g>oN`Aj0b>UOdAc0_>^OdAXNm@]1R13T>8oAAR>Bh@LY1;R>Fh@0Z13o=Li@2\\1Jn=1i@7[1Ao=5i@;\\1YOl=;j@=]1QOl=?j@a0Ua0\\On^Of0o`0ZOR_Og0m`0ZOR_Og0m`0YOS_Oh0k`0ZOU_Og0i`0YOW_Oh0g`0ZOX_Og0Ta0lNl^OV1Ta0iNk^OX1Ua0gNk^OZ1Ta0gNk^OZ1Ua0eNk^O]1Ta0cNk^O^1Ta0bNl^O_1Ua0`Nj^Oa1Ya0[Ng^Og1ia00O1O1O2N1O1O1O2N1O1O101N1O1O1O2N1O1O1O2V@aMo=`2oAdMn=]2QBfMl=[2RBiMl=X2RBkMk=V2TBkMk=V2SBjMn=W2QBhMP>Z2nAeMS>\\2kAdMV>]2iAbMX>`2dAaM]>`2_AbMb>`2ZAZMn>i2l@XMW?j2c@WM_?l2\\@TMf?o2T@RMn?^32N2N2N2N2N21O1O100O1O1O1^OZLh@f3V?]Li@c3V?^Lj@b3U?`Lj@`3U?bLj@^3V?cLi@]3W?cLi@]3W?dLh@\\3X?eLh@Y3Y?hLf@X3Z?hLf@X3Z?iLe@V3\\?kLc@U3]?kLc@U3^?kLa@U3_?lL`@S3a?nL^@P3d?PM\\@n2f?SMY@j2j?WMV@g2k?ZMT@i2i?WMW@l2f?UMY@n2d?SM[@P3b?k0L3M3M100O2NPLk@V3T?jLm@V3R?jLPAU3o>kLRAU3m>kLUAT3j>lLWAT3h>lLZAT3d>kL^AU3a>kLaAT3^>lLcAT3\\>lLfAS3Y>mLhAS3W>lLlAS3T>lLmAU3Q>jLRBU3m=jLUBV3j=jLXBU3g=jL[BV3d=jL\\BW3c=fL`B[3_=_LgBb3X=XLnBh3e>00001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O01O010O0010O00O2O001O000O2O001O0O2O00001N101O00001N101O001N10001O0O2O002N1O1N3N1O1O2M2O1O2N1N2M4L3N2M4M2M5K5L3L5L4K5K4M4K5K5L4K4MR[T1"}, "label": "a white road bike"}]} {"id": 74663, "image": "COCO_train2014_000000074663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe nearest the camera\"."}], "task": "refering", "answer_template": "The \"giraffe nearest the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 62, 63, 64, 77, 78, 79, 80, 81, 82, 92, 93, 94, 95, 96, 97, 98, 99, 108, 109, 110, 111, 112, 113, 115, 116, 123, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 145, 155, 156, 157, 158, 159, 160, 161, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194, 204, 205, 206, 207, 208, 209, 210, 211, 221, 222, 223, 224, 225, 226, 227, 228, 239, 240, 241, 243, 244, 245, 256, 257, 258, 260, 261, 273, 274, 277, 278, 290, 291, 294, 295, 307, 308, 311, 312, 324, 325, 328, 329, 341, 342, 345, 346, 358, 359, 362, 363], "bbox": [0.032551440329218105, 0.096625, 0.8641152263374485, 0.964046875], "iscrowd": 0, "area": 61087.23030000001, "rle": {"size": [640, 486], "counts": "dZ:4jc05K5J7J5J6K5J6K5K5J6K5J6L5J5L4L=C>B>A?B:F4L5J6K4L5K5K5J5L5K5K4K6K5K5bE^IU7f6XHjIg7X9N3N1O1O2M2OnJiF]OV9?XGWOh8d0hGPOY8l0VHgNk7U1XJlLi5P3e5L4K4M4K6K4KeAQ1P8kN]Gm1c8nM`GU2^8fMfG\\2Z8_MiGc2V8XMoGj2P8TMPHn2o7RMPHl2S8TMlGa2a8jMSGQO9QOP9W2]FfN` in this image.", "objects": [{"patches": [46, 47, 62, 63, 64, 77, 78, 79, 80, 81, 82, 92, 93, 94, 95, 96, 97, 98, 99, 108, 109, 110, 111, 112, 113, 115, 116, 123, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 145, 155, 156, 157, 158, 159, 160, 161, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194, 204, 205, 206, 207, 208, 209, 210, 211, 221, 222, 223, 224, 225, 226, 227, 228, 239, 240, 241, 243, 244, 245, 256, 257, 258, 260, 261, 273, 274, 277, 278, 290, 291, 294, 295, 307, 308, 311, 312, 324, 325, 328, 329, 341, 342, 345, 346, 358, 359, 362, 363], "bbox": [0.032551440329218105, 0.096625, 0.8641152263374485, 0.964046875], "iscrowd": 0, "area": 61087.23030000001, "rle": {"size": [640, 486], "counts": "dZ:4jc05K5J7J5J6K5J6K5K5J6K5J6L5J5L4L=C>B>A?B:F4L5J6K4L5K5K5J5L5K5K4K6K5K5bE^IU7f6XHjIg7X9N3N1O1O2M2OnJiF]OV9?XGWOh8d0hGPOY8l0VHgNk7U1XJlLi5P3e5L4K4M4K6K4KeAQ1P8kN]Gm1c8nM`GU2^8fMfG\\2Z8_MiGc2V8XMoGj2P8TMPHn2o7RMPHl2S8TMlGa2a8jMSGQO9QOP9W2]FfN` in this image.", "objects": [{"patches": [41, 42, 57, 58, 59, 60, 74, 75, 76, 77, 78, 93, 94, 96, 97, 98, 113, 114, 115, 116, 130, 131, 132, 133, 134, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 215, 216, 217, 218, 219, 220, 233, 236, 249, 250, 253, 254, 267, 270, 271, 287], "bbox": [0.3598559670781893, 0.08671875, 0.9712551440329218, 0.714640625], "iscrowd": 0, "area": 29818.398400000002, "rle": {"size": [640, 486], "counts": "\\_]32mc03N2M3M3M3M3M3M3M3M100O001O1O001O001O10O01O0O2N2N101N1O2N2N1O2N1O2N2N1O2N1O2N20O2O0O2O1O0O2O1N101O1N10?A02N2M2O2N2N2M2O2N2M3N1O2M3N2N1N3N2N2M2O2N2M3N1OhMj^Oh1Sa0QNV_OP2]a0O1N2O1N2N2N2O1N2N2N2O1N2N1O2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N101N2N2N2O1N2N2N2O`U==SjB0O1O100O1O1P^OS1m>mNa@f1]?[Nc@e1\\?\\Nd@d1\\?\\Nd@e1Z?\\Nf@d1Y?]Ng@c1Y?]Ng@c1X?^Nh@d1U?]Nk@g1Q?YNo@l1k>UNUAo1f>RNZAS2a>mM^AX2]>iMcA\\2W>eMiA_2S>aMmAd2m=]MSBg2h=ZMXBk2c=UM]Bo2^=RMbBQ3Z=PMfBP3Z=PMfBQ3X=PMhBP3OoM^:ROcEo2OPN]:QOdEo2NQN]:QOdEP3NoM]:ROeEo2NoM]:ROeEo2MPN]:ROfEn2MPN\\:SOgEn2LoM]:SOgEn2LPN[:SOiEm2KQN[:SOjEl2KQN[:SOjEl2KQNZ:TOkEl2JPNZ:UOlEk2IQN[:TOlEk2IQN[:TOlEk2IRN[:QOmEn2FRN]:POlEo2GQN^:oNkEP3GQN^:oNkEP3GQN_:nNjER3EQNa:mNjER3EQNb:lNiES3EQNb:lNiES3ERNb:jNiET3DSNc:iNiEU3CRNe:hNhEV3CRNf:gNgEW3BSNg:fNgEW3BdMW;TOWEY3AnLm;IaDZ3BXLc<=kC[3S=eLmB[3T=dLlB]3S=cLmBQ3`=nL`Bb2P>^MPBR2a>mM_Ad1P?\\NPAd1Q?[No@e1Q?[No@e1R?ZNn@f1R?ZNn@f1S?YNm@f1T?ZNk@g1V?XNj@h1V?XNj@h1W?WNi@i1W?WNi@i1X?VNh@j1X?VNh@i1Z?VNf@j1Z?VNf@j1[?UNe@k1[?UNe@k1\\?TNd@l1]?SNc@n1\\?RNc@Q2\\?nMd@U2Y?kMg@W2X?hMh@Z2V?fMj@\\2U?cMk@_2S?aMm@b2Q?]Mo@e2o>[MQAg2n>XMRAj2l>UMUAn2i>QMWAQ3g>oLYAS3f>lLYAW3e>iL[AY3d>fL\\A]3a>cL_A_3`>`L`Ab3^>^LbAd3]>[LcAg3[>YLeAj3Y>ULgAm3W>SLiAo3V>PLjAR4T>nKlAT4S>kKmAX4P>hKPBZ4o=eKPB^4n=bKRB_4n=`KRB`4o=_KQBa4:`Kb in this image.", "objects": [{"patches": [41, 42, 57, 58, 59, 60, 74, 75, 76, 77, 78, 93, 94, 96, 97, 98, 113, 114, 115, 116, 130, 131, 132, 133, 134, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 215, 216, 217, 218, 219, 220, 233, 236, 249, 250, 253, 254, 267, 270, 271, 287], "bbox": [0.3598559670781893, 0.08671875, 0.9712551440329218, 0.714640625], "iscrowd": 0, "area": 29818.398400000002, "rle": {"size": [640, 486], "counts": "\\_]32mc03N2M3M3M3M3M3M3M3M100O001O1O001O001O10O01O0O2N2N101N1O2N2N1O2N1O2N2N1O2N1O2N20O2O0O2O1O0O2O1N101O1N10?A02N2M2O2N2N2M2O2N2M3N1O2M3N2N1N3N2N2M2O2N2M3N1OhMj^Oh1Sa0QNV_OP2]a0O1N2O1N2N2N2O1N2N2N2O1N2N1O2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N101N2N2N2O1N2N2N2O`U==SjB0O1O100O1O1P^OS1m>mNa@f1]?[Nc@e1\\?\\Nd@d1\\?\\Nd@e1Z?\\Nf@d1Y?]Ng@c1Y?]Ng@c1X?^Nh@d1U?]Nk@g1Q?YNo@l1k>UNUAo1f>RNZAS2a>mM^AX2]>iMcA\\2W>eMiA_2S>aMmAd2m=]MSBg2h=ZMXBk2c=UM]Bo2^=RMbBQ3Z=PMfBP3Z=PMfBQ3X=PMhBP3OoM^:ROcEo2OPN]:QOdEo2NQN]:QOdEP3NoM]:ROeEo2NoM]:ROeEo2MPN]:ROfEn2MPN\\:SOgEn2LoM]:SOgEn2LPN[:SOiEm2KQN[:SOjEl2KQN[:SOjEl2KQNZ:TOkEl2JPNZ:UOlEk2IQN[:TOlEk2IQN[:TOlEk2IRN[:QOmEn2FRN]:POlEo2GQN^:oNkEP3GQN^:oNkEP3GQN_:nNjER3EQNa:mNjER3EQNb:lNiES3EQNb:lNiES3ERNb:jNiET3DSNc:iNiEU3CRNe:hNhEV3CRNf:gNgEW3BSNg:fNgEW3BdMW;TOWEY3AnLm;IaDZ3BXLc<=kC[3S=eLmB[3T=dLlB]3S=cLmBQ3`=nL`Bb2P>^MPBR2a>mM_Ad1P?\\NPAd1Q?[No@e1Q?[No@e1R?ZNn@f1R?ZNn@f1S?YNm@f1T?ZNk@g1V?XNj@h1V?XNj@h1W?WNi@i1W?WNi@i1X?VNh@j1X?VNh@i1Z?VNf@j1Z?VNf@j1[?UNe@k1[?UNe@k1\\?TNd@l1]?SNc@n1\\?RNc@Q2\\?nMd@U2Y?kMg@W2X?hMh@Z2V?fMj@\\2U?cMk@_2S?aMm@b2Q?]Mo@e2o>[MQAg2n>XMRAj2l>UMUAn2i>QMWAQ3g>oLYAS3f>lLYAW3e>iL[AY3d>fL\\A]3a>cL_A_3`>`L`Ab3^>^LbAd3]>[LcAg3[>YLeAj3Y>ULgAm3W>SLiAo3V>PLjAR4T>nKlAT4S>kKmAX4P>hKPBZ4o=eKPB^4n=bKRB_4n=`KRB`4o=_KQBa4:`Kb in this image.", "objects": [{"patches": [107, 108, 124, 125, 128, 129, 130, 131, 146, 147, 148, 149, 150, 151, 152, 153, 154, 169, 170, 171, 172, 173, 174, 175, 176, 177, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 219, 220, 237, 238, 239, 240, 242, 243, 261, 262, 263, 264, 265, 266], "bbox": [0.332875, 0.2936768149882904, 0.7386250000000001, 0.8075878220140515], "iscrowd": 0, "area": 21556.370300000006, "rle": {"size": [427, 640], "counts": "TSi23k<=I7M3N2O1DTOWDn0R<1FQOVDQ1j;oNUDR1j;oNUDR1k;nNSDT1l;lNTDU1l;kNRDW1m;400O100O100OO2O0N3LdDZOZ:c0eEDY:h1bNkMdF5=S2k8aNPGa1m8eNmF]1P9b1L3M2O2M3OROhGjLV8Q3oGQMo7m2SHTMl7i2eGlL3QGHm88PGMn83oF2P9V2N2N1O2OUOWGUMf8f2cG[MY8^2QHcMj7Y2\\HjM`7Q2fHQNV7l1nHVNo6g1UI[Ni6a1\\I^Nc6`1aI_N_6^1eIaNZ6]1jIbNU6[1oIeNP6Y1TJfNk5X1ZJfNf5V1_JhNa5X1cJcN^5^1fJ[N\\5e1U31O1O001O010001NeGSN[5k1bJVNb5g1]JZNg5b1XJ_NR6R1QJoNR6j0SJVOn5e0VJ[Ol5>XJBi900O010N2O0000O10000mMHSG9i8MUG3f84WGMe88ZGHb8>[GD`8a0_G_O]8g0`GZOS8T1kGmNh7b1VH^N^7P2_HRN_7P2_HQNa7Q2\\HPNc7S2ZHnMf7S2YHmMf7V2UHnMj7S2QHQNo7Q2lGRNT8P2eGUN[8]3001N2N2O1N2N2O1N2N2O1N23M3M3L3M4M3[GnKj7n40LOnMSHiNk7T1_HgN_7U1lHfNS7V1WIdNh6Z1`IbN^6_1hI[NW6f1gJ\\MX5d2jJZMV5e2nJYMR5f2PKYMo4g2TKWMl4g2WKXMh4h2[KVMe4h2P3N1O2N1O2N2O1N2N101N2N2O1N00O1O1O2001O0O101O001N101O001N2O001N2N1O2O1N1010000000O1O1N2N2O1N3N3O0001O1O1O001O1N10N101N8I0O1N2N2N2M3c0]O5K2N3M2N3dNWDo0S<01O1N1O2O0^OiCOY in this image.", "objects": [{"patches": [107, 108, 124, 125, 128, 129, 130, 131, 146, 147, 148, 149, 150, 151, 152, 153, 154, 169, 170, 171, 172, 173, 174, 175, 176, 177, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 219, 220, 237, 238, 239, 240, 242, 243, 261, 262, 263, 264, 265, 266], "bbox": [0.332875, 0.2936768149882904, 0.7386250000000001, 0.8075878220140515], "iscrowd": 0, "area": 21556.370300000006, "rle": {"size": [427, 640], "counts": "TSi23k<=I7M3N2O1DTOWDn0R<1FQOVDQ1j;oNUDR1j;oNUDR1k;nNSDT1l;lNTDU1l;kNRDW1m;400O100O100OO2O0N3LdDZOZ:c0eEDY:h1bNkMdF5=S2k8aNPGa1m8eNmF]1P9b1L3M2O2M3OROhGjLV8Q3oGQMo7m2SHTMl7i2eGlL3QGHm88PGMn83oF2P9V2N2N1O2OUOWGUMf8f2cG[MY8^2QHcMj7Y2\\HjM`7Q2fHQNV7l1nHVNo6g1UI[Ni6a1\\I^Nc6`1aI_N_6^1eIaNZ6]1jIbNU6[1oIeNP6Y1TJfNk5X1ZJfNf5V1_JhNa5X1cJcN^5^1fJ[N\\5e1U31O1O001O010001NeGSN[5k1bJVNb5g1]JZNg5b1XJ_NR6R1QJoNR6j0SJVOn5e0VJ[Ol5>XJBi900O010N2O0000O10000mMHSG9i8MUG3f84WGMe88ZGHb8>[GD`8a0_G_O]8g0`GZOS8T1kGmNh7b1VH^N^7P2_HRN_7P2_HQNa7Q2\\HPNc7S2ZHnMf7S2YHmMf7V2UHnMj7S2QHQNo7Q2lGRNT8P2eGUN[8]3001N2N2O1N2N2O1N2N2O1N23M3M3L3M4M3[GnKj7n40LOnMSHiNk7T1_HgN_7U1lHfNS7V1WIdNh6Z1`IbN^6_1hI[NW6f1gJ\\MX5d2jJZMV5e2nJYMR5f2PKYMo4g2TKWMl4g2WKXMh4h2[KVMe4h2P3N1O2N1O2N2O1N2N101N2N2O1N00O1O1O2001O0O101O001N101O001N2O001N2N1O2O1N1010000000O1O1N2N2O1N3N3O0001O1O1O001O1N10N101N8I0O1N2N2N2M3c0]O5K2N3M2N3dNWDo0S<01O1N1O2O0^OiCOY in this image.", "objects": [{"patches": [154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 246, 247, 248, 249], "bbox": [0.644796875, 0.4164788732394366, 0.882171875, 0.7307746478873238], "iscrowd": 0, "area": 14967.34325, "rle": {"size": [426, 640], "counts": "nP\\5c0e<5K5L2M4M2`DPOd:Q1ZEVO_:l0`ESOa:o0\\EROc:P1\\EPOd:Q1ZEPOf:Q1YEoNf:T1WEmNi:T1VElNj:U1TElNl:U1SEkNm:W1PEiNQ;X1nDhNR;h10000000000000O100000000001^EjMn9W2QFmMk9T2SFRNh9P2VFSNg9n1XFRNh9o1VFRNi9Q2UFoMj9S2UFmMj9U2TFlMk9W2SFiMl9Y2RFhMn9j2O1O1O1O1O1N2N2M2O2N2N2N200O1O100O1O100O100O1O1001O0O100000000000000000000O2O00000O10001O000O101O00000O2O0000001N1000001O0O10001N10001N100O101O0O10001N10000O2O0O101O0O2O000O2N100O2N1O101N1O1O1O101N1O1O1^Ob0]Oc0mN\\D?e;[OeD?];]OjD?W;^OPE in this image.", "objects": [{"patches": [154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 246, 247, 248, 249], "bbox": [0.644796875, 0.4164788732394366, 0.882171875, 0.7307746478873238], "iscrowd": 0, "area": 14967.34325, "rle": {"size": [426, 640], "counts": "nP\\5c0e<5K5L2M4M2`DPOd:Q1ZEVO_:l0`ESOa:o0\\EROc:P1\\EPOd:Q1ZEPOf:Q1YEoNf:T1WEmNi:T1VElNj:U1TElNl:U1SEkNm:W1PEiNQ;X1nDhNR;h10000000000000O100000000001^EjMn9W2QFmMk9T2SFRNh9P2VFSNg9n1XFRNh9o1VFRNi9Q2UFoMj9S2UFmMj9U2TFlMk9W2SFiMl9Y2RFhMn9j2O1O1O1O1O1N2N2M2O2N2N2N200O1O100O1O100O100O1O1001O0O100000000000000000000O2O00000O10001O000O101O00000O2O0000001N1000001O0O10001N10001N100O101O0O10001N10000O2O0O101O0O2O000O2N100O2N1O101N1O1O1O101N1O1O1^Ob0]Oc0mN\\D?e;[OeD?];]OjD?W;^OPE in this image.", "objects": [{"patches": [89, 111, 112, 113, 134, 135, 136, 156, 157, 178, 179, 180, 201, 202, 203, 224, 225, 247, 248, 270, 271], "bbox": [0.744921875, 0.21783333333333335, 0.93284375, 0.6896249999999999], "iscrowd": 0, "area": 7151.045550000005, "rle": {"size": [480, 640], "counts": "jio6?B=Ch0XO2N1O2N2Mh0YO5K5K5K4L5KN1L5XOh05K4L5K8H8H7I3N0O]NjE_NT:[1[F]Nd9b1aF\\N]9e1fFXNY9i1jFUNT9l1oFRNo8o1VGlMi8U2P25K4L4La1_N4L4L4L3M1OVNWF\\Nk9a1k1N2N2N2M3N3M2O1N2O1N2O1H8C=F:N2M3N2O1O1O1O1O1O1O1O1O1O1ON2O1N2O2N2O010000O1000000O1000000O10N2N2N2011N101N10100O01000O010O10O0100O4M3L3N2N1N101N101L3L5KgPd0"}, "label": "giraffe in front of barn door"}]} {"id": 50101, "image": "COCO_train2014_000000050101.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe closest to the door\"."}], "task": "refering", "answer_template": "The \"a giraffe closest to the door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 111, 112, 113, 134, 135, 136, 156, 157, 178, 179, 180, 201, 202, 203, 224, 225, 247, 248, 270, 271], "bbox": [0.744921875, 0.21783333333333335, 0.93284375, 0.6896249999999999], "iscrowd": 0, "area": 7151.045550000005, "rle": {"size": [480, 640], "counts": "jio6?B=Ch0XO2N1O2N2Mh0YO5K5K5K4L5KN1L5XOh05K4L5K8H8H7I3N0O]NjE_NT:[1[F]Nd9b1aF\\N]9e1fFXNY9i1jFUNT9l1oFRNo8o1VGlMi8U2P25K4L4La1_N4L4L4L3M1OVNWF\\Nk9a1k1N2N2N2M3N3M2O1N2O1N2O1H8C=F:N2M3N2O1O1O1O1O1O1O1O1O1O1ON2O1N2O2N2O010000O1000000O1000000O10N2N2N2011N101N10100O01000O010O10O0100O4M3L3N2N1N101N101L3L5KgPd0"}, "label": "a giraffe closest to the door"}]} {"id": 500662, "image": "COCO_train2014_000000500662.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue piece of luggage with a blue pattern on it\"."}], "task": "refering", "answer_template": "The \"a blue piece of luggage with a blue pattern on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 219, 220, 221, 222, 234, 235, 236, 237, 238, 239, 240, 252, 253, 254, 255, 256, 257, 258, 270, 271, 272, 273, 274, 275, 276, 288, 289, 290, 291, 292, 293, 294, 306, 307, 308, 309, 310, 311, 312, 324, 325, 326, 327, 328, 329, 330, 331, 342, 343, 344, 345, 346, 347, 348, 349, 360, 361, 362, 363, 364, 365, 366, 367, 378, 379, 380], "bbox": [0.005748502994011976, 0.5370781250000001, 0.4018962075848303, 0.9213437500000001], "iscrowd": 0, "area": 43783.952999999994, "rle": {"size": [640, 501], "counts": "d\\2j1d?a2_Mb2XOh0000000000000000000001O000000000O1000000000000000001O0000000000000O100000000000001O00000000000000000O1000000000001O0000000000000000000O100000001O00000000000000000000000O1000001O0000000000000000000000000O101O00000000000000000000000000001N1000000000000000000000000000001O0O100000000000000000000000001O00000O1000000000000000000000001O00000O10000O10000O4M;D in this image.", "objects": [{"patches": [216, 217, 218, 219, 220, 221, 222, 234, 235, 236, 237, 238, 239, 240, 252, 253, 254, 255, 256, 257, 258, 270, 271, 272, 273, 274, 275, 276, 288, 289, 290, 291, 292, 293, 294, 306, 307, 308, 309, 310, 311, 312, 324, 325, 326, 327, 328, 329, 330, 331, 342, 343, 344, 345, 346, 347, 348, 349, 360, 361, 362, 363, 364, 365, 366, 367, 378, 379, 380], "bbox": [0.005748502994011976, 0.5370781250000001, 0.4018962075848303, 0.9213437500000001], "iscrowd": 0, "area": 43783.952999999994, "rle": {"size": [640, 501], "counts": "d\\2j1d?a2_Mb2XOh0000000000000000000001O000000000O1000000000000000001O0000000000000O100000000000001O00000000000000000O1000000000001O0000000000000000000O100000001O00000000000000000000000O1000001O0000000000000000000000000O101O00000000000000000000000000001N1000000000000000000000000000001O0O100000000000000000000000001O00000O1000000000000000000000001O00000O10000O10000O4M;D in this image.", "objects": [{"patches": [47, 48, 49, 50, 51, 52, 53, 65, 66, 67, 68, 69, 70, 71, 83, 84, 85, 86, 87, 88, 89, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 143], "bbox": [0.6370658682634731, 0.096109375, 0.9987225548902195, 0.348046875], "iscrowd": 0, "area": 27106.882000000005, "rle": {"size": [640, 501], "counts": "T_W6Z3d`03L3M4L4L3N3L3M4L4M201O001O1O001O001O1N101O001O001O1O001O00000000000O1000O1000000000000000000000O10000000000000000000000000000O10000000O100000000000000000O10000000000000000000000000000O100000000000O10000000000000O100000000000000000000000000O1000000000000000O100000000000O100000000000000000000000000O10000000000000000000O10000000O1000000000000000000000000f0ZO^1bN^1bN`>"}, "label": "luggage with colorful stripes on a green table"}]} {"id": 500662, "image": "COCO_train2014_000000500662.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a striped suitcase\"."}], "task": "refering", "answer_template": "The \"a striped suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 51, 52, 53, 65, 66, 67, 68, 69, 70, 71, 83, 84, 85, 86, 87, 88, 89, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 143], "bbox": [0.6370658682634731, 0.096109375, 0.9987225548902195, 0.348046875], "iscrowd": 0, "area": 27106.882000000005, "rle": {"size": [640, 501], "counts": "T_W6Z3d`03L3M4L4L3N3L3M4L4M201O001O1O001O001O1N101O001O001O1O001O00000000000O1000O1000000000000000000000O10000000000000000000000000000O10000000O100000000000000000O10000000000000000000000000000O100000000000O10000000000000O100000000000000000000000000O1000000000000000O100000000000O100000000000000000000000000O10000000000000000000O10000000O1000000000000000000000000f0ZO^1bN^1bN`>"}, "label": "a striped suitcase"}]} {"id": 500662, "image": "COCO_train2014_000000500662.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a suitcase with blue and red squares\"."}], "task": "refering", "answer_template": "The \"a suitcase with blue and red squares\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 240, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 295, 296, 297, 298, 299, 313, 314, 315, 316, 317, 331, 332, 333, 334, 335], "bbox": [0.10672654690618762, 0.48782812499999995, 0.669441117764471, 0.8296093750000001], "iscrowd": 0, "area": 31658.931800000002, "rle": {"size": [640, 501], "counts": "mRR12mc01N2N2O1N2N2O1N2N2O1N2N2O1O1O10000000000O1000000000000O10000000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O10000000000000000O100000000000000O10000000000000000O100000000000000O10000000000000000O100000000000000O1000000000001O8H9F9H9G8H9G8H9G8H9G8I8G8H9G8H9G8H9G8H9G8H1O00O10000000000O100000001O0O10000000000O10000000000O10000000000O10001O000O10000000000O10000000000O10000000000O10001O00000O100000000O10000000000O10000000000O10001O00000O10000000000O10001O1O1O0O2O1O1O001N2O1O001O1O1N101O1O001O1N2O004L;E:F;D in this image.", "objects": [{"patches": [200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 240, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 295, 296, 297, 298, 299, 313, 314, 315, 316, 317, 331, 332, 333, 334, 335], "bbox": [0.10672654690618762, 0.48782812499999995, 0.669441117764471, 0.8296093750000001], "iscrowd": 0, "area": 31658.931800000002, "rle": {"size": [640, 501], "counts": "mRR12mc01N2N2O1N2N2O1N2N2O1N2N2O1O1O10000000000O1000000000000O10000000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O10000000000000000O100000000000000O10000000000000000O100000000000000O10000000000000000O100000000000000O1000000000001O8H9F9H9G8H9G8H9G8H9G8I8G8H9G8H9G8H9G8H9G8H1O00O10000000000O100000001O0O10000000000O10000000000O10000000000O10001O000O10000000000O10000000000O10000000000O10001O00000O100000000O10000000000O10000000000O10001O00000O10000000000O10001O1O1O0O2O1O1O001N2O1O001O1O1N101O1O001O1N2O004L;E:F;D in this image.", "objects": [{"patches": [176, 177, 199, 200, 201, 222, 223, 224, 225, 246, 247, 248, 249, 269, 270, 271, 272, 273, 274, 293, 294, 295, 296, 297], "bbox": [0.661515625, 0.40847916666666667, 0.931625, 0.7673333333333333], "iscrowd": 0, "area": 12366.949550000003, "rle": {"size": [480, 640], "counts": "V`V62i>a0_O8H7L4M3L4L5L3L4L4L4M4K4L4N2N2O0O1O1O2O0O1O1O2O0O1O2O0O1O1O2O0O1O1O2O0O1O2N100O2N1O21N4M2M3M3NO010O00010O010O01O001N1000001N10000O10001O0O1000000O101O0O1000000O100O100O10O0100O10000O10O0100O100O10O0100O100O100O010O100O1000O0100O100O100O001O1O1N2O1O0O2O1O1N2O1O001N2O1O0O2O1O001O1O010O001O1O001O1O001O001N101O001O0O2O001O001N2O1N2O1N2N2O1N2N2OaYd0"}, "label": "the fry on top of the pile of fries"}]} {"id": 205757, "image": "COCO_train2014_000000205757.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the french fry on top of the stack with 4 darker lines on it\"."}], "task": "refering", "answer_template": "The \"the french fry on top of the stack with 4 darker lines on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 199, 200, 201, 222, 223, 224, 225, 246, 247, 248, 249, 269, 270, 271, 272, 273, 274, 293, 294, 295, 296, 297], "bbox": [0.661515625, 0.40847916666666667, 0.931625, 0.7673333333333333], "iscrowd": 0, "area": 12366.949550000003, "rle": {"size": [480, 640], "counts": "V`V62i>a0_O8H7L4M3L4L5L3L4L4L4M4K4L4N2N2O0O1O1O2O0O1O1O2O0O1O2O0O1O1O2O0O1O1O2O0O1O2N100O2N1O21N4M2M3M3NO010O00010O010O01O001N1000001N10000O10001O0O1000000O101O0O1000000O100O100O10O0100O10000O10O0100O100O10O0100O100O100O010O100O1000O0100O100O100O001O1O1N2O1O0O2O1O1N2O1O001N2O1O0O2O1O001O1O010O001O1O001O1O001O001N101O001O0O2O001O001N2O1N2O1N2N2O1N2N2OaYd0"}, "label": "the french fry on top of the stack with 4 darker lines on it"}]} {"id": 156608, "image": "COCO_train2014_000000156608.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in light jeans\"."}], "task": "refering", "answer_template": "The \"man in light jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 31, 32, 33, 34, 35, 36, 53, 54, 55, 56, 57, 58, 76, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 104, 123, 124, 125, 126, 127, 146, 147, 148, 149, 169, 170, 171, 172, 192, 193, 194], "bbox": [0.32984375, 0.006370023419203748, 0.58921875, 0.5748946135831382], "iscrowd": 0, "area": 27051.62035, "rle": {"size": [427, 640], "counts": "lQh27R=f0[OL3M3N3L3ETKQHQ5l7PKoGV5o75N3M2jI_J^4c5^KaJ`4b5[KaJd4a5XKbJg4a5SKcJl4_5oJeJP5]5lJfJS5]5gJgJX5[5cJiJ\\5Y5`JiJ`5Z5ZJiJf5^6O10000000O100000000000000O100O100O100O100O10000O100O1O1O1O1O1O1O1O1O1N200O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1N2O1N2N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2N2O1N2N2WO]GhLe8T3k0M3L4L4L4M3L4L4M3L4L4L4K5L4L4L4L4L4L4L4L4K5L4L4L4K5L4K5L4K_e]3"}, "label": "man in light jeans"}]} {"id": 156608, "image": "COCO_train2014_000000156608.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing stone wash jeans sitting next to a guy with two blue bud light beers and a red candle that is lit\"."}], "task": "refering", "answer_template": "The \"a man wearing stone wash jeans sitting next to a guy with two blue bud light beers and a red candle that is lit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 31, 32, 33, 34, 35, 36, 53, 54, 55, 56, 57, 58, 76, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 104, 123, 124, 125, 126, 127, 146, 147, 148, 149, 169, 170, 171, 172, 192, 193, 194], "bbox": [0.32984375, 0.006370023419203748, 0.58921875, 0.5748946135831382], "iscrowd": 0, "area": 27051.62035, "rle": {"size": [427, 640], "counts": "lQh27R=f0[OL3M3N3L3ETKQHQ5l7PKoGV5o75N3M2jI_J^4c5^KaJ`4b5[KaJd4a5XKbJg4a5SKcJl4_5oJeJP5]5lJfJS5]5gJgJX5[5cJiJ\\5Y5`JiJ`5Z5ZJiJf5^6O10000000O100000000000000O100O100O100O100O10000O100O1O1O1O1O1O1O1O1O1N200O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1N2O1N2N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2N2O1N2N2WO]GhLe8T3k0M3L4L4L4M3L4L4M3L4L4L4K5L4L4L4L4L4L4L4L4K5L4L4L4K5L4K5L4K_e]3"}, "label": "a man wearing stone wash jeans sitting next to a guy with two blue bud light beers and a red candle that is lit"}]} {"id": 156608, "image": "COCO_train2014_000000156608.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man drinking beer\"."}], "task": "refering", "answer_template": "The \"a man drinking beer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 36, 37, 38, 39, 40, 41, 42, 58, 59, 60, 61, 62, 63, 64, 65, 81, 82, 83, 84, 85, 86, 104, 105, 106, 108, 109, 110, 127, 128, 131, 132, 133, 150, 151, 153, 154, 155, 173, 174, 176, 177, 196, 197, 199, 200, 219, 220, 222, 223], "bbox": [0.5277499999999999, 0.0038173302107728335, 0.848609375, 0.659976580796253], "iscrowd": 0, "area": 29317.15364999999, "rle": {"size": [427, 640], "counts": "SU]4:g<=B>C>G8H8I7I7H6K6J5J6J7G8G9H9F9ZOf0B>G9M2M3N2N3M2M3N2N3L3N2N2M4L3M3M3M4L3L4M3M4L3M3M3M4LgMXLUKe3k4fLmJV3S5UMfJk2W5_McJ`2Y5hMcJX2Z5nMcJR2Y5TNeJl1X5YNgJf1\\4S4N2N2M3N2N2N2M3N2N2N2N2000000O10000O10000O100000PJ^LW2b3gMaLX2^3gMeLX2[3fMhLY2X3eMkLZ2U3dMmL\\2S3bMPM]2P3aMSM^2l2aMWM^2i2`MYM`2g2^M\\Ma2d2]M^Mc2b2[M`Me2_2ZMdMe2\\2YMfMg2Y2XMiMh2V2WMmMh2R2XMoMh2P2WMRNi2m1WMUNh2k1VMWNj2h1VMYNcM^Ni4X3dM[NaM`Nj4T3dM_N`M_Nk4S3cM_N`M`Nm4R3aM_N`MbNn4P3_MaN`MbNP5m2_MROa2o0\\MTOc2m0[MUOd2l0ZMUOf2l0WMWOh2j0VMXOi2h0VMdNnM_Nl4n2SMcNTN_Nh4o2RMbNYN]Ne4R3oLaN_N]Na4S3nL`NdN[N^4V3kL_NjN[NZ4V3kL_NnNYNW4Y3hL^NTOYNS4Z3fL^NZOVNP4]3cL]N@VNl3^3aL\\NGTNh3a3^L[NMTNd3b3\\LZN4RN`3d3ZLZN9QN]3f3XLXN?QNX3h3VLWNe0PNU3j3SLVNl0oMP3l3QLUNR1nMm2n3nKSNZ1nMg2P4lKRN`1mMd2R4iKQNZ6P2dIPN]6Q2`IPNa6Q2\\IoMf6R2WIoMj6X41N2O1O1N2O1O1N2O1`LWI=j6AYI in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 36, 37, 38, 39, 40, 41, 42, 58, 59, 60, 61, 62, 63, 64, 65, 81, 82, 83, 84, 85, 86, 104, 105, 106, 108, 109, 110, 127, 128, 131, 132, 133, 150, 151, 153, 154, 155, 173, 174, 176, 177, 196, 197, 199, 200, 219, 220, 222, 223], "bbox": [0.5277499999999999, 0.0038173302107728335, 0.848609375, 0.659976580796253], "iscrowd": 0, "area": 29317.15364999999, "rle": {"size": [427, 640], "counts": "SU]4:g<=B>C>G8H8I7I7H6K6J5J6J7G8G9H9F9ZOf0B>G9M2M3N2N3M2M3N2N3L3N2N2M4L3M3M3M4L3L4M3M4L3M3M3M4LgMXLUKe3k4fLmJV3S5UMfJk2W5_McJ`2Y5hMcJX2Z5nMcJR2Y5TNeJl1X5YNgJf1\\4S4N2N2M3N2N2N2M3N2N2N2N2000000O10000O10000O100000PJ^LW2b3gMaLX2^3gMeLX2[3fMhLY2X3eMkLZ2U3dMmL\\2S3bMPM]2P3aMSM^2l2aMWM^2i2`MYM`2g2^M\\Ma2d2]M^Mc2b2[M`Me2_2ZMdMe2\\2YMfMg2Y2XMiMh2V2WMmMh2R2XMoMh2P2WMRNi2m1WMUNh2k1VMWNj2h1VMYNcM^Ni4X3dM[NaM`Nj4T3dM_N`M_Nk4S3cM_N`M`Nm4R3aM_N`MbNn4P3_MaN`MbNP5m2_MROa2o0\\MTOc2m0[MUOd2l0ZMUOf2l0WMWOh2j0VMXOi2h0VMdNnM_Nl4n2SMcNTN_Nh4o2RMbNYN]Ne4R3oLaN_N]Na4S3nL`NdN[N^4V3kL_NjN[NZ4V3kL_NnNYNW4Y3hL^NTOYNS4Z3fL^NZOVNP4]3cL]N@VNl3^3aL\\NGTNh3a3^L[NMTNd3b3\\LZN4RN`3d3ZLZN9QN]3f3XLXN?QNX3h3VLWNe0PNU3j3SLVNl0oMP3l3QLUNR1nMm2n3nKSNZ1nMg2P4lKRN`1mMd2R4iKQNZ6P2dIPN]6Q2`IPNa6Q2\\IoMf6R2WIoMj6X41N2O1O1N2O1O1N2O1`LWI=j6AYI in this image.", "objects": [{"patches": [154, 155, 156, 157, 176, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 270, 271], "bbox": [0.688125, 0.44304449648711947, 0.9529062500000001, 0.7614285714285715], "iscrowd": 0, "area": 16508.52345, "rle": {"size": [427, 640], "counts": "RSh52X=3M5K6K5J6J6J6J6J6J5K6K5J8I8H8G8I8H8G10O100O1O100O1O1O010O1O1O11O2N2N2N3M2N2N2N2N2N2M3N0000O10000O10002N2M2O2N2N1N3N2N1N_OdLWGZ3i8lLRGS3n8SMnFk2S9ZMiFc2X9cMgFX2Y9m0000O10000O101O000O1000000O10000O2O00000O1000000O10001N1000000O2O3M2N2M4M1O0O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O10001O1N2O1O1O1bM^EP2d:nM]ER2d:lM]ET2d:jM^EU2c:hM_EX2k:N10001O001N10001O0_NlDj0V;ROnDm0R;POREo0n:nNUER1l:jNXEU1h:hN\\EV1^;O001O1O5J:G9G:FXW<"}, "label": "a skewered half sandwich with its other half between it and a glass of red wine"}]} {"id": 66518, "image": "COCO_train2014_000000066518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of a sandwich to the right of another half\"."}], "task": "refering", "answer_template": "The \"half of a sandwich to the right of another half\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 176, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 270, 271], "bbox": [0.688125, 0.44304449648711947, 0.9529062500000001, 0.7614285714285715], "iscrowd": 0, "area": 16508.52345, "rle": {"size": [427, 640], "counts": "RSh52X=3M5K6K5J6J6J6J6J6J5K6K5J8I8H8G8I8H8G10O100O1O100O1O1O010O1O1O11O2N2N2N3M2N2N2N2N2N2M3N0000O10000O10002N2M2O2N2N1N3N2N1N_OdLWGZ3i8lLRGS3n8SMnFk2S9ZMiFc2X9cMgFX2Y9m0000O10000O101O000O1000000O10000O2O00000O1000000O10001N1000000O2O3M2N2M4M1O0O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O10001O1N2O1O1O1bM^EP2d:nM]ER2d:lM]ET2d:jM^EU2c:hM_EX2k:N10001O001N10001O0_NlDj0V;ROnDm0R;POREo0n:nNUER1l:jNXEU1h:hN\\EV1^;O001O1O5J:G9G:FXW<"}, "label": "half of a sandwich to the right of another half"}]} {"id": 66518, "image": "COCO_train2014_000000066518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the slightly smaller portion of a sandwich , with a darker crust\"."}], "task": "refering", "answer_template": "The \"the slightly smaller portion of a sandwich , with a darker crust\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246], "bbox": [0.48946875, 0.4734660421545667, 0.716890625, 0.7353161592505856], "iscrowd": 0, "area": 13176.250000000005, "rle": {"size": [427, 640], "counts": "VhR43W=8H9G6QDWOQ;j0lDYOS;i0iD[OU;g0gD\\OX;e0eD_OY;c0cD@\\;a0aDC];V1N3N1O2N1O2N1O2N2N1O2N1O2N1O2N1N3O0O2O0O2O0O101N100O100O2O0O100O101N100O100O2O000O10001O000000000000000000000000000000000O10000000000000000000000000O2O0000000000000000000000000001O000000000000001O01O0000000001O000000000000001O0003M3M3M3M4L3N2M3M3M4L3M3M3N3L6J6J6H9E:E;F:FWV[2"}, "label": "the slightly smaller portion of a sandwich , with a darker crust"}]} {"id": 66518, "image": "COCO_train2014_000000066518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left half of the sandwhich\"."}], "task": "refering", "answer_template": "The \"the left half of the sandwhich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246], "bbox": [0.48946875, 0.4734660421545667, 0.716890625, 0.7353161592505856], "iscrowd": 0, "area": 13176.250000000005, "rle": {"size": [427, 640], "counts": "VhR43W=8H9G6QDWOQ;j0lDYOS;i0iD[OU;g0gD\\OX;e0eD_OY;c0cD@\\;a0aDC];V1N3N1O2N1O2N1O2N2N1O2N1O2N1O2N1N3O0O2O0O2O0O101N100O100O2O0O100O101N100O100O2O000O10001O000000000000000000000000000000000O10000000000000000000000000O2O0000000000000000000000000001O000000000000001O01O0000000001O000000000000001O0003M3M3M3M4L3N2M3M3M4L3M3M3N3L6J6J6H9E:E;F:FWV[2"}, "label": "the left half of the sandwhich"}]} {"id": 312282, "image": "COCO_train2014_000000312282.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grilled sandwich cut in half\"."}], "task": "refering", "answer_template": "The \"a grilled sandwich cut in half\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 37, 38, 39, 40, 41, 55, 56, 57, 58, 59, 60, 61, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 93, 94, 95, 96, 97, 98, 99, 100, 101, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 133, 134, 135, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 194], "bbox": [0.04222, 0.11749333333333334, 0.88514, 0.7804], "iscrowd": 0, "area": 38964.009550000024, "rle": {"size": [375, 500], "counts": "ki71d;3L3N3L3N3L3N3L3N2M4M2O2O001N101O001O001O001O0AjN]FV1a9lN_FU1]9oNbFQ1[9ROeFo0Y9SOfFm0Z9SOfFn0Y9SOfFm0Z9SOfFn0X9TOgFl0V9WOjFj0R9ZOmFf0Q9_OlFb0R9R1N2O2N1N3N1N3N1N2O2O010O00010O0010O01O01O01O0010O01O01O01O010O00010O0010O01O01O01O010O0010O00010O01O010O00010O0010O0001O010O0010O00010O01O010O100O100O100O2O0O100O100O01O10O01O010O001O010O0010O01O010O0010O01O010O0010O01O1O010O0010O01O010O0010O01O010O0010O01O0010O01O010O1O010O0010O01O010O0010O01O010O0100O010O100O01000O010O10O0100O100O010O100O10O0100O100O010O100O010O100O10O0100O100O010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O01O010O00010O0010O0010O0010O00010O0010O00010O0001O01O01O010O00010O00010O001O01O01O01O01O010O000010O00010O00010O01O01O01O00010O0010O0001O0O1N3M2N3M2O1N3M2N2O2M2H8F;O001O00001O00001O001O00001O0N2M5L5K4K6J6J6J5K6Jodd0"}, "label": "a grilled sandwich cut in half"}]} {"id": 312282, "image": "COCO_train2014_000000312282.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwish on a plate , on the table\"."}], "task": "refering", "answer_template": "The \"a sandwish on a plate , on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 37, 38, 39, 40, 41, 55, 56, 57, 58, 59, 60, 61, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 93, 94, 95, 96, 97, 98, 99, 100, 101, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 133, 134, 135, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 194], "bbox": [0.04222, 0.11749333333333334, 0.88514, 0.7804], "iscrowd": 0, "area": 38964.009550000024, "rle": {"size": [375, 500], "counts": "ki71d;3L3N3L3N3L3N3L3N2M4M2O2O001N101O001O001O001O0AjN]FV1a9lN_FU1]9oNbFQ1[9ROeFo0Y9SOfFm0Z9SOfFn0Y9SOfFm0Z9SOfFn0X9TOgFl0V9WOjFj0R9ZOmFf0Q9_OlFb0R9R1N2O2N1N3N1N3N1N2O2O010O00010O0010O01O01O01O0010O01O01O01O010O00010O0010O01O01O01O010O0010O00010O01O010O00010O0010O0001O010O0010O00010O01O010O100O100O100O2O0O100O100O01O10O01O010O001O010O0010O01O010O0010O01O010O0010O01O1O010O0010O01O010O0010O01O010O0010O01O0010O01O010O1O010O0010O01O010O0010O01O010O0100O010O100O01000O010O10O0100O100O010O100O10O0100O100O010O100O010O100O10O0100O100O010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O01O010O00010O0010O0010O0010O00010O0010O00010O0001O01O01O010O00010O00010O001O01O01O01O01O010O000010O00010O00010O01O01O01O00010O0010O0001O0O1N3M2N3M2O1N3M2N2O2M2H8F;O001O00001O00001O001O00001O0N2M5L5K4K6J6J6J5K6Jodd0"}, "label": "a sandwish on a plate , on the table"}]} {"id": 246753, "image": "COCO_train2014_000000246753.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman holding the puppy\"."}], "task": "refering", "answer_template": "The \"the woman holding the puppy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 239, 240, 241, 242, 262, 263, 264], "bbox": [0.281875, 0.25843091334894613, 0.58621875, 0.7910070257611241], "iscrowd": 0, "area": 26699.718000000004, "rle": {"size": [427, 640], "counts": "ee[23W=1N2O2M2N2O2M2O1N3M2O1N3N1N3M2O2M2O2E:F;E:G:N1O2O0O1O2N1O2N101N1O2N1O2O1N1O2O0O2N2O0O2O1N1O2O1N101N2O0O2N2O0O2O1N2N4M3L4M3L3N3L4L4M3M3M2N3M3M3M3M2N3M3M3M3M3M2N3M3M3M3M2N3M3M3M3M3M2N3M3M00O10000000000O1000001O000O2O001O001O0O101O001O000O2O001O001O000N3I6J7I6L5M2N2N3M2N3M20001N101O0O2O0O101O0O2O2N3L3N2N2M4M2M3N2N3L3N2N2M2O2N1N2O2M2O2N1N3N1O2M2O2N1N3N1N3M2N3N1N2N3M2O2M2N3M2O2M2N4Ma0^O>B4KiX^3"}, "label": "the woman holding the puppy"}]} {"id": 246753, "image": "COCO_train2014_000000246753.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hispanic female holding puppy\"."}], "task": "refering", "answer_template": "The \"hispanic female holding puppy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 239, 240, 241, 242, 262, 263, 264], "bbox": [0.281875, 0.25843091334894613, 0.58621875, 0.7910070257611241], "iscrowd": 0, "area": 26699.718000000004, "rle": {"size": [427, 640], "counts": "ee[23W=1N2O2M2N2O2M2O1N3M2O1N3N1N3M2O2M2O2E:F;E:G:N1O2O0O1O2N1O2N101N1O2N1O2O1N1O2O0O2N2O0O2O1N1O2O1N101N2O0O2N2O0O2O1N2N4M3L4M3L3N3L4L4M3M3M2N3M3M3M3M2N3M3M3M3M3M2N3M3M3M3M2N3M3M3M3M3M2N3M3M00O10000000000O1000001O000O2O001O001O0O101O001O000O2O001O001O000N3I6J7I6L5M2N2N3M2N3M20001N101O0O2O0O101O0O2O2N3L3N2N2M4M2M3N2N3L3N2N2M2O2N1N2O2M2O2N1N3N1O2M2O2N1N3N1N3M2N3N1N2N3M2O2M2N3M2O2M2N4Ma0^O>B4KiX^3"}, "label": "hispanic female holding puppy"}]} {"id": 246753, "image": "COCO_train2014_000000246753.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the white tshirt\"."}], "task": "refering", "answer_template": "The \"the man in the white tshirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 23, 24, 25, 26, 46, 47, 48, 49, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 96, 97, 98, 99, 115, 116, 117, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332], "bbox": [0.0015, 0.010140515222482436, 0.4583125, 0.9898594847775175], "iscrowd": 0, "area": 79142.84844999998, "rle": {"size": [427, 640], "counts": "^>c0R in this image.", "objects": [{"patches": [0, 1, 2, 3, 23, 24, 25, 26, 46, 47, 48, 49, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 96, 97, 98, 99, 115, 116, 117, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332], "bbox": [0.0015, 0.010140515222482436, 0.4583125, 0.9898594847775175], "iscrowd": 0, "area": 79142.84844999998, "rle": {"size": [427, 640], "counts": "^>c0R in this image.", "objects": [{"patches": [276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.0, 0.7265208333333334, 0.752328125, 0.9974791666666667], "iscrowd": 0, "area": 53927.4708, "rle": {"size": [480, 640], "counts": "m:f1Z=[2eM000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000001O000000000000001O0000000000001O0000001O1O000000000000000000001O00000000000000000000000001O000000000001O0000000000000000000000000000000000001O000000000000000000000000000000000000001O000000000000000000000000000000000000001O00000000000000000000000000000000001O00000000000000000000000000000000001O00000000000000000000000000000000001O00000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000001O001O00001O00001O00001O001O00001O001O1O001O1O1O001O1O0_NoDSOR;n0WEgNj:Y1_10O101O0O10O10O01O1O1O1O1O1O1O2N1O1O1O1O100O2N1O1O1M3M3M3M3N2M3M2N3N2M3M3MQTZ2"}, "label": "a well made bed , with a turquoise towel on it"}]} {"id": 205794, "image": "COCO_train2014_000000205794.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bed with a green towel rolled up on it\"."}], "task": "refering", "answer_template": "The \"a bed with a green towel rolled up on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.0, 0.7265208333333334, 0.752328125, 0.9974791666666667], "iscrowd": 0, "area": 53927.4708, "rle": {"size": [480, 640], "counts": "m:f1Z=[2eM000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000001O000000000000001O0000000000001O0000001O1O000000000000000000001O00000000000000000000000001O000000000001O0000000000000000000000000000000000001O000000000000000000000000000000000000001O000000000000000000000000000000000000001O00000000000000000000000000000000001O00000000000000000000000000000000001O00000000000000000000000000000000001O00000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000001O001O00001O00001O00001O001O00001O001O1O001O1O1O001O1O0_NoDSOR;n0WEgNj:Y1_10O101O0O10O10O01O1O1O1O1O1O1O2N1O1O1O1O100O2N1O1O1M3M3M3M3N2M3M2N3N2M3M3MQTZ2"}, "label": "a bed with a green towel rolled up on it"}]} {"id": 205794, "image": "COCO_train2014_000000205794.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an unmade bed with white sheets and brown blankets\"."}], "task": "refering", "answer_template": "The \"an unmade bed with white sheets and brown blankets\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 149, 150, 151, 152, 153, 154, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270], "bbox": [0.1483125, 0.37714583333333335, 0.7938125, 0.7277083333333334], "iscrowd": 0, "area": 48371.8819, "rle": {"size": [480, 640], "counts": "ji\\14k>7I7J7H7I8H6K0O1O2N100O1O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1O100O2N1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O1O1O1O1N2O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O10000000000O1000001O00000000000000000O1000000000000O100O10000O100O100O100O101O0O100O10000000000000000000000O10000000000000000000000O100000000000000000000000O01000000000000000000000000O1000000000000N2O1O1O1N2O1O1N2O1O1O10000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O00O10000000000000000000O10000000000000O100000000000000000000000000000000O10000000000000000000000000000000000O100000000000000000000000000000000O10000000000000000000000000000000000O100000000000000000000000000000O1M2N3M3M3M3M3M4N20000010O2N2N2N2N3N0O2N2N1O2N101N2N1OO2G:E;F;E:E;F:E;K6L3L4M3M3L5L3L4M3M4K4M3M3L4M4L3L4MScm1"}, "label": "an unmade bed with white sheets and brown blankets"}]} {"id": 205794, "image": "COCO_train2014_000000205794.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the matteress of the bed with pillows and blankets\"."}], "task": "refering", "answer_template": "The \"the matteress of the bed with pillows and blankets\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 149, 150, 151, 152, 153, 154, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270], "bbox": [0.1483125, 0.37714583333333335, 0.7938125, 0.7277083333333334], "iscrowd": 0, "area": 48371.8819, "rle": {"size": [480, 640], "counts": "ji\\14k>7I7J7H7I8H6K0O1O2N100O1O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1O100O2N1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O1O1O1O1O1N2O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O10000000000O1000001O00000000000000000O1000000000000O100O10000O100O100O100O101O0O100O10000000000000000000000O10000000000000000000000O100000000000000000000000O01000000000000000000000000O1000000000000N2O1O1O1N2O1O1N2O1O1O10000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O00O10000000000000000000O10000000000000O100000000000000000000000000000000O10000000000000000000000000000000000O100000000000000000000000000000000O10000000000000000000000000000000000O100000000000000000000000000000O1M2N3M3M3M3M3M4N20000010O2N2N2N2N3N0O2N2N1O2N101N2N1OO2G:E;F;E:E;F:E;K6L3L4M3M3L5L3L4M3M4K4M3M3L4M4L3L4MScm1"}, "label": "the matteress of the bed with pillows and blankets"}]} {"id": 9185, "image": "COCO_train2014_000000009185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the reddish orange in the back from the right\"."}], "task": "refering", "answer_template": "The \"the reddish orange in the back from the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 131, 132, 133, 153, 154, 155, 176, 177, 178, 199, 200, 201], "bbox": [0.6637187499999999, 0.2705140186915888, 0.875765625, 0.5571028037383178], "iscrowd": 0, "area": 6777.459550000003, "rle": {"size": [428, 640], "counts": "Wka5>j<4M3L4M3L4L4N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N3N1N2O1O1O1N2O2N1O1O1O1O1O1N2O2N1O1O1O1O1O1N2O2O0O1O1O100O1O1O2O0O1O1O1O100O1O1O100O1O1O100O1OjNVGlMi8R2Z1M3M2N3M3L3N3M3M2N3M3L4N101O1O001O1O001O1O0O2O1O001O1O001O1O1O1O1O1O001O1O1O100O100O1O100O1O100O1O100O1O100O1O100O100O1O100OigQ1"}, "label": "the reddish orange in the back from the right"}]} {"id": 9185, "image": "COCO_train2014_000000009185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"apple furthest to the right\"."}], "task": "refering", "answer_template": "The \"apple furthest to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 131, 132, 133, 153, 154, 155, 176, 177, 178, 199, 200, 201], "bbox": [0.6637187499999999, 0.2705140186915888, 0.875765625, 0.5571028037383178], "iscrowd": 0, "area": 6777.459550000003, "rle": {"size": [428, 640], "counts": "Wka5>j<4M3L4M3L4L4N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1O1N3N1N2O1O1O1N2O2N1O1O1O1O1O1N2O2N1O1O1O1O1O1N2O2O0O1O1O100O1O1O2O0O1O1O1O100O1O1O100O1O1O100O1OjNVGlMi8R2Z1M3M2N3M3L3N3M3M2N3M3L4N101O1O001O1O001O1O0O2O1O001O1O001O1O1O1O1O1O001O1O1O100O100O1O100O1O100O1O100O1O100O1O100O100O1O100OigQ1"}, "label": "apple furthest to the right"}]} {"id": 9185, "image": "COCO_train2014_000000009185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an apple in the center of oranges\"."}], "task": "refering", "answer_template": "The \"an apple in the center of oranges\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 237, 238], "bbox": [0.33201562500000004, 0.41883177570093455, 0.645140625, 0.6964719626168224], "iscrowd": 0, "area": 13857.713199999997, "rle": {"size": [428, 640], "counts": "iYi24U=4M3L4L4M3L4M3L4L4cNoN[FR1e9ROUFQ1j9TOoEn0R:VOhEl0W:XOcEj0]:[O]Ef0d:]OVEf0i:n001O010O00010O001O010O1O00100O1O100O001O100O1O010O1O1O10O01O100O1000O10O100000O10O100000O10O10000000O01000000O010000000O0100001O2M3N1O00000O01O1N2N2O0O2N2N2O1N101O10001O0O10000O1000000O101O000O10000O1000001N10000O1000001N10001O0O2O001N101O000O2O001O0O2O000O2O001O0O2O00001N101O0O2O00001N101O001N10001N101O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O1Omdn2"}, "label": "an apple in the center of oranges"}]} {"id": 9185, "image": "COCO_train2014_000000009185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the apple in the middle\"."}], "task": "refering", "answer_template": "The \"the apple in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 237, 238], "bbox": [0.33201562500000004, 0.41883177570093455, 0.645140625, 0.6964719626168224], "iscrowd": 0, "area": 13857.713199999997, "rle": {"size": [428, 640], "counts": "iYi24U=4M3L4L4M3L4M3L4L4cNoN[FR1e9ROUFQ1j9TOoEn0R:VOhEl0W:XOcEj0]:[O]Ef0d:]OVEf0i:n001O010O00010O001O010O1O00100O1O100O001O100O1O010O1O1O10O01O100O1000O10O100000O10O100000O10O10000000O01000000O010000000O0100001O2M3N1O00000O01O1N2N2O0O2N2N2O1N101O10001O0O10000O1000000O101O000O10000O1000001N10000O1000001N10001O0O2O001N101O000O2O001O0O2O000O2O001O0O2O00001N101O0O2O00001N101O001N10001N101O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O1Omdn2"}, "label": "the apple in the middle"}]} {"id": 517095, "image": "COCO_train2014_000000517095.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chestnut horse with a white stripe down its face\"."}], "task": "refering", "answer_template": "The \"a chestnut horse with a white stripe down its face\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 126, 127, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 264, 265, 266, 285], "bbox": [0.291953125, 0.3111875, 0.60234375, 0.7205833333333334], "iscrowd": 0, "area": 20099.478700000003, "rle": {"size": [480, 640], "counts": "ekg28f>8H3M3N1N3M3N0N2N3N1N2iNZOSDi0m;EcC<^ in this image.", "objects": [{"patches": [122, 123, 124, 126, 127, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 264, 265, 266, 285], "bbox": [0.291953125, 0.3111875, 0.60234375, 0.7205833333333334], "iscrowd": 0, "area": 20099.478700000003, "rle": {"size": [480, 640], "counts": "ekg28f>8H3M3N1N3M3N0N2N3N1N2iNZOSDi0m;EcC<^ in this image.", "objects": [{"patches": [119, 120, 121, 141, 142, 143, 144, 164, 165, 166, 167, 168, 169, 189, 190, 191, 192, 213, 215, 236, 237, 238, 259, 260], "bbox": [0.16325, 0.28002083333333333, 0.3874375, 0.7166458333333333], "iscrowd": 0, "area": 9376.428300000007, "rle": {"size": [480, 640], "counts": "_n`11m>2N3lNMVC7gT:AoE=Q:ATF in this image.", "objects": [{"patches": [45, 66, 67, 68, 83, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 294, 295, 296, 297, 318, 319, 320, 321, 340, 341, 342, 343, 344, 367], "bbox": [0.6235937500000001, 0.08989583333333333, 1.0, 0.892125], "iscrowd": 0, "area": 23949.114999999998, "rle": {"size": [480, 640], "counts": "mbk52o>1N3N1O2N1O1O2N1O2M2O2N1O2N1OO10000000000000000O10000000000000000O11O1O1O1N2O1O1O001O1O1O1N2O1O1O001O1O1O1N2N2M3N2M2O2M3M3N0O100O10O0100O100O010O100O1O010O100O10O0100O100O010O1O100O010O100O10O0100O1O10O0100O100O010O100O1O010O100O10OUHcMT3^2mLbMR3^2nLcMP3^2PMcMn2^2SMbM[M1VNMj4_2FdMQM8_NEi4`2GdMlLg3i18iMPL>h3i18jMoK=h3j19jMnKjMgKkMfK;k3k1?kMeK:l3j1a0lMaK;m3j1b0lM`K:n3j1b0lM`K:m3k1d0kM^K:n3j1e0lM]K:m3k1f0kM]K:m3k1g0kMZK;n3k1h0jMZK;m3k1j0jMYK;m3k1k0jMWK;m3l1l0iMWK;m3l1m0hMUK=m3k1o0iMSKn3l1Y1fMiJ>m3l1\\1fMfJ>n3l1\\1fMfJ?l3l1_1eMdJ?m3l1_1fMbJ?n3k1a1gM`J>n3l1b1gM_J=o3l1b1hM^J in this image.", "objects": [{"patches": [45, 66, 67, 68, 83, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 294, 295, 296, 297, 318, 319, 320, 321, 340, 341, 342, 343, 344, 367], "bbox": [0.6235937500000001, 0.08989583333333333, 1.0, 0.892125], "iscrowd": 0, "area": 23949.114999999998, "rle": {"size": [480, 640], "counts": "mbk52o>1N3N1O2N1O1O2N1O2M2O2N1O2N1OO10000000000000000O10000000000000000O11O1O1O1N2O1O1O001O1O1O1N2O1O1O001O1O1O1N2N2M3N2M2O2M3M3N0O100O10O0100O100O010O100O1O010O100O10O0100O100O010O1O100O010O100O10O0100O1O10O0100O100O010O100O1O010O100O10OUHcMT3^2mLbMR3^2nLcMP3^2PMcMn2^2SMbM[M1VNMj4_2FdMQM8_NEi4`2GdMlLg3i18iMPL>h3i18jMoK=h3j19jMnKjMgKkMfK;k3k1?kMeK:l3j1a0lMaK;m3j1b0lM`K:n3j1b0lM`K:m3k1d0kM^K:n3j1e0lM]K:m3k1f0kM]K:m3k1g0kMZK;n3k1h0jMZK;m3k1j0jMYK;m3k1k0jMWK;m3l1l0iMWK;m3l1m0hMUK=m3k1o0iMSKn3l1Y1fMiJ>m3l1\\1fMfJ>n3l1\\1fMfJ?l3l1_1eMdJ?m3l1_1fMbJ?n3k1a1gM`J>n3l1b1gM_J=o3l1b1hM^J in this image.", "objects": [{"patches": [148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 381, 382, 383], "bbox": [0.2561875, 0.3595416666666667, 0.8612343750000001, 0.9483125000000001], "iscrowd": 0, "area": 58171.22344999999, "rle": {"size": [480, 640], "counts": "jX]2c0X>:H6K3M3L4J6O1N2O1O1N2O1N2O1N10001O0O1000001N100000001N100000O01000O10O10O1000O01O100O1ODPNiCP2WhDV12VN99l:Z2kD^M6;o:V2kD`M3U;Q2jDZNV;GiDV10TOX;c1hD^NW;EiDU10WOW;CiDV10XOW;BiDU10ZOV;AjD\\2V;dMjD\\2V;cMjD]2V;cMkD]2U;cMjD^2V;aMkD_2U;`MlD`2S;aMmD_2S;`MnD`2R;`MnD_2R;aMnD`2R;_MoDa2Q;_MoDa2P;_MQEa2o:_MQEa2o:^MQEc2n:]MSEc2m:]MSEc2m:\\MSEe2l:[MUEe2k:[MUEe2k:ZMUEg2j:ZMVEf2j:YMWEg2i:XMWEi2h:XMWEi2i:VMVEl2j:TMUEm2j:SMVEn2h8gLQI:WNP3c8mLTI3WNR3`8SMVIJZNT3Z8ZMZIV3a6RM\\In2_6YM_Ig2]6`M`Ia2Z6fMdIZ2W6UN_Ik1^6cNWI^1f6kNSIU1k6ROPIn0n6YOmHg0P7g3M2N2N2O1O100O1O100O1O100O1O100O1O1O100O1O100O101N10001O00001N10001O00001O000000000000000000000000000000000000000000000000000000001O000000001O00000010O0000000000001O01O000000000001O000000010O0000001O0000001O0000001O0000001O0000001O0cKfI4Z6LgI3Y6LiI3X6LjI2V6NkI1U6OlI0U6NmI1S6OmI1S6OnI0S6OnI0R6OoI1Q6OPJ0Q6OPJ0P60PJ0P6ORJ0o5ORJ0n50RJ0n50SJ0m5NUJ1k5OVJ0j5NYJ1h5L\\J2d5M^J2b5LaJ3_5KdJ4]5IfJPOQO]NY6a2iJiN_O]Ni5g2kJcNFdN_5g2oJaNFfN\\5f2QKcNHbNW5h2UKdNe6Y1^IeNc6Z1`IcNb6S1hIkNY6l0PJTOQ6k0PJSOQ6l0QJROQ6m0oIROR6m0PJROQ6k0QJWOm5h0UJZOi5c0YJ@d5?^JC`5:cJG[58fJK[5OhJ3Z5FhJ0000000000000O10000O1N200O1O1N2N2O1N200O1O1O1O1N2O1O1O1O1O1O1N2O1O1O2N1O1O2O0O1O2N1O101N1O101N2O2M3M2O2M3M4M2M3N1N3M2O2M2O2M2O1N3MQ[Y1"}, "label": "a small , one - year - old , female human holding food and smashing said food on their face"}]} {"id": 173032, "image": "COCO_train2014_000000173032.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl celebrating her birthday holding cake with icing smeared all over her face\"."}], "task": "refering", "answer_template": "The \"a girl celebrating her birthday holding cake with icing smeared all over her face\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 381, 382, 383], "bbox": [0.2561875, 0.3595416666666667, 0.8612343750000001, 0.9483125000000001], "iscrowd": 0, "area": 58171.22344999999, "rle": {"size": [480, 640], "counts": "jX]2c0X>:H6K3M3L4J6O1N2O1O1N2O1N2O1N10001O0O1000001N100000001N100000O01000O10O10O1000O01O100O1ODPNiCP2WhDV12VN99l:Z2kD^M6;o:V2kD`M3U;Q2jDZNV;GiDV10TOX;c1hD^NW;EiDU10WOW;CiDV10XOW;BiDU10ZOV;AjD\\2V;dMjD\\2V;cMjD]2V;cMkD]2U;cMjD^2V;aMkD_2U;`MlD`2S;aMmD_2S;`MnD`2R;`MnD_2R;aMnD`2R;_MoDa2Q;_MoDa2P;_MQEa2o:_MQEa2o:^MQEc2n:]MSEc2m:]MSEc2m:\\MSEe2l:[MUEe2k:[MUEe2k:ZMUEg2j:ZMVEf2j:YMWEg2i:XMWEi2h:XMWEi2i:VMVEl2j:TMUEm2j:SMVEn2h8gLQI:WNP3c8mLTI3WNR3`8SMVIJZNT3Z8ZMZIV3a6RM\\In2_6YM_Ig2]6`M`Ia2Z6fMdIZ2W6UN_Ik1^6cNWI^1f6kNSIU1k6ROPIn0n6YOmHg0P7g3M2N2N2O1O100O1O100O1O100O1O100O1O1O100O1O100O101N10001O00001N10001O00001O000000000000000000000000000000000000000000000000000000001O000000001O00000010O0000000000001O01O000000000001O000000010O0000001O0000001O0000001O0000001O0000001O0cKfI4Z6LgI3Y6LiI3X6LjI2V6NkI1U6OlI0U6NmI1S6OmI1S6OnI0S6OnI0R6OoI1Q6OPJ0Q6OPJ0P60PJ0P6ORJ0o5ORJ0n50RJ0n50SJ0m5NUJ1k5OVJ0j5NYJ1h5L\\J2d5M^J2b5LaJ3_5KdJ4]5IfJPOQO]NY6a2iJiN_O]Ni5g2kJcNFdN_5g2oJaNFfN\\5f2QKcNHbNW5h2UKdNe6Y1^IeNc6Z1`IcNb6S1hIkNY6l0PJTOQ6k0PJSOQ6l0QJROQ6m0oIROR6m0PJROQ6k0QJWOm5h0UJZOi5c0YJ@d5?^JC`5:cJG[58fJK[5OhJ3Z5FhJ0000000000000O10000O1N200O1O1N2N2O1N200O1O1O1O1N2O1O1O1O1O1O1N2O1O1O2N1O1O2O0O1O2N1O101N1O101N2O2M3M2O2M3M4M2M3N1N3M2O2M2O2M2O1N3MQ[Y1"}, "label": "a girl celebrating her birthday holding cake with icing smeared all over her face"}]} {"id": 173032, "image": "COCO_train2014_000000173032.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man standing in the backside with ash color pant and blue color t - shirt\"."}], "task": "refering", "answer_template": "The \"man standing in the backside with ash color pant and blue color t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 60, 61, 62, 63, 64, 65, 66, 67, 83, 84, 85, 86, 87, 88, 107, 108, 109, 137, 154, 155, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 292, 293, 295, 296, 297, 298, 315], "bbox": [0.5644843749999999, 0.0, 1.0, 0.7868541666666666], "iscrowd": 0, "area": 47183.45195, "rle": {"size": [480, 640], "counts": "PWY51O001O001O001O001O001O001O001O001O001O1O1O1O001O1O1O001O1O1O2Na0_O1O0000O10000O10000B>O100O1OVCSOc;l0ZDYODIZ;n0oD]OCI];i0nDBAJ_;c0nDG@Ia;?mDL^OId;;jD1_OHe;6jDo0U;POiDT1V;lNgDY1W;fNgD^1X;aN_Dk1_;l0O2N2N3M1O001O1O001O001QKYLcNh3Y1\\LfNd3V1aLiN`3S1dLlN\\3P1iLoNX3m0lLROU3i0PMVOP3g0UMWOl2d0YM[Oh2a0\\M^Oe2=`MBa2:cME^26gMIZ22lMLU20oMOR2LSN3n1IVN6k1EZN:g1B]N=d1^OaNa0`1oNQOo0o0oNTOP1k0mM:R2FhMa0W2_OdMg0[2YOeMi0Y2VOgMl0X2TOhMm0W2SOhMo0W2POjMR1T2nNkMT1T2lNlMU1S2jNnMX1P2hNoMZ1P2fNPN[1o1dNQN_1m1aNSN`1l1`NSNb1l1^NTNd1j1[NWNf1h1ZNYNf1f1ZNZNh1d1WN]Nj1b1VN^Nf1f1YN[N]1o1cNQNT1X2kNiMk0a2UO`Ma0i2^OXM8R3GoL0Z30fLGc38^L_Ok3a0ULWOS4h0nKoN[4Q1eKfNd4Y1^K]Nk4c1UKTNT5k1mJlM\\5T2dJcMe5\\2\\J]Mk5c2UJXMP6g2QJTMT6k2nInLX6R3hIiL]6V3dIdLb6\\3^I_Lg6`3ZI[Lk6e3X2O100O100O1O100O100G9CeKWMYMf2e2`MZM^2b2hM^MV2_2PN`Mc1g2cNYM\\1c2jN\\MU1_2QOaMn0[2XOdMg0W2_OiMa0R2EmM:n1LRN4i11WNOc18\\NH^1>bNBY1d0fN\\OT1j0lNUOg0Z1XOfN9i1GWNIZ26fM[Oi2e0WMZOk2e0UM[Ok2e0UMZOl2f0UMYOl2f0TMYOm2g0SMYOn2f0RMYOo2g0QMYOP3f0PMYOQ3g0oLYOR3f0nLYOS3g0mLYOT3f0nLWOS3i0oLUOQ3k0[MhNg2W1^MdNb2\\1cM^N_2a1aM_N_2a1aM_N`2`1_M`Nb2`1^M`Nc2_1\\MbNd2^1\\MbNe2]1ZMdNf2\\1ZMdNg2[1XMfNh2Z1XMfNi2Y1WMgNi2Y1VMgNk2Y1UMgNl2X1SMiNm2W1SMiNn2V1QMkNo2U1QMkNP3T1oLmNQ3S1oLmNR3R1mLoNS3Q1mLnNU3Q1jLPOV3P1jLPOV3P1jLPOW3o0hLROX3n0hLROX3n0gLSOY3m0gLSOZ3l0eLUO[3k0eLUO[3k0dLUO]3k0cLUO^3j0aLWO_3i0aLWO_3i0aLWO_3i0`LXO`3h0`LXOa3g0^LZOb3f0^LYOc3g0\\LZOd3f0\\LZOe3e0ZL\\Of3d0ZL\\Of3d0YL]Og3c0YL\\Oi3c0VL^Oj3b0VL^Oj3b0VL^Oj3b0UL_Ol3`0SLAm3?RLAo3?oKCQ4=nKDR4 in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 60, 61, 62, 63, 64, 65, 66, 67, 83, 84, 85, 86, 87, 88, 107, 108, 109, 137, 154, 155, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 292, 293, 295, 296, 297, 298, 315], "bbox": [0.5644843749999999, 0.0, 1.0, 0.7868541666666666], "iscrowd": 0, "area": 47183.45195, "rle": {"size": [480, 640], "counts": "PWY51O001O001O001O001O001O001O001O001O001O1O1O1O001O1O1O001O1O1O2Na0_O1O0000O10000O10000B>O100O1OVCSOc;l0ZDYODIZ;n0oD]OCI];i0nDBAJ_;c0nDG@Ia;?mDL^OId;;jD1_OHe;6jDo0U;POiDT1V;lNgDY1W;fNgD^1X;aN_Dk1_;l0O2N2N3M1O001O1O001O001QKYLcNh3Y1\\LfNd3V1aLiN`3S1dLlN\\3P1iLoNX3m0lLROU3i0PMVOP3g0UMWOl2d0YM[Oh2a0\\M^Oe2=`MBa2:cME^26gMIZ22lMLU20oMOR2LSN3n1IVN6k1EZN:g1B]N=d1^OaNa0`1oNQOo0o0oNTOP1k0mM:R2FhMa0W2_OdMg0[2YOeMi0Y2VOgMl0X2TOhMm0W2SOhMo0W2POjMR1T2nNkMT1T2lNlMU1S2jNnMX1P2hNoMZ1P2fNPN[1o1dNQN_1m1aNSN`1l1`NSNb1l1^NTNd1j1[NWNf1h1ZNYNf1f1ZNZNh1d1WN]Nj1b1VN^Nf1f1YN[N]1o1cNQNT1X2kNiMk0a2UO`Ma0i2^OXM8R3GoL0Z30fLGc38^L_Ok3a0ULWOS4h0nKoN[4Q1eKfNd4Y1^K]Nk4c1UKTNT5k1mJlM\\5T2dJcMe5\\2\\J]Mk5c2UJXMP6g2QJTMT6k2nInLX6R3hIiL]6V3dIdLb6\\3^I_Lg6`3ZI[Lk6e3X2O100O100O1O100O100G9CeKWMYMf2e2`MZM^2b2hM^MV2_2PN`Mc1g2cNYM\\1c2jN\\MU1_2QOaMn0[2XOdMg0W2_OiMa0R2EmM:n1LRN4i11WNOc18\\NH^1>bNBY1d0fN\\OT1j0lNUOg0Z1XOfN9i1GWNIZ26fM[Oi2e0WMZOk2e0UM[Ok2e0UMZOl2f0UMYOl2f0TMYOm2g0SMYOn2f0RMYOo2g0QMYOP3f0PMYOQ3g0oLYOR3f0nLYOS3g0mLYOT3f0nLWOS3i0oLUOQ3k0[MhNg2W1^MdNb2\\1cM^N_2a1aM_N_2a1aM_N`2`1_M`Nb2`1^M`Nc2_1\\MbNd2^1\\MbNe2]1ZMdNf2\\1ZMdNg2[1XMfNh2Z1XMfNi2Y1WMgNi2Y1VMgNk2Y1UMgNl2X1SMiNm2W1SMiNn2V1QMkNo2U1QMkNP3T1oLmNQ3S1oLmNR3R1mLoNS3Q1mLnNU3Q1jLPOV3P1jLPOV3P1jLPOW3o0hLROX3n0hLROX3n0gLSOY3m0gLSOZ3l0eLUO[3k0eLUO[3k0dLUO]3k0cLUO^3j0aLWO_3i0aLWO_3i0aLWO_3i0`LXO`3h0`LXOa3g0^LZOb3f0^LYOc3g0\\LZOd3f0\\LZOe3e0ZL\\Of3d0ZL\\Of3d0YL]Og3c0YL\\Oi3c0VL^Oj3b0VL^Oj3b0VL^Oj3b0UL_Ol3`0SLAm3?RLAo3?oKCQ4=nKDR4 in this image.", "objects": [{"patches": [39, 40, 41, 42, 62, 63, 64, 65, 86, 87, 109, 110, 111], "bbox": [0.698828125, 0.07439678284182305, 0.872609375, 0.3760053619302949], "iscrowd": 0, "area": 7028.161650000002, "rle": {"size": [373, 640], "counts": "fkR52`;6J6K5J5L200010O00010O000010O01O01O01O00010O06J7J6I8I6I:Fm0TO3L0001O01O00000O1000000000000O1000000000000O1000000000000O0100000000000001O000000000000000000O1000iNmGlNR8g0]HXOc7:lHES75VIIj67WIHi67ZIGe69^IEb6;_IDa6;bIC]6>eI@[6?gI@X6a0jI]OV6b0mI\\OS6d0nI[OQ6e0RJXOo5h0SJVOl5j0WJTOi5l0`2O010O101N2M3M3N2M3N2M3N2M3Nejm0"}, "label": "the head of the white animal looking at the dog"}]} {"id": 222199, "image": "COCO_train2014_000000222199.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the head of an out of focus white lamb\"."}], "task": "refering", "answer_template": "The \"the head of an out of focus white lamb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 42, 62, 63, 64, 65, 86, 87, 109, 110, 111], "bbox": [0.698828125, 0.07439678284182305, 0.872609375, 0.3760053619302949], "iscrowd": 0, "area": 7028.161650000002, "rle": {"size": [373, 640], "counts": "fkR52`;6J6K5J5L200010O00010O000010O01O01O01O00010O06J7J6I8I6I:Fm0TO3L0001O01O00000O1000000000000O1000000000000O1000000000000O0100000000000001O000000000000000000O1000iNmGlNR8g0]HXOc7:lHES75VIIj67WIHi67ZIGe69^IEb6;_IDa6;bIC]6>eI@[6?gI@X6a0jI]OV6b0mI\\OS6d0nI[OQ6e0RJXOo5h0SJVOl5j0WJTOi5l0`2O010O101N2M3M3N2M3N2M3N2M3Nejm0"}, "label": "the head of an out of focus white lamb"}]} {"id": 222199, "image": "COCO_train2014_000000222199.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back end of a sheep\"."}], "task": "refering", "answer_template": "The \"the back end of a sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 247, 249, 250, 252, 266, 267, 270, 273], "bbox": [0.5790625, 0.37252010723860585, 1.0, 0.9068632707774799], "iscrowd": 0, "area": 36294.089949999994, "rle": {"size": [373, 640], "counts": "f]W4Z1Z:1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1OJ6E;0UNfN\\I[1P6YOoIg0h5DVJ in this image.", "objects": [{"patches": [109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 247, 249, 250, 252, 266, 267, 270, 273], "bbox": [0.5790625, 0.37252010723860585, 1.0, 0.9068632707774799], "iscrowd": 0, "area": 36294.089949999994, "rle": {"size": [373, 640], "counts": "f]W4Z1Z:1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1OJ6E;0UNfN\\I[1P6YOoIg0h5DVJ in this image.", "objects": [{"patches": [111, 112, 113, 119, 129, 130, 131, 133, 134, 135, 136, 137, 138, 139, 141, 142, 143, 144, 147, 148, 149, 151, 152, 153, 154, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.28, 1.0, 0.9858878504672898], "iscrowd": 0, "area": 167452.02674999993, "rle": {"size": [428, 640], "counts": "T7b0o3JX4\\1UJ]O_5h1oHPOf6Q4dHjJZ6Z5fIfJZ6Z5fIfJZ6[5eIeJZ6\\5fIdJZ6\\5fIdJZ6]5eIcJ[6]5eIcJZ6^5fIbJZ6^5fIbJZ6_5eIaJZ6`5fI`JZ6`5fI`JZ6Y600O1000000O10000O1000000O11O0000001O0000001O000000001O0000001O00000000001O00000000001O00000000002N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O^ORJhIl5V6]JdIb5[6h0N2O1N2O1N2O1N2O1N2O1N200O1O100O1O100O100O1O100O1O010O1O100O1O100O1O100O1O100O1O100O1O1001O000000001O0000001O000000001O0000001O000000001O0000001O0000001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O2N4L4L5K4L4L5K3M00O100O1O100O100O100O10000000000000000O10000000000000000O10000000000000000O100000000000000N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2N2O1O100O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O1001O001O001O001O00001O001O001O001O001O00001O001O0O3N1O1]JRJf3P6WLQJi3P6ULQJk3P6SLQJm3Q6PLPJP4Q6nKPJR4Q6lKPJT4R6iKoIW4R6gKoIY4R6eKoI[4S6cKmI]4T6dKjI\\4W6_KlIb4V6XKnIh4S6SKPJn4Q6mJSJS5o5gJTJZ5m5aJWJ_5j610000O100O10000O100O10000O1O1O1O100O1O1O1O1O1O100POP1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O10000001O000000000000000000001O0000000000000000001O000000000000000000001O0000000000001O2N2N2N2N2N3M2N2N2N2N2N2N2N3SJiI_4Y6[KkIe4W6UKnIj4T6QKoIo4S6kJRJT5P6fJTJZ5n5`JWJ_5k601O001O00001O001O001O001O00B>N2N2O1N2N2N2N2N2N2N2N2N2N2N2O1N2O1N2O1N2N2[Oe0M3L4M3L4M3L4O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O00100O10000O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O100000000001O00000000001O00000000001O00001O1O001O1O1O1O1O001O1O1O1O1O001O2N2N2N2N3M2N2N2N3M2N2N3dIeJa4`5WKgJe4^5SKiJi4\\5oJkJm4Z5kJmJQ5X5gJoJU5V5cJQKY5e6L3M3M3M2N3M3M[1eN>B=C=C=C=C=C=C=CTC"}, "label": "all of the apples that are in the box"}]} {"id": 517123, "image": "COCO_train2014_000000517123.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white sheep on the right\"."}], "task": "refering", "answer_template": "The \"the white sheep on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 105, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 202, 203, 204], "bbox": [0.54634375, 0.022625, 1.0, 0.539875], "iscrowd": 0, "area": 40384.37479999999, "rle": {"size": [480, 640], "counts": "RUT52m>1N2O2M2N3N1N3N2M3M3N3L3N3L3N3L3N3M1O2N2N2N2N1O2N2N2N1O1O001O001O00001O001O01O01O001O01O01O010O001O010O0010O01O010O001O010O00100O001O010O00100O1O2O0O1001O1O2O0O2NO1000001O0000000jDjMb9W2YFnMf9R2WFQNh9P2WETN61b:k1WEVN23g:g1VEWN24g:e1VEYN13i:e1TEYN32i:e1SE[N22j:]2UEcMk:]2UEdMj:Y301N101O0O2O001N2O2M3N2N3L3N2M4M2M3N2N3L3000O1N2O1N1O010O00010O001O01O01O010O000010O01O01O01OO2O000O101O0O101N10003L7J6I7J6J3L2O000O101N10000O2O000O101NkNlG`KT8a4UHUKk7l4^HjJb7V5hHaJW7`5T11O1O1O001O1O000000000O10001O000O100000000]G[Jj7f5lG[K]7T60O100O010O100O100O1^O\\HhId7W6eH`I\\7^6d0N101N2O1N2N2N2N2O1N2N2N1O2N2O1N2N2O001O1O1O00100O1O1O001O1O1O001O1000O1000001N1000000O101O000O10000O101N1O100O1000001O000000000010O0000000000c0^OP2oMQ2oMoJ"}, "label": "the white sheep on the right"}]} {"id": 517123, "image": "COCO_train2014_000000517123.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white sheep\"."}], "task": "refering", "answer_template": "The \"a white sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 105, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 202, 203, 204], "bbox": [0.54634375, 0.022625, 1.0, 0.539875], "iscrowd": 0, "area": 40384.37479999999, "rle": {"size": [480, 640], "counts": "RUT52m>1N2O2M2N3N1N3N2M3M3N3L3N3L3N3L3N3M1O2N2N2N2N1O2N2N2N1O1O001O001O00001O001O01O01O001O01O01O010O001O010O0010O01O010O001O010O00100O001O010O00100O1O2O0O1001O1O2O0O2NO1000001O0000000jDjMb9W2YFnMf9R2WFQNh9P2WETN61b:k1WEVN23g:g1VEWN24g:e1VEYN13i:e1TEYN32i:e1SE[N22j:]2UEcMk:]2UEdMj:Y301N101O0O2O001N2O2M3N2N3L3N2M4M2M3N2N3L3000O1N2O1N1O010O00010O001O01O01O010O000010O01O01O01OO2O000O101O0O101N10003L7J6I7J6J3L2O000O101N10000O2O000O101NkNlG`KT8a4UHUKk7l4^HjJb7V5hHaJW7`5T11O1O1O001O1O000000000O10001O000O100000000]G[Jj7f5lG[K]7T60O100O010O100O100O1^O\\HhId7W6eH`I\\7^6d0N101N2O1N2N2N2N2O1N2N2N1O2N2O1N2N2O001O1O1O00100O1O1O001O1O1O001O1000O1000001N1000000O101O000O10000O101N1O100O1000001O000000000010O0000000000c0^OP2oMQ2oMoJ"}, "label": "a white sheep"}]} {"id": 263176, "image": "COCO_train2014_000000263176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white and light blue boat on the right\"."}], "task": "refering", "answer_template": "The \"the white and light blue boat on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 288, 289, 290], "bbox": [0.1989375, 0.3438262910798122, 0.80921875, 0.8629107981220658], "iscrowd": 0, "area": 56225.90575, "rle": {"size": [426, 640], "counts": "dQe1?k<1N2O1O1O1N2O1O1N2O1O1O1N3N1O2N1N3N1O2M2O2N1O2M2O2N1O1N3N1N3N1O2M2O2M2O2N1N3N1O2M2O1N3N1O2M2O2M2O2N1N3N1N3N1O2N101N1O101N1O2N101N1O2O1N8H7J7H8H6JO2O00001O0000001O0O101O0000001O000O2O00000000O100O10O0100O100O100O100O010O10000O100O10O01O1O1O1O1O1O1N2O001O1O1O1O100O10O10000000000000000O1000O100000000000O100000000000O10O0100O100O10000000O10000000000000000O1000O1000000000O10000000000000000O10000000000000000O100000000000000O1000O1N2O1N2O1N2N101N2O1N2O1N2O1O1O0100000O10000O10O1000O10000O100000O010000O1000000O010O10000O100O100O010O10000O100O1000O100000000000000O10000000000000000O10000000001O00001N1000001O000O101N100O2N100O101N1O101N100O1O2O0O100O2N100O2O0O1O101N10>A3M2O1N2O1N2N3N1N2O1N2N2O2M2O1N2O1N3M2O1N2O1N2N3N1N2O1M3L4L3N3L3M3M4L3M4L3M4L3N3L3M4L3M4L3M4L3M4M2M4L3M4L3M3M4L3M4M\\ab1"}, "label": "the white and light blue boat on the right"}]} {"id": 263176, "image": "COCO_train2014_000000263176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and blue sail boat parked at the dock\"."}], "task": "refering", "answer_template": "The \"a white and blue sail boat parked at the dock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 288, 289, 290], "bbox": [0.1989375, 0.3438262910798122, 0.80921875, 0.8629107981220658], "iscrowd": 0, "area": 56225.90575, "rle": {"size": [426, 640], "counts": "dQe1?k<1N2O1O1O1N2O1O1N2O1O1O1N3N1O2N1N3N1O2M2O2N1O2M2O2N1O1N3N1N3N1O2M2O2M2O2N1N3N1O2M2O1N3N1O2M2O2M2O2N1N3N1N3N1O2N101N1O101N1O2N101N1O2O1N8H7J7H8H6JO2O00001O0000001O0O101O0000001O000O2O00000000O100O10O0100O100O100O100O010O10000O100O10O01O1O1O1O1O1O1N2O001O1O1O1O100O10O10000000000000000O1000O100000000000O100000000000O10O0100O100O10000000O10000000000000000O1000O1000000000O10000000000000000O10000000000000000O100000000000000O1000O1N2O1N2O1N2N101N2O1N2O1N2O1O1O0100000O10000O10O1000O10000O100000O010000O1000000O010O10000O100O100O010O10000O100O1000O100000000000000O10000000000000000O10000000001O00001N1000001O000O101N100O2N100O101N1O101N100O1O2O0O100O2N100O2O0O1O101N10>A3M2O1N2O1N2N3N1N2O1N2N2O2M2O1N2O1N3M2O1N2O1N2N3N1N2O1M3L4L3N3L3M3M4L3M4L3M4L3N3L3M4L3M4L3M4L3M4M2M4L3M4L3M3M4L3M4M\\ab1"}, "label": "a white and blue sail boat parked at the dock"}]} {"id": 402448, "image": "COCO_train2014_000000402448.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown chair with a cat looking at it\"."}], "task": "refering", "answer_template": "The \"brown chair with a cat looking at it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 111, 112, 113, 114, 134, 135, 136, 137, 156, 157, 158, 159, 180, 181], "bbox": [0.7341875, 0.011207729468599032, 1.0, 0.496207729468599], "iscrowd": 0, "area": 18897.636200000004, "rle": {"size": [414, 640], "counts": "QRn51m<1M3M3M3M3L4M3M3M3M3M3M3N2N2N2N4M4K5K4L5K5L3L3M1O2N101O0O2O001N101O0O2O001O0O2O001O001N101O001O1O1O1O1eHUMa4l2YKZMf4g2WK]Mg4c2XK^Mh4c2VK^Mj4c2TK^Ml4c2RK_Mm4b2QK_Mo4b2PK^MP5c2nJ^MR5c2lJ^MT5b2kJ_MU5b2iJ`MV5`2iJaMW5`2hJ`MY5`2eJaM[5_2dJbM\\5_2bJbM_5]2`JeM_5\\2^JfMb5Z2]JgMc5Z2ZJhMg5W2XJjMh5W2VJjMj5V2TJmMl5S2RJnMn5R2PJPNP6Q2nIPNR6P2lIRNU6m1jITNW6l1fIWNZ6h1eIYN[6h1bIZN_6e1`I\\Na6d1\\I^Ne6a1ZI`Nf6a1WIbNi6_1TIbNl6_1RIbNn6`1nHbNQ7f30001N10000O10001N1000000O101O00000000N2O1N2O1N2N2O1N2N2O1O1aMdH[O^7a0gH\\OZ7a0kH[OW7b0nH[OS7d0PIYOR7e0RIWOo6h0RIWOo6h0RIVOP7i0RIUOo6j0RIUOP7i0QIVOP7i0RIUOP7j0PITOS7j0dHlM]OY1P8j0bH@a7>^HBe7=[HAh7=XHBk7 in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 111, 112, 113, 114, 134, 135, 136, 137, 156, 157, 158, 159, 180, 181], "bbox": [0.7341875, 0.011207729468599032, 1.0, 0.496207729468599], "iscrowd": 0, "area": 18897.636200000004, "rle": {"size": [414, 640], "counts": "QRn51m<1M3M3M3M3L4M3M3M3M3M3M3N2N2N2N4M4K5K4L5K5L3L3M1O2N101O0O2O001N101O0O2O001O0O2O001O001N101O001O1O1O1O1eHUMa4l2YKZMf4g2WK]Mg4c2XK^Mh4c2VK^Mj4c2TK^Ml4c2RK_Mm4b2QK_Mo4b2PK^MP5c2nJ^MR5c2lJ^MT5b2kJ_MU5b2iJ`MV5`2iJaMW5`2hJ`MY5`2eJaM[5_2dJbM\\5_2bJbM_5]2`JeM_5\\2^JfMb5Z2]JgMc5Z2ZJhMg5W2XJjMh5W2VJjMj5V2TJmMl5S2RJnMn5R2PJPNP6Q2nIPNR6P2lIRNU6m1jITNW6l1fIWNZ6h1eIYN[6h1bIZN_6e1`I\\Na6d1\\I^Ne6a1ZI`Nf6a1WIbNi6_1TIbNl6_1RIbNn6`1nHbNQ7f30001N10000O10001N1000000O101O00000000N2O1N2O1N2N2O1N2N2O1O1aMdH[O^7a0gH\\OZ7a0kH[OW7b0nH[OS7d0PIYOR7e0RIWOo6h0RIWOo6h0RIVOP7i0RIUOo6j0RIUOP7i0QIVOP7i0RIUOP7j0PITOS7j0dHlM]OY1P8j0bH@a7>^HBe7=[HAh7=XHBk7 in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 27, 28, 29, 30, 31, 33, 50, 51, 52, 53, 54, 73, 74, 75, 76, 77, 96, 97, 98, 99, 119], "bbox": [0.177, 0.0, 0.4848906250000001, 0.36227053140096616], "iscrowd": 0, "area": 17814.18355, "rle": {"size": [414, 640], "counts": "ag]19R in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 27, 28, 29, 30, 31, 33, 50, 51, 52, 53, 54, 73, 74, 75, 76, 77, 96, 97, 98, 99, 119], "bbox": [0.177, 0.0, 0.4848906250000001, 0.36227053140096616], "iscrowd": 0, "area": 17814.18355, "rle": {"size": [414, 640], "counts": "ag]19R in this image.", "objects": [{"patches": [211, 212, 213, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 300, 301, 302], "bbox": [0.06103125, 0.546723044397463, 0.307359375, 0.7723467230443976], "iscrowd": 0, "area": 8026.69505, "rle": {"size": [473, 640], "counts": "g[b01e>4L3M3M4L3M3M4M2M3M4L3N2M3O1001O0000000000000000000000000000000000000000000000001O0000000000000O100000000000000O10000000000000001N10000000000O100O100O100ON3L3N3L3N3L4M2M4M2M4M2M4M2M4M2M4O01O100O001O1O10O01O1O10O0100O10O010000O10O0100O10O1000O10O100000O10O100000O01000O1N6K2N1O1O1N2O1O1O2N1N2O1O1O1O1N2O1O1O1OV1iN3N1OeZ\\6"}, "label": "red flowers to the left of a blue teddy bear"}]} {"id": 484369, "image": "COCO_train2014_000000484369.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red and purple wild flowers surrounded by greenery\"."}], "task": "refering", "answer_template": "The \"red and purple wild flowers surrounded by greenery\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 300, 301, 302], "bbox": [0.06103125, 0.546723044397463, 0.307359375, 0.7723467230443976], "iscrowd": 0, "area": 8026.69505, "rle": {"size": [473, 640], "counts": "g[b01e>4L3M3M4L3M3M4M2M3M4L3N2M3O1001O0000000000000000000000000000000000000000000000001O0000000000000O100000000000000O10000000000000001N10000000000O100O100O100ON3L3N3L3N3L4M2M4M2M4M2M4M2M4M2M4O01O100O001O1O10O01O1O10O0100O10O010000O10O0100O10O1000O10O100000O10O100000O01000O1N6K2N1O1O1N2O1O1O2N1N2O1O1O1O1N2O1O1O1OV1iN3N1OeZ\\6"}, "label": "red and purple wild flowers surrounded by greenery"}]} {"id": 484369, "image": "COCO_train2014_000000484369.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"grass and other vegetation below the blue doll , parts of the feet of the doll\"."}], "task": "refering", "answer_template": "The \"grass and other vegetation below the blue doll , parts of the feet of the doll\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [279, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326, 327, 347, 348, 349, 350, 351, 369, 370, 371, 372, 373, 374, 375], "bbox": [0.033515625, 0.7536997885835095, 0.319921875, 0.9998942917547569], "iscrowd": 0, "area": 10364.835899999998, "rle": {"size": [473, 640], "counts": "Za:7]>5K6L301O001O001ON2L4K50O3N3L3M4L3M2O2N101N1O1YDlNg9V1VFkNj9V1TFkNl9W1PFlNo9U1oElNQ:V1lEkNU:U1iElNX:U1dEmN]:S1aEnN`:S1]EnNd:R1ZEoNg:R1UEPOl:P1REQOn:P1PEQOo:R1nDoNR;R1lDoNS;T1iDnNV;V1fDkNZ;W1cDjN\\;Z1`DgN`;\\1[DfNd;X200O1O100O1O1N2O1O1O2M2O1O1O1O1O1000000000000002N1O1O1O2N1O001O1O1O1O001O1O1O3M3M3M3M3M00O10000O10000O10M3M3N2O1O100O100O1O100O1O2O0kNaCOa<0`CNa<2`CMa<2_COah in this image.", "objects": [{"patches": [186, 187, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 299, 322], "bbox": [0.003328125, 0.4853911205073996, 0.207609375, 0.8471881606765328], "iscrowd": 0, "area": 10843.288350000003, "rle": {"size": [473, 640], "counts": "^V1V3X;j0@001O001O001O001O001O001O001O001OO1L4L4L4L5K5K5K5O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O0N3L4L4L4N200000000000000001O00000000000000000000000000000000001_OUNRDk1m;a000O100O10jMSD`1l;\\NXDe1h;WN[Dj1d;RNaDn1_;nMdDS2Q<0O1O100O2O0O1O100O100O1O10000001O10O01O1O001O1O001O1K4K6J6I7J5K6K5M2N3M3M2O2M3M2N3M9GimY7"}, "label": "wooden flower container"}]} {"id": 443410, "image": "COCO_train2014_000000443410.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"middle - aged dark brown or black hair woman with read dress and glasses\"."}], "task": "refering", "answer_template": "The \"middle - aged dark brown or black hair woman with read dress and glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 155, 176, 177, 178, 198, 199, 200, 201, 202, 221, 222, 223, 224, 225, 244, 245, 246, 247, 267, 268, 269, 291, 292], "bbox": [0.610734375, 0.3571875, 0.8020156249999999, 0.7486875000000001], "iscrowd": 0, "area": 13014.250049999999, "rle": {"size": [480, 640], "counts": "Vbg54h>5K6J5K5K6J5K6J5J6J7I6J6J7I6L5M2N2N3M2N2N3N1N3M2N2N3M2N2N3M2O2M2N2M4J5K5K6K4K6J5N2O2M101O0O2O0O2O0O1000O0100O0100O010O010O010O10O10O010O010O10O010O10O10O010OO4E;F9F;E;E:F;E;E:F;J5O2N2N101N2N1O2O1N1O3M3M3N1N3M3M3N2M3M3M3M3N1N3M3M3N2M3M3M2O2M2NfXk1"}, "label": "middle - aged dark brown or black hair woman with read dress and glasses"}]} {"id": 443410, "image": "COCO_train2014_000000443410.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the skirt\"."}], "task": "refering", "answer_template": "The \"the woman in the skirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 155, 176, 177, 178, 198, 199, 200, 201, 202, 221, 222, 223, 224, 225, 244, 245, 246, 247, 267, 268, 269, 291, 292], "bbox": [0.610734375, 0.3571875, 0.8020156249999999, 0.7486875000000001], "iscrowd": 0, "area": 13014.250049999999, "rle": {"size": [480, 640], "counts": "Vbg54h>5K6J5K5K6J5K6J5J6J7I6J6J7I6L5M2N2N3M2N2N3N1N3M2N2N3M2N2N3M2O2M2N2M4J5K5K6K4K6J5N2O2M101O0O2O0O2O0O1000O0100O0100O010O010O010O10O10O010O010O10O010O10O10O010OO4E;F9F;E;E:F;E;E:F;J5O2N2N101N2N1O2O1N1O3M3M3N1N3M3M3N2M3M3M3M3N1N3M3M3N2M3M3M2O2M2NfXk1"}, "label": "the woman in the skirt"}]} {"id": 443410, "image": "COCO_train2014_000000443410.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the middle sitting on the couch\"."}], "task": "refering", "answer_template": "The \"the woman in the middle sitting on the couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 171, 172, 173, 194, 195, 196, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 287, 288, 289], "bbox": [0.391015625, 0.39325, 0.584828125, 0.7617916666666666], "iscrowd": 0, "area": 12315.664050000001, "rle": {"size": [480, 640], "counts": "i_e31n>2M4L3N2M3M4M2M3N2N3M2O1O1N2O000O10000O1D_OQ;n0bD_Of0^Ol:n0_DDk0YOi:o0_DEi<6]CEg<6^CEg<6^CFf<5`CFd<6W1KQ]l3"}, "label": "the woman in the middle sitting on the couch"}]} {"id": 443410, "image": "COCO_train2014_000000443410.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in middle\"."}], "task": "refering", "answer_template": "The \"woman in middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 171, 172, 173, 194, 195, 196, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 287, 288, 289], "bbox": [0.391015625, 0.39325, 0.584828125, 0.7617916666666666], "iscrowd": 0, "area": 12315.664050000001, "rle": {"size": [480, 640], "counts": "i_e31n>2M4L3N2M3M4M2M3N2N3M2O1O1N2O000O10000O1D_OQ;n0bD_Of0^Ol:n0_DDk0YOi:o0_DEi<6]CEg<6^CEg<6^CFf<5`CFd<6W1KQ]l3"}, "label": "woman in middle"}]} {"id": 443410, "image": "COCO_train2014_000000443410.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a blue shirt\"."}], "task": "refering", "answer_template": "The \"a woman in a blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 165, 166, 167, 188, 189, 190, 191, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261, 262, 280, 281, 282, 283, 284, 285, 305, 306, 307, 308], "bbox": [0.185390625, 0.39325, 0.41460937500000006, 0.79325], "iscrowd": 0, "area": 17722.492599999998, "rle": {"size": [480, 640], "counts": "gPh1Z1c=8J6J6J4L1O2O1bCPNi;Q2VD^N[;d1bD_N\\;c1aD_N^;b1aD`N];b1`D`N_;b2N1O2BaLXE`3d:eLZE]3a:hL]EZ3^:lL_EU3^:i0N1O2M2O2N1N3N1O2M3O001O001N10001O001O001O0O2O001O001O010O010O010O0100O010O010]K]Fh3d9SL`Fl3b9oKaFR4a9hKcFX4_9cKcF]4Q:0O10O0100O101N100O10O001O001O001N10001O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N3N2N2M3N2N2M3N2N4K5L3M;D3N1O2M2O1O0O2O1O1O1O1O1O1N2O1O1O2N1O1O1O1O2M2O1O2N2N1O2N:F=B7J6J6Jd]_5"}, "label": "a woman in a blue shirt"}]} {"id": 443410, "image": "COCO_train2014_000000443410.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in blue shirt sitting on couch\"."}], "task": "refering", "answer_template": "The \"woman in blue shirt sitting on couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 165, 166, 167, 188, 189, 190, 191, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261, 262, 280, 281, 282, 283, 284, 285, 305, 306, 307, 308], "bbox": [0.185390625, 0.39325, 0.41460937500000006, 0.79325], "iscrowd": 0, "area": 17722.492599999998, "rle": {"size": [480, 640], "counts": "gPh1Z1c=8J6J6J4L1O2O1bCPNi;Q2VD^N[;d1bD_N\\;c1aD_N^;b1aD`N];b1`D`N_;b2N1O2BaLXE`3d:eLZE]3a:hL]EZ3^:lL_EU3^:i0N1O2M2O2N1N3N1O2M3O001O001N10001O001O001O0O2O001O001O010O010O010O0100O010O010]K]Fh3d9SL`Fl3b9oKaFR4a9hKcFX4_9cKcF]4Q:0O10O0100O101N100O10O001O001O001N10001O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N3N2N2M3N2N2M3N2N4K5L3M;D3N1O2M2O1O0O2O1O1O1O1O1O1N2O1O1O2N1O1O1O1O2M2O1O2N2N1O2N:F=B7J6J6Jd]_5"}, "label": "woman in blue shirt sitting on couch"}]} {"id": 328727, "image": "COCO_train2014_000000328727.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue felt chair in a class room\"."}], "task": "refering", "answer_template": "The \"a blue felt chair in a class room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 198, 199, 200, 220, 221, 222, 223, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269, 288, 289, 290, 291, 292, 311], "bbox": [0.5427500000000001, 0.49743750000000003, 0.7284375000000001, 0.8132291666666667], "iscrowd": 0, "area": 12171.82955, "rle": {"size": [480, 640], "counts": "e]S57U2NZ:8VE1j:6gD0[;6VD2l;4eC3\\ in this image.", "objects": [{"patches": [197, 198, 199, 200, 220, 221, 222, 223, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269, 288, 289, 290, 291, 292, 311], "bbox": [0.5427500000000001, 0.49743750000000003, 0.7284375000000001, 0.8132291666666667], "iscrowd": 0, "area": 12171.82955, "rle": {"size": [480, 640], "counts": "e]S57U2NZ:8VE1j:6gD0[;6VD2l;4eC3\\ in this image.", "objects": [{"patches": [240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 261, 262, 270, 271, 272, 273, 274, 275, 276, 277, 278, 285, 286, 287, 288, 289, 290, 291, 292, 293, 300, 301, 302, 303, 304, 305, 306, 307, 308, 315, 316, 317, 318, 319, 320, 321, 322, 323, 330, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.010141176470588235, 0.701125, 0.6023529411764705, 1.0], "iscrowd": 0, "area": 39322.15539999999, "rle": {"size": [640, 425], "counts": "on2Q5n>1O1O1N101O1O1O1O1O001O1O1O1O1N2O0010000O1000000000000O10000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O10000000000O100000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00000000000000001O00000000000000000000000000000000001O00000000000000000000000000000000001O0000000000000000001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O001O00001O00001O1O1O1O1O1O1O1O2N1O1O1O1N2O1O1N2O1N2O1O2M2O1O1N2O1N2O1O1N2N2M4M4K7I8H7I8H7I7I8H7I8H7I8I6I]QY3"}, "label": "the first pizza from the front"}]} {"id": 566301, "image": "COCO_train2014_000000566301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza at the bottom left with the green and yellow toppings\"."}], "task": "refering", "answer_template": "The \"pizza at the bottom left with the green and yellow toppings\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 261, 262, 270, 271, 272, 273, 274, 275, 276, 277, 278, 285, 286, 287, 288, 289, 290, 291, 292, 293, 300, 301, 302, 303, 304, 305, 306, 307, 308, 315, 316, 317, 318, 319, 320, 321, 322, 323, 330, 331, 332, 333, 334, 335, 336, 337, 338], "bbox": [0.010141176470588235, 0.701125, 0.6023529411764705, 1.0], "iscrowd": 0, "area": 39322.15539999999, "rle": {"size": [640, 425], "counts": "on2Q5n>1O1O1N101O1O1O1O1O001O1O1O1O1N2O0010000O1000000000000O10000000000O1000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O10000000000O100000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00000000000000001O00000000000000000000000000000000001O00000000000000000000000000000000001O0000000000000000001O00001O001O00001O00001O001O00001O00001O001O00001O00001O001O00001O001O00001O00001O1O1O1O1O1O1O1O2N1O1O1O1N2O1O1N2O1N2O1O2M2O1O1N2O1N2O1O1N2N2M4M4K7I8H7I8H7I7I8H7I8H7I8I6I]QY3"}, "label": "pizza at the bottom left with the green and yellow toppings"}]} {"id": 566301, "image": "COCO_train2014_000000566301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kid looking down at something\"."}], "task": "refering", "answer_template": "The \"a kid looking down at something\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 15, 16, 30, 31, 45, 46, 47, 48, 49, 60, 61, 62, 63, 64, 75, 76, 77, 78, 90, 91, 92, 105, 106, 107, 108, 120, 121, 122, 123, 135, 136, 137, 150, 151, 152], "bbox": [0.0, 0.00021875000000000003, 0.2964705882352941, 0.46554687499999997], "iscrowd": 0, "area": 20348.624449999996, "rle": {"size": [640, 425], "counts": "4d1\\b0W3iLX3hLP1PO3MO1O1O001O1O1O1O100O1O1O001kImD\\3T;dLlD[3U;dLlD\\3T;cLmD]3S;bLnD^3R;aLoD_3Q;aLoD_3Q;`LQE`3n:_LSEa3l:_LUEa3k:^LVEb3j:^LVEc3i:YL[Eg3e:QLcEo3]:_JUGa5k8[JYGe5g8WJ]Gj5b8TJaGk5^8SJeGm5[8PJhGP6T;001O000000010O000000001O01O0hLVB:j=CYB>g=_O[Ba0e=\\O^Bc0c=\\O^Bc0d=\\O]Bb0d=^O\\Ba0e=_O[B`0f=_O[B`0f=@ZB?h=@YB>h=AYB?g=AYB>h=AZB=g=BZB=h=BXB=i=BYB in this image.", "objects": [{"patches": [0, 1, 15, 16, 30, 31, 45, 46, 47, 48, 49, 60, 61, 62, 63, 64, 75, 76, 77, 78, 90, 91, 92, 105, 106, 107, 108, 120, 121, 122, 123, 135, 136, 137, 150, 151, 152], "bbox": [0.0, 0.00021875000000000003, 0.2964705882352941, 0.46554687499999997], "iscrowd": 0, "area": 20348.624449999996, "rle": {"size": [640, 425], "counts": "4d1\\b0W3iLX3hLP1PO3MO1O1O001O1O1O1O100O1O1O001kImD\\3T;dLlD[3U;dLlD\\3T;cLmD]3S;bLnD^3R;aLoD_3Q;aLoD_3Q;`LQE`3n:_LSEa3l:_LUEa3k:^LVEb3j:^LVEc3i:YL[Eg3e:QLcEo3]:_JUGa5k8[JYGe5g8WJ]Gj5b8TJaGk5^8SJeGm5[8PJhGP6T;001O000000010O000000001O01O0hLVB:j=CYB>g=_O[Ba0e=\\O^Bc0c=\\O^Bc0d=\\O]Bb0d=^O\\Ba0e=_O[B`0f=_O[B`0f=@ZB?h=@YB>h=AYB?g=AYB>h=AZB=g=BZB=h=BXB=i=BYB in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 16, 17, 18, 19, 20, 30, 31, 32, 33, 34, 35, 36, 45, 46, 47, 48, 49, 50, 51, 64, 65, 66, 78, 79, 80, 81, 82, 92, 93, 94, 95, 108, 109], "bbox": [0.05075294117647059, 0.0, 0.5008235294117647, 0.32134375], "iscrowd": 0, "area": 20342.918949999996, "rle": {"size": [640, 425], "counts": "[j=b0[c03M3M3I7G9G9K5L3N3L3@a0N1O2N1O20O100O1O100O100000000000000000001O000000000000000000000000000000hA]MS;c2kD`MT;`2kDbMT;_2jDcMU;]2iDeMW;[2hDgMW;Y2hDiMW;W2gDlMX;T2gDnMX;R2gDPNX;Q2gDPNX;P2gDRNX;n1gDSNY;m1gDTNX;l1gDVNX;j1gDXNX;i1fDYNY;g1gDZNX;f1gD\\NX;d1gD]NY;c1gD^NX;b1gD`NX;`1gDbNX;_1fDcNY;]1fDdNZ;\\1eDeN[;[1dDfN\\;Z1cDgN];Y1bDhN^;X1aDiN_;X1_DiNa;W1^DjNb;V1]DkNc;U1\\DlNd;T1[DmNe;S1ZDmNg;S1XDnNh;R1WDnNj;S1UDmNk;S1TDmNm;T1QDmNo;T1oClNR in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 16, 17, 18, 19, 20, 30, 31, 32, 33, 34, 35, 36, 45, 46, 47, 48, 49, 50, 51, 64, 65, 66, 78, 79, 80, 81, 82, 92, 93, 94, 95, 108, 109], "bbox": [0.05075294117647059, 0.0, 0.5008235294117647, 0.32134375], "iscrowd": 0, "area": 20342.918949999996, "rle": {"size": [640, 425], "counts": "[j=b0[c03M3M3I7G9G9K5L3N3L3@a0N1O2N1O20O100O1O100O100000000000000000001O000000000000000000000000000000hA]MS;c2kD`MT;`2kDbMT;_2jDcMU;]2iDeMW;[2hDgMW;Y2hDiMW;W2gDlMX;T2gDnMX;R2gDPNX;Q2gDPNX;P2gDRNX;n1gDSNY;m1gDTNX;l1gDVNX;j1gDXNX;i1fDYNY;g1gDZNX;f1gD\\NX;d1gD]NY;c1gD^NX;b1gD`NX;`1gDbNX;_1fDcNY;]1fDdNZ;\\1eDeN[;[1dDfN\\;Z1cDgN];Y1bDhN^;X1aDiN_;X1_DiNa;W1^DjNb;V1]DkNc;U1\\DlNd;T1[DmNe;S1ZDmNg;S1XDnNh;R1WDnNj;S1UDmNk;S1TDmNm;T1QDmNo;T1oClNR in this image.", "objects": [{"patches": [151, 152, 153, 154, 155, 156, 157, 165, 166, 167, 168, 169, 170, 171, 172, 173, 180, 181, 182, 183, 184, 185, 186, 187, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213], "bbox": [0.0, 0.43990625000000005, 0.5494117647058824, 0.6178281250000001], "iscrowd": 0, "area": 19688.024050000007, "rle": {"size": [640, 425], "counts": "S:U2ia02N2N2M4N11O000000000001O000O100O1O1O101N1O100O1000000000000N2M3M3M3M3N2O100O1000000O1000000O1000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000O0100000^M]_Og1c`0XN^_Og1c`0YN]_Og1c`0XN^_Oh1a`0YN__Og1a`0XN`_Oh1``0WNa_Oh1``0XN`_Oh1``0WNa_Oi1_`0VNb_Oj1]`0WNc_Oi1]`0VNd_Oi1]`0WNc_Oi1]`0VNd_Oj1\\`0UNe_Ok1[`0UNe_Ok1Z`0UNf_Ok1[`0gMS@Y2m?gMS@Y2m?gMS@Y2m?gMS@Y2m?gMS@X2n?hMQ@Y2P`0fMP@Z2Q`0eMo_O[2R`0dMn_O\\2S`0cMm_O\\2l`00O100000O10O100000O1000O1000O1000O1000O10O100000O01000O10O010000O10O0100O10O010000O010O10O010O00100O0010O01O0100O1000O0100O100O100O01000O100O100O010O10000O101N100O1O1N3N1O2N1O2N1O2N2M2O2N2N1N3M2N3M3M2N3M2N3J6J5Kefg3"}, "label": "the middle pizza with a piece missing"}]} {"id": 566301, "image": "COCO_train2014_000000566301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza on a white plate in between two other pizzas on plates\"."}], "task": "refering", "answer_template": "The \"pizza on a white plate in between two other pizzas on plates\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 154, 155, 156, 157, 165, 166, 167, 168, 169, 170, 171, 172, 173, 180, 181, 182, 183, 184, 185, 186, 187, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213], "bbox": [0.0, 0.43990625000000005, 0.5494117647058824, 0.6178281250000001], "iscrowd": 0, "area": 19688.024050000007, "rle": {"size": [640, 425], "counts": "S:U2ia02N2N2M4N11O000000000001O000O100O1O1O101N1O100O1000000000000N2M3M3M3M3N2O100O1000000O1000000O1000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000O0100000^M]_Og1c`0XN^_Og1c`0YN]_Og1c`0XN^_Oh1a`0YN__Og1a`0XN`_Oh1``0WNa_Oh1``0XN`_Oh1``0WNa_Oi1_`0VNb_Oj1]`0WNc_Oi1]`0VNd_Oi1]`0WNc_Oi1]`0VNd_Oj1\\`0UNe_Ok1[`0UNe_Ok1Z`0UNf_Ok1[`0gMS@Y2m?gMS@Y2m?gMS@Y2m?gMS@Y2m?gMS@X2n?hMQ@Y2P`0fMP@Z2Q`0eMo_O[2R`0dMn_O\\2S`0cMm_O\\2l`00O100000O10O100000O1000O1000O1000O1000O10O100000O01000O10O010000O10O0100O10O010000O010O10O010O00100O0010O01O0100O1000O0100O100O100O01000O100O100O010O10000O101N100O1O1N3N1O2N1O2N1O2N2M2O2N2N1N3M2N3M3M2N3M2N3J6J5Kefg3"}, "label": "pizza on a white plate in between two other pizzas on plates"}]} {"id": 418847, "image": "COCO_train2014_000000418847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bird stands behind two brown birds\"."}], "task": "refering", "answer_template": "The \"a white bird stands behind two brown birds\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 127, 128, 150, 151, 152, 153, 173, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201], "bbox": [0.510546875, 0.27675, 0.80525, 0.5303125000000001], "iscrowd": 0, "area": 8516.235449999996, "rle": {"size": [480, 640], "counts": "ali41o>1O0O101O0O10001N10001O0O101O00000O2O002N3M3L4M3L4M4K9G9G8H9H8G4M2N2N2O1N1O1O010O1O1O1O10O01O1O100O010001O000O5L6J2N001O1N1015KN10OL401O001O00010O001O00010O001O00010O0010O01O01O01O010O000N3O001N10001O0O2O001N10001O0O2O001O0O101O0O2O1O3L5L3M3L5L1O0O10000O100O100JQB[Oo=e0RBYOo=f0SBYOm=g060000O10001O0O10000O2O00000O2O00000O2O000O1G`A2`>MaA3_>LbA4^>KcA4^>KcA5]>JdA6c>00O10000O1000000O1000000O10000O1000000O100000TZk1"}, "label": "a white bird stands behind two brown birds"}]} {"id": 418847, "image": "COCO_train2014_000000418847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white bird standing in water\"."}], "task": "refering", "answer_template": "The \"white bird standing in water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 127, 128, 150, 151, 152, 153, 173, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201], "bbox": [0.510546875, 0.27675, 0.80525, 0.5303125000000001], "iscrowd": 0, "area": 8516.235449999996, "rle": {"size": [480, 640], "counts": "ali41o>1O0O101O0O10001N10001O0O101O00000O2O002N3M3L4M3L4M4K9G9G8H9H8G4M2N2N2O1N1O1O010O1O1O1O10O01O1O100O010001O000O5L6J2N001O1N1015KN10OL401O001O00010O001O00010O001O00010O0010O01O01O01O010O000N3O001N10001O0O2O001N10001O0O2O001O0O101O0O2O1O3L5L3M3L5L1O0O10000O100O100JQB[Oo=e0RBYOo=f0SBYOm=g060000O10001O0O10000O2O00000O2O00000O2O000O1G`A2`>MaA3_>LbA4^>KcA4^>KcA5]>JdA6c>00O10000O1000000O1000000O10000O1000000O100000TZk1"}, "label": "white bird standing in water"}]} {"id": 418847, "image": "COCO_train2014_000000418847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown birds with the long beak are in the water\"."}], "task": "refering", "answer_template": "The \"the brown birds with the long beak are in the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 193, 194, 195, 196, 197, 217, 218, 219, 220, 221, 222, 241, 242, 243, 244, 264, 265, 266, 288], "bbox": [0.352953125, 0.41602083333333334, 0.66271875, 0.7252708333333333], "iscrowd": 0, "area": 7700.212150000003, "rle": {"size": [480, 640], "counts": "SeZ31nm000000PA010O010O0010O010O01O010O01O010O01O010O010O0010O010O01O010O010O0010O010O010O01O01O01O001O001O001O001O0O2O0O2N101N1O2O0O2SBCU=>bBOZ=2aB7Z=k0L3M2O2M2O1O1O2N1O1O2N101N101O0O20O1N2O0jM^Cm1lOjA0W>OjA0V>OlA0U>OlA0T>1lAMV>2jANV>2jAMX>3hAKY>5gAJ[>49MdoT3"}, "label": "the brown birds with the long beak are in the water"}]} {"id": 295970, "image": "COCO_train2014_000000295970.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"rotten fruit with a face carved into it\"."}], "task": "refering", "answer_template": "The \"rotten fruit with a face carved into it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 194, 195, 196, 209, 210, 211, 212, 213, 214, 226, 227, 228, 229, 230, 231, 232, 242, 243, 244, 245, 246, 247, 248, 249, 259, 260, 261, 262, 263, 264, 265, 277, 278, 279, 280, 281, 282, 294, 295, 296, 297, 298], "bbox": [0.27939583333333334, 0.48021874999999997, 0.6539583333333333, 0.7844687499999999], "iscrowd": 0, "area": 26253.071100000005, "rle": {"size": [640, 480], "counts": "aTd23ic0 in this image.", "objects": [{"patches": [193, 194, 195, 196, 209, 210, 211, 212, 213, 214, 226, 227, 228, 229, 230, 231, 232, 242, 243, 244, 245, 246, 247, 248, 249, 259, 260, 261, 262, 263, 264, 265, 277, 278, 279, 280, 281, 282, 294, 295, 296, 297, 298], "bbox": [0.27939583333333334, 0.48021874999999997, 0.6539583333333333, 0.7844687499999999], "iscrowd": 0, "area": 26253.071100000005, "rle": {"size": [640, 480], "counts": "aTd23ic0 in this image.", "objects": [{"patches": [97, 118, 119, 120, 141, 164, 187, 210, 233, 234, 257], "bbox": [0.14040625, 0.32065268065268066, 0.2568125, 0.8156410256410256], "iscrowd": 0, "area": 2139.85005, "rle": {"size": [429, 640], "counts": "hlU1f0ch4B^K=]4CiK;T4DSL9i3F^L9]3GiL7S3ISM6h2J_M3]2MoMLl14f51O01O001O1O010O100O010O10O0100O010O1000O0100O010O10O0100O10O01O10O01O1O010O1O1O2OQYW6"}, "label": "the back of the chair that the girl in the purple t - shirt is sitting in"}]} {"id": 58403, "image": "COCO_train2014_000000058403.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair with a metal frame\"."}], "task": "refering", "answer_template": "The \"a chair with a metal frame\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 118, 119, 120, 141, 164, 187, 210, 233, 234, 257], "bbox": [0.14040625, 0.32065268065268066, 0.2568125, 0.8156410256410256], "iscrowd": 0, "area": 2139.85005, "rle": {"size": [429, 640], "counts": "hlU1f0ch4B^K=]4CiK;T4DSL9i3F^L9]3GiL7S3ISM6h2J_M3]2MoMLl14f51O01O001O1O010O100O010O10O0100O010O1000O0100O010O10O0100O10O01O10O01O1O010O1O1O2OQYW6"}, "label": "a chair with a metal frame"}]} {"id": 58403, "image": "COCO_train2014_000000058403.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in blue shirt\"."}], "task": "refering", "answer_template": "The \"woman in blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 61, 62, 63, 64, 65, 84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 112, 128, 129, 130, 131, 132, 133, 134, 135, 151, 152, 153, 154, 155, 156, 157, 158, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247], "bbox": [0.55365625, 0.08009324009324009, 0.90884375, 0.7242424242424242], "iscrowd": 0, "area": 42915.61189999999, "rle": {"size": [429, 640], "counts": "and47T=6I7J6J7I6J4L3M3M3M3M3M3K5K5K5J6K5K5K5K5J6K5VFmLT9P4D6I5M2O2N2N2N1O2N2N2N2N3M200O100O1O100O101N101N2O1N2O0O010O1O10O0100O100O010O100O10O0100O10O010000O10O10O1000O010000O0]OTJgIl5R6`JiI`5Q6jJjIW5o5W1J501O1O1O010O1O1O1O001O1O1O001O1O1O1O010O1O1O001O1O1O1O001O1O1O01000000O100000000O1000000O100000000O1000000O100000000O1000000O1000001O0O2O001O001N101O1O001N101O001N101O1O001N4M2N3M2M3N3M2M4M2cJ\\Im3h6gK`IX4i6XK`Ig4d7O1N101N2O1O0O2O1O1N101O1N2O001N2O10O1000000O1000O10ON3K5J8H7J7H7I8H7J7H8H7J7H7I8H7J\\Th0"}, "label": "woman in blue shirt"}]} {"id": 58403, "image": "COCO_train2014_000000058403.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl wearing blue colored top holding a spoon in her hands\"."}], "task": "refering", "answer_template": "The \"a girl wearing blue colored top holding a spoon in her hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 61, 62, 63, 64, 65, 84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 112, 128, 129, 130, 131, 132, 133, 134, 135, 151, 152, 153, 154, 155, 156, 157, 158, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247], "bbox": [0.55365625, 0.08009324009324009, 0.90884375, 0.7242424242424242], "iscrowd": 0, "area": 42915.61189999999, "rle": {"size": [429, 640], "counts": "and47T=6I7J6J7I6J4L3M3M3M3M3M3K5K5K5J6K5K5K5K5J6K5VFmLT9P4D6I5M2O2N2N2N1O2N2N2N2N3M200O100O1O100O101N101N2O1N2O0O010O1O10O0100O100O010O100O10O0100O10O010000O10O10O1000O010000O0]OTJgIl5R6`JiI`5Q6jJjIW5o5W1J501O1O1O010O1O1O1O001O1O1O001O1O1O1O010O1O1O001O1O1O1O001O1O1O01000000O100000000O1000000O100000000O1000000O100000000O1000000O1000001O0O2O001O001N101O1O001N101O001N101O1O001N4M2N3M2M3N3M2M4M2cJ\\Im3h6gK`IX4i6XK`Ig4d7O1N101N2O1O0O2O1O1N101O1N2O001N2O10O1000000O1000O10ON3K5J8H7J7H7I8H7J7H8H7J7H7I8H7J\\Th0"}, "label": "a girl wearing blue colored top holding a spoon in her hands"}]} {"id": 58403, "image": "COCO_train2014_000000058403.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl in purple t - shirt\"."}], "task": "refering", "answer_template": "The \"girl in purple t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 31, 51, 52, 53, 54, 55, 74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 195, 210, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 303, 304, 305, 306, 307, 308, 326, 327, 328, 329, 330], "bbox": [0.15584375, 0.058135198135198136, 0.49024999999999996, 0.9881585081585081], "iscrowd": 0, "area": 57617.44599999999, "rle": {"size": [429, 640], "counts": "ZSZ17P=;F:E;F:F:E;J6J6K5J6J6J6K5J6J6K4K6J6J6K5J6J6K5J6J6J6K5J6J6J6K5J6J6K5J6J6J6aJ`Hg4S8O1O1O1O100O1O1O100O1O1O100O1^McG3^8KRHFo79[H]Of7a0_H[Ob7d0bHWO`7h0cHUO^7i0gHRO[7m0iHoNX7o0lHmNV7R1mHkNS7T1RIgNP7X1SIeNn6Y1WIcNj6\\1YI`Ni6^1[I_Nf6a1c2O10000O100O100O100O100O100O10000O100O100O01000O0100000O01000O0100000O01000O0100000O01000O0100000O01000O010001O1N2O001N2O1O1O0O2O1O1N101O1O1N101O1N2O001O1N2O001O1N7J9G9C=B>A?B>B;E;D in this image.", "objects": [{"patches": [28, 29, 30, 31, 51, 52, 53, 54, 55, 74, 75, 76, 77, 78, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 195, 210, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 303, 304, 305, 306, 307, 308, 326, 327, 328, 329, 330], "bbox": [0.15584375, 0.058135198135198136, 0.49024999999999996, 0.9881585081585081], "iscrowd": 0, "area": 57617.44599999999, "rle": {"size": [429, 640], "counts": "ZSZ17P=;F:E;F:F:E;J6J6K5J6J6J6K5J6J6K4K6J6J6K5J6J6K5J6J6J6K5J6J6J6K5J6J6K5J6J6J6aJ`Hg4S8O1O1O1O100O1O1O100O1O1O100O1^McG3^8KRHFo79[H]Of7a0_H[Ob7d0bHWO`7h0cHUO^7i0gHRO[7m0iHoNX7o0lHmNV7R1mHkNS7T1RIgNP7X1SIeNn6Y1WIcNj6\\1YI`Ni6^1[I_Nf6a1c2O10000O100O100O100O100O100O10000O100O100O01000O0100000O01000O0100000O01000O0100000O01000O0100000O01000O010001O1N2O001N2O1O1O0O2O1O1N101O1O1N101O1N2O001O1N2O001O1N7J9G9C=B>A?B>B;E;D in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 113, 114, 115, 116, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 137, 138, 139, 143, 144, 145, 146, 147, 148, 149, 150, 152, 153, 154, 155, 156, 157, 160, 161, 167, 168, 169, 170, 171, 172, 173, 175, 176, 177, 178, 179, 180, 183, 184, 190, 191, 192, 193, 201, 202, 203, 207], "bbox": [0.0016406250000000002, 0.0, 1.0, 0.6641920374707261], "iscrowd": 0, "area": 101001.38200000001, "rle": {"size": [427, 640], "counts": "\\e0i0n:a1^Nb1_Na1^Nb1dN\\13M3M3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M3M3M3N2M3M3M3N3L3M3M3M3N2M3M3M3M3N2M3M3M3N2M3M3M3M4M2M3M3M3M3N2N20000O10000O10000O10001N10000O1000000O10000O10000O10000O10001N10000O1000000O10000O10000O10000O101O0O10000O10000O1000000O1000000001O100O1O1O1O1O001O1O1O1O1O1O0iG_Nd4b1YKcNd4^1XKhNe4Y1XKlNe4U1XKPOe4Q1XKTOf4l0WKYOf4h0VK^Og4e0TK_Oj4Z1[JjNc5P2aIUN\\6d2iH`MU7]4N2N2N2N2N3M2N2N2N2N2N3N1N2N2N2N2N2N3M2N2N2N2N1OO10000O10000O2O000O10000O10000O10000O10000O10000O2O000O10000O10000O10000O10000O10000O2O000O10000O10000O10000O10000O10000O2O000O10000O10000O10000O1000000000001O0O100000000000000000000000000000000O2O0000000000000000000000000000000O101O00000000000000000000000000000O10001O000000000000000000000000000O1000001O000000000000000000000I7\\Od0]Oc0J6J6J6I7J7O000001O00001O0000001O00001O00000010O0001O0000001O00001O0000001O005K5K4M4K4L5K5K4L5K2N1O1O1O2N1O1O1O1O1O2N100O1O1O1O2N1O1O00001O0000000000000001O000000000000000001O000000000000000001O000000000000000001O000000001O000010O0001O00001O0000001O00001O00001O00010O00001O0000001O00001O00001O00000010O0001O0000001O0000001O0000lLiITOW67`JH_51jJMV53kJLU53nJLR52QKLo43TKKl45VKIj46XKJg45\\KId46_KHa48aKF_49cKG\\48gKFY4:iKDW4;lKDS4;PLAR4>PL_OR4a0PL[OR4d0QLXOQ4g0RLUOP4k0QLROQ4m0RLPOo3o0cL^N`3`1V400O1O100001O001O00001O001O00001O010O001O00001O001O00001O001O001O000010O01O00001O001O001O00001O1O2N1O2N1O101N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O4L4M3L3M4L4L4L4L4L4L3M2N2N2N2N2N2N2N2N2N2O1N2N2N2NWMf1"}, "label": "the women is wearing blazer"}]} {"id": 230436, "image": "COCO_train2014_000000230436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"window on a train with a man standing outside\"."}], "task": "refering", "answer_template": "The \"window on a train with a man standing outside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 113, 114, 115, 116, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 137, 138, 139, 143, 144, 145, 146, 147, 148, 149, 150, 152, 153, 154, 155, 156, 157, 160, 161, 167, 168, 169, 170, 171, 172, 173, 175, 176, 177, 178, 179, 180, 183, 184, 190, 191, 192, 193, 201, 202, 203, 207], "bbox": [0.0016406250000000002, 0.0, 1.0, 0.6641920374707261], "iscrowd": 0, "area": 101001.38200000001, "rle": {"size": [427, 640], "counts": "\\e0i0n:a1^Nb1_Na1^Nb1dN\\13M3M3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M3M3M3N2M3M3M3N3L3M3M3M3N2M3M3M3M3N2M3M3M3N2M3M3M3M4M2M3M3M3M3N2N20000O10000O10000O10001N10000O1000000O10000O10000O10000O10001N10000O1000000O10000O10000O10000O101O0O10000O10000O1000000O1000000001O100O1O1O1O1O001O1O1O1O1O1O0iG_Nd4b1YKcNd4^1XKhNe4Y1XKlNe4U1XKPOe4Q1XKTOf4l0WKYOf4h0VK^Og4e0TK_Oj4Z1[JjNc5P2aIUN\\6d2iH`MU7]4N2N2N2N2N3M2N2N2N2N2N3N1N2N2N2N2N2N3M2N2N2N2N1OO10000O10000O2O000O10000O10000O10000O10000O10000O2O000O10000O10000O10000O10000O10000O2O000O10000O10000O10000O10000O10000O2O000O10000O10000O10000O1000000000001O0O100000000000000000000000000000000O2O0000000000000000000000000000000O101O00000000000000000000000000000O10001O000000000000000000000000000O1000001O000000000000000000000I7\\Od0]Oc0J6J6J6I7J7O000001O00001O0000001O00001O00000010O0001O0000001O00001O0000001O005K5K4M4K4L5K5K4L5K2N1O1O1O2N1O1O1O1O1O2N100O1O1O1O2N1O1O00001O0000000000000001O000000000000000001O000000000000000001O000000000000000001O000000001O000010O0001O00001O0000001O00001O00001O00010O00001O0000001O00001O00001O00000010O0001O0000001O0000001O0000lLiITOW67`JH_51jJMV53kJLU53nJLR52QKLo43TKKl45VKIj46XKJg45\\KId46_KHa48aKF_49cKG\\48gKFY4:iKDW4;lKDS4;PLAR4>PL_OR4a0PL[OR4d0QLXOQ4g0RLUOP4k0QLROQ4m0RLPOo3o0cL^N`3`1V400O1O100001O001O00001O001O00001O010O001O00001O001O00001O001O001O000010O01O00001O001O001O00001O1O2N1O2N1O101N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O4L4M3L3M4L4L4L4L4L4L3M2N2N2N2N2N2N2N2N2N2O1N2N2N2NWMf1"}, "label": "window on a train with a man standing outside"}]} {"id": 230436, "image": "COCO_train2014_000000230436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lap top that a man is using\"."}], "task": "refering", "answer_template": "The \"a lap top that a man is using\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 150, 151, 173, 174, 193, 194, 195, 196, 197, 216, 217], "bbox": [0.3868125, 0.33707259953161595, 0.61171875, 0.6112412177985949], "iscrowd": 0, "area": 4930.64245, "rle": {"size": [427, 640], "counts": "YeW33X=0O1O1N2O1O1O1O2M2O1O1O1O1N2O1O1O1O0O2O1000000000O10000000000O10000000000O10O10001O000000001O000000001O000000O1000O1000O10000O10000O1000000O1000O010000O11O00000000001O0000000O2O0000000N3M2N2N2N2L4L5K4K5M3N2M4M2K5I6J4L5M3000O0100O001O12M3M3M3M3M3L5L3M3M3M3M3L4M3M3M3M3L4M4L3M3M3M4K6K5K4LcXW3"}, "label": "a lap top that a man is using"}]} {"id": 230436, "image": "COCO_train2014_000000230436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver laptop being used by a man\"."}], "task": "refering", "answer_template": "The \"a silver laptop being used by a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 150, 151, 173, 174, 193, 194, 195, 196, 197, 216, 217], "bbox": [0.3868125, 0.33707259953161595, 0.61171875, 0.6112412177985949], "iscrowd": 0, "area": 4930.64245, "rle": {"size": [427, 640], "counts": "YeW33X=0O1O1N2O1O1O1O2M2O1O1O1O1N2O1O1O1O0O2O1000000000O10000000000O10000000000O10O10001O000000001O000000001O000000O1000O1000O10000O10000O1000000O1000O010000O11O00000000001O0000000O2O0000000N3M2N2N2N2L4L5K4K5M3N2M4M2K5I6J4L5M3000O0100O001O12M3M3M3M3M3L5L3M3M3M3M3L4M3M3M3M3L4M4L3M3M3M4K6K5K4LcXW3"}, "label": "a silver laptop being used by a man"}]} {"id": 230436, "image": "COCO_train2014_000000230436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue and yellow train\"."}], "task": "refering", "answer_template": "The \"a blue and yellow train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 110, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 143, 144, 145, 146, 147, 149, 150, 152, 153, 154, 155, 156, 167, 168, 169, 170, 172, 173, 175, 176, 177, 179], "bbox": [0.21590625000000002, 0.2853864168618267, 0.8231093749999999, 0.4786416861826698], "iscrowd": 0, "area": 21945.206850000002, "rle": {"size": [427, 640], "counts": "]fi11Z=1N3N2N1N3N2N2M2O2N2M2O2N2M3N1O2M3N1O2M3N2N1N3N2N1M3L3N2M3M3N2KnMnDU2R;2O1000001O0000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000N2N1O2lNT1K5K5L4K5M300O100000000O1001O1O001O1O1O1O2N2N3M2N2N2N2N2N2N2N3M2M3N2N2N2N2N3M2N2N2N2N2N2N3M2N2N1O000000000000O1000000000000O1000000000O10O1000000000000N2L4L4L4L4L4L4L4L4L4L4L4L4L4O1O100O10000O1000000O1000O10O100000000O1000000O100?Aa0_O`0@`0@00000000000000000000000000000000000O1000000000000000000000000000000000001O0000000000000000000000000000000O10000000000000000000000000000000000000K5_Oa0_Oa0_Oa0_Ojl:o0bQEV1000000000O2O00000000000O2O00000000000O2O000000000oM_F8b9CmQ_1"}, "label": "a blue and yellow train"}]} {"id": 230436, "image": "COCO_train2014_000000230436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue train in the background\"."}], "task": "refering", "answer_template": "The \"the blue train in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 110, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 143, 144, 145, 146, 147, 149, 150, 152, 153, 154, 155, 156, 167, 168, 169, 170, 172, 173, 175, 176, 177, 179], "bbox": [0.21590625000000002, 0.2853864168618267, 0.8231093749999999, 0.4786416861826698], "iscrowd": 0, "area": 21945.206850000002, "rle": {"size": [427, 640], "counts": "]fi11Z=1N3N2N1N3N2N2M2O2N2M2O2N2M3N1O2M3N1O2M3N2N1N3N2N1M3L3N2M3M3N2KnMnDU2R;2O1000001O0000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000N2N1O2lNT1K5K5L4K5M300O100000000O1001O1O001O1O1O1O2N2N3M2N2N2N2N2N2N2N3M2M3N2N2N2N2N3M2N2N2N2N2N2N3M2N2N1O000000000000O1000000000000O1000000000O10O1000000000000N2L4L4L4L4L4L4L4L4L4L4L4L4L4O1O100O10000O1000000O1000O10O100000000O1000000O100?Aa0_O`0@`0@00000000000000000000000000000000000O1000000000000000000000000000000000001O0000000000000000000000000000000O10000000000000000000000000000000000000K5_Oa0_Oa0_Oa0_Ojl:o0bQEV1000000000O2O00000000000O2O00000000000O2O000000000oM_F8b9CmQ_1"}, "label": "the blue train in the background"}]} {"id": 230436, "image": "COCO_train2014_000000230436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in black jacket working on a laptop\"."}], "task": "refering", "answer_template": "The \"a man in black jacket working on a laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 48, 49, 50, 51, 52, 71, 72, 73, 74, 75, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.0, 0.08990632318501171, 0.577625, 0.9870023419203747], "iscrowd": 0, "area": 74328.48835, "rle": {"size": [427, 640], "counts": "X8`2e:c2]M6J6K5J6J6J6_Oa0M3M3M3M3M3N2N2M3N2M3N2N2M3N2M3N2O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O100O1O1O100M3]Nc1O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1O100O1O100O1O100O100O100O100O100O100O100O100O11O0000001O0000001O0000001O0000001O0000001O0000001O0kGaN^4_1bKeNZ4[1fKhNW4Y1hKkNT4U1lKnNQ4R1oKPOo3P1QLROm3o0RLSOl3m0SLVOk3k0TLVOk3k0TLWOj3k0TLWOj3j0ULXOi3i0VLXOi3j0TLYOj3h0SL\\Ok3f0RL]Ol3d0QL@m3b0PLCl3>RLGj3;RLLi35UL1f31VL5f3LXL:c3G[LP7CPI=n6DSI]5_OfJa0Y5]OjJc0U5[OnJe0R5WORKi0m4UOVKk0i4SOZKm0f4PO]KP1b4nNaKR1_4mNcKR1\\4nNeKR1[4mNfKS1Y4mNhKS1X4lNiKT1W4kNjKU1U4kNlKU1T4jNmKV1S4iNnKW1Q4iNPLW1P4hNQLX1o3gNSLX1l3hNULX1k3gNVLY1j3fNWLZ1h3fNYLZ1g3eNZL[1f3dN[L\\1d3dN]L\\1c3cN^L]1e3_N\\La1f3\\N[Ld1g3YN[Lf1g3WNZLi1h3TNYLl1i3RNWLn1l3nMULR2m3kMTLU2n3hMSLX2`7000001O000000000000000000001O0000000000000000001O000000000000000000001O000000000000000000001O0000000000000000001O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1L4M3M4K6K4L4L4K5L4L_V`3"}, "label": "a man in black jacket working on a laptop"}]} {"id": 230436, "image": "COCO_train2014_000000230436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in suit looking at laptop\"."}], "task": "refering", "answer_template": "The \"man in suit looking at laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 48, 49, 50, 51, 52, 71, 72, 73, 74, 75, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.0, 0.08990632318501171, 0.577625, 0.9870023419203747], "iscrowd": 0, "area": 74328.48835, "rle": {"size": [427, 640], "counts": "X8`2e:c2]M6J6K5J6J6J6_Oa0M3M3M3M3M3N2N2M3N2M3N2N2M3N2M3N2O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O100O1O1O100M3]Nc1O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1O100O1O100O1O100O100O100O100O100O100O100O100O11O0000001O0000001O0000001O0000001O0000001O0000001O0kGaN^4_1bKeNZ4[1fKhNW4Y1hKkNT4U1lKnNQ4R1oKPOo3P1QLROm3o0RLSOl3m0SLVOk3k0TLVOk3k0TLWOj3k0TLWOj3j0ULXOi3i0VLXOi3j0TLYOj3h0SL\\Ok3f0RL]Ol3d0QL@m3b0PLCl3>RLGj3;RLLi35UL1f31VL5f3LXL:c3G[LP7CPI=n6DSI]5_OfJa0Y5]OjJc0U5[OnJe0R5WORKi0m4UOVKk0i4SOZKm0f4PO]KP1b4nNaKR1_4mNcKR1\\4nNeKR1[4mNfKS1Y4mNhKS1X4lNiKT1W4kNjKU1U4kNlKU1T4jNmKV1S4iNnKW1Q4iNPLW1P4hNQLX1o3gNSLX1l3hNULX1k3gNVLY1j3fNWLZ1h3fNYLZ1g3eNZL[1f3dN[L\\1d3dN]L\\1c3cN^L]1e3_N\\La1f3\\N[Ld1g3YN[Lf1g3WNZLi1h3TNYLl1i3RNWLn1l3nMULR2m3kMTLU2n3hMSLX2`7000001O000000000000000000001O0000000000000000001O000000000000000000001O000000000000000000001O0000000000000000001O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1L4M3M4K6K4L4L4K5L4L_V`3"}, "label": "man in suit looking at laptop"}]} {"id": 230436, "image": "COCO_train2014_000000230436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman wearing a pink scarf working on a laptop computer\"."}], "task": "refering", "answer_template": "The \"woman wearing a pink scarf working on a laptop computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 110, 111, 112, 113, 133, 134, 135, 136, 137, 157, 158, 159, 160, 177, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.571234375, 0.20381733021077283, 1.0, 0.9858313817330212], "iscrowd": 0, "area": 47952.04034999998, "rle": {"size": [427, 640], "counts": "lPi42V=4L3N2M3N2M3N2N21O001O001O00O100000000000000O10000000000O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2N2N2M3N2N2N2N2N2N2N2@`0O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1QMkLgKV3Y4kLeKV3[4kLbKW3^4iLaKX3_4iL_KX3a4iL]KX3b4jL\\KW3d4jLZKW3f4iLYKX3g4iLWKX3j4hLSKZ3m4gLQKZ3o4fLPK[3Q5eLmJ\\3S5eLlJ[3T5fLkJZ3U5fLkJZ3V5eLjJ[3V5fLiJZ3W5fLiJZ3X5eLgJ\\3Y5dLgJ\\3Y5eLfJ[3Z5eLfJ[3[5dLeJ\\3[5eLdJ[3\\5eLdJ[3]5dLcJ\\3]5dLcJ\\3]5eLbJ[3_5dLaJ\\3_5dLaJ\\3_5eL`J[3`5eL_J\\3b5cL^J]3b5cL^J]3b5dL]J\\3d5cL\\J]3d5cL\\J]3d5dL[J\\3e5fLYJZ3h5fLWJZ3i5gLWJX3i5iLVJW3k5jLSJV3m5kLRJU3n5lLRJS3n5nLQJR3P6nLoIR3Q6PMmIP3S6QMlIo2U6QMkIn2U6SMjIm2W6TMgIl2Y6UMfIk2[6UMeIj2\\6VMcIj2]6WMbIi2_6XM_Ih2a6YM^Ig2c6YM]If2d6ZM[If2e6\\MYId2h6\\MWId2i6]MWIb2j6^MUIb2l6_MRIa2n6`MQI`2P7aMoH^2R7bMmH^2S7dMkH\\2V7dMiH\\2W7fMgHZ2U4aLYNU1`M[2U4dLZNR1[M^2X4dL\\No0WMa2[4dL]Nl0RMd2^4eL_Nf4_1^K_Nb4_1aKaN^4]1gK_NZ4^1mK]NT4a1QL[NP4S1eLiN\\3W1eLgN\\3X1fLfN[3Y1fLeN\\3Z1eLeN\\3Z1eLeN\\3Z1eLeN\\3[1dLdN]3[1dLdN]3[1dLdN]3[1eLcN\\3\\1eLcN\\3]1dLaN^3^1cL`N_3_1bLZNe3e1\\LTNk3k1VLoMP4P2RLiMT4W2lKcMZ4\\2gK_M^4`2cKZMc4f2]KTMi4k2XKPMm4P3g200000000O10000000000O10000000000O10000000000O10000000000O1000000001O001O001O001O001O001O001O001O2N3M4L3M4L3M4L4L4L4L4L3M4L4L4L4L4L3M2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2N2N2N2NV1jNi3WLhD"}, "label": "woman wearing a pink scarf working on a laptop computer"}]} {"id": 230436, "image": "COCO_train2014_000000230436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady using a laptop\"."}], "task": "refering", "answer_template": "The \"a lady using a laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 110, 111, 112, 113, 133, 134, 135, 136, 137, 157, 158, 159, 160, 177, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.571234375, 0.20381733021077283, 1.0, 0.9858313817330212], "iscrowd": 0, "area": 47952.04034999998, "rle": {"size": [427, 640], "counts": "lPi42V=4L3N2M3N2M3N2N21O001O001O00O100000000000000O10000000000O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2N2N2M3N2N2N2N2N2N2N2@`0O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1QMkLgKV3Y4kLeKV3[4kLbKW3^4iLaKX3_4iL_KX3a4iL]KX3b4jL\\KW3d4jLZKW3f4iLYKX3g4iLWKX3j4hLSKZ3m4gLQKZ3o4fLPK[3Q5eLmJ\\3S5eLlJ[3T5fLkJZ3U5fLkJZ3V5eLjJ[3V5fLiJZ3W5fLiJZ3X5eLgJ\\3Y5dLgJ\\3Y5eLfJ[3Z5eLfJ[3[5dLeJ\\3[5eLdJ[3\\5eLdJ[3]5dLcJ\\3]5dLcJ\\3]5eLbJ[3_5dLaJ\\3_5dLaJ\\3_5eL`J[3`5eL_J\\3b5cL^J]3b5cL^J]3b5dL]J\\3d5cL\\J]3d5cL\\J]3d5dL[J\\3e5fLYJZ3h5fLWJZ3i5gLWJX3i5iLVJW3k5jLSJV3m5kLRJU3n5lLRJS3n5nLQJR3P6nLoIR3Q6PMmIP3S6QMlIo2U6QMkIn2U6SMjIm2W6TMgIl2Y6UMfIk2[6UMeIj2\\6VMcIj2]6WMbIi2_6XM_Ih2a6YM^Ig2c6YM]If2d6ZM[If2e6\\MYId2h6\\MWId2i6]MWIb2j6^MUIb2l6_MRIa2n6`MQI`2P7aMoH^2R7bMmH^2S7dMkH\\2V7dMiH\\2W7fMgHZ2U4aLYNU1`M[2U4dLZNR1[M^2X4dL\\No0WMa2[4dL]Nl0RMd2^4eL_Nf4_1^K_Nb4_1aKaN^4]1gK_NZ4^1mK]NT4a1QL[NP4S1eLiN\\3W1eLgN\\3X1fLfN[3Y1fLeN\\3Z1eLeN\\3Z1eLeN\\3Z1eLeN\\3[1dLdN]3[1dLdN]3[1dLdN]3[1eLcN\\3\\1eLcN\\3]1dLaN^3^1cL`N_3_1bLZNe3e1\\LTNk3k1VLoMP4P2RLiMT4W2lKcMZ4\\2gK_M^4`2cKZMc4f2]KTMi4k2XKPMm4P3g200000000O10000000000O10000000000O10000000000O10000000000O1000000001O001O001O001O001O001O001O001O2N3M4L3M4L3M4L4L4L4L4L3M4L4L4L4L4L3M2N2N2N2N3M2N2N2N2N2N2N2N2N2N2N2N2N2N2N2NV1jNi3WLhD"}, "label": "a lady using a laptop"}]} {"id": 230436, "image": "COCO_train2014_000000230436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the laptop on which the woman is working\"."}], "task": "refering", "answer_template": "The \"the laptop on which the woman is working\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 128, 129, 151, 152, 174, 175, 197, 198, 199, 200, 201, 220, 221, 222, 223], "bbox": [0.5726093750000001, 0.324192037470726, 0.78125, 0.6140281030444965], "iscrowd": 0, "area": 4384.155399999998, "rle": {"size": [427, 640], "counts": "cUi41Z=00001PEV1]7jN`HY1`7gN]H\\1c7dN[H^1d7cNYHa1f7`NVHc1j7]NTHe1l7[NQHh1o7XNnGk1R8VNjGm1V8SNhGo1W8RNfGQ2Z8oMcGT2]8lMaGV2_8kM]GX2c8hMZG[2f8eMXG]2h8cMUGa2i8aMSGb2m8[MSGh2m8k000000000O1\\1dN]1cN2N2N2N2O0O0000000000000000O10000000001O000000000O1000000000000000000O1000000000000000000O1O100O1O11O3M2N2N3M2N1O1O001N10001O001O000000O1000000O10O100O00100O00100000O1001O0000001O0000001O00000N3N1O1O2N1O1O^Tj1"}, "label": "the laptop on which the woman is working"}]} {"id": 230436, "image": "COCO_train2014_000000230436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the laptop being used by the woman\"."}], "task": "refering", "answer_template": "The \"the laptop being used by the woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 128, 129, 151, 152, 174, 175, 197, 198, 199, 200, 201, 220, 221, 222, 223], "bbox": [0.5726093750000001, 0.324192037470726, 0.78125, 0.6140281030444965], "iscrowd": 0, "area": 4384.155399999998, "rle": {"size": [427, 640], "counts": "cUi41Z=00001PEV1]7jN`HY1`7gN]H\\1c7dN[H^1d7cNYHa1f7`NVHc1j7]NTHe1l7[NQHh1o7XNnGk1R8VNjGm1V8SNhGo1W8RNfGQ2Z8oMcGT2]8lMaGV2_8kM]GX2c8hMZG[2f8eMXG]2h8cMUGa2i8aMSGb2m8[MSGh2m8k000000000O1\\1dN]1cN2N2N2N2O0O0000000000000000O10000000001O000000000O1000000000000000000O1000000000000000000O1O100O1O11O3M2N2N3M2N1O1O001N10001O001O000000O1000000O10O100O00100O00100000O1001O0000001O0000001O00000N3N1O1O2N1O1O^Tj1"}, "label": "the laptop being used by the woman"}]} {"id": 238630, "image": "COCO_train2014_000000238630.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a made bed to the right of the photo\"."}], "task": "refering", "answer_template": "The \"a made bed to the right of the photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.44701562499999997, 0.6067708333333334, 1.0, 0.9999791666666666], "iscrowd": 0, "area": 49379.22469999999, "rle": {"size": [480, 640], "counts": "i`V47\\>=J6J6J6J6H8H8K5M3N2N2M3N2M3N2N2O1N2O1O1N2O1O1N2O100O100O100O100O100O100O100O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O100000000O10000000000O10000000000O1000000000000O10000000000O100000000O10000O10000O10000O10000O10000O10000O1000000O1000000O100000000O1000000O100000000O1000000000000000000O1000000000000000000O10000000000O1000000O100000000O1000000O10000000000O10000KeKeE[4[:fKdEZ4[:hKdEX4[:7O1O1O100O1O1O1O1M3N2M3M3L4L4L4M3N2N2N2M30000001O00000000000000001O001O001O001O000000000000000000M300001O00000000000000001O000000001O0000000000000000O1001O000000001O0000001O000000001O00000000000000000000000000O10000000000000000O10000000000O100000000000\\JnFZ5Q9cJSG]5Y9O001O00001O001O"}, "label": "a made bed to the right of the photo"}]} {"id": 238630, "image": "COCO_train2014_000000238630.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed closest to the viewer\"."}], "task": "refering", "answer_template": "The \"the bed closest to the viewer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.44701562499999997, 0.6067708333333334, 1.0, 0.9999791666666666], "iscrowd": 0, "area": 49379.22469999999, "rle": {"size": [480, 640], "counts": "i`V47\\>=J6J6J6J6H8H8K5M3N2N2M3N2M3N2N2O1N2O1O1N2O1O1N2O100O100O100O100O100O100O100O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O100000000O10000000000O10000000000O1000000000000O10000000000O100000000O10000O10000O10000O10000O10000O10000O1000000O1000000O100000000O1000000O100000000O1000000000000000000O1000000000000000000O10000000000O1000000O100000000O1000000O10000000000O10000KeKeE[4[:fKdEZ4[:hKdEX4[:7O1O1O100O1O1O1O1M3N2M3M3L4L4L4M3N2N2N2M30000001O00000000000000001O001O001O001O000000000000000000M300001O00000000000000001O000000001O0000000000000000O1001O000000001O0000001O000000001O00000000000000000000000000O10000000000000000O10000000000O100000000000\\JnFZ5Q9cJSG]5Y9O001O00001O001O"}, "label": "the bed closest to the viewer"}]} {"id": 238630, "image": "COCO_train2014_000000238630.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed that is whiter than the other one\"."}], "task": "refering", "answer_template": "The \"the bed that is whiter than the other one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 222, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 303, 304, 305, 306, 307, 308, 309, 310, 311, 328, 329, 330, 331, 332, 352, 353, 354], "bbox": [0.208984375, 0.5528125, 0.69775, 0.9325833333333333], "iscrowd": 0, "area": 30795.51414999999, "rle": {"size": [480, 640], "counts": "oTo1k0R>8H4M3L4L4M2O100O1O1O00100O1O1O1O010O1O1O100O001O1O100O1O1O2N100O101O1N2O1O1N2O1O0O2O1O1N3N2M2O2N1N2O1O1O1O1O1O1N2O1O1O1O1O1O0O2O1O1O1O1O1O001N2O1O1O1O001O1N2O1O1O1O1O1O1O1N2O1O001O001O001O001O1O000000000_EdK^:^41O10O0100O1O100O1O100O100O1O100O1O1O100O1O1O100O1O010O100O100O100O1O100O100O1O100O00100O1O1O1O1O1O1O10O10O10000O1000000O01000O1N2N2M3M2N3M2O2M3M2100O10O10O100O100O01000O1O1N2O0O2O1N2O1O1O001O1O001O1O00100O001O1O001O11N10000O10000O10000O2O000O10000O2O000O10000O10000O10000O10000O10000O10000000000O2O0000000O101O1O1O1O1N2O1O00000O10000000000O100000000O10O10000O10O1000O10O10O01000O0100O10O10O10O01000A9H2N2M2O2NWcj2"}, "label": "the bed that is whiter than the other one"}]} {"id": 238630, "image": "COCO_train2014_000000238630.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white coloured bedspread on a be in the room\"."}], "task": "refering", "answer_template": "The \"a white coloured bedspread on a be in the room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 222, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 303, 304, 305, 306, 307, 308, 309, 310, 311, 328, 329, 330, 331, 332, 352, 353, 354], "bbox": [0.208984375, 0.5528125, 0.69775, 0.9325833333333333], "iscrowd": 0, "area": 30795.51414999999, "rle": {"size": [480, 640], "counts": "oTo1k0R>8H4M3L4L4M2O100O1O1O00100O1O1O1O010O1O1O100O001O1O100O1O1O2N100O101O1N2O1O1N2O1O0O2O1O1N3N2M2O2N1N2O1O1O1O1O1O1N2O1O1O1O1O1O0O2O1O1O1O1O1O001N2O1O1O1O001O1N2O1O1O1O1O1O1O1N2O1O001O001O001O001O1O000000000_EdK^:^41O10O0100O1O100O1O100O100O1O100O1O1O100O1O1O100O1O010O100O100O100O1O100O100O1O100O00100O1O1O1O1O1O1O10O10O10000O1000000O01000O1N2N2M3M2N3M2O2M3M2100O10O10O100O100O01000O1O1N2O0O2O1N2O1O1O001O1O001O1O00100O001O1O001O11N10000O10000O10000O2O000O10000O2O000O10000O10000O10000O10000O10000O10000000000O2O0000000O101O1O1O1O1N2O1O00000O10000000000O100000000O10O10000O10O1000O10O10O01000O0100O10O10O10O01000A9H2N2M2O2NWcj2"}, "label": "a white coloured bedspread on a be in the room"}]} {"id": 255016, "image": "COCO_train2014_000000255016.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person in the black shirt who ' s shoulder is only seen\"."}], "task": "refering", "answer_template": "The \"the person in the black shirt who ' s shoulder is only seen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [183, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 341, 342, 343, 344, 364, 365, 366, 367, 388, 389, 390], "bbox": [0.857859375, 0.43145833333333333, 1.0, 0.9820208333333333], "iscrowd": 0, "area": 17402.184600000008, "rle": {"size": [480, 640], "counts": "\\hQ82k>4L4L4L4M3L4L4L4L4L4L4M3L4iNW1\\Nd1J6J6K5J6K5M3M3M3M3N2M3M3N2M3M3N2M3M3N2M3M3M3N2M3M3N2M200O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1;Ef3ZLf3ZLiA"}, "label": "the person in the black shirt who ' s shoulder is only seen"}]} {"id": 255016, "image": "COCO_train2014_000000255016.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black shirt and half of an arm\"."}], "task": "refering", "answer_template": "The \"black shirt and half of an arm\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [183, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 341, 342, 343, 344, 364, 365, 366, 367, 388, 389, 390], "bbox": [0.857859375, 0.43145833333333333, 1.0, 0.9820208333333333], "iscrowd": 0, "area": 17402.184600000008, "rle": {"size": [480, 640], "counts": "\\hQ82k>4L4L4L4M3L4L4L4L4L4L4M3L4iNW1\\Nd1J6J6K5J6K5M3M3M3M3N2M3M3N2M3M3N2M3M3N2M3M3M3N2M3M3N2M200O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1;Ef3ZLf3ZLiA"}, "label": "black shirt and half of an arm"}]} {"id": 255016, "image": "COCO_train2014_000000255016.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a light blue shirt - dress wearing a large crocheted looking backpack\"."}], "task": "refering", "answer_template": "The \"a woman in a light blue shirt - dress wearing a large crocheted looking backpack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 92, 93, 94, 115, 116, 117, 138, 139, 140, 161, 162, 184, 185, 207, 208, 230, 232, 253, 255], "bbox": [0.001203125, 0.178125, 0.12262499999999998, 0.7012916666666666], "iscrowd": 0, "area": 7364.9653, "rle": {"size": [480, 640], "counts": "ic0W5U:I7TNoEUNY:]1nE_N^:S1gEiNe:i0`ETOl:?ZE\\OR;6UEEW;NlD1_;BdDB6J4L3M4L3M4L3M4L3M3M4L in this image.", "objects": [{"patches": [69, 70, 92, 93, 94, 115, 116, 117, 138, 139, 140, 161, 162, 184, 185, 207, 208, 230, 232, 253, 255], "bbox": [0.001203125, 0.178125, 0.12262499999999998, 0.7012916666666666], "iscrowd": 0, "area": 7364.9653, "rle": {"size": [480, 640], "counts": "ic0W5U:I7TNoEUNY:]1nE_N^:S1gEiNe:i0`ETOl:?ZE\\OR;6UEEW;NlD1_;BdDB6J4L3M4L3M4L3M4L3M3M4L in this image.", "objects": [{"patches": [95, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 254, 255, 256], "bbox": [0.0008281250000000001, 0.28429166666666666, 0.222265625, 0.6938333333333334], "iscrowd": 0, "area": 17214.739849999998, "rle": {"size": [480, 640], "counts": "Te0R2j<5N2N4L4M3Ln0RO2O1N2N2O1N101N2N2O0O100O100O100O100O100O100O101O001N101O001N101O0O2O1O001O001O000000001O00001O001O:F000000001O00000000000000O1000001O000O10O1O1QORKjGP5U8QKjGo4V8SKhGn4X8m0O010O02O000O1O1N3M2^OiFYKY9f4kFWKV9h4mFTKT9k4nFSKT9e4hFVK53T9d4VG\\Kk8Y4aFjKg0Kj8Y4`FkKg0Kj8X4aFmKY:R4hEnKY:P4hEoKY:P4hEPLX:n3jEQLX:m3>O1O2O1N2N4MO01lM\\ELd:IPFMQ:0VFLk90ZFLh91\\FMf91\\FMe94[FIh96YFFQ:3PFHQ:;PFCn9`0SF_On9?UF_Ok9`0WF_Oj9`0WF^Oj9a0XF^Oh9a0ZF^Og9`0[F_Oe9?^F@c9=a2N3M2O2O0O2O1N1O2OnXY7"}, "label": "the smaller of the two clay pots"}]} {"id": 566319, "image": "COCO_train2014_000000566319.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left pot in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the left pot in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 254, 255, 256], "bbox": [0.0008281250000000001, 0.28429166666666666, 0.222265625, 0.6938333333333334], "iscrowd": 0, "area": 17214.739849999998, "rle": {"size": [480, 640], "counts": "Te0R2j<5N2N4L4M3Ln0RO2O1N2N2O1N101N2N2O0O100O100O100O100O100O100O101O001N101O001N101O0O2O1O001O001O000000001O00001O001O:F000000001O00000000000000O1000001O000O10O1O1QORKjGP5U8QKjGo4V8SKhGn4X8m0O010O02O000O1O1N3M2^OiFYKY9f4kFWKV9h4mFTKT9k4nFSKT9e4hFVK53T9d4VG\\Kk8Y4aFjKg0Kj8Y4`FkKg0Kj8X4aFmKY:R4hEnKY:P4hEoKY:P4hEPLX:n3jEQLX:m3>O1O2O1N2N4MO01lM\\ELd:IPFMQ:0VFLk90ZFLh91\\FMf91\\FMe94[FIh96YFFQ:3PFHQ:;PFCn9`0SF_On9?UF_Ok9`0WF_Oj9`0WF^Oj9a0XF^Oh9a0ZF^Og9`0[F_Oe9?^F@c9=a2N3M2O2O0O2O1N1O2OnXY7"}, "label": "the left pot in the right hand picture"}]} {"id": 566319, "image": "COCO_train2014_000000566319.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a reddish planter filled with soil and plants just beginning to emerge\"."}], "task": "refering", "answer_template": "The \"a reddish planter filled with soil and plants just beginning to emerge\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 148, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 283, 284, 285, 286], "bbox": [0.266296875, 0.3662916666666666, 0.5106718750000001, 0.7662916666666667], "iscrowd": 0, "area": 23755.641400000008, "rle": {"size": [480, 640], "counts": "jo_21f>;E:L5L3lMG\\E>_:OTE6g:=gDFV;U2K4M4K4M3L4M2N3N2N2O0O2O1N101N2N2O0O2O0O2O1N2O2M2N2O1N2O1N3N1N2O1N2O1N1O2O1N101O1O001O001O1O001O000000000O010000000O10000000000O100000000O100000000O10O1000001O0O101O001N101O001O0O101O001O0O2O00001O0O2O0O2O0O2O0O2O1N2O0O2O1O1N2N1N3N2M3M3M201N2O1O2M2O1N2O1N2O2N1N2O1N2O1O2M2O1N3N1N6H`0_Oa0@_1\\N9M4L4K4M4L`]b4"}, "label": "a reddish planter filled with soil and plants just beginning to emerge"}]} {"id": 83000, "image": "COCO_train2014_000000083000.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a metal bowl with dough in it\"."}], "task": "refering", "answer_template": "The \"a metal bowl with dough in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 290, 306, 307, 308, 309, 310, 311, 312, 313, 329, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.328671875, 0.7799063231850116, 0.6555156249999999, 1.0], "iscrowd": 0, "area": 15172.234800000002, "rle": {"size": [427, 640], "counts": "Wog24S=6K4K5L4L4M2N2N2N2N2N2N2O1N2N2N2N2N2O1O100O1O1O1O1O1O1O100O1O1O100O1O1O1O100O1O1O10O0100O100O100O100O100O100O100O10000O100O100O10000O10000O10000O10000O10000O100000000O10000000000O100000000O100000000O100000000O100000000O100000000O0100000000000000000000000000000000000000000000000001O0000001O000000001O00001O001O001O001O001O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O001O2N1O2N1O2N1O2N1O2N1O2N2FiC\\OX<`0mC^OiTl2"}, "label": "a metal bowl with dough in it"}]} {"id": 83005, "image": "COCO_train2014_000000083005.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse with red wrapped around its legs carrying a rider holding a rope\"."}], "task": "refering", "answer_template": "The \"a horse with red wrapped around its legs carrying a rider holding a rope\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 108, 109, 131, 132, 153, 154, 155, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 243, 244, 245, 246, 266, 267, 268, 288, 289, 290, 291, 311, 312, 313], "bbox": [0.54125, 0.22023419203747074, 0.779640625, 0.943817330210773], "iscrowd": 0, "area": 17710.89985, "rle": {"size": [427, 640], "counts": "QRa43W=2M3N3M2N3M2N0O101O001O0O0100O1O100O001O01N1O1O2WD^Oh:c0YE@c:`0]ED^:=bEGY:;fEIU:8lEJP:7PFIP:8oEIo98QFIm8_ORGi02Gf8MPG=:G_89PG1`0GY8e0oFEh0FS8P2nGPNn7T2QHmMl7U2THkMk7V2UHkMh7W2YHiMe7Y2ZHgMd7[2\\HfMa7\\2_HeM_7]2aHbM^7_2bHbM^7]2cHcM^7[2cHdM_7[2`HfM`7Y2aHgM`7W2`HiMb7V2^HjMb7U2_HkMb7S2^HmMd7Q2]HoMc7Q2\\HPNe7n1\\HQNf7m1[HSNe7m1ZHSNg7m1YHSNg7n1XHRNh7o1VHQNl7P2RHPNn7Q2PHPNP8Q2oGoMQ8R2mGoMS8R2lGnMT8S2kGmMV8S2hGnMX8S2gGmMY8T2eGmM[8T2dGlM\\8U2cGkM^8V2_GkMa8V2^GjMb8W2\\GjMd8W2[GiMe8X2ZGhMg8X2WGhMj8Y2UGgMk8\\300001O000000001O0000O1N2O1O0O2O1N101O1N1L5E;G9F9L500O010O10O0100O011N100000000O2O00000002N3M2N2N2O0O2N1O1O2N1N2N3N1N2N3M2N2RNiGfNZ8mNdHa0EOk7]OhH in this image.", "objects": [{"patches": [85, 86, 108, 109, 131, 132, 153, 154, 155, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 243, 244, 245, 246, 266, 267, 268, 288, 289, 290, 291, 311, 312, 313], "bbox": [0.54125, 0.22023419203747074, 0.779640625, 0.943817330210773], "iscrowd": 0, "area": 17710.89985, "rle": {"size": [427, 640], "counts": "QRa43W=2M3N3M2N3M2N0O101O001O0O0100O1O100O001O01N1O1O2WD^Oh:c0YE@c:`0]ED^:=bEGY:;fEIU:8lEJP:7PFIP:8oEIo98QFIm8_ORGi02Gf8MPG=:G_89PG1`0GY8e0oFEh0FS8P2nGPNn7T2QHmMl7U2THkMk7V2UHkMh7W2YHiMe7Y2ZHgMd7[2\\HfMa7\\2_HeM_7]2aHbM^7_2bHbM^7]2cHcM^7[2cHdM_7[2`HfM`7Y2aHgM`7W2`HiMb7V2^HjMb7U2_HkMb7S2^HmMd7Q2]HoMc7Q2\\HPNe7n1\\HQNf7m1[HSNe7m1ZHSNg7m1YHSNg7n1XHRNh7o1VHQNl7P2RHPNn7Q2PHPNP8Q2oGoMQ8R2mGoMS8R2lGnMT8S2kGmMV8S2hGnMX8S2gGmMY8T2eGmM[8T2dGlM\\8U2cGkM^8V2_GkMa8V2^GjMb8W2\\GjMd8W2[GiMe8X2ZGhMg8X2WGhMj8Y2UGgMk8\\300001O000000001O0000O1N2O1O0O2O1N101O1N1L5E;G9F9L500O010O10O0100O011N100000000O2O00000002N3M2N2N2O0O2N1O1O2N1N2N3N1N2N3M2N2RNiGfNZ8mNdHa0EOk7]OhH in this image.", "objects": [{"patches": [101, 123, 124, 125, 126, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 192, 193, 194, 195, 214, 215, 216, 217, 218, 237, 238, 239, 240, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286], "bbox": [0.244828125, 0.32402810304449653, 0.526875, 0.8669555035128806], "iscrowd": 0, "area": 17304.02495, "rle": {"size": [427, 640], "counts": "[jQ24V=1O1O1O1O1O1O1O2M2N2N2O1N2N2N2O2N1O100O1O100O1O100O1O100O1O1O1N2O1O1O1N2OK5K4L5K5N2O001O1O1OfCCh;=WDKc;5\\DMb;3_DN`;1aD0];0eD0Z;OgD2X;NhD3V;OjD1U;OjD3T;NkD4T;MkD4S;MlD5R;MmD4n:0QE2k:1TE1g:4XEMc:7\\EKb:7\\EKb:6]ELa:`1N2O1N2N2N1ON3M2M3N2N202O0O10N2K4K6@`0M2N3RO`LoGb3Q8dLgG_3Z7XLTIb4m6`KoHb4Q7`KkHa4V7aKfH`4[7aKbHa4^7bK]H_4d7eKVH\\4l7a00O1O100O101O`HYK\\6f4aI_K^6a4^IeK`6Z4]IkKb6U4[IoKe6o3YITLg6l3VIWLj6T510001O00000O101O0cNSI\\Lm6c3WIZLi6e3ZIYLf6f3^IVLc6i3dIQL]6m3iInKW6Q4PJiKP6W4VJcKj5\\4[J`Ke5_4`J\\K`5d4d1O1O1O1O1O100O001O1O1O3M4M4K6J6J6J00O2O00000000001O01O001O1O100O1O1N2N2N2M3N2N3M3L4M3M4L3J9F in this image.", "objects": [{"patches": [120, 121, 122, 123, 143, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 211, 212, 213, 214, 233, 234, 235, 236, 237, 255, 256, 257, 258, 278, 279, 280], "bbox": [0.11359375000000001, 0.3297423887587822, 0.39829687500000005, 0.8599063231850117], "iscrowd": 0, "area": 15580.3761, "rle": {"size": [427, 640], "counts": "mhn0:o<2N2O1O1N2O1O1N2O1O0O100O10O010000O01000000O10O100000O10O1O1TDXOU;h0iD]OT;d0hDAV;`0gDDX; in this image.", "objects": [{"patches": [120, 121, 122, 123, 143, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 211, 212, 213, 214, 233, 234, 235, 236, 237, 255, 256, 257, 258, 278, 279, 280], "bbox": [0.11359375000000001, 0.3297423887587822, 0.39829687500000005, 0.8599063231850117], "iscrowd": 0, "area": 15580.3761, "rle": {"size": [427, 640], "counts": "mhn0:o<2N2O1O1N2O1O1N2O1O0O100O10O010000O01000000O10O100000O10O1O1TDXOU;h0iD]OT;d0hDAV;`0gDDX; in this image.", "objects": [{"patches": [103, 104, 105, 119, 120, 121, 122, 123, 124, 137, 138, 139, 140, 141, 142, 155, 156, 157], "bbox": [0.6247, 0.41818666666666665, 0.9034200000000001, 0.69152], "iscrowd": 0, "area": 7980.121550000002, "rle": {"size": [375, 500], "counts": "o^b31c;5YO2^E3`:d0M3N2M3@kNYFX1e9?N2N2N2N2N2N2N2N2N1O2N2N2N2M10001O0000000000000000001O00000O1000000000eNPGOP91QGNo81RGOn81SGNm82TGMk84VGKi86XGGi8:XGDh8=XGAi8`0XG]Oi8d0XGWOk8j0VGROl8o0UGmNn8S1UGfNn8[1f01O2N1O100O1O1O10000001O0000001O0001O01O0000001O0000001O00001O0000010O00001O00001O00001O001O00001O01O01O001O001O1O001O1O1O001O1O1O010O1O1O001O1O1O1N1O2M3M3M2L5J6J6Je\\a0"}, "label": "a dark colored couch"}]} {"id": 533568, "image": "COCO_train2014_000000533568.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark colored couch sitting in front of a red rug\"."}], "task": "refering", "answer_template": "The \"a dark colored couch sitting in front of a red rug\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 119, 120, 121, 122, 123, 124, 137, 138, 139, 140, 141, 142, 155, 156, 157], "bbox": [0.6247, 0.41818666666666665, 0.9034200000000001, 0.69152], "iscrowd": 0, "area": 7980.121550000002, "rle": {"size": [375, 500], "counts": "o^b31c;5YO2^E3`:d0M3N2M3@kNYFX1e9?N2N2N2N2N2N2N2N2N1O2N2N2N2M10001O0000000000000000001O00000O1000000000eNPGOP91QGNo81RGOn81SGNm82TGMk84VGKi86XGGi8:XGDh8=XGAi8`0XG]Oi8d0XGWOk8j0VGROl8o0UGmNn8S1UGfNn8[1f01O2N1O100O1O1O10000001O0000001O0001O01O0000001O0000001O00001O0000010O00001O00001O00001O001O00001O01O01O001O001O1O001O1O1O001O1O1O010O1O1O001O1O1O1N1O2M3M3M2L5J6J6Je\\a0"}, "label": "a dark colored couch sitting in front of a red rug"}]} {"id": 533568, "image": "COCO_train2014_000000533568.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown arm chair\"."}], "task": "refering", "answer_template": "The \"brown arm chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 226, 227, 228, 229, 230, 231], "bbox": [0.57812, 0.5455733333333334, 0.98438, 0.9882933333333334], "iscrowd": 0, "area": 21372.1642, "rle": {"size": [375, 500], "counts": "dTZ3:Y;c0^Oc0\\O in this image.", "objects": [{"patches": [142, 143, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 226, 227, 228, 229, 230, 231], "bbox": [0.57812, 0.5455733333333334, 0.98438, 0.9882933333333334], "iscrowd": 0, "area": 21372.1642, "rle": {"size": [375, 500], "counts": "dTZ3:Y;c0^Oc0\\O in this image.", "objects": [{"patches": [59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 230, 231, 232, 233, 234, 253, 254, 255, 276], "bbox": [0.001484375, 0.145260663507109, 0.9979687500000001, 0.8141943127962086], "iscrowd": 0, "area": 118040.56629999999, "rle": {"size": [422, 640], "counts": "\\d0a3S6b30O100O10O01O100O10O0100O100O010O100O1O010O100O100O010O100O10O01O100O10O0100O100O10O0100O1O10O10O100O10O10O10000O010O10000O10O10O100O10O10O100O10O10O10000O10O010000O1000O0100O10000O01000O100O10O10O10000O10O010000O10O10O100O1000O010000O100O01000O100O10O10O10000O10O010000O1000O0100O10000O01000O100O10O10O10000O10O010000O10000O100O10000O100O10000O100O100O10000O100O100O2O000O100O10000O100O100O10000O100O100O10000O100O1000000000000O100000000000000O10000000000000000O1000001O0000000O10000000000000000O100000000000000O100000000000000000000001O0000001O0000001O0000001O00000010O000001O0000001O0000001O0000001O0O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O1000O1000000000O1000000000000O10000000000000000000O10000000O10000000000000000000000000000000000000O1000O1000000000000000000000000000000000000000O10O1000000000000000000000000000000000000000000O0100000000000000000000000000000000000000000O10O1000000000000000000000000000000000000000O1001O0O2O001O001O001O001O001O1O001O001O001O001O001O0O2O00k:"}, "label": "a zebra doesn ' t have birds on its back"}]} {"id": 164935, "image": "COCO_train2014_000000164935.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a zebra behind two birds\"."}], "task": "refering", "answer_template": "The \"the back of a zebra behind two birds\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 230, 231, 232, 233, 234, 253, 254, 255, 276], "bbox": [0.001484375, 0.145260663507109, 0.9979687500000001, 0.8141943127962086], "iscrowd": 0, "area": 118040.56629999999, "rle": {"size": [422, 640], "counts": "\\d0a3S6b30O100O10O01O100O10O0100O100O010O100O1O010O100O100O010O100O10O01O100O10O0100O100O10O0100O1O10O10O100O10O10O10000O010O10000O10O10O100O10O10O100O10O10O10000O10O010000O1000O0100O10000O01000O100O10O10O10000O10O010000O10O10O100O1000O010000O100O01000O100O10O10O10000O10O010000O1000O0100O10000O01000O100O10O10O10000O10O010000O10000O100O10000O100O10000O100O100O10000O100O100O2O000O100O10000O100O100O10000O100O100O10000O100O1000000000000O100000000000000O10000000000000000O1000001O0000000O10000000000000000O100000000000000O100000000000000000000001O0000001O0000001O0000001O00000010O000001O0000001O0000001O0000001O0O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O100000000000000O1000O1000000000O1000000000000O10000000000000000000O10000000O10000000000000000000000000000000000000O1000O1000000000000000000000000000000000000000O10O1000000000000000000000000000000000000000000O0100000000000000000000000000000000000000000O10O1000000000000000000000000000000000000000O1001O0O2O001O001O001O001O001O1O001O001O001O001O001O0O2O00k:"}, "label": "the back of a zebra behind two birds"}]} {"id": 238667, "image": "COCO_train2014_000000238667.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man horse rideing and take helment\"."}], "task": "refering", "answer_template": "The \"a man horse rideing and take helment\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 100, 101, 102, 103, 122, 123, 124, 125, 126, 127, 147, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 239, 240, 241, 244, 262], "bbox": [0.33646875, 0.18905077262693157, 0.639609375, 0.713664459161148], "iscrowd": 0, "area": 15240.5333, "rle": {"size": [453, 640], "counts": "ZXo21S>3RB0b=:N2O1N2O1O1N2O1O1N2O1N2O1N2SOCQD>l;o0N2M3M3O100O01O010O010O1O2O1N2N2O0011O2Nf0ZO1jHWMl2k2PMWMP3j2lLZMS3i2hLYMX3k2aLXM_3l2[LVMe3n2ULTMk3o2oKTMQ4m2lKUMT4l2hKWMX4j2eKXM[4i2bKYM]4j2^KYMb4h2RJfLId0U6g2dIVM06\\6e2UIfM:Ga6Z4[IhKe6Z4VIiKj6Y4RIiKn6g500000_O`0_IhHQ6j7J6I7J7I6000001O00001O0\\OVGmKi8P4ZGPLf8l3^GTLc8g3bGXL^8e3eGZL\\8b3hG^LY8^3jGbLV8Z3nGfLR8Y3oGgLQ8j2UGPMk05R8i2TGQMi07T8f2UGRMf09V8d2TGSMg08V8d2TGSMg08W8c2SGTMf09W8c2SGTMg08V8c2UGUMd08W8c2VGRMf0:T8d2VGPMi0:Q8g2\\HXMe7h2\\HVMe7i2]HTMe7l2\\HRMf7m2j1O001N2O1N101N2O0OcD]M[;b2eDfMT;X2lDoMa1@j7`2eFTNZ1@P8[2fFZNS1^OX8V2fFaNk0\\O_8R2eFgNe0[Of8l1fFUO5ROU9i1eF1\\9LeF4\\9KcF7^9FcF;]9DbF;b9B_F=d9@\\F?g9@YFN=VN^9i1UFKc:4]EIh:4XEeN9?d:j0SEaNg09[:T1QFeNT:Y1]1N7J in this image.", "objects": [{"patches": [153, 154, 175, 176, 177, 198, 199, 200, 220, 221, 222, 223, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 294, 311, 312, 313, 314, 315, 316, 317, 334, 335, 336, 337, 338, 339, 340, 357, 358, 359, 360, 361, 362, 363, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.54059375, 0.39499999999999996, 0.8061562500000001, 0.9981041666666667], "iscrowd": 0, "area": 35361.0383, "rle": {"size": [480, 640], "counts": "iaR55c>:C=C=H8F9eCaNj:j1kDbNi:o2L5K5K5K5M3N2_FgKX8[4[GQLc8R4VGTLi8R5N2O1O1N2O1N2O1N2O1O1N2O1N2O1O1O1O100O1O1O100O1O1O100O1O1O1O1O1O1N2O1O1O1N2O1O1M3N2M3^Ob0K5M3M3M3M3O1O1O1O2N1N2O1O1O100O10000O1000000O1000000O1000000O10000000000001O000000001O0000001O00001O00001O001[IQJW4o5fKYJU4h5hKZJX4g5eK\\JZ4e5bK^J^4d5^K`J`4b5[KcJc4`5XKdJf4^5WKeJg4^5TKeJk4c5kJaJS5b5hJaJW5b5eJaJY5a5cJbJ\\5`5aJcJ]5a7O2N1O1O1O1O1O001O2N1O1O2N1O2N1O2N2N3M2N3M2N3M3M2N3M2N3M3M2M4K4L5K?Ag0ZO7I8G8I8C=C>Affi1"}, "label": "an overweight guy in a white sweater playing wii"}]} {"id": 492638, "image": "COCO_train2014_000000492638.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large man in white and gold sweater on the right\"."}], "task": "refering", "answer_template": "The \"large man in white and gold sweater on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 175, 176, 177, 198, 199, 200, 220, 221, 222, 223, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 294, 311, 312, 313, 314, 315, 316, 317, 334, 335, 336, 337, 338, 339, 340, 357, 358, 359, 360, 361, 362, 363, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.54059375, 0.39499999999999996, 0.8061562500000001, 0.9981041666666667], "iscrowd": 0, "area": 35361.0383, "rle": {"size": [480, 640], "counts": "iaR55c>:C=C=H8F9eCaNj:j1kDbNi:o2L5K5K5K5M3N2_FgKX8[4[GQLc8R4VGTLi8R5N2O1O1N2O1N2O1N2O1O1N2O1N2O1O1O1O100O1O1O100O1O1O100O1O1O1O1O1O1N2O1O1O1N2O1O1M3N2M3^Ob0K5M3M3M3M3O1O1O1O2N1N2O1O1O100O10000O1000000O1000000O1000000O10000000000001O000000001O0000001O00001O00001O001[IQJW4o5fKYJU4h5hKZJX4g5eK\\JZ4e5bK^J^4d5^K`J`4b5[KcJc4`5XKdJf4^5WKeJg4^5TKeJk4c5kJaJS5b5hJaJW5b5eJaJY5a5cJbJ\\5`5aJcJ]5a7O2N1O1O1O1O1O001O2N1O1O2N1O2N1O2N2N3M2N3M2N3M3M2N3M2N3M3M2M4K4L5K?Ag0ZO7I8G8I8C=C>Affi1"}, "label": "large man in white and gold sweater on the right"}]} {"id": 492638, "image": "COCO_train2014_000000492638.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the red shirt\"."}], "task": "refering", "answer_template": "The \"the man in the red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 28, 29, 30, 31, 51, 52, 53, 54, 74, 75, 76, 77, 95, 96, 97, 98, 99, 116, 117, 118, 119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.0017500000000000003, 0.04004166666666666, 0.509375, 0.9768125], "iscrowd": 0, "area": 95248.59525000001, "rle": {"size": [480, 640], "counts": "dh0j0ZB9G4L4L5L3L4L5K4L4L5K4L5K5K5L5J5K6J5K7I8H7I8H7I8H8I6I8H`gb4"}, "label": "the man in the red shirt"}]} {"id": 492638, "image": "COCO_train2014_000000492638.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"old man in red sweater playing wii\"."}], "task": "refering", "answer_template": "The \"old man in red sweater playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 28, 29, 30, 31, 51, 52, 53, 54, 74, 75, 76, 77, 95, 96, 97, 98, 99, 116, 117, 118, 119, 120, 121, 122, 123, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.0017500000000000003, 0.04004166666666666, 0.509375, 0.9768125], "iscrowd": 0, "area": 95248.59525000001, "rle": {"size": [480, 640], "counts": "dh0j0ZB9G4L4L5L3L4L5K4L4L5K4L5K5K5L5J5K6J5K7I8H7I8H7I8H8I6I8H`gb4"}, "label": "old man in red sweater playing wii"}]} {"id": 574563, "image": "COCO_train2014_000000574563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young dark haired asian girl about to take a bite out of a donut\"."}], "task": "refering", "answer_template": "The \"a young dark haired asian girl about to take a bite out of a donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 45, 46, 47, 48, 49, 63, 64, 65, 66, 67, 68, 80, 81, 82, 83, 84, 85, 86, 98, 99, 100, 101, 102, 103, 104, 114, 115, 116, 117, 118, 119, 120, 121, 122, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.27027999999999996, 0.11037333333333334, 0.9645199999999999, 0.98648], "iscrowd": 0, "area": 70403.061, "rle": {"size": [375, 500], "counts": "Vfa12b;5K4K6\\OBhEb0U:BfEc0V:AgEb0U:BhE`0U:d0M2O2N2M2O2N2bHSN`40`In1n1SNa42_Im1l1SNd43^Ik1k1TNg43\\Ik1i1TNj44ZIk1h1SNn44XIj1h1SNo4`2lJcMT5^2mI[ML in this image.", "objects": [{"patches": [28, 29, 30, 45, 46, 47, 48, 49, 63, 64, 65, 66, 67, 68, 80, 81, 82, 83, 84, 85, 86, 98, 99, 100, 101, 102, 103, 104, 114, 115, 116, 117, 118, 119, 120, 121, 122, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.27027999999999996, 0.11037333333333334, 0.9645199999999999, 0.98648], "iscrowd": 0, "area": 70403.061, "rle": {"size": [375, 500], "counts": "Vfa12b;5K4K6\\OBhEb0U:BfEc0V:AgEb0U:BhE`0U:d0M2O2N2M2O2N2bHSN`40`In1n1SNa42_Im1l1SNd43^Ik1k1TNg43\\Ik1i1TNj44ZIk1h1SNn44XIj1h1SNo4`2lJcMT5^2mI[ML in this image.", "objects": [{"patches": [71, 89, 106, 107, 124, 125, 142, 143, 159, 160, 161, 178, 179, 196, 197, 214, 215, 233], "bbox": [0.8848400000000001, 0.26741333333333334, 1.0, 0.9887733333333333], "iscrowd": 0, "area": 10609.620749999995, "rle": {"size": [375, 500], "counts": "nRR55[;V1kN7H8I7H8I7H8I7H8I7M3M3N2kH^LQ6d3kIcLQ6_3iIhLS6[3hIlLT6V3fIQMW6Q3dITMY6X4M3L3N3M3L3N3L4M2M4M3L4L4M2M4L4L4M3L4L4M3L4L4L3N3L4L2O1O1O1O1O100O1O1O1O1OmMT2iKR1"}, "label": "the person with their back to the camera"}]} {"id": 574563, "image": "COCO_train2014_000000574563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back of a person wearing a plaid shirt\"."}], "task": "refering", "answer_template": "The \"back of a person wearing a plaid shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 89, 106, 107, 124, 125, 142, 143, 159, 160, 161, 178, 179, 196, 197, 214, 215, 233], "bbox": [0.8848400000000001, 0.26741333333333334, 1.0, 0.9887733333333333], "iscrowd": 0, "area": 10609.620749999995, "rle": {"size": [375, 500], "counts": "nRR55[;V1kN7H8I7H8I7H8I7H8I7M3M3N2kH^LQ6d3kIcLQ6_3iIhLS6[3hIlLT6V3fIQMW6Q3dITMY6X4M3L3N3M3L3N3L4M2M4M3L4L4M2M4L4L4M3L4L4M3L4L4L3N3L4L2O1O1O1O1O100O1O1O1O1OmMT2iKR1"}, "label": "back of a person wearing a plaid shirt"}]} {"id": 320611, "image": "COCO_train2014_000000320611.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white sofa with patterned pillows upon it\"."}], "task": "refering", "answer_template": "The \"a white sofa with patterned pillows upon it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 294, 295, 312, 313, 314, 315, 316, 317, 319, 335, 336, 337, 338, 339, 340, 359, 360, 361, 362, 363, 383, 384, 385, 386, 387, 388], "bbox": [0.55121875, 0.6175259875259874, 0.903109375, 0.9863409563409562], "iscrowd": 0, "area": 20576.9835, "rle": {"size": [481, 640], "counts": "QUV5=Q>e0\\Oc0M3O2M2O1N3N1O1N3N1N2O1O2M2O1N3N1O1N3N1N2O2N1N2O1N3N1O1N3N1J6O2N1O1O2N100O2N1O1O1O2N1O1O2N1O1O2N100O2N1O1O2N1O1O1O2N1O1O2O0O1O2N1O1O2N1O1O1O2N100O2O00001OO10O01O001O00001O010O00001O001O00010O00001O001O00010O001O00001O0010O0001O0XNoE`NR:^1WFZNi9d1aFTN_9j1lFmMU9P2UGiMj8V2_GbMb8[2hG]MX8c2S2O2O0O2O0OcEbMY8^2bGgM_8W2\\GoMd8Q2VGUNj8j1RG[No8c1lFcNT9]1iFfNX9X1hFhNY9W1iFhNW9U1lFkNU9o0QGPOo8k0WGTOj8f0\\GYOd8e0_GZOa8f0`GYOa8f0_GZOa8f0_GZOb8e0^G[Ob8e0_GZOa8f0_GZOb8e0^GZOc8f0]GZOd8e0\\G[Od8e0\\G[Od8e0]GZOd8e0\\G[Od8h0YGXOh8m0RGSOn8U1jFkNV9Y1gFfNZ9Z1eFfN[9g0VEUO_14[9f0WEVO^14\\9d0XEXO[13^9d0YEXOZ12_9d0XE[OX1Bo9R1kD[Oh`CA`<>bCA_<>aCC^<=gC^OY in this image.", "objects": [{"patches": [243, 244, 245, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 294, 295, 312, 313, 314, 315, 316, 317, 319, 335, 336, 337, 338, 339, 340, 359, 360, 361, 362, 363, 383, 384, 385, 386, 387, 388], "bbox": [0.55121875, 0.6175259875259874, 0.903109375, 0.9863409563409562], "iscrowd": 0, "area": 20576.9835, "rle": {"size": [481, 640], "counts": "QUV5=Q>e0\\Oc0M3O2M2O1N3N1O1N3N1N2O1O2M2O1N3N1O1N3N1N2O2N1N2O1N3N1O1N3N1J6O2N1O1O2N100O2N1O1O1O2N1O1O2N1O1O2N100O2N1O1O2N1O1O1O2N1O1O2O0O1O2N1O1O2N1O1O1O2N100O2O00001OO10O01O001O00001O010O00001O001O00010O00001O001O00010O001O00001O0010O0001O0XNoE`NR:^1WFZNi9d1aFTN_9j1lFmMU9P2UGiMj8V2_GbMb8[2hG]MX8c2S2O2O0O2O0OcEbMY8^2bGgM_8W2\\GoMd8Q2VGUNj8j1RG[No8c1lFcNT9]1iFfNX9X1hFhNY9W1iFhNW9U1lFkNU9o0QGPOo8k0WGTOj8f0\\GYOd8e0_GZOa8f0`GYOa8f0_GZOa8f0_GZOb8e0^G[Ob8e0_GZOa8f0_GZOb8e0^GZOc8f0]GZOd8e0\\G[Od8e0\\G[Od8e0]GZOd8e0\\G[Od8h0YGXOh8m0RGSOn8U1jFkNV9Y1gFfNZ9Z1eFfN[9g0VEUO_14[9f0WEVO^14\\9d0XEXO[13^9d0YEXOZ12_9d0XE[OX1Bo9R1kD[Oh`CA`<>bCA_<>aCC^<=gC^OY in this image.", "objects": [{"patches": [277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 346, 347, 348, 349, 350, 369, 370, 373], "bbox": [0.0175625, 0.7012889812889813, 0.269390625, 0.9844074844074844], "iscrowd": 0, "area": 15532.47905, "rle": {"size": [481, 640], "counts": "[a52n>4L4K5L4K5L4L4K5L4L4L4M3L4L4M3L4L4M3L4M3L4L4M3L4M3L5K4M3L2O1N1O100OJ6G:0O100O1O100O100O00100O100O1O100O1O100O100O1O100O100O1O010O11O000000000000001O0000000000000O101O00003M3M2N2N000000000000000000000O10000000O10000000000000000000000000O10000000000000000003M3M4L3M3M3M2N0000O1000000O1000000O1ZOhDcMX;Z2SE^Mm:`2]EXMc:f2m0M3N2M9H9G9F:G9F:G9F;FQ^k6"}, "label": "a single white cousioned chair farthest away from the french doors"}]} {"id": 66669, "image": "COCO_train2014_000000066669.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange cat stuffed animal\"."}], "task": "refering", "answer_template": "The \"an orange cat stuffed animal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 23, 24, 25, 26, 41, 42, 43, 44, 56, 57, 58, 59, 60, 61, 62, 73, 74, 75, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 165, 166, 167], "bbox": [0.0, 0.0, 0.4972, 0.7685333333333333], "iscrowd": 0, "area": 34314.578250000006, "rle": {"size": [375, 500], "counts": "^63d;5J6K5K5J6K5J7J1O0O2O1N2O001N2O001N2O1N101O1N101N2O1O0O2O1N101O0O10O10O100O1000O0100bNlMXIU2f6PNWIP2h6TNUIl1i6YNSIg1m6]NmHf1Q7_NcHj1[7ZNZHo1e7R1N2N101N2N2O1N2N101N2N2O1N2O0O2N20000O0100000O1000O1000O100000O01000000O10O1000O101O1O2M2O1O1O1N2O2N1O1N3N1O2N1N2O2N1O1N3N1O2N1N2O2N1O1N2O00000O1000000O1000000O1000000O1000000O1000000JfMhJaNS5d1[KnM_4W2PL[Mk3j2]LnL]3X3dLgLW3]3kL`LQ3f3QMWLn2j3TMSLl2o3UMnKk2T4VMiKj2X4XMeKg2^4[M^Ke2d4\\MYKd2h4^MVKa2l4`MQK`2P5cMlJ]2V5dMgJ]2Z5dMcJ]2]5n11O1O001O001O1O001O1O1O1O001O1O1O1O10000O01000O10000O1000O010000O10000O10O1\\O_JRK`5f4lJWKT5`4XK]Kg4[4eKbK[4]4iK`KW4`4a1N2N1O2N2N101N2N1O2N2O0O2N2N1O2N2O0O2O1O001O1N101O1O1O0O2O1O001O1O0O2POjG]NW8Y1THeNl7S1^HkNc7m0fHQO\\7g0lHVOX7a0PI]OS7;TICo64YIJj6M^I1V9NXUl2"}, "label": "an orange cat stuffed animal"}]} {"id": 66669, "image": "COCO_train2014_000000066669.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the teddy bear that is as large as the baby\"."}], "task": "refering", "answer_template": "The \"the teddy bear that is as large as the baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 23, 24, 25, 26, 41, 42, 43, 44, 56, 57, 58, 59, 60, 61, 62, 73, 74, 75, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 165, 166, 167], "bbox": [0.0, 0.0, 0.4972, 0.7685333333333333], "iscrowd": 0, "area": 34314.578250000006, "rle": {"size": [375, 500], "counts": "^63d;5J6K5K5J6K5J7J1O0O2O1N2O001N2O001N2O1N101O1N101N2O1O0O2O1N101O0O10O10O100O1000O0100bNlMXIU2f6PNWIP2h6TNUIl1i6YNSIg1m6]NmHf1Q7_NcHj1[7ZNZHo1e7R1N2N101N2N2O1N2N101N2N2O1N2O0O2N20000O0100000O1000O1000O100000O01000000O10O1000O101O1O2M2O1O1O1N2O2N1O1N3N1O2N1N2O2N1O1N3N1O2N1N2O2N1O1N2O00000O1000000O1000000O1000000O1000000O1000000JfMhJaNS5d1[KnM_4W2PL[Mk3j2]LnL]3X3dLgLW3]3kL`LQ3f3QMWLn2j3TMSLl2o3UMnKk2T4VMiKj2X4XMeKg2^4[M^Ke2d4\\MYKd2h4^MVKa2l4`MQK`2P5cMlJ]2V5dMgJ]2Z5dMcJ]2]5n11O1O001O001O1O001O1O1O1O001O1O1O1O10000O01000O10000O1000O010000O10000O10O1\\O_JRK`5f4lJWKT5`4XK]Kg4[4eKbK[4]4iK`KW4`4a1N2N1O2N2N101N2N1O2N2O0O2N2N1O2N2O0O2O1O001O1N101O1O1O0O2O1O001O1O0O2POjG]NW8Y1THeNl7S1^HkNc7m0fHQO\\7g0lHVOX7a0PI]OS7;TICo64YIJj6M^I1V9NXUl2"}, "label": "the teddy bear that is as large as the baby"}]} {"id": 66669, "image": "COCO_train2014_000000066669.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the group of stuffed animals the child is facing\"."}], "task": "refering", "answer_template": "The \"the group of stuffed animals the child is facing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 120, 121, 122, 136, 137, 138, 139, 140, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 206, 207, 208, 209, 210, 211, 224, 225, 226, 227, 228], "bbox": [0.47022, 0.40861333333333333, 0.81742, 0.9906400000000001], "iscrowd": 0, "area": 19540.534449999996, "rle": {"size": [375, 500], "counts": "j\\f23];;E:G:L3L5K2O0O1O1O100O1O100O1O100O1O1M3L3N3L4M300000000000000000000000O1000000000000000000000000O10001O001O001OK5fMlMlJY2R5kMgJZ2W5iMcJ\\2[5hM^J\\2a5hMZJ[2d5hMYJZ2g5hMUJZ2k5hMPJ[2P6fMmI\\2S6fMhI\\2Y6fMcI\\2^6bMaI`2_6`M`I`2a6`M]Ib2c6^M\\Ic2d6]MZId2g6\\MWIf2i6ZMVIg2j6XMUIi2m6VMRIk2n6UMPIm2P7SMoHm2R7SMmHn2S7RMmHn2S7RMlHn2U7QMkHP3T7QMlHo2T7QMkHo2V7i0O100O00100O001O1O0N3L4L3N3M3N2N2M4M3L4G9J501O1N0100O1O1O1O1O100O001O1O100O1O00000010O000001O01O00100O1O0011O1N3N1SOQJPLQ6j3dJeK]5U4W1J5J7J5J6L4N2O1N2N2O1N2N2lLnG`2j8K5J6J6K4K4L4L4J6J5K6J6J5K6K5JjRQ1"}, "label": "the group of stuffed animals the child is facing"}]} {"id": 66669, "image": "COCO_train2014_000000066669.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"infant feed her teddy bear\"."}], "task": "refering", "answer_template": "The \"infant feed her teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 120, 121, 122, 136, 137, 138, 139, 140, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 206, 207, 208, 209, 210, 211, 224, 225, 226, 227, 228], "bbox": [0.47022, 0.40861333333333333, 0.81742, 0.9906400000000001], "iscrowd": 0, "area": 19540.534449999996, "rle": {"size": [375, 500], "counts": "j\\f23];;E:G:L3L5K2O0O1O1O100O1O100O1O100O1O1M3L3N3L4M300000000000000000000000O1000000000000000000000000O10001O001O001OK5fMlMlJY2R5kMgJZ2W5iMcJ\\2[5hM^J\\2a5hMZJ[2d5hMYJZ2g5hMUJZ2k5hMPJ[2P6fMmI\\2S6fMhI\\2Y6fMcI\\2^6bMaI`2_6`M`I`2a6`M]Ib2c6^M\\Ic2d6]MZId2g6\\MWIf2i6ZMVIg2j6XMUIi2m6VMRIk2n6UMPIm2P7SMoHm2R7SMmHn2S7RMmHn2S7RMlHn2U7QMkHP3T7QMlHo2T7QMkHo2V7i0O100O00100O001O1O0N3L4L3N3M3N2N2M4M3L4G9J501O1N0100O1O1O1O1O100O001O1O100O1O00000010O000001O01O00100O1O0011O1N3N1SOQJPLQ6j3dJeK]5U4W1J5J7J5J6L4N2O1N2N2O1N2N2lLnG`2j8K5J6J6K4K4L4L4J6J5K6J6J5K6K5JjRQ1"}, "label": "infant feed her teddy bear"}]} {"id": 99451, "image": "COCO_train2014_000000099451.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wii remotes are sitting on the laps of two children\"."}], "task": "refering", "answer_template": "The \"wii remotes are sitting on the laps of two children\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 188, 198, 199, 200, 206, 207, 224], "bbox": [0.03716, 0.79504, 0.51014, 0.9459466666666667], "iscrowd": 0, "area": 1898.9688999999998, "rle": {"size": [375, 500], "counts": "PY7=Y;1O1O1O1O1O100O1O1O1O1O1O001O0000001O0001O010O1O1O100O1O1O1O100O1O1O100O1O1O100O1H_]l12hbSN6I6J7J6I6J100O2KiNfEX1Y:hNhEX1W:iNhEX1X:gNhE[1V:fNjEZ1U:gNjEZ1U:fNkE[1U:5N1O100OO101N8I8G8I7I8G8IX\\i2"}, "label": "wii remotes are sitting on the laps of two children"}]} {"id": 99451, "image": "COCO_train2014_000000099451.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy with black and gray shirt , longer brown hair\"."}], "task": "refering", "answer_template": "The \"boy with black and gray shirt , longer brown hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 68, 83, 84, 85, 86, 101, 102, 103, 104, 105, 119, 120, 121, 122, 135, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 224, 225, 229, 230, 231, 232, 233], "bbox": [0.43456, 0.2709066666666667, 0.95974, 0.9867999999999999], "iscrowd": 0, "area": 38086.704, "rle": {"size": [375, 500], "counts": "jh_24`;=DXGAf8a0ZG_Oe8b0\\G]Ob8e0_GZO`8f0aGZO]8h0dGWO[8j0fGTOZ8m0gGROW8P1iGPOV8P1lGoNR8S1oGlNP8U1PHkNn7W1SHhNl7X1VHfNj7[1WHdNg7^1ZH`Nf7a1[H^Nc7c1_H[Na7f1`HXN_7j1bHUN]7l1dHRN[7P2fHoMY7R2hHlMX7U2iHjMU7X2lHfMT7[2mHdMQ7^2S11N2O1O1N2O1N2O1N2O1N2O1O1N2O1O100O10000O100O10000O10000O100O10000O2O0O10000O10000O100O10000O10000O100O10000O100O1[NfLmJZ3o4UMfJk2U5fM^J[2]5UNXJk1c5W2M3O1N2N2O1N2N2O1N2N2O1N2O1O101O0O10001O0O10000O2O00000O2O00000O2O000O10001O0O10001N1000001N100000001O0001O01O01O0000\\LoHb2R7]MnHd2Q7\\MoHd2Q7\\MoHd2Q7\\MoHe2Q7ZMoHf2Q7ZMoHf2Q7ZMoHf2Q7ZMoHg2Q7WMPIi2P7WMPIi2P7WMPIi2P7WMPIj2o6VMQIj2P7UMPIk2P7UMPIl2o6TMQIl2P3_Ld0 in this image.", "objects": [{"patches": [58, 59, 60, 75, 76, 77, 78, 91, 92, 93, 94, 95, 96, 109, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 180, 181, 182, 183, 184, 185, 186, 198, 199, 200, 201, 202, 203, 204, 205, 216, 217, 218, 219, 220, 221, 222, 223], "bbox": [0.00066, 0.22922666666666666, 0.4479, 1.0], "iscrowd": 0, "area": 44052.106750000006, "rle": {"size": [375, 500], "counts": "e:8^;j0VO1O1O100bNH\\G9b8NXG3f84TGMj8a1M3N2N2N2N2M3L4M3L4VObLWIb3g6dLQI`3l6g0M3M3L4M3M3L4M3L4N2N2M3N2N2N2N2N2N3L3O1O1N2O1O1O1O1N20000O1000000O1000000O1000000O1000000O1000000O1000000000000O100000000000000O10000000000000000O1000000000000000000000000PO_IjLa6o2gInL[6n2jImLZ6R3hIjL[6U3gIfL]6Y3fIbL]6]3o0O1O1O1O1O100O1O1O100O1O1O1O1O1O1O1O1O1O1O100O100O100O100O1O100O100O100000000O100000000O1000000001O00000gI\\M_3d2`L]M`3d2]L_Mb3a2]L`Mc3`2[LcMd3]2[LdMe3\\2YLgMf3Z2WLhMi3X2VLjMi3V2ULlMk3T2TLnMk3S2RLoMn3Q2QLQNn3P2oKRNQ4n1nKTNQ4m1lKUNT4k1jKXNU4i1iKXNW4i1fKZNY4f1eK\\N[4e1bK^NV2eN`Nn2WO_Ni1TOoN^2VO`Ni1UOoN]2TO`Nl1UOoN]2QOaNn1TOPO]2nNbNP2SOQO]2lNaNS2TOoN]2jNbNU2SOPO\\2gNdNX2ROPO\\2aNhN]2nNQO\\2\\NjNb2mNPOj4n0XKQOh4n0ZKQOg4n0ZKQOi4k0XKUOn4d0SK\\OR5>oJBV58lJGb80000000000000001O0000000000003M;E:F;D:D2NPgT3"}, "label": "a boy with blue pants and dark blue t shirt smiling at camera"}]} {"id": 271488, "image": "COCO_train2014_000000271488.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a closed , double wall oven\"."}], "task": "refering", "answer_template": "The \"a closed , double wall oven\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298], "bbox": [0.8047187499999999, 0.3095208333333333, 0.99765625, 0.7482083333333333], "iscrowd": 0, "area": 23875.411800000013, "rle": {"size": [480, 640], "counts": "lea7k1P9YN\\G]5c8`0000000000001O0000000000001O0000000000001O000000000000001O00000O1000001O0000000000001O0000000000001O0000000000001O0000000000001O000000000000001O0000000000001N1000000000001O0000000000001O0000000000001O0000000000001O0000000000N2_Oa0XOh0YOg0XOh0XOh0YOg0XOh0YOg0XOS;"}, "label": "a closed , double wall oven"}]} {"id": 271488, "image": "COCO_train2014_000000271488.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stainless steel double oven with tinted glass\"."}], "task": "refering", "answer_template": "The \"a stainless steel double oven with tinted glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298], "bbox": [0.8047187499999999, 0.3095208333333333, 0.99765625, 0.7482083333333333], "iscrowd": 0, "area": 23875.411800000013, "rle": {"size": [480, 640], "counts": "lea7k1P9YN\\G]5c8`0000000000001O0000000000001O0000000000001O000000000000001O00000O1000001O0000000000001O0000000000001O0000000000001O0000000000001O000000000000001O0000000000001N1000000000001O0000000000001O0000000000001O0000000000001O0000000000N2_Oa0XOh0YOg0XOh0XOh0YOg0XOh0YOg0XOS;"}, "label": "a stainless steel double oven with tinted glass"}]} {"id": 246927, "image": "COCO_train2014_000000246927.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue table that has the man with striped shirt in the foreground\"."}], "task": "refering", "answer_template": "The \"the blue table that has the man with striped shirt in the foreground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.049984375, 0.540841121495327, 0.999171875, 1.0], "iscrowd": 0, "area": 81297.97535000002, "rle": {"size": [428, 640], "counts": "[i=1Z=1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1K5K5K5O1N2O1O1N2O1O1N2O1O1N2O1N2O100O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O100000000000000O1000000000000O100000000000000O100000000000000000000000000000000LjKRGV4l8mKTGR4k8PLTGP4k89O1O1O1O1O1O1N20000B_KQHa4n7?O1O1O1O1O1O10000O10000O1000000O10000O10000O1000000O10000O100O1O1O1O1O1O1O10000000000004L3M000000000000000000001O000000000000000000000000000000000000000000000000000000000000O1000000000000O1009G:F3MO100O1O100O100O100000000001O0000000000000000001O000000000000O1O1O1O1O1O1H8O1001O1O001O001OO100O10000O10000O100O100001O001O001O1O001O1O001O001O1O001O1OO100O10000O10000O10000O1N2M3N2M3N2M3N2M3N2N2001O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O5K6J5K5K1O00001O00001O00001O001O001O00001O0000O1000000O100000000E;O1O1O1O1O10000O10000000000000000000000000000000000000000000000000000001O000000000000001O0000001O000000001O0000001O1ON2N2N2001O001nJlGj4T8UKnGj4S8SKPHl4X801O1O1O1]KaGW4_8fKgGW4Z8fKkGW4e801O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000n3"}, "label": "the blue table that has the man with striped shirt in the foreground"}]} {"id": 246927, "image": "COCO_train2014_000000246927.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table of food\"."}], "task": "refering", "answer_template": "The \"the table of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.049984375, 0.540841121495327, 0.999171875, 1.0], "iscrowd": 0, "area": 81297.97535000002, "rle": {"size": [428, 640], "counts": "[i=1Z=1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1K5K5K5O1N2O1O1N2O1O1N2O1O1N2O1N2O100O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O100000000000000O1000000000000O100000000000000O100000000000000000000000000000000LjKRGV4l8mKTGR4k8PLTGP4k89O1O1O1O1O1O1N20000B_KQHa4n7?O1O1O1O1O1O10000O10000O1000000O10000O10000O1000000O10000O100O1O1O1O1O1O1O10000000000004L3M000000000000000000001O000000000000000000000000000000000000000000000000000000000000O1000000000000O1009G:F3MO100O1O100O100O100000000001O0000000000000000001O000000000000O1O1O1O1O1O1H8O1001O1O001O001OO100O10000O10000O100O100001O001O001O1O001O1O001O001O1O001O1OO100O10000O10000O10000O1N2M3N2M3N2M3N2M3N2N2001O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O1O1O5K6J5K5K1O00001O00001O00001O001O001O00001O0000O1000000O100000000E;O1O1O1O1O10000O10000000000000000000000000000000000000000000000000000001O000000000000001O0000001O000000001O0000001O1ON2N2N2001O001nJlGj4T8UKnGj4S8SKPHl4X801O1O1O1]KaGW4_8fKgGW4Z8fKkGW4e801O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000n3"}, "label": "the table of food"}]} {"id": 181393, "image": "COCO_train2014_000000181393.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a zebra in front of another one\"."}], "task": "refering", "answer_template": "The \"this is a zebra in front of another one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 155, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 219, 220, 221, 222, 243, 245, 266, 267, 268], "bbox": [0.5553125, 0.3876875, 0.763546875, 0.6924375], "iscrowd": 0, "area": 9179.0872, "rle": {"size": [480, 640], "counts": "YdV5m0g= in this image.", "objects": [{"patches": [153, 154, 155, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 219, 220, 221, 222, 243, 245, 266, 267, 268], "bbox": [0.5553125, 0.3876875, 0.763546875, 0.6924375], "iscrowd": 0, "area": 9179.0872, "rle": {"size": [480, 640], "counts": "YdV5m0g= in this image.", "objects": [{"patches": [54, 55, 56, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194], "bbox": [0.042234375000000005, 0.15202083333333333, 0.486484375, 0.5101249999999999], "iscrowd": 0, "area": 33564.38105, "rle": {"size": [480, 640], "counts": "dk<2m>3M3M4K4M3M4L3M3M3M2N1O2N1O1O1O1O1O1N2O1O2N1O1O1O1O1O1O1O2N1O1O1O1O100O1O2O0O1O100O1O100O2N100O1O100O1O1O2O0O1O100O1O100O2N100O1O100O1O101N1O1N2B>N200000001N10000000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O1000000O100O1O100O1O100O1O100O1O100O1O100O1O100O100O10000O10000O10000O10000O10000O10000O1000000O10000O100O1O2N100O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1000000O100000000O100000000O1000000O100000O10O10O10000O10000O10000O10000O10000O10000O10000O10O10O10000O10000O10000O10000O10000O10000O10000O\\1eNc0]O7H7HC>BbPj4"}, "label": "a few slices of pizza among a complete pizza"}]} {"id": 468117, "image": "COCO_train2014_000000468117.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two slices of pizza sit next to a slice that has the tip cut off\"."}], "task": "refering", "answer_template": "The \"two slices of pizza sit next to a slice that has the tip cut off\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194], "bbox": [0.042234375000000005, 0.15202083333333333, 0.486484375, 0.5101249999999999], "iscrowd": 0, "area": 33564.38105, "rle": {"size": [480, 640], "counts": "dk<2m>3M3M4K4M3M4L3M3M3M2N1O2N1O1O1O1O1O1N2O1O2N1O1O1O1O1O1O1O2N1O1O1O1O100O1O2O0O1O100O1O100O2N100O1O100O1O1O2O0O1O100O1O100O2N100O1O100O1O101N1O1N2B>N200000001N10000000000O100000000O100000000O10000000000O100000000O100000000O100000000O10000000000O100000000O1000000O100O1O100O1O100O1O100O1O100O1O100O1O100O100O10000O10000O10000O10000O10000O10000O1000000O10000O100O1O2N100O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1000000O100000000O100000000O1000000O100000O10O10O10000O10000O10000O10000O10000O10000O10000O10O10O10000O10000O10000O10000O10000O10000O10000O\\1eNc0]O7H7HC>BbPj4"}, "label": "two slices of pizza sit next to a slice that has the tip cut off"}]} {"id": 468117, "image": "COCO_train2014_000000468117.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hands with a fork in background\"."}], "task": "refering", "answer_template": "The \"hands with a fork in background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 74], "bbox": [0.232421875, 0.0, 0.83475, 0.18635416666666668], "iscrowd": 0, "area": 26593.729599999995, "rle": {"size": [480, 640], "counts": "]mU23M3L4M3eN_OTDd0k;_OnCf0R<[OfCk0g;jN[D in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 74], "bbox": [0.232421875, 0.0, 0.83475, 0.18635416666666668], "iscrowd": 0, "area": 26593.729599999995, "rle": {"size": [480, 640], "counts": "]mU23M3L4M3eN_OTDd0k;_OnCf0R<[OfCk0g;jN[D in this image.", "objects": [{"patches": [227, 228, 229, 243, 244, 245, 246, 259, 260, 261, 262, 275, 276, 277, 278, 291, 292, 293, 294, 307, 308, 309, 310, 323, 324, 326], "bbox": [0.1847737556561086, 0.6347968749999999, 0.44000000000000006, 0.88025], "iscrowd": 0, "area": 11704.655850000001, "rle": {"size": [640, 442], "counts": "mdc12lc03N2N2M300[^OK^?4_@4]?L_@;]?F^@b0_?^O]@i0`?WO\\@o0b?ROY@T1e?lNV@[1h?eNT@a1j?_NR@h1j?YNQ@o1l?QNP@V2l?mMn_Oa2f?eMQ@m2a?m0K100O1O010O010O10O100gNfAlLZ>T3Y100000000000001O00m_OlLZ?V3`@RM\\?Q3^@YM[?i2`@dMW?b3000000001O0O01O010OPLi@o2`?oLf@d2c?[M`@_2d?`M^@[2g?dMY@X2k?fMX@X2i?fMZ@Z2m?[MW@e2a`01000000O1000000O10000000001O2N1O2N2N1O2N?A;E01O0O2O0O2O001N101N10001N2O001O1O001O1O001O1O1O001]Oc0POo0LQbj4"}, "label": "baby sheep with other sheep"}]} {"id": 156823, "image": "COCO_train2014_000000156823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the goat on the left\"."}], "task": "refering", "answer_template": "The \"the goat on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [227, 228, 229, 243, 244, 245, 246, 259, 260, 261, 262, 275, 276, 277, 278, 291, 292, 293, 294, 307, 308, 309, 310, 323, 324, 326], "bbox": [0.1847737556561086, 0.6347968749999999, 0.44000000000000006, 0.88025], "iscrowd": 0, "area": 11704.655850000001, "rle": {"size": [640, 442], "counts": "mdc12lc03N2N2M300[^OK^?4_@4]?L_@;]?F^@b0_?^O]@i0`?WO\\@o0b?ROY@T1e?lNV@[1h?eNT@a1j?_NR@h1j?YNQ@o1l?QNP@V2l?mMn_Oa2f?eMQ@m2a?m0K100O1O010O010O10O100gNfAlLZ>T3Y100000000000001O00m_OlLZ?V3`@RM\\?Q3^@YM[?i2`@dMW?b3000000001O0O01O010OPLi@o2`?oLf@d2c?[M`@_2d?`M^@[2g?dMY@X2k?fMX@X2i?fMZ@Z2m?[MW@e2a`01000000O1000000O10000000001O2N1O2N2N1O2N?A;E01O0O2O0O2O001N101N10001N2O001O1O001O1O001O1O1O001]Oc0POo0LQbj4"}, "label": "the goat on the left"}]} {"id": 156823, "image": "COCO_train2014_000000156823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sheep in middlr of picture\"."}], "task": "refering", "answer_template": "The \"sheep in middlr of picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 231, 232, 233, 247, 248, 249, 250, 251, 263, 264, 265, 266, 267, 280, 281, 282, 283, 296, 297, 298, 299, 312, 313], "bbox": [0.4365384615384615, 0.5917656250000001, 0.7306787330316742, 0.86153125], "iscrowd": 0, "area": 12964.607750000001, "rle": {"size": [640, 442], "counts": "YQi32jc06K4K5L4K5L3N2N3M2N2N2N101N2O1O1N2O100O10001N6K5K5J6K6I6c^OhMh`0k2M3L4M3M3M3M3N2M3M3M2N100O1O2N100O2O6I9H9G2M01O1O001O1O1O1O1O1O1F:C=B>K42O0S@kLn>U3g@YMV?h31O1O1O1O1O1O1O1O1L4K5K5J6L4L4M3L4M3L4M2000001O01O00000000O10001O0O100001O3M3M3N3L3N2MN3M2N3N1N3L3L5K4M6I6K5K6J5G:F9F;F9G9FUnY2"}, "label": "sheep in middlr of picture"}]} {"id": 156823, "image": "COCO_train2014_000000156823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middled sized sheep to the right of the left most sheep\"."}], "task": "refering", "answer_template": "The \"the middled sized sheep to the right of the left most sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 231, 232, 233, 247, 248, 249, 250, 251, 263, 264, 265, 266, 267, 280, 281, 282, 283, 296, 297, 298, 299, 312, 313], "bbox": [0.4365384615384615, 0.5917656250000001, 0.7306787330316742, 0.86153125], "iscrowd": 0, "area": 12964.607750000001, "rle": {"size": [640, 442], "counts": "YQi32jc06K4K5L4K5L3N2N3M2N2N2N101N2O1O1N2O100O10001N6K5K5J6K6I6c^OhMh`0k2M3L4M3M3M3M3N2M3M3M2N100O1O2N100O2O6I9H9G2M01O1O001O1O1O1O1O1O1F:C=B>K42O0S@kLn>U3g@YMV?h31O1O1O1O1O1O1O1O1L4K5K5J6L4L4M3L4M3L4M2000001O01O00000000O10001O0O100001O3M3M3N3L3N2MN3M2N3N1N3L3L5K4M6I6K5K6J5G:F9F;F9G9FUnY2"}, "label": "the middled sized sheep to the right of the left most sheep"}]} {"id": 156823, "image": "COCO_train2014_000000156823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult sheep in the background of the two lambs\"."}], "task": "refering", "answer_template": "The \"an adult sheep in the background of the two lambs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 185, 186, 187, 188, 201, 202, 203, 204, 205, 217, 218, 219, 220, 221, 233, 234, 235, 236, 237, 250, 251, 252, 253, 267, 268, 284], "bbox": [0.5623981900452489, 0.45621875, 0.8358371040723982, 0.765234375], "iscrowd": 0, "area": 13684.362599999991, "rle": {"size": [640, 442], "counts": "jok43jc05L4L4K5L3M4L4kMQORAT1l>mNQAW1m>jNPAZ1n>gNn@]1R?cNk@a1S?`Nj@d1T?]Nh@h1V?ZNf@j1X?WNe@l1[?TNb@P2\\?QN`@T2^?mM_@W2_?jM^@Z2`?gM\\@^2c?cMY@_2g?bMV@`2j?aMR@c2m?^MP@d2P`0]Mm_Oe2T`0d0O0000001O00001O0000001O00010O00001O0000001O00001O00001O01O0001O1O1O2N1O2N1O2M2O1N3N1N3N1N3NhKn@c3P?[LSAe3l>ZLWAe3i>XLZAh3e>VL_Ai3a>TLbAk3^>TLeAk3[>RLhAn3W>PLmAn3S?01O04M3L4M3L4M3L4M3L4M1N10O001hNPA^MP?a2TA\\Mm>c2VAZMj>f2YAWMh>g2\\AVMe>i2_ASMa>m2bAPM_>n2eAoL[>Q3hAlLY>S3jAkLU>T3PBhLQ>W3RBfLn=Y3^1O001N101O0O2fNY1@a0@?AiY]1"}, "label": "an adult sheep in the background of the two lambs"}]} {"id": 156823, "image": "COCO_train2014_000000156823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mother sheep looking at the camera\"."}], "task": "refering", "answer_template": "The \"a mother sheep looking at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [169, 170, 171, 185, 186, 187, 188, 201, 202, 203, 204, 205, 217, 218, 219, 220, 221, 233, 234, 235, 236, 237, 250, 251, 252, 253, 267, 268, 284], "bbox": [0.5623981900452489, 0.45621875, 0.8358371040723982, 0.765234375], "iscrowd": 0, "area": 13684.362599999991, "rle": {"size": [640, 442], "counts": "jok43jc05L4L4K5L3M4L4kMQORAT1l>mNQAW1m>jNPAZ1n>gNn@]1R?cNk@a1S?`Nj@d1T?]Nh@h1V?ZNf@j1X?WNe@l1[?TNb@P2\\?QN`@T2^?mM_@W2_?jM^@Z2`?gM\\@^2c?cMY@_2g?bMV@`2j?aMR@c2m?^MP@d2P`0]Mm_Oe2T`0d0O0000001O00001O0000001O00010O00001O0000001O00001O00001O01O0001O1O1O2N1O2N1O2M2O1N3N1N3N1N3NhKn@c3P?[LSAe3l>ZLWAe3i>XLZAh3e>VL_Ai3a>TLbAk3^>TLeAk3[>RLhAn3W>PLmAn3S?01O04M3L4M3L4M3L4M3L4M1N10O001hNPA^MP?a2TA\\Mm>c2VAZMj>f2YAWMh>g2\\AVMe>i2_ASMa>m2bAPM_>n2eAoL[>Q3hAlLY>S3jAkLU>T3PBhLQ>W3RBfLn=Y3^1O001N101O0O2fNY1@a0@?AiY]1"}, "label": "a mother sheep looking at the camera"}]} {"id": 386211, "image": "COCO_train2014_000000386211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman wearing pink shirt\"."}], "task": "refering", "answer_template": "The \"woman wearing pink shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 82, 83, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143], "bbox": [0.41216, 0.32613333333333333, 1.0, 0.59792], "iscrowd": 0, "area": 19370.3229, "rle": {"size": [375, 500], "counts": "[c[23b;4M4K5K5K4M2M4L3N2N2N2O1N2N3M2O1N3M3M3N2M2N3LPN]FR2a94N2N2N2M2O1O001O01O0000001O0000001O00001O001O001O1O1O1O1O1O1O10O0001O0000000000005K3M2N001O1O1O1O001O1O1O00_N`NlH_1S7eNjH[1W7gNfHY1\\7hNaHX1`7jN]HV1d7kNZHU1h7lNUHT1l7QOmGP1U8VOaGl0a8UO[Gl0g8TOUGn0l8RORGo0P9k0N2M3N2O1O110O01O001O1O001O1O1O0000000000001O1O1O2N1O2N3M4L2N1O001O1O001O1O001O1O001O1O001O001O0000001O0000000000000000000000001O0000000000000000000000000000000O1000O100000000000000000000000000000000000000000000O100000000O10000O100O10000O10000O100O100000000O1000000000000O100000000O100000000O1000000O100000000O100000000O1000000O100000000O1000000O100000000O100000000O100000000O100000O10O100000000O100000000 in this image.", "objects": [{"patches": [80, 82, 83, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143], "bbox": [0.41216, 0.32613333333333333, 1.0, 0.59792], "iscrowd": 0, "area": 19370.3229, "rle": {"size": [375, 500], "counts": "[c[23b;4M4K5K5K4M2M4L3N2N2N2O1N2N3M2O1N3M3M3N2M2N3LPN]FR2a94N2N2N2M2O1O001O01O0000001O0000001O00001O001O001O1O1O1O1O1O1O10O0001O0000000000005K3M2N001O1O1O1O001O1O1O00_N`NlH_1S7eNjH[1W7gNfHY1\\7hNaHX1`7jN]HV1d7kNZHU1h7lNUHT1l7QOmGP1U8VOaGl0a8UO[Gl0g8TOUGn0l8RORGo0P9k0N2M3N2O1O110O01O001O1O001O1O1O0000000000001O1O1O2N1O2N3M4L2N1O001O1O001O1O001O1O001O1O001O001O0000001O0000000000000000000000001O0000000000000000000000000000000O1000O100000000000000000000000000000000000000000000O100000000O10000O100O10000O10000O100O100000000O1000000000000O100000000O100000000O1000000O100000000O100000000O1000000O100000000O1000000O100000000O100000000O100000000O100000O10O100000000O100000000 in this image.", "objects": [{"patches": [8, 9, 25, 26, 27, 28, 43, 44, 45, 46, 61, 62, 63, 64, 65, 78, 79, 80, 81, 82, 83, 96, 97, 99, 100, 112, 113, 114, 115, 129, 130, 131, 132, 133, 134, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 182, 183, 186, 199, 200, 203, 204, 205, 217, 218, 221, 222], "bbox": [0.08764, 0.02472, 0.63202, 0.9842666666666666], "iscrowd": 0, "area": 25268.733299999993, "rle": {"size": [375, 500], "counts": "b^`0?W;L4M31O1O001O1G9A?M3N2N2M3N2M3N2M3N2M3N2N0O100O10000O10000O10VO`F_O_9c0dFYO\\9g0hFUOW9m0mFnNR9T1QGhNo8X1VGcNi8_1e02O0O2N2O1N2O2N100000001O0O10001O000O101O0000000O010001O00000000001O0O10000N2M4K4M3M3O11O01O01O01O10O01O010O0010O011N4M3L3N3bIYNf2j1VMYNg2k1UMXNi2k1SMXNj2k1TMUNl2n1PMSNo2Q2mLPNS3S2iLnMW3T2fLmMY3W2cLjM]3X2`LhMa3Z2\\LgMd3\\2XLeMg3^2VLcMj3_2TLaMl3a2QL`Mo3b2nK_MQ4d2lK]MT4e2hK\\MY4f2dK[M\\4g2`K[M`4g2]KZMb4i2ZKYMf4i2WKWMj4R3kJoLV5\\3^JeLa5h41O1O1O0PLQJY2P6fMSJX2m5hMVJT2k5lMWJQ2i5oMZJn1g5RN[Jk1f5UN\\Jh1e5WN^Jf1b5[N`Jc1`5]NbJ`1^5`NeJ]1[5dNgJY1k3RMfLf1FP1]3dMjL\\1j0OY2jNlLW1n0KW2oNiLW1S1FT2VOfLT1Y1CR2ZOULb1l1POP2_OQLc1R2jNn1ElKb1Z2eNj1KiKb1`2_Nh10eKc1e2ZNg15`Kb1m2UNc1l2_NQMb1o2`NnLa1Q3aNmL`1S3bNjL^1W3cNgL^1X3eNeL\\1[3fNbLZ1_3_3`NkG]OU8b0nG]OR8b0QH\\Oo7b0UH[Ok7d0YHZOf7f0]HXO`7i0cHUO\\7l0gHROW7o0lHoNQ7S1RIkNl6V1WIhNf6Z1\\IfNa6[1`IfN]6[1eIdNZ6\\1hIdNb4A^Kl11cN_4b0`Jk0S1cNZ4R3hKmLV4T3kKlLS4U3oKjLo3W3SLgLl3Z3ULfLh3\\3ZLbLe3_3]L`La3a3`L_L^3b3dL\\L\\3d3fL[LZ3f3eLYL\\3g3eLXL[3h3fLVL[3j3eLVL[3j3fLULZ3k3gLSLZ3m3fLSLZ3m3hLPLY3Q4hLmKX3S4iLkKX3U4T210000000001O000001O01O1O1O1O1O100001N1O101N1O1O2O0O1O2O0O1O2N10XHULc7P41O00001O2N1O1O2N1N3N1N2^NnG[OU8TOlGi02Oh82YGKi84ZGIf87]GEd8<]GAd8?_G^Ob8a0hGTOY8m0oGiNR8W1X1O01O01O01O0010O01O0103M4K6K4L5J6K4L5J5LPXS2"}, "label": "the girl sitting on the bench"}]} {"id": 386211, "image": "COCO_train2014_000000386211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in black sitting on the bench\"."}], "task": "refering", "answer_template": "The \"the woman in black sitting on the bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 25, 26, 27, 28, 43, 44, 45, 46, 61, 62, 63, 64, 65, 78, 79, 80, 81, 82, 83, 96, 97, 99, 100, 112, 113, 114, 115, 129, 130, 131, 132, 133, 134, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 182, 183, 186, 199, 200, 203, 204, 205, 217, 218, 221, 222], "bbox": [0.08764, 0.02472, 0.63202, 0.9842666666666666], "iscrowd": 0, "area": 25268.733299999993, "rle": {"size": [375, 500], "counts": "b^`0?W;L4M31O1O001O1G9A?M3N2N2M3N2M3N2M3N2M3N2N0O100O10000O10000O10VO`F_O_9c0dFYO\\9g0hFUOW9m0mFnNR9T1QGhNo8X1VGcNi8_1e02O0O2N2O1N2O2N100000001O0O10001O000O101O0000000O010001O00000000001O0O10000N2M4K4M3M3O11O01O01O01O10O01O010O0010O011N4M3L3N3bIYNf2j1VMYNg2k1UMXNi2k1SMXNj2k1TMUNl2n1PMSNo2Q2mLPNS3S2iLnMW3T2fLmMY3W2cLjM]3X2`LhMa3Z2\\LgMd3\\2XLeMg3^2VLcMj3_2TLaMl3a2QL`Mo3b2nK_MQ4d2lK]MT4e2hK\\MY4f2dK[M\\4g2`K[M`4g2]KZMb4i2ZKYMf4i2WKWMj4R3kJoLV5\\3^JeLa5h41O1O1O0PLQJY2P6fMSJX2m5hMVJT2k5lMWJQ2i5oMZJn1g5RN[Jk1f5UN\\Jh1e5WN^Jf1b5[N`Jc1`5]NbJ`1^5`NeJ]1[5dNgJY1k3RMfLf1FP1]3dMjL\\1j0OY2jNlLW1n0KW2oNiLW1S1FT2VOfLT1Y1CR2ZOULb1l1POP2_OQLc1R2jNn1ElKb1Z2eNj1KiKb1`2_Nh10eKc1e2ZNg15`Kb1m2UNc1l2_NQMb1o2`NnLa1Q3aNmL`1S3bNjL^1W3cNgL^1X3eNeL\\1[3fNbLZ1_3_3`NkG]OU8b0nG]OR8b0QH\\Oo7b0UH[Ok7d0YHZOf7f0]HXO`7i0cHUO\\7l0gHROW7o0lHoNQ7S1RIkNl6V1WIhNf6Z1\\IfNa6[1`IfN]6[1eIdNZ6\\1hIdNb4A^Kl11cN_4b0`Jk0S1cNZ4R3hKmLV4T3kKlLS4U3oKjLo3W3SLgLl3Z3ULfLh3\\3ZLbLe3_3]L`La3a3`L_L^3b3dL\\L\\3d3fL[LZ3f3eLYL\\3g3eLXL[3h3fLVL[3j3eLVL[3j3fLULZ3k3gLSLZ3m3fLSLZ3m3hLPLY3Q4hLmKX3S4iLkKX3U4T210000000001O000001O01O1O1O1O1O100001N1O101N1O1O2O0O1O2O0O1O2N10XHULc7P41O00001O2N1O1O2N1N3N1N2^NnG[OU8TOlGi02Oh82YGKi84ZGIf87]GEd8<]GAd8?_G^Ob8a0hGTOY8m0oGiNR8W1X1O01O01O01O0010O01O0103M4K6K4L5J6K4L5J5LPXS2"}, "label": "the woman in black sitting on the bench"}]} {"id": 230570, "image": "COCO_train2014_000000230570.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a perfectly sliced piece of sandwich sits in a wrapper\"."}], "task": "refering", "answer_template": "The \"a perfectly sliced piece of sandwich sits in a wrapper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 263, 264, 265, 266, 267, 275, 276, 277, 278, 279, 280, 281, 282, 283, 290, 291, 292, 293, 294, 295, 296, 297, 305, 306, 307, 308, 309, 310, 311, 312], "bbox": [0.37231850117096016, 0.6997656250000001, 0.8808665105386416, 0.9071250000000001], "iscrowd": 0, "area": 24006.175999999996, "rle": {"size": [640, 427], "counts": "TmS36_c0@`0@`0B>N2M3N2N2N2N2N2N2N2N3M2N1N2O1O1O1O1O1O2N1O1O1O1N2000001O00000O1000001O00000O100000001O000O10000000001O0O1000000000001O0000000O1000001O000000000000001O0000000O1000001O000000000000001OO1000000O100000001N100000000O1000000O1000001O0O1000000O100000000O10001O000O1000000O100000000O101O00000O1000000O100000000O2O00000O100000000O100000001N1000000O1003M00001O000100O101N1O100O2N1N2N2M4M2M3M3N3L3N2M3M4M2M3N2M4L3N2M3N3L3M3N2M4M8G;E in this image.", "objects": [{"patches": [246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 263, 264, 265, 266, 267, 275, 276, 277, 278, 279, 280, 281, 282, 283, 290, 291, 292, 293, 294, 295, 296, 297, 305, 306, 307, 308, 309, 310, 311, 312], "bbox": [0.37231850117096016, 0.6997656250000001, 0.8808665105386416, 0.9071250000000001], "iscrowd": 0, "area": 24006.175999999996, "rle": {"size": [640, 427], "counts": "TmS36_c0@`0@`0B>N2M3N2N2N2N2N2N2N2N3M2N1N2O1O1O1O1O1O2N1O1O1O1N2000001O00000O1000001O00000O100000001O000O10000000001O0O1000000000001O0000000O1000001O000000000000001O0000000O1000001O000000000000001OO1000000O100000001N100000000O1000000O1000001O0O1000000O100000000O10001O000O1000000O100000000O101O00000O1000000O100000000O2O00000O100000000O100000001N1000000O1003M00001O000100O101N1O100O2N1N2N2M4M2M3M3N3L3N2M3M4M2M3N2M4L3N2M3N3L3M3N2M4M8G;E in this image.", "objects": [{"patches": [271, 272, 273, 274, 285, 286, 287, 288, 289, 290, 300, 301, 302, 303, 304, 305, 306, 315, 316, 317, 318, 319, 320, 321, 331, 332, 333, 334, 335], "bbox": [0.013114754098360656, 0.78878125, 0.451943793911007, 0.9939375], "iscrowd": 0, "area": 16686.24299999999, "rle": {"size": [640, 427], "counts": "lY44hc04L4O1O1O1m\\OCQb04l]O:2Hha05S^O44Kba07W^ON6N`a09V^OJ9Naa0k0]^OWOba0j0]^OXOba0h0\\^OZOca0g0\\^OZOca0g0[^O\\Oda0d0[^O]Oda0a1O0O101N100O2O0O2O7H7J4K1O2N101N1O2N101N1O2O0O2N1O2O0O2N1O2O0O2N1O100O100000001O00000000000001O0000000000000001O000000000001O00001O1O001N101O1N101O1O0O10000O1000000O1000000O10000O10002N1N3N2N2M2O1O001N2O1O1N01O10O01O010O00100O0101N101N101O0O2O1N101N101O0O2O0O2O>B2N3M2N3M2N3M2N3M1O001O001O1O001O002N2N2N1O2N2N1O2N2N1O1O001O1O1N1N3L4M3M3M2N3M^ea4"}, "label": "smaller portion of a sandwich on tin foil"}]} {"id": 418989, "image": "COCO_train2014_000000418989.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe whose head is not visible\"."}], "task": "refering", "answer_template": "The \"the giraffe whose head is not visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 276, 277, 299, 300], "bbox": [0.025312499999999998, 0.5147652582159624, 0.14890625, 0.9465258215962442], "iscrowd": 0, "area": 7263.428650000002, "rle": {"size": [426, 640], "counts": "[Q73m<:G:jLCSIg0g6N`HB>A`0Am`R7"}, "label": "the giraffe whose head is not visible"}]} {"id": 418989, "image": "COCO_train2014_000000418989.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a giraffe standing behind another giraffe\"."}], "task": "refering", "answer_template": "The \"the back of a giraffe standing behind another giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 276, 277, 299, 300], "bbox": [0.025312499999999998, 0.5147652582159624, 0.14890625, 0.9465258215962442], "iscrowd": 0, "area": 7263.428650000002, "rle": {"size": [426, 640], "counts": "[Q73m<:G:jLCSIg0g6N`HB>A`0Am`R7"}, "label": "the back of a giraffe standing behind another giraffe"}]} {"id": 418989, "image": "COCO_train2014_000000418989.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall giraffe eating\"."}], "task": "refering", "answer_template": "The \"a tall giraffe eating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 133, 134, 135, 142, 143, 144, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 195, 210, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 261, 262, 263, 279, 280, 281, 284, 285, 286, 302, 303, 304, 307, 308, 325, 326, 327, 330, 331], "bbox": [0.139109375, 0.13446009389671362, 0.879515625, 0.9906338028169014], "iscrowd": 0, "area": 61681.87640000004, "rle": {"size": [426, 640], "counts": "f\\U14a9:cINm5m0`I[OWN]OQ7S2TJiNcNWOm6b2VJ_NfNQOP7h2SJ_NfNkNV7j2nIhNQ6\\1iIhNU6]1dIfN[6^1_IeN`6`1YIcNf6a1TIcNj6o300O001O1O001O1O001O1O10O01O1O00^Ob0L401N100008I1N2N2O1N2N2O1N3M2O1N2N2O1N2N2N2O1N2N2O1N2N100O2N100O1O1O100O1O100O2OQOdJPJ[5U5eKgJ[4k4WLQKh3k4bLPK]3l4lLPKT3k4g2M4K4L4L5L3L5L3101N2N2O1N2O1N2N2O1N101N2O1N2N2O1N2OO01O010O10O010O01O010O010O1O010O001O1O001O001O1O001O1O001O001O1O001O1O0cKlJe0U5QOUKn0k4iN_KV1b4dNdK[1\\4bNgK^1Y4^NlKa1T4\\NPLc1Q4XNTLg1l3VNXLi1h3SN]Ll1c3nMcLR2]3gMkLX2V3`MRM_2n2ZMZMe2f2UMaMj2_2nLjMQ3V2`LYN`3i50A>2N2O0O`0@100O100O1O100O100O1O100O1O100O100O1O100OO1L5L4K5L4K5K5L4K5I7I7H8K5M3L4L4L4M3L4L4L4L4M3L4L4N2M3M3M3M3M3M3M3M3N2M3M3M3dMUHVOn7j0VHoNm7P1YHhNj7W1\\HaNg7_1^HYNe7f1aHSNa7m1cHcMi7\\2c100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O10000O10O10O1000000O10000O10O10O10000O10000O10O10O1000000O10000O01000O10000O10000O10O10O1000000O10000O01000O10000O10000O01000O1000000O10000O010000000O100000O10O100000O010000000O0100000000O010000000O010000000O0100000O1000O1000O1000O1000O1000O1000O100001N10001N101O0O2O001N10001N101O0O2O001D;01N10001N101O1O1N2O2N1N2O1O1O1N2O1O2M2g0YO1O0000001O01O01N1O1O2N1O1O1O2N1O1O2N1O4L4L4L5K4L4L4L5K4L4L4L5K4L4Lmko0"}, "label": "a tall giraffe eating"}]} {"id": 418989, "image": "COCO_train2014_000000418989.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe eating leaves from a tree along side another giraffe\"."}], "task": "refering", "answer_template": "The \"a giraffe eating leaves from a tree along side another giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 133, 134, 135, 142, 143, 144, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 195, 210, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 261, 262, 263, 279, 280, 281, 284, 285, 286, 302, 303, 304, 307, 308, 325, 326, 327, 330, 331], "bbox": [0.139109375, 0.13446009389671362, 0.879515625, 0.9906338028169014], "iscrowd": 0, "area": 61681.87640000004, "rle": {"size": [426, 640], "counts": "f\\U14a9:cINm5m0`I[OWN]OQ7S2TJiNcNWOm6b2VJ_NfNQOP7h2SJ_NfNkNV7j2nIhNQ6\\1iIhNU6]1dIfN[6^1_IeN`6`1YIcNf6a1TIcNj6o300O001O1O001O1O001O1O10O01O1O00^Ob0L401N100008I1N2N2O1N2N2O1N3M2O1N2N2O1N2N2N2O1N2N2O1N2N100O2N100O1O1O100O1O100O2OQOdJPJ[5U5eKgJ[4k4WLQKh3k4bLPK]3l4lLPKT3k4g2M4K4L4L5L3L5L3101N2N2O1N2O1N2N2O1N101N2O1N2N2O1N2OO01O010O10O010O01O010O010O1O010O001O1O001O001O1O001O1O001O001O1O001O1O0cKlJe0U5QOUKn0k4iN_KV1b4dNdK[1\\4bNgK^1Y4^NlKa1T4\\NPLc1Q4XNTLg1l3VNXLi1h3SN]Ll1c3nMcLR2]3gMkLX2V3`MRM_2n2ZMZMe2f2UMaMj2_2nLjMQ3V2`LYN`3i50A>2N2O0O`0@100O100O1O100O100O1O100O1O100O100O1O100OO1L5L4K5L4K5K5L4K5I7I7H8K5M3L4L4L4M3L4L4L4L4M3L4L4N2M3M3M3M3M3M3M3M3N2M3M3M3dMUHVOn7j0VHoNm7P1YHhNj7W1\\HaNg7_1^HYNe7f1aHSNa7m1cHcMi7\\2c100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O10000O10O10O1000000O10000O10O10O10000O10000O10O10O1000000O10000O01000O10000O10000O10O10O1000000O10000O01000O10000O10000O01000O1000000O10000O010000000O100000O10O100000O010000000O0100000000O010000000O010000000O0100000O1000O1000O1000O1000O1000O1000O100001N10001N101O0O2O001N10001N101O0O2O001D;01N10001N101O1O1N2O2N1N2O1O1O1N2O1O2M2g0YO1O0000001O01O01N1O1O2N1O1O1O2N1O1O2N1O4L4L4L5K4L4L4L5K4L4L4L5K4L4Lmko0"}, "label": "a giraffe eating leaves from a tree along side another giraffe"}]} {"id": 74945, "image": "COCO_train2014_000000074945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing beige dress holding umbrella at a picnic\"."}], "task": "refering", "answer_template": "The \"a woman wearing beige dress holding umbrella at a picnic\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 259, 260, 261, 262, 263, 284, 285, 286, 308, 309], "bbox": [0.254359375, 0.30599609375, 0.45618749999999997, 0.7547070312499999], "iscrowd": 0, "area": 14572.5139, "rle": {"size": [512, 640], "counts": "oga2X1c=U100O1O001O1O1O1O1O00100O1O1O001O1O1O1O001O100G9^Ob0[DZLm:h3TEZLg:i3XEZLd:h3\\EZL_:i3`EZL\\:h3dEYL[:f3eE\\LZ:d3fE\\LZ:d3eE^LZ:e3cE\\L[:g3bE[L]:h3`EYL_:j3]EXLb:k3[EVLd:d4O1O001O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1OcKWFd2h9ZM\\Fe2c9YM`Fg2`9VMdFi2\\9TMgFl2Y9RMjFm2U9PMoFP3Q9mLSGR3m8jLWGV3i8fL\\GY3c8eL`G[3`8aLdG_3\\8^LhGb1bNGf9cNlGe1^NIe9_NRHf1ZNKd9\\NUHg1YNMb9XNZHi1TN0b9TN]Hk1RN1i;MWD4i;JYD6h;HYD8g;GYD:h;^O_Db0]<[NjCe1X=10000O100O2O000O01O1O0O2O1O0O2O001O1N101O1N101O0O101O001N10001N101[Od0J7L3L4L5L3Lge]5"}, "label": "a woman wearing beige dress holding umbrella at a picnic"}]} {"id": 74945, "image": "COCO_train2014_000000074945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman wearing glasses and white dress\"."}], "task": "refering", "answer_template": "The \"woman wearing glasses and white dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 259, 260, 261, 262, 263, 284, 285, 286, 308, 309], "bbox": [0.254359375, 0.30599609375, 0.45618749999999997, 0.7547070312499999], "iscrowd": 0, "area": 14572.5139, "rle": {"size": [512, 640], "counts": "oga2X1c=U100O1O001O1O1O1O1O00100O1O1O001O1O1O1O001O100G9^Ob0[DZLm:h3TEZLg:i3XEZLd:h3\\EZL_:i3`EZL\\:h3dEYL[:f3eE\\LZ:d3fE\\LZ:d3eE^LZ:e3cE\\L[:g3bE[L]:h3`EYL_:j3]EXLb:k3[EVLd:d4O1O001O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1OcKWFd2h9ZM\\Fe2c9YM`Fg2`9VMdFi2\\9TMgFl2Y9RMjFm2U9PMoFP3Q9mLSGR3m8jLWGV3i8fL\\GY3c8eL`G[3`8aLdG_3\\8^LhGb1bNGf9cNlGe1^NIe9_NRHf1ZNKd9\\NUHg1YNMb9XNZHi1TN0b9TN]Hk1RN1i;MWD4i;JYD6h;HYD8g;GYD:h;^O_Db0]<[NjCe1X=10000O100O2O000O01O1O0O2O1O0O2O001O1N101O1N101O0O101O001N10001N101[Od0J7L3L4L5L3Lge]5"}, "label": "woman wearing glasses and white dress"}]} {"id": 74945, "image": "COCO_train2014_000000074945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young girl in a pink dress and a pink bonnet\"."}], "task": "refering", "answer_template": "The \"a young girl in a pink dress and a pink bonnet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 117, 118, 119, 120, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 185, 186, 187, 188, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281, 282, 283, 284, 300, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328], "bbox": [0.06281250000000001, 0.274140625, 0.396296875, 0.8107421875], "iscrowd": 0, "area": 32865.756299999994, "rle": {"size": [512, 640], "counts": "T\\d02i?8I7H5L4KXOQAi0k>8jNV1XLjMZIc2b6aMUIg2i6[M\\H_3b7eLoGi3o7YLgGQ4W8QLdGT4Z8oK_GW4_8kK\\GZ4b8iKXG]4e8fKUG_4i8fKlFb4Q9X1N2M4L3L4L4L5M2N2N2N3L3O1O2O0O1O2O000O2O000O101O0O1000001O0000001O000000001O0000001O01O001O1O001O001O00O2N1O1O101O01O010O001O010O2O0O10O00fI`HU4_7lK`HT4`7mK`HR4`7oK`Ho3a7QL`Hn3`7SL`Hk3b7UL^Hj3b7VL`Hg3a7ZL_HQNH^5i7bL_HnMJ^5h7dLfHZ3Z7gLgHPNAm4j7RMfHmMFn4d7UMTIg2m6YMWIb2j6_MXI^2h6bMZI\\2f6dM\\IY2e6gM^IV2c6iM_IT2b6lMZJX1f5hN[JW1f5hN[JW1f5hNZJW1g5iNZJV1g5iNYJW1g5^MSH;V2W2h5ZMWH in this image.", "objects": [{"patches": [95, 117, 118, 119, 120, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 185, 186, 187, 188, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 236, 237, 254, 255, 256, 257, 258, 259, 260, 277, 278, 279, 280, 281, 282, 283, 284, 300, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328], "bbox": [0.06281250000000001, 0.274140625, 0.396296875, 0.8107421875], "iscrowd": 0, "area": 32865.756299999994, "rle": {"size": [512, 640], "counts": "T\\d02i?8I7H5L4KXOQAi0k>8jNV1XLjMZIc2b6aMUIg2i6[M\\H_3b7eLoGi3o7YLgGQ4W8QLdGT4Z8oK_GW4_8kK\\GZ4b8iKXG]4e8fKUG_4i8fKlFb4Q9X1N2M4L3L4L4L5M2N2N2N3L3O1O2O0O1O2O000O2O000O101O0O1000001O0000001O000000001O0000001O01O001O1O001O001O00O2N1O1O101O01O010O001O010O2O0O10O00fI`HU4_7lK`HT4`7mK`HR4`7oK`Ho3a7QL`Hn3`7SL`Hk3b7UL^Hj3b7VL`Hg3a7ZL_HQNH^5i7bL_HnMJ^5h7dLfHZ3Z7gLgHPNAm4j7RMfHmMFn4d7UMTIg2m6YMWIb2j6_MXI^2h6bMZI\\2f6dM\\IY2e6gM^IV2c6iM_IT2b6lMZJX1f5hN[JW1f5hN[JW1f5hNZJW1g5iNZJV1g5iNYJW1g5^MSH;V2W2h5ZMWH in this image.", "objects": [{"patches": [110, 111, 112, 133, 134, 135, 136, 156, 157, 158, 159, 180, 181, 182, 203, 204, 205, 226, 227, 228, 249, 250, 251, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343], "bbox": [0.690125, 0.2344921875, 0.984796875, 0.792890625], "iscrowd": 0, "area": 27248.644500000002, "rle": {"size": [512, 640], "counts": "^[m6h0W?1N2N3M2M4M2O101N1O2N1O1O2O0O2N1O101O000000000O2O000000000O10001O0O1000000O10001O0O1000000O101O0O10000O101O000XJQN^MP2a2TN\\Ml1d2VNZMj1e2XNYMi1g2WNYMj1e2WN[Mi1e2WNZMj1e2WN[Mi1e2WN[Mi1^2^NaMc1U2gNkMY1n1nNRNS1l1nNSNS1m1POPNP1P2SOmMm0S2VOiMk0W2XOfMh0Z2[ObMg0]2\\O`Md0`2_O]Mb0b2AZM`0f2BXM>h2DUM=k2FRM:n2HoL9Q3ImL7R3KmL5S3LnJgMH^2Z5LkJhMK\\2Z5LhJlMLY2\\5LeJnMMW2^5LaJPNOV2`5K^JRN0T2b5K[JSN1T2d5IXJWN2Q2f5IUJYN2P2i5GRJVOZOT1d6GoIc2Q6]MlIf2T6ZMjIh2U6YMhIj2X6WMeIk2[6VMaIm2_6SM_Io2a6RM[IQ3e6oLYIS3g6mLWIU3i6kLTIX3l6iLQIY3o6gLnH\\3R7i20000O10000000000000000000O100000O10000000000003M5K6I6K2N00001O0TIYH_5g7^J]Ha5c7hIZH16V6`7iI_HI5]6]7hIcHD3d6Z7hIWIW6i6hIYIW6i6fIYIX6l6`IYI_6T801O3M3M3L3N2N1O2N2N2M2O2N2N2N1O2N2N2N1M4hLZEQ1_;QNiDl1hERB9R>CQB:S>BPB;m>L3M4M[d4"}, "label": "a kneeling person in a black top hat"}]} {"id": 74945, "image": "COCO_train2014_000000074945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in black sitting down\"."}], "task": "refering", "answer_template": "The \"the woman in black sitting down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 133, 134, 135, 136, 156, 157, 158, 159, 180, 181, 182, 203, 204, 205, 226, 227, 228, 249, 250, 251, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343], "bbox": [0.690125, 0.2344921875, 0.984796875, 0.792890625], "iscrowd": 0, "area": 27248.644500000002, "rle": {"size": [512, 640], "counts": "^[m6h0W?1N2N3M2M4M2O101N1O2N1O1O2O0O2N1O101O000000000O2O000000000O10001O0O1000000O10001O0O1000000O101O0O10000O101O000XJQN^MP2a2TN\\Ml1d2VNZMj1e2XNYMi1g2WNYMj1e2WN[Mi1e2WNZMj1e2WN[Mi1e2WN[Mi1^2^NaMc1U2gNkMY1n1nNRNS1l1nNSNS1m1POPNP1P2SOmMm0S2VOiMk0W2XOfMh0Z2[ObMg0]2\\O`Md0`2_O]Mb0b2AZM`0f2BXM>h2DUM=k2FRM:n2HoL9Q3ImL7R3KmL5S3LnJgMH^2Z5LkJhMK\\2Z5LhJlMLY2\\5LeJnMMW2^5LaJPNOV2`5K^JRN0T2b5K[JSN1T2d5IXJWN2Q2f5IUJYN2P2i5GRJVOZOT1d6GoIc2Q6]MlIf2T6ZMjIh2U6YMhIj2X6WMeIk2[6VMaIm2_6SM_Io2a6RM[IQ3e6oLYIS3g6mLWIU3i6kLTIX3l6iLQIY3o6gLnH\\3R7i20000O10000000000000000000O100000O10000000000003M5K6I6K2N00001O0TIYH_5g7^J]Ha5c7hIZH16V6`7iI_HI5]6]7hIcHD3d6Z7hIWIW6i6hIYIW6i6fIYIX6l6`IYI_6T801O3M3M3L3N2N1O2N2N2M2O2N2N2N1O2N2N2N1M4hLZEQ1_;QNiDl1hERB9R>CQB:S>BPB;m>L3M4M[d4"}, "label": "the woman in black sitting down"}]} {"id": 558276, "image": "COCO_train2014_000000558276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown bear\"."}], "task": "refering", "answer_template": "The \"a brown bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 115, 123, 124, 125, 126, 127, 128, 129, 130, 138, 139, 140, 141, 142, 143, 144, 145, 146, 154, 155, 156, 157, 158, 159, 160, 161, 168, 169, 174, 175, 176, 177, 190, 191, 192, 205, 206, 207, 221, 222], "bbox": [0.21220140515222483, 0.287640625, 0.8689929742388759, 0.617984375], "iscrowd": 0, "area": 27105.896500000006, "rle": {"size": [640, 427], "counts": "kSi1i0ib0>N2N2O1N2O2M2O1N2O1N2N2O1N2O1N2O1N2O1V_ORN`?P2\\@TNc?m1X@XNf?j1V@[Ng?g1T@^Nk?c1Q@aNm?b1m_OcNR`0^1j_OgNS`0d2O1N11N2O001O1N2O1O1O0O2O1O1O1N2O0010000O1000O1000O100000O0100000000O0100000O1000O10O10000000O10000000O10000000000000O1000O10000000001O00001N1000001O0000001O000O2O0000001O0000001O00001O0O10001O0000001O00001O0000001O000O101O00001O0000001O0000001O000O2O0000010O01O1O100O1O100O1O1O010O1O1O100O1O100O001O100O1O1O100O1O010O001O10O01O0010O01O10O01O0010O01O11N101N1015J9H8G9H3L0010O001O001O0010O01O001O0010O01O001O0O2N2N2N2O1N3M2N2O1N3M2N2O1N3M2N2O1N3M2N2O2M2N2N3N1N3M2N2N3M2M4L3M4L4L3L5L4L3M4L4L3M4L3M4L4L3M4L4K]dR1"}, "label": "a brown bear"}]} {"id": 558276, "image": "COCO_train2014_000000558276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown stuffed teddy bear that is on top of a white teddy bear\"."}], "task": "refering", "answer_template": "The \"a brown stuffed teddy bear that is on top of a white teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 115, 123, 124, 125, 126, 127, 128, 129, 130, 138, 139, 140, 141, 142, 143, 144, 145, 146, 154, 155, 156, 157, 158, 159, 160, 161, 168, 169, 174, 175, 176, 177, 190, 191, 192, 205, 206, 207, 221, 222], "bbox": [0.21220140515222483, 0.287640625, 0.8689929742388759, 0.617984375], "iscrowd": 0, "area": 27105.896500000006, "rle": {"size": [640, 427], "counts": "kSi1i0ib0>N2N2O1N2O2M2O1N2O1N2N2O1N2O1N2O1N2O1V_ORN`?P2\\@TNc?m1X@XNf?j1V@[Ng?g1T@^Nk?c1Q@aNm?b1m_OcNR`0^1j_OgNS`0d2O1N11N2O001O1N2O1O1O0O2O1O1O1N2O0010000O1000O1000O100000O0100000000O0100000O1000O10O10000000O10000000O10000000000000O1000O10000000001O00001N1000001O0000001O000O2O0000001O0000001O00001O0O10001O0000001O00001O0000001O000O101O00001O0000001O0000001O000O2O0000010O01O1O100O1O100O1O1O010O1O1O100O1O100O001O100O1O1O100O1O010O001O10O01O0010O01O10O01O0010O01O11N101N1015J9H8G9H3L0010O001O001O0010O01O001O0010O01O001O0O2N2N2N2O1N3M2N2O1N3M2N2O1N3M2N2O1N3M2N2O2M2N2N3N1N3M2N2N3M2M4L3M4L4L3L5L4L3M4L4L3M4L3M4L4L3M4L4K]dR1"}, "label": "a brown stuffed teddy bear that is on top of a white teddy bear"}]} {"id": 558276, "image": "COCO_train2014_000000558276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white stuffed animal beneath a brown stuffed animal\"."}], "task": "refering", "answer_template": "The \"a white stuffed animal beneath a brown stuffed animal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 158, 159, 169, 170, 171, 172, 173, 174, 175, 184, 185, 186, 187, 188, 189, 190, 198, 199, 200, 201, 202, 203, 204, 205, 206, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 273, 274, 275, 277, 278, 279, 280, 281, 282, 283, 296, 297], "bbox": [0.13386416861826697, 0.44378124999999996, 0.9047072599531615, 0.84640625], "iscrowd": 0, "area": 54524.201150000015, "rle": {"size": [640, 427], "counts": "PgT14jc02N2N2M3N2SOBV^O`0ia0BS^Oa0ka0AS^Oa0ka0BR^Oa0la0AQ^Oa0ma0AP^Ob0na0Ao]Oa0oa0Bn]O`0Qb0Al]Ob0Rb0g0N2O2M4L4M3L5K4L4M3L4L4M2M2N2O1N2N2N2O1N2N3N1N2N2O1N2N2N2O1N3M2O1N2N2N2O1N2N2O1N2N2O1N2N2N2M3M3L4M3L4M3L4O1O1N2O1N2O1O1N2O1O1N101O1N101N2O1O0O2O1O0O200000O0100000O1000O01000000O10O10O1000000O0100000O11O0O10001O00000O2O000000001N1000001O000O10001O00000O2O000000000O100000000000O2O000000000000001N11O2N1O2N2N2N101N2N1O1O1O1O100O1O1O1O1O2N1O01O0O1000000000001O0000001O1O001O1O1O1O1O001O1O1O1O0O2O1O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O0N3M3N2M4L5L4K6J5L4K5K5L4K5K5K5L5J3N2N3M2N2M3NRAZLe=d3VBbLk=\\3oAkLS>S3gATMY>k2aA[M`>d2ZAbMg>]2RAkMn>c31O2N100O1O100O2N100O1O1O100O2N100O1N2O2N1N3N2N1N3N1O2M2O2N1N3N1O2M2O2N1N3N1O2M2O2N1N3M2N3M2M4M2N3M2N3M3L3N3M2N3M2N3L3M4I6J7I6J7I6J7I^Ui0"}, "label": "a white stuffed animal beneath a brown stuffed animal"}]} {"id": 558276, "image": "COCO_train2014_000000558276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white stuffed animal on the bottom\"."}], "task": "refering", "answer_template": "The \"the white stuffed animal on the bottom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 158, 159, 169, 170, 171, 172, 173, 174, 175, 184, 185, 186, 187, 188, 189, 190, 198, 199, 200, 201, 202, 203, 204, 205, 206, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 273, 274, 275, 277, 278, 279, 280, 281, 282, 283, 296, 297], "bbox": [0.13386416861826697, 0.44378124999999996, 0.9047072599531615, 0.84640625], "iscrowd": 0, "area": 54524.201150000015, "rle": {"size": [640, 427], "counts": "PgT14jc02N2N2M3N2SOBV^O`0ia0BS^Oa0ka0AS^Oa0ka0BR^Oa0la0AQ^Oa0ma0AP^Ob0na0Ao]Oa0oa0Bn]O`0Qb0Al]Ob0Rb0g0N2O2M4L4M3L5K4L4M3L4L4M2M2N2O1N2N2N2O1N2N3N1N2N2O1N2N2N2O1N3M2O1N2N2N2O1N2N2O1N2N2O1N2N2N2M3M3L4M3L4M3L4O1O1N2O1N2O1O1N2O1O1N101O1N101N2O1O0O2O1O0O200000O0100000O1000O01000000O10O10O1000000O0100000O11O0O10001O00000O2O000000001N1000001O000O10001O00000O2O000000000O100000000000O2O000000000000001N11O2N1O2N2N2N101N2N1O1O1O1O100O1O1O1O1O2N1O01O0O1000000000001O0000001O1O001O1O1O1O1O001O1O1O1O0O2O1O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O0N3M3N2M4L5L4K6J5L4K5K5L4K5K5K5L5J3N2N3M2N2M3NRAZLe=d3VBbLk=\\3oAkLS>S3gATMY>k2aA[M`>d2ZAbMg>]2RAkMn>c31O2N100O1O100O2N100O1O1O100O2N100O1N2O2N1N3N2N1N3N1O2M2O2N1N3N1O2M2O2N1N3N1O2M2O2N1N3M2N3M2M4M2N3M2N3M3L3N3M2N3M2N3L3M4I6J7I6J7I6J7I^Ui0"}, "label": "the white stuffed animal on the bottom"}]} {"id": 33991, "image": "COCO_train2014_000000033991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby looks at an elephant\"."}], "task": "refering", "answer_template": "The \"a baby looks at an elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 116, 117, 118, 119, 134, 135, 136, 137, 138, 151, 152, 153, 154, 155, 156, 157, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 205, 206, 207, 208, 209, 210, 211, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 258, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 294, 295, 296, 297, 298, 299, 312, 313, 314, 316, 330, 331, 332, 348, 349, 350, 367, 368], "bbox": [0.3469667318982388, 0.251125, 0.77573385518591, 0.884015625], "iscrowd": 0, "area": 45167.04175, "rle": {"size": [640, 511], "counts": "\\R_31kc0:F;D;F:F:E;l^O\\N\\?n1f_OkNU`0b2L4K5L5K4M3100O001O100O1O010O1O1O100O001Ob0^O1O00oKbLaG]3[8gLeGY3W8kLjGT3R8PMoGo2m7VMRHj2m7WMTHh2l7XMTHh2k7YMVHf2j7ZMWHe2h7]MWHe2g6eLUFf0U3g2\\6lL\\F>X3i2S6PMdF7X1ZOMa3m7WMkFOY1]ON^3f7^MQGHY1_ON^3_7bMXGFV1]O3X3X7oM]G_OT1^O6R3P7[NdGZOQ1\\O;m2h6gNjGTOm0\\Ob0f2^6UOPHnNk0ZOg0`2Y6_OTHkNg0YOl0[2X6DRHmNe0XOP1U2X6HRHjNe0]OQ1n1Y6MnGhNh0@Q1i1X6j1gH`LQ1d1W6k1iHeLo0]1Y6m1hHiLo0X1X6o1hHmLP1R1W6Q2iHPMP1m0V6S2jHSMP1g0W6U2iHXMo0a0W6V2jH]Mo0;V6X2kH_MP16V6Z2jHcMP11U6\\2kHeMQ1MS6^2kHhMS1GR6a2kHkMS1BR6b2kHnMT1^OP6c2lHRNV1XOm5f2mHTNW1SOl5i2lHXNX1mNl5j2lH[NY1iNj5l2lH^N^1`Nf5R3lHaNb1VNc5X3jHeNi1hM`5b3hHhNm1\\M^5l3dHlNS2mL]5V4`HoNX2aL\\5_4[HSOc:l0\\EWOc:h0]E[Oa:e0]E^Oa:b0_E@`:`0^ED`:<\\EIc:7YENf:2VE4h:KUE:i:`4O1O2N1O1O1O1N2O1O1O1O1O1N2O1O1O2N1O1O1N2O1O1O1O1O1O100O1O1O100O1O10O01O1O100OM3M3N3L3M3M4L3M3M3M4M2M3M3H9F9H8M4M3L4M3L4M3L4M3L4M3M3L4M3L4M3L4M4K4M3L4M3L4K6J5K5L4K5K5K6Kj0WO3O000O101O0O10001N10002M3N2N2N2M3N2N2M3N2N2M3N2N2M3N2N1N3N2N2M3L4L4L3N3L4L4L4L6J:F:F9H]bW2"}, "label": "a baby looks at an elephant"}]} {"id": 33991, "image": "COCO_train2014_000000033991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small child in an orange shirt\"."}], "task": "refering", "answer_template": "The \"a small child in an orange shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 116, 117, 118, 119, 134, 135, 136, 137, 138, 151, 152, 153, 154, 155, 156, 157, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 205, 206, 207, 208, 209, 210, 211, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 258, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 294, 295, 296, 297, 298, 299, 312, 313, 314, 316, 330, 331, 332, 348, 349, 350, 367, 368], "bbox": [0.3469667318982388, 0.251125, 0.77573385518591, 0.884015625], "iscrowd": 0, "area": 45167.04175, "rle": {"size": [640, 511], "counts": "\\R_31kc0:F;D;F:F:E;l^O\\N\\?n1f_OkNU`0b2L4K5L5K4M3100O001O100O1O010O1O1O100O001Ob0^O1O00oKbLaG]3[8gLeGY3W8kLjGT3R8PMoGo2m7VMRHj2m7WMTHh2l7XMTHh2k7YMVHf2j7ZMWHe2h7]MWHe2g6eLUFf0U3g2\\6lL\\F>X3i2S6PMdF7X1ZOMa3m7WMkFOY1]ON^3f7^MQGHY1_ON^3_7bMXGFV1]O3X3X7oM]G_OT1^O6R3P7[NdGZOQ1\\O;m2h6gNjGTOm0\\Ob0f2^6UOPHnNk0ZOg0`2Y6_OTHkNg0YOl0[2X6DRHmNe0XOP1U2X6HRHjNe0]OQ1n1Y6MnGhNh0@Q1i1X6j1gH`LQ1d1W6k1iHeLo0]1Y6m1hHiLo0X1X6o1hHmLP1R1W6Q2iHPMP1m0V6S2jHSMP1g0W6U2iHXMo0a0W6V2jH]Mo0;V6X2kH_MP16V6Z2jHcMP11U6\\2kHeMQ1MS6^2kHhMS1GR6a2kHkMS1BR6b2kHnMT1^OP6c2lHRNV1XOm5f2mHTNW1SOl5i2lHXNX1mNl5j2lH[NY1iNj5l2lH^N^1`Nf5R3lHaNb1VNc5X3jHeNi1hM`5b3hHhNm1\\M^5l3dHlNS2mL]5V4`HoNX2aL\\5_4[HSOc:l0\\EWOc:h0]E[Oa:e0]E^Oa:b0_E@`:`0^ED`:<\\EIc:7YENf:2VE4h:KUE:i:`4O1O2N1O1O1O1N2O1O1O1O1O1N2O1O1O2N1O1O1N2O1O1O1O1O1O100O1O1O100O1O10O01O1O100OM3M3N3L3M3M4L3M3M3M4M2M3M3H9F9H8M4M3L4M3L4M3L4M3L4M3M3L4M3L4M3L4M4K4M3L4M3L4K6J5K5L4K5K5K6Kj0WO3O000O101O0O10001N10002M3N2N2N2M3N2N2M3N2N2M3N2N2M3N2N1N3N2N2M3L4L4L3N3L4L4L4L6J:F:F9H]bW2"}, "label": "a small child in an orange shirt"}]} {"id": 33991, "image": "COCO_train2014_000000033991.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a green hat holding a child in her arms\"."}], "task": "refering", "answer_template": "The \"a woman in a green hat holding a child in her arms\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 68, 69, 70, 71, 84, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 229, 230, 231, 232, 233, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 301, 302, 303, 304, 305, 319, 320, 321, 322, 323, 336, 337, 338, 339, 340, 341, 354, 355, 356, 357, 358, 359, 369, 370, 371, 372, 373, 374, 375, 376, 377, 386, 387, 388, 389, 390, 391, 392, 393, 394, 406, 407, 408, 409, 410, 411, 412], "bbox": [0.44469667318982387, 0.141578125, 1.0, 0.9820156250000001], "iscrowd": 0, "area": 76883.83744999998, "rle": {"size": [640, 511], "counts": "Y^^41oc03M3M3M3M3M2N0000001O0000001O0O10001O0000001O0000001O0000001O00000O101O0000001O0000001O0000001O0000001O0O10001O0000001O0000001O000000001O000O101O0000001O0000001O0000001O0000001N1000001O0000001O000000001O00000O2M2N2N2N2N2N2N2N2oASNm9o1QFVNj9l1TFZNf9h1YF\\Nb9f1\\F_N_9c1_FbN\\9`1bFeNY9]1fFgNU9[1iFkNQ9W1mFnNn8T1PGQOk8Q1TGSOg8o0YGTOb8n0^GVO\\8l0eGVOV8l0jGWOP8l0PHXOh7l0XHWOb7l0^HXO[7k0fHWOT7l0lHWOn6l0RIa0^5BbJa0X5BhJa0R5BoJ`0k4CUK`0e4C[K`0_4CaK`0Y4CgK`0S4CnK>m3ESL>g3EYL5j3NVL4e3O\\L1`32aLOZ34gLLU37lLJo29RMGj2 in this image.", "objects": [{"patches": [66, 67, 68, 69, 70, 71, 84, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 229, 230, 231, 232, 233, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 301, 302, 303, 304, 305, 319, 320, 321, 322, 323, 336, 337, 338, 339, 340, 341, 354, 355, 356, 357, 358, 359, 369, 370, 371, 372, 373, 374, 375, 376, 377, 386, 387, 388, 389, 390, 391, 392, 393, 394, 406, 407, 408, 409, 410, 411, 412], "bbox": [0.44469667318982387, 0.141578125, 1.0, 0.9820156250000001], "iscrowd": 0, "area": 76883.83744999998, "rle": {"size": [640, 511], "counts": "Y^^41oc03M3M3M3M3M2N0000001O0000001O0O10001O0000001O0000001O0000001O00000O101O0000001O0000001O0000001O0000001O0O10001O0000001O0000001O000000001O000O101O0000001O0000001O0000001O0000001N1000001O0000001O000000001O00000O2M2N2N2N2N2N2N2N2oASNm9o1QFVNj9l1TFZNf9h1YF\\Nb9f1\\F_N_9c1_FbN\\9`1bFeNY9]1fFgNU9[1iFkNQ9W1mFnNn8T1PGQOk8Q1TGSOg8o0YGTOb8n0^GVO\\8l0eGVOV8l0jGWOP8l0PHXOh7l0XHWOb7l0^HXO[7k0fHWOT7l0lHWOn6l0RIa0^5BbJa0X5BhJa0R5BoJ`0k4CUK`0e4C[K`0_4CaK`0Y4CgK`0S4CnK>m3ESL>g3EYL5j3NVL4e3O\\L1`32aLOZ34gLLU37lLJo29RMGj2 in this image.", "objects": [{"patches": [69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143, 163, 164, 165], "bbox": [0.012140624999999999, 0.17325, 0.259546875, 0.43516666666666665], "iscrowd": 0, "area": 12565.565150000002, "rle": {"size": [480, 640], "counts": "_Z41o>0O1O100O1O100O1I7CDVB in this image.", "objects": [{"patches": [69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143, 163, 164, 165], "bbox": [0.012140624999999999, 0.17325, 0.259546875, 0.43516666666666665], "iscrowd": 0, "area": 12565.565150000002, "rle": {"size": [480, 640], "counts": "_Z41o>0O1O100O1O100O1I7CDVB in this image.", "objects": [{"patches": [12, 13, 14, 35, 36, 37, 38, 58, 59, 60, 61, 81, 82, 83, 84, 85, 86, 105, 106, 107], "bbox": [0.523734375, 0.004875, 0.747984375, 0.251875], "iscrowd": 0, "area": 9714.68725, "rle": {"size": [480, 640], "counts": "_Qm4?A3L3N3L4M3M3L4M3M2M6K6J6I7J3L3N2N2M3N2N3L3N2M2O1O1N2O1O0O2O1O1N2O0O]LiD_3^;N3N1N1O1O001O0010O01O001O1O001O00001O001O001O00001O001O00001O001O001O001O1O001N2O001O1O001O1O1O001O1O001O1O0O2O1O1N4M2N3M2M4M2N3L3N3M2N2M4M2N2M3N3M2N2N3M1O0000O10O1000O1000O010000O100O01000O1001O0000001N10002N3L3J6IS\\[2"}, "label": "dark chicken closest to the fence"}]} {"id": 33992, "image": "COCO_train2014_000000033992.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chicken on right\"."}], "task": "refering", "answer_template": "The \"chicken on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 133, 134, 135, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 246, 247, 269, 270, 271], "bbox": [0.543546875, 0.23441666666666666, 0.9758125, 0.6967708333333333], "iscrowd": 0, "area": 29491.589599999996, "rle": {"size": [480, 640], "counts": "hXS51o>1N2O1N101N2O1N2N101N2O1N101N2O1N2O2M2N2O1N2O1O1O1O1N2O001O0O2O001O001N101O001N101O001O0O2O001N2`CdNZ;]1_DmN];U1`DPO\\;Q1cDSOX;o0gDUOU;m0iDWOS;j0lDXOQ;j0nDXOP;i0nDZOP;h0nDZOo:h0PEZOn:g0QE[Om:g0QE[Ol:]201O0O101O0O101O0O2O000O2O000O2O1O0N300O1M20100M201O1NXKkEh4T:201N2O0O2O1O1N101O1N102N2M3N2N1N3N2N2M3N2N2N2[GQJU8Q6eGTJZ8n5`GWJ_8U601O1O1O001O1O1O00fNlGiKS8U4UHeKk7Z4YHdKf7[4]HcKc7[4aHdK^7[4dHdK\\7Z4hHeKW7Z4lHdKT7\\4mHdKR7[4oHeKQ7[4oHfKP7Y4QIgKo6Y4QIhKn6W4SIiKm6W4SIiKm6V4TIjKl6V4UIiKk6V4VIjKj6V4VIiKk6W4UIiKk6V4VIjKj6V4VIjKj6U4WIjKj6V4VIjKj6U4WIkKi6U4XIjKh6U4YI`KcN1T8_4YI`KdNOT8`4YIaKeNMR8b4YIaKfNLQ8b4ZIbKfNKP8c4ZIbKhNIn7d4[IcKo6]4RIbKn6]4SIcKm6]4SIcKm6\\4TIdKm6[4TIdKl6\\4UIcKk6]4UIcKl6\\4UIcKk6]4VIbKk6]4UIcKk6]4VIbKj6^4VIbKk6]4VIbKj6^4WIaKi6_4WIaKj6^4l101O00001O01O01O001O000010O01O00001O01O01O001N1L4M4L3M4M2M3M4M2N2O2M2O3L5L3M3N2M3M3M3N2M3M3M2N2O1N1O2N2O0O2N2O1N1O2N2O0O2O1O00100O001O100O00100O001O10O01O10O01O10O01O00100O00100O001O01N100000000001O2M2F;K4L4L5KQk6"}, "label": "chicken on right"}]} {"id": 33992, "image": "COCO_train2014_000000033992.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chicken to the right of another chicken\"."}], "task": "refering", "answer_template": "The \"a chicken to the right of another chicken\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 133, 134, 135, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 246, 247, 269, 270, 271], "bbox": [0.543546875, 0.23441666666666666, 0.9758125, 0.6967708333333333], "iscrowd": 0, "area": 29491.589599999996, "rle": {"size": [480, 640], "counts": "hXS51o>1N2O1N101N2O1N2N101N2O1N101N2O1N2O2M2N2O1N2O1O1O1O1N2O001O0O2O001O001N101O001N101O001O0O2O001N2`CdNZ;]1_DmN];U1`DPO\\;Q1cDSOX;o0gDUOU;m0iDWOS;j0lDXOQ;j0nDXOP;i0nDZOP;h0nDZOo:h0PEZOn:g0QE[Om:g0QE[Ol:]201O0O101O0O101O0O2O000O2O000O2O1O0N300O1M20100M201O1NXKkEh4T:201N2O0O2O1O1N101O1N102N2M3N2N1N3N2N2M3N2N2N2[GQJU8Q6eGTJZ8n5`GWJ_8U601O1O1O001O1O1O00fNlGiKS8U4UHeKk7Z4YHdKf7[4]HcKc7[4aHdK^7[4dHdK\\7Z4hHeKW7Z4lHdKT7\\4mHdKR7[4oHeKQ7[4oHfKP7Y4QIgKo6Y4QIhKn6W4SIiKm6W4SIiKm6V4TIjKl6V4UIiKk6V4VIjKj6V4VIiKk6W4UIiKk6V4VIjKj6V4VIjKj6U4WIjKj6V4VIjKj6U4WIkKi6U4XIjKh6U4YI`KcN1T8_4YI`KdNOT8`4YIaKeNMR8b4YIaKfNLQ8b4ZIbKfNKP8c4ZIbKhNIn7d4[IcKo6]4RIbKn6]4SIcKm6]4SIcKm6\\4TIdKm6[4TIdKl6\\4UIcKk6]4UIcKl6\\4UIcKk6]4VIbKk6]4UIcKk6]4VIbKj6^4VIbKk6]4VIbKj6^4WIaKi6_4WIaKj6^4l101O00001O01O01O001O000010O01O00001O01O01O001N1L4M4L3M4M2M3M4M2N2O2M2O3L5L3M3N2M3M3M3N2M3M3M2N2O1N1O2N2O0O2N2O1N1O2N2O0O2O1O00100O001O100O00100O001O10O01O10O01O10O01O00100O00100O001O01N100000000001O2M2F;K4L4L5KQk6"}, "label": "a chicken to the right of another chicken"}]} {"id": 33992, "image": "COCO_train2014_000000033992.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown chicken in front of more chickens\"."}], "task": "refering", "answer_template": "The \"the brown chicken in front of more chickens\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 146, 147, 148, 149, 162, 163, 168, 169, 170, 171, 172, 185, 186, 187, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.014484374999999999, 0.3253125, 0.53271875, 0.9884166666666668], "iscrowd": 0, "area": 71388.54305, "rle": {"size": [480, 640], "counts": "ec42k>7J7I7I6J7H7J7I7I6J7RNPN_FW2Z9PN`FT2Z9TN`FP2[9VN`Fn1Z9ZN`Fi1\\9]N^Fh1\\9`N^Fd1]9bN^Fa1IkLa9k1`Fa1[9T2M3M3L4M3M3M3L4M3M3M3L4iN[IPJh6n5\\ImIg6Q6]IiIg6T6_IeIe6Y6_IbId6\\6`I^Id6`6o0N2M3N2N2N2002N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001ON2N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O10000O10000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000000000000000M3G900O1O100O1O10`JjK6U4DYL5f3GfL2Z3IXMJg22mM_OU2?PN\\OR2b0RNZOP2d0TNXOl1h0XNTOh1l0]NoNb1R1bNjN^1V1fNfNZ1Z1jNbNU1_1BjM>V2a0kL_OU3U602M5K4L4L5L3L5K4L4L5J5G:M2N2O2M2N2N3N1Nk0UO100O001O100O1O100O0M4K5Kdm[4"}, "label": "the brown chicken in front of more chickens"}]} {"id": 33992, "image": "COCO_train2014_000000033992.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown chicken in a barnyard is looking at the camera suspiciously\"."}], "task": "refering", "answer_template": "The \"a brown chicken in a barnyard is looking at the camera suspiciously\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 146, 147, 148, 149, 162, 163, 168, 169, 170, 171, 172, 185, 186, 187, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.014484374999999999, 0.3253125, 0.53271875, 0.9884166666666668], "iscrowd": 0, "area": 71388.54305, "rle": {"size": [480, 640], "counts": "ec42k>7J7I7I6J7H7J7I7I6J7RNPN_FW2Z9PN`FT2Z9TN`FP2[9VN`Fn1Z9ZN`Fi1\\9]N^Fh1\\9`N^Fd1]9bN^Fa1IkLa9k1`Fa1[9T2M3M3L4M3M3M3L4M3M3M3L4iN[IPJh6n5\\ImIg6Q6]IiIg6T6_IeIe6Y6_IbId6\\6`I^Id6`6o0N2M3N2N2N2002N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001ON2N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O10000O10000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000000000000000M3G900O1O100O1O10`JjK6U4DYL5f3GfL2Z3IXMJg22mM_OU2?PN\\OR2b0RNZOP2d0TNXOl1h0XNTOh1l0]NoNb1R1bNjN^1V1fNfNZ1Z1jNbNU1_1BjM>V2a0kL_OU3U602M5K4L4L5L3L5K4L4L5J5G:M2N2O2M2N2N3N1Nk0UO100O001O100O1O100O0M4K5Kdm[4"}, "label": "a brown chicken in a barnyard is looking at the camera suspiciously"}]} {"id": 320721, "image": "COCO_train2014_000000320721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle bicycle in a group of three\"."}], "task": "refering", "answer_template": "The \"the middle bicycle in a group of three\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 75, 95, 96, 97, 98, 99, 100, 121, 122, 123, 144, 145, 146, 167, 168, 169, 191, 192, 215, 238, 239, 261, 262, 284, 285, 307, 308, 330, 331], "bbox": [0.138140625, 0.21608899297423886, 0.417578125, 0.9901873536299766], "iscrowd": 0, "area": 10352.003350000003, "rle": {"size": [427, 640], "counts": "^iT1?l<00O20O0000001O000000001O0000001O0000001O000000010O0001O010O000010O01O000010O01O00UCL_<`01O001O0001O01O00000001O000000000010O0000001O01O0001O00010O000001M2M3M3M4K4M3M3M4O000TEaNm9^1SFjNf9V1YFnNc9R1]FQOa9n0_FTO`9k0`FWO^9j0aFXO^9g0bF\\O\\9c0dF_OZ9a0fFAY9?fFCY9@]8f0iGAH70F\\8`0nGYO2`0AL\\89YIIXN3\\82aIIQN9[8LhIJjM?Z8ERJJaMf0Z8^OZJJZMl0Y8YO`JV1\\5kNeJX1W5gNkJ[1R5dNPK^1m4aNVK`1f4aN[Kb1a4]NaKj1W4UNkKQ2m3PNULU2d3jM^L\\2[3cMiLb2n2^MYMc2`2\\MgMe2R2ZMUNn2]1PMkNe39[LMR4AmKc0o3^OPLg0k3YOULk0g3VOWLo0e3ROZLS1a3mN_LW1]3jNbLZ1Z3gNdL`1V3`4J5K6J5K6J5K6J9Gh0XO8G8IUUk4"}, "label": "the middle bicycle in a group of three"}]} {"id": 320721, "image": "COCO_train2014_000000320721.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle bicycle\"."}], "task": "refering", "answer_template": "The \"the middle bicycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 75, 95, 96, 97, 98, 99, 100, 121, 122, 123, 144, 145, 146, 167, 168, 169, 191, 192, 215, 238, 239, 261, 262, 284, 285, 307, 308, 330, 331], "bbox": [0.138140625, 0.21608899297423886, 0.417578125, 0.9901873536299766], "iscrowd": 0, "area": 10352.003350000003, "rle": {"size": [427, 640], "counts": "^iT1?l<00O20O0000001O000000001O0000001O0000001O000000010O0001O010O000010O01O000010O01O00UCL_<`01O001O0001O01O00000001O000000000010O0000001O01O0001O00010O000001M2M3M3M4K4M3M3M4O000TEaNm9^1SFjNf9V1YFnNc9R1]FQOa9n0_FTO`9k0`FWO^9j0aFXO^9g0bF\\O\\9c0dF_OZ9a0fFAY9?fFCY9@]8f0iGAH70F\\8`0nGYO2`0AL\\89YIIXN3\\82aIIQN9[8LhIJjM?Z8ERJJaMf0Z8^OZJJZMl0Y8YO`JV1\\5kNeJX1W5gNkJ[1R5dNPK^1m4aNVK`1f4aN[Kb1a4]NaKj1W4UNkKQ2m3PNULU2d3jM^L\\2[3cMiLb2n2^MYMc2`2\\MgMe2R2ZMUNn2]1PMkNe39[LMR4AmKc0o3^OPLg0k3YOULk0g3VOWLo0e3ROZLS1a3mN_LW1]3jNbLZ1Z3gNdL`1V3`4J5K6J5K6J5K6J9Gh0XO8G8IUUk4"}, "label": "the middle bicycle"}]} {"id": 443604, "image": "COCO_train2014_000000443604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a old man who wearing glass looking a young women\"."}], "task": "refering", "answer_template": "The \"a old man who wearing glass looking a young women\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.596765625, 0.45383529411764706, 1.0, 0.9898588235294118], "iscrowd": 0, "area": 32872.1107, "rle": {"size": [425, 640], "counts": "bnn41W=1O100O1O1O1O1O100O100000000000O100000000000000000O1O100O100O100O100O1O100O100O1O1N2O1O1N2O1O1O1O1O1O1O1O1O1O1O1O100O1O100O100O1O100O100O1O100O100O1O100O100O10000O10000O10000O1000000O10000O1O1O1O1O1O1O1N2O1O1M3N2M3O1O1O1O100O1O100^NPNjGQ2Q8VNlGk1R8YNlGg1R8\\NlGe1]7gMTHg0=c1\\7VObHk0[7YOdHg0Y7\\OfHe0W7_OgHb0Y7_OfHa0Z7@dHa0[7AcH`0]7@bHa0^7@aH`0^7AaH`0_7A_H`0a7@_H`0`7B^H?b7A]H`0b7B]H>c7EYH\\IAd6>]IBd6<^ICc6;^IEb6:_IFb68`IGa68_IHa67aIH`66aIJ`64bIK^63dIM]60eI0\\6MgI2Y6LiI4]6DfI;a6\\OaId0e6oNdIl0PO"}, "label": "a old man who wearing glass looking a young women"}]} {"id": 443604, "image": "COCO_train2014_000000443604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old man is sitting near a woman holding an umbrella\"."}], "task": "refering", "answer_template": "The \"an old man is sitting near a woman holding an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.596765625, 0.45383529411764706, 1.0, 0.9898588235294118], "iscrowd": 0, "area": 32872.1107, "rle": {"size": [425, 640], "counts": "bnn41W=1O100O1O1O1O1O100O100000000000O100000000000000000O1O100O100O100O100O1O100O100O1O1N2O1O1N2O1O1O1O1O1O1O1O1O1O1O1O100O1O100O100O1O100O100O1O100O100O1O100O100O10000O10000O10000O1000000O10000O1O1O1O1O1O1O1N2O1O1M3N2M3O1O1O1O100O1O100^NPNjGQ2Q8VNlGk1R8YNlGg1R8\\NlGe1]7gMTHg0=c1\\7VObHk0[7YOdHg0Y7\\OfHe0W7_OgHb0Y7_OfHa0Z7@dHa0[7AcH`0]7@bHa0^7@aH`0^7AaH`0_7A_H`0a7@_H`0`7B^H?b7A]H`0b7B]H>c7EYH\\IAd6>]IBd6<^ICc6;^IEb6:_IFb68`IGa68_IHa67aIH`66aIJ`64bIK^63dIM]60eI0\\6MgI2Y6LiI4]6DfI;a6\\OaId0e6oNdIl0PO"}, "label": "an old man is sitting near a woman holding an umbrella"}]} {"id": 419028, "image": "COCO_train2014_000000419028.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black ancient greek vase with figures\"."}], "task": "refering", "answer_template": "The \"a black ancient greek vase with figures\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 158, 179, 180, 181, 182, 202, 203, 204, 205, 224, 225, 226, 227, 228, 247, 248, 249, 250, 251, 270, 271, 272, 273, 294, 295, 296, 317, 318, 319, 340, 341, 342, 363, 364, 365], "bbox": [0.763484375, 0.38877083333333334, 0.92696875, 0.9146041666666667], "iscrowd": 0, "area": 19321.554099999998, "rle": {"size": [480, 640], "counts": "i_U7=]>=B=G8I8I6I7I7J3L5K5I6iESMR8U3]GkLnN6f8l3[HoKkN>a8h3bHkKiNg0Z8c3lHiLi6]3UIkL`6Y3_IjL\\6Y3cIgL[6]3cIcL[6`3dI`LY6e3eI[LZ6g3eIYLZ6i3eIWLZ6k3eIULZ6n3dIRL[6P4dIoK]6R4bInK]6T4bIlK^6V4`IjKa6V4^IjKb6X4\\IhKe6X4YIiKh6X4TIjKm6V4QIkKo6W4mHjKU7P61O1O001O1O001O1OO1O1N2O1O01000O100O10000O100O100000000003M4L3PNdHVL\\7h3hHVLX7j3iHULV7k3mHSLS7l3PIQLP7o3TInKk6R4YIkKf6U4_IgK`6Y4dIdK[6[4kI`KV6_4oI]KP6c4TJZKl5e4VJZKj5d4YJ[Kg5d4[JZKe5f4^JXKb5g4`JXK`5g4bJXK^5f4eJYK[5f4iJSK[5k4^2N3M2N2K5K8I7I7J6I7J6J7H7iMXDm0n;iNcDi0d;nNfDi0o in this image.", "objects": [{"patches": [156, 157, 158, 179, 180, 181, 182, 202, 203, 204, 205, 224, 225, 226, 227, 228, 247, 248, 249, 250, 251, 270, 271, 272, 273, 294, 295, 296, 317, 318, 319, 340, 341, 342, 363, 364, 365], "bbox": [0.763484375, 0.38877083333333334, 0.92696875, 0.9146041666666667], "iscrowd": 0, "area": 19321.554099999998, "rle": {"size": [480, 640], "counts": "i_U7=]>=B=G8I8I6I7I7J3L5K5I6iESMR8U3]GkLnN6f8l3[HoKkN>a8h3bHkKiNg0Z8c3lHiLi6]3UIkL`6Y3_IjL\\6Y3cIgL[6]3cIcL[6`3dI`LY6e3eI[LZ6g3eIYLZ6i3eIWLZ6k3eIULZ6n3dIRL[6P4dIoK]6R4bInK]6T4bIlK^6V4`IjKa6V4^IjKb6X4\\IhKe6X4YIiKh6X4TIjKm6V4QIkKo6W4mHjKU7P61O1O001O1O001O1OO1O1N2O1O01000O100O10000O100O100000000003M4L3PNdHVL\\7h3hHVLX7j3iHULV7k3mHSLS7l3PIQLP7o3TInKk6R4YIkKf6U4_IgK`6Y4dIdK[6[4kI`KV6_4oI]KP6c4TJZKl5e4VJZKj5d4YJ[Kg5d4[JZKe5f4^JXKb5g4`JXK`5g4bJXK^5f4eJYK[5f4iJSK[5k4^2N3M2N2K5K8I7I7J6I7J6J7H7iMXDm0n;iNcDi0d;nNfDi0o in this image.", "objects": [{"patches": [138, 139, 141, 142, 143, 161, 162, 163, 164, 165, 166, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 254, 255, 256, 257, 277, 278, 279, 280, 301, 302, 303, 323, 324, 325, 326, 346, 347, 348, 349], "bbox": [0.035390624999999995, 0.3701041666666667, 0.22921874999999997, 0.9004375000000001], "iscrowd": 0, "area": 22109.659000000003, "rle": {"size": [480, 640], "counts": "nn:h0W>5dCAn9d0lECn9a0mEBQ:c0hE@W:e0cE]O\\:h0^EZOa:c2O1O1O1O1O8I7I7I7I7J6I4L1O2O0O1O2ZHTKT5m4iJUKX5k4fJVKZ5k4dJVK\\5l4`JVK`5k4^JVKb5k4[JWKf5i4XJXKh5j4UJWKo5f4nI\\K[6\\4cIeKc6W4YIkKg6V4WIkKi6V4UIkKk6V4RIlKn6U4PIlKP7U4mHnKR7R4mHoKS7R4kHoKU7Q4jHPLV7Q4jHnKU7S4kHmKU7T4jHlKV7U4iHkKW7U4iHkKW7V4iHiKW7W4iHiKW7X4hHhKX7Y4gHgKY7Y4hHfKX7[4gHeKY7\\4fHdKY7^4fHbKZ7_4\\HjKd7i500000000000000O1000000000001O0000000OUN`HXL`7h3mHkKS7T4nHlKR7S4PIlKP7S4QImKo6S4QImKf6[4[IeKc6\\4_IcK_6^4bIbK\\6`4dI`K[6`4fI`KZ6_4gIaKY6^4hIbKX6^4hIbKX6]4iIcKW6\\4jIdKV6[4kIeKU6Z4lIfKT6Z4lIdKV6[4kIcKW6\\4jIbKX6]4iIaKY6_4gI_K[6`4fI^K\\6`4fI^K\\6`4fI_K\\6]4gIaK[6]4fIbK]6[4eIcK]6Z4fIdK]6Y4eIeK]6X4fIfK]6W4eIgK]6W4X2L3hMbEDi:7\\E[OQ;a0TEWOT;e0PEYOQ;c0TE[On:`0WE^Oj:>ZE@i:<[EBf::`ECb:8iE@^:6lEC[:4oEDX:2lnV7"}, "label": "a gold , intricate vase to the left of three other vases"}]} {"id": 419028, "image": "COCO_train2014_000000419028.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a golden piece of pottery\"."}], "task": "refering", "answer_template": "The \"a golden piece of pottery\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 141, 142, 143, 161, 162, 163, 164, 165, 166, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 254, 255, 256, 257, 277, 278, 279, 280, 301, 302, 303, 323, 324, 325, 326, 346, 347, 348, 349], "bbox": [0.035390624999999995, 0.3701041666666667, 0.22921874999999997, 0.9004375000000001], "iscrowd": 0, "area": 22109.659000000003, "rle": {"size": [480, 640], "counts": "nn:h0W>5dCAn9d0lECn9a0mEBQ:c0hE@W:e0cE]O\\:h0^EZOa:c2O1O1O1O1O8I7I7I7I7J6I4L1O2O0O1O2ZHTKT5m4iJUKX5k4fJVKZ5k4dJVK\\5l4`JVK`5k4^JVKb5k4[JWKf5i4XJXKh5j4UJWKo5f4nI\\K[6\\4cIeKc6W4YIkKg6V4WIkKi6V4UIkKk6V4RIlKn6U4PIlKP7U4mHnKR7R4mHoKS7R4kHoKU7Q4jHPLV7Q4jHnKU7S4kHmKU7T4jHlKV7U4iHkKW7U4iHkKW7V4iHiKW7W4iHiKW7X4hHhKX7Y4gHgKY7Y4hHfKX7[4gHeKY7\\4fHdKY7^4fHbKZ7_4\\HjKd7i500000000000000O1000000000001O0000000OUN`HXL`7h3mHkKS7T4nHlKR7S4PIlKP7S4QImKo6S4QImKf6[4[IeKc6\\4_IcK_6^4bIbK\\6`4dI`K[6`4fI`KZ6_4gIaKY6^4hIbKX6^4hIbKX6]4iIcKW6\\4jIdKV6[4kIeKU6Z4lIfKT6Z4lIdKV6[4kIcKW6\\4jIbKX6]4iIaKY6_4gI_K[6`4fI^K\\6`4fI^K\\6`4fI_K\\6]4gIaK[6]4fIbK]6[4eIcK]6Z4fIdK]6Y4eIeK]6X4fIfK]6W4eIgK]6W4X2L3hMbEDi:7\\E[OQ;a0TEWOT;e0PEYOQ;c0TE[On:`0WE^Oj:>ZE@i:<[EBf::`ECb:8iE@^:6lEC[:4oEDX:2lnV7"}, "label": "a golden piece of pottery"}]} {"id": 566488, "image": "COCO_train2014_000000566488.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bench the skateboarder is sitting on\"."}], "task": "refering", "answer_template": "The \"a bench the skateboarder is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 171, 172, 181, 182, 187, 188, 189, 190, 198, 199, 204, 205, 206, 207, 215, 216, 221, 222, 223, 224, 232, 233, 241, 249, 250, 258], "bbox": [0.01240343347639485, 0.43059375, 0.747145922746781, 0.6901875], "iscrowd": 0, "area": 9967.87785, "rle": {"size": [640, 466], "counts": "nQ4o0Pc05K4L5K5K5K1O100000000O100000000O1000000O100000000O100000000O1000000O100000000O100000000O100000000O1000000O2O0000000O100O100O2O2M3M3N2M4M2M3M3N2M4M24L:F:F5d0YOM3N2N2M3N2M3N3L:G9DK]ALd>0hA]Oc>?a2L4L`_Y2"}, "label": "a bench the skateboarder is sitting on"}]} {"id": 50410, "image": "COCO_train2014_000000050410.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white surfboard laying in the sand with black straps on it\"."}], "task": "refering", "answer_template": "The \"white surfboard laying in the sand with black straps on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 150, 160, 161, 162, 163, 172, 173, 174, 175, 176, 185, 186, 187, 188, 197, 198, 199, 200, 201, 210, 211, 212, 213, 223, 224, 225, 226, 236, 237, 238, 249, 250], "bbox": [0.15908333333333335, 0.50484375, 0.5687222222222222, 0.849359375], "iscrowd": 0, "area": 17377.601099999993, "rle": {"size": [640, 360], "counts": "VRT12]1:g`0Ia^O`1k`0dNg^Om1Ra0a0O2N1O1O2N1O1O2N1N2O2M2O1N3M2O1N3M101N2O1N2N2O1N2N2O1N2N101N2N2O0O2N100O1O100O101N1O100O1O100O2N100O1O10O01O10O01O10O0100000O01000O1000O0100000O010000O01000O2O00000O101N100O101N100O101N2O1N101N2O1N2O1N2N1O2M3N2N2M3N2N2M2O2N2M3N3M2M3N2N2M3N2N3L3N2N2M3N2N2M4M2N2N2M3N3M2M4L3M3M4K4M4L3L7J8H8G]aP3"}, "label": "white surfboard laying in the sand with black straps on it"}]} {"id": 50410, "image": "COCO_train2014_000000050410.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white surfboard with a greenish - yellow striped design at the end\"."}], "task": "refering", "answer_template": "The \"a white surfboard with a greenish - yellow striped design at the end\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 150, 160, 161, 162, 163, 172, 173, 174, 175, 176, 185, 186, 187, 188, 197, 198, 199, 200, 201, 210, 211, 212, 213, 223, 224, 225, 226, 236, 237, 238, 249, 250], "bbox": [0.15908333333333335, 0.50484375, 0.5687222222222222, 0.849359375], "iscrowd": 0, "area": 17377.601099999993, "rle": {"size": [640, 360], "counts": "VRT12]1:g`0Ia^O`1k`0dNg^Om1Ra0a0O2N1O1O2N1O1O2N1N2O2M2O1N3M2O1N3M101N2O1N2N2O1N2N2O1N2N101N2N2O0O2N100O1O100O101N1O100O1O100O2N100O1O10O01O10O01O10O0100000O01000O1000O0100000O010000O01000O2O00000O101N100O101N100O101N2O1N101N2O1N2O1N2N1O2M3N2N2M3N2N2M2O2N2M3N3M2M3N2N2M3N2N3L3N2N2M3N2N2M4M2N2N2M3N3M2M4L3M3M4K4M4L3L7J8H8G]aP3"}, "label": "a white surfboard with a greenish - yellow striped design at the end"}]} {"id": 1261, "image": "COCO_train2014_000000001261.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass bowl of food that is sitting on a whicker circle holder\"."}], "task": "refering", "answer_template": "The \"a glass bowl of food that is sitting on a whicker circle holder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 85, 86, 87, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 361, 362, 363], "bbox": [0.434828125, 0.20831249999999998, 1.0, 0.9049375000000001], "iscrowd": 0, "area": 95788.1471, "rle": {"size": [480, 640], "counts": "UaR43j>9F9G:F:F9H9F:F9G:F9H9F:F8H6M2N3L3N3L3N3L3N3M3L3N3L3N3M2M4M2M4M2N3L3N3L3N3M3L3N3M2N3N1N2N2O1N101N2O1N2O1N101N2N2O1N101N2O1N2O1N101N2O1N2N101N2O1N2O1N101N2O1N2N101N2O1N2O1N101N2O1N2O0O2N2O1O1N2O001O1O1O1O001N2O1O1O1O001O1O1O1N101O1O0000001O0O10001O0000001O000O101O0000001O000O2O0000001O0000001N1000001O0000001O000O101O0000001O0000001O000O101O0000001O00001O00000O2O0000001O0000010O00000010O0000010O0000010O0000010O000001O01O0001O01O0000001N1000001N1000000O2O000O101O000O101O000O101O0O10001O0O10001O0O10001N1000001N10001O0O10001N1000001N2O1O001N2O1O1N101O1O1N2O001O1N2O1O0O2O1O1O1N2O001O1N2O1O001N2O1O1N101O1O1N2O001O1N2O1O1N101O1O1N2O001O1N2O1O0O2O1O1O1N101N2M3N2N1O2N2N2N2M2O2N2N2N2N1O2M3N3M2N2N3M2M4M2N3M2N3M2N3L3N3M2N3M2NbI"}, "label": "a glass bowl of food that is sitting on a whicker circle holder"}]} {"id": 1261, "image": "COCO_train2014_000000001261.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bowl on right\"."}], "task": "refering", "answer_template": "The \"bowl on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 85, 86, 87, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 361, 362, 363], "bbox": [0.434828125, 0.20831249999999998, 1.0, 0.9049375000000001], "iscrowd": 0, "area": 95788.1471, "rle": {"size": [480, 640], "counts": "UaR43j>9F9G:F:F9H9F:F9G:F9H9F:F8H6M2N3L3N3L3N3L3N3M3L3N3L3N3M2M4M2M4M2N3L3N3L3N3M3L3N3M2N3N1N2N2O1N101N2O1N2O1N101N2N2O1N101N2O1N2O1N101N2O1N2N101N2O1N2O1N101N2O1N2N101N2O1N2O1N101N2O1N2O0O2N2O1O1N2O001O1O1O1O001N2O1O1O1O001O1O1O1N101O1O0000001O0O10001O0000001O000O101O0000001O000O2O0000001O0000001N1000001O0000001O000O101O0000001O0000001O000O101O0000001O00001O00000O2O0000001O0000010O00000010O0000010O0000010O0000010O000001O01O0001O01O0000001N1000001N1000000O2O000O101O000O101O000O101O0O10001O0O10001O0O10001N1000001N10001O0O10001N1000001N2O1O001N2O1O1N101O1O1N2O001O1N2O1O0O2O1O1O1N2O001O1N2O1O001N2O1O1N101O1O1N2O001O1N2O1O1N101O1O1N2O001O1N2O1O0O2O1O1O1N101N2M3N2N1O2N2N2N2M2O2N2N2N2N1O2M3N3M2N2N3M2M4M2N3M2N3M2N3L3N3M2N3M2NbI"}, "label": "bowl on right"}]} {"id": 1261, "image": "COCO_train2014_000000001261.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plate with fruit on it behind two other dishes\"."}], "task": "refering", "answer_template": "The \"a plate with fruit on it behind two other dishes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 34, 35, 36, 37, 52, 53, 54, 59, 60, 61, 74, 75, 84, 97, 98, 99, 121, 122, 123, 146, 147, 148], "bbox": [0.251125, 0.0516875, 0.687640625, 0.38202083333333337], "iscrowd": 0, "area": 6863.3622, "rle": {"size": [480, 640], "counts": "Rb[2`0^>;F4K4M4K5K5L4K4M2M3N2M3M3N1N3N2M3MfNbC7\\hGBW8?jGAU8?kGAU8?lG@S8a0nG_OQ8a0oG_OP8b0QH]Oo7c0QH^On7c0RH\\Om7e0THZOm7e0SH[Oo7c0RH\\OP8b0PH^OR8`0nGAT8 in this image.", "objects": [{"patches": [30, 31, 34, 35, 36, 37, 52, 53, 54, 59, 60, 61, 74, 75, 84, 97, 98, 99, 121, 122, 123, 146, 147, 148], "bbox": [0.251125, 0.0516875, 0.687640625, 0.38202083333333337], "iscrowd": 0, "area": 6863.3622, "rle": {"size": [480, 640], "counts": "Rb[2`0^>;F4K4M4K5K5L4K4M2M3N2M3M3N1N3N2M3MfNbC7\\hGBW8?jGAU8?kGAU8?lG@S8a0nG_OQ8a0oG_OP8b0QH]Oo7c0QH^On7c0RH\\Om7e0THZOm7e0SH[Oo7c0RH\\OP8b0PH^OR8`0nGAT8 in this image.", "objects": [{"patches": [186, 208, 209, 231, 232, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 302, 303, 325, 326], "bbox": [0.009765625, 0.5900702576112412, 0.202625, 0.9980327868852459], "iscrowd": 0, "area": 8364.551399999998, "rle": {"size": [427, 640], "counts": "ak23S=;E;E;F:E8H4L4N2O10000O100O100O10WOXEPOg:R1_EgNa:Z1f02O100O4L7I7J6I3M000010O0001O000_OfMSFZ2e9PNXFP2b9YN[Fg1_9bN^F_1a9dN]F[1d9gNYFY1g9V10O00100O0010O01O010O0010O01O101N1O2O0O2N101VNnE8S:FPF8R:FoE:Q:EQF9Q:DRF:o9ESF;m9CUF=j9BXF?g9^O\\Fc0d9ZO^Fg0f9RO\\Fn0h9kN[FV1i9cNYF^1j9\\NXFd1h:1O1O3M5K5K3L10001O000002N4L4M3L4L2N1O100O1O1O1O1O100O2O11OO2L3M3M4L3M3M4L3M3M4L3M3M2N1O2N\\ed6"}, "label": "man wearing gray wetsuit carrying a yellow surfboard watching another man on waves"}]} {"id": 206062, "image": "COCO_train2014_000000206062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man holding a surf board in hand\"."}], "task": "refering", "answer_template": "The \"man holding a surf board in hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 208, 209, 231, 232, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 302, 303, 325, 326], "bbox": [0.009765625, 0.5900702576112412, 0.202625, 0.9980327868852459], "iscrowd": 0, "area": 8364.551399999998, "rle": {"size": [427, 640], "counts": "ak23S=;E;E;F:E8H4L4N2O10000O100O100O10WOXEPOg:R1_EgNa:Z1f02O100O4L7I7J6I3M000010O0001O000_OfMSFZ2e9PNXFP2b9YN[Fg1_9bN^F_1a9dN]F[1d9gNYFY1g9V10O00100O0010O01O010O0010O01O101N1O2O0O2N101VNnE8S:FPF8R:FoE:Q:EQF9Q:DRF:o9ESF;m9CUF=j9BXF?g9^O\\Fc0d9ZO^Fg0f9RO\\Fn0h9kN[FV1i9cNYF^1j9\\NXFd1h:1O1O3M5K5K3L10001O000002N4L4M3L4L2N1O100O1O1O1O1O100O2O11OO2L3M3M4L3M3M4L3M3M4L3M3M2N1O2N\\ed6"}, "label": "man holding a surf board in hand"}]} {"id": 206062, "image": "COCO_train2014_000000206062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow surfboard with swell written on it\"."}], "task": "refering", "answer_template": "The \"a yellow surfboard with swell written on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [256, 257, 258, 259, 260, 261, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328], "bbox": [0.0015, 0.7272833723653396, 0.380609375, 0.9867915690866511], "iscrowd": 0, "area": 15251.647100000002, "rle": {"size": [427, 640], "counts": "oi03W=7I7I2N1O1O1O100O1O1O1O1O2N1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100003M4L4L3M4L4L3M4L3M1O2N2N1O2N1O2N2NTN]OQGb0o8APG=o8FPG9P9JoF4Q9NoF0P94nFKR97nFGQ9=mFBS9`0mF^OS9e0lFYOS9j0lFUOT9n0kFPOU9R1jFmNU9W1^12N3L3N3M2M4M2N3L3N3M3L3N3M2M4M2N3L3N1O0O10000O10000O10000O10000O10000O10000O10000O10000O10000O100O10000O10000000000O1000000000O10O10000000O1O1O010O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O100O1O1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2M2N3M3L3N3M3M3L3N3M3M2M4M3M3M2NfiT5"}, "label": "a yellow surfboard with swell written on it"}]} {"id": 206062, "image": "COCO_train2014_000000206062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow surfboard being held by a man\"."}], "task": "refering", "answer_template": "The \"a yellow surfboard being held by a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [256, 257, 258, 259, 260, 261, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328], "bbox": [0.0015, 0.7272833723653396, 0.380609375, 0.9867915690866511], "iscrowd": 0, "area": 15251.647100000002, "rle": {"size": [427, 640], "counts": "oi03W=7I7I2N1O1O1O100O1O1O1O1O2N1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100003M4L4L3M4L4L3M4L3M1O2N2N1O2N1O2N2NTN]OQGb0o8APG=o8FPG9P9JoF4Q9NoF0P94nFKR97nFGQ9=mFBS9`0mF^OS9e0lFYOS9j0lFUOT9n0kFPOU9R1jFmNU9W1^12N3L3N3M2M4M2N3L3N3M3L3N3M2M4M2N3L3N1O0O10000O10000O10000O10000O10000O10000O10000O10000O10000O100O10000O10000000000O1000000000O10O10000000O1O1O010O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O100O1O1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2M2N3M3L3N3M3M3L3N3M3M2M4M3M3M2NfiT5"}, "label": "a yellow surfboard being held by a man"}]} {"id": 541938, "image": "COCO_train2014_000000541938.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bottle\"."}], "task": "refering", "answer_template": "The \"a bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 251, 252], "bbox": [0.886078125, 0.2606791569086651, 1.0, 0.7123653395784544], "iscrowd": 0, "area": 9612.943649999996, "rle": {"size": [427, 640], "counts": "ae\\74Q=7I8H7I7I7H8I8H7I7I7H8I8H7I7I7I7H9H7I7J6K5K6L3O10000O2O0001O101M3N1N100O1O100O100O010O100O1O100O10O0100O100O1O100O10O0100O100O1O100O100O100O1O10lJ"}, "label": "a bottle"}]} {"id": 541938, "image": "COCO_train2014_000000541938.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue and whit spotted jar holding butter on the right\"."}], "task": "refering", "answer_template": "The \"a blue and whit spotted jar holding butter on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 251, 252], "bbox": [0.886078125, 0.2606791569086651, 1.0, 0.7123653395784544], "iscrowd": 0, "area": 9612.943649999996, "rle": {"size": [427, 640], "counts": "ae\\74Q=7I8H7I7I7H8I8H7I7I7H8I8H7I7I7I7H9H7I7J6K5K6L3O10000O2O0001O101M3N1N100O1O100O100O010O100O1O100O10O0100O100O1O100O10O0100O100O1O100O100O100O1O10lJ"}, "label": "a blue and whit spotted jar holding butter on the right"}]} {"id": 541938, "image": "COCO_train2014_000000541938.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white cup of chocolate with four streaks falling from its lip\"."}], "task": "refering", "answer_template": "The \"a white cup of chocolate with four streaks falling from its lip\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 98, 99, 100, 101, 102, 103, 104, 120, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 281, 282, 283, 284, 285, 286, 287, 288, 289, 304, 305, 306, 307, 308, 309, 310, 311, 312, 327, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.17584375000000002, 0.2586416861826698, 0.60340625, 1.0], "iscrowd": 0, "area": 71276.86265, "rle": {"size": [427, 640], "counts": "fZ_19P=6J6J6J6J6J6J6J6J6J6J6I7H8I7I7J6J6J7I6J6K5J6J6J6J6J6J6K5J6J6J6J6J6J6L4K5L5J5K5L4K5L4K5L4K4M3L4M3L4M3L4M0O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100000000000000O10000000000000000000000000000O10000000000000000000000000000O100000000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O1001O1O001O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O001O1O001O1O001O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001H8B>A?B=C>A?B>B>A>C>Aa0@a0A?H8H8G9H9G8H8G9H8H8GjfY3"}, "label": "a white cup of chocolate with four streaks falling from its lip"}]} {"id": 541938, "image": "COCO_train2014_000000541938.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chocolate in a white mug\"."}], "task": "refering", "answer_template": "The \"chocolate in a white mug\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 98, 99, 100, 101, 102, 103, 104, 120, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 281, 282, 283, 284, 285, 286, 287, 288, 289, 304, 305, 306, 307, 308, 309, 310, 311, 312, 327, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.17584375000000002, 0.2586416861826698, 0.60340625, 1.0], "iscrowd": 0, "area": 71276.86265, "rle": {"size": [427, 640], "counts": "fZ_19P=6J6J6J6J6J6J6J6J6J6J6I7H8I7I7J6J6J7I6J6K5J6J6J6J6J6J6K5J6J6J6J6J6J6L4K5L5J5K5L4K5L4K5L4K4M3L4M3L4M3L4M0O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100000000000000O10000000000000000000000000000O10000000000000000000000000000O100000000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O1001O1O001O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O001O1O001O1O001O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001H8B>A?B=C>A?B>B>A>C>Aa0@a0A?H8H8G9H9G8H8G9H8H8GjfY3"}, "label": "chocolate in a white mug"}]} {"id": 541938, "image": "COCO_train2014_000000541938.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cup on the right in the right hand picture\"."}], "task": "refering", "answer_template": "The \"cup on the right in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 17, 18, 33, 34, 35, 36, 37, 38, 39, 40, 56, 57, 58, 59, 60, 61, 62, 63, 79, 80, 81, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 177, 197, 198, 199, 200, 220, 221, 222, 223, 243, 244, 245], "bbox": [0.44829687500000004, 0.001873536299765808, 0.7961250000000001, 0.7277049180327869], "iscrowd": 0, "area": 43181.66849999999, "rle": {"size": [427, 640], "counts": "ieg3P2Y;\\1eN2O001O001O00001O001O001O001O00001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O2N1O1jKYKFh4UO\\Lk0f3PO\\LQ1f3iN^LW1g3aN\\L_1h3ZN[Lf1j3RNYLn1l3jMWLV2m3cMVL]2o3ZMTLg2Q4QMRLo2S4iLPLW3T4bLoK^3V4ZLmKf3l6000O1000000000000O10000000000O1000000000000O1000000O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O101N1O100O100O1O100O100O1O100O1N2N2N2M3N2N2M3N2K5K5K5J6K5K5K5K5J6K5K5K5J6K5K5K5J6K5J6K5J6J6K5J6K5J6J6K5J6K5J6J6K5J6K5J6F:A?J6M3N2N2M3N6J7Ihcf1"}, "label": "cup on the right in the right hand picture"}]} {"id": 541938, "image": "COCO_train2014_000000541938.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pot with lighter chocolate\"."}], "task": "refering", "answer_template": "The \"pot with lighter chocolate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 17, 18, 33, 34, 35, 36, 37, 38, 39, 40, 56, 57, 58, 59, 60, 61, 62, 63, 79, 80, 81, 82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 177, 197, 198, 199, 200, 220, 221, 222, 223, 243, 244, 245], "bbox": [0.44829687500000004, 0.001873536299765808, 0.7961250000000001, 0.7277049180327869], "iscrowd": 0, "area": 43181.66849999999, "rle": {"size": [427, 640], "counts": "ieg3P2Y;\\1eN2O001O001O00001O001O001O001O00001O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O2N1O1jKYKFh4UO\\Lk0f3PO\\LQ1f3iN^LW1g3aN\\L_1h3ZN[Lf1j3RNYLn1l3jMWLV2m3cMVL]2o3ZMTLg2Q4QMRLo2S4iLPLW3T4bLoK^3V4ZLmKf3l6000O1000000000000O10000000000O1000000000000O1000000O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O101N1O100O100O1O100O100O1O100O1N2N2N2M3N2N2M3N2K5K5K5J6K5K5K5K5J6K5K5K5J6K5K5K5J6K5J6K5J6J6K5J6K5J6J6K5J6K5J6J6K5J6K5J6F:A?J6M3N2N2M3N6J7Ihcf1"}, "label": "pot with lighter chocolate"}]} {"id": 468219, "image": "COCO_train2014_000000468219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"young man getting his tie tied\"."}], "task": "refering", "answer_template": "The \"young man getting his tie tied\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 99, 100, 101, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 167, 168, 169, 170, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 286, 287, 305, 306, 307, 308, 329, 330, 331, 352, 353, 354, 355], "bbox": [0.145640625, 0.2359641255605381, 0.49173437499999995, 0.9887668161434978], "iscrowd": 0, "area": 35689.25639999999, "rle": {"size": [446, 640], "counts": "TiX11k=5L4K5K5L4K5L4K6J5L4K5M3L4M3L3N0O2O0O2O0O2O0O2O0O2N100O2O0O2O0O2O0O2O0O2O0O2O0O101N1O2O0O2O0O2O0O2O0O2O0O101N101N101N1O2O0O2O0O2O0O2O0O101N101N101N101N1O2N1O2N1N2O2N1O2N1N3N1O2N1N3N1bN_K^Ic4]6cK`I^4\\6hKaIY4Z6nKbIU4Y6PLeIQ4W6ULfIm3U6YLhIh3T6^LiId3R6bLjIh3j5]LTJl3`5ZL]Jo3W5WLfJR4Q5QLkJX4m4e2G9H8G8H1O1O1O1O100O1O1O1O1O1O100O1O1O1000000000000000000000kKdLgM\\3Y2iLcMW3\\2nL`MR3_2RM^Mn2a2WM[Mi2d2[MYMe2f2^MXMb2g2aMWM_2h2eMUM[2j2hMTMX2k2kMSMU2l2nMRMR2m2RNPMn1o2UNoLk1Q3WNmLi1R3[NkLe1V3\\NhLc1Z3^NdLb1]3_NaLa1a3`N\\L`1e3aNYL_1h3bNVL_1j3cNSL^1k3eNSL\\1k3gNSLZ1j3kNSLV1j3nNTLS1j3POTLQ1i3SOULn0i3VOTLk0i3YOULg0j3\\OTLe0j3_OSLb0l3@RLa0m3AQL`0n3BPL?o3DnK=P4EoKm0@Z6OiHa0n0_O\\6NeHd0o0]OQ8d0nG\\OR8d0oG[OQ8f0d20O0013M6I7J6J^V]4"}, "label": "young man getting his tie tied"}]} {"id": 468219, "image": "COCO_train2014_000000468219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a suit getting his tie fitted for a special event\"."}], "task": "refering", "answer_template": "The \"a man in a suit getting his tie fitted for a special event\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 99, 100, 101, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 167, 168, 169, 170, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 286, 287, 305, 306, 307, 308, 329, 330, 331, 352, 353, 354, 355], "bbox": [0.145640625, 0.2359641255605381, 0.49173437499999995, 0.9887668161434978], "iscrowd": 0, "area": 35689.25639999999, "rle": {"size": [446, 640], "counts": "TiX11k=5L4K5K5L4K5L4K6J5L4K5M3L4M3L3N0O2O0O2O0O2O0O2O0O2N100O2O0O2O0O2O0O2O0O2O0O2O0O101N1O2O0O2O0O2O0O2O0O2O0O101N101N101N1O2O0O2O0O2O0O2O0O101N101N101N101N1O2N1O2N1N2O2N1O2N1N3N1O2N1N3N1bN_K^Ic4]6cK`I^4\\6hKaIY4Z6nKbIU4Y6PLeIQ4W6ULfIm3U6YLhIh3T6^LiId3R6bLjIh3j5]LTJl3`5ZL]Jo3W5WLfJR4Q5QLkJX4m4e2G9H8G8H1O1O1O1O100O1O1O1O1O1O100O1O1O1000000000000000000000kKdLgM\\3Y2iLcMW3\\2nL`MR3_2RM^Mn2a2WM[Mi2d2[MYMe2f2^MXMb2g2aMWM_2h2eMUM[2j2hMTMX2k2kMSMU2l2nMRMR2m2RNPMn1o2UNoLk1Q3WNmLi1R3[NkLe1V3\\NhLc1Z3^NdLb1]3_NaLa1a3`N\\L`1e3aNYL_1h3bNVL_1j3cNSL^1k3eNSL\\1k3gNSLZ1j3kNSLV1j3nNTLS1j3POTLQ1i3SOULn0i3VOTLk0i3YOULg0j3\\OTLe0j3_OSLb0l3@RLa0m3AQL`0n3BPL?o3DnK=P4EoKm0@Z6OiHa0n0_O\\6NeHd0o0]OQ8d0nG\\OR8d0oG[OQ8f0d20O0013M6I7J6J^V]4"}, "label": "a man in a suit getting his tie fitted for a special event"}]} {"id": 468219, "image": "COCO_train2014_000000468219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man without his face showing facing two men putting on a tie\"."}], "task": "refering", "answer_template": "The \"a man without his face showing facing two men putting on a tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 165, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 351, 352], "bbox": [0.0, 0.2854035874439462, 0.311640625, 0.9842600896860987], "iscrowd": 0, "area": 39504.49255, "rle": {"size": [446, 640], "counts": "c4R9l40O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100001O00000000001O0000000010O0000000001O00000000001O00000000001O000000001O00000000001O000000001O1O2N2N2N1O2N2N1O2N2N1O2N2]KlJ8U5DQK9Q5AUK=m4^OYK?h4\\O^Kb0d4XOcKe0`4UOeKi0_4oNhKn0\\4jNjKT1Z4eNlKX1Y4_NmK_1W4VNSLg1Q4fM`LX2d3YMiLe2\\3SMiLk2[3oLiLo2[3jLiLU3R7N001O001O1O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O00100O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O1O001O001O001O003Md0\\Ob0^O6J6J6Jldo5"}, "label": "a man without his face showing facing two men putting on a tie"}]} {"id": 468219, "image": "COCO_train2014_000000468219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a man to the left of two other men\"."}], "task": "refering", "answer_template": "The \"the back of a man to the left of two other men\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 165, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 351, 352], "bbox": [0.0, 0.2854035874439462, 0.311640625, 0.9842600896860987], "iscrowd": 0, "area": 39504.49255, "rle": {"size": [446, 640], "counts": "c4R9l40O1O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100001O00000000001O0000000010O0000000001O00000000001O00000000001O000000001O00000000001O000000001O1O2N2N2N1O2N2N1O2N2N1O2N2]KlJ8U5DQK9Q5AUK=m4^OYK?h4\\O^Kb0d4XOcKe0`4UOeKi0_4oNhKn0\\4jNjKT1Z4eNlKX1Y4_NmK_1W4VNSLg1Q4fM`LX2d3YMiLe2\\3SMiLk2[3oLiLo2[3jLiLU3R7N001O001O1O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O00100O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O1O001O001O001O003Md0\\Ob0^O6J6J6Jldo5"}, "label": "the back of a man to the left of two other men"}]} {"id": 468219, "image": "COCO_train2014_000000468219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall man fixing another man ' s tie\"."}], "task": "refering", "answer_template": "The \"a tall man fixing another man ' s tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 226, 227, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364], "bbox": [0.35471875, 0.08219730941704036, 0.875796875, 0.9898654708520179], "iscrowd": 0, "area": 77266.66940000001, "rle": {"size": [446, 640], "counts": "\\US32k=5K4L4L5K4L4L4L5K4L4L4L3M3O11O1O10N1N2M3M3M3M3M3M3M4L3M3N2M3M3M3M3M3M4L30O7I7I7I7I5K0000O010^NfM_GZ2_8jM_GV2`8mM]GT2b8oM\\GQ2c8RN[Gn1c8VN[Gj1d8YNYGg1g8\\NWGd1h8_NVGa1h8cNVG]1i8fNTG[1k8hNSGX1l8kNRGU1l8nNSGR1l8nNTGR1l8oNTGQ1k8oNVGQ1i8oNXGQ1h8nNXGS1h8mNXGS1g8mNZGS1f8lN[GT1d8lN\\GT1e8lN[GT1d8lN]GT1c8kN^GU1b8jN^GW1a8jN_GV1a8iN`GW1_8iNaGX1_8gNbGX1^8iNbGW1^8hNcGX1]8gNcGZ1\\8fNeGZ1[8fNeGZ1Z8fNgGZ1Y8eNgG\\1Y8cNhG]1X8cNeG_1\\8`NbGb1^8^N`Gd1a8[N]Gg1d8YNYGi1h8VNVGl1k8SNSGo1n8PNPGR2Q9nMlFT2U9kMiFW2X9hMfFZ2[9eMdF\\2]9dM`F^2a9aM]Fa2c9m01O1O1O1O1O1O1O1OG9E;F:E;F:F:N2O1O1O1O1O1O1O1SKkJP1V5oNmJo0S5oNQKn0P5QORKm0o4QOUKl0l4SOWKj0j4TOYKj0h4UO[KUN]Ni1Y61]KoM_No1U60_KkMaNT2Q60aKeMcNZ2m5OWL0j3OVL2i3MYL2h3MYL2h3L_LNb31dLI]35jLEW3:oL@R3>QM@P3?RM_Oo2?TM_Om2`0UM^Ol2`0WM_Oi2`0YM^Oh2a0ZM]Og2c0ZM[Og2e0ZMYOg2f0[MXOf2h0[MVOf2i0\\MUOe2k0\\MSOe2m0\\MQOe2n0]MPOd2P1\\MPOd2P1\\MoNe2P1\\MoNe2Q1[MnNf2Q1[MnNf2R1ZMmNg2S1YMlNh2S1YMlNh2T1XMkNi2T1XMkNi2U1WMjNj2V1VMjNj2U1WMjNj2V1VMiNk2W1UMhNl2W1UMhNl2X1TMgNm2X1TMgNm2Y1SMfNn2Z1RMeNo2Z1QMfNP3Z1PMeNQ3Z1PMfNP3Z1j400000000000000000000000000001O001O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O00c0]OS1mNc0]O2N2N2N1O2N2N2N2N1O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O2N2N2N2N2N2N2N2N3M2N2N2N2N2N2eM[2RNo1F9G9H8G9G9G9GWUR1"}, "label": "a tall man fixing another man ' s tie"}]} {"id": 468219, "image": "COCO_train2014_000000468219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a shaved head and white jacket on helps another man with his tie\"."}], "task": "refering", "answer_template": "The \"a man with a shaved head and white jacket on helps another man with his tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 58, 59, 60, 61, 62, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 226, 227, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364], "bbox": [0.35471875, 0.08219730941704036, 0.875796875, 0.9898654708520179], "iscrowd": 0, "area": 77266.66940000001, "rle": {"size": [446, 640], "counts": "\\US32k=5K4L4L5K4L4L4L5K4L4L4L3M3O11O1O10N1N2M3M3M3M3M3M3M4L3M3N2M3M3M3M3M3M4L30O7I7I7I7I5K0000O010^NfM_GZ2_8jM_GV2`8mM]GT2b8oM\\GQ2c8RN[Gn1c8VN[Gj1d8YNYGg1g8\\NWGd1h8_NVGa1h8cNVG]1i8fNTG[1k8hNSGX1l8kNRGU1l8nNSGR1l8nNTGR1l8oNTGQ1k8oNVGQ1i8oNXGQ1h8nNXGS1h8mNXGS1g8mNZGS1f8lN[GT1d8lN\\GT1e8lN[GT1d8lN]GT1c8kN^GU1b8jN^GW1a8jN_GV1a8iN`GW1_8iNaGX1_8gNbGX1^8iNbGW1^8hNcGX1]8gNcGZ1\\8fNeGZ1[8fNeGZ1Z8fNgGZ1Y8eNgG\\1Y8cNhG]1X8cNeG_1\\8`NbGb1^8^N`Gd1a8[N]Gg1d8YNYGi1h8VNVGl1k8SNSGo1n8PNPGR2Q9nMlFT2U9kMiFW2X9hMfFZ2[9eMdF\\2]9dM`F^2a9aM]Fa2c9m01O1O1O1O1O1O1O1OG9E;F:E;F:F:N2O1O1O1O1O1O1O1SKkJP1V5oNmJo0S5oNQKn0P5QORKm0o4QOUKl0l4SOWKj0j4TOYKj0h4UO[KUN]Ni1Y61]KoM_No1U60_KkMaNT2Q60aKeMcNZ2m5OWL0j3OVL2i3MYL2h3MYL2h3L_LNb31dLI]35jLEW3:oL@R3>QM@P3?RM_Oo2?TM_Om2`0UM^Ol2`0WM_Oi2`0YM^Oh2a0ZM]Og2c0ZM[Og2e0ZMYOg2f0[MXOf2h0[MVOf2i0\\MUOe2k0\\MSOe2m0\\MQOe2n0]MPOd2P1\\MPOd2P1\\MoNe2P1\\MoNe2Q1[MnNf2Q1[MnNf2R1ZMmNg2S1YMlNh2S1YMlNh2T1XMkNi2T1XMkNi2U1WMjNj2V1VMjNj2U1WMjNj2V1VMiNk2W1UMhNl2W1UMhNl2X1TMgNm2X1TMgNm2Y1SMfNn2Z1RMeNo2Z1QMfNP3Z1PMeNQ3Z1PMfNP3Z1j400000000000000000000000000001O001O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O00c0]OS1mNc0]O2N2N2N1O2N2N2N2N1O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O2N2N2N2N2N2N2N2N3M2N2N2N2N2N2eM[2RNo1F9G9H8G9G9G9GWUR1"}, "label": "a man with a shaved head and white jacket on helps another man with his tie"}]} {"id": 541949, "image": "COCO_train2014_000000541949.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in a black dress with white flowers on the skirt\"."}], "task": "refering", "answer_template": "The \"the woman in a black dress with white flowers on the skirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 44, 64, 65, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 130, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 297, 313, 314, 315, 316, 317, 318, 319, 320, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 423, 424, 425, 426, 428, 429, 430, 431, 432, 433, 444, 445, 446, 447, 448, 452, 453, 454, 455, 466, 467, 468, 469, 475, 476, 477, 478, 488, 489, 490, 491, 492, 498, 499, 500], "bbox": [0.21703125, 0.06516339869281046, 0.99353125, 0.99968954248366], "iscrowd": 0, "area": 96663.423, "rle": {"size": [612, 640], "counts": "mWd22Rc01O1O2N1O1O2N2N1O2N2N2N2N001O000000000000000000000000000000000000000O1O1O1O1O101N1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O10O01O0O101N101N0010O010ZOf_ObNZ`0^1o_OZNQ`0e1X@RNh?o1f0O010O02O0O101N101N101N1O101N101N101N100O2O0O2O0O2O0O101N101N100O2N100O101N100O2O0O100O2O0O10001N10000O2O00000O10000O10000O10000O2O00000O10000O10000O100001O001O1O1O1N1O2M3M3N2M3M2O2M3M3N2M2N3N2N2O1O00100O1O1O1O0012N2N2N2M3N1O2N1O2M2O2N1O2N1O2M2O2N2N1O2M2O1O1O1O1N2O2N1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O001N101O001O001dIgKjMZ4V2gKiMZ4W2fKhMZ4Y2fKfM[4Y2gKeMY4\\2gKbM[4l1bKcH3`5[4l1dKdH2^5[4g1^KbH781]5Z4YJTLe6^OZI5:O]5c5`0gJSJFZ5i5>cJWJEY5j5>cJXJDX5k5?bJXJCo1a05[5d3bJWJCn1`07]5c3aJWJCm1`08^5c3aJWJAl1`09a5d3PKRL_O:c5d3nJoKAe5e3iJlKB?g5d3hJhKEb0g5e3dJdKHg0f5d3cJ`KKk0d5e3`J]KNn0d5e3^JXK1S1c5d3]JTK4V1b5f3ZJoJ7[1`5f3ZJjJ9`1_5f3XJfJi1[5i3WJYJa0n1X5e3]JXJ?Q2U5b3gLiKULROT7P5kLmK[LkNj6S5oLQLk3n3ULQLm3o3RLPLP4P4oKPLR4R4kKmKW4T4gKkK[4V4cKiK_4X4dGUKT3b0Z5P5eJPK\\5o4dJQK]5o4bJQK_5P5_JPKb5Q5]JnJd5S5ZJmJg5S5XJmJi5P5YJPKh5m4ZJSKg5l4YJTKh5j4YJUKi5j4WJUKk5i4VJVKl5i4TJVKn5i4RJVKP6i4PJVKR6i4U4O1O1O1N200O10000OROaJYD_5a;iJ]DV5];RKbDn4W;[KgDe4X;]KgDb4Z;`KdD`4\\;aKcD^4^;dK`D\\4`;fK^DY4c;hK\\DX4e;iKYDW4g;jKXDV4h;lKVDT4j;nKTDR4l;PLRDo3o;SLoCm3QX1cAYNa>c1nAmMW>o1d1I7J6J7nN[^O8na0^OZ^O in this image.", "objects": [{"patches": [12, 13, 14, 15, 35, 36, 37, 38, 58, 59, 60, 61, 81, 82, 83, 84, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 235, 236, 237, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 286, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 314, 332, 333, 334, 335, 355, 356, 404, 427, 428, 450, 451, 452, 473, 474, 475, 496, 497, 498], "bbox": [0.24815625, 0.013643790849673203, 0.69740625, 0.9937091503267973], "iscrowd": 0, "area": 59786.1725, "rle": {"size": [612, 640], "counts": "jYo26mb0:F:F2OO01O10O01O1O010O1O010O1O00100O0010O01O00010O0010O01O0010O01O010O001O010O0010O01O00010O0010O01O010O001O010O0010O01O0010O01O010O000010O01O010O001O010O0010O01O0010O01O010O000010O01O010O001O010O001O0N3L3J7G8G:N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1010O10O0100O010O1O010O\\AmLY=S3iA_Nb=V4\\Od0\\Od0[Od0]Od0\\Od0[Oe0\\O1ON101N2O0O2O1N101N2O1N101N2O0O2O1N101N2O0O2O1N101N2O0O2O1N101N2O0OeNnGhGQ8m7dHjG[7j7ZImGg6g7mIQHS6h7g201N2N2O1N2N2N2O0O2N2N2O1N2N2O10O10000000000000YKVDT2i;nMaDg1_;ZNkD[1U;fNUEo0k:RO_Eb0b:_OgE7Y:JQFKo96[F_Oe9c0dFRO[9P1oFeNQ9\\1YGYNg8h1cGmM]8T2mGaMS8`2mG_MS8b2lG^MT8c2kG]MU8d2jG\\MU8f2jGZMV8h2hGWMY8j2fGVMZ8k2dGVM\\8k2cGUM]8l2bGTM^8m2aGSM_8n2`GRM`8o2_GQM`8Q3_GoLa8R3^GnLb8S3]GmLc8U3[GkLe8V3ZGjLf8W3XGjLh8W3WGiLi8X3VGgLj8[3UGeLk8\\3TGdLl8]3SGcLm8^3RGbLn8_3QGaLo8`3PG`LP9b3nF^LR9c3mF]LR9e3lF\\LT9e3kF[LU9f3jFZLV9g3iFYLW9h3hFXLX9i3gFVLZ9k3eFUL[9l3dFULT1F]4X4]JSLT1J[4T4`JSLR1OZ4o3cJRLR14W4k3eJSLR17U4g3hJSLP1 in this image.", "objects": [{"patches": [12, 13, 14, 15, 35, 36, 37, 38, 58, 59, 60, 61, 81, 82, 83, 84, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 235, 236, 237, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 286, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 314, 332, 333, 334, 335, 355, 356, 404, 427, 428, 450, 451, 452, 473, 474, 475, 496, 497, 498], "bbox": [0.24815625, 0.013643790849673203, 0.69740625, 0.9937091503267973], "iscrowd": 0, "area": 59786.1725, "rle": {"size": [612, 640], "counts": "jYo26mb0:F:F2OO01O10O01O1O010O1O010O1O00100O0010O01O00010O0010O01O0010O01O010O001O010O0010O01O00010O0010O01O010O001O010O0010O01O0010O01O010O000010O01O010O001O010O0010O01O0010O01O010O000010O01O010O001O010O001O0N3L3J7G8G:N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1010O10O0100O010O1O010O\\AmLY=S3iA_Nb=V4\\Od0\\Od0[Od0]Od0\\Od0[Oe0\\O1ON101N2O0O2O1N101N2O1N101N2O0O2O1N101N2O0O2O1N101N2O0O2O1N101N2O0OeNnGhGQ8m7dHjG[7j7ZImGg6g7mIQHS6h7g201N2N2O1N2N2N2O0O2N2N2O1N2N2O10O10000000000000YKVDT2i;nMaDg1_;ZNkD[1U;fNUEo0k:RO_Eb0b:_OgE7Y:JQFKo96[F_Oe9c0dFRO[9P1oFeNQ9\\1YGYNg8h1cGmM]8T2mGaMS8`2mG_MS8b2lG^MT8c2kG]MU8d2jG\\MU8f2jGZMV8h2hGWMY8j2fGVMZ8k2dGVM\\8k2cGUM]8l2bGTM^8m2aGSM_8n2`GRM`8o2_GQM`8Q3_GoLa8R3^GnLb8S3]GmLc8U3[GkLe8V3ZGjLf8W3XGjLh8W3WGiLi8X3VGgLj8[3UGeLk8\\3TGdLl8]3SGcLm8^3RGbLn8_3QGaLo8`3PG`LP9b3nF^LR9c3mF]LR9e3lF\\LT9e3kF[LU9f3jFZLV9g3iFYLW9h3hFXLX9i3gFVLZ9k3eFUL[9l3dFULT1F]4X4]JSLT1J[4T4`JSLR1OZ4o3cJRLR14W4k3eJSLR17U4g3hJSLP1 in this image.", "objects": [{"patches": [52, 53, 54, 55, 75, 76, 77, 78, 98, 99, 100, 101, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 282, 283, 284, 285, 286, 305, 306, 307, 308, 309, 328, 329, 330, 331, 332, 350, 351, 352, 353, 354, 355, 373, 374, 375, 376, 377, 378, 397, 398, 399, 400, 420, 421, 422, 423, 425, 443, 444, 445, 447, 448, 449, 466, 467, 469, 470, 471, 472, 492, 493, 494, 495], "bbox": [0.013734374999999998, 0.08816993464052288, 0.5499218750000001, 1.0], "iscrowd": 0, "area": 71357.24805000002, "rle": {"size": [612, 640], "counts": "fd53Pc05K4K6K5K4L5K4K5L1O2N1O1O2N1O1O2N100O2N100O2N100O2N100O2N100O2N1O101N1O101N1O101N1O101N1O101N1O100O1O1O100O1O100O2N100O1O100O1O100O1O1O100O1O100O1O100O2N100O1O100O1O1O100O1O10O01O100O1O10O01O100O100O1O010O1O100O1O010O1O100O1O010O1O100O1O010O1O100O1O010O100O100O010O100O100O00100O100O10O0_O_AaLa>^3b010O010O010O010O010OiD\\Lo6e3mHXMo5QOYFg3e3VNjNXN]3OSJd3a3ROlMoMS4_O^Ja3`3NkLcMn4POgJ`3[3m1j1hJiJ]3X3n1l1jJjJZ3U3P2l1kJoJV3P3S2l1oNPNS1m1ROnMR1m1SOoMP1i1ZORNj0f1_O^LSJQ1a6Z2EdLlIm0b6[2GfLjIi0b6_2HgLhIh0`6_2JhLiIf0]6a2MhLgIe0]6`20iLfIe0Y6a24hLfId0W6a27jLdIb0U6c2TOPMjJI0b0R6c2UOVMgJD4`0P6e2WO[MaJ_O:`0n5f2VO_MbJXO:b0n5g2VOdM\\KBa5i2TOfMZK^Oe5k2QOiMYKXOi5P3nNjMWKTOm5Q3lNmMVKPOo5S3kNmMWKmNQ6U3iNnMVKkNR6W3hNnMb2R2^MnMa2R2_MnMb2Q2^MPNa2P2`MoMa2Q2^MPNa2P2`MoMa2P2_MQN`2o1`MQNa2o1_MQN`2o1`MQN`2o1aMQN^2o1bMQN^2P2aMQN^2P2bMoM^2R2aMnM`2R2`MnM_2S2`MmM`2T2_MmM`2T2`MkM`2V2_MkM`2V2_MjMa2W2_MiM`2X2_MhMa2Y2_MgMa2Y2^MgMb2Z2]MgMb2Z2^MeMb2\\2]MeMb2\\2^McMb2^2]McMb2^2]MbMc2_2]MaMb2`2]M`Mc2a2]M_Mc2a2\\M_Md2b2[M_Md2b2\\M]Md2d2[M\\Me2e2[M[Md2f2[MZMe2g2ZMZMe2h2ZMWMe2l2YMUMf2m2YMRMg2P3WMQMg2Q3XMoLh2S3WMmLg2V3WMjLi2Z1PHYO\\3TN\\NY1W6Z1RHZOY3TN]NX1W6[1THZOU3TN`NX1V6Z1VH[OQ3UNcNU1V6[1WH[Oo2VNcNU1U6[1ZH[Ok2WNeNS1V6[1[H\\Oh2VNhNS1S6\\1^H\\Od2WNjNQ1T6\\1_H]Oa2WNlNP1S6\\1aH^O]2WNoNo0R6]1cH^OX2YNROm0R6\\1eH_OS2[NVOi0Q6^1gH_On1]NYOf0R6m0aHRN7P2h1]N_Od0P6k0]I5P1_NBa0P6j0aI7i0aNE?P6g0eI;a0cNI:P6g0iI=:fNK7Q6e0lI?4hNO3Q6f0mIa0OiN11Q6g0oI`0LjNSOjN in this image.", "objects": [{"patches": [52, 53, 54, 55, 75, 76, 77, 78, 98, 99, 100, 101, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 282, 283, 284, 285, 286, 305, 306, 307, 308, 309, 328, 329, 330, 331, 332, 350, 351, 352, 353, 354, 355, 373, 374, 375, 376, 377, 378, 397, 398, 399, 400, 420, 421, 422, 423, 425, 443, 444, 445, 447, 448, 449, 466, 467, 469, 470, 471, 472, 492, 493, 494, 495], "bbox": [0.013734374999999998, 0.08816993464052288, 0.5499218750000001, 1.0], "iscrowd": 0, "area": 71357.24805000002, "rle": {"size": [612, 640], "counts": "fd53Pc05K4K6K5K4L5K4K5L1O2N1O1O2N1O1O2N100O2N100O2N100O2N100O2N100O2N1O101N1O101N1O101N1O101N1O101N1O100O1O1O100O1O100O2N100O1O100O1O100O1O1O100O1O100O1O100O2N100O1O100O1O1O100O1O10O01O100O1O10O01O100O100O1O010O1O100O1O010O1O100O1O010O1O100O1O010O1O100O1O010O100O100O010O100O100O00100O100O10O0_O_AaLa>^3b010O010O010O010O010OiD\\Lo6e3mHXMo5QOYFg3e3VNjNXN]3OSJd3a3ROlMoMS4_O^Ja3`3NkLcMn4POgJ`3[3m1j1hJiJ]3X3n1l1jJjJZ3U3P2l1kJoJV3P3S2l1oNPNS1m1ROnMR1m1SOoMP1i1ZORNj0f1_O^LSJQ1a6Z2EdLlIm0b6[2GfLjIi0b6_2HgLhIh0`6_2JhLiIf0]6a2MhLgIe0]6`20iLfIe0Y6a24hLfId0W6a27jLdIb0U6c2TOPMjJI0b0R6c2UOVMgJD4`0P6e2WO[MaJ_O:`0n5f2VO_MbJXO:b0n5g2VOdM\\KBa5i2TOfMZK^Oe5k2QOiMYKXOi5P3nNjMWKTOm5Q3lNmMVKPOo5S3kNmMWKmNQ6U3iNnMVKkNR6W3hNnMb2R2^MnMa2R2_MnMb2Q2^MPNa2P2`MoMa2Q2^MPNa2P2`MoMa2P2_MQN`2o1`MQNa2o1_MQN`2o1`MQN`2o1aMQN^2o1bMQN^2P2aMQN^2P2bMoM^2R2aMnM`2R2`MnM_2S2`MmM`2T2_MmM`2T2`MkM`2V2_MkM`2V2_MjMa2W2_MiM`2X2_MhMa2Y2_MgMa2Y2^MgMb2Z2]MgMb2Z2^MeMb2\\2]MeMb2\\2^McMb2^2]McMb2^2]MbMc2_2]MaMb2`2]M`Mc2a2]M_Mc2a2\\M_Md2b2[M_Md2b2\\M]Md2d2[M\\Me2e2[M[Md2f2[MZMe2g2ZMZMe2h2ZMWMe2l2YMUMf2m2YMRMg2P3WMQMg2Q3XMoLh2S3WMmLg2V3WMjLi2Z1PHYO\\3TN\\NY1W6Z1RHZOY3TN]NX1W6[1THZOU3TN`NX1V6Z1VH[OQ3UNcNU1V6[1WH[Oo2VNcNU1U6[1ZH[Ok2WNeNS1V6[1[H\\Oh2VNhNS1S6\\1^H\\Od2WNjNQ1T6\\1_H]Oa2WNlNP1S6\\1aH^O]2WNoNo0R6]1cH^OX2YNROm0R6\\1eH_OS2[NVOi0Q6^1gH_On1]NYOf0R6m0aHRN7P2h1]N_Od0P6k0]I5P1_NBa0P6j0aI7i0aNE?P6g0eI;a0cNI:P6g0iI=:fNK7Q6e0lI?4hNO3Q6f0mIa0OiN11Q6g0oI`0LjNSOjN in this image.", "objects": [{"patches": [75, 97, 98, 99, 120, 121, 122, 143, 166, 167, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262, 280, 281, 282, 283, 284, 285, 303, 304, 305, 306, 307, 308, 326, 327, 328, 329, 330, 349, 350, 351, 352, 353, 372, 373, 374, 375, 376, 395, 396, 397, 398, 399, 418, 419, 420, 421, 422, 441, 442, 443, 444, 445, 463, 464, 465, 466, 467, 468, 486, 487, 488, 489, 490, 491], "bbox": [0.13709375, 0.1727463651050081, 0.40198437499999995, 0.9973344103392568], "iscrowd": 0, "area": 49379.761999999995, "rle": {"size": [619, 640], "counts": "Qie12Xc04L3M4L4L3M1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1lJfNVH[1b7POZHQ1^7[O]Hf0[7EaH in this image.", "objects": [{"patches": [75, 97, 98, 99, 120, 121, 122, 143, 166, 167, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262, 280, 281, 282, 283, 284, 285, 303, 304, 305, 306, 307, 308, 326, 327, 328, 329, 330, 349, 350, 351, 352, 353, 372, 373, 374, 375, 376, 395, 396, 397, 398, 399, 418, 419, 420, 421, 422, 441, 442, 443, 444, 445, 463, 464, 465, 466, 467, 468, 486, 487, 488, 489, 490, 491], "bbox": [0.13709375, 0.1727463651050081, 0.40198437499999995, 0.9973344103392568], "iscrowd": 0, "area": 49379.761999999995, "rle": {"size": [619, 640], "counts": "Qie12Xc04L3M4L4L3M1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1lJfNVH[1b7POZHQ1^7[O]Hf0[7EaH in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 190, 191, 192, 193, 194], "bbox": [0.009875, 0.06233552631578947, 0.44539062500000004, 0.7506578947368421], "iscrowd": 0, "area": 55381.225999999995, "rle": {"size": [304, 640], "counts": "hi1h0h8U5kJ00000000001O00000000000000000000000000000000001O0000000000000000000000000000000000001O000O1000000000000000000000000000001O0000000000000000000000000000000000001O0000000000000000000000000000000O101O0000000000000000000000000000000000001O00000000000000000000000000000000001O0000000000000000000000000O10000000001O0000000000000000000000000000000000001O00000000000000000000000000000000001O00000000000000000O1000000000000000001O00000000000000000000000000000000001O0000000000000000000000000000000000001O0000000O100000000000000000000000001O00000000000000000000000m[Y3"}, "label": "left side monitor"}]} {"id": 124169, "image": "COCO_train2014_000000124169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the monitor on the left\"."}], "task": "refering", "answer_template": "The \"the monitor on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 190, 191, 192, 193, 194], "bbox": [0.009875, 0.06233552631578947, 0.44539062500000004, 0.7506578947368421], "iscrowd": 0, "area": 55381.225999999995, "rle": {"size": [304, 640], "counts": "hi1h0h8U5kJ00000000001O00000000000000000000000000000000001O0000000000000000000000000000000000001O000O1000000000000000000000000000001O0000000000000000000000000000000000001O0000000000000000000000000000000O101O0000000000000000000000000000000000001O00000000000000000000000000000000001O0000000000000000000000000O10000000001O0000000000000000000000000000000000001O00000000000000000000000000000000001O00000000000000000O1000000000000000001O00000000000000000000000000000000001O0000000000000000000000000000000000001O0000000O100000000000000000000000001O00000000000000000000000m[Y3"}, "label": "the monitor on the left"}]} {"id": 124169, "image": "COCO_train2014_000000124169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dell monitor running a multimedia suite\"."}], "task": "refering", "answer_template": "The \"a dell monitor running a multimedia suite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224], "bbox": [0.4461875, 0.026973684210526313, 0.9852187499999999, 0.8629276315789474], "iscrowd": 0, "area": 77828.60335000002, "rle": {"size": [304, 640], "counts": "bmd2d6l200000000000000000O10000000000000001O000000000000000000000000000000000000001O0000000000000O1000000000000000000000001O0000000000000000000000000000000000001O0000000O1000000000000000000000000000001O000000000000000000000000000000000000001N100000000000000000000000000000000000001O0000000000000000b0^O0000001O00000O10001O000000001O00000000001O00000000001O00000000001O000000001O00000000N1]Od00000001O0000000000000000000000000000000000000000000000000000000000000O1000001O000000000000000000000000000000000000000000000000000000000000000O1000001O000000000000000000000000000000000000000000000000000000000000000O10001O00000000000000000000000000000000000000000000000000000000000000000O10001O0000ZOf0fNZ1fNZ1fNZ1fNZ1eNcf2"}, "label": "a dell monitor running a multimedia suite"}]} {"id": 124169, "image": "COCO_train2014_000000124169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"desktop screen which is displaying options for viewing the pictures + videos\"."}], "task": "refering", "answer_template": "The \"desktop screen which is displaying options for viewing the pictures + videos\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224], "bbox": [0.4461875, 0.026973684210526313, 0.9852187499999999, 0.8629276315789474], "iscrowd": 0, "area": 77828.60335000002, "rle": {"size": [304, 640], "counts": "bmd2d6l200000000000000000O10000000000000001O000000000000000000000000000000000000001O0000000000000O1000000000000000000000001O0000000000000000000000000000000000001O0000000O1000000000000000000000000000001O000000000000000000000000000000000000001N100000000000000000000000000000000000001O0000000000000000b0^O0000001O00000O10001O000000001O00000000001O00000000001O00000000001O000000001O00000000N1]Od00000001O0000000000000000000000000000000000000000000000000000000000000O1000001O000000000000000000000000000000000000000000000000000000000000000O1000001O000000000000000000000000000000000000000000000000000000000000000O10001O00000000000000000000000000000000000000000000000000000000000000000O10001O0000ZOf0fNZ1fNZ1fNZ1fNZ1eNcf2"}, "label": "desktop screen which is displaying options for viewing the pictures + videos"}]} {"id": 304408, "image": "COCO_train2014_000000304408.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man carrying a young girl\"."}], "task": "refering", "answer_template": "The \"a man carrying a young girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 83, 84, 107, 108, 110, 111, 129, 130, 131, 134, 152, 153, 154, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 198, 201, 202, 203, 221], "bbox": [0.6167812500000001, 0.12620833333333334, 0.877578125, 0.5677291666666666], "iscrowd": 0, "area": 10934.825199999996, "rle": {"size": [480, 640], "counts": "j[i56m0Mg<ZCG]mGo1I^MY8P1bG`1<\\MP8W1bG\\1d9fN\\FW1e9lNYFS1g9oNXFR1f9QOXFP1f9SOYFl0g9UOXFl0h9UOVFl0i9UOVFl0i9VOVFi0k9XOSFi0l9XOSFi0l9ZORFe0o9\\OoE`0U:BiE:[:HdE3a:Q24K6K5K5K5K4SMgDR2\\;mMUEa1m:^NXE\\1j:dNWEY1k:fNXEV19`NR99hFT12mNQ9NPGQ1OVOn8ITGn0O[Ol8FXGk0M@k8D[Gi0IEl8B]Ge0HJk8@`Gb0GOh8^OeG>E4g8]OfG;E8f8[OhG:C;f8YOjG8B`0c8XOmG4Bd0b8VOoG2Bg0`8UOQH1@j0_8UOjIk0U:01O001O10O01O0010O01O0001O0000001O0000000000000000000000000000000WB_OS=a0dBI[=o0O1O001XM]NQHd1o7eNgG\\1X8dNhG]1X8cNgG]1Y8cNfG]1[8dNdGY1`8fN_GX1d8iN[GT1i8kNVGS1m8nNRGP1Q9POmFn0V9ROjFk0Z9UOdFj0^9WOaFf0b9[O\\Fd0g9[OYFb0j9_OTF`0o9@oE>T:BkEkFCU9`0fFB[9?aFC_9`0\\FBe9?WFCi9`0RFBP:>mEBU:`0fEB[:?aECa:`0TEFm:>hDGZ;>ZDHg;h12N1N2O2N1O1O2N1O2N4K4L4L5K4L4@a0[Od0\\On\\T1"}, "label": "a man carrying a young girl"}]} {"id": 304408, "image": "COCO_train2014_000000304408.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a old man holding a baby girl near jiraffe\"."}], "task": "refering", "answer_template": "The \"a old man holding a baby girl near jiraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 83, 84, 107, 108, 110, 111, 129, 130, 131, 134, 152, 153, 154, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 198, 201, 202, 203, 221], "bbox": [0.6167812500000001, 0.12620833333333334, 0.877578125, 0.5677291666666666], "iscrowd": 0, "area": 10934.825199999996, "rle": {"size": [480, 640], "counts": "j[i56m0Mg<ZCG]mGo1I^MY8P1bG`1<\\MP8W1bG\\1d9fN\\FW1e9lNYFS1g9oNXFR1f9QOXFP1f9SOYFl0g9UOXFl0h9UOVFl0i9UOVFl0i9VOVFi0k9XOSFi0l9XOSFi0l9ZORFe0o9\\OoE`0U:BiE:[:HdE3a:Q24K6K5K5K5K4SMgDR2\\;mMUEa1m:^NXE\\1j:dNWEY1k:fNXEV19`NR99hFT12mNQ9NPGQ1OVOn8ITGn0O[Ol8FXGk0M@k8D[Gi0IEl8B]Ge0HJk8@`Gb0GOh8^OeG>E4g8]OfG;E8f8[OhG:C;f8YOjG8B`0c8XOmG4Bd0b8VOoG2Bg0`8UOQH1@j0_8UOjIk0U:01O001O10O01O0010O01O0001O0000001O0000000000000000000000000000000WB_OS=a0dBI[=o0O1O001XM]NQHd1o7eNgG\\1X8dNhG]1X8cNgG]1Y8cNfG]1[8dNdGY1`8fN_GX1d8iN[GT1i8kNVGS1m8nNRGP1Q9POmFn0V9ROjFk0Z9UOdFj0^9WOaFf0b9[O\\Fd0g9[OYFb0j9_OTF`0o9@oE>T:BkEkFCU9`0fFB[9?aFC_9`0\\FBe9?WFCi9`0RFBP:>mEBU:`0fEB[:?aECa:`0TEFm:>hDGZ;>ZDHg;h12N1N2O2N1O1O2N1O2N4K4L4L5K4L4@a0[Od0\\On\\T1"}, "label": "a old man holding a baby girl near jiraffe"}]} {"id": 304408, "image": "COCO_train2014_000000304408.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe looking at a kid and older man\"."}], "task": "refering", "answer_template": "The \"a giraffe looking at a kid and older man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 299, 300, 301, 302, 322, 323, 324, 325, 345, 346, 347, 368, 369, 370], "bbox": [0.0, 0.36358333333333337, 0.548359375, 1.0], "iscrowd": 0, "area": 41966.75050000001, "rle": {"size": [480, 640], "counts": "`<`2^<2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O0O1O2O0O1O2O00001N10001O0O101OO1O001O1O1N2O001O1O1O001O1O1O001O1O1O00aMUKlJj4i4dKUK[4m4fKQKY4Q5hKlJX4V5iKhJV4Z5kKdJT4^5lKaJS4a5nK]JQ4e5PLXJP4j5QLTJm3P6SLnIl3T6ULjIj3X6WLeIi3]6WLbIh3`6YL^If3d6[LZId3h6]LUIc3m6^LQIa3Q7`LmH_3U7bLiH]3Y7cLeH]3^7cL`H\\3b7eL\\HZ3f7gLXHX3j7iLTHV3n7kLoGU3S8kLlGT3V8nLgGQ3[8QMbGn2`8SM]Gm2e8UMXGk2j8VMSGj2n8WMPGh2R9b11O1O1N2O1O1N2O2N1O1N2O1O1N2O1O1O1N3N1O1N2O1O1O1N2O1O1N3N1O1O1N2O1O1N2O1O1O1N3N1O1N2O10000O100000O1O1O1O100O001O100000000001O0O1000001O000000000O2O000000001O00000O101O000000001O0O1000001O000O10001N100O10001N100O10001N100O10001N100O101O0O100O101O0O10000O2O0O10000O2O0O10000O2O0O10001N100O10001N100O10001N10000O101N10000O2O0O10000O2O0O10001N100O10001N100O101O0O100O101O0O100O2O000O10001O0O1000001O00000O2O000000001O0O10001O000[OiBAW=?iBAW=>jBAX==jBBV=>jBBV==lBBT==mBCT=;mBES=;nBDR=;oBDR=;PCDP= in this image.", "objects": [{"patches": [141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 299, 300, 301, 302, 322, 323, 324, 325, 345, 346, 347, 368, 369, 370], "bbox": [0.0, 0.36358333333333337, 0.548359375, 1.0], "iscrowd": 0, "area": 41966.75050000001, "rle": {"size": [480, 640], "counts": "`<`2^<2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O0O1O2O0O1O2O00001N10001O0O101OO1O001O1O1N2O001O1O1O001O1O1O001O1O1O00aMUKlJj4i4dKUK[4m4fKQKY4Q5hKlJX4V5iKhJV4Z5kKdJT4^5lKaJS4a5nK]JQ4e5PLXJP4j5QLTJm3P6SLnIl3T6ULjIj3X6WLeIi3]6WLbIh3`6YL^If3d6[LZId3h6]LUIc3m6^LQIa3Q7`LmH_3U7bLiH]3Y7cLeH]3^7cL`H\\3b7eL\\HZ3f7gLXHX3j7iLTHV3n7kLoGU3S8kLlGT3V8nLgGQ3[8QMbGn2`8SM]Gm2e8UMXGk2j8VMSGj2n8WMPGh2R9b11O1O1N2O1O1N2O2N1O1N2O1O1N2O1O1O1N3N1O1N2O1O1O1N2O1O1N3N1O1O1N2O1O1N2O1O1O1N3N1O1N2O10000O100000O1O1O1O100O001O100000000001O0O1000001O000000000O2O000000001O00000O101O000000001O0O1000001O000O10001N100O10001N100O10001N100O10001N100O101O0O100O101O0O10000O2O0O10000O2O0O10000O2O0O10001N100O10001N100O10001N10000O101N10000O2O0O10000O2O0O10001N100O10001N100O101O0O100O101O0O100O2O000O10001O0O1000001O00000O2O000000001O0O10001O000[OiBAW=?iBAW=>jBAX==jBBV=>jBBV==lBBT==mBCT=;mBES=;nBDR=;oBDR=;PCDP= in this image.", "objects": [{"patches": [62, 63, 84, 85, 86, 87, 108, 109, 110, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 179, 180, 202], "bbox": [0.6619843750000001, 0.1360625, 0.8532500000000001, 0.5304166666666666], "iscrowd": 0, "area": 10420.920200000002, "rle": {"size": [480, 640], "counts": "lnV61o>0O1O1O1O1O100K5B?J5M3M3M3M3M3M3[MkN\\GW1[8UO`Gk0b8XOYGh0i8[ORGe0Q9]OjFa0Z9C`F;d9HXF4n9NoE1R:2kEOT:4hEMX:6eEK[:7bEI^:;]EFc:=ZEDf:>WECh:?UECj:?TEBl:?RECm:>PEDP;V201O00001O00001O0O101O0000001O00001O000000000O11O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1aMgDU1Z;bNoD]1R;YNXEf1j:oM_EQ2i;O1O1O0O_DXNm9g1PF\\NQ:c1lE`NU:_1jEbNV:]1lEbNU:]1lEcNT:\\1kEeNU:[1hEhNY:W1eEkN\\:S1bEPO_:o0_ESOa:m0]EUOd:j0YEYOh:e0WE]Oj:a0TEDj:0`E0c:M\\E3h:NTE0Q;1jDOY;3cDKb;5ZDJk;7QDHS<9hCH\\<[15K4L5K4L5K4L4L5L3L5K4L4L5K5K7I8HQl[1"}, "label": "the child looking at the giraffe"}]} {"id": 304408, "image": "COCO_train2014_000000304408.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young girl wearing a brown shirt , held up by an older man wearing glasses\"."}], "task": "refering", "answer_template": "The \"a young girl wearing a brown shirt , held up by an older man wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 84, 85, 86, 87, 108, 109, 110, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 179, 180, 202], "bbox": [0.6619843750000001, 0.1360625, 0.8532500000000001, 0.5304166666666666], "iscrowd": 0, "area": 10420.920200000002, "rle": {"size": [480, 640], "counts": "lnV61o>0O1O1O1O1O100K5B?J5M3M3M3M3M3M3[MkN\\GW1[8UO`Gk0b8XOYGh0i8[ORGe0Q9]OjFa0Z9C`F;d9HXF4n9NoE1R:2kEOT:4hEMX:6eEK[:7bEI^:;]EFc:=ZEDf:>WECh:?UECj:?TEBl:?RECm:>PEDP;V201O00001O00001O0O101O0000001O00001O000000000O11O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1aMgDU1Z;bNoD]1R;YNXEf1j:oM_EQ2i;O1O1O0O_DXNm9g1PF\\NQ:c1lE`NU:_1jEbNV:]1lEbNU:]1lEcNT:\\1kEeNU:[1hEhNY:W1eEkN\\:S1bEPO_:o0_ESOa:m0]EUOd:j0YEYOh:e0WE]Oj:a0TEDj:0`E0c:M\\E3h:NTE0Q;1jDOY;3cDKb;5ZDJk;7QDHS<9hCH\\<[15K4L5K4L5K4L4L5L3L5K4L4L5K5K7I8HQl[1"}, "label": "a young girl wearing a brown shirt , held up by an older man wearing glasses"}]} {"id": 427301, "image": "COCO_train2014_000000427301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing blue jeans\"."}], "task": "refering", "answer_template": "The \"a man wearing blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 107, 108, 109, 124, 125, 126, 127, 141, 142, 143, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 204, 205, 206, 207, 208, 209, 210, 211, 212, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 244, 245, 255, 256, 257, 258, 259, 260, 261, 262, 272, 273, 274, 275, 276, 277, 291, 292, 307, 308, 309, 323, 324, 325, 326], "bbox": [0.0, 0.23528125000000003, 0.49572916666666667, 0.859609375], "iscrowd": 0, "area": 51361.695649999994, "rle": {"size": [640, 480], "counts": "[>9Tc0c0]Ob0^Oc0]Oc0]Oc0^Oa0J7N2N101N2N101N2N2O0O2N2N101N2N2\\C\\Kk9d4TF^Kj9d4UF]Ki9e4UF]Kj9c4UF_Kk9a4TF`Kl9`4RFbKn9]4QFeKo9[4PFfKP:Z4nEhKR:X4lEjKS:V4lElKT:T4kEmKU:S4iEoKW:P4hERLX:n3gESLY:m3eEUL[:j3dEXL\\:h3cEYL]:g3aE[L_:e3_E]La:b3_E_L`:b3^E`Lb:`3\\EbLd:]3\\EdLd:\\3ZEfLf:Z3XEhLh:X3WEiLi:V3VElLj:T3TEnLl:R3SEnLn:Q3QEQMn:P3PERMP;n2oDSMQ;l2nDVMR;j2lDXMS;i2lDXMS;i2kDYMT;g2kD[MS;g2kD[MT;f2kD[MT;e2kD]MT;d2jD^MU;c2jD^MT;c2kD_MT;b2jD`MU;a2jD`MU;a2iDaMV;_2iDcMV;^2iDcMU;_2iDcMV;]2iDeMV;\\2iDeMV;\\2hDfMW;[2gDgMW;Z2iDgMV;Z2hDhMW;Y2gDiMX;W2hDjMW;W2gDkMX;V2fDlMZ;S2fDnMZ;R2dDPN[;[5000000O10000OM4L4L3L5L4K4M4K5L3L5L4M3M2N3M3M2N3M3M2N3L4O010O1O010O10O01O100O010O1O010O10O0100O00100O010O01O001UNWJ^Ei5Y:QKVEP5`:kKnDU4n:a2O1O001O010O1O001O001O00100O001O001O1O00100O1O10000000000000000000000000000000O10000000000000000000001O001O1O001O001O001ZNXEkIP;S6[E]Ik:b6aElHg:R7T1O2N2N2N4M3gKjBa1Z=[NhBc1[=ZNgBd1]=XNfBe1^=WNdBg1`=UNcBh1a=TNaBj1c=RN`Bl1c=PN_Bn1e=mM_BP2e=lM]BR2g=jMZBU2j=gMXBW2k=fMVBY2n=cMTB[2i?O2O0N3G8I8H7H9oNP1C>Ckof4"}, "label": "a man wearing blue jeans"}]} {"id": 427301, "image": "COCO_train2014_000000427301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the guy on the left making sarcastic and funny faces for the photo\"."}], "task": "refering", "answer_template": "The \"the guy on the left making sarcastic and funny faces for the photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 107, 108, 109, 124, 125, 126, 127, 141, 142, 143, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 204, 205, 206, 207, 208, 209, 210, 211, 212, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 244, 245, 255, 256, 257, 258, 259, 260, 261, 262, 272, 273, 274, 275, 276, 277, 291, 292, 307, 308, 309, 323, 324, 325, 326], "bbox": [0.0, 0.23528125000000003, 0.49572916666666667, 0.859609375], "iscrowd": 0, "area": 51361.695649999994, "rle": {"size": [640, 480], "counts": "[>9Tc0c0]Ob0^Oc0]Oc0]Oc0^Oa0J7N2N101N2N101N2N2O0O2N2N101N2N2\\C\\Kk9d4TF^Kj9d4UF]Ki9e4UF]Kj9c4UF_Kk9a4TF`Kl9`4RFbKn9]4QFeKo9[4PFfKP:Z4nEhKR:X4lEjKS:V4lElKT:T4kEmKU:S4iEoKW:P4hERLX:n3gESLY:m3eEUL[:j3dEXL\\:h3cEYL]:g3aE[L_:e3_E]La:b3_E_L`:b3^E`Lb:`3\\EbLd:]3\\EdLd:\\3ZEfLf:Z3XEhLh:X3WEiLi:V3VElLj:T3TEnLl:R3SEnLn:Q3QEQMn:P3PERMP;n2oDSMQ;l2nDVMR;j2lDXMS;i2lDXMS;i2kDYMT;g2kD[MS;g2kD[MT;f2kD[MT;e2kD]MT;d2jD^MU;c2jD^MT;c2kD_MT;b2jD`MU;a2jD`MU;a2iDaMV;_2iDcMV;^2iDcMU;_2iDcMV;]2iDeMV;\\2iDeMV;\\2hDfMW;[2gDgMW;Z2iDgMV;Z2hDhMW;Y2gDiMX;W2hDjMW;W2gDkMX;V2fDlMZ;S2fDnMZ;R2dDPN[;[5000000O10000OM4L4L3L5L4K4M4K5L3L5L4M3M2N3M3M2N3M3M2N3L4O010O1O010O10O01O100O010O1O010O10O0100O00100O010O01O001UNWJ^Ei5Y:QKVEP5`:kKnDU4n:a2O1O001O010O1O001O001O00100O001O001O1O00100O1O10000000000000000000000000000000O10000000000000000000001O001O1O001O001O001ZNXEkIP;S6[E]Ik:b6aElHg:R7T1O2N2N2N4M3gKjBa1Z=[NhBc1[=ZNgBd1]=XNfBe1^=WNdBg1`=UNcBh1a=TNaBj1c=RN`Bl1c=PN_Bn1e=mM_BP2e=lM]BR2g=jMZBU2j=gMXBW2k=fMVBY2n=cMTB[2i?O2O0N3G8I8H7H9oNP1C>Ckof4"}, "label": "the guy on the left making sarcastic and funny faces for the photo"}]} {"id": 427301, "image": "COCO_train2014_000000427301.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in striped shirt\"."}], "task": "refering", "answer_template": "The \"man in striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 77, 78, 79, 80, 94, 95, 96, 97, 111, 112, 113, 114, 128, 129, 130, 131, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 178, 179, 180, 181, 182, 183, 184, 185, 195, 196, 197, 198, 199, 200, 201, 202, 203, 212, 213, 214, 215, 216, 217, 218, 219, 220, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 326, 327, 328, 334, 335, 336, 343, 344, 345, 351, 352, 353, 354, 360, 361, 362, 368, 369, 370, 371, 377, 378, 385, 386, 387, 388], "bbox": [0.179125, 0.16678125, 0.9779166666666668, 1.0], "iscrowd": 0, "area": 96413.19604999997, "rle": {"size": [640, 480], "counts": "h[f15bc09G9G9G9G9G9G:F9G9G9G9I7M3M3M3M3M3M3M3M3M3M4M2M3M3E;F:O1O1O100O1O1O1O1O1O2N1O100O100O1O10N1O101N100O2N100O2N101N1O2N2N2M3N2N1N3N2N2N2M3N2N2M3N2N2N2K5J6J6J6J6J5K6J6J6L40000000000000000001O0000000O100000000000000000000000000000000O10000eN\\MZAd2e>eMSA[2l>QNi@o1U?_N_@a1`?dN\\@\\1c?jNX@V1f?c1O1O1O1N2O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1@`0_Oa0_O`0_Ob0POP1O1O10O01O100O00100O10O01O100O010O1WMiGhIX8R6]H_Ic7\\6lHYIT7c6YIUIg6e6bIWI_6d6hIZIX6`6QJ\\Io5`6XJ^Ih5]6`J_Ia5^6dJ_I]5`6gJ]IX5d6jJYIW5g6lJVIT5i6oJTIR5l6QKPIo4P7TKnHl4R7TKmHm4S7RKmHo4R7QKoHn4R7QKnHP5Q7QKnHP5R7oJoHQ5Q7nJoHS5P7mJPIS5Q7TKgHm4X7n3000O0100O10O10O100O01000O100O10000O100001O001O001O001O[L`CK_<0gCOY in this image.", "objects": [{"patches": [61, 77, 78, 79, 80, 94, 95, 96, 97, 111, 112, 113, 114, 128, 129, 130, 131, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 178, 179, 180, 181, 182, 183, 184, 185, 195, 196, 197, 198, 199, 200, 201, 202, 203, 212, 213, 214, 215, 216, 217, 218, 219, 220, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 326, 327, 328, 334, 335, 336, 343, 344, 345, 351, 352, 353, 354, 360, 361, 362, 368, 369, 370, 371, 377, 378, 385, 386, 387, 388], "bbox": [0.179125, 0.16678125, 0.9779166666666668, 1.0], "iscrowd": 0, "area": 96413.19604999997, "rle": {"size": [640, 480], "counts": "h[f15bc09G9G9G9G9G9G:F9G9G9G9I7M3M3M3M3M3M3M3M3M3M4M2M3M3E;F:O1O1O100O1O1O1O1O1O2N1O100O100O1O10N1O101N100O2N100O2N101N1O2N2N2M3N2N1N3N2N2N2M3N2N2M3N2N2N2K5J6J6J6J6J5K6J6J6L40000000000000000001O0000000O100000000000000000000000000000000O10000eN\\MZAd2e>eMSA[2l>QNi@o1U?_N_@a1`?dN\\@\\1c?jNX@V1f?c1O1O1O1N2O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1@`0_Oa0_O`0_Ob0POP1O1O10O01O100O00100O10O01O100O010O1WMiGhIX8R6]H_Ic7\\6lHYIT7c6YIUIg6e6bIWI_6d6hIZIX6`6QJ\\Io5`6XJ^Ih5]6`J_Ia5^6dJ_I]5`6gJ]IX5d6jJYIW5g6lJVIT5i6oJTIR5l6QKPIo4P7TKnHl4R7TKmHm4S7RKmHo4R7QKoHn4R7QKnHP5Q7QKnHP5R7oJoHQ5Q7nJoHS5P7mJPIS5Q7TKgHm4X7n3000O0100O10O10O100O01000O100O10000O100001O001O001O001O[L`CK_<0gCOY in this image.", "objects": [{"patches": [260, 261, 262, 263, 264, 266, 267, 268, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.213421875, 0.736924882629108, 0.9745625, 0.9857981220657277], "iscrowd": 0, "area": 29658.311149999994, "rle": {"size": [426, 640], "counts": "[[i16R=4L4L5K4N0O2N1O2M3N101N1O2O000O10000000O01000000000O1000000000O0100000001O000O100000O100O100O100O10O0100O010O10O0100O010O101O2M2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N3N1N2O1N2O1N2O1N2O1O1N100O100O10000O10000000000000000000000O10000O1000000O10O010000O100O10000001O000000001O0000001O00001O001O001O001O001O001O00000000001O001O1O001O1O1O002N1O1O2N1O2N3M001O1O1O0O2O1O0000001O00001O00001O00001O0000001O00000000000000000000000000000000O100O1O1O100O1O100O1O100O1O100O1O100O1O100O0100000O10000O1000000O10000O1000000O1000000O10000O1000000O100001O00000000001O00000000001O00000000001O0000001O00001N101O00001O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O001O1O0000001O0O10001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O000O101O000000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O00000O2O0000001O0000001O001O2N2N2N2N2NWX6"}, "label": "a man in white pants"}]} {"id": 419110, "image": "COCO_train2014_000000419110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in white trousers\"."}], "task": "refering", "answer_template": "The \"a man in white trousers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [260, 261, 262, 263, 264, 266, 267, 268, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.213421875, 0.736924882629108, 0.9745625, 0.9857981220657277], "iscrowd": 0, "area": 29658.311149999994, "rle": {"size": [426, 640], "counts": "[[i16R=4L4L5K4N0O2N1O2M3N101N1O2O000O10000000O01000000000O1000000000O0100000001O000O100000O100O100O100O10O0100O010O10O0100O010O101O2M2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N3N1N2O1N2O1N2O1N2O1O1N100O100O10000O10000000000000000000000O10000O1000000O10O010000O100O10000001O000000001O0000001O00001O001O001O001O001O001O00000000001O001O1O001O1O1O002N1O1O2N1O2N3M001O1O1O0O2O1O0000001O00001O00001O00001O0000001O00000000000000000000000000000000O100O1O1O100O1O100O1O100O1O100O1O100O1O100O0100000O10000O1000000O10000O1000000O1000000O10000O1000000O100001O00000000001O00000000001O00000000001O0000001O00001N101O00001O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O001O1O0000001O0O10001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O000O101O000000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O00000O2O0000001O0000001O001O2N2N2N2N2NWX6"}, "label": "a man in white trousers"}]} {"id": 419110, "image": "COCO_train2014_000000419110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman wearing green and brown dress\"."}], "task": "refering", "answer_template": "The \"woman wearing green and brown dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 109, 110, 111, 112, 113, 114, 125, 126, 127, 131, 132, 133, 134, 135, 136, 137, 146, 147, 148, 149, 150, 151, 152, 154, 155, 156, 157, 158, 159, 160, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 292, 293, 294, 295, 296, 297, 298, 318, 319, 320, 321, 344], "bbox": [0.327546875, 0.24943661971830988, 1.0, 0.9514788732394366], "iscrowd": 0, "area": 79893.74059999998, "rle": {"size": [426, 640], "counts": "obg2>k<5K4L3M4M3L4L3M4L4L3M4L2N2M3N2N3L3N2M3N2M4M2M3N2M3M3N1N3N1N2O2M2O1N3N1N2O2N1O1O2O0O101N1O2O0O2N2O0O2N101N2O0O2N101N1O2O00000O2O000O101O0O101O000O101O00001N1000001N1000001O0O101O0000001O00001O001O00001O001N101O0O2O2N2M3N2N2M3N2N2M3N0000O1O100O1O100O1O100O100O11O001O001O010O001O010O0010O01O0010O01O010O10O010O1O1N2N2N1O2N2N2N2N1O1O1O2N1O1O2N1O1O1O1O1C=\\Od0K5J6O100O1O100O2N101N100100O1O100O1O100O1d0\\O3MO100O100O100O10000O100O100O3N1N3N1N3N1O2M2O2M2O1F8M4L4M3N1N3N2M3N1N3N2M3N1N3N2M3N1O2N2N2N1O2O1M3M2N3M3L4M2N3M3M3M2O2N1O2N100O1O2N1O1O1O2N1O100O2N1O1O1O2N1O1O101N1O1O1O2N100O1O2N100O1O2N100O1O2N100O1O2O0O1O1O2O0O1O1O2O0O1O1O2O0O1O1O2O0O1O101N1O1O10001O0000001O0001O0001O0000001O000000001O000000001O0000001O00001O00010O00001O00001O00001O00001O00001O00001O00001O00000010O0001O00001O00001O00001O00001O000012M3M3N2M3MO2O00001O0O101O00001N10001O000O2O00001O0O101O000YL"}, "label": "woman wearing green and brown dress"}]} {"id": 419110, "image": "COCO_train2014_000000419110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman being touched by the elephant trunk\"."}], "task": "refering", "answer_template": "The \"the woman being touched by the elephant trunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 109, 110, 111, 112, 113, 114, 125, 126, 127, 131, 132, 133, 134, 135, 136, 137, 146, 147, 148, 149, 150, 151, 152, 154, 155, 156, 157, 158, 159, 160, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 292, 293, 294, 295, 296, 297, 298, 318, 319, 320, 321, 344], "bbox": [0.327546875, 0.24943661971830988, 1.0, 0.9514788732394366], "iscrowd": 0, "area": 79893.74059999998, "rle": {"size": [426, 640], "counts": "obg2>k<5K4L3M4M3L4L3M4L4L3M4L2N2M3N2N3L3N2M3N2M4M2M3N2M3M3N1N3N1N2O2M2O1N3N1N2O2N1O1O2O0O101N1O2O0O2N2O0O2N101N2O0O2N101N1O2O00000O2O000O101O0O101O000O101O00001N1000001N1000001O0O101O0000001O00001O001O00001O001N101O0O2O2N2M3N2N2M3N2N2M3N0000O1O100O1O100O1O100O100O11O001O001O010O001O010O0010O01O0010O01O010O10O010O1O1N2N2N1O2N2N2N2N1O1O1O2N1O1O2N1O1O1O1O1C=\\Od0K5J6O100O1O100O2N101N100100O1O100O1O100O1d0\\O3MO100O100O100O10000O100O100O3N1N3N1N3N1O2M2O2M2O1F8M4L4M3N1N3N2M3N1N3N2M3N1N3N2M3N1O2N2N2N1O2O1M3M2N3M3L4M2N3M3M3M2O2N1O2N100O1O2N1O1O1O2N1O100O2N1O1O1O2N1O1O101N1O1O1O2N100O1O2N100O1O2N100O1O2N100O1O2O0O1O1O2O0O1O1O2O0O1O1O2O0O1O1O2O0O1O101N1O1O10001O0000001O0001O0001O0000001O000000001O000000001O0000001O00001O00010O00001O00001O00001O00001O00001O00001O00001O00000010O0001O00001O00001O00001O00001O000012M3M3N2M3MO2O00001O0O101O00001N10001O000O2O00001O0O101O000YL"}, "label": "the woman being touched by the elephant trunk"}]} {"id": 369961, "image": "COCO_train2014_000000369961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a yellow t - shirt runs with a frisbee in a farm ' s empty field\"."}], "task": "refering", "answer_template": "The \"a boy in a yellow t - shirt runs with a frisbee in a farm ' s empty field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 198, 199, 200, 220, 221, 222, 223, 245, 246, 267, 268, 269, 270, 290, 291, 292, 293, 314, 315], "bbox": [0.589140625, 0.4699297423887588, 0.75046875, 0.8947540983606558], "iscrowd": 0, "area": 8034.685250000004, "rle": {"size": [427, 640], "counts": "l^m42X=2]NOiE1X:OhE1Y:OfE1Z:OeE2[:NeE2[:OdE2[:NeE8U:HlE9R:GnE9R:GoE8Q:HoE9P:IoE6o9OgDHS18S:;jEET:a0iE_OU:c0kE\\OS:g0lEZOS:f0mEZOR:g0oEXOP:i0PFXOo9h0QFXOo9h0QFYOm9h0SFXOm9h0TFWOl9j0]ERO65[:j0ZEYO8M^:j0XED1Cf:U2oFaMV7_2gHdMY7]2bHhM^7W2^HmMb7S2[HPNe7P2WHTNj7k1SHYNl7g1PH]NQ8b1lGaNU8]1jGfNV8Y1gGPOT8o0jGTOV8k0fGZO^8a0]GEf8VNRGl142o8VNkFa11?S9lMkF12n3X91O2N1O2O1N1O2VGgK^8Z4_GiK`8Y4]GhKd8X4ZGiKf8a4MiN]GfMd8Y2`GcM`8]2dG_M]8`2gG\\MY8d2kGXMU8f2oGXMR8e2QHZMo7e2THXMn7f2UHXMk7h2WHVMi7i2ZHUMg7i2\\HUMd7k2^HRMd7m2[HTMe7k2\\HUMd7k2\\HUMe7j2\\HTMe7k2_HRMa7_2TGmM[:P2fEQNZ:m1gESN[:k1fEUN[:j1dEZNY:f1fE[N[:d1eEZN^:^1hEcN[:Q1mEQOX:g0jEZO[:`0eEB`:7_EKg:NYE3h:XOgD`0a09k:EUE>k:^OWEc0i:[OWEj0f:SO\\EQ1b:lN_EU1a:hN`EZ1X;1O0001N2N4K7J6Go\\R2"}, "label": "a boy in a yellow t - shirt runs with a frisbee in a farm ' s empty field"}]} {"id": 369961, "image": "COCO_train2014_000000369961.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy wearing a yellow shirt\"."}], "task": "refering", "answer_template": "The \"a boy wearing a yellow shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 198, 199, 200, 220, 221, 222, 223, 245, 246, 267, 268, 269, 270, 290, 291, 292, 293, 314, 315], "bbox": [0.589140625, 0.4699297423887588, 0.75046875, 0.8947540983606558], "iscrowd": 0, "area": 8034.685250000004, "rle": {"size": [427, 640], "counts": "l^m42X=2]NOiE1X:OhE1Y:OfE1Z:OeE2[:NeE2[:OdE2[:NeE8U:HlE9R:GnE9R:GoE8Q:HoE9P:IoE6o9OgDHS18S:;jEET:a0iE_OU:c0kE\\OS:g0lEZOS:f0mEZOR:g0oEXOP:i0PFXOo9h0QFXOo9h0QFYOm9h0SFXOm9h0TFWOl9j0]ERO65[:j0ZEYO8M^:j0XED1Cf:U2oFaMV7_2gHdMY7]2bHhM^7W2^HmMb7S2[HPNe7P2WHTNj7k1SHYNl7g1PH]NQ8b1lGaNU8]1jGfNV8Y1gGPOT8o0jGTOV8k0fGZO^8a0]GEf8VNRGl142o8VNkFa11?S9lMkF12n3X91O2N1O2O1N1O2VGgK^8Z4_GiK`8Y4]GhKd8X4ZGiKf8a4MiN]GfMd8Y2`GcM`8]2dG_M]8`2gG\\MY8d2kGXMU8f2oGXMR8e2QHZMo7e2THXMn7f2UHXMk7h2WHVMi7i2ZHUMg7i2\\HUMd7k2^HRMd7m2[HTMe7k2\\HUMd7k2\\HUMe7j2\\HTMe7k2_HRMa7_2TGmM[:P2fEQNZ:m1gESN[:k1fEUN[:j1dEZNY:f1fE[N[:d1eEZN^:^1hEcN[:Q1mEQOX:g0jEZO[:`0eEB`:7_EKg:NYE3h:XOgD`0a09k:EUE>k:^OWEc0i:[OWEj0f:SO\\EQ1b:lN_EU1a:hN`EZ1X;1O0001N2N4K7J6Go\\R2"}, "label": "a boy wearing a yellow shirt"}]} {"id": 320834, "image": "COCO_train2014_000000320834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a plaid shirt with blue pants\"."}], "task": "refering", "answer_template": "The \"a person in a plaid shirt with blue pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 18, 19, 20, 21, 22, 23, 36, 37, 38, 39, 40, 41], "bbox": [0.0, 5.333333333333333e-05, 0.29994, 0.22221333333333332], "iscrowd": 0, "area": 8484.905100000002, "rle": {"size": [375, 500], "counts": "2`0U;?C2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N4L3M3M1O00O100O100VNUGh0k8SOaGe0`8ZOhG?X8AkG>S8BmG`0Q8@PH`0o7@RHa0l7_OTHb0k7^OVHc0h7]OYHd0e7\\O\\Hd0c7\\O]Hf0a7ZO`Hf0_7ZObHe0^7[OcHd0]7\\OdHc0\\7]OeHb0[7^OeHb0[7^OfHa0Z7_OgH`0Y7@iH>W7BlH;T7EPI7P7ISI4m6LVI1j6O_20000000XF0i70\\F0f15j71SH5h7MTH:g7HVH=f7GUH:k7JPH7P8MkG4U80fG1Z83aGN_86\\GKd89WGHi8c1000000000000000000000O1000000000001O0000000000000000000000000O100000000000000000000000000000000000000O100000000hMlFm1T9RNoFl1Q9SNQGl1o8TNQGl1o8SNSGl1m8TNSGl1m8SNTGm1l8SNUGl1k8SNVGm1j8SNVGm1j8RNWGn1Y90001O00000O101O1O1O1O1O3BRFmNT:l0oEROV:h0nEUOX:d0eTP4"}, "label": "a person in a plaid shirt with blue pants"}]} {"id": 320834, "image": "COCO_train2014_000000320834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white color shirt\"."}], "task": "refering", "answer_template": "The \"white color shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 18, 19, 20, 21, 22, 23, 36, 37, 38, 39, 40, 41], "bbox": [0.0, 5.333333333333333e-05, 0.29994, 0.22221333333333332], "iscrowd": 0, "area": 8484.905100000002, "rle": {"size": [375, 500], "counts": "2`0U;?C2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N4L3M3M1O00O100O100VNUGh0k8SOaGe0`8ZOhG?X8AkG>S8BmG`0Q8@PH`0o7@RHa0l7_OTHb0k7^OVHc0h7]OYHd0e7\\O\\Hd0c7\\O]Hf0a7ZO`Hf0_7ZObHe0^7[OcHd0]7\\OdHc0\\7]OeHb0[7^OeHb0[7^OfHa0Z7_OgH`0Y7@iH>W7BlH;T7EPI7P7ISI4m6LVI1j6O_20000000XF0i70\\F0f15j71SH5h7MTH:g7HVH=f7GUH:k7JPH7P8MkG4U80fG1Z83aGN_86\\GKd89WGHi8c1000000000000000000000O1000000000001O0000000000000000000000000O100000000000000000000000000000000000000O100000000hMlFm1T9RNoFl1Q9SNQGl1o8TNQGl1o8SNSGl1m8TNSGl1m8SNTGm1l8SNUGl1k8SNVGm1j8SNVGm1j8RNWGn1Y90001O00000O101O1O1O1O1O3BRFmNT:l0oEROV:h0nEUOX:d0eTP4"}, "label": "white color shirt"}]} {"id": 42308, "image": "COCO_train2014_000000042308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person holding the snowboard\"."}], "task": "refering", "answer_template": "The \"the person holding the snowboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 99, 100, 116, 117, 118, 134, 135, 136, 152, 153, 154, 170, 171, 172, 188, 189, 190, 206, 207, 208], "bbox": [0.47522000000000003, 0.38652694610778443, 0.60672, 0.9805389221556886], "iscrowd": 0, "area": 8343.4027, "rle": {"size": [334, 500], "counts": "ni]26f3Ki2=PMDo2?nLCQ3=mLFQ3;nLFQ3;jK@lN7Y5:aK3kNCe59YK>nN[Of5:TKd0kNkNK;R69_JZOM`1CVOm57]JVOK`2d5aN[JSOL`2f5l1N2O100O10OCXKUKi4e4_KYKa4e4cKVK`4i4c0\\KbJV4`5fKfJW4Z5hKjJU4W5jKjJU4W5jKjJU4W5jKjJV4V5iKkJW4e5101O000O1000000O2O1O1O2M2N2N`0@1O2O0M3L4L5K4nNcJnLc5i2V1G9F:E;E;F:N2M7I8H8H7J7H7IURP2"}, "label": "the person holding the snowboard"}]} {"id": 42308, "image": "COCO_train2014_000000042308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing with a snowboard\"."}], "task": "refering", "answer_template": "The \"a man standing with a snowboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 99, 100, 116, 117, 118, 134, 135, 136, 152, 153, 154, 170, 171, 172, 188, 189, 190, 206, 207, 208], "bbox": [0.47522000000000003, 0.38652694610778443, 0.60672, 0.9805389221556886], "iscrowd": 0, "area": 8343.4027, "rle": {"size": [334, 500], "counts": "ni]26f3Ki2=PMDo2?nLCQ3=mLFQ3;nLFQ3;jK@lN7Y5:aK3kNCe59YK>nN[Of5:TKd0kNkNK;R69_JZOM`1CVOm57]JVOK`2d5aN[JSOL`2f5l1N2O100O10OCXKUKi4e4_KYKa4e4cKVK`4i4c0\\KbJV4`5fKfJW4Z5hKjJU4W5jKjJU4W5jKjJU4W5jKjJV4V5iKkJW4e5101O000O1000000O2O1O1O2M2N2N`0@1O2O0M3L4L5K4nNcJnLc5i2V1G9F:E;E;F:N2M7I8H8H7J7H7IURP2"}, "label": "a man standing with a snowboard"}]} {"id": 116040, "image": "COCO_train2014_000000116040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bottom airplane\"."}], "task": "refering", "answer_template": "The \"bottom airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [280, 295, 296, 297, 298, 299, 300, 301, 302, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 353, 354, 355, 371, 372], "bbox": [0.41834990059642146, 0.647015625, 0.8107355864811133, 0.900984375], "iscrowd": 0, "area": 10149.92825, "rle": {"size": [640, 503], "counts": "^ZT41oc0001O001O001O001O1O001O001O001N2O001O001O001N101O1O001O0O2O001O001O1O0O2O001O001O1O0O2O001O001O001N2O001O001O001N2O001O001M2N3L3N3M3L3N3M2M4M2N3L3N3M3L3N3M2O21N2O1O1O1O100O100O1O100O100O1O100O2O0O1O100O1O100O100O1O10N1O2N1O1O2N1O1O2N1O2N1O1N3G8H9G8H8O20O0001O001ON2O1bNm_OmNT`0S1m_OkNT`0U1m_OiNT`0W1n_OfNS`0Z1n_OdNS`0\\1n_OaNT`0^1n_O`NS`0`1o_O]NR`0c1o_O[NR`0e1o_OYNR`0g1P@VNQ`0j1P@TNR`0j1P@TNQ`0l1Q@PNQ`0P2P@nMQ`0R2P@lMQ`0T2Q@iMP`0W2Q@gMo?Z2k0001N2OaNb^O=]a0BP_O3o`0L]_OJa`06l_O^OT`0a0X@UOg?j0h1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001Ocak1"}, "label": "bottom airplane"}]} {"id": 116040, "image": "COCO_train2014_000000116040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lowest plane in the sky\"."}], "task": "refering", "answer_template": "The \"the lowest plane in the sky\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [280, 295, 296, 297, 298, 299, 300, 301, 302, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 353, 354, 355, 371, 372], "bbox": [0.41834990059642146, 0.647015625, 0.8107355864811133, 0.900984375], "iscrowd": 0, "area": 10149.92825, "rle": {"size": [640, 503], "counts": "^ZT41oc0001O001O001O001O1O001O001O001N2O001O001O001N101O1O001O0O2O001O001O1O0O2O001O001O1O0O2O001O001O001N2O001O001O001N2O001O001M2N3L3N3M3L3N3M2M4M2N3L3N3M3L3N3M2O21N2O1O1O1O100O100O1O100O100O1O100O2O0O1O100O1O100O100O1O10N1O2N1O1O2N1O1O2N1O2N1O1N3G8H9G8H8O20O0001O001ON2O1bNm_OmNT`0S1m_OkNT`0U1m_OiNT`0W1n_OfNS`0Z1n_OdNS`0\\1n_OaNT`0^1n_O`NS`0`1o_O]NR`0c1o_O[NR`0e1o_OYNR`0g1P@VNQ`0j1P@TNR`0j1P@TNQ`0l1Q@PNQ`0P2P@nMQ`0R2P@lMQ`0T2Q@iMP`0W2Q@gMo?Z2k0001N2OaNb^O=]a0BP_O3o`0L]_OJa`06l_O^OT`0a0X@UOg?j0h1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001Ocak1"}, "label": "the lowest plane in the sky"}]} {"id": 116040, "image": "COCO_train2014_000000116040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fighter jet in the middle\"."}], "task": "refering", "answer_template": "The \"a fighter jet in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 183, 184, 185, 186, 187, 188, 189, 190, 202, 203, 204, 205, 206, 207, 208, 209, 210, 222, 223, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 262, 263], "bbox": [0.17727634194831013, 0.385828125, 0.7148111332007953, 0.6577343750000001], "iscrowd": 0, "area": 15700.322999999993, "rle": {"size": [640, 503], "counts": "Uah11oc0001O001O001O001O001O1O001O001O001O001O001O0010O01O1O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O1O001O001O001O001O001O0F;mNR1M400O001O010O0010O01O0010O01O00100O1O2O0O1O1O100O1O2O0O1O1O11O2O1N2N1O2N2N2O0O2N2N1O2N2O1N1ON2M3M3M4M2O1O1O1O1O1OROP_OnNn`0S1R_OmNn`0R1S_OnNl`0R1U_OnNj`0S1W_OlNh`0T1Y_OjNh`0W1X_OhNh`0X1Y_OgNg`0Z1Y_OdNh`0\\1Z_ObNf`0_1Z_O`Nf`0`1[_O_Ne`0b1[_O\\Nf`0d1\\_OZNd`0f1]_OYNc`0h1]_OVNd`0j1]_OUNc`0l1^_ORNc`0m1^_ORNb`0o1^_OoMc`0Q2^_OnMc`0R2^_OlMb`0T2__OjMc`0V2]_OiMd`0V2c00100O01O1O1O1O1O1O1O1O1O1O1OWNb^OQ1]a0oNc^OP1^a0POc^Ok0aa0UO_^Oh0da0WO]^Oi0ca0WO^^Oi0aa0WO_^Ok0_a0TOc^Ol0\\a0TOd^Om0[a0SOe^On0Za0QOh^Oo0Wa0QOi^OP1Va0POj^OQ1Ua0nNm^OR1Ra0nNn^OS1Qa0mNo^OT1Pa0kNR_OU1m`0kNS_OV1l`0jNT_OW1k`0iNV_OW1i`0hNX_OY1g`0gNY_OZ1f`0fN[_OZ1d`0eN]_O[1d`0dN\\_O]1c`0cN^_O]1a`0bN`_O_1``0`N`_Oa1_`0_Nb_Oa1]`0^Nc_Od1]`0[Na_Oh1^`0XNa_Oj1^`0UNb_Om1^`0RNa_OP2^`0PNa_OQ2_`0oM`_OS2``0kM`_OW2_`0iM`_OY2_`0gM`_O[2``0cM`_O_2_`0aM`_Oa2_`0_M__Od2a`0ZM__Oh2``0XM__Oj2``0VM`_Ok2``0SMb_Ol2^`0TMc_Ol2\\`0TMe_Ol2Z`0SMh_Om2f`001O00000000000O1N1L4L4L4L4L4M3L4L4L4L4L4L4L3M4L4L4M3L4N20000O10000O10000O10000O100O10000O10000O1000O010003L5L3M_Qi2"}, "label": "a fighter jet in the middle"}]} {"id": 116040, "image": "COCO_train2014_000000116040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"2 air force jets in between the outer jets\"."}], "task": "refering", "answer_template": "The \"2 air force jets in between the outer jets\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 183, 184, 185, 186, 187, 188, 189, 190, 202, 203, 204, 205, 206, 207, 208, 209, 210, 222, 223, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 262, 263], "bbox": [0.17727634194831013, 0.385828125, 0.7148111332007953, 0.6577343750000001], "iscrowd": 0, "area": 15700.322999999993, "rle": {"size": [640, 503], "counts": "Uah11oc0001O001O001O001O001O1O001O001O001O001O001O0010O01O1O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O1O001O001O001O001O001O0F;mNR1M400O001O010O0010O01O0010O01O00100O1O2O0O1O1O100O1O2O0O1O1O11O2O1N2N1O2N2N2O0O2N2N1O2N2O1N1ON2M3M3M4M2O1O1O1O1O1OROP_OnNn`0S1R_OmNn`0R1S_OnNl`0R1U_OnNj`0S1W_OlNh`0T1Y_OjNh`0W1X_OhNh`0X1Y_OgNg`0Z1Y_OdNh`0\\1Z_ObNf`0_1Z_O`Nf`0`1[_O_Ne`0b1[_O\\Nf`0d1\\_OZNd`0f1]_OYNc`0h1]_OVNd`0j1]_OUNc`0l1^_ORNc`0m1^_ORNb`0o1^_OoMc`0Q2^_OnMc`0R2^_OlMb`0T2__OjMc`0V2]_OiMd`0V2c00100O01O1O1O1O1O1O1O1O1O1O1OWNb^OQ1]a0oNc^OP1^a0POc^Ok0aa0UO_^Oh0da0WO]^Oi0ca0WO^^Oi0aa0WO_^Ok0_a0TOc^Ol0\\a0TOd^Om0[a0SOe^On0Za0QOh^Oo0Wa0QOi^OP1Va0POj^OQ1Ua0nNm^OR1Ra0nNn^OS1Qa0mNo^OT1Pa0kNR_OU1m`0kNS_OV1l`0jNT_OW1k`0iNV_OW1i`0hNX_OY1g`0gNY_OZ1f`0fN[_OZ1d`0eN]_O[1d`0dN\\_O]1c`0cN^_O]1a`0bN`_O_1``0`N`_Oa1_`0_Nb_Oa1]`0^Nc_Od1]`0[Na_Oh1^`0XNa_Oj1^`0UNb_Om1^`0RNa_OP2^`0PNa_OQ2_`0oM`_OS2``0kM`_OW2_`0iM`_OY2_`0gM`_O[2``0cM`_O_2_`0aM`_Oa2_`0_M__Od2a`0ZM__Oh2``0XM__Oj2``0VM`_Ok2``0SMb_Ol2^`0TMc_Ol2\\`0TMe_Ol2Z`0SMh_Om2f`001O00000000000O1N1L4L4L4L4L4M3L4L4L4L4L4L4L3M4L4L4M3L4N20000O10000O10000O10000O100O10000O10000O1000O010003L5L3M_Qi2"}, "label": "2 air force jets in between the outer jets"}]} {"id": 116040, "image": "COCO_train2014_000000116040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white red and blue airplane in the sky next to 2 other planes\"."}], "task": "refering", "answer_template": "The \"a white red and blue airplane in the sky next to 2 other planes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 94, 95, 96, 97, 111, 112, 113, 114, 115, 116, 117, 118, 132, 133, 134, 151, 152, 170, 171, 189], "bbox": [0.17310139165009938, 0.1859375, 0.5914711729622266, 0.44898437500000005], "iscrowd": 0, "area": 8696.0406, "rle": {"size": [640, 503], "counts": "Xaf11oc000001O00001O001N10001O00001O001O000O2O00001O001O00001O000O2O001O00001O00001O0O2O00001O001O00001O0O101O001O00001O00001N101O00001O00001O001N10001O00001H7I8H7I7H9I61O010O010O0010O10000O100O13M3M3M3M3M3M2N3M3M3M00O1O1O1O1O1O1O100O1O2N1O1O1O1O100O1O1O1O1O1O1I7H81O0000000000001O00000000001O0001O00000O1O1O1O1O1O1O1N2O1O1O1ORN\\_Oa0c`0YOi_Ob0V`0XOV@c0i?WOc@d0\\?XOn@c0Q?]OUA>j>B\\A9c>GcA4\\>KkA0T>0RBKl=6ZBEe=;P3O1000000O1000000O1000000O0100000O10000O1000000O1000000O0100000O1000000O100NebP4"}, "label": "a white red and blue airplane in the sky next to 2 other planes"}]} {"id": 132430, "image": "COCO_train2014_000000132430.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a faded and worn light wood bench with iron legs that also form arm braces\"."}], "task": "refering", "answer_template": "The \"a faded and worn light wood bench with iron legs that also form arm braces\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 81, 82, 92, 93, 95, 96, 97, 98, 99, 100, 101, 104, 105, 110, 111, 113, 120, 121, 122, 123, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 146, 147, 149, 151, 152, 153, 154, 155, 156, 157, 158, 159, 164, 167, 174, 175, 177, 192, 195], "bbox": [0.13019999999999998, 0.38615615615615617, 0.89498, 0.9142342342342342], "iscrowd": 0, "area": 18590.6348, "rle": {"size": [333, 500], "counts": "`Ye01[:5eFKU8:gGHW8;eGGY8>cGC[8b0`GA]8d0_G]O`8g0[G[O2[1b6_OXIU1g6Y100OM3M3M4L3M4L4L4L4K5J6K5J6K5M300001O00000002N6J6J7I6J6J5L0O1O1O1O1O1O002N2N2N0001O000000000O10000000001N1000000000000000000bNGhH:S70gH3R72mH1k64VINc67\\IM\\68cIKV69jIJS67lIMP63PJ0m51SJ1j50UJ4g5LYJ7d5J[J9b5G^J=^5DaJVOWOc0U67dJVO\\Oa0m59hJUO_O?h5mIBS6>mIBS6>mIBS6>mIBS6>mIBS6>mIBS6>mIBS6>nIAR6?nIAR6?nIAR6?nIBQ6>oIBR6=nICR6VJCj5=VJDi5=VJCj5=VJCj5=VJCj5=VJCj5=VJCj5=VJCj5=VJCj5=VJCj5=VJCj5=VJCj5=VJCj5=VJCk5 in this image.", "objects": [{"patches": [77, 78, 79, 80, 81, 82, 92, 93, 95, 96, 97, 98, 99, 100, 101, 104, 105, 110, 111, 113, 120, 121, 122, 123, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 146, 147, 149, 151, 152, 153, 154, 155, 156, 157, 158, 159, 164, 167, 174, 175, 177, 192, 195], "bbox": [0.13019999999999998, 0.38615615615615617, 0.89498, 0.9142342342342342], "iscrowd": 0, "area": 18590.6348, "rle": {"size": [333, 500], "counts": "`Ye01[:5eFKU8:gGHW8;eGGY8>cGC[8b0`GA]8d0_G]O`8g0[G[O2[1b6_OXIU1g6Y100OM3M3M4L3M4L4L4L4K5J6K5J6K5M300001O00000002N6J6J7I6J6J5L0O1O1O1O1O1O002N2N2N0001O000000000O10000000001N1000000000000000000bNGhH:S70gH3R72mH1k64VINc67\\IM\\68cIKV69jIJS67lIMP63PJ0m51SJ1j50UJ4g5LYJ7d5J[J9b5G^J=^5DaJVOWOc0U67dJVO\\Oa0m59hJUO_O?h5mIBS6>mIBS6>mIBS6>mIBS6>mIBS6>mIBS6>mIBS6>nIAR6?nIAR6?nIAR6?nIBQ6>oIBR6=nICR6VJCj5=VJDi5=VJCj5=VJCj5=VJCj5=VJCj5=VJCj5=VJCj5=VJCj5=VJCj5=VJCj5=VJCj5=VJCj5=VJCk5 in this image.", "objects": [{"patches": [145, 146, 151, 152, 153, 154, 155, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 232, 233, 234, 236, 237, 238, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 260, 261, 268], "bbox": [0.003, 0.3901401869158878, 0.8696718749999999, 0.744392523364486], "iscrowd": 0, "area": 37147.804599999996, "rle": {"size": [428, 640], "counts": "hR1:R=6J0000000001O000001O000000000000000000001O0000000000000000010O0000000000000000001O0000000000000000001O0001O0000000000000001O0000000mC^O\\;b0_DD`;<[DIe;7VDNj;2RD2n;f0000001O000001O0001O0000000000001O0000000000001O000000000010O0000000POSDa0m;_OUD?k;BWDR;CmD=S;CmD=S;CmD=S;CmD=S;CmD=S;DlD in this image.", "objects": [{"patches": [145, 146, 151, 152, 153, 154, 155, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 232, 233, 234, 236, 237, 238, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 260, 261, 268], "bbox": [0.003, 0.3901401869158878, 0.8696718749999999, 0.744392523364486], "iscrowd": 0, "area": 37147.804599999996, "rle": {"size": [428, 640], "counts": "hR1:R=6J0000000001O000001O000000000000000000001O0000000000000000010O0000000000000000001O0000000000000000001O0001O0000000000000001O0000000mC^O\\;b0_DD`;<[DIe;7VDNj;2RD2n;f0000001O000001O0001O0000000000001O0000000000001O000000000010O0000000POSDa0m;_OUD?k;BWDR;CmD=S;CmD=S;CmD=S;CmD=S;CmD=S;DlD in this image.", "objects": [{"patches": [30, 31, 32, 42, 43, 44, 45, 54, 55, 56, 57, 66, 67, 68, 69, 78, 79, 80, 81, 82, 89, 90, 91, 92, 93, 94, 95, 101, 102, 103, 104, 105, 106, 107, 113, 114, 115, 116, 117, 118, 119, 124, 125, 126, 127, 128, 129, 130, 131, 136, 137, 138, 139, 140, 141, 142, 143, 148, 149, 150, 151, 152, 153, 154, 155, 160, 161, 162, 163, 164, 165, 166, 167, 172, 173, 174, 175, 176, 177, 178, 179, 184, 185, 186, 187, 188, 189, 190, 191, 196, 197, 198, 199, 200, 201, 202, 203, 208, 209, 210, 211, 212, 213, 214, 215], "bbox": [0.3319033232628399, 0.12, 1.0, 1.0], "iscrowd": 0, "area": 70400.5303, "rle": {"size": [500, 331], "counts": "RUf17X?8G9H9F9H8G8I5J6K6I6K5J6K5J7J5J6K5J6E<^Ob0I7H8I8H7\\FWJU9P6fFTJT9Y6I7J6J6J6J6J6J6L4L4M3L4L4L4L4L4M3N2O100O100O1O100O1O100O1O100O1O100O1gLkF`M4i1Q9e0RGYM:h1e8m0SHQOm7m0WHPOj7n0ZHPOf7n0^HoNc7n0bHPO^7n0eHPO\\7n0hHPOX7n0lHPOT7n0PIoNQ7o0SIoNm6Q1UIlNl6S1WIkNi6U1YIhNh6W1ZIhNf6W1]IfNd6Z1^IdNb6[1aIbN`6^1bI`N^6_1eIWNc6i1d3O100O100O1O100O100O100O100O100O1O100O100001O001O001O00001O001O00001O001O001O00001O001O001O1O1O1O1O1O1O1O1O001REbMd8_2YGdMf8]2WGgMg8Z2UGjMj8W2SGlMl8U2PGoMo8R2nFQNQ9Z2bFjM\\9e2RF_Mm9Q3aERM^:S4O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O001O001O001O001O001O001O001O001O001O001O001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O000IjI"}, "label": "a man who wears glasses with a dark suit coat , white shirt , dark brown tie and khaki pants"}]} {"id": 443741, "image": "COCO_train2014_000000443741.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a blue suit jacket and glasses smiling\"."}], "task": "refering", "answer_template": "The \"a man wearing a blue suit jacket and glasses smiling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 42, 43, 44, 45, 54, 55, 56, 57, 66, 67, 68, 69, 78, 79, 80, 81, 82, 89, 90, 91, 92, 93, 94, 95, 101, 102, 103, 104, 105, 106, 107, 113, 114, 115, 116, 117, 118, 119, 124, 125, 126, 127, 128, 129, 130, 131, 136, 137, 138, 139, 140, 141, 142, 143, 148, 149, 150, 151, 152, 153, 154, 155, 160, 161, 162, 163, 164, 165, 166, 167, 172, 173, 174, 175, 176, 177, 178, 179, 184, 185, 186, 187, 188, 189, 190, 191, 196, 197, 198, 199, 200, 201, 202, 203, 208, 209, 210, 211, 212, 213, 214, 215], "bbox": [0.3319033232628399, 0.12, 1.0, 1.0], "iscrowd": 0, "area": 70400.5303, "rle": {"size": [500, 331], "counts": "RUf17X?8G9H9F9H8G8I5J6K6I6K5J6K5J7J5J6K5J6E<^Ob0I7H8I8H7\\FWJU9P6fFTJT9Y6I7J6J6J6J6J6J6L4L4M3L4L4L4L4L4M3N2O100O100O1O100O1O100O1O100O1O100O1gLkF`M4i1Q9e0RGYM:h1e8m0SHQOm7m0WHPOj7n0ZHPOf7n0^HoNc7n0bHPO^7n0eHPO\\7n0hHPOX7n0lHPOT7n0PIoNQ7o0SIoNm6Q1UIlNl6S1WIkNi6U1YIhNh6W1ZIhNf6W1]IfNd6Z1^IdNb6[1aIbN`6^1bI`N^6_1eIWNc6i1d3O100O100O1O100O100O100O100O100O1O100O100001O001O001O00001O001O00001O001O001O00001O001O001O1O1O1O1O1O1O1O1O001REbMd8_2YGdMf8]2WGgMg8Z2UGjMj8W2SGlMl8U2PGoMo8R2nFQNQ9Z2bFjM\\9e2RF_Mm9Q3aERM^:S4O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O001O001O001O001O001O001O001O001O001O001O001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O000IjI"}, "label": "a man wearing a blue suit jacket and glasses smiling"}]} {"id": 492894, "image": "COCO_train2014_000000492894.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"traffic signal machine\"."}], "task": "refering", "answer_template": "The \"traffic signal machine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 56, 57, 58, 59, 60, 61, 62, 63, 64, 79, 80, 81, 82, 83, 84, 85, 86, 87, 102, 103, 104, 105, 106, 107, 108, 109, 110, 125, 126, 127, 128, 129, 130, 131, 132, 133], "bbox": [0.468125, 0.14116666666666666, 0.807890625, 0.41633333333333333], "iscrowd": 0, "area": 20049.19535, "rle": {"size": [360, 640], "counts": "j`Y3a2g82N2N2N2N2N1O0O1000000000000000O100000000000O1000000000000000000000000000000O100000000000000000O1000000000O10000000000000000000000000000O100000000000000000000000O100000O10000000000000000000000000000O1000000000000000000000000000O10000000000000000000001O00000000000000000000000000000000000000000O101O0000000000000000000000000000000000000000001O000000000000000O100000000000000000000000001O0000000000000000000000000000000000000O10001J5YNgW[1"}, "label": "traffic signal machine"}]} {"id": 492894, "image": "COCO_train2014_000000492894.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the 3 lights in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the 3 lights in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 56, 57, 58, 59, 60, 61, 62, 63, 64, 79, 80, 81, 82, 83, 84, 85, 86, 87, 102, 103, 104, 105, 106, 107, 108, 109, 110, 125, 126, 127, 128, 129, 130, 131, 132, 133], "bbox": [0.468125, 0.14116666666666666, 0.807890625, 0.41633333333333333], "iscrowd": 0, "area": 20049.19535, "rle": {"size": [360, 640], "counts": "j`Y3a2g82N2N2N2N2N1O0O1000000000000000O100000000000O1000000000000000000000000000000O100000000000000000O1000000000O10000000000000000000000000000O100000000000000000000000O100000O10000000000000000000000000000O1000000000000000000000000000O10000000000000000000001O00000000000000000000000000000000000000000O101O0000000000000000000000000000000000000000001O000000000000000O100000000000000000000000001O0000000000000000000000000000000000000O10001J5YNgW[1"}, "label": "the 3 lights in the right hand picture"}]} {"id": 492894, "image": "COCO_train2014_000000492894.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white base of the signal\"."}], "task": "refering", "answer_template": "The \"white base of the signal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178], "bbox": [0.525, 0.3481388888888889, 0.836109375, 0.5925833333333332], "iscrowd": 0, "area": 10376.95015, "rle": {"size": [360, 640], "counts": "UYf33T;6K6I6J6K4K6K5J5K3N1O00001O001O001O0O10000000000000000000O100000O1000000000000000O1000000000O100000000000O100000000000O1000000000O1000000000000000O100000O10000000000000000000O10O10000000000000000000000O010000000000000O1000000000O010000000000O1000000000000O01000000000O10000000000O1000O100000O10000000000O100000O1000O1000000000000O100000O1O10O0100O1O100O00100O100O1O001O1O100O1O001O1O1O1O1O001O11YOPjT1"}, "label": "white base of the signal"}]} {"id": 517492, "image": "COCO_train2014_000000517492.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a asian woman wholding her hand to her face\"."}], "task": "refering", "answer_template": "The \"a asian woman wholding her hand to her face\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 66, 67, 68, 89, 90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 203, 204, 205, 206, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 319, 320, 321, 342, 343, 344], "bbox": [0.701671875, 0.08763466042154568, 1.0, 0.9820140515222482], "iscrowd": 0, "area": 33056.428400000004, "rle": {"size": [427, 640], "counts": "iak55U= in this image.", "objects": [{"patches": [44, 45, 66, 67, 68, 89, 90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 203, 204, 205, 206, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 319, 320, 321, 342, 343, 344], "bbox": [0.701671875, 0.08763466042154568, 1.0, 0.9820140515222482], "iscrowd": 0, "area": 33056.428400000004, "rle": {"size": [427, 640], "counts": "iak55U= in this image.", "objects": [{"patches": [46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281], "bbox": [0.0, 0.13220140515222484, 0.274640625, 0.8324355971896955], "iscrowd": 0, "area": 39625.302, "rle": {"size": [427, 640], "counts": "X7:l2On2`2VKmMh1Lh2Y2^KQNf1Ej2\\2`KSNb1Am2a2\\KTNb1\\Oo2f2ZKTNc1UOR3l2WKSNc1QOU3R3RKSNe1jNW3Y3PKRNd1eN[3_5aLaJ]3f5_LZJ_3k5]LUJc3P6XLPJh3R6WLnIg3U6WLlIh3W6ULjIj3X6ULhIj3Z6TLgIk3[6SLfIl3]6RLdIl3^6RLcIm3_6RLaIm3a6QL`Io3b6nK_IQ4c6nK]IQ4e6mK\\IR4f6lK[IT4g6jKYIU4i6iKXIV4j6hKWIW4k6aK\\I_4i7N1O1O1000000000000000001O0000000000001O00001O00001O00001O0O101O00100O1O1O1O01O0O1000000O2O000O1000001O00000O101O00000000001N100000001O000O10001O1O1O1O1O1N2O1O1O1O1O1O1N2O002N5aImJ\\4W5^KoJ^4V5[KPK_4W5[KnJ`4X5ZKlJa4Z5ZKiJYOXOP5U6cKdJXO\\OP5X6cKgJX4a5cKaJ\\4c5_K_Jb4c5YK_Ji4b5RK`JP5m61O2N00O1O1O1O001O2N2N2N2N2N2N2N2N2N2N2N2N2N4L3M3M4L3L5L3M4L6Jk0UO5K5K4L0O101N100O4L6K5J6I>@fUQ6"}, "label": "chinese woman with black hair playing with her fingers looking of the screen"}]} {"id": 517492, "image": "COCO_train2014_000000517492.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in black looks at a woman in a white shirt\"."}], "task": "refering", "answer_template": "The \"a woman in black looks at a woman in a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281], "bbox": [0.0, 0.13220140515222484, 0.274640625, 0.8324355971896955], "iscrowd": 0, "area": 39625.302, "rle": {"size": [427, 640], "counts": "X7:l2On2`2VKmMh1Lh2Y2^KQNf1Ej2\\2`KSNb1Am2a2\\KTNb1\\Oo2f2ZKTNc1UOR3l2WKSNc1QOU3R3RKSNe1jNW3Y3PKRNd1eN[3_5aLaJ]3f5_LZJ_3k5]LUJc3P6XLPJh3R6WLnIg3U6WLlIh3W6ULjIj3X6ULhIj3Z6TLgIk3[6SLfIl3]6RLdIl3^6RLcIm3_6RLaIm3a6QL`Io3b6nK_IQ4c6nK]IQ4e6mK\\IR4f6lK[IT4g6jKYIU4i6iKXIV4j6hKWIW4k6aK\\I_4i7N1O1O1000000000000000001O0000000000001O00001O00001O00001O0O101O00100O1O1O1O01O0O1000000O2O000O1000001O00000O101O00000000001N100000001O000O10001O1O1O1O1O1N2O1O1O1O1O1O1N2O002N5aImJ\\4W5^KoJ^4V5[KPK_4W5[KnJ`4X5ZKlJa4Z5ZKiJYOXOP5U6cKdJXO\\OP5X6cKgJX4a5cKaJ\\4c5_K_Jb4c5YK_Ji4b5RK`JP5m61O2N00O1O1O1O001O2N2N2N2N2N2N2N2N2N2N2N2N2N4L3M3M4L3L5L3M4L6Jk0UO5K5K4L0O101N100O4L6K5J6I>@fUQ6"}, "label": "a woman in black looks at a woman in a white shirt"}]} {"id": 206198, "image": "COCO_train2014_000000206198.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bench on baseball dugout with people sitting on it\"."}], "task": "refering", "answer_template": "The \"bench on baseball dugout with people sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 195, 196, 197, 198, 199, 200, 207, 208, 209, 210, 211, 212, 213, 215, 216, 217, 218, 219, 220, 221, 241, 242, 243, 244, 245], "bbox": [0.0, 0.5223887587822014, 0.70534375, 0.6945667447306791], "iscrowd": 0, "area": 10526.7059, "rle": {"size": [427, 640], "counts": "S7k1\\;40001O000000L4L400G`DhN_;V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1eDjN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDjN];V1cDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;U1dDkN\\;5cD?1\\O\\;5PE2EH[;6TENAL[;6TENAMZ;5UENAMZ;5UENAMZ;5VEM@NZ;5VEM@NZ;5VEM@NZ;5VEM@NZ;5VEM@NZ;5VEM@NZ;5VEM@NZ;5VEM@NZ;5VEN_OM[;5VEN_OM[;5VEN_OM[;5VEN_OM[;5VEN_OM[;5VEN_OM[;5VEN_OM[;5VEN_OM\\;4UEO_OM\\;4UEO_OM\\;4UEO_OM\\;4UEO_OM\\;4UEO_OM\\;4UEO_ON[;3VEO_ON[;3VEO_ON[;3VEO_ON[;3VEO_ON[;3VEO_ON[;3VEO_ON[;3VEO_ON[;3VEO_ON[;3VEO_ON[;3VEO_ON[;3VEO_ON[;3VEO_ON[;3VEO_ON[;3VEO_ON[;3VEO_ON[;4VEM@NZ;5VEM@NZ;5VEM@NZ;5VEM@NZ;5VEM@NZ;5VEM@NZ;5VEM@NZ;5VEM@NZ;5VEM@OY;4WEM@OY;4WEM@OY;4WEM@OY;4WEM@OY;4WEM@OY;4WEM@OY;4WEN_ONZ;4WEN_ONZ;4WEN_ONZ;4WEN_ONZ;4WEN_ONZ;4WEN_ONZ;4WEN_ONZ;4WEN_ONZ;4WEN_ONZ;4WEN_ONZ;4WEN_ONZ;4WEN_ONZ;4WEN_ONZ;4WEN_ONZ;4WEN_ONZ;4WEN_ON[;3VEO_ON[;3VEO_OOZ;2WEO_OOZ;2WEO_OOZ;2WEO_OOZ;2WEO_OOZ;2WEO_OOZ;2XEN^O0Z;2XEN^O0Z;2XEN^O0Z;2XEN^O0Z;2XEN_OOY;3XEN_OOY;3XEN_OOY;3XEN_OOY;3XEN_OOY;3XEN_OOY;3XEN_OOY;3XEN_OOY;3XEN_OOY;3XEN_OOY;3XEN_OOY;3XEN_OOY;3XENW;2iDMX;3hDLY;4gDKZ;5fDJ[;6eDI\\;7dDH];8cDG^;9bDE`;;h00000000000000000000000000000000000000001O0001O00000000000000O1O1O1O_DIP:5RFKn94RFMn92SFNn90RF1d;0O100O100O1000000000000000000000000000000001O000000000000000000000000000000000001O00000003M2N`e;0`ZD1O1@OdC2[<1bCO_<2_CO`<1aCN_<3`CM`<3`CMa<3ZC1f<;2M in this image.", "objects": [{"patches": [6, 7, 8, 20, 21, 22, 35, 36, 49, 50, 51, 64, 65, 66, 80], "bbox": [0.28899532710280373, 0.0063593750000000004, 0.5645327102803739, 0.22209374999999998], "iscrowd": 0, "area": 7313.022499999998, "rle": {"size": [640, 428], "counts": "cc]25gc04L5M2N3N1N3M2O2M2O2M2O2M2N3N2M4M2M4M2M4M2M4L2O1N2N3L3N2M3M3N2M3N2M3M4M2M3N2M3M3N2M3O1N3N1O1O1O1N2O1N1N3N1O1O2O0100O1O1VN[@YOe?e0i@nNX?Q1VAbNj>]1cAVN^>h1n1O1O1O1O1O1O1O1O1O101N1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1N2N2N2N2N2N2N2N2N2N2N2N2N3M2N2Nlkd3"}, "label": "a man in a black suit sitting"}]} {"id": 1407, "image": "COCO_train2014_000000001407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man to the far left of the guy in the wheelchair\"."}], "task": "refering", "answer_template": "The \"the man to the far left of the guy in the wheelchair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 20, 21, 22, 35, 36, 49, 50, 51, 64, 65, 66, 80], "bbox": [0.28899532710280373, 0.0063593750000000004, 0.5645327102803739, 0.22209374999999998], "iscrowd": 0, "area": 7313.022499999998, "rle": {"size": [640, 428], "counts": "cc]25gc04L5M2N3N1N3M2O2M2O2M2O2M2N3N2M4M2M4M2M4M2M4L2O1N2N3L3N2M3M3N2M3N2M3M4M2M3N2M3M3N2M3O1N3N1O1O1O1N2O1N1N3N1O1O2O0100O1O1VN[@YOe?e0i@nNX?Q1VAbNj>]1cAVN^>h1n1O1O1O1O1O1O1O1O1O101N1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1N2N2N2N2N2N2N2N2N2N2N2N2N3M2N2Nlkd3"}, "label": "the man to the far left of the guy in the wheelchair"}]} {"id": 1407, "image": "COCO_train2014_000000001407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man playing tennis reaching far back to serve\"."}], "task": "refering", "answer_template": "The \"a man playing tennis reaching far back to serve\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 127, 128, 142, 143, 145, 157, 158, 159, 160, 172, 173, 174, 175, 186, 187, 188, 189, 190, 200, 201, 202, 203, 215, 216, 217, 230, 231, 232, 244, 245, 246, 247, 248, 259, 260, 261, 262, 263, 275, 276, 278, 279, 290, 291, 292, 294], "bbox": [0.32469626168224297, 0.31074999999999997, 0.7337616822429907, 0.865609375], "iscrowd": 0, "area": 20449.62315, "rle": {"size": [640, 428], "counts": "\\Zg2b0Qc0e0E3L4L4M3L5K4M3L4M3L4L5L3N3N4L4Z@WMn=l2mAXMQ>l2kAVMS>m2jAUMT>n2iAUMT>m2jATMT>P3jAPMT>T3iAmLT>X3jAhLT>[3kAfLR>^3kAcLR>b3lA^LR>e3mA[LQ>R2hAhN4WOS>Q2jAiN2VOR>Q2oAiNMWOS>o1RBkNJVOR>o1YBiNDYOo=o1cBeN\\O]On=P2kB]OR=d0TCXOi;bAY4`>f0ZAfJb>\\501N100O2N1O2B=N3M5E:J6J7I6J7I5LkMlCdLQc1`N^8o0SF?c1_NZ8o0UFc0b1^NX8l0XFf0b1]NV8j0ZFi0a1_NS8d0^Fn0b1_Nm7`0^F]NROd2e2bNg7:eFbNlNc2i2cNc74lF\\1c1bN^7NSG_1a1dNZ7IYGb1_1dNX7G\\Ge1]1cNW7F^Gf1]1dNT7CbGh1\\1dNR7BdGj1[1dNP7@gGk1[1dNn6_OjGk1[1eNj6^OmGm1[1cNi6]OnGo1\\1bNf6\\OQHQ2\\1bNb6XOWHV2Y1`Na6RO^H\\2T1bNR9]1QGbNn8^1TGbNk8\\1XGdNf8[1]GeNb8Z1`GeN`8Y1cGgN\\8W1gGiNW8W1kGiNS8V1PHjNo7U1SHlNj7S1ZHlNd76gCIg4Ob77hCJh4Ma77hCKj4L^77jCMj4K[77lCNk4IZ77lCOl4IX77mC0l4HW77nC1k4HX74oC3k4HV74PD4k4G\\7MjC in this image.", "objects": [{"patches": [113, 127, 128, 142, 143, 145, 157, 158, 159, 160, 172, 173, 174, 175, 186, 187, 188, 189, 190, 200, 201, 202, 203, 215, 216, 217, 230, 231, 232, 244, 245, 246, 247, 248, 259, 260, 261, 262, 263, 275, 276, 278, 279, 290, 291, 292, 294], "bbox": [0.32469626168224297, 0.31074999999999997, 0.7337616822429907, 0.865609375], "iscrowd": 0, "area": 20449.62315, "rle": {"size": [640, 428], "counts": "\\Zg2b0Qc0e0E3L4L4M3L5K4M3L4M3L4L5L3N3N4L4Z@WMn=l2mAXMQ>l2kAVMS>m2jAUMT>n2iAUMT>m2jATMT>P3jAPMT>T3iAmLT>X3jAhLT>[3kAfLR>^3kAcLR>b3lA^LR>e3mA[LQ>R2hAhN4WOS>Q2jAiN2VOR>Q2oAiNMWOS>o1RBkNJVOR>o1YBiNDYOo=o1cBeN\\O]On=P2kB]OR=d0TCXOi;bAY4`>f0ZAfJb>\\501N100O2N1O2B=N3M5E:J6J7I6J7I5LkMlCdLQc1`N^8o0SF?c1_NZ8o0UFc0b1^NX8l0XFf0b1]NV8j0ZFi0a1_NS8d0^Fn0b1_Nm7`0^F]NROd2e2bNg7:eFbNlNc2i2cNc74lF\\1c1bN^7NSG_1a1dNZ7IYGb1_1dNX7G\\Ge1]1cNW7F^Gf1]1dNT7CbGh1\\1dNR7BdGj1[1dNP7@gGk1[1dNn6_OjGk1[1eNj6^OmGm1[1cNi6]OnGo1\\1bNf6\\OQHQ2\\1bNb6XOWHV2Y1`Na6RO^H\\2T1bNR9]1QGbNn8^1TGbNk8\\1XGdNf8[1]GeNb8Z1`GeN`8Y1cGgN\\8W1gGiNW8W1kGiNS8V1PHjNo7U1SHlNj7S1ZHlNd76gCIg4Ob77hCJh4Ma77hCKj4L^77jCMj4K[77lCNk4IZ77lCOl4IX77mC0l4HW77nC1k4HX74oC3k4HV74PD4k4G\\7MjC in this image.", "objects": [{"patches": [9, 10, 11, 24, 25, 26, 38, 39, 40, 41, 52, 53, 54, 55, 67, 68, 69, 82, 83], "bbox": [0.46380841121495325, 0.0, 0.789696261682243, 0.241546875], "iscrowd": 0, "area": 8922.419399999995, "rle": {"size": [640, 428], "counts": "[`l34hc05K5K5L4M3M2O2M3N2M3N1N2N2O1N2O2N1O1O100O1O1O1O2N1O1O1O1O1O100O1O2N001O1O1O010O1O001O1O1N101N2N2N101N2NO1O2N1O1N3N1M3M4K4M3N32M3M2N3L3N3L3N3L4L2O2N101O001O001B=010O010O010O1O100O100O1O1N2N2N2N2N2N2N2O100O100O100O100O2O0O2O0O101O2M4PNX^Ob1ja0\\NW^Ob1la0[NW^Ob1ma0[NT^Oc1Wb0N1N3M2N3M2N3M3M2N3M2N3M2N3M2M4M2M4L3N]Vh1"}, "label": "the man in the wheelchair"}]} {"id": 1407, "image": "COCO_train2014_000000001407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the wheelchair watching the tennis game\"."}], "task": "refering", "answer_template": "The \"the man in the wheelchair watching the tennis game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 24, 25, 26, 38, 39, 40, 41, 52, 53, 54, 55, 67, 68, 69, 82, 83], "bbox": [0.46380841121495325, 0.0, 0.789696261682243, 0.241546875], "iscrowd": 0, "area": 8922.419399999995, "rle": {"size": [640, 428], "counts": "[`l34hc05K5K5L4M3M2O2M3N2M3N1N2N2O1N2O2N1O1O100O1O1O1O2N1O1O1O1O1O100O1O2N001O1O1O010O1O001O1O1N101N2N2N101N2NO1O2N1O1N3N1M3M4K4M3N32M3M2N3L3N3L3N3L4L2O2N101O001O001B=010O010O010O1O100O100O1O1N2N2N2N2N2N2N2O100O100O100O100O2O0O2O0O101O2M4PNX^Ob1ja0\\NW^Ob1la0[NW^Ob1ma0[NT^Oc1Wb0N1N3M2N3M2N3M3M2N3M2N3M2N3M2M4M2M4L3N]Vh1"}, "label": "the man in the wheelchair watching the tennis game"}]} {"id": 189836, "image": "COCO_train2014_000000189836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tan cow that is hidden behind the white lattice fencing\"."}], "task": "refering", "answer_template": "The \"the tan cow that is hidden behind the white lattice fencing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 266, 286, 287], "bbox": [0.3675625, 0.43244140625, 0.61621875, 0.7162109375], "iscrowd": 0, "area": 11590.92165, "rle": {"size": [512, 640], "counts": "kge35c??J6I7I7I2N0000O100000O100O01000O10000000O1000001O000O10000000000OdAmNm=S1QBROl=n0SBVOj=j0TB[Oj=d0UBAg=?WBFf=:YBJd=5[B0b=0]B4`=Z1L3M3M4L4L4L4L4K5L4L4K3NM3M2N3L4M2N3M3M2N3M3M2N3M3M2M4M3M3M3M3M2N3M3N26J6J7H7J6J6J7I6J6I2O001O00001N10001O00SNUDEk;:iDTOU;m0]EaNc:^1P200O100O010O100O100O010O100O100O010O100O100O010O13M3M4L3L4M4L3M3M3M4K4M3M3M4L8H8G2OM3M3K5I7I7I6J7I7I7I7I7I6J7I7I7I7I7IYij3"}, "label": "the tan cow that is hidden behind the white lattice fencing"}]} {"id": 189836, "image": "COCO_train2014_000000189836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown cow standing behind the fence\"."}], "task": "refering", "answer_template": "The \"a brown cow standing behind the fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 192, 193, 194, 195, 196, 197, 198, 215, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 266, 286, 287], "bbox": [0.3675625, 0.43244140625, 0.61621875, 0.7162109375], "iscrowd": 0, "area": 11590.92165, "rle": {"size": [512, 640], "counts": "kge35c??J6I7I7I2N0000O100000O100O01000O10000000O1000001O000O10000000000OdAmNm=S1QBROl=n0SBVOj=j0TB[Oj=d0UBAg=?WBFf=:YBJd=5[B0b=0]B4`=Z1L3M3M4L4L4L4L4K5L4L4K3NM3M2N3L4M2N3M3M2N3M3M2N3M3M2M4M3M3M3M3M2N3M3N26J6J7H7J6J6J7I6J6I2O001O00001N10001O00SNUDEk;:iDTOU;m0]EaNc:^1P200O100O010O100O100O010O100O100O010O100O100O010O13M3M4L3L4M4L3M3M3M4K4M3M3M4L8H8G2OM3M3K5I7I7I6J7I7I7I7I7I6J7I7I7I7I7IYij3"}, "label": "a brown cow standing behind the fence"}]} {"id": 189836, "image": "COCO_train2014_000000189836.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and brown cow standing slightly behind a fence\"."}], "task": "refering", "answer_template": "The \"a white and brown cow standing slightly behind a fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 255, 256, 257, 259, 278, 279, 280, 281, 282], "bbox": [0.057203125, 0.467890625, 0.32315625, 0.72853515625], "iscrowd": 0, "area": 9884.715649999996, "rle": {"size": [512, 640], "counts": "[hb07i?1O1O1O001O001O0O100000007H4M2N3L=D0O100O10000O10000000000ZChNd:X1UEoNk:Q1nDVOR;j0kDYOU;g0iD[OW;e0hD\\OW;e0gD]OY;c0eD^O\\;c0SDhNCf0ZM4L4L3N3M3N1N3N21N10001N101O0O1000O10O100`CRNn:m1oDVNQ;k1lDXNT;h1hD[NY;e1cD_N\\;b1aDaN_;^1YDjNh;V1lCVOS<[21000O01O1M3N2N2K5J6J5K6M3M3M3N2M3M3N2M3M3O1N1O2000000O10000000000O1000O100000O100000000O10O10O10000O100OhCZNX:f1dE^N\\:b1aEaN_:^1^EfNa:_1XEdNh:_1REdNn:`1jDeNU;_1dDcN];a1[DbNf;a1nCfNT\\O in this image.", "objects": [{"patches": [187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 255, 256, 257, 259, 278, 279, 280, 281, 282], "bbox": [0.057203125, 0.467890625, 0.32315625, 0.72853515625], "iscrowd": 0, "area": 9884.715649999996, "rle": {"size": [512, 640], "counts": "[hb07i?1O1O1O001O001O0O100000007H4M2N3L=D0O100O10000O10000000000ZChNd:X1UEoNk:Q1nDVOR;j0kDYOU;g0iD[OW;e0hD\\OW;e0gD]OY;c0eD^O\\;c0SDhNCf0ZM4L4L3N3M3N1N3N21N10001N101O0O1000O10O100`CRNn:m1oDVNQ;k1lDXNT;h1hD[NY;e1cD_N\\;b1aDaN_;^1YDjNh;V1lCVOS<[21000O01O1M3N2N2K5J6J5K6M3M3M3N2M3M3N2M3M3O1N1O2000000O10000000000O1000O100000O100000000O10O10O10000O100OhCZNX:f1dE^N\\:b1aEaN_:^1^EfNa:_1XEdNh:_1REdNn:`1jDeNU;_1dDcN];a1[DbNf;a1nCfNT\\O in this image.", "objects": [{"patches": [103, 104, 105, 121, 122, 123, 137, 138, 139, 140, 155, 156, 157, 158, 174, 175, 176, 193], "bbox": [0.6536799999999999, 0.45749999999999996, 0.8429, 0.861656626506024], "iscrowd": 0, "area": 7999.883500000005, "rle": {"size": [332, 500], "counts": "]WZ3:R:2N2Mg0ZO1O1O1O1N6K001@iNiGW1V8`000O2O0OO1O2O11O0O2O1O001O1O0O2O1O001O1O001N2O001O1O001N2O001O1O001VOdMaI`2T6V1I1O1N101O1O100000O1N2M3M30000O100000000000000000000000000O1L4K5L4K5L4K5L4L4K5H9F9G9F:G9G9F:GT_i0"}, "label": "a brown sofa chair sitting next to window curtains"}]} {"id": 247181, "image": "COCO_train2014_000000247181.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair on the far right partially obscured by the pole\"."}], "task": "refering", "answer_template": "The \"the chair on the far right partially obscured by the pole\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 121, 122, 123, 137, 138, 139, 140, 155, 156, 157, 158, 174, 175, 176, 193], "bbox": [0.6536799999999999, 0.45749999999999996, 0.8429, 0.861656626506024], "iscrowd": 0, "area": 7999.883500000005, "rle": {"size": [332, 500], "counts": "]WZ3:R:2N2Mg0ZO1O1O1O1N6K001@iNiGW1V8`000O2O0OO1O2O11O0O2O1O001O1O0O2O1O001O1O001N2O001O1O001N2O001O1O001VOdMaI`2T6V1I1O1N101O1O100000O1N2M3M30000O100000000000000000000000000O1L4K5L4K5L4K5L4L4K5H9F9G9F:G9G9F:GT_i0"}, "label": "the chair on the far right partially obscured by the pole"}]} {"id": 247181, "image": "COCO_train2014_000000247181.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a leather couch with a plaid throw on the back\"."}], "task": "refering", "answer_template": "The \"a leather couch with a plaid throw on the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 180, 181, 182, 183, 184, 185, 198, 199, 200, 201, 202, 203], "bbox": [0.0, 0.5483734939759036, 0.33127999999999996, 1.0], "iscrowd": 0, "area": 19761.91775, "rle": {"size": [332, 500], "counts": "l6j0a9c1]NU1lN0O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O100O1O100O100O1O100O100O1O100O100O1O100O100O100O1O1002N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O000000000000O10000000000000000000000000000O100000000000000000000000000O1000000000000000000000000O1000000000000000000000000000000001O000QMYIY2g6eM\\IZ2d6dM_I[2b6bMaI]2_6aMdI^2]6_MfI`2Z6^MiIa2X6\\MkIc2U6[MnId2o6N1O1O2N1SOn0XOh0XOlQ\\3"}, "label": "a leather couch with a plaid throw on the back"}]} {"id": 247181, "image": "COCO_train2014_000000247181.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown leather sofa with a brown , red , and white blanket laying on the back of it\"."}], "task": "refering", "answer_template": "The \"a brown leather sofa with a brown , red , and white blanket laying on the back of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 180, 181, 182, 183, 184, 185, 198, 199, 200, 201, 202, 203], "bbox": [0.0, 0.5483734939759036, 0.33127999999999996, 1.0], "iscrowd": 0, "area": 19761.91775, "rle": {"size": [332, 500], "counts": "l6j0a9c1]NU1lN0O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O100O1O100O100O1O100O100O1O100O100O1O100O100O100O1O1002N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O000000000000O10000000000000000000000000000O100000000000000000000000000O1000000000000000000000000O1000000000000000000000000000000001O000QMYIY2g6eM\\IZ2d6dM_I[2b6bMaI]2_6aMdI^2]6_MfI`2Z6^MiIa2X6\\MkIc2U6[MnId2o6N1O1O2N1SOn0XOh0XOlQ\\3"}, "label": "a brown leather sofa with a brown , red , and white blanket laying on the back of it"}]} {"id": 468373, "image": "COCO_train2014_000000468373.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown teddy bear along with a black teddy bear\"."}], "task": "refering", "answer_template": "The \"the brown teddy bear along with a black teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 55, 56, 57, 58, 59, 72, 73, 74, 75, 76, 77, 89, 90, 91, 92, 93, 94, 104, 105, 106, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125, 126, 127, 128, 137, 138, 139, 140, 141, 142, 143, 144, 154, 155, 156, 157, 158, 159, 160, 161, 162, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 205, 206, 207, 208, 209, 210, 211, 212, 213, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 262, 273, 274, 275, 276, 277, 278, 279, 280, 289, 290, 291, 292, 293, 294, 295, 296, 297, 306, 307, 308, 309, 310, 312, 313, 314, 323, 324, 325, 326, 327, 341, 342], "bbox": [0.03895833333333333, 0.083140625, 0.6172291666666666, 0.8876406250000001], "iscrowd": 0, "area": 85063.01365000001, "rle": {"size": [640, 480], "counts": "o[8d5[5nI[J`09c5[5nIYJb0:b5\\5lIYJd09a5^5lIVJf0:`5`5jISJi0<^5`5jIRJj0<^5a5iIPJl0=\\5d5hIlIo0>[5e5gIkIP1>[5f5fIiIR1=]5[2TISO>RNU1;]5^2SITO;QNW19`5`2QISO9QN[16_5d2PISO7QN]12b5g2mHUO4oMb1Oa5k2mHTO1oMf1La5o2jHTO0oMi1Gb5T3iHSOMoMm1Db5W3gHUOJmMR2Aa5[3fHUOX3[MW4^3dHTOg;j0\\DUOe;h0_DUOb;i0aDUOa;i0aDUO`;i0aDXO_;f0cDYO^;e0cD[O_;a0cD@];>dDB];;fDD[;9gDG[;5gDLY;2hDNY;OjD0W;MkD4V;HlD8U;EmD;T;CnD;T;BnD>\\;SOiDl0VO0O1O100O2N1O100O2N100O1O2O0O1O101N1O4L5L3L4L4L5J5L4K5L5K4K5LZfb3"}, "label": "the brown teddy bear along with a black teddy bear"}]} {"id": 468373, "image": "COCO_train2014_000000468373.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light brown teddy bear sitting next to a dark brown teddy bear on a chair\"."}], "task": "refering", "answer_template": "The \"a light brown teddy bear sitting next to a dark brown teddy bear on a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 55, 56, 57, 58, 59, 72, 73, 74, 75, 76, 77, 89, 90, 91, 92, 93, 94, 104, 105, 106, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125, 126, 127, 128, 137, 138, 139, 140, 141, 142, 143, 144, 154, 155, 156, 157, 158, 159, 160, 161, 162, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 205, 206, 207, 208, 209, 210, 211, 212, 213, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 262, 273, 274, 275, 276, 277, 278, 279, 280, 289, 290, 291, 292, 293, 294, 295, 296, 297, 306, 307, 308, 309, 310, 312, 313, 314, 323, 324, 325, 326, 327, 341, 342], "bbox": [0.03895833333333333, 0.083140625, 0.6172291666666666, 0.8876406250000001], "iscrowd": 0, "area": 85063.01365000001, "rle": {"size": [640, 480], "counts": "o[8d5[5nI[J`09c5[5nIYJb0:b5\\5lIYJd09a5^5lIVJf0:`5`5jISJi0<^5`5jIRJj0<^5a5iIPJl0=\\5d5hIlIo0>[5e5gIkIP1>[5f5fIiIR1=]5[2TISO>RNU1;]5^2SITO;QNW19`5`2QISO9QN[16_5d2PISO7QN]12b5g2mHUO4oMb1Oa5k2mHTO1oMf1La5o2jHTO0oMi1Gb5T3iHSOMoMm1Db5W3gHUOJmMR2Aa5[3fHUOX3[MW4^3dHTOg;j0\\DUOe;h0_DUOb;i0aDUOa;i0aDUO`;i0aDXO_;f0cDYO^;e0cD[O_;a0cD@];>dDB];;fDD[;9gDG[;5gDLY;2hDNY;OjD0W;MkD4V;HlD8U;EmD;T;CnD;T;BnD>\\;SOiDl0VO0O1O100O2N1O100O2N100O1O2O0O1O101N1O4L5L3L4L4L5J5L4K5L5K4K5LZfb3"}, "label": "a light brown teddy bear sitting next to a dark brown teddy bear on a chair"}]} {"id": 468373, "image": "COCO_train2014_000000468373.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark brown teddy bear\"."}], "task": "refering", "answer_template": "The \"a dark brown teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 42, 43, 44, 45, 46, 47, 48, 59, 60, 61, 62, 63, 64, 65, 77, 78, 79, 80, 81, 82, 94, 95, 96, 97, 98, 111, 112, 113, 114, 115, 116, 127, 128, 129, 130, 131, 132, 133, 134, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 180, 181, 182, 183, 184, 185, 186, 196, 197, 198, 199, 200, 201, 202, 203, 212, 213, 214, 215, 216, 217, 218, 219, 220, 229, 230, 231, 232, 233, 234, 235, 236, 237, 246, 247, 248, 249, 250, 253, 254, 263, 264, 265, 266, 267, 270, 271, 280, 281, 282, 283, 287, 288, 299], "bbox": [0.47147916666666667, 0.07320312500000001, 1.0, 0.7421093750000001], "iscrowd": 0, "area": 70055.1109, "rle": {"size": [640, 480], "counts": "]Z]41nc05L4VGIbM=X2GeM=V2GfM>U2FgM>U2FhM:W2JeM8ASOUKi0V56jKQO14kNi0W52lKTO04iNj0W50nKUO04gNk0W5MQLVO04eNm0W5JRLXO04cNn0W5GULZOO4aNl0[5GSL\\ON5`Ni0^5GSL^OM5^Ng0a5HRL_OM5\\Ne0e5HPLAM4[Nd0g5HPLCL4YNb0j5IoKDK5XN?n5HnKGJ5VN=Q6ImKHJ5SN`6CjKOH4hM=f6AhK1H4dM=k6@gK2H^2`4aMgK3H\\2a4cMdK4IY2b4kM]KOOV2c4SNVKI5U2e4XNoJDN1N3N1I8[Oe0[Od0]Od0[Od0\\OoD"}, "label": "a dark brown teddy bear"}]} {"id": 468373, "image": "COCO_train2014_000000468373.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark brown bear\"."}], "task": "refering", "answer_template": "The \"a dark brown bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 42, 43, 44, 45, 46, 47, 48, 59, 60, 61, 62, 63, 64, 65, 77, 78, 79, 80, 81, 82, 94, 95, 96, 97, 98, 111, 112, 113, 114, 115, 116, 127, 128, 129, 130, 131, 132, 133, 134, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 180, 181, 182, 183, 184, 185, 186, 196, 197, 198, 199, 200, 201, 202, 203, 212, 213, 214, 215, 216, 217, 218, 219, 220, 229, 230, 231, 232, 233, 234, 235, 236, 237, 246, 247, 248, 249, 250, 253, 254, 263, 264, 265, 266, 267, 270, 271, 280, 281, 282, 283, 287, 288, 299], "bbox": [0.47147916666666667, 0.07320312500000001, 1.0, 0.7421093750000001], "iscrowd": 0, "area": 70055.1109, "rle": {"size": [640, 480], "counts": "]Z]41nc05L4VGIbM=X2GeM=V2GfM>U2FgM>U2FhM:W2JeM8ASOUKi0V56jKQO14kNi0W52lKTO04iNj0W50nKUO04gNk0W5MQLVO04eNm0W5JRLXO04cNn0W5GULZOO4aNl0[5GSL\\ON5`Ni0^5GSL^OM5^Ng0a5HRL_OM5\\Ne0e5HPLAM4[Nd0g5HPLCL4YNb0j5IoKDK5XN?n5HnKGJ5VN=Q6ImKHJ5SN`6CjKOH4hM=f6AhK1H4dM=k6@gK2H^2`4aMgK3H\\2a4cMdK4IY2b4kM]KOOV2c4SNVKI5U2e4XNoJDN1N3N1I8[Oe0[Od0]Od0[Od0\\OoD"}, "label": "a dark brown bear"}]} {"id": 550308, "image": "COCO_train2014_000000550308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sausage on a roll on the far right of the plate\"."}], "task": "refering", "answer_template": "The \"the sausage on a roll on the far right of the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 387], "bbox": [0.606734375, 0.4134791666666667, 1.0, 0.9528125000000001], "iscrowd": 0, "area": 41935.945199999995, "rle": {"size": [480, 640], "counts": "Rcf56f>7I6K5J6J6J6K5J6J6J7K4M3N2N2N2M3N2N2M5L5K4L4K6K4L4K6K4L2N2M3N2N2M3N2N2N2M3N2N2M3N2N2N2M3N2N2M2O2N2N2M3N2N2M3N2N2N2M3N2N2M3N2N2N2M3N2N2M101O1O0O2O1O0O2O1O001N200O00100O00100O010O1O010O1O010O1O010O1O010O1O010O1O010O1O01O00010O000010O0001O00010O000010O0001O01O01O000010O0001O01O01O00010O00001O01O01O00010O001O000O2N1O101N1O1O2O0O1O2N100O2N1O101N2N2N3N2M3M2N3N2M3M3M2O2M3M3M2O2M3M3M3N1N3M3M3N2M3M2N3M3N2M3M3M3N2M2N3M3N2M3M3M3N1N3M3M3N2M3M3M3N1N3M3M3N2M3M3M2O2M3M3M3N2M3M3M2O2M3M3M3N2M3M2N3N2M3M3M3N2M3M2N[D"}, "label": "the sausage on a roll on the far right of the plate"}]} {"id": 550308, "image": "COCO_train2014_000000550308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hot dog in burger\"."}], "task": "refering", "answer_template": "The \"hot dog in burger\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 387], "bbox": [0.606734375, 0.4134791666666667, 1.0, 0.9528125000000001], "iscrowd": 0, "area": 41935.945199999995, "rle": {"size": [480, 640], "counts": "Rcf56f>7I6K5J6J6J6K5J6J6J7K4M3N2N2N2M3N2N2M5L5K4L4K6K4L4K6K4L2N2M3N2N2M3N2N2N2M3N2N2M3N2N2N2M3N2N2M2O2N2N2M3N2N2M3N2N2N2M3N2N2M3N2N2N2M3N2N2M101O1O0O2O1O0O2O1O001N200O00100O00100O010O1O010O1O010O1O010O1O010O1O010O1O010O1O01O00010O000010O0001O00010O000010O0001O01O01O000010O0001O01O01O00010O00001O01O01O00010O001O000O2N1O101N1O1O2O0O1O2N100O2N1O101N2N2N3N2M3M2N3N2M3M3M2O2M3M3M2O2M3M3M3N1N3M3M3N2M3M2N3M3N2M3M3M3N2M2N3M3N2M3M3M3N1N3M3M3N2M3M3M3N1N3M3M3N2M3M3M2O2M3M3M3N2M3M3M2O2M3M3M3N2M3M2N3N2M3M3M3N2M3M2N[D"}, "label": "hot dog in burger"}]} {"id": 550308, "image": "COCO_train2014_000000550308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"burger in a plate with cheese\"."}], "task": "refering", "answer_template": "The \"burger in a plate with cheese\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 48, 49, 50, 51, 52, 70, 71, 72, 73, 74, 75, 76, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 162, 163, 164, 165], "bbox": [0.032015625, 0.09625, 0.313484375, 0.44906250000000003], "iscrowd": 0, "area": 22382.360299999993, "rle": {"size": [480, 640], "counts": "ga91i>7H8H8L4N2N2fNSO]DP1a;UOZDm0d;YOUDj0i;\\OQDf0m;@lCc0SA?Ale]6"}, "label": "burger in a plate with cheese"}]} {"id": 550308, "image": "COCO_train2014_000000550308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"first sandwich on the left just beneath the fork\"."}], "task": "refering", "answer_template": "The \"first sandwich on the left just beneath the fork\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 48, 49, 50, 51, 52, 70, 71, 72, 73, 74, 75, 76, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 162, 163, 164, 165], "bbox": [0.032015625, 0.09625, 0.313484375, 0.44906250000000003], "iscrowd": 0, "area": 22382.360299999993, "rle": {"size": [480, 640], "counts": "ga91i>7H8H8L4N2N2fNSO]DP1a;UOZDm0d;YOUDj0i;\\OQDf0m;@lCc0SA?Ale]6"}, "label": "first sandwich on the left just beneath the fork"}]} {"id": 550308, "image": "COCO_train2014_000000550308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a breakfast sandwich with egg\"."}], "task": "refering", "answer_template": "The \"a breakfast sandwich with egg\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241], "bbox": [0.20973437499999997, 0.25535416666666666, 0.585578125, 0.6513333333333333], "iscrowd": 0, "area": 33025.87994999999, "rle": {"size": [480, 640], "counts": "con11o>4L4K4M4L4K5L3L3N1O1N01000^DQOR9P1hFXOoMGP;Q1kFEU9;fFJY97bEQO4n0Z:1aETONP1a:L`EVOMo0b:LaEVOLn0c:L`EXOLl0d:L_E[OKi0e:L`E]OIh0g:K_E_OIf0h:K_E@Hf0g:K`E@Hf0h:J_EBGe0j:I^ECGe0j:I_ECDf0m:G^EDDf0n:F]EW1b:jN\\EX1d:hN[EY1e:fNZE\\1e:eNZE\\1f:fNVE\\1i:fNUE\\1j:]100O10000O10000O10001N10000O10000EfKoE[4Q:gKmEY4S:iKkEW4T:lKiEU4W::O2O000O10000O10000O10000O2O000O100O100O100O100O100O100O100O100O100O1O100O100O10000O10000O1000000O10000O01000O01000O0100O10O10O10O10001O2N1O1O1O100O1O2N1O100O1O000001O0O10000O2O00000O100O2O0O1O100O2O0O1O101N100O2N101N100O2N100O2O0O1O2O0O101O0O101N10000O10000O100O1O1O2N1O100O1O1O1O1O100O1O2N1O001O10O01[MZEP1f:nN\\ER1d:lN^ET1b:jNaEU1_:hNdEX1]:eNeER1CPNg:m0hEP1FPNb:o0iEP1InM_:P1jEP1d:oN]EP1d:PO\\Eo0f:oN[EP1f:oN[EP1f:oNZEQ1h:mNYEQ1i:oNWEP1j:oNVEQ1k:mNWER1k:kNWET1k:iNVEW1R;_NQE`1Y in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241], "bbox": [0.20973437499999997, 0.25535416666666666, 0.585578125, 0.6513333333333333], "iscrowd": 0, "area": 33025.87994999999, "rle": {"size": [480, 640], "counts": "con11o>4L4K4M4L4K5L3L3N1O1N01000^DQOR9P1hFXOoMGP;Q1kFEU9;fFJY97bEQO4n0Z:1aETONP1a:L`EVOMo0b:LaEVOLn0c:L`EXOLl0d:L_E[OKi0e:L`E]OIh0g:K_E_OIf0h:K_E@Hf0g:K`E@Hf0h:J_EBGe0j:I^ECGe0j:I_ECDf0m:G^EDDf0n:F]EW1b:jN\\EX1d:hN[EY1e:fNZE\\1e:eNZE\\1f:fNVE\\1i:fNUE\\1j:]100O10000O10000O10001N10000O10000EfKoE[4Q:gKmEY4S:iKkEW4T:lKiEU4W::O2O000O10000O10000O10000O2O000O100O100O100O100O100O100O100O100O100O1O100O100O10000O10000O1000000O10000O01000O01000O0100O10O10O10O10001O2N1O1O1O100O1O2N1O100O1O000001O0O10000O2O00000O100O2O0O1O100O2O0O1O101N100O2N101N100O2N100O2O0O1O2O0O101O0O101N10000O10000O100O1O1O2N1O100O1O1O1O1O100O1O2N1O001O10O01[MZEP1f:nN\\ER1d:lN^ET1b:jNaEU1_:hNdEX1]:eNeER1CPNg:m0hEP1FPNb:o0iEP1InM_:P1jEP1d:oN]EP1d:PO\\Eo0f:oN[EP1f:oN[EP1f:oNZEQ1h:mNYEQ1i:oNWEP1j:oNVEQ1k:mNWER1k:kNWET1k:iNVEW1R;_NQE`1Y in this image.", "objects": [{"patches": [135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 251, 252, 274, 275, 298], "bbox": [0.8764062499999999, 0.3236041666666667, 1.0, 0.7685416666666666], "iscrowd": 0, "area": 11443.400999999994, "rle": {"size": [480, 640], "counts": "_VW86Z>a0XOi0WOh0_Ob0N2N1N3N1O2N2N1O2N1O2N2N1N3N2N1O2N1O2N2N1O2M2O2N2N1O2N2N1O2N1O2M3N1O2N1O2N2N1O2N1N3N2N1O2N2N1O2N1O2N2M201N1O2N2O0O2N2N101N1O2O1N1O2N101N2N101N2N1O2OjKSO"}, "label": "the top piece of bread for the sandwich furthest to the right"}]} {"id": 550308, "image": "COCO_train2014_000000550308.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top bun of the sandwich on the right\"."}], "task": "refering", "answer_template": "The \"the top bun of the sandwich on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 251, 252, 274, 275, 298], "bbox": [0.8764062499999999, 0.3236041666666667, 1.0, 0.7685416666666666], "iscrowd": 0, "area": 11443.400999999994, "rle": {"size": [480, 640], "counts": "_VW86Z>a0XOi0WOh0_Ob0N2N1N3N1O2N2N1O2N1O2N2N1N3N2N1O2N1O2N2N1O2M2O2N2N1O2N2N1O2N1O2M3N1O2N1O2N2N1O2N1N3N2N1O2N2N1O2N1O2N2M201N1O2N2O0O2N2N101N1O2O1N1O2N101N2N101N2N1O2OjKSO"}, "label": "the top bun of the sandwich on the right"}]} {"id": 50601, "image": "COCO_train2014_000000050601.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black dog sitting outside a chainlinked fence\"."}], "task": "refering", "answer_template": "The \"a black dog sitting outside a chainlinked fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250, 251], "bbox": [0.565234375, 0.5056206088992974, 0.9370625, 0.7236065573770493], "iscrowd": 0, "area": 15984.080100000003, "rle": {"size": [427, 640], "counts": "dWg43S=6J5K5M3N2N2N2N2N2N2N2N2N2O1O1O2N1O100O1O1O1O1O1O1O100O100O100O10000O010O100O100O01000O100O100O0100000O1000O100000O10000000000O1000O100000O10000000000O10O100000O010000O1000O10O1000000O0100000O01000O10O1000000O1000000O10000000001N10001O00001O0O1000>B1O000O2O00001O00001O000O2O000000O10000000O1000O100000000000O01000000000000000O1000000000000000000O100000O10001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O0O100O2O000O2O0O10002M2O2N1N3N1N3N1O2M2M3M4M2M4L3N3L3M4ZO^^`0"}, "label": "a black dog sitting outside a chainlinked fence"}]} {"id": 50601, "image": "COCO_train2014_000000050601.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black animal right next to a fencepost\"."}], "task": "refering", "answer_template": "The \"a black animal right next to a fencepost\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250, 251], "bbox": [0.565234375, 0.5056206088992974, 0.9370625, 0.7236065573770493], "iscrowd": 0, "area": 15984.080100000003, "rle": {"size": [427, 640], "counts": "dWg43S=6J5K5M3N2N2N2N2N2N2N2N2N2O1O1O2N1O100O1O1O1O1O1O1O100O100O100O10000O010O100O100O01000O100O100O0100000O1000O100000O10000000000O1000O100000O10000000000O10O100000O010000O1000O10O1000000O0100000O01000O10O1000000O1000000O10000000001N10001O00001O0O1000>B1O000O2O00001O00001O000O2O000000O10000000O1000O100000000000O01000000000000000O1000000000000000000O100000O10001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O0O100O2O000O2O0O10002M2O2N1N3N1N3N1O2M2M3M4M2M4L3N3L3M4ZO^^`0"}, "label": "a black animal right next to a fencepost"}]} {"id": 50601, "image": "COCO_train2014_000000050601.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"water tub in zoo\"."}], "task": "refering", "answer_template": "The \"water tub in zoo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 300, 301, 302, 303, 324, 325, 326], "bbox": [0.0531875, 0.6778454332552694, 0.47748437500000007, 0.9789695550351287], "iscrowd": 0, "area": 15630.814399999994, "rle": {"size": [427, 640], "counts": "j_>1P=3WCOd<9WCHe in this image.", "objects": [{"patches": [231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 300, 301, 302, 303, 324, 325, 326], "bbox": [0.0531875, 0.6778454332552694, 0.47748437500000007, 0.9789695550351287], "iscrowd": 0, "area": 15630.814399999994, "rle": {"size": [427, 640], "counts": "j_>1P=3WCOd<9WCHe in this image.", "objects": [{"patches": [59, 60, 61, 62, 72, 73, 74, 75, 76, 85, 86, 87, 88, 89, 90, 91, 99, 100, 101, 102, 103, 104, 105, 113, 114, 115, 116, 117, 118, 119, 127, 128, 129, 130, 131, 132, 141, 142, 143, 144, 145, 146, 155, 156, 157, 158, 159, 168, 169, 170, 171, 172, 173, 174, 177, 178, 179, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 252, 253, 254, 255, 256, 257, 258, 259, 266, 267, 268, 269, 270], "bbox": [0.026975, 0.18651666666666666, 0.85955, 0.9483166666666667], "iscrowd": 0, "area": 82903.39975000001, "rle": {"size": [600, 400], "counts": "ll6Q3f`5]4jIPKk0`0\\5^4mImJk0c0Y5^4PJiJm0g0S5^4UJfJl0i0P5_4XJdJk0k0n4a4YJ`Jn0k0j4e4ZJ]Jo0k0h4g4\\JZJP1l0e4j4^JUJQ1n0b4m4_JQJT1n0^4Q5`JnIV1l0\\4V5`JjIZ1i0Y4\\5`JgI\\1d0Z4e5[JdIa1<[4P6TJbIg13\\4[6mIaI]9_6^100O2O0000000O1000000O1O1O1O100O1O1O100O1O1O100O1O1O1O100O2N1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O2N1O1O100O1O1O100000000000O1O1N2O1N2O1N2O1N2O1O1N2O1N2N2O1N2O1N2O1N2O1N2O1O00100000000O100000000OO2O1O1O1O001N2O1O1O1ObIWJNi50^JLa53fJHX58nJDQ5:VKBi4=^K]Ob4b0dKZO\\4c0lKXOT4f0RLVOn3h0XLTOh3i0[LVOf3h0]LWOc3f0`LZO`3d0cL[O]3c0eL]O[3`0iL_OW3?kL@W3=lLBT3;PMDQ39QMGo26UMIl24VMKk24WMKj22XMNh21ZMNg2O\\M0e2N\\M1e2N]M1d2L^M4b2K`M4a2PLUJm2\\3S1_2kK^Jm2U3V1_2fKfJn2n2[1\\2bKoJn2f2`1f3ZN]Le1g3VNZLj1i3RNYLm1k3mMWLR2n3iMTLV2P4dMSL[2Q4`MPL`2T4ZMnKf2U4VMlKi2Y4aLYL_3`9000000O10000000001N100000000O100000000O100000000O10000000O10O10O10000O010O10000O01O1O001O0O2O1O001O1N101O1O0O2O001O1N101O1O0O2O1O001N101O00001N1000001N10001O0O10001O0O0100000O1000000O100000000O01002N1N3N1O2L3M4L3L5L4L3M4K4M4L3M4M3M2O2M2O2M2O2M2N3N2M2O2M2O2M2O2M2N3N1N3N2M2O_QQ1"}, "label": "the smaller polar bear on the left"}]} {"id": 181681, "image": "COCO_train2014_000000181681.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby polar bear\"."}], "task": "refering", "answer_template": "The \"a baby polar bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 72, 73, 74, 75, 76, 85, 86, 87, 88, 89, 90, 91, 99, 100, 101, 102, 103, 104, 105, 113, 114, 115, 116, 117, 118, 119, 127, 128, 129, 130, 131, 132, 141, 142, 143, 144, 145, 146, 155, 156, 157, 158, 159, 168, 169, 170, 171, 172, 173, 174, 177, 178, 179, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 252, 253, 254, 255, 256, 257, 258, 259, 266, 267, 268, 269, 270], "bbox": [0.026975, 0.18651666666666666, 0.85955, 0.9483166666666667], "iscrowd": 0, "area": 82903.39975000001, "rle": {"size": [600, 400], "counts": "ll6Q3f`5]4jIPKk0`0\\5^4mImJk0c0Y5^4PJiJm0g0S5^4UJfJl0i0P5_4XJdJk0k0n4a4YJ`Jn0k0j4e4ZJ]Jo0k0h4g4\\JZJP1l0e4j4^JUJQ1n0b4m4_JQJT1n0^4Q5`JnIV1l0\\4V5`JjIZ1i0Y4\\5`JgI\\1d0Z4e5[JdIa1<[4P6TJbIg13\\4[6mIaI]9_6^100O2O0000000O1000000O1O1O1O100O1O1O100O1O1O100O1O1O1O100O2N1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O2N1O1O100O1O1O100000000000O1O1N2O1N2O1N2O1N2O1O1N2O1N2N2O1N2O1N2O1N2O1N2O1O00100000000O100000000OO2O1O1O1O001N2O1O1O1ObIWJNi50^JLa53fJHX58nJDQ5:VKBi4=^K]Ob4b0dKZO\\4c0lKXOT4f0RLVOn3h0XLTOh3i0[LVOf3h0]LWOc3f0`LZO`3d0cL[O]3c0eL]O[3`0iL_OW3?kL@W3=lLBT3;PMDQ39QMGo26UMIl24VMKk24WMKj22XMNh21ZMNg2O\\M0e2N\\M1e2N]M1d2L^M4b2K`M4a2PLUJm2\\3S1_2kK^Jm2U3V1_2fKfJn2n2[1\\2bKoJn2f2`1f3ZN]Le1g3VNZLj1i3RNYLm1k3mMWLR2n3iMTLV2P4dMSL[2Q4`MPL`2T4ZMnKf2U4VMlKi2Y4aLYL_3`9000000O10000000001N100000000O100000000O100000000O10000000O10O10O10000O010O10000O01O1O001O0O2O1O001O1N101O1O0O2O001O1N101O1O0O2O1O001N101O00001N1000001N10001O0O10001O0O0100000O1000000O100000000O01002N1N3N1O2L3M4L3L5L4L3M4K4M4L3M4M3M2O2M2O2M2O2M2N3N2M2O2M2O2M2O2M2N3N1N3N2M2O_QQ1"}, "label": "a baby polar bear"}]} {"id": 181681, "image": "COCO_train2014_000000181681.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mother white polar bear with her baby nuzzling her\"."}], "task": "refering", "answer_template": "The \"a mother white polar bear with her baby nuzzling her\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 22, 23, 24, 25, 26, 27, 35, 36, 37, 38, 39, 40, 41, 49, 50, 51, 52, 53, 54, 55, 62, 63, 64, 65, 66, 67, 68, 69, 76, 77, 78, 79, 80, 81, 82, 83, 91, 92, 93, 94, 95, 96, 97, 105, 106, 107, 108, 109, 110, 111, 118, 119, 120, 121, 122, 123, 124, 125, 132, 133, 134, 135, 136, 137, 138, 139, 146, 147, 148, 149, 150, 151, 152, 153, 159, 160, 161, 162, 163, 164, 165, 166, 167, 174, 175, 176, 177, 178, 179, 180, 181, 188, 189, 190, 191, 193, 194, 195, 206, 207, 208, 209, 220, 221, 222, 223, 233, 234, 235, 236, 237, 246, 247, 248, 249, 250, 251, 257, 258, 259, 260, 261, 262, 263, 264, 265, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279], "bbox": [0.28315, 0.002666666666666667, 0.9573, 0.9528166666666668], "iscrowd": 0, "area": 91496.85840000003, "rle": {"size": [600, 400], "counts": "chR21fb01O1N2O1O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O001O1O1N2O1O1O1O1O1O100O1O1O100O1O100O1000000O1000000O100mIfNhJZ1Q5oNmJQ1Q5SOmJm0P5XOmJi0Q5\\OlJd0R5@lJ`0Q5EmJ;Q5ImJ7Q5NlJ2Q53lJNR56lJJR5:lJFR5?kJAR5d0lJ\\OR5h0kJYOS5k0kJUOR5Q1kJoNS5U1kJkNS5W1mJiNP5[1oJeNo4]1PKdNn4^1RKbNk4a1UK_Ni4d1VK\\Nh4f1XKZNe4i1[KWNc4k1\\KVNb4m1]KSNm0_NSNa3P1PNj0fNPN\\3V1nMg0nNlMV3]1lMd0UOhMS3c1hMa0]OfMm2h1gM?DbMg2o1eMhH@Z7`0fH^O\\7b0dH\\O^7c0cH[O`7d0`HZOb7f0^HXOd7h0\\HVOf7j0ZHTOi7k0WHSOk7l0VHROl7n0THPOn7P1RHnNQ8Q1oGmNS8S1mGkNU8T1lGjNV8V1jGhNY8W1gGgN[8Y1eGeN]8[1cGcN_8]1aGaNa8^1`G`Na8a1_G]Nc8c1]G[Ne8e1[GYNg8g1YGWNh8i1YGUNi8j1XGTNj8l1VGRNk8n1VGQNk8n1VGQNk8o1UGPNl8o1UGPNk8Q2UGnMl8Q2UGnMl8Q2UGoMk8Q2UGnMk8R2VGmMk8S2UGlMl8S2UGlMk8T2VGkMk8U2UGjMl8U2UGkMk8T2VGjMk8V2VGiMk8V2VGiMk8U2WGjMi8V2XGiMi8V2XGiMi8V2XGiMi8U2YGjMg8V2ZGhMh8W2YGhMh8W2YGhMh8W2YGhMg8W2[GhMf8W2[GhMf8W2[GhMe8X2\\GfMf8X2\\GgMe8X2\\GgMe8X2\\GgMd8Y2]GfMd8Y2]GfMb8Z2`GeM`8[2aGeM^8[2cGeM\\8[2eGeMZ8Z2hGfMW8Z2jGfMU8Z2lGfMR8[2oGeMP8[2QHeMn7Z2THfMk7Z2VHfMi7Z2XHfMg7Z2ZHfMe7Y2]HgMb7Y2_HgM_7Z2bHfM]7Y2eHgMZ7Y2gHgMX7X2jHiMT7W2mHiMR7W2oHiMP7V2RIjMn6U2SIkMl6T2VIlMj6S2WImMh6R2ZInMf6Q2[IoMd6P2^IPNa6P2`IPN`6n1bIRN]6n1dIRN\\6l1fITNY6l1hITNX6j1jIVNV6i1kIWNU6g1mIYNS6e1oI[NR6b1PJ_No5_1SJaNm5]1UJcNk5[1WJeNi5X1ZJhNf5V1\\JjNd5T1^JlNb5R1`JnN`5P1bJPO^5m0eJSO[5k0gJUOZ5h0hJXOY5e0iJ[OW5c0kJ]OV5`0lJ@U5=mJCT5:nJFR59oJGR56PKJQ53QKNn40TK0m4NTK2m4KUK5l4HVK8j4FXK:i4DXK in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 22, 23, 24, 25, 26, 27, 35, 36, 37, 38, 39, 40, 41, 49, 50, 51, 52, 53, 54, 55, 62, 63, 64, 65, 66, 67, 68, 69, 76, 77, 78, 79, 80, 81, 82, 83, 91, 92, 93, 94, 95, 96, 97, 105, 106, 107, 108, 109, 110, 111, 118, 119, 120, 121, 122, 123, 124, 125, 132, 133, 134, 135, 136, 137, 138, 139, 146, 147, 148, 149, 150, 151, 152, 153, 159, 160, 161, 162, 163, 164, 165, 166, 167, 174, 175, 176, 177, 178, 179, 180, 181, 188, 189, 190, 191, 193, 194, 195, 206, 207, 208, 209, 220, 221, 222, 223, 233, 234, 235, 236, 237, 246, 247, 248, 249, 250, 251, 257, 258, 259, 260, 261, 262, 263, 264, 265, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279], "bbox": [0.28315, 0.002666666666666667, 0.9573, 0.9528166666666668], "iscrowd": 0, "area": 91496.85840000003, "rle": {"size": [600, 400], "counts": "chR21fb01O1N2O1O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O001O1O1N2O1O1O1O1O1O100O1O1O100O1O100O1000000O1000000O100mIfNhJZ1Q5oNmJQ1Q5SOmJm0P5XOmJi0Q5\\OlJd0R5@lJ`0Q5EmJ;Q5ImJ7Q5NlJ2Q53lJNR56lJJR5:lJFR5?kJAR5d0lJ\\OR5h0kJYOS5k0kJUOR5Q1kJoNS5U1kJkNS5W1mJiNP5[1oJeNo4]1PKdNn4^1RKbNk4a1UK_Ni4d1VK\\Nh4f1XKZNe4i1[KWNc4k1\\KVNb4m1]KSNm0_NSNa3P1PNj0fNPN\\3V1nMg0nNlMV3]1lMd0UOhMS3c1hMa0]OfMm2h1gM?DbMg2o1eMhH@Z7`0fH^O\\7b0dH\\O^7c0cH[O`7d0`HZOb7f0^HXOd7h0\\HVOf7j0ZHTOi7k0WHSOk7l0VHROl7n0THPOn7P1RHnNQ8Q1oGmNS8S1mGkNU8T1lGjNV8V1jGhNY8W1gGgN[8Y1eGeN]8[1cGcN_8]1aGaNa8^1`G`Na8a1_G]Nc8c1]G[Ne8e1[GYNg8g1YGWNh8i1YGUNi8j1XGTNj8l1VGRNk8n1VGQNk8n1VGQNk8o1UGPNl8o1UGPNk8Q2UGnMl8Q2UGnMl8Q2UGoMk8Q2UGnMk8R2VGmMk8S2UGlMl8S2UGlMk8T2VGkMk8U2UGjMl8U2UGkMk8T2VGjMk8V2VGiMk8V2VGiMk8U2WGjMi8V2XGiMi8V2XGiMi8V2XGiMi8U2YGjMg8V2ZGhMh8W2YGhMh8W2YGhMh8W2YGhMg8W2[GhMf8W2[GhMf8W2[GhMe8X2\\GfMf8X2\\GgMe8X2\\GgMe8X2\\GgMd8Y2]GfMd8Y2]GfMb8Z2`GeM`8[2aGeM^8[2cGeM\\8[2eGeMZ8Z2hGfMW8Z2jGfMU8Z2lGfMR8[2oGeMP8[2QHeMn7Z2THfMk7Z2VHfMi7Z2XHfMg7Z2ZHfMe7Y2]HgMb7Y2_HgM_7Z2bHfM]7Y2eHgMZ7Y2gHgMX7X2jHiMT7W2mHiMR7W2oHiMP7V2RIjMn6U2SIkMl6T2VIlMj6S2WImMh6R2ZInMf6Q2[IoMd6P2^IPNa6P2`IPN`6n1bIRN]6n1dIRN\\6l1fITNY6l1hITNX6j1jIVNV6i1kIWNU6g1mIYNS6e1oI[NR6b1PJ_No5_1SJaNm5]1UJcNk5[1WJeNi5X1ZJhNf5V1\\JjNd5T1^JlNb5R1`JnN`5P1bJPO^5m0eJSO[5k0gJUOZ5h0hJXOY5e0iJ[OW5c0kJ]OV5`0lJ@U5=mJCT5:nJFR59oJGR56PKJQ53QKNn40TK0m4NTK2m4KUK5l4HVK8j4FXK:i4DXK in this image.", "objects": [{"patches": [66, 67, 68, 88, 89, 90, 91, 110, 111, 112, 113, 114, 132, 133, 134, 135, 154, 155, 156, 157, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 286, 287, 288], "bbox": [0.004493464052287581, 0.1370751633986928, 0.28988562091503267, 0.6202287581699346], "iscrowd": 0, "area": 33433.17105, "rle": {"size": [612, 612], "counts": "dP2c4m9d400000O1000000O10000O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O10001N3N1O1O1N2O2N1O1N2O1O2N1N2O1O1N2O2N1O1N2O1O2N1N2O1O1O2M2O1O1O1N3fKTDX1m;cNXD\\1i;^N]D`1e;[N`Dd1a;WNdDh1];SNhDl1Z;nMkDP2W;jMoDU2R;fMSEY2n:bMWE]2j:^M[E`2h:ZM]Ee2d:UMbEj2_:QMfEn2[:mLjEQ3X:jLmEU3U:eLPFZ3Q:`LUF^3j in this image.", "objects": [{"patches": [66, 67, 68, 88, 89, 90, 91, 110, 111, 112, 113, 114, 132, 133, 134, 135, 154, 155, 156, 157, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 286, 287, 288], "bbox": [0.004493464052287581, 0.1370751633986928, 0.28988562091503267, 0.6202287581699346], "iscrowd": 0, "area": 33433.17105, "rle": {"size": [612, 612], "counts": "dP2c4m9d400000O1000000O10000O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O10001N3N1O1O1N2O2N1O1N2O1O2N1N2O1O1N2O2N1O1N2O1O2N1N2O1O1O2M2O1O1O1N3fKTDX1m;cNXD\\1i;^N]D`1e;[N`Dd1a;WNdDh1];SNhDl1Z;nMkDP2W;jMoDU2R;fMSEY2n:bMWE]2j:^M[E`2h:ZM]Ee2d:UMbEj2_:QMfEn2[:mLjEQ3X:jLmEU3U:eLPFZ3Q:`LUF^3j in this image.", "objects": [{"patches": [11, 12, 30, 31, 32, 33, 34, 35, 50, 51, 52, 53, 54, 55, 56, 57, 58, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 460, 461, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479], "bbox": [0.05393790849673202, 0.036111111111111115, 1.0, 0.9821732026143791], "iscrowd": 0, "area": 231636.01920000004, "rle": {"size": [612, 612], "counts": "jYd03mb04K5K5K5K5K5L4K5K5K5K5K5L4K5K5K5K5K5L4K5K5K5K5K5L4K5K5K5K5K5L4XO\\KkBi4T=[KdBj4[=`0O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1cHcGR4^8mKSHb3n7^LbHQ3_7nLRIa2o6_MaIP2`6oMQJ`1P6`N[LTOf3k0Z700O100O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O1000000O1001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001c@RNb=n1\\BXN`=i1]B\\N`=e1]B`N`=`1^BfN^=[1_BjN^=V1`BoN]=R1`BTO\\=l0bBYO[=h0bB]O[=d0bBBZ=>dBGY=:dBKY=5eB1W=0fB54cM[ in this image.", "objects": [{"patches": [11, 12, 30, 31, 32, 33, 34, 35, 50, 51, 52, 53, 54, 55, 56, 57, 58, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 460, 461, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479], "bbox": [0.05393790849673202, 0.036111111111111115, 1.0, 0.9821732026143791], "iscrowd": 0, "area": 231636.01920000004, "rle": {"size": [612, 612], "counts": "jYd03mb04K5K5K5K5K5L4K5K5K5K5K5L4K5K5K5K5K5L4K5K5K5K5K5L4K5K5K5K5K5L4XO\\KkBi4T=[KdBj4[=`0O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1cHcGR4^8mKSHb3n7^LbHQ3_7nLRIa2o6_MaIP2`6oMQJ`1P6`N[LTOf3k0Z700O100O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O1000000O1001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001c@RNb=n1\\BXN`=i1]B\\N`=e1]B`N`=`1^BfN^=[1_BjN^=V1`BoN]=R1`BTO\\=l0bBYO[=h0bB]O[=d0bBBZ=>dBGY=:dBKY=5eB1W=0fB54cM[ in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 76, 77, 78, 79, 80, 81, 82, 101, 102, 103, 104, 105, 106, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 220, 221, 222, 223, 243, 244, 245, 246, 266, 267, 268, 269, 290, 291, 292, 313, 314, 337], "bbox": [0.0, 0.0022482435597189696, 0.752640625, 0.9505620608899297], "iscrowd": 0, "area": 52066.19710000002, "rle": {"size": [427, 640], "counts": "12Y=5K4L5K2N001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O00100O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O0000O100000000O1000000O100000000O1000000000000000000O10000000001O000000000O10000000000000000000000O10000000000000000000000O11O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00001O00001O00010O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O00001O001O00001O001O00001O001O00001O001O001O1O1O001O1O1O001O1O1O1O001O1O1O00100O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O001O100O01000O10O10O1000O01000000O01000O10O10O1000O0101O0O10001N100O10001N10000O2O000O101N10000O2O01O2N1O101N1O2N1O101N1O1O2N1O1O2O0O00001O000010O00000oFUMa7k2ZH[Me7e2VH`Mj7a2oGfMo7l301O1N101O0O3N2N2M3N2N3L4M3M3L3N1O0O10001N10000O101O0O1001O01O01O01O3N3L4L4XJSIj4P7hJaIQ5a7L01N10001N10001N1000001N101O001O001O001O0O2O0O101N1O2O0O2O0O2M2O00N2O1O0O2O1ON1O2N2O0O4L4L4L4L4L4SOm0E;F:G9F:F:YLPGj2X9oLnFk2Y9nLmFl2l9J6J6M3M3M100O2N1O100AWEbNe:\\1dEaNX:^1PF_Nl9_1Q1N1N2N3N3L4N2N2N2N2D<@egQ2"}, "label": "girrafe in the back"}]} {"id": 427461, "image": "COCO_train2014_000000427461.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe bends over to stick its nose into a cage\"."}], "task": "refering", "answer_template": "The \"a giraffe bends over to stick its nose into a cage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 23, 24, 46, 47, 48, 49, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 194, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 282, 283, 284, 285, 305, 306, 307, 308, 328, 329, 330], "bbox": [0.0, 0.040491803278688523, 0.47614062500000004, 0.9845667447306792], "iscrowd": 0, "area": 42943.1896, "rle": {"size": [427, 640], "counts": "a2P2[9Q20O100O010O10O0100O010O10O0100O10O0100O010O10O10O10O0100O10O0100O010O10O01000O010000O010000O01000O10O10O1000O10O1000O01000O010000O1000000O10O10O10000O10000O10000O10000O1O100O100O1O010O1O100O100O1O100O1O100O100O1000000000O100000000000001O00000001O00001O00001O010O00001O00001O00001O000010O0001O001O00001O00001O01O01O00001O01O01O00U1kN3N1N3M3M2N3M2O2M2N3M2N3N1N2N3M2N2O1N2N3M2N2N2O1N3M00O1O2N1O1O1N2O1O2N1O1O1N2O1O2N1O1O1N2O1O1O2N1O1O1N2O1O1O1O2N1N2O1O1O1O1O1N3N1O1O1O1O1O1N3N1O1O1O2N2M3N1O2N2N1L5K5K5K4M4K5K5K4L4L5K4L4L4L4M3M3O1O2M2O1N2O1O1NO2N1O2N1OnNQ1M4L4L4L4L4O11O11N2O1O1N2O1O2M2O1O1N2O2N1N2O1O1N3N1N2O1O4K6KS^[4"}, "label": "a giraffe bends over to stick its nose into a cage"}]} {"id": 427461, "image": "COCO_train2014_000000427461.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the neck and face of a giraffe with its head in a feeding trough\"."}], "task": "refering", "answer_template": "The \"the neck and face of a giraffe with its head in a feeding trough\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 23, 24, 46, 47, 48, 49, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 194, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 282, 283, 284, 285, 305, 306, 307, 308, 328, 329, 330], "bbox": [0.0, 0.040491803278688523, 0.47614062500000004, 0.9845667447306792], "iscrowd": 0, "area": 42943.1896, "rle": {"size": [427, 640], "counts": "a2P2[9Q20O100O010O10O0100O010O10O0100O10O0100O010O10O10O10O0100O10O0100O010O10O01000O010000O010000O01000O10O10O1000O10O1000O01000O010000O1000000O10O10O10000O10000O10000O10000O1O100O100O1O010O1O100O100O1O100O1O100O100O1000000000O100000000000001O00000001O00001O00001O010O00001O00001O00001O000010O0001O001O00001O00001O01O01O00001O01O01O00U1kN3N1N3M3M2N3M2O2M2N3M2N3N1N2N3M2N2O1N2N3M2N2N2O1N3M00O1O2N1O1O1N2O1O2N1O1O1N2O1O2N1O1O1N2O1O1O2N1O1O1N2O1O1O1O2N1N2O1O1O1O1O1N3N1O1O1O1O1O1N3N1O1O1O2N2M3N1O2N2N1L5K5K5K4M4K5K5K4L4L5K4L4L4L4M3M3O1O2M2O1N2O1O1NO2N1O2N1OnNQ1M4L4L4L4L4O11O11N2O1O1N2O1O2M2O1O1N2O2N1N2O1O1N3N1N2O1O4K6KS^[4"}, "label": "the neck and face of a giraffe with its head in a feeding trough"}]} {"id": 427467, "image": "COCO_train2014_000000427467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white dog on back of truck\"."}], "task": "refering", "answer_template": "The \"white dog on back of truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 193, 214, 215, 216, 217, 218, 219, 220, 237, 238, 239, 240, 241, 242, 243, 244, 245, 267], "bbox": [0.29426562500000003, 0.46699999999999997, 0.659453125, 0.6500833333333332], "iscrowd": 0, "area": 10204.939550000005, "rle": {"size": [480, 640], "counts": "X[h2:f>2O2M2N2O0O100O1O3N6I2O1N2N2N1O1O001O1O001O2M3N1O2N1O1O2N2N3L4M5K3M1O1O0000000000000000000O1N2N2OSN`C\\1^ in this image.", "objects": [{"patches": [190, 191, 192, 193, 214, 215, 216, 217, 218, 219, 220, 237, 238, 239, 240, 241, 242, 243, 244, 245, 267], "bbox": [0.29426562500000003, 0.46699999999999997, 0.659453125, 0.6500833333333332], "iscrowd": 0, "area": 10204.939550000005, "rle": {"size": [480, 640], "counts": "X[h2:f>2O2M2N2O0O100O1O3N6I2O1N2N2N1O1O001O1O001O2M3N1O2N1O1O2N2N3L4M5K3M1O1O0000000000000000000O1N2N2OSN`C\\1^ in this image.", "objects": [{"patches": [197, 208, 209, 210, 211, 220, 221, 228, 230, 231, 232, 233, 234, 235, 242, 243, 244, 245, 246, 250, 251, 252, 253, 254, 255, 256, 257, 258, 263, 264, 265, 266, 267, 268, 269, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.017953125, 0.546056338028169, 0.9961875000000001, 0.9797652582159624], "iscrowd": 0, "area": 61270.380200000014, "rle": {"size": [426, 640], "counts": "\\l48Q=_2aM1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N100O2N1O1O2N1O2N1O1O1O100000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O100000000O1000000O100000000O1000000O1002N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N1O00O1000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000O10000000000000000000000000000000000O100O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100gNjL]HW3b7]MkGc2T8Z1O1O1N2O1O001O1N2O1O1O1O1N2O1O1O1N2O1O1O12N2N3M3M3M3M3M3M3M3M3M3M3M3M2M4M001O1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O00O01000000000O100000000O100000000O100001O3M3M3M3M3M3M3M3M3M3M3L4M3M3M3M3M3M3M3M3M3M3M3M2N3M3M1O0000001O0000000000000000000O100000000000000000000000000000000000000000000000000000001O0O10000000000000000000000000000000000000000000000000000000000O100000000000001O0000000000000000000000000000000000^MnNlHR1S7XOdHh0[7C[H=d7NRH2m78jGGV8d0`G\\O_8n0XGROg8Y1oFgNP9c1gF]NX9k2O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O11O2N1O1O1O1O2M2O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1D in this image.", "objects": [{"patches": [197, 208, 209, 210, 211, 220, 221, 228, 230, 231, 232, 233, 234, 235, 242, 243, 244, 245, 246, 250, 251, 252, 253, 254, 255, 256, 257, 258, 263, 264, 265, 266, 267, 268, 269, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.017953125, 0.546056338028169, 0.9961875000000001, 0.9797652582159624], "iscrowd": 0, "area": 61270.380200000014, "rle": {"size": [426, 640], "counts": "\\l48Q=_2aM1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N100O2N1O1O2N1O2N1O1O1O100000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O100000000O1000000O100000000O1000000O1002N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N1O00O1000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000O10000000000000000000000000000000000O100O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100gNjL]HW3b7]MkGc2T8Z1O1O1N2O1O001O1N2O1O1O1O1N2O1O1O1N2O1O1O12N2N3M3M3M3M3M3M3M3M3M3M3M3M2M4M001O1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O00O01000000000O100000000O100000000O100001O3M3M3M3M3M3M3M3M3M3M3L4M3M3M3M3M3M3M3M3M3M3M3M2N3M3M1O0000001O0000000000000000000O100000000000000000000000000000000000000000000000000000001O0O10000000000000000000000000000000000000000000000000000000000O100000000000001O0000000000000000000000000000000000^MnNlHR1S7XOdHh0[7C[H=d7NRH2m78jGGV8d0`G\\O_8n0XGROg8Y1oFgNP9c1gF]NX9k2O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O11O2N1O1O1O1O2M2O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1D in this image.", "objects": [{"patches": [77, 78, 79, 100, 101, 102, 123, 124, 125, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 284, 285], "bbox": [0.292015625, 0.20215962441314556, 0.54825, 0.8064553990610329], "iscrowd": 0, "area": 26780.063850000002, "rle": {"size": [426, 640], "counts": "mQ^27P=9F6J5M3N2M3M3M3M3M3M2N2N3M2N2N2N3M2N2N2N3XE[Ng9g2K6J5K5K5L2N3M2N2N2N3N1N2O1N3N1N2O2M200O1O2N1O1O100O1O1O1O1UNnJ]KS5^4YKZKh4b4bKWK_4e4kKTKV4h4YLlJh3Q5]LkJe3U5\\LjJd3U5^LiJc3W5^LhJb3W5`LgJa3X5aLgJ_3Y5bLeJ_3Z5dLaJ_3_5V2O100O1O100O2O0O100O01000O100O1000O010000000000000O1000001O2N1O1N3N1O1O1O2N1O1N2O2N1N2O1O2M2O1O1O2M2O1O1`JeIn3c6cKgI[4d6QKfIn4]7N1N10000O2O000O2O00000O2O000O101O000O2O001N2O2N1O2M3N1M4SMRGX1Q9ZN`G`1h8mMjGl1P:J5J7I5K5K4L5K4M4K:F7I5K5KVng3"}, "label": "a man wearing a green jacket"}]} {"id": 361939, "image": "COCO_train2014_000000361939.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the passenger holding a bicycle with the ' wrangler ' jacket on\"."}], "task": "refering", "answer_template": "The \"the passenger holding a bicycle with the ' wrangler ' jacket on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 100, 101, 102, 123, 124, 125, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 284, 285], "bbox": [0.292015625, 0.20215962441314556, 0.54825, 0.8064553990610329], "iscrowd": 0, "area": 26780.063850000002, "rle": {"size": [426, 640], "counts": "mQ^27P=9F6J5M3N2M3M3M3M3M3M2N2N3M2N2N2N3M2N2N2N3XE[Ng9g2K6J5K5K5L2N3M2N2N2N3N1N2O1N3N1N2O2M200O1O2N1O1O100O1O1O1O1UNnJ]KS5^4YKZKh4b4bKWK_4e4kKTKV4h4YLlJh3Q5]LkJe3U5\\LjJd3U5^LiJc3W5^LhJb3W5`LgJa3X5aLgJ_3Y5bLeJ_3Z5dLaJ_3_5V2O100O1O100O2O0O100O01000O100O1000O010000000000000O1000001O2N1O1N3N1O1O1O2N1O1N2O2N1N2O1O2M2O1O1O2M2O1O1`JeIn3c6cKgI[4d6QKfIn4]7N1N10000O2O000O2O00000O2O000O101O000O2O001N2O2N1O2M3N1M4SMRGX1Q9ZN`G`1h8mMjGl1P:J5J7I5K5K4L5K4M4K:F7I5K5KVng3"}, "label": "the passenger holding a bicycle with the ' wrangler ' jacket on"}]} {"id": 361939, "image": "COCO_train2014_000000361939.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a green hat holding a red jacket\"."}], "task": "refering", "answer_template": "The \"a woman wearing a green hat holding a red jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 87, 88, 89, 110, 111, 112, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 292, 293, 294, 295, 296, 316, 317, 318, 319, 320, 339, 340, 342, 343], "bbox": [0.6880625, 0.18492957746478875, 0.9573750000000001, 0.9888967136150234], "iscrowd": 0, "area": 33851.16140000001, "rle": {"size": [426, 640], "counts": "gZg51W=6J6K5J5_Ob0@?YOh0E:G9G:J5L4L3L4M3M3M3M3L4M2N2N3M2M3N2M3K6J5K5K4M4L3N3M3L3O2M2O2N2nImIi4d7F2N1O2N2N1O1O1O2N10000O10000O10000O10000RNjGhNV8W1oGeNQ8Z1PHeNQ8[1PHdNo7\\1RHdNm7]1THbNk7^1VHbNi7_1XH`Nf7a1[H[MJ9i6EeIh2IXMO2Q6_4QJ^Ki6g4XIVKg6m4YIoJf6W5`06K5J6K6I6J6K6O01O100N2N2N2N101N2N2N2N2N2N3M4M3M3M3M3M3M2N3M4L7I6K5J6K6I6K5J5L3L3M100O010O10O010lGhLU6W3iIoLT6Q3iIRMV6P3\\IhLnN9g7P3VIlLoN6l7o2PI`MQ7a2kHaMU7a2gHaMZ7`2bHbMj3mNWOc3jLbMm3oNWOa3hLcMn3POYO^3dLeMS4oNXO]3aLfMW4oNVOj4>jJ_L`0o2f4a0kJ`Ld0k2`4e0mJ_Lg0h2]4g0nJ`Lj0d2X4?jJoL6Mm0a2R4b0nJnL4NP1^2n3e0PKnL3Nf1h1X3[1QKmL10^2P1`2R2SKlL10`2P1\\2S2TKlL10`2P1\\2R2TKnL1O`2P1\\2R2SKoL2N`2Q1[2P2TKQMQ3n0l1P2SKRMR3m0l1o1SKSMS3n0m1j1RKWMR3n0R2d1mJ^MR3m0X2]1gJfMR3l0^2V1aJnMR3k0c2`0lIPO`0ER3j0i26aNIg1FcN9X7O1N2O1O1O1O1O1N2O2N1O1O1N2M3N2Ncj:"}, "label": "a woman wearing a green hat holding a red jacket"}]} {"id": 361939, "image": "COCO_train2014_000000361939.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a hat and holding a red blanket\"."}], "task": "refering", "answer_template": "The \"a woman wearing a hat and holding a red blanket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 87, 88, 89, 110, 111, 112, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 292, 293, 294, 295, 296, 316, 317, 318, 319, 320, 339, 340, 342, 343], "bbox": [0.6880625, 0.18492957746478875, 0.9573750000000001, 0.9888967136150234], "iscrowd": 0, "area": 33851.16140000001, "rle": {"size": [426, 640], "counts": "gZg51W=6J6K5J5_Ob0@?YOh0E:G9G:J5L4L3L4M3M3M3M3L4M2N2N3M2M3N2M3K6J5K5K4M4L3N3M3L3O2M2O2N2nImIi4d7F2N1O2N2N1O1O1O2N10000O10000O10000O10000RNjGhNV8W1oGeNQ8Z1PHeNQ8[1PHdNo7\\1RHdNm7]1THbNk7^1VHbNi7_1XH`Nf7a1[H[MJ9i6EeIh2IXMO2Q6_4QJ^Ki6g4XIVKg6m4YIoJf6W5`06K5J6K6I6J6K6O01O100N2N2N2N101N2N2N2N2N2N3M4M3M3M3M3M3M2N3M4L7I6K5J6K6I6K5J5L3L3M100O010O10O010lGhLU6W3iIoLT6Q3iIRMV6P3\\IhLnN9g7P3VIlLoN6l7o2PI`MQ7a2kHaMU7a2gHaMZ7`2bHbMj3mNWOc3jLbMm3oNWOa3hLcMn3POYO^3dLeMS4oNXO]3aLfMW4oNVOj4>jJ_L`0o2f4a0kJ`Ld0k2`4e0mJ_Lg0h2]4g0nJ`Lj0d2X4?jJoL6Mm0a2R4b0nJnL4NP1^2n3e0PKnL3Nf1h1X3[1QKmL10^2P1`2R2SKlL10`2P1\\2S2TKlL10`2P1\\2R2TKnL1O`2P1\\2R2SKoL2N`2Q1[2P2TKQMQ3n0l1P2SKRMR3m0l1o1SKSMS3n0m1j1RKWMR3n0R2d1mJ^MR3m0X2]1gJfMR3l0^2V1aJnMR3k0c2`0lIPO`0ER3j0i26aNIg1FcN9X7O1N2O1O1O1O1O1N2O2N1O1O1N2M3N2Ncj:"}, "label": "a woman wearing a hat and holding a red blanket"}]} {"id": 222676, "image": "COCO_train2014_000000222676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a tennis racket across his chest\"."}], "task": "refering", "answer_template": "The \"a man holding a tennis racket across his chest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 80, 81, 102, 103, 104, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 194, 195, 196, 197, 216, 217, 218, 219, 220], "bbox": [0.420734375, 0.1430246913580247, 0.5951875, 0.5844650205761317], "iscrowd": 0, "area": 16600.8073, "rle": {"size": [486, 640], "counts": "gko3:m=P1I7I7I7dDWNS9P2eFWNU9P2dFVNZ9m1aFWN^9k1_FXN`9i1\\FZNc9h1ZFZNe9h1WF[Ni9f1TF\\Nk9f1PF^No9d1`EmN_:m2O1O100O2N100O1O100O1O100O1O010O1O1O1L4L4L4]Oc0N2M3N2N2N1N3N2O1O1000000O1000000000O2O00001O0000001O00001O0000001jIYGk5T9M3M3M3M3iJjFX4S:O1O001O001O001O001O00001O001O001O0O1XNi1N2O1N2O1N2O0O2O4L8G8I7H7J7H7J7H9DZgj3"}, "label": "a man holding a tennis racket across his chest"}]} {"id": 222676, "image": "COCO_train2014_000000222676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man between other two holding a badminton racket on his chest\"."}], "task": "refering", "answer_template": "The \"a man between other two holding a badminton racket on his chest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 80, 81, 102, 103, 104, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 194, 195, 196, 197, 216, 217, 218, 219, 220], "bbox": [0.420734375, 0.1430246913580247, 0.5951875, 0.5844650205761317], "iscrowd": 0, "area": 16600.8073, "rle": {"size": [486, 640], "counts": "gko3:m=P1I7I7I7dDWNS9P2eFWNU9P2dFVNZ9m1aFWN^9k1_FXN`9i1\\FZNc9h1ZFZNe9h1WF[Ni9f1TF\\Nk9f1PF^No9d1`EmN_:m2O1O100O2N100O1O100O1O100O1O010O1O1O1L4L4L4]Oc0N2M3N2N2N1N3N2O1O1000000O1000000000O2O00001O0000001O00001O0000001jIYGk5T9M3M3M3M3iJjFX4S:O1O001O001O001O001O00001O001O001O0O1XNi1N2O1N2O1N2O0O2O4L8G8I7H7J7H7J7H9DZgj3"}, "label": "a man between other two holding a badminton racket on his chest"}]} {"id": 222676, "image": "COCO_train2014_000000222676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy scared of balls\"."}], "task": "refering", "answer_template": "The \"a guy scared of balls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 189, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 258, 259, 260, 281, 282, 283, 304, 305, 306, 307, 327, 328, 329, 330, 350, 351, 352, 353, 375, 376], "bbox": [0.149828125, 0.24664609053497943, 0.384796875, 0.9865432098765432], "iscrowd": 0, "area": 29037.259049999997, "rle": {"size": [486, 640], "counts": "Vg]1h0]>?B8G1O2N2O1O001O10O10O1000O1000B[NYCf1d<^N[Ca1bY4P2jHlMg2fKGX4 in this image.", "objects": [{"patches": [96, 97, 98, 118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 189, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 258, 259, 260, 281, 282, 283, 304, 305, 306, 307, 327, 328, 329, 330, 350, 351, 352, 353, 375, 376], "bbox": [0.149828125, 0.24664609053497943, 0.384796875, 0.9865432098765432], "iscrowd": 0, "area": 29037.259049999997, "rle": {"size": [486, 640], "counts": "Vg]1h0]>?B8G1O2N2O1O001O10O10O1000O1000B[NYCf1d<^N[Ca1bY4P2jHlMg2fKGX4 in this image.", "objects": [{"patches": [84, 85, 107, 108, 129, 130, 131, 152, 153, 154, 155, 175, 176, 177, 178, 179, 198, 199, 200, 201, 221, 222, 223, 244, 245, 246, 247, 267, 268, 269, 270, 290, 291, 292, 293, 294, 312, 313, 314, 315, 316, 317, 335, 336, 337, 338, 339, 340, 358, 359, 362, 363], "bbox": [0.587453125, 0.21364197530864198, 0.83375, 0.9487860082304526], "iscrowd": 0, "area": 27459.72415000001, "rle": {"size": [486, 640], "counts": "Ukb58k>b0^O=C:G:E:F4L3N2M2N3M3N2M3VMdMgH_2S7WNZHk1`7lNnGV1l7_ObGd0X8X3I7J6J5J7J6J6I^OfHfIT7_6PI]Ik6h6ZITI_6R7eIkHU6Z7PJbHj5c7ZJZH_5l7h05M40O0000O100O100O100O2OO0N3M2N3M2N3M2N3M2NK5J36K5K6I6K5K6O03M3N1N3M3M3N1N3N2N2O1N101N2O1N2N101N2O1N2O0OZIWIQ5g6lJ^IT5a6gJeIY5Y6dJlI]5R6^JTJc5i5ZJ\\Jh5a5SJeJn5X5oImJR6Q5iIUKY6g4dI^Ka2_Ng0P6dLgK^2`NP1f5^LnKT1RNJb0e2[5ZLTLW1SN@f0P3W7`MVHWOk0[3k6^M^HnNn0e3_6_MfHcNS1o3S6_M_Jc2[5_MgJb2U5_MmJc2o4^MRKc2k4]MWKd2f4\\M]Ke2_4[McKe2[4[MgKc2Y4]MiKa2W4_MkK`2T4cLXI5f2V3Q4dL[I6f2T3o3dL]I8f2S3l3cL`I:f2Q3j3dLaI;g2o2h3eLbI=g2m2h3dLaI?i2k2i3bL_Ic0j2i2j3`L]Ig0k2h2j3]L\\Ik0l2f2V4ZMlKd2T4\\MnKc2Q4]MQLa2o3_MSL_2l3bMXL[2g3eM]LW2c3jMaLQ2_3nMfLo1Y3QNkLk1U3UNoLg1Q3XNTMe1k2[NYMa1g2^N^M^1c2aNaM\\1_2bNfMY1\\2fNiMU1X2jNXNf0i1XOfN9\\1FeN9\\1EfN:\\1DeN:^1CeN;]1CdN;_1BcN=_1AbN>`1@aN>U8O1O1N2O1O1N3N1O1N2O1OT[b1"}, "label": "the person to the right of the man not swinging his tennis racket"}]} {"id": 222676, "image": "COCO_train2014_000000222676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue jacket playing tennis\"."}], "task": "refering", "answer_template": "The \"a man in a blue jacket playing tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 107, 108, 129, 130, 131, 152, 153, 154, 155, 175, 176, 177, 178, 179, 198, 199, 200, 201, 221, 222, 223, 244, 245, 246, 247, 267, 268, 269, 270, 290, 291, 292, 293, 294, 312, 313, 314, 315, 316, 317, 335, 336, 337, 338, 339, 340, 358, 359, 362, 363], "bbox": [0.587453125, 0.21364197530864198, 0.83375, 0.9487860082304526], "iscrowd": 0, "area": 27459.72415000001, "rle": {"size": [486, 640], "counts": "Ukb58k>b0^O=C:G:E:F4L3N2M2N3M3N2M3VMdMgH_2S7WNZHk1`7lNnGV1l7_ObGd0X8X3I7J6J5J7J6J6I^OfHfIT7_6PI]Ik6h6ZITI_6R7eIkHU6Z7PJbHj5c7ZJZH_5l7h05M40O0000O100O100O100O2OO0N3M2N3M2N3M2N3M2NK5J36K5K6I6K5K6O03M3N1N3M3M3N1N3N2N2O1N101N2O1N2N101N2O1N2O0OZIWIQ5g6lJ^IT5a6gJeIY5Y6dJlI]5R6^JTJc5i5ZJ\\Jh5a5SJeJn5X5oImJR6Q5iIUKY6g4dI^Ka2_Ng0P6dLgK^2`NP1f5^LnKT1RNJb0e2[5ZLTLW1SN@f0P3W7`MVHWOk0[3k6^M^HnNn0e3_6_MfHcNS1o3S6_M_Jc2[5_MgJb2U5_MmJc2o4^MRKc2k4]MWKd2f4\\M]Ke2_4[McKe2[4[MgKc2Y4]MiKa2W4_MkK`2T4cLXI5f2V3Q4dL[I6f2T3o3dL]I8f2S3l3cL`I:f2Q3j3dLaI;g2o2h3eLbI=g2m2h3dLaI?i2k2i3bL_Ic0j2i2j3`L]Ig0k2h2j3]L\\Ik0l2f2V4ZMlKd2T4\\MnKc2Q4]MQLa2o3_MSL_2l3bMXL[2g3eM]LW2c3jMaLQ2_3nMfLo1Y3QNkLk1U3UNoLg1Q3XNTMe1k2[NYMa1g2^N^M^1c2aNaM\\1_2bNfMY1\\2fNiMU1X2jNXNf0i1XOfN9\\1FeN9\\1EfN:\\1DeN:^1CeN;]1CdN;_1BcN=_1AbN>`1@aN>U8O1O1N2O1O1N3N1O1N2O1OT[b1"}, "label": "a man in a blue jacket playing tennis"}]} {"id": 189915, "image": "COCO_train2014_000000189915.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bench with a woman in a blue sweater sitting on it\"."}], "task": "refering", "answer_template": "The \"a bench with a woman in a blue sweater sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 187, 188, 189, 190, 191, 192, 197, 198, 199, 207, 208, 209, 210, 211, 212, 213, 214, 215, 220, 221, 222, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 243, 244, 245, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 265, 266, 267, 268, 276, 277, 282, 283, 289, 290, 299, 300, 305, 306, 312, 313], "bbox": [0.0015, 0.5505607476635513, 0.6672499999999999, 0.9235981308411214], "iscrowd": 0, "area": 30398.79745, "rle": {"size": [428, 640], "counts": "hg01[:i1iFXNX9n2000000000O100000000O100000000O100008Hl0TO>B00O10000O10000O100dN\\10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000005K:F:F:F:F:F3M1O0000000000001O00000000001O000000000000VMcGf0\\8YOgGe0Y8YOjGf0V8YOlGf0T8XOoGg0Q8XOoGiNJT1W81QHkN1k0n79RHlN8c0f7`0SHmN`09^7h0THoNg00U7P1THPOQ1Fl6Y1THQOQ1Ek6Z1THQOR1Ck6\\1SHQOR1Bl6]1RHQOS1Ak6]1RHSOT1^Ok6_1QHSOT1^Ok6_1QHSOU1\\Ok6a1PHSOV1[Oj6a1QHTOU1ZOk6b1oGUOk9k0UFUOk9j0VFVOj98dE@b08j94hED=9k90lEF9:k9MoEH7:k9JSFK2;k9GVFNO;k9DZF0J=l9_O^F3G>k9^O`F3E?k9]OaF4D?k9\\OcF4B`0k9[OeF3@c0k9YOfF4_Oc0k9XOhF4]Od0k9WOiF5\\Od0U;ZOlDf0T;ZOkDg0V;XOjDh0V;WOkDi0V;VOjDj0V;VOjDi0l;1001O3M2N00001O000000001O0000001N100000001O0000001O0000001O00000O101O0000001O5K5KfRi00YmVO100O1O1O100O1O100O1O100O00100O00100O00100O00100O0oC7`:m1^O0O1O1gMeMaIZ2`6WNnHj1R7`NcHa1]7`NbH_1_7bN_H_1a7bN^H^1b7cN]H]1c7dN\\H[1e7iNWHW1i7mNSHS1m7QOoGo0Q8UOkGj0V8ZOfGgNIY1a81eGfNi0:b7P1eGfNZ:[1eEeN[:[1fEdNZ:]1eEcN[:]1eEcN[:^1dEbN\\:^1dEbN\\:_1bEbN^:^1`EdN`:]1^EdNb:T2O0000000000000000000000O100O1O100O2O0O1O10c0\\Ob1^N9HQgh2"}, "label": "a bench with a woman in a blue sweater sitting on it"}]} {"id": 189915, "image": "COCO_train2014_000000189915.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black park bench woman holding book sitting on\"."}], "task": "refering", "answer_template": "The \"black park bench woman holding book sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 187, 188, 189, 190, 191, 192, 197, 198, 199, 207, 208, 209, 210, 211, 212, 213, 214, 215, 220, 221, 222, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 243, 244, 245, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 265, 266, 267, 268, 276, 277, 282, 283, 289, 290, 299, 300, 305, 306, 312, 313], "bbox": [0.0015, 0.5505607476635513, 0.6672499999999999, 0.9235981308411214], "iscrowd": 0, "area": 30398.79745, "rle": {"size": [428, 640], "counts": "hg01[:i1iFXNX9n2000000000O100000000O100000000O100008Hl0TO>B00O10000O10000O100dN\\10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000005K:F:F:F:F:F3M1O0000000000001O00000000001O000000000000VMcGf0\\8YOgGe0Y8YOjGf0V8YOlGf0T8XOoGg0Q8XOoGiNJT1W81QHkN1k0n79RHlN8c0f7`0SHmN`09^7h0THoNg00U7P1THPOQ1Fl6Y1THQOQ1Ek6Z1THQOR1Ck6\\1SHQOR1Bl6]1RHQOS1Ak6]1RHSOT1^Ok6_1QHSOT1^Ok6_1QHSOU1\\Ok6a1PHSOV1[Oj6a1QHTOU1ZOk6b1oGUOk9k0UFUOk9j0VFVOj98dE@b08j94hED=9k90lEF9:k9MoEH7:k9JSFK2;k9GVFNO;k9DZF0J=l9_O^F3G>k9^O`F3E?k9]OaF4D?k9\\OcF4B`0k9[OeF3@c0k9YOfF4_Oc0k9XOhF4]Od0k9WOiF5\\Od0U;ZOlDf0T;ZOkDg0V;XOjDh0V;WOkDi0V;VOjDj0V;VOjDi0l;1001O3M2N00001O000000001O0000001N100000001O0000001O0000001O00000O101O0000001O5K5KfRi00YmVO100O1O1O100O1O100O1O100O00100O00100O00100O00100O0oC7`:m1^O0O1O1gMeMaIZ2`6WNnHj1R7`NcHa1]7`NbH_1_7bN_H_1a7bN^H^1b7cN]H]1c7dN\\H[1e7iNWHW1i7mNSHS1m7QOoGo0Q8UOkGj0V8ZOfGgNIY1a81eGfNi0:b7P1eGfNZ:[1eEeN[:[1fEdNZ:]1eEcN[:]1eEcN[:^1dEbN\\:^1dEbN\\:_1bEbN^:^1`EdN`:]1^EdNb:T2O0000000000000000000000O100O1O100O2O0O1O10c0\\Ob1^N9HQgh2"}, "label": "black park bench woman holding book sitting on"}]} {"id": 435681, "image": "COCO_train2014_000000435681.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"car\"."}], "task": "refering", "answer_template": "The \"car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73], "bbox": [0.0, 0.03189695550351288, 0.231, 0.2512646370023419], "iscrowd": 0, "area": 11265.100149999998, "rle": {"size": [427, 640], "counts": "P37f<=D=B>C=B>D<000000000000000000000O1000000000000000000000000000000O1000000000000000000000O100000O1000000000000001O2N1O2N2N00000000000000000000O10000O1O100O1O100O1O100O1K5L40O10O1000000000000O1000000000000O100000000000000O1008H00000000000000000000001O01O2N2N2N2N3M2N2N2N2L4I7J7K4M3N2M3N2M3M3N3L3DbS]6"}, "label": "car"}]} {"id": 435681, "image": "COCO_train2014_000000435681.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue van behind woman\"."}], "task": "refering", "answer_template": "The \"blue van behind woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73], "bbox": [0.0, 0.03189695550351288, 0.231, 0.2512646370023419], "iscrowd": 0, "area": 11265.100149999998, "rle": {"size": [427, 640], "counts": "P37f<=D=B>C=B>D<000000000000000000000O1000000000000000000000000000000O1000000000000000000000O100000O1000000000000001O2N1O2N2N00000000000000000000O10000O1O100O1O100O1O100O1K5L40O10O1000000000000O1000000000000O100000000000000O1008H00000000000000000000001O01O2N2N2N2N3M2N2N2N2L4I7J7K4M3N2M3N2M3M3N3L3DbS]6"}, "label": "blue van behind woman"}]} {"id": 288234, "image": "COCO_train2014_000000288234.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy playing baseball wearing a blue and white uniform\"."}], "task": "refering", "answer_template": "The \"a boy playing baseball wearing a blue and white uniform\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 34, 35, 36, 56, 57, 58, 59, 78, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 243, 244, 245, 260, 261, 262, 266, 267, 268, 283, 284, 290, 291, 304, 305, 306, 307, 313, 314, 327, 328, 329, 336, 337, 338], "bbox": [0.23778125, 0.0016981132075471698, 0.71634375, 0.9881603773584906], "iscrowd": 0, "area": 48200.958849999995, "rle": {"size": [424, 640], "counts": "XZo12T=3M4L3M4M2O2N2N1O2N1O2N1O2N001O000000001O0001O010O00001O01O01O0001O0O10000O10000O01000O1N2O1N2N2O1N2O0Em2Y4nMPKYOi0g2Y4YOkKc0V4\\OnK`0T4^OQL=Q4ATL:m3EWL7k3F[L5g3I]L3d3LaLOa3OdLL^3b0VLZOn3b0WLYOn3a0XLZOl3`0ZL\\Ok3=^L\\Og3=cL\\O`3>jLZO[3?nLZOX3;UM]O`8K1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N2N_nZ2"}, "label": "a boy playing baseball wearing a blue and white uniform"}]} {"id": 288234, "image": "COCO_train2014_000000288234.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"child player morgan\"."}], "task": "refering", "answer_template": "The \"child player morgan\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 34, 35, 36, 56, 57, 58, 59, 78, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 243, 244, 245, 260, 261, 262, 266, 267, 268, 283, 284, 290, 291, 304, 305, 306, 307, 313, 314, 327, 328, 329, 336, 337, 338], "bbox": [0.23778125, 0.0016981132075471698, 0.71634375, 0.9881603773584906], "iscrowd": 0, "area": 48200.958849999995, "rle": {"size": [424, 640], "counts": "XZo12T=3M4L3M4M2O2N2N1O2N1O2N1O2N001O000000001O0001O010O00001O01O01O0001O0O10000O10000O01000O1N2O1N2N2O1N2O0Em2Y4nMPKYOi0g2Y4YOkKc0V4\\OnK`0T4^OQL=Q4ATL:m3EWL7k3F[L5g3I]L3d3LaLOa3OdLL^3b0VLZOn3b0WLYOn3a0XLZOl3`0ZL\\Ok3=^L\\Og3=cL\\O`3>jLZO[3?nLZOX3;UM]O`8K1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N2N_nZ2"}, "label": "child player morgan"}]} {"id": 288234, "image": "COCO_train2014_000000288234.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right leg of a person playing a baseman in a baseball game\"."}], "task": "refering", "answer_template": "The \"the right leg of a person playing a baseman in a baseball game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 138, 161, 184, 185, 207, 208, 230, 231, 253, 254, 255, 256, 277, 278, 279], "bbox": [0.0, 0.3350707547169811, 0.1583125, 0.8597405660377357], "iscrowd": 0, "area": 6968.905450000002, "rle": {"size": [424, 640], "counts": "m6j1P9`21O000000000000000000001O1O1O001O1N2M3M3M3N2O001Of0ZO1O1O1O001O10fElM]9S2`FnMc9j24L3M3M4L3M3M3M4K4M3M4L3L4M3M4L4L4K5L4L4L2M2O001O1O0O2O1O001O0O101O00001N10001O000O101O00000O2O0000000O2O0O101O0O10001N100O2N2M2N3M3MVkn6"}, "label": "the right leg of a person playing a baseman in a baseball game"}]} {"id": 288234, "image": "COCO_train2014_000000288234.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a leg sticks out ahead of the base\"."}], "task": "refering", "answer_template": "The \"a leg sticks out ahead of the base\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 138, 161, 184, 185, 207, 208, 230, 231, 253, 254, 255, 256, 277, 278, 279], "bbox": [0.0, 0.3350707547169811, 0.1583125, 0.8597405660377357], "iscrowd": 0, "area": 6968.905450000002, "rle": {"size": [424, 640], "counts": "m6j1P9`21O000000000000000000001O1O1O001O1N2M3M3M3N2O001Of0ZO1O1O1O001O10fElM]9S2`FnMc9j24L3M3M4L3M3M3M4K4M3M4L3L4M3M4L4L4K5L4L4L2M2O001O1O0O2O1O001O0O101O00001N10001O000O101O00000O2O0000000O2O0O101O0O10001N100O2N2M2N3M3MVkn6"}, "label": "a leg sticks out ahead of the base"}]} {"id": 476651, "image": "COCO_train2014_000000476651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white t - shirt with both of his hands on a tennis net\"."}], "task": "refering", "answer_template": "The \"a man in a white t - shirt with both of his hands on a tennis net\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 147, 148, 149, 169, 170, 171, 172, 173, 192, 193, 194, 196, 215], "bbox": [0.350078125, 0.3655841121495327, 0.5733125, 0.6176168224299066], "iscrowd": 0, "area": 5574.115, "rle": {"size": [428, 640], "counts": "Qlm23Y=000000000O10000O101kBKQ=5oBKP=6oBKQ=8O100O1N2M3N0D=N1N3M2O10O1NO12O1N101N3N1N3O10O017QD\\Oh:e1O1N2N2O1N2O1O0010000O10000O1000000O10O1000O100L4H8O1N2O100O10O10000000O11O00001O001O0TOdEgN]:U1PF^NU:^1o0M3L5L4K5iNoCm0Z in this image.", "objects": [{"patches": [125, 126, 147, 148, 149, 169, 170, 171, 172, 173, 192, 193, 194, 196, 215], "bbox": [0.350078125, 0.3655841121495327, 0.5733125, 0.6176168224299066], "iscrowd": 0, "area": 5574.115, "rle": {"size": [428, 640], "counts": "Qlm23Y=000000000O10000O101kBKQ=5oBKP=6oBKQ=8O100O1N2M3N0D=N1N3M2O10O1NO12O1N101N3N1N3O10O017QD\\Oh:e1O1N2N2O1N2O1O0010000O10000O1000000O10O1000O100L4H8O1N2O100O10O10000000O11O00001O001O0TOdEgN]:U1PF^NU:^1o0M3L5L4K5iNoCm0Z in this image.", "objects": [{"patches": [129, 152, 153, 175, 176, 177, 198, 199, 200, 221, 222, 223, 244, 245, 246, 267, 268, 269, 290, 291, 313, 314, 335, 336, 337], "bbox": [0.595109375, 0.38651869158878505, 0.712328125, 0.9730373831775702], "iscrowd": 0, "area": 9471.172300000004, "rle": {"size": [428, 640], "counts": "fdo42X=5L2N1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2N2J6J6K5J6J6J6mJUN`NP2W1ZNbNk1[1[NTK^OU2]2e2YNPKBQ2Z2n2XNkJFk1Z2X3TNgJJe1Y2c3QNbJN_1Y2P4XOdKo0]4UOVKT1k4e32O0O2O0O2O1O0010O01O10O01O00100O00102M4L4L5L3L4WOfIXJ_6\\5P1D9H9F9H9F:G8F;D;D=D;D=C=]Ob0XO`f\\2"}, "label": "a man wearing sports dress and a bat in his hand"}]} {"id": 476651, "image": "COCO_train2014_000000476651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a yellow sweatband on his head\"."}], "task": "refering", "answer_template": "The \"a man wearing a yellow sweatband on his head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 152, 153, 175, 176, 177, 198, 199, 200, 221, 222, 223, 244, 245, 246, 267, 268, 269, 290, 291, 313, 314, 335, 336, 337], "bbox": [0.595109375, 0.38651869158878505, 0.712328125, 0.9730373831775702], "iscrowd": 0, "area": 9471.172300000004, "rle": {"size": [428, 640], "counts": "fdo42X=5L2N1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2N2J6J6K5J6J6J6mJUN`NP2W1ZNbNk1[1[NTK^OU2]2e2YNPKBQ2Z2n2XNkJFk1Z2X3TNgJJe1Y2c3QNbJN_1Y2P4XOdKo0]4UOVKT1k4e32O0O2O0O2O1O0010O01O10O01O00100O00102M4L4L5L3L4WOfIXJ_6\\5P1D9H9F9H9F:G8F;D;D=D;D=C=]Ob0XO`f\\2"}, "label": "a man wearing a yellow sweatband on his head"}]} {"id": 230893, "image": "COCO_train2014_000000230893.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a truck\"."}], "task": "refering", "answer_template": "The \"the back of a truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.55108, 0.733342175066313, 1.0, 1.0], "iscrowd": 0, "area": 18000.600099999992, "rle": {"size": [377, 500], "counts": "YlU37b;8H8H8H8H7I8H8H8H8H8H2N00000000000000001O00000000000000000000000000000000001O000000000000000000000000000000001O000000000000000000000000000O1000001O00000001O0000001O0000001O0000001O01O0001O0000001O0000001O0000001O0000010O0000001O0000001O0000001O000000010O0000001O0000001O0000001O0000001O01O0001O0000001O0000001O0000001O0000010O0000001O0000001O0000001O0000001O01O0001O0000001O0000001O0000001O0001O01O0000001O0000001O0000001O0000O101N10000O100O10000"}, "label": "the back of a truck"}]} {"id": 58864, "image": "COCO_train2014_000000058864.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair on the right\"."}], "task": "refering", "answer_template": "The \"the chair on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 34, 35, 36, 37, 38, 39, 57, 58, 59, 60, 61, 62, 80, 81, 82, 84, 85, 103, 104, 105, 107, 108], "bbox": [0.48707812500000003, 0.0022500000000000003, 0.717984375, 0.2741666666666667], "iscrowd": 0, "area": 14775.708300000006, "rle": {"size": [480, 640], "counts": "lYb47^> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 34, 35, 36, 37, 38, 39, 57, 58, 59, 60, 61, 62, 80, 81, 82, 84, 85, 103, 104, 105, 107, 108], "bbox": [0.48707812500000003, 0.0022500000000000003, 0.717984375, 0.2741666666666667], "iscrowd": 0, "area": 14775.708300000006, "rle": {"size": [480, 640], "counts": "lYb47^> in this image.", "objects": [{"patches": [180, 181, 182, 183, 184, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 222, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 308, 309, 310, 311, 312, 313, 314, 315, 316, 326, 327, 328, 329, 330, 331, 332, 333, 334, 344, 345, 346, 347, 348, 349, 350, 351, 352, 362, 363, 364, 365, 366, 367, 368, 369, 370, 382, 383, 384, 385, 386, 387, 403, 404, 405], "bbox": [0.019783889980353634, 0.4494375, 0.5905304518664047, 0.97753125], "iscrowd": 0, "area": 62926.83815000001, "rle": {"size": [640, 509], "counts": "na6U1ib04L4L4L4L4L4V@UNY=o1cBSN[=R2`BQN]=S2_BPN_=S2]BPNb=R2ZBQNd=R2YBPNf=R2VBQNh=R2TBQNk=Q2SBoMm=R2QBPNm=S2PBoMo=S2oAnMo=T2oAnMP>T2nAmMQ>U2lAnMQ>T2nAmMQ>U2lAmMR>V2kAlMT>U2hAoMV>T2eAPNY>n3M1O2O1N2O1O001O1O1O1O1O001O1O1O000000000000000000000000000000007I>B>B>B=C>B>B:F00001O0000001O0000001O00010O0001O01O01O0010O0001O01O01O0010O00010O0001O010O000010O0001O010O000010O00010O01O00010O000010O01O1O010O1O1O010O1O100O001O100O1O00100O1O1O010O1O1O10O01O100O1O00100O1O00100O1O1O010O1O1O10O01O100O1O00100O1O1O010O1O1O2O0O1O1O100O1O100O2N1O100O1O1O100O1O2N100O1O1O2O1N3M3N1N3M2N3N2M2N3M2O2M3M2N3N1N3M3M01O0O10001N100O2O0O10001N100O101N10000O2N1N2O1O0O01O004L6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6JRSR4"}, "label": "a chair with a white & black color cover"}]} {"id": 280051, "image": "COCO_train2014_000000280051.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white chair wrapped in a black sash\"."}], "task": "refering", "answer_template": "The \"a white chair wrapped in a black sash\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 183, 184, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 222, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 308, 309, 310, 311, 312, 313, 314, 315, 316, 326, 327, 328, 329, 330, 331, 332, 333, 334, 344, 345, 346, 347, 348, 349, 350, 351, 352, 362, 363, 364, 365, 366, 367, 368, 369, 370, 382, 383, 384, 385, 386, 387, 403, 404, 405], "bbox": [0.019783889980353634, 0.4494375, 0.5905304518664047, 0.97753125], "iscrowd": 0, "area": 62926.83815000001, "rle": {"size": [640, 509], "counts": "na6U1ib04L4L4L4L4L4V@UNY=o1cBSN[=R2`BQN]=S2_BPN_=S2]BPNb=R2ZBQNd=R2YBPNf=R2VBQNh=R2TBQNk=Q2SBoMm=R2QBPNm=S2PBoMo=S2oAnMo=T2oAnMP>T2nAmMQ>U2lAnMQ>T2nAmMQ>U2lAmMR>V2kAlMT>U2hAoMV>T2eAPNY>n3M1O2O1N2O1O001O1O1O1O1O001O1O1O000000000000000000000000000000007I>B>B>B=C>B>B:F00001O0000001O0000001O00010O0001O01O01O0010O0001O01O01O0010O00010O0001O010O000010O0001O010O000010O00010O01O00010O000010O01O1O010O1O1O010O1O100O001O100O1O00100O1O1O010O1O1O10O01O100O1O00100O1O00100O1O1O010O1O1O10O01O100O1O00100O1O1O010O1O1O2O0O1O1O100O1O100O2N1O100O1O1O100O1O2N100O1O1O2O1N3M3N1N3M2N3N2M2N3M2O2M3M2N3N1N3M3M01O0O10001N100O2O0O10001N100O101N10000O2N1N2O1O0O01O004L6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6J6JRSR4"}, "label": "a white chair wrapped in a black sash"}]} {"id": 148983, "image": "COCO_train2014_000000148983.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing gree flies a kite\"."}], "task": "refering", "answer_template": "The \"a woman wearing gree flies a kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [291, 292, 307, 308, 309, 324, 325, 326, 327, 328, 341, 342, 343, 344, 358, 359, 360, 361, 375, 376, 377, 378], "bbox": [0.0656875, 0.753578125, 0.3020416666666667, 1.0], "iscrowd": 0, "area": 10728.3771, "rle": {"size": [640, 480], "counts": "Xad09ac0:H7K2O1N2O1N2N2O1N2O100O1O1X^OQO_`0P1[_O]O\\`0d0`_OGX`0;c_OOV`0Q2M3L5K100O1O1O1O1O1O1O1N2O100O1DoKh@R4T?ULh@l3T?\\Lh@d3W?c0O1O1O100O1O1O100O100000000001O001O00001O001O001O00004bKl@h3i?M1O1O1O1O1O1O1O1O001O001O00001O0O2K4J7H7J7fNU_OWOSa0g0R_OnNVa0P1o^OfNYa0W1k0N1N2N2N101N1O2NOO0010N1M23N2N2N3N1O2N2M4L4J6KRk`6"}, "label": "a woman wearing gree flies a kite"}]} {"id": 148983, "image": "COCO_train2014_000000148983.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman on the beach with a kite\"."}], "task": "refering", "answer_template": "The \"a woman on the beach with a kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [291, 292, 307, 308, 309, 324, 325, 326, 327, 328, 341, 342, 343, 344, 358, 359, 360, 361, 375, 376, 377, 378], "bbox": [0.0656875, 0.753578125, 0.3020416666666667, 1.0], "iscrowd": 0, "area": 10728.3771, "rle": {"size": [640, 480], "counts": "Xad09ac0:H7K2O1N2O1N2N2O1N2O100O1O1X^OQO_`0P1[_O]O\\`0d0`_OGX`0;c_OOV`0Q2M3L5K100O1O1O1O1O1O1O1N2O100O1DoKh@R4T?ULh@l3T?\\Lh@d3W?c0O1O1O100O1O1O100O100000000001O001O00001O001O001O00004bKl@h3i?M1O1O1O1O1O1O1O1O001O001O00001O0O2K4J7H7J7fNU_OWOSa0g0R_OnNVa0P1o^OfNYa0W1k0N1N2N2N101N1O2NOO0010N1M23N2N2N3N1O2N2M4L4J6KRk`6"}, "label": "a woman on the beach with a kite"}]} {"id": 419324, "image": "COCO_train2014_000000419324.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with a grey shirt\"."}], "task": "refering", "answer_template": "The \"man with a grey shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 240, 241, 262, 263, 264, 265, 285, 286, 287, 288, 308, 310, 354, 377], "bbox": [0.388671875, 0.5716249999999999, 0.53515625, 0.9752708333333333], "iscrowd": 0, "area": 5799.21875, "rle": {"size": [480, 640], "counts": "hTe33m>3M2gAKc=8nA4R>=iMUOYFk0g9UOYFk0h8XOlEN[1k0g8DcEAf1k0e8GdE_Of1j0e8IcE^Oh1i0c8KeE\\Oh1i0a8MgEZOh1i0`8NgE[Oh1g0_80iEYOg1h0_80jEXOg1h0]82kEWOh1g0[84mEVOg1f0[85mEVOh1e0Y87oETOh1e0W89QFROh1e0V8:QFSOg1d0X89QFSOg1d0W8:RFROg1d0W8:QFSOh1c0V8;RFQOi1d0U8 in this image.", "objects": [{"patches": [217, 218, 240, 241, 262, 263, 264, 265, 285, 286, 287, 288, 308, 310, 354, 377], "bbox": [0.388671875, 0.5716249999999999, 0.53515625, 0.9752708333333333], "iscrowd": 0, "area": 5799.21875, "rle": {"size": [480, 640], "counts": "hTe33m>3M2gAKc=8nA4R>=iMUOYFk0g9UOYFk0h8XOlEN[1k0g8DcEAf1k0e8GdE_Of1j0e8IcE^Oh1i0c8KeE\\Oh1i0a8MgEZOh1i0`8NgE[Oh1g0_80iEYOg1h0_80jEXOg1h0]82kEWOh1g0[84mEVOg1f0[85mEVOh1e0Y87oETOh1e0W89QFROh1e0V8:QFSOg1d0X89QFSOg1d0W8:RFROg1d0W8:QFSOh1c0V8;RFQOi1d0U8 in this image.", "objects": [{"patches": [156, 157, 178, 179, 180, 201, 202, 203, 204, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 270, 271, 272, 273, 293, 294, 295, 316, 317, 318], "bbox": [0.7283593749999999, 0.37454166666666666, 0.924640625, 0.7817708333333333], "iscrowd": 0, "area": 14018.559399999996, "rle": {"size": [480, 640], "counts": "efj62k>6K5J5K5L3L5K5L3L5L3N3M3M2N3M2M4M3]DWNj9k1hEcNV:`1ZEoNd:h2N2M3O2M2N2O0O2O0O1O1ROjKRGV4j8SLnFP4o8TLnFn3P9VLkFm3S9VLiFm3U9Q1N1O100O1O100ON3M2M3N3M2N2N36J7I6J2N01O0O1000000O2O00001O1O1O1O001O1O1O3M3M2N3M2N3M2N3M2XKZFR4Z:01O00001ZN]EZNV;e1T1]MdC`2\\<`MeC_2_ in this image.", "objects": [{"patches": [156, 157, 178, 179, 180, 201, 202, 203, 204, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 270, 271, 272, 273, 293, 294, 295, 316, 317, 318], "bbox": [0.7283593749999999, 0.37454166666666666, 0.924640625, 0.7817708333333333], "iscrowd": 0, "area": 14018.559399999996, "rle": {"size": [480, 640], "counts": "efj62k>6K5J5K5L3L5K5L3L5L3N3M3M2N3M2M4M3]DWNj9k1hEcNV:`1ZEoNd:h2N2M3O2M2N2O0O2O0O1O1ROjKRGV4j8SLnFP4o8TLnFn3P9VLkFm3S9VLiFm3U9Q1N1O100O1O100ON3M2M3N3M2N2N36J7I6J2N01O0O1000000O2O00001O1O1O1O001O1O1O3M3M2N3M2N3M2N3M2XKZFR4Z:01O00001ZN]EZNV;e1T1]MdC`2\\<`MeC_2_ in this image.", "objects": [{"patches": [129, 130, 131, 152, 153, 154, 175, 176, 177, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 244, 245, 246, 267, 268, 269, 270, 290, 291, 292, 293, 313, 314], "bbox": [0.43695312499999994, 0.3026875, 0.748, 0.7725833333333334], "iscrowd": 0, "area": 16574.04325, "rle": {"size": [480, 640], "counts": "U`S41o>00000O10000O10000O1000000O4M6J1N1000001O0000001O1O1O1O1O1O1O001O00001O0000O10000O100O100O10000O100O2O00000O10O1O010O10O01O10O0100O00100O10O01O10O0100O00100O010O100O0010O0100O0010O01O010O1O010O001O10O01O010O1N101N101N2N101N101N2TOhNoCX1P in this image.", "objects": [{"patches": [129, 130, 131, 152, 153, 154, 175, 176, 177, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 244, 245, 246, 267, 268, 269, 270, 290, 291, 292, 293, 313, 314], "bbox": [0.43695312499999994, 0.3026875, 0.748, 0.7725833333333334], "iscrowd": 0, "area": 16574.04325, "rle": {"size": [480, 640], "counts": "U`S41o>00000O10000O10000O1000000O4M6J1N1000001O0000001O1O1O1O1O1O1O001O00001O0000O10000O100O100O10000O100O2O00000O10O1O010O10O01O10O0100O00100O10O01O10O0100O00100O010O100O0010O0100O0010O01O010O1O010O001O10O01O010O1N101N101N2N101N101N2TOhNoCX1P in this image.", "objects": [{"patches": [66, 67, 68, 84, 85, 86, 87, 101, 102, 103, 104, 105, 121, 122, 123], "bbox": [0.65628, 0.24670000000000003, 0.89098, 0.4631], "iscrowd": 0, "area": 7439.683000000001, "rle": {"size": [400, 500], "counts": "iXP42Y<6J5J7J5J6K6I6K6O000O2O0O2O000O2O001N101O0O101N101O0O101O0O2O001N100O2O001N10001O001O001O01O01O001O000010O01O00001O010O001O000010O01O000010O01O01O00001O001N10001O00001O0O101O001O00001N10001O00001O0O2O00001O00001N10002N2N2iNW1L3N3L4M3L^[e0"}, "label": "a black microwave"}]} {"id": 157190, "image": "COCO_train2014_000000157190.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black microwave on the counter to the right of a man wearing a grey shirt\"."}], "task": "refering", "answer_template": "The \"a black microwave on the counter to the right of a man wearing a grey shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 68, 84, 85, 86, 87, 101, 102, 103, 104, 105, 121, 122, 123], "bbox": [0.65628, 0.24670000000000003, 0.89098, 0.4631], "iscrowd": 0, "area": 7439.683000000001, "rle": {"size": [400, 500], "counts": "iXP42Y<6J5J7J5J6K6I6K6O000O2O0O2O000O2O001N101O0O101N101O0O101O0O2O001N100O2O001N10001O001O001O01O01O001O000010O01O00001O010O001O000010O01O000010O01O01O00001O001N10001O00001O0O101O001O00001N10001O00001O0O2O00001O00001N10002N2N2iNW1L3N3L4M3L^[e0"}, "label": "a black microwave on the counter to the right of a man wearing a grey shirt"}]} {"id": 443916, "image": "COCO_train2014_000000443916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a metal chair to the right\"."}], "task": "refering", "answer_template": "The \"a metal chair to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 245, 268, 269, 270], "bbox": [0.661125, 0.33345833333333336, 0.8877812500000001, 0.7150416666666667], "iscrowd": 0, "area": 607.6213000000005, "rle": {"size": [480, 640], "counts": "hbV61l>4O00000jAN^=2aB0^=0aB3]=MbB5]=KbB7]=IbB6`=K^B4d=L[B3g=MXB3i=MUB3m=MQB3Q>MnA2T>NjA2X>NfA3[>72N2N2O1N20010O000000010O00001O01O01O000010O0001O001O01O01O000010O0001O00010O003MmUP1OUjoN0O100O100O1O100000001O00000004L_bQ1"}, "label": "a metal chair to the right"}]} {"id": 443916, "image": "COCO_train2014_000000443916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair the bear on the right is sitting in\"."}], "task": "refering", "answer_template": "The \"the chair the bear on the right is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 245, 268, 269, 270], "bbox": [0.661125, 0.33345833333333336, 0.8877812500000001, 0.7150416666666667], "iscrowd": 0, "area": 607.6213000000005, "rle": {"size": [480, 640], "counts": "hbV61l>4O00000jAN^=2aB0^=0aB3]=MbB5]=KbB7]=IbB6`=K^B4d=L[B3g=MXB3i=MUB3m=MQB3Q>MnA2T>NjA2X>NfA3[>72N2N2O1N20010O000000010O00001O01O01O000010O0001O001O01O01O000010O0001O00010O003MmUP1OUjoN0O100O100O1O100000001O00000004L_bQ1"}, "label": "the chair the bear on the right is sitting in"}]} {"id": 443916, "image": "COCO_train2014_000000443916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giant teddybear with a happy halloween sign\"."}], "task": "refering", "answer_template": "The \"a giant teddybear with a happy halloween sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 84, 85, 86, 87, 88, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 288, 289, 290, 291, 293, 294, 295, 296, 312, 313, 315, 316, 317, 318, 319, 339, 340, 341, 363, 364], "bbox": [0.5561875, 0.15954166666666666, 0.9421406250000001, 0.9303333333333333], "iscrowd": 0, "area": 53141.5089, "rle": {"size": [480, 640], "counts": "nVW52l>5K5K5jMBbEc0Y:G^E=^:LXE9d:0RE5i:5mDNQ;;eDGZ;c0\\D^Oc;Q2M2O2N1N3N1O2M2N3N1N3M2N3M2N2O2M2M4M2N3M2N3M2M4M2N3M2N2N1N3N1O1O2N1O1N3N1O2N100O2N101N1O101N1O2O0O1O2O0O2N10N2O1N3M2N3oNXGgKj8U4eG^K^8]4T1L]NaKnH[4U7hKjHY4S7iKlHX4S7jKlHW4Q7lKmHU4Q7nKnHR4P7PLoHR4P7oKoHQ4Q7PLmHR4Q7TLjHl3V7P2O001N101O00001N101O001N101O001O0O2O00001N101O001N101O0O2O0RJhHT4X7iKoHR4S7jKTIR4l6lKZIP4g6lK`Io3a6nKfIn3[6oKlIk3U6RLRJj3o5RLXJj3h5TL\\Ji3f5SL]Jm3c5PLaJP4`5lKbJT4`5gKdJY4\\5cKgJ]4[5_KhJ`4m70O01N1O2N2N1O2N1O2O1O0O2O001O1O0O2O001O001N2O001O0011O0O2O1N101OO001O001O001O1O001O001O10O10O01O1N1O2O0]FQLT8Q4fGULX8m3cGXL[8i3cGZL[8h3aG\\L^8e3_G^L_8c3_G`L_8b3^GaL`8a3\\GcLc8]3[GfLc8\\3ZGgLd8[3XGiLf8X3XGkLf8W3WGlLh8g4N1O2M3B>A?H8H8H9G8H9G8H9H7H9G9H7M4K4L5M3M2N3M2N3M2N3M3M2N3M2N3M2N3M3M2N3M2N3M2L5L4K4M4L3L5L3L5L4L in this image.", "objects": [{"patches": [63, 64, 84, 85, 86, 87, 88, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 288, 289, 290, 291, 293, 294, 295, 296, 312, 313, 315, 316, 317, 318, 319, 339, 340, 341, 363, 364], "bbox": [0.5561875, 0.15954166666666666, 0.9421406250000001, 0.9303333333333333], "iscrowd": 0, "area": 53141.5089, "rle": {"size": [480, 640], "counts": "nVW52l>5K5K5jMBbEc0Y:G^E=^:LXE9d:0RE5i:5mDNQ;;eDGZ;c0\\D^Oc;Q2M2O2N1N3N1O2M2N3N1N3M2N3M2N2O2M2M4M2N3M2N3M2M4M2N3M2N2N1N3N1O1O2N1O1N3N1O2N100O2N101N1O101N1O2O0O1O2O0O2N10N2O1N3M2N3oNXGgKj8U4eG^K^8]4T1L]NaKnH[4U7hKjHY4S7iKlHX4S7jKlHW4Q7lKmHU4Q7nKnHR4P7PLoHR4P7oKoHQ4Q7PLmHR4Q7TLjHl3V7P2O001N101O00001N101O001N101O001O0O2O00001N101O001N101O0O2O0RJhHT4X7iKoHR4S7jKTIR4l6lKZIP4g6lK`Io3a6nKfIn3[6oKlIk3U6RLRJj3o5RLXJj3h5TL\\Ji3f5SL]Jm3c5PLaJP4`5lKbJT4`5gKdJY4\\5cKgJ]4[5_KhJ`4m70O01N1O2N2N1O2N1O2O1O0O2O001O1O0O2O001O001N2O001O0011O0O2O1N101OO001O001O001O1O001O001O10O10O01O1N1O2O0]FQLT8Q4fGULX8m3cGXL[8i3cGZL[8h3aG\\L^8e3_G^L_8c3_G`L_8b3^GaL`8a3\\GcLc8]3[GfLc8\\3ZGgLd8[3XGiLf8X3XGkLf8W3WGlLh8g4N1O2M3B>A?H8H8H9G8H9G8H9H7H9G9H7M4K4L5M3M2N3M2N3M2N3M3M2N3M2N3M2N3M3M2N3M2N3M2L5L4K4M4L3L5L3L5L4L in this image.", "objects": [{"patches": [50, 71, 72, 73, 74, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 326, 327, 328, 350, 351], "bbox": [0.0471875, 0.16179166666666667, 0.443265625, 0.903375], "iscrowd": 0, "area": 49760.3815, "rle": {"size": [480, 640], "counts": "oZ>2g>9WKFaJb0]5EVJd0h5BkIf0S6A`Ih0]6[O]Ij0a6WO\\In0a6TOZIQ1e6POWIU1f6mNUIX1i6jNSI[1j6gNQI^1m6dNoH`1o6bNmHc1Q7^NlHe1R7]NlHf1Q7\\NmHf1Q7\\NmHg1P7[NmHh1R7YNlHi1S7XNkHk1S7VNkHl1U7SNiHP2V7QNhHR2W7nMgHT2X7mMeHW2Y7jMeHX2Y7jMeHX2Y7iMfHX2X7kMeHW2Z7kMdHV2[7lMbHW2\\7kMbHV2]7lM`HV2^7lM`HV2_7PN[HQ2d7VNTHl1k7e2N2O1O2N1000000O101O001O0O2O001O001N101O001O0O2O001O001N101O001N1O2N101N1O3M3N3L3M3N3L3M3M4M2M2N3N2M2N3M3N1O2N2N101N1O2O1N1O2N2O1N3M3N2M2N2N1001O4L4L5K4L5K00001O00000O10001O000O100O1O2N1O100O1O2N100jKWJSOj5k0[JQOf5m0_JoNb5o0bJnN`5o0eJnN[5P1jJlNW5R1nJjNS5T1PKjNR5S1PKlNQ5R1QKnNo4P1TKnNm4o0VKPOl4m0VKROk4l0WKSOj4l0XKROi4m0XKoNl4P1UKmNn4R1TKiNo4X1QKeNR5Z1oJbNT5^1mJ_NV5`1lJ[NW5e1jJXNY5h1gJTN\\5l1fJPN]5o1dJmM_5S2bJjMa5V2_JfMd5Z2^JbMd5^2]J^Mg5a2ZJ\\Mh5e2YJVMk5i2VJTMl5l2UJPMo5o2RJnLP6R3RJiLR6W3nIgLS6Y3nIeLT6Z3nIbLT6^3mI`LU6`3kI^LV6b3kI\\LW6c3kIZLV6f3c2O010O1O0010O01O1O010O001O1O01jE^Lm8a3fFmLY9S3ZFZMg9e2mEhMS:b30010O0101N2O2M2O2M3M2O2M3N1N3N1N2O00O0100O010O10O0100O010O10O010g0XO2O1N2O1N2O1N2O1N2O1N2M3M3M3M3M3N2MU1kN:F:F:FVaV5"}, "label": "stuffed black bear wearing yellow t shirt with orange pumkin on shirt"}]} {"id": 443916, "image": "COCO_train2014_000000443916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the huge toy bear seated with a jack - o - lantern on one side of its body\"."}], "task": "refering", "answer_template": "The \"the huge toy bear seated with a jack - o - lantern on one side of its body\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 71, 72, 73, 74, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 326, 327, 328, 350, 351], "bbox": [0.0471875, 0.16179166666666667, 0.443265625, 0.903375], "iscrowd": 0, "area": 49760.3815, "rle": {"size": [480, 640], "counts": "oZ>2g>9WKFaJb0]5EVJd0h5BkIf0S6A`Ih0]6[O]Ij0a6WO\\In0a6TOZIQ1e6POWIU1f6mNUIX1i6jNSI[1j6gNQI^1m6dNoH`1o6bNmHc1Q7^NlHe1R7]NlHf1Q7\\NmHf1Q7\\NmHg1P7[NmHh1R7YNlHi1S7XNkHk1S7VNkHl1U7SNiHP2V7QNhHR2W7nMgHT2X7mMeHW2Y7jMeHX2Y7jMeHX2Y7iMfHX2X7kMeHW2Z7kMdHV2[7lMbHW2\\7kMbHV2]7lM`HV2^7lM`HV2_7PN[HQ2d7VNTHl1k7e2N2O1O2N1000000O101O001O0O2O001O001N101O001O0O2O001O001N101O001N1O2N101N1O3M3N3L3M3N3L3M3M4M2M2N3N2M2N3M3N1O2N2N101N1O2O1N1O2N2O1N3M3N2M2N2N1001O4L4L5K4L5K00001O00000O10001O000O100O1O2N1O100O1O2N100jKWJSOj5k0[JQOf5m0_JoNb5o0bJnN`5o0eJnN[5P1jJlNW5R1nJjNS5T1PKjNR5S1PKlNQ5R1QKnNo4P1TKnNm4o0VKPOl4m0VKROk4l0WKSOj4l0XKROi4m0XKoNl4P1UKmNn4R1TKiNo4X1QKeNR5Z1oJbNT5^1mJ_NV5`1lJ[NW5e1jJXNY5h1gJTN\\5l1fJPN]5o1dJmM_5S2bJjMa5V2_JfMd5Z2^JbMd5^2]J^Mg5a2ZJ\\Mh5e2YJVMk5i2VJTMl5l2UJPMo5o2RJnLP6R3RJiLR6W3nIgLS6Y3nIeLT6Z3nIbLT6^3mI`LU6`3kI^LV6b3kI\\LW6c3kIZLV6f3c2O010O1O0010O01O1O010O001O1O01jE^Lm8a3fFmLY9S3ZFZMg9e2mEhMS:b30010O0101N2O2M2O2M3M2O2M3N1N3N1N2O00O0100O010O10O0100O010O10O010g0XO2O1N2O1N2O1N2O1N2O1N2M3M3M3M3M3N2MU1kN:F:F:FVaV5"}, "label": "the huge toy bear seated with a jack - o - lantern on one side of its body"}]} {"id": 443916, "image": "COCO_train2014_000000443916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a teddy bear in a yellow shirt with an orange pumpkin basket\"."}], "task": "refering", "answer_template": "The \"a teddy bear in a yellow shirt with an orange pumpkin basket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 71, 72, 73, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120, 121, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 261, 262, 279, 280, 281, 282, 284, 285, 303, 304, 305, 306, 307, 308, 326, 327, 328, 329, 350, 351], "bbox": [0.04153125, 0.1628125, 0.440640625, 0.9063125], "iscrowd": 0, "area": 48089.38649999998, "rle": {"size": [480, 640], "counts": "Pi<3m>3M3M3L4M3fE^OP6c0iIET6=dIK[66^I2`6N[I9c6HWI?f6BUId0j6]OPIj0n6VOmHQ1Q7POiHV1U7kNfH\\1X7eNcHa1[7_N`Hh1]7ZN`Hi1_7WNaHk1]7VNaHm1]7TNaHn1]7SNbHP2\\7QNbHR2\\7nMdHS2[7nMcHU2Z7mMdHV2Z7jMeHY2Y7hMfHY2W7jMgHY2W7hMgH[2V7gMhH[2W7fMgH]2V7eMiH]2U7dMiH^2U7dMdHd2Z7]M_Hk2_7VM`Hl2]7VMaHl2^7VM`Hj2_7XM_Hi2a7XM^Hh2a7ZM]Hg2c7ZM\\Hf2d7[M[Hd2e7^MYHc2g7^MXHb2g7`MWHa2i7`MVH`2i7bMUH_2k7cMSH]2m7[2O010O100O10000O010O100O10000O101N2O2M3N1O2M2O2M2O2M3N1O2M2O2M2O2M3N1O2M2N3N1N3N2L3M4L3M4L3N3L4N3N3M2N2N001O1O001O0015J6K5K1N100O010O10000O1O1O1N2N101N2N2[GVJQ8k5fG\\JZ8U6001O000010O0001O00001O01O01O00001OcLeIUN[6d1nIZNS6_1UJ^Nl5_1YJ^Nh5a1[J]Ne5c1]JZNd5e1_JYNa5g1bJUN`5i1cJUN]5k1eJRN\\5m1gJQNY5o1hJoMZ5o1iJoMW5Q2jJmMW5R2lJkMV5T2lJjMT5U2nJiMS5W2oJdMU5Z2nJaMU5_2lJ]MX5a2kJZMX5f2jJUMY5j2iJRM[5m2gJnL\\5Q3fJkL]5U3eJeL`5Y3cJbL`5]3bJ_Lb5`3`J[Lc5d3`JWLc5i3k2N101N2O1N2O1N2N2O1N2O1N2O1N2O1N3N1N2OTD]M\\;\\3H7I8H7I9G9H5KN101N2O1N1O2O1N101N2O0O2O1N1O2O1N101N2O0O2N2M2O2M2O1N2O1N2O1N2O2M2O1N2O1N2O1nMcD>_;\\OiDa0Y;WOPEe0R;UOVEh0k:QO^Ek0d:nNeEm0^:mNjEm0Y in this image.", "objects": [{"patches": [50, 71, 72, 73, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120, 121, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 261, 262, 279, 280, 281, 282, 284, 285, 303, 304, 305, 306, 307, 308, 326, 327, 328, 329, 350, 351], "bbox": [0.04153125, 0.1628125, 0.440640625, 0.9063125], "iscrowd": 0, "area": 48089.38649999998, "rle": {"size": [480, 640], "counts": "Pi<3m>3M3M3L4M3fE^OP6c0iIET6=dIK[66^I2`6N[I9c6HWI?f6BUId0j6]OPIj0n6VOmHQ1Q7POiHV1U7kNfH\\1X7eNcHa1[7_N`Hh1]7ZN`Hi1_7WNaHk1]7VNaHm1]7TNaHn1]7SNbHP2\\7QNbHR2\\7nMdHS2[7nMcHU2Z7mMdHV2Z7jMeHY2Y7hMfHY2W7jMgHY2W7hMgH[2V7gMhH[2W7fMgH]2V7eMiH]2U7dMiH^2U7dMdHd2Z7]M_Hk2_7VM`Hl2]7VMaHl2^7VM`Hj2_7XM_Hi2a7XM^Hh2a7ZM]Hg2c7ZM\\Hf2d7[M[Hd2e7^MYHc2g7^MXHb2g7`MWHa2i7`MVH`2i7bMUH_2k7cMSH]2m7[2O010O100O10000O010O100O10000O101N2O2M3N1O2M2O2M2O2M3N1O2M2O2M2O2M3N1O2M2N3N1N3N2L3M4L3M4L3N3L4N3N3M2N2N001O1O001O0015J6K5K1N100O010O10000O1O1O1N2N101N2N2[GVJQ8k5fG\\JZ8U6001O000010O0001O00001O01O01O00001OcLeIUN[6d1nIZNS6_1UJ^Nl5_1YJ^Nh5a1[J]Ne5c1]JZNd5e1_JYNa5g1bJUN`5i1cJUN]5k1eJRN\\5m1gJQNY5o1hJoMZ5o1iJoMW5Q2jJmMW5R2lJkMV5T2lJjMT5U2nJiMS5W2oJdMU5Z2nJaMU5_2lJ]MX5a2kJZMX5f2jJUMY5j2iJRM[5m2gJnL\\5Q3fJkL]5U3eJeL`5Y3cJbL`5]3bJ_Lb5`3`J[Lc5d3`JWLc5i3k2N101N2O1N2O1N2N2O1N2O1N2O1N2O1N3N1N2OTD]M\\;\\3H7I8H7I9G9H5KN101N2O1N1O2O1N101N2O0O2O1N1O2O1N101N2O0O2N2M2O2M2O1N2O1N2O1N2O2M2O1N2O1N2O1nMcD>_;\\OiDa0Y;WOPEe0R;UOVEh0k:QO^Ek0d:nNeEm0^:mNjEm0Y in this image.", "objects": [{"patches": [166, 167, 188, 189, 190, 211, 212, 213, 214, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331, 332, 350, 353, 354, 373, 376, 377], "bbox": [0.178546875, 0.43829166666666663, 0.462734375, 0.9957916666666667], "iscrowd": 0, "area": 25815.474799999993, "rle": {"size": [480, 640], "counts": "gee12n>5J6J6K4K6K5K5K5K7I8G9F:F9G:F:F:G9G9Ge0[Oc0^O3L5K5K4M4K5K4L6J6K4K6K3M2M4M2N2M0100000O01000O01N2dNoHcJW7[5UIUJP7l5R1001N10000O2O000O2O000O101N10001N12N1O100O1O100O2N2O2M4L4L3N3L3M4M3L3M4L3N3L3M4M3L3M4L3N3L3M4M2M0001O0000010O00001O0000010O00001O0001O01O00001O00010O00015J8H8H6K4K4L5K2O0O0001O0001O01O000WNaFQN`9l1mFjMR9T2ZGbMf8\\2gGYMY8e2THoLn7n2S2N2N3M2N2N2N3M2N2N3M2N2N2O2M2N2N3M2N2N2N3MQDcN_:o2001O0000001O0O100aN`1ROn0J8GRnP5"}, "label": "a wooden chair with a black bag sitting on top of it next to a toilet"}]} {"id": 140821, "image": "COCO_train2014_000000140821.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wood chair with the bag on it\"."}], "task": "refering", "answer_template": "The \"the wood chair with the bag on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [166, 167, 188, 189, 190, 211, 212, 213, 214, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331, 332, 350, 353, 354, 373, 376, 377], "bbox": [0.178546875, 0.43829166666666663, 0.462734375, 0.9957916666666667], "iscrowd": 0, "area": 25815.474799999993, "rle": {"size": [480, 640], "counts": "gee12n>5J6J6K4K6K5K5K5K7I8G9F:F9G:F:F:G9G9Ge0[Oc0^O3L5K5K4M4K5K4L6J6K4K6K3M2M4M2N2M0100000O01000O01N2dNoHcJW7[5UIUJP7l5R1001N10000O2O000O2O000O101N10001N12N1O100O1O100O2N2O2M4L4L3N3L3M4M3L3M4L3N3L3M4M3L3M4L3N3L3M4M2M0001O0000010O00001O0000010O00001O0001O01O00001O00010O00015J8H8H6K4K4L5K2O0O0001O0001O01O000WNaFQN`9l1mFjMR9T2ZGbMf8\\2gGYMY8e2THoLn7n2S2N2N3M2N2N2N3M2N2N3M2N2N2O2M2N2N3M2N2N2N3MQDcN_:o2001O0000001O0O100aN`1ROn0J8GRnP5"}, "label": "the wood chair with the bag on it"}]} {"id": 140821, "image": "COCO_train2014_000000140821.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden chair sitting in the corner next to a few rolls of toilet paper\"."}], "task": "refering", "answer_template": "The \"a wooden chair sitting in the corner next to a few rolls of toilet paper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 167, 168, 169, 190, 191, 192, 214, 215, 216, 217, 218, 237, 238, 239, 240, 241, 263, 264, 286, 287, 309], "bbox": [0.29304687500000004, 0.36264583333333333, 0.49768749999999995, 0.8319583333333334], "iscrowd": 0, "area": 7961.132499999995, "rle": {"size": [480, 640], "counts": "[Zh28f>8G8I8H8I7J5J3M3M3N2M2N3M5L3L3M4L3N3L3M3M4M2M4L5L5J5K2N100O100O100003M4M3L4L4K5L5K4L4K5L4L4L4K6K4L4K5L4L4K5L1O0O101N10000O10000O101O0O100O100O1O2O0O1O100O1O1O2O0O10000O1000001O00000O1000001O00fB@Tm in this image.", "objects": [{"patches": [145, 146, 167, 168, 169, 190, 191, 192, 214, 215, 216, 217, 218, 237, 238, 239, 240, 241, 263, 264, 286, 287, 309], "bbox": [0.29304687500000004, 0.36264583333333333, 0.49768749999999995, 0.8319583333333334], "iscrowd": 0, "area": 7961.132499999995, "rle": {"size": [480, 640], "counts": "[Zh28f>8G8I8H8I7J5J3M3M3N2M2N3M5L3L3M4L3N3L3M3M4M2M4L5L5J5K2N100O100O100003M4M3L4L4K5L5K4L4K5L4L4L4K6K4L4K5L4L4K5L1O0O101N10000O10000O101O0O100O100O1O2O0O1O100O1O1O2O0O10000O1000001O00000O1000001O00fB@Tm in this image.", "objects": [{"patches": [60, 83, 105, 106, 107, 109, 110, 129, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 175, 176, 177, 178, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 291, 292, 293, 294], "bbox": [0.48426562500000003, 0.09213457076566126, 0.838390625, 0.8224825986078885], "iscrowd": 0, "area": 23131.694700000022, "rle": {"size": [431, 640], "counts": "hiR42^=3M4K4M3L10O00O10000O10000O10000O10000000000000000001O00000000nJ@UM`0k2ATM?l2ATM?m2@SM`0m2@SM`0m2ARM?n2ARM?n2ARM?n2ARM?n2ARM?n2ARM?n2ARM?n2ARM?n2ARM?n2ARM?n2ARM?n2ARM?n2ARM?n2BQM>o2BQM>o2BQM>X1@`K2Y3=W1A`K2Y3=X1@_K3Y3=X1@_K3Y3=X1@_K3Y3=X1A^K2Z3fLe5@SKNiNl3`0fLd5AZKj3ROeLd5AYKk3SOdLe5@WKm3TOcLe5@VKn3UObLZ7_3fHaLY7`3gH`LX7a3hH_LW7b3iH^LV7c3jH]LU7c3lH]LS7d3mH\\LR7e3nH[LR7e3nH[LQ7f3oHZLP7g3PIYLo6h3PIYLo6h3QIXLn6i3RIWLm6j3SIVLl6j3UIVLj6k3VIULi6l3WITLh6m3XISLg6n3YIRLf6o3ZIQLe6P4[IPLd6Q4\\IoKd6Q4\\IoKc6R4]InKb6R4_InK`6S4`ImK_6T4aIlK^6U4bIkK]6V4cIjK\\6W4dIiK[6X4eIhKZ6Y4eIhKZ6Y4fIgKY6Z4gIfKX6Z4iIfKV6[4jIeKU6\\4kIdKU6\\4kIdKT6]4lIcKS6^4mIbKR6_4nIaKQ6`4oI`KP6a4PJ_Ko5b4QJ^Kn5b4SJ^Kl5c4TJ]Kk5d4UJ\\Kj5e4VJ[Ki5f4WJZKh5g4XJYKh5g4XJYKk5d4UJ\\Km5?YIn2i0dLP6<[Il2e0hLP6<_Ig2b0mLo5QMo5 in this image.", "objects": [{"patches": [60, 83, 105, 106, 107, 109, 110, 129, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 175, 176, 177, 178, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 291, 292, 293, 294], "bbox": [0.48426562500000003, 0.09213457076566126, 0.838390625, 0.8224825986078885], "iscrowd": 0, "area": 23131.694700000022, "rle": {"size": [431, 640], "counts": "hiR42^=3M4K4M3L10O00O10000O10000O10000O10000000000000000001O00000000nJ@UM`0k2ATM?l2ATM?m2@SM`0m2@SM`0m2ARM?n2ARM?n2ARM?n2ARM?n2ARM?n2ARM?n2ARM?n2ARM?n2ARM?n2ARM?n2ARM?n2ARM?n2ARM?n2BQM>o2BQM>o2BQM>X1@`K2Y3=W1A`K2Y3=X1@_K3Y3=X1@_K3Y3=X1@_K3Y3=X1A^K2Z3fLe5@SKNiNl3`0fLd5AZKj3ROeLd5AYKk3SOdLe5@WKm3TOcLe5@VKn3UObLZ7_3fHaLY7`3gH`LX7a3hH_LW7b3iH^LV7c3jH]LU7c3lH]LS7d3mH\\LR7e3nH[LR7e3nH[LQ7f3oHZLP7g3PIYLo6h3PIYLo6h3QIXLn6i3RIWLm6j3SIVLl6j3UIVLj6k3VIULi6l3WITLh6m3XISLg6n3YIRLf6o3ZIQLe6P4[IPLd6Q4\\IoKd6Q4\\IoKc6R4]InKb6R4_InK`6S4`ImK_6T4aIlK^6U4bIkK]6V4cIjK\\6W4dIiK[6X4eIhKZ6Y4eIhKZ6Y4fIgKY6Z4gIfKX6Z4iIfKV6[4jIeKU6\\4kIdKU6\\4kIdKT6]4lIcKS6^4mIbKR6_4nIaKQ6`4oI`KP6a4PJ_Ko5b4QJ^Kn5b4SJ^Kl5c4TJ]Kk5d4UJ\\Kj5e4VJ[Ki5f4WJZKh5g4XJYKh5g4XJYKk5d4UJ\\Km5?YIn2i0dLP6<[Il2e0hLP6<_Ig2b0mLo5QMo5 in this image.", "objects": [{"patches": [159, 180, 181, 182, 183, 202, 203, 204, 205, 206, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 296, 297], "bbox": [0.817421875, 0.39931249999999996, 0.9893281249999999, 0.7139166666666665], "iscrowd": 0, "area": 12494.183099999995, "rle": {"size": [480, 640], "counts": "c^e72a<1dE6Y:3ZE3e:5mD2Q;8aDO\\;:WDLg;=kCJS in this image.", "objects": [{"patches": [159, 180, 181, 182, 183, 202, 203, 204, 205, 206, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 296, 297], "bbox": [0.817421875, 0.39931249999999996, 0.9893281249999999, 0.7139166666666665], "iscrowd": 0, "area": 12494.183099999995, "rle": {"size": [480, 640], "counts": "c^e72a<1dE6Y:3ZE3e:5mD2Q;8aDO\\;:WDLg;=kCJS in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 92, 93, 94, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 115, 116, 117, 121, 122, 123, 124, 125, 126, 138, 139, 140, 146, 147, 148, 162, 163], "bbox": [0.001953125, 0.0051875, 0.6818125, 0.4285625], "iscrowd": 0, "area": 53148.9449, "rle": {"size": [480, 640], "counts": "Sd0c0`LS`o2"}, "label": "a silver and blue van is parked next to a motorcycle"}]} {"id": 288310, "image": "COCO_train2014_000000288310.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ash color van named ' service department ' is parked in a parking bay\"."}], "task": "refering", "answer_template": "The \"a ash color van named ' service department ' is parked in a parking bay\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 92, 93, 94, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 115, 116, 117, 121, 122, 123, 124, 125, 126, 138, 139, 140, 146, 147, 148, 162, 163], "bbox": [0.001953125, 0.0051875, 0.6818125, 0.4285625], "iscrowd": 0, "area": 53148.9449, "rle": {"size": [480, 640], "counts": "Sd0c0`LS`o2"}, "label": "a ash color van named ' service department ' is parked in a parking bay"}]} {"id": 566847, "image": "COCO_train2014_000000566847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffee with the old giraffee in the farm\"."}], "task": "refering", "answer_template": "The \"a giraffee with the old giraffee in the farm\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 159, 174, 175, 190, 191, 205, 206, 220, 221, 235, 236, 250, 251, 265, 266], "bbox": [0.582576112412178, 0.418765625, 0.8062529274004684, 0.76925], "iscrowd": 0, "area": 6644.6158, "rle": {"size": [640, 427], "counts": "clk42nc01N2N200O100O10000O1FGm\\O:Pc0Mi\\O6Tc0>M3M2NO3N1O2N2N3M01O2N101N1O16K3L01O010O00001O0T^O@W`0e0`_OGX`0>__ONZ`05^_O7Z`0N]_O>[`0G\\_Oe0\\`0h1H8H8IS1lN6J10N3M4\\MdAjM1c0^>X1TBbMMo0R>X1WCaNm<_1VCoMRNLk>V2aCgMcX40010O01hNg@gMZ?P2WAhMh>l1gAQNZ>o1hAjM]>Y2c13N2l_O\\Mg>g2o@_MQ?d2`@iM`?k0R@n0n?VOR@g0n?\\OS@XOO6o?e0S@nN;0d?V1Q@eNf0H]?f1n_O[NZa0g1h^OSNYa0P2h^OjM[a0Y254K in this image.", "objects": [{"patches": [144, 159, 174, 175, 190, 191, 205, 206, 220, 221, 235, 236, 250, 251, 265, 266], "bbox": [0.582576112412178, 0.418765625, 0.8062529274004684, 0.76925], "iscrowd": 0, "area": 6644.6158, "rle": {"size": [640, 427], "counts": "clk42nc01N2N200O100O10000O1FGm\\O:Pc0Mi\\O6Tc0>M3M2NO3N1O2N2N3M01O2N101N1O16K3L01O010O00001O0T^O@W`0e0`_OGX`0>__ONZ`05^_O7Z`0N]_O>[`0G\\_Oe0\\`0h1H8H8IS1lN6J10N3M4\\MdAjM1c0^>X1TBbMMo0R>X1WCaNm<_1VCoMRNLk>V2aCgMcX40010O01hNg@gMZ?P2WAhMh>l1gAQNZ>o1hAjM]>Y2c13N2l_O\\Mg>g2o@_MQ?d2`@iM`?k0R@n0n?VOR@g0n?\\OS@XOO6o?e0S@nN;0d?V1Q@eNf0H]?f1n_O[NZa0g1h^OSNYa0P2h^OjM[a0Y254K in this image.", "objects": [{"patches": [77, 78, 92, 93, 108, 109, 123, 124, 125, 126, 139, 140, 141, 142, 155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 175, 185, 186, 187, 188, 189, 190, 201, 202, 203, 204, 216, 217, 219, 231, 232, 233, 234, 246, 247, 248, 249, 261, 263, 264], "bbox": [0.15674473067915692, 0.23895312500000002, 0.6769789227166276, 0.749984375], "iscrowd": 0, "area": 20624.609200000006, "rle": {"size": [640, 427], "counts": "PQZ15kc03M2O2N1O001O001N8\\Ob0J001O1O1O0:GO1O1FlNe]OV1Zb0nNb]OT1[b0:O1N2O20O2N1O00O2NkNi]Ob0Wb0\\Om]Ob0Qb0^OQ^Ob0na0\\OU^Od0ja0[OX^Of0fa0YO\\^Og0ca0XO_^Oh0`a0WOa^Oj0_a0TOc^Ol0`a0oNb^OR1Vb0000O101O0000000O2O0000000O2O0000001O0O1000001O0O1000001O0O1000001O000O101O00000O101O0000000`^O\\Nb`0f1[_O]Nd`0c1Y_ObNe`0^1X_OgNf`0Y1X_OjNg`0W1U_OnNh`0S1U_OQOj`0P1S_OTOk`0l0R_OXOm`0h0P_O]Om`0n1N1N3M2O2M3M2O2M2N3M3N1N6J5L5J5XC_Kg9g4QF[KP:i4iEYKW:l4bEVK^:P5ZERKg:R5REPKn:X5hDjJX;V6fClIZ^DXOc;i0eDmN[;V1kDcMS<_2f23N1N3M2O2M2N3O1N1O2O0O2O0O3N3L[Na@kN\\?T1i@kNT?T1QAkNl>T1YAjNd>V1aAiN\\>V1iAiNT>V1RBgNl=X1YBgNd=X1aBgN\\=X1iBfNU=Y1nBhNn_2]A`Mi>d2n@[MX?a35M4MaAoKS=n3kBWLV=e3jB^LV=`3iBcLX=Y3iBiLW=T3iBoLW=[3]BgLc=d3oA_LQ>h41O0O2`NeAXM^>c2iAWM[>d2lAWMW>d2PBWMS>d2dBgL^=U3j1K5K5G9_O`0A`0E;E;EQne2"}, "label": "a larger giraffe that has a smaller giraffe leaning on it"}]} {"id": 566847, "image": "COCO_train2014_000000566847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large giraffe standing next to baby\"."}], "task": "refering", "answer_template": "The \"a large giraffe standing next to baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 92, 93, 108, 109, 123, 124, 125, 126, 139, 140, 141, 142, 155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 175, 185, 186, 187, 188, 189, 190, 201, 202, 203, 204, 216, 217, 219, 231, 232, 233, 234, 246, 247, 248, 249, 261, 263, 264], "bbox": [0.15674473067915692, 0.23895312500000002, 0.6769789227166276, 0.749984375], "iscrowd": 0, "area": 20624.609200000006, "rle": {"size": [640, 427], "counts": "PQZ15kc03M2O2N1O001O001N8\\Ob0J001O1O1O0:GO1O1FlNe]OV1Zb0nNb]OT1[b0:O1N2O20O2N1O00O2NkNi]Ob0Wb0\\Om]Ob0Qb0^OQ^Ob0na0\\OU^Od0ja0[OX^Of0fa0YO\\^Og0ca0XO_^Oh0`a0WOa^Oj0_a0TOc^Ol0`a0oNb^OR1Vb0000O101O0000000O2O0000000O2O0000001O0O1000001O0O1000001O0O1000001O000O101O00000O101O0000000`^O\\Nb`0f1[_O]Nd`0c1Y_ObNe`0^1X_OgNf`0Y1X_OjNg`0W1U_OnNh`0S1U_OQOj`0P1S_OTOk`0l0R_OXOm`0h0P_O]Om`0n1N1N3M2O2M3M2O2M2N3M3N1N6J5L5J5XC_Kg9g4QF[KP:i4iEYKW:l4bEVK^:P5ZERKg:R5REPKn:X5hDjJX;V6fClIZ^DXOc;i0eDmN[;V1kDcMS<_2f23N1N3M2O2M2N3O1N1O2O0O2O0O3N3L[Na@kN\\?T1i@kNT?T1QAkNl>T1YAjNd>V1aAiN\\>V1iAiNT>V1RBgNl=X1YBgNd=X1aBgN\\=X1iBfNU=Y1nBhNn_2]A`Mi>d2n@[MX?a35M4MaAoKS=n3kBWLV=e3jB^LV=`3iBcLX=Y3iBiLW=T3iBoLW=[3]BgLc=d3oA_LQ>h41O0O2`NeAXM^>c2iAWM[>d2lAWMW>d2PBWMS>d2dBgL^=U3j1K5K5G9_O`0A`0E;E;EQne2"}, "label": "a large giraffe standing next to baby"}]} {"id": 263744, "image": "COCO_train2014_000000263744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an apple between two oranges\"."}], "task": "refering", "answer_template": "The \"an apple between two oranges\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 64, 83, 99, 100, 101, 102, 118, 119], "bbox": [0.54232, 0.18792, 0.7014400000000001, 0.51744], "iscrowd": 0, "area": 2451.923749999999, "rle": {"size": [375, 500], "counts": "T]S31b;5N2bMIZI7e6KZI6e6KYI7e6J[I7c6J\\I8b6H_I8a6H^I:`6G`I:^6FcI:]6FbI<\\6DeI=Y6DfI=Z6CfI>X6BhI?W6BiI?V6@kIa0S6@lIa0T6_OlIb0R6^OnId0Q6\\OoId0P6]OoIe0P6ZOQJg0n5YOQJi0m5WOSJj0m5VORJl0l5UOTJl0k5SOUJo0i5ROVJo0j5POWJQ1g5POXJQ1g5POXJQ1h5nNXJS1g5nNYJQ1h5nNXJR1i5nNVJR1k5nNUJQ1l5nNTJR1n5mNQJR1Q6nNnIR1S6mNnIR1T6mNkIS1V6lNjIT1W6lNiIS1Y6lNfIT1[6kNeIU1\\6jNdIV1_8O01O01O00010O3M3N1N3M3M2O2N2N1O2N1O1O2N1O1O1O1O1O2N1O1N2O1Ochf1"}, "label": "an apple between two oranges"}]} {"id": 263744, "image": "COCO_train2014_000000263744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange with a sticker in a black and white photo , under a banana\"."}], "task": "refering", "answer_template": "The \"an orange with a sticker in a black and white photo , under a banana\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 66, 67, 68, 69, 70, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 121, 122, 123], "bbox": [0.65764, 0.17176, 0.906, 0.4890133333333333], "iscrowd": 0, "area": 10280.337650000001, "rle": {"size": [375, 500], "counts": "hch32d;4L4L3M3M2N2N2N3M2N2N2N2N2N3N1N2N2N2M3N3M2M3N2M2O1O2M2O2N1N3N1O2M2O2N1N3N1RO^MfHc2[7]MbHf2]7[MaHf2_7[M_Hg2`7[M\\Hg2d7ZMXHj2g7WMRHo2n7>O001O00010O000000000000000001O000000000000000000001O00000000000000000O101O001N101O1O0O2O1O001N2O001O1N101O1O0O2N2N101N2N1O2O1N1O3M3N2M3M3L4L4K4L5K5L4K5K5B>AeSa0"}, "label": "an orange with a sticker in a black and white photo , under a banana"}]} {"id": 263744, "image": "COCO_train2014_000000263744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange with a sticker on it\"."}], "task": "refering", "answer_template": "The \"orange with a sticker on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 66, 67, 68, 69, 70, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 121, 122, 123], "bbox": [0.65764, 0.17176, 0.906, 0.4890133333333333], "iscrowd": 0, "area": 10280.337650000001, "rle": {"size": [375, 500], "counts": "hch32d;4L4L3M3M2N2N2N3M2N2N2N2N2N3N1N2N2N2M3N3M2M3N2M2O1O2M2O2N1N3N1O2M2O2N1N3N1RO^MfHc2[7]MbHf2]7[MaHf2_7[M_Hg2`7[M\\Hg2d7ZMXHj2g7WMRHo2n7>O001O00010O000000000000000001O000000000000000000001O00000000000000000O101O001N101O1O0O2O1O001N2O001O1N101O1O0O2N2N101N2N1O2O1N1O3M3N2M3M3L4L4K4L5K5L4K5K5B>AeSa0"}, "label": "orange with a sticker on it"}]} {"id": 263744, "image": "COCO_train2014_000000263744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an upside down apple beneath an orange\"."}], "task": "refering", "answer_template": "The \"an upside down apple beneath an orange\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 119, 120, 121, 122, 137, 138, 139, 140, 155, 156, 157, 158, 173, 174, 175, 176, 191], "bbox": [0.61154, 0.45282666666666666, 0.8229399999999999, 0.7826933333333334], "iscrowd": 0, "area": 8998.843600000002, "rle": {"size": [375, 500], "counts": "nW`37^;3M4M2N3M2N2N3L3N3M2cGPOP6R1jIUOS6n0gIXOV6j0eIZOZ6i0`I[O`6f0[I^Od6e0VI_Oj6b0nHFP7 in this image.", "objects": [{"patches": [7, 8, 24, 25, 26, 27, 41, 42, 43, 44, 45, 59, 60, 61, 62, 63], "bbox": [0.32294, 0.061173333333333337, 0.55058, 0.3153066666666667], "iscrowd": 0, "area": 8516.792650000001, "rle": {"size": [375, 500], "counts": "lPk11d;?A;E4L4L4L4L4L4L4L4M3M2N2N1O1O2N1O2O0O1O2N1O2N1O101N100O2O0O100O10O0100O100O10O0100O100000O10O10000000O010000000O10O100000O10001O0O1000000O2O00000O10000O2O00000O10001O0O1000000O101O000O10000O101N100O1O2N3N2M3M3M2N3N2L4L8G:G9GhZb2"}, "label": "orange behind the bannana"}]} {"id": 263744, "image": "COCO_train2014_000000263744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange in a fruit bowl\"."}], "task": "refering", "answer_template": "The \"an orange in a fruit bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 24, 25, 26, 27, 41, 42, 43, 44, 45, 59, 60, 61, 62, 63], "bbox": [0.32294, 0.061173333333333337, 0.55058, 0.3153066666666667], "iscrowd": 0, "area": 8516.792650000001, "rle": {"size": [375, 500], "counts": "lPk11d;?A;E4L4L4L4L4L4L4L4M3M2N2N1O1O2N1O2O0O1O2N1O2N1O101N100O2O0O100O10O0100O100O10O0100O100000O10O10000000O010000000O10O100000O10001O0O1000000O2O00000O10000O2O00000O10001O0O1000000O101O000O10000O101N100O1O2N3N2M3M3M2N3N2L4L8G:G9GhZb2"}, "label": "an orange in a fruit bowl"}]} {"id": 419396, "image": "COCO_train2014_000000419396.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue and gold vase\"."}], "task": "refering", "answer_template": "The \"blue and gold vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 28, 29, 47, 48, 49, 50, 51, 52, 71, 72, 73, 74, 93, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 163, 164, 165, 166, 187, 188, 189, 209, 210, 211, 212, 233], "bbox": [0.035953125, 0.143125, 0.303375, 0.91615625], "iscrowd": 0, "area": 25488.366800000007, "rle": {"size": [320, 640], "counts": "lY71n92N2N3N1N3M2N2TGAU8a0fGDY8P1a1P4nMWK in this image.", "objects": [{"patches": [24, 25, 26, 27, 28, 29, 47, 48, 49, 50, 51, 52, 71, 72, 73, 74, 93, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 163, 164, 165, 166, 187, 188, 189, 209, 210, 211, 212, 233], "bbox": [0.035953125, 0.143125, 0.303375, 0.91615625], "iscrowd": 0, "area": 25488.366800000007, "rle": {"size": [320, 640], "counts": "lY71n92N2N3N1N3M2N2TGAU8a0fGDY8P1a1P4nMWK in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 30, 31, 32, 33, 34, 35, 36, 37, 38, 53, 54, 55, 56, 57, 58, 59, 60, 61, 76, 77, 78, 79, 80, 81, 82, 83, 84, 99, 100, 101, 102, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 129, 130, 145, 146, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 237, 238, 239, 240, 241, 242, 243, 244, 245], "bbox": [0.323375, 0.013000000000000001, 0.670125, 0.9818125], "iscrowd": 0, "area": 64673.104849999996, "rle": {"size": [320, 640], "counts": "omP2?m8o0mNS1lMS2mMQ2nMj1TOl000000000O1000000O1000000O1001O0000000000001O000000000000001O0000000000001O0000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O10000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O000001O0001O000000004LY1gNY1gNX1hNY1gNY1gNX1iNX1hNadQ2"}, "label": "the floral pattern detail close - up"}]} {"id": 419396, "image": "COCO_train2014_000000419396.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zoomed in section of vase\"."}], "task": "refering", "answer_template": "The \"zoomed in section of vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 30, 31, 32, 33, 34, 35, 36, 37, 38, 53, 54, 55, 56, 57, 58, 59, 60, 61, 76, 77, 78, 79, 80, 81, 82, 83, 84, 99, 100, 101, 102, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 129, 130, 145, 146, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 237, 238, 239, 240, 241, 242, 243, 244, 245], "bbox": [0.323375, 0.013000000000000001, 0.670125, 0.9818125], "iscrowd": 0, "area": 64673.104849999996, "rle": {"size": [320, 640], "counts": "omP2?m8o0mNS1lMS2mMQ2nMj1TOl000000000O1000000O1000000O1001O0000000000001O000000000000001O0000000000001O0000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O10000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O000001O0001O000000004LY1gNY1gNX1hNY1gNY1gNX1iNX1hNadQ2"}, "label": "zoomed in section of vase"}]} {"id": 460362, "image": "COCO_train2014_000000460362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing blue dress serving some sweets\"."}], "task": "refering", "answer_template": "The \"a man wearing blue dress serving some sweets\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 30, 31, 32, 33, 53, 54, 55, 56, 76, 77, 78, 79, 99, 100, 101, 102, 122, 123, 124, 146, 147, 169, 170, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 221, 222, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 287, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 326, 327, 328, 329, 330, 331], "bbox": [0.18671875, 0.030070588235294116, 0.662296875, 0.9849647058823529], "iscrowd": 0, "area": 42596.687650000014, "rle": {"size": [425, 640], "counts": "dVb17j<7J7H8I7H8K5N2O1O1O1N2O1O1O1N2O1O1O1N101O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O001N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N101O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O0O2]IQLk3P4kK_Lo3b3gKmLS4T3kKSMo3n2nKZMl3g2RL`Mh3a2VLiNb2W1\\MlNb2U1\\MnNb2S1[MQOc2P1[MROd2o0ZMTOe2l0YMWOe2j0YMYOe2h0ZMZOd2g0[M[Od2e0\\M\\Ob2d0^M^O`2c0`M^O^2c0aM_O]2b0cM_O\\2a0cMA[2`0eMAY2`0fMBX2?hMAW2`0hMBW2>iMCU2>jMDT2=lMDR2=mMER2;nMEQ2_NhNKj0f1a0\\NgNLi0e1b0^NgNLh0c1d0`NeNLg0c1e0aNdNLh0`1f0cNdNLg0^1h0eNbNLf0^1i0fNaNLg0[1j0iN`NKg0Y1l0jN^NMg0W1l0lN^NLf0V1n0nN]NKf0R1R1QOYNMf0m0U1VOVNMd0h0[1[OQNMg0a0]1@nMNm05Z1MiMNU1HW1:eMM]1\\OS1f0aMMR2bNb0`1]MNa62`IMa62_INa61aIN`61aIN_62aIN_62bIM_61bIO^61cIN]62dIM]61dIO\\61eIN[62eIN\\61eIN[62eIOZ61gINZ62fIMZ63fIMZ63gILZ63fIMZ63gILY64hIKY64gILY64hIKX65hIKY65gIJY66hIIX67hIIY66hIIX67hIIX67iIHX67hIIX67iIHX67iIHW68iIIV67kIHV67jIIV67kIHU68lIGU68kIHU68lIGT69lIGT69mIFS6:mIFS6:nIER6;oIDQ6QJBo5>RJAn5?RJAn5?RJAn5?RJAn5?SJ@m5`0SJAk5`0UJ@k5`0UJ@k5`0VJ_Oj5a0VJ_Oj5a0VJ_Oj5a0VJ_Oj5a0WJ^Oi5b0WJ^Oh5c0XJ]Oh5c0XJ]Oh5c0YJ\\Og5d0YJ\\Og5d0YJ\\Og5d0YJ\\Og5d0ZJ[Of5e0ZJ[Oe5f0[JZOe5f0[JZOe5f0\\JYOd5g0\\JZOc5f0]JZOc5f0]JZOc5f0^JYOb5g0^JYOb5g0^JYOb5g0^JYOa5h0`JWO`5i0`JWO`5i0`JWO`5i0`30000000000O1000O100000O100000000O10000000001N101O00001O0O2O001O0O1N3M2O2M2O2M6K5J6Je\\i2"}, "label": "a man wearing blue dress serving some sweets"}]} {"id": 460362, "image": "COCO_train2014_000000460362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the bald head\"."}], "task": "refering", "answer_template": "The \"the man with the bald head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 30, 31, 32, 33, 53, 54, 55, 56, 76, 77, 78, 79, 99, 100, 101, 102, 122, 123, 124, 146, 147, 169, 170, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 221, 222, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 287, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 326, 327, 328, 329, 330, 331], "bbox": [0.18671875, 0.030070588235294116, 0.662296875, 0.9849647058823529], "iscrowd": 0, "area": 42596.687650000014, "rle": {"size": [425, 640], "counts": "dVb17j<7J7H8I7H8K5N2O1O1O1N2O1O1O1N2O1O1O1N101O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O001N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N101O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O0O2]IQLk3P4kK_Lo3b3gKmLS4T3kKSMo3n2nKZMl3g2RL`Mh3a2VLiNb2W1\\MlNb2U1\\MnNb2S1[MQOc2P1[MROd2o0ZMTOe2l0YMWOe2j0YMYOe2h0ZMZOd2g0[M[Od2e0\\M\\Ob2d0^M^O`2c0`M^O^2c0aM_O]2b0cM_O\\2a0cMA[2`0eMAY2`0fMBX2?hMAW2`0hMBW2>iMCU2>jMDT2=lMDR2=mMER2;nMEQ2_NhNKj0f1a0\\NgNLi0e1b0^NgNLh0c1d0`NeNLg0c1e0aNdNLh0`1f0cNdNLg0^1h0eNbNLf0^1i0fNaNLg0[1j0iN`NKg0Y1l0jN^NMg0W1l0lN^NLf0V1n0nN]NKf0R1R1QOYNMf0m0U1VOVNMd0h0[1[OQNMg0a0]1@nMNm05Z1MiMNU1HW1:eMM]1\\OS1f0aMMR2bNb0`1]MNa62`IMa62_INa61aIN`61aIN_62aIN_62bIM_61bIO^61cIN]62dIM]61dIO\\61eIN[62eIN\\61eIN[62eIOZ61gINZ62fIMZ63fIMZ63gILZ63fIMZ63gILY64hIKY64gILY64hIKX65hIKY65gIJY66hIIX67hIIY66hIIX67hIIX67iIHX67hIIX67iIHX67iIHW68iIIV67kIHV67jIIV67kIHU68lIGU68kIHU68lIGT69lIGT69mIFS6:mIFS6:nIER6;oIDQ6QJBo5>RJAn5?RJAn5?RJAn5?RJAn5?SJ@m5`0SJAk5`0UJ@k5`0UJ@k5`0VJ_Oj5a0VJ_Oj5a0VJ_Oj5a0VJ_Oj5a0WJ^Oi5b0WJ^Oh5c0XJ]Oh5c0XJ]Oh5c0YJ\\Og5d0YJ\\Og5d0YJ\\Og5d0YJ\\Og5d0ZJ[Of5e0ZJ[Oe5f0[JZOe5f0[JZOe5f0\\JYOd5g0\\JZOc5f0]JZOc5f0]JZOc5f0^JYOb5g0^JYOb5g0^JYOb5g0^JYOa5h0`JWO`5i0`JWO`5i0`JWO`5i0`30000000000O1000O100000O100000000O10000000001N101O00001O0O2O001O0O1N3M2O2M2O2M6K5J6Je\\i2"}, "label": "the man with the bald head"}]} {"id": 460362, "image": "COCO_train2014_000000460362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"guy clicking picture of pizza with dslr camera\"."}], "task": "refering", "answer_template": "The \"guy clicking picture of pizza with dslr camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 53, 69, 70, 71, 72, 73, 74, 75, 76, 92, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 322, 323, 324, 325], "bbox": [0.0015, 0.024258823529411767, 0.39723437500000003, 0.9881411764705882], "iscrowd": 0, "area": 71046.78005, "rle": {"size": [425, 640], "counts": "Z?R;W2000001O2N2N1O001O1O1O1O1O1O1O000000O100N2M3M3N2M3M3M3L4M3M3M3O1O1O1O100O1O2N1O100O1O1O1O100O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1000000000000000000000ABSD>l;DSDl4@UK>n4_OSKa0o4]OQKc0Q5[OoJe0S5YOnJf0T5XOlJg0W5WOiJi0Y5UOgJk0[5SOeJm0]5RObJn0`5PO`Jo0c5oN]JQ1e5mN[JS1g5kNYJU1i5iNWJTNFU2U6EVJSNJU2R6FVJQNMV2o5GVJoM0W2l5HUJnM4W2i5JTJkM8X2f5^OeKa0]4nNTLP1n3fM[MY2k600O2N1O1O1O001O1O001O1O00100O001O001O1O1O1O1O1O1O00100O1O1O1O1O001OO1O100O101N10N1N4L5L3L4M4K4L6K4K5K6K4KfPP5"}, "label": "guy clicking picture of pizza with dslr camera"}]} {"id": 460362, "image": "COCO_train2014_000000460362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man taking the photography of the pizza\"."}], "task": "refering", "answer_template": "The \"a man taking the photography of the pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 53, 69, 70, 71, 72, 73, 74, 75, 76, 92, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 322, 323, 324, 325], "bbox": [0.0015, 0.024258823529411767, 0.39723437500000003, 0.9881411764705882], "iscrowd": 0, "area": 71046.78005, "rle": {"size": [425, 640], "counts": "Z?R;W2000001O2N2N1O001O1O1O1O1O1O1O000000O100N2M3M3N2M3M3M3L4M3M3M3O1O1O1O100O1O2N1O100O1O1O1O100O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1000000000000000000000ABSD>l;DSDl4@UK>n4_OSKa0o4]OQKc0Q5[OoJe0S5YOnJf0T5XOlJg0W5WOiJi0Y5UOgJk0[5SOeJm0]5RObJn0`5PO`Jo0c5oN]JQ1e5mN[JS1g5kNYJU1i5iNWJTNFU2U6EVJSNJU2R6FVJQNMV2o5GVJoM0W2l5HUJnM4W2i5JTJkM8X2f5^OeKa0]4nNTLP1n3fM[MY2k600O2N1O1O1O001O1O001O1O00100O001O001O1O1O1O1O1O1O00100O1O1O1O1O001OO1O100O101N10N1N4L5L3L4M4K4L6K4K5K6K4KfPP5"}, "label": "a man taking the photography of the pizza"}]} {"id": 17997, "image": "COCO_train2014_000000017997.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe standing in the middle\"."}], "task": "refering", "answer_template": "The \"the giraffe standing in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 74, 75, 76, 77, 78, 100, 101, 102, 103, 124, 125, 126, 127, 147, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 216, 217, 218, 219, 241, 242], "bbox": [0.251453125, 0.1674617737003058, 0.532234375, 0.8579510703363915], "iscrowd": 0, "area": 11341.658749999997, "rle": {"size": [327, 640], "counts": "Y`c12R:3M3M3N2O1N2O1N2O1N1O2O1N2O0O2N2O1O0100000000O0100O1N101N2O1N1011O1O1N2oNTGe0W90000001O1N2O001O1O1O01000000O1000000O1000000O1000000O1000000O1000O100O1O10O01O100O1O10O01O100O100O1O010O1O100^GmNn7T1lGTOQ8l0kG[OS8e0gGCV8Z1N2M3M3N2M3N2M3M3N2M3N2M3N2M;F=B9H9F9H8G3NO00010O0hMY26I7J5K6J6J6J6J5K6J6J2N00O1XN^KjMb4Y1hLZNX3g1c2101N2N2O0O2N2O0O2O1N1O2O1N1O2O1O001O1O1O001O1O004K8I6J7I7I6J3M00O1O1O100O1fMnI=S6ROeJf0\\5iN\\KP1d4eNnKS1S4hNZLo0\\7TOiVo2"}, "label": "the giraffe standing in the middle"}]} {"id": 17997, "image": "COCO_train2014_000000017997.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe in between two other giraffes\"."}], "task": "refering", "answer_template": "The \"a giraffe in between two other giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 74, 75, 76, 77, 78, 100, 101, 102, 103, 124, 125, 126, 127, 147, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 216, 217, 218, 219, 241, 242], "bbox": [0.251453125, 0.1674617737003058, 0.532234375, 0.8579510703363915], "iscrowd": 0, "area": 11341.658749999997, "rle": {"size": [327, 640], "counts": "Y`c12R:3M3M3N2O1N2O1N2O1N1O2O1N2O0O2N2O1O0100000000O0100O1N101N2O1N1011O1O1N2oNTGe0W90000001O1N2O001O1O1O01000000O1000000O1000000O1000000O1000000O1000O100O1O10O01O100O1O10O01O100O100O1O010O1O100^GmNn7T1lGTOQ8l0kG[OS8e0gGCV8Z1N2M3M3N2M3N2M3M3N2M3N2M3N2M;F=B9H9F9H8G3NO00010O0hMY26I7J5K6J6J6J6J5K6J6J2N00O1XN^KjMb4Y1hLZNX3g1c2101N2N2O0O2N2O0O2O1N1O2O1N1O2O1O001O1O1O001O1O004K8I6J7I7I6J3M00O1O1O100O1fMnI=S6ROeJf0\\5iN\\KP1d4eNnKS1S4hNZLo0\\7TOiVo2"}, "label": "a giraffe in between two other giraffes"}]} {"id": 17997, "image": "COCO_train2014_000000017997.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe closest to the gate\"."}], "task": "refering", "answer_template": "The \"the giraffe closest to the gate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 86, 87, 88, 107, 108, 109, 110, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 177, 197, 198, 200, 201, 220, 221, 223, 224, 243, 244, 246, 247], "bbox": [0.559375, 0.17990825688073395, 0.847453125, 0.9112844036697249], "iscrowd": 0, "area": 12685.696649999994, "rle": {"size": [327, 640], "counts": "[ab31T:4L3M4L3N3L3M000000O2O00000O10`H_O\\5a0`JE_5:^JL_54]J3`5L]J:a5F\\J`0a5@[Jf0d5YOYJm0d5ROYJR1g5nNVJS1l5mNPJV1Q6jNlIW1V6hNjIX1V6iNjIV1W6jNiIT1Y6nNeIQ1\\6ROaIl0a6XOZIh0g6\\OUIb0m6API>Q7FkH8W7LeH3\\70aHN`7[11O35J6J6J7I6J6J6J6J6J6K4KUN\\JUOc5h0kJnNT5o0ZKgNf4U1hKaNX4[1VL[Ni3b1eLTN[3h1d2M3000O10O100000O10O10000000O01000000O0101O00000N2O1O1O1N2O1O1O2N1hHNY44[K8d4g2N2N1N3OTO_KRLa4h3PLnKo3m3bLhK]3S4^1K5J6K5K5M33M3M3M2N3N2M3M3M_NRJlNk5R1[JmNb5R1cJmNZ5Q1lJnNQ5P1UKnNi4P1]KoN_4P1hKnNT4Q1RLnNj3Q1\\LnN`3OlJMj13V3OVKHj18l20_KBk1=b21hK\\Ol1a0Y23PLWOm1d0o15YLQOn1i0e16_OI=6IH480EO;4CK=8AG?<@B`0a0^O^Oa0f0]OYOc0j0[OUOe0n0ZOPOf0S1XOlNh0W1VOhNj0[1TOdNl0_1RO`Nm0d1ROZNn0_500000000O10000000000O10O10000O10O01O1O1O3M2N34L3L10L4L3M4O10O2O1O1\\OTGHo84UGJl82YGLi80[GNf8O^GOc8NaG0a8LcG1`_o0"}, "label": "the giraffe closest to the gate"}]} {"id": 534107, "image": "COCO_train2014_000000534107.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"picture of a giraffe\"."}], "task": "refering", "answer_template": "The \"picture of a giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 51, 52, 53, 54, 55, 56, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 219, 220, 223, 224, 242, 243, 246, 247, 265, 266, 269, 270, 288, 289, 292, 293, 312, 315, 316, 339], "bbox": [0.19954687499999998, 0.0886416861826698, 0.7863125, 0.9414051522248243], "iscrowd": 0, "area": 33541.477950000015, "rle": {"size": [427, 640], "counts": "m^e18R=3M2O1N2M2N3M2N2N1O2O1O001O001O1O000O101N2O1M3L4O1O1O11N2O1O1O00O1O10O01N101O0O2O001N101O1N101O00iN_Df0a;WOcDg0^;UOgDj0Z;ROjDm0k;O0O10000O012O0O2N1O01O00001O00001O00010O00001O00001O000010O0001O00010O00001O01O01O00001O010O00001O01O01O000010O0001O001O01O01O000010O0001O00010O001O000010O0001O00010O00001O010O00001O00010O00001O00010O001O0010O0001O001O010O001O00001O010O001O001O00001O001O010O1O1O1O1O1O2N1O1O1O100WE]Nh9d1UF`Ni9a1TFdNi9]1TFgNj9[1RFjNk9W1RFmNl9T1QFPOm9Z2M2PK_L2c3L^L3d3J]L6d3H]L8e3E]L:f3B[L>i3]OXLc0k3YOWLf0l3VOULk0o3PORLo0S4jNoKV1W4bNkK^1]4YNeKf1f4mM\\KS2o4`MRKa2Y5nLmJR3[5ULUKl3m4dKaK\\4f6100O1HVKoGj4o7:O1L4D in this image.", "objects": [{"patches": [29, 51, 52, 53, 54, 55, 56, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 107, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 219, 220, 223, 224, 242, 243, 246, 247, 265, 266, 269, 270, 288, 289, 292, 293, 312, 315, 316, 339], "bbox": [0.19954687499999998, 0.0886416861826698, 0.7863125, 0.9414051522248243], "iscrowd": 0, "area": 33541.477950000015, "rle": {"size": [427, 640], "counts": "m^e18R=3M2O1N2M2N3M2N2N1O2O1O001O001O1O000O101N2O1M3L4O1O1O11N2O1O1O00O1O10O01N101O0O2O001N101O1N101O00iN_Df0a;WOcDg0^;UOgDj0Z;ROjDm0k;O0O10000O012O0O2N1O01O00001O00001O00010O00001O00001O000010O0001O00010O00001O01O01O00001O010O00001O01O01O000010O0001O001O01O01O000010O0001O00010O001O000010O0001O00010O00001O010O00001O00010O00001O00010O001O0010O0001O001O010O001O00001O010O001O001O00001O001O010O1O1O1O1O1O2N1O1O1O100WE]Nh9d1UF`Ni9a1TFdNi9]1TFgNj9[1RFjNk9W1RFmNl9T1QFPOm9Z2M2PK_L2c3L^L3d3J]L6d3H]L8e3E]L:f3B[L>i3]OXLc0k3YOWLf0l3VOULk0o3PORLo0S4jNoKV1W4bNkK^1]4YNeKf1f4mM\\KS2o4`MRKa2Y5nLmJR3[5ULUKl3m4dKaK\\4f6100O1HVKoGj4o7:O1L4D in this image.", "objects": [{"patches": [151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 243, 244, 245, 246, 247, 248, 249, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 292, 293, 294, 295], "bbox": [0.494359375, 0.39329166666666665, 0.8729374999999999, 0.7608125], "iscrowd": 0, "area": 34723.18884999999, "rle": {"size": [480, 640], "counts": "\\\\d4_1_=4L5oNYNaDk1\\;YN_Dk1];[N]Di1`;[NUDP2g;g0M3M4K4M3M3M4L3M3M4M2M3N2N3L3N2M100O100O10000O10O0100O100O100O1O100O100O00100O1O100O1O100O100O1O10O01O100O10000000000O1000O10000000O1000000000000O10O1000000000O100000000000O0100000000000000O10000000O10000000000000000000000O10000000000000000000000O1000000000000000000000000O10000000000000000000000O10001O0000001O0000001O000O101O00001O1O1O001O1O1O0O2O1O1O001O1O1O001O1O1O001N2O1O1O2N1O1O1O2N1O1O1O2M2O1O1O2N1O1O1O2N1O1O2M2M3M3M4L3L4M4L4L4L4K5L4L4L4L4K5L4L4L5K5J6K5K4L5K5KQUV1"}, "label": "a donut to the right of two other donuts"}]} {"id": 9822, "image": "COCO_train2014_000000009822.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the doughnut in the bottom left corner\"."}], "task": "refering", "answer_template": "The \"the doughnut in the bottom left corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 170, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 301, 302, 303, 304, 305], "bbox": [0.0625, 0.3705, 0.46790624999999997, 0.7916666666666666], "iscrowd": 0, "area": 38545.994399999996, "rle": {"size": [480, 640], "counts": "[Qc03j>6J7I7J6I6K6J6K4K6K5J5L5J6VOj0K3N2M2N2O1N2N3N1N2N2O1N2N2O0O1O2O0O1O2O0O2O0O1O2O0O101N100O2N100O2O0O100O100O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O100O10001O000O100000001N100000000O2O0000000O10O100000O1000O1000O100000O1000O100000O10O100000O1000O100000O10O10000000O10O0100O100O1O010O100O010O01O010O010O00100O00100O00100O010O1O0101O1N2O001N2O1O1N2O1O1N2O1O1O001O1O1O1O1N2O00000000000000000O100000O1000001O1O100O1O1O1O1O1O100O1O1O0000O1O2N1O1O1O100O2N2N3M2N2N2N2O2M2N2N2N2N3M2N2O2M5J5J9G=C in this image.", "objects": [{"patches": [6, 7, 26, 27, 28, 29, 30, 31, 32, 49, 50, 51, 52, 53, 54, 55, 56, 72, 73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147], "bbox": [0.136828125, 0.051791666666666666, 0.4695937499999999, 0.39864583333333337], "iscrowd": 0, "area": 28896.017699999997, "rle": {"size": [480, 640], "counts": "c[Y1 in this image.", "objects": [{"patches": [6, 7, 26, 27, 28, 29, 30, 31, 32, 49, 50, 51, 52, 53, 54, 55, 56, 72, 73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147], "bbox": [0.136828125, 0.051791666666666666, 0.4695937499999999, 0.39864583333333337], "iscrowd": 0, "area": 28896.017699999997, "rle": {"size": [480, 640], "counts": "c[Y1 in this image.", "objects": [{"patches": [158, 159, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 294, 295, 296, 297, 298, 318, 319, 320, 321, 342, 343, 344], "bbox": [0.8203125, 0.39466666666666667, 1.0, 0.845625], "iscrowd": 0, "area": 20553.853750000002, "rle": {"size": [480, 640], "counts": "b]f76e>H8G9XOULPFT4k9d0N2M3N1O1O1O1N101O1O1O0O2O0O1N3M2N3M2N2N3M2N3N1O1O2O0O1O2N101N1O1O2O0O2N1O10010O0000001O0000001O0001O010O1O1O00100O1O00100O1O002N1O1O2N1O2N1O2N1O2N1O001O001O001O001O1N101N101N2O1N2O1N2O1N3N1N2O1N2O1N2O2M2Ok0TOfG"}, "label": "the all blue backpack that the man with a dark green textured hoodie is wearing"}]} {"id": 411238, "image": "COCO_train2014_000000411238.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and yellow bus with an ad on the side\"."}], "task": "refering", "answer_template": "The \"a red and yellow bus with an ad on the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 69, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 119, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170], "bbox": [0.0050625, 0.035958333333333335, 0.667421875, 0.4179791666666667], "iscrowd": 0, "area": 48932.87769999999, "rle": {"size": [480, 640], "counts": "o]1h0XOh0XOf0ZON2N3M2POeCZO\\ in this image.", "objects": [{"patches": [7, 8, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 69, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 119, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170], "bbox": [0.0050625, 0.035958333333333335, 0.667421875, 0.4179791666666667], "iscrowd": 0, "area": 48932.87769999999, "rle": {"size": [480, 640], "counts": "o]1h0XOh0XOf0ZON2N3M2POeCZO\\ in this image.", "objects": [{"patches": [238, 239, 240, 260, 261, 262, 263, 282, 283, 284, 285, 286, 305, 306, 307, 309], "bbox": [0.26603125, 0.6883333333333334, 0.47890625, 0.93962441314554], "iscrowd": 0, "area": 7281.815499999998, "rle": {"size": [426, 640], "counts": "XSW22V=6ZOOhC9P in this image.", "objects": [{"patches": [238, 239, 240, 260, 261, 262, 263, 282, 283, 284, 285, 286, 305, 306, 307, 309], "bbox": [0.26603125, 0.6883333333333334, 0.47890625, 0.93962441314554], "iscrowd": 0, "area": 7281.815499999998, "rle": {"size": [426, 640], "counts": "XSW22V=6ZOOhC9P in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196], "bbox": [0.6505599999999999, 0.48314666666666667, 0.99102, 0.8045066666666667], "iscrowd": 0, "area": 14606.342199999997, "rle": {"size": [375, 500], "counts": "lXg31f;3M4L3M3M3L4M3M3M4L1O1N2O1O1O2NO10O10000000000O1000O10000000O1H8I7I7I7I6I8I7I7I7I7I70000000O2O0000000000000000000O100000001O000000000O1000000000000000000O2O00000000000000000O10000000001O0000000O1000000000000000000O2O00000000000000000O10000000001O0000000O1000000000000000000O101O000000000000000O10000O1K5K5K6J5K5K5K5K5L4K5K5K5K51OK5J6J6K6Ikd1"}, "label": "a silver apple laptop with a sticker on the corner"}]} {"id": 304757, "image": "COCO_train2014_000000304757.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"apple laptop\"."}], "task": "refering", "answer_template": "The \"apple laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196], "bbox": [0.6505599999999999, 0.48314666666666667, 0.99102, 0.8045066666666667], "iscrowd": 0, "area": 14606.342199999997, "rle": {"size": [375, 500], "counts": "lXg31f;3M4L3M3M3L4M3M3M4L1O1N2O1O1O2NO10O10000000000O1000O10000000O1H8I7I7I7I6I8I7I7I7I7I70000000O2O0000000000000000000O100000001O000000000O1000000000000000000O2O00000000000000000O10000000001O0000000O1000000000000000000O2O00000000000000000O10000000001O0000000O1000000000000000000O101O000000000000000O10000O1K5K5K6J5K5K5K5K5L4K5K5K5K51OK5J6J6K6Ikd1"}, "label": "apple laptop"}]} {"id": 304757, "image": "COCO_train2014_000000304757.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a laptop with a black back\"."}], "task": "refering", "answer_template": "The \"a laptop with a black back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 111, 112, 113, 114, 115, 127, 128, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 168, 169, 181, 182, 183, 184, 185, 186, 187, 200, 201, 202, 203, 204, 205], "bbox": [0.03886, 0.46282666666666666, 0.44764, 0.8614933333333333], "iscrowd": 0, "area": 25951.825100000002, "rle": {"size": [375, 500], "counts": "j_75b;5K5K4L5K5K5K5K5K4L5K5K5K5K5K4L5K5K5K5K5K5K4L5K5K5K4L1O001O in this image.", "objects": [{"patches": [108, 109, 110, 111, 112, 113, 114, 115, 127, 128, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 168, 169, 181, 182, 183, 184, 185, 186, 187, 200, 201, 202, 203, 204, 205], "bbox": [0.03886, 0.46282666666666666, 0.44764, 0.8614933333333333], "iscrowd": 0, "area": 25951.825100000002, "rle": {"size": [375, 500], "counts": "j_75b;5K5K4L5K5K5K5K5K4L5K5K5K5K5K4L5K5K5K5K5K5K4L5K5K5K4L1O001O in this image.", "objects": [{"patches": [177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.6493125, 0.4767205542725173, 0.9944062500000002, 0.9733718244803696], "iscrowd": 0, "area": 38884.8405, "rle": {"size": [433, 640], "counts": "eW`56Z=6I8I6J7I6J7I6I7J6K10O101O000000J6C=B>C=H8I7I7J6I7I7L4M5J6J6J3N2M3M3N2M3M3N200O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O1000000O10000O10000O10000O10000O10000O10000O10000O10000O100O10000O1000000O10000O11O0O10001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O1O2N1O1O1O1O1O1O1O1O2N2N2N2N2N2N2N2N2N2N3M3M2N3M3M3M2Ng0YO^1bN=_O`0@a0_OnY1"}, "label": "the silhouette of a person ' s head"}]} {"id": 468602, "image": "COCO_train2014_000000468602.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"head in shadow\"."}], "task": "refering", "answer_template": "The \"head in shadow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.6493125, 0.4767205542725173, 0.9944062500000002, 0.9733718244803696], "iscrowd": 0, "area": 38884.8405, "rle": {"size": [433, 640], "counts": "eW`56Z=6I8I6J7I6J7I6I7J6K10O101O000000J6C=B>C=H8I7I7J6I7I7L4M5J6J6J3N2M3M3N2M3M3N200O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O1000000O10000O10000O10000O10000O10000O10000O10000O10000O100O10000O1000000O10000O11O0O10001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O1O2N1O1O1O1O1O1O1O1O2N2N2N2N2N2N2N2N2N2N3M3M2N3M3M3M2Ng0YO^1bN=_O`0@a0_OnY1"}, "label": "head in shadow"}]} {"id": 329339, "image": "COCO_train2014_000000329339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white car in front of a city bus\"."}], "task": "refering", "answer_template": "The \"a white car in front of a city bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 342, 343, 344], "bbox": [0.68896875, 0.5548125, 1.0, 0.8610208333333332], "iscrowd": 0, "area": 22573.497200000005, "rle": {"size": [480, 640], "counts": "\\Q_6e0X>4M3M3L4M3L4M3L5L3F:N2O1N2O1N2O3L4M3L4M2N3M1O1O1O1O1O1O1O1O000O1O1O100O1O1O100O1O1O1O100O1O00000001O001O001N1011N100O10000O100O10001O000000000O100000001O000O1000000000000O2O0000000000000O10001O0000000O10000000001O0000000O10000000001O0000000O10000000001O001O1O1N2O1O4M3L2N2N1O1O2N000000001O000000000000000000O2O0O100O100O2O0O100O1L4M4L300001O0001O0001O000001O0001O000001O01O000000010O000000001m0ROR2nM^D"}, "label": "a white car in front of a city bus"}]} {"id": 329339, "image": "COCO_train2014_000000329339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light - colored sedan in front of a bus\"."}], "task": "refering", "answer_template": "The \"a light - colored sedan in front of a bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 315, 316, 317, 318, 319, 320, 321, 342, 343, 344], "bbox": [0.68896875, 0.5548125, 1.0, 0.8610208333333332], "iscrowd": 0, "area": 22573.497200000005, "rle": {"size": [480, 640], "counts": "\\Q_6e0X>4M3M3L4M3L4M3L5L3F:N2O1N2O1N2O3L4M3L4M2N3M1O1O1O1O1O1O1O1O000O1O1O100O1O1O100O1O1O1O100O1O00000001O001O001N1011N100O10000O100O10001O000000000O100000001O000O1000000000000O2O0000000000000O10001O0000000O10000000001O0000000O10000000001O0000000O10000000001O001O1O1N2O1O4M3L2N2N1O1O2N000000001O000000000000000000O2O0O100O100O2O0O100O1L4M4L300001O0001O0001O000001O0001O000001O01O000000010O000000001m0ROR2nM^D"}, "label": "a light - colored sedan in front of a bus"}]} {"id": 280191, "image": "COCO_train2014_000000280191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy holding a girl in the chari\"."}], "task": "refering", "answer_template": "The \"a guy holding a girl in the chari\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 103, 104, 105, 106, 126, 127, 128, 129, 130, 132, 148, 149, 150, 151, 152, 155, 156, 171, 172, 173, 174, 175, 179, 180, 194, 195, 196, 197, 198, 202, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 247, 263, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 313, 333, 334, 335, 336], "bbox": [0.45081249999999995, 0.24424242424242423, 0.836296875, 0.9989743589743588], "iscrowd": 0, "area": 34008.51120000001, "rle": {"size": [429, 640], "counts": "d[i3?l<3L5L3L4M4L3L4M4CC=eNZ1K6J6J5K5M2O2N2M2O2N1O2M2O2N1N3N1O2O0O2O1N1SOjJcIW5P6WKmIj4X1hJg2b0oKg4n0ZKj22ULe4f0mKj2B^Lb4C[On:7\\Ea0DYOQ;6YEb0FXOR;7UEc0GWOU;6REe0GVOX;5mDh0KSOY;5gDV1Z;>1N2N2O1N2O1N3K4I7J6I7I7J6I7Ibi[1"}, "label": "a guy holding a girl in the chari"}]} {"id": 280191, "image": "COCO_train2014_000000280191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man sits on a bench with his arm around a smiling girl\"."}], "task": "refering", "answer_template": "The \"a young man sits on a bench with his arm around a smiling girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 103, 104, 105, 106, 126, 127, 128, 129, 130, 132, 148, 149, 150, 151, 152, 155, 156, 171, 172, 173, 174, 175, 179, 180, 194, 195, 196, 197, 198, 202, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 247, 263, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 313, 333, 334, 335, 336], "bbox": [0.45081249999999995, 0.24424242424242423, 0.836296875, 0.9989743589743588], "iscrowd": 0, "area": 34008.51120000001, "rle": {"size": [429, 640], "counts": "d[i3?l<3L5L3L4M4L3L4M4CC=eNZ1K6J6J5K5M2O2N2M2O2N1O2M2O2N1N3N1O2O0O2O1N1SOjJcIW5P6WKmIj4X1hJg2b0oKg4n0ZKj22ULe4f0mKj2B^Lb4C[On:7\\Ea0DYOQ;6YEb0FXOR;7UEc0GWOU;6REe0GVOX;5mDh0KSOY;5gDV1Z;>1N2N2O1N2O1N3K4I7J6I7I7J6I7Ibi[1"}, "label": "a young man sits on a bench with his arm around a smiling girl"}]} {"id": 280191, "image": "COCO_train2014_000000280191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl sitting on a bench near a boy\"."}], "task": "refering", "answer_template": "The \"a girl sitting on a bench near a boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 106, 107, 108, 130, 131, 132, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 198, 199, 200, 201, 202, 221, 222, 223, 224, 225, 245, 246, 247, 248, 268, 269, 270, 271, 291, 292, 293, 294, 314, 315, 316, 317, 337, 338, 340, 341], "bbox": [0.614703125, 0.23543123543123542, 0.8446562500000001, 0.9887878787878788], "iscrowd": 0, "area": 26386.157750000002, "rle": {"size": [429, 640], "counts": "PkT51[=[MX11a1l0gLi1`0\\Me1Q7]OgHhN=h1l6^OVJ?i5@\\J=d5AaJ<_5CdJ;\\5CiJ:W5DnJ9Q5\\O^Ka0j8M3M3M3M3M3M2N3M3M3M3M3M3M`_Y1"}, "label": "a girl sitting on a bench near a boy"}]} {"id": 280191, "image": "COCO_train2014_000000280191.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young girl sitting and smiling on a bench\"."}], "task": "refering", "answer_template": "The \"a young girl sitting and smiling on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 106, 107, 108, 130, 131, 132, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 198, 199, 200, 201, 202, 221, 222, 223, 224, 225, 245, 246, 247, 248, 268, 269, 270, 271, 291, 292, 293, 294, 314, 315, 316, 317, 337, 338, 340, 341], "bbox": [0.614703125, 0.23543123543123542, 0.8446562500000001, 0.9887878787878788], "iscrowd": 0, "area": 26386.157750000002, "rle": {"size": [429, 640], "counts": "PkT51[=[MX11a1l0gLi1`0\\Me1Q7]OgHhN=h1l6^OVJ?i5@\\J=d5AaJ<_5CdJ;\\5CiJ:W5DnJ9Q5\\O^Ka0j8M3M3M3M3M3M2N3M3M3M3M3M3M`_Y1"}, "label": "a young girl sitting and smiling on a bench"}]} {"id": 444033, "image": "COCO_train2014_000000444033.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older woman in a black suit looks upon a spread - out buffet of pizza and other foods\"."}], "task": "refering", "answer_template": "The \"an older woman in a black suit looks upon a spread - out buffet of pizza and other foods\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 118, 119, 120, 141, 142, 143, 164, 165, 166, 167, 187, 188, 189, 190, 210, 211, 212, 213, 233, 234, 235, 236, 257], "bbox": [0.1436875, 0.31354460093896713, 0.28346875, 0.7456338028169014], "iscrowd": 0, "area": 10548.051649999996, "rle": {"size": [426, 640], "counts": "g_V14R=8E;F9G:G7M3M2O2M3M2^EcNj7CbGOf0m1DjNQ8^OcGOc0l1_OPOX8YOcGMe0U3e7\\MYHi2`7]M]He2a7^M[Hc2d7`MYHa2g7`MXH`2h7bMUH`2j7bMTH^2k7dMSH]2m7dMRH^2l7bMSHd2h7\\MWHi2e7WM[Hi2d7_10O1000O0100O10O10O100O10001N2O0O2O0jKTHn2m7PMUHn2l7QMUHo2l7oLVHo2k7QMUHo2l7oLUHQ3l7fL]HX3e7eL]H[3k7ZLXHe3h8O000O2O0O100O2O000O2O0O10000O100O101O3L4M2^NRFCR::QFCQ:;SFAP:=RF_OS:;SFAP:;TF^OS:=SF]OR:?RF]OR:>TF\\OQ:`0TF[OQ:`0`1K3N3L4MWgn5"}, "label": "an older woman in a black suit looks upon a spread - out buffet of pizza and other foods"}]} {"id": 444033, "image": "COCO_train2014_000000444033.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in black shirt with gray hair standing next to a man\"."}], "task": "refering", "answer_template": "The \"a woman in black shirt with gray hair standing next to a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 118, 119, 120, 141, 142, 143, 164, 165, 166, 167, 187, 188, 189, 190, 210, 211, 212, 213, 233, 234, 235, 236, 257], "bbox": [0.1436875, 0.31354460093896713, 0.28346875, 0.7456338028169014], "iscrowd": 0, "area": 10548.051649999996, "rle": {"size": [426, 640], "counts": "g_V14R=8E;F9G:G7M3M2O2M3M2^EcNj7CbGOf0m1DjNQ8^OcGOc0l1_OPOX8YOcGMe0U3e7\\MYHi2`7]M]He2a7^M[Hc2d7`MYHa2g7`MXH`2h7bMUH`2j7bMTH^2k7dMSH]2m7dMRH^2l7bMSHd2h7\\MWHi2e7WM[Hi2d7_10O1000O0100O10O10O100O10001N2O0O2O0jKTHn2m7PMUHn2l7QMUHo2l7oLVHo2k7QMUHo2l7oLUHQ3l7fL]HX3e7eL]H[3k7ZLXHe3h8O000O2O0O100O2O000O2O0O10000O100O101O3L4M2^NRFCR::QFCQ:;SFAP:=RF_OS:;SFAP:;TF^OS:=SF]OR:?RF]OR:>TF\\OQ:`0TF[OQ:`0`1K3N3L4MWgn5"}, "label": "a woman in black shirt with gray hair standing next to a man"}]} {"id": 444033, "image": "COCO_train2014_000000444033.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue shirt holding a plate in a food line at a potluck\"."}], "task": "refering", "answer_template": "The \"a man in a blue shirt holding a plate in a food line at a potluck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 115, 116, 117, 138, 139, 140, 161, 162, 163, 184, 185, 186, 208, 209, 231, 232], "bbox": [0.00010937500000000002, 0.26539906103286387, 0.123046875, 0.711338028169014], "iscrowd": 0, "area": 9306.096099999995, "rle": {"size": [426, 640], "counts": "c41g0h0V:XOjE:Dk1LkMW9`0eF1Oi16SNh8V3XGkLg8U3XGlLg8U3YGkLg8T3ZGlLf8T3ZGkLf8V3ZGjLf8V3ZGjLf8V3ZGjLe8V3\\GiLe8W3[GiLd8X3\\GhLd8X3\\GhLd8Y3[GgLd8Z3\\GeLe8\\3ZGdLf8]3YGcLh8\\3XGdLh8\\3XGcLj8\\3VGdLj8\\3VGdLj8\\3VGdLk8[3UGeLk8[3UGeLd0IY7a3SHfLb09m6c3QI`Ln6`3QIaLP7^3PIcLo6]3QIdLn6\\3QIfLo6Z3PIgLo6Y3PIhLS7U3mHkLV7R3jHmLZ7P3eHQM_7k2aHUMb7]OQHk2=hMS8W2lGjMT8V2lGjMU8U2kGjMV8V2iGkMX8T2hGlMX8T2hGlMY8S2fGnMZ8S2eGmM\\8R2dGnM\\8R2cGoM^8P2bGPN^8Q2aGoM`8P2_GQNa8o1_GQNb8n1^GQNc8o1\\GRNe8n1ZGRNf8n1ZGRNg8m1XGTNh8l1TGXNm8W1YGSOh8k0^GPOc8n0bGnN_8o0gGmN[8P1jGlNW8R1oGiNR8T1THgNn7W1WHeNj7Y1\\HbNe7[1aH_Nb7_1W2L3N2M3N6J6H8G9G9FTVY7"}, "label": "a man in a blue shirt holding a plate in a food line at a potluck"}]} {"id": 444033, "image": "COCO_train2014_000000444033.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in blue shirt\"."}], "task": "refering", "answer_template": "The \"man in blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 115, 116, 117, 138, 139, 140, 161, 162, 163, 184, 185, 186, 208, 209, 231, 232], "bbox": [0.00010937500000000002, 0.26539906103286387, 0.123046875, 0.711338028169014], "iscrowd": 0, "area": 9306.096099999995, "rle": {"size": [426, 640], "counts": "c41g0h0V:XOjE:Dk1LkMW9`0eF1Oi16SNh8V3XGkLg8U3XGlLg8U3YGkLg8T3ZGlLf8T3ZGkLf8V3ZGjLf8V3ZGjLf8V3ZGjLe8V3\\GiLe8W3[GiLd8X3\\GhLd8X3\\GhLd8Y3[GgLd8Z3\\GeLe8\\3ZGdLf8]3YGcLh8\\3XGdLh8\\3XGcLj8\\3VGdLj8\\3VGdLj8\\3VGdLk8[3UGeLk8[3UGeLd0IY7a3SHfLb09m6c3QI`Ln6`3QIaLP7^3PIcLo6]3QIdLn6\\3QIfLo6Z3PIgLo6Y3PIhLS7U3mHkLV7R3jHmLZ7P3eHQM_7k2aHUMb7]OQHk2=hMS8W2lGjMT8V2lGjMU8U2kGjMV8V2iGkMX8T2hGlMX8T2hGlMY8S2fGnMZ8S2eGmM\\8R2dGnM\\8R2cGoM^8P2bGPN^8Q2aGoM`8P2_GQNa8o1_GQNb8n1^GQNc8o1\\GRNe8n1ZGRNf8n1ZGRNg8m1XGTNh8l1TGXNm8W1YGSOh8k0^GPOc8n0bGnN_8o0gGmN[8P1jGlNW8R1oGiNR8T1THgNn7W1WHeNj7Y1\\HbNe7[1aH_Nb7_1W2L3N2M3N6J6H8G9G9FTVY7"}, "label": "man in blue shirt"}]} {"id": 444033, "image": "COCO_train2014_000000444033.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in silver holding plate of veggies\"."}], "task": "refering", "answer_template": "The \"man in silver holding plate of veggies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 229, 250, 251, 252, 273, 274, 275, 296, 297, 298, 320, 321, 343, 344], "bbox": [0.877, 0.5495539906103287, 0.9999375, 0.9752347417840377], "iscrowd": 0, "area": 7877.407200000005, "rle": {"size": [426, 640], "counts": "leY76R=3N3M2N2M4L3F;L3M3L5L:H;F1O1O3M3M2N2N1O2M3N1O2N2N101O0O100O1O010O1O10O01O1O100M1O2N2N2M2O2N2N1N3N2M2N2N2O0O2N2N1O2N2N1O2N2N2N1O2N2M2N3M3M3M3M2O2M3M3M3M31Oh2XMkE"}, "label": "man in silver holding plate of veggies"}]} {"id": 444033, "image": "COCO_train2014_000000444033.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the gray and blue shirt\"."}], "task": "refering", "answer_template": "The \"the gray and blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 229, 250, 251, 252, 273, 274, 275, 296, 297, 298, 320, 321, 343, 344], "bbox": [0.877, 0.5495539906103287, 0.9999375, 0.9752347417840377], "iscrowd": 0, "area": 7877.407200000005, "rle": {"size": [426, 640], "counts": "leY76R=3N3M2N2M4L3F;L3M3L5L:H;F1O1O3M3M2N2N1O2M3N1O2N2N101O0O100O1O010O1O10O01O1O100M1O2N2N2M2O2N2N1N3N2M2N2N2O0O2N2N1O2N2N1O2N2N2N1O2N2M2N3M3M3M3M2O2M3M3M3M31Oh2XMkE"}, "label": "the gray and blue shirt"}]} {"id": 444033, "image": "COCO_train2014_000000444033.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with hands on the pizza box\"."}], "task": "refering", "answer_template": "The \"woman with hands on the pizza box\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 128, 129, 150, 151, 152, 153, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 223, 224, 241, 242, 243, 244, 247, 248, 265, 266], "bbox": [0.51903125, 0.30561032863849763, 0.8002499999999999, 0.7618075117370892], "iscrowd": 0, "area": 16302.666050000002, "rle": {"size": [426, 640], "counts": "_\\Z4\\1i;4L4L5K4L4YNfN[G^1]8[OnFi0S9f10010O2O0O100O2O0O101N1O101N100O2O0O101N100[OoKPHS4m7f0N2N3M2N2N3M2N2O2M101O001O010O001O010O11N100O2O0O2O0O101N10001N100O2O0O101N101O1N4M2M4M3L7hK`GZ3h8[L[Ge3m8QLVGn3S90001M2N2cN^1F9D=0O10O100O101O0O100O10001N100O10000O2O0O100O2O000O2O0O101O0O1010O00010O000010O01O01O01O01O01O010O00010O01O01O01O01O010O00010O0010O00001N100O2O0O2O1N2N2O1N101N2O1N2O1N2O1N1O2O1N2N2O1N2N2N10Tnd1"}, "label": "woman with hands on the pizza box"}]} {"id": 444033, "image": "COCO_train2014_000000444033.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman standing behind multiple boxes of pizza reaching for 3 / 8 of a pepperoni pizza\"."}], "task": "refering", "answer_template": "The \"woman standing behind multiple boxes of pizza reaching for 3 / 8 of a pepperoni pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 128, 129, 150, 151, 152, 153, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 223, 224, 241, 242, 243, 244, 247, 248, 265, 266], "bbox": [0.51903125, 0.30561032863849763, 0.8002499999999999, 0.7618075117370892], "iscrowd": 0, "area": 16302.666050000002, "rle": {"size": [426, 640], "counts": "_\\Z4\\1i;4L4L5K4L4YNfN[G^1]8[OnFi0S9f10010O2O0O100O2O0O101N1O101N100O2O0O101N100[OoKPHS4m7f0N2N3M2N2N3M2N2O2M101O001O010O001O010O11N100O2O0O2O0O101N10001N100O2O0O101N101O1N4M2M4M3L7hK`GZ3h8[L[Ge3m8QLVGn3S90001M2N2cN^1F9D=0O10O100O101O0O100O10001N100O10000O2O0O100O2O000O2O0O101O0O1010O00010O000010O01O01O01O01O01O010O00010O01O01O01O01O010O00010O0010O00001N100O2O0O2O1N2N2O1N101N2O1N2O1N2O1N1O2O1N2N2O1N2N2N10Tnd1"}, "label": "woman standing behind multiple boxes of pizza reaching for 3 / 8 of a pepperoni pizza"}]} {"id": 444036, "image": "COCO_train2014_000000444036.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown dish of food\"."}], "task": "refering", "answer_template": "The \"a brown dish of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 31, 32, 33, 34, 35, 54, 55, 56, 57, 58, 59, 77, 78, 79, 80, 81, 82, 83, 99, 100, 101, 102, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 284, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 332, 333, 334, 335], "bbox": [0.319484375, 0.044145833333333336, 0.76753125, 0.8857083333333333], "iscrowd": 0, "area": 74447.86489999999, "rle": {"size": [480, 640], "counts": "Uio22k>a0_Oa0_Oa0_Oa0_Oa0_Oa0^O;F5K6J6J6J6J6I7J6J6J6J6J6J6J5J7J6J6J6J6J6K5J6J6K5J6K4K5K5L4K5L4K5L4K5M3N2M3N2N3L3N2M3N2M3N2M3N2M3N2N2M3N2M2O1N101N2O1O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1O1N101N2O1N101N2O001O001O00100O001O1O001O0000000000001O0000000001O1O001O1O0O2O001O1O0O2O1O001N2O001O1O0O2O1O1N101O1N2O1O0O2O1O1N2O1O0O2O1O1N2O1O1N3N1O1N2O1O1N2O1O1N2O1N2O1N2O1N3N1N3M2O2M2N3M2N3M2N3M2N3M2N3M2N3M2N3L3N3L3M4M2M4M2M4M2M4M2M4M2M3N3L3M4K4K5L5J5L4K6J5L4K5L5J5K5J6K6I6K5J6J6K6I6K5K4K6K4K5L5K4L5K4L5K4L5K4L4L4L5M2O1N2O2N1N2O1N2O1O1O10000O10000O1000000O1000O00100O00100O1OUOcB4^=JeB5Z=JjB3W=KlB4Y=EkB8Z=ChB

M2O2M2N3N2MSeU2"}, "label": "a brown dish of food"}]} {"id": 444036, "image": "COCO_train2014_000000444036.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown plate on white place mat with two pieces of toast framing a cheese dish\"."}], "task": "refering", "answer_template": "The \"a brown plate on white place mat with two pieces of toast framing a cheese dish\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 31, 32, 33, 34, 35, 54, 55, 56, 57, 58, 59, 77, 78, 79, 80, 81, 82, 83, 99, 100, 101, 102, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 284, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 332, 333, 334, 335], "bbox": [0.319484375, 0.044145833333333336, 0.76753125, 0.8857083333333333], "iscrowd": 0, "area": 74447.86489999999, "rle": {"size": [480, 640], "counts": "Uio22k>a0_Oa0_Oa0_Oa0_Oa0_Oa0^O;F5K6J6J6J6J6I7J6J6J6J6J6J6J5J7J6J6J6J6J6K5J6J6K5J6K4K5K5L4K5L4K5L4K5M3N2M3N2N3L3N2M3N2M3N2M3N2M3N2N2M3N2M2O1N101N2O1O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1O1N101N2O1N101N2O001O001O00100O001O1O001O0000000000001O0000000001O1O001O1O0O2O001O1O0O2O1O001N2O001O1O0O2O1O1N101O1N2O1O0O2O1O1N2O1O0O2O1O1N2O1O1N3N1O1N2O1O1N2O1O1N2O1N2O1N2O1N3N1N3M2O2M2N3M2N3M2N3M2N3M2N3M2N3M2N3L3N3L3M4M2M4M2M4M2M4M2M4M2M3N3L3M4K4K5L5J5L4K6J5L4K5L5J5K5J6K6I6K5J6J6K6I6K5K4K6K4K5L5K4L5K4L5K4L5K4L4L4L5M2O1N2O2N1N2O1N2O1O1O10000O10000O1000000O1000O00100O00100O1OUOcB4^=JeB5Z=JjB3W=KlB4Y=EkB8Z=ChB

M2O2M2N3N2MSeU2"}, "label": "a brown plate on white place mat with two pieces of toast framing a cheese dish"}]} {"id": 444036, "image": "COCO_train2014_000000444036.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver bread bowl with a white napkin and roll inside\"."}], "task": "refering", "answer_template": "The \"a silver bread bowl with a white napkin and roll inside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 158, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 270, 271, 272, 273, 274, 293, 294, 295, 296], "bbox": [0.7213437500000001, 0.34606250000000005, 0.9488749999999999, 0.7573124999999999], "iscrowd": 0, "area": 21193.666699999998, "rle": {"size": [480, 640], "counts": "^ih6g1X= in this image.", "objects": [{"patches": [156, 157, 158, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 270, 271, 272, 273, 274, 293, 294, 295, 296], "bbox": [0.7213437500000001, 0.34606250000000005, 0.9488749999999999, 0.7573124999999999], "iscrowd": 0, "area": 21193.666699999998, "rle": {"size": [480, 640], "counts": "^ih6g1X= in this image.", "objects": [{"patches": [76, 77, 78, 81, 82, 99, 100, 101, 103, 104, 105, 122, 123, 124, 126, 127, 128, 145, 150, 151, 168, 173, 174, 196, 197, 219, 242], "bbox": [0.31262500000000004, 0.23819248826291078, 0.604296875, 0.7146009389671361], "iscrowd": 0, "area": 10954.59045, "rle": {"size": [426, 640], "counts": "fZc2;jA5N3O00001O001O0000G9G9H8N2O1O1N2O1N2O1O1N2O2N1N2O1O1N2O1O1N2O1N2O1O1N2O1O100O100O100O100O100O100O100O100O100O102Mb0_OYZb0Gme]O3N3L3O2N1O2N1O2N1M4L3L5EQOUDS1k;70O01O001O010O001O0010O01O001O011N2N3M3N1BWNWEP2b:RN\\EU2]:mMaE[2V:hMhE`2P:bMoEe3i8]LUGn3_8f0I7I7H8I7I7I5J11O0000000G9]Oc0A?C=B?B=B>K5L4L4M3O100O1014K5K5YFTMf8n3J5K5K6J5;Ek0hNW1cN]1cNjQY3"}, "label": "the chair the blonde woman in black is on"}]} {"id": 452229, "image": "COCO_train2014_000000452229.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white fence used as room divider\"."}], "task": "refering", "answer_template": "The \"white fence used as room divider\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 81, 82, 99, 100, 101, 103, 104, 105, 122, 123, 124, 126, 127, 128, 145, 150, 151, 168, 173, 174, 196, 197, 219, 242], "bbox": [0.31262500000000004, 0.23819248826291078, 0.604296875, 0.7146009389671361], "iscrowd": 0, "area": 10954.59045, "rle": {"size": [426, 640], "counts": "fZc2;jA5N3O00001O001O0000G9G9H8N2O1O1N2O1N2O1O1N2O2N1N2O1O1N2O1O1N2O1N2O1O1N2O1O100O100O100O100O100O100O100O100O100O102Mb0_OYZb0Gme]O3N3L3O2N1O2N1O2N1M4L3L5EQOUDS1k;70O01O001O010O001O0010O01O001O011N2N3M3N1BWNWEP2b:RN\\EU2]:mMaE[2V:hMhE`2P:bMoEe3i8]LUGn3_8f0I7I7H8I7I7I5J11O0000000G9]Oc0A?C=B?B=B>K5L4L4M3O100O1014K5K5YFTMf8n3J5K5K6J5;Ek0hNW1cN]1cNjQY3"}, "label": "white fence used as room divider"}]} {"id": 198277, "image": "COCO_train2014_000000198277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow taxi , which is to the left of another taxi\"."}], "task": "refering", "answer_template": "The \"a yellow taxi , which is to the left of another taxi\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 286, 287, 288, 289, 290, 291, 308, 309, 310, 311, 312, 313, 331, 335], "bbox": [0.02246732026143791, 0.48277777777777775, 0.2741503267973856, 0.6962581699346405], "iscrowd": 0, "area": 15883.808150000006, "rle": {"size": [612, 612], "counts": "\\f8c0\\b0i0YOg0ZO6J4L4K4M4L4L4L4K6K5K6J2O1000000000O0100000000000000000000O1000000000M3L4L4L4N2000O100000001O0000000000000O10000000000000000000000O100000001O0000001O000O2O001O00001O001O00001O0O100000000000O100O1O1O1O010O1O1O1O1O100O1O11O1O0O2O1O001O1O1O1O1O1O1O1N2O100O1O1O01O0fNj@gNU?R1UAkNk>o0]APOc>h0fAVOZ>e0nAXOR>d0TB[Ol=a0YB]Og=`0^B_Oa=>eB_O[=>jB@W=R3M3M4L3L4MURY8"}, "label": "a yellow taxi , which is to the left of another taxi"}]} {"id": 198277, "image": "COCO_train2014_000000198277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow cab closest to the buildings on the left\"."}], "task": "refering", "answer_template": "The \"the yellow cab closest to the buildings on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 286, 287, 288, 289, 290, 291, 308, 309, 310, 311, 312, 313, 331, 335], "bbox": [0.02246732026143791, 0.48277777777777775, 0.2741503267973856, 0.6962581699346405], "iscrowd": 0, "area": 15883.808150000006, "rle": {"size": [612, 612], "counts": "\\f8c0\\b0i0YOg0ZO6J4L4K4M4L4L4L4K6K5K6J2O1000000000O0100000000000000000000O1000000000M3L4L4L4N2000O100000001O0000000000000O10000000000000000000000O100000001O0000001O000O2O001O00001O001O00001O0O100000000000O100O1O1O1O010O1O1O1O1O100O1O11O1O0O2O1O001O1O1O1O1O1O1O1N2O100O1O1O01O0fNj@gNU?R1UAkNk>o0]APOc>h0fAVOZ>e0nAXOR>d0TB[Ol=a0YB]Og=`0^B_Oa=>eB_O[=>jB@W=R3M3M4L3L4MURY8"}, "label": "the yellow cab closest to the buildings on the left"}]} {"id": 198277, "image": "COCO_train2014_000000198277.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow taxi cab follows behind a bus\"."}], "task": "refering", "answer_template": "The \"a yellow taxi cab follows behind a bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 229, 230, 247, 248, 249, 250, 251, 252, 253, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 313, 314, 315, 316, 317, 318, 319, 336, 337, 339, 340, 341], "bbox": [0.24383986928104573, 0.47683006535947714, 0.5301960784313726, 0.7050163398692811], "iscrowd": 0, "area": 19813.051149999996, "rle": {"size": [612, 612], "counts": "`]i23jb09G8J7K5J6J6K5J6K5J6K5J5K6J6K5J5K5M3N2N2N2N2O1N2N2N2N2O1N2N2N2N2O2N1O00O010O100O1O100O010O100O100O100O00100O100O100O010O100000000O1000000O1000000O1000000O10001O0O100000000O10000001O000000001O000000001O00001O001O00001O00001O001O00001O001O001O001O00001O001O001O001O0000000000000000000002N2N2N2N2N1O2N2N2M3N2N2N2N2N2N2N2N1O1O1N2N2N2N4L4L5K5K4L5J5E in this image.", "objects": [{"patches": [226, 227, 228, 229, 230, 247, 248, 249, 250, 251, 252, 253, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 313, 314, 315, 316, 317, 318, 319, 336, 337, 339, 340, 341], "bbox": [0.24383986928104573, 0.47683006535947714, 0.5301960784313726, 0.7050163398692811], "iscrowd": 0, "area": 19813.051149999996, "rle": {"size": [612, 612], "counts": "`]i23jb09G8J7K5J6J6K5J6K5J6K5J5K6J6K5J5K5M3N2N2N2N2O1N2N2N2N2O1N2N2N2N2O2N1O00O010O100O1O100O010O100O100O100O00100O100O100O010O100000000O1000000O1000000O1000000O10001O0O100000000O10000001O000000001O000000001O00001O001O00001O00001O001O00001O001O001O001O00001O001O001O001O0000000000000000000002N2N2N2N2N1O2N2N2M3N2N2N2N2N2N2N2N1O1O1N2N2N2N4L4L5K5K4L5J5E in this image.", "objects": [{"patches": [19, 20, 21, 22, 23, 33, 34, 35, 36, 37, 38, 39, 48, 49, 50, 51, 52, 53, 54, 63, 64, 65, 66, 67, 68, 69, 78, 79, 80, 81, 82, 83, 84, 94, 95, 96, 97, 98, 109, 110, 111, 112, 113, 126, 127], "bbox": [0.2121077283372365, 0.041281250000000005, 0.6362997658079625, 0.359671875], "iscrowd": 0, "area": 27512.61155, "rle": {"size": [640, 427], "counts": "hnh19ec0;E;E:F8H4L4L4L4L4L4L4L5K4L4L4M3M3M3M4L3M3L4M3M3M3M3M2M4M2N3M2N3M2N3N1N3N1O2N1O2M2O2N1O2N1N3N1O2N1O0O2O001O0O101O001O0O2O00001O0O2O00001N101O00001N101O001O0O101O001N10010O001O01O01O0010O01O000010O01OO101O0000010O000000001N10000O2O000O100O2O000O101N2O1O0O2O1N101O1N101N2O1O0O2O1N101O1N2O0O2M3M2M4M3M3L3N3M3M2N3M3M3N1N3M3M2O2M3M3M5L3L5K4L4M4K4L5L3L5K4L4I8_O`0D=C in this image.", "objects": [{"patches": [19, 20, 21, 22, 23, 33, 34, 35, 36, 37, 38, 39, 48, 49, 50, 51, 52, 53, 54, 63, 64, 65, 66, 67, 68, 69, 78, 79, 80, 81, 82, 83, 84, 94, 95, 96, 97, 98, 109, 110, 111, 112, 113, 126, 127], "bbox": [0.2121077283372365, 0.041281250000000005, 0.6362997658079625, 0.359671875], "iscrowd": 0, "area": 27512.61155, "rle": {"size": [640, 427], "counts": "hnh19ec0;E;E:F8H4L4L4L4L4L4L4L5K4L4L4M3M3M3M4L3M3L4M3M3M3M3M2M4M2N3M2N3M2N3N1N3N1O2N1O2M2O2N1O2N1N3N1O2N1O0O2O001O0O101O001O0O2O00001O0O2O00001N101O00001N101O001O0O101O001N10010O001O01O01O0010O01O000010O01OO101O0000010O000000001N10000O2O000O100O2O000O101N2O1O0O2O1N101O1N101N2O1O0O2O1N101O1N2O0O2M3M2M4M3M3L3N3M3M2N3M3M3N1N3M3M2O2M3M3M5L3L5K4L4M4K4L5L3L5K4L4I8_O`0D=C in this image.", "objects": [{"patches": [71, 72, 84, 85, 86, 87, 88, 89, 98, 99, 100, 101, 102, 103, 104, 113, 114, 115, 116, 117, 118, 119, 128, 129, 130, 131, 132, 133, 134, 143, 144, 145, 146, 147, 148, 149, 159, 160, 161, 162, 163, 164, 174, 175, 176, 177, 178], "bbox": [0.5642857142857143, 0.210625, 0.9956908665105386, 0.5204531250000001], "iscrowd": 0, "area": 29414.921099999996, "rle": {"size": [640, 427], "counts": "ajf4=_c0k0UOd0]O5J6L3M4L4L4L4L3M4L4L4L4L4L3M4M3L3N2N2M3N2M3N2M3N2N2M3N2N2O1N2O1N2O1N2O1N101N2O1N100O2O0O2O001N10001O0O2O00001O0O2O001O000O2O001O0O101O00000O101O00000000001O00000000001O000000000010O000000000001O00000000001O0001O000001O0001O000001O0001O00000001O01O0O2O0O2N1O101N1O2O0O2N1O101N1O2O0O2N100O2N1O2O0O2N2O1N1O2N2O1N2N101M3M3N3L3M4M2M4M2M4L4M2M4L3N3L3M4M2M4M2M4L3N3VOi0nNQS1"}, "label": "the first lemon from the right"}]} {"id": 132746, "image": "COCO_train2014_000000132746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an aeroplane is standing behind another one\"."}], "task": "refering", "answer_template": "The \"an aeroplane is standing behind another one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 249, 250, 251, 252, 268, 269, 273], "bbox": [0.533015625, 0.5156828193832599, 1.0, 0.7301321585903084], "iscrowd": 0, "area": 11329.574450000002, "rle": {"size": [454, 640], "counts": "T_g42o=7H8I7L3O2N2N2N00001O00001O00001O0000001O00001OO1000000000000O1000000000000O1000000000000O1000000000000O01000000O1000O10O1000000000O01000001O0000000000001N1000000000001O000000000O101O001O6J5K3M1O1N2O1O00@jCQOU in this image.", "objects": [{"patches": [97, 98, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 129, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 247, 248, 249, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 270, 271, 272, 273, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 292, 293, 294, 295, 296, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 315, 316, 317, 318, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 338, 339, 340, 341, 342, 349, 350, 351, 361, 362, 363, 364, 365], "bbox": [0.0025781249999999997, 0.2888766519823789, 1.0, 1.0], "iscrowd": 0, "area": 130969.59300000001, "rle": {"size": [454, 640], "counts": "SQ1d5c8O00000000000000001N100000PJbJX3^5bLgJ_3Y5[LmJe3S5VLRKj3n4PLXKP4h4jK^KW4a4cKeK]4[4^KiKc4W4WKoKi4Q4QKULP5j3kJ[LU5i3aJ]L_5U61O0O101O00001O000O10000000O1000O1000002N2N1O0O100000O100000O1000000000O10O1000000000O10O1000000000O10O1000000000O0100000000O1000000011N3M2M4M2N3M3M2N0000000O1000000000001O000O1000000000000000000000000000000000001O0000000O100O100O1O100O1O100O1O1O2N1O100O1O1O1O2N1O100O1O1O1O10000O10000O10000O02O3M2N6J1O2N2N1O001O0O2O001O00O100002N001N2N2NO100001O1O1O010000000000000000000000O100O100O1O10O0100O100O1N2N2O1N2O1N20000000000000O2N1N2O1N2O1N2O1O2M2O1N2000000O1000001N10000O1000000O10001O0O10000O1000000O101O000O1000000O10000O2O00000O1000000O10001O0O10000O1000000O101O000O10000O1000000O2O00000O1000000O10001N1000000O1000000O01MeNaIUK_6f4nIRKR6i4[JnJf5m4gJkJY5P5TKhJl4S5V2K5J6K6J7I7I7L6L8G8I8H3M1O001O001O001O1O001O000000000O01000000O10000O1000000O1000000DTM]Em2c:;0O10001O000000000000000000000000001O0O1000000000001O00000000000000001O000O10000000001O00000000000000001O00000O100000001O00000000000000^JVMIi2KcM5]2@nM`0R2TOZNl0f1jNdNV1\\1eNiN[1V1aNoN_1Q1[NUOe1k0VNZOj1f0RN]OP2b0lMBT2>iMEW2;eMI[27bML^24^M0b2O\\M4d2LYM7g2IUM;l2DQM?o2AmLc0S3\\OkLg0U3YOgLk0Y3UOdLn0\\3RO`LR1`3nN]LU1c3jNZLZ1f3fNTL`1m3_NmKf1T4ZNfKl1Z4SNaKS2_4mM[KY2e4gMUK_2k4aMoJe2Q5[MiJk2W5TMdJR3]5mL]JY3c5e2000000000000001N100000000000001OO1000000O1000000O10O1000O1000000O1000000O1000000O100000O01000000O1000000OeMbJPL^5P4dJnK\\5Q4hJlKX5:bJY19ZNT5>kJP13`NR5`0RKh00eNn4c0YKa0LiNk4f0`K:GnNh4h0iK3BQOf4l0PLJ]OXOc4n0WLCYO\\O`4Q1^L\\OUO@\\4U1fLTOPOEZ4V1lLoNnNHV4Y1RMiNkNKS4\\1XMcNgNOP4_1_M[NeN3l3b1eMUNbN6i3e1jMPN_N9g3f1QNkM[NnMBR2`0lM_OT2e0iM[OW2h0fMXOZ2j0cMWO]2l0`MSOa2Y600O10O100000O01000000O0100000O1000O10O100000O01000000O10O1000O1000O10O100000O01000000O10O1000O1000O10O1000000O0100000O10O1000O1000O10OQO]L"}, "label": "airplane"}]} {"id": 132746, "image": "COCO_train2014_000000132746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white picture of a steel airplane\"."}], "task": "refering", "answer_template": "The \"a black and white picture of a steel airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 100, 101, 102, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 129, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 247, 248, 249, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 270, 271, 272, 273, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 292, 293, 294, 295, 296, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 315, 316, 317, 318, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 338, 339, 340, 341, 342, 349, 350, 351, 361, 362, 363, 364, 365], "bbox": [0.0025781249999999997, 0.2888766519823789, 1.0, 1.0], "iscrowd": 0, "area": 130969.59300000001, "rle": {"size": [454, 640], "counts": "SQ1d5c8O00000000000000001N100000PJbJX3^5bLgJ_3Y5[LmJe3S5VLRKj3n4PLXKP4h4jK^KW4a4cKeK]4[4^KiKc4W4WKoKi4Q4QKULP5j3kJ[LU5i3aJ]L_5U61O0O101O00001O000O10000000O1000O1000002N2N1O0O100000O100000O1000000000O10O1000000000O10O1000000000O10O1000000000O0100000000O1000000011N3M2M4M2N3M3M2N0000000O1000000000001O000O1000000000000000000000000000000000001O0000000O100O100O1O100O1O100O1O1O2N1O100O1O1O1O2N1O100O1O1O1O10000O10000O10000O02O3M2N6J1O2N2N1O001O0O2O001O00O100002N001N2N2NO100001O1O1O010000000000000000000000O100O100O1O10O0100O100O1N2N2O1N2O1N20000000000000O2N1N2O1N2O1N2O1O2M2O1N2000000O1000001N10000O1000000O10001O0O10000O1000000O101O000O1000000O10000O2O00000O1000000O10001O0O10000O1000000O101O000O10000O1000000O2O00000O1000000O10001N1000000O1000000O01MeNaIUK_6f4nIRKR6i4[JnJf5m4gJkJY5P5TKhJl4S5V2K5J6K6J7I7I7L6L8G8I8H3M1O001O001O001O1O001O000000000O01000000O10000O1000000O1000000DTM]Em2c:;0O10001O000000000000000000000000001O0O1000000000001O00000000000000001O000O10000000001O00000000000000001O00000O100000001O00000000000000^JVMIi2KcM5]2@nM`0R2TOZNl0f1jNdNV1\\1eNiN[1V1aNoN_1Q1[NUOe1k0VNZOj1f0RN]OP2b0lMBT2>iMEW2;eMI[27bML^24^M0b2O\\M4d2LYM7g2IUM;l2DQM?o2AmLc0S3\\OkLg0U3YOgLk0Y3UOdLn0\\3RO`LR1`3nN]LU1c3jNZLZ1f3fNTL`1m3_NmKf1T4ZNfKl1Z4SNaKS2_4mM[KY2e4gMUK_2k4aMoJe2Q5[MiJk2W5TMdJR3]5mL]JY3c5e2000000000000001N100000000000001OO1000000O1000000O10O1000O1000000O1000000O1000000O100000O01000000O1000000OeMbJPL^5P4dJnK\\5Q4hJlKX5:bJY19ZNT5>kJP13`NR5`0RKh00eNn4c0YKa0LiNk4f0`K:GnNh4h0iK3BQOf4l0PLJ]OXOc4n0WLCYO\\O`4Q1^L\\OUO@\\4U1fLTOPOEZ4V1lLoNnNHV4Y1RMiNkNKS4\\1XMcNgNOP4_1_M[NeN3l3b1eMUNbN6i3e1jMPN_N9g3f1QNkM[NnMBR2`0lM_OT2e0iM[OW2h0fMXOZ2j0cMWO]2l0`MSOa2Y600O10O100000O01000000O0100000O1000O10O100000O01000000O10O1000O1000O10O100000O01000000O10O1000O1000O10O1000000O0100000O10O1000O1000O10OQO]L"}, "label": "a black and white picture of a steel airplane"}]} {"id": 534155, "image": "COCO_train2014_000000534155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bottom of the silver and yellow plane on the right\"."}], "task": "refering", "answer_template": "The \"the bottom of the silver and yellow plane on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 42, 43, 44, 45, 66, 67, 68, 89, 90, 91, 113, 114, 137], "bbox": [0.8161406250000001, 0.005166666666666667, 1.0, 0.41597222222222224], "iscrowd": 0, "area": 10429.38, "rle": {"size": [360, 640], "counts": "jkg52V;1O2N2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2O0O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O1O001O00001O001O001O00001O001O001O00001O001O001L"}, "label": "the bottom of the silver and yellow plane on the right"}]} {"id": 534155, "image": "COCO_train2014_000000534155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"body of the huge silver plane , behind the red and blue\"."}], "task": "refering", "answer_template": "The \"body of the huge silver plane , behind the red and blue\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 42, 43, 44, 45, 66, 67, 68, 89, 90, 91, 113, 114, 137], "bbox": [0.8161406250000001, 0.005166666666666667, 1.0, 0.41597222222222224], "iscrowd": 0, "area": 10429.38, "rle": {"size": [360, 640], "counts": "jkg52V;1O2N2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2O0O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O1O001O00001O001O001O00001O001O001O00001O001O001L"}, "label": "body of the huge silver plane , behind the red and blue"}]} {"id": 534155, "image": "COCO_train2014_000000534155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue and red airplane\"."}], "task": "refering", "answer_template": "The \"a blue and red airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 26, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 86, 87, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 161, 162, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 184, 185, 188, 189, 190, 191, 195, 196, 198, 199, 200, 201, 202, 211, 212, 213, 214, 236, 237], "bbox": [0.015140625, 0.12133333333333333, 0.90428125, 0.8119166666666667], "iscrowd": 0, "area": 61814.67029999998, "rle": {"size": [360, 640], "counts": "Qh33S;2N2mKOlL4T3MiL6U3KhL8X3IaL=_3C[Lc0d3_OULg0k3YOoKm0P4UOiKR1V4nNdKX1[4jN_K[1a4eNYKa1f4aNSKe1m4[NmJk1R5WNgJn1Z5RN`JT2`5mMYJY2f5hMVJ[2k5fMRJ\\2n5dMoI^2R6cMkI]2W6cMfI^2\\6cMaI]2`6dM]I]2e6dMXI\\2T5YMlK;mN]2\\5ZMbK:oN\\2c5]MYK]3k4fLPKW3U5lLfJQ3`5QMZJm2k5UMQJi2S6ZMhIc2^6o05K5K6K4K5K6J5L5J6J6J3N1N1O2N2O1N2O03N2N1O2N2M2O2N1O2N1N3N2N1O1O2M2O2N1O2N1N2O2N1O2N1N3N2N1ObLPIg2o6WMSIi2m6VMTIk2k6SMWIm2i6QMYIP3f6nL\\IS3c6kL_IU3a6iLaIX3^6fLdIZ3\\6dLgI\\3X6cLiI]3W6aLkI`3n61O00000001O000001O0000000001O0000QJjLl3W3RLlLl3T3RLPMl3P3SLSMk3m2TLTMl3l2RLVMo3i2PLXMP4h2oKYMQ4g2nKZMR4f2nKZMR4g2lKZMT4f2lK[MS4e2mK[MS4e2lK\\MT4d2lK\\MU4c2jK^MV4b2lJ`Lb0n0b4b2lJdL=k0g4a2kJhL;g0j4b2jJjL8e0n4a2iJnL6a0Q5a2iJRM1=W5a2hJUMN9\\5a2eJZMK5a5a2dJ[N]5e1bJ[N_5e1aJZN`5f1_JZNc5f1\\JYNe5g1ZJXNh5h1XJUNk5k1TJTNo5k1QJRNR6`32M4M2M3M3N3L3N2M40O001O010O0001O0000001O0000010O00001O0000001O00010ObJhKl3X4oKoKo3Q4mKULQ4k3mKYLR4f3lK`LP4`3nKcLQ4]3nKdLR4\\3mKfLS4Y3lKhLT4X3lKiLS4X3kKiLV4V3iKkLW4U3hKmLW4S3gKoLY4Q3`KVMa4i2WK_Mi4a4O01O01O0001O00001O000O101O000O1gNPKZLQ5[3[KcLe4R3hKjLZ4o2PLlLS4S3oKiLS4W3PLdLR4\\3PL`LR4`3QL\\LQ4c3QLYLQ4g3h101O00000O2O00001O0000001O0000001O00001O0000001O00001O00001O00001O00001O0000001O0O100O010\\OTHkMm7T2VHjMj7V2XHhMg7Y2[HdMf7\\2[HcMe7]2\\HaMd7`2]H_Mc7`2`H]Ma7c2b00O1000O10O10000O10001O0O101O0O1000000O1000000O2O000O1000000000001O000001O01O000000001O000000000fHXMT6i2^IdMb6\\2ZIiMe6X2VImMj6R2SIQNm6U30000000000000N2O1O100O1O2N1DQI_LP7_3=O2N1O2N10001O010O010O010O001O01O01O001O001O001O000010O01O001O001O010O000010O01O010O0010O0001O010O001O010O00001O001O001O001O00001O001O001O001@eGPN\\8o1gGoMY8P2iGoMX8P2iGoMW8Q2kGmMV8R2kGlMV8T2kGkMV8U2>0001O010O010O010OEPN^GP2`8SN_Gn1`8RNaGm1_8TN`Gm1^8TNcGk1]8VNbGj1]8WNdGi1[8XNdGh1\\8XNeGg1Z8[NeGf1Z8ZNgGe1Y8\\NeGe1[8[NeGe1[8e00O1O010O001O10O01O010O1OO2M2O2M2N3N1N3N1N3N1N3N1N3N2M3N1N3M3N1N3O1O1O001O1O000010O000001O000000M3M3M3M3M3N200O10001O009G;Eb0^OO1O1O001N2O1O1O1O001O1O00VNbGk0^8UOcGk0\\8TOfGk0Z8UOgGKL6]8OhGKL4\\81jGJK3[83kGKJ1Z84mGKJOY86nGKJNW86QHLIMT88THKHLT89VHJGLR8:XHKFIR8;ZHLDIQ8;\\HLDHo7;_HMCFn7=`HMECj7?dHNF]Of7e0eHNo71SHOl71UHOk7OWH1h7OYH1f7O[H2d7L_H3`7MaH3_7KcH5]7JdH6]7GeH:[7DfH<[7AhH>X7AiH?X7^OjHb0W7\\OjHd0V7ZOlHg0j801O000001O0001O00100O1O1O2N1O100O2N7I6JWXe0"}, "label": "a blue and red airplane"}]} {"id": 534155, "image": "COCO_train2014_000000534155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small red white and blue airplane that says bud light\"."}], "task": "refering", "answer_template": "The \"small red white and blue airplane that says bud light\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 26, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 86, 87, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 161, 162, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 184, 185, 188, 189, 190, 191, 195, 196, 198, 199, 200, 201, 202, 211, 212, 213, 214, 236, 237], "bbox": [0.015140625, 0.12133333333333333, 0.90428125, 0.8119166666666667], "iscrowd": 0, "area": 61814.67029999998, "rle": {"size": [360, 640], "counts": "Qh33S;2N2mKOlL4T3MiL6U3KhL8X3IaL=_3C[Lc0d3_OULg0k3YOoKm0P4UOiKR1V4nNdKX1[4jN_K[1a4eNYKa1f4aNSKe1m4[NmJk1R5WNgJn1Z5RN`JT2`5mMYJY2f5hMVJ[2k5fMRJ\\2n5dMoI^2R6cMkI]2W6cMfI^2\\6cMaI]2`6dM]I]2e6dMXI\\2T5YMlK;mN]2\\5ZMbK:oN\\2c5]MYK]3k4fLPKW3U5lLfJQ3`5QMZJm2k5UMQJi2S6ZMhIc2^6o05K5K6K4K5K6J5L5J6J6J3N1N1O2N2O1N2O03N2N1O2N2M2O2N1O2N1N3N2N1O1O2M2O2N1O2N1N2O2N1O2N1N3N2N1ObLPIg2o6WMSIi2m6VMTIk2k6SMWIm2i6QMYIP3f6nL\\IS3c6kL_IU3a6iLaIX3^6fLdIZ3\\6dLgI\\3X6cLiI]3W6aLkI`3n61O00000001O000001O0000000001O0000QJjLl3W3RLlLl3T3RLPMl3P3SLSMk3m2TLTMl3l2RLVMo3i2PLXMP4h2oKYMQ4g2nKZMR4f2nKZMR4g2lKZMT4f2lK[MS4e2mK[MS4e2lK\\MT4d2lK\\MU4c2jK^MV4b2lJ`Lb0n0b4b2lJdL=k0g4a2kJhL;g0j4b2jJjL8e0n4a2iJnL6a0Q5a2iJRM1=W5a2hJUMN9\\5a2eJZMK5a5a2dJ[N]5e1bJ[N_5e1aJZN`5f1_JZNc5f1\\JYNe5g1ZJXNh5h1XJUNk5k1TJTNo5k1QJRNR6`32M4M2M3M3N3L3N2M40O001O010O0001O0000001O0000010O00001O0000001O00010ObJhKl3X4oKoKo3Q4mKULQ4k3mKYLR4f3lK`LP4`3nKcLQ4]3nKdLR4\\3mKfLS4Y3lKhLT4X3lKiLS4X3kKiLV4V3iKkLW4U3hKmLW4S3gKoLY4Q3`KVMa4i2WK_Mi4a4O01O01O0001O00001O000O101O000O1gNPKZLQ5[3[KcLe4R3hKjLZ4o2PLlLS4S3oKiLS4W3PLdLR4\\3PL`LR4`3QL\\LQ4c3QLYLQ4g3h101O00000O2O00001O0000001O0000001O00001O0000001O00001O00001O00001O00001O0000001O0O100O010\\OTHkMm7T2VHjMj7V2XHhMg7Y2[HdMf7\\2[HcMe7]2\\HaMd7`2]H_Mc7`2`H]Ma7c2b00O1000O10O10000O10001O0O101O0O1000000O1000000O2O000O1000000000001O000001O01O000000001O000000000fHXMT6i2^IdMb6\\2ZIiMe6X2VImMj6R2SIQNm6U30000000000000N2O1O100O1O2N1DQI_LP7_3=O2N1O2N10001O010O010O010O001O01O01O001O001O001O000010O01O001O001O010O000010O01O010O0010O0001O010O001O010O00001O001O001O001O00001O001O001O001@eGPN\\8o1gGoMY8P2iGoMX8P2iGoMW8Q2kGmMV8R2kGlMV8T2kGkMV8U2>0001O010O010O010OEPN^GP2`8SN_Gn1`8RNaGm1_8TN`Gm1^8TNcGk1]8VNbGj1]8WNdGi1[8XNdGh1\\8XNeGg1Z8[NeGf1Z8ZNgGe1Y8\\NeGe1[8[NeGe1[8e00O1O010O001O10O01O010O1OO2M2O2M2N3N1N3N1N3N1N3N1N3N2M3N1N3M3N1N3O1O1O001O1O000010O000001O000000M3M3M3M3M3N200O10001O009G;Eb0^OO1O1O001N2O1O1O1O001O1O00VNbGk0^8UOcGk0\\8TOfGk0Z8UOgGKL6]8OhGKL4\\81jGJK3[83kGKJ1Z84mGKJOY86nGKJNW86QHLIMT88THKHLT89VHJGLR8:XHKFIR8;ZHLDIQ8;\\HLDHo7;_HMCFn7=`HMECj7?dHNF]Of7e0eHNo71SHOl71UHOk7OWH1h7OYH1f7O[H2d7L_H3`7MaH3_7KcH5]7JdH6]7GeH:[7DfH<[7AhH>X7AiH?X7^OjHb0W7\\OjHd0V7ZOlHg0j801O000001O0001O00100O1O1O2N1O100O2N7I6JWXe0"}, "label": "small red white and blue airplane that says bud light"}]} {"id": 83605, "image": "COCO_train2014_000000083605.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearinf a light blue t - shirt and jeans with his arms extended\"."}], "task": "refering", "answer_template": "The \"a man wearinf a light blue t - shirt and jeans with his arms extended\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 18, 19, 32, 33, 34, 46, 47, 48, 49, 59, 60, 61, 62, 74, 76, 77], "bbox": [0.26844999999999997, 0.029299999999999996, 0.536925, 0.29116666666666663], "iscrowd": 0, "area": 6252.101449999997, "rle": {"size": [600, 400], "counts": "l\\o13cb04M3M3M3M3RM_O`Cd0\\<_OaCe0Z<_OdCd0X<@eCc0X<_OfCe0W<]OfCf0Y_OlAa0V>@hA`0Z>_OdAa0_>_O^Ab0c>_OZAa0i>^OUAc0l>^ORAa0Q?_Ol@b0U?^Oi@b0Z?^Oc@c0^?^O_@c0c?\\O\\@d0f?\\OW@f0j?ZOS@g0o?XOo_Oi0S`0WOj_Ok0W`0j0101N2O1N2N2N1O2N2M2O2N1O1N2N2N2N2N2N2N2N1Ok^OfNn`0X1R_OoNh`0Q1W_OWOc`0h0\\_O_O_`0`0`_OHY`08g_ONT`01k_O1U`0Nj_O5T`0Kk_O7U`0W1O0O2O0O2O001N10]NT@;k?A[@=d?Aa@>^?^Oh@`0X?]Om@b0Q?\\OTAb0l>^OVAa0i>_OXA`0g>@[ALX?4j@3n>LSA;f>D[AB]A=d>B]AD]A;d>D]A;d>C_A:c>E^A:c>E^A:c>E^A9g>B[A=a`0O1N3N1O2M9HcW\\3"}, "label": "a man wearinf a light blue t - shirt and jeans with his arms extended"}]} {"id": 83605, "image": "COCO_train2014_000000083605.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man on a skateboard\"."}], "task": "refering", "answer_template": "The \"a man on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 18, 19, 32, 33, 34, 46, 47, 48, 49, 59, 60, 61, 62, 74, 76, 77], "bbox": [0.26844999999999997, 0.029299999999999996, 0.536925, 0.29116666666666663], "iscrowd": 0, "area": 6252.101449999997, "rle": {"size": [600, 400], "counts": "l\\o13cb04M3M3M3M3RM_O`Cd0\\<_OaCe0Z<_OdCd0X<@eCc0X<_OfCe0W<]OfCf0Y_OlAa0V>@hA`0Z>_OdAa0_>_O^Ab0c>_OZAa0i>^OUAc0l>^ORAa0Q?_Ol@b0U?^Oi@b0Z?^Oc@c0^?^O_@c0c?\\O\\@d0f?\\OW@f0j?ZOS@g0o?XOo_Oi0S`0WOj_Ok0W`0j0101N2O1N2N2N1O2N2M2O2N1O1N2N2N2N2N2N2N2N1Ok^OfNn`0X1R_OoNh`0Q1W_OWOc`0h0\\_O_O_`0`0`_OHY`08g_ONT`01k_O1U`0Nj_O5T`0Kk_O7U`0W1O0O2O0O2O001N10]NT@;k?A[@=d?Aa@>^?^Oh@`0X?]Om@b0Q?\\OTAb0l>^OVAa0i>_OXA`0g>@[ALX?4j@3n>LSA;f>D[AB]A=d>B]AD]A;d>D]A;d>C_A:c>E^A:c>E^A:c>E^A9g>B[A=a`0O1N3N1O2M9HcW\\3"}, "label": "a man on a skateboard"}]} {"id": 485014, "image": "COCO_train2014_000000485014.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women wearing black dress holding a phone in the right hand\"."}], "task": "refering", "answer_template": "The \"a women wearing black dress holding a phone in the right hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 80, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 166, 167, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 195, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 292, 293, 294, 295, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 338, 339, 340, 341, 342, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 379, 380, 381, 382, 383, 384, 385, 387, 388, 389], "bbox": [0.24270312500000002, 0.13033333333333333, 0.9202187500000001, 0.9887708333333334], "iscrowd": 0, "area": 79437.12594999999, "rle": {"size": [480, 640], "counts": "QlX21o>2M3N3L3N3L3N2M4M2M4M1N2O0O2O0O2O0O2O0O2O0O2O001O010O010O0100O010O010O010O010O010O00100O010O0100O100O10O010000000000000O1000000000000O100000000000000O101O00000000000O100000000000000O1000000000000000O1O1O001N2O1O1O1O1O1N2O1O1O001O1N2O1O1O1O00000O1000000000O010000000000O1000O100000O10000000O10_Oa0O1O1O1O1O1N2O0`MbM_H_2]7jM]HW2_7RN[Ho1a7ZNYHg1R2ZMe1X1SL\\2Z1aL\\2[1TLk2h0PLl2^1VLd2l0RLg2c1WL]2P1ULa2g1YLU2T1ZL\\2h1[LP2W1\\LV2n1]Lh1[1[Li1[2lL\\1Y1ZL\\1i2ZMo0W1ZL\\1i2^Mn0T1ZL]1i2_Mm0S1[L]1h2`Mo0Q1ZL]1h2cMo0m0[L_1g2dMo0k0[L_1g2gMo0h0[L`1g2hMn0f0]L`1f2kMn0c0]La1f2lMn0a0]Lb1e2nMo0>]Lb1e2PNP1;]Ld1d2QNP19]Ld1d2TNo07^Ld1d2UNo05^Ld1d2XNo01^Lg1c2YNP1N^Lh1c2ZN[1ASLS2c2]NY3c1hL[NY3e1gL[NY3e1hLZNX3e1jLZNV3f1kLYNU3g1kLYNU3g1lLXNT3h1mLWNS3i1nLUNS3k1mLUNS3k1nLTNR3l1oLSNQ3m1PMRNP3n1PMRNP3m1WMmMi2S2R500000000000000000000O100000000000000000000O10000000000000000000000001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O001O00001O0gG^MT4b2kK_MU4b2jK_MU4a2jKaMU4`2jK`MV4a2iK`MV4a2hK`MX4a2gK`MX4`2gKbMX4_2gKaMY4`2eKbMZ4_2eKaM[4`2dKaM[4_2_J^M[N5U7^2]JiNc5X1YJlNf5U1WJmNi5T1SJPOl5Q1QJROn5n0oIUOQ6l0lIWOS6j0iIYOW6h0eI\\OZ6e0cI]O]6c0`IA_6`0^ICa6>[IEe6 in this image.", "objects": [{"patches": [58, 59, 60, 61, 80, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 166, 167, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 195, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 292, 293, 294, 295, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 338, 339, 340, 341, 342, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 379, 380, 381, 382, 383, 384, 385, 387, 388, 389], "bbox": [0.24270312500000002, 0.13033333333333333, 0.9202187500000001, 0.9887708333333334], "iscrowd": 0, "area": 79437.12594999999, "rle": {"size": [480, 640], "counts": "QlX21o>2M3N3L3N3L3N2M4M2M4M1N2O0O2O0O2O0O2O0O2O0O2O001O010O010O0100O010O010O010O010O010O00100O010O0100O100O10O010000000000000O1000000000000O100000000000000O101O00000000000O100000000000000O1000000000000000O1O1O001N2O1O1O1O1O1N2O1O1O001O1N2O1O1O1O00000O1000000000O010000000000O1000O100000O10000000O10_Oa0O1O1O1O1O1N2O0`MbM_H_2]7jM]HW2_7RN[Ho1a7ZNYHg1R2ZMe1X1SL\\2Z1aL\\2[1TLk2h0PLl2^1VLd2l0RLg2c1WL]2P1ULa2g1YLU2T1ZL\\2h1[LP2W1\\LV2n1]Lh1[1[Li1[2lL\\1Y1ZL\\1i2ZMo0W1ZL\\1i2^Mn0T1ZL]1i2_Mm0S1[L]1h2`Mo0Q1ZL]1h2cMo0m0[L_1g2dMo0k0[L_1g2gMo0h0[L`1g2hMn0f0]L`1f2kMn0c0]La1f2lMn0a0]Lb1e2nMo0>]Lb1e2PNP1;]Ld1d2QNP19]Ld1d2TNo07^Ld1d2UNo05^Ld1d2XNo01^Lg1c2YNP1N^Lh1c2ZN[1ASLS2c2]NY3c1hL[NY3e1gL[NY3e1hLZNX3e1jLZNV3f1kLYNU3g1kLYNU3g1lLXNT3h1mLWNS3i1nLUNS3k1mLUNS3k1nLTNR3l1oLSNQ3m1PMRNP3n1PMRNP3m1WMmMi2S2R500000000000000000000O100000000000000000000O10000000000000000000000001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O001O00001O0gG^MT4b2kK_MU4b2jK_MU4a2jKaMU4`2jK`MV4a2iK`MV4a2hK`MX4a2gK`MX4`2gKbMX4_2gKaMY4`2eKbMZ4_2eKaM[4`2dKaM[4_2_J^M[N5U7^2]JiNc5X1YJlNf5U1WJmNi5T1SJPOl5Q1QJROn5n0oIUOQ6l0lIWOS6j0iIYOW6h0eI\\OZ6e0cI]O]6c0`IA_6`0^ICa6>[IEe6 in this image.", "objects": [{"patches": [46, 69, 92, 115, 116, 138, 139, 161, 162, 184, 185, 207, 208, 230, 231, 253, 254, 276, 277, 278, 299, 300, 301, 322, 323, 324, 347], "bbox": [0.000703125, 0.12677083333333333, 0.12821875, 0.8963541666666667], "iscrowd": 0, "area": 14269.0021, "rle": {"size": [480, 640], "counts": "i;m0S5P9TOl0100O001O001ZFiKa8X4\\GnK`8R4_GQL_8P4_GSL_8Q4YGTLf8U5N1O2N1O2N2O0O2N2N101N2N2O0O2N2N102M2N3M2O2RJfIP3\\6lLjIP3[6kLhIR3^6gLfIW3^6cLeI[3^6aLeI]3^6_LdI`3_6[LeId3^6WLeIg3c6nKbIP4i8N2O0eLRFb1P:nMeFk1]9PNjFm1V9oMoFo1S9mMQGQ2S9hMRGV2m:01N2N2O1N2O2M3M3N2M3M3N2N2M3N2N2M2O2M2O1N2O1N2N2O2M2O1N2O5JoTU8"}, "label": "the back of a striped shirt facing away from an overjoyed woman playing a nintendo wii game"}]} {"id": 485014, "image": "COCO_train2014_000000485014.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a striped shirt\"."}], "task": "refering", "answer_template": "The \"a man in a striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 69, 92, 115, 116, 138, 139, 161, 162, 184, 185, 207, 208, 230, 231, 253, 254, 276, 277, 278, 299, 300, 301, 322, 323, 324, 347], "bbox": [0.000703125, 0.12677083333333333, 0.12821875, 0.8963541666666667], "iscrowd": 0, "area": 14269.0021, "rle": {"size": [480, 640], "counts": "i;m0S5P9TOl0100O001O001ZFiKa8X4\\GnK`8R4_GQL_8P4_GSL_8Q4YGTLf8U5N1O2N1O2N2O0O2N2N101N2N2O0O2N2N102M2N3M2O2RJfIP3\\6lLjIP3[6kLhIR3^6gLfIW3^6cLeI[3^6aLeI]3^6_LdI`3_6[LeId3^6WLeIg3c6nKbIP4i8N2O0eLRFb1P:nMeFk1]9PNjFm1V9oMoFo1S9mMQGQ2S9hMRGV2m:01N2N2O1N2O2M3M3N2M3M3N2N2M3N2N2M2O2M2O1N2O1N2N2O2M2O1N2O5JoTU8"}, "label": "a man in a striped shirt"}]} {"id": 485014, "image": "COCO_train2014_000000485014.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an office chair closest to the corner\"."}], "task": "refering", "answer_template": "The \"an office chair closest to the corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 209, 210, 211, 212, 231, 232, 233, 234, 254, 255, 256, 257, 277, 278, 279, 280, 301, 302, 303, 324, 325, 326, 327], "bbox": [0.06759375, 0.5005416666666667, 0.230125, 0.8641249999999999], "iscrowd": 0, "area": 10940.1861, "rle": {"size": [480, 640], "counts": "nmd03m>3M3kNJZC8c in this image.", "objects": [{"patches": [79, 101, 102, 103, 104, 125, 126, 127, 128, 129, 149, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 265, 266, 267, 288, 289, 290, 313, 314, 335, 336, 337, 358, 359, 360], "bbox": [0.42303125, 0.21797916666666667, 0.684265625, 0.9123541666666667], "iscrowd": 0, "area": 23354.525200000007, "rle": {"size": [480, 640], "counts": "iTo35j>8H7J4K2N3M2N2N3N1N3M2N2N3M2N2O0O2N2N101O001O0010O01O1O0010O01O1O3M3N2M3NO1N2O1N2O1OoEQNR7m1mHUNT7j1jHYNV7X1VFgNc24W7o0ZISOf6n0VIUOj6l0RIWOo6i0mHZOS7g0jH[OV7f0fH]OZ7d0bH_O^7a0_HBa7?]HBc7?[HAf7`0XHAh7`0VHAj7`0SHBl7`0RHAm7a0QH@n7b0PH_Oo7c0oG^OP8d0mG^OR8c0kG@T8b0iG@V8b0fGAY8a0cGB:`NU6P2]IB>_NT6T2WIBb0ZNW6Y2QI@e0YNY6]2jH_Oj0UN[6a2dH^On0RN^6f2\\H]OS1nM`6Z5]IhJb6[5[IfJd6]5YIcJg6_5VIcJi6`5TIaJk6a5RI`Jn6c5oH]JR7e5kH[JU7g5hHZJX7i5eHWJ[7k5bHVJ_7e61O001O1O001O1O1OjInH^4R7aKPI_4o6_KTIa4k6^KWIb4h6]KZIb4f6^K[Ib4d6^K\\Ic4c6\\K_Id4`6\\KaIc4_6]KbIc4]6\\KeId4Z6\\KgId4X6\\KhId4X6\\KhIe4W6ZKjIg4U6YKlIf4T6ZKlIg4S6XKnIi4Q6WKoIj4P6VKPJj4P6VKQJj4n5UKSJl4l5TKTJm4k5SKUJm4k5SKVJm4i5RKXJo4g5QKYJP5f5PKZJQ5h5kJYJV5j5fJWJZ5k5cJUJ^5n5^JRJc5Q6XJPJi5R6TJoIl5T6PJlIQ6W6jIjIW6i3_IfM7aN[6e3dIkMM`N`6b3hImNZ6P1kImNV6o0PJnNQ6o0SJoNn5n0WJoNj5m0\\JPOe5m0`JoNb5n0cJoN^5m0hJPOY5m0kJQOV5l0oJQOR5l0SKQOn4k0XKQOj4l0]KoNd4n0bKnN_4n0hKnNY4o0mKmNT4P1SLjNo3S1WLiNj3S1]LiNd3T1cLgN^3V1hLfNZ3V1mLeNU3V1SMdNo2Y1WMcNk2Y1\\MbNf2Y1bMbN`2Z1fMbN[2[1h5L4K5L3M4L4K?BP_n2"}, "label": "man in black clothes on a skateboard"}]} {"id": 124569, "image": "COCO_train2014_000000124569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man on the skateboard is wearing knee and elbow pads\"."}], "task": "refering", "answer_template": "The \"the man on the skateboard is wearing knee and elbow pads\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 101, 102, 103, 104, 125, 126, 127, 128, 129, 149, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 265, 266, 267, 288, 289, 290, 313, 314, 335, 336, 337, 358, 359, 360], "bbox": [0.42303125, 0.21797916666666667, 0.684265625, 0.9123541666666667], "iscrowd": 0, "area": 23354.525200000007, "rle": {"size": [480, 640], "counts": "iTo35j>8H7J4K2N3M2N2N3N1N3M2N2N3M2N2O0O2N2N101O001O0010O01O1O0010O01O1O3M3N2M3NO1N2O1N2O1OoEQNR7m1mHUNT7j1jHYNV7X1VFgNc24W7o0ZISOf6n0VIUOj6l0RIWOo6i0mHZOS7g0jH[OV7f0fH]OZ7d0bH_O^7a0_HBa7?]HBc7?[HAf7`0XHAh7`0VHAj7`0SHBl7`0RHAm7a0QH@n7b0PH_Oo7c0oG^OP8d0mG^OR8c0kG@T8b0iG@V8b0fGAY8a0cGB:`NU6P2]IB>_NT6T2WIBb0ZNW6Y2QI@e0YNY6]2jH_Oj0UN[6a2dH^On0RN^6f2\\H]OS1nM`6Z5]IhJb6[5[IfJd6]5YIcJg6_5VIcJi6`5TIaJk6a5RI`Jn6c5oH]JR7e5kH[JU7g5hHZJX7i5eHWJ[7k5bHVJ_7e61O001O1O001O1O1OjInH^4R7aKPI_4o6_KTIa4k6^KWIb4h6]KZIb4f6^K[Ib4d6^K\\Ic4c6\\K_Id4`6\\KaIc4_6]KbIc4]6\\KeId4Z6\\KgId4X6\\KhId4X6\\KhIe4W6ZKjIg4U6YKlIf4T6ZKlIg4S6XKnIi4Q6WKoIj4P6VKPJj4P6VKQJj4n5UKSJl4l5TKTJm4k5SKUJm4k5SKVJm4i5RKXJo4g5QKYJP5f5PKZJQ5h5kJYJV5j5fJWJZ5k5cJUJ^5n5^JRJc5Q6XJPJi5R6TJoIl5T6PJlIQ6W6jIjIW6i3_IfM7aN[6e3dIkMM`N`6b3hImNZ6P1kImNV6o0PJnNQ6o0SJoNn5n0WJoNj5m0\\JPOe5m0`JoNb5n0cJoN^5m0hJPOY5m0kJQOV5l0oJQOR5l0SKQOn4k0XKQOj4l0]KoNd4n0bKnN_4n0hKnNY4o0mKmNT4P1SLjNo3S1WLiNj3S1]LiNd3T1cLgN^3V1hLfNZ3V1mLeNU3V1SMdNo2Y1WMcNk2Y1\\MbNf2Y1bMbN`2Z1fMbN[2[1h5L4K5L3M4L4K?BP_n2"}, "label": "the man on the skateboard is wearing knee and elbow pads"}]} {"id": 411289, "image": "COCO_train2014_000000411289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in dark athletic pants and sweatshirt getting supplies ready for fishing\"."}], "task": "refering", "answer_template": "The \"a man in dark athletic pants and sweatshirt getting supplies ready for fishing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 95, 96, 97, 118, 119, 120, 121, 141, 142, 143, 144, 165, 166, 188, 189, 211, 212, 213, 234, 235], "bbox": [0.164328125, 0.2629166666666667, 0.279359375, 0.8516944444444445], "iscrowd": 0, "area": 9633.41065, "rle": {"size": [360, 640], "counts": "_QU1T1R:4L3L4N3M2N3M]OSGnNk8V1a04L4LW1dNZMjGg2U87N101OU2kM5J6K6J5J5L00O01O100H8N101N2O1000O10000UNk1B>3N2M3N2PKdKn2_4XLXLi3j3]KnLc4l4O100OO2O009G4L9G001N1N2M4L3N2nL]JU5_O[KOj4MiK^O\\4`0VLkNP4Q1W3M3K5K5K5L4K5K5K5KUmQ5"}, "label": "a man in dark athletic pants and sweatshirt getting supplies ready for fishing"}]} {"id": 411289, "image": "COCO_train2014_000000411289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with black pants whose back side can only be seen\"."}], "task": "refering", "answer_template": "The \"a man with black pants whose back side can only be seen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 95, 96, 97, 118, 119, 120, 121, 141, 142, 143, 144, 165, 166, 188, 189, 211, 212, 213, 234, 235], "bbox": [0.164328125, 0.2629166666666667, 0.279359375, 0.8516944444444445], "iscrowd": 0, "area": 9633.41065, "rle": {"size": [360, 640], "counts": "_QU1T1R:4L3L4N3M2N3M]OSGnNk8V1a04L4LW1dNZMjGg2U87N101OU2kM5J6K6J5J5L00O01O100H8N101N2O1000O10000UNk1B>3N2M3N2PKdKn2_4XLXLi3j3]KnLc4l4O100OO2O009G4L9G001N1N2M4L3N2nL]JU5_O[KOj4MiK^O\\4`0VLkNP4Q1W3M3K5K5K5L4K5K5K5KUmQ5"}, "label": "a man with black pants whose back side can only be seen"}]} {"id": 411289, "image": "COCO_train2014_000000411289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boat on the left\"."}], "task": "refering", "answer_template": "The \"the boat on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 123, 124, 138, 139, 140, 146, 147, 161, 162, 163, 164, 165, 167, 169, 184, 185, 186, 187, 188, 190, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234], "bbox": [0.0, 0.3536944444444444, 0.40896875, 0.8369166666666668], "iscrowd": 0, "area": 15778.236800000002, "rle": {"size": [360, 640], "counts": "_54e2m0l4YObIP2^6]110O001O010O00010O000001O00001O0000001O000O2O0000001O00001O0000000000001O000000000000001O00000000000000001N100000000000001O0000001O0000001N1000001O0000001O0000001O0O10001O0000001O00000O10000000000000001N10000O1000000O10000fNhGCY8K6N4K5LShT4"}, "label": "the boat on the left"}]} {"id": 411289, "image": "COCO_train2014_000000411289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boat on the let\"."}], "task": "refering", "answer_template": "The \"the boat on the let\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 123, 124, 138, 139, 140, 146, 147, 161, 162, 163, 164, 165, 167, 169, 184, 185, 186, 187, 188, 190, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234], "bbox": [0.0, 0.3536944444444444, 0.40896875, 0.8369166666666668], "iscrowd": 0, "area": 15778.236800000002, "rle": {"size": [360, 640], "counts": "_54e2m0l4YObIP2^6]110O001O010O00010O000001O00001O0000001O000O2O0000001O00001O0000000000001O000000000000001O00000000000000001N100000000000001O0000001O0000001N1000001O0000001O0000001O0O10001O0000001O00000O10000000000000001N10000O1000000O10000fNhGCY8K6N4K5LShT4"}, "label": "the boat on the let"}]} {"id": 411289, "image": "COCO_train2014_000000411289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boat with a fisherman standing in it\"."}], "task": "refering", "answer_template": "The \"a boat with a fisherman standing in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 113, 128, 129, 131, 132, 133, 134, 135, 136, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273], "bbox": [0.56503125, 0.2943888888888889, 0.969515625, 0.8831388888888889], "iscrowd": 0, "area": 36398.335450000006, "rle": {"size": [360, 640], "counts": "]^o37k:8G9H7I8M3M3M2N3N2M3M2N3M3M3M3M2N3M3N2N3M3M3M3N2M3M3M3N2M3M3M3N2M3M2N3N0O001O100O1O1O010O1O1O100O001O1002N2N1N3N2M2O2M3M2O0O010O010O10O01O10O010O0100O010O0000O100O2O0O100O100O2O0O100O100O2O0O100fNcLhJ^3W5hLbJZ3^5kLXJ[3g5iLoI]3Q6m0000000O2O0000000000001N10000000001O0O100000001O000O1000001O00000000001N10000000001O00000000001O000000000O2O00000000001O00000000001O00000000O1000000O1000000O2O000O2O00001N10001O0O101O000O2O000O2OO100O010O1000O0100O10O01000O0100O10O10O100O010O100N101N2O0O2N2O1N101O00N2M3N2N2M3N2N2N2aKVJ^3m5PLkJh3b6G8H8H9F9H9G8G:F9G9H9F9G:F_j6"}, "label": "a boat with a fisherman standing in it"}]} {"id": 411289, "image": "COCO_train2014_000000411289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boat on the right\"."}], "task": "refering", "answer_template": "The \"the boat on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 113, 128, 129, 131, 132, 133, 134, 135, 136, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273], "bbox": [0.56503125, 0.2943888888888889, 0.969515625, 0.8831388888888889], "iscrowd": 0, "area": 36398.335450000006, "rle": {"size": [360, 640], "counts": "]^o37k:8G9H7I8M3M3M2N3N2M3M2N3M3M3M3M2N3M3N2N3M3M3M3N2M3M3M3N2M3M3M3N2M3M2N3N0O001O100O1O1O010O1O1O100O001O1002N2N1N3N2M2O2M3M2O0O010O010O10O01O10O010O0100O010O0000O100O2O0O100O100O2O0O100O100O2O0O100fNcLhJ^3W5hLbJZ3^5kLXJ[3g5iLoI]3Q6m0000000O2O0000000000001N10000000001O0O100000001O000O1000001O00000000001N10000000001O00000000001O000000000O2O00000000001O00000000001O00000000O1000000O1000000O2O000O2O00001N10001O0O101O000O2O000O2OO100O010O1000O0100O10O01000O0100O10O10O100O010O100N101N2O0O2N2O1N101O00N2M3N2N2M3N2N2N2aKVJ^3m5PLkJh3b6G8H8H9F9H9G8G:F9G9H9F9G:F_j6"}, "label": "the boat on the right"}]} {"id": 280228, "image": "COCO_train2014_000000280228.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red headed girl in a yellow blouse holding a knife in each hand posed as if she is about to stab a guy laying on the table\"."}], "task": "refering", "answer_template": "The \"a red headed girl in a yellow blouse holding a knife in each hand posed as if she is about to stab a guy laying on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 38, 39, 40, 41, 42, 43, 55, 56, 57, 58, 59, 60, 61, 72, 73, 74, 75, 76, 77, 78], "bbox": [0.24735416666666668, 0.14376175548589343, 0.5990208333333334, 0.46887147335423196], "iscrowd": 0, "area": 9588.142150000003, "rle": {"size": [319, 480], "counts": "nTU14i95dGHi6>mHIP7:lHJQ7:jHJm6?oHEP7 in this image.", "objects": [{"patches": [23, 24, 38, 39, 40, 41, 42, 43, 55, 56, 57, 58, 59, 60, 61, 72, 73, 74, 75, 76, 77, 78], "bbox": [0.24735416666666668, 0.14376175548589343, 0.5990208333333334, 0.46887147335423196], "iscrowd": 0, "area": 9588.142150000003, "rle": {"size": [319, 480], "counts": "nTU14i95dGHi6>mHIP7:lHJQ7:jHJm6?oHEP7 in this image.", "objects": [{"patches": [122, 144, 145, 146, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 218, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.07628125, 0.3865258215962441, 0.52203125, 0.9865258215962441], "iscrowd": 0, "area": 48085.9282, "rle": {"size": [426, 640], "counts": "jgd06P=:F9G6J5K5K5L4L4L4L3M4M3M3L4M3M3L3K6\\OXM`Fl2a9YMUFk2m982N1O2O0O2N2N1O2O0O2N2N1O2N2OO0O1O1O1N3N1O1O1N2O2N1O1N2O1O2WFlLY9U3`FRM^9^3O1O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O100O100O100O100O100O100O100O100O010O100O100O100O10000O100O100O100O100O100O100O100O1SOeJkI[5n5XKeIi4T6_KkIa4o5fKoI[4P6gKoIY4P6jKmIW4Q6lKnIT4Q6nKmIS4R6oKmIQ4R6QLlIP4R6e1O1O1O1N2O1O1O1N2O1N2N2M3N2000000000000O100000000000O0100000000000000O100000000000000001O001O00001O00001O001O00002N1O1O2N1O1O2N1O1O2N1O2N2N3M2N3M2N2N1O001O1O1O1O1O1O1O1O1O1O1O1N2O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001lJdHX4\\7dKjHZ4W7`KPI^4Q7]KUIa4k6[K[Ic4h7N1O2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N:F:F:F;E:E in this image.", "objects": [{"patches": [122, 144, 145, 146, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 218, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.07628125, 0.3865258215962441, 0.52203125, 0.9865258215962441], "iscrowd": 0, "area": 48085.9282, "rle": {"size": [426, 640], "counts": "jgd06P=:F9G6J5K5K5L4L4L4L3M4M3M3L4M3M3L3K6\\OXM`Fl2a9YMUFk2m982N1O2O0O2N2N1O2O0O2N2N1O2N2OO0O1O1O1N3N1O1O1N2O2N1O1N2O1O2WFlLY9U3`FRM^9^3O1O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O100O100O100O100O100O100O100O100O010O100O100O100O10000O100O100O100O100O100O100O100O1SOeJkI[5n5XKeIi4T6_KkIa4o5fKoI[4P6gKoIY4P6jKmIW4Q6lKnIT4Q6nKmIS4R6oKmIQ4R6QLlIP4R6e1O1O1O1N2O1O1O1N2O1N2N2M3N2000000000000O100000000000O0100000000000000O100000000000000001O001O00001O00001O001O00002N1O1O2N1O1O2N1O1O2N1O2N2N3M2N3M2N2N1O001O1O1O1O1O1O1O1O1O1O1O1N2O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001lJdHX4\\7dKjHZ4W7`KPI^4Q7]KUIa4k6[K[Ic4h7N1O2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N:F:F:F;E:E in this image.", "objects": [{"patches": [69, 70, 92, 93, 115, 116, 117, 138, 139, 140, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 230, 231, 253], "bbox": [0.0, 0.21974178403755867, 0.150734375, 0.7488732394366197], "iscrowd": 0, "area": 14473.873, "rle": {"size": [426, 640], "counts": "T3f1c;\\3eLe1ZN2O1O0O2O001N101N0100O010O100O101N100O2O0O1O2O0O1O2N100O2N100O2N100O2N1O101N1O4M7H7I8I6I7I8H7J2M1O2O1N2N101N2N2N2O0O2N2O1N1O2N2O1N1O2O1N2N2O0O2N2N2O0O2L5L5J6K5J6K5J6K5J6K5J6K5J6K5J6K5J7JRSR7"}, "label": "man with beard and glasses , wearing a navy blue hooded sweatshirt"}]} {"id": 181929, "image": "COCO_train2014_000000181929.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing blue sweater and wearing glasses\"."}], "task": "refering", "answer_template": "The \"a person wearing blue sweater and wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 92, 93, 115, 116, 117, 138, 139, 140, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 230, 231, 253], "bbox": [0.0, 0.21974178403755867, 0.150734375, 0.7488732394366197], "iscrowd": 0, "area": 14473.873, "rle": {"size": [426, 640], "counts": "T3f1c;\\3eLe1ZN2O1O0O2O001N101N0100O010O100O101N100O2O0O1O2O0O1O2N100O2N100O2N100O2N1O101N1O4M7H7I8I6I7I8H7J2M1O2O1N2N101N2N2N2O0O2N2O1N1O2N2O1N1O2O1N2N2O0O2N2N2O0O2L5L5J6K5J6K5J6K5J6K5J6K5J6K5J6K5J7JRSR7"}, "label": "a person wearing blue sweater and wearing glasses"}]} {"id": 181929, "image": "COCO_train2014_000000181929.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with phone\"."}], "task": "refering", "answer_template": "The \"man with phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 109, 110, 111, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 291, 292, 293, 294, 315, 316, 317, 338, 339], "bbox": [0.668625, 0.23490610328638495, 0.880484375, 0.9482394366197183], "iscrowd": 0, "area": 27019.829749999997, "rle": {"size": [426, 640], "counts": "YYb53Q=7H7J6I7J6eETOY71[GR1Y1ZOP7MZGo0c1]Ok6Y1RIoNf6T1WITOb6o0ZIYO_6i0^I^O^6b0_IE\\62N3M3O010000O01000O100N2N2N2N2M2O2N2N2N2N2N2M3N2N1O2N2N2N2M3N2N2N2N1O2N2M3N2N5I6XKZId2m6XMVIa2R7\\MPI^2V7_MmH[2Z7bMgHY2`7cMcHW2c7gM_HS2h7jMZHo1m7nMUHm1R8PNPHj1W8RNlGh1Z8VNhGd1_8XNdGa1d8\\N]G_1i8^NZG\\1m8aNUGV1U9fNnFP1\\9nNeFi0f9SO]Fb0n9\\OTF:d;Fe]o0"}, "label": "man with phone"}]} {"id": 181929, "image": "COCO_train2014_000000181929.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a brown jacket holding a camera\"."}], "task": "refering", "answer_template": "The \"a man in a brown jacket holding a camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 109, 110, 111, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 291, 292, 293, 294, 315, 316, 317, 338, 339], "bbox": [0.668625, 0.23490610328638495, 0.880484375, 0.9482394366197183], "iscrowd": 0, "area": 27019.829749999997, "rle": {"size": [426, 640], "counts": "YYb53Q=7H7J6I7J6eETOY71[GR1Y1ZOP7MZGo0c1]Ok6Y1RIoNf6T1WITOb6o0ZIYO_6i0^I^O^6b0_IE\\62N3M3O010000O01000O100N2N2N2N2M2O2N2N2N2N2N2M3N2N1O2N2N2N2M3N2N2N2N1O2N2M3N2N5I6XKZId2m6XMVIa2R7\\MPI^2V7_MmH[2Z7bMgHY2`7cMcHW2c7gM_HS2h7jMZHo1m7nMUHm1R8PNPHj1W8RNlGh1Z8VNhGd1_8XNdGa1d8\\N]G_1i8^NZG\\1m8aNUGV1U9fNnFP1\\9nNeFi0f9SO]Fb0n9\\OTF:d;Fe]o0"}, "label": "a man in a brown jacket holding a camera"}]} {"id": 181929, "image": "COCO_train2014_000000181929.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an asian girl with glasses\"."}], "task": "refering", "answer_template": "The \"an asian girl with glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 141, 142, 143, 164, 165, 166, 187, 188, 189, 209, 210, 211, 232, 233, 234], "bbox": [0.11295312500000002, 0.33711267605633805, 0.263546875, 0.7307746478873239], "iscrowd": 0, "area": 9086.849399999997, "rle": {"size": [426, 640], "counts": "bWn0;m<4L4L4L4L3M3L3N2N2N2N2N2M3N2N2cM_NnHd1Q7aNgHc1X7bN`Ha1`7cNYHa1e7cNUH`1e7iNWHY1a7QO[HP1^7[O]Hg0c7[OYHf0g7]OUHe0j7^ORHc0o7_OmGc0R8_OkGb0V8@fGb0Y8_OeGb0\\8^ObGd0]8T21O000O2O001O1O1N101N2M3M3M3M3M3N2N2N2N2N2N2O100O100O100O100O100O101N2O0O2O1N101N2O1N101N4M2M3N2N2M3N3L3N2QNXEW1k:_NgEW1\\;I4M3L4L3N3L4M2L5I]TT6"}, "label": "an asian girl with glasses"}]} {"id": 181929, "image": "COCO_train2014_000000181929.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with glasses standing beside a man with glasses\"."}], "task": "refering", "answer_template": "The \"a woman with glasses standing beside a man with glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 141, 142, 143, 164, 165, 166, 187, 188, 189, 209, 210, 211, 232, 233, 234], "bbox": [0.11295312500000002, 0.33711267605633805, 0.263546875, 0.7307746478873239], "iscrowd": 0, "area": 9086.849399999997, "rle": {"size": [426, 640], "counts": "bWn0;m<4L4L4L4L3M3L3N2N2N2N2N2M3N2N2cM_NnHd1Q7aNgHc1X7bN`Ha1`7cNYHa1e7cNUH`1e7iNWHY1a7QO[HP1^7[O]Hg0c7[OYHf0g7]OUHe0j7^ORHc0o7_OmGc0R8_OkGb0V8@fGb0Y8_OeGb0\\8^ObGd0]8T21O000O2O001O1O1N101N2M3M3M3M3M3N2N2N2N2N2N2O100O100O100O100O100O101N2O0O2O1N101N2O1N101N4M2M3N2N2M3N3L3N2QNXEW1k:_NgEW1\\;I4M3L4L3N3L4M2L5I]TT6"}, "label": "a woman with glasses standing beside a man with glasses"}]} {"id": 18093, "image": "COCO_train2014_000000018093.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the inside of an oven\"."}], "task": "refering", "answer_template": "The \"the inside of an oven\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 111, 112, 113, 114, 115, 116, 117, 118, 134, 135, 136, 137, 138, 139, 140, 141, 157, 158, 159, 160, 161, 162, 163, 164, 181, 182, 183, 184, 185, 186, 187, 204, 205, 206, 207, 208, 209, 210, 227, 228, 229, 230, 231, 232, 233, 250, 251, 252, 253, 254, 255, 256, 257, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0015, 0.004637002341920374, 1.0, 0.968688524590164], "iscrowd": 0, "area": 139025.76864999998, "rle": {"size": [427, 640], "counts": "a=QR9BoF>P9BQG>o8ARG?m8BSG>m8ATG?k8AVG?i8AXG?h8@YG`0f8@[G`0e8_O\\Ga0c8@]G`0c8_O^Ga0a8_O`Ga0_8_ObGa0^8^OcGb0\\8^OeGb0b:0hM^OgGb0X8^OiGb0`:0000000000000000000cLIfI7f5`0WJ@U5X1gJhNe4o1XKQNT4e2jK[Mb3X3_LhLm2m3RMSLZ2a4fM_Kf1V5YNjJS1j5mNVJ`0^6_ObIa0^6_ObI`0`6_O`Ia0`6_O`I`0b6_O^Ia0b6_O^I`0d6_O\\I`0e6@[I`0f6_OZI`0g6@YI`0h6_OXI`0i6@WI?k6@UI`0k6@UI?m6@SI`0m6@SI?o6@QI?P7API?Q7@oH?R7AnH?S7@mH?T7AlH?U7@kH?V7AjH>W7BiH>W7BiH=X7ChH=X7ChH_8BaG>_8BaG>_8BaG>_8BaG?^8AbG?^8AbG?^8AbG?^8AbG?^8AbG?^8AbG?_8@aGa0^8_ObGa0^8_ObGa0^8_ObGa0^8_ObGa0^8_ObGa0]8@cG`0]8@cG`0]8@cGa0\\8_OdGa0\\8_OdGa0\\8_OdGa0\\8_OdGa0\\8_OdGa0\\8_OdGa0[8@eGa0Z8_OfGa0Z8_OfGa0Z8_OfGa0Z8_OfGa0Z8_OfGa0Z8_OfGa0Z8_OfGa0Y8@gGa0X8_OhGa0X8_OhGa0X8_OgGb0Y8^OgGb0Y8^OgGb0Y8^OgGb0Y8^OgGb0Y8^OgGc0W8^OiGb0W8^OiGb0W8^OiGb0W8^OiGb0W8^OiGb0W8^OiGb0W8^OiGc0V8]OjGc0V8]OjGc0U8^OkGb0U8^OkGb0U8^OkGb0U8^OkGb0U8^OkGb0U8^OkGc0T8]OlGc0T8]OlGc0S8^OmGb0S8^OmGb0S8^OmGb0S8^OmGb0S8^OmGb0S8^OmGc0R8]OnGc0R8]OnGc0R8]OnGc0Q8^OoGb0Q8^OoGb0Q8^OoGb0Q8^OoGc0P8]OPHc0P8]OPHc0P8]OPHc0P8]OPHc0P8]OPHc0o7^OQHb0o7^OQHb0o7^OQHc0n7]ORHc0n7]ORHc0n7]ORHc0n7]ORHc0n7]ORHc0m7^OSHb0m7^OSHb0m7^OSHc0l7]OTHc0l7]OTHc0l7]OTHc0l7]OTHc0l7]OTHc0l7]OTHc0k7^OUHc0j7]OVHc0j7]OVHc0j7]OVHc0j7]OVHc0j7]OVHc0j7]OVHc0j7]OVHc0j7]OVHd0h7]OXHc0h7]OXHc0h7]OXHc0h7]OXHc0h7]OXHc0h7]OXHc0h7]OXHd0g7\\OYHd0f7]OZHc0f7]OZHc0f7]OZHc0f7]OZHc0f7]OZHc0f7]OZHc0f7]OZHd0e7\\O[Hd0e7\\O[Hd0d7]O\\Hc0d7]O\\Hc0d7]O\\Hc0d7]O\\Hc0d7]O\\Hc0d7]O\\Hd0c7\\O]Hd0c7\\O]Hd0c7\\O]Hd0b7]O^Hc0b7]O^Hc0b7]O^Hc0b7]O^Hd0a7\\O_Hd0a7\\O_Hd0a7\\O_Hd0a7\\O_Hd0`7]O`Hc0`7]O`Hc0`7]O`Hc0`7]O`Hd0_7\\OaHd0_7\\OaHd0_7\\OaHd0_7\\OaHd0_7\\OaHd0^7]ObHc0^7]ObHc0^7]ObHd0]7\\OcHd0]7\\OcHd0]7\\OcHd0]7\\OcHd0]7\\OcHd0]7\\OcHd0\\7]OdHd0[7\\OeHd0[7\\OeHd0[7\\OeHd0[7\\OeHd0[7\\OeHd0[7\\OeHd0Z7]OfHc0Z7]OfHd0X7]OhHc0X7]OhHc0W7^OiHb0W7^OiHb0V7_OjHa0V7_OjH`0@XMd7Y2lH>@[Mc7W2mHc0Q7^OoH<@_M`7U2PI;@bM^7T2RI9@eM\\7S2TI8@fM[7R2UI7@iMY7Q2WI5@mMW7n1YI4@PNU7m1[I2ARNS7l1\\I2@TNR7k1^I0@WNQ7i1_IOAYNn6i1aIMA\\Nm6g1bIM@^Nl6f1dIK@aNj6e1fIIAdNg6c1hIHAgNe6b1jIG@iNe6`1kIFAkNb6`1mIDAnNa6^1nICAQO_6]1PJB@SO_6[1QJAAUO\\6[1SJ_OA\\OW6U1XJ^OAFn5m0aJ\\OBOd5f0jJ[OA8]5=RKZOAb0T55[KXOAl0l4LcKWOBU1b4ElKVOA^1[4\\OTLUOAh1R4TO\\LTOCP2i3lNdLSOCZ2`3dNmLROBc2Y3[NUMQOBm2P3SN^MoNCV3f2lMgMmNC`3^2cMoMlNCj3U2[MXNkNCR4m1SM`NjNC\\4d1kLiNhNCf4\\1bLQOgNCP5S1ZLZOfNCX5k0RLBeNCb5b0jKKcNCl5:aK3bNDU60ZKA?B7"}, "label": "the inside of an oven"}]} {"id": 18093, "image": "COCO_train2014_000000018093.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the inside of the oven\"."}], "task": "refering", "answer_template": "The \"the inside of the oven\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 111, 112, 113, 114, 115, 116, 117, 118, 134, 135, 136, 137, 138, 139, 140, 141, 157, 158, 159, 160, 161, 162, 163, 164, 181, 182, 183, 184, 185, 186, 187, 204, 205, 206, 207, 208, 209, 210, 227, 228, 229, 230, 231, 232, 233, 250, 251, 252, 253, 254, 255, 256, 257, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0015, 0.004637002341920374, 1.0, 0.968688524590164], "iscrowd": 0, "area": 139025.76864999998, "rle": {"size": [427, 640], "counts": "a=QR9BoF>P9BQG>o8ARG?m8BSG>m8ATG?k8AVG?i8AXG?h8@YG`0f8@[G`0e8_O\\Ga0c8@]G`0c8_O^Ga0a8_O`Ga0_8_ObGa0^8^OcGb0\\8^OeGb0b:0hM^OgGb0X8^OiGb0`:0000000000000000000cLIfI7f5`0WJ@U5X1gJhNe4o1XKQNT4e2jK[Mb3X3_LhLm2m3RMSLZ2a4fM_Kf1V5YNjJS1j5mNVJ`0^6_ObIa0^6_ObI`0`6_O`Ia0`6_O`I`0b6_O^Ia0b6_O^I`0d6_O\\I`0e6@[I`0f6_OZI`0g6@YI`0h6_OXI`0i6@WI?k6@UI`0k6@UI?m6@SI`0m6@SI?o6@QI?P7API?Q7@oH?R7AnH?S7@mH?T7AlH?U7@kH?V7AjH>W7BiH>W7BiH=X7ChH=X7ChH_8BaG>_8BaG>_8BaG>_8BaG?^8AbG?^8AbG?^8AbG?^8AbG?^8AbG?^8AbG?_8@aGa0^8_ObGa0^8_ObGa0^8_ObGa0^8_ObGa0^8_ObGa0]8@cG`0]8@cG`0]8@cGa0\\8_OdGa0\\8_OdGa0\\8_OdGa0\\8_OdGa0\\8_OdGa0\\8_OdGa0[8@eGa0Z8_OfGa0Z8_OfGa0Z8_OfGa0Z8_OfGa0Z8_OfGa0Z8_OfGa0Z8_OfGa0Y8@gGa0X8_OhGa0X8_OhGa0X8_OgGb0Y8^OgGb0Y8^OgGb0Y8^OgGb0Y8^OgGb0Y8^OgGc0W8^OiGb0W8^OiGb0W8^OiGb0W8^OiGb0W8^OiGb0W8^OiGb0W8^OiGc0V8]OjGc0V8]OjGc0U8^OkGb0U8^OkGb0U8^OkGb0U8^OkGb0U8^OkGb0U8^OkGc0T8]OlGc0T8]OlGc0S8^OmGb0S8^OmGb0S8^OmGb0S8^OmGb0S8^OmGb0S8^OmGc0R8]OnGc0R8]OnGc0R8]OnGc0Q8^OoGb0Q8^OoGb0Q8^OoGb0Q8^OoGc0P8]OPHc0P8]OPHc0P8]OPHc0P8]OPHc0P8]OPHc0o7^OQHb0o7^OQHb0o7^OQHc0n7]ORHc0n7]ORHc0n7]ORHc0n7]ORHc0n7]ORHc0m7^OSHb0m7^OSHb0m7^OSHc0l7]OTHc0l7]OTHc0l7]OTHc0l7]OTHc0l7]OTHc0l7]OTHc0k7^OUHc0j7]OVHc0j7]OVHc0j7]OVHc0j7]OVHc0j7]OVHc0j7]OVHc0j7]OVHc0j7]OVHd0h7]OXHc0h7]OXHc0h7]OXHc0h7]OXHc0h7]OXHc0h7]OXHc0h7]OXHd0g7\\OYHd0f7]OZHc0f7]OZHc0f7]OZHc0f7]OZHc0f7]OZHc0f7]OZHc0f7]OZHd0e7\\O[Hd0e7\\O[Hd0d7]O\\Hc0d7]O\\Hc0d7]O\\Hc0d7]O\\Hc0d7]O\\Hc0d7]O\\Hd0c7\\O]Hd0c7\\O]Hd0c7\\O]Hd0b7]O^Hc0b7]O^Hc0b7]O^Hc0b7]O^Hd0a7\\O_Hd0a7\\O_Hd0a7\\O_Hd0a7\\O_Hd0`7]O`Hc0`7]O`Hc0`7]O`Hc0`7]O`Hd0_7\\OaHd0_7\\OaHd0_7\\OaHd0_7\\OaHd0_7\\OaHd0^7]ObHc0^7]ObHc0^7]ObHd0]7\\OcHd0]7\\OcHd0]7\\OcHd0]7\\OcHd0]7\\OcHd0]7\\OcHd0\\7]OdHd0[7\\OeHd0[7\\OeHd0[7\\OeHd0[7\\OeHd0[7\\OeHd0[7\\OeHd0Z7]OfHc0Z7]OfHd0X7]OhHc0X7]OhHc0W7^OiHb0W7^OiHb0V7_OjHa0V7_OjH`0@XMd7Y2lH>@[Mc7W2mHc0Q7^OoH<@_M`7U2PI;@bM^7T2RI9@eM\\7S2TI8@fM[7R2UI7@iMY7Q2WI5@mMW7n1YI4@PNU7m1[I2ARNS7l1\\I2@TNR7k1^I0@WNQ7i1_IOAYNn6i1aIMA\\Nm6g1bIM@^Nl6f1dIK@aNj6e1fIIAdNg6c1hIHAgNe6b1jIG@iNe6`1kIFAkNb6`1mIDAnNa6^1nICAQO_6]1PJB@SO_6[1QJAAUO\\6[1SJ_OA\\OW6U1XJ^OAFn5m0aJ\\OBOd5f0jJ[OA8]5=RKZOAb0T55[KXOAl0l4LcKWOBU1b4ElKVOA^1[4\\OTLUOAh1R4TO\\LTOCP2i3lNdLSOCZ2`3dNmLROBc2Y3[NUMQOBm2P3SN^MoNCV3f2lMgMmNC`3^2cMoMlNCj3U2[MXNkNCR4m1SM`NjNC\\4d1kLiNhNCf4\\1bLQOgNCP5S1ZLZOfNCX5k0RLBeNCb5b0jKKcNCl5:aK3bNDU60ZKA?B7"}, "label": "the inside of the oven"}]} {"id": 18093, "image": "COCO_train2014_000000018093.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the door of the oven\"."}], "task": "refering", "answer_template": "The \"the door of the oven\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 224, 235, 236, 237, 244, 245, 246, 247, 258, 259, 260, 267, 268, 269, 270, 271, 282, 283, 289, 290, 291, 292, 293], "bbox": [0.21662499999999998, 0.550655737704918, 0.81621875, 0.8519437939110069], "iscrowd": 0, "area": 15413.953899999997, "rle": {"size": [427, 640], "counts": "ZYj11X=2N3M2O1N2N2O1N3M2N2O1N2N2O1N3M2O1N2N2N2O1N3M2O1N2N2N2O1N3M2O1N2N2001O100O1O1O2N1O100O1O1O1O1O2N100O1O1O00001O01O01O0XOWDNj;0YDNg;0^DMb;2bDK_;2eDMZ;2jDKV;3oDJR;3REKn:4U1N3NUdo0Om[PO1O1O1O1O1O1N2O1O1O1O1O1O1O00001O00000000000000001O00000000000O101O000000000000000000001O000000000000000000001O0O1000000000000000001O0000000000000000001O00000000000O100000001O0000000000000000000aFXO[6i0cIWO^6i0`IYO`6g0^I[Ob6e0\\I]Od6c0[I^Oe6b0YI_Oh6c0TI_Ol6c0QI^Oo6d0mH^OS7d0iH^OV7e0gH\\OY7e0dH\\O]7f0_H\\Oa7f0\\H[Od7g0XH[Oh7g0TH[Ol7g0QHYOP8j0kGXOU8k0fGWOZ8l0_GXOa8k0YGXOg8k0RGXOo8l0iFXOW9k0bFYO^9Z200000O100000000O1000O1000O100000000O10000000000O100000000O100000000O1000O100000O100000002M2O2N1O1O2M2O2N1O2N1N2O2N1O2N1O2M2O2N1O1O2M2O2N1O2N1N2O2N1O2N1O2M2O2N1O1O2M2O2N1O2N1N2O2N1O2N1O2M2O2N1O1O2M2O2N1O2N1N2O2N1O2N1N3N1O1O2N1O2MSl`1"}, "label": "the door of the oven"}]} {"id": 18093, "image": "COCO_train2014_000000018093.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the oven door\"."}], "task": "refering", "answer_template": "The \"the oven door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 224, 235, 236, 237, 244, 245, 246, 247, 258, 259, 260, 267, 268, 269, 270, 271, 282, 283, 289, 290, 291, 292, 293], "bbox": [0.21662499999999998, 0.550655737704918, 0.81621875, 0.8519437939110069], "iscrowd": 0, "area": 15413.953899999997, "rle": {"size": [427, 640], "counts": "ZYj11X=2N3M2O1N2N2O1N3M2N2O1N2N2O1N3M2O1N2N2N2O1N3M2O1N2N2N2O1N3M2O1N2N2001O100O1O1O2N1O100O1O1O1O1O2N100O1O1O00001O01O01O0XOWDNj;0YDNg;0^DMb;2bDK_;2eDMZ;2jDKV;3oDJR;3REKn:4U1N3NUdo0Om[PO1O1O1O1O1O1N2O1O1O1O1O1O1O00001O00000000000000001O00000000000O101O000000000000000000001O000000000000000000001O0O1000000000000000001O0000000000000000001O00000000000O100000001O0000000000000000000aFXO[6i0cIWO^6i0`IYO`6g0^I[Ob6e0\\I]Od6c0[I^Oe6b0YI_Oh6c0TI_Ol6c0QI^Oo6d0mH^OS7d0iH^OV7e0gH\\OY7e0dH\\O]7f0_H\\Oa7f0\\H[Od7g0XH[Oh7g0TH[Ol7g0QHYOP8j0kGXOU8k0fGWOZ8l0_GXOa8k0YGXOg8k0RGXOo8l0iFXOW9k0bFYO^9Z200000O100000000O1000O1000O100000000O10000000000O100000000O100000000O1000O100000O100000002M2O2N1O1O2M2O2N1O2N1N2O2N1O2N1O2M2O2N1O1O2M2O2N1O2N1N2O2N1O2N1O2M2O2N1O1O2M2O2N1O2N1N2O2N1O2N1O2M2O2N1O1O2M2O2N1O2N1N2O2N1O2N1N3N1O1O2N1O2MSl`1"}, "label": "the oven door"}]} {"id": 321209, "image": "COCO_train2014_000000321209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bench that is close to the mopads\"."}], "task": "refering", "answer_template": "The \"a bench that is close to the mopads\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 160, 176, 177, 192, 193, 194, 195, 210, 211, 212, 213], "bbox": [0.66812, 0.6966257668711656, 0.89962, 0.9887730061349692], "iscrowd": 0, "area": 4247.397650000001, "rle": {"size": [326, 500], "counts": "bcZ39m96J1O00100O1O0000000000100O1O1O2N1O1O0001O01OM3L4O1001O0010O01cFCo8j000000000001O000000001O0000001N10000O100O10000O010O10O2N101N1O1O2J5K6J5K5K6J5J6K5K5K5K5L41O0000001O000@kHoMU7o1QImMP7Q2UIkMk6S2e0N2O2M2N2M4K4M3M3M4M2M3N2M4M2M3N2M4M2N201N1O1O101N2N2N3N1Nke?"}, "label": "a bench that is close to the mopads"}]} {"id": 321209, "image": "COCO_train2014_000000321209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bench with three men sitting on it\"."}], "task": "refering", "answer_template": "The \"a bench with three men sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 160, 176, 177, 192, 193, 194, 195, 210, 211, 212, 213], "bbox": [0.66812, 0.6966257668711656, 0.89962, 0.9887730061349692], "iscrowd": 0, "area": 4247.397650000001, "rle": {"size": [326, 500], "counts": "bcZ39m96J1O00100O1O0000000000100O1O1O2N1O1O0001O01OM3L4O1001O0010O01cFCo8j000000000001O000000001O0000001N10000O100O10000O010O10O2N101N1O1O2J5K6J5K5K6J5J6K5K5K5K5L41O0000001O000@kHoMU7o1QImMP7Q2UIkMk6S2e0N2O2M2N2M4K4M3M3M4M2M3N2M4M2M3N2M4M2N201N1O1O101N2N2N3N1Nke?"}, "label": "a bench with three men sitting on it"}]} {"id": 321209, "image": "COCO_train2014_000000321209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the motor scooter that is in the middle of the other scooters and behind the man with the cane that is sticking straight out\"."}], "task": "refering", "answer_template": "The \"the motor scooter that is in the middle of the other scooters and behind the man with the cane that is sticking straight out\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 105, 107, 122, 123, 124, 125, 140, 141, 159, 160], "bbox": [0.79698, 0.39671779141104296, 1.0, 0.727239263803681], "iscrowd": 0, "area": 3442.2362499999995, "rle": {"size": [326, 500], "counts": "jVo34P:5K4L4L3M4L4M2M4L3M4L01K5N101O1O10O2YGBe7>ZHEd7<[HEc7<\\HFc7:]HGb7:\\HIa78`HH_79`HH_78bHH]79bHIk6QO]IV1HIi6WOZIQ1MHh6ZOYIn0NIi6ZOXIm0OIi6\\OVIk01Jh6\\OVIj01Ki6^OSIg04Kj6h0VIXOj6h0VIYOi6g0WIYOi6g0XIXOS7=mHCW79iHmNM=[7d0hHjN6>T7e0gHiNb100000001O00001O1O101O01O1O010O001O0O2N101N100000@`F:`9FaF9_9GbF8^9HbF7^9JbF5^9LbF3^9NcFN_93aFL`95_FK`96`FJ_97bFH]9:bFF^9:bFF^9:bFF^9;aFE_9;bFD^9`FB`9=aFC`9<`FD`9<6O002N4LlJ"}, "label": "the motor scooter that is in the middle of the other scooters and behind the man with the cane that is sticking straight out"}]} {"id": 321209, "image": "COCO_train2014_000000321209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black motor scooter parked on the curb\"."}], "task": "refering", "answer_template": "The \"black motor scooter parked on the curb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 142, 143, 160, 161, 177, 178, 179, 195, 196, 197, 214], "bbox": [0.8620800000000001, 0.5427914110429447, 1.0, 0.9390184049079755], "iscrowd": 0, "area": 4489.332450000002, "rle": {"size": [326, 500], "counts": "f_Y43S9NiG8S8KhG;U8GfG?V8DeGb0W8@eGc0Z8@bGc0\\8d0O2N1O2N1O2O001O0010OO2K4M3O001N100O2OO1O0O20O2O001N10000O100O10000O_O_H`Na7b1?3BoMdHT2W7QNfHQ2V7SNgHP2U7TNiHn1U7TNiHn1V7b0N2O1O0O10eNRI^On6?XI^Og6`0_I^O`6?fI^O[6>kI_OV6=PJ@P6=VJ@k5 in this image.", "objects": [{"patches": [125, 142, 143, 160, 161, 177, 178, 179, 195, 196, 197, 214], "bbox": [0.8620800000000001, 0.5427914110429447, 1.0, 0.9390184049079755], "iscrowd": 0, "area": 4489.332450000002, "rle": {"size": [326, 500], "counts": "f_Y43S9NiG8S8KhG;U8GfG?V8DeGb0W8@eGc0Z8@bGc0\\8d0O2N1O2N1O2O001O0010OO2K4M3O001N100O2OO1O0O20O2O001N10000O100O10000O_O_H`Na7b1?3BoMdHT2W7QNfHQ2V7SNgHP2U7TNiHn1U7TNiHn1V7b0N2O1O0O10eNRI^On6?XI^Og6`0_I^O`6?fI^O[6>kI_OV6=PJ@P6=VJ@k5 in this image.", "objects": [{"patches": [258, 259, 260, 281, 282, 283, 284, 285, 304, 305, 306, 307, 308, 309, 328, 329, 330, 331, 332, 352, 353, 354, 355], "bbox": [0.21459375, 0.6795208333333334, 0.470953125, 0.9425416666666667], "iscrowd": 0, "area": 11867.777650000002, "rle": {"size": [480, 640], "counts": "fbP24V>g0M2N3M3M3M2N3O1N101O1O1N101O1N101O1N101O1N2O001N2O001N2O1O0O2O1O0O2O1O1O0O2O1O0O2O1O1N101O1N101O1N1010000O010O10O10O1000O0100O0100O010O00100O010O010O0010O010O00010O010O010O01O01O010O010O01O01O010O010O01O010O01O010O01O010O01O01O010O010O010O00100O010O10O01O10O0100O1O001N2O1N2O1O2M2O1N2N2M4M2N2M3N2N2N3L3N2N1N10001O0O2O4L8G9H8HT`n4"}, "label": "a fake long john donut in a package"}]} {"id": 83641, "image": "COCO_train2014_000000083641.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plastic eclair toy\"."}], "task": "refering", "answer_template": "The \"a plastic eclair toy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [258, 259, 260, 281, 282, 283, 284, 285, 304, 305, 306, 307, 308, 309, 328, 329, 330, 331, 332, 352, 353, 354, 355], "bbox": [0.21459375, 0.6795208333333334, 0.470953125, 0.9425416666666667], "iscrowd": 0, "area": 11867.777650000002, "rle": {"size": [480, 640], "counts": "fbP24V>g0M2N3M3M3M2N3O1N101O1O1N101O1N101O1N101O1N2O001N2O001N2O1O0O2O1O0O2O1O1O0O2O1O0O2O1O1N101O1N101O1N1010000O010O10O10O1000O0100O0100O010O00100O010O010O0010O010O00010O010O010O01O01O010O010O01O01O010O010O01O010O01O010O01O010O01O01O010O010O010O00100O010O10O01O10O0100O1O001N2O1N2O1O2M2O1N2N2M4M2N2M3N2N2N3L3N2N1N10001O0O2O4L8G9H8HT`n4"}, "label": "a plastic eclair toy"}]} {"id": 149180, "image": "COCO_train2014_000000149180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"german shephard dog sleeping\"."}], "task": "refering", "answer_template": "The \"german shephard dog sleeping\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 78, 79, 80, 81, 82, 83, 84, 85, 101, 102, 103, 104, 105, 106, 107, 108, 109, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 191, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 271, 272, 273], "bbox": [0.295390625, 0.18933333333333333, 0.999375, 0.9358055555555556], "iscrowd": 0, "area": 57555.883050000004, "rle": {"size": [360, 640], "counts": "PeR24m:7N3N1O2M2O1O2N1O2N1N3O0O2N1O2N00000001N1000000000000O2O000000000O10001O00000O101O001O001O001O010O001O001O001O001OB_OVFb0f9A[F?b9e0L4M3L3M4M3L4L4lGSNj6P2RIUNk6l1RIYNm6g1QI\\Nm6e1RI]Nm6c1RI`Nk6a1TIaNj6_1VIcNi6]1UIfNi6[1VIgNi6X1TInNi6S1QIROo6o0mHUOS7k0lHUOT7k0mHUOR7l0mHTO0bNl6\\2SISO0cNk6Z2UISO0eNh6Z2WIRO0gNf6X2YIQO1iNc6W2\\IQO0jNb6V2^IoN1kNa6V2]IoN2mN`6T2^ImN3QO^6Q2`IlN4TO[6P2aIjN6WOX6o1bIgN9ZOT6]1hI\\NLi0<_Oo5\\1WJPOOEi5\\1WJkN4Jd5[1XJkN4Jc5\\1YJiN4Mb5Z1ZJhN5O`5Z1YJhN7O^5Z1[JfN80]5Z1[JeN92Z5[1\\JbN:5Y5Y1]JbN:6X5X1^JaN;7V5Y1_J_N;:U5X1_J]N= in this image.", "objects": [{"patches": [57, 58, 59, 60, 78, 79, 80, 81, 82, 83, 84, 85, 101, 102, 103, 104, 105, 106, 107, 108, 109, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 191, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 271, 272, 273], "bbox": [0.295390625, 0.18933333333333333, 0.999375, 0.9358055555555556], "iscrowd": 0, "area": 57555.883050000004, "rle": {"size": [360, 640], "counts": "PeR24m:7N3N1O2M2O1O2N1O2N1N3O0O2N1O2N00000001N1000000000000O2O000000000O10001O00000O101O001O001O001O010O001O001O001O001OB_OVFb0f9A[F?b9e0L4M3L3M4M3L4L4lGSNj6P2RIUNk6l1RIYNm6g1QI\\Nm6e1RI]Nm6c1RI`Nk6a1TIaNj6_1VIcNi6]1UIfNi6[1VIgNi6X1TInNi6S1QIROo6o0mHUOS7k0lHUOT7k0mHUOR7l0mHTO0bNl6\\2SISO0cNk6Z2UISO0eNh6Z2WIRO0gNf6X2YIQO1iNc6W2\\IQO0jNb6V2^IoN1kNa6V2]IoN2mN`6T2^ImN3QO^6Q2`IlN4TO[6P2aIjN6WOX6o1bIgN9ZOT6]1hI\\NLi0<_Oo5\\1WJPOOEi5\\1WJkN4Jd5[1XJkN4Jc5\\1YJiN4Mb5Z1ZJhN5O`5Z1YJhN7O^5Z1[JfN80]5Z1[JeN92Z5[1\\JbN:5Y5Y1]JbN:6X5X1^JaN;7V5Y1_J_N;:U5X1_J]N= in this image.", "objects": [{"patches": [24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 52, 53, 54, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 115, 116], "bbox": [0.00228125, 0.10627777777777778, 0.373484375, 0.3916944444444444], "iscrowd": 0, "area": 16318.962049999996, "rle": {"size": [360, 640], "counts": "]?4o9U1^Ob0O2O0O2N100O2N2O0O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O01000O01000O0100O10O10O10O10O10O10O100O01000O01000002M2O1O1N2O1O1N0100000O10000O10000O10O1000O100000000001ClGdMT8\\2lGdMT8[2mGdMT8\\2lGdMT8\\2lGdMT8[2mGeMS8[2mGeMS8[2mGeMS8Z2mGgMS8Y2mGgMS8Y2mGgMS8X2nGhMR8X2nGhMR8W2oGiMQ8W2oGhMS8W2mGiMS8W2mGiMS8W2mGiMS8X2lGhMT8X2lGhMT8X2lGhMT8X2lGhMT8X2lGhMT8X2lGhMT8X2kGiMU8W2kGhMV8X2jGhMV8X2jGhMV8X2jGhMV8X2jGgMW8Y2iGgMW8Y2iGfMX8Z2hGeMY8[2gGdMZ8\\2fGcM[8]2fGbMZ8^2fGaM[8_2eG`M\\8Y2kGfMV8X2lGhMT8X2>1O000000000O100000000000002N2N2N101N2O1O001N0100O01000O10O0100O10O10O1O1O001O2N2N2N2N2N2N2N2XOh0J4L4K5L4L4M300001O000000001O0000010O000000001O0000001OO100O1O1O100O101O00000000000000001O0000001O000000001O00000Nll\\4"}, "label": "golden dog in back"}]} {"id": 149180, "image": "COCO_train2014_000000149180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smaller dog with eyes open\"."}], "task": "refering", "answer_template": "The \"a smaller dog with eyes open\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 52, 53, 54, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 115, 116], "bbox": [0.00228125, 0.10627777777777778, 0.373484375, 0.3916944444444444], "iscrowd": 0, "area": 16318.962049999996, "rle": {"size": [360, 640], "counts": "]?4o9U1^Ob0O2O0O2N100O2N2O0O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O01000O01000O0100O10O10O10O10O10O10O100O01000O01000002M2O1O1N2O1O1N0100000O10000O10000O10O1000O100000000001ClGdMT8\\2lGdMT8[2mGdMT8\\2lGdMT8\\2lGdMT8[2mGeMS8[2mGeMS8[2mGeMS8Z2mGgMS8Y2mGgMS8Y2mGgMS8X2nGhMR8X2nGhMR8W2oGiMQ8W2oGhMS8W2mGiMS8W2mGiMS8W2mGiMS8X2lGhMT8X2lGhMT8X2lGhMT8X2lGhMT8X2lGhMT8X2lGhMT8X2kGiMU8W2kGhMV8X2jGhMV8X2jGhMV8X2jGhMV8X2jGgMW8Y2iGgMW8Y2iGfMX8Z2hGeMY8[2gGdMZ8\\2fGcM[8]2fGbMZ8^2fGaM[8_2eG`M\\8Y2kGfMV8X2lGhMT8X2>1O000000000O100000000000002N2N2N101N2O1O001N0100O01000O10O0100O10O10O1O1O001O2N2N2N2N2N2N2N2XOh0J4L4K5L4L4M300001O000000001O0000010O000000001O0000001OO100O1O1O100O101O00000000000000001O0000001O000000001O00000Nll\\4"}, "label": "a smaller dog with eyes open"}]} {"id": 100034, "image": "COCO_train2014_000000100034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl in the very back with the yellow and white shirt wearing a navy hat\"."}], "task": "refering", "answer_template": "The \"girl in the very back with the yellow and white shirt wearing a navy hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 131, 132, 154, 155, 176, 177, 178, 199, 200, 201, 223, 224, 245, 246, 247, 268, 269, 270, 291, 292, 314, 317, 340, 363, 364, 365, 387, 388], "bbox": [0.6451875, 0.22157643312101913, 0.8854218750000001, 0.7430732484076433], "iscrowd": 0, "area": 10670.742199999997, "rle": {"size": [628, 640], "counts": "icm72`c04L3M4K4M3M4PL_OgDd0X;D^D`0`;A]Db0b;_O[De0c;\\O[Df0d;\\OYDg0e;ZOYDj0d;WOZDl0e;TOYDo0e;ROZDQ1c;PO[DS1d;mNZDV1e;jNYDZ1e;fNYD\\1h;cNVD\\1m;dNPD\\1X<`NeC`1_<^N^Cb1e<^NYCb1h<`NTC`1o<`NoB`1S=`NiBa1Y=aNbBa1\\=cN`B^1`=fNZB]1f=fNUB[1l=hNiA`1W>cNhAZ1hni?@[@=g?@\\@=g?@\\@=g?@\\@>^a0N3MoAHi85VG0j8MTG8k8FTG>m=5K5K5K5K5N2O11O1O2N1O2N2N1M4M2N3M2M4M3M2M4M1O00O0j\\OLmb0>O1O2O0O1OO20O3M4M0O1O01O000010O0001O00001O002O1N1O2N1O2N1O2N2N1O2N1O2O1N1KZZ\\1"}, "label": "girl in the very back with the yellow and white shirt wearing a navy hat"}]} {"id": 100034, "image": "COCO_train2014_000000100034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two kids to the right\"."}], "task": "refering", "answer_template": "The \"two kids to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 131, 132, 154, 155, 176, 177, 178, 199, 200, 201, 223, 224, 245, 246, 247, 268, 269, 270, 291, 292, 314, 317, 340, 363, 364, 365, 387, 388], "bbox": [0.6451875, 0.22157643312101913, 0.8854218750000001, 0.7430732484076433], "iscrowd": 0, "area": 10670.742199999997, "rle": {"size": [628, 640], "counts": "icm72`c04L3M4K4M3M4PL_OgDd0X;D^D`0`;A]Db0b;_O[De0c;\\O[Df0d;\\OYDg0e;ZOYDj0d;WOZDl0e;TOYDo0e;ROZDQ1c;PO[DS1d;mNZDV1e;jNYDZ1e;fNYD\\1h;cNVD\\1m;dNPD\\1X<`NeC`1_<^N^Cb1e<^NYCb1h<`NTC`1o<`NoB`1S=`NiBa1Y=aNbBa1\\=cN`B^1`=fNZB]1f=fNUB[1l=hNiA`1W>cNhAZ1hni?@[@=g?@\\@=g?@\\@=g?@\\@>^a0N3MoAHi85VG0j8MTG8k8FTG>m=5K5K5K5K5N2O11O1O2N1O2N2N1M4M2N3M2M4M3M2M4M1O00O0j\\OLmb0>O1O2O0O1OO20O3M4M0O1O01O000010O0001O00001O002O1N1O2N1O2N1O2N2N1O2N1O2O1N1KZZ\\1"}, "label": "two kids to the right"}]} {"id": 100034, "image": "COCO_train2014_000000100034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"child on the right wearing a blue shirt\"."}], "task": "refering", "answer_template": "The \"child on the right wearing a blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 155, 156, 157, 178, 179, 180, 201, 202, 203, 204, 224, 225, 226, 247, 248, 249, 270, 271, 272, 292, 293, 294, 295, 315, 316, 317, 318, 319, 338, 339, 340, 341, 342, 361, 362, 364, 365, 383, 384, 388, 389, 406, 407, 411, 412, 413], "bbox": [0.672546875, 0.24875796178343948, 0.9658125, 0.8128025477707007], "iscrowd": 0, "area": 24474.06510000001, "rle": {"size": [628, 640], "counts": "dUX81ac0:G=B5K5K2N2N3M2N2N2N2O1N2N2N2K5H8H8H8H8H8M3N2N2N3M2M3N2NkNn@lMP?T2W10000O10000000000OJ7D in this image.", "objects": [{"patches": [132, 133, 134, 155, 156, 157, 178, 179, 180, 201, 202, 203, 204, 224, 225, 226, 247, 248, 249, 270, 271, 272, 292, 293, 294, 295, 315, 316, 317, 318, 319, 338, 339, 340, 341, 342, 361, 362, 364, 365, 383, 384, 388, 389, 406, 407, 411, 412, 413], "bbox": [0.672546875, 0.24875796178343948, 0.9658125, 0.8128025477707007], "iscrowd": 0, "area": 24474.06510000001, "rle": {"size": [628, 640], "counts": "dUX81ac0:G=B5K5K2N2N3M2N2N2N2O1N2N2N2K5H8H8H8H8H8M3N2N2N3M2M3N2NkNn@lMP?T2W10000O10000000000OJ7D in this image.", "objects": [{"patches": [107, 108, 109, 121, 130, 131, 144, 145, 150, 151, 152, 153, 154, 167, 168, 173, 174, 175, 176, 196, 197, 198, 199, 200, 223], "bbox": [0.27846875, 0.21170382165605095, 0.766875, 0.43468152866242044], "iscrowd": 0, "area": 8300.172100000002, "rle": {"size": [628, 640], "counts": "\\Z]3j0gb0?A7H7J501N000000O101O1N3N2N1N3N2N1N3N5K7H9H4M100O1L4L4M4KRd\\21m[cM2N3N2M3N2M4M2M4L4M3L3N3L101N2O1N2N2O0O2O1N2O1N2O0O2O10O000001O01O0000001O000000010O000000001O000000010O000000001O0001O01O1O001O10O01O1O1O010O1O001O1O010O1O001O1O00dNhN]@X1a?kN^@T1b?oN\\@P1c?TOZ@l0e?XOY@g0f?\\OY@c0f?AU@a0j?BS@?m?Co_O?P`0Dk_O?T`0Dg_OBF7b`0;b_O@O3^`0Y1e_OfNY`0Y1k_OeNT`0[1o_OdNP`0Z1T@dNk?[1Y@dNe?[1_@cN`?\\1d@cNZ?\\1j@bNU?^1n@aNQ?]1SAaNl>^1WAbNg>_1[A_Nd>a1^A_N`>b1aA]N^>c1cA^N\\>b1eA^NY>b1iA_NU>a1kA`NT>n0Y@VOe1MP>k0^@WOc1Om=h0c@XO`12k=d0h@XO_13i=c0k@YO]14g=b0n@XO\\17e=?RAYOZ18c=OdAGj0:b=OeAFj0;`=OgADk0=^=NhADj0>^=MlC3S in this image.", "objects": [{"patches": [79, 80, 81, 102, 103, 104, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 168, 169, 170, 173, 174, 190, 191, 192, 193, 194, 196, 197, 198, 214, 216, 217, 220, 239, 240, 241, 263, 264, 286, 287, 309, 310, 332, 333], "bbox": [0.299890625, 0.1370859872611465, 0.617421875, 0.6629140127388536], "iscrowd": 0, "area": 18828.215699999997, "rle": {"size": [628, 640], "counts": "Toe39Yc0b0^O2N2N3N1N3N110O0010O01O01O01O010O010O0010O00010O010N10KjNl]OU1Tb0kNl]OU1Tb0kNk]OW1Tb0jNk]OV1Ub0jNj]OW1Vb0iNj]OX1Ub030001O00001O000O101O00001O00001O00001O2N4L3M3M3M3M4L3M3M3M3M3M4L3MoMn_Od0o?\\OS@e0j?[OX@f0e?YO^@h0_?XOc@h0[?XOg@h0W?XOk@i0R?VORAi0l>WOVAj0g>VO[Aj0c>UO`Ak0^>UOdAl0Y>TOiAl0U>SOnAm0P>SORBn0k=ROWBn0g=QO]Bo0`=QObBo0\\=QOfBo0X=ROiBo0T=QOnBo0P=ROQCX1m0mMh8k0]Fb1?dMS9k0_Fc19eMV9h0cFd14eMX9h0fFc1OfMZ9g0iFd1IgM]9f0kFd1EgM^9f0oFd1@hM_9d0SGf1YOhMc9c0UGf1UOhMe9b0XGg1POhMg9b0ZGg1lNiMh9`0^Gh1fNjMk9?aGg1]NoMQ::dGh1TNSNW:6fGi1kMVN^:1iGj1aM\\Nd:KlGj1YM`Nj:FPH\\3o7fLQHY3n7hLSHV3m7lLSHR3m6oMTIo1j6TNXIi1f6[NZIc1c6aN^I]1`6fNaIX1]6lNdIQ1\\6POeIn0\\6ROeIl0\\6UOdIi0]6WOdIg0]6YOeId0\\6]OdIa0]6_OfIYOkLbN`9U2jIROTM\\NS9c2mIkN]MUNg8P3PJfNV7Z1oH`NR7a1QIYNQ7g1VIQNk6o1[IjMf6W2`I`Mb6`2dIYM]6g2jIQMW6P3V500001O001O101O0O2O0O2O0O2O1N101O0O2O1NRMg@b1e`02N103L4L4M3L4L4M3L4L4M3L4M1N1O10O01O100O001001O2N2M3N2N2N2M3N2M3L4L5K5K6J6J5K6J6J6J6J7I7I6J7I8HZPf4"}, "label": "the man with the white shirt in the back holding a tennis racket"}]} {"id": 100034, "image": "COCO_train2014_000000100034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy in the back wearing a white shirt\"."}], "task": "refering", "answer_template": "The \"the boy in the back wearing a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 102, 103, 104, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 168, 169, 170, 173, 174, 190, 191, 192, 193, 194, 196, 197, 198, 214, 216, 217, 220, 239, 240, 241, 263, 264, 286, 287, 309, 310, 332, 333], "bbox": [0.299890625, 0.1370859872611465, 0.617421875, 0.6629140127388536], "iscrowd": 0, "area": 18828.215699999997, "rle": {"size": [628, 640], "counts": "Toe39Yc0b0^O2N2N3N1N3N110O0010O01O01O01O010O010O0010O00010O010N10KjNl]OU1Tb0kNl]OU1Tb0kNk]OW1Tb0jNk]OV1Ub0jNj]OW1Vb0iNj]OX1Ub030001O00001O000O101O00001O00001O00001O2N4L3M3M3M3M4L3M3M3M3M3M4L3MoMn_Od0o?\\OS@e0j?[OX@f0e?YO^@h0_?XOc@h0[?XOg@h0W?XOk@i0R?VORAi0l>WOVAj0g>VO[Aj0c>UO`Ak0^>UOdAl0Y>TOiAl0U>SOnAm0P>SORBn0k=ROWBn0g=QO]Bo0`=QObBo0\\=QOfBo0X=ROiBo0T=QOnBo0P=ROQCX1m0mMh8k0]Fb1?dMS9k0_Fc19eMV9h0cFd14eMX9h0fFc1OfMZ9g0iFd1IgM]9f0kFd1EgM^9f0oFd1@hM_9d0SGf1YOhMc9c0UGf1UOhMe9b0XGg1POhMg9b0ZGg1lNiMh9`0^Gh1fNjMk9?aGg1]NoMQ::dGh1TNSNW:6fGi1kMVN^:1iGj1aM\\Nd:KlGj1YM`Nj:FPH\\3o7fLQHY3n7hLSHV3m7lLSHR3m6oMTIo1j6TNXIi1f6[NZIc1c6aN^I]1`6fNaIX1]6lNdIQ1\\6POeIn0\\6ROeIl0\\6UOdIi0]6WOdIg0]6YOeId0\\6]OdIa0]6_OfIYOkLbN`9U2jIROTM\\NS9c2mIkN]MUNg8P3PJfNV7Z1oH`NR7a1QIYNQ7g1VIQNk6o1[IjMf6W2`I`Mb6`2dIYM]6g2jIQMW6P3V500001O001O101O0O2O0O2O0O2O1N101O0O2O1NRMg@b1e`02N103L4L4M3L4L4M3L4L4M3L4M1N1O10O01O100O001001O2N2M3N2N2N2M3N2M3L4L5K5K6J6J5K6J6J6J6J7I7I6J7I8HZPf4"}, "label": "the boy in the back wearing a white shirt"}]} {"id": 100034, "image": "COCO_train2014_000000100034.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the closest girl\"."}], "task": "refering", "answer_template": "The \"the closest girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 170, 171, 172, 173, 193, 194, 195, 196, 218, 219, 220, 241, 242, 243, 244, 264, 265, 266, 267, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 333, 334, 335, 336, 337, 357, 358, 359, 380, 381, 382, 403, 404, 405, 406, 426, 427, 428, 429, 449, 450, 451, 452, 471, 472, 473, 474, 494, 495], "bbox": [0.406640625, 0.2871656050955414, 0.6961562499999999, 0.9583280254777071], "iscrowd": 0, "area": 32289.155400000003, "rle": {"size": [628, 640], "counts": "Teo41_c07J5L5L3N3M1O1N2O1O1N2N2N1O2M3N2N3M2N2N2M4M2N2N2N3M2N2O1O0O2O1O1O1O1O1O1O000O11O000001fCfMX7[2dHhM]7X2_HlM`7U2\\HnMd7S2YHPNf7R2WHPNh7Q2VHPNk7P2THQNk7P2SHQNm7P2QHRNn7o1PHRNP8o1eDbMY2`0Q9R3kFoLV9S3fFoLY9U3bFlL^9X3^FiLa9^3WFcLi9c3PF^LQ:e3jE\\Lb3oNW1f4UK[Lc3POY1f4QK[Le3PO[1f4mJ[Lg3PO]1f4jJZLh3QO_1f4fJZLi3ROb1e4bJZLk3ROd1e4_JYLl3SOf1d4\\JZLm3SOj1b4WJ[Ln3TOm1`4SJ\\LQ4TOm1`4PJ\\L5kNf2:V3_4nI[L5^OV2Ii3^4iI[L7Hn1@S4]4fIZL9Nk1[OW4]4dIYL95g1VO]4]4`IXL: in this image.", "objects": [{"patches": [148, 149, 170, 171, 172, 173, 193, 194, 195, 196, 218, 219, 220, 241, 242, 243, 244, 264, 265, 266, 267, 287, 288, 289, 290, 291, 310, 311, 312, 313, 314, 333, 334, 335, 336, 337, 357, 358, 359, 380, 381, 382, 403, 404, 405, 406, 426, 427, 428, 429, 449, 450, 451, 452, 471, 472, 473, 474, 494, 495], "bbox": [0.406640625, 0.2871656050955414, 0.6961562499999999, 0.9583280254777071], "iscrowd": 0, "area": 32289.155400000003, "rle": {"size": [628, 640], "counts": "Teo41_c07J5L5L3N3M1O1N2O1O1N2N2N1O2M3N2N3M2N2N2M4M2N2N2N3M2N2O1O0O2O1O1O1O1O1O1O000O11O000001fCfMX7[2dHhM]7X2_HlM`7U2\\HnMd7S2YHPNf7R2WHPNh7Q2VHPNk7P2THQNk7P2SHQNm7P2QHRNn7o1PHRNP8o1eDbMY2`0Q9R3kFoLV9S3fFoLY9U3bFlL^9X3^FiLa9^3WFcLi9c3PF^LQ:e3jE\\Lb3oNW1f4UK[Lc3POY1f4QK[Le3PO[1f4mJ[Lg3PO]1f4jJZLh3QO_1f4fJZLi3ROb1e4bJZLk3ROd1e4_JYLl3SOf1d4\\JZLm3SOj1b4WJ[Ln3TOm1`4SJ\\LQ4TOm1`4PJ\\L5kNf2:V3_4nI[L5^OV2Ii3^4iI[L7Hn1@S4]4fIZL9Nk1[OW4]4dIYL95g1VO]4]4`IXL: in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 58, 76, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 104, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 169, 170], "bbox": [0.32014062499999996, 0.11175, 0.5594531249999999, 0.4205416666666667], "iscrowd": 0, "area": 18072.939950000007, "rle": {"size": [480, 640], "counts": "[WP37c>9G8H9G9G9G9G9F:G9G9G9K5M3M3M3M1N2O001O001N10001O001O0O2O001O001N101O00001O0O2O00O1000O100000O1000O100000O10O10000000O10O10000000O01000000000O0100000000O01000000000O010000000O01000000000O0100000000O010000000O101O1O001O1N2O001O1O0O2O1O1O001N2O1O001O1O0O2O1O1N1O2N2N2N1O2N2N101N2N2N1O2N2N4L:F:F;E:F:G9F:F;ERQT4"}, "label": "a white cake with blue , green , and yellow poke - a - dots on it next to three other decorated cakes on a platter"}]} {"id": 149202, "image": "COCO_train2014_000000149202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cake with the yellow circles\"."}], "task": "refering", "answer_template": "The \"the cake with the yellow circles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 58, 76, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 104, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 169, 170], "bbox": [0.32014062499999996, 0.11175, 0.5594531249999999, 0.4205416666666667], "iscrowd": 0, "area": 18072.939950000007, "rle": {"size": [480, 640], "counts": "[WP37c>9G8H9G9G9G9G9F:G9G9G9K5M3M3M3M1N2O001O001N10001O001O0O2O001O001N101O00001O0O2O00O1000O100000O1000O100000O10O10000000O10O10000000O01000000000O0100000000O01000000000O010000000O01000000000O0100000000O010000000O101O1O001O1N2O001O1O0O2O1O1O001N2O1O001O1O0O2O1O1N1O2N2N2N1O2N2N101N2N2N1O2N2N4L:F:F;E:F:G9F:F;ERQT4"}, "label": "the cake with the yellow circles"}]} {"id": 149202, "image": "COCO_train2014_000000149202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"mini cake with a black bow on top and black and blue stripes around it\"."}], "task": "refering", "answer_template": "The \"mini cake with a black bow on top and black and blue stripes around it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237], "bbox": [0.10570312500000001, 0.3178958333333333, 0.365765625, 0.6445208333333333], "iscrowd": 0, "area": 22004.4455, "rle": {"size": [480, 640], "counts": "jQP1?_>a0@`0_Oa0@a0_O`0_O:G2M2O2M3N1O2M3N2M2O2N2M3N1N3N2M201O1O1O001O1O0O2O1O1O00001O0000001O0000001O000000001O0000001O0000000000000000000000O10000000000000000O1000000000000000000O10000000000000000O100000000000001O000O1000001O000000000O101O00000000001O000O1000001O00000000000O2O00000000001O000O10004L3M3M3M3H8E;F;E:F:G9K5K5K5L5J5K5K5K5K5K5K5K6JiRn5"}, "label": "mini cake with a black bow on top and black and blue stripes around it"}]} {"id": 149202, "image": "COCO_train2014_000000149202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"mini cale with brown bow an blue stripes\"."}], "task": "refering", "answer_template": "The \"mini cale with brown bow an blue stripes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237], "bbox": [0.10570312500000001, 0.3178958333333333, 0.365765625, 0.6445208333333333], "iscrowd": 0, "area": 22004.4455, "rle": {"size": [480, 640], "counts": "jQP1?_>a0@`0_Oa0@a0_O`0_O:G2M2O2M3N1O2M3N2M2O2N2M3N1N3N2M201O1O1O001O1O0O2O1O1O00001O0000001O0000001O000000001O0000001O0000000000000000000000O10000000000000000O1000000000000000000O10000000000000000O100000000000001O000O1000001O000000000O101O00000000001O000O1000001O00000000000O2O00000000001O000O10004L3M3M3M3H8E;F;E:F:G9K5K5K5L5J5K5K5K5K5K5K5K6JiRn5"}, "label": "mini cale with brown bow an blue stripes"}]} {"id": 149202, "image": "COCO_train2014_000000149202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cupcake that looks like a hat with a bow in front\"."}], "task": "refering", "answer_template": "The \"a cupcake that looks like a hat with a bow in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 247], "bbox": [0.550671875, 0.25, 0.8445937499999999, 0.5946041666666666], "iscrowd": 0, "area": 23742.54475, "rle": {"size": [480, 640], "counts": "feU54i>3M3M3M4L3N2M5L3L4M2N3M3N2M3N2N1L5J6WOi0M3L4M3L3N3O1O1O1N2O1O1O1O2N1O5J7J2N2N2N2N2N2N0O2O0000001O000O101O00001O001O0O101O001O00001N10001O001O000O2O00000000001O0O1000000000000O101O00000000000O10001O0000000O10000000001O0O100000000000001N1000000001O00001O010O00001O001O00001O0O2O000O2O0O101N101N100O2O0O2O0O101N102M2O1N2O1O1N3N1N2O1N2O2M3N2M3N3L3N2M3N2M3L5L3M8G in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 247], "bbox": [0.550671875, 0.25, 0.8445937499999999, 0.5946041666666666], "iscrowd": 0, "area": 23742.54475, "rle": {"size": [480, 640], "counts": "feU54i>3M3M3M4L3N2M5L3L4M2N3M3N2M3N2N1L5J6WOi0M3L4M3L3N3O1O1O1N2O1O1O1O2N1O5J7J2N2N2N2N2N2N0O2O0000001O000O101O00001O001O0O101O001O00001N10001O001O000O2O00000000001O0O1000000000000O101O00000000000O10001O0000000O10000000001O0O100000000000001N1000000001O00001O010O00001O001O00001O0O2O000O2O0O101N101N100O2O0O2O0O101N102M2O1N2O1O1N3N1N2O1N2O2M3N2M3N3L3N2M3N2M3L5L3M8G in this image.", "objects": [{"patches": [170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 263, 264, 265, 266, 267, 283, 284, 285, 286, 287, 288, 289, 290, 307, 308, 309, 310, 311, 312, 313, 332, 333, 334, 335], "bbox": [0.32359375, 0.43370833333333336, 0.6539375, 0.8449375], "iscrowd": 0, "area": 32279.3765, "rle": {"size": [480, 640], "counts": "mZQ36h>f0ZO`0A2M4L3N2M3M3N3L3N2J6I8G8I7H8I7H9H7M3N2M3N3M1O1O2M2O1O2N1O1N2O2N1O1M4O0O101N100O101N100O2O0O100O2O0O101N100O101O0O101O00000O2O00001O0O10001O000O2O000000000O101O00000O100000000O101O0000000O10000001O0001O00000001O0000000001O0001O000000000000010O000000000000001O0001O00000001O0000000001O0001O001O001O001O10OO2O1N1O2N2O0O2N1O2O1N1O2N2O0O2N1O2O1N2N3M3N2M3M3M3N2M3M3M3N2M3M3M3N2M4K6J7I6J6J6J7J5J6J7I6J4L5K4L4L4L5K4L4L4L5K4LegW3"}, "label": "a small black and white cake with a ribbon on top"}]} {"id": 149202, "image": "COCO_train2014_000000149202.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pastry with a bow - tie and spiral dots\"."}], "task": "refering", "answer_template": "The \"a pastry with a bow - tie and spiral dots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 263, 264, 265, 266, 267, 283, 284, 285, 286, 287, 288, 289, 290, 307, 308, 309, 310, 311, 312, 313, 332, 333, 334, 335], "bbox": [0.32359375, 0.43370833333333336, 0.6539375, 0.8449375], "iscrowd": 0, "area": 32279.3765, "rle": {"size": [480, 640], "counts": "mZQ36h>f0ZO`0A2M4L3N2M3M3N3L3N2J6I8G8I7H8I7H9H7M3N2M3N3M1O1O2M2O1O2N1O1N2O2N1O1M4O0O101N100O101N100O2O0O100O2O0O101N100O101O0O101O00000O2O00001O0O10001O000O2O000000000O101O00000O100000000O101O0000000O10000001O0001O00000001O0000000001O0001O000000000000010O000000000000001O0001O00000001O0000000001O0001O001O001O001O10OO2O1N1O2N2O0O2N1O2O1N1O2N2O0O2N1O2O1N2N3M3N2M3M3M3N2M3M3M3N2M3M3M3N2M4K6J7I6J6J6J7J5J6J7I6J4L5K4L4L4L5K4L4L4L5K4LegW3"}, "label": "a pastry with a bow - tie and spiral dots"}]} {"id": 558804, "image": "COCO_train2014_000000558804.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red sox player with both arms crossed in front of him and a bat pointing away from his body\"."}], "task": "refering", "answer_template": "The \"red sox player with both arms crossed in front of him and a bat pointing away from his body\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 59, 60, 61, 62, 82, 83, 84, 85, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 294, 311, 312, 313, 314, 315, 316, 317, 334, 335, 336, 337, 338, 339, 340, 358, 359, 360, 361, 362, 363], "bbox": [0.51284375, 0.06419028340080972, 0.791, 0.8930161943319838], "iscrowd": 0, "area": 54844.08935, "rle": {"size": [494, 640], "counts": "m\\n4W1k=j0_Of0YO5K5K5K5L3L:F in this image.", "objects": [{"patches": [36, 37, 38, 59, 60, 61, 62, 82, 83, 84, 85, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 294, 311, 312, 313, 314, 315, 316, 317, 334, 335, 336, 337, 338, 339, 340, 358, 359, 360, 361, 362, 363], "bbox": [0.51284375, 0.06419028340080972, 0.791, 0.8930161943319838], "iscrowd": 0, "area": 54844.08935, "rle": {"size": [494, 640], "counts": "m\\n4W1k=j0_Of0YO5K5K5K5L3L:F in this image.", "objects": [{"patches": [55, 56, 77, 78, 79, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 142, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 242, 259, 260, 261, 262, 263, 264, 265, 282, 283, 284, 286, 287, 288, 305, 306, 307, 308, 309, 310, 311, 328, 329, 330, 331, 332, 333, 334, 352, 353, 354, 356, 357, 375, 376, 379, 380], "bbox": [0.18660937500000002, 0.13625506072874494, 0.5564218750000001, 0.9142914979757085], "iscrowd": 0, "area": 49899.849650000004, "rle": {"size": [494, 640], "counts": "bai11]?4L4K5L4K5L2N1N3N1N2O1O2M2O1N2O2N1N2O2M2O1O2M3N1N3N2N1N3N2M2O2N2O0O2N2N1O2N2N1O2N2N2N1O2N2N2N2N2N2N2lEmLj7T3iF[NS9g1hFaNS9a1gFhNT9Z1gFmNU9U1fFSOU9h3K5K5K5K4L2N2N2N2N2N2N2N2N2O2M4L4L4L4L4L2N3M3M2N3M3M3M3M3M3M4L3N3L3M1O1O001O0O2O000O2O001N1000000O10000O01000O1000000O1mNYF_Lg9`3^F\\La9c3eFXL\\9g3jFTLV9j3PGRLP9m3UGnKl8P4ZGlKf8S4[1N2O100O001O]LoLkJP3S5WMiJi2T5^MhJa2W5dMeJ\\2Y5kMcJT2\\5SN^Jn1_5[NZJe1e5dNcIm1^6a30O101O0O10001O0O10001N10001O00000O2O0000001O00000100O1O10O01O11O2N2M4M2N4L3M3L5L3M3M3L5L5M4L4L4K6K5KhF]Jf8S61O1O100O1O1O1O100O000002N2O0O2N1O2iIkIP3V6jLoIU3T6dLQJ[3Q6_LTJa3Y6mKmIQ4g6VK_Ij4P7bJUI]5a8N2M3N2M3N5J5K6J6J=CY1RMPDLc in this image.", "objects": [{"patches": [55, 56, 77, 78, 79, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 142, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 242, 259, 260, 261, 262, 263, 264, 265, 282, 283, 284, 286, 287, 288, 305, 306, 307, 308, 309, 310, 311, 328, 329, 330, 331, 332, 333, 334, 352, 353, 354, 356, 357, 375, 376, 379, 380], "bbox": [0.18660937500000002, 0.13625506072874494, 0.5564218750000001, 0.9142914979757085], "iscrowd": 0, "area": 49899.849650000004, "rle": {"size": [494, 640], "counts": "bai11]?4L4K5L4K5L2N1N3N1N2O1O2M2O1N2O2N1N2O2M2O1O2M3N1N3N2N1N3N2M2O2N2O0O2N2N1O2N2N1O2N2N2N1O2N2N2N2N2N2N2lEmLj7T3iF[NS9g1hFaNS9a1gFhNT9Z1gFmNU9U1fFSOU9h3K5K5K5K4L2N2N2N2N2N2N2N2N2O2M4L4L4L4L4L2N3M3M2N3M3M3M3M3M3M4L3N3L3M1O1O001O0O2O000O2O001N1000000O10000O01000O1000000O1mNYF_Lg9`3^F\\La9c3eFXL\\9g3jFTLV9j3PGRLP9m3UGnKl8P4ZGlKf8S4[1N2O100O001O]LoLkJP3S5WMiJi2T5^MhJa2W5dMeJ\\2Y5kMcJT2\\5SN^Jn1_5[NZJe1e5dNcIm1^6a30O101O0O10001O0O10001N10001O00000O2O0000001O00000100O1O10O01O11O2N2M4M2N4L3M3L5L3M3M3L5L5M4L4L4K6K5KhF]Jf8S61O1O100O1O1O1O100O000002N2O0O2N1O2iIkIP3V6jLoIU3T6dLQJ[3Q6_LTJa3Y6mKmIQ4g6VK_Ij4P7bJUI]5a8N2M3N2M3N5J5K6J6J=CY1RMPDLc in this image.", "objects": [{"patches": [27, 28, 49, 50, 51, 52, 72, 73, 74, 75, 96, 97, 98, 119, 120, 121, 142, 143, 144, 164, 165, 167, 168, 187, 188, 190, 191], "bbox": [0.147984375, 0.14223484848484846, 0.319234375, 0.9443181818181818], "iscrowd": 0, "area": 10341.964, "rle": {"size": [264, 640], "counts": "\\bh03T84^LKjN9Q1JlN:o0IQO8k0LTO4j0NUO4h0OWO1i0OXO1f01YOOf02ZOOd03[OMd05[OLc05]OKb07]OJa07_OI`08@I>9AH=9CG<:DG:;FD9=HC6>JC3`0M_O1c00]OLf04ZOIj07VOEm0 in this image.", "objects": [{"patches": [27, 28, 49, 50, 51, 52, 72, 73, 74, 75, 96, 97, 98, 119, 120, 121, 142, 143, 144, 164, 165, 167, 168, 187, 188, 190, 191], "bbox": [0.147984375, 0.14223484848484846, 0.319234375, 0.9443181818181818], "iscrowd": 0, "area": 10341.964, "rle": {"size": [264, 640], "counts": "\\bh03T84^LKjN9Q1JlN:o0IQO8k0LTO4j0NUO4h0OWO1i0OXO1f01YOOf02ZOOd03[OMd05[OLc05]OKb07]OJa07_OI`08@I>9AH=9CG<:DG:;FD9=HC6>JC3`0M_O1c00]OLf04ZOIj07VOEm0 in this image.", "objects": [{"patches": [12, 13, 14, 35, 36, 37, 59, 60, 61, 82, 83, 84, 105, 106, 107, 128, 129, 151, 152, 153, 175, 176, 199], "bbox": [0.5475156250000001, 0.04231060606060606, 0.675078125, 0.904621212121212], "iscrowd": 0, "area": 9178.226649999997, "rle": {"size": [264, 640], "counts": "Xaj24P84L4L5K4L4M3M3O100O10000O2O0O100O100SLlNe0U1XOTOb0l0[O\\O?e0[MkNW1i0X1<^MPOU1l0W15aMSOT1Q1U1MdMVOR1U1U1EfMZOR1W1S1@iM]OP1X1T1[OiMAP1Y1T1VOjMEn0Z1U1QOjMIo0Z1T1mNjMNn0Z1U12hNMZ16bNJ`17\\NJf18TNHP29jMF\\2<^MBh2i26J6J6N22M3N3M2dNmJ_OU5>lJCU5:lJFV57lJHW54jJMW51iJOY5NhJ2Z5KgJ6Z5GgJ9\\5CeJ?[5^OfJe0Y5YOhJi0W5TOjJn0W5nNjJU1U5hNlJZ1T5cNmJ`1R5]NoJe1Q5XNPKk1P5QNRKQ2a53M3ZMYNeNj1Q1`NlNc1l0eNQO^1h0jNTOY1c0POZOS19ZOEg00EM>GN65^O7?LTOa0i0BkNj0R1[3K5K5L4K6Il\\e1"}, "label": "a person with yellow boots"}]} {"id": 337621, "image": "COCO_train2014_000000337621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skier wearing a bright purple and yellow - green suit along with a fuzzy wig\"."}], "task": "refering", "answer_template": "The \"skier wearing a bright purple and yellow - green suit along with a fuzzy wig\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 35, 36, 37, 59, 60, 61, 82, 83, 84, 105, 106, 107, 128, 129, 151, 152, 153, 175, 176, 199], "bbox": [0.5475156250000001, 0.04231060606060606, 0.675078125, 0.904621212121212], "iscrowd": 0, "area": 9178.226649999997, "rle": {"size": [264, 640], "counts": "Xaj24P84L4L5K4L4M3M3O100O10000O2O0O100O100SLlNe0U1XOTOb0l0[O\\O?e0[MkNW1i0X1<^MPOU1l0W15aMSOT1Q1U1MdMVOR1U1U1EfMZOR1W1S1@iM]OP1X1T1[OiMAP1Y1T1VOjMEn0Z1U1QOjMIo0Z1T1mNjMNn0Z1U12hNMZ16bNJ`17\\NJf18TNHP29jMF\\2<^MBh2i26J6J6N22M3N3M2dNmJ_OU5>lJCU5:lJFV57lJHW54jJMW51iJOY5NhJ2Z5KgJ6Z5GgJ9\\5CeJ?[5^OfJe0Y5YOhJi0W5TOjJn0W5nNjJU1U5hNlJZ1T5cNmJ`1R5]NoJe1Q5XNPKk1P5QNRKQ2a53M3ZMYNeNj1Q1`NlNc1l0eNQO^1h0jNTOY1c0POZOS19ZOEg00EM>GN65^O7?LTOa0i0BkNj0R1[3K5K5L4K6Il\\e1"}, "label": "skier wearing a bright purple and yellow - green suit along with a fuzzy wig"}]} {"id": 337621, "image": "COCO_train2014_000000337621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"child in orange\"."}], "task": "refering", "answer_template": "The \"child in orange\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 58, 80, 81, 82, 103, 104, 105, 127, 128, 150, 151, 172, 173, 174, 195, 196, 197], "bbox": [0.482046875, 0.16136363636363638, 0.612609375, 0.9162878787878788], "iscrowd": 0, "area": 6660.880649999998, "rle": {"size": [264, 640], "counts": "Zl_22S84M2O2O1O0O101O0000001WMDdM=[2FbM:^2GaM:^2GaM:^2GaM9_2H_M:^1CPM5_19`1CPM6]1:a1BQM5\\1:a1CSM4Y1<`1DUM1Z1<_1EWM0o0e0h1]OXM0c0o0S2TOXMOc0n0P1_NdNl0HHb0k0V1eNYNl0LEd0i0X1mNQNh01Cc0g0]1ROjMS1HZOk0?e1e2fMXM]2l3>N110000000000001O2N2N1N3N^M\\Kb1b4^N_Kc1_4\\NbKe1\\4[NeKg1Y4XNhKj11_Mg3f0YLm1IgMl38\\Lf2=iLX2?[Mj27nL^24\\Mo22PMd2N[MT3LPMm2IWMX3IPMS3EUM]3CPMn3d3[LPLR1Lb0R4]NWLl0Lc0l3aN^Lg0Kc0f3gNlL9Ck0`3lNZNP1e1PO\\NP1c1PO_No0_1ROcNm0\\1TOgNi0X1WOkNg0T1YOoNe0P1[OTOb0k0_OWO?h0A[O=d0C@:`0EH47M1H27P4O1O2N2O2M3M2N3Nijo1"}, "label": "child in orange"}]} {"id": 337621, "image": "COCO_train2014_000000337621.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child wearing a white and red striped hat and orange coat\"."}], "task": "refering", "answer_template": "The \"a child wearing a white and red striped hat and orange coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 58, 80, 81, 82, 103, 104, 105, 127, 128, 150, 151, 172, 173, 174, 195, 196, 197], "bbox": [0.482046875, 0.16136363636363638, 0.612609375, 0.9162878787878788], "iscrowd": 0, "area": 6660.880649999998, "rle": {"size": [264, 640], "counts": "Zl_22S84M2O2O1O0O101O0000001WMDdM=[2FbM:^2GaM:^2GaM:^2GaM9_2H_M:^1CPM5_19`1CPM6]1:a1BQM5\\1:a1CSM4Y1<`1DUM1Z1<_1EWM0o0e0h1]OXM0c0o0S2TOXMOc0n0P1_NdNl0HHb0k0V1eNYNl0LEd0i0X1mNQNh01Cc0g0]1ROjMS1HZOk0?e1e2fMXM]2l3>N110000000000001O2N2N1N3N^M\\Kb1b4^N_Kc1_4\\NbKe1\\4[NeKg1Y4XNhKj11_Mg3f0YLm1IgMl38\\Lf2=iLX2?[Mj27nL^24\\Mo22PMd2N[MT3LPMm2IWMX3IPMS3EUM]3CPMn3d3[LPLR1Lb0R4]NWLl0Lc0l3aN^Lg0Kc0f3gNlL9Ck0`3lNZNP1e1PO\\NP1c1PO_No0_1ROcNm0\\1TOgNi0X1WOkNg0T1YOoNe0P1[OTOb0k0_OWO?h0A[O=d0C@:`0EH47M1H27P4O1O2N2O2M3M2N3Nijo1"}, "label": "a child wearing a white and red striped hat and orange coat"}]} {"id": 558824, "image": "COCO_train2014_000000558824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in dark clothing holding an umbrella in the background\"."}], "task": "refering", "answer_template": "The \"a person in dark clothing holding an umbrella in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 220, 221, 222, 244, 245, 266, 267, 268, 289, 290, 291, 312, 313, 314, 337], "bbox": [0.57140625, 0.4353629976580796, 0.794796875, 0.945456674473068], "iscrowd": 0, "area": 11986.932299999997, "rle": {"size": [427, 640], "counts": "ejh42\\=2N2M3N2N3YOKnC7n;1mCNPSJHl5?hIGX6`0[IFd6a0oHFQ7a0bHE^7S301000000O1000O1000O100000001N1000001N1O100O101N1O100O2O0O100O1O2OnNSIfKl6Y4[IbKe6\\4cI^K\\6b4kIYKS6f4RJWKn5g4TJYKk5g4WJXKi5a3^K_La4k2WLUMh3j2YLVMf3j2\\LUMc3k2^LUMb3j2`LUM_3k2cLUM\\3j2eLVMZ3j2hLUMW3k2jLUMV3e2QM[MR3\\2SMdMn2U2YMiMj2P2[MoMg2j1`MUNb2d1cM[N_2^1hM`N[2Y1o4K6N1O2M4M2N2N3M2N3M00001O000O1000000000000000000O100000000000001O00000O100000000000000000000O100000001N1O1O1N2O1O1O^jf1"}, "label": "a person in dark clothing holding an umbrella in the background"}]} {"id": 558824, "image": "COCO_train2014_000000558824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing capris and holding an umbrella\"."}], "task": "refering", "answer_template": "The \"a woman wearing capris and holding an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 220, 221, 222, 244, 245, 266, 267, 268, 289, 290, 291, 312, 313, 314, 337], "bbox": [0.57140625, 0.4353629976580796, 0.794796875, 0.945456674473068], "iscrowd": 0, "area": 11986.932299999997, "rle": {"size": [427, 640], "counts": "ejh42\\=2N2M3N2N3YOKnC7n;1mCNPSJHl5?hIGX6`0[IFd6a0oHFQ7a0bHE^7S301000000O1000O1000O100000001N1000001N1O100O101N1O100O2O0O100O1O2OnNSIfKl6Y4[IbKe6\\4cI^K\\6b4kIYKS6f4RJWKn5g4TJYKk5g4WJXKi5a3^K_La4k2WLUMh3j2YLVMf3j2\\LUMc3k2^LUMb3j2`LUM_3k2cLUM\\3j2eLVMZ3j2hLUMW3k2jLUMV3e2QM[MR3\\2SMdMn2U2YMiMj2P2[MoMg2j1`MUNb2d1cM[N_2^1hM`N[2Y1o4K6N1O2M4M2N2N3M2N3M00001O000O1000000000000000000O100000000000001O00000O100000000000000000000O100000001N1O1O1N2O1O1O^jf1"}, "label": "a woman wearing capris and holding an umbrella"}]} {"id": 558824, "image": "COCO_train2014_000000558824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady with curly hair in ponyail under umbrella and looking to her left\"."}], "task": "refering", "answer_template": "The \"lady with curly hair in ponyail under umbrella and looking to her left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 143, 144, 145, 146, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 264, 280, 281, 282, 283, 284, 285, 286, 287, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331], "bbox": [0.172421875, 0.3928805620608899, 0.48726562500000004, 0.990632318501171], "iscrowd": 0, "area": 36352.880450000004, "rle": {"size": [427, 640], "counts": "nX^12i<`0@`0@`0@`0@`0D in this image.", "objects": [{"patches": [122, 143, 144, 145, 146, 166, 167, 168, 169, 170, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 264, 280, 281, 282, 283, 284, 285, 286, 287, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331], "bbox": [0.172421875, 0.3928805620608899, 0.48726562500000004, 0.990632318501171], "iscrowd": 0, "area": 36352.880450000004, "rle": {"size": [427, 640], "counts": "nX^12i<`0@`0@`0@`0@`0D in this image.", "objects": [{"patches": [25, 26, 43, 44, 45, 60, 61, 62, 63, 64, 77, 78, 79, 80, 81, 82, 95, 96, 97, 98, 99, 100, 113, 114, 115, 116, 117, 118, 131, 132, 133, 134, 135, 136, 149, 150, 151, 152, 153, 168, 169, 170, 171, 186, 187, 188, 189], "bbox": [0.31162, 0.09439252336448599, 0.59584, 0.9887538940809968], "iscrowd": 0, "area": 28531.7021, "rle": {"size": [321, 500], "counts": "YSa1:_9>B?A>A`0A>B>B:UOk0XOg0N3N2N1N3N00O010O010O10O010O3dLoJd1R5RNWKn1j4gM`KX2b4^MgKb2[4SMnKm2c5O100O10000O100O10000O10000O100O10000O100`NZIEf6:bI^O_6b0gIXOY6g0mISOT6m0QJnNo5R1VJiNj5V1a100O100O100O10000O100O11O0000001O0000001O0000001O0000005K7I7Ia0_O9G001O001O00001O001O001O001O001O00001K4A`0A>A`0K410O0100O0010O0100O01ROfKTLY4j3\\LcKe3Z4R1N2N3M2N3M2N3M3L9H9G9G8H9G8H9G9G in this image.", "objects": [{"patches": [25, 26, 43, 44, 45, 60, 61, 62, 63, 64, 77, 78, 79, 80, 81, 82, 95, 96, 97, 98, 99, 100, 113, 114, 115, 116, 117, 118, 131, 132, 133, 134, 135, 136, 149, 150, 151, 152, 153, 168, 169, 170, 171, 186, 187, 188, 189], "bbox": [0.31162, 0.09439252336448599, 0.59584, 0.9887538940809968], "iscrowd": 0, "area": 28531.7021, "rle": {"size": [321, 500], "counts": "YSa1:_9>B?A>A`0A>B>B:UOk0XOg0N3N2N1N3N00O010O010O10O010O3dLoJd1R5RNWKn1j4gM`KX2b4^MgKb2[4SMnKm2c5O100O10000O100O10000O10000O100O10000O100`NZIEf6:bI^O_6b0gIXOY6g0mISOT6m0QJnNo5R1VJiNj5V1a100O100O100O10000O100O11O0000001O0000001O0000001O0000005K7I7Ia0_O9G001O001O00001O001O001O001O001O00001K4A`0A>A`0K410O0100O0010O0100O01ROfKTLY4j3\\LcKe3Z4R1N2N3M2N3M2N3M3L9H9G9G8H9G8H9G9G in this image.", "objects": [{"patches": [31, 32, 33, 49, 50, 51, 66, 67, 68, 69, 70, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 192, 193, 194, 195], "bbox": [0.63334, 0.11012461059190032, 0.98248, 0.988753894080997], "iscrowd": 0, "area": 32249.98215, "rle": {"size": [321, 500], "counts": "_bS3`0Y9a0@ in this image.", "objects": [{"patches": [31, 32, 33, 49, 50, 51, 66, 67, 68, 69, 70, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 192, 193, 194, 195], "bbox": [0.63334, 0.11012461059190032, 0.98248, 0.988753894080997], "iscrowd": 0, "area": 32249.98215, "rle": {"size": [321, 500], "counts": "_bS3`0Y9a0@ in this image.", "objects": [{"patches": [38, 39, 56, 57, 72, 73, 74, 75, 76, 90, 91, 92, 93, 94, 108, 109, 110, 111, 112, 126, 127, 128, 129, 130, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 180, 181, 182, 183, 184], "bbox": [0.0028799999999999997, 0.17753894080996885, 0.26112, 0.9887538940809968], "iscrowd": 0, "area": 24279.581049999993, "rle": {"size": [321, 500], "counts": "Yb02l8a0]GLZ7o1I6M4L8I:E7I5K5K5L4K5K5K4L4M3L4M3O1eKTKa3l4ZLYKe3h4VL]Kj3c4QLbKo3^4lKgKS4T50000O10000O10000O10000O10000O10000O10000O10000O1O100O100O1nNSJoMn5m1XJQNh5l1]JQNd5l1`JSN`5j1cJUN^5h1fJSN^5j1eJoMb5o1Y100O100O100O1O100O100001O001O00001O00001O001O00001O00001O001^IYNh4g1SK`Nk4`1TKdNi4]1UKeNj4]1SKeNl4\\1RKeNn4]1oJeNP5\\1nJeNR5]1kJeNP5`1lJcNP5c1kJaNm4]3J70O01O001O001O0010O01O00001O0010ON3M2Na0_Oc0]Oc0]Oc0]Oc0]Oc0\\Od0]Ob0^OQ]c3"}, "label": "man on far left"}]} {"id": 501493, "image": "COCO_train2014_000000501493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a long white boat\"."}], "task": "refering", "answer_template": "The \"a long white boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156], "bbox": [0.20992187499999998, 0.53672, 0.9140625, 0.7189200000000001], "iscrowd": 0, "area": 8355.035249999997, "rle": {"size": [250, 640], "counts": "]SQ11h73N1N2O1N2O1N2O1O1001O001OO1O0010O01O1O001O010O001O1O001O001O1O000000001O000000000001O000001O0000000000000000001O0000000000000000000000000000000000O10000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000O101O00001O5K3MO10000000000000000O100000000000000L4O1O1O1O10000001O000000001O00004L1O0000000O01000000O100000000O1O1N200000000005K1O00000000000000000000O1N2O1O1O1N1100000000000000000000000000000000000008H3M0000O100000000000000O10000GKmH5S791O00000000001O0000001O0000001O0000001OO100O10000O10000O10000O10000O100000000O10000000000000000001O00000000001N100000001O00000000001O00000000000000001O00O1000O0100O10000O100O100000000000000000O10O100000000000000000000000000000O100000000000000O100000000000O010005K5K1O000O10001OO10000O01000O10000O10O10O1B>O1O1O1O001O1O1O1O1O1000O1000000000000O1000000O10000O100O100O10000O10O0100O10000O100O100O10000O2O0N2J6M4N1N2N`Y="}, "label": "a long white boat"}]} {"id": 501493, "image": "COCO_train2014_000000501493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boat\"."}], "task": "refering", "answer_template": "The \"boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156], "bbox": [0.20992187499999998, 0.53672, 0.9140625, 0.7189200000000001], "iscrowd": 0, "area": 8355.035249999997, "rle": {"size": [250, 640], "counts": "]SQ11h73N1N2O1N2O1N2O1O1001O001OO1O0010O01O1O001O010O001O1O001O001O1O000000001O000000000001O000001O0000000000000000001O0000000000000000000000000000000000O10000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000O101O00001O5K3MO10000000000000000O100000000000000L4O1O1O1O10000001O000000001O00004L1O0000000O01000000O100000000O1O1N200000000005K1O00000000000000000000O1N2O1O1O1N1100000000000000000000000000000000000008H3M0000O100000000000000O10000GKmH5S791O00000000001O0000001O0000001O0000001OO100O10000O10000O10000O10000O100000000O10000000000000000001O00000000001N100000001O00000000001O00000000000000001O00O1000O0100O10000O100O100000000000000000O10O100000000000000000000000000000O100000000000000O100000000000O010005K5K1O000O10001OO10000O01000O10000O10O10O1B>O1O1O1O001O1O1O1O1O1000O1000000000000O1000000O10000O100O100O10000O10O0100O10000O100O100O10000O2O0N2J6M4N1N2N`Y="}, "label": "boat"}]} {"id": 222977, "image": "COCO_train2014_000000222977.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white photo of a man dressed in business attire on skis .. posing\"."}], "task": "refering", "answer_template": "The \"a black and white photo of a man dressed in business attire on skis .. posing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 36, 37, 38, 39, 58, 59, 60, 61, 64, 65, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 125, 126, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 270, 286, 287, 288, 290, 291, 292, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 335, 336, 337, 354, 355, 356, 357, 358, 359, 360], "bbox": [0.40412499999999996, 0.008473804100227791, 0.8500624999999999, 0.9904100227790433], "iscrowd": 0, "area": 53023.81085, "rle": {"size": [439, 640], "counts": "Y^_3?U=3M3M3M3M3M3M3M3M3O001N2O1O1N2O1O1O1hGcNT4_1jKcNU4^1iKeNU4\\1hKhNU4Z1iKhNV4Y1hKjNV4W1gKmNW4T1gKnNW4T1gKoNW4R1gKQOW4P1hKQOV4Q1iKQOU4P1jKROT4o0kKSOR4o0mKROQ4P1nKROo3P1PLROn3o0QLROm3P1SLQOj3Q1VLPOg3R1YLnNe3T1[LmNb3U1^LlN`3U1_LlN_3V1aLkN\\3W1dLjNY3W1hLiNV3Y1jLhNS3Z1mLgNQ3Z1nLgNP3[1PMeNn2]1RMbNm2`1RMaNl2a1TM_Nj2c1VM\\Ni2f1VM[Ni2f1WMZNg2h1XMXNg2j1YMVNe2\\1aJ[Mj2X1d2^1eJXMf2[1c2]1jJWMc2\\1b2]1mJVMa2\\1a2_1QKSM]2_1`2^1VKRMZ2`1^2_1[KoLV2b1^2_1_KnLS2c1\\2`1cKlLQ2c1_OPOO`2eNkLl1f1^OSOM]2kNiLj1g1\\OXOJX2i1`N[N^OGS2o1_NXNBEo1U2_NSNHCj1Z2^NRNM@d1`2_NmM2_O`1e2]NkM8[O[1k2^NiM;WOX1Q3\\NhM?oNY1[3XNeMc0bN^1j3oMdMa4^2_KaMa4`2^K`Ma4b2_K]M`4e2_K[M`4f2aKYM^4i2aKWM^4k2bKTM]4n2bKfLi4\\3WK\\Lo4e3QKZLP5g3PKVLQ5l3nJRLS5R4kJkKV5Y4Q23M2N2N2N2N2OPMfGe0X8[OlGc0R8]ORHa0l7_OXH?f7A^H=`7CdH;Z7EiH:U7FmH:Q7FQI9n6GSI:k6FWI:g6F[I:d6D_I<_6DcI<[6EfI;X6EjI;T6EnI;P6ERJ;l5FUJ:i5FXJ:g5F[J:b5H_J8^5IdJ7Z5IhJ6V5KlJ5Q5MPK3n4MTK2j4OXKTOaNiNT6U2[KlNiNnNj5V2_KiNjNQOd5W2dKeNlNRO^5[2gK`NnNUOY5\\2jK\\NPOWOZ5Y2gK]NROXO\\5W2cK`NSOWO]5V2aKaNTOXO_5S2^KdNTOWOb5S2ZKdNWOWOc5Q2VKhN^6T1cIjNa6S1`IlNd6P1]InNg6o0ZIPOj6l0WISOm6i0TIVOo6g0RIWOS7e0nHZOU7c0lH\\OX7`0hH@[7=fHB^7:cHF`76aHJa73`HMc7O^H1d:000000000O10000]OClC=QUDCh;?XDCe;?ZDBc;`0]DAa;a0^D@_;b0`D@j5ZODX1aJ_Og5[OGY1aJ]Od5[OLY1`J]O`5\\OOZ1`J\\O[5h0jNMkK_O`1YOi1`1h0KnK@R1AT2Z1h0FRLCj0BX2Z1h0P1gLlM]2Y1h0`2TOfMh0\\2SOjMi0X2TOnMh0S2UORNg0Q2UOUNg0l1VOYNf0j1UO\\Nh0e1UOaNf0a1YObNd0`1ZOdNa0^1_OdN>]1AgN:\\1EfN8[1GiN4Y1LiN1X1NkNNV13mNIT16oNFS19QOCP1=RO_OP1`0TO\\Om0d0UOXOl0h0WOTOk0l0XOPOi0o0ZOmNh0S1[OiNg0U1\\OgNf0Y1]OcNe0[1^OaNe0]1^O`Nc0`1k5N2O1N101N2O0000001N10000000001O0000000000000000000000000000CnNUDR1j;QOTDo0k;UORDk0l;YORDg0m;b0O1N2O1O100000000001O0O100O100O2O1N1O2O1N2O1N101N2N2O1L4K4IbRY1"}, "label": "a black and white photo of a man dressed in business attire on skis .. posing"}]} {"id": 222977, "image": "COCO_train2014_000000222977.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man is holding ski poles\"."}], "task": "refering", "answer_template": "The \"the man is holding ski poles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 36, 37, 38, 39, 58, 59, 60, 61, 64, 65, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 125, 126, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 269, 270, 286, 287, 288, 290, 291, 292, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 335, 336, 337, 354, 355, 356, 357, 358, 359, 360], "bbox": [0.40412499999999996, 0.008473804100227791, 0.8500624999999999, 0.9904100227790433], "iscrowd": 0, "area": 53023.81085, "rle": {"size": [439, 640], "counts": "Y^_3?U=3M3M3M3M3M3M3M3M3O001N2O1O1N2O1O1O1hGcNT4_1jKcNU4^1iKeNU4\\1hKhNU4Z1iKhNV4Y1hKjNV4W1gKmNW4T1gKnNW4T1gKoNW4R1gKQOW4P1hKQOV4Q1iKQOU4P1jKROT4o0kKSOR4o0mKROQ4P1nKROo3P1PLROn3o0QLROm3P1SLQOj3Q1VLPOg3R1YLnNe3T1[LmNb3U1^LlN`3U1_LlN_3V1aLkN\\3W1dLjNY3W1hLiNV3Y1jLhNS3Z1mLgNQ3Z1nLgNP3[1PMeNn2]1RMbNm2`1RMaNl2a1TM_Nj2c1VM\\Ni2f1VM[Ni2f1WMZNg2h1XMXNg2j1YMVNe2\\1aJ[Mj2X1d2^1eJXMf2[1c2]1jJWMc2\\1b2]1mJVMa2\\1a2_1QKSM]2_1`2^1VKRMZ2`1^2_1[KoLV2b1^2_1_KnLS2c1\\2`1cKlLQ2c1_OPOO`2eNkLl1f1^OSOM]2kNiLj1g1\\OXOJX2i1`N[N^OGS2o1_NXNBEo1U2_NSNHCj1Z2^NRNM@d1`2_NmM2_O`1e2]NkM8[O[1k2^NiM;WOX1Q3\\NhM?oNY1[3XNeMc0bN^1j3oMdMa4^2_KaMa4`2^K`Ma4b2_K]M`4e2_K[M`4f2aKYM^4i2aKWM^4k2bKTM]4n2bKfLi4\\3WK\\Lo4e3QKZLP5g3PKVLQ5l3nJRLS5R4kJkKV5Y4Q23M2N2N2N2N2OPMfGe0X8[OlGc0R8]ORHa0l7_OXH?f7A^H=`7CdH;Z7EiH:U7FmH:Q7FQI9n6GSI:k6FWI:g6F[I:d6D_I<_6DcI<[6EfI;X6EjI;T6EnI;P6ERJ;l5FUJ:i5FXJ:g5F[J:b5H_J8^5IdJ7Z5IhJ6V5KlJ5Q5MPK3n4MTK2j4OXKTOaNiNT6U2[KlNiNnNj5V2_KiNjNQOd5W2dKeNlNRO^5[2gK`NnNUOY5\\2jK\\NPOWOZ5Y2gK]NROXO\\5W2cK`NSOWO]5V2aKaNTOXO_5S2^KdNTOWOb5S2ZKdNWOWOc5Q2VKhN^6T1cIjNa6S1`IlNd6P1]InNg6o0ZIPOj6l0WISOm6i0TIVOo6g0RIWOS7e0nHZOU7c0lH\\OX7`0hH@[7=fHB^7:cHF`76aHJa73`HMc7O^H1d:000000000O10000]OClC=QUDCh;?XDCe;?ZDBc;`0]DAa;a0^D@_;b0`D@j5ZODX1aJ_Og5[OGY1aJ]Od5[OLY1`J]O`5\\OOZ1`J\\O[5h0jNMkK_O`1YOi1`1h0KnK@R1AT2Z1h0FRLCj0BX2Z1h0P1gLlM]2Y1h0`2TOfMh0\\2SOjMi0X2TOnMh0S2UORNg0Q2UOUNg0l1VOYNf0j1UO\\Nh0e1UOaNf0a1YObNd0`1ZOdNa0^1_OdN>]1AgN:\\1EfN8[1GiN4Y1LiN1X1NkNNV13mNIT16oNFS19QOCP1=RO_OP1`0TO\\Om0d0UOXOl0h0WOTOk0l0XOPOi0o0ZOmNh0S1[OiNg0U1\\OgNf0Y1]OcNe0[1^OaNe0]1^O`Nc0`1k5N2O1N101N2O0000001N10000000001O0000000000000000000000000000CnNUDR1j;QOTDo0k;UORDk0l;YORDg0m;b0O1N2O1O100000000001O0O100O100O2O1N1O2O1N2O1N101N2N2O1L4K4IbRY1"}, "label": "the man is holding ski poles"}]} {"id": 222977, "image": "COCO_train2014_000000222977.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man wearing black shirt and with a letters nh\"."}], "task": "refering", "answer_template": "The \"the man wearing black shirt and with a letters nh\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 28, 29, 30, 51, 52, 53, 54, 73, 74, 75, 76, 77, 96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 258, 259, 260, 261, 280, 281, 282, 283, 284, 305, 306, 307, 327, 328, 329, 330, 350, 351, 352, 353], "bbox": [0.17628125, 0.011184510250569477, 0.41385937500000003, 1.0], "iscrowd": 0, "area": 39847.66814999998, "rle": {"size": [439, 640], "counts": "ha`1?V=e0ZOa0_O4M3L4M3L3M4M3N2N2N2PJ`Mf0b2UK\\MR3:a1\\2[KbMm28b1W2^KjMh22g1V2_KPNc2Lk1V2`KVN]2GP2U2`K]NY2_OU2V2`KcNS2ZOZ2U2`KjNo1SO_2T2`KPOj1lNf2U2^KXOe1cNl2W2]K^O_1[NT3Y2ZKFZ1PN]3\\2VK9h0\\MR4e6ZKZIf4l70O100O001O100O00fM[GDe8OhG1W8OjGlNIG]8]1kGeN3JQ8a1mG_N`Lc2X4[MjKb2Y4\\MhKb2[4\\MgKa2\\4]MfK`2\\4_MeK_2^4_MdK^2^4aMcK]2`4aMbK\\2`4cMaKZ2c4dM_KY2R8M2N3M2N3M2N3M7I8H7I8G9H7I8H\\cP5"}, "label": "the man wearing black shirt and with a letters nh"}]} {"id": 222977, "image": "COCO_train2014_000000222977.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man having nh letters on his tea shirt\"."}], "task": "refering", "answer_template": "The \"a man having nh letters on his tea shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 28, 29, 30, 51, 52, 53, 54, 73, 74, 75, 76, 77, 96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 258, 259, 260, 261, 280, 281, 282, 283, 284, 305, 306, 307, 327, 328, 329, 330, 350, 351, 352, 353], "bbox": [0.17628125, 0.011184510250569477, 0.41385937500000003, 1.0], "iscrowd": 0, "area": 39847.66814999998, "rle": {"size": [439, 640], "counts": "ha`1?V=e0ZOa0_O4M3L4M3L3M4M3N2N2N2PJ`Mf0b2UK\\MR3:a1\\2[KbMm28b1W2^KjMh22g1V2_KPNc2Lk1V2`KVN]2GP2U2`K]NY2_OU2V2`KcNS2ZOZ2U2`KjNo1SO_2T2`KPOj1lNf2U2^KXOe1cNl2W2]K^O_1[NT3Y2ZKFZ1PN]3\\2VK9h0\\MR4e6ZKZIf4l70O100O001O100O00fM[GDe8OhG1W8OjGlNIG]8]1kGeN3JQ8a1mG_N`Lc2X4[MjKb2Y4\\MhKb2[4\\MgKa2\\4]MfK`2\\4_MeK_2^4_MdK^2^4aMcK]2`4aMbK\\2`4cMaKZ2c4dM_KY2R8M2N3M2N3M2N3M7I8H7I8G9H7I8H\\cP5"}, "label": "a man having nh letters on his tea shirt"}]} {"id": 222977, "image": "COCO_train2014_000000222977.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old man wearing a tie and spectacles\"."}], "task": "refering", "answer_template": "The \"an old man wearing a tie and spectacles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 46, 47, 69, 70, 71, 92, 93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 299, 300, 301, 322, 323, 324, 345, 346, 347], "bbox": [0.0, 0.04849658314350797, 0.16609374999999998, 0.9976765375854214], "iscrowd": 0, "area": 33000.09885000002, "rle": {"size": [439, 640], "counts": "e16_6g3_LdLPNn8c1]GYNh8^1_G_Ng8Z1\\GeNh8U1[GiN`9:eFC\\9=eFB[9>fF@[9?gF_OZ9a0hF\\OY9c0jF[OV9e0lFVOX9i0k1O10000O1000000000O0100001N1000001N10000010O0100000O10O1O0O2bDWOl9j0SF@c9b0[FAb9`0]F@c9b0ZF@e9a0YF@f9b0XF_Oh9b0UF@k9a0RFAn9`0PFAP:`0mECQ:`0kEBU:`0gEBY:?eEB[:=eED[:B5K6J5K6J6[MkH`N[7\\1VIQNo6k1_IfMg6W2_2M3M3L3N3M3M3K4K6K;C=C=CihT7"}, "label": "an old man wearing a tie and spectacles"}]} {"id": 222977, "image": "COCO_train2014_000000222977.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a tie and glasses\"."}], "task": "refering", "answer_template": "The \"a man wearing a tie and glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 46, 47, 69, 70, 71, 92, 93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 299, 300, 301, 322, 323, 324, 345, 346, 347], "bbox": [0.0, 0.04849658314350797, 0.16609374999999998, 0.9976765375854214], "iscrowd": 0, "area": 33000.09885000002, "rle": {"size": [439, 640], "counts": "e16_6g3_LdLPNn8c1]GYNh8^1_G_Ng8Z1\\GeNh8U1[GiN`9:eFC\\9=eFB[9>fF@[9?gF_OZ9a0hF\\OY9c0jF[OV9e0lFVOX9i0k1O10000O1000000000O0100001N1000001N10000010O0100000O10O1O0O2bDWOl9j0SF@c9b0[FAb9`0]F@c9b0ZF@e9a0YF@f9b0XF_Oh9b0UF@k9a0RFAn9`0PFAP:`0mECQ:`0kEBU:`0gEBY:?eEB[:=eED[:B5K6J5K6J6[MkH`N[7\\1VIQNo6k1_IfMg6W2_2M3M3L3N3M3M3K4K6K;C=C=CihT7"}, "label": "a man wearing a tie and glasses"}]} {"id": 313091, "image": "COCO_train2014_000000313091.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bloody sheep on the grassy pasture who just had a baby\"."}], "task": "refering", "answer_template": "The \"a bloody sheep on the grassy pasture who just had a baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 177, 178, 194, 195, 196, 197, 198, 199, 200, 201, 217, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 264, 265, 268, 269, 287, 288, 290, 291, 292], "bbox": [0.39015625, 0.27018796992481203, 0.762875, 0.6717669172932331], "iscrowd": 0, "area": 25650.273349999996, "rle": {"size": [532, 640], "counts": "RRR4:W`04N1N3N2M2O1O1N2O1N3N100O1O1O100O1N2O2HkNn@V1P?8M30101O1O1O1N3N1O2M3N2M4M2M3N1N3K4M4L3L4L4L5J5N2M3N2M4M2O11O1O001O00O1O1N10001O0mDWLY9j3cFZL]9g3_F\\L`9e3kEZL^O5h:a3cEUM^:l2YE[Mg:f2SE_Mn:S41O001O1N2O1YOg0K5J6M3NaDQL[:o3XE^Li:g4O01O001O10O01O1O001UOk0I7UOj0L5N2O010O10O1O001O001O001O010O001O00001O00001O00001O00001O010O00001O00001O00001O00001O00000001O0001O000000000000001O000000000000000000001UDbL^:^3_EgL_:Y3^EkLa:U3\\EmLe:S3ZEnLg:Q3WEQMi:o2VEQMk:o2SESMm:m2RETMn:l2PEUMQ;k2nDVMR;k2`DaMa;i30N2N2N2M3N2M3N2M4L3N2M4M2N2M4L3M3N3M2M3O2N100O2O0O1100O2N1SDlLY:U3aEPM^:Q3^ERMb:o2YEUMh:S3kDQMU;Z3\\DjLe;R40O101O001O001O0O2H7K6J5UOaC]Md in this image.", "objects": [{"patches": [124, 125, 126, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 177, 178, 194, 195, 196, 197, 198, 199, 200, 201, 217, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 264, 265, 268, 269, 287, 288, 290, 291, 292], "bbox": [0.39015625, 0.27018796992481203, 0.762875, 0.6717669172932331], "iscrowd": 0, "area": 25650.273349999996, "rle": {"size": [532, 640], "counts": "RRR4:W`04N1N3N2M2O1O1N2O1N3N100O1O1O100O1N2O2HkNn@V1P?8M30101O1O1O1N3N1O2M3N2M4M2M3N1N3K4M4L3L4L4L5J5N2M3N2M4M2O11O1O001O00O1O1N10001O0mDWLY9j3cFZL]9g3_F\\L`9e3kEZL^O5h:a3cEUM^:l2YE[Mg:f2SE_Mn:S41O001O1N2O1YOg0K5J6M3NaDQL[:o3XE^Li:g4O01O001O10O01O1O001UOk0I7UOj0L5N2O010O10O1O001O001O001O010O001O00001O00001O00001O00001O010O00001O00001O00001O00001O00000001O0001O000000000000001O000000000000000000001UDbL^:^3_EgL_:Y3^EkLa:U3\\EmLe:S3ZEnLg:Q3WEQMi:o2VEQMk:o2SESMm:m2RETMn:l2PEUMQ;k2nDVMR;k2`DaMa;i30N2N2N2M3N2M3N2M4L3N2M4M2N2M4L3M3N3M2M3O2N100O2O0O1100O2N1SDlLY:U3aEPM^:Q3^ERMb:o2YEUMh:S3kDQMU;Z3\\DjLe;R40O101O001O001O0O2H7K6J5UOaC]Md in this image.", "objects": [{"patches": [103, 104, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 171, 172, 173, 174, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 264, 265], "bbox": [0.40059374999999997, 0.3071662763466042, 0.604484375, 0.7628337236533959], "iscrowd": 0, "area": 15784.247999999998, "rle": {"size": [427, 640], "counts": "SP[31Q=S1QOiNk87dGM`0Q1TOjNe8=dGIc0n0WOkNb8`0aGIe0k0ZOlN_8a0aGHf0i0^OlN\\8b0`GIg0f0@nNY8c0`GIk0;CYOS8b0^GLo0LHGk7a0^GLS1@L3d7`0]GMZ13jNA_8?\\GNY16jN]Oa8?\\GOU1;kNWOe8>[G0T1>iNUOh8>ZGOT1a0hNROk8>WG0V1c0dNPOQ9kG]OV8a0PHZOQ8e0UHUOm7i0XHROi7m0[2O1O2MQ[Y3"}, "label": "a police bike without red and blue lights on"}]} {"id": 395013, "image": "COCO_train2014_000000395013.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a police motorcycle to the left of the other motorcycle\"."}], "task": "refering", "answer_template": "The \"a police motorcycle to the left of the other motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 120, 121, 122, 123, 143, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 259], "bbox": [0.23284375000000002, 0.30234192037470725, 0.380609375, 0.7408196721311475], "iscrowd": 0, "area": 13469.743500000002, "rle": {"size": [427, 640], "counts": "^]n15n<;I6nM_OSGd0g7^O]G4g0a0j7MjFHX1=l7h0oGZOP8j0kGYOR8l0iGVOV8n0eGTOY8Q1aGSO]8Q1^GQO`8S1\\GoNc8o0^GTO^8m0bGUO\\8R1]GPO`8o2N3M2M3N3N1O1O2N1O1O1N3N1O1O2N1O1O1O2M2O1O2N1O1O1000O2O0O10000O100O100O2O0O100O10000O100O2O0O100O10000O101N100O100O3N3L3N3M2M3I8H7I8H7I8E:C>CCBRnT5"}, "label": "a police motorcycle to the left of the other motorcycle"}]} {"id": 395013, "image": "COCO_train2014_000000395013.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a police officer driving a motorcycle on the right side of the road\"."}], "task": "refering", "answer_template": "The \"a police officer driving a motorcycle on the right side of the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 120, 121, 122, 123, 143, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 259], "bbox": [0.23284375000000002, 0.30234192037470725, 0.380609375, 0.7408196721311475], "iscrowd": 0, "area": 13469.743500000002, "rle": {"size": [427, 640], "counts": "^]n15n<;I6nM_OSGd0g7^O]G4g0a0j7MjFHX1=l7h0oGZOP8j0kGYOR8l0iGVOV8n0eGTOY8Q1aGSO]8Q1^GQO`8S1\\GoNc8o0^GTO^8m0bGUO\\8R1]GPO`8o2N3M2M3N3N1O1O2N1O1O1N3N1O1O2N1O1O1O2M2O1O2N1O1O1000O2O0O10000O100O100O2O0O100O10000O100O2O0O100O10000O101N100O100O3N3L3N3M2M3I8H7I8H7I8E:C>CCBRnT5"}, "label": "a police officer driving a motorcycle on the right side of the road"}]} {"id": 124694, "image": "COCO_train2014_000000124694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sheep under arm of woman wearing dark top\"."}], "task": "refering", "answer_template": "The \"sheep under arm of woman wearing dark top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 304, 307, 308, 330], "bbox": [0.21667187499999999, 0.44945564516129033, 0.49499999999999994, 0.7849395161290323], "iscrowd": 0, "area": 15924.831950000002, "rle": {"size": [496, 640], "counts": "mcS2>i>?A>E9lBiNe;_1TDeNh;a1RDbNk;d2L4L3N2M2010000O010OA`0H7M4M3L3N3L3O20O2O1O1O1O1O1O1O1O1O00001O00001O00001O0000001O000010O002O0O1O1O01O0001O0000000000001O000000000000000000O2O1O1O1O1O1O1O1N2O2M4L4M3L4M3L5L`0_O100O101N1O100O100O1O100O1O01O0010O0001O00001O001O0_OUE_Lj:W3cEfL^:R3kElLU:S3oEjLQ:U3RFiLo9U3V10001N10001O00001O0O101N2O1O1N2O1N2N1O2N2N2UO_CdNd<\\1`C\\Nd in this image.", "objects": [{"patches": [192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 304, 307, 308, 330], "bbox": [0.21667187499999999, 0.44945564516129033, 0.49499999999999994, 0.7849395161290323], "iscrowd": 0, "area": 15924.831950000002, "rle": {"size": [496, 640], "counts": "mcS2>i>?A>E9lBiNe;_1TDeNh;a1RDbNk;d2L4L3N2M2010000O010OA`0H7M4M3L3N3L3O20O2O1O1O1O1O1O1O1O1O00001O00001O00001O0000001O000010O002O0O1O1O01O0001O0000000000001O000000000000000000O2O1O1O1O1O1O1O1N2O2M4L4M3L4M3L5L`0_O100O101N1O100O100O1O100O1O01O0010O0001O00001O001O0_OUE_Lj:W3cEfL^:R3kElLU:S3oEjLQ:U3RFiLo9U3V10001N10001O00001O0O101N2O1O1N2O1N2N1O2N2N2UO_CdNd<\\1`C\\Nd in this image.", "objects": [{"patches": [61, 62, 84, 85, 86, 107, 108, 109, 129, 130, 131, 132, 133, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 198, 200, 201, 202, 221, 223, 224, 225, 246, 247, 248, 268, 269, 270, 271, 272, 291, 292, 293, 294, 295, 315, 316], "bbox": [0.610296875, 0.12550403225806453, 0.8304843749999999, 0.7518951612903225], "iscrowd": 0, "area": 22990.4893, "rle": {"size": [496, 640], "counts": "Tdm55Z?4L5K3M1N2O1O1E:01O0O00O1000O3XO`0M8H9F:1O1O2N2O00100O0010O0dNTNVEk1i:XNUEh1j:[NTEf1W:kM\\Eb0a1W:TNXE=`0_1V:WNXE<`0]1Y:WNVE>`0[1Y:YNTE>b0Z1Z:[OeEe0Z:\\OeEe0[:[OcEg0\\:ZObEh0^:YO`Eh0`:YO^Eh0b:YO\\Eg0e:ZOYEg0U3eMP4f1hHf0T3jMQ4b1iHe0R3oMR4^1jHd0Q3RNS4o3iKWLS4k3hK\\LT4f3RJYKd0Y1W5_3VJYK;_1_5W3VJ[K5e1c5Q3WJ[KOl1h5i2YJ]KFS2P6`2YJ^K_O[2V6X2ZJ^K^O]2V6U2\\J_K]O^2V6S2\\J`K\\O`2V6Q2]J`K[Oa2W6P2]J_K[Oc2X6m1]J`KZOe2X6l1]J_KYOg2Z6k1\\J^KYOi2[6h1\\JeNc5\\1]JdNb5]1]JcNd5]1[JcNe5_1YJaNg5a1WJ_Ni5d1TJ\\Nm5e1QJ[No5l1jITNV6c51O011N2N2O1N1O2OO0100O1O10000O0dH\\Im5d6QJ]IP6c6oI]IQ6d6mI]IT6c6jI^IV6c6hI_IW6b6eIaI[6`6aIdI_6]6]IeIc6^6WIfIh6]6RIfIo6^7O001O010O001O1O1K5K4M4K5PKeGh2]8QMlGl2U8nLSHo2n7lLXHQ3i7lL\\HR3e7jL`HT3a7gLeHW3\\7eLiHX3\\7`LkH]3Z7[LkHc3c9N2M3N3M3Mc0\\O5L4L4K5L4L5K6I7J6I7G9Gooc1"}, "label": "woman on right"}]} {"id": 124694, "image": "COCO_train2014_000000124694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with black hair rubbing a sheep\"."}], "task": "refering", "answer_template": "The \"a woman with black hair rubbing a sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 84, 85, 86, 107, 108, 109, 129, 130, 131, 132, 133, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 198, 200, 201, 202, 221, 223, 224, 225, 246, 247, 248, 268, 269, 270, 271, 272, 291, 292, 293, 294, 295, 315, 316], "bbox": [0.610296875, 0.12550403225806453, 0.8304843749999999, 0.7518951612903225], "iscrowd": 0, "area": 22990.4893, "rle": {"size": [496, 640], "counts": "Tdm55Z?4L5K3M1N2O1O1E:01O0O00O1000O3XO`0M8H9F:1O1O2N2O00100O0010O0dNTNVEk1i:XNUEh1j:[NTEf1W:kM\\Eb0a1W:TNXE=`0_1V:WNXE<`0]1Y:WNVE>`0[1Y:YNTE>b0Z1Z:[OeEe0Z:\\OeEe0[:[OcEg0\\:ZObEh0^:YO`Eh0`:YO^Eh0b:YO\\Eg0e:ZOYEg0U3eMP4f1hHf0T3jMQ4b1iHe0R3oMR4^1jHd0Q3RNS4o3iKWLS4k3hK\\LT4f3RJYKd0Y1W5_3VJYK;_1_5W3VJ[K5e1c5Q3WJ[KOl1h5i2YJ]KFS2P6`2YJ^K_O[2V6X2ZJ^K^O]2V6U2\\J_K]O^2V6S2\\J`K\\O`2V6Q2]J`K[Oa2W6P2]J_K[Oc2X6m1]J`KZOe2X6l1]J_KYOg2Z6k1\\J^KYOi2[6h1\\JeNc5\\1]JdNb5]1]JcNd5]1[JcNe5_1YJaNg5a1WJ_Ni5d1TJ\\Nm5e1QJ[No5l1jITNV6c51O011N2N2O1N1O2OO0100O1O10000O0dH\\Im5d6QJ]IP6c6oI]IQ6d6mI]IT6c6jI^IV6c6hI_IW6b6eIaI[6`6aIdI_6]6]IeIc6^6WIfIh6]6RIfIo6^7O001O010O001O1O1K5K4M4K5PKeGh2]8QMlGl2U8nLSHo2n7lLXHQ3i7lL\\HR3e7jL`HT3a7gLeHW3\\7eLiHX3\\7`LkH]3Z7[LkHc3c9N2M3N3M3Mc0\\O5L4L4K5L4L5K6I7J6I7G9Gooc1"}, "label": "a woman with black hair rubbing a sheep"}]} {"id": 124694, "image": "COCO_train2014_000000124694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sheep on the right\"."}], "task": "refering", "answer_template": "The \"the sheep on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 312, 313, 314], "bbox": [0.47296875, 0.43070564516129034, 0.731859375, 0.7626209677419354], "iscrowd": 0, "area": 16454.110899999992, "rle": {"size": [496, 640], "counts": "YQc4S1[>8I7I7H8I7H7C>G9F9H4M3M3M3N2M3MaNXE[Nf:d1dERN\\:P2kEhMT:Z2SF]Mo9c2_10O1000001O1O1N2O2N1O2N1N3N00001O0O2O00001O00001O0000001O00000000001O00000000001O0001O000000010O00001O001O0001WDiLX;W3_DSM_;a3O1N2N2O1N2N010O00010O001OoNkDgMU;X2R1N2O1O1O]OUMbDk2_;YM[Dh2f;a00000000001O00000O1000000O2aD^Ln:b3lDiLo:W3nDPMn:k3M2O1O1N2O2N1N2O1O1N10lNbEUM_:h2lEnLU:Q3YFbLh9\\3W1M2O2L4M3L4M3L3M4M3L4L4L4L4L4M3M3N2N3M2N3M3M2N3M3M3N2N3M2N3L3N3M2M4J5L5KhRc2"}, "label": "the sheep on the right"}]} {"id": 124694, "image": "COCO_train2014_000000124694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sheep on the right side with the woman in white top\"."}], "task": "refering", "answer_template": "The \"the sheep on the right side with the woman in white top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 312, 313, 314], "bbox": [0.47296875, 0.43070564516129034, 0.731859375, 0.7626209677419354], "iscrowd": 0, "area": 16454.110899999992, "rle": {"size": [496, 640], "counts": "YQc4S1[>8I7I7H8I7H7C>G9F9H4M3M3M3N2M3MaNXE[Nf:d1dERN\\:P2kEhMT:Z2SF]Mo9c2_10O1000001O1O1N2O2N1O2N1N3N00001O0O2O00001O00001O0000001O00000000001O00000000001O0001O000000010O00001O001O0001WDiLX;W3_DSM_;a3O1N2N2O1N2N010O00010O001OoNkDgMU;X2R1N2O1O1O]OUMbDk2_;YM[Dh2f;a00000000001O00000O1000000O2aD^Ln:b3lDiLo:W3nDPMn:k3M2O1O1N2O2N1N2O1O1N10lNbEUM_:h2lEnLU:Q3YFbLh9\\3W1M2O2L4M3L4M3L3M4M3L4L4L4L4L4M3M3N2N3M2N3M3M2N3M3M3N2N3M2N3L3N3M2M4J5L5KhRc2"}, "label": "the sheep on the right side with the woman in white top"}]} {"id": 132889, "image": "COCO_train2014_000000132889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"slice of pizza on plate with cheese stringing from pan\"."}], "task": "refering", "answer_template": "The \"slice of pizza on plate with cheese stringing from pan\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 154, 155, 173, 174, 175, 176, 177, 178, 194, 195, 196, 197, 198, 199, 200, 201, 218], "bbox": [0.440359375, 0.4377570093457944, 0.757671875, 0.6191121495327103], "iscrowd": 0, "area": 6246.11795, "rle": {"size": [428, 640], "counts": "eSf32Z=2N0O10000000000000000000O10O10000001O0O1100O010O1O010O0010O010O01O010O1O01O0O10O1O010O1O001N2M2O1O2N1N3M2L4K6I6I7N11jNUDk0m;UOWDd0k;]OYD=h;D_D3c;Ln02O0O1O101O000O101O00000O20O01O10O01O10O01O10O01O1O001O1O001O1O001O1O001N101O1N101O1N101O1O001N2O001O001NBRDAm;?WD]Oi;c0[DZOd;f0`DWO_;i0b00O10001O1O1N3O1N1O2N1O2N2N001O10O01O001O001O1O0010O01O0000001O0000000010O00000001O000000001O0O100O1O2O00000000001O0O10000K6N1O100O100O100O12N1O2N1N2O2N1M4I6K5J;Ec0^O[bP2"}, "label": "slice of pizza on plate with cheese stringing from pan"}]} {"id": 132889, "image": "COCO_train2014_000000132889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza to the right in plate\"."}], "task": "refering", "answer_template": "The \"pizza to the right in plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 154, 155, 173, 174, 175, 176, 177, 178, 194, 195, 196, 197, 198, 199, 200, 201, 218], "bbox": [0.440359375, 0.4377570093457944, 0.757671875, 0.6191121495327103], "iscrowd": 0, "area": 6246.11795, "rle": {"size": [428, 640], "counts": "eSf32Z=2N0O10000000000000000000O10O10000001O0O1100O010O1O010O0010O010O01O010O1O01O0O10O1O010O1O001N2M2O1O2N1N3M2L4K6I6I7N11jNUDk0m;UOWDd0k;]OYD=h;D_D3c;Ln02O0O1O101O000O101O00000O20O01O10O01O10O01O10O01O1O001O1O001O1O001O1O001N101O1N101O1N101O1O001N2O001O001NBRDAm;?WD]Oi;c0[DZOd;f0`DWO_;i0b00O10001O1O1N3O1N1O2N1O2N2N001O10O01O001O001O1O0010O01O0000001O0000000010O00000001O000000001O0O100O1O2O00000000001O0O10000K6N1O100O100O100O12N1O2N1N2O2N1M4I6K5J;Ec0^O[bP2"}, "label": "pizza to the right in plate"}]} {"id": 132889, "image": "COCO_train2014_000000132889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"table with cloth\"."}], "task": "refering", "answer_template": "The \"table with cloth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 219, 220, 221, 238, 239, 240, 241, 242, 243, 244, 245, 246, 254, 255, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339], "bbox": [0.02409375, 0.6103504672897196, 0.777203125, 0.9887383177570094], "iscrowd": 0, "area": 53395.6709, "rle": {"size": [428, 640], "counts": "Re61Y=7J6I7I7J5J2N2O1N2O1N2N2O1N2N2O1N2N2O1N2O1N2N2[O]N`Ed1]:g01O0000000000001O000000000000001O0000000000001O0000000000001O0000000000001O0000000000001O0000000000001O00000000000000O100000000O1000000O1000000O1000000O100000000O1000000O1000000O100000000O1000000O1000000O1000000O100000000O1000000O1000000O100000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O100000000O1000000O1000000O100000000O1000000O1000000O1000000O100000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O100O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O100O1O100O100O100O100O100O100O100O10000O1000000000000000000000000O1000000000000000000000000O1000000000000000000000000O100000000000000000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O1000000O100000000O11O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O002N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1AhEhMY:V2mEeMT:X2SFcMo9Z2c0M3N2N2N3N1N2N2O1N3N1N2O1N2O2M2N2O1N2O2Ii]k1"}, "label": "table with cloth"}]} {"id": 132889, "image": "COCO_train2014_000000132889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the deep dish pizza pie with the slice missing\"."}], "task": "refering", "answer_template": "The \"the deep dish pizza pie with the slice missing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259], "bbox": [0.04659375, 0.4757242990654206, 0.443328125, 0.7700934579439253], "iscrowd": 0, "area": 23216.398300000004, "rle": {"size": [428, 640], "counts": "_i<6U=:F:F:F:F2O1N1O2N1O2N101N1O2N1O2N101N1O2N1O2N101N1O2N1O2N101N2N1O2N1O2O0O2N1O2N1O2O0O2N1O2N101N1O2N1O2N010O1O1O1O100000000000O1000O100000000000000000O1000O100000000000000000O1000O1000000000000000O100000O1000000000000000O100000O10000000000000O10000000O1000000000O1000000O1000O10O10000O1000000O1000000O100000O01000000O1000000O1000000O100000O01000000O1000000O1000000O1000000O0SNlEd0T:\\OlEd0T:\\OlEd0T:[OmEd0T:\\OlEd0T:[OnEd0R:\\OnEd0R:\\OnEc0S:\\OnEd0R:\\OnEd0R:\\OnEc0S:\\OnEd0R:\\OoEc0Q:\\OPFd0P:\\OPFc0Q:]OoEc0P:]OQFc0o9]OQFc0o9]OQFb0P:]ORFb0n9^ORFb0n9]OSFc0m9]OSFb0n9^ORFb0n9]OSFc0m9]OSFc0m9]OSFb0n9]OTFb0l9^OTFb0l9]OUFc0k9]OUFb0l9^OTFb0l9]OUFc0l9\\OTFc0m9]OTFa0n9]OSFb0n9^ORFa0P:]OQFb0P:^OPFa0R:^OnEa0S:^OnEa0T:^OlE`0W:_OiE`0X:_OiE`0Y:_OgE`0Z:_OgE`0[:_OeE`0\\:@dE?^:_OcE`0`:^O`Ea0f;O1O1O1O1N2O1O1O1O1O1O1O1O1O3M5K5J\\ad4"}, "label": "the deep dish pizza pie with the slice missing"}]} {"id": 132889, "image": "COCO_train2014_000000132889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza sitting up on a stand\"."}], "task": "refering", "answer_template": "The \"the pizza sitting up on a stand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259], "bbox": [0.04659375, 0.4757242990654206, 0.443328125, 0.7700934579439253], "iscrowd": 0, "area": 23216.398300000004, "rle": {"size": [428, 640], "counts": "_i<6U=:F:F:F:F2O1N1O2N1O2N101N1O2N1O2N101N1O2N1O2N101N1O2N1O2N101N2N1O2N1O2O0O2N1O2N1O2O0O2N1O2N101N1O2N1O2N010O1O1O1O100000000000O1000O100000000000000000O1000O100000000000000000O1000O1000000000000000O100000O1000000000000000O100000O10000000000000O10000000O1000000000O1000000O1000O10O10000O1000000O1000000O100000O01000000O1000000O1000000O100000O01000000O1000000O1000000O1000000O0SNlEd0T:\\OlEd0T:\\OlEd0T:[OmEd0T:\\OlEd0T:[OnEd0R:\\OnEd0R:\\OnEc0S:\\OnEd0R:\\OnEd0R:\\OnEc0S:\\OnEd0R:\\OoEc0Q:\\OPFd0P:\\OPFc0Q:]OoEc0P:]OQFc0o9]OQFc0o9]OQFb0P:]ORFb0n9^ORFb0n9]OSFc0m9]OSFb0n9^ORFb0n9]OSFc0m9]OSFc0m9]OSFb0n9]OTFb0l9^OTFb0l9]OUFc0k9]OUFb0l9^OTFb0l9]OUFc0l9\\OTFc0m9]OTFa0n9]OSFb0n9^ORFa0P:]OQFb0P:^OPFa0R:^OnEa0S:^OnEa0T:^OlE`0W:_OiE`0X:_OiE`0Y:_OgE`0Z:_OgE`0[:_OeE`0\\:@dE?^:_OcE`0`:^O`Ea0f;O1O1O1O1N2O1O1O1O1O1O1O1O1O3M5K5J\\ad4"}, "label": "the pizza sitting up on a stand"}]} {"id": 345882, "image": "COCO_train2014_000000345882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small girl wearing a brown dress with embroidered flowers\"."}], "task": "refering", "answer_template": "The \"a small girl wearing a brown dress with embroidered flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 128, 129, 130, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 266, 267, 268, 289, 290, 291, 312, 313, 336], "bbox": [0.5337500000000001, 0.2859484777517564, 0.700171875, 0.9601170960187354], "iscrowd": 0, "area": 18996.018399999997, "rle": {"size": [427, 640], "counts": "Wk^4l0^<3M1N2O1N2O100@kNjDU1T;nNkDR1T;oNlDQ1R;ROmDn0Q;UOnDk0P;WOPEi0>mNc9=nEf0;ROd9;PFc08XOd97TFa04]Oe0_Ok7f0[G>1Ce0\\Om7e0\\G=MGf0ZOo7d0^G;ILg0XOo7i0[GY1a0RNR8V3kGmLf0BT6b3RIoLd0EY6_3oHoLb0G_6[3kHRM`0Hd6W4VIoKj6Y5O1O10000O2O0eNSJWKn5g4\\1O000000000O100000000001N10ZJSIP5g7L101N101N101N1O2O0O2O0O2NF;N1N3N1O2M2O1N3N1O2M4M2O101N100O2O0O101N100O2O0O101N3N3L4UJ_JQ3EmKo5m0cJe2NYLc5m0bJ]2;bLV5l0cJT2X7gMlHk1c7PN`Hc1o7XNUHZ1Z8aNjGR1d8iN_GS1`:J6K4L5J6K4L5Kmj_2"}, "label": "a small girl wearing a brown dress with embroidered flowers"}]} {"id": 345882, "image": "COCO_train2014_000000345882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl wearing a brown dress with flowers\"."}], "task": "refering", "answer_template": "The \"a little girl wearing a brown dress with flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 128, 129, 130, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 266, 267, 268, 289, 290, 291, 312, 313, 336], "bbox": [0.5337500000000001, 0.2859484777517564, 0.700171875, 0.9601170960187354], "iscrowd": 0, "area": 18996.018399999997, "rle": {"size": [427, 640], "counts": "Wk^4l0^<3M1N2O1N2O100@kNjDU1T;nNkDR1T;oNlDQ1R;ROmDn0Q;UOnDk0P;WOPEi0>mNc9=nEf0;ROd9;PFc08XOd97TFa04]Oe0_Ok7f0[G>1Ce0\\Om7e0\\G=MGf0ZOo7d0^G;ILg0XOo7i0[GY1a0RNR8V3kGmLf0BT6b3RIoLd0EY6_3oHoLb0G_6[3kHRM`0Hd6W4VIoKj6Y5O1O10000O2O0eNSJWKn5g4\\1O000000000O100000000001N10ZJSIP5g7L101N101N101N1O2O0O2O0O2NF;N1N3N1O2M2O1N3N1O2M4M2O101N100O2O0O101N100O2O0O101N3N3L4UJ_JQ3EmKo5m0cJe2NYLc5m0bJ]2;bLV5l0cJT2X7gMlHk1c7PN`Hc1o7XNUHZ1Z8aNjGR1d8iN_GS1`:J6K4L5J6K4L5Kmj_2"}, "label": "a little girl wearing a brown dress with flowers"}]} {"id": 345882, "image": "COCO_train2014_000000345882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man leaning down\"."}], "task": "refering", "answer_template": "The \"a man leaning down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 25, 26, 27, 28, 29, 30, 31, 32, 49, 50, 51, 52, 53, 54, 73, 74, 209, 210, 211, 231, 232, 233, 234, 254, 255, 256, 257, 277, 278, 279, 280, 281, 282, 300, 301, 302, 303, 304, 305, 323, 324, 325], "bbox": [0.046484375, 0.004496487119437939, 0.439296875, 1.0], "iscrowd": 0, "area": 28881.03525, "rle": {"size": [427, 640], "counts": "`i<;P=?@`0A?A?@?B6I5L5J5L5K4K5L5J5L5J3N000O100O100O10000O100O100O100[G`K_8`4aGaK]8`4cGbK[8^4eGcKY8^4gGdKW8\\4iGfKT8[4lGfKS8Z4mGhKQ8X4oGjKn7W4RHjKm7V4SHlKj7U4VHmKh7S4XHnKf7S4ZHoKd7Q4\\HQLb7o3^HRL`7o3`HSL^7m3bHUL[7l3eHULZ7k3fHWLW7j3iHXLU7h3kHYLT7g3lH[LQ7f3oH\\Lo6d3QI]Lm6d3SI^Lk6b3UI`Lh6a3XI`Lg6`3YIbLe6^3[IdLb6]3^IdLa6\\3_IfL^6[3bIfL]6Z3cIhLZ6Y3fIiLX6W3gIkLX6U3hImLU6T3kImLT6S3lIoLQ6R3oIPMo5P3QJQMn5o2RJSMk5n2UJSMj5m2VJUMg5h2]JYMb5b2cJ`MZ5\\2kJfMS5U2RKlMm4o1XKSNe4j1_KWN`4d1eK^NX4`1kKaNT4^1mKdNP4]1PLeNn3[1RLfNm3Z1SLhNj3Y1VLhNi3X1WLjNf3W1ZLjNe3V1[LkNc3V1]LkNb3U1^LlNa3T1_LlN`3T1aLmN^3S1bLnN\\3S1dLnN[3TN_La06\\1Y3mLiMf1nN^1h5aNXJ_1j5_NVJb1j5]NVJd1k5ZNUJg1m5VNSJk1n5SNRJn1o5oMRJQ2P6mMPJT2R6iMnIX2S6fMmI[2T6cMlI^2V6_MjIa2X6]MhIa2\\6]MdIb2_6\\MaIc2c6ZM]Ie2f6YMZIf2i6WMXIh2l6UMTIj2m6VMSIi2n6WMRIg2P7YMPIf2Q7ZMoHe2Q7\\MoHc2R7]MnHb2S7^MmHa2T7_MlH`2T7aMlH^2U7aMlH^2U7bMkH\\2W7dMiH[2W7fMiHY2X7gMhHX2Y7hMgHW2Z7iMfHV2[7jMeHU2[7lMeHS2\\7mMdHQ2^7oMbHP2_7oMbHP2^7QNbHn1_7RNaHm1`7SN`Hk1b7UN^Hj1b7WN^Hh1c7XN]Hg1d7YN\\Hf1e7YN\\He1e7[N\\Hd1e7[N\\Hd1e7\\N[Hc1f7\\N[Hc1f7\\N[Hb1h7\\NYHc1i7[NXHd1j7[NVHd1m7YNTHf1n7XNSHf1P8XNQHg1Q8XNoGg1k900001O001O1O001O001O1O001O001O1O001O1O00O100O10O0100O100O100O1O100O100O100O100O11O1O001O1O001OO1N2O1N2N2N2N2O1O1O1O100O1O1O1O1O1O1O100O1O100001O001O001O001O1O00O1N2O1O5J5L5J5L4L5J5Ldee4"}, "label": "a man leaning down"}]} {"id": 345882, "image": "COCO_train2014_000000345882.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a button down shirt and jeans\"."}], "task": "refering", "answer_template": "The \"a man wearing a button down shirt and jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 25, 26, 27, 28, 29, 30, 31, 32, 49, 50, 51, 52, 53, 54, 73, 74, 209, 210, 211, 231, 232, 233, 234, 254, 255, 256, 257, 277, 278, 279, 280, 281, 282, 300, 301, 302, 303, 304, 305, 323, 324, 325], "bbox": [0.046484375, 0.004496487119437939, 0.439296875, 1.0], "iscrowd": 0, "area": 28881.03525, "rle": {"size": [427, 640], "counts": "`i<;P=?@`0A?A?@?B6I5L5J5L5K4K5L5J5L5J3N000O100O100O10000O100O100O100[G`K_8`4aGaK]8`4cGbK[8^4eGcKY8^4gGdKW8\\4iGfKT8[4lGfKS8Z4mGhKQ8X4oGjKn7W4RHjKm7V4SHlKj7U4VHmKh7S4XHnKf7S4ZHoKd7Q4\\HQLb7o3^HRL`7o3`HSL^7m3bHUL[7l3eHULZ7k3fHWLW7j3iHXLU7h3kHYLT7g3lH[LQ7f3oH\\Lo6d3QI]Lm6d3SI^Lk6b3UI`Lh6a3XI`Lg6`3YIbLe6^3[IdLb6]3^IdLa6\\3_IfL^6[3bIfL]6Z3cIhLZ6Y3fIiLX6W3gIkLX6U3hImLU6T3kImLT6S3lIoLQ6R3oIPMo5P3QJQMn5o2RJSMk5n2UJSMj5m2VJUMg5h2]JYMb5b2cJ`MZ5\\2kJfMS5U2RKlMm4o1XKSNe4j1_KWN`4d1eK^NX4`1kKaNT4^1mKdNP4]1PLeNn3[1RLfNm3Z1SLhNj3Y1VLhNi3X1WLjNf3W1ZLjNe3V1[LkNc3V1]LkNb3U1^LlNa3T1_LlN`3T1aLmN^3S1bLnN\\3S1dLnN[3TN_La06\\1Y3mLiMf1nN^1h5aNXJ_1j5_NVJb1j5]NVJd1k5ZNUJg1m5VNSJk1n5SNRJn1o5oMRJQ2P6mMPJT2R6iMnIX2S6fMmI[2T6cMlI^2V6_MjIa2X6]MhIa2\\6]MdIb2_6\\MaIc2c6ZM]Ie2f6YMZIf2i6WMXIh2l6UMTIj2m6VMSIi2n6WMRIg2P7YMPIf2Q7ZMoHe2Q7\\MoHc2R7]MnHb2S7^MmHa2T7_MlH`2T7aMlH^2U7aMlH^2U7bMkH\\2W7dMiH[2W7fMiHY2X7gMhHX2Y7hMgHW2Z7iMfHV2[7jMeHU2[7lMeHS2\\7mMdHQ2^7oMbHP2_7oMbHP2^7QNbHn1_7RNaHm1`7SN`Hk1b7UN^Hj1b7WN^Hh1c7XN]Hg1d7YN\\Hf1e7YN\\He1e7[N\\Hd1e7[N\\Hd1e7\\N[Hc1f7\\N[Hc1f7\\N[Hb1h7\\NYHc1i7[NXHd1j7[NVHd1m7YNTHf1n7XNSHf1P8XNQHg1Q8XNoGg1k900001O001O1O001O001O1O001O001O1O001O1O00O100O10O0100O100O100O1O100O100O100O100O11O1O001O1O001OO1N2O1N2N2N2N2O1O1O1O100O1O1O1O1O1O1O100O1O100001O001O001O001O1O00O1N2O1O5J5L5J5L4L5J5Ldee4"}, "label": "a man wearing a button down shirt and jeans"}]} {"id": 272155, "image": "COCO_train2014_000000272155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a laptop on tabule\"."}], "task": "refering", "answer_template": "The \"a laptop on tabule\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 164, 165, 166, 167, 168, 169, 182, 183, 184, 185, 186, 187, 200, 201, 202, 203, 204, 218, 219, 220], "bbox": [0.12972, 0.6539204545454546, 0.42713999999999996, 0.9550568181818182], "iscrowd": 0, "area": 9894.889100000002, "rle": {"size": [352, 500], "counts": "icf02m:5L4L4L4K5L4I7F:G8I8L4K5L1O001N101O000O2O001O001N1000000O1000000O1000000O1000000O1000O10O010000O10O0100O10O0100O10O10O100O010O100O010O1000O0100O100O010O1000O0100O10O0100O10O010000O010O100O010O100O01000O100O010O100O010O1000O0100O10O0100O10O010000O010O100O10O0100O10O10O1K5L3L5L4L5J5L4L4K5L4K5L5Kl]R3"}, "label": "a laptop on tabule"}]} {"id": 272155, "image": "COCO_train2014_000000272155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a laptop sitting on a table being used by a man\"."}], "task": "refering", "answer_template": "The \"a laptop sitting on a table being used by a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 164, 165, 166, 167, 168, 169, 182, 183, 184, 185, 186, 187, 200, 201, 202, 203, 204, 218, 219, 220], "bbox": [0.12972, 0.6539204545454546, 0.42713999999999996, 0.9550568181818182], "iscrowd": 0, "area": 9894.889100000002, "rle": {"size": [352, 500], "counts": "icf02m:5L4L4L4K5L4I7F:G8I8L4K5L1O001N101O000O2O001O001N1000000O1000000O1000000O1000000O1000O10O010000O10O0100O10O0100O10O10O100O010O100O010O1000O0100O100O010O1000O0100O10O0100O10O010000O010O100O010O100O01000O100O010O100O010O1000O0100O10O0100O10O010000O010O100O10O0100O10O10O1K5L3L5L4L5J5L4L4K5L4K5L5Kl]R3"}, "label": "a laptop sitting on a table being used by a man"}]} {"id": 272155, "image": "COCO_train2014_000000272155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a blue t - shirt walking out of the room\"."}], "task": "refering", "answer_template": "The \"a man wearing a blue t - shirt walking out of the room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 46, 47, 64, 65, 82, 83, 99, 100, 101, 117, 118, 119, 135, 137], "bbox": [0.5292, 0.10326704545454546, 0.6595, 0.6044034090909091], "iscrowd": 0, "area": 5185.88375, "rle": {"size": [352, 500], "counts": "`Xk2;a:6L3O2N001O0O101B=N1OO100O1O100O1O010OO2N2\\MkN]KX1c4jNPJN4Y1m5kNZIM08?S1V6jNPI6868l0b6YOTI4Of0n6WOnHP2T7m01O1N101O10000O10000O1002N:G8G:QJcKg4f4lJ^KT5b0^J`30SLc5:`Jn3`5PLcJP4W6O1O1O1O100O1O1O1gLkJe0V5VOXKa0i4\\ObK:`4CiK5Y4InK0T4MSLMo30XLJj33^LGi31e^j1"}, "label": "a man wearing a blue t - shirt walking out of the room"}]} {"id": 272155, "image": "COCO_train2014_000000272155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a dark shirt and jeans walking out of the room\"."}], "task": "refering", "answer_template": "The \"a man in a dark shirt and jeans walking out of the room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 46, 47, 64, 65, 82, 83, 99, 100, 101, 117, 118, 119, 135, 137], "bbox": [0.5292, 0.10326704545454546, 0.6595, 0.6044034090909091], "iscrowd": 0, "area": 5185.88375, "rle": {"size": [352, 500], "counts": "`Xk2;a:6L3O2N001O0O101B=N1OO100O1O100O1O010OO2N2\\MkN]KX1c4jNPJN4Y1m5kNZIM08?S1V6jNPI6868l0b6YOTI4Of0n6WOnHP2T7m01O1N101O10000O10000O1002N:G8G:QJcKg4f4lJ^KT5b0^J`30SLc5:`Jn3`5PLcJP4W6O1O1O1O100O1O1O1gLkJe0V5VOXKa0i4\\ObK:`4CiK5Y4InK0T4MSLMo30XLJj33^LGi31e^j1"}, "label": "a man in a dark shirt and jeans walking out of the room"}]} {"id": 272155, "image": "COCO_train2014_000000272155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a green short sleeved shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing a green short sleeved shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 74, 75, 76, 92, 93, 94, 95, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 180, 181, 182], "bbox": [0.013699999999999999, 0.2610795454545455, 0.3535, 0.7956818181818183], "iscrowd": 0, "area": 15546.202049999996, "rle": {"size": [352, 500], "counts": "ld24h:6J5K6J6J6J6K5L3M4K4M2N2M3N3M2N2N2M3N3L3I7N2M4M2M3N2N2N3N1N2N2O1O100O2OmNfH`NZ7]1jHbNU7_1kHaNU7_1lH`NT7_1mHaNR7`1oH_NQ7a1oH_NQ7a1PI^No6c1QI]No6c1RI\\Nm6d1TI\\Nl6d1UI[Nk6e1UI[Nj6f1WIZNh6f1XIZNh6e1YI[Nf6f1[IYNe6g1[IYNd6h1]IWNc6i1]IWNc6h1^IXNa6i1`IVN`6j1`IVN`6j1`IVN_6k1aIUN_6j1bIVNa5A_JY21UN[5IaJS24UNT5OgJl15UNS52eJj18TNQ55fJf1;TNn49eJc1=TNm4;dJb1?`MD4W5m0eJ_1`0_MJ0P5T1eJ]1a0_M]5U1RJ[1h6gNWIY1h6jNVIV1j6kNUIU1j6mNTIT1k6nNTIQ1m6a1O1O1000O100000O1N1O2N2O1000O100000000O101O001O000O2O001O001N2O1O1O1O1N2O1O1O1O1N3_MeHY1\\7dNfH\\1[7aNgH_1Z7_NhH`1Z7\\NhHc1[7YNgHg1\\7UNeHk1]7QNeHo1V800O100H8F:01O10O010000O101N11000O`G^Na7c1\\H^Nf7b1VH`Nk7`1RHaNQ8k1_GWNb8S22N2M3N3M4L2M3N2N1O2M3N2N2dN^FV1h9M3N1O2N2L4K6J7I7HVj^3"}, "label": "a man wearing a green short sleeved shirt"}]} {"id": 272155, "image": "COCO_train2014_000000272155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a green shirt staring at his open laptop\"."}], "task": "refering", "answer_template": "The \"a man in a green shirt staring at his open laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 74, 75, 76, 92, 93, 94, 95, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 180, 181, 182], "bbox": [0.013699999999999999, 0.2610795454545455, 0.3535, 0.7956818181818183], "iscrowd": 0, "area": 15546.202049999996, "rle": {"size": [352, 500], "counts": "ld24h:6J5K6J6J6J6K5L3M4K4M2N2M3N3M2N2N2M3N3L3I7N2M4M2M3N2N2N3N1N2N2O1O100O2OmNfH`NZ7]1jHbNU7_1kHaNU7_1lH`NT7_1mHaNR7`1oH_NQ7a1oH_NQ7a1PI^No6c1QI]No6c1RI\\Nm6d1TI\\Nl6d1UI[Nk6e1UI[Nj6f1WIZNh6f1XIZNh6e1YI[Nf6f1[IYNe6g1[IYNd6h1]IWNc6i1]IWNc6h1^IXNa6i1`IVN`6j1`IVN`6j1`IVN_6k1aIUN_6j1bIVNa5A_JY21UN[5IaJS24UNT5OgJl15UNS52eJj18TNQ55fJf1;TNn49eJc1=TNm4;dJb1?`MD4W5m0eJ_1`0_MJ0P5T1eJ]1a0_M]5U1RJ[1h6gNWIY1h6jNVIV1j6kNUIU1j6mNTIT1k6nNTIQ1m6a1O1O1000O100000O1N1O2N2O1000O100000000O101O001O000O2O001O001N2O1O1O1O1N2O1O1O1O1N3_MeHY1\\7dNfH\\1[7aNgH_1Z7_NhH`1Z7\\NhHc1[7YNgHg1\\7UNeHk1]7QNeHo1V800O100H8F:01O10O010000O101N11000O`G^Na7c1\\H^Nf7b1VH`Nk7`1RHaNQ8k1_GWNb8S22N2M3N3M4L2M3N2N1O2M3N2N2dN^FV1h9M3N1O2N2L4K6J7I7HVj^3"}, "label": "a man in a green shirt staring at his open laptop"}]} {"id": 67356, "image": "COCO_train2014_000000067356.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of sandwich with more visible meat\"."}], "task": "refering", "answer_template": "The \"half of sandwich with more visible meat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 85, 86, 87, 88, 103, 104, 105, 106, 107, 108, 109, 110, 111, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 241, 242, 243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 289, 290, 291], "bbox": [0.45775, 0.2434894613583138, 0.9125156249999999, 0.8219672131147542], "iscrowd": 0, "area": 51311.592549999994, "rle": {"size": [427, 640], "counts": "lZj31Z=2N2M3N1N3M3L4M3M3L4M3L3N3M3L5L3M4K5L3L5L3M4M3L3M4M3L3N3bF^Ml7e2nGjMg7Y2RHVNb7n1XHXNe7h3N2O2M2O2M2O2]H`JV7b5hH`JW7a5gH`JX7c5eH_JZ7j5N1O2O1N10O010O010O0012M2O1N3M2O2M2O2M2N101N2N101N1O2O0O2O1N1O2O001N101O001O1O001O001O001O1O0O2O001O0000000000000000000O100000000000000000000000O0100000O100000000O1000000O0100000O100000000O1000000O10000O1O10O0100O100O1O100O100O100O1O100O100O1O100O100O100O1O100O100O010O1O100O100O1O100O100O100O1O100O100O1O100O101N100O1O100O100O100O1O100O100UKmH]3T7cLlH\\3V7bLlH]3T7cLlH\\3U7cLlH]3T7cLlH\\3U7cLlH]3T7cLlH\\3U7cLlH\\3U7dLlH[3T7dLmH[3T7eLlH[3T7dLmH[3T7eLlH[3T7eLkH[3V7dLkH[3V7eLjH[3V7dLkH[3V7eLjH[3W7cLjH\\3W7dLiH[3X7dLiH\\3W7dLiH[3X7dLiH\\3W7dLhH\\3Z7bLgH^3Y7bLgH]3Z7bLgH]3[7bLeH^3[7aLfH^3[7bLgH\\3Z7cLgH[3Z7dLiHZ3W7fLjHX3X7fLkHW3V7iLQIP3P7nL]Ie2d6[MeI\\2[6cMgI[2[6dMeI[2\\6dMeI\\2[6dMeI[2]6cMdI]2\\6cMdI\\2^6bMcI^2]6bMcI]2^6`MeI_2]6]MfIc2Z6ZMiIe2_800O100O1O100O100O1XNYEk0h:VOXEi0h:WOYEg0h:YOYEf0g:ZOZEd0g:\\OZEc0f:POXEL2R1g:QOZEK0R1g:RO\\EJMT1g:PO_EKLR1h:PO_EMIR1k:lN`E0FQ1c;nN^DP1d;oN]Do0e;QO[Dm0f;SO\\Di0U in this image.", "objects": [{"patches": [82, 83, 84, 85, 86, 87, 88, 103, 104, 105, 106, 107, 108, 109, 110, 111, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 241, 242, 243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 269, 270, 289, 290, 291], "bbox": [0.45775, 0.2434894613583138, 0.9125156249999999, 0.8219672131147542], "iscrowd": 0, "area": 51311.592549999994, "rle": {"size": [427, 640], "counts": "lZj31Z=2N2M3N1N3M3L4M3M3L4M3L3N3M3L5L3M4K5L3L5L3M4M3L3M4M3L3N3bF^Ml7e2nGjMg7Y2RHVNb7n1XHXNe7h3N2O2M2O2M2O2]H`JV7b5hH`JW7a5gH`JX7c5eH_JZ7j5N1O2O1N10O010O010O0012M2O1N3M2O2M2O2M2N101N2N101N1O2O0O2O1N1O2O001N101O001O1O001O001O001O1O0O2O001O0000000000000000000O100000000000000000000000O0100000O100000000O1000000O0100000O100000000O1000000O10000O1O10O0100O100O1O100O100O100O1O100O100O1O100O100O100O1O100O100O010O1O100O100O1O100O100O100O1O100O100O1O100O101N100O1O100O100O100O1O100O100UKmH]3T7cLlH\\3V7bLlH]3T7cLlH\\3U7cLlH]3T7cLlH\\3U7cLlH]3T7cLlH\\3U7cLlH\\3U7dLlH[3T7dLmH[3T7eLlH[3T7dLmH[3T7eLlH[3T7eLkH[3V7dLkH[3V7eLjH[3V7dLkH[3V7eLjH[3W7cLjH\\3W7dLiH[3X7dLiH\\3W7dLiH[3X7dLiH\\3W7dLhH\\3Z7bLgH^3Y7bLgH]3Z7bLgH]3[7bLeH^3[7aLfH^3[7bLgH\\3Z7cLgH[3Z7dLiHZ3W7fLjHX3X7fLkHW3V7iLQIP3P7nL]Ie2d6[MeI\\2[6cMgI[2[6dMeI[2\\6dMeI\\2[6dMeI[2]6cMdI]2\\6cMdI\\2^6bMcI^2]6bMcI]2^6`MeI_2]6]MfIc2Z6ZMiIe2_800O100O1O100O100O1XNYEk0h:VOXEi0h:WOYEg0h:YOYEf0g:ZOZEd0g:\\OZEc0f:POXEL2R1g:QOZEK0R1g:RO\\EJMT1g:PO_EKLR1h:PO_EMIR1k:lN`E0FQ1c;nN^DP1d;oN]Do0e;QO[Dm0f;SO\\Di0U in this image.", "objects": [{"patches": [239, 240, 241, 261, 262, 263, 282, 283, 284, 285, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 346, 347, 348, 349, 350, 351, 368, 369, 370, 371, 372, 373, 390, 391, 392, 393, 394, 395, 412, 413, 414, 415, 416, 417, 434, 435, 436, 437, 438, 439, 456, 457, 458, 459, 460, 461, 481, 482, 483], "bbox": [0.72359477124183, 0.4629248366013072, 1.0, 0.9640522875816994], "iscrowd": 0, "area": 40153.8103, "rle": {"size": [612, 612], "counts": "gVY87ib0:E:G:F:E:G:F:E;F9G:E;G8J7I7I4L2N3M2N2O1N2N2N3M2N2N2N2N2N3M2N2N2N2N2O2M2N2N2N2N2N1O1O1O1O1O2N100O1O1O1O2N1O1O1O1O1O1O2N100O1O1O1O2N1O1O1O1O1O1O2M2O1O1O1N2O2N1O1N2O1O1N2O2\\Oc0N2N2N2N2N2N2N3M2N2N2N2N2N2O1O101N10000O10000O101N10000O10000000001O00000000000000001O000000000000000000000000000001O001O001O1O001O001O1O001O001O1O001O001O1O001O001SJbL"}, "label": "vase with small blue flowers"}]} {"id": 427805, "image": "COCO_train2014_000000427805.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass vase to the right of two other vases\"."}], "task": "refering", "answer_template": "The \"a glass vase to the right of two other vases\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [239, 240, 241, 261, 262, 263, 282, 283, 284, 285, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 346, 347, 348, 349, 350, 351, 368, 369, 370, 371, 372, 373, 390, 391, 392, 393, 394, 395, 412, 413, 414, 415, 416, 417, 434, 435, 436, 437, 438, 439, 456, 457, 458, 459, 460, 461, 481, 482, 483], "bbox": [0.72359477124183, 0.4629248366013072, 1.0, 0.9640522875816994], "iscrowd": 0, "area": 40153.8103, "rle": {"size": [612, 612], "counts": "gVY87ib0:E:G:F:E:G:F:E;F9G:E;G8J7I7I4L2N3M2N2O1N2N2N3M2N2N2N2N2N3M2N2N2N2N2O2M2N2N2N2N2N1O1O1O1O1O2N100O1O1O1O2N1O1O1O1O1O1O2N100O1O1O1O2N1O1O1O1O1O1O2M2O1O1O1N2O2N1O1N2O1O1N2O2\\Oc0N2N2N2N2N2N2N3M2N2N2N2N2N2O1O101N10000O10000O101N10000O10000000001O00000000000000001O000000000000000000000000000001O001O001O1O001O001O1O001O001O1O001O001O1O001O001SJbL"}, "label": "a glass vase to the right of two other vases"}]} {"id": 427805, "image": "COCO_train2014_000000427805.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass vase with some white flowers in it\"."}], "task": "refering", "answer_template": "The \"a glass vase with some white flowers in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 308, 309, 310, 311, 312, 313, 314, 330, 331, 332, 333, 334, 335, 336, 352, 353, 354, 355, 356, 357, 358, 359, 374, 375, 376, 377, 378, 379, 380, 381, 396, 397, 398, 399, 400, 401, 402, 418, 419, 420, 421, 422, 423, 424, 441, 442, 443, 444, 445, 446, 465, 466], "bbox": [0.013513071895424836, 0.558562091503268, 0.3243300653594771, 0.9684640522875818], "iscrowd": 0, "area": 37909.09425, "rle": {"size": [612, 612], "counts": "^Y52kb08G9H8H8H8G9H8H8H8G9H8H8H8H8G8I8H8H8G9H8H8H8H8G9L4O1O1O1N2O1O1O1O1O1O001O1O001O1O001O1N2O001O1O001O1O001O1O001O1O001N2O001O1O001O1O001O1O1O001O1O0O2O1O001O1O001O1O00O10000000O0100000000O100000000O10000000O0100000000O10000O10000O100000O10O10000000000O1000001O0O2O1O1O1O1O1O001N2O1O1O1N2O001O1N2O1O1O0O2O1O1N2O1O1O3L3N2N3M2M3N3M2M3N3M2N2M4J5K6J5J6K6J6J6K5K5K5J7J5K5K5K7I7EC[of7"}, "label": "a glass vase with some white flowers in it"}]} {"id": 427805, "image": "COCO_train2014_000000427805.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the glass vase with a lot purple flowers\"."}], "task": "refering", "answer_template": "The \"the glass vase with a lot purple flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 308, 309, 310, 311, 312, 313, 314, 330, 331, 332, 333, 334, 335, 336, 352, 353, 354, 355, 356, 357, 358, 359, 374, 375, 376, 377, 378, 379, 380, 381, 396, 397, 398, 399, 400, 401, 402, 418, 419, 420, 421, 422, 423, 424, 441, 442, 443, 444, 445, 446, 465, 466], "bbox": [0.013513071895424836, 0.558562091503268, 0.3243300653594771, 0.9684640522875818], "iscrowd": 0, "area": 37909.09425, "rle": {"size": [612, 612], "counts": "^Y52kb08G9H8H8H8G9H8H8H8G9H8H8H8H8G8I8H8H8G9H8H8H8H8G9L4O1O1O1N2O1O1O1O1O1O001O1O001O1O001O1N2O001O1O001O1O001O1O001O1O001N2O001O1O001O1O001O1O1O001O1O0O2O1O001O1O001O1O00O10000000O0100000000O100000000O10000000O0100000000O10000O10000O100000O10O10000000000O1000001O0O2O1O1O1O1O1O001N2O1O1O1N2O001O1N2O1O1O0O2O1O1N2O1O1O3L3N2N3M2M3N3M2M3N3M2N2M4J5K6J5J6K6J6J6K5K5K5J7J5K5K5K7I7EC[of7"}, "label": "the glass vase with a lot purple flowers"}]} {"id": 427805, "image": "COCO_train2014_000000427805.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vase that the pink flowers are sitting in\"."}], "task": "refering", "answer_template": "The \"the vase that the pink flowers are sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [298, 299, 316, 319, 320, 321, 322, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 382, 383, 384, 385, 386, 387, 388, 404, 405, 406, 407, 408, 409, 410, 426, 427, 428, 429, 430, 431, 432, 448, 449, 450, 451, 452, 453], "bbox": [0.3686111111111111, 0.5913888888888889, 0.6598202614379085, 0.9503431372549019], "iscrowd": 0, "area": 28390.06455, "rle": {"size": [612, 612], "counts": "\\_W4d0\\b0g0YOg0YOd0^Oa0B9G2N3N2M3M2N3N2M1N2N2N2O0O2N2N2N2N100O1O1O001O10O01O1O1O001O1O00100O10O0001O0001O01O0000000O2O0000001O000000001O01O0001O01O000000O10O100000O10000O100O100O100O10000O100O1O1O100O1O100O1O1O100O1O1O100O100O00100O100M3M2M4E;H8000O10001O001O0O100000O10O10O010O0100O010O01M2M40O2O2M2N3M3N1N3M2N3M2N3M3M2N3N1N3M3M2N3M2N3M2M4L4L3M4M2M5K6J6J5K6J6J6J`0YOg0XOT1lNUkk3"}, "label": "the vase that the pink flowers are sitting in"}]} {"id": 427805, "image": "COCO_train2014_000000427805.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a vase with red and pink flowers in it\"."}], "task": "refering", "answer_template": "The \"a vase with red and pink flowers in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [298, 299, 316, 319, 320, 321, 322, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 382, 383, 384, 385, 386, 387, 388, 404, 405, 406, 407, 408, 409, 410, 426, 427, 428, 429, 430, 431, 432, 448, 449, 450, 451, 452, 453], "bbox": [0.3686111111111111, 0.5913888888888889, 0.6598202614379085, 0.9503431372549019], "iscrowd": 0, "area": 28390.06455, "rle": {"size": [612, 612], "counts": "\\_W4d0\\b0g0YOg0YOd0^Oa0B9G2N3N2M3M2N3N2M1N2N2N2O0O2N2N2N2N100O1O1O001O10O01O1O1O001O1O00100O10O0001O0001O01O0000000O2O0000001O000000001O01O0001O01O000000O10O100000O10000O100O100O100O10000O100O1O1O100O1O100O1O1O100O1O1O100O100O00100O100M3M2M4E;H8000O10001O001O0O100000O10O10O010O0100O010O01M2M40O2O2M2N3M3N1N3M2N3M2N3M3M2N3N1N3M3M2N3M2N3M2M4L4L3M4M2M5K6J6J5K6J6J6J`0YOg0XOT1lNUkk3"}, "label": "a vase with red and pink flowers in it"}]} {"id": 329502, "image": "COCO_train2014_000000329502.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting on a chair\"."}], "task": "refering", "answer_template": "The \"a man sitting on a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 66, 67, 68, 107, 108, 109, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 197, 198], "bbox": [0.590671875, 0.009875, 1.0, 0.5336875], "iscrowd": 0, "area": 13195.963750000003, "rle": {"size": [480, 640], "counts": "j[a5?Q>g0I6J7H8I5K3L3N3M3L4M3M2M4M00O001O1O001O1O001O1O001O0O1O100O10O0100O1O10O0100O1O01O002O0O1O2O0O100100O001O10O01O10O01O1O10O01O1O010O1O010O1O00100OYMkMdHU2\\7lMdHT2[7mMeHR2Z7oMgHP2Y7oMiHP2W7oMkHQ2T7oMmHP2S7PNnHo1R7RNnHm1R7TNnHl1Q7TNPIk1Q7UNoHj1R7VNnHj1Q7WNTFKh2m1U7WNTFMf2k1V7YNSFOd2h1Z7YNRF0c2g1[7YNQF1d2e1[7ZNRF1c2e1[7ZNQF2d2d1[7ZNQF2d2d1Z7dNfH\\1Z7ZNRF3d2c1Z7dNfHW1^7jNbHl0h7SOYHc0Q8]OoG:Y8GgGOc81]GEm8;l200000000O100O100O100O100O2O000000000000000000000000000000000000010O000001O00001O0000001O0000001O00001O0001O01O00001O0000001O00001O00001O001O001bB[O_QCCo<=PCDP==nBDR= in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 66, 67, 68, 107, 108, 109, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 197, 198], "bbox": [0.590671875, 0.009875, 1.0, 0.5336875], "iscrowd": 0, "area": 13195.963750000003, "rle": {"size": [480, 640], "counts": "j[a5?Q>g0I6J7H8I5K3L3N3M3L4M3M2M4M00O001O1O001O1O001O1O001O0O1O100O10O0100O1O10O0100O1O01O002O0O1O2O0O100100O001O10O01O10O01O1O10O01O1O010O1O010O1O00100OYMkMdHU2\\7lMdHT2[7mMeHR2Z7oMgHP2Y7oMiHP2W7oMkHQ2T7oMmHP2S7PNnHo1R7RNnHm1R7TNnHl1Q7TNPIk1Q7UNoHj1R7VNnHj1Q7WNTFKh2m1U7WNTFMf2k1V7YNSFOd2h1Z7YNRF0c2g1[7YNQF1d2e1[7ZNRF1c2e1[7ZNQF2d2d1[7ZNQF2d2d1Z7dNfH\\1Z7ZNRF3d2c1Z7dNfHW1^7jNbHl0h7SOYHc0Q8]OoG:Y8GgGOc81]GEm8;l200000000O100O100O100O100O2O000000000000000000000000000000000000010O000001O00001O0000001O0000001O00001O0001O01O00001O0000001O00001O00001O001O001bB[O_QCCo<=PCDP==nBDR= in this image.", "objects": [{"patches": [10, 11, 12, 13, 33, 34, 35, 36, 55, 56, 57, 58, 59, 79, 80, 81, 82, 101, 102, 103, 104, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 308, 309, 310, 311, 312, 331, 332, 333, 334, 356, 357, 358, 379, 380], "bbox": [0.358984375, 0.013479166666666665, 0.62190625, 0.9685416666666667], "iscrowd": 0, "area": 48543.78679999999, "rle": {"size": [480, 640], "counts": "^Q\\38a>:E;E;XOi0TOk0L3L5L4K5L4L3L5L4K7J6J7H7J6J5J7J5J7J6J5J5L4K4M4L4K5L3M4K5L4K4M4K6J6K6I7I6J6K4L5bNRGaLS9U3UGiLo8n2XGoLm8g2ZGWMj8a2]G\\Mh8[2_GcMa8W2gGfMY8Y2jGdMX8[2kGaMW8]2mG`MT8^2QH]MP8c2THXMn7f2W2O1N101N2N2O1N3N1N3M2O2M201O00100O1O3M3M6K5J6J6J1O01O01O001O0O101O001O00001N101O00001O0O2O00O100O100O010O100O1XJTO\\Ml0d2UOZMk0f2XOXMh0hMWOa13d0g0hMDX1Gm0g0hMFZ1Ej0g0jMG^1@f0j0iMJb1\\Oc0j0iMMe1XO`0k0iM1i1SOf2bNeLl0cN9o1;j2bN]Lo0fN7P2;m2gNQLk0oN7P2:P3nNcKd2Z1cNQ3SOWK_2e1dNQ3S3kLTMR3l2jL[MS3Y3VLmLh3X3nKoLo3X3gKdM_3R7M3M3M3M3K5J6K5K5J6K5K5J6K4L6I9H9G6I6K5K5J6J[Ta3"}, "label": "a toddler in gray pants and a striped shirt"}]} {"id": 329502, "image": "COCO_train2014_000000329502.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy wearing a striped shirt and gray pants\"."}], "task": "refering", "answer_template": "The \"a boy wearing a striped shirt and gray pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 33, 34, 35, 36, 55, 56, 57, 58, 59, 79, 80, 81, 82, 101, 102, 103, 104, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 308, 309, 310, 311, 312, 331, 332, 333, 334, 356, 357, 358, 379, 380], "bbox": [0.358984375, 0.013479166666666665, 0.62190625, 0.9685416666666667], "iscrowd": 0, "area": 48543.78679999999, "rle": {"size": [480, 640], "counts": "^Q\\38a>:E;E;XOi0TOk0L3L5L4K5L4L3L5L4K7J6J7H7J6J5J7J5J7J6J5J5L4K4M4L4K5L3M4K5L4K4M4K6J6K6I7I6J6K4L5bNRGaLS9U3UGiLo8n2XGoLm8g2ZGWMj8a2]G\\Mh8[2_GcMa8W2gGfMY8Y2jGdMX8[2kGaMW8]2mG`MT8^2QH]MP8c2THXMn7f2W2O1N101N2N2O1N3N1N3M2O2M201O00100O1O3M3M6K5J6J6J1O01O01O001O0O101O001O00001N101O00001O0O2O00O100O100O010O100O1XJTO\\Ml0d2UOZMk0f2XOXMh0hMWOa13d0g0hMDX1Gm0g0hMFZ1Ej0g0jMG^1@f0j0iMJb1\\Oc0j0iMMe1XO`0k0iM1i1SOf2bNeLl0cN9o1;j2bN]Lo0fN7P2;m2gNQLk0oN7P2:P3nNcKd2Z1cNQ3SOWK_2e1dNQ3S3kLTMR3l2jL[MS3Y3VLmLh3X3nKoLo3X3gKdM_3R7M3M3M3M3K5J6K5K5J6K5K5J6K4L6I9H9G6I6K5K5J6J[Ta3"}, "label": "a boy wearing a striped shirt and gray pants"}]} {"id": 329502, "image": "COCO_train2014_000000329502.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man kneeling on floor\"."}], "task": "refering", "answer_template": "The \"man kneeling on floor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 29, 30, 31, 32, 51, 52, 53, 54, 55, 73, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 235, 236, 237, 238, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 305], "bbox": [0.17825, 0.0, 0.46076562499999996, 0.7950625], "iscrowd": 0, "area": 38840.3496, "rle": {"size": [480, 640], "counts": "hde15e>8G:G9G9G9H8G9H8H8G8I8G9H8H7H7J7I7H8I7H8I7I7L3N3M3L4M3M3M3M3M2M4M3M3M3M3L4M3M2N3M3M3L4M3M3M3M3L9H7SJkGb4]8iJRHU5o8M2M4M2N1O1O1O2M2O1O2N1O1O1O2N1O1O1O1O1O001O00001O001O00001O00O1M3M3UOgKmF\\4T9fKgF\\4[9eK_F^4b9dKXF_4j9:1O2N1O1A`0O001O01O01O001O001O00001O001O001OgKPLbMP4Z2ZL`Me3]2fL]MY3_2QM[Mo2a2[MZMd2b2gMWMY2e2QNUMo1g2[NTMd1h2aNWM_1e2fN[MY1`2nN^MR1^2SOaMm0]2VOcMi0Z2[OeMe0Y2^OfMb0X2BgM=W2FhM:U2JkM5S2NlM2R22lMNR25nMJP29oMGn1=RNBl1U1_MkN_2\\1\\MdNb2a1\\M^Nb2h1ZMXNc2n1[MQNc2U2YMkMe2Z2XMfMf2_2XM`Me2g2WMYMg2l2WMSMh2Q3UMoLi2W3SMiLk2\\3SMcLk2c3QM]Lm2h3QMWLn2m3oLSLo2S4mLmKQ3X4mLgKQ3_4kLaKS3W8O1N2N2N2N2kNlC]OU<`0SDYOo;c0YDWOi;f0_DSOc;i0fDPO\\;m0lDmNU;P1_1M3L4M3M3M3M3M3N2O1O1N2O1N2OQVR5"}, "label": "man kneeling on floor"}]} {"id": 329502, "image": "COCO_train2014_000000329502.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black t - shirt\"."}], "task": "refering", "answer_template": "The \"a man in a black t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 29, 30, 31, 32, 51, 52, 53, 54, 55, 73, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 235, 236, 237, 238, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 305], "bbox": [0.17825, 0.0, 0.46076562499999996, 0.7950625], "iscrowd": 0, "area": 38840.3496, "rle": {"size": [480, 640], "counts": "hde15e>8G:G9G9G9H8G9H8H8G8I8G9H8H7H7J7I7H8I7H8I7I7L3N3M3L4M3M3M3M3M2M4M3M3M3M3L4M3M2N3M3M3L4M3M3M3M3L9H7SJkGb4]8iJRHU5o8M2M4M2N1O1O1O2M2O1O2N1O1O1O2N1O1O1O1O1O001O00001O001O00001O00O1M3M3UOgKmF\\4T9fKgF\\4[9eK_F^4b9dKXF_4j9:1O2N1O1A`0O001O01O01O001O001O00001O001O001OgKPLbMP4Z2ZL`Me3]2fL]MY3_2QM[Mo2a2[MZMd2b2gMWMY2e2QNUMo1g2[NTMd1h2aNWM_1e2fN[MY1`2nN^MR1^2SOaMm0]2VOcMi0Z2[OeMe0Y2^OfMb0X2BgM=W2FhM:U2JkM5S2NlM2R22lMNR25nMJP29oMGn1=RNBl1U1_MkN_2\\1\\MdNb2a1\\M^Nb2h1ZMXNc2n1[MQNc2U2YMkMe2Z2XMfMf2_2XM`Me2g2WMYMg2l2WMSMh2Q3UMoLi2W3SMiLk2\\3SMcLk2c3QM]Lm2h3QMWLn2m3oLSLo2S4mLmKQ3X4mLgKQ3_4kLaKS3W8O1N2N2N2N2kNlC]OU<`0SDYOo;c0YDWOi;f0_DSOc;i0fDPO\\;m0lDmNU;P1_1M3L4M3M3M3M3M3N2O1O1N2O1N2OQVR5"}, "label": "a man in a black t - shirt"}]} {"id": 370461, "image": "COCO_train2014_000000370461.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra behind another zebra\"."}], "task": "refering", "answer_template": "The \"a zebra behind another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 246, 247, 249, 269, 270, 272, 273, 293, 295], "bbox": [0.717828125, 0.435211970074813, 0.94115625, 0.8738902743142145], "iscrowd": 0, "area": 12999.282199999996, "rle": {"size": [401, 640], "counts": "lnc52]"}, "label": "a zebra behind another zebra"}]} {"id": 370461, "image": "COCO_train2014_000000370461.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zibra with one zibera\"."}], "task": "refering", "answer_template": "The \"zibra with one zibera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 119, 120, 121, 122, 123, 124, 126, 127, 128, 129, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 211, 212, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 237, 238, 239, 243, 244, 245, 246, 261, 262, 267, 268, 284, 285, 290, 291, 307, 308, 313, 314], "bbox": [0.16924999999999998, 0.31169576059850373, 0.719328125, 0.9844139650872819], "iscrowd": 0, "area": 45406.06555000004, "rle": {"size": [401, 640], "counts": "QaZ11[<8H9G8G:G8F;F9F;E7I7I7I7I7I7I7I7L400O00100O100O100O1O100O100O1O100O10N1O2N1O1O2N101N2N2N2N2N3M2N2N2N200O1O100O2N1O100O1O100O1O1O2O01O001O001O001O01O01O001O001O001O00100O2N1O2N1O2N2O0O2N101N2N1O2O0O2N3N1N3M2N3N1N2N3N1N3N1N3N1N3N1O2M2YKkJo1_5_MhJa2j5gL\\JZ3V6mKPJS4W71O001O001O001O001O001O001O1O001O001O001O001O001O001O001O0O2O1O0bLlJ^OV5a0_KjNa4V1TLUNm3j1gLaMZ3_2[MlLf2S3nMYLR2g3]3O001N101O001O001O1O001O0O2O001O001O001O001N101O000000O10000000O10O1000000000000O10O10000000O1000000000O10O10000000000O10O10000000O1000O10000000O0100000000O010000000O1000O10O10000000O0100000000O10O1000O100000O201N2N1O2O1N2N2N101N2N2N2O1N1O2N2O1N2N101N2N2N2O1N1O2N2O1N6J\\E^Oa:b0fEZOW:f0k0001O000000N2H8HYWV2"}, "label": "zibra with one zibera"}]} {"id": 91933, "image": "COCO_train2014_000000091933.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bench facing the mountains\"."}], "task": "refering", "answer_template": "The \"bench facing the mountains\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 242, 243, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 271, 272, 273, 274, 275, 276, 277, 278, 279, 286, 287, 293, 301, 302, 308, 322, 323, 338], "bbox": [0.05096244131455399, 0.718421875, 0.6657981220657276, 0.9623750000000001], "iscrowd": 0, "area": 14396.209449999998, "rle": {"size": [640, 426], "counts": "kV>3lc06J5K3N1N2N2N2O1N2N2N2O0O2O00001O001O0O101O001O00001N10S1mN^1bN001O001N101OO1O100O1O1O001O1O100O1O1O1\\Od0UOk0]Oc0001O00001O0O2O000010O01O01O001O01O01O000010O0001O0010O0001O00010O00001O010O00001O00001O00010O001O0000001O0001O00000000000000000010O000000000000000001O00000001O00000000000000000000000010O000000000000000000000000000000001O000000000000000000001O000000000000000000001O000000000000000000000000O1000000O10X@bNn<^1eBoN[=Q1XB\\Oh=c0lAJT>6`A6`>JSAb0n>^Oe@o0[?POZ@\\1f?_10000000000000O1G9F:F:F:F:E;F:F:G900O100O101O0O2O0O2O001N101N10001N101N101O0O2O0O2O000O2O0O2O0O2O1O4K6I6JXXh2"}, "label": "bench facing the mountains"}]} {"id": 517920, "image": "COCO_train2014_000000517920.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"car parked furthest from meters\"."}], "task": "refering", "answer_template": "The \"car parked furthest from meters\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 45, 46, 47, 48, 52, 53, 60, 61, 62, 75, 76, 77, 90, 91, 92, 105, 106, 107, 120, 121, 122, 135, 136, 137, 150, 151, 152, 165, 166, 167, 181, 182, 196], "bbox": [0.0018266978922716628, 0.033218750000000005, 0.8671662763466041, 0.6001875], "iscrowd": 0, "area": 33818.7317, "rle": {"size": [640, 427], "counts": "gg0i1Ub0a2_MZ1fN4L4L4L3M4L3M3M3M2N3M3M3M2N3M3M3M2N3M3M3M2N3M3M3M2N3L4M7H8I8SEcGi9]9K5L4K5L4N2N2M3N2N1N100000000000000O1000N2M3M3M3M3M3M3L3N3M3M3I7D in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 45, 46, 47, 48, 52, 53, 60, 61, 62, 75, 76, 77, 90, 91, 92, 105, 106, 107, 120, 121, 122, 135, 136, 137, 150, 151, 152, 165, 166, 167, 181, 182, 196], "bbox": [0.0018266978922716628, 0.033218750000000005, 0.8671662763466041, 0.6001875], "iscrowd": 0, "area": 33818.7317, "rle": {"size": [640, 427], "counts": "gg0i1Ub0a2_MZ1fN4L4L4L3M4L3M3M3M2N3M3M3M2N3M3M3M2N3M3M3M2N3M3M3M2N3L4M7H8I8SEcGi9]9K5L4K5L4N2N2M3N2N1N100000000000000O1000N2M3M3M3M3M3M3L3N3M3M3I7D in this image.", "objects": [{"patches": [39, 40, 41, 42, 43, 44, 53, 54, 55, 56, 57, 58, 59, 68, 69, 70, 71, 72, 73, 74, 83, 84, 85, 86, 87, 88, 89, 98, 99, 100, 101, 102, 103, 104, 113, 114, 115, 116, 117, 118, 119, 128, 129, 130, 131, 132, 133, 134, 143, 144, 145, 146, 147, 148, 149, 158, 159, 160, 161, 162, 163, 164, 173, 174, 175, 176, 177, 178, 179, 188, 189, 190, 191, 192, 193, 194, 203, 204, 205, 206, 207, 208, 209, 218, 219, 220, 221, 222, 223, 224, 232, 233, 234, 235, 236, 237, 238, 239, 247, 248, 249, 250, 251, 252, 253, 254, 262, 263, 264, 265, 266, 267, 268, 269, 276, 277, 278, 279, 280, 281, 282, 283, 284, 292, 293, 294, 295, 296, 297, 298, 299, 308, 309, 310, 311, 312, 313, 314, 324, 325, 326, 327, 328, 329, 340, 341, 342, 343, 344], "bbox": [0.4547072599531616, 0.098875, 1.0, 0.9887656249999999], "iscrowd": 0, "area": 106410.6388, "rle": {"size": [640, 427], "counts": "lgi31mc03M3M3M3M3XNEg_O>W`0N\\_O4d`07o^OLo`0`0d^OC[a0b1N2O1N2O0O2O1N2O1N2O1N101N2O1N2N2L4L3M4M3L4L4M3L4L4aJZL`Ji3P4gMlK]2`1gKnKW2]2V2c1gKiKX2a2T2e1hKaK[2e2Q2h1hK\\K\\2h2Q2i1gKWK^2m2n1k1hKPK`2P3l1n1hKjJb2T3j1P2hKeJd2W3g1S2iK]Jf2\\3e1U2iKXJh2^3c1X2GdM=[2H^M;`2MVM7i21lL3R36dLMZ3;\\LIc3?RLEl3c0kK@S4i0bK[O]4m0XKWOf4Q1QKSOl4R8O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2N3N1N2O1O1O1O1O1O1O1N3N2N1O2N2N1O2M3N1O2N1O2N2N1N3N2N1O2N2N1O2N2M2O2N000000000O1000000000000O100000000000000O100000000000O01000000000000O100000000000000O1000000000000O1000000000000O100000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000O100000O1000000000000000QN"}, "label": "a carp arked next to a parking meter"}]} {"id": 517920, "image": "COCO_train2014_000000517920.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the car closes to the parking meter\"."}], "task": "refering", "answer_template": "The \"the car closes to the parking meter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 42, 43, 44, 53, 54, 55, 56, 57, 58, 59, 68, 69, 70, 71, 72, 73, 74, 83, 84, 85, 86, 87, 88, 89, 98, 99, 100, 101, 102, 103, 104, 113, 114, 115, 116, 117, 118, 119, 128, 129, 130, 131, 132, 133, 134, 143, 144, 145, 146, 147, 148, 149, 158, 159, 160, 161, 162, 163, 164, 173, 174, 175, 176, 177, 178, 179, 188, 189, 190, 191, 192, 193, 194, 203, 204, 205, 206, 207, 208, 209, 218, 219, 220, 221, 222, 223, 224, 232, 233, 234, 235, 236, 237, 238, 239, 247, 248, 249, 250, 251, 252, 253, 254, 262, 263, 264, 265, 266, 267, 268, 269, 276, 277, 278, 279, 280, 281, 282, 283, 284, 292, 293, 294, 295, 296, 297, 298, 299, 308, 309, 310, 311, 312, 313, 314, 324, 325, 326, 327, 328, 329, 340, 341, 342, 343, 344], "bbox": [0.4547072599531616, 0.098875, 1.0, 0.9887656249999999], "iscrowd": 0, "area": 106410.6388, "rle": {"size": [640, 427], "counts": "lgi31mc03M3M3M3M3XNEg_O>W`0N\\_O4d`07o^OLo`0`0d^OC[a0b1N2O1N2O0O2O1N2O1N2O1N101N2O1N2N2L4L3M4M3L4L4M3L4L4aJZL`Ji3P4gMlK]2`1gKnKW2]2V2c1gKiKX2a2T2e1hKaK[2e2Q2h1hK\\K\\2h2Q2i1gKWK^2m2n1k1hKPK`2P3l1n1hKjJb2T3j1P2hKeJd2W3g1S2iK]Jf2\\3e1U2iKXJh2^3c1X2GdM=[2H^M;`2MVM7i21lL3R36dLMZ3;\\LIc3?RLEl3c0kK@S4i0bK[O]4m0XKWOf4Q1QKSOl4R8O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2N3N1N2O1O1O1O1O1O1O1N3N2N1O2N2N1O2M3N1O2N1O2N2N1N3N2N1O2N2N1O2N2M2O2N000000000O1000000000000O100000000000000O100000000000O01000000000000O100000000000000O1000000000000O1000000000000O100000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000O100000O1000000000000000QN"}, "label": "the car closes to the parking meter"}]} {"id": 223023, "image": "COCO_train2014_000000223023.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bathtub\"."}], "task": "refering", "answer_template": "The \"a bathtub\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 230, 231, 232, 233, 254, 255, 256, 277, 278], "bbox": [0.0, 0.3919014084507042, 0.201765625, 0.8396244131455398], "iscrowd": 0, "area": 15902.584499999999, "rle": {"size": [426, 640], "counts": "b5a0i in this image.", "objects": [{"patches": [138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 230, 231, 232, 233, 254, 255, 256, 277, 278], "bbox": [0.0, 0.3919014084507042, 0.201765625, 0.8396244131455398], "iscrowd": 0, "area": 15902.584499999999, "rle": {"size": [426, 640], "counts": "b5a0i in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 56, 74, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 193, 194, 213, 214, 215, 216, 236, 237, 238, 239, 259, 260, 261, 262, 263, 282, 283, 284, 285, 286, 305, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332], "bbox": [0.219875, 0.1573004694835681, 0.48464062499999994, 0.9865258215962441], "iscrowd": 0, "area": 36100.80095, "rle": {"size": [426, 640], "counts": "fgj11Y=2N2M2O2N2N2M2O2N2N1N3N2N2N1N3N2N2N1N3N2N1O2M3N2`EcNo8^1mFgNQ9[1jFiNa1Eo4d1ZIlNd1Do4b1YIoNd1BR5a1UIQOf1AT5`1PITOi1@U5^1mHWOj1^OW5]1kHYOk1]OY5h2eJZMZ5h2bJ\\M\\5e2aJ^M^5c2_J`M`5b2]J`Mb5a2\\JbMP1b0;m1cNcMl0e0`0k1aNbMj0h0d0g1`NdMe0j0k0c1^NeMd0k0m0b1\\NeMf0i0n0c1[NeMf0i0n0c1ZNgMe0g0Q1d1WNgMg0f0Q1c1WNiMf0e0R1c1WNjMf0d0R1c1VNkMg0b0S1c1UNlMg0b0T1b1UNmMf0a0U1b1TNoMe0a0V1`1TNQNe0?W1`1SNSNd0>X1a1RNSNe0_NcNEi0l1a0bNeNHe0f1c0eNfNGg0d1`0hNhNFg0b1=lNjNDi0`1:oNlNBj0_17ROmNAl0]14UOoN_Om0\\12WOoN_Oo0Z12WOmNAQ1X11XOkNCT1U11XOiNFU1R12XOgNHW1P11YOfNIY1n01YOcNL\\1l00ZO`NL`1j0O]O]NKd1h0O@XNJi1f0OBSNKn1c0NFoMIS2a0NHkMJV2>OJgMJZ2 in this image.", "objects": [{"patches": [51, 52, 53, 54, 55, 56, 74, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 193, 194, 213, 214, 215, 216, 236, 237, 238, 239, 259, 260, 261, 262, 263, 282, 283, 284, 285, 286, 305, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332], "bbox": [0.219875, 0.1573004694835681, 0.48464062499999994, 0.9865258215962441], "iscrowd": 0, "area": 36100.80095, "rle": {"size": [426, 640], "counts": "fgj11Y=2N2M2O2N2N2M2O2N2N1N3N2N2N1N3N2N2N1N3N2N1O2M3N2`EcNo8^1mFgNQ9[1jFiNa1Eo4d1ZIlNd1Do4b1YIoNd1BR5a1UIQOf1AT5`1PITOi1@U5^1mHWOj1^OW5]1kHYOk1]OY5h2eJZMZ5h2bJ\\M\\5e2aJ^M^5c2_J`M`5b2]J`Mb5a2\\JbMP1b0;m1cNcMl0e0`0k1aNbMj0h0d0g1`NdMe0j0k0c1^NeMd0k0m0b1\\NeMf0i0n0c1[NeMf0i0n0c1ZNgMe0g0Q1d1WNgMg0f0Q1c1WNiMf0e0R1c1WNjMf0d0R1c1VNkMg0b0S1c1UNlMg0b0T1b1UNmMf0a0U1b1TNoMe0a0V1`1TNQNe0?W1`1SNSNd0>X1a1RNSNe0_NcNEi0l1a0bNeNHe0f1c0eNfNGg0d1`0hNhNFg0b1=lNjNDi0`1:oNlNBj0_17ROmNAl0]14UOoN_Om0\\12WOoN_Oo0Z12WOmNAQ1X11XOkNCT1U11XOiNFU1R12XOgNHW1P11YOfNIY1n01YOcNL\\1l00ZO`NL`1j0O]O]NKd1h0O@XNJi1f0OBSNKn1c0NFoMIS2a0NHkMJV2>OJgMJZ2 in this image.", "objects": [{"patches": [149, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 242, 243, 244, 265, 266, 267], "bbox": [0.49770312499999997, 0.4320422535211268, 0.6629375, 0.7891079812206573], "iscrowd": 0, "area": 10686.214950000001, "rle": {"size": [426, 640], "counts": "elT4i0a<5K6J5K5K6J4L3M2N3M2QEVN[:e2LM2NN2O1N101O1N2M2M4M3G9L4N1O2O0004LO2N1O1O1O2N100O1O2N1CRL[GP4d8RLYGP4f8RLWGQ4g8901O000O13M2O2M2N1O1O1O100OO1O100O2N1ChKgGY4Y8iKdGX4]8iK_GY4b862N1O2N2N2O1N1O2N2N2O1aNWGZNk8d1^101O1N2O1O1O1N2O1O1N2O1O1O2M2N2N2N2M4M3M3M3L5L3M3M3L4J[di2"}, "label": "a bidet next to a toilet"}]} {"id": 223023, "image": "COCO_train2014_000000223023.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bidet beside a toilet\"."}], "task": "refering", "answer_template": "The \"a bidet beside a toilet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 152, 172, 173, 174, 175, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 242, 243, 244, 265, 266, 267], "bbox": [0.49770312499999997, 0.4320422535211268, 0.6629375, 0.7891079812206573], "iscrowd": 0, "area": 10686.214950000001, "rle": {"size": [426, 640], "counts": "elT4i0a<5K6J5K5K6J4L3M2N3M2QEVN[:e2LM2NN2O1N101O1N2M2M4M3G9L4N1O2O0004LO2N1O1O1O2N100O1O2N1CRL[GP4d8RLYGP4f8RLWGQ4g8901O000O13M2O2M2N1O1O1O100OO1O100O2N1ChKgGY4Y8iKdGX4]8iK_GY4b862N1O2N2N2O1N1O2N2N2O1aNWGZNk8d1^101O1N2O1O1O1N2O1O1N2O1O1O2M2N2N2N2M4M3M3M3L5L3M3M3L4J[di2"}, "label": "a bidet beside a toilet"}]} {"id": 509746, "image": "COCO_train2014_000000509746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a red tie hugging a girl\"."}], "task": "refering", "answer_template": "The \"a man in a red tie hugging a girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 39, 40, 41, 42, 43, 44, 61, 62, 63, 64, 65, 66, 67, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390, 405, 406, 407, 408, 409, 410, 411, 412, 413, 429, 430, 431, 432, 433, 434, 435, 436], "bbox": [0.47945312500000004, 0.024274952919020716, 1.0, 0.9881544256120528], "iscrowd": 0, "area": 104047.56374999999, "rle": {"size": [531, 640], "counts": "Xao45W`07K6N1O2O0O1O2N101N1O1O2O0O2N100O2N1O2O0O1O2O0O2N1O101N1O2N100O2N101N1O1O2O0O1O2O0O2N1O101M2O2N1O1O2N1O2N1O2N2M3N2N1O2N2N2N2N2N2N2M3N2N2N2N2N2N1O2N2N2M3N2N2N2N2N2N2N2N2N2M2O3M4L4L5K4L5K4L4L5J5L5K4L5K:F:F;E:F:F9G3N1O1O1N2O1O1N2O2N1N2O1O1N2O1O2M2O1O000O10000O10000O100lJSGb2m8lLgGQ3Z8nLiGP3W8PMkGn2U8QMnGm2R8SMPHj2Q8UMRHi2n7VMUHh2k7XMWHe2j7ZMYHd2g7\\M[Hb2e7]M^H`2c7_M`H_2`7aMbH]2^7bMfHZ2[7eMQIP2o6PN]Id1c6[NiIZ1W6fNPJR1Q6mNSJo0n5POWJl0i5TOZJh0g5WO]Jf0c5YOaJc0`5]OcJ`0]5_OgJ=Z5CiJ9X5FlJ7T5HPK4Q5LQK2o4MUKOl40WKNi42YKKh45[KGf49\\KEd4:`KBa4>aK@_4`0cK]O^4c0eKZO[4f0gKWOZ4i0iKSOX4l0kKROU4n0nKnNS4R1g50000O10000000000O100000000O10000000000O100000000O10000000000O100000000O01000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O1000000001O1O1O1O1O1TDYOj7h0TH\\Oi7e0UH_Oh7b0UHDg7=WHGf7:XHKd76YHOd72ZH2c7O[H6a7K\\H:a7G]H>_7C_Ha0^7@_He0^7]O_Hh0]7ZO`Hj0]7XO_Hl0_7VO^Hl0a7VO\\Hm0b7UOZHn0d7UOXHo0f7SOVHP1i7ROSHR1k7POQHS1n7POmGT1P8oNlGT1S8oNiGT1U8SObGP1]8VO[Gm0d8ZOSGi0k8^OmFe0R9BeFa0Z9]3O1O1N2O1O1O1O1N2O1O1O1N2O1XNfHUJ[7Q5dIhJ^6^4bJ[K_5k3`KnKb4W3_LaLd3d2\\MRMj2U2W6@a0_O`0@a0^OeH"}, "label": "a man in a red tie hugging a girl"}]} {"id": 509746, "image": "COCO_train2014_000000509746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dorky dude she can do way better then\"."}], "task": "refering", "answer_template": "The \"the dorky dude she can do way better then\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 39, 40, 41, 42, 43, 44, 61, 62, 63, 64, 65, 66, 67, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 382, 383, 384, 385, 386, 387, 388, 389, 390, 405, 406, 407, 408, 409, 410, 411, 412, 413, 429, 430, 431, 432, 433, 434, 435, 436], "bbox": [0.47945312500000004, 0.024274952919020716, 1.0, 0.9881544256120528], "iscrowd": 0, "area": 104047.56374999999, "rle": {"size": [531, 640], "counts": "Xao45W`07K6N1O2O0O1O2N101N1O1O2O0O2N100O2N1O2O0O1O2O0O2N1O101N1O2N100O2N101N1O1O2O0O1O2O0O2N1O101M2O2N1O1O2N1O2N1O2N2M3N2N1O2N2N2N2N2N2N2M3N2N2N2N2N2N1O2N2N2M3N2N2N2N2N2N2N2N2N2M2O3M4L4L5K4L5K4L4L5J5L5K4L5K:F:F;E:F:F9G3N1O1O1N2O1O1N2O2N1N2O1O1N2O1O2M2O1O000O10000O10000O100lJSGb2m8lLgGQ3Z8nLiGP3W8PMkGn2U8QMnGm2R8SMPHj2Q8UMRHi2n7VMUHh2k7XMWHe2j7ZMYHd2g7\\M[Hb2e7]M^H`2c7_M`H_2`7aMbH]2^7bMfHZ2[7eMQIP2o6PN]Id1c6[NiIZ1W6fNPJR1Q6mNSJo0n5POWJl0i5TOZJh0g5WO]Jf0c5YOaJc0`5]OcJ`0]5_OgJ=Z5CiJ9X5FlJ7T5HPK4Q5LQK2o4MUKOl40WKNi42YKKh45[KGf49\\KEd4:`KBa4>aK@_4`0cK]O^4c0eKZO[4f0gKWOZ4i0iKSOX4l0kKROU4n0nKnNS4R1g50000O10000000000O100000000O10000000000O100000000O10000000000O100000000O01000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O10000000000O100000000O1000000001O1O1O1O1O1TDYOj7h0TH\\Oi7e0UH_Oh7b0UHDg7=WHGf7:XHKd76YHOd72ZH2c7O[H6a7K\\H:a7G]H>_7C_Ha0^7@_He0^7]O_Hh0]7ZO`Hj0]7XO_Hl0_7VO^Hl0a7VO\\Hm0b7UOZHn0d7UOXHo0f7SOVHP1i7ROSHR1k7POQHS1n7POmGT1P8oNlGT1S8oNiGT1U8SObGP1]8VO[Gm0d8ZOSGi0k8^OmFe0R9BeFa0Z9]3O1O1N2O1O1O1O1N2O1O1O1N2O1XNfHUJ[7Q5dIhJ^6^4bJ[K_5k3`KnKb4W3_LaLd3d2\\MRMj2U2W6@a0_O`0@a0^OeH"}, "label": "the dorky dude she can do way better then"}]} {"id": 509746, "image": "COCO_train2014_000000509746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the little girl with a black dress\"."}], "task": "refering", "answer_template": "The \"the little girl with a black dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 428], "bbox": [0.11187499999999999, 0.2921280602636535, 0.645109375, 0.9842749529190207], "iscrowd": 0, "area": 78895.07365, "rle": {"size": [531, 640], "counts": "]iU1X1[?>B000O10O100000000000O100000000000000N2O1O1N2O1N2O1N2O1N2O1N2O1WNlMQEV2d:XNXEi1^:dN]E^1X:POdEQ1R:\\OjEe0o9DmE>n9JnE7m91nE1m97oEJl9>oEDk9d0RF]Oi9k0RFWOi9k3L5K5J6K5K5K5K5J6K5K5K5K5K5K5N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2O100O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O100O1O1O100000000000000000O1000000000000000000000000000000000000000000000000000000000O10000000000000000000O10001O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O001O2N2M2PJbIi1`6SNcIl1^6QNeIn1]6nMfIQ2\\6jMhIU2Y6gMjIY2X6bMlI]2V6_MmI`2T6\\MPJc2R6XMRJg2o5UMTJk2n5PMVJo2l5lLXJS3i5hL\\JW3f5aLaJ^3a5YLfJg3[5VLhJi3Z5TLhJk3Y5RLjJm3X5PLiJP4Y5lKjJS4W5fKoJZ4S5^KSKb4o4VKXKi4i4PK]KP5e4hJaKX5n7O00001O0000001O00001O00001N10001O00001O0000001O00001O00001O00001O0000001O00001O001O00001O00001O001O00001O0O2O00001O0iNX1O01000O2O001O1N101O1O0O2O001N2O001O1N101O0O2O1O001N2O001O1N2O001N2O1O001N2O1O0O2O1O1O1N101O1O3Ln0SOl0TOQWe3"}, "label": "the little girl with a black dress"}]} {"id": 509746, "image": "COCO_train2014_000000509746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"female wearing dress and makeup\"."}], "task": "refering", "answer_template": "The \"female wearing dress and makeup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 428], "bbox": [0.11187499999999999, 0.2921280602636535, 0.645109375, 0.9842749529190207], "iscrowd": 0, "area": 78895.07365, "rle": {"size": [531, 640], "counts": "]iU1X1[?>B000O10O100000000000O100000000000000N2O1O1N2O1N2O1N2O1N2O1N2O1WNlMQEV2d:XNXEi1^:dN]E^1X:POdEQ1R:\\OjEe0o9DmE>n9JnE7m91nE1m97oEJl9>oEDk9d0RF]Oi9k0RFWOi9k3L5K5J6K5K5K5K5J6K5K5K5K5K5K5N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2O100O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O100O1O1O100000000000000000O1000000000000000000000000000000000000000000000000000000000O10000000000000000000O10001O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O001O2N2M2PJbIi1`6SNcIl1^6QNeIn1]6nMfIQ2\\6jMhIU2Y6gMjIY2X6bMlI]2V6_MmI`2T6\\MPJc2R6XMRJg2o5UMTJk2n5PMVJo2l5lLXJS3i5hL\\JW3f5aLaJ^3a5YLfJg3[5VLhJi3Z5TLhJk3Y5RLjJm3X5PLiJP4Y5lKjJS4W5fKoJZ4S5^KSKb4o4VKXKi4i4PK]KP5e4hJaKX5n7O00001O0000001O00001O00001N10001O00001O0000001O00001O00001O00001O0000001O00001O001O00001O00001O001O00001O0O2O00001O0iNX1O01000O2O001O1N101O1O0O2O001N2O001O1N101O0O2O1O001N2O001O1N2O001N2O1O001N2O1O0O2O1O1O1N101O1O3Ln0SOl0TOQWe3"}, "label": "female wearing dress and makeup"}]} {"id": 485173, "image": "COCO_train2014_000000485173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the head and shoulders and one leg of a goat closest to the dog\"."}], "task": "refering", "answer_template": "The \"the head and shoulders and one leg of a goat closest to the dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 262, 263, 264, 284, 285, 286, 287, 307, 308, 309, 310, 329, 330, 331, 332, 333, 352, 353, 354, 355, 356, 376, 377, 379], "bbox": [0.31521875, 0.435741127348643, 0.5576875, 0.9886638830897705], "iscrowd": 0, "area": 19235.641799999987, "rle": {"size": [479, 640], "counts": "lgn22l>3M2M3N2M3N3L3N2M3`FXOe4l0WKXOe4k0YKYOa4k0\\KYOX4G_GR1V4[Ol30mGg0T4^Ol3KoGj0R4^Om3GRHl0n3Am3DTHm0n3@m3BVHo0k3@o3AUHR1i3_OQ4_OVHS1g3@R4\\OWHV1f3_OR4[OXHW1d3_OS4ZOYHY1b3_OT4XOYH[1a3_OU4VOZH\\1`3_OU4TO[H_1^3_OV4RO\\H`1\\3]OZ4ROZHc1Z3ZO]4TOYHc1Y3WO`4Y1_KfNc4VOUHf1V3TOf4\\1XKcNj4VORHj1S3nNl4YOQHj1Q3lNP5ZOoGk1o2mNQ5WOQHm1m2POn4SOUHm1l2VOj4lN[Ho1i2ZOg4hN`Ho1g2^Oe4cNdHP2d2Cd4\\NiHR2`2Hc4VNmHS2\\2Mc4PNRIT2W20f4lMSIU2S23i4hMTIU2[1eMoNb2a6cMVIW2X1m0`5mLXIW2V1m0b5kLYIX2T1o0b5iLZIY2S1o0b5hL[IY2R1Q1a5fL^IZ2P1P1b5fL]I[2Q1o0b5fL]I\\2P1m0c5gL^I]2o0k0c5hL^I]2o0j0d5iL]I^2n0h0f5jL\\I`2m0ZOTOmNb6YO^Ia2k0TO^OmNZ6^O]Ib2j0mNIoNP6B]Ic2j0iNOkNm5IZId2i0gN5dNl50WIf2g0eN<]Ni58TIh2f0bNe7fNdGj2f0_Ng7gNbGl2f0\\Ni7gN^GR3i0UNj7iNZGV3k0QNk7\\2THbMm7_2RH`Mo7a2PH^MQ8c2oG\\MQ8e2nGZMS8g2lGWMV8j2jGTMX8l2gGRM[8o2dGoL^8R3bGlL_8U3`GiLb8X3]GgLd8Z3\\GcLf8]3ZGbLg8_3YG^Li8c3VG\\Lk8e3TGYLn8h3RGVLo8k3PGYLl8h3TG\\Lg8e3XG_Ld8b3[GcL`8^3`GfLZ8\\3dGiLX8X3eGnLW8R3gGUMT8l2iGZMS8g2jG_MR8k4L4L3M4M3L4L4M3LmJkH^2T7aMPI]2n6cMVI\\2g6cM\\J]1c5bNmJP1Q5PO^Kb0a4]OfK=[4BkK8V4GPL3Q4KVLOk30_LFb39gL_OY3?RMWOo2h0UMTOl2k0WMROj26XIjNR4m0g25]IkNP4l0d29_IfNQ4o0a2;WNBj1?XN]Oi1c0YNZOi1f0YNVOh1k0[NPOf1P1V60O10O10001N10001N101O8F9H1O0O2O1O0N3N1O1N2O2M2O1O2M2OnRT4"}, "label": "the head and shoulders and one leg of a goat closest to the dog"}]} {"id": 485173, "image": "COCO_train2014_000000485173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a goat which is fully white\"."}], "task": "refering", "answer_template": "The \"a goat which is fully white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 212, 213, 214, 233, 235, 236, 237, 238, 257, 258, 259, 260, 261, 262, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 325, 326, 327, 328, 329, 330, 347, 348, 349, 350, 351, 352, 369, 370, 371, 372, 373, 374, 375], "bbox": [0.050546875000000005, 0.5350730688935282, 0.462671875, 0.9844258872651357], "iscrowd": 0, "area": 25545.75565, "rle": {"size": [479, 640], "counts": "g]?1n>0O1O1O1O1O100O1O1O1O1O10O01O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1hKcNmJ^1Q5dNnJ]1o4gNoJZ1n4jNPKW1n4kNRKU1k4oNSKR1j4QOUKP1i4ROVKo0i4QOWKP1i4POVKQ1j4nNWKR1i4mNWKT1i4lNWKT1i4kNWKV1i4iNXJ_OQNh1h7hNTJCSNf1i7fNRJGTNd1j7dNPJKVNa1j7dNmINXN_1k7bNlI0ZN^1j7aNjI4[N\\1k7`NgI7^NY1k7_NeI;_NW1l7]NcI?`NU1m7\\NaIa0bNS1m7[N_Ie0cNQ1n7ZN\\Ig0gNo0m7ZN[Ih0fNP1n7YN[Ih0fNP1o7XNYIj0fNP1Q8VNXIk0fNP1R8UNWIl0eNQ1T8SNVIm0eNQ1U8RNTIn0fNR1U8RNSIm0gNR1V8QNRIn0fNS1X8oMPIP1gNR1Y8nMoHQ1fNS1[8lMnHR1fNS1\\8kMmHS1fNS1\\8kMlHT1fNS1^8iMkHT1gNT1^8hMjHU1fNU1`8fMiHU1gNV1`8eMgHW1gNV1b8cMfHW1hNW1a8cMfHV1hNY1b8aMeHV4Z7kKdHW4\\7iKcHX4\\7iKcHX4\\7iKbHY4^7gKaHZ4_7fK`H[4`7eK_H\\4b7cK^H]4b7cK]H^4d7aK[H`4e7`KZHa4f7`KXHa4i7`KSHb4m7_KQHb4P8_KmGb4S8_KkGb4U8`KhGa4Y8_KeGb4[8Q1001O001O00001O001O00001O00001O001O0O101O001O00001O4L3MO1O1O1N2O1L4K5K5K5L4O10000000000000000005K5K5K6J5K1O0lN^GkKc8i3SHlKm7n3f1N1O101N1O2O0O2N101N2N2O0O2N2O2M2N3N2M3M2O2M3M3N2GmCaMU<]27O1N2O1eN_C7b in this image.", "objects": [{"patches": [210, 211, 212, 213, 214, 233, 235, 236, 237, 238, 257, 258, 259, 260, 261, 262, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 325, 326, 327, 328, 329, 330, 347, 348, 349, 350, 351, 352, 369, 370, 371, 372, 373, 374, 375], "bbox": [0.050546875000000005, 0.5350730688935282, 0.462671875, 0.9844258872651357], "iscrowd": 0, "area": 25545.75565, "rle": {"size": [479, 640], "counts": "g]?1n>0O1O1O1O1O100O1O1O1O1O10O01O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1hKcNmJ^1Q5dNnJ]1o4gNoJZ1n4jNPKW1n4kNRKU1k4oNSKR1j4QOUKP1i4ROVKo0i4QOWKP1i4POVKQ1j4nNWKR1i4mNWKT1i4lNWKT1i4kNWKV1i4iNXJ_OQNh1h7hNTJCSNf1i7fNRJGTNd1j7dNPJKVNa1j7dNmINXN_1k7bNlI0ZN^1j7aNjI4[N\\1k7`NgI7^NY1k7_NeI;_NW1l7]NcI?`NU1m7\\NaIa0bNS1m7[N_Ie0cNQ1n7ZN\\Ig0gNo0m7ZN[Ih0fNP1n7YN[Ih0fNP1o7XNYIj0fNP1Q8VNXIk0fNP1R8UNWIl0eNQ1T8SNVIm0eNQ1U8RNTIn0fNR1U8RNSIm0gNR1V8QNRIn0fNS1X8oMPIP1gNR1Y8nMoHQ1fNS1[8lMnHR1fNS1\\8kMmHS1fNS1\\8kMlHT1fNS1^8iMkHT1gNT1^8hMjHU1fNU1`8fMiHU1gNV1`8eMgHW1gNV1b8cMfHW1hNW1a8cMfHV1hNY1b8aMeHV4Z7kKdHW4\\7iKcHX4\\7iKcHX4\\7iKbHY4^7gKaHZ4_7fK`H[4`7eK_H\\4b7cK^H]4b7cK]H^4d7aK[H`4e7`KZHa4f7`KXHa4i7`KSHb4m7_KQHb4P8_KmGb4S8_KkGb4U8`KhGa4Y8_KeGb4[8Q1001O001O00001O001O00001O00001O001O0O101O001O00001O4L3MO1O1O1N2O1L4K5K5K5L4O10000000000000000005K5K5K6J5K1O0lN^GkKc8i3SHlKm7n3f1N1O101N1O2O0O2N101N2N2O0O2N2O2M2N3N2M3M2O2M3M3N2GmCaMU<]27O1N2O1eN_C7b in this image.", "objects": [{"patches": [162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 299, 300, 301, 322, 323, 324, 325, 345, 346, 347, 348, 368, 369], "bbox": [0.0019375, 0.44703549060542797, 0.190515625, 0.9821085594989561], "iscrowd": 0, "area": 19483.508499999996, "rle": {"size": [479, 640], "counts": "Qi0a0^>i1WNi1WNa0_O000mN]KjGc4U8`KhGa4W8cKeG^4Y8fKdG[4[8iKaGX4^8kK_GV4_8oK]GR4b8QL[GP4d8TLXGm3f8WLWGj3h8W1O1N2O1O1N2O1O1BmInGU6g7g0G900O100O100O100O01O000010O01O001O010O00001O010O1O1O1O010O1O1O1O010O1O1O1O010O1O1O1O10O01O1O1O10O01O1RMQHkNQ8R1WHgNj7W1^HcNb7[1fH^N\\7_1lHZNU7d1SIUNn6i1[IoMf6o1dIiM]6T2lIdMU6Z2TJ^Mm5`2\\JXMf5e2dJRM]5l2lJmLT5i1XHgNW;W1nDcNT;\\1QE^NQ;a1X1M2O1O2N1N2O11001N10001O0O1O2N1N2O2N1O1N3N1O1O2M2O2N8H7HeQb7"}, "label": "all brown goat next to the white goats"}]} {"id": 485173, "image": "COCO_train2014_000000485173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown goat\"."}], "task": "refering", "answer_template": "The \"a brown goat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 299, 300, 301, 322, 323, 324, 325, 345, 346, 347, 348, 368, 369], "bbox": [0.0019375, 0.44703549060542797, 0.190515625, 0.9821085594989561], "iscrowd": 0, "area": 19483.508499999996, "rle": {"size": [479, 640], "counts": "Qi0a0^>i1WNi1WNa0_O000mN]KjGc4U8`KhGa4W8cKeG^4Y8fKdG[4[8iKaGX4^8kK_GV4_8oK]GR4b8QL[GP4d8TLXGm3f8WLWGj3h8W1O1N2O1O1N2O1O1BmInGU6g7g0G900O100O100O100O01O000010O01O001O010O00001O010O1O1O1O010O1O1O1O010O1O1O1O010O1O1O1O10O01O1O1O10O01O1RMQHkNQ8R1WHgNj7W1^HcNb7[1fH^N\\7_1lHZNU7d1SIUNn6i1[IoMf6o1dIiM]6T2lIdMU6Z2TJ^Mm5`2\\JXMf5e2dJRM]5l2lJmLT5i1XHgNW;W1nDcNT;\\1QE^NQ;a1X1M2O1O2N1N2O11001N10001O0O1O2N1N2O2N1O1N3N1O1O2M2O2N8H7HeQb7"}, "label": "a brown goat"}]} {"id": 26421, "image": "COCO_train2014_000000026421.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"middle zebra\"."}], "task": "refering", "answer_template": "The \"middle zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 218], "bbox": [0.28928125, 0.4158313817330211, 0.5768125, 0.6363934426229508], "iscrowd": 0, "area": 9021.357399999999, "rle": {"size": [427, 640], "counts": "bh]21Y=1O001O1N101O1O001O1O001O1N101O1O001O1O001O1N101O1O00100O01000O010O10O0101O1O1O1O001O1O1O100O000000O100O10001N1000WDVOQ;i0nD\\Oo:e0mDAP;a0kDDS;>hDGV; in this image.", "objects": [{"patches": [145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 218], "bbox": [0.28928125, 0.4158313817330211, 0.5768125, 0.6363934426229508], "iscrowd": 0, "area": 9021.357399999999, "rle": {"size": [427, 640], "counts": "bh]21Y=1O001O1N101O1O001O1O001O1N101O1O001O1O001O1N101O1O00100O01000O010O10O0101O1O1O1O001O1O1O100O000000O100O10001N1000WDVOQ;i0nD\\Oo:e0mDAP;a0kDDS;>hDGV; in this image.", "objects": [{"patches": [173, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 224, 225, 226, 227, 240, 241, 242, 243, 244, 245, 247, 248, 249, 264, 265, 266, 267, 268, 270, 271, 272, 287, 288, 289, 290, 294, 295, 296, 311, 312, 313, 317, 318, 319, 334, 335, 336, 341, 342], "bbox": [0.46178125000000003, 0.45775175644028104, 0.886078125, 0.9790866510538643], "iscrowd": 0, "area": 31489.58155, "rle": {"size": [427, 640], "counts": "Tgk31Z=1O2N1O1O2N1O1O2M2O1O2N1O1O2N1O1O4L5J5L4K6J5L2M2O2M2O2M2O1M4K4L4L5K4L4K6K4L6J6J6J6I7J7I6J6J6K6K4L4L4K5L1O2N1N2O1O1O2M2O1O1O1N3N1O1O101O0O1000001NYNTIQMl6R3VIhLk6\\3WIZLl6j3WIlKm6W4UI_KP7e4RIQKR7R5a06K8H9F9H4L000O10000O10000O1000000O10000O1000000O10000O0100000OmLeIXO[6h0WJfNi5[1^K]Ma4d2RLiLn3X3VLcLj3]3Q32O001O0O2O001N101O001N101O1O001O001O001ON2M3M3N2M3M3N2M3L4@`0A?@`0A?@`0@]c4KW]K7I7J6J5J7J5KM2N3N2M2L5HdF\\Oa6;eIH[60jI2e93L3N2N3M2N3L3N3M2N3M2M4M4L4L5J6K4L5K4L5J6H7I8I6I8VOi001N4L5K4M4K5K4L5K5L3L5K00010O00TN[HRNe7o1_HlMa7U2bHhM^7X2eHdM[7]2hH_MX7a2lH\\MS7e2PIWMQ7i2SIRMm6o2VImLj6T3YIiLf6W3^IeLb6]3`I_La6b3aI[L^6g3eITL[6n3gInKY6T4_10O2N2N200O1O1O10O01O1O101N2N2UH\\Kk6e4jHdKV7]4`HkKa7P50O2O0O100lL]H>e7\\OeH?[7\\OoH?Q7\\OYI>i6\\OaI?`6[OjI?X6[OSJ?n5\\O\\J=f5]OfJ<[5^OQK:Q5@ZK:h4_OdK8_4CkK0]4JkZn0"}, "label": "zebra behind a tree with its mouth to the ground"}]} {"id": 296760, "image": "COCO_train2014_000000296760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black and white dog with pointy ears\"."}], "task": "refering", "answer_template": "The \"black and white dog with pointy ears\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 44, 60, 61, 62, 63, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100], "bbox": [0.3668, 0.17085, 0.6204200000000001, 0.417875], "iscrowd": 0, "area": 6075.106399999999, "rle": {"size": [400, 500], "counts": "jaW21[<4O2O0100OlCOj;1RD5k;:O2Nj0VO4K6K2N0O1000001N1N2O1O1O1O1O1O0O2O000O010O10O10O010O10O010O01N2OVNUFP1m9lNWFS1Q:bNRF]1b:O02O2M2OO001O001O0010O000001O000010O0]EaNU:_1hEdNX:\\1eEgN\\:Y1cEgN]:Y1cEgN]:Y1bEhN^:X1bEhN_:W1aEiN_:X1`EhN`:X1`EhNa:W1^EjNb:V1^EiNc:X1\\EhNd:X1\\EhNe:W1[EiNe:`1101O000O101O000O101O0O1N3I6O10010O01O01N1000001N10001O0O2O00001N101N1O101N1O2O2M2N3M2N3L3N3M2N3MjRZ2"}, "label": "black and white dog with pointy ears"}]} {"id": 296760, "image": "COCO_train2014_000000296760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black and white dog nearest the wall\"."}], "task": "refering", "answer_template": "The \"the black and white dog nearest the wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 44, 60, 61, 62, 63, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100], "bbox": [0.3668, 0.17085, 0.6204200000000001, 0.417875], "iscrowd": 0, "area": 6075.106399999999, "rle": {"size": [400, 500], "counts": "jaW21[<4O2O0100OlCOj;1RD5k;:O2Nj0VO4K6K2N0O1000001N1N2O1O1O1O1O1O0O2O000O010O10O10O010O10O010O01N2OVNUFP1m9lNWFS1Q:bNRF]1b:O02O2M2OO001O001O0010O000001O000010O0]EaNU:_1hEdNX:\\1eEgN\\:Y1cEgN]:Y1cEgN]:Y1bEhN^:X1bEhN_:W1aEiN_:X1`EhN`:X1`EhNa:W1^EjNb:V1^EiNc:X1\\EhNd:X1\\EhNe:W1[EiNe:`1101O000O101O000O101O0O1N3I6O10010O01O01N1000001N10001O0O2O00001N101N1O101N1O2O2M2N3M2N3L3N3M2N3MjRZ2"}, "label": "the black and white dog nearest the wall"}]} {"id": 296760, "image": "COCO_train2014_000000296760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black dog\"."}], "task": "refering", "answer_template": "The \"black dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 115, 116, 117, 118, 119, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 205, 206, 207, 208, 209, 210, 211, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246], "bbox": [0.3803, 0.4152, 0.77636, 1.0], "iscrowd": 0, "area": 31248.399199999993, "rle": {"size": [400, 500], "counts": "c]Z2`0l;6N1N3M2O1N3M2O1N2O1N2N2O1N2gFoNP7S1eH^OT7c0`HOX73_H in this image.", "objects": [{"patches": [99, 100, 115, 116, 117, 118, 119, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 205, 206, 207, 208, 209, 210, 211, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246], "bbox": [0.3803, 0.4152, 0.77636, 1.0], "iscrowd": 0, "area": 31248.399199999993, "rle": {"size": [400, 500], "counts": "c]Z2`0l;6N1N3M2O1N3M2O1N2O1N2N2O1N2gFoNP7S1eH^OT7c0`HOX73_H in this image.", "objects": [{"patches": [119, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 211, 212, 213, 214, 229, 230, 231, 232, 247, 248, 249, 250], "bbox": [0.61764, 0.4215, 1.0, 0.9914000000000001], "iscrowd": 0, "area": 27313.556649999995, "rle": {"size": [400, 500], "counts": "_mh3=X in this image.", "objects": [{"patches": [119, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 211, 212, 213, 214, 229, 230, 231, 232, 247, 248, 249, 250], "bbox": [0.61764, 0.4215, 1.0, 0.9914000000000001], "iscrowd": 0, "area": 27313.556649999995, "rle": {"size": [400, 500], "counts": "_mh3=X in this image.", "objects": [{"patches": [72, 73, 74, 75, 76, 77, 90, 91, 92, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 115, 116, 126, 127, 128, 129, 130, 131, 132, 145, 146, 147, 148, 163, 164], "bbox": [0.0032400000000000003, 0.301, 0.4509, 0.70505], "iscrowd": 0, "area": 19267.258000000005, "rle": {"size": [400, 500], "counts": "cm07U<6M2N2O0O2N2O0O2N101N1O2N2O0O2N101iDEV: in this image.", "objects": [{"patches": [72, 73, 74, 75, 76, 77, 90, 91, 92, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 115, 116, 126, 127, 128, 129, 130, 131, 132, 145, 146, 147, 148, 163, 164], "bbox": [0.0032400000000000003, 0.301, 0.4509, 0.70505], "iscrowd": 0, "area": 19267.258000000005, "rle": {"size": [400, 500], "counts": "cm07U<6M2N2O0O2N2O0O2N101N1O2N2O0O2N101iDEV: in this image.", "objects": [{"patches": [35, 36, 58, 59, 60, 81, 82, 83, 103, 104, 105, 106, 107, 123, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 170, 171, 173, 174, 175, 176, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 264, 265, 266, 267, 268, 287, 288, 290, 291, 310, 311, 312, 313, 314, 333, 334, 335, 336, 337, 356, 357, 358, 359, 378, 379, 381, 382], "bbox": [0.372609375, 0.09597916666666667, 0.681640625, 1.0], "iscrowd": 0, "area": 34793.62015, "rle": {"size": [480, 640], "counts": "Wg_31n>a0_O3L4M2N3M110O010000O10O02O0O010O0000010O000010O000001O01O000001O000000000000000000^JYOkLg0P3_OoLa0Q3_OoLa0P3@PM`0o2AQM?n2BRM>m2CSM=l2DTMe2LWOMkLZ1n3jNTO^3k0cLSO_3k0bLoNd3Q1\\LiNk3U1VLhNn3V1SLhNP4V1QLgNR4Y1nKeNT4Z1mKcNV4\\1jKbNY4]1hKaN[4]1fK`N]4a1bK]N`4c1`KXNe4h1[KSNj4m1UKQNn4P2QKoMQ5P2oJPNQ5P2oJoMR5R2nJlMR5U2PKhMQ5X2PKfMP5[2RKbMf4h2ZKVMb4o2_KoL]4V3eKhLU4^3lK`Ln3g3SLWLh3o3ZLnKb2X5^MdJ]2d5dMYJ\\2i5eMSJ[2P6fMmIU2Z6lMbIV2_6lM_IQ2P5ZLUKf1Hm1U5bLPKb1Jh1P5[LRJ?S1_1Bi1^5RMnJW1\\Oj1j5YM`JY5b5iJ\\JU5g5mJWJP5k5RKTJl4n5VKoIj4Q6XKnIg4R6\\KkIe4U6\\KjIc4V6`KgI`4Z6aKdI_4\\6dK\\Ia4e6l1001O1N2O1O1N101N2O1O1N2O1N2O2M200O2N10001S2H8HZ_o2"}, "label": "a boy wearing a grey shirt"}]} {"id": 436025, "image": "COCO_train2014_000000436025.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man playing the wii with a woman in the living room\"."}], "task": "refering", "answer_template": "The \"a man playing the wii with a woman in the living room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 58, 59, 60, 81, 82, 83, 103, 104, 105, 106, 107, 123, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 170, 171, 173, 174, 175, 176, 196, 197, 198, 199, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245, 264, 265, 266, 267, 268, 287, 288, 290, 291, 310, 311, 312, 313, 314, 333, 334, 335, 336, 337, 356, 357, 358, 359, 378, 379, 381, 382], "bbox": [0.372609375, 0.09597916666666667, 0.681640625, 1.0], "iscrowd": 0, "area": 34793.62015, "rle": {"size": [480, 640], "counts": "Wg_31n>a0_O3L4M2N3M110O010000O10O02O0O010O0000010O000010O000001O01O000001O000000000000000000^JYOkLg0P3_OoLa0Q3_OoLa0P3@PM`0o2AQM?n2BRM>m2CSM=l2DTMe2LWOMkLZ1n3jNTO^3k0cLSO_3k0bLoNd3Q1\\LiNk3U1VLhNn3V1SLhNP4V1QLgNR4Y1nKeNT4Z1mKcNV4\\1jKbNY4]1hKaN[4]1fK`N]4a1bK]N`4c1`KXNe4h1[KSNj4m1UKQNn4P2QKoMQ5P2oJPNQ5P2oJoMR5R2nJlMR5U2PKhMQ5X2PKfMP5[2RKbMf4h2ZKVMb4o2_KoL]4V3eKhLU4^3lK`Ln3g3SLWLh3o3ZLnKb2X5^MdJ]2d5dMYJ\\2i5eMSJ[2P6fMmIU2Z6lMbIV2_6lM_IQ2P5ZLUKf1Hm1U5bLPKb1Jh1P5[LRJ?S1_1Bi1^5RMnJW1\\Oj1j5YM`JY5b5iJ\\JU5g5mJWJP5k5RKTJl4n5VKoIj4Q6XKnIg4R6\\KkIe4U6\\KjIc4V6`KgI`4Z6aKdI_4\\6dK\\Ia4e6l1001O1N2O1O1N101N2O1O1N2O1N2O2M200O2N10001S2H8HZ_o2"}, "label": "a man playing the wii with a woman in the living room"}]} {"id": 436025, "image": "COCO_train2014_000000436025.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair sitting beside the lamp\"."}], "task": "refering", "answer_template": "The \"the chair sitting beside the lamp\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 233, 234, 235, 236, 237, 256, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 352, 353, 354], "bbox": [0.153984375, 0.5416666666666666, 0.47651562499999994, 0.9279166666666666], "iscrowd": 0, "area": 25824.879, "rle": {"size": [480, 640], "counts": "[f^13m>4L4L4K5L3M4L4L4L3M4VOj0K5L4K4M4L4K5L3L5L4L4K4M4K5oE\\L^8h3gFnLZ9\\4O10O010O10O010O10O10O010O10O10O010O01000O100O100010O00001O00010O00001O000010O01O00001O01O01O00001O00010O001O00001O01O01O00001O01O02N1O2N1O2N101N2N1O2N1O2N1O2N101N2N1O2N1O2N1O2N101N2N1O00000000O101O00000000000000001O00000O10000000001O00000000000000001N1000000000000000O1N2O1O1N2O1N101N2O1O1N2O1N3N5K4L4L4L4L4L1O1O2O0O1O1O1O1O1O2N1O1O2N6K5J6J5K6J6J6J6J5K6J6J6Kgdl4"}, "label": "the chair sitting beside the lamp"}]} {"id": 436025, "image": "COCO_train2014_000000436025.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plaided chair with a white object on the back of it\"."}], "task": "refering", "answer_template": "The \"a plaided chair with a white object on the back of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 233, 234, 235, 236, 237, 256, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 352, 353, 354], "bbox": [0.153984375, 0.5416666666666666, 0.47651562499999994, 0.9279166666666666], "iscrowd": 0, "area": 25824.879, "rle": {"size": [480, 640], "counts": "[f^13m>4L4L4K5L3M4L4L4L3M4VOj0K5L4K4M4L4K5L3L5L4L4K4M4K5oE\\L^8h3gFnLZ9\\4O10O010O10O010O10O10O010O10O10O010O01000O100O100010O00001O00010O00001O000010O01O00001O01O01O00001O00010O001O00001O01O01O00001O01O02N1O2N1O2N101N2N1O2N1O2N1O2N101N2N1O2N1O2N1O2N101N2N1O00000000O101O00000000000000001O00000O10000000001O00000000000000001N1000000000000000O1N2O1O1N2O1N101N2O1O1N2O1N3N5K4L4L4L4L4L1O1O2O0O1O1O1O1O1O2N1O1O2N6K5J6J5K6J6J6J6J5K6J6J6Kgdl4"}, "label": "a plaided chair with a white object on the back of it"}]} {"id": 436025, "image": "COCO_train2014_000000436025.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman playing wii\"."}], "task": "refering", "answer_template": "The \"a woman playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 85, 86, 87, 88, 109, 110, 111, 132, 133, 134, 135, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 248, 249, 250, 251, 271, 272, 273, 274, 294, 295, 296, 297, 317, 318, 319, 320, 340, 341, 342, 363, 364, 365, 366, 386, 387, 388, 389], "bbox": [0.69028125, 0.131375, 0.9555, 0.9955416666666667], "iscrowd": 0, "area": 36311.60824999999, "rle": {"size": [480, 640], "counts": "`]_63i>5J6L400O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O0nKVOnIj0n5@lI?Q6KiI5S64hILT6:jIES6a0kI_OR6f0lIZOP6l0nISOR6P1lIPOT6R1iIoNV6T1hIkNY6V1fIjNY6Y1eIgN[6Z1cIfN]6\\1bIdN^6^1_IcN`6_1_I`Nb6`1]IaNb6`1]IaNc6_1[IcNd6_1ZIaNg6_1XIaNh6`1oHhNQ7Y1THBl7>SHBn7>PHCQ8=nGCS8=kGDV8m5QMmKb2TNb0g5PMUL`2PNm0]5fLcLm6g2WIYMQ8\\1RHdNP8V1THkNn7n0VHROm7h0VHXOm7a0WH@k7:XHFk73YHMj7LYH6i7A]H?f7XO`Hh0e7nN`HS1d7bNbH^1d7WNaHj1c7kMcHU2T:1O001O001O001O001O1O001O1O1O2N1O1O2N1WEYMW9j2dFYM[9l2^FVMb9o2WFTMh9Q3QFQMo9T3gERMX:S3]ESMc:f3O001O001O001O0000001O0O10001O0000001O00001O1O001O1O002N2N1O2N2N1O2N2kMlJZJU5b5bKjI`4Q6hKjI\\4n5lKnIY4j5nKRJV4g5PLVJU4b5RLZJR4_5UL]JP4[5VLbJn3W5WLgJn3Q5WLmJm3k4YLSKl3e4YLYKT4W4QLgK_4f3fKXLf4X3`KfLf4Q3_KmLg4j2P4F9G:E:G8H7I7H8I8H7I7ITj<"}, "label": "a woman playing wii"}]} {"id": 436025, "image": "COCO_train2014_000000436025.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman playing wii\"."}], "task": "refering", "answer_template": "The \"a woman playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 85, 86, 87, 88, 109, 110, 111, 132, 133, 134, 135, 155, 156, 157, 158, 159, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 248, 249, 250, 251, 271, 272, 273, 274, 294, 295, 296, 297, 317, 318, 319, 320, 340, 341, 342, 363, 364, 365, 366, 386, 387, 388, 389], "bbox": [0.69028125, 0.131375, 0.9555, 0.9955416666666667], "iscrowd": 0, "area": 36311.60824999999, "rle": {"size": [480, 640], "counts": "`]_63i>5J6L400O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O0nKVOnIj0n5@lI?Q6KiI5S64hILT6:jIES6a0kI_OR6f0lIZOP6l0nISOR6P1lIPOT6R1iIoNV6T1hIkNY6V1fIjNY6Y1eIgN[6Z1cIfN]6\\1bIdN^6^1_IcN`6_1_I`Nb6`1]IaNb6`1]IaNc6_1[IcNd6_1ZIaNg6_1XIaNh6`1oHhNQ7Y1THBl7>SHBn7>PHCQ8=nGCS8=kGDV8m5QMmKb2TNb0g5PMUL`2PNm0]5fLcLm6g2WIYMQ8\\1RHdNP8V1THkNn7n0VHROm7h0VHXOm7a0WH@k7:XHFk73YHMj7LYH6i7A]H?f7XO`Hh0e7nN`HS1d7bNbH^1d7WNaHj1c7kMcHU2T:1O001O001O001O001O1O001O1O1O2N1O1O2N1WEYMW9j2dFYM[9l2^FVMb9o2WFTMh9Q3QFQMo9T3gERMX:S3]ESMc:f3O001O001O001O0000001O0O10001O0000001O00001O1O001O1O002N2N1O2N2N1O2N2kMlJZJU5b5bKjI`4Q6hKjI\\4n5lKnIY4j5nKRJV4g5PLVJU4b5RLZJR4_5UL]JP4[5VLbJn3W5WLgJn3Q5WLmJm3k4YLSKl3e4YLYKT4W4QLgK_4f3fKXLf4X3`KfLf4Q3_KmLg4j2P4F9G:E:G8H7I7H8I8H7I7ITj<"}, "label": "a woman playing wii"}]} {"id": 141121, "image": "COCO_train2014_000000141121.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"smallest boy in the image\"."}], "task": "refering", "answer_template": "The \"smallest boy in the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 101, 102, 103, 119, 120, 121, 137, 138, 139, 155, 156, 157, 173, 174, 191, 192], "bbox": [0.6234, 0.4381993569131833, 0.7464, 0.9618006430868168], "iscrowd": 0, "area": 5965.048500000001, "rle": {"size": [311, 500], "counts": "dln24c95K2N0O2O1O001N2O001O1TG@b14k4=aIBZ1:S56aIBS1a0b0ZOl2d0^KC1a1a1[NP3c0[Kc1b1lMS3o2jLTMT3n2iLTMW3m2fLUMZ3n2PLeMP4^2eKjM\\4h3000000000000001O00000000000000?A1O1O001O1O1I6E in this image.", "objects": [{"patches": [83, 84, 101, 102, 103, 119, 120, 121, 137, 138, 139, 155, 156, 157, 173, 174, 191, 192], "bbox": [0.6234, 0.4381993569131833, 0.7464, 0.9618006430868168], "iscrowd": 0, "area": 5965.048500000001, "rle": {"size": [311, 500], "counts": "dln24c95K2N0O2O1O001N2O001O1TG@b14k4=aIBZ1:S56aIBS1a0b0ZOl2d0^KC1a1a1[NP3c0[Kc1b1lMS3o2jLTMT3n2iLTMW3m2fLUMZ3n2PLeMP4^2eKjM\\4h3000000000000001O00000000000000?A1O1O001O1O1I6E in this image.", "objects": [{"patches": [2, 3, 4, 20, 21, 22, 39, 40, 41, 57, 58, 59, 75, 76, 77, 93, 94, 95, 111, 112, 113, 129, 130, 147, 148, 165, 166, 167, 183, 184, 185], "bbox": [0.15516, 0.04508038585209003, 0.32008, 0.9462057877813504], "iscrowd": 0, "area": 13393.799499999994, "rle": {"size": [311, 500], "counts": "lfg09Z96M4nFDb8j0N2iNnNgIS1W6nNhIT1V6mNiIT1W6lNiIU1U6mNjIT1S6nNlIS1S6nNmIR1S6nNlIS1S6nNmIR1R6oNnIR1BQO`5MnJR1]OZOl28]N]OZOQ1[O_Of2:cNVO\\OR1WOCc2K4^OhG]O]8:g0FWUW3"}, "label": "man in blue jacket with red gloves"}]} {"id": 141121, "image": "COCO_train2014_000000141121.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"father on ski ' s\"."}], "task": "refering", "answer_template": "The \"father on ski ' s\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 20, 21, 22, 39, 40, 41, 57, 58, 59, 75, 76, 77, 93, 94, 95, 111, 112, 113, 129, 130, 147, 148, 165, 166, 167, 183, 184, 185], "bbox": [0.15516, 0.04508038585209003, 0.32008, 0.9462057877813504], "iscrowd": 0, "area": 13393.799499999994, "rle": {"size": [311, 500], "counts": "lfg09Z96M4nFDb8j0N2iNnNgIS1W6nNhIT1V6mNiIT1W6lNiIU1U6mNjIT1S6nNlIS1S6nNmIR1S6nNlIS1S6nNmIR1R6oNnIR1BQO`5MnJR1]OZOl28]N]OZOQ1[O_Of2:cNVO\\OR1WOCc2K4^OhG]O]8:g0FWUW3"}, "label": "father on ski ' s"}]} {"id": 329543, "image": "COCO_train2014_000000329543.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old oven with the door open\"."}], "task": "refering", "answer_template": "The \"an old oven with the door open\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 292, 293, 294, 295, 296, 315, 316, 317, 318, 319, 341], "bbox": [0.695734375, 0.5838562091503268, 0.8818125, 0.9040958605664489], "iscrowd": 0, "area": 15199.993699999999, "rle": {"size": [459, 640], "counts": "[gW69R>P1POo0QOo0lDjLU9W4O00000000N2O1O1O1N2O1O1N20000000000001O00000000001O000001O0001O00000000001O00000000001O00000000001O00000000010O00000000001O00000000001O00000000001O0000000O2O1O001O1O1O1O001O1O1O1O1O001OO1O1O1H8K5O001O100O1O1O1O100O1O1O2N3O3L6K5KTiQ1"}, "label": "an old oven with the door open"}]} {"id": 329543, "image": "COCO_train2014_000000329543.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stove to the right of a woman and under an imperial sign\"."}], "task": "refering", "answer_template": "The \"a stove to the right of a woman and under an imperial sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 292, 293, 294, 295, 296, 315, 316, 317, 318, 319, 341], "bbox": [0.695734375, 0.5838562091503268, 0.8818125, 0.9040958605664489], "iscrowd": 0, "area": 15199.993699999999, "rle": {"size": [459, 640], "counts": "[gW69R>P1POo0QOo0lDjLU9W4O00000000N2O1O1O1N2O1O1N20000000000001O00000000001O000001O0001O00000000001O00000000001O00000000001O00000000010O00000000001O00000000001O00000000001O0000000O2O1O001O1O1O1O001O1O1O1O1O001OO1O1O1H8K5O001O100O1O1O1O100O1O1O2N3O3L6K5KTiQ1"}, "label": "a stove to the right of a woman and under an imperial sign"}]} {"id": 329543, "image": "COCO_train2014_000000329543.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the appliance to the left of the girl in the black\"."}], "task": "refering", "answer_template": "The \"the appliance to the left of the girl in the black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 259, 260, 280, 281, 282, 283, 303, 304, 306, 326, 327, 329], "bbox": [0.14534375, 0.5920479302832244, 0.33614062499999997, 0.9071895424836601], "iscrowd": 0, "area": 9750.472999999996, "rle": {"size": [459, 640], "counts": "dnY14W>6J000000000000000000000000000000000O100000000000000000000000000000Z3fL00001O00001NN32Nf0ZO0000000000gN]FjMc9V2mFZMS9f2Y10000O1000000000000000O1000O10000000000000000000000O10000000000000O100000O100000000000000000000O100000000000O10000000O11O7I7I8H7I7I1O0000000cWn5"}, "label": "the appliance to the left of the girl in the black"}]} {"id": 329543, "image": "COCO_train2014_000000329543.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small oven is left side of the woman\"."}], "task": "refering", "answer_template": "The \"a small oven is left side of the woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 259, 260, 280, 281, 282, 283, 303, 304, 306, 326, 327, 329], "bbox": [0.14534375, 0.5920479302832244, 0.33614062499999997, 0.9071895424836601], "iscrowd": 0, "area": 9750.472999999996, "rle": {"size": [459, 640], "counts": "dnY14W>6J000000000000000000000000000000000O100000000000000000000000000000Z3fL00001O00001NN32Nf0ZO0000000000gN]FjMc9V2mFZMS9f2Y10000O1000000000000000O1000O10000000000000000000000O10000000000000O100000O100000000000000000000O100000000000O10000000O11O7I7I8H7I7I1O0000000cWn5"}, "label": "a small oven is left side of the woman"}]} {"id": 264016, "image": "COCO_train2014_000000264016.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"3 teddy bears sitting on a bed\"."}], "task": "refering", "answer_template": "The \"3 teddy bears sitting on a bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 24, 25, 26, 27, 42, 43, 44, 45, 47, 48, 49, 60, 61, 62, 63, 64, 65, 66, 67, 68, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196], "bbox": [0.22582, 0.06756024096385542, 0.92422, 0.9009036144578314], "iscrowd": 0, "area": 64856.2279, "rle": {"size": [332, 500], "counts": "PkT12W:6J6J5K6J3M3M3M3N2M3jGgNV7\\1iHiNP7Z1oHlNj6V1UIPOc6S1\\IRO]6R1aIQOZ6R1eIQOW6Q1hIQOU6Q1iIROU6o0jITOS6m0lIUOR6l0mIVOQ6k0nIXOo5i0PJYOP5cN`KU2_OZOm4eNbKR2@\\Oj4dNeKQ2@]Oh4dNgKP2@]Og4eNhKo1@]Of4fNiKn1@]Of4fNiKn1@]Og4fNgKn1A]Oh4eNfKo1A]Oi4dNeKP2A^Oh4cNfKP2A^Oi4bNeKR2A\\Oj4cNbKR4^4nKaKS4_4mK`KT4_4mK`KT4`4lK_KU4a4lK\\KW4c4iK\\KX4c4iK\\KX4d4hK[KY4e4gKYK[4g4e000O101O00000O1000000\\OdJXL\\5e3g0L4cN`IkNc6T1`IgNc6X1`IbNd6\\1`I_Nb6a1aIZNb6e1aIVNb6i1aIQNc6m1aInMb6Q2m0O1O0O2O11O0000001O0O10001O000000001N1000001O002N4L3L2O0000000000000O100000000000000O1000O100O10N12N2O1N2N2O2M2O1N001000000001O00001O000O2O00003M4L5K4L4L3M3eJoLU3S3hLnLW3U3eLmL[3V3aLlL^3V3^LlLb3W3YLkLg3W3TLmLk3U3QLmLo3V3lKmLS4U3hKnLX4\\3\\KfLd4b4O0O2O1O1O1O1O001O1O1O2N1N3N1O2N1O2NN2M3M4XOcK\\K`4b4f0O1O1O1O1N2O1O1O1O1N2O1O1O1O2QOSJ[Mo5d2eJhL\\5X3l000O100O10000O100O100O10001N100O100001O001O1O001O001O1O0010O01O1O001O001O1O001O001O1O001O10O00000000000000000000000000000O2O000O10000O10000O10000002N1O2dK^Lj1d3RN_Ln1a3oMaLQ2a3lMaLS2a3iMbLV2_3gMdLX2^3dMeL[2]3aMeL_2]3]MfLb2`3WMbLh2d3PM_Lo2g3lLXLT3k3kLTLT3o3kLPLT3R4lLmKS3V4iLlKV3W4cLnK\\3U4]LPLb3R4XLTLf3Z5O100O1O1O1O1O1O2N100O1O1O1O1O2O1N3M2N2N3M2N2O2M2N2N3M2N2UNRHU1P8dNUH[1m7^NXHb1Z8O1O1O0O2O1N101N2N101N2O0O2O1N101N2O1N3N2M3N2M3N2M3M5LdQ<"}, "label": "3 teddy bears sitting on a bed"}]} {"id": 264016, "image": "COCO_train2014_000000264016.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"group of three teddy bears sitting\"."}], "task": "refering", "answer_template": "The \"group of three teddy bears sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 24, 25, 26, 27, 42, 43, 44, 45, 47, 48, 49, 60, 61, 62, 63, 64, 65, 66, 67, 68, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196], "bbox": [0.22582, 0.06756024096385542, 0.92422, 0.9009036144578314], "iscrowd": 0, "area": 64856.2279, "rle": {"size": [332, 500], "counts": "PkT12W:6J6J5K6J3M3M3M3N2M3jGgNV7\\1iHiNP7Z1oHlNj6V1UIPOc6S1\\IRO]6R1aIQOZ6R1eIQOW6Q1hIQOU6Q1iIROU6o0jITOS6m0lIUOR6l0mIVOQ6k0nIXOo5i0PJYOP5cN`KU2_OZOm4eNbKR2@\\Oj4dNeKQ2@]Oh4dNgKP2@]Og4eNhKo1@]Of4fNiKn1@]Of4fNiKn1@]Og4fNgKn1A]Oh4eNfKo1A]Oi4dNeKP2A^Oh4cNfKP2A^Oi4bNeKR2A\\Oj4cNbKR4^4nKaKS4_4mK`KT4_4mK`KT4`4lK_KU4a4lK\\KW4c4iK\\KX4c4iK\\KX4d4hK[KY4e4gKYK[4g4e000O101O00000O1000000\\OdJXL\\5e3g0L4cN`IkNc6T1`IgNc6X1`IbNd6\\1`I_Nb6a1aIZNb6e1aIVNb6i1aIQNc6m1aInMb6Q2m0O1O0O2O11O0000001O0O10001O000000001N1000001O002N4L3L2O0000000000000O100000000000000O1000O100O10N12N2O1N2N2O2M2O1N001000000001O00001O000O2O00003M4L5K4L4L3M3eJoLU3S3hLnLW3U3eLmL[3V3aLlL^3V3^LlLb3W3YLkLg3W3TLmLk3U3QLmLo3V3lKmLS4U3hKnLX4\\3\\KfLd4b4O0O2O1O1O1O1O001O1O1O2N1N3N1O2N1O2NN2M3M4XOcK\\K`4b4f0O1O1O1O1N2O1O1O1O1N2O1O1O1O2QOSJ[Mo5d2eJhL\\5X3l000O100O10000O100O100O10001N100O100001O001O1O001O001O1O0010O01O1O001O001O1O001O001O1O001O10O00000000000000000000000000000O2O000O10000O10000O10000002N1O2dK^Lj1d3RN_Ln1a3oMaLQ2a3lMaLS2a3iMbLV2_3gMdLX2^3dMeL[2]3aMeL_2]3]MfLb2`3WMbLh2d3PM_Lo2g3lLXLT3k3kLTLT3o3kLPLT3R4lLmKS3V4iLlKV3W4cLnK\\3U4]LPLb3R4XLTLf3Z5O100O1O1O1O1O1O2N100O1O1O1O1O2O1N3M2N2N3M2N2O2M2N2N3M2N2UNRHU1P8dNUH[1m7^NXHb1Z8O1O1O0O2O1N101N2N101N2O0O2O1N101N2O1N3N2M3N2M3N2M3M5LdQ<"}, "label": "group of three teddy bears sitting"}]} {"id": 264016, "image": "COCO_train2014_000000264016.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small white bear with little brown bow tie\"."}], "task": "refering", "answer_template": "The \"a small white bear with little brown bow tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 118, 119, 120, 121, 136, 137, 138, 139, 154, 155, 156, 157, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 208, 209, 210, 211], "bbox": [0.55034, 0.4628313253012048, 0.81206, 0.9740963855421686], "iscrowd": 0, "area": 15781.364799999994, "rle": {"size": [332, 500], "counts": "W]i22S:8I8I6N3L3N3mFSOb8o0\\GUO`8m0^GWO^8l0_GXO]8j0aGYO[8a1K4L4L5K4L5eNXMhJm2f4dMWK`2`4iM\\KY2b4jM[KW2d4lMYKU2f4oMVKR2h4RNTKP2k4SNRKn1m4VNoJk1P5XNkJk1T5g1N2O100000000000000000000000000000000000000000000000TOoJ]LR5a3PK^LP5b3RKZLP5f3RKWLo4i3SKTLn4k3UKoKo4Q4RKjKR5V4b000O2O00000O10TJlK`5T4^JoKa5P4]JTLb5l3\\JWLd5h3ZJ\\Ld5W4O1O2N1O1O1O0001O000000000000010O01O001O01jLdJZ1\\5dNgJ\\1Y5bNiJ]1W5aNlJ_1T5_NoJ_1Q5_NRKa1n4]NTKb1l4\\NWKd1i4ZNZKd1f4ZN]Kf1c4XN_Kg1a4WNbKh1_4UNdKh1^4VNfKg1\\4VNgKg1[4WNhKg1]4SNfKk1_4nMfKQ1\\N_OU6YObKW1[N]OR8c0nG[OT8d0nGYOU8f0lGWOV8h0kGmN`8R1>00O2O0O2N101N100O2M2N3M2N3FUWn0"}, "label": "a small white bear with little brown bow tie"}]} {"id": 264016, "image": "COCO_train2014_000000264016.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smallest white teddy bear sitting in front of the other white bear\"."}], "task": "refering", "answer_template": "The \"the smallest white teddy bear sitting in front of the other white bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 118, 119, 120, 121, 136, 137, 138, 139, 154, 155, 156, 157, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 208, 209, 210, 211], "bbox": [0.55034, 0.4628313253012048, 0.81206, 0.9740963855421686], "iscrowd": 0, "area": 15781.364799999994, "rle": {"size": [332, 500], "counts": "W]i22S:8I8I6N3L3N3mFSOb8o0\\GUO`8m0^GWO^8l0_GXO]8j0aGYO[8a1K4L4L5K4L5eNXMhJm2f4dMWK`2`4iM\\KY2b4jM[KW2d4lMYKU2f4oMVKR2h4RNTKP2k4SNRKn1m4VNoJk1P5XNkJk1T5g1N2O100000000000000000000000000000000000000000000000TOoJ]LR5a3PK^LP5b3RKZLP5f3RKWLo4i3SKTLn4k3UKoKo4Q4RKjKR5V4b000O2O00000O10TJlK`5T4^JoKa5P4]JTLb5l3\\JWLd5h3ZJ\\Ld5W4O1O2N1O1O1O0001O000000000000010O01O001O01jLdJZ1\\5dNgJ\\1Y5bNiJ]1W5aNlJ_1T5_NoJ_1Q5_NRKa1n4]NTKb1l4\\NWKd1i4ZNZKd1f4ZN]Kf1c4XN_Kg1a4WNbKh1_4UNdKh1^4VNfKg1\\4VNgKg1[4WNhKg1]4SNfKk1_4nMfKQ1\\N_OU6YObKW1[N]OR8c0nG[OT8d0nGYOU8f0lGWOV8h0kGmN`8R1>00O2O0O2N101N100O2M2N3M2N3FUWn0"}, "label": "the smallest white teddy bear sitting in front of the other white bear"}]} {"id": 517985, "image": "COCO_train2014_000000517985.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the green vegetable in the left\"."}], "task": "refering", "answer_template": "The \"the green vegetable in the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279], "bbox": [0.0, 0.3077916666666667, 0.26356250000000003, 0.7464583333333333], "iscrowd": 0, "area": 21507.712550000007, "rle": {"size": [480, 640], "counts": "X8d0[>Y1gNm0SO2O0O1O1O1O100O1O2N100O1N2N2N2M3oNR1N100O1O100O1O1O1O2M2O1O1O1N2N2N2M3N2N2O1N2O1N2N2O1O2N1N2O1O1O1O1O1O1O1O1O100O100O100O1O100O100O100O100O101N100001O1O00N2M3N20000O10000000000O10cNjGPLV8o3SHiKm7U4]HcKc7\\4fH\\K\\7`4nHXKS7f4e1N3L3N2N2M4M2N2M4M2N2M3N2N2M3N2N2M3N2N2M4M3L6K4L1N2O1O2M2O1O1N2O1N2O1O1N101N2O0O2O1N3N1N3N1O2O1N1O2O0O2O0O2N2O0O2N2O1N2O2M2N2O2L3N3M3L4M4K5L3MXQm6"}, "label": "the green vegetable in the left"}]} {"id": 517985, "image": "COCO_train2014_000000517985.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green vegetable by brown food\"."}], "task": "refering", "answer_template": "The \"green vegetable by brown food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 230, 231, 232, 233, 253, 254, 255, 256, 276, 277, 278, 279], "bbox": [0.0, 0.3077916666666667, 0.26356250000000003, 0.7464583333333333], "iscrowd": 0, "area": 21507.712550000007, "rle": {"size": [480, 640], "counts": "X8d0[>Y1gNm0SO2O0O1O1O1O100O1O2N100O1N2N2N2M3oNR1N100O1O100O1O1O1O2M2O1O1O1N2N2N2M3N2N2O1N2O1N2N2O1O2N1N2O1O1O1O1O1O1O1O1O100O100O100O1O100O100O100O100O101N100001O1O00N2M3N20000O10000000000O10cNjGPLV8o3SHiKm7U4]HcKc7\\4fH\\K\\7`4nHXKS7f4e1N3L3N2N2M4M2N2M4M2N2M3N2N2M3N2N2M3N2N2M4M3L6K4L1N2O1O2M2O1O1N2O1N2O1O1N101N2O0O2O1N3N1N3N1O2O1N1O2O0O2O0O2N2O0O2N2O1N2O2M2N2O2L3N3M3L4M4K5L3MXQm6"}, "label": "green vegetable by brown food"}]} {"id": 517985, "image": "COCO_train2014_000000517985.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chinese food rests in a blue dish with broccoli and tiny corn\"."}], "task": "refering", "answer_template": "The \"chinese food rests in a blue dish with broccoli and tiny corn\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 35, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 353, 354, 355, 356, 357], "bbox": [0.0, 0.07416666666666667, 1.0, 0.8921250000000001], "iscrowd": 0, "area": 189058.49285, "rle": {"size": [480, 640], "counts": "Y6k2U<;D=D in this image.", "objects": [{"patches": [32, 33, 34, 35, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 353, 354, 355, 356, 357], "bbox": [0.0, 0.07416666666666667, 1.0, 0.8921250000000001], "iscrowd": 0, "area": 189058.49285, "rle": {"size": [480, 640], "counts": "Y6k2U<;D=D in this image.", "objects": [{"patches": [61, 62, 63, 64, 65, 83, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203], "bbox": [0.54353125, 0.12235416666666667, 0.9688281249999999, 0.494125], "iscrowd": 0, "area": 30132.639149999995, "rle": {"size": [480, 640], "counts": "TZS54e>8J5N2N2M3N2M3M4L3M3L5K4L5M2N2M5L5K4M3L4L5L3L4M1N2N2O1O0O2O1O1O1O1O001O100O10O0100O100O10O0100O100O010O100O100O001N2O1O1O001O1O100O100O2N100O2O0O2O0O2N100O2O0O1O100N3N1N2O1O1O2N1O1O100O2N2N101N1O2O1N101N101N101N10000O101N10000O1O2N1O100O101O000000001N10000000001O0000000000001O000001O1O010O001O00001O000O2O0000000O2O0000000O101O00000O101O00000O101O000000001O0O10001O1O001N2O001O001O1N101O0jMiECX:8mEGW:2mEMX:LkE2Z:HiE7[:ChE<^:\\OeEc0b:TO`El0f:lN]ES1i:eNYEZ1n:^NUEa1U<01O001O001O001O001O0000001O00000O10001O00000000001N10000000001O000O100O101O0O2O001N100O2O001N101N10001N101N101O0O2O0O101O0O2O1N3I6GVV9"}, "label": "broccoli above the piece of baby corn"}]} {"id": 517985, "image": "COCO_train2014_000000517985.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"broccoli on top of the corn\"."}], "task": "refering", "answer_template": "The \"broccoli on top of the corn\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 64, 65, 83, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203], "bbox": [0.54353125, 0.12235416666666667, 0.9688281249999999, 0.494125], "iscrowd": 0, "area": 30132.639149999995, "rle": {"size": [480, 640], "counts": "TZS54e>8J5N2N2M3N2M3M4L3M3L5K4L5M2N2M5L5K4M3L4L5L3L4M1N2N2O1O0O2O1O1O1O1O001O100O10O0100O100O10O0100O100O010O100O100O001N2O1O1O001O1O100O100O2N100O2O0O2O0O2N100O2O0O1O100N3N1N2O1O1O2N1O1O100O2N2N101N1O2O1N101N101N101N10000O101N10000O1O2N1O100O101O000000001N10000000001O0000000000001O000001O1O010O001O00001O000O2O0000000O2O0000000O101O00000O101O00000O101O000000001O0O10001O1O001N2O001O001O1N101O0jMiECX:8mEGW:2mEMX:LkE2Z:HiE7[:ChE<^:\\OeEc0b:TO`El0f:lN]ES1i:eNYEZ1n:^NUEa1U<01O001O001O001O001O0000001O00000O10001O00000000001N10000000001O000O100O101O0O2O001N100O2O001N101N10001N101N101O0O2O0O101O0O2O1N3I6GVV9"}, "label": "broccoli on top of the corn"}]} {"id": 313186, "image": "COCO_train2014_000000313186.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young blond haired boy wearing a hoodie and holding a blue kite with a whale on it\"."}], "task": "refering", "answer_template": "The \"a young blond haired boy wearing a hoodie and holding a blue kite with a whale on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 43, 44, 57, 58, 59, 60, 61, 74, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 113, 125, 126, 127, 128, 129, 130, 142, 143, 144, 145, 146, 147, 158, 159, 160, 161, 162, 163, 164, 174, 175, 176, 177, 178, 179, 180, 181, 182, 191, 192, 193, 194, 195, 196, 197, 198, 199, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 306, 307, 308, 309, 312, 313, 314, 315, 316, 317, 318, 319, 320, 329, 330, 331, 332, 333, 334, 335, 336, 346, 347, 348, 349, 350, 351, 352, 353, 363, 364, 365, 366, 367, 368, 369, 370, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.003, 0.09176562499999999, 0.904875, 0.9749062500000001], "iscrowd": 0, "area": 120249.3314, "rle": {"size": [640, 480], "counts": "[g13lc02M3N2M3N2M3N2M3N3L3M3N2M3N2M3N2M3N200O1O1O1O1O1O1O010O000000000001O000000000000010O0000000000001OO1O1O100O1O1O2N1O100O1O1O1O100O1O1O00010O00001O010O00001O01O01O001O00010O00001O010O00001O01O01O001O01O01O000010O01O00001M2H9F9H8H9F9H8H9G8G:G8H8G:J500010O01O00010O000010O01O00010O00010O001O01O01O00010O0010O0001O01O01O3M3N3L3M4M2M4L3ZKaIfKc6U4`JjJb5R5aKlIc4P6^LoHe3m6^MQHf2j7k4L5kKkC7Y<^OPDa0T in this image.", "objects": [{"patches": [40, 41, 42, 43, 44, 57, 58, 59, 60, 61, 74, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 113, 125, 126, 127, 128, 129, 130, 142, 143, 144, 145, 146, 147, 158, 159, 160, 161, 162, 163, 164, 174, 175, 176, 177, 178, 179, 180, 181, 182, 191, 192, 193, 194, 195, 196, 197, 198, 199, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 306, 307, 308, 309, 312, 313, 314, 315, 316, 317, 318, 319, 320, 329, 330, 331, 332, 333, 334, 335, 336, 346, 347, 348, 349, 350, 351, 352, 353, 363, 364, 365, 366, 367, 368, 369, 370, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.003, 0.09176562499999999, 0.904875, 0.9749062500000001], "iscrowd": 0, "area": 120249.3314, "rle": {"size": [640, 480], "counts": "[g13lc02M3N2M3N2M3N2M3N3L3M3N2M3N2M3N2M3N200O1O1O1O1O1O1O010O000000000001O000000000000010O0000000000001OO1O1O100O1O1O2N1O100O1O1O1O100O1O1O00010O00001O010O00001O01O01O001O00010O00001O010O00001O01O01O001O01O01O000010O01O00001M2H9F9H8H9F9H8H9G8G:G8H8G:J500010O01O00010O000010O01O00010O00010O001O01O01O00010O0010O0001O01O01O3M3N3L3M4M2M4L3ZKaIfKc6U4`JjJb5R5aKlIc4P6^LoHe3m6^MQHf2j7k4L5kKkC7Y<^OPDa0T in this image.", "objects": [{"patches": [124, 126, 127, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 286, 287, 288, 289, 290, 309], "bbox": [0.364859375, 0.3277083333333334, 0.65540625, 0.7736458333333334], "iscrowd": 0, "area": 24542.199399999998, "rle": {"size": [480, 640], "counts": "Xk]3e0[>f0ZO5K4L3M3M4L3M3M4L3M3M4L3M3M4L3M3M3M3M3M3M3M4L3M3M3M3M3M3M4L3M3M3M3M3M3M3WGQK[7R5aHoJ_7S5]HoJc7R5[HoJe7R5YHoJg7S5VHnJj7S5THnJl7S5QHoJo7R5oGoJQ8S5lGnJT8S5jGmJW8T5gGmJY8T5eGmJ[8U5aGmJ`8S5^GnJb8S5\\GnJd8g500000000000O100000000000O10000O1000O010000O1000000O10000O10O10O1000000O10000O10000O01000O100000QNPY4gKXmK in this image.", "objects": [{"patches": [124, 126, 127, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 286, 287, 288, 289, 290, 309], "bbox": [0.364859375, 0.3277083333333334, 0.65540625, 0.7736458333333334], "iscrowd": 0, "area": 24542.199399999998, "rle": {"size": [480, 640], "counts": "Xk]3e0[>f0ZO5K4L3M3M4L3M3M4L3M3M4L3M3M4L3M3M3M3M3M3M3M4L3M3M3M3M3M3M4L3M3M3M3M3M3M3WGQK[7R5aHoJ_7S5]HoJc7R5[HoJe7R5YHoJg7S5VHnJj7S5THnJl7S5QHoJo7R5oGoJQ8S5lGnJT8S5jGmJW8T5gGmJY8T5eGmJ[8U5aGmJ`8S5^GnJb8S5\\GnJd8g500000000000O100000000000O10000O1000O010000O1000000O10000O10O10O1000000O10000O10000O01000O100000QNPY4gKXmK in this image.", "objects": [{"patches": [89, 90, 112, 113, 134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 250, 251, 252, 274, 275, 297, 298, 319, 320, 321, 343, 344], "bbox": [0.8268281249999999, 0.18589583333333334, 1.0, 0.8813125], "iscrowd": 0, "area": 23229.450050000003, "rle": {"size": [480, 640], "counts": "iTh75i>:F3L5L6I7J9YCjNS;]1gDhNR;^1iDgNR;^1iDgNS;\\1jDgNS;[1jDiNS;Z1iDiNV;X1cDPO[;\\2N1O2N2N2O001O1N2O010O1O0010OnNmLVFU3d9RMYFo2d9X1M4M2N3N1O1O1O2O0O1O1O101O0O100O2O000kIWJR3k5kL[JQ3f5kL_JS3b5iLdJT3]5hLhJV3Y5eLlJZ3T5cLPK[3R5`LRK`3n4ZLYKd3h4TLaKl3^4iKmKW4S4eKRL[4m3bKWL^4h3_K[Lb4d3\\K_Ld4`3YKdLg4[3VKjLh4V3VKmLj4S3RKRMm4m2PKXMo4h2lJ]MT5Y6O10bGiJP7X5mHiJU7W5iHiJX7W5fHjJZ7Y5bHhJ_7Y5^HhJb7[5ZHfJf7^5UHcJl7`5oGaJQ8d5iG]JX8U60O101O0O101N101O0O100001N1lNgIaIZ6^6gIaIZ6^6gIaIY6_6hI`IX6`6hI`IX6`6iI_IW6a6iI^IX6b6hI^IX6b6iI\\IW6e6iI[IY6b6hI^IY6a6gI^I[6`6gI_IZ6_6gI`I[6^6fIbI[6[6hI`Id1"}, "label": "a woman sitting with arms folded"}]} {"id": 550760, "image": "COCO_train2014_000000550760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women\"."}], "task": "refering", "answer_template": "The \"a women\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 112, 113, 134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 250, 251, 252, 274, 275, 297, 298, 319, 320, 321, 343, 344], "bbox": [0.8268281249999999, 0.18589583333333334, 1.0, 0.8813125], "iscrowd": 0, "area": 23229.450050000003, "rle": {"size": [480, 640], "counts": "iTh75i>:F3L5L6I7J9YCjNS;]1gDhNR;^1iDgNR;^1iDgNS;\\1jDgNS;[1jDiNS;Z1iDiNV;X1cDPO[;\\2N1O2N2N2O001O1N2O010O1O0010OnNmLVFU3d9RMYFo2d9X1M4M2N3N1O1O1O2O0O1O1O101O0O100O2O000kIWJR3k5kL[JQ3f5kL_JS3b5iLdJT3]5hLhJV3Y5eLlJZ3T5cLPK[3R5`LRK`3n4ZLYKd3h4TLaKl3^4iKmKW4S4eKRL[4m3bKWL^4h3_K[Lb4d3\\K_Ld4`3YKdLg4[3VKjLh4V3VKmLj4S3RKRMm4m2PKXMo4h2lJ]MT5Y6O10bGiJP7X5mHiJU7W5iHiJX7W5fHjJZ7Y5bHhJ_7Y5^HhJb7[5ZHfJf7^5UHcJl7`5oGaJQ8d5iG]JX8U60O101O0O101N101O0O100001N1lNgIaIZ6^6gIaIZ6^6gIaIY6_6hI`IX6`6hI`IX6`6iI_IW6a6iI^IX6b6hI^IX6b6iI\\IW6e6iI[IY6b6hI^IY6a6gI^I[6`6gI_IZ6_6gI`I[6^6fIbI[6[6hI`Id1"}, "label": "a women"}]} {"id": 550760, "image": "COCO_train2014_000000550760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man seated on far end of couch , to left of man in plaid\"."}], "task": "refering", "answer_template": "The \"man seated on far end of couch , to left of man in plaid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 142, 143, 144, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 213, 214, 215, 237], "bbox": [0.1938125, 0.303375, 0.40449999999999997, 0.6696666666666666], "iscrowd": 0, "area": 9638.038050000001, "rle": {"size": [480, 640], "counts": "lXj15j>9H8G8I7H7I6K6I6K6I7J5K6I4M00001N10001N101O1O0O2O1O1O001O1O001O1O1O1O1O1O00100O1O1O1000O1000O1000O10000O2O3M3L4M2N1N2O1O101N1000000aMXDc1b<2N2N1O2N2N2N1O2M3N2N1O2N2N2N2N2N1O2N2N2NGVD^Mh;\\2a0J6M3N2N2N101O1N2O1O1O1O1N2O1L4L4L4N2M3N2N2N2N2N2N2N2N2N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1NYkb5"}, "label": "man seated on far end of couch , to left of man in plaid"}]} {"id": 550760, "image": "COCO_train2014_000000550760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in middle with solid grey shirt\"."}], "task": "refering", "answer_template": "The \"man in middle with solid grey shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 142, 143, 144, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 213, 214, 215, 237], "bbox": [0.1938125, 0.303375, 0.40449999999999997, 0.6696666666666666], "iscrowd": 0, "area": 9638.038050000001, "rle": {"size": [480, 640], "counts": "lXj15j>9H8G8I7H7I6K6I6K6I7J5K6I4M00001N10001N101O1O0O2O1O1O001O1O001O1O1O1O1O1O00100O1O1O1000O1000O1000O10000O2O3M3L4M2N1N2O1O101N1000000aMXDc1b<2N2N1O2N2N2N1O2M3N2N1O2N2N2N2N2N1O2N2N2NGVD^Mh;\\2a0J6M3N2N2N101O1N2O1O1O1O1N2O1L4L4L4N2M3N2N2N2N2N2N2N2N2N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1NYkb5"}, "label": "man in middle with solid grey shirt"}]} {"id": 550760, "image": "COCO_train2014_000000550760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little boy getting his teeth brushed\"."}], "task": "refering", "answer_template": "The \"little boy getting his teeth brushed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 103, 104, 105, 106, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 215, 216, 219, 220, 221, 238, 242, 243, 244, 261, 264, 265, 266, 267, 268, 284, 286, 287, 288, 289, 290, 291, 292, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 334, 335, 337, 338, 339, 355, 356, 357], "bbox": [0.348875, 0.18652083333333333, 0.768546875, 0.9460625], "iscrowd": 0, "area": 38582.197349999995, "rle": {"size": [480, 640], "counts": "gjX38X>k0B=C5H8G9J6J6J4M3M3N1O2N2OoMUDn0j;SOoD5o:KZENd:3dEE[:oKEo3:PLIo37PLLn33QL1l30SL3k3LTL7j3JUL9h3GXLQNCn1?oMCQ2=nMCR2>lMCT2?jMBU2?iMBX2?fMAZ2`0dMB[2?dMA\\2a0aM@`2`0_M@a2b0]M_Ob2b0\\M_Od2c0ZM]Og2c0WM_Oh2d0UM\\Ok2f0RM[On2h0oLXOR3j0jLXOU3k0hLUOX3n0dLSO\\3P1_LSOa3o0YLTOg3n0SLVOm3m0mKVOS4m0fKXOZ4j0`KYO`4c51O010O1O10O01O1O00100O1O001O1O1O010O1O1O1O2N2N1dJkJl0W5SOjJl0W5ROlJk0U5UOmJh0U5WOlJg0V5XOlJe0U5[OlJc0V5\\OlJ`0W5_OjJ?W5@lJPK@Q5=RKCn4;UKEj48YKHg45]KJe41^K0c4KaK4a4GcK8_4BeK?\\4\\OhKc0Z4XOiKh0Y4SOkKm0V4nNmKR1U4hNPLW1_4WNdKj1j8O001O1O01O01O001O010O00001O010O001O00010N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N100O2N1O4L6K4K5K0010O01O001O001O00001O0O2O001O001N101O001O00]oT2"}, "label": "little boy getting his teeth brushed"}]} {"id": 550760, "image": "COCO_train2014_000000550760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde boy with an orange shirt\"."}], "task": "refering", "answer_template": "The \"a blonde boy with an orange shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 103, 104, 105, 106, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 215, 216, 219, 220, 221, 238, 242, 243, 244, 261, 264, 265, 266, 267, 268, 284, 286, 287, 288, 289, 290, 291, 292, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 334, 335, 337, 338, 339, 355, 356, 357], "bbox": [0.348875, 0.18652083333333333, 0.768546875, 0.9460625], "iscrowd": 0, "area": 38582.197349999995, "rle": {"size": [480, 640], "counts": "gjX38X>k0B=C5H8G9J6J6J4M3M3N1O2N2OoMUDn0j;SOoD5o:KZENd:3dEE[:oKEo3:PLIo37PLLn33QL1l30SL3k3LTL7j3JUL9h3GXLQNCn1?oMCQ2=nMCR2>lMCT2?jMBU2?iMBX2?fMAZ2`0dMB[2?dMA\\2a0aM@`2`0_M@a2b0]M_Ob2b0\\M_Od2c0ZM]Og2c0WM_Oh2d0UM\\Ok2f0RM[On2h0oLXOR3j0jLXOU3k0hLUOX3n0dLSO\\3P1_LSOa3o0YLTOg3n0SLVOm3m0mKVOS4m0fKXOZ4j0`KYO`4c51O010O1O10O01O1O00100O1O001O1O1O010O1O1O1O2N2N1dJkJl0W5SOjJl0W5ROlJk0U5UOmJh0U5WOlJg0V5XOlJe0U5[OlJc0V5\\OlJ`0W5_OjJ?W5@lJPK@Q5=RKCn4;UKEj48YKHg45]KJe41^K0c4KaK4a4GcK8_4BeK?\\4\\OhKc0Z4XOiKh0Y4SOkKm0V4nNmKR1U4hNPLW1_4WNdKj1j8O001O1O01O01O001O010O00001O010O001O00010N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N100O2N1O4L6K4K5K0010O01O001O001O00001O0O2O001O001N101O001O00]oT2"}, "label": "a blonde boy with an orange shirt"}]} {"id": 550760, "image": "COCO_train2014_000000550760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with glasses holding a boy\"."}], "task": "refering", "answer_template": "The \"a man with glasses holding a boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 69, 70, 71, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 189, 194, 195, 207, 208, 209, 210, 211, 212, 213, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 292, 293, 294, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 313, 314, 315, 316, 317, 318, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 336, 337, 338, 339, 340, 341, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 356, 357, 358, 359, 360, 361, 362, 363, 364, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.0, 0.14035416666666667, 0.896296875, 1.0], "iscrowd": 0, "area": 103341.4742, "rle": {"size": [480, 640], "counts": "W2c<[23M301N1O1O0001O0000001O000000001O000000001O0000001O000000001O000000001O000000001O00001O001O001O001O001O001O1O001O001O001O001O001O001O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O001O001O1O001O001O0000000000001O001O1O001O0mEULQ9m3lFYLo8i3mF^Ln8d3nFcLm8_3oFhLl8j4L5K6J6J5K6J5K6J2N1O001O1O001O1O1O001O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O002N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O2N2N2N2N2N2N2N2N2N2N2N2N2Na0_O7I1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1OO1O1O1O1O1O1O1M3M3L4M3M3L4MLhEfKS:]4oEcKe9f4]F[KV9P5lFPKg8Z5\\GfJb8Z5_GgJ_8Y5cGgJ\\8W5gGiJX8V5iGkJV8S5mGmJQ8S5QHmJn7R5SHoJl7o4WHQKh7n4ZHRKd7m4^HTKa7k4aHUK^7i4eHWKZ7h4hHXKW7g4jHZKU7d4nH\\KQ7c4QI]Kn6a4TI`Kk6_4WIaKh6]4[IcKd6\\4]IeKb6Z4`IfK_6X4dIhK[6W4fIjKY6U4iIkKV6U4kIkKT6T4mImKR6R4PJnKo5R4RJnKm5Q4TJPLl5o3UJQLj5n3XJRLh5l3YJULf5k3[JULd5j3^JVLb5h3_JYL`5g3aJYL^5f3dJZL\\5d3fJ\\LY5d3gJ]LY5a3iJ_LV5a3kJ_LT5`3mJaLS5]3oJcLP5]3QKcLn4\\3SKeLm4Y3UKgLj4Y3WKgLh4X3YKiLg4V3ZKjLe4U3]KkLc4S3^KnLa4R3`KnL`4Q3aKoLa4m2`KTMa4j2`KVMb4g2^KZMd4c2]K]Me4`2\\K`Me4^2[KcMg4Y2[KgMe4X2\\KhMc4X2]KiMb4W2_KiMa4V2`KjM_4V2aKkM^4U2cKkM\\4T2fKlMZ4S2fKnMY4R2hKnMW4R2iKoMV4Q2kKoMU4P2lKPNS4o1nKRNR4m1oKSNR4k1oKUNR4i1nKXNS4f1nKZNS4d1nK\\NS4b1mK_NT4^1nKbNS4\\1nKdNS4Z1mKgNT4W1mKiNT4U1lKlNU4R1lKnNU4o0mKQOT4m0lKTOU4i0mKWOT4f0nKZOR4d0oK]OR4`0PL@R4 in this image.", "objects": [{"patches": [46, 47, 69, 70, 71, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 189, 194, 195, 207, 208, 209, 210, 211, 212, 213, 216, 217, 218, 219, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 292, 293, 294, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 313, 314, 315, 316, 317, 318, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 336, 337, 338, 339, 340, 341, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 356, 357, 358, 359, 360, 361, 362, 363, 364, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.0, 0.14035416666666667, 0.896296875, 1.0], "iscrowd": 0, "area": 103341.4742, "rle": {"size": [480, 640], "counts": "W2c<[23M301N1O1O0001O0000001O000000001O000000001O0000001O000000001O000000001O000000001O00001O001O001O001O001O001O1O001O001O001O001O001O001O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O001O001O1O001O001O0000000000001O001O1O001O0mEULQ9m3lFYLo8i3mF^Ln8d3nFcLm8_3oFhLl8j4L5K6J6J5K6J5K6J2N1O001O1O001O1O1O001O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O002N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O2N2N2N2N2N2N2N2N2N2N2N2N2Na0_O7I1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1OO1O1O1O1O1O1O1M3M3L4M3M3L4MLhEfKS:]4oEcKe9f4]F[KV9P5lFPKg8Z5\\GfJb8Z5_GgJ_8Y5cGgJ\\8W5gGiJX8V5iGkJV8S5mGmJQ8S5QHmJn7R5SHoJl7o4WHQKh7n4ZHRKd7m4^HTKa7k4aHUK^7i4eHWKZ7h4hHXKW7g4jHZKU7d4nH\\KQ7c4QI]Kn6a4TI`Kk6_4WIaKh6]4[IcKd6\\4]IeKb6Z4`IfK_6X4dIhK[6W4fIjKY6U4iIkKV6U4kIkKT6T4mImKR6R4PJnKo5R4RJnKm5Q4TJPLl5o3UJQLj5n3XJRLh5l3YJULf5k3[JULd5j3^JVLb5h3_JYL`5g3aJYL^5f3dJZL\\5d3fJ\\LY5d3gJ]LY5a3iJ_LV5a3kJ_LT5`3mJaLS5]3oJcLP5]3QKcLn4\\3SKeLm4Y3UKgLj4Y3WKgLh4X3YKiLg4V3ZKjLe4U3]KkLc4S3^KnLa4R3`KnL`4Q3aKoLa4m2`KTMa4j2`KVMb4g2^KZMd4c2]K]Me4`2\\K`Me4^2[KcMg4Y2[KgMe4X2\\KhMc4X2]KiMb4W2_KiMa4V2`KjM_4V2aKkM^4U2cKkM\\4T2fKlMZ4S2fKnMY4R2hKnMW4R2iKoMV4Q2kKoMU4P2lKPNS4o1nKRNR4m1oKSNR4k1oKUNR4i1nKXNS4f1nKZNS4d1nK\\NS4b1mK_NT4^1nKbNS4\\1nKdNS4Z1mKgNT4W1mKiNT4U1lKlNU4R1lKnNU4o0mKQOT4m0lKTOU4i0mKWOT4f0nKZOR4d0oK]OR4`0PL@R4 in this image.", "objects": [{"patches": [78, 79, 80, 95, 96, 97, 112, 113, 114, 128, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 151, 161, 162, 163, 164, 165, 166, 167, 168, 178, 179, 180, 181, 182, 183, 184, 185, 195, 196, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 247, 248, 249, 250, 251, 265, 266, 267, 268, 282, 283, 284, 285, 299, 300, 301, 302, 316, 317, 318, 319, 334, 335, 336, 351, 352, 368, 369, 386], "bbox": [0.5107740585774059, 0.207109375, 0.9248535564853557, 0.97009375], "iscrowd": 0, "area": 48882.93950000001, "rle": {"size": [640, 478], "counts": "\\hh4k1Rb06K4L5K4K6L4K4L5L4K4M4K5L3L5L4K4M4K4M4Kg0ZO1N2O1N2N2O1N2O1N2O0O3N2M3N2M3N2M3N2N2N2M3N2N2N2M3ROZI_Dh6[;_IcDb6X;dIfD]6T;jIjDa6g:eIWE[6f:jIXEW6a0oHh8KZFS1:S6c0XI_8CcFV19P6d0_IW8[OkFY19m5d0eIQ8VORGZ18l5c0jIl7QOXG\\18i5c0oIi7kN\\G_15j5d0QJg7:bGg5f0TJc7h7[H\\Hb7e7\\H`H`7`7_HeH]7\\7aHhH\\7X7dHlHX7U7dHUIU7k6gHbIP7_6lHmIk6T6QIUKi5h9I1O1O1O1O2N1O1O1O1O1O1O1O1O1O001O001O1O001O001O1O001O1O0010O10000O100_BTKl;m4PDWKP in this image.", "objects": [{"patches": [78, 79, 80, 95, 96, 97, 112, 113, 114, 128, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 151, 161, 162, 163, 164, 165, 166, 167, 168, 178, 179, 180, 181, 182, 183, 184, 185, 195, 196, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 247, 248, 249, 250, 251, 265, 266, 267, 268, 282, 283, 284, 285, 299, 300, 301, 302, 316, 317, 318, 319, 334, 335, 336, 351, 352, 368, 369, 386], "bbox": [0.5107740585774059, 0.207109375, 0.9248535564853557, 0.97009375], "iscrowd": 0, "area": 48882.93950000001, "rle": {"size": [640, 478], "counts": "\\hh4k1Rb06K4L5K4K6L4K4L5L4K4M4K5L3L5L4K4M4K4M4Kg0ZO1N2O1N2N2O1N2O1N2O0O3N2M3N2M3N2M3N2N2N2M3N2N2N2M3ROZI_Dh6[;_IcDb6X;dIfD]6T;jIjDa6g:eIWE[6f:jIXEW6a0oHh8KZFS1:S6c0XI_8CcFV19P6d0_IW8[OkFY19m5d0eIQ8VORGZ18l5c0jIl7QOXG\\18i5c0oIi7kN\\G_15j5d0QJg7:bGg5f0TJc7h7[H\\Hb7e7\\H`H`7`7_HeH]7\\7aHhH\\7X7dHlHX7U7dHUIU7k6gHbIP7_6lHmIk6T6QIUKi5h9I1O1O1O1O2N1O1O1O1O1O1O1O1O1O001O001O1O001O001O1O001O1O0010O10000O100_BTKl;m4PDWKP in this image.", "objects": [{"patches": [73, 89, 90, 91, 106, 107, 108, 123, 124, 125, 126, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 224, 225, 226, 227, 228, 229, 230, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 276, 277, 278, 279, 293, 294, 295, 296, 310, 311, 312, 313, 327, 328, 329, 330, 345, 346, 347, 362, 363, 379, 380], "bbox": [0.1918410041841004, 0.212578125, 0.5414225941422595, 0.9702187500000001], "iscrowd": 0, "area": 46700.74655, "rle": {"size": [640, 478], "counts": "Uji1j2T?T2K4M4K4L5K5K4L5K4M4L3O1N1O1O1O1O1O1O1O1O100O1O1O1O1ON2I7HcNnI]Ej5k:XJSEe5o:_JnDh5j:\\JTEk5d:XJYEP6^:TJ_ES6Z:oIdEY6S:kIjE]6n9fIoEa6i9cITFe8c7_GZHf8a7\\G[Hj8`7ZG[Hm8`7VGZHR9`7RG[HT9`7k1K5L4L4L4L4L4L5K4L4L4L5WJQBW5S>cJQB[5S>`JoA_5[>N1O1O2N1N3N1O1O2N1O2N0000001O01O001O010UNTK`Dl4X;aKcD`4U;YK_C4W1c4R;\\KmC6k0_4b7kNhGkLb0Y4k7bNiGYM7U4U8WNkGhMKR4_8kMkGYN@l3j8aMlGgNUOi3S9VMnGUOkNd3\\9mLnGDaN`3e9bLPHm5]7hIiH[6X7[InHh6S7nHRIU7P7aHVIb7k6SH\\IP8X94M3M in this image.", "objects": [{"patches": [73, 89, 90, 91, 106, 107, 108, 123, 124, 125, 126, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 224, 225, 226, 227, 228, 229, 230, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 276, 277, 278, 279, 293, 294, 295, 296, 310, 311, 312, 313, 327, 328, 329, 330, 345, 346, 347, 362, 363, 379, 380], "bbox": [0.1918410041841004, 0.212578125, 0.5414225941422595, 0.9702187500000001], "iscrowd": 0, "area": 46700.74655, "rle": {"size": [640, 478], "counts": "Uji1j2T?T2K4M4K4L5K5K4L5K4M4L3O1N1O1O1O1O1O1O1O1O100O1O1O1O1ON2I7HcNnI]Ej5k:XJSEe5o:_JnDh5j:\\JTEk5d:XJYEP6^:TJ_ES6Z:oIdEY6S:kIjE]6n9fIoEa6i9cITFe8c7_GZHf8a7\\G[Hj8`7ZG[Hm8`7VGZHR9`7RG[HT9`7k1K5L4L4L4L4L4L5K4L4L4L5WJQBW5S>cJQB[5S>`JoA_5[>N1O1O2N1N3N1O1O2N1O2N0000001O01O001O010UNTK`Dl4X;aKcD`4U;YK_C4W1c4R;\\KmC6k0_4b7kNhGkLb0Y4k7bNiGYM7U4U8WNkGhMKR4_8kMkGYN@l3j8aMlGgNUOi3S9VMnGUOkNd3\\9mLnGDaN`3e9bLPHm5]7hIiH[6X7[InHh6S7nHRIU7P7aHVIb7k6SH\\IP8X94M3M in this image.", "objects": [{"patches": [27, 49, 50, 51, 72, 73, 74, 95, 96, 97, 118, 119, 140, 141, 142, 163, 164, 165, 166, 187, 189, 190], "bbox": [0.10571875, 0.1856701030927835, 0.26890625, 0.9057731958762888], "iscrowd": 0, "area": 8096.122899999997, "rle": {"size": [291, 640], "counts": "V`c03o84L3M3M3M4L3M3M4L3M3M1O2N2O1N1O2N2O0]IdN^5^1\\JfNd5_1VJcNj5`1QJaNo5c1lI^NT6Y2O00^NcI`0[6BhI:V4eNeMQ1YN7m3oNgMk0^N3k3UOdMi0cN0i3XObMi0gNLg3^OhL^1DPOd3DbLa1MhNa3I\\Lb16bN^3e2cLZM]3f2dLXM\\3i2fLUMY3j2jLTMW3l2kLRMT3P3lLnLU3S3kLkLU3W3lLgLS3\\3lLbLU3`3jL^LV3e3jLWLW3l3o02M3M4K6K3M2N2N2M3M2O1N2NSOQLkLi3X3bL_L]3a3eL^LZ3a3iL^LV3a3lL`LS3^3PMaLP3]3RMcLn2[3UMdLk2[3WMdLj2Y3XMhLg2V3\\MiLd2T3`MkL`2n2gMSMX2i0[Lc0d1dNP2g0_L>h1kNh1N\\L876]2Eo0IaL:62`2KW12[LOb2OS10]LNc22o0NbLLb26k0MhLG_2=h0K75II97FH<8CG>:A@f0`0o3000O101N101N2O1N3N5J3NO000O2O0M4J6K7I6JdhT4"}, "label": "a man wearing a blue tshirt & catch the plate"}]} {"id": 34680, "image": "COCO_train2014_000000034680.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a jumping boy in a blue shirt and khakis\"."}], "task": "refering", "answer_template": "The \"a jumping boy in a blue shirt and khakis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 49, 50, 51, 72, 73, 74, 95, 96, 97, 118, 119, 140, 141, 142, 163, 164, 165, 166, 187, 189, 190], "bbox": [0.10571875, 0.1856701030927835, 0.26890625, 0.9057731958762888], "iscrowd": 0, "area": 8096.122899999997, "rle": {"size": [291, 640], "counts": "V`c03o84L3M3M3M4L3M3M4L3M3M1O2N2O1N1O2N2O0]IdN^5^1\\JfNd5_1VJcNj5`1QJaNo5c1lI^NT6Y2O00^NcI`0[6BhI:V4eNeMQ1YN7m3oNgMk0^N3k3UOdMi0cN0i3XObMi0gNLg3^OhL^1DPOd3DbLa1MhNa3I\\Lb16bN^3e2cLZM]3f2dLXM\\3i2fLUMY3j2jLTMW3l2kLRMT3P3lLnLU3S3kLkLU3W3lLgLS3\\3lLbLU3`3jL^LV3e3jLWLW3l3o02M3M4K6K3M2N2N2M3M2O1N2NSOQLkLi3X3bL_L]3a3eL^LZ3a3iL^LV3a3lL`LS3^3PMaLP3]3RMcLn2[3UMdLk2[3WMdLj2Y3XMhLg2V3\\MiLd2T3`MkL`2n2gMSMX2i0[Lc0d1dNP2g0_L>h1kNh1N\\L876]2Eo0IaL:62`2KW12[LOb2OS10]LNc22o0NbLLb26k0MhLG_2=h0K75II97FH<8CG>:A@f0`0o3000O101N101N2O1N3N5J3NO000O2O0M4J6K7I6JdhT4"}, "label": "a jumping boy in a blue shirt and khakis"}]} {"id": 313209, "image": "COCO_train2014_000000313209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bread sandwich in table\"."}], "task": "refering", "answer_template": "The \"bread sandwich in table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 115, 116, 117, 118, 130, 131, 132, 133, 134, 135, 136, 137, 138, 148, 149, 150, 151, 152, 153, 154, 155, 156, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 220, 221, 222, 223, 224, 225, 226, 227, 228], "bbox": [0.20562, 0.48989333333333335, 0.74326, 0.9887733333333334], "iscrowd": 0, "area": 41938.017850000004, "rle": {"size": [375, 500], "counts": "iPV16\\;9F:F9H9F:G9F9I8I6K3L5L3L5L4K4L5L4K4M4K5M2N2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O100O1O100O100O1O100O1O100O100O1O100O100O010O1O100O100O100O100O1O100O100O1000000O10000O10000O10000O10000O100O100O10000O100O10000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000001O000000001O000000001O000000001O000000001O000000001O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1N101O1O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O1O001O1O2N1N3M2N3N1N3M2N3N2K4K6J5K6J5K6J5K6J5K6H7EU1kNTc^1"}, "label": "bread sandwich in table"}]} {"id": 313209, "image": "COCO_train2014_000000313209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of fried snack and tomato sauce in a green spoon\"."}], "task": "refering", "answer_template": "The \"a bowl of fried snack and tomato sauce in a green spoon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 115, 116, 117, 118, 130, 131, 132, 133, 134, 135, 136, 137, 138, 148, 149, 150, 151, 152, 153, 154, 155, 156, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 220, 221, 222, 223, 224, 225, 226, 227, 228], "bbox": [0.20562, 0.48989333333333335, 0.74326, 0.9887733333333334], "iscrowd": 0, "area": 41938.017850000004, "rle": {"size": [375, 500], "counts": "iPV16\\;9F:F9H9F:G9F9I8I6K3L5L3L5L4K4L5L4K4M4K5M2N2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O100O1O100O100O1O100O1O100O100O1O100O100O010O1O100O100O100O100O1O100O100O1000000O10000O10000O10000O10000O100O100O10000O100O10000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000001O000000001O000000001O000000001O000000001O000000001O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1N101O1O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O1O001O1O2N1N3M2N3N1N3M2N3N2K4K6J5K6J5K6J5K6J5K6H7EU1kNTc^1"}, "label": "a bowl of fried snack and tomato sauce in a green spoon"}]} {"id": 313209, "image": "COCO_train2014_000000313209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bowl of carrots\"."}], "task": "refering", "answer_template": "The \"the bowl of carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 70, 71, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 213, 214, 215], "bbox": [0.6668, 0.24655999999999997, 0.99928, 0.8781333333333333], "iscrowd": 0, "area": 30715.363849999998, "rle": {"size": [375, 500], "counts": "ZUj31_;8H9H8G9J5M4L4K5L3L5L4K5L3M4K5L3L5L4L4L3M4L4L4L3M4L4L4M4K4L4M3M3M3N3L3M3M3N2M3M2N2N101N1O1O2N101N1O1O2N100O2N1O2O0O1O2N1O101N1O2N1O101N1O2N10001O0O101O00001N10001O0O101O00001N10001O0O101O00001N10001O0O10001O000O2O00001N10001O0O101O00001N10001O0O101O00001N1000000O11O0O10001O0O101O00000O2O0000001N10001O0O101O001O0O2O00001O0O2O0F;lLS3mLK"}, "label": "the bowl of carrots"}]} {"id": 313209, "image": "COCO_train2014_000000313209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small bowl full of seasoned carrots located on the right side\"."}], "task": "refering", "answer_template": "The \"a small bowl full of seasoned carrots located on the right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 70, 71, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 213, 214, 215], "bbox": [0.6668, 0.24655999999999997, 0.99928, 0.8781333333333333], "iscrowd": 0, "area": 30715.363849999998, "rle": {"size": [375, 500], "counts": "ZUj31_;8H9H8G9J5M4L4K5L3L5L4K5L3M4K5L3L5L4L4L3M4L4L4L3M4L4L4M4K4L4M3M3M3N3L3M3M3N2M3M2N2N101N1O1O2N101N1O1O2N100O2N1O2O0O1O2N1O101N1O2N1O101N1O2N10001O0O101O00001N10001O0O101O00001N10001O0O101O00001N10001O0O10001O000O2O00001N10001O0O101O00001N10001O0O101O00001N1000000O11O0O10001O0O101O00000O2O0000001N10001O0O101O001O0O2O00001O0O2O0F;lLS3mLK"}, "label": "a small bowl full of seasoned carrots located on the right side"}]} {"id": 313209, "image": "COCO_train2014_000000313209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a corrot\"."}], "task": "refering", "answer_template": "The \"a corrot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 89, 103, 104, 107, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 157, 158, 159, 160, 161, 176, 177, 178, 179, 196, 197], "bbox": [0.7283, 0.33738666666666667, 1.0, 0.7860266666666667], "iscrowd": 0, "area": 12365.393200000002, "rle": {"size": [375, 500], "counts": "c_U41e;6K4K5K6K4TOXO`Fm0Z9VOdFo0U9UOjFT1k8PORG[1b8hN]G]1\\8fNcG\\1Z8fNeG[1Y8fNgG\\1FoM]8e0nG]1DnM^8e0nG`1Q8_NPHb1o7^NQHc1n7]NQHd1o7[NRHf1m7ZNSHg1l7YNTHg1l7XNUHi1i7nMaHS2^7mMbHT2]7kMdHU2\\7kMdHV2Z7jMkHS2T7mMPIP2o6PNTIm1l6RNYIk1e6VN_Ig1`6XNdIe1\\6[NhIb1W6]NmIa1Q6`NSJ]1l5bNYJZ1g5fN\\JX1c5gNbJV1\\5kNgJR1Y5mNlJP1S5POQKm0m4SOTKn0k4ROSKP1m4POQKS1n4lNPKX1n4iNPKZ1o4eNoJ]1R5cNlJ^1T5bNkJ`1U5`NhJc1X5\\NgJg1W5ZNfJj1Y5UNfJn1Y5RNeJQ2Y5oMeJV2Y5jMeJZ2Y5eMeJ^2Z5cMdJ`2[5`McJb2]5]MbJe2]5\\M`Jh2_5WM`Jk2`5UM^Jm2a5SM^Jo2b5QM[JS3d5lL[JV3d5kLZJW3f5iLZJW3f5hL[JY3c5hL\\JY3d5fL]JZ3b5gL^JY3b5fL_J[3_5fLaJZ3_5eLbJ[3]5fLcJZ3]5eLdJ[3\\5eLdJ\\3Z5eLfJ[3Z5dLgJ\\3X5eLhJ[3X5dLiJ\\3V5eLkJY3V5fLkJY3V5gLkJW3U5iLnJU3R5kLPKT3n4mLTKR3k4nLWKR3g4nL[KQ3d4oL]KQ3b4nL_KS3`4mL`KT3_4lLaKT3_4lLaKT3_4lLaKT3_4lLaKT3_4lLaKS3b6000001O1O1O1O001N2O1OO1M3L3O2N2O1N2N1O2nNgLeIZ3Y6S1O1O001O1OgMRIKn62UINk6OXI0h6N[I2e6K^I5b6HaI8^6FeI:[6DgI;Z6FeI:Z6HdI9\\6HbIeN>1g1"}, "label": "a corrot"}]} {"id": 264058, "image": "COCO_train2014_000000264058.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bed in a large room near a chair\"."}], "task": "refering", "answer_template": "The \"a bed in a large room near a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317], "bbox": [0.0, 0.537125, 0.834734375, 0.989025], "iscrowd": 0, "area": 68242.2033, "rle": {"size": [400, 640], "counts": "T7c0S1k2Y6f1000001N1000000000000O101O000000000O10000000000O1000000000000O1000000000000O1000000000000O100000000000000000000O100000001O0000000000000O100000000000000000000O10000000000000000000000O1000000O100000000O10000002N1O00001O00001O001O00001O00001O001O00001O00001O001O1O001O001O1O001O001O001O1O001O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O1O001O1O001O001O1O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000001O00000000000000000O1000000000000000000000000O10O100000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000001O0000000000000000000000001O0000000000000000000000000000000000000000000000000000001O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N2N2N2N1O1O001O001O001O1O001O001O001O001O1O001O1O1O1O1O1O2N1O1O1O1O1O1O1O100O1O1O2N1O1O1OePY1"}, "label": "a bed in a large room near a chair"}]} {"id": 264058, "image": "COCO_train2014_000000264058.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a colorful bed\"."}], "task": "refering", "answer_template": "The \"a colorful bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317], "bbox": [0.0, 0.537125, 0.834734375, 0.989025], "iscrowd": 0, "area": 68242.2033, "rle": {"size": [400, 640], "counts": "T7c0S1k2Y6f1000001N1000000000000O101O000000000O10000000000O1000000000000O1000000000000O1000000000000O100000000000000000000O100000001O0000000000000O100000000000000000000O10000000000000000000000O1000000O100000000O10000002N1O00001O00001O001O00001O00001O001O00001O00001O001O1O001O001O1O001O001O001O1O001O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O1O001O1O001O001O1O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000001O00000000000000000O1000000000000000000000000O10O100000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000001O0000000000000000000000001O0000000000000000000000000000000000000000000000000000001O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N2N2N2N1O1O001O001O001O1O001O001O001O001O1O001O1O1O1O1O1O2N1O1O1O1O1O1O1O100O1O1O2N1O1O1OePY1"}, "label": "a colorful bed"}]} {"id": 345981, "image": "COCO_train2014_000000345981.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with yellow backpack\"."}], "task": "refering", "answer_template": "The \"a man with yellow backpack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [244, 245, 260, 261, 262, 277, 278, 279, 294, 295, 296, 311, 312, 313, 328, 329, 330, 345, 346, 347, 361, 362, 363, 378, 379, 380], "bbox": [0.28241666666666665, 0.610125, 0.4298125, 1.0], "iscrowd": 0, "area": 10388.211550000002, "rle": {"size": [640, 480], "counts": "^ce28`c08H9G8H9M2N3M2M4M2N3TMYN[Ci1d<_NSCc1lCWA?h>]2O2N1I7G9G9F:E;@a0N100O1O100O1O2O0O100O1O100M4K4M3LkNZ1M2N2N2N2N2N2N2N2N2N2N3M2YMjABY>;SBYOo=oNeA>h08e=WOPB4g0:[=@YBKg0:Q=_ObABl0>n04gV1M\\6m2M]iZ5"}, "label": "a man with yellow backpack"}]} {"id": 345981, "image": "COCO_train2014_000000345981.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man on skis wearing a red head covering\"."}], "task": "refering", "answer_template": "The \"a man on skis wearing a red head covering\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [244, 245, 260, 261, 262, 277, 278, 279, 294, 295, 296, 311, 312, 313, 328, 329, 330, 345, 346, 347, 361, 362, 363, 378, 379, 380], "bbox": [0.28241666666666665, 0.610125, 0.4298125, 1.0], "iscrowd": 0, "area": 10388.211550000002, "rle": {"size": [640, 480], "counts": "^ce28`c08H9G8H9M2N3M2M4M2N3TMYN[Ci1d<_NSCc1lCWA?h>]2O2N1I7G9G9F:E;@a0N100O1O100O1O2O0O100O1O100M4K4M3LkNZ1M2N2N2N2N2N2N2N2N2N2N3M2YMjABY>;SBYOo=oNeA>h08e=WOPB4g0:[=@YBKg0:Q=_ObABl0>n04gV1M\\6m2M]iZ5"}, "label": "a man on skis wearing a red head covering"}]} {"id": 419714, "image": "COCO_train2014_000000419714.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an officer riding a horse and wearing a helmet\"."}], "task": "refering", "answer_template": "The \"an officer riding a horse and wearing a helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 44, 45, 62, 63, 79, 80, 81, 82, 98, 99, 100, 116, 117, 135, 136, 153, 170, 171, 188, 189, 190], "bbox": [0.43428, 0.08314732142857142, 0.596, 0.6561830357142858], "iscrowd": 0, "area": 9191.173599999995, "rle": {"size": [448, 500], "counts": "iQo21k=5L4L7bCEk:i0iDZOS;n0eDVOW;o0eDTOX;o0dDUOY;o0cDTOZ;k1M3N1N3M3K5L4K5O1O1N2O1O1N2O1N2`HeLX4]3cKgL\\4[3`KhL^4[3]KiLn3EnIe3o1iLQ4EnId3l1kLT4DnIc3i1mLX4BmIc3f1oL[4AlIc3e1oL^4i3]K[Lc4f3XK^Lg4d3UK^Ll4d3nJ`LQ5b3jJbLV5_3eJeL[5\\3cJeL\\5]3bJdL^5]3_JfL`5[3^JfLb5[3\\JfLc5o1QJYM:i0e5m1TJZM3k0i5k1UJZM0O6WOf5^3WJ[MKNh00_Nj4e0WK30e01cNi4`0[K8Ka05fNg4>\\K:I>6jNf4;_KC^N?X1d0?hN^4?_KE^N?W1a0m6VObIH\\Nb0S1 in this image.", "objects": [{"patches": [26, 27, 44, 45, 62, 63, 79, 80, 81, 82, 98, 99, 100, 116, 117, 135, 136, 153, 170, 171, 188, 189, 190], "bbox": [0.43428, 0.08314732142857142, 0.596, 0.6561830357142858], "iscrowd": 0, "area": 9191.173599999995, "rle": {"size": [448, 500], "counts": "iQo21k=5L4L7bCEk:i0iDZOS;n0eDVOW;o0eDTOX;o0dDUOY;o0cDTOZ;k1M3N1N3M3K5L4K5O1O1N2O1O1N2O1N2`HeLX4]3cKgL\\4[3`KhL^4[3]KiLn3EnIe3o1iLQ4EnId3l1kLT4DnIc3i1mLX4BmIc3f1oL[4AlIc3e1oL^4i3]K[Lc4f3XK^Lg4d3UK^Ll4d3nJ`LQ5b3jJbLV5_3eJeL[5\\3cJeL\\5]3bJdL^5]3_JfL`5[3^JfLb5[3\\JfLc5o1QJYM:i0e5m1TJZM3k0i5k1UJZM0O6WOf5^3WJ[MKNh00_Nj4e0WK30e01cNi4`0[K8Ka05fNg4>\\K:I>6jNf4;_KC^N?X1d0?hN^4?_KE^N?W1a0m6VObIH\\Nb0S1 in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 40, 41, 42, 43, 44, 63, 64, 65, 66, 67, 86, 87, 88, 89, 90, 109, 110, 111, 112, 113, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 178, 179, 180, 181, 201, 202, 203, 204, 227], "bbox": [0.73865625, 0.0016250000000000001, 0.9379062499999999, 0.5458958333333332], "iscrowd": 0, "area": 27654.678149999992, "rle": {"size": [480, 640], "counts": "lmm6?k=f0ZOg0YOf0ZOf0ZOf0ZOf0ZOg0B=N2M3M3N3L3M3N2M3M4M2M3M3N2M4L3N2M3M3N3N100000001O000000001O000000001O000000001O0000001O00000O101O00001O0000001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00N2H8F:G9G9G9G9F:G9G9G9F:G=Ca0_Ob0^Oa0^Oc0^Oa0_Ob0^Oa0_OYfb0"}, "label": "pinot noir bottle"}]} {"id": 329616, "image": "COCO_train2014_000000329616.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bottle of vignette on the right\"."}], "task": "refering", "answer_template": "The \"the bottle of vignette on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 40, 41, 42, 43, 44, 63, 64, 65, 66, 67, 86, 87, 88, 89, 90, 109, 110, 111, 112, 113, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 178, 179, 180, 181, 201, 202, 203, 204, 227], "bbox": [0.73865625, 0.0016250000000000001, 0.9379062499999999, 0.5458958333333332], "iscrowd": 0, "area": 27654.678149999992, "rle": {"size": [480, 640], "counts": "lmm6?k=f0ZOg0YOf0ZOf0ZOf0ZOf0ZOg0B=N2M3M3N3L3M3N2M3M4M2M3M3N2M4L3N2M3M3N3N100000001O000000001O000000001O000000001O0000001O00000O101O00001O0000001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00N2H8F:G9G9G9G9F:G9G9G9F:G=Ca0_Ob0^Oa0^Oc0^Oa0_Ob0^Oa0_OYfb0"}, "label": "the bottle of vignette on the right"}]} {"id": 329616, "image": "COCO_train2014_000000329616.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"aceto\"."}], "task": "refering", "answer_template": "The \"aceto\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 36, 37, 38, 39, 40, 59, 60, 61, 62, 63, 82, 83, 84, 85, 86, 105, 106, 107, 108, 109, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 177, 178, 200], "bbox": [0.60109375, 0.0, 0.757515625, 0.4776458333333334], "iscrowd": 0, "area": 21227.087850000004, "rle": {"size": [480, 640], "counts": "_ad5^4m9e0\\Od0[Of0I6O1O1O1O100O1O1O1O1O101N1O1000000000000000001O00000000000000000000001O000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000001O010O001O001O001O0D=VOj0VOi0WOj0VOi0WOj0VOi0WOj0VO_eX2"}, "label": "aceto"}]} {"id": 329616, "image": "COCO_train2014_000000329616.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bottle of aceton balsemita de modena\"."}], "task": "refering", "answer_template": "The \"bottle of aceton balsemita de modena\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 36, 37, 38, 39, 40, 59, 60, 61, 62, 63, 82, 83, 84, 85, 86, 105, 106, 107, 108, 109, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 177, 178, 200], "bbox": [0.60109375, 0.0, 0.757515625, 0.4776458333333334], "iscrowd": 0, "area": 21227.087850000004, "rle": {"size": [480, 640], "counts": "_ad5^4m9e0\\Od0[Of0I6O1O1O1O100O1O1O1O1O101N1O1000000000000000001O00000000000000000000001O000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000001O010O001O001O001O0D=VOj0VOi0WOj0VOi0WOj0VOi0WOj0VO_eX2"}, "label": "bottle of aceton balsemita de modena"}]} {"id": 337808, "image": "COCO_train2014_000000337808.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl in a white coat with her hood up , sits in a rocking chair reading\"."}], "task": "refering", "answer_template": "The \"girl in a white coat with her hood up , sits in a rocking chair reading\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 108, 109, 110, 111, 131, 132, 133, 134, 154, 155, 156, 157, 172, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 285, 286, 287, 289, 290, 291, 292, 293, 308, 309, 310, 314, 332, 333], "bbox": [0.42471875, 0.17527083333333332, 0.87134375, 0.8786458333333333], "iscrowd": 0, "area": 39953.7936, "rle": {"size": [480, 640], "counts": "cjo3V1i=5K4L2N3L3N3M[NlBb1Q=\\NQCf1mG8O1O1O2N101O0010O010O010O010O01O010O010O010O0M4L3N3N1\\NfC`0]<_OdC>^g:BVE`0j:ASEb0m:]OQEf0o:ZOnDh0S;WOkDl0U;TOhDo0W;QOhDQ1X;oNhDP1Y;oNgDR1Y;nNgDR1X;nNhDR1Y;mNhDS1X;mNgDS1Z;lNfDU1Z;jNgDV1X;kNgDU1Z;jNgDV1Y;iNgDX1Y;hNgDW1Z;hNfDY1Y;gNhDY1X;fNhDZ1X;gNgDZ1X:oMWFh0BX1R:WNXFc0FW1P:YNWFb0IU1o9[NTFb0MS1n9^NRF`01R1l9`NoEa04P1k9bNnE?8n0i9fNkE>a0k0e92ZFNf92ZFOd92]FMb95]FLa95`FJ_97aFI_98aFG^9:bFF]9;dFC\\9>dFB[9`0dF@[9a0fF^OY9c0gF]OX9e0hFZOW9g0iFYOV9h0kFWOU9j0jFVOU9k0lFSOT9n0lFROS9o0mFQO^8ZN]Gg25nN\\8`N[Gc29mNZ8dNZG`2=kNX8iNWG^2`0iNc6gNmI6mN[2d0gN_6lNmIV3DnM\\6POmIU3GiMZ6WOkIR3KgMY6YOiIR3OdMW6\\OhIQ31cMV6_OeIQ34`MU6BdIP38]MS6EaIQ3 in this image.", "objects": [{"patches": [85, 86, 87, 108, 109, 110, 111, 131, 132, 133, 134, 154, 155, 156, 157, 172, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 285, 286, 287, 289, 290, 291, 292, 293, 308, 309, 310, 314, 332, 333], "bbox": [0.42471875, 0.17527083333333332, 0.87134375, 0.8786458333333333], "iscrowd": 0, "area": 39953.7936, "rle": {"size": [480, 640], "counts": "cjo3V1i=5K4L2N3L3N3M[NlBb1Q=\\NQCf1mG8O1O1O2N101O0010O010O010O010O01O010O010O010O0M4L3N3N1\\NfC`0]<_OdC>^g:BVE`0j:ASEb0m:]OQEf0o:ZOnDh0S;WOkDl0U;TOhDo0W;QOhDQ1X;oNhDP1Y;oNgDR1Y;nNgDR1X;nNhDR1Y;mNhDS1X;mNgDS1Z;lNfDU1Z;jNgDV1X;kNgDU1Z;jNgDV1Y;iNgDX1Y;hNgDW1Z;hNfDY1Y;gNhDY1X;fNhDZ1X;gNgDZ1X:oMWFh0BX1R:WNXFc0FW1P:YNWFb0IU1o9[NTFb0MS1n9^NRF`01R1l9`NoEa04P1k9bNnE?8n0i9fNkE>a0k0e92ZFNf92ZFOd92]FMb95]FLa95`FJ_97aFI_98aFG^9:bFF]9;dFC\\9>dFB[9`0dF@[9a0fF^OY9c0gF]OX9e0hFZOW9g0iFYOV9h0kFWOU9j0jFVOU9k0lFSOT9n0lFROS9o0mFQO^8ZN]Gg25nN\\8`N[Gc29mNZ8dNZG`2=kNX8iNWG^2`0iNc6gNmI6mN[2d0gN_6lNmIV3DnM\\6POmIU3GiMZ6WOkIR3KgMY6YOiIR3OdMW6\\OhIQ31cMV6_OeIQ34`MU6BdIP38]MS6EaIQ3 in this image.", "objects": [{"patches": [176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272], "bbox": [0.5937187500000001, 0.5101170960187353, 0.835109375, 0.7707962529274005], "iscrowd": 0, "area": 13579.116550000004, "rle": {"size": [427, 640], "counts": "Pfn42h04c;0VD;a;JXDb0_;BYDj0_;d0M0O10001O1N2O1O1O1N2O1O1O1N2O001O001N2O001O1N101O1O0O2O001O1N101O001N101O1O0O2O001O0O2O00000O1000000O1000001N100000000O10001O0O1000000O101O000000000000000000000010O00000000000000000000000000000TOUFXNk9h1YFTNh9k1\\FQNd9n1^FQNb9o1_FPNa9P2_FQN`9o1aFPN_9o1bFQN^9o1cFPN]9P2dFoM\\9Q2dFoM\\9P2fFoMZ9P2gFPNY9P2hFoMX9P2jFoMV9P2lFoMT9P2nFoMR9P2PGoMQ9P2PGoMP9P2RGlMQ9T2S11O1O1O1O1O1O001O2N2N3M2N2N2O1O12O1N2N2N2M3L4M3L4L4L4L4L4M3L4L4L_m[1"}, "label": "an orange bus in between two other buses"}]} {"id": 141207, "image": "COCO_train2014_000000141207.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bus in the middle\"."}], "task": "refering", "answer_template": "The \"the bus in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272], "bbox": [0.5937187500000001, 0.5101170960187353, 0.835109375, 0.7707962529274005], "iscrowd": 0, "area": 13579.116550000004, "rle": {"size": [427, 640], "counts": "Pfn42h04c;0VD;a;JXDb0_;BYDj0_;d0M0O10001O1N2O1O1O1N2O1O1O1N2O001O001N2O001O1N101O1O0O2O001O1N101O001N101O1O0O2O001O0O2O00000O1000000O1000001N100000000O10001O0O1000000O101O000000000000000000000010O00000000000000000000000000000TOUFXNk9h1YFTNh9k1\\FQNd9n1^FQNb9o1_FPNa9P2_FQN`9o1aFPN_9o1bFQN^9o1cFPN]9P2dFoM\\9Q2dFoM\\9P2fFoMZ9P2gFPNY9P2hFoMX9P2jFoMV9P2lFoMT9P2nFoMR9P2PGoMQ9P2PGoMP9P2RGlMQ9T2S11O1O1O1O1O1O001O2N2N3M2N2N2O1O12O1N2N2N2M3L4M3L4L4L4L4L4M3L4L4L_m[1"}, "label": "the bus in the middle"}]} {"id": 378791, "image": "COCO_train2014_000000378791.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby giraffe milking\"."}], "task": "refering", "answer_template": "The \"baby giraffe milking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 73, 74, 75, 84, 85, 86, 87, 88, 96, 97, 98, 99, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125, 133, 134, 135, 136, 137, 138, 146, 147, 148, 149, 150, 159, 160, 161, 162, 163, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 224, 225, 226, 227], "bbox": [0.18465181058495825, 0.23596, 0.8356545961002786, 0.98876], "iscrowd": 0, "area": 34948.072199999995, "rle": {"size": [500, 359], "counts": "RdP11b?4L4M3L4L3M7J7H8H6J100O001O1O0cLkN]HU1a7FeG:[8;PGFn8i0eFWOY9X1YFhNg9]1UFcNi9c1SFcNe9c1WFeN`9`1\\F^Od8h0XGXOf8l3O1N2N2O1N2N2O1N3M2O1N2N2O1OZO`HZI_7b6l0`NYGnKk8l3_GnKc8m3gGlK[8o3nGkKT8P4UHjKl7S4\\HgKf7T4cHgK^7U4R2L4L4L4L3M4M3M34L4L4WH]LY3h3]L_Lc3c3SLfLl3]3iKkLW4X3iIPL7R1P6Q3]I^LKR1g6c2RIROn6Q1fHXOZ7b4O00QOn02N101N1O2N1O2N3M4L3M4L4L3M4L4L4L3M2N2N1O1OYO\\I`Hc6W7mIdHQ6T7]JgHb5V7hJdHX5X7RKbHm4\\7e1N100O1O2O2M2K6J6I7J6I7M36THnJk4W5eJjKe4[4jJRLU5T4ZJXLe5l6O2N10000O1000000O10ZOf0K5PNhI_J\\6^5QJTJT6i5WJkIm5Q6`JaIe5\\6fJWI_5f6i1L4M3L4M4L3M3M3M3M3L4M3M4]LlEa1W:^NlE]1W:bNmEX1V:hNlES1W:lNmEm0W:ROlEh0X:WOkEd0Y:ZOkE?Y:@jE:Z:EjE5Y:JjE0Z:OjEJP:`0RFZOQ:f0SFSOP:m0ZFfNi9Z1aFYNa9h1U20000001O00000000000001O01O0001O0001O01O0001N1O1O1O2O0O1N2N3N1N21O100O100000000000001O000000000000000000000000002N5WN^BV1f=cN_B[1n=O2N2N2N2N1O2N2N2N2N1O2N2H8H8Hidl0"}, "label": "baby giraffe milking"}]} {"id": 378791, "image": "COCO_train2014_000000378791.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby giraffe\"."}], "task": "refering", "answer_template": "The \"a baby giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 73, 74, 75, 84, 85, 86, 87, 88, 96, 97, 98, 99, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125, 133, 134, 135, 136, 137, 138, 146, 147, 148, 149, 150, 159, 160, 161, 162, 163, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 224, 225, 226, 227], "bbox": [0.18465181058495825, 0.23596, 0.8356545961002786, 0.98876], "iscrowd": 0, "area": 34948.072199999995, "rle": {"size": [500, 359], "counts": "RdP11b?4L4M3L4L3M7J7H8H6J100O001O1O0cLkN]HU1a7FeG:[8;PGFn8i0eFWOY9X1YFhNg9]1UFcNi9c1SFcNe9c1WFeN`9`1\\F^Od8h0XGXOf8l3O1N2N2O1N2N2O1N3M2O1N2N2O1OZO`HZI_7b6l0`NYGnKk8l3_GnKc8m3gGlK[8o3nGkKT8P4UHjKl7S4\\HgKf7T4cHgK^7U4R2L4L4L4L3M4M3M34L4L4WH]LY3h3]L_Lc3c3SLfLl3]3iKkLW4X3iIPL7R1P6Q3]I^LKR1g6c2RIROn6Q1fHXOZ7b4O00QOn02N101N1O2N1O2N3M4L3M4L4L3M4L4L4L3M2N2N1O1OYO\\I`Hc6W7mIdHQ6T7]JgHb5V7hJdHX5X7RKbHm4\\7e1N100O1O2O2M2K6J6I7J6I7M36THnJk4W5eJjKe4[4jJRLU5T4ZJXLe5l6O2N10000O1000000O10ZOf0K5PNhI_J\\6^5QJTJT6i5WJkIm5Q6`JaIe5\\6fJWI_5f6i1L4M3L4M4L3M3M3M3M3L4M3M4]LlEa1W:^NlE]1W:bNmEX1V:hNlES1W:lNmEm0W:ROlEh0X:WOkEd0Y:ZOkE?Y:@jE:Z:EjE5Y:JjE0Z:OjEJP:`0RFZOQ:f0SFSOP:m0ZFfNi9Z1aFYNa9h1U20000001O00000000000001O01O0001O0001O01O0001N1O1O1O2O0O1N2N3N1N21O100O100000000000001O000000000000000000000000002N5WN^BV1f=cN_B[1n=O2N2N2N2N1O2N2N2N2N1O2N2H8H8Hidl0"}, "label": "a baby giraffe"}]} {"id": 378791, "image": "COCO_train2014_000000378791.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the side of a giraffe\"."}], "task": "refering", "answer_template": "The \"the side of a giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 21, 22, 23, 24, 25, 34, 35, 36, 37, 38, 47, 48, 49, 50, 51, 60, 61, 62, 63, 64, 73, 75, 76, 77, 88, 89, 90, 100, 101, 102, 103, 113, 114, 116, 126, 127, 129, 139, 140, 142, 152, 153, 155, 165, 166, 168, 178, 179, 181, 191, 192, 194, 204, 205, 207, 217, 218, 220, 230, 231], "bbox": [0.6353481894150418, 0.0, 1.0, 0.98652], "iscrowd": 0, "area": 32649.41494999997, "rle": {"size": [500, 359], "counts": "g^_32P>9YCIU<\\1UCgNY in this image.", "objects": [{"patches": [9, 10, 11, 12, 21, 22, 23, 24, 25, 34, 35, 36, 37, 38, 47, 48, 49, 50, 51, 60, 61, 62, 63, 64, 73, 75, 76, 77, 88, 89, 90, 100, 101, 102, 103, 113, 114, 116, 126, 127, 129, 139, 140, 142, 152, 153, 155, 165, 166, 168, 178, 179, 181, 191, 192, 194, 204, 205, 207, 217, 218, 220, 230, 231], "bbox": [0.6353481894150418, 0.0, 1.0, 0.98652], "iscrowd": 0, "area": 32649.41494999997, "rle": {"size": [500, 359], "counts": "g^_32P>9YCIU<\\1UCgNY in this image.", "objects": [{"patches": [133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 176, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 344], "bbox": [0.61184375, 0.35873536299765807, 0.996296875, 0.9730913348946135], "iscrowd": 0, "area": 35837.770049999985, "rle": {"size": [427, 640], "counts": "XgS51X=100O2N100O1O2N100O1N3M1O1O001O1O003M5N1N2O1O1N2O1O1`DROa:o0[ETOd:n0XEVOg:k0UEYOj:h0RE[Om:g0oD]OP;e0lD^OS;\\11M2O2M3N2M2N3N2M4M2M31O1O100O001O1O100O1O1O001O100O1O1O1O10O01O1O1O100O1O1O001gETNQ9m1kFWNU9i1hFYNX9h1dF\\NZ9f1bF\\N_9e1]F\\Ne9d1XF]Ni9e1TFZNo9g1mEZNU:g1gEYN[:Z22M3N1N3N20O2O001N2O0N3N2M3N1O2M3N1N3N2M3M2N3L4M2M4M3L4L3N3L4M2M4L4M3L3N3N2N1N3N2N2N1O2M3N1O2N2M3O001O1O010O1O1O001O1O001O1O1O001O1VJPJX4o5^KZJc4g5oJfJQ5[5_JSKb5[60O100O100O100O100O2N100O1L4M3M3M3M3L4M3O11O0010O01O1O001O1O001O1O001O1O010O001O1O001O1O001O1O001O10O01O001O1O001O1O001O1O00100O001O1O001O001O1O001O1O010O1O001O1O001O001O1O001O1O010O1O001S7lH`>"}, "label": "the bench the bird is walking under"}]} {"id": 346026, "image": "COCO_train2014_000000346026.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second bench in the right\"."}], "task": "refering", "answer_template": "The \"the second bench in the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 197, 198, 220], "bbox": [0.502640625, 0.3137939110070258, 0.798828125, 0.6097423887587823], "iscrowd": 0, "area": 12895.160050000004, "rle": {"size": [427, 640], "counts": "Z^V43V=3L4L4O1WCE` in this image.", "objects": [{"patches": [105, 106, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 197, 198, 220], "bbox": [0.502640625, 0.3137939110070258, 0.798828125, 0.6097423887587823], "iscrowd": 0, "area": 12895.160050000004, "rle": {"size": [427, 640], "counts": "Z^V43V=3L4L4O1WCE` in this image.", "objects": [{"patches": [106, 107, 108, 109, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 225, 240, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 292, 293, 294, 314, 315, 316, 337, 338, 339], "bbox": [0.431796875, 0.267423887587822, 0.8381093750000002, 0.9505620608899297], "iscrowd": 0, "area": 42877.30065000001, "rle": {"size": [427, 640], "counts": "\\[c31Z=3L4L5K3N0O1O2N1O100O1O2O000O10001O0O1000001N10000mDYOb9g0\\F^O`9d0^FA^9?`FG\\98cFMZ93eF2V9NiF7T9IjF=R9BnFc0n8]OPGf0n8ZOQGh0o8XOQGh0o8WORGi0m8XOSGh0l8YOTGg0l8XOUGh0j8YOVGg0i8YOXGg0h8YOXGg0g8YOZGg0e8ZO\\Ge0d8ZO]Gf0b8[O^Ge0b8ZO_Gf0`8[O`Ge0_8[ObGe0^8[ObGe0]8\\OcGd0\\8\\OeGd0[8\\OeGd0Z8\\OgGd0X8]OhGc0X8]OhGc0W8^OiGb0V8_OkG`0U8@kG`0T8AlG?S8BmG>S8AnG?Q8BoG>Q8BoG>P8CPH=o7DQH in this image.", "objects": [{"patches": [106, 107, 108, 109, 128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 225, 240, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 289, 290, 291, 292, 293, 294, 314, 315, 316, 337, 338, 339], "bbox": [0.431796875, 0.267423887587822, 0.8381093750000002, 0.9505620608899297], "iscrowd": 0, "area": 42877.30065000001, "rle": {"size": [427, 640], "counts": "\\[c31Z=3L4L5K3N0O1O2N1O100O1O2O000O10001O0O1000001N10000mDYOb9g0\\F^O`9d0^FA^9?`FG\\98cFMZ93eF2V9NiF7T9IjF=R9BnFc0n8]OPGf0n8ZOQGh0o8XOQGh0o8WORGi0m8XOSGh0l8YOTGg0l8XOUGh0j8YOVGg0i8YOXGg0h8YOXGg0g8YOZGg0e8ZO\\Ge0d8ZO]Gf0b8[O^Ge0b8ZO_Gf0`8[O`Ge0_8[ObGe0^8[ObGe0]8\\OcGd0\\8\\OeGd0[8\\OeGd0Z8\\OgGd0X8]OhGc0X8]OhGc0W8^OiGb0V8_OkG`0U8@kG`0T8AlG?S8BmG>S8AnG?Q8BoG>Q8BoG>P8CPH=o7DQH in this image.", "objects": [{"patches": [34, 35, 55, 56, 57, 58, 59, 60, 78, 79, 80, 81, 82, 83, 84, 100, 101, 102, 103, 104, 105, 106, 107, 121, 122, 123, 124, 125, 126, 127, 128, 144, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 285, 286], "bbox": [0.233890625, 0.11236533957845432, 0.6776875, 0.8404449648711944], "iscrowd": 0, "area": 45620.47725000002, "rle": {"size": [427, 640], "counts": "]kn18l<8G:K4M3L4M3M3L4M4K4M3M3N1O1O1O1O1O2O0WMSNSJn1j5\\NnIe1n5eNkI\\1Q6nNhIS1T6WOeIj0Z6XOdIi0\\6XObIi0]6YOaIi0^6WOaIj0^6XO`Ii0`6WO_Ij0`6XO^Ii0b6XO\\Ii0c6XO\\Ii0d6XOZIi0e6XOZIi0f6XOXIj0f6WOYIj0f6XOXIi0h6XOVIi0i6XOWIh0i6YOUIh0j6YOUIh0k6YOSIh0l6ZORIg0n6YOPIj0n6XOPIi0P7WOnHk0Q7h200O1O100O101N00100O100O1O100O010O010O010O10O010O010O10O010OO2O0O2N101N2O0O2M2N3L3M4L3M4M3NmHSKk5m4TJYKh5f4VJaKf5^4YJgKc5Y4[JkKd5T4[JnKe5l5O2O000O101N10001N1SO^IQKb6m4`ISK`6l4bIRK`6k4bIUK^6i4eIVK[6h4gIWK[6g4gIXKY6f4iIZKW6e4kIYKW6e4jI[KV6e4kIZKU6e4lIVKY6j4U1N100O2O0O100OPMhKPMV4Q3kKnLT4S3mKlLR4U3oKkLo3V3SLhLl3Y3ULfLj3Z3XLfLf3[3[LdLd3]3]LbLb3_3`L_L_3b3bL^L\\3c3eL\\LZ3e3gLZLX3g3iLWLW3i3lLULS3l3mLTLS3l3mLTLR3m3nLSLQ3n3nLSLR3m3nLRLR3o3nLQLR3o3nLQLQ3P4oLXKh3h4YLVKi3j4WLSKk3n4ULPKm3P5SLnJn3R5SLlJn3U5RLiJP4V5QLiJo3X5PLhJP4Y5PLgJP4X5QLhJn3Y5RLgJn3X5SLhJl3Y5TLgJk3Y5VLgJj3Y5VLgJi3Z5WLfJi3Y5XLgJh3Y5XLgJh3X5YLhJg3X5XLhJi3X5WLhJi3W5XLhJi3X5WLhJi3W5XLhJi3X5WLgJj3Y5VLgJj3X5XLfJi3Z5XLeJh3Z5ZLdJg3\\5ZL_Jj3`5WLYJP4g5QLQJV4o5c1O2O0O10000O101N100O10000O2O0O101O0O2O0O2O1O0O2O0O2O0O2O001N101N10001N101oNlGhLU8S3VHgLj7T3aHeLa7U3jHeLV7V3UIdLl6V3g1I7H9N101N1O1O2N2O1N1O2N2N101N2N1O2N2O0O2N1O2O0O1O2N1O2O0O2N1O1]N^D]1j;M2N3M2O2M2O2M2O001N2O1O1N101N2O1O0O2O1N2O1O0O2O4K6KSie2"}, "label": "a toy teddy bear holding a girl baby toy"}]} {"id": 34739, "image": "COCO_train2014_000000034739.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"some figures made on rock\"."}], "task": "refering", "answer_template": "The \"some figures made on rock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 55, 56, 57, 58, 59, 60, 78, 79, 80, 81, 82, 83, 84, 100, 101, 102, 103, 104, 105, 106, 107, 121, 122, 123, 124, 125, 126, 127, 128, 144, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 285, 286], "bbox": [0.233890625, 0.11236533957845432, 0.6776875, 0.8404449648711944], "iscrowd": 0, "area": 45620.47725000002, "rle": {"size": [427, 640], "counts": "]kn18l<8G:K4M3L4M3M3L4M4K4M3M3N1O1O1O1O1O2O0WMSNSJn1j5\\NnIe1n5eNkI\\1Q6nNhIS1T6WOeIj0Z6XOdIi0\\6XObIi0]6YOaIi0^6WOaIj0^6XO`Ii0`6WO_Ij0`6XO^Ii0b6XO\\Ii0c6XO\\Ii0d6XOZIi0e6XOZIi0f6XOXIj0f6WOYIj0f6XOXIi0h6XOVIi0i6XOWIh0i6YOUIh0j6YOUIh0k6YOSIh0l6ZORIg0n6YOPIj0n6XOPIi0P7WOnHk0Q7h200O1O100O101N00100O100O1O100O010O010O010O10O010O010O10O010OO2O0O2N101N2O0O2M2N3L3M4L3M4M3NmHSKk5m4TJYKh5f4VJaKf5^4YJgKc5Y4[JkKd5T4[JnKe5l5O2O000O101N10001N1SO^IQKb6m4`ISK`6l4bIRK`6k4bIUK^6i4eIVK[6h4gIWK[6g4gIXKY6f4iIZKW6e4kIYKW6e4jI[KV6e4kIZKU6e4lIVKY6j4U1N100O2O0O100OPMhKPMV4Q3kKnLT4S3mKlLR4U3oKkLo3V3SLhLl3Y3ULfLj3Z3XLfLf3[3[LdLd3]3]LbLb3_3`L_L_3b3bL^L\\3c3eL\\LZ3e3gLZLX3g3iLWLW3i3lLULS3l3mLTLS3l3mLTLR3m3nLSLQ3n3nLSLR3m3nLRLR3o3nLQLR3o3nLQLQ3P4oLXKh3h4YLVKi3j4WLSKk3n4ULPKm3P5SLnJn3R5SLlJn3U5RLiJP4V5QLiJo3X5PLhJP4Y5PLgJP4X5QLhJn3Y5RLgJn3X5SLhJl3Y5TLgJk3Y5VLgJj3Y5VLgJi3Z5WLfJi3Y5XLgJh3Y5XLgJh3X5YLhJg3X5XLhJi3X5WLhJi3W5XLhJi3X5WLhJi3W5XLhJi3X5WLgJj3Y5VLgJj3X5XLfJi3Z5XLeJh3Z5ZLdJg3\\5ZL_Jj3`5WLYJP4g5QLQJV4o5c1O2O0O10000O101N100O10000O2O0O101O0O2O0O2O1O0O2O0O2O0O2O001N101N10001N101oNlGhLU8S3VHgLj7T3aHeLa7U3jHeLV7V3UIdLl6V3g1I7H9N101N1O1O2N2O1N1O2N2N101N2N1O2N2O0O2N1O2O0O1O2N1O2O0O2N1O1]N^D]1j;M2N3M2O2M2O2M2O001N2O1O1N101N2O1O0O2O1N2O1O0O2O4K6KSie2"}, "label": "some figures made on rock"}]} {"id": 444344, "image": "COCO_train2014_000000444344.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large pizza sitting on stand on the left side of the table\"."}], "task": "refering", "answer_template": "The \"large pizza sitting on stand on the left side of the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257], "bbox": [0.0015625, 0.4791666666666667, 0.3359375, 0.8847222222222222], "iscrowd": 0, "area": 23181.0, "rle": {"size": [360, 640], "counts": "mb0h0U:ZOUF_1S9dNiF\\1W9hNeFW1]9:1O0O01O1O0\\OWNiGj1U8YNiGh1V8YNiGh1U8ZNjGf1U8]NiGd1U8^NjGc1U8_NjG`1U8bNiG`1U8bNjG_1T8cNjG^1V8cNhG_1V8cNiG^1U8dNiG]1V8l0N2O0O1O1O101N1O1O100O2N100O1O100O2N100O1O101O00000O101O000O1000001O0O1000001N100000000O2OO1000O1000000O10000O01000O1000000O10000O01000O1000000O10000O10O10O1000000O10000O10O10O1000000O100000O01000O10O10O100O01000O10O10O100O01000O10000O100O10001N10000O100O2O000O10000O2O0O10000O1O2O0O1O2N1O2O0O2N1O2O0O2N1O1O2O0O2N1O2O0O2N1O2N101N1O2N1O2O0O2N1N2N3M2N3N1N3M2N4L5L5J5K5K6J5K5L5J5K5K6JQVe4"}, "label": "large pizza sitting on stand on the left side of the table"}]} {"id": 444344, "image": "COCO_train2014_000000444344.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza with green vegetables\"."}], "task": "refering", "answer_template": "The \"pizza with green vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257], "bbox": [0.0015625, 0.4791666666666667, 0.3359375, 0.8847222222222222], "iscrowd": 0, "area": 23181.0, "rle": {"size": [360, 640], "counts": "mb0h0U:ZOUF_1S9dNiF\\1W9hNeFW1]9:1O0O01O1O0\\OWNiGj1U8YNiGh1V8YNiGh1U8ZNjGf1U8]NiGd1U8^NjGc1U8_NjG`1U8bNiG`1U8bNjG_1T8cNjG^1V8cNhG_1V8cNiG^1U8dNiG]1V8l0N2O0O1O1O101N1O1O100O2N100O1O100O2N100O1O101O00000O101O000O1000001O0O1000001N100000000O2OO1000O1000000O10000O01000O1000000O10000O01000O1000000O10000O10O10O1000000O10000O10O10O1000000O100000O01000O10O10O100O01000O10O10O100O01000O10000O100O10001N10000O100O2O000O10000O2O0O10000O1O2O0O1O2N1O2O0O2N1O2O0O2N1O1O2O0O2N1O2O0O2N1O2N101N1O2N1O2O0O2N1N2N3M2N3N1N3M2N4L5L5J5K5K6J5K5L5J5K5K6JQVe4"}, "label": "pizza with green vegetables"}]} {"id": 444344, "image": "COCO_train2014_000000444344.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza with black olives\"."}], "task": "refering", "answer_template": "The \"a pizza with black olives\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222], "bbox": [0.4110625, 0.4165, 0.726421875, 0.7525], "iscrowd": 0, "area": 18800.946750000003, "rle": {"size": [360, 640], "counts": "ddl2`0d:>D4L3N3M2M3N3M2M4M2N2M3N2N2M3O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O001N10001N101O001N101O001N101O001N101O000O2O001O0O1000001N1000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001O00000010O000000000O10000000000O2O0000001O0O101O0000001N1000001O00000O2O0000001O0O10001O00001N101O001O001N2O001O001O0O2O001O1O001O0O2O001O001O0O2O1N1O2N1O2N1O2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2M3N2M3M8I7H8I6ImYm1"}, "label": "a pizza with black olives"}]} {"id": 444344, "image": "COCO_train2014_000000444344.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza with black olives and pepperonis\"."}], "task": "refering", "answer_template": "The \"pizza with black olives and pepperonis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 193, 194, 195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222], "bbox": [0.4110625, 0.4165, 0.726421875, 0.7525], "iscrowd": 0, "area": 18800.946750000003, "rle": {"size": [360, 640], "counts": "ddl2`0d:>D4L3N3M2M3N3M2M4M2N2M3N2N2M3O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O001N10001N101O001N101O001N101O001N101O000O2O001O0O1000001N1000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001O00000010O000000000O10000000000O2O0000001O0O101O0000001N1000001O00000O2O0000001O0O10001O00001N101O001O001N2O001O001O0O2O001O1O001O0O2O001O001O0O2O1N1O2N1O2N1O2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N2M3N2M3M8I7H8I6ImYm1"}, "label": "pizza with black olives and pepperonis"}]} {"id": 444344, "image": "COCO_train2014_000000444344.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty chair\"."}], "task": "refering", "answer_template": "The \"an empty chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 39, 40, 41, 42, 43, 62, 63, 64, 65, 66, 86, 87, 88, 110, 111], "bbox": [0.6958593750000001, 0.019944444444444445, 0.9085625, 0.38], "iscrowd": 0, "area": 7118.387749999998, "rle": {"size": [360, 640], "counts": "g`l41Q;5K6I7J501O10O100O1M3M3O12N2N1O2N2N1O1O0000000000001O000000001O0O2M2M4L3M4L3N3L3M4L3N3L3M3M4L3N2M4O00010O00001O0000001O01OjG]Mg7d2nGhMP8j2N2NM3N2M3N2M4L3N2M31O000000010bH^MT6b2jI`MV6`2gIbMZ6^2cIeM]6[2aIfM`6[2\\IgMe6Y2YIiMg6KcHR2c0UNk6HeHQ2=ZNo6DeHQ2:\\NR7BfHR24_NW7^OgHQ2OcN]7YOQIh1@POX8P1eGSO\\8l0bGSOa8h13L3N2N3L3N3M2M3N3L3N3M2M3N3L3N3M2MPOmNRHP1Q8POoGl0T8UOkGi0W8XOiGe0Y8\\OgG`0]8@cG=_8DaG8b8I]G5e8L[G1h8OXGMk84UGIm88Q11O1O1O100O1O1O1O00M3M3L4M[gd0"}, "label": "an empty chair"}]} {"id": 444344, "image": "COCO_train2014_000000444344.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black bar chair in the background\"."}], "task": "refering", "answer_template": "The \"a black bar chair in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 39, 40, 41, 42, 43, 62, 63, 64, 65, 66, 86, 87, 88, 110, 111], "bbox": [0.6958593750000001, 0.019944444444444445, 0.9085625, 0.38], "iscrowd": 0, "area": 7118.387749999998, "rle": {"size": [360, 640], "counts": "g`l41Q;5K6I7J501O10O100O1M3M3O12N2N1O2N2N1O1O0000000000001O000000001O0O2M2M4L3M4L3N3L3M4L3N3L3M3M4L3N2M4O00010O00001O0000001O01OjG]Mg7d2nGhMP8j2N2NM3N2M3N2M4L3N2M31O000000010bH^MT6b2jI`MV6`2gIbMZ6^2cIeM]6[2aIfM`6[2\\IgMe6Y2YIiMg6KcHR2c0UNk6HeHQ2=ZNo6DeHQ2:\\NR7BfHR24_NW7^OgHQ2OcN]7YOQIh1@POX8P1eGSO\\8l0bGSOa8h13L3N2N3L3N3M2M3N3L3N3M2M3N3L3N3M2MPOmNRHP1Q8POoGl0T8UOkGi0W8XOiGe0Y8\\OgG`0]8@cG=_8DaG8b8I]G5e8L[G1h8OXGMk84UGIm88Q11O1O1O100O1O1O1O00M3M3L4M[gd0"}, "label": "a black bar chair in the background"}]} {"id": 444346, "image": "COCO_train2014_000000444346.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe looking to the left side\"."}], "task": "refering", "answer_template": "The \"giraffe looking to the left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 192, 193, 194, 195, 214, 215, 216, 217, 218, 237, 238, 240, 241, 242, 263, 264, 265, 266, 287, 288, 289, 290, 311, 312, 313, 314, 315, 334, 335, 336, 337, 338, 339, 358, 359, 360, 361, 362, 363, 364, 365, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.312078125, 0.407375, 0.91275, 0.9890208333333333], "iscrowd": 0, "area": 26761.525699999995, "rle": {"size": [480, 640], "counts": "RQn23l>5J5L3M1N2O1O1N101O1N2O1O1N2O1O1N2O1N2O1O0O101O0O101O0O2O000O2O000O20O01O001O001O1O1O001O1O00100O1O001O1O001O1O1O001O1O010H8I7O001O1O001O1O10O01O1`0_O=D001O0O20O101N101N2O0O2O1N101N1O2N2O0O2N2N1O2O1N1O2N101O10O01O10O01O100O100O1O010O1O100O1O100O1O100O1O100O100O00100O1O100O1O100O1O100O1O100O1O010O100O1O100O1O100O00000001N1000000000001O0000000000001O0O10000000001O0000000000001c0\\O6JO1O1O1O2N1O1O1O1O1O1O1O001O001O001O001O001O001O001N101O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O00001O00001N10001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1OXii0"}, "label": "giraffe looking to the left side"}]} {"id": 444346, "image": "COCO_train2014_000000444346.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two giraffes in wilderness crossing their necks\"."}], "task": "refering", "answer_template": "The \"two giraffes in wilderness crossing their necks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 192, 193, 194, 195, 214, 215, 216, 217, 218, 237, 238, 240, 241, 242, 263, 264, 265, 266, 287, 288, 289, 290, 311, 312, 313, 314, 315, 334, 335, 336, 337, 338, 339, 358, 359, 360, 361, 362, 363, 364, 365, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.312078125, 0.407375, 0.91275, 0.9890208333333333], "iscrowd": 0, "area": 26761.525699999995, "rle": {"size": [480, 640], "counts": "RQn23l>5J5L3M1N2O1O1N101O1N2O1O1N2O1O1N2O1N2O1O0O101O0O101O0O2O000O2O000O20O01O001O001O1O1O001O1O00100O1O001O1O001O1O1O001O1O010H8I7O001O1O001O1O10O01O1`0_O=D001O0O20O101N101N2O0O2O1N101N1O2N2O0O2N2N1O2O1N1O2N101O10O01O10O01O100O100O1O010O1O100O1O100O1O100O1O100O100O00100O1O100O1O100O1O100O1O100O1O010O100O1O100O1O100O00000001N1000000000001O0000000000001O0O10000000001O0000000000001c0\\O6JO1O1O1O2N1O1O1O1O1O1O1O001O001O001O001O001O001O001N101O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O00001O00001N10001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1OXii0"}, "label": "two giraffes in wilderness crossing their necks"}]} {"id": 444346, "image": "COCO_train2014_000000444346.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe who ' s head is back of another giraffe\"."}], "task": "refering", "answer_template": "The \"a giraffe who ' s head is back of another giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 150, 151, 152, 153, 173, 174, 175, 176, 177, 196, 197, 198, 219, 220, 242, 243, 262, 263, 264, 284, 285, 286, 287, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332, 333, 350, 351, 352, 353, 354, 355, 356, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.20056250000000003, 0.305625, 0.701125, 0.9865208333333333], "iscrowd": 0, "area": 25720.396399999998, "rle": {"size": [480, 640], "counts": "cml13j>3M3M3M3N200O1O1O1O2N1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O2N1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O2O0O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O100O1O1O1O101N1O1O1O100O001O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O001O1O1O1O1O1O1O1O1O100O100O100O100O100O100O100O010O100O100O11O1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2TNREBo:;[E\\Oh:`0aEYO`:d0iEUOY:g0QFQOQ:k0XFnNj9n0`FjNb9R1gFgN[9U1[2L3M4L4L4L4L4L3M4L4LoK;^3>B=C=J6L4L4L5K4L4L4L2N2N2N2N2N2N2N2N2N1O2SOjLmEX3S:oLdES3[:TM]En2c:h0ON3L4M2N3L3N3M2N3L3N3M2N3L3N3M1N2O1O1O1N101O1O1N2O1O1OO10O01002N4K5L4LESNiCi1Z<\\NbC_1c in this image.", "objects": [{"patches": [27, 28, 29, 48, 49, 50, 51, 52, 53, 54, 55, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260], "bbox": [0.015, 0.1146135831381733, 0.515765625, 0.7573067915690866], "iscrowd": 0, "area": 69149.12244999998, "rle": {"size": [427, 640], "counts": "bZ4g0`<=D:E;E in this image.", "objects": [{"patches": [27, 28, 29, 48, 49, 50, 51, 52, 53, 54, 55, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260], "bbox": [0.015, 0.1146135831381733, 0.515765625, 0.7573067915690866], "iscrowd": 0, "area": 69149.12244999998, "rle": {"size": [427, 640], "counts": "bZ4g0`<=D:E;E in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.46178125000000003, 0.2853864168618267, 1.0, 0.9865105386416863], "iscrowd": 0, "area": 86519.92010000002, "rle": {"size": [427, 640], "counts": "^ek3k0XBa0^Ob0_O in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.46178125000000003, 0.2853864168618267, 1.0, 0.9865105386416863], "iscrowd": 0, "area": 86519.92010000002, "rle": {"size": [427, 640], "counts": "^ek3k0XBa0^Ob0_O in this image.", "objects": [{"patches": [21, 22, 44, 45, 67, 68, 90, 91, 113, 114, 136, 137, 159, 160, 182, 183, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344, 365, 366, 367, 388, 389, 390], "bbox": [0.90203125, 0.0, 1.0, 1.0], "iscrowd": 0, "area": 24310.447950000005, "rle": {"size": [480, 640], "counts": "]m^84d=Y1iNW1POQ1oNP1QOo0POQ1oNP1@`0G9H9G8G9H7H8dLdFm0c9QOiH^N^7_1S3N2M3N2M3N2N2M3N2M3N2M3N2M3N2M3N2M3N2N2M3N2MP]9f4`]Fc4ZLg3YL"}, "label": "the back of a plaid shirt"}]} {"id": 1994, "image": "COCO_train2014_000000001994.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back side of a white and blue striped shirt\"."}], "task": "refering", "answer_template": "The \"the back side of a white and blue striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 44, 45, 67, 68, 90, 91, 113, 114, 136, 137, 159, 160, 182, 183, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344, 365, 366, 367, 388, 389, 390], "bbox": [0.90203125, 0.0, 1.0, 1.0], "iscrowd": 0, "area": 24310.447950000005, "rle": {"size": [480, 640], "counts": "]m^84d=Y1iNW1POQ1oNP1QOo0POQ1oNP1@`0G9H9G8G9H7H8dLdFm0c9QOiH^N^7_1S3N2M3N2M3N2N2M3N2M3N2M3N2M3N2M3N2M3N2N2M3N2MP]9f4`]Fc4ZLg3YL"}, "label": "the back side of a white and blue striped shirt"}]} {"id": 1994, "image": "COCO_train2014_000000001994.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the guy with a camera\"."}], "task": "refering", "answer_template": "The \"the guy with a camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 42, 43, 44, 65, 66, 67, 87, 88, 89, 90, 110, 111, 112, 113, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 179, 180, 181, 182, 201, 202, 203, 204, 224, 225, 226, 227, 247, 248, 249, 250, 270, 271, 272, 273, 293, 294, 295, 296, 316, 317, 318, 319, 340, 341, 363, 364, 365, 386, 387], "bbox": [0.758265625, 0.04660416666666667, 0.94740625, 0.9582291666666667], "iscrowd": 0, "area": 30602.767850000004, "rle": {"size": [480, 640], "counts": "_fS75h>4K4M4K4M4kLZOjGj0S8[OhGi0i5YOaI2`0i0j5^O_I2=e0P6@]I53d0[6@ZI7Jc0h6\\OXI9Cc0Q7YOTI=@d0X7ROQId0[Od0_7lNnHe3n6_LPIc3k6aLUI`3g6cLXI_3g0jJa4j1fJ`3a0mJf4d1iJc3=kJf4c1mJg39hJf4c1QKi34fJi4f1nJi35cJi4[8mJgGo4X9M3L3M3N2M3M3N1N3N2O1O1N101O1N2O1N101O1N100O10000O01TN[L]Hd3b7bLZH^3e7gL^GA2g3`8jL]GA1e3`8lL`G@Nc3b8oL_G@Ma3c8PMaG_OK`3e8RM_GCG[3j8RM`Gc3a8^L^Gb3b8^L_Ga3b8_L]G`3d8`L]G_3c8bL\\G]3e8cL\\G\\3e8dLZGR2@XNW9FZGo1EWNQ9KYGk1MUNl8OXGi13SNe85WGe1:RN`89WGa1=UN\\8;VG_1`0TN\\8=TG]1b0TN[8?SG\\1d0RN[8c0QGZ1e0QN[8e0PGY1g0oM\\8k0gFW1o0lM[8Q1_FU1X1hMZ8W1WFT1[;W11O1O10O01O1OO2M2O1N2N3WOh0UOk0XOj0_O`0_Oa0_Oa0E;G9F:G9G:E:G9G9_MoFiN[95UHTOT8i0S3N3L3M3M3N2M3M3N2M4M2N2MYm?"}, "label": "the guy with a camera"}]} {"id": 1994, "image": "COCO_train2014_000000001994.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with vertical striped shirt\"."}], "task": "refering", "answer_template": "The \"man with vertical striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 42, 43, 44, 65, 66, 67, 87, 88, 89, 90, 110, 111, 112, 113, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 179, 180, 181, 182, 201, 202, 203, 204, 224, 225, 226, 227, 247, 248, 249, 250, 270, 271, 272, 273, 293, 294, 295, 296, 316, 317, 318, 319, 340, 341, 363, 364, 365, 386, 387], "bbox": [0.758265625, 0.04660416666666667, 0.94740625, 0.9582291666666667], "iscrowd": 0, "area": 30602.767850000004, "rle": {"size": [480, 640], "counts": "_fS75h>4K4M4K4M4kLZOjGj0S8[OhGi0i5YOaI2`0i0j5^O_I2=e0P6@]I53d0[6@ZI7Jc0h6\\OXI9Cc0Q7YOTI=@d0X7ROQId0[Od0_7lNnHe3n6_LPIc3k6aLUI`3g6cLXI_3g0jJa4j1fJ`3a0mJf4d1iJc3=kJf4c1mJg39hJf4c1QKi34fJi4f1nJi35cJi4[8mJgGo4X9M3L3M3N2M3M3N1N3N2O1O1N101O1N2O1N101O1N100O10000O01TN[L]Hd3b7bLZH^3e7gL^GA2g3`8jL]GA1e3`8lL`G@Nc3b8oL_G@Ma3c8PMaG_OK`3e8RM_GCG[3j8RM`Gc3a8^L^Gb3b8^L_Ga3b8_L]G`3d8`L]G_3c8bL\\G]3e8cL\\G\\3e8dLZGR2@XNW9FZGo1EWNQ9KYGk1MUNl8OXGi13SNe85WGe1:RN`89WGa1=UN\\8;VG_1`0TN\\8=TG]1b0TN[8?SG\\1d0RN[8c0QGZ1e0QN[8e0PGY1g0oM\\8k0gFW1o0lM[8Q1_FU1X1hMZ8W1WFT1[;W11O1O10O01O1OO2M2O1N2N3WOh0UOk0XOj0_O`0_Oa0_Oa0E;G9F:G9G:E:G9G9_MoFiN[95UHTOT8i0S3N3L3M3M3N2M3M3N2M4M2N2MYm?"}, "label": "man with vertical striped shirt"}]} {"id": 1994, "image": "COCO_train2014_000000001994.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing checked shirt bearing goblet\"."}], "task": "refering", "answer_template": "The \"a man wearing checked shirt bearing goblet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 336, 337, 345, 346, 347, 348, 349, 350, 351, 352, 353, 358, 359, 360, 361, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 380, 381, 382, 383, 384], "bbox": [0.0, 0.0, 0.7130156249999999, 1.0], "iscrowd": 0, "area": 88326.8621, "rle": {"size": [480, 640], "counts": "W8e6V7U1nHR70000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0YJm0aKSOT3Y2lLfM[2U3cMkLX2\\3fMdLT2d3jM\\LQ2k3mMULm1S4QNmKi1[4UNeKf1b4XN^Kb1i4]NWK]1Q5aNoJZ1X5dNhJV1`5hN`JR1h5lNXJo0o5oNQJk0W6SOiIh0]6WOcIc0e6[O[I?m6_OSI`F\\Oh9b0XFUOT:h0lEoN`:n0`EiNk:U1UEbNW;[1iD[Nd;b1\\D\\Ni;a1XD]Nk;a1UD^No;_1QD_NT<^1lCaNX<\\1hCcN[<[1eCcN_<[1aCdNa<[1_CdNb<\\1]CdNe<[1[CcNg<]1YCbNh<^1XCaNj<^1VC`Nl<`1TC_Nn<`1QC`NP=`1PC^NR=b1nB]NT=b1lB]NU=c15O00001O001O00001O001O00001O01O01O001O000gN`BR1a=kNbBT1^=kNcBU1e=O001O001O001O001O001O001O001O001O001O1O1O001O1O1O1O1O0M4K5K5JVbT10n]kN3N2M3M3N1N100O1O010O100O1O10O01O10O001O00001N101O00001N101O00001OO3N3M2N2N3M2N2N2N2N2N2N2N2M3N2N2N2N2N2N3M2N2N2N1O1O1N200O1O100O1O100O100O1O100O100O1O100O1O100001O1O1O1O1O001O1O1O1O1O001O1N2O1O1O1N4M3M3M2M4K5L3L5K3N1M3N3L3M3N3L3N2M4LXje2"}, "label": "a man wearing checked shirt bearing goblet"}]} {"id": 1994, "image": "COCO_train2014_000000001994.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man ' s plaid shirt and hand , the parts not shown behind the wine glass\"."}], "task": "refering", "answer_template": "The \"man ' s plaid shirt and hand , the parts not shown behind the wine glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 336, 337, 345, 346, 347, 348, 349, 350, 351, 352, 353, 358, 359, 360, 361, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 380, 381, 382, 383, 384], "bbox": [0.0, 0.0, 0.7130156249999999, 1.0], "iscrowd": 0, "area": 88326.8621, "rle": {"size": [480, 640], "counts": "W8e6V7U1nHR70000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0YJm0aKSOT3Y2lLfM[2U3cMkLX2\\3fMdLT2d3jM\\LQ2k3mMULm1S4QNmKi1[4UNeKf1b4XN^Kb1i4]NWK]1Q5aNoJZ1X5dNhJV1`5hN`JR1h5lNXJo0o5oNQJk0W6SOiIh0]6WOcIc0e6[O[I?m6_OSI`F\\Oh9b0XFUOT:h0lEoN`:n0`EiNk:U1UEbNW;[1iD[Nd;b1\\D\\Ni;a1XD]Nk;a1UD^No;_1QD_NT<^1lCaNX<\\1hCcN[<[1eCcN_<[1aCdNa<[1_CdNb<\\1]CdNe<[1[CcNg<]1YCbNh<^1XCaNj<^1VC`Nl<`1TC_Nn<`1QC`NP=`1PC^NR=b1nB]NT=b1lB]NU=c15O00001O001O00001O001O00001O01O01O001O000gN`BR1a=kNbBT1^=kNcBU1e=O001O001O001O001O001O001O001O001O001O1O1O001O1O1O1O1O0M4K5K5JVbT10n]kN3N2M3M3N1N100O1O010O100O1O10O01O10O001O00001N101O00001N101O00001OO3N3M2N2N3M2N2N2N2N2N2N2N2M3N2N2N2N2N2N3M2N2N2N1O1O1N200O1O100O1O100O100O1O100O100O1O100O1O100001O1O1O1O1O001O1O1O1O1O001O1N2O1O1O1N4M3M3M2M4K5L3L5K3N1M3N3L3M3N3L3N2M4LXje2"}, "label": "man ' s plaid shirt and hand , the parts not shown behind the wine glass"}]} {"id": 108499, "image": "COCO_train2014_000000108499.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair close to yellow bag\"."}], "task": "refering", "answer_template": "The \"chair close to yellow bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 225, 226, 246, 247, 248, 249, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 318, 336, 337, 338, 339, 340, 341, 359, 360, 361, 364], "bbox": [0.6360625, 0.5790990990990991, 0.8605625, 0.9983108108108109], "iscrowd": 0, "area": 15840.118900000003, "rle": {"size": [444, 640], "counts": "Rj`5g0U=^1aNh0YO00001N1000000O100000PNo12O00000O2O00000O2O00000O101O000O101O000O10001O0O10001O0O1000001N1000001N100N2b?Lk]Og2dN[110000000000O1000000J6D in this image.", "objects": [{"patches": [223, 224, 225, 226, 246, 247, 248, 249, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 318, 336, 337, 338, 339, 340, 341, 359, 360, 361, 364], "bbox": [0.6360625, 0.5790990990990991, 0.8605625, 0.9983108108108109], "iscrowd": 0, "area": 15840.118900000003, "rle": {"size": [444, 640], "counts": "Rj`5g0U=^1aNh0YO00001N1000000O100000PNo12O00000O2O00000O2O00000O101O000O101O000O10001O0O10001O0O1000001N1000001N100N2b?Lk]Og2dN[110000000000O1000000J6D in this image.", "objects": [{"patches": [266, 267, 289, 290, 291, 312, 313, 314, 315, 334, 335, 336, 337, 338, 339, 357, 358, 359, 360, 361, 362, 380, 381, 382, 383, 384, 385, 404, 405, 406, 407, 427, 428, 429, 449, 450, 451, 452, 471, 472, 473, 474, 475, 495, 496, 497, 498, 518, 519, 520, 521], "bbox": [0.51690625, 0.48070312499999995, 0.7737031249999999, 1.0], "iscrowd": 0, "area": 27981.491100000007, "rle": {"size": [640, 640], "counts": "n]_67ec0:E:G9F:F:G:H7J6J6J4\\LYMhEh2V:]MfEd2W:aMfEa2W:cMfE^2W:hMeEY2Y:kMdEV2Y:oMdER2Z:SNbEo1Z:VNcEk1[:YNbEh1[:]NbEd1\\:^NcEd1Y:_NfEb1X:`NgEa1V:bNiE_1U:cNjE^1S:eNlE]1P:fNoE[1n9hNQFY1l9jNSFW1j9lNUFV1g9mNXFT1e9oNYFS1d9POZFR1c9QO[FQ1b9RO]FP1^9TO`Fn0]9UOaFJZNUNi:]2kF^OaNRNf9^3gG`NhNPN`9`3fG`NoNmM[9c3cGaNWOiMU9g3bG`N]OgMQ9i3`G`NDdMl8l3^G`NKbMe8W4VGWN:_M`8i4dFiMP1\\M\\8T8iGiGW8U8nGiGP8V8UHgGk7W8P2N1O2N2N2N1O2N2N2N2N3M3M2N3007I7H8I7I7I7I7H8I7I7I6IO2N3M2N3L4M2N3M3L4M3M3M3M3L4M3M3M3L5L3M2N2N2M3N2N2N3M2M3N2N2N2DkMSBgM]>V2oARM_>m2kA\\Lc>b3o0N2N2N2O1N2N2N2N2N2O1M3N1O2M3N2M3N2M3N2N2M2O2M3N2N1N3N2M3N1O2M3N1N3N2N2M2O2M3N2N1N3N2M3N1O2M3K4K6JUWj2"}, "label": "indian man in button up colored shirt and blue pants walking"}]} {"id": 452565, "image": "COCO_train2014_000000452565.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing a white shirt\"."}], "task": "refering", "answer_template": "The \"man wearing a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [266, 267, 289, 290, 291, 312, 313, 314, 315, 334, 335, 336, 337, 338, 339, 357, 358, 359, 360, 361, 362, 380, 381, 382, 383, 384, 385, 404, 405, 406, 407, 427, 428, 429, 449, 450, 451, 452, 471, 472, 473, 474, 475, 495, 496, 497, 498, 518, 519, 520, 521], "bbox": [0.51690625, 0.48070312499999995, 0.7737031249999999, 1.0], "iscrowd": 0, "area": 27981.491100000007, "rle": {"size": [640, 640], "counts": "n]_67ec0:E:G9F:F:G:H7J6J6J4\\LYMhEh2V:]MfEd2W:aMfEa2W:cMfE^2W:hMeEY2Y:kMdEV2Y:oMdER2Z:SNbEo1Z:VNcEk1[:YNbEh1[:]NbEd1\\:^NcEd1Y:_NfEb1X:`NgEa1V:bNiE_1U:cNjE^1S:eNlE]1P:fNoE[1n9hNQFY1l9jNSFW1j9lNUFV1g9mNXFT1e9oNYFS1d9POZFR1c9QO[FQ1b9RO]FP1^9TO`Fn0]9UOaFJZNUNi:]2kF^OaNRNf9^3gG`NhNPN`9`3fG`NoNmM[9c3cGaNWOiMU9g3bG`N]OgMQ9i3`G`NDdMl8l3^G`NKbMe8W4VGWN:_M`8i4dFiMP1\\M\\8T8iGiGW8U8nGiGP8V8UHgGk7W8P2N1O2N2N2N1O2N2N2N2N3M3M2N3007I7H8I7I7I7I7H8I7I7I6IO2N3M2N3L4M2N3M3L4M3M3M3M3L4M3M3M3L5L3M2N2N2M3N2N2N3M2M3N2N2N2DkMSBgM]>V2oARM_>m2kA\\Lc>b3o0N2N2N2O1N2N2N2N2N2O1M3N1O2M3N2M3N2M3N2N2M2O2M3N2N1N3N2M3N1O2M3N1N3N2N2M2O2M3N2N1N3N2M3N1O2M3K4K6JUWj2"}, "label": "man wearing a white shirt"}]} {"id": 452565, "image": "COCO_train2014_000000452565.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in cultural clothing\"."}], "task": "refering", "answer_template": "The \"man in cultural clothing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 261, 262, 283, 284, 285, 286, 306, 307, 308, 309, 329, 330, 331, 332, 333, 352, 353, 354, 355, 356, 375, 376, 377, 378, 398, 399, 400, 401, 421, 422, 423, 424, 444, 445, 446, 447, 448, 467, 468, 469, 470, 471, 490, 491, 492, 493, 494, 514, 515, 516, 517], "bbox": [0.305625, 0.4494375, 0.52134375, 0.9887656249999999], "iscrowd": 0, "area": 28980.82269999999, "rle": {"size": [640, 640], "counts": "mPk39bc09F:G9bLlNbC^1X in this image.", "objects": [{"patches": [238, 239, 261, 262, 283, 284, 285, 286, 306, 307, 308, 309, 329, 330, 331, 332, 333, 352, 353, 354, 355, 356, 375, 376, 377, 378, 398, 399, 400, 401, 421, 422, 423, 424, 444, 445, 446, 447, 448, 467, 468, 469, 470, 471, 490, 491, 492, 493, 494, 514, 515, 516, 517], "bbox": [0.305625, 0.4494375, 0.52134375, 0.9887656249999999], "iscrowd": 0, "area": 28980.82269999999, "rle": {"size": [640, 640], "counts": "mPk39bc09F:G9bLlNbC^1X in this image.", "objects": [{"patches": [201, 202, 224, 225, 226, 247, 248, 249, 269, 270, 271, 293, 294, 316, 339], "bbox": [0.73253125, 0.4653333333333334, 0.8680468749999999, 0.8426875], "iscrowd": 0, "area": 5819.330849999998, "rle": {"size": [480, 640], "counts": "nTl63i>5L4J6RCBT;c0jD_OS;d0kD^OS;d0lD]OQ;e0nD]Oo:g0nD[OP;h0nDXOP;l0kDXOR;l0hDYOV;j0eDZOX;j0cDZO[;S1VDQOc;h2[O6I7M3N3L3M3MDjETLQ:a3`F^L\\9f3eFYLZ9g3iFWLV9i3lFVLS9j3oFVLo8k3RGTLl8m3VGRLi8n3YGQLf8o3\\GPLc8Q4^GnKa8R4V1000O1000001O01O0VN^EWOc:d0gEVOZ:h0mEROS:o0QFlNP:T1UFgNk9Y1`FYNc9g1j120dNWC?j<_OXC`0i<^OZC`0g<_O[C?f8kAGV>8jAHW>7jAHW>8iAGX>89O100O100O1O100O1O010O1N2OlaW1"}, "label": "a woman wearing shorts and a white sweater pushing a baby carriage"}]} {"id": 100312, "image": "COCO_train2014_000000100312.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a sweater and shorts pushing a baby stroller\"."}], "task": "refering", "answer_template": "The \"a woman wearing a sweater and shorts pushing a baby stroller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 202, 224, 225, 226, 247, 248, 249, 269, 270, 271, 293, 294, 316, 339], "bbox": [0.73253125, 0.4653333333333334, 0.8680468749999999, 0.8426875], "iscrowd": 0, "area": 5819.330849999998, "rle": {"size": [480, 640], "counts": "nTl63i>5L4J6RCBT;c0jD_OS;d0kD^OS;d0lD]OQ;e0nD]Oo:g0nD[OP;h0nDXOP;l0kDXOR;l0hDYOV;j0eDZOX;j0cDZO[;S1VDQOc;h2[O6I7M3N3L3M3MDjETLQ:a3`F^L\\9f3eFYLZ9g3iFWLV9i3lFVLS9j3oFVLo8k3RGTLl8m3VGRLi8n3YGQLf8o3\\GPLc8Q4^GnKa8R4V1000O1000001O01O0VN^EWOc:d0gEVOZ:h0mEROS:o0QFlNP:T1UFgNk9Y1`FYNc9g1j120dNWC?j<_OXC`0i<^OZC`0g<_O[C?f8kAGV>8jAHW>7jAHW>8iAGX>89O100O100O1O100O1O010O1N2OlaW1"}, "label": "a woman wearing a sweater and shorts pushing a baby stroller"}]} {"id": 509914, "image": "COCO_train2014_000000509914.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting on bike\"."}], "task": "refering", "answer_template": "The \"a man sitting on bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 95, 96, 97, 98, 118, 119, 120, 121, 141, 142, 143, 144, 145, 166, 167, 168, 190, 191, 213, 214, 237, 238, 260, 261, 284], "bbox": [0.1424375, 0.24245327102803738, 0.38428125, 0.8608878504672897], "iscrowd": 0, "area": 11136.8223, "rle": {"size": [428, 640], "counts": "XVV11Z=7J5K5K6J1O000000000O1eNWOSFi0m9[OoEe0R:]OkEc0U:AgE`0X:DdE<\\:GaE9^:J`E6`:L^E4b:N[E3d:0ZE0f:2XENg:5WEKi:7UEIk:9SEGl:PECn:`0PE@P;^1O10000O10O01O1O100O1O1O1O100O1O1001N101O001O001O1O001O001O0101000O2O001hMVEj1V;0O2N2N1O2N2N1O2N2N2O1N2N3M2N2N2N2N2N3M2N3N1010O100O01000O10O001N101O1O001N101O0011N2O1O1N2O1N2O1N2O1N2O1N2O0O2N2N2O0O2N2N2O1N1O2N2O1N1O2N2O1N1O2N2L3M4L4L3L5L3M4L4L3M4K5L3M4L4L3L^ZT5"}, "label": "a man sitting on bike"}]} {"id": 509914, "image": "COCO_train2014_000000509914.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man hanging out on a motorcycle wearing a ball cap\"."}], "task": "refering", "answer_template": "The \"a man hanging out on a motorcycle wearing a ball cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 95, 96, 97, 98, 118, 119, 120, 121, 141, 142, 143, 144, 145, 166, 167, 168, 190, 191, 213, 214, 237, 238, 260, 261, 284], "bbox": [0.1424375, 0.24245327102803738, 0.38428125, 0.8608878504672897], "iscrowd": 0, "area": 11136.8223, "rle": {"size": [428, 640], "counts": "XVV11Z=7J5K5K6J1O000000000O1eNWOSFi0m9[OoEe0R:]OkEc0U:AgE`0X:DdE<\\:GaE9^:J`E6`:L^E4b:N[E3d:0ZE0f:2XENg:5WEKi:7UEIk:9SEGl:PECn:`0PE@P;^1O10000O10O01O1O100O1O1O1O100O1O1001N101O001O001O1O001O001O0101000O2O001hMVEj1V;0O2N2N1O2N2N1O2N2N2O1N2N3M2N2N2N2N2N3M2N3N1010O100O01000O10O001N101O1O001N101O0011N2O1O1N2O1N2O1N2O1N2O1N2O0O2N2N2O0O2N2N2O1N1O2N2O1N1O2N2O1N1O2N2L3M4L4L3L5L3M4L4L3M4K5L3M4L4L3L^ZT5"}, "label": "a man hanging out on a motorcycle wearing a ball cap"}]} {"id": 239596, "image": "COCO_train2014_000000239596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue mug in the right hand picture\"."}], "task": "refering", "answer_template": "The \"blue mug in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 203], "bbox": [0.6851406250000001, 0.0440983606557377, 1.0, 0.5418501170960187], "iscrowd": 0, "area": 30361.573949999998, "rle": {"size": [427, 640], "counts": "Ujf51X=4L5K4L5SOl0L5K4K5L3M2M4M3L4M2N3L4M3M2M4M3M2M4M3L4M2N3L4M3M2M3N3cNoK\\IS4c6PLXIS4f6RLUIP4j6SLQIQ4l6TLoHn3P7ULlHm3R7WLiHl3V7XLeHk3Y7XLcHj3[7T1O101O00000O2O00000O2O00000O101O000O101O000O101O0O10001O0O10001O0000001O000000001O0000001O0000001O0000001O0000001O01O0000000O100O10000O100O100O101O0O100O101N10000O101N100O10001N100O100O2O0N2N2O1N3M2K5K5L4N3M2N3M2N3M2O2N1O2N101N1O2O0nN\\G\\Mf8[2bGeM^8Z2cGfM^8X2bGhM_8V2cGjM^8T2cGlM]8R2eGnM\\8P2eGoM\\8P2eGPN\\8m1eGTN[8k1fGTN\\8i1fGWNZ8g1hGYNY8d1iG\\NW8b1kG^NV8_1lGaNU8]1kGdNU8Z1mGfNT8W1nGiNR8U1PHkNQ8R1QHnNP8o0RHQOn7m0SHTOn7i0THWOl7g0VHYOk7d0WH\\Oj7b0VH_Oj7`0WH@j7?VH@k7?UHBl7=THCn7:SHFR85mGLV81jGOU81kG0U80kG0U8OlG1T8OkG2T8OlG1T8NmG2S8NlG3S8MnG3R8MmG4S8L\\HEc7;mHVO[5"}, "label": "blue mug in the right hand picture"}]} {"id": 239596, "image": "COCO_train2014_000000239596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue coffed mug with old taos guesthouse on it\"."}], "task": "refering", "answer_template": "The \"a blue coffed mug with old taos guesthouse on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 203], "bbox": [0.6851406250000001, 0.0440983606557377, 1.0, 0.5418501170960187], "iscrowd": 0, "area": 30361.573949999998, "rle": {"size": [427, 640], "counts": "Ujf51X=4L5K4L5SOl0L5K4K5L3M2M4M3L4M2N3L4M3M2M4M3M2M4M3L4M2N3L4M3M2M3N3cNoK\\IS4c6PLXIS4f6RLUIP4j6SLQIQ4l6TLoHn3P7ULlHm3R7WLiHl3V7XLeHk3Y7XLcHj3[7T1O101O00000O2O00000O2O00000O101O000O101O000O101O0O10001O0O10001O0000001O000000001O0000001O0000001O0000001O0000001O01O0000000O100O10000O100O100O101O0O100O101N10000O101N100O10001N100O100O2O0N2N2O1N3M2K5K5L4N3M2N3M2N3M2O2N1O2N101N1O2O0nN\\G\\Mf8[2bGeM^8Z2cGfM^8X2bGhM_8V2cGjM^8T2cGlM]8R2eGnM\\8P2eGoM\\8P2eGPN\\8m1eGTN[8k1fGTN\\8i1fGWNZ8g1hGYNY8d1iG\\NW8b1kG^NV8_1lGaNU8]1kGdNU8Z1mGfNT8W1nGiNR8U1PHkNQ8R1QHnNP8o0RHQOn7m0SHTOn7i0THWOl7g0VHYOk7d0WH\\Oj7b0VH_Oj7`0WH@j7?VH@k7?UHBl7=THCn7:SHFR85mGLV81jGOU81kG0U80kG0U8OlG1T8OkG2T8OlG1T8NmG2S8NlG3S8MnG3R8MmG4S8L\\HEc7;mHVO[5"}, "label": "a blue coffed mug with old taos guesthouse on it"}]} {"id": 239596, "image": "COCO_train2014_000000239596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"place mat under the food\"."}], "task": "refering", "answer_template": "The \"place mat under the food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 120, 121, 137, 138, 139, 140, 141, 142, 143, 159, 160, 161, 162, 163, 164, 165, 180, 181, 182, 183, 184, 185, 186, 187, 188, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 225, 226, 227, 228, 229, 230, 231, 232, 233, 248, 249, 250, 251, 252, 253, 254, 255, 256, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.006, 0.32466042154566743, 1.0, 0.9988056206088993], "iscrowd": 0, "area": 66309.6877, "rle": {"size": [427, 640], "counts": "_o1S3j5^4fNZ100001O00001O00001O0000001O00001O000010O000001O00000000000000000000000000000000000000000000O1002M3N2N2M3N2N3M2M3N3M3L3N3M3L3N2N1N2O1O1O2N1O100O1O1O1O1O1O0001O00000O100000001O000000ZLgI:X6DmI:S6DPJ;P6CSJW2@kM`0U2^OnMa0R2]OQNb0n1\\OUNe0j1YOYNf0f1YO]Nf0c1XO_Nh0`1WOcNh0\\1WOgNh0X1VOlNi0S1VOoNj0Q1TOROk0m0TOVOk0i0TOYOl0f0SO]Ol0b0SOAl0>ROEn0:QOIn06QOMn01RO1o0MQO5n0JQO9n0FRO;n0DRO?l0_OUOc0j0\\OVOg0h0XOXOj0g0UOYOn0e0QO[OQ1d0mN]OU1b0jN^OY1`0fN@\\1?cNA`1=_NCc1<[NEh19WNFl19SNGo18PNHS27kMIW26hMJ[24eMK]24cMK^25cMI^27bMH`27`MHa28_MGb29^MFd29\\MFe2:\\MDe2<[MCg2WMAk2>VM@k2`0UM_Ol2a0TM_Om2a0n4000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O000000001O00000000001O000000001O0000000O2O0O100O1O10gfm10XYRN8I00O0100O100O10000O100O10000O100O100O1000O0100O100O10000O100O10000O100O100O1000O0100O100O10000O100O10000O100O100O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1M3L4M3M3L4M3M3L4M3L4kLVMfKm2Z4SMcKP3X74M3YMiLVK[3j4eLSK^3m4bLoJb3Q5^LlJe3T5\\LhJg3Y5]L^Jg3b5aLSJb3m5fLgI^3Y6jL\\IY3d6PMoHT3Q7TMdHo2\\7`11OO2O0000000O1000000O100000000O100000000O100000000O100000000O1000000O101O000O100O100O100O100O100O100O100O100O10000O100O100N2N2O1N2N2N2N2N3M200O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O100O1O2O0O10O0100O1O10O0100O10O0100O1O010O1000N2O1O1O1N2O1O8ULUN"}, "label": "place mat under the food"}]} {"id": 239596, "image": "COCO_train2014_000000239596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table mat where plate of snacks are placed\"."}], "task": "refering", "answer_template": "The \"a table mat where plate of snacks are placed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 120, 121, 137, 138, 139, 140, 141, 142, 143, 159, 160, 161, 162, 163, 164, 165, 180, 181, 182, 183, 184, 185, 186, 187, 188, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 225, 226, 227, 228, 229, 230, 231, 232, 233, 248, 249, 250, 251, 252, 253, 254, 255, 256, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.006, 0.32466042154566743, 1.0, 0.9988056206088993], "iscrowd": 0, "area": 66309.6877, "rle": {"size": [427, 640], "counts": "_o1S3j5^4fNZ100001O00001O00001O0000001O00001O000010O000001O00000000000000000000000000000000000000000000O1002M3N2N2M3N2N3M2M3N3M3L3N3M3L3N2N1N2O1O1O2N1O100O1O1O1O1O1O0001O00000O100000001O000000ZLgI:X6DmI:S6DPJ;P6CSJW2@kM`0U2^OnMa0R2]OQNb0n1\\OUNe0j1YOYNf0f1YO]Nf0c1XO_Nh0`1WOcNh0\\1WOgNh0X1VOlNi0S1VOoNj0Q1TOROk0m0TOVOk0i0TOYOl0f0SO]Ol0b0SOAl0>ROEn0:QOIn06QOMn01RO1o0MQO5n0JQO9n0FRO;n0DRO?l0_OUOc0j0\\OVOg0h0XOXOj0g0UOYOn0e0QO[OQ1d0mN]OU1b0jN^OY1`0fN@\\1?cNA`1=_NCc1<[NEh19WNFl19SNGo18PNHS27kMIW26hMJ[24eMK]24cMK^25cMI^27bMH`27`MHa28_MGb29^MFd29\\MFe2:\\MDe2<[MCg2WMAk2>VM@k2`0UM_Ol2a0TM_Om2a0n4000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O000000001O00000000001O000000001O0000000O2O0O100O1O10gfm10XYRN8I00O0100O100O10000O100O10000O100O100O1000O0100O100O10000O100O10000O100O100O1000O0100O100O10000O100O10000O100O100O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1M3L4M3M3L4M3M3L4M3L4kLVMfKm2Z4SMcKP3X74M3YMiLVK[3j4eLSK^3m4bLoJb3Q5^LlJe3T5\\LhJg3Y5]L^Jg3b5aLSJb3m5fLgI^3Y6jL\\IY3d6PMoHT3Q7TMdHo2\\7`11OO2O0000000O1000000O100000000O100000000O100000000O100000000O1000000O101O000O100O100O100O100O100O100O100O100O10000O100O100N2N2O1N2N2N2N2N3M200O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O100O1O2O0O10O0100O1O10O0100O10O0100O1O010O1000N2O1O1O1N2O1O8ULUN"}, "label": "a table mat where plate of snacks are placed"}]} {"id": 239596, "image": "COCO_train2014_000000239596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass of juice on the table\"."}], "task": "refering", "answer_template": "The \"a glass of juice on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 24, 25, 26, 27, 28, 47, 48, 49, 50, 51, 70, 71, 72, 73, 74, 93, 94, 95, 96, 97, 117, 118, 119, 120, 140, 141, 142, 143, 164], "bbox": [0.05248437500000001, 0.03213114754098361, 0.262375, 0.474824355971897], "iscrowd": 0, "area": 18787.7421, "rle": {"size": [427, 640], "counts": "lV>;o<;E5K4K6K4L5K4K6K4L5J5L5K4L5L3M4L3M4L3M4L3L5L3M4L3M4L3M4L3M4L3L5L3M4L3M4L3M4L3M4K2O001O00001O0000001N1000001O00001O000O101O0000O1000000000O010000000000O1000O100000O100000001O001N101O001N101O1N101N101O0O2O1O0O2O0O2O001N3N4L5J5L4L6I8I6I8I7H7I8H7I8H8I6I8H8H7I8I7H7I8H8HmdT6"}, "label": "a glass of juice on the table"}]} {"id": 239596, "image": "COCO_train2014_000000239596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass of orange juice\"."}], "task": "refering", "answer_template": "The \"a glass of orange juice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 24, 25, 26, 27, 28, 47, 48, 49, 50, 51, 70, 71, 72, 73, 74, 93, 94, 95, 96, 97, 117, 118, 119, 120, 140, 141, 142, 143, 164], "bbox": [0.05248437500000001, 0.03213114754098361, 0.262375, 0.474824355971897], "iscrowd": 0, "area": 18787.7421, "rle": {"size": [427, 640], "counts": "lV>;o<;E5K4K6K4L5K4K6K4L5J5L5K4L5L3M4L3M4L3M4L3L5L3M4L3M4L3M4L3M4L3L5L3M4L3M4L3M4L3M4K2O001O00001O0000001N1000001O00001O000O101O0000O1000000000O010000000000O1000O100000O100000001O001N101O001N101O1N101N101O0O2O1O0O2O0O2O001N3N4L5J5L4L6I8I6I8I7H7I8H7I8H8I6I8H8H7I8I7H7I8H8HmdT6"}, "label": "a glass of orange juice"}]} {"id": 239596, "image": "COCO_train2014_000000239596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a placemat is empty behind a placemat that is full\"."}], "task": "refering", "answer_template": "The \"a placemat is empty behind a placemat that is full\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 97, 98, 105, 106, 107, 108, 121, 129, 130, 153], "bbox": [0.22840625, 0.0033723653395784543, 0.8369843750000001, 0.442576112412178], "iscrowd": 0, "area": 33514.09545, "rle": {"size": [427, 640], "counts": "oml12Y=4L5J6K5K4K6K5K5K4K6K5K5K4K6K5K5K4K6K5K4K1000000O1000000O100000N1N3L4L3M4L4M3L3M4O100O00eNhMeGX2[8jMcGU2^8nM_GR2a8PN\\GP2f8RNWGn1i8TNUGk1l8XNPGi1P9ZNmFe1T9]NjFc1V9`NgF_1Z9cNcF^1]9eN`FZ1a9hN]FX1c9W11O100O100O100O100O100O100O100O1O100O100O100O1000000O2O00000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O10001O0O10000O1000000O1000000001O000000001O00000000001O000000001O000000001O01O00000001O000000001O000000001O00000000001O000000001O000000001O00000000010O000000001O00000000001O000000001O000000001O1O001O1O001O001O1O001O1O010O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O4L7I8H7I7I8H2N0010O01O001O00001O001O001O00001O001O001O001O00001O001O001O00oNZH]Lf7a3_H]L`7`3fH]LZ7a3kH]LT7`3RI]Ln6a3WI]Lh6`3^I]Lb6a3dI[L]6b3hI\\LW6a3oI\\LQ6b3TJ\\Lk5a3Q2N2M3N2N2M3N2M3N2M3N2N2M3N2M3N2M3N2M3N2L4lNkDNY;2jDGZ;8jDAZ;>kDZOY;f0f0O100O1O100O1O2O0O100O1O100O1O100O100O1O100O1O100O100O1O100O100O10000000000O100000000O2O0000000O10000000000O10000000^c\\1"}, "label": "a placemat is empty behind a placemat that is full"}]} {"id": 329724, "image": "COCO_train2014_000000329724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza in the opened box has toppings that consist of tomatoes and cheese\"."}], "task": "refering", "answer_template": "The \"the pizza in the opened box has toppings that consist of tomatoes and cheese\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164], "bbox": [0.20882903981264636, 0.1483125, 1.0, 0.462921875], "iscrowd": 0, "area": 55046.62950000001, "rle": {"size": [640, 427], "counts": "Rkg17Sc0h0J6L4M3L3M4L4L4K5L3M4L4L4L4L3M4L3M4L4L3M4L4L3M4M3M2N2N1O2O1N1O2N1O2O1N1O2N2N100O1O1O1O1O100O100O10000O100O100O100O100O100O10000O100O100O100O100O100O10000O100O100O2O000O10000000000O10000000000O10000000000O1000001O000O10000000000000000000000000000000000000001O00000000000000000O1O1O100O1O1O1O1O1O100O100O101N10000O100O100O10000O100O10000O100O100O10000O2O00000000000000000000000000000000000000000000001O00000000000001O00001O00001O00001O0000001O000010O0001O00001O0000001O00001O00001O00001O00001O01O0001O00001O00001O00001O0000001O00001O00001O00010O00001O00001O001O00001O00001O00001O00001O00010O001O1O001O1O001O1O001O1O001O001O1N101N2O0O2O1N101N101N2O0O2L4D;E in this image.", "objects": [{"patches": [53, 54, 55, 56, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164], "bbox": [0.20882903981264636, 0.1483125, 1.0, 0.462921875], "iscrowd": 0, "area": 55046.62950000001, "rle": {"size": [640, 427], "counts": "Rkg17Sc0h0J6L4M3L3M4L4L4K5L3M4L4L4L4L3M4L3M4L4L3M4L4L3M4M3M2N2N1O2O1N1O2N1O2O1N1O2N2N100O1O1O1O1O100O100O10000O100O100O100O100O100O10000O100O100O100O100O100O10000O100O100O2O000O10000000000O10000000000O10000000000O1000001O000O10000000000000000000000000000000000000001O00000000000000000O1O1O100O1O1O1O1O1O100O100O101N10000O100O100O10000O100O10000O100O100O10000O2O00000000000000000000000000000000000000000000001O00000000000001O00001O00001O00001O0000001O000010O0001O00001O0000001O00001O00001O00001O00001O01O0001O00001O00001O00001O0000001O00001O00001O00010O00001O00001O001O00001O00001O00001O00001O00010O001O1O001O1O001O1O001O1O001O001O1N101N2O0O2O1N101N101N2O0O2L4D;E in this image.", "objects": [{"patches": [184, 185, 186, 187, 188, 189, 190, 191, 192, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.01009367681498829, 0.532578125, 1.0, 0.9820156250000001], "iscrowd": 0, "area": 95503.17979999998, "rle": {"size": [640, 427], "counts": "hP35]b0^1bN]1POQ10000O100000O01000000O1000O10O10000O1000O10O1000000O10O10O1000000O10O1000O10000O10O1000O1000000O01000O10000O1O001O100O1O1O1O001O1O100O1O1O001O1O100eC\\KY9e4fF\\KY9e4fF\\KY9d4gF]KX9d4gF]KX9d4gF]KX9d4fF^KZ9b4eF_KZ9b4eF_KZ9b4eF_KZ9a4fF`KY9a4fF`KY9a4eFaKZ9`4eFaK[9_4dFbK[9_4dFbK[9_4dFbK[9^4eFcKZ9^4dFdK[9]4dFdK[9]4dFdK\\9\\4cFeK]9[4bFfK^9Z4aFgK^9Y4aFiK_9W4`FjK`9V4_FkK`9V4_FkKa9U4^FlKb9T4]FmKc9S4[FoKd9Q4\\FPLd9P4[FQLe9o3ZFRLe9o3ZFRLf9n3YFSLg9m3WFULi9k3VFVLi9j3WFWLi9i3VFXLj9h3UFYLj9h3UFYLk9g3SF[Lm9e3RF\\Ln9d3QF]Ln9c3RF^Ln9b3QF_Lo9a3PF`Lo9a3oEaLQ:_3nEbLR:^3mEcLS:]3lEdLS:\\3mEeLS:[3lEfLT:Z3jEhLU:Y3jEhLV:X3iEiLW:W3hEjLW:V3iEkLW:U3hElLX:T3fEnLZ:R3eEoLZ:R3eEoL[:Q3dEPM\\:P3cEQM\\:o2dERM\\:n2bETM^:l2aEUM_:k2`EVM_:k2`EVM`:j2_EWMa:i2^EXMa:h2^EZMb:f2]E[Mc:e2\\E\\Md:d2[E]Md:d2[E]Me:c2ZE^Mf:b2XE`Mg:`2YEaMg:_2XEbMh:^2WEcMi:]2VEdMi:]2VEdMj:\\2TEfMl:Z2SEgMm:X2SEiMm:W2REjMn:V2QEkMo:U2PElMo:U2oDmMQ;S2nDnMR;R2mDoMS;P2mDQNS;^50001O000000000000000O1000000000000000000000000000000O1000000000000000000000000000000000000001O0000000000000000000000001O000000000000000000000000001O0000000000000000000000001O00000000000000000000001O000000000000001O00000000000000001O000000000000001O000000000000001O00000000000000001O000000000000001O000000000000001O00000000000000001O00001O1O1O001O1O1O1N2O1N2N1O2N2N2O1N2N2N1O2N2O1N2N2N2N1O2O1N2N2N2N2N2O0O2N2N2N2N2O1N1O2N2N2N2O1N2N2N1O2N2O1N2N2N2N1O2O1N2N2N2N2N2O0O2N2N2N2N2O1N1O2N2N2N9H>A`0@?A`0@?A`0A>A`0@?Ae^O"}, "label": "pizza pie in a box with one slice missing"}]} {"id": 329724, "image": "COCO_train2014_000000329724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza with a slice missing\"."}], "task": "refering", "answer_template": "The \"pizza with a slice missing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 187, 188, 189, 190, 191, 192, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.01009367681498829, 0.532578125, 1.0, 0.9820156250000001], "iscrowd": 0, "area": 95503.17979999998, "rle": {"size": [640, 427], "counts": "hP35]b0^1bN]1POQ10000O100000O01000000O1000O10O10000O1000O10O1000000O10O10O1000000O10O1000O10000O10O1000O1000000O01000O10000O1O001O100O1O1O1O001O1O100O1O1O001O1O100eC\\KY9e4fF\\KY9e4fF\\KY9d4gF]KX9d4gF]KX9d4gF]KX9d4fF^KZ9b4eF_KZ9b4eF_KZ9b4eF_KZ9a4fF`KY9a4fF`KY9a4eFaKZ9`4eFaK[9_4dFbK[9_4dFbK[9_4dFbK[9^4eFcKZ9^4dFdK[9]4dFdK[9]4dFdK\\9\\4cFeK]9[4bFfK^9Z4aFgK^9Y4aFiK_9W4`FjK`9V4_FkK`9V4_FkKa9U4^FlKb9T4]FmKc9S4[FoKd9Q4\\FPLd9P4[FQLe9o3ZFRLe9o3ZFRLf9n3YFSLg9m3WFULi9k3VFVLi9j3WFWLi9i3VFXLj9h3UFYLj9h3UFYLk9g3SF[Lm9e3RF\\Ln9d3QF]Ln9c3RF^Ln9b3QF_Lo9a3PF`Lo9a3oEaLQ:_3nEbLR:^3mEcLS:]3lEdLS:\\3mEeLS:[3lEfLT:Z3jEhLU:Y3jEhLV:X3iEiLW:W3hEjLW:V3iEkLW:U3hElLX:T3fEnLZ:R3eEoLZ:R3eEoL[:Q3dEPM\\:P3cEQM\\:o2dERM\\:n2bETM^:l2aEUM_:k2`EVM_:k2`EVM`:j2_EWMa:i2^EXMa:h2^EZMb:f2]E[Mc:e2\\E\\Md:d2[E]Md:d2[E]Me:c2ZE^Mf:b2XE`Mg:`2YEaMg:_2XEbMh:^2WEcMi:]2VEdMi:]2VEdMj:\\2TEfMl:Z2SEgMm:X2SEiMm:W2REjMn:V2QEkMo:U2PElMo:U2oDmMQ;S2nDnMR;R2mDoMS;P2mDQNS;^50001O000000000000000O1000000000000000000000000000000O1000000000000000000000000000000000000001O0000000000000000000000001O000000000000000000000000001O0000000000000000000000001O00000000000000000000001O000000000000001O00000000000000001O000000000000001O000000000000001O00000000000000001O000000000000001O000000000000001O00000000000000001O00001O1O1O001O1O1O1N2O1N2N1O2N2N2O1N2N2N1O2N2O1N2N2N2N1O2O1N2N2N2N2N2O0O2N2N2N2N2O1N1O2N2N2N2O1N2N2N1O2N2O1N2N2N2N1O2O1N2N2N2N2N2O0O2N2N2N2N2O1N1O2N2N2N9H>A`0@?A`0@?A`0A>A`0@?Ae^O"}, "label": "pizza with a slice missing"}]} {"id": 83968, "image": "COCO_train2014_000000083968.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"three seat brown couch on the right of room\"."}], "task": "refering", "answer_template": "The \"three seat brown couch on the right of room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 179, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 268, 269, 270, 271], "bbox": [0.6143125, 0.6172136222910217, 0.837890625, 0.9892569659442726], "iscrowd": 0, "area": 12207.62305, "rle": {"size": [323, 640], "counts": "PVl3f0]9c0]O0O107IM2013M1O0O2O001N2O0O2L4N1O2O1N101N1O2O1N101N2N101N2O0O2N2O0O2O0N3L4M2M4000O10000O100O10000O10000O10O1O001O00100O001O1O001O10O01O001O1O000010O000000000000000000001O0000000001O0000000000000000000000000O1000001OO10O0100O100O10O0100O1O100O010`MQIk1P7SNSIl1m6QNVIn1k6PNXIo1h6nM\\IP2d6oM^IQ2b6lMbIR2_6lMcIS2W700O100O101N1E;E;D in this image.", "objects": [{"patches": [176, 177, 179, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 268, 269, 270, 271], "bbox": [0.6143125, 0.6172136222910217, 0.837890625, 0.9892569659442726], "iscrowd": 0, "area": 12207.62305, "rle": {"size": [323, 640], "counts": "PVl3f0]9c0]O0O107IM2013M1O0O2O001N2O0O2L4N1O2O1N101N1O2O1N101N2N101N2O0O2N2O0O2O0N3L4M2M4000O10000O100O10000O10000O10O1O001O00100O001O1O001O10O01O001O1O000010O000000000000000000001O0000000001O0000000000000000000000000O1000001OO10O0100O100O10O0100O1O100O010`MQIk1P7SNSIl1m6QNVIn1k6PNXIo1h6nM\\IP2d6oM^IQ2b6lMbIR2_6lMcIS2W700O100O101N1E;E;D in this image.", "objects": [{"patches": [103, 104, 126, 127, 128, 148, 149, 150, 151, 170, 171, 172, 173, 193, 194, 195, 196, 217], "bbox": [0.41421875, 0.27084309133489465, 0.5892812500000001, 0.6637939110070258], "iscrowd": 0, "area": 8181.979549999997, "rle": {"size": [427, 640], "counts": "\\g^31X=4L4L4L4M3M000O2O001O001O0^D[O_:e0^E_O`:c0^E@^:b0bE@Z:c0fE^OV:d0kE]Od9R1]FoN_9R1cFnNY9U1dFPOY9Q1eFSOZ9k0dFYO\\9f0aF^O_9a0_FC_9<_FHa9Q2000001N10000000O1O1N2A?N2M4M1N2NhNiMaGU2a8QNYGm1h8[NQGb1R9dNhFZ1Y9iNeFX1Z9kNcFV1[9mNcFU1[9mNdFS1[9QOaFP1^9UO^Fl0a9_1O101N100O2N10000001O010O001O0010O01O001jNcF\\N]9b1gF\\NZ9a1kF\\NX9_1lF_NV9^1nF_NU9^1mF`NV9]1lF`NW9]1oF^NS9`1RG[No8c1VGYNl8e1_1N2M3N2N1O1N101O001O0O2O00001O0O2O1O1N2N2N1O2O1N2N2N2I^`]3"}, "label": "a man in a green shirt sitting in the library"}]} {"id": 321539, "image": "COCO_train2014_000000321539.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person wearing hoodie on the computer\"."}], "task": "refering", "answer_template": "The \"person wearing hoodie on the computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 126, 127, 128, 148, 149, 150, 151, 170, 171, 172, 173, 193, 194, 195, 196, 217], "bbox": [0.41421875, 0.27084309133489465, 0.5892812500000001, 0.6637939110070258], "iscrowd": 0, "area": 8181.979549999997, "rle": {"size": [427, 640], "counts": "\\g^31X=4L4L4L4M3M000O2O001O001O0^D[O_:e0^E_O`:c0^E@^:b0bE@Z:c0fE^OV:d0kE]Od9R1]FoN_9R1cFnNY9U1dFPOY9Q1eFSOZ9k0dFYO\\9f0aF^O_9a0_FC_9<_FHa9Q2000001N10000000O1O1N2A?N2M4M1N2NhNiMaGU2a8QNYGm1h8[NQGb1R9dNhFZ1Y9iNeFX1Z9kNcFV1[9mNcFU1[9mNdFS1[9QOaFP1^9UO^Fl0a9_1O101N100O2N10000001O010O001O0010O01O001jNcF\\N]9b1gF\\NZ9a1kF\\NX9_1lF_NV9^1nF_NU9^1mF`NV9]1lF`NW9]1oF^NS9`1RG[No8c1VGYNl8e1_1N2M3N2N1O1N101O001O0O2O00001O0O2O1O1N2N2N1O2O1N2N2N2I^`]3"}, "label": "person wearing hoodie on the computer"}]} {"id": 321539, "image": "COCO_train2014_000000321539.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt , looking overwhelmed , as he gazes at his computer screen\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt , looking overwhelmed , as he gazes at his computer screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 72, 73, 74, 95, 96, 97, 98, 118, 141, 142, 165], "bbox": [0.13403125, 0.1424590163934426, 0.28049999999999997, 0.5537470725995316], "iscrowd": 0, "area": 5765.368099999999, "rle": {"size": [427, 640], "counts": "VPT17j<:]CGj;b0WDFY;b0gD@T;c0lD^Oo:e0RE[Ok:h0VEWOg:k0ZEVOb:m0^ESO_:P1aEPO^:P1dEoN[:R1eEoNY:Q1iEnNV:S1jEmNU:T1lElNR:T1oElNP:U1QFjNn9W1RFiNm9W1TFjNj9T1ZFkNe9Q1`FoN`9P1bFPO\\9Q1dFoN[9Q1gFnNY9R1gFoNW9R1jFmNV9S1jFmNU9T1kFlNT9U1mFkNR9U1nFkNHA^8d1lGjNDF^8`1nGkNBG^8_1QHiN_OK^8]1SHhN]OO^8Y1VHgNZO3^8W1XHgNWO5U8@dGe1Q1eNTO9T8CcG_1U1eNRO=R8]1kHWNQO>Q8^1nHSNoNb0R8[1oHSNmNf0R8X1QIQNkNj0R8V1SImMnNn0n7T1VIlM[Od0]7`1XIlMJ6l6o1[IjM7J]6\\2\\IiMV8W2kGhMU8X2kGhMT8Y2lGfMU8Z2lGeMT8[2lGeMT8[2mGcMT8]2lGcMT8]2^1O00000000001N1M3M3N5M5K5L3L3M2N2N3L2O1O1]OQDHP<4RDLP<2QDNP<0PD1P in this image.", "objects": [{"patches": [50, 51, 72, 73, 74, 95, 96, 97, 98, 118, 141, 142, 165], "bbox": [0.13403125, 0.1424590163934426, 0.28049999999999997, 0.5537470725995316], "iscrowd": 0, "area": 5765.368099999999, "rle": {"size": [427, 640], "counts": "VPT17j<:]CGj;b0WDFY;b0gD@T;c0lD^Oo:e0RE[Ok:h0VEWOg:k0ZEVOb:m0^ESO_:P1aEPO^:P1dEoN[:R1eEoNY:Q1iEnNV:S1jEmNU:T1lElNR:T1oElNP:U1QFjNn9W1RFiNm9W1TFjNj9T1ZFkNe9Q1`FoN`9P1bFPO\\9Q1dFoN[9Q1gFnNY9R1gFoNW9R1jFmNV9S1jFmNU9T1kFlNT9U1mFkNR9U1nFkNHA^8d1lGjNDF^8`1nGkNBG^8_1QHiN_OK^8]1SHhN]OO^8Y1VHgNZO3^8W1XHgNWO5U8@dGe1Q1eNTO9T8CcG_1U1eNRO=R8]1kHWNQO>Q8^1nHSNoNb0R8[1oHSNmNf0R8X1QIQNkNj0R8V1SImMnNn0n7T1VIlM[Od0]7`1XIlMJ6l6o1[IjM7J]6\\2\\IiMV8W2kGhMU8X2kGhMT8Y2lGfMU8Z2lGeMT8[2lGeMT8[2mGcMT8]2lGcMT8]2^1O00000000001N1M3M3N5M5K5L3L3M2N2N3L2O1O1]OQDHP<4RDLP<2QDNP<0PD1P in this image.", "objects": [{"patches": [50, 51, 72, 73, 74, 95, 96, 97, 98, 118, 141, 142, 165], "bbox": [0.13403125, 0.1424590163934426, 0.28049999999999997, 0.5537470725995316], "iscrowd": 0, "area": 5765.368099999999, "rle": {"size": [427, 640], "counts": "VPT17j<:]CGj;b0WDFY;b0gD@T;c0lD^Oo:e0RE[Ok:h0VEWOg:k0ZEVOb:m0^ESO_:P1aEPO^:P1dEoN[:R1eEoNY:Q1iEnNV:S1jEmNU:T1lElNR:T1oElNP:U1QFjNn9W1RFiNm9W1TFjNj9T1ZFkNe9Q1`FoN`9P1bFPO\\9Q1dFoN[9Q1gFnNY9R1gFoNW9R1jFmNV9S1jFmNU9T1kFlNT9U1mFkNR9U1nFkNHA^8d1lGjNDF^8`1nGkNBG^8_1QHiN_OK^8]1SHhN]OO^8Y1VHgNZO3^8W1XHgNWO5U8@dGe1Q1eNTO9T8CcG_1U1eNRO=R8]1kHWNQO>Q8^1nHSNoNb0R8[1oHSNmNf0R8X1QIQNkNj0R8V1SImMnNn0n7T1VIlM[Od0]7`1XIlMJ6l6o1[IjM7J]6\\2\\IiMV8W2kGhMU8X2kGhMT8Y2lGfMU8Z2lGeMT8[2lGeMT8[2mGcMT8]2lGcMT8]2^1O00000000001N1M3M3N5M5K5L3L3M2N2N3L2O1O1]OQDHP<4RDLP<2QDNP<0PD1P in this image.", "objects": [{"patches": [221, 238, 239, 240, 241, 255, 256, 257, 258, 259, 260, 273, 274, 275, 276, 277, 278, 293, 294, 295], "bbox": [0.20172, 0.7422314049586777, 0.47162, 0.9991942148760331], "iscrowd": 0, "area": 9032.98, "rle": {"size": [484, 500], "counts": "RU`12Q?2N1O2M3N2N2N1N3N2N2N2M2O2N2N2M3N2N1O2M3N2N2N1N3N2N2N2M2O2N2N2N2M3N1O2N2M3N2N1O2M3N2N2N1N3N2N2N2M3N1O2N2M3N2N1O2M300O10O100000O01O1O1O1O001O1O1O1O001O1O1O10O01O1O1N1O2N2N2N2N1O2N2N2N2N1N3N2N2N2M2O2N2N2M3N1O2N2M3N1O2N2N2M3N1O2N2M3N2N1O2M3N2N2N1N3N2N2N2M2O2N2N2M3N1O2NRkl3"}, "label": "children ' s book or a picture featuring a large pig and several baby pigs"}]} {"id": 411653, "image": "COCO_train2014_000000411653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pink pig cartoon\"."}], "task": "refering", "answer_template": "The \"a pink pig cartoon\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 238, 239, 240, 241, 255, 256, 257, 258, 259, 260, 273, 274, 275, 276, 277, 278, 293, 294, 295], "bbox": [0.20172, 0.7422314049586777, 0.47162, 0.9991942148760331], "iscrowd": 0, "area": 9032.98, "rle": {"size": [484, 500], "counts": "RU`12Q?2N1O2M3N2N2N1N3N2N2N2M2O2N2N2M3N2N1O2M3N2N2N1N3N2N2N2M2O2N2N2N2M3N1O2N2M3N2N1O2M3N2N2N1N3N2N2N2M3N1O2N2M3N2N1O2M300O10O100000O01O1O1O1O001O1O1O1O001O1O1O10O01O1O1N1O2N2N2N2N1O2N2N2N2N1N3N2N2N2M2O2N2N2M3N1O2N2M3N1O2N2N2M3N1O2N2M3N2N1O2M3N2N2N1N3N2N2N2M2O2N2N2M3N1O2NRkl3"}, "label": "a pink pig cartoon"}]} {"id": 501773, "image": "COCO_train2014_000000501773.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the inside of an empty bathtub\"."}], "task": "refering", "answer_template": "The \"the inside of an empty bathtub\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [311, 312, 313, 314, 315, 333, 334, 335, 336, 337, 338, 339, 355, 356, 357, 358, 359, 360, 361, 362, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.4526875, 0.8039791666666667, 0.758, 1.0], "iscrowd": 0, "area": 14554.1003, "rle": {"size": [480, 640], "counts": "a\\X43k>4M3L3N2M3N2M3N1N3N2M3N1N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O10000O10000O1000000O1000000O10000O1000000O1000000O1000N2O1N2M3M2N3M3N2O1O1O100O1O1O1O100O1O1O1OjCUNY;k1cDYN];g1`D\\N_;e1_D]Na;c1^D^Nb;b1]D_Nc;a1[DaNe;_1ZDbNf;^1YDcNg;]1XDdNh;\\1VDfNj;Z1UDgNk;Y1TDhNl;X1SDiNm;W1RDjNn;V1PDlNP in this image.", "objects": [{"patches": [311, 312, 313, 314, 315, 333, 334, 335, 336, 337, 338, 339, 355, 356, 357, 358, 359, 360, 361, 362, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.4526875, 0.8039791666666667, 0.758, 1.0], "iscrowd": 0, "area": 14554.1003, "rle": {"size": [480, 640], "counts": "a\\X43k>4M3L3N2M3N2M3N1N3N2M3N1N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O10000O10000O1000000O1000000O10000O1000000O1000000O1000N2O1N2M3M2N3M3N2O1O1O100O1O1O1O100O1O1O1OjCUNY;k1cDYN];g1`D\\N_;e1_D]Na;c1^D^Nb;b1]D_Nc;a1[DaNe;_1ZDbNf;^1YDcNg;]1XDdNh;\\1VDfNj;Z1UDgNk;Y1TDhNl;X1SDiNm;W1RDjNn;V1PDlNP in this image.", "objects": [{"patches": [65, 66, 78, 79, 91, 92, 104, 105, 117, 130, 143, 156], "bbox": [0.0, 0.32282, 0.13920000000000002, 0.7238600000000001], "iscrowd": 0, "area": 5599.730700000001, "rle": {"size": [500, 375], "counts": "[6Z1X>=PCWOk9XOYEm1HKg:R3G8H9G9G3NN2O1N2N2K5K5J6L4L5L4K5L4M3M3M3M2M4M3M3VO`D]Mc;a2i0N2N2N2N3N1O1N2O1N4K5L4M3N3kN^B2g=I^B1h=K]BOh=LZB2U>@lA=h>N2M3N2N3Lf_m4"}, "label": "man with his back to the camera in black jacket"}]} {"id": 370727, "image": "COCO_train2014_000000370727.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in red holding a coke and hot dog\"."}], "task": "refering", "answer_template": "The \"a man in red holding a coke and hot dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 29, 30, 31, 32, 42, 43, 44, 45, 46, 55, 56, 57, 58, 59, 69, 70, 71, 72, 82, 83, 84, 85, 93, 94, 95, 96, 97, 98, 99, 100, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 223, 224, 225, 226, 227, 228, 229, 230, 231], "bbox": [0.0, 0.062240000000000004, 0.9947466666666666, 0.98358], "iscrowd": 0, "area": 103034.48060000001, "rle": {"size": [500, 375], "counts": "Qm01c?1N1`NN^C3^<8ZCIbc0oA]OT>?mAAV>O_A1l>CUA in this image.", "objects": [{"patches": [16, 17, 18, 19, 29, 30, 31, 32, 42, 43, 44, 45, 46, 55, 56, 57, 58, 59, 69, 70, 71, 72, 82, 83, 84, 85, 93, 94, 95, 96, 97, 98, 99, 100, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 223, 224, 225, 226, 227, 228, 229, 230, 231], "bbox": [0.0, 0.062240000000000004, 0.9947466666666666, 0.98358], "iscrowd": 0, "area": 103034.48060000001, "rle": {"size": [500, 375], "counts": "Qm01c?1N1`NN^C3^<8ZCIbc0oA]OT>?mAAV>O_A1l>CUA in this image.", "objects": [{"patches": [54, 55, 67, 68, 69, 80, 81, 82, 92, 93, 94, 104, 105, 106, 117, 118, 130, 143, 156, 169], "bbox": [0.0010666666666666667, 0.24547999999999998, 0.3553866666666667, 0.76314], "iscrowd": 0, "area": 6864.8930500000015, "rle": {"size": [500, 375], "counts": "k;2e>m0TOk0J6M2N1O001O001O1O00L4K6M2O2N1M4L3N4L4M4K4N3O1O00100O1O1O001O1O100O1O1O1O1O2O00000000010O00000000001O00000010O000dN@eC`0S in this image.", "objects": [{"patches": [54, 55, 67, 68, 69, 80, 81, 82, 92, 93, 94, 104, 105, 106, 117, 118, 130, 143, 156, 169], "bbox": [0.0010666666666666667, 0.24547999999999998, 0.3553866666666667, 0.76314], "iscrowd": 0, "area": 6864.8930500000015, "rle": {"size": [500, 375], "counts": "k;2e>m0TOk0J6M2N1O001O001O1O00L4K6M2O2N1M4L3N4L4M4K4N3O1O00100O1O1O001O1O100O1O1O1O1O2O00000000010O00000000001O00000010O000dN@eC`0S in this image.", "objects": [{"patches": [138, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.0, 0.34924999999999995, 0.5605, 0.9766458333333332], "iscrowd": 0, "area": 72336.4545, "rle": {"size": [480, 640], "counts": "Y<\\2n7f4eM[22N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O]ObHeI]7[6fHcIY7]6iH`IX7`6jH]IW7b6lH[IU7e6mHXIT7h6b000O10000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000O10O1000000O10000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O100001O1O1O2N1O1O1O2N;E>B5K000000O100000000000000O10000001O2N3M3M2N3M3M2N3M3M2N3M3M2N3M3M2N2N1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O3M3M3M3M3M3M3M2N1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O1O2N1O]XS4"}, "label": "a sofa with pink and white stripes with a white / off - white pillow"}]} {"id": 18473, "image": "COCO_train2014_000000018473.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pink and white striped sofa with soft pastel pillow and a brown dog chair with red floral rug\"."}], "task": "refering", "answer_template": "The \"a pink and white striped sofa with soft pastel pillow and a brown dog chair with red floral rug\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.0, 0.34924999999999995, 0.5605, 0.9766458333333332], "iscrowd": 0, "area": 72336.4545, "rle": {"size": [480, 640], "counts": "Y<\\2n7f4eM[22N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O]ObHeI]7[6fHcIY7]6iH`IX7`6jH]IW7b6lH[IU7e6mHXIT7h6b000O10000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000O10O1000000O10000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O100001O1O1O2N1O1O1O2N;E>B5K000000O100000000000000O10000001O2N3M3M2N3M3M2N3M3M2N3M3M2N3M3M2N2N1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O3M3M3M3M3M3M3M2N1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O1O2N1O]XS4"}, "label": "a pink and white striped sofa with soft pastel pillow and a brown dog chair with red floral rug"}]} {"id": 149556, "image": "COCO_train2014_000000149556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the mouse nearest the large black object\"."}], "task": "refering", "answer_template": "The \"the mouse nearest the large black object\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 28, 29, 30, 42, 43, 44, 45, 46, 47, 48, 49, 50, 60, 61, 62, 63, 64, 65, 66, 67, 68, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 98, 99, 100, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123], "bbox": [0.34044, 0.11685333333333334, 0.8713399999999999, 0.5190933333333333], "iscrowd": 0, "area": 27733.250149999996, "rle": {"size": [375, 500], "counts": "S[n14[;:F:J6M4L3M3M3M3M2N2N2N2N3M2N2N2N2N3M2N2O1N1O2N2O0O2N2N101N2N1O2O0O2O1N101O001O001O1O000O2O000000001O000000001O00000000001O000000001O001O001O001O1O001O1O001O001O1O00001O00001O00001O0000001O000010O0001O0000001O001O01O01O00001O00001O000010O0001O001O00001O00010O00001O00001O0010O0001O00001O01O01O00001O000010O01O00001O0001O01O0000001O01O01O0000001O01O0001O00010O0001O010O001O010O00010O001O010O00010O001O010O1O001O001O001O0O2O1O001O0O2O001O0O2O1O1O1O1N2O2N1N2O1O2M2O1N2O2N1N2O1O2M2O1O1N3N1N2O1O2M2O1O2M2O1O1N3N2M3M3M4L3L4M3L5L3M3J7B=DkYg0"}, "label": "the mouse nearest the large black object"}]} {"id": 149556, "image": "COCO_train2014_000000149556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right most mouse\"."}], "task": "refering", "answer_template": "The \"the right most mouse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 28, 29, 30, 42, 43, 44, 45, 46, 47, 48, 49, 50, 60, 61, 62, 63, 64, 65, 66, 67, 68, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 98, 99, 100, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123], "bbox": [0.34044, 0.11685333333333334, 0.8713399999999999, 0.5190933333333333], "iscrowd": 0, "area": 27733.250149999996, "rle": {"size": [375, 500], "counts": "S[n14[;:F:J6M4L3M3M3M3M2N2N2N2N3M2N2N2N2N3M2N2O1N1O2N2O0O2N2N101N2N1O2O0O2O1N101O001O001O1O000O2O000000001O000000001O00000000001O000000001O001O001O001O1O001O1O001O001O1O00001O00001O00001O0000001O000010O0001O0000001O001O01O01O00001O00001O000010O0001O001O00001O00010O00001O00001O0010O0001O00001O01O01O00001O000010O01O00001O0001O01O0000001O01O01O0000001O01O0001O00010O0001O010O001O010O00010O001O010O00010O001O010O1O001O001O001O0O2O1O001O0O2O001O0O2O1O1O1O1N2O2N1N2O1O2M2O1N2O2N1N2O1O2M2O1O1N3N1N2O1O2M2O1O2M2O1O1N3N2M3M3M4L3L4M3L5L3M3J7B=DkYg0"}, "label": "the right most mouse"}]} {"id": 149556, "image": "COCO_train2014_000000149556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the mouse that is closest\"."}], "task": "refering", "answer_template": "The \"the mouse that is closest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 91, 92, 93, 94, 95, 96, 97, 98, 99, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 172, 173, 185, 186, 187, 188, 189, 190, 191, 206, 207, 208], "bbox": [0.059840000000000004, 0.3617066666666666, 0.66822, 0.8829866666666667], "iscrowd": 0, "area": 38996.51425000001, "rle": {"size": [375, 500], "counts": "[U;;[; in this image.", "objects": [{"patches": [75, 76, 77, 91, 92, 93, 94, 95, 96, 97, 98, 99, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 172, 173, 185, 186, 187, 188, 189, 190, 191, 206, 207, 208], "bbox": [0.059840000000000004, 0.3617066666666666, 0.66822, 0.8829866666666667], "iscrowd": 0, "area": 38996.51425000001, "rle": {"size": [375, 500], "counts": "[U;;[; in this image.", "objects": [{"patches": [282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 377, 378], "bbox": [0.0050625, 0.7078749999999999, 0.47359375000000004, 0.9887708333333334], "iscrowd": 0, "area": 24284.923350000005, "rle": {"size": [480, 640], "counts": "gZ21o>0O1O100O1O1O100mMNWE3i:4oDMQ;;gDEY;`0aDA_;?`DB`;>_DCa;=_DCa;=^DDb;<]DEc;;]DEc;;\\DFd;:[DGf;8YDIg;7YDIg;7XDJh;6WDKi;5WDKi;5VDLj;4UDMk;3TDNm;1SDOm;1RD0n;0QD1o;OQD1o;OPD2P^^ in this image.", "objects": [{"patches": [282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 368, 369, 370, 371, 372, 373, 374, 377, 378], "bbox": [0.0050625, 0.7078749999999999, 0.47359375000000004, 0.9887708333333334], "iscrowd": 0, "area": 24284.923350000005, "rle": {"size": [480, 640], "counts": "gZ21o>0O1O100O1O1O100mMNWE3i:4oDMQ;;gDEY;`0aDA_;?`DB`;>_DCa;=_DCa;=^DDb;<]DEc;;]DEc;;\\DFd;:[DGf;8YDIg;7YDIg;7XDJh;6WDKi;5WDKi;5VDLj;4UDMk;3TDNm;1SDOm;1RD0n;0QD1o;OQD1o;OPD2P^^ in this image.", "objects": [{"patches": [55, 56, 57, 74, 75, 76, 77, 78, 79, 80, 81, 97, 98, 99, 100, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 150, 151, 165, 166, 167, 168, 169, 170, 173, 174, 188, 189, 190, 191, 214], "bbox": [0.170640625, 0.19033333333333333, 0.616859375, 0.7004444444444443], "iscrowd": 0, "area": 26190.354049999998, "rle": {"size": [360, 640], "counts": "faV11T;3M4L3M3M3M4M2N2N2N3M2N2N2N2N3N1N2N2N3M2N2N2N1gNfNdH[1[7UOUHk0k7X1O100O2O0O100O2O0O100O2O0001O2O0O10000O2O000O10001O01O1O1O1O100O1O1O1O2O0O2N2N3N1N3N1O2M2O2M101O0O2O1O0O2O1]IkKP6V4lIXLj5i3QJbLh5_3SJeLm5[400001O0000000000000000000000000O1O1O2N1O1O1O1M3L4K5K5J6K5M3N2O1N2N2O1N2O1N2O1N2N2O1N12O1O1O1O1O1O1O1O0000O1000O00100O010O010000O01000O10000O10O0100O100O1O1O010O1O1O1O100O001N2O1N2N2N1N3M2NO13O1N2O1O1O1O1N2O100O1O100O1O100O1O101N10001O00010O001O01O01O00010O01O01O100O101N1O101N100O1O100O100O12N1O1O2N6J7I3M2N2M3N2N2N00O100000000O10003M3M3L1000000O1QNZH9h7D`H6a7GgH2[7MiHOW71kHLW73jHLW73kHIW76o1O001O001O000O10001O0000000O2O0000001O00T[f2"}, "label": "elephant to the left"}]} {"id": 550972, "image": "COCO_train2014_000000550972.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant on left side\"."}], "task": "refering", "answer_template": "The \"elephant on left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 74, 75, 76, 77, 78, 79, 80, 81, 97, 98, 99, 100, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 150, 151, 165, 166, 167, 168, 169, 170, 173, 174, 188, 189, 190, 191, 214], "bbox": [0.170640625, 0.19033333333333333, 0.616859375, 0.7004444444444443], "iscrowd": 0, "area": 26190.354049999998, "rle": {"size": [360, 640], "counts": "faV11T;3M4L3M3M3M4M2N2N2N3M2N2N2N2N3N1N2N2N3M2N2N2N1gNfNdH[1[7UOUHk0k7X1O100O2O0O100O2O0O100O2O0001O2O0O10000O2O000O10001O01O1O1O1O100O1O1O1O2O0O2N2N3N1N3N1O2M2O2M101O0O2O1O0O2O1]IkKP6V4lIXLj5i3QJbLh5_3SJeLm5[400001O0000000000000000000000000O1O1O2N1O1O1O1M3L4K5K5J6K5M3N2O1N2N2O1N2O1N2O1N2N2O1N12O1O1O1O1O1O1O1O0000O1000O00100O010O010000O01000O10000O10O0100O100O1O1O010O1O1O1O100O001N2O1N2N2N1N3M2NO13O1N2O1O1O1O1N2O100O1O100O1O100O1O101N10001O00010O001O01O01O00010O01O01O100O101N1O101N100O1O100O100O12N1O1O2N6J7I3M2N2M3N2N2N00O100000000O10003M3M3L1000000O1QNZH9h7D`H6a7GgH2[7MiHOW71kHLW73jHLW73kHIW76o1O001O001O000O10001O0000000O2O0000001O00T[f2"}, "label": "elephant on left side"}]} {"id": 75841, "image": "COCO_train2014_000000075841.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy that is serving a ball\"."}], "task": "refering", "answer_template": "The \"the boy that is serving a ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 23, 36, 37, 50, 51, 65, 79, 93, 94, 107, 108, 120, 121, 122, 123, 134, 135, 136, 137, 144, 149, 150, 151, 158, 159, 160, 162, 163, 164, 165, 173, 174, 175, 176, 177, 178, 179, 188, 189, 190, 191, 192, 193, 203, 204, 205, 206, 207, 208, 218, 219, 220, 221, 222, 232, 233, 234, 235, 236, 245, 246, 247, 248, 249, 250, 258, 259, 260, 261, 262, 263, 272, 273, 274, 275, 276, 277, 286, 287, 289, 290, 291, 301, 302, 304, 305, 315, 316, 318, 319], "bbox": [0.3247704081632653, 0.073484375, 0.890841836734694, 0.98734375], "iscrowd": 0, "area": 46911.81665, "rle": {"size": [640, 392], "counts": "ge_22lc0>B4M2M2N3M3M3M21000000O1000O10O1000000O100000O01000000000000O0100000000000O1000O10000000000000000O100000000nAROQ9n0mFVOQ9j0lF\\OQ9d0mFAP9?mFFP9;nFJo86nF0o80oF4o8LoF9n8GoF?n8APGd0m8\\OPGi0n8WOPGn0m8ROPGT1l8mNRGW1l8iNQG]1l8cNRGb1k8^NRGg1l8YNRGl1k8TNSGQ2j8oMSGV2k8jMSG[2j8dMTGb2h8^MWGf2g8YMWGm2UMRLV;P1dGS3nLQL_;k0aG]3_8cL`G^3_8cL_G_3`8bL_G_3`8bL^G`3`8bL_G_3`8bL^G`3a8`L_Ga3`8`L^GS1[M3V;kN^GR1`M0P;PO^GR1eMKl:TO^GQ1jMHg:XO]GQ1PNDb:\\O]GQ1TN@]:A]GP1ZN\\OX:E]GP1^NXOT:I\\GP1dNTOo9M\\Go0iNQOj91[GP1nNlNe96\\Gn0TOhN_9;[Gn0ZOdNZ9`0ZGm0_OaNU9c0ZGm0E]N1nNV7j1SIk0JZNMTOR7h1UIl0OUNJZOn6g1WIj05RNF@j6e1TIP1?hMCFf6c1PIV1k0^M_OLa6c1lH\\1V1RM^O1\\6b1gHd1b1fL[O6Y6a1bHk1m1[LXOB?H7M3ML4L0OO1N2N2N3M2L00O0O2O001N2O0O2O00O0O2O001O1O001O1O001O0mMUIQGl6n8XInFi6Q9[IkFe6U9_IgFb6W9cIeFd6T9`IhFf6R9^IjFh6P9\\IkFl6m8YIoFm6k8XIPGP7h8TITGW7a8mH[G]7Z8hHbG_7W8[2J5K6I7J5K6J6I6K6lNUCgJQ=Q5UCkJP=m4XCmJo in this image.", "objects": [{"patches": [22, 23, 36, 37, 50, 51, 65, 79, 93, 94, 107, 108, 120, 121, 122, 123, 134, 135, 136, 137, 144, 149, 150, 151, 158, 159, 160, 162, 163, 164, 165, 173, 174, 175, 176, 177, 178, 179, 188, 189, 190, 191, 192, 193, 203, 204, 205, 206, 207, 208, 218, 219, 220, 221, 222, 232, 233, 234, 235, 236, 245, 246, 247, 248, 249, 250, 258, 259, 260, 261, 262, 263, 272, 273, 274, 275, 276, 277, 286, 287, 289, 290, 291, 301, 302, 304, 305, 315, 316, 318, 319], "bbox": [0.3247704081632653, 0.073484375, 0.890841836734694, 0.98734375], "iscrowd": 0, "area": 46911.81665, "rle": {"size": [640, 392], "counts": "ge_22lc0>B4M2M2N3M3M3M21000000O1000O10O1000000O100000O01000000000000O0100000000000O1000O10000000000000000O100000000nAROQ9n0mFVOQ9j0lF\\OQ9d0mFAP9?mFFP9;nFJo86nF0o80oF4o8LoF9n8GoF?n8APGd0m8\\OPGi0n8WOPGn0m8ROPGT1l8mNRGW1l8iNQG]1l8cNRGb1k8^NRGg1l8YNRGl1k8TNSGQ2j8oMSGV2k8jMSG[2j8dMTGb2h8^MWGf2g8YMWGm2UMRLV;P1dGS3nLQL_;k0aG]3_8cL`G^3_8cL_G_3`8bL_G_3`8bL^G`3`8bL_G_3`8bL^G`3a8`L_Ga3`8`L^GS1[M3V;kN^GR1`M0P;PO^GR1eMKl:TO^GQ1jMHg:XO]GQ1PNDb:\\O]GQ1TN@]:A]GP1ZN\\OX:E]GP1^NXOT:I\\GP1dNTOo9M\\Go0iNQOj91[GP1nNlNe96\\Gn0TOhN_9;[Gn0ZOdNZ9`0ZGm0_OaNU9c0ZGm0E]N1nNV7j1SIk0JZNMTOR7h1UIl0OUNJZOn6g1WIj05RNF@j6e1TIP1?hMCFf6c1PIV1k0^M_OLa6c1lH\\1V1RM^O1\\6b1gHd1b1fL[O6Y6a1bHk1m1[LXOB?H7M3ML4L0OO1N2N2N3M2L00O0O2O001N2O0O2O00O0O2O001O1O001O1O001O0mMUIQGl6n8XInFi6Q9[IkFe6U9_IgFb6W9cIeFd6T9`IhFf6R9^IjFh6P9\\IkFl6m8YIoFm6k8XIPGP7h8TITGW7a8mH[G]7Z8hHbG_7W8[2J5K6I7J5K6J6I6K6lNUCgJQ=Q5UCkJP=m4XCmJo in this image.", "objects": [{"patches": [90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 340, 341, 342, 344, 363, 364, 367], "bbox": [0.783953125, 0.23751693002257335, 1.0, 0.9858239277652371], "iscrowd": 0, "area": 23016.793850000005, "rle": {"size": [443, 640], "counts": "Uci61h=2O1N2O1N3N1N2O1N2O1N2N101N2O1N2O1N2O1N2O1C=N2N2M3N2O1O1N101G9N2O1O1O1O1O1O1O100O0000000O010O10000O100fIdMU1]2jNgMS1X2nNkMo0S2SOPNi0P2WOTNf0e1A]NVOEZ1`1i3]NeKa0UOGY1Z1n3`NaKb0TOJX1U1T4_N_Kd0ROMW1o0Y4bN[KT27[O_4aNYKX2K@m4YNUK\\2I[OS5ZNRK`2FWOY5ZN\\JTO9d3MoN_5YNYJ\\32\\Ne5\\5100O001O10O01O001OO2N1O1N2O1O2N1O11O100O001O1O100O001O1O11O3L3N3M3M3M4K5L4L4L3M1N2O1O2N1O1O1O1O1O1O001OlHiKe4V4ZKoK8Kd1U4RNUL5Lf1n3TNYL20f1f3VN]L13e1`3YN`LN6f1Y3ZNeLL4k1W3WNgLK5n1T3UNjLI5R2Q3RNmLI5U2o2oMoLH5EkNX1Q4XOSMG4AQO^1i3XOdMVOENUOc11"}, "label": "the little boy wearing red and white swinging a bat"}]} {"id": 305219, "image": "COCO_train2014_000000305219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little leaguer wearing his uniform , swinging a baseball bat\"."}], "task": "refering", "answer_template": "The \"a little leaguer wearing his uniform , swinging a baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 340, 341, 342, 344, 363, 364, 367], "bbox": [0.783953125, 0.23751693002257335, 1.0, 0.9858239277652371], "iscrowd": 0, "area": 23016.793850000005, "rle": {"size": [443, 640], "counts": "Uci61h=2O1N2O1N3N1N2O1N2O1N2N101N2O1N2O1N2O1N2O1C=N2N2M3N2O1O1N101G9N2O1O1O1O1O1O1O100O0000000O010O10000O100fIdMU1]2jNgMS1X2nNkMo0S2SOPNi0P2WOTNf0e1A]NVOEZ1`1i3]NeKa0UOGY1Z1n3`NaKb0TOJX1U1T4_N_Kd0ROMW1o0Y4bN[KT27[O_4aNYKX2K@m4YNUK\\2I[OS5ZNRK`2FWOY5ZN\\JTO9d3MoN_5YNYJ\\32\\Ne5\\5100O001O10O01O001OO2N1O1N2O1O2N1O11O100O001O1O100O001O1O11O3L3N3M3M3M4K5L4L4L3M1N2O1O2N1O1O1O1O1O1O001OlHiKe4V4ZKoK8Kd1U4RNUL5Lf1n3TNYL20f1f3VN]L13e1`3YN`LN6f1Y3ZNeLL4k1W3WNgLK5n1T3UNjLI5R2Q3RNmLI5U2o2oMoLH5EkNX1Q4XOSMG4AQO^1i3XOdMVOENUOc11"}, "label": "a little leaguer wearing his uniform , swinging a baseball bat"}]} {"id": 305219, "image": "COCO_train2014_000000305219.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the young boy spectator holds a blue umbrella while watching a softball game\"."}], "task": "refering", "answer_template": "The \"the young boy spectator holds a blue umbrella while watching a softball game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 95, 96, 117, 118, 119, 120, 140, 141, 142, 143, 163, 164, 165, 166, 186, 187, 188, 189, 209, 210, 211, 212, 232, 233, 234, 235, 255, 256, 257, 258, 279, 280, 281, 302, 303, 304, 324, 325, 326, 327], "bbox": [0.09959375000000001, 0.1899097065462754, 0.26209375, 0.914785553047404], "iscrowd": 0, "area": 19422.067299999988, "rle": {"size": [443, 640], "counts": "fkk06c=8H4K6PIYOd0k0ZOWOb0m0]LSOH2h3P1VLUO1Me3S1RLWO8Fd3W1lKYO?Ac3Z1eK\\Oh0ZOa3]1`K@m0TOa3_1ZKDT1nN_3b1VKF[1hN]3d1RKK_1bN]3f1lJOg1[N[3i1eJ5n1SN\\3j1[J=X2jM]3d4cL\\K\\3f4ULhKk3X4[KbLd4_3bJZM_5a5O100O1OiIaJT4_5`KlJa4S5ZKSKf4m4eJ\\JLg0`5l4^JdJ1`0b5l4VJkJ7:c5j4PJSK<3e5d5ZJ\\Jg5d5YJfIL8i5m5SJhI728OX5X6_JgI02R6X6UJeIH2n5Z6`JeI@1j5]5QJ[Kl0VOXO1f5_5`JRKg0]OSO1`5a5PKjJ`0J[4]5_KaJ;0Q4e5UMXJf2n5[MQJ_2V6\\27I7I6J7I7I6K6J66J6J6J6J6J6J7J6I7J6J6JTI[Jg5_5WJeJm5U5QJmJU6m4iIUK]6e4bI[Ke6_4XIdKn6W4mHlKZ7o3bHTL^7m3]HWLc7j3XHYLh7i3THZLl7g3oG]LP8g410O010O10OWM_JXMb5f2gJSMX5m2PKkLQ5T3YKcLg4\\3jKSLV4m3ZLcKg3\\4jLTKV3k4m201O2N3M3M3L4M3]LfFU1 in this image.", "objects": [{"patches": [72, 73, 95, 96, 117, 118, 119, 120, 140, 141, 142, 143, 163, 164, 165, 166, 186, 187, 188, 189, 209, 210, 211, 212, 232, 233, 234, 235, 255, 256, 257, 258, 279, 280, 281, 302, 303, 304, 324, 325, 326, 327], "bbox": [0.09959375000000001, 0.1899097065462754, 0.26209375, 0.914785553047404], "iscrowd": 0, "area": 19422.067299999988, "rle": {"size": [443, 640], "counts": "fkk06c=8H4K6PIYOd0k0ZOWOb0m0]LSOH2h3P1VLUO1Me3S1RLWO8Fd3W1lKYO?Ac3Z1eK\\Oh0ZOa3]1`K@m0TOa3_1ZKDT1nN_3b1VKF[1hN]3d1RKK_1bN]3f1lJOg1[N[3i1eJ5n1SN\\3j1[J=X2jM]3d4cL\\K\\3f4ULhKk3X4[KbLd4_3bJZM_5a5O100O1OiIaJT4_5`KlJa4S5ZKSKf4m4eJ\\JLg0`5l4^JdJ1`0b5l4VJkJ7:c5j4PJSK<3e5d5ZJ\\Jg5d5YJfIL8i5m5SJhI728OX5X6_JgI02R6X6UJeIH2n5Z6`JeI@1j5]5QJ[Kl0VOXO1f5_5`JRKg0]OSO1`5a5PKjJ`0J[4]5_KaJ;0Q4e5UMXJf2n5[MQJ_2V6\\27I7I6J7I7I6K6J66J6J6J6J6J6J7J6I7J6J6JTI[Jg5_5WJeJm5U5QJmJU6m4iIUK]6e4bI[Ke6_4XIdKn6W4mHlKZ7o3bHTL^7m3]HWLc7j3XHYLh7i3THZLl7g3oG]LP8g410O010O10OWM_JXMb5f2gJSMX5m2PKkLQ5T3YKcLg4\\3jKSLV4m3ZLcKg3\\4jLTKV3k4m201O2N3M3M3L4M3]LfFU1 in this image.", "objects": [{"patches": [37, 38, 39, 40, 41, 42, 59, 60, 61, 62, 63, 64, 65, 82, 83, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226], "bbox": [0.5945625, 0.08445833333333333, 0.868515625, 0.5404166666666668], "iscrowd": 0, "area": 35870.74555, "rle": {"size": [480, 640], "counts": "Vib5V1P=o0QOk0UOk0UOk0UOk0G9N1O2N2N2O0O2N1O1O2N101N1O2N1O2N1O101N1O2N1O2O001O001O00000000000000000001O000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O0000000000000000001O1O2N1N3L3N2N3M2N2M4M2N2N3M2M3N3M2N3L3N2WL[]W1"}, "label": "a sweet foo in an orange bowl"}]} {"id": 510027, "image": "COCO_train2014_000000510027.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the snack in the orange container\"."}], "task": "refering", "answer_template": "The \"the snack in the orange container\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 40, 41, 42, 59, 60, 61, 62, 63, 64, 65, 82, 83, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226], "bbox": [0.5945625, 0.08445833333333333, 0.868515625, 0.5404166666666668], "iscrowd": 0, "area": 35870.74555, "rle": {"size": [480, 640], "counts": "Vib5V1P=o0QOk0UOk0UOk0UOk0G9N1O2N2N2O0O2N1O1O2N101N1O2N1O2N1O101N1O2N1O2O001O001O00000000000000000001O000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O0000000000000000001O1O2N1N3L3N2N3M2N2M4M2N2N3M2M3N3M2N3L3N2WL[]W1"}, "label": "the snack in the orange container"}]} {"id": 510027, "image": "COCO_train2014_000000510027.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green bowl of red apple slices\"."}], "task": "refering", "answer_template": "The \"a green bowl of red apple slices\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 295, 312, 313, 314, 315, 316, 317, 318, 335, 336, 337, 338, 339, 340, 341, 358, 359, 360, 361, 362, 363, 364, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.5843906249999999, 0.5377916666666667, 0.863484375, 0.9933333333333333], "iscrowd": 0, "area": 35859.18915, "rle": {"size": [480, 640], "counts": "^g_5 in this image.", "objects": [{"patches": [220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 295, 312, 313, 314, 315, 316, 317, 318, 335, 336, 337, 338, 339, 340, 341, 358, 359, 360, 361, 362, 363, 364, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.5843906249999999, 0.5377916666666667, 0.863484375, 0.9933333333333333], "iscrowd": 0, "area": 35859.18915, "rle": {"size": [480, 640], "counts": "^g_5 in this image.", "objects": [{"patches": [266, 267, 268, 289, 290, 291, 292, 293, 294, 313, 314, 315, 316, 317, 318, 337, 338, 339, 340, 362], "bbox": [0.6010625, 0.6725833333333333, 0.8348125, 0.8925833333333332], "iscrowd": 0, "area": 8891.6608, "rle": {"size": [480, 640], "counts": "Sid5h0X>2N101N2N101N101N2N101N1O2O1N1O2O0O2N2O0O2N100O1O100O1O100O00100O1O100O1O010O1O1O100O1O010O1O100O1O100O00100O1O100O1O00100O1O100O1O010O1O010O0001O00000010O000001O0000010O0000001O01O0001O00001O01O0001O00000010O000001O00000O2O0O100O2N100O2O0O101N1O2O0O101N101N100O2N101N1O1N3M2O2M2O1N3M2FRBAP>;TBBn=;VBCk=;?J5LbZa1"}, "label": "a large apple slice in a green container between two other apple slices"}]} {"id": 510027, "image": "COCO_train2014_000000510027.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an apple which cutted into piecies\"."}], "task": "refering", "answer_template": "The \"an apple which cutted into piecies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [266, 267, 268, 289, 290, 291, 292, 293, 294, 313, 314, 315, 316, 317, 318, 337, 338, 339, 340, 362], "bbox": [0.6010625, 0.6725833333333333, 0.8348125, 0.8925833333333332], "iscrowd": 0, "area": 8891.6608, "rle": {"size": [480, 640], "counts": "Sid5h0X>2N101N2N101N101N2N101N1O2O1N1O2O0O2N2O0O2N100O1O100O1O100O00100O1O100O1O010O1O1O100O1O010O1O100O1O100O00100O1O100O1O00100O1O100O1O010O1O010O0001O00000010O000001O0000010O0000001O01O0001O00001O01O0001O00000010O000001O00000O2O0O100O2N100O2O0O101N1O2O0O101N101N100O2N101N1O1N3M2O2M2O1N3M2FRBAP>;TBBn=;VBCk=;?J5LbZa1"}, "label": "an apple which cutted into piecies"}]} {"id": 510027, "image": "COCO_train2014_000000510027.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cake cut into a triange in the orange colored container block\"."}], "task": "refering", "answer_template": "The \"cake cut into a triange in the orange colored container block\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 84, 85, 86, 87, 107, 108, 109, 110, 130, 131, 132, 133, 152, 153, 154, 155, 175, 176, 177, 198, 199], "bbox": [0.621484375, 0.1586875, 0.823453125, 0.514625], "iscrowd": 0, "area": 10684.841750000003, "rle": {"size": [480, 640], "counts": "Tij58_>NcA4U>1kA3o=0QB2l=c0M4L3M2N3M3L3L4M3M3M3N3M2N2M3N2M3L3M2N3N2N2M3N2N2N1N3N2M3M3M3N1O2N2O0O2N2K5K5L4N2001O00001O2N100O1O2N1N2O1N3N1O1N2O2N1O1O1O2N100O1O2N2M2O2N2M2O2M2O2N101N1O1O2O0O2N1O1O1O1O1O1O1N2O1O2N1O1O2N1O101N1O1O2N101N1O2N2N3N2M3M4L3M3M2M3N2M3N2M3N2M3N3L3N3L4L3NSkd1"}, "label": "cake cut into a triange in the orange colored container block"}]} {"id": 510027, "image": "COCO_train2014_000000510027.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the triangle shaped treat in the orange box\"."}], "task": "refering", "answer_template": "The \"the triangle shaped treat in the orange box\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 84, 85, 86, 87, 107, 108, 109, 110, 130, 131, 132, 133, 152, 153, 154, 155, 175, 176, 177, 198, 199], "bbox": [0.621484375, 0.1586875, 0.823453125, 0.514625], "iscrowd": 0, "area": 10684.841750000003, "rle": {"size": [480, 640], "counts": "Tij58_>NcA4U>1kA3o=0QB2l=c0M4L3M2N3M3L3L4M3M3M3N3M2N2M3N2M3L3M2N3N2N2M3N2N2N1N3N2M3M3M3N1O2N2O0O2N2K5K5L4N2001O00001O2N100O1O2N1N2O1N3N1O1N2O2N1O1O1O2N100O1O2N2M2O2N2M2O2M2O2N101N1O1O2O0O2N1O1O1O1O1O1O1N2O1O2N1O1O2N1O101N1O1O2N101N1O2N2N3N2M3M4L3M3M2M3N2M3N2M3N2M3N3L3N3L4L3NSkd1"}, "label": "the triangle shaped treat in the orange box"}]} {"id": 510027, "image": "COCO_train2014_000000510027.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of an egg salad sandwich with the other half below it\"."}], "task": "refering", "answer_template": "The \"half of an egg salad sandwich with the other half below it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82], "bbox": [0.10618749999999999, 0.103375, 0.586515625, 0.22472916666666667], "iscrowd": 0, "area": 15111.832, "rle": {"size": [480, 640], "counts": "fmo01n>3M3M3M3N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N0000000000001O0000000000000000000000001O000000000000000000000000001O0000000000000000000000001O0000000000000000000000001O000000000000000000000000001O0000000000000000000000001O0000000000000001O00000001O000000000000000000000000001O0O10000000000000000000001O0000000000000000000000001O0000000000000000O1000000000000000000000000000000O10000000000000000000001O000000000O1000000000000000000000001O00000O10000000000000000000000000001O0O100000000000000000000000000000000O2O00000000000000000000000O10000000001O000000000000000O100M3N2M2O20000000000000O100000000000000O100000000005La0^OeTl3"}, "label": "half of an egg salad sandwich with the other half below it"}]} {"id": 510027, "image": "COCO_train2014_000000510027.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the very top piece of bread on the top slice of sandwich\"."}], "task": "refering", "answer_template": "The \"the very top piece of bread on the top slice of sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82], "bbox": [0.10618749999999999, 0.103375, 0.586515625, 0.22472916666666667], "iscrowd": 0, "area": 15111.832, "rle": {"size": [480, 640], "counts": "fmo01n>3M3M3M3N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N0000000000001O0000000000000000000000001O000000000000000000000000001O0000000000000000000000001O0000000000000000000000001O000000000000000000000000001O0000000000000000000000001O0000000000000001O00000001O000000000000000000000000001O0O10000000000000000000001O0000000000000000000000001O0000000000000000O1000000000000000000000000000000O10000000000000000000001O000000000O1000000000000000000000001O00000O10000000000000000000000000001O0O100000000000000000000000000000000O2O00000000000000000000000O10000000001O000000000000000O100M3N2M2O20000000000000O100000000000000O100000000005La0^OeTl3"}, "label": "the very top piece of bread on the top slice of sandwich"}]} {"id": 510027, "image": "COCO_train2014_000000510027.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the container with carrots\"."}], "task": "refering", "answer_template": "The \"the container with carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.062171875, 0.5207291666666667, 0.5751250000000001, 0.9689166666666666], "iscrowd": 0, "area": 65819.53994999999, "rle": {"size": [480, 640], "counts": "]Rc0S3f9X2O1N3N1O1N2O1O1N2O1O1N2O1O2M2O1N2O1O1N2O1O1N2O1O2M100O10000O10000001O000000000000001O01O00000000000001O000000000000001O000000000000001O00000000000000001O000000000000001O000000000000001O000000000000000010O00000000000001O000000000000001O00000000000000001O000000000000001O000000000000001O00000000000000001O000000000000001O01O00000000000000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000001N100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000001O00000O1000000000000O100000000000000O10000001O1O1O1O1O1O1N2N2M4M2M3M3N2M3N2M3N2M3N3L3N2M3N2M3N2Ml2TMbTo3"}, "label": "the container with carrots"}]} {"id": 510027, "image": "COCO_train2014_000000510027.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a portion of food with carrots and cheese slices\"."}], "task": "refering", "answer_template": "The \"a portion of food with carrots and cheese slices\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.062171875, 0.5207291666666667, 0.5751250000000001, 0.9689166666666666], "iscrowd": 0, "area": 65819.53994999999, "rle": {"size": [480, 640], "counts": "]Rc0S3f9X2O1N3N1O1N2O1O1N2O1O1N2O1O2M2O1N2O1O1N2O1O1N2O1O2M100O10000O10000001O000000000000001O01O00000000000001O000000000000001O000000000000001O00000000000000001O000000000000001O000000000000001O000000000000000010O00000000000001O000000000000001O00000000000000001O000000000000001O000000000000001O00000000000000001O000000000000001O01O00000000000000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000001N100000000000000O100000000000000O1000000000000O100000000000000O100000000000000O1000000000000O100000000000000O100000001O00000O1000000000000O100000000000000O10000001O1O1O1O1O1O1N2N2M4M2M3M3N2M3N2M3N2M3N3L3N2M3N2M3N2Ml2TMbTo3"}, "label": "a portion of food with carrots and cheese slices"}]} {"id": 510027, "image": "COCO_train2014_000000510027.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the piece of bread second from the top\"."}], "task": "refering", "answer_template": "The \"the piece of bread second from the top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127], "bbox": [0.079953125, 0.1903125, 0.6034375, 0.3315624999999999], "iscrowd": 0, "area": 14547.31235, "rle": {"size": [480, 640], "counts": "kog04l>2N2N2N2N2N2N2N1O1O001O100O1O9G3M2N2N2N1O2O0O1O2OO000010O0001O0001O01O001O010O000000000001O000000000000001O000001O00000001O000000000000001O000000000000000001O0000000000000000000000000000000000000000010O00000000000000000000000000000000001O00000000000000001O000000000000010O00000000000000001O00000000000000001O000001O00000001O0000000000000000000000000000000001O0000000O10000000000000000000000000000000001O00000O10000000000000000000000000000O1000001O00000000000000000000000O100000000000000000000000000000000O1000000000001O0000000000000000000000000000000O100000000000000000000000001O0000001O00001O00001O0000001O0000N2O2N1O1N3L3L4L5K4M4M2N2100O010O010OO2N1O2N1O2M3M3N2M3M3N\\nf3"}, "label": "the piece of bread second from the top"}]} {"id": 510027, "image": "COCO_train2014_000000510027.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the half part of the sandwich nearest to the carrots\"."}], "task": "refering", "answer_template": "The \"the half part of the sandwich nearest to the carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216], "bbox": [0.06390625, 0.3004583333333333, 0.5885624999999999, 0.5426041666666667], "iscrowd": 0, "area": 32365.289350000014, "rle": {"size": [480, 640], "counts": "m\\c01n>101N101N1O2O0O2N101YBFe< in this image.", "objects": [{"patches": [117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216], "bbox": [0.06390625, 0.3004583333333333, 0.5885624999999999, 0.5426041666666667], "iscrowd": 0, "area": 32365.289350000014, "rle": {"size": [480, 640], "counts": "m\\c01n>101N101N1O2O0O2N101YBFe< in this image.", "objects": [{"patches": [159, 160, 173, 174, 187, 188, 200, 201, 202, 214, 215, 216, 228, 229, 230, 242, 243, 244, 256, 257, 269, 270, 271, 283, 284, 285], "bbox": [0.2616710182767624, 0.48475, 0.5033420365535248, 0.910984375], "iscrowd": 0, "area": 12019.37535, "rle": {"size": [640, 383], "counts": "bQo12ic05J7J5J7J5K6I6K6J5J7J5K6I6K6I6K6J5J7J5K6I6K6I6K6J5J7K4M4K4M4K4M4K4M4L3L5L3L5L2M2O0O2O001O001O1O001O001O00N^O]InCa6W in this image.", "objects": [{"patches": [159, 160, 173, 174, 187, 188, 200, 201, 202, 214, 215, 216, 228, 229, 230, 242, 243, 244, 256, 257, 269, 270, 271, 283, 284, 285], "bbox": [0.2616710182767624, 0.48475, 0.5033420365535248, 0.910984375], "iscrowd": 0, "area": 12019.37535, "rle": {"size": [640, 383], "counts": "bQo12ic05J7J5J7J5K6I6K6J5J7J5K6I6K6I6K6J5J7J5K6I6K6I6K6J5J7K4M4K4M4K4M4K4M4L3L5L3L5L2M2O0O2O001O001O1O001O001O00N^O]InCa6W in this image.", "objects": [{"patches": [106, 107, 108, 109, 120, 121, 122, 123, 134, 135, 136, 137, 148, 149, 150, 151, 161, 162, 163, 164, 165, 166, 174, 175, 176, 177, 178, 179, 180, 188, 189, 190, 191, 192, 193, 194, 195, 201, 202, 203, 204, 205, 206, 207, 208, 209, 218, 219, 220, 221, 222, 232, 233, 234, 235, 236, 245, 246, 247, 248, 249, 250, 251, 259, 260, 261, 262, 263, 264, 265, 273, 274, 275, 276, 277, 278, 279, 287, 288, 289, 290, 291, 292, 293, 301, 302, 303, 304, 305, 306, 307, 316, 317, 318, 319, 320, 321], "bbox": [0.38302872062663185, 0.303375, 0.9838381201044386, 0.968546875], "iscrowd": 0, "area": 53836.19875, "rle": {"size": [640, 383], "counts": "mXl21nc01N2O1N2O1N2O1N2O1N101N2O1N2O1N2O1N2O1N2O1N2O1N2O0O2O1N2O1N2O1N2O1N2O1N2O1N2O1N101N2O1N2O1O1N2O1N2O1O1N2O1O1O1O001O1O1O1N2O1O1O1O1O1cDiMP5X2UJcNj5^1[I]Oc6e0cH5\\7LiGo0V8QOQGi1n8XNlFn1S9SNjFP2U9QNhFR2W9oMeFV2Y9kMdFX2Z9jMcFY2\\9hM`F\\2_9eM^F^2Q7]JdJV3XN`2T7_J^JR3[Nb2V7bJYJm2]Ne2Y7dJTJh2`Ng2[7gJoIc2cNi2]7iJjI`2fNj2_7lJeIZ2iNn2a7nJ`Id8_6bG[I_8d6gGVIZ8i6lGQIU8n6QHlHP8S7UHhHl7X7XHbHj7]7VHcHk7\\7UHdHl7[7]2O1O1O1O1O1O1O101N1O1O1O1O1O1O1O1O1O100000000000000000000000000000000000001O00000000000000PFYIg6g6VI\\Ij6d6TI_Ik6a6RIbIn6^6PIdIP7\\6mHgIS7Y6kHiIU7W6hHlIX7T6fHoIY7Q6dHRJm1WOP1g6PMUJP2UOo0f6oLWJQ2TOP1e6lLZJT2ROo0d6kL\\JU2ROo0b6iL`JW2nNP1c6fLaJY2nNP1j6[L[Je2lNo0Q7RLUJn2lNo0X7gKoIZ3iNo0a7\\KhIe3hNn0h7RKcIo3gNn0g9QOYFP1f9oN\\FP1d9PO]Fo0c9PO_Fo0a9QO_Fo0a9POaFo0`9POaFo0_9PObFP1^9POcFo0]9POeFo0[9QOfFn0Z9QOgFo0Y9QOhFn0Y9POiFo0W9QOiFo0W9POkFo0U9QOlFn0T9QOnFn0R9ROnFn0S9POoFo0Q9QOPGn0P9QOQGo0o8QORGn0n8ROSGm0m8ROTGn0o8oNRGP1P9mNRGR1P9lNQGS1R9iNoFW1S9gNnFY1T9cNnF\\1T9bNlF^1W9^NkFa1W9]NjFb1Y9ZNiFe1Y9YNgFg1\\9UNfFj1\\9TNeFk1]9RNdFn1_9oMbFP2`9mMbFR2a9kM`FT2b9iM_FW2d9fM]FY2e9dM]F[2f9bMZF^2h9_MZF`2i9]MXFb2j9[MXFd2j9ZMVFf2m9WMTFh2n9UMTFj2o9SMQFm2Q:PMQFo2R:nLoEQ3S:lLoES3T:jLlEV3V:gLlEX3W:eLjE[3W:bLjE^3X:`LiE_3Z:]LhEb3Z:\\LgEc3\\:YLeEg3]:WLdEh3[=\\NZ@lNf?Q1g1\\NoY3"}, "label": "a little girl"}]} {"id": 354391, "image": "COCO_train2014_000000354391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child sitting on his dad ' s lap\"."}], "task": "refering", "answer_template": "The \"a child sitting on his dad ' s lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 109, 120, 121, 122, 123, 134, 135, 136, 137, 148, 149, 150, 151, 161, 162, 163, 164, 165, 166, 174, 175, 176, 177, 178, 179, 180, 188, 189, 190, 191, 192, 193, 194, 195, 201, 202, 203, 204, 205, 206, 207, 208, 209, 218, 219, 220, 221, 222, 232, 233, 234, 235, 236, 245, 246, 247, 248, 249, 250, 251, 259, 260, 261, 262, 263, 264, 265, 273, 274, 275, 276, 277, 278, 279, 287, 288, 289, 290, 291, 292, 293, 301, 302, 303, 304, 305, 306, 307, 316, 317, 318, 319, 320, 321], "bbox": [0.38302872062663185, 0.303375, 0.9838381201044386, 0.968546875], "iscrowd": 0, "area": 53836.19875, "rle": {"size": [640, 383], "counts": "mXl21nc01N2O1N2O1N2O1N2O1N101N2O1N2O1N2O1N2O1N2O1N2O1N2O0O2O1N2O1N2O1N2O1N2O1N2O1N2O1N101N2O1N2O1O1N2O1N2O1O1N2O1O1O1O001O1O1O1N2O1O1O1O1O1cDiMP5X2UJcNj5^1[I]Oc6e0cH5\\7LiGo0V8QOQGi1n8XNlFn1S9SNjFP2U9QNhFR2W9oMeFV2Y9kMdFX2Z9jMcFY2\\9hM`F\\2_9eM^F^2Q7]JdJV3XN`2T7_J^JR3[Nb2V7bJYJm2]Ne2Y7dJTJh2`Ng2[7gJoIc2cNi2]7iJjI`2fNj2_7lJeIZ2iNn2a7nJ`Id8_6bG[I_8d6gGVIZ8i6lGQIU8n6QHlHP8S7UHhHl7X7XHbHj7]7VHcHk7\\7UHdHl7[7]2O1O1O1O1O1O1O101N1O1O1O1O1O1O1O1O1O100000000000000000000000000000000000001O00000000000000PFYIg6g6VI\\Ij6d6TI_Ik6a6RIbIn6^6PIdIP7\\6mHgIS7Y6kHiIU7W6hHlIX7T6fHoIY7Q6dHRJm1WOP1g6PMUJP2UOo0f6oLWJQ2TOP1e6lLZJT2ROo0d6kL\\JU2ROo0b6iL`JW2nNP1c6fLaJY2nNP1j6[L[Je2lNo0Q7RLUJn2lNo0X7gKoIZ3iNo0a7\\KhIe3hNn0h7RKcIo3gNn0g9QOYFP1f9oN\\FP1d9PO]Fo0c9PO_Fo0a9QO_Fo0a9POaFo0`9POaFo0_9PObFP1^9POcFo0]9POeFo0[9QOfFn0Z9QOgFo0Y9QOhFn0Y9POiFo0W9QOiFo0W9POkFo0U9QOlFn0T9QOnFn0R9ROnFn0S9POoFo0Q9QOPGn0P9QOQGo0o8QORGn0n8ROSGm0m8ROTGn0o8oNRGP1P9mNRGR1P9lNQGS1R9iNoFW1S9gNnFY1T9cNnF\\1T9bNlF^1W9^NkFa1W9]NjFb1Y9ZNiFe1Y9YNgFg1\\9UNfFj1\\9TNeFk1]9RNdFn1_9oMbFP2`9mMbFR2a9kM`FT2b9iM_FW2d9fM]FY2e9dM]F[2f9bMZF^2h9_MZF`2i9]MXFb2j9[MXFd2j9ZMVFf2m9WMTFh2n9UMTFj2o9SMQFm2Q:PMQFo2R:nLoEQ3S:lLoES3T:jLlEV3V:gLlEX3W:eLjE[3W:bLjE^3X:`LiE_3Z:]LhEb3Z:\\LgEc3\\:YLeEg3]:WLdEh3[=\\NZ@lNf?Q1g1\\NoY3"}, "label": "a child sitting on his dad ' s lap"}]} {"id": 354391, "image": "COCO_train2014_000000354391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a fedora is hugging a girl\"."}], "task": "refering", "answer_template": "The \"a man in a fedora is hugging a girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 63, 64, 74, 75, 76, 77, 78, 79, 87, 88, 89, 90, 91, 92, 93, 101, 102, 103, 104, 105, 106, 114, 115, 116, 117, 118, 119, 120, 127, 128, 129, 130, 131, 132, 133, 134, 141, 142, 143, 144, 145, 146, 147, 148, 154, 155, 156, 157, 158, 159, 160, 161, 168, 169, 170, 171, 172, 173, 174, 175, 182, 183, 184, 185, 186, 187, 188, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 224, 225, 226, 227, 228, 229, 230, 238, 239, 240, 241, 242, 243, 244, 245, 252, 253, 254, 255, 256, 257, 258, 259, 266, 267, 268, 269, 270, 271, 272, 273, 280, 281, 282, 283, 284, 285, 286, 287, 294, 295, 296, 297, 298, 299, 300, 310, 311], "bbox": [0.033785900783289816, 0.17093750000000002, 0.657154046997389, 0.9641875000000001], "iscrowd": 0, "area": 79307.27475000001, "rle": {"size": [640, 383], "counts": "We8i0Rc06J7I6VIYOSJn0i5ZOmIm0o5[OfIk0V6]O_Ik0]6]OSIn0j6F[EiN[1k1V9i1lEbNP:i5L4M4K4L5L3N2M4M2M3M4M2M3L4L3N2M4L3M3N3L3M3O2N101N100O2O0O101O0O2O0O101N10001N100O100O10O010000O100O100O010O10000O100O100O01000O100O100N2N2M2O2M3N2I7H8G9H8H8G8O2O1O1N2O1O1O1N2O001N2O1O1O1N2O1O001N2O1O1N2O1O1O1N101O1N2O1N2N2O1N2O0O2O1O1O1O100O100O00100O100O1O100O10O2O0000000000000^IgK\\MZ4b2ULQMk3m2dLfL\\3X3SM[Ln2b3^MTLb2j3bMTL^2j3eMUL[2j3hMTLX2j3kMULU2i3oMTLR2j3QNULo1i3UNULk1i3XNVLh1h3\\NVLd1h3_NWLa1g3cNWL^1f3fNXLZ1f3iNYLW1e3mNXLT1g3nNXLR1f3ROXLn0f3UOYLk0e3YOYLg0e3\\OZLd0d3@YLb0d3AULe0i3_OnKh0P4\\OgKl0V4WOcKo0[4UOUKY1i4jNeJh1Y5[NTJV2j5mM`Ii2]6^5O1N3N1N2O2M2O2N1N2O2M2O1O2M2O2M2O1M4M2N2M4M2N3L3N3M2N3L4M3M3L4K5K4K6K5K5J6K5K4K6K5K5J6K5K5J5L5K5K5J6K5K5E:XOSha2"}, "label": "a man in a fedora is hugging a girl"}]} {"id": 354391, "image": "COCO_train2014_000000354391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing hat holding child\"."}], "task": "refering", "answer_template": "The \"man wearing hat holding child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 63, 64, 74, 75, 76, 77, 78, 79, 87, 88, 89, 90, 91, 92, 93, 101, 102, 103, 104, 105, 106, 114, 115, 116, 117, 118, 119, 120, 127, 128, 129, 130, 131, 132, 133, 134, 141, 142, 143, 144, 145, 146, 147, 148, 154, 155, 156, 157, 158, 159, 160, 161, 168, 169, 170, 171, 172, 173, 174, 175, 182, 183, 184, 185, 186, 187, 188, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 224, 225, 226, 227, 228, 229, 230, 238, 239, 240, 241, 242, 243, 244, 245, 252, 253, 254, 255, 256, 257, 258, 259, 266, 267, 268, 269, 270, 271, 272, 273, 280, 281, 282, 283, 284, 285, 286, 287, 294, 295, 296, 297, 298, 299, 300, 310, 311], "bbox": [0.033785900783289816, 0.17093750000000002, 0.657154046997389, 0.9641875000000001], "iscrowd": 0, "area": 79307.27475000001, "rle": {"size": [640, 383], "counts": "We8i0Rc06J7I6VIYOSJn0i5ZOmIm0o5[OfIk0V6]O_Ik0]6]OSIn0j6F[EiN[1k1V9i1lEbNP:i5L4M4K4L5L3N2M4M2M3M4M2M3L4L3N2M4L3M3N3L3M3O2N101N100O2O0O101O0O2O0O101N10001N100O100O10O010000O100O100O010O10000O100O100O01000O100O100N2N2M2O2M3N2I7H8G9H8H8G8O2O1O1N2O1O1O1N2O001N2O1O1O1N2O1O001N2O1O1N2O1O1O1N101O1N2O1N2N2O1N2O0O2O1O1O1O100O100O00100O100O1O100O10O2O0000000000000^IgK\\MZ4b2ULQMk3m2dLfL\\3X3SM[Ln2b3^MTLb2j3bMTL^2j3eMUL[2j3hMTLX2j3kMULU2i3oMTLR2j3QNULo1i3UNULk1i3XNVLh1h3\\NVLd1h3_NWLa1g3cNWL^1f3fNXLZ1f3iNYLW1e3mNXLT1g3nNXLR1f3ROXLn0f3UOYLk0e3YOYLg0e3\\OZLd0d3@YLb0d3AULe0i3_OnKh0P4\\OgKl0V4WOcKo0[4UOUKY1i4jNeJh1Y5[NTJV2j5mM`Ii2]6^5O1N3N1N2O2M2O2N1N2O2M2O1O2M2O2M2O1M4M2N2M4M2N3L3N3M2N3L4M3M3L4K5K4K6K5K5J6K5K4K6K5K5J6K5K5J5L5K5K5J6K5K5E:XOSha2"}, "label": "man wearing hat holding child"}]} {"id": 313437, "image": "COCO_train2014_000000313437.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small lamb in man ' s right hand\"."}], "task": "refering", "answer_template": "The \"small lamb in man ' s right hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [283, 305, 306, 307, 308, 327, 329, 330, 331, 332, 350, 351, 352, 353, 354, 355, 356, 373, 374, 375, 376, 396, 397, 398, 399, 419, 420, 421, 422, 442, 443, 444], "bbox": [0.22484375, 0.5569531249999999, 0.49035937500000004, 0.84175], "iscrowd": 0, "area": 11544.93615, "rle": {"size": [640, 640], "counts": "k\\j23kc05K6J5J6T^OAo?e0j_OMh?6Q@7f?LS@a0d?CT@g0e?]OT@g0l?\\Om_Oh0S`0c100O10000O10O1000O1000000O01OcNULfBj3Y=YLfBf3Z=\\LdBd3\\=^LcBa3\\=aLdB^3\\=dLbBd0kNf1b>iMbB4]On1Q>PNaBLIn1e=XNbBF1Z1kNoNb>2aBB;X1fNoN]>8cB]Oa0Y1cNmNY>>^Db0\\MlNU>c0cD>l;BXD;g;F\\D7b;JdD0\\;1kDGU;9RE@m:a0VE\\Oj:e0WEXOj:h0XEVOh:k0XETOg:m0V400000O2O00000O101c]OoNia0Q1W^OTOda03k]O?`0CKf2X>XM_A>>Q3S>_LbA?:T3S>\\LeA?8U3S>[LgA`04W3S>XLlA`01X3S>WLnA`0N[3S>TLQB`0L\\3g>cLXA_3h>aLWA_3j>`LUAb3k>aLQA_3P?i0000\\NVBPMk=n2e1M1O1O1hN\\_OoNk`0h0Z15K5K5L3L4L4L4L4M:F`1`NVM\\@Y1d?`Nf@\\1[?]Nn@`1S?\\NSAb1m>]NWA_1j>`NYA]1h>bN[A\\1e>cN_AY1b>fNaAW1`>hNhAQ1X>nNPBj0Q>UOWBd0i=[O^B>c=AdB8\\=IiB2W=MPCLo<5XCDg<=_C^O^ in this image.", "objects": [{"patches": [148, 149, 164, 165, 166, 180, 181, 182, 183, 184, 185, 197, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 219, 230, 232, 233, 234, 235, 249, 250, 251, 252, 266, 267, 268, 269, 283, 284, 285, 286, 300, 301, 302, 303, 317, 318, 319, 320, 334, 335, 336, 337], "bbox": [0.5348958333333333, 0.34831249999999997, 0.9207500000000001, 0.850296875], "iscrowd": 0, "area": 22758.37020000001, "rle": {"size": [640, 480], "counts": "hoP51nc01O1O1O1O1O1N2O1O2N1O1O2N1O1O2N1O1O00N2O1N2N2000000OO2O1O1N2O1O001N2O1O1O0O2O1O1N2O1O001N2O1000O1000000000\\CQOX6n0hIVOU6j0jIYOT6g0kI]OR6c0mI@Q6`0mICR6=nIDQ6;PJGo58QJIn57RJJm56SIZObKa0Z;5nHFbK6_;3jH3bKJc;3eH?bK@g;3`Hf1_7aNTHf1k7aNiEeN8Q3n9aN[E5ROa1b;aNnDh3Q;_LaDi3^;]2O1O1O1O001O1O1^Ob0O1O001N101O001OPNiEXJW:f5hF^IW9`6T201N2N2N2N2N2N2N101N2N2N2N2N2N3M5L41O2N10001NYJ`Bm4`=jJfBZ5n=3N2M4L7J7H8WCVJl:R6bDZJ];U7O2O0O2N2O0O1O01O000010O0001O00001mLdE`L\\:\\3SFYLm9b3cFRL_9h3QGlKP9o3bGbK`8X4h3F9G:ROn0N2O1N2O1N2N2N2N2N2N2N2N2N2N2N2N2N2L4L4K5L4L4L4M3N2M3M2N3N2M2N3N2M2N3L4K\\^g0"}, "label": "a man with blue color shirt standing along with him friend"}]} {"id": 116832, "image": "COCO_train2014_000000116832.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white man with a blue shirt and grey pants touching a yellow hydrant next to another man\"."}], "task": "refering", "answer_template": "The \"a white man with a blue shirt and grey pants touching a yellow hydrant next to another man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 164, 165, 166, 180, 181, 182, 183, 184, 185, 197, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 219, 230, 232, 233, 234, 235, 249, 250, 251, 252, 266, 267, 268, 269, 283, 284, 285, 286, 300, 301, 302, 303, 317, 318, 319, 320, 334, 335, 336, 337], "bbox": [0.5348958333333333, 0.34831249999999997, 0.9207500000000001, 0.850296875], "iscrowd": 0, "area": 22758.37020000001, "rle": {"size": [640, 480], "counts": "hoP51nc01O1O1O1O1O1N2O1O2N1O1O2N1O1O2N1O1O00N2O1N2N2000000OO2O1O1N2O1O001N2O1O1O0O2O1O1N2O1O001N2O1000O1000000000\\CQOX6n0hIVOU6j0jIYOT6g0kI]OR6c0mI@Q6`0mICR6=nIDQ6;PJGo58QJIn57RJJm56SIZObKa0Z;5nHFbK6_;3jH3bKJc;3eH?bK@g;3`Hf1_7aNTHf1k7aNiEeN8Q3n9aN[E5ROa1b;aNnDh3Q;_LaDi3^;]2O1O1O1O001O1O1^Ob0O1O001N101O001OPNiEXJW:f5hF^IW9`6T201N2N2N2N2N2N2N101N2N2N2N2N2N3M5L41O2N10001NYJ`Bm4`=jJfBZ5n=3N2M4L7J7H8WCVJl:R6bDZJ];U7O2O0O2N2O0O1O01O000010O0001O00001mLdE`L\\:\\3SFYLm9b3cFRL_9h3QGlKP9o3bGbK`8X4h3F9G:ROn0N2O1N2O1N2N2N2N2N2N2N2N2N2N2N2N2N2L4L4K5L4L4L4M3N2M3M2N3N2M2N3N2M2N3L4K\\^g0"}, "label": "a white man with a blue shirt and grey pants touching a yellow hydrant next to another man"}]} {"id": 116832, "image": "COCO_train2014_000000116832.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a baby blue shirt standing next to a yellow fire hydrant\"."}], "task": "refering", "answer_template": "The \"a man wearing a baby blue shirt standing next to a yellow fire hydrant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 158, 159, 160, 173, 174, 175, 176, 177, 190, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 225, 226, 227, 228, 229, 243, 244, 245, 246, 260, 261, 262, 277, 278, 279, 280, 294, 295, 296, 297, 311, 312, 313, 314, 328, 329, 330, 331], "bbox": [0.2130625, 0.37707812500000004, 0.5195000000000001, 0.8585312500000001], "iscrowd": 0, "area": 24867.859549999997, "rle": {"size": [640, 480], "counts": "hQP21mc04M3L3M4M3L4L4L4M2M4L4M2M4N1N2O2N1N3N1O1N3N1O1N3N1O2M2O1O2M2O1O2M2O2N1N2Ol0SO5L3M4K5L3M4K5L3M4K5L4L3L5L4L7H8I7I8G8I7I7H in this image.", "objects": [{"patches": [142, 143, 158, 159, 160, 173, 174, 175, 176, 177, 190, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 225, 226, 227, 228, 229, 243, 244, 245, 246, 260, 261, 262, 277, 278, 279, 280, 294, 295, 296, 297, 311, 312, 313, 314, 328, 329, 330, 331], "bbox": [0.2130625, 0.37707812500000004, 0.5195000000000001, 0.8585312500000001], "iscrowd": 0, "area": 24867.859549999997, "rle": {"size": [640, 480], "counts": "hQP21mc04M3L3M4M3L4L4L4M2M4L4M2M4N1N2O2N1N3N1O1N3N1O1N3N1O2M2O1O2M2O1O2M2O2N1N2Ol0SO5L3M4K5L3M4K5L3M4K5L4L3L5L4L7H8I7I8G8I7I7H in this image.", "objects": [{"patches": [27, 28, 29, 30, 31, 45, 46, 47, 48, 49, 63, 64, 65, 66, 67], "bbox": [0.50966, 0.09437037037037037, 0.74994, 0.3932592592592593], "iscrowd": 0, "area": 7392.009050000001, "rle": {"size": [270, 500], "counts": "SZS26R87H8L4M2N3mHSOb6o0\\IROb6Q1[IQOc6R1[IoNc6T1ZInNd6U1ZIlNe6_1N110O1O1O100O1O1O1O10O01O1O100O1O1O100000O100000O100000000000O01000000000000O1000O1000000000O10000000O1000O1000000000000O10O1000000000O10000000O100000O100000000000O01000000000000O1000O10000000O12N2M4J6J5J7J6J5K6G9G8HWmP1"}, "label": "a man with a black head wrap talking on a phone in a blue car"}]} {"id": 428149, "image": "COCO_train2014_000000428149.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in a car talking on a cellphone\"."}], "task": "refering", "answer_template": "The \"man in a car talking on a cellphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 28, 29, 30, 31, 45, 46, 47, 48, 49, 63, 64, 65, 66, 67], "bbox": [0.50966, 0.09437037037037037, 0.74994, 0.3932592592592593], "iscrowd": 0, "area": 7392.009050000001, "rle": {"size": [270, 500], "counts": "SZS26R87H8L4M2N3mHSOb6o0\\IROb6Q1[IQOc6R1[IoNc6T1ZInNd6U1ZIlNe6_1N110O1O1O100O1O1O1O10O01O1O100O1O1O100000O100000O100000000000O01000000000000O1000O1000000000O10000000O1000O1000000000000O10O1000000000O10000000O100000O100000000000O01000000000000O1000O10000000O12N2M4J6J5J7J6J5K6G9G8HWmP1"}, "label": "man in a car talking on a cellphone"}]} {"id": 428149, "image": "COCO_train2014_000000428149.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"arm holding cell phone to ear\"."}], "task": "refering", "answer_template": "The \"arm holding cell phone to ear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 81, 82, 83, 84, 98, 99, 100, 101, 102, 103, 116, 117, 118, 119, 120, 121, 134, 135, 136, 137, 138, 139], "bbox": [0.44812, 0.3601481481481481, 0.76848, 0.7892222222222222], "iscrowd": 0, "area": 13794.786500000002, "rle": {"size": [270, 500], "counts": "bXk12Y83N2M3N3L3N2M3N2M3N2M3M3N2M3N2M3N2K5K5K5K4L5M3M3M3M3M3M3M3M3M3L4M3000000000000O100000000000000000O010000000000000000O10000000000000000O100000000000000000O010000000000000000O1000000000000000000O10000000000000000O0100000000000000000O10000000000000000O1000000000000000ZMYKe1g4RNhKg1X4TNSLg1m3TNYLk1g3PN_Lo1Q5O2N1O1O1O1O1O001O1O1O1N2N2N2N1O3M4L3M4L4F9Ec]n0"}, "label": "arm holding cell phone to ear"}]} {"id": 428149, "image": "COCO_train2014_000000428149.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bue shirt\"."}], "task": "refering", "answer_template": "The \"a bue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 81, 82, 83, 84, 98, 99, 100, 101, 102, 103, 116, 117, 118, 119, 120, 121, 134, 135, 136, 137, 138, 139], "bbox": [0.44812, 0.3601481481481481, 0.76848, 0.7892222222222222], "iscrowd": 0, "area": 13794.786500000002, "rle": {"size": [270, 500], "counts": "bXk12Y83N2M3N3L3N2M3N2M3N2M3M3N2M3N2M3N2K5K5K5K4L5M3M3M3M3M3M3M3M3M3L4M3000000000000O100000000000000000O010000000000000000O10000000000000000O100000000000000000O010000000000000000O1000000000000000000O10000000000000000O0100000000000000000O10000000000000000O1000000000000000ZMYKe1g4RNhKg1X4TNSLg1m3TNYLk1g3PN_Lo1Q5O2N1O1O1O1O1O001O1O1O1N2N2N2N1O3M4L3M4L4F9Ec]n0"}, "label": "a bue shirt"}]} {"id": 198785, "image": "COCO_train2014_000000198785.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with long hair on the left\"."}], "task": "refering", "answer_template": "The \"woman with long hair on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 38, 39, 40, 55, 56, 57, 71, 72, 73, 74, 75, 87, 88, 89, 90, 91, 92, 104, 105, 106, 107, 108, 109, 121, 122, 123, 124, 125, 126, 139, 140, 141, 142, 156, 157, 158, 159, 173, 174, 175, 176, 190, 191, 192, 193, 207, 208, 209, 210, 225, 226, 227, 242, 243, 244, 260], "bbox": [0.1476875, 0.07896875, 0.45591666666666664, 0.6728906250000001], "iscrowd": 0, "area": 30331.58125, "rle": {"size": [640, 480], "counts": "o`\\16hc07H8I5L5L4L4L3L5K2N108G2N1O100O3M3M3N2M3M5K4K2O2N1O2N1j_O]Mo>d2k@hMm>Z2m@QNl>P2o@]Nh>e1n@nNg>`3F:G9F:E9H3L3N2M3N2M3M3N2M2O2M3M1O0100O100O10ON2L4McNg1E;H9H8H6J6J7I3M10000SGcEV8^:eGkEW8W:cGnE\\8k:O2O1O1N2O1O1N2O001N1000000001O0O101O0N3M2O1K6H72O1N1O2OO000010O04L4kNVGlFo8h8\\GTGm8^8^G[Gl8W8d1H7gNfDdIa;Y6`DcIe;`4UD`L7mNi;V4]DkLKkNn;V4ZDmLHjNS in this image.", "objects": [{"patches": [22, 38, 39, 40, 55, 56, 57, 71, 72, 73, 74, 75, 87, 88, 89, 90, 91, 92, 104, 105, 106, 107, 108, 109, 121, 122, 123, 124, 125, 126, 139, 140, 141, 142, 156, 157, 158, 159, 173, 174, 175, 176, 190, 191, 192, 193, 207, 208, 209, 210, 225, 226, 227, 242, 243, 244, 260], "bbox": [0.1476875, 0.07896875, 0.45591666666666664, 0.6728906250000001], "iscrowd": 0, "area": 30331.58125, "rle": {"size": [640, 480], "counts": "o`\\16hc07H8I5L5L4L4L3L5K2N108G2N1O100O3M3M3N2M3M5K4K2O2N1O2N1j_O]Mo>d2k@hMm>Z2m@QNl>P2o@]Nh>e1n@nNg>`3F:G9F:E9H3L3N2M3N2M3M3N2M2O2M3M1O0100O100O10ON2L4McNg1E;H9H8H6J6J7I3M10000SGcEV8^:eGkEW8W:cGnE\\8k:O2O1O1N2O1O1N2O001N1000000001O0O101O0N3M2O1K6H72O1N1O2OO000010O04L4kNVGlFo8h8\\GTGm8^8^G[Gl8W8d1H7gNfDdIa;Y6`DcIe;`4UD`L7mNi;V4]DkLKkNn;V4ZDmLHjNS in this image.", "objects": [{"patches": [25, 26, 27, 42, 43, 44, 59, 60, 61, 62, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 143, 144, 145, 146, 147, 160, 161, 162, 163, 164, 177, 178, 193, 194, 195, 210, 211, 212, 227, 244, 245, 260, 261, 262, 278], "bbox": [0.35072916666666665, 0.061953125, 0.6957708333333333, 0.7222187499999999], "iscrowd": 0, "area": 25804.641450000003, "rle": {"size": [640, 480], "counts": "X]Y3k0Rc06I6K5J6K5K3L4L4L4L4K5J6M3M3L3N3M2O2O1N101O0O2O0SOhLRAZ3k>iLTAW3j>lLQAX3l>lLk@f0Fb1^?kMl@b0NY1V?XNm@=5Q1n>eNm@8V11m=Jm@4`1Fc=9m@0g1]O_=d0j@N`a04`^OJca05]^OJea05[^OJga05Y^OJia05W^OIla06T^OId`0@Y@e0SOJc`0C[@b0SOIb`0G[@?SOIa`0K\\@;SOJ_`0O]@6TOJ_`02]@3TOK]`05_@OTOK\\`09a@JSOMZ`0;d@GROMd=K_Dc0kNDRON_=>TD3[O@RON[=R1hCCK[OSO0U=`1cCVO6YOSO0o in this image.", "objects": [{"patches": [25, 26, 27, 42, 43, 44, 59, 60, 61, 62, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 109, 110, 111, 112, 113, 126, 127, 128, 129, 130, 143, 144, 145, 146, 147, 160, 161, 162, 163, 164, 177, 178, 193, 194, 195, 210, 211, 212, 227, 244, 245, 260, 261, 262, 278], "bbox": [0.35072916666666665, 0.061953125, 0.6957708333333333, 0.7222187499999999], "iscrowd": 0, "area": 25804.641450000003, "rle": {"size": [640, 480], "counts": "X]Y3k0Rc06I6K5J6K5K3L4L4L4L4K5J6M3M3L3N3M2O2O1N101O0O2O0SOhLRAZ3k>iLTAW3j>lLQAX3l>lLk@f0Fb1^?kMl@b0NY1V?XNm@=5Q1n>eNm@8V11m=Jm@4`1Fc=9m@0g1]O_=d0j@N`a04`^OJca05]^OJea05[^OJga05Y^OJia05W^OIla06T^OId`0@Y@e0SOJc`0C[@b0SOIb`0G[@?SOIa`0K\\@;SOJ_`0O]@6TOJ_`02]@3TOK]`05_@OTOK\\`09a@JSOMZ`0;d@GROMd=K_Dc0kNDRON_=>TD3[O@RON[=R1hCCK[OSO0U=`1cCVO6YOSO0o in this image.", "objects": [{"patches": [164, 178, 179, 180, 181, 193, 194, 195, 196, 208, 209, 210, 211, 224, 225, 226, 227, 240, 241, 242, 256, 257, 258, 272, 273, 274, 289], "bbox": [0.02782894736842105, 0.46460937500000005, 0.3306359649122807, 0.791984375], "iscrowd": 0, "area": 10976.508399999995, "rle": {"size": [640, 456], "counts": "bc85ec05PMMlAk0h@^OU?h2O0NSLR@c3W`0Gj3P1L4M2@a0L4L4K5N2M3N2N2N2M3M3M3N1N3fN^^O3ea0L_^ONda01`^OGea08_^O@fa0`0]^OXOga0h0i000O10O0100O100O010O10O10O010O10O0100O01@XOh]Og0Zb0ZOd]Of0]b0[O`]Of0`b0\\O^]Oc0db0_OY]Oa0gb0;2NN3M2N30O101N1O2O0O3L4L5L4K4L2O1O01000Fa\\O4^c0Ld\\O2\\c0Nf\\O1Yc0Oj\\ONVc02;12M3NVZn5"}, "label": "the giraffe on the left in the right hand picture"}]} {"id": 411778, "image": "COCO_train2014_000000411778.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffie in back\"."}], "task": "refering", "answer_template": "The \"giraffie in back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 178, 179, 180, 181, 193, 194, 195, 196, 208, 209, 210, 211, 224, 225, 226, 227, 240, 241, 242, 256, 257, 258, 272, 273, 274, 289], "bbox": [0.02782894736842105, 0.46460937500000005, 0.3306359649122807, 0.791984375], "iscrowd": 0, "area": 10976.508399999995, "rle": {"size": [640, 456], "counts": "bc85ec05PMMlAk0h@^OU?h2O0NSLR@c3W`0Gj3P1L4M2@a0L4L4K5N2M3N2N2N2M3M3M3N1N3fN^^O3ea0L_^ONda01`^OGea08_^O@fa0`0]^OXOga0h0i000O10O0100O100O010O10O10O010O10O0100O01@XOh]Og0Zb0ZOd]Of0]b0[O`]Of0`b0\\O^]Oc0db0_OY]Oa0gb0;2NN3M2N30O101N1O2O0O3L4L5L4K4L2O1O01000Fa\\O4^c0Ld\\O2\\c0Nf\\O1Yc0Oj\\ONVc02;12M3NVZn5"}, "label": "giraffie in back"}]} {"id": 100485, "image": "COCO_train2014_000000100485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown dog paw is holding onto an empty water bottle\"."}], "task": "refering", "answer_template": "The \"a brown dog paw is holding onto an empty water bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 115, 116, 117, 138, 139, 140, 141, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.0013281249999999999, 0.2633018867924528, 0.544953125, 0.9891037735849056], "iscrowd": 0, "area": 58807.14795, "rle": {"size": [424, 640], "counts": "Rc0d6U5a1VOi001N2O001O0O2O1O001N2O00O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001bL[GW2f8gM\\GX2d8fM_GY2b8eM`GZ2a8cMbG\\2_8bMcG]2^8bMcG]2^8aMcG_2^8`McG_2^8_MdG`2]8_MdG`2]8_MdG`2]8^MeGa2\\8^MeGa2\\8]MfGb2[8]MfGb2[8\\MgGc2Z8\\MgGc2Z8[MhGd2Y8[MhGd2Y8[MgGe2Z8YMhGf2^9O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001N2O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1OgRh3"}, "label": "a brown dog paw is holding onto an empty water bottle"}]} {"id": 100485, "image": "COCO_train2014_000000100485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown dog paw\"."}], "task": "refering", "answer_template": "The \"brown dog paw\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 115, 116, 117, 138, 139, 140, 141, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.0013281249999999999, 0.2633018867924528, 0.544953125, 0.9891037735849056], "iscrowd": 0, "area": 58807.14795, "rle": {"size": [424, 640], "counts": "Rc0d6U5a1VOi001N2O001O0O2O1O001N2O00O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O001O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O001O1O1O1O001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001bL[GW2f8gM\\GX2d8fM_GY2b8eM`GZ2a8cMbG\\2_8bMcG]2^8bMcG]2^8aMcG_2^8`McG_2^8_MdG`2]8_MdG`2]8_MdG`2]8^MeGa2\\8^MeGa2\\8]MfGb2[8]MfGb2[8\\MgGc2Z8\\MgGc2Z8[MhGd2Y8[MhGd2Y8[MgGe2Z8YMhGf2^9O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O001N2O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1OgRh3"}, "label": "brown dog paw"}]} {"id": 100485, "image": "COCO_train2014_000000100485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dogs nose\"."}], "task": "refering", "answer_template": "The \"a dogs nose\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 56, 57, 58, 59, 80, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205], "bbox": [0.448125, 0.11516509433962263, 0.9870468750000001, 0.5960613207547171], "iscrowd": 0, "area": 33351.0823, "rle": {"size": [424, 640], "counts": "alf31W=1N2O1N3N1O1N2O1N2O1O1N2O1N3N1O1N2O1N2O12NZDWOP;e0mD_OU;?fDG\\;5`D0b;M[D8g;DUDa0n;=1NN3M3M2N3M2N3M2N3M2N3M1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1N2N2O1N2N2O1N101N1O101N1O101O01O100[MQFo1P:oMSFo1n9oMTFP2l9oMVFQ2j9lMZFR2g9lM[FS2f9kM\\FU2d9hM`FV2a9hMaFW2X:10O0010O01O010O0010O0000001O0O101O1N2O1O1O2M2O1O1O1N2O2N1N2O1O1O2M2O1O1O1N3N1O1N2O1O2N1O1O1O1O2N1O1O1O1O001O00001O00001O00001O001O01O01O00001O`LjFn2V9QMkFo2U9PMlFQ3S9nLoFQ3Q9nLPGR3P9mLQGT3n8kLSGU3m8jLTGV3l8iLUGX3j8gLXGX3h8gLYGY3g8fLZG[3e8dL\\G\\3d8cL]G^3b8aL_G_3a8`LaG_3_8`LbGa3R900001O01O01O000000010O0000000001O000000000010O000001O0000001O000010O000001O0000001O0000010O0000000000O10000O101O0O100O10000O10000O10001N10000O100O10000O1O1O2N1O100O1O100O100O1O101N100O1O010O100O00100O010O1O10O0100O0101O1N101O1O1N101O1N2O001O1N2O001N2O1O001N2O001N101O0O2O00001N1M4L3N3L3N2M4L3N3L3N3L3N2N3M2N3M2N3M2N3M2N2N3M2N3M2N6J;E:F:FiT3"}, "label": "a dogs nose"}]} {"id": 100485, "image": "COCO_train2014_000000100485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pink and black snout of an animal with whiskers coming off it\"."}], "task": "refering", "answer_template": "The \"pink and black snout of an animal with whiskers coming off it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 56, 57, 58, 59, 80, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205], "bbox": [0.448125, 0.11516509433962263, 0.9870468750000001, 0.5960613207547171], "iscrowd": 0, "area": 33351.0823, "rle": {"size": [424, 640], "counts": "alf31W=1N2O1N3N1O1N2O1N2O1O1N2O1N3N1O1N2O1N2O12NZDWOP;e0mD_OU;?fDG\\;5`D0b;M[D8g;DUDa0n;=1NN3M3M2N3M2N3M2N3M2N3M1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1N2N2O1N2N2O1N101N1O101N1O101O01O100[MQFo1P:oMSFo1n9oMTFP2l9oMVFQ2j9lMZFR2g9lM[FS2f9kM\\FU2d9hM`FV2a9hMaFW2X:10O0010O01O010O0010O0000001O0O101O1N2O1O1O2M2O1O1O1N2O2N1N2O1O1O2M2O1O1O1N3N1O1N2O1O2N1O1O1O1O2N1O1O1O1O001O00001O00001O00001O001O01O01O00001O`LjFn2V9QMkFo2U9PMlFQ3S9nLoFQ3Q9nLPGR3P9mLQGT3n8kLSGU3m8jLTGV3l8iLUGX3j8gLXGX3h8gLYGY3g8fLZG[3e8dL\\G\\3d8cL]G^3b8aL_G_3a8`LaG_3_8`LbGa3R900001O01O01O000000010O0000000001O000000000010O000001O0000001O000010O000001O0000001O0000010O0000000000O10000O101O0O100O10000O10000O10001N10000O100O10000O1O1O2N1O100O1O100O100O1O101N100O1O010O100O00100O010O1O10O0100O0101O1N101O1O1N101O1N2O001O1N2O001N2O1O001N2O001N101O0O2O00001N1M4L3N3L3N2M4L3N3L3N3L3N2N3M2N3M2N3M2N3M2N2N3M2N3M2N6J;E:F:FiT3"}, "label": "pink and black snout of an animal with whiskers coming off it"}]} {"id": 100485, "image": "COCO_train2014_000000100485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown part of the dogs face\"."}], "task": "refering", "answer_template": "The \"the brown part of the dogs face\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 82, 83, 84, 85, 86, 87, 88, 89, 108, 109, 110, 111], "bbox": [0.34759375000000003, 0.0025943396226415097, 0.965359375, 0.32207547169811324], "iscrowd": 0, "area": 33362.50644999999, "rle": {"size": [424, 640], "counts": "lZl21W=2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O000000O100000000O10000000000O10000000000O100000000O100000000001O3M2N2N2N2N2N3M2N2N2N2N2N1O00000000000000000000O1000000000000000O10000000000000O100000000000000001O00001O001O001O001O001O001O00001O001O001O001O001O006J6J6J2N00000000O10000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000001O00001O0000001O00001O0000001O00001O00001O000aFdLR9\\3lFfLT9[3iFhLU9Y3iFiLW9W3gFkLY9c3O001O00001O00001O001O00001O001O00001O00001O001O0000000000000000000000000000000000000000000000000000000000000000O100O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O001N2O1O1O1O1O1O1O1N2O1O1O1N2M3M3N2M3M3N2M3M3N2M3N2M3M3N2M3M3N2M3M3N2M3N2MbS9"}, "label": "the brown part of the dogs face"}]} {"id": 100485, "image": "COCO_train2014_000000100485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog is trying to chew on an empty water bottle\"."}], "task": "refering", "answer_template": "The \"a dog is trying to chew on an empty water bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 82, 83, 84, 85, 86, 87, 88, 89, 108, 109, 110, 111], "bbox": [0.34759375000000003, 0.0025943396226415097, 0.965359375, 0.32207547169811324], "iscrowd": 0, "area": 33362.50644999999, "rle": {"size": [424, 640], "counts": "lZl21W=2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O000000O100000000O10000000000O10000000000O100000000O100000000001O3M2N2N2N2N2N3M2N2N2N2N2N1O00000000000000000000O1000000000000000O10000000000000O100000000000000001O00001O001O001O001O001O001O00001O001O001O001O001O006J6J6J2N00000000O10000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000001O00001O0000001O00001O0000001O00001O00001O000aFdLR9\\3lFfLT9[3iFhLU9Y3iFiLW9W3gFkLY9c3O001O00001O00001O001O00001O001O00001O00001O001O0000000000000000000000000000000000000000000000000000000000000000O100O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O001N2O1O1O1O1O1O1O1N2O1O1O1N2M3M3N2M3M3N2M3M3N2M3N2M3M3N2M3M3N2M3M3N2M3N2MbS9"}, "label": "a dog is trying to chew on an empty water bottle"}]} {"id": 182406, "image": "COCO_train2014_000000182406.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue and black backpack in front of a man on the phone\"."}], "task": "refering", "answer_template": "The \"blue and black backpack in front of a man on the phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [316, 317, 333, 334, 335, 336, 337, 338, 339, 340, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.4971875, 0.801875, 1.0, 0.9906458333333333], "iscrowd": 0, "area": 18342.748500000005, "rle": {"size": [480, 640], "counts": "X_e41k>7J5K5J7J5J7M2N2N3M2N2N3L3N2N3M2N2N3M2N2M4M1O000000000000O010000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000000000001O000O101O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O00000000O10000O100O10000O100O10000O100O10000O100O10000O100O01000O100O10000O100O10000O100O10000O100O10000O1001O001O1O001O1O001O001O1O001O001O3M:F9G1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O0O2O1O001O1O1O001O1O001O1O001O1O001O00000000001O0000000000001O00000000001O00000000001O00000000001O00000000001O00001O001O001O001O1O001O001N2O001O001O1O001O001O1O001O001O001O1O00:"}, "label": "blue and black backpack in front of a man on the phone"}]} {"id": 182406, "image": "COCO_train2014_000000182406.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue and black booksack or duffel bag next to a guy on the phone\"."}], "task": "refering", "answer_template": "The \"a blue and black booksack or duffel bag next to a guy on the phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [316, 317, 333, 334, 335, 336, 337, 338, 339, 340, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.4971875, 0.801875, 1.0, 0.9906458333333333], "iscrowd": 0, "area": 18342.748500000005, "rle": {"size": [480, 640], "counts": "X_e41k>7J5K5J7J5J7M2N2N3M2N2N3L3N2N3M2N2N3M2N2M4M1O000000000000O010000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000000000001O000O101O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O00000000O10000O100O10000O100O10000O100O10000O100O10000O100O01000O100O10000O100O10000O100O10000O100O10000O1001O001O1O001O1O001O001O1O001O001O3M:F9G1O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O0O2O1O001O1O1O001O1O001O1O001O1O001O00000000001O0000000000001O00000000001O00000000001O00000000001O00000000001O00001O001O001O001O1O001O001N2O001O001O1O001O001O1O001O001O001O1O00:"}, "label": "a blue and black booksack or duffel bag next to a guy on the phone"}]} {"id": 493707, "image": "COCO_train2014_000000493707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass wirh a juice\"."}], "task": "refering", "answer_template": "The \"a glass wirh a juice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 24, 25, 26, 27, 28, 48, 49, 50, 51, 71, 72, 73, 74, 94, 95, 96], "bbox": [0.0643125, 0.0, 0.23782812499999997, 0.3146009389671362], "iscrowd": 0, "area": 10737.507150000001, "rle": {"size": [426, 640], "counts": "]Ra01Y=3L3N3L4M2N3L4M3L3N3M3L3N3M3L4M2M4M3M2M4M3L4M2N3L4M2M4M3M3L3N3L4M2N3L4M3L3N3M3L3N1N101O0O2O001O1O001O00O100O10000O100O100O10000O100O100O10000O100O010O10000O100O100O10000O100O100O10000O100O100O10000O100O8I8G8I8H8G:G`0_Ob0_O`0_OamZ6"}, "label": "a glass wirh a juice"}]} {"id": 493707, "image": "COCO_train2014_000000493707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow - colored beverage in a glass\"."}], "task": "refering", "answer_template": "The \"a yellow - colored beverage in a glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 24, 25, 26, 27, 28, 48, 49, 50, 51, 71, 72, 73, 74, 94, 95, 96], "bbox": [0.0643125, 0.0, 0.23782812499999997, 0.3146009389671362], "iscrowd": 0, "area": 10737.507150000001, "rle": {"size": [426, 640], "counts": "]Ra01Y=3L3N3L4M2N3L4M3L3N3M3L3N3M3L4M2M4M3M2M4M3L4M2N3L4M2M4M3M3L3N3L4M2N3L4M3L3N3M3L3N1N101O0O2O001O1O001O00O100O10000O100O100O10000O100O100O10000O100O010O10000O100O100O10000O100O100O10000O100O100O10000O100O8I8G8I8H8G:G`0_Ob0_O`0_OamZ6"}, "label": "a yellow - colored beverage in a glass"}]} {"id": 354444, "image": "COCO_train2014_000000354444.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the flying seagull\"."}], "task": "refering", "answer_template": "The \"the flying seagull\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 21, 22, 23, 40, 41, 42, 48, 49, 60, 61, 62, 63, 64, 65, 66, 67, 79, 80, 81, 82, 98, 99, 100], "bbox": [0.2089676113360324, 0.032265625, 0.791255060728745, 0.266671875], "iscrowd": 0, "area": 8747.524150000001, "rle": {"size": [640, 494], "counts": "e\\P21oc07J6J5J7J6J1N10O10O0100O010O10O0100O010O10O0100O010O10O0100O010O10O0100O010O10O0100O010O10O0100O010O10O0100O010O10O0100O010O10O0100O010O10O01O10O0100O010O1O010O10O0100O00100O010O10O01O010O01O010O010O01O01O010O010O01O01O010T]O@Vb0`0j]OAVb0?h]OCWb0=i]ODWb0;h]OGWb0:h]OGWb09h]OIXb07g]OJXb06g]OLYb04e]ONZb03e]ON[b01d]O1[b00d]O1[b00c]O2]b0Nb]O3]b0Na]O4^b0Ma]O4^b0h0O1O3M4L3M4L3M3M3MO1N2N2O1N2N2N2O1N2N2O1N2N2O12N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2NG9G9O10O01O100O1O100O010O1O100O1YO`]ON`b02a]OL_b05c]OH^b07e]OG[b09f]OE[b0;g]OBZb0>h]O@Xb0?k]O^OVb0b0k]O]OTb0d0n]OYOSb0f0d000000O01000O01000O10O1000O01000O10O1000O01000O10O1000O01000O10O1000O01000O10O10O10O1000O10O10O10O1000O10O10O10O0100O10O0100O010O100O01O1N101O0O2O0O2O0O2O001N101N101O1N101N101O0O2O0O2O00WnP2"}, "label": "the flying seagull"}]} {"id": 354444, "image": "COCO_train2014_000000354444.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a seagul flying over seaguls sitting on an old car\"."}], "task": "refering", "answer_template": "The \"a seagul flying over seaguls sitting on an old car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 21, 22, 23, 40, 41, 42, 48, 49, 60, 61, 62, 63, 64, 65, 66, 67, 79, 80, 81, 82, 98, 99, 100], "bbox": [0.2089676113360324, 0.032265625, 0.791255060728745, 0.266671875], "iscrowd": 0, "area": 8747.524150000001, "rle": {"size": [640, 494], "counts": "e\\P21oc07J6J5J7J6J1N10O10O0100O010O10O0100O010O10O0100O010O10O0100O010O10O0100O010O10O0100O010O10O0100O010O10O0100O010O10O0100O010O10O0100O010O10O01O10O0100O010O1O010O10O0100O00100O010O10O01O010O01O010O010O01O01O010O010O01O01O010T]O@Vb0`0j]OAVb0?h]OCWb0=i]ODWb0;h]OGWb0:h]OGWb09h]OIXb07g]OJXb06g]OLYb04e]ONZb03e]ON[b01d]O1[b00d]O1[b00c]O2]b0Nb]O3]b0Na]O4^b0Ma]O4^b0h0O1O3M4L3M4L3M3M3MO1N2N2O1N2N2N2O1N2N2O1N2N2O12N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2NG9G9O10O01O100O1O100O010O1O100O1YO`]ON`b02a]OL_b05c]OH^b07e]OG[b09f]OE[b0;g]OBZb0>h]O@Xb0?k]O^OVb0b0k]O]OTb0d0n]OYOSb0f0d000000O01000O01000O10O1000O01000O10O1000O01000O10O1000O01000O10O1000O01000O10O10O10O1000O10O10O10O1000O10O10O10O0100O10O0100O010O100O01O1N101O0O2O0O2O0O2O001N101N101O1N101N101O0O2O0O2O00WnP2"}, "label": "a seagul flying over seaguls sitting on an old car"}]} {"id": 354444, "image": "COCO_train2014_000000354444.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bird with its wings spread standing on top of the vehicle\"."}], "task": "refering", "answer_template": "The \"a bird with its wings spread standing on top of the vehicle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 74, 75, 76, 77, 91, 92, 93, 94, 95, 110, 111, 112, 127, 128, 129, 130, 145, 146, 147], "bbox": [0.05860323886639676, 0.15371875, 0.3040283400809717, 0.384375], "iscrowd": 0, "area": 8217.516599999999, "rle": {"size": [640, 494], "counts": "oZb04fc06O1000O010000O100000O010000O10000O10000O1`NJV_O6a`04^_OL``06`_OJ_`08j^OZO5>Qa0:e^O^O88Qa0 in this image.", "objects": [{"patches": [58, 59, 74, 75, 76, 77, 91, 92, 93, 94, 95, 110, 111, 112, 127, 128, 129, 130, 145, 146, 147], "bbox": [0.05860323886639676, 0.15371875, 0.3040283400809717, 0.384375], "iscrowd": 0, "area": 8217.516599999999, "rle": {"size": [640, 494], "counts": "oZb04fc06O1000O010000O100000O010000O10000O10000O1`NJV_O6a`04^_OL``06`_OJ_`08j^OZO5>Qa0:e^O^O88Qa0 in this image.", "objects": [{"patches": [99, 100, 101, 106, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269], "bbox": [0.131265625, 0.25329166666666664, 0.7458437499999999, 0.7050208333333334], "iscrowd": 0, "area": 60575.525000000016, "rle": {"size": [480, 640], "counts": "ZdW14f>8H8H8I7L4L4L4M3M3N2M3M3M3N2M1O00100O1O00100O1O100O100O1O1N2L4K5UNSN^FR2a9RNWFS2i9mMRFX2n9hMmE]2U:bMbEf2`:l02N2N2N1O000010BnKlER4T:QLiEP4U:ULgEk3X:`0N1O2O0O2N2O000O10000000000000000000O1001O1O001O1N2O1O001O1O1O1O001O0000001O000000001N100000001O0000001O00O1O1O1O1O1O1O00100O1O1O1O1002N2N2N7I2N2N2M3N3M2N2N2N2N2N0000O10000O10O10O10000O10000O1000000O10O1000O10000O1000000O1000000O2N100O1O100O1O2iNRGYLo8g3\\GnKd8R4fGcK[8]4T1001O01O000001O0_FgK[8Y4ZGTLd8l3QGaLm8l4N1N2N000001O00001O0000001N100O2N100O101N101N3eJlFe4W9XKjFh4X9TKjFk4Y9RKhFn4e9O1O2J5J6K5J60000000000000001N01O1O1O1O2N4L4M3L4M3L3N3M3L4M0O10000O10O010000O100O10000O100O01000O10000O100O1000O0100O10000O100000000O0100000000000O10O100O100O1O10O0100000000000O10000000000_OdJ^G\\5a8fJ^GZ5a8hJ]GY5b8iJ]GW5c8jJ\\GV5d8kJ[GU5e8kJXGX5h8`0001O2N2VJTG\\5m8[J[Ge5P90001O0000000000010O001O1O1O1N10001O00000000001O000O10000TK_FY4b9fKbFU4_9kKeFQ4[9oKjFl3V9TLnFh3R9XLSGb3Q9[LSGa3S9YLRGb3T9XLPGd3Z:K6K4L5K0O10O1O1O001O11O2N2M3N2N3M2M3N2N2N2M2O2N2N2N2M3N2N1O5J6K6J5I7G:F9GUT\\2"}, "label": "3 farm tractors"}]} {"id": 395425, "image": "COCO_train2014_000000395425.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"farm equipment being transported\"."}], "task": "refering", "answer_template": "The \"farm equipment being transported\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 106, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269], "bbox": [0.131265625, 0.25329166666666664, 0.7458437499999999, 0.7050208333333334], "iscrowd": 0, "area": 60575.525000000016, "rle": {"size": [480, 640], "counts": "ZdW14f>8H8H8I7L4L4L4M3M3N2M3M3M3N2M1O00100O1O00100O1O100O100O1O1N2L4K5UNSN^FR2a9RNWFS2i9mMRFX2n9hMmE]2U:bMbEf2`:l02N2N2N1O000010BnKlER4T:QLiEP4U:ULgEk3X:`0N1O2O0O2N2O000O10000000000000000000O1001O1O001O1N2O1O001O1O1O1O001O0000001O000000001N100000001O0000001O00O1O1O1O1O1O1O00100O1O1O1O1002N2N2N7I2N2N2M3N3M2N2N2N2N2N0000O10000O10O10O10000O10000O1000000O10O1000O10000O1000000O1000000O2N100O1O100O1O2iNRGYLo8g3\\GnKd8R4fGcK[8]4T1001O01O000001O0_FgK[8Y4ZGTLd8l3QGaLm8l4N1N2N000001O00001O0000001N100O2N100O101N101N3eJlFe4W9XKjFh4X9TKjFk4Y9RKhFn4e9O1O2J5J6K5J60000000000000001N01O1O1O1O2N4L4M3L4M3L3N3M3L4M0O10000O10O010000O100O10000O100O01000O10000O100O1000O0100O10000O100000000O0100000000000O10O100O100O1O10O0100000000000O10000000000_OdJ^G\\5a8fJ^GZ5a8hJ]GY5b8iJ]GW5c8jJ\\GV5d8kJ[GU5e8kJXGX5h8`0001O2N2VJTG\\5m8[J[Ge5P90001O0000000000010O001O1O1O1N10001O00000000001O000O10000TK_FY4b9fKbFU4_9kKeFQ4[9oKjFl3V9TLnFh3R9XLSGb3Q9[LSGa3S9YLRGb3T9XLPGd3Z:K6K4L5K0O10O1O1O001O11O2N2M3N2N3M2M3N2N2N2M2O2N2N2N2M3N2N1O5J6K6J5I7G:F9GUT\\2"}, "label": "farm equipment being transported"}]} {"id": 362657, "image": "COCO_train2014_000000362657.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left side of the parking meter\"."}], "task": "refering", "answer_template": "The \"the left side of the parking meter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 47, 48, 49, 50, 51, 52, 62, 63, 64, 65, 66, 67, 77, 78, 79, 80, 81, 82, 93, 94, 95, 96, 97, 108, 109, 110, 111, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 153, 154, 155, 156, 157, 168, 169, 170, 171, 183, 184, 185, 186, 198, 199, 200], "bbox": [0.17757075471698114, 0.10515624999999999, 0.5043396226415096, 0.596921875], "iscrowd": 0, "area": 30769.272899999996, "rle": {"size": [640, 424], "counts": "ZP_16[c0c0]Oc0^Ob0K3N1N3M3M2N3N1N3M2N3M3N1N3M2N3N1N3oAYMd:j2SElM]:V2[EUNa:m1WE\\Ne:f1SEbNk:a1mDgNQ;Z1gDnNW;U1aDRO];P1[DYOb;j0VD]Oi;e0oCCn;?PDCn;`0oCBP in this image.", "objects": [{"patches": [40, 54, 55, 56, 68, 69, 70, 71, 72, 83, 84, 85, 86, 98, 99, 100, 101, 113, 114, 115, 116, 128, 129, 130, 131, 143, 144, 145, 146, 158, 159, 160, 161, 174, 175, 176, 189, 190, 191, 204, 205], "bbox": [0.5596698113207548, 0.12321875, 0.810683962264151, 0.592890625], "iscrowd": 0, "area": 22129.645549999997, "rle": {"size": [640, 424], "counts": "XXd41mc05L5J6e_OAk in this image.", "objects": [{"patches": [40, 54, 55, 56, 68, 69, 70, 71, 72, 83, 84, 85, 86, 98, 99, 100, 101, 113, 114, 115, 116, 128, 129, 130, 131, 143, 144, 145, 146, 158, 159, 160, 161, 174, 175, 176, 189, 190, 191, 204, 205], "bbox": [0.5596698113207548, 0.12321875, 0.810683962264151, 0.592890625], "iscrowd": 0, "area": 22129.645549999997, "rle": {"size": [640, 424], "counts": "XXd41mc05L5J6e_OAk in this image.", "objects": [{"patches": [188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0016875000000000002, 0.5066041666666666, 1.0, 0.9864375], "iscrowd": 0, "area": 141245.63809999998, "rle": {"size": [480, 640], "counts": "Wg0b6^80O10000O100O100O100O13M00000000000000000000O100000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O00000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001YOf0fMZ2gMnI"}, "label": "a toasted sandwich is shown on a plate in an outdoor dining area"}]} {"id": 92323, "image": "COCO_train2014_000000092323.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich on a tabletop\"."}], "task": "refering", "answer_template": "The \"a sandwich on a tabletop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0016875000000000002, 0.5066041666666666, 1.0, 0.9864375], "iscrowd": 0, "area": 141245.63809999998, "rle": {"size": [480, 640], "counts": "Wg0b6^80O10000O100O100O100O13M00000000000000000000O100000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O00000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001YOf0fMZ2gMnI"}, "label": "a sandwich on a tabletop"}]} {"id": 395432, "image": "COCO_train2014_000000395432.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small giraffe on the right\"."}], "task": "refering", "answer_template": "The \"small giraffe on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 153, 154, 155, 156, 175, 176, 177, 196, 197, 198, 199, 219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 268, 287, 289, 290, 291, 309, 310, 313, 333, 336, 337], "bbox": [0.458953125, 0.325784543325527, 0.814078125, 0.9914754098360656], "iscrowd": 0, "area": 13668.297100000003, "rle": {"size": [427, 640], "counts": "Zoj33U=5L3L5L4K5L0O2O0O2O0O10O1N101N2O0O2O2N14aNZOTFi0i9\\OQFQ1c9TOWF]1Y9iNaFf1P9_NjFm1k8[1N1001QNfFLZ92ZG[Og8b0nGkNT8R1VHeNk7X1YHfNh7W1[HhNf7V1\\HiNe7T1^HkNc7S1`HkNa7R1bHmN`7P1dHmN]7Q1a1PObEo0Z:XOdEf0Z:^OfE`0Z:BfE<[:EfE8[:IeE5\\:LdE3\\:MeE1\\:0dEN\\:4dEJ[:9eEE[:=eEA[:a0eE^OZ:d0fEZOZ:h0jEROV:P1P13N1O2N2N2N1N3N2L4K4PEXNY:_2O1O010O1O00100O001O010O1O0010O01O010O10O010O010O01lE^Mg1Ka5g2aHWNU1YOY6`2]HaNo0VOd6i3XIWLg6o3TIPLl6U4PIjKQ7Z4kHeKU7a4fH^K[7f4`HZK`7i4\\HXKe7f4\\HZKd7f4\\H[Kd7d4\\H\\Kd7k2WIbMUOJ]7e2kISMgN`0W7_2_JdM^5]2aJfM\\5\\2bJgM[5[2cJmLaN8k6T2UIiM]1KiN5c6P2`IPNR1KoN4\\6P2fIPNn0KUO3U6Q2lIPNh0L\\OOo5U2oIPNd0K@1k5S2TJQN>LEOh5]1`IjNg0M;LIOd5]1eIgNf016LLOb5[1kIfNd031L1O_5Z1oIdNc07KM6M\\5Y1UJbNa0;EM;MY5X1YJaN`0LV5X1hKnNaNOc0JS5Y1lKmN\\N0f0JQ5X1PLnNVN0l0Im4Y1SLmNSN1o0Hj4Z1VLmNnM2S1Gh4Z1ZLlNhM4Y1Ed4Z1]LoNbM2_1Da4[1`LROZM0h1A]4]1cLBX5>jJ@W5?kJ_OU5a0mJ]OS5c0oJ[OQ5e0QKXOQ5g0QKWOo4i0SKUOm4k0UKSOk4m0WKQOi4o0YKoNh4o0`KjN`4V1j30001O000O100O100O100O100O100O1O100O100O010OCPObDP1_;RO^Dm0d;TOZDk0g;XOWDg0j;[OSDd0o;]OoCc0Q<:2O1N2O002N101N101N2N101N2O0OH[OkCf0U<\\OhCd0X<^OgCb0Y<8O2N20O100O1O100O10O01O100O1O101N3M2O2M1O1O100O001O100O001O1O1O002N3M3M3MW^a1"}, "label": "small giraffe on the right"}]} {"id": 395432, "image": "COCO_train2014_000000395432.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is a shorter giraffe shown in the picture\"."}], "task": "refering", "answer_template": "The \"there is a shorter giraffe shown in the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 153, 154, 155, 156, 175, 176, 177, 196, 197, 198, 199, 219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 268, 287, 289, 290, 291, 309, 310, 313, 333, 336, 337], "bbox": [0.458953125, 0.325784543325527, 0.814078125, 0.9914754098360656], "iscrowd": 0, "area": 13668.297100000003, "rle": {"size": [427, 640], "counts": "Zoj33U=5L3L5L4K5L0O2O0O2O0O10O1N101N2O0O2O2N14aNZOTFi0i9\\OQFQ1c9TOWF]1Y9iNaFf1P9_NjFm1k8[1N1001QNfFLZ92ZG[Og8b0nGkNT8R1VHeNk7X1YHfNh7W1[HhNf7V1\\HiNe7T1^HkNc7S1`HkNa7R1bHmN`7P1dHmN]7Q1a1PObEo0Z:XOdEf0Z:^OfE`0Z:BfE<[:EfE8[:IeE5\\:LdE3\\:MeE1\\:0dEN\\:4dEJ[:9eEE[:=eEA[:a0eE^OZ:d0fEZOZ:h0jEROV:P1P13N1O2N2N2N1N3N2L4K4PEXNY:_2O1O010O1O00100O001O010O1O0010O01O010O10O010O010O01lE^Mg1Ka5g2aHWNU1YOY6`2]HaNo0VOd6i3XIWLg6o3TIPLl6U4PIjKQ7Z4kHeKU7a4fH^K[7f4`HZK`7i4\\HXKe7f4\\HZKd7f4\\H[Kd7d4\\H\\Kd7k2WIbMUOJ]7e2kISMgN`0W7_2_JdM^5]2aJfM\\5\\2bJgM[5[2cJmLaN8k6T2UIiM]1KiN5c6P2`IPNR1KoN4\\6P2fIPNn0KUO3U6Q2lIPNh0L\\OOo5U2oIPNd0K@1k5S2TJQN>LEOh5]1`IjNg0M;LIOd5]1eIgNf016LLOb5[1kIfNd031L1O_5Z1oIdNc07KM6M\\5Y1UJbNa0;EM;MY5X1YJaN`0LV5X1hKnNaNOc0JS5Y1lKmN\\N0f0JQ5X1PLnNVN0l0Im4Y1SLmNSN1o0Hj4Z1VLmNnM2S1Gh4Z1ZLlNhM4Y1Ed4Z1]LoNbM2_1Da4[1`LROZM0h1A]4]1cLBX5>jJ@W5?kJ_OU5a0mJ]OS5c0oJ[OQ5e0QKXOQ5g0QKWOo4i0SKUOm4k0UKSOk4m0WKQOi4o0YKoNh4o0`KjN`4V1j30001O000O100O100O100O100O100O1O100O100O010OCPObDP1_;RO^Dm0d;TOZDk0g;XOWDg0j;[OSDd0o;]OoCc0Q<:2O1N2O002N101N101N2N101N2O0OH[OkCf0U<\\OhCd0X<^OgCb0Y<8O2N20O100O1O100O10O01O100O1O101N3M2O2M1O1O100O001O100O001O1O1O002N3M3M3MW^a1"}, "label": "there is a shorter giraffe shown in the picture"}]} {"id": 395432, "image": "COCO_train2014_000000395432.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"male giraffe alongside another giraffe\"."}], "task": "refering", "answer_template": "The \"male giraffe alongside another giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 81, 82, 83, 84, 104, 105, 126, 127, 148, 149, 150, 170, 171, 172, 173, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 236, 237, 238, 239, 240, 241, 242, 259, 260, 261, 262, 265, 283, 284, 285, 288, 289, 306, 307, 308, 309, 311, 312, 329, 330, 332], "bbox": [0.27790625, 0.15723653395784543, 0.6731875, 0.982576112412178], "iscrowd": 0, "area": 22373.56980000001, "rle": {"size": [427, 640], "counts": "SaZ23V=3M3L4M3L4M3L4M1N2O1O0O2O10000000O100000000O100000000O100000\\GROl4n0TKSOk4m0TKUOk4k0SKXOl4i0PK[Oo4e0nJ_OP5b0nJAQ5?lHQOf0d0]6;eHD`05j67_H78FX73XHj02WOe7OQH^1KgNS8n3O1O1O00O1N2O1O1O1N2O1G9^Ob0^Ob0_Oa0M3O1O1O2O0O2N1O2N1OnEdMX9[2YFXNe9h2N2O2M2N3N1N3N1N2N3N1N3N1N2N3N1N3N1N3M2O1N[NdG_NZ8`1mG[NQ8e1WHVNf7j1bHPN]7n1jHoMS7Q2PInMn6R2UImMh6T2[IlMa6U2cIiMZ6X2iIhMS6Y2PJfMm5\\2UJcMh5_2ZJaMb5a2`J^M]5d2eJ\\MW5f2kJYMR5i2PKWMl4k2WKSMg4n2[KQMb4Q3`KoL\\4S3fKlLW4U3lKiLS4n2_IjL`26o3Q3bIiLa23m3T3dIhLa21k3X3dIgLc2Oh3Z3gIfLd2Le3^3iIeLd2Kb3a3jIdLR9\\3PGcLo8]3RGcLm8^3TGaLk8_3VGaLi8_3YG`Le8b3[G^Ld8b3^G]La8c3`G]L_8d3bG[L]8f3c01O1O1O1O1O2N1N2O2N2N2N2N2N2N1O2N2N2M3N2N2N2N2N2N1O]MlG2R8MSH1k7NYH1e7MaHSO_O^On7^1gHoN_OCg7^1nHjN_OHa7]1UIdN@NY7\\1\\IaN_O3S7[1bI]N_O7n6[1hIXN^O=h6Y1oITN^Oc0a6X1UJPN^Oh0[6W1aKiN]4V1fKjNX4T1lKlNR4S1QLmNm3R1VLnNg3R1\\LnN^N[OP3e1fNoNQNFW3Z1kN_OUM@n3Q1POS1o0kNUOU1i0jNZOU1d0kN_OT1`0kNCU1:jNJU15jNNV10iN3V1KjN9T1FjNh0l0VOSOV1c0hN]Oe19YNET21jMOc2F\\M9P3^OnLa0_3TO_Lk0n3lN]KLWMW1n:mNmDR1U;ROeDm0^;WO]Dh0e;=3M2M3N2N2N2M3N3M2N2001O000O1F:1O00100O001O001O10O01O001O00100O001O001O010O1O001O001O011N2N2N2N1O2N2N2N2N2N2N2N2N2N2N[Qg2"}, "label": "male giraffe alongside another giraffe"}]} {"id": 395432, "image": "COCO_train2014_000000395432.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a larger giraffe climbing up behind a smaller giraffe\"."}], "task": "refering", "answer_template": "The \"a larger giraffe climbing up behind a smaller giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 81, 82, 83, 84, 104, 105, 126, 127, 148, 149, 150, 170, 171, 172, 173, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219, 236, 237, 238, 239, 240, 241, 242, 259, 260, 261, 262, 265, 283, 284, 285, 288, 289, 306, 307, 308, 309, 311, 312, 329, 330, 332], "bbox": [0.27790625, 0.15723653395784543, 0.6731875, 0.982576112412178], "iscrowd": 0, "area": 22373.56980000001, "rle": {"size": [427, 640], "counts": "SaZ23V=3M3L4M3L4M3L4M1N2O1O0O2O10000000O100000000O100000000O100000\\GROl4n0TKSOk4m0TKUOk4k0SKXOl4i0PK[Oo4e0nJ_OP5b0nJAQ5?lHQOf0d0]6;eHD`05j67_H78FX73XHj02WOe7OQH^1KgNS8n3O1O1O00O1N2O1O1O1N2O1G9^Ob0^Ob0_Oa0M3O1O1O2O0O2N1O2N1OnEdMX9[2YFXNe9h2N2O2M2N3N1N3N1N2N3N1N3N1N2N3N1N3N1N3M2O1N[NdG_NZ8`1mG[NQ8e1WHVNf7j1bHPN]7n1jHoMS7Q2PInMn6R2UImMh6T2[IlMa6U2cIiMZ6X2iIhMS6Y2PJfMm5\\2UJcMh5_2ZJaMb5a2`J^M]5d2eJ\\MW5f2kJYMR5i2PKWMl4k2WKSMg4n2[KQMb4Q3`KoL\\4S3fKlLW4U3lKiLS4n2_IjL`26o3Q3bIiLa23m3T3dIhLa21k3X3dIgLc2Oh3Z3gIfLd2Le3^3iIeLd2Kb3a3jIdLR9\\3PGcLo8]3RGcLm8^3TGaLk8_3VGaLi8_3YG`Le8b3[G^Ld8b3^G]La8c3`G]L_8d3bG[L]8f3c01O1O1O1O1O2N1N2O2N2N2N2N2N2N1O2N2N2M3N2N2N2N2N2N1O]MlG2R8MSH1k7NYH1e7MaHSO_O^On7^1gHoN_OCg7^1nHjN_OHa7]1UIdN@NY7\\1\\IaN_O3S7[1bI]N_O7n6[1hIXN^O=h6Y1oITN^Oc0a6X1UJPN^Oh0[6W1aKiN]4V1fKjNX4T1lKlNR4S1QLmNm3R1VLnNg3R1\\LnN^N[OP3e1fNoNQNFW3Z1kN_OUM@n3Q1POS1o0kNUOU1i0jNZOU1d0kN_OT1`0kNCU1:jNJU15jNNV10iN3V1KjN9T1FjNh0l0VOSOV1c0hN]Oe19YNET21jMOc2F\\M9P3^OnLa0_3TO_Lk0n3lN]KLWMW1n:mNmDR1U;ROeDm0^;WO]Dh0e;=3M2M3N2N2N2M3N3M2N2001O000O1F:1O00100O001O001O10O01O001O00100O001O001O010O1O001O001O011N2N2N2N1O2N2N2N2N2N2N2N2N2N2N[Qg2"}, "label": "a larger giraffe climbing up behind a smaller giraffe"}]} {"id": 493760, "image": "COCO_train2014_000000493760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark brown horse with a green leg wrap\"."}], "task": "refering", "answer_template": "The \"a dark brown horse with a green leg wrap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 210, 211, 212, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 278, 279, 282, 283, 301, 302, 306, 324, 325, 329], "bbox": [0.097296875, 0.46021653543307084, 0.43187499999999995, 0.8172047244094488], "iscrowd": 0, "area": 13741.5209, "rle": {"size": [508, 640], "counts": "STo07\\?e0B5K5L5J5K5K3M2\\NbN]DE@j1l;CSD?h;DXD>f;BZD`0d;A[D^OH2k;`0^D_O2F^;k0`DB in this image.", "objects": [{"patches": [192, 193, 210, 211, 212, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 239, 255, 256, 257, 258, 259, 260, 278, 279, 282, 283, 301, 302, 306, 324, 325, 329], "bbox": [0.097296875, 0.46021653543307084, 0.43187499999999995, 0.8172047244094488], "iscrowd": 0, "area": 13741.5209, "rle": {"size": [508, 640], "counts": "STo07\\?e0B5K5L5J5K5K3M2\\NbN]DE@j1l;CSD?h;DXD>f;BZD`0d;A[D^OH2k;`0^D_O2F^;k0`DB in this image.", "objects": [{"patches": [218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 246, 262, 263, 264, 265, 266, 267, 268, 269, 285, 286, 287, 288, 289, 290, 291, 313], "bbox": [0.40393749999999995, 0.5196456692913386, 0.713703125, 0.7297834645669292], "iscrowd": 0, "area": 13378.114900000006, "rle": {"size": [508, 640], "counts": "UjP4:a?1O100O1O1O1O1N2O1N2N2M3M3O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1003M3M2N1O1O001O1O1O1O001O1O001O1O00O100_NPOVDP1j;QOUDo0k;SORDn0n;_OeCa0[ in this image.", "objects": [{"patches": [218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 246, 262, 263, 264, 265, 266, 267, 268, 269, 285, 286, 287, 288, 289, 290, 291, 313], "bbox": [0.40393749999999995, 0.5196456692913386, 0.713703125, 0.7297834645669292], "iscrowd": 0, "area": 13378.114900000006, "rle": {"size": [508, 640], "counts": "UjP4:a?1O100O1O1O1O1N2O1N2N2M3M3O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1003M3M2N1O1O001O1O1O1O001O1O001O1O00O100_NPOVDP1j;QOUDo0k;SORDn0n;_OeCa0[ in this image.", "objects": [{"patches": [110, 132, 133, 134, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 200], "bbox": [0.642140625, 0.3443888888888889, 0.926546875, 0.6320277777777777], "iscrowd": 0, "area": 8355.37785, "rle": {"size": [360, 640], "counts": "Pf`41V;2N101O1O0O2O1O000O101O0O10001O0O10001N100O2O1N2O1O1N2O1N2O3L3N2N2M3N0O100O10000O100O100O1000O010O010O1O010O010O010O10O010N101O001O0O2O001O001O0O101O001O0ON3L4M3L4N2O0O2N2O1N2N101N2O15K6J5J6K6J5K5K6I6K5K4L4L4L4NO1O1O100O1O1O1O001O1O1O1O1O1N2O1O1O1O1O1O0O2O1O1O1O1O1O1O1N2O1O1O1O001O1N2O1O1O1O1O1O1N2O1O1O001O1O1O1N2O1O1O1O1O1O1N2O001N2M3M3M3N2M3M3N2M3M3N1N2O2N1OiZ`0"}, "label": "gray naval boat parked on far right"}]} {"id": 84162, "image": "COCO_train2014_000000084162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the grey ship furthest right\"."}], "task": "refering", "answer_template": "The \"the grey ship furthest right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 132, 133, 134, 153, 154, 155, 156, 157, 158, 159, 175, 176, 177, 178, 179, 180, 181, 182, 200], "bbox": [0.642140625, 0.3443888888888889, 0.926546875, 0.6320277777777777], "iscrowd": 0, "area": 8355.37785, "rle": {"size": [360, 640], "counts": "Pf`41V;2N101O1O0O2O1O000O101O0O10001O0O10001N100O2O1N2O1O1N2O1N2O3L3N2N2M3N0O100O10000O100O100O1000O010O010O1O010O010O010O10O010N101O001O0O2O001O001O0O101O001O0ON3L4M3L4N2O0O2N2O1N2N101N2O15K6J5J6K6J5K5K6I6K5K4L4L4L4NO1O1O100O1O1O1O001O1O1O1O1O1N2O1O1O1O1O1O0O2O1O1O1O1O1O1O1N2O1O1O1O001O1N2O1O1O1O1O1O1N2O1O1O001O1O1O1N2O1O1O1O1O1O1N2O001N2M3M3M3N2M3M3N2M3M3N1N2O2N1OiZ`0"}, "label": "the grey ship furthest right"}]} {"id": 84162, "image": "COCO_train2014_000000084162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this picture shows a ferry boat transporting people\"."}], "task": "refering", "answer_template": "The \"this picture shows a ferry boat transporting people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 177, 178, 179, 181, 182, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275], "bbox": [0.602953125, 0.4494444444444445, 0.9973281249999999, 0.9123611111111113], "iscrowd": 0, "area": 21388.428149999996, "rle": {"size": [360, 640], "counts": "doW42V;2M4M3M2N3M2N3L4M2N3M2N3L2O000000000O100000000O100000000FlN`FT1`9:O100O1N2N2M3N2N2O10000UORN[Ho1e7QN[Ho1e7RNZHn1f7SNYHm1g7TNXHl1h7UNWHk1i7VNVHj1j7VNUHk1j7WNjGT2V8?00000000000000000000O10000000000000000000000O10000000000N2L4L4L4L4L40000000001O000000000000001O000000000000000O1H8G9H8G9H7M48H8H7I8H5K2N3L3N2N3M2N3M2N2M4M2N2N3PMgGk2[8000000O100000000000000O1000000000000O100000000000001N100000000000000O1000000000000O1000O1000000000O10O100000000000O10O1B>0000O1000O100000001O2N1N3N1O2N1O1O00000O1000000000O1000O1000000000000O01000000001O2N2M2O2N2N2N1O1OO1O001O1O1005Kb0]Od0]Oc0\\OZ="}, "label": "this picture shows a ferry boat transporting people"}]} {"id": 84162, "image": "COCO_train2014_000000084162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white tourist ship\"."}], "task": "refering", "answer_template": "The \"a white tourist ship\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 177, 178, 179, 181, 182, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275], "bbox": [0.602953125, 0.4494444444444445, 0.9973281249999999, 0.9123611111111113], "iscrowd": 0, "area": 21388.428149999996, "rle": {"size": [360, 640], "counts": "doW42V;2M4M3M2N3M2N3L4M2N3M2N3L2O000000000O100000000O100000000FlN`FT1`9:O100O1N2N2M3N2N2O10000UORN[Ho1e7QN[Ho1e7RNZHn1f7SNYHm1g7TNXHl1h7UNWHk1i7VNVHj1j7VNUHk1j7WNjGT2V8?00000000000000000000O10000000000000000000000O10000000000N2L4L4L4L4L40000000001O000000000000001O000000000000000O1H8G9H8G9H7M48H8H7I8H5K2N3L3N2N3M2N3M2N2M4M2N2N3PMgGk2[8000000O100000000000000O1000000000000O100000000000001N100000000000000O1000000000000O1000O1000000000O10O100000000000O10O1B>0000O1000O100000001O2N1N3N1O2N1O1O00000O1000000000O1000O1000000000000O01000000001O2N2M2O2N2N2N1O1OO1O001O1O1005Kb0]Od0]Oc0\\OZ="}, "label": "a white tourist ship"}]} {"id": 84162, "image": "COCO_train2014_000000084162.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a decorated passenger ship is waiting for departure in a port\"."}], "task": "refering", "answer_template": "The \"a decorated passenger ship is waiting for departure in a port\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 55, 56, 57, 77, 78, 79, 80, 81, 98, 99, 100, 101, 102, 103, 104, 120, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198], "bbox": [0.1288125, 0.08880555555555555, 0.64521875, 0.6853888888888889], "iscrowd": 0, "area": 35668.1759, "rle": {"size": [360, 640], "counts": "\\\\m03T;5L4K4L1O2O0O1O101N3M2N3N2M2N3M3N1N3M3M3N0O1O1O100O1O100O1O1O100O1O1O100O1O1O010O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O00100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O001O100O1O1O100O1O1O010O1O100O1O1O100O1O00100O1O1O100O1O100O001O100O1O1O100O1O1O010O1O1O100O1O100O1O00100O1O1O100O1O1O100O001O100O1O100SOVK[Kk4e4l0000O3N1O2N1O2N1O2N1O2N2M2O2N1O2N1O2N1O2N1O2M3N1O2N1O2N1O2N1O2N1O2N2M2O2N1O2N1O2N1O2N1O2N2N1N3N1O2N1O2N1O2N1O2N2N1O2M2O2N1O2N1O2N1O2N2N1O2N1O2M2O2N1O2N1O2N2N1O2N1O2N1N3N1O2N1O2N2N1O2N1O2N1O2M2O2N1O2N2N1O00000000000O001L4M3M3M3M3L4M3M3M3M3L4M2NTd_2"}, "label": "a decorated passenger ship is waiting for departure in a port"}]} {"id": 387293, "image": "COCO_train2014_000000387293.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"smaller giraffe looking larger giraffe\"."}], "task": "refering", "answer_template": "The \"smaller giraffe looking larger giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 35, 36, 58, 59, 81, 104, 126, 127, 149, 150, 171, 172, 173, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 285, 287, 288], "bbox": [0.38803125, 0.03800469483568075, 0.5954843750000001, 0.8653521126760563], "iscrowd": 0, "area": 14363.219600000008, "rle": {"size": [426, 640], "counts": "Y`W31V=5K5K6J4L2N2N2O1N2O1N10000O100O10000O2O0O10000O10000O100O10000O1O100O1O1O1O1O1O1O1O1O1O100aEoNf8R1[GmNd8T1\\GmNb8T1^GmN`8T1`GmN^8T1bGlN]8U1dGkNZ8V1fGkNX8V1hGjNW8W1iGjNW8U1iGlNV8T1kGlNU8U1iGkNM]OS7k1lHiNKCW7i1iHZOVOROQ8g1fHLY77cHK\\78`HJ^7:^HGb7<[HEd7>XHDg7?UHCi7o20K5^Oa0\\Oe000h0WOi0XO;D5K3N1N101N101N2\\LfI6[6JgI4Z6KhI2Y6NiI0c5d0^JXOQ5^1nJZNf4[2YK]M_4S3aKeLb4a3^KVLf4`0TKj05^Nj4j0TKi01UNn4S1UKh0MmMR5\\1TKg0KcMU5h1SKd0I\\MY5o1RKd0FUM\\5W2RKd0BmLa5^2PK1ZOhM6Ae5e2oJN^OgMl5Z2jJLAeMc5^2PKO_O^Mg5Z2nJ;k5\\OXJf0l5POWJS1m5cNRJc1S6SNlIT2a1jLe1i0iLd2P1^MR2DnLT3>RN`2`NSMb3LiNX4Z1]K[OV4h0jKCh3`0XL4Q3@WLbLh0d4X2lNWNd1W1_NiNo1f0SN[O\\23jMJc2FdM6i2ZO^Mb0o2nN\\Mj0e7L2O2N2N1O2N2N2M2O2N2NWf[3"}, "label": "smaller giraffe looking larger giraffe"}]} {"id": 387293, "image": "COCO_train2014_000000387293.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the shorter of the fighting giraffes\"."}], "task": "refering", "answer_template": "The \"the shorter of the fighting giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 35, 36, 58, 59, 81, 104, 126, 127, 149, 150, 171, 172, 173, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 285, 287, 288], "bbox": [0.38803125, 0.03800469483568075, 0.5954843750000001, 0.8653521126760563], "iscrowd": 0, "area": 14363.219600000008, "rle": {"size": [426, 640], "counts": "Y`W31V=5K5K6J4L2N2N2O1N2O1N10000O100O10000O2O0O10000O10000O100O10000O1O100O1O1O1O1O1O1O1O1O1O100aEoNf8R1[GmNd8T1\\GmNb8T1^GmN`8T1`GmN^8T1bGlN]8U1dGkNZ8V1fGkNX8V1hGjNW8W1iGjNW8U1iGlNV8T1kGlNU8U1iGkNM]OS7k1lHiNKCW7i1iHZOVOROQ8g1fHLY77cHK\\78`HJ^7:^HGb7<[HEd7>XHDg7?UHCi7o20K5^Oa0\\Oe000h0WOi0XO;D5K3N1N101N101N2\\LfI6[6JgI4Z6KhI2Y6NiI0c5d0^JXOQ5^1nJZNf4[2YK]M_4S3aKeLb4a3^KVLf4`0TKj05^Nj4j0TKi01UNn4S1UKh0MmMR5\\1TKg0KcMU5h1SKd0I\\MY5o1RKd0FUM\\5W2RKd0BmLa5^2PK1ZOhM6Ae5e2oJN^OgMl5Z2jJLAeMc5^2PKO_O^Mg5Z2nJ;k5\\OXJf0l5POWJS1m5cNRJc1S6SNlIT2a1jLe1i0iLd2P1^MR2DnLT3>RN`2`NSMb3LiNX4Z1]K[OV4h0jKCh3`0XL4Q3@WLbLh0d4X2lNWNd1W1_NiNo1f0SN[O\\23jMJc2FdM6i2ZO^Mb0o2nN\\Mj0e7L2O2N2N1O2N2N2M2O2N2NWf[3"}, "label": "the shorter of the fighting giraffes"}]} {"id": 223459, "image": "COCO_train2014_000000223459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the number one jet in the blue angles\"."}], "task": "refering", "answer_template": "The \"the number one jet in the blue angles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 245, 246, 247, 268, 269, 270, 292, 293, 294, 316, 317, 318, 341], "bbox": [0.618765625, 0.48912213740458016, 0.8544218749999999, 0.771145038167939], "iscrowd": 0, "area": 4600.91465, "rle": {"size": [524, 640], "counts": "bm[61\\`00\\`00d_O000O01000O1000O010000O10O10O2O1O0O2O1O0O2O1N2O006I100O10000O1M4L2M4M3L4M3M3L4M3L3N3L4M3O10000O10O10000000j0VO2NN1O2N2N2N2N2N2N1O2N2N2N2N2N2N1010oNfAOY>NkA2T>KoA6Q>FSB:l=CXB00O010O1000O0100O10O1000000O10000O1000000O1000O10O1000000O1000000O1000000O10O1000O010002N2M3N3M2N2M4M000O2O00001N101O1N3N1O2N1N3N1OXV_1"}, "label": "the number one jet in the blue angles"}]} {"id": 223459, "image": "COCO_train2014_000000223459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a jet third from the right in a group of identical jets\"."}], "task": "refering", "answer_template": "The \"a jet third from the right in a group of identical jets\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [264, 266, 288, 289, 290, 311, 312, 313, 335, 336, 337, 338, 360, 361, 383, 384], "bbox": [0.49917187500000004, 0.6037977099236641, 0.7225781250000001, 0.8670992366412213], "iscrowd": 0, "area": 4805.407900000002, "rle": {"size": [524, 640], "counts": "_bT51[`0100O10000O10000O100O10000O100O2O000O10000O100O100O1O1N2O1N2O2N10000000000001OM3J6J6J6J7I6O100O101O0O100003M=D:EN2N2O1N2N2N3M2N2N2O1N2N2N1O1O10O01O1O1O2N2000lNgA4Y>IjA7V>FmA:S>BPB?P>^OSBb0m=[OVBe0j=XOYBh0g=UO\\Bk0d=QO`Bo0^>2O000000000O10000000000O10001O00001N1O1O1O2N1O1O1010O0001O0000010O00O1O100O2N1O100TORA8o>DVA:k>CYA;i>A[A=]?N1N2N2N2O1N2N2Nddj2"}, "label": "a jet third from the right in a group of identical jets"}]} {"id": 223459, "image": "COCO_train2014_000000223459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plane to far left\"."}], "task": "refering", "answer_template": "The \"plane to far left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [264, 286, 287, 288, 309, 310, 311, 312, 333, 334, 335, 357, 358, 359, 381, 382, 405], "bbox": [0.44476562499999994, 0.6195419847328244, 0.6481874999999999, 0.9093129770992366], "iscrowd": 0, "area": 4845.500650000001, "rle": {"size": [524, 640], "counts": "gUb43Y`04M2M2O1N101N2O1O0O2O1OO0100O1000O0100O1O10O01O100O10O01O100O1O0kNVOhBk0W=WOgBj0X=WOgBj0X=XOgBh0Y=XOgBg0Z=YOeBh0[=YOdBg0\\=YOdBg0\\=ZO_Bj0a=VOXBQ1h=oNPBX1Q>b0000000000000000000000ZNhAW1X>hNRBo0S>lNUBl0c>3N2N2N2N2N2N101N1O2N1O1O2N1010O10000O10000lN\\A?d>\\OaAd0_>WOfAi0Z>SOjAm0V>nNoAS1e>00000000001O000001O00000000000001O0002N00O1O1N2N2M3N2O0011O0000001O0oNTAa0l>^OZA=f>C\\A;e>C]AD_A9c>F_A8a>H`A7a>HaA6`>IbA5^>KcA4^>JeA1_ec3"}, "label": "plane to far left"}]} {"id": 223459, "image": "COCO_train2014_000000223459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the jet that is under all the other jets\"."}], "task": "refering", "answer_template": "The \"the jet that is under all the other jets\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [264, 286, 287, 288, 309, 310, 311, 312, 333, 334, 335, 357, 358, 359, 381, 382, 405], "bbox": [0.44476562499999994, 0.6195419847328244, 0.6481874999999999, 0.9093129770992366], "iscrowd": 0, "area": 4845.500650000001, "rle": {"size": [524, 640], "counts": "gUb43Y`04M2M2O1N101N2O1O0O2O1OO0100O1000O0100O1O10O01O100O10O01O100O1O0kNVOhBk0W=WOgBj0X=WOgBj0X=XOgBh0Y=XOgBg0Z=YOeBh0[=YOdBg0\\=YOdBg0\\=ZO_Bj0a=VOXBQ1h=oNPBX1Q>b0000000000000000000000ZNhAW1X>hNRBo0S>lNUBl0c>3N2N2N2N2N2N101N1O2N1O1O2N1010O10000O10000lN\\A?d>\\OaAd0_>WOfAi0Z>SOjAm0V>nNoAS1e>00000000001O000001O00000000000001O0002N00O1O1N2N2M3N2O0011O0000001O0oNTAa0l>^OZA=f>C\\A;e>C]AD_A9c>F_A8a>H`A7a>HaA6`>IbA5^>KcA4^>JeA1_ec3"}, "label": "the jet that is under all the other jets"}]} {"id": 223459, "image": "COCO_train2014_000000223459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"four blue jets flying close to each other\"."}], "task": "refering", "answer_template": "The \"four blue jets flying close to each other\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [244, 245, 266, 267, 268, 290, 291, 292, 314, 315, 316, 338, 339, 362], "bbox": [0.576265625, 0.5505725190839694, 0.7805, 0.8134923664122137], "iscrowd": 0, "area": 4162.355950000002, "rle": {"size": [524, 640], "counts": "Xdm52Y`02O1O1O1O1O1O1O1O10000O100O10001N10O010000O100O1000O0100O101O0L4L5L3L4L4L5L3L4M3M4L3O1O2N1O1O1O10O3N2N3Ld0]O7IO010O10000O100O1000001O000O10000000001O0000000O1000001O0000000000O10000O1000000O10000O10000O10001O0O10000O1000001N2O1O1N2O1O1O1N2O1O1N2O1O1O1N3N4L3LggW2"}, "label": "four blue jets flying close to each other"}]} {"id": 223459, "image": "COCO_train2014_000000223459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a second plane which is not forming foam\"."}], "task": "refering", "answer_template": "The \"a second plane which is not forming foam\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [244, 245, 266, 267, 268, 290, 291, 292, 314, 315, 316, 338, 339, 362], "bbox": [0.576265625, 0.5505725190839694, 0.7805, 0.8134923664122137], "iscrowd": 0, "area": 4162.355950000002, "rle": {"size": [524, 640], "counts": "Xdm52Y`02O1O1O1O1O1O1O1O10000O100O10001N10O010000O100O1000O0100O101O0L4L5L3L4L4L5L3L4M3M4L3O1O2N1O1O1O10O3N2N3Ld0]O7IO010O10000O100O1000001O000O10000000001O0000000O1000001O0000000000O10000O1000000O10000O10000O10001O0O10000O1000001N2O1O1N2O1O1O1N2O1O1N2O1O1O1N3N4L3LggW2"}, "label": "a second plane which is not forming foam"}]} {"id": 559331, "image": "COCO_train2014_000000559331.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboarder doing a trick\"."}], "task": "refering", "answer_template": "The \"a skateboarder doing a trick\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 19, 20, 21, 22, 23, 33, 34, 35, 46, 47, 48, 57, 58, 59, 60, 61, 62, 70, 71, 72, 73, 74, 75, 85, 86, 87], "bbox": [0.4289487870619946, 0.03048, 0.8001078167115903, 0.38488], "iscrowd": 0, "area": 12144.838549999999, "rle": {"size": [500, 371], "counts": "dh]22a?8I6I00100O010O10O01O10O0100O00100O10O010O01000O010O010O10O10O010OfME\\EY;DeD>Y;CfD?Y;BeD?Z;CcD`0\\;@cDb0\\;_ObDc0];^OaDd0^;\\OSDS1m;nNQDT1n;mNPDU1`0iMi:R1gDV18QNo:j0gDo1W;Y1N2O1N2N2O1N2N2N10O0000001N102N2N1O2M2O2N2N1O1N3N1O1O2N1N1000000O100O100O010O01O001O001N101N2O0O2O1N101N101_OQL[EQ4d:?01OdL\\Ej1c:VNcEd1\\:\\NjE`1U:`NQFZ1o9eNWFV1k9gN[FU1e9jNaFP1a9mNeFn0]9POiFj0Y9SOkFj0V9UOmFh0U9UOnFi0T9UOnFi0S9UOPGi0R9UOPGi0R9TOQGj0T9QOnFm0W9mNlFQ1Z9iNhFU1n;M4L4L4L4E;DYPT1"}, "label": "a skateboarder doing a trick"}]} {"id": 559331, "image": "COCO_train2014_000000559331.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skateboarder in the air\"."}], "task": "refering", "answer_template": "The \"the skateboarder in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 19, 20, 21, 22, 23, 33, 34, 35, 46, 47, 48, 57, 58, 59, 60, 61, 62, 70, 71, 72, 73, 74, 75, 85, 86, 87], "bbox": [0.4289487870619946, 0.03048, 0.8001078167115903, 0.38488], "iscrowd": 0, "area": 12144.838549999999, "rle": {"size": [500, 371], "counts": "dh]22a?8I6I00100O010O10O01O10O0100O00100O10O010O01000O010O010O10O10O010OfME\\EY;DeD>Y;CfD?Y;BeD?Z;CcD`0\\;@cDb0\\;_ObDc0];^OaDd0^;\\OSDS1m;nNQDT1n;mNPDU1`0iMi:R1gDV18QNo:j0gDo1W;Y1N2O1N2N2O1N2N2N10O0000001N102N2N1O2M2O2N2N1O1N3N1O1O2N1N1000000O100O100O010O01O001O001N101N2O0O2O1N101N101_OQL[EQ4d:?01OdL\\Ej1c:VNcEd1\\:\\NjE`1U:`NQFZ1o9eNWFV1k9gN[FU1e9jNaFP1a9mNeFn0]9POiFj0Y9SOkFj0V9UOmFh0U9UOnFi0T9UOnFi0S9UOPGi0R9UOPGi0R9TOQGj0T9QOnFm0W9mNlFQ1Z9iNhFU1n;M4L4L4L4E;DYPT1"}, "label": "the skateboarder in the air"}]} {"id": 321766, "image": "COCO_train2014_000000321766.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe to the right with its neck behind the giraffe on the left\"."}], "task": "refering", "answer_template": "The \"a giraffe to the right with its neck behind the giraffe on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 122, 123, 146, 170, 171, 192, 193, 194, 215, 216, 217, 218, 238, 239, 240, 241, 261, 262, 263, 264, 285, 286, 287, 308, 309, 310, 331, 332, 333, 354, 355, 356, 377, 378, 379], "bbox": [0.309109375, 0.24497916666666666, 0.51928125, 0.98], "iscrowd": 0, "area": 16589.51105, "rle": {"size": [480, 640], "counts": "Znl21o>1N2O1O1O1N2O1O1O10O01fAIi=7UB3c=NZB4g=KWB6j=KSB7m=>N2L4K4M4M3000O1000015J2N05J6J5K6JcNSCn0gfMj1`8hMhIb0bMl1c8cMiI^3S6cLkIa3T6^LjIg3T6VLlIP4X1SLm0ImM[4h0ULg3\\4SJUK>g0]5e4[IPKP1b0c5U5TJRKj5Q5jIZKU6i6001O1O1O1O1O001O1O1WNWIYKj6\\4hI\\KY6Y4YJ_Kh5R3ZI_LW2E`4m3bLnK_3S4cLiK^3X4d32O1O1O1O1O100O103gGiK`5[4UJlKl5X4gIoKZ6V4UIVLl6m3ZHhLg7Y51O0O101N1O2N1O2N1O1N3N1O1O1YLgGe0[8]OjG^N`0Fg7P2mGVN:Fk7X2nGoM6En7`2nGhM3DP8l2lG^M1CV8V3gGPM6FU8a3dGZLj9n3a08I6I?A1O01O0fM`FSO`9d0mFWOU9?WG^Oj87dGD]83PHDV82WHFn7N]HNh7EdH6b7]OjH?o:L5J5L4LXn_4"}, "label": "a giraffe to the right with its neck behind the giraffe on the left"}]} {"id": 321766, "image": "COCO_train2014_000000321766.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two zibras standing on road\"."}], "task": "refering", "answer_template": "The \"two zibras standing on road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 123, 124, 125, 146, 147, 148, 149, 168, 169, 170, 190, 191, 192, 213, 214, 215, 235, 236, 237, 238, 258, 259, 260, 261, 281, 282, 283, 284, 304, 305, 306, 307, 328, 329, 330, 351, 352, 353, 376], "bbox": [0.233546875, 0.24410416666666668, 0.48928125, 0.9715208333333334], "iscrowd": 0, "area": 17355.7725, "rle": {"size": [480, 640], "counts": "RdV2:`>9G9G:H7K5K6J5K5K5K6J5K5K6J:F`0@`0@`0@2O0000N1N3L4M3L4M2M4M3M31N2O1O2N1N2O1O1eFRLc7P4mG_LQ8c3`GmL^8l4N2N2M3N2N2M2O2N1O1N2O1O1O2MTO_H]J`7`5Q1N1N3N2M3N1M4I7I7I7I7I7S1mNa0_Oa0_Oa0_O`0@=C1O1OSOZIcId6T6jIhIU6o5YJnIe5i5iJSJV5d5XKXJg4T2[In0[2kLX4P2iIj0Z2RMl3m1UJg0j8ROaGd0e8XOdG=b8BcG4d8JaGF\\7kN[H^1?TOd7_OlG[1f0dNl72\\GZ1i:hNTEX1m:iNQEW1P;jNnDV1T;kNiDU1Y;kNeDU1];kNaDT1b;mN[DS1g;mNWDS1k;mNSDS1o;m02N2M3N1O2N1O1O1O1N3N1O1O1O1OO2N2N1O2N21O2N1O1N200010O001O00010O001O010O0100O2O1N101N101N2N2O0O2O1N2N1O2N2N2N5K9G4K3N2N2N3M2N2NZSi4"}, "label": "two zibras standing on road"}]} {"id": 321766, "image": "COCO_train2014_000000321766.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe looking to the right\"."}], "task": "refering", "answer_template": "The \"a giraffe looking to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 123, 124, 125, 146, 147, 148, 149, 168, 169, 170, 190, 191, 192, 213, 214, 215, 235, 236, 237, 238, 258, 259, 260, 261, 281, 282, 283, 284, 304, 305, 306, 307, 328, 329, 330, 351, 352, 353, 376], "bbox": [0.233546875, 0.24410416666666668, 0.48928125, 0.9715208333333334], "iscrowd": 0, "area": 17355.7725, "rle": {"size": [480, 640], "counts": "RdV2:`>9G9G:H7K5K6J5K5K5K6J5K5K6J:F`0@`0@`0@2O0000N1N3L4M3L4M2M4M3M31N2O1O2N1N2O1O1eFRLc7P4mG_LQ8c3`GmL^8l4N2N2M3N2N2M2O2N1O1N2O1O1O2MTO_H]J`7`5Q1N1N3N2M3N1M4I7I7I7I7I7S1mNa0_Oa0_Oa0_O`0@=C1O1OSOZIcId6T6jIhIU6o5YJnIe5i5iJSJV5d5XKXJg4T2[In0[2kLX4P2iIj0Z2RMl3m1UJg0j8ROaGd0e8XOdG=b8BcG4d8JaGF\\7kN[H^1?TOd7_OlG[1f0dNl72\\GZ1i:hNTEX1m:iNQEW1P;jNnDV1T;kNiDU1Y;kNeDU1];kNaDT1b;mN[DS1g;mNWDS1k;mNSDS1o;m02N2M3N1O2N1O1O1O1N3N1O1O1O1OO2N2N1O2N21O2N1O1N200010O001O00010O001O010O0100O2O1N101N101N2N2O0O2O1N2N1O2N2N2N5K9G4K3N2N2N3M2N2NZSi4"}, "label": "a giraffe looking to the right"}]} {"id": 346344, "image": "COCO_train2014_000000346344.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"traffic lights not above the bus\"."}], "task": "refering", "answer_template": "The \"traffic lights not above the bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 36, 37, 38, 49, 50, 51, 62, 63, 64, 75, 76, 77, 89, 90], "bbox": [0.7795733333333332, 0.09566, 0.9625866666666666, 0.38270000000000004], "iscrowd": 0, "area": 7219.782349999999, "rle": {"size": [500, 375], "counts": "Rf^42n=OZCe4_OZ;9\\D;2D`;6XD91Ge;5UD6OMj;2QD41Nl;3nC05Nk;Z1TDgNj;[1UDeNj;\\1UDfNj;[1UDfNj;Z1VDfNj;[1TDgNk;Z1TDfNl;Z1TDgNk;Z1SDhNl;Y1SDgNm;a2O001O00100O1O7I1O00000000000000001O000000010O001O001O1O001O000100O00100O001aLhDa2X;XMQEe2o:TMZEj2g:oLaEP3d;`MkC[1X=O2O1N2NR1oN8G8Hmd6"}, "label": "traffic lights not above the bus"}]} {"id": 346344, "image": "COCO_train2014_000000346344.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a traffic light\"."}], "task": "refering", "answer_template": "The \"a traffic light\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 36, 37, 38, 49, 50, 51, 62, 63, 64, 75, 76, 77, 89, 90], "bbox": [0.7795733333333332, 0.09566, 0.9625866666666666, 0.38270000000000004], "iscrowd": 0, "area": 7219.782349999999, "rle": {"size": [500, 375], "counts": "Rf^42n=OZCe4_OZ;9\\D;2D`;6XD91Ge;5UD6OMj;2QD41Nl;3nC05Nk;Z1TDgNj;[1UDeNj;\\1UDfNj;[1UDfNj;Z1VDfNj;[1TDgNk;Z1TDfNl;Z1TDgNk;Z1SDhNl;Y1SDgNm;a2O001O00100O1O7I1O00000000000000001O000000010O001O001O1O001O000100O00100O001aLhDa2X;XMQEe2o:TMZEj2g:oLaEP3d;`MkC[1X=O2O1N2NR1oN8G8Hmd6"}, "label": "a traffic light"}]} {"id": 67835, "image": "COCO_train2014_000000067835.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboarder\"."}], "task": "refering", "answer_template": "The \"a skateboarder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 30, 31, 53, 54, 55, 56, 57, 58, 59, 76, 77, 78, 79, 80, 81, 82, 83, 100, 101, 102, 103, 104, 105, 106, 111, 112, 123, 124, 125, 126, 127, 128, 129, 133, 134, 135, 136, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 309, 310, 324, 325, 327, 328, 348], "bbox": [0.10965625000000001, 0.03146484375, 0.9204531249999999, 0.8426953125000001], "iscrowd": 0, "area": 60461.890999999996, "rle": {"size": [512, 640], "counts": "h\\S1d0\\?000000001O0O100000000000001N100002N2N2N2N2O1N2N2N2N2N2N2N_ob00^P]O3N2N2M3N2M3N2M3N2N2M3N2M3N1N3N2N2M3N2M3hNX1O1O1O1O1O1O1O1O1O100O1O1O001O1O1O1O1O1O100O1ON2N2M3M3N1N3M2O2M3M2011O0O2O1RITNGm16WNGk17XNGh18[NEg19\\NEe18_NFb18aNEa19bNE^19fNE[19hNDZ1:iNDX1HnL^Km1h4U1J_O3c0M^O1c0O]OOe01\\OMe03\\OKd06]OGe09\\OEe0;\\OCe0<]OAe0?\\O_Od0b0]O\\Od0d0]OYOe0g0\\OWOe0i0\\OTOe0m0\\OQOe0o0\\OnNf0Q1\\OlNf0T1]MTK?f3V2V1WM[Ka07cNX2g3W2PMbKd0HnNZ2`3\\2jLiKf0YOZO]2X3a2cLPLl1Z1c1f2]LWLo1o0f1j2WL^LS2c0h1n2QLgLW26j1S3kKmL\\2Kk1X3dKUM`2^Oo1h5PNTJR2o5kMoIW2S6hMjIZ2W6fMfI\\2Z6dMdI^2]6bM`I`2a6`M\\Ib2e6]MYIe2h6[MTIg2m6ZMPIh2Q7XMlHj2U7UMiHm2X7SMeHo2\\7QMaHQ3X6bKdI]12S3Y6VNdIl1\\6UN`In1`6RN^IP2b6g3000000000000000000O10000000000000000000000O100000001O0]JcFc4]9ZKfFf4Z9WKiFi4W9TKkFm4U9PKnFP5R9mJQGS5P9iJSGW5m8eJWGZ5j8cJYG]5g8`J[Ga5b9000000000O1000000000O10O10000000000O101O0000000O2O000000001N100O1G9F;F9F:K5O1O2N100O1O1O2N100O1M3@`0_Ob0H7N2N2N2N3M2N2N2N2M4M2N2N2N2N3M2N2N2N2M4M2N2N2O1O2M2O1O1O1O2N1O1O1O1N3N1O2N2N2N2N3M2O1O1O1cMcD5_;GdD8];EfD:[;BiD=X;_OkD`0X;ZOmDe0T;UOREj0o:POWEo0j:kN\\ET1e:gN`EX1b:aNcE_1^:[NhEd1Y:VNmEi1V in this image.", "objects": [{"patches": [7, 8, 30, 31, 53, 54, 55, 56, 57, 58, 59, 76, 77, 78, 79, 80, 81, 82, 83, 100, 101, 102, 103, 104, 105, 106, 111, 112, 123, 124, 125, 126, 127, 128, 129, 133, 134, 135, 136, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 309, 310, 324, 325, 327, 328, 348], "bbox": [0.10965625000000001, 0.03146484375, 0.9204531249999999, 0.8426953125000001], "iscrowd": 0, "area": 60461.890999999996, "rle": {"size": [512, 640], "counts": "h\\S1d0\\?000000001O0O100000000000001N100002N2N2N2N2O1N2N2N2N2N2N2N_ob00^P]O3N2N2M3N2M3N2M3N2N2M3N2M3N1N3N2N2M3N2M3hNX1O1O1O1O1O1O1O1O1O100O1O1O001O1O1O1O1O1O100O1ON2N2M3M3N1N3M2O2M3M2011O0O2O1RITNGm16WNGk17XNGh18[NEg19\\NEe18_NFb18aNEa19bNE^19fNE[19hNDZ1:iNDX1HnL^Km1h4U1J_O3c0M^O1c0O]OOe01\\OMe03\\OKd06]OGe09\\OEe0;\\OCe0<]OAe0?\\O_Od0b0]O\\Od0d0]OYOe0g0\\OWOe0i0\\OTOe0m0\\OQOe0o0\\OnNf0Q1\\OlNf0T1]MTK?f3V2V1WM[Ka07cNX2g3W2PMbKd0HnNZ2`3\\2jLiKf0YOZO]2X3a2cLPLl1Z1c1f2]LWLo1o0f1j2WL^LS2c0h1n2QLgLW26j1S3kKmL\\2Kk1X3dKUM`2^Oo1h5PNTJR2o5kMoIW2S6hMjIZ2W6fMfI\\2Z6dMdI^2]6bM`I`2a6`M\\Ib2e6]MYIe2h6[MTIg2m6ZMPIh2Q7XMlHj2U7UMiHm2X7SMeHo2\\7QMaHQ3X6bKdI]12S3Y6VNdIl1\\6UN`In1`6RN^IP2b6g3000000000000000000O10000000000000000000000O100000001O0]JcFc4]9ZKfFf4Z9WKiFi4W9TKkFm4U9PKnFP5R9mJQGS5P9iJSGW5m8eJWGZ5j8cJYG]5g8`J[Ga5b9000000000O1000000000O10O10000000000O101O0000000O2O000000001N100O1G9F;F9F:K5O1O2N100O1O1O2N100O1M3@`0_Ob0H7N2N2N2N3M2N2N2N2M4M2N2N2N2N3M2N2N2N2M4M2N2N2O1O2M2O1O1O1O2N1O1O1O1N3N1O2N2N2N2N3M2O1O1O1cMcD5_;GdD8];EfD:[;BiD=X;_OkD`0X;ZOmDe0T;UOREj0o:POWEo0j:kN\\ET1e:gN`EX1b:aNcE_1^:[NhEd1Y:VNmEi1V in this image.", "objects": [{"patches": [139, 140, 141, 156, 157, 158, 173, 174, 175, 176, 190, 191, 192, 193, 207, 208, 224, 225, 226, 227, 241, 242, 243, 244], "bbox": [0.17189583333333333, 0.364390625, 0.417125, 0.64259375], "iscrowd": 0, "area": 9292.884500000004, "rle": {"size": [640, 480], "counts": "PVd1:`c0;E;F5O1O1ROjNP_OX1k`0nNS_OS1g`0TOW_Om0d`0YOZ_Oh0a`0^O\\_Od0c`0^OW_Og0h`0[OQ_Oj0o`0T1O1i_OVMW?k2j@WMR?j2n@YMn>h2SAZMi>g2WA\\Me>e2[A^Ma>c2`A_M\\>b2dAaMY>_2gAaMX>`2iA`MU>a2kA`MS>b2lA^MS>c2mA^MQ>c2oA^MP>b2QB^Mm=d2RB\\Mn=h2nAYMQ>j2lAWMS>l2jATMW>k2iAVMV>k2iAVMV>j2kAVMT>j2lAVMU>j2jAWMU>i2lAVMT>j2lAVMT>k2kAVMU>i2lAVMT>k2kAUMU>k2kAUMU>k2kAUMU>l2kASMV>l2jATMV>l2jATMV>m2jARMV>m2lARMU>l2lATMT>j2oAUMQ>j2QBUMo=i2SBWMn=g2TBXMm=R2TAjMQ14k=R2WAfMP17j=Q2ZAcMn0g=Q2^A\\Mn0b0e=P2jBPNV=P2kBoMV=P2jBPNX=m1jBRNY=k1gBUN\\=h1fBVN\\=g1hBVN[=g1hBVN[=f1hBXN[=e1hBWNb=`1j2D5L1O0O2O1O1N101O1N101O1N[_O^On=b0mAER>:\\A9d>EZA?f>@YAb0f>]O[Ac0f>\\OZAd0g>[OXAe0i>ZOXAf0i>YOVAg0l>XOTAh0m>VOTAj0l>VOSAj0o>UOQAk0P?SOQAm0o>SOPAm0R?QOo@o0R?POn@P1S?oNl@Q1U?nNl@R1U?mNk@S1V?lNi@T1Ra00OYOY]O7gb0F\\]O:Vc0O1O2M2O1O1O2N4L4K`c^5"}, "label": "a messy baby eating while sitting in a stroller"}]} {"id": 100611, "image": "COCO_train2014_000000100611.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde baby with a yellow shirt , bib , and messy face , sitting in a grey stroller\"."}], "task": "refering", "answer_template": "The \"a blonde baby with a yellow shirt , bib , and messy face , sitting in a grey stroller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 156, 157, 158, 173, 174, 175, 176, 190, 191, 192, 193, 207, 208, 224, 225, 226, 227, 241, 242, 243, 244], "bbox": [0.17189583333333333, 0.364390625, 0.417125, 0.64259375], "iscrowd": 0, "area": 9292.884500000004, "rle": {"size": [640, 480], "counts": "PVd1:`c0;E;F5O1O1ROjNP_OX1k`0nNS_OS1g`0TOW_Om0d`0YOZ_Oh0a`0^O\\_Od0c`0^OW_Og0h`0[OQ_Oj0o`0T1O1i_OVMW?k2j@WMR?j2n@YMn>h2SAZMi>g2WA\\Me>e2[A^Ma>c2`A_M\\>b2dAaMY>_2gAaMX>`2iA`MU>a2kA`MS>b2lA^MS>c2mA^MQ>c2oA^MP>b2QB^Mm=d2RB\\Mn=h2nAYMQ>j2lAWMS>l2jATMW>k2iAVMV>k2iAVMV>j2kAVMT>j2lAVMU>j2jAWMU>i2lAVMT>j2lAVMT>k2kAVMU>i2lAVMT>k2kAUMU>k2kAUMU>k2kAUMU>l2kASMV>l2jATMV>l2jATMV>m2jARMV>m2lARMU>l2lATMT>j2oAUMQ>j2QBUMo=i2SBWMn=g2TBXMm=R2TAjMQ14k=R2WAfMP17j=Q2ZAcMn0g=Q2^A\\Mn0b0e=P2jBPNV=P2kBoMV=P2jBPNX=m1jBRNY=k1gBUN\\=h1fBVN\\=g1hBVN[=g1hBVN[=f1hBXN[=e1hBWNb=`1j2D5L1O0O2O1O1N101O1N101O1N[_O^On=b0mAER>:\\A9d>EZA?f>@YAb0f>]O[Ac0f>\\OZAd0g>[OXAe0i>ZOXAf0i>YOVAg0l>XOTAh0m>VOTAj0l>VOSAj0o>UOQAk0P?SOQAm0o>SOPAm0R?QOo@o0R?POn@P1S?oNl@Q1U?nNl@R1U?mNk@S1V?lNi@T1Ra00OYOY]O7gb0F\\]O:Vc0O1O2M2O1O1O2N4L4K`c^5"}, "label": "a blonde baby with a yellow shirt , bib , and messy face , sitting in a grey stroller"}]} {"id": 100611, "image": "COCO_train2014_000000100611.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady sitting next to a toddler wiping her mouth\"."}], "task": "refering", "answer_template": "The \"a lady sitting next to a toddler wiping her mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 42, 43, 44, 58, 59, 60, 61, 74, 75, 76, 77, 78, 91, 92, 93, 94, 95, 108, 109, 110, 111, 112, 125, 126], "bbox": [0.39887500000000004, 0.06029687500000001, 0.6375416666666667, 0.31853125000000004], "iscrowd": 0, "area": 10806.5016, "rle": {"size": [640, 480], "counts": "[bg31ic06cN\\1I8^Ob0N2M2O2O1N2O0010000O0100O10O10O10O10O1O001O001O010O10O01O010O1O001O1O001O100N101N2O001N2O1N2O1O1NjN`NP@_1n?gNo_OX1Q`0kNl_OU1S`0oNk_OP1R`0WOk_Oh0U`0ZOj_Od0W`0^Og_Ob0Y`0Ae_O>]`0Bb_O=_`0D]_O>c`0_1001O0100O101N2O1N2O2M:G4K5L3L5L4K101N101O0000000000O100O101N1bMd@7]?Hh@2Z?Mh@1Y?Ni@0Y?Ni@OY?0i@NY?0i@NX?1j@MX?1j@LX?3j@KX?3j@I[?2i@KZ?1j@J[?2l@GX?5PABU?;RA]OT?>XAUOo>e0Z2L;D4Lje\\3"}, "label": "a lady sitting next to a toddler wiping her mouth"}]} {"id": 100611, "image": "COCO_train2014_000000100611.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with red hair wiping her chin\"."}], "task": "refering", "answer_template": "The \"a woman with red hair wiping her chin\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [25, 26, 27, 42, 43, 44, 58, 59, 60, 61, 74, 75, 76, 77, 78, 91, 92, 93, 94, 95, 108, 109, 110, 111, 112, 125, 126], "bbox": [0.39887500000000004, 0.06029687500000001, 0.6375416666666667, 0.31853125000000004], "iscrowd": 0, "area": 10806.5016, "rle": {"size": [640, 480], "counts": "[bg31ic06cN\\1I8^Ob0N2M2O2O1N2O0010000O0100O10O10O10O10O1O001O001O010O10O01O010O1O001O1O001O100N101N2O001N2O1N2O1O1NjN`NP@_1n?gNo_OX1Q`0kNl_OU1S`0oNk_OP1R`0WOk_Oh0U`0ZOj_Od0W`0^Og_Ob0Y`0Ae_O>]`0Bb_O=_`0D]_O>c`0_1001O0100O101N2O1N2O2M:G4K5L3L5L4K101N101O0000000000O100O101N1bMd@7]?Hh@2Z?Mh@1Y?Ni@0Y?Ni@OY?0i@NY?0i@NX?1j@MX?1j@LX?3j@KX?3j@I[?2i@KZ?1j@J[?2l@GX?5PABU?;RA]OT?>XAUOo>e0Z2L;D4Lje\\3"}, "label": "a woman with red hair wiping her chin"}]} {"id": 100611, "image": "COCO_train2014_000000100611.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black chair with no one sitting in it\"."}], "task": "refering", "answer_template": "The \"a black chair with no one sitting in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 235, 236, 237, 251, 252, 253, 254, 267, 268, 269, 270, 271, 284, 285, 286, 287, 288, 300, 301, 302, 303, 304, 305, 318, 319, 320, 321, 322, 334, 335, 337, 338, 339, 350, 351, 356, 367, 368, 384], "bbox": [0.5935833333333334, 0.546078125, 1.0, 0.9733750000000001], "iscrowd": 0, "area": 21734.837500000005, "rle": {"size": [640, 480], "counts": "Zgb51mc02N3M2O2M2N2O2M2N3M100O0010O0001O010O0010O01O0010O01O010O001O01O01O010O001O010O0010O01O0010O01O0\\NVOT@j0k?ZOS@f0l?]OQ@c0o?Am_O`0R`0Cl_OX_OBg`0b0V_O_Oi`0d0T_O\\Ol`0h0Q_OXOn`0k0o^OUOQa0l110O0010O010O010O01O010O010O101N3N1N3M2O2M2O2O1O001O001O001O001O001O001O001O001O1O001O001O001O001O001N1O2M2N3N2M2N3N1N2N2O2M2N2O2M2N2O2M2N2O2M2O1N3M2O1N3M2O1N3M2O1N3M2O1O101N100O2N100O2O0O101O00001O00001O00001O00001O0000001O00001O00001O00001O00001O000_1bNQ2nMR2nM`_O"}, "label": "a black chair with no one sitting in it"}]} {"id": 100611, "image": "COCO_train2014_000000100611.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black patio chair outside on the sidewalk unoccupied\"."}], "task": "refering", "answer_template": "The \"black patio chair outside on the sidewalk unoccupied\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 235, 236, 237, 251, 252, 253, 254, 267, 268, 269, 270, 271, 284, 285, 286, 287, 288, 300, 301, 302, 303, 304, 305, 318, 319, 320, 321, 322, 334, 335, 337, 338, 339, 350, 351, 356, 367, 368, 384], "bbox": [0.5935833333333334, 0.546078125, 1.0, 0.9733750000000001], "iscrowd": 0, "area": 21734.837500000005, "rle": {"size": [640, 480], "counts": "Zgb51mc02N3M2O2M2N2O2M2N3M100O0010O0001O010O0010O01O0010O01O010O001O01O01O010O001O010O0010O01O0010O01O0\\NVOT@j0k?ZOS@f0l?]OQ@c0o?Am_O`0R`0Cl_OX_OBg`0b0V_O_Oi`0d0T_O\\Ol`0h0Q_OXOn`0k0o^OUOQa0l110O0010O010O010O01O010O010O101N3N1N3M2O2M2O2O1O001O001O001O001O001O001O001O001O1O001O001O001O001O001N1O2M2N3N2M2N3N1N2N2O2M2N2O2M2N2O2M2N2O2M2O1N3M2O1N3M2O1N3M2O1N3M2O1O101N100O2N100O2O0O101O00001O00001O00001O00001O0000001O00001O00001O00001O00001O000_1bNQ2nMR2nM`_O"}, "label": "black patio chair outside on the sidewalk unoccupied"}]} {"id": 125193, "image": "COCO_train2014_000000125193.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person skateboarding on a half - pipe\"."}], "task": "refering", "answer_template": "The \"a person skateboarding on a half - pipe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 29, 30, 51, 52, 53, 54, 73, 74, 75, 77, 78, 95, 96, 118, 119], "bbox": [0.13825, 0.048430913348946135, 0.39820312500000005, 0.39751756440281033], "iscrowd": 0, "area": 6846.593049999999, "rle": {"size": [427, 640], "counts": "mjT12iO2O00001O001N10001O001O0O2O0O101N101N101O0O1010O100O1O100O100O100O1O100O1O100O100SOhNfEY1Y:iNeEX1Y:kNfEU1Y:nNdES1Z:POeEP1Z:QOeEP1Z:QOeEP1Y:QOhEo0W:ROhEo0V:ROjEo0U:ROjEo0T:ROlEo0T:QOkEP1U:oNkER1V:mNiET1W:kNiEU1Y:jNfEV1[:iNeEW1]:hNbEX1_:jN^EV1d:kNYET1i:mNVER1l:oNQEQ1P;d02N1O2M2O2N1O2N1O2N1N3N2N1O2N1O2N2O0O2O0O2O1N1O20O10O10O010O11N5L4L4K4M1N10000O2O000O10001N100O101O0O10O1O010O001O1OO10001O00002N4L4L[^P5"}, "label": "a person skateboarding on a half - pipe"}]} {"id": 125193, "image": "COCO_train2014_000000125193.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboarder riding a pipe\"."}], "task": "refering", "answer_template": "The \"a skateboarder riding a pipe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 29, 30, 51, 52, 53, 54, 73, 74, 75, 77, 78, 95, 96, 118, 119], "bbox": [0.13825, 0.048430913348946135, 0.39820312500000005, 0.39751756440281033], "iscrowd": 0, "area": 6846.593049999999, "rle": {"size": [427, 640], "counts": "mjT12iO2O00001O001N10001O001O0O2O0O101N101N101O0O1010O100O1O100O100O100O1O100O1O100O100SOhNfEY1Y:iNeEX1Y:kNfEU1Y:nNdES1Z:POeEP1Z:QOeEP1Z:QOeEP1Y:QOhEo0W:ROhEo0V:ROjEo0U:ROjEo0T:ROlEo0T:QOkEP1U:oNkER1V:mNiET1W:kNiEU1Y:jNfEV1[:iNeEW1]:hNbEX1_:jN^EV1d:kNYET1i:mNVER1l:oNQEQ1P;d02N1O2M2O2N1O2N1O2N1N3N2N1O2N1O2N2O0O2O0O2O1N1O20O10O10O010O11N5L4L4K4M1N10000O2O000O10001N100O101O0O10O1O010O001O1OO10001O00002N4L4L[^P5"}, "label": "a skateboarder riding a pipe"}]} {"id": 190732, "image": "COCO_train2014_000000190732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a red shirt\"."}], "task": "refering", "answer_template": "The \"a man in a red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 27, 28, 29, 30, 44, 45, 46, 47, 48, 49, 62, 63, 64, 65, 66, 67, 79, 80, 81, 82, 83, 84, 85, 86, 98, 99, 100, 101, 102, 103, 104, 105, 117, 118, 119, 120, 122, 123, 135, 136, 137, 138, 139, 154, 155, 156, 157, 158, 174, 175, 176, 193, 194, 211], "bbox": [0.4301785714285714, 0.0005637982195845697, 0.8829761904761905, 0.9418694362017805], "iscrowd": 0, "area": 30442.41725, "rle": {"size": [337, 504], "counts": "XaW22]:3M3N2M3SOI\\G:b8KYG7d80UG3i82RG0k8l0O2M3N2M3N2M3N2N2M4M2M3N2M3N2M3N3L3N2M3N3M3L5L3L5L3L5L3L5M2N3L3N3M2N2N3L3N2N2N2M3N2N2N1O0O2O1O1O001N2O1O001N2O1O001O1N2O001O1O1N2O001O1N2O001F:E;I6N3M3M3M2N3N2O1O100O2N1O1O2N1O1O1O000001N1000000000000O1000001O000O100001O001O001O010O001OjIZMa4f2]K\\Mc4d2[K^Mf4a2WKbMi4^2UKdMn0N]1]2cMgMn07V1R2iMkMm0>S1g1mMmMn0g0l0\\1SNPNo0n0g0Q1XNSNn0T1e0i0[NUNn0V1f0e0XNXNS1S1d0g0UNXNX1R1b0f0SNZN\\1Q1?g0PN[Nb1n0=h0nM]Ne1l0[2AeMc0X2\\OiMg0T2XOmMk0P2TOPNU1h1iNZNd1Y1[NgNf1Y1YNhNf1Y1YNgNf1Z1ZNgNe1Z1ZNgNe1Z1ZNfNf1[1YNfNe1\\1ZNdNf1]1YNdNe1]1[NdNb1^1^NbN_1b1`N_N]1c1cN]N[1e1eN\\NX1f1hNZNV1h1iNZNS1j1lNWNQ1k1oNUNo0m1QOTNk0o1UOQNi0Q2WOPNe0T2ZOlMc0W2]OjM`0X2@hM=[2CfM9]2GcM7`2HaM4b2L^M1e2O\\MNf22ZMKj24WMHl28TMFn29SMCQ3=PM]OU3c0kLXO[3g0X30001O001O001O1O001O1O001N10000N2O1N101O1N101O1N2O0O2O1O0O2O1O]bc0"}, "label": "a man in a red shirt"}]} {"id": 190732, "image": "COCO_train2014_000000190732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"soccer player in red\"."}], "task": "refering", "answer_template": "The \"soccer player in red\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 27, 28, 29, 30, 44, 45, 46, 47, 48, 49, 62, 63, 64, 65, 66, 67, 79, 80, 81, 82, 83, 84, 85, 86, 98, 99, 100, 101, 102, 103, 104, 105, 117, 118, 119, 120, 122, 123, 135, 136, 137, 138, 139, 154, 155, 156, 157, 158, 174, 175, 176, 193, 194, 211], "bbox": [0.4301785714285714, 0.0005637982195845697, 0.8829761904761905, 0.9418694362017805], "iscrowd": 0, "area": 30442.41725, "rle": {"size": [337, 504], "counts": "XaW22]:3M3N2M3SOI\\G:b8KYG7d80UG3i82RG0k8l0O2M3N2M3N2M3N2N2M4M2M3N2M3N2M3N3L3N2M3N3M3L5L3L5L3L5L3L5M2N3L3N3M2N2N3L3N2N2N2M3N2N2N1O0O2O1O1O001N2O1O001N2O1O001O1N2O001O1O1N2O001O1N2O001F:E;I6N3M3M3M2N3N2O1O100O2N1O1O2N1O1O1O000001N1000000000000O1000001O000O100001O001O001O010O001OjIZMa4f2]K\\Mc4d2[K^Mf4a2WKbMi4^2UKdMn0N]1]2cMgMn07V1R2iMkMm0>S1g1mMmMn0g0l0\\1SNPNo0n0g0Q1XNSNn0T1e0i0[NUNn0V1f0e0XNXNS1S1d0g0UNXNX1R1b0f0SNZN\\1Q1?g0PN[Nb1n0=h0nM]Ne1l0[2AeMc0X2\\OiMg0T2XOmMk0P2TOPNU1h1iNZNd1Y1[NgNf1Y1YNhNf1Y1YNgNf1Z1ZNgNe1Z1ZNgNe1Z1ZNfNf1[1YNfNe1\\1ZNdNf1]1YNdNe1]1[NdNb1^1^NbN_1b1`N_N]1c1cN]N[1e1eN\\NX1f1hNZNV1h1iNZNS1j1lNWNQ1k1oNUNo0m1QOTNk0o1UOQNi0Q2WOPNe0T2ZOlMc0W2]OjM`0X2@hM=[2CfM9]2GcM7`2HaM4b2L^M1e2O\\MNf22ZMKj24WMHl28TMFn29SMCQ3=PM]OU3c0kLXO[3g0X30001O001O001O1O001O1O001N10000N2O1N101O1N101O1N2O0O2O1O0O2O1O]bc0"}, "label": "soccer player in red"}]} {"id": 190732, "image": "COCO_train2014_000000190732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two soccer player battling for the ball\"."}], "task": "refering", "answer_template": "The \"two soccer player battling for the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 38, 39, 40, 41, 42, 43, 44, 57, 58, 59, 60, 61, 62, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 98, 110, 111, 112, 113, 114, 115, 116, 117, 129, 130, 131, 132, 133, 134, 135, 136, 147, 149, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 171, 172, 173, 183, 185, 186, 190, 191, 192, 203, 204, 205, 209, 210], "bbox": [0.14125, 0.08329376854599406, 0.7137301587301587, 0.9866765578635015], "iscrowd": 0, "area": 35445.865650000014, "rle": {"size": [337, 504], "counts": "aag01Z:6K5K5K5K5fNB\\Hc0b5TObKS2X4WNcKj1X4^NcKd1W4dNeK^1Y4fNcK\\1n1dMTOS1jNZ1Q2dMUOU1fNY1S2dMVOV1dN[1P2aM\\OV1`Na1k1`M@R1aNe1h1[MGR1]Nk1d1UMOR1YNP2`1QM7Q1UNV2\\1kL?Q1QN[2Y1fLe0`5ROcJn0b5kN_JV1g5bNZJ_1k5ZNVJg1P6PNQJR2T6gMlI[2T719GO2L3L4L5K4L5L3O2M2O2M2O1UOWMmIk2X3ZNgLHNP2\\3McL5\\3LaL6`3I_L9a3G\\L;e3X3100O100000O100000000000000O11O000006J5LO0O1O1O1fLeL\\O\\3?jL_OX3nNnLmN0R2S3]NcM[O_OV2Q3XNgM\\O^OY2n2SNlM]O\\O^2l2mMoM_OZOb2j2hM^OU2e0dMAZ2^4M2N2N1O00000000000O1000001O0O1000000O100000000O1000000O10jNTNgIl1S6\\NkId1o5cNPJ]1n5fNQJY1o5jNoIV1o5mNPJS1o5oNPJQ1n5ROQJn0n5UOPJk0o5WOPJj0m5YORJg0m5[ORJe0m5^OQJb0m5ARJ?m5CRJ>l5DSJYJCd5`0[J@c5h2N2N3M2N2N2N3M2O1N2N3O0UMeJa0Z5\\OjJd0V5WOoJh0Q5UORKl0m4QOWKo0h4nN\\KR1d4jN_KW1`4SNhJDl0Y2[4PN\\LP2d3lM_LU2`3hMdLX2[3eMiL[2W3aMmL^2S3_MPMb2o2[MUMe2\\5N1O2N2M2O2N20O01O10O01O10O01O100O2N2O1N2O1N2N2O1N2O1O1O001O1O1O100O1O2N1O00000000000000000000000000000000000000000001O0000000000000000000000000000000000000001O1O2N1O1O2N1O2N1O2N1O1O2O0O2N1O2O0OO1O2N1O1O1O2N1N2O1N3N1N2O2M2O0OO2O2M2O1N2N3M2M3M4L3L4M4K4M3L4McS_1"}, "label": "two soccer player battling for the ball"}]} {"id": 190732, "image": "COCO_train2014_000000190732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"soccer referee fight\"."}], "task": "refering", "answer_template": "The \"soccer referee fight\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 38, 39, 40, 41, 42, 43, 44, 57, 58, 59, 60, 61, 62, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 98, 110, 111, 112, 113, 114, 115, 116, 117, 129, 130, 131, 132, 133, 134, 135, 136, 147, 149, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 171, 172, 173, 183, 185, 186, 190, 191, 192, 203, 204, 205, 209, 210], "bbox": [0.14125, 0.08329376854599406, 0.7137301587301587, 0.9866765578635015], "iscrowd": 0, "area": 35445.865650000014, "rle": {"size": [337, 504], "counts": "aag01Z:6K5K5K5K5fNB\\Hc0b5TObKS2X4WNcKj1X4^NcKd1W4dNeK^1Y4fNcK\\1n1dMTOS1jNZ1Q2dMUOU1fNY1S2dMVOV1dN[1P2aM\\OV1`Na1k1`M@R1aNe1h1[MGR1]Nk1d1UMOR1YNP2`1QM7Q1UNV2\\1kL?Q1QN[2Y1fLe0`5ROcJn0b5kN_JV1g5bNZJ_1k5ZNVJg1P6PNQJR2T6gMlI[2T719GO2L3L4L5K4L5L3O2M2O2M2O1UOWMmIk2X3ZNgLHNP2\\3McL5\\3LaL6`3I_L9a3G\\L;e3X3100O100000O100000000000000O11O000006J5LO0O1O1O1fLeL\\O\\3?jL_OX3nNnLmN0R2S3]NcM[O_OV2Q3XNgM\\O^OY2n2SNlM]O\\O^2l2mMoM_OZOb2j2hM^OU2e0dMAZ2^4M2N2N1O00000000000O1000001O0O1000000O100000000O1000000O10jNTNgIl1S6\\NkId1o5cNPJ]1n5fNQJY1o5jNoIV1o5mNPJS1o5oNPJQ1n5ROQJn0n5UOPJk0o5WOPJj0m5YORJg0m5[ORJe0m5^OQJb0m5ARJ?m5CRJ>l5DSJYJCd5`0[J@c5h2N2N3M2N2N2N3M2O1N2N3O0UMeJa0Z5\\OjJd0V5WOoJh0Q5UORKl0m4QOWKo0h4nN\\KR1d4jN_KW1`4SNhJDl0Y2[4PN\\LP2d3lM_LU2`3hMdLX2[3eMiL[2W3aMmL^2S3_MPMb2o2[MUMe2\\5N1O2N2M2O2N20O01O10O01O10O01O100O2N2O1N2O1N2N2O1N2O1O1O001O1O1O100O1O2N1O00000000000000000000000000000000000000000001O0000000000000000000000000000000000000001O1O2N1O1O2N1O2N1O2N1O1O2O0O2N1O2O0OO1O2N1O1O1O2N1N2O1N3N1N2O2M2O0OO2O2M2O1N2N3M2M3M4L3L4M4K4M3L4McS_1"}, "label": "soccer referee fight"}]} {"id": 141581, "image": "COCO_train2014_000000141581.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black car behind bags of ice\"."}], "task": "refering", "answer_template": "The \"a black car behind bags of ice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [184, 185, 186, 187, 207, 208, 209, 210, 211, 231, 232, 233, 234, 235, 258], "bbox": [0.0016875000000000002, 0.4719166666666667, 0.259546875, 0.6921250000000001], "iscrowd": 0, "area": 6802.0084, "rle": {"size": [480, 640], "counts": "mf0g0_=k0O00000000O101N100O100O10001N100O1000000O2O00000000000O2O00000000000100O2N2N1O2N1O2O1N1O2N2N1O2N2lBYNk in this image.", "objects": [{"patches": [184, 185, 186, 187, 207, 208, 209, 210, 211, 231, 232, 233, 234, 235, 258], "bbox": [0.0016875000000000002, 0.4719166666666667, 0.259546875, 0.6921250000000001], "iscrowd": 0, "area": 6802.0084, "rle": {"size": [480, 640], "counts": "mf0g0_=k0O00000000O101N100O100O10001N100O1000000O2O00000000000O2O00000000000100O2N2N1O2N1O2O1N1O2N2N1O2N2lBYNk in this image.", "objects": [{"patches": [33, 34, 55, 56, 57, 78, 79, 80, 81, 101, 102, 103, 104, 124, 125, 126, 147, 148, 149, 170, 171, 172, 194, 195, 216, 217, 218], "bbox": [0.395125, 0.07997933884297521, 0.5368906250000001, 0.5945247933884298], "iscrowd": 0, "area": 10845.756500000001, "rle": {"size": [484, 640], "counts": "jeg3:c>8K4O1N3O00001N001O00100O0nFTOm3m0QLUOn3k0RLWOl3j0[GXOd1c0K^OV7f0XGGY1:4]OZ7c0UG3Q12>YO[7i1PHSOc0UO\\7i1`GoMMW1U1RO^7g1]GJS1`N_7g1ZGNU1]N_7Q4_HnKb7S4_HkK`7W4bHfK^7\\4cHaK\\7e4_14K6K7H7J6I3N3M in this image.", "objects": [{"patches": [33, 34, 55, 56, 57, 78, 79, 80, 81, 101, 102, 103, 104, 124, 125, 126, 147, 148, 149, 170, 171, 172, 194, 195, 216, 217, 218], "bbox": [0.395125, 0.07997933884297521, 0.5368906250000001, 0.5945247933884298], "iscrowd": 0, "area": 10845.756500000001, "rle": {"size": [484, 640], "counts": "jeg3:c>8K4O1N3O00001N001O00100O0nFTOm3m0QLUOn3k0RLWOl3j0[GXOd1c0K^OV7f0XGGY1:4]OZ7c0UG3Q12>YO[7i1PHSOc0UO\\7i1`GoMMW1U1RO^7g1]GJS1`N_7g1ZGNU1]N_7Q4_HnKb7S4_HkK`7W4bHfK^7\\4cHaK\\7e4_14K6K7H7J6I3N3M in this image.", "objects": [{"patches": [121, 122, 123, 124, 144, 145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 193, 213, 214, 215, 216, 236, 237, 238, 260], "bbox": [0.248296875, 0.3280985915492958, 0.43528124999999995, 0.7573004694835681], "iscrowd": 0, "area": 12282.212, "rle": {"size": [426, 640], "counts": "UbS21U20n80RG1m8OSG1m80RG0n80oF3Q9MkF:R9GjF?S9AcFl0Z9TOeFP1o8nNkE3T1U1d8QOWFJT1[1`8lNZFJU1\\1`8jN[FKS1]1b8hNZFKS1_1b8fN[FKR1a1b8eN[FKR1a1b8nN^GS1b8mN]GT1b8mN]GT1b8mN]GT1c8lN\\GU1c8lN\\GU1c8lN\\GV1b8jN^GX1a8hN_GX1_8jN`GX1^8]OnFW1\\8jNdGX1Z8iNfGW1X8jNhGX1V8iNhGZ1U8gNjG\\1S8eNlG\\1T8eNjG]1T8dNlG\\1T8eNjG\\1U8eNjG]1U8l1O101O01O01OO101O0000001N1000N2O1N2OO10O010O10O01000O011N3N2N1N3N2N2L3N3L4M3M2M4M3O10O01O1O1O001O1_NgFoNY9Q1hFnNY9Q1iFmNX9R1iFmNX9R1iFlN^9n0cFQOg9e0ZFYOP:>QFAP:VG]Ol8b0o1N2N1O2N2N2N101N2N2N1O2Nl\\f4"}, "label": "a blue motorcycle"}]} {"id": 141603, "image": "COCO_train2014_000000141603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cyan blue motorcycle\"."}], "task": "refering", "answer_template": "The \"cyan blue motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 144, 145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 193, 213, 214, 215, 216, 236, 237, 238, 260], "bbox": [0.248296875, 0.3280985915492958, 0.43528124999999995, 0.7573004694835681], "iscrowd": 0, "area": 12282.212, "rle": {"size": [426, 640], "counts": "UbS21U20n80RG1m8OSG1m80RG0n80oF3Q9MkF:R9GjF?S9AcFl0Z9TOeFP1o8nNkE3T1U1d8QOWFJT1[1`8lNZFJU1\\1`8jN[FKS1]1b8hNZFKS1_1b8fN[FKR1a1b8eN[FKR1a1b8nN^GS1b8mN]GT1b8mN]GT1b8mN]GT1c8lN\\GU1c8lN\\GU1c8lN\\GV1b8jN^GX1a8hN_GX1_8jN`GX1^8]OnFW1\\8jNdGX1Z8iNfGW1X8jNhGX1V8iNhGZ1U8gNjG\\1S8eNlG\\1T8eNjG]1T8dNlG\\1T8eNjG\\1U8eNjG]1U8l1O101O01O01OO101O0000001N1000N2O1N2OO10O010O10O01000O011N3N2N1N3N2N2L3N3L4M3M2M4M3O10O01O1O1O001O1_NgFoNY9Q1hFnNY9Q1iFmNX9R1iFmNX9R1iFlN^9n0cFQOg9e0ZFYOP:>QFAP:VG]Ol8b0o1N2N1O2N2N2N101N2N2N1O2Nl\\f4"}, "label": "cyan blue motorcycle"}]} {"id": 469293, "image": "COCO_train2014_000000469293.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"empty chair back\"."}], "task": "refering", "answer_template": "The \"empty chair back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 183, 203, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275, 297, 298], "bbox": [0.8338749999999999, 0.41075, 1.0, 0.7268749999999999], "iscrowd": 0, "area": 10655.252150000004, "rle": {"size": [480, 640], "counts": "U`j72n>O18I1N3M2N2N3M2N4L3M4M3L3M3M2N3M2N2N3M2N2O2M2N2N2N3M2N2N2N2O1N1O2N2N2N1O4L5K5K5L4K4L5K5K5K4L001O010O001O001O001O001O1O010O001O000000001O000001O000001O0000000000010O00000001O0001O0001O000000010O7I5K3M2N3[M^Ef0d:SOkEd0Y:SO\\G\\Om86XGJS1"}, "label": "empty chair back"}]} {"id": 469293, "image": "COCO_train2014_000000469293.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown chair behind an old asian woman\"."}], "task": "refering", "answer_template": "The \"a brown chair behind an old asian woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 182, 183, 203, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275, 297, 298], "bbox": [0.8338749999999999, 0.41075, 1.0, 0.7268749999999999], "iscrowd": 0, "area": 10655.252150000004, "rle": {"size": [480, 640], "counts": "U`j72n>O18I1N3M2N2N3M2N4L3M4M3L3M3M2N3M2N2N3M2N2O2M2N2N2N3M2N2N2N2O1N1O2N2N2N1O4L5K5K5L4K4L5K5K5K4L001O010O001O001O001O001O1O010O001O000000001O000001O000001O0000000000010O00000001O0001O0001O000000010O7I5K3M2N3[M^Ef0d:SOkEd0Y:SO\\G\\Om86XGJS1"}, "label": "a brown chair behind an old asian woman"}]} {"id": 100667, "image": "COCO_train2014_000000100667.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a banana on the right\"."}], "task": "refering", "answer_template": "The \"a banana on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 89, 107, 108, 109, 110, 111, 112, 129, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 248, 249, 250, 265, 266, 267, 270, 271, 272, 273, 293, 294, 295], "bbox": [0.5561875, 0.19341666666666668, 0.9168593749999999, 0.7641875000000001], "iscrowd": 0, "area": 33353.44874999999, "rle": {"size": [480, 640], "counts": "jUW52l>5K3UNHjD in this image.", "objects": [{"patches": [87, 88, 89, 107, 108, 109, 110, 111, 112, 129, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 181, 197, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 248, 249, 250, 265, 266, 267, 270, 271, 272, 273, 293, 294, 295], "bbox": [0.5561875, 0.19341666666666668, 0.9168593749999999, 0.7641875000000001], "iscrowd": 0, "area": 33353.44874999999, "rle": {"size": [480, 640], "counts": "jUW52l>5K3UNHjD in this image.", "objects": [{"patches": [11, 12, 13, 14, 34, 35, 36, 37, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 173, 174, 308, 309, 310, 330, 331, 332, 351, 352, 353, 354, 355, 373, 374, 375, 376, 377, 378], "bbox": [0.23490625, 0.018104166666666664, 0.677859375, 0.9756041666666667], "iscrowd": 0, "area": 29709.923100000004, "rle": {"size": [480, 640], "counts": "cWW21n>1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O1O100O100O1O100N200O100O100O100O1O100O100O100O100O100O100O100O100O1O100O100O100PHUNj2l1hLeNU3[1hLkNT3V1jLoNS3Q1kLSOR3n0mLVOP3j0oLXOo2i0PMYOo2g0PM[On2f0QM\\On2d0QM^Om2c0RM_Ol2b0SM_Om2a0RMAl2`0SMBl2>TMCj2>UMDj2 in this image.", "objects": [{"patches": [11, 12, 13, 14, 34, 35, 36, 37, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 101, 102, 103, 104, 105, 106, 107, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 173, 174, 308, 309, 310, 330, 331, 332, 351, 352, 353, 354, 355, 373, 374, 375, 376, 377, 378], "bbox": [0.23490625, 0.018104166666666664, 0.677859375, 0.9756041666666667], "iscrowd": 0, "area": 29709.923100000004, "rle": {"size": [480, 640], "counts": "cWW21n>1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O100O1O100O100O1O100O100O1O100O100O1O100N200O100O100O100O1O100O100O100O100O100O100O100O100O1O100O100O100PHUNj2l1hLeNU3[1hLkNT3V1jLoNS3Q1kLSOR3n0mLVOP3j0oLXOo2i0PMYOo2g0PM[On2f0QM\\On2d0QM^Om2c0RM_Ol2b0SM_Om2a0RMAl2`0SMBl2>TMCj2>UMDj2 in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 112, 113, 114, 132, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 201, 204, 205, 206, 224, 225, 227, 228, 229, 246, 247, 248, 250, 251, 252, 267, 268, 269, 270, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.44662499999999994, 0.0022500000000000003, 0.9994375, 0.9887708333333333], "iscrowd": 0, "area": 84004.20559999999, "rle": {"size": [480, 640], "counts": "d`V43j>3N3L3M4L4M2M3M3M3N2M3M3N2M3M3M3N2M3M3M3N1N3M3M3N2M3M3M3N2M3M3M3N2O1O1O1O1O1O1O1O1O1O1O100O10000O100O10000O100O10000O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O10000O1000000O1000000O1000000eG_KV6b4]ITLZ6l3bI`LV6`3fIlLR6T3jIYMl5h2QJcMg5]2UJ`NR5`1lJcNS5]1lJdNS5]1lJeNS5[1lJgNS5Y1\\J\\K1^3a5W1^JaKJZ3f5V1`JYO^5h0bJXO]5i0cJXO[5i0eJXOY5i0gJXOV5i0kJXOS5i0mJWOR5j0nJWOP5j0PKWOn4j0RKWOl4j0TKWOj4j0VKWOh4j0XKVOg4k0YKVOd4l0\\KUOb4l0^KUO`4l0`KUO^4l0bKUO\\4l0dKTO[4m0eKSOZ4n0fKQOZ4P1fKPOX4R1hKnNW4S1iKmNV4T1jKlNU4U1kKkNT4V1lKjNS4W1mKiNR4Y1mKfNS4[1mKeNR4\\1nKcNQ4_1oKaNP3eKcMk5]O_NP3iK`Mi5@^NP3jK^Mi5A]NQ3lK\\Mh5C[NR3nKYMh5EZNQ3QLWMj1@d18`0R3SLTMf1Jc10d0Q3ULSMb12b1Jf0R3WLoL_1=`1Bj0Q3ZLmLZ1f0_1\\Ol0R3\\LjLV1l0b1XOl0Q3^LiLR1Q1d1UOk0R3`LfLn0W1g1QOk0Q3TNRNQ1mNj0R3PNVNV1hNj0W3fMVN`1cNi0^3]MSNj1_Nh0d3TMRNT2ZNf0l3lLoM]2UNf0S4bLmMh2PNd0[4ZLiMR3lMc0S7]OmHa0U7_OkH`0V7@jH>X7BhH=Y7CgH;[7EeH:]7EcH:^7FbH9_7GaH7`7J`H5`7L`H3a7M_H2a7O_H0a71_HNa73_HLa75_HJa77_HGc79]HFc7;]HDc7=]HBc7?\\HAd7`0\\H_Oe7a0[H^Oe7c0[H\\Oe7e0[HZOe7g0[HYOd7h0\\HXO`2QMX2h3XKWO_2SMX2f3YKWO^2UMW2e3[KVO^2VMU2e3]KUO]2XMT2d3_KTO\\2ZMS2c3aKSO[2\\MR2b3cKSOY2]MS2`3dKSOX2_MR2_3fKSOV2`MR2^3hKSOT2aMR2]3jKROS2cMQ2\\3lKROP2eMQ2[3oKQOm1gMR2Y3QLQOj1iMR2X3TLoNi1kMQ2W3VLoNf1mMR2U3XLoNc1oMR2T3[LmNa1RNR2R3]LmN^1TNR2Q3`LlN\\1UNR2P3bLkNZ1XNQ2o2eLjNW1ZNR2n2fLiNU1\\NR2m2iLgNS1_NR2k2kLgNQ1`NQ2k2nLfNn0bNR2i2PMeNl0eNQ2h2SMdNi0gNR2f2UMdNf0iNR2e2XMbNe0kNQ2d2ZMbNb0mNQ2c2]MaN?oNQ2b2_M`N>QOo1b2cM_N:WOj1]2lM_N6\\Od1X2VN^N2C^1R2`N]NNJX1l1jN\\NJ1R1f1TOa2b0bM^Oe28^MHi2OYM1k8000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000001O0000000000000Y7eBhN"}, "label": "a man in a green shirt eating a banana"}]} {"id": 100667, "image": "COCO_train2014_000000100667.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man eating banana in green shirt\"."}], "task": "refering", "answer_template": "The \"man eating banana in green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 112, 113, 114, 132, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 201, 204, 205, 206, 224, 225, 227, 228, 229, 246, 247, 248, 250, 251, 252, 267, 268, 269, 270, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.44662499999999994, 0.0022500000000000003, 0.9994375, 0.9887708333333333], "iscrowd": 0, "area": 84004.20559999999, "rle": {"size": [480, 640], "counts": "d`V43j>3N3L3M4L4M2M3M3M3N2M3M3N2M3M3M3N2M3M3M3N1N3M3M3N2M3M3M3N2M3M3M3N2O1O1O1O1O1O1O1O1O1O1O100O10000O100O10000O100O10000O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O10000O1000000O1000000O1000000eG_KV6b4]ITLZ6l3bI`LV6`3fIlLR6T3jIYMl5h2QJcMg5]2UJ`NR5`1lJcNS5]1lJdNS5]1lJeNS5[1lJgNS5Y1\\J\\K1^3a5W1^JaKJZ3f5V1`JYO^5h0bJXO]5i0cJXO[5i0eJXOY5i0gJXOV5i0kJXOS5i0mJWOR5j0nJWOP5j0PKWOn4j0RKWOl4j0TKWOj4j0VKWOh4j0XKVOg4k0YKVOd4l0\\KUOb4l0^KUO`4l0`KUO^4l0bKUO\\4l0dKTO[4m0eKSOZ4n0fKQOZ4P1fKPOX4R1hKnNW4S1iKmNV4T1jKlNU4U1kKkNT4V1lKjNS4W1mKiNR4Y1mKfNS4[1mKeNR4\\1nKcNQ4_1oKaNP3eKcMk5]O_NP3iK`Mi5@^NP3jK^Mi5A]NQ3lK\\Mh5C[NR3nKYMh5EZNQ3QLWMj1@d18`0R3SLTMf1Jc10d0Q3ULSMb12b1Jf0R3WLoL_1=`1Bj0Q3ZLmLZ1f0_1\\Ol0R3\\LjLV1l0b1XOl0Q3^LiLR1Q1d1UOk0R3`LfLn0W1g1QOk0Q3TNRNQ1mNj0R3PNVNV1hNj0W3fMVN`1cNi0^3]MSNj1_Nh0d3TMRNT2ZNf0l3lLoM]2UNf0S4bLmMh2PNd0[4ZLiMR3lMc0S7]OmHa0U7_OkH`0V7@jH>X7BhH=Y7CgH;[7EeH:]7EcH:^7FbH9_7GaH7`7J`H5`7L`H3a7M_H2a7O_H0a71_HNa73_HLa75_HJa77_HGc79]HFc7;]HDc7=]HBc7?\\HAd7`0\\H_Oe7a0[H^Oe7c0[H\\Oe7e0[HZOe7g0[HYOd7h0\\HXO`2QMX2h3XKWO_2SMX2f3YKWO^2UMW2e3[KVO^2VMU2e3]KUO]2XMT2d3_KTO\\2ZMS2c3aKSO[2\\MR2b3cKSOY2]MS2`3dKSOX2_MR2_3fKSOV2`MR2^3hKSOT2aMR2]3jKROS2cMQ2\\3lKROP2eMQ2[3oKQOm1gMR2Y3QLQOj1iMR2X3TLoNi1kMQ2W3VLoNf1mMR2U3XLoNc1oMR2T3[LmNa1RNR2R3]LmN^1TNR2Q3`LlN\\1UNR2P3bLkNZ1XNQ2o2eLjNW1ZNR2n2fLiNU1\\NR2m2iLgNS1_NR2k2kLgNQ1`NQ2k2nLfNn0bNR2i2PMeNl0eNQ2h2SMdNi0gNR2f2UMdNf0iNR2e2XMbNe0kNQ2d2ZMbNb0mNQ2c2]MaN?oNQ2b2_M`N>QOo1b2cM_N:WOj1]2lM_N6\\Od1X2VN^N2C^1R2`N]NNJX1l1jN\\NJ1R1f1TOa2b0bM^Oe28^MHi2OYM1k8000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000001O0000000000000Y7eBhN"}, "label": "man eating banana in green shirt"}]} {"id": 100667, "image": "COCO_train2014_000000100667.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on left bite banana with teeth\"."}], "task": "refering", "answer_template": "The \"man on left bite banana with teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 96, 97, 119, 120, 121, 122, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 191, 192, 208, 209, 210, 212, 213, 214, 215, 231, 232, 233, 254, 255, 256, 277, 278, 279, 300, 301, 302, 323, 324, 325, 347, 348], "bbox": [0.044125, 0.20941666666666667, 0.478234375, 0.9082291666666666], "iscrowd": 0, "area": 23012.371999999996, "rle": {"size": [480, 640], "counts": "c_=3f>;E:F;E;F:E;E9H8K5J5L5J6K4K6K5L3L5L3L4M4K4M3L4M4K4M3L3N2M4M2M3M3N2N3M2O1N2N2O1N3N1O1O100O2N1O1O1O1O1O2N1O1O1O101O0O100O2O0O[KiGc2W8ZMWH[2i7bMeHT2Y7iM]Je0c5XOTK4k4J_K0`4MfK0Z4MkK1T4NQLMQ40ULLk32[LJf33`LI`35eLH\\36iLFX37nLER3:RMCn2lFBS9?mFAS9>oFAP9`0PG_OP9b0PG^Oo8b0RG^Om8c0TG[Ol8f0TGZOl8e0UGZOl8j0PGVOP9Q1jFnNV9X1dFgN\\9\\1cFcN]9^1bFaN_9`1aF]Na9d1_FZNb9h1]FVNd9k1\\FSNe9n1f11O2M2O1O2N:F1O00000100O1O1O1O1O1O1O2N1MfDjMP:S2`E^Na:_1[EgNg:U1WEoNj:n0TEVOn:f0PE^OR;?jDFW;7gDM[;ObD6_;G_D=c;@ZDd0g;YOWDk0l;PORDT1Q in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 121, 122, 123, 138, 139, 140, 141, 142, 161, 162, 184, 185, 207, 208, 230, 231, 253, 254, 276, 277, 299, 322], "bbox": [0.0, 0.003958333333333333, 0.49265625, 0.8527083333333333], "iscrowd": 0, "area": 52741.94395, "rle": {"size": [480, 640], "counts": "2g in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 121, 122, 123, 138, 139, 140, 141, 142, 161, 162, 184, 185, 207, 208, 230, 231, 253, 254, 276, 277, 299, 322], "bbox": [0.0, 0.003958333333333333, 0.49265625, 0.8527083333333333], "iscrowd": 0, "area": 52741.94395, "rle": {"size": [480, 640], "counts": "2g in this image.", "objects": [{"patches": [21, 22, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.051546875000000006, 0.0, 1.0, 1.0], "iscrowd": 0, "area": 143676.44939999998, "rle": {"size": [480, 640], "counts": "om?1n>1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1O1000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000XNhIhJX6T5lIlJT6P5QJoJo5m4UJSKk5i4ZJVKf5f4_JYKa5c4cJ]K]5_4hJ`KX5\\4lJdKT5X4QKgKo4T4WKkKi4Q4[KoKe4m3`KRL`4i3eKWL[4e3jKZLV4a3PL^LP4^3TL_Lo3\\3WL\\LP4`3TLYLS4b3SLULU4g3PLQLW4j3nKoKY4m3_3K5L4L4K5L4K5L4K5L4K5L4N200O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O10Pd5"}, "label": "bench closest to camera"}]} {"id": 338242, "image": "COCO_train2014_000000338242.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bench that is first\"."}], "task": "refering", "answer_template": "The \"the bench that is first\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.051546875000000006, 0.0, 1.0, 1.0], "iscrowd": 0, "area": 143676.44939999998, "rle": {"size": [480, 640], "counts": "om?1n>1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1O1000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000XNhIhJX6T5lIlJT6P5QJoJo5m4UJSKk5i4ZJVKf5f4_JYKa5c4cJ]K]5_4hJ`KX5\\4lJdKT5X4QKgKo4T4WKkKi4Q4[KoKe4m3`KRL`4i3eKWL[4e3jKZLV4a3PL^LP4^3TL_Lo3\\3WL\\LP4`3TLYLS4b3SLULU4g3PLQLW4j3nKoKY4m3_3K5L4L4K5L4K5L4K5L4K5L4N200O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O10Pd5"}, "label": "the bench that is first"}]} {"id": 551244, "image": "COCO_train2014_000000551244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass with napkins and utensils inside of it sitting near a pizza\"."}], "task": "refering", "answer_template": "The \"a glass with napkins and utensils inside of it sitting near a pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 54, 55, 56, 57, 58, 69, 70, 71, 72, 73, 84, 85, 86, 87, 88, 99, 100, 101, 102, 103, 114, 115, 116, 117, 118, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 159, 160, 161, 162, 174, 175, 176, 177], "bbox": [0.6309579439252336, 0.11809375, 0.9387616822429907, 0.5189999999999999], "iscrowd": 0, "area": 25051.067699999996, "rle": {"size": [640, 428], "counts": "_mX5j0db0j0B>A?m@bMn;k2gCZMQ in this image.", "objects": [{"patches": [43, 54, 55, 56, 57, 58, 69, 70, 71, 72, 73, 84, 85, 86, 87, 88, 99, 100, 101, 102, 103, 114, 115, 116, 117, 118, 129, 130, 131, 132, 133, 144, 145, 146, 147, 148, 159, 160, 161, 162, 174, 175, 176, 177], "bbox": [0.6309579439252336, 0.11809375, 0.9387616822429907, 0.5189999999999999], "iscrowd": 0, "area": 25051.067699999996, "rle": {"size": [640, 428], "counts": "_mX5j0db0j0B>A?m@bMn;k2gCZMQ in this image.", "objects": [{"patches": [48, 49, 50, 51, 63, 64, 65, 66, 78, 79, 80, 81, 93, 94, 95, 96, 109, 110, 111, 124, 125, 126, 138, 139, 140, 141, 154, 155], "bbox": [0.22628504672897196, 0.14896875, 0.4461682242990654, 0.446984375], "iscrowd": 0, "area": 13960.692350000003, "rle": {"size": [640, 428], "counts": "Ygl1[1eb03L5K5L4K4L5L4L6I8I3M2N3XASMVS401N2O0000001O00001N10000000001O000O10000001O00000001O000000000000O100000000O2O00000O101O0O1O1O101N1O1O1O2O0O1N2hN`AnLc>a2TBUMo=g2[BPMh=l2h1L5L3L4M3M4K4L4L4K6K5K6J5I8GiPd4"}, "label": "a glass of juice is placed on the table near pizza"}]} {"id": 551244, "image": "COCO_train2014_000000551244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a coca - cola glass sitting on a table half filled with an orange drink\"."}], "task": "refering", "answer_template": "The \"a coca - cola glass sitting on a table half filled with an orange drink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 63, 64, 65, 66, 78, 79, 80, 81, 93, 94, 95, 96, 109, 110, 111, 124, 125, 126, 138, 139, 140, 141, 154, 155], "bbox": [0.22628504672897196, 0.14896875, 0.4461682242990654, 0.446984375], "iscrowd": 0, "area": 13960.692350000003, "rle": {"size": [640, 428], "counts": "Ygl1[1eb03L5K5L4K4L5L4L6I8I3M2N3XASMVS401N2O0000001O00001N10000000001O000O10000001O00000001O000000000000O100000000O2O00000O101O0O1O1O101N1O1O1O2O0O1N2hN`AnLc>a2TBUMo=g2[BPMh=l2h1L5L3L4M3M4K4L4L4K6K5K6J5I8GiPd4"}, "label": "a coca - cola glass sitting on a table half filled with an orange drink"}]} {"id": 2411, "image": "COCO_train2014_000000002411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white piece of cake\"."}], "task": "refering", "answer_template": "The \"the white piece of cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 119, 120, 121, 141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282, 302, 303, 304], "bbox": [0.036843749999999995, 0.2780334728033473, 0.28803124999999996, 0.7847489539748954], "iscrowd": 0, "area": 24803.099099999996, "rle": {"size": [478, 640], "counts": "o^;3l>1N3M3M3N1N3M3M3D in this image.", "objects": [{"patches": [97, 98, 119, 120, 121, 141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282, 302, 303, 304], "bbox": [0.036843749999999995, 0.2780334728033473, 0.28803124999999996, 0.7847489539748954], "iscrowd": 0, "area": 24803.099099999996, "rle": {"size": [478, 640], "counts": "o^;3l>1N3M3M3N1N3M3M3D in this image.", "objects": [{"patches": [36, 37, 38, 59, 60, 61, 62, 63, 64, 82, 83, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 111, 112, 128, 129, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 200, 201], "bbox": [0.5640468750000001, 0.09244769874476987, 0.898578125, 0.4882845188284519], "iscrowd": 0, "area": 25117.39015, "rle": {"size": [478, 640], "counts": "\\dX52i>4L4M3L4L4L4L5L3L4L4YNhNcE\\1]:dN_E`1`:`N^Ec1a:^N[Ef1d:ZNYEj1f:WNVEm1i:TNSEP2m:oMQET2n:o0O2N1O100O1O1O1O1O100O1O1O1O2N100O1O1O1O1O100O1O1O1O1O101N1O1O1O100O1O1O1O1O100O1O2N1O100O1O100O1O100O1O1O101N1O100O1O100O1O100O1O1O101N1O100O1O100O1O100O1O1O10O00001O00001O000O2O0000001O00001O0O2O00001O001O00001O0O101O001O00001O000O2O001N10001N101O0O101O0O101O0O2O000O2O001N10001N10001N101O0O101O0N2N3M2N3M3M3M2N3M3M3M3M3M2N3M3M3M3M3M2N3M3M3M3M2N3M3M3M3M3M2N3M3M3M3M3M2N3M3M3M3M3M2N3M3M3MRWn0"}, "label": "a piece of chocolate layer cake on a floral patterned dish"}]} {"id": 2411, "image": "COCO_train2014_000000002411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chocolate dessert cake on a plate\"."}], "task": "refering", "answer_template": "The \"chocolate dessert cake on a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 59, 60, 61, 62, 63, 64, 82, 83, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 111, 112, 128, 129, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 200, 201], "bbox": [0.5640468750000001, 0.09244769874476987, 0.898578125, 0.4882845188284519], "iscrowd": 0, "area": 25117.39015, "rle": {"size": [478, 640], "counts": "\\dX52i>4L4M3L4L4L4L5L3L4L4YNhNcE\\1]:dN_E`1`:`N^Ec1a:^N[Ef1d:ZNYEj1f:WNVEm1i:TNSEP2m:oMQET2n:o0O2N1O100O1O1O1O1O100O1O1O1O2N100O1O1O1O1O100O1O1O1O1O101N1O1O1O100O1O1O1O1O100O1O2N1O100O1O100O1O100O1O1O101N1O100O1O100O1O100O1O1O101N1O100O1O100O1O100O1O1O10O00001O00001O000O2O0000001O00001O0O2O00001O001O00001O0O101O001O00001O000O2O001N10001N101O0O101O0O101O0O2O000O2O001N10001N10001N101O0O101O0N2N3M2N3M3M3M2N3M3M3M3M3M2N3M3M3M3M3M2N3M3M3M3M2N3M3M3M3M3M2N3M3M3M3M3M2N3M3M3M3M3M2N3M3M3MRWn0"}, "label": "chocolate dessert cake on a plate"}]} {"id": 133486, "image": "COCO_train2014_000000133486.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing over a counter making food\"."}], "task": "refering", "answer_template": "The \"a man standing over a counter making food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 22, 23, 24, 25, 26, 27, 36, 37, 38, 39, 40, 41, 42, 43, 51, 52, 53, 54, 55, 56, 57, 58, 59, 67, 68, 69, 70, 71, 72, 73, 74, 82, 83, 84, 85, 86, 87, 88, 89, 98, 99, 100, 101, 102, 103, 104, 113, 114, 115, 116, 117, 118, 119, 131, 132, 133, 134, 146, 147, 148, 149, 160, 161, 162, 163, 164, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 203, 204, 205, 206, 207, 208, 209, 216, 217, 218, 219, 220, 221, 222, 223, 224, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 294, 295, 296, 297, 298, 299, 308, 309, 310, 311, 312, 313, 314, 322, 323, 324, 325, 326, 327, 328, 329, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.32334894613583137, 0.013109375000000001, 1.0, 0.9794062499999999], "iscrowd": 0, "area": 108956.55635000001, "rle": {"size": [640, 427], "counts": "bff2`0bb0n0M3O10001O00000000001O0O1000001O00000000001N10000000001O0O1N2O1N3M2N2N2N2O1N3M2N2N2N2O1N3N1O100SFWMc2i2[M\\Ma2f2]M]Ma2c2^MaM_2_2`MdM]2]2bMfM\\2Z2cMjMZ2V2eMmMX2U2fMoMW2Q2hMRNV2n1iMVNS2k1lMXNR2h1mM[NQ2e1nM_Nn1c1PN`Nn1`1QNdNl1\\1RNhNk1Y1TNkNi1U1VNnNh1S1VNPOg1Q1XNSOe1m0ZNVOc1k0\\NYOa1g0^N\\O`1d0_N@]1b0aNA]1?bNE[1;dNHY19`LhLUNR3Y56aLiLUNU3W53aLjLVNW3V50bLkLWNX3U5McLlLWNY3U5KbLnLXNZ3S5IcLoLYN[3R5FdLPMYN]3Q5DcLQM[N]3P5CcLRM\\N^3o4@dLSM\\N`3n4]OdLUM]Na3l4[OeLVM^Na3l4YOdLXM_Nb3k4WOdLXM`Nd3i4UOeLYMaNe3h4ROeL[MbNe3h4POeL\\MaNh3g4mNfL]MbNi3f4kNeL^MdNi3e4jNfL^MdNk3d4gNfL`MeNl3c4dNfLbMfNm3a4bNhLbMfNn3a4`NgLdMgNn3a4_NeLeMiNn3`4^NfLeMiNn3a4]NdLgMjNm3b4\\NbLiMkNm3a4[NcLiMkNm3b4ZNaLkMlNl3c4ZN^LlMmNm3c4XN^LmMnNl3d4WN]LnMnNm3d4UN\\LPNoNl3d4UN[LQNQOj3d4VNYLQNROk3d4TNXLSNTOi3c4UNWLUNUOf3d4UNVLVNVOf3c4TNULXNXOd3b4UNTLYNZOc3a4UNSLYN[Oc3b4TNQL[N]Oa3a4UNPL\\N_O`3`4TNPL]N@_3_4UNoK^NB]3_4UNmK`ND\\3^4UNlK`NF[3]4VNkKaNGZ3\\4WNkKaNIY3Z4WNkKbNKW3X4YNlKaNLW3U4[NlK`NOU3S4]NlK`N1S3R4^NlK`N2S3o3_NmK`N3R3n3`NmK`N5P3l3bNnK_N6P3j3cNmK_N9n2h3eNmK_N;m2e3fNoK^N in this image.", "objects": [{"patches": [8, 9, 10, 11, 22, 23, 24, 25, 26, 27, 36, 37, 38, 39, 40, 41, 42, 43, 51, 52, 53, 54, 55, 56, 57, 58, 59, 67, 68, 69, 70, 71, 72, 73, 74, 82, 83, 84, 85, 86, 87, 88, 89, 98, 99, 100, 101, 102, 103, 104, 113, 114, 115, 116, 117, 118, 119, 131, 132, 133, 134, 146, 147, 148, 149, 160, 161, 162, 163, 164, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 203, 204, 205, 206, 207, 208, 209, 216, 217, 218, 219, 220, 221, 222, 223, 224, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 294, 295, 296, 297, 298, 299, 308, 309, 310, 311, 312, 313, 314, 322, 323, 324, 325, 326, 327, 328, 329, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.32334894613583137, 0.013109375000000001, 1.0, 0.9794062499999999], "iscrowd": 0, "area": 108956.55635000001, "rle": {"size": [640, 427], "counts": "bff2`0bb0n0M3O10001O00000000001O0O1000001O00000000001N10000000001O0O1N2O1N3M2N2N2N2O1N3M2N2N2N2O1N3N1O100SFWMc2i2[M\\Ma2f2]M]Ma2c2^MaM_2_2`MdM]2]2bMfM\\2Z2cMjMZ2V2eMmMX2U2fMoMW2Q2hMRNV2n1iMVNS2k1lMXNR2h1mM[NQ2e1nM_Nn1c1PN`Nn1`1QNdNl1\\1RNhNk1Y1TNkNi1U1VNnNh1S1VNPOg1Q1XNSOe1m0ZNVOc1k0\\NYOa1g0^N\\O`1d0_N@]1b0aNA]1?bNE[1;dNHY19`LhLUNR3Y56aLiLUNU3W53aLjLVNW3V50bLkLWNX3U5McLlLWNY3U5KbLnLXNZ3S5IcLoLYN[3R5FdLPMYN]3Q5DcLQM[N]3P5CcLRM\\N^3o4@dLSM\\N`3n4]OdLUM]Na3l4[OeLVM^Na3l4YOdLXM_Nb3k4WOdLXM`Nd3i4UOeLYMaNe3h4ROeL[MbNe3h4POeL\\MaNh3g4mNfL]MbNi3f4kNeL^MdNi3e4jNfL^MdNk3d4gNfL`MeNl3c4dNfLbMfNm3a4bNhLbMfNn3a4`NgLdMgNn3a4_NeLeMiNn3`4^NfLeMiNn3a4]NdLgMjNm3b4\\NbLiMkNm3a4[NcLiMkNm3b4ZNaLkMlNl3c4ZN^LlMmNm3c4XN^LmMnNl3d4WN]LnMnNm3d4UN\\LPNoNl3d4UN[LQNQOj3d4VNYLQNROk3d4TNXLSNTOi3c4UNWLUNUOf3d4UNVLVNVOf3c4TNULXNXOd3b4UNTLYNZOc3a4UNSLYN[Oc3b4TNQL[N]Oa3a4UNPL\\N_O`3`4TNPL]N@_3_4UNoK^NB]3_4UNmK`ND\\3^4UNlK`NF[3]4VNkKaNGZ3\\4WNkKaNIY3Z4WNkKbNKW3X4YNlKaNLW3U4[NlK`NOU3S4]NlK`N1S3R4^NlK`N2S3o3_NmK`N3R3n3`NmK`N5P3l3bNnK_N6P3j3cNmK_N9n2h3eNmK_N;m2e3fNoK^N in this image.", "objects": [{"patches": [125, 126, 127, 147, 148, 149, 150, 170, 171, 172, 173, 174, 175, 176, 179, 180, 181, 185, 186, 187, 191, 192, 193, 194, 195, 196, 197, 198, 199, 201, 202, 203, 204, 208, 209, 210, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 246, 247, 248, 257, 258, 261, 262, 263, 264, 265, 266, 283, 284, 285, 286, 287, 288, 289, 290, 306, 307, 308, 309, 310, 311, 312, 313, 314, 329, 330, 331, 332, 333, 334, 335, 336, 337, 352, 353, 354, 355, 356, 357, 358, 359, 360, 375, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.053937500000000006, 0.3438125, 0.896625, 0.9887708333333334], "iscrowd": 0, "area": 65483.67590000001, "rle": {"size": [480, 640], "counts": "ld`05k>7I7I7I7I7J5J4L00O100000000O1000O1000O10000000O0100000000O10O100000O10000000O0100000000O10O100000O100000O10O100000000O010000000O1002N2M3N2N2N2N]OkB]OV=`0lB@X=9kBGX=4jBLY=NiB3Z=GiB9[=AgB?n=00000000000N2M3N0O0010O010O00010O01O010O01O01O010O0010O010O00010O01O010O010O00010O0010O01O010O01O01O010O0010O0010O0010O01O01O010O01O010O00010O001O010O00001O010O001O00010O001O010O1O1O2N100O1O1_F^Oc4c0\\Ib1_6`N_Id1^6\\NeGfNe1R3a6YNgGiNf1R3_6`N_Id1]6_NaIe1[6\\NcIh1Y6[NdIi1X6YNgIi1W6YNfIh1Z6YNdIh1[6ZNdIf1\\6[NbIf1^6ZNaIg1^6YNbIh1^6XNbIh1^6WNbIj1]6WNbIj1^6UNcIk1]6UNbIl1]6TNcIm1]6SNcIm1]6RNcIo1\\6QNdIP2\\6nMfIR2Z6kMhIV2W6gMlIZ2T6cMoI]2Q6_MRJb2m5\\MUJe2k5WMXJj2h5RM\\Jn2c5TM[Jm2e5UMXJl2h5UMWJk2h5XMUJi2k5XMSJi2m5XMRJh2m5ZMQJg2o5ZMoIg2Q6ZMnIf2d4bK]Kj1Me2f4dKZKh1Ne2g4gKXKe10d2g4kKUKc12c2i4mKRKa13c2j4PLPK^14c2k4SLlJ`15]2o4VLgJd14W2T5YLcJg13Q2Y5\\L_Jk12i1_5_LZJo11c1d5a4O100O1O100O1O100O1O1O100O1O100O1O100O1001O1O001O001O1O001O1O001O001O1O001O001O1O00O100O1O100O100O1O100O1O100O100O1O100O100O1000000001O000000000000001oGlJT6T5kInJT6R5kIPKT6P5lIRKR6n4mITKR6l4mIVKR6j4mIVKT6j4lIVKT6j4kIVKV6k4hIVKX6j4gIVKZ6j4eIWK[6i4eIVK\\6j4cIWK]6i4bIWK_6m4YIWKg6m4QIVKP7o4fHVKZ7n4^HUKc7i5]HYIc7g6;1O00001O001O00001OO1N2O1N2N2O1N2O1NYJUJ]2j5dMWJ[2i5eMYJY2g5gMZJX2f5hM[JW2e5iM]JU2d5kM\\JT2d5lM]JS2c5mM_JQ2a5oM`JP2`5PNaJo1_5QNcJm1^5RNcJm1]5SNdJl1\\5TNfJj1Z5WNfJh1Z5XNgJg1Y5ZNhJd1Y5[NhJd1X5]NhJb1X5^NjJ`1V5aNjJ^1V5bNkJ]1U5dNlJZ1U5fNkJY1U5gNlJX1T5iNmJU1S5kNnJT1R5mNnJR1R5nNPKP1Q5POoJo0Q5oNRKP1n4oNUKo0k4^N\\I[Nk1V3j4^N`IUNj1[3f4`NdIPNh1_3e4_NlKa1T4^NnK`1S4_NnK`1R4`NoK_1R4`NPL^1Q4`NQL_1P4`NQL_1o3aNSL]1n3bNSL\\NnNl0P5g0aNoN`1o0kNgNU1Y1UO]Nl0b1Z6O010O1O10O01O100O00100O10OgNkBg0h=O010O0010O010O01O010O0010O010O01O010O1O010O010O0010O01O010O0010O010O01O010O0010O010O01O010O0010O010O01O01POZOeCe0X<_OgCa0VQjCAZ<^14L4M3L4M3L5K4M3N20010O00000001O01O0001O000001O01O000000010O0000000010O00000001O01O000001OM3M4L3M3M3M3N2M4L3M3M3M3M3M4M2M3MWgn0"}, "label": "the motorcycle closest to the camera"}]} {"id": 182642, "image": "COCO_train2014_000000182642.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the motorcycle handlebars and gas tank in the front\"."}], "task": "refering", "answer_template": "The \"the motorcycle handlebars and gas tank in the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 147, 148, 149, 150, 170, 171, 172, 173, 174, 175, 176, 179, 180, 181, 185, 186, 187, 191, 192, 193, 194, 195, 196, 197, 198, 199, 201, 202, 203, 204, 208, 209, 210, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 246, 247, 248, 257, 258, 261, 262, 263, 264, 265, 266, 283, 284, 285, 286, 287, 288, 289, 290, 306, 307, 308, 309, 310, 311, 312, 313, 314, 329, 330, 331, 332, 333, 334, 335, 336, 337, 352, 353, 354, 355, 356, 357, 358, 359, 360, 375, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.053937500000000006, 0.3438125, 0.896625, 0.9887708333333334], "iscrowd": 0, "area": 65483.67590000001, "rle": {"size": [480, 640], "counts": "ld`05k>7I7I7I7I7J5J4L00O100000000O1000O1000O10000000O0100000000O10O100000O10000000O0100000000O10O100000O100000O10O100000000O010000000O1002N2M3N2N2N2N]OkB]OV=`0lB@X=9kBGX=4jBLY=NiB3Z=GiB9[=AgB?n=00000000000N2M3N0O0010O010O00010O01O010O01O01O010O0010O010O00010O01O010O010O00010O0010O01O010O01O01O010O0010O0010O0010O01O01O010O01O010O00010O001O010O00001O010O001O00010O001O010O1O1O2N100O1O1_F^Oc4c0\\Ib1_6`N_Id1^6\\NeGfNe1R3a6YNgGiNf1R3_6`N_Id1]6_NaIe1[6\\NcIh1Y6[NdIi1X6YNgIi1W6YNfIh1Z6YNdIh1[6ZNdIf1\\6[NbIf1^6ZNaIg1^6YNbIh1^6XNbIh1^6WNbIj1]6WNbIj1^6UNcIk1]6UNbIl1]6TNcIm1]6SNcIm1]6RNcIo1\\6QNdIP2\\6nMfIR2Z6kMhIV2W6gMlIZ2T6cMoI]2Q6_MRJb2m5\\MUJe2k5WMXJj2h5RM\\Jn2c5TM[Jm2e5UMXJl2h5UMWJk2h5XMUJi2k5XMSJi2m5XMRJh2m5ZMQJg2o5ZMoIg2Q6ZMnIf2d4bK]Kj1Me2f4dKZKh1Ne2g4gKXKe10d2g4kKUKc12c2i4mKRKa13c2j4PLPK^14c2k4SLlJ`15]2o4VLgJd14W2T5YLcJg13Q2Y5\\L_Jk12i1_5_LZJo11c1d5a4O100O1O100O1O100O1O1O100O1O100O1O100O1001O1O001O001O1O001O1O001O001O1O001O001O1O00O100O1O100O100O1O100O1O100O100O1O100O100O1000000001O000000000000001oGlJT6T5kInJT6R5kIPKT6P5lIRKR6n4mITKR6l4mIVKR6j4mIVKT6j4lIVKT6j4kIVKV6k4hIVKX6j4gIVKZ6j4eIWK[6i4eIVK\\6j4cIWK]6i4bIWK_6m4YIWKg6m4QIVKP7o4fHVKZ7n4^HUKc7i5]HYIc7g6;1O00001O001O00001OO1N2O1N2N2O1N2O1NYJUJ]2j5dMWJ[2i5eMYJY2g5gMZJX2f5hM[JW2e5iM]JU2d5kM\\JT2d5lM]JS2c5mM_JQ2a5oM`JP2`5PNaJo1_5QNcJm1^5RNcJm1]5SNdJl1\\5TNfJj1Z5WNfJh1Z5XNgJg1Y5ZNhJd1Y5[NhJd1X5]NhJb1X5^NjJ`1V5aNjJ^1V5bNkJ]1U5dNlJZ1U5fNkJY1U5gNlJX1T5iNmJU1S5kNnJT1R5mNnJR1R5nNPKP1Q5POoJo0Q5oNRKP1n4oNUKo0k4^N\\I[Nk1V3j4^N`IUNj1[3f4`NdIPNh1_3e4_NlKa1T4^NnK`1S4_NnK`1R4`NoK_1R4`NPL^1Q4`NQL_1P4`NQL_1o3aNSL]1n3bNSL\\NnNl0P5g0aNoN`1o0kNgNU1Y1UO]Nl0b1Z6O010O1O10O01O100O00100O10OgNkBg0h=O010O0010O010O01O010O0010O010O01O010O1O010O010O0010O01O010O0010O010O01O010O0010O010O01O010O0010O010O01O01POZOeCe0X<_OgCa0VQjCAZ<^14L4M3L4M3L5K4M3N20010O00000001O01O0001O000001O01O000000010O0000000010O00000001O01O000001OM3M4L3M3M3M3N2M4L3M3M3M3M3M4M2M3MWgn0"}, "label": "the motorcycle handlebars and gas tank in the front"}]} {"id": 182642, "image": "COCO_train2014_000000182642.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the backside of the bike with a rider on it\"."}], "task": "refering", "answer_template": "The \"the backside of the bike with a rider on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 55, 56, 57, 58, 59, 60, 78, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 124, 127, 128, 129, 146, 151, 169], "bbox": [0.35225, 0.056187499999999994, 0.62528125, 0.46291666666666664], "iscrowd": 0, "area": 15070.478800000003, "rle": {"size": [480, 640], "counts": "XeY31n>6I7J6J6I3N0O2O0O2O001N101N101N101OM2M4L3L5L3M4L2NO4L5K4L4L4L4L4O2N1O100O1O1O1O2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2N2O1N3M2N3O1O100O101O0O10001N10000O101N10000O101O0O10001N100O10001N10000O2O000000001O0000001O000000001O0000001O000000001O0000001O000000010000000O1000000000O10000000O100000000000O100001N3N3M2N2N3M2N3M2NN2L3M4L4L4L4M3L4L4L4L6J6J6J6J7I6J6[Oj]`3"}, "label": "the backside of the bike with a rider on it"}]} {"id": 182642, "image": "COCO_train2014_000000182642.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back end of motorcycle\"."}], "task": "refering", "answer_template": "The \"back end of motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 55, 56, 57, 58, 59, 60, 78, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 124, 127, 128, 129, 146, 151, 169], "bbox": [0.35225, 0.056187499999999994, 0.62528125, 0.46291666666666664], "iscrowd": 0, "area": 15070.478800000003, "rle": {"size": [480, 640], "counts": "XeY31n>6I7J6J6I3N0O2O0O2O001N101N101N101OM2M4L3L5L3M4L2NO4L5K4L4L4L4L4O2N1O100O1O1O1O2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2N2O1N3M2N3O1O100O101O0O10001N10000O101N10000O101O0O10001N100O10001N10000O2O000000001O0000001O000000001O0000001O000000001O0000001O000000010000000O1000000000O10000000O100000000000O100001N3N3M2N2N3M2N3M2NN2L3M4L4L4L4M3L4L4L4L6J6J6J6J7I6J6[Oj]`3"}, "label": "back end of motorcycle"}]} {"id": 502134, "image": "COCO_train2014_000000502134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man on a motorcycle wearing a santa costume\"."}], "task": "refering", "answer_template": "The \"a man on a motorcycle wearing a santa costume\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 79, 80, 99, 100, 101, 102, 103, 122, 123, 124, 125, 145, 146, 147, 148, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 196, 197, 213, 214, 215, 216, 217, 218, 219, 220, 236, 237, 238, 239, 240, 242, 243, 260, 261, 262, 263, 283, 284, 285, 286, 307, 308, 309, 310, 330, 331, 332, 333], "bbox": [0.27135937499999996, 0.20454332552693208, 0.6034062499999999, 0.9979156908665106], "iscrowd": 0, "area": 36278.42410000002, "rle": {"size": [427, 640], "counts": "fiX28k<9N2M4K4L4TDVOS;m0gDXOV;a1L4L3M4L3N3L3M3M4L3M4L3N3L4L5K4L4L4K5M3M3M3M4M2M4L4M2RObKkHb4k0XKb4=aJ^4h0XKa4a0dJZ4g0YK^4d0iJV4d0YK]4k0hJQ4f0XK`4V6\\KmIc4V6XKnIf4U6TKPJj4R6RKQJm4R6lJTJR5S7O2N1O1O3M4M2M4L4M3L4L7J7H3N1O2M2O1N3N1O1N3N2N1O2N1O1O0000000000000000001O00000O100O1000000O101O000000000000000000001O000O1001\\GYMb6g2QIhMm6Y2QIiMn6W2RIjMm6W2RIjMm6V2RIlMm6U2RIlMn6S2QIPNm6Q2RIPNm6P2RIRNm6o1RISNl6n1SIZNl2lNAj2bM`Nf2kNEf2dMcNb2jNHc2fMgN\\2kNK_2gMkNY2iNN\\2iMoNS2lNOV2lMQOQ2PONo1QNTOl1RO1j1RNVOe1VOQNhN\\1l2mNXOb1XOoMkN`1f2nNXOZOgMn1`1UNROc1_2oNZOXOgMQ2_1oMXOg1Y2oN[OWOgMT2`2C>QOEZ10CZOA_18iN?FZOa1h0QN6;UOc1f4[N\\Kd1f4YN]Kf1d4VN_Kj1Q801O001N100010O01O000010O10O01O010O10O01O010O10O01O010O10O001O001O00001O0000001O000000001O00000000001O0N2N2N2M3N2N2N2N200O2N100O100O1WOkC;W in this image.", "objects": [{"patches": [76, 77, 78, 79, 80, 99, 100, 101, 102, 103, 122, 123, 124, 125, 145, 146, 147, 148, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 196, 197, 213, 214, 215, 216, 217, 218, 219, 220, 236, 237, 238, 239, 240, 242, 243, 260, 261, 262, 263, 283, 284, 285, 286, 307, 308, 309, 310, 330, 331, 332, 333], "bbox": [0.27135937499999996, 0.20454332552693208, 0.6034062499999999, 0.9979156908665106], "iscrowd": 0, "area": 36278.42410000002, "rle": {"size": [427, 640], "counts": "fiX28k<9N2M4K4L4TDVOS;m0gDXOV;a1L4L3M4L3N3L3M3M4L3M4L3N3L4L5K4L4L4K5M3M3M3M4M2M4L4M2RObKkHb4k0XKb4=aJ^4h0XKa4a0dJZ4g0YK^4d0iJV4d0YK]4k0hJQ4f0XK`4V6\\KmIc4V6XKnIf4U6TKPJj4R6RKQJm4R6lJTJR5S7O2N1O1O3M4M2M4L4M3L4L7J7H3N1O2M2O1N3N1O1N3N2N1O2N1O1O0000000000000000001O00000O100O1000000O101O000000000000000000001O000O1001\\GYMb6g2QIhMm6Y2QIiMn6W2RIjMm6W2RIjMm6V2RIlMm6U2RIlMn6S2QIPNm6Q2RIPNm6P2RIRNm6o1RISNl6n1SIZNl2lNAj2bM`Nf2kNEf2dMcNb2jNHc2fMgN\\2kNK_2gMkNY2iNN\\2iMoNS2lNOV2lMQOQ2PONo1QNTOl1RO1j1RNVOe1VOQNhN\\1l2mNXOb1XOoMkN`1f2nNXOZOgMn1`1UNROc1_2oNZOXOgMQ2_1oMXOg1Y2oN[OWOgMT2`2C>QOEZ10CZOA_18iN?FZOa1h0QN6;UOc1f4[N\\Kd1f4YN]Kf1d4VN_Kj1Q801O001N100010O01O000010O10O01O010O10O01O010O10O01O010O10O001O001O00001O0000001O000000001O00000000001O0N2N2N2M3N2N2N2N200O2N100O100O1WOkC;W\"."}], "task": "refering", "answer_template": "The \"a lady wearing a mrs . claus suit sitting on the back of a motorcycle with a helmet on >\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 70, 71, 72, 73, 93, 94, 95, 96, 116, 117, 118, 119, 139, 140, 141, 142, 163, 164, 165, 186, 187, 188, 209, 210, 211, 212, 232, 233, 234, 235, 236, 237, 257, 258, 259, 260, 282, 283, 305, 306, 328, 329], "bbox": [0.07196875, 0.12807962529274003, 0.32984375, 0.9820140515222482], "iscrowd": 0, "area": 23036.559000000012, "rle": {"size": [427, 640], "counts": "WYc0>i<7I6J7H7jDPOk9U1kE]Oj9f0mEEm9>kEJR:j1M3M4L3M3M3M3M4M2M3M2N3M2N2N3M2O1N3N1kIZKb3h4[LZKd3i4YLYKf3h4WLZKh3h4`K`KgNJh5i4]KnKc4S4YKQLe4Q4XKQLh4P4TKSLk4o3SKRLl4P4RKQLn4o3QKSLm4o3QKRLn4P4PKQLP5P4nJQLQ5Q4mJPLS5Q4kJQLS5Q4kJPLT5U4gJlKY5S6O2O00000O101O000O10001O0O10001O0O1000001N1000000010O2YKiJV1X5fNoJV1R5gNSKV1o4eNWKY1i4dNZK[1g4aN]K^1e4WNeKi1\\4iMQLV2Q4aMWL^2k3ZM\\Le2f3SMaLm2`3lLfLS3]3dLkLZ3d6O1SMeFi1[9TNkFh1V9UNPGg1Q9VNVGe1k8XNZGf1g8VN]Gh1i8PN[Gn1j8kMYGU2P:N2N2N101N2N1O2O0O1O1O100O1O100O1O1O100O001O10O01O1O010O1O10O01O01jFbN^6]1`IeN`6\\1VImNk6R1kHYOT7h0bHA_7>fG=Z8Z21O00001O001O001O001O001O1O1O2N1O1O1O1O1O1O1O1O1O1O3M2N2N3RNm1H9H"}]} {"id": 502134, "image": "COCO_train2014_000000502134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a female in a santa claus costume riding as passenger behind a santa - clad man on a motorcycle\"."}], "task": "refering", "answer_template": "The \"a female in a santa claus costume riding as passenger behind a santa - clad man on a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 70, 71, 72, 73, 93, 94, 95, 96, 116, 117, 118, 119, 139, 140, 141, 142, 163, 164, 165, 186, 187, 188, 209, 210, 211, 212, 232, 233, 234, 235, 236, 237, 257, 258, 259, 260, 282, 283, 305, 306, 328, 329], "bbox": [0.07196875, 0.12807962529274003, 0.32984375, 0.9820140515222482], "iscrowd": 0, "area": 23036.559000000012, "rle": {"size": [427, 640], "counts": "WYc0>i<7I6J7H7jDPOk9U1kE]Oj9f0mEEm9>kEJR:j1M3M4L3M3M3M3M4M2M3M2N3M2N2N3M2O1N3N1kIZKb3h4[LZKd3i4YLYKf3h4WLZKh3h4`K`KgNJh5i4]KnKc4S4YKQLe4Q4XKQLh4P4TKSLk4o3SKRLl4P4RKQLn4o3QKSLm4o3QKRLn4P4PKQLP5P4nJQLQ5Q4mJPLS5Q4kJQLS5Q4kJPLT5U4gJlKY5S6O2O00000O101O000O10001O0O10001O0O1000001N1000000010O2YKiJV1X5fNoJV1R5gNSKV1o4eNWKY1i4dNZK[1g4aN]K^1e4WNeKi1\\4iMQLV2Q4aMWL^2k3ZM\\Le2f3SMaLm2`3lLfLS3]3dLkLZ3d6O1SMeFi1[9TNkFh1V9UNPGg1Q9VNVGe1k8XNZGf1g8VN]Gh1i8PN[Gn1j8kMYGU2P:N2N2N101N2N1O2O0O1O1O100O1O100O1O1O100O001O10O01O1O010O1O10O01O01jFbN^6]1`IeN`6\\1VImNk6R1kHYOT7h0bHA_7>fG=Z8Z21O00001O001O001O001O001O1O1O2N1O1O1O1O1O1O1O1O1O1O3M2N2N3RNm1H9H in this image.", "objects": [{"patches": [89, 112, 113, 134, 135, 136, 157, 158, 159, 180, 181, 182, 203, 204, 205, 226, 227, 228, 250, 251, 273, 274, 297], "bbox": [0.844109375, 0.25618266978922716, 0.9460625, 0.8202341920374707], "iscrowd": 0, "area": 10452.011949999998, "rle": {"size": [427, 640], "counts": "l\\Q74U;d0oEGi9j0bECU:P2I7H9H7K6I4M1N2O2M2O1@`0J6K6K4K5L4O1N3N1N2O1O1N01G8H9G9G8108G9H8G9I7I7H8I7I7I7IO10O10000000O012N5K5K3M0O2L3]M`JXMb5b2gJYM\\5`2lJ\\MV5^2SK]MQ5[2XK`Mm4W2\\KdMi4S2`KhMe4o1dKlMa4k1]a>"}, "label": "a woman taking a picture with a light blue phone"}]} {"id": 502134, "image": "COCO_train2014_000000502134.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a santa hat holding up her blue phone\"."}], "task": "refering", "answer_template": "The \"a woman wearing a santa hat holding up her blue phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 112, 113, 134, 135, 136, 157, 158, 159, 180, 181, 182, 203, 204, 205, 226, 227, 228, 250, 251, 273, 274, 297], "bbox": [0.844109375, 0.25618266978922716, 0.9460625, 0.8202341920374707], "iscrowd": 0, "area": 10452.011949999998, "rle": {"size": [427, 640], "counts": "l\\Q74U;d0oEGi9j0bECU:P2I7H9H7K6I4M1N2O2M2O1@`0J6K6K4K5L4O1N3N1N2O1O1N01G8H9G9G8108G9H8G9I7I7H8I7I7I7IO10O10000000O012N5K5K3M0O2L3]M`JXMb5b2gJYM\\5`2lJ\\MV5^2SK]MQ5[2XK`Mm4W2\\KdMi4S2`KhMe4o1dKlMa4k1]a>"}, "label": "a woman wearing a santa hat holding up her blue phone"}]} {"id": 428407, "image": "COCO_train2014_000000428407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bunch of apples in a sack\"."}], "task": "refering", "answer_template": "The \"a bunch of apples in a sack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 79, 96, 97, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331, 350, 351, 352, 353, 354, 374, 375], "bbox": [0.082578125, 0.18839583333333335, 0.588203125, 0.9659166666666666], "iscrowd": 0, "area": 79320.07475000001, "rle": {"size": [480, 640], "counts": "nSi02k>4L4L4L4L4L4L3M4L4L4L5K4K5L4L4L4L5K4L4L4L4L5K4L4L4L4M3L5L3M3L4M3M4K4M3L4M3M3L5L3M3L4M3L4M4L3L4M2N3L4M3L4M3M2O2N2N2O1N2N101N2O1N2N2O1N101N2N2O1N2O2M7I8I6I8I6I7I8hIWGk5P9N3N2M3M3N2M3M3N2M2N3N2M3M3M3N2M3M2N3N2M1O1O100O1O1O100O1000000O10000000000O100000000O10002N6J3L1000O100O010O100O1000O0100O100O010O1000O0100O10O010000O010O100O010O1000O0100O10O0100O10O10O2O0O100O10000O100O100O101N10000O2O0O101N101N10001N100O2O0O2O000O2O0O101CYK^Fh4a9YK_Fh4_9ZK_Fg4`9ZK`Ff4_9[KaFf4\\9\\KcFe4\\9\\KdFe4Z9\\KeFe4Z9\\KfFd4Y9]KfFe4X9\\KhFd4W9^KgFd4W9]KiFc4V9^KjFc4T9^KkFc4T9^KlFb4S9_KlFc4R9^KnFb4Q9h0N1O2N1O2N1N2O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2N1O1N2O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1N3L3N2N3M2N3M2N3M3M3K5_Oa0_Oa0J5L5L4K5L4K5L4K5K5L4K5L3L5K5K5K4M4K5K5L4M2N3M3M3M2N3M3M3M3M2N3M3M3M5K6J7I7I7I`bk3"}, "label": "a bunch of apples in a sack"}]} {"id": 428407, "image": "COCO_train2014_000000428407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bag of green apples\"."}], "task": "refering", "answer_template": "The \"a bag of green apples\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 79, 96, 97, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 303, 304, 305, 306, 307, 308, 309, 326, 327, 328, 329, 330, 331, 350, 351, 352, 353, 354, 374, 375], "bbox": [0.082578125, 0.18839583333333335, 0.588203125, 0.9659166666666666], "iscrowd": 0, "area": 79320.07475000001, "rle": {"size": [480, 640], "counts": "nSi02k>4L4L4L4L4L4L3M4L4L4L5K4K5L4L4L4L5K4L4L4L4L5K4L4L4L4M3L5L3M3L4M3M4K4M3L4M3M3L5L3M3L4M3L4M4L3L4M2N3L4M3L4M3M2O2N2N2O1N2N101N2O1N2N2O1N101N2N2O1N2O2M7I8I6I8I6I7I8hIWGk5P9N3N2M3M3N2M3M3N2M2N3N2M3M3M3N2M3M2N3N2M1O1O100O1O1O100O1000000O10000000000O100000000O10002N6J3L1000O100O010O100O1000O0100O100O010O1000O0100O10O010000O010O100O010O1000O0100O10O0100O10O10O2O0O100O10000O100O100O101N10000O2O0O101N101N10001N100O2O0O2O000O2O0O101CYK^Fh4a9YK_Fh4_9ZK_Fg4`9ZK`Ff4_9[KaFf4\\9\\KcFe4\\9\\KdFe4Z9\\KeFe4Z9\\KfFd4Y9]KfFe4X9\\KhFd4W9^KgFd4W9]KiFc4V9^KjFc4T9^KkFc4T9^KlFb4S9_KlFc4R9^KnFb4Q9h0N1O2N1O2N1N2O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2N1O1N2O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1N3L3N2N3M2N3M2N3M3M3K5_Oa0_Oa0J5L5L4K5L4K5L4K5K5L4K5L3L5K5K5K4M4K5K5L4M2N3M3M3M2N3M3M3M3M2N3M3M3M5K6J7I7I7I`bk3"}, "label": "a bag of green apples"}]} {"id": 526713, "image": "COCO_train2014_000000526713.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small giraffe\"."}], "task": "refering", "answer_template": "The \"a small giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 166, 167, 190, 191, 213, 214, 215, 216, 236, 237, 238, 239, 259, 260, 261, 262, 282, 283, 284, 285, 307], "bbox": [0.24040625000000002, 0.43284037558685445, 0.40418750000000003, 0.8797887323943661], "iscrowd": 0, "area": 7267.643599999999, "rle": {"size": [426, 640], "counts": "eXP23T=4L3M4L4M3M2N3L4M3M3O001O1O3M3M3M3M3M3M3M4L3M4O02N2M10O1O1O2WGSNo5n1hI\\NV6f1`IaNa6`1WIeNk6\\1lHkNU7W1aHoNa7R1VHUOk7m0kGXOX8i0`G[Oc8d23N1L5G9G9G9F:G8D;H90000001O00000O3N1O2N2N2N2O010M3M2N3M[NfE=W:EbFC\\9=fFBX9?jFAS9`0oF@o8`0SG_Ol8a0UG^Oj8c0WG\\Oi8d0YG[Oe8f0\\GYOd8g0]GXOb8i0_GWO`8i0aGVO^8k0dGSO[8n0fGROY8n0hGPOX8R1hGlNY8T1hGjNX8W1iGgNX8Y1hGfNX8[1hGdNY8\\1gGcNY8_1fG_N\\8a1dG\\N^8e1c1100O2O0O10W1hN3N1NjMTFo0l9hN^FV1b9cNhF7[:BnE8W:@SF0lan4"}, "label": "a small giraffe"}]} {"id": 526713, "image": "COCO_train2014_000000526713.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young giraffe walking with other giraffes\"."}], "task": "refering", "answer_template": "The \"a young giraffe walking with other giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 166, 167, 190, 191, 213, 214, 215, 216, 236, 237, 238, 239, 259, 260, 261, 262, 282, 283, 284, 285, 307], "bbox": [0.24040625000000002, 0.43284037558685445, 0.40418750000000003, 0.8797887323943661], "iscrowd": 0, "area": 7267.643599999999, "rle": {"size": [426, 640], "counts": "eXP23T=4L3M4L4M3M2N3L4M3M3O001O1O3M3M3M3M3M3M3M4L3M4O02N2M10O1O1O2WGSNo5n1hI\\NV6f1`IaNa6`1WIeNk6\\1lHkNU7W1aHoNa7R1VHUOk7m0kGXOX8i0`G[Oc8d23N1L5G9G9G9F:G8D;H90000001O00000O3N1O2N2N2N2O010M3M2N3M[NfE=W:EbFC\\9=fFBX9?jFAS9`0oF@o8`0SG_Ol8a0UG^Oj8c0WG\\Oi8d0YG[Oe8f0\\GYOd8g0]GXOb8i0_GWO`8i0aGVO^8k0dGSO[8n0fGROY8n0hGPOX8R1hGlNY8T1hGjNX8W1iGgNX8Y1hGfNX8[1hGdNY8\\1gGcNY8_1fG_N\\8a1dG\\N^8e1c1100O2O0O10W1hN3N1NjMTFo0l9hN^FV1b9cNhF7[:BnE8W:@SF0lan4"}, "label": "a young giraffe walking with other giraffes"}]} {"id": 526713, "image": "COCO_train2014_000000526713.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe looking at camera\"."}], "task": "refering", "answer_template": "The \"giraffe looking at camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 57, 80, 81, 103, 104, 126, 127, 149, 150, 151, 172, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 242, 243, 244, 265, 266, 267, 287, 288, 289, 290], "bbox": [0.469234375, 0.10471830985915492, 0.6521250000000001, 0.8659859154929578], "iscrowd": 0, "area": 15277.421000000004, "rle": {"size": [426, 640], "counts": "hXm33V=2O1O1O2N1O1O2N1O1O3M3L4L4K6K4K6K4M4L4L;E;E;ZHWMU4S3ZKYMa4R3nJZMm4P3aJ\\MZ5o2oIbMn5Q5M4\\LWI in this image.", "objects": [{"patches": [34, 57, 80, 81, 103, 104, 126, 127, 149, 150, 151, 172, 173, 174, 175, 195, 196, 197, 198, 218, 219, 220, 221, 242, 243, 244, 265, 266, 267, 287, 288, 289, 290], "bbox": [0.469234375, 0.10471830985915492, 0.6521250000000001, 0.8659859154929578], "iscrowd": 0, "area": 15277.421000000004, "rle": {"size": [426, 640], "counts": "hXm33V=2O1O1O2N1O1O2N1O1O3M3L4L4K6K4K6K4M4L4L;E;E;ZHWMU4S3ZKYMa4R3nJZMm4P3aJ\\MZ5o2oIbMn5Q5M4\\LWI in this image.", "objects": [{"patches": [24, 25, 47, 48, 70, 71, 72, 94, 95, 117, 118, 119, 141, 142, 143, 163, 164, 165, 166, 187, 188, 189, 190, 210, 211, 212, 213, 233, 234, 235, 256, 257, 258], "bbox": [0.043375000000000004, 0.10786384976525823, 0.273734375, 0.811244131455399], "iscrowd": 0, "area": 13318.088150000001, "rle": {"size": [426, 640], "counts": "_g;8P=4L2N2O1N1O2N2N2N1N3M2O2O0000001O00000010O000002N100O2N1O101N12O1N2N2O1O02N3N2M2N3N1N3M2O2N101O010O001O1VFYNl7g1oG]NS8c1gGaNZ8_1aGeN`8\\1ZGiNf8W1UGmNl8T1nFPOS9P1hFTOQ3POd2l1UJXOQ3POm2h1nI[Oo2QOV3e1eI^OP3PO]3Q3]LSMd3o2VLTMk3m2oKWMR4j2iKYMY4g2aK]M`4d2ZK`Mg4a2TKbMm4_2mJeMT5\\2gJgMZ5m2mIWMT6n42K6J5K5J6K5K5J6K5K5K6I6K5N22N100O2NZNXGjNh8T1[GkNd8S1`GmN^8Q1fGnNY8P1jGPOV8n0mGROQ8l0SHSOl7k0XHTOh7j0[HVOc7h0aHWO^7h0eHWO[7h0gHXOW7i0iHWOV7j0kHUOU7j0mHVOQ7k0PITOo6m0RIROn6m0TISOj6n0WIQOh6P1YImNi6R1YIkNh6V1XIhNi6Y1XIcNk6\\1WIaNi6a1XI[Ni6g1XIVNi6j1YIRNh6P2YImMg6U2ZIgMh6Y2fHcMTO2V8l3kGPLV8S4`02M[LoFS3P9kLRGW3k8hLWGY3g8eL\\G]3b8aLaG_3]8`LeGb3X8]LjGd3U8ZLmGh3P8VLSHi3m7VLUHh3m7VLVHh3j7WLXHg3g8N2O1N3aN_1UOl0I7J6I6K6J6IhVQ6"}, "label": "giraffe to the far left"}]} {"id": 526713, "image": "COCO_train2014_000000526713.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe furthest left walking toward the photographer\"."}], "task": "refering", "answer_template": "The \"the giraffe furthest left walking toward the photographer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 47, 48, 70, 71, 72, 94, 95, 117, 118, 119, 141, 142, 143, 163, 164, 165, 166, 187, 188, 189, 190, 210, 211, 212, 213, 233, 234, 235, 256, 257, 258], "bbox": [0.043375000000000004, 0.10786384976525823, 0.273734375, 0.811244131455399], "iscrowd": 0, "area": 13318.088150000001, "rle": {"size": [426, 640], "counts": "_g;8P=4L2N2O1N1O2N2N2N1N3M2O2O0000001O00000010O000002N100O2N1O101N12O1N2N2O1O02N3N2M2N3N1N3M2O2N101O010O001O1VFYNl7g1oG]NS8c1gGaNZ8_1aGeN`8\\1ZGiNf8W1UGmNl8T1nFPOS9P1hFTOQ3POd2l1UJXOQ3POm2h1nI[Oo2QOV3e1eI^OP3PO]3Q3]LSMd3o2VLTMk3m2oKWMR4j2iKYMY4g2aK]M`4d2ZK`Mg4a2TKbMm4_2mJeMT5\\2gJgMZ5m2mIWMT6n42K6J5K5J6K5K5J6K5K5K6I6K5N22N100O2NZNXGjNh8T1[GkNd8S1`GmN^8Q1fGnNY8P1jGPOV8n0mGROQ8l0SHSOl7k0XHTOh7j0[HVOc7h0aHWO^7h0eHWO[7h0gHXOW7i0iHWOV7j0kHUOU7j0mHVOQ7k0PITOo6m0RIROn6m0TISOj6n0WIQOh6P1YImNi6R1YIkNh6V1XIhNi6Y1XIcNk6\\1WIaNi6a1XI[Ni6g1XIVNi6j1YIRNh6P2YImMg6U2ZIgMh6Y2fHcMTO2V8l3kGPLV8S4`02M[LoFS3P9kLRGW3k8hLWGY3g8eL\\G]3b8aLaG_3]8`LeGb3X8]LjGd3U8ZLmGh3P8VLSHi3m7VLUHh3m7VLVHh3j7WLXHg3g8N2O1N3aN_1UOl0I7J6I6K6J6IhVQ6"}, "label": "the giraffe furthest left walking toward the photographer"}]} {"id": 313724, "image": "COCO_train2014_000000313724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bear on right\"."}], "task": "refering", "answer_template": "The \"bear on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 82, 83, 84, 85, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 248, 249, 250], "bbox": [0.5337500000000001, 0.1685480093676815, 0.908578125, 0.7236065573770493], "iscrowd": 0, "area": 36918.08460000001, "rle": {"size": [427, 640], "counts": "Wh^4?i<5L3M2O2N1N2O2N1N2O2N1N3TFkN_7V1]HRO^7o0]HZO^7h0\\HA_7`0]HH^7:\\HN`73[H3c7NWH:f7MnG:P8ObG9Y8`2K4L5K4L5K4M4K4M4L4M3M2M4M3L4M2N1N2O2M2O2N10001N10001O0O101O0O101O0O2O00000O0100O010O100O10000O100O100O10000O100O100O100O10000O100O101N100O2O000O101N100O2O0O101O0O101N100100O1O010001O000O100jIRIn5o6PJRIP6n6PJRIP6n6oISIP6n6oITIP6m6nITIR6l6nITIR6S700O100O2O0O1O100O100O2N100O100O100O2N100O100O1O1O2N1O1O1O100O1O2N1O1O1O1O1O2N1O2N1O2N2O0O2N1O2N1O2N2M2O2M2O2N1N3N2M2O2M2O2M2O2N2N1N3N1O2N1O2M3N1O2N1O2M2O2N1O2N2M2O2N1O2N1N3N1O2N1O2M3N1O2N1O2M2O2N1O2M3N1N3N1N3N1O2M4M4K5L4K5L4K5L4K^Zh0"}, "label": "bear on right"}]} {"id": 313724, "image": "COCO_train2014_000000313724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bear attacking\"."}], "task": "refering", "answer_template": "The \"a bear attacking\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 82, 83, 84, 85, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 248, 249, 250], "bbox": [0.5337500000000001, 0.1685480093676815, 0.908578125, 0.7236065573770493], "iscrowd": 0, "area": 36918.08460000001, "rle": {"size": [427, 640], "counts": "Wh^4?i<5L3M2O2N1N2O2N1N2O2N1N3TFkN_7V1]HRO^7o0]HZO^7h0\\HA_7`0]HH^7:\\HN`73[H3c7NWH:f7MnG:P8ObG9Y8`2K4L5K4L5K4M4K4M4L4M3M2M4M3L4M2N1N2O2M2O2N10001N10001O0O101O0O101O0O2O00000O0100O010O100O10000O100O100O10000O100O100O100O10000O100O101N100O2O000O101N100O2O0O101O0O101N100100O1O010001O000O100jIRIn5o6PJRIP6n6PJRIP6n6oISIP6n6oITIP6m6nITIR6l6nITIR6S700O100O2O0O1O100O100O2N100O100O100O2N100O100O1O1O2N1O1O1O100O1O2N1O1O1O1O1O2N1O2N1O2N2O0O2N1O2N1O2N2M2O2M2O2N1N3N2M2O2M2O2M2O2N2N1N3N1O2N1O2M3N1O2N1O2M2O2N1O2N2M2O2N1O2N1N3N1O2N1O2M3N1O2N1O2M2O2N1O2M3N1N3N1N3N1O2M4M4K5L4K5L4K5L4K^Zh0"}, "label": "a bear attacking"}]} {"id": 313724, "image": "COCO_train2014_000000313724.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bear closer to beach , on left\"."}], "task": "refering", "answer_template": "The \"bear closer to beach , on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 96, 97, 98, 99, 100, 101, 102, 103, 118, 119, 120, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 256, 257, 258, 261, 264, 265, 279, 280, 281], "bbox": [0.10504687500000001, 0.23140515222482436, 0.57125, 0.829016393442623], "iscrowd": 0, "area": 49060.9902, "rle": {"size": [427, 640], "counts": "^Tl04U=>B>B in this image.", "objects": [{"patches": [77, 78, 96, 97, 98, 99, 100, 101, 102, 103, 118, 119, 120, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 256, 257, 258, 261, 264, 265, 279, 280, 281], "bbox": [0.10504687500000001, 0.23140515222482436, 0.57125, 0.829016393442623], "iscrowd": 0, "area": 49060.9902, "rle": {"size": [427, 640], "counts": "^Tl04U=>B>B in this image.", "objects": [{"patches": [40, 62, 63, 64, 85, 86, 87, 107, 108, 109, 110, 130, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 179, 200, 201], "bbox": [0.6715312499999999, 0.1236, 0.82075, 0.5528], "iscrowd": 0, "area": 11640.344649999997, "rle": {"size": [425, 640], "counts": "aeb54Q=5L4E;D in this image.", "objects": [{"patches": [40, 62, 63, 64, 85, 86, 87, 107, 108, 109, 110, 130, 131, 132, 133, 153, 154, 155, 156, 176, 177, 178, 179, 200, 201], "bbox": [0.6715312499999999, 0.1236, 0.82075, 0.5528], "iscrowd": 0, "area": 11640.344649999997, "rle": {"size": [425, 640], "counts": "aeb54Q=5L4E;D in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 16, 17, 18, 19, 20, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 40, 41, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 64, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 87, 88, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 110, 111, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 133, 134, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297], "bbox": [0.0416875, 0.0033704735376044567, 0.917203125, 0.9876044568245124], "iscrowd": 0, "area": 151052.25385, "rle": {"size": [359, 640], "counts": "`d94Q;4L4L4L4L4N2XHYOl4h0QKAg4`0VKJc46ZK3`4M^KgH_O\\7R1THjNo7X1oGeNT8\\1kGbNW85UGP1c0hN[8d1c01OdN`FR1_9nNaFR1_9nNbFQ1^9oNbFm0b9SO^Fj0e9VO\\Fe0h9\\OWF`0m9@TFe9B[Fb0a9^O_Ff0]9ZOcFj0Y9VOgFn0U9ROkFR1Q9nNoFV1m8jNSGZ1i8fNWG^1e8bNZGc1b8\\N_Gh1]8XNcGl1Y8TNgGP2U8PNkGT2Q8lMoGX2m7hMSH]2h7cMXHa2d7_M\\He2`7[M`Hi2\\7WMdHm2X7SMhHQ3T7oLkHV3Q7jLoHZ3m6fLSI^3i6bLWIb3e6^L[If3a6ZL_Ij3]6VLcIn3Y6RLgIR4U6nKkIV4Q6jKoIZ4m5fKSJ^4i5bKWJb4e5^K[Jf4a5ZK^Jk4^5UKbJm4\\5SKdJm4\\5SKdJm4\\5SKdJm4\\5SKdJm4\\5SKdJm4\\5SKdJm4\\5SKdJm4\\5SKdJm4\\5SKdJm4\\5SKdJm4\\5SKcJn45mJj45QKn42RKk4\\5QKjJm4V5PKPKm4g5N2N2N3M8H8H9G8H8H8H8H8H8H8H8H8H8H9G8H8H8H8HfWb0"}, "label": "the back view of a chevrolet truck , including racks in the truck bed"}]} {"id": 502153, "image": "COCO_train2014_000000502153.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue truck\"."}], "task": "refering", "answer_template": "The \"a blue truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 16, 17, 18, 19, 20, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 40, 41, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 64, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 87, 88, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 110, 111, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 133, 134, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297], "bbox": [0.0416875, 0.0033704735376044567, 0.917203125, 0.9876044568245124], "iscrowd": 0, "area": 151052.25385, "rle": {"size": [359, 640], "counts": "`d94Q;4L4L4L4L4N2XHYOl4h0QKAg4`0VKJc46ZK3`4M^KgH_O\\7R1THjNo7X1oGeNT8\\1kGbNW85UGP1c0hN[8d1c01OdN`FR1_9nNaFR1_9nNbFQ1^9oNbFm0b9SO^Fj0e9VO\\Fe0h9\\OWF`0m9@TFe9B[Fb0a9^O_Ff0]9ZOcFj0Y9VOgFn0U9ROkFR1Q9nNoFV1m8jNSGZ1i8fNWG^1e8bNZGc1b8\\N_Gh1]8XNcGl1Y8TNgGP2U8PNkGT2Q8lMoGX2m7hMSH]2h7cMXHa2d7_M\\He2`7[M`Hi2\\7WMdHm2X7SMhHQ3T7oLkHV3Q7jLoHZ3m6fLSI^3i6bLWIb3e6^L[If3a6ZL_Ij3]6VLcIn3Y6RLgIR4U6nKkIV4Q6jKoIZ4m5fKSJ^4i5bKWJb4e5^K[Jf4a5ZK^Jk4^5UKbJm4\\5SKdJm4\\5SKdJm4\\5SKdJm4\\5SKdJm4\\5SKdJm4\\5SKdJm4\\5SKdJm4\\5SKdJm4\\5SKdJm4\\5SKdJm4\\5SKcJn45mJj45QKn42RKk4\\5QKjJm4V5PKPKm4g5N2N2N3M8H8H9G8H8H8H8H8H8H8H8H8H8H9G8H8H8H8HfWb0"}, "label": "a blue truck"}]} {"id": 502153, "image": "COCO_train2014_000000502153.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back window of a truck\"."}], "task": "refering", "answer_template": "The \"the back window of a truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129], "bbox": [0.13046875, 0.2313649025069638, 0.653796875, 0.4378272980501393], "iscrowd": 0, "area": 20779.2846, "rle": {"size": [359, 640], "counts": "Wcm02R;3M3M3N2M3M3M3N2M3M3M3M3N2M3M3M3N2M3M3M3N2M300O1000000O10000O10000O1000O010000O10000000000O100000000000000000000000000000000000000000000000000000000000000O10000000000000000000O10000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000O1000O10000000000000000000001O000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000001O0000000000000000000000000000001O000001O000001O0000000000001O000000000000001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O002N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N2O0O2N1O2N2N1O2Nge]2"}, "label": "the back window of a truck"}]} {"id": 477580, "image": "COCO_train2014_000000477580.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the carrot on the very bottom\"."}], "task": "refering", "answer_template": "The \"the carrot on the very bottom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [289, 290, 291, 292, 307, 308, 309, 310, 311, 312, 313, 314, 315, 329, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.3109375, 0.8423529411764706, 0.71875, 0.9858823529411764], "iscrowd": 0, "area": 8143.5, "rle": {"size": [425, 640], "counts": "f\\c22W=1N2O1O1O1O1N2O1O2N1O1O1N2O1O1O1O1O1N10000000000O100000000O10000000O10O10000000000O100000000O10000000000O10000000000O100000000O1000O100000O10000000000O10000000000O100000O10O10000000O10O1000000000O0100000000O01000000000O01000000000O10O100000O1000O100000O1000O100000O100000O1000O100000O10O10000000O10O1000000000O010000000O010000000O10O100000O1000O1000O1000O1000O100000O01000000000O10000000000O100000000O10000000000O10000000000O100000000O10000000000O10000001O00001O00001O00001O00001N100O1O2N1O1O101N1O1O2N1O1N2N3L3M3M4L3MW[Z2"}, "label": "the carrot on the very bottom"}]} {"id": 477580, "image": "COCO_train2014_000000477580.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a group of broccoli\"."}], "task": "refering", "answer_template": "The \"a group of broccoli\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 278, 279], "bbox": [0.0, 0.5060705882352942, 0.24728124999999998, 0.8269176470588236], "iscrowd": 0, "area": 15116.198299999996, "rle": {"size": [425, 640], "counts": "f9o0[:n1QOo02O0O2N1O2O000O2O001O0O2O001N101O0O2O000O2O001N101O000O2O000O2O000010O1O00100O1O010O1O1O0O2O1O1N2O001O1N10001N10001O0O2O00001N10001N101O000O2O000010O01O01O000O1O2N1O1O2N1O1O101N1O1O1O2N1O1O2N100O1O2N1O1O2N1O1O101N1O1O1O2N1O1O2O0O101O00001N10001O000O2O00001O0O3N1O2N1O2M2O2N1O2N1N3N1O2N1O2M5L3M3M4L3L5L3M4L4K5K6IShW6"}, "label": "a group of broccoli"}]} {"id": 477580, "image": "COCO_train2014_000000477580.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a head of broccoli sits on the table\"."}], "task": "refering", "answer_template": "The \"a head of broccoli sits on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 278, 279], "bbox": [0.0, 0.5060705882352942, 0.24728124999999998, 0.8269176470588236], "iscrowd": 0, "area": 15116.198299999996, "rle": {"size": [425, 640], "counts": "f9o0[:n1QOo02O0O2N1O2O000O2O001O0O2O001N101O0O2O000O2O001N101O000O2O000O2O000010O1O00100O1O010O1O1O0O2O1O1N2O001O1N10001N10001O0O2O00001N10001N101O000O2O000010O01O01O000O1O2N1O1O2N1O1O101N1O1O1O2N1O1O2N100O1O2N1O1O2N1O1O101N1O1O1O2N1O1O2O0O101O00001N10001O000O2O00001O0O3N1O2N1O2M2O2N1O2N1N3N1O2N1O2M5L3M3M4L3L5L3M4L4K5K6IShW6"}, "label": "a head of broccoli sits on the table"}]} {"id": 477580, "image": "COCO_train2014_000000477580.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bottle of wine laying down amongst the veggies\"."}], "task": "refering", "answer_template": "The \"a bottle of wine laying down amongst the veggies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225], "bbox": [0.477625, 0.484, 0.8528281249999999, 0.6756], "iscrowd": 0, "area": 10628.95475, "rle": {"size": [425, 640], "counts": "SXo33V=4K4L5L4K4L5L3L3M10O01O10O0100O0010O01O10O01O10O01O100O100O1O100O1O010O1000000O1000000O1000O10O1000000O1000000O10O1000O100000000O100000O1000O10000000O10O10000000O10O100000O1000O100000O1000O100000O1000O100000O1000O1000O1000O100000O1000O100000O1000O1000O100000O1000O100000O1000O100000O10O10000000O10O10000000O10O10000000O010000000000O01000000000O10O100000000000O10O1000000000000O10000000O1000O10000000000001O1O001N101O1O001O001O1O001O1N2M3M3M3N3M2O1O1O1N2O2N1N2O1O1N3N1O1N2O1O2NVVW1"}, "label": "a bottle of wine laying down amongst the veggies"}]} {"id": 477580, "image": "COCO_train2014_000000477580.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bottle of wine between the vegetables\"."}], "task": "refering", "answer_template": "The \"a bottle of wine between the vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 178, 179, 195, 196, 197, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225], "bbox": [0.477625, 0.484, 0.8528281249999999, 0.6756], "iscrowd": 0, "area": 10628.95475, "rle": {"size": [425, 640], "counts": "SXo33V=4K4L5L4K4L5L3L3M10O01O10O0100O0010O01O10O01O10O01O100O100O1O100O1O010O1000000O1000000O1000O10O1000000O1000000O10O1000O100000000O100000O1000O10000000O10O10000000O10O100000O1000O100000O1000O100000O1000O100000O1000O1000O1000O100000O1000O100000O1000O1000O100000O1000O100000O1000O100000O10O10000000O10O10000000O10O10000000O010000000000O01000000000O10O100000000000O10O1000000000000O10000000O1000O10000000000001O1O001N101O1O001O001O1O001O1N2M3M3M3N3M2O1O1O1N2O2N1N2O1O1N3N1O1N2O1O2NVVW1"}, "label": "a bottle of wine between the vegetables"}]} {"id": 477580, "image": "COCO_train2014_000000477580.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the carrot to the left of the broccol\"."}], "task": "refering", "answer_template": "The \"the carrot to the left of the broccol\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 236, 237, 238, 257, 258, 259, 260, 278, 279, 280, 281, 282, 300, 301, 302, 303, 322, 323, 324, 325], "bbox": [0.028125, 0.6517647058823529, 0.36875, 0.9858823529411764], "iscrowd": 0, "area": 9401.0, "rle": {"size": [425, 640], "counts": "ik72T=5L4L4K5L3L3N2O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O101N1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O0001O000010O0001O0010O01O001O1O010O001O0010O01O001O10O01O001O0010O01O001O10O01O001O0010O01O001O10O01O001O010O0010O010O01O10O010O01O010O010O00100O010O0010O010O01O010O10O01O010O010O0010O0100O0010O01O010O010O00100O010O0010O010O01O01O01O010O001O001O001O001O001O001O01O01O001O001O001O001O001O0010O01O00001O001O001O001O00P]W5"}, "label": "the carrot to the left of the broccol"}]} {"id": 477580, "image": "COCO_train2014_000000477580.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a carrot going across the others\"."}], "task": "refering", "answer_template": "The \"a carrot going across the others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 236, 237, 238, 257, 258, 259, 260, 278, 279, 280, 281, 282, 300, 301, 302, 303, 322, 323, 324, 325], "bbox": [0.028125, 0.6517647058823529, 0.36875, 0.9858823529411764], "iscrowd": 0, "area": 9401.0, "rle": {"size": [425, 640], "counts": "ik72T=5L4L4K5L3L3N2O1O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O101N1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O0001O000010O0001O0010O01O001O1O010O001O0010O01O001O10O01O001O0010O01O001O10O01O001O0010O01O001O10O01O001O010O0010O010O01O10O010O01O010O010O00100O010O0010O010O01O010O10O01O010O010O0010O0100O0010O01O010O010O00100O010O0010O010O01O01O01O010O001O001O001O001O001O001O01O01O001O001O001O001O001O0010O01O00001O001O001O001O00P]W5"}, "label": "a carrot going across the others"}]} {"id": 477580, "image": "COCO_train2014_000000477580.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bottle of red wine , labeled chocovine , which has a picture of a windmill on it\"."}], "task": "refering", "answer_template": "The \"a bottle of red wine , labeled chocovine , which has a picture of a windmill on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 31, 32, 33, 34, 54, 55, 56, 57, 58, 77, 78, 79, 80, 81, 100, 101, 102, 103, 104, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 193, 194, 195, 196, 216], "bbox": [0.366515625, 0.001811764705882353, 0.532515625, 0.6148470588235294], "iscrowd": 0, "area": 23370.684999999994, "rle": {"size": [425, 640], "counts": "bfQ3c1n9h1XNh1A>K6M3N2N2N2N3M4L3L5L4L2N2N2N1O2M2O1O0010O2O1N2O1OgJZIQ4e6mK^IS4b6jKaIV4_6gKdIY4[6dKhI]4X6`KkI`4U6]KnIc4Q6[KRJe4n5YKSJh4m5UKVJk4i5SKZJm4f5PKZJS5f5jJXJ[5h5bJWJb5c65O1000O10001O0000000000000000001O0000001O2N1O1O2N1O2N1O0000O100O10000O100O10000O10000O100O10000mHdJX6]5gIjJS6V5lIoJo5R5QJPKm5P5RJTKk5m4SJWKi5m4SJVKk5m4QJWKk5m4QJWKl5U6TKfHe3]7ULhHh3\\7gK]HO;Y4Z7`KdH24]4Y800jNV1[Nd1dNbQl3"}, "label": "a bottle of red wine , labeled chocovine , which has a picture of a windmill on it"}]} {"id": 477580, "image": "COCO_train2014_000000477580.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bottle of wine which is kept in the middle of vegetables\"."}], "task": "refering", "answer_template": "The \"a bottle of wine which is kept in the middle of vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 31, 32, 33, 34, 54, 55, 56, 57, 58, 77, 78, 79, 80, 81, 100, 101, 102, 103, 104, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 193, 194, 195, 196, 216], "bbox": [0.366515625, 0.001811764705882353, 0.532515625, 0.6148470588235294], "iscrowd": 0, "area": 23370.684999999994, "rle": {"size": [425, 640], "counts": "bfQ3c1n9h1XNh1A>K6M3N2N2N2N3M4L3L5L4L2N2N2N1O2M2O1O0010O2O1N2O1OgJZIQ4e6mK^IS4b6jKaIV4_6gKdIY4[6dKhI]4X6`KkI`4U6]KnIc4Q6[KRJe4n5YKSJh4m5UKVJk4i5SKZJm4f5PKZJS5f5jJXJ[5h5bJWJb5c65O1000O10001O0000000000000000001O0000001O2N1O1O2N1O2N1O0000O100O10000O100O10000O10000O100O10000mHdJX6]5gIjJS6V5lIoJo5R5QJPKm5P5RJTKk5m4SJWKi5m4SJVKk5m4QJWKk5m4QJWKl5U6TKfHe3]7ULhHh3\\7gK]HO;Y4Z7`KdH24]4Y800jNV1[Nd1dNbQl3"}, "label": "a bottle of wine which is kept in the middle of vegetables"}]} {"id": 477580, "image": "COCO_train2014_000000477580.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a carrot stacked in the middle of two other carrots\"."}], "task": "refering", "answer_template": "The \"a carrot stacked in the middle of two other carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 303, 304, 305, 306, 307, 308, 309, 310], "bbox": [0.17489062500000002, 0.820564705882353, 0.639515625, 0.9315764705882353], "iscrowd": 0, "area": 6214.0416000000005, "rle": {"size": [425, 640], "counts": "oX_11X=0O001O0010O01O1O0010O01O1O010O001O00100O001O00100O001O00100O0010O100O10000O10O01001O000000000O100000001O0000000O10000000001O001N10001O001O001O001N101O1O1O1O1O1O2M2O1O1O1O1O1O00O01000O10000O10O100000O100000O0100000000O10O100000O100000O0100000000O100000000O10O1000O100000000O100000O100000000000000000000O101O00O1000000O100000O01000000O1000000O1000O1000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O10000000000O10000000000000000O10000000000000000O2O000000000000000O2O000000000000000O2O000000000000000O2O000000000000000O2O00000000001O0O101O00001O00001N10001O00001O0000Y[P3"}, "label": "a carrot stacked in the middle of two other carrots"}]} {"id": 477580, "image": "COCO_train2014_000000477580.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle carrot that ' s laying down with two other carrots , on either side\"."}], "task": "refering", "answer_template": "The \"the middle carrot that ' s laying down with two other carrots , on either side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 303, 304, 305, 306, 307, 308, 309, 310], "bbox": [0.17489062500000002, 0.820564705882353, 0.639515625, 0.9315764705882353], "iscrowd": 0, "area": 6214.0416000000005, "rle": {"size": [425, 640], "counts": "oX_11X=0O001O0010O01O1O0010O01O1O010O001O00100O001O00100O001O00100O0010O100O10000O10O01001O000000000O100000001O0000000O10000000001O001N10001O001O001O001N101O1O1O1O1O1O2M2O1O1O1O1O1O00O01000O10000O10O100000O100000O0100000000O10O100000O100000O0100000000O100000000O10O1000O100000000O100000O100000000000000000000O101O00O1000000O100000O01000000O1000000O1000O1000O1000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O10000000000O10000000000000000O10000000000000000O2O000000000000000O2O000000000000000O2O000000000000000O2O000000000000000O2O00000000001O0O101O00001O00001N10001O00001O0000Y[P3"}, "label": "the middle carrot that ' s laying down with two other carrots , on either side"}]} {"id": 215436, "image": "COCO_train2014_000000215436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zara shopping bag\"."}], "task": "refering", "answer_template": "The \"the zara shopping bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [135, 136, 147, 148, 149, 150, 160, 161, 162, 163, 173, 174, 175, 176, 186, 187, 188, 189], "bbox": [0.3532533333333333, 0.6, 0.5922133333333333, 0.8155800000000001], "iscrowd": 0, "area": 8101.094350000003, "rle": {"size": [500, 375], "counts": "ahP21c?b0^Oa0^Ob0_O`0@1O1O1O2N1O1O2M2O1O1O2N1O1O2N000O10O10J6I70000O10000O1000O10O1000000O1000000O10O2O1O1O1O1O1O2N1O1O1O1N2O000000000000O100000O100000000000000000O1000O10000000000000000000O0o\\Z2"}, "label": "the zara shopping bag"}]} {"id": 215436, "image": "COCO_train2014_000000215436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zara bag that the man is holding\"."}], "task": "refering", "answer_template": "The \"zara bag that the man is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [135, 136, 147, 148, 149, 150, 160, 161, 162, 163, 173, 174, 175, 176, 186, 187, 188, 189], "bbox": [0.3532533333333333, 0.6, 0.5922133333333333, 0.8155800000000001], "iscrowd": 0, "area": 8101.094350000003, "rle": {"size": [500, 375], "counts": "ahP21c?b0^Oa0^Ob0_O`0@1O1O1O2N1O1O2M2O1O1O2N1O1O2N000O10O10J6I70000O10000O1000O10O1000000O1000000O10O2O1O1O1O1O1O2N1O1O1O1N2O000000000000O100000O100000000000000000O1000O10000000000000000000O0o\\Z2"}, "label": "zara bag that the man is holding"}]} {"id": 215436, "image": "COCO_train2014_000000215436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow messenger bag being held by a man\"."}], "task": "refering", "answer_template": "The \"a yellow messenger bag being held by a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 108, 109, 110, 121, 122, 123, 134, 135], "bbox": [0.3077066666666667, 0.40904, 0.55408, 0.61486], "iscrowd": 0, "area": 3928.7467500000007, "rle": {"size": [500, 375], "counts": "_\\h14_?4L4L4L3M4L4L4L3M4L3M3M3M2N3M3M3M3M3M2N3M3M3M3M31N101O1M2N3N2M2O2M3M2N3L4CUBkNm=m0m0A6J4L5Ld\\12\\cN3L3N1On0SO9F3M2N2N2O0N1O1O2N00O1O2N2O1N2N2N2O1N2N2N2O1N2N2N2N3M2N2N3M2N2N3M2N2N3M2N2N3L3N2NWia2"}, "label": "a yellow messenger bag being held by a man"}]} {"id": 215436, "image": "COCO_train2014_000000215436.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow bag pinned beneath the forearm of a man sitting on a yellow fire hydrant .\\\"."}], "task": "refering", "answer_template": "The \"a yellow bag pinned beneath the forearm of a man sitting on a yellow fire hydrant .\\\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 108, 109, 110, 121, 122, 123, 134, 135], "bbox": [0.3077066666666667, 0.40904, 0.55408, 0.61486], "iscrowd": 0, "area": 3928.7467500000007, "rle": {"size": [500, 375], "counts": "_\\h14_?4L4L4L3M4L4L4L3M4L3M3M3M2N3M3M3M3M3M2N3M3M3M3M31N101O1M2N3N2M2O2M3M2N3L4CUBkNm=m0m0A6J4L5Ld\\12\\cN3L3N1On0SO9F3M2N2N2O0N1O1O2N00O1O2N2O1N2N2N2O1N2N2N2O1N2N2N2N3M2N2N3M2N2N3M2N2N3M2N2N3L3N2NWia2"}, "label": "a yellow bag pinned beneath the forearm of a man sitting on a yellow fire hydrant .\\"}]} {"id": 141711, "image": "COCO_train2014_000000141711.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue jacket skiing down a mountain\"."}], "task": "refering", "answer_template": "The \"a man in a blue jacket skiing down a mountain\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 59, 60, 81, 82, 83, 84, 104, 105, 106, 107, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 173, 174, 175, 176, 196, 197, 198, 219, 220, 221, 242, 243, 244, 265, 266, 267, 288, 289, 290], "bbox": [0.52353125, 0.09876953125, 0.73115625, 0.7229296875], "iscrowd": 0, "area": 21238.047349999997, "rle": {"size": [512, 640], "counts": "hiW58g?i0WO7H8I7SK]NPKf1g4fNSK[1f4nNVKU1c4RO[KP1`4UO^Kl0_4YO]Kj0_4ZO_Kh0\\4^OaKc0\\4AaKb0Z4CcK`0Y4E]Kc0`4AWITNa0f2T6[OYIZNIk2n6kNXIdNYOi2^7eNVId2j6\\MUIe2k6\\MSIe2l6]MRId2n6\\MPIf2o6\\MoHe2Q7gMbHZ2]7U30000O100O100O10O1MO1L6L6O0O100^OXGkIi8T6b0O2N1O2N2M2O2N2N1O1O10000000iNQJ]Ho5a7WJ[Hi5b7^JZHb5d7dJXH\\5f7iJWHW5f7mJXHT5f7oJYHQ5d7SK[Hn4b7UK]Hk4`7YK_Hh4^7\\K`He4]7^K`He4\\7_KaHd4i6PJWI_1Nb4e6XLXIk3`6P3I8I6K6I7J:E in this image.", "objects": [{"patches": [37, 59, 60, 81, 82, 83, 84, 104, 105, 106, 107, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 173, 174, 175, 176, 196, 197, 198, 219, 220, 221, 242, 243, 244, 265, 266, 267, 288, 289, 290], "bbox": [0.52353125, 0.09876953125, 0.73115625, 0.7229296875], "iscrowd": 0, "area": 21238.047349999997, "rle": {"size": [512, 640], "counts": "hiW58g?i0WO7H8I7SK]NPKf1g4fNSK[1f4nNVKU1c4RO[KP1`4UO^Kl0_4YO]Kj0_4ZO_Kh0\\4^OaKc0\\4AaKb0Z4CcK`0Y4E]Kc0`4AWITNa0f2T6[OYIZNIk2n6kNXIdNYOi2^7eNVId2j6\\MUIe2k6\\MSIe2l6]MRId2n6\\MPIf2o6\\MoHe2Q7gMbHZ2]7U30000O100O100O10O1MO1L6L6O0O100^OXGkIi8T6b0O2N1O2N2M2O2N2N1O1O10000000iNQJ]Ho5a7WJ[Hi5b7^JZHb5d7dJXH\\5f7iJWHW5f7mJXHT5f7oJYHQ5d7SK[Hn4b7UK]Hk4`7YK_Hh4^7\\K`He4]7^K`He4\\7_KaHd4i6PJWI_1Nb4e6XLXIk3`6P3I8I6K6I7J:E in this image.", "objects": [{"patches": [106, 129, 130, 152, 153, 154, 176, 177, 178, 200, 201, 202, 203, 224, 225, 226, 227], "bbox": [0.6106874999999999, 0.27523474178403756, 0.9170624999999999, 0.6355868544600939], "iscrowd": 0, "area": 5223.698399999999, "rle": {"size": [426, 640], "counts": "ciR53U=4M3K5K5K5J6L4K4O1O2M2O1O2N1O1N3N1O1O2N1O1N25L1N2N2O1N1O2N2N2N1O1O000001O0001O000000001O000O100000001O000O1000000000000O101O000000000O10000O10001N10lCWOf;i0XD\\Oe;c0ZDAd;?ZDDe;<[DDe;;\\DFc;:^DEb;;_DE`;:bDF];:dDE[;iDAW;`0hDAX;?hD@Y;`0fD@[;`0eD_O\\;a0cD@\\;a0cD_O^;a0bD^O_;b0`D^Oa;b0_D]Ob;c0]D^Oc;b0]D]Oe;b0ZD^Oh;a0XD^Ok;`0SDBn;=QDCQ< in this image.", "objects": [{"patches": [106, 129, 130, 152, 153, 154, 176, 177, 178, 200, 201, 202, 203, 224, 225, 226, 227], "bbox": [0.6106874999999999, 0.27523474178403756, 0.9170624999999999, 0.6355868544600939], "iscrowd": 0, "area": 5223.698399999999, "rle": {"size": [426, 640], "counts": "ciR53U=4M3K5K5K5J6L4K4O1O2M2O1O2N1O1N3N1O1O2N1O1N25L1N2N2O1N1O2N2N2N1O1O000001O0001O000000001O000O100000001O000O1000000000000O101O000000000O10000O10001N10lCWOf;i0XD\\Oe;c0ZDAd;?ZDDe;<[DDe;;\\DFc;:^DEb;;_DE`;:bDF];:dDE[;iDAW;`0hDAX;?hD@Y;`0fD@[;`0eD_O\\;a0cD@\\;a0cD_O^;a0bD^O_;b0`D^Oa;b0_D]Ob;c0]D^Oc;b0]D]Oe;b0ZD^Oh;a0XD^Ok;`0SDBn;=QDCQ< in this image.", "objects": [{"patches": [234, 235, 236, 256, 257, 258, 259, 260, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 307, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 352, 353, 354, 355, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.149140625, 0.6194791666666667, 0.467984375, 0.9995416666666668], "iscrowd": 0, "area": 24240.91165, "rle": {"size": [480, 640], "counts": "bk\\14i>g0ZO4K5K6K4K5K5L5J5M3M4L3M3M4L3N2M`0@7I6J7I6K5J6K5K6I6[FSKm8R5kFQKU9T5dFnJ[9\\50000O10000O10000O11O0000000000001O000000000000001O0000000000001O000000000000001O00001O001O001O1O001O001O00100O001O001O001O1O001O001O001O1O001O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N2N2N2N2N2N2N2N3M2N2N2N3M4M3L4L4L4L4L4L4L4L4L4L2N00O010O10000O100O100O100O1000O0100O01N101O1O0O2O001O001N2O001O0O2O001O01000000000000000000O0100005K7I8H3M01O5K6J6J8H>@`0^Oc0]Oamn4"}, "label": "chair under girl with green shorts"}]} {"id": 149921, "image": "COCO_train2014_000000149921.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old chair\"."}], "task": "refering", "answer_template": "The \"an old chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [234, 235, 236, 256, 257, 258, 259, 260, 279, 280, 281, 282, 283, 302, 303, 304, 305, 306, 307, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 352, 353, 354, 355, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.149140625, 0.6194791666666667, 0.467984375, 0.9995416666666668], "iscrowd": 0, "area": 24240.91165, "rle": {"size": [480, 640], "counts": "bk\\14i>g0ZO4K5K6K4K5K5L5J5M3M4L3M3M4L3N2M`0@7I6J7I6K5J6K5K6I6[FSKm8R5kFQKU9T5dFnJ[9\\50000O10000O10000O11O0000000000001O000000000000001O0000000000001O000000000000001O00001O001O001O1O001O001O00100O001O001O001O1O001O001O001O1O001O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N2N2N2N2N2N2N2N3M2N2N2N3M4M3L4L4L4L4L4L4L4L4L4L2N00O010O10000O100O100O100O1000O0100O01N101O1O0O2O001O001N2O001O0O2O001O01000000000000000000O0100005K7I8H3M01O5K6J6J8H>@`0^Oc0]Oamn4"}, "label": "an old chair"}]} {"id": 149921, "image": "COCO_train2014_000000149921.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty picnic table bench\"."}], "task": "refering", "answer_template": "The \"an empty picnic table bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 276, 277, 278, 299, 300, 301], "bbox": [0.0, 0.516875, 0.21234375, 0.8129791666666667], "iscrowd": 0, "area": 4770.5238500000005, "rle": {"size": [480, 640], "counts": "f91o>1^B0Q<3mC3m;NRD9g;HYD9d;I[D8d;JZD6e;L[D3e;NZD2f;0XD1f;1ZDNf;3YDMg;4XDMf;6YDIg;8YDGf;;YDFf;XDBg;>[DBd;>]DAc;>^DBa;>aDB^;>bDB];>eDA[;?fDAY;>hDBW;?jD@V;?kDAT;`0mD@R;?oDAQ;?oDAP;?REAm:>TEBl:>TEBk:>VECi:=WECh:=YECg:=YEDf:;[EEd:<]ECc:=]EDb:>\\EBc:`0\\E@d:b0ZE_Od:d0ZE\\Of:e0YE[Og:g0XEXOg:j0XEWOg:k0WEUOh:R1REnNn:X1lDhNT;^1fDbNY;e1aD[N_;i1]DXNa;m1[DSNe;9VDm01jNi;9YD]1f;cN^DZ1b;fNaDW1^;jNeDS1[;lNiDQ1V;POmDn0R;QOREl0m:UOVEh0j:WOZEf0f:ZO]Ec0b:]ObE9e:GY2010O01O010O10O010O01O010O010O10O10O0100O010O10O10O10O0100O010O1000O0100O010O10O01000O01O10O01O10O01O10O010^Oa0O101N100O2O0O101N100O2O0O10YP\\7"}, "label": "an empty picnic table bench"}]} {"id": 149921, "image": "COCO_train2014_000000149921.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"empty bench\"."}], "task": "refering", "answer_template": "The \"empty bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 276, 277, 278, 299, 300, 301], "bbox": [0.0, 0.516875, 0.21234375, 0.8129791666666667], "iscrowd": 0, "area": 4770.5238500000005, "rle": {"size": [480, 640], "counts": "f91o>1^B0Q<3mC3m;NRD9g;HYD9d;I[D8d;JZD6e;L[D3e;NZD2f;0XD1f;1ZDNf;3YDMg;4XDMf;6YDIg;8YDGf;;YDFf;XDBg;>[DBd;>]DAc;>^DBa;>aDB^;>bDB];>eDA[;?fDAY;>hDBW;?jD@V;?kDAT;`0mD@R;?oDAQ;?oDAP;?REAm:>TEBl:>TEBk:>VECi:=WECh:=YECg:=YEDf:;[EEd:<]ECc:=]EDb:>\\EBc:`0\\E@d:b0ZE_Od:d0ZE\\Of:e0YE[Og:g0XEXOg:j0XEWOg:k0WEUOh:R1REnNn:X1lDhNT;^1fDbNY;e1aD[N_;i1]DXNa;m1[DSNe;9VDm01jNi;9YD]1f;cN^DZ1b;fNaDW1^;jNeDS1[;lNiDQ1V;POmDn0R;QOREl0m:UOVEh0j:WOZEf0f:ZO]Ec0b:]ObE9e:GY2010O01O010O10O010O01O010O010O10O10O0100O010O10O10O10O0100O010O1000O0100O010O10O01000O01O10O01O10O01O10O010^Oa0O101N100O2O0O101N100O2O0O10YP\\7"}, "label": "empty bench"}]} {"id": 149921, "image": "COCO_train2014_000000149921.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty seat at on a bench\"."}], "task": "refering", "answer_template": "The \"an empty seat at on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [268, 269, 270, 290, 291, 292, 293, 312, 313, 314, 315, 335, 336, 337, 338, 359, 360, 383], "bbox": [0.5749375, 0.6750833333333334, 0.7742812499999999, 1.0], "iscrowd": 0, "area": 7746.115900000001, "rle": {"size": [480, 640], "counts": "jl\\51l>5L4K4M4L3M3N1O2M2O1O1O2N1N2O1O1O2N1N2O1O2N1O1O1N3N1O1O1O2M2O1O1O2N1O1N2O1O2N1O1N2O2N1O1O1N3N1O1O1O2N2M3N2N3M2N2M3N2N2N2N2M3N3aE]L\\9e3aF]L^9m2kEcMd0C_9j2PFcM=Eb9g2UFcM5If9c2XFdMNLi9^2]FeMGOl9[2`FfM@1P:W2dFgMYO4S:T2hF\\NX9b1mF[NS9d1QGYNo8e1VGXNj8g1ZGVNf8h1_GUNa8j1[2N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O2M3N2M3N1N2O1O2M2O1N2O1N3N1N2O1N2O2M2O1N2O2M2O1N2OedS2"}, "label": "an empty seat at on a bench"}]} {"id": 149921, "image": "COCO_train2014_000000149921.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table bench with two people on\"."}], "task": "refering", "answer_template": "The \"the table bench with two people on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [268, 269, 270, 290, 291, 292, 293, 312, 313, 314, 315, 335, 336, 337, 338, 359, 360, 383], "bbox": [0.5749375, 0.6750833333333334, 0.7742812499999999, 1.0], "iscrowd": 0, "area": 7746.115900000001, "rle": {"size": [480, 640], "counts": "jl\\51l>5L4K4M4L3M3N1O2M2O1O1O2N1N2O1O1O2N1N2O1O2N1O1O1N3N1O1O1O2M2O1O1O2N1O1N2O1O2N1O1N2O2N1O1O1N3N1O1O1O2N2M3N2N3M2N2M3N2N2N2N2M3N3aE]L\\9e3aF]L^9m2kEcMd0C_9j2PFcM=Eb9g2UFcM5If9c2XFdMNLi9^2]FeMGOl9[2`FfM@1P:W2dFgMYO4S:T2hF\\NX9b1mF[NS9d1QGYNo8e1VGXNj8g1ZGVNf8h1_GUNa8j1[2N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O2M3N2M3N1N2O1O2M2O1N2O1N3N1N2O1N2O2M2O1N2O2M2O1N2OedS2"}, "label": "the table bench with two people on"}]} {"id": 100777, "image": "COCO_train2014_000000100777.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair of the man with the bald spot\"."}], "task": "refering", "answer_template": "The \"the chair of the man with the bald spot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [391, 392, 393, 394, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 437, 438, 439, 440, 441, 442, 443, 444, 445, 460, 461, 462, 463, 464, 465, 466, 467, 468], "bbox": [0.012640624999999999, 0.83220367278798, 0.42159375, 0.9898664440734559], "iscrowd": 0, "area": 17319.81919999999, "rle": {"size": [599, 640], "counts": "[U53db07J7I6I7J7I6I7J6I8H7I7J5J0000000001O000000001O00000000001O000000001O0000000000000000000001O000O10000000000000000O100O10000O101N10000O10000O100O1001O001O001O001O0010O01O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O010O1O1O1O1O1O1O1O00000000000000000000010O0000000000000000O1000000000000000000000000000000O10000000000O100000000O10000000000O100000000O1000000000000O10000000000O1000000000000000000O10000000000000000000000000000O100000000000000O1000000000O100000000O100000000O100N2N2N2M3M2N3M3L4B>L4M3L3O2O100O1O00100O1O1O010N2N3N1N2N3M2Na`h6"}, "label": "the chair of the man with the bald spot"}]} {"id": 100777, "image": "COCO_train2014_000000100777.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden chair being sat on by a man in the foreground\"."}], "task": "refering", "answer_template": "The \"a wooden chair being sat on by a man in the foreground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [391, 392, 393, 394, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 437, 438, 439, 440, 441, 442, 443, 444, 445, 460, 461, 462, 463, 464, 465, 466, 467, 468], "bbox": [0.012640624999999999, 0.83220367278798, 0.42159375, 0.9898664440734559], "iscrowd": 0, "area": 17319.81919999999, "rle": {"size": [599, 640], "counts": "[U53db07J7I6I7J7I6I7J6I8H7I7J5J0000000001O000000001O00000000001O000000001O0000000000000000000001O000O10000000000000000O100O10000O101N10000O10000O100O1001O001O001O001O0010O01O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O010O1O1O1O1O1O1O1O00000000000000000000010O0000000000000000O1000000000000000000000000000000O10000000000O100000000O10000000000O100000000O1000000000000O10000000000O1000000000000000000O10000000000000000000000000000O100000000000000O1000000000O100000000O100000000O100N2N2N2M3M2N3M3L4B>L4M3L3O2O100O1O00100O1O1O010N2N3N1N2N3M2Na`h6"}, "label": "a wooden chair being sat on by a man in the foreground"}]} {"id": 371114, "image": "COCO_train2014_000000371114.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the light streak\"."}], "task": "refering", "answer_template": "The \"the light streak\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 260, 261, 262, 276, 277], "bbox": [0.0, 0.3634583333333333, 0.418921875, 0.7318958333333333], "iscrowd": 0, "area": 35273.43980000001, "rle": {"size": [480, 640], "counts": "h5i2W in this image.", "objects": [{"patches": [138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 260, 261, 262, 276, 277], "bbox": [0.0, 0.3634583333333333, 0.418921875, 0.7318958333333333], "iscrowd": 0, "area": 35273.43980000001, "rle": {"size": [480, 640], "counts": "h5i2W in this image.", "objects": [{"patches": [132, 133, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 175, 176, 193, 194, 216, 239, 241, 242, 244, 245, 246, 265, 266, 267, 269, 288, 289], "bbox": [0.4055625, 0.37462616822429906, 0.824453125, 0.8193457943925234], "iscrowd": 0, "area": 5669.858799999999, "rle": {"size": [428, 640], "counts": "gn\\35R=5K5^OBQDd0l;]ORDf0k;]OUDc0i;_OWD`0h;BXDZ4BiK;W4FjK7W4IlK4T4LnK1S4OoKOR40QLMo33e400001O0000000001O01O000000000000001SGId46[KMc43]KOa41^K2_4OaK2^4NVKI`LX8FYKM^LT5BkJ>V5BjJ=W5ChJ=Y5DfJ;P5DaG1_3:Q51nJOR52nJLT54kJLU55kJJV56iJIX58gJHZ59cJH_57`JHd56[JJg55XJKk53SJMQ61nIOT60kIOY6OfI1\\6ObI1_6O_I1c6O\\I1d60[I0f60[IMg63YILh64XIJj67VIGj6:VIEk6;UICm6=TIAm6?SI@n6`0n201000000O0100000O01O100O010O100O10O10O100O0cF_OQ6a0oIDl5j4BWKMX53iJ[Oi5e0^310O010O10000O01000O10000O1000O010000O10000O01000O10000O100O010O100O100O010O10O0100O10O010O010O0100O010O100O10O0100O100000000O2O2N1O1O1O1NVR_1"}, "label": "a horse behind the another horse"}]} {"id": 182706, "image": "COCO_train2014_000000182706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse that is fully visible\"."}], "task": "refering", "answer_template": "The \"the horse that is fully visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 135, 149, 150, 154, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 217, 218, 219, 220, 221, 222, 223, 224, 225, 239, 240, 241, 242, 243, 244, 245, 246, 247, 262, 263, 264, 269, 270, 285, 286, 287, 288, 291, 292, 293, 310, 311, 315, 316, 338, 339], "bbox": [0.40475000000000005, 0.3547196261682243, 0.9294062499999999, 0.9548364485981309], "iscrowd": 0, "area": 36911.62614999999, "rle": {"size": [428, 640], "counts": "Zb\\34U=:F:F9H9F3M2N2O2M2N3M2O1N3M2N2N3N1N2N3lNS1N3M2O1N3NF9]Oc0L59F:F9H9H8I7H8I7I7H3N1N2O001N101N101N2O001N101N101O1N101N101O0O100O101OVN_HiM`7V2dHhM\\7V2iHgMV7X2nHfMR7Y2QIeMo6Y2UIfMj6X2[IeMe6Y2_IeMa6Y2cIeM]6Z2gIcMY6[2kIcMU6[2oIdMQ6Y2TJdMl5Z2XJdMh5[2[JcMe5[2_JcMa5[2dJbM\\5\\2hJbMX5\\2lJcMT5[2PKbMP5\\2TKbMl4\\2XKbMh4\\2]KaMc4]2aKaM_4^2dKaM[4]2iKaMW4]2nK`MS4]2_3O1O10000O1100O001O1O001O10O01O1O001O1O001O10O01O1O001O1O001O1O010O001O000000000010O0000000001O00000001O01O000000001O0001O000001O00000000001O00000O10001O00000000001eGPMX6P3dITM\\6l2_IYMa6g2\\I\\Md6d2[I]Me6c2[I]Me6d2ZI\\Mf6d2YI\\Mh6d2XI\\Mg6e2YI[Mg6e2XI\\Mh6e2WI[Mi6e2WI[Mi6f2UI[Mk6e2UI[Mk6f2SI[Ml6f2TIYMm6g2SIYMl6i2RIXMn6h2RIXMm6j2RIVMn6j2QIWMo6j2PIVMo6R3jHnLV7Z3aHfL_7c3YH]Lg7l3PHTLo7e406J9F9H2M4M2M4M3M2M4M3L3N1N10000O^NdIUL]6j3fITLY6l3jIRLV6m3mIQLR6o3QJoKo5P4UJmKj5S4YJkKg5T4\\JjKc5V4`JhK`5W4cJgK\\5Y4hJdKX5j2gIlMT1XOT5l2mIgMR1[OQ5m2QJdMQ1]Om4P3UJ_MQ1_Oj4Q3YJ\\MQ1@e4T3_JWMo0Cb4V3bJSMo0E^4X3gJoLn0G[4Y3kJlLm0IW4\\3PKfLm0KS4^3TKcLl0Mo3`3f3O010O0010O01O010O0010O01O0O2M2N3M3L4M3M3M3M3O1N2O1N2O1N2O1N2O1N2O0O2O1N2N2O1N2O1N2O1O11O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1M3M3M3L4M3L4M4O01O1O001N2N2N1N3N2N1O2N2N2M2O2N2N3M6J6J6I7J:F;E in this image.", "objects": [{"patches": [133, 134, 135, 149, 150, 154, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 217, 218, 219, 220, 221, 222, 223, 224, 225, 239, 240, 241, 242, 243, 244, 245, 246, 247, 262, 263, 264, 269, 270, 285, 286, 287, 288, 291, 292, 293, 310, 311, 315, 316, 338, 339], "bbox": [0.40475000000000005, 0.3547196261682243, 0.9294062499999999, 0.9548364485981309], "iscrowd": 0, "area": 36911.62614999999, "rle": {"size": [428, 640], "counts": "Zb\\34U=:F:F9H9F3M2N2O2M2N3M2O1N3M2N2N3N1N2N3lNS1N3M2O1N3NF9]Oc0L59F:F9H9H8I7H8I7I7H3N1N2O001N101N101N2O001N101N101O1N101N101O0O100O101OVN_HiM`7V2dHhM\\7V2iHgMV7X2nHfMR7Y2QIeMo6Y2UIfMj6X2[IeMe6Y2_IeMa6Y2cIeM]6Z2gIcMY6[2kIcMU6[2oIdMQ6Y2TJdMl5Z2XJdMh5[2[JcMe5[2_JcMa5[2dJbM\\5\\2hJbMX5\\2lJcMT5[2PKbMP5\\2TKbMl4\\2XKbMh4\\2]KaMc4]2aKaM_4^2dKaM[4]2iKaMW4]2nK`MS4]2_3O1O10000O1100O001O1O001O10O01O1O001O1O001O10O01O1O001O1O001O1O010O001O000000000010O0000000001O00000001O01O000000001O0001O000001O00000000001O00000O10001O00000000001eGPMX6P3dITM\\6l2_IYMa6g2\\I\\Md6d2[I]Me6c2[I]Me6d2ZI\\Mf6d2YI\\Mh6d2XI\\Mg6e2YI[Mg6e2XI\\Mh6e2WI[Mi6e2WI[Mi6f2UI[Mk6e2UI[Mk6f2SI[Ml6f2TIYMm6g2SIYMl6i2RIXMn6h2RIXMm6j2RIVMn6j2QIWMo6j2PIVMo6R3jHnLV7Z3aHfL_7c3YH]Lg7l3PHTLo7e406J9F9H2M4M2M4M3M2M4M3L3N1N10000O^NdIUL]6j3fITLY6l3jIRLV6m3mIQLR6o3QJoKo5P4UJmKj5S4YJkKg5T4\\JjKc5V4`JhK`5W4cJgK\\5Y4hJdKX5j2gIlMT1XOT5l2mIgMR1[OQ5m2QJdMQ1]Om4P3UJ_MQ1_Oj4Q3YJ\\MQ1@e4T3_JWMo0Cb4V3bJSMo0E^4X3gJoLn0G[4Y3kJlLm0IW4\\3PKfLm0KS4^3TKcLl0Mo3`3f3O010O0010O01O010O0010O01O0O2M2N3M3L4M3M3M3M3O1N2O1N2O1N2O1N2O1N2O0O2O1N2N2O1N2O1N2O1O11O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1M3M3M3L4M3L4M4O01O1O001N2N2N1N3N2N1O2N2N2M2O2N2N3M6J6J6I7J:F;E in this image.", "objects": [{"patches": [142, 143, 165, 166, 167, 188, 189, 190, 191, 210, 211, 212, 213, 234, 235, 236, 257, 258, 259, 280, 281, 282, 303, 304], "bbox": [0.154046875, 0.38535416666666666, 0.34315625, 0.7775625], "iscrowd": 0, "area": 8106.447750000001, "rle": {"size": [480, 640], "counts": "Ze^14l>1O1O1O0O100O100O10O0010O01O01HIfA8Y>JfA5Y>MfA4W>0hA0U>?M3M2N2N2O10O01SEoNk7R1oG]Oh7c0TH@l7a0RHEi7c0nGAQ8a0kGBS8b0gGAY8Q1_FoMHS1o8UOmFl17RNHo0R9WOlFj16WNFk0U9YOkFh14Jo8R1kFSOU9]3O0O10\\OoFXKQ9f4VGVKj8h4[GUKe8i4^GUKb8j4aGUK_8i4dGVK\\8g4gGXK[8e4hGZKX8d4S1L5L3M3M4L3L4RLVEc3V;L4M3gLaDQ3l0UMd9h2TF_Mk9c2lEeMT:[2hEiMW:Y2cElM]:[301O100O0001eLbER2^:kMiEQ2X:gMTFT2l9dM`FX2`9iMcFS2^9mMdFP2]9PNfFl1Z9UNhFh1Y9XNjFd1W9[NkFc1V9]NkFn0`NPOe:2mFf0jNROZ:7mFb0POTOT:9nF>QOYOQ:9oF:SO]Oo98PG6UOAk9:QG0WOFi99QGMYOJf99SGHZOOd98cHDa7;b3010O01O010O10O010O010O0100O010O010O10O0100O00100O2N2N2NVSU6"}, "label": "the man whipping the elephant"}]} {"id": 27070, "image": "COCO_train2014_000000027070.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in camo\"."}], "task": "refering", "answer_template": "The \"man in camo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 165, 166, 167, 188, 189, 190, 191, 210, 211, 212, 213, 234, 235, 236, 257, 258, 259, 280, 281, 282, 303, 304], "bbox": [0.154046875, 0.38535416666666666, 0.34315625, 0.7775625], "iscrowd": 0, "area": 8106.447750000001, "rle": {"size": [480, 640], "counts": "Ze^14l>1O1O1O0O100O100O10O0010O01O01HIfA8Y>JfA5Y>MfA4W>0hA0U>?M3M2N2N2O10O01SEoNk7R1oG]Oh7c0TH@l7a0RHEi7c0nGAQ8a0kGBS8b0gGAY8Q1_FoMHS1o8UOmFl17RNHo0R9WOlFj16WNFk0U9YOkFh14Jo8R1kFSOU9]3O0O10\\OoFXKQ9f4VGVKj8h4[GUKe8i4^GUKb8j4aGUK_8i4dGVK\\8g4gGXK[8e4hGZKX8d4S1L5L3M3M4L3L4RLVEc3V;L4M3gLaDQ3l0UMd9h2TF_Mk9c2lEeMT:[2hEiMW:Y2cElM]:[301O100O0001eLbER2^:kMiEQ2X:gMTFT2l9dM`FX2`9iMcFS2^9mMdFP2]9PNfFl1Z9UNhFh1Y9XNjFd1W9[NkFc1V9]NkFn0`NPOe:2mFf0jNROZ:7mFb0POTOT:9nF>QOYOQ:9oF:SO]Oo98PG6UOAk9:QG0WOFi99QGMYOJf99SGHZOOd98cHDa7;b3010O01O010O10O010O010O0100O010O010O10O0100O00100O2N2N2NVSU6"}, "label": "man in camo"}]} {"id": 27070, "image": "COCO_train2014_000000027070.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smaller elephant\"."}], "task": "refering", "answer_template": "The \"the smaller elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 126, 127, 128, 129, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 217, 218, 219, 220, 240, 241, 242, 243, 264, 265, 266], "bbox": [0.421015625, 0.2557083333333333, 0.6176875, 0.6773541666666667], "iscrowd": 0, "area": 13015.3952, "rle": {"size": [480, 640], "counts": "_hn34l>3M2N3M3M3M3M3M30O2O0016I8I7H8IM2M3N3L3M4M2M4L3N3L3M4M2M4L3N3L3MmU2>diMf0ZO`0@:G:E:G9G9G9H8G:F9G9G7J4K6J5K5K5L4K6I5L000000O10000O1000000O1000O10O10000O1000000O1\\N]GiLe8`2^HRMd7W2_I[Mc6^2S300001O000000001O2N1jF`Mk5a2SJ`Ml5b2RJ_Mm5b2QJ`Mo5a2oI`MP6b2nI_MR6a2mI`MR6b2lI^MT6X3VIhLk6P4\\HPLd7j4bGUK_8f51O000O2O0O4K5L4K5K5L4K5K5@`0hNW1gMZD8e<_Oc1I8Hneb3"}, "label": "the smaller elephant"}]} {"id": 27070, "image": "COCO_train2014_000000027070.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant with a long trunk stands next to another elephant\"."}], "task": "refering", "answer_template": "The \"an elephant with a long trunk stands next to another elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 79, 96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 142, 143, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 190, 191, 192, 194, 195, 213, 214, 215, 217, 218, 236, 237, 238, 240, 241, 260, 261, 262, 263, 264, 284, 287], "bbox": [0.187078125, 0.17527083333333332, 0.514046875, 0.7303333333333334], "iscrowd": 0, "area": 30430.465549999994, "rle": {"size": [480, 640], "counts": "P]h13k>5K4K5L4L5K4K5L5K4L4K5L5K3M0O01O2N2N2O12O1N1O2O1N2N2O1N1O2N2O1N2N2O0O2N2N2O1N2N101N2N2O1N1O2N2O1N2N2O0O2N2N2^EmLU9T3gFVMQ9l2jF`Mm8a2oFjMj8W2RGSNf8o1UG]Nb8e1ZGeN_8\\1\\GoN\\8P4J0010O00010YNkGbLU8_3oG\\LQ8e3THVLk7k3YHPLh7P4\\HkKe7U4`HfK`7Z4dHaK^7^4gH\\KZ7e4iHWKW7i4mHRKT7n4_1000gG]KU6c4`IiK_6W4WITLh6l3nH_LR7`3cHlL\\7]5O1O1O1O2N1O1O100O1O1O1O1O1O1O1O100O2N1O1O1O1O1O1O01O0O100O100O10000`MPIoLR7j2XIRMh6h2cIRM^6i2lIRMT6h2WJRMj5h2`JTM`5g2jJTMV5f2UKTMl4g2^KTMb4i2eKRM]4l2iKoLW4Q3nKiLS4V3SLeLm3[3WL`Lj3_3\\L\\Ld3d3aLWL_3i3l3O101N100O10000O100O101N10000O100O100O1010O3M4M3M2N3L3N3M3M2N3M2N3M3M2M4M2N3M3M2N3M2N3L4M2N3M2NX1hN00O2O:F;D;F;E:E in this image.", "objects": [{"patches": [56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 102, 103, 104, 105, 106, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 217, 218], "bbox": [0.433625, 0.19238961038961036, 0.6485, 0.6721558441558441], "iscrowd": 0, "area": 22610.6268, "rle": {"size": [385, 640], "counts": "PdX3V4g76I6K6I6K5J7L3N2O001O0O2O001O001O00001O001O001O001OO1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O10000O100000000O010000000O100000000O1000000000000001O00001O001O00001O00001O00001O00001O00001OO10000000000000000000000O100000000000000000000O1N2O1N20003M3M3L4M4L3M3M3M4L3L4M3Mo2QM in this image.", "objects": [{"patches": [56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 102, 103, 104, 105, 106, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 217, 218], "bbox": [0.433625, 0.19238961038961036, 0.6485, 0.6721558441558441], "iscrowd": 0, "area": 22610.6268, "rle": {"size": [385, 640], "counts": "PdX3V4g76I6K6I6K5J7L3N2O001O0O2O001O001O00001O001O001O001OO1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O10000O100000000O010000000O100000000O1000000000000001O00001O001O00001O00001O00001O00001O00001OO10000000000000000000000O100000000000000000000O1N2O1N20003M3M3L4M4L3M3M3M4L3L4M3Mo2QM in this image.", "objects": [{"patches": [108, 109, 110, 111, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 200, 201, 202], "bbox": [0.64909375, 0.30664935064935067, 0.8545468750000002, 0.6081298701298701], "iscrowd": 0, "area": 11631.085149999992, "rle": {"size": [385, 640], "counts": "QVl41l;5N3M1aDOn:2PEOP;ORE1n:MTE3o:JQE6R;GmD2J26;IC\\:NPF5Ck0X:YOlET1o9`0M2O1N2N2N1O00002OOYFjMe9X211O1O2O0O10000O10?@2O1O0O100O100O10000OK6L310000O100O100O1O1O100O2N1O1O1O1O1O100O100000000000000000000000O1000O01001O2NO1O100001O0000001O2N2N1O1O000000000000000001O1O1O1O1O1O1N1O1O1O2N1O1L4O103M3M3M4L3M3L4M4L3L4M6I;E;^Oa0@SiR1"}, "label": "the back of a white van"}]} {"id": 346562, "image": "COCO_train2014_000000346562.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white van\"."}], "task": "refering", "answer_template": "The \"the white van\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 111, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 200, 201, 202], "bbox": [0.64909375, 0.30664935064935067, 0.8545468750000002, 0.6081298701298701], "iscrowd": 0, "area": 11631.085149999992, "rle": {"size": [385, 640], "counts": "QVl41l;5N3M1aDOn:2PEOP;ORE1n:MTE3o:JQE6R;GmD2J26;IC\\:NPF5Ck0X:YOlET1o9`0M2O1N2N2N1O00002OOYFjMe9X211O1O2O0O10000O10?@2O1O0O100O100O10000OK6L310000O100O100O1O1O100O2N1O1O1O1O1O100O100000000000000000000000O1000O01001O2NO1O100001O0000001O2N2N1O1O000000000000000001O1O1O1O1O1O1N1O1O1O2N1O1L4O103M3M3M4L3M3L4M4L3L4M6I;E;^Oa0@SiR1"}, "label": "the white van"}]} {"id": 215493, "image": "COCO_train2014_000000215493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white horse\"."}], "task": "refering", "answer_template": "The \"the white horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 68, 90, 91, 112, 113, 114, 134, 135, 136, 137, 154, 155, 157, 158, 159, 160, 177, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 265, 271, 272, 273, 274, 275, 287, 288, 289, 294, 295, 296, 297, 298, 310, 311, 312, 313, 318, 319, 320, 335, 336, 341, 342, 343, 354, 355, 356, 357, 358, 359, 360, 361, 362, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.40540624999999997, 0.06758333333333333, 1.0, 0.9864791666666667], "iscrowd": 0, "area": 41156.84595, "rle": {"size": [480, 640], "counts": "^ki31m>2N2N2N2N2N2O1O1O1O100O100O2O0O10000O100O100O100O10000O100O100O101O0O1000000000000000000000000O100000001O0O10000000000O1000000fMoNeFQ1[9RObFn0^9UO^Fl0b9WOZFj0f9YOWFg0h9_ORFb0m9@QFa0o9_OPFb0o9@PFa0n9@QFa0o9_OPFb0o9@oEa0R9>nFBR9>mFCT9h:QOVEb01=d;E\\D:c;H\\D8d;H]D7b;K]D5c;L\\D4c;N]D1c;O]D1b;1]DOc;2]DMb;5^DJb;6`DH_;:aDE_; in this image.", "objects": [{"patches": [45, 68, 90, 91, 112, 113, 114, 134, 135, 136, 137, 154, 155, 157, 158, 159, 160, 177, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 265, 271, 272, 273, 274, 275, 287, 288, 289, 294, 295, 296, 297, 298, 310, 311, 312, 313, 318, 319, 320, 335, 336, 341, 342, 343, 354, 355, 356, 357, 358, 359, 360, 361, 362, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.40540624999999997, 0.06758333333333333, 1.0, 0.9864791666666667], "iscrowd": 0, "area": 41156.84595, "rle": {"size": [480, 640], "counts": "^ki31m>2N2N2N2N2N2O1O1O1O100O100O2O0O10000O100O100O100O10000O100O100O101O0O1000000000000000000000000O100000001O0O10000000000O1000000fMoNeFQ1[9RObFn0^9UO^Fl0b9WOZFj0f9YOWFg0h9_ORFb0m9@QFa0o9_OPFb0o9@PFa0n9@QFa0o9_OPFb0o9@oEa0R9>nFBR9>mFCT9h:QOVEb01=d;E\\D:c;H\\D8d;H]D7b;K]D5c;L\\D4c;N]D1c;O]D1b;1]DOc;2]DMb;5^DJb;6`DH_;:aDE_; in this image.", "objects": [{"patches": [36, 53, 54, 55, 56, 57, 58, 59, 60, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 289, 290, 291, 292, 293, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 313, 314, 315, 316, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 337, 338, 339, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 361, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0, 0.1096875, 0.77096875, 0.9957083333333333], "iscrowd": 0, "area": 148208.5634, "rle": {"size": [480, 640], "counts": "n5U1j=T4lK8H9G8I8G8H9G8I7I8H7J7H7I7I2O1N2N2N2O1N2N2N2O0O0001O0000001O0000O1O100O1O100O1O100O100O2O0O100O100O100O100O100O1O1N2O1N2N2O1N2N2O1N2N2O1N2N2N2O1N2N3N1N2N2O1N2N200O100O100O100O100O100O100O100O100O100O100O100O101N100O100O100O100O100O100O100O100O100O100O100O100O100O100O101N100O100O100O100O100O100O100O100O100O100O100O1O100O10000O100O2O000O100O10000O10000O100O10000O100O10O10O10000O100O1000O0100O10000O100000O10O100000000O1000O100000O10000000O0100000000O100000O1000O1001O0000001N10001O00001O00001N10001O000000AjCUNU in this image.", "objects": [{"patches": [36, 53, 54, 55, 56, 57, 58, 59, 60, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 289, 290, 291, 292, 293, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 313, 314, 315, 316, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 337, 338, 339, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 361, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0, 0.1096875, 0.77096875, 0.9957083333333333], "iscrowd": 0, "area": 148208.5634, "rle": {"size": [480, 640], "counts": "n5U1j=T4lK8H9G8I8G8H9G8I7I8H7J7H7I7I2O1N2N2N2O1N2N2N2O0O0001O0000001O0000O1O100O1O100O1O100O100O2O0O100O100O100O100O100O1O1N2O1N2N2O1N2N2O1N2N2O1N2N2N2O1N2N3N1N2N2O1N2N200O100O100O100O100O100O100O100O100O100O100O100O101N100O100O100O100O100O100O100O100O100O100O100O100O100O100O101N100O100O100O100O100O100O100O100O100O100O100O1O100O10000O100O2O000O100O10000O10000O100O10000O100O10O10O10000O100O1000O0100O10000O100000O10O100000000O1000O100000O10000000O0100000000O100000O1000O1001O0000001N10001O00001O00001N10001O000000AjCUNU in this image.", "objects": [{"patches": [149, 150, 151, 152, 163, 164, 165, 166, 167, 168, 169, 180, 181, 182, 183, 184, 185, 186, 203, 220], "bbox": [0.6137708333333334, 0.341609375, 0.9963541666666667, 0.5328593749999999], "iscrowd": 0, "area": 9525.640300000001, "rle": {"size": [640, 480], "counts": "lch5b0^c0>B000O10001O00000000001N31O5K4K1OO0000001O0000000000001O0000000000001O000O10000000001O00000000000`NROn_Oo0Q`0QOo_Oo0Q`0QOo_Oo0aa0O100O10aNSOk_Ol0V`0TOj_Ok0W`0UOi_Ok0W`0UOi_Oj0X`0VOh_Oj0X`0VOh_Oi0Y`0WOg_Oi0Y`0WOg_Oh0[`0XOd_Oh0\\`0XOd_Og0]`0YOc_Og0]`0YOc_Of0^`0ZOb_Of0^`0\\O`_Oc0a`0_O]_Oa0c`0BZ_O=g`0FV_O:j`0IS_O6n`0Mo^O3Qa00l^OOVa02h^ONXa05e^OJ\\a09a^OF`a0=]^OCca0`0Y^O@ha0c0U^O]Oka0e0S^OZOna0i0o]OXOPb0\\10000001O000001O0000000000001O01O0000000001O000000BP^OlNPb0S1S^OkNma0T1V^OjNka0T1X^OkNga0T1]^OiNca0W1_^OgNaa0X1d0O1O1O12N1O2N1O2O0O2N2N1O2N1O2N1O2N0000010O000000001O0000001O0001O0001O0000001O000b^OXNd`0h1V_OgNa`0Z1]_OnN\\`0R1c_OoN^`0P1a_ORO^`0n0a_OSO_`0P1]_OROb`0V201O001O001O001O010O001O00O1O100O1O10jP1"}, "label": "the black couch with the green pillow on it"}]} {"id": 100812, "image": "COCO_train2014_000000100812.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black couch in the back side of the image\"."}], "task": "refering", "answer_template": "The \"black couch in the back side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 152, 163, 164, 165, 166, 167, 168, 169, 180, 181, 182, 183, 184, 185, 186, 203, 220], "bbox": [0.6137708333333334, 0.341609375, 0.9963541666666667, 0.5328593749999999], "iscrowd": 0, "area": 9525.640300000001, "rle": {"size": [640, 480], "counts": "lch5b0^c0>B000O10001O00000000001N31O5K4K1OO0000001O0000000000001O0000000000001O000O10000000001O00000000000`NROn_Oo0Q`0QOo_Oo0Q`0QOo_Oo0aa0O100O10aNSOk_Ol0V`0TOj_Ok0W`0UOi_Ok0W`0UOi_Oj0X`0VOh_Oj0X`0VOh_Oi0Y`0WOg_Oi0Y`0WOg_Oh0[`0XOd_Oh0\\`0XOd_Og0]`0YOc_Og0]`0YOc_Of0^`0ZOb_Of0^`0\\O`_Oc0a`0_O]_Oa0c`0BZ_O=g`0FV_O:j`0IS_O6n`0Mo^O3Qa00l^OOVa02h^ONXa05e^OJ\\a09a^OF`a0=]^OCca0`0Y^O@ha0c0U^O]Oka0e0S^OZOna0i0o]OXOPb0\\10000001O000001O0000000000001O01O0000000001O000000BP^OlNPb0S1S^OkNma0T1V^OjNka0T1X^OkNga0T1]^OiNca0W1_^OgNaa0X1d0O1O1O12N1O2N1O2O0O2N2N1O2N1O2N1O2N0000010O000000001O0000001O0001O0001O0000001O000b^OXNd`0h1V_OgNa`0Z1]_OnN\\`0R1c_OoN^`0P1a_ORO^`0n0a_OSO_`0P1]_OROb`0V201O001O001O001O010O001O00O1O100O1O10jP1"}, "label": "black couch in the back side of the image"}]} {"id": 100812, "image": "COCO_train2014_000000100812.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white chair on the side of the table closest to the cat\"."}], "task": "refering", "answer_template": "The \"a white chair on the side of the table closest to the cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 222, 223, 224, 225, 238, 239, 240, 241, 242, 255, 256, 257, 258, 273, 274, 275, 276, 277, 278, 292, 293, 294, 295, 296, 309, 310, 311, 312, 313, 326, 327, 328, 329], "bbox": [0.025812500000000002, 0.555453125, 0.4444375, 0.8565156250000001], "iscrowd": 0, "area": 15410.302199999993, "rle": {"size": [640, 480], "counts": "jl71mc03L5L3L5K4N3N1N3N1O2M2O1N2O1N2O1N2O1N2O0O2O1N2O1N2O1O1N2O1N2O001O0O101O001N10001N10001O0O2O00001N10001N1000000O1000000O1O1O010O1O100O1O1TA`MX<`2gCaMX<`2fCcMX<^2gCcMX<_2fCbMZ<_2dCbM[<_2dCcMZ<_2dCbM\\<^2bCdM]<^2aCcM^<_2`CcM_<]2`CdM_<^2_CcM`<^2^CeM`<]2^CdMb<]2\\CdMc<]2\\CdMc<^2[CdMd<\\2ZCfMe<\\2YCeMf<]2XCdMh<\\2WCfMg<\\2UCgMj<[2QCiMo01O0000000000001O0000000000000000001O000000000000000000000O2O000O10000O10000O10000O10000O10000O10000O100O10000O1000O010000O100O1O1O1O1O1N2O2N1O2N1N3N1O2N1N3N1O1O2N1N3N1L5H7J7I6JZ]V5"}, "label": "a white chair on the side of the table closest to the cat"}]} {"id": 100812, "image": "COCO_train2014_000000100812.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white chair on the left\"."}], "task": "refering", "answer_template": "The \"the white chair on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 222, 223, 224, 225, 238, 239, 240, 241, 242, 255, 256, 257, 258, 273, 274, 275, 276, 277, 278, 292, 293, 294, 295, 296, 309, 310, 311, 312, 313, 326, 327, 328, 329], "bbox": [0.025812500000000002, 0.555453125, 0.4444375, 0.8565156250000001], "iscrowd": 0, "area": 15410.302199999993, "rle": {"size": [640, 480], "counts": "jl71mc03L5L3L5K4N3N1N3N1O2M2O1N2O1N2O1N2O1N2O0O2O1N2O1N2O1O1N2O1N2O001O0O101O001N10001N10001O0O2O00001N10001N1000000O1000000O1O1O010O1O100O1O1TA`MX<`2gCaMX<`2fCcMX<^2gCcMX<_2fCbMZ<_2dCbM[<_2dCcMZ<_2dCbM\\<^2bCdM]<^2aCcM^<_2`CcM_<]2`CdM_<^2_CcM`<^2^CeM`<]2^CdMb<]2\\CdMc<]2\\CdMc<^2[CdMd<\\2ZCfMe<\\2YCeMf<]2XCdMh<\\2WCfMg<\\2UCgMj<[2QCiMo01O0000000000001O0000000000000000001O000000000000000000000O2O000O10000O10000O10000O10000O10000O10000O100O10000O1000O010000O100O1O1O1O1O1N2O2N1O2N1N3N1O2N1N3N1O1O2N1N3N1L5H7J7I6JZ]V5"}, "label": "the white chair on the left"}]} {"id": 305624, "image": "COCO_train2014_000000305624.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tugboat tied up at a dock in chilly weather\"."}], "task": "refering", "answer_template": "The \"a tugboat tied up at a dock in chilly weather\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 182, 183], "bbox": [0.76240625, 0.0835042735042735, 0.9925, 0.5613675213675213], "iscrowd": 0, "area": 11538.424999999997, "rle": {"size": [351, 640], "counts": "W^W54j:2N2N1K5H8_O]OfFc0Z9^OdFc0[9^O`Fg0`9<010kN_Fj0b9TO]Fn0c9QO^Fo0i91O00POTFm0o900001O01O100O000001N1L4O101N1O100O1O100O1O100O1O1O001N2O1O1O1O1ZO]NnGe1Q8\\NnGe1R8\\NkGg1S8ZNlGg1S8ZNkGh1U8XNjGi1U8XNiGj1V8YNfGi1Z8WNeGj1Z8WNdGk1\\8UNcGl1\\8 in this image.", "objects": [{"patches": [44, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 182, 183], "bbox": [0.76240625, 0.0835042735042735, 0.9925, 0.5613675213675213], "iscrowd": 0, "area": 11538.424999999997, "rle": {"size": [351, 640], "counts": "W^W54j:2N2N1K5H8_O]OfFc0Z9^OdFc0[9^O`Fg0`9<010kN_Fj0b9TO]Fn0c9QO^Fo0i91O00POTFm0o900001O01O100O000001N1L4O101N1O100O1O100O1O100O1O1O001N2O1O1O1O1ZO]NnGe1Q8\\NnGe1R8\\NkGg1S8ZNlGg1S8ZNkGh1U8XNjGi1U8XNiGj1V8YNfGi1Z8WNeGj1Z8WNdGk1\\8UNcGl1\\8 in this image.", "objects": [{"patches": [207, 208, 209, 210, 211, 212, 213, 214, 215, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 289, 290], "bbox": [0.006140625, 0.7154985754985754, 0.657875, 0.9621367521367521], "iscrowd": 0, "area": 21641.594600000004, "rle": {"size": [351, 640], "counts": "]e18W:`0_Oa0@`0K500000000001O000000000000000000000000000000000000001O000000000000000000000000000000000000001O00000000000000000000000000001O00001O001O00001O00001O001O00001O001O00001O001O00001O00001O001O00000000000000000000001O00000000000000000000000000000000000000000000000000O1000O100000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000001O00000000000000O1000000000000O10000000000O1000000000000001O00001O00001O001O00001O001O00001OO1000000000001O000000000O10000000000000000000001N10000000000000000000000O100O1N2N2O2M2O1000000000000000000O10000000001O00000000000000000000000000000O101O000000000000000000000000000000001O0O1000000000000000000000000000001O000000000O10000000000000000000001O1O>B1O00001O001N1000000000CjGQNV8l1THmMm7o1d0L4N101O1O1O1O1O1N101O1O1O1O1O1N2O1O2N1O2N1J7I7I6J7H_iZ2"}, "label": "a bench on a pier with a couple sitting on it"}]} {"id": 305624, "image": "COCO_train2014_000000305624.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bench with people sitting on it\"."}], "task": "refering", "answer_template": "The \"a bench with people sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 210, 211, 212, 213, 214, 215, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 289, 290], "bbox": [0.006140625, 0.7154985754985754, 0.657875, 0.9621367521367521], "iscrowd": 0, "area": 21641.594600000004, "rle": {"size": [351, 640], "counts": "]e18W:`0_Oa0@`0K500000000001O000000000000000000000000000000000000001O000000000000000000000000000000000000001O00000000000000000000000000001O00001O001O00001O00001O001O00001O001O00001O001O00001O00001O001O00000000000000000000001O00000000000000000000000000000000000000000000000000O1000O100000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000001O00000000000000O1000000000000O10000000000O1000000000000001O00001O00001O001O00001O001O00001OO1000000000001O000000000O10000000000000000000001N10000000000000000000000O100O1N2N2O2M2O1000000000000000000O10000000001O00000000000000000000000000000O101O000000000000000000000000000000001O0O1000000000000000000000000000001O000000000O10000000000000000000001O1O>B1O00001O001N1000000000CjGQNV8l1THmMm7o1d0L4N101O1O1O1O1O1N101O1O1O1O1O1N2O1O2N1O2N1J7I7I6J7H_iZ2"}, "label": "a bench with people sitting on it"}]} {"id": 305624, "image": "COCO_train2014_000000305624.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty bench to the right of another bench that has people on it\"."}], "task": "refering", "answer_template": "The \"an empty bench to the right of another bench that has people on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 298], "bbox": [0.627796875, 0.7332193732193732, 0.99734375, 0.9810541310541311], "iscrowd": 0, "area": 13076.760449999994, "rle": {"size": [351, 640], "counts": "cQZ41n:1O2N1O1O1O2N1O1O2N1O2N5K4L5K4L5K4L5K5K6J6J4L4L5K4L1OO100O100O100O100XOaGoN`8Q1eGjN[8V1iGeNX8[1lGaNT8_1QH[NP8e1e0000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000002N5K4L4L4L1O00001O0000001N10XOhGiNX8W1QHaNn7_1i00008H`0@`0@e<"}, "label": "an empty bench to the right of another bench that has people on it"}]} {"id": 305624, "image": "COCO_train2014_000000305624.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bench on the right\"."}], "task": "refering", "answer_template": "The \"the bench on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 298], "bbox": [0.627796875, 0.7332193732193732, 0.99734375, 0.9810541310541311], "iscrowd": 0, "area": 13076.760449999994, "rle": {"size": [351, 640], "counts": "cQZ41n:1O2N1O1O1O2N1O1O2N1O2N5K4L5K4L5K4L5K5K6J6J4L4L5K4L1OO100O100O100O100XOaGoN`8Q1eGjN[8V1iGeNX8[1lGaNT8_1QH[NP8e1e0000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000002N5K4L4L4L1O00001O0000001N10XOhGiNX8W1QHaNn7_1i00008H`0@`0@e<"}, "label": "the bench on the right"}]} {"id": 305624, "image": "COCO_train2014_000000305624.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty bench in a park along a waterfront with people and boats\"."}], "task": "refering", "answer_template": "The \"an empty bench in a park along a waterfront with people and boats\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 298], "bbox": [0.627796875, 0.7332193732193732, 0.99734375, 0.9810541310541311], "iscrowd": 0, "area": 13076.760449999994, "rle": {"size": [351, 640], "counts": "cQZ41n:1O2N1O1O1O2N1O1O2N1O2N5K4L5K4L5K4L5K5K6J6J4L4L5K4L1OO100O100O100O100XOaGoN`8Q1eGjN[8V1iGeNX8[1lGaNT8_1QH[NP8e1e0000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000002N5K4L4L4L1O00001O0000001N10XOhGiNX8W1QHaNn7_1i00008H`0@`0@e<"}, "label": "an empty bench in a park along a waterfront with people and boats"}]} {"id": 305624, "image": "COCO_train2014_000000305624.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couple sitting close together on the bench\"."}], "task": "refering", "answer_template": "The \"the couple sitting close together on the bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 119, 120, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 186, 187, 188, 189, 191, 209, 210, 211, 212], "bbox": [0.104609375, 0.33, 0.33390624999999996, 0.7555270655270655], "iscrowd": 0, "area": 10544.124399999997, "rle": {"size": [351, 640], "counts": "WVg04j:5K4L3M2N2N2N2N3M2N2M3N2N2M3M3M3L3N3M2N2N3M2N3M2N3M2M3N2N2N2N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1O100O1000000O10000O10000O10000O1^ObLcI^3Y6iLcIX3Y6oLbIS3\\6RM_IP3`6h0O1O1O1O100O1O1000000000N101POPJPMQ6m2R1N2M3O2M2N3N1N3M2O2M2O2_NiG6[8GgG7[8GfG9\\8DfG<[8AfG?[8^OhGa0Z8ZOiGg0X9O0010O0100O1O100O00100O1O100O10O01O100O1O100O001O01O01O000010O0001O00010O003M2N2N2O2M2N2N3MWja4"}, "label": "the couple sitting close together on the bench"}]} {"id": 305624, "image": "COCO_train2014_000000305624.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couple hugging each other\"."}], "task": "refering", "answer_template": "The \"the couple hugging each other\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 119, 120, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 186, 187, 188, 189, 191, 209, 210, 211, 212], "bbox": [0.104609375, 0.33, 0.33390624999999996, 0.7555270655270655], "iscrowd": 0, "area": 10544.124399999997, "rle": {"size": [351, 640], "counts": "WVg04j:5K4L3M2N2N2N2N3M2N2M3N2N2M3M3M3L3N3M2N2N3M2N3M2N3M2M3N2N2N2N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1O100O1000000O10000O10000O10000O1^ObLcI^3Y6iLcIX3Y6oLbIS3\\6RM_IP3`6h0O1O1O1O100O1O1000000000N101POPJPMQ6m2R1N2M3O2M2N3N1N3M2O2M2O2_NiG6[8GgG7[8GfG9\\8DfG<[8AfG?[8^OhGa0Z8ZOiGg0X9O0010O0100O1O100O00100O1O100O10O01O100O1O100O001O01O01O000010O0001O00010O003M2N2N2O2M2N2N3MWja4"}, "label": "the couple hugging each other"}]} {"id": 158195, "image": "COCO_train2014_000000158195.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"laptop the person with black suit is operating\"."}], "task": "refering", "answer_template": "The \"laptop the person with black suit is operating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 211, 212, 213, 214, 215, 216, 217, 226, 227, 228, 229, 230, 231, 232, 233, 241, 242, 243, 244, 245, 246, 247, 248, 258, 259, 260, 261, 262, 274, 275], "bbox": [0.11761682242990655, 0.582015625, 0.6048598130841122, 0.8022500000000001], "iscrowd": 0, "area": 14629.479000000008, "rle": {"size": [640, 428], "counts": "mdo02Q10`a0b0g]OOoa0W1E:F:001O0010O01O10O00100O010O0010O010O01O010O100O1O2O0O100O010O0010O0010O0010O10O10O10O01000O010O10O10O10O10O10O10O10000O01000O10000O10000O10000O10000O10O010000O10000O10000O1000O001N2O0O10O010O0100O01000O010O010O10O001O010O00100O0010O01O0010O010O010O10O10O10O010O0100O010O10O010O01000O0100O103L3N3L4M3L3N3L4M2N3L4M3L3N3L4M3L3N1N100O10O10O100O100O010O100O100O010O01O1O001O001O010O001O001O001O12M4M3L4M2M4M3L4M3L4MfUY3"}, "label": "laptop the person with black suit is operating"}]} {"id": 158195, "image": "COCO_train2014_000000158195.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a laptop that the man is using\"."}], "task": "refering", "answer_template": "The \"a laptop that the man is using\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 211, 212, 213, 214, 215, 216, 217, 226, 227, 228, 229, 230, 231, 232, 233, 241, 242, 243, 244, 245, 246, 247, 248, 258, 259, 260, 261, 262, 274, 275], "bbox": [0.11761682242990655, 0.582015625, 0.6048598130841122, 0.8022500000000001], "iscrowd": 0, "area": 14629.479000000008, "rle": {"size": [640, 428], "counts": "mdo02Q10`a0b0g]OOoa0W1E:F:001O0010O01O10O00100O010O0010O010O01O010O100O1O2O0O100O010O0010O0010O0010O10O10O10O01000O010O10O10O10O10O10O10O10000O01000O10000O10000O10000O10000O10O010000O10000O10000O1000O001N2O0O10O010O0100O01000O010O010O10O001O010O00100O0010O01O0010O010O010O10O10O10O010O0100O010O10O010O01000O0100O103L3N3L4M3L3N3L4M2N3L4M3L3N3L4M3L3N1N100O10O10O100O100O010O100O100O010O01O1O001O001O010O001O001O001O12M4M3L4M2M4M3L4M3L4MfUY3"}, "label": "a laptop that the man is using"}]} {"id": 35322, "image": "COCO_train2014_000000035322.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a potted tree leaning to the right\"."}], "task": "refering", "answer_template": "The \"a potted tree leaning to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 210, 211, 212, 213], "bbox": [0.14060937499999998, 0.0, 0.824171875, 0.6337089201877933], "iscrowd": 0, "area": 67035.16189999999, "rle": {"size": [426, 640], "counts": "P`U11Y=0O2O0O2N2O0O2O0O2O1N101PIEn0Q5BYJ0d0a0R5@YJOd0b0S5b1lJ`NS5a1lJ`NT5`1kJaNT5`1jJbNV5]1jJeNU5[1jJfNU5[1fJjNZ5U1cJoN]5Q1_JSOU5iLoJc7Q5;01O0O10000O10001N1000000O101O0O1000O0O2O1O001N2O1O001O1N101O1O1O2M2O1O2N1N2O1O2N1O1N3N1O100O2O000O101O0O101N10000O2O0O101O0O10001N100O2O000O101N100O2O0O100O100O100O010O10000O010O100O010O100O105J3N1N10000O100O2O0O100O010O01O010O0010O01000O100O2O0O101N100O2O0O101N1O101N100O10O010O0100O010O10O010O02O0O100O2O0O100O2O0O101N100O101N100O101N100O101N100O2O0O100O2O0O101N100O2O0O2N101N100O2O0O2O0O2O0O101N10L3E<0O100O010O100O1O010O100O10O0100O10O001N1O2O1O5L4L4L4L4L4L00O1O1O1N2O1O1O1O1O1N2O1O1O1O12N1O1O1O2N1O1O2N1O1O1O2N1OM3L4O1ZOlGWLS8i3nGVLR8j3oGULQ8l3oGSLQ8m3PHRLP8n3RHPLn7P4SHoKm7R4SHmKm7S4THlKl7T4UHkKk7V4UHiKk7W4WHhKh7X4YHgKg7Y4ZHfKf7[4ZHdKg7[4ZHdKf7\\4\\HbKd7^4]HaKc7`4]H_Kc7a4^H^Kb7b4c01O0000001O0000001O00001O00A_GULa8k3cGQL]8o3gGnKX8R4mGiKS8W4`000eMPG in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 210, 211, 212, 213], "bbox": [0.14060937499999998, 0.0, 0.824171875, 0.6337089201877933], "iscrowd": 0, "area": 67035.16189999999, "rle": {"size": [426, 640], "counts": "P`U11Y=0O2O0O2N2O0O2O0O2O1N101PIEn0Q5BYJ0d0a0R5@YJOd0b0S5b1lJ`NS5a1lJ`NT5`1kJaNT5`1jJbNV5]1jJeNU5[1jJfNU5[1fJjNZ5U1cJoN]5Q1_JSOU5iLoJc7Q5;01O0O10000O10001N1000000O101O0O1000O0O2O1O001N2O1O001O1N101O1O1O2M2O1O2N1N2O1O2N1O1N3N1O100O2O000O101O0O101N10000O2O0O101O0O10001N100O2O000O101N100O2O0O100O100O100O010O10000O010O100O010O100O105J3N1N10000O100O2O0O100O010O01O010O0010O01000O100O2O0O101N100O2O0O101N1O101N100O10O010O0100O010O10O010O02O0O100O2O0O100O2O0O101N100O101N100O101N100O101N100O2O0O100O2O0O101N100O2O0O2N101N100O2O0O2O0O2O0O101N10L3E<0O100O010O100O1O010O100O10O0100O10O001N1O2O1O5L4L4L4L4L4L00O1O1O1N2O1O1O1O1O1N2O1O1O1O12N1O1O1O2N1O1O2N1O1O1O2N1OM3L4O1ZOlGWLS8i3nGVLR8j3oGULQ8l3oGSLQ8m3PHRLP8n3RHPLn7P4SHoKm7R4SHmKm7S4THlKl7T4UHkKk7V4UHiKk7W4WHhKh7X4YHgKg7Y4ZHfKf7[4ZHdKg7[4ZHdKf7\\4\\HbKd7^4]HaKc7`4]H_Kc7a4^H^Kb7b4c01O0000001O0000001O00001O00A_GULa8k3cGQL]8o3gGnKX8R4mGiKS8W4`000eMPG in this image.", "objects": [{"patches": [145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 184, 185, 186, 187], "bbox": [0.0, 0.3589583333333334, 0.74440625, 0.4940625], "iscrowd": 0, "area": 9450.401949999998, "rle": {"size": [480, 640], "counts": "b6g0Y>4L0000O100000000O10O10000000O100000000000000000000O10000000000000000000000O10O100000000000000000000O010000000000000000000O10O100000000000000000O1000O10000000000000O10000000O100000000000O1000000000O1000000000O10000000000000O100000O1000000000000000O10O10000000000000000000O0100000000000000000000O010000000000000000000O01J6K5J6Khi:0ZVE3M3M2M4M3N20000O10000000O10000000O100000000000O1000O10000000000000O1000O10000000000000O10O10000000000000000O0100000000000000000O01000000000000000O10O10000000000000O0100000000000000O0100000000000O100000O10000000O10000000O1000O100000000000O10O1000000000000O10O100000000000O1000O1000000000O100000O100000O1000000000O1000O1000000000000O010000000000000O10O1000000000000000O0100000000000000000O1000O10000000000000O100000O100000000000O10000000O1000000000O1000000000O10000000O100000000000O100000O1000000000000000O10O101O000000000000000O1000000000000000000O2O00000000000000000O1000006J\\f\\2"}, "label": "3rd train from front"}]} {"id": 92685, "image": "COCO_train2014_000000092685.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"orange train on furthest track\"."}], "task": "refering", "answer_template": "The \"orange train on furthest track\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 184, 185, 186, 187], "bbox": [0.0, 0.3589583333333334, 0.74440625, 0.4940625], "iscrowd": 0, "area": 9450.401949999998, "rle": {"size": [480, 640], "counts": "b6g0Y>4L0000O100000000O10O10000000O100000000000000000000O10000000000000000000000O10O100000000000000000000O010000000000000000000O10O100000000000000000O1000O10000000000000O10000000O100000000000O1000000000O1000000000O10000000000000O100000O1000000000000000O10O10000000000000000000O0100000000000000000000O010000000000000000000O01J6K5J6Khi:0ZVE3M3M2M4M3N20000O10000000O10000000O100000000000O1000O10000000000000O1000O10000000000000O10O10000000000000000O0100000000000000000O01000000000000000O10O10000000000000O0100000000000000O0100000000000O100000O10000000O10000000O1000O100000000000O10O1000000000000O10O100000000000O1000O1000000000O100000O100000O1000000000O1000O1000000000000O010000000000000O10O1000000000000000O0100000000000000000O1000O10000000000000O100000O100000000000O10000000O1000000000O1000000000O10000000O100000000000O100000O1000000000000000O10O101O000000000000000O1000000000000000000O2O00000000000000000O1000006J\\f\\2"}, "label": "orange train on furthest track"}]} {"id": 92685, "image": "COCO_train2014_000000092685.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a train between two other trains\"."}], "task": "refering", "answer_template": "The \"a train between two other trains\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 155, 156, 157, 158, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 207, 208, 209, 210, 211, 212], "bbox": [0.00010937500000000002, 0.3917291666666667, 0.90228125, 0.5913333333333334], "iscrowd": 0, "area": 20510.672749999998, "rle": {"size": [480, 640], "counts": "]72n>4L4K6K4L4L4L5K4L4L5K2NO01000000000O10000000O0100000000O1000000000O0100000000000000O10O100000000000O100000O10000000O1000000000O1000O100000000000000O010000000000000O1000O1000000000O10000000O100000O100000000000O10O100000000000000O10O100000000000O100000O10000000O1000000000O1000O10000000000000O0100000000000000O1000O1000000000O106J in this image.", "objects": [{"patches": [138, 139, 140, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 253, 254, 255, 276, 277, 278, 299, 300, 301, 322, 323, 345, 346, 368, 369, 370], "bbox": [0.0, 0.35870833333333335, 0.49507812500000004, 1.0], "iscrowd": 0, "area": 30041.2269, "rle": {"size": [480, 640], "counts": "`5S2l in this image.", "objects": [{"patches": [138, 139, 140, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 253, 254, 255, 276, 277, 278, 299, 300, 301, 322, 323, 345, 346, 368, 369, 370], "bbox": [0.0, 0.35870833333333335, 0.49507812500000004, 1.0], "iscrowd": 0, "area": 30041.2269, "rle": {"size": [480, 640], "counts": "`5S2l in this image.", "objects": [{"patches": [50, 51, 52, 53, 71, 72, 73, 74, 75, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 169, 170, 171, 172, 189, 191, 192, 193], "bbox": [0.020921875, 0.11391666666666667, 0.5030781249999999, 0.498875], "iscrowd": 0, "area": 23990.278599999998, "rle": {"size": [480, 640], "counts": "_g63k>4M2M4L3N3L3N1N2O1N2O1N2I7H8M3O001O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O0O2O1O2N1O2N1O1O2M201O001O001O001O0010O01O00001O001O0O2O001O1O001O001O1N101O1O001O1O001O0O2O1O001O1O001O1O0O2O001O00001N1O1O2N1O1O2O0O101O000O1O2M2O1N3N1N2O2M3M2O2M3M3N1N3M3N1N3M3N1N3M3N1N3N2O001O00001O0000001O0000O001O1O1N2O1O1O1O1N2O1UMZF:h9FXF9i9GXF6j9IXF5i9KXF3i9LXF2j9NWF0j90WFMk92WFLk93UFKn93TFKn94SFJo94SFLm93SFMm93TFMl91VFOj90WFOj9OWF2i9MXF2i9LYF4g9KZF4g9JZF7f9_OcEhNh0h1f9]OeEiNg0j1c9[OiEjNe0j1c9YOlEjNb0n1b9UOnFj0S9QORGo0n8kNXGU1h8dN^G\\1c8]NdGc1\\8UNnGh1h:N1N3N2M2O2M3N1N3N2N1O1N0100O10O0100O010O100O010O10O0100O10O010O0000001N100000000O2O00O10O100000000O100000000O10000000O01000000O100000000O100000000O100000000O1000000O100000000O1000001O0O100000000O1000000O100000000O100000001M2O1N2O1N2O1N3M2O1N2O1Ngkd4"}, "label": "man in brown shirt and gray pants playing wii"}]} {"id": 27149, "image": "COCO_train2014_000000027149.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man playing wii\"."}], "task": "refering", "answer_template": "The \"the man playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 53, 71, 72, 73, 74, 75, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 169, 170, 171, 172, 189, 191, 192, 193], "bbox": [0.020921875, 0.11391666666666667, 0.5030781249999999, 0.498875], "iscrowd": 0, "area": 23990.278599999998, "rle": {"size": [480, 640], "counts": "_g63k>4M2M4L3N3L3N1N2O1N2O1N2I7H8M3O001O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O0O2O1O2N1O2N1O1O2M201O001O001O001O0010O01O00001O001O0O2O001O1O001O001O1N101O1O001O1O001O0O2O1O001O1O001O1O0O2O001O00001N1O1O2N1O1O2O0O101O000O1O2M2O1N3N1N2O2M3M2O2M3M3N1N3M3N1N3M3N1N3M3N1N3N2O001O00001O0000001O0000O001O1O1N2O1O1O1O1N2O1UMZF:h9FXF9i9GXF6j9IXF5i9KXF3i9LXF2j9NWF0j90WFMk92WFLk93UFKn93TFKn94SFJo94SFLm93SFMm93TFMl91VFOj90WFOj9OWF2i9MXF2i9LYF4g9KZF4g9JZF7f9_OcEhNh0h1f9]OeEiNg0j1c9[OiEjNe0j1c9YOlEjNb0n1b9UOnFj0S9QORGo0n8kNXGU1h8dN^G\\1c8]NdGc1\\8UNnGh1h:N1N3N2M2O2M3N1N3N2N1O1N0100O10O0100O010O100O010O10O0100O10O010O0000001N100000000O2O00O10O100000000O100000000O10000000O01000000O100000000O100000000O100000000O1000000O100000000O1000001O0O100000000O1000000O100000000O100000001M2O1N2O1N2O1N3M2O1N2O1Ngkd4"}, "label": "the man playing wii"}]} {"id": 313873, "image": "COCO_train2014_000000313873.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy with a brown hat and gray hooded jacket\"."}], "task": "refering", "answer_template": "The \"a boy with a brown hat and gray hooded jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 22, 23, 24, 25, 26, 44, 45, 46, 47, 48, 66, 67, 68, 69, 88, 89, 90, 91, 92, 110, 111, 112, 113, 114, 115, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 330, 331, 332, 333, 334, 352, 353, 354, 355, 374, 375, 376, 396, 397, 398, 418, 419], "bbox": [0.03145424836601307, 0.03370915032679738, 0.24944444444444444, 0.8808986928104575], "iscrowd": 0, "area": 49468.58350000001, "rle": {"size": [612, 612], "counts": "hk;;da0U1kNU1bM_2gEX:O1O1O1O2N1O1O1O1O2N1O1O101N1O1O1O1O001O001O1O0010O100O100O100O1J6D in this image.", "objects": [{"patches": [1, 2, 3, 22, 23, 24, 25, 26, 44, 45, 46, 47, 48, 66, 67, 68, 69, 88, 89, 90, 91, 92, 110, 111, 112, 113, 114, 115, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 330, 331, 332, 333, 334, 352, 353, 354, 355, 374, 375, 376, 396, 397, 398, 418, 419], "bbox": [0.03145424836601307, 0.03370915032679738, 0.24944444444444444, 0.8808986928104575], "iscrowd": 0, "area": 49468.58350000001, "rle": {"size": [612, 612], "counts": "hk;;da0U1kNU1bM_2gEX:O1O1O1O2N1O1O1O1O2N1O1O101N1O1O1O1O001O001O1O0010O100O100O100O1J6D in this image.", "objects": [{"patches": [202, 203, 204, 224, 225, 226, 227, 246, 247, 248, 249, 250, 269, 270, 271, 291, 292, 293, 294, 313, 314, 315, 316, 335, 336, 337, 338, 357, 358, 359, 360, 380, 381, 382, 383, 402, 403, 404, 405, 406, 424, 425, 426, 427, 428, 446, 447, 448, 449, 469, 470, 471], "bbox": [0.2, 0.4044934640522876, 0.47191176470588236, 0.9685457516339869], "iscrowd": 0, "area": 27032.91385, "rle": {"size": [612, 612], "counts": "^UY21Sc05J5L5J5K6K5J5L5J5K6K5J5L5J6J5L5J5L5J6J5L5J5L5J6K4L5K4M4K5L3L5K5L3L5L3L5K5L3L5K4M4K5L3L5K4M4K5L3L5K5L3L5K4M4K5L3L5K4M4K5L3N3M3N1N3M2N3M3M2N3M2N1ON2N2N2N2N2N2M3N3M2N2N2N2N2N1OO100O100O100O100O10THVF^6i9]IaF_6_9`IfF\\6Y9eIkFW6V9gIoFU6R9jISGQ6n8mIWGo5j8PJZGk5h8SJ]Gi5d8VJ`Gf5a8WJeGe5]8WJjGd5W8YJoGc5R8[JSHa5n7\\JXH`5j:L5K4L4L4L4L4L5K4L4L4L4L4L5K4L4L4L4K6K4L4L4L4L4L5K4L4L4LO100ON3L301N101N101N1A^^O]Oca0c0a^OWO`a0j0d^OQO]a0o0;100O010O12M2N3M2M4M3M3M in this image.", "objects": [{"patches": [202, 203, 204, 224, 225, 226, 227, 246, 247, 248, 249, 250, 269, 270, 271, 291, 292, 293, 294, 313, 314, 315, 316, 335, 336, 337, 338, 357, 358, 359, 360, 380, 381, 382, 383, 402, 403, 404, 405, 406, 424, 425, 426, 427, 428, 446, 447, 448, 449, 469, 470, 471], "bbox": [0.2, 0.4044934640522876, 0.47191176470588236, 0.9685457516339869], "iscrowd": 0, "area": 27032.91385, "rle": {"size": [612, 612], "counts": "^UY21Sc05J5L5J5K6K5J5L5J5K6K5J5L5J6J5L5J5L5J6J5L5J5L5J6K4L5K4M4K5L3L5K5L3L5L3L5K5L3L5K4M4K5L3L5K4M4K5L3L5K5L3L5K4M4K5L3L5K4M4K5L3N3M3N1N3M2N3M3M2N3M2N1ON2N2N2N2N2N2M3N3M2N2N2N2N2N1OO100O100O100O100O10THVF^6i9]IaF_6_9`IfF\\6Y9eIkFW6V9gIoFU6R9jISGQ6n8mIWGo5j8PJZGk5h8SJ]Gi5d8VJ`Gf5a8WJeGe5]8WJjGd5W8YJoGc5R8[JSHa5n7\\JXH`5j:L5K4L4L4L4L4L5K4L4L4L4L4L5K4L4L4L4K6K4L4L4L4L4L5K4L4L4LO100ON3L301N101N101N1A^^O]Oca0c0a^OWO`a0j0d^OQO]a0o0;100O010O12M2N3M2M4M3M3M in this image.", "objects": [{"patches": [31, 32, 51, 52, 53, 54, 73, 74, 75, 76, 77, 95, 96, 97, 98, 99, 100, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 160, 161, 162, 163, 164, 165, 166, 167, 182, 183, 184, 185, 186, 187, 188, 189, 204, 205, 206, 207, 208, 209, 210, 211, 226, 227, 228, 229, 230, 231, 232, 249, 250, 251, 252, 253, 254, 268, 269, 272, 273, 274, 275, 276, 277, 290, 291, 296, 297, 298, 312, 313, 319, 320, 335, 341, 342, 363, 364, 385, 386, 408], "bbox": [0.20225490196078433, 0.07415032679738562, 0.6089869281045751, 0.87640522875817], "iscrowd": 0, "area": 51733.05015, "rle": {"size": [612, 612], "counts": "f^Z23ob05L4L4K4M4K5L4K5L4K5L4L4K4M3O03N2M3M2O2M3M3M3N2M3M3N2M3M3N2M2N3N2M3M3N2M3M3NQQ1NlnN3M3M2N3M3M2N3M3M2N3M3M2N5F9EA?B>B>B>B>A?B>B=L5L4L4L3M4L4L3M4M2aM`DXMc;^2iD]MZ;Z2QEaMQ;Y2WEcMl:X2[EcMg:Z2_EaMd:[2bE`M`:]2fEYMb:c2dEnLi:m2^EdLn:Y3a2L3M4K5L3K6WOjf^4"}, "label": "a boy in a red sweatshirt"}]} {"id": 313873, "image": "COCO_train2014_000000313873.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl wearing the red shirt\"."}], "task": "refering", "answer_template": "The \"the girl wearing the red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 51, 52, 53, 54, 73, 74, 75, 76, 77, 95, 96, 97, 98, 99, 100, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 160, 161, 162, 163, 164, 165, 166, 167, 182, 183, 184, 185, 186, 187, 188, 189, 204, 205, 206, 207, 208, 209, 210, 211, 226, 227, 228, 229, 230, 231, 232, 249, 250, 251, 252, 253, 254, 268, 269, 272, 273, 274, 275, 276, 277, 290, 291, 296, 297, 298, 312, 313, 319, 320, 335, 341, 342, 363, 364, 385, 386, 408], "bbox": [0.20225490196078433, 0.07415032679738562, 0.6089869281045751, 0.87640522875817], "iscrowd": 0, "area": 51733.05015, "rle": {"size": [612, 612], "counts": "f^Z23ob05L4L4K4M4K5L4K5L4K5L4L4K4M3O03N2M3M2O2M3M3M3N2M3M3N2M3M3N2M2N3N2M3M3N2M3M3NQQ1NlnN3M3M2N3M3M2N3M3M2N3M3M2N5F9EA?B>B>B>B>A?B>B=L5L4L4L3M4L4L3M4M2aM`DXMc;^2iD]MZ;Z2QEaMQ;Y2WEcMl:X2[EcMg:Z2_EaMd:[2bE`M`:]2fEYMb:c2dEnLi:m2^EdLn:Y3a2L3M4K5L3K6WOjf^4"}, "label": "the girl wearing the red shirt"}]} {"id": 215569, "image": "COCO_train2014_000000215569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a row of glasses with a womens hand\"."}], "task": "refering", "answer_template": "The \"a row of glasses with a womens hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 306, 307, 308, 309, 310, 311, 312, 313, 314, 330, 331, 332, 333, 334, 335, 336, 337, 355, 356, 357, 358, 359, 360, 380, 381, 382], "bbox": [0.33464062499999997, 0.6433958333333333, 0.7365468749999999, 0.9662499999999999], "iscrowd": 0, "area": 28238.23115, "rle": {"size": [480, 640], "counts": "beT32k>5J6K5K4K6L4M3L4L4L4M3L4M3M3N2N2M3N001O0O101O0O2O001N101O0O2O001N101O0O2O001N10001N101O0O2O0O2O001N101O0O2O001N10001N101O0O2O0O2O001N101O0O2O001N101O0O101O0O2O0O2O1O0O2O001O001O001O1O001O001O001O001O1N101O001O001O001O001O1O001O001O001O001O1O001O0O2O001O001O001O1O001O0000000000000000000000000O10000000000000000000000000000000000O1000000000000000000000000O100000O1O1O100O1O1O1O1O1O1O001O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O101N1O1O1O1O100O1O1O1O1O100O1O1O1O1O2L3VOj0UOk0L4N2N2O1N2N2O1N2N2N2O1N2N3M2O1N2N2N2O1N2N3N3L4LUm^2"}, "label": "a row of glasses with a womens hand"}]} {"id": 215569, "image": "COCO_train2014_000000215569.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a row of four wine glasses\"."}], "task": "refering", "answer_template": "The \"a row of four wine glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [261, 262, 263, 264, 265, 266, 267, 268, 269, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 306, 307, 308, 309, 310, 311, 312, 313, 314, 330, 331, 332, 333, 334, 335, 336, 337, 355, 356, 357, 358, 359, 360, 380, 381, 382], "bbox": [0.33464062499999997, 0.6433958333333333, 0.7365468749999999, 0.9662499999999999], "iscrowd": 0, "area": 28238.23115, "rle": {"size": [480, 640], "counts": "beT32k>5J6K5K4K6L4M3L4L4L4M3L4M3M3N2N2M3N001O0O101O0O2O001N101O0O2O001N101O0O2O001N10001N101O0O2O0O2O001N101O0O2O001N10001N101O0O2O0O2O001N101O0O2O001N101O0O101O0O2O0O2O1O0O2O001O001O001O1O001O001O001O001O1N101O001O001O001O001O1O001O001O001O001O1O001O0O2O001O001O001O1O001O0000000000000000000000000O10000000000000000000000000000000000O1000000000000000000000000O100000O1O1O100O1O1O1O1O1O1O001O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O101N1O1O1O1O100O1O1O1O1O100O1O1O1O1O2L3VOj0UOk0L4N2N2O1N2N2O1N2N2N2O1N2N3M2O1N2N2N2O1N2N3N3L4LUm^2"}, "label": "a row of four wine glasses"}]} {"id": 43543, "image": "COCO_train2014_000000043543.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person snowboarding\"."}], "task": "refering", "answer_template": "The \"a person snowboarding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 23, 24, 25, 26, 27, 28, 41, 42, 43, 44, 45, 46, 58, 59, 60, 61, 62, 63, 64, 76, 77, 79, 80, 81, 82, 94, 95, 98, 99, 100, 101, 117, 118, 119, 136, 137, 154, 155, 156, 172, 173, 174, 190, 191, 192, 208], "bbox": [0.2184, 0.016373333333333333, 0.71614, 0.8606133333333332], "iscrowd": 0, "area": 24208.18919999999, "rle": {"size": [375, 500], "counts": "fQX11d;5K4L3M3M2N3M3M2N3M3N1O2N1O2O1N1O1O2O0O1O1O101N1O1O100N3N1O1N2O1O1N10O1O0C>E;0O10001N10000O10000O101O0O10010WNIQH7P8IoG8Q8HnG8R8ImG8S8JjG6V8MgG4Y8NeG1\\82`GO_84^GLc8d1O0100O0100O0100O010O10O02O1N2O001N2O1O10O1000000O10000000021N2OO1N00O2N1O1O2N1O1O2O0O1O2N1O1O2N100O2O0O101N10001N100O100O010O01O10O010O1O1O2N1O1O1O2M2O1O2N1O1O1O2N1O1O1O2N1O100O1O2O0O1O100O2N1O100eL\\KAe4=\\KCf4;ZKFf48\\KGe47\\KIf45ZKLg41[KNh4NYK2i4KXK5j4IVK8k4EWK:k4CVK=l4@UKa0l4]OTKc0R1gL\\2c2dLe0n0lL^2\\2eLh0l0PM^2U2gLl0h0SMa2o1gLn0f0XMb2g1jLP1>aMh2\\1kLQ3U3kLlLW3T3gLlL[3U3aLmLa3R3\\LnLg3S3TLoLo3P3mKRMU4o2gKQM]4n2_KTMc4m2YKTMj4j44M3L4M3M2M4M3L4M3M3L3N3L3N3M3L3N3L4M2N3L4M2N3N2M2O2M3M3N1N3N2M2O2M3L3N3M3L3N3M3M2M3N1O2N2M3N1O2M3N3M3M3L5L3M3G:E:GYgc1"}, "label": "a person snowboarding"}]} {"id": 43543, "image": "COCO_train2014_000000043543.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"snowboarder\"."}], "task": "refering", "answer_template": "The \"snowboarder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 23, 24, 25, 26, 27, 28, 41, 42, 43, 44, 45, 46, 58, 59, 60, 61, 62, 63, 64, 76, 77, 79, 80, 81, 82, 94, 95, 98, 99, 100, 101, 117, 118, 119, 136, 137, 154, 155, 156, 172, 173, 174, 190, 191, 192, 208], "bbox": [0.2184, 0.016373333333333333, 0.71614, 0.8606133333333332], "iscrowd": 0, "area": 24208.18919999999, "rle": {"size": [375, 500], "counts": "fQX11d;5K4L3M3M2N3M3M2N3M3N1O2N1O2O1N1O1O2O0O1O1O101N1O1O100N3N1O1N2O1O1N10O1O0C>E;0O10001N10000O10000O101O0O10010WNIQH7P8IoG8Q8HnG8R8ImG8S8JjG6V8MgG4Y8NeG1\\82`GO_84^GLc8d1O0100O0100O0100O010O10O02O1N2O001N2O1O10O1000000O10000000021N2OO1N00O2N1O1O2N1O1O2O0O1O2N1O1O2N100O2O0O101N10001N100O100O010O01O10O010O1O1O2N1O1O1O2M2O1O2N1O1O1O2N1O1O1O2N1O100O1O2O0O1O100O2N1O100eL\\KAe4=\\KCf4;ZKFf48\\KGe47\\KIf45ZKLg41[KNh4NYK2i4KXK5j4IVK8k4EWK:k4CVK=l4@UKa0l4]OTKc0R1gL\\2c2dLe0n0lL^2\\2eLh0l0PM^2U2gLl0h0SMa2o1gLn0f0XMb2g1jLP1>aMh2\\1kLQ3U3kLlLW3T3gLlL[3U3aLmLa3R3\\LnLg3S3TLoLo3P3mKRMU4o2gKQM]4n2_KTMc4m2YKTMj4j44M3L4M3M2M4M3L4M3M3L3N3L3N3M3L3N3L4M2N3L4M2N3N2M2O2M3M3N1N3N2M2O2M3L3N3M3L3N3M3M2M3N1O2N2M3N1O2M3N3M3M3L5L3M3G:E:GYgc1"}, "label": "snowboarder"}]} {"id": 412190, "image": "COCO_train2014_000000412190.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty sink\"."}], "task": "refering", "answer_template": "The \"an empty sink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [328, 329, 330, 331, 347, 348, 349, 350, 351, 352, 353, 354, 355, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.032015625, 0.8539375, 0.482015625, 0.9797708333333333], "iscrowd": 0, "area": 9620.629100000004, "rle": {"size": [480, 640], "counts": "eh:1o>0O100O100O100O100O100O100O100O100O10000O1000000O10O10O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O1000O10O10000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O0100000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O10O10O1000000O10000000000O10000000000O10000000000O10000000000O10000000000O100000O1000O100000000O10000000000O10000000000O1000000000000001O000000001O000000001O0O100000001O000000001O000000001O000000001O000000001O2N1O1O1O1O1O1O1O1O1O1O1O2N1N2O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O2N1O1ObUk4"}, "label": "an empty sink"}]} {"id": 412190, "image": "COCO_train2014_000000412190.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white sink\"."}], "task": "refering", "answer_template": "The \"white sink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [328, 329, 330, 331, 347, 348, 349, 350, 351, 352, 353, 354, 355, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.032015625, 0.8539375, 0.482015625, 0.9797708333333333], "iscrowd": 0, "area": 9620.629100000004, "rle": {"size": [480, 640], "counts": "eh:1o>0O100O100O100O100O100O100O100O100O10000O1000000O10O10O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O1000O10O10000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O0100000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O10O10O1000000O10000000000O10000000000O10000000000O10000000000O10000000000O100000O1000O100000000O10000000000O10000000000O1000000000000001O000000001O000000001O0O100000001O000000001O000000001O000000001O000000001O2N1O1O1O1O1O1O1O1O1O1O1O2N1N2O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O2N1O1ObUk4"}, "label": "white sink"}]} {"id": 109095, "image": "COCO_train2014_000000109095.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black train behind a man\"."}], "task": "refering", "answer_template": "The \"a black train behind a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 180, 183], "bbox": [0.645453125, 0.20987447698744768, 0.9995156249999999, 0.4271548117154812], "iscrowd": 0, "area": 17923.032600000002, "rle": {"size": [478, 640], "counts": "nlP63g>6K5N2N2N3M2N2N2N2N2N3M2N2M3N2N3M2N2N2N2N2N3M2N2N1O0O10000000000000000000000O10001O0000000000000000000O1000000000000000000000001N1000000000000000000000000O1000002N2N2N2N2N2N1O2N1O0O1000001O000000001O0000000O101O000000001O000000000O2O000000001O00000000001O0O1000001O00000000001O00000O0100O100O100O10000O100O100O10000O10O0100O100O10000O100O100O100O10000O100O100O100000000000000000000000000001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1N101O1OG9QOo0ROn0QOkM"}, "label": "a black train behind a man"}]} {"id": 109095, "image": "COCO_train2014_000000109095.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cargo car on a train is sitting on the snowy tracks\"."}], "task": "refering", "answer_template": "The \"a cargo car on a train is sitting on the snowy tracks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 180, 183], "bbox": [0.645453125, 0.20987447698744768, 0.9995156249999999, 0.4271548117154812], "iscrowd": 0, "area": 17923.032600000002, "rle": {"size": [478, 640], "counts": "nlP63g>6K5N2N2N3M2N2N2N2N2N3M2N2M3N2N3M2N2N2N2N2N3M2N2N1O0O10000000000000000000000O10001O0000000000000000000O1000000000000000000000001N1000000000000000000000000O1000002N2N2N2N2N2N1O2N1O0O1000001O000000001O0000000O101O000000001O000000000O2O000000001O00000000001O0O1000001O00000000001O00000O0100O100O100O10000O100O100O10000O10O0100O100O10000O100O100O100O10000O100O100O100000000000000000000000000001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1N101O1OG9QOo0ROn0QOkM"}, "label": "a cargo car on a train is sitting on the snowy tracks"}]} {"id": 109095, "image": "COCO_train2014_000000109095.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man pulling a lever near a track\"."}], "task": "refering", "answer_template": "The \"man pulling a lever near a track\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 174, 175, 176, 197, 198, 199, 221, 222, 223, 243, 244, 245, 246, 266, 267, 268, 269, 289, 290, 291, 292, 312, 313, 314, 315], "bbox": [0.47465624999999995, 0.2342050209205021, 0.719625, 0.804581589958159], "iscrowd": 0, "area": 16715.13355, "rle": {"size": [478, 640], "counts": "VR^42k>2L4M3L3N3O1O00101N2N2O0O10O10O01O1O00001O01O0001O00O101O00000000000O1000000000000O1000000O1000000O1000000O10000O0100000O1000000O10O1000O1jGXOV2h0dM_OZ2b0eM@Z2`0eMCY2=fMDY2h6]OaIn0J>c2lNWMS6^M[IT2>b0X6[MYIS2:g0]68^IMb64YI1g6R4100O010O1N2O001O1O1O2M2N2M4L3M4L5K5L3L5K;F6I5D<2M2O2M2N2O2M2N2N3N101N10O1N2M4M3L4M1001O1c0]O2N2Kc0]O7I3VM`Eo0c:lNeEo0]:lNjEP1Y:kNlEQ1Y:gNnEV1W:^NSF_1j;K4M3M3L4M2J7J5I8F9G:GUWc2"}, "label": "man pulling a lever near a track"}]} {"id": 109095, "image": "COCO_train2014_000000109095.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person dressed in all black in the foreground\"."}], "task": "refering", "answer_template": "The \"person dressed in all black in the foreground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 174, 175, 176, 197, 198, 199, 221, 222, 223, 243, 244, 245, 246, 266, 267, 268, 269, 289, 290, 291, 292, 312, 313, 314, 315], "bbox": [0.47465624999999995, 0.2342050209205021, 0.719625, 0.804581589958159], "iscrowd": 0, "area": 16715.13355, "rle": {"size": [478, 640], "counts": "VR^42k>2L4M3L3N3O1O00101N2N2O0O10O10O01O1O00001O01O0001O00O101O00000000000O1000000000000O1000000O1000000O1000000O10000O0100000O1000000O10O1000O1jGXOV2h0dM_OZ2b0eM@Z2`0eMCY2=fMDY2h6]OaIn0J>c2lNWMS6^M[IT2>b0X6[MYIS2:g0]68^IMb64YI1g6R4100O010O1N2O001O1O1O2M2N2M4L3M4L5K5L3L5K;F6I5D<2M2O2M2N2O2M2N2N3N101N10O1N2M4M3L4M1001O1c0]O2N2Kc0]O7I3VM`Eo0c:lNeEo0]:lNjEP1Y:kNlEQ1Y:gNnEV1W:^NSF_1j;K4M3M3L4M2J7J5I8F9G:GUWc2"}, "label": "person dressed in all black in the foreground"}]} {"id": 412220, "image": "COCO_train2014_000000412220.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front bus in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the front bus in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 305, 306, 307, 308, 309, 310, 328, 329, 330], "bbox": [0.205625, 0.08089583333333333, 0.9471875000000001, 0.8606666666666667], "iscrowd": 0, "area": 129495.92345, "rle": {"size": [480, 640], "counts": "VTn1Y196`;h1K5K6I6K5K5J6K5K5J6K5K5M3M4M2M3M3N2M3M3N2M3N2M3M4M2M3M3N2M3M3N2M3N3L5K5L4K8H7H9H7H9H8H7H7J3L4M3M4K4M3O1O2N1O100O2N1O1O2N1O1O1O2N1O100O2N1O1O1O2N1O1O2N1O1N2O2N1O1O001O000000001O000O1000001O000001O100O2N1O1O1O100O2N1O1O100O2N1O1O100O2N1O1O1O101N100O100O2O0O10000O2O0O100O101O000000001N100000001O0000000O101O000000001O000O10001O000000001O0O1000001O000000001N100000001O0000000O2O00000000001O000O10001O000000001O0O1000001O000000001N100000001O0000000O2O000000001O00000O101O00000000001O0O1000001O000000001N100000001O0000000O2O000000001O00000O101O000000001O000O10001O00000000001N100000001O0000000O2O0000001O00001O0O10001O0000001O000O2O0000001O0000001N10001O0000001O000O101O0000001O00001N1000001O0000001O0O101O0000001O00000O20O1O100O001O100O1O1O10O01O1O10O00001O000O101O0000001O0O101O0000001N1000001O00001N10000O2N100O1O2N100O2N100O1O2N100O1O2O0O1O1O2O0O101N100O101N100O101N10001N100O101N100O101N2O1N2N2I7J6J6J6J6I7J6J6RNmFhMY9W2i1O1N2O1O1O1N2O1O1N2O1O;E>A`0A?A?@]g?"}, "label": "the front bus in the right hand picture"}]} {"id": 412220, "image": "COCO_train2014_000000412220.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"fully visible double decker bus\"."}], "task": "refering", "answer_template": "The \"fully visible double decker bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 305, 306, 307, 308, 309, 310, 328, 329, 330], "bbox": [0.205625, 0.08089583333333333, 0.9471875000000001, 0.8606666666666667], "iscrowd": 0, "area": 129495.92345, "rle": {"size": [480, 640], "counts": "VTn1Y196`;h1K5K6I6K5K5J6K5K5J6K5K5M3M4M2M3M3N2M3M3N2M3N2M3M4M2M3M3N2M3M3N2M3N3L5K5L4K8H7H9H7H9H8H7H7J3L4M3M4K4M3O1O2N1O100O2N1O1O2N1O1O1O2N1O100O2N1O1O1O2N1O1O2N1O1N2O2N1O1O001O000000001O000O1000001O000001O100O2N1O1O1O100O2N1O1O100O2N1O1O100O2N1O1O1O101N100O100O2O0O10000O2O0O100O101O000000001N100000001O0000000O101O000000001O000O10001O000000001O0O1000001O000000001N100000001O0000000O2O00000000001O000O10001O000000001O0O1000001O000000001N100000001O0000000O2O000000001O00000O101O00000000001O0O1000001O000000001N100000001O0000000O2O000000001O00000O101O000000001O000O10001O00000000001N100000001O0000000O2O0000001O00001O0O10001O0000001O000O2O0000001O0000001N10001O0000001O000O101O0000001O00001N1000001O0000001O0O101O0000001O00000O20O1O100O001O100O1O1O10O01O1O10O00001O000O101O0000001O0O101O0000001N1000001O00001N10000O2N100O1O2N100O2N100O1O2N100O1O2O0O1O1O2O0O101N100O101N100O101N10001N100O101N100O101N2O1N2N2I7J6J6J6J6I7J6J6RNmFhMY9W2i1O1N2O1O1O1N2O1O1N2O1O;E>A`0A?A?@]g?"}, "label": "fully visible double decker bus"}]} {"id": 412220, "image": "COCO_train2014_000000412220.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red double - decker with a wrigley ' s ad\"."}], "task": "refering", "answer_template": "The \"a red double - decker with a wrigley ' s ad\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 305, 306, 307, 308, 309, 310, 328, 329, 330], "bbox": [0.205625, 0.08089583333333333, 0.9471875000000001, 0.8606666666666667], "iscrowd": 0, "area": 129495.92345, "rle": {"size": [480, 640], "counts": "VTn1Y196`;h1K5K6I6K5K5J6K5K5J6K5K5M3M4M2M3M3N2M3M3N2M3N2M3M4M2M3M3N2M3M3N2M3N3L5K5L4K8H7H9H7H9H8H7H7J3L4M3M4K4M3O1O2N1O100O2N1O1O2N1O1O1O2N1O100O2N1O1O1O2N1O1O2N1O1N2O2N1O1O001O000000001O000O1000001O000001O100O2N1O1O1O100O2N1O1O100O2N1O1O100O2N1O1O1O101N100O100O2O0O10000O2O0O100O101O000000001N100000001O0000000O101O000000001O000O10001O000000001O0O1000001O000000001N100000001O0000000O2O00000000001O000O10001O000000001O0O1000001O000000001N100000001O0000000O2O000000001O00000O101O00000000001O0O1000001O000000001N100000001O0000000O2O000000001O00000O101O000000001O000O10001O00000000001N100000001O0000000O2O0000001O00001O0O10001O0000001O000O2O0000001O0000001N10001O0000001O000O101O0000001O00001N1000001O0000001O0O101O0000001O00000O20O1O100O001O100O1O1O10O01O1O10O00001O000O101O0000001O0O101O0000001N1000001O00001N10000O2N100O1O2N100O2N100O1O2N100O1O2O0O1O1O2O0O101N100O101N100O101N10001N100O101N100O101N2O1N2N2I7J6J6J6J6I7J6J6RNmFhMY9W2i1O1N2O1O1O1N2O1O1N2O1O;E>A`0A?A?@]g?"}, "label": "a red double - decker with a wrigley ' s ad"}]} {"id": 436797, "image": "COCO_train2014_000000436797.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby sheep walking amongst the grass\"."}], "task": "refering", "answer_template": "The \"a baby sheep walking amongst the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [272, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 319, 336, 337, 338, 339, 340, 341, 342, 343, 362, 363, 364, 365], "bbox": [0.62415625, 0.6973012552301255, 0.9198124999999999, 0.9217991631799163], "iscrowd": 0, "area": 10406.977750000007, "rle": {"size": [478, 640], "counts": "]dj51k>3L4L4M3L4L4M3L4M3M2N2N2N2N200O1O10O010O01000O10000O01O10O0100O0010O101O1N3N1O1N2O000010O01O00010O00001O01O00000000001O0000000O2O00001O0O2O0O101O1N2O1O2M2O1ZC]Nn;e1oC\\NRC3N2M01O100O00102M1OO1O2N1O2N1O2C^BZOd==eBA\\=8nlg0"}, "label": "a baby sheep walking amongst the grass"}]} {"id": 436797, "image": "COCO_train2014_000000436797.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white little lamb\"."}], "task": "refering", "answer_template": "The \"the white little lamb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [272, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 319, 336, 337, 338, 339, 340, 341, 342, 343, 362, 363, 364, 365], "bbox": [0.62415625, 0.6973012552301255, 0.9198124999999999, 0.9217991631799163], "iscrowd": 0, "area": 10406.977750000007, "rle": {"size": [478, 640], "counts": "]dj51k>3L4L4M3L4L4M3L4M3M2N2N2N2N200O1O10O010O01000O10000O01O10O0100O0010O101O1N3N1O1N2O000010O01O00010O00001O01O00000000001O0000000O2O00001O0O2O0O101O1N2O1O2M2O1ZC]Nn;e1oC\\NRC3N2M01O100O00102M1OO1O2N1O2N1O2C^BZOd==eBA\\=8nlg0"}, "label": "the white little lamb"}]} {"id": 436797, "image": "COCO_train2014_000000436797.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a big female sheep and her baby lamb\"."}], "task": "refering", "answer_template": "The \"a big female sheep and her baby lamb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 143, 144, 145, 146, 147, 152, 153, 154, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 308, 309, 331, 332, 354, 355], "bbox": [0.23174999999999998, 0.34487447698744766, 0.8262968749999999, 0.9236192468619248], "iscrowd": 0, "area": 57050.61105000004, "rle": {"size": [478, 640], "counts": "]gU25h>1O2O1N101O1N101O1O001O00001N100000001O00000001O0OSB4j in this image.", "objects": [{"patches": [122, 143, 144, 145, 146, 147, 152, 153, 154, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 308, 309, 331, 332, 354, 355], "bbox": [0.23174999999999998, 0.34487447698744766, 0.8262968749999999, 0.9236192468619248], "iscrowd": 0, "area": 57050.61105000004, "rle": {"size": [478, 640], "counts": "]gU25h>1O2O1N101O1N101O1O001O00001N100000001O00000001O0OSB4j in this image.", "objects": [{"patches": [22, 44, 45, 67, 68, 89, 90, 91, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 177, 178, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 315, 316, 317, 318, 319, 338, 339, 340, 341, 342], "bbox": [0.6971093749999999, 0.04300469483568075, 1.0, 0.9870892018779343], "iscrowd": 0, "area": 36475.55630000002, "rle": {"size": [426, 640], "counts": "YUj52X=1O2SIM?5\\OMe04WOOi03RO0n01nN2R10iN3W1OdN4\\10]N3c1L\\N6d1IZN:f1EXN=i1BUNa0k1^OSNe0m1ZOQNi0P2UOnMn0R2QOlMR1T2mNjMV1X2gNlIKo2a1X3aNeI4n2^1S4hNgKZ1V4mNdKV1Z4oN`KT1_4oN\\KT1c4POWKS1h4POTKR1k4ROoJQ1P5SOlJn0S5UOkJk0U5XOhJh0W5[OgJd0Y5@cJa0\\5BbJ>]5F]J=b5GWJ=h5FVJ:j5IRJ8n5JPJ6P6MlI4S60jIOW63gIMY66cIK]67aII_6:]IGc6;[IEe6>XIBh6a0TI@m6a0QI_Oo6d0mH\\OU7e0iH[OW7\\3001O001lKeHW2[7fMhHZ2Y7bMjH^2V7_MlHa2V7\\MlHd2T7YMoHg2R7WMoHi2Q7VMPIj2R7RMPIn2Q7PMoHQ3S7lLnHT3S7jLmHW3U7fLkH[3V7cLjH]3Y7`LfHb3[7\\LeHe3]7WLdHj3]7TLcHm3^7QLaHQ4a7lK_HU4b7iK^HX4d7eK[H]4f7aKZH_4X81N2O1O1N2O1K5L`LhGV2S8mMPHQ2m7QNUHo1f7TN]Hk1_7WNdHh1]7UNeHk1^7PNeHA@]1n7mNeHS2AfLe7R1mHW2[OjLk7j0lHd2W7WMlHh2W7SMlH5ROl1Y5[M_M?YNQ3W4eL`M4XNZ3V4gLbMJVNc3V4iLcM_OVNR1SOe1R5oM^N?`Lb1P5UNXN=hL_1m4ZNTNT3k1PMoMS3o1SMiM6]MW1h4iNdM3dMT1f4nN_M2kMP1d4TOZMk2c2ZMWMLXNk0b4^OnLK`Nd0c4GfLHgN?c4N`LFmN9d47WLDUO2e4`0oK@\\ONf4g0hK^OBHh4o0^K]OJBh4V1XK[O0\\Oi4T4XKiKi4Y4VKeKk4]4UKaKk4a4UK\\Km4e4SKXKn4i4SKTKn4n4QKPKP5R5PKlJP5V5PKgJQ5Z5PKcJR5^5nJ_JS5c5mJ[JS5g5lJWJU5k5kJRJV5o5kJnIW5S6iJkIW5W6iJfIX5\\6gJbIZ5_6gJ^IZ5d6fJZIZ5`6eJUI1:Z5e6fJmH0=Y5l6hJTIV5`7O1O2N100O2N1O1O2N100O2N1O1O2N100O2N1O000001O0O100000001O000000001O00002N2N2N3M2N3M2N2N3M2N3M2M3N3M2N3MYO"}, "label": "a brown chair with a white padding , on which a grey cat is standing on"}]} {"id": 322121, "image": "COCO_train2014_000000322121.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair that the cat is standing on\"."}], "task": "refering", "answer_template": "The \"the chair that the cat is standing on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 44, 45, 67, 68, 89, 90, 91, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 177, 178, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 292, 293, 294, 295, 296, 315, 316, 317, 318, 319, 338, 339, 340, 341, 342], "bbox": [0.6971093749999999, 0.04300469483568075, 1.0, 0.9870892018779343], "iscrowd": 0, "area": 36475.55630000002, "rle": {"size": [426, 640], "counts": "YUj52X=1O2SIM?5\\OMe04WOOi03RO0n01nN2R10iN3W1OdN4\\10]N3c1L\\N6d1IZN:f1EXN=i1BUNa0k1^OSNe0m1ZOQNi0P2UOnMn0R2QOlMR1T2mNjMV1X2gNlIKo2a1X3aNeI4n2^1S4hNgKZ1V4mNdKV1Z4oN`KT1_4oN\\KT1c4POWKS1h4POTKR1k4ROoJQ1P5SOlJn0S5UOkJk0U5XOhJh0W5[OgJd0Y5@cJa0\\5BbJ>]5F]J=b5GWJ=h5FVJ:j5IRJ8n5JPJ6P6MlI4S60jIOW63gIMY66cIK]67aII_6:]IGc6;[IEe6>XIBh6a0TI@m6a0QI_Oo6d0mH\\OU7e0iH[OW7\\3001O001lKeHW2[7fMhHZ2Y7bMjH^2V7_MlHa2V7\\MlHd2T7YMoHg2R7WMoHi2Q7VMPIj2R7RMPIn2Q7PMoHQ3S7lLnHT3S7jLmHW3U7fLkH[3V7cLjH]3Y7`LfHb3[7\\LeHe3]7WLdHj3]7TLcHm3^7QLaHQ4a7lK_HU4b7iK^HX4d7eK[H]4f7aKZH_4X81N2O1O1N2O1K5L`LhGV2S8mMPHQ2m7QNUHo1f7TN]Hk1_7WNdHh1]7UNeHk1^7PNeHA@]1n7mNeHS2AfLe7R1mHW2[OjLk7j0lHd2W7WMlHh2W7SMlH5ROl1Y5[M_M?YNQ3W4eL`M4XNZ3V4gLbMJVNc3V4iLcM_OVNR1SOe1R5oM^N?`Lb1P5UNXN=hL_1m4ZNTNT3k1PMoMS3o1SMiM6]MW1h4iNdM3dMT1f4nN_M2kMP1d4TOZMk2c2ZMWMLXNk0b4^OnLK`Nd0c4GfLHgN?c4N`LFmN9d47WLDUO2e4`0oK@\\ONf4g0hK^OBHh4o0^K]OJBh4V1XK[O0\\Oi4T4XKiKi4Y4VKeKk4]4UKaKk4a4UK\\Km4e4SKXKn4i4SKTKn4n4QKPKP5R5PKlJP5V5PKgJQ5Z5PKcJR5^5nJ_JS5c5mJ[JS5g5lJWJU5k5kJRJV5o5kJnIW5S6iJkIW5W6iJfIX5\\6gJbIZ5_6gJ^IZ5d6fJZIZ5`6eJUI1:Z5e6fJmH0=Y5l6hJTIV5`7O1O2N100O2N1O1O2N100O2N1O1O2N100O2N1O000001O0O100000001O000000001O00002N2N2N3M2N3M2N2N3M2N3M2M3N3M2N3MYO"}, "label": "the chair that the cat is standing on"}]} {"id": 559700, "image": "COCO_train2014_000000559700.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large standing elephant\"."}], "task": "refering", "answer_template": "The \"a large standing elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 36, 37, 38, 39, 40, 41, 42, 43, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 139, 140, 141, 143, 144, 145, 146, 147, 148, 149, 153, 154, 155, 163, 164, 168, 170, 179], "bbox": [0.21413145539906103, 0.07528125, 1.0, 0.52171875], "iscrowd": 0, "area": 56969.03310000002, "rle": {"size": [640, 426], "counts": "lUi11mc0101N2O1N101O100O00100O1O010O1O10O0100O0001O000001N100O10000O101N100O10000O2O0O100O1N3\\Oc0]Oc0B>D in this image.", "objects": [{"patches": [24, 25, 26, 27, 36, 37, 38, 39, 40, 41, 42, 43, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 139, 140, 141, 143, 144, 145, 146, 147, 148, 149, 153, 154, 155, 163, 164, 168, 170, 179], "bbox": [0.21413145539906103, 0.07528125, 1.0, 0.52171875], "iscrowd": 0, "area": 56969.03310000002, "rle": {"size": [640, 426], "counts": "lUi11mc0101N2O1N101O100O00100O1O010O1O10O0100O0001O000001N100O10000O101N100O10000O2O0O100O1N3\\Oc0]Oc0B>D in this image.", "objects": [{"patches": [222, 223, 231, 232, 233, 234, 235, 236, 237, 238, 239, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 302, 303, 304, 306, 307, 308, 309, 310, 311, 312, 313, 314], "bbox": [0.013497652582159625, 0.640453125, 0.9993192488262911, 0.912359375], "iscrowd": 0, "area": 48133.983700000004, "rle": {"size": [640, 426], "counts": "aX42lc03M2N3M2N3M2O2N1O2N1O2N1O2N1O2N1O2O1N1O2N1O2N1O2N1O2N1O2N101N100O2O0O101O0O101N100O2O000O2O0O101N10001N100O2O0O101O0O101N100O2O000O2O0O101N10001N100O2O000O2O0O101N10001N100O2O000O2O0O101O0O100O01000O10O010000O010O1000O0100O10O10O100O01000O100O0100000O01000000O0100000O0100000O10O1000O1000O10O100000O10O100000O10001O000O101O00000O2O0000001N100000001N1000001N1000001O0O1000001O00000O2O0000001O0000001N100000001O0000001N1000001O0000001N@S@oLm?Q3U@mLk?T3V@jLj?V3X@hLh?X3Z@gLe?Y3\\@fLd?Z3^@dLb?\\3`@bL`?^3b@aL\\?`3f@^LZ?c3a00000000000001N1000000000001O000O10000000001O0O100000000O2O0000000O10001O000O1000000O2O0000000O1000001O0O100000000O2O00000O1000001O0O100000000O101O00000O10001O0O100000001N100000000O2O0000000O101O000O10001O000O1000001O0O100000001N1000000O101O00000O10001O000O1000001O0O1000001N100000000O2O00000000010O0001O000010O000001O00010O00001O00010O00001O01O0001O000010O0001O000010O000004LU1kNU1lNS1lNU1kNS^O"}, "label": "elephant lying on the ground"}]} {"id": 559700, "image": "COCO_train2014_000000559700.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant laying on its side with two other elephants above it\"."}], "task": "refering", "answer_template": "The \"an elephant laying on its side with two other elephants above it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 231, 232, 233, 234, 235, 236, 237, 238, 239, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 302, 303, 304, 306, 307, 308, 309, 310, 311, 312, 313, 314], "bbox": [0.013497652582159625, 0.640453125, 0.9993192488262911, 0.912359375], "iscrowd": 0, "area": 48133.983700000004, "rle": {"size": [640, 426], "counts": "aX42lc03M2N3M2N3M2O2N1O2N1O2N1O2N1O2N1O2O1N1O2N1O2N1O2N1O2N1O2N101N100O2O0O101O0O101N100O2O000O2O0O101N10001N100O2O0O101O0O101N100O2O000O2O0O101N10001N100O2O000O2O0O101N10001N100O2O000O2O0O101O0O100O01000O10O010000O010O1000O0100O10O10O100O01000O100O0100000O01000000O0100000O0100000O10O1000O1000O10O100000O10O100000O10001O000O101O00000O2O0000001N100000001N1000001N1000001O0O1000001O00000O2O0000001O0000001N100000001O0000001N1000001O0000001N@S@oLm?Q3U@mLk?T3V@jLj?V3X@hLh?X3Z@gLe?Y3\\@fLd?Z3^@dLb?\\3`@bL`?^3b@aL\\?`3f@^LZ?c3a00000000000001N1000000000001O000O10000000001O0O100000000O2O0000000O10001O000O1000000O2O0000000O1000001O0O100000000O2O00000O1000001O0O100000000O101O00000O10001O0O100000001N100000000O2O0000000O101O000O10001O000O1000001O0O100000001N1000000O101O00000O10001O000O1000001O0O1000001N100000000O2O00000000010O0001O000010O000001O00010O00001O00010O00001O01O0001O000010O0001O000010O000004LU1kNU1lNS1lNU1kNS^O"}, "label": "an elephant laying on its side with two other elephants above it"}]} {"id": 559700, "image": "COCO_train2014_000000559700.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby elephant between a sleeping elephant and a standing large elephant\"."}], "task": "refering", "answer_template": "The \"a baby elephant between a sleeping elephant and a standing large elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 140, 141, 142, 143, 144, 145, 146, 147, 148, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 199, 200, 203, 204, 205, 206, 207, 208, 209, 215, 218, 219, 220, 221, 222, 223, 224, 233, 234, 235, 236], "bbox": [0.31734741784037557, 0.384265625, 1.0, 0.6921406250000001], "iscrowd": 0, "area": 38075.68075, "rle": {"size": [640, 426], "counts": "ofd27fc0e0ZO9G8I6I7I8H7J7I4M2M4M2N2M4M2N2M4M1OdNd_OXOY`0k0Q@jNn?X1\\@^Nb?c1i@RNU?Q2[12N1N3N2N1N3N20000O100O1000O010000O10000O100O01000O10000O100O10O10O10000O10000O10O010000O10000O100O10O10O10000O10000O10O010000O10000O100O10O1g2XM3N2N2O0O00O101O00001O0O10001O000O2O0000001N10001O000O101O00001N1000001O000O2O0000001N10001O000O101O00001N10001O001N10001O001N10001O001N10001O001N10001O0O2O0O101N10O0100O100O010O1O100O0100000000O010000000O10000000O0100000O10O1000O1000O1000O1000O10O100000O0100000001N101O1O1N110O1O001O1O10O01O1O00100O001O1O1O010O1O001O10O01O1O1O010O1O001O1O010O1O1O00100K4M4K5K5L3L5L4K5K8I7H8I7H8H8I7H7J7H8H8I\\E"}, "label": "a baby elephant between a sleeping elephant and a standing large elephant"}]} {"id": 559700, "image": "COCO_train2014_000000559700.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a calf\"."}], "task": "refering", "answer_template": "The \"a calf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 140, 141, 142, 143, 144, 145, 146, 147, 148, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 199, 200, 203, 204, 205, 206, 207, 208, 209, 215, 218, 219, 220, 221, 222, 223, 224, 233, 234, 235, 236], "bbox": [0.31734741784037557, 0.384265625, 1.0, 0.6921406250000001], "iscrowd": 0, "area": 38075.68075, "rle": {"size": [640, 426], "counts": "ofd27fc0e0ZO9G8I6I7I8H7J7I4M2M4M2N2M4M2N2M4M1OdNd_OXOY`0k0Q@jNn?X1\\@^Nb?c1i@RNU?Q2[12N1N3N2N1N3N20000O100O1000O010000O10000O100O01000O10000O100O10O10O10000O10000O10O010000O10000O100O10O10O10000O10000O10O010000O10000O100O10O1g2XM3N2N2O0O00O101O00001O0O10001O000O2O0000001N10001O000O101O00001N1000001O000O2O0000001N10001O000O101O00001N10001O001N10001O001N10001O001N10001O001N10001O0O2O0O101N10O0100O100O010O1O100O0100000000O010000000O10000000O0100000O10O1000O1000O1000O1000O10O100000O0100000001N101O1O1N110O1O001O1O10O01O1O00100O001O1O1O010O1O001O10O01O1O1O010O1O001O1O010O1O1O00100K4M4K5K5L3L5L4K5K8I7H8I7H8H8I7H7J7H8H8I\\E"}, "label": "a calf"}]} {"id": 223831, "image": "COCO_train2014_000000223831.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small bench covered with a green cloth\"."}], "task": "refering", "answer_template": "The \"a small bench covered with a green cloth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 162, 163, 164, 165, 166, 180, 181, 182, 183, 184, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220, 234, 235, 236, 237, 238], "bbox": [0.00174, 0.6187305699481865, 0.26830000000000004, 0.9841968911917098], "iscrowd": 0, "area": 14161.82965, "rle": {"size": [386, 500], "counts": "Yg03k;5L3L5K4L5L3L5K4M4K4L5K4M4K4L5K4M4K4L5L3L5K4L5L3L5K4L5L3L5K4M4K4O2O010O01O001O00001O001O00001O001O00001O001O00001O001O001O000O2O001O00001O00001O000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001N10002N:F9G9G9G9G:F9G9G9G9G9G:FXcY4"}, "label": "a small bench covered with a green cloth"}]} {"id": 223831, "image": "COCO_train2014_000000223831.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green , unoccupied couch\"."}], "task": "refering", "answer_template": "The \"a green , unoccupied couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 162, 163, 164, 165, 166, 180, 181, 182, 183, 184, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220, 234, 235, 236, 237, 238], "bbox": [0.00174, 0.6187305699481865, 0.26830000000000004, 0.9841968911917098], "iscrowd": 0, "area": 14161.82965, "rle": {"size": [386, 500], "counts": "Yg03k;5L3L5K4L5L3L5K4M4K4L5K4M4K4L5K4M4K4L5L3L5K4L5L3L5K4L5L3L5K4M4K4O2O010O01O001O00001O001O00001O001O00001O001O00001O001O001O000O2O001O00001O00001O000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001N10002N:F9G9G9G9G:F9G9G9G9G9G:FXcY4"}, "label": "a green , unoccupied couch"}]} {"id": 92760, "image": "COCO_train2014_000000092760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man skateboarding with grey pants and dark grey t - shirt on\"."}], "task": "refering", "answer_template": "The \"a man skateboarding with grey pants and dark grey t - shirt on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 30, 31, 32, 33, 48, 49, 50, 51, 66, 67, 68, 85, 86, 104], "bbox": [0.66804, 0.03159159159159159, 0.88022, 0.44255255255255255], "iscrowd": 0, "area": 5831.499599999998, "rle": {"size": [333, 500], "counts": "`e\\369O]9c0M3L4M3N2N2N2N1O2N1O2N1O2O0O1O1O001mG^NZ7c1`HcN_7a1YHeNe7R2O2N1O2N1O2N2N1O2N1OO1O2M3M2N3L3N3L4N1101N1000POUH^Ok7`0WH@j7>XHBg7>ZHAf7>[HBf7<\\HDc7<]HDc7<^HD`7=`HC?De5j0mIB6Kl5d0nIl0P6UOQJj0n5WORJj0l5WOUJh0k5XOUJi0j5XOVJh0i5XOWJh0h5YOXJh0g5jNhI^Ob0h1e5jNjI^O`0h1f5iNlI^O?i1d5iNnI]O>k1b5hNRJ\\O=k1a5iNRJ\\O>j1`5iNSJ^O=R1@UOo5:UJ_O=k0HYOf5WJA=<<[OV5e0SJC<8l63iHF>0n67eHI`0IQ7:aHMc80_G0c8L_G4a8K`G5`8K`G5`8K`G5`8L`G3`8M`G3`8M`G3`8M`G3`8M`G3`8N_G2a8N_G2a8N^G3b8M]G4c8L\\G5f8JYG6m8EQG<_9010O0100O2O0O100O100O4M2Nj^c0"}, "label": "a man skateboarding with grey pants and dark grey t - shirt on"}]} {"id": 92760, "image": "COCO_train2014_000000092760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboarder\"."}], "task": "refering", "answer_template": "The \"a skateboarder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 30, 31, 32, 33, 48, 49, 50, 51, 66, 67, 68, 85, 86, 104], "bbox": [0.66804, 0.03159159159159159, 0.88022, 0.44255255255255255], "iscrowd": 0, "area": 5831.499599999998, "rle": {"size": [333, 500], "counts": "`e\\369O]9c0M3L4M3N2N2N2N1O2N1O2N1O2O0O1O1O001mG^NZ7c1`HcN_7a1YHeNe7R2O2N1O2N1O2N2N1O2N1OO1O2M3M2N3L3N3L4N1101N1000POUH^Ok7`0WH@j7>XHBg7>ZHAf7>[HBf7<\\HDc7<]HDc7<^HD`7=`HC?De5j0mIB6Kl5d0nIl0P6UOQJj0n5WORJj0l5WOUJh0k5XOUJi0j5XOVJh0i5XOWJh0h5YOXJh0g5jNhI^Ob0h1e5jNjI^O`0h1f5iNlI^O?i1d5iNnI]O>k1b5hNRJ\\O=k1a5iNRJ\\O>j1`5iNSJ^O=R1@UOo5:UJ_O=k0HYOf5WJA=<<[OV5e0SJC<8l63iHF>0n67eHI`0IQ7:aHMc80_G0c8L_G4a8K`G5`8K`G5`8K`G5`8L`G3`8M`G3`8M`G3`8M`G3`8M`G3`8N_G2a8N_G2a8N^G3b8M]G4c8L\\G5f8JYG6m8EQG<_9010O0100O2O0O100O100O4M2Nj^c0"}, "label": "a skateboarder"}]} {"id": 43609, "image": "COCO_train2014_000000043609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown truck\"."}], "task": "refering", "answer_template": "The \"brown truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 122, 123, 124, 125, 126, 127, 128, 129, 133, 134, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 218, 219, 220, 221, 222, 241, 242, 243], "bbox": [0.284828125, 0.31303738317757007, 0.866390625, 0.695607476635514], "iscrowd": 0, "area": 37101.043, "rle": {"size": [428, 640], "counts": "lg\\21Z=100O1O1iN?cDAW;f0gD[OS;l0kDTOQ;R1nDnNo:V1oDkNP;W1nDjNP;Y1nDhNQ;[1mDeNQ;^1mDcNR;_1lDbNS;`1lD`NR;n1O001N2O1O1N2O1N2O1O1N2O001N2O1O1N2O1N2O1O2M2M3K5K5O1O1O1O100O1O1O1O1O1O100O1000000O10000O10000000000O100000000000001O00000O100000000000O10000O100O100O100O10000O100O1000000O1000000000000O10000000000000000000001O0000000000000000000b0^O5K5K2N2N2N2N1O00001O00001O00001O001O00001O00001O000000000000000000000000000000000001O1O1O2M2N2O2M2N2N3M4J6H8H7I8M3L4L3M1O10O010000O10O010000O10O101O2N1O1O2M2O1O1O001N101O0000O10O01O1O100O1XORFQNQ:m1PFQNR:n1nERNT:l1mERNU:l1lESNV:l1kESNW:AhEU21YN\\:h1eEVN\\:j1dEUN]:j1eETN\\:l1dESN]:m1dEQN]:o1cEPN^:o1dEoM]:Q2?N200000O101O0000000000001O00000O10001O0000000000001O000O1000001O00000000001O00000O1000001O00000000001O000O100000000000000000O1000O100000O1000000000000O10O10000000O1000000000000O10O10jNZEFf:9[EFf::ZEFf:9\\EFd::\\EEe:;[EEe:;[EEe:;[EEe:;[EDf:;[EEe:;\\EDc:=]EBd:>\\EBd:>\\EBd:>]EAc:?]E@d:?]EAc:?]EAc:?^E_Oc:a0]E_Oc:a0]E_Oc:a0]E_Oc:a0^E]Ob:c0_E]Oa:c0`E\\O`:d0`E[Oa:e0`EZO`:f0`EZO`:f0aEXO`:g0aEYO_:g0aEYO_:g0bEXO^:h0bEWO_:i0aEWO_:i0aEVO_:k0bERO`:m0aEQOl:d0TEYO[;;ViS1"}, "label": "brown truck"}]} {"id": 43609, "image": "COCO_train2014_000000043609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old rusty flatbed pick up truck\"."}], "task": "refering", "answer_template": "The \"an old rusty flatbed pick up truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 122, 123, 124, 125, 126, 127, 128, 129, 133, 134, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 218, 219, 220, 221, 222, 241, 242, 243], "bbox": [0.284828125, 0.31303738317757007, 0.866390625, 0.695607476635514], "iscrowd": 0, "area": 37101.043, "rle": {"size": [428, 640], "counts": "lg\\21Z=100O1O1iN?cDAW;f0gD[OS;l0kDTOQ;R1nDnNo:V1oDkNP;W1nDjNP;Y1nDhNQ;[1mDeNQ;^1mDcNR;_1lDbNS;`1lD`NR;n1O001N2O1O1N2O1N2O1O1N2O001N2O1O1N2O1N2O1O2M2M3K5K5O1O1O1O100O1O1O1O1O1O100O1000000O10000O10000000000O100000000000001O00000O100000000000O10000O100O100O100O10000O100O1000000O1000000000000O10000000000000000000001O0000000000000000000b0^O5K5K2N2N2N2N1O00001O00001O00001O001O00001O00001O000000000000000000000000000000000001O1O1O2M2N2O2M2N2N3M4J6H8H7I8M3L4L3M1O10O010000O10O010000O10O101O2N1O1O2M2O1O1O001N101O0000O10O01O1O100O1XORFQNQ:m1PFQNR:n1nERNT:l1mERNU:l1lESNV:l1kESNW:AhEU21YN\\:h1eEVN\\:j1dEUN]:j1eETN\\:l1dESN]:m1dEQN]:o1cEPN^:o1dEoM]:Q2?N200000O101O0000000000001O00000O10001O0000000000001O000O1000001O00000000001O00000O1000001O00000000001O000O100000000000000000O1000O100000O1000000000000O10O10000000O1000000000000O10O10jNZEFf:9[EFf::ZEFf:9\\EFd::\\EEe:;[EEe:;[EEe:;[EEe:;[EDf:;[EEe:;\\EDc:=]EBd:>\\EBd:>\\EBd:>]EAc:?]E@d:?]EAc:?]EAc:?^E_Oc:a0]E_Oc:a0]E_Oc:a0]E_Oc:a0^E]Ob:c0_E]Oa:c0`E\\O`:d0`E[Oa:e0`EZO`:f0`EZO`:f0aEXO`:g0aEYO_:g0aEYO_:g0bEXO^:h0bEWO_:i0aEWO_:i0aEVO_:k0bERO`:m0aEQOl:d0TEYO[;;ViS1"}, "label": "an old rusty flatbed pick up truck"}]} {"id": 240225, "image": "COCO_train2014_000000240225.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl wearing a yellow shirt with a ladybug on it\"."}], "task": "refering", "answer_template": "The \"a little girl wearing a yellow shirt with a ladybug on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 37, 57, 58, 59, 60, 61, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 133, 147, 148, 149, 150, 151, 152, 153, 155, 156, 173, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 265, 266, 267, 268, 289, 290, 291, 312, 313, 314, 335, 336], "bbox": [0.43032812500000006, 0.0623125, 0.814390625, 0.8588333333333333], "iscrowd": 0, "area": 32632.219950000002, "rle": {"size": [480, 640], "counts": "bRQ4:f>0O1O100O100O1O100O100O10O1000O10O10000O010000O010O1O001O100O001O1O010O1O1O00100O001O1O10O01O1O00100O1O001VMYOZGg0d8HPG8m8:cFF[9`0aF@_9c0^F]Ob9g0ZFYOf9k0WFTOi9n0TFTOj9n0UFROk9o0SFROn9P1oEPOR:Q1kEPOU:R1iEnNX:S1eEoN[:S1bEmN^:U1_ElNb:U1\\EkNb4lNR1Z2YJkN[4WOZ1P2YJiNR4Dd1e1VJiNk3Mo1[1TJhNb3:X2Y2^MSNa2l1TMaNk2^1kLmNT3T1`LYO_3f0YLDf3;RLOm30kK9U4HaKb0^4^OYKl0f4TORKU1l4lNkJ]1U5cNbJg1]5o3O1O001N2O1O001O1N2O001hK\\FQ3f9iLaFU3`9dLhFZ3X9`LPG^3Q9\\LUGc3l8XLYGg3Q:O001O1O001O3N3L6J1O1O10O01O1O10N2O1N3N1N2N2O1N2N3N2M3WFWLW8j3hGVLW8m3fGTLY8o3dGRL\\8P4bGPL]8S4`GoK_8R4`GnK_8U4^GlKa8V4]GkKc8W4[GiKd8Y4ZGiKe8X4ZGhKe8[4XGfKg8\\4WGeKi8]4UGdKi8^4UGcKj8_4UGaKc8i4ZGXKe5N^LR5kMQKe5:nKn4[NiJc5h0^Kh4mNaJc5R8\\JnGa5j8N3L3N3L3K5F;E:gMRFVOY:d0_F_Nk9]1n1K6K4L5N100O2O0O2O00001O001O00001O010O010O00010O010O0100O10O0101N1O100001O00000O10000000000000O100000000000O100O2O0O101O001O01O01O001O001O01O01O0000000O2M2M3N2N2N6J7IYcg1"}, "label": "a little girl wearing a yellow shirt with a ladybug on it"}]} {"id": 240225, "image": "COCO_train2014_000000240225.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small girl wearing pink and yellow dress playing food ball\"."}], "task": "refering", "answer_template": "The \"a small girl wearing pink and yellow dress playing food ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 37, 57, 58, 59, 60, 61, 81, 82, 83, 84, 85, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 133, 147, 148, 149, 150, 151, 152, 153, 155, 156, 173, 174, 175, 176, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 265, 266, 267, 268, 289, 290, 291, 312, 313, 314, 335, 336], "bbox": [0.43032812500000006, 0.0623125, 0.814390625, 0.8588333333333333], "iscrowd": 0, "area": 32632.219950000002, "rle": {"size": [480, 640], "counts": "bRQ4:f>0O1O100O100O1O100O100O10O1000O10O10000O010000O010O1O001O100O001O1O010O1O1O00100O001O1O10O01O1O00100O1O001VMYOZGg0d8HPG8m8:cFF[9`0aF@_9c0^F]Ob9g0ZFYOf9k0WFTOi9n0TFTOj9n0UFROk9o0SFROn9P1oEPOR:Q1kEPOU:R1iEnNX:S1eEoN[:S1bEmN^:U1_ElNb:U1\\EkNb4lNR1Z2YJkN[4WOZ1P2YJiNR4Dd1e1VJiNk3Mo1[1TJhNb3:X2Y2^MSNa2l1TMaNk2^1kLmNT3T1`LYO_3f0YLDf3;RLOm30kK9U4HaKb0^4^OYKl0f4TORKU1l4lNkJ]1U5cNbJg1]5o3O1O001N2O1O001O1N2O001hK\\FQ3f9iLaFU3`9dLhFZ3X9`LPG^3Q9\\LUGc3l8XLYGg3Q:O001O1O001O3N3L6J1O1O10O01O1O10N2O1N3N1N2N2O1N2N3N2M3WFWLW8j3hGVLW8m3fGTLY8o3dGRL\\8P4bGPL]8S4`GoK_8R4`GnK_8U4^GlKa8V4]GkKc8W4[GiKd8Y4ZGiKe8X4ZGhKe8[4XGfKg8\\4WGeKi8]4UGdKi8^4UGcKj8_4UGaKc8i4ZGXKe5N^LR5kMQKe5:nKn4[NiJc5h0^Kh4mNaJc5R8\\JnGa5j8N3L3N3L3K5F;E:gMRFVOY:d0_F_Nk9]1n1K6K4L5N100O2O0O2O00001O001O00001O010O010O00010O010O0100O10O0101N1O100001O00000O10000000000000O100000000000O100O2O0O101O001O01O01O001O001O01O01O0000000O2M2M3N2N2N6J7IYcg1"}, "label": "a small girl wearing pink and yellow dress playing food ball"}]} {"id": 240225, "image": "COCO_train2014_000000240225.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the toddler boy kicking the soccer ball\"."}], "task": "refering", "answer_template": "The \"the toddler boy kicking the soccer ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 26, 27, 28, 29, 49, 50, 51, 52, 70, 71, 72, 73, 74, 75, 95, 96, 97, 98, 118, 119, 120, 121, 141, 142, 143, 164, 165, 166, 187, 188, 189, 210, 211, 212, 213, 235], "bbox": [0.080625, 0.0, 0.302203125, 0.5973333333333334], "iscrowd": 0, "area": 19961.058750000004, "rle": {"size": [480, 640], "counts": "S_h06i>5L2O0O1HFhA;W>EiA;W>EjA:V>GiA9W>GiA9V>HjA7W>IiA7W>KgA4Z>91N100O100O100O101N0001O00001N101O00001O00001O0000001O0O100M4L3M3UEROb7Q1UHA_7b0XHJ^7=YH0[78[H5WO]Ni6c1cIa0@nMk6a1VIS1M^Mj6a1kHa17QMk6c5SI^Jk6e5RI\\JV6AiIW6NYJX6BhIW6MYJZ6AgIY6KXJ^6AdIZ6KWJ`6AcI[6HVJe6AaIW7_6jH`IV7`6kH_IU7a6lH^IT7b6mH\\IT7d6nHZIR7f6oHYIQ7g6PIXIP7h6d00O100O10000O100O100O100O100O100O100O10000O1000RIdHX6[7dIPIV6o7N2N2N0000000000\\GPJX8P6gGRJY8m5gGUJW8k5iGVJV8j5iGXJ_5JlLo5dMYJ^5LkLk5eM[J`5MhLh5gM]J`5NeLf5jM^J`5OcLc5lM_J[58eLY5oMaJQ5c0lLb5l2iJgKhN?`6`3SLXLn3`3\\L^Le3X3gLeLZ3U3mLhLT3T3]M_Ld2Z3[IgLk:o1VE`N9E`:g1`F[N_9f1^F\\Nc9f1YF\\Ng9f1TF]Nk9f1QF\\No9f1lE\\NT:g1hEZNY:i1aEYN`:P2UEQNk:R33L5L4K4L5L6I7J7H8E;Ce0WOY^a6"}, "label": "the toddler boy kicking the soccer ball"}]} {"id": 240225, "image": "COCO_train2014_000000240225.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy in a red & white outfit playing with a soccer ball\"."}], "task": "refering", "answer_template": "The \"a young boy in a red & white outfit playing with a soccer ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 26, 27, 28, 29, 49, 50, 51, 52, 70, 71, 72, 73, 74, 75, 95, 96, 97, 98, 118, 119, 120, 121, 141, 142, 143, 164, 165, 166, 187, 188, 189, 210, 211, 212, 213, 235], "bbox": [0.080625, 0.0, 0.302203125, 0.5973333333333334], "iscrowd": 0, "area": 19961.058750000004, "rle": {"size": [480, 640], "counts": "S_h06i>5L2O0O1HFhA;W>EiA;W>EjA:V>GiA9W>GiA9V>HjA7W>IiA7W>KgA4Z>91N100O100O100O101N0001O00001N101O00001O00001O0000001O0O100M4L3M3UEROb7Q1UHA_7b0XHJ^7=YH0[78[H5WO]Ni6c1cIa0@nMk6a1VIS1M^Mj6a1kHa17QMk6c5SI^Jk6e5RI\\JV6AiIW6NYJX6BhIW6MYJZ6AgIY6KXJ^6AdIZ6KWJ`6AcI[6HVJe6AaIW7_6jH`IV7`6kH_IU7a6lH^IT7b6mH\\IT7d6nHZIR7f6oHYIQ7g6PIXIP7h6d00O100O10000O100O100O100O100O100O100O10000O1000RIdHX6[7dIPIV6o7N2N2N0000000000\\GPJX8P6gGRJY8m5gGUJW8k5iGVJV8j5iGXJ_5JlLo5dMYJ^5LkLk5eM[J`5MhLh5gM]J`5NeLf5jM^J`5OcLc5lM_J[58eLY5oMaJQ5c0lLb5l2iJgKhN?`6`3SLXLn3`3\\L^Le3X3gLeLZ3U3mLhLT3T3]M_Ld2Z3[IgLk:o1VE`N9E`:g1`F[N_9f1^F\\Nc9f1YF\\Ng9f1TF]Nk9f1QF\\No9f1lE\\NT:g1hEZNY:i1aEYN`:P2UEQNk:R33L5L4K4L5L6I7J7H8E;Ce0WOY^a6"}, "label": "a young boy in a red & white outfit playing with a soccer ball"}]} {"id": 551524, "image": "COCO_train2014_000000551524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a red shirt\"."}], "task": "refering", "answer_template": "The \"a woman in a red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372, 373, 374], "bbox": [0.008421874999999999, 0.18891666666666668, 0.31854687499999995, 0.9889166666666667], "iscrowd": 0, "area": 52552.702150000005, "rle": {"size": [480, 640], "counts": "ci23T70@R1BK^OV1:QOeNQ2T1UNkMn2n1WMRMj3g2\\LXLf4d3]K\\Kd5a4_2L4M3L4L4M3L4M3O1N2N2O1N2O1N2N2O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O10000O10001O000O1000000O100000000O1000000O100000000O100000000O100000000001O00001O00001O00010O00001O00001O00001O001O2N1O1O1mHfL^2\\3]MgLc2Z3[MgLe2Z3YMhLf2Y3WMiLi2Y3TMhLl2Y3RMiLm2X3QMiLo2X3nLjLR3W3lLkLS3W3jLjLV3W3gLlLX3U3fLlLZ3V3cLkL]3V3aLlL^3U3_LmLa3U3\\LlLd3U3ZLmLe3T3YLmLg3U3VLmLi3T3ULnLj3S3TLoLk3S3RLoLm3R3QLoLP4Q3nKQMQ4Q3lKQMS4P3kKRMT4P4iJRLV5S4dJoK[5V4_JlK`5X4[JiKe5\\4mInKR6a6O1O1O1O001O1O3M2N2N2N2N3M2N2N2N2N1O1O1O1O2N1O1O1O1PN[FcNf9\\1aF]N`9a1gFYNZ9f1mFSNT9l1RGnMo8Q2XGhMj8U2]GfMc8Y2dG`M]8_2iG[MX8c2PHVMQ8h2V2N2M3M3N2M2N3N2M3M3N2M3M3M3N2M3M3N1N3M3M3L4M3M3L4M3M2M4M3M3L4MXP\\6"}, "label": "a woman in a red shirt"}]} {"id": 551524, "image": "COCO_train2014_000000551524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women with long hair and a marron shirt\"."}], "task": "refering", "answer_template": "The \"a women with long hair and a marron shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372, 373, 374], "bbox": [0.008421874999999999, 0.18891666666666668, 0.31854687499999995, 0.9889166666666667], "iscrowd": 0, "area": 52552.702150000005, "rle": {"size": [480, 640], "counts": "ci23T70@R1BK^OV1:QOeNQ2T1UNkMn2n1WMRMj3g2\\LXLf4d3]K\\Kd5a4_2L4M3L4L4M3L4M3O1N2N2O1N2O1N2N2O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O10000O10001O000O1000000O100000000O1000000O100000000O100000000O100000000001O00001O00001O00010O00001O00001O00001O001O2N1O1O1mHfL^2\\3]MgLc2Z3[MgLe2Z3YMhLf2Y3WMiLi2Y3TMhLl2Y3RMiLm2X3QMiLo2X3nLjLR3W3lLkLS3W3jLjLV3W3gLlLX3U3fLlLZ3V3cLkL]3V3aLlL^3U3_LmLa3U3\\LlLd3U3ZLmLe3T3YLmLg3U3VLmLi3T3ULnLj3S3TLoLk3S3RLoLm3R3QLoLP4Q3nKQMQ4Q3lKQMS4P3kKRMT4P4iJRLV5S4dJoK[5V4_JlK`5X4[JiKe5\\4mInKR6a6O1O1O1O001O1O3M2N2N2N2N3M2N2N2N2N1O1O1O1O2N1O1O1O1PN[FcNf9\\1aF]N`9a1gFYNZ9f1mFSNT9l1RGnMo8Q2XGhMj8U2]GfMc8Y2dG`M]8_2iG[MX8c2PHVMQ8h2V2N2M3M3N2M2N3N2M3M3N2M3M3M3N2M3M3N1N3M3M3L4M3M3L4M3M2M4M3M3L4MXP\\6"}, "label": "a women with long hair and a marron shirt"}]} {"id": 551524, "image": "COCO_train2014_000000551524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing red color tshirt pouring wine in glass\"."}], "task": "refering", "answer_template": "The \"a man wearing red color tshirt pouring wine in glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 109, 110, 111, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 269, 270, 271], "bbox": [0.432515625, 0.21229166666666668, 0.8863749999999999, 0.7083333333333334], "iscrowd": 0, "area": 30402.05660000001, "rle": {"size": [480, 640], "counts": "PQR46\\>0lA0R>4kAMS>5mALo=8oAHP>8QBIm=7UBHj=9VBGi=:XBFh=:WBFi=;VBEk=;UBEk=;TBEl=;TBEm=;SBEl=f00O010000O010000O100002N1O1O1WBRO^=o0^BTOb=T11O101N1O100O1N2O1O001N2O1O0O10N2001N101O0O2001O1O1O1O00001O0000001O0O2O1O0O2O1O1O0O1000000O1000000000000000000000000000001O00001O00001O00001O0O1000000010O001O001O0001O0000000001O00000000000000000001O001O1O1O1REXOb7i0WH^Oh7c0QHCo7?hGHX8:`GLa84XG3g8OQG7o8KiF;W9FbF`0^9BZFd0f9\\2O0000001O000000001O0000001O000000001O00000010O00000001O0000001O00000001O003M00WFSK_9m4_FVKa9i4^FYKa9g4_FZK`9f4_F\\K`9d4`F\\K`9S5N1O1D`JVGb5g8dJTG^5i8?K4M3N2M3N2N2M3O2N101N101N010O100O1O2O0O2O^OUHoIj7S6VHmIi7S6XHlIg7U6ZHkIe7U6\\HjIc7W6^HhIb7X6`HgI^7Z6bHfI]7[6cHfI\\7[6cHeI]7[6cHdI^7\\6bHdI^7\\6bHdI^7\\6bHdI^7\\6bHcI_7]6aHcI_7]6aHcI^7_6aH`I`7_6aHaI_7^6bHbI^7^6bHaI_7^6bHbI^7^6bHbI^7]6cHbI^7]6dHbI]7]6cHcI^7[6cHdI^7[6cHdI_7[6aHdIa7X6cHfI_7V6dHiI]7T6fHkI\\7Q6l0L3N3N2N1O2O1N1O2N2N2K4M4L4K5_LZFc1j9YN[Fb1h9[N\\Fa1h9\\N\\F^1i9`NZF[1j9bNZFY1l9bNWFY1o9bNUFZ1Q:[NXFa1f;K5L4L5K4L4L4L3M3L4M3M3M3M3M4LmPR1"}, "label": "a man wearing red color tshirt pouring wine in glass"}]} {"id": 551524, "image": "COCO_train2014_000000551524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man behind the counter pouring a drink into a glass\"."}], "task": "refering", "answer_template": "The \"a man behind the counter pouring a drink into a glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 109, 110, 111, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 269, 270, 271], "bbox": [0.432515625, 0.21229166666666668, 0.8863749999999999, 0.7083333333333334], "iscrowd": 0, "area": 30402.05660000001, "rle": {"size": [480, 640], "counts": "PQR46\\>0lA0R>4kAMS>5mALo=8oAHP>8QBIm=7UBHj=9VBGi=:XBFh=:WBFi=;VBEk=;UBEk=;TBEl=;TBEm=;SBEl=f00O010000O010000O100002N1O1O1WBRO^=o0^BTOb=T11O101N1O100O1N2O1O001N2O1O0O10N2001N101O0O2001O1O1O1O00001O0000001O0O2O1O0O2O1O1O0O1000000O1000000000000000000000000000001O00001O00001O00001O0O1000000010O001O001O0001O0000000001O00000000000000000001O001O1O1O1REXOb7i0WH^Oh7c0QHCo7?hGHX8:`GLa84XG3g8OQG7o8KiF;W9FbF`0^9BZFd0f9\\2O0000001O000000001O0000001O000000001O00000010O00000001O0000001O00000001O003M00WFSK_9m4_FVKa9i4^FYKa9g4_FZK`9f4_F\\K`9d4`F\\K`9S5N1O1D`JVGb5g8dJTG^5i8?K4M3N2M3N2N2M3O2N101N101N010O100O1O2O0O2O^OUHoIj7S6VHmIi7S6XHlIg7U6ZHkIe7U6\\HjIc7W6^HhIb7X6`HgI^7Z6bHfI]7[6cHfI\\7[6cHeI]7[6cHdI^7\\6bHdI^7\\6bHdI^7\\6bHdI^7\\6bHcI_7]6aHcI_7]6aHcI^7_6aH`I`7_6aHaI_7^6bHbI^7^6bHaI_7^6bHbI^7^6bHbI^7]6cHbI^7]6dHbI]7]6cHcI^7[6cHdI^7[6cHdI_7[6aHdIa7X6cHfI_7V6dHiI]7T6fHkI\\7Q6l0L3N3N2N1O2O1N1O2N2N2K4M4L4K5_LZFc1j9YN[Fb1h9[N\\Fa1h9\\N\\F^1i9`NZF[1j9bNZFY1l9bNWFY1o9bNUFZ1Q:[NXFa1f;K5L4L5K4L4L4L3M3L4M3M3M3M3M4LmPR1"}, "label": "a man behind the counter pouring a drink into a glass"}]} {"id": 551524, "image": "COCO_train2014_000000551524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown hair of the person standing on the far right\"."}], "task": "refering", "answer_template": "The \"the brown hair of the person standing on the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [91, 113, 114, 135, 136, 137, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 341, 342, 343, 344, 363, 364, 365, 366, 367, 386, 387, 388, 389, 390], "bbox": [0.773546875, 0.18597916666666667, 0.998875, 0.9797083333333333], "iscrowd": 0, "area": 34833.93375, "rle": {"size": [480, 640], "counts": "e_X71n>1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1ZMYOQGi0h8ATG`0d8JXG8a82[GO^8:^GH[8a0bG@V8k0eGWOT8R1iGoNP8Z1lGhNm7b1oG_Nj7j1RHWNg7T2UHmMd7\\2XHfMa7c2\\H^Ma7h2ZHZMd7j2YHWMe7m2WHUMg7P3UHQMi7S3SHoLk7V3PHlLn7T5N2N2M3N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2M3N1O2N2N2N2N2N2N2N2O1N2N2N2N2N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1OdN\\1WMi2WMi2WMP?"}, "label": "the brown hair of the person standing on the far right"}]} {"id": 551524, "image": "COCO_train2014_000000551524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hair of a woman standing mainly out of view\"."}], "task": "refering", "answer_template": "The \"the hair of a woman standing mainly out of view\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [91, 113, 114, 135, 136, 137, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 341, 342, 343, 344, 363, 364, 365, 366, 367, 386, 387, 388, 389, 390], "bbox": [0.773546875, 0.18597916666666667, 0.998875, 0.9797083333333333], "iscrowd": 0, "area": 34833.93375, "rle": {"size": [480, 640], "counts": "e_X71n>1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1ZMYOQGi0h8ATG`0d8JXG8a82[GO^8:^GH[8a0bG@V8k0eGWOT8R1iGoNP8Z1lGhNm7b1oG_Nj7j1RHWNg7T2UHmMd7\\2XHfMa7c2\\H^Ma7h2ZHZMd7j2YHWMe7m2WHUMg7P3UHQMi7S3SHoLk7V3PHlLn7T5N2N2M3N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2M3N1O2N2N2N2N2N2N2N2O1N2N2N2N2N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1OdN\\1WMi2WMi2WMP?"}, "label": "the hair of a woman standing mainly out of view"}]} {"id": 174700, "image": "COCO_train2014_000000174700.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden chair leg surrounded by cats\"."}], "task": "refering", "answer_template": "The \"a wooden chair leg surrounded by cats\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 8, 9, 10, 11, 12, 13, 28, 29, 30, 31, 32, 33, 34, 52, 53, 54, 55, 75, 76, 77, 78, 99, 100, 101, 122, 123, 124, 145, 146, 168, 169, 192], "bbox": [0.20410937499999998, 0.0, 0.58721875, 0.4882426778242678], "iscrowd": 0, "area": 15625.643200000002, "rle": {"size": [478, 640], "counts": "nTm15e>5O2N1O2N1O2N1O100O1O1O001O1O1O1O1O1O1O1O1O00100O1O1O1O1O1O1O1O2N1O1O1O1O2N100O5K4L5K4L4L3M3M2N3M2NeMgCR2X in this image.", "objects": [{"patches": [4, 5, 6, 8, 9, 10, 11, 12, 13, 28, 29, 30, 31, 32, 33, 34, 52, 53, 54, 55, 75, 76, 77, 78, 99, 100, 101, 122, 123, 124, 145, 146, 168, 169, 192], "bbox": [0.20410937499999998, 0.0, 0.58721875, 0.4882426778242678], "iscrowd": 0, "area": 15625.643200000002, "rle": {"size": [478, 640], "counts": "nTm15e>5O2N1O2N1O2N1O100O1O1O001O1O1O1O1O1O1O1O1O00100O1O1O1O1O1O1O1O2N1O1O1O1O2N100O5K4L5K4L4L3M3M2N3M2NeMgCR2X in this image.", "objects": [{"patches": [48, 49, 50, 51, 52, 71, 72, 73, 74, 75, 76, 93, 94, 95, 96, 97, 98, 99, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 231, 232, 233, 254], "bbox": [0.0, 0.1532426778242678, 0.32784375, 0.6571338912133892], "iscrowd": 0, "area": 26824.77645, "rle": {"size": [478, 640], "counts": "S7;U>?M3M4M2M3M3N3L3M3M3N3L3M3M3N3L3M3M3N3L3M3M4M2M2N3M2O1N3M2O1O2M2O1O2N1N2O2eE[LZ9f3bF_L\\9c3_FbL^9`3^FfL_9[3^FiL`9Y3[FlLc9U3YFPMd9R3XFSMf9T4N2N2N1N2O001O1N101O1000O2O000O2O0O1O2N1O2N1O1O2N1O1O2N1O2N1O1OPNWKnIh4R6\\KjIc4W6aKeI^4\\6gK]IZ4d6jKVIW4k6nKoHR4R7RLhHo3Y7VL`Hk3a7YLYHh3h7]LQHd3P8`LjGa3W8b11O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1K5I7J6J6J6J5N3O100O1O1O1O100O1O1O100O1O1O1O100O1O1O1O101N1O2N101N1O1O2N101N1O2N1O101N1O2N101N1O1O2N101N1O2N3M3N1N3M3M2M4L4L3M4L4L4M2M4L4L3M4L4L3M4M3L4L3M4LXcX6"}, "label": "cat on left side"}]} {"id": 174700, "image": "COCO_train2014_000000174700.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white cat sleeping to the left of two other cats\"."}], "task": "refering", "answer_template": "The \"a black and white cat sleeping to the left of two other cats\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 51, 52, 71, 72, 73, 74, 75, 76, 93, 94, 95, 96, 97, 98, 99, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 231, 232, 233, 254], "bbox": [0.0, 0.1532426778242678, 0.32784375, 0.6571338912133892], "iscrowd": 0, "area": 26824.77645, "rle": {"size": [478, 640], "counts": "S7;U>?M3M4M2M3M3N3L3M3M3N3L3M3M3N3L3M3M3N3L3M3M4M2M2N3M2O1N3M2O1O2M2O1O2N1N2O2eE[LZ9f3bF_L\\9c3_FbL^9`3^FfL_9[3^FiL`9Y3[FlLc9U3YFPMd9R3XFSMf9T4N2N2N1N2O001O1N101O1000O2O000O2O0O1O2N1O2N1O1O2N1O1O2N1O2N1O1OPNWKnIh4R6\\KjIc4W6aKeI^4\\6gK]IZ4d6jKVIW4k6nKoHR4R7RLhHo3Y7VL`Hk3a7YLYHh3h7]LQHd3P8`LjGa3W8b11O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1K5I7J6J6J6J5N3O100O1O1O1O100O1O1O100O1O1O1O100O1O1O1O101N1O2N101N1O1O2N101N1O2N1O101N1O2N101N1O1O2N101N1O2N3M3N1N3M3M2M4L4L3M4L4L4M2M4L4L3M4L4L3M4M3L4L3M4LXcX6"}, "label": "a black and white cat sleeping to the left of two other cats"}]} {"id": 174700, "image": "COCO_train2014_000000174700.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey cat on a red suitcase between two black cats\"."}], "task": "refering", "answer_template": "The \"a grey cat on a red suitcase between two black cats\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 102, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 284, 285, 286, 308, 309], "bbox": [0.34253125, 0.21274058577405858, 0.629921875, 0.8011087866108787], "iscrowd": 0, "area": 30503.11085, "rle": {"size": [478, 640], "counts": "Z_V33j>8H9G9G8G:G4L4L3L5L4L4L3M4K5L4L3M4L4K4M4L4J6K4K6J6J6K4K6J5K5L4K6J5K5L4L5K4K5L3M2N1N3N2N2N2M3O0O2O1O1N2O00nLaHgN]7[1gHaNX7`1jH^NU7c1mH[NQ7g1PIWNP7j1RITNm6l1VIRNh6P2nHnKEP2\\7S2nHQLEj1\\7X2lHRLFe1\\7\\2kHSLI^1\\7a2iHULJX1\\7e2hHVLLS1Z7j2gHXLNl0Z7n2fHYLOh0Z7Q3eH[L0UOf8i551O1N2O1O1N2N2N2M3M3N2M2N3M3N2O100O1000000O1000000O01000000O1N2O1N2O1N2O1M3J6M31O1O1O001O1O1O1O1O1O001O1O1O01O000O10000O10000O10000O100O10001N10000O100N2N2O1N2N2N3M2jLSG:n8AXG in this image.", "objects": [{"patches": [79, 80, 81, 82, 102, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 284, 285, 286, 308, 309], "bbox": [0.34253125, 0.21274058577405858, 0.629921875, 0.8011087866108787], "iscrowd": 0, "area": 30503.11085, "rle": {"size": [478, 640], "counts": "Z_V33j>8H9G9G8G:G4L4L3L5L4L4L3M4K5L4L3M4L4K4M4L4J6K4K6J6J6K4K6J5K5L4K6J5K5L4L5K4K5L3M2N1N3N2N2N2M3O0O2O1O1N2O00nLaHgN]7[1gHaNX7`1jH^NU7c1mH[NQ7g1PIWNP7j1RITNm6l1VIRNh6P2nHnKEP2\\7S2nHQLEj1\\7X2lHRLFe1\\7\\2kHSLI^1\\7a2iHULJX1\\7e2hHVLLS1Z7j2gHXLNl0Z7n2fHYLOh0Z7Q3eH[L0UOf8i551O1N2O1O1N2N2N2M3M3N2M2N3M3N2O100O1000000O1000000O01000000O1N2O1N2O1N2O1M3J6M31O1O1O001O1O1O1O1O1O001O1O1O01O000O10000O10000O10000O100O10001N10000O100N2N2O1N2N2N3M2jLSG:n8AXG in this image.", "objects": [{"patches": [184, 207, 230, 231, 253, 254, 276, 277, 278, 299, 300, 301, 302, 324, 325, 326, 348, 349], "bbox": [0.0, 0.503305439330544, 0.20564062500000002, 0.9414644351464435], "iscrowd": 0, "area": 7962.468750000002, "rle": {"size": [478, 640], "counts": "b7h2U in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 317], "bbox": [0.524046875, 0.00037656903765690374, 0.9999218750000001, 0.7712133891213389], "iscrowd": 0, "area": 89059.41940000001, "rle": {"size": [478, 640], "counts": "mll46g>5J5L3L4M3L4M2N2O0O2N1O1N3M2N2O2M2N3M2N2N3M2N2N2N3M2N4L4L4L4RJ[M]Oi2ZMWM12c2j2YMWM12c2k2YMTM32b2l2YMUM21c2m2YMSM22e2k2WMVMAa0X3Y2UMXM@b0Z3W2UMYM\\Od0_3S2SM[M_NGBn0l4P2QM^M`NKYOj0V5n1oL^MbN^1^4T1nL`MdN\\1]4U1mLbMeNY1^4U1kLdMgNW1^4U1iLfMhNV1_4T1hLhMhNT1`4U1dLjMkNR1a4T1bLlMlNQ1c4R1_LPNmNn0d4R1\\LSNoNl0e4R1YLTNQOk0f4R1VLVNSOh0g4S1TLVNTOh0h4T1PLVNXOf0h4U1nKWNYOd0i4W1kKVN[Od0j4W1hKWN^Ob0j4X1fKYN^O`0k4Y1dK]N[O in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 317], "bbox": [0.524046875, 0.00037656903765690374, 0.9999218750000001, 0.7712133891213389], "iscrowd": 0, "area": 89059.41940000001, "rle": {"size": [478, 640], "counts": "mll46g>5J5L3L4M3L4M2N2O0O2N1O1N3M2N2O2M2N3M2N2N3M2N2N2N3M2N4L4L4L4RJ[M]Oi2ZMWM12c2j2YMWM12c2k2YMTM32b2l2YMUM21c2m2YMSM22e2k2WMVMAa0X3Y2UMXM@b0Z3W2UMYM\\Od0_3S2SM[M_NGBn0l4P2QM^M`NKYOj0V5n1oL^MbN^1^4T1nL`MdN\\1]4U1mLbMeNY1^4U1kLdMgNW1^4U1iLfMhNV1_4T1hLhMhNT1`4U1dLjMkNR1a4T1bLlMlNQ1c4R1_LPNmNn0d4R1\\LSNoNl0e4R1YLTNQOk0f4R1VLVNSOh0g4S1TLVNTOh0h4T1PLVNXOf0h4U1nKWNYOd0i4W1kKVN[Od0j4W1hKWN^Ob0j4X1fKYN^O`0k4Y1dK]N[O in this image.", "objects": [{"patches": [140, 141, 157, 158, 159, 174, 175, 176, 190, 191, 192, 193, 207, 208, 209, 210, 225, 226, 227, 242, 243, 244, 259, 260, 261], "bbox": [0.1920625, 0.36328125, 0.4140625, 0.6995], "iscrowd": 0, "area": 14660.044800000001, "rle": {"size": [640, 480], "counts": "hji13ic07J7K3M1O1O1O1O100O100O100O100O2M2O1O1O1OOUNPOh@Q1S?XOi@h0V?]Of@c0`0ROdaNmBT1VO0o=gNSCS1ZOVOSO1c>@YCS1IYOR?a0TAZOQ?b0SAZOQ?b0[2K4M3L5KTY_5"}, "label": "a teddy bear wearing a black dress"}]} {"id": 264821, "image": "COCO_train2014_000000264821.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a teddy bear wearing a purple jacket stands next to a bear in a dress\"."}], "task": "refering", "answer_template": "The \"a teddy bear wearing a purple jacket stands next to a bear in a dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 155, 156, 157, 172, 173, 174, 188, 189, 190, 191, 205, 206, 207, 208, 223, 224, 225, 240, 241, 242, 258], "bbox": [0.08202083333333333, 0.352046875, 0.27591666666666664, 0.66240625], "iscrowd": 0, "area": 9711.592999999997, "rle": {"size": [640, 480], "counts": "cZi02lc03N1N3N1O1L4I8N1000O10O10O10N101N2O0O2O0O2O000O2WNBo_O>n0ZOd<e3nA\\LR>e3lA\\LT>d3kA]LT>h40000O1O100O2O1N2O_MVBlNj=R1[BlNd=S1_BlN_=S1eBlNZ=S1iBkNW=T1kBkNU=U1mBiNS=V1nBjNQ=V1PCjNQ=T1PClNP=T1QCkNoZBS1k0ORO_Ni==]BS1i0OTO`Ng=<^BR1i00VO`Nc=>`Bo0g02YO`Na=>eBi0=9C_N[=?lBa01>T>POPB=K<^>WOhA;K5f>@`A9JOo>HXA7JHW?0QA6IA_?9i@4^`0Lg_ONZ`02g_OL[`02g_OLZ`04g_OJZ`05h_OI[`02i_OLlhi6"}, "label": "a teddy bear wearing a purple jacket stands next to a bear in a dress"}]} {"id": 10870, "image": "COCO_train2014_000000010870.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the laptop behind the beer bottle\"."}], "task": "refering", "answer_template": "The \"the laptop behind the beer bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 32, 33, 34, 35, 53, 55, 56, 57, 58, 76, 77, 78, 79, 80, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 148, 167, 168, 169, 171], "bbox": [0.2455625, 0.03906542056074766, 0.56078125, 0.5286448598130841], "iscrowd": 0, "area": 16197.749849999997, "rle": {"size": [428, 640], "counts": "SiQ21[=2M3N3L3N2M3N3L3N2N3L3N2M3N3L3N2N1N101N100O2O0O101O0O101N100O2O0O10001N100O2O0O101N10001N100O2O0O100O2O000O2O0^Ob0XOi0VOi0A?5L4K6J5L4K6J5K6K4K5K6K4K5K6K4K6J7J7H9G9H8G8H9G9H7Hbg21]XM5K5K5J6K5K5K5K5K6J5K5J6K5K5K5K5K5K5K5J6K5K5K5K5K5K5K5J6K5K5K5K5K5K5K3L0100O1O100O1O100O00100O1O1M3M3M3L3N3M3L4M3M3L4M3M2M4M3M3L4M3M3L4M2N3L4M3M3M3L4M3M3L3N3M3L4M3M3L4M3M2M4M3M3L4O100000O10O100000003M4K5L4L4L4L4L4K5L4Ld\\e3"}, "label": "the laptop behind the beer bottle"}]} {"id": 10870, "image": "COCO_train2014_000000010870.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a laptop with a blank screen\"."}], "task": "refering", "answer_template": "The \"a laptop with a blank screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 32, 33, 34, 35, 53, 55, 56, 57, 58, 76, 77, 78, 79, 80, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 148, 167, 168, 169, 171], "bbox": [0.2455625, 0.03906542056074766, 0.56078125, 0.5286448598130841], "iscrowd": 0, "area": 16197.749849999997, "rle": {"size": [428, 640], "counts": "SiQ21[=2M3N3L3N2M3N3L3N2N3L3N2M3N3L3N2N1N101N100O2O0O101O0O101N100O2O0O10001N100O2O0O101N10001N100O2O0O100O2O000O2O0^Ob0XOi0VOi0A?5L4K6J5L4K6J5K6K4K5K6K4K5K6K4K6J7J7H9G9H8G8H9G9H7Hbg21]XM5K5K5J6K5K5K5K5K6J5K5J6K5K5K5K5K5K5K5J6K5K5K5K5K5K5K5J6K5K5K5K5K5K5K3L0100O1O100O1O100O00100O1O1M3M3M3L3N3M3L4M3M3L4M3M2M4M3M3L4M3M3L4M2N3L4M3M3M3L4M3M3L3N3M3L4M3M3L4M3M2M4M3M3L4O100000O10O100000003M4K5L4L4L4L4L4K5L4Ld\\e3"}, "label": "a laptop with a blank screen"}]} {"id": 10870, "image": "COCO_train2014_000000010870.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a partially taken wooden table\"."}], "task": "refering", "answer_template": "The \"a partially taken wooden table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [204, 205, 206, 227, 228, 229, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343, 344], "bbox": [0.8546875, 0.5686214953271028, 1.0, 1.0], "iscrowd": 0, "area": 15764.512200000003, "rle": {"size": [428, 640], "counts": "dPU79gB in this image.", "objects": [{"patches": [54, 55, 56, 57, 77, 78, 79, 80, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 146, 147, 148, 149, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520, 521], "bbox": [0.12255086071987481, 0.101171875, 0.8177464788732394, 1.0], "iscrowd": 0, "area": 152214.32165, "rle": {"size": [640, 639], "counts": "mTa11ec0=H7K5K6J5K5J7J5K5L5K4M3L5K4L4L5K4M3L3M3M3M3M3N2M3M3M3M3M3N1N1O2N2N2N2N2N1O2N2N2N2N2N1O2N2M3N2N2N1O2O1N2N2O1N101N2N2O1N2O0O2N2O1N2N1O2N2N2N2N1TCmIk;U6SDPJh;R6VDSJe;o5YDVJb;l5\\DYJ_;i5`D[J\\;e5cD`JY;a5eDcJY;]5eDhJW;W7mGeGT4h8PKhGP5d8TJlGk5a8YIoGg6]8]HSHb7V:00O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O100O100O100O100O100O100O100O1[MPCZNP=c1]CRNd3jAAV>=XBWOi=f0gBlNZ=R1dCPN^ in this image.", "objects": [{"patches": [54, 55, 56, 57, 77, 78, 79, 80, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 146, 147, 148, 149, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520, 521], "bbox": [0.12255086071987481, 0.101171875, 0.8177464788732394, 1.0], "iscrowd": 0, "area": 152214.32165, "rle": {"size": [640, 639], "counts": "mTa11ec0=H7K5K6J5K5J7J5K5L5K4M3L5K4L4L5K4M3L3M3M3M3M3N2M3M3M3M3M3N1N1O2N2N2N2N2N1O2N2N2N2N2N1O2N2M3N2N2N1O2O1N2N2O1N101N2N2O1N2O0O2N2O1N2N1O2N2N2N2N1TCmIk;U6SDPJh;R6VDSJe;o5YDVJb;l5\\DYJ_;i5`D[J\\;e5cD`JY;a5eDcJY;]5eDhJW;W7mGeGT4h8PKhGP5d8TJlGk5a8YIoGg6]8]HSHb7V:00O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O1O100O100O100O100O100O100O100O100O100O100O1[MPCZNP=c1]CRNd3jAAV>=XBWOi=f0gBlNZ=R1dCPN^ in this image.", "objects": [{"patches": [90, 98, 102, 103, 111, 112, 113, 114, 115, 116, 124, 125, 126, 127, 139, 140], "bbox": [0.54784, 0.36418, 0.99488, 0.5877], "iscrowd": 0, "area": 5166.011699999999, "rle": {"size": [500, 375], "counts": "QZT31a?4L5K4L4L3N2O2N101N101N101N101N101N1O2O0O2O0O2O0O101N1O2O0O2O0O2O0O2O0O2N01O0010ON3K4M4O00001N101O001N101O000O2O001OO001O1O2N1O2N2N2N1O2N2N1O2N2O010O10O1001N2O1N2N2OH\\B`Nb=\\1cBgNY=V1lBhNT=X1nBeNS=[1oBbNS=]1b01O00N2N3M2N2N2N200O2N100O100O2N100O1O100O2O0O1O100O2N100O1O100O2O0O1O100O2N100OmNHlB7T=LjB4V=MiB2W=0iBOV=3iBLW=5iBJW=5jBLV=3kBLU=3lBMT=1oBOP=ORC1n<0RC0m<0SC0n<0RCOn<1RC0m<1RCOn<1SCNm<3RCNn<2PCOP=3jB1X=NgB3Y=MgB2[=MdB4]=JcB6^=JaB6a=H_B9b=F^B9^>0100O1O10O0100O1O100O00eh0"}, "label": "there is chair on which a man is resting"}]} {"id": 141952, "image": "COCO_train2014_000000141952.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green chair seats a homeless man in a white hat\"."}], "task": "refering", "answer_template": "The \"a green chair seats a homeless man in a white hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 98, 102, 103, 111, 112, 113, 114, 115, 116, 124, 125, 126, 127, 139, 140], "bbox": [0.54784, 0.36418, 0.99488, 0.5877], "iscrowd": 0, "area": 5166.011699999999, "rle": {"size": [500, 375], "counts": "QZT31a?4L5K4L4L3N2O2N101N101N101N101N101N1O2O0O2O0O2O0O101N1O2O0O2O0O2O0O2O0O2N01O0010ON3K4M4O00001N101O001N101O000O2O001OO001O1O2N1O2N2N2N1O2N2N1O2N2O010O10O1001N2O1N2N2OH\\B`Nb=\\1cBgNY=V1lBhNT=X1nBeNS=[1oBbNS=]1b01O00N2N3M2N2N2N200O2N100O100O2N100O1O100O2O0O1O100O2N100O1O100O2O0O1O100O2N100OmNHlB7T=LjB4V=MiB2W=0iBOV=3iBLW=5iBJW=5jBLV=3kBLU=3lBMT=1oBOP=ORC1n<0RC0m<0SC0n<0RCOn<1RC0m<1RCOn<1SCNm<3RCNn<2PCOP=3jB1X=NgB3Y=MgB2[=MdB4]=JcB6^=JaB6a=H_B9b=F^B9^>0100O1O10O0100O1O100O00eh0"}, "label": "a green chair seats a homeless man in a white hat"}]} {"id": 141952, "image": "COCO_train2014_000000141952.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting indian style in front of a pizza wearing a red jacket\"."}], "task": "refering", "answer_template": "The \"a man sitting indian style in front of a pizza wearing a red jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 108, 109, 110, 111, 121, 122, 123, 124, 134, 135, 136, 137, 147, 148, 149, 150, 160, 161, 162, 163], "bbox": [0.29962666666666665, 0.38201999999999997, 0.56928, 0.7078599999999999], "iscrowd": 0, "area": 10263.796150000004, "rle": {"size": [500, 375], "counts": "]of11`?>C4K4aN[OkCh0SgEQ1CoN?BV:=kEf0KZO5BU:=WF2KLKBU:?aF^OJ`0AAd:0^G`B_Oc=?j0M4K4Ldi^2"}, "label": "a man sitting indian style in front of a pizza wearing a red jacket"}]} {"id": 141952, "image": "COCO_train2014_000000141952.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in red shirt offering a piece of pizza to her lady love\"."}], "task": "refering", "answer_template": "The \"a man in red shirt offering a piece of pizza to her lady love\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 108, 109, 110, 111, 121, 122, 123, 124, 134, 135, 136, 137, 147, 148, 149, 150, 160, 161, 162, 163], "bbox": [0.29962666666666665, 0.38201999999999997, 0.56928, 0.7078599999999999], "iscrowd": 0, "area": 10263.796150000004, "rle": {"size": [500, 375], "counts": "]of11`?>C4K4aN[OkCh0SgEQ1CoN?BV:=kEf0KZO5BU:=WF2KLKBU:?aF^OJ`0AAd:0^G`B_Oc=?j0M4K4Ldi^2"}, "label": "a man in red shirt offering a piece of pizza to her lady love"}]} {"id": 141952, "image": "COCO_train2014_000000141952.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a standing woman handing a sitting man a plate\"."}], "task": "refering", "answer_template": "The \"a standing woman handing a sitting man a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 15, 16, 17, 18, 28, 29, 30, 31, 40, 41, 42, 43, 53, 54, 55, 56, 66, 67, 68, 69, 79, 80, 81, 82, 92, 93, 94, 95, 104, 105, 106, 117, 118, 119, 130, 131, 143, 144, 156, 157], "bbox": [0.0, 0.0022400000000000002, 0.45543999999999996, 0.71686], "iscrowd": 0, "area": 25359.552850000004, "rle": {"size": [500, 375], "counts": "]::g>c0A`0L3L4M4L3M4K4M3M4K4M3M4L3L5L3M3M4K4M3M4L3L5L3M3L5L3YNSLdGP4Y8eLRG_3j8g1M4M2M3M4L3M3M4LYN]GcL`8_3bGaLZ8`3hG`LU8a3nG^Lo7c3SH]Lj7e3XHZLe7g3]HXLa7i3bHVL[7k3hHTLT7o3mHPLQ7Q4RInKk6S4WIhKk6Y4XIbKi6_4YI\\Kh6g4ZITKg6m4[InJg6S5\\IhJe6Y5^IaJd6a5]I[Jd6f5_ITJb6n5`11O1O2N1001O00100O1O1O1O100O1O1O1O100O1O1O1O100O1O001O100O001O12M3N3M2M4M3M1N2O1O1N2O1O1N2O1O1N2O1O1O0100O100O100O1RNTGSMm8m2XGnLh8R3]GhLd8X3aGcL_8]3fG]L[8c3jGXLV8h3oGRLR8n3SHmKm7R4VHkKk7U4WHhKj7X4WHgKi7Y4XHeKi7[4YHcKg7]4ZHaKg7_4ZH`Kf7`4\\H]Ke7b4g1iMW20000001O0000001O0000001O000000001O00J6J6I7J6L4O1O103L3N3L4M3K5K4K6K5K5K5JcbS3"}, "label": "a standing woman handing a sitting man a plate"}]} {"id": 141952, "image": "COCO_train2014_000000141952.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a scarf is handing a plate to a man who is sitting on the side walk with a pizza , pop up chairs , and another man who seems to be waiting for something\"."}], "task": "refering", "answer_template": "The \"a woman wearing a scarf is handing a plate to a man who is sitting on the side walk with a pizza , pop up chairs , and another man who seems to be waiting for something\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 15, 16, 17, 18, 28, 29, 30, 31, 40, 41, 42, 43, 53, 54, 55, 56, 66, 67, 68, 69, 79, 80, 81, 82, 92, 93, 94, 95, 104, 105, 106, 117, 118, 119, 130, 131, 143, 144, 156, 157], "bbox": [0.0, 0.0022400000000000002, 0.45543999999999996, 0.71686], "iscrowd": 0, "area": 25359.552850000004, "rle": {"size": [500, 375], "counts": "]::g>c0A`0L3L4M4L3M4K4M3M4K4M3M4L3L5L3M3M4K4M3M4L3L5L3M3L5L3YNSLdGP4Y8eLRG_3j8g1M4M2M3M4L3M3M4LYN]GcL`8_3bGaLZ8`3hG`LU8a3nG^Lo7c3SH]Lj7e3XHZLe7g3]HXLa7i3bHVL[7k3hHTLT7o3mHPLQ7Q4RInKk6S4WIhKk6Y4XIbKi6_4YI\\Kh6g4ZITKg6m4[InJg6S5\\IhJe6Y5^IaJd6a5]I[Jd6f5_ITJb6n5`11O1O2N1001O00100O1O1O1O100O1O1O1O100O1O1O1O100O1O001O100O001O12M3N3M2M4M3M1N2O1O1N2O1O1N2O1O1N2O1O1O0100O100O100O1RNTGSMm8m2XGnLh8R3]GhLd8X3aGcL_8]3fG]L[8c3jGXLV8h3oGRLR8n3SHmKm7R4VHkKk7U4WHhKj7X4WHgKi7Y4XHeKi7[4YHcKg7]4ZHaKg7_4ZH`Kf7`4\\H]Ke7b4g1iMW20000001O0000001O0000001O000000001O00J6J6I7J6L4O1O103L3N3L4M3K5K4K6K5K5K5JcbS3"}, "label": "a woman wearing a scarf is handing a plate to a man who is sitting on the side walk with a pizza , pop up chairs , and another man who seems to be waiting for something"}]} {"id": 141952, "image": "COCO_train2014_000000141952.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow chair with a blanket\"."}], "task": "refering", "answer_template": "The \"a yellow chair with a blanket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 177, 178, 179, 181, 189, 190, 191, 192, 193, 194, 202, 203, 204, 205, 206, 207, 216, 217, 218, 219, 220, 229, 230, 231, 232, 233], "bbox": [0.5830666666666667, 0.6499400000000001, 1.0, 1.0], "iscrowd": 0, "area": 17903.502050000006, "rle": {"size": [500, 375], "counts": "`Z[3=V?2N2N2N2O1N2N2N2N2D=M2O2ZCcNQ;_1lDbNU;c1dD^N\\;i1]DWNd;n1UDSNl;R2nCnM_ON[<^2oCjMl;]2QDdMk;b2RD_Mj;g2QD[Ml;Y3L3N4L3N3L3M3M3N1N3M2O1N2N2HYKaEi4R:XKnET5Q:8001O001O1O1O4L2N1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O001O1O001O001O001O1O001O001O00O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N2d0\\O"}, "label": "a yellow chair with a blanket"}]} {"id": 141952, "image": "COCO_train2014_000000141952.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow lawn chair with nobody on it\"."}], "task": "refering", "answer_template": "The \"a yellow lawn chair with nobody on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 177, 178, 179, 181, 189, 190, 191, 192, 193, 194, 202, 203, 204, 205, 206, 207, 216, 217, 218, 219, 220, 229, 230, 231, 232, 233], "bbox": [0.5830666666666667, 0.6499400000000001, 1.0, 1.0], "iscrowd": 0, "area": 17903.502050000006, "rle": {"size": [500, 375], "counts": "`Z[3=V?2N2N2N2O1N2N2N2N2D=M2O2ZCcNQ;_1lDbNU;c1dD^N\\;i1]DWNd;n1UDSNl;R2nCnM_ON[<^2oCjMl;]2QDdMk;b2RD_Mj;g2QD[Ml;Y3L3N4L3N3L3M3M3N1N3M2O1N2N2HYKaEi4R:XKnET5Q:8001O001O1O1O4L2N1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O001O001O1O001O001O001O1O001O001O00O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N2d0\\O"}, "label": "a yellow lawn chair with nobody on it"}]} {"id": 141952, "image": "COCO_train2014_000000141952.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white hat relaxing in a lawn chair\"."}], "task": "refering", "answer_template": "The \"a man in a white hat relaxing in a lawn chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 90, 96, 97, 98, 99, 100, 101, 102, 103, 112, 113, 114, 115, 116], "bbox": [0.43488000000000004, 0.3207, 1.0, 0.46812], "iscrowd": 0, "area": 5134.71635, "rle": {"size": [500, 375], "counts": "]h_24_?5K6J6I in this image.", "objects": [{"patches": [83, 84, 85, 90, 96, 97, 98, 99, 100, 101, 102, 103, 112, 113, 114, 115, 116], "bbox": [0.43488000000000004, 0.3207, 1.0, 0.46812], "iscrowd": 0, "area": 5134.71635, "rle": {"size": [500, 375], "counts": "]h_24_?5K6J6I in this image.", "objects": [{"patches": [21, 22, 23, 38, 39, 40, 41, 56, 57, 58, 74, 75, 76, 93, 108, 109, 126], "bbox": [0.2810416666666667, 0.12809701492537312, 0.5345, 0.7348134328358209], "iscrowd": 0, "area": 5731.918699999999, "rle": {"size": [268, 480], "counts": "W\\S14P8:L4M2O2M2O2N1N3N1N100WITOR6l0hIYOX6h0cI^O\\6b0bIA\\6`0cIB\\6>cID\\6[1Hm3HhJa0[1Eo3IgJa0Z1EQ4JeJa0Z1DS4IdJc0X1CV4IcJe0V1AX4IcJg0S1]O_4J_Jj0Q7UOoHl0Y71O1O1OO100O1O001O1O1O1O1O1O1O1O001O100O1O1O1O1O001O1O1O1L4L3M]Wj1"}, "label": "a man riding a bike in desert runes"}]} {"id": 551553, "image": "COCO_train2014_000000551553.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person riding on a yellow dirt bike\"."}], "task": "refering", "answer_template": "The \"person riding on a yellow dirt bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 23, 38, 39, 40, 41, 56, 57, 58, 74, 75, 76, 93, 108, 109, 126], "bbox": [0.2810416666666667, 0.12809701492537312, 0.5345, 0.7348134328358209], "iscrowd": 0, "area": 5731.918699999999, "rle": {"size": [268, 480], "counts": "W\\S14P8:L4M2O2M2O2N1N3N1N100WITOR6l0hIYOX6h0cI^O\\6b0bIA\\6`0cIB\\6>cID\\6[1Hm3HhJa0[1Eo3IgJa0Z1EQ4JeJa0Z1DS4IdJc0X1CV4IcJe0V1AX4IcJg0S1]O_4J_Jj0Q7UOoHl0Y71O1O1OO100O1O001O1O1O1O1O1O1O1O001O100O1O1O1O1O001O1O1O1L4L3M]Wj1"}, "label": "person riding on a yellow dirt bike"}]} {"id": 43655, "image": "COCO_train2014_000000043655.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty chair directly to the left of the leftmost of two cats and behind a second empty chair that is closer to the foreground\"."}], "task": "refering", "answer_template": "The \"an empty chair directly to the left of the leftmost of two cats and behind a second empty chair that is closer to the foreground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 115, 138, 139, 161, 162, 184, 185, 207, 208, 209, 230, 231, 232, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302], "bbox": [0.0, 0.31613577023498696, 0.17759375, 0.9892428198433421], "iscrowd": 0, "area": 15822.0785, "rle": {"size": [383, 640], "counts": "U:d1R7Y3QMo20000O10000O10000O10000O1003M2N3M3M3M3N2M3M3M3M2N3M3M3M3M3M3M3M3M2N3M3M3M3M3M3M3M3M2N3M3M3M3M3M3M3M2N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N2N3M2O2M2N3M2N3M2N3M2N2N3M2N3M2N3M2N3M2N2N3M2NkgT6"}, "label": "an empty chair directly to the left of the leftmost of two cats and behind a second empty chair that is closer to the foreground"}]} {"id": 43655, "image": "COCO_train2014_000000043655.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat in a bed to the right of another cat\"."}], "task": "refering", "answer_template": "The \"a cat in a bed to the right of another cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 223, 224, 225], "bbox": [0.604515625, 0.3643080939947781, 0.8590625, 0.660757180156658], "iscrowd": 0, "area": 12436.184650000001, "rle": {"size": [383, 640], "counts": "\\m`43k;6J5K5K7I6J2NO11O3M3M2N3M2N2N1O2N1O2N1O2N2N2N3M3M3M3M3N1N3N1N2O0O2O0O2O1N101N101N2O0O2O1O00010O00010O010O00010O00010O00010O001O01O01O00010O00010O001O00001O0010O0001O001O00000001O00001N10001O001O00001O001O00001O001N10001O001O00001O00001O00001N100O2N100O2O0O101N1O101N100O2O0O101N1Oa0@3L4M3L4M1N1O2O0O2N1O2N1O2N1O2N1O2N1O5K6J7I\\^Q1"}, "label": "a cat in a bed to the right of another cat"}]} {"id": 43655, "image": "COCO_train2014_000000043655.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat , sitting in a cushion , with all its paws inside the cushion\"."}], "task": "refering", "answer_template": "The \"a cat , sitting in a cushion , with all its paws inside the cushion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 223, 224, 225], "bbox": [0.604515625, 0.3643080939947781, 0.8590625, 0.660757180156658], "iscrowd": 0, "area": 12436.184650000001, "rle": {"size": [383, 640], "counts": "\\m`43k;6J5K5K7I6J2NO11O3M3M2N3M2N2N1O2N1O2N1O2N2N2N3M3M3M3M3N1N3N1N2O0O2O0O2O1N101N101N2O0O2O1O00010O00010O010O00010O00010O00010O001O01O01O00010O00010O001O00001O0010O0001O001O00000001O00001N10001O001O00001O001O00001O001N10001O001O00001O00001O00001N100O2N100O2O0O101N1O101N100O2O0O101N1Oa0@3L4M3L4M1N1O2O0O2N1O2N1O2N1O2N1O2N1O5K6J7I\\^Q1"}, "label": "a cat , sitting in a cushion , with all its paws inside the cushion"}]} {"id": 43655, "image": "COCO_train2014_000000043655.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair sits empty behind two cats\"."}], "task": "refering", "answer_template": "The \"a chair sits empty behind two cats\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 59, 77, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 124, 125, 126, 127, 128], "bbox": [0.379171875, 0.1566840731070496, 0.608328125, 0.42822454308093993], "iscrowd": 0, "area": 12877.086650000003, "rle": {"size": [383, 640], "counts": "Ynj2:e; in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 59, 77, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 124, 125, 126, 127, 128], "bbox": [0.379171875, 0.1566840731070496, 0.608328125, 0.42822454308093993], "iscrowd": 0, "area": 12877.086650000003, "rle": {"size": [383, 640], "counts": "Ynj2:e; in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 215, 216, 217, 239, 240], "bbox": [0.294734375, 0.37921671018276765, 0.543328125, 0.7582245430809399], "iscrowd": 0, "area": 14088.30965, "rle": {"size": [383, 640], "counts": "YlV23j;5K4L5L3L5K4L4M4K4L3M3N1N3M3M2N3N2M3M2N3N2M2N3N2O0O2O1O1O001N2O001O1N101O1O1O0O2O1O001N1000001N10001O000O2O00000O2O00001O0O3N2N2N2M3N2N2N2M3N2N2O1N2N2N2N2N2O1N2N0000O2O000O10001N100O100cN[H`Ne7`1^H\\Nd7c1`HYN`7g1cHVN]7k1eHQN\\7o1hHmMX7S2kHjMV7U2mHgMT7Y2oHdMQ7\\2SI_Mn6a2UI\\Mk6d2XIYMi6f2X1001O000001O0000001O001O100O001O1O1O001O1O1O00100O1O1O001O1O1O001O100O001O1O1O001O1O1O00100O1O4L8H8G9C=C=BPP]3"}, "label": "cat with one paw outside his bed"}]} {"id": 43655, "image": "COCO_train2014_000000043655.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat on the left side sitting next to another cat\"."}], "task": "refering", "answer_template": "The \"a cat on the left side sitting next to another cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 215, 216, 217, 239, 240], "bbox": [0.294734375, 0.37921671018276765, 0.543328125, 0.7582245430809399], "iscrowd": 0, "area": 14088.30965, "rle": {"size": [383, 640], "counts": "YlV23j;5K4L5L3L5K4L4M4K4L3M3N1N3M3M2N3N2M3M2N3N2M2N3N2O0O2O1O1O001N2O001O1N101O1O1O0O2O1O001N1000001N10001O000O2O00000O2O00001O0O3N2N2N2M3N2N2N2M3N2N2O1N2N2N2N2N2O1N2N0000O2O000O10001N100O100cN[H`Ne7`1^H\\Nd7c1`HYN`7g1cHVN]7k1eHQN\\7o1hHmMX7S2kHjMV7U2mHgMT7Y2oHdMQ7\\2SI_Mn6a2UI\\Mk6d2XIYMi6f2X1001O000001O0000001O001O100O001O1O1O001O1O1O00100O1O1O001O1O1O001O100O001O1O1O001O1O1O00100O1O4L8H8G9C=C=BPP]3"}, "label": "a cat on the left side sitting next to another cat"}]} {"id": 51851, "image": "COCO_train2014_000000051851.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white toilet with a black seat\"."}], "task": "refering", "answer_template": "The \"a white toilet with a black seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [248, 249, 250, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 314, 315, 316, 317, 318, 338, 339, 340], "bbox": [0.671921875, 0.6584222737819024, 0.8989218749999999, 0.9797679814385151], "iscrowd": 0, "area": 10513.607000000007, "rle": {"size": [431, 640], "counts": "T[e5`0`dD^O_;b0bD[O`;f0`DVOR< in this image.", "objects": [{"patches": [248, 249, 250, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 314, 315, 316, 317, 318, 338, 339, 340], "bbox": [0.671921875, 0.6584222737819024, 0.8989218749999999, 0.9797679814385151], "iscrowd": 0, "area": 10513.607000000007, "rle": {"size": [431, 640], "counts": "T[e5`0`dD^O_;b0bD[O`;f0`DVOR< in this image.", "objects": [{"patches": [57, 60, 80, 82, 83, 87, 88, 103, 104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 151, 152, 153, 154, 174, 175, 176, 177, 197, 198, 199, 200, 220, 221, 222, 223, 243, 244, 245, 246, 266, 267, 268, 269, 289, 290, 291, 292, 312, 313, 314, 335, 336], "bbox": [0.49173437499999995, 0.12298780487804878, 0.8663281250000001, 0.8089634146341463], "iscrowd": 0, "area": 25384.867899999997, "rle": {"size": [492, 640], "counts": "Q]g49S?>C;D:G9F:G9F:G2OO4K4M4K4L5L3L5L3L5K4M4K5L3L5L2M00O2O001O001O00001N101O00001O001O0O2O000cFRO\\4o0_KYO]4g0`K@\\4a0`KF\\4:`KN[44aK3[4MaK;Y3aNbIT1R3b0W3\\NfIS1o2h0V3WNjIS1j2n0W3QNnIS1f2S1W3lMRJS1b2X1W3gMVJS1]2[1[3dMVJT1Z2\\1^3bMWJT1U2`1a3^MYJT1Q2b1d3\\MZJT1m1d1g3ZM[JT1h1h1j3VM]JT1d1j1m3TM^JT1_1m1Q4QM^JU1\\1o1S4QM]JR1[1Q2V4SMZJn0Z1T2Z4TMWJj0Z1W2]4[O]Kj0b4XOYKl0f4VOUKo0i4SORKQ1m4QOoJR1P5oNmJZ1l4hNoJ[1Q5gNkJ[1U5gNgJZ1Z5hNbJnN5oNY5Y5oJaJQ5]5VK^Jj4a5\\KZJd4d5dKVJ\\4h5kKSJU4l5QLoIo3o5YLkIg3S6_LiIa3V6eLeIZ3Z6lLbIT3\\6SM_Im2`6XM\\Ih2b6_LYInN0c4e6_L^IQOI`4h6^LbI`4\\6`KhI^4V6bKmIj4e5UK_JY5o4fJUKg5[4YJiKe5V4ZJmKe5Q4[JSLc5k3]JVLd5h3[JZLf5e3YJ\\Lh5c3WJ^Lj5`3VJaLk5^3TJcLl5]3RJeLo5Z3PJgLQ6X3nIiLS6U3mIlLT6S3kInLV6Q3iIPMX6o2fISM[6n2bISM_6n2]ITMd6m2WIVMj6k2QIXMP7j2jHYMW7h2eHZM[7h2`H[Ma7f2ZH]Mg7d2TH_Mm7o41O1O1O1O1O1O1N2M2N3L4M3M3M3M3M3L4aLmFb0V9UOVGe0m8RO_Gh0d8oNhGk0[8lNQHn0Q8jN[Ho0i7kN`Ho0c7POaHj0b7VOaHd0b7[ObH?a7AbH9a7GbH2b7MbHMa73bHGa78cHB`7>cH\\O`7c0dHWO_7i0dHPO_7P1_300O10000O100O100O100O010O1O10O01O010O1O010O0010O01O10O010O01O010O1O010O0010O01O10O01O010O00100O0010O01O010O10O01O010O0001O0O101N100100O1O100O100O0100003M2M4M3M3MSWY1"}, "label": "a white andble unfirom with the number 14 on it"}]} {"id": 289425, "image": "COCO_train2014_000000289425.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"playing man\"."}], "task": "refering", "answer_template": "The \"playing man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 60, 80, 82, 83, 87, 88, 103, 104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 151, 152, 153, 154, 174, 175, 176, 177, 197, 198, 199, 200, 220, 221, 222, 223, 243, 244, 245, 246, 266, 267, 268, 269, 289, 290, 291, 292, 312, 313, 314, 335, 336], "bbox": [0.49173437499999995, 0.12298780487804878, 0.8663281250000001, 0.8089634146341463], "iscrowd": 0, "area": 25384.867899999997, "rle": {"size": [492, 640], "counts": "Q]g49S?>C;D:G9F:G9F:G2OO4K4M4K4L5L3L5L3L5K4M4K5L3L5L2M00O2O001O001O00001N101O00001O001O0O2O000cFRO\\4o0_KYO]4g0`K@\\4a0`KF\\4:`KN[44aK3[4MaK;Y3aNbIT1R3b0W3\\NfIS1o2h0V3WNjIS1j2n0W3QNnIS1f2S1W3lMRJS1b2X1W3gMVJS1]2[1[3dMVJT1Z2\\1^3bMWJT1U2`1a3^MYJT1Q2b1d3\\MZJT1m1d1g3ZM[JT1h1h1j3VM]JT1d1j1m3TM^JT1_1m1Q4QM^JU1\\1o1S4QM]JR1[1Q2V4SMZJn0Z1T2Z4TMWJj0Z1W2]4[O]Kj0b4XOYKl0f4VOUKo0i4SORKQ1m4QOoJR1P5oNmJZ1l4hNoJ[1Q5gNkJ[1U5gNgJZ1Z5hNbJnN5oNY5Y5oJaJQ5]5VK^Jj4a5\\KZJd4d5dKVJ\\4h5kKSJU4l5QLoIo3o5YLkIg3S6_LiIa3V6eLeIZ3Z6lLbIT3\\6SM_Im2`6XM\\Ih2b6_LYInN0c4e6_L^IQOI`4h6^LbI`4\\6`KhI^4V6bKmIj4e5UK_JY5o4fJUKg5[4YJiKe5V4ZJmKe5Q4[JSLc5k3]JVLd5h3[JZLf5e3YJ\\Lh5c3WJ^Lj5`3VJaLk5^3TJcLl5]3RJeLo5Z3PJgLQ6X3nIiLS6U3mIlLT6S3kInLV6Q3iIPMX6o2fISM[6n2bISM_6n2]ITMd6m2WIVMj6k2QIXMP7j2jHYMW7h2eHZM[7h2`H[Ma7f2ZH]Mg7d2TH_Mm7o41O1O1O1O1O1O1N2M2N3L4M3M3M3M3M3L4aLmFb0V9UOVGe0m8RO_Gh0d8oNhGk0[8lNQHn0Q8jN[Ho0i7kN`Ho0c7POaHj0b7VOaHd0b7[ObH?a7AbH9a7GbH2b7MbHMa73bHGa78cHB`7>cH\\O`7c0dHWO_7i0dHPO_7P1_300O10000O100O100O100O010O1O10O01O010O1O010O0010O01O10O010O01O010O1O010O0010O01O10O01O010O00100O0010O01O010O10O01O010O0001O0O101N100100O1O100O100O0100003M2M4M3M3MSWY1"}, "label": "playing man"}]} {"id": 289425, "image": "COCO_train2014_000000289425.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is one person wearing white dress and helmet looking at somewhere\"."}], "task": "refering", "answer_template": "The \"there is one person wearing white dress and helmet looking at somewhere\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 119, 120, 141, 142, 143, 144, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 210, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 259, 260, 261, 280, 281, 282, 283, 303, 304, 305, 306, 326, 327, 328, 329, 348, 349, 351, 352, 371, 372, 373, 374, 375], "bbox": [0.118, 0.2286178861788618, 0.354, 0.9204268292682928], "iscrowd": 0, "area": 26264.695549999997, "rle": {"size": [492, 640], "counts": "jgT14U?4M4K4L5L4K4M4K4M4K5K4M4K5L3L5L3M2O100O100O100O1_HfMP2Z2lMjMS2W2jMlMV2T2gMoMX2R2dMSN[2m1bMVN]2k1`MXN`2i1\\M[Nb2h1YM[Ng2g1TM\\Nk2g1QM[No2f1mL^NQ3e1jL^NV3d1fL^NY3d1fIRMP2]1Z4c1nHhMc2i0]4_2_KcMa4_2ZKdMe4_2VKdMj4^2QKfMm4]2oJeMo4_2lJbMU5a2fJ_M\\5d2_J\\Mb5h2XJYMj5j2QJVMc4YNYKe4NSMk4gNhJY48PMS5lN_JW48oL[5lNYJY47nL_5lNWJY44nLf5jNTJ[41mLk5o60O10O01O01OI8N1O2N1O1O2N1O2N2N2N2N2N2N2N3M4K4M3M4L3M3M4L3[IZH`5`2gIa2e0UK^5X2QJe2 in this image.", "objects": [{"patches": [96, 97, 119, 120, 141, 142, 143, 144, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 210, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 259, 260, 261, 280, 281, 282, 283, 303, 304, 305, 306, 326, 327, 328, 329, 348, 349, 351, 352, 371, 372, 373, 374, 375], "bbox": [0.118, 0.2286178861788618, 0.354, 0.9204268292682928], "iscrowd": 0, "area": 26264.695549999997, "rle": {"size": [492, 640], "counts": "jgT14U?4M4K4L5L4K4M4K4M4K5K4M4K5L3L5L3M2O100O100O100O1_HfMP2Z2lMjMS2W2jMlMV2T2gMoMX2R2dMSN[2m1bMVN]2k1`MXN`2i1\\M[Nb2h1YM[Ng2g1TM\\Nk2g1QM[No2f1mL^NQ3e1jL^NV3d1fL^NY3d1fIRMP2]1Z4c1nHhMc2i0]4_2_KcMa4_2ZKdMe4_2VKdMj4^2QKfMm4]2oJeMo4_2lJbMU5a2fJ_M\\5d2_J\\Mb5h2XJYMj5j2QJVMc4YNYKe4NSMk4gNhJY48PMS5lN_JW48oL[5lNYJY47nL_5lNWJY44nLf5jNTJ[41mLk5o60O10O01O01OI8N1O2N1O1O2N1O2N2N2N2N2N2N2N3M4K4M3M4L3M3M4L3[IZH`5`2gIa2e0UK^5X2QJe2 in this image.", "objects": [{"patches": [57, 70, 71, 83, 84, 97, 98, 110, 111, 112, 113, 124, 125, 126, 139, 152, 153, 165, 166, 178, 179, 192], "bbox": [0.41648, 0.22696, 0.806, 0.80898], "iscrowd": 0, "area": 7446.659450000002, "rle": {"size": [500, 375], "counts": "fY\\27h>i0N2M3M3M3M3N2M3M3M3N2M3N2N2NO1N2O1N2O1N2N2O1N2O2M2N2O1N2O1N2NSBkN]=V1a01O2N1O1O1ON2N2N2N2OdAXOk=Y11O1O1O1O1N2O1O1O1O1O1O1O1OO100O010O100O100O100O100O100O010O100O100O100O010O01O001O010O001O010O00000010O0000010O0001O01O00100O1O1O100O1O100O1O1O100O1O101N1O1O18H8h0XO9G3NL3M3ML5K4LO1O1N2O1N2O1O1N2O1O1NW1jN in this image.", "objects": [{"patches": [57, 70, 71, 83, 84, 97, 98, 110, 111, 112, 113, 124, 125, 126, 139, 152, 153, 165, 166, 178, 179, 192], "bbox": [0.41648, 0.22696, 0.806, 0.80898], "iscrowd": 0, "area": 7446.659450000002, "rle": {"size": [500, 375], "counts": "fY\\27h>i0N2M3M3M3M3N2M3M3M3N2M3N2N2NO1N2O1N2O1N2N2O1N2O2M2N2O1N2O1N2NSBkN]=V1a01O2N1O1O1ON2N2N2N2OdAXOk=Y11O1O1O1O1N2O1O1O1O1O1O1O1OO100O010O100O100O100O100O100O010O100O100O100O010O01O001O010O001O010O00000010O0000010O0001O01O00100O1O1O100O1O100O1O1O100O1O101N1O1O18H8h0XO9G3NL3M3ML5K4LO1O1N2O1N2O1O1N2O1O1NW1jN in this image.", "objects": [{"patches": [97, 109, 110, 111, 123, 124, 125, 136, 137, 138, 139, 150, 151, 152, 163, 165, 176, 178], "bbox": [0.4351466666666667, 0.4403, 0.7592533333333334, 0.78038], "iscrowd": 0, "area": 7730.628700000001, "rle": {"size": [500, 375], "counts": "ii_21c?3M3M3M3M3M2N1O1O1O2N000000000fA^Ob=b0\\BBb=>[BGc=9ZBLd=4ZB0d=0YB5e=KZB8d=HZBG:F9GTb[1"}, "label": "a giraffe that ' s the smallest in the picture"}]} {"id": 510611, "image": "COCO_train2014_000000510611.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smallest giraffe with it ' s head in the feed cage\"."}], "task": "refering", "answer_template": "The \"the smallest giraffe with it ' s head in the feed cage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 109, 110, 111, 123, 124, 125, 136, 137, 138, 139, 150, 151, 152, 163, 165, 176, 178], "bbox": [0.4351466666666667, 0.4403, 0.7592533333333334, 0.78038], "iscrowd": 0, "area": 7730.628700000001, "rle": {"size": [500, 375], "counts": "ii_21c?3M3M3M3M3M2N1O1O1O2N000000000fA^Ob=b0\\BBb=>[BGc=9ZBLd=4ZB0d=0YB5e=KZB8d=HZBG:F9GTb[1"}, "label": "the smallest giraffe with it ' s head in the feed cage"}]} {"id": 510611, "image": "COCO_train2014_000000510611.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe on the far left\"."}], "task": "refering", "answer_template": "The \"the giraffe on the far left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 79, 80, 81, 92, 93, 104, 105, 106, 117, 118, 130, 131, 143, 144, 156, 157, 170], "bbox": [0.00039999999999999996, 0.31282, 0.26744, 0.7475], "iscrowd": 0, "area": 6787.7198499999995, "rle": {"size": [500, 375], "counts": "R8g1l=1O001O1O1O001O1O1O001O1O1O001O1O001N2O1O00105J6J6JA9G7I7J6I2N0010O0I7I6J6I7L4O2M2O1O1O1O2M2O000O2lNfClN] in this image.", "objects": [{"patches": [67, 79, 80, 81, 92, 93, 104, 105, 106, 117, 118, 130, 131, 143, 144, 156, 157, 170], "bbox": [0.00039999999999999996, 0.31282, 0.26744, 0.7475], "iscrowd": 0, "area": 6787.7198499999995, "rle": {"size": [500, 375], "counts": "R8g1l=1O001O1O1O001O1O1O001O1O1O001O1O001N2O1O00105J6J6JA9G7I7J6I2N0010O0I7I6J6I7L4O2M2O1O1O1O2M2O000O2lNfClN] in this image.", "objects": [{"patches": [92, 115, 116, 138, 139, 140, 161, 162, 163, 184, 185, 186, 207, 208, 230, 231, 253, 254, 255, 276, 277, 278, 300, 301], "bbox": [0.0, 0.2733411214953271, 0.10023437500000001, 0.8952803738317757], "iscrowd": 0, "area": 11324.612200000001, "rle": {"size": [428, 640], "counts": "e34l0g0Z5VOUOm1oK^N95c0BV3n1bKSO^O>]1cNe3P2RK\\1P1fLP4Q2dJe1V1\\LW4S6bKoI`4T6ZKnIg4U6RKmIP5Z6eJhI]5_6YJcIh5o62O0O3N3L4M4L3L4M1N2O1O1[NbIYL`6]3mI\\LW6_3QJVLW6m3jIgK^6[4cIZKe6h4o02M3N3M2N3M3M3M2N3M3M3M3M2O2M9G;FN1O1O1O2N1O1O2kLSIGn62]IHe62dII_6LnIOT6GZJ3h5CeJ8]5^OPK=S5XO[Kb0g4UOeKf0]4POPLk0S4jN[LP1U8K=BT]`7"}, "label": "a man wearing a black shirt and black watch standing at the back left of a commercial kitchen"}]} {"id": 174749, "image": "COCO_train2014_000000174749.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man all in black\"."}], "task": "refering", "answer_template": "The \"a man all in black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 115, 116, 138, 139, 140, 161, 162, 163, 184, 185, 186, 207, 208, 230, 231, 253, 254, 255, 276, 277, 278, 300, 301], "bbox": [0.0, 0.2733411214953271, 0.10023437500000001, 0.8952803738317757], "iscrowd": 0, "area": 11324.612200000001, "rle": {"size": [428, 640], "counts": "e34l0g0Z5VOUOm1oK^N95c0BV3n1bKSO^O>]1cNe3P2RK\\1P1fLP4Q2dJe1V1\\LW4S6bKoI`4T6ZKnIg4U6RKmIP5Z6eJhI]5_6YJcIh5o62O0O3N3L4M4L3L4M1N2O1O1[NbIYL`6]3mI\\LW6_3QJVLW6m3jIgK^6[4cIZKe6h4o02M3N3M2N3M3M3M2N3M3M3M3M2O2M9G;FN1O1O1O2N1O1O2kLSIGn62]IHe62dII_6LnIOT6GZJ3h5CeJ8]5^OPK=S5XO[Kb0g4UOeKf0]4POPLk0S4jN[LP1U8K=BT]`7"}, "label": "a man all in black"}]} {"id": 174749, "image": "COCO_train2014_000000174749.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chef wearing a black apron\"."}], "task": "refering", "answer_template": "The \"a chef wearing a black apron\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 30, 31, 32, 52, 53, 54, 55, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 118, 119, 120, 121, 122, 123, 124, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 326, 327, 328, 329, 330, 331], "bbox": [0.145328125, 0.04123831775700934, 0.47620312499999995, 0.9996728971962616], "iscrowd": 0, "area": 58987.737550000005, "rle": {"size": [428, 640], "counts": "jQW14V=8G8I7H9H7H9H7hGVNa4R2YKRNb4V2XKmMc4Z2nI`MD;W6^2gI`M15S6b2`IaM;1o5f2YIcMf0Jl5o3RJULh5o3WJULc5n3ZJWL`5m3^JXL]5j3`J]LZ5e3dJ`LW5b3gJcLT5^3jJiLP5Y3nJmLm4S3QKSMj4n2UKXMe4j2YK\\Ma4e2]KbM]4`2aKeMZ4\\2dKgMZ4[2dKfM[4[2dKfM[4f5O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O100O100O1O101N1O100O100O1O1O1O1O1O1O2O0O1O1O1O1N2N2N2N2N2N2L4L4L4L4M3L4N2N2N2N2N2N2O1N2N2N2N101000000O10000O10000O1000000O10000O10000000000000000000QG^OT5b0kJ@T5`0lJAS5?lJCS5=lJES5;mJFR5:mJHR58mJJR56mJLR54nJMQ53nJ0P50oJ3o4NPK4n4MPK6n4KPK8n4HQK;m4FRKl4CRK`0l4@TKb0j4_OTKd0j4]OTKf0_3bNfJi0i1h0]3bNiJg0h1j0\\3aNkJf0f1n0Z3^NPKe0d1Q1W3]NTKc0b1U1U3[NXKa0a1X1R3ZN\\K?`1\\1o2WN`K?]1_1n2UNaK`0_1_1k2TNaKb0b1^1h2SNaKd0d1^1f2QN`Kg0h1\\1c2PN`Ki0k1[1a2oM^Kl0m1Z1`23]M2^21^M5]2N_M8\\2KaM:[2HaM>[2F_Mf0X2^OYMa1R2b5[O`0_O6K3L2O0O2O0O2N2O001O001O001O1O001N101O1O007I:F9Goe[4"}, "label": "a chef wearing a black apron"}]} {"id": 174749, "image": "COCO_train2014_000000174749.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white and black striped apron\"."}], "task": "refering", "answer_template": "The \"a man in a white and black striped apron\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 30, 31, 32, 52, 53, 54, 55, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 118, 119, 120, 121, 122, 123, 124, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 286, 303, 304, 305, 306, 307, 308, 326, 327, 328, 329, 330, 331], "bbox": [0.145328125, 0.04123831775700934, 0.47620312499999995, 0.9996728971962616], "iscrowd": 0, "area": 58987.737550000005, "rle": {"size": [428, 640], "counts": "jQW14V=8G8I7H9H7H9H7hGVNa4R2YKRNb4V2XKmMc4Z2nI`MD;W6^2gI`M15S6b2`IaM;1o5f2YIcMf0Jl5o3RJULh5o3WJULc5n3ZJWL`5m3^JXL]5j3`J]LZ5e3dJ`LW5b3gJcLT5^3jJiLP5Y3nJmLm4S3QKSMj4n2UKXMe4j2YK\\Ma4e2]KbM]4`2aKeMZ4\\2dKgMZ4[2dKfM[4[2dKfM[4f5O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O100O100O1O101N1O100O100O1O1O1O1O1O1O2O0O1O1O1O1N2N2N2N2N2N2L4L4L4L4M3L4N2N2N2N2N2N2O1N2N2N2N101000000O10000O10000O1000000O10000O10000000000000000000QG^OT5b0kJ@T5`0lJAS5?lJCS5=lJES5;mJFR5:mJHR58mJJR56mJLR54nJMQ53nJ0P50oJ3o4NPK4n4MPK6n4KPK8n4HQK;m4FRKl4CRK`0l4@TKb0j4_OTKd0j4]OTKf0_3bNfJi0i1h0]3bNiJg0h1j0\\3aNkJf0f1n0Z3^NPKe0d1Q1W3]NTKc0b1U1U3[NXKa0a1X1R3ZN\\K?`1\\1o2WN`K?]1_1n2UNaK`0_1_1k2TNaKb0b1^1h2SNaKd0d1^1f2QN`Kg0h1\\1c2PN`Ki0k1[1a2oM^Kl0m1Z1`23]M2^21^M5]2N_M8\\2KaM:[2HaM>[2F_Mf0X2^OYMa1R2b5[O`0_O6K3L2O0O2O0O2N2O001O001O001O1O001N101O1O007I:F9Goe[4"}, "label": "a man in a white and black striped apron"}]} {"id": 371361, "image": "COCO_train2014_000000371361.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skier wearing black\"."}], "task": "refering", "answer_template": "The \"the skier wearing black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 81, 82, 103, 104, 105, 106, 126, 127, 128, 129, 149, 150, 151, 152, 172, 173, 174, 175, 196, 197, 198, 218, 219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267], "bbox": [0.49437499999999995, 0.18143192488262913, 0.64328125, 0.7787323943661972], "iscrowd": 0, "area": 16876.363000000005, "rle": {"size": [426, 640], "counts": "XdS41T=8H8H8H8H8I7H8H8cG_Nh4g1oJkNc4[1RKlNm4X1jJoNS5V1dJPO[5T1\\JROb5S1TJTOk5Q1kIUOT6o0cIXOZ6m0\\IZOc6j0TI\\Oj6c3O1O2M2O1N2O1N3XOg0O1O100O2N1O1O1O1O2O00I7J6I7I7I7I7I7I7I7I7I7o1QN1N101O1O1O1O1O10O10000000000O1000O01N2VI^Ji5d5lIgJR6[5cIPK[6Q6M4M2N2kLRIKP74YIBj6=_IXOc6g0fInN]6Q1lIeNV6Z1TJ[No5c1ZJSNh5l1bJiM`5V2iJ`MY5`2j2N1O1O1O2N1O100O?Ad0VOj0ROXen2"}, "label": "the skier wearing black"}]} {"id": 371361, "image": "COCO_train2014_000000371361.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing black ski gear riding a t - bar lift\"."}], "task": "refering", "answer_template": "The \"a person wearing black ski gear riding a t - bar lift\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 81, 82, 103, 104, 105, 106, 126, 127, 128, 129, 149, 150, 151, 152, 172, 173, 174, 175, 196, 197, 198, 218, 219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267], "bbox": [0.49437499999999995, 0.18143192488262913, 0.64328125, 0.7787323943661972], "iscrowd": 0, "area": 16876.363000000005, "rle": {"size": [426, 640], "counts": "XdS41T=8H8H8H8H8I7H8H8cG_Nh4g1oJkNc4[1RKlNm4X1jJoNS5V1dJPO[5T1\\JROb5S1TJTOk5Q1kIUOT6o0cIXOZ6m0\\IZOc6j0TI\\Oj6c3O1O2M2O1N2O1N3XOg0O1O100O2N1O1O1O1O2O00I7J6I7I7I7I7I7I7I7I7I7o1QN1N101O1O1O1O1O10O10000000000O1000O01N2VI^Ji5d5lIgJR6[5cIPK[6Q6M4M2N2kLRIKP74YIBj6=_IXOc6g0fInN]6Q1lIeNV6Z1TJ[No5c1ZJSNh5l1bJiM`5V2iJ`MY5`2j2N1O1O1O2N1O100O?Ad0VOj0ROXen2"}, "label": "a person wearing black ski gear riding a t - bar lift"}]} {"id": 199331, "image": "COCO_train2014_000000199331.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue shirt and blue shorts holding a green cone\"."}], "task": "refering", "answer_template": "The \"a man in a blue shirt and blue shorts holding a green cone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 129, 130, 131, 152, 153, 154, 176, 177, 198, 199, 200, 221, 222, 223, 243, 244, 246, 269], "bbox": [0.5821562499999999, 0.2742857142857143, 0.7371718749999999, 0.7467447306791569], "iscrowd": 0, "area": 7649.003300000004, "rle": {"size": [427, 640], "counts": "ejk42W=2N3N101N101O0O101N101N101N101O0O2O0O2O0O2O001N1O1M3L4L4M3L4^KdNQM`1P3VOVLk0m3WOlK?c4CWK?k4CnJ`0T5BfJ`0\\5A`J`0b5BXJ`0j5CoI>T6EeI<^6H[I8h6KQI6R7IlH7X7GfH9]7FaH:b7E\\H3P8KnG5U8IjG7d7oN]Gi0m08g7RO\\Gd0k0:j7VOYG?k02V8BnF;j01[8HjF8f02`8HiFT2V9oMiFo1X9SNhFk1X9WNgFh1Y9T12O1O1N3N100001O00000002N2N3cLYFP3f1gLZ65RHT3b1hLe6Y3YIiLf6Y3WIiLh6[3SIgLl6a1jG2U1_NP7_1lG5o0^NU7\\1lG;i0ZN[7[1mG?b0XN`7Y1oGd0:UNf7W1QHi02QNm7V1QHb1P8^NQH`1o7`NRH_1n7aNSHoNMc1P8^OTHkN1e1_9lMaF:6h1h9TN]Fj1e9QN`Fm1a:N2N2N2N7I:F9E in this image.", "objects": [{"patches": [106, 107, 129, 130, 131, 152, 153, 154, 176, 177, 198, 199, 200, 221, 222, 223, 243, 244, 246, 269], "bbox": [0.5821562499999999, 0.2742857142857143, 0.7371718749999999, 0.7467447306791569], "iscrowd": 0, "area": 7649.003300000004, "rle": {"size": [427, 640], "counts": "ejk42W=2N3N101N101O0O101N101N101N101O0O2O0O2O0O2O001N1O1M3L4L4M3L4^KdNQM`1P3VOVLk0m3WOlK?c4CWK?k4CnJ`0T5BfJ`0\\5A`J`0b5BXJ`0j5CoI>T6EeI<^6H[I8h6KQI6R7IlH7X7GfH9]7FaH:b7E\\H3P8KnG5U8IjG7d7oN]Gi0m08g7RO\\Gd0k0:j7VOYG?k02V8BnF;j01[8HjF8f02`8HiFT2V9oMiFo1X9SNhFk1X9WNgFh1Y9T12O1O1N3N100001O00000002N2N3cLYFP3f1gLZ65RHT3b1hLe6Y3YIiLf6Y3WIiLh6[3SIgLl6a1jG2U1_NP7_1lG5o0^NU7\\1lG;i0ZN[7[1mG?b0XN`7Y1oGd0:UNf7W1QHi02QNm7V1QHb1P8^NQH`1o7`NRH_1n7aNSHoNMc1P8^OTHkN1e1_9lMaF:6h1h9TN]Fj1e9QN`Fm1a:N2N2N2N7I:F9E in this image.", "objects": [{"patches": [40, 41, 63, 64, 86, 87, 88, 109, 110, 111, 112, 132, 133, 134, 135, 155, 156, 157, 158, 178, 179, 180, 181, 201, 202, 203, 224, 225, 226, 227, 248, 250, 251, 252, 271, 272, 274, 275, 293, 294, 295, 297], "bbox": [0.7351875, 0.07850117096018736, 0.9698281249999999, 0.8269320843091335], "iscrowd": 0, "area": 19044.949249999994, "rle": {"size": [427, 640], "counts": "d^T659M_<7]CL_<:]CI``JbN49h0T1b42RJXO=Cj0Y1e4LQJC9XOk0`1i4EPJM6oNk0e1m4[1QJPMo0j1m4h1QK\\NT1gMa1m3YM`NT1dMb1l3XMdNT1bMc1i3WMjNR1_Mf1g3VMnNR1\\Mg1f3TMSOS1XMh1f3RMVOS1VMk1f3mLXOV1SMl1g3jLZOX1oLn1i3cL_O\\1iLR2a7lM_HT2c7iM^HX2c7fM]H[2d7bM]H_2d7]M^Hd2c7YM^Hh2k5_LUKf0QOl2g5bLXK?ROR3a5cL\\K9TO^3T5]LhKOXOn3e4VLSLE]OV4^4WLVL]O@]4W4ZL]Me3c2^LZMc3d2aLZM_3e2dLXM]3g2gLVMY3h2mLSMT3l2SMnLn2o2[MiLg2U3]MgLe2\\3YMbLh2b3VMZLl2i3SMTLn2P4PMmKQ3U4nLiKT3X4lLdKV3]4jLaKW3`4iL^KY3b4fL]K[3e4dLXK^3h4cLVK^3j4bLUK_3l4aLRK`3m4bLPK`3P5_LPKb3P5^LoJd3o4]LPKd3P5\\LnJf3Q5[LnJf3R5ZLlJh3T5XLjJj3U5WLhJl3X5TLfJn3Z5RLdJP4[5PLdJR4\\5nKbJT4^5n1N10000gN^HgLc7S3cHmL]7m2iHSMW7d2QI\\MP7c2QI]MR7_2oHaMT7[2lHeMX7V2jHjMY7R2gHoM\\7m1eHRN^7k1cHUN`7F_GZ1Q1QOd7]OdG\\1i0WOf7VOkG[1`0_Of8;[GEj85VGLn8OSG1Q9JPG5U9FlF:S;01O0O2O00001N10001O0O2O00010O00000101N2N2N3M3M?A2M3N1OO0O2O1N100O2O0O100O101N1O100O1O2M2N3M2ChCD\\<9 in this image.", "objects": [{"patches": [40, 41, 63, 64, 86, 87, 88, 109, 110, 111, 112, 132, 133, 134, 135, 155, 156, 157, 158, 178, 179, 180, 181, 201, 202, 203, 224, 225, 226, 227, 248, 250, 251, 252, 271, 272, 274, 275, 293, 294, 295, 297], "bbox": [0.7351875, 0.07850117096018736, 0.9698281249999999, 0.8269320843091335], "iscrowd": 0, "area": 19044.949249999994, "rle": {"size": [427, 640], "counts": "d^T659M_<7]CL_<:]CI``JbN49h0T1b42RJXO=Cj0Y1e4LQJC9XOk0`1i4EPJM6oNk0e1m4[1QJPMo0j1m4h1QK\\NT1gMa1m3YM`NT1dMb1l3XMdNT1bMc1i3WMjNR1_Mf1g3VMnNR1\\Mg1f3TMSOS1XMh1f3RMVOS1VMk1f3mLXOV1SMl1g3jLZOX1oLn1i3cL_O\\1iLR2a7lM_HT2c7iM^HX2c7fM]H[2d7bM]H_2d7]M^Hd2c7YM^Hh2k5_LUKf0QOl2g5bLXK?ROR3a5cL\\K9TO^3T5]LhKOXOn3e4VLSLE]OV4^4WLVL]O@]4W4ZL]Me3c2^LZMc3d2aLZM_3e2dLXM]3g2gLVMY3h2mLSMT3l2SMnLn2o2[MiLg2U3]MgLe2\\3YMbLh2b3VMZLl2i3SMTLn2P4PMmKQ3U4nLiKT3X4lLdKV3]4jLaKW3`4iL^KY3b4fL]K[3e4dLXK^3h4cLVK^3j4bLUK_3l4aLRK`3m4bLPK`3P5_LPKb3P5^LoJd3o4]LPKd3P5\\LnJf3Q5[LnJf3R5ZLlJh3T5XLjJj3U5WLhJl3X5TLfJn3Z5RLdJP4[5PLdJR4\\5nKbJT4^5n1N10000gN^HgLc7S3cHmL]7m2iHSMW7d2QI\\MP7c2QI]MR7_2oHaMT7[2lHeMX7V2jHjMY7R2gHoM\\7m1eHRN^7k1cHUN`7F_GZ1Q1QOd7]OdG\\1i0WOf7VOkG[1`0_Of8;[GEj85VGLn8OSG1Q9JPG5U9FlF:S;01O0O2O00001N10001O0O2O00010O00000101N2N2N3M3M?A2M3N1OO0O2O1N100O2O0O100O101N1O100O1O2M2N3M2ChCD\\<9 in this image.", "objects": [{"patches": [79, 80, 102, 103, 124, 125, 126, 127, 147, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 215, 216, 217, 218, 219, 238, 239, 242, 260, 261], "bbox": [0.311671875, 0.2546838407494145, 0.5675156250000001, 0.7765573770491804], "iscrowd": 0, "area": 14552.376700000003, "rle": {"size": [427, 640], "counts": "aYc21X=6K4L1N2O100000000O100000000O010000000O010000000O1M2O2M3N2M3N1HWOmCl0Q<8N2N2O0O2N1O10O01OO1I8M2000001O00001O10O2UNQOYGo0d8]ORGd0l8EmFlFCS9`0iFAV9c0gF]OX9f0eF\\OY9h0dFXO\\9k0aFUO`9[2O01O0100O100O2O0O100O100O1mNTLgHk3Y7_L]Ha3c7eLXHZ3i7kLRHT3n7RMmGm2S8XMiGg2X8U1001O0O101O1O1O1O1O001O1O1N2O1O1O12N101N1O2N101N1OWLdGe2Z8ZMhGf2X8YMiGh2U8WMmGi2R8VMQHj2m7UMUHk2k7SMVHo2R91O2M2O2N00000001O001O0O101O001O05L6I7J1N1O1000UNXGQOj8m0XGROi84VGjN4P1g8IaGVOJP1f8[OnGC_OP1e8ZOnGE_OP1d8XOoGG_OP1c8WOoGH@P1b8UOPHIAP1a8TOUHF]OS1`8TOUIj0h9N7I7I6J1O1O1O3M4L5L`Vc3"}, "label": "man withyellow shirt with a blue vest on top , black shorts with the number 79 playing soccer"}]} {"id": 199331, "image": "COCO_train2014_000000199331.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with white and orange t - shirt , black trousers playing football\"."}], "task": "refering", "answer_template": "The \"a man with white and orange t - shirt , black trousers playing football\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 102, 103, 124, 125, 126, 127, 147, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 215, 216, 217, 218, 219, 238, 239, 242, 260, 261], "bbox": [0.311671875, 0.2546838407494145, 0.5675156250000001, 0.7765573770491804], "iscrowd": 0, "area": 14552.376700000003, "rle": {"size": [427, 640], "counts": "aYc21X=6K4L1N2O100000000O100000000O010000000O010000000O1M2O2M3N2M3N1HWOmCl0Q<8N2N2O0O2N1O10O01OO1I8M2000001O00001O10O2UNQOYGo0d8]ORGd0l8EmFlFCS9`0iFAV9c0gF]OX9f0eF\\OY9h0dFXO\\9k0aFUO`9[2O01O0100O100O2O0O100O100O1mNTLgHk3Y7_L]Ha3c7eLXHZ3i7kLRHT3n7RMmGm2S8XMiGg2X8U1001O0O101O1O1O1O1O001O1O1N2O1O1O12N101N1O2N101N1OWLdGe2Z8ZMhGf2X8YMiGh2U8WMmGi2R8VMQHj2m7UMUHk2k7SMVHo2R91O2M2O2N00000001O001O0O101O001O05L6I7J1N1O1000UNXGQOj8m0XGROi84VGjN4P1g8IaGVOJP1f8[OnGC_OP1e8ZOnGE_OP1d8XOoGG_OP1c8WOoGH@P1b8UOPHIAP1a8TOUHF]OS1`8TOUIj0h9N7I7I6J1O1O1O3M4L5L`Vc3"}, "label": "a man with white and orange t - shirt , black trousers playing football"}]} {"id": 404139, "image": "COCO_train2014_000000404139.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with a beard in black shirt and blue jeans\"."}], "task": "refering", "answer_template": "The \"man with a beard in black shirt and blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 144, 145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 213, 214, 215, 236, 237, 238, 259, 260, 261, 283, 284, 306, 307, 328, 329, 330, 352, 353, 354, 376, 377], "bbox": [0.268453125, 0.3111875, 0.427859375, 0.9532291666666667], "iscrowd": 0, "area": 14685.580100000003, "rle": {"size": [480, 640], "counts": "hi`22m>4^HK37IK69EI::CG<=_OE`0>QJFd3NZ2`0SIe0Z4mNc2`0oHl0U4fNk2b0lHP1Q4_NR3d0jHU1k3YNZ3S3]LoLc3Y3SLiLm3Z3nKhLR4[3jKfLV4]3eKeL[4]3aKeL_4^3\\KdLd4^3XKcLi4`3RKbLn4`3oJaLQ5b3jJ`LV5b3fJ`LZ5c3aJ_L_5d3\\J^Le5c3XJ^Lj5c3QJ_LP6b3lI`LU6b3fI`L\\6`3`IbLa6`3[IaLf6`3VIbLl6_3oHcLR7_3iHcLX7^3dHdL^7]3^HdLc7\\51N3N1O1O in this image.", "objects": [{"patches": [124, 144, 145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 213, 214, 215, 236, 237, 238, 259, 260, 261, 283, 284, 306, 307, 328, 329, 330, 352, 353, 354, 376, 377], "bbox": [0.268453125, 0.3111875, 0.427859375, 0.9532291666666667], "iscrowd": 0, "area": 14685.580100000003, "rle": {"size": [480, 640], "counts": "hi`22m>4^HK37IK69EI::CG<=_OE`0>QJFd3NZ2`0SIe0Z4mNc2`0oHl0U4fNk2b0lHP1Q4_NR3d0jHU1k3YNZ3S3]LoLc3Y3SLiLm3Z3nKhLR4[3jKfLV4]3eKeL[4]3aKeL_4^3\\KdLd4^3XKcLi4`3RKbLn4`3oJaLQ5b3jJ`LV5b3fJ`LZ5c3aJ_L_5d3\\J^Le5c3XJ^Lj5c3QJ_LP6b3lI`LU6b3fI`L\\6`3`IbLa6`3[IaLf6`3VIbLl6_3oHcLR7_3iHcLX7^3dHdL^7]3^HdLc7\\51N3N1O1O in this image.", "objects": [{"patches": [122, 123, 124, 139, 140, 141, 142, 157, 158, 159, 160, 175, 176, 177, 178, 193, 194, 195, 196, 211, 212, 213, 229, 230, 231], "bbox": [0.73456, 0.48624, 0.92316, 0.9655199999999999], "iscrowd": 0, "area": 12429.842950000002, "rle": {"size": [375, 500], "counts": "ddV43a;;F:E8lF^Og6h0SI[Oi6l0RIVOk6S1lHPOP7a1aHaN]7c1^H_N`7g1ZH\\Nd7h1WH[Nf7m2N2M2O1M4K4M3L4M3L5L3O1O1O1O2N1O1O1O1O10O0100O100O100O10O010000O10O1O1O010O1M3K4N3N_I]KR6a4oIeKm5X4UJiKj5U4WJlKi5Q4ZJoKf5o3[JRLe5l3[JVLe5i3WJ\\Li5d3SJ`Lm5`3PJdLo5\\3mIhLS6^40000001O010O1O0O2N2M2O2YOeIiKb6S2\\ISOe0dNT6W2`IkN_7S1gHeN\\7Z1gHaN]7]1hH[N]7c1kHPN[7o1X1M2N4L3K5L5J5L4K6I6H:G:Flf="}, "label": "a man in an orange shirt bending down on a tennis court"}]} {"id": 494257, "image": "COCO_train2014_000000494257.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person in the orange shirt\"."}], "task": "refering", "answer_template": "The \"the person in the orange shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 139, 140, 141, 142, 157, 158, 159, 160, 175, 176, 177, 178, 193, 194, 195, 196, 211, 212, 213, 229, 230, 231], "bbox": [0.73456, 0.48624, 0.92316, 0.9655199999999999], "iscrowd": 0, "area": 12429.842950000002, "rle": {"size": [375, 500], "counts": "ddV43a;;F:E8lF^Og6h0SI[Oi6l0RIVOk6S1lHPOP7a1aHaN]7c1^H_N`7g1ZH\\Nd7h1WH[Nf7m2N2M2O1M4K4M3L4M3L5L3O1O1O1O2N1O1O1O1O10O0100O100O100O10O010000O10O1O1O010O1M3K4N3N_I]KR6a4oIeKm5X4UJiKj5U4WJlKi5Q4ZJoKf5o3[JRLe5l3[JVLe5i3WJ\\Li5d3SJ`Lm5`3PJdLo5\\3mIhLS6^40000001O010O1O0O2N2M2O2YOeIiKb6S2\\ISOe0dNT6W2`IkN_7S1gHeN\\7Z1gHaN]7]1hH[N]7c1kHPN[7o1X1M2N4L3K5L5J5L4K6I6H:G:Flf="}, "label": "the person in the orange shirt"}]} {"id": 494257, "image": "COCO_train2014_000000494257.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl on the tennis court lagging while her friend is in a bag\"."}], "task": "refering", "answer_template": "The \"a girl on the tennis court lagging while her friend is in a bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 57, 58, 74, 75, 76, 92, 93, 94, 110, 111, 112, 128, 129, 130, 147, 148, 165, 166, 183, 184, 185], "bbox": [0.13784, 0.21738666666666664, 0.30004, 0.85008], "iscrowd": 0, "area": 10450.973199999999, "rle": {"size": [375, 500], "counts": "Y]i0K5J6K5K6_FgNY8^1aGdN^8^1_GdN`8^1]GdNa8_1[GdNd8^1YGdNg8]1VGeNi8\\1UGfNj8T2O1IVM`Gk2m7\\MWHKKj2l7iMSHZ2i7iMUH^2i0lL<0X4V4TKWL7Fb4]5VKeJh4b5QKaJm4f5jJ]JT5T6O1O2N10000001^OQKZJo4b5WK\\Ji4`5_K\\Ja4l4WKUKh5h4f0L3J6C=H71O100O1O012N4L4M3L4L4M2TKUKf2o4TMSKl2Q5nLRKR3f5SL\\Jm3f5oK]JP4e5lK]JV4c5gK^J[4f5_K[Jb4l5UKTJm4Z62lIRK\\5R5]JPKc5U5VJnJi5`510O1nK[JP2e5oM]JQ2c5lM_JK94X5iNTKH^O8_1Q1o3jNlK3S17Q3ARL6n09U5@oJa0R5YORKg0V8J4K3N2M2O000M2Nkko3"}, "label": "a girl on the tennis court lagging while her friend is in a bag"}]} {"id": 494257, "image": "COCO_train2014_000000494257.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman smile and stand in a tennis court\"."}], "task": "refering", "answer_template": "The \"a woman smile and stand in a tennis court\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 57, 58, 74, 75, 76, 92, 93, 94, 110, 111, 112, 128, 129, 130, 147, 148, 165, 166, 183, 184, 185], "bbox": [0.13784, 0.21738666666666664, 0.30004, 0.85008], "iscrowd": 0, "area": 10450.973199999999, "rle": {"size": [375, 500], "counts": "Y]i0K5J6K5K6_FgNY8^1aGdN^8^1_GdN`8^1]GdNa8_1[GdNd8^1YGdNg8]1VGeNi8\\1UGfNj8T2O1IVM`Gk2m7\\MWHKKj2l7iMSHZ2i7iMUH^2i0lL<0X4V4TKWL7Fb4]5VKeJh4b5QKaJm4f5jJ]JT5T6O1O2N10000001^OQKZJo4b5WK\\Ji4`5_K\\Ja4l4WKUKh5h4f0L3J6C=H71O100O1O012N4L4M3L4L4M2TKUKf2o4TMSKl2Q5nLRKR3f5SL\\Jm3f5oK]JP4e5lK]JV4c5gK^J[4f5_K[Jb4l5UKTJm4Z62lIRK\\5R5]JPKc5U5VJnJi5`510O1nK[JP2e5oM]JQ2c5lM_JK94X5iNTKH^O8_1Q1o3jNlK3S17Q3ARL6n09U5@oJa0R5YORKg0V8J4K3N2M2O000M2Nkko3"}, "label": "a woman smile and stand in a tennis court"}]} {"id": 174774, "image": "COCO_train2014_000000174774.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"right monitor in the right hand picture\"."}], "task": "refering", "answer_template": "The \"right monitor in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228], "bbox": [0.829515625, 0.0, 1.0, 0.630936768149883], "iscrowd": 0, "area": 24943.68170000001, "rle": {"size": [427, 640], "counts": "T_m67S=:F:F:F:F;D;F:F:F:F:F:F:F:F:E;F:F:F:F:F:F:F;E:E3N1O1O1O1O1O2N1O1N2O1O1O1O1O1O1O1O1N101O1O1O01000000O1000000O1000000O100000000O1000000O1000000O1000000O100000000O1000000O101O000O1000000O100000000O1000000O1000000O100000O"}, "label": "right monitor in the right hand picture"}]} {"id": 174774, "image": "COCO_train2014_000000174774.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black computer screen on the right\"."}], "task": "refering", "answer_template": "The \"the black computer screen on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228], "bbox": [0.829515625, 0.0, 1.0, 0.630936768149883], "iscrowd": 0, "area": 24943.68170000001, "rle": {"size": [427, 640], "counts": "T_m67S=:F:F:F:F;D;F:F:F:F:F:F:F:F:E;F:F:F:F:F:F:F;E:E3N1O1O1O1O1O2N1O1N2O1O1O1O1O1O1O1O1N101O1O1O01000000O1000000O1000000O100000000O1000000O1000000O1000000O100000000O1000000O101O000O1000000O100000000O1000000O1000000O100000O"}, "label": "the black computer screen on the right"}]} {"id": 174774, "image": "COCO_train2014_000000174774.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the computer screen with the image of the dinosaur\"."}], "task": "refering", "answer_template": "The \"the computer screen with the image of the dinosaur\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 59, 60, 61, 62, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 108, 125, 126, 127, 128, 129, 130, 131, 148, 149, 150, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244], "bbox": [0.44582812499999996, 0.06053864168618267, 0.740734375, 0.711288056206089], "iscrowd": 0, "area": 38539.243650000004, "rle": {"size": [427, 640], "counts": "oSg3`0Z8a40O01O1O100O1O1O010O1O1O1O10O01O1O100O1O1O010O1O1O1O10O01O1O100O1O1O010O1O1O1O10O01O1O100O1O1O010O1O1O100O001O1O100O1O1O010O1O1O100O001O1O100O1O1O010O1O1O100O001O1O100O1O1O010O1O1O100O001O1O100O1O1O010O1O1O100O001O100O1O1O1O010O1O1O100O001O100O1O1O1O010O1O1O100O001O100O1O1O1O010O1O1O100O1O00100O1O1O1O010O100001O001N101O00001O0O2O00001O001L3K5K6J5K8H?A?A?Aa0_Oa0^Oc0^Oa0_Oa0_Oa0_Oa0_Oa0_Oa0\\OjQU2"}, "label": "the computer screen with the image of the dinosaur"}]} {"id": 174774, "image": "COCO_train2014_000000174774.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"television with a raptor on the screen\"."}], "task": "refering", "answer_template": "The \"television with a raptor on the screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 59, 60, 61, 62, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 107, 108, 125, 126, 127, 128, 129, 130, 131, 148, 149, 150, 151, 152, 153, 154, 171, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244], "bbox": [0.44582812499999996, 0.06053864168618267, 0.740734375, 0.711288056206089], "iscrowd": 0, "area": 38539.243650000004, "rle": {"size": [427, 640], "counts": "oSg3`0Z8a40O01O1O100O1O1O010O1O1O1O10O01O1O100O1O1O010O1O1O1O10O01O1O100O1O1O010O1O1O1O10O01O1O100O1O1O010O1O1O100O001O1O100O1O1O010O1O1O100O001O1O100O1O1O010O1O1O100O001O1O100O1O1O010O1O1O100O001O1O100O1O1O010O1O1O100O001O100O1O1O1O010O1O1O100O001O100O1O1O1O010O1O1O100O001O100O1O1O1O010O1O1O100O1O00100O1O1O1O010O100001O001N101O00001O0O2O00001O001L3K5K6J5K8H?A?A?Aa0_Oa0^Oc0^Oa0_Oa0_Oa0_Oa0_Oa0_Oa0\\OjQU2"}, "label": "television with a raptor on the screen"}]} {"id": 461494, "image": "COCO_train2014_000000461494.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white truck\"."}], "task": "refering", "answer_template": "The \"a white truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 226], "bbox": [0.72315625, 0.3214375, 0.999046875, 0.5609375], "iscrowd": 0, "area": 13032.081700000004, "rle": {"size": [480, 640], "counts": "mUi6;e>:F:F9G:F:F:E4M0000001O00000000001O00000000001O00000000001O00000O101O00000000001O00000000001O00000000001O000000000O2O00001O001O1O001O010O1O001O2N3M3M3ZDlL[;\\30001O000000010O000000001O0001O0POP1000000000010O0000000N2G901O000000000000001O0 in this image.", "objects": [{"patches": [131, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 226], "bbox": [0.72315625, 0.3214375, 0.999046875, 0.5609375], "iscrowd": 0, "area": 13032.081700000004, "rle": {"size": [480, 640], "counts": "mUi6;e>:F:F9G:F:F:E4M0000001O00000000001O00000000001O00000000001O00000O101O00000000001O00000000001O00000000001O000000000O2O00001O001O1O001O010O1O001O2N3M3M3ZDlL[;\\30001O000000010O000000001O0001O0POP1000000000010O0000000N2G901O000000000000001O0 in this image.", "objects": [{"patches": [200, 201, 221, 222, 223, 224, 225, 244, 245, 246, 247, 267, 268, 269, 270, 290, 291, 292, 293, 313, 314, 315, 316, 336, 337, 338], "bbox": [0.6042500000000001, 0.5832860520094563, 0.8002343750000002, 0.9890780141843971], "iscrowd": 0, "area": 12132.495899999998, "rle": {"size": [423, 640], "counts": "WTP53R=3M4lCHU;:fDJ[;8_DJb;:XDHh;;UDDa2Ja6d0jFCa2Me6l0XIWOh6j0TIZOl6f0PI]OQ7b0lHBS7?iHDX7;eHG\\7:`HFd79YHFk7;QHES8;iGHW88fGKZ86bGN^81_G2a8O[G4e8LXG8g8IUG:l8FPG>o8BnFa0R9@jFc0V9h1N10001N10001N10001N10001N1^OVLgGj3Y8YLdGg3\\8[LaGf3`8YL]Gj3c8=00000000000000K50001O000000000O10000O100O101O0O10000O103L7J5K0O100O1000VMSH:m7FVHE2XNh7S2ZH@4XNa7Y2^H\\O5WN]7\\2cHXOQ8h0SHSOn7m0UHoNl7Q1XHkNg7V1]HeNd7Z1`HbNa7^1cH]N^7c1eHZN[7f1iHUNX7k1lHPNU7o1oHmMS7R2QIjMo6V2TIhMk6X2R2O10001O001K5K4L5K5K5K5K4L4L4K5L4L4L4M3N2O1N2O1N2O1ORdd1"}, "label": "a wicker chair standing on the edge of the counter across from the cardenza"}]} {"id": 363190, "image": "COCO_train2014_000000363190.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair in between two other chairs at the counter\"."}], "task": "refering", "answer_template": "The \"the chair in between two other chairs at the counter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [204, 225, 226, 227, 247, 248, 249, 250, 270, 271, 272, 273, 293, 294, 295, 296, 316, 317, 318, 319], "bbox": [0.726328125, 0.5956973995271867, 0.91109375, 0.9225768321513003], "iscrowd": 0, "area": 10405.219000000005, "rle": {"size": [423, 640], "counts": "SYQ61V=0fW10ZhN000000010O00000mN1nDOn:8oDIm:=PECl:e0PE[Om:k0PEUOf:W1YEhNc:]1\\EdNa:_1^EaN_:d1_E\\NW:o1hEQNS:V2kEjMP:\\2oEeMl9a2RF_Mj9h2TFWMh9P3UFPMg9W3VFiLi9\\3O0O1O10001O00000000000000000000000000000O1D in this image.", "objects": [{"patches": [204, 225, 226, 227, 247, 248, 249, 250, 270, 271, 272, 273, 293, 294, 295, 296, 316, 317, 318, 319], "bbox": [0.726328125, 0.5956973995271867, 0.91109375, 0.9225768321513003], "iscrowd": 0, "area": 10405.219000000005, "rle": {"size": [423, 640], "counts": "SYQ61V=0fW10ZhN000000010O00000mN1nDOn:8oDIm:=PECl:e0PE[Om:k0PEUOf:W1YEhNc:]1\\EdNa:_1^EaN_:d1_E\\NW:o1hEQNS:V2kEjMP:\\2oEeMl9a2RF_Mj9h2TFWMh9P3UFPMg9W3VFiLi9\\3O0O1O10001O00000000000000000000000000000O1D in this image.", "objects": [{"patches": [320, 321, 340, 341, 343, 344, 358, 359, 360, 361, 362, 363, 364, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.587109375, 0.7728541666666667, 1.0, 1.0], "iscrowd": 0, "area": 10894.909399999999, "rle": {"size": [480, 640], "counts": "ce`51n>2O2N2N2N1O2N2N2N1O9G in this image.", "objects": [{"patches": [320, 321, 340, 341, 343, 344, 358, 359, 360, 361, 362, 363, 364, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.587109375, 0.7728541666666667, 1.0, 1.0], "iscrowd": 0, "area": 10894.909399999999, "rle": {"size": [480, 640], "counts": "ce`51n>2O2N2N2N1O2N2N2N1O9G in this image.", "objects": [{"patches": [260, 283, 286, 306, 307, 308, 309, 310, 329, 330, 331, 332, 333, 351, 352, 353, 354, 355, 356, 374, 375, 376, 377, 378], "bbox": [0.26275, 0.6962083333333333, 0.495015625, 0.9972708333333333], "iscrowd": 0, "area": 8334.02635, "rle": {"size": [480, 640], "counts": "hV_24j>101N2N2O1N2N2N2O1N2N2N3M2N2N2M3O1N2O1N2O1N3N1N2O1O1O1\\MnNZGR1`8WO\\Gj0a8ZO^Gf0b8ZO]Gg0d8YOZGi0e8XOZGh0f8XOYGi0g8XOXGh0h8lNfECa1a1d9^N]Fa1c9`N\\F`1e9_N[Fa1f9_NYFa1g9`NXF`1i9_NWFa1i9`NTFc1l9\\NhE4jNb1_;ZNbES2_:nM]EU2d:jMYEY2g:hMTE\\2m:dMoD_2Q;aMlDb2U;^MgDe2Y;[MeDg2[;ZMbDh2^;YM_Di2a;WM]Dk2d;8000000001VN^DDR<=XDTOl;l0]DgNg;Z1o010O01O008H8I2M1O2N100O2N1O2N01O000000001O000000001O000O1O1L4H7HdC[No;1]CZ1j0hNf;OcCS1i0QOa;LiCn0h0_OV;DUDg0h0FQ;C[Da0f0Nn:A^DP in this image.", "objects": [{"patches": [260, 283, 286, 306, 307, 308, 309, 310, 329, 330, 331, 332, 333, 351, 352, 353, 354, 355, 356, 374, 375, 376, 377, 378], "bbox": [0.26275, 0.6962083333333333, 0.495015625, 0.9972708333333333], "iscrowd": 0, "area": 8334.02635, "rle": {"size": [480, 640], "counts": "hV_24j>101N2N2O1N2N2N2O1N2N2N3M2N2N2M3O1N2O1N2O1N3N1N2O1O1O1\\MnNZGR1`8WO\\Gj0a8ZO^Gf0b8ZO]Gg0d8YOZGi0e8XOZGh0f8XOYGi0g8XOXGh0h8lNfECa1a1d9^N]Fa1c9`N\\F`1e9_N[Fa1f9_NYFa1g9`NXF`1i9_NWFa1i9`NTFc1l9\\NhE4jNb1_;ZNbES2_:nM]EU2d:jMYEY2g:hMTE\\2m:dMoD_2Q;aMlDb2U;^MgDe2Y;[MeDg2[;ZMbDh2^;YM_Di2a;WM]Dk2d;8000000001VN^DDR<=XDTOl;l0]DgNg;Z1o010O01O008H8I2M1O2N100O2N1O2N01O000000001O000000001O000O1O1L4H7HdC[No;1]CZ1j0hNf;OcCS1i0QOa;LiCn0h0_OV;DUDg0h0FQ;C[Da0f0Nn:A^DP in this image.", "objects": [{"patches": [183, 197, 198, 199, 228, 229, 250, 251, 252, 266, 273, 289, 296, 312, 313, 318, 319], "bbox": [0.573375, 0.5101058201058201, 1.0, 0.9887566137566137], "iscrowd": 0, "area": 4661.9366000000055, "rle": {"size": [378, 640], "counts": "_fW4;_;0000O1000O100000000NFdD:\\;20OmFO_61bI?m5ATJo0[5QOfJZ1n4fNSKY1m4gNTKX1l4hNUKW1k4iNVKV1j4jNWKU1i4kN^Kn0b4ROmK?S4A[L1e3OS4O10000000001O0000000000000000000O1000000000000000000000000000000O100000000000000000000000000O001O1O001O1O001O00neT12kYkN6N2N2L4M3M3K5K5K5K5K5K5K5J6K5K5K5K5K5K5J6M30L4J6K5K5J5L5J6K5J6K5J6K4K6K5J6K5M3000O1000O10000000O1000O10000000O10O1000000000O01000nMBhH>S7HlH8T7HlH8T7HlH7V7IhH8X7HhH8X7HhH8X7HhH8X7IgH7Z7HfH7[7IeH7[7IdH8\\7IbH8^7HaH9`7F^Hh7BVH`0j7@UHa0l7^ORHd0n7]OoGe0Q8[OmGg0S8YOkGh0V8YOgGi0Z8VOeGk0[8UOcG`0d2"}, "label": "a brown color wooden frame for the sofa"}]} {"id": 535234, "image": "COCO_train2014_000000535234.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black jacket and baseball cap\"."}], "task": "refering", "answer_template": "The \"a man wearing a black jacket and baseball cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 86, 87, 88, 103, 104, 105, 106, 107, 121, 122, 123, 124, 125, 139, 140, 141, 142, 157, 158, 159, 160, 175, 176, 177, 178, 192, 193, 195, 196, 210, 211, 213, 214, 215], "bbox": [0.69148, 0.2702801120448179, 0.97934, 0.912156862745098], "iscrowd": 0, "area": 16615.462449999995, "rle": {"size": [357, 500], "counts": "fmh39i:4M2K5I7M3N2N2N2M3M3M2N3M3M3M3M3]M\\NaKh1Y4`NcKb1Y4dNbK_1[4eNbK]1[4hNaKZ1\\4jN`KY1]4lN_KV1_4nN]KT1a4nN]KP1f4QOUKP1n4QOkJR1W5QOeJn0^5SO^Jn0d5SOXJm0l5SOoIo0S6SOgIn0]6RO_In0d6g13N1N3M2O1N3M2O2N1O2O0O2O010O010O0100O100O10001O0O10001O2M3K5L3E in this image.", "objects": [{"patches": [69, 86, 87, 88, 103, 104, 105, 106, 107, 121, 122, 123, 124, 125, 139, 140, 141, 142, 157, 158, 159, 160, 175, 176, 177, 178, 192, 193, 195, 196, 210, 211, 213, 214, 215], "bbox": [0.69148, 0.2702801120448179, 0.97934, 0.912156862745098], "iscrowd": 0, "area": 16615.462449999995, "rle": {"size": [357, 500], "counts": "fmh39i:4M2K5I7M3N2N2N2M3M3M2N3M3M3M3M3]M\\NaKh1Y4`NcKb1Y4dNbK_1[4eNbK]1[4hNaKZ1\\4jN`KY1]4lN_KV1_4nN]KT1a4nN]KP1f4QOUKP1n4QOkJR1W5QOeJn0^5SO^Jn0d5SOXJm0l5SOoIo0S6SOgIn0]6RO_In0d6g13N1N3M2O1N3M2O2N1O2O0O2O010O010O0100O100O10001O0O10001O2M3K5L3E in this image.", "objects": [{"patches": [56, 57, 58, 59, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 109, 110, 111, 112, 127, 128, 129, 130, 144, 145, 146, 147, 148, 162, 163, 165, 180, 183], "bbox": [0.01124, 0.2269747899159664, 0.45246000000000003, 0.8382072829131653], "iscrowd": 0, "area": 13650.504199999998, "rle": {"size": [357, 500], "counts": "j[23P;4M2M4M3L4L3N3L1O01O00O2E:F:F;O00001O01O000O1N3M2E;1O1O100O001O100000000O1000000000000000O1000000001N101O001O1O0RO^NZHc1e7`NSHe1l7^NmGh1R8ZNhGj1X8b0N101YOYMoHg2R7ZMlHg2T7ZMjHf2V7[MhHg2X7ZMfHg2Z7ZMdHf2\\7\\MbHe2^7\\M`Hd2a7]M]Hd2b7^M\\Hb2e7^MYHd2g7<10O0100O010OlJfL_2e3RM_Ln2S4\\LQLc3P4YLTLg3l3TLZLk3e3RL_Ln3a3nKcLR4]3jKgLV4Y3eKlL[4T3`KQM`4o2ZKWMe4j2TK^Mk4^400000000000001N100001O0UOl0iNV1iNW1]Od0N101N1O101N1O2N101NKTNSGk1m8YNPGg1o87O101O0O101O0O101O00001O10O0100O10O0100000O1hNYG2g8M[G2e8L^G2c8M`G1a8MaG2_8McG2]8MdG2]8MdG3\\8LfG3[8KfG5[8IfG6[8HgG8Z8FgG:Z8DgG in this image.", "objects": [{"patches": [56, 57, 58, 59, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 109, 110, 111, 112, 127, 128, 129, 130, 144, 145, 146, 147, 148, 162, 163, 165, 180, 183], "bbox": [0.01124, 0.2269747899159664, 0.45246000000000003, 0.8382072829131653], "iscrowd": 0, "area": 13650.504199999998, "rle": {"size": [357, 500], "counts": "j[23P;4M2M4M3L4L3N3L1O01O00O2E:F:F;O00001O01O000O1N3M2E;1O1O100O001O100000000O1000000000000000O1000000001N101O001O1O0RO^NZHc1e7`NSHe1l7^NmGh1R8ZNhGj1X8b0N101YOYMoHg2R7ZMlHg2T7ZMjHf2V7[MhHg2X7ZMfHg2Z7ZMdHf2\\7\\MbHe2^7\\M`Hd2a7]M]Hd2b7^M\\Hb2e7^MYHd2g7<10O0100O010OlJfL_2e3RM_Ln2S4\\LQLc3P4YLTLg3l3TLZLk3e3RL_Ln3a3nKcLR4]3jKgLV4Y3eKlL[4T3`KQM`4o2ZKWMe4j2TK^Mk4^400000000000001N100001O0UOl0iNV1iNW1]Od0N101N1O101N1O2N101NKTNSGk1m8YNPGg1o87O101O0O101O0O101O00001O10O0100O10O0100000O1hNYG2g8M[G2e8L^G2c8M`G1a8MaG2_8McG2]8MdG2]8MdG3\\8LfG3[8KfG5[8IfG6[8HgG8Z8FgG:Z8DgG in this image.", "objects": [{"patches": [98, 99, 100, 116, 117, 118, 134, 135, 136, 152, 153, 154, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 206, 207], "bbox": [0.29936, 0.4259103641456583, 0.60194, 0.8720728291316527], "iscrowd": 0, "area": 10986.74745, "rle": {"size": [357, 500], "counts": "Ybd13R;3M1O001N2O001O1O001O1O001O000O2O00001O001O00001O00000O10001O000000000000000O101O0O100O100O100O100O1O100O100O100O100O100O100O10000O100O1eLlNeLT1Z3nNeLR1Z3QOcLQ1[3QOdLo0[3TObLm0]3UObLl0\\3WObLi0^3XO`Lj0_3XO_Lh0a3YO]Lh0c3XO]Li0b3XO]Lh0c3XO\\Lj0d3UO[Ll0e3UOZLk0f3UOQJEb1X1\\4SOoIJb1T1_4SOmILb1R1a4QOlIOa1Q1c4POjI2a1o0e4POhI3b1n0f4JeJbNAe1j5J_JQO[OV1V6JZJQ1g5oNTJU1l5lNoIW1S6hNiI\\1W6eNeI]1\\6f10O100O100O100O100O100O100O100O100O100O100O2O000O100000001O0O10001O000O101O00000O2M2N2N3M7I6I8I7I7I5K6Ib1_N?AWTU2"}, "label": "a baseball player sliding into a base"}]} {"id": 535234, "image": "COCO_train2014_000000535234.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baseball player stealing base , wearing the number 12 sliding into 2nd base\"."}], "task": "refering", "answer_template": "The \"baseball player stealing base , wearing the number 12 sliding into 2nd base\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 116, 117, 118, 134, 135, 136, 152, 153, 154, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 206, 207], "bbox": [0.29936, 0.4259103641456583, 0.60194, 0.8720728291316527], "iscrowd": 0, "area": 10986.74745, "rle": {"size": [357, 500], "counts": "Ybd13R;3M1O001N2O001O1O001O1O001O000O2O00001O001O00001O00000O10001O000000000000000O101O0O100O100O100O100O1O100O100O100O100O100O100O10000O100O1eLlNeLT1Z3nNeLR1Z3QOcLQ1[3QOdLo0[3TObLm0]3UObLl0\\3WObLi0^3XO`Lj0_3XO_Lh0a3YO]Lh0c3XO]Li0b3XO]Lh0c3XO\\Lj0d3UO[Ll0e3UOZLk0f3UOQJEb1X1\\4SOoIJb1T1_4SOmILb1R1a4QOlIOa1Q1c4POjI2a1o0e4POhI3b1n0f4JeJbNAe1j5J_JQO[OV1V6JZJQ1g5oNTJU1l5lNoIW1S6hNiI\\1W6eNeI]1\\6f10O100O100O100O100O100O100O100O100O100O100O2O000O100000001O0O10001O000O101O00000O2M2N2N3M7I6I8I7I7I5K6Ib1_N?AWTU2"}, "label": "baseball player stealing base , wearing the number 12 sliding into 2nd base"}]} {"id": 445127, "image": "COCO_train2014_000000445127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"monitor closer to the mouse\"."}], "task": "refering", "answer_template": "The \"monitor closer to the mouse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 41, 42, 43, 44, 59, 60, 61, 62, 63, 64, 65, 66, 67, 82, 83, 84, 85, 86, 87, 88, 89, 90, 105, 106, 107, 108, 109, 110, 111, 112, 113, 128, 129, 130, 131, 132, 133, 134, 135, 151, 152, 153, 154, 155, 156, 157, 158], "bbox": [0.57640625, 0.092125, 0.9353906249999999, 0.39325], "iscrowd": 0, "area": 30233.0714, "rle": {"size": [480, 640], "counts": "bT]56]>=C=B>C=C>B=C=C=C=C=M30000000000000000000000000000001O0000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000001O00000000000000000000000000O1H8I7H8H8H8I7H8H8I7H8H8H8I8G8H8I7Hgdc0"}, "label": "monitor closer to the mouse"}]} {"id": 445127, "image": "COCO_train2014_000000445127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the computer monitor by the corner of the desk\"."}], "task": "refering", "answer_template": "The \"the computer monitor by the corner of the desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 41, 42, 43, 44, 59, 60, 61, 62, 63, 64, 65, 66, 67, 82, 83, 84, 85, 86, 87, 88, 89, 90, 105, 106, 107, 108, 109, 110, 111, 112, 113, 128, 129, 130, 131, 132, 133, 134, 135, 151, 152, 153, 154, 155, 156, 157, 158], "bbox": [0.57640625, 0.092125, 0.9353906249999999, 0.39325], "iscrowd": 0, "area": 30233.0714, "rle": {"size": [480, 640], "counts": "bT]56]>=C=B>C=C>B=C=C=C=C=M30000000000000000000000000000001O0000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000001O00000000000000000000000000O1H8I7H8H8H8I7H8H8I7H8H8H8I8G8H8I7Hgdc0"}, "label": "the computer monitor by the corner of the desk"}]} {"id": 445127, "image": "COCO_train2014_000000445127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a monitor to the left of another monitor\"."}], "task": "refering", "answer_template": "The \"a monitor to the left of another monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 50, 51, 52, 53, 54, 55, 56, 57, 73, 74, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 189, 190, 191], "bbox": [0.200625, 0.09660416666666666, 0.5138906249999999, 0.5066666666666667], "iscrowd": 0, "area": 31723.838799999998, "rle": {"size": [480, 640], "counts": "QRl12n>`0@`0@a0_O`0@a0_O`0@`0_Ob0_O?A000000000000000O2O0000000000000000001N101O002N1O1O1O8H;E3M0O010000O1000000O1000O010000O1000000O10000O01000O10000O1000000O10O10O10000O1000000O10O10O10000O10000O100000O010000O10000O1000000O01000O10000O1000000O01000O10000O10000O1000O10O10000O10000O100000O010000O10000O100000O010000O10000O10000O10O1000O10000O10000O1000O10O10000O10000O1000O10O10000O10000O10000O0100000O10000O100^Ob0_Oa00O2aga4"}, "label": "a monitor to the left of another monitor"}]} {"id": 445127, "image": "COCO_train2014_000000445127.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a monitor with two speakers\"."}], "task": "refering", "answer_template": "The \"a monitor with two speakers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 50, 51, 52, 53, 54, 55, 56, 57, 73, 74, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 125, 126, 142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 189, 190, 191], "bbox": [0.200625, 0.09660416666666666, 0.5138906249999999, 0.5066666666666667], "iscrowd": 0, "area": 31723.838799999998, "rle": {"size": [480, 640], "counts": "QRl12n>`0@`0@a0_O`0@a0_O`0@`0_Ob0_O?A000000000000000O2O0000000000000000001N101O002N1O1O1O8H;E3M0O010000O1000000O1000O010000O1000000O10000O01000O10000O1000000O10O10O10000O1000000O10O10O10000O10000O100000O010000O10000O1000000O01000O10000O1000000O01000O10000O10000O1000O10O10000O10000O100000O010000O10000O100000O010000O10000O10000O10O1000O10000O10000O1000O10O10000O10000O1000O10O10000O10000O10000O0100000O10000O100^Ob0_Oa00O2aga4"}, "label": "a monitor with two speakers"}]} {"id": 436941, "image": "COCO_train2014_000000436941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white zebra standing sideways facing another zebra\"."}], "task": "refering", "answer_template": "The \"a black and white zebra standing sideways facing another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 219, 237, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.27137500000000003, 0.39775175644028105, 1.0, 0.9887587822014052], "iscrowd": 0, "area": 53986.4196, "rle": {"size": [427, 640], "counts": "kiX23X=1N3M2O2M3M2N3N1N3M3M2O2M2N3M2O1N2N2N2O0O2N2O1N2N2N010O1O00100O00100O001O010O1O010O1O00100O0010O01O1O010O1O010O001O10O01WORNTFo1j9WNQFi1n9]NnEc1P:dNjE]1T:iNgEW1X:o0O0O4L4L4L45K5L0O0001O00001L3N2M4M2M3N3L3N2N3N1O1O1O2N1O1O1O1O000O2O001N101N101N101N101N101N101N101N101N101N101N101N101N1O2O0O2O0O2O001O00001O001O1O001O1O001O1O001O1O00O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2M2O2N1O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O0000001O00000000000000000000000001O000000000000000000000001O00000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000001O00000000001O000000001O00000000001O000000001O0001O000001O000000001O00000000001O000000001O00000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O00000000001O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O00^E"}, "label": "a black and white zebra standing sideways facing another zebra"}]} {"id": 436941, "image": "COCO_train2014_000000436941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra with it ' s nose on the other zebra\"."}], "task": "refering", "answer_template": "The \"a zebra with it ' s nose on the other zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 219, 237, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.27137500000000003, 0.39775175644028105, 1.0, 0.9887587822014052], "iscrowd": 0, "area": 53986.4196, "rle": {"size": [427, 640], "counts": "kiX23X=1N3M2O2M3M2N3N1N3M3M2O2M2N3M2O1N2N2N2O0O2N2O1N2N2N010O1O00100O00100O001O010O1O010O1O00100O0010O01O1O010O1O010O001O10O01WORNTFo1j9WNQFi1n9]NnEc1P:dNjE]1T:iNgEW1X:o0O0O4L4L4L45K5L0O0001O00001L3N2M4M2M3N3L3N2N3N1O1O1O2N1O1O1O1O000O2O001N101N101N101N101N101N101N101N101N101N101N101N101N1O2O0O2O0O2O001O00001O001O1O001O1O001O1O001O1O00O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2M2O2N1O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O0000001O00000000000000000000000001O000000000000000000000001O00000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000001O00000000001O000000001O00000000001O000000001O0001O000001O000000001O00000000001O000000001O00000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O00000000001O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O00^E"}, "label": "a zebra with it ' s nose on the other zebra"}]} {"id": 436941, "image": "COCO_train2014_000000436941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small zebra being nibbled on by a larger zebra\"."}], "task": "refering", "answer_template": "The \"a small zebra being nibbled on by a larger zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 188, 189, 211, 212, 213, 234, 235, 236, 256, 257, 258, 259, 260, 278, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 306, 323, 324, 325, 326, 327, 328, 329], "bbox": [0.06896875, 0.5142388758782201, 0.34524999999999995, 1.0], "iscrowd": 0, "area": 20374.555099999994, "rle": {"size": [427, 640], "counts": "Yhb01W=3M3M3M3M3M3L4M3M3M3M3M3M3O1N2O1O1N2O1O1N2O1O1N2O001N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1XNTLoIm3P6[LiIf3U6aLdIa3Z6fL^I]3`6jLWIZ3g6mLQIV3m6RMiHR3U7c1N2N200001O001O001O001O010O001O1O2N3M2N2N2N3N1N2N00O100O10000000001O000000000002N2N2N2N2O0O2N2N2N2N2N2N2N2N3M2O1N2N2N2N2N3M2N2N2N2O1fLgF^2`9ZMbFe2i9nLYFR3Q:O001O1O001O1O010O1O001O1E;M3N3L3M4M2M4L3N3L3M4M2M4]Oa\\^5"}, "label": "a small zebra being nibbled on by a larger zebra"}]} {"id": 436941, "image": "COCO_train2014_000000436941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra that is standing to the left of another zebra\"."}], "task": "refering", "answer_template": "The \"a zebra that is standing to the left of another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 188, 189, 211, 212, 213, 234, 235, 236, 256, 257, 258, 259, 260, 278, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 306, 323, 324, 325, 326, 327, 328, 329], "bbox": [0.06896875, 0.5142388758782201, 0.34524999999999995, 1.0], "iscrowd": 0, "area": 20374.555099999994, "rle": {"size": [427, 640], "counts": "Yhb01W=3M3M3M3M3M3L4M3M3M3M3M3M3O1N2O1O1N2O1O1N2O1O1N2O001N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1XNTLoIm3P6[LiIf3U6aLdIa3Z6fL^I]3`6jLWIZ3g6mLQIV3m6RMiHR3U7c1N2N200001O001O001O001O010O001O1O2N3M2N2N2N3N1N2N00O100O10000000001O000000000002N2N2N2N2O0O2N2N2N2N2N2N2N2N3M2O1N2N2N2N2N3M2N2N2N2O1fLgF^2`9ZMbFe2i9nLYFR3Q:O001O1O001O1O010O1O001O1E;M3N3L3M4M2M4L3N3L3M4M2M4]Oa\\^5"}, "label": "a zebra that is standing to the left of another zebra"}]} {"id": 68305, "image": "COCO_train2014_000000068305.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white sink\"."}], "task": "refering", "answer_template": "The \"the white sink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [214, 215, 216, 217, 218, 219, 236, 237, 238, 239, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 266, 283, 284, 285, 286, 287], "bbox": [0.277640625, 0.5268125, 0.5788437500000001, 0.7432916666666666], "iscrowd": 0, "area": 14285.490549999999, "rle": {"size": [480, 640], "counts": "kgc23k>4M3M3L5L3M3L4M3L3N1O2M3N2M3N2N2N2N2O1N2O000O1O2N100O2N1O1O1O2N1O101N1O1O2O0O100O2O0O101N10000O2O0O101N10000000000O100000001O0000000O100000000000000O10000000000000000J[MPDf2m;8O1O100000001O00000O100000000000O10O100000O10000000O10O100RMVDf2j;ZMVDf2j;YMXDf2h;ZMXDe2j;YMWDg2j;XMVDh2j;WMWDi2j;VMVDj2P in this image.", "objects": [{"patches": [214, 215, 216, 217, 218, 219, 236, 237, 238, 239, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 266, 283, 284, 285, 286, 287], "bbox": [0.277640625, 0.5268125, 0.5788437500000001, 0.7432916666666666], "iscrowd": 0, "area": 14285.490549999999, "rle": {"size": [480, 640], "counts": "kgc23k>4M3M3L5L3M3L4M3L3N1O2M3N2M3N2N2N2N2O1N2O000O1O2N100O2N1O1O1O2N1O101N1O1O2O0O100O2O0O101N10000O2O0O101N10000000000O100000001O0000000O100000000000000O10000000000000000J[MPDf2m;8O1O100000001O00000O100000000000O10O100000O10000000O10O100RMVDf2j;ZMVDf2j;YMXDf2h;ZMXDe2j;YMWDg2j;XMVDh2j;WMWDi2j;VMVDj2P in this image.", "objects": [{"patches": [106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 295], "bbox": [0.592, 0.24802083333333333, 0.844734375, 0.7623333333333334], "iscrowd": 0, "area": 35935.09835, "rle": {"size": [480, 640], "counts": "nla5k0e=W1hNW1jNW1iNg0SOf0ZOf0E;01O0000000O100000000000001O0O10000000000000000O101O0000000000000O100000001O0000000O100000000000001O0O10000000000000000O10001O00000000000O10000000000000000O100000001O0000000O10000000000000000O1000000000001O000O10000000000000000O10000000000000001N10000000000000000O10000002N6J7I7I6J7I6K6I6J7I8H=C in this image.", "objects": [{"patches": [106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 295], "bbox": [0.592, 0.24802083333333333, 0.844734375, 0.7623333333333334], "iscrowd": 0, "area": 35935.09835, "rle": {"size": [480, 640], "counts": "nla5k0e=W1hNW1jNW1iNg0SOf0ZOf0E;01O0000000O100000000000001O0O10000000000000000O101O0000000000000O100000001O0000000O100000000000001O0O10000000000000000O10001O00000000000O10000000000000000O100000001O0000000O10000000000000000O1000000000001O000O10000000000000000O10000000000000001N10000000000000000O10000002N6J7I7I6J7I6K6I6J7I8H=C in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 237, 238, 239, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 266, 283, 284, 285, 286, 287], "bbox": [0.3036875, 0.30222916666666666, 0.5973124999999999, 0.7205833333333334], "iscrowd": 0, "area": 31973.792799999996, "rle": {"size": [480, 640], "counts": "dXk21o>0O1O1O1O1B>]Oc0\\Od0\\Od0]Oc0\\Od0\\Od0]Oc0\\Od0^Ob000000000000001O000000000000000000000000000000000000000000001O000O1000000000000000000000000000000000000000001O000000000000000000000000000000000000000000001O000O1000000000000000000000000000000000000000001O0001O1O1O1O1N101O1N2O1N2O1O1N2O1N2O0O2O1O1N2O1N2O1O1N2O1N101N2O1O1N2O1N2O1N2O1O0O2O1N2O1O1N2O1N2O1N2O001N2O1N2O1O1N2O1N2O0O2O1O1N2O1N2B>TOl0UOk0TO[hh3"}, "label": "a two decker red bus is parked between two other buses"}]} {"id": 125658, "image": "COCO_train2014_000000125658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red bus in between 2 other red buses\"."}], "task": "refering", "answer_template": "The \"a red bus in between 2 other red buses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 237, 238, 239, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 266, 283, 284, 285, 286, 287], "bbox": [0.3036875, 0.30222916666666666, 0.5973124999999999, 0.7205833333333334], "iscrowd": 0, "area": 31973.792799999996, "rle": {"size": [480, 640], "counts": "dXk21o>0O1O1O1O1B>]Oc0\\Od0\\Od0]Oc0\\Od0\\Od0]Oc0\\Od0^Ob000000000000001O000000000000000000000000000000000000000000001O000O1000000000000000000000000000000000000000001O000000000000000000000000000000000000000000001O000O1000000000000000000000000000000000000000001O0001O1O1O1O1N101O1N2O1N2O1O1N2O1N2O0O2O1O1N2O1N2O1O1N2O1N101N2O1O1N2O1N2O1N2O1O0O2O1N2O1O1N2O1N2O1N2O001N2O1N2O1O1N2O1N2O0O2O1O1N2O1N2B>TOl0UOk0TO[hh3"}, "label": "a red bus in between 2 other red buses"}]} {"id": 125658, "image": "COCO_train2014_000000125658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red double decker bus with large plain white sign with no writing\"."}], "task": "refering", "answer_template": "The \"red double decker bus with large plain white sign with no writing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 119, 120, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260], "bbox": [0.083765625, 0.32727083333333334, 0.32337499999999997, 0.6909166666666666], "iscrowd": 0, "area": 23877.637399999996, "rle": {"size": [480, 640], "counts": "hci0=T>`0@?A?A?A?A?A?B>K5L4L4L4L4K5L4O2O0000000000000O100000000000001O000000000000000O1000000000001O0000000000000000000O10000000001O0000000000000000000O100000001O00000000000000000000000O10001O000000000001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001M2L5J5^Oc0]Ob0]Od0]Ob0^Oc0]Ob0^OjiZ6"}, "label": "red double decker bus with large plain white sign with no writing"}]} {"id": 125658, "image": "COCO_train2014_000000125658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of red bus parked\"."}], "task": "refering", "answer_template": "The \"the back of red bus parked\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 119, 120, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260], "bbox": [0.083765625, 0.32727083333333334, 0.32337499999999997, 0.6909166666666666], "iscrowd": 0, "area": 23877.637399999996, "rle": {"size": [480, 640], "counts": "hci0=T>`0@?A?A?A?A?A?B>K5L4L4L4L4K5L4O2O0000000000000O100000000000001O000000000000000O1000000000001O0000000000000000000O10000000001O0000000000000000000O100000001O00000000000000000000000O10001O000000000001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001M2L5J5^Oc0]Ob0]Od0]Ob0^Oc0]Ob0^OjiZ6"}, "label": "the back of red bus parked"}]} {"id": 183022, "image": "COCO_train2014_000000183022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"space shuttle being taxied by a jumbo jet\"."}], "task": "refering", "answer_template": "The \"space shuttle being taxied by a jumbo jet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 96, 97, 98, 99, 100, 101, 107, 120, 121, 122, 123, 124, 125, 126, 128, 129, 130, 131, 145, 146, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 176, 177, 197, 198, 199, 200], "bbox": [0.19465625, 0.25392857142857145, 0.743265625, 0.5955], "iscrowd": 0, "area": 16669.495300000002, "rle": {"size": [420, 640], "counts": "Y\\c18l<9G1N101O001O001O1N101O001O1O0O2O001O001O1O0O2O001O1O001O0O2O001O1O001O0O2O001O001O0O2O001O001O001O001O001N101O001O001O001O001N101O001O001O001O1O001N101O001O001O001O001O0010O010O010O010O010O010O010O010O010O010O01O010O010O010O010O010O010O010O010O010O010O00100O010O010O010O01O001O001N101O001O001O00001O001O001O001O001O001N101O001O001O001O000YOeDE\\;:dDF\\;;dDD];;cDE];;dDD];;dDD\\;bDB_;=bDB^;>bDB_;=bDB^;>cDA^;>bDB^;>cDA^;>bDB^;>cDA];?cDA^;?bD@^;`0bD@_;?bD@^;`0cD_O^;`0bD@^;`0cD_O^;`0bD_O_;a0bD^O_;a0aD_O_;a0bD^O_;b0`D^O`;b0aD]O`;b0`D^O`;b0aD]O`;b0a001O001O0010O0001O001O001O001O010O000000O10000O100O2O0O100O100O100O100O2O0O100UEYOY9g0gFZOW9g0iFYOW9g0hF[OV9g0iFYOW9g0iFZOU9g0jFZOV9f0jF[OT9f0lFZOT9g0jF[OT9f0lFZOT9f0lFZOS9g0lF[OS9f0lFZOS9g0mFZOR9f0mF[OR9f0nF[OQ9e0oF[OQ9f0mF\\OQ9d0PG\\OP9d0PG\\Oo8e0PG]Oo8c0QG]On8c0SG^Ol8b0SG_Ol8b0TG_Ok8a0UG_Oj8a0VG@j8`0VG@i8a0WG_Oi8a0VG@i8`0XG@h8`0XG@g8a0XG@h8`0XG@h8?YG@g8a0XG@h8?YGAf8`0ZG@f8?[GAd8`0[GAe8>\\GBc8?]GAc8>]GCb8>^GBb8=^GDa8=_GCa8<_GDa8<`GD`8;`GF_8:bGF^89bGH^87cGI\\87dGJ\\85eGKZ85fGLZ83gGMX83hGNX81iGOV81jGOW80jG0U80kG1U8NlG2U8LkG5U8JlG6U8HkG9U8FlG:U8DlGU8@lG`0T8_OlGb0U8\\OlGc0U8\\OkGe0V8YOkGg0V8WOjGj0V8UOkGk0V8SOjGn0V8QOkGo0V8POkGo0U8POmGo0T8oNnGP1R8oNPHP1Q8nNRHP1R:N2O1O1O1O2N1O1O2N1O2N1O1N3N1O2N1O1O2N1O2N1O1OmPS2"}, "label": "space shuttle being taxied by a jumbo jet"}]} {"id": 183022, "image": "COCO_train2014_000000183022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"ship on top of airplane\"."}], "task": "refering", "answer_template": "The \"ship on top of airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 96, 97, 98, 99, 100, 101, 107, 120, 121, 122, 123, 124, 125, 126, 128, 129, 130, 131, 145, 146, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 176, 177, 197, 198, 199, 200], "bbox": [0.19465625, 0.25392857142857145, 0.743265625, 0.5955], "iscrowd": 0, "area": 16669.495300000002, "rle": {"size": [420, 640], "counts": "Y\\c18l<9G1N101O001O001O1N101O001O1O0O2O001O001O1O0O2O001O1O001O0O2O001O1O001O0O2O001O001O0O2O001O001O001O001O001N101O001O001O001O001N101O001O001O001O1O001N101O001O001O001O001O0010O010O010O010O010O010O010O010O010O010O01O010O010O010O010O010O010O010O010O010O010O00100O010O010O010O01O001O001N101O001O001O00001O001O001O001O001O001N101O001O001O001O000YOeDE\\;:dDF\\;;dDD];;cDE];;dDD];;dDD\\;bDB_;=bDB^;>bDB_;=bDB^;>cDA^;>bDB^;>cDA^;>bDB^;>cDA];?cDA^;?bD@^;`0bD@_;?bD@^;`0cD_O^;`0bD@^;`0cD_O^;`0bD_O_;a0bD^O_;a0aD_O_;a0bD^O_;b0`D^O`;b0aD]O`;b0`D^O`;b0aD]O`;b0a001O001O0010O0001O001O001O001O010O000000O10000O100O2O0O100O100O100O100O2O0O100UEYOY9g0gFZOW9g0iFYOW9g0hF[OV9g0iFYOW9g0iFZOU9g0jFZOV9f0jF[OT9f0lFZOT9g0jF[OT9f0lFZOT9f0lFZOS9g0lF[OS9f0lFZOS9g0mFZOR9f0mF[OR9f0nF[OQ9e0oF[OQ9f0mF\\OQ9d0PG\\OP9d0PG\\Oo8e0PG]Oo8c0QG]On8c0SG^Ol8b0SG_Ol8b0TG_Ok8a0UG_Oj8a0VG@j8`0VG@i8a0WG_Oi8a0VG@i8`0XG@h8`0XG@g8a0XG@h8`0XG@h8?YG@g8a0XG@h8?YGAf8`0ZG@f8?[GAd8`0[GAe8>\\GBc8?]GAc8>]GCb8>^GBb8=^GDa8=_GCa8<_GDa8<`GD`8;`GF_8:bGF^89bGH^87cGI\\87dGJ\\85eGKZ85fGLZ83gGMX83hGNX81iGOV81jGOW80jG0U80kG1U8NlG2U8LkG5U8JlG6U8HkG9U8FlG:U8DlGU8@lG`0T8_OlGb0U8\\OlGc0U8\\OkGe0V8YOkGg0V8WOjGj0V8UOkGk0V8SOjGn0V8QOkGo0V8POkGo0U8POmGo0T8oNnGP1R8oNPHP1Q8nNRHP1R:N2O1O1O1O2N1O1O2N1O2N1O1N3N1O2N1O1O2N1O2N1O1OmPS2"}, "label": "ship on top of airplane"}]} {"id": 346876, "image": "COCO_train2014_000000346876.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the broccoli closest to the camera furthest from the fork\"."}], "task": "refering", "answer_template": "The \"the broccoli closest to the camera furthest from the fork\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 236, 237, 238, 239, 260, 261, 262, 283, 284], "bbox": [0.280921875, 0.5631024930747922, 0.47212499999999996, 0.9497229916897507], "iscrowd": 0, "area": 10608.428500000002, "rle": {"size": [361, 640], "counts": "Rfo1k0\\:6J6K5J5K6K5J5K6J6K5J3M3N3L3M3M3O2M2O1N2N2O2M2O1N2O2M2N2O1O1O2N1O1O101N1O1O1O1O2N1O1O1O01N2O1O1O001O1O1O0O2O=C1O1O001N2O1O1O001M2N2N3M2N3F94L3N3L1O01N101O00001O000000001N10000000001cNcGK^84dGG`88cGC`8 in this image.", "objects": [{"patches": [176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 292, 293, 294, 295, 296, 316, 317, 318, 319, 340, 341], "bbox": [0.6674218749999999, 0.43652083333333336, 0.9016874999999999, 0.8410208333333333], "iscrowd": 0, "area": 21958.71290000001, "rle": {"size": [480, 640], "counts": "j[X65K5K5K5K5K5L4K5K5L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4M3L4L4L4L4L4L4L4L1O001O00001O00001O00010O001O00001O001O00001O001O00001O0010O0001O00001O001O00001O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O001O001O0010O0001O001O00001O0000N2N2O1N2N2O1N2O2O000O10000O100O1O1O100O1O100O1O100O1O2O0O1O1O100O1O1O1Oj0WOl0SOm0SOm0SOm0SOoUm0"}, "label": "a bus with the number 17 in its window"}]} {"id": 51965, "image": "COCO_train2014_000000051965.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a passing bus with the number 17 on it\"."}], "task": "refering", "answer_template": "The \"a passing bus with the number 17 on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 250, 269, 270, 271, 272, 273, 292, 293, 294, 295, 296, 316, 317, 318, 319, 340, 341], "bbox": [0.6674218749999999, 0.43652083333333336, 0.9016874999999999, 0.8410208333333333], "iscrowd": 0, "area": 21958.71290000001, "rle": {"size": [480, 640], "counts": "j[X65K5K5K5K5K5L4K5K5L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4M3L4L4L4L4L4L4L4L1O001O00001O00001O00010O001O00001O001O00001O001O00001O0010O0001O00001O001O00001O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O001O001O0010O0001O001O00001O0000N2N2O1N2N2O1N2O2O000O10000O100O1O1O100O1O100O1O100O1O2O0O1O1O100O1O1O1Oj0WOl0SOm0SOm0SOm0SOoUm0"}, "label": "a passing bus with the number 17 on it"}]} {"id": 51965, "image": "COCO_train2014_000000051965.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bus reading nis training vehicle\"."}], "task": "refering", "answer_template": "The \"bus reading nis training vehicle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 126, 127, 128, 129, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 351, 352, 353], "bbox": [0.146625, 0.28695833333333337, 0.72134375, 0.9004375000000001], "iscrowd": 0, "area": 86288.47970000001, "rle": {"size": [480, 640], "counts": "fZ\\1c0n=k0UOj0VOk0ZOf0C=C=C;E3M3M3M3M3M3M3M3M3M3M3M2N1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O100O2N1O1O1O101N100O100O2O0O2O1N101N2O0O2O1N101N2O0O2O1N101N2O0O2O1N101N2O0O2O0O100O100O100O101N10000O101N2O0O2O1N101O0001O00001N10001O001O000O2O00001O00001N101O00001O000O101O00000000000O100000001O000O1000000000000O101O00000000000O100000001O000O1000000000000O10001O00000000000O100000001O00000O010000O1000000O01000O1000000O10O1000O10000O100000O010000O1000000O01000O1000000O10O10O1000000O1000O01000000O10000O10O1000O10000O1000000O101O0O1000000O101O0O1000000O10001N1000000O10001N10000001O1O1O10O01O1O1O1O1O00100O1O1O001O1O100O1O001O1O1O1O010O1O1O2N6J5K6I7I6J7I6J7I7I6J6J5K5K5K5K5K5K5K5K5K5L4K5K6J5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K6J5K5K5K]ac2"}, "label": "bus reading nis training vehicle"}]} {"id": 51965, "image": "COCO_train2014_000000051965.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"nis training vehicle bus number 2565\"."}], "task": "refering", "answer_template": "The \"nis training vehicle bus number 2565\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 126, 127, 128, 129, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 351, 352, 353], "bbox": [0.146625, 0.28695833333333337, 0.72134375, 0.9004375000000001], "iscrowd": 0, "area": 86288.47970000001, "rle": {"size": [480, 640], "counts": "fZ\\1c0n=k0UOj0VOk0ZOf0C=C=C;E3M3M3M3M3M3M3M3M3M3M3M2N1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O100O2N1O1O1O101N100O100O2O0O2O1N101N2O0O2O1N101N2O0O2O1N101N2O0O2O1N101N2O0O2O0O100O100O100O101N10000O101N2O0O2O1N101O0001O00001N10001O001O000O2O00001O00001N101O00001O000O101O00000000000O100000001O000O1000000000000O101O00000000000O100000001O000O1000000000000O10001O00000000000O100000001O00000O010000O1000000O01000O1000000O10O1000O10000O100000O010000O1000000O01000O1000000O10O10O1000000O1000O01000000O10000O10O1000O10000O1000000O101O0O1000000O101O0O1000000O10001N1000000O10001N10000001O1O1O10O01O1O1O1O1O00100O1O1O001O1O100O1O001O1O1O1O010O1O1O2N6J5K6I7I6J7I6J7I7I6J6J5K5K5K5K5K5K5K5K5K5L4K5K6J5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K5K6J5K5K5K]ac2"}, "label": "nis training vehicle bus number 2565"}]} {"id": 396042, "image": "COCO_train2014_000000396042.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glass window pane on the side of a cat sitting in bottom of a flower pot\"."}], "task": "refering", "answer_template": "The \"glass window pane on the side of a cat sitting in bottom of a flower pot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 50, 51, 63, 64, 76, 77, 88, 89, 90, 101, 102, 103, 114, 115, 116], "bbox": [0.7972800000000001, 0.1247, 1.0, 0.49416], "iscrowd": 0, "area": 7769.548050000002, "rle": {"size": [500, 375], "counts": "nUb41a?3N2M2dAIY=:dBIZ=9cBJZ=8dBKY=8dBJ[=8bBK[=7bBM[=6bBM\\=5aBN]=4`BN`=2^B1a=0\\B3c=P1O001O1N2O1L3J7J6N2nMYMTGh2l8YMRGg2n8\\MoFd2R9]MlFd2T9]MjFb2W9`MgF`2Z9bMcFj1aN\\Nn:K_Fh1hNWNl:2[Ff1mNQNk: in this image.", "objects": [{"patches": [38, 50, 51, 63, 64, 76, 77, 88, 89, 90, 101, 102, 103, 114, 115, 116], "bbox": [0.7972800000000001, 0.1247, 1.0, 0.49416], "iscrowd": 0, "area": 7769.548050000002, "rle": {"size": [500, 375], "counts": "nUb41a?3N2M2dAIY=:dBIZ=9cBJZ=8dBKY=8dBJ[=8bBK[=7bBM[=6bBM\\=5aBN]=4`BN`=2^B1a=0\\B3c=P1O001O1N2O1L3J7J6N2nMYMTGh2l8YMRGg2n8\\MoFd2R9]MlFd2T9]MjFb2W9`MgF`2Z9bMcFj1aN\\Nn:K_Fh1hNWNl:2[Ff1mNQNk: in this image.", "objects": [{"patches": [96, 97, 114, 115, 116, 132, 133, 134, 150, 151, 152, 169, 170, 187, 188], "bbox": [0.34793999999999997, 0.426006006006006, 0.46888, 0.8923723723723724], "iscrowd": 0, "area": 5719.601049999999, "rle": {"size": [333, 500], "counts": "ghh15U:8H9F8I5L4L4K3N2O00100AcNnGk1b7[NYHj1`7]N\\He1`7\\N`Hg1^7XNaHl1\\7i0dImLR5Z3eJhLZ5]3`JfL_5]3]JdLb5`3YJcLf5a3TJaLk5R4O3L3N3N00O1O1O2N100O2hNbJ]Me5o1jJeMe5W2]JdMe5c2VJXMn5n2e07Id0]O4K0001O001O2N2N1N3N2N2N2jMWId0k6VO`I7EkNm6i0SKVOo4d0VK[Ol4>[K_Oh4:^KEd45m\\f2"}, "label": "a men in brown jacket"}]} {"id": 101140, "image": "COCO_train2014_000000101140.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in orange coat standing beside a truck with two men standing in front of him\"."}], "task": "refering", "answer_template": "The \"person in orange coat standing beside a truck with two men standing in front of him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 114, 115, 116, 132, 133, 134, 150, 151, 152, 169, 170, 187, 188], "bbox": [0.34793999999999997, 0.426006006006006, 0.46888, 0.8923723723723724], "iscrowd": 0, "area": 5719.601049999999, "rle": {"size": [333, 500], "counts": "ghh15U:8H9F8I5L4L4K3N2O00100AcNnGk1b7[NYHj1`7]N\\He1`7\\N`Hg1^7XNaHl1\\7i0dImLR5Z3eJhLZ5]3`JfL_5]3]JdLb5`3YJcLf5a3TJaLk5R4O3L3N3N00O1O1O2N100O2hNbJ]Me5o1jJeMe5W2]JdMe5c2VJXMn5n2e07Id0]O4K0001O001O2N2N1N3N2N2N2jMWId0k6VO`I7EkNm6i0SKVOo4d0VK[Ol4>[K_Oh4:^KEd45m\\f2"}, "label": "person in orange coat standing beside a truck with two men standing in front of him"}]} {"id": 101140, "image": "COCO_train2014_000000101140.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow truck behind the truck loaded with bags\"."}], "task": "refering", "answer_template": "The \"the yellow truck behind the truck loaded with bags\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 35, 50, 51, 52, 53, 67, 68, 69, 70, 71, 85, 86, 87, 88, 89, 107, 125, 143, 161], "bbox": [0.74, 0.14813813813813814, 1.0, 0.7357957957957958], "iscrowd": 0, "area": 9922.508149999998, "rle": {"size": [333, 500], "counts": "m]h36Q:0O100O10000O10000000000000000000000000000000000000000000000000000000000000000001O000000001O00000000001O000000003M2N3M2N2N3M2N3M2N2N2N3M2N2aKnL_1T3ZNQMf1Q3nMXMS2k2TMkMl2W2\\L_Nd3U40O100000000O1000000O100000000O10_N"}, "label": "the yellow truck behind the truck loaded with bags"}]} {"id": 101140, "image": "COCO_train2014_000000101140.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hidden truck\"."}], "task": "refering", "answer_template": "The \"a hidden truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 35, 50, 51, 52, 53, 67, 68, 69, 70, 71, 85, 86, 87, 88, 89, 107, 125, 143, 161], "bbox": [0.74, 0.14813813813813814, 1.0, 0.7357957957957958], "iscrowd": 0, "area": 9922.508149999998, "rle": {"size": [333, 500], "counts": "m]h36Q:0O100O10000O10000000000000000000000000000000000000000000000000000000000000000001O000000001O00000000001O000000003M2N3M2N2N3M2N3M2N2N2N3M2N2aKnL_1T3ZNQMf1Q3nMXMS2k2TMkMl2W2\\L_Nd3U40O100000000O1000000O100000000O10_N"}, "label": "a hidden truck"}]} {"id": 330516, "image": "COCO_train2014_000000330516.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown horse bent over eating grass\"."}], "task": "refering", "answer_template": "The \"a brown horse bent over eating grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 167, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 241], "bbox": [0.270609375, 0.3892916666666667, 0.5066875, 0.6388750000000001], "iscrowd": 0, "area": 7241.742550000001, "rle": {"size": [480, 640], "counts": "cha21o>1O1O1O1O1O00001O00001O00001O0O101O00001fAFl=:QBIn=e00O10000000O10O100000O1000O100000O10O100000O1000O100000O0100O100O010O01O001O001O102M4L5K4L5K5K5aCaNU;c1`DkNY;Y1[DoNe;7XDf0h;\\OZD`0e;B]D@0Id;h0^D^O7_O[;T1aDZO5@Z;W1cDWOS in this image.", "objects": [{"patches": [146, 147, 148, 167, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 241], "bbox": [0.270609375, 0.3892916666666667, 0.5066875, 0.6388750000000001], "iscrowd": 0, "area": 7241.742550000001, "rle": {"size": [480, 640], "counts": "cha21o>1O1O1O1O1O00001O00001O00001O0O101O00001fAFl=:QBIn=e00O10000000O10O100000O1000O100000O10O100000O1000O100000O0100O100O010O01O001O001O102M4L5K4L5K5K5aCaNU;c1`DkNY;Y1[DoNe;7XDf0h;\\OZD`0e;B]D@0Id;h0^D^O7_O[;T1aDZO5@Z;W1cDWOS in this image.", "objects": [{"patches": [125, 126, 147, 148, 149, 169, 170, 171, 172, 173, 192, 193, 194, 195, 196, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 262, 263, 264, 285, 286, 287, 308, 309, 310, 331, 332, 333], "bbox": [0.366546875, 0.3130625, 0.543921875, 0.871625], "iscrowd": 0, "area": 18781.317450000006, "rle": {"size": [480, 640], "counts": "X\\^39d> in this image.", "objects": [{"patches": [125, 126, 147, 148, 149, 169, 170, 171, 172, 173, 192, 193, 194, 195, 196, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 262, 263, 264, 285, 286, 287, 308, 309, 310, 331, 332, 333], "bbox": [0.366546875, 0.3130625, 0.543921875, 0.871625], "iscrowd": 0, "area": 18781.317450000006, "rle": {"size": [480, 640], "counts": "X\\^39d> in this image.", "objects": [{"patches": [130, 131, 132, 153, 154, 155, 156, 176, 177, 178, 179, 199, 200, 201, 202, 222, 223, 224, 225, 245, 246, 247, 248, 268, 269, 270, 271, 291, 292, 293, 294, 314, 315, 316, 337, 338, 339], "bbox": [0.658984375, 0.298875, 0.8072968749999999, 0.8539375], "iscrowd": 0, "area": 16659.101349999994, "rle": {"size": [480, 640], "counts": "WVV6:f>1O000WNJhD7S;1iDO^8EfHd0hNG_8HdHg0mNA\\8KbHi0QO\\OZ8O`Hk0TOVOZ82]Hm0XOROW84\\HP1]OkNS8:[HQ1_OfNS8=YHR1CaNQ8a0VHT1H\\Nm7b0YHX1GWNm7a0ZH^1HQN_6CnIm0He1JkM]6IjIk0Jh1LeM]6NgIg0Ll1O_M[64dIc0NQ21XM[69`Ia01S21TM[6>\\I>6U2OQM]6`0YI<9X2MnLa6a0SI<;[2KlLf6P2]IY1GkLl6k1VI`2j6k20000O10000000000O100000000O2O0000001O000010O1O100K4L5K5L4K5K7H:G9F;E:G4O2O1O001N101O1O0O2O001O1N101O1XL_Fo1b9nMbFP2_9nMeFn1]9oMgFo1Z9oMjFn1W9PNkFo1V9oMmFn1W9nMkFQ2W9mMkFQ2W9lMlFR2V9kMlFS2W9eMQGY2P;KY1gN?@`0A?AnYi1"}, "label": "woman standing at sink in grey dress"}]} {"id": 174876, "image": "COCO_train2014_000000174876.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a woman in a grey dress\"."}], "task": "refering", "answer_template": "The \"this is a woman in a grey dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 153, 154, 155, 156, 176, 177, 178, 179, 199, 200, 201, 202, 222, 223, 224, 225, 245, 246, 247, 248, 268, 269, 270, 271, 291, 292, 293, 294, 314, 315, 316, 337, 338, 339], "bbox": [0.658984375, 0.298875, 0.8072968749999999, 0.8539375], "iscrowd": 0, "area": 16659.101349999994, "rle": {"size": [480, 640], "counts": "WVV6:f>1O000WNJhD7S;1iDO^8EfHd0hNG_8HdHg0mNA\\8KbHi0QO\\OZ8O`Hk0TOVOZ82]Hm0XOROW84\\HP1]OkNS8:[HQ1_OfNS8=YHR1CaNQ8a0VHT1H\\Nm7b0YHX1GWNm7a0ZH^1HQN_6CnIm0He1JkM]6IjIk0Jh1LeM]6NgIg0Ll1O_M[64dIc0NQ21XM[69`Ia01S21TM[6>\\I>6U2OQM]6`0YI<9X2MnLa6a0SI<;[2KlLf6P2]IY1GkLl6k1VI`2j6k20000O10000000000O100000000O2O0000001O000010O1O100K4L5K5L4K5K7H:G9F;E:G4O2O1O001N101O1O0O2O001O1N101O1XL_Fo1b9nMbFP2_9nMeFn1]9oMgFo1Z9oMjFn1W9PNkFo1V9oMmFn1W9nMkFQ2W9mMkFQ2W9lMlFR2V9kMlFS2W9eMQGY2P;KY1gN?@`0A?AnYi1"}, "label": "this is a woman in a grey dress"}]} {"id": 92957, "image": "COCO_train2014_000000092957.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cup on the right in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the cup on the right in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 68, 91, 113, 114, 136, 137, 159, 160, 182, 183, 206, 252, 275, 296, 297, 298, 320, 321], "bbox": [0.89125, 0.08968749999999999, 1.0, 0.8183958333333333], "iscrowd": 0, "area": 8076.607999999993, "rle": {"size": [480, 640], "counts": "mo[82m>2O2M2O1N3N1N2N2O2M2O001O1O001O0O2O001O1O001O001jIWORNj0l1BjM?R2OcM1Z2=[MDa2i0UMXOh2U1mLlNo2Y1oLhNm2\\1RMeNj2_1TMcNi2`1VMaNf2c1YM]Nd2g1[MZNa2i1_MXN^2k1aMVN[2n1dMRNY2R2eMPNW2T2hMmMT2W2kMiMS2Z2lMgMP2]2oMcMn1`2RN`Mk1d2TN]Mi1f2VNZMg1j2WNWMf1m2YNSMd1Q3[NPMb1S3]NmL`1W3^NjL_1Z3]NiL`1Z3\\NjLb1Y3ZNkLb1Y3ZNjLc1Z3YNiLd1[3XNhLe1\\3VNhLh1[3TNhLi1\\3SNhLk1Y3RNjLn1W3mMmLR2U3jMoLU2R3gMQMY2S2kNfMS1ZOR3BZ7mNXD"}, "label": "the cup on the right in the right hand picture"}]} {"id": 92957, "image": "COCO_train2014_000000092957.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass of wine that we can only see a little bit off on the side of the picture\"."}], "task": "refering", "answer_template": "The \"a glass of wine that we can only see a little bit off on the side of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 68, 91, 113, 114, 136, 137, 159, 160, 182, 183, 206, 252, 275, 296, 297, 298, 320, 321], "bbox": [0.89125, 0.08968749999999999, 1.0, 0.8183958333333333], "iscrowd": 0, "area": 8076.607999999993, "rle": {"size": [480, 640], "counts": "mo[82m>2O2M2O1N3N1N2N2O2M2O001O1O001O0O2O001O1O001O001jIWORNj0l1BjM?R2OcM1Z2=[MDa2i0UMXOh2U1mLlNo2Y1oLhNm2\\1RMeNj2_1TMcNi2`1VMaNf2c1YM]Nd2g1[MZNa2i1_MXN^2k1aMVN[2n1dMRNY2R2eMPNW2T2hMmMT2W2kMiMS2Z2lMgMP2]2oMcMn1`2RN`Mk1d2TN]Mi1f2VNZMg1j2WNWMf1m2YNSMd1Q3[NPMb1S3]NmL`1W3^NjL_1Z3]NiL`1Z3\\NjLb1Y3ZNkLb1Y3ZNjLc1Z3YNiLd1[3XNhLe1\\3VNhLh1[3TNhLi1\\3SNhLk1Y3RNjLn1W3mMmLR2U3jMoLU2R3gMQMY2S2kNfMS1ZOR3BZ7mNXD"}, "label": "a glass of wine that we can only see a little bit off on the side of the picture"}]} {"id": 92957, "image": "COCO_train2014_000000092957.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wineglass with some red wine in it and a tag with the number 64 directly behind it\"."}], "task": "refering", "answer_template": "The \"a wineglass with some red wine in it and a tag with the number 64 directly behind it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 35, 36, 37, 38, 39, 40, 41, 58, 59, 60, 61, 62, 63, 64, 81, 82, 83, 84, 85, 86, 87, 104, 105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 150, 151, 152, 153, 154, 155, 156, 173, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 221, 222, 244, 245, 267, 268, 290, 312, 313, 314, 334, 335, 336, 337, 338, 357, 358, 359, 360, 361, 381, 382, 383, 384], "bbox": [0.52753125, 0.0, 0.8190937500000001, 0.9730416666666667], "iscrowd": 0, "area": 51271.94090000002, "rle": {"size": [480, 640], "counts": "obn4;\\>=VI[OhNR1R1SOaNZ1[1iNZNd1`1aNTNk1g1YNnMT2m1PNgM\\2T2iMaMc2Y2aM_Mi2\\2[M\\Mm2_2WMYMS3a2PMYMX3c2iLWMa3c2bLWMg3e2YLUMP4g2RLSMU4k2lKoLZ4P3fKjLa4U3`KeLf4Y3[KaLk4_3VK[LP5d3PKVLW5i3jJQL\\5m3eJnKa5Q4`JnKa5Q4`JnKa5P4`JPL`5P4aJoK`5P4`JPLa5o3`JPLa5n3`JRLa5m3`JRLa5m3_JSLb5l3_JSLa5l3aJSL`5l3`JTLa5k3`JTLa5k3_JULb5i3`JVLa5i3_JWLb5h3_JWLb5h3^JYLa5f3`JZLa5e3_J[Lb5d3^J\\Lc5c3]J]Ld5a3^J^Lc5a3]J_Lc5a3]J_Ld5`3\\J`Ld5_3]JaLd5^3\\JbLd5^3]JaLd5^3\\JbLd5]3]JcLd5\\3\\JdLd5\\3\\JdLe5[3\\JdLd5[3]JeLd5Z3\\JfLd5X3^JhLc5R3bJnL^5n2fJRM[5i2jJVMV5e2oJ[MR5`2RK`Mn4[2WKeMj4V2ZKjMf4R2_KmMb4a1oK_NQ4n0bLRO_3;SMEm2HfM8a801O\\MCeF>R9JnF6j83UGMc8;]GEZ8d0fG\\OR8m0nGROj7V1VHjNa7_1_HaN]7d1bH\\NZ7h1fHXNU7n1kHQNQ7S2oHmMm6W2SIiMi6\\2VIdMf6`2YIaMb6e2]I[M_6i2aIWM[6n2cISMY6Q3gIoLU6V3jIjLR6Z3nIfLm5`3QJaLn5`3RJ`Lm5b3RJ^Ln5b3QJ_Ln5c3QJ]Ln5d3RJ\\Ln5e3PJ\\Lo5e3QJ[Lo5f3PJZLo5g3PJZLP6g3oIYLP6h3PJXLP6i3nIXLQ6i3oIWLP6k3oIVLP6j3oIWLP6k3oIULQ6k3oIULP6m3nITLR6l3nITLQ6n3nIRLQ6o3nIRLR6o3mIQLR6P4nIPLR6Q4lIPLS6Q4mIoKS6R4lInKS6S4lInKS6U4kIkKU6W4iIiKV6Z4gIgKY6[4dIfK[6]4`IfK`6\\4[IgKd6\\4WIgKh6\\4SIgKm6Z4oHiKP7Z4kHiKU7Y4fHjKY7Y4cHiK]7m5O100O1O100O100O101N1O100O100O100O1O100O100O100O1O1L4L4L4L4L4L4L4L4L4L4L4L4L4L4L:F;E in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 35, 36, 37, 38, 39, 40, 41, 58, 59, 60, 61, 62, 63, 64, 81, 82, 83, 84, 85, 86, 87, 104, 105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 150, 151, 152, 153, 154, 155, 156, 173, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 221, 222, 244, 245, 267, 268, 290, 312, 313, 314, 334, 335, 336, 337, 338, 357, 358, 359, 360, 361, 381, 382, 383, 384], "bbox": [0.52753125, 0.0, 0.8190937500000001, 0.9730416666666667], "iscrowd": 0, "area": 51271.94090000002, "rle": {"size": [480, 640], "counts": "obn4;\\>=VI[OhNR1R1SOaNZ1[1iNZNd1`1aNTNk1g1YNnMT2m1PNgM\\2T2iMaMc2Y2aM_Mi2\\2[M\\Mm2_2WMYMS3a2PMYMX3c2iLWMa3c2bLWMg3e2YLUMP4g2RLSMU4k2lKoLZ4P3fKjLa4U3`KeLf4Y3[KaLk4_3VK[LP5d3PKVLW5i3jJQL\\5m3eJnKa5Q4`JnKa5Q4`JnKa5P4`JPL`5P4aJoK`5P4`JPLa5o3`JPLa5n3`JRLa5m3`JRLa5m3_JSLb5l3_JSLa5l3aJSL`5l3`JTLa5k3`JTLa5k3_JULb5i3`JVLa5i3_JWLb5h3_JWLb5h3^JYLa5f3`JZLa5e3_J[Lb5d3^J\\Lc5c3]J]Ld5a3^J^Lc5a3]J_Lc5a3]J_Ld5`3\\J`Ld5_3]JaLd5^3\\JbLd5^3]JaLd5^3\\JbLd5]3]JcLd5\\3\\JdLd5\\3\\JdLe5[3\\JdLd5[3]JeLd5Z3\\JfLd5X3^JhLc5R3bJnL^5n2fJRM[5i2jJVMV5e2oJ[MR5`2RK`Mn4[2WKeMj4V2ZKjMf4R2_KmMb4a1oK_NQ4n0bLRO_3;SMEm2HfM8a801O\\MCeF>R9JnF6j83UGMc8;]GEZ8d0fG\\OR8m0nGROj7V1VHjNa7_1_HaN]7d1bH\\NZ7h1fHXNU7n1kHQNQ7S2oHmMm6W2SIiMi6\\2VIdMf6`2YIaMb6e2]I[M_6i2aIWM[6n2cISMY6Q3gIoLU6V3jIjLR6Z3nIfLm5`3QJaLn5`3RJ`Lm5b3RJ^Ln5b3QJ_Ln5c3QJ]Ln5d3RJ\\Ln5e3PJ\\Lo5e3QJ[Lo5f3PJZLo5g3PJZLP6g3oIYLP6h3PJXLP6i3nIXLQ6i3oIWLP6k3oIVLP6j3oIWLP6k3oIULQ6k3oIULP6m3nITLR6l3nITLQ6n3nIRLQ6o3nIRLR6o3mIQLR6P4nIPLR6Q4lIPLS6Q4mIoKS6R4lInKS6S4lInKS6U4kIkKU6W4iIiKV6Z4gIgKY6[4dIfK[6]4`IfK`6\\4[IgKd6\\4WIgKh6\\4SIgKm6Z4oHiKP7Z4kHiKU7Y4fHjKY7Y4cHiK]7m5O100O1O100O100O101N1O100O100O100O1O100O100O100O1O1L4L4L4L4L4L4L4L4L4L4L4L4L4L4L:F;E in this image.", "objects": [{"patches": [10, 11, 12, 13, 27, 28, 29, 30, 31, 32, 45, 46, 47, 48, 49, 50, 63, 64, 65, 66, 67, 68, 81, 82, 83, 84, 85, 86, 100, 101, 102, 103, 104], "bbox": [0.54102, 0.05168, 0.82078, 0.44944], "iscrowd": 0, "area": 19141.29245, "rle": {"size": [375, 500], "counts": "nXS3g1o99G9G8H9G8H9G9G8H9G8H1O1O1O001O1000000000000000000000O1000000000000000000000000000000001O0O1000000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O000000000000O1000000000000000O100000000001O1O001O1O1O001O1O1O001O1O1O001O1O1O001O7I>ZOe0A?A`0@?A`0A>AajP1"}, "label": "red fruit pur \u00e9 e in a pink container"}]} {"id": 174892, "image": "COCO_train2014_000000174892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"apple sauce\"."}], "task": "refering", "answer_template": "The \"apple sauce\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 27, 28, 29, 30, 31, 32, 45, 46, 47, 48, 49, 50, 63, 64, 65, 66, 67, 68, 81, 82, 83, 84, 85, 86, 100, 101, 102, 103, 104], "bbox": [0.54102, 0.05168, 0.82078, 0.44944], "iscrowd": 0, "area": 19141.29245, "rle": {"size": [375, 500], "counts": "nXS3g1o99G9G8H9G8H9G9G8H9G8H1O1O1O001O1000000000000000000000O1000000000000000000000000000000001O0O1000000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O000000000000O1000000000000000O100000000001O1O001O1O1O001O1O1O001O1O1O001O1O1O001O7I>ZOe0A?A`0@?A`0A>AajP1"}, "label": "apple sauce"}]} {"id": 174892, "image": "COCO_train2014_000000174892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of bananas , honey and blueberries\"."}], "task": "refering", "answer_template": "The \"a bowl of bananas , honey and blueberries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 23, 24, 25, 26, 27, 37, 38, 39, 40, 41, 42, 43, 44, 45, 55, 56, 57, 58, 59, 60, 61, 62, 63, 73, 74, 75, 76, 77, 78, 79, 80, 81, 91, 92, 93, 94, 95, 96, 97, 98, 99], "bbox": [0.07584, 0.07864, 0.54776, 0.4629333333333334], "iscrowd": 0, "area": 30490.627749999996, "rle": {"size": [375, 500], "counts": "RR>4];7H8H8H8H8H7I8H8H8I7H8H7I8J6N2N2N2N1O2N2M2O2N1O1O001O1O1O1O1O1O100O10000O100O100000000000000000000000000000000O1000O100000000000000000000000000000000000000000O1000000000000000000000O100000000000000000000000O1000000000000000000000000000000000000000O100000O10000000000000000000000000000000000000000000000O100000000000O1000000000000000000000000000000000O100000000000000000000000000000O1000000000000000O10000001O000000001O000000000O2O0O1O2O0O2N101N2N1O2M2O2M3N1N3M2O2Mo0ROP1oNR1oNZdb2"}, "label": "a bowl of bananas , honey and blueberries"}]} {"id": 174892, "image": "COCO_train2014_000000174892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue pot containing bananas , honey and blackcurrants\"."}], "task": "refering", "answer_template": "The \"a blue pot containing bananas , honey and blackcurrants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 23, 24, 25, 26, 27, 37, 38, 39, 40, 41, 42, 43, 44, 45, 55, 56, 57, 58, 59, 60, 61, 62, 63, 73, 74, 75, 76, 77, 78, 79, 80, 81, 91, 92, 93, 94, 95, 96, 97, 98, 99], "bbox": [0.07584, 0.07864, 0.54776, 0.4629333333333334], "iscrowd": 0, "area": 30490.627749999996, "rle": {"size": [375, 500], "counts": "RR>4];7H8H8H8H8H7I8H8H8I7H8H7I8J6N2N2N2N1O2N2M2O2N1O1O001O1O1O1O1O1O100O10000O100O100000000000000000000000000000000O1000O100000000000000000000000000000000000000000O1000000000000000000000O100000000000000000000000O1000000000000000000000000000000000000000O100000O10000000000000000000000000000000000000000000000O100000000000O1000000000000000000000000000000000O100000000000000000000000000000O1000000000000000O10000001O000000001O000000000O2O0O1O2O0O2N101N2N1O2M2O2M3N1N3M2O2Mo0ROP1oNR1oNZdb2"}, "label": "a blue pot containing bananas , honey and blackcurrants"}]} {"id": 174892, "image": "COCO_train2014_000000174892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"container with bananas and blueberries\"."}], "task": "refering", "answer_template": "The \"container with bananas and blueberries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 23, 24, 26, 27, 38, 39, 40, 41, 42, 43, 44, 45, 56, 57, 58, 59, 60, 61, 62, 63, 74, 75, 76, 77, 78, 79, 80, 81, 92, 93, 94, 95, 96, 97, 98, 99], "bbox": [0.11084000000000001, 0.10968000000000001, 0.5375, 0.44632], "iscrowd": 0, "area": 20495.916350000003, "rle": {"size": [375, 500], "counts": "UXd03c;d0\\O8G9H6I6E:M1N100TORNnGP2Q8QNmGP2S8QNlGo1T8RNjGo1U8SNjGm1V8TNiGm1V8TNiGl1W8TNhGm1W8TNiGl1W8TNiGl1W8SNjGm1U8TNkGl1U8TNkGl1U8SNlGm1S8TNmGl1Q8UNPHk1o7VNPHk1o7VNQHj1n7VNSHj1k7XNUHh1j7XNWHh1h7YNXHg1g7YNZHg1e7ZN[Hf1c7[N^He1b7[N\\Hg1c7ZN\\Hg1c7YN]Hh1c7XN\\Hi1c7XN[Hj1e7WNYHj1f7XNXHi1h7XNUHj1j7XNTHi1l7XNQHj1n7XNoGj1Q8n0O1O100O100O1000O2O1O001O1O001O1O001O001O1OO1O1O1O11O001O1O001O1O001O00O1O1O1O1O1O1O1O1O100O10000O10000O1000000O10000O10000O10000O11O2N2N2M3N1O2M2A=10]MjGe1V8[NjGd1V8\\NlGc1T8\\NmGd1R8]NnGc1R8\\NPHb1P8_NPHa1P8^NRHa1m7`NSH`1m7_NTH`1l7`NVH_1j7aNVH`1h7`NZH`1e7`N[Ha1c7^N`Hb1a7ZNbHf1^7XNcHi1^7SNeHm1c80O101O0O100O10000O10000O01O1O00O0O2N2YOkMjGW2T8mMhGU2V8oMeGT2Z8oMbGS2\\8QN`GQ2_8d0N2O0O2O1N2O100000000000O1000O10000000000000O100000O100000000O10O1000O01O100O010O1O100O105TN_G7f8E_G3i8H\\G0l8JfGCe83n1@V`d2"}, "label": "container with bananas and blueberries"}]} {"id": 174892, "image": "COCO_train2014_000000174892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of banana slices and blueberries\"."}], "task": "refering", "answer_template": "The \"a bowl of banana slices and blueberries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 23, 24, 26, 27, 38, 39, 40, 41, 42, 43, 44, 45, 56, 57, 58, 59, 60, 61, 62, 63, 74, 75, 76, 77, 78, 79, 80, 81, 92, 93, 94, 95, 96, 97, 98, 99], "bbox": [0.11084000000000001, 0.10968000000000001, 0.5375, 0.44632], "iscrowd": 0, "area": 20495.916350000003, "rle": {"size": [375, 500], "counts": "UXd03c;d0\\O8G9H6I6E:M1N100TORNnGP2Q8QNmGP2S8QNlGo1T8RNjGo1U8SNjGm1V8TNiGm1V8TNiGl1W8TNhGm1W8TNiGl1W8TNiGl1W8SNjGm1U8TNkGl1U8TNkGl1U8SNlGm1S8TNmGl1Q8UNPHk1o7VNPHk1o7VNQHj1n7VNSHj1k7XNUHh1j7XNWHh1h7YNXHg1g7YNZHg1e7ZN[Hf1c7[N^He1b7[N\\Hg1c7ZN\\Hg1c7YN]Hh1c7XN\\Hi1c7XN[Hj1e7WNYHj1f7XNXHi1h7XNUHj1j7XNTHi1l7XNQHj1n7XNoGj1Q8n0O1O100O100O1000O2O1O001O1O001O1O001O001O1OO1O1O1O11O001O1O001O1O001O00O1O1O1O1O1O1O1O1O100O10000O10000O1000000O10000O10000O10000O11O2N2N2M3N1O2M2A=10]MjGe1V8[NjGd1V8\\NlGc1T8\\NmGd1R8]NnGc1R8\\NPHb1P8_NPHa1P8^NRHa1m7`NSH`1m7_NTH`1l7`NVH_1j7aNVH`1h7`NZH`1e7`N[Ha1c7^N`Hb1a7ZNbHf1^7XNcHi1^7SNeHm1c80O101O0O100O10000O10000O01O1O00O0O2N2YOkMjGW2T8mMhGU2V8oMeGT2Z8oMbGS2\\8QN`GQ2_8d0N2O0O2O1N2O100000000000O1000O10000000000000O100000O100000000O10O1000O01O100O010O1O100O105TN_G7f8E_G3i8H\\G0l8JfGCe83n1@V`d2"}, "label": "a bowl of banana slices and blueberries"}]} {"id": 174892, "image": "COCO_train2014_000000174892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the slice of orange that is touching the bottom right corner of its container\"."}], "task": "refering", "answer_template": "The \"the slice of orange that is touching the bottom right corner of its container\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 173, 174, 175, 191, 192, 193, 194, 210, 211, 212], "bbox": [0.63352, 0.6141066666666667, 0.82588, 0.90352], "iscrowd": 0, "area": 6263.431350000001, "rle": {"size": [375, 500], "counts": "P\\d35`;3E;ROn0I8J5O1N2O1O2M2O2M101N101N2O0O2O1N2O0O2O1N2O2M2O2M2O2N1O1N2O1O100O1O1O1O0002N2N101N2N6Ja0_O3M2N3M2N1O1O001O1O1O00001O001N101O001O001N101O001O001N101O0O2O001N2O001N2O2N1N3M2N2N4L3M3L5K5J5Kjao0"}, "label": "the slice of orange that is touching the bottom right corner of its container"}]} {"id": 174892, "image": "COCO_train2014_000000174892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange slice closest to the camera\"."}], "task": "refering", "answer_template": "The \"the orange slice closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 173, 174, 175, 191, 192, 193, 194, 210, 211, 212], "bbox": [0.63352, 0.6141066666666667, 0.82588, 0.90352], "iscrowd": 0, "area": 6263.431350000001, "rle": {"size": [375, 500], "counts": "P\\d35`;3E;ROn0I8J5O1N2O1O2M2O2M101N101N2O0O2O1N2O0O2O1N2O2M2O2M2O2N1O1N2O1O100O1O1O1O0002N2N101N2N6Ja0_O3M2N3M2N1O1O001O1O1O00001O001N101O001O001N101O001O001N101O0O2O001N2O001N2O2N1N3M2N2N4L3M3L5K5J5Kjao0"}, "label": "the orange slice closest to the camera"}]} {"id": 174892, "image": "COCO_train2014_000000174892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"peeled nectarines sitting inside a green plastic container\"."}], "task": "refering", "answer_template": "The \"peeled nectarines sitting inside a green plastic container\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 135, 136, 137, 138, 139, 140, 153, 154, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213], "bbox": [0.5427000000000001, 0.43594666666666665, 0.84606, 0.92136], "iscrowd": 0, "area": 25353.8151, "rle": {"size": [375, 500], "counts": "`aS3>S8W3H9G8H8H8J6N2N2N3L3N2N2N2N2N3M2N2N10N3O00000000000000000000000000000000000000001O0000000000000000000000000001O000000000001O00000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O00000000000000000000000000000000000001O01O001O1O1O2M2N2O2M2N3N1N2N3M2N3M3M3Lb0_Ob0^Oa0^Oa0@?A`0_Oa0@`0@alk0"}, "label": "peeled nectarines sitting inside a green plastic container"}]} {"id": 174892, "image": "COCO_train2014_000000174892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green cup with oranges in it\"."}], "task": "refering", "answer_template": "The \"a green cup with oranges in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 135, 136, 137, 138, 139, 140, 153, 154, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213], "bbox": [0.5427000000000001, 0.43594666666666665, 0.84606, 0.92136], "iscrowd": 0, "area": 25353.8151, "rle": {"size": [375, 500], "counts": "`aS3>S8W3H9G8H8H8J6N2N2N3L3N2N2N2N2N3M2N2N10N3O00000000000000000000000000000000000000001O0000000000000000000000000001O000000000001O00000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O00000000000000000000000000000000000001O01O001O1O1O2M2N2O2M2N3N1N2N3M2N3M3M3Lb0_Ob0^Oa0^Oa0@?A`0_Oa0@`0@alk0"}, "label": "a green cup with oranges in it"}]} {"id": 174892, "image": "COCO_train2014_000000174892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"food in an orange container\"."}], "task": "refering", "answer_template": "The \"food in an orange container\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 98, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207], "bbox": [0.02022, 0.45618666666666663, 0.53932, 0.89888], "iscrowd": 0, "area": 40348.80695, "rle": {"size": [375, 500], "counts": "Vo36o:c0^Oc0]Ob0H8H9G7I5J6K6J7I7I7I7I7I7K5O10000O1000000O10000O2O0000000O10000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O0O100000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000001O0000000000000000000000000O10000000000000000000001O000000000000000000000000000000000000000000000000001O000000000000000000000000000000O1000001O00000O2O001O2N1O2K4M3L`1aNi2VMdmc2"}, "label": "food in an orange container"}]} {"id": 174892, "image": "COCO_train2014_000000174892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"food in an orange contaier\"."}], "task": "refering", "answer_template": "The \"food in an orange contaier\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 98, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207], "bbox": [0.02022, 0.45618666666666663, 0.53932, 0.89888], "iscrowd": 0, "area": 40348.80695, "rle": {"size": [375, 500], "counts": "Vo36o:c0^Oc0]Ob0H8H9G7I5J6K6J7I7I7I7I7I7K5O10000O1000000O10000O2O0000000O10000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O0O100000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000001O0000000000000000000000000O10000000000000000000001O000000000000000000000000000000000000000000000000001O000000000000000000000000000000O1000001O00000O2O001O2N1O2K4M3L`1aNi2VMdmc2"}, "label": "food in an orange contaier"}]} {"id": 486193, "image": "COCO_train2014_000000486193.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a male skier wearing a grey jacket and black trousers\"."}], "task": "refering", "answer_template": "The \"a male skier wearing a grey jacket and black trousers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 103, 104, 105, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 195, 196, 197, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265], "bbox": [0.408078125, 0.20540983606557375, 0.644890625, 0.7997658079625293], "iscrowd": 0, "area": 16395.969100000002, "rle": {"size": [427, 640], "counts": "SU]34U=3N1N2O2M2N3ChHBW7`0hH@W7b0hH^OW7d0hH\\OW7f0iHYOV76eGlNR1o0X75jGjNi0V1\\70mGiNf0W1_7OmGhNa0\\1b7DhGfN<3JT2R8POnHo0S7POmHQ1T7mNmHR1T7mNkHU1U7kNjHU1X7jNgHW1Y7iNeHX1\\7hNVHnNCZ2X8hNTHTO_OR2^8jNQHYO^Om1a8kNhGCE`1e8mN`GR2`8nM[GV2g8V10O101N1O2O0O2N101NdHlKd5S4[JQLd5n3YJVLg5h3XJ\\Lf5c3nIoKUOa0m6^3oITLPO`0Q7[3PJWLkN`0T7X3PJSMP6k2nI[M`5POeIc3h0cM_5POfI[3g0mMY5QOoIQ3c0UNZ5f2cJaMZ5^2dJgM[5Y2eJhMZ5Z2YJdKNS2h5[2WJSNg5i400O101O00XO]JhIc5W6`JgI`5Y6bJeI^5X6fJeI\\5[6fJZId5f6`001O1O1O1N3D]IlIf6n5a0I6J5^J[HT5W8I3D in this image.", "objects": [{"patches": [81, 82, 103, 104, 105, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 195, 196, 197, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265], "bbox": [0.408078125, 0.20540983606557375, 0.644890625, 0.7997658079625293], "iscrowd": 0, "area": 16395.969100000002, "rle": {"size": [427, 640], "counts": "SU]34U=3N1N2O2M2N3ChHBW7`0hH@W7b0hH^OW7d0hH\\OW7f0iHYOV76eGlNR1o0X75jGjNi0V1\\70mGiNf0W1_7OmGhNa0\\1b7DhGfN<3JT2R8POnHo0S7POmHQ1T7mNmHR1T7mNkHU1U7kNjHU1X7jNgHW1Y7iNeHX1\\7hNVHnNCZ2X8hNTHTO_OR2^8jNQHYO^Om1a8kNhGCE`1e8mN`GR2`8nM[GV2g8V10O101N1O2O0O2N101NdHlKd5S4[JQLd5n3YJVLg5h3XJ\\Lf5c3nIoKUOa0m6^3oITLPO`0Q7[3PJWLkN`0T7X3PJSMP6k2nI[M`5POeIc3h0cM_5POfI[3g0mMY5QOoIQ3c0UNZ5f2cJaMZ5^2dJgM[5Y2eJhMZ5Z2YJdKNS2h5[2WJSNg5i400O101O00XO]JhIc5W6`JgI`5Y6bJeI^5X6fJeI\\5[6fJZId5f6`001O1O1O1N3D]IlIf6n5a0I6J5^J[HT5W8I3D in this image.", "objects": [{"patches": [72, 73, 86, 87, 88, 89, 100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 130, 131, 132, 133, 134, 146, 147, 148, 149], "bbox": [0.703559718969555, 0.20635937499999998, 1.0, 0.42940625], "iscrowd": 0, "area": 14578.648300000003, "rle": {"size": [640, 427], "counts": "Zfk51fc0`0A>B?F6K4M4L3M4L3M3M4L3N2M1O2N2N1O2N2O0O2N2N1O2N1O2N2O0O2N2N1O2N2N1O2O0O2N2N1O2N1O2O001O001N101O001O001N101O001O001O0O2O001O00O10O1000000000O01000000000O10O10000000O10O10000001N101O001O1O001N101O1O001O0O2N1O2O1N1O2N101N2N1O2O0O2N1O2O1N1O2O0O2N1O2O1Nb0^OfJ"}, "label": "the face of a large coco tower in a city"}]} {"id": 92988, "image": "COCO_train2014_000000092988.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"face of clock on tower closest to camera\"."}], "task": "refering", "answer_template": "The \"face of clock on tower closest to camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 86, 87, 88, 89, 100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 130, 131, 132, 133, 134, 146, 147, 148, 149], "bbox": [0.703559718969555, 0.20635937499999998, 1.0, 0.42940625], "iscrowd": 0, "area": 14578.648300000003, "rle": {"size": [640, 427], "counts": "Zfk51fc0`0A>B?F6K4M4L3M4L3M3M4L3N2M1O2N2N1O2N2O0O2N2N1O2N1O2N2O0O2N2N1O2N2N1O2O0O2N2N1O2N1O2O001O001N101O001O001N101O001O001O0O2O001O00O10O1000000000O01000000000O10O10000000O10O10000001N101O001O1O001N101O1O001O0O2N1O2O1N1O2N101N2N1O2O0O2N1O2O1N1O2O0O2N1O2O1Nb0^OfJ"}, "label": "face of clock on tower closest to camera"}]} {"id": 199485, "image": "COCO_train2014_000000199485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant in the front of the line\"."}], "task": "refering", "answer_template": "The \"the elephant in the front of the line\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 58, 59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 124, 125, 126, 127, 128, 129, 130, 131, 132, 147, 148, 149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 178, 194, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 293, 310, 311, 312, 313, 314, 315, 316, 333, 334, 335, 336, 337, 338], "bbox": [0.42279687499999996, 0.11686182669789227, 0.7856249999999999, 0.986510538641686], "iscrowd": 0, "area": 63161.32310000001, "rle": {"size": [427, 640], "counts": "^Ua3:o<9F;E:G4K5L5J5K6K4K5K6K4K5L5J5K6M2M3N3M2N2N1O1O2N1O1O2O09G9H6IO2O00001O000fNUMSHl2j7lM`GT2_8]1N3M2N4L6J5K5K5K5K6J5K5K7Il0TOk0UOl0YKPG\\3j9F0O2N1O1O1O1O1O1O1O1O1O1O1N2O1N2N2O1N2O1N2O1N3N1N2N2O1N2O1N2O1N2O1N2O100000I7G8I8H8G9H7I8H8G9I6J7J6I7J6M210001O0O1000001O005K5K5K6J5K5K6K4K5mM_F2g9]OdFc0a9mNjFS1[9]NPGc1\\:0000000000001O00000000000000000000001O000000000000O1000000000000000000000000O100000000000000O100000000000000O1000000001O00001O00001O00001iNUNlFk1T9[NfFe1Z9bN_F_1`9hNYFX1g9nNSFR1m9UOlEl0S:V100O2O000jJbN_N_1_1gN\\NY1b1nNYNR1f1SOVNn0h1SOXNm0g1ROWNR1g1oNUNW1i1iNTN[1k1dNRNb1l1^NQNh1k1\\NoMk1k1YNoMo1g1YNTNn1b1[NWNm1^1\\N]Nk1Y1]NaNk1U1]NbNm1U1\\N`NP2U1YNfNn1d0fNVOb1K\\ONT7KPkh1"}, "label": "the elephant in the front of the line"}]} {"id": 199485, "image": "COCO_train2014_000000199485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant used for transport\"."}], "task": "refering", "answer_template": "The \"an elephant used for transport\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 58, 59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 124, 125, 126, 127, 128, 129, 130, 131, 132, 147, 148, 149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 178, 194, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 292, 293, 310, 311, 312, 313, 314, 315, 316, 333, 334, 335, 336, 337, 338], "bbox": [0.42279687499999996, 0.11686182669789227, 0.7856249999999999, 0.986510538641686], "iscrowd": 0, "area": 63161.32310000001, "rle": {"size": [427, 640], "counts": "^Ua3:o<9F;E:G4K5L5J5K6K4K5K6K4K5L5J5K6M2M3N3M2N2N1O1O2N1O1O2O09G9H6IO2O00001O000fNUMSHl2j7lM`GT2_8]1N3M2N4L6J5K5K5K5K6J5K5K7Il0TOk0UOl0YKPG\\3j9F0O2N1O1O1O1O1O1O1O1O1O1O1N2O1N2N2O1N2O1N2O1N3N1N2N2O1N2O1N2O1N2O1N2O100000I7G8I8H8G9H7I8H8G9I6J7J6I7J6M210001O0O1000001O005K5K5K6J5K5K6K4K5mM_F2g9]OdFc0a9mNjFS1[9]NPGc1\\:0000000000001O00000000000000000000001O000000000000O1000000000000000000000000O100000000000000O100000000000000O1000000001O00001O00001O00001iNUNlFk1T9[NfFe1Z9bN_F_1`9hNYFX1g9nNSFR1m9UOlEl0S:V100O2O000jJbN_N_1_1gN\\NY1b1nNYNR1f1SOVNn0h1SOXNm0g1ROWNR1g1oNUNW1i1iNTN[1k1dNRNb1l1^NQNh1k1\\NoMk1k1YNoMo1g1YNTNn1b1[NWNm1^1\\N]Nk1Y1]NaNk1U1]NbNm1U1\\N`NP2U1YNfNn1d0fNVOb1K\\ONT7KPkh1"}, "label": "an elephant used for transport"}]} {"id": 199485, "image": "COCO_train2014_000000199485.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant between two other elephants\"."}], "task": "refering", "answer_template": "The \"an elephant between two other elephants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 97, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 282, 283, 284, 285, 286, 305, 306, 307, 308, 309, 328, 329], "bbox": [0.238390625, 0.21362997658079624, 0.46328125, 0.968688524590164], "iscrowd": 0, "area": 35294.984, "rle": {"size": [427, 640], "counts": "nmo1m0\\ in this image.", "objects": [{"patches": [75, 76, 77, 78, 97, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 282, 283, 284, 285, 286, 305, 306, 307, 308, 309, 328, 329], "bbox": [0.238390625, 0.21362997658079624, 0.46328125, 0.968688524590164], "iscrowd": 0, "area": 35294.984, "rle": {"size": [427, 640], "counts": "nmo1m0\\ in this image.", "objects": [{"patches": [283, 306, 329, 352, 353, 375, 376, 399, 422, 423, 425, 445, 446, 447, 448, 449], "bbox": [0.308265625, 0.6021364452423699, 0.561734375, 1.0], "iscrowd": 0, "area": 6124.471150000001, "rle": {"size": [557, 640], "counts": "f_[33Za08H8H8G:G7I5K5K4K6K5K5K4L5J6K5K\\McA]2X>bMmA^2o=aMVB`2d=`MaBa2Z=^MjB2^OV2c=gMTCLB[2V=hMmCV2o;iMXDT2d;lMcDQ2Z;nMlDP2Q;oMVEn1g:QN`El1S=O000O10000000g@[Ng>e1UA\\Nn>d1n@^NT?m12M3O12N2N2N2N2N2NN2M2N2O2M2N2N3M2N2N2N3M2O0O2N2N1O2N2N1O2N5L4K5K5K4L1O00000001O0000000O1O100O101N100O100O1O100O100O100O100O1O100O010O100O10000O10000O100001O1O001O001O001O001O001N2O002N2N2N2N0000O100O100O10000000000O1000000000O1000002N1O2N2N2N2N2N2NmYh4"}, "label": "man ' s black shoulder bag"}]} {"id": 535358, "image": "COCO_train2014_000000535358.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bag over the shoulder of the man closest to the horse\"."}], "task": "refering", "answer_template": "The \"the bag over the shoulder of the man closest to the horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [283, 306, 329, 352, 353, 375, 376, 399, 422, 423, 425, 445, 446, 447, 448, 449], "bbox": [0.308265625, 0.6021364452423699, 0.561734375, 1.0], "iscrowd": 0, "area": 6124.471150000001, "rle": {"size": [557, 640], "counts": "f_[33Za08H8H8G:G7I5K5K4K6K5K5K4L5J6K5K\\McA]2X>bMmA^2o=aMVB`2d=`MaBa2Z=^MjB2^OV2c=gMTCLB[2V=hMmCV2o;iMXDT2d;lMcDQ2Z;nMlDP2Q;oMVEn1g:QN`El1S=O000O10000000g@[Ng>e1UA\\Nn>d1n@^NT?m12M3O12N2N2N2N2N2NN2M2N2O2M2N2N3M2N2N2N3M2O0O2N2N1O2N2N1O2N5L4K5K5K4L1O00000001O0000000O1O100O101N100O100O1O100O100O100O100O1O100O010O100O10000O10000O100001O1O001O001O001O001O001N2O002N2N2N2N0000O100O100O10000000000O1000000000O1000002N1O2N2N2N2N2N2NmYh4"}, "label": "the bag over the shoulder of the man closest to the horse"}]} {"id": 535358, "image": "COCO_train2014_000000535358.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the car behind the horse\"."}], "task": "refering", "answer_template": "The \"the car behind the horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [363, 364, 386, 387, 388, 389, 390, 409, 410, 411, 412, 413, 431, 432, 433, 434, 435, 436, 454, 455, 456, 457, 458, 459], "bbox": [0.73140625, 0.7591741472172352, 1.0, 1.0], "iscrowd": 0, "area": 14820.239549999998, "rle": {"size": [557, 640], "counts": "`So71Za02N2N2N2N2N2N2N2N2N2N2O1N2O1N2O1N2O1N200O1000000O10000O1TMkN`EU1]:oNaER1^:PO_ER1a:oN]ER1c:nN[ET1e:mNYET1g:mNVEU1j:lNTEU1m:kNPEW1P;iNoDX1Q;iNlDY1T;hNjDY1V;hNhDY1X;gNfD[1Z;fNdD[1\\;fNbD[1_;eN_D\\1a;dN]D^1c;cN[D^1e;cNWD`1i;aNSDb1m;_NPDc1P<^NmCd1S<]NiCf1X<[NdCg1\\<[N`Cg1`<\\N[Cf1e in this image.", "objects": [{"patches": [363, 364, 386, 387, 388, 389, 390, 409, 410, 411, 412, 413, 431, 432, 433, 434, 435, 436, 454, 455, 456, 457, 458, 459], "bbox": [0.73140625, 0.7591741472172352, 1.0, 1.0], "iscrowd": 0, "area": 14820.239549999998, "rle": {"size": [557, 640], "counts": "`So71Za02N2N2N2N2N2N2N2N2N2N2O1N2O1N2O1N2O1N200O1000000O10000O1TMkN`EU1]:oNaER1^:PO_ER1a:oN]ER1c:nN[ET1e:mNYET1g:mNVEU1j:lNTEU1m:kNPEW1P;iNoDX1Q;iNlDY1T;hNjDY1V;hNhDY1X;gNfD[1Z;fNdD[1\\;fNbD[1_;eN_D\\1a;dN]D^1c;cN[D^1e;cNWD`1i;aNSDb1m;_NPDc1P<^NmCd1S<]NiCf1X<[NdCg1\\<[N`Cg1`<\\N[Cf1e in this image.", "objects": [{"patches": [172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 224, 239, 240, 241, 242, 243, 244, 245, 246, 247, 262, 263, 264, 265, 266, 267, 268, 269, 285, 286, 287, 288, 289, 291, 292, 308, 309, 310, 311, 312, 314, 315], "bbox": [0.399625, 0.39225352112676054, 0.748640625, 0.760784708249497], "iscrowd": 0, "area": 25625.4777, "rle": {"size": [497, 640], "counts": "Xcl32]?3M2O2M3M2fNE\\C>a^ZVUTSXD@j;a0VD\\Ol;e0TDYOm;h0SDUOo;l0QDQOQ\\Og9_3kERMd0\\Oa9b3kEPMi0[O\\9e3kEnLS;R3mDlLU;T3lDhLW;Y3hDeLZ;[3fDcL\\;]3dDaL^;_3bD_L`;a3`D\\Lc;d35000O100000001O00000000000000000000001O00000000000O10000O10001N10000O10000`DTLZ;m3fDVLW;j3iDYLT;g3lD]Lo:P2QEUO0nNl:m1XEROKUOi:i1_EoNIZOd:i1fEiNFAa:f1lEfNCH]:b1TFbN_OOY:`1[F^N\\O2Y:`1^F[NYO5Y:`1aFXNVO9Y:^1dFVNSO in this image.", "objects": [{"patches": [172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 224, 239, 240, 241, 242, 243, 244, 245, 246, 247, 262, 263, 264, 265, 266, 267, 268, 269, 285, 286, 287, 288, 289, 291, 292, 308, 309, 310, 311, 312, 314, 315], "bbox": [0.399625, 0.39225352112676054, 0.748640625, 0.760784708249497], "iscrowd": 0, "area": 25625.4777, "rle": {"size": [497, 640], "counts": "Xcl32]?3M2O2M3M2fNE\\C>a^ZVUTSXD@j;a0VD\\Ol;e0TDYOm;h0SDUOo;l0QDQOQ\\Og9_3kERMd0\\Oa9b3kEPMi0[O\\9e3kEnLS;R3mDlLU;T3lDhLW;Y3hDeLZ;[3fDcL\\;]3dDaL^;_3bD_L`;a3`D\\Lc;d35000O100000001O00000000000000000000001O00000000000O10000O10001N10000O10000`DTLZ;m3fDVLW;j3iDYLT;g3lD]Lo:P2QEUO0nNl:m1XEROKUOi:i1_EoNIZOd:i1fEiNFAa:f1lEfNCH]:b1TFbN_OOY:`1[F^N\\O2Y:`1^F[NYO5Y:`1aFXNVO9Y:^1dFVNSO in this image.", "objects": [{"patches": [189, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 258, 259, 260, 261, 262, 281, 282, 283, 284, 305, 306], "bbox": [0.228609375, 0.4696579476861167, 0.41184374999999995, 0.770784708249497], "iscrowd": 0, "area": 10540.308149999999, "rle": {"size": [497, 640], "counts": "^TW21W?O001O1O10O01O1hNbC[O^<1YDLh;AmDo:@VE=k:_O[E?n in this image.", "objects": [{"patches": [189, 190, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 258, 259, 260, 261, 262, 281, 282, 283, 284, 305, 306], "bbox": [0.228609375, 0.4696579476861167, 0.41184374999999995, 0.770784708249497], "iscrowd": 0, "area": 10540.308149999999, "rle": {"size": [497, 640], "counts": "^TW21W?O001O1O10O01O1hNbC[O^<1YDLh;AmDo:@VE=k:_O[E?n in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 126, 127, 128, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 254, 255, 256, 257, 266, 267, 268, 269, 270, 271, 291, 292, 293], "bbox": [0.025343749999999998, 0.3074375, 0.9459531250000002, 0.7331041666666667], "iscrowd": 0, "area": 78415.01124999998, "rle": {"size": [480, 640], "counts": "lg73j>6J7H7J7H7J7H7J5K3L5L3L5M2N3O0O2O00001N101O2N2M4M3M3M2M4M3M2M4M3M2N2M10001O0O101O000O2O001O0O101O000O2O000O2O001O0O101O00001O00001O0O1000O100O100O10000O10O0100O100O100O100O1000O0100O100O100O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1000000000O100O100O100O2N100O100O100O1O100O100O100O1O100O100O100O1O2O0O100O100O1O100O100O100O1O100O100O100O1O100O2O0O100O1O100O100O100O1O100O100O1000000000001N1000000000000000000000000000000000000O10001O0000000000000000000000000000000O1000000000001O0000000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O0000000001O0000000000000000000000000000000001O0000000000000000000000000000000000000000001O000001O00000000000000000000000000000000000001O0001O0000008H=C in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 126, 127, 128, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 254, 255, 256, 257, 266, 267, 268, 269, 270, 271, 291, 292, 293], "bbox": [0.025343749999999998, 0.3074375, 0.9459531250000002, 0.7331041666666667], "iscrowd": 0, "area": 78415.01124999998, "rle": {"size": [480, 640], "counts": "lg73j>6J7H7J7H7J7H7J5K3L5L3L5M2N3O0O2O00001N101O2N2M4M3M3M2M4M3M2M4M3M2N2M10001O0O101O000O2O001O0O101O000O2O000O2O001O0O101O00001O00001O0O1000O100O100O10000O10O0100O100O100O100O1000O0100O100O100O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1000000000O100O100O100O2N100O100O100O1O100O100O100O1O100O100O100O1O2O0O100O100O1O100O100O100O1O100O100O100O1O100O2O0O100O1O100O100O100O1O100O100O1000000000001N1000000000000000000000000000000000000O10001O0000000000000000000000000000000O1000000000001O0000000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O0000000001O0000000000000000000000000000000001O0000000000000000000000000000000000000000001O000001O00000000000000000000000000000000000001O0001O0000008H=C in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 291, 292, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.0015, 0.0011267605633802818, 0.7195937499999999, 0.9898591549295774], "iscrowd": 0, "area": 131585.62924999994, "rle": {"size": [426, 640], "counts": "Rb04P=7I8I7H7I8H8H7J7H7M2O1N3N1O1N2O2M2O1O2M2O2N1N3N1N3N2N1N3N1O2M2O2M2O2N5J6K6J5J7J5K6I6K6I6K5K6I6K6I6K6K4K6K4K6K4L5J6K5J6K5J5L5J6K5J6K5K5J5L5J6K5J6K5J6K5J5L5J6K5J6K5J6K4L5J6K0O10000O100O10000O100O10000O100O10000O100O10000O100O10000O1O100O1O1O100O100O1O100O100O100O100O1O100O100O100O1000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000001O00000hI9mNGo0c0kN]OR1i0kNWOQ1Q1kNoNl0]1QOcNb0k1]OUNa0l1@TN?m1ASN>n1BRNZO_Oj0?WO@l0>TOAn0=SOBo0=QOBQ1=oNBT1;mNDU1;kNDW1:jNEX1:hNE[17gNH[1KRO3P1CYOl0_OVO?l0@TO?n0_OSO`0o0^OSO?Q1^OPOa0R1]OPOa0S1\\OoNb0S1]OnNa0U1\\OlNc0W1ZOkNZNVLR2Q5BSOP1_OSO?o0_OSO>o0@RO`0P1]OQOc0Q1XOQOh0S1ROPOn0S1mNoNS1T1gNoNX1V1aNmN_1W1ZNlNf1[6011O2M4M2N2M3N2N2N1N3N1O2N1O2N100O1O001O1OROdEoN[:Q1gEnNX:R1jElNW:S1kElNT:T1oEjNP:W1PFiNP:V1PFkNP:T1QFlNn9U1RFjNo9U1RFkNn9U1RFkNn9T1SFlNm9S1TFlNl9U1SFlNo9Q1RFnNS:n0mEROS:m0nEROR:c0aEoN>>R:b0YF]Og9b0ZF^Og9a0ZF^Of9a0[F_Of9`0[F_Oe9`0\\F@e9?\\F@d9`0]F_Od9`0]F_Oc9`0_F_Ob9`0^F_Oc9a0^F]Od9c0]FYOf9g0ZFWOh9h0YFUOj9k0[1O1O100O014K5K4JhUZ2"}, "label": "a girl holding a donut"}]} {"id": 2892, "image": "COCO_train2014_000000002892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman who is wearing a blue shirt\"."}], "task": "refering", "answer_template": "The \"the woman who is wearing a blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 291, 292, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.0015, 0.0011267605633802818, 0.7195937499999999, 0.9898591549295774], "iscrowd": 0, "area": 131585.62924999994, "rle": {"size": [426, 640], "counts": "Rb04P=7I8I7H7I8H8H7J7H7M2O1N3N1O1N2O2M2O1O2M2O2N1N3N1N3N2N1N3N1O2M2O2M2O2N5J6K6J5J7J5K6I6K6I6K5K6I6K6I6K6K4K6K4K6K4L5J6K5J6K5J5L5J6K5J6K5K5J5L5J6K5J6K5J6K5J5L5J6K5J6K5J6K4L5J6K0O10000O100O10000O100O10000O100O10000O100O10000O100O10000O1O100O1O1O100O100O1O100O100O100O100O1O100O100O100O1000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000001O00000hI9mNGo0c0kN]OR1i0kNWOQ1Q1kNoNl0]1QOcNb0k1]OUNa0l1@TN?m1ASN>n1BRNZO_Oj0?WO@l0>TOAn0=SOBo0=QOBQ1=oNBT1;mNDU1;kNDW1:jNEX1:hNE[17gNH[1KRO3P1CYOl0_OVO?l0@TO?n0_OSO`0o0^OSO?Q1^OPOa0R1]OPOa0S1\\OoNb0S1]OnNa0U1\\OlNc0W1ZOkNZNVLR2Q5BSOP1_OSO?o0_OSO>o0@RO`0P1]OQOc0Q1XOQOh0S1ROPOn0S1mNoNS1T1gNoNX1V1aNmN_1W1ZNlNf1[6011O2M4M2N2M3N2N2N1N3N1O2N1O2N100O1O001O1OROdEoN[:Q1gEnNX:R1jElNW:S1kElNT:T1oEjNP:W1PFiNP:V1PFkNP:T1QFlNn9U1RFjNo9U1RFkNn9U1RFkNn9T1SFlNm9S1TFlNl9U1SFlNo9Q1RFnNS:n0mEROS:m0nEROR:c0aEoN>>R:b0YF]Og9b0ZF^Og9a0ZF^Of9a0[F_Of9`0[F_Oe9`0\\F@e9?\\F@d9`0]F_Od9`0]F_Oc9`0_F_Ob9`0^F_Oc9a0^F]Od9c0]FYOf9g0ZFWOh9h0YFUOj9k0[1O1O100O014K5K4JhUZ2"}, "label": "the woman who is wearing a blue shirt"}]} {"id": 125774, "image": "COCO_train2014_000000125774.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the arm of a dark skinned person sitting next to a man holding a teddy bear\"."}], "task": "refering", "answer_template": "The \"the arm of a dark skinned person sitting next to a man holding a teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [352, 374, 375, 396, 397, 398, 418, 419, 420, 440, 441, 442, 443, 462, 463, 464, 465, 484, 485, 486, 487], "bbox": [0.0, 0.687546875, 0.1702618657937807, 1.0], "iscrowd": 0, "area": 13279.521049999996, "rle": {"size": [640, 611], "counts": "h=X6i=O1O1O1O1O2N2N3M6J3M2N2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M3K6J6K4K6J:G;D in this image.", "objects": [{"patches": [352, 374, 375, 396, 397, 398, 418, 419, 420, 440, 441, 442, 443, 462, 463, 464, 465, 484, 485, 486, 487], "bbox": [0.0, 0.687546875, 0.1702618657937807, 1.0], "iscrowd": 0, "area": 13279.521049999996, "rle": {"size": [640, 611], "counts": "h=X6i=O1O1O1O1O2N2N3M6J3M2N2N1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M3K6J6K4K6J:G;D in this image.", "objects": [{"patches": [6, 7, 8, 27, 28, 29, 30, 31, 32, 49, 50, 51, 52, 53, 54, 55, 71, 72, 73, 74, 75, 76, 77, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 135, 136, 137, 138, 139, 140, 141, 142, 155, 156, 157, 158, 159, 160, 161, 162, 163, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 279, 280, 286, 287, 288, 289, 290, 291, 292, 300, 301, 302, 308, 309, 310, 311, 312, 313, 314, 320, 321, 322, 323, 324, 330, 331, 332, 333, 334, 335, 336, 337, 338, 340, 341, 342, 343, 344, 345, 346, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 375, 376, 377, 378, 379, 380, 381, 382, 384, 385, 386, 387, 388, 397, 398, 399, 400, 401, 402, 403, 404, 406, 407, 408, 409, 420, 421, 422, 423, 424, 425, 429, 430, 443, 444, 445, 446, 447, 448, 451, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 481, 482, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 503, 504, 505], "bbox": [0.0, 0.032265625, 0.9847626841243864, 1.0], "iscrowd": 0, "area": 130088.26434999998, "rle": {"size": [640, 611], "counts": "_8Z5d>3N2M3N2N2M3N2M=D2N2M3N2M3N2N2M3N2M3N3M2M3N2M3N2N2M3O1N2N2N2N2N2N2N2N3N1N2N2N2N2N2N2N2N3M2O1N2N2N2N3M2N2N2O1N2O2N1N2O1O1N2O1O2M2O1O1N2O1O2M2O1O1N2O1O1N3N1N2O1O1N2O2N1N3M4M3L4M2M4M3L4M3L4M2M5K6K4K5L4K5L4K5L5J5L4L1N10001N101O0O101O0O101O0O2O0O101N1N3N1N2O2M2O1N3N1O2M2O1N3N10QFfJX5Z5fJlJW5S5gJSKU5l4kJXKS5\\2gFROT4hNQ5R2hG]NU3Go4h1QH`No2Ln4`1WHbNj24k4V1_HfNe29h4m0hHiN^2?h4c0oHmNW2f0f4:WInNS2m0c40_IROl1T1a4GfIUOg1Y1`4AkIUOd1_1^4[OoISOd1f1Z4VOTJoNd1m1Y4ROVJlNc1T2V4oN^JcN_1`2R4lNnNW1R1hNmNY1S1fNmN\\1R1cNnN_1R1_NmNd1R1\\NmNe1S1ZNmNh1R1WNoNj1P1UNQOl1n0TNROl1n0SNSOn1l0QNTOQ2k0nMVOS2i0mMWOS2i0lMXOU2g0jMZOW2e0hM[OZ2d0eM]O[2c0eM]O\\2b0cM_O^2`0aMA_2?`MAb2>^MBc2=\\MDe2;ZMFh28WMIj26VMIm25RMLo23PMNR30mL1T3NlL2V3LiL4Y3KgL4\\3JcL6_3IaL6b3H^L7d3H\\L6h3HXL7j3HVL7m3GSL8o3GQL7S4GmK8U4GkK8X4FhK8\\4FcK:_4EaK:b4D^K;d4D\\K:h4DXK;i4EWK:k4EUK9m4GSK8n4HRK7o4IQK5R5JnJ5S5KlJ5U5KkJ3W5MiJ2Y5MgJ2Z5NfJ0\\50dJO]51cJN_51aJMa53_JLb5aLeJFIi3b5`LgJFFk3d5]LiJFCm3d5[LlJG@n3d5ZLnJG^Oo3d5YLQKF[OQ4e5WLRKGYOQ4f5WLSKGWOR4f5ULWKGSOT4f5TLYKGQOU4g5RLZKHoNV4g5QL]KGlNX4g5oK`KHiNY4g5nKbKHgNZ4h5lKdKHdN\\4h5kKfKHcN\\4g5kKhKHaN]4g5iKkKI^N^4h5gKmKI\\N_4g5gKoKIZN`4g5fKQLIXNa4g5eKTLHUNc4h5bKVLJSNc4g5aKYLKPNd4g5^K^LLkMf4g5[KbLNgMf4i5YKcL0eMf4h5WKhL1`Mh4i5SKkL4\\Mi4k5oJlL7ZMi4k5lJPM9UMk4l5iJRM;RMl4n5cJUM`0mLm4o5\\J\\Me0fLn4P6TJbMm0^Lo4Q6mIgMT1XLo4i:QKWEo4i:QKXEm4i:SKWEa4UOaJd;n0WE`4XO`Jb;o0WE_4YOaJ`;P1WE]4\\ObJ];Q1WE\\4_OaJZ;S1WE[4AaJX;T1XEX4CcJU;U1XEW4FbJR;W1XEU4`;jKaDT4`;lK`DS4a;mK_DQ4c;oK]DP4d;PL]Dm3e;SL[Dl3j7nJQKV1UMk3k7PKPKT1UMj3l7TKnJR1WMh3l7WKlJQ1XMg3l7ZKkJo0YMe3n7]KhJn0[Mc3m7aKgJl0\\Ma3o7dKeJj0\\Ma3P8fKbJj0^M_3P8iKaJh0`M]3P8lK^Jh0bM[3P8oK\\Jg0dMX3R8RLYJf0fMV3Q8VLWJe0hMT3R8XLUJd0iMS3R8[LSJc0kMQ3S8]LPJc0nMm2S8bLnIa0oMl2T8dLkIa0QNj2T8gLjI?RNi2`7RLiGg0c2?UNf2Z7ZLmGc0b2>XNc2Y7^LlGb0b2=ZNb2X7_LkGd0a2<]N_2W7cLjGc0`2<`N\\2W7eLhGe0`2:bN[2V7gLgGe0_2:eNX2U7jLfGf0_28gNV2U7mLdGf0^28jNS2T7PMcGg0]27mNQ2S7RMbGg0]26nNP2S7TMaGh0\\25QOm1S7WM_Gh0\\24SOk1S7YM]Gj0[23VOi1Q7oNgIYOYOf1Q7ROeIXO[Od1P7VOcIWO^Oa1P7YOaIVO@`1n6\\O`IUOC]1n6_O]IUOFZ1n6B[ITOHX1m6FYISOKV1l6HXIROMT1k6LVIQO0P1l60SIPO2n0l63PIPO5k0k67oHnN7i0k69mHoN9e0k6=lHnN9d0l6=kHPO:a0k6`0kHoN;?k6b0iHPO=;k6e0iHPO=9k6g0gHQO?6j6j0gHPO`04j6k0hHPO?2j6o0VGTMX1i1i02j6P1VGXMU1e1l01i6S1UG\\MT1^1W1Ia6]1SG]MV1\\1a1YO[6n1nF^MX1Y1i1kNW6^2hF^MY1X1S2[NQ6P3bF^M\\1U1e;\\1oB`M^1R1b;_1oB`M`1P1a;`1oB`Ma1o0_;a1PCaMb1m0^;b1oBbMd1k0\\;d1oBaMg1j0Z;d1PCbMg1i0X;f1PCaMj1h0V;g1oBaMm1g0S;h1PCbMn1e0R;i1PCaMP2e0o:^2SEaMm:_2TE`Mk:`2WE_Mi:`2YE_Mg:`2[E_Md:b2]E]Mc:b2_E]M`:c2bE\\M^:d2cE[M\\:e2fEZMZ:e2hEZMW:f2kEYMU:g2lEXMT:g2nEXMR:g2PFXMo9i2RFVMn9i2TFVMl9i2WFUMi9j2YFUMg9k2ZFTMe9l2]FSMc9l2_FSMa9m2`FRM`9m2bFRM^9m2dFRM[9o2fFPMZ9o2hFPMX9o2jFPMV9o2lFPMT9o2nFPMQ9P3QGoLo8P3SGoLm8P3UGoLj8Q3XGnLh8Q3ZGnLe8R3]GmLc8R3_GmLa8R3aGmL^8S3X40O01O001O1O001O001O1OO1N2O100O100000000O100000010O01O001O001O00001O001O001O001O001OM3_Oa01O1O100O1O1O101N1O100O1O1O100O1O100O2N100O1O2N101N2N2M3N2N2M3N2N3L3N3L3N3M3L3N3M2M4K4ZOgkX1E\\bfNHSb0`0a]OC`b0e0S]O_Omb0j0O100O100O1000000O100O100O10000001O0000001O00001O001O001O001O001O001O001O001O001O001O001O001O1O2N1O1O1O2N1O1O2N1O1O2N3M3MTd5"}, "label": "a man sitting the chair and holding the brown color teddy"}]} {"id": 125774, "image": "COCO_train2014_000000125774.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a car shirt and santa hat\"."}], "task": "refering", "answer_template": "The \"a man wearing a car shirt and santa hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 27, 28, 29, 30, 31, 32, 49, 50, 51, 52, 53, 54, 55, 71, 72, 73, 74, 75, 76, 77, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 135, 136, 137, 138, 139, 140, 141, 142, 155, 156, 157, 158, 159, 160, 161, 162, 163, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 264, 265, 266, 267, 268, 269, 279, 280, 286, 287, 288, 289, 290, 291, 292, 300, 301, 302, 308, 309, 310, 311, 312, 313, 314, 320, 321, 322, 323, 324, 330, 331, 332, 333, 334, 335, 336, 337, 338, 340, 341, 342, 343, 344, 345, 346, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 375, 376, 377, 378, 379, 380, 381, 382, 384, 385, 386, 387, 388, 397, 398, 399, 400, 401, 402, 403, 404, 406, 407, 408, 409, 420, 421, 422, 423, 424, 425, 429, 430, 443, 444, 445, 446, 447, 448, 451, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 481, 482, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 503, 504, 505], "bbox": [0.0, 0.032265625, 0.9847626841243864, 1.0], "iscrowd": 0, "area": 130088.26434999998, "rle": {"size": [640, 611], "counts": "_8Z5d>3N2M3N2N2M3N2M=D2N2M3N2M3N2N2M3N2M3N3M2M3N2M3N2N2M3O1N2N2N2N2N2N2N2N3N1N2N2N2N2N2N2N2N3M2O1N2N2N2N3M2N2N2O1N2O2N1N2O1O1N2O1O2M2O1O1N2O1O2M2O1O1N2O1O1N3N1N2O1O1N2O2N1N3M4M3L4M2M4M3L4M3L4M2M5K6K4K5L4K5L4K5L5J5L4L1N10001N101O0O101O0O101O0O2O0O101N1N3N1N2O2M2O1N3N1O2M2O1N3N10QFfJX5Z5fJlJW5S5gJSKU5l4kJXKS5\\2gFROT4hNQ5R2hG]NU3Go4h1QH`No2Ln4`1WHbNj24k4V1_HfNe29h4m0hHiN^2?h4c0oHmNW2f0f4:WInNS2m0c40_IROl1T1a4GfIUOg1Y1`4AkIUOd1_1^4[OoISOd1f1Z4VOTJoNd1m1Y4ROVJlNc1T2V4oN^JcN_1`2R4lNnNW1R1hNmNY1S1fNmN\\1R1cNnN_1R1_NmNd1R1\\NmNe1S1ZNmNh1R1WNoNj1P1UNQOl1n0TNROl1n0SNSOn1l0QNTOQ2k0nMVOS2i0mMWOS2i0lMXOU2g0jMZOW2e0hM[OZ2d0eM]O[2c0eM]O\\2b0cM_O^2`0aMA_2?`MAb2>^MBc2=\\MDe2;ZMFh28WMIj26VMIm25RMLo23PMNR30mL1T3NlL2V3LiL4Y3KgL4\\3JcL6_3IaL6b3H^L7d3H\\L6h3HXL7j3HVL7m3GSL8o3GQL7S4GmK8U4GkK8X4FhK8\\4FcK:_4EaK:b4D^K;d4D\\K:h4DXK;i4EWK:k4EUK9m4GSK8n4HRK7o4IQK5R5JnJ5S5KlJ5U5KkJ3W5MiJ2Y5MgJ2Z5NfJ0\\50dJO]51cJN_51aJMa53_JLb5aLeJFIi3b5`LgJFFk3d5]LiJFCm3d5[LlJG@n3d5ZLnJG^Oo3d5YLQKF[OQ4e5WLRKGYOQ4f5WLSKGWOR4f5ULWKGSOT4f5TLYKGQOU4g5RLZKHoNV4g5QL]KGlNX4g5oK`KHiNY4g5nKbKHgNZ4h5lKdKHdN\\4h5kKfKHcN\\4g5kKhKHaN]4g5iKkKI^N^4h5gKmKI\\N_4g5gKoKIZN`4g5fKQLIXNa4g5eKTLHUNc4h5bKVLJSNc4g5aKYLKPNd4g5^K^LLkMf4g5[KbLNgMf4i5YKcL0eMf4h5WKhL1`Mh4i5SKkL4\\Mi4k5oJlL7ZMi4k5lJPM9UMk4l5iJRM;RMl4n5cJUM`0mLm4o5\\J\\Me0fLn4P6TJbMm0^Lo4Q6mIgMT1XLo4i:QKWEo4i:QKXEm4i:SKWEa4UOaJd;n0WE`4XO`Jb;o0WE_4YOaJ`;P1WE]4\\ObJ];Q1WE\\4_OaJZ;S1WE[4AaJX;T1XEX4CcJU;U1XEW4FbJR;W1XEU4`;jKaDT4`;lK`DS4a;mK_DQ4c;oK]DP4d;PL]Dm3e;SL[Dl3j7nJQKV1UMk3k7PKPKT1UMj3l7TKnJR1WMh3l7WKlJQ1XMg3l7ZKkJo0YMe3n7]KhJn0[Mc3m7aKgJl0\\Ma3o7dKeJj0\\Ma3P8fKbJj0^M_3P8iKaJh0`M]3P8lK^Jh0bM[3P8oK\\Jg0dMX3R8RLYJf0fMV3Q8VLWJe0hMT3R8XLUJd0iMS3R8[LSJc0kMQ3S8]LPJc0nMm2S8bLnIa0oMl2T8dLkIa0QNj2T8gLjI?RNi2`7RLiGg0c2?UNf2Z7ZLmGc0b2>XNc2Y7^LlGb0b2=ZNb2X7_LkGd0a2<]N_2W7cLjGc0`2<`N\\2W7eLhGe0`2:bN[2V7gLgGe0_2:eNX2U7jLfGf0_28gNV2U7mLdGf0^28jNS2T7PMcGg0]27mNQ2S7RMbGg0]26nNP2S7TMaGh0\\25QOm1S7WM_Gh0\\24SOk1S7YM]Gj0[23VOi1Q7oNgIYOYOf1Q7ROeIXO[Od1P7VOcIWO^Oa1P7YOaIVO@`1n6\\O`IUOC]1n6_O]IUOFZ1n6B[ITOHX1m6FYISOKV1l6HXIROMT1k6LVIQO0P1l60SIPO2n0l63PIPO5k0k67oHnN7i0k69mHoN9e0k6=lHnN9d0l6=kHPO:a0k6`0kHoN;?k6b0iHPO=;k6e0iHPO=9k6g0gHQO?6j6j0gHPO`04j6k0hHPO?2j6o0VGTMX1i1i02j6P1VGXMU1e1l01i6S1UG\\MT1^1W1Ia6]1SG]MV1\\1a1YO[6n1nF^MX1Y1i1kNW6^2hF^MY1X1S2[NQ6P3bF^M\\1U1e;\\1oB`M^1R1b;_1oB`M`1P1a;`1oB`Ma1o0_;a1PCaMb1m0^;b1oBbMd1k0\\;d1oBaMg1j0Z;d1PCbMg1i0X;f1PCaMj1h0V;g1oBaMm1g0S;h1PCbMn1e0R;i1PCaMP2e0o:^2SEaMm:_2TE`Mk:`2WE_Mi:`2YE_Mg:`2[E_Md:b2]E]Mc:b2_E]M`:c2bE\\M^:d2cE[M\\:e2fEZMZ:e2hEZMW:f2kEYMU:g2lEXMT:g2nEXMR:g2PFXMo9i2RFVMn9i2TFVMl9i2WFUMi9j2YFUMg9k2ZFTMe9l2]FSMc9l2_FSMa9m2`FRM`9m2bFRM^9m2dFRM[9o2fFPMZ9o2hFPMX9o2jFPMV9o2lFPMT9o2nFPMQ9P3QGoLo8P3SGoLm8P3UGoLj8Q3XGnLh8Q3ZGnLe8R3]GmLc8R3_GmLa8R3aGmL^8S3X40O01O001O1O001O001O1OO1N2O100O100000000O100000010O01O001O001O00001O001O001O001O001OM3_Oa01O1O100O1O1O101N1O100O1O1O100O1O100O2N100O1O2N101N2N2M3N2N2M3N2N3L3N3L3N3M3L3N3M2M4K4ZOgkX1E\\bfNHSb0`0a]OC`b0e0S]O_Omb0j0O100O100O1000000O100O100O10000001O0000001O00001O001O001O001O001O001O001O001O001O001O001O001O1O2N1O1O1O2N1O1O2N1O1O2N3M3MTd5"}, "label": "a man wearing a car shirt and santa hat"}]} {"id": 68430, "image": "COCO_train2014_000000068430.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the handle bars and seat of one bicycle and the back half of another\"."}], "task": "refering", "answer_template": "The \"the handle bars and seat of one bicycle and the back half of another\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [27, 43, 44, 45, 57, 58, 60, 61, 62, 74, 75, 78, 79, 91, 92, 93], "bbox": [0.38939583333333333, 0.078359375, 0.7037916666666667, 0.26834375], "iscrowd": 0, "area": 4022.3452000000007, "rle": {"size": [640, 480], "counts": "YPe39gc07I0TOEm]O`0Rb0Cj]O`0ea0AW^OM3b0ia0BP^ON6?na0Bi]O26>Rb0Nk]O4Vb0Lf]O6[b0g0101OO2N1002[Nf]O7Nn0bb0lNd]OR1fb0O1O01O000001O1O1N101O1O2]Om\\O5Vc0Fn\\O8^c0O2N2N1O2N10O2V\\OKec0801O001O1000001O0O12D]\\O6ac0Hd\\O6dc0N3N^he04^WZO:F:F9G:F:F3M00O1000000O100gNj]Oh0Ub0XOl]Og0Ub0WOn]Oh0Rb0WOo]Oi0Qb0VOQ^Oh0oa0XOS^Og0ma0XOT^OLIb0Rb0Aa^Oa0]a0_Oc^Oc0Za0]Og^Oe0Wa0ZOk^Oh0Sa0WOm^Ok0Ra0SOo^Oo0Pa0POP_OR1o`0lNQ_OW1n`0hNQ_OY1o`0fNQ_O[1ja0O01O1O010O1O001000O2O000O10000O2O000O10000O2O0N2N2N2M3N6J7I6J6ISeh2"}, "label": "the handle bars and seat of one bicycle and the back half of another"}]} {"id": 68430, "image": "COCO_train2014_000000068430.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bread sandwich in plate\"."}], "task": "refering", "answer_template": "The \"bread sandwich in plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [260, 261, 262, 277, 278, 279, 280, 294, 295, 296, 297, 307, 308, 309, 310, 311, 312, 313, 314, 324, 325, 326, 327, 328, 329, 330, 331, 342, 343, 344, 345, 346, 347, 361, 362], "bbox": [0.10233333333333333, 0.6856718749999999, 0.5267083333333333, 0.947515625], "iscrowd": 0, "area": 16755.671649999997, "rle": {"size": [640, 480], "counts": "gTo0;`c08H8J4M3M2O1N3M2O1N3M2N2O2O0O1010O1O00100O1O010O1O100O001O100O00100O1O1O010O1O1O010O1O1O10O0100000O100000001N101O1O1O0O2N2L3N3M3M3M2M4M2N1O2M2O2N1O1O2M2O2N1O1O2M2O2N1O1N3N1O2N1O1N2O00001N1000001O0O101O00001N1000001N100mNbLeA_3[>gL^AY3c>nLUAR3l>TMl@n2S?o01O1O100O1O010O1O100O1O00100O1O100O001O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O1O2O0O1O2O0WN[_O6g`0G[_O8f`0G\\_O8d`0G]_O8e`0TO\\_OB1Y1d`0cNS@M[O^1Ub001N1O101N1O101N1O2N100O2N100O2N1N3L3M3M4L4M9FV\\]4"}, "label": "bread sandwich in plate"}]} {"id": 68430, "image": "COCO_train2014_000000068430.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hoagie sandwich that has been cut in three\"."}], "task": "refering", "answer_template": "The \"a hoagie sandwich that has been cut in three\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [260, 261, 262, 277, 278, 279, 280, 294, 295, 296, 297, 307, 308, 309, 310, 311, 312, 313, 314, 324, 325, 326, 327, 328, 329, 330, 331, 342, 343, 344, 345, 346, 347, 361, 362], "bbox": [0.10233333333333333, 0.6856718749999999, 0.5267083333333333, 0.947515625], "iscrowd": 0, "area": 16755.671649999997, "rle": {"size": [640, 480], "counts": "gTo0;`c08H8J4M3M2O1N3M2O1N3M2N2O2O0O1010O1O00100O1O010O1O100O001O100O00100O1O1O010O1O1O010O1O1O10O0100000O100000001N101O1O1O0O2N2L3N3M3M3M2M4M2N1O2M2O2N1O1O2M2O2N1O1O2M2O2N1O1N3N1O2N1O1N2O00001N1000001O0O101O00001N1000001N100mNbLeA_3[>gL^AY3c>nLUAR3l>TMl@n2S?o01O1O100O1O010O1O100O1O00100O1O100O001O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O1O2O0O1O2O0WN[_O6g`0G[_O8f`0G\\_O8d`0G]_O8e`0TO\\_OB1Y1d`0cNS@M[O^1Ub001N1O101N1O101N1O2N100O2N100O2N1N3L3M3M4L4M9FV\\]4"}, "label": "a hoagie sandwich that has been cut in three"}]} {"id": 68430, "image": "COCO_train2014_000000068430.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a black outfit to the left of a girl in a black dress\"."}], "task": "refering", "answer_template": "The \"a person in a black outfit to the left of a girl in a black dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 17, 18, 34, 35, 51, 52, 68, 69, 85, 86, 102, 103, 119, 120, 136, 137, 138, 153, 154, 155, 170, 171, 172, 187, 188, 189], "bbox": [0.014979166666666667, 0.0, 0.1498125, 0.52809375], "iscrowd": 0, "area": 12928.822450000001, "rle": {"size": [640, 480], "counts": "P\\4^2ca0Z7eHh0YOO001O00100O001VNi1101N10001N101N101N101N10001NmHVEQ5j:lJ^EP5`:mJiEn4W:oJQFm4m9QK[Fj4e9RKcFk4\\9RKlFi4S9UKTGg4l8UK[Gi4c8UKdGg4\\8VKjGh4j;M4M2M3M4M5J6K4K6K5J6J6I7H8I6I8I7H8I7H8I6I8I7I7H8I7H7J7H8IPfn7"}, "label": "a person in a black outfit to the left of a girl in a black dress"}]} {"id": 68430, "image": "COCO_train2014_000000068430.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in black next to the girl with flags\"."}], "task": "refering", "answer_template": "The \"person in black next to the girl with flags\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 17, 18, 34, 35, 51, 52, 68, 69, 85, 86, 102, 103, 119, 120, 136, 137, 138, 153, 154, 155, 170, 171, 172, 187, 188, 189], "bbox": [0.014979166666666667, 0.0, 0.1498125, 0.52809375], "iscrowd": 0, "area": 12928.822450000001, "rle": {"size": [640, 480], "counts": "P\\4^2ca0Z7eHh0YOO001O00100O001VNi1101N10001N101N101N101N10001NmHVEQ5j:lJ^EP5`:mJiEn4W:oJQFm4m9QK[Fj4e9RKcFk4\\9RKlFi4S9UKTGg4l8UK[Gi4c8UKdGg4\\8VKjGh4j;M4M2M3M4M5J6K4K6K5J6J6I7H8I6I8I7H8I7H8I6I8I7I7H8I7H7J7H8IPfn7"}, "label": "person in black next to the girl with flags"}]} {"id": 453475, "image": "COCO_train2014_000000453475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe which is standing on the right side of the other two giraffes\"."}], "task": "refering", "answer_template": "The \"a giraffe which is standing on the right side of the other two giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 64, 65, 66, 87, 88, 111, 134, 135, 157, 158, 180, 181, 182, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342], "bbox": [0.778390625, 0.07433255269320843, 1.0, 0.9639578454332554], "iscrowd": 0, "area": 19961.9797, "rle": {"size": [427, 640], "counts": "]g_62U=5N101N2N101N2O00001N100[CD[ in this image.", "objects": [{"patches": [41, 42, 64, 65, 66, 87, 88, 111, 134, 135, 157, 158, 180, 181, 182, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342], "bbox": [0.778390625, 0.07433255269320843, 1.0, 0.9639578454332554], "iscrowd": 0, "area": 19961.9797, "rle": {"size": [427, 640], "counts": "]g_62U=5N101N2N101N2O00001N100[CD[ in this image.", "objects": [{"patches": [175, 176, 177, 199, 200, 222, 223, 224, 246, 247, 248, 269, 270, 271], "bbox": [0.64, 0.4741217798594847, 0.8229843750000001, 0.7888992974238876], "iscrowd": 0, "area": 4664.5828, "rle": {"size": [427, 640], "counts": "mU[52W=2O1O2N100O100O1O1O010O1O1O010O1O1O1O1O1O1O1O1O201N1O1O[O`Cf0Z in this image.", "objects": [{"patches": [175, 176, 177, 199, 200, 222, 223, 224, 246, 247, 248, 269, 270, 271], "bbox": [0.64, 0.4741217798594847, 0.8229843750000001, 0.7888992974238876], "iscrowd": 0, "area": 4664.5828, "rle": {"size": [427, 640], "counts": "mU[52W=2O1O2N100O100O1O1O010O1O1O010O1O1O1O1O1O1O1O1O201N1O1O[O`Cf0Z in this image.", "objects": [{"patches": [127, 128, 129, 139, 140, 141, 142, 152, 153, 154, 155, 164, 165, 166, 167, 168, 177, 178, 179, 180, 181, 189, 190, 191, 192, 193, 194, 203, 204, 205, 206, 217, 218, 219], "bbox": [0.5865122615803815, 0.51032, 0.998882833787466, 0.91614], "iscrowd": 0, "area": 18739.392200000002, "rle": {"size": [500, 367], "counts": "S\\Y32`?5K5K3M3N2M2N3M3M2N3M3N2M2N3M3M3M2N3N2M2N3M3M3M2N3M3N2M2N3M3M2N3M3N2M2N3M3M2N3M3N2M2N3M3M3M2N3M2O1N2N3M2N2N2N2O1N2N2N3M2N2N2N2O1N2N3M2N2N2N2O1N2N2N3M2O2N1O2N1O2N1O2N1O2N1O2N02O0O1O2O0O2O0O2N101N1O2N1N2N3N1N3M2N3M2N3M2N3M2O2M2N2N3M2N3M2N3M2O2M2N3M2N2N3M2N3N1N3M2N3M2N3M2N2N3N1N3M2N3M2N3M2N3N1N2I8^Oa0_Ob0^Oa0_OeG"}, "label": "the device with a a macbreak weekly screen on it"}]} {"id": 215908, "image": "COCO_train2014_000000215908.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"macbreak weekly screen using the phone\"."}], "task": "refering", "answer_template": "The \"macbreak weekly screen using the phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 139, 140, 141, 142, 152, 153, 154, 155, 164, 165, 166, 167, 168, 177, 178, 179, 180, 181, 189, 190, 191, 192, 193, 194, 203, 204, 205, 206, 217, 218, 219], "bbox": [0.5865122615803815, 0.51032, 0.998882833787466, 0.91614], "iscrowd": 0, "area": 18739.392200000002, "rle": {"size": [500, 367], "counts": "S\\Y32`?5K5K3M3N2M2N3M3M2N3M3N2M2N3M3M3M2N3N2M2N3M3M3M2N3M3N2M2N3M3M2N3M3N2M2N3M3M2N3M3N2M2N3M3M3M2N3M2O1N2N3M2N2N2N2O1N2N2N3M2N2N2N2O1N2N3M2N2N2N2O1N2N2N3M2O2N1O2N1O2N1O2N1O2N1O2N02O0O1O2O0O2O0O2N101N1O2N1N2N3N1N3M2N3M2N3M2N3M2O2M2N2N3M2N3M2N3M2O2M2N3M2N2N3M2N3N1N3M2N3M2N3M2N2N3N1N3M2N3M2N3M2N3N1N2I8^Oa0_Ob0^Oa0_OeG"}, "label": "macbreak weekly screen using the phone"}]} {"id": 215908, "image": "COCO_train2014_000000215908.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a flip phone displaying the time 7 : 45\"."}], "task": "refering", "answer_template": "The \"a flip phone displaying the time 7 : 45\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 30, 31, 32, 43, 44, 45, 56, 57, 58, 69, 70, 71, 82, 83, 84, 96, 97], "bbox": [0.30591280653950953, 0.0587, 0.5283378746594005, 0.40494], "iscrowd": 0, "area": 10290.3115, "rle": {"size": [500, 367], "counts": "Uhf12a?8I7VAKg=a1_O6K4K6K4L4K6K4K6K4L4K6K2N3L3N2N2N2N1O0O2O1O1O1O1O1O001N101O00001O001O000O100000000000000O1000O100000VOnKRFR4m9k0000000000O010O100O10O06K6I7J6I7J7kKbDe3l;I6K6I7J5K6I;F<@?^OY[d2"}, "label": "a flip phone displaying the time 7 : 45"}]} {"id": 215908, "image": "COCO_train2014_000000215908.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver cell phone\"."}], "task": "refering", "answer_template": "The \"a silver cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 30, 31, 32, 43, 44, 45, 56, 57, 58, 69, 70, 71, 82, 83, 84, 96, 97], "bbox": [0.30591280653950953, 0.0587, 0.5283378746594005, 0.40494], "iscrowd": 0, "area": 10290.3115, "rle": {"size": [500, 367], "counts": "Uhf12a?8I7VAKg=a1_O6K4K6K4L4K6K4K6K4L4K6K2N3L3N2N2N2N1O0O2O1O1O1O1O1O001N101O00001O001O000O100000000000000O1000O100000VOnKRFR4m9k0000000000O010O100O10O06K6I7J6I7J7kKbDe3l;I6K6I7J5K6I;F<@?^OY[d2"}, "label": "a silver cell phone"}]} {"id": 166762, "image": "COCO_train2014_000000166762.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe with its head disappearing behind a tree\"."}], "task": "refering", "answer_template": "The \"a giraffe with its head disappearing behind a tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 170, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 299, 300, 302, 322, 324, 325], "bbox": [0.0015, 0.5058644859813084, 0.40367187499999996, 1.0], "iscrowd": 0, "area": 17007.111599999997, "rle": {"size": [428, 640], "counts": "_g0`1c0F_9;aFE^9dFB[9?eFAZ9`0fF@Z9`0fF@Y9a0gF_OX9b0hF^OW9c0iF]OV9d0jFZOW9g0iFWOY9i0gFTO[9m0eFQO]9o0cFnN_9S1aFkNa9U1_FiNb9W1_FgNc9Y1]FeNe9[1[FcNf9^1ZF`Nh9`1XF^Ni9c1WF[Nk9e1UFYNl9h1TFVNn9j1RFTNo9m1QFQNQ:o1f00O100O100O10000O100O100]EkMn9V2RFnMi9S2WFQNd9P2\\FTN`9l1`FWN\\9j1eFUNZ9l1lFmMT9T2RGfMm8[2Q1100O1O1O1O100O1O1O1O100O1O1O1O1O1O1O1N2O1N2O1O1O10000O100O10000O1`FkMi7U2WHkMh7V2XHjMh7V2XHjMh7V2XHkMf7V2YHkMg7U2YHkMg7U2YHkMf7V2ZHjMf7V2ZHjMf7V2ZHjMf7V2ZHjMf7V2ZHkMe7U2[HkMe7U2[HkMd7V2\\HjMd7V2\\HjMd7V2\\HjMd7V2\\HjMd7V2\\HjMd7V2\\HkMb7V2^HjMb7V2^HjMb7V2^HjMb7V2^HgMe7Y2[H_Mm7a2SH[MP8f2PHUMU8k2kGQMY8o2gGlL^8S3cGhLa8X3`GdLd8[3g0O001O1O001O1N1O2hNlEVOV:h0nESOT:m0oEoNS:P1PFlNR:R1RFjNo9V1TFbNR:]1o0O0O2O1O001O1O1O010O10O0100O00100O010O1O010O10O01O10O0100O010O00100O0010O0100O00100O0010O0100O00100O0010O0100O0010O01O10O0100O0010O01O10O01O10OKgCXOZ in this image.", "objects": [{"patches": [168, 169, 170, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 299, 300, 302, 322, 324, 325], "bbox": [0.0015, 0.5058644859813084, 0.40367187499999996, 1.0], "iscrowd": 0, "area": 17007.111599999997, "rle": {"size": [428, 640], "counts": "_g0`1c0F_9;aFE^9dFB[9?eFAZ9`0fF@Z9`0fF@Y9a0gF_OX9b0hF^OW9c0iF]OV9d0jFZOW9g0iFWOY9i0gFTO[9m0eFQO]9o0cFnN_9S1aFkNa9U1_FiNb9W1_FgNc9Y1]FeNe9[1[FcNf9^1ZF`Nh9`1XF^Ni9c1WF[Nk9e1UFYNl9h1TFVNn9j1RFTNo9m1QFQNQ:o1f00O100O100O10000O100O100]EkMn9V2RFnMi9S2WFQNd9P2\\FTN`9l1`FWN\\9j1eFUNZ9l1lFmMT9T2RGfMm8[2Q1100O1O1O1O100O1O1O1O100O1O1O1O1O1O1O1N2O1N2O1O1O10000O100O10000O1`FkMi7U2WHkMh7V2XHjMh7V2XHjMh7V2XHkMf7V2YHkMg7U2YHkMg7U2YHkMf7V2ZHjMf7V2ZHjMf7V2ZHjMf7V2ZHjMf7V2ZHkMe7U2[HkMe7U2[HkMd7V2\\HjMd7V2\\HjMd7V2\\HjMd7V2\\HjMd7V2\\HjMd7V2\\HkMb7V2^HjMb7V2^HjMb7V2^HjMb7V2^HgMe7Y2[H_Mm7a2SH[MP8f2PHUMU8k2kGQMY8o2gGlL^8S3cGhLa8X3`GdLd8[3g0O001O1O001O1N1O2hNlEVOV:h0nESOT:m0oEoNS:P1PFlNR:R1RFjNo9V1TFbNR:]1o0O0O2O1O001O1O1O010O10O0100O00100O010O1O010O10O01O10O0100O010O00100O0010O0100O00100O0010O0100O00100O0010O0100O0010O01O10O0100O0010O01O10O01O10OKgCXOZ in this image.", "objects": [{"patches": [151, 173, 174, 175, 196, 197, 198, 199, 221, 222, 223, 224, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 338, 339, 341, 342], "bbox": [0.5285624999999999, 0.43717289719626173, 0.9122343749999999, 0.9887616822429908], "iscrowd": 0, "area": 19692.41375, "rle": {"size": [428, 640], "counts": "Q`]41Z=2N3M2N2N3M2N100O1O010O1O100O100O0O2M3M3N2O100O1O1O100O1O100N2O1N101O1N2O1O1N2O1O0O2001O001O000fN^Dj0R<2N3M2N2N2M2O2N2N2N2N1O2M3N2N2N1OVNPEZ1n:aNXE`1Z;1O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O001O1O00h0XO9G5J5L4K5L4K6KO001O00001O00[GoKP8P4PHWLi7i3WHZLg7e3XH\\Lh7d3XH\\Li7c3WH]Li7c3WH]Li7c3WH]Lj7b3VH^Lj7b3VH^Lk7a3UH_Lk7a3UH_Ll7`3TH`Ll7`3TH`Lm7_3SH`Ln7`3RH`Ln7`3RH`Lo7_3QHaLo7_3QHaLP8^3PHbLP8^3PHbLQ8]3oGcLQ8]3oGcLR8\\3nGdLR8\\3nGdLR8\\3nGdLS8[3mGeLS8[3lGfLU8Y3kGgLU8Y3kGgLV8X3jGhLV8X3jGhLW8W3iGhLX8X3hGhLX8X3hGhLY8W3gGiLY8W3gGiLZ8V3fGjLZ8V3fGjL[8U3eGkL\\8T3dGlL]8S3cGmL^8R3bGnL^8R3bGnL_8Q3aGoL`8P3aGoL`8P3bGnL_8Q3cGmL^8R3cGmL^8R3dGkL^8S3dGlL\\8T3o0O1O1O1O000000001O0000001O0000001O000000001O0000001O0001O01O000000SFUMY9k2gFWMX9h2hFYMW9g2iFZMV9f2jF\\MU9c2kF^MT9b2lF`MS9_2mFbMR9^2nFcMQ9]2oFcMR9\\2mFeMS9[2mFeMT9[2kFeMU9[2kFeMV9Z2jFfMV9Z2jFeMW9[2iFeMX9Z2hFfMX9Z2hFfMY9Y2gFgMY9Y2gFgMZ9X2fFhMZ9X2fFhM[9W2eFiM[9W2eFiM\\9V2dFiM]9W2cFiM^9V2bFjM^9V2cFiM^9U2cFkM^9T2cFkM]9T2eFkM\\9T2dFlM\\9S2fFlM[9f1RGZNn8a1XG^Ni8\\1]GbNe8X1j1K77H`0AO1O1O3M4L4L4LT1lN6JjPg0"}, "label": "animal on the right"}]} {"id": 166762, "image": "COCO_train2014_000000166762.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lighter giraffe whose face is visible\"."}], "task": "refering", "answer_template": "The \"the lighter giraffe whose face is visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 173, 174, 175, 196, 197, 198, 199, 221, 222, 223, 224, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 338, 339, 341, 342], "bbox": [0.5285624999999999, 0.43717289719626173, 0.9122343749999999, 0.9887616822429908], "iscrowd": 0, "area": 19692.41375, "rle": {"size": [428, 640], "counts": "Q`]41Z=2N3M2N2N3M2N100O1O010O1O100O100O0O2M3M3N2O100O1O1O100O1O100N2O1N101O1N2O1O1N2O1O0O2001O001O000fN^Dj0R<2N3M2N2N2M2O2N2N2N2N1O2M3N2N2N1OVNPEZ1n:aNXE`1Z;1O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O001O1O00h0XO9G5J5L4K5L4K6KO001O00001O00[GoKP8P4PHWLi7i3WHZLg7e3XH\\Lh7d3XH\\Li7c3WH]Li7c3WH]Li7c3WH]Lj7b3VH^Lj7b3VH^Lk7a3UH_Lk7a3UH_Ll7`3TH`Ll7`3TH`Lm7_3SH`Ln7`3RH`Ln7`3RH`Lo7_3QHaLo7_3QHaLP8^3PHbLP8^3PHbLQ8]3oGcLQ8]3oGcLR8\\3nGdLR8\\3nGdLR8\\3nGdLS8[3mGeLS8[3lGfLU8Y3kGgLU8Y3kGgLV8X3jGhLV8X3jGhLW8W3iGhLX8X3hGhLX8X3hGhLY8W3gGiLY8W3gGiLZ8V3fGjLZ8V3fGjL[8U3eGkL\\8T3dGlL]8S3cGmL^8R3bGnL^8R3bGnL_8Q3aGoL`8P3aGoL`8P3bGnL_8Q3cGmL^8R3cGmL^8R3dGkL^8S3dGlL\\8T3o0O1O1O1O000000001O0000001O0000001O000000001O0000001O0001O01O000000SFUMY9k2gFWMX9h2hFYMW9g2iFZMV9f2jF\\MU9c2kF^MT9b2lF`MS9_2mFbMR9^2nFcMQ9]2oFcMR9\\2mFeMS9[2mFeMT9[2kFeMU9[2kFeMV9Z2jFfMV9Z2jFeMW9[2iFeMX9Z2hFfMX9Z2hFfMY9Y2gFgMY9Y2gFgMZ9X2fFhMZ9X2fFhM[9W2eFiM[9W2eFiM\\9V2dFiM]9W2cFiM^9V2bFjM^9V2cFiM^9U2cFkM^9T2cFkM]9T2eFkM\\9T2dFlM\\9S2fFlM[9f1RGZNn8a1XG^Ni8\\1]GbNe8X1j1K77H`0AO1O1O3M4L4L4LT1lN6JjPg0"}, "label": "the lighter giraffe whose face is visible"}]} {"id": 322411, "image": "COCO_train2014_000000322411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of pasta next to a bowl of salad\"."}], "task": "refering", "answer_template": "The \"a bowl of pasta next to a bowl of salad\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 109, 110, 111, 112, 113, 114, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 181, 182, 183, 184, 185, 186, 199, 200, 201, 202, 203, 204, 219, 220], "bbox": [0.038759999999999996, 0.4426933333333333, 0.41124, 0.9415733333333334], "iscrowd": 0, "area": 27111.970599999993, "rle": {"size": [375, 500], "counts": "aV72c;?@?A;E5K5K5K5L4K5K5K5K5K5K5L4K5K5M2N2N2N2N2N3M2N2N2N2N2N3M2N2N2N2N2N3M2N1O2N1O2N1O1O2O0O2O001O0O101O001N10001O0O2O001O0O101O001N101O0O101O001N10001O0O1000000O1000000O1000000O100000000000000000000001O0000000000000000000000000O100O100O100O101N100O100O100O2O0O2O0O2O1N101N101N101N101N2O0O2O0O2O0O2O0N3N2M2O2M2O2N1N3N2M2O2M3N2M3N2M3N3L3N2M3N2M3N4K:G7H6J5J7I7I7I7I6K6I7I7I7IRi[3"}, "label": "a bowl of pasta next to a bowl of salad"}]} {"id": 322411, "image": "COCO_train2014_000000322411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bowl with noodles in it\"."}], "task": "refering", "answer_template": "The \"the bowl with noodles in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 109, 110, 111, 112, 113, 114, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 181, 182, 183, 184, 185, 186, 199, 200, 201, 202, 203, 204, 219, 220], "bbox": [0.038759999999999996, 0.4426933333333333, 0.41124, 0.9415733333333334], "iscrowd": 0, "area": 27111.970599999993, "rle": {"size": [375, 500], "counts": "aV72c;?@?A;E5K5K5K5L4K5K5K5K5K5K5L4K5K5M2N2N2N2N2N3M2N2N2N2N2N3M2N2N2N2N2N3M2N1O2N1O2N1O1O2O0O2O001O0O101O001N10001O0O2O001O0O101O001N101O0O101O001N10001O0O1000000O1000000O1000000O100000000000000000000001O0000000000000000000000000O100O100O100O101N100O100O100O2O0O2O0O2O1N101N101N101N101N2O0O2O0O2O0O2O0N3N2M2O2M2O2N1N3N2M2O2M3N2M3N2M3N3L3N2M3N2M3N4K:G7H6J5J7I7I7I7I6K6I7I7I7IRi[3"}, "label": "the bowl with noodles in it"}]} {"id": 322411, "image": "COCO_train2014_000000322411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of onion cooked broccoli and tomatoes sitting next to other food in bowls\"."}], "task": "refering", "answer_template": "The \"a bowl of onion cooked broccoli and tomatoes sitting next to other food in bowls\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 22, 23, 24, 25, 26, 27, 28, 29, 40, 41, 42, 43, 44, 45, 46, 47, 58, 59, 60, 61, 62, 63, 64, 65, 76, 77, 78, 79, 80, 81, 82, 83, 95, 96, 97, 98, 99, 100, 114, 115, 116, 117], "bbox": [0.22922, 0.0, 0.6438200000000001, 0.5213599999999999], "iscrowd": 0, "area": 29686.48015, "rle": {"size": [375, 500], "counts": "dVZ14a;5L4L4K5L4L4L3L5L4J6J6J6J6J5K4K6K4L4L4L4L4L4L4N2N2O1N2N2N2N2N2N2O1N2N2N2N2N3M2O1N2N2N2N2N2O1N2O1O1N2O1N2O1O1N10001N100O2O001N100O2O000O2O0O101O0O2O000O2O0O101O0O101O001O0O101O00001O0O101O001OO01000000000O0100000000O0100000000O0100000000O01000001O001N10001O001O0O2O00001O001N101O00001N101O001N100O2N101N1O2O0O2O0O1O2O0O2N101N101N2N2O1N1O2O1N2O1N2N2O1M2N3N2M3M3M3N2M2N3M3M3N2M3M3M4L5L5J5K5K6J5L4K6J5J7I6I7J7I6J6I8IPSQ2"}, "label": "a bowl of onion cooked broccoli and tomatoes sitting next to other food in bowls"}]} {"id": 322411, "image": "COCO_train2014_000000322411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cup of broccoli and tomatoes\"."}], "task": "refering", "answer_template": "The \"a cup of broccoli and tomatoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 22, 23, 24, 25, 26, 27, 28, 29, 40, 41, 42, 43, 44, 45, 46, 47, 58, 59, 60, 61, 62, 63, 64, 65, 76, 77, 78, 79, 80, 81, 82, 83, 95, 96, 97, 98, 99, 100, 114, 115, 116, 117], "bbox": [0.22922, 0.0, 0.6438200000000001, 0.5213599999999999], "iscrowd": 0, "area": 29686.48015, "rle": {"size": [375, 500], "counts": "dVZ14a;5L4L4K5L4L4L3L5L4J6J6J6J6J5K4K6K4L4L4L4L4L4L4N2N2O1N2N2N2N2N2N2O1N2N2N2N2N3M2O1N2N2N2N2N2O1N2O1O1N2O1N2O1O1N10001N100O2O001N100O2O000O2O0O101O0O2O000O2O0O101O0O101O001O0O101O00001O0O101O001OO01000000000O0100000000O0100000000O0100000000O01000001O001N10001O001O0O2O00001O001N101O00001N101O001N100O2N101N1O2O0O2O0O1O2O0O2N101N101N2N2O1N1O2O1N2O1N2N2O1M2N3N2M3M3M3N2M2N3M3M3N2M3M3M4L5L5J5K5K6J5L4K6J5J7I6I7J7I6J6I8IPSQ2"}, "label": "a cup of broccoli and tomatoes"}]} {"id": 322411, "image": "COCO_train2014_000000322411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bowl with banana slices , raisins , and cookies\"."}], "task": "refering", "answer_template": "The \"bowl with banana slices , raisins , and cookies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 119, 120, 134, 135, 136, 137, 138, 139, 151, 152, 153, 154, 155, 156, 157, 158, 169, 170, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 192, 193, 194, 205, 206, 207, 208, 209, 210, 211, 224, 225, 226, 227, 228, 229], "bbox": [0.41798, 0.47416, 0.7955, 0.9887733333333333], "iscrowd": 0, "area": 28398.44815, "rle": {"size": [375, 500], "counts": "li\\26Z;=D in this image.", "objects": [{"patches": [116, 117, 118, 119, 120, 134, 135, 136, 137, 138, 139, 151, 152, 153, 154, 155, 156, 157, 158, 169, 170, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 192, 193, 194, 205, 206, 207, 208, 209, 210, 211, 224, 225, 226, 227, 228, 229], "bbox": [0.41798, 0.47416, 0.7955, 0.9887733333333333], "iscrowd": 0, "area": 28398.44815, "rle": {"size": [375, 500], "counts": "li\\26Z;=D in this image.", "objects": [{"patches": [31, 32, 33, 47, 48, 49, 50, 51, 52, 65, 66, 67, 68, 69, 70, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124], "bbox": [0.6236, 0.12808, 0.9438200000000001, 0.54608], "iscrowd": 0, "area": 19326.6013, "rle": {"size": [375, 500], "counts": "e\\b35];7J7H7I8I6I8H7I8H7I7I8K4M2M3N1N3N1N3N2M2O2M3N1N3N2M2O2N2M201O1N101O1N101O1N101O1N101O1N101O1N10000O2O000O10001N10000O101O0O10001N10000O101O000000001O0000000010O0N2N2000000O1000000O100000000O1000001N1000000O1000001O1N101O001N101O1O0O2O001N101N1O2N2O0O2N1O2O0O2N1O2O1N1O2O0O2N2N3N2M3M3M3N1M4L4M3L4L4L4L4L4L4L4L8I9F;E:F;EjS:"}, "label": "bowl of watermelon"}]} {"id": 322411, "image": "COCO_train2014_000000322411.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of cut watermelon chunks\"."}], "task": "refering", "answer_template": "The \"a bowl of cut watermelon chunks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 47, 48, 49, 50, 51, 52, 65, 66, 67, 68, 69, 70, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124], "bbox": [0.6236, 0.12808, 0.9438200000000001, 0.54608], "iscrowd": 0, "area": 19326.6013, "rle": {"size": [375, 500], "counts": "e\\b35];7J7H7I8I6I8H7I8H7I7I8K4M2M3N1N3N1N3N2M2O2M3N1N3N2M2O2N2M201O1N101O1N101O1N101O1N101O1N101O1N10000O2O000O10001N10000O101O0O10001N10000O101O000000001O0000000010O0N2N2000000O1000000O100000000O1000001N1000000O1000001O1N101O001N101O1O0O2O001N101N1O2N2O0O2N1O2O0O2N1O2O1N1O2O0O2N2N3N2M3M3M3N1M4L4M3L4L4L4L4L4L4L4L8I9F;E:F;EjS:"}, "label": "a bowl of cut watermelon chunks"}]} {"id": 240500, "image": "COCO_train2014_000000240500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden bench with a giant white bird standing on it\"."}], "task": "refering", "answer_template": "The \"a wooden bench with a giant white bird standing on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 243, 244, 245, 246, 247, 251, 266, 267, 268, 269, 273, 276, 288, 289, 290, 295, 296, 297, 298, 299, 310, 311, 312, 317, 318, 319, 320, 321, 332, 333, 334, 339, 340, 341, 342, 343, 354, 355, 356, 361, 362, 364, 365, 376, 377, 378, 379, 380, 383, 384, 398, 399, 400, 401, 402], "bbox": [0.07600323624595469, 0.45884375000000005, 0.6149514563106796, 0.8036406250000001], "iscrowd": 0, "area": 23706.33515, "rle": {"size": [640, 618], "counts": "[em02mc07J6J5K6I6K4L4oAUOe8o0WGQOh8T1TGlNl8X1oFiNQ9[1kFeNT9`1hF`NX9d1cF]N]9g1_FZN`9j1[FWNd9n1XFRNh9R2TFnMl9V2oEkMQ:Y2kEgMV:]2dEdM^:`2\\E`Mf:c2UE]Mm:g2lDZMU;k2eDUM];n2]DSMe;Q3nCWMS4nANXO`17ROb>]OSBOXOi2e>SMXB2WOj2_>nLbB6ROj2j?UMY@j2g?SM]@k2d?SM_@l2`?SMd@j2[`0O1N2O1N2OM2O2N1O2M2O2N1O2N1N2OVNd_O3Z`0MV@Di?=e@VOX?l0WAeNh>\\1m1001N22M4M3M3M3L4M3M2N0O101O000O2O1O1O0O2O1OmN_Oh^O`0Wa0Hc^O6]a02]^OMba0;W^OEha0=X^OBga0?Z^O_Ofa0c0Y^O]Ofa0d0Z^O\\Oea0f0Z^OYOfa0h0Z^OXOea0i0[^OWOda0k0[^OUOda0l0\\^OSOda0o0[^OQOda0P1\\^OPOca0R1[^OoNda0R1Z^OoNfa0R1Y^OoNfa0S1W^OoNha0R1V^OPOja0Q1T^OoNla0c1O1O1O001Y_OoM^?R2U@[Nj?f1i_OgNV`0b2O110O01O00001O010O00h0XO3M4M2M4L3M4L3M4M2M4L3M4L=Dg0XOnYd4"}, "label": "a wooden bench with a giant white bird standing on it"}]} {"id": 240500, "image": "COCO_train2014_000000240500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bench with a large bird on it and a person in a long black coat in front of it\"."}], "task": "refering", "answer_template": "The \"bench with a large bird on it and a person in a long black coat in front of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 243, 244, 245, 246, 247, 251, 266, 267, 268, 269, 273, 276, 288, 289, 290, 295, 296, 297, 298, 299, 310, 311, 312, 317, 318, 319, 320, 321, 332, 333, 334, 339, 340, 341, 342, 343, 354, 355, 356, 361, 362, 364, 365, 376, 377, 378, 379, 380, 383, 384, 398, 399, 400, 401, 402], "bbox": [0.07600323624595469, 0.45884375000000005, 0.6149514563106796, 0.8036406250000001], "iscrowd": 0, "area": 23706.33515, "rle": {"size": [640, 618], "counts": "[em02mc07J6J5K6I6K4L4oAUOe8o0WGQOh8T1TGlNl8X1oFiNQ9[1kFeNT9`1hF`NX9d1cF]N]9g1_FZN`9j1[FWNd9n1XFRNh9R2TFnMl9V2oEkMQ:Y2kEgMV:]2dEdM^:`2\\E`Mf:c2UE]Mm:g2lDZMU;k2eDUM];n2]DSMe;Q3nCWMS4nANXO`17ROb>]OSBOXOi2e>SMXB2WOj2_>nLbB6ROj2j?UMY@j2g?SM]@k2d?SM_@l2`?SMd@j2[`0O1N2O1N2OM2O2N1O2M2O2N1O2N1N2OVNd_O3Z`0MV@Di?=e@VOX?l0WAeNh>\\1m1001N22M4M3M3M3L4M3M2N0O101O000O2O1O1O0O2O1OmN_Oh^O`0Wa0Hc^O6]a02]^OMba0;W^OEha0=X^OBga0?Z^O_Ofa0c0Y^O]Ofa0d0Z^O\\Oea0f0Z^OYOfa0h0Z^OXOea0i0[^OWOda0k0[^OUOda0l0\\^OSOda0o0[^OQOda0P1\\^OPOca0R1[^OoNda0R1Z^OoNfa0R1Y^OoNfa0S1W^OoNha0R1V^OPOja0Q1T^OoNla0c1O1O1O001Y_OoM^?R2U@[Nj?f1i_OgNV`0b2O110O01O00001O010O00h0XO3M4M2M4L3M4L3M4M2M4L3M4L=Dg0XOnYd4"}, "label": "bench with a large bird on it and a person in a long black coat in front of it"}]} {"id": 240500, "image": "COCO_train2014_000000240500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a swan on a bench\"."}], "task": "refering", "answer_template": "The \"a swan on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 165, 166, 188, 209, 210, 230, 231, 232, 251, 252, 253, 254, 255, 273, 274, 275, 276, 295, 296, 297, 319, 320], "bbox": [0.4346116504854369, 0.29114062500000004, 0.5945631067961165, 0.645671875], "iscrowd": 0, "area": 9756.564550000003, "rle": {"size": [640, 618], "counts": "n^X56cc08K4M3N2N2N2N3M2N3j]OSOUa0o0e^OXOYa0j0e^OWOZa0k0d^OWOYa0o0c^ORO[a0R1a^OPO]a0U1^^OlN`a0l1N2NO001N11O1O1O101N1O1O1O2O0001O010O0010O01O0010O1O0a_OYMg?g2X@[Mf?f2X@\\Mg?j2S@WMm?o2l_OSMR`0T3g_OmLX`0^301N1O100OSOm_OnMS`0S2P@jMP`0U2T@hMk?X2Z@dMe?]2^@`Ma?`2c@\\M]?d2g@ZMV?h2k@XMR?i2o@XMo>h2RAYMk>i2UAXMh>i2YAXMc>j2^AXMcUMXZMbB[1_=aNhBY1]=bNhBZ1[=aNkB[1Y=_NmB]1m?M4L4L`0@fbl4"}, "label": "a swan on a bench"}]} {"id": 240500, "image": "COCO_train2014_000000240500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white pelican standing on top of a bench\"."}], "task": "refering", "answer_template": "The \"white pelican standing on top of a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 165, 166, 188, 209, 210, 230, 231, 232, 251, 252, 253, 254, 255, 273, 274, 275, 276, 295, 296, 297, 319, 320], "bbox": [0.4346116504854369, 0.29114062500000004, 0.5945631067961165, 0.645671875], "iscrowd": 0, "area": 9756.564550000003, "rle": {"size": [640, 618], "counts": "n^X56cc08K4M3N2N2N2N3M2N3j]OSOUa0o0e^OXOYa0j0e^OWOZa0k0d^OWOYa0o0c^ORO[a0R1a^OPO]a0U1^^OlN`a0l1N2NO001N11O1O1O101N1O1O1O2O0001O010O0010O01O0010O1O0a_OYMg?g2X@[Mf?f2X@\\Mg?j2S@WMm?o2l_OSMR`0T3g_OmLX`0^301N1O100OSOm_OnMS`0S2P@jMP`0U2T@hMk?X2Z@dMe?]2^@`Ma?`2c@\\M]?d2g@ZMV?h2k@XMR?i2o@XMo>h2RAYMk>i2UAXMh>i2YAXMc>j2^AXMcUMXZMbB[1_=aNhBY1]=bNhBZ1[=aNkB[1Y=_NmB]1m?M4L4L`0@fbl4"}, "label": "white pelican standing on top of a bench"}]} {"id": 240500, "image": "COCO_train2014_000000240500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a picture of an older man with a black tench coat on standing next to a bench\"."}], "task": "refering", "answer_template": "The \"a picture of an older man with a black tench coat on standing next to a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 160, 161, 162, 163, 181, 182, 183, 184, 185, 203, 204, 205, 206, 207, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 334, 335, 336, 337, 338, 339, 357, 358, 359, 360, 361, 379, 380, 381, 382, 383, 402, 403, 404, 405, 424, 425, 426, 427], "bbox": [0.2246116504854369, 0.2684375, 0.45155339805825245, 0.8588906250000001], "iscrowd": 0, "area": 38827.51394999999, "rle": {"size": [640, 618], "counts": "SYg2j0Yb0n0G9H7H9G9G9H7H9G9G9H7H9G9G9H8G8H9G9H8G8H9K:H;D in this image.", "objects": [{"patches": [139, 140, 141, 160, 161, 162, 163, 181, 182, 183, 184, 185, 203, 204, 205, 206, 207, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 334, 335, 336, 337, 338, 339, 357, 358, 359, 360, 361, 379, 380, 381, 382, 383, 402, 403, 404, 405, 424, 425, 426, 427], "bbox": [0.2246116504854369, 0.2684375, 0.45155339805825245, 0.8588906250000001], "iscrowd": 0, "area": 38827.51394999999, "rle": {"size": [640, 618], "counts": "SYg2j0Yb0n0G9H7H9G9G9H7H9G9G9H7H9G9G9H8G8H9G9H8G8H9K:H;D in this image.", "objects": [{"patches": [126, 127, 147, 148, 149, 150, 170, 171, 172, 192, 193, 194, 214, 215, 216, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 323, 324, 325, 326, 327, 345, 346, 347, 348, 349, 367, 368, 369, 370, 371, 391, 392, 413, 414, 415, 434, 435, 436, 437, 457, 458, 459], "bbox": [0.6832200647249191, 0.256703125, 0.8903398058252427, 0.9008593749999999], "iscrowd": 0, "area": 33398.0376, "rle": {"size": [640, 618], "counts": "XRX82mc06J6J6J5J7J6J6J6J5J7J6J6J6J5K6I7J3M2hKlLaGV3lLlL]:3bHS3mLoL^:2`HR3nLRM^:0`HP3nLUM_:0^Hm2oLXM`:O]Hk2oL[Mb:M[Hj2oL_Mb:KZHi2PMaMc:KVHU4h7oKRHT4n7oKkGU4U8nKdGV4\\8nK\\GV4d8lKWGW4i8jKQG[4o8fKeFf4Z9\\KZFn4f9SKnEX5R:jJgE[5Y:fJdE\\5\\:eJ`E^5P2fH[6[9eIhFX6X9hIjFV6V9jIkFU6V9jIkFU6U9kIkFU6U9kIkFU6U9kIlFT6T9lIlFT6T9lImFS6S9mImFS6S9mImFS6T9lImFS6S9mImFS6S9nImFQ6X9jIhFfNIh6d9]JcFjNMf6e9[J_FmN0e6f9YJZFQO2e6i9UJVFUO1f6n9PJQFYO2g6R:kIlE]O3h6V:fIgEA4i6Z:aIbEE4k6_:[I]EI5l6c:VIXEN5l6h:QISE26n6k:kHoD67o6Z;oHgDR7X;mHhDT7X;kHiDU7W;kHiDV7V;iHkDW7U;hHlDY7l;10001O001N101O00001N101O00001N1`DaH_:`7^EdH`:\\7^EhH`:X7^ElH`:U7]EoH`:R7_ERI^:o6_EUI_:k6_EYI_:g6_E]I_:d6^E`I_:a6_EdI^:]6_EgI_:o7N2N2N3L5L6J6J6J6I7cL\\ERMj:h2fEhL`:R3oE_LW:Z3YFWLm9c3cFlKd9m3lFdKZ9V4UG[KQ9_4^3I7J6J6I9H8D<[Oe0[Oe0[Oe0[OYTZ1"}, "label": "woman in coat standing outside"}]} {"id": 240500, "image": "COCO_train2014_000000240500.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady wearing a black and white coat , holding a purse\"."}], "task": "refering", "answer_template": "The \"a lady wearing a black and white coat , holding a purse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 147, 148, 149, 150, 170, 171, 172, 192, 193, 194, 214, 215, 216, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 323, 324, 325, 326, 327, 345, 346, 347, 348, 349, 367, 368, 369, 370, 371, 391, 392, 413, 414, 415, 434, 435, 436, 437, 457, 458, 459], "bbox": [0.6832200647249191, 0.256703125, 0.8903398058252427, 0.9008593749999999], "iscrowd": 0, "area": 33398.0376, "rle": {"size": [640, 618], "counts": "XRX82mc06J6J6J5J7J6J6J6J5J7J6J6J6J5K6I7J3M2hKlLaGV3lLlL]:3bHS3mLoL^:2`HR3nLRM^:0`HP3nLUM_:0^Hm2oLXM`:O]Hk2oL[Mb:M[Hj2oL_Mb:KZHi2PMaMc:KVHU4h7oKRHT4n7oKkGU4U8nKdGV4\\8nK\\GV4d8lKWGW4i8jKQG[4o8fKeFf4Z9\\KZFn4f9SKnEX5R:jJgE[5Y:fJdE\\5\\:eJ`E^5P2fH[6[9eIhFX6X9hIjFV6V9jIkFU6V9jIkFU6U9kIkFU6U9kIkFU6U9kIlFT6T9lIlFT6T9lImFS6S9mImFS6S9mImFS6T9lImFS6S9mImFS6S9nImFQ6X9jIhFfNIh6d9]JcFjNMf6e9[J_FmN0e6f9YJZFQO2e6i9UJVFUO1f6n9PJQFYO2g6R:kIlE]O3h6V:fIgEA4i6Z:aIbEE4k6_:[I]EI5l6c:VIXEN5l6h:QISE26n6k:kHoD67o6Z;oHgDR7X;mHhDT7X;kHiDU7W;kHiDV7V;iHkDW7U;hHlDY7l;10001O001N101O00001N101O00001N1`DaH_:`7^EdH`:\\7^EhH`:X7^ElH`:U7]EoH`:R7_ERI^:o6_EUI_:k6_EYI_:g6_E]I_:d6^E`I_:a6_EdI^:]6_EgI_:o7N2N2N3L5L6J6J6J6I7cL\\ERMj:h2fEhL`:R3oE_LW:Z3YFWLm9c3cFlKd9m3lFdKZ9V4UG[KQ9_4^3I7J6J6I9H8D<[Oe0[Oe0[Oe0[OYTZ1"}, "label": "a lady wearing a black and white coat , holding a purse"}]} {"id": 281464, "image": "COCO_train2014_000000281464.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bowl with the white stuff in it\"."}], "task": "refering", "answer_template": "The \"the bowl with the white stuff in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 72, 73, 74, 75, 76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 304, 305, 306], "bbox": [0.02190625, 0.16629166666666664, 0.4971875, 0.7752708333333334], "iscrowd": 0, "area": 68165.25699999998, "rle": {"size": [480, 640], "counts": "Si63j>7I7I7I6J7I7I6J7I7I7I6J7I7I6J7I7I7I6J7I7I4L3O2N2N2N101N2N1O2N2N2O0O2N2N1O2O1N2N1O2N2N101N2N2N1O2N2O0O2N2N2N1O2O1N1O2N2N2N101N2N1O2N2N2O0O2N2N1O2N2O1N1O2N2N1O2O1N1O1O2N1O101N10001N101N10001N100O2O000O2O0O2O000O2O0O101O0O101N10001N101N10001N100O2O000O2O0O101O0O2O0O101O0O101N10001N100O2O001N100O10000O100O10000O100000000001O0000000000001O01O0000000001O000000000000001O0000000000001O0000000000001O000000000O101N1O100O100O101N1O100O100O101N100O1O100O101N100O100O1O101N100O100O1O101N2O1N2O1N2O1N2N2O1N2N2N2O2M2N2N2N2O1N2N2N2O1N2N2N2N3N1N2N2N2N2O1N2N2N2N2O1N2N2N3N1N2N2N2K5J6J7I7I8I6I7I8H7I7I8I6K5L5J5L4L5J5L4Kf0[Okgf4"}, "label": "the bowl with the white stuff in it"}]} {"id": 281464, "image": "COCO_train2014_000000281464.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vanilla dessert with a cashew nose\"."}], "task": "refering", "answer_template": "The \"the vanilla dessert with a cashew nose\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 72, 73, 74, 75, 76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 304, 305, 306], "bbox": [0.02190625, 0.16629166666666664, 0.4971875, 0.7752708333333334], "iscrowd": 0, "area": 68165.25699999998, "rle": {"size": [480, 640], "counts": "Si63j>7I7I7I6J7I7I6J7I7I7I6J7I7I6J7I7I7I6J7I7I4L3O2N2N2N101N2N1O2N2N2O0O2N2N1O2O1N2N1O2N2N101N2N2N1O2N2O0O2N2N2N1O2O1N1O2N2N2N101N2N1O2N2N2O0O2N2N1O2N2O1N1O2N2N1O2O1N1O1O2N1O101N10001N101N10001N100O2O000O2O0O2O000O2O0O101O0O101N10001N101N10001N100O2O000O2O0O101O0O2O0O101O0O101N10001N100O2O001N100O10000O100O10000O100000000001O0000000000001O01O0000000001O000000000000001O0000000000001O0000000000001O000000000O101N1O100O100O101N1O100O100O101N100O1O100O101N100O100O1O101N100O100O1O101N2O1N2O1N2O1N2N2O1N2N2N2O2M2N2N2N2O1N2N2N2O1N2N2N2N3N1N2N2N2N2O1N2N2N2N2O1N2N2N3N1N2N2N2K5J6J7I7I8I6I7I8H7I7I8I6K5L5J5L4L5J5L4Kf0[Okgf4"}, "label": "the vanilla dessert with a cashew nose"}]} {"id": 281464, "image": "COCO_train2014_000000281464.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl right of another\"."}], "task": "refering", "answer_template": "The \"a bowl right of another\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 111, 112, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295], "bbox": [0.5397968750000001, 0.1874375, 0.987109375, 0.7300416666666666], "iscrowd": 0, "area": 57505.3196, "rle": {"size": [480, 640], "counts": "fmQ51h>;F;D;F:E in this image.", "objects": [{"patches": [83, 84, 85, 86, 87, 88, 105, 106, 107, 108, 109, 110, 111, 112, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 292, 293, 294, 295], "bbox": [0.5397968750000001, 0.1874375, 0.987109375, 0.7300416666666666], "iscrowd": 0, "area": 57505.3196, "rle": {"size": [480, 640], "counts": "fmQ51h>;F;D;F:E in this image.", "objects": [{"patches": [335, 336, 337, 338, 339, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385], "bbox": [0.5654375, 0.8316874999999999, 0.998859375, 0.9996875], "iscrowd": 0, "area": 11376.939600000005, "rle": {"size": [480, 640], "counts": "hSZ56f>8I7K2M4L4M2M4M3L3M3N2N2O1N2O0O2N2O1N2O0O2N101N101N1O101N101N1O100O100O100000000O1000000000000O100000001O000O100000000000000O100000000000000O10000001O0000001O0000001O00001O0000001O0000001O001O1O001O1O001O001O1O001O1O001O1O0O2N2N1O2N2N2N2N1O2M3M3M3N1N3M3M3M3O0O2O1N2N100O100O10000O1000000000000000000000O10O1000000000000000000000000O10000000000000000000O100000O10000000000000O1000000000O100000O10000000000000000O1000000000000000000O10000000000000000O10000000000000000O1000000000001O00000O100000000000001N100000000000000O100000000000000O100000000000001N1l1"}, "label": "the spoon on the bottom right under the bowl with the black stuff in it"}]} {"id": 281464, "image": "COCO_train2014_000000281464.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a spoon in front of a cup which has black cream in it\"."}], "task": "refering", "answer_template": "The \"a spoon in front of a cup which has black cream in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [335, 336, 337, 338, 339, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385], "bbox": [0.5654375, 0.8316874999999999, 0.998859375, 0.9996875], "iscrowd": 0, "area": 11376.939600000005, "rle": {"size": [480, 640], "counts": "hSZ56f>8I7K2M4L4M2M4M3L3M3N2N2O1N2O0O2N2O1N2O0O2N101N101N1O101N101N1O100O100O100000000O1000000000000O100000001O000O100000000000000O100000000000000O10000001O0000001O0000001O00001O0000001O0000001O001O1O001O1O001O001O1O001O1O001O1O0O2N2N1O2N2N2N2N1O2M3M3M3N1N3M3M3M3O0O2O1N2N100O100O10000O1000000000000000000000O10O1000000000000000000000000O10000000000000000000O100000O10000000000000O1000000000O100000O10000000000000000O1000000000000000000O10000000000000000O10000000000000000O1000000000001O00000O100000000000001N100000000000000O100000000000000O100000000000001N1l1"}, "label": "a spoon in front of a cup which has black cream in it"}]} {"id": 281464, "image": "COCO_train2014_000000281464.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"spoon on the left\"."}], "task": "refering", "answer_template": "The \"spoon on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [299, 322, 323, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 368, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.0005, 0.79725, 0.492484375, 0.9918750000000001], "iscrowd": 0, "area": 9688.170749999994, "rle": {"size": [480, 640], "counts": "d>8[<\\2I80O1O1O2O0O2N100O2N1O2O0O2N1O101N1O2O0O2N1O2O0O1O2O1N2N2N2O1N2O2M2O1N2N2O1N2O2M2O1N2O1N2O2M2O2N1O2N1O2M3O0100N000000000000000001O0000000000000000001O000000000000000001O000000000000000000000001O0000000001O0000000000000000001O000000000000000000001O0000000000000001O01O000000000000000000001O0000000000000000001O000000000000000000001O00000001O000001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O000000000000001O01O0000000000000000000001O000000000000000000000000000000000000001O000000000000000000000000000000000000001O000000000000000000000001O001N2O0O2O2M4M2MRmg4"}, "label": "spoon on the left"}]} {"id": 281464, "image": "COCO_train2014_000000281464.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the spoon the left\"."}], "task": "refering", "answer_template": "The \"the spoon the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [299, 322, 323, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 368, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.0005, 0.79725, 0.492484375, 0.9918750000000001], "iscrowd": 0, "area": 9688.170749999994, "rle": {"size": [480, 640], "counts": "d>8[<\\2I80O1O1O2O0O2N100O2N1O2O0O2N1O101N1O2O0O2N1O2O0O1O2O1N2N2N2O1N2O2M2O1N2N2O1N2O2M2O1N2O1N2O2M2O2N1O2N1O2M3O0100N000000000000000001O0000000000000000001O000000000000000001O000000000000000000000001O0000000001O0000000000000000001O000000000000000000001O0000000000000001O01O000000000000000000001O0000000000000000001O000000000000000000001O00000001O000001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O000000000000001O0000000000001O0000000000001O000000000000001O01O0000000000000000000001O000000000000000000000000000000000000001O000000000000000000000000000000000000001O000000000000000000000001O001N2O0O2O2M4M2MRmg4"}, "label": "the spoon the left"}]} {"id": 535418, "image": "COCO_train2014_000000535418.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person ' s shoulder and arm wearing blue and black stripes\"."}], "task": "refering", "answer_template": "The \"the person ' s shoulder and arm wearing blue and black stripes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 138, 139, 140, 161, 162, 163, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326, 345, 346, 347, 348, 349, 368, 369, 370, 371, 372], "bbox": [0.0, 0.28881496881496876, 0.201765625, 1.0], "iscrowd": 0, "area": 33259.593100000006, "rle": {"size": [481, 640], "counts": "e9[5\\4Z51O1O001O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O2N1O1O1O2N1O1O1O1O2N1O100O2N1O1O1O2N1O1O1O1O2N1O1O1O2N3M3M3M3M3M4L3M3M3M3M4L3M3M3M3M4L3M3M3M3M4L3M3M3M3M4L3M3M3M3M4L3iL^Fi0e9SObFg0a9VOdFe0_9XOgFb0]9YOiFb0Z9[OlF?W9^OoF in this image.", "objects": [{"patches": [115, 116, 138, 139, 140, 161, 162, 163, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326, 345, 346, 347, 348, 349, 368, 369, 370, 371, 372], "bbox": [0.0, 0.28881496881496876, 0.201765625, 1.0], "iscrowd": 0, "area": 33259.593100000006, "rle": {"size": [481, 640], "counts": "e9[5\\4Z51O1O001O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O2N1O1O1O2N1O1O1O1O2N1O100O2N1O1O1O2N1O1O1O1O2N1O1O1O2N3M3M3M3M3M4L3M3M3M3M4L3M3M3M3M4L3M3M3M3M4L3M3M3M3M4L3M3M3M3M4L3iL^Fi0e9SObFg0a9VOdFe0_9XOgFb0]9YOiFb0Z9[OlF?W9^OoF in this image.", "objects": [{"patches": [37, 38, 54, 55, 56, 57, 72, 73, 74, 75, 91, 92, 93, 94, 110, 111, 112], "bbox": [0.02866, 0.17528000000000002, 0.25642000000000004, 0.5370666666666666], "iscrowd": 0, "area": 5789.745899999998, "rle": {"size": [375, 500], "counts": "oW52X;>N1N2N3M2N3N1N3M2N2O2N101N101N101N2O1N2O1N2O1N2O1N2K5G8N3N1O1O2M2O2N101O00001O010O002N=C=C=CVLkHT3k71O100O1O101N01O000O2O0O2QOjG[NW8m0cHPO^7o0eHmN]7R1gHjNZ7U1iHhNX7W1kHeNW7Z1d1O1O1O002N1O2N1O2N2YOYE:h:DZE;h:BYE`0f:^O[Ee0k:1N2N3N1N2N2N3N1N2N4O3M2OL3M4M2M3M4L3N3L6J3M1O100O1O1O001N2N2O1N2N2OdQX4"}, "label": "a security guard sitting in a chair"}]} {"id": 568187, "image": "COCO_train2014_000000568187.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the officer is sitting in his chair next to the dalmation\"."}], "task": "refering", "answer_template": "The \"the officer is sitting in his chair next to the dalmation\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 54, 55, 56, 57, 72, 73, 74, 75, 91, 92, 93, 94, 110, 111, 112], "bbox": [0.02866, 0.17528000000000002, 0.25642000000000004, 0.5370666666666666], "iscrowd": 0, "area": 5789.745899999998, "rle": {"size": [375, 500], "counts": "oW52X;>N1N2N3M2N3N1N3M2N2O2N101N101N101N2O1N2O1N2O1N2O1N2K5G8N3N1O1O2M2O2N101O00001O010O002N=C=C=CVLkHT3k71O100O1O101N01O000O2O0O2QOjG[NW8m0cHPO^7o0eHmN]7R1gHjNZ7U1iHhNX7W1kHeNW7Z1d1O1O1O002N1O2N1O2N2YOYE:h:DZE;h:BYE`0f:^O[Ee0k:1N2N3N1N2N2N3N1N2N4O3M2OL3M4M2M3M4L3N3L6J3M1O100O1O1O001N2N2O1N2N2OdQX4"}, "label": "the officer is sitting in his chair next to the dalmation"}]} {"id": 338819, "image": "COCO_train2014_000000338819.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the ram on the left who is head butting another ram standing in the middle of the road\"."}], "task": "refering", "answer_template": "The \"the ram on the left who is head butting another ram standing in the middle of the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 237, 238, 239, 240, 241, 242, 257, 258, 259, 261, 262, 263, 284, 285], "bbox": [0.19878125, 0.4127916666666666, 0.5460625, 0.7515833333333333], "iscrowd": 0, "area": 16985.1573, "rle": {"size": [480, 640], "counts": "Wkk11d=]1L4M4L3L4M3`NWNdEk1P:bNlEa1R:cNjE^1U:eNhE\\1W:hNdE11BZ:`0bEL:AS:g0_EFc0Am9l0XEGP1ZOg9U2]FhMc9Y2`FcM_9_2`12N2N2N2N3L5M3M2M4M3M2M6K:F1N3N00POVEjMi:S2_EiMa:S2fEkMY:R2nEjMQ:U2UFgMk9X2XFiMe9U2^FlM`9R2cFmM]9R2fFmMY9Q2jFnMV9P2mFoMS9P2oFoMQ9P2RGnMn8R2SGmMm8S2TGlMl8U2UGiMk8W2WGgMi8Z2XGdMh8\\2ZGbMf8^2o10O0001O00010O00001O00010O00001O01O0001O01O01O0000010O000010O0001O01O0001O01O01O0000TEgMR9Y2kFkMS9U2kFnMT9T2hFZNm8h1nFZNQ9i1hF\\NX9g1aF]N_9e1ZFaNe9]30000000]OfE`LZ:Y3PFdLo9V3[FgLe9U3bFhL^9U3kFeLU9Y3Z1I7C=002N6J5TDbMl:[3N1O1O2N2N2N1OWN^E\\N5C]:P2cEjM?5n9Q2gFmMY9R2iFmMW9S2jFkMX9S2kFkMU9U2lFjMU9U2mFiMT9U2nFiMT9V2oFgMQ9Y2l1N2O1N2O1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O1N2O00SCcNX<\\1`CnN_ in this image.", "objects": [{"patches": [165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 237, 238, 239, 240, 241, 242, 257, 258, 259, 261, 262, 263, 284, 285], "bbox": [0.19878125, 0.4127916666666666, 0.5460625, 0.7515833333333333], "iscrowd": 0, "area": 16985.1573, "rle": {"size": [480, 640], "counts": "Wkk11d=]1L4M4L3L4M3`NWNdEk1P:bNlEa1R:cNjE^1U:eNhE\\1W:hNdE11BZ:`0bEL:AS:g0_EFc0Am9l0XEGP1ZOg9U2]FhMc9Y2`FcM_9_2`12N2N2N2N3L5M3M2M4M3M2M6K:F1N3N00POVEjMi:S2_EiMa:S2fEkMY:R2nEjMQ:U2UFgMk9X2XFiMe9U2^FlM`9R2cFmM]9R2fFmMY9Q2jFnMV9P2mFoMS9P2oFoMQ9P2RGnMn8R2SGmMm8S2TGlMl8U2UGiMk8W2WGgMi8Z2XGdMh8\\2ZGbMf8^2o10O0001O00010O00001O00010O00001O01O0001O01O01O0000010O000010O0001O01O0001O01O01O0000TEgMR9Y2kFkMS9U2kFnMT9T2hFZNm8h1nFZNQ9i1hF\\NX9g1aF]N_9e1ZFaNe9]30000000]OfE`LZ:Y3PFdLo9V3[FgLe9U3bFhL^9U3kFeLU9Y3Z1I7C=002N6J5TDbMl:[3N1O1O2N2N2N1OWN^E\\N5C]:P2cEjM?5n9Q2gFmMY9R2iFmMW9S2jFkMX9S2kFkMU9U2lFjMU9U2mFiMT9U2nFiMT9V2oFgMQ9Y2l1N2O1N2O1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O1N2O00SCcNX<\\1`CnN_ in this image.", "objects": [{"patches": [195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 289, 290, 292, 293, 294, 312, 313, 315, 316, 317, 338, 339, 340], "bbox": [0.490453125, 0.4794375, 0.7991406249999999, 0.8503333333333333], "iscrowd": 0, "area": 15335.327350000012, "rle": {"size": [480, 640], "counts": "a]c47g>6K2N2O1N2O1O2O0O10O10O01O1O1O2N1O1O100O00O2O0100O0001M2O1N3_B@_SF`0TO[Ok:L[Fa1g9UNcFi1_9nMjFP2T;O2N2N2WNQC[1a=F:G9I7I7IkRl1"}, "label": "the right goat of the ones butting heads"}]} {"id": 338819, "image": "COCO_train2014_000000338819.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white animal closest to the yellow line in the road\"."}], "task": "refering", "answer_template": "The \"white animal closest to the yellow line in the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 199, 200, 218, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 289, 290, 292, 293, 294, 312, 313, 315, 316, 317, 338, 339, 340], "bbox": [0.490453125, 0.4794375, 0.7991406249999999, 0.8503333333333333], "iscrowd": 0, "area": 15335.327350000012, "rle": {"size": [480, 640], "counts": "a]c47g>6K2N2O1N2O1O2O0O10O10O01O1O1O2N1O1O100O00O2O0100O0001M2O1N3_B@_SF`0TO[Ok:L[Fa1g9UNcFi1_9nMjFP2T;O2N2N2WNQC[1a=F:G9I7I7IkRl1"}, "label": "white animal closest to the yellow line in the road"}]} {"id": 240521, "image": "COCO_train2014_000000240521.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man is smiling with his couple and both are wearing white cloth\"."}], "task": "refering", "answer_template": "The \"a man is smiling with his couple and both are wearing white cloth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 72, 83, 84, 85, 86, 87, 88, 98, 99, 100, 101, 102, 103, 113, 114, 115, 116, 117, 118, 128, 129, 130, 131, 132, 133, 143, 144, 145, 146, 147, 148, 158, 159, 160, 161, 162, 163, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 203, 204, 205, 206, 207, 209, 218, 219, 220, 221, 222, 224, 233, 234, 235, 236, 237, 239, 248, 249, 250, 251, 254, 265, 266, 269, 283, 284, 297, 298, 299, 308, 309, 310, 311, 312, 313, 314, 323, 324, 325, 326, 327, 328, 329, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.539316037735849, 0.2, 1.0, 0.9887656249999999], "iscrowd": 0, "area": 60005.99044999998, "rle": {"size": [640, 424], "counts": "bg_47h56d72[HZ1R6nNmIY1FUN>j0KT1CWN:j02R1AXN8h06S1^O[N7c0:U1[O]N6`0>V1XO_N5=b0W1VO`N3;f0X1SOcN26j0Z1POeN13n0[1nNfNO1R1\\1kNiNMMW1]1hNkNMIZ1_1eNmNLF^1`1cNnNJDb1a1`NoNKAd1f1ZNlN1^Od1e2[MoMQ1\\Oc1Q3PMeM\\1YOd1X3iLaMb1XOd1\\3dL^Mh1VOc1j7\\NVHc1k7]NUHb1m7]NSHb1n7]NSHb1o7]NQHb1P8^NPHa1Q8_NoG`1S8_NmG`1T8_NmG`1U8_NkGa1U8_NkG`1W8_NiG`1X8`NhG_1Y8`NhG_1Z8`NfG_1[8aNeG^1]8aNcG^1^8bNbG]1`8aNaG^1`8bN`G]1a8cN_G\\1c8cN]G]1c8cN]G]1d8aN]G_1c8aN]G_1d8`N\\G`1d8`N\\G_1e8aN[G_1f8`NZG`1f8_N[Ga1f8^NZGb1f8^NZGb1f8^NZGb1g8]NYGc1g8\\NZGd1g8[NYGd1h8\\NXGd1i8[NWGe1i8[NWGe1i8ZNXGf1i8YNWGg1i8YNWGg1j8XNVGh1j8XNVGg1l8WNUGi1k8WNUGi1k8WNUGi1l8VNTGj1l8VNTGj1m8TNTGl1l8TNTGl1m8SNSGm1m8SNSGl1n8TNRGl1o8SNQGm1o8RNRGn1o8QNQGo1o8QNQGo1P9PNPGP2n8RNRGn1l8SNUGm1i8UNWGj1h8XNXGh1f8ZNZGf1c8]N]Gc1a8^N_Gc1_8_NaGa1]8aNcG_1[8cNeG]1Y8eNgGZ1X8gNiGY1U8iNkGW1S8kNmGU1Q8mNoGS1o7oNQHQ1m7POTHP1j7ROVHn0h7TOXHl0f7VOZHi0e7YO[Hg0c7[O]He0a7\\O`Hd0^7^ObHb0\\7@dHa0Y7AgH`0V7BjH>T7CmH>P7DPI=m6ESIjJCS5?mJBP5?QKBk4a0UK_Oi4c0WK^Oe4e0[K\\Ob4f0^K[O^4g0cKYO[4i0eKXOW4i0kKXOR4h0PLXOn3h0TLYOh3h0ZLYOc3g0_LZO]3g0eLYOY3g0iLZOS3g0oLZOn2f0TM[Oh2f0ZMZOd2f0^M[O_2e0cM\\OY2e0iM\\OT2c0oM]On1d0TN]Oi1c0YN^Oc1c0_N^O^1b0dNCHfJhMh5b2MROkJSNX5m28[NQK_Ng4W3d0fMUKiNX4c3n0oL[1S3POYLW1i3^72N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3G9G9G9G9G9G9G9G9GC=lNT1lNT1oL\\H"}, "label": "a man is smiling with his couple and both are wearing white cloth"}]} {"id": 240521, "image": "COCO_train2014_000000240521.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man wearing the white vest\"."}], "task": "refering", "answer_template": "The \"the man wearing the white vest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 72, 83, 84, 85, 86, 87, 88, 98, 99, 100, 101, 102, 103, 113, 114, 115, 116, 117, 118, 128, 129, 130, 131, 132, 133, 143, 144, 145, 146, 147, 148, 158, 159, 160, 161, 162, 163, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 203, 204, 205, 206, 207, 209, 218, 219, 220, 221, 222, 224, 233, 234, 235, 236, 237, 239, 248, 249, 250, 251, 254, 265, 266, 269, 283, 284, 297, 298, 299, 308, 309, 310, 311, 312, 313, 314, 323, 324, 325, 326, 327, 328, 329, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.539316037735849, 0.2, 1.0, 0.9887656249999999], "iscrowd": 0, "area": 60005.99044999998, "rle": {"size": [640, 424], "counts": "bg_47h56d72[HZ1R6nNmIY1FUN>j0KT1CWN:j02R1AXN8h06S1^O[N7c0:U1[O]N6`0>V1XO_N5=b0W1VO`N3;f0X1SOcN26j0Z1POeN13n0[1nNfNO1R1\\1kNiNMMW1]1hNkNMIZ1_1eNmNLF^1`1cNnNJDb1a1`NoNKAd1f1ZNlN1^Od1e2[MoMQ1\\Oc1Q3PMeM\\1YOd1X3iLaMb1XOd1\\3dL^Mh1VOc1j7\\NVHc1k7]NUHb1m7]NSHb1n7]NSHb1o7]NQHb1P8^NPHa1Q8_NoG`1S8_NmG`1T8_NmG`1U8_NkGa1U8_NkG`1W8_NiG`1X8`NhG_1Y8`NhG_1Z8`NfG_1[8aNeG^1]8aNcG^1^8bNbG]1`8aNaG^1`8bN`G]1a8cN_G\\1c8cN]G]1c8cN]G]1d8aN]G_1c8aN]G_1d8`N\\G`1d8`N\\G_1e8aN[G_1f8`NZG`1f8_N[Ga1f8^NZGb1f8^NZGb1f8^NZGb1g8]NYGc1g8\\NZGd1g8[NYGd1h8\\NXGd1i8[NWGe1i8[NWGe1i8ZNXGf1i8YNWGg1i8YNWGg1j8XNVGh1j8XNVGg1l8WNUGi1k8WNUGi1k8WNUGi1l8VNTGj1l8VNTGj1m8TNTGl1l8TNTGl1m8SNSGm1m8SNSGl1n8TNRGl1o8SNQGm1o8RNRGn1o8QNQGo1o8QNQGo1P9PNPGP2n8RNRGn1l8SNUGm1i8UNWGj1h8XNXGh1f8ZNZGf1c8]N]Gc1a8^N_Gc1_8_NaGa1]8aNcG_1[8cNeG]1Y8eNgGZ1X8gNiGY1U8iNkGW1S8kNmGU1Q8mNoGS1o7oNQHQ1m7POTHP1j7ROVHn0h7TOXHl0f7VOZHi0e7YO[Hg0c7[O]He0a7\\O`Hd0^7^ObHb0\\7@dHa0Y7AgH`0V7BjH>T7CmH>P7DPI=m6ESIjJCS5?mJBP5?QKBk4a0UK_Oi4c0WK^Oe4e0[K\\Ob4f0^K[O^4g0cKYO[4i0eKXOW4i0kKXOR4h0PLXOn3h0TLYOh3h0ZLYOc3g0_LZO]3g0eLYOY3g0iLZOS3g0oLZOn2f0TM[Oh2f0ZMZOd2f0^M[O_2e0cM\\OY2e0iM\\OT2c0oM]On1d0TN]Oi1c0YN^Oc1c0_N^O^1b0dNCHfJhMh5b2MROkJSNX5m28[NQK_Ng4W3d0fMUKiNX4c3n0oL[1S3POYLW1i3^72N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3G9G9G9G9G9G9G9G9GC=lNT1lNT1oL\\H"}, "label": "the man wearing the white vest"}]} {"id": 445323, "image": "COCO_train2014_000000445323.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tie which weared by a person wearing head phone\"."}], "task": "refering", "answer_template": "The \"a tie which weared by a person wearing head phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 175, 177, 178, 198, 199, 200, 221, 222, 244, 245, 267, 268, 290, 291, 312, 313, 314], "bbox": [0.603890625, 0.45392307692307693, 0.78465625, 1.0], "iscrowd": 0, "area": 6878.6196, "rle": {"size": [390, 640], "counts": "`\\c41e;`0A?A?N2K5K5K6K4K5N2N2N2N2N2N2M3N2N00N10TNZMeHc0m0R2_6^MdHf0e0j1i6cMaHi0=c1S7gMaHl02]1^7jM_HX3i7lLVHc2\\8\\MdG]2d8cM\\GY2i8eMWGY2]1iM_5d2^JbMi5U2TJPNU6d1hIbN[6X1cImNc6f1^H`Nh7l26J8HO1O100O1O1O1N2N2O1N2N2N2OZMiH?U7@^I0`6NTJBj5 in this image.", "objects": [{"patches": [155, 175, 177, 178, 198, 199, 200, 221, 222, 244, 245, 267, 268, 290, 291, 312, 313, 314], "bbox": [0.603890625, 0.45392307692307693, 0.78465625, 1.0], "iscrowd": 0, "area": 6878.6196, "rle": {"size": [390, 640], "counts": "`\\c41e;`0A?A?N2K5K5K6K4K5N2N2N2N2N2N2M3N2N00N10TNZMeHc0m0R2_6^MdHf0e0j1i6cMaHi0=c1S7gMaHl02]1^7jM_HX3i7lLVHc2\\8\\MdG]2d8cM\\GY2i8eMWGY2]1iM_5d2^JbMi5U2TJPNU6d1hIbN[6X1cImNc6f1^H`Nh7l26J8HO1O100O1O1O1N2N2O1N2N2N2OZMiH?U7@^I0`6NTJBj5 in this image.", "objects": [{"patches": [52, 53, 56, 75, 76, 77, 78, 79, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 124, 144, 145, 146, 147, 166, 167, 169, 170, 188, 189, 190, 192, 193, 216], "bbox": [0.169015625, 0.17187793427230047, 0.46900000000000003, 0.6188028169014085], "iscrowd": 0, "area": 10983.884799999998, "rle": {"size": [426, 640], "counts": "iR]18R=1O1O1N100000000000O01O001O01000000O10000O1000VEFc8:]GI`88`GI^87bGJ^86aGK^86aGK^85cGK]85cGL[84eGMZ84fGLZ84eGMZ84fGLZ83fGNY83gGMY83fGMZ84fGLZ84eGM[83dGN[83cGO]80cG1]8NbG4^8KcG5\\8LcG5\\8KeG5Z8KgG5X8LgGASNb0U:MlG2S8OkG3S8NmG3S8MkG5T8KlG6T8JjG8U8HjG:V8FiG;V8EiG=W8CgG?X8AgGa0Y8_OeGc0T7lNXHf0c0`0S7lN\\G4d0a0k0a0S7lN]G4c0`0k0b0S7WOQH6k07cN^O_87RH5j06oNVOS8b0RH2k05VOTOj7h0UHOi06e7LaGNi07f7L`GNh07g7N_GLg08j7N^GGi0Fd0j6mNdI;Dg0h6mNgI9Dh0e6oNiI5Hh0^6SOmI2Ih0Z6VOoIOLg0U6YOULe0k3[OWLb0j3]OZL`0e3A]L=c3C_L;a3E`L:`3FaL9^3GcL9]3GdL7]3IcL7]3IdL6[3KfL4Z3LfL4Y3MhL2W3OjL0U30lL0T30lL0S31mLOR32nLNQ33PMLo25QMKm27SMHm28UMGj2:WM^OP3b0l4^O\\C>c in this image.", "objects": [{"patches": [52, 53, 56, 75, 76, 77, 78, 79, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 124, 144, 145, 146, 147, 166, 167, 169, 170, 188, 189, 190, 192, 193, 216], "bbox": [0.169015625, 0.17187793427230047, 0.46900000000000003, 0.6188028169014085], "iscrowd": 0, "area": 10983.884799999998, "rle": {"size": [426, 640], "counts": "iR]18R=1O1O1N100000000000O01O001O01000000O10000O1000VEFc8:]GI`88`GI^87bGJ^86aGK^86aGK^85cGK]85cGL[84eGMZ84fGLZ84eGMZ84fGLZ83fGNY83gGMY83fGMZ84fGLZ84eGM[83dGN[83cGO]80cG1]8NbG4^8KcG5\\8LcG5\\8KeG5Z8KgG5X8LgGASNb0U:MlG2S8OkG3S8NmG3S8MkG5T8KlG6T8JjG8U8HjG:V8FiG;V8EiG=W8CgG?X8AgGa0Y8_OeGc0T7lNXHf0c0`0S7lN\\G4d0a0k0a0S7lN]G4c0`0k0b0S7WOQH6k07cN^O_87RH5j06oNVOS8b0RH2k05VOTOj7h0UHOi06e7LaGNi07f7L`GNh07g7N_GLg08j7N^GGi0Fd0j6mNdI;Dg0h6mNgI9Dh0e6oNiI5Hh0^6SOmI2Ih0Z6VOoIOLg0U6YOULe0k3[OWLb0j3]OZL`0e3A]L=c3C_L;a3E`L:`3FaL9^3GcL9]3GdL7]3IcL7]3IdL6[3KfL4Z3LfL4Y3MhL2W3OjL0U30lL0T30lL0S31mLOR32nLNQ33PMLo25QMKm27SMHm28UMGj2:WM^OP3b0l4^O\\C>c in this image.", "objects": [{"patches": [4, 25, 26, 27, 28, 29, 48, 49, 50, 51, 52, 70, 71, 72, 73, 74, 75, 93, 94, 95, 96, 97, 98, 99, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 277, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.0, 0.050633802816901406, 0.45881249999999996, 1.0], "iscrowd": 0, "area": 73185.81519999998, "rle": {"size": [426, 640], "counts": "o85m<9F9H9F:J6I6K6J6J5K6J7I6M4M2M4L3M4L3M4L3M4L3M4M2M4L3M3M4J5L5K5J5L5J5L5K4K6K5J5L5K3L5L3M4L3M4L4L3M4M2M4L4L3M4L3M4RKVH[1GX1j9aM[F]2Z:N2M3N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2M3N2N2N2N2N2N2O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1000000000000000000000000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O00001O1O1O1O1O1O2N1O1O1O1O2N3M3M3M3M4L5K5K5K5K5K5K5K5K5K6J5K5K5kG]LY6h3hHTMV7c4M2N2N2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2`JnHd4S7[KQIa4o6`KSI]4n6bKVIZ4k6fKXIV4i6jKYIS4g6mK]Io3d6QL^Il3c6SLaIi3_6XLcIf0@U1m6UNgI4Nc1[6ZNjI@>R2h5_NgK]1Z4bNjKZ1V4gNnKT1R4lNRLP1n3QOULk0l3TOXLh0h3YO[Lc0e3^O^L>b3BcL9]3HfL4[3KiL1W30lLLT34PMHP39SMCm2>WM]Oj2b0ZMZOf2g0]MUOc2k0n4100O001O0010O01O1OO1O2N1O1O1N2O2N1O1O1O1O1O1O2N1N2O1O1O1O1O2N1O1O1O1OkS`4"}, "label": "a woman with sunglasses in her hair"}]} {"id": 543642, "image": "COCO_train2014_000000543642.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman\"."}], "task": "refering", "answer_template": "The \"a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 25, 26, 27, 28, 29, 48, 49, 50, 51, 52, 70, 71, 72, 73, 74, 75, 93, 94, 95, 96, 97, 98, 99, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 277, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306, 307, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.0, 0.050633802816901406, 0.45881249999999996, 1.0], "iscrowd": 0, "area": 73185.81519999998, "rle": {"size": [426, 640], "counts": "o85m<9F9H9F:J6I6K6J6J5K6J7I6M4M2M4L3M4L3M4L3M4L3M4M2M4L3M3M4J5L5K5J5L5J5L5K4K6K5J5L5K3L5L3M4L3M4L4L3M4M2M4L4L3M4L3M4RKVH[1GX1j9aM[F]2Z:N2M3N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2M3N2N2N2N2N2N2O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1000000000000000000000000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O00001O1O1O1O1O1O2N1O1O1O1O2N3M3M3M3M4L5K5K5K5K5K5K5K5K5K6J5K5K5kG]LY6h3hHTMV7c4M2N2N2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2`JnHd4S7[KQIa4o6`KSI]4n6bKVIZ4k6fKXIV4i6jKYIS4g6mK]Io3d6QL^Il3c6SLaIi3_6XLcIf0@U1m6UNgI4Nc1[6ZNjI@>R2h5_NgK]1Z4bNjKZ1V4gNnKT1R4lNRLP1n3QOULk0l3TOXLh0h3YO[Lc0e3^O^L>b3BcL9]3HfL4[3KiL1W30lLLT34PMHP39SMCm2>WM]Oj2b0ZMZOf2g0]MUOc2k0n4100O001O0010O01O1OO1O2N1O1O1N2O2N1O1O1O1O1O1O2N1N2O1O1O1O1O2N1O1O1O1OkS`4"}, "label": "a woman"}]} {"id": 543642, "image": "COCO_train2014_000000543642.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the purse strap on the woman\"."}], "task": "refering", "answer_template": "The \"the purse strap on the woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 166, 167, 190, 191, 213, 214, 237], "bbox": [0.17770312500000002, 0.3932629107981221, 0.340828125, 0.7359859154929577], "iscrowd": 0, "area": 2327.656099999999, "rle": {"size": [426, 640], "counts": "VP`11Y=1O1O100O1O1O1O1O100O1O1O1O10000O100000000O1000000O1000000000000O10000000000000000O100000000000000O10000000000000001N2O1O2N1O1O2N100O2O0O1O2O0O101N100O2N100O2O0O101N100100O2O0O1001O001ON3K4K5L4L4L4L5J5L4LbX_5"}, "label": "the purse strap on the woman"}]} {"id": 543642, "image": "COCO_train2014_000000543642.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown purse strap\"."}], "task": "refering", "answer_template": "The \"brown purse strap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 166, 167, 190, 191, 213, 214, 237], "bbox": [0.17770312500000002, 0.3932629107981221, 0.340828125, 0.7359859154929577], "iscrowd": 0, "area": 2327.656099999999, "rle": {"size": [426, 640], "counts": "VP`11Y=1O1O100O1O1O1O1O100O1O1O1O10000O100000000O1000000O1000000000000O10000000000000000O100000000000000O10000000000000001N2O1O2N1O1O2N100O2O0O1O2O0O101N100O2N100O2O0O101N100100O2O0O1001O001ON3K4K5L4L4L4L5J5L4LbX_5"}, "label": "brown purse strap"}]} {"id": 527267, "image": "COCO_train2014_000000527267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a solid green shirt\"."}], "task": "refering", "answer_template": "The \"a person in a solid green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 119, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 159, 160, 161, 162, 163, 164, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 205, 206, 207, 208, 209, 221, 222, 223, 224, 234, 235, 236, 237, 238, 239, 249, 250, 251, 252, 253, 254, 263, 264, 265, 266, 267, 268, 269, 277, 278, 279, 280, 281, 282, 283, 284, 291, 292, 293, 294, 295, 296, 297, 298, 299, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.3505399061032864, 0.32032812499999996, 1.0, 0.9987187500000001], "iscrowd": 0, "area": 71040.02485000002, "rle": {"size": [640, 426], "counts": "ffm2o0ob03N3L3M4M2M3M3M2O1N2N2O1N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O2M2O1O1O1N2O1O1O1N2O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2N2N2N2M3N2N2N2N2N2N2N2N2M3N2N2O100O10000O1O1O1O1O1O1O1O1O1eKlIiJU6V5oIgJQ6X5RJeJo5Z5UJcJk5\\5XJaJi5^5[J_Je5^5`J_Ja5Y5jJdJV5U5SKhJn4P5_KkJa4n4mKjJT4R5XLfJh3X5aL`J`3_5eL]J[3b5hL\\JX3c5lLYJU3f5oLWJQ3h5RMVJn2i5VMSJk2k5ZMRJf2m5]MQJc2n5aMnI`2Q6dMlI\\2S6gMkIY2T6kMhIV2W6kMhIV2V6mMhIT2W6nMhIR2W6oMhIR2W6PNgIQ2X6QNfIP2Y6RNeIo1Z6RNeIo1Y6TNeIm1[6TNdIl1[6VNcIk1\\6VNcIk1]6VNaIk1^6WN`Ij1_6WN`Ij1_6XN_Ii1a6XN^Ih1a6ZN]Ig1b6ZN^If1b6[N\\If1c6\\N\\Id1c6]N\\Id1c6^N[Ic1e6^NZIb1e6`NYIa1f6`NZI`1e6bNYI_1g6cNWI]1h6n5O100O1O1O100O1O100O1O100000000O100000000O100000000O100000000O100000000O100000000O100000000001O000O1N2N2N2N2N3M2N2M3N2N2N3M2N2N2N3M2N2N2N3M2M3N2N3L3K5L4K6K4K5L4K6L3L4L4M4K4L4M3L5L3J6cL`ETMf:l2[EmLl:R3VEfLP;Y3SE_LS;SNQHa1mL_OXb0oNmC"}, "label": "a person in a solid green shirt"}]} {"id": 527267, "image": "COCO_train2014_000000527267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a short haired girl in glasses standing in a green shirt with arms crossed\"."}], "task": "refering", "answer_template": "The \"a short haired girl in glasses standing in a green shirt with arms crossed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 119, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 159, 160, 161, 162, 163, 164, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 205, 206, 207, 208, 209, 221, 222, 223, 224, 234, 235, 236, 237, 238, 239, 249, 250, 251, 252, 253, 254, 263, 264, 265, 266, 267, 268, 269, 277, 278, 279, 280, 281, 282, 283, 284, 291, 292, 293, 294, 295, 296, 297, 298, 299, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.3505399061032864, 0.32032812499999996, 1.0, 0.9987187500000001], "iscrowd": 0, "area": 71040.02485000002, "rle": {"size": [640, 426], "counts": "ffm2o0ob03N3L3M4M2M3M3M2O1N2N2O1N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O2M2O1O1O1N2O1O1O1N2O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2N2N2N2M3N2N2N2N2N2N2N2N2M3N2N2O100O10000O1O1O1O1O1O1O1O1O1eKlIiJU6V5oIgJQ6X5RJeJo5Z5UJcJk5\\5XJaJi5^5[J_Je5^5`J_Ja5Y5jJdJV5U5SKhJn4P5_KkJa4n4mKjJT4R5XLfJh3X5aL`J`3_5eL]J[3b5hL\\JX3c5lLYJU3f5oLWJQ3h5RMVJn2i5VMSJk2k5ZMRJf2m5]MQJc2n5aMnI`2Q6dMlI\\2S6gMkIY2T6kMhIV2W6kMhIV2V6mMhIT2W6nMhIR2W6oMhIR2W6PNgIQ2X6QNfIP2Y6RNeIo1Z6RNeIo1Y6TNeIm1[6TNdIl1[6VNcIk1\\6VNcIk1]6VNaIk1^6WN`Ij1_6WN`Ij1_6XN_Ii1a6XN^Ih1a6ZN]Ig1b6ZN^If1b6[N\\If1c6\\N\\Id1c6]N\\Id1c6^N[Ic1e6^NZIb1e6`NYIa1f6`NZI`1e6bNYI_1g6cNWI]1h6n5O100O1O1O100O1O100O1O100000000O100000000O100000000O100000000O100000000O100000000O100000000001O000O1N2N2N2N2N3M2N2M3N2N2N3M2N2N2N3M2N2N2N3M2M3N2N3L3K5L4K6K4K5L4K6L3L4L4M4K4L4M3L5L3J6cL`ETMf:l2[EmLl:R3VEfLP;Y3SE_LS;SNQHa1mL_OXb0oNmC"}, "label": "a short haired girl in glasses standing in a green shirt with arms crossed"}]} {"id": 527267, "image": "COCO_train2014_000000527267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing green and holding bananas\"."}], "task": "refering", "answer_template": "The \"a person wearing green and holding bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 32, 33, 34, 35, 47, 48, 49, 50, 51, 62, 63, 64, 65, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 96, 106, 107, 108, 109, 110, 111, 112, 120, 121, 122, 123, 124, 125, 126, 127, 135, 136, 137, 138, 139, 140, 141, 142, 150, 151, 152, 153, 154, 155, 156, 165, 166, 167, 168, 169, 170, 171, 181, 182, 183, 184, 185, 186, 196, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 227, 228, 229, 230, 231, 242, 243, 244, 245, 257, 258, 259, 260, 271, 272, 273, 274, 275, 286, 287, 288, 289], "bbox": [0.03406103286384977, 0.06953125, 0.48715962441314553, 0.8450468749999999], "iscrowd": 0, "area": 60808.028900000005, "rle": {"size": [640, 426], "counts": "Zd9>_c0>B>B6J7H8I6J7I6I8I9G8H4L4K6L3L4L4L5K4L4L4L5L3L4eFPKa3T5\\LRK\\3T5aLnJZ3W5dLjJX3[5eLgJW3_5fLbJV3c5hL^JT3g5iL[JS3k5jLVJR3o5lLRJP3S6mLnIP3X6mLiIP3[6nLfIP3^6lLdIR3_6lLbIQ3b6mL_IQ3e6kL]IS3g6jLZIS3l6iLUIU3o6gLSIW3R7eLoHX3V7eLkHY3Z7bLhH\\3\\7_LgH^3^7^LdH`3a7ZLbHd3b7XL`He3f7UL]Hi3g7RL\\Hl3h7PLZHn3k7lKXHP4n7lKTHQ4R8iKQHT4T8hKnGT4Y8fKjGW4]8cKeGZ4a8`KbG\\4e8_K\\G`4i8ZKZGe4k8UKWGj4m8PKVGo4n8lJTGS5P9gJSGX5R9aJQG^5i;O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O100000000000O1O2O0O1O100O1O100O2N1O100O1O100O2N2O0O2N101N2N1O2O0O2N2O0O2N2O0O2N101cDXJ_8j5^GZJ_8g5^G^J`8c5\\GcJ`8^5^GfJ_8\\5]GjJW7^O^Gj5X1lJS7D`G`5Z1QKo6J`GW5^1RKn60]GP5EkIi1X1R76ZGg4KmIe1Y1S7:WGb41kIa1\\1T7c6[GRH^1^1T7c6]GPH[1`1U7d6`GmGW1a1V7e6cGkGS1d1W7d6fGhGo0g1X7d6iGgGk0h1Y7n6cHUIZ7o6cHSIZ7Q7bHSIZ7R7bHPI\\7S7`HQI\\7S7aHoH\\7U7`HnH]7m9L3N3^M`ChLZ=T3oB_LY=]3h1L3L5K4N2N3M2N3M2N3M2N2N3M2N4L4L5K5K5J5H9G9H8G9G8B?[OlQX4"}, "label": "a person wearing green and holding bananas"}]} {"id": 527267, "image": "COCO_train2014_000000527267.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady holding bananas\"."}], "task": "refering", "answer_template": "The \"a lady holding bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 32, 33, 34, 35, 47, 48, 49, 50, 51, 62, 63, 64, 65, 76, 77, 78, 79, 80, 91, 92, 93, 94, 95, 96, 106, 107, 108, 109, 110, 111, 112, 120, 121, 122, 123, 124, 125, 126, 127, 135, 136, 137, 138, 139, 140, 141, 142, 150, 151, 152, 153, 154, 155, 156, 165, 166, 167, 168, 169, 170, 171, 181, 182, 183, 184, 185, 186, 196, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 227, 228, 229, 230, 231, 242, 243, 244, 245, 257, 258, 259, 260, 271, 272, 273, 274, 275, 286, 287, 288, 289], "bbox": [0.03406103286384977, 0.06953125, 0.48715962441314553, 0.8450468749999999], "iscrowd": 0, "area": 60808.028900000005, "rle": {"size": [640, 426], "counts": "Zd9>_c0>B>B6J7H8I6J7I6I8I9G8H4L4K6L3L4L4L5K4L4L4L5L3L4eFPKa3T5\\LRK\\3T5aLnJZ3W5dLjJX3[5eLgJW3_5fLbJV3c5hL^JT3g5iL[JS3k5jLVJR3o5lLRJP3S6mLnIP3X6mLiIP3[6nLfIP3^6lLdIR3_6lLbIQ3b6mL_IQ3e6kL]IS3g6jLZIS3l6iLUIU3o6gLSIW3R7eLoHX3V7eLkHY3Z7bLhH\\3\\7_LgH^3^7^LdH`3a7ZLbHd3b7XL`He3f7UL]Hi3g7RL\\Hl3h7PLZHn3k7lKXHP4n7lKTHQ4R8iKQHT4T8hKnGT4Y8fKjGW4]8cKeGZ4a8`KbG\\4e8_K\\G`4i8ZKZGe4k8UKWGj4m8PKVGo4n8lJTGS5P9gJSGX5R9aJQG^5i;O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O100000000000O1O2O0O1O100O1O100O2N1O100O1O100O2N2O0O2N101N2N1O2O0O2N2O0O2N2O0O2N101cDXJ_8j5^GZJ_8g5^G^J`8c5\\GcJ`8^5^GfJ_8\\5]GjJW7^O^Gj5X1lJS7D`G`5Z1QKo6J`GW5^1RKn60]GP5EkIi1X1R76ZGg4KmIe1Y1S7:WGb41kIa1\\1T7c6[GRH^1^1T7c6]GPH[1`1U7d6`GmGW1a1V7e6cGkGS1d1W7d6fGhGo0g1X7d6iGgGk0h1Y7n6cHUIZ7o6cHSIZ7Q7bHSIZ7R7bHPI\\7S7`HQI\\7S7aHoH\\7U7`HnH]7m9L3N3^M`ChLZ=T3oB_LY=]3h1L3L5K4N2N3M2N3M2N3M2N2N3M2N4L4L5K5K5J5H9G9H8G9G8B?[OlQX4"}, "label": "a lady holding bananas"}]} {"id": 551869, "image": "COCO_train2014_000000551869.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small baby eating banana with her mom\"."}], "task": "refering", "answer_template": "The \"small baby eating banana with her mom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 38, 39, 40, 41, 47, 49, 50, 51, 52, 58, 59, 60, 61, 62, 67, 68, 69, 70, 71, 72, 73, 77, 78, 79, 81, 82, 83, 84], "bbox": [0.0035000000000000005, 0.2735416666666667, 0.79484375, 0.8754583333333333], "iscrowd": 0, "area": 14308.553349999993, "rle": {"size": [240, 320], "counts": "e==S700000O10000000000O100O1O100O1O100O1O100O1O100O1N2O1O1O1N2O1O1O1N2O1O1O1O1O1O1O100O1O1O1O100O1O1O100O10000000000000000000fNTJ<2Lj5A\\J?NOf5A\\Ja0OMo53RJGIJV6>RJGT67nIGR6:nIES6NgI47LS60hI2h6MZI2f6N[I0f60=1N1000000100O2NSm00iRO4O2O1N1N2N3M2O1O100O1M3O01O1K410L401O1N2M5Lk00UO3M2N2WIGY6;dIGV6?hICV6?hICV6?hIBX6`0eIBV6n0O2O0O101O01O0010LgNoI[1n55O1N3N1O2N100O2N101O001O01O01O05J6J2N1O2N2N2N1O1O001O00001O00POlNdKT1V4VOcKm0U4_OeKc0[4@`Kb0_4GWK;[4eNlKb2Q4cMlK^2P4c0K5N2M3L4N2O1O1O1O1O1O1O1O1O1O001O100O1O100O100O1O100O100O100O10000000000O10000000000000000000000001O000XNQMoNo2o0UMnNm2P1VMiNo2V1UMbNP3]1SMmMa3R2fLeM^3Y2m0O1N3N1O1O1N3N1N2O2N1O1O2N2N2M3N2N2N2M3N2M3M4K5L4L4L4K5L4LR\\?"}, "label": "small baby eating banana with her mom"}]} {"id": 551869, "image": "COCO_train2014_000000551869.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby eating a banana\"."}], "task": "refering", "answer_template": "The \"a baby eating a banana\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 38, 39, 40, 41, 47, 49, 50, 51, 52, 58, 59, 60, 61, 62, 67, 68, 69, 70, 71, 72, 73, 77, 78, 79, 81, 82, 83, 84], "bbox": [0.0035000000000000005, 0.2735416666666667, 0.79484375, 0.8754583333333333], "iscrowd": 0, "area": 14308.553349999993, "rle": {"size": [240, 320], "counts": "e==S700000O10000000000O100O1O100O1O100O1O100O1O100O1N2O1O1O1N2O1O1O1N2O1O1O1O1O1O1O100O1O1O1O100O1O1O100O10000000000000000000fNTJ<2Lj5A\\J?NOf5A\\Ja0OMo53RJGIJV6>RJGT67nIGR6:nIES6NgI47LS60hI2h6MZI2f6N[I0f60=1N1000000100O2NSm00iRO4O2O1N1N2N3M2O1O100O1M3O01O1K410L401O1N2M5Lk00UO3M2N2WIGY6;dIGV6?hICV6?hICV6?hIBX6`0eIBV6n0O2O0O101O01O0010LgNoI[1n55O1N3N1O2N100O2N101O001O01O01O05J6J2N1O2N2N2N1O1O001O00001O00POlNdKT1V4VOcKm0U4_OeKc0[4@`Kb0_4GWK;[4eNlKb2Q4cMlK^2P4c0K5N2M3L4N2O1O1O1O1O1O1O1O1O1O001O100O1O100O100O1O100O100O100O10000000000O10000000000000000000000001O000XNQMoNo2o0UMnNm2P1VMiNo2V1UMbNP3]1SMmMa3R2fLeM^3Y2m0O1N3N1O1O1N3N1N2O2N1O1O2N2N2M3N2N2N2M3N2M3M4K5L4L4L4K5L4LR\\?"}, "label": "a baby eating a banana"}]} {"id": 551869, "image": "COCO_train2014_000000551869.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a baby eating a banana\"."}], "task": "refering", "answer_template": "The \"a woman with a baby eating a banana\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 29, 30, 31, 32, 41, 42, 43, 52, 53, 54, 62, 63, 64, 65, 68, 69, 73, 74, 75, 76, 79, 80, 81, 84, 85, 86, 87], "bbox": [0.19421875, 0.1115, 1.0, 0.8793749999999999], "iscrowd": 0, "area": 18410.945450000003, "rle": {"size": [240, 320], "counts": "mf>7Y74L1O2N8H0O1C[OmIg0R6]OjId0V6^OgIc0Y6:0000000000KlNnIT1Q6nNnIR1Q6QOmIo0R68N2O1O1O1001O1O1O001O1O1O001O1O001O1O1O001O1O1O0O2O1O001O001O001O001O001O001O001O1O001O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O2N[P<2coC2M4M2N2N1O2N1O1O2N101N1O1O2N1O2N1O1O2N1O1]LQOd0P1YOUOc0l0[OXOdNFa0S1h0\\O]NHk0l0g0@UNJT1f0f0DjMO`1R1BnN in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 29, 30, 31, 32, 41, 42, 43, 52, 53, 54, 62, 63, 64, 65, 68, 69, 73, 74, 75, 76, 79, 80, 81, 84, 85, 86, 87], "bbox": [0.19421875, 0.1115, 1.0, 0.8793749999999999], "iscrowd": 0, "area": 18410.945450000003, "rle": {"size": [240, 320], "counts": "mf>7Y74L1O2N8H0O1C[OmIg0R6]OjId0V6^OgIc0Y6:0000000000KlNnIT1Q6nNnIR1Q6QOmIo0R68N2O1O1O1001O1O1O001O1O1O001O1O001O1O1O001O1O1O0O2O1O001O001O001O001O001O001O001O1O001O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O2N[P<2coC2M4M2N2N1O2N1O1O2N101N1O1O2N1O2N1O1O2N1O1]LQOd0P1YOUOc0l0[OXOdNFa0S1h0\\O]NHk0l0g0@UNJT1f0f0DjMO`1R1BnN in this image.", "objects": [{"patches": [56, 57, 58, 59, 78, 79, 80, 81, 82, 101, 102, 103, 104, 124, 125, 126, 127, 148, 149, 150, 172, 173], "bbox": [0.3910625, 0.14777083333333335, 0.59559375, 0.43975000000000003], "iscrowd": 0, "area": 9673.96, "rle": {"size": [480, 640], "counts": "TZe31m>2N2N2N2N3M2N3M2N3M3M3N3M2N2M3N2N2M3N2N2N1N2O2N1O1O1O1O101N1O100O100O1O100O1O101O0nCQNR;P2lDVNP;j1lD[NR;f1fDbNZ;^1eDdNZ;\\1dDfN[;[1dDgN[;Y1dDhN[;Y1dDiN[;W1dDjN\\;W1`DmN^;[200O2O0000001O000O101O001O1nDZLh:h3RE]Lm:l3O1O1O0O1000000O101O000O100000]MZEl0e:SO]Em0c:QO`En0`:nNdER1\\:jNiEU1V:kNkEU1U:jNlEV1T:iNnEW1R:hNnEa0ChN_:f0PF>DlN\\:e0QF=FmNZ:e0PF;IaNC^Od9k0oEC`0Ba9k0oEAb0D_9k0PF]Oe0F[9n0RFXOe0JY9n0UFROe00V9n0QHQOn7P1THmNm7S1P30O0100O010O2O1N2N2O0O2N2N2O1N2N2N2N2N6J6KSZi3"}, "label": "a young man is doing a trick in the air while skateboarding"}]} {"id": 543678, "image": "COCO_train2014_000000543678.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy jumping in mid air\"."}], "task": "refering", "answer_template": "The \"a boy jumping in mid air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 59, 78, 79, 80, 81, 82, 101, 102, 103, 104, 124, 125, 126, 127, 148, 149, 150, 172, 173], "bbox": [0.3910625, 0.14777083333333335, 0.59559375, 0.43975000000000003], "iscrowd": 0, "area": 9673.96, "rle": {"size": [480, 640], "counts": "TZe31m>2N2N2N2N3M2N3M2N3M3M3N3M2N2M3N2N2M3N2N2N1N2O2N1O1O1O1O101N1O100O100O1O100O1O101O0nCQNR;P2lDVNP;j1lD[NR;f1fDbNZ;^1eDdNZ;\\1dDfN[;[1dDgN[;Y1dDhN[;Y1dDiN[;W1dDjN\\;W1`DmN^;[200O2O0000001O000O101O001O1nDZLh:h3RE]Lm:l3O1O1O0O1000000O101O000O100000]MZEl0e:SO]Em0c:QO`En0`:nNdER1\\:jNiEU1V:kNkEU1U:jNlEV1T:iNnEW1R:hNnEa0ChN_:f0PF>DlN\\:e0QF=FmNZ:e0PF;IaNC^Od9k0oEC`0Ba9k0oEAb0D_9k0PF]Oe0F[9n0RFXOe0JY9n0UFROe00V9n0QHQOn7P1THmNm7S1P30O0100O010O2O1N2N2O0O2N2N2O1N2N2N2N2N6J6KSZi3"}, "label": "a boy jumping in mid air"}]} {"id": 183236, "image": "COCO_train2014_000000183236.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the largest monitor in between two smaller ones\"."}], "task": "refering", "answer_template": "The \"the largest monitor in between two smaller ones\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 111, 112, 113, 125, 126, 127, 128, 129, 130, 142, 143, 144, 145, 146, 147, 159, 160, 161, 162, 163, 164, 175, 176, 177, 178, 179, 180, 181, 192, 193, 194, 195, 196, 197, 198], "bbox": [0.3378125, 0.29809375, 0.6905833333333333, 0.49989062500000003], "iscrowd": 0, "area": 19935.03435, "rle": {"size": [640, 480], "counts": "maU33dc09H8G9G9G9H8G9G9G9G9H8G9G9G81000000000O1000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000O100000O10000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000O1000O1000000000000000000000000000000000000000000000000000O100000000000000G9SOm0SOm0ROonl2"}, "label": "the largest monitor in between two smaller ones"}]} {"id": 183236, "image": "COCO_train2014_000000183236.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the large mac computer on the hdesk\"."}], "task": "refering", "answer_template": "The \"the large mac computer on the hdesk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 111, 112, 113, 125, 126, 127, 128, 129, 130, 142, 143, 144, 145, 146, 147, 159, 160, 161, 162, 163, 164, 175, 176, 177, 178, 179, 180, 181, 192, 193, 194, 195, 196, 197, 198], "bbox": [0.3378125, 0.29809375, 0.6905833333333333, 0.49989062500000003], "iscrowd": 0, "area": 19935.03435, "rle": {"size": [640, 480], "counts": "maU33dc09H8G9G9G9H8G9G9G9G9H8G9G9G81000000000O1000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000O100000O10000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000O1000O1000000000000000000000000000000000000000000000000000O100000000000000G9SOm0SOm0ROonl2"}, "label": "the large mac computer on the hdesk"}]} {"id": 256965, "image": "COCO_train2014_000000256965.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a truck with the side doors open parked beside pop up tent\"."}], "task": "refering", "answer_template": "The \"a truck with the side doors open parked beside pop up tent\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 106, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 234, 235, 236, 240], "bbox": [0.12276562499999999, 0.2583098591549296, 0.685265625, 0.686056338028169], "iscrowd": 0, "area": 51920.96839999999, "rle": {"size": [426, 640], "counts": "STQ1b0T in this image.", "objects": [{"patches": [74, 75, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 106, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 234, 235, 236, 240], "bbox": [0.12276562499999999, 0.2583098591549296, 0.685265625, 0.686056338028169], "iscrowd": 0, "area": 51920.96839999999, "rle": {"size": [426, 640], "counts": "STQ1b0T in this image.", "objects": [{"patches": [123, 124, 145, 146, 147, 168, 169, 170, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 304, 305, 306, 307], "bbox": [0.22174999999999997, 0.3639344262295082, 0.48579687499999996, 0.9252459016393444], "iscrowd": 0, "area": 21086.26705, "rle": {"size": [427, 640], "counts": "_bk1;a in this image.", "objects": [{"patches": [123, 124, 145, 146, 147, 168, 169, 170, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 304, 305, 306, 307], "bbox": [0.22174999999999997, 0.3639344262295082, 0.48579687499999996, 0.9252459016393444], "iscrowd": 0, "area": 21086.26705, "rle": {"size": [427, 640], "counts": "_bk1;a in this image.", "objects": [{"patches": [172, 173, 174, 195, 196, 197, 218, 219, 220, 241, 242, 243, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 305, 306, 307, 308, 309, 310, 311, 312, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.29021875, 0.5114285714285715, 0.6073437500000001, 0.9912646370023419], "iscrowd": 0, "area": 21817.7879, "rle": {"size": [427, 640], "counts": "cm]2k0_R:CkE=V:[12N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1100O2N101N1O2N1O2O1N1O2N1O2N1O2O0O2N1O2N000001O01O00000000000F:F:F:G9F:F:G9J6I7I7I7M3O10000O100O100O10000O100O10000O100O10000O20O000000000000000000001O00000000000000001O001O001O001O0O2M2O2M2XKaIc2a6XMeId2]6YMhId2Z6WMmId2V6WMPJe2[6oLjIm2]8K5L5J5L4K5K5L5J5L4K5K5L4KoZX3"}, "label": "baby being worn in an orange printed sling"}]} {"id": 314319, "image": "COCO_train2014_000000314319.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baby in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the baby in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 195, 196, 197, 218, 219, 220, 241, 242, 243, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 305, 306, 307, 308, 309, 310, 311, 312, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.29021875, 0.5114285714285715, 0.6073437500000001, 0.9912646370023419], "iscrowd": 0, "area": 21817.7879, "rle": {"size": [427, 640], "counts": "cm]2k0_R:CkE=V:[12N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1100O2N101N1O2N1O2O1N1O2N1O2N1O2O0O2N1O2N000001O01O00000000000F:F:F:G9F:F:G9J6I7I7I7M3O10000O100O100O10000O100O10000O100O10000O20O000000000000000000001O00000000000000001O001O001O001O0O2M2O2M2XKaIc2a6XMeId2]6YMhId2Z6WMmId2V6WMPJe2[6oLjIm2]8K5L5J5L4K5K5L5J5L4K5K5L4KoZX3"}, "label": "the baby in the right hand picture"}]} {"id": 420823, "image": "COCO_train2014_000000420823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorized bicycle behind a man in blue tracksuit bottoms\"."}], "task": "refering", "answer_template": "The \"a motorized bicycle behind a man in blue tracksuit bottoms\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [179, 180, 202, 203, 204, 205, 221, 222, 224, 225, 226, 227, 228, 244, 245, 247, 248, 249, 250, 251, 272, 273, 274], "bbox": [0.60965625, 0.4473333333333333, 0.9425000000000001, 0.7074374999999999], "iscrowd": 0, "area": 10579.95175, "rle": {"size": [480, 640], "counts": "jRg54`>>I8I6L4M4K4M3M3L3N1N3N21O4K5L5K4L3L5L2M0O10O100O010O100O10O01N2L4M2M4M3M3L3O24L5I6K6I7I6J7J5Jcii00XVVO7I6I8I7J5K6I6K6JIPCjNj<\\1WCbNg in this image.", "objects": [{"patches": [179, 180, 202, 203, 204, 205, 221, 222, 224, 225, 226, 227, 228, 244, 245, 247, 248, 249, 250, 251, 272, 273, 274], "bbox": [0.60965625, 0.4473333333333333, 0.9425000000000001, 0.7074374999999999], "iscrowd": 0, "area": 10579.95175, "rle": {"size": [480, 640], "counts": "jRg54`>>I8I6L4M4K4M3M3L3N1N3N21O4K5L5K4L3L5L2M0O10O100O010O100O10O01N2L4M2M4M3M3L3O24L5I6K6I7I6J7J5Jcii00XVVO7I6I8I7J5K6I6K6JIPCjNj<\\1WCbNg in this image.", "objects": [{"patches": [3, 4, 23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 70, 71, 72, 73, 74, 93, 94, 95, 96, 97, 117, 118, 119, 120, 140, 141, 142, 163, 164, 186, 187, 209, 210], "bbox": [0.003, 0.06030444964871194, 0.298359375, 0.6310772833723653], "iscrowd": 0, "area": 21472.7422, "rle": {"size": [427, 640], "counts": "Pl06T=2O1O1N2O1O1N2O1O2M2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1O2M4M3M3M3M2N3M3M3M3M3M3M3M3M3M3M3M3M4L5WIbL[3b3_LaL`3c3ZLaLd3d3UL`Li3d3QL_Ln3f3kK^LR4g3gK]LX4g3bK]L\\4h3]K\\La4j3WKYLh4m3PKULP5R4fJQLZ5U4^JnKa5n5000001O0O1000001O0000001O0000001N1000001O0000001O00000O10O1[NeJYK\\5b4nJXKS5h4RKRKo4m4UKPKk4o4XKnJi4Q5YKkJj4U5XKfJk4Y5XKbJk4]5b1N2O1N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2O1N2N2O100O1O100O100O1O101N100O100O1O100O100O1O100O1H8G9G9G9H9F9G9G9G9I7O2N1O1N2O1O1O2M2O1O1O1O1N3N1O1O10000O2O0O100O10000O2O0O10000O[Vk5"}, "label": "a floral vase to the left of a larger vase"}]} {"id": 330716, "image": "COCO_train2014_000000330716.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the floral vase in the back left of the picture\"."}], "task": "refering", "answer_template": "The \"the floral vase in the back left of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 70, 71, 72, 73, 74, 93, 94, 95, 96, 97, 117, 118, 119, 120, 140, 141, 142, 163, 164, 186, 187, 209, 210], "bbox": [0.003, 0.06030444964871194, 0.298359375, 0.6310772833723653], "iscrowd": 0, "area": 21472.7422, "rle": {"size": [427, 640], "counts": "Pl06T=2O1O1N2O1O1N2O1O2M2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1O2M4M3M3M3M2N3M3M3M3M3M3M3M3M3M3M3M3M4L5WIbL[3b3_LaL`3c3ZLaLd3d3UL`Li3d3QL_Ln3f3kK^LR4g3gK]LX4g3bK]L\\4h3]K\\La4j3WKYLh4m3PKULP5R4fJQLZ5U4^JnKa5n5000001O0O1000001O0000001O0000001N1000001O0000001O00000O10O1[NeJYK\\5b4nJXKS5h4RKRKo4m4UKPKk4o4XKnJi4Q5YKkJj4U5XKfJk4Y5XKbJk4]5b1N2O1N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2O1N2N2O100O1O100O100O1O101N100O100O1O100O100O1O100O1H8G9G9G9H9F9G9G9G9I7O2N1O1N2O1O1O2M2O1O1O1O1N3N1O1O10000O2O0O100O10000O2O0O10000O[Vk5"}, "label": "the floral vase in the back left of the picture"}]} {"id": 330716, "image": "COCO_train2014_000000330716.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pitcher with gold handle\"."}], "task": "refering", "answer_template": "The \"a pitcher with gold handle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 32, 33, 34, 35, 36, 37, 38, 40, 41, 42, 43, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 266, 267, 268], "bbox": [0.3988125, 0.022482435597189696, 0.8890781249999999, 0.7460655737704919], "iscrowd": 0, "area": 66675.24979999999, "rle": {"size": [427, 640], "counts": "k[Z3`0d in this image.", "objects": [{"patches": [9, 10, 11, 12, 32, 33, 34, 35, 36, 37, 38, 40, 41, 42, 43, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 266, 267, 268], "bbox": [0.3988125, 0.022482435597189696, 0.8890781249999999, 0.7460655737704919], "iscrowd": 0, "area": 66675.24979999999, "rle": {"size": [427, 640], "counts": "k[Z3`0d in this image.", "objects": [{"patches": [75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 191, 192, 193], "bbox": [0.22200000000000003, 0.20941176470588235, 0.507, 0.5694117647058823], "iscrowd": 0, "area": 18058.752, "rle": {"size": [408, 640], "counts": "Yhh11e<3N1N2N3M2N3N1N3M2N3M3N1N3M3M2N3N1N3M3mNR1O2N2O0O2N101N2N1O2O1N1O2O0O2N2N101N2N101N1O2N2O0O2N2N101N1O2O1N1O2N2O0O2N101N2N1O101N1O11O10O01O0010O01O010O1O0010O01O0010O01O00100O0010O01O0010O01O1O010O0000000O100O10O0100O100O10O0100O100O10O0100O100O1O010O100O100O010O100O100O010O100O101N101N101N101N1O2N1O1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O3M6J5K5K5K5K6J5K5K5K5K6J5K5Kigm3"}, "label": "green color vegetable in between potato and carrot"}]} {"id": 158686, "image": "COCO_train2014_000000158686.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bundle of broccoli\"."}], "task": "refering", "answer_template": "The \"bundle of broccoli\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 191, 192, 193], "bbox": [0.22200000000000003, 0.20941176470588235, 0.507, 0.5694117647058823], "iscrowd": 0, "area": 18058.752, "rle": {"size": [408, 640], "counts": "Yhh11e<3N1N2N3M2N3N1N3M2N3M3N1N3M3M2N3N1N3M3mNR1O2N2O0O2N101N2N1O2O1N1O2O0O2N2N101N2N101N1O2N2O0O2N2N101N1O2O1N1O2N2O0O2N101N2N1O101N1O11O10O01O0010O01O010O1O0010O01O0010O01O00100O0010O01O0010O01O1O010O0000000O100O10O0100O100O10O0100O100O10O0100O100O1O010O100O100O010O100O100O010O100O101N101N101N101N1O2N1O1O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O3M6J5K5K5K5K6J5K5K5K5K6J5K5Kigm3"}, "label": "bundle of broccoli"}]} {"id": 158686, "image": "COCO_train2014_000000158686.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"fresh kale on wooden tray with other veggies\"."}], "task": "refering", "answer_template": "The \"fresh kale on wooden tray with other veggies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 185, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 277, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306], "bbox": [0.014859375, 0.5002696078431373, 0.399671875, 0.9147549019607844], "iscrowd": 0, "area": 29692.480349999994, "rle": {"size": [408, 640], "counts": "PY42d<3L4M3M2N3M3M3M2N3M3M3L4M2N3M3M3M2N3M3M3M3M2N3M3M3M2N3M3M3N2M2N3M2N2N2N1O2N200O10000O10O0100O100O101N3N3L4M3L4M3L4M3L3N2N0O2O0O101N101N101N100O2O0O2O001O00001O0O2O001O00001O0O2O001O00001O001N101O000000000O1000000000000O100000000000O10O100001O00001O00001O001O00001O00001O00001O00001N10001O00001O001O00001O00001O00001O00001O00001O001O0O101O00001O00001O00001O000O2O0O2N100O2N100O2O0O1O2O0O101N2N2O1N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2N2O1N2O3K4L4K6K4K6K4K5L5K4K5L5J5L4K6Kdfh4"}, "label": "fresh kale on wooden tray with other veggies"}]} {"id": 158686, "image": "COCO_train2014_000000158686.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"leafy green vegetable to the left of the yams and just below the broccolli\"."}], "task": "refering", "answer_template": "The \"leafy green vegetable to the left of the yams and just below the broccolli\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 185, 186, 187, 188, 189, 190, 191, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 277, 278, 279, 280, 281, 282, 283, 284, 301, 302, 303, 304, 305, 306], "bbox": [0.014859375, 0.5002696078431373, 0.399671875, 0.9147549019607844], "iscrowd": 0, "area": 29692.480349999994, "rle": {"size": [408, 640], "counts": "PY42d<3L4M3M2N3M3M3M2N3M3M3L4M2N3M3M3M2N3M3M3M3M2N3M3M3M2N3M3M3N2M2N3M2N2N2N1O2N200O10000O10O0100O100O101N3N3L4M3L4M3L4M3L3N2N0O2O0O101N101N101N100O2O0O2O001O00001O0O2O001O00001O0O2O001O00001O001N101O000000000O1000000000000O100000000000O10O100001O00001O00001O001O00001O00001O00001O00001N10001O00001O001O00001O00001O00001O00001O00001O001O0O101O00001O00001O00001O000O2O0O2N100O2N100O2O0O1O2O0O101N2N2O1N2O1N2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2N2O1N2N2O1N2O3K4L4K6K4K6K4K5L5K4K5L5J5L4K6Kdfh4"}, "label": "leafy green vegetable to the left of the yams and just below the broccolli"}]} {"id": 437224, "image": "COCO_train2014_000000437224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the screen on the back of the seat that says spoil yourself\"."}], "task": "refering", "answer_template": "The \"the screen on the back of the seat that says spoil yourself\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 170, 171, 190, 191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264], "bbox": [0.26834375, 0.5152693208430914, 0.525671875, 0.7752459016393444], "iscrowd": 0, "area": 16260.083799999997, "rle": {"size": [427, 640], "counts": "PnW26U=8H8H9G8H8H8H9G8I7H9G8H8H1O000000000000001O00000000000001O0000000000000000000000000000000000001O00000001O0000000000000000000000000001O00000000000000000000000000000001O0001O0000000000000000000000000000000000001O000000000000000001O000000000000000001O0000000000000000000000000000000000001O0001O0000000000000000000000000000I7WOi0XOi0VOTbn3"}, "label": "the screen on the back of the seat that says spoil yourself"}]} {"id": 437224, "image": "COCO_train2014_000000437224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the screen with text clearly on it\"."}], "task": "refering", "answer_template": "The \"the screen with text clearly on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 170, 171, 190, 191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264], "bbox": [0.26834375, 0.5152693208430914, 0.525671875, 0.7752459016393444], "iscrowd": 0, "area": 16260.083799999997, "rle": {"size": [427, 640], "counts": "PnW26U=8H8H9G8H8H8H9G8I7H9G8H8H1O000000000000001O00000000000001O0000000000000000000000000000000000001O00000001O0000000000000000000000000001O00000000000000000000000000000001O0001O0000000000000000000000000000000000001O000000000000000001O000000000000000001O0000000000000000000000000000000000001O0001O0000000000000000000000000000I7WOi0XOi0VOTbn3"}, "label": "the screen with text clearly on it"}]} {"id": 527345, "image": "COCO_train2014_000000527345.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sausage sitting on top of the beans and french fries\"."}], "task": "refering", "answer_template": "The \"the sausage sitting on top of the beans and french fries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 176, 177, 178, 179, 198, 199, 200, 201, 202, 221, 222, 223, 224, 243, 244, 245, 246, 247, 266, 267, 268, 269, 289, 290, 291, 292, 312, 313, 314, 315, 334, 335, 336, 337, 357, 358, 359, 360, 380, 381, 382, 383, 403, 404, 405, 406, 426, 427, 428, 449, 450, 451], "bbox": [0.5350625, 0.127265625, 0.9610312499999999, 0.8805156249999999], "iscrowd": 0, "area": 40222.80169999998, "rle": {"size": [640, 640], "counts": "WXf61ac0?A?A?A?B>J6J6I7J5K6J6J6J6J6J6I7J6J6J5K6J6J6J6J6I7J6J6J5K6J6M3M3M3M3M3M3M3M2N3M3M3M3M3M3M3M3M3M2N1N2O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001OO1K5K5K5K5K5K5K5J6N2N2N2N2M3N2N2N3M2N2N3M2M3N3M2N2N3M2N2N3L3N2N3M2N2O2O0O101O0O101N10001N100O2O0O101O0O101N100O101O0O101N2O1N3N1O1N2O1N3N1O1N2O1N3N1N2O1O1N3N1N2O1N2O2N1N2O1N2O2N1O1O1O1O2N1O1O1O1O2N1O100O1O2N1O1O1O1O2N1O1O2N2O0O2N2N1O2N2N1O2N2N1O2O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O001O1O001O1O001O100O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O101N2O2M2O1N2O2M2O1N2O2M2M3B\\a?"}, "label": "the sausage sitting on top of the beans and french fries"}]} {"id": 527345, "image": "COCO_train2014_000000527345.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"weiner on the right\"."}], "task": "refering", "answer_template": "The \"weiner on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 176, 177, 178, 179, 198, 199, 200, 201, 202, 221, 222, 223, 224, 243, 244, 245, 246, 247, 266, 267, 268, 269, 289, 290, 291, 292, 312, 313, 314, 315, 334, 335, 336, 337, 357, 358, 359, 360, 380, 381, 382, 383, 403, 404, 405, 406, 426, 427, 428, 449, 450, 451], "bbox": [0.5350625, 0.127265625, 0.9610312499999999, 0.8805156249999999], "iscrowd": 0, "area": 40222.80169999998, "rle": {"size": [640, 640], "counts": "WXf61ac0?A?A?A?B>J6J6I7J5K6J6J6J6J6J6I7J6J6J5K6J6J6J6J6I7J6J6J5K6J6M3M3M3M3M3M3M3M2N3M3M3M3M3M3M3M3M3M2N1N2O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O001OO1K5K5K5K5K5K5K5J6N2N2N2N2M3N2N2N3M2N2N3M2M3N3M2N2N3M2N2N3L3N2N3M2N2O2O0O101O0O101N10001N100O2O0O101O0O101N100O101O0O101N2O1N3N1O1N2O1N3N1O1N2O1N3N1N2O1O1N3N1N2O1N2O2N1N2O1N2O2N1O1O1O1O2N1O1O1O1O2N1O100O1O2N1O1O1O1O2N1O1O2N2O0O2N2N1O2N2N1O2N2N1O2O1O001O1O001O1O001O1O001O1O1O001O1O001O1O001O1O001O1O001O1O001O1O001O100O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O101N2O2M2O1N2O2M2O1N2O2M2M3B\\a?"}, "label": "weiner on the right"}]} {"id": 527345, "image": "COCO_train2014_000000527345.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hotdog on the left\"."}], "task": "refering", "answer_template": "The \"the hotdog on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 118, 119, 120, 121, 141, 142, 143, 144, 164, 165, 166, 187, 188, 189, 210, 211, 212, 213, 233, 234, 235, 236, 256, 257, 258, 259, 260, 280, 281, 282, 283, 303, 304, 305, 306, 307, 326, 327, 328, 329, 330, 350, 351, 352, 353, 354, 374, 375, 376, 377, 378, 397, 398, 399, 400, 401, 402, 421, 422, 423, 424, 425, 426, 445, 446, 447, 448, 449, 450], "bbox": [0.14640625000000002, 0.180515625, 0.590578125, 0.86290625], "iscrowd": 0, "area": 39617.25264999999, "rle": {"size": [640, 640], "counts": "Qoj1:bc0?@?B>A?B?@;F6J7H7J6I8I6J7H7J6I8I6I8L3M3N3L4L5L4K5L4K5K6K4K5L4K5K5L4K5L4K5K5L5J5L2M3M2O2M2N3N1N2O2M2N3N1N2O2M2N3N1N3N1N2N3N1N3N1N3N1O1O2N1O2N[IdFg2[9gLZGW3e8WLPHh3o7gKcHX4\\7gKhHX4V7fKnHY4Q7eKTIZ4j6eKZIY4e6eK_IZ4`6dKeI[4Y6dKkIZ4U6cKoI]4o5bKUJ\\4j5bK[J\\4d5bK`J^4^5aKfJ]4Y5aKkJ_4S5`KQK^4n4`KWK_4h4^K\\Ka4c4^KaK`4^4^KgKa4W4^KlKa4S4]KRLb4l3\\KYLb4f3]K]Lc4a3[KdLc4\\3[KgLd4X3ZKmLe4Q3YKTMe4k2ZKXMf4f2XK_Me4a2ZKcMd4\\2ZKhMe4X2YKlMe4U2WKoMh4Q2VKSNh4m1UKXNi4h1UK[Nj4f1SK^Nj4_8N1O2N2N1O2N1O2N1N1000001O0000000O2O00000000001O0O1000001O0000000O2O000000001O000O10001O000000001N100000001O00000O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O100O2N1O1O1O1O2N1O001O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O001O1O1O1O001O1OijS5"}, "label": "the hotdog on the left"}]} {"id": 527345, "image": "COCO_train2014_000000527345.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sausage that is on top of the beans\"."}], "task": "refering", "answer_template": "The \"sausage that is on top of the beans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 118, 119, 120, 121, 141, 142, 143, 144, 164, 165, 166, 187, 188, 189, 210, 211, 212, 213, 233, 234, 235, 236, 256, 257, 258, 259, 260, 280, 281, 282, 283, 303, 304, 305, 306, 307, 326, 327, 328, 329, 330, 350, 351, 352, 353, 354, 374, 375, 376, 377, 378, 397, 398, 399, 400, 401, 402, 421, 422, 423, 424, 425, 426, 445, 446, 447, 448, 449, 450], "bbox": [0.14640625000000002, 0.180515625, 0.590578125, 0.86290625], "iscrowd": 0, "area": 39617.25264999999, "rle": {"size": [640, 640], "counts": "Qoj1:bc0?@?B>A?B?@;F6J7H7J6I8I6J7H7J6I8I6I8L3M3N3L4L5L4K5L4K5K6K4K5L4K5K5L4K5L4K5K5L5J5L2M3M2O2M2N3N1N2O2M2N3N1N2O2M2N3N1N3N1N2N3N1N3N1N3N1O1O2N1O2N[IdFg2[9gLZGW3e8WLPHh3o7gKcHX4\\7gKhHX4V7fKnHY4Q7eKTIZ4j6eKZIY4e6eK_IZ4`6dKeI[4Y6dKkIZ4U6cKoI]4o5bKUJ\\4j5bK[J\\4d5bK`J^4^5aKfJ]4Y5aKkJ_4S5`KQK^4n4`KWK_4h4^K\\Ka4c4^KaK`4^4^KgKa4W4^KlKa4S4]KRLb4l3\\KYLb4f3]K]Lc4a3[KdLc4\\3[KgLd4X3ZKmLe4Q3YKTMe4k2ZKXMf4f2XK_Me4a2ZKcMd4\\2ZKhMe4X2YKlMe4U2WKoMh4Q2VKSNh4m1UKXNi4h1UK[Nj4f1SK^Nj4_8N1O2N2N1O2N1O2N1N1000001O0000000O2O00000000001O0O1000001O0000000O2O000000001O000O10001O000000001N100000001O00000O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O100O2N1O1O1O1O2N1O001O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O001O1O1O1O001O1OijS5"}, "label": "sausage that is on top of the beans"}]} {"id": 510976, "image": "COCO_train2014_000000510976.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the mobile phone with a number 2125 towards the top right side\"."}], "task": "refering", "answer_template": "The \"the mobile phone with a number 2125 towards the top right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 25, 26, 27, 28, 29, 30, 48, 49, 50, 51, 52, 53, 71, 72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 303, 304], "bbox": [0.042453125, 0.06292307692307691, 0.32181249999999995, 0.9460769230769231], "iscrowd": 0, "area": 45854.187, "rle": {"size": [390, 640], "counts": "cb:b0d:Q1H9G9F9H9F9H9G8E in this image.", "objects": [{"patches": [4, 5, 25, 26, 27, 28, 29, 30, 48, 49, 50, 51, 52, 53, 71, 72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 303, 304], "bbox": [0.042453125, 0.06292307692307691, 0.32181249999999995, 0.9460769230769231], "iscrowd": 0, "area": 45854.187, "rle": {"size": [390, 640], "counts": "cb:b0d:Q1H9G9F9H9F9H9G8E in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297], "bbox": [0.313, 0.14838461538461536, 0.9760937500000001, 0.9311025641025641], "iscrowd": 0, "area": 114580.69880000001, "rle": {"size": [390, 640], "counts": "Y_\\22d;d0[Of0\\Oc0E;E;D9H9G8H8H9J5L4POPLdIV4P6U1F;E;E:G:E;G8L5L4L3M4K5L3N3M2N2O2M2N200O2N100O2N1O100O101N10000O10000O10000O10000O101O000000000O10000000000000001N1000000000000000000O1000000000000000000O1000000000000000000O10000000001O00000O1000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000O100000000000O100000000000000000000000000000000000000000000O10000000000000000000O100000000000000000000000O100000000000000000000000000000000000000000000O10000000O100000000000000000000000000000000000O100000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000O10000000000000000000000000000O2O0000000000000O1000001O0000000O100000001O00001N1000001O00001O0O101O001O1O001O1N2O1O1O1O1O1O0O2O2N1O1O1O1O1O2M2O1O2N2N2N2N2M3N4K4M4L3L4M4K4M4L3L4M4K4M5J?Bk0TO3N2M2O2M3N2M3N2M4L5J5L5K4L5K4L5K4L5K5K6J6I6J7I7J5J7I7H8G:F9Gh]5"}, "label": "phone that slides up"}]} {"id": 510976, "image": "COCO_train2014_000000510976.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"phone on right\"."}], "task": "refering", "answer_template": "The \"phone on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297], "bbox": [0.313, 0.14838461538461536, 0.9760937500000001, 0.9311025641025641], "iscrowd": 0, "area": 114580.69880000001, "rle": {"size": [390, 640], "counts": "Y_\\22d;d0[Of0\\Oc0E;E;D9H9G8H8H9J5L4POPLdIV4P6U1F;E;E:G:E;G8L5L4L3M4K5L3N3M2N2O2M2N200O2N100O2N1O100O101N10000O10000O10000O10000O101O000000000O10000000000000001N1000000000000000000O1000000000000000000O1000000000000000000O10000000001O00000O1000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000O100000000000O100000000000000000000000000000000000000000000O10000000000000000000O100000000000000000000000O100000000000000000000000000000000000000000000O10000000O100000000000000000000000000000000000O100000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000O10000000000000000000000000000O2O0000000000000O1000001O0000000O100000001O00001N1000001O00001O0O101O001O1O001O1N2O1O1O1O1O1O0O2O2N1O1O1O1O1O2M2O1O2N2N2N2N2M3N4K4M4L3L4M4K4M4L3L4M4K4M5J?Bk0TO3N2M2O2M3N2M3N2M4L5J5L5K4L5K4L5K4L5K5K6J6I6J7I7J5J7I7H8G:F9Gh]5"}, "label": "phone on right"}]} {"id": 76802, "image": "COCO_train2014_000000076802.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white couch with the various colored pillows on it\"."}], "task": "refering", "answer_template": "The \"the white couch with the various colored pillows on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 239, 244, 245], "bbox": [0.412265625, 0.522927400468384, 0.6895625000000001, 0.7139812646370024], "iscrowd": 0, "area": 9123.471699999996, "rle": {"size": [427, 640], "counts": "`Z^39Q=>B:F2N2N1SDnN`;T1YDROg;Y1O1O1O00MKeNaDX1c;621N100000000O100000000O101OO1O010O1O1O010O1O100O1O100O1O1000000000000000000001O000000000000000001O0FWDROi;n0YDlNk;T160001N10000000000O20O000000001O01O00kCTOk;W1O0000000O10000O100O100O10000O20O1O2N00001OO1000000000000O2O000000000000000000000000000dDcNj:^1oDhNQ;i100O1000000000000000000000000000000001O0000000000001O00000001O0001O0iNjD`0P;^OQEa0P;^OREa0o:]OREb0o:^ORE`0P;^OQEe0l:[OUEMc;2k001O1O1JXCGi<77OPhb2"}, "label": "the white couch with the various colored pillows on it"}]} {"id": 76802, "image": "COCO_train2014_000000076802.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the large couch between the lights\"."}], "task": "refering", "answer_template": "The \"the large couch between the lights\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 239, 244, 245], "bbox": [0.412265625, 0.522927400468384, 0.6895625000000001, 0.7139812646370024], "iscrowd": 0, "area": 9123.471699999996, "rle": {"size": [427, 640], "counts": "`Z^39Q=>B:F2N2N1SDnN`;T1YDROg;Y1O1O1O00MKeNaDX1c;621N100000000O100000000O101OO1O010O1O1O010O1O100O1O100O1O1000000000000000000001O000000000000000001O0FWDROi;n0YDlNk;T160001N10000000000O20O000000001O01O00kCTOk;W1O0000000O10000O100O100O10000O20O1O2N00001OO1000000000000O2O000000000000000000000000000dDcNj:^1oDhNQ;i100O1000000000000000000000000000000001O0000000000001O00000001O0001O0iNjD`0P;^OQEa0P;^OREa0o:]OREb0o:^ORE`0P;^OQEe0l:[OUEMc;2k001O1O1JXCGi<77OPhb2"}, "label": "the large couch between the lights"}]} {"id": 429059, "image": "COCO_train2014_000000429059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow fabric on a shelf behind a man\"."}], "task": "refering", "answer_template": "The \"yellow fabric on a shelf behind a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 143, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 212, 213, 214, 215, 216, 236, 237, 238, 239], "bbox": [0.22528125000000002, 0.39220140515222485, 0.421109375, 0.7324590163934426], "iscrowd": 0, "area": 12179.516099999997, "rle": {"size": [427, 640], "counts": "mVl11X=8I6J6I7K5K5K6J5K5K5K5K6J5K5K5K5K5K6J5K5K5K5K5K6J5K5K2N000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000001O;E:F;iLZFS2P:`MVF]2_:N2M2N3M3M3M3M2N3M3M3M2M4M2M3M2O1N3N1N2N2O2M2N2O2M2N2O1N3M2O1N3M2O1N2N3NZ]j4"}, "label": "yellow fabric on a shelf behind a man"}]} {"id": 429059, "image": "COCO_train2014_000000429059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow bag with part of mans eyes and ear in the foreground\"."}], "task": "refering", "answer_template": "The \"yellow bag with part of mans eyes and ear in the foreground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 143, 144, 145, 146, 166, 167, 168, 169, 189, 190, 191, 192, 212, 213, 214, 215, 216, 236, 237, 238, 239], "bbox": [0.22528125000000002, 0.39220140515222485, 0.421109375, 0.7324590163934426], "iscrowd": 0, "area": 12179.516099999997, "rle": {"size": [427, 640], "counts": "mVl11X=8I6J6I7K5K5K6J5K5K5K5K6J5K5K5K5K5K6J5K5K5K5K5K6J5K5K2N000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000001O;E:F;iLZFS2P:`MVF]2_:N2M2N3M3M3M3M2N3M3M3M2M4M2M3M2O1N3N1N2N2O2M2N2O2M2N2O1N3M2O1N3M2O1N2N3NZ]j4"}, "label": "yellow bag with part of mans eyes and ear in the foreground"}]} {"id": 429059, "image": "COCO_train2014_000000429059.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a purple folder on the bottom row\"."}], "task": "refering", "answer_template": "The \"a purple folder on the bottom row\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 221, 222, 223, 224, 225, 246, 247, 248], "bbox": [0.59834375, 0.3784543325526932, 0.828375, 0.7214519906323185], "iscrowd": 0, "area": 15042.6999, "rle": {"size": [427, 640], "counts": "Zno41l:b0mFAQ9`0mFCR9>lFDR9=lFET9 in this image.", "objects": [{"patches": [128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 221, 222, 223, 224, 225, 246, 247, 248], "bbox": [0.59834375, 0.3784543325526932, 0.828375, 0.7214519906323185], "iscrowd": 0, "area": 15042.6999, "rle": {"size": [427, 640], "counts": "Zno41l:b0mFAQ9`0mFCR9>lFDR9=lFET9 in this image.", "objects": [{"patches": [227, 228, 229, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 273, 274, 275, 276, 277, 278, 289, 290, 291, 292, 293, 294, 295, 306, 307, 308, 309, 310, 311, 323, 324, 325, 326, 327, 339, 340, 341, 342], "bbox": [0.13049180327868853, 0.668234375, 0.8640281030444966, 0.9858750000000001], "iscrowd": 0, "area": 20678.003600000007, "rle": {"size": [640, 427], "counts": "Y^S17cc0:G:I4L3M3M3M3M3M3N2N2O1O1O1O1N101O1O1O1N2O1O1O1O1O1O1O1O100O1O1O1O1O100O1O1O00100O001O1O010O1O001O10O01O1O001O10O10000O10O10O10000O10000O10O10O10000O10000O1000O010000O100O10000O10O10O10000O10000O10O10O0100O010O10O0100O010O10O0100O01000O010O10O0100O010O10O0100O010O01O001O1O001O0010O01O001O001O001O001O001O010O1O001O001O001O001O0010O01O001O010000O100O10O0100O100O100O10O0100O100O10000O010O100O100O100O10O0100O100O100O10O10O100O100O100O010O100O100O100O010O10000O100O010O0102M3M4L3M3M3M3M3M3M3M00000O1000000000000000000000001O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O]dS1"}, "label": "a long carrot that is slightly curved"}]} {"id": 175116, "image": "COCO_train2014_000000175116.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"2 carrots\"."}], "task": "refering", "answer_template": "The \"2 carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [229, 230, 231, 232, 244, 245, 246, 247, 248, 249, 261, 262, 263, 264, 265, 266, 277, 278, 279, 280, 281, 282, 283, 294, 295, 296, 297, 298, 310, 311, 312, 313, 327, 328], "bbox": [0.31665105386416864, 0.65634375, 0.9147775175644028, 0.9286406250000001], "iscrowd": 0, "area": 16558.34645000001, "rle": {"size": [640, 427], "counts": "^id2g0Yc01O1N101O0O2O1N101O0O2O1O0O2O0O2O1O0O2O001N2O001O10O010O10O10OO2M2N3N20O10O010O10O10O10O010O10O10O010O10O10O01000O0100O01000O010O10O10O10O01000O0100O01000O01000O010O10O10O10O01000O0100O010O10O0100O00100O010O1O010O1O010O010O0010O010O01O010O01O01O010O010O0010O01O010O010O0010O010O01O010O010O0010O01O010O01O01O010O010O1O3N2M2O2M3M2O2M3N0O1O010O010O0010O010O01O010O010O00010O010O0010O010O01O010O010O0010O010O00010O010O01O010O0010O010O01O01O001O000O2O00001N101O00001N10001O0O2O00001N101O00001N10001O0O2O000M4Ha0^Od0]Oc0\\OPoe0"}, "label": "2 carrots"}]} {"id": 175116, "image": "COCO_train2014_000000175116.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the two carrots in front\"."}], "task": "refering", "answer_template": "The \"the two carrots in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [229, 230, 231, 232, 244, 245, 246, 247, 248, 249, 261, 262, 263, 264, 265, 266, 277, 278, 279, 280, 281, 282, 283, 294, 295, 296, 297, 298, 310, 311, 312, 313, 327, 328], "bbox": [0.31665105386416864, 0.65634375, 0.9147775175644028, 0.9286406250000001], "iscrowd": 0, "area": 16558.34645000001, "rle": {"size": [640, 427], "counts": "^id2g0Yc01O1N101O0O2O1N101O0O2O1O0O2O0O2O1O0O2O001N2O001O10O010O10O10OO2M2N3N20O10O010O10O10O10O010O10O10O010O10O10O01000O0100O01000O010O10O10O10O01000O0100O01000O01000O010O10O10O10O01000O0100O010O10O0100O00100O010O1O010O1O010O010O0010O010O01O010O01O01O010O010O0010O01O010O010O0010O010O01O010O010O0010O01O010O01O01O010O010O1O3N2M2O2M3M2O2M3N0O1O010O010O0010O010O01O010O010O00010O010O0010O010O01O010O010O0010O010O00010O010O01O010O0010O010O01O01O001O000O2O00001N101O00001N10001O0O2O00001N101O00001N10001O0O2O000M4Ha0^Od0]Oc0\\OPoe0"}, "label": "the two carrots in front"}]} {"id": 355345, "image": "COCO_train2014_000000355345.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black strap on a white shirt\"."}], "task": "refering", "answer_template": "The \"a black strap on a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [259, 260, 275, 276, 291, 292, 308, 324, 325, 341], "bbox": [0.10165271966527198, 0.6680937499999999, 0.33705020920502093, 0.8956875], "iscrowd": 0, "area": 1862.2617499999994, "rle": {"size": [640, 478], "counts": "[Uo0a0^c02M2O0O010O01O0000O10O01O010O1D;O2O03M3M2N3M3M2N3M3M2N1O1O02O100O001O1O1O00100O1O1O01000000O1000000000000000000000001O00000000000000000000000000000000000000000000001O010O10O0100O01000O010O10O0100O10000O100O100O101O0O100O1OdfU6"}, "label": "a black strap on a white shirt"}]} {"id": 355345, "image": "COCO_train2014_000000355345.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umbrella the person in the pink hat is holding\"."}], "task": "refering", "answer_template": "The \"the umbrella the person in the pink hat is holding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 141, 142, 143, 153, 154, 155, 156, 157, 158, 159, 160, 174, 191], "bbox": [0.0, 0.31721875, 0.4765481171548117, 0.5151875], "iscrowd": 0, "area": 10202.109550000001, "rle": {"size": [640, 478], "counts": "b84gc0401O1N2O001N2O1O001N2O1O001N2O1O001N2O1O0O2O1O1O0O2O1O1O0O2O1O10O01O100O00100O1O010O1O100O1O100O100O1O100O1O100O1O100O1O100O1000000O100000000000000O1000000000000000O01000000000000O100000000O100000000O0100000000000000000000005K9G8HZOh^OlNPa0T1X_OlN``0T1i_OkNo?U1Y@kN_?U1j@jNn>V1ZAdNd>\\1eA[N[>e1P200000000O1000000000000000000000O10O1000000000000000001O001O0O2O001O001O001O001O001O001N101O001O001O001O001O001O0O2O001O001O001O001O1O1O001N2O1O1O1O1O1O1O001O1O1O1O1N2O1O1O001O1O1O1O1O1O1O1O0O2O1O1O1O1O1O1O00PPl4"}, "label": "the umbrella the person in the pink hat is holding"}]} {"id": 355345, "image": "COCO_train2014_000000355345.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"light blue umbrella bottom left\"."}], "task": "refering", "answer_template": "The \"light blue umbrella bottom left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 141, 142, 143, 153, 154, 155, 156, 157, 158, 159, 160, 174, 191], "bbox": [0.0, 0.31721875, 0.4765481171548117, 0.5151875], "iscrowd": 0, "area": 10202.109550000001, "rle": {"size": [640, 478], "counts": "b84gc0401O1N2O001N2O1O001N2O1O001N2O1O001N2O1O0O2O1O1O0O2O1O1O0O2O1O10O01O100O00100O1O010O1O100O1O100O100O1O100O1O100O1O100O1O100O1000000O100000000000000O1000000000000000O01000000000000O100000000O100000000O0100000000000000000000005K9G8HZOh^OlNPa0T1X_OlN``0T1i_OkNo?U1Y@kN_?U1j@jNn>V1ZAdNd>\\1eA[N[>e1P200000000O1000000000000000000000O10O1000000000000000001O001O0O2O001O001O001O001O001O001N101O001O001O001O001O001O0O2O001O001O001O001O1O1O001N2O1O1O1O1O1O1O001O1O1O1O1N2O1O1O001O1O1O1O1O1O1O1O0O2O1O1O1O1O1O1O00PPl4"}, "label": "light blue umbrella bottom left"}]} {"id": 11282, "image": "COCO_train2014_000000011282.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person behind the brown donut\"."}], "task": "refering", "answer_template": "The \"person behind the brown donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 48, 49, 50, 51, 53, 54, 55, 56, 59, 60, 61, 62, 63, 64, 65, 66, 67, 71, 72, 73, 77, 83, 84, 85, 86, 87, 88, 89, 90, 91, 94, 95, 107, 108, 109, 110, 111, 112, 113, 114, 117, 130, 131, 132, 133, 134, 135, 136, 137, 140, 153, 154, 155, 156, 157, 158, 159, 160, 163, 176, 177, 178, 179, 181, 182, 183, 186, 199, 200, 201, 202, 203, 205, 209, 210, 222, 223, 224, 232], "bbox": [0.09775, 0.0, 1.0, 0.6224791666666667], "iscrowd": 0, "area": 70622.28854999998, "rle": {"size": [480, 640], "counts": "hgm0b2a9l2UMk2UOk01O1O1N2O1O1O1jLnI]MS6`2VJYMk5c2^JWMc5e2gJTMZ5i2oJPMR5l2WKnLj4o2_KjLb4R3hKgLY4U3QLdLP4Y3XLaLi3[3aL^L`3_3iLZLX3b3QMXLP3d3ZMULg2h3bMQL_2k3]4M3L4L4M3L4M3M3O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100001O001O0O101O00001O00001O00001O00001O00001O00001O00001O00001O00001O001O001O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O00000000000000000000O100000000000000000000O100000000000000000000O100000000O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O010O1O1O1O1O1001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O3M3M3M3M3M4L3M3M3M3M3M4L3M3M3UIUL[2n3^MVLb2m3VMXLj2k3nLZLR3j3fLZLZ3i3^L\\Lb3g3VL]Lk3f3mK_LS4d3fK`LZ4c3^KbLb4b3UKcLk4`3nJdLR5_3fJfLZ5]3^JgLc5\\3VJhLj5[3nIjLR6Z3eIkL[6X3^IlLb6i50O10000000000O1000000000000O10000000000O10000000000O10000000O10O1000000000000O10000000000O10000000000O10000000000O1000000VMUI_Mk6_2YI_Mg6_2]I_Mc6_2`I_Ma6_2cI_M]6_2gI_MY6_2jI`MV6^2nI`MR6^2RJ`Mn5_2TJ_Mm5b2TJ\\Ml5e2UJYMk5h2UJWMk5j2VJTMj5m2WJQMi5P3WJnLj5S3WJkLi5V3XJhLh5Y3XJfLh5[3YJcLg5^3ZJ`Lf5b3YJ]Lg5d3ZJYLg5h3ZJVLf5k3[JSLe5n3[JQLe5P4\\JnKd5S4]JkKc5V4]JhKd5Y4]JcKe5^4\\J^Kf5c4ZJZKh5g4YJUKi5l4XJoJk5R5UJkJm5V5TJfJn5[5SJ`JP6a5g11O1O1O1O1O1O100000000001O000000001O00000000010O00000000001O000000001O00000000001O000000001O00000000001O00000000001O0O1N2N2N3N1N2N2O1N2N3M2O1N2N2O2M2N2N2O1N2N3N1N2N2N2O1N3M2O1N2N2N3N1N2VMiDh1Y;VNUE]1l:aNbEQ1`:lNnEg0U:VOYF in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 48, 49, 50, 51, 53, 54, 55, 56, 59, 60, 61, 62, 63, 64, 65, 66, 67, 71, 72, 73, 77, 83, 84, 85, 86, 87, 88, 89, 90, 91, 94, 95, 107, 108, 109, 110, 111, 112, 113, 114, 117, 130, 131, 132, 133, 134, 135, 136, 137, 140, 153, 154, 155, 156, 157, 158, 159, 160, 163, 176, 177, 178, 179, 181, 182, 183, 186, 199, 200, 201, 202, 203, 205, 209, 210, 222, 223, 224, 232], "bbox": [0.09775, 0.0, 1.0, 0.6224791666666667], "iscrowd": 0, "area": 70622.28854999998, "rle": {"size": [480, 640], "counts": "hgm0b2a9l2UMk2UOk01O1O1N2O1O1O1jLnI]MS6`2VJYMk5c2^JWMc5e2gJTMZ5i2oJPMR5l2WKnLj4o2_KjLb4R3hKgLY4U3QLdLP4Y3XLaLi3[3aL^L`3_3iLZLX3b3QMXLP3d3ZMULg2h3bMQL_2k3]4M3L4L4M3L4M3M3O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100001O001O0O101O00001O00001O00001O00001O00001O00001O00001O00001O00001O001O001O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O00000000000000000000O100000000000000000000O100000000000000000000O100000000O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O010O1O1O1O1O1001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O3M3M3M3M3M4L3M3M3M3M3M4L3M3M3UIUL[2n3^MVLb2m3VMXLj2k3nLZLR3j3fLZLZ3i3^L\\Lb3g3VL]Lk3f3mK_LS4d3fK`LZ4c3^KbLb4b3UKcLk4`3nJdLR5_3fJfLZ5]3^JgLc5\\3VJhLj5[3nIjLR6Z3eIkL[6X3^IlLb6i50O10000000000O1000000000000O10000000000O10000000000O10000000O10O1000000000000O10000000000O10000000000O10000000000O1000000VMUI_Mk6_2YI_Mg6_2]I_Mc6_2`I_Ma6_2cI_M]6_2gI_MY6_2jI`MV6^2nI`MR6^2RJ`Mn5_2TJ_Mm5b2TJ\\Ml5e2UJYMk5h2UJWMk5j2VJTMj5m2WJQMi5P3WJnLj5S3WJkLi5V3XJhLh5Y3XJfLh5[3YJcLg5^3ZJ`Lf5b3YJ]Lg5d3ZJYLg5h3ZJVLf5k3[JSLe5n3[JQLe5P4\\JnKd5S4]JkKc5V4]JhKd5Y4]JcKe5^4\\J^Kf5c4ZJZKh5g4YJUKi5l4XJoJk5R5UJkJm5V5TJfJn5[5SJ`JP6a5g11O1O1O1O1O1O100000000001O000000001O00000000010O00000000001O000000001O00000000001O000000001O00000000001O00000000001O0O1N2N2N3N1N2N2O1N2N3M2O1N2N2O2M2N2N2O1N2N3N1N2N2N2O1N3M2O1N2N2N3N1N2VMiDh1Y;VNUE]1l:aNbEQ1`:lNnEg0U:VOYF in this image.", "objects": [{"patches": [210, 232, 233, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 284, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.0016875000000000002, 0.5720624999999999, 0.5945937499999999, 0.9887291666666667], "iscrowd": 0, "area": 41082.5064, "rle": {"size": [480, 640], "counts": "gm03g>6K5J6K5J6K5J6K5K5J6K5J6K5J6K5J6K5J5L5M3N2N2N2N2N2M3N2N2N2N2N2M3N2O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1002N2N1O2N1O2N2N1O2nLRGNo8KVG6l8FTGSECn::SEGm:7TEJl:6SEKm:5RELo:2QEOo:1PE0P;OPE2P;NoD3Q;LoD6P;JoD7Q;InD8R;GnD:R;FmD;S;DmD=S;ClD>T;AlD`0T;@kDa0U;_OjDb0V;]OjDd0V;\\OiDe0W;ZOiDg0W;YOhDh0X;WOhDj0X;VOgDk0Y;UOfDl0Z;SOfDn0Z;ROeDo0[;POeDQ1[;oNdDR1\\;mNdDT1\\;lNcDU1];jNcDX1\\;hNcDY1];gNbDZ1^;eNbD\\1^;dNaD]1^;cNbD^1\\;dNcD]1[;dNeD]1Y;eNfD\\1X;fNgD[1Y;dNgD]1[;aNdD`1^;]NbDd1`;ZN_Dg1c;VN]Dk1e;SN[Dm1f;RNYDo1[<1O1O1O1O1O1O1O2N1O0000O1000000O1000000O10000O1000000O1000001N1000000O1000000O10000O1000000O1000000O1000001N101O1O1N101O1O1N101O1N2O002N1N3N1O1O2M2N3N1N2N3M2O1N3M8I6Ia`i3"}, "label": "a hand holding a chocolate covered doughnut with sprinkles"}]} {"id": 11282, "image": "COCO_train2014_000000011282.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hand holding a chocolate donut\"."}], "task": "refering", "answer_template": "The \"hand holding a chocolate donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 232, 233, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 284, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.0016875000000000002, 0.5720624999999999, 0.5945937499999999, 0.9887291666666667], "iscrowd": 0, "area": 41082.5064, "rle": {"size": [480, 640], "counts": "gm03g>6K5J6K5J6K5J6K5K5J6K5J6K5J6K5J6K5J5L5M3N2N2N2N2N2M3N2N2N2N2N2M3N2O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1002N2N1O2N1O2N2N1O2nLRGNo8KVG6l8FTGSECn::SEGm:7TEJl:6SEKm:5RELo:2QEOo:1PE0P;OPE2P;NoD3Q;LoD6P;JoD7Q;InD8R;GnD:R;FmD;S;DmD=S;ClD>T;AlD`0T;@kDa0U;_OjDb0V;]OjDd0V;\\OiDe0W;ZOiDg0W;YOhDh0X;WOhDj0X;VOgDk0Y;UOfDl0Z;SOfDn0Z;ROeDo0[;POeDQ1[;oNdDR1\\;mNdDT1\\;lNcDU1];jNcDX1\\;hNcDY1];gNbDZ1^;eNbD\\1^;dNaD]1^;cNbD^1\\;dNcD]1[;dNeD]1Y;eNfD\\1X;fNgD[1Y;dNgD]1[;aNdD`1^;]NbDd1`;ZN_Dg1c;VN]Dk1e;SN[Dm1f;RNYDo1[<1O1O1O1O1O1O1O2N1O0000O1000000O1000000O10000O1000000O1000001N1000000O1000000O10000O1000000O1000000O1000001N101O1O1N101O1O1N101O1N2O002N1N3N1O1O2M2N3N1N2N3M2O1N3M8I6Ia`i3"}, "label": "hand holding a chocolate donut"}]} {"id": 265235, "image": "COCO_train2014_000000265235.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cake made of cupcakes shaped like a baseball bat\"."}], "task": "refering", "answer_template": "The \"a cake made of cupcakes shaped like a baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 62, 63, 64, 65, 83, 84, 85, 86, 87, 88, 104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 133, 148, 149, 150, 151, 152, 153, 154, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 215, 216, 217, 218, 219, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306], "bbox": [0.0124375, 0.12467362924281984, 0.8471406250000001, 0.9844125326370756], "iscrowd": 0, "area": 51520.70135000001, "rle": {"size": [383, 640], "counts": "X[38b;5K5K5K5M3O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O100O1O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O10000O100O100O100O100O10O0100O100O10000O100O100O100O100O10000O100O100O100O100O10000O100O100O100O10000O100O100O100O100O10000O100O100O100O100O10000O100O100O100O100O100O1O100O1O100O2N100O1O100O2O0O1O10O01O010O0010O01O010O00010O0010O01O010O0010O01O010O010O0010O01O01O01O010O0010O01O010O0010O01O010O0010O00010O010O01O010O0010O01O010O0010O01O010O00010O0010O01O010O0010O001N100O2O000O2O0O101N100O2O000O101N100O101O0O100O2O0O100O2O000O100100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1O01O0000001O0000001O0000001O000000001O0000001O0000001O00000010O001O0010O0001O010O001O01O01O010O010O00010O010O010O010O01O010O01O010O01O010O010O010O00001N101O001N101O0O2O000O2O001N101O0O2O001O0O20O0001O001O001O001O001O01O01O001O001O001O010O010O00010O0010O010O010O010O1O010O010O10O010O010O0100O010O010O10O010O010O010O0100O010O010O010O010O010O10O010O010O010O010O10O010O10O0010O010O01O010O010O010O0010O0100O1O010O100O00100O2O1N3M2O1N3N1N2N2O1N2M3M3M3M3M3M3N2M3M3M3N2M5L3L5L4K4K6G\\bT1"}, "label": "a cake made of cupcakes shaped like a baseball bat"}]} {"id": 265235, "image": "COCO_train2014_000000265235.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cupcake cake made into a baseball bat\"."}], "task": "refering", "answer_template": "The \"a cupcake cake made into a baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 62, 63, 64, 65, 83, 84, 85, 86, 87, 88, 104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 133, 148, 149, 150, 151, 152, 153, 154, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 215, 216, 217, 218, 219, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306], "bbox": [0.0124375, 0.12467362924281984, 0.8471406250000001, 0.9844125326370756], "iscrowd": 0, "area": 51520.70135000001, "rle": {"size": [383, 640], "counts": "X[38b;5K5K5K5M3O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O100O1O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O10000O100O100O100O100O10O0100O100O10000O100O100O100O100O10000O100O100O100O100O10000O100O100O100O10000O100O100O100O100O10000O100O100O100O100O10000O100O100O100O100O100O1O100O1O100O2N100O1O100O2O0O1O10O01O010O0010O01O010O00010O0010O01O010O0010O01O010O010O0010O01O01O01O010O0010O01O010O0010O01O010O0010O00010O010O01O010O0010O01O010O0010O01O010O00010O0010O01O010O0010O001N100O2O000O2O0O101N100O2O000O101N100O101O0O100O2O0O100O2O000O100100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1O01O0000001O0000001O0000001O000000001O0000001O0000001O00000010O001O0010O0001O010O001O01O01O010O010O00010O010O010O010O01O010O01O010O01O010O010O010O00001N101O001N101O0O2O000O2O001N101O0O2O001O0O20O0001O001O001O001O001O01O01O001O001O001O010O010O00010O0010O010O010O010O1O010O010O10O010O010O0100O010O010O10O010O010O010O0100O010O010O010O010O010O10O010O010O010O010O10O010O10O0010O010O01O010O010O010O0010O0100O1O010O100O00100O2O1N3M2O1N3N1N2N2O1N2M3M3M3M3M3M3N2M3M3M3N2M5L3L5L4K4K6G\\bT1"}, "label": "a cupcake cake made into a baseball bat"}]} {"id": 265235, "image": "COCO_train2014_000000265235.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown mit cake\"."}], "task": "refering", "answer_template": "The \"the brown mit cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 205, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320], "bbox": [0.596890625, 0.40180156657963445, 1.0, 0.9810182767624022], "iscrowd": 0, "area": 43135.9703, "rle": {"size": [383, 640], "counts": "SS_41m;3L3cEMk86RG2f80XG9_8J^G>Y8FdG`0V8BiG?U8DhG>V8DhG>U8FhG in this image.", "objects": [{"patches": [132, 133, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 205, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320], "bbox": [0.596890625, 0.40180156657963445, 1.0, 0.9810182767624022], "iscrowd": 0, "area": 43135.9703, "rle": {"size": [383, 640], "counts": "SS_41m;3L3cEMk86RG2f80XG9_8J^G>Y8FdG`0V8BiG?U8DhG>V8DhG>U8FhG in this image.", "objects": [{"patches": [49, 50, 51, 52, 53, 71, 72, 73, 74, 75, 76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 101, 117, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213], "bbox": [0.072625, 0.16180156657963446, 0.44109375, 0.6898955613577024], "iscrowd": 0, "area": 33973.19435, "rle": {"size": [383, 640], "counts": "P]a02e;`0G8H4L4L3M3N2mNhNZG[1`8lN]GV1]8QO`GQ1Z8UOdGl0W8[OfGf0X8^OeGc0Y8BcG?[8EbG<[8I_G;_8c1O2N2N2N2N2N1O2N2N2oNULkIm3n5\\LnIf3P6]LlIe3S6^LjId3T6_LiIc3U6`LhIb3U6cLgI^3X6eLeI\\3Z6gLcI[3[6U1O1N2O2N1O1O1N3N1000001N100000000O10O100000000000O01000000000000O1O1O1O1O1O1O2BkIoJV6Q5<0001N10001O000010O0010O01O1O001N101O001O2N1O1N2O001O001O000O01000O1O100O100O100O010O100O1O100O0100000000O1000000O101O0O10000O10001N100O2M2O2N2N1N3N1O2O1N100O100O101N1M3K5I7I7I7N3O1O001O1O1O00;E00000000000O10001O000000000000O1O1O100O1O1000000001O01O010O0010O01O0010O01O010O00O2NoNMV9]ORHb0lNO_:MeE2\\:JiE3Y:FnE8n:O2M2N2N3NlfU4"}, "label": "the cupcakes with a baseball desing over them"}]} {"id": 265235, "image": "COCO_train2014_000000265235.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cupcakes in the shape of a baseball\"."}], "task": "refering", "answer_template": "The \"cupcakes in the shape of a baseball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 52, 53, 71, 72, 73, 74, 75, 76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 101, 117, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 147, 162, 163, 164, 165, 166, 167, 168, 169, 170, 185, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213], "bbox": [0.072625, 0.16180156657963446, 0.44109375, 0.6898955613577024], "iscrowd": 0, "area": 33973.19435, "rle": {"size": [383, 640], "counts": "P]a02e;`0G8H4L4L3M3N2mNhNZG[1`8lN]GV1]8QO`GQ1Z8UOdGl0W8[OfGf0X8^OeGc0Y8BcG?[8EbG<[8I_G;_8c1O2N2N2N2N2N1O2N2N2oNULkIm3n5\\LnIf3P6]LlIe3S6^LjId3T6_LiIc3U6`LhIb3U6cLgI^3X6eLeI\\3Z6gLcI[3[6U1O1N2O2N1O1O1N3N1000001N100000000O10O100000000000O01000000000000O1O1O1O1O1O1O2BkIoJV6Q5<0001N10001O000010O0010O01O1O001N101O001O2N1O1N2O001O001O000O01000O1O100O100O100O010O100O1O100O0100000000O1000000O101O0O10000O10001N100O2M2O2N2N1N3N1O2O1N100O100O101N1M3K5I7I7I7N3O1O001O1O1O00;E00000000000O10001O000000000000O1O1O100O1O1000000001O01O010O0010O01O0010O01O010O00O2NoNMV9]ORHb0lNO_:MeE2\\:JiE3Y:FnE8n:O2M2N2N3NlfU4"}, "label": "cupcakes in the shape of a baseball"}]} {"id": 298014, "image": "COCO_train2014_000000298014.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top of a moving car\"."}], "task": "refering", "answer_template": "The \"the top of a moving car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [337, 338, 339, 340, 341, 342, 343, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.4971875, 0.8606666666666667, 1.0, 0.9842708333333333], "iscrowd": 0, "area": 14517.859649999999, "rle": {"size": [480, 640], "counts": "f^f41n>100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O2O0O100O1O100O100O100O100O1O100O100O100O100O1O100LQOVBP1m=0LPOWBQ1i=oNWBQ1i=oNVBR1i=300O10000O100O10000O100O10000O100O10000O100000000O10000000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O100000000000000O10000000000000000001O00000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O000000000000000000000000000000000000001O0000001O000000001O000000001O0000001O000000001O000000001O0000001O000000001O00001O0000001O0000001O00001O0000001O00001O4L?A?A_A"}, "label": "the top of a moving car"}]} {"id": 298014, "image": "COCO_train2014_000000298014.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black car going through an intersection\"."}], "task": "refering", "answer_template": "The \"black car going through an intersection\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [337, 338, 339, 340, 341, 342, 343, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.4971875, 0.8606666666666667, 1.0, 0.9842708333333333], "iscrowd": 0, "area": 14517.859649999999, "rle": {"size": [480, 640], "counts": "f^f41n>100O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O2O0O100O1O100O100O100O100O1O100O100O100O100O1O100LQOVBP1m=0LPOWBQ1i=oNWBQ1i=oNVBR1i=300O10000O100O10000O100O10000O100O10000O100000000O10000000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O100000000000000O10000000000000000O100000000000000O100000000000000O10000000000000000001O00000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O000000000000000000000000000000000000001O0000001O000000001O000000001O0000001O000000001O000000001O0000001O000000001O00001O0000001O0000001O00001O0000001O00001O4L?A?A_A"}, "label": "black car going through an intersection"}]} {"id": 27679, "image": "COCO_train2014_000000027679.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"inside of the oven\"."}], "task": "refering", "answer_template": "The \"inside of the oven\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.0050625, 0.18427083333333333, 0.9960625000000001, 0.7460625000000001], "iscrowd": 0, "area": 166457.22065, "rle": {"size": [480, 640], "counts": "ne1U2g:T2lMT2RNn100000000000000001O00000000000000000000000001O00000000000000001O00000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000O2O000000000000000000000O1000000000000000000000000O10000000000000000000000O10000000d0\\OX1hNW1iNX1hNX1hNX1hNUS1"}, "label": "inside of the oven"}]} {"id": 27679, "image": "COCO_train2014_000000027679.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"apple pie inside of an oven\"."}], "task": "refering", "answer_template": "The \"apple pie inside of an oven\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298], "bbox": [0.0050625, 0.18427083333333333, 0.9960625000000001, 0.7460625000000001], "iscrowd": 0, "area": 166457.22065, "rle": {"size": [480, 640], "counts": "ne1U2g:T2lMT2RNn100000000000000001O00000000000000000000000001O00000000000000001O00000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000O2O000000000000000000000O1000000000000000000000000O10000000000000000000000O10000000d0\\OX1hNW1iNX1hNX1hNX1hNUS1"}, "label": "apple pie inside of an oven"}]} {"id": 27679, "image": "COCO_train2014_000000027679.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bottom portion of the oven door\"."}], "task": "refering", "answer_template": "The \"the bottom portion of the oven door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0050625, 0.7367083333333333, 1.0, 1.0], "iscrowd": 0, "area": 77146.83000000002, "rle": {"size": [480, 640], "counts": "[h1c3];000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000"}, "label": "the bottom portion of the oven door"}]} {"id": 27679, "image": "COCO_train2014_000000027679.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lower front panel of a prince oven\"."}], "task": "refering", "answer_template": "The \"a lower front panel of a prince oven\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0050625, 0.7367083333333333, 1.0, 1.0], "iscrowd": 0, "area": 77146.83000000002, "rle": {"size": [480, 640], "counts": "[h1c3];000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000"}, "label": "a lower front panel of a prince oven"}]} {"id": 273439, "image": "COCO_train2014_000000273439.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bus that is close to the camera\"."}], "task": "refering", "answer_template": "The \"a bus that is close to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 59, 60, 61, 62, 63, 79, 80, 81, 82, 83, 84, 85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 110, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 293], "bbox": [0.41792187500000005, 0.16541666666666666, 0.8571562500000001, 0.8454398148148147], "iscrowd": 0, "area": 66956.76495, "rle": {"size": [432, 640], "counts": "_m`31]=;E in this image.", "objects": [{"patches": [56, 57, 58, 59, 60, 61, 62, 63, 79, 80, 81, 82, 83, 84, 85, 86, 102, 103, 104, 105, 106, 107, 108, 109, 110, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 287, 288, 293], "bbox": [0.41792187500000005, 0.16541666666666666, 0.8571562500000001, 0.8454398148148147], "iscrowd": 0, "area": 66956.76495, "rle": {"size": [432, 640], "counts": "_m`31]=;E in this image.", "objects": [{"patches": [266, 267, 268, 269, 285, 286, 287, 288, 304, 305, 306, 307, 323, 324, 325, 326, 342, 343, 344, 361, 362, 363, 364, 380, 381, 382, 383, 384, 399, 400, 401, 402, 403, 404, 418, 419, 420, 421, 422, 423], "bbox": [0.0, 0.62978125, 0.303314606741573, 1.0], "iscrowd": 0, "area": 24720.231649999998, "rle": {"size": [640, 534], "counts": "Y>7o1X1W>UO`An2Y<_M^Cb2^N1O2N1O2N1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O2N1O2N1O0000O01000O010O10O01000O101O0O100O10000O100N2M6J6I7J6IVmW7"}, "label": "a woman in a grey shirt with a purse walking by a pile of plush bears"}]} {"id": 330785, "image": "COCO_train2014_000000330785.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with short bob haircut and glasses with a purse on her right shoulder\"."}], "task": "refering", "answer_template": "The \"woman with short bob haircut and glasses with a purse on her right shoulder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [266, 267, 268, 269, 285, 286, 287, 288, 304, 305, 306, 307, 323, 324, 325, 326, 342, 343, 344, 361, 362, 363, 364, 380, 381, 382, 383, 384, 399, 400, 401, 402, 403, 404, 418, 419, 420, 421, 422, 423], "bbox": [0.0, 0.62978125, 0.303314606741573, 1.0], "iscrowd": 0, "area": 24720.231649999998, "rle": {"size": [640, 534], "counts": "Y>7o1X1W>UO`An2Y<_M^Cb2^N1O2N1O2N1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O2N1O2N1O0000O01000O010O10O01000O101O0O100O10000O100N2M6J6I7J6IVmW7"}, "label": "woman with short bob haircut and glasses with a purse on her right shoulder"}]} {"id": 298017, "image": "COCO_train2014_000000298017.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman holding umbrella\"."}], "task": "refering", "answer_template": "The \"woman holding umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 110, 111, 112, 113, 121, 122, 123, 124, 125, 126, 134, 137, 138, 139, 140, 150, 151, 152, 153, 163, 164, 165, 166, 176, 177, 178, 179, 180, 190, 191, 192, 193], "bbox": [0.32432, 0.38426, 0.8657333333333332, 0.8352], "iscrowd": 0, "area": 16055.677599999983, "rle": {"size": [500, 375], "counts": "hjk15^?6K2M100O1O100O100O100O1O010O100O010O1O010O10O0100O010O1O010O10O10O10O10000O1000O10O1000O010O10O10O10O01000O0100O01000O001N2N2O0O2O10O0100O013M2M4M2N3L3N3@dNbB^1R=mNlBW1nXEl05dNe:`0XEi04eNg:a0WEg03eNi:e0UEd04cNj:h0TEc06_Nh:n0TEb0h;^OYD`0g;@[D>f;A\\Da;@`D`0b;]O_De0`;YObDh0T=4M3L4M1O2N00O001O1O1N2N2M2N3M3I7L7H7JaQh0"}, "label": "woman holding umbrella"}]} {"id": 298017, "image": "COCO_train2014_000000298017.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a long , unhooded coat , poses with her upside - down umbrella on a sidewalk\"."}], "task": "refering", "answer_template": "The \"a woman in a long , unhooded coat , poses with her upside - down umbrella on a sidewalk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 110, 111, 112, 113, 121, 122, 123, 124, 125, 126, 134, 137, 138, 139, 140, 150, 151, 152, 153, 163, 164, 165, 166, 176, 177, 178, 179, 180, 190, 191, 192, 193], "bbox": [0.32432, 0.38426, 0.8657333333333332, 0.8352], "iscrowd": 0, "area": 16055.677599999983, "rle": {"size": [500, 375], "counts": "hjk15^?6K2M100O1O100O100O100O1O010O100O010O1O010O10O0100O010O1O010O10O10O10O10000O1000O10O1000O010O10O10O10O01000O0100O01000O001N2N2O0O2O10O0100O013M2M4M2N3L3N3@dNbB^1R=mNlBW1nXEl05dNe:`0XEi04eNg:a0WEg03eNi:e0UEd04cNj:h0TEc06_Nh:n0TEb0h;^OYD`0g;@[D>f;A\\Da;@`D`0b;]O_De0`;YObDh0T=4M3L4M1O2N00O001O1O1N2N2M2N3M3I7L7H7JaQh0"}, "label": "a woman in a long , unhooded coat , poses with her upside - down umbrella on a sidewalk"}]} {"id": 175142, "image": "COCO_train2014_000000175142.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bench that the gentleman wearing a light blue hat is sitting on\"."}], "task": "refering", "answer_template": "The \"the bench that the gentleman wearing a light blue hat is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 126, 127, 132, 133, 134, 135, 136, 137, 144, 150, 151, 152, 153, 154, 155, 156, 164, 168, 169, 170, 171, 172, 173, 174, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 209, 210], "bbox": [0.0015, 0.57996996996997, 0.72, 0.9898798798798799], "iscrowd": 0, "area": 20230.496249999997, "rle": {"size": [333, 500], "counts": "``0d1i8OgHgNa5W1ZJoNf5P1[JPOe5o0\\JQOd5o0\\JQOd5n0]JROc5m0^JSOb5l0_JTOa5k0`JUO`5k0_JVOa5i0`JWO`5i0`JWO`5i0`JWO`5i0`JWO`5i0`JWO`5h0`JYO`5g0`JYO`5g0`JYO`5g0`JYO`5f0aJZO_5f0aJZO_5f0`J[O`5d0aJ\\O_5c0bJ]O^5a0dJ_O]5?dJA\\5=fJCZ5;gJFY58iJHW56kJJU54mJLS51PKOP50QK0o4ORK1n4NRK3n4MRK3n4LSK4m4KTK5l4JUK6k4IVK7j4IUK8k4GVK9j4EXK;h4DYKeIB[6>eIB[6?dIA\\6?dIA\\6`0cI@]6`0cI@]6a0bI_O^6a0bI_O^6b0aI^O_6c0aI\\O_6d0bI[O^6f0bIYO^6g0bIYO^6h0bIWO^6i0cIVO]6k0cITO]6l0dISO\\6n0cIRO]6n0dIQO\\6P1dIoN\\6R1dImN\\6S1eIlN[6T1eIkN\\6U1eIjN[6V1fIiNZ6X1gIfNY6Z1hIeNX6[1iIdNW6\\1jIcNV6]1kIbNU6^1lIaNT6_1mI`NS6a1nI]NR6c1oI\\NQ6d1PJ[NP6e1PJ[NP6e1PJ[NP6e1PJ[NP6e1PJ[NP6e1PJ[NP6e1PJ[NP6e1PJ[NP6e1PJ[NP6e1PJ[NP6e1PJ[NP6e1PJ[NP6e1PJZNQ6f1oIZNQ6f1oIZNQ6f1oIZNQ6f1oIZNQ6f1oIZNQ6f1mI\\NS6d1lI]NT6d1iI^NW6b1hI_NX6b1eI`N[6a1cI`N]6`1aIbN_6_1_IbNa6_1\\IcNd6]1[IdNe6]1XIeNh6\\1VIeNj6[1TIgNl6Z1RIgNn6Z1oHhNQ7X1mHjNS7X2000O10000000000000000000000000000000O1000000000000000O100000000000000000000000000000000000fImLZ5S3^JUMb5k2VJ]Mj5c2mIfMS6Z35K6K5C=B>A?A=LO1O001O100O1O1O1O1O100000000000000001O000000000O100000001O12N:F:F7IO1O1O2O4]NhHG\\7NSIJR7I]I0g6EhI3a8HgV]1"}, "label": "the bench that the gentleman wearing a light blue hat is sitting on"}]} {"id": 19501, "image": "COCO_train2014_000000019501.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman body surfing\"."}], "task": "refering", "answer_template": "The \"woman body surfing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 119, 123, 124, 125, 127, 128, 129, 140, 141, 142, 143, 144, 146, 147, 148, 149, 150, 151, 152, 153, 155, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 239, 240, 241, 242, 243, 244], "bbox": [0.09081249999999999, 0.30493750000000003, 0.8139062499999999, 0.6322916666666667], "iscrowd": 0, "area": 36913.66445000001, "rle": {"size": [480, 640], "counts": "^jk01n>101O0O2O0O101O0O2O0O2O000O2O001N101N100O2O001N101N10001N101N100O2O001N2O1N101N2O1O0O2O1N2O00100O1O00100O1O10O10O100O01000O10O10O100O01000O10O0100O010O100O00100O010O1O010O100O00100O010O1O010O1O010O10O01O10O010O01O10O10O10O100O10000O10O10O100O10000O10O10O100O10000O10O10O10O000000001O000000001O0000000010O00000001O000000001O00001O1O001O011`NVNfEl1W:UNiEm1U:SNkEo1R:RNnEP2o9RNPFn1o9SNQFn1m9SNSFm1l9UNSFl1l9UNSFk1l9WNSFj1l9VNTFj1k9YNSFh1l9YNSFh1k9ZNTFf1l9[NSFe1l9]NSFd1l9^NRFb1n9_NQFa1n9aNQF_1o9bNPF_1n9cNQF]1o9eNPFZ1o9hNPFY1o9hNPFX1o9jNPFV1P:kNoEU1P:mNoET1P:nNnER1Q:POnEP1R:QOlEP1S:SOjEo0V:QOiEo0X:ROeEo0\\:QObEQ1^:PO`EP1a:PO]EQ1c:QO[Eo0f:f10O1O010O1O00100O001O10O01O1O010O1O001O010O1O2N101N2N2N2N101N2N2N101N2N1OO1N3N1O1N2O2M2O1N2O1N3N11O000001O00000010O00000001OO10001O0000000000000O2ON2L4L4L4L40000O10O10O10000O10000O1000O01000000000000001O0O1000000000000000001N100000000O100O100O2O001N2O0O2O0O2O1O0O2O0O2O0O2M3L3N3M2M4M3L4M3M3N3L3M4L3M4M1N3M3O001O001O10O01O001O1O001O001O10O01O001O1O001O000010O00000001O0000001O01O000001O00000000O101O0O10000O10000O10000O10000O10000O10000O10000001O001O1O001O001O1O0O2O001O001O1O1O2N1O1N2O1O2N1M3M3N2M4M2M3M3N2M=Diag1"}, "label": "woman body surfing"}]} {"id": 19501, "image": "COCO_train2014_000000019501.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a blue bikini\"."}], "task": "refering", "answer_template": "The \"a woman in a blue bikini\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 119, 123, 124, 125, 127, 128, 129, 140, 141, 142, 143, 144, 146, 147, 148, 149, 150, 151, 152, 153, 155, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 239, 240, 241, 242, 243, 244], "bbox": [0.09081249999999999, 0.30493750000000003, 0.8139062499999999, 0.6322916666666667], "iscrowd": 0, "area": 36913.66445000001, "rle": {"size": [480, 640], "counts": "^jk01n>101O0O2O0O101O0O2O0O2O000O2O001N101N100O2O001N101N10001N101N100O2O001N2O1N101N2O1O0O2O1N2O00100O1O00100O1O10O10O100O01000O10O10O100O01000O10O0100O010O100O00100O010O1O010O100O00100O010O1O010O1O010O10O01O10O010O01O10O10O10O100O10000O10O10O100O10000O10O10O100O10000O10O10O10O000000001O000000001O0000000010O00000001O000000001O00001O1O001O011`NVNfEl1W:UNiEm1U:SNkEo1R:RNnEP2o9RNPFn1o9SNQFn1m9SNSFm1l9UNSFl1l9UNSFk1l9WNSFj1l9VNTFj1k9YNSFh1l9YNSFh1k9ZNTFf1l9[NSFe1l9]NSFd1l9^NRFb1n9_NQFa1n9aNQF_1o9bNPF_1n9cNQF]1o9eNPFZ1o9hNPFY1o9hNPFX1o9jNPFV1P:kNoEU1P:mNoET1P:nNnER1Q:POnEP1R:QOlEP1S:SOjEo0V:QOiEo0X:ROeEo0\\:QObEQ1^:PO`EP1a:PO]EQ1c:QO[Eo0f:f10O1O010O1O00100O001O10O01O1O010O1O001O010O1O2N101N2N2N2N101N2N2N101N2N1OO1N3N1O1N2O2M2O1N2O1N3N11O000001O00000010O00000001OO10001O0000000000000O2ON2L4L4L4L40000O10O10O10000O10000O1000O01000000000000001O0O1000000000000000001N100000000O100O100O2O001N2O0O2O0O2O1O0O2O0O2O0O2M3L3N3M2M4M3L4M3M3N3L3M4L3M4M1N3M3O001O001O10O01O001O1O001O001O10O01O001O1O001O000010O00000001O0000001O01O000001O00000000O101O0O10000O10000O10000O10000O10000O10000O10000001O001O1O001O001O1O0O2O001O001O1O1O2N1O1N2O1O2N1M3M3N2M4M2M3M3N2M=Diag1"}, "label": "a woman in a blue bikini"}]} {"id": 19501, "image": "COCO_train2014_000000019501.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person cheering in the white t - shirt\"."}], "task": "refering", "answer_template": "The \"the person cheering in the white t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 34, 35, 36, 37, 38, 58, 59, 60, 61, 62, 82, 83, 84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 203, 204, 205, 206, 227, 228, 229, 251, 252, 274, 275, 296, 297, 298, 320, 321], "bbox": [0.47865624999999995, 0.008979166666666667, 0.994375, 0.7775208333333333], "iscrowd": 0, "area": 36018.99479999999, "rle": {"size": [480, 640], "counts": "Tm_42n>1O1O2N100O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O2N10001N1000001N10001O0O101O000O10O10O100000O01000000O0100000O01000O1000O10O100000O01000000O01000O1000O10O1000O10O100000O010000O10O1000O1000O10O1000O01000000O0100000O10O1000O1000O01000000O0100000O0100000O10O10O100000O01000000O0100000O10O10O1000O10O100000O010000O10O01O1O100O00100O1O00100O1O1O010O1O1O10O01O1O100O001O100O1O00100O1O00100O1O1O010O1O1O10O01O1O100O00100O1O00100O1O1O010O1O1O10O01O1O100O001O100O1O00100O1O00100O1O1O010O1aHkKQ4V4eKSL[4n3ZK\\Lg4c3PKfLP5[3fJnLZ5S3[JWMf5i2PJ`MP6a2fIiMY6W2^IRNb6o1SI[Nn6e1hHdNX7b400001O01O0001O0000001O0001O01O0000001O0000010O0000001O00000010O000001O00001O0001O01O00P1PO\\3dL\\c1"}, "label": "the person cheering in the white t - shirt"}]} {"id": 19501, "image": "COCO_train2014_000000019501.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the arm of someone in a white shirt\"."}], "task": "refering", "answer_template": "The \"the arm of someone in a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 34, 35, 36, 37, 38, 58, 59, 60, 61, 62, 82, 83, 84, 85, 86, 87, 88, 106, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 203, 204, 205, 206, 227, 228, 229, 251, 252, 274, 275, 296, 297, 298, 320, 321], "bbox": [0.47865624999999995, 0.008979166666666667, 0.994375, 0.7775208333333333], "iscrowd": 0, "area": 36018.99479999999, "rle": {"size": [480, 640], "counts": "Tm_42n>1O1O2N100O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O2N10001N1000001N10001O0O101O000O10O10O100000O01000000O0100000O01000O1000O10O100000O01000000O01000O1000O10O1000O10O100000O010000O10O1000O1000O10O1000O01000000O0100000O10O1000O1000O01000000O0100000O0100000O10O10O100000O01000000O0100000O10O10O1000O10O100000O010000O10O01O1O100O00100O1O00100O1O1O010O1O1O10O01O1O100O001O100O1O00100O1O00100O1O1O010O1O1O10O01O1O100O00100O1O00100O1O1O010O1O1O10O01O1O100O001O100O1O00100O1O00100O1O1O010O1aHkKQ4V4eKSL[4n3ZK\\Lg4c3PKfLP5[3fJnLZ5S3[JWMf5i2PJ`MP6a2fIiMY6W2^IRNb6o1SI[Nn6e1hHdNX7b400001O01O0001O0000001O0001O01O0000001O0000010O0000001O00000010O000001O00001O0001O01O00P1PO\\3dL\\c1"}, "label": "the arm of someone in a white shirt"}]} {"id": 240689, "image": "COCO_train2014_000000240689.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a white t - shirt and multi - colored shorts posing at the beach\"."}], "task": "refering", "answer_template": "The \"a boy in a white t - shirt and multi - colored shorts posing at the beach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 221, 222, 223, 224, 244, 245, 246, 247, 248, 269, 292], "bbox": [0.636203125, 0.5624, 0.827328125, 0.8285411764705882], "iscrowd": 0, "area": 3621.425600000001, "rle": {"size": [425, 640], "counts": "WVY51V=6J3M3N1001O1O010O001O1O010O0000O101OO01000000O1VOEeD<[;HaD8^;L^D4b;1ZDOe;6VDKi;6VDKi;6WDJh;7WDJh;7XDIh;8VDHj;9UDFm;:SDEn;l0N2O1O1O1N2O1O2M2O11O1O1O1OeEbNk8]1jFmN_9[20O01O1O001ONLYMjEb2W:56K5J6J6WNTEn0R;mNSEn0R;mNSEm0S;nNREn0f;01O100O1O1O010O1O100O001O1O1N2O1O001N2O1O1O0O1000000O100000000O10000000000000000000001O00100O010000O10O1L3Mlh]1"}, "label": "a boy in a white t - shirt and multi - colored shorts posing at the beach"}]} {"id": 240689, "image": "COCO_train2014_000000240689.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"young boy balancing on one leg\"."}], "task": "refering", "answer_template": "The \"young boy balancing on one leg\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 221, 222, 223, 224, 244, 245, 246, 247, 248, 269, 292], "bbox": [0.636203125, 0.5624, 0.827328125, 0.8285411764705882], "iscrowd": 0, "area": 3621.425600000001, "rle": {"size": [425, 640], "counts": "WVY51V=6J3M3N1001O1O010O001O1O010O0000O101OO01000000O1VOEeD<[;HaD8^;L^D4b;1ZDOe;6VDKi;6VDKi;6WDJh;7WDJh;7XDIh;8VDHj;9UDFm;:SDEn;l0N2O1O1O1N2O1O2M2O11O1O1O1OeEbNk8]1jFmN_9[20O01O1O001ONLYMjEb2W:56K5J6J6WNTEn0R;mNSEn0R;mNSEm0S;nNREn0f;01O100O1O1O010O1O100O001O1O1N2O1O001N2O1O1O0O1000000O100000000O10000000000000000000001O00100O010000O10O1L3Mlh]1"}, "label": "young boy balancing on one leg"}]} {"id": 560180, "image": "COCO_train2014_000000560180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"several tennis racquets held in a chain link fence\"."}], "task": "refering", "answer_template": "The \"several tennis racquets held in a chain link fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 124, 125, 126, 145, 146, 147, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 203, 204, 205, 206, 207, 208, 209, 210, 224, 225, 226, 227, 228, 229, 230, 231, 232, 246, 247, 248, 249, 250, 251, 252, 253, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 310, 311, 312, 313, 314, 315, 316, 317, 318, 332, 333, 334, 335, 336, 337, 338, 339, 354, 355, 356, 357, 358, 359, 360, 361, 377, 378, 379, 380, 381], "bbox": [0.11684640522875818, 0.18426470588235294, 0.7797712418300653, 0.8134803921568627], "iscrowd": 0, "area": 53513.04965000001, "rle": {"size": [612, 612], "counts": "[^[17ib07H7I8H8I6I8H7I8I7H7I8H8I6I8K5L3L5L1N2O2N1N2O2M2O2M2O1O2M2O1N3N1O1N3N1N3N1O1N3N1N2O2M2O2N1N2O2M2O1O2M2O1N3N1N3N1O1N3N1N2O2N1N3N1N2O2M2O1O2O0O1O1O00100O1O1O1O100O1O1O001O100O1O1O1O100O1O1O00100O1O1O1O1O100O1O001000000000000O1000O10000000O100000000000O10O01O100O1O100O001O100O1O1O100O00100O1O1O100O001O100O1O1O100O001O100O1O1O10O01O1O100O1O1O10O01O1O100O1O0010O01O10O01O1O010O001O10O01O00100N1O2N2O0O2N1O2N2N101N1O2N2N1O2N2O0O2N1O2N2N101N2N1O2N1O2O1N1O2N2N2N101N2N1O2N2N101M3L4M2M4M3L3N3L4M3L3N3L4M2M4M3L3N3L4N2N101O1N101N2O1O0O2O1N101O1N101N2O1O0O2O1N101O1N2O000O100O101O0O100O10001N10000O1000001N1000000O10001O0O1000000O2O00000O1000001N1000000O10001O0O100000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001O00000000001O00O1000000000000000000000000O01000001O2N2N2N2N2N2N2N1O2N2N2N2N2N2Noa`2"}, "label": "several tennis racquets held in a chain link fence"}]} {"id": 560180, "image": "COCO_train2014_000000560180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow tennis racket behind a pink tennis racket\"."}], "task": "refering", "answer_template": "The \"a yellow tennis racket behind a pink tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 124, 125, 126, 145, 146, 147, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 203, 204, 205, 206, 207, 208, 209, 210, 224, 225, 226, 227, 228, 229, 230, 231, 232, 246, 247, 248, 249, 250, 251, 252, 253, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 310, 311, 312, 313, 314, 315, 316, 317, 318, 332, 333, 334, 335, 336, 337, 338, 339, 354, 355, 356, 357, 358, 359, 360, 361, 377, 378, 379, 380, 381], "bbox": [0.11684640522875818, 0.18426470588235294, 0.7797712418300653, 0.8134803921568627], "iscrowd": 0, "area": 53513.04965000001, "rle": {"size": [612, 612], "counts": "[^[17ib07H7I8H8I6I8H7I8I7H7I8H8I6I8K5L3L5L1N2O2N1N2O2M2O2M2O1O2M2O1N3N1O1N3N1N3N1O1N3N1N2O2M2O2N1N2O2M2O1O2M2O1N3N1N3N1O1N3N1N2O2N1N3N1N2O2M2O1O2O0O1O1O00100O1O1O1O100O1O1O001O100O1O1O1O100O1O1O00100O1O1O1O1O100O1O001000000000000O1000O10000000O100000000000O10O01O100O1O100O001O100O1O1O100O00100O1O1O100O001O100O1O1O100O001O100O1O1O10O01O1O100O1O1O10O01O1O100O1O0010O01O10O01O1O010O001O10O01O00100N1O2N2O0O2N1O2N2N101N1O2N2N1O2N2O0O2N1O2N2N101N2N1O2N1O2O1N1O2N2N2N101N2N1O2N2N101M3L4M2M4M3L3N3L4M3L3N3L4M2M4M3L3N3L4N2N101O1N101N2O1O0O2O1N101O1N101N2O1O0O2O1N101O1N2O000O100O101O0O100O10001N10000O1000001N1000000O10001O0O1000000O2O00000O1000001N1000000O10001O0O100000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001O00000000001O00O1000000000000000000000000O01000001O2N2N2N2N2N2N2N1O2N2N2N2N2N2Noa`2"}, "label": "a yellow tennis racket behind a pink tennis racket"}]} {"id": 560180, "image": "COCO_train2014_000000560180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white zebra design tennis bat\"."}], "task": "refering", "answer_template": "The \"a black and white zebra design tennis bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 122, 123, 124, 125, 143, 144, 145, 158, 159, 160, 161, 162, 163, 164, 165, 166, 179, 180, 181, 182, 183, 184, 185, 186, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 264, 265, 266, 267, 287, 288, 289, 309, 310], "bbox": [0.03145424836601307, 0.1820261437908497, 0.707859477124183, 0.67640522875817], "iscrowd": 0, "area": 24750.661000000007, "rle": {"size": [612, 612], "counts": "oe;2ob07I7I7I7I7I7J6I7I6J4M4L4M3M3M3L4M2N3M3L4M3M3M3L3N3M1O1N3M2O1N3M2N2O2M2N2N3N1N2O2N1O1O2N1O1O2N1O2N100O2N1ON2N3M2N2N2M3N3M2N2N2N2N3L3N2N2N2N3M2N2M3O2O0O1000001N10000O2O001O0O101O001N101O0O2O001O0O2O001N101O000O2O001O1N2O001N2O1O001N2O1O001N2O001N2O1O001N2O1O001N2O1O0O2000O1000O10O1000O10000000O01000000O100000000O1000O10O100000000O1000000O010000000O1000000O1000O1000O1000000O10000000O01000000O100000000O10O10O100O10000O100O100O01000O100O100O10000O010O100O10000O100O100O01000O100O100O10000O010O100O1000O010000O1000O010000O1000O010000O10O10O10000O10O10O10000O10O10O10000O0100O00100O00100O010O1O01O00001O001O00001O00010O00001O001O00001O000010O01O01O01O01O01O01O01O010O00010O00010O001O01O01O01O01O01O00010O00010O000010O00010O00010O0001O01O00010O0001O01O01O00010O00001O00001O01O01O00001O000010O000001O000012M3M3N2L4M2N2N2N2N2N2N3L3N[kZ3"}, "label": "a black and white zebra design tennis bat"}]} {"id": 560180, "image": "COCO_train2014_000000560180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and white racquet next to a yellow racquet in a fence\"."}], "task": "refering", "answer_template": "The \"a black and white racquet next to a yellow racquet in a fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 122, 123, 124, 125, 143, 144, 145, 158, 159, 160, 161, 162, 163, 164, 165, 166, 179, 180, 181, 182, 183, 184, 185, 186, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 264, 265, 266, 267, 287, 288, 289, 309, 310], "bbox": [0.03145424836601307, 0.1820261437908497, 0.707859477124183, 0.67640522875817], "iscrowd": 0, "area": 24750.661000000007, "rle": {"size": [612, 612], "counts": "oe;2ob07I7I7I7I7I7J6I7I6J4M4L4M3M3M3L4M2N3M3L4M3M3M3L3N3M1O1N3M2O1N3M2N2O2M2N2N3N1N2O2N1O1O2N1O1O2N1O2N100O2N1ON2N3M2N2N2M3N3M2N2N2N2N3L3N2N2N2N3M2N2M3O2O0O1000001N10000O2O001O0O101O001N101O0O2O001O0O2O001N101O000O2O001O1N2O001N2O1O001N2O1O001N2O001N2O1O001N2O1O001N2O1O0O2000O1000O10O1000O10000000O01000000O100000000O1000O10O100000000O1000000O010000000O1000000O1000O1000O1000000O10000000O01000000O100000000O10O10O100O10000O100O100O01000O100O100O10000O010O100O10000O100O100O01000O100O100O10000O010O100O1000O010000O1000O010000O1000O010000O10O10O10000O10O10O10000O10O10O10000O0100O00100O00100O010O1O01O00001O001O00001O00010O00001O001O00001O000010O01O01O01O01O01O01O01O010O00010O00010O001O01O01O01O01O01O00010O00010O000010O00010O00010O0001O01O00010O0001O01O01O00010O00001O00001O01O01O00001O000010O000001O000012M3M3N2L4M2N2N2N2N2N2N3L3N[kZ3"}, "label": "a black and white racquet next to a yellow racquet in a fence"}]} {"id": 560180, "image": "COCO_train2014_000000560180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black tennis racket in the back\"."}], "task": "refering", "answer_template": "The \"the black tennis racket in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 99, 100, 101, 120, 121, 122, 135, 136, 137, 138, 139, 140, 141, 142, 156, 157, 158, 159, 160, 161, 162, 163, 177, 178, 179, 199, 200, 220, 221], "bbox": [0.04238562091503268, 0.13866013071895425, 0.6497222222222222, 0.4969771241830065], "iscrowd": 0, "area": 12970.788849999997, "rle": {"size": [612, 612], "counts": "ji?j0Sb08L4M3L4M000O100O11O0000000010O1O1O010O100000O010000000O100000000O10000000010O00000001O01O00000001O01O000001O01O001O01O01O010O010O0010O010O010O010O00010O010O010O01O010O010O010O1000O0100O10O10O100O10O10O100O10O10O100O10O10O100O10O1000O10O100000000000O1000000000000O100000001O000O10000000000000000000000O100000000000000000000000000000001O00000O100000001O000O10000000001O000000000000001O000000000000001O00000O100000O1000000000000O10000000000O10O1000000000O010000000O10O0100O10000O010O100O10N101N1O2O0O2N101O1O001O010O001O001O001O00L_Og]O`0Yb0Be]O?Zb05000001O00000010O0001O10O01O10O0001O01O01O01O01O00010O00010O000010O00010O00010O0001O01O01O01O0001O01O00010O0000010O000010O000010O000010O000010O01O010O0010O0010O001O0O2O0010O00100O100O2O0O102M2N2N3M2N3M4M2M3Mg\\P4"}, "label": "the black tennis racket in the back"}]} {"id": 560180, "image": "COCO_train2014_000000560180.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tennis racket far from the camera\"."}], "task": "refering", "answer_template": "The \"tennis racket far from the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 99, 100, 101, 120, 121, 122, 135, 136, 137, 138, 139, 140, 141, 142, 156, 157, 158, 159, 160, 161, 162, 163, 177, 178, 179, 199, 200, 220, 221], "bbox": [0.04238562091503268, 0.13866013071895425, 0.6497222222222222, 0.4969771241830065], "iscrowd": 0, "area": 12970.788849999997, "rle": {"size": [612, 612], "counts": "ji?j0Sb08L4M3L4M000O100O11O0000000010O1O1O010O100000O010000000O100000000O10000000010O00000001O01O00000001O01O000001O01O001O01O01O010O010O0010O010O010O010O00010O010O010O01O010O010O010O1000O0100O10O10O100O10O10O100O10O10O100O10O10O100O10O1000O10O100000000000O1000000000000O100000001O000O10000000000000000000000O100000000000000000000000000000001O00000O100000001O000O10000000001O000000000000001O000000000000001O00000O100000O1000000000000O10000000000O10O1000000000O010000000O10O0100O10000O010O100O10N101N1O2O0O2N101O1O001O010O001O001O001O00L_Og]O`0Yb0Be]O?Zb05000001O00000010O0001O10O01O10O0001O01O01O01O01O00010O00010O000010O00010O00010O0001O01O01O01O0001O01O00010O0000010O000010O000010O000010O000010O01O010O0010O0010O001O0O2O0010O00100O100O2O0O102M2N2N3M2N3M4M2M3Mg\\P4"}, "label": "tennis racket far from the camera"}]} {"id": 199743, "image": "COCO_train2014_000000199743.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing blue t - shirt riding bike on the sidewalk and waiting for traffic\"."}], "task": "refering", "answer_template": "The \"man wearing blue t - shirt riding bike on the sidewalk and waiting for traffic\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 99, 100, 101, 122, 123, 124, 144, 145, 146, 147, 148, 167, 168, 169, 171, 190, 191, 192, 214, 215, 216, 237, 238, 239, 261, 262, 285, 286], "bbox": [0.277734375, 0.24707259953161592, 0.48446875, 0.8608665105386416], "iscrowd": 0, "area": 12408.816699999998, "rle": {"size": [427, 640], "counts": "h]Z27R=6J7I6J3M2N2N2N2N2N2N3M2N3L4M2N3M2M4M2N3M2M3N2M4L4L4J6hGTMm5o2PJTMm5o2PJSMo5Q3mIPMQ6S3lIoLS6T3jImLU6V3gIlLX6V3fIkLY6X3bIkL^6W3ZIoLe6S3XIoLg6T3UInLk6T3QInLo6U3fHSMZ7o2ZH[Mf7U4O1000000O1MVOgHhKo6a4UI\\Kf6h4\\IVKd6j4^IUKc6i4_IVK\\6HRIP5d0WKY6o4iIPKV6P5lIoJT6P5nIoJR6R5nIlJR6X5kIhJU6Z5nIaJR6a5k0:FkMiHZNS7^1VIdNg6Y1]IjN_6T1eInNW6o0mIWOl5f0XJ[Of5c0^J]O`5:jJGU55PKKn42WKN^O`MS4^2dL2ROeM[4U2gL8iNeMa4`0gJe0V2T1]NjM\\OFZ5c0iJg0Y2U1SNnMe5l0\\LQ2e3mM`LP2_3oMfLm1[3QNhLm1X3SNiLl1X3SNmLh1T3WNPMe1Q3ZNRMc1o2\\NdMP1^2POdMm0]2SOfMi0[2WOPN=Q2DPN9R2FQN6P2JRN3P2MRNOo11TNKn14UNGn19SNDo1;QNDQ2;PNCR2=mMBU2=jMCX2=fMC\\2gLBZ3?dLA^3?aL@`3a0^L_Oc3a0\\L^Og3a0YL^Oh3c0VL]Ol3c0SL[OP4d0PLZOS4f0lKWOX4h0V410O2M2N3L3N2N2M2O1N2O0O2Nb\\Y4"}, "label": "man wearing blue t - shirt riding bike on the sidewalk and waiting for traffic"}]} {"id": 199743, "image": "COCO_train2014_000000199743.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man on the bike\"."}], "task": "refering", "answer_template": "The \"the man on the bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 99, 100, 101, 122, 123, 124, 144, 145, 146, 147, 148, 167, 168, 169, 171, 190, 191, 192, 214, 215, 216, 237, 238, 239, 261, 262, 285, 286], "bbox": [0.277734375, 0.24707259953161592, 0.48446875, 0.8608665105386416], "iscrowd": 0, "area": 12408.816699999998, "rle": {"size": [427, 640], "counts": "h]Z27R=6J7I6J3M2N2N2N2N2N2N3M2N3L4M2N3M2M4M2N3M2M3N2M4L4L4J6hGTMm5o2PJTMm5o2PJSMo5Q3mIPMQ6S3lIoLS6T3jImLU6V3gIlLX6V3fIkLY6X3bIkL^6W3ZIoLe6S3XIoLg6T3UInLk6T3QInLo6U3fHSMZ7o2ZH[Mf7U4O1000000O1MVOgHhKo6a4UI\\Kf6h4\\IVKd6j4^IUKc6i4_IVK\\6HRIP5d0WKY6o4iIPKV6P5lIoJT6P5nIoJR6R5nIlJR6X5kIhJU6Z5nIaJR6a5k0:FkMiHZNS7^1VIdNg6Y1]IjN_6T1eInNW6o0mIWOl5f0XJ[Of5c0^J]O`5:jJGU55PKKn42WKN^O`MS4^2dL2ROeM[4U2gL8iNeMa4`0gJe0V2T1]NjM\\OFZ5c0iJg0Y2U1SNnMe5l0\\LQ2e3mM`LP2_3oMfLm1[3QNhLm1X3SNiLl1X3SNmLh1T3WNPMe1Q3ZNRMc1o2\\NdMP1^2POdMm0]2SOfMi0[2WOPN=Q2DPN9R2FQN6P2JRN3P2MRNOo11TNKn14UNGn19SNDo1;QNDQ2;PNCR2=mMBU2=jMCX2=fMC\\2gLBZ3?dLA^3?aL@`3a0^L_Oc3a0\\L^Og3a0YL^Oh3c0VL]Ol3c0SL[OP4d0PLZOS4f0lKWOX4h0V410O2M2N3L3N2N2M2O1N2O0O2Nb\\Y4"}, "label": "the man on the bike"}]} {"id": 166975, "image": "COCO_train2014_000000166975.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man standing in front of bus\"."}], "task": "refering", "answer_template": "The \"man standing in front of bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 100, 101, 102, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 192, 193, 194, 215, 216, 217, 238, 239, 240, 260, 261, 262, 263, 283, 284, 285, 286, 306, 307, 308, 309, 329, 330, 331, 332], "bbox": [0.318125, 0.21594847775175643, 0.5415625000000001, 0.9580327868852458], "iscrowd": 0, "area": 23514.243650000004, "rle": {"size": [427, 640], "counts": "]^e25R=5J5fJDdM>e19ZNG]1b0cN_OZ1d0eN]OW1g0hNYOU1j0lNVOQ1m0nNSOo0Q1POPOm0R1mNTOP1P1bN^OZ1f0YNFe1=mM1o12eM:Y2IZLPNVO]2\\4GVLVNYOV2^4FSL^NZOm1`4IoKdN\\Oe1a4KlKiN^O^1d4KgKRO@T1f4NcKXOAk0k40]KZ2b4hMWK^2i4R30O0100O10O01000M2L5K4L5J5L5K4L5K5JB?N25K4L5J5L4L5K4L5J5L5K4L5K4L51Nc0]Oc0^Ob0]Oc0^Oa0@a0^O`0A0O01O00001O0010O01O001O001YGYLh7g3QH`LP8_3iGiLV8X3bGPM\\8Q4N1O2N2N2N1O2N2N2N1O2N2gMULoJm3g1gJ8P2kMZ3l1oJHQ2ZNR3l1VKYOQ2jNk2l1\\KiNS2XOd2m1bKZNS2E_2Q2[5N2N2O0O2N2O1N101N2N2O0O2N2O1N1O2O1N2O0O2N2O1N1O2O1N1O2O1N2O0O2N2O1N1O2O1N2O0O2N2Kf]j3"}, "label": "man standing in front of bus"}]} {"id": 166975, "image": "COCO_train2014_000000166975.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with black shirt standing in front of bus\"."}], "task": "refering", "answer_template": "The \"man with black shirt standing in front of bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 100, 101, 102, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 192, 193, 194, 215, 216, 217, 238, 239, 240, 260, 261, 262, 263, 283, 284, 285, 286, 306, 307, 308, 309, 329, 330, 331, 332], "bbox": [0.318125, 0.21594847775175643, 0.5415625000000001, 0.9580327868852458], "iscrowd": 0, "area": 23514.243650000004, "rle": {"size": [427, 640], "counts": "]^e25R=5J5fJDdM>e19ZNG]1b0cN_OZ1d0eN]OW1g0hNYOU1j0lNVOQ1m0nNSOo0Q1POPOm0R1mNTOP1P1bN^OZ1f0YNFe1=mM1o12eM:Y2IZLPNVO]2\\4GVLVNYOV2^4FSL^NZOm1`4IoKdN\\Oe1a4KlKiN^O^1d4KgKRO@T1f4NcKXOAk0k40]KZ2b4hMWK^2i4R30O0100O10O01000M2L5K4L5J5L5K4L5K5JB?N25K4L5J5L4L5K4L5J5L5K4L5K4L51Nc0]Oc0^Ob0]Oc0^Oa0@a0^O`0A0O01O00001O0010O01O001O001YGYLh7g3QH`LP8_3iGiLV8X3bGPM\\8Q4N1O2N2N2N1O2N2N2N1O2N2gMULoJm3g1gJ8P2kMZ3l1oJHQ2ZNR3l1VKYOQ2jNk2l1\\KiNS2XOd2m1bKZNS2E_2Q2[5N2N2O0O2N2O1N101N2N2O0O2N2O1N1O2O1N2O0O2N2O1N1O2O1N1O2O1N2O0O2N2O1N1O2O1N2O0O2N2Kf]j3"}, "label": "man with black shirt standing in front of bus"}]} {"id": 470085, "image": "COCO_train2014_000000470085.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant with his trunk touching the mouth of the other elephant\"."}], "task": "refering", "answer_template": "The \"an elephant with his trunk touching the mouth of the other elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 230, 231, 232, 233, 234, 240, 241, 242, 243, 244, 245, 246, 247, 253, 254, 255, 256, 257, 264, 265, 266, 267, 268, 269, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327], "bbox": [0.0015, 0.006604215456674473, 0.788359375, 0.9819437939110071], "iscrowd": 0, "area": 117396.40129999998, "rle": {"size": [427, 640], "counts": "nc0`6;`60000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000kIHWO8i0NQO2n04mNLR1:iNFW1?dNA[1e0`N[O_1k0\\NUOd1P1WNPOh1V1SNjNl1\\1oMdNP2b1kM^NU2h1eMXNZ2n1aMRN^2T2]MlMc2Y2WMhMh2^2SMbMn2b2mL^MS3g2hLYMX3l2cLTM]3Q3^LoLb3V3YLjLg3\\3SLdLm3a3nK_LR4f3iKZLX4j3cKVL]4o3^KQLb4T4YKlKg4Y4TKgKl4f4gJZKY5S5ZJmJf5`5mI`JT6g5dIYJ]6Z6001O1O001O1O001O001O1O0000001O00000000000010O000000000001O0000000000001O0000001O0000001O0000001O00001O001O001O001O001O001O001O00001O000000000000000000000000O100001O00000000001O00000000001O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O101N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O001000O1jGRLn6o3nHZLl6f3PIcLj6^3RIkLi6V3RIRMj6n2RIXMk6i2RI[Mm6f2nH_MP7b2lHcMS7]2iHhMV7Y4M2O1O1N2O1O2M2O1O1N2O2N1N2O1O1N3N1O1N3N1O2M3N1O2M2O2N2M2O2N1N3N2M2O2N1N3M3M2N3M2N3M2N3M2N3M2N3M2O2M2N3M2N3M2N3M2N3M2N3M7I7I5K1O1O1O100O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O100O00O10000000O100O10O10O10000O100O10O10O10000O1000O0100O10000O1000O0100O10000O1000O0100O10000O1000O010000O100O10O10O10000O100O10000O010O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O001O1O1O1N2O1N2N2O1N101N7J6I7JjPh1"}, "label": "an elephant with his trunk touching the mouth of the other elephant"}]} {"id": 470085, "image": "COCO_train2014_000000470085.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant with short tusks and trunk extended\"."}], "task": "refering", "answer_template": "The \"elephant with short tusks and trunk extended\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 184, 185, 186, 187, 188, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 230, 231, 232, 233, 234, 240, 241, 242, 243, 244, 245, 246, 247, 253, 254, 255, 256, 257, 264, 265, 266, 267, 268, 269, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327], "bbox": [0.0015, 0.006604215456674473, 0.788359375, 0.9819437939110071], "iscrowd": 0, "area": 117396.40129999998, "rle": {"size": [427, 640], "counts": "nc0`6;`60000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000kIHWO8i0NQO2n04mNLR1:iNFW1?dNA[1e0`N[O_1k0\\NUOd1P1WNPOh1V1SNjNl1\\1oMdNP2b1kM^NU2h1eMXNZ2n1aMRN^2T2]MlMc2Y2WMhMh2^2SMbMn2b2mL^MS3g2hLYMX3l2cLTM]3Q3^LoLb3V3YLjLg3\\3SLdLm3a3nK_LR4f3iKZLX4j3cKVL]4o3^KQLb4T4YKlKg4Y4TKgKl4f4gJZKY5S5ZJmJf5`5mI`JT6g5dIYJ]6Z6001O1O001O1O001O001O1O0000001O00000000000010O000000000001O0000000000001O0000001O0000001O0000001O00001O001O001O001O001O001O001O00001O000000000000000000000000O100001O00000000001O00000000001O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O101N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O001000O1jGRLn6o3nHZLl6f3PIcLj6^3RIkLi6V3RIRMj6n2RIXMk6i2RI[Mm6f2nH_MP7b2lHcMS7]2iHhMV7Y4M2O1O1N2O1O2M2O1O1N2O2N1N2O1O1N3N1O1N3N1O2M3N1O2M2O2N2M2O2N1N3N2M2O2N1N3M3M2N3M2N3M2N3M2N3M2N3M2O2M2N3M2N3M2N3M2N3M2N3M7I7I5K1O1O1O100O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O100O00O10000000O100O10O10O10000O100O10O10O10000O1000O0100O10000O1000O0100O10000O1000O0100O10000O1000O010000O100O10O10O10000O100O10000O010O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O001O1O1O1N2O1N2N2O1N101N7J6I7JjPh1"}, "label": "elephant with short tusks and trunk extended"}]} {"id": 470085, "image": "COCO_train2014_000000470085.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant with it ' s trunk curled under itself and no tusk showing\"."}], "task": "refering", "answer_template": "The \"an elephant with it ' s trunk curled under itself and no tusk showing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 61, 62, 63, 64, 65, 66, 67, 68, 83, 84, 85, 86, 87, 88, 89, 90, 91, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 340, 341, 342], "bbox": [0.617703125, 0.0022482435597189696, 1.0, 0.9573067915690866], "iscrowd": 0, "area": 75755.72240000001, "rle": {"size": [427, 640], "counts": "a\\U55R=4L4L4K5L5K4L4L4L4L4L4M3M3XFXNk7j1THXNh7k1XHWNd7j1]HWN`7k1`HVNW4OnLn1jNVNU42jLj1QOUNR48eLf1YOSNP4>_La1ASNm3b0ZL^1HRNk3k0RLU13RNh3S1iKm0`0QNd3]1aKd0j0QNc3d1XK>U1PN`3l1PK6`1oM]3n4cLSK[3n4dLUKX3m4hLTKU3n4jLTKS3m4nLUKo2l4QMUKl2m4SMVKi2l4WMUKf2m4YMUKd2m4\\MUK`2m4_MUK_2l4aMUK\\2l4dMWKX2k4hMVKU2l4jMWKR2k4nMVKo1l4PNVKm1l4SNVKj1k4UNWKh1k4XNVKe1k4[NXKb1i4^NXK`1i4`NWK_1j4`NXK^1i4bNWK]1j4bNWK]1j4cNWK[1i4eNXKZ1i4fNWKY1j4fNWKY1j4gNWKW1i4iNXKV1i4jNWKU1j4jNXKT1h4mNXK=lMdNm6n0XKoGBo7`0PHAn7a0RH_Ol7c0TH]Ok7d0UH\\Oj7e0VH[Oj7e0VH[Oi7f0WHZOi7e0XH[Og7f0XH[Oh7e0XH[Og7f0YHZOg7f0YHZOf7g0ZHYOf7f0[HZOd7g0\\HYOc7h0]HXOc7h0\\HYOc7h0]HXOc7h0]HXOb7h0_HXOa7h0_HXO`7i0`HWO`7i0`HWO_7j0aHVO_7j0aHVO^7j0bHWO^7i0bHWO^7i0bHWO]7j0cHVO]7j0cHVO]7j0cHVO\\7j0dHWO\\7i0dHWO\\7i0dHWO[7j0eHVO[7j0eHVOZ7k0fHUOY7k0gHVOY7j0gHVOX7k0hHUOW7l0iHTOW7l0iHTOV7m0iHTOW7l0iHTOV7m0jHSOU7n0kHROU7n0kHROT7o0kHROU7n0kHROT7o0lHQOS7P1mHPOS7P1mHPOR7Q1mHPOS7P1mHPOR7Q1nHoNR7Q1nHoNQ7R1oHnNQ7Q1oHPOP7Q1PIoNP7Q1PIoNo6R1PIoNP7;\\GZOd1;P7:^GZOb1P76eG[O[1?o66hGZOY1`0o64kG[OU1b0o63nGZOS1c0o61QH[Oo0e0o60THZOm0f0o6NWH[Oj0g0n6M[H[Of0i0o6K]H[Od0k0m6IbH[O`0m0n6GfHYO in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 21, 22, 39, 40, 41, 42, 43, 44, 45, 61, 62, 63, 64, 65, 66, 67, 68, 83, 84, 85, 86, 87, 88, 89, 90, 91, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 340, 341, 342], "bbox": [0.617703125, 0.0022482435597189696, 1.0, 0.9573067915690866], "iscrowd": 0, "area": 75755.72240000001, "rle": {"size": [427, 640], "counts": "a\\U55R=4L4L4K5L5K4L4L4L4L4L4M3M3XFXNk7j1THXNh7k1XHWNd7j1]HWN`7k1`HVNW4OnLn1jNVNU42jLj1QOUNR48eLf1YOSNP4>_La1ASNm3b0ZL^1HRNk3k0RLU13RNh3S1iKm0`0QNd3]1aKd0j0QNc3d1XK>U1PN`3l1PK6`1oM]3n4cLSK[3n4dLUKX3m4hLTKU3n4jLTKS3m4nLUKo2l4QMUKl2m4SMVKi2l4WMUKf2m4YMUKd2m4\\MUK`2m4_MUK_2l4aMUK\\2l4dMWKX2k4hMVKU2l4jMWKR2k4nMVKo1l4PNVKm1l4SNVKj1k4UNWKh1k4XNVKe1k4[NXKb1i4^NXK`1i4`NWK_1j4`NXK^1i4bNWK]1j4bNWK]1j4cNWK[1i4eNXKZ1i4fNWKY1j4fNWKY1j4gNWKW1i4iNXKV1i4jNWKU1j4jNXKT1h4mNXK=lMdNm6n0XKoGBo7`0PHAn7a0RH_Ol7c0TH]Ok7d0UH\\Oj7e0VH[Oj7e0VH[Oi7f0WHZOi7e0XH[Og7f0XH[Oh7e0XH[Og7f0YHZOg7f0YHZOf7g0ZHYOf7f0[HZOd7g0\\HYOc7h0]HXOc7h0\\HYOc7h0]HXOc7h0]HXOb7h0_HXOa7h0_HXO`7i0`HWO`7i0`HWO_7j0aHVO_7j0aHVO^7j0bHWO^7i0bHWO^7i0bHWO]7j0cHVO]7j0cHVO]7j0cHVO\\7j0dHWO\\7i0dHWO\\7i0dHWO[7j0eHVO[7j0eHVOZ7k0fHUOY7k0gHVOY7j0gHVOX7k0hHUOW7l0iHTOW7l0iHTOV7m0iHTOW7l0iHTOV7m0jHSOU7n0kHROU7n0kHROT7o0kHROU7n0kHROT7o0lHQOS7P1mHPOS7P1mHPOR7Q1mHPOS7P1mHPOR7Q1nHoNR7Q1nHoNQ7R1oHnNQ7Q1oHPOP7Q1PIoNP7Q1PIoNo6R1PIoNP7;\\GZOd1;P7:^GZOb1P76eG[O[1?o66hGZOY1`0o64kG[OU1b0o63nGZOS1c0o61QH[Oo0e0o60THZOm0f0o6NWH[Oj0g0n6M[H[Of0i0o6K]H[Od0k0m6IbH[O`0m0n6GfHYO in this image.", "objects": [{"patches": [164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 210, 211, 212, 213, 214, 233, 234, 235, 236, 256, 257, 258, 259, 280, 281, 282], "bbox": [0.14575, 0.4634793187347932, 0.324703125, 0.8522384428223844], "iscrowd": 0, "area": 13020.681550000001, "rle": {"size": [411, 640], "counts": "`cU15Z<>C>A>B>B?A=L5N1O1O1O1O1O1O1O1O1O2N1O1O1N2N2N2N2N2N2N3M2N2N2N2N2M3N2O0O2O0O101N1O101N100O101O00001O000O2O0000001O00001N1000001O00001O000O2O0000001O0000000O11N101O1O001O0O2O1O001iMYHXOg7g0WInMX7P2k1M2O2N2N1N3N1N3M3N1N4M4K5L4K4L5L4K5L3L5L4K5K4MRU]5"}, "label": "a man in a white shirt and a black tie looking down"}]} {"id": 363593, "image": "COCO_train2014_000000363593.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a nab wearing a white shirt and tie\"."}], "task": "refering", "answer_template": "The \"a nab wearing a white shirt and tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 210, 211, 212, 213, 214, 233, 234, 235, 236, 256, 257, 258, 259, 280, 281, 282], "bbox": [0.14575, 0.4634793187347932, 0.324703125, 0.8522384428223844], "iscrowd": 0, "area": 13020.681550000001, "rle": {"size": [411, 640], "counts": "`cU15Z<>C>A>B>B?A=L5N1O1O1O1O1O1O1O1O1O2N1O1O1N2N2N2N2N2N2N3M2N2N2N2N2M3N2O0O2O0O101N1O101N100O101O00001O000O2O0000001O00001N1000001O00001O000O2O0000001O0000000O11N101O1O001O0O2O1O001iMYHXOg7g0WInMX7P2k1M2O2N2N1N3N1N3M3N1N4M4K5L4K4L5L4K5L3L5L4K5K4MRU]5"}, "label": "a nab wearing a white shirt and tie"}]} {"id": 363593, "image": "COCO_train2014_000000363593.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white - haired man talking on a phone\"."}], "task": "refering", "answer_template": "The \"a white - haired man talking on a phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 283, 284, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 306, 307, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.32328125, 0.24489051094890513, 0.9974999999999999, 0.9841362530413625], "iscrowd": 0, "area": 78001.83320000001, "rle": {"size": [411, 640], "counts": "W]c29WkJDU5Q6BnI?R6AmI`0S6@lIa0T6_OkIb0U6^OiId0W6\\OhIe0X6[OgIf0Y6ZOfIg0Z6YOeIh0[6XOcIj0]6VObIk0^6UOaIl0_6TO`Im0`6SO_In0a6RO^Io0c6PO[IR1e6oNYIR1g6nNXIS1h6mNWIT1i6lNVIU1j6kNTIW1l6iNSIX1m6hNRIY1n6gNQIZ1o6fNPI[1P7eNoH\\1Q7dNoH\\1Q7dNnH]1R7cNmH^1S7bNlH_1T7aNlH_1T7aNkH`1U7`NjHa1V7_NiHb1W7^NiHb1W7^NhHc1X7]NgHd1Y7]NfHc1Z7]NeHd1[7\\NdHe1\\7[NcHf1]7ZNcHf1]7ZNbHg1^7YNaHh1_7XN`Hi1`7WN`Hi1`7WN_Hj1a7VN^Hk1b7UN^Hk1b7UN]Hl1c7TN\\Hm1d7SN[Hn1e7RN[Hn1e7RNZHo1f7QNYHP2g7PNXHQ2h7oMXHQ2h7oMWHR2i7nMVHS2j7nMUHR2k7nMTHS2l7mMSHT2m7lMRHU2n7kMRHU2n7\\100000000000000001O00000000000000000000000000000000000000N2N2N2M3N2N2M3N2N2M3N2N2N2N200O1O1BWJiIj5V6YJgIh5X6[JfIe5Z6]JcId5\\6>O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1N2J6J6K5L4M3L4M3L4M3L400O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O10000O10000000000O10000000000O10000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O10000000000O10000000000O1000000000000001O00001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O001O00001Ol0TOPe0"}, "label": "a white - haired man talking on a phone"}]} {"id": 363593, "image": "COCO_train2014_000000363593.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white haired man talking on a cellphone\"."}], "task": "refering", "answer_template": "The \"a white haired man talking on a cellphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 260, 261, 262, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 283, 284, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 306, 307, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.32328125, 0.24489051094890513, 0.9974999999999999, 0.9841362530413625], "iscrowd": 0, "area": 78001.83320000001, "rle": {"size": [411, 640], "counts": "W]c29WkJDU5Q6BnI?R6AmI`0S6@lIa0T6_OkIb0U6^OiId0W6\\OhIe0X6[OgIf0Y6ZOfIg0Z6YOeIh0[6XOcIj0]6VObIk0^6UOaIl0_6TO`Im0`6SO_In0a6RO^Io0c6PO[IR1e6oNYIR1g6nNXIS1h6mNWIT1i6lNVIU1j6kNTIW1l6iNSIX1m6hNRIY1n6gNQIZ1o6fNPI[1P7eNoH\\1Q7dNoH\\1Q7dNnH]1R7cNmH^1S7bNlH_1T7aNlH_1T7aNkH`1U7`NjHa1V7_NiHb1W7^NiHb1W7^NhHc1X7]NgHd1Y7]NfHc1Z7]NeHd1[7\\NdHe1\\7[NcHf1]7ZNcHf1]7ZNbHg1^7YNaHh1_7XN`Hi1`7WN`Hi1`7WN_Hj1a7VN^Hk1b7UN^Hk1b7UN]Hl1c7TN\\Hm1d7SN[Hn1e7RN[Hn1e7RNZHo1f7QNYHP2g7PNXHQ2h7oMXHQ2h7oMWHR2i7nMVHS2j7nMUHR2k7nMTHS2l7mMSHT2m7lMRHU2n7kMRHU2n7\\100000000000000001O00000000000000000000000000000000000000N2N2N2M3N2N2M3N2N2M3N2N2N2N200O1O1BWJiIj5V6YJgIh5X6[JfIe5Z6]JcId5\\6>O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1N2J6J6K5L4M3L4M3L4M3L400O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O10000O10000000000O10000000000O10000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O10000000000O10000000000O1000000000000001O00001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O00001O001O00001O00001O00001O00001O001O00001Ol0TOPe0"}, "label": "a white haired man talking on a cellphone"}]} {"id": 166985, "image": "COCO_train2014_000000166985.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing next to a girl on a surfboard\"."}], "task": "refering", "answer_template": "The \"a man standing next to a girl on a surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 139, 140, 153, 154, 155, 156, 168, 169, 170, 171, 183, 184, 185, 186, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 227, 228, 229, 230, 231, 243, 244, 245, 246], "bbox": [0.15868852459016394, 0.367, 0.4666978922716628, 0.7235625], "iscrowd": 0, "area": 16859.0862, "rle": {"size": [640, 427], "counts": "WmZ13jc05K4L4L5K4L5L3K5K4K5L4K5L4K4M1NL6H90O100O2K4K5L4L3L5L3L5N12O2M2N3N1N2O2Mb_OVOk=i0RB\\On=?SBEl=7TBNl=LUB9j=CWBa0i=ZOWBk0h=AiAc0W>c2O100O1POQKoBo4a in this image.", "objects": [{"patches": [124, 125, 139, 140, 153, 154, 155, 156, 168, 169, 170, 171, 183, 184, 185, 186, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 227, 228, 229, 230, 231, 243, 244, 245, 246], "bbox": [0.15868852459016394, 0.367, 0.4666978922716628, 0.7235625], "iscrowd": 0, "area": 16859.0862, "rle": {"size": [640, 427], "counts": "WmZ13jc05K4L4L5K4L5L3K5K4K5L4K5L4K4M1NL6H90O100O2K4K5L4L3L5L3L5N12O2M2N3N1N2O2Mb_OVOk=i0RB\\On=?SBEl=7TBNl=LUB9j=CWBa0i=ZOWBk0h=AiAc0W>c2O100O1POQKoBo4a in this image.", "objects": [{"patches": [157, 158, 159, 172, 173, 174, 187, 188, 189, 202, 203, 204, 205, 217, 218, 219, 220, 232, 233, 234, 235, 245, 246, 247, 248, 249, 250, 251, 260, 261, 262, 264, 265, 266], "bbox": [0.3705854800936768, 0.454296875, 0.782903981264637, 0.76896875], "iscrowd": 0, "area": 17213.19775, "rle": {"size": [640, 427], "counts": "aVS34lc0=C1N10000O101O0O10000O10000O10000O10000O10000O101N10O10O10000O001N2N2N2M2O2N2N2N2M2N3DcLhA^3T>kLfAV3W>QMdAP3Y>XMbAg2[>aM`A`2]>c1O1O1O001O1O1O1O00100O1O001O1O1000O100000000O100000000O10001O000O2O001O000O2O001O001O001O1O2N1O1O2M20010O100O2O0O101N109F5L3L4M0O000O001O001O001O3M5K5K5K5K5K4L5K5K5PNl^Oj0Ya0mNl^OR1Ya0dNm^O[1ma0O2M2O1O2N1O1O1O2N1N2O2N1O1O1O2N1N2O2N1O1O2M2O1O2N1N2O2N1O1O2M2O1Oldi1"}, "label": "girl sitting on board"}]} {"id": 166985, "image": "COCO_train2014_000000166985.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman sitting on surf board in water\"."}], "task": "refering", "answer_template": "The \"woman sitting on surf board in water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 172, 173, 174, 187, 188, 189, 202, 203, 204, 205, 217, 218, 219, 220, 232, 233, 234, 235, 245, 246, 247, 248, 249, 250, 251, 260, 261, 262, 264, 265, 266], "bbox": [0.3705854800936768, 0.454296875, 0.782903981264637, 0.76896875], "iscrowd": 0, "area": 17213.19775, "rle": {"size": [640, 427], "counts": "aVS34lc0=C1N10000O101O0O10000O10000O10000O10000O10000O101N10O10O10000O001N2N2N2M2O2N2N2N2M2N3DcLhA^3T>kLfAV3W>QMdAP3Y>XMbAg2[>aM`A`2]>c1O1O1O001O1O1O1O00100O1O001O1O1000O100000000O100000000O10001O000O2O001O000O2O001O001O001O1O2N1O1O2M20010O100O2O0O101N109F5L3L4M0O000O001O001O001O3M5K5K5K5K5K4L5K5K5PNl^Oj0Ya0mNl^OR1Ya0dNm^O[1ma0O2M2O1O2N1O1O1O2N1N2O2N1O1O1O2N1N2O2N1O1O2M2O1O2N1N2O2N1O1O2M2O1Oldi1"}, "label": "woman sitting on surf board in water"}]} {"id": 257102, "image": "COCO_train2014_000000257102.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair in kitchen near counter top\"."}], "task": "refering", "answer_template": "The \"chair in kitchen near counter top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 219, 220, 221, 222, 223, 242, 243, 244, 245, 246, 267, 268, 269, 270, 288, 290, 291, 292, 293, 311, 314, 315, 316, 337], "bbox": [0.525203125, 0.4949882903981265, 0.74684375, 0.9600468384074943], "iscrowd": 0, "area": 6139.221050000001, "rle": {"size": [427, 640], "counts": "U_\\42R=7I8H7iM[OaGm0]8UO]GP1a8XOYGh0f8EmF=R9DmF;T9ElF5Z9JgF0_90aFIf98YFBm9>RF\\OT:f0kEUOZ:l0R1O1L4M31O00001N101O00001O00001O001O000O2O001O00001O001O00001O001TNnN]GR1_8ROaGn0_8ROaGo0_8POaGP1_8QO_GP1a8PO_GQ1`8oN`GQ1`8oN`GR1_8nNaGR1_8nNaGi1h7WNSHX2c7hM^HX2a7hM_HY2`7gM`HZ2_7fMaH[2^7eMcH[2]7eMbHQ2h7oMXHf1S8ZNnGZ1]8fNcGW1`8iN`GW1`8iN`GW1`8iNaGW1^8iNbGW1^8iNbGW1^8iNbGW1^8iNcGW1\\8iNdGW1\\8jNcGV1]8jNcGW1]8hNdGW1\\8iNdGW1\\8iNdGW1\\8iNeGW1Z8iNfGP4a5PL_JP4a5PL_JP4a5PLbJn3]5RLjJg3V5YLQK`3o4aLWKX3i4hL^KQ3b4oLeKe2`4[MoKf1b4YNX500000000000000001O000001O00000000eD0[90dF8U9HkF?n8AQGf0i8[OUGg0j8YO[Gb0e8^O\\G`0e8@[G?f8AZG?f8A[G=g8BYG=h8CYGo8BQG?n8ARGb0k8^OUGU1X8kNhGi1e7VN[H]2R7cMnHo2`6QM`In2a6RM_Im2b6SMPJ[2P6fMfJb1[5^NTKR1P5kN`Ke0S9_OWVS2"}, "label": "chair in kitchen near counter top"}]} {"id": 257102, "image": "COCO_train2014_000000257102.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair pushed up to table with nothing in front of it\"."}], "task": "refering", "answer_template": "The \"chair pushed up to table with nothing in front of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 167, 188, 189, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 260, 280, 281, 304], "bbox": [0.181203125, 0.49543325526932086, 0.337765625, 0.8971662763466043], "iscrowd": 0, "area": 5794.712099999999, "rle": {"size": [427, 640], "counts": "ff`16i<=D;E;G9@`0C7NO2M3M2N00F:91O10000005K1O00001O01O000000001O000000001O001O001_GnMg5R2RIVOm6k0oG8Q8a200O01M3E;XOh0A?[Oe0G9UNRF8o9\\O\\Fd0e9\\O[Fd0e9[O[Ff0e9ZO[Ff0e9ZOZFg0f9YOZFf0g9ZOYFf0g9ZOXFg0h9XOYFh0h9WOWFj0U;0O010O10000O01000000000O01000000002Nk0VO in this image.", "objects": [{"patches": [165, 166, 167, 188, 189, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 260, 280, 281, 304], "bbox": [0.181203125, 0.49543325526932086, 0.337765625, 0.8971662763466043], "iscrowd": 0, "area": 5794.712099999999, "rle": {"size": [427, 640], "counts": "ff`16i<=D;E;G9@`0C7NO2M3M2N00F:91O10000005K1O00001O01O000000001O000000001O001O001_GnMg5R2RIVOm6k0oG8Q8a200O01M3E;XOh0A?[Oe0G9UNRF8o9\\O\\Fd0e9\\O[Fd0e9[O[Ff0e9ZO[Ff0e9ZOZFg0f9YOZFf0g9ZOYFf0g9ZOXFg0h9XOYFh0h9WOWFj0U;0O010O10000O01000000000O01000000002Nk0VO in this image.", "objects": [{"patches": [191, 192, 193, 194, 214, 215, 216, 217, 237, 238, 239, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 266, 283, 284, 285, 286, 287, 288, 289, 306, 308, 309, 312, 329, 331, 332, 335], "bbox": [0.31396875, 0.5282903981264637, 0.5949375, 1.0], "iscrowd": 0, "area": 14406.552949999996, "rle": {"size": [427, 640], "counts": "UWd26k<:E in this image.", "objects": [{"patches": [191, 192, 193, 194, 214, 215, 216, 217, 237, 238, 239, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 266, 283, 284, 285, 286, 287, 288, 289, 306, 308, 309, 312, 329, 331, 332, 335], "bbox": [0.31396875, 0.5282903981264637, 0.5949375, 1.0], "iscrowd": 0, "area": 14406.552949999996, "rle": {"size": [427, 640], "counts": "UWd26k<:E in this image.", "objects": [{"patches": [230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 306], "bbox": [0.0, 0.7097882352941177, 0.364953125, 0.8916705882352942], "iscrowd": 0, "area": 13702.09905, "rle": {"size": [425, 640], "counts": "`9l1];O100000000000000000000O1000000000000003M;E3M00000000000000000000000000A?O1000000O10000000000000000000000000000000000000000O100000000000000000O100000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000`0@0000000000000000000000O10000QNjEk0Z;@`0000000O100N2_OUVX5"}, "label": "a stone bench that doesnt have an animal in front of it"}]} {"id": 109654, "image": "COCO_train2014_000000109654.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a low bench on the left side of an entrance gate\"."}], "task": "refering", "answer_template": "The \"a low bench on the left side of an entrance gate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 306], "bbox": [0.0, 0.7097882352941177, 0.364953125, 0.8916705882352942], "iscrowd": 0, "area": 13702.09905, "rle": {"size": [425, 640], "counts": "`9l1];O100000000000000000000O1000000000000003M;E3M00000000000000000000000000A?O1000000O10000000000000000000000000000000000000000O100000000000000000O100000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000`0@0000000000000000000000O10000QNjEk0Z;@`0000000O100N2_OUVX5"}, "label": "a low bench on the left side of an entrance gate"}]} {"id": 216150, "image": "COCO_train2014_000000216150.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bike with a laid shirt on its seat resting on a pole\"."}], "task": "refering", "answer_template": "The \"a bike with a laid shirt on its seat resting on a pole\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [144, 145, 146, 162, 163, 164, 165, 166, 180, 181, 182, 183, 184, 185, 200, 201, 202, 203, 218, 219, 220, 221, 236, 237, 238], "bbox": [0.0, 0.5334889434889435, 0.30512, 0.9032186732186732], "iscrowd": 0, "area": 9521.755450000004, "rle": {"size": [407, 500], "counts": "Y8a0k:\\1J_NnEh0R:WOoE`0]O]Oc:4oE2^9MjE0:0`01\\9OkEO91a00[90lEN91a00[91jEO:OT;0cD090kf:2UeE1O1]OMbD3?Ml91eE2<1n9NeE1;3P:MdE097R:JdEO7;T:GdEN5?V:DdEM4a0X:CbEM3d0Z:@bEL1h0\\:]ObEKOl0]:[OcEIMP1_:XObEIMQ1a:WO`ET16_Nn9=kEh1m9g0J2N3M2N3M2NaMeFe1X9WNlFk1R9QNRGR2j8kMZGW2d8gM^G[2`8dMaG]2^8cMbG^2]8aMdG_2\\8aMdG`2[8_MfGb2Y8^MgGc2X8]MhGd2W8[MjGf2V8YMiGh2X8VMiGk2V8UMjGl2V8SMjGn2o8001O00001O001O00001O001O00001O001O000000000000O1000000000000000000000000000000nLdFi2\\9UMiFh2W9VMmFh2S9VMPGi2`9N2O1N2O0O2O1N2N2N2N2N2N1N3N2N2N2N2M5L4L3M4L4K5D=D in this image.", "objects": [{"patches": [81, 98, 99, 100, 101, 116, 117, 118, 119, 135, 136, 137, 152, 153, 154, 155, 170, 171, 172, 173, 188, 189, 190, 205, 206, 207, 208, 223, 224, 225, 226, 241, 242, 243, 244, 261, 262], "bbox": [0.42012, 0.3247420147420147, 0.65996, 0.984938574938575], "iscrowd": 0, "area": 17726.869400000003, "rle": {"size": [407, 500], "counts": "hhc25b<9F:G9G9G9F;F4L00O01000O01000OL5J6J5QLaMaMf2]2\\M\\Mk2c2VMUMQ3j2QMoLV3P3kLiL\\3U3fLcLa3]3`L\\Lg3b3\\LVLk3i3VLVLj3i3XLULj3j3WLTLk3j3WLULi3k3XLSLj3k3XLTLi3k3YLRLh3m3ZLQLh3n3YLQLh3m3ZLQLg3o3ZLPLg3n3[LPLg3o3ZLPLf3o3]LnKe3Q4\\LmKf3S4ZLlKf3U4ZLiKh3W4XLhKi3Y4VLeKk3\\4ULbKm3^4SLaKT4Y4WLZKQ4`4X2K6J6J5K1O0000O101O000000000O1000000000000O2O000000001O`KlG^4V1cKd5_4\\JfK]5]4bJhKX5Y4hJiK_22_NX4ROfK\\2g0mMf3FdKZ2j0lMd3JcKW2m0kMc3MaKT2Q1kM`31`KQ2T1kM^33_Ko1W1jM]36]Km1Z1iM[3:[Kk1^1gMZ3=YKh1b1gMW3b0WKd1f1fMV3e0UKb1i1fMT3g0TK`1l1eMR3k0SK]1o1dMQ3n0QK[1j7eNWHW1m7hNTHU1o7jNQHT1Q8mNoGP1n9M3O1N2N2O0O2N1O2O0O2N101N1O2O0O2N3Mc0^O6I7I]ZS2"}, "label": "a woman in a purple dress"}]} {"id": 216150, "image": "COCO_train2014_000000216150.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a long purple sarong type dress\"."}], "task": "refering", "answer_template": "The \"a woman in a long purple sarong type dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 98, 99, 100, 101, 116, 117, 118, 119, 135, 136, 137, 152, 153, 154, 155, 170, 171, 172, 173, 188, 189, 190, 205, 206, 207, 208, 223, 224, 225, 226, 241, 242, 243, 244, 261, 262], "bbox": [0.42012, 0.3247420147420147, 0.65996, 0.984938574938575], "iscrowd": 0, "area": 17726.869400000003, "rle": {"size": [407, 500], "counts": "hhc25b<9F:G9G9G9F;F4L00O01000O01000OL5J6J5QLaMaMf2]2\\M\\Mk2c2VMUMQ3j2QMoLV3P3kLiL\\3U3fLcLa3]3`L\\Lg3b3\\LVLk3i3VLVLj3i3XLULj3j3WLTLk3j3WLULi3k3XLSLj3k3XLTLi3k3YLRLh3m3ZLQLh3n3YLQLh3m3ZLQLg3o3ZLPLg3n3[LPLg3o3ZLPLf3o3]LnKe3Q4\\LmKf3S4ZLlKf3U4ZLiKh3W4XLhKi3Y4VLeKk3\\4ULbKm3^4SLaKT4Y4WLZKQ4`4X2K6J6J5K1O0000O101O000000000O1000000000000O2O000000001O`KlG^4V1cKd5_4\\JfK]5]4bJhKX5Y4hJiK_22_NX4ROfK\\2g0mMf3FdKZ2j0lMd3JcKW2m0kMc3MaKT2Q1kM`31`KQ2T1kM^33_Ko1W1jM]36]Km1Z1iM[3:[Kk1^1gMZ3=YKh1b1gMW3b0WKd1f1fMV3e0UKb1i1fMT3g0TK`1l1eMR3k0SK]1o1dMQ3n0QK[1j7eNWHW1m7hNTHU1o7jNQHT1Q8mNoGP1n9M3O1N2N2O0O2N1O2O0O2N101N1O2O0O2N3Mc0^O6I7I]ZS2"}, "label": "a woman in a long purple sarong type dress"}]} {"id": 429143, "image": "COCO_train2014_000000429143.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sailboat with a yellow base\"."}], "task": "refering", "answer_template": "The \"a sailboat with a yellow base\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 18, 19, 20, 31, 32, 33, 34, 44, 45, 46, 47, 56, 57, 58, 59, 60, 69, 70, 71, 72, 73, 74, 82, 83, 84, 85, 86, 87, 95, 96, 97, 98, 99, 100, 108, 109, 110, 111, 112, 113, 114, 120, 121, 122, 123, 124, 125, 126, 127, 133, 134, 135, 136, 137, 138, 139, 140, 146, 147, 148, 149, 150, 151, 152, 153, 154, 159, 160, 161, 162, 163, 164, 165, 166, 167, 172, 173, 174, 175, 176, 177, 178, 179], "bbox": [0.25021333333333334, 0.0, 0.9283999999999999, 0.7432000000000001], "iscrowd": 0, "area": 54509.8149, "rle": {"size": [500, 375], "counts": "UX^13\\?5K5L4K5K5K5J6K5K5K5K5K5J6K5K5K5K5J6K5K5K6J5K5J6K5K5K5K5K5J6K5K5K5K5K5J6K6J5K5K5K5J6K5K5K5K5K5J6K5K5K5K5K6I6K5K5L4K5L4K5L4L4K5L4K5L4L4K5L4K6K4L4K5N11000000000000000001O0000000000000000000000000000000000001O000O10000000000000000000000000003M3M2N3M3M3M3M2N3M3M3M3M3M2N3M3M3L4M2N3M3M3M3M2N3M3M3M3M2N3M3M3M3M2M4M3M3M3M2N3M3M3M3M3L3N3M3M3M3M2N3M3M3M3M2M4M3M3M3M2N3M3M3M3M2M4M3M3M3M3M2N3M3M3M3M2M4M3M3M3M2N3M3M3M3M2M4M3M3M3M2N3M3M3M3M2N3L4M3M2N3M3M3M3M2M4M3M3L4M2N3L4M3M2M4M3M3L2O1O1N101O1N101O0O2O1O0O2O1O0O2O1KZk<"}, "label": "a sailboat with a yellow base"}]} {"id": 429143, "image": "COCO_train2014_000000429143.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow remote control boat on a lake\"."}], "task": "refering", "answer_template": "The \"a yellow remote control boat on a lake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 18, 19, 20, 31, 32, 33, 34, 44, 45, 46, 47, 56, 57, 58, 59, 60, 69, 70, 71, 72, 73, 74, 82, 83, 84, 85, 86, 87, 95, 96, 97, 98, 99, 100, 108, 109, 110, 111, 112, 113, 114, 120, 121, 122, 123, 124, 125, 126, 127, 133, 134, 135, 136, 137, 138, 139, 140, 146, 147, 148, 149, 150, 151, 152, 153, 154, 159, 160, 161, 162, 163, 164, 165, 166, 167, 172, 173, 174, 175, 176, 177, 178, 179], "bbox": [0.25021333333333334, 0.0, 0.9283999999999999, 0.7432000000000001], "iscrowd": 0, "area": 54509.8149, "rle": {"size": [500, 375], "counts": "UX^13\\?5K5L4K5K5K5J6K5K5K5K5K5J6K5K5K5K5J6K5K5K6J5K5J6K5K5K5K5K5J6K5K5K5K5K5J6K6J5K5K5K5J6K5K5K5K5K5J6K5K5K5K5K6I6K5K5L4K5L4K5L4L4K5L4K5L4L4K5L4K6K4L4K5N11000000000000000001O0000000000000000000000000000000000001O000O10000000000000000000000000003M3M2N3M3M3M3M2N3M3M3M3M3M2N3M3M3L4M2N3M3M3M3M2N3M3M3M3M2N3M3M3M3M2M4M3M3M3M2N3M3M3M3M3L3N3M3M3M3M2N3M3M3M3M2M4M3M3M3M2N3M3M3M3M2M4M3M3M3M3M2N3M3M3M3M2M4M3M3M3M2N3M3M3M3M2M4M3M3M3M2N3M3M3M3M2N3L4M3M2N3M3M3M3M2M4M3M3L4M2N3L4M3M2M4M3M3L2O1O1N101O1N101O0O2O1O0O2O1O0O2O1KZk<"}, "label": "a yellow remote control boat on a lake"}]} {"id": 543838, "image": "COCO_train2014_000000543838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the vase from the right\"."}], "task": "refering", "answer_template": "The \"the vase from the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 154, 155, 156, 157, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 291, 292, 293, 294, 295, 296, 297, 315, 316, 317, 318, 319, 320, 340, 341], "bbox": [0.65353125, 0.3029497907949791, 0.95884375, 0.8323430962343097], "iscrowd": 0, "area": 33473.2836, "rle": {"size": [478, 640], "counts": "c\\S66V>e0I7J6J7I6J6J6J6K5J6J7J5K5L3M3M3L4M2N3M3L4M3M3M3M2M4M3M3M3L4M3M2N3L4M3M3M3M2M3N2N2N2N2N1O2N2N2M3N2N2N2N2N2N2N1O2N2N2N2N1O1N3L3M3O110O01O0010O01O010O4L3M3M1O100O0000N2O1N2O1O1N3N1N2O1O1N2O1F:I73N3L4L6J0000O010000000O100001O4L3M2N1N2O1O1O2N1O1O1O1O1O1O1N2O1O1O2N1O1O1O1O1O1O1O1N2O1O2N1O1O1O1O1N2O1N3N1O1N3N1N2M3N3M2N2N3M2N2N3M2N2N2N3L3N2N3M3M3M3M2N3M3L3M4L3M4L3M4L3M4L3M4L3M4L3M6J6H8H9F9E;B>B^k;"}, "label": "the vase from the right"}]} {"id": 543838, "image": "COCO_train2014_000000543838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"teardrop shaped vase on the right\"."}], "task": "refering", "answer_template": "The \"teardrop shaped vase on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 154, 155, 156, 157, 176, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 291, 292, 293, 294, 295, 296, 297, 315, 316, 317, 318, 319, 320, 340, 341], "bbox": [0.65353125, 0.3029497907949791, 0.95884375, 0.8323430962343097], "iscrowd": 0, "area": 33473.2836, "rle": {"size": [478, 640], "counts": "c\\S66V>e0I7J6J7I6J6J6J6K5J6J7J5K5L3M3M3L4M2N3M3L4M3M3M3M2M4M3M3M3L4M3M2N3L4M3M3M3M2M3N2N2N2N2N1O2N2N2M3N2N2N2N2N2N2N1O2N2N2N2N1O1N3L3M3O110O01O0010O01O010O4L3M3M1O100O0000N2O1N2O1O1N3N1N2O1O1N2O1F:I73N3L4L6J0000O010000000O100001O4L3M2N1N2O1O1O2N1O1O1O1O1O1O1N2O1O1O2N1O1O1O1O1O1O1O1N2O1O2N1O1O1O1O1N2O1N3N1O1N3N1N2M3N3M2N2N3M2N2N3M2N2N2N3L3N2N3M3M3M3M2N3M3L3M4L3M4L3M4L3M4L3M4L3M4L3M6J6H8H9F9E;B>B^k;"}, "label": "teardrop shaped vase on the right"}]} {"id": 543838, "image": "COCO_train2014_000000543838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a flower vase between two others\"."}], "task": "refering", "answer_template": "The \"a flower vase between two others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 124, 125, 126, 127, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 237, 238, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 312], "bbox": [0.334921875, 0.28251046025104604, 0.611234375, 0.8161506276150627], "iscrowd": 0, "area": 34992.85850000001, "rle": {"size": [478, 640], "counts": "iUT31g>9G9F9H9G9G9G9G9F:G9I6M4K5K5L4K5L4K3M4M3L4L3N3L4L4M2M4M3L3M4M3L4L3L5L4K5L3M4K5L4K4M4K5L3L3N2N21O1O2N2N1O1O00O10000O1000000O100008H1O1OoNXImIg6R6[ImIe6S6\\ImIc6R6_InI`6R6aInI^6Q6dImI]6S6cIlI^6S6dIlI\\6T6eIjI\\6V6eIiI[6W6fIgI[6Y6fIfIZ6Z6gIdIZ6\\6gIcIY6]6jI_IW6a6mI[IS6e6PJWIQ6i6SJRIo5n6P1000000001O00000000001O000000001O00000000001O00000000000000000000000000000000000000000000001O001O1O1O1O001O1O1O1O2M2dIbHT5b7dJdHZ5`7`JdH^5_7`JbH^5b7^J`H`5`8N2N2N2N2N2N6J2N01O0000O1N2O2M6K5J6K5J8I8G8H8I7H8I8Gg0ZOV1iNRoc3"}, "label": "a flower vase between two others"}]} {"id": 543838, "image": "COCO_train2014_000000543838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"flower bottle in middle like a mango shape\"."}], "task": "refering", "answer_template": "The \"flower bottle in middle like a mango shape\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 124, 125, 126, 127, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 237, 238, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 266, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310, 311, 312], "bbox": [0.334921875, 0.28251046025104604, 0.611234375, 0.8161506276150627], "iscrowd": 0, "area": 34992.85850000001, "rle": {"size": [478, 640], "counts": "iUT31g>9G9F9H9G9G9G9G9F:G9I6M4K5K5L4K5L4K3M4M3L4L3N3L4L4M2M4M3L3M4M3L4L3L5L4K5L3M4K5L4K4M4K5L3L3N2N21O1O2N2N1O1O00O10000O1000000O100008H1O1OoNXImIg6R6[ImIe6S6\\ImIc6R6_InI`6R6aInI^6Q6dImI]6S6cIlI^6S6dIlI\\6T6eIjI\\6V6eIiI[6W6fIgI[6Y6fIfIZ6Z6gIdIZ6\\6gIcIY6]6jI_IW6a6mI[IS6e6PJWIQ6i6SJRIo5n6P1000000001O00000000001O000000001O00000000001O00000000000000000000000000000000000000000000001O001O1O1O1O001O1O1O1O2M2dIbHT5b7dJdHZ5`7`JdH^5_7`JbH^5b7^J`H`5`8N2N2N2N2N2N6J2N01O0000O1N2O2M6K5J6K5J8I8G8H8I7H8I8Gg0ZOV1iNRoc3"}, "label": "flower bottle in middle like a mango shape"}]} {"id": 543838, "image": "COCO_train2014_000000543838.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a thin flower vase place beside other fat flower vase\"."}], "task": "refering", "answer_template": "The \"a thin flower vase place beside other fat flower vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 117, 118, 119, 139, 140, 141, 142, 162, 163, 164, 165, 185, 186, 187, 188, 189, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 277, 278, 279, 280, 281, 282, 300, 301, 302, 303, 304, 326], "bbox": [0.041953125, 0.2764016736401674, 0.266859375, 0.8292050209205021], "iscrowd": 0, "area": 26520.44595, "rle": {"size": [478, 640], "counts": "bl in this image.", "objects": [{"patches": [207, 208, 209, 224, 225, 226, 227, 241, 242, 243, 244, 259, 260, 261, 262, 277, 278, 279, 280, 295, 296], "bbox": [0.21133333333333332, 0.548546875, 0.4858333333333333, 0.7619374999999999], "iscrowd": 0, "area": 9157.916449999999, "rle": {"size": [640, 480], "counts": "PTP29`c09G9G8I8I7N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O1N101O1O1O1O1O001O1O1O1O1O001O1O1O100000O10O10000000O0100000000O01000000000O010000000O1000O1000O100000O101O00001O0O101O001O00001N10001O001O0O101N1O1N3N1N3N1N2O2M2O2M2O1N3N1N2O2M2O2M2O1N3N1N2O2M2O2M2O1N3N1N2O2M2O2M2O1N3Nemi4"}, "label": "keyboard for xbox controller"}]} {"id": 142431, "image": "COCO_train2014_000000142431.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small keyboard peripheral for an xbox 360 controller\"."}], "task": "refering", "answer_template": "The \"a small keyboard peripheral for an xbox 360 controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 224, 225, 226, 227, 241, 242, 243, 244, 259, 260, 261, 262, 277, 278, 279, 280, 295, 296], "bbox": [0.21133333333333332, 0.548546875, 0.4858333333333333, 0.7619374999999999], "iscrowd": 0, "area": 9157.916449999999, "rle": {"size": [640, 480], "counts": "PTP29`c09G9G8I8I7N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O1N101O1O1O1O1O001O1O1O1O1O001O1O1O100000O10O10000000O0100000000O01000000000O010000000O1000O1000O100000O101O00001O0O101O001O00001N10001O001O0O101N1O1N3N1N3N1N2O2M2O2M2O1N3N1N2O2M2O2M2O1N3N1N2O2M2O2M2O1N3N1N2O2M2O2M2O1N3Nemi4"}, "label": "a small keyboard peripheral for an xbox 360 controller"}]} {"id": 142431, "image": "COCO_train2014_000000142431.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cracked clay in a gentleman ' s left hand\"."}], "task": "refering", "answer_template": "The \"cracked clay in a gentleman ' s left hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [232, 233, 234, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 285, 286, 294, 295, 296, 297, 298, 299, 300, 301, 302, 311, 312, 313, 314, 315, 329, 330], "bbox": [0.33820833333333333, 0.5862499999999999, 0.8688333333333332, 0.8519375], "iscrowd": 0, "area": 18765.855799999998, "rle": {"size": [640, 480], "counts": "a[V3:ac08H8J7L3M4L3M3M3M2N3M2N2N001O0O100O100O100O100O101N100O100O100O100O10cNa^O9^a0Hn^OLQa04Q_OLn`04R_OLm`04U_OKk`05V_OJi`06Y_OIf`08[_OGe`08]_OGb`0:__OE``0a_OA^`0`0c_O^O^`0c0b_O\\O]`0f0c_OXO]`0j0b_OVO^`0k0b_OSO^`0n0c_OQO]`0P1c_OnN]`0T1c_OkN\\`0W1d_OhN\\`0X1U12N2O0O2O0O2N2O0O2N101NO2O00001O001O00010O000010O010O100O100O2O0O100O100O2O0O100O101N100O2O0O101N100O2O0O2N1O100O1O1O1O001O1O1O1O1O1O1O1O1O00100O1O1O1O1O1O1O1O001O1O1O001O001O01O01O0010O0010O00100O010O10O01O10O0100O00100O10O0100O00100O010O10O01O10O1000O01000000O10000O10000O10000O1000000000000O10O10000000000001O1O1O1O1O001O1O1O1O1O1O001O1N2O1N3M3M3N2M3M3M3M3N2M3M2N3N2M3M3M4L3N2M4L3M4M2M3M4L;FmmV1"}, "label": "cracked clay in a gentleman ' s left hand"}]} {"id": 142431, "image": "COCO_train2014_000000142431.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"it appears to be a wiimote , with a boy ' s fingers holding it up . additionally , half of an xbox 360 controller\"."}], "task": "refering", "answer_template": "The \"it appears to be a wiimote , with a boy ' s fingers holding it up . additionally , half of an xbox 360 controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [232, 233, 234, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 279, 280, 281, 282, 283, 284, 285, 286, 294, 295, 296, 297, 298, 299, 300, 301, 302, 311, 312, 313, 314, 315, 329, 330], "bbox": [0.33820833333333333, 0.5862499999999999, 0.8688333333333332, 0.8519375], "iscrowd": 0, "area": 18765.855799999998, "rle": {"size": [640, 480], "counts": "a[V3:ac08H8J7L3M4L3M3M3M2N3M2N2N001O0O100O100O100O100O101N100O100O100O100O10cNa^O9^a0Hn^OLQa04Q_OLn`04R_OLm`04U_OKk`05V_OJi`06Y_OIf`08[_OGe`08]_OGb`0:__OE``0a_OA^`0`0c_O^O^`0c0b_O\\O]`0f0c_OXO]`0j0b_OVO^`0k0b_OSO^`0n0c_OQO]`0P1c_OnN]`0T1c_OkN\\`0W1d_OhN\\`0X1U12N2O0O2O0O2N2O0O2N101NO2O00001O001O00010O000010O010O100O100O2O0O100O100O2O0O100O101N100O2O0O101N100O2O0O2N1O100O1O1O1O001O1O1O1O1O1O1O1O1O00100O1O1O1O1O1O1O1O001O1O1O001O001O01O01O0010O0010O00100O010O10O01O10O0100O00100O10O0100O00100O010O10O01O10O1000O01000000O10000O10000O10000O1000000000000O10O10000000000001O1O1O1O1O001O1O1O1O1O1O001O1N2O1N3M3M3N2M3M3M3M3N2M3M2N3N2M3M3M4L3N2M4L3M4M2M3M4L;FmmV1"}, "label": "it appears to be a wiimote , with a boy ' s fingers holding it up . additionally , half of an xbox 360 controller"}]} {"id": 142431, "image": "COCO_train2014_000000142431.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a playstation controller\"."}], "task": "refering", "answer_template": "The \"a playstation controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 139, 140, 141, 142, 143, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 188, 189, 190, 205, 206, 207, 222, 223, 224], "bbox": [0.04547916666666666, 0.33014062499999997, 0.4671666666666667, 0.6107968749999999], "iscrowd": 0, "area": 15392.782899999998, "rle": {"size": [640, 480], "counts": "RQ>8_c00O100000O01000000O010000000O01000O1000O010000O10000O10001N10000O10000O10000O10000O10000O1000000O10000O10001N10001N3N1O2M3N2N2M3N1O2M3N2N2N2N1O2N2N2N2N2N1O2N2N2N2N2N1O2N2O1N2N1O2N2N2N2N2NXgo4"}, "label": "a playstation controller"}]} {"id": 142431, "image": "COCO_train2014_000000142431.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sony playstation controller on the top right of the collection\"."}], "task": "refering", "answer_template": "The \"sony playstation controller on the top right of the collection\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 139, 140, 141, 142, 143, 153, 154, 155, 156, 157, 158, 159, 160, 170, 171, 172, 173, 174, 175, 176, 177, 188, 189, 190, 205, 206, 207, 222, 223, 224], "bbox": [0.04547916666666666, 0.33014062499999997, 0.4671666666666667, 0.6107968749999999], "iscrowd": 0, "area": 15392.782899999998, "rle": {"size": [640, 480], "counts": "RQ>8_c00O100000O01000000O010000000O01000O1000O010000O10000O10001N10000O10000O10000O10000O10000O1000000O10000O10001N10001N3N1O2M3N2N2M3N1O2M3N2N2N2N1O2N2N2N2N2N1O2N2N2N2N2N1O2N2O1N2N1O2N2N2N2N2NXgo4"}, "label": "sony playstation controller on the top right of the collection"}]} {"id": 142431, "image": "COCO_train2014_000000142431.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an xbox video game controller\"."}], "task": "refering", "answer_template": "The \"an xbox video game controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 226, 227, 228, 229, 230, 231, 244, 245, 246, 247, 248, 262, 263, 264, 265, 280], "bbox": [0.21460416666666668, 0.43987499999999996, 0.6458541666666666, 0.7038125], "iscrowd": 0, "area": 17574.045000000006, "rle": {"size": [640, 480], "counts": "dfP22jc04L5M2O2M2O2N1N2O2M2O2M2O1O2M2O101O000O2O000O101O000O2O000O101O000O101O000O2O0O1O101N100O2O1N2O2M2O1N2O1N2O1N2O1O1O1N3N1N2N2O2M2N2O1N3N1N2N2O2M2N2O1N2N3N1N2N2O1N2O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O100010O00001O001O0O101O00001O0O2O1O1N2N2M3N2M3N2M3N1O2M3N2M2O0O2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N101N2O001N101N2O0O2N1O2O1N1O2N101N1O2N2O0O2N1O2O1N1O2N1O2O1N1O2N1O2N1O2N2N1O2N1O4L5K5K5K5K6J5KckY3"}, "label": "an xbox video game controller"}]} {"id": 142431, "image": "COCO_train2014_000000142431.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the game controller in the middle\"."}], "task": "refering", "answer_template": "The \"the game controller in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 226, 227, 228, 229, 230, 231, 244, 245, 246, 247, 248, 262, 263, 264, 265, 280], "bbox": [0.21460416666666668, 0.43987499999999996, 0.6458541666666666, 0.7038125], "iscrowd": 0, "area": 17574.045000000006, "rle": {"size": [640, 480], "counts": "dfP22jc04L5M2O2M2O2N1N2O2M2O2M2O1O2M2O101O000O2O000O101O000O2O000O101O000O101O000O2O0O1O101N100O2O1N2O2M2O1N2O1N2O1N2O1O1O1N3N1N2N2O2M2N2O1N3N1N2N2O2M2N2O1N2N3N1N2N2O1N2O1O1O1O2N1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O100010O00001O001O0O101O00001O0O2O1O1N2N2M3N2M3N2M3N1O2M3N2M2O0O2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N101N2O001N101N2O0O2N1O2O1N1O2N101N1O2N2O0O2N1O2O1N1O2N1O2O1N1O2N1O2N1O2N2N1O2N1O4L5K5K5K5K6J5KckY3"}, "label": "the game controller in the middle"}]} {"id": 183392, "image": "COCO_train2014_000000183392.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sausage under the other sausages nearest to the camera\"."}], "task": "refering", "answer_template": "The \"the sausage under the other sausages nearest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [314, 315, 316, 317, 318, 319, 330, 331, 332, 333, 334, 335, 336, 348, 349, 350, 351, 352], "bbox": [0.4551975051975052, 0.7717968749999999, 0.8224740124740125, 0.9124375], "iscrowd": 0, "area": 8127.518199999998, "rle": {"size": [640, 481], "counts": "e\\Y42lc04M4K4L4L4L4L4N2N2N2N2O1N2N2N3M2N2N2N2N0010O01O001O001O001O010O00001O001O00000001O000000000001O0001O00000000O100O1O1O100O1O100O1O010O1O00100O0010O01O1O010O1O010O1O0G:2N3L2O1N2O001N2O1O1N101O1N2O001N2O1N2O001N2O1O0O2O000O2O000O101N10001N10000O2N1O100O2N1O101N1O100O2N1O100O2N100O2N1O100O2N1O100O0001O000O101O00001O00000O2O0000001O0000000O100000000000K5Fkdd1"}, "label": "the sausage under the other sausages nearest to the camera"}]} {"id": 183392, "image": "COCO_train2014_000000183392.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"underneath sausage closest to red pepper\"."}], "task": "refering", "answer_template": "The \"underneath sausage closest to red pepper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [314, 315, 316, 317, 318, 319, 330, 331, 332, 333, 334, 335, 336, 348, 349, 350, 351, 352], "bbox": [0.4551975051975052, 0.7717968749999999, 0.8224740124740125, 0.9124375], "iscrowd": 0, "area": 8127.518199999998, "rle": {"size": [640, 481], "counts": "e\\Y42lc04M4K4L4L4L4L4N2N2N2N2O1N2N2N3M2N2N2N2N0010O01O001O001O001O010O00001O001O00000001O000000000001O0001O00000000O100O1O1O100O1O100O1O010O1O00100O0010O01O1O010O1O010O1O0G:2N3L2O1N2O001N2O1O1N101O1N2O001N2O1N2O001N2O1O0O2O000O2O000O101N10001N10000O2N1O100O2N1O101N1O100O2N1O100O2N100O2N1O100O2N1O100O0001O000O101O00001O00000O2O0000001O0000000O100000000000K5Fkdd1"}, "label": "underneath sausage closest to red pepper"}]} {"id": 183392, "image": "COCO_train2014_000000183392.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the whie egg holder on the white plate with water in it\"."}], "task": "refering", "answer_template": "The \"the whie egg holder on the white plate with water in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 183, 184, 185, 186, 200, 201, 202, 203, 217, 218, 219, 220, 234, 235, 236, 237, 251, 252, 253, 254, 268, 269, 270, 271, 286, 287, 288, 304], "bbox": [0.783014553014553, 0.41820312499999995, 1.0, 0.745109375], "iscrowd": 0, "area": 16953.879100000002, "rle": {"size": [640, 481], "counts": "mm[7\\1Ub0`0M4L4L3N3M2N3M3M2N3S@aMT>b2cAfM[>\\2cAfM[>]2bAfM[>]2bAeM\\>]2bAfM[>]2bAeM\\>^2aAdM^>^2_AeM^>^2_AdM_>^2_AeM^>^2_AdM_>_2^AdM_>_2^AcM`>`2]AbMa>Q4M2N3M200O1O1O1O1O100O1O1O1OfJVB_4j=^K\\B_4c=^KeB^4Z=`KlB]4S=aKTC[4k in this image.", "objects": [{"patches": [167, 168, 169, 183, 184, 185, 186, 200, 201, 202, 203, 217, 218, 219, 220, 234, 235, 236, 237, 251, 252, 253, 254, 268, 269, 270, 271, 286, 287, 288, 304], "bbox": [0.783014553014553, 0.41820312499999995, 1.0, 0.745109375], "iscrowd": 0, "area": 16953.879100000002, "rle": {"size": [640, 481], "counts": "mm[7\\1Ub0`0M4L4L3N3M2N3M3M2N3S@aMT>b2cAfM[>\\2cAfM[>]2bAfM[>]2bAeM\\>]2bAfM[>]2bAeM\\>^2aAdM^>^2_AeM^>^2_AdM_>^2_AeM^>^2_AdM_>_2^AdM_>_2^AcM`>`2]AbMa>Q4M2N3M200O1O1O1O1O100O1O1O1OfJVB_4j=^K\\B_4c=^KeB^4Z=`KlB]4S=aKTC[4k in this image.", "objects": [{"patches": [230, 231, 246, 247, 248, 249, 264, 265, 266, 267, 282, 283, 284, 285, 299, 300, 301, 302, 317, 318, 319], "bbox": [0.5143243243243243, 0.5882031249999999, 0.8194594594594594, 0.8280937500000001], "iscrowd": 0, "area": 11243.243500000002, "rle": {"size": [640, 481], "counts": "h\\k42lc03M3N1N3M3N2M2N3N2M2O2M3M3N1N3M3O1O001N2O001N2O1O001N2O1O1O0O2O1O1O1O1O1O001O1O1O1O1O1O007H8I7I8I6I7J0O2O1N2O1N101N2N2O1N100O1001N101O0O2O0O2O00O010000O10000O100O10000O10000O100O100O1O10O01O100O1O100O1O100O1O1O2N1O1O2O0O1O2N1O1O2M2N2N3M2N2N3M2N2O2M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M5K5K5K5K4L5Kale1"}, "label": "the top right sausage in the pile"}]} {"id": 183392, "image": "COCO_train2014_000000183392.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass of rice\"."}], "task": "refering", "answer_template": "The \"a glass of rice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 43, 44, 45, 56, 57, 58, 59, 60, 61, 62, 63, 73, 74, 75, 76, 77, 78, 79, 80, 81, 90, 91, 92, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 147, 159, 160, 161, 162, 163, 164, 177, 178, 179, 180, 181, 194, 195, 196, 197, 198, 211, 214], "bbox": [0.3192099792099792, 0.10223437500000002, 0.7816216216216216, 0.534984375], "iscrowd": 0, "area": 42102.33314999999, "rle": {"size": [640, 481], "counts": "h[P37cc0:G9G8M4M2I8I6J7L4L3M4K4M3M3L4M3M3L3N3M3M3L4M2N3L4M3M3L4M2N3M3L4M3M3L3N3M3L4M3M4K6K5K5J6K5J6K5K4K6K5J6K5K:E in this image.", "objects": [{"patches": [41, 42, 43, 44, 45, 56, 57, 58, 59, 60, 61, 62, 63, 73, 74, 75, 76, 77, 78, 79, 80, 81, 90, 91, 92, 93, 94, 95, 96, 97, 108, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 142, 143, 144, 145, 146, 147, 159, 160, 161, 162, 163, 164, 177, 178, 179, 180, 181, 194, 195, 196, 197, 198, 211, 214], "bbox": [0.3192099792099792, 0.10223437500000002, 0.7816216216216216, 0.534984375], "iscrowd": 0, "area": 42102.33314999999, "rle": {"size": [640, 481], "counts": "h[P37cc0:G9G8M4M2I8I6J7L4L3M4K4M3M3L4M3M3L3N3M3M3L4M2N3L4M3M3L4M2N3M3L4M3M3L3N3M3L4M3M4K6K5K5J6K5J6K5K4K6K5J6K5K:E in this image.", "objects": [{"patches": [216, 217, 218, 219, 239, 240, 241, 242, 243, 244, 245, 246, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 428, 429, 430, 431, 432, 433, 434, 435, 436, 454, 455, 456, 457, 458, 480, 481], "bbox": [0.305625, 0.408984375, 0.9842656249999999, 0.885390625], "iscrowd": 0, "area": 86723.10824999998, "rle": {"size": [640, 640], "counts": "Pmj3b0[c03M4M3L3M4L4L4M2M4L4L4L3N3L4L3M4L3N2M4L3N2N3M2N2N3M2O1N3M2N2N3M2O1N2N3M2N2N3N1N2N3M2N2O2M2N2N3N1N2N3M2O1N3M2O1N3M2N2O2M2O1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1N2O2N1O1O2N1O1O2N1000001O00010O000010O0001O01O01O000010O0001O01O01O0000010O000010O0001O00010O000010O0001O01O01O000010O0001O01O0001O00010O000010O0001O00010O000010O0001O01O01O0000010O000010O0001O00010O000010O0001O00010O000010O0000010O0001O00010O000010O0001O00010O000010O0001O00010O0000010O000010O0001O00010O000010O0001O00010O000010O000001O01O01O00010O000010O0001O00010O000010O0001O00010O0000010O00001O01O01O00010O000010O0001O00010O000010O000001O01O01O00010O00001O01O01O00010O000010O0001O0001O01O00010O00001O01O01O00010O00001O01O01O00010O000010O000001O01O01O00010O00001O01O01O00010O00001O01O01O0001O01O00010O00001O01O01O00010O00001O01O01O00010OO1O100O2N1O100O2N100O1O2O0O1M3L4K6L3M3M3M4M2M3M3N2M4L3N2M3M4L3N2M3Ml0UOd1[Ne1[Nci5"}, "label": "a blackberry cellphone sitting on a table alongside another brand of cellphone"}]} {"id": 445540, "image": "COCO_train2014_000000445540.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey blackberry phone sitting on a talble\"."}], "task": "refering", "answer_template": "The \"a grey blackberry phone sitting on a talble\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 219, 239, 240, 241, 242, 243, 244, 245, 246, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 428, 429, 430, 431, 432, 433, 434, 435, 436, 454, 455, 456, 457, 458, 480, 481], "bbox": [0.305625, 0.408984375, 0.9842656249999999, 0.885390625], "iscrowd": 0, "area": 86723.10824999998, "rle": {"size": [640, 640], "counts": "Pmj3b0[c03M4M3L3M4L4L4M2M4L4L4L3N3L4L3M4L3N2M4L3N2N3M2N2N3M2O1N3M2N2N3M2O1N2N3M2N2N3N1N2N3M2N2O2M2N2N3N1N2N3M2O1N3M2O1N3M2N2O2M2O1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1N2O2N1O1O2N1O1O2N1000001O00010O000010O0001O01O01O000010O0001O01O01O0000010O000010O0001O00010O000010O0001O01O01O000010O0001O01O0001O00010O000010O0001O00010O000010O0001O01O01O0000010O000010O0001O00010O000010O0001O00010O000010O0000010O0001O00010O000010O0001O00010O000010O0001O00010O0000010O000010O0001O00010O000010O0001O00010O000010O000001O01O01O00010O000010O0001O00010O000010O0001O00010O0000010O00001O01O01O00010O000010O0001O00010O000010O000001O01O01O00010O00001O01O01O00010O000010O0001O0001O01O00010O00001O01O01O00010O00001O01O01O00010O000010O000001O01O01O00010O00001O01O01O00010O00001O01O01O0001O01O00010O00001O01O01O00010O00001O01O01O00010OO1O100O2N1O100O2N100O1O2O0O1M3L4K6L3M3M3M4M2M3M3N2M4L3N2M3M4L3N2M3Ml0UOd1[Ne1[Nci5"}, "label": "a grey blackberry phone sitting on a talble"}]} {"id": 445540, "image": "COCO_train2014_000000445540.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white rectangular mp3 player\"."}], "task": "refering", "answer_template": "The \"a white rectangular mp3 player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 99, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 324, 325, 326], "bbox": [0.031718750000000004, 0.16985937499999998, 0.45251562500000003, 0.6428906249999999], "iscrowd": 0, "area": 48817.94655, "rle": {"size": [640, 640], "counts": "ef<4lc0n0QOo0QOP1POo0QOP1PO:F2N1O2N1O2O0O2N1O2N2N1O2N1O2N101N1O2N1O2N1O2N1O2N101N1O2N1O2N1O2N1O2N2N101N1O2N1O2N1O2N1O2N101N1O2N1O2N1O2N1O2N101N1O2N2N1O2N1O1O1O1O100O001O1O1O1O1O001O100O1O00O10000O10000O101O0O10000O10000O10001N10000O10000O1000001O001O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2M2O1O2N1O1N3N1O1O2N1N2O2N1O1O2M2O1O2N1O1N3N1O1O2N1N2O2N1O1O2M2O1O2N1O1N3N1O1O2N1N2O2N1O1O2M2O1O2N1O1N3N1O1O2N1N2O2N1O1O2M2O1O2N1O1O2M2O1O2N1O1N3N1O1O2N1N2O2N1O1O2M2O2N1O1O2M2O2N1O1O2M2O2N1O1O2M2O2N1O1O2M2O2N1O1O2M2O1O2N1O2M2O1O2N1O2M2O1O2N1O2M2O1O2N2N2M3N2N2N3M2M3N6J8H9GUbj6"}, "label": "a white rectangular mp3 player"}]} {"id": 445540, "image": "COCO_train2014_000000445540.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white mp3 player sitting on left side of phones\"."}], "task": "refering", "answer_template": "The \"white mp3 player sitting on left side of phones\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 99, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 324, 325, 326], "bbox": [0.031718750000000004, 0.16985937499999998, 0.45251562500000003, 0.6428906249999999], "iscrowd": 0, "area": 48817.94655, "rle": {"size": [640, 640], "counts": "ef<4lc0n0QOo0QOP1POo0QOP1PO:F2N1O2N1O2O0O2N1O2N2N1O2N1O2N101N1O2N1O2N1O2N1O2N101N1O2N1O2N1O2N1O2N2N101N1O2N1O2N1O2N1O2N101N1O2N1O2N1O2N1O2N101N1O2N2N1O2N1O1O1O1O100O001O1O1O1O1O001O100O1O00O10000O10000O101O0O10000O10000O10001N10000O10000O1000001O001O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2M2O1O2N1O1N3N1O1O2N1N2O2N1O1O2M2O1O2N1O1N3N1O1O2N1N2O2N1O1O2M2O1O2N1O1N3N1O1O2N1N2O2N1O1O2M2O1O2N1O1N3N1O1O2N1N2O2N1O1O2M2O1O2N1O1O2M2O1O2N1O1N3N1O1O2N1N2O2N1O1O2M2O2N1O1O2M2O2N1O1O2M2O2N1O1O2M2O2N1O1O2M2O2N1O1O2M2O1O2N1O2M2O1O2N1O2M2O1O2N1O2M2O1O2N2N2M3N2N2N3M2M3N6J8H9GUbj6"}, "label": "white mp3 player sitting on left side of phones"}]} {"id": 445540, "image": "COCO_train2014_000000445540.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older generation apple iphone\"."}], "task": "refering", "answer_template": "The \"an older generation apple iphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229], "bbox": [0.43640625, 0.210796875, 0.9914531249999999, 0.41740625], "iscrowd": 0, "area": 40583.16684999999, "rle": {"size": [640, 640], "counts": "\\c^51hc07J7H9H7H9G8I7H9H7H7I8I6I7I8M20001N100O2O000O2O000O10000O100O1000000000000000000000001O0001O00000000000000000000000000000000000001O000000000001O00000000000000000000000000000001O00000000000000000001O0000000000000000000001O0000000000000000000000000001O00000000000001O000000000000000000000000000000000001O00000001O00000000000000000000000000000000000000000010O000000000000000000000000000000000000000001O00000001O000000000000000000000000000000000001O00000000000001O0000000000000000000000000001O000000000000000000000001O000000000000000001O00000000000000000000000000000001O000000000001O00000000000000000000000000000000000001O0001O00000000000000000000000000000001O1O001O1O100O1O001O1O1O1O1O001O1O1O1O9G9G:D in this image.", "objects": [{"patches": [103, 104, 105, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229], "bbox": [0.43640625, 0.210796875, 0.9914531249999999, 0.41740625], "iscrowd": 0, "area": 40583.16684999999, "rle": {"size": [640, 640], "counts": "\\c^51hc07J7H9H7H9G8I7H9H7H7I8I6I7I8M20001N100O2O000O2O000O10000O100O1000000000000000000000001O0001O00000000000000000000000000000000000001O000000000001O00000000000000000000000000000001O00000000000000000001O0000000000000000000001O0000000000000000000000000001O00000000000001O000000000000000000000000000000000001O00000001O00000000000000000000000000000000000000000010O000000000000000000000000000000000000000001O00000001O000000000000000000000000000000000001O00000000000001O0000000000000000000000000001O000000000000000000000001O000000000000000001O00000000000000000000000000000001O000000000001O00000000000000000000000000000000000001O0001O00000000000000000000000000000001O1O001O1O100O1O001O1O1O1O1O001O1O1O1O9G9G:D in this image.", "objects": [{"patches": [1, 17, 18, 19, 35, 36, 37, 53, 54, 55, 70, 71, 72, 87, 88, 89, 90, 104, 106, 107], "bbox": [0.050958333333333335, 0.030453124999999998, 0.3110625, 0.28556250000000005], "iscrowd": 0, "area": 7517.196549999999, "rle": {"size": [640, 480], "counts": "fP?2nc0?A1O1O2N1N3N1O2N1O2N1O2N1O1O2M2O2N2N2N2N2M3O1N2O1N2O1O1N2O1N2O1O1N2O1O1O1OPN_^Oa1`a0_Ne^O]1Za0bNk^OZ1Ua0fNk^O[1Ta0dNl^O`2Q`0`Mn_OS3`?lL`@^3W?bLh@e3Q?[Lo@e3R?[Lm@e3S?\\Ll@d3T?\\Ll@d3U?\\Lj@e3U?[Lk@e3U?\\Lj@U3f?jLZ@[2a`0fM^_Ol1Pa0TNP_Ok1Ra0UNm^OQ2m`0oMS_O\\2c`0dM\\_Og2Y`0YMg_Oo2Q`0RMn_On2S`0mLQ@S3^`0000000000000000000\\Od0B>A?N20000O100O2O1O1N2O1O1N2O002N4K5L4L4K5O1OO2L4M4K4M3L4M4K4M3L5L3M3M3L5L3M31O1L4J5L5K5K5J6K5K4L6I7J6J6Jbf^6"}, "label": "a green chair with a boy in red sitting on it"}]} {"id": 35945, "image": "COCO_train2014_000000035945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green chair the child is sitting in\"."}], "task": "refering", "answer_template": "The \"green chair the child is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 17, 18, 19, 35, 36, 37, 53, 54, 55, 70, 71, 72, 87, 88, 89, 90, 104, 106, 107], "bbox": [0.050958333333333335, 0.030453124999999998, 0.3110625, 0.28556250000000005], "iscrowd": 0, "area": 7517.196549999999, "rle": {"size": [640, 480], "counts": "fP?2nc0?A1O1O2N1N3N1O2N1O2N1O2N1O1O2M2O2N2N2N2N2M3O1N2O1N2O1O1N2O1N2O1O1N2O1O1O1OPN_^Oa1`a0_Ne^O]1Za0bNk^OZ1Ua0fNk^O[1Ta0dNl^O`2Q`0`Mn_OS3`?lL`@^3W?bLh@e3Q?[Lo@e3R?[Lm@e3S?\\Ll@d3T?\\Ll@d3U?\\Lj@e3U?[Lk@e3U?\\Lj@U3f?jLZ@[2a`0fM^_Ol1Pa0TNP_Ok1Ra0UNm^OQ2m`0oMS_O\\2c`0dM\\_Og2Y`0YMg_Oo2Q`0RMn_On2S`0mLQ@S3^`0000000000000000000\\Od0B>A?N20000O100O2O1O1N2O1O1N2O002N4K5L4L4K5O1OO2L4M4K4M3L4M4K4M3L5L3M3M3L5L3M31O1L4J5L5K5K5J6K5K4L6I7J6J6Jbf^6"}, "label": "green chair the child is sitting in"}]} {"id": 35945, "image": "COCO_train2014_000000035945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green , unoccupied chair near a colorful fire hydrant\"."}], "task": "refering", "answer_template": "The \"a green , unoccupied chair near a colorful fire hydrant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 27, 28, 29, 30, 42, 43, 44, 45, 46, 59, 60, 61, 62, 63, 76, 77, 78, 79, 80, 94, 95, 96, 97, 112, 113, 114, 129, 130, 131, 146, 147], "bbox": [0.46902083333333333, 0.024890625, 0.7800833333333334, 0.378546875], "iscrowd": 0, "area": 14604.437050000002, "rle": {"size": [640, 480], "counts": "Yg\\43dc0>C:I500O1O2O0O100O10000O2O001O1O1N2O1O1O1`]OcN[b0c1M2O1O1O001N2O1O1O001N2O1O00001N101O001O0O101O001O0O2O001M3M2N3N2M2N3M3N1O2O1O001N2O002M3N4L6J5J5L4L3L5L4L4L4L4L4_A]K\\=d5000O10000O10O1000O1000000O10O10O10N2O1N101N2OPMo22O1N2O2M5M4K5K6K4K5L4K5L4K5L4K5K6K4KmMbAjNZ>S1PBjNk=T1_BgN_=V1kBbNU=\\1UC]Nj in this image.", "objects": [{"patches": [12, 13, 27, 28, 29, 30, 42, 43, 44, 45, 46, 59, 60, 61, 62, 63, 76, 77, 78, 79, 80, 94, 95, 96, 97, 112, 113, 114, 129, 130, 131, 146, 147], "bbox": [0.46902083333333333, 0.024890625, 0.7800833333333334, 0.378546875], "iscrowd": 0, "area": 14604.437050000002, "rle": {"size": [640, 480], "counts": "Yg\\43dc0>C:I500O1O2O0O100O10000O2O001O1O1N2O1O1O1`]OcN[b0c1M2O1O1O001N2O1O1O001N2O1O00001N101O001O0O101O001O0O2O001M3M2N3N2M2N3M3N1O2O1O001N2O002M3N4L6J5J5L4L3L5L4L4L4L4L4_A]K\\=d5000O10000O10O1000O1000000O10O10O10N2O1N101N2OPMo22O1N2O2M5M4K5K6K4K5L4K5L4K5L4K5K6K4KmMbAjNZ>S1PBjNk=T1_BgN_=V1kBbNU=\\1UC]Nj in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 27, 28, 30, 31, 32, 46, 47, 48, 63, 64], "bbox": [0.6158541666666667, 0.00021875000000000003, 0.9417083333333333, 0.168390625], "iscrowd": 0, "area": 8192.82755, "rle": {"size": [640, 480], "counts": "ZPi54kc06K4K6K5J5L5J3N001O1N101O1O1O001O1O0O2O1O001O1O1O00N101O1N2O1N2O1N2O0O2O1N2O1N2O1O1N101N2O1N200O100000O100000O1000Z_O[OQ>e0lA^OT>b0hABX>>eAE[>;bAH^>7_AMa>3\\A0d>0YA3g>MUA7k>IRA9o>Gn@ in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 27, 28, 30, 31, 32, 46, 47, 48, 63, 64], "bbox": [0.6158541666666667, 0.00021875000000000003, 0.9417083333333333, 0.168390625], "iscrowd": 0, "area": 8192.82755, "rle": {"size": [640, 480], "counts": "ZPi54kc06K4K6K5J5L5J3N001O1N101O1O1O001O1O0O2O1O001O1O1O00N101O1N2O1N2O1N2O0O2O1N2O1N2O1O1N101N2O1N200O100000O100000O1000Z_O[OQ>e0lA^OT>b0hABX>>eAE[>;bAH^>7_AMa>3\\A0d>0YA3g>MUA7k>IRA9o>Gn@ in this image.", "objects": [{"patches": [3, 4, 20, 21, 37, 38, 39, 54, 55, 56, 57, 72, 73, 74, 90, 91], "bbox": [0.17247916666666668, 0.0, 0.37595833333333334, 0.25609375], "iscrowd": 0, "area": 6970.058550000001, "rle": {"size": [640, 480], "counts": "Vnc18Uc0OR]O6gb01X]O0bb06^]O0Va0]Og_Oe0SO0Ta0^Oc_Oe0YOORa0^Ob_Od0\\O1o`0\\Ob_Oe0_O1m`0[Oa_Of0B2j`0ZOa_Oe0F2g`0ZO`_Of0I2e`0ZO^_Of0M1d`0ZOW_OD1S14Od`0f0\\_OZOd`0f0\\_OZOd`0f0\\_OZOd`0f0\\_OZOd`0f0\\_OZOd`0f0\\_OYOe`0g0[_OYOe`0g0\\_OXOd`0h0\\_OWOe`0j0[_OUOe`0k0^_OROb`0n0`_OPO``0Q1b_OkN_`0V1c_OgN]`0[1e_OaN[`0a1f_O\\NZ`0g1g_OUNY`0n1g_OnMZ`0U2f_OhMZ`0]2b04LUNd^OV1Xa0jNh^OZ1Ta0fNm^O]1o`0cNQ_O`1l`0`NU_O`1j`0`NW_O^1j`0bNX_O\\1h`0cN\\_OZ1d`0eN`_OX1``0TNT_O=a0]1``0_Ne_O_1\\`0[Nm^OMk0f1^a0N3M3OO2N3M3N2N2M3M3N1N3M3M3M5J5L5K0000O3M10O0100O1O00001O000010O0001O1O1N2]Ob^O_Nca0_1^^O\\Nha0c1W^O[Nma0c1S^O[NRb0b1n]O[NYb0a18K4K6J`0@9F3N4L2Nh[k5"}, "label": "the boy wearing a maroon shirt and blue jeans"}]} {"id": 35945, "image": "COCO_train2014_000000035945.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little boy in a red sweatshirt sitting at a table with two other , older , people\"."}], "task": "refering", "answer_template": "The \"a little boy in a red sweatshirt sitting at a table with two other , older , people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 20, 21, 37, 38, 39, 54, 55, 56, 57, 72, 73, 74, 90, 91], "bbox": [0.17247916666666668, 0.0, 0.37595833333333334, 0.25609375], "iscrowd": 0, "area": 6970.058550000001, "rle": {"size": [640, 480], "counts": "Vnc18Uc0OR]O6gb01X]O0bb06^]O0Va0]Og_Oe0SO0Ta0^Oc_Oe0YOORa0^Ob_Od0\\O1o`0\\Ob_Oe0_O1m`0[Oa_Of0B2j`0ZOa_Oe0F2g`0ZO`_Of0I2e`0ZO^_Of0M1d`0ZOW_OD1S14Od`0f0\\_OZOd`0f0\\_OZOd`0f0\\_OZOd`0f0\\_OZOd`0f0\\_OZOd`0f0\\_OYOe`0g0[_OYOe`0g0\\_OXOd`0h0\\_OWOe`0j0[_OUOe`0k0^_OROb`0n0`_OPO``0Q1b_OkN_`0V1c_OgN]`0[1e_OaN[`0a1f_O\\NZ`0g1g_OUNY`0n1g_OnMZ`0U2f_OhMZ`0]2b04LUNd^OV1Xa0jNh^OZ1Ta0fNm^O]1o`0cNQ_O`1l`0`NU_O`1j`0`NW_O^1j`0bNX_O\\1h`0cN\\_OZ1d`0eN`_OX1``0TNT_O=a0]1``0_Ne_O_1\\`0[Nm^OMk0f1^a0N3M3OO2N3M3N2N2M3M3N1N3M3M3M5J5L5K0000O3M10O0100O1O00001O000010O0001O1O1N2]Ob^O_Nca0_1^^O\\Nha0c1W^O[Nma0c1S^O[NRb0b1n]O[NYb0a18K4K6J`0@9F3N4L2Nh[k5"}, "label": "a little boy in a red sweatshirt sitting at a table with two other , older , people"}]} {"id": 3178, "image": "COCO_train2014_000000003178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby giraffe to the right of its mother\"."}], "task": "refering", "answer_template": "The \"baby giraffe to the right of its mother\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 60, 61, 62, 63, 64, 65, 83, 84, 85, 86, 87, 88, 89, 107, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 269, 270, 271, 272, 292, 293, 294, 295, 315, 316, 317, 318], "bbox": [0.644671875, 0.06530474040632055, 0.9189999999999999, 0.8416252821670429], "iscrowd": 0, "area": 35252.725249999996, "rle": {"size": [443, 640], "counts": "Xhb57\\=8I7H7I8I7H8I7M30O1000lDTNY:l1cE[NZ:e1aEdN[:\\1aEkN\\:U1`ERO]:m0`E[O[:f0`EB]:R2L3M4L3M4L3M4L3M4L4K6K4L4L4L5K4L4L4L4L5J5_LQJnNS6k0PJUOT6d0oI\\OV6=lICX66kIJY6OjI1Z6HiI8\\6AfI?^6gNXJY1l5eMRKZ2X80000000000001O00001O001O0O1O2O0O2N1O1O2O0I8]Ob0]OQMZGo1S8^NTHb1Y7iNnHW1d6QObIP1Q6UOVJk0^5[OiJd0l4AZK`0Y4EnK;f3K`L6V3LRM3m2HYM9d2BcM?Z2]OlMd0R2WOUNh0i1SO^Nn0_1nNgNS1W1gNPOZ1m0bNYO_1d0\\NDc1;XNHl1P1UMQOQ3g56J6J5K6J6J6J5K6J6J6I7J5K6J6J6J6J5K6J6J6J5fNWI]Ko6_4X1N3L3M4M2M4L3N3L3M3N3L3M`GhLd6U3eHUOj5h0_Ic1Q5[4XO01O00001O001O001O001C<_Oa0_Ob0]Ob0_Ob0^Oa0B?H7J7H7I7J7H7K67H8I6J7K5K5J6K5K4LN1N3ZMiGCY83[H@h77kH\\OW7;]IYOe6=oIVOT6a0_JSOc5c0RKoNP5i0W4F:Fn[f0"}, "label": "baby giraffe to the right of its mother"}]} {"id": 3178, "image": "COCO_train2014_000000003178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffe to the right of two other giraffes\"."}], "task": "refering", "answer_template": "The \"giraffe to the right of two other giraffes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 60, 61, 62, 63, 64, 65, 83, 84, 85, 86, 87, 88, 89, 107, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 269, 270, 271, 272, 292, 293, 294, 295, 315, 316, 317, 318], "bbox": [0.644671875, 0.06530474040632055, 0.9189999999999999, 0.8416252821670429], "iscrowd": 0, "area": 35252.725249999996, "rle": {"size": [443, 640], "counts": "Xhb57\\=8I7H7I8I7H8I7M30O1000lDTNY:l1cE[NZ:e1aEdN[:\\1aEkN\\:U1`ERO]:m0`E[O[:f0`EB]:R2L3M4L3M4L3M4L3M4L4K6K4L4L4L5K4L4L4L4L5J5_LQJnNS6k0PJUOT6d0oI\\OV6=lICX66kIJY6OjI1Z6HiI8\\6AfI?^6gNXJY1l5eMRKZ2X80000000000001O00001O001O0O1O2O0O2N1O1O2O0I8]Ob0]OQMZGo1S8^NTHb1Y7iNnHW1d6QObIP1Q6UOVJk0^5[OiJd0l4AZK`0Y4EnK;f3K`L6V3LRM3m2HYM9d2BcM?Z2]OlMd0R2WOUNh0i1SO^Nn0_1nNgNS1W1gNPOZ1m0bNYO_1d0\\NDc1;XNHl1P1UMQOQ3g56J6J5K6J6J6J5K6J6J6I7J5K6J6J6J6J5K6J6J6J5fNWI]Ko6_4X1N3L3M4M2M4L3N3L3M3N3L3M`GhLd6U3eHUOj5h0_Ic1Q5[4XO01O00001O001O001O001C<_Oa0_Ob0]Ob0_Ob0^Oa0B?H7J7H7I7J7H7K67H8I6J7K5K5J6K5K4LN1N3ZMiGCY83[H@h77kH\\OW7;]IYOe6=oIVOT6a0_JSOc5c0RKoNP5i0W4F:Fn[f0"}, "label": "giraffe to the right of two other giraffes"}]} {"id": 3178, "image": "COCO_train2014_000000003178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the mother giraffe in the middle\"."}], "task": "refering", "answer_template": "The \"the mother giraffe in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 53, 54, 55, 56, 57, 58, 59, 60, 61, 77, 78, 79, 80, 81, 82, 83, 104, 105, 106, 122, 123, 127, 128, 129, 145, 146, 151, 152, 168, 169, 174, 175, 191, 192, 197, 198, 214, 215, 220, 221, 237, 238, 243, 244, 260, 261, 266, 267, 283, 284, 289, 290, 305, 306, 307, 308, 312, 313], "bbox": [0.24504687500000003, 0.0009932279909706547, 0.660328125, 0.8627313769751693], "iscrowd": 0, "area": 40367.424800000015, "rle": {"size": [443, 640], "counts": "\\nS28\\==G6K3M3M2M3N3M2N2N3M20000010O000001O00001O00001O0nM]NlGc1R8aNlG`1P8cNPH]1n7eNRH\\1j7gNVHY1h7iNXHW1f7kNZHV1c7lN]HT1b7mN^HS1a7nN_HS1^7oNbHR1\\7nNeHR1Y7POgHQ1W7POiHQ1T7QOlHP1P7SOPIm0m6VOSIk0i6XOWIi0d6[O\\Ie0a6^O_Ic0T1[Nk1U1QMa0P1dN_1T1aM9l0mNQ1T1TNOh05E6CFd0>WO75\\Oa0e0fN=i0nN=m0PNk0c1YN9i4nMYIj1o14d4[N[Ia1R2Of4bNVI`1T2Nf4cNTI`1W2Mc4h0^KWOb4i0^KWOb4i0_KWO^4k0cKTO\\4m0eKSOY4n0gKROX4o0iKPOQ4V1PLjNU3P2kLPNb2c2_M\\MS2Q1_KSO_2Ld1[1PLiN]2K_1]1WLhNZ2Le1R1TLROX2KQ2c0jKBV2Ja20\\K6S2K`3lN`JY1Q2Jg3cNZJc1P2Im3ZNVJm1n1If47ZKIf47[KHe48[KId47]KHc48]KIb47_KH`49`KH_48bKG^49cKG\\49dKGcMYOZ5P1TMg1XNkL[2^1]O[4FeK;Z5eNfJ[1]5bNcJ_1`5\\NaJd1b5aM^H7P2Y2X6aMhI_2[6^MeIc2[6\\MeIe2[6ZMeIf2\\6XMeIi2Z6WMfIi2[6UMfIk2e81O00001O001O00001O0000001O000000001O0000N2I7L4M3O100001O0000001O00001O0000001O0000001O00001O000000001O00000000001O0000001O0000001O0000001O1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O1O2M3N2N1O2N2N1O2N2N2N1O2N2N2N1O2N2N2N1O2N2N2N1O2N2N2N2N3M4L5K4L4eMSJ^J0m1Q6_3QJcJOn1T6X3PJiJMo1W6Q3nIoJMP2W6m2TJSMo5h2SJXMP6d2QJ\\MR6_2PJaMR6\\2nIeMU6V2mIjMV6R2kInMW6n1kIRN]6d1eI\\Nf6ROaKn0i4VNiKj1T8000000000[HoM_3R2bLRNY3n1hLWNR3i1nL\\Nm2e1SM^Ni2c1VM_Nh2b1bMVN[2k1fMUNX2l1hMVNU2k1lMUN_1DaJX2Q4UN`NDD2jMV2R4UNRN?AXOZNU2T4TNdMZ1^O^NaNJO\\2\\3oMTM4Y1Q2UOhMPOK1Z2V3VN[MLP1e2QO^M\\Ol1Q3]N`MFc0S3YOF]2hMaMEa0P7f1cIdMAa0R7c1eIfM_Ob0Q7b1kJXN[5c1oJSNV5h1b3K6I7J6I7H9Ga0@ajm2"}, "label": "the mother giraffe in the middle"}]} {"id": 3178, "image": "COCO_train2014_000000003178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tan and white giraffe with a small giraffe under it\"."}], "task": "refering", "answer_template": "The \"a tan and white giraffe with a small giraffe under it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 53, 54, 55, 56, 57, 58, 59, 60, 61, 77, 78, 79, 80, 81, 82, 83, 104, 105, 106, 122, 123, 127, 128, 129, 145, 146, 151, 152, 168, 169, 174, 175, 191, 192, 197, 198, 214, 215, 220, 221, 237, 238, 243, 244, 260, 261, 266, 267, 283, 284, 289, 290, 305, 306, 307, 308, 312, 313], "bbox": [0.24504687500000003, 0.0009932279909706547, 0.660328125, 0.8627313769751693], "iscrowd": 0, "area": 40367.424800000015, "rle": {"size": [443, 640], "counts": "\\nS28\\==G6K3M3M2M3N3M2N2N3M20000010O000001O00001O00001O0nM]NlGc1R8aNlG`1P8cNPH]1n7eNRH\\1j7gNVHY1h7iNXHW1f7kNZHV1c7lN]HT1b7mN^HS1a7nN_HS1^7oNbHR1\\7nNeHR1Y7POgHQ1W7POiHQ1T7QOlHP1P7SOPIm0m6VOSIk0i6XOWIi0d6[O\\Ie0a6^O_Ic0T1[Nk1U1QMa0P1dN_1T1aM9l0mNQ1T1TNOh05E6CFd0>WO75\\Oa0e0fN=i0nN=m0PNk0c1YN9i4nMYIj1o14d4[N[Ia1R2Of4bNVI`1T2Nf4cNTI`1W2Mc4h0^KWOb4i0^KWOb4i0_KWO^4k0cKTO\\4m0eKSOY4n0gKROX4o0iKPOQ4V1PLjNU3P2kLPNb2c2_M\\MS2Q1_KSO_2Ld1[1PLiN]2K_1]1WLhNZ2Le1R1TLROX2KQ2c0jKBV2Ja20\\K6S2K`3lN`JY1Q2Jg3cNZJc1P2Im3ZNVJm1n1If47ZKIf47[KHe48[KId47]KHc48]KIb47_KH`49`KH_48bKG^49cKG\\49dKGcMYOZ5P1TMg1XNkL[2^1]O[4FeK;Z5eNfJ[1]5bNcJ_1`5\\NaJd1b5aM^H7P2Y2X6aMhI_2[6^MeIc2[6\\MeIe2[6ZMeIf2\\6XMeIi2Z6WMfIi2[6UMfIk2e81O00001O001O00001O0000001O000000001O0000N2I7L4M3O100001O0000001O00001O0000001O0000001O00001O000000001O00000000001O0000001O0000001O0000001O1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O1O2M3N2N1O2N2N1O2N2N2N1O2N2N2N1O2N2N2N1O2N2N2N1O2N2N2N2N3M4L5K4L4eMSJ^J0m1Q6_3QJcJOn1T6X3PJiJMo1W6Q3nIoJMP2W6m2TJSMo5h2SJXMP6d2QJ\\MR6_2PJaMR6\\2nIeMU6V2mIjMV6R2kInMW6n1kIRN]6d1eI\\Nf6ROaKn0i4VNiKj1T8000000000[HoM_3R2bLRNY3n1hLWNR3i1nL\\Nm2e1SM^Ni2c1VM_Nh2b1bMVN[2k1fMUNX2l1hMVNU2k1lMUN_1DaJX2Q4UN`NDD2jMV2R4UNRN?AXOZNU2T4TNdMZ1^O^NaNJO\\2\\3oMTM4Y1Q2UOhMPOK1Z2V3VN[MLP1e2QO^M\\Ol1Q3]N`MFc0S3YOF]2hMaMEa0P7f1cIdMAa0R7c1eIfM_Ob0Q7b1kJXN[5c1oJSNV5h1b3K6I7J6I7H9Ga0@ajm2"}, "label": "a tan and white giraffe with a small giraffe under it"}]} {"id": 3178, "image": "COCO_train2014_000000003178.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby giraffe nurses from its mother\"."}], "task": "refering", "answer_template": "The \"a baby giraffe nurses from its mother\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 49, 50, 51, 52, 53, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 211, 212, 213, 214, 215, 231, 232, 234, 235, 237, 238, 254, 255, 258, 260, 261, 277, 281, 283, 284, 300, 304, 305, 306, 307, 323, 327], "bbox": [0.046234375, 0.09663656884875847, 0.5288593749999999, 0.9303386004514672], "iscrowd": 0, "area": 50102.018249999994, "rle": {"size": [443, 640], "counts": "oW=7f27h72oG2VN@Q9h0aHJXNCS9n0]HBXNFV9R1oG8f4GiKW3Q4`MXKe2c4iMoJ\\2k4RNhJS2S5]3M3O2N1O1N2O1O2M2OUO]K`Hb4V7kKhHS4P7YLnHf3h6gLVIX3a6SM_Ik2X6aMfI]2W6jMgIU2W6PNgIo1X6TNgIj1X6[NgIc1W6bNgI]1X6gNfIX1X6U301N2O0O2O1N2N101N2O1N3N1N3N1N3M2O2M2O2M2O2M2O2M3N101N101N101N101N1O2O0O2O0O2O0O2O1N101N101N101N1O2O0O2O0O2O010O0100O1000O0100O10O0101N3M2N2O3Lb0^Oi1WN3N2M2N3M3N2M3M6iK_Hi1h7jMaHT2d7^MgH_2^9L:F6K5J5L000O100O1000O0100O01000O01hMYM\\If2[6lM]IT2Z6]N]Id1[6lN]IS1\\6]O\\Ic0]6LZI4_6 in this image.", "objects": [{"patches": [211, 212, 213, 228, 229, 230, 245, 246, 247, 262, 263, 264, 279, 280, 281, 282, 295, 296, 297, 298, 299, 312, 313, 314, 315, 329, 330, 331, 332, 347, 348, 349, 364, 365, 366, 382, 383], "bbox": [0.381375, 0.52478125, 0.6366458333333334, 0.9797343749999999], "iscrowd": 0, "area": 21408.19185, "rle": {"size": [640, 480], "counts": "gkb34jc06J6J6J7I6J7I7I7I6J7I7H8XNbMbAe2V>gM_A_2[>lMZA[2^>QNWAV2b>j1H7J7H8I7YOd0ZOe0[Oe0M4M2O2M2N100O00010O010O0010O0100O1O100000000000000000000000004L5L5J2N01O00001N101O1O001O1O001O1O001O1O001O1O0SIfDT5[;gJREP5Q;iJWES5l:fJYEY5k:`JYE_5m:WJYEg5a<\\L]Bi0d=ROgBg0[=TOoBe0R=VOQDJQ<2XDFl;9XDAj;>[D[Oh;d0^DUOd;k0`DoNb;P1dDhN`;V1eDdN];[1iD]NZ;b1kDXNX;f1c3N2N2N2N3M2N2N3M3M2N3M3L4M2N3M3M1N3N2N1O2N2N1O2LdX\\3"}, "label": "a small child"}]} {"id": 339051, "image": "COCO_train2014_000000339051.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy looking in a refrigerator\"."}], "task": "refering", "answer_template": "The \"a boy looking in a refrigerator\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 228, 229, 230, 245, 246, 247, 262, 263, 264, 279, 280, 281, 282, 295, 296, 297, 298, 299, 312, 313, 314, 315, 329, 330, 331, 332, 347, 348, 349, 364, 365, 366, 382, 383], "bbox": [0.381375, 0.52478125, 0.6366458333333334, 0.9797343749999999], "iscrowd": 0, "area": 21408.19185, "rle": {"size": [640, 480], "counts": "gkb34jc06J6J6J7I6J7I7I7I6J7I7H8XNbMbAe2V>gM_A_2[>lMZA[2^>QNWAV2b>j1H7J7H8I7YOd0ZOe0[Oe0M4M2O2M2N100O00010O010O0010O0100O1O100000000000000000000000004L5L5J2N01O00001N101O1O001O1O001O1O001O1O001O1O0SIfDT5[;gJREP5Q;iJWES5l:fJYEY5k:`JYE_5m:WJYEg5a<\\L]Bi0d=ROgBg0[=TOoBe0R=VOQDJQ<2XDFl;9XDAj;>[D[Oh;d0^DUOd;k0`DoNb;P1dDhN`;V1eDdN];[1iD]NZ;b1kDXNX;f1c3N2N2N2N3M2N2N3M3M2N3M3L4M2N3M3M1N3N2N1O2N2N1O2LdX\\3"}, "label": "a boy looking in a refrigerator"}]} {"id": 339051, "image": "COCO_train2014_000000339051.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman or girl in a red shirt with brown hair\"."}], "task": "refering", "answer_template": "The \"a woman or girl in a red shirt with brown hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [202, 217, 218, 219, 220, 234, 235, 236, 237, 250, 251, 252, 253, 254, 266, 267, 268, 269, 270, 271, 283, 284, 285, 286, 287, 288, 300, 301, 302, 303, 304, 305, 317, 318, 319, 320, 321, 322, 335, 336, 337, 338, 339, 352, 353, 354, 355, 356, 370, 371, 372, 373, 387, 388, 389, 390], "bbox": [0.6636666666666666, 0.512390625, 1.0, 0.989859375], "iscrowd": 0, "area": 37790.08625, "rle": {"size": [640, 480], "counts": "cjW6<[c0=DB>D in this image.", "objects": [{"patches": [202, 217, 218, 219, 220, 234, 235, 236, 237, 250, 251, 252, 253, 254, 266, 267, 268, 269, 270, 271, 283, 284, 285, 286, 287, 288, 300, 301, 302, 303, 304, 305, 317, 318, 319, 320, 321, 322, 335, 336, 337, 338, 339, 352, 353, 354, 355, 356, 370, 371, 372, 373, 387, 388, 389, 390], "bbox": [0.6636666666666666, 0.512390625, 1.0, 0.989859375], "iscrowd": 0, "area": 37790.08625, "rle": {"size": [640, 480], "counts": "cjW6<[c0=DB>D in this image.", "objects": [{"patches": [158, 159, 180, 181, 182, 183, 203, 204, 205, 206, 227, 228, 229, 249, 250, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344], "bbox": [0.84975, 0.44887587822014047, 1.0, 0.9963231850117095], "iscrowd": 0, "area": 16417.8921, "rle": {"size": [427, 640], "counts": "^QS7m04Bc;V1N1O2N2O1RFVNY8k1eGVN[8k1bGXN]8g1cG[N[8f1cG]N\\8c1bG`N]8`1aGcN^8]1aGeN]8\\1aGfN_8Z1_GiN`8W1_GkN`8U1^GnNa8R1]GQOa8P1]GSOb8l0^GVOa8j0^GXOa8h0]G\\O`8e0_G]O`8c0_G_O`8a0^GBa8>^GG]8;aGIi0`Nn5g1WINd0iNk5Y1`I2=TOi5k0hI67D`56WJ;2Cf54VJ=LDn5OUJh2k5XMSJj2m5WMQJj2n5WMQJj2o5Y2000000001O001O001O001O00001O001O001O1O1O1O1O001O1O1O1O1O001O1O1O1O2N2N1O2N2N1O2N2N1O2N2N1O2N1O1O1O1O1J6J6J6J6J6RNiGaN]8W1PH^N[8U1SH_N\\8m0THcN1"}, "label": "a man with long hair and a black shirt drinking from a glass"}]} {"id": 535666, "image": "COCO_train2014_000000535666.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with long hair and a beard\"."}], "task": "refering", "answer_template": "The \"a man with long hair and a beard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 180, 181, 182, 183, 203, 204, 205, 206, 227, 228, 229, 249, 250, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344], "bbox": [0.84975, 0.44887587822014047, 1.0, 0.9963231850117095], "iscrowd": 0, "area": 16417.8921, "rle": {"size": [427, 640], "counts": "^QS7m04Bc;V1N1O2N2O1RFVNY8k1eGVN[8k1bGXN]8g1cG[N[8f1cG]N\\8c1bG`N]8`1aGcN^8]1aGeN]8\\1aGfN_8Z1_GiN`8W1_GkN`8U1^GnNa8R1]GQOa8P1]GSOb8l0^GVOa8j0^GXOa8h0]G\\O`8e0_G]O`8c0_G_O`8a0^GBa8>^GG]8;aGIi0`Nn5g1WINd0iNk5Y1`I2=TOi5k0hI67D`56WJ;2Cf54VJ=LDn5OUJh2k5XMSJj2m5WMQJj2n5WMQJj2o5Y2000000001O001O001O001O00001O001O001O1O1O1O1O001O1O1O1O1O001O1O1O1O2N2N1O2N2N1O2N2N1O2N2N1O2N1O1O1O1O1J6J6J6J6J6RNiGaN]8W1PH^N[8U1SH_N\\8m0THcN1"}, "label": "a man with long hair and a beard"}]} {"id": 535666, "image": "COCO_train2014_000000535666.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in the background looking at her phone\"."}], "task": "refering", "answer_template": "The \"woman in the background looking at her phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 186, 207, 208, 209, 210, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 299, 300, 301, 302, 303, 325, 326], "bbox": [0.0, 0.5617330210772834, 0.217921875, 0.9867915690866511], "iscrowd": 0, "area": 15571.426250000004, "rle": {"size": [427, 640], "counts": "P<8j;:bD5d:h1D9L3N2N2O010O010O01O01O010O01N1N\\Oc02N1O100O1O2N1I8E in this image.", "objects": [{"patches": [185, 186, 207, 208, 209, 210, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 299, 300, 301, 302, 303, 325, 326], "bbox": [0.0, 0.5617330210772834, 0.217921875, 0.9867915690866511], "iscrowd": 0, "area": 15571.426250000004, "rle": {"size": [427, 640], "counts": "P<8j;:bD5d:h1D9L3N2N2O010O010O01O01O010O01N1N\\Oc02N1O100O1O2N1I8E in this image.", "objects": [{"patches": [8, 9, 10, 30, 31, 32, 33, 34, 53, 54, 55, 56, 57, 77, 78, 79, 80, 100, 101, 102, 103, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 197, 213, 214, 215, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 263, 264, 265, 281, 282, 283, 284, 286, 287, 304, 305, 306, 307, 308, 328, 329, 330, 331, 332], "bbox": [0.23151562499999997, 0.013278688524590163, 0.6170781249999999, 0.997072599531616], "iscrowd": 0, "area": 52153.42675000001, "rle": {"size": [427, 640], "counts": "XRn13n<[MK]2T1n1gNXN:]MK^2S1k1mNjKIj1;TNL^2R1i1QObKNR23UNK`2Q1f1HdM\\OWNJ^2S1d1JgMXOXNK\\2T1U1:VNgNYNJ]2T1n0c0[N_NZNJ]2T1m0d0\\N]N[NK`2P1h0j0\\N[N]NFg2P1`0Q1[NXNn1b0GW1[NWNP2 in this image.", "objects": [{"patches": [8, 9, 10, 30, 31, 32, 33, 34, 53, 54, 55, 56, 57, 77, 78, 79, 80, 100, 101, 102, 103, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 197, 213, 214, 215, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 263, 264, 265, 281, 282, 283, 284, 286, 287, 304, 305, 306, 307, 308, 328, 329, 330, 331, 332], "bbox": [0.23151562499999997, 0.013278688524590163, 0.6170781249999999, 0.997072599531616], "iscrowd": 0, "area": 52153.42675000001, "rle": {"size": [427, 640], "counts": "XRn13n<[MK]2T1n1gNXN:]MK^2S1k1mNjKIj1;TNL^2R1i1QObKNR23UNK`2Q1f1HdM\\OWNJ^2S1d1JgMXOXNK\\2T1U1:VNgNYNJ]2T1n0c0[N_NZNJ]2T1m0d0\\N]N[NK`2P1h0j0\\N[N]NFg2P1`0Q1[NXNn1b0GW1[NWNP2 in this image.", "objects": [{"patches": [163, 164, 178, 179, 192, 193, 194, 207, 208, 209, 222, 223, 224, 237, 238, 239, 252, 253, 254, 267, 268, 269, 282], "bbox": [0.8372535211267605, 0.43820312499999997, 1.0, 0.8022500000000001], "iscrowd": 0, "area": 10529.377949999998, "rle": {"size": [640, 426], "counts": "Qco6c0ib0e0[Oe0\\Od0[Oe0\\Od0G:E:G9G9F:G9F:G5O04M3L3M4M3L4L4M2M4L4M3L4L4M2M4L01L3M3N3N15K4M3L`BQLW;k3PDSMo;j2WCSNi in this image.", "objects": [{"patches": [163, 164, 178, 179, 192, 193, 194, 207, 208, 209, 222, 223, 224, 237, 238, 239, 252, 253, 254, 267, 268, 269, 282], "bbox": [0.8372535211267605, 0.43820312499999997, 1.0, 0.8022500000000001], "iscrowd": 0, "area": 10529.377949999998, "rle": {"size": [640, 426], "counts": "Qco6c0ib0e0[Oe0\\Od0[Oe0\\Od0G:E:G9G9F:G9F:G5O04M3L3M4M3L4L4M2M4L4M3L4L4M2M4L01L3M3N3N15K4M3L`BQLW;k3PDSMo;j2WCSNi in this image.", "objects": [{"patches": [95, 96, 97, 106, 107, 108, 110, 111, 112, 121, 122, 123, 124, 125, 126, 127, 128, 139, 140, 141, 142, 143, 144, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 189, 190, 191, 199, 200, 201, 203, 204, 205, 206, 218, 219, 220, 221, 233, 234, 235, 236, 237, 248, 249, 250, 251, 252, 263, 264, 265, 266, 278, 279, 280, 281, 293, 294, 295, 308, 309, 310, 311, 322, 323, 324, 325, 326, 339, 340], "bbox": [0.05518779342723005, 0.273125, 0.8064084507042253, 0.9716250000000001], "iscrowd": 0, "area": 43678.644150000015, "rle": {"size": [640, 426], "counts": "_V?1oc01O001O0000001O0a\\OOob01i\\O8ob0EQ]Oe0ob0]Oo\\Oc0Qc0_Om\\Oa0Sc06010O00000010O0000010O000001O001O1O1O001N2O001O1O001O1O010O0010O001O01O01O0010O00010O01O00010O001O01O01O001O01O01O00001O0010O0001O001O000010O01O00001O00010Y@YOSl4101N100O2O0O101N101N101N10XIfJlNW5T1kJmNS5Q1oJROn4m0TKTOk4i0XKXOf4h0[KYOd4f0]K[Oa4e0_KA[4>fKHU47lKJR46nKJS45mKLR43oKNP42PLNQ41oK0K`Hj3`7\\L0HbHl3T7mKPI`0\\7CcHo3Q7oKoH?^7@fHQ4k6RLPI=k7`3TOULPI:m7b3ROULoH:P8`3POXLoH8R8_3nNlLS1R3mNoLS1Q3mNoLT1P3kNQMV1n2iNSMX1l2hNSMY1m2`HZLY5j0X2k2\\HjLo4 in this image.", "objects": [{"patches": [95, 96, 97, 106, 107, 108, 110, 111, 112, 121, 122, 123, 124, 125, 126, 127, 128, 139, 140, 141, 142, 143, 144, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 189, 190, 191, 199, 200, 201, 203, 204, 205, 206, 218, 219, 220, 221, 233, 234, 235, 236, 237, 248, 249, 250, 251, 252, 263, 264, 265, 266, 278, 279, 280, 281, 293, 294, 295, 308, 309, 310, 311, 322, 323, 324, 325, 326, 339, 340], "bbox": [0.05518779342723005, 0.273125, 0.8064084507042253, 0.9716250000000001], "iscrowd": 0, "area": 43678.644150000015, "rle": {"size": [640, 426], "counts": "_V?1oc01O001O0000001O0a\\OOob01i\\O8ob0EQ]Oe0ob0]Oo\\Oc0Qc0_Om\\Oa0Sc06010O00000010O0000010O000001O001O1O1O001N2O001O1O001O1O010O0010O001O01O01O0010O00010O01O00010O001O01O01O001O01O01O00001O0010O0001O001O000010O01O00001O00010Y@YOSl4101N100O2O0O101N101N101N10XIfJlNW5T1kJmNS5Q1oJROn4m0TKTOk4i0XKXOf4h0[KYOd4f0]K[Oa4e0_KA[4>fKHU47lKJR46nKJS45mKLR43oKNP42PLNQ41oK0K`Hj3`7\\L0HbHl3T7mKPI`0\\7CcHo3Q7oKoH?^7@fHQ4k6RLPI=k7`3TOULPI:m7b3ROULoH:P8`3POXLoH8R8_3nNlLS1R3mNoLS1Q3mNoLT1P3kNQMV1n2iNSMX1l2hNSMY1m2`HZLY5j0X2k2\\HjLo4 in this image.", "objects": [{"patches": [54, 55, 56, 57, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 142, 143, 147], "bbox": [0.171109375, 0.16234192037470724, 0.524328125, 0.42117096018735356], "iscrowd": 0, "area": 13983.31745, "rle": {"size": [427, 640], "counts": "mP^1>k<2N2N2N2N2N2N2O1N2N2N2O1O1N2O1O001O1O1O1O1O1O00100O1O1O1O1O001O1O1O100O1OO_DZNa;f120O1000O010000O1000O010000O10O10O10000O10O10O10000O01000O10000O10000O10000O1000O010000O10000O10000O1O100O100O100O1O100O100O1O100O100O1O101N100O100O1O2O0O100O1O100O2O0O1O100O101N100O1O2O0O2O001O001O0O2O001O00001O0O2O1O0000O1N2O0O2O1O1N2O1K5C in this image.", "objects": [{"patches": [54, 55, 56, 57, 75, 76, 77, 78, 79, 80, 96, 97, 98, 99, 100, 101, 102, 103, 119, 120, 121, 122, 123, 124, 142, 143, 147], "bbox": [0.171109375, 0.16234192037470724, 0.524328125, 0.42117096018735356], "iscrowd": 0, "area": 13983.31745, "rle": {"size": [427, 640], "counts": "mP^1>k<2N2N2N2N2N2N2O1N2N2N2O1O1N2O1O001O1O1O1O1O1O00100O1O1O1O1O001O1O1O100O1OO_DZNa;f120O1000O010000O1000O010000O10O10O10000O10O10O10000O01000O10000O10000O10000O1000O010000O10000O10000O1O100O100O100O1O100O100O1O100O100O1O101N100O100O1O2O0O100O1O100O2O0O1O100O101N100O1O2O0O2O001O001O0O2O001O00001O0O2O1O0000O1N2O0O2O1O1N2O1K5C in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68], "bbox": [0.61771875, 0.0, 1.0, 0.19824355971896956], "iscrowd": 0, "area": 17189.518450000003, "rle": {"size": [427, 640], "counts": "ifT51Z==C=C=C9G001O000000001O000000001O0000001O000000001O000000001O000000001O000000001O000000001O000000001O0000001O000000001O000000001O0000001O0000001O0000001O00001O0000001O0000001O0000001O0000001O00001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O00000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000L4K5K5L4K5K5K5K5K5K5K5L4K5K5K5K5K5"}, "label": "a chair at the front of a table"}]} {"id": 183435, "image": "COCO_train2014_000000183435.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wall is striped with the colors brown and white\"."}], "task": "refering", "answer_template": "The \"a wall is striped with the colors brown and white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68], "bbox": [0.61771875, 0.0, 1.0, 0.19824355971896956], "iscrowd": 0, "area": 17189.518450000003, "rle": {"size": [427, 640], "counts": "ifT51Z==C=C=C9G001O000000001O000000001O0000001O000000001O000000001O000000001O000000001O000000001O000000001O0000001O000000001O000000001O0000001O0000001O0000001O00001O0000001O0000001O0000001O0000001O00001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O00000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000L4K5K5L4K5K5K5K5K5K5K5L4K5K5K5K5K5"}, "label": "a wall is striped with the colors brown and white"}]} {"id": 183435, "image": "COCO_train2014_000000183435.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair on the left of the fruit\"."}], "task": "refering", "answer_template": "The \"the chair on the left of the fruit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 92, 115, 116, 138, 139, 161, 162, 184, 185, 186, 207, 208, 209, 230, 231, 232, 253, 254, 276], "bbox": [0.0034687500000000005, 0.1922014051522248, 0.11957812500000001, 0.8857142857142857], "iscrowd": 0, "area": 10600.934400000002, "rle": {"size": [427, 640], "counts": "Ym0W9V4L5K4M4K4L5K5K4L5K4L5K4L5K5K4L5K4L5K4L5K5K4L5K4L5K4L5K4M4L4L3M4L3M4L3M4L4L3M4L3M4L3M4L4L3M4L3M4M2M4L3M4L4L3M4L3M4L3M4L4L3M4L3M4L3M4K4L5L4K4M4K4L5L3LS`Z7"}, "label": "the chair on the left of the fruit"}]} {"id": 183435, "image": "COCO_train2014_000000183435.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair near to the table full of vegetables\"."}], "task": "refering", "answer_template": "The \"a chair near to the table full of vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 92, 115, 116, 138, 139, 161, 162, 184, 185, 186, 207, 208, 209, 230, 231, 232, 253, 254, 276], "bbox": [0.0034687500000000005, 0.1922014051522248, 0.11957812500000001, 0.8857142857142857], "iscrowd": 0, "area": 10600.934400000002, "rle": {"size": [427, 640], "counts": "Ym0W9V4L5K4M4K4L5K5K4L5K4L5K4L5K5K4L5K4L5K4L5K5K4L5K4L5K4L5K4M4L4L3M4L3M4L3M4L4L3M4L3M4L3M4L4L3M4L3M4M2M4L3M4L4L3M4L3M4L3M4L4L3M4L3M4L3M4K4L5L4K4M4K4L5L3LS`Z7"}, "label": "a chair near to the table full of vegetables"}]} {"id": 183435, "image": "COCO_train2014_000000183435.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a group of six oranges on a wooden table surrounded by several other fruits and vegetables\"."}], "task": "refering", "answer_template": "The \"a group of six oranges on a wooden table surrounded by several other fruits and vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251, 270], "bbox": [0.601796875, 0.41498829039812646, 0.93409375, 0.7459953161592505], "iscrowd": 0, "area": 18113.687699999995, "rle": {"size": [427, 640], "counts": "liP52W=6J7I6J6J3M3M3M3M3N101N2O1O0O2O1O1N101O1N2O1N101O1N2O001N10000O1JQNPEo1o:TNnDm1Q;7O001O1O1O1O1O001O1O1O1O001O1O1O1O001O100O10O10O01N2N2N1O2N2O0O2N2N22M2IeM\\E]2`:hM]EY2`:;M4L3M4M2M4M2M4M2M4M2N3L3O2M2O1N100O100O100O100O101O0O1000000000000000000000001N10O100O101N10000O2O0O101O0O101N1N2N2N3M3M4L3M3M3M3M3N2M3M40O1O1OZMoER2e:2M2O1O1N2O1O1N2O1N01O2N2N2N2N2N2N2N2O0O2N001O0000001O0000001O0000001O000O101O0O100O2O0O10001N2O1N2O001N2O1N2N1N3M3M3M3M3N1M6K5K5J6K6IjWa0"}, "label": "a group of six oranges on a wooden table surrounded by several other fruits and vegetables"}]} {"id": 183435, "image": "COCO_train2014_000000183435.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"stack of oranges\"."}], "task": "refering", "answer_template": "The \"stack of oranges\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 204, 220, 221, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251, 270], "bbox": [0.601796875, 0.41498829039812646, 0.93409375, 0.7459953161592505], "iscrowd": 0, "area": 18113.687699999995, "rle": {"size": [427, 640], "counts": "liP52W=6J7I6J6J3M3M3M3M3N101N2O1O0O2O1O1N101O1N2O1N101O1N2O001N10000O1JQNPEo1o:TNnDm1Q;7O001O1O1O1O1O001O1O1O1O001O1O1O1O001O100O10O10O01N2N2N1O2N2O0O2N2N22M2IeM\\E]2`:hM]EY2`:;M4L3M4M2M4M2M4M2M4M2N3L3O2M2O1N100O100O100O100O101O0O1000000000000000000000001N10O100O101N10000O2O0O101O0O101N1N2N2N3M3M4L3M3M3M3M3N2M3M40O1O1OZMoER2e:2M2O1O1N2O1O1N2O1N01O2N2N2N2N2N2N2N2O0O2N001O0000001O0000001O0000001O000O101O0O100O2O0O10001N2O1N2O001N2O1N2N1N3M3M3M3M3N1M6K5K5J6K6IjWa0"}, "label": "stack of oranges"}]} {"id": 183435, "image": "COCO_train2014_000000183435.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bag of vegetables behind all of the other fruits and vegetables\"."}], "task": "refering", "answer_template": "The \"the bag of vegetables behind all of the other fruits and vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 35, 36, 37, 38, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 80, 81, 82, 83, 84], "bbox": [0.3773125, 0.09679156908665106, 0.7819062499999999, 0.2636065573770492], "iscrowd": 0, "area": 9483.706400000003, "rle": {"size": [427, 640], "counts": "oaT31Z=0O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O1000000000001O0O2O00001O001O00001O001O001O00001O001O00001O0O2O001O001O1O1O1O001O1O1O1O1O001O1O1O1O001O6I;F2N01O0O1000000000001O000000000000001O000000000000001O000000000000001O0000000000000O2O000000000000001O00000O100O10001N100O100O100O101N100O10001N100O100O2O0O100O101O0O100O2O0O100O101N100O10001N100O100O2O0O100O101N10000O101N100O100O2O0O10001O000000001O000O10001O000000001O000000001O000000001O000000001O0000001N100000001O000000001O000000nYj1"}, "label": "the bag of vegetables behind all of the other fruits and vegetables"}]} {"id": 183435, "image": "COCO_train2014_000000183435.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the farthest away greens\"."}], "task": "refering", "answer_template": "The \"the farthest away greens\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 34, 35, 36, 37, 38, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 80, 81, 82, 83, 84], "bbox": [0.3773125, 0.09679156908665106, 0.7819062499999999, 0.2636065573770492], "iscrowd": 0, "area": 9483.706400000003, "rle": {"size": [427, 640], "counts": "oaT31Z=0O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O1000000000001O0O2O00001O001O00001O001O001O00001O001O00001O0O2O001O001O1O1O1O001O1O1O1O1O001O1O1O1O001O6I;F2N01O0O1000000000001O000000000000001O000000000000001O000000000000001O0000000000000O2O000000000000001O00000O100O10001N100O100O100O101N100O10001N100O100O2O0O100O101O0O100O2O0O100O101N100O10001N100O100O2O0O100O101N10000O101N100O100O2O0O10001O000000001O000O10001O000000001O000000001O000000001O000000001O0000001N100000001O000000001O000000nYj1"}, "label": "the farthest away greens"}]} {"id": 11426, "image": "COCO_train2014_000000011426.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue van\"."}], "task": "refering", "answer_template": "The \"the blue van\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 58, 59, 60, 61, 62, 63, 64, 65, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 107, 109, 110, 111, 125, 126, 127, 128], "bbox": [0.45168749999999996, 0.10074999999999999, 0.8561875000000001, 0.32545833333333335], "iscrowd": 0, "area": 16978.403949999996, "rle": {"size": [480, 640], "counts": "ZcW44e>8G8I7H8N2N3L3N2N2M4N10000O2O000O101O0O10001N102N1N2O2N1N2O000O10O10O1000O010000O10O10O1O1O10O01O100O1O010O1O100O00100O1O10O01O100O1O00100O1O010O1O100O00100O1O10O01O100000O010000000O100000O10O10000SOjCPOVB?A?A?A`0@hoZ1"}, "label": "the blue van"}]} {"id": 11426, "image": "COCO_train2014_000000011426.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue van\"."}], "task": "refering", "answer_template": "The \"blue van\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 58, 59, 60, 61, 62, 63, 64, 65, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 107, 109, 110, 111, 125, 126, 127, 128], "bbox": [0.45168749999999996, 0.10074999999999999, 0.8561875000000001, 0.32545833333333335], "iscrowd": 0, "area": 16978.403949999996, "rle": {"size": [480, 640], "counts": "ZcW44e>8G8I7H8N2N3L3N2N2M4N10000O2O000O101O0O10001N102N1N2O2N1N2O000O10O10O1000O010000O10O10O1O1O10O01O100O1O010O1O100O00100O1O10O01O100O1O00100O1O010O1O100O00100O1O10O01O100000O010000000O100000O10O10000SOjCPOVB?A?A?A`0@hoZ1"}, "label": "blue van"}]} {"id": 11426, "image": "COCO_train2014_000000011426.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black motorcycle on green grass near a sign that reads 141 parked near other vehicles\"."}], "task": "refering", "answer_template": "The \"a black motorcycle on green grass near a sign that reads 141 parked near other vehicles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 106, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 295, 301, 302, 303, 304, 305, 306, 307, 310, 311, 312, 325, 326, 327, 328, 331, 332, 333, 334], "bbox": [0.059125000000000004, 0.1790625, 0.9307500000000001, 0.8637291666666667], "iscrowd": 0, "area": 113728.42169999999, "rle": {"size": [480, 640], "counts": "PSb0Q1k=j0UO7I7J6I7J6J6L4M3M3L3N3M3M3L4M3VORL]FQ4^9SL_Fo3^9TLaFn3[9VLcFl3\\9TLcFn3[9SLcFo3\\9SLbFo3]9QLbFQ4\\9QLbFQ4\\9PLbFR4]9oKbFS4]9nKaFT4]9mKaFU4^9h0O1N2N2O1N2N101N2N2[OSJUHn5j7WJPHj5P8ZJkGh5T8[JiGf5V8b00100O001O1O000010O00001O001O0010O0000000000000O2O000000000O10000000000O101O0000000O10000000000O10001O00000O100000O10E;I7N2O1O1O100O1O1O1O010O100000000000000000O1000000000000000000000001O0O100000001O000000001O0000000O101O0O100O2O000O2O0O2O0O101N10001N100O2O0O2O0O1O2O0O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1O1O1O1O1N2O1O1O1O1O1N2O10000000000000000001O00000000001O00000000001O000fHlIh5T6XJmIg5S6XJnIh5R6XJnIh5S6WJmIi5S6VJnIj5R6VJnIj5R6VJnIj5R6VJnIj5R6UJoIk5Q6UJoIj5R6VJnIj5R6UJoIk5Q6UJPJj5P6VJPJj5P6UJQJk5o5UJQJk5o5UJQJk5o5TJRJl5n5TJRJl5n5TJRJl5n5SJSJm5m5SJSJm5m5SJSJm5m5RJUJm5k5SJUJl5l5TJTJl5l5SJUJm5k5SJUJm5k5SJUJm5k5SJUJl5l5SJUJm5k5SJUJm5k5SJUJm5k5RJVJn5j5RJVJm5k5SJVJl5k5RJUJo5l5PJTJP6m5oISJQ6o5lIQJT6Q6kIoIU6R6jInIV6S6hImIY6T6fIlIZ6U6eIkIZ6X6dIgI]6\\6_IeIa6]6]IcIc6`6ZI_Ig6c6VI^Ij6f6RIYIn6]7000O1000O010000O10O10O10000O010O10000O10O10O10000O01000O1000O0100O1000O01fNiHkJW7U5lHhJT7W5QIdJP7\\5RIbJm6_5UI_Jk6`5XI]Ji6c5YI[Jg6e5[IXJf6g5]IUJe6k5]IRJc6o5_InIb6Q6aIkIa6U6`IiIa6W6aIfI`6Y6S1100000O10O10O10000O01000OWOkIiHU6V7lIjHT6T7nIlHR6R7PJmHP6S7QJmHo5Q7SJoHm5o6UJQIj5n6XJRIh5n6XJQIi5o6WJQIh5P7YJoHg5Q7YJoHg5Q7YJoHf5R7ZJmHg5S7YJmHg5S7YJmHf5T7ZJlHf5S7[JmHe5S7[JlHf5T7ZJlHe5U7[JkHe5U7[JkHe5U7\\JjHc5W7]JhHd5X7\\JhHd5X7\\JhHc5Y7]JgHc5Y7]JgHc5X7^JhHa5Y7_JfHb5Z7^JfHb5Z7_JeHa5Z7`JeHh4LRK_77[Hn4 in this image.", "objects": [{"patches": [83, 84, 85, 86, 106, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 295, 301, 302, 303, 304, 305, 306, 307, 310, 311, 312, 325, 326, 327, 328, 331, 332, 333, 334], "bbox": [0.059125000000000004, 0.1790625, 0.9307500000000001, 0.8637291666666667], "iscrowd": 0, "area": 113728.42169999999, "rle": {"size": [480, 640], "counts": "PSb0Q1k=j0UO7I7J6I7J6J6L4M3M3L3N3M3M3L4M3VORL]FQ4^9SL_Fo3^9TLaFn3[9VLcFl3\\9TLcFn3[9SLcFo3\\9SLbFo3]9QLbFQ4\\9QLbFQ4\\9PLbFR4]9oKbFS4]9nKaFT4]9mKaFU4^9h0O1N2N2O1N2N101N2N2[OSJUHn5j7WJPHj5P8ZJkGh5T8[JiGf5V8b00100O001O1O000010O00001O001O0010O0000000000000O2O000000000O10000000000O101O0000000O10000000000O10001O00000O100000O10E;I7N2O1O1O100O1O1O1O010O100000000000000000O1000000000000000000000001O0O100000001O000000001O0000000O101O0O100O2O000O2O0O2O0O101N10001N100O2O0O2O0O1O2O0O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1O1O1O1O1N2O1O1O1O1O1N2O10000000000000000001O00000000001O00000000001O000fHlIh5T6XJmIg5S6XJnIh5R6XJnIh5S6WJmIi5S6VJnIj5R6VJnIj5R6VJnIj5R6VJnIj5R6UJoIk5Q6UJoIj5R6VJnIj5R6UJoIk5Q6UJPJj5P6VJPJj5P6UJQJk5o5UJQJk5o5UJQJk5o5TJRJl5n5TJRJl5n5TJRJl5n5SJSJm5m5SJSJm5m5SJSJm5m5RJUJm5k5SJUJl5l5TJTJl5l5SJUJm5k5SJUJm5k5SJUJm5k5SJUJl5l5SJUJm5k5SJUJm5k5SJUJm5k5RJVJn5j5RJVJm5k5SJVJl5k5RJUJo5l5PJTJP6m5oISJQ6o5lIQJT6Q6kIoIU6R6jInIV6S6hImIY6T6fIlIZ6U6eIkIZ6X6dIgI]6\\6_IeIa6]6]IcIc6`6ZI_Ig6c6VI^Ij6f6RIYIn6]7000O1000O010000O10O10O10000O010O10000O10O10O10000O01000O1000O0100O1000O01fNiHkJW7U5lHhJT7W5QIdJP7\\5RIbJm6_5UI_Jk6`5XI]Ji6c5YI[Jg6e5[IXJf6g5]IUJe6k5]IRJc6o5_InIb6Q6aIkIa6U6`IiIa6W6aIfI`6Y6S1100000O10O10O10000O01000OWOkIiHU6V7lIjHT6T7nIlHR6R7PJmHP6S7QJmHo5Q7SJoHm5o6UJQIj5n6XJRIh5n6XJQIi5o6WJQIh5P7YJoHg5Q7YJoHg5Q7YJoHf5R7ZJmHg5S7YJmHg5S7YJmHf5T7ZJlHf5S7[JmHe5S7[JlHf5T7ZJlHe5U7[JkHe5U7[JkHe5U7\\JjHc5W7]JhHd5X7\\JhHd5X7\\JhHc5Y7]JgHc5Y7]JgHc5X7^JhHa5Y7_JfHb5Z7^JfHb5Z7_JeHa5Z7`JeHh4LRK_77[Hn4 in this image.", "objects": [{"patches": [0, 23, 24, 46, 47, 69, 70, 71, 92, 93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 207, 208, 230, 231, 232, 253, 254, 255], "bbox": [0.0017968749999999999, 0.0471875, 0.172578125, 0.6606835937500001], "iscrowd": 0, "area": 17124.700099999995, "rle": {"size": [512, 640], "counts": "]a02W1o3[9YLaFj3Y9\\LcFh3V9`LfFc3T9dLgF`3T9eLiF^3T9dLiF_3T9eLiF^3S9fLiF^3T9dLjF_3R9eLjF^3S9eLkF^3Q9l1M3M3M3N2M3M2N3^IcH[4`7[KhHd4[7QKmHo4V7gJQIY5R7\\JVId5m6RJZIn5\\8001N2O2M2M3M3M3M3M6J5L5Jo0QO5K5K4L3M3M2N3MgGVI`7h6^H^Ia7`6]HeIa7b6UHaIl7e6jG^IV8V71O001kMlG\\LT8^3UH_Ll7Z3]HcLc7V3fHhL[7R3mHkLS7Q3TIlLm6n2[IoLf6k2bIQM`6GcGo1T2XNY6KjG`1T2bNS6OPHQ1U2mNl53UHc0W2VOf58ZH4X2A^5 in this image.", "objects": [{"patches": [0, 23, 24, 46, 47, 69, 70, 71, 92, 93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 207, 208, 230, 231, 232, 253, 254, 255], "bbox": [0.0017968749999999999, 0.0471875, 0.172578125, 0.6606835937500001], "iscrowd": 0, "area": 17124.700099999995, "rle": {"size": [512, 640], "counts": "]a02W1o3[9YLaFj3Y9\\LcFh3V9`LfFc3T9dLgF`3T9eLiF^3T9dLiF_3T9eLiF^3S9fLiF^3T9dLjF_3R9eLjF^3S9eLkF^3Q9l1M3M3M3N2M3M2N3^IcH[4`7[KhHd4[7QKmHo4V7gJQIY5R7\\JVId5m6RJZIn5\\8001N2O2M2M3M3M3M3M6J5L5Jo0QO5K5K4L3M3M2N3MgGVI`7h6^H^Ia7`6]HeIa7b6UHaIl7e6jG^IV8V71O001kMlG\\LT8^3UH_Ll7Z3]HcLc7V3fHhL[7R3mHkLS7Q3TIlLm6n2[IoLf6k2bIQM`6GcGo1T2XNY6KjG`1T2bNS6OPHQ1U2mNl53UHc0W2VOf58ZH4X2A^5 in this image.", "objects": [{"patches": [106, 128, 129, 130, 131, 151, 152, 153, 154, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 240, 241, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 312, 313, 314, 315, 316, 335, 336, 337, 338, 339, 340, 361, 362], "bbox": [0.44751562500000003, 0.26080078125, 0.7913593750000001, 0.8741992187500001], "iscrowd": 0, "area": 34666.14625, "rle": {"size": [512, 640], "counts": "_i_43j?4M2N3M3M3N2N101O1O1O1OO100O100O100O10O01O100O100O100O100000000001O00000000001O2N2N3M2N9F:G4L1O1N2M3N2N2N2N2N2M3N2N2N2N2N2M3M3M3M3M3M3M3M3M3M3L4M3L4M4K4M3L4M3L4M3L4M3J6XNSK]G in this image.", "objects": [{"patches": [106, 128, 129, 130, 131, 151, 152, 153, 154, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 219, 220, 221, 222, 223, 224, 240, 241, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 312, 313, 314, 315, 316, 335, 336, 337, 338, 339, 340, 361, 362], "bbox": [0.44751562500000003, 0.26080078125, 0.7913593750000001, 0.8741992187500001], "iscrowd": 0, "area": 34666.14625, "rle": {"size": [512, 640], "counts": "_i_43j?4M2N3M3M3N2N101O1O1O1OO100O100O100O10O01O100O100O100O100000000001O00000000001O2N2N3M2N9F:G4L1O1N2M3N2N2N2N2N2M3N2N2N2N2N2M3M3M3M3M3M3M3M3M3M3L4M3L4M4K4M3L4M3L4M3L4M3J6XNSK]G in this image.", "objects": [{"patches": [139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 210, 211, 212, 213, 234, 235, 236, 258, 259, 282], "bbox": [0.02890625, 0.4821111111111111, 0.28907812499999996, 0.9531944444444443], "iscrowd": 0, "area": 11720.612400000002, "rle": {"size": [360, 640], "counts": "Zl61W;000O100O100N2O1N2O1N2O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O2N100O101N100O2O001O1O001O1O1N2O1O2N1O1O1O1O1O1O1O1O00101N1O2N2N3N2M2N001O100O10O01O10O0100O010O1O10O0100O1O001O1O001O001O1O001O001O1O001O2N1O1O1O0010O01O0010O01O1O100O1O100O1O100O100O100O010O1O100O100O2O000O10O100O10000O10O001O1O00100O1O1O1O@YHeMa7[2aHiMY7W2iHjMT7V2oHiMo6W2SIjMj6U2YIkMf6T2\\IlMb6T2aIlM\\6T2fIlMZ6R2gIlM\\6Q2fImM\\6R2hIhM\\6V2iIbM[6]2_1Ic1\\N[fo4"}, "label": "a bird standing on the cage"}]} {"id": 445628, "image": "COCO_train2014_000000445628.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small bird hanging on the net\"."}], "task": "refering", "answer_template": "The \"a small bird hanging on the net\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 210, 211, 212, 213, 234, 235, 236, 258, 259, 282], "bbox": [0.02890625, 0.4821111111111111, 0.28907812499999996, 0.9531944444444443], "iscrowd": 0, "area": 11720.612400000002, "rle": {"size": [360, 640], "counts": "Zl61W;000O100O100N2O1N2O1N2O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N2O1O2N100O101N100O2O001O1O001O1O1N2O1O2N1O1O1O1O1O1O1O1O00101N1O2N2N3N2M2N001O100O10O01O10O0100O010O1O10O0100O1O001O1O001O001O1O001O001O1O001O2N1O1O1O0010O01O0010O01O1O100O1O100O1O100O100O100O010O1O100O100O2O000O10O100O10000O10O001O1O00100O1O1O1O@YHeMa7[2aHiMY7W2iHjMT7V2oHiMo6W2SIjMj6U2YIkMf6T2\\IlMb6T2aIlM\\6T2fIlMZ6R2gIlM\\6Q2fImM\\6R2hIhM\\6V2iIbM[6]2_1Ic1\\N[fo4"}, "label": "a small bird hanging on the net"}]} {"id": 445628, "image": "COCO_train2014_000000445628.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black bird sitting on top of a piece of wood looking off the screen\"."}], "task": "refering", "answer_template": "The \"black bird sitting on top of a piece of wood looking off the screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 203, 204, 226], "bbox": [0.828640625, 0.40391666666666665, 0.98475, 0.7246111111111111], "iscrowd": 0, "area": 6164.6362500000005, "rle": {"size": [360, 640], "counts": "V`j52T;k0VOj0VOi0XO2M100O10O0100O010O10OO2N2O0O2O0O2O10O01O00010O000010O01O00010O01O010O001O1O001O001O0010O01O002N1O1O1O1N3N1N2O1N2M3M4L3M3M3G9K5N3N1N2N3N1O1N3N1O1O2N101N1O1O2O1O001N2O1O1N101O1N2O1O1O0O2O1O1N2OX[3"}, "label": "black bird sitting on top of a piece of wood looking off the screen"}]} {"id": 445628, "image": "COCO_train2014_000000445628.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bird on the right is so dark it looks like a shadow\"."}], "task": "refering", "answer_template": "The \"the bird on the right is so dark it looks like a shadow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 203, 204, 226], "bbox": [0.828640625, 0.40391666666666665, 0.98475, 0.7246111111111111], "iscrowd": 0, "area": 6164.6362500000005, "rle": {"size": [360, 640], "counts": "V`j52T;k0VOj0VOi0XO2M100O10O0100O010O10OO2N2O0O2O0O2O10O01O00010O000010O01O00010O01O010O001O1O001O001O0010O01O002N1O1O1O1N3N1N2O1N2M3M4L3M3M3G9K5N3N1N2N3N1O1N3N1O1O2N101N1O1O2O1O001N2O1O1N101O1N2O1O1O0O2O1O1N2OX[3"}, "label": "the bird on the right is so dark it looks like a shadow"}]} {"id": 273600, "image": "COCO_train2014_000000273600.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the palm on which the device is been displaying\"."}], "task": "refering", "answer_template": "The \"the palm on which the device is been displaying\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 29, 30, 47, 48, 65, 66, 83, 84, 85, 93, 94, 95, 96, 101, 102, 103, 109, 110, 111, 112, 113, 114, 120, 121, 126, 127, 128, 129, 130, 131, 132, 138, 139, 144, 145, 146, 147, 148, 149, 150, 156, 162, 163, 164, 165, 166, 167, 168, 174, 180, 181, 182, 183, 184, 185, 186, 192, 198, 199, 200, 201, 202, 203, 204, 210, 211, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230], "bbox": [0.0016799999999999999, 0.053919999999999996, 0.7837, 0.9887733333333333], "iscrowd": 0, "area": 42610.61390000001, "rle": {"size": [375, 500], "counts": "kb06`;:[Fc1]5fNZJ[1f5kNSJV1l5kNSJV1l5kNSJV1l5kNSJV1m5jNRJW1m5jNRJW1m5kNRJU1m5lNRJU1n5kNQJV1n5kNQJV1n5kNQJV1n5kNQJV1n5kNQJV1o5jNPJW1o5jNPJW1o5jNQJV1n5lNPJU1P6kNoIV1P6kNoIV1P6kNoIV1P6kNoIV1Q6jNnIW1Q6jNnIW1Q6jNnIW1Q6jNoIV1Q6jNnIW1Q6kNmIV1R6kNmIV1R6kNmIV1S6jNlIW1S6jNlIW1S6jNlIW1S6jNlIW1T6iNkIX1T6iNlIW1S6jNlIW1S6jNlIW1T6jNjIW1U6jNjIW1U6jNjIW1V6iNiIX1V6iNiIX1W6hNhIY1W6hNiIX1W6T2O100O100O100O100O100O10000O10000O10000O100000001N100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O100O10000000000000000O100000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O100000000000000002Ne0[Oe0[Of0ZOU1kN\\1dNP1PO7I4L1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000O10000O1000000O1000000O10001N10000dE^Oc9b0^F^O`9c0`F]O`9c0`F^O_9b0aF_O^9a0cF_O[9b0eF_OZ9a0fF_OZ9a0fF@Y9`0hF@V9a0jF@U9`0kFAT9?lFBR9?nFAR9?nFGl89TG9Z8GfGj0h7WOXH[1V7eNjHm1d6SN\\Ik2f5UMZJi3g4XLYKf4i3ZKWLd5j2]JWM`6j1aIXN[7j0eHXOS1YO^4o0`JIo0BX4f0iJJk0BZ4e0lJJf0D]4c0mJKb0D`4a0oJMSK10Dk4=UK1LDm4 in this image.", "objects": [{"patches": [11, 29, 30, 47, 48, 65, 66, 83, 84, 85, 93, 94, 95, 96, 101, 102, 103, 109, 110, 111, 112, 113, 114, 120, 121, 126, 127, 128, 129, 130, 131, 132, 138, 139, 144, 145, 146, 147, 148, 149, 150, 156, 162, 163, 164, 165, 166, 167, 168, 174, 180, 181, 182, 183, 184, 185, 186, 192, 198, 199, 200, 201, 202, 203, 204, 210, 211, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230], "bbox": [0.0016799999999999999, 0.053919999999999996, 0.7837, 0.9887733333333333], "iscrowd": 0, "area": 42610.61390000001, "rle": {"size": [375, 500], "counts": "kb06`;:[Fc1]5fNZJ[1f5kNSJV1l5kNSJV1l5kNSJV1l5kNSJV1m5jNRJW1m5jNRJW1m5kNRJU1m5lNRJU1n5kNQJV1n5kNQJV1n5kNQJV1n5kNQJV1n5kNQJV1o5jNPJW1o5jNPJW1o5jNQJV1n5lNPJU1P6kNoIV1P6kNoIV1P6kNoIV1P6kNoIV1Q6jNnIW1Q6jNnIW1Q6jNnIW1Q6jNoIV1Q6jNnIW1Q6kNmIV1R6kNmIV1R6kNmIV1S6jNlIW1S6jNlIW1S6jNlIW1S6jNlIW1T6iNkIX1T6iNlIW1S6jNlIW1S6jNlIW1T6jNjIW1U6jNjIW1U6jNjIW1V6iNiIX1V6iNiIX1W6hNhIY1W6hNiIX1W6T2O100O100O100O100O100O10000O10000O10000O100000001N100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O100O10000000000000000O100000000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O100000000000000002Ne0[Oe0[Of0ZOU1kN\\1dNP1PO7I4L1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000O10000O1000000O1000000O10001N10000dE^Oc9b0^F^O`9c0`F]O`9c0`F^O_9b0aF_O^9a0cF_O[9b0eF_OZ9a0fF_OZ9a0fF@Y9`0hF@V9a0jF@U9`0kFAT9?lFBR9?nFAR9?nFGl89TG9Z8GfGj0h7WOXH[1V7eNjHm1d6SN\\Ik2f5UMZJi3g4XLYKf4i3ZKWLd5j2]JWM`6j1aIXN[7j0eHXOS1YO^4o0`JIo0BX4f0iJJk0BZ4e0lJJf0D]4c0mJKb0D`4a0oJMSK10Dk4=UK1LDm4 in this image.", "objects": [{"patches": [22, 23, 24, 25, 26, 36, 37, 38, 39, 40, 41, 42, 43, 50, 51, 52, 53, 54, 55, 56, 57, 58, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 127, 128, 129, 130, 131, 132, 133, 134, 142, 143, 144, 145, 146, 147, 148, 149, 158, 159, 160, 161, 162, 163, 164, 173, 174, 175, 176, 177, 178, 179, 188, 189, 190, 191, 192, 193, 194, 202, 203, 204, 205, 206, 207, 208, 209, 217, 218, 219, 220, 221, 222, 223, 224, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 291, 292, 293, 294, 295, 296, 297, 298, 299, 303, 304, 305, 306, 309, 310, 312, 313, 314, 318, 319, 320, 321, 329, 334, 335, 336], "bbox": [0.22533957845433256, 0.053421874999999994, 1.0, 0.9736249999999999], "iscrowd": 0, "area": 119489.39369999999, "rle": {"size": [640, 427], "counts": "[bl16_c0=M4XLZOoCi0l;BjCb0QWLfAi3d>mK_Ao3d>nKcAl3[>VLkAd3T>]LRB\\3m=fLXBU3f=mL\\BQ3b=RM\\Bn2d=RM\\Bo2b=RM^Bo2`=RM`Bn2`=RM`Bo2^=RMbBo2\\=RMeBm2Z=TMfBm2Y=SMgBn2W=SMiBn2U=SMkBn2S=SMmBo2Q=QMoBP3o in this image.", "objects": [{"patches": [22, 23, 24, 25, 26, 36, 37, 38, 39, 40, 41, 42, 43, 50, 51, 52, 53, 54, 55, 56, 57, 58, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 127, 128, 129, 130, 131, 132, 133, 134, 142, 143, 144, 145, 146, 147, 148, 149, 158, 159, 160, 161, 162, 163, 164, 173, 174, 175, 176, 177, 178, 179, 188, 189, 190, 191, 192, 193, 194, 202, 203, 204, 205, 206, 207, 208, 209, 217, 218, 219, 220, 221, 222, 223, 224, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 291, 292, 293, 294, 295, 296, 297, 298, 299, 303, 304, 305, 306, 309, 310, 312, 313, 314, 318, 319, 320, 321, 329, 334, 335, 336], "bbox": [0.22533957845433256, 0.053421874999999994, 1.0, 0.9736249999999999], "iscrowd": 0, "area": 119489.39369999999, "rle": {"size": [640, 427], "counts": "[bl16_c0=M4XLZOoCi0l;BjCb0QWLfAi3d>mK_Ao3d>nKcAl3[>VLkAd3T>]LRB\\3m=fLXBU3f=mL\\BQ3b=RM\\Bn2d=RM\\Bo2b=RM^Bo2`=RM`Bn2`=RM`Bo2^=RMbBo2\\=RMeBm2Z=TMfBm2Y=SMgBn2W=SMiBn2U=SMkBn2S=SMmBo2Q=QMoBP3o in this image.", "objects": [{"patches": [277, 278, 281, 282, 292, 293, 294, 296, 297, 306, 307, 308, 309, 310, 311, 312, 313, 321, 322, 323, 324, 325, 326, 327, 328, 329, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.4172599531615925, 0.7805156249999999, 1.0, 1.0], "iscrowd": 0, "area": 19801.213550000004, "rle": {"size": [640, 427], "counts": "mk_33dc09G9H8G9G9G9K5N2M3N2M3N2M3N1N3N2N200L4L4K5O11O001O00001O001O00001O001O00001O001O00001N10iN`Mo@_2R?YNU@g1l?V11O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1N2O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O0O2O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O00O1M3N2N1O2N2M3N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2M3N2N2N2N2N2M3N2N2N2N1N3N2N2O1000000001O0000001O000000001O4L?A?A?A?A;E0000001N10000000001O00000000001O00000000001O00000000001O0000000O101O00000000001O00000000001O00000000001O00000000001O001N6K5K5K5K5K5K5K5K[\\O"}, "label": "an open book resting on the ground"}]} {"id": 101573, "image": "COCO_train2014_000000101573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an open book sitting on the ground in front of a young lady wearing sandals and blue jeans\"."}], "task": "refering", "answer_template": "The \"an open book sitting on the ground in front of a young lady wearing sandals and blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [277, 278, 281, 282, 292, 293, 294, 296, 297, 306, 307, 308, 309, 310, 311, 312, 313, 321, 322, 323, 324, 325, 326, 327, 328, 329, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.4172599531615925, 0.7805156249999999, 1.0, 1.0], "iscrowd": 0, "area": 19801.213550000004, "rle": {"size": [640, 427], "counts": "mk_33dc09G9H8G9G9G9K5N2M3N2M3N2M3N1N3N2N200L4L4K5O11O001O00001O001O00001O001O00001O001O00001N10iN`Mo@_2R?YNU@g1l?V11O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1N2O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O0O2O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O00O1M3N2N1O2N2M3N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2M3N2N2N2N2N2M3N2N2N2N1N3N2N2O1000000001O0000001O000000001O4L?A?A?A?A;E0000001N10000000001O00000000001O00000000001O00000000001O0000000O101O00000000001O00000000001O00000000001O00000000001O001N6K5K5K5K5K5K5K5K[\\O"}, "label": "an open book sitting on the ground in front of a young lady wearing sandals and blue jeans"}]} {"id": 101573, "image": "COCO_train2014_000000101573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl laying on the ground\"."}], "task": "refering", "answer_template": "The \"girl laying on the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 136, 137, 138, 139, 140, 141, 151, 152, 153, 154, 155, 156, 166, 167, 168, 169, 170, 171, 172, 173, 180, 181, 182, 183, 184, 185, 186, 187, 188, 195, 196, 197, 198, 199, 200, 201, 202, 203, 210, 211, 212, 213, 214, 215, 216, 225, 226, 227, 228, 240, 241, 242, 243, 255, 256, 257, 258], "bbox": [0.0, 0.34575, 0.5908665105386417, 0.7590468749999999], "iscrowd": 0, "area": 36358.917799999996, "rle": {"size": [640, 427], "counts": "_=T1lb00000OG>^Oa0J7I6J3M4K4K6K4K6K4L4K6K4K6L3N2M4M2M3N2M3N2N2M3N2M3O1O1N2O1OlNYAQMe>Q3\\AnLd>b2ZAjL3d0b>c2[AiL3d0a>d2\\AhL3d0a>d2\\AgL4e0]=YOVC[3YOgL4e0[=^OTCW3]OcL7h0U=BVCR3@_L9l0P=EWCP3@[L=P1jiCV1Z1\\Nk:`0kCV1X1ZNl:b0jCW1W1XNn:c0jCW1V1VNo:e0jCW1U1TNP;g0jCW1T1RNQ;k0gCV1U1PNS;m0fCU1U1nMT;X4jDhKT;\\4jDdKU;^4jDbKU;a4hD`KV;d4hD\\KX;f4fDZKZ;h4dDXK[;k4cDUK];m4`DTK_;o4_DQKa;P5^DPKa;Q5_DoJa;Q5_DoJ`;R5_DoJa;Q5_DoJ`;R5`DnJ`;R5`DnJ_;S5aDmJ_;S5`DnJ_;S5aDmJ^;S5cDmJ];S5cDmJ\\;T5cDmJ\\;T5dDlJ\\;S5eDmJZ;T5fDlJZ;T5fDlJZ;T5eDmJZ;S5gDmJY;S5gDmJY;S5gDmJY;S5gDmJY;S5fDnJZ;Q5gDoJX;R5hDnJX;R5hDnJX;Q5iDoJW;Q5hDPKX;P5hDoJY;P5hDPKW;Q5iDoJW;Q5iDoI40S;P6iDnI[ in this image.", "objects": [{"patches": [122, 123, 124, 125, 136, 137, 138, 139, 140, 141, 151, 152, 153, 154, 155, 156, 166, 167, 168, 169, 170, 171, 172, 173, 180, 181, 182, 183, 184, 185, 186, 187, 188, 195, 196, 197, 198, 199, 200, 201, 202, 203, 210, 211, 212, 213, 214, 215, 216, 225, 226, 227, 228, 240, 241, 242, 243, 255, 256, 257, 258], "bbox": [0.0, 0.34575, 0.5908665105386417, 0.7590468749999999], "iscrowd": 0, "area": 36358.917799999996, "rle": {"size": [640, 427], "counts": "_=T1lb00000OG>^Oa0J7I6J3M4K4K6K4K6K4L4K6K4K6L3N2M4M2M3N2M3N2N2M3N2M3O1O1N2O1OlNYAQMe>Q3\\AnLd>b2ZAjL3d0b>c2[AiL3d0a>d2\\AhL3d0a>d2\\AgL4e0]=YOVC[3YOgL4e0[=^OTCW3]OcL7h0U=BVCR3@_L9l0P=EWCP3@[L=P1jiCV1Z1\\Nk:`0kCV1X1ZNl:b0jCW1W1XNn:c0jCW1V1VNo:e0jCW1U1TNP;g0jCW1T1RNQ;k0gCV1U1PNS;m0fCU1U1nMT;X4jDhKT;\\4jDdKU;^4jDbKU;a4hD`KV;d4hD\\KX;f4fDZKZ;h4dDXK[;k4cDUK];m4`DTK_;o4_DQKa;P5^DPKa;Q5_DoJa;Q5_DoJ`;R5_DoJa;Q5_DoJ`;R5`DnJ`;R5`DnJ_;S5aDmJ_;S5`DnJ_;S5aDmJ^;S5cDmJ];S5cDmJ\\;T5cDmJ\\;T5dDlJ\\;S5eDmJZ;T5fDlJZ;T5fDlJZ;T5eDmJZ;S5gDmJY;S5gDmJY;S5gDmJY;S5gDmJY;S5fDnJZ;Q5gDoJX;R5hDnJX;R5hDnJX;Q5iDoJW;Q5hDPKX;P5hDoJY;P5hDPKW;Q5iDoJW;Q5iDoI40S;P6iDnI[ in this image.", "objects": [{"patches": [215, 216, 217, 218, 226, 227, 228, 229, 230, 241], "bbox": [0.0807728337236534, 0.6097812499999999, 0.6057377049180327, 0.7102187499999999], "iscrowd": 0, "area": 2309.4817, "rle": {"size": [640, 427], "counts": "Vfe01ic0510O10O10@KT]O6lb0JT]O6lb0JT]O5mb0KS]O5mb0KR]O6nb0JR]O6nb0JR]O5ob0KP]O6ob0KQ]O5ob0KQ]O5ob0KP]O5Qc0Ko\\O5Qc0Ko\\O5Qc0Kn\\O5Sc0Km\\O5Sc0Km\\O5Rc0Lm\\O5Sc0Km\\O4Tc0Ll\\O4Tc0Lk\\O5Uc0Kk\\O5Uc0Kk\\O4Vc0Li\\O5Wc0Ki\\O5Vc0Lj\\O4Vc0Lj\\O3Wc0Mh\\O4Xc0Lh\\O4Xc0Lh\\O3Yc0Mf\\O4Zc0Lf\\O4Zc0Lf\\O4Yc0Mf\\O3[c0Me\\O3[c0Me\\O3[c0Md\\O4\\c0Ld\\O3]c0Mc\\O3]c0Mb\\O4^c0Lb\\O4]c0Mc\\O2^c0Na\\O3_c0Ma\\O3_c0Ma\\O2`c0700000O0100000O1000000O01000O1000000O10O1000O1000000O010000000001N10001O00O1000O0100O10000O10O0100O10O10O010O10O010O10O010O10O010O10O10O010O010O10O10O010O01000O010O010O10O10O010O010O10O10O010O01000O010O010O01000O010O010O01000O010O0100O010000001O0000000000000O10001O0000000000000000001O0000000000000000001O0000000000000000001N1000000000000000001N1OeWY3"}, "label": "notebook with papaer"}]} {"id": 101573, "image": "COCO_train2014_000000101573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a notebook being held by one of the women\"."}], "task": "refering", "answer_template": "The \"a notebook being held by one of the women\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 217, 218, 226, 227, 228, 229, 230, 241], "bbox": [0.0807728337236534, 0.6097812499999999, 0.6057377049180327, 0.7102187499999999], "iscrowd": 0, "area": 2309.4817, "rle": {"size": [640, 427], "counts": "Vfe01ic0510O10O10@KT]O6lb0JT]O6lb0JT]O5mb0KS]O5mb0KR]O6nb0JR]O6nb0JR]O5ob0KP]O6ob0KQ]O5ob0KQ]O5ob0KP]O5Qc0Ko\\O5Qc0Ko\\O5Qc0Kn\\O5Sc0Km\\O5Sc0Km\\O5Rc0Lm\\O5Sc0Km\\O4Tc0Ll\\O4Tc0Lk\\O5Uc0Kk\\O5Uc0Kk\\O4Vc0Li\\O5Wc0Ki\\O5Vc0Lj\\O4Vc0Lj\\O3Wc0Mh\\O4Xc0Lh\\O4Xc0Lh\\O3Yc0Mf\\O4Zc0Lf\\O4Zc0Lf\\O4Yc0Mf\\O3[c0Me\\O3[c0Me\\O3[c0Md\\O4\\c0Ld\\O3]c0Mc\\O3]c0Mb\\O4^c0Lb\\O4]c0Mc\\O2^c0Na\\O3_c0Ma\\O3_c0Ma\\O2`c0700000O0100000O1000000O01000O1000000O10O1000O1000000O010000000001N10001O00O1000O0100O10000O10O0100O10O10O010O10O010O10O010O10O010O10O10O010O010O10O10O010O01000O010O010O10O10O010O010O10O10O010O01000O010O010O01000O010O010O01000O010O0100O010000001O0000000000000O10001O0000000000000000001O0000000000000000001O0000000000000000001N1000000000000000001N1OeWY3"}, "label": "a notebook being held by one of the women"}]} {"id": 257243, "image": "COCO_train2014_000000257243.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and blue and orange double dekker bus on the road\"."}], "task": "refering", "answer_template": "The \"a white and blue and orange double dekker bus on the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 145, 146, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 215, 216, 217, 218, 219, 220, 221, 222], "bbox": [0.321046875, 0.2561875, 0.69928125, 0.5784166666666667], "iscrowd": 0, "area": 31197.586549999996, "rle": {"size": [480, 640], "counts": "cZP32]>a0^Oc0^Oa0^Ob0_Oa0K6O0001O0001O0000001O00001N3N1O2N1O2N1N3N000000000O010000000O10O10000000O10O1001O000O1000001O000O1000001O000O1000001O000O100000001O0O100000001O0O100000001O0O100000001N10000000000O2O000000000O2O000000000O2O001O1O001N101O001O001N101O1O001OO010O10000O10O01000000000O2O000000001N100000001N10000000000O10000000000O10000000000O10000000000O10000000000O1000O100000O1000O10001O0000000O101O000000001O0O1000001O000000001O0000000O2O000000001O00000000001O0O1000001O000000001O000mMiF]NW9^1X2[Of0ZOe0[O_li2"}, "label": "a white and blue and orange double dekker bus on the road"}]} {"id": 257243, "image": "COCO_train2014_000000257243.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white blue and orange double decker bus\"."}], "task": "refering", "answer_template": "The \"a white blue and orange double decker bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 145, 146, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 215, 216, 217, 218, 219, 220, 221, 222], "bbox": [0.321046875, 0.2561875, 0.69928125, 0.5784166666666667], "iscrowd": 0, "area": 31197.586549999996, "rle": {"size": [480, 640], "counts": "cZP32]>a0^Oc0^Oa0^Ob0_Oa0K6O0001O0001O0000001O00001N3N1O2N1O2N1N3N000000000O010000000O10O10000000O10O1001O000O1000001O000O1000001O000O1000001O000O100000001O0O100000001O0O100000001O0O100000001N10000000000O2O000000000O2O000000000O2O001O1O001N101O001O001N101O1O001OO010O10000O10O01000000000O2O000000001N100000001N10000000000O10000000000O10000000000O10000000000O10000000000O1000O100000O1000O10001O0000000O101O000000001O0O1000001O000000001O0000000O2O000000001O00000000001O0O1000001O000000001O000mMiF]NW9^1X2[Of0ZOe0[O_li2"}, "label": "a white blue and orange double decker bus"}]} {"id": 421086, "image": "COCO_train2014_000000421086.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the slice of pizza that is being lifted off of the sheet on the right\"."}], "task": "refering", "answer_template": "The \"the slice of pizza that is being lifted off of the sheet on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 194, 195, 196, 197, 217, 218, 219, 220, 221, 242, 243, 244, 245, 266, 267, 268], "bbox": [0.4449375, 0.45841666666666664, 0.687640625, 0.6943750000000001], "iscrowd": 0, "area": 7264.798050000004, "rle": {"size": [480, 640], "counts": "bjU45e>7I7L4O1O1N2O1O1N2O1O0O2O1O0O2O1O0O2O1O0O2O001N2O001O10O01O10O01O1O010O1O00100O00100O001O010O1O00100O00100O001O10O01O1O010O1O010O1O1O010O1000O100000000O100000000O1000000O100000000O1000000O100000000O100000000O1000000O1O1O1O100O1O1O1O100O1O1O1O00100O1O1O1O1O10O01O1O1O100O1O001O1O100O1O1O1O00100O1O1O1O1L4J5K6J]^m2"}, "label": "the slice of pizza that is being lifted off of the sheet on the right"}]} {"id": 421086, "image": "COCO_train2014_000000421086.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza slice being held\"."}], "task": "refering", "answer_template": "The \"the pizza slice being held\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 194, 195, 196, 197, 217, 218, 219, 220, 221, 242, 243, 244, 245, 266, 267, 268], "bbox": [0.4449375, 0.45841666666666664, 0.687640625, 0.6943750000000001], "iscrowd": 0, "area": 7264.798050000004, "rle": {"size": [480, 640], "counts": "bjU45e>7I7L4O1O1N2O1O1N2O1O0O2O1O0O2O1O0O2O1O0O2O001N2O001O10O01O10O01O1O010O1O00100O00100O001O010O1O00100O00100O001O10O01O1O010O1O010O1O1O010O1000O100000000O100000000O1000000O100000000O1000000O100000000O100000000O1000000O1O1O1O100O1O1O1O100O1O1O1O00100O1O1O1O1O10O01O1O1O100O1O001O1O100O1O1O1O00100O1O1O1O1L4J5K6J]^m2"}, "label": "the pizza slice being held"}]} {"id": 421086, "image": "COCO_train2014_000000421086.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person wearing the black shirt\"."}], "task": "refering", "answer_template": "The \"the person wearing the black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 54, 55, 56, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 76, 77, 78, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 99, 100, 101, 104, 105, 106, 107, 108, 109, 110, 111, 112, 122, 123, 124, 125, 127, 128, 129, 130, 131, 132, 133, 146, 147, 148, 150, 151, 152, 153, 154, 155, 169, 170, 171, 172, 176, 177, 192, 193, 194, 195, 217], "bbox": [0.340296875, 0.003958333333333333, 1.0, 0.5818333333333333], "iscrowd": 0, "area": 64425.659299999985, "rle": {"size": [480, 640], "counts": "SZV3>T>d0]Oc0\\Oc0]Od0@`0G9G9G9G5K6J5K6J5K6J5K6J5K6J5K6J5K6J5K5K6J4N2O001O001O1O001O001O001O1O001O00_LcH@\\7>gHBX7;lHDT7:nHGQ76SIIm65VIKi62ZINf60]I0b6MbI3]6KfI4Z6IiI8V6FmI9S6DQJl5_OXJa0g5]O\\Jc0c5ZOaJe0_5YOcJh0\\5UOhJj0X5SOlJm0S5QOPKn0P5oNSKR1l4kNXKU1g4iN\\KV1d4gN_KZ1`4dNcK[1]4bNgK^1X4_NlKV1XMWMk6a1RLP1YM^Me6_1XLk0YMfM^6]1^Ld0[MnMW6]1bL=]MVNP6[1hL;ZMZNm5Z1mL9WM\\Nl5Z1QM7SM`Nk5X1VM5PMbNj5X1ZM3mLeNh5W1_M1jLhNf5V1cM0hLiNe5U1gM0eLkNc5T1jM0dLkNb5T1lM0bLmNa5R1oM0aLnN_5Q1RN0`LnN^5Q1UNO^LPO\\5o0YN0\\LPO[5o0[N0[LQOY5n0^N0ZLQOX5n0`N0XLSOW5l0dNOVLUOU5k0gNOULUOT5k0iNOTLVOR5i0mN0RLVOQ5i0oN0QLWOo4h0SOOoKYOm4g0VOOmKZOm4f0XOOlK[Ok4e0[ONX20kMOU20nMNR21PNMQ22QNMo12SNMm11VNMk12XNLh13ZNLf13\\NKe14]NKc15^NJb16`NGa19c600001_Cd2\\7\\McHg2[7cNZGa1c8aNZGb1d8aNXGb1f8`NWGd1f8^NWGd1h8^NUGc1k8_NRGa1o8aNnF`1R9W20O100O10000O10000O100O10000O100O10000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O1001O5K4L5K4L2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O001O001O001O00001O001O001O001O0000000000O100O100O100O100O1O100O100O100O100O1O100O100O1O1O100O1O100O1O1O100O1O100O1O1O1O1O1O100O1O1O1O1E;L4K5K5M3O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1O1M3N2M3N200O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O10000O100O100O100O10000O100OO"}, "label": "the person wearing the black shirt"}]} {"id": 421086, "image": "COCO_train2014_000000421086.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man cutting slices of pizza wearing gloves and a black shirt\"."}], "task": "refering", "answer_template": "The \"a man cutting slices of pizza wearing gloves and a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 54, 55, 56, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 76, 77, 78, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 99, 100, 101, 104, 105, 106, 107, 108, 109, 110, 111, 112, 122, 123, 124, 125, 127, 128, 129, 130, 131, 132, 133, 146, 147, 148, 150, 151, 152, 153, 154, 155, 169, 170, 171, 172, 176, 177, 192, 193, 194, 195, 217], "bbox": [0.340296875, 0.003958333333333333, 1.0, 0.5818333333333333], "iscrowd": 0, "area": 64425.659299999985, "rle": {"size": [480, 640], "counts": "SZV3>T>d0]Oc0\\Oc0]Od0@`0G9G9G9G5K6J5K6J5K6J5K6J5K6J5K6J5K6J5K5K6J4N2O001O001O1O001O001O001O1O001O00_LcH@\\7>gHBX7;lHDT7:nHGQ76SIIm65VIKi62ZINf60]I0b6MbI3]6KfI4Z6IiI8V6FmI9S6DQJl5_OXJa0g5]O\\Jc0c5ZOaJe0_5YOcJh0\\5UOhJj0X5SOlJm0S5QOPKn0P5oNSKR1l4kNXKU1g4iN\\KV1d4gN_KZ1`4dNcK[1]4bNgK^1X4_NlKV1XMWMk6a1RLP1YM^Me6_1XLk0YMfM^6]1^Ld0[MnMW6]1bL=]MVNP6[1hL;ZMZNm5Z1mL9WM\\Nl5Z1QM7SM`Nk5X1VM5PMbNj5X1ZM3mLeNh5W1_M1jLhNf5V1cM0hLiNe5U1gM0eLkNc5T1jM0dLkNb5T1lM0bLmNa5R1oM0aLnN_5Q1RN0`LnN^5Q1UNO^LPO\\5o0YN0\\LPO[5o0[N0[LQOY5n0^N0ZLQOX5n0`N0XLSOW5l0dNOVLUOU5k0gNOULUOT5k0iNOTLVOR5i0mN0RLVOQ5i0oN0QLWOo4h0SOOoKYOm4g0VOOmKZOm4f0XOOlK[Ok4e0[ONX20kMOU20nMNR21PNMQ22QNMo12SNMm11VNMk12XNLh13ZNLf13\\NKe14]NKc15^NJb16`NGa19c600001_Cd2\\7\\McHg2[7cNZGa1c8aNZGb1d8aNXGb1f8`NWGd1f8^NWGd1h8^NUGc1k8_NRGa1o8aNnF`1R9W20O100O10000O10000O100O10000O100O10000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O1001O5K4L5K4L2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O001O001O001O00001O001O001O001O0000000000O100O100O100O100O1O100O100O100O100O1O100O100O1O1O100O1O100O1O1O100O1O100O1O1O1O1O1O100O1O1O1O1E;L4K5K5M3O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1O1M3N2M3N200O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O10000O100O100O100O10000O100OO"}, "label": "a man cutting slices of pizza wearing gloves and a black shirt"}]} {"id": 421086, "image": "COCO_train2014_000000421086.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white apron\"."}], "task": "refering", "answer_template": "The \"white apron\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 23, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 53, 69, 70, 71, 72, 73, 74, 75, 76, 92, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118], "bbox": [0.013453125, 0.0, 0.344734375, 0.3026875], "iscrowd": 0, "area": 28268.210749999995, "rle": {"size": [480, 640], "counts": "cW4n3c:?L4000000000000000000000000000000000000000000000O100000000000000000000000001O0000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000001O000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000O0100000000000000O10000003M4L3M4L3M4K4L5K4L5K4L5K4L5K4L5K4L5K4L5K4L5K4K6K5K:F:F:F^YT6"}, "label": "white apron"}]} {"id": 421086, "image": "COCO_train2014_000000421086.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white dressed chef standing other side of a table\"."}], "task": "refering", "answer_template": "The \"a white dressed chef standing other side of a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 23, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 53, 69, 70, 71, 72, 73, 74, 75, 76, 92, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118], "bbox": [0.013453125, 0.0, 0.344734375, 0.3026875], "iscrowd": 0, "area": 28268.210749999995, "rle": {"size": [480, 640], "counts": "cW4n3c:?L4000000000000000000000000000000000000000000000O100000000000000000000000001O0000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000001O000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000O0100000000000000O10000003M4L3M4L3M4K4L5K4L5K4L5K4L5K4L5K4L5K4L5K4L5K4K6K5K:F:F:F^YT6"}, "label": "a white dressed chef standing other side of a table"}]} {"id": 421086, "image": "COCO_train2014_000000421086.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza in front of white apron\"."}], "task": "refering", "answer_template": "The \"pizza in front of white apron\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239], "bbox": [0.0016875000000000002, 0.28539583333333335, 0.40112499999999995, 0.6314583333333333], "iscrowd": 0, "area": 35403.49865000001, "rle": {"size": [480, 640], "counts": "[g0T1d;X2^Nb10000O100000000000O100000000000000000000000000000000O10000000000000000000O1000000001O0O10000000000000000000001O0000000O1000000000000000001O0000000000001N10000000001O00000000001O0000000O10001O00000000001O00000000000000O010000000000000000000O1000000000O1000000000O1000000000000000000O10O10000000000000O100000000O100000O01000000O100000000O1000000O100000O10O100000000O10000000000O100000O10O10000001O001O001O001O001O00001N101O001O001O001O00003M7I8H7I7I7H8I8H4L2N3M2N3M3M2N3M3M3M3M4L3M3M4K4M3M4L3M3M4L3M4L3M3M4K4MRXc5"}, "label": "pizza in front of white apron"}]} {"id": 421086, "image": "COCO_train2014_000000421086.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cheese pizza on a pan\"."}], "task": "refering", "answer_template": "The \"cheese pizza on a pan\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239], "bbox": [0.0016875000000000002, 0.28539583333333335, 0.40112499999999995, 0.6314583333333333], "iscrowd": 0, "area": 35403.49865000001, "rle": {"size": [480, 640], "counts": "[g0T1d;X2^Nb10000O100000000000O100000000000000000000000000000000O10000000000000000000O1000000001O0O10000000000000000000001O0000000O1000000000000000001O0000000000001N10000000001O00000000001O0000000O10001O00000000001O00000000000000O010000000000000000000O1000000000O1000000000O1000000000000000000O10O10000000000000O100000000O100000O01000000O100000000O1000000O100000O10O100000000O10000000000O100000O10O10000001O001O001O001O001O00001N101O001O001O001O00003M7I8H7I7I7H8I8H4L2N3M2N3M3M2N3M3M3M3M4L3M3M4K4M3M4L3M3M4L3M4L3M3M4K4MRXc5"}, "label": "cheese pizza on a pan"}]} {"id": 363752, "image": "COCO_train2014_000000363752.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player sitting in the dugout\"."}], "task": "refering", "answer_template": "The \"a baseball player sitting in the dugout\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 121, 122, 135, 136, 137, 149, 150, 151, 163, 164, 165, 166, 178, 179, 180], "bbox": [0.6514567901234567, 0.43062, 0.9300987654320986, 0.67924], "iscrowd": 0, "area": 6728.577850000001, "rle": {"size": [500, 405], "counts": "WUQ4U1]>6J6J4N1N2SO`NcCb1U in this image.", "objects": [{"patches": [107, 121, 122, 135, 136, 137, 149, 150, 151, 163, 164, 165, 166, 178, 179, 180], "bbox": [0.6514567901234567, 0.43062, 0.9300987654320986, 0.67924], "iscrowd": 0, "area": 6728.577850000001, "rle": {"size": [500, 405], "counts": "WUQ4U1]>6J6J4N1N2SO`NcCb1U in this image.", "objects": [{"patches": [5, 6, 7, 15, 16, 17, 18, 19, 20, 21, 29, 30, 31, 32, 33, 34, 35, 36, 44, 45, 46, 47, 48, 49, 50, 57, 58, 59, 60, 61, 62, 63, 64, 74, 75, 76, 77, 78, 88, 89, 90, 91, 92, 102, 103, 104, 105, 106, 116, 117, 118, 119, 120, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 159, 160, 161, 162, 163, 174, 175, 176, 177, 178, 186, 187, 188, 189, 190, 191, 192, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 212, 213, 214, 215, 219, 220, 221, 226, 227, 233, 234, 235, 236, 241, 248, 249, 250], "bbox": [0.10795061728395061, 0.0294, 0.9134814814814816, 0.96888], "iscrowd": 0, "area": 60294.06449999999, "rle": {"size": [500, 405], "counts": "Qae0:U?6N2N2N2N2N2N3jKTO_In0`6TO]In0a6TO^Im0lMROg63\\Kk0fMXOn6N[Kk0aM\\OT7J[Kj0[MAZ7GYKi0WMF^7C[Kg0oLMf7^OYKf0ZLb0]8iNYK]2g4dMXK]2g4fMWKY2i4kMSKU2l4oMRKP2n4SNoJn1P5VNlJj1T5YNjJf1V5]NgJc1X5bNeJ^1Z5eNcJ[1]5hNaJW1_5mN]JT1b5oN\\JP1c5TOZJl0f5WOXJh0h5XOXJh0h5YOXJf0h5ZOZJd0e5^O[Ja0e5_O\\J`0d5A\\J>d5C]J;c5E^J:b5G^J9`5HbJ6^5KeJ1[5OjJLU56oJEQ5;SKAl4a0UK]Ok4c0VK\\Oj4e0UK[Oj4g0VKXOj4h0VKXOi4j0WKUOi4k0WKUOi4l0WKSOh4n0XKROh4n0YKQOf4P1QJ`MTO_1k6Q1oIhMPOW1Q7Q1mIkMROS1P7S1lImMSOP1Q7S1kIoMTOm0P7T1kIRNTOj0Q7OWH_Oc1JTOh0Q70[HZOa1OQOh0S7O^HWO^12POi0T7MbHTO[17mNi0U7MfHPOX1;kNi0W7KSK]OdMi0X7KTK]ObMi0Z7JTK^O`Mi0\\7HUK@]Mi0]7HVK_O\\Mj0^7FWKAYMn0[7B\\KAWMR1Y7]O`KAWMV1U7XOeKBVM[1o6TOkKAVM_1k6oNPLBUMc1f6lNULAUMh1a6gN[L@TMm1\\6cNaL@SMo1Z6aNcL@SMR2W6]NgLARMT2T6\\NjL@RMW2Q6YNmL@RMY2Y2gLR1`1D@QM\\2P2kLZ1Y1E@QM^2i1oL`1R1GAPMa2`1SMg1l0HAQMc2W1WMo1e0IAQMe2o0\\MU2>LAPMo2`0XMc28MBoLZ3OSMT30OCnLR5Q3\\K1BmLT5Q3ZK2BmLU5o2YK5BlLV5n2XK6BlLW5l2WK9CjLW5l2VK:CjLX5k2UK;CjLY5i2TK>ChL[5i2RK>DiL[5g2QKa0DhL\\5f2PKb0EgL\\5f2oJc0EgL]5d2nJf0EfL^5c2mJf0FgL^5a2lJi0FeL`5a2jJi0a6WO_Ih0c6VO]Ij0e6UO[Ij0g6TOYIl0i6SOWIl0j6SOWIl0k6SOTIm0n6RORIn0o6POQIP1Q7oNoHQ1Q7nNPIQ1R7nNmHS1V7jNjHU1Z7gNfHZ1]7cNcH\\1a7`N`H_1e7]NZHd1i7YNWHf1m7VNSHk1P8RNPHm1U8nMlGR2W8kMhGU2\\8gMeGY2_8cM`G^2c8_M]Ga2f8[M[Ge2i8WMVGj2P9oLQGR3S9iLlFX3h:000O100O10001N100O10000O100O2O000O101O0O2O1N2O001N2O0O2O1O1N101N2O1O0O2O1cElKW9V4eFmK[9S4bFPL^9S4\\FQLb9U4WFmKi9P5N10001N100O2O001M2M3M4TJTJk1o5aMnJX2P5UMlKd2S4XMYL`2f3^MdL\\2Z3cMPMU2P3hM[MQ2c2nMeMm1Z2PNjMo1U2PNnMP2P2nMTNQ2l1mMWNS2g1kM]NU2a1jMbNU2`1fMdNZ2]1bMfN]2\\1^MhNb2X1[MkNe2V1VMnNi2T1SMoNm2R1nLROQ3P1kLSOU3n0fLVOZ3k0bLXO]3j0^LZOb3g0ZL\\Oe3f0VL^Oj3W600O100O01000O100O10000O100O01000O100O10000O100O02O1O001O1O1O1O00O1N1O2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N1O2N2N2N5K5K5K5K4L3M00001O00001O0O100O2O000O2O0O100O2N1O1O1O2N1O1O2N1O1O1O2N1O101N1O1O1OUTa0"}, "label": "a baseball player swinging a bat"}]} {"id": 363752, "image": "COCO_train2014_000000363752.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player with 760 next to him\"."}], "task": "refering", "answer_template": "The \"a baseball player with 760 next to him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 15, 16, 17, 18, 19, 20, 21, 29, 30, 31, 32, 33, 34, 35, 36, 44, 45, 46, 47, 48, 49, 50, 57, 58, 59, 60, 61, 62, 63, 64, 74, 75, 76, 77, 78, 88, 89, 90, 91, 92, 102, 103, 104, 105, 106, 116, 117, 118, 119, 120, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 159, 160, 161, 162, 163, 174, 175, 176, 177, 178, 186, 187, 188, 189, 190, 191, 192, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 212, 213, 214, 215, 219, 220, 221, 226, 227, 233, 234, 235, 236, 241, 248, 249, 250], "bbox": [0.10795061728395061, 0.0294, 0.9134814814814816, 0.96888], "iscrowd": 0, "area": 60294.06449999999, "rle": {"size": [500, 405], "counts": "Qae0:U?6N2N2N2N2N2N3jKTO_In0`6TO]In0a6TO^Im0lMROg63\\Kk0fMXOn6N[Kk0aM\\OT7J[Kj0[MAZ7GYKi0WMF^7C[Kg0oLMf7^OYKf0ZLb0]8iNYK]2g4dMXK]2g4fMWKY2i4kMSKU2l4oMRKP2n4SNoJn1P5VNlJj1T5YNjJf1V5]NgJc1X5bNeJ^1Z5eNcJ[1]5hNaJW1_5mN]JT1b5oN\\JP1c5TOZJl0f5WOXJh0h5XOXJh0h5YOXJf0h5ZOZJd0e5^O[Ja0e5_O\\J`0d5A\\J>d5C]J;c5E^J:b5G^J9`5HbJ6^5KeJ1[5OjJLU56oJEQ5;SKAl4a0UK]Ok4c0VK\\Oj4e0UK[Oj4g0VKXOj4h0VKXOi4j0WKUOi4k0WKUOi4l0WKSOh4n0XKROh4n0YKQOf4P1QJ`MTO_1k6Q1oIhMPOW1Q7Q1mIkMROS1P7S1lImMSOP1Q7S1kIoMTOm0P7T1kIRNTOj0Q7OWH_Oc1JTOh0Q70[HZOa1OQOh0S7O^HWO^12POi0T7MbHTO[17mNi0U7MfHPOX1;kNi0W7KSK]OdMi0X7KTK]ObMi0Z7JTK^O`Mi0\\7HUK@]Mi0]7HVK_O\\Mj0^7FWKAYMn0[7B\\KAWMR1Y7]O`KAWMV1U7XOeKBVM[1o6TOkKAVM_1k6oNPLBUMc1f6lNULAUMh1a6gN[L@TMm1\\6cNaL@SMo1Z6aNcL@SMR2W6]NgLARMT2T6\\NjL@RMW2Q6YNmL@RMY2Y2gLR1`1D@QM\\2P2kLZ1Y1E@QM^2i1oL`1R1GAPMa2`1SMg1l0HAQMc2W1WMo1e0IAQMe2o0\\MU2>LAPMo2`0XMc28MBoLZ3OSMT30OCnLR5Q3\\K1BmLT5Q3ZK2BmLU5o2YK5BlLV5n2XK6BlLW5l2WK9CjLW5l2VK:CjLX5k2UK;CjLY5i2TK>ChL[5i2RK>DiL[5g2QKa0DhL\\5f2PKb0EgL\\5f2oJc0EgL]5d2nJf0EfL^5c2mJf0FgL^5a2lJi0FeL`5a2jJi0a6WO_Ih0c6VO]Ij0e6UO[Ij0g6TOYIl0i6SOWIl0j6SOWIl0k6SOTIm0n6RORIn0o6POQIP1Q7oNoHQ1Q7nNPIQ1R7nNmHS1V7jNjHU1Z7gNfHZ1]7cNcH\\1a7`N`H_1e7]NZHd1i7YNWHf1m7VNSHk1P8RNPHm1U8nMlGR2W8kMhGU2\\8gMeGY2_8cM`G^2c8_M]Ga2f8[M[Ge2i8WMVGj2P9oLQGR3S9iLlFX3h:000O100O10001N100O10000O100O2O000O101O0O2O1N2O001N2O0O2O1O1N101N2O1O0O2O1cElKW9V4eFmK[9S4bFPL^9S4\\FQLb9U4WFmKi9P5N10001N100O2O001M2M3M4TJTJk1o5aMnJX2P5UMlKd2S4XMYL`2f3^MdL\\2Z3cMPMU2P3hM[MQ2c2nMeMm1Z2PNjMo1U2PNnMP2P2nMTNQ2l1mMWNS2g1kM]NU2a1jMbNU2`1fMdNZ2]1bMfN]2\\1^MhNb2X1[MkNe2V1VMnNi2T1SMoNm2R1nLROQ3P1kLSOU3n0fLVOZ3k0bLXO]3j0^LZOb3g0ZL\\Oe3f0VL^Oj3W600O100O01000O100O10000O100O01000O100O10000O100O02O1O001O1O1O1O00O1N1O2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N1O2N2N2N5K5K5K5K4L3M00001O00001O0O100O2O000O2O0O100O2N1O1O1O2N1O1O2N1O1O1O2N1O101N1O1O1OUTa0"}, "label": "a baseball player with 760 next to him"}]} {"id": 44266, "image": "COCO_train2014_000000044266.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown wicker chair sitting in front of a brown desk\"."}], "task": "refering", "answer_template": "The \"a brown wicker chair sitting in front of a brown desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 106, 107, 108, 130, 131, 153, 154, 155, 156, 176, 177, 178, 179, 198, 199, 200, 201, 222], "bbox": [0.6183125, 0.1779375, 0.8190625, 0.54625], "iscrowd": 0, "area": 10400.00685, "rle": {"size": [480, 640], "counts": "Yki51o>1N101N2O0O2O1N101N100O0iKGiI9V6ObI1_67YIIg6>RIAo6f0iHZOX7n0`HRO`7U1XHkNi7\\1PHdNP8d1gG\\NZ8k1_GUNa8R2WGnMj8Y2oFgMQ9a2fF_MZ9i2_FWMa9P3WFPMj9X3nEgLS:`3eEaL[:f3^EYLc:R40O100O19G;D;F3M001O00000000001O000O10D in this image.", "objects": [{"patches": [83, 84, 85, 106, 107, 108, 130, 131, 153, 154, 155, 156, 176, 177, 178, 179, 198, 199, 200, 201, 222], "bbox": [0.6183125, 0.1779375, 0.8190625, 0.54625], "iscrowd": 0, "area": 10400.00685, "rle": {"size": [480, 640], "counts": "Yki51o>1N101N2O0O2O1N101N100O0iKGiI9V6ObI1_67YIIg6>RIAo6f0iHZOX7n0`HRO`7U1XHkNi7\\1PHdNP8d1gG\\NZ8k1_GUNa8R2WGnMj8Y2oFgMQ9a2fF_MZ9i2_FWMa9P3WFPMj9X3nEgLS:`3eEaL[:f3^EYLc:R40O100O19G;D;F3M001O00000000001O000O10D in this image.", "objects": [{"patches": [166, 189, 190, 213, 214, 215, 236, 237, 238, 239, 259, 260, 261, 262, 263, 282, 283, 284, 285, 286, 287, 288, 305, 306, 307, 308, 309, 310, 311, 328, 329, 330, 331, 332, 333, 334, 351, 352, 353, 354, 355, 356, 357, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.233875, 0.4215, 0.569359375, 0.9871041666666667], "iscrowd": 0, "area": 26686.44945, "rle": {"size": [480, 640], "counts": "[`V28i>3M1O2N1O1O1O2N1O1O1O1O1O10O01O1`HWOf0j0VOYOj0h0RO[On0f0nN^OQ1c0kN@U1a0gNBY1?aNG^19^NId18ZNHg1;UNEl1>PNAR2b0jMPOjJN]7U1eMmNUKGW7_1`MkNn2W1oLiNR3Z1`H5m1aNd5]1PHd0U2oMl5Q4oIoKR6T4jIlKW6W4eIiK\\6[4_IeKb6^4ZIbKg6a4UI`Kk6b4RI^Ko6b4PI^KQ7c4mH]KT7c4kH]KV7c4iH]KX7c4gH]KZ7d4dH\\K]7d4bH\\K_7d4`H]K`7c4_H]Kb7d4\\H\\Ke7d4ZH\\Kf7e4YH[Kg7g4WHYKj7g4UHYKk7h4THXKm7h4RHXKn7j4PHWKP8i4oGWKQ8j4nGVKS8j4kGWKU8k4iGUKW8l4hGTKY8l4fGTKZ8m4eGSK\\8l4dGTK\\8l4dGTK]8k4cGVK\\8k4cGUK]8k4cGUK^8j4bGVK^8k4aGUK`8j4`GVK`8j4`GVKa8i4_GWKa8j4^GVKc8i4]GXKb8h4^GXKb8i4]GWKd8h4\\GXKd8h4\\GXKe8g4[GYKe8h4ZGXKg8g4YGYKg8g4YGYKh8g4WGYKi8g4WGYKi8g4WGYKj8f4VGZKj8g4UGYKl8f4TGZKl8f4TGZKm8f4RGZKn8f4RGZKo8e4QG[KP9d4PG\\KQ9d4nF\\KS9c4mF]K7A`8Q5YG^K6Bb8o4XG_K4De8m4VG_K3Fh8j4UG`K1Hl8f4SGbK0In8e4QGbKOKQ9b4PGPLQ9o3oFQLR9n3nFRLS9n3lFRLU9m3kFSLW9k3iFULX9i41O1O1O1O2N1O1O1O1O1O1O2N7I6J6J7I00O1O100O1O100OCgEULX:l3kEQLU:o3nEnKQ:R4SFkKl9V4?01O1N101O1O0O2O1O001N2O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O0O2O1O0aNVDHk;5[DGf;5aDG_;7gDEZ;7iDIX;4jDLV;2lDNU;1kDOV;OkD1U;NkD3V;KkD5V;IkD7U;HlD8U;FlD:U;DlD;U;DlD;V;CkD:Y;DhD9[;FeD8_;GaD7b;G_D6d;I]D4g;JYD4k;JVD3m;LSD2Q in this image.", "objects": [{"patches": [166, 189, 190, 213, 214, 215, 236, 237, 238, 239, 259, 260, 261, 262, 263, 282, 283, 284, 285, 286, 287, 288, 305, 306, 307, 308, 309, 310, 311, 328, 329, 330, 331, 332, 333, 334, 351, 352, 353, 354, 355, 356, 357, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.233875, 0.4215, 0.569359375, 0.9871041666666667], "iscrowd": 0, "area": 26686.44945, "rle": {"size": [480, 640], "counts": "[`V28i>3M1O2N1O1O1O2N1O1O1O1O1O10O01O1`HWOf0j0VOYOj0h0RO[On0f0nN^OQ1c0kN@U1a0gNBY1?aNG^19^NId18ZNHg1;UNEl1>PNAR2b0jMPOjJN]7U1eMmNUKGW7_1`MkNn2W1oLiNR3Z1`H5m1aNd5]1PHd0U2oMl5Q4oIoKR6T4jIlKW6W4eIiK\\6[4_IeKb6^4ZIbKg6a4UI`Kk6b4RI^Ko6b4PI^KQ7c4mH]KT7c4kH]KV7c4iH]KX7c4gH]KZ7d4dH\\K]7d4bH\\K_7d4`H]K`7c4_H]Kb7d4\\H\\Ke7d4ZH\\Kf7e4YH[Kg7g4WHYKj7g4UHYKk7h4THXKm7h4RHXKn7j4PHWKP8i4oGWKQ8j4nGVKS8j4kGWKU8k4iGUKW8l4hGTKY8l4fGTKZ8m4eGSK\\8l4dGTK\\8l4dGTK]8k4cGVK\\8k4cGUK]8k4cGUK^8j4bGVK^8k4aGUK`8j4`GVK`8j4`GVKa8i4_GWKa8j4^GVKc8i4]GXKb8h4^GXKb8i4]GWKd8h4\\GXKd8h4\\GXKe8g4[GYKe8h4ZGXKg8g4YGYKg8g4YGYKh8g4WGYKi8g4WGYKi8g4WGYKj8f4VGZKj8g4UGYKl8f4TGZKl8f4TGZKm8f4RGZKn8f4RGZKo8e4QG[KP9d4PG\\KQ9d4nF\\KS9c4mF]K7A`8Q5YG^K6Bb8o4XG_K4De8m4VG_K3Fh8j4UG`K1Hl8f4SGbK0In8e4QGbKOKQ9b4PGPLQ9o3oFQLR9n3nFRLS9n3lFRLU9m3kFSLW9k3iFULX9i41O1O1O1O2N1O1O1O1O1O1O2N7I6J6J7I00O1O100O1O100OCgEULX:l3kEQLU:o3nEnKQ:R4SFkKl9V4?01O1N101O1O0O2O1O001N2O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O0O2O1O0aNVDHk;5[DGf;5aDG_;7gDEZ;7iDIX;4jDLV;2lDNU;1kDOV;OkD1U;NkD3V;KkD5V;IkD7U;HlD8U;FlD:U;DlD;U;DlD;V;CkD:Y;DhD9[;FeD8_;GaD7b;G_D6d;I]D4g;JYD4k;JVD3m;LSD2Q in this image.", "objects": [{"patches": [195, 196, 211, 212, 213, 214, 215, 216, 228, 229, 230, 231, 232, 233, 234, 235, 236, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 295, 296, 297, 298, 299, 300, 301, 302, 303, 313, 314, 315, 316, 317, 318, 319, 320, 333, 334, 335, 336], "bbox": [0.35614583333333333, 0.513140625, 0.9274791666666666, 0.85684375], "iscrowd": 0, "area": 42337.57395, "rle": {"size": [640, 480], "counts": "mY[33kc04L5L3L4L5K5K6J6J6J6J6J5K6J6J6J3M3M2N2O2M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3N1N2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2O1N3M2N201N10001N10001N100O2O000O101O0O101N100O2O0O110O00001O01O01O00010O000010O0001O00010O000010O0000010O0001O00010O0000010O0000010O000000010O0000010O0000010O000000010O0000010O0000010O00000010O00001O01O01O0001O01O00010O00001O01O0001O01O01O00010O00001O01O0001O01O01O00010O0000001O0O101N100O2O0O100O2O0O101N100O2O000O101N100O2O0O101N100O101N100O2O0N2M4M2N2M4M2M3N2M4M2M3N3L3N2M4M2M3N2M4M2M3N3L3N2M4M2M3N2N3L3N2M4M3L4M3L4M3L4M3L4M4K6K6I6K6I7J5JZ_e0"}, "label": "the small red container with slices of seasoned sausage"}]} {"id": 503022, "image": "COCO_train2014_000000503022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sliced bananas in a red bowl\"."}], "task": "refering", "answer_template": "The \"sliced bananas in a red bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 211, 212, 213, 214, 215, 216, 228, 229, 230, 231, 232, 233, 234, 235, 236, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 295, 296, 297, 298, 299, 300, 301, 302, 303, 313, 314, 315, 316, 317, 318, 319, 320, 333, 334, 335, 336], "bbox": [0.35614583333333333, 0.513140625, 0.9274791666666666, 0.85684375], "iscrowd": 0, "area": 42337.57395, "rle": {"size": [640, 480], "counts": "mY[33kc04L5L3L4L5K5K6J6J6J6J6J5K6J6J6J3M3M2N2O2M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3N1N2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2O1N3M2N201N10001N10001N100O2O000O101O0O101N100O2O0O110O00001O01O01O00010O000010O0001O00010O000010O0000010O0001O00010O0000010O0000010O000000010O0000010O0000010O000000010O0000010O0000010O00000010O00001O01O01O0001O01O00010O00001O01O0001O01O01O00010O00001O01O0001O01O01O00010O0000001O0O101N100O2O0O100O2O0O101N100O2O000O101N100O2O0O101N100O101N100O2O0N2M4M2N2M4M2M3N2M4M2M3N3L3N2M4M2M3N2M4M2M3N3L3N2M4M2M3N2N3L3N2M4M3L4M3L4M3L4M3L4M4K6K6I6K6I7J5JZ_e0"}, "label": "sliced bananas in a red bowl"}]} {"id": 503022, "image": "COCO_train2014_000000503022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"purple box with salad greens , sliced almonds and purple flowers\"."}], "task": "refering", "answer_template": "The \"purple box with salad greens , sliced almonds and purple flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 97, 98, 99, 100, 101, 113, 114, 115, 116, 117, 118, 129, 130, 131, 132, 133, 134, 135, 146, 147, 148, 149, 150, 151, 152, 162, 163, 164, 165, 166, 167, 168, 169, 178, 179, 180, 181, 182, 183, 184, 185, 186, 195, 196, 197, 198, 199, 200, 201, 202, 203, 214, 215, 216, 217, 218, 219, 220, 234, 235, 236, 237], "bbox": [0.5077291666666667, 0.20810937499999999, 1.0, 0.61178125], "iscrowd": 0, "area": 39277.56870000001, "rle": {"size": [640, 480], "counts": "cih41mc03N2M3N1N3N2M3M3N1N3N2M3N2M2O2M3N2M3N1N3N1N2O2M2O1N3N1N2O2M2N2O2M2O2M2O1N3N1N2O2M2O1N3N1N2O2M2O2N1N2O2M2O1O2M2O1O2M2O1N3N1O1N3N1O1N3N1N2O2N1N2O2N1N2O2M2O1O2M2O1O2M2O1N3N1O1N3N1O1N3N1N2O2N1N2O2N1N2O2M2O1O1N3N1O1N3N1N2O2N1N2O2N1N2O2M2O1O2M2O1O2N11O01O00010O000010O0001O01O01O00010O00001O01O01O00010O000010O0001O01O01O00010O0000010O000010O0001O010O000010O01O00010O001O01O01O0010O0001O010O00001O010O00003N2MN2N2M4M2M3N2M4M2N2M3N3L3N2M3N3L3N2M3N3L3N2N2M4M2M3N2M4M2M3N2M6K6I7JgJ"}, "label": "purple box with salad greens , sliced almonds and purple flowers"}]} {"id": 503022, "image": "COCO_train2014_000000503022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"purple container with purple flowers , green leaves and sliced almonds\"."}], "task": "refering", "answer_template": "The \"purple container with purple flowers , green leaves and sliced almonds\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 97, 98, 99, 100, 101, 113, 114, 115, 116, 117, 118, 129, 130, 131, 132, 133, 134, 135, 146, 147, 148, 149, 150, 151, 152, 162, 163, 164, 165, 166, 167, 168, 169, 178, 179, 180, 181, 182, 183, 184, 185, 186, 195, 196, 197, 198, 199, 200, 201, 202, 203, 214, 215, 216, 217, 218, 219, 220, 234, 235, 236, 237], "bbox": [0.5077291666666667, 0.20810937499999999, 1.0, 0.61178125], "iscrowd": 0, "area": 39277.56870000001, "rle": {"size": [640, 480], "counts": "cih41mc03N2M3N1N3N2M3M3N1N3N2M3N2M2O2M3N2M3N1N3N1N2O2M2O1N3N1N2O2M2N2O2M2O2M2O1N3N1N2O2M2O1N3N1N2O2M2O2N1N2O2M2O1O2M2O1O2M2O1N3N1O1N3N1O1N3N1N2O2N1N2O2N1N2O2M2O1O2M2O1O2M2O1N3N1O1N3N1O1N3N1N2O2N1N2O2N1N2O2M2O1O1N3N1O1N3N1N2O2N1N2O2N1N2O2M2O1O2M2O1O2N11O01O00010O000010O0001O01O01O00010O00001O01O01O00010O000010O0001O01O01O00010O0000010O000010O0001O010O000010O01O00010O001O01O01O0010O0001O010O00001O010O00003N2MN2N2M4M2M3N2M4M2N2M3N3L3N2M3N3L3N2M3N3L3N2N2M4M2M3N2M4M2M3N2M6K6I7JgJ"}, "label": "purple container with purple flowers , green leaves and sliced almonds"}]} {"id": 503022, "image": "COCO_train2014_000000503022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sliced fruit in a blue bowl , bottom left of the tray\"."}], "task": "refering", "answer_template": "The \"sliced fruit in a blue bowl , bottom left of the tray\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 170, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 192, 193, 194, 195, 204, 205, 206, 207, 208, 209, 210, 211, 221, 222, 223, 224, 225, 226, 227, 228, 238, 239, 240, 241, 242, 243, 244, 255, 256, 257, 258, 259, 260, 261, 274, 275, 276, 277, 293], "bbox": [0.0058125, 0.420296875, 0.48412499999999997, 0.7441093750000001], "iscrowd": 0, "area": 31872.647900000004, "rle": {"size": [640, 480], "counts": "QV2W1hb0h1XNb0^O1O1O1O1O1O1O1N3N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1N2O1O1O1O1O1O2N1O1O1O100O10000O100O10000O100O10000O10000O101N2O2N1N3N1N3N1O2N101N1O2O0O2N101N1O2O0O2N1O2O0O1O01O01O01O00001O0000001O01O01O00001O00000010O0001O00001O00001O01O0001O00001O0O101N1O1O1O2N1O1O2N1O1O2O0O1O2N1O1O1O2N1O1O2N100O2N1O1O1O2N1O1O2N1O101N1O1O2N1O1O1O2N1O1O2N1N2N3M2O1N3M2N2O1N3M2N2O2M2N2N3N1N2N1O3N1N3M2N3N1N3M2O2M2N3M2O2M2N3M2O2M2N3M2O2M2N3M2O2M2N3M2O2M2N3M2O2M2N3M2O2M2N3M2O2M2Nmej4"}, "label": "sliced fruit in a blue bowl , bottom left of the tray"}]} {"id": 503022, "image": "COCO_train2014_000000503022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue container with chopped fruit inside it\"."}], "task": "refering", "answer_template": "The \"a blue container with chopped fruit inside it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 170, 171, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 192, 193, 194, 195, 204, 205, 206, 207, 208, 209, 210, 211, 221, 222, 223, 224, 225, 226, 227, 228, 238, 239, 240, 241, 242, 243, 244, 255, 256, 257, 258, 259, 260, 261, 274, 275, 276, 277, 293], "bbox": [0.0058125, 0.420296875, 0.48412499999999997, 0.7441093750000001], "iscrowd": 0, "area": 31872.647900000004, "rle": {"size": [640, 480], "counts": "QV2W1hb0h1XNb0^O1O1O1O1O1O1O1N3N1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O2N1O1N2O1O1O1O1O1O2N1O1O1O100O10000O100O10000O100O10000O10000O101N2O2N1N3N1N3N1O2N101N1O2O0O2N101N1O2O0O2N1O2O0O1O01O01O01O00001O0000001O01O01O00001O00000010O0001O00001O00001O01O0001O00001O0O101N1O1O1O2N1O1O2N1O1O2O0O1O2N1O1O1O2N1O1O2N100O2N1O1O1O2N1O1O2N1O101N1O1O2N1O1O1O2N1O1O2N1N2N3M2O1N3M2N2O1N3M2N2O2M2N2N3N1N2N1O3N1N3M2N3N1N3M2O2M2N3M2O2M2N3M2O2M2N3M2O2M2N3M2O2M2N3M2O2M2N3M2O2M2N3M2O2M2N3M2O2M2Nmej4"}, "label": "a blue container with chopped fruit inside it"}]} {"id": 503022, "image": "COCO_train2014_000000503022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"carrots and cauliflower in a clear container\"."}], "task": "refering", "answer_template": "The \"carrots and cauliflower in a clear container\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 91, 92, 93, 94, 95, 96, 107, 108, 109, 110, 111, 112, 113, 123, 124, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 143, 144, 145, 146, 156, 157, 158, 159, 160, 161, 162, 174, 175, 176, 177, 178, 179, 194, 195], "bbox": [0.19574999999999998, 0.19007812500000001, 0.6992499999999999, 0.525859375], "iscrowd": 0, "area": 29227.25245, "rle": {"size": [640, 480], "counts": "\\Pk13kc03L4L5D:O2N1N2O2N1O1N3N1O1N3N1O1O2M2O1O2N1N2O2N1N2O2N1O2M2O1O2N1N2O2N1N2O2N1O1N3N1O1O2M2O1O2M2O1O2N1N2O2N1O1N3N1O1N3N1O1O2M2O2N1O1N3N1O1N3N1O1O2M2O1O2N1N2O2N1N2O2N1O1N3N1O1O2M2O1O2M2O1O2N1O2N100O2O0O101N100O2O0O101N100O2O000O2O0O10001N100O2O000O2O00001O1O1O2N1O1O1O1O002N3M4L3MO100000000O10001O000O1WOeBmJ\\=P5gBoJY=o4jBPKV=o4lBoJV=n4oBnJR=Q5RCiJQ=U5TCeJo in this image.", "objects": [{"patches": [74, 75, 76, 77, 78, 91, 92, 93, 94, 95, 96, 107, 108, 109, 110, 111, 112, 113, 123, 124, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 143, 144, 145, 146, 156, 157, 158, 159, 160, 161, 162, 174, 175, 176, 177, 178, 179, 194, 195], "bbox": [0.19574999999999998, 0.19007812500000001, 0.6992499999999999, 0.525859375], "iscrowd": 0, "area": 29227.25245, "rle": {"size": [640, 480], "counts": "\\Pk13kc03L4L5D:O2N1N2O2N1O1N3N1O1N3N1O1O2M2O1O2N1N2O2N1N2O2N1O2M2O1O2N1N2O2N1N2O2N1O1N3N1O1O2M2O1O2M2O1O2N1N2O2N1O1N3N1O1N3N1O1O2M2O2N1O1N3N1O1N3N1O1O2M2O1O2N1N2O2N1N2O2N1O1N3N1O1O2M2O1O2M2O1O2N1O2N100O2O0O101N100O2O0O101N100O2O000O2O0O10001N100O2O000O2O00001O1O1O2N1O1O1O1O002N3M4L3MO100000000O10001O000O1WOeBmJ\\=P5gBoJY=o4jBPKV=o4lBoJV=n4oBnJR=Q5RCiJQ=U5TCeJo in this image.", "objects": [{"patches": [219, 220, 238, 239, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 282, 283, 284, 285, 286, 287, 288, 306, 307], "bbox": [0.28646875, 0.606734693877551, 0.579125, 0.8337188208616779], "iscrowd": 0, "area": 9444.51095, "rle": {"size": [441, 640], "counts": "Xb_21g=3N1O1O1N3N1O1N101O001N2O001O0O2O1O0O2O001O1O001N101O1O001O001O1O0O2O001O1O001O001N2O00000O1M3N1O2M3N2N1N3N2N2I610O10O1000O10O1000O10O1000O10O1000O10O1000O0100000O0100000O0100000O0100000O010000O0100000O0100000O0100000O0100000O01000O10O1000O10O1000O10O1000O10O10O1000O10O10O1000O10O1000O10O1000O10O10O1000O10O1000O10O1000O10O100N1N3M3M3M3M3M3M3M2N3M3M3M3M3M3M2N3M3M3M3M3MePd3"}, "label": "the black laptop that the woman is holding"}]} {"id": 281840, "image": "COCO_train2014_000000281840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black laptop\"."}], "task": "refering", "answer_template": "The \"the black laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 238, 239, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 282, 283, 284, 285, 286, 287, 288, 306, 307], "bbox": [0.28646875, 0.606734693877551, 0.579125, 0.8337188208616779], "iscrowd": 0, "area": 9444.51095, "rle": {"size": [441, 640], "counts": "Xb_21g=3N1O1O1N3N1O1N101O001N2O001O0O2O1O0O2O001O1O001N101O1O001O001O1O0O2O001O1O001O001N2O00000O1M3N1O2M3N2N1N3N2N2I610O10O1000O10O1000O10O1000O10O1000O10O1000O0100000O0100000O0100000O0100000O010000O0100000O0100000O0100000O0100000O01000O10O1000O10O1000O10O1000O10O10O1000O10O10O1000O10O1000O10O1000O10O10O1000O10O1000O10O1000O10O100N1N3M3M3M3M3M3M3M2N3M3M3M3M3M3M2N3M3M3M3M3MePd3"}, "label": "the black laptop"}]} {"id": 421105, "image": "COCO_train2014_000000421105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the nose of a red , small airplane with sparky written on it\"."}], "task": "refering", "answer_template": "The \"the nose of a red , small airplane with sparky written on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 29, 30, 52, 75, 98, 120, 121, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 205, 206, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 228, 229, 233, 234, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 255, 256, 257, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 280, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 302, 303, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 338, 339, 340, 341, 342, 343], "bbox": [0.1244375, 0.022482435597189696, 1.0, 0.9820140515222484], "iscrowd": 0, "area": 80654.52935000003, "rle": {"size": [427, 640], "counts": "ZeQ1k0]<5K3N2N2]MjNoHX1l6nNPIU1k6QORIQ1j6TOSIn0k6UOSIk0m6WOPIj0P7XOmHh0S7[OkHe0U7]OhHc0Z7^OcHc0^7^O`Hb0`7@]H`0e7AXH`0i7BTH=m7EPH in this image.", "objects": [{"patches": [7, 29, 30, 52, 75, 98, 120, 121, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 205, 206, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 228, 229, 233, 234, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 255, 256, 257, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 280, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 302, 303, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 338, 339, 340, 341, 342, 343], "bbox": [0.1244375, 0.022482435597189696, 1.0, 0.9820140515222484], "iscrowd": 0, "area": 80654.52935000003, "rle": {"size": [427, 640], "counts": "ZeQ1k0]<5K3N2N2]MjNoHX1l6nNPIU1k6QORIQ1j6TOSIn0k6UOSIk0m6WOPIj0P7XOmHh0S7[OkHe0U7]OhHc0Z7^OcHc0^7^O`Hb0`7@]H`0e7AXH`0i7BTH=m7EPH in this image.", "objects": [{"patches": [66, 80, 81, 82, 95, 96, 97, 109, 110, 111, 112, 113, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 153, 154, 155, 156, 157, 158, 159, 169, 170, 171, 172, 173, 174, 183, 184, 185, 186, 187, 188, 189, 190, 197, 198, 199, 200, 201, 202, 203, 204, 205, 211, 212, 213, 214, 215, 216, 217, 218, 220, 221, 229, 230, 231, 232, 233, 235, 236, 245, 246, 247, 248, 259, 260, 261, 274, 275, 276, 277, 288, 289, 290, 291, 292, 302, 303, 304, 306, 307, 308, 317, 318, 319], "bbox": [0.11896713615023474, 0.207046875, 0.7804694835680751, 0.9534218750000001], "iscrowd": 0, "area": 50324.13489999997, "rle": {"size": [640, 426], "counts": "iXP13jc03M3M3O1O100O1O100O100O1@B]]O?cb0EY]O;fb0JU]O7kb0?O1O100O100O1PCjNV7V1gHnNW7S1gHPOW7P1gHTOX7l0eHYOX7g0gH\\OX7d0eH@Z7?dHE[7:dHH\\78bHK]74bHN^72`H0`7O_H3a7L]H7c7I[H9d7G[H;e7EYH=f7CYH?e7CYH`0e7@[Ha0c7@]Ha0b7_O^Hb0`7_O`Hb0_7^OaHc0]7]OdHd0Z7]OeHe0Z7[OfHf0X7ZOiHh0U7XOkHi0S7WOnHj0Q7VOoHk0o6VOQIk0n6TOSIm0k6TOUIm0j6ROVIP1h6QOXIQ1f6oNZIR1]5hMlFV1h3S1S5QNSGl0j3T1i4ZN[Gc0l3T1a4bNaG9o3V1W4jNiG0P4YKnN^8X1VH\\OS4>\\KkNT8c1\\HTOS4?`KgNm7l1]HnNV4`0cKcNk7o1[HnNV4a0gK_Ni7R2YHnNV4b0kK[Ng7V2VHlNa0PO^2c1WMXNd7Z2UHkN`0\\OU2X1Q6b0WGjNb0AR2T1V6a0UGjN`0FETOj1l1l6b0SGhN`0KDROj1i1P7b0QGgN`00BQOk1f1S7c0nFfN?6BnNj1d1X7b0kFgN?9AmNk1a1Z7d0iFdN`0?_OjNl1_1]7d0gFeN=c0BgNi1^1b7d0dFdN=Z1[1^Oe7k1mFkN\\1[Og7j1kFmN]1ZOi7h1gFRO^1XOk7e1dFWO_1UOm7g4QHZKP8e4oG\\KR8d4lG]Kf7nMWGe6Q1^Kh7R5VHoJj7S5THmJl7V5QHjJo7X5oGhJa6mMgI]7EgJc6SNbIY7HdJf6YN\\IU7LbJg6aNVIP7O`Jj6gNQIk63^Jk6nNlHf67\\Jm6TOgHb68[JP7]7nHbHR7a7lH^HT7f7iHYHX7i7eHWH]7k7`HTHa7S:3N1O2N2N1O2N2N1102N1N3N1N3N1O2M2O2N1N3N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N\\MgIbGX6R8VJnGh5Q8\\JmGc5Q8aJnG^5P8fJnG\\5o7hJoG[5l7hJRH[5k7hJRH\\5j7fJUH\\5h7gJUH]5g7fJVH]5g7fJWHb5`7`J^Hl5V7WJhHS6m6PJPI[6d6iIYIY6e6jIYIV6f6lIXIT6h6mIXIS6g6nIXIR6h6oIWIR6h6PJWIP6h6QJWIo5h6SJWIn5h6SJXIl5h6UJWIl5h6UJXIj5h6WJWIj5h6WJWIj5h6WJXIh5h6YJWIh5h6YJXIf5g6]JWId5h6]JVId5j6]JSIf5l6[JRIg5m6ZJPIh5Y6[GgIn2Ni5W6^GiIj2Mj5V6cGjId2Nk5T6fGkIa2Oj5R6kGlI]2Ok5Q6lGoIZ2Nl5o5mGRJY2Ll5n5oGUJX2Jk5Z5dHhJl:n3^FPLe9U2mChN_?[1a@`N`?c1`@YNb?j1]@RNd?Q2\\@kMf?W2Z@eMg?_2Y@[Mj?g2g03M2O2M2O1O001N2O1O001N2O1OO10O0100O10O01O100O1dMa_OU1_`0hNn_Om0T`0oNR@n0o?nNT@Q1m?mNU@S1l?iNV@W1k?fNX@Z1i?bNZ@]1g?`N\\@`1Sa0O100O100O100O100O100O2O1N2N101N2N2M3N2M2O1N101N2O0O2OO101N1P]O^Oab0c0V]ODjb0j00O01000O0100OO2N1O2N1O2N2N1O2N1O2M3N5K6JT[j1"}, "label": "a girl wearing white t - shirt who is kicking the foot ball"}]} {"id": 552184, "image": "COCO_train2014_000000552184.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl , who is not wearing a headband , and is attempting to kick the soccer ball\"."}], "task": "refering", "answer_template": "The \"a girl , who is not wearing a headband , and is attempting to kick the soccer ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 80, 81, 82, 95, 96, 97, 109, 110, 111, 112, 113, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 153, 154, 155, 156, 157, 158, 159, 169, 170, 171, 172, 173, 174, 183, 184, 185, 186, 187, 188, 189, 190, 197, 198, 199, 200, 201, 202, 203, 204, 205, 211, 212, 213, 214, 215, 216, 217, 218, 220, 221, 229, 230, 231, 232, 233, 235, 236, 245, 246, 247, 248, 259, 260, 261, 274, 275, 276, 277, 288, 289, 290, 291, 292, 302, 303, 304, 306, 307, 308, 317, 318, 319], "bbox": [0.11896713615023474, 0.207046875, 0.7804694835680751, 0.9534218750000001], "iscrowd": 0, "area": 50324.13489999997, "rle": {"size": [640, 426], "counts": "iXP13jc03M3M3O1O100O1O100O100O1@B]]O?cb0EY]O;fb0JU]O7kb0?O1O100O100O1PCjNV7V1gHnNW7S1gHPOW7P1gHTOX7l0eHYOX7g0gH\\OX7d0eH@Z7?dHE[7:dHH\\78bHK]74bHN^72`H0`7O_H3a7L]H7c7I[H9d7G[H;e7EYH=f7CYH?e7CYH`0e7@[Ha0c7@]Ha0b7_O^Hb0`7_O`Hb0_7^OaHc0]7]OdHd0Z7]OeHe0Z7[OfHf0X7ZOiHh0U7XOkHi0S7WOnHj0Q7VOoHk0o6VOQIk0n6TOSIm0k6TOUIm0j6ROVIP1h6QOXIQ1f6oNZIR1]5hMlFV1h3S1S5QNSGl0j3T1i4ZN[Gc0l3T1a4bNaG9o3V1W4jNiG0P4YKnN^8X1VH\\OS4>\\KkNT8c1\\HTOS4?`KgNm7l1]HnNV4`0cKcNk7o1[HnNV4a0gK_Ni7R2YHnNV4b0kK[Ng7V2VHlNa0PO^2c1WMXNd7Z2UHkN`0\\OU2X1Q6b0WGjNb0AR2T1V6a0UGjN`0FETOj1l1l6b0SGhN`0KDROj1i1P7b0QGgN`00BQOk1f1S7c0nFfN?6BnNj1d1X7b0kFgN?9AmNk1a1Z7d0iFdN`0?_OjNl1_1]7d0gFeN=c0BgNi1^1b7d0dFdN=Z1[1^Oe7k1mFkN\\1[Og7j1kFmN]1ZOi7h1gFRO^1XOk7e1dFWO_1UOm7g4QHZKP8e4oG\\KR8d4lG]Kf7nMWGe6Q1^Kh7R5VHoJj7S5THmJl7V5QHjJo7X5oGhJa6mMgI]7EgJc6SNbIY7HdJf6YN\\IU7LbJg6aNVIP7O`Jj6gNQIk63^Jk6nNlHf67\\Jm6TOgHb68[JP7]7nHbHR7a7lH^HT7f7iHYHX7i7eHWH]7k7`HTHa7S:3N1O2N2N1O2N2N1102N1N3N1N3N1O2M2O2N1N3N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N\\MgIbGX6R8VJnGh5Q8\\JmGc5Q8aJnG^5P8fJnG\\5o7hJoG[5l7hJRH[5k7hJRH\\5j7fJUH\\5h7gJUH]5g7fJVH]5g7fJWHb5`7`J^Hl5V7WJhHS6m6PJPI[6d6iIYIY6e6jIYIV6f6lIXIT6h6mIXIS6g6nIXIR6h6oIWIR6h6PJWIP6h6QJWIo5h6SJWIn5h6SJXIl5h6UJWIl5h6UJXIj5h6WJWIj5h6WJWIj5h6WJXIh5h6YJWIh5h6YJXIf5g6]JWId5h6]JVId5j6]JSIf5l6[JRIg5m6ZJPIh5Y6[GgIn2Ni5W6^GiIj2Mj5V6cGjId2Nk5T6fGkIa2Oj5R6kGlI]2Ok5Q6lGoIZ2Nl5o5mGRJY2Ll5n5oGUJX2Jk5Z5dHhJl:n3^FPLe9U2mChN_?[1a@`N`?c1`@YNb?j1]@RNd?Q2\\@kMf?W2Z@eMg?_2Y@[Mj?g2g03M2O2M2O1O001N2O1O001N2O1OO10O0100O10O01O100O1dMa_OU1_`0hNn_Om0T`0oNR@n0o?nNT@Q1m?mNU@S1l?iNV@W1k?fNX@Z1i?bNZ@]1g?`N\\@`1Sa0O100O100O100O100O100O2O1N2N101N2N2M3N2M2O1N101N2O0O2OO101N1P]O^Oab0c0V]ODjb0j00O01000O0100OO2N1O2N1O2N2N1O2N1O2M3N5K6JT[j1"}, "label": "a girl , who is not wearing a headband , and is attempting to kick the soccer ball"}]} {"id": 3320, "image": "COCO_train2014_000000003320.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man dressed in all black with a black backpack\"."}], "task": "refering", "answer_template": "The \"a man dressed in all black with a black backpack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 213, 214, 215, 236, 237, 238, 259, 260, 261, 262, 282, 283, 284, 305, 306, 307, 328, 329, 330, 351, 352, 353, 375, 376], "bbox": [0.2674375, 0.4780416666666667, 0.400890625, 0.9983125], "iscrowd": 0, "area": 13212.171749999996, "rle": {"size": [480, 640], "counts": "]]`2Z1f==C7H7J3L4M4XFaMg6b2UI_Mj6f2PI]Mo6f2lH\\MT7f2hH]MV7g2eH\\MZ7f2bH\\M^7f2]H^Mb7d2XHaMg7a2THcMj7a2oGdMP8`2iGdMV8`2bGeM]8_2kFVNT9l3O1O0O2O001O0O1O2N1O2N1O1O3M2M7J6I6J4M1N2M3N3N1O1O1O1O1OKdHhH\\7U7eHmHZ7P7iHQIV7l6lHVIS7h6oHYIP7d6SI]Il6a6WI_Ii6Y6_IgIa6b5hHcJl0L\\6[5[JeJe5U5aJkJ_5m4iJTKV5i4mJWKS5f4oJ\\KP5`4TK`Km4\\4WKdKh4V4`K]K\\M6U7V4mKgKS4V4SLgKn3U4`3N3M3M2L5UMRE_1R;_NWEX1k:fNYEW1h:gN\\EU1g:hN]EU1d:iN`ES1b:kNcEo0a:nNdEj0a:TOeEa0b:]OdE8d:DbE2e:LZ2MVhc5"}, "label": "a man dressed in all black with a black backpack"}]} {"id": 3320, "image": "COCO_train2014_000000003320.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the black backpack strapped to his chest\"."}], "task": "refering", "answer_template": "The \"the man with the black backpack strapped to his chest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 213, 214, 215, 236, 237, 238, 259, 260, 261, 262, 282, 283, 284, 305, 306, 307, 328, 329, 330, 351, 352, 353, 375, 376], "bbox": [0.2674375, 0.4780416666666667, 0.400890625, 0.9983125], "iscrowd": 0, "area": 13212.171749999996, "rle": {"size": [480, 640], "counts": "]]`2Z1f==C7H7J3L4M4XFaMg6b2UI_Mj6f2PI]Mo6f2lH\\MT7f2hH]MV7g2eH\\MZ7f2bH\\M^7f2]H^Mb7d2XHaMg7a2THcMj7a2oGdMP8`2iGdMV8`2bGeM]8_2kFVNT9l3O1O0O2O001O0O1O2N1O2N1O1O3M2M7J6I6J4M1N2M3N3N1O1O1O1O1OKdHhH\\7U7eHmHZ7P7iHQIV7l6lHVIS7h6oHYIP7d6SI]Il6a6WI_Ii6Y6_IgIa6b5hHcJl0L\\6[5[JeJe5U5aJkJ_5m4iJTKV5i4mJWKS5f4oJ\\KP5`4TK`Km4\\4WKdKh4V4`K]K\\M6U7V4mKgKS4V4SLgKn3U4`3N3M3M2L5UMRE_1R;_NWEX1k:fNYEW1h:gN\\EU1g:hN]EU1d:iN`ES1b:kNcEo0a:nNdEj0a:TOeEa0b:]OdE8d:DbE2e:LZ2MVhc5"}, "label": "the man with the black backpack strapped to his chest"}]} {"id": 175370, "image": "COCO_train2014_000000175370.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man skating on the wooden rod\"."}], "task": "refering", "answer_template": "The \"a man skating on the wooden rod\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 100, 101, 102, 103, 104, 125, 126, 148, 149, 171, 172, 173, 194, 195, 196, 217, 219, 240, 242], "bbox": [0.34945312500000003, 0.20866323907455014, 0.545890625, 0.7438303341902314], "iscrowd": 0, "area": 7762.387449999998, "rle": {"size": [389, 640], "counts": "PWe21U<0O1oCOl;2RDOn;3100000VDLc;3]DNd;1\\DOc;2\\DNe;3YDMh;3WDNi;61O10O10000000O10O100000000000000000000000000000001O0O1000001O00000000O01mHE\\2;cMG\\28dMI\\27cMJ]26bMK]26aML_23aMN^23aMN^23SLDTMb1i6iNQLHTMa1k6gNoKJTM`1n6fNiKd1X4\\NbKj1_4VN\\Kn1e4SNWKn1k4SNPKX1nMoNT7JjJX1o5iNlIY1V6hNfIY1[6mN]IT1e6TORIk0P7[OhHe0h6[NgHV195W7nN\\HS3b7TMYHk2f7ZMWHf2`7aLfHY4[7kKbHS4`7mK`HQ4b7PL]Hm3e7ULZHi3h790O2O2M2N2O1N30O0010O010000001OO101O1O1O1N3N1O2N_M\\H`0d7VOkHe0U7ROZIh0e6POhIk0Y6ROlIl0[NbN]7`0\\JX1d5fN`J[1]5cNhJ_1R5`NRKR2Z4lMjKQ2W4mMlKQ2U4mMTLl1l3RN]Lf1c3YNeL`1\\3^NlL=nLHW6HmL8ZMJj5L]N1d1M]N2d1L^N3k6O00fX^3"}, "label": "a man skating on the wooden rod"}]} {"id": 175370, "image": "COCO_train2014_000000175370.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man performing skateboard trick\"."}], "task": "refering", "answer_template": "The \"a man performing skateboard trick\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 100, 101, 102, 103, 104, 125, 126, 148, 149, 171, 172, 173, 194, 195, 196, 217, 219, 240, 242], "bbox": [0.34945312500000003, 0.20866323907455014, 0.545890625, 0.7438303341902314], "iscrowd": 0, "area": 7762.387449999998, "rle": {"size": [389, 640], "counts": "PWe21U<0O1oCOl;2RDOn;3100000VDLc;3]DNd;1\\DOc;2\\DNe;3YDMh;3WDNi;61O10O10000000O10O100000000000000000000000000000001O0O1000001O00000000O01mHE\\2;cMG\\28dMI\\27cMJ]26bMK]26aML_23aMN^23aMN^23SLDTMb1i6iNQLHTMa1k6gNoKJTM`1n6fNiKd1X4\\NbKj1_4VN\\Kn1e4SNWKn1k4SNPKX1nMoNT7JjJX1o5iNlIY1V6hNfIY1[6mN]IT1e6TORIk0P7[OhHe0h6[NgHV195W7nN\\HS3b7TMYHk2f7ZMWHf2`7aLfHY4[7kKbHS4`7mK`HQ4b7PL]Hm3e7ULZHi3h790O2O2M2N2O1N30O0010O010000001OO101O1O1O1N3N1O2N_M\\H`0d7VOkHe0U7ROZIh0e6POhIk0Y6ROlIl0[NbN]7`0\\JX1d5fN`J[1]5cNhJ_1R5`NRKR2Z4lMjKQ2W4mMlKQ2U4mMTLl1l3RN]Lf1c3YNeL`1\\3^NlL=nLHW6HmL8ZMJj5L]N1d1M]N2d1L^N3k6O00fX^3"}, "label": "a man performing skateboard trick"}]} {"id": 240911, "image": "COCO_train2014_000000240911.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front train cart that ' s yellow and gray with a navy blue wavy banner painted on it\"."}], "task": "refering", "answer_template": "The \"the front train cart that ' s yellow and gray with a navy blue wavy banner painted on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290], "bbox": [0.017984375, 0.49437939110070256, 0.72265625, 0.8539344262295082], "iscrowd": 0, "area": 50320.76345, "rle": {"size": [427, 640], "counts": "SZ55Q=6K6N1N2O1gC]Og;e0WD\\Oi;e0UD]Oi;S101O0O101N101O0O2O0O1G9D=C in this image.", "objects": [{"patches": [163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290], "bbox": [0.017984375, 0.49437939110070256, 0.72265625, 0.8539344262295082], "iscrowd": 0, "area": 50320.76345, "rle": {"size": [427, 640], "counts": "SZ55Q=6K6N1N2O1gC]Og;e0WD\\Oi;e0UD]Oi;S101O0O101N101O0O2O0O1G9D=C in this image.", "objects": [{"patches": [28, 29, 30, 31, 32, 33, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 120, 121, 123, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 153, 169, 172, 173, 174, 175, 176, 192, 214], "bbox": [0.10785937500000001, 0.060671875, 0.6696562500000001, 0.43370312499999997], "iscrowd": 0, "area": 34253.86669999999, "rle": {"size": [640, 640], "counts": "ZW[11mc02O1N2O2M2O2N1N3N1N3N1O2M2O2M2O1O2N101N101N101N101N101N101N101N1O2O0O101N101N101N101N101N101N1O2O0O2O0O2O0O2O0O2O0O2O0O101N101N101N101N101O0O2O0O2O0O2O001N101N101O0O101N101N101O0O2O0O2O001N101N101O001O001N10001O001OmNW@oMi?l1`@QN_?k1j@PNV?o1]110001O000000000O2O000000001O00000000001O00000000001O000000001\\CcMc7]2ZHfMf7Z2VHgMm7Y2oGhMT8X2iGhMZ8X2bGhMb8Y2ZGgMi8Y2SGhMP9X2mFhMV9X2fFhM^9X2_FhMd9X2XFiMk9X2QFhMR:X2jEhMZ:X2bEiMa:W2\\EiMg:W2UEjMn:V2oDiMU;X2fDiM];X2_DhMd;Z2VDgMm;\\2mCcMX<_2aCbMb`3aA`Lb>b3XA_Lk>b3QA`LP?b3k@`LW?`3f@aL[?P42O1N3N1O1N3N1N2O1N3N1O11O010O1O0010O01O1O010O001O10O01O00100O001O010O0000010O00010O0BlKPAT4P?QLk@P4T?ULh@j3W?`001O0?A010O0010O01O0010O01O010O0010O010O10O0100O100O01O1O001O001O001O010O001O001O001O001O001O0010O01O001O1O001O001O001O0010O01O001O001O001O1O001O1O10O01O1O1O1O001O1O1O100O00100O1O100O010O1O100O00100O1O100O00102M3M3N2M3N2M3M2O0O2N100O2N101N1O101N101N1O101N1N3M2N2N3L4M6J6J6J7I4L4J7EoeS4"}, "label": "a black umbrella which a manis carrying"}]} {"id": 298262, "image": "COCO_train2014_000000298262.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black umbrella carried by the man in grey shirt\"."}], "task": "refering", "answer_template": "The \"black umbrella carried by the man in grey shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 31, 32, 33, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 120, 121, 123, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 153, 169, 172, 173, 174, 175, 176, 192, 214], "bbox": [0.10785937500000001, 0.060671875, 0.6696562500000001, 0.43370312499999997], "iscrowd": 0, "area": 34253.86669999999, "rle": {"size": [640, 640], "counts": "ZW[11mc02O1N2O2M2O2N1N3N1N3N1O2M2O2M2O1O2N101N101N101N101N101N101N101N1O2O0O101N101N101N101N101N101N1O2O0O2O0O2O0O2O0O2O0O2O0O101N101N101N101N101O0O2O0O2O0O2O001N101N101O0O101N101N101O0O2O0O2O001N101N101O001O001N10001O001OmNW@oMi?l1`@QN_?k1j@PNV?o1]110001O000000000O2O000000001O00000000001O00000000001O000000001\\CcMc7]2ZHfMf7Z2VHgMm7Y2oGhMT8X2iGhMZ8X2bGhMb8Y2ZGgMi8Y2SGhMP9X2mFhMV9X2fFhM^9X2_FhMd9X2XFiMk9X2QFhMR:X2jEhMZ:X2bEiMa:W2\\EiMg:W2UEjMn:V2oDiMU;X2fDiM];X2_DhMd;Z2VDgMm;\\2mCcMX<_2aCbMb`3aA`Lb>b3XA_Lk>b3QA`LP?b3k@`LW?`3f@aL[?P42O1N3N1O1N3N1N2O1N3N1O11O010O1O0010O01O1O010O001O10O01O00100O001O010O0000010O00010O0BlKPAT4P?QLk@P4T?ULh@j3W?`001O0?A010O0010O01O0010O01O010O0010O010O10O0100O100O01O1O001O001O001O010O001O001O001O001O001O0010O01O001O1O001O001O001O0010O01O001O001O001O1O001O1O10O01O1O1O1O001O1O1O100O00100O1O100O010O1O100O00100O1O100O00102M3M3N2M3N2M3M2O0O2N100O2N101N1O101N101N1O101N1N3M2N2N3L4M6J6J6J7I4L4J7EoeS4"}, "label": "black umbrella carried by the man in grey shirt"}]} {"id": 298262, "image": "COCO_train2014_000000298262.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing bow tie holding black umbrella\"."}], "task": "refering", "answer_template": "The \"man wearing bow tie holding black umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 121, 122, 123, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 304, 305, 306, 307, 308, 327, 328, 329, 330, 331, 350, 351, 352, 353, 354, 374, 375, 376, 397, 398, 399, 420, 421, 422, 443, 444, 445, 466, 467, 468, 489, 490, 513], "bbox": [0.18834375, 0.17478125, 0.46637500000000004, 0.9662499999999999], "iscrowd": 0, "area": 46767.08404999999, "rle": {"size": [640, 640], "counts": "[l[25fc06K4K6J6K5J6J6K4M4L4L4L4L3M4L4M3L4L3M4L4L4L4L4L3M4L4N2O1eB\\LW:e3cEdLZ:]3_ElL^:U3[EUM`:m2ZE[Mc:f2VEdMf:\\2TEmMh:U2QEUNk:l1nD]No:d1jDfNQ;\\1iDlNT;U1eDUOW;k0eD\\OW;f0eDAW;`0eDFX;;eDKV;7fDOW;a4L3L4M4L3L5L3L5L3aGbFV6b9]IkF^6Y9VISGn5XO]Ii99ZGo5YOdI`91cGo5ZOkI\\;j5PERJT;g5g1M3M3M3M4L3M3M4L4L3M4L3M4L2N001O1O1N100000O100O10000O100O10000O10000O100O102N1TOdKQB_4l=fKPB[4o=jKkAY4S>kKiAW4U>mKgAT4W>m0N1O2WO\\JQCg5f in this image.", "objects": [{"patches": [98, 99, 100, 121, 122, 123, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 234, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 304, 305, 306, 307, 308, 327, 328, 329, 330, 331, 350, 351, 352, 353, 354, 374, 375, 376, 397, 398, 399, 420, 421, 422, 443, 444, 445, 466, 467, 468, 489, 490, 513], "bbox": [0.18834375, 0.17478125, 0.46637500000000004, 0.9662499999999999], "iscrowd": 0, "area": 46767.08404999999, "rle": {"size": [640, 640], "counts": "[l[25fc06K4K6J6K5J6J6K4M4L4L4L4L3M4L4M3L4L3M4L4L4L4L4L3M4L4N2O1eB\\LW:e3cEdLZ:]3_ElL^:U3[EUM`:m2ZE[Mc:f2VEdMf:\\2TEmMh:U2QEUNk:l1nD]No:d1jDfNQ;\\1iDlNT;U1eDUOW;k0eD\\OW;f0eDAW;`0eDFX;;eDKV;7fDOW;a4L3L4M4L3L5L3L5L3aGbFV6b9]IkF^6Y9VISGn5XO]Ii99ZGo5YOdI`91cGo5ZOkI\\;j5PERJT;g5g1M3M3M3M4L3M3M4L4L3M4L3M4L2N001O1O1N100000O100O10000O100O10000O10000O100O102N1TOdKQB_4l=fKPB[4o=jKkAY4S>kKiAW4U>mKgAT4W>m0N1O2WO\\JQCg5f in this image.", "objects": [{"patches": [214, 215, 216, 217, 218, 219, 220, 221, 236, 237, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 263, 264, 265, 266, 267], "bbox": [0.264359375, 0.6308899297423888, 0.64515625, 0.7676814988290397], "iscrowd": 0, "area": 9811.64765, "rle": {"size": [427, 640], "counts": "\\hV29Q=101N2N1O2N100O2O001O001O0O101O001O0O101N1O100O100O100O100O100O100O101N1001O00000000000000000000001O000000000001O0001O00000000000000001O000000000000000010O000O10000000000000000000000000000000000O100000000000000O101O000000000O1000000000001N10000O2O000O2O0O10001N10001O0O101O00000O1000001O0O1000000O10001O000O100000000O2O00000000000000000000000000000O100000000000000001O000000001O0O1000001O000000001O0O1000001O000000001O0O1000001O001O001O001N10001O001O001O0O2M2O1O2N1N2O1O3L3N3M3L3N3M4K6Ke[n2"}, "label": "carrot sitting on cutting board"}]} {"id": 199958, "image": "COCO_train2014_000000199958.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the carrot between the bottom carrot and the green onion\"."}], "task": "refering", "answer_template": "The \"the carrot between the bottom carrot and the green onion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [214, 215, 216, 217, 218, 219, 220, 221, 236, 237, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 263, 264, 265, 266, 267], "bbox": [0.264359375, 0.6308899297423888, 0.64515625, 0.7676814988290397], "iscrowd": 0, "area": 9811.64765, "rle": {"size": [427, 640], "counts": "\\hV29Q=101N2N1O2N100O2O001O001O0O101O001O0O101N1O100O100O100O100O100O100O101N1001O00000000000000000000001O000000000001O0001O00000000000000001O000000000000000010O000O10000000000000000000000000000000000O100000000000000O101O000000000O1000000000001N10000O2O000O2O0O10001N10001O0O101O00000O1000001O0O1000000O10001O000O100000000O2O00000000000000000000000000000O100000000000000001O000000001O0O1000001O000000001O0O1000001O000000001O0O1000001O001O001O001N10001O001O001O0O2M2O1O2N1N2O1O3L3N3M3L3N3M4K6Ke[n2"}, "label": "the carrot between the bottom carrot and the green onion"}]} {"id": 199958, "image": "COCO_train2014_000000199958.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white tub of yogurt\"."}], "task": "refering", "answer_template": "The \"a white tub of yogurt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 77, 78, 79, 80, 81, 100, 101, 102, 103, 104, 124, 125, 126, 127, 147, 148, 149, 150, 171, 172], "bbox": [0.3881875, 0.17142857142857143, 0.570140625, 0.49091334894613586], "iscrowd": 0, "area": 12059.558849999998, "rle": {"size": [427, 640], "counts": "]`W3l0^<1O2N11OO2N1N3N1O2N1O2M2O:F:F:F:F9H9F7J1O1N2O001N2O001N101O0O2O001N101O0O2O000O2O001N10001O00001O00001O001N0100000000O100000000O10O100000O100000000O101O0O10000O100O100O100O10000O100O100O101N2N3M2M3N2M3N3M2M3QOXFXNj9]1hFYN\\9[1`1I7N3M3L5L3M5K4M1N1OVbb3"}, "label": "a white tub of yogurt"}]} {"id": 199958, "image": "COCO_train2014_000000199958.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yogurt behind onions\"."}], "task": "refering", "answer_template": "The \"yogurt behind onions\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 77, 78, 79, 80, 81, 100, 101, 102, 103, 104, 124, 125, 126, 127, 147, 148, 149, 150, 171, 172], "bbox": [0.3881875, 0.17142857142857143, 0.570140625, 0.49091334894613586], "iscrowd": 0, "area": 12059.558849999998, "rle": {"size": [427, 640], "counts": "]`W3l0^<1O2N11OO2N1N3N1O2N1O2M2O:F:F:F:F9H9F7J1O1N2O001N2O001N101O0O2O001N101O0O2O000O2O001N10001O00001O00001O001N0100000000O100000000O10O100000O100000000O101O0O10000O100O100O100O10000O100O100O101N2N3M2M3N2M3N3M2M3QOXFXNj9]1hFYN\\9[1`1I7N3M3L5L3M5K4M1N1OVbb3"}, "label": "yogurt behind onions"}]} {"id": 52518, "image": "COCO_train2014_000000052518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white couch\"."}], "task": "refering", "answer_template": "The \"white couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 277, 278, 279, 280, 281, 282, 283, 284, 300, 301, 302, 303, 304, 305], "bbox": [0.040234375, 0.48283333333333334, 0.44259375, 0.8111666666666667], "iscrowd": 0, "area": 29654.6683, "rle": {"size": [480, 640], "counts": "ba<6b>8H8I7RNYO\\En0d:ZOmDm0R;[OaDk0_;\\OSDk0l;U100O100K5D`CWOk in this image.", "objects": [{"patches": [185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 277, 278, 279, 280, 281, 282, 283, 284, 300, 301, 302, 303, 304, 305], "bbox": [0.040234375, 0.48283333333333334, 0.44259375, 0.8111666666666667], "iscrowd": 0, "area": 29654.6683, "rle": {"size": [480, 640], "counts": "ba<6b>8H8I7RNYO\\En0d:ZOmDm0R;[OaDk0_;\\OSDk0l;U100O100K5D`CWOk in this image.", "objects": [{"patches": [2, 25, 44, 45, 48, 67, 68, 82, 89, 90, 112], "bbox": [0.10234375, 0.014018691588785047, 0.97265625, 0.5607476635514018], "iscrowd": 0, "area": 1460.125, "rle": {"size": [214, 640], "counts": "ei=6^68I6J5K5L4M3M3Na0^O10M2O1O2gNlKLVcl10XbSN1O001O1O1O001OL4Ll[X1M[dgN2M3M2N3NO000010O000000000O1O2O1N2N2O1N2N2O10000OIXJBi5>ZJ_Of5a0610001N4M2N2M01O1O001O1N101O1O001O2N3M4K4M3MRg3"}, "label": "3 tennis rackets pointing upwards"}]} {"id": 535874, "image": "COCO_train2014_000000535874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an arm chair with spotted cushions\"."}], "task": "refering", "answer_template": "The \"an arm chair with spotted cushions\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 304, 305, 306], "bbox": [0.169359375, 0.4623541666666667, 0.44692187499999997, 0.8255833333333332], "iscrowd": 0, "area": 18589.90205, "rle": {"size": [480, 640], "counts": "V[c18h>4L1N2O1O1N101O1O1N101O1]BE^<=WC6`O001N2O001O1NO2001N10001N10001N10001O0O101O0O10001N10001N10001O0O101O0O101O0O101O0O101O000O7J7I1N101O0O2O001N101O000O01000O01000O0100O0I[FUKe9k4aFoJ`9o4;M3M3M3H8M3N3L5L3L4M3M3L5L3L4M3L5L3M0O101N10001N100O2O0O101O0O100O2O0O101O0O101N100O2O000O2O0O101N10001N100O2O000O21N2N100O0010O01O10O0O101O0O101O0O2O001O0O2I6O7I7I^dU5"}, "label": "an arm chair with spotted cushions"}]} {"id": 535874, "image": "COCO_train2014_000000535874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tacky animal print arm chair in a room with other tacky animal print items\"."}], "task": "refering", "answer_template": "The \"a tacky animal print arm chair in a room with other tacky animal print items\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 211, 212, 213, 214, 215, 234, 235, 236, 237, 238, 239, 240, 257, 258, 259, 260, 261, 262, 263, 280, 281, 282, 283, 284, 285, 286, 304, 305, 306], "bbox": [0.169359375, 0.4623541666666667, 0.44692187499999997, 0.8255833333333332], "iscrowd": 0, "area": 18589.90205, "rle": {"size": [480, 640], "counts": "V[c18h>4L1N2O1O1N101O1O1N101O1]BE^<=WC6`O001N2O001O1NO2001N10001N10001N10001O0O101O0O10001N10001N10001O0O101O0O101O0O101O0O101O000O7J7I1N101O0O2O001N101O000O01000O01000O0100O0I[FUKe9k4aFoJ`9o4;M3M3M3H8M3N3L5L3L4M3M3L5L3L4M3L5L3M0O101N10001N100O2O0O101O0O100O2O0O101O0O101N100O2O000O2O0O101N10001N100O2O000O21N2N100O0010O01O10O0O101O0O101O0O2O001O0O2I6O7I7I^dU5"}, "label": "a tacky animal print arm chair in a room with other tacky animal print items"}]} {"id": 535874, "image": "COCO_train2014_000000535874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two black and white color pillows on the purple color table toy\"."}], "task": "refering", "answer_template": "The \"two black and white color pillows on the purple color table toy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 274, 294, 295, 296, 318], "bbox": [0.602765625, 0.5323541666666667, 0.9370937500000001, 0.7942916666666666], "iscrowd": 0, "area": 13367.381449999999, "rle": {"size": [480, 640], "counts": "TWe54j>3N1O2BHPB:n=IPB9n=IPB8o=IPB9n=IPB8o==O2N>C001N101N101O0O101O0O2O0O101O001O0O101O001N101O00001N100000O01000O1000O010000O1O010O1O1O1001O00010O00001O01O0001O01O01O000010O0001O0001O01O00010O00001O01O0001O00010O000010O0001O0001O01O000010O0001O01O01O0000^NUCo0lLg]b0"}, "label": "two black and white color pillows on the purple color table toy"}]} {"id": 535874, "image": "COCO_train2014_000000535874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white cheetah print cushions on a bench\"."}], "task": "refering", "answer_template": "The \"white cheetah print cushions on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 274, 294, 295, 296, 318], "bbox": [0.602765625, 0.5323541666666667, 0.9370937500000001, 0.7942916666666666], "iscrowd": 0, "area": 13367.381449999999, "rle": {"size": [480, 640], "counts": "TWe54j>3N1O2BHPB:n=IPB9n=IPB8o=IPB9n=IPB8o==O2N>C001N101N101O0O101O0O2O0O101O001O0O101O001N101O00001N100000O01000O1000O010000O1O010O1O1O1001O00010O00001O01O0001O01O01O000010O0001O0001O01O00010O00001O01O0001O00010O000010O0001O0001O01O000010O0001O01O01O0000^NUCo0lLg]b0"}, "label": "white cheetah print cushions on a bench"}]} {"id": 200010, "image": "COCO_train2014_000000200010.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair with the white cushion on the left\"."}], "task": "refering", "answer_template": "The \"chair with the white cushion on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 210, 211, 212, 233, 234, 235, 236, 237, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 306, 326, 327], "bbox": [0.11881250000000002, 0.5354929577464789, 0.35785937500000004, 0.9806338028169014], "iscrowd": 0, "area": 14071.058149999997, "rle": {"size": [426, 640], "counts": "Qko08Q=3hGGZ3XLDg3?ULCk3?QLBP4`0lKAV4`0fKA[4`0aKAa4a0[K@f4b0VK^Ol4d0PK]OQ5e0kJ\\OV5f0fJZO]5g0_JZOb5h0[JWOg5k0VJUOk5n0QJROP6T1iInNV6X1dIiN[6]1^IdNc6a1VIaNi6e1PI\\NP7j1iHXNV7m1eHTNZ7R2_HoMa7S2\\HoMd7Q2ZHPNf7Q2YHPNf7R2WHoMi7R2UHPNj7Q2THQNk7P2SHQNm7P2RHQNn7P2oGQNQ8P2mGRNR8o1lGSNS8n1kGSNU8n1jGSNU8n1iGSNW8a300O0010O01O10O01O010O10O01O101N2O1N2001O0O2L4L3M4L3M4K5L3M4L4L3M4K4M4L4L3M4K4M4L1O001O1N2O1O001O1O1N2O001O100O100O101N10000O100O2O0O100O100O101O0O100O100O101ZOnDQOS;n0PEPOP;o0REPOn:o0TEoNm:o0VEPOj:o0XEoNi:P1YEoNh:o0ZEoNg:P1[EoNf:n0]EQOd:m0^ESOa:l0aETO_:j0bEWO^:g0dEYO[:e0hE[OX:c0iE^OW:`0kE@T:?mECR:;PFEP:9QFHn98SFHm96XFFh99_FBa9=f1O101N1O2O2M3N3LaeZ5"}, "label": "chair with the white cushion on the left"}]} {"id": 200010, "image": "COCO_train2014_000000200010.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair with the white cusion on the back nearest the brick wall and plants\"."}], "task": "refering", "answer_template": "The \"the chair with the white cusion on the back nearest the brick wall and plants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 210, 211, 212, 233, 234, 235, 236, 237, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 306, 326, 327], "bbox": [0.11881250000000002, 0.5354929577464789, 0.35785937500000004, 0.9806338028169014], "iscrowd": 0, "area": 14071.058149999997, "rle": {"size": [426, 640], "counts": "Qko08Q=3hGGZ3XLDg3?ULCk3?QLBP4`0lKAV4`0fKA[4`0aKAa4a0[K@f4b0VK^Ol4d0PK]OQ5e0kJ\\OV5f0fJZO]5g0_JZOb5h0[JWOg5k0VJUOk5n0QJROP6T1iInNV6X1dIiN[6]1^IdNc6a1VIaNi6e1PI\\NP7j1iHXNV7m1eHTNZ7R2_HoMa7S2\\HoMd7Q2ZHPNf7Q2YHPNf7R2WHoMi7R2UHPNj7Q2THQNk7P2SHQNm7P2RHQNn7P2oGQNQ8P2mGRNR8o1lGSNS8n1kGSNU8n1jGSNU8n1iGSNW8a300O0010O01O10O01O010O10O01O101N2O1N2001O0O2L4L3M4L3M4K5L3M4L4L3M4K4M4L4L3M4K4M4L1O001O1N2O1O001O1O1N2O001O100O100O101N10000O100O2O0O100O100O101O0O100O100O101ZOnDQOS;n0PEPOP;o0REPOn:o0TEoNm:o0VEPOj:o0XEoNi:P1YEoNh:o0ZEoNg:P1[EoNf:n0]EQOd:m0^ESOa:l0aETO_:j0bEWO^:g0dEYO[:e0hE[OX:c0iE^OW:`0kE@T:?mECR:;PFEP:9QFHn98SFHm96XFFh99_FBa9=f1O101N1O2O2M3N3LaeZ5"}, "label": "the chair with the white cusion on the back nearest the brick wall and plants"}]} {"id": 200010, "image": "COCO_train2014_000000200010.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lawn chair , its back is directly facing the camera\"."}], "task": "refering", "answer_template": "The \"a lawn chair , its back is directly facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 331, 332, 333, 334, 335], "bbox": [0.37298437500000003, 0.6724178403755868, 0.6182812500000001, 1.0], "iscrowd": 0, "area": 16411.061450000005, "rle": {"size": [426, 640], "counts": "dfS35V=5J3M7I8H7I7I7J7H7I7I7I8H7I7J6I001O001jFfLZ8Z3eGfL\\8[3aGfLa8Y3]GhLd8Y3ZGgLg8Y3WGhLj8Y3TGfLn8Z3PGgLQ9Z3lFgLU9Y3jFgLX9Y3eFhL\\9X3bFiL_9X3_FhLb9^32O1O1O2N1O100001O0001O0001O00000010O00000001O0001O0001O000000010O0000001O000001O01O0000000010O00000001O0001O01O0000000010O00000001O0001O0001O00000010O00000001O0001O0001O000000010O0000001O0001O0001O000000010O000O1O1O2N1O1O1O2N1d0\\O3E8F;F9F;E:F:E in this image.", "objects": [{"patches": [238, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 331, 332, 333, 334, 335], "bbox": [0.37298437500000003, 0.6724178403755868, 0.6182812500000001, 1.0], "iscrowd": 0, "area": 16411.061450000005, "rle": {"size": [426, 640], "counts": "dfS35V=5J3M7I8H7I7I7J7H7I7I7I8H7I7J6I001O001jFfLZ8Z3eGfL\\8[3aGfLa8Y3]GhLd8Y3ZGgLg8Y3WGhLj8Y3TGfLn8Z3PGgLQ9Z3lFgLU9Y3jFgLX9Y3eFhL\\9X3bFiL_9X3_FhLb9^32O1O1O2N1O100001O0001O0001O00000010O00000001O0001O0001O000000010O0000001O000001O01O0000000010O00000001O0001O01O0000000010O00000001O0001O0001O00000010O00000001O0001O0001O000000010O0000001O0001O0001O000000010O000O1O1O2N1O1O1O2N1d0\\O3E8F;F9F;E:F:E in this image.", "objects": [{"patches": [203, 204, 205, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 292, 293, 294, 295, 316, 317, 318], "bbox": [0.7081875, 0.5242253521126761, 0.948453125, 0.9290140845070423], "iscrowd": 0, "area": 12708.2841, "rle": {"size": [426, 640], "counts": "Whl52P=:G9M2O2N2N1O2N2N2N1N3N2dDjN`:W1[EnNd:T1VEPOj:R1PESOo:o0kDUOT;^11N2M4L3002N2O01000O10000TFUNY8j1_G`N_8a1_GaNa8a1[GaNe8a1WGbNh8a1SGaNl8b1QG`Nn8b1oF_NQ9b1mF_NS9c1kF^NT9c1jFQNEL`9V2hFoMNEZ9^2fFlMi9T2UFlMl9V2QFkMo9V2PFiMQ:h20O1O100O1O10001O000000000000000000001O0000000001O0000000L4N2N2N2N2N2N2N2NO1G91O1O1O1O1O1O2N2N1O2N2N2N2N1O2N200000O2I6M301O0O10001O0O101E]FmLc9R3;O2M2O101O0O101O0O100O2O0FcEgM]:T2iEjMY:R2jEnMV:o1nEPNR:n1PFQNR:k1RFTNn9i1UFWNk9g1XFWNj9e1ZFZNf9c1R1N3L3M3N3L3N2M4L3N2M4L3N2M4L3N2M4Ll_="}, "label": "an outdoor patio chair placed closest to the fence"}]} {"id": 339283, "image": "COCO_train2014_000000339283.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a boy wearing white pants and holding a baseball bat\"."}], "task": "refering", "answer_template": "The \"this is a boy wearing white pants and holding a baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 84, 85, 86, 87, 88, 107, 108, 109, 110, 130, 131, 132, 153, 154, 155, 176, 177, 178, 199, 200, 201, 202, 222, 223, 224, 225, 244, 245, 246, 247, 248, 267, 268, 270, 271, 289, 290, 294, 295, 312, 313, 317, 318], "bbox": [0.579546875, 0.15767058823529412, 0.8537968750000001, 0.9239058823529411], "iscrowd": 0, "area": 21968.019299999996, "rle": {"size": [425, 640], "counts": "[[j45m<7M3L5K4L4L4L4O1O1O1O1O1O1O1O1O1O1M3M3L4M3M3M3O1N2N2O1N2NO2M2N2N3M2N2N201O00001O000010O001O0O2O0O2O001_NnNjFS1U9a1M2N3M2O2M2N3N1N3_LiLnLY3S3lLfLV3[3oLZLU3h3oLmKW3T4mLaKX3`4h21N2O0O2O1N101O1N101N2O0O2O1N2O0O2O1N101N2O0O2O1N10100100O100O100O101N2N3N1N2O1N2N2N2O1N2N2N2O1NXIXJQ6f5iIcJW6Z5fIkJZ6T5cIPK]6n4bIUK^6i4aIZK_6d4_I`Ka6_4]IbKh1SOn2\\6eLoJ`2n4SMeKj2X4UMlKOmMj1U6VNQLMnMl1n5WNWLJnMo1h5WN]LFPNR2a5WNbLDQNT2Z5XNhLARNW2R5XNoL]OSN[2l4WNTM[OTN`2b4UN]MXOUN`3]3XMIf27YMMd22\\M2b2L^M8_2GaM<]2DbM`0\\2^OdMf0Y2YOgMk0V2UOiMn0U2QOkMS1S2kNmMY1P2gNoM\\1o1cNRN`1k1_NVNd1h1[NXNh1e1WN\\Nk1b1TN_No1_1oMbNT2[1lMeNW2X1hMhN[2V1aMnNb2P1^MoNe2n0]MPOe2n0]MPOf2n0ZMQOi2l0YMoNm2n0UMlNQ3R1PMhNX3V1iLdN^3Y1\\4M3M2N3L4L4L3M4L4L4L3M4MbdV1"}, "label": "this is a boy wearing white pants and holding a baseball bat"}]} {"id": 339283, "image": "COCO_train2014_000000339283.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kid hitting a baseball with his bat\"."}], "task": "refering", "answer_template": "The \"a kid hitting a baseball with his bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 84, 85, 86, 87, 88, 107, 108, 109, 110, 130, 131, 132, 153, 154, 155, 176, 177, 178, 199, 200, 201, 202, 222, 223, 224, 225, 244, 245, 246, 247, 248, 267, 268, 270, 271, 289, 290, 294, 295, 312, 313, 317, 318], "bbox": [0.579546875, 0.15767058823529412, 0.8537968750000001, 0.9239058823529411], "iscrowd": 0, "area": 21968.019299999996, "rle": {"size": [425, 640], "counts": "[[j45m<7M3L5K4L4L4L4O1O1O1O1O1O1O1O1O1O1M3M3L4M3M3M3O1N2N2O1N2NO2M2N2N3M2N2N201O00001O000010O001O0O2O0O2O001_NnNjFS1U9a1M2N3M2O2M2N3N1N3_LiLnLY3S3lLfLV3[3oLZLU3h3oLmKW3T4mLaKX3`4h21N2O0O2O1N101O1N101N2O0O2O1N2O0O2O1N101N2O0O2O1N10100100O100O100O101N2N3N1N2O1N2N2N2O1N2N2N2O1NXIXJQ6f5iIcJW6Z5fIkJZ6T5cIPK]6n4bIUK^6i4aIZK_6d4_I`Ka6_4]IbKh1SOn2\\6eLoJ`2n4SMeKj2X4UMlKOmMj1U6VNQLMnMl1n5WNWLJnMo1h5WN]LFPNR2a5WNbLDQNT2Z5XNhLARNW2R5XNoL]OSN[2l4WNTM[OTN`2b4UN]MXOUN`3]3XMIf27YMMd22\\M2b2L^M8_2GaM<]2DbM`0\\2^OdMf0Y2YOgMk0V2UOiMn0U2QOkMS1S2kNmMY1P2gNoM\\1o1cNRN`1k1_NVNd1h1[NXNh1e1WN\\Nk1b1TN_No1_1oMbNT2[1lMeNW2X1hMhN[2V1aMnNb2P1^MoNe2n0]MPOe2n0]MPOf2n0ZMQOi2l0YMoNm2n0UMlNQ3R1PMhNX3V1iLdN^3Y1\\4M3M2N3L4L4L3M4L4L4L3M4MbdV1"}, "label": "a kid hitting a baseball with his bat"}]} {"id": 404826, "image": "COCO_train2014_000000404826.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green table cloth spread on the table\"."}], "task": "refering", "answer_template": "The \"a green table cloth spread on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 90, 91, 114], "bbox": [0.486265625, 0.0037083333333333334, 1.0, 0.2776666666666667], "iscrowd": 0, "area": 16647.832499999993, "rle": {"size": [480, 640], "counts": "eXb41o>00000O100000000O1000001O0O100000000O100000001N100000000O2O000000001N100000000O2O0000000O2O0000000O101O00000O101O00000O10001O000O10001O000O1000001O000O10001O000O1000001O0O1000001O0O100000001N100000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000001O01O000000001O0000001O001O001O001O001O001O001O001O001O001O00001O001O001O001O0000000000001O00000000001O00000000001O0000000000001O00000000001O0000000000001O00000000010O0000001O0000001O0000001O0000001O0000001O00001O001O1O2N1O1O2O0O1O2N1O1O2N1O1O2N100O1O1O1O1O001O1O1O100O1O1O1O001O1O1O1O100O1O1O001O1O1O1O1O100O1O001O1O1O1O1O1OB?[Od0\\Od0]Oc0\\Od0\\O:"}, "label": "a green table cloth spread on the table"}]} {"id": 314725, "image": "COCO_train2014_000000314725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra on the left in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the zebra on the left in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 99, 112, 113, 114, 117, 129, 130, 131, 147, 148, 149, 165, 166, 167, 183, 184, 185], "bbox": [0.1819, 0.47525525525525525, 0.5741, 0.9318618618618618], "iscrowd": 0, "area": 6732.246849999996, "rle": {"size": [333, 500], "counts": "\\lm09o94L4L4XOA_Gc0`8GTG=k8b0O100O10000N3LO1L44K6K4YOg0J6J6K5O1O2N1N2OD in this image.", "objects": [{"patches": [84, 85, 101, 102, 103, 104, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 149, 150, 151, 152, 153, 154, 155, 167, 168, 169, 170, 171, 172, 173, 185, 186, 187, 190], "bbox": [0.30278, 0.38450450450450446, 0.8180599999999999, 0.9215015015015016], "iscrowd": 0, "area": 21512.96764999998, "rle": {"size": [333, 500], "counts": "i[a13T:=iMIiI9bN_OW7f1lHmNIGm6e1WIUOd6Q1XIROe6[201O001OO2N1N2O1N2N3N1O2M2O1011N2O1N101N@QIdMo6[2SIdMl6\\2c0O1NQIeMo5[2nIjMQ6T2lIRNS6m1gI[NW6g1bIaN\\6i2O2N3M3M3M1O1OO100000000\\OWJdLi5T3eJfL[5X3n0N2N2N2N2M3M3M3N2M4M2N2N2N2N2001O010O001O001O010O001O001O0100O1O101N101NO1L4M30000001O0000000000000000000000000O1000001O1O1O1O1O1O00JiMWHU2j7mMTHS2l77000O1000000O1000gIfM_4Y2^KkMb4U2\\KmMd4S2[KmMf4S2XKoMg4Q2YKPNg4P2WKRNi4n1QKWNP5i1gJ`NY5`1bJdN_5\\1^JfNc5Z1ZJhNg5X1VJkNi5V1TJkNn5V1hIQOZ6`21O1O15KO1N2N2O1N1O2O1M3N2L4H8J6N2N2N2N00001O0001O02N100O2N10001O01O01O001N1O2N1O2O0O2M2N3O001O001O001O001O010O1O001O1OGSNTHl1k7WNSHi1n7YNPHf1R8\\NkGc1V892M4M1N3M2O1O2N10O2N1O2O010O00100O010O01000O1000O01001O0WN_Ge1`8YNdGe1c83NO010N100O2N1O1O1O1bNVGW1R9N2N2N2N2N2M3N2M4L4K5L6Ik\\m0"}, "label": "a zebra with two birds perched on its side"}]} {"id": 314725, "image": "COCO_train2014_000000314725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra with a bunch of birds hanging out on it ' s back\"."}], "task": "refering", "answer_template": "The \"a zebra with a bunch of birds hanging out on it ' s back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 101, 102, 103, 104, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 149, 150, 151, 152, 153, 154, 155, 167, 168, 169, 170, 171, 172, 173, 185, 186, 187, 190], "bbox": [0.30278, 0.38450450450450446, 0.8180599999999999, 0.9215015015015016], "iscrowd": 0, "area": 21512.96764999998, "rle": {"size": [333, 500], "counts": "i[a13T:=iMIiI9bN_OW7f1lHmNIGm6e1WIUOd6Q1XIROe6[201O001OO2N1N2O1N2N3N1O2M2O1011N2O1N101N@QIdMo6[2SIdMl6\\2c0O1NQIeMo5[2nIjMQ6T2lIRNS6m1gI[NW6g1bIaN\\6i2O2N3M3M3M1O1OO100000000\\OWJdLi5T3eJfL[5X3n0N2N2N2N2M3M3M3N2M4M2N2N2N2N2001O010O001O001O010O001O001O0100O1O101N101NO1L4M30000001O0000000000000000000000000O1000001O1O1O1O1O1O00JiMWHU2j7mMTHS2l77000O1000000O1000gIfM_4Y2^KkMb4U2\\KmMd4S2[KmMf4S2XKoMg4Q2YKPNg4P2WKRNi4n1QKWNP5i1gJ`NY5`1bJdN_5\\1^JfNc5Z1ZJhNg5X1VJkNi5V1TJkNn5V1hIQOZ6`21O1O15KO1N2N2O1N1O2O1M3N2L4H8J6N2N2N2N00001O0001O02N100O2N10001O01O01O001N1O2N1O2O0O2M2N3O001O001O001O001O010O1O001O1OGSNTHl1k7WNSHi1n7YNPHf1R8\\NkGc1V892M4M1N3M2O1O2N10O2N1O2O010O00100O010O01000O1000O01001O0WN_Ge1`8YNdGe1c83NO010N100O2N1O1O1O1bNVGW1R9N2N2N2N2N2M3N2M4L4K5L6Ik\\m0"}, "label": "a zebra with a bunch of birds hanging out on it ' s back"}]} {"id": 208243, "image": "COCO_train2014_000000208243.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two sporty - looking motorcycles next to each other on a road\"."}], "task": "refering", "answer_template": "The \"two sporty - looking motorcycles next to each other on a road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 64, 83, 84, 85, 86, 87, 104, 105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 143, 144, 145, 146, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 272, 273, 280, 281, 282, 283, 284, 285, 286, 287, 288, 304, 305], "bbox": [0.127140625, 0.13454965357967666, 0.9414374999999999, 0.8727944572748269], "iscrowd": 0, "area": 88821.43964999997, "rle": {"size": [433, 640], "counts": "i`R17Y=1N3N1O2@DhC=W in this image.", "objects": [{"patches": [61, 62, 63, 64, 83, 84, 85, 86, 87, 104, 105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 143, 144, 145, 146, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 272, 273, 280, 281, 282, 283, 284, 285, 286, 287, 288, 304, 305], "bbox": [0.127140625, 0.13454965357967666, 0.9414374999999999, 0.8727944572748269], "iscrowd": 0, "area": 88821.43964999997, "rle": {"size": [433, 640], "counts": "i`R17Y=1N3N1O2@DhC=W in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 75, 92, 93], "bbox": [0.019109375, 0.0023556581986143186, 0.29609375, 0.28942263279445724], "iscrowd": 0, "area": 16812.422450000002, "rle": {"size": [433, 640], "counts": "WU5R1o9`2000000000000000000O100000000000000000000000000O100000000000000000000000000O10O1O1O1O1O1N2O1O1O1O1O1O1O001O1O1O1O1O1O1O10000O100000O1000O10000000000O1000000000O010000000000O10000000000O10O10000000O10000000000O100000O1000O10000000000O1000000000O010000000000O10000000000O10O10000000O10001O1O1O1O1001O000000001O0000O1O1O1O1O1O1O1O1O1O1O1O2N1O1N8C=D in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 75, 92, 93], "bbox": [0.019109375, 0.0023556581986143186, 0.29609375, 0.28942263279445724], "iscrowd": 0, "area": 16812.422450000002, "rle": {"size": [433, 640], "counts": "WU5R1o9`2000000000000000000O100000000000000000000000000O100000000000000000000000000O10O1O1O1O1O1N2O1O1O1O1O1O1O001O1O1O1O1O1O1O10000O100000O1000O10000000000O1000000000O010000000000O10000000000O10O10000000O10000000000O100000O1000O10000000000O1000000000O010000000000O10000000000O10O10000000O10001O1O1O1O1001O000000001O0000O1O1O1O1O1O1O1O1O1O1O1O2N1O1N8C=D in this image.", "objects": [{"patches": [11, 12, 33, 34, 35, 54, 55, 56, 57, 58, 59, 77, 78, 79, 80, 81, 82, 83, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 169, 170, 171, 186, 187], "bbox": [0.03496875, 0.022471131639722863, 0.6370312499999999, 0.5954965357967668], "iscrowd": 0, "area": 35766.00935000001, "rle": {"size": [433, 640], "counts": "^k92^=3M200O1O2N100O2N4L3N3L4L4L3N3L4L5K6K5J6J4L3N2M3M3M3L3M4M3L4L4M3L4L3N1N2O1N2N3N1N2O1N2O1N2N2O1N2O1N1O100O100O100O1O100OM4N101O0O101N101N2O1N2O1O1N2N2N2M4M2N2M3O10000000000000001O00000000000000000000001O0ROgEcNZ:]1gEbNY:]1iEaNX:_1iE`NW:_1kE_NV:`1lE_NT:a1mE]NT:b1mE^NS:b1nE\\NS:c1oE[NR:e1oEZNQ:e1QFYNP:f1n000O100O11O1O1O1O2N1O1O2N1O1O2N100010O0010O01O0O1N2N2O1N101N2N2O1N2O1N1000000000001N100000000000000O1000000000O10O0100O100O10000O010O100O100O10O0100O100O10000O010O100O10000O101O0O2O1O0O2O1O0O2O1N2O001N2O001M3VO\\MdFe2Y9^MfFd2V9^MjFd2S9^MlFc2R9^MmFe2Q9\\MnFf2o8\\MPGe2n8\\MRGf2k8\\MTGf2j8ZMUGi2h8YMWGh2h8YMWGi2g8WMYGk2e8VMYGl2f8TMZGm2e8TMZGl2g8SMYGn2f8SMYGn2f8RMYGo2g8QMWGR3i8nLUGT3k8kLTGV3m8g01O000O10000O1000000O10000O01O1O010O1O1O001O1O10O01O1O1O1O0O2O1O1N101O1O1N2O0I8M3N2N2N1O2N2N2O0010000000O100000O01N2O1N2O1N2O1N1_KmGe3T8ZLnGc3T8[LoGc3R8\\LQH`3Q8^LRH_3P8`LRHR2J^NU8_OSHQ23TNl7ITHQ2;kMa74VHn1e0bMW7>WHn1U9QNmFm1T9RNnFk1T9SNoFk1R9TNQGi1P9UNSGh1o8WNSGg1m8YNUGf1k8XNXGR2]8mMeG]2P8aMSH^2m7aMUH^2_901O0O1000001O000000001O000O101O2N3L3N2M2O0O010O100O10O0100O1O010O103L5L3L4gNmCn0] in this image.", "objects": [{"patches": [11, 12, 33, 34, 35, 54, 55, 56, 57, 58, 59, 77, 78, 79, 80, 81, 82, 83, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 169, 170, 171, 186, 187], "bbox": [0.03496875, 0.022471131639722863, 0.6370312499999999, 0.5954965357967668], "iscrowd": 0, "area": 35766.00935000001, "rle": {"size": [433, 640], "counts": "^k92^=3M200O1O2N100O2N4L3N3L4L4L3N3L4L5K6K5J6J4L3N2M3M3M3L3M4M3L4L4M3L4L3N1N2O1N2N3N1N2O1N2O1N2N2O1N2O1N1O100O100O100O1O100OM4N101O0O101N101N2O1N2O1O1N2N2N2M4M2N2M3O10000000000000001O00000000000000000000001O0ROgEcNZ:]1gEbNY:]1iEaNX:_1iE`NW:_1kE_NV:`1lE_NT:a1mE]NT:b1mE^NS:b1nE\\NS:c1oE[NR:e1oEZNQ:e1QFYNP:f1n000O100O11O1O1O1O2N1O1O2N1O1O2N100010O0010O01O0O1N2N2O1N101N2N2O1N2O1N1000000000001N100000000000000O1000000000O10O0100O100O10000O010O100O100O10O0100O100O10000O010O100O10000O101O0O2O1O0O2O1O0O2O1N2O001N2O001M3VO\\MdFe2Y9^MfFd2V9^MjFd2S9^MlFc2R9^MmFe2Q9\\MnFf2o8\\MPGe2n8\\MRGf2k8\\MTGf2j8ZMUGi2h8YMWGh2h8YMWGi2g8WMYGk2e8VMYGl2f8TMZGm2e8TMZGl2g8SMYGn2f8SMYGn2f8RMYGo2g8QMWGR3i8nLUGT3k8kLTGV3m8g01O000O10000O1000000O10000O01O1O010O1O1O001O1O10O01O1O1O1O0O2O1O1N101O1O1N2O0I8M3N2N2N1O2N2N2O0010000000O100000O01N2O1N2O1N2O1N1_KmGe3T8ZLnGc3T8[LoGc3R8\\LQH`3Q8^LRH_3P8`LRHR2J^NU8_OSHQ23TNl7ITHQ2;kMa74VHn1e0bMW7>WHn1U9QNmFm1T9RNnFk1T9SNoFk1R9TNQGi1P9UNSGh1o8WNSGg1m8YNUGf1k8XNXGR2]8mMeG]2P8aMSH^2m7aMUH^2_901O0O1000001O000000001O000O101O2N3L3N2M2O0O010O100O10O0100O1O010O103L5L3L4gNmCn0] in this image.", "objects": [{"patches": [211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 305, 306, 310, 311, 312, 313, 328, 329, 351, 352], "bbox": [0.0, 0.5685497835497836, 0.73971875, 0.9820129870129871], "iscrowd": 0, "area": 35565.46029999999, "rle": {"size": [462, 640], "counts": "f9f1g<1N2O1O2O03M000001O0O100000000O100000000O100000000O10000000001N100000000O100000000O100000000O100000000O2O000000000O100000000O100000000O100000000O2O0000000O100000000O10000000000O100000000O2O0000000O100000000O100000000O10000000000O101O00000O100000000O100000000O100000000O10001N1O100O100O100O100O1O100O100O100O1O100O101N100O100O1O100O100O100O100O10000000000000001O0001OmFUMc6k2bHPN^7P2hGiNY8f30000000000001O0000000000000000001O00000]MnGjNR8V1bHVN^7j1VIbMj6^2jInLV6R3c20001O000000010O001O00001O001O001O000010O01O00001O001O001O00001O010O00001O001O00001O001O0010O0001O001O00001O001O001O00010O001O00001O00001O001O00010O00001O001O00001O00001O010O00001O00001O001O00001O01O01O001O00001O00001O001O01O01O00001O001O00001O00001O001O00001O00001O001O00001O000O2O001O0000001O0000001O00001O0000001O0000001O0000001O0000001O0000001O0000001O00001O0000001N1000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O001N10001O001N101O001N10001O0O2O001O000O2O001O0O2O00001O0O2O001O0O2O00001O0O2O001O0O101O001OjoZ2"}, "label": "straw - based umbrellas sit near a waterway"}]} {"id": 380284, "image": "COCO_train2014_000000380284.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umbrella in front\"."}], "task": "refering", "answer_template": "The \"umbrella in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 305, 306, 310, 311, 312, 313, 328, 329, 351, 352], "bbox": [0.0, 0.5685497835497836, 0.73971875, 0.9820129870129871], "iscrowd": 0, "area": 35565.46029999999, "rle": {"size": [462, 640], "counts": "f9f1g<1N2O1O2O03M000001O0O100000000O100000000O100000000O10000000001N100000000O100000000O100000000O100000000O2O000000000O100000000O100000000O100000000O2O0000000O100000000O10000000000O100000000O2O0000000O100000000O100000000O10000000000O101O00000O100000000O100000000O100000000O10001N1O100O100O100O100O1O100O100O100O1O100O101N100O100O1O100O100O100O100O10000000000000001O0001OmFUMc6k2bHPN^7P2hGiNY8f30000000000001O0000000000000000001O00000]MnGjNR8V1bHVN^7j1VIbMj6^2jInLV6R3c20001O000000010O001O00001O001O001O000010O01O00001O001O001O00001O010O00001O001O00001O001O0010O0001O001O00001O001O001O00010O001O00001O00001O001O00010O00001O001O00001O00001O010O00001O00001O001O00001O01O01O001O00001O00001O001O01O01O00001O001O00001O00001O001O00001O00001O001O00001O000O2O001O0000001O0000001O00001O0000001O0000001O0000001O0000001O0000001O0000001O00001O0000001N1000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O001N10001O001N101O001N10001O0O2O001O000O2O001O0O2O00001O0O2O001O0O2O00001O0O2O001O0O101O001OjoZ2"}, "label": "umbrella in front"}]} {"id": 306561, "image": "COCO_train2014_000000306561.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in a white long sleeve shirt wit a tie on\"."}], "task": "refering", "answer_template": "The \"man in a white long sleeve shirt wit a tie on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 79, 80, 81, 97, 98, 99, 100, 115, 116, 117, 118, 132, 133, 134, 135, 149, 150, 151, 152, 167, 168, 169, 185, 186, 203, 204], "bbox": [0.3098, 0.29663663663663664, 0.5792, 0.9887687687687687], "iscrowd": 0, "area": 11859.5736, "rle": {"size": [333, 500], "counts": "bnb1`2a0bMi6^2VIcMj6]2UIdMj6]2UIdMk6\\2TIeMk6\\2TIeMk6\\2TIeMl6[2SIfMl6[2SIeMm6\\2RIeMn6[2QIfMn6[2PIgMo6Z2oHhMQ7j2O1O100O1O100O1O010O1O1O1N2O1O1N2OjNlIUNS6\\1`JaN^5Q1UKlNi4T1\\KiNb4Y1aKcN^4_1dK_NZ4c1iKZNU4h1nKUNP4m1V22N2N2M3K5L4L4L4L4G9O100O1O1SOkLbJU3\\5oLaJR3]5RM`Jo2^5VM_Jj2_5[M]Jf2a5`MZJa2d5eMXJ[2g5V1N2NM3K7M2O1N3M2N3N2M3N2N2O2M2O1N2O1N2O2O0O1O1O2O1N2N2O1N2N3N2M3M3jMYHb1j7YNYHe1m7TNUHl1W8O010O001O010O001O010O001O0010O01O001O001N2O0O2O1N2N101N2O1N102M3M3N1N3N2M3TOoTT2"}, "label": "man in a white long sleeve shirt wit a tie on"}]} {"id": 306561, "image": "COCO_train2014_000000306561.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white shirt and black pants\"."}], "task": "refering", "answer_template": "The \"a man in a white shirt and black pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 79, 80, 81, 97, 98, 99, 100, 115, 116, 117, 118, 132, 133, 134, 135, 149, 150, 151, 152, 167, 168, 169, 185, 186, 203, 204], "bbox": [0.3098, 0.29663663663663664, 0.5792, 0.9887687687687687], "iscrowd": 0, "area": 11859.5736, "rle": {"size": [333, 500], "counts": "bnb1`2a0bMi6^2VIcMj6]2UIdMj6]2UIdMk6\\2TIeMk6\\2TIeMk6\\2TIeMl6[2SIfMl6[2SIeMm6\\2RIeMn6[2QIfMn6[2PIgMo6Z2oHhMQ7j2O1O100O1O100O1O010O1O1O1N2O1O1N2OjNlIUNS6\\1`JaN^5Q1UKlNi4T1\\KiNb4Y1aKcN^4_1dK_NZ4c1iKZNU4h1nKUNP4m1V22N2N2M3K5L4L4L4L4G9O100O1O1SOkLbJU3\\5oLaJR3]5RM`Jo2^5VM_Jj2_5[M]Jf2a5`MZJa2d5eMXJ[2g5V1N2NM3K7M2O1N3M2N3N2M3N2N2O2M2O1N2O1N2O2O0O1O1O2O1N2N2O1N2N3N2M3M3jMYHb1j7YNYHe1m7TNUHl1W8O010O001O010O001O010O001O0010O01O001O001N2O0O2O1N2N101N2O1N102M3M3N1N3N2M3TOoTT2"}, "label": "a man in a white shirt and black pants"}]} {"id": 306561, "image": "COCO_train2014_000000306561.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman have a beer cup\"."}], "task": "refering", "answer_template": "The \"woman have a beer cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 92, 93, 94, 109, 110, 111, 112, 128, 129, 130, 146, 147, 148, 164, 165, 166], "bbox": [0.10326, 0.39774774774774774, 0.25742000000000004, 0.8381981981981982], "iscrowd": 0, "area": 6925.573050000002, "rle": {"size": [333, 500], "counts": "oRa06U:8H5K4M3L4L4L4M3M2N3M3M2N3M3M3PIUNY5V2UJRNj5Y2dIoMZ6S3O1N2N2N2N2N2N2O1O2N1N2O1O1O1O1OFTJULk5f3^JWLa5c3g0N3N1000001O001O1O1O1O1O11O001O001O001OlNUIfNk6Z1VIfNi6Z1YIdNg6[1[IeNd6[1]IdNc6\\1^IdNb6[1_IdNd6Y1]ITOX6i0iIWOY6f0hIZO[6c0eI^O]6>cICb67_IfN93[6R1R2K4K4L5K5L4K4Ld^h3"}, "label": "woman have a beer cup"}]} {"id": 306561, "image": "COCO_train2014_000000306561.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady holding a cup and listening the conversation in a house\"."}], "task": "refering", "answer_template": "The \"a lady holding a cup and listening the conversation in a house\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 92, 93, 94, 109, 110, 111, 112, 128, 129, 130, 146, 147, 148, 164, 165, 166], "bbox": [0.10326, 0.39774774774774774, 0.25742000000000004, 0.8381981981981982], "iscrowd": 0, "area": 6925.573050000002, "rle": {"size": [333, 500], "counts": "oRa06U:8H5K4M3L4L4L4M3M2N3M3M2N3M3M3PIUNY5V2UJRNj5Y2dIoMZ6S3O1N2N2N2N2N2N2O1O2N1N2O1O1O1O1OFTJULk5f3^JWLa5c3g0N3N1000001O001O1O1O1O1O11O001O001O001OlNUIfNk6Z1VIfNi6Z1YIdNg6[1[IeNd6[1]IdNc6\\1^IdNb6[1_IdNd6Y1]ITOX6i0iIWOY6f0hIZO[6c0eI^O]6>cICb67_IfN93[6R1R2K4K4L5K5L4K4Ld^h3"}, "label": "a lady holding a cup and listening the conversation in a house"}]} {"id": 445829, "image": "COCO_train2014_000000445829.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark brown leather sofa behind a foot stool that has a laptop computer on it\"."}], "task": "refering", "answer_template": "The \"a dark brown leather sofa behind a foot stool that has a laptop computer on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [208, 209, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 265, 266, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 360, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.07418749999999999, 0.5271250000000001, 0.727421875, 0.9873333333333334], "iscrowd": 0, "area": 63146.2347, "rle": {"size": [480, 640], "counts": "`Yf0Q4oK>B010O001N101O00001O001O001O001O001O00001O001O001O001N1O2M2N2O2M2N3M2O2M2N3O001O00001O001O001O001O001N101O00001O001O001O001O001O00001O010O001O001O0010O0001O001O0010O01O001O001O01O01O001O001O010O001O00001O010O0000000001O00000000001O00001O00001O00001O00001O00001O00001O00001O00001O00001N10001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O1O001O1O001O1N2O001O1O001O1O001O1O001O1O1O001O1O001O1O004L6J6J6J6J4L00O100000000O10000000000O100000000O10000000000O10000000O010O10000O10O010000O100O01000O100O01000O100O10O10O100O10O10O100O1000O01N2N2N2O0O2N2N2N2O1N1O2N2O1N2N1O2N2O1N2O11N100000000O12N3M4L3M3M3M000001O0001O000000000000001O000001O0000000001O000000000000001O01O000000000001O00000000000O1N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2O1N3N1O1N2O1O1O5J5L5K4K6K4L[Va2"}, "label": "a dark brown leather sofa behind a foot stool that has a laptop computer on it"}]} {"id": 445829, "image": "COCO_train2014_000000445829.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black leather couch and ottoman in the center of the room\"."}], "task": "refering", "answer_template": "The \"the black leather couch and ottoman in the center of the room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [208, 209, 210, 211, 212, 213, 214, 231, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 265, 266, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 360, 372, 373, 374, 375, 376, 377, 378, 379, 380], "bbox": [0.07418749999999999, 0.5271250000000001, 0.727421875, 0.9873333333333334], "iscrowd": 0, "area": 63146.2347, "rle": {"size": [480, 640], "counts": "`Yf0Q4oK>B010O001N101O00001O001O001O001O001O00001O001O001O001N1O2M2N2O2M2N3M2O2M2N3O001O00001O001O001O001O001N101O00001O001O001O001O001O00001O010O001O001O0010O0001O001O0010O01O001O001O01O01O001O001O010O001O00001O010O0000000001O00000000001O00001O00001O00001O00001O00001O00001O00001O00001O00001N10001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O1O001O1O001O1N2O001O1O001O1O001O1O001O1O1O001O1O001O1O004L6J6J6J6J4L00O100000000O10000000000O100000000O10000000000O10000000O010O10000O10O010000O100O01000O100O01000O100O10O10O100O10O10O100O1000O01N2N2N2O0O2N2N2N2O1N1O2N2O1N2N1O2N2O1N2O11N100000000O12N3M4L3M3M3M000001O0001O000000000000001O000001O0000000001O000000000000001O01O000000000001O00000000000O1N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2O1N3N1O1N2O1O1O5J5L5K4K6K4L[Va2"}, "label": "the black leather couch and ottoman in the center of the room"}]} {"id": 355717, "image": "COCO_train2014_000000355717.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man playing tennis\"."}], "task": "refering", "answer_template": "The \"man playing tennis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 82, 83, 84, 100, 101, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 309, 310, 311, 312, 331, 332, 333, 334, 335], "bbox": [0.3328125, 0.16741784037558685, 0.67159375, 0.9902816901408451], "iscrowd": 0, "area": 38871.80770000002, "rle": {"size": [426, 640], "counts": "Thh21Y=0O2N101N1O2N101N1O101N1O2O0O2N1O2O0O2N3N2M4N1N2N2N2O2M1O0001O0001O01O0001O01O0001O01O000001O01O03M3M3N3L3M4L3N3PIQN]2R2\\MmNi1S1RNMS14hNc0f0\\OVOh0j0XOROl0n0UOlNP1T1POhNT1Y1kNcNY1]1hN^N\\1b1dNZN`1f1`NUNe1k1\\NPNh1Q2WNkMl1V2TNfMP2Z2QN`MT2`2lM\\MX2e2hMVM\\2j2cMSMa2m2_MnLf2R3ZMjLj2W3TMfLP3Z3PMbLT3^3lL]LY3c3fLZL^3g3aLVLb3j3]LULe3k3[LSLg3m3YLQLi3P4ULnKm3S4SLkKm3W4SLgKn3Z4QLeKo3]4QLaKP4a4oK]KQ4d4oK[KQ4g4oKWKR4i4oKUKQ4m4nKRKR4o4oKoJR4R5nKlJR4U5nKjJS4V5o12N101N1O28Gg0ZO0O1O100O100O100O100O1O100O11O00000O10000000001O00000000000000001O000000000000001O00ZOWGkLi8S3^GhLb8V3eGeL[8Z3kGaLU8]3m0N00O1O1O100O1O2N1O1O1O100O1N3L5L3L5K5K4M4K5K4L5K5L4K4L5K5L3L5dN\\1N1N3N2N1O2M3N2N2M3N3M3L4M3M2M4]M^G:e8C_G9d8C`G9c8F`G6c8H`G4c8J`G1d8L_G1d8M_GOc80`GLc82`GId85^GHe85Y2O2L5J6I8I7H7IfXg2"}, "label": "man playing tennis"}]} {"id": 355717, "image": "COCO_train2014_000000355717.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man holding a black ball bat , wearing white shorts and a white shirt with a yellow and blue stripe running diagonal down it\"."}], "task": "refering", "answer_template": "The \"the man holding a black ball bat , wearing white shorts and a white shirt with a yellow and blue stripe running diagonal down it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 82, 83, 84, 100, 101, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 285, 286, 287, 288, 289, 308, 309, 310, 311, 312, 331, 332, 333, 334, 335], "bbox": [0.3328125, 0.16741784037558685, 0.67159375, 0.9902816901408451], "iscrowd": 0, "area": 38871.80770000002, "rle": {"size": [426, 640], "counts": "Thh21Y=0O2N101N1O2N101N1O101N1O2O0O2N1O2O0O2N3N2M4N1N2N2N2O2M1O0001O0001O01O0001O01O0001O01O000001O01O03M3M3N3L3M4L3N3PIQN]2R2\\MmNi1S1RNMS14hNc0f0\\OVOh0j0XOROl0n0UOlNP1T1POhNT1Y1kNcNY1]1hN^N\\1b1dNZN`1f1`NUNe1k1\\NPNh1Q2WNkMl1V2TNfMP2Z2QN`MT2`2lM\\MX2e2hMVM\\2j2cMSMa2m2_MnLf2R3ZMjLj2W3TMfLP3Z3PMbLT3^3lL]LY3c3fLZL^3g3aLVLb3j3]LULe3k3[LSLg3m3YLQLi3P4ULnKm3S4SLkKm3W4SLgKn3Z4QLeKo3]4QLaKP4a4oK]KQ4d4oK[KQ4g4oKWKR4i4oKUKQ4m4nKRKR4o4oKoJR4R5nKlJR4U5nKjJS4V5o12N101N1O28Gg0ZO0O1O100O100O100O100O1O100O11O00000O10000000001O00000000000000001O000000000000001O00ZOWGkLi8S3^GhLb8V3eGeL[8Z3kGaLU8]3m0N00O1O1O100O1O2N1O1O1O100O1N3L5L3L5K5K4M4K5K4L5K5L4K4L5K5L3L5dN\\1N1N3N2N1O2M3N2N2M3N3M3L4M3M2M4]M^G:e8C_G9d8C`G9c8F`G6c8H`G4c8J`G1d8L_G1d8M_GOc80`GLc82`GId85^GHe85Y2O2L5J6I8I7H7IfXg2"}, "label": "the man holding a black ball bat , wearing white shorts and a white shirt with a yellow and blue stripe running diagonal down it"}]} {"id": 355717, "image": "COCO_train2014_000000355717.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in the background jumping to return the tennis ball\"."}], "task": "refering", "answer_template": "The \"a man in the background jumping to return the tennis ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 129, 130, 131, 153, 154, 155, 176, 177, 178, 199, 200, 201, 222, 223, 224, 245, 246, 268, 269, 291, 292, 313, 314], "bbox": [0.643234375, 0.27354460093896715, 0.785015625, 0.9103051643192488], "iscrowd": 0, "area": 11687.739099999997, "rle": {"size": [426, 640], "counts": "Xa[53W73`N1Z14cN0V17gNLW16eNOW14eN0W14eN0W12fN9P1JmNe1D]N=e1^O^Na0e1ZO]Ng0f1SO]Nl0f1iNbNX1`1SM2m21gKY1Y4k3100O101N1O1N3M2O1N2N3M2O1O2N2M3N1O2M3J5L5K5K5J5L5K5A?^Oc0SOm0N2O1N2N101N2O1N1O2O1N2O3L6J7J6I4L3N3L3M4L3N3L3M3N3L3M3M3N3L3M3M2O1N2N3N1N2N3M3N1N3M2Ne0\\O1N2NXTi1"}, "label": "a man in the background jumping to return the tennis ball"}]} {"id": 355717, "image": "COCO_train2014_000000355717.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the player that is out of focus\"."}], "task": "refering", "answer_template": "The \"the player that is out of focus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 129, 130, 131, 153, 154, 155, 176, 177, 178, 199, 200, 201, 222, 223, 224, 245, 246, 268, 269, 291, 292, 313, 314], "bbox": [0.643234375, 0.27354460093896715, 0.785015625, 0.9103051643192488], "iscrowd": 0, "area": 11687.739099999997, "rle": {"size": [426, 640], "counts": "Xa[53W73`N1Z14cN0V17gNLW16eNOW14eN0W14eN0W12fN9P1JmNe1D]N=e1^O^Na0e1ZO]Ng0f1SO]Nl0f1iNbNX1`1SM2m21gKY1Y4k3100O101N1O1N3M2O1N2N3M2O1O2N2M3N1O2M3J5L5K5K5J5L5K5A?^Oc0SOm0N2O1N2N101N2O1N1O2O1N2O3L6J7J6I4L3N3L3M4L3N3L3M3N3L3M3M3N3L3M3M2O1N2N3N1N2N3M3N1N3M2Ne0\\O1N2NXTi1"}, "label": "the player that is out of focus"}]} {"id": 290185, "image": "COCO_train2014_000000290185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person in the background wearing a gray sweater and a pair of black slacks and a pair of black shoes\"."}], "task": "refering", "answer_template": "The \"the person in the background wearing a gray sweater and a pair of black slacks and a pair of black shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 43, 44, 45, 66, 67, 68, 89, 90, 91, 112, 113, 114, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 227], "bbox": [0.8590781249999999, 0.006743697478991596, 1.0, 0.5640546218487394], "iscrowd": 0, "area": 17209.991050000004, "rle": {"size": [476, 640], "counts": "Ulo7?j=f0B>I7I7I7I7I4L5K5K5J5L5K5J6K4L5J6K4K6K5K5J5I8QOo0mMkIYKZ6f4gIUK]6k4dIPK]6S5dIgJ^6\\5cI_J^6d5cIWJ^6l5cIoI^6T6cIgI^6\\6cI^I^6f6cIUI\\6P7eIkHZ6Z7gIaHZ6b7`00000000000000000000O1000000001O1O1O1O001O1O1O1O001O1O1O1O1O001O001O000000cNoHPKP7m4XInJh6P5_IkJa6R5dIlJ\\6Q5hIoJW6n4nIPKR6n4QJQKo5l4UJSKk5j4YJVKf5h4]JWKc5h4`JVK`5h4cJWK]5h4eJXKZ5f4iJYKW5f4kJYKU5f4nJXKo4"}, "label": "the person in the background wearing a gray sweater and a pair of black slacks and a pair of black shoes"}]} {"id": 290185, "image": "COCO_train2014_000000290185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in grey suit jacket and black pants standing near bar\"."}], "task": "refering", "answer_template": "The \"man in grey suit jacket and black pants standing near bar\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 43, 44, 45, 66, 67, 68, 89, 90, 91, 112, 113, 114, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 227], "bbox": [0.8590781249999999, 0.006743697478991596, 1.0, 0.5640546218487394], "iscrowd": 0, "area": 17209.991050000004, "rle": {"size": [476, 640], "counts": "Ulo7?j=f0B>I7I7I7I7I4L5K5K5J5L5K5J6K4L5J6K4K6K5K5J5I8QOo0mMkIYKZ6f4gIUK]6k4dIPK]6S5dIgJ^6\\5cI_J^6d5cIWJ^6l5cIoI^6T6cIgI^6\\6cI^I^6f6cIUI\\6P7eIkHZ6Z7gIaHZ6b7`00000000000000000000O1000000001O1O1O1O001O1O1O1O001O1O1O1O1O001O001O000000cNoHPKP7m4XInJh6P5_IkJa6R5dIlJ\\6Q5hIoJW6n4nIPKR6n4QJQKo5l4UJSKk5j4YJVKf5h4]JWKc5h4`JVK`5h4cJWK]5h4eJXKZ5f4iJYKW5f4kJYKU5f4nJXKo4"}, "label": "man in grey suit jacket and black pants standing near bar"}]} {"id": 290185, "image": "COCO_train2014_000000290185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"napkins stacked on a round , brown table\"."}], "task": "refering", "answer_template": "The \"napkins stacked on a round , brown table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 138, 139, 161, 162, 185, 186, 208, 209, 231, 232, 254, 255, 277, 278, 300, 301, 302, 303, 322, 323, 324, 325, 326, 345, 346, 347, 348, 369, 370], "bbox": [0.0, 0.05088235294117647, 0.26575000000000004, 0.9749159663865545], "iscrowd": 0, "area": 25774.763649999994, "rle": {"size": [476, 640], "counts": "o0W4b:6K4L21O00001O001O000ULjEP3W:iLTFR3l9hL`FR3a9gLjFT3`6mLXLP3h3SMVLR3c3RMYLU3`3nL^LW3\\3mLaLY3Y3jLcL]3V3gLgL^3S3fLiLa3P3bLnLd3l2_LQMg3h2]LTMi3f2[LWMk3b2YL[Mm3_2UL^MQ4\\2SLaMS4X2QLeMU4T2mKjMZ4P2fKPN_4j1aKVNf4c1[K]Nk4\\1VKdNo4W1PKiNW5P1jJPO\\5i0eJWO`5c0aJ\\Of5>YJCl57UJIQ60PJOV6LjI4\\6EdI;W1dMT2j1fLb0V1iMT2`1fLg0U1oMV2S1eLo0U1SNV2h0fLU1S1YNX2 in this image.", "objects": [{"patches": [23, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 138, 139, 161, 162, 185, 186, 208, 209, 231, 232, 254, 255, 277, 278, 300, 301, 302, 303, 322, 323, 324, 325, 326, 345, 346, 347, 348, 369, 370], "bbox": [0.0, 0.05088235294117647, 0.26575000000000004, 0.9749159663865545], "iscrowd": 0, "area": 25774.763649999994, "rle": {"size": [476, 640], "counts": "o0W4b:6K4L21O00001O001O000ULjEP3W:iLTFR3l9hL`FR3a9gLjFT3`6mLXLP3h3SMVLR3c3RMYLU3`3nL^LW3\\3mLaLY3Y3jLcL]3V3gLgL^3S3fLiLa3P3bLnLd3l2_LQMg3h2]LTMi3f2[LWMk3b2YL[Mm3_2UL^MQ4\\2SLaMS4X2QLeMU4T2mKjMZ4P2fKPN_4j1aKVNf4c1[K]Nk4\\1VKdNo4W1PKiNW5P1jJPO\\5i0eJWO`5c0aJ\\Of5>YJCl57UJIQ60PJOV6LjI4\\6EdI;W1dMT2j1fLb0V1iMT2`1fLg0U1oMV2S1eLo0U1SNV2h0fLU1S1YNX2 in this image.", "objects": [{"patches": [57, 78, 79, 80, 81, 82, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 200, 213, 215, 216, 217, 218, 219, 220, 221, 222, 223, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 352, 353, 354, 355, 356, 357, 358, 376, 377], "bbox": [0.19634374999999998, 0.1653361344537815, 0.74771875, 0.9760084033613445], "iscrowd": 0, "area": 77295.27565000001, "rle": {"size": [476, 640], "counts": "jkj1d0V>2O2M3M3N2M3N1N3N2M3N2N101N3M3N2M3M3M3N2M3M3N2M3N2N2N2N2M2O1O1O1N2O1O1O1O0O2O1O1O2N1N3N1O2N1N3N1O2N1O2M2O2N1O2N1N3N1O2N1N3N1O1O1O1O1000000O100000000O100001O001O1O001O001O1O001O001O001N2O001O\\Od0oNQ1L4L4L4L3M4L4L4L4L4M3L3M4L4M3L4L4L3N3L4L4L4cN]1L4K4L4L4M3M3N2M3N1N3M3N2M3N2M3O001O1O0000O00100O100O1O100O010O1O1O1O1O1O2N2O0O2O1N101N2O03N2N1N3N2N1N3N2N1N3N2N0O2O001N101O0O101O001O001O00001O0010O0001O001O000001O0O2O001O0O2O001O0O2O001N101O001N101O001N101O001N101O0O2O001O0O2O001O0O2O0O2N1O2N1O2N1O2N1O2N1O2N1O3M4L4L3M4L4L4L4L4oH]Jb4g5RKbJm4b5gJgJX5]5\\JmJb5W5RJRKm5Q5gIYKX6g6N1O1O1O2N1O1O2N1O101O000010O0001O00010O0000010O000010O0001O01O01O00010O0000100O2N1O100O2N1O1O2O0O1O2N100O2N1O100O2N1O3N2M3M3RM[F?h9YOWGNk8JjGFY83QHGR81XHJb;L3L5K5K3N0O001O100O1O1O00100O1O001O1O00elZ2"}, "label": "a bride with pizza cheese hanging off pizza"}]} {"id": 290185, "image": "COCO_train2014_000000290185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman holds up a piece of pizza , she is wearing a dress\"."}], "task": "refering", "answer_template": "The \"a woman holds up a piece of pizza , she is wearing a dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 78, 79, 80, 81, 82, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 175, 176, 192, 193, 194, 195, 196, 197, 198, 199, 200, 213, 215, 216, 217, 218, 219, 220, 221, 222, 223, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 352, 353, 354, 355, 356, 357, 358, 376, 377], "bbox": [0.19634374999999998, 0.1653361344537815, 0.74771875, 0.9760084033613445], "iscrowd": 0, "area": 77295.27565000001, "rle": {"size": [476, 640], "counts": "jkj1d0V>2O2M3M3N2M3N1N3N2M3N2N101N3M3N2M3M3M3N2M3M3N2M3N2N2N2N2M2O1O1O1N2O1O1O1O0O2O1O1O2N1N3N1O2N1N3N1O2N1O2M2O2N1O2N1N3N1O2N1N3N1O1O1O1O1000000O100000000O100001O001O1O001O001O1O001O001O001N2O001O\\Od0oNQ1L4L4L4L3M4L4L4L4L4M3L3M4L4M3L4L4L3N3L4L4L4cN]1L4K4L4L4M3M3N2M3N1N3M3N2M3N2M3O001O1O0000O00100O100O1O100O010O1O1O1O1O1O2N2O0O2O1N101N2O03N2N1N3N2N1N3N2N1N3N2N0O2O001N101O0O101O001O001O00001O0010O0001O001O000001O0O2O001O0O2O001O0O2O001N101O001N101O001N101O001N101O0O2O001O0O2O001O0O2O0O2N1O2N1O2N1O2N1O2N1O2N1O3M4L4L3M4L4L4L4L4oH]Jb4g5RKbJm4b5gJgJX5]5\\JmJb5W5RJRKm5Q5gIYKX6g6N1O1O1O2N1O1O2N1O101O000010O0001O00010O0000010O000010O0001O01O01O00010O0000100O2N1O100O2N1O1O2O0O1O2N100O2N1O100O2N1O3N2M3M3RM[F?h9YOWGNk8JjGFY83QHGR81XHJb;L3L5K5K3N0O001O100O1O1O00100O1O001O1O00elZ2"}, "label": "a woman holds up a piece of pizza , she is wearing a dress"}]} {"id": 290185, "image": "COCO_train2014_000000290185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a colorful bouquet on a table with a bottle of beer on a table before a beautiful , smiling woman\"."}], "task": "refering", "answer_template": "The \"a colorful bouquet on a table with a bottle of beer on a table before a beautiful , smiling woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [290, 303, 304, 305, 306, 313, 314, 325, 326, 327, 328, 329, 330, 332, 335, 336, 337, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.05682812499999999, 0.7507142857142857, 0.9309375, 0.9821638655462185], "iscrowd": 0, "area": 28225.75805, "rle": {"size": [476, 640], "counts": "hSb01j>1O100O1O100O1O1O100O1O100O1O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O2N1O100O1O100O1O1O100O1O100O1O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1001O00000000001O0000000000001O00000000001O00000000001O0000000000001O00000000001O002N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O0000000000O10000000000000000O100000000000000O1000000O1N2N2M3N2N2M3N2N2N2M3N2001O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O001O1O001O1O010OO1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O10000O100O1N2N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2M3L4L4L4L4L4L4L4N200000000O100000000O100000000O100000000003M=C;E4L4L4L4L4L6J2N1O00001O001O00001O0000001O00000000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O0000000001O0001O000000000000002N2N3M2N3M3M2N3M2N3M3M2N3Mmoc0"}, "label": "a colorful bouquet on a table with a bottle of beer on a table before a beautiful , smiling woman"}]} {"id": 290185, "image": "COCO_train2014_000000290185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table with a beer and a flower bouquet on it\"."}], "task": "refering", "answer_template": "The \"a table with a beer and a flower bouquet on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [290, 303, 304, 305, 306, 313, 314, 325, 326, 327, 328, 329, 330, 332, 335, 336, 337, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.05682812499999999, 0.7507142857142857, 0.9309375, 0.9821638655462185], "iscrowd": 0, "area": 28225.75805, "rle": {"size": [476, 640], "counts": "hSb01j>1O100O1O100O1O1O100O1O100O1O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O2N1O100O1O100O1O1O100O1O100O1O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1001O00000000001O0000000000001O00000000001O00000000001O0000000000001O00000000001O002N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O0000000000O10000000000000000O100000000000000O1000000O1N2N2M3N2N2M3N2N2N2M3N2001O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O001O1O001O1O010OO1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O10000O100O1N2N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2M3L4L4L4L4L4L4L4N200000000O100000000O100000000O100000000003M=C;E4L4L4L4L4L6J2N1O00001O001O00001O0000001O00000000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O0000000001O0001O000000000000002N2N3M2N3M3M2N3M2N3M3M2N3Mmoc0"}, "label": "a table with a beer and a flower bouquet on it"}]} {"id": 568725, "image": "COCO_train2014_000000568725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow section of a car and the wheel to the right of it\"."}], "task": "refering", "answer_template": "The \"yellow section of a car and the wheel to the right of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 56, 57, 58, 59, 61, 62, 63, 64, 65, 66, 67, 68, 80, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 180, 181], "bbox": [0.305140625, 0.0060563380281690145, 0.9961874999999999, 0.48246478873239435], "iscrowd": 0, "area": 47414.22944999999, "rle": {"size": [426, 640], "counts": "o[b21Y=00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O001O3M3M3M3M3M3M4L3M3M3M3M3M3M6J6J6J00000000000000000000O1000000000000000000O1000000000000000000O10000000000000000O1000000000000000000O1000000000000000000O100000000O1N2O1N2N2O1N2O1N2N2O1O1O100O100O100O10000O100O100O100O100O100O1000000001O0000000000000000000000001O000000000000000000000000001O00000000000000001O2N1O2N1O1O2N1O2N1O1O2N1O2N1O2dFkM_7V2XHUNe7l1QH`Nl7b1jGjNR8W1dGUOY8m0]G^O`8j2L3M3M3M1O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O001O00001O001O001O00001O001O00001O001O001OO10000000000000000O100000000000000O10000000000000000O100000000000000O100000000000000O101O0000000000000O100000000000000O10000000000000000O100000000000000O100000000000000O10000000`j0"}, "label": "yellow section of a car and the wheel to the right of it"}]} {"id": 568725, "image": "COCO_train2014_000000568725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow car\"."}], "task": "refering", "answer_template": "The \"the yellow car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 56, 57, 58, 59, 61, 62, 63, 64, 65, 66, 67, 68, 80, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 180, 181], "bbox": [0.305140625, 0.0060563380281690145, 0.9961874999999999, 0.48246478873239435], "iscrowd": 0, "area": 47414.22944999999, "rle": {"size": [426, 640], "counts": "o[b21Y=00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O0000001O001O3M3M3M3M3M3M4L3M3M3M3M3M3M6J6J6J00000000000000000000O1000000000000000000O1000000000000000000O10000000000000000O1000000000000000000O1000000000000000000O100000000O1N2O1N2N2O1N2O1N2N2O1O1O100O100O100O10000O100O100O100O100O100O1000000001O0000000000000000000000001O000000000000000000000000001O00000000000000001O2N1O2N1O1O2N1O2N1O1O2N1O2N1O2dFkM_7V2XHUNe7l1QH`Nl7b1jGjNR8W1dGUOY8m0]G^O`8j2L3M3M3M1O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O001O00001O001O00001O001O001O00001O001O001O00001O001O00001O001O001OO10000000000000000O100000000000000O10000000000000000O100000000000000O100000000000000O101O0000000000000O100000000000000O10000000000000000O100000000000000O100000000000000O10000000`j0"}, "label": "the yellow car"}]} {"id": 568725, "image": "COCO_train2014_000000568725.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red box behind the teddy bear\"."}], "task": "refering", "answer_template": "The \"a red box behind the teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 69, 70, 71, 72, 73, 74, 75, 76, 79, 92, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327], "bbox": [0.0, 0.0, 0.471171875, 0.9797652582159624], "iscrowd": 0, "area": 84834.3028, "rle": {"size": [426, 640], "counts": "4]6m6^6bI0000000000000000000000000O1000000000000000000000001O000000000000000000000000000000000000000000000000000000001O000000000000000O1000000000000000000000000000000000000000000000O10000000000O10000000000O10000000000O10000000000O100000O1000O1000000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O1000000O00100O1O1O1O100O1G9\\Od0L4M3M3M3M3M3M3M3M3\\MSNhIP2X5SObJP1^43[K1_48[KKa4 in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 69, 70, 71, 72, 73, 74, 75, 76, 79, 92, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327], "bbox": [0.0, 0.0, 0.471171875, 0.9797652582159624], "iscrowd": 0, "area": 84834.3028, "rle": {"size": [426, 640], "counts": "4]6m6^6bI0000000000000000000000000O1000000000000000000000001O000000000000000000000000000000000000000000000000000000001O000000000000000O1000000000000000000000000000000000000000000000O10000000000O10000000000O10000000000O10000000000O100000O1000O1000000000000O10000000000O10000000000O10000000000O10000000000O10000000000O10000000000O1000000O00100O1O1O1O100O1G9\\Od0L4M3M3M3M3M3M3M3M3\\MSNhIP2X5SObJP1^43[K1_48[KKa4 in this image.", "objects": [{"patches": [26, 27, 28, 29, 43, 44, 45, 46, 60, 77, 94, 111, 128, 144, 145, 161, 162, 178, 179, 195, 196, 212, 213, 229, 230, 246, 247, 262, 263, 276, 277, 278, 279, 280, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 327, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349], "bbox": [0.28197916666666667, 0.038609375, 0.7215416666666667, 0.8965781250000001], "iscrowd": 0, "area": 24867.704049999986, "rle": {"size": [640, 480], "counts": "eld23Tc0n0QOo0ROn0C=K5K6J4L00000000000O1000001O0000000000000O100000000000000000000O100000000000000000000O101O00000000000000000000000000000000000000000O1000000000000000001O00000000000000000H7G:G9G9G9G8G:I7J6I6K6I7J6I7J5J7J6J6I6K6I7J6I7J5J7H8H8H7IVMnFVKe8d4oG[Kb7`4SI_K^6]4VJaK\\5[4XKdKY4W4[LiKV3S4^MlKS2P4aNoKP1l3ERL4c3`0\\L\\O]3o0cLlNX3_1gL]Nl2U2RMgMb2k2]MQMV2a3iMZLk1W4UNdK_1n4`NnJS1d5kNXJi0Z6VObI=P7BkH6b7JZH6k7HQH8U8aNmH_1g<1O10O2N6I7J5J7J6E;F:000O0100000000O010000000O10O1000O100000O10O10000000O01000000O0000O100O100O100O101N10O001O10O01O0010O01O0010O03M4L5L3L4LUfc2"}, "label": "the black and yellow mercedes lift"}]} {"id": 11673, "image": "COCO_train2014_000000011673.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"yellow crane\"."}], "task": "refering", "answer_template": "The \"yellow crane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 29, 43, 44, 45, 46, 60, 77, 94, 111, 128, 144, 145, 161, 162, 178, 179, 195, 196, 212, 213, 229, 230, 246, 247, 262, 263, 276, 277, 278, 279, 280, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 327, 328, 329, 330, 331, 332, 345, 346, 347, 348, 349], "bbox": [0.28197916666666667, 0.038609375, 0.7215416666666667, 0.8965781250000001], "iscrowd": 0, "area": 24867.704049999986, "rle": {"size": [640, 480], "counts": "eld23Tc0n0QOo0ROn0C=K5K6J4L00000000000O1000001O0000000000000O100000000000000000000O100000000000000000000O101O00000000000000000000000000000000000000000O1000000000000000001O00000000000000000H7G:G9G9G9G8G:I7J6I6K6I7J6I7J5J7J6J6I6K6I7J6I7J5J7H8H8H7IVMnFVKe8d4oG[Kb7`4SI_K^6]4VJaK\\5[4XKdKY4W4[LiKV3S4^MlKS2P4aNoKP1l3ERL4c3`0\\L\\O]3o0cLlNX3_1gL]Nl2U2RMgMb2k2]MQMV2a3iMZLk1W4UNdK_1n4`NnJS1d5kNXJi0Z6VObI=P7BkH6b7JZH6k7HQH8U8aNmH_1g<1O10O2N6I7J5J7J6E;F:000O0100000000O010000000O10O1000O100000O10O10000000O01000000O0000O100O100O100O101N10O001O10O01O0010O01O0010O03M4L5L3L4LUfc2"}, "label": "yellow crane"}]} {"id": 265625, "image": "COCO_train2014_000000265625.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman looking at a phone\"."}], "task": "refering", "answer_template": "The \"a woman looking at a phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 127, 128, 129, 130, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 225, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 286, 287, 288, 289, 291, 292, 293, 294, 295, 309, 310, 311, 312, 314, 315, 316, 317, 318, 333, 334, 337, 338, 339, 340, 363], "bbox": [0.43870312499999997, 0.2525, 0.845140625, 0.8897708333333333], "iscrowd": 0, "area": 42495.46469999999, "rle": {"size": [480, 640], "counts": "eRT41l>3N2K6K4L4K5L4L4K5M3L4M3L4M3M3L4M3M3L4N2O1N3M2N2NW1jN8G2N3N1N2O1O1N2O1O001O1N2O1O1O10O01O1O100O1N2O0O2N10N1MnM[KPJg4Y5SLcJn3\\5VLaJk3\\5YLcJe3]5_LaJ`3^5fL]JY3b5mL[JR3d5TMXJk2f5bMoI]2Q6mMgIR2X6SNeIm1Z6WNbIj1]6YNaIg1^6[NaId1_6^N_Ic1`6`N^Ia1`6aN_I_1`6cN_I]1_6fN_I[1`6gN_IY1a6gN^IZ1a6gN^IZ1b6fN]I[1b6fN]I[1c6a3O100O2O0O10000O10000000O00100O100O0N3N2N1H9DC=E;E;J6L4I7E_S^1"}, "label": "a woman looking at a phone"}]} {"id": 265625, "image": "COCO_train2014_000000265625.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman on her phone at a diner table\"."}], "task": "refering", "answer_template": "The \"a woman on her phone at a diner table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 127, 128, 129, 130, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 225, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 286, 287, 288, 289, 291, 292, 293, 294, 295, 309, 310, 311, 312, 314, 315, 316, 317, 318, 333, 334, 337, 338, 339, 340, 363], "bbox": [0.43870312499999997, 0.2525, 0.845140625, 0.8897708333333333], "iscrowd": 0, "area": 42495.46469999999, "rle": {"size": [480, 640], "counts": "eRT41l>3N2K6K4L4K5L4L4K5M3L4M3L4M3M3L4M3M3L4N2O1N3M2N2NW1jN8G2N3N1N2O1O1N2O1O001O1N2O1O1O10O01O1O100O1N2O0O2N10N1MnM[KPJg4Y5SLcJn3\\5VLaJk3\\5YLcJe3]5_LaJ`3^5fL]JY3b5mL[JR3d5TMXJk2f5bMoI]2Q6mMgIR2X6SNeIm1Z6WNbIj1]6YNaIg1^6[NaId1_6^N_Ic1`6`N^Ia1`6aN_I_1`6cN_I]1_6fN_I[1`6gN_IY1a6gN^IZ1a6gN^IZ1b6fN]I[1b6fN]I[1c6a3O100O2O0O10000O10000000O00100O100O0N3N2N1H9DC=E;E;J6L4I7E_S^1"}, "label": "a woman on her phone at a diner table"}]} {"id": 265625, "image": "COCO_train2014_000000265625.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt holding an umbrella\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt holding an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 148, 149, 161, 162, 163, 164, 165, 166, 167, 171, 172, 184, 185, 186, 187, 188, 189, 190, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 216, 217, 218, 230, 231, 232, 233, 234, 235, 239, 240, 253, 254, 255, 256, 257, 258, 259, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 352, 353, 368, 369, 370, 371, 372, 373, 374], "bbox": [0.0016875000000000002, 0.21508333333333332, 0.5320937499999999, 0.9853541666666668], "iscrowd": 0, "area": 68922.59080000003, "rle": {"size": [480, 640], "counts": "je0j1V=d3\\La2_M00000000000000000001O000000000000000000000000001O000000000000001O000000M3K5J6J6H8G9H8H8M3M3M3N2M3M3N2M3N2O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N200O100O1O100O1O100O1O100O100O1O100OnN_LgF`3W9cLiF]3T9fLlFY3R9jLnFV3o8mLQGS3k8PMVGP3g8QM[Gn2c8TM^Gl2_8UMcGk2Z8VMhGj2U8WMmGh2P8[MQHe2l7]MUHc2h7_MYHa2d7bM\\H]2b7eM_H[2]7hMdHX2Y7jMhHV2V7lMjHT2X7jMhHV2Y7iMgHW2Z7gMgHY2Z7fMfHZ2[7eMeH[2\\7dMdH\\2^7aMcH_2^7`MbH`2_7`M`H`2a7_M_Ha2b7_M]Ha2d7^M\\Hb2e7]M[Hc2g7\\MXHd2i7[MWHe2j3cLeNh0`Me2j3eLfNe0`Mf2j3fLfNc0`Mg2j3gLfNb0_Mg2k3hLfN`0_Mi2i3iLgN>`Mi2i3jLfN>`Mh2j3kLeN>`Mg2k3lLdN>`Mf2k3nLdN=`Me2l3oLcN>_Md2m3oLcN>_Mc2n3PMbN>_Mb2n3RMbN=_Ma2o3SMaN=_M`2P4TM`N=_M_2Q4UM_N=_M_2Q4UM_N=_M^2R4VM]N>`M\\2S4WM\\N>`M[2T4XM[N>`MZ2U4YMYN?aMX2V4ZMXN?aMW2W4ZMXNa0_MV2X4ZMXNa0_MU2Y4[MVNb0`MS2Z4\\MUNb0`MR2[4]MTNb0`MQ2\\4]MTNc0_MP2]4^MSNc0_MP2]4^MTNb0^MP2^4_MSNb0^MP2^4^MUNb0\\MQ2]4_MVNa0\\MQ2]4_MWN`0[MR2]4_MWN`0[MQ2^4aMVN?ZMQ2_4eMSN:]MR2_4iMoM6aMR2^4kMPN3aMS2^4lMoM2bMS2^4mMoM0bMS2_4oMmMOcMS2_4PNmMMcMT2^4QNnMLcMT2]4RNPNJbMU2[4TNSNGaMV2Y4VNUNEaMV2X4WNWNe3g1_LWNb3f1aLZN_3c1dL\\N]3a1fL_NZ3]1jLcNV3Z1mLeNT3X1PMgNP3U1TMjNl2k0`MUO`2h0cMWO^2g0eMXO[2f0gMYOZ2f0gMZOY2d0iM\\OW27nIlMl3m1W25QJjMj3P2U24VJiMf3S2T2VOTJXOV4b1g1mN\\J]OQ4e1d1kN^J\\OR4i1`1iN`JZOT4l1]1hNaJYOT4P2_1aN_J[OV4S2c1YNZOf1m0QNUOP2n6O1O010O1O10O01O10O01O100O010O1O010O1O010O00O2M2N2N3N1N2N3M2O1N3M2N2O1N2N2NO2L3N200100O2N2N101N1OARDPNl;o1ZDnMd;R2aDjM^;U2iDgMU;Y2h00O0100001N3N1O1O2N1N3N1O2N2N2M3N2N1O2N2M3N2mNkLYFW3e9PMTFQ3k9WMgEQ3W:m0M3N2N00N2N2N1O2O1N2N3M3M3M3N2M3M3M3M4M2M3M3O1O1O2O1N1O2N2N2N2N2N2N2N2N200O100O100O2O0O100O1O4M2M4M3L3N3K5K4L4L5K4K6K4L5K4K;@jm[4"}, "label": "a man in a black shirt holding an umbrella"}]} {"id": 265625, "image": "COCO_train2014_000000265625.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding an umbrella over a woman\"."}], "task": "refering", "answer_template": "The \"a man holding an umbrella over a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 148, 149, 161, 162, 163, 164, 165, 166, 167, 171, 172, 184, 185, 186, 187, 188, 189, 190, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 216, 217, 218, 230, 231, 232, 233, 234, 235, 239, 240, 253, 254, 255, 256, 257, 258, 259, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 352, 353, 368, 369, 370, 371, 372, 373, 374], "bbox": [0.0016875000000000002, 0.21508333333333332, 0.5320937499999999, 0.9853541666666668], "iscrowd": 0, "area": 68922.59080000003, "rle": {"size": [480, 640], "counts": "je0j1V=d3\\La2_M00000000000000000001O000000000000000000000000001O000000000000001O000000M3K5J6J6H8G9H8H8M3M3M3N2M3M3N2M3N2O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N200O100O1O100O1O100O1O100O100O1O100OnN_LgF`3W9cLiF]3T9fLlFY3R9jLnFV3o8mLQGS3k8PMVGP3g8QM[Gn2c8TM^Gl2_8UMcGk2Z8VMhGj2U8WMmGh2P8[MQHe2l7]MUHc2h7_MYHa2d7bM\\H]2b7eM_H[2]7hMdHX2Y7jMhHV2V7lMjHT2X7jMhHV2Y7iMgHW2Z7gMgHY2Z7fMfHZ2[7eMeH[2\\7dMdH\\2^7aMcH_2^7`MbH`2_7`M`H`2a7_M_Ha2b7_M]Ha2d7^M\\Hb2e7]M[Hc2g7\\MXHd2i7[MWHe2j3cLeNh0`Me2j3eLfNe0`Mf2j3fLfNc0`Mg2j3gLfNb0_Mg2k3hLfN`0_Mi2i3iLgN>`Mi2i3jLfN>`Mh2j3kLeN>`Mg2k3lLdN>`Mf2k3nLdN=`Me2l3oLcN>_Md2m3oLcN>_Mc2n3PMbN>_Mb2n3RMbN=_Ma2o3SMaN=_M`2P4TM`N=_M_2Q4UM_N=_M_2Q4UM_N=_M^2R4VM]N>`M\\2S4WM\\N>`M[2T4XM[N>`MZ2U4YMYN?aMX2V4ZMXN?aMW2W4ZMXNa0_MV2X4ZMXNa0_MU2Y4[MVNb0`MS2Z4\\MUNb0`MR2[4]MTNb0`MQ2\\4]MTNc0_MP2]4^MSNc0_MP2]4^MTNb0^MP2^4_MSNb0^MP2^4^MUNb0\\MQ2]4_MVNa0\\MQ2]4_MWN`0[MR2]4_MWN`0[MQ2^4aMVN?ZMQ2_4eMSN:]MR2_4iMoM6aMR2^4kMPN3aMS2^4lMoM2bMS2^4mMoM0bMS2_4oMmMOcMS2_4PNmMMcMT2^4QNnMLcMT2]4RNPNJbMU2[4TNSNGaMV2Y4VNUNEaMV2X4WNWNe3g1_LWNb3f1aLZN_3c1dL\\N]3a1fL_NZ3]1jLcNV3Z1mLeNT3X1PMgNP3U1TMjNl2k0`MUO`2h0cMWO^2g0eMXO[2f0gMYOZ2f0gMZOY2d0iM\\OW27nIlMl3m1W25QJjMj3P2U24VJiMf3S2T2VOTJXOV4b1g1mN\\J]OQ4e1d1kN^J\\OR4i1`1iN`JZOT4l1]1hNaJYOT4P2_1aN_J[OV4S2c1YNZOf1m0QNUOP2n6O1O010O1O10O01O10O01O100O010O1O010O1O010O00O2M2N2N3N1N2N3M2O1N3M2N2O1N2N2NO2L3N200100O2N2N101N1OARDPNl;o1ZDnMd;R2aDjM^;U2iDgMU;Y2h00O0100001N3N1O1O2N1N3N1O2N2N2M3N2N1O2N2M3N2mNkLYFW3e9PMTFQ3k9WMgEQ3W:m0M3N2N00N2N2N1O2O1N2N3M3M3M3N2M3M3M3M4M2M3M3O1O1O2O1N1O2N2N2N2N2N2N2N2N200O100O100O2O0O100O1O4M2M4M3L3N3K5K4L4L5K4K6K4L5K4K;@jm[4"}, "label": "a man holding an umbrella over a woman"}]} {"id": 19874, "image": "COCO_train2014_000000019874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black car with one tire showing\"."}], "task": "refering", "answer_template": "The \"the black car with one tire showing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 113, 114, 115, 116, 117, 118, 126, 127, 128, 134, 135, 142, 143, 144, 152, 159, 160, 161, 169, 177, 178, 186], "bbox": [0.38606250000000003, 0.25740625, 0.9996666666666666, 0.462328125], "iscrowd": 0, "area": 9293.77275, "rle": {"size": [640, 480], "counts": "ekc35gc0`0_O8K4L4L4L4N1N2O1N3N3L3N2M3N2M2O2M3N00000O1000000O2O00000O100000000O10000000000O101O00000O01O1O1O1O0O2M3N110000O1O1O1O1O1OROj^OTOUa0b0^_OVOb`0c0i_OYOV`0e0P@XOo?f0f1O1N2N2O2N1O1O1O1O2N1O100000001O0000001O0001O010O001O010O010O10O10O010O01000O0100O010O10O10O100000000000O100000000000000O1000001O0000000O1000001O000000000O10001OG9O100O2Nok60RTI2M1000O10000000000000000Oog=0RXB00000000000000000000000000O10O01Ook6;hSI6I010O010O00001O001O1O001N2O001O1O1O001O1O1O1O1O1O1O1O2M2O1O2N1O2N1O2N2N2N2N2N2N2N6i_OoMS>W2[ATNe>S2j@WNW?Y30O1O1O100O1O1N2N2N3N2N3N1O1N3N1O1N3N1O1G:jM_L"}, "label": "the black car with one tire showing"}]} {"id": 19874, "image": "COCO_train2014_000000019874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black car that is behind the green car\"."}], "task": "refering", "answer_template": "The \"black car that is behind the green car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 113, 114, 115, 116, 117, 118, 126, 127, 128, 134, 135, 142, 143, 144, 152, 159, 160, 161, 169, 177, 178, 186], "bbox": [0.38606250000000003, 0.25740625, 0.9996666666666666, 0.462328125], "iscrowd": 0, "area": 9293.77275, "rle": {"size": [640, 480], "counts": "ekc35gc0`0_O8K4L4L4L4N1N2O1N3N3L3N2M3N2M2O2M3N00000O1000000O2O00000O100000000O10000000000O101O00000O01O1O1O1O0O2M3N110000O1O1O1O1O1OROj^OTOUa0b0^_OVOb`0c0i_OYOV`0e0P@XOo?f0f1O1N2N2O2N1O1O1O1O2N1O100000001O0000001O0001O010O001O010O010O10O10O010O01000O0100O010O10O10O100000000000O100000000000000O1000001O0000000O1000001O000000000O10001OG9O100O2Nok60RTI2M1000O10000000000000000Oog=0RXB00000000000000000000000000O10O01Ook6;hSI6I010O010O00001O001O1O001N2O001O1O1O001O1O1O1O1O1O1O1O2M2O1O2N1O2N1O2N2N2N2N2N2N2N6i_OoMS>W2[ATNe>S2j@WNW?Y30O1O1O100O1O1N2N2N3N2N3N1O1N3N1O1N3N1O1G:jM_L"}, "label": "black car that is behind the green car"}]} {"id": 19874, "image": "COCO_train2014_000000019874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gold mid - sized vehicle\"."}], "task": "refering", "answer_template": "The \"a gold mid - sized vehicle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 107, 108, 119, 120, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 141, 142, 153, 154, 155, 156, 157, 158, 159], "bbox": [0.0022708333333333335, 0.28535937499999997, 0.38808333333333334, 0.43862500000000004], "iscrowd": 0, "area": 13927.656299999999, "rle": {"size": [640, 480], "counts": "Sk0P1nb03N2M?B1O1O100000O01000000O1000O10O1000000N2L30000001O2N1O2ON2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1000000O10000000000000000000000000000000000000000000000000000000000O10000000000000002N1O2N1O2N1O2N1O2N2N00O10000000000O10000000O10O1N2M3M3O10000O10000000000O10000000000O1000000000000O1000O100000O10000000000O10000000000001O1O1O1O1O1O1O1O001OO103M3L5L3M3L4M4L3L4M3L4M4L3H8DUag5"}, "label": "a gold mid - sized vehicle"}]} {"id": 19874, "image": "COCO_train2014_000000019874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light green car\"."}], "task": "refering", "answer_template": "The \"a light green car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 106, 107, 108, 119, 120, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 141, 142, 153, 154, 155, 156, 157, 158, 159], "bbox": [0.0022708333333333335, 0.28535937499999997, 0.38808333333333334, 0.43862500000000004], "iscrowd": 0, "area": 13927.656299999999, "rle": {"size": [640, 480], "counts": "Sk0P1nb03N2M?B1O1O100000O01000000O1000O10O1000000N2L30000001O2N1O2ON2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1000000O10000000000000000000000000000000000000000000000000000000000O10000000000000002N1O2N1O2N1O2N1O2N2N00O10000000000O10000000O10O1N2M3M3O10000O10000000000O10000000000O1000000000000O1000O100000O10000000000O10000000000001O1O1O1O1O1O1O1O001OO103M3L5L3M3L4M4L3L4M3L4M4L3H8DUag5"}, "label": "a light green car"}]} {"id": 404899, "image": "COCO_train2014_000000404899.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two stacked microwaves\"."}], "task": "refering", "answer_template": "The \"two stacked microwaves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349, 350, 351, 352, 368, 369, 370, 371, 372, 373, 374, 375, 392, 393], "bbox": [0.0303125, 0.5396666666666667, 0.340296875, 0.9587254901960786], "iscrowd": 0, "area": 38340.80345000001, "rle": {"size": [510, 640], "counts": "eg96W56n4c1kFk0T92oE0P:0PF1o90PF1n90SFOm91SF0l91SF0l90UF0j90VF1i9OXF0h91WF0h90XF1g9OZF1e90ZF0e91\\FOc91]F0b90_F0`91_F0`90`F0`90aF0^91aF0^90cF0\\90dF0\\91cF0\\90eF0Z90fF1Y9OhF1W90hF0X90hF0X90iFOW92hFMY93hFLX94hFLX94iFKW96hFJX96hFJX96hFJX96hFIY97gFIY97gFIY97gFIX98gFIY97gFIY97gFIY97gFHZ98fFHZ98fFHZ97gFIY97gFIY97gFIY97gFIY97fFI[97eFI[97eFI[97eFI[97eFI[97eFI[97eFI[97eFH\\97eFI[97dFJ\\96dFJ\\96dFJ\\96dFJ\\96dFJ\\96dFI]97cFI]97cFI\\98dFH\\98dFH\\98cFI]96dFJ\\96dFI]97cFI]97cFI]97cFI]97cFI]97cFI]97cFH^98aFI_97aFI_97aFI_96bFJ^96bFJ^96bFJ^96bFI_97aFI_97aFI_97`FJ`96`FJ`96`FJ`96`FJ`96`FIa97_FIa96`FJ_97aFI_97aFI_97aFI_97`FJ`96`FIa97_FIa97_FIa97_FIa97_FIa97_FIa97_FIa96`FIa97^FJb96^FJb96^FJb96^FJb96^FJb96^FJb96^FIc97]FIc97]FIc97\\FJd96\\FJd95]FKc95]FKc95]FJd96\\FJc97]FIc97]FIc97]FIc97]FIc97\\FJd96\\FIe97[FIe97[FIe96\\FJd96\\FJd96\\FJd96\\FJd96\\FIe97ZFJf96ZFJf96ZFJf96ZFJf96ZFJf96ZFIg96ZFJf96ZFJf96ZFJf96YFKg95YFKg95YFKg95YFJh96XFJg97YFIg97YFIg97YFIg97YFIg96ZFJf96YFJh96XFJh96XFJh96XFJh96XFJi95WFKi95WFKi95WFJj96VFJj96UFKl94TFLl93UFMk93UFMk93UFMl92TFMm93SFMm93SFMm93SFMm93RFNo91QFOo91QFOo91QFNP:2PFNP:1QFOP:0PF0P:0PF0P:0PF0P:0PF0P:0oE0S:OmE1S:OmE1S:OmE1S:OmE1T:NlE2T:NlE2T:MmE2T:NlE2T:NkE3V:LjE4V:LjE4V:LjE4\\:FdE9S in this image.", "objects": [{"patches": [208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 322, 323, 324, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349, 350, 351, 352, 368, 369, 370, 371, 372, 373, 374, 375, 392, 393], "bbox": [0.0303125, 0.5396666666666667, 0.340296875, 0.9587254901960786], "iscrowd": 0, "area": 38340.80345000001, "rle": {"size": [510, 640], "counts": "eg96W56n4c1kFk0T92oE0P:0PF1o90PF1n90SFOm91SF0l91SF0l90UF0j90VF1i9OXF0h91WF0h90XF1g9OZF1e90ZF0e91\\FOc91]F0b90_F0`91_F0`90`F0`90aF0^91aF0^90cF0\\90dF0\\91cF0\\90eF0Z90fF1Y9OhF1W90hF0X90hF0X90iFOW92hFMY93hFLX94hFLX94iFKW96hFJX96hFJX96hFJX96hFIY97gFIY97gFIY97gFIX98gFIY97gFIY97gFIY97gFHZ98fFHZ98fFHZ97gFIY97gFIY97gFIY97gFIY97fFI[97eFI[97eFI[97eFI[97eFI[97eFI[97eFI[97eFH\\97eFI[97dFJ\\96dFJ\\96dFJ\\96dFJ\\96dFJ\\96dFI]97cFI]97cFI\\98dFH\\98dFH\\98cFI]96dFJ\\96dFI]97cFI]97cFI]97cFI]97cFI]97cFI]97cFH^98aFI_97aFI_97aFI_96bFJ^96bFJ^96bFJ^96bFI_97aFI_97aFI_97`FJ`96`FJ`96`FJ`96`FJ`96`FIa97_FIa96`FJ_97aFI_97aFI_97aFI_97`FJ`96`FIa97_FIa97_FIa97_FIa97_FIa97_FIa97_FIa96`FIa97^FJb96^FJb96^FJb96^FJb96^FJb96^FJb96^FIc97]FIc97]FIc97\\FJd96\\FJd95]FKc95]FKc95]FJd96\\FJc97]FIc97]FIc97]FIc97]FIc97\\FJd96\\FIe97[FIe97[FIe96\\FJd96\\FJd96\\FJd96\\FJd96\\FIe97ZFJf96ZFJf96ZFJf96ZFJf96ZFJf96ZFIg96ZFJf96ZFJf96ZFJf96YFKg95YFKg95YFKg95YFJh96XFJg97YFIg97YFIg97YFIg97YFIg96ZFJf96YFJh96XFJh96XFJh96XFJh96XFJi95WFKi95WFKi95WFJj96VFJj96UFKl94TFLl93UFMk93UFMk93UFMl92TFMm93SFMm93SFMm93SFMm93RFNo91QFOo91QFOo91QFNP:2PFNP:1QFOP:0PF0P:0PF0P:0PF0P:0PF0P:0oE0S:OmE1S:OmE1S:OmE1S:OmE1T:NlE2T:NlE2T:MmE2T:NlE2T:NkE3V:LjE4V:LjE4V:LjE4\\:FdE9S in this image.", "objects": [{"patches": [222, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 274, 290, 291, 292, 293, 294, 295, 296, 297, 313, 314, 315, 316, 317, 318, 319, 320], "bbox": [0.61959375, 0.5460588235294118, 0.9580468750000002, 0.7775294117647059], "iscrowd": 0, "area": 22397.519199999995, "rle": {"size": [510, 640], "counts": "SSV63h?4dM3bD0Z;=ZDFb;k0nCYOm;]1^CfN^<_2O1O1O2N1O1O1000001O0000000000001O0000000000001O00000000001O000001O000001O00000000000010O000000000001O000001O0001O0000000001O01O0000000000010O0000000000000001O00O10000000000000000000000O10000000000000000000000O100000000000000O100000000000O010000000000O1000000000000O1000000000000O10000000000O10000000000000000000000000O1_OkCYMU in this image.", "objects": [{"patches": [222, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 274, 290, 291, 292, 293, 294, 295, 296, 297, 313, 314, 315, 316, 317, 318, 319, 320], "bbox": [0.61959375, 0.5460588235294118, 0.9580468750000002, 0.7775294117647059], "iscrowd": 0, "area": 22397.519199999995, "rle": {"size": [510, 640], "counts": "SSV63h?4dM3bD0Z;=ZDFb;k0nCYOm;]1^CfN^<_2O1O1O2N1O1O1000001O0000000000001O0000000000001O00000000001O000001O000001O00000000000010O000000000001O000001O0001O0000000001O01O0000000000010O0000000000000001O00O10000000000000000000000O10000000000000000000000O100000000000000O100000000000O010000000000O1000000000000O1000000000000O10000000000O10000000000000000000000000O1_OkCYMU in this image.", "objects": [{"patches": [0, 1, 2, 17, 18, 19, 34, 35, 36, 37, 38, 52, 53, 54, 55, 69, 70, 71, 72, 86, 87], "bbox": [0.0, 0.0, 0.26778242677824265, 0.238203125], "iscrowd": 0, "area": 9188.805800000002, "rle": {"size": [640, 478], "counts": "3:fc0e0[Oe0[O7I0000000O10001O0000000000000000001O0000000000000000001O00003L7J6J6J6J6J6J6J6J6J7I6J6J6J5K0000000000000000000000dKi@a3d?`L]@b2``0]Mc_Oe1Ya0ZNi^O]1_a0bNd^O^1Za0aNh^O_1Wa0`Nl^O`1Ra0_NP_Ob1n`0]NU_Oc1i`0\\NZ_Oc1e`0\\N]_Oe1a`0[Na_Oe1]`0ZNe_Of1Z`0YNi_Og1Xa010O0100O10O0100O01000O010O100O010O10O0100O010O1N01O01O00001O00001O0001O0O100000000000000O1M4I6K5J6J6J6J6J6J6Jmej6"}, "label": "a dark table in the far left hand background"}]} {"id": 404905, "image": "COCO_train2014_000000404905.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white chair near table\"."}], "task": "refering", "answer_template": "The \"a white chair near table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 38, 39, 40, 41, 42, 43, 44, 45, 55, 56, 57, 58, 59, 60, 61, 62, 72, 73, 74, 75, 76, 77, 78, 79, 89, 90, 91, 92, 93, 94, 95], "bbox": [0.23497907949790794, 0.082703125, 0.6986610878661087, 0.238546875], "iscrowd": 0, "area": 18982.51485, "rle": {"size": [640, 478], "counts": "`hV24hc04M3L4M3L4H8F:G9G9G9G9G9G9O100000000000001O0000000000000001O0000000000000000000000000000000000000000000000000000000000000000010O00000000000000000000000000000000000000000000000000000000000000001O000000000000000001O000000000000000000000000000000000000000000000001O000000000000000000000000000000000001O0000000000000000000000000001O000000000000000000000000000000000000000000000000000001O00000000L4M4K4L4L4M3L4L4M3L4L4M3L4L4M3L4L4M3L4L4M3L4L4M3LYni2"}, "label": "a white chair near table"}]} {"id": 404905, "image": "COCO_train2014_000000404905.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white chair at table\"."}], "task": "refering", "answer_template": "The \"white chair at table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 38, 39, 40, 41, 42, 43, 44, 45, 55, 56, 57, 58, 59, 60, 61, 62, 72, 73, 74, 75, 76, 77, 78, 79, 89, 90, 91, 92, 93, 94, 95], "bbox": [0.23497907949790794, 0.082703125, 0.6986610878661087, 0.238546875], "iscrowd": 0, "area": 18982.51485, "rle": {"size": [640, 478], "counts": "`hV24hc04M3L4M3L4H8F:G9G9G9G9G9G9O100000000000001O0000000000000001O0000000000000000000000000000000000000000000000000000000000000000010O00000000000000000000000000000000000000000000000000000000000000001O000000000000000001O000000000000000000000000000000000000000000000001O000000000000000000000000000000000001O0000000000000000000000000001O000000000000000000000000000000000000000000000000000001O00000000L4M4K4L4L4M3L4L4M3L4L4M3L4L4M3L4L4M3L4L4M3L4L4M3LYni2"}, "label": "white chair at table"}]} {"id": 11690, "image": "COCO_train2014_000000011690.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the motorcycle the man is riding\"."}], "task": "refering", "answer_template": "The \"the motorcycle the man is riding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 76, 77, 78, 99, 100, 101, 121, 122, 123, 124, 125, 131, 144, 145, 146, 147, 148, 149, 151, 152, 153, 154, 166, 167, 168, 169, 170, 171, 173, 174, 175, 176, 177, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 259, 267, 268], "bbox": [0.194453125, 0.19774647887323943, 0.746390625, 0.7483098591549295], "iscrowd": 0, "area": 39961.87700000001, "rle": {"size": [426, 640], "counts": "Tkc12T=:F;E;E9F7J5K5M4K4L5L3L5L2M4N100O10001O0O10001O0O10001O000000001O0000001O000000001O0000000000000000000000001O0000000000000bNSM_Hm2]7[M_He2]7bM_H_2]7hM`HX2\\7PN`HP2^7UN_Hk1a7o1O1O100O001O100O11O000000000000000O10PNmJjKR5V4QKgKo4Y4SKeKm4[4VKbKj4^4XK`Kh4`4ZK^Ke4c4^KZKb4f4`KXK_4i4dKnJb4R5`KYJS5g5W11WM^JjMb5U2bJgM_5Y2dJdM\\5]2gJ^MZ5b2iJ[MW5f2kJVMV5k2lJRMT5o2oJlLR52fIf1]1SNm47hIe1^1oMk4;iIf1^1lMi4>kIe1^1iMh4a0mIe1]1gMf4d0oId1]1dMe4g0QJQ2k7oMWHU2c7jM^H\\2\\7dMdHb2V7^MjHg2Q7YMoHh2P7XMPIj2n6VMRIl2l6SMUIn2j6RMVIP3h6PMYIQ3e6oL[IS3c6mL^IS3a6mLaIR3_6lLcIT3\\6lLeIT3Z6lLhIS3W6mLjIT3T6kLnIU3R6jLPJU3o5kLRJU3m5kLSJV3l5iLUJX3k5gLUJY3k5gLUJY3k5gLUJZ3j5fLWJY3i5gLWJY3i5gLWJY3i5gLXJY3g5gLYJY3g5hLXJX3h5hLXJX3h5gLZJX3f5gL[JZ3d5eL]J[3c5cL`J\\3c5`L^J`3e5\\L\\Jd3g5XLZJh3j5SLXJl3k5oKWJQ4a701O00001O1]NjFQOU9h0TGVOm8b0[G]Of8=`GB`8=cGA^8>cGA^8=eGA[8?fG@[8>gGAY8>jG@W8?jG@W8>lG@T8`0mG_OT8?nG@S8?oG_OQ8`0RH^Oo7a0TH\\Om7b0WH[Oj7c0YH[Og7e0\\HXOe7f0_HWOb7h0aHUO`7i0cHUO^7e0hH[OX7`0nH^OR7=TIB[:O2N1O1Od^21ZaM2N3M4L4L4L4L4L3L5L4K5L2M3N2M4M2M3M3N1N2O1N3N1N2O1N2O1N2O1O1N2O1N3N1N3N2M3N2M3N1N3N2N2M3N2M2O2M3N2N2N2O0O2N1O2O0O1O2N100O2O0000001O000010O0001O0000001O00001O00000000000000000000000O1000000O10000O0100O10O0100O010O100O010O10O0100O10O0100O010O100O100O100O1SMoFa1Q9^NRG_1o8aNRG^1n8aNUG\\1l8cNWG[1i8dNYGZ1h8eN[GY1f8fN\\GW1e8WNoFH>P2c8XNXG^O7X2b8ZNoGe1Q8ZNQHd1P8\\NPHd1P8\\NQHb1P8^NQHa1o7^NRHa1o7_NRH`1n7`NSH^1n7aNSH^1o7aNRH]1o7bNRH]1o7bNSH\\1n7bNUH\\1l7cNUH\\1l7bNWH\\1j7cNaHQ1a7mNmHf0T7YOoHc0S7[OQIb0P7]OSI`0n9L5LXTS2"}, "label": "the motorcycle the man is riding"}]} {"id": 11690, "image": "COCO_train2014_000000011690.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red motor cycle\"."}], "task": "refering", "answer_template": "The \"the red motor cycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 76, 77, 78, 99, 100, 101, 121, 122, 123, 124, 125, 131, 144, 145, 146, 147, 148, 149, 151, 152, 153, 154, 166, 167, 168, 169, 170, 171, 173, 174, 175, 176, 177, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 259, 267, 268], "bbox": [0.194453125, 0.19774647887323943, 0.746390625, 0.7483098591549295], "iscrowd": 0, "area": 39961.87700000001, "rle": {"size": [426, 640], "counts": "Tkc12T=:F;E;E9F7J5K5M4K4L5L3L5L2M4N100O10001O0O10001O0O10001O000000001O0000001O000000001O0000000000000000000000001O0000000000000bNSM_Hm2]7[M_He2]7bM_H_2]7hM`HX2\\7PN`HP2^7UN_Hk1a7o1O1O100O001O100O11O000000000000000O10PNmJjKR5V4QKgKo4Y4SKeKm4[4VKbKj4^4XK`Kh4`4ZK^Ke4c4^KZKb4f4`KXK_4i4dKnJb4R5`KYJS5g5W11WM^JjMb5U2bJgM_5Y2dJdM\\5]2gJ^MZ5b2iJ[MW5f2kJVMV5k2lJRMT5o2oJlLR52fIf1]1SNm47hIe1^1oMk4;iIf1^1lMi4>kIe1^1iMh4a0mIe1]1gMf4d0oId1]1dMe4g0QJQ2k7oMWHU2c7jM^H\\2\\7dMdHb2V7^MjHg2Q7YMoHh2P7XMPIj2n6VMRIl2l6SMUIn2j6RMVIP3h6PMYIQ3e6oL[IS3c6mL^IS3a6mLaIR3_6lLcIT3\\6lLeIT3Z6lLhIS3W6mLjIT3T6kLnIU3R6jLPJU3o5kLRJU3m5kLSJV3l5iLUJX3k5gLUJY3k5gLUJY3k5gLUJZ3j5fLWJY3i5gLWJY3i5gLWJY3i5gLXJY3g5gLYJY3g5hLXJX3h5hLXJX3h5gLZJX3f5gL[JZ3d5eL]J[3c5cL`J\\3c5`L^J`3e5\\L\\Jd3g5XLZJh3j5SLXJl3k5oKWJQ4a701O00001O1]NjFQOU9h0TGVOm8b0[G]Of8=`GB`8=cGA^8>cGA^8=eGA[8?fG@[8>gGAY8>jG@W8?jG@W8>lG@T8`0mG_OT8?nG@S8?oG_OQ8`0RH^Oo7a0TH\\Om7b0WH[Oj7c0YH[Og7e0\\HXOe7f0_HWOb7h0aHUO`7i0cHUO^7e0hH[OX7`0nH^OR7=TIB[:O2N1O1Od^21ZaM2N3M4L4L4L4L4L3L5L4K5L2M3N2M4M2M3M3N1N2O1N3N1N2O1N2O1N2O1O1N2O1N3N1N3N2M3N2M3N1N3N2N2M3N2M2O2M3N2N2N2O0O2N1O2O0O1O2N100O2O0000001O000010O0001O0000001O00001O00000000000000000000000O1000000O10000O0100O10O0100O010O100O010O10O0100O10O0100O010O100O100O100O1SMoFa1Q9^NRG_1o8aNRG^1n8aNUG\\1l8cNWG[1i8dNYGZ1h8eN[GY1f8fN\\GW1e8WNoFH>P2c8XNXG^O7X2b8ZNoGe1Q8ZNQHd1P8\\NPHd1P8\\NQHb1P8^NQHa1o7^NRHa1o7_NRH`1n7`NSH^1n7aNSH^1o7aNRH]1o7bNRH]1o7bNSH\\1n7bNUH\\1l7cNUH\\1l7bNWH\\1j7cNaHQ1a7mNmHf0T7YOoHc0S7[OQIb0P7]OSI`0n9L5LXTS2"}, "label": "the red motor cycle"}]} {"id": 11690, "image": "COCO_train2014_000000011690.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white shirt , riding a motorcycle , with red helmet\"."}], "task": "refering", "answer_template": "The \"a man in a white shirt , riding a motorcycle , with red helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 56, 57, 58, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 125, 126, 127, 128, 149, 150, 151, 171, 172, 173, 194, 195, 217, 218, 241, 263, 264, 265], "bbox": [0.36709375, 0.08124413145539906, 0.587609375, 0.7986150234741785], "iscrowd": 0, "area": 15155.083050000005, "rle": {"size": [426, 640], "counts": "^lQ33U=5K4M2N101N100O10O010O2N101N101O001O001O01OO1000O101N101N101N101N101O0O2O001N10001N10000O2O0O101N100O2O0O100cFWOY6i0eI[OKJb3l0`L@JEf3k0^LEHBU2GQMS1P1HG_OW2HQMR1m0LH[OX2IRMQ1l00GVOZ2JSMP1j04FSO\\2JSMP1i07EPO]2LTMm0g0=DkN`2LUMl0e0a0CiNa2KWMk0c0e0BfNb2MWMj0b0g0BcNd2MXMi0?m0A^Ng2MYMh0=R1_OZNk2LYMg0 in this image.", "objects": [{"patches": [33, 34, 35, 56, 57, 58, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 125, 126, 127, 128, 149, 150, 151, 171, 172, 173, 194, 195, 217, 218, 241, 263, 264, 265], "bbox": [0.36709375, 0.08124413145539906, 0.587609375, 0.7986150234741785], "iscrowd": 0, "area": 15155.083050000005, "rle": {"size": [426, 640], "counts": "^lQ33U=5K4M2N101N100O10O010O2N101N101O001O001O01OO1000O101N101N101N101N101O0O2O001N10001N10000O2O0O101N100O2O0O100cFWOY6i0eI[OKJb3l0`L@JEf3k0^LEHBU2GQMS1P1HG_OW2HQMR1m0LH[OX2IRMQ1l00GVOZ2JSMP1j04FSO\\2JSMP1i07EPO]2LTMm0g0=DkN`2LUMl0e0a0CiNa2KWMk0c0e0BfNb2MWMj0b0g0BcNd2MXMi0?m0A^Ng2MYMh0=R1_OZNk2LYMg0 in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71], "bbox": [0.005059999999999999, 0.09437333333333334, 0.99944, 0.2921333333333333], "iscrowd": 0, "area": 30033.195050000002, "rle": {"size": [375, 500], "counts": "fT1o0h:h0XO0000001O000000010O000000001O0000001O0000000010O00000001O00000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000O10000000000000000000000000000000000O1000000000000000000000000000000000O10O10000000000000000000000000000000000O100000000000000000000000000000000O10000O10000O100O10000O10000O10000000O10O1000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000O10000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000O1000000000000000O10000000000000000000000000000000000000000000000000O10000001O00001O0000001O00001O0000001O00001O0000001O00001O0O10001O000000000000001O00000000000000001O0000000000000000001O0000000000000000001O000000000000000000000000O100000O1000000000001O0000000000000000U1kN[M"}, "label": "a plate of food that is furthest from the viewer"}]} {"id": 421298, "image": "COCO_train2014_000000421298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"background of table top with plate on top\"."}], "task": "refering", "answer_template": "The \"background of table top with plate on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71], "bbox": [0.005059999999999999, 0.09437333333333334, 0.99944, 0.2921333333333333], "iscrowd": 0, "area": 30033.195050000002, "rle": {"size": [375, 500], "counts": "fT1o0h:h0XO0000001O000000010O000000001O0000001O0000000010O00000001O00000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000O10000000000000000000000000000000000O1000000000000000000000000000000000O10O10000000000000000000000000000000000O100000000000000000000000000000000O10000O10000O100O10000O10000O10000000O10O1000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000O10000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000O1000000000000000O10000000000000000000000000000000000000000000000000O10000001O00001O0000001O00001O0000001O00001O0000001O00001O0O10001O000000000000001O00000000000000001O0000000000000000001O0000000000000000001O000000000000000000000000O100000O1000000000001O0000000000000000U1kN[M"}, "label": "background of table top with plate on top"}]} {"id": 421298, "image": "COCO_train2014_000000421298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of a sandwich in back of another half\"."}], "task": "refering", "answer_template": "The \"half of a sandwich in back of another half\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 66, 67, 68, 69, 70, 81, 82, 83, 84, 85, 86, 87, 88, 100, 101, 102, 117, 118, 136], "bbox": [0.52622, 0.28229333333333334, 0.9359400000000001, 0.5641866666666666], "iscrowd": 0, "area": 7118.5513, "rle": {"size": [375, 500], "counts": "S^P31d;3M2N3M3M2N3M3ZF]OT8e0kG_OP8c0oGBk7`0UHCf7?YHFc7:\\HJa76^HO^72`H2]7NbH6[7JdH9Z7=oGDQ8 in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 26, 27, 28, 29, 32], "bbox": [0.43426, 0.00224, 0.8427, 0.1311466666666667], "iscrowd": 0, "area": 5766.135200000001, "rle": {"size": [375, 500], "counts": "a__25W;=L4M3M3N2N2N3M2N2N2N2N2N00001O0000000001N1O1O1N200O100000000O2N1O100O1O1O1O1O1O1O1O101N1O1O10000000000000001O00000000000001O000001O0000000000000001O000001O0000000001O000000O100000O10O1001O1O001O0O2O000000001O00000O0100000000O1000O10O10000O10000O10O100000O10000000000000O010000000000000000O1000000000000000000000000000O10001O001O001O001O001O1O001O001O1O001O00001O0000001N10001O00001O00M3N2N2M3N2N2M3M3L4M3Ioml0"}, "label": "a plastic chair behind the far table"}]} {"id": 421298, "image": "COCO_train2014_000000421298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwich\"."}], "task": "refering", "answer_template": "The \"the sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 89, 100, 101, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212], "bbox": [0.56838, 0.3085333333333333, 1.0, 0.8556], "iscrowd": 0, "area": 36362.164650000006, "rle": {"size": [375, 500], "counts": "^XX32a91\\H5^71\\H5^71]H4]71^H5\\71^H5\\71^H5^7N^H7_7L[H:c7GXH?e7DUH`0j7CQHa0n7@mGe0Q8b1N2O1N2O1N1O2O1N2O1N2N2O1N1O2O1N2O1N2N2O1N1O2O1N2O1N2N2O1N101N2N100O1O1000000O1000001N1000000O1000000O1000000O10001O000O1000000O1000000O1000000O101O000O1000000O100000000O1000001N1000000O1000000O1000000O1000001N1000000O10O100000O1000O10O1000000O0100000O100000O01000000O10O100000O1000O10O1000000O0100000O1000O10O1000000O10O100O10O0100O10O0100O10O0100O10O0100O100O2O0O2O0O2O0O2N101N101N100O2O0O2O0O2O0O2O0Om0TOVK"}, "label": "the sandwich"}]} {"id": 421298, "image": "COCO_train2014_000000421298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wheat bread sandwhich\"."}], "task": "refering", "answer_template": "The \"a wheat bread sandwhich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 89, 100, 101, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212], "bbox": [0.56838, 0.3085333333333333, 1.0, 0.8556], "iscrowd": 0, "area": 36362.164650000006, "rle": {"size": [375, 500], "counts": "^XX32a91\\H5^71\\H5^71]H4]71^H5\\71^H5\\71^H5^7N^H7_7L[H:c7GXH?e7DUH`0j7CQHa0n7@mGe0Q8b1N2O1N2O1N1O2O1N2O1N2N2O1N1O2O1N2O1N2N2O1N1O2O1N2O1N2N2O1N101N2N100O1O1000000O1000001N1000000O1000000O1000000O10001O000O1000000O1000000O1000000O101O000O1000000O100000000O1000001N1000000O1000000O1000000O1000001N1000000O10O100000O1000O10O1000000O0100000O100000O01000000O10O100000O1000O10O1000000O0100000O1000O10O1000000O10O100O10O0100O10O0100O10O0100O10O0100O100O2O0O2O0O2O0O2N101N101N100O2O0O2O0O2O0O2O0Om0TOVK"}, "label": "a wheat bread sandwhich"}]} {"id": 421298, "image": "COCO_train2014_000000421298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"salad on a plate\"."}], "task": "refering", "answer_template": "The \"salad on a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 81, 92, 93, 94, 95, 96, 97, 98, 99, 100, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226], "bbox": [0.0033599999999999997, 0.3276266666666667, 0.58184, 0.98456], "iscrowd": 0, "area": 62011.02464999999, "rle": {"size": [375, 500], "counts": "_m0`5W60O100O100O100O100O10000O100O100O100O10000O100O100O100O100O10000O100O100O100O10000O100O100O100O100O10000O100O100O100O10000O100O100O100O100O10000O100O100O100O1000000O100000000O1000000O1000000O1000000O100000000O1000000O1000000O100000000O1000000O0100000O1000000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O10000O100000000O10000000000O10000000000O10000000000O100000000O10000000000O10000000000O100000000000000000000O10000000000000000000000O100000000000000000000001O0000001O000000001O000`HULT7X4O0000001iNbKSK^4h4iKVKW4d4QLZKP4`4WL^Ki3\\4^LcKb3X4eLfK\\3T4kLjKU3P4SMnKm2P4WMnKj2o3YMjKTNHd4[4[MhKVNL`4X4]MgKYN0[4c2XKnN`7k0jHoN]7f0nHSO\\7?nHZOb73iHFi9J6J7HTY\\2"}, "label": "salad on a plate"}]} {"id": 421298, "image": "COCO_train2014_000000421298.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a side dish consisting of radishes and lettuce\"."}], "task": "refering", "answer_template": "The \"a side dish consisting of radishes and lettuce\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 81, 92, 93, 94, 95, 96, 97, 98, 99, 100, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226], "bbox": [0.0033599999999999997, 0.3276266666666667, 0.58184, 0.98456], "iscrowd": 0, "area": 62011.02464999999, "rle": {"size": [375, 500], "counts": "_m0`5W60O100O100O100O100O10000O100O100O100O10000O100O100O100O100O10000O100O100O100O10000O100O100O100O100O10000O100O100O100O10000O100O100O100O100O10000O100O100O100O1000000O100000000O1000000O1000000O1000000O100000000O1000000O1000000O100000000O1000000O0100000O1000000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O10000O100000000O10000000000O10000000000O10000000000O100000000O10000000000O10000000000O100000000000000000000O10000000000000000000000O100000000000000000000001O0000001O000000001O000`HULT7X4O0000001iNbKSK^4h4iKVKW4d4QLZKP4`4WL^Ki3\\4^LcKb3X4eLfK\\3T4kLjKU3P4SMnKm2P4WMnKj2o3YMjKTNHd4[4[MhKVNL`4X4]MgKYN0[4c2XKnN`7k0jHoN]7f0nHSO\\7?nHZOb73iHFi9J6J7HTY\\2"}, "label": "a side dish consisting of radishes and lettuce"}]} {"id": 69047, "image": "COCO_train2014_000000069047.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right half of a long sandwich with banana peppers and green spices on the side\"."}], "task": "refering", "answer_template": "The \"the right half of a long sandwich with banana peppers and green spices on the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 218, 219, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247], "bbox": [0.5039374999999999, 0.22304166666666667, 0.9067343749999999, 0.6207916666666667], "iscrowd": 0, "area": 34876.737499999996, "rle": {"size": [480, 640], "counts": "Tcg46e>7H7J7H7J7I>oC_Nk9X2YETNa:X3L4N2O2M2N2O1N3N1N2O1N3N1N2O1N3N1N2N2O2M2O1N2O2M2O1KhJ`FZ5`94O0000001O0000010O00001O0000001O0000001O0000001O0000010O0000001O0000001O0000001O0000001O01O0001O0000001O0000000000000000000001O0000000001O00000000000000000000000001O000000000000000001O000000000001O001O1O1O001O1O1O1O001O1O1O001O100O1O001O1O1O001O1O1O1O0O2N2N2N1O2O1N2N2N1O2N2O1N1O2N2N2N2O0O2N2O1N101O1O1N2O001O1N2O001O1N2O1O1O3L4M3L4L4L5L3L4L4L4L4M0O1O1O1O100O001O1O100O001O1O1O1O00O14M3L4L4L;E1O2N1O1O2N1O2N1O2N1O1O2O0O2N1O1O2N1O2N1O2N1O1O2N1OQmk0"}, "label": "the right half of a long sandwich with banana peppers and green spices on the side"}]} {"id": 69047, "image": "COCO_train2014_000000069047.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"muffaleta with bites missing\"."}], "task": "refering", "answer_template": "The \"muffaleta with bites missing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 99, 100, 101, 102, 103, 104, 120, 121, 122, 123, 124, 125, 126, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 325, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 352, 353, 354, 355], "bbox": [0.042046875, 0.20616666666666666, 0.575109375, 0.9438333333333333], "iscrowd": 0, "area": 70650.25170000001, "rle": {"size": [480, 640], "counts": "il<2n>2N2N2M3N2N2M3N2N2M3N2N2N2M3N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N2M3N2N2M3N2N3L3N2N2M3N2M3N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3M3K5K5L4K5K5K5L4KTJgGY5T8jJlGX5o7kJQHW5k7kJUHW5f7lJZHV5a7mJ_HV5[7mJeHU5V7nJjHT5R7nJnHT5m6oJSIS5j6nJVIT5h6lJXIV5e6kJ[IW5b6jJ^IX5`6hJ`IX5_6hJbIX5^6hJbIW5^6jJbIU5^6lJbIT5^6lJaIT5_6mJaIR5_6oJaIP5`6PK`IP5_6QKaIn4`6RK`Im4`6TK`Il4_6UKaIj4`6VK`Ii4`6XK`Ig4a6XK`Ih4_6YKaIf4_6[KaId4`6\\K`Id4_6]KaIb4`6^K`Ia4`6`K`I_4`6bK`I^4`6bK`I]4`6dK`I[4a6eK_I[4`6fK`IY4`6hK`IW4a6iK`IU4`6lK`IT4_6lKbIS4_6mKaIR4_6oKaIQ4_6oKaIP4_6QLbIn3]6SLcIm3]6SLcIm3\\6TLdIl3\\6TLdIl3[6ULeIk3Z6VLgIj3X6VLhIj3W6WLiIi3W6WLiIi3V6WLkIi3T6XLlIh3T6XLmIg3R6ZLnIf3R6ZLnIf3Q6[LoIe3P6\\LPJd3P6\\LPJd3o5]LQJc3n5^LSJa3m5_LSJa3l5`LTJ`3l5`LTJ`3k5aLUJ_3j5bLVJ^3j5aLXJ^3g5cLYJ]3g5cLYJ^3e5cL[J]3d5dL\\J\\3d5dL\\J\\3c5eL^JZ3b5fL^JZ3a5gL_JY3`5hL`JX3`5hL`JX3_5iLaJW3_5iLbJV3^5jLbJV3^5iLcJW3\\5jLdJV3\\5jLdJV3\\5jLdJW3[5iLeJW3Z5jLgJV3X5jLhJV3X5jLhJV3X5jLhJW3V5jLjJV3V5jLjJW3U5iLlJV3T5jLlJW3S5iLmJW3R5jLnJV3R5iLoJX3P5hLPKX3P5hLQKX3m4iLSKW3m4iLSKX3l4hLTKX3l4hLTKY3j4hLVKX3j4hLWKW3i4iLWKX3h4hLXKX3g4iLYKX3f4hLZKX3f4hLZKY3e4fL\\KZ3c4gL^KX3b4hL^KY3a4gL_KY3a4gL_KZ3_4gLaKY3_4gLaKZ3^4fLcKY3]4gLcKY3\\4hLdKY3[4gLeKY3[4gLeKZ3Z4fLgKY3X4hLhKY3W4gLiKY3W4fLkKY3U4gLkKZ3T4fLlKZ3S4gLnKY3Q4gLoKY3Q4gLoKZ3P4fLPLZ3o3gLRLY3m3gLSLY3m3gLSLY3m3gLTLY3j3hLVLX3j3hLVLY3i3gLXLX3h3gLYLZ3e3gL[LY3e3gL\\LW3e3iL[LW3e3iL[LV3e3kL\\LT3d3lL\\LS3e3mL[LS3e3mL[LR3e3oL\\LP3d3PM\\Lo2e3QM[Lo2e3QM\\Lm2d3TM\\Ll2d3TM\\Lk2e3TM]Lk2c3UM]Lj2c3WM]Li2c3WM^Lg2c3YM]Lg2c3YM]Lf2d3ZM]Le2b3\\M^Lc2c3]M]Lc2c3]M^La2c3_M]La2b3`M^L_2c3aM]L_2c3aM^L]2c3bM^L^2a3cM_L\\2b3dM_L\\1`4dN`K[1a4eN_KZ1a4gN`KX1`4hN`KW1a4iN_KV1b4jN_KU1`4lN`KS1a4mN_KR1b4nN_KQ1a4oN_KP1a4QO_Kn0b4RO^Kn0b4QO`Km0a4SO_Kl0a4UO_Kk0a4UO`Ki0a4WO_Kh0b4XO^Kh0a4YO`Ke0a4[O_Kd0b4\\O^Kd0b4\\O_Kb0a4_O_K`0b4@^K`0b4@_K>b4B^K=c4B^K>a4C_K in this image.", "objects": [{"patches": [79, 80, 81, 99, 100, 101, 102, 103, 104, 120, 121, 122, 123, 124, 125, 126, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 325, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 352, 353, 354, 355], "bbox": [0.042046875, 0.20616666666666666, 0.575109375, 0.9438333333333333], "iscrowd": 0, "area": 70650.25170000001, "rle": {"size": [480, 640], "counts": "il<2n>2N2N2M3N2N2M3N2N2M3N2N2N2M3N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N2M3N2N2M3N2N2M3N2M3N2N2M3N2N3L3N2N2M3N2M3N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3M3K5K5L4K5K5K5L4KTJgGY5T8jJlGX5o7kJQHW5k7kJUHW5f7lJZHV5a7mJ_HV5[7mJeHU5V7nJjHT5R7nJnHT5m6oJSIS5j6nJVIT5h6lJXIV5e6kJ[IW5b6jJ^IX5`6hJ`IX5_6hJbIX5^6hJbIW5^6jJbIU5^6lJbIT5^6lJaIT5_6mJaIR5_6oJaIP5`6PK`IP5_6QKaIn4`6RK`Im4`6TK`Il4_6UKaIj4`6VK`Ii4`6XK`Ig4a6XK`Ih4_6YKaIf4_6[KaId4`6\\K`Id4_6]KaIb4`6^K`Ia4`6`K`I_4`6bK`I^4`6bK`I]4`6dK`I[4a6eK_I[4`6fK`IY4`6hK`IW4a6iK`IU4`6lK`IT4_6lKbIS4_6mKaIR4_6oKaIQ4_6oKaIP4_6QLbIn3]6SLcIm3]6SLcIm3\\6TLdIl3\\6TLdIl3[6ULeIk3Z6VLgIj3X6VLhIj3W6WLiIi3W6WLiIi3V6WLkIi3T6XLlIh3T6XLmIg3R6ZLnIf3R6ZLnIf3Q6[LoIe3P6\\LPJd3P6\\LPJd3o5]LQJc3n5^LSJa3m5_LSJa3l5`LTJ`3l5`LTJ`3k5aLUJ_3j5bLVJ^3j5aLXJ^3g5cLYJ]3g5cLYJ^3e5cL[J]3d5dL\\J\\3d5dL\\J\\3c5eL^JZ3b5fL^JZ3a5gL_JY3`5hL`JX3`5hL`JX3_5iLaJW3_5iLbJV3^5jLbJV3^5iLcJW3\\5jLdJV3\\5jLdJV3\\5jLdJW3[5iLeJW3Z5jLgJV3X5jLhJV3X5jLhJV3X5jLhJW3V5jLjJV3V5jLjJW3U5iLlJV3T5jLlJW3S5iLmJW3R5jLnJV3R5iLoJX3P5hLPKX3P5hLQKX3m4iLSKW3m4iLSKX3l4hLTKX3l4hLTKY3j4hLVKX3j4hLWKW3i4iLWKX3h4hLXKX3g4iLYKX3f4hLZKX3f4hLZKY3e4fL\\KZ3c4gL^KX3b4hL^KY3a4gL_KY3a4gL_KZ3_4gLaKY3_4gLaKZ3^4fLcKY3]4gLcKY3\\4hLdKY3[4gLeKY3[4gLeKZ3Z4fLgKY3X4hLhKY3W4gLiKY3W4fLkKY3U4gLkKZ3T4fLlKZ3S4gLnKY3Q4gLoKY3Q4gLoKZ3P4fLPLZ3o3gLRLY3m3gLSLY3m3gLSLY3m3gLTLY3j3hLVLX3j3hLVLY3i3gLXLX3h3gLYLZ3e3gL[LY3e3gL\\LW3e3iL[LW3e3iL[LV3e3kL\\LT3d3lL\\LS3e3mL[LS3e3mL[LR3e3oL\\LP3d3PM\\Lo2e3QM[Lo2e3QM\\Lm2d3TM\\Ll2d3TM\\Lk2e3TM]Lk2c3UM]Lj2c3WM]Li2c3WM^Lg2c3YM]Lg2c3YM]Lf2d3ZM]Le2b3\\M^Lc2c3]M]Lc2c3]M^La2c3_M]La2b3`M^L_2c3aM]L_2c3aM^L]2c3bM^L^2a3cM_L\\2b3dM_L\\1`4dN`K[1a4eN_KZ1a4gN`KX1`4hN`KW1a4iN_KV1b4jN_KU1`4lN`KS1a4mN_KR1b4nN_KQ1a4oN_KP1a4QO_Kn0b4RO^Kn0b4QO`Km0a4SO_Kl0a4UO_Kk0a4UO`Ki0a4WO_Kh0b4XO^Kh0a4YO`Ke0a4[O_Kd0b4\\O^Kd0b4\\O_Kb0a4_O_K`0b4@^K`0b4@_K>b4B^K=c4B^K>a4C_K in this image.", "objects": [{"patches": [37, 51, 52, 53, 66, 67, 68, 69, 78, 79, 80, 81, 82, 83, 84, 92, 93, 94, 95, 96, 107, 108, 109, 110, 121, 122, 123, 124, 125, 126, 136, 137, 138, 140, 141, 150, 151, 152, 153, 154, 166, 167, 168, 169, 181, 182, 183, 184, 197, 198, 199, 211, 212, 213, 214, 215, 226, 227, 228, 229, 230, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291], "bbox": [0.05346604215456674, 0.12357812500000001, 0.6188758782201405, 0.86953125], "iscrowd": 0, "area": 43040.962199999994, "rle": {"size": [640, 427], "counts": "Te>2kc04M3L5K4M3L4M3L4M3L4L4M3L4mBoNS7U1kHSOj6R1YH^NTLf0\\;P1jGVOaLLa;S1jG>Q8GjG>R8FiG>T8EhG`0T8DhG`0U8CfGa0W8CdGb0X8BcGc0Z8@aGe0[8_OaGd0]8^O_Gf0^8^O]Gf0a8]OZGh0b8\\OYGh0d8\\OXGg0f8\\OUGi0g8[OTGj0i8XOSGm0i8WORGn0k8UOoFP1n8TOmFP1P9TOkFP1S9SOhFQ1U9SOeFR1Y9l4M4M1N3N1N2O2N1N3N1O2N1O2N1N101O001O001N101O001O001N101O001O1O001N101O001O001N2O001O100000fFZIZ5f6eJ[I[5d6eJ]I[5c6dJ^I\\5b6cJ_I]5a6bJ`I^5_6bJbI^5^6aJcI_5]6_JeI`5\\6_JeIa5Z6_JgIa5Y6^JhIb5X6^JhIb5X6]JiIc5V6]JkIc5U6]JkIc5U6\\JlId5T6[JmIe5R6\\JnId5R6[JoIe5Q6ZJPJf5P6ZJPJe5Q6ZJPJPNYNX7f7hJRJnM]NW7a7kJSJkMaNW7\\7mJTJjMeNV7W7oJnJn4Q5SKSKmMfNT5W6nLPMn2P3QMVMj2j2VM[Me2d2[MbM`2^2_M]ImMQ3d3mKSMe7\\1YIQNW3]3mKTMd7]1YIRNX3Z3nKUM`7_1[ISNW3Z1VLKKVO^7b1ZISNY3?QL\\Oe0P1[O\\O_7e1XITNY3=_M4dNB^7h1WIUNZ3:iNDh4n1UITN\\36kNDf4R2RIUN^33o6g1cEWN`3On6j1bEWNa3Lo6m1_EXNd3Hm6Q2^EXN\\?h1d@XN\\?g1d@[N[?e1e@[NZ?f1e@[N[?e1d@\\N\\?b1f@^NZ?_1h@bNX?]1i@cNV?]1j@dNV?\\1i@eNW?[1i@eNW?[1h@fNW?W1l@kNR?R1RAmNo>P1SAQOj7\\OjMc1]JPOh7_OjM`1_JROe7@kM]1aJTOc7@jM]1dJROb7BeM`1iJoNb7@aMd1oJkNa7A[Mh1TKhNa7@VMl1ZKdN`7@PMQ2`K_Na7_OlLU2dK\\N`7_OkLT2fK^N_7^OjLU2gK\\N`7^OiLV2gK]N`7[OjLX2gK\\N_7[OkLY2gK\\N_=d1aB]N^=d1bB[N^=e1cB[N[=f1fBZNY=f1hBTNVNDR?X2lBPNTNEQ?\\2kBoMX=Q2iBnMW=S2iBmMV=S2kBgMPNOU?[2RCdMm<\\2[21O1O001O001O10O100000O1000O1O10O01O1O1O10O0100O10000O2O0O100O110O000000001O001O1O1O1O1N2O001O1N2O1O1M4M2N2O2O0O2N100N3M2N2N3M4L4L4K5I7FZhU3"}, "label": "a young boy is helping a little girl try to stand on a skateboard"}]} {"id": 101832, "image": "COCO_train2014_000000101832.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy helping a girl ride a skateboard\"."}], "task": "refering", "answer_template": "The \"a boy helping a girl ride a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 51, 52, 53, 66, 67, 68, 69, 78, 79, 80, 81, 82, 83, 84, 92, 93, 94, 95, 96, 107, 108, 109, 110, 121, 122, 123, 124, 125, 126, 136, 137, 138, 140, 141, 150, 151, 152, 153, 154, 166, 167, 168, 169, 181, 182, 183, 184, 197, 198, 199, 211, 212, 213, 214, 215, 226, 227, 228, 229, 230, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291], "bbox": [0.05346604215456674, 0.12357812500000001, 0.6188758782201405, 0.86953125], "iscrowd": 0, "area": 43040.962199999994, "rle": {"size": [640, 427], "counts": "Te>2kc04M3L5K4M3L4M3L4M3L4L4M3L4mBoNS7U1kHSOj6R1YH^NTLf0\\;P1jGVOaLLa;S1jG>Q8GjG>R8FiG>T8EhG`0T8DhG`0U8CfGa0W8CdGb0X8BcGc0Z8@aGe0[8_OaGd0]8^O_Gf0^8^O]Gf0a8]OZGh0b8\\OYGh0d8\\OXGg0f8\\OUGi0g8[OTGj0i8XOSGm0i8WORGn0k8UOoFP1n8TOmFP1P9TOkFP1S9SOhFQ1U9SOeFR1Y9l4M4M1N3N1N2O2N1N3N1O2N1O2N1N101O001O001N101O001O001N101O001O1O001N101O001O001N2O001O100000fFZIZ5f6eJ[I[5d6eJ]I[5c6dJ^I\\5b6cJ_I]5a6bJ`I^5_6bJbI^5^6aJcI_5]6_JeI`5\\6_JeIa5Z6_JgIa5Y6^JhIb5X6^JhIb5X6]JiIc5V6]JkIc5U6]JkIc5U6\\JlId5T6[JmIe5R6\\JnId5R6[JoIe5Q6ZJPJf5P6ZJPJe5Q6ZJPJPNYNX7f7hJRJnM]NW7a7kJSJkMaNW7\\7mJTJjMeNV7W7oJnJn4Q5SKSKmMfNT5W6nLPMn2P3QMVMj2j2VM[Me2d2[MbM`2^2_M]ImMQ3d3mKSMe7\\1YIQNW3]3mKTMd7]1YIRNX3Z3nKUM`7_1[ISNW3Z1VLKKVO^7b1ZISNY3?QL\\Oe0P1[O\\O_7e1XITNY3=_M4dNB^7h1WIUNZ3:iNDh4n1UITN\\36kNDf4R2RIUN^33o6g1cEWN`3On6j1bEWNa3Lo6m1_EXNd3Hm6Q2^EXN\\?h1d@XN\\?g1d@[N[?e1e@[NZ?f1e@[N[?e1d@\\N\\?b1f@^NZ?_1h@bNX?]1i@cNV?]1j@dNV?\\1i@eNW?[1i@eNW?[1h@fNW?W1l@kNR?R1RAmNo>P1SAQOj7\\OjMc1]JPOh7_OjM`1_JROe7@kM]1aJTOc7@jM]1dJROb7BeM`1iJoNb7@aMd1oJkNa7A[Mh1TKhNa7@VMl1ZKdN`7@PMQ2`K_Na7_OlLU2dK\\N`7_OkLT2fK^N_7^OjLU2gK\\N`7^OiLV2gK]N`7[OjLX2gK\\N_7[OkLY2gK\\N_=d1aB]N^=d1bB[N^=e1cB[N[=f1fBZNY=f1hBTNVNDR?X2lBPNTNEQ?\\2kBoMX=Q2iBnMW=S2iBmMV=S2kBgMPNOU?[2RCdMm<\\2[21O1O001O001O10O100000O1000O1O10O01O1O1O10O0100O10000O2O0O100O110O000000001O001O1O1O1O1N2O001O1N2O1O1M4M2N2O2O0O2N100N3M2N2N3M4L4L4K5I7FZhU3"}, "label": "a boy helping a girl ride a skateboard"}]} {"id": 101832, "image": "COCO_train2014_000000101832.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl on a skateboard\"."}], "task": "refering", "answer_template": "The \"a girl on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 96, 97, 98, 99, 109, 110, 111, 112, 113, 114, 124, 125, 126, 127, 128, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 169, 170, 171, 172, 173, 184, 185, 186, 187, 199, 200, 201, 202, 215, 216, 217, 230, 231, 232, 245, 246, 247, 260, 261, 262, 275, 276, 277, 290, 291, 292], "bbox": [0.2706791569086651, 0.24548437500000003, 0.6259250585480093, 0.8752812500000001], "iscrowd": 0, "area": 33943.42715, "rle": {"size": [640, 427], "counts": "UgX27hc08H9G8I8G8H4L4M3L4L4M4K4L4L4M3L4L4L5LO101N10000O101N100O10001N102M3N2M3J6K4L5K5K5K5K5K5K5J6K5KSjS3"}, "label": "a girl on a skateboard"}]} {"id": 101832, "image": "COCO_train2014_000000101832.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl wearing jeans and a paisley shirt\"."}], "task": "refering", "answer_template": "The \"a girl wearing jeans and a paisley shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 96, 97, 98, 99, 109, 110, 111, 112, 113, 114, 124, 125, 126, 127, 128, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 169, 170, 171, 172, 173, 184, 185, 186, 187, 199, 200, 201, 202, 215, 216, 217, 230, 231, 232, 245, 246, 247, 260, 261, 262, 275, 276, 277, 290, 291, 292], "bbox": [0.2706791569086651, 0.24548437500000003, 0.6259250585480093, 0.8752812500000001], "iscrowd": 0, "area": 33943.42715, "rle": {"size": [640, 427], "counts": "UgX27hc08H9G8I8G8H4L4M3L4L4M4K4L4L4M3L4L4L5LO101N10000O101N100O10001N102M3N2M3J6K4L5K5K5K5K5K5K5J6K5KSjS3"}, "label": "a girl wearing jeans and a paisley shirt"}]} {"id": 347596, "image": "COCO_train2014_000000347596.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a surfboard being carried by a man walking right\"."}], "task": "refering", "answer_template": "The \"a surfboard being carried by a man walking right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 148, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 233], "bbox": [0.12501562500000002, 0.43236533957845436, 0.450453125, 0.6844262295081968], "iscrowd": 0, "area": 8630.252150000002, "rle": {"size": [427, 640], "counts": "ncQ12V=5L4L4_CBo;`0oCBn;a0QD@m;b0QD_On;d0PD]On;e0PD]OP in this image.", "objects": [{"patches": [145, 146, 147, 148, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 233], "bbox": [0.12501562500000002, 0.43236533957845436, 0.450453125, 0.6844262295081968], "iscrowd": 0, "area": 8630.252150000002, "rle": {"size": [427, 640], "counts": "ncQ12V=5L4L4_CBo;`0oCBn;a0QD@m;b0QD_On;d0PD]On;e0PD]OP in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 21, 37, 38, 39, 40, 41, 42, 43, 59, 60, 61, 62, 63, 65, 66, 81, 82, 83, 84, 85, 86, 87, 88, 89, 104, 105, 106, 107, 108, 111, 112, 127, 128, 129, 130, 131, 150, 151, 153, 173, 174], "bbox": [0.5293281249999999, 6.25e-05, 0.923203125, 0.4365625], "iscrowd": 0, "area": 24061.309699999998, "rle": {"size": [480, 640], "counts": "oQo47b>8cN0`C:^ in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 21, 37, 38, 39, 40, 41, 42, 43, 59, 60, 61, 62, 63, 65, 66, 81, 82, 83, 84, 85, 86, 87, 88, 89, 104, 105, 106, 107, 108, 111, 112, 127, 128, 129, 130, 131, 150, 151, 153, 173, 174], "bbox": [0.5293281249999999, 6.25e-05, 0.923203125, 0.4365625], "iscrowd": 0, "area": 24061.309699999998, "rle": {"size": [480, 640], "counts": "oQo47b>8cN0`C:^ in this image.", "objects": [{"patches": [191, 192, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 300, 301, 302, 303, 304, 305, 306, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 352, 368, 369, 370, 371, 372], "bbox": [0.01978125, 0.5207916666666667, 0.447265625, 0.9873124999999999], "iscrowd": 0, "area": 30404.10835, "rle": {"size": [480, 640], "counts": "ha62k>3M3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3M3N2M3N2N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1M3N2M3N2M3N2O1O1O100O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1OM3K5K5O2O1O1N2O1O0O2O1N2O1N2N2O0O2O1N2O1N2O1N1O2O1N2O1N2O1N101O1N2O1O1N2O001O1N2O1O1N2O001N2O1i0WO00O1O0010N1O2N2N3M3M3N1O2O1O1O0O2O1O1O001N2O1O001NoNbFcL]9X3kFeLU9V3RGhLn8P3\\GnLc8Q3aGmL_8Q3eGmL[8Q3iGmLW8Q3mGnLQ8Q3QHoLo7o2RHRMn7l2SHUMm7j2SHWMl7h2UHYMk7d2WH^Mh7\\2]HeMc7U2bHlM]7n1hHTNX7g1lHZNT7b1oH_NQ7]1RIeNm6X1UIiNj6W1VIjNj6U1VIlNj6T1UIkNm6U1SIjNm6X1RIgNo6Z1PIdNR7]1nHaNS7`1lH_NU7a1kH^NV7c1iHZNZ7g1eHTNa7l1^HoMg7R2YHiMk7W2UHdMP8\\2Z2O1O100O1O2N1O100O100O100O101N100O100O2O1O1O001N2O1O1O1O001O1O1N2O1O001O1O2N1N3N1O2N1O7I001N2O1O1L3N3L4M3L3N3LodU5"}, "label": "dog with purple collar looking in the mirror"}]} {"id": 568788, "image": "COCO_train2014_000000568788.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog eting\"."}], "task": "refering", "answer_template": "The \"a dog eting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 300, 301, 302, 303, 304, 305, 306, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 352, 368, 369, 370, 371, 372], "bbox": [0.01978125, 0.5207916666666667, 0.447265625, 0.9873124999999999], "iscrowd": 0, "area": 30404.10835, "rle": {"size": [480, 640], "counts": "ha62k>3M3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3M3N2M3N2N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1M3N2M3N2M3N2O1O1O100O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1OM3K5K5O2O1O1N2O1O0O2O1N2O1N2N2O0O2O1N2O1N2O1N1O2O1N2O1N2O1N101O1N2O1O1N2O001O1N2O1O1N2O001N2O1i0WO00O1O0010N1O2N2N3M3M3N1O2O1O1O0O2O1O1O001N2O1O001NoNbFcL]9X3kFeLU9V3RGhLn8P3\\GnLc8Q3aGmL_8Q3eGmL[8Q3iGmLW8Q3mGnLQ8Q3QHoLo7o2RHRMn7l2SHUMm7j2SHWMl7h2UHYMk7d2WH^Mh7\\2]HeMc7U2bHlM]7n1hHTNX7g1lHZNT7b1oH_NQ7]1RIeNm6X1UIiNj6W1VIjNj6U1VIlNj6T1UIkNm6U1SIjNm6X1RIgNo6Z1PIdNR7]1nHaNS7`1lH_NU7a1kH^NV7c1iHZNZ7g1eHTNa7l1^HoMg7R2YHiMk7W2UHdMP8\\2Z2O1O100O1O2N1O100O100O100O101N100O100O2O1O1O001N2O1O1O1O001O1O1N2O1O001O1O2N1N3N1O2N1O7I001N2O1O1L3N3L4M3L3N3LodU5"}, "label": "a dog eting"}]} {"id": 323030, "image": "COCO_train2014_000000323030.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman who is a tennis player\"."}], "task": "refering", "answer_template": "The \"woman who is a tennis player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 124, 125, 126, 127, 147, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 266, 282, 283, 284, 285, 286, 287, 288, 289, 305, 306, 307, 308, 309, 310, 311, 312, 328, 329, 330, 331, 332, 333, 334, 351, 352, 353, 354, 355, 356, 357, 375, 376, 377, 378, 379, 380], "bbox": [0.264796875, 0.2599166666666667, 0.605046875, 1.0], "iscrowd": 0, "area": 47479.13665, "rle": {"size": [480, 640], "counts": "db_21m>6J6J5K6J6J6J6XBTOR=c1I3M3M2M4M3L3N3M3N2N1O2M3N1O2N1O1`EQM`8_3_GaLT8l3kGULR8o3mGQLQ8Q4oGoKn7T4RHlKm7U4SHkKl7W4SHiKm7W4RHjKm7W4SHiKm7W4SHiKl7X4THhKl7X4RHjKm7W4oGmKQ8T4kGoKT8R4hGRLX8n3eGULZ8l3bGXL]8i3`GZL_8g3]G]Lb8e3ZG^Le8c3WGaLh8l40000O1000000O1000000O1000000O1000000O100000000O10000000000O100000000O1dMiIoKW6n3PJnKP6n3XJnKh5n3`JnK`5n3lJjKT5S4XKcKi4[4cK[K]4c4iKYKW4e4nKXKR4f4SLWKm3g4XLUKi3i4\\LTKd3k4_LSKa3l4cLQK]3m4hLPKX3o4]3O1N2O1O1N2N2M3N2N2N2O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O100000000000000O1000000000000001O2N2N2N2N2N2N001O00001O001O001O00001O001]IeK_2[4]MjKb2V4YMPLf2Q4TMTLl2l3PMYLo2h3kL]LU3d3eLbLZ3`3dLaL[3a3bL`L^3b3`L_L_3b3`L_L^2XO[K[4U2]L[2B[KT4V2\\LY2F`KR4S2YL[2FbKT4P2VL^2FbKW4l1UL`2EdKY4i1RLb2FeKc4^1hKl2HcKd4]1eKn2KbKc4\\1cKQ3P6lLQJR3S6kLnIT3U6TL`Jk3d5oK`Jo3d5jKaJT4V8G:F9A`0D in this image.", "objects": [{"patches": [102, 103, 124, 125, 126, 127, 147, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 215, 216, 217, 218, 219, 237, 238, 239, 240, 241, 242, 243, 259, 260, 261, 262, 263, 264, 265, 266, 282, 283, 284, 285, 286, 287, 288, 289, 305, 306, 307, 308, 309, 310, 311, 312, 328, 329, 330, 331, 332, 333, 334, 351, 352, 353, 354, 355, 356, 357, 375, 376, 377, 378, 379, 380], "bbox": [0.264796875, 0.2599166666666667, 0.605046875, 1.0], "iscrowd": 0, "area": 47479.13665, "rle": {"size": [480, 640], "counts": "db_21m>6J6J5K6J6J6J6XBTOR=c1I3M3M2M4M3L3N3M3N2N1O2M3N1O2N1O1`EQM`8_3_GaLT8l3kGULR8o3mGQLQ8Q4oGoKn7T4RHlKm7U4SHkKl7W4SHiKm7W4RHjKm7W4SHiKm7W4SHiKl7X4THhKl7X4RHjKm7W4oGmKQ8T4kGoKT8R4hGRLX8n3eGULZ8l3bGXL]8i3`GZL_8g3]G]Lb8e3ZG^Le8c3WGaLh8l40000O1000000O1000000O1000000O1000000O100000000O10000000000O100000000O1dMiIoKW6n3PJnKP6n3XJnKh5n3`JnK`5n3lJjKT5S4XKcKi4[4cK[K]4c4iKYKW4e4nKXKR4f4SLWKm3g4XLUKi3i4\\LTKd3k4_LSKa3l4cLQK]3m4hLPKX3o4]3O1N2O1O1N2N2M3N2N2N2O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O100000000000000O1000000000000001O2N2N2N2N2N2N001O00001O001O001O00001O001]IeK_2[4]MjKb2V4YMPLf2Q4TMTLl2l3PMYLo2h3kL]LU3d3eLbLZ3`3dLaL[3a3bL`L^3b3`L_L_3b3`L_L^2XO[K[4U2]L[2B[KT4V2\\LY2F`KR4S2YL[2FbKT4P2VL^2FbKW4l1UL`2EdKY4i1RLb2FeKc4^1hKl2HcKd4]1eKn2KbKc4\\1cKQ3P6lLQJR3S6kLnIT3U6TL`Jk3d5oK`Jo3d5jKaJT4V8G:F9A`0D in this image.", "objects": [{"patches": [60, 61, 82, 83, 84, 85, 86, 105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.5525156250000001, 0.169375, 1.0, 0.9985833333333333], "iscrowd": 0, "area": 77465.74369999998, "rle": {"size": [480, 640], "counts": "SSV57c>9H8H7I8G9J6I6K6I7K5K4L5K5K5K4K5L4L3M4L3M4M3N1N3M2mH[Lh2h3UMhL\\2Z3`MXMo1j2oM^Ni0e1SOaNi0`1UOdNg0_1VOfNe0]1WOlNa0V1]OUO9m0CVO;l0CWO:l0BWO=j0AYOPNDP2 in this image.", "objects": [{"patches": [60, 61, 82, 83, 84, 85, 86, 105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.5525156250000001, 0.169375, 1.0, 0.9985833333333333], "iscrowd": 0, "area": 77465.74369999998, "rle": {"size": [480, 640], "counts": "SSV57c>9H8H7I8G9J6I6K6I7K5K4L5K5K5K4K5L4L3M4L3M4M3N1N3M2mH[Lh2h3UMhL\\2Z3`MXMo1j2oM^Ni0e1SOaNi0`1UOdNg0_1VOfNe0]1WOlNa0V1]OUO9m0CVO;l0CWO:l0BWO=j0AYOPNDP2 in this image.", "objects": [{"patches": [199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 312, 313, 314, 315, 316, 317, 318, 319, 320, 335, 336, 337, 338, 339, 340, 341, 342, 343, 358, 359, 360, 361, 362, 363, 364, 365, 366, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.5709531250000001, 0.503375, 0.952703125, 1.0], "iscrowd": 0, "area": 49209.671449999994, "rle": {"size": [480, 640], "counts": "f_[56\\=g1D in this image.", "objects": [{"patches": [199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 297, 312, 313, 314, 315, 316, 317, 318, 319, 320, 335, 336, 337, 338, 339, 340, 341, 342, 343, 358, 359, 360, 361, 362, 363, 364, 365, 366, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.5709531250000001, 0.503375, 0.952703125, 1.0], "iscrowd": 0, "area": 49209.671449999994, "rle": {"size": [480, 640], "counts": "f_[56\\=g1D in this image.", "objects": [{"patches": [240, 241, 242, 260, 263, 264, 265, 282, 283, 285, 286, 287, 306, 308, 309, 310, 331, 332], "bbox": [0.27809375, 0.6756908665105386, 0.5586718749999999, 0.9996252927400467], "iscrowd": 0, "area": 6581.910000000005, "rle": {"size": [427, 640], "counts": "^oZ22X=2O0O2N101N2N1O2O1N1O2O1N1O2O0QD\\OV;f0hD]OU;e0hD_OV;a0iD@V;b0hD_OW;c0iD\\OW;d0iD[OW;g0hDYOX;g0hDXOX;j0gDUOZ;l0eDSO[;n0eDQO\\;P1dDnN\\;T1cDkN^;U1;2O1N101N10M2M4L3M4L4J5L5K4K6KS]=1PcB3M4L3M4L3M:F2N2N3N1N2N2N2N3LfNYDU1c;nNaDo0\\;ROiDk0T;WOoDg0n:ZORE0C:X;HUEND9U;JXEME7m:2^EGF7b:;hE^OH5\\:`0mE[OH4W:e0QFVOK3Q:i0UFTOK2l9n0YFPOM0g9R1]FnNNNa9X1aFjNOM]9[1eFhN0KY9_1gFfN1JU9b1jFeN3Ho8f1nFbN4Gl8h1QGaN5Eg8m1TG^Ne9c1\\F]Nd9c1\\F]Nc9c1^F]Na9d1_F\\N`9e1`FZN`9f1aFZN_9f1aFZN^9f1cFZN\\9g1cFZN\\9f1eFZN[9f1eFZNZ9f1gFZNY9f1gFZNX9f1iFZNW9f1iFZNW9e1jF[NU9f1kFZNU9e1lF[NT9e1lFZNT9f1mFZNS9f1mFZNS9e1mF\\NR9e1nF[NR9e1nF[NR9d1oF\\NQ9d1oF\\NQ9c1PG]NP9c1PG]No8c1RG]Nn8c1RG]Nn8b1SG^Nm8b1SG^Nm8b1SG]Nn8b1RG_Nn8a1RG_Nn8`1SG`Nl8a1TG_Nl8`1UG`Nk8`1UG`Nk8_1VGaNj8Y1\\GbNi8W1^GdNg8U1k1M3O1O100O100O100O1O100O100O1O100O1M3L4K5L4K5LQZe3"}, "label": "chair under the man in blue shirt"}]} {"id": 224734, "image": "COCO_train2014_000000224734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair is holding the african american man as he rests\"."}], "task": "refering", "answer_template": "The \"the chair is holding the african american man as he rests\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [240, 241, 242, 260, 263, 264, 265, 282, 283, 285, 286, 287, 306, 308, 309, 310, 331, 332], "bbox": [0.27809375, 0.6756908665105386, 0.5586718749999999, 0.9996252927400467], "iscrowd": 0, "area": 6581.910000000005, "rle": {"size": [427, 640], "counts": "^oZ22X=2O0O2N101N2N1O2O1N1O2O1N1O2O0QD\\OV;f0hD]OU;e0hD_OV;a0iD@V;b0hD_OW;c0iD\\OW;d0iD[OW;g0hDYOX;g0hDXOX;j0gDUOZ;l0eDSO[;n0eDQO\\;P1dDnN\\;T1cDkN^;U1;2O1N101N10M2M4L3M4L4J5L5K4K6KS]=1PcB3M4L3M4L3M:F2N2N3N1N2N2N2N3LfNYDU1c;nNaDo0\\;ROiDk0T;WOoDg0n:ZORE0C:X;HUEND9U;JXEME7m:2^EGF7b:;hE^OH5\\:`0mE[OH4W:e0QFVOK3Q:i0UFTOK2l9n0YFPOM0g9R1]FnNNNa9X1aFjNOM]9[1eFhN0KY9_1gFfN1JU9b1jFeN3Ho8f1nFbN4Gl8h1QGaN5Eg8m1TG^Ne9c1\\F]Nd9c1\\F]Nc9c1^F]Na9d1_F\\N`9e1`FZN`9f1aFZN_9f1aFZN^9f1cFZN\\9g1cFZN\\9f1eFZN[9f1eFZNZ9f1gFZNY9f1gFZNX9f1iFZNW9f1iFZNW9e1jF[NU9f1kFZNU9e1lF[NT9e1lFZNT9f1mFZNS9f1mFZNS9e1mF\\NR9e1nF[NR9e1nF[NR9d1oF\\NQ9d1oF\\NQ9c1PG]NP9c1PG]No8c1RG]Nn8c1RG]Nn8b1SG^Nm8b1SG^Nm8b1SG]Nn8b1RG_Nn8a1RG_Nn8`1SG`Nl8a1TG_Nl8`1UG`Nk8`1UG`Nk8_1VGaNj8Y1\\GbNi8W1^GdNg8U1k1M3O1O100O100O100O1O100O100O1O100O1M3L4K5L4K5LQZe3"}, "label": "the chair is holding the african american man as he rests"}]} {"id": 224734, "image": "COCO_train2014_000000224734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the gentleman sitting in the chair wearing a light blue short sleeved oxford shirt and black slacks\"."}], "task": "refering", "answer_template": "The \"the gentleman sitting in the chair wearing a light blue short sleeved oxford shirt and black slacks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 102, 103, 104, 124, 125, 126, 127, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 215, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 306, 307, 308, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.165671875, 0.22861826697892274, 0.6045, 1.0], "iscrowd": 0, "area": 43422.744399999996, "rle": {"size": [427, 640], "counts": "\\b\\1]1`:^1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1oNYL^Hg3a7dLTH]3k7POkGnNT8R1THgNl7Y1ZHaNe7_1cHZN\\7f1lHSNS7m1UIlMj6U2]IdMb6\\2fI]MZ6b2nIWMQ6j2UJPMj5P3X2000010O0001O001O010O001O001O01O01O010O0100O010O010O010O010O0010fGjL_6U3^IVMZ6k2bI_MX6`2eIjMT6W2hITNQ6k1mI\\Nn5e1nI_NR6`1kIdNT6]1hIgNX6X1eIlNZ6U1bIoN^6P1_ITOa6k0\\IYOc6h0YI\\Og6c0VIAi6`0SIDm6;PIIo6Z300O010O10000O1O1N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N1O1O1O001O1O00100O101NF:C?O1O1N101N2O0O2O1O0O2O1N101O00100O010O1N1N3LjNTJUKi5k4`JnJ]5T5kJfJR5Z5VKaJf4a5bKXJ\\4h5mKRJo3P6XLkIe3W6d1400000000000O100000000000000O1000000000O100O010O100O1O100O10O0100O1O100O101N2O2M2N3N2M2O2lJQIg3Q7ULSIi3P7SLRIm3R7mKQIQ4S7jKPIT4T7fKPIY4T7aKnH^4V7]KmHa4P8N2O0O2N2O0O2N2N2O0O2N2N2O0O2N2O0O2N2N2O0O2N2O1N1O2N2O1N4M4K6K4K6]Ob0J7K5K4L5J5L5K4K6K4L5K4K6K4LiWY3"}, "label": "the gentleman sitting in the chair wearing a light blue short sleeved oxford shirt and black slacks"}]} {"id": 224734, "image": "COCO_train2014_000000224734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a blue shirt and sitting on a chair\"."}], "task": "refering", "answer_template": "The \"a man wearing a blue shirt and sitting on a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 102, 103, 104, 124, 125, 126, 127, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 215, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 306, 307, 308, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.165671875, 0.22861826697892274, 0.6045, 1.0], "iscrowd": 0, "area": 43422.744399999996, "rle": {"size": [427, 640], "counts": "\\b\\1]1`:^1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O1oNYL^Hg3a7dLTH]3k7POkGnNT8R1THgNl7Y1ZHaNe7_1cHZN\\7f1lHSNS7m1UIlMj6U2]IdMb6\\2fI]MZ6b2nIWMQ6j2UJPMj5P3X2000010O0001O001O010O001O001O01O01O010O0100O010O010O010O010O0010fGjL_6U3^IVMZ6k2bI_MX6`2eIjMT6W2hITNQ6k1mI\\Nn5e1nI_NR6`1kIdNT6]1hIgNX6X1eIlNZ6U1bIoN^6P1_ITOa6k0\\IYOc6h0YI\\Og6c0VIAi6`0SIDm6;PIIo6Z300O010O10000O1O1N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N1O1O1O001O1O00100O101NF:C?O1O1N101N2O0O2O1O0O2O1N101O00100O010O1N1N3LjNTJUKi5k4`JnJ]5T5kJfJR5Z5VKaJf4a5bKXJ\\4h5mKRJo3P6XLkIe3W6d1400000000000O100000000000000O1000000000O100O010O100O1O100O10O0100O1O100O101N2O2M2N3N2M2O2lJQIg3Q7ULSIi3P7SLRIm3R7mKQIQ4S7jKPIT4T7fKPIY4T7aKnH^4V7]KmHa4P8N2O0O2N2O0O2N2N2O0O2N2N2O0O2N2O0O2N2N2O0O2N2O1N1O2N2O1N4M4K6K4K6]Ob0J7K5K4L5J5L5K4K6K4L5K4K6K4LiWY3"}, "label": "a man wearing a blue shirt and sitting on a chair"}]} {"id": 224734, "image": "COCO_train2014_000000224734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sitting person in background\"."}], "task": "refering", "answer_template": "The \"sitting person in background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 117, 118, 139, 140, 141, 142, 143, 163, 164, 165, 166, 188, 189, 211, 212, 235], "bbox": [0.067015625, 0.26362997658079623, 0.254296875, 0.680655737704918], "iscrowd": 0, "area": 6964.4460500000005, "rle": {"size": [427, 640], "counts": "[Qb01Y=2O2N2N2O1dDIa98\\FKd96XFMg93WF0h90UF4j9LTF8j9HTF=i9DTFa0j9^OSFg0k9SOUEOn0R1l9lNWE3k0T1m9hNYE6f0T1U:oNfEU1Y:lNcEW1]:kN^EY1`:iN\\E[1b:hNYE[1g:f0N1O2OO00000O2O000000001O0O101O001O1O001N10_OfMRFY2m9jMQFV2n9mMQFR2m9RNQFn1n9UNUFf1j9n0N24LN2O1N3NlLQFn2n9oLUFR3T:N3WMfE[2Y:dMkEZ2X:`MmE_2n:mMnDo0h;O0101N3M2O2M2O1N3M2O1N3N1N3M2O1N3NO01O010O01PNQEc1n:^NUE^1k:bNYE[1g:dN]EX1c:iN`EU1_:jNeET1Y:lNkEQ1T:POoEn0o9ROSFn0l9QOWFm0h9TOYFl0e9TO]Fl0a9TObFj0^9VOcFj0[9VOgFi0X9WOkFh0S9YOnFg0P9YORGf0n8YOVGe0h8\\O[Gb0c8^OaG>_8BfG9[8GgG6Y8JhG5X8KjG3V8NjG1W8NkGOV81kGNU83lGKT85oGHR87RHEn7;a2N2M4L3NlfV6"}, "label": "sitting person in background"}]} {"id": 224734, "image": "COCO_train2014_000000224734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in back sitting\"."}], "task": "refering", "answer_template": "The \"person in back sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 117, 118, 139, 140, 141, 142, 143, 163, 164, 165, 166, 188, 189, 211, 212, 235], "bbox": [0.067015625, 0.26362997658079623, 0.254296875, 0.680655737704918], "iscrowd": 0, "area": 6964.4460500000005, "rle": {"size": [427, 640], "counts": "[Qb01Y=2O2N2N2O1dDIa98\\FKd96XFMg93WF0h90UF4j9LTF8j9HTF=i9DTFa0j9^OSFg0k9SOUEOn0R1l9lNWE3k0T1m9hNYE6f0T1U:oNfEU1Y:lNcEW1]:kN^EY1`:iN\\E[1b:hNYE[1g:f0N1O2OO00000O2O000000001O0O101O001O1O001N10_OfMRFY2m9jMQFV2n9mMQFR2m9RNQFn1n9UNUFf1j9n0N24LN2O1N3NlLQFn2n9oLUFR3T:N3WMfE[2Y:dMkEZ2X:`MmE_2n:mMnDo0h;O0101N3M2O2M2O1N3M2O1N3N1N3M2O1N3NO01O010O01PNQEc1n:^NUE^1k:bNYE[1g:dN]EX1c:iN`EU1_:jNeET1Y:lNkEQ1T:POoEn0o9ROSFn0l9QOWFm0h9TOYFl0e9TO]Fl0a9TObFj0^9VOcFj0[9VOgFi0X9WOkFh0S9YOnFg0P9YORGf0n8YOVGe0h8\\O[Gb0c8^OaG>_8BfG9[8GgG6Y8JhG5X8KjG3V8NjG1W8NkGOV81kGNU83lGKT85oGHR87RHEn7;a2N2M4L3NlfV6"}, "label": "person in back sitting"}]} {"id": 372198, "image": "COCO_train2014_000000372198.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small girl starring at something , along with her elder sister\"."}], "task": "refering", "answer_template": "The \"a small girl starring at something , along with her elder sister\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 79, 80, 81, 97, 98, 99, 114, 115, 116, 117, 132, 133, 134, 150, 151], "bbox": [0.33462000000000003, 0.3010810810810811, 0.51816, 0.7090090090090091], "iscrowd": 0, "area": 6602.707299999999, "rle": {"size": [333, 500], "counts": "Rjf13Z:2N2M3N2M3N2N3L3O1N2N2N2N2N2N2N2N2N2L4L2N3M2N3M2N3M2N3M2M4M2N3bNQNZJQ2d5_NlId1P6_NmId1R6^NlId1S6]NlIc1T6^NkI`1V6bNhI]1Z6cNfI\\1[6eNcIZ1^6hNaIW1`6Y12N101N2O0O2O1N1O2O0O2O1N2N2O1N2O2M2N2O1N2O1N2N20000O1000001O2N2N1N3N1O2N101O2M3K6J6J5Kb0^O9G6J7I6JgV^2"}, "label": "a small girl starring at something , along with her elder sister"}]} {"id": 372198, "image": "COCO_train2014_000000372198.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little girl sitting with toy in her hand\"."}], "task": "refering", "answer_template": "The \"little girl sitting with toy in her hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 79, 80, 81, 97, 98, 99, 114, 115, 116, 117, 132, 133, 134, 150, 151], "bbox": [0.33462000000000003, 0.3010810810810811, 0.51816, 0.7090090090090091], "iscrowd": 0, "area": 6602.707299999999, "rle": {"size": [333, 500], "counts": "Rjf13Z:2N2M3N2M3N2N3L3O1N2N2N2N2N2N2N2N2N2L4L2N3M2N3M2N3M2N3M2M4M2N3bNQNZJQ2d5_NlId1P6_NmId1R6^NlId1S6]NlIc1T6^NkI`1V6bNhI]1Z6cNfI\\1[6eNcIZ1^6hNaIW1`6Y12N101N2O0O2O1N1O2O0O2O1N2N2O1N2O2M2N2O1N2O1N2N20000O1000001O2N2N1N3N1O2N101O2M3K6J6J5Kb0^O9G6J7I6JgV^2"}, "label": "little girl sitting with toy in her hand"}]} {"id": 372198, "image": "COCO_train2014_000000372198.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in pajamas in a suitcase\"."}], "task": "refering", "answer_template": "The \"a boy in pajamas in a suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 82, 83, 99, 100, 101, 116, 117, 118, 119, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155], "bbox": [0.41128, 0.30810810810810807, 0.66096, 0.7451051051051051], "iscrowd": 0, "area": 9762.992050000003, "rle": {"size": [333, 500], "counts": "QWS22Z:1N3M2O1N2N3N1N3M2O2M2N2N3M2N3M2O2M2N3M3M2N3N1O2M2O1O1O1O001000O100000O010000000001N2ON2N2N2O0O2N2N2O0O2N1O2M2O2O1O001O001O001N2O001N1F`MgHb2V7^6@fI9^6EfI2b6L`I0c6NaINb60aILb61bIKb61aILb61bIJc62eIF^68jI_OZ6=n1L3H^Ug1"}, "label": "a boy in pajamas in a suitcase"}]} {"id": 372198, "image": "COCO_train2014_000000372198.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two young girls sitting in a black suitcase on the floor\"."}], "task": "refering", "answer_template": "The \"two young girls sitting in a black suitcase on the floor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 82, 83, 99, 100, 101, 116, 117, 118, 119, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155], "bbox": [0.41128, 0.30810810810810807, 0.66096, 0.7451051051051051], "iscrowd": 0, "area": 9762.992050000003, "rle": {"size": [333, 500], "counts": "QWS22Z:1N3M2O1N2N3N1N3M2O2M2N2N3M2N3M2O2M2N3M3M2N3N1O2M2O1O1O1O001000O100000O010000000001N2ON2N2N2O0O2N2N2O0O2N1O2M2O2O1O001O001O001N2O001N1F`MgHb2V7^6@fI9^6EfI2b6L`I0c6NaINb60aILb61bIKb61aILb61bIJc62eIF^68jI_OZ6=n1L3H^Ug1"}, "label": "two young girls sitting in a black suitcase on the floor"}]} {"id": 142825, "image": "COCO_train2014_000000142825.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small child wearing blue and black socks\"."}], "task": "refering", "answer_template": "The \"a small child wearing blue and black socks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 113, 114, 115, 128, 129, 130, 131, 143, 144, 145, 146, 158, 159, 160, 161, 173, 174, 175, 189, 203, 204, 219], "bbox": [0.5665734265734266, 0.289953125, 0.7585314685314685, 0.64778125], "iscrowd": 0, "area": 8548.226099999998, "rle": {"size": [640, 429], "counts": "VTh47ec0?A?Ac0^O?C4K6SO`MP@e2i?`MT@d2f?bMW@`2d?eMZ@k0F2X1iNP=?QBf02Hk0YOi<>XBb0`0\\O=Gh<=ZB`0c0[O9Jh<=[B?e0XO60g<;\\B>h0UO54e<<\\B;l0SO28d<<\\B:o0PO1=a<;^B8R1nNNa0` in this image.", "objects": [{"patches": [99, 100, 113, 114, 115, 128, 129, 130, 131, 143, 144, 145, 146, 158, 159, 160, 161, 173, 174, 175, 189, 203, 204, 219], "bbox": [0.5665734265734266, 0.289953125, 0.7585314685314685, 0.64778125], "iscrowd": 0, "area": 8548.226099999998, "rle": {"size": [640, 429], "counts": "VTh47ec0?A?Ac0^O?C4K6SO`MP@e2i?`MT@d2f?bMW@`2d?eMZ@k0F2X1iNP=?QBf02Hk0YOi<>XBb0`0\\O=Gh<=ZB`0c0[O9Jh<=[B?e0XO60g<;\\B>h0UO54e<<\\B;l0SO28d<<\\B:o0PO1=a<;^B8R1nNNa0` in this image.", "objects": [{"patches": [47, 48, 49, 50, 51, 52, 62, 63, 64, 65, 66, 67, 68, 76, 77, 78, 79, 80, 81, 82, 83, 84, 91, 92, 93, 94, 95, 96, 97, 98, 99, 106, 107, 108, 109, 110, 111, 112, 113, 114, 121, 122, 123, 124, 125, 126, 127, 128, 136, 137, 138, 139, 140, 141, 142, 143, 151, 152, 153, 154, 155, 156, 157, 158, 166, 167, 168, 169, 170, 173, 181, 182, 184, 185], "bbox": [0.09202797202797203, 0.14825, 0.644988344988345, 0.5574375], "iscrowd": 0, "area": 40824.4653, "rle": {"size": [640, 429], "counts": "eWi04gc04L4L5L3L4L4L4L4J6G9G9G9Z@SOQi2UBaLl=Y3a1K4L5J6K5K5J6K5K4K6L4L4L4L4L4L4L3M4L4L4L4M3L4L3M4L4L4L4Lnjn2"}, "label": "gray elephant standing on dirt and leaves"}]} {"id": 142825, "image": "COCO_train2014_000000142825.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a eliphant in the place\"."}], "task": "refering", "answer_template": "The \"a eliphant in the place\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 49, 50, 51, 52, 62, 63, 64, 65, 66, 67, 68, 76, 77, 78, 79, 80, 81, 82, 83, 84, 91, 92, 93, 94, 95, 96, 97, 98, 99, 106, 107, 108, 109, 110, 111, 112, 113, 114, 121, 122, 123, 124, 125, 126, 127, 128, 136, 137, 138, 139, 140, 141, 142, 143, 151, 152, 153, 154, 155, 156, 157, 158, 166, 167, 168, 169, 170, 173, 181, 182, 184, 185], "bbox": [0.09202797202797203, 0.14825, 0.644988344988345, 0.5574375], "iscrowd": 0, "area": 40824.4653, "rle": {"size": [640, 429], "counts": "eWi04gc04L4L5L3L4L4L4L4J6G9G9G9Z@SOQi2UBaLl=Y3a1K4L5J6K5K5J6K5K4K6L4L4L4L4L4L4L3M4L4L4L4M3L4L3M4L4L4L4Lnjn2"}, "label": "a eliphant in the place"}]} {"id": 142825, "image": "COCO_train2014_000000142825.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in blue holding child\"."}], "task": "refering", "answer_template": "The \"man in blue holding child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 131, 132, 133, 146, 147, 148, 158, 159, 160, 161, 162, 163, 174, 175, 176, 177, 178, 189, 190, 191, 192, 193, 204, 205, 206, 207, 218, 219, 220, 221, 222, 234, 235, 236, 237, 249, 250, 251, 264, 265, 266, 279, 280, 281, 294, 295, 296, 309, 310, 311, 325, 326, 340, 341], "bbox": [0.574032634032634, 0.302234375, 0.9044522144522145, 0.98903125], "iscrowd": 0, "area": 29964.507850000005, "rle": {"size": [640, 429], "counts": "kTj42jc04L5K4K6K4lLYOnBi0P=ZOoBe0o<@oB`0okBBU=?jBAV=`0iB_OY=a0gB^O?@h9S1hE\\OO6V:?jE[O^Oj0f:KlEZOmN`1T;WOnEYO\\NU2c;cNQFVOkMl2QjN_AP1b>nNfAk0[>ROmAf0V>XORBBSB2T>KTBJR>4VB@P>>WBWOo=f0ZBnNm=o0h2H7J7I7H8IV]i0"}, "label": "man in blue holding child"}]} {"id": 142825, "image": "COCO_train2014_000000142825.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a little boy , both with dark hair , the boy is wearing blue striped socks and the man is wearing jeans that go just below his knees\"."}], "task": "refering", "answer_template": "The \"a man holding a little boy , both with dark hair , the boy is wearing blue striped socks and the man is wearing jeans that go just below his knees\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 117, 118, 131, 132, 133, 146, 147, 148, 158, 159, 160, 161, 162, 163, 174, 175, 176, 177, 178, 189, 190, 191, 192, 193, 204, 205, 206, 207, 218, 219, 220, 221, 222, 234, 235, 236, 237, 249, 250, 251, 264, 265, 266, 279, 280, 281, 294, 295, 296, 309, 310, 311, 325, 326, 340, 341], "bbox": [0.574032634032634, 0.302234375, 0.9044522144522145, 0.98903125], "iscrowd": 0, "area": 29964.507850000005, "rle": {"size": [640, 429], "counts": "kTj42jc04L5K4K6K4lLYOnBi0P=ZOoBe0o<@oB`0okBBU=?jBAV=`0iB_OY=a0gB^O?@h9S1hE\\OO6V:?jE[O^Oj0f:KlEZOmN`1T;WOnEYO\\NU2c;cNQFVOkMl2QjN_AP1b>nNfAk0[>ROmAf0V>XORBBSB2T>KTBJR>4VB@P>>WBWOo=f0ZBnNm=o0h2H7J7I7H8IV]i0"}, "label": "a man holding a little boy , both with dark hair , the boy is wearing blue striped socks and the man is wearing jeans that go just below his knees"}]} {"id": 183788, "image": "COCO_train2014_000000183788.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bottom end of a book with a leather book cover\"."}], "task": "refering", "answer_template": "The \"the bottom end of a book with a leather book cover\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 29, 30, 31, 32, 33, 34, 35, 51, 52, 53, 54, 55, 56, 57, 58, 77, 78, 79, 80, 81], "bbox": [0.252046875, 0.0, 0.5583125, 0.2004375], "iscrowd": 0, "area": 16557.496199999998, "rle": {"size": [480, 640], "counts": "[a[21h>;E7I7J6I8H7I7J6I7I7J600O10001O000000000000000000001O000000000000000000001O0000000000000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O00000000000000000000001O00000000000000O1O2N1O1N2I7I7J6H8C=D in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 12, 29, 30, 31, 32, 33, 34, 35, 51, 52, 53, 54, 55, 56, 57, 58, 77, 78, 79, 80, 81], "bbox": [0.252046875, 0.0, 0.5583125, 0.2004375], "iscrowd": 0, "area": 16557.496199999998, "rle": {"size": [480, 640], "counts": "[a[21h>;E7I7J6I8H7I7J6I7I7J600O10001O000000000000000000001O000000000000000000001O0000000000000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O00000000000000000000001O00000000000000O1O2N1O1N2I7I7J6H8C=D in this image.", "objects": [{"patches": [36, 37, 38, 39, 59, 60, 61, 62, 63, 64, 81, 82, 83, 84, 85, 86, 87, 104, 105, 106, 107, 108, 109, 110, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 221, 222, 223], "bbox": [0.483703125, 0.056187499999999994, 0.812359375, 0.5797708333333332], "iscrowd": 0, "area": 31352.103199999998, "rle": {"size": [480, 640], "counts": "d`a43k>3L3M4M3L3N3L3jAXOQ>l0O3L4M2M4L3N3L3N2M4L3N3L3N3L3M4M2M3N3L3M4M2M4M2M3M4M2M4M2M4L3N2M4M2M4L3N3L3N3L3M3N3L3N3L3M4M2M3N3L3M4M2M4M2M4L3N2M4M2M4L3N3L3N2O20O01O010O00010O0010O01O010O00010O0010O01O01O01O010O0010O00010O010O01O010O00010O0010O01O01O01O010O0010O00010O01O010O0010O00010O01O010O01M2M3M4L3M3N3L3M3M4L3M3M4L3M3M4M2M3M4L3M3M4L3M3M4L3M3N3L3M3M4L3M4L3M3M4L3N2M4L3M3M4L3M3M4L3M3N3L3M3M4L3M3M4L3M3M4L3N2M4L3M3M4L3M3M4L3M4M2M3MiUh1"}, "label": "a dark blue passport - sized book"}]} {"id": 183788, "image": "COCO_train2014_000000183788.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue covered book near diary in a table\"."}], "task": "refering", "answer_template": "The \"a blue covered book near diary in a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 59, 60, 61, 62, 63, 64, 81, 82, 83, 84, 85, 86, 87, 104, 105, 106, 107, 108, 109, 110, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 221, 222, 223], "bbox": [0.483703125, 0.056187499999999994, 0.812359375, 0.5797708333333332], "iscrowd": 0, "area": 31352.103199999998, "rle": {"size": [480, 640], "counts": "d`a43k>3L3M4M3L3N3L3jAXOQ>l0O3L4M2M4L3N3L3N2M4L3N3L3N3L3M4M2M3N3L3M4M2M4M2M3M4M2M4M2M4L3N2M4M2M4L3N3L3N3L3M3N3L3N3L3M4M2M3N3L3M4M2M4M2M4L3N2M4M2M4L3N3L3N2O20O01O010O00010O0010O01O010O00010O0010O01O01O01O010O0010O00010O010O01O010O00010O0010O01O01O01O010O0010O00010O01O010O0010O00010O01O010O01M2M3M4L3M3N3L3M3M4L3M3M4L3M3M4M2M3M4L3M3M4L3M3M4L3M3N3L3M3M4L3M4L3M3M4L3N2M4L3M3M4L3M3M4L3M3N3L3M3M4L3M3M4L3M3M4L3N2M4L3M3M4L3M3M4L3M4M2M3MiUh1"}, "label": "a blue covered book near diary in a table"}]} {"id": 183788, "image": "COCO_train2014_000000183788.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"drawings of saints in a sketch book\"."}], "task": "refering", "answer_template": "The \"drawings of saints in a sketch book\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209], "bbox": [0.0, 0.0, 0.277015625, 0.5423958333333334], "iscrowd": 0, "area": 42180.37595000001, "rle": {"size": [480, 640], "counts": ";l5i8X2SN0000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O100000000O1000000O1000B?A>A`0A>B;E;D in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209], "bbox": [0.0, 0.0, 0.277015625, 0.5423958333333334], "iscrowd": 0, "area": 42180.37595000001, "rle": {"size": [480, 640], "counts": ";l5i8X2SN0000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O100000000O1000000O1000B?A>A`0A>B;E;D in this image.", "objects": [{"patches": [133, 134, 154, 155, 173, 174, 175, 176, 193, 194, 195, 196, 197, 214, 215, 216, 217, 236, 237, 238, 257, 258, 259, 278, 279, 280, 299, 300], "bbox": [0.23138333333333336, 0.41796875, 0.4083333333333333, 0.94140625], "iscrowd": 0, "area": 11925.695799999998, "rle": {"size": [448, 600], "counts": "\\Rm1o0m<;F9F;H8I6I6K5J5K6K5J5K7J6hFXLk7o3nGSLQ8S4hGoKV8X4bGjK]8]4\\GeKc8a4VG`Ki8m4N2O001O0O2O1O001O001O1O001ON2O1N2nN`GcLa8\\3fG\\L\\8b3kGXLV8f3RHSLo7l3WHmKk7Q4S1N2N1O2K5K4L4K5L4L4K5L3L5L4001O0VFcMR8^2dGmMZ8S2]GWNb8j1TG`Nj8b1lFhNS9X1dFRO[9j2O1O1O1O0O01O100O1]Oc0M3N1O2N2N2N2N2N00001O000O2N1L8H8G9H8H8H7K6I6M3M4K4M4L4L4L4J6K5J6JlSk4"}, "label": "the horse on the left closest to the camera"}]} {"id": 126447, "image": "COCO_train2014_000000126447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse with multiple brands\"."}], "task": "refering", "answer_template": "The \"a horse with multiple brands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 154, 155, 173, 174, 175, 176, 193, 194, 195, 196, 197, 214, 215, 216, 217, 236, 237, 238, 257, 258, 259, 278, 279, 280, 299, 300], "bbox": [0.23138333333333336, 0.41796875, 0.4083333333333333, 0.94140625], "iscrowd": 0, "area": 11925.695799999998, "rle": {"size": [448, 600], "counts": "\\Rm1o0m<;F9F;H8I6I6K5J5K6K5J5K7J6hFXLk7o3nGSLQ8S4hGoKV8X4bGjK]8]4\\GeKc8a4VG`Ki8m4N2O001O0O2O1O001O001O1O001ON2O1N2nN`GcLa8\\3fG\\L\\8b3kGXLV8f3RHSLo7l3WHmKk7Q4S1N2N1O2K5K4L4K5L4L4K5L3L5L4001O0VFcMR8^2dGmMZ8S2]GWNb8j1TG`Nj8b1lFhNS9X1dFRO[9j2O1O1O1O0O01O100O1]Oc0M3N1O2N2N2N2N2N00001O000O2N1L8H8G9H8H8H7K6I6M3M4K4M4L4L4L4J6K5J6JlSk4"}, "label": "a horse with multiple brands"}]} {"id": 126447, "image": "COCO_train2014_000000126447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"horse getting ready to get behind the truck to travel long distance\"."}], "task": "refering", "answer_template": "The \"horse getting ready to get behind the truck to travel long distance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 116, 117, 118, 137, 138, 139, 140, 159, 160, 161, 162, 163, 180, 181, 182, 183, 184, 185, 201, 202, 203, 204, 205, 206, 207, 222, 223, 243, 244, 268, 269, 289, 290], "bbox": [0.5436333333333333, 0.28089285714285717, 0.8741833333333333, 0.8471874999999999], "iscrowd": 0, "area": 14818.668600000003, "rle": {"size": [448, 600], "counts": "eh^4:f=1N3N1O1N2O1O1N2O1O2M2O1O1N2O1O1N3N1O1N2O1N2N2N2N3M2N2N2N2N2mDUNS21d4l1VIYNP2Mi4l1SI]No1Im4l1QIaNk1FS5k1oHeNh1AY5l1kHjNf1\\O_5k1hHoNc1XOd5k1eHSOb1TOh5k1cHWO^1QOP6h1^H_OZ1lNZ6l3WIWLj6U4fHmK\\7[52N1O2N1O2N1N2O2N1O001O1O1N2O001M3M4L4K5L4L4K5L4L4L4K5L4oNR1J6M2N3M3M1O010O001O00100O001O010O0010O01O1O010O0010O01O1O010O001O010O0010O00O2O001O0O2O001O0O2O001O001N1000001N1000001N1000001AiCROX in this image.", "objects": [{"patches": [96, 116, 117, 118, 137, 138, 139, 140, 159, 160, 161, 162, 163, 180, 181, 182, 183, 184, 185, 201, 202, 203, 204, 205, 206, 207, 222, 223, 243, 244, 268, 269, 289, 290], "bbox": [0.5436333333333333, 0.28089285714285717, 0.8741833333333333, 0.8471874999999999], "iscrowd": 0, "area": 14818.668600000003, "rle": {"size": [448, 600], "counts": "eh^4:f=1N3N1O1N2O1O1N2O1O2M2O1O1N2O1O1N3N1O1N2O1N2N2N2N3M2N2N2N2N2mDUNS21d4l1VIYNP2Mi4l1SI]No1Im4l1QIaNk1FS5k1oHeNh1AY5l1kHjNf1\\O_5k1hHoNc1XOd5k1eHSOb1TOh5k1cHWO^1QOP6h1^H_OZ1lNZ6l3WIWLj6U4fHmK\\7[52N1O2N1O2N1N2O2N1O001O1O1N2O001M3M4L4K5L4L4K5L4L4L4K5L4oNR1J6M2N3M3M1O010O001O00100O001O010O0010O01O1O010O0010O01O1O010O001O010O0010O00O2O001O0O2O001O0O2O001O001N1000001N1000001N1000001AiCROX in this image.", "objects": [{"patches": [243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.007075471698113208, 0.71503125, 1.0, 0.987328125], "iscrowd": 0, "area": 65477.22325, "rle": {"size": [640, 424], "counts": "h^2P1Pa0P2I700O100O100O10000O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O01000O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N20000000000000000000000000000000000000000000000001O001O00001O00001O001O00001O001O00001O00001O001O00001O00001OO100000000000000000000000000000000O100000000000000000000000000000000O10000000000000000000000000000000O0100000000000000000000000000000000O100000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O10000000000000000000000000000000000000000001O000000000000001O0000000000001O000000000000001O0000000000001O000000000000001O0000000000001O000000000000001O0000000000001O000000000000001O0000000000001O000000000000001O0000000000001O0000000000VA"}, "label": "a red table"}]} {"id": 298481, "image": "COCO_train2014_000000298481.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red table fartherest away from the windows\"."}], "task": "refering", "answer_template": "The \"the red table fartherest away from the windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.007075471698113208, 0.71503125, 1.0, 0.987328125], "iscrowd": 0, "area": 65477.22325, "rle": {"size": [640, 424], "counts": "h^2P1Pa0P2I700O100O100O10000O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O100O01000O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O100O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1N20000000000000000000000000000000000000000000000001O001O00001O00001O001O00001O001O00001O00001O001O00001O00001OO100000000000000000000000000000000O100000000000000000000000000000000O10000000000000000000000000000000O0100000000000000000000000000000000O100000000000000000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O100000000000000000000000000000000O10000000000000000000000000000000000000000001O000000000000001O0000000000001O000000000000001O0000000000001O000000000000001O0000000000001O000000000000001O0000000000001O000000000000001O0000000000001O000000000000001O0000000000001O0000000000VA"}, "label": "the red table fartherest away from the windows"}]} {"id": 298481, "image": "COCO_train2014_000000298481.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table under the chandelier\"."}], "task": "refering", "answer_template": "The \"a table under the chandelier\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 181, 195, 196, 210, 225, 227, 228, 240, 241, 242, 243, 255], "bbox": [0.0, 0.54259375, 0.2420990566037736, 0.774140625], "iscrowd": 0, "area": 5608.938050000001, "rle": {"size": [640, 424], "counts": "n6_BUOnMGb?Q1eBVOmMF^?R1hBWOc=f0aBXO`=b0hB\\OX=d0iBZOX=e0jBYOW=g0jBXOV=i0jBTOX=n0l21O1O1O1O2N1O10O00000001O0000B>A?EXo11RQNg0YO3M0000000O100000O100000000O100000O10O1000000O100000000O10O100000O100000000O10000000O010000ZOQWX6"}, "label": "a table under the chandelier"}]} {"id": 28154, "image": "COCO_train2014_000000028154.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady on left\"."}], "task": "refering", "answer_template": "The \"lady on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 105, 106, 107, 108, 122, 123, 124, 125, 139, 140, 141, 142, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 178, 189, 190, 191, 192, 193, 204, 205, 206, 207, 208, 209, 210, 211, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 262, 273, 274, 275, 276, 277, 278, 289, 290, 291, 292, 293, 294, 306, 307, 308, 323, 324, 325, 326, 340, 341, 342], "bbox": [0.0, 0.2555, 0.5093541666666667, 0.8959374999999999], "iscrowd": 0, "area": 48157.44045000002, "rle": {"size": [640, 480], "counts": "f:1nc02O2N2k\\ONWb05`]O3_b0O`]O1_b01_]OOab02_]ON_b05_]OKab07]]OJbb07^]OI`b0:^]OFbb0;]]OFbb0;^]OEab0=]]OCcb0>]]OBcb0m0VBAT8?eGHZ88`GNa82fFf0[9ZO]Fm0c9VOZFk0f9VOXFj0i9UOWFk0i9UOWFk0j9UOUFl0k9SOmEU1S:kN`Eb1a:]NSEo1n:PNbDa2];`MUDm2l;RMRDP3n;PMPDR3PDRB9l`0N2N2N3MQTc4"}, "label": "lady on left"}]} {"id": 28154, "image": "COCO_train2014_000000028154.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a black top holding a video game remote\"."}], "task": "refering", "answer_template": "The \"a woman in a black top holding a video game remote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 105, 106, 107, 108, 122, 123, 124, 125, 139, 140, 141, 142, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 178, 189, 190, 191, 192, 193, 204, 205, 206, 207, 208, 209, 210, 211, 221, 222, 223, 224, 225, 226, 227, 228, 229, 239, 240, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 262, 273, 274, 275, 276, 277, 278, 289, 290, 291, 292, 293, 294, 306, 307, 308, 323, 324, 325, 326, 340, 341, 342], "bbox": [0.0, 0.2555, 0.5093541666666667, 0.8959374999999999], "iscrowd": 0, "area": 48157.44045000002, "rle": {"size": [640, 480], "counts": "f:1nc02O2N2k\\ONWb05`]O3_b0O`]O1_b01_]OOab02_]ON_b05_]OKab07]]OJbb07^]OI`b0:^]OFbb0;]]OFbb0;^]OEab0=]]OCcb0>]]OBcb0m0VBAT8?eGHZ88`GNa82fFf0[9ZO]Fm0c9VOZFk0f9VOXFj0i9UOWFk0i9UOWFk0j9UOUFl0k9SOmEU1S:kN`Eb1a:]NSEo1n:PNbDa2];`MUDm2l;RMRDP3n;PMPDR3PDRB9l`0N2N2N3MQTc4"}, "label": "a woman in a black top holding a video game remote"}]} {"id": 28154, "image": "COCO_train2014_000000028154.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the legs of a person sitting on the couch\"."}], "task": "refering", "answer_template": "The \"the legs of a person sitting on the couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [269, 270, 271, 285, 286, 287, 288, 302, 303, 304, 305, 318, 319, 320, 321, 322, 334, 335, 336, 337, 338, 339, 350, 351, 352, 353, 354, 355, 356, 367, 368, 369, 370, 371, 372, 373, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.5543125, 0.662921875, 1.0, 0.986515625], "iscrowd": 0, "area": 27489.710450000006, "rle": {"size": [640, 480], "counts": "e_W52mc01O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O0O2O000O101O0O10001N100O101O0O10001N10001N10000O2O0i@kKd>i4N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O100O100O10000O10000O10000O100O10000O10000O10000O100O10000O10000O100O10000O10O10O10000O11O00001O001O001O00001O001O00001O001O001O00001O001O001OQ3oL[_O"}, "label": "the legs of a person sitting on the couch"}]} {"id": 28154, "image": "COCO_train2014_000000028154.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the legs of the person who isn ' t in the shot\"."}], "task": "refering", "answer_template": "The \"the legs of the person who isn ' t in the shot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [269, 270, 271, 285, 286, 287, 288, 302, 303, 304, 305, 318, 319, 320, 321, 322, 334, 335, 336, 337, 338, 339, 350, 351, 352, 353, 354, 355, 356, 367, 368, 369, 370, 371, 372, 373, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.5543125, 0.662921875, 1.0, 0.986515625], "iscrowd": 0, "area": 27489.710450000006, "rle": {"size": [640, 480], "counts": "e_W52mc01O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O0O2O000O101O0O10001N100O101O0O10001N10001N10000O2O0i@kKd>i4N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O100O100O10000O10000O10000O100O10000O10000O10000O100O10000O10000O100O10000O10O10O10000O11O00001O001O001O00001O001O00001O001O001O00001O001O001OQ3oL[_O"}, "label": "the legs of the person who isn ' t in the shot"}]} {"id": 28154, "image": "COCO_train2014_000000028154.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl wearing earrings and sitting in the middle of the couch is playing wii\"."}], "task": "refering", "answer_template": "The \"a girl wearing earrings and sitting in the middle of the couch is playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 134, 148, 149, 150, 151, 165, 166, 167, 168, 169, 177, 181, 182, 183, 184, 185, 186, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 229, 230, 231, 232, 233, 234, 235, 236, 237, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 262, 263, 264, 265, 266, 267, 268, 269, 278, 279, 280, 281, 282, 283, 284, 295, 296, 297, 298, 299, 300, 301, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 326, 327, 328, 329, 330, 331, 332, 343, 344, 345, 346, 347, 348, 349, 361, 362, 363, 364, 365, 379, 380, 381, 382], "bbox": [0.18354166666666666, 0.303890625, 1.0, 0.9870156250000001], "iscrowd": 0, "area": 70283.05050000001, "rle": {"size": [640, 480], "counts": "Q`g11oc03L4L4L4M3L3M4L4M2M3M3M3N2M3M3M3N2M4L3N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N3N1O1N2O1O00O010O10O0100O10O0100O010O100O010O10O0100O10O0100O010O10O0100O10N1M4L3M4L3M4L3N3L3M3M3M3M2N3M3M3M3M3M3UKaKRJb4l5`KQJc4n5_KnId4P6^KmIe4R6\\KkIg4S6\\KiIg4V6ZKgIi4W6ZKeIi4Z6XKcIk4[6WKbIl4]6VK`Il4^6VK_Im4_6VK]Im4b6TK[Io4c6TKYIo4f6RKWIQ5g6QKVIR5i6PKSIS5k6UJYF>i2`5m6SJ[F=d2c5o6QJ_F;_2g5Q7oIaF9[2k5R7nIdF7V2n5T7lIhF5Q2R6V7jIjF4l1U6[7fIjF4h1Y6^7bIlF5c1[6a7`ImF4a1]6c7^ImF4_1_6d7\\IoF5[1`6g7ZIoF5Z1a6g7ZIPG5W1b6i7YIQG4U1d6k7VIRG5R1f6l7UISG5o0g6o7SISG5n0h6o7RIUG6j0i6Q8QIVG5h0k6S8oHVG6e0l6U8mHXG6b0n6W8kHWG8`0n6Y8jHWG9?l6[8jHWG<;k6_8hHVG>9k6a8gHVG?7k6d8dHWGa03l6f8cHWGc00k6i8bHWGd0Oj6l9WIRFj6n9WIPFj6P:WInEj6R:XIkEi6V:WIiEi6W:XIgEi6Y:XIeEi6[:YIbEh6^:YI`Eh6`:YI^Eh6b:YI]Eg6d:ZIYEg6i9WHnFS1WOg6k9WHmFS1VOg6m9WHlFP9T9PGlFP9T9QGkFo8U9RGjFn8V9SGhFi8]9XGbF^8h9bGWFV8S:jGlEl7^:UH`Ec7i:^HVEX7T;iHjDn6`;RI`De6i;\\IUD[6U in this image.", "objects": [{"patches": [131, 132, 133, 134, 148, 149, 150, 151, 165, 166, 167, 168, 169, 177, 181, 182, 183, 184, 185, 186, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 229, 230, 231, 232, 233, 234, 235, 236, 237, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 262, 263, 264, 265, 266, 267, 268, 269, 278, 279, 280, 281, 282, 283, 284, 295, 296, 297, 298, 299, 300, 301, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 326, 327, 328, 329, 330, 331, 332, 343, 344, 345, 346, 347, 348, 349, 361, 362, 363, 364, 365, 379, 380, 381, 382], "bbox": [0.18354166666666666, 0.303890625, 1.0, 0.9870156250000001], "iscrowd": 0, "area": 70283.05050000001, "rle": {"size": [640, 480], "counts": "Q`g11oc03L4L4L4M3L3M4L4M2M3M3M3N2M3M3M3N2M4L3N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N3N1O1N2O1O00O010O10O0100O10O0100O010O100O010O10O0100O10O0100O010O10O0100O10N1M4L3M4L3M4L3N3L3M3M3M3M2N3M3M3M3M3M3UKaKRJb4l5`KQJc4n5_KnId4P6^KmIe4R6\\KkIg4S6\\KiIg4V6ZKgIi4W6ZKeIi4Z6XKcIk4[6WKbIl4]6VK`Il4^6VK_Im4_6VK]Im4b6TK[Io4c6TKYIo4f6RKWIQ5g6QKVIR5i6PKSIS5k6UJYF>i2`5m6SJ[F=d2c5o6QJ_F;_2g5Q7oIaF9[2k5R7nIdF7V2n5T7lIhF5Q2R6V7jIjF4l1U6[7fIjF4h1Y6^7bIlF5c1[6a7`ImF4a1]6c7^ImF4_1_6d7\\IoF5[1`6g7ZIoF5Z1a6g7ZIPG5W1b6i7YIQG4U1d6k7VIRG5R1f6l7UISG5o0g6o7SISG5n0h6o7RIUG6j0i6Q8QIVG5h0k6S8oHVG6e0l6U8mHXG6b0n6W8kHWG8`0n6Y8jHWG9?l6[8jHWG<;k6_8hHVG>9k6a8gHVG?7k6d8dHWGa03l6f8cHWGc00k6i8bHWGd0Oj6l9WIRFj6n9WIPFj6P:WInEj6R:XIkEi6V:WIiEi6W:XIgEi6Y:XIeEi6[:YIbEh6^:YI`Eh6`:YI^Eh6b:YI]Eg6d:ZIYEg6i9WHnFS1WOg6k9WHmFS1VOg6m9WHlFP9T9PGlFP9T9QGkFo8U9RGjFn8V9SGhFi8]9XGbF^8h9bGWFV8S:jGlEl7^:UH`Ec7i:^HVEX7T;iHjDn6`;RI`De6i;\\IUD[6U in this image.", "objects": [{"patches": [54, 55, 56, 77, 78, 79, 80, 99, 100, 101, 102, 103, 123, 124, 125, 126, 145, 146, 147, 148, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 261, 262, 263, 264, 265, 280, 281, 282, 283, 284, 285, 304, 305, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332, 333, 351, 352, 353, 354, 355, 356, 374, 375, 376, 377, 378, 379], "bbox": [0.17953125, 0.15010416666666665, 0.546984375, 0.9890208333333333], "iscrowd": 0, "area": 54691.69875000001, "rle": {"size": [480, 640], "counts": "nVf12m>4K4L5L4K5L4K5L4K4L4M2M4M2M3N3L3M4M2M3N3L3N2M4L3M4L3M3M4L3M3M3M3M2N3L4M3M2L5L4K5K4L5L4L4O0O2N2O1N1O2O1N2N101N2eHlJg4b5fJdJZ5a7O1O100O100O1O100O1O100O1O100O1O100O1O100O100O100O100O100O100O100O100O10O0100O100O100O100O1kM]GmMc8P2hGgMY8V2QHbMP8[2ZH^Mf7_2eHXM\\7d2oHTMR7i2RIUMo6h2UIWMk6f2YIXMh6e2]IXMd6e2`IYMa6d2cI[M]6a2iI\\MX6a2lI]MU6b2mI\\MT6c2nI\\MR6d2PJYMQ6f2U300O100O100O100O100O100O100O100O100O10000001O000iKbMlK^2S4eMkK[2T4gMkKZ2S4iMkKW2T4kMkKU2T4nMjKR2U4PNjKQ2U4PNjKP2U4SNiKm1V4UNiKk1V4XNhKh1W4ZNhKg1V4\\NhKd1W4^NhKb1W4`NhK`1X4bNfK^1Y4dNfK]1X4eNgK[1X4gNgKY1Y4gNgKY1Y4hNfKY1Y4gNgKY1Y4hNfKX1Z4hNfKX1Z4iNeKW1[4iNeKX1Z4iNeKX1Z4hNfKY1Y4hNfKY1Y4gNgKZ1X4gNgKZ1X4fNhK[1W4fNhK[1W4eNiK\\1V4eNiK[1W4eNiK\\1V4eNiK\\1V4dNjK]1U4dNjK]1U4cNkK^1:UM6^1_O^13[M>W1_O_1N^Mc0T1^O_1KaMf0P1_O`1FdMk0m0^O`1BhMo0h0_Oa1^OjMS1e0_Ob1YOmMX1b0^Oa1VORN[1=_Ob1ROTN_1;^Ob1nNXNb17@b1hN\\Nh13_Od1aN^NP20]OS2mMSNe27QOW2dMgM[3?dNT4\\1ZLVNg3i1_LQNa3o1eLkM\\3T2iLgMX3X2nLbMT3\\2QM_MP3`2VMZMl2e2YMUMi2i2]MQMd2o2h4N101O1N101N2O0O2O1N101N2Oc0\\O]1dN1N2O2M2O1N2O1N2O2M2N2NacW4"}, "label": "man holding two plates"}]} {"id": 11774, "image": "COCO_train2014_000000011774.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"short guy with smores\"."}], "task": "refering", "answer_template": "The \"short guy with smores\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 77, 78, 79, 80, 99, 100, 101, 102, 103, 123, 124, 125, 126, 145, 146, 147, 148, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 261, 262, 263, 264, 265, 280, 281, 282, 283, 284, 285, 304, 305, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332, 333, 351, 352, 353, 354, 355, 356, 374, 375, 376, 377, 378, 379], "bbox": [0.17953125, 0.15010416666666665, 0.546984375, 0.9890208333333333], "iscrowd": 0, "area": 54691.69875000001, "rle": {"size": [480, 640], "counts": "nVf12m>4K4L5L4K5L4K5L4K4L4M2M4M2M3N3L3M4M2M3N3L3N2M4L3M4L3M3M4L3M3M3M3M2N3L4M3M2L5L4K5K4L5L4L4O0O2N2O1N1O2O1N2N101N2eHlJg4b5fJdJZ5a7O1O100O100O1O100O1O100O1O100O1O100O1O100O100O100O100O100O100O100O100O10O0100O100O100O100O1kM]GmMc8P2hGgMY8V2QHbMP8[2ZH^Mf7_2eHXM\\7d2oHTMR7i2RIUMo6h2UIWMk6f2YIXMh6e2]IXMd6e2`IYMa6d2cI[M]6a2iI\\MX6a2lI]MU6b2mI\\MT6c2nI\\MR6d2PJYMQ6f2U300O100O100O100O100O100O100O100O100O10000001O000iKbMlK^2S4eMkK[2T4gMkKZ2S4iMkKW2T4kMkKU2T4nMjKR2U4PNjKQ2U4PNjKP2U4SNiKm1V4UNiKk1V4XNhKh1W4ZNhKg1V4\\NhKd1W4^NhKb1W4`NhK`1X4bNfK^1Y4dNfK]1X4eNgK[1X4gNgKY1Y4gNgKY1Y4hNfKY1Y4gNgKY1Y4hNfKX1Z4hNfKX1Z4iNeKW1[4iNeKX1Z4iNeKX1Z4hNfKY1Y4hNfKY1Y4gNgKZ1X4gNgKZ1X4fNhK[1W4fNhK[1W4eNiK\\1V4eNiK[1W4eNiK\\1V4eNiK\\1V4dNjK]1U4dNjK]1U4cNkK^1:UM6^1_O^13[M>W1_O_1N^Mc0T1^O_1KaMf0P1_O`1FdMk0m0^O`1BhMo0h0_Oa1^OjMS1e0_Ob1YOmMX1b0^Oa1VORN[1=_Ob1ROTN_1;^Ob1nNXNb17@b1hN\\Nh13_Od1aN^NP20]OS2mMSNe27QOW2dMgM[3?dNT4\\1ZLVNg3i1_LQNa3o1eLkM\\3T2iLgMX3X2nLbMT3\\2QM_MP3`2VMZMl2e2YMUMi2i2]MQMd2o2h4N101O1N101N2O0O2O1N101N2Oc0\\O]1dN1N2O2M2O1N2O1N2O2M2N2NacW4"}, "label": "short guy with smores"}]} {"id": 11774, "image": "COCO_train2014_000000011774.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a red shirt\"."}], "task": "refering", "answer_template": "The \"a man in a red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 59, 60, 61, 62, 82, 83, 84, 85, 106, 107, 108, 129, 130, 131, 132, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 295, 312, 313, 314, 315, 316, 317, 318, 335, 336, 337, 338, 339, 340, 357, 358, 359, 360, 361, 362, 363, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.541015625, 0.096625, 0.884828125, 0.9842708333333333], "iscrowd": 0, "area": 61090.31834999999, "rle": {"size": [480, 640], "counts": "_]R54k>;D4M3M4L3M4K4M3M4ZHkNZ1X1ZNROd1R1PNXOo1l0dM^O[2e0YMEf2?mLKR38cL1[33aLO^35]LMb38WLKh3:RLHl3?lKDS4a0`KH_4>oJOP56`J6^50QJ=n5[4oKjGW1W8hNmGT1T8kNPHQ1Q8nNTHl0n7SOVHi0k7VOYHf0h7XO^Hc0c7\\OaH?a7@bH;a7DaH9`7GcH5_7JdH2^7LeH1]7NfHM]72eHK]74fHG]78eHD^7;eHA]7>fH]O]7a0fH\\O\\7c0gHXO\\7g0fHUO]7j0fHRO\\7m0fHoN]7P1fHkN]7S1fHjN\\7U1gHfN\\7Y1gHcN[7\\1gH`N\\7`1fH[N]7e1o200000000O11O0O101O00001O001O00001O00001O00001O001O00001O00001O001O00001N101O1O1O1O1O1O1O1O1O1O1O1O1O2N1O3M4L3M3M3M3Mm0SO>B5K001O1O0O2O001O1O001O1O001O1O001O1O0000001O1O001O1O001O1O001O1O1O0O2O1O001O1O001O1O1O001O1O001N2K4M4L4K5L3M4K5L3L5L4L3L5L4K5mL]HdNf7\\1`H[Nd7c1dHRNa7m1fHiM]7W2iH`M[7^2mHWMX7h2nHoLU7P3SIeLR7Z3TI]LP7b3WITLm6j3U201N2O0O2O3M2M4M3L3N3L4M2N3L3N3L4M2N3L4M6I8I6I8I7I7H7J7H8I6Ig\\R1"}, "label": "a man in a red shirt"}]} {"id": 11774, "image": "COCO_train2014_000000011774.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a red shirt eating sweets with his friend\"."}], "task": "refering", "answer_template": "The \"a man in a red shirt eating sweets with his friend\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 59, 60, 61, 62, 82, 83, 84, 85, 106, 107, 108, 129, 130, 131, 132, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 295, 312, 313, 314, 315, 316, 317, 318, 335, 336, 337, 338, 339, 340, 357, 358, 359, 360, 361, 362, 363, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.541015625, 0.096625, 0.884828125, 0.9842708333333333], "iscrowd": 0, "area": 61090.31834999999, "rle": {"size": [480, 640], "counts": "_]R54k>;D4M3M4L3M4K4M3M4ZHkNZ1X1ZNROd1R1PNXOo1l0dM^O[2e0YMEf2?mLKR38cL1[33aLO^35]LMb38WLKh3:RLHl3?lKDS4a0`KH_4>oJOP56`J6^50QJ=n5[4oKjGW1W8hNmGT1T8kNPHQ1Q8nNTHl0n7SOVHi0k7VOYHf0h7XO^Hc0c7\\OaH?a7@bH;a7DaH9`7GcH5_7JdH2^7LeH1]7NfHM]72eHK]74fHG]78eHD^7;eHA]7>fH]O]7a0fH\\O\\7c0gHXO\\7g0fHUO]7j0fHRO\\7m0fHoN]7P1fHkN]7S1fHjN\\7U1gHfN\\7Y1gHcN[7\\1gH`N\\7`1fH[N]7e1o200000000O11O0O101O00001O001O00001O00001O00001O001O00001O00001O001O00001N101O1O1O1O1O1O1O1O1O1O1O1O1O2N1O3M4L3M3M3M3Mm0SO>B5K001O1O0O2O001O1O001O1O001O1O001O1O0000001O1O001O1O001O1O001O1O1O0O2O1O001O1O001O1O1O001O1O001N2K4M4L4K5L3M4K5L3L5L4L3L5L4K5mL]HdNf7\\1`H[Nd7c1dHRNa7m1fHiM]7W2iH`M[7^2mHWMX7h2nHoLU7P3SIeLR7Z3TI]LP7b3WITLm6j3U201N2O0O2O3M2M4M3L3N3L4M2N3L3N3L4M2N3L4M6I8I6I8I7I7H7J7H8I6Ig\\R1"}, "label": "a man in a red shirt eating sweets with his friend"}]} {"id": 454144, "image": "COCO_train2014_000000454144.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with blue full sleeves and hand inside pocket\"."}], "task": "refering", "answer_template": "The \"a man with blue full sleeves and hand inside pocket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 89, 90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 319, 320, 321, 342, 343, 344], "bbox": [0.854078125, 0.187018779342723, 1.0, 0.9870187793427231], "iscrowd": 0, "area": 26455.267, "rle": {"size": [426, 640], "counts": "QlS7:h<;E in this image.", "objects": [{"patches": [67, 68, 89, 90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 319, 320, 321, 342, 343, 344], "bbox": [0.854078125, 0.187018779342723, 1.0, 0.9870187793427231], "iscrowd": 0, "area": 26455.267, "rle": {"size": [426, 640], "counts": "QlS7:h<;E in this image.", "objects": [{"patches": [124, 145, 146, 147, 148, 168, 169, 170, 171, 192, 193, 194, 215, 216, 217, 238, 239, 240, 261, 262, 284, 285], "bbox": [0.3254375, 0.3886384976525822, 0.48015625, 0.8718309859154929], "iscrowd": 0, "area": 11183.286850000006, "rle": {"size": [426, 640], "counts": "^gf23V=7I6I2O2N2M3N2N2M3N2N2N2N1O2N2Nb0^O1O2QFVN[8l1_G[N^8h1_G\\N^8e1aG]N\\8e1cG^NY8e1dG\\N[8f1dGXN]8j1aGTNa8P2XGRNi8R2oFQNR9R2hFVNS9U3K5K5K5K5L4K5L4L4L4K5L4L4K5L2N2N2M3N2N1NWNZIoLg6o2\\IoLe6o2_ImLc6Q3`ImLa6S3aIiLa6V3`IgLc6Y3^IbLf6^3ZIULS7j3oHiK]7W4k0O10000O10000O101O1N2O1O001O1O1O3L4M3M1O2UOj0D=J66J;^E`Mb9V3O01O2N1M2N2M3M4M4K5L3^OZEZNj:[1cE_Na:Y1iEaNZ:f0U[Z4"}, "label": "a woman bending over to pick something up off a table"}]} {"id": 454144, "image": "COCO_train2014_000000454144.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman leaning over wearing a white jacket\"."}], "task": "refering", "answer_template": "The \"a woman leaning over wearing a white jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 145, 146, 147, 148, 168, 169, 170, 171, 192, 193, 194, 215, 216, 217, 238, 239, 240, 261, 262, 284, 285], "bbox": [0.3254375, 0.3886384976525822, 0.48015625, 0.8718309859154929], "iscrowd": 0, "area": 11183.286850000006, "rle": {"size": [426, 640], "counts": "^gf23V=7I6I2O2N2M3N2N2M3N2N2N2N1O2N2Nb0^O1O2QFVN[8l1_G[N^8h1_G\\N^8e1aG]N\\8e1cG^NY8e1dG\\N[8f1dGXN]8j1aGTNa8P2XGRNi8R2oFQNR9R2hFVNS9U3K5K5K5K5L4K5L4L4L4K5L4L4K5L2N2N2M3N2N1NWNZIoLg6o2\\IoLe6o2_ImLc6Q3`ImLa6S3aIiLa6V3`IgLc6Y3^IbLf6^3ZIULS7j3oHiK]7W4k0O10000O10000O101O1N2O1O001O1O1O3L4M3M1O2UOj0D=J66J;^E`Mb9V3O01O2N1M2N2M3M4M4K5L3^OZEZNj:[1cE_Na:Y1iEaNZ:f0U[Z4"}, "label": "a woman leaning over wearing a white jacket"}]} {"id": 454144, "image": "COCO_train2014_000000454144.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man dressed in a black leather jacket and denim pants is standing while looking downward\"."}], "task": "refering", "answer_template": "The \"a man dressed in a black leather jacket and denim pants is standing while looking downward\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 103, 104, 105, 106, 127, 128, 129, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 218, 219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 288, 289, 290, 311, 312, 313, 334, 335, 336], "bbox": [0.49293750000000003, 0.19988262910798124, 0.66759375, 1.0], "iscrowd": 0, "area": 26682.99930000001, "rle": {"size": [426, 640], "counts": "ZZS41V=:G9F:B>]Ob0_O>C in this image.", "objects": [{"patches": [80, 81, 82, 103, 104, 105, 106, 127, 128, 129, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 218, 219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 288, 289, 290, 311, 312, 313, 334, 335, 336], "bbox": [0.49293750000000003, 0.19988262910798124, 0.66759375, 1.0], "iscrowd": 0, "area": 26682.99930000001, "rle": {"size": [426, 640], "counts": "ZZS41V=:G9F:B>]Ob0_O>C in this image.", "objects": [{"patches": [199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 318, 319, 339, 340, 341, 342], "bbox": [0.6405937500000001, 0.529037558685446, 0.891046875, 1.0], "iscrowd": 0, "area": 22408.05334999999, "rle": {"size": [426, 640], "counts": "cmZ58ciIAX6:mIES66RJIo53VJLj50ZJ0f5K`J4`5HdJ8]5BiJ=W5_OmJa0S5ZOSKe0m4WOWKi0i4RO]Km0c4oNbKP1_4kNeKU1[4fNkKY1U4cNoK\\1R4_NTL`1k701O001O1O001O001O001O0O2O1O00?A`0@Pgl0"}, "label": "a black sofa with a brown or gold pattern"}]} {"id": 454144, "image": "COCO_train2014_000000454144.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the long black couch\"."}], "task": "refering", "answer_template": "The \"the long black couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 313, 314, 315, 316, 317, 318, 319, 339, 340, 341, 342], "bbox": [0.6405937500000001, 0.529037558685446, 0.891046875, 1.0], "iscrowd": 0, "area": 22408.05334999999, "rle": {"size": [426, 640], "counts": "cmZ58ciIAX6:mIES66RJIo53VJLj50ZJ0f5K`J4`5HdJ8]5BiJ=W5_OmJa0S5ZOSKe0m4WOWKi0i4RO]Km0c4oNbKP1_4kNeKU1[4fNkKY1U4cNoK\\1R4_NTL`1k701O001O1O001O001O001O0O2O1O00?A`0@Pgl0"}, "label": "the long black couch"}]} {"id": 454144, "image": "COCO_train2014_000000454144.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is a grey colored sofa chair in front of three people\"."}], "task": "refering", "answer_template": "The \"there is a grey colored sofa chair in front of three people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [256, 278, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 308, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.12453125000000001, 0.7914084507042253, 0.415125, 0.9892488262910797], "iscrowd": 0, "area": 12998.710750000004, "rle": {"size": [426, 640], "counts": "PdQ1c0d<7K4M4K4L4L3N2M3M4L3N2M3M3M3M3N2M4M2N2O1N1O0000000000000001O0000000000000000001O00000000000000001O0000000000000000001O0000000000000000001O00000000000000001O0000000000000000001O0000000000000000001O00000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O00000000000000001O000000001O001O00001O001O00001O001O001O00001O001O00001O3M6J5K5K5K5K5K5K5K6J5KVck4"}, "label": "there is a grey colored sofa chair in front of three people"}]} {"id": 454144, "image": "COCO_train2014_000000454144.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"steal blue chair headrest in room\"."}], "task": "refering", "answer_template": "The \"steal blue chair headrest in room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [256, 278, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 308, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.12453125000000001, 0.7914084507042253, 0.415125, 0.9892488262910797], "iscrowd": 0, "area": 12998.710750000004, "rle": {"size": [426, 640], "counts": "PdQ1c0d<7K4M4K4L4L3N2M3M4L3N2M3M3M3M3N2M4M2N2O1N1O0000000000000001O0000000000000000001O00000000000000001O0000000000000000001O0000000000000000001O00000000000000001O0000000000000000001O0000000000000000001O00000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O00000000000000001O000000001O001O00001O001O00001O001O001O00001O001O00001O3M6J5K5K5K5K5K5K5K6J5KVck4"}, "label": "steal blue chair headrest in room"}]} {"id": 101892, "image": "COCO_train2014_000000101892.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"beach umbrella on the beach ,\"."}], "task": "refering", "answer_template": "The \"beach umbrella on the beach ,\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [240, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 262, 263, 264, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301], "bbox": [0.20787109375, 0.5780937500000001, 0.7837109375, 0.7016875], "iscrowd": 0, "area": 14780.291949999995, "rle": {"size": [640, 512], "counts": "dXS21oc01N101O001N2O001O0O2O001N2O001O0O2O1O0O2O001O1N101O0O2O001O1N101O0O2O1O001N101O0O2O1O001N101O1N1000001N10000O1000000O10001O0O10000O1000000O2O000O1000000O10001N1000000O10000O101O000O10000O1000000O2O000O1000000O10001N1000000O1000000O2O0000000000000000001O00000000000000001O00000000000000001O0000000000000000001O000O100000000000000000000000000000000000001O000001O001O001O001O001O001O001O001O001O001O001O001O001O010O001O001O1O001O001O001O001O001O001O001O001O001O001O0010O01O001O001O001O001O001O001O1O001O001O001O001O001O001O010O001O001O001O001O001O001O001N1O2N1O2N1O2M2O2N1O2N2N1O2N1O2N1O2NdnT2"}, "label": "beach umbrella on the beach ,"}]} {"id": 347655, "image": "COCO_train2014_000000347655.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bare legs and feet\"."}], "task": "refering", "answer_template": "The \"the bare legs and feet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 20, 21, 22, 38, 39, 40, 41, 42, 44, 45, 62, 63, 64, 65, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 135, 136, 137], "bbox": [0.62821875, 0.0002347417840375587, 0.99765625, 0.3702582159624413], "iscrowd": 0, "area": 16946.97425, "rle": {"size": [426, 640], "counts": "PeW51Y=1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O1O2N1O2N1O101O0010O0001O010O2mEYM^9h2]F\\Mc9f2XF\\Mj9d2QF`Mo9n2000O2O000O2O00000N3N1N2O1N3N1N2O1O2N1O100O2N1O1O2O0O2N3M2N2O1N2N2N2O1N2N2M3N1O1O1N2O1O1O]MbNZI]1g6dNYI[1g6fNYIY1g6hNYIW1g6jNYIT1h6mNXIR1h6oNWIQ1j6oNVIo0k6ROUIm0k6TOUIk0k6WOTIh0l6YOTIe0m6\\OSIc0m6^OSIa0m6@SI>n6CSI;m6FSI9WNYOk7`0mI5XN]Oj7?oI2UNBk7=SJMQNHk7=XJDlM2k7<\\J]OhM:j7:bJWObMa0l7:QKGo4;mJFT5;iJGW5;fJFZ5WJCi5?TJBl5`0PJAR6a0jI@V6b0gI_OY6d0bI^O^6d0_I]Oa6e0[I]Oe6f0WI[Oi6g0TIZOl6h0PIZOP7h0mHXOT7j0gHYOY7i0bHZO^7h0\\H\\Od7f0WH]Oi7e0PH@P8c0gGCY8f21000000O10000O1000001N1000000O1000000O1000000O4M4L3M4K4M4L3M4K4M4LT<"}, "label": "the bare legs and feet"}]} {"id": 347655, "image": "COCO_train2014_000000347655.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person ' s sandy feet\"."}], "task": "refering", "answer_template": "The \"a person ' s sandy feet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 20, 21, 22, 38, 39, 40, 41, 42, 44, 45, 62, 63, 64, 65, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 135, 136, 137], "bbox": [0.62821875, 0.0002347417840375587, 0.99765625, 0.3702582159624413], "iscrowd": 0, "area": 16946.97425, "rle": {"size": [426, 640], "counts": "PeW51Y=1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O1O2N1O2N1O101O0010O0001O010O2mEYM^9h2]F\\Mc9f2XF\\Mj9d2QF`Mo9n2000O2O000O2O00000N3N1N2O1N3N1N2O1O2N1O100O2N1O1O2O0O2N3M2N2O1N2N2N2O1N2N2M3N1O1O1N2O1O1O]MbNZI]1g6dNYI[1g6fNYIY1g6hNYIW1g6jNYIT1h6mNXIR1h6oNWIQ1j6oNVIo0k6ROUIm0k6TOUIk0k6WOTIh0l6YOTIe0m6\\OSIc0m6^OSIa0m6@SI>n6CSI;m6FSI9WNYOk7`0mI5XN]Oj7?oI2UNBk7=SJMQNHk7=XJDlM2k7<\\J]OhM:j7:bJWObMa0l7:QKGo4;mJFT5;iJGW5;fJFZ5WJCi5?TJBl5`0PJAR6a0jI@V6b0gI_OY6d0bI^O^6d0_I]Oa6e0[I]Oe6f0WI[Oi6g0TIZOl6h0PIZOP7h0mHXOT7j0gHYOY7i0bHZO^7h0\\H\\Od7f0WH]Oi7e0PH@P8c0gGCY8f21000000O10000O1000001N1000000O1000000O1000000O4M4L3M4K4M4L3M4K4M4LT<"}, "label": "a person ' s sandy feet"}]} {"id": 347655, "image": "COCO_train2014_000000347655.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a female holding a knife crouched down and cutting a designer cake\"."}], "task": "refering", "answer_template": "The \"a female holding a knife crouched down and cutting a designer cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 23, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 53, 54, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 147, 148, 149, 150, 151, 152, 153, 154, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326], "bbox": [0.003, 0.0022535211267605635, 0.7180937500000001, 0.9819718309859155], "iscrowd": 0, "area": 83557.42399999998, "rle": {"size": [426, 640], "counts": "ej0SR:BmE?R:CmE=S:ClE>T:BlE>T:CjE>U:CjE>V:BjE>V:ChE>X:BgE?Z8\\OaH5UO?o7HkHJTO?P8HlHITO?o7ImHHSO`0o7InHHQO`0P8JnHFRO`0n7LPIDQOa0n7LQICPOb0n7MQIBPOa0n7NRIAoNb0n7NSI@nNc0m70TI]OoNc0l71UI\\OnNd0l71VI[OnNd0k73VIYOnNe0k73WIYOmNd0j75YIWOlNe0j76YIUOmNe0i77ZITOlNf0i78ZIROlNg0i78[IQOlNg0g7;\\InNlNh0g7<\\ImNlNg0g7>\\IkNlNh0g7>]IjNlNh0e7a0^IgNlNi0e7b0^IeNmNi0d7e0]IbNnNj0d7f0]I`NnNk0d7h0\\I^NoNj0c7l0\\IZNPOk0c7m0\\IXNQOk0b7o0\\IHc69\\IHc69]IGa6<^IDa6=^IDa6=_IC`6?^IBa6?_IA_6a0aI_O^6b0aI_O^6b0bI^O]6c0bI^O]6c0cI]O\\6d0dI\\O[6e0dI\\O[6e0eI[OZ6f0fIZOY6g0fIZOY6g0gIYOX6h0gIYOX6h0hIYOV6h0jIYOT6h0kIZOS6f0nI[OP6f0oI\\OP6d0PJ]On5d0QJ^Om5c0SJ^Ok5c0TJ_Oj5b0UJ@i5a0WJ@g5a0XJAf5`0ZJAd5?\\JCb5>^JC`5>_JD_5>`JC^5>aJD]5=bJE\\5 in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 23, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 53, 54, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 147, 148, 149, 150, 151, 152, 153, 154, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326], "bbox": [0.003, 0.0022535211267605635, 0.7180937500000001, 0.9819718309859155], "iscrowd": 0, "area": 83557.42399999998, "rle": {"size": [426, 640], "counts": "ej0SR:BmE?R:CmE=S:ClE>T:BlE>T:CjE>U:CjE>V:BjE>V:ChE>X:BgE?Z8\\OaH5UO?o7HkHJTO?P8HlHITO?o7ImHHSO`0o7InHHQO`0P8JnHFRO`0n7LPIDQOa0n7LQICPOb0n7MQIBPOa0n7NRIAoNb0n7NSI@nNc0m70TI]OoNc0l71UI\\OnNd0l71VI[OnNd0k73VIYOnNe0k73WIYOmNd0j75YIWOlNe0j76YIUOmNe0i77ZITOlNf0i78ZIROlNg0i78[IQOlNg0g7;\\InNlNh0g7<\\ImNlNg0g7>\\IkNlNh0g7>]IjNlNh0e7a0^IgNlNi0e7b0^IeNmNi0d7e0]IbNnNj0d7f0]I`NnNk0d7h0\\I^NoNj0c7l0\\IZNPOk0c7m0\\IXNQOk0b7o0\\IHc69\\IHc69]IGa6<^IDa6=^IDa6=_IC`6?^IBa6?_IA_6a0aI_O^6b0aI_O^6b0bI^O]6c0bI^O]6c0cI]O\\6d0dI\\O[6e0dI\\O[6e0eI[OZ6f0fIZOY6g0fIZOY6g0gIYOX6h0gIYOX6h0hIYOV6h0jIYOT6h0kIZOS6f0nI[OP6f0oI\\OP6d0PJ]On5d0QJ^Om5c0SJ^Ok5c0TJ_Oj5b0UJ@i5a0WJ@g5a0XJAf5`0ZJAd5?\\JCb5>^JC`5>_JD_5>`JC^5>aJD]5=bJE\\5 in this image.", "objects": [{"patches": [46, 47, 63, 64, 69, 70, 86, 87, 88, 92, 110, 111, 112, 115, 161, 184, 185, 186, 187, 207, 208, 230, 253, 276], "bbox": [0.0, 0.12360416666666667, 0.8949374999999999, 0.7595416666666667], "iscrowd": 0, "area": 7696.119499999995, "rle": {"size": [480, 640], "counts": "X:S1e;U2YKQMSL\\4n3aL[KX3f4bMbJ[2_5eMdJW2]5iMfJS2[5mMhJo1Y5PNjJm1W5SNlJi1U5WNnJe1S5ZNQKb1P5^NRK_1o4aNTK[1m4eNVKW1k4hNYKT1h4lNZKQ1g4oN\\Km0e4RO_Kj0b4VOaKg0_4YOcKd0^4\\OeK`0\\4_OgK>Z4BhK;Z4DiK8Z4EiK8Z4FhK7[4GhK5[4IgK4\\4IgK4\\4JgK2\\4LfK2\\4KgK4Z4JhK5Y4IjK6V4HlK7U4FnK9S4EPL:P4DRL;o3BTL=m3AVL>j3@XL?i3_OYL`0h3]O[Lc0e3[O]Ld0d3ZO^Le0c3XO`Lh0`3VOcLh0^3VOdLi0]3UOeLk0h8O1O100O1O2O0O1O100O100O1O101N1O100O1O100O101O0O10000O10000O2O000O100O10000O101O0O10000O1000lWe50ThZJ2N1O1O1O1O1O1N2O1O1O1O1O1O1O1O1N2O1O10000000000000000O1000000000000000000000001O00000O100000000000000000O100O1O1O100O001O100O1O1O100O1O1O1O100O1O1O10O01O1O100O1O1O100O1O1O1O10ggo0"}, "label": "the table holding the trays of food and coffee"}]} {"id": 568840, "image": "COCO_train2014_000000568840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the part of the table in the top right corner\"."}], "task": "refering", "answer_template": "The \"the part of the table in the top right corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 87, 88, 89, 90, 91, 111, 112, 113, 114, 136, 137, 160], "bbox": [0.5553125, 0.006208333333333333, 1.0, 0.3718958333333333], "iscrowd": 0, "area": 24269.746900000006, "rle": {"size": [480, 640], "counts": "SlV54l>3M4L3M4L3M4L3M4L3M4L3M3M00000000000000000000000000000000000000O1000000000000000000000000000000000000000000O1000000000000000000000000000001O000000000O10000000000001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O1O1O1O2N1O1O1O2N1O1O2N1O000000O10000O10000O10000O10000O10000O11O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O100O1O1O1O001O1O1O1O1O1O1O1O001O1O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O1O001O001O1O001O001O1O001O001O1O001O001O1O001O001f1ZN^2aMeK"}, "label": "the part of the table in the top right corner"}]} {"id": 568840, "image": "COCO_train2014_000000568840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"donut with chocolate and vanilla frosting\"."}], "task": "refering", "answer_template": "The \"donut with chocolate and vanilla frosting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 241, 257, 258, 259, 260, 261, 262, 263, 264, 280, 281, 282, 283, 284, 285, 286, 287, 303, 304, 305, 306, 307, 308, 309, 310, 326, 327, 328, 329, 330, 331, 332, 333, 349, 350, 351, 352, 353, 354, 355, 356, 373, 374, 375, 376, 377, 378], "bbox": [0.17528125, 0.5528125, 0.515734375, 0.9887708333333334], "iscrowd": 0, "area": 36079.40015, "rle": {"size": [480, 640], "counts": "ljd14k>>A>C=B4M3M4K4M3L5L3M3L5L3M3L5L3L5L3M3L5L3L4M4L3L4M4L3L4M4K4M2N1N3N1N3N2N101N1O2O0O2N100O2N101N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O2O00001O0000O10000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000O1000000O100O1O100O2N2O1N2N101N2N2O1N1O2O1N2N2O0O2N2O1N2N101N2N2O1N1O2O1N2N2O0O2N2O1N2N101N2N2O0O2N2O1N2L3M4L4L4L4L3M4L4L4L4L3M4L4I7I7ICBZo`4"}, "label": "donut with chocolate and vanilla frosting"}]} {"id": 568840, "image": "COCO_train2014_000000568840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a doughnut with icing in the shape of a spiderweb\"."}], "task": "refering", "answer_template": "The \"a doughnut with icing in the shape of a spiderweb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 241, 257, 258, 259, 260, 261, 262, 263, 264, 280, 281, 282, 283, 284, 285, 286, 287, 303, 304, 305, 306, 307, 308, 309, 310, 326, 327, 328, 329, 330, 331, 332, 333, 349, 350, 351, 352, 353, 354, 355, 356, 373, 374, 375, 376, 377, 378], "bbox": [0.17528125, 0.5528125, 0.515734375, 0.9887708333333334], "iscrowd": 0, "area": 36079.40015, "rle": {"size": [480, 640], "counts": "ljd14k>>A>C=B4M3M4K4M3L5L3M3L5L3M3L5L3L5L3M3L5L3L4M4L3L4M4L3L4M4K4M2N1N3N1N3N2N101N1O2O0O2N100O2N101N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O101N1O2O00001O0000O10000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000O1000000O100O1O100O2N2O1N2N101N2N2O1N1O2O1N2N2O0O2N2O1N2N101N2N2O1N1O2O1N2N2O0O2N2O1N2N101N2N2O0O2N2O1N2L3M4L4L4L4L3M4L4L4L4L3M4L4I7I7ICBZo`4"}, "label": "a doughnut with icing in the shape of a spiderweb"}]} {"id": 568840, "image": "COCO_train2014_000000568840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a krispy creme coffee cup that is filled nearly to the brim\"."}], "task": "refering", "answer_template": "The \"a krispy creme coffee cup that is filled nearly to the brim\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 338, 339, 340], "bbox": [0.637078125, 0.23820833333333333, 1.0, 0.8584166666666666], "iscrowd": 0, "area": 50464.34039999999, "rle": {"size": [480, 640], "counts": "`ao5;Z>?B>A`0@?A?A`0A>A?A`0@?A?B?@?A?A`0F7M2N1O2O1N1O2O1N1O2N101N2N101N2N1O2O1N1O2O1N101N101O1O001N2O001O0O10001O000O101O00001N1000001O0O101O000O1010O000001O00001O0000010O00001O0000O100O101N100O100O100O100O101O0O100O100O100O100O2O0O100O100O1O1O1O1N3N1N2O1N2O1N2O1O1N2O2M2O1N2O1N3N1O1N3N1N2O2M2O1N3N1O2M2O1N3N1N2O2M2O2N1N2O2M2O1N3N1N3N1O1N3N1N2O2M2O1N3N1O2M2O1N3N1N2O2N1N3N1N2O2M2O1N3N1O2M2O1N3N1N3N2M2O2N2M2O2M3N1N3N2M2O2N2M2O2M3N1N3N2M2O2N2M2O2M3N1N3N2M2O2N2M2O2M3N1N3NY1fNbI"}, "label": "a krispy creme coffee cup that is filled nearly to the brim"}]} {"id": 568840, "image": "COCO_train2014_000000568840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the coffee cup closest to the chocolate glazed donut with the spiderweb design\"."}], "task": "refering", "answer_template": "The \"the coffee cup closest to the chocolate glazed donut with the spiderweb design\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 338, 339, 340], "bbox": [0.637078125, 0.23820833333333333, 1.0, 0.8584166666666666], "iscrowd": 0, "area": 50464.34039999999, "rle": {"size": [480, 640], "counts": "`ao5;Z>?B>A`0@?A?A`0A>A?A`0@?A?B?@?A?A`0F7M2N1O2O1N1O2O1N1O2N101N2N101N2N1O2O1N1O2O1N101N101O1O001N2O001O0O10001O000O101O00001N1000001O0O101O000O1010O000001O00001O0000010O00001O0000O100O101N100O100O100O100O101O0O100O100O100O100O2O0O100O100O1O1O1O1N3N1N2O1N2O1N2O1O1N2O2M2O1N2O1N3N1O1N3N1N2O2M2O1N3N1O2M2O1N3N1N2O2M2O2N1N2O2M2O1N3N1N3N1O1N3N1N2O2M2O1N3N1O2M2O1N3N1N2O2N1N3N1N2O2M2O1N3N1O2M2O1N3N1N3N2M2O2N2M2O2M3N1N3N2M2O2N2M2O2M3N1N3N2M2O2N2M2O2M3N1N3N2M2O2N2M2O2M3N1N3NY1fNbI"}, "label": "the coffee cup closest to the chocolate glazed donut with the spiderweb design"}]} {"id": 568840, "image": "COCO_train2014_000000568840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"coffee cup on left side in the back\"."}], "task": "refering", "answer_template": "The \"coffee cup on left side in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 25, 26, 27, 28, 29, 49, 50, 51, 52, 72, 73, 74, 75, 95, 96, 97, 98], "bbox": [0.112296875, 0.0, 0.30281250000000004, 0.28075], "iscrowd": 0, "area": 13086.11585, "rle": {"size": [480, 640], "counts": "ShQ12:1W>2fA1W>2fA1W>2fA1V>a0L3M3M4L3M3M4L3M3M3M4L3M3M4L3M3M3M4L3M3M4L3M3M3M4L3M3M3M2N2N1O001O00001O001O001O00001O001N101O00001O001O001O00001O0000O100000000000000O100000000000000O100000000000000O1000000O10O01O1O1O1O1O100O1O1O1O1O100O1O1O1O10000O10000O7J5ZOd0ZOc0]Oc0]Oc0]OURa6"}, "label": "coffee cup on left side in the back"}]} {"id": 568840, "image": "COCO_train2014_000000568840.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cup of coffee at the back\"."}], "task": "refering", "answer_template": "The \"a cup of coffee at the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 25, 26, 27, 28, 29, 49, 50, 51, 52, 72, 73, 74, 75, 95, 96, 97, 98], "bbox": [0.112296875, 0.0, 0.30281250000000004, 0.28075], "iscrowd": 0, "area": 13086.11585, "rle": {"size": [480, 640], "counts": "ShQ12:1W>2fA1W>2fA1W>2fA1V>a0L3M3M4L3M3M4L3M3M3M4L3M3M4L3M3M3M4L3M3M4L3M3M3M4L3M3M3M2N2N1O001O00001O001O001O00001O001N101O00001O001O001O00001O0000O100000000000000O100000000000000O100000000000000O1000000O10O01O1O1O1O1O100O1O1O1O1O100O1O1O1O10000O10000O7J5ZOd0ZOc0]Oc0]Oc0]OURa6"}, "label": "a cup of coffee at the back"}]} {"id": 241170, "image": "COCO_train2014_000000241170.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy looking away from camera\"."}], "task": "refering", "answer_template": "The \"boy looking away from camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 207, 208, 209, 229, 230, 231, 232, 233, 252, 253, 254, 255, 256, 274, 275, 276, 277, 278, 279, 296, 297, 298, 299, 300, 301, 319, 320, 321, 322, 323, 324, 343, 344, 345, 346, 347, 365, 366, 367, 368, 369, 388, 389, 390, 391, 392, 411, 412, 413, 414, 415, 434, 435, 436, 437, 457, 458, 459, 460, 479, 480, 481, 482], "bbox": [0.4237581699346405, 0.3970751633986928, 0.9490686274509803, 0.9998692810457518], "iscrowd": 0, "area": 33267.35060000004, "rle": {"size": [612, 612], "counts": "_Rk41Qc04L1O1O2\\Oc0K501O011N2N2O1N1O2O1N2N2O1N2O1N2O1N2O0\\_OYNj?i1R@dNd?]1X@gNg?[1T@jNi?X1U@kNi?V1U@lNj?U1T@nNj?S1T@POj?V2N3N2N1O2N2N1O2N1O2O1N1O2N101N2n@PLl>Q4SAQLk>W4N1O1O1N2O1O01O1O100O1O3N1M4SLUAZ3Z?1O00100O000O10O1oNe@YN[?h1g@UNY?l1i@PNX?Q2j@lMV?V2k@fMV?[2l@`MV?a2l@YMW?h2a03N2N2N2N1O2N2N2N2N2N2N3M2N2N2N3M1O2N1O2N101N1O2N101N101O001O00001O001O001O1O1O010O1O1N101O1O1O001O1O001O1N101O1O001O0000001O00000000mKPBR3o=mLQBT3o=kLRBU3o=hLSBX3m=gLSBZ3m=eLTB[3l=dLUB]3j=bLVB_3j=_LXBa3h=^LYBb3h=\\LXBe3h=ZLYBf3g=YLZBh3e=WL[Bj3k=mKXBS4\\>10000O2O000O10000OO2O1N2O1N2O1O1N2O2O01O01O01O00001O001O00010O001O0000001OWNRA[On>d0SA\\On>c0SA\\Om>c0TA]Ol>c0UA]Oj>b0WA^Oi>c0WA\\Oi>d0WA\\Oi>e0WAZOi>f0WA[Oh>f0XAYOh>g0YAXOh>g0XAYOh>g0XAZOg>f0ZAYOf>g0ZAYOf>h0YAXOg>h0YAXOg>h0ZAXOe>h0[AXOe>h0[AXOe>h0\\AXOc>h0]AXOc>h0^AWOb>h0_AYOa>f0`AYO`>g0`AYO`>g0aAYO^>[OVA=<8^>YOXA?;7]>XOZAa098\\>TO^Ac088Z>SO`Ae068Z>QObAg049Y>nNeAi038X>mNhAj009X>kNjAl0N:W>hNmAn0M9W>eNoAQ1K:V>cNQBS1I:V>aNSBU1G:U?Fl@9T?Gl@9T?Gl@9T?Gm@8S?Hm@8S?Go@8Q?In@7R?Jn@5R?Ln@3R?Mo@2Q?On@1R?0n@OR?2n@MR?3n@MR?4n@KR?6n@JR?5n@KR?6n@IR?7n@HS?8n@GR?9n@FS?:n@ER?;n@DS?n@@S?a0l@^OU?b0l@\\OU?d0k@[OV?e0k@YOV?g0j@XOW?i0i@UOX?k0i@SOX?m0h@ROY?n0h@POZ?P1e@oN\\?Q1e@mN\\?S1d@lN^?T1b@jN`?U1`@jNb?U1_@iNc?W1]@gNe?X1]@eNe?Z1T100O1O100O1O1O10000000000000O100OO100O2N1O1O1N2N3M2N2O1N3M2N2N2O2M2O1N2O2M2N2O[na0"}, "label": "boy looking away from camera"}]} {"id": 241170, "image": "COCO_train2014_000000241170.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"torso of the boy holding kite\"."}], "task": "refering", "answer_template": "The \"torso of the boy holding kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 207, 208, 209, 229, 230, 231, 232, 233, 252, 253, 254, 255, 256, 274, 275, 276, 277, 278, 279, 296, 297, 298, 299, 300, 301, 319, 320, 321, 322, 323, 324, 343, 344, 345, 346, 347, 365, 366, 367, 368, 369, 388, 389, 390, 391, 392, 411, 412, 413, 414, 415, 434, 435, 436, 437, 457, 458, 459, 460, 479, 480, 481, 482], "bbox": [0.4237581699346405, 0.3970751633986928, 0.9490686274509803, 0.9998692810457518], "iscrowd": 0, "area": 33267.35060000004, "rle": {"size": [612, 612], "counts": "_Rk41Qc04L1O1O2\\Oc0K501O011N2N2O1N1O2O1N2N2O1N2O1N2O1N2O0\\_OYNj?i1R@dNd?]1X@gNg?[1T@jNi?X1U@kNi?V1U@lNj?U1T@nNj?S1T@POj?V2N3N2N1O2N2N1O2N1O2O1N1O2N101N2n@PLl>Q4SAQLk>W4N1O1O1N2O1O01O1O100O1O3N1M4SLUAZ3Z?1O00100O000O10O1oNe@YN[?h1g@UNY?l1i@PNX?Q2j@lMV?V2k@fMV?[2l@`MV?a2l@YMW?h2a03N2N2N2N1O2N2N2N2N2N2N3M2N2N2N3M1O2N1O2N101N1O2N101N101O001O00001O001O001O1O1O010O1O1N101O1O1O001O1O001O1N101O1O001O0000001O00000000mKPBR3o=mLQBT3o=kLRBU3o=hLSBX3m=gLSBZ3m=eLTB[3l=dLUB]3j=bLVB_3j=_LXBa3h=^LYBb3h=\\LXBe3h=ZLYBf3g=YLZBh3e=WL[Bj3k=mKXBS4\\>10000O2O000O10000OO2O1N2O1N2O1O1N2O2O01O01O01O00001O001O00010O001O0000001OWNRA[On>d0SA\\On>c0SA\\Om>c0TA]Ol>c0UA]Oj>b0WA^Oi>c0WA\\Oi>d0WA\\Oi>e0WAZOi>f0WA[Oh>f0XAYOh>g0YAXOh>g0XAYOh>g0XAZOg>f0ZAYOf>g0ZAYOf>h0YAXOg>h0YAXOg>h0ZAXOe>h0[AXOe>h0[AXOe>h0\\AXOc>h0]AXOc>h0^AWOb>h0_AYOa>f0`AYO`>g0`AYO`>g0aAYO^>[OVA=<8^>YOXA?;7]>XOZAa098\\>TO^Ac088Z>SO`Ae068Z>QObAg049Y>nNeAi038X>mNhAj009X>kNjAl0N:W>hNmAn0M9W>eNoAQ1K:V>cNQBS1I:V>aNSBU1G:U?Fl@9T?Gl@9T?Gl@9T?Gm@8S?Hm@8S?Go@8Q?In@7R?Jn@5R?Ln@3R?Mo@2Q?On@1R?0n@OR?2n@MR?3n@MR?4n@KR?6n@JR?5n@KR?6n@IR?7n@HS?8n@GR?9n@FS?:n@ER?;n@DS?n@@S?a0l@^OU?b0l@\\OU?d0k@[OV?e0k@YOV?g0j@XOW?i0i@UOX?k0i@SOX?m0h@ROY?n0h@POZ?P1e@oN\\?Q1e@mN\\?S1d@lN^?T1b@jN`?U1`@jNb?U1_@iNc?W1]@gNe?X1]@eNe?Z1T100O1O100O1O1O10000000000000O100OO100O2N1O1O1N2N3M2N2O1N3M2N2N2O2M2O1N2O2M2N2O[na0"}, "label": "torso of the boy holding kite"}]} {"id": 77332, "image": "COCO_train2014_000000077332.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a female in a long blue jacket standing next to the table\"."}], "task": "refering", "answer_template": "The \"a female in a long blue jacket standing next to the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 101, 102, 123, 124, 125, 126, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 192, 193, 194, 195, 196, 215, 216, 217, 218, 239, 240, 241, 242, 262, 263, 264, 265, 285, 286, 287, 308, 309, 310, 331, 332, 333], "bbox": [0.3565, 0.22420833333333334, 0.5566093750000001, 0.8834166666666667], "iscrowd": 0, "area": 24571.9902, "rle": {"size": [480, 640], "counts": "gQ[3U2j<1N3N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1N3N1FQ1nNT1kNV1ROm0UOk0VOj0UO>C000O1001O001O001O1O001O001O1O001O001O1O1O1N3N1O101O000O2O00001O000O2O00001OTJWGa5h8^JZGc5e8\\J[Ge5e8ZJ\\Gf5k0YJd6OcHi5=cJP7CdHj51oJZ7VOfHP7Y7oHiHR7V7mHkHS7a70O2O001O00001O001O0O101O001O00002SJfHP4[7nKjHm3Y7PLlHl3U7RLPIj3R7ULQIg3P7XLTId3n6[LUIa3l6_LWI]3k6aLYIX3k6hLXIl2R7SMQIb2W7^MlHV2^7hMfHl1c7TN`Ha1h7_N[HU1o7iNUHk0T8UOoG?[8@hG5`8JdGJf85]G_Ol8a0WGSOS9k0b201O001N101O0O2O001N101L3K6J5J^mT4"}, "label": "a female in a long blue jacket standing next to the table"}]} {"id": 77332, "image": "COCO_train2014_000000077332.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a blue coat holding a camera\"."}], "task": "refering", "answer_template": "The \"a woman wearing a blue coat holding a camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 101, 102, 123, 124, 125, 126, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 192, 193, 194, 195, 196, 215, 216, 217, 218, 239, 240, 241, 242, 262, 263, 264, 265, 285, 286, 287, 308, 309, 310, 331, 332, 333], "bbox": [0.3565, 0.22420833333333334, 0.5566093750000001, 0.8834166666666667], "iscrowd": 0, "area": 24571.9902, "rle": {"size": [480, 640], "counts": "gQ[3U2j<1N3N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1N3N1FQ1nNT1kNV1ROm0UOk0VOj0UO>C000O1001O001O001O1O001O001O1O001O001O1O1O1N3N1O101O000O2O00001O000O2O00001OTJWGa5h8^JZGc5e8\\J[Ge5e8ZJ\\Gf5k0YJd6OcHi5=cJP7CdHj51oJZ7VOfHP7Y7oHiHR7V7mHkHS7a70O2O001O00001O001O0O101O001O00002SJfHP4[7nKjHm3Y7PLlHl3U7RLPIj3R7ULQIg3P7XLTId3n6[LUIa3l6_LWI]3k6aLYIX3k6hLXIl2R7SMQIb2W7^MlHV2^7hMfHl1c7TN`Ha1h7_N[HU1o7iNUHk0T8UOoG?[8@hG5`8JdGJf85]G_Ol8a0WGSOS9k0b201O001N101O0O2O001N101L3K6J5J^mT4"}, "label": "a woman wearing a blue coat holding a camera"}]} {"id": 77332, "image": "COCO_train2014_000000077332.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a button up shirt and jeans\"."}], "task": "refering", "answer_template": "The \"a man in a button up shirt and jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 72, 73, 94, 95, 96, 117, 118, 119, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236, 255, 256, 257, 258, 259, 279, 280, 281, 282, 302, 303, 304, 305, 306, 325, 326, 327, 328, 329, 348, 349, 350, 351, 352, 371, 372, 373, 374, 375], "bbox": [0.067265625, 0.15695833333333334, 0.32959375, 1.0], "iscrowd": 0, "area": 35435.14655, "rle": {"size": [480, 640], "counts": "W[d03k>7J7I7I6J7I6J7I7I6I8I6J7I7I6J6J2N3M2M4M2N3M3M2N3M2N3M3iNXLPGk3m8W1M2M4M2M4M3L3N3L3N3L4M2M4M2M4M2M4M3L8I8G8I7H8I8H7I7Jm0ROk0UO4L3M4M3L3M4L3M4L4M2M4L3M4L3M4M2M1O1O1O1O12N2N2N2N2N2N2N2N2PFUMg7m2THXMj7j2PH]Mm7e2nG`MP8b2jGdMT8^2fGhMX8Z2cGlMZ8V2`GPN^8Q2^GTN`8n1ZGYNn5_NmKZ3oM]NP6aNkKT3PN`NQ6dNiKn2PNdNU6dNeKV5Z4nJ`KT5a4oJYKS5g4QKSKQ5n4RKlJQ5S5SKgJo4Z5SKaJo4`5TKZJn4f5VKTJm4l5VKnIl4R6XKhIj4Y6o1001O001O001O001O10O01O001O0011N2O2M2O2M2O2N1N3N1L5K5K5K5K5K5K5J6K5K5K5K4L5K4L5K4L5K4L5K4L5K4L5J5L5K4L5K:F9G9G9G9G9GWdX6"}, "label": "a man in a button up shirt and jeans"}]} {"id": 77332, "image": "COCO_train2014_000000077332.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older man wearing jeans , a black belt and a striped shirt\"."}], "task": "refering", "answer_template": "The \"an older man wearing jeans , a black belt and a striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 72, 73, 94, 95, 96, 117, 118, 119, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 185, 186, 187, 188, 189, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236, 255, 256, 257, 258, 259, 279, 280, 281, 282, 302, 303, 304, 305, 306, 325, 326, 327, 328, 329, 348, 349, 350, 351, 352, 371, 372, 373, 374, 375], "bbox": [0.067265625, 0.15695833333333334, 0.32959375, 1.0], "iscrowd": 0, "area": 35435.14655, "rle": {"size": [480, 640], "counts": "W[d03k>7J7I7I6J7I6J7I7I6I8I6J7I7I6J6J2N3M2M4M2N3M3M2N3M2N3M3iNXLPGk3m8W1M2M4M2M4M3L3N3L3N3L4M2M4M2M4M2M4M3L8I8G8I7H8I8H7I7Jm0ROk0UO4L3M4M3L3M4L3M4L4M2M4L3M4L3M4M2M1O1O1O1O12N2N2N2N2N2N2N2N2PFUMg7m2THXMj7j2PH]Mm7e2nG`MP8b2jGdMT8^2fGhMX8Z2cGlMZ8V2`GPN^8Q2^GTN`8n1ZGYNn5_NmKZ3oM]NP6aNkKT3PN`NQ6dNiKn2PNdNU6dNeKV5Z4nJ`KT5a4oJYKS5g4QKSKQ5n4RKlJQ5S5SKgJo4Z5SKaJo4`5TKZJn4f5VKTJm4l5VKnIl4R6XKhIj4Y6o1001O001O001O001O10O01O001O0011N2O2M2O2M2O2N1N3N1L5K5K5K5K5K5K5J6K5K5K5K4L5K4L5K4L5K4L5K4L5K4L5J5L5K4L5K:F9G9G9G9G9GWdX6"}, "label": "an older man wearing jeans , a black belt and a striped shirt"}]} {"id": 77332, "image": "COCO_train2014_000000077332.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall wooden chair with a green cushion\"."}], "task": "refering", "answer_template": "The \"a tall wooden chair with a green cushion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 195, 196, 218, 219, 220, 241, 242, 243, 264, 265, 266, 267, 287, 288, 289, 290, 310, 311, 312, 313, 334, 335], "bbox": [0.49907812500000004, 0.43945833333333334, 0.63653125, 0.8565], "iscrowd": 0, "area": 7939.1381999999985, "rle": {"size": [480, 640], "counts": "Qhe41j> in this image.", "objects": [{"patches": [172, 195, 196, 218, 219, 220, 241, 242, 243, 264, 265, 266, 267, 287, 288, 289, 290, 310, 311, 312, 313, 334, 335], "bbox": [0.49907812500000004, 0.43945833333333334, 0.63653125, 0.8565], "iscrowd": 0, "area": 7939.1381999999985, "rle": {"size": [480, 640], "counts": "Qhe41j> in this image.", "objects": [{"patches": [41, 42, 43, 44, 62, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135], "bbox": [0.70978125, 0.17976284584980237, 0.9700625, 0.6269565217391304], "iscrowd": 0, "area": 15182.951200000003, "rle": {"size": [253, 640], "counts": "_W`37f7f0ZOg0XOa0@1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1OH7N3000O1000000O10000O1000000O100000?@01000O1000O10O100000O010000O10O1000O10O1000O10O10O10O1000O10O1000O10O1000001O000000000O1000000000000000000000000O1000000002N1O2N1O2N1O2N1O2N1O0000000000O100000000O10001O0O1BbK]M^4c2eKZM[4f2hKWMX4i2lKSMT4m2>O1000000000000000000000000O10000000000000000000000O9H9G9G9G9G8H9G9G9G9GPb4"}, "label": "a bus was going on the road"}]} {"id": 396825, "image": "COCO_train2014_000000396825.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bus with the number 3501\"."}], "task": "refering", "answer_template": "The \"bus with the number 3501\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 43, 44, 62, 63, 64, 65, 66, 67, 68, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135], "bbox": [0.70978125, 0.17976284584980237, 0.9700625, 0.6269565217391304], "iscrowd": 0, "area": 15182.951200000003, "rle": {"size": [253, 640], "counts": "_W`37f7f0ZOg0XOa0@1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1OH7N3000O1000000O10000O1000000O100000?@01000O1000O10O100000O010000O10O1000O10O1000O10O10O10O1000O10O1000O10O1000001O000000000O1000000000000000000000000O1000000002N1O2N1O2N1O2N1O2N1O0000000000O100000000O10001O0O1BbK]M^4c2eKZM[4f2hKWMX4i2lKSMT4m2>O1000000000000000000000000O10000000000000000000000O9H9G9G9G9G8H9G9G9G9GPb4"}, "label": "bus with the number 3501"}]} {"id": 396825, "image": "COCO_train2014_000000396825.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front of a beat - up red and yellow bus\"."}], "task": "refering", "answer_template": "The \"the front of a beat - up red and yellow bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 23, 24, 25, 26, 27, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 138, 139], "bbox": [0.016875, 0.03371541501976284, 0.208765625, 0.7258498023715415], "iscrowd": 0, "area": 17125.34385, "rle": {"size": [253, 640], "counts": "Xk2l0j5X1hNX1hNX1D=N1O001O1O1O1O1O1O1O1O001O001O000000001O000000001O0000001OO100O100O100O100O100O1O1N2O1O1O1N2O1O1O100O10000000000000000000000O1000000000000O1000000O1O1N2N2O1N2O1O2N1O1O1O1O1O10000O104L3L4M3M4K4M3L4M4L3L4M3M4K4M3M3L4M4L3L4M4ZOSJSOR6g0RJVOR6e0QJXOU6b0oIZOW6`0h0I7J5K[ml3"}, "label": "the front of a beat - up red and yellow bus"}]} {"id": 396825, "image": "COCO_train2014_000000396825.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red front of a bus\"."}], "task": "refering", "answer_template": "The \"the red front of a bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 23, 24, 25, 26, 27, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 92, 93, 94, 95, 96, 115, 116, 117, 118, 138, 139], "bbox": [0.016875, 0.03371541501976284, 0.208765625, 0.7258498023715415], "iscrowd": 0, "area": 17125.34385, "rle": {"size": [253, 640], "counts": "Xk2l0j5X1hNX1hNX1D=N1O001O1O1O1O1O1O1O1O001O001O000000001O000000001O0000001OO100O100O100O100O100O1O1N2O1O1O1N2O1O1O100O10000000000000000000000O1000000000000O1000000O1O1N2N2O1N2O1O2N1O1O1O1O1O10000O104L3L4M3M4K4M3L4M4L3L4M3M4K4M3M3L4M4L3L4M4ZOSJSOR6g0RJVOR6e0QJXOU6b0oIZOW6`0h0I7J5K[ml3"}, "label": "the red front of a bus"}]} {"id": 396825, "image": "COCO_train2014_000000396825.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the swaybacked cow in the left image that is in front of the smaller cow\"."}], "task": "refering", "answer_template": "The \"the swaybacked cow in the left image that is in front of the smaller cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 98, 99, 100, 101, 102, 122, 123, 124, 125, 145, 146, 147], "bbox": [0.295484375, 0.4234387351778656, 0.466546875, 0.7221739130434782], "iscrowd": 0, "area": 4216.742200000001, "rle": {"size": [253, 640], "counts": "Qj^18a79J5K5K4L4dISOa5U1TJPOk5a1OO21N3MBVJmNh5m0cJPO[5o0iJoNV5P1mJPOP5P1SKnNm4o0WKoNj4R1VKmNj4S1VKlNl4Q1WKmNk4Q1Q100100O0010O01O010O1O01O001O01O000000000000O1000000000O011O4L4Ld0\\O4L1N101N\\OgJeNZ5S1RKiNm4U1WKjNi4V1XKhNi4X1m0f0[O4M1N3NUO]JZOc5g0dJQO\\5P1jJiNW5W1jJgNW5X1kJfNW5X1jJgNW5X1iJgNZ5W1g0O2M101N2O1N1O2N100O101O0O100010O001OO2N100100O1O1O1N2O1N101NUOVIe0j6ZOYId0g6[O\\I1K;i6ChI8X6IkI4U6JTUd2"}, "label": "the swaybacked cow in the left image that is in front of the smaller cow"}]} {"id": 396825, "image": "COCO_train2014_000000396825.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white ox standing in front of other ox and a bus\"."}], "task": "refering", "answer_template": "The \"a white ox standing in front of other ox and a bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 98, 99, 100, 101, 102, 122, 123, 124, 125, 145, 146, 147], "bbox": [0.295484375, 0.4234387351778656, 0.466546875, 0.7221739130434782], "iscrowd": 0, "area": 4216.742200000001, "rle": {"size": [253, 640], "counts": "Qj^18a79J5K5K4L4dISOa5U1TJPOk5a1OO21N3MBVJmNh5m0cJPO[5o0iJoNV5P1mJPOP5P1SKnNm4o0WKoNj4R1VKmNj4S1VKlNl4Q1WKmNk4Q1Q100100O0010O01O010O1O01O001O01O000000000000O1000000000O011O4L4Ld0\\O4L1N101N\\OgJeNZ5S1RKiNm4U1WKjNi4V1XKhNi4X1m0f0[O4M1N3NUO]JZOc5g0dJQO\\5P1jJiNW5W1jJgNW5X1kJfNW5X1jJgNW5X1iJgNZ5W1g0O2M101N2O1N1O2N100O101O0O100010O001OO2N100100O1O1O1N2O1N101NUOVIe0j6ZOYId0g6[O\\I1K;i6ChI8X6IkI4U6JTUd2"}, "label": "a white ox standing in front of other ox and a bus"}]} {"id": 85549, "image": "COCO_train2014_000000085549.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black suitcase next to a woman on her phone\"."}], "task": "refering", "answer_template": "The \"a black suitcase next to a woman on her phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [269, 270, 271, 285, 286, 287, 288, 301, 302, 303, 304, 305, 317, 318, 319, 320, 321, 322, 333, 334, 335, 336, 337, 338, 339, 350, 351, 352, 353, 354, 368, 369, 370, 385, 386], "bbox": [0.6232377049180328, 0.65540625, 0.9954303278688524, 0.981984375], "iscrowd": 0, "area": 19970.645900000007, "rle": {"size": [640, 488], "counts": "ean52kc04M4K5K4L5K5K4L5K5K4L5K5K4L5K5L3L5M3M2N3N2M2N3M3M2O2M3M2N3M3N1N3M3M2N3N1N3M1O01O0001O0001O0001O0001O01O000001O01O0000010O000000010O00000010O000000010O000001O01O000001O01O0001O0001O0001O0001O01O000001O01O0000010O000001O01O0000010O000000010O00000010O000000010O000001O01O000001O05K2N101N2N1O2N2O0O2N2N101N2N1O2N2O0O2N2N101N2N1O2N2O0O2N2N101N2N1O2N2O0O2N2N101N2N1K6_N]k0"}, "label": "a black suitcase next to a woman on her phone"}]} {"id": 85549, "image": "COCO_train2014_000000085549.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black suit case\"."}], "task": "refering", "answer_template": "The \"black suit case\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [269, 270, 271, 285, 286, 287, 288, 301, 302, 303, 304, 305, 317, 318, 319, 320, 321, 322, 333, 334, 335, 336, 337, 338, 339, 350, 351, 352, 353, 354, 368, 369, 370, 385, 386], "bbox": [0.6232377049180328, 0.65540625, 0.9954303278688524, 0.981984375], "iscrowd": 0, "area": 19970.645900000007, "rle": {"size": [640, 488], "counts": "ean52kc04M4K5K4L5K5K4L5K5K4L5K5K4L5K5L3L5M3M2N3N2M2N3M3M2O2M3M2N3M3N1N3M3M2N3N1N3M1O01O0001O0001O0001O0001O01O000001O01O0000010O000000010O00000010O000000010O000001O01O000001O01O0001O0001O0001O0001O01O000001O01O0000010O000001O01O0000010O000000010O00000010O000000010O000001O01O000001O05K2N101N2N1O2N2O0O2N2N101N2N1O2N2O0O2N2N101N2N1O2N2O0O2N2N101N2N1O2N2O0O2N2N101N2N1K6_N]k0"}, "label": "black suit case"}]} {"id": 290354, "image": "COCO_train2014_000000290354.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plaid umbrella held by woman in black boots\"."}], "task": "refering", "answer_template": "The \"plaid umbrella held by woman in black boots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 28, 29, 30, 31, 32, 33, 34, 35, 51, 52, 53, 54, 55, 56, 57, 58, 78, 101, 124, 147], "bbox": [0.221, 0.0040705882352941175, 0.55478125, 0.41070588235294114], "iscrowd": 0, "area": 9037.10905, "rle": {"size": [425, 640], "counts": "ioj11W=1O1O1O1O1O001O1O1O1O100O1O1O1O1O1O1O2N1O100O1O1O1O2N1O1O1O2O0O1O2O0O101N100O100O1O100O100O010O100O1O100O100O100O100O100O10000O1000000O10000O1000000O1000000O10000O1000000O2O00000O10000O1000000O101O000O10000000000O10000000000000007Ii0WOPOiEF5^NI in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 28, 29, 30, 31, 32, 33, 34, 35, 51, 52, 53, 54, 55, 56, 57, 58, 78, 101, 124, 147], "bbox": [0.221, 0.0040705882352941175, 0.55478125, 0.41070588235294114], "iscrowd": 0, "area": 9037.10905, "rle": {"size": [425, 640], "counts": "ioj11W=1O1O1O1O1O001O1O1O1O100O1O1O1O1O1O1O2N1O100O1O1O1O2N1O1O1O2O0O1O2O0O101N100O100O1O100O100O010O100O1O100O100O100O100O100O10000O1000000O10000O1000000O1000000O10000O1000000O2O00000O10000O1000000O101O000O10000000000O10000000000000007Ii0WOPOiEF5^NI in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 61, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87], "bbox": [0.648546875, 0.0, 1.0, 0.2164871194379391], "iscrowd": 0, "area": 19140.521249999998, "rle": {"size": [427, 640], "counts": "eQ]5b0i in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 19, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 61, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87], "bbox": [0.648546875, 0.0, 1.0, 0.2164871194379391], "iscrowd": 0, "area": 19140.521249999998, "rle": {"size": [427, 640], "counts": "eQ]5b0i in this image.", "objects": [{"patches": [19, 20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275, 296, 297, 298, 320, 321], "bbox": [0.7672656250000001, 0.0, 0.9995156250000001, 0.7262213740458016], "iscrowd": 0, "area": 39477.099600000016, "rle": {"size": [524, 640], "counts": "U^k71Y`05K4M3L4M3L5L3L4M3L4M4K4L4M3L4M4K1O0N3L4M2M4L3N3L4M2M4YDXMh9k2VFYMd9j2\\FZM]9j2aF[MY9h2eF\\MU9g2kF]Mo8f2oF_Mj8d2VG_Me8d2YGaMa8b2]GcM]8_2cGeMX8]2fGgMU8\\2jGhMP8Z2oGkMk7X2THkMh7W2VHnMd7T2[HQN_7R2ZHWN`7l1THdNe7`1nGPOj7m4J7H8I7I6O2M2O1N101N100O2O0O2O0O2O0O101N101O0O3N2M3N2M2O2M3O1N2N2O1N2N2N2O1N2N01O01O00010O0001O01O00010O0001O01O00010O00001N1O1O1O2N1O1O2N1O1O1O2N1O101O000000100O001O001O001O1O0010O01O001O001O1O001O010R7mHjH"}, "label": "grey / black bag touching striped cat"}]} {"id": 527925, "image": "COCO_train2014_000000527925.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black bag next to the lighter cat\"."}], "task": "refering", "answer_template": "The \"the black bag next to the lighter cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [19, 20, 21, 22, 42, 43, 44, 45, 65, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275, 296, 297, 298, 320, 321], "bbox": [0.7672656250000001, 0.0, 0.9995156250000001, 0.7262213740458016], "iscrowd": 0, "area": 39477.099600000016, "rle": {"size": [524, 640], "counts": "U^k71Y`05K4M3L4M3L5L3L4M3L4M4K4L4M3L4M4K1O0N3L4M2M4L3N3L4M2M4YDXMh9k2VFYMd9j2\\FZM]9j2aF[MY9h2eF\\MU9g2kF]Mo8f2oF_Mj8d2VG_Me8d2YGaMa8b2]GcM]8_2cGeMX8]2fGgMU8\\2jGhMP8Z2oGkMk7X2THkMh7W2VHnMd7T2[HQN_7R2ZHWN`7l1THdNe7`1nGPOj7m4J7H8I7I6O2M2O1N101N100O2O0O2O0O2O0O101N101O0O3N2M3N2M2O2M3O1N2N2O1N2N2N2O1N2N01O01O00010O0001O01O00010O0001O01O00010O00001N1O1O1O2N1O1O2N1O1O1O2N1O101O000000100O001O001O001O1O0010O01O001O001O1O001O010R7mHjH"}, "label": "the black bag next to the lighter cat"}]} {"id": 527925, "image": "COCO_train2014_000000527925.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back pack on left side\"."}], "task": "refering", "answer_template": "The \"the back pack on left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 92, 93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 161, 162, 163, 184, 185, 207], "bbox": [0.0, 0.0, 0.23332812500000003, 0.4911832061068702], "iscrowd": 0, "area": 23694.1427, "rle": {"size": [524, 640], "counts": "0o7]8000000001O00000000001O0001O00000O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2J5K5K5K5K5M3M3N2N2N2M3N2N2N2N2M3N2N2N2M3N2N2N2M3N2N2N2N3L3N2N2N2M3N2N2M3M3M3N2M3M3M3M3M3M3M3M3N2M3M3O1001O000010O0001O0000jNhA9W>@QB?o=[OYBc0g=VOaBj0a>O2N1O2O0O2N101N1O101N1K6K]Vk7"}, "label": "the back pack on left side"}]} {"id": 527925, "image": "COCO_train2014_000000527925.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue backpack near the black cat\"."}], "task": "refering", "answer_template": "The \"blue backpack near the black cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 92, 93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 161, 162, 163, 184, 185, 207], "bbox": [0.0, 0.0, 0.23332812500000003, 0.4911832061068702], "iscrowd": 0, "area": 23694.1427, "rle": {"size": [524, 640], "counts": "0o7]8000000001O00000000001O0001O00000O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2J5K5K5K5K5M3M3N2N2N2M3N2N2N2N2M3N2N2N2M3N2N2N2M3N2N2N2N3L3N2N2N2M3N2N2M3M3M3N2M3M3M3M3M3M3M3M3N2M3M3O1001O000010O0001O0000jNhA9W>@QB?o=[OYBc0g=VOaBj0a>O2N1O2O0O2N101N1O101N1K6K]Vk7"}, "label": "blue backpack near the black cat"}]} {"id": 323133, "image": "COCO_train2014_000000323133.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed and pillow next to the window on the right\"."}], "task": "refering", "answer_template": "The \"the bed and pillow next to the window on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 296, 297, 298, 320, 321], "bbox": [0.769265625, 0.5398125, 0.9973281249999999, 0.7839375000000001], "iscrowd": 0, "area": 10446.758200000002, "rle": {"size": [480, 640], "counts": "UmV7n0R>O1O010O1O1O10O01O100O1O00100O1000O1000O1000O1000O100000O10O100000O010000000O010000000O011BZO`Bh0]=ZObBg0\\=ZOdBg0[=ZOcBh0[=YOeBi0X=[OeBf0Y=_OcBb0\\=c0M2N2O1O2M3N1O2M3N1O2M3N2N1N3N1O1N101O001O001O00001O001O001O001O0000001O00001WDUM[;k2cDXM\\;h2cDZM\\;g2aD\\M^;d2`D^M`;b2_D`M`;`2^DcMa;o2O001O0000001O00001O00001O0000001O00001O00000000000000O100000000000000O10000000000000000O102Ab0[NVg0"}, "label": "the bed and pillow next to the window on the right"}]} {"id": 323133, "image": "COCO_train2014_000000323133.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray ottoman\"."}], "task": "refering", "answer_template": "The \"a gray ottoman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 296, 297, 298, 320, 321], "bbox": [0.769265625, 0.5398125, 0.9973281249999999, 0.7839375000000001], "iscrowd": 0, "area": 10446.758200000002, "rle": {"size": [480, 640], "counts": "UmV7n0R>O1O010O1O1O10O01O100O1O00100O1000O1000O1000O1000O100000O10O100000O010000000O010000000O011BZO`Bh0]=ZObBg0\\=ZOdBg0[=ZOcBh0[=YOeBi0X=[OeBf0Y=_OcBb0\\=c0M2N2O1O2M3N1O2M3N1O2M3N2N1N3N1O1N101O001O001O00001O001O001O001O0000001O00001WDUM[;k2cDXM\\;h2cDZM\\;g2aD\\M^;d2`D^M`;b2_D`M`;`2^DcMa;o2O001O0000001O00001O00001O0000001O00001O00000000000000O100000000000000O10000000000000000O102Ab0[NVg0"}, "label": "a gray ottoman"}]} {"id": 323133, "image": "COCO_train2014_000000323133.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch with the cat on it\"."}], "task": "refering", "answer_template": "The \"the couch with the cat on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.080640625, 0.48870833333333336, 0.783875, 0.9919375], "iscrowd": 0, "area": 83880.4424, "rle": {"size": [480, 640], "counts": "Vdh0^4b:2N2N2N2N2O1N2N2O1N2N2O1N2O1N3M2O1N2N2O1N2N2L4K5L4K5L4K5K5L4K6K4K5L2M3000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O00000000O1000001O0000000000000O1000000000000000000O10000000000000000001O4L5K3M000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000001O0000000000000000000000000000000002N2N3M2N2N3M2N3M2N3M2N2N3M2N3M2N2N3M2N3M2N2N0000O10000O100O100O10000O100O100O100O10000O10000001O001O001O001O001O1O1O1O1O1O001O1O1O1O1O1O1O004L5K0000000000O100000001O0000000O100000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O100000000000000O100000000001O001O001O001O00001O1O2N1O2N2N1O2N2N1O2NO10000000000000001O0O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O11O0000001O0000001O00001O0000001O00001O0000001O0000001Oa[P2"}, "label": "the couch with the cat on it"}]} {"id": 323133, "image": "COCO_train2014_000000323133.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black sofa with the cat on it\"."}], "task": "refering", "answer_template": "The \"the black sofa with the cat on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.080640625, 0.48870833333333336, 0.783875, 0.9919375], "iscrowd": 0, "area": 83880.4424, "rle": {"size": [480, 640], "counts": "Vdh0^4b:2N2N2N2N2O1N2N2O1N2N2O1N2O1N3M2O1N2N2O1N2N2L4K5L4K5L4K5K5L4K6K4K5L2M3000000000000001O000000000000001O000000000000001O000000000000001O000000000000001O00000000O1000001O0000000000000O1000000000000000000O10000000000000000001O4L5K3M000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000001O0000000000000000000000000000000002N2N3M2N2N3M2N3M2N3M2N2N3M2N3M2N2N3M2N3M2N2N0000O10000O100O100O10000O100O100O100O10000O10000001O001O001O001O001O1O1O1O1O1O001O1O1O1O1O1O1O004L5K0000000000O100000001O0000000O100000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O100000000000000O100000000001O001O001O001O00001O1O2N1O2N2N1O2N2N1O2NO10000000000000001O0O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O11O0000001O0000001O00001O0000001O00001O0000001O0000001Oa[P2"}, "label": "the black sofa with the cat on it"}]} {"id": 233022, "image": "COCO_train2014_000000233022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra standing behind the side of a smaller zebra\"."}], "task": "refering", "answer_template": "The \"a zebra standing behind the side of a smaller zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 76, 77, 78, 79, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 159, 160, 165, 166, 167, 168, 169, 170, 171, 179, 180, 181, 182, 183, 188, 189, 190, 191, 192, 193, 202, 203, 204, 205, 206, 222, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.19553125, 0.17695550351288056, 1.0, 0.98903981264637], "iscrowd": 0, "area": 64272.36094999998, "rle": {"size": [427, 640], "counts": "jZd14U=;E7I2N3M2N2N2N2N3M2N2O1O1O2N1O1O1O1O2N1O100O1O2N1O1O1O2N1O1O1O1O2N1O1O100O2N1O1O1O1O2N001O1O1O1O1N2O1O1O1O1O001O1O1O1O1O1O1O1O1O001O100O001O1O010O1O00100VNhL`IX3b6iL[IX3e6jLXIW3i6jLTIV3n6kLPIU3P7mLmHT3T7oLgHQ3[7RM`Ho2`7TM\\Hm2e7VMWHi2k7ZMPHg2P8]MlGc2U8_MhG`2Z8bMcG^2V8jMiGV2T8oMjGQ2W8PNfGP2\\8PNcGP2]8_11N2O1O0O2O1O1N101O0O101O0O10000O11O1O001O1O001O1O1O2N1O1O2N1O1O000000000O10000000O100O100O100O1O100O2N1O2M2O2N1O2N1O2N1O2N1O2N1N3N1O1O2N1O2N1O2N1O2N1N3N2N2N2N2N2N2N2N2O1N3M2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2O1N2N2N3M2N2N2N2N2O1N2N1O2N2O1N2N2O1N1OlJ[OeMc0Z2G_M8_2M_M1`21aMM^25cMJ[28gMEX2=iMAV2a0kM]OT2f0lMYOR2i0PNTOP2m0QNQOn1Q1SNmNl1U1TNkNj1W1WNgNh1[1R52N2N2N2N1O2N2N2N2N2O1N1O2N2N2O1O1O1O001N2O1O1O1O1O0O2O1O1O1O1O0O2O1O1O1O1O1O0O101O00001O0O101O1O1O2N1N2O2N1O1N3M2O1N3M2O1N2N3N1N2N3N1N2O1N2O11O0001O0001O0000001O0001O0001O0000001O01O000001O0000001O01O000001O00000010O000001O000000001O0000001O00001O001O00001O001O00001O001O00001O001O00001O001O0O[MWLkKh3U4YLjKg3V4ZLiKf3W4[LiKd3W4]LhKc3W4^LjKa3V4`LiK`3W4aLhK_3X4bLhK]3X4dLgK\\3Y4eLgKZ3Y4gLfKY3Y4iLfKW3Z4iLgKV3Y4kLfKU3Z4lLfKS3Z4nLeKR3[4oLeKP3[4QMdKo2\\4RMcKn2\\4TMdKk2\\4UMdKk2\\4VMdKi2\\4XMcKh2]4ZMbKe2^4]M`Kc2`4^M_Kb2`4aM_K^2a4dM]K\\2c4eM]KZ2c4hM[KX2e4jMZKU2f4lMYKT2g4nMWKQ2j4QNUKn1j4TNUKl1k4VNQKl1o4VNlJm1T5TNgJo1Z5SNaJo1`5SN[JP2e5RNVJP2k5QNPJR2P6QNkIQ2V6QNdIS2\\6nM_IU2b6mMXIV2i6V20O1O00100O1O100O1O1O010O00100O001O010O0010O01O010O001O010O0010O01O1O010O0010O2O`0_O=D0O100O2O0O10001O01O0001O0000001O00000010O000001O000000M3J6L5L3L4L4M3L4L4N2N5L4K4L5K5J5L5K5H8CbI"}, "label": "a zebra standing behind the side of a smaller zebra"}]} {"id": 233022, "image": "COCO_train2014_000000233022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra looking to the left\"."}], "task": "refering", "answer_template": "The \"a zebra looking to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 76, 77, 78, 79, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 159, 160, 165, 166, 167, 168, 169, 170, 171, 179, 180, 181, 182, 183, 188, 189, 190, 191, 192, 193, 202, 203, 204, 205, 206, 222, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.19553125, 0.17695550351288056, 1.0, 0.98903981264637], "iscrowd": 0, "area": 64272.36094999998, "rle": {"size": [427, 640], "counts": "jZd14U=;E7I2N3M2N2N2N2N3M2N2O1O1O2N1O1O1O1O2N1O100O1O2N1O1O1O2N1O1O1O1O2N1O1O100O2N1O1O1O1O2N001O1O1O1O1N2O1O1O1O1O001O1O1O1O1O1O1O1O1O001O100O001O1O010O1O00100VNhL`IX3b6iL[IX3e6jLXIW3i6jLTIV3n6kLPIU3P7mLmHT3T7oLgHQ3[7RM`Ho2`7TM\\Hm2e7VMWHi2k7ZMPHg2P8]MlGc2U8_MhG`2Z8bMcG^2V8jMiGV2T8oMjGQ2W8PNfGP2\\8PNcGP2]8_11N2O1O0O2O1O1N101O0O101O0O10000O11O1O001O1O001O1O1O2N1O1O2N1O1O000000000O10000000O100O100O100O1O100O2N1O2M2O2N1O2N1O2N1O2N1O2N1N3N1O1O2N1O2N1O2N1O2N1N3N2N2N2N2N2N2N2N2O1N3M2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2O1N2N2N3M2N2N2N2N2O1N2N1O2N2O1N2N2O1N1OlJ[OeMc0Z2G_M8_2M_M1`21aMM^25cMJ[28gMEX2=iMAV2a0kM]OT2f0lMYOR2i0PNTOP2m0QNQOn1Q1SNmNl1U1TNkNj1W1WNgNh1[1R52N2N2N2N1O2N2N2N2N2O1N1O2N2N2O1O1O1O001N2O1O1O1O1O0O2O1O1O1O1O0O2O1O1O1O1O1O0O101O00001O0O101O1O1O2N1N2O2N1O1N3M2O1N3M2O1N2N3N1N2N3N1N2O1N2O11O0001O0001O0000001O0001O0001O0000001O01O000001O0000001O01O000001O00000010O000001O000000001O0000001O00001O001O00001O001O00001O001O00001O001O00001O001O0O[MWLkKh3U4YLjKg3V4ZLiKf3W4[LiKd3W4]LhKc3W4^LjKa3V4`LiK`3W4aLhK_3X4bLhK]3X4dLgK\\3Y4eLgKZ3Y4gLfKY3Y4iLfKW3Z4iLgKV3Y4kLfKU3Z4lLfKS3Z4nLeKR3[4oLeKP3[4QMdKo2\\4RMcKn2\\4TMdKk2\\4UMdKk2\\4VMdKi2\\4XMcKh2]4ZMbKe2^4]M`Kc2`4^M_Kb2`4aM_K^2a4dM]K\\2c4eM]KZ2c4hM[KX2e4jMZKU2f4lMYKT2g4nMWKQ2j4QNUKn1j4TNUKl1k4VNQKl1o4VNlJm1T5TNgJo1Z5SNaJo1`5SN[JP2e5RNVJP2k5QNPJR2P6QNkIQ2V6QNdIS2\\6nM_IU2b6mMXIV2i6V20O1O00100O1O100O1O1O010O00100O001O010O0010O01O010O001O010O0010O01O1O010O0010O2O`0_O=D0O100O2O0O10001O01O0001O0000001O00000010O000001O000000M3J6L5L3L4L4M3L4L4N2N5L4K4L5K5J5L5K5H8CbI"}, "label": "a zebra looking to the left"}]} {"id": 233022, "image": "COCO_train2014_000000233022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra facing right\"."}], "task": "refering", "answer_template": "The \"a zebra facing right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 163, 171, 172, 173, 174, 175, 176, 177, 178, 179, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 247, 248, 249, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.0, 0.2966276346604215, 0.86509375, 0.9887587822014052], "iscrowd": 0, "area": 91925.89129999997, "rle": {"size": [427, 640], "counts": "\\:h2j7h211N100O100O101N100O1O100O2O0O100O01N1O2M2O2M2O1O2M3N2M3N2N1N3N2M3N2001O1N2O1O1O001O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1O0001O00001O0000001O00001O00001O0000001O00001O00001O0000001O000010O0001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O0000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O1O1O1O1OLjHUJU7k5oHRJP7n561O0000O1000000O1001O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O2N2N1O2N2N2N2N1O2N2N100O2N100O1O2O0O1O101N000001O01O01N1O1O2N1O1O2N1O1O1O2N1O101N1O1O2N1O1O2N1O1O2N1O1O2N1O2O0O2N1O2O0O2O0O2O001N10@VJaIh5a6\\JZId5g6^JWIa5j6aJSI^5o6>1O1N2O1N2O1O1N3N1N3N3M2M4M2N3L3N2M3N1O0O2O1O1N101N10O100O1002N1O1N2O2N1O1O1O1O2N100000010O0000010O0001O01O010O1O010O001O010O1O010O1O2O0O2N2O0TLmF^3U9]LnFc3^9O10O01O100O00100O001O100O00100O00100O1O00100O00100O1O100O100O100O100O10O10O2O1N3N1N2O1N2O1N3N000O100O100O100O100O10000O100O100O2O0O101N1O2N1O1N3N1O1N3N1N2K6K4K5M4K5L3L5K4M4K5L3L5K4MTcS1"}, "label": "a zebra facing right"}]} {"id": 233022, "image": "COCO_train2014_000000233022.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front most zebra\"."}], "task": "refering", "answer_template": "The \"the front most zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 163, 171, 172, 173, 174, 175, 176, 177, 178, 179, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 247, 248, 249, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 322, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.0, 0.2966276346604215, 0.86509375, 0.9887587822014052], "iscrowd": 0, "area": 91925.89129999997, "rle": {"size": [427, 640], "counts": "\\:h2j7h211N100O100O101N100O1O100O2O0O100O01N1O2M2O2M2O1O2M3N2M3N2N1N3N2M3N2001O1N2O1O1O001O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1O0001O00001O0000001O00001O00001O0000001O00001O00001O0000001O000010O0001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O0000000000001O000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O1O1O1O1OLjHUJU7k5oHRJP7n561O0000O1000000O1001O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O2N2N1O2N2N2N2N1O2N2N100O2N100O1O2O0O1O101N000001O01O01N1O1O2N1O1O2N1O1O1O2N1O101N1O1O2N1O1O2N1O1O2N1O1O2N1O2O0O2N1O2O0O2O0O2O001N10@VJaIh5a6\\JZId5g6^JWIa5j6aJSI^5o6>1O1N2O1N2O1O1N3N1N3N3M2M4M2N3L3N2M3N1O0O2O1O1N101N10O100O1002N1O1N2O2N1O1O1O1O2N100000010O0000010O0001O01O010O1O010O001O010O1O010O1O2O0O2N2O0TLmF^3U9]LnFc3^9O10O01O100O00100O001O100O00100O00100O1O00100O00100O1O100O100O100O100O10O10O2O1N3N1N2O1N2O1N3N000O100O100O100O100O10000O100O100O2O0O101N1O2N1O1N3N1O1N3N1N2K6K4K5M4K5L3L5K4M4K5L3L5K4MTcS1"}, "label": "the front most zebra"}]} {"id": 159299, "image": "COCO_train2014_000000159299.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"french fries occupy half of a white bowl\"."}], "task": "refering", "answer_template": "The \"french fries occupy half of a white bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 136], "bbox": [0.737453125, 0.017694117647058822, 0.999640625, 0.3356470588235294], "iscrowd": 0, "area": 18209.826449999997, "rle": {"size": [425, 640], "counts": "SnS63b in this image.", "objects": [{"patches": [17, 18, 19, 20, 21, 22, 40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 136], "bbox": [0.737453125, 0.017694117647058822, 0.999640625, 0.3356470588235294], "iscrowd": 0, "area": 18209.826449999997, "rle": {"size": [425, 640], "counts": "SnS63b in this image.", "objects": [{"patches": [197, 198, 199, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 270, 271, 272, 289, 290, 294, 295], "bbox": [0.5363749999999999, 0.5280921052631579, 0.8389843749999999, 0.8179824561403509], "iscrowd": 0, "area": 10506.729800000003, "rle": {"size": [456, 640], "counts": "RQi42S>6J5L5J6J6J6J5K6J6J6K4M4L4M3L010O01O001O0010O01O001O00YElM\\9T2aFoM^9R2_FQNa9P2[FSNd9n1YFUNg9k1VFXNi9i1TF[Nk9f1QF]Nn9d1nE`NR:`1kEcNT:^1iEeNW:\\1eEgNZ:Z1cEgN_:]23L5K4M4K4L5L3010O01O001O010O001O010OXEjM_9V2oE\\NR:l2001O1O1N2O1O2N1N2L4L4M3L5K4M3L4L4M4K4L4M3L4K5H9H7I700000001O0O101O001O001O1O001O1O001O1O001O10O010O01O10O01O10O01O10O0100O0010O01O10O0100O00100O0101O0O101N10001N10001N10000O2O00001O00001O001O010O1O1O1O1O1O010O1O1O1O1O1O01O0O1O1O100O1O1O0001O0000000010O3M6J6K5J5K6J6Jf`]1"}, "label": "a giraffe in the back of a row of three giraffes"}]} {"id": 77380, "image": "COCO_train2014_000000077380.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe with its head the lowest\"."}], "task": "refering", "answer_template": "The \"the giraffe with its head the lowest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 198, 199, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 248, 265, 266, 267, 268, 270, 271, 272, 289, 290, 294, 295], "bbox": [0.5363749999999999, 0.5280921052631579, 0.8389843749999999, 0.8179824561403509], "iscrowd": 0, "area": 10506.729800000003, "rle": {"size": [456, 640], "counts": "RQi42S>6J5L5J6J6J6J5K6J6J6K4M4L4M3L010O01O001O0010O01O001O00YElM\\9T2aFoM^9R2_FQNa9P2[FSNd9n1YFUNg9k1VFXNi9i1TF[Nk9f1QF]Nn9d1nE`NR:`1kEcNT:^1iEeNW:\\1eEgNZ:Z1cEgN_:]23L5K4M4K4L5L3010O01O001O010O001O010OXEjM_9V2oE\\NR:l2001O1O1N2O1O2N1N2L4L4M3L5K4M3L4L4M4K4L4M3L4K5H9H7I700000001O0O101O001O001O1O001O1O001O1O001O10O010O01O10O01O10O01O10O0100O0010O01O10O0100O00100O0101O0O101N10001N10001N10000O2O00001O00001O001O010O1O1O1O1O1O010O1O1O1O1O1O01O0O1O1O100O1O1O0001O0000000010O3M6J6K5J5K6J6Jf`]1"}, "label": "the giraffe with its head the lowest"}]} {"id": 77380, "image": "COCO_train2014_000000077380.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe with its head up\"."}], "task": "refering", "answer_template": "The \"a giraffe with its head up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 57, 58, 59, 79, 80, 81, 82, 101, 102, 103, 123, 124, 125, 145, 146, 147, 148, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262, 280, 281, 283, 284, 303, 304, 306, 307, 326, 327, 329, 330], "bbox": [0.172921875, 0.10111842105263158, 0.59721875, 0.9258333333333334], "iscrowd": 0, "area": 26415.68910000001, "rle": {"size": [456, 640], "counts": "Pja1a0`=7H8I6J7I6I8M2N3M2O2M10N100O01N101N1K6J6J6K5J7RFoL`8W3[GQM_8W3WGRMg8X4O2N1O2N1O101N1O2NO1O1O1O2N1O1L4L4L4L4L5K4L4I7H8H8L401O0O100O10001N10000O10001N10000O11O100O1O1O1O2N1O2O0O2N1N2O00001NK6N21N2O1O1O1N2O1O1O100O100nGfMU4d2SKkMm4g2QJiMn5Y500O10O010M2E in this image.", "objects": [{"patches": [34, 35, 57, 58, 59, 79, 80, 81, 82, 101, 102, 103, 123, 124, 125, 145, 146, 147, 148, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 262, 280, 281, 283, 284, 303, 304, 306, 307, 326, 327, 329, 330], "bbox": [0.172921875, 0.10111842105263158, 0.59721875, 0.9258333333333334], "iscrowd": 0, "area": 26415.68910000001, "rle": {"size": [456, 640], "counts": "Pja1a0`=7H8I6J7I6I8M2N3M2O2M10N100O01N101N1K6J6J6K5J7RFoL`8W3[GQM_8W3WGRMg8X4O2N1O2N1O101N1O2NO1O1O1O2N1O1L4L4L4L4L5K4L4I7H8H8L401O0O100O10001N10000O10001N10000O11O100O1O1O1O2N1O2O0O2N1N2O00001NK6N21N2O1O1O1N2O1O1O100O100nGfMU4d2SKkMm4g2QJiMn5Y500O10O010M2E in this image.", "objects": [{"patches": [149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 178, 193, 194, 195, 196, 197, 200, 201, 216, 217, 218, 219, 238, 239, 240, 241, 242, 261, 262, 263, 264, 265, 284, 285, 286, 287, 288, 306, 307, 308, 309, 310, 311, 331], "bbox": [0.340625, 0.3828508771929825, 0.761921875, 0.9006578947368422], "iscrowd": 0, "area": 18783.914450000004, "rle": {"size": [456, 640], "counts": "c^Q32Q>5J5K5L4K5L4K5L4K6KO0N2O20O31O3L4J6G9H8G^U33cjL=B>B=fCdNT;Z2K5K5K5K5iNnLTGW3l8kLmFV2LUNW9HeFT2c:nMVES2n:oMkDR2a:eMRFf2Q:_MjE_2Y:eMbEZ2_:hM_EW2b:lM\\ER2f:oMXEP2i:RNUEm1l:VNQEi1Q;e01O1O2N1O2N1O1O2N1O1O2N11O010O1N101O1N101O1N101O1N2O1O1O100O1000`0_Om0TOb0^O8G8I8G6K001N10000O100O2OC=[OfGQKd8\\4n0\\Od0F:2iGZLh5h3SJ[Ll5g3PJ\\LP6f3kI]LU6d3iI]LW6e3fI\\LZ6f3cI[L]6f3aI[L_6g3^IZLb6g3\\IZLd6h3YIYLf6i3XIWLi6k3TIULm6m3PISLQ7n3mHSLS7o3jHQLW7Q4fHoK[7T4aHlK`7V4]HkKc7X4YHhKg7[4VHeKk7U1hGD1=9jNo7S1lG_O4c0MlNS8P1QH[O6g0CnNW8n0THWO9l0YOoN[8l0XHROjCVOY<5cCOQ=0PCOR=OoB0R=OPCOR=OoB0S=NnB0T=OnBOd[T2"}, "label": "middle giraffe"}]} {"id": 77380, "image": "COCO_train2014_000000077380.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe in the middle\"."}], "task": "refering", "answer_template": "The \"the giraffe in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 178, 193, 194, 195, 196, 197, 200, 201, 216, 217, 218, 219, 238, 239, 240, 241, 242, 261, 262, 263, 264, 265, 284, 285, 286, 287, 288, 306, 307, 308, 309, 310, 311, 331], "bbox": [0.340625, 0.3828508771929825, 0.761921875, 0.9006578947368422], "iscrowd": 0, "area": 18783.914450000004, "rle": {"size": [456, 640], "counts": "c^Q32Q>5J5K5L4K5L4K5L4K6KO0N2O20O31O3L4J6G9H8G^U33cjL=B>B=fCdNT;Z2K5K5K5K5iNnLTGW3l8kLmFV2LUNW9HeFT2c:nMVES2n:oMkDR2a:eMRFf2Q:_MjE_2Y:eMbEZ2_:hM_EW2b:lM\\ER2f:oMXEP2i:RNUEm1l:VNQEi1Q;e01O1O2N1O2N1O1O2N1O1O2N11O010O1N101O1N101O1N101O1N2O1O1O100O1000`0_Om0TOb0^O8G8I8G6K001N10000O100O2OC=[OfGQKd8\\4n0\\Od0F:2iGZLh5h3SJ[Ll5g3PJ\\LP6f3kI]LU6d3iI]LW6e3fI\\LZ6f3cI[L]6f3aI[L_6g3^IZLb6g3\\IZLd6h3YIYLf6i3XIWLi6k3TIULm6m3PISLQ7n3mHSLS7o3jHQLW7Q4fHoK[7T4aHlK`7V4]HkKc7X4YHhKg7[4VHeKk7U1hGD1=9jNo7S1lG_O4c0MlNS8P1QH[O6g0CnNW8n0THWO9l0YOoN[8l0XHROjCVOY<5cCOQ=0PCOR=OoB0R=OPCOR=OoB0S=NnB0T=OnBOd[T2"}, "label": "the giraffe in the middle"}]} {"id": 28230, "image": "COCO_train2014_000000028230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman doing acrobatics on a street corner\"."}], "task": "refering", "answer_template": "The \"a woman doing acrobatics on a street corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 101, 123, 124, 146, 147, 166, 167, 169, 170, 171, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 194, 195, 196, 197, 199, 213, 214, 215, 216, 217, 237, 238, 261, 262], "bbox": [0.22140624999999997, 0.22185011709601873, 0.70009375, 0.7913583138173302], "iscrowd": 0, "area": 15059.872600000004, "rle": {"size": [427, 640], "counts": "e]k12Y=1N2N2N2N2N3M2N21O0O1O1O2N1O1O1O1OO100O101N10010O1O010O0010O01kCGV;:\\D4d;K[D7d;JZD8e;HZD:f;FWD=h;CWD?h;BVD`0j;a0O1O100O1O10N10001O00001O0000001O000100O1O1O1O1O2M2O1N2O1O1N2O1N2O1N3N000O0100O1O1003M2O2M2N2N3M2N2N3M2SNfMeH]2W7fMgHi0kN3[8WOiHe0nN4V8ZOjHb0SO4o7]OlH?WO4j7@mH<[O4e7CnH8A5]7FPI5E5X7IPI3J2U7NoH0ONR75mHM3Lo6;kHI8Jk6a0kHE=Gd6j0mH_Oa0E^6R1oHYOf0AW6]1QIROj0_OQ6e1RImNo0ZOm5o1RIgNT1TOg5]2SI_NX1oNb5j2TIWN\\1jN\\5e4gJTKW5R5Z17I7001N2O001O1O001QMTH=l7UOcHj0^7gNRIW1o6[N_Id1a6oMmIP2T6bMZJ]2[8O1O1O1O2N1O1O1O1N2O1O2N1O1O1O1O1O1O2M2O1N2N2O1N2O1N3N1N10O01O010O10O010O01O010O010O010O00100O0O2O001N110O1O100O010O1O100O010O1O100O00100O100O00100O100O00100O10001O000001O01O00000001O01O0000000010O00000001O01O00000000001OO10000O100O10000O10000O100O10000O10000O100O10000O10000O100O1001O1O1O1O1O1O1O1O1O1O00O001O001O001O010O001O001O001O2N2N1O2N1O2N1O2N1O2N1Obk_2"}, "label": "a woman doing acrobatics on a street corner"}]} {"id": 28230, "image": "COCO_train2014_000000028230.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl doing pole tricks\"."}], "task": "refering", "answer_template": "The \"a girl doing pole tricks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 101, 123, 124, 146, 147, 166, 167, 169, 170, 171, 172, 173, 174, 175, 176, 189, 190, 191, 192, 193, 194, 195, 196, 197, 199, 213, 214, 215, 216, 217, 237, 238, 261, 262], "bbox": [0.22140624999999997, 0.22185011709601873, 0.70009375, 0.7913583138173302], "iscrowd": 0, "area": 15059.872600000004, "rle": {"size": [427, 640], "counts": "e]k12Y=1N2N2N2N2N3M2N21O0O1O1O2N1O1O1O1OO100O101N10010O1O010O0010O01kCGV;:\\D4d;K[D7d;JZD8e;HZD:f;FWD=h;CWD?h;BVD`0j;a0O1O100O1O10N10001O00001O0000001O000100O1O1O1O1O2M2O1N2O1O1N2O1N2O1N3N000O0100O1O1003M2O2M2N2N3M2N2N3M2SNfMeH]2W7fMgHi0kN3[8WOiHe0nN4V8ZOjHb0SO4o7]OlH?WO4j7@mH<[O4e7CnH8A5]7FPI5E5X7IPI3J2U7NoH0ONR75mHM3Lo6;kHI8Jk6a0kHE=Gd6j0mH_Oa0E^6R1oHYOf0AW6]1QIROj0_OQ6e1RImNo0ZOm5o1RIgNT1TOg5]2SI_NX1oNb5j2TIWN\\1jN\\5e4gJTKW5R5Z17I7001N2O001O1O001QMTH=l7UOcHj0^7gNRIW1o6[N_Id1a6oMmIP2T6bMZJ]2[8O1O1O1O2N1O1O1O1N2O1O2N1O1O1O1O1O1O2M2O1N2N2O1N2O1N3N1N10O01O010O10O010O01O010O010O010O00100O0O2O001N110O1O100O010O1O100O010O1O100O00100O100O00100O100O00100O10001O000001O01O00000001O01O0000000010O00000001O01O00000000001OO10000O100O10000O10000O100O10000O10000O100O10000O10000O100O1001O1O1O1O1O1O1O1O1O1O00O001O001O001O010O001O001O001O2N2N1O2N1O2N1O2N1O2N1Obk_2"}, "label": "a girl doing pole tricks"}]} {"id": 126537, "image": "COCO_train2014_000000126537.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sunexpress airplane on a runway\"."}], "task": "refering", "answer_template": "The \"sunexpress airplane on a runway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 111, 112, 133, 134, 135, 155, 156, 157, 158, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 233, 240, 241, 242, 243], "bbox": [0.061328125, 0.24662763466042156, 0.9932968750000001, 0.6995550351288057], "iscrowd": 0, "area": 37557.54049999999, "rle": {"size": [427, 640], "counts": "\\``07T=0O2N101N1O2O0O2O1N1O2O0O2N101N2N101N101N1O2O1N1O2O0O2N101N2N101N101N1O2O1N101O0O2O001N2O001N101O1N101O0O2O1O0O2O1O0O2O001N2O001N01000O10O10O@XE_Nh:a1a00O10000O10000O10000O100000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O000O101O000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O0100000O10000O10O1000O1000000O1000O10O1000000O100000O01000000O1000000O0100000O1000000O10O1000O1000000O10O1000O1000000O1000O10O1000000O10000O0100000O100000O010O100O010O100O010O100O010O100O010O100O010O100O01000O10O0100O10O0100O10O0100O10O0100O10O01O1O1O0O2O1N2O0O2O1N101N2O1N101N2O1N101N2O1N101O0O2O1N101N2O0O2O0O2O1N101N2O0O2O1N101O0O2O1N101O10O010O1000O01XOdFaM\\9Z2hFgMW9U2nFjMS9Q2RGoMn8l1WGTNi8g1\\GXNd8e1`G[N`8`1eG_N\\8\\1iGdNW8W1nGiNQ8S1THlNm7o0XHQOh7j0]HVOc7e0bHZO^7b0gH^OY7=lHBU79oHHQ73TIMk60YIOh6L]I4c6GbI9]6ChI in this image.", "objects": [{"patches": [89, 111, 112, 133, 134, 135, 155, 156, 157, 158, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 233, 240, 241, 242, 243], "bbox": [0.061328125, 0.24662763466042156, 0.9932968750000001, 0.6995550351288057], "iscrowd": 0, "area": 37557.54049999999, "rle": {"size": [427, 640], "counts": "\\``07T=0O2N101N1O2O0O2O1N1O2O0O2N101N2N101N101N1O2O1N1O2O0O2N101N2N101N101N1O2O1N101O0O2O001N2O001N101O1N101O0O2O1O0O2O1O0O2O001N2O001N01000O10O10O@XE_Nh:a1a00O10000O10000O10000O100000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O000O101O000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O0100000O10000O10O1000O1000000O1000O10O1000000O100000O01000000O1000000O0100000O1000000O10O1000O1000000O10O1000O1000000O1000O10O1000000O10000O0100000O100000O010O100O010O100O010O100O010O100O010O100O010O100O01000O10O0100O10O0100O10O0100O10O0100O10O01O1O1O0O2O1N2O0O2O1N101N2O1N101N2O1N101N2O1N101O0O2O1N101N2O0O2O0O2O1N101N2O0O2O1N101O0O2O1N101O10O010O1000O01XOdFaM\\9Z2hFgMW9U2nFjMS9Q2RGoMn8l1WGTNi8g1\\GXNd8e1`G[N`8`1eG_N\\8\\1iGdNW8W1nGiNQ8S1THlNm7o0XHQOh7j0]HVOc7e0bHZO^7b0gH^OY7=lHBU79oHHQ73TIMk60YIOh6L]I4c6GbI9]6ChI in this image.", "objects": [{"patches": [140, 141, 162, 163, 164, 185, 186, 187, 207, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282], "bbox": [0.03671875, 0.4497897196261682, 0.289921875, 0.8461214953271028], "iscrowd": 0, "area": 15252.296999999993, "rle": {"size": [428, 640], "counts": "_Y:6U=8H7H8I7I6J4L4L5K4L4L4L4K6K4L4M1N2O1O1N3N1O1N2POVM\\Gk2]8]MaGd2[8`McGc2Y8bMdG_2Z8eMcG\\2[8hMaGZ2\\8kMaGV2]8nM`GS2]8`1N100O10000O100O10000O100O10000O10000O100O10001O0001O001O00001O00001O001bLhGf1Y8YNhGf1Y8XNiGg1X8XNjGf1W8XNjGh1W8WNjGh1W8WNiGi1Y8hMTHW2n7PMjHP3i8000001O0000001O00000M4J4K60O3N2M4M2M3M2O2M2O1O1O1O000001O000000001O0000001N3N2N2N2N2N2N2N6J6J7SOlDTOX;>REBn: in this image.", "objects": [{"patches": [140, 141, 162, 163, 164, 185, 186, 187, 207, 208, 209, 210, 211, 212, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282], "bbox": [0.03671875, 0.4497897196261682, 0.289921875, 0.8461214953271028], "iscrowd": 0, "area": 15252.296999999993, "rle": {"size": [428, 640], "counts": "_Y:6U=8H7H8I7I6J4L4L5K4L4L4L4K6K4L4M1N2O1O1N3N1O1N2POVM\\Gk2]8]MaGd2[8`McGc2Y8bMdG_2Z8eMcG\\2[8hMaGZ2\\8kMaGV2]8nM`GS2]8`1N100O10000O100O10000O100O10000O10000O100O10001O0001O001O00001O00001O001bLhGf1Y8YNhGf1Y8XNiGg1X8XNjGf1W8XNjGh1W8WNjGh1W8WNiGi1Y8hMTHW2n7PMjHP3i8000001O0000001O00000M4J4K60O3N2M4M2M3M2O2M2O1O1O1O000001O000000001O0000001N3N2N2N2N2N2N2N6J6J7SOlDTOX;>REBn: in this image.", "objects": [{"patches": [10, 17, 18, 33, 34, 36, 37, 39, 40, 57, 58, 59, 60, 61, 62, 81, 82, 83, 84, 104, 105, 106, 127, 128, 129, 150, 151, 152, 173, 174, 175, 196, 197, 198, 219, 220, 221, 242, 243, 244, 265, 266, 267], "bbox": [0.43581250000000005, 0.01796728971962617, 0.790484375, 0.7685280373831775], "iscrowd": 0, "area": 18868.515000000003, "rle": {"size": [428, 640], "counts": "fdd34T=8G8L20O0000001O00001O00M3N1100000001O3N3M2N1O0000O10000000000O10000000000O01000000000O10000000000O10000000000O100000O1001N101N1O1^E[O[8g0XGKb85RG=g8DmFQ1i8QOjFm1c8TNQGi2^8e1]Ob0mHhJn4m6N2O1N2N2N101N2N2N2SOUHVLh7j3^HRL_7o3gHmKW7S4oHiKn6X4XIeKe6[4U10O101N3O1000001O0000000;E in this image.", "objects": [{"patches": [10, 17, 18, 33, 34, 36, 37, 39, 40, 57, 58, 59, 60, 61, 62, 81, 82, 83, 84, 104, 105, 106, 127, 128, 129, 150, 151, 152, 173, 174, 175, 196, 197, 198, 219, 220, 221, 242, 243, 244, 265, 266, 267], "bbox": [0.43581250000000005, 0.01796728971962617, 0.790484375, 0.7685280373831775], "iscrowd": 0, "area": 18868.515000000003, "rle": {"size": [428, 640], "counts": "fdd34T=8G8L20O0000001O00001O00M3N1100000001O3N3M2N1O0000O10000000000O10000000000O01000000000O10000000000O10000000000O100000O1001N101N1O1^E[O[8g0XGKb85RG=g8DmFQ1i8QOjFm1c8TNQGi2^8e1]Ob0mHhJn4m6N2O1N2N2N101N2N2N2SOUHVLh7j3^HRL_7o3gHmKW7S4oHiKn6X4XIeKe6[4U10O101N3O1000001O0000000;E in this image.", "objects": [{"patches": [65, 66, 83, 84, 85, 101, 102, 103, 104, 105, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215], "bbox": [0.6196, 0.29582582582582584, 0.96382, 0.9887687687687687], "iscrowd": 0, "area": 25281.034499999998, "rle": {"size": [333, 500], "counts": "hmT3e0d96L5M3L4L2O1N1O2O1N2N2O0O2N2O0O2N2O0O2O1N101O1N101O0O2PInMc5R2WJYNd5h1UJcNe5^1UJmNf5T1TJVOg5k0WJ[Oc5g0[J^Oh1YNc1Z2cLCY1cNP2k1eLMc0kNb2Z1iLP2S3QNkLS2R3nMlLS2T3nMjLS2W3a21O001O1O001O1O001O1O001O1O001O2N1O2N1O2N2N1O2N5K7I1O000000000000000000000000001O00000000000000001O0000001O00001O0000001O0000001O00001O00001O001O00001O001O001O001O00001O003M3M3M2N3M3M3M3M3M2N3M3M3M3M3M3M2N3M3M3M3M3M3M2N3M3M3M2N2N2N1O2N2N1O2N2N8H8H7I8H8H8H9G8H9G8HVa5"}, "label": "the blonde woman in the black coat with a water bottle in her pocket"}]} {"id": 44637, "image": "COCO_train2014_000000044637.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl in black coat with red hair\"."}], "task": "refering", "answer_template": "The \"girl in black coat with red hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 83, 84, 85, 101, 102, 103, 104, 105, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215], "bbox": [0.6196, 0.29582582582582584, 0.96382, 0.9887687687687687], "iscrowd": 0, "area": 25281.034499999998, "rle": {"size": [333, 500], "counts": "hmT3e0d96L5M3L4L2O1N1O2O1N2N2O0O2N2O0O2N2O0O2O1N101O1N101O0O2PInMc5R2WJYNd5h1UJcNe5^1UJmNf5T1TJVOg5k0WJ[Oc5g0[J^Oh1YNc1Z2cLCY1cNP2k1eLMc0kNb2Z1iLP2S3QNkLS2R3nMlLS2T3nMjLS2W3a21O001O1O001O1O001O1O001O1O001O2N1O2N1O2N2N1O2N5K7I1O000000000000000000000000001O00000000000000001O0000001O00001O0000001O0000001O00001O00001O001O00001O001O001O001O00001O003M3M3M2N3M3M3M3M3M2N3M3M3M3M3M3M2N3M3M3M3M3M3M2N3M3M3M2N2N2N1O2N2N1O2N2N8H8H7I8H8H8H9G8H9G8HVa5"}, "label": "girl in black coat with red hair"}]} {"id": 44637, "image": "COCO_train2014_000000044637.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young brunette woman hugging a red haired woman\"."}], "task": "refering", "answer_template": "The \"a young brunette woman hugging a red haired woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 63, 64, 65, 66, 81, 82, 83, 97, 98, 99, 100, 101, 116, 117, 118, 119, 120, 134, 135, 136, 137, 138, 153, 154, 155, 156, 171, 172, 173, 174, 188, 189, 190, 191, 192, 206, 207, 208, 209, 210], "bbox": [0.41906, 0.20672672672672673, 0.6989200000000001, 1.0], "iscrowd": 0, "area": 22689.122450000006, "rle": {"size": [333, 500], "counts": "X^T21[:2O1N2N2O1N2N2O1N2O1N2O1N2O1N2O2M2N2O1N2O1N2ZKmN0T1JQO5R1ERO;o0@UO?m0[OXOe0i0VO[Oi0g0RO]On0d0oN^OP1d0mN^OR1e0kN]OT1d0jN]OU1f0gN\\OY1e0eN\\OZ1g0cNZO\\1h0bNYO]1j0_NXO_1l0^NUOa1m0\\NUOc1m0YNVOf1m0TNWOj1l0QNXOn1j0nMYOQ2j0iM[OU2g0gM\\OX2f0cM^O\\2e0_M^O`2d0\\LWMc0Y2P3Q1iLROW3R1aLRO^3S1[LPOd3m3O1L4M3M3L4M3L4M3N2O1N2O1O1N2O1N2O1O1N200O100O100O100O100O100O100O100O100O100O100000000001O000000000nIoMh3R2PLXNm3h1lKaNR4_1gKjNW4V1bKTO[4l0dKWOZ4i0fKZOW4f0iK[OV4f0hK[OX4e0hK\\OW4d0iK\\OW4d0hK^OW4b0iK^OW4b0hK@W4a0hK_OX4b0gK_OX4c0eK^O[4c0dK^O[4c0dK^O[4d0bK]O^4d0aK]O^4d0`K]O`4e0^K\\Oa4e0^K[Ob4f0\\K\\Oc4f0[KZOe0oMk2i2^LYOc0WNi2c2`LWOe0^Ne2]2dLVOd0fNa2W2gLTOf0mN]2Q2kLSOd0UO[2P4cM]LR2d3kMjLo0QNSOU5LXMM_N3\\4JjN3Y1HmN4j5LQi`1"}, "label": "a young brunette woman hugging a red haired woman"}]} {"id": 44637, "image": "COCO_train2014_000000044637.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a coat trimmed with fur\"."}], "task": "refering", "answer_template": "The \"a woman in a coat trimmed with fur\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 63, 64, 65, 66, 81, 82, 83, 97, 98, 99, 100, 101, 116, 117, 118, 119, 120, 134, 135, 136, 137, 138, 153, 154, 155, 156, 171, 172, 173, 174, 188, 189, 190, 191, 192, 206, 207, 208, 209, 210], "bbox": [0.41906, 0.20672672672672673, 0.6989200000000001, 1.0], "iscrowd": 0, "area": 22689.122450000006, "rle": {"size": [333, 500], "counts": "X^T21[:2O1N2N2O1N2N2O1N2O1N2O1N2O1N2O2M2N2O1N2O1N2ZKmN0T1JQO5R1ERO;o0@UO?m0[OXOe0i0VO[Oi0g0RO]On0d0oN^OP1d0mN^OR1e0kN]OT1d0jN]OU1f0gN\\OY1e0eN\\OZ1g0cNZO\\1h0bNYO]1j0_NXO_1l0^NUOa1m0\\NUOc1m0YNVOf1m0TNWOj1l0QNXOn1j0nMYOQ2j0iM[OU2g0gM\\OX2f0cM^O\\2e0_M^O`2d0\\LWMc0Y2P3Q1iLROW3R1aLRO^3S1[LPOd3m3O1L4M3M3L4M3L4M3N2O1N2O1O1N2O1N2O1O1N200O100O100O100O100O100O100O100O100O100O100000000001O000000000nIoMh3R2PLXNm3h1lKaNR4_1gKjNW4V1bKTO[4l0dKWOZ4i0fKZOW4f0iK[OV4f0hK[OX4e0hK\\OW4d0iK\\OW4d0hK^OW4b0iK^OW4b0hK@W4a0hK_OX4b0gK_OX4c0eK^O[4c0dK^O[4c0dK^O[4d0bK]O^4d0aK]O^4d0`K]O`4e0^K\\Oa4e0^K[Ob4f0\\K\\Oc4f0[KZOe0oMk2i2^LYOc0WNi2c2`LWOe0^Ne2]2dLVOd0fNa2W2gLTOf0mN]2Q2kLSOd0UO[2P4cM]LR2d3kMjLo0QNSOU5LXMM_N3\\4JjN3Y1HmN4j5LQi`1"}, "label": "a woman in a coat trimmed with fur"}]} {"id": 454246, "image": "COCO_train2014_000000454246.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse in the background with a cover on it ' s back\"."}], "task": "refering", "answer_template": "The \"the horse in the background with a cover on it ' s back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 54, 55, 56, 57, 58, 59, 60, 77, 78, 81, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 134, 135, 149, 150, 154, 155, 156, 172, 173, 177, 178, 179, 180, 195], "bbox": [0.366453125, 0.17419014084507042, 0.89321875, 0.8258098591549295], "iscrowd": 0, "area": 17823.86355, "rle": {"size": [284, 640], "counts": "hXQ21i82N2N2N2N2N2M3M3N2M3M3N2O1N2O1N2O1N2N2N2M2J6100O1O1O2O1N2O0O2O001N100O2O000O10000O100O2O000O100O10000O10000O101N10000O10000O100O2O000O10001N1000jJ@Q2>QNFj1:VNJg15ZNOa10`NOmL;h3E\\OMQM6h3MWOKUM4f30WOIXM2d35SOG\\M0d39oNF`MNd3:mNFn1:SNFm18UNGk19VNGj18VNIi16YNJf16[NIjM3_23GKfM9_2KVOLkM0<R21O01O010O01O010O010O0010O0010O0010O001M2O1N3M2O2M2Nlob0"}, "label": "the horse in the background with a cover on it ' s back"}]} {"id": 454246, "image": "COCO_train2014_000000454246.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"one of two horses running past each other in a dirt field . he has a black covering over his body\"."}], "task": "refering", "answer_template": "The \"one of two horses running past each other in a dirt field . he has a black covering over his body\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 54, 55, 56, 57, 58, 59, 60, 77, 78, 81, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 134, 135, 149, 150, 154, 155, 156, 172, 173, 177, 178, 179, 180, 195], "bbox": [0.366453125, 0.17419014084507042, 0.89321875, 0.8258098591549295], "iscrowd": 0, "area": 17823.86355, "rle": {"size": [284, 640], "counts": "hXQ21i82N2N2N2N2N2M3M3N2M3M3N2O1N2O1N2O1N2N2N2M2J6100O1O1O2O1N2O0O2O001N100O2O000O10000O100O2O000O100O10000O10000O101N10000O10000O100O2O000O10001N1000jJ@Q2>QNFj1:VNJg15ZNOa10`NOmL;h3E\\OMQM6h3MWOKUM4f30WOIXM2d35SOG\\M0d39oNF`MNd3:mNFn1:SNFm18UNGk19VNGj18VNIi16YNJf16[NIjM3_23GKfM9_2KVOLkM0<R21O01O010O01O010O010O0010O0010O0010O001M2O1N3M2O2M2Nlob0"}, "label": "one of two horses running past each other in a dirt field . he has a black covering over his body"}]} {"id": 454246, "image": "COCO_train2014_000000454246.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown horse with white hooves is next to the horse outside\"."}], "task": "refering", "answer_template": "The \"the brown horse with white hooves is next to the horse outside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 96, 97, 98, 99, 100, 101, 102, 103, 105, 119, 120, 121, 122, 123, 124, 125, 126, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 171, 172, 173, 174, 188, 189, 190, 194, 195], "bbox": [0.186234375, 0.19721830985915492, 0.612890625, 0.8610211267605634], "iscrowd": 0, "area": 22477.906549999996, "rle": {"size": [284, 640], "counts": "nTQ19j7k0J7I6J6J7L3M2SKoMm2S2[LdNc3^1TLjNl3W1oKmNP4V1kKmNT4U1gKPOX4Q1dKRO[4Q1aKQO_44_KSO2h0^4JmK^ODi0]4HSL_O_Oh0^4GVLA\\Og0]4G[L_OXOl0\\4D^L^OWO92AY4g0bL]OROKf0Of3g0eL]OPOLg0Od3f0iL\\OlN0i0Lb3f0lL]OiN0l0K_3g0oL[OfN4n0H]3h0QMWOhN8l0H\\3g0]MAYOE[3i0\\MBl3=iKYOhN:_5mMCP2?PNBn1>SNCk1>VNBg1`0YN@f1`0[NAc1`0^N@_1a0cN_O[1b0eN_OX1c0iN\\OU1e0mNZOQ1h0POXOm0j0TOUOj0l0XOTOf0m0ZOSOd0o0]OPOb0P1@nN`0S1AjN?X1BfN>[1CbN`0\\1AbNb0]1_OaNc0^1^O_Ne0`1]O]Ne0b1\\3O1N2O001N2N2O1N2N2N3M2N3M2N4L3M3L5J5K6J^eT2"}, "label": "the brown horse with white hooves is next to the horse outside"}]} {"id": 454246, "image": "COCO_train2014_000000454246.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two horses on a plot of dirt\"."}], "task": "refering", "answer_template": "The \"two horses on a plot of dirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 96, 97, 98, 99, 100, 101, 102, 103, 105, 119, 120, 121, 122, 123, 124, 125, 126, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 171, 172, 173, 174, 188, 189, 190, 194, 195], "bbox": [0.186234375, 0.19721830985915492, 0.612890625, 0.8610211267605634], "iscrowd": 0, "area": 22477.906549999996, "rle": {"size": [284, 640], "counts": "nTQ19j7k0J7I6J6J7L3M2SKoMm2S2[LdNc3^1TLjNl3W1oKmNP4V1kKmNT4U1gKPOX4Q1dKRO[4Q1aKQO_44_KSO2h0^4JmK^ODi0]4HSL_O_Oh0^4GVLA\\Og0]4G[L_OXOl0\\4D^L^OWO92AY4g0bL]OROKf0Of3g0eL]OPOLg0Od3f0iL\\OlN0i0Lb3f0lL]OiN0l0K_3g0oL[OfN4n0H]3h0QMWOhN8l0H\\3g0]MAYOE[3i0\\MBl3=iKYOhN:_5mMCP2?PNBn1>SNCk1>VNBg1`0YN@f1`0[NAc1`0^N@_1a0cN_O[1b0eN_OX1c0iN\\OU1e0mNZOQ1h0POXOm0j0TOUOj0l0XOTOf0m0ZOSOd0o0]OPOb0P1@nN`0S1AjN?X1BfN>[1CbN`0\\1AbNb0]1_OaNc0^1^O_Ne0`1]O]Ne0b1\\3O1N2O001N2N2O1N2N2N3M2N3M2N4L3M3L5J5K6J^eT2"}, "label": "two horses on a plot of dirt"}]} {"id": 560750, "image": "COCO_train2014_000000560750.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the reflection of the toothbrush in the mirror\"."}], "task": "refering", "answer_template": "The \"the reflection of the toothbrush in the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 107, 108, 109, 129, 130, 131, 132, 153, 154, 155, 176, 177, 178, 200, 201, 223, 224, 246, 247, 269, 270, 292, 293, 316], "bbox": [0.644703125, 0.2375175644028103, 0.7796406250000001, 0.8937002341920375], "iscrowd": 0, "area": 10191.0501, "rle": {"size": [427, 640], "counts": "l[\\56S=8G8I7I7G9\\Od0_O=N2N3M2N3M2N3L3N2N3M2N3M2N3L300O1O00010O00001O01O0001O01O01O00010O0000010O000010O000?Aa0_Of1[N5J7I6M3M3M3L5L3M3M3L4M00cJ^Jk2e5PMeJh2^5RMlJf2W5UMTKc2o4WM[Kb2Z5gLPKP3h5VLbJc3j7H9H7I8G8I7I8G8I7H8I7I7H8I7I7Ho]j1"}, "label": "the reflection of the toothbrush in the mirror"}]} {"id": 560750, "image": "COCO_train2014_000000560750.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"toothbrush to the right\"."}], "task": "refering", "answer_template": "The \"toothbrush to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 107, 108, 109, 129, 130, 131, 132, 153, 154, 155, 176, 177, 178, 200, 201, 223, 224, 246, 247, 269, 270, 292, 293, 316], "bbox": [0.644703125, 0.2375175644028103, 0.7796406250000001, 0.8937002341920375], "iscrowd": 0, "area": 10191.0501, "rle": {"size": [427, 640], "counts": "l[\\56S=8G8I7I7G9\\Od0_O=N2N3M2N3M2N3L3N2N3M2N3M2N3L300O1O00010O00001O01O0001O01O01O00010O0000010O000010O000?Aa0_Of1[N5J7I6M3M3M3L5L3M3M3L4M00cJ^Jk2e5PMeJh2^5RMlJf2W5UMTKc2o4WM[Kb2Z5gLPKP3h5VLbJc3j7H9H7I8G8I7I8G8I7H8I7I7H8I7I7Ho]j1"}, "label": "toothbrush to the right"}]} {"id": 560750, "image": "COCO_train2014_000000560750.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white toothbrush\"."}], "task": "refering", "answer_template": "The \"a white toothbrush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 103, 104, 105, 106, 126, 127, 128, 129, 149, 150, 151, 172, 173, 174, 194, 195, 196, 217, 218, 240], "bbox": [0.45578124999999997, 0.25775175644028103, 0.6387031249999999, 0.6824590163934425], "iscrowd": 0, "area": 8868.030999999999, "rle": {"size": [427, 640], "counts": "[Qj34R=5K5J5L5J6K5J5L5J6K4K6K5J6K4K6K5L3N3M3M3MO1N2N2M3N2N2N2N2N3M2M3N2N2004L4N2N2M3N2N2N1N3N2N2M3N2N2N2M3N2N2M3N2N2N2M3N2N1O0100O2N1O1O1O2O0O1O1O1O2N100O1O1O2L9G3N3L4L3M4M3L3M4M2N3M3L3N3M3M2M4M3M2N3M3M2N3N2M2O2M3N1N3M2O2M3N1N3N2M2O2M3N1NoUP3"}, "label": "a white toothbrush"}]} {"id": 560750, "image": "COCO_train2014_000000560750.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the head of a toothbrush in sticking out of a jar\"."}], "task": "refering", "answer_template": "The \"the head of a toothbrush in sticking out of a jar\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 103, 104, 105, 106, 126, 127, 128, 129, 149, 150, 151, 172, 173, 174, 194, 195, 196, 217, 218, 240], "bbox": [0.45578124999999997, 0.25775175644028103, 0.6387031249999999, 0.6824590163934425], "iscrowd": 0, "area": 8868.030999999999, "rle": {"size": [427, 640], "counts": "[Qj34R=5K5J5L5J6K5J5L5J6K4K6K5J6K4K6K5L3N3M3M3MO1N2N2M3N2N2N2N2N3M2M3N2N2004L4N2N2M3N2N2N1N3N2N2M3N2N2N2M3N2N2M3N2N2N2M3N2N1O0100O2N1O1O1O2O0O1O1O1O2N100O1O1O2L9G3N3L4L3M4M3L3M4M2N3M3L3N3M3M2M4M3M2N3M3M2N3N2M2O2M3N1N3M2O2M3N1N3N2M2O2M3N1NoUP3"}, "label": "the head of a toothbrush in sticking out of a jar"}]} {"id": 69231, "image": "COCO_train2014_000000069231.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"luxurious black cat snuggled on the white blanket\"."}], "task": "refering", "answer_template": "The \"luxurious black cat snuggled on the white blanket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 85, 102, 104, 105, 106, 107], "bbox": [0.387640625, 0.09045833333333334, 0.709546875, 0.2702291666666667], "iscrowd": 0, "area": 8980.922149999999, "rle": {"size": [480, 640], "counts": "lZd31o>2M3N2N2N3M2M10001O000001O0001O01O00010O01O0010O01O0010O01O010O000001O0001M2M3M3O1000001O00O101O2M3N2N1O0O01N2O001O1O001O10O1001O00000001O000000cNYOaDg0_;\\O_Dc0a;@\\D`0e;BPDf0Q<\\OjCf0V<^OcCf0]<[O_Cg0b<[OYCg0g<\\OTCg0l in this image.", "objects": [{"patches": [34, 57, 58, 59, 60, 61, 78, 79, 80, 81, 82, 83, 84, 85, 102, 104, 105, 106, 107], "bbox": [0.387640625, 0.09045833333333334, 0.709546875, 0.2702291666666667], "iscrowd": 0, "area": 8980.922149999999, "rle": {"size": [480, 640], "counts": "lZd31o>2M3N2N2N3M2M10001O000001O0001O01O00010O01O0010O01O0010O01O010O000001O0001M2M3M3O1000001O00O101O2M3N2N1O0O01N2O001O1O001O10O1001O00000001O000000cNYOaDg0_;\\O_Dc0a;@\\D`0e;BPDf0Q<\\OjCf0V<^OcCf0]<[O_Cg0b<[OYCg0g<\\OTCg0l in this image.", "objects": [{"patches": [106, 129, 130, 149, 150, 151, 152, 153, 157, 158, 170, 171, 172, 173, 174, 175, 176, 177, 179, 180, 181, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 326, 327, 328, 331, 332, 354, 355], "bbox": [0.066265625, 0.26185416666666667, 0.910015625, 0.9191874999999999], "iscrowd": 0, "area": 63508.66394999997, "rle": {"size": [480, 640], "counts": "j^d01o>001O1O001O1O001O1O00100O001O1O001O1O001O1O001O1O001O1O2N2N3M2N3M2N2N3N1N3M2N3M2N3M2N1O2N1O2N2N1N3N1N3M2O2M3N1N3N0O2O1N1O2O1N101N2O001O1N101O1N101O0O2O001O1N101O0O2O001N2O0O2N1O2N101N2N1O101N1O1O2O0O1O2N100O1O2N100O2N1O101N1O100O100O100O1O100O010O100O1O100O100O10O01O000001O01O00000001O01O0000N3M2O1N3M201O0O101N101O0O101O0010O1O100O100O1O10O0100O1O100O100O1O100O100O10O01O100O100O1O01N1O2N1O1O2N1100O2N10001N10001N101O0O101O5K6I7ZGlJT7W5gHkJY7X5cHjJ[7Z5aHhJ^7Z5^HiJa7Z5[HhJd7[5XHgJf7[6O1O1N2O000O2O00000O10001N10000O10001O0O100000O001UNVH`Lj7`3WH^Lj7b3WH]Li7c3XH[Lh7g3XHWLi7i3XHVLh7j3YHTLh7l3[HPLe7Q4bHfK`7Z4dH`K^7`4eHZK^7f4eHUK]7k4fHPK[7Q5gHkJ[7V5gHeJ[7]5S11O1N2O1N2N2O1N2O1N2N2OiMSG\\Nk8e1VGZNj8f1WGYNh8h1YGWNf8j1[GUNe8k1\\GTNc8m1^GRNb8n1_GQN`8P2aGoM^8R2cGmM]8S2dGkM\\8V2eGhM\\8X2eGfM[8[2eGeMZ8\\2gGbMZ8^2gG`MY8a2hG]MX8d2iGZMX8f2iGXMW8i2kGUMU8k2mGRMS8o2oGnLQ8S3QHjLP8V3RHgLn7Z3THdLk7]3WH`Lj7`3YH\\Lg7e3j10O100000000000000O100000000000000O1000000000000000000000000000000000000001O0000000000000000000000000O2O00000O10ROfKVGZ4c8oK\\GP4\\8[LaGd3Z8dLdG\\3Z8iLcGW3\\8mLaGS3]8QMaGn2_8VM_Gi2b8XM\\Gh2e8ZMXGf2i8\\MTGc2o8]MoFc2R9_MlF`2U9e12N1N2O1O2N2M3N2N2N2N2N1O2N2N3M2N3M2N3M2N2N3M2N3L3N3M2N3M4L4L4L4L4L3M4L4L3M3M2N3L3M3M100O1O00100O1O1O00100O1O1O0100000000O1000000O1000000O1000000O10000O1000000O10000O1000000O10000O01O100O00100O1O010O100O1O010kNYC0h in this image.", "objects": [{"patches": [106, 129, 130, 149, 150, 151, 152, 153, 157, 158, 170, 171, 172, 173, 174, 175, 176, 177, 179, 180, 181, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 326, 327, 328, 331, 332, 354, 355], "bbox": [0.066265625, 0.26185416666666667, 0.910015625, 0.9191874999999999], "iscrowd": 0, "area": 63508.66394999997, "rle": {"size": [480, 640], "counts": "j^d01o>001O1O001O1O001O1O00100O001O1O001O1O001O1O001O1O001O1O2N2N3M2N3M2N2N3N1N3M2N3M2N3M2N1O2N1O2N2N1N3N1N3M2O2M3N1N3N0O2O1N1O2O1N101N2O001O1N101O1N101O0O2O001O1N101O0O2O001N2O0O2N1O2N101N2N1O101N1O1O2O0O1O2N100O1O2N100O2N1O101N1O100O100O100O1O100O010O100O1O100O100O10O01O000001O01O00000001O01O0000N3M2O1N3M201O0O101N101O0O101O0010O1O100O100O1O10O0100O1O100O100O1O100O100O10O01O100O100O1O01N1O2N1O1O2N1100O2N10001N10001N101O0O101O5K6I7ZGlJT7W5gHkJY7X5cHjJ[7Z5aHhJ^7Z5^HiJa7Z5[HhJd7[5XHgJf7[6O1O1N2O000O2O00000O10001N10000O10001O0O100000O001UNVH`Lj7`3WH^Lj7b3WH]Li7c3XH[Lh7g3XHWLi7i3XHVLh7j3YHTLh7l3[HPLe7Q4bHfK`7Z4dH`K^7`4eHZK^7f4eHUK]7k4fHPK[7Q5gHkJ[7V5gHeJ[7]5S11O1N2O1N2N2O1N2O1N2N2OiMSG\\Nk8e1VGZNj8f1WGYNh8h1YGWNf8j1[GUNe8k1\\GTNc8m1^GRNb8n1_GQN`8P2aGoM^8R2cGmM]8S2dGkM\\8V2eGhM\\8X2eGfM[8[2eGeMZ8\\2gGbMZ8^2gG`MY8a2hG]MX8d2iGZMX8f2iGXMW8i2kGUMU8k2mGRMS8o2oGnLQ8S3QHjLP8V3RHgLn7Z3THdLk7]3WH`Lj7`3YH\\Lg7e3j10O100000000000000O100000000000000O1000000000000000000000000000000000000001O0000000000000000000000000O2O00000O10ROfKVGZ4c8oK\\GP4\\8[LaGd3Z8dLdG\\3Z8iLcGW3\\8mLaGS3]8QMaGn2_8VM_Gi2b8XM\\Gh2e8ZMXGf2i8\\MTGc2o8]MoFc2R9_MlF`2U9e12N1N2O1O2N2M3N2N2N2N2N1O2N2N3M2N3M2N3M2N2N3M2N3L3N3M2N3M4L4L4L4L4L3M4L4L3M3M2N3L3M3M100O1O00100O1O1O00100O1O1O0100000000O1000000O1000000O1000000O10000O1000000O10000O1000000O10000O01O100O00100O1O010O100O1O010kNYC0h in this image.", "objects": [{"patches": [72, 73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 103, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 193, 194, 196, 197, 198, 210, 211, 212, 213, 216, 217, 219, 220, 233, 234, 235, 236, 239, 240, 242, 243, 256, 257, 259, 260, 263, 265, 266], "bbox": [0.14275, 0.1981967213114754, 0.632625, 0.7995550351288055], "iscrowd": 0, "area": 42799.63785, "rle": {"size": [427, 640], "counts": "aSV11Uc0Dk8NdF7o0A^87eF3W;LjDN[;2fDH_;7k001N100O101N2O1O1N3NZVR3"}, "label": "a black and white zebra eating grass in a field"}]} {"id": 413297, "image": "COCO_train2014_000000413297.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"one zebra closest to the camera\"."}], "task": "refering", "answer_template": "The \"one zebra closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 103, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 193, 194, 196, 197, 198, 210, 211, 212, 213, 216, 217, 219, 220, 233, 234, 235, 236, 239, 240, 242, 243, 256, 257, 259, 260, 263, 265, 266], "bbox": [0.14275, 0.1981967213114754, 0.632625, 0.7995550351288055], "iscrowd": 0, "area": 42799.63785, "rle": {"size": [427, 640], "counts": "aSV11Uc0Dk8NdF7o0A^87eF3W;LjDN[;2fDH_;7k001N100O101N2O1O1N3NZVR3"}, "label": "one zebra closest to the camera"}]} {"id": 413297, "image": "COCO_train2014_000000413297.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra behind another zebra\"."}], "task": "refering", "answer_template": "The \"a zebra behind another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 133, 150, 151, 152, 154, 155, 156, 169, 170, 174, 175, 177, 178, 179, 192, 193, 198, 201, 202, 215, 216, 221, 222, 224, 225, 238, 239], "bbox": [0.36796875, 0.16665105386416862, 0.8106874999999999, 0.6968618266978923], "iscrowd": 0, "area": 20958.827299999994, "rle": {"size": [427, 640], "counts": "kcR35S=7J6JW1iN9H3N2M3N2N8G8I0O101OiNhE[OW:b0oE\\OR:`0TF]OY6TOkLX1RMBm97YFGk5XOUMj0VMLe5ZOXMd0XM0d9IcF4X5_OaM6\\M9S5A_O<`0EB7`0HC5=LF0;0_6100O0100000000000O10000000000000000O100000001O1O1O2N1N2O2N1O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N100O2N1O2N1O1O2N1O2N1N3N1O1O1N1000O01000000O10O1000O100000O01000000O10O10O1bFZMW8f2YGjMg8]31O01O0001O0001O01O000000010O01XGhK[8X4eGkKX8V4fGoKW8P4iGTLS8m3kGXLQ8h3oG^Lk7c3SHiLc7V3]HlLa7U3]HmLb7S3^HmLb7_40WO`HfK_7c3YI\\Lg6m2RJRMn5`2aJ_M_5^2eJaM[5\\2iJdMU5Z2nJfMR5W2RKhMn4U2VKjMj4S2ZKlMe4T2\\KkMe4T2]KjMd4U2^KjMa4V2aKhM`4V2bKiM_4V2cKgM_4X2cKfM]4Z2[3N2O1O001N2O001N2O1O0O2O1O1O00100O10O0100O010O1O10O101O1O0O2O1O0O2O1O001N2O001O1O010O1O010O>B7JO000O101O01O010O1O010O001O010O00100O004M5JO1O100O0010O00000O2O0O101O01O100O0O2N2M2O9dN_E\\Of:XO[Ej08H`:[O^Ed0;K[:\\O`Eb0 in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 133, 150, 151, 152, 154, 155, 156, 169, 170, 174, 175, 177, 178, 179, 192, 193, 198, 201, 202, 215, 216, 221, 222, 224, 225, 238, 239], "bbox": [0.36796875, 0.16665105386416862, 0.8106874999999999, 0.6968618266978923], "iscrowd": 0, "area": 20958.827299999994, "rle": {"size": [427, 640], "counts": "kcR35S=7J6JW1iN9H3N2M3N2N8G8I0O101OiNhE[OW:b0oE\\OR:`0TF]OY6TOkLX1RMBm97YFGk5XOUMj0VMLe5ZOXMd0XM0d9IcF4X5_OaM6\\M9S5A_O<`0EB7`0HC5=LF0;0_6100O0100000000000O10000000000000000O100000001O1O1O2N1N2O2N1O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N100O2N1O2N1O1O2N1O2N1N3N1O1O1N1000O01000000O10O1000O100000O01000000O10O10O1bFZMW8f2YGjMg8]31O01O0001O0001O01O000000010O01XGhK[8X4eGkKX8V4fGoKW8P4iGTLS8m3kGXLQ8h3oG^Lk7c3SHiLc7V3]HlLa7U3]HmLb7S3^HmLb7_40WO`HfK_7c3YI\\Lg6m2RJRMn5`2aJ_M_5^2eJaM[5\\2iJdMU5Z2nJfMR5W2RKhMn4U2VKjMj4S2ZKlMe4T2\\KkMe4T2]KjMd4U2^KjMa4V2aKhM`4V2bKiM_4V2cKgM_4X2cKfM]4Z2[3N2O1O001N2O001N2O1O0O2O1O1O00100O10O0100O010O1O10O101O1O0O2O1O0O2O1O001N2O001O1O010O1O010O>B7JO000O101O01O010O1O010O001O010O00100O004M5JO1O100O0010O00000O2O0O101O01O100O0O2N2M2O9dN_E\\Of:XO[Ej08H`:[O^Ed0;K[:\\O`Eb0 in this image.", "objects": [{"patches": [150, 151, 152, 166, 167, 168, 169, 183, 184, 185, 186, 200, 201, 202, 203, 217, 218, 219, 220, 235, 236, 237, 253, 254, 270, 271, 288], "bbox": [0.8017291666666666, 0.369765625, 1.0, 0.716703125], "iscrowd": 0, "area": 14484.3522, "rle": {"size": [640, 480], "counts": "fn`78]c0;D;F;DWOi0N2N2N2N2N2N2N2N2N2N2O1O1N2O1O1O1O1O1O1O1O1O1O2N1O1O2N1O1O2N1O2N2N2N3M2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N3N1O1O1O2N00000000O1000000O1N2N2N2N2N2nMVBkMl=U2]B`Me=`2dBTM_=l2jBiLX=W3QC^LQ=b3XCSLj in this image.", "objects": [{"patches": [150, 151, 152, 166, 167, 168, 169, 183, 184, 185, 186, 200, 201, 202, 203, 217, 218, 219, 220, 235, 236, 237, 253, 254, 270, 271, 288], "bbox": [0.8017291666666666, 0.369765625, 1.0, 0.716703125], "iscrowd": 0, "area": 14484.3522, "rle": {"size": [640, 480], "counts": "fn`78]c0;D;F;DWOi0N2N2N2N2N2N2N2N2N2N2O1O1N2O1O1O1O1O1O1O1O1O1O2N1O1O2N1O1O2N1O2N2N2N3M2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1N2N2N2N3N1O1O1O2N00000000O1000000O1N2N2N2N2N2nMVBkMl=U2]B`Me=`2dBTM_=l2jBiLX=W3QC^LQ=b3XCSLj in this image.", "objects": [{"patches": [216, 217, 233, 234, 235, 250, 251, 252, 267, 268, 269, 283, 284, 285, 286, 300, 301, 302, 303, 317, 318, 319, 334, 335, 351, 352, 353, 368, 370, 385], "bbox": [0.6552708333333332, 0.5329375, 0.8831458333333333, 1.0], "iscrowd": 0, "area": 15427.276799999996, "rle": {"size": [640, 480], "counts": "P\\U64hc05L4L5J5f]O:U`0Jh_On1_>VN^Ak1_>[N\\Ag1b>]N[Ad1b>bNZA_1d>eNYA\\1e>iNWAY1e>mNWAT1g>S2N2N2OaM\\AEc>;_AE_>;dAC[>>fAAY>?iA@V>`0mAlMd?V2\\@jMb?X2_@gM_?[2a@eM]?]2c@cM[?_2f@aMW?`2j@`MT?b2m@]MQ?e2o@[Mo>g2QAZMl>h2UAWMi>j2XAVMf>l2[ASMc>o2]ARM`>P3`APM_>Q3bAnL\\>T3dAlL[>T3gAkLW>W3iAjLT>X3lAhLS>Y3nAfLP>\\3PBeLn=[3TBeLi=]3WBdLS<]OcDS4ZOaLP in this image.", "objects": [{"patches": [216, 217, 233, 234, 235, 250, 251, 252, 267, 268, 269, 283, 284, 285, 286, 300, 301, 302, 303, 317, 318, 319, 334, 335, 351, 352, 353, 368, 370, 385], "bbox": [0.6552708333333332, 0.5329375, 0.8831458333333333, 1.0], "iscrowd": 0, "area": 15427.276799999996, "rle": {"size": [640, 480], "counts": "P\\U64hc05L4L5J5f]O:U`0Jh_On1_>VN^Ak1_>[N\\Ag1b>]N[Ad1b>bNZA_1d>eNYA\\1e>iNWAY1e>mNWAT1g>S2N2N2OaM\\AEc>;_AE_>;dAC[>>fAAY>?iA@V>`0mAlMd?V2\\@jMb?X2_@gM_?[2a@eM]?]2c@cM[?_2f@aMW?`2j@`MT?b2m@]MQ?e2o@[Mo>g2QAZMl>h2UAWMi>j2XAVMf>l2[ASMc>o2]ARM`>P3`APM_>Q3bAnL\\>T3dAlL[>T3gAkLW>W3iAjLT>X3lAhLS>Y3nAfLP>\\3PBeLn=[3TBeLi=]3WBdLS<]OcDS4ZOaLP in this image.", "objects": [{"patches": [125, 136, 137, 138, 139, 148, 149, 150, 151, 152, 153, 162, 163, 164, 165, 166, 176, 177, 178, 179, 189, 190, 191, 192, 193], "bbox": [0.5460313315926892, 0.37834375, 0.9998172323759791, 0.583015625], "iscrowd": 0, "area": 10887.973800000003, "rle": {"size": [640, 383], "counts": "^oR41nc01O101O0O100O100000000000000000000000001O00010O0000000iNi0e]OWOVb0P1i]OPOQb0W1m]OiNna0^1Q^OaNja0e1U^O[Nha0i1X^OVNga0l1X^OTNha0m1X^ORNga0T201O00O100O101NL5N1O101N10O2O0O2O010O000011N3N1N2O1N2N2O2M2N2O1N2O1N00010O01O000100O01N1100O6K1N2O0O2N2O0O100O00010OO2N1O2N100001O01O00001O001N100000001N1O2O1N2N2N2O1N3N1N3N1O1N3N1N3N1O2M2O2M3N001N2O1N2O2N4K2O1N2O2N001N10000O1000000O10001O1ON10100001O1N2O2M2O1N2O1N2O1O1N3N1N^H"}, "label": "the sandwich standing next to the man ' s thumb"}]} {"id": 339579, "image": "COCO_train2014_000000339579.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half a sandwich with lettuce and bacon on very white colored bread\"."}], "task": "refering", "answer_template": "The \"half a sandwich with lettuce and bacon on very white colored bread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 106, 107, 108, 109, 110, 120, 121, 122, 123, 124, 134, 135, 136, 137, 148, 149], "bbox": [0.5788250652741515, 0.256703125, 0.8889556135770236, 0.4720625], "iscrowd": 0, "area": 10333.957300000002, "rle": {"size": [640, 383], "counts": "[oZ4_1oa0c0M3M3M3L4M3M3L4M3M3L4M3M3M3L4M3M3L4N2O1N10O010O0010ON3N1N3N1O20O00010O010O01O010O0010O0100O2N100O100O1O101N1O100O100O1O2O0O2O1N2O1N2O1N2O1N2O0O2O0O100O100O100O2O0O100O1O100O100O101M2O1O2N3L5L3M3L2O2N1O2M2O2N1N3N1O2N1N3N1O1N5L5K5K5J6K[ej0"}, "label": "half a sandwich with lettuce and bacon on very white colored bread"}]} {"id": 339579, "image": "COCO_train2014_000000339579.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hero sandwich on a white plate with lettuce and cold cuts on it\"."}], "task": "refering", "answer_template": "The \"a hero sandwich on a white plate with lettuce and cold cuts on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 106, 107, 108, 109, 110, 120, 121, 122, 123, 124, 134, 135, 136, 137, 148, 149], "bbox": [0.5788250652741515, 0.256703125, 0.8889556135770236, 0.4720625], "iscrowd": 0, "area": 10333.957300000002, "rle": {"size": [640, 383], "counts": "[oZ4_1oa0c0M3M3M3L4M3M3L4M3M3L4M3M3M3L4M3M3L4N2O1N10O010O0010ON3N1N3N1O20O00010O010O01O010O0010O0100O2N100O100O1O101N1O100O100O1O2O0O2O1N2O1N2O1N2O1N2O0O2O0O100O100O100O2O0O100O1O100O100O101M2O1O2N3L5L3M3L2O2N1O2M2O2N1N3N1O2N1N3N1O1N5L5K5K5J6K[ej0"}, "label": "a hero sandwich on a white plate with lettuce and cold cuts on it"}]} {"id": 339579, "image": "COCO_train2014_000000339579.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baked chicken on the plate\"."}], "task": "refering", "answer_template": "The \"the baked chicken on the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [160, 161, 162, 173, 174, 175, 176, 186, 187, 188, 189, 190, 200, 201, 202, 203, 204, 214, 215, 216, 217, 218, 228, 229, 230, 231, 232, 242, 243], "bbox": [0.2973890339425588, 0.48267187500000003, 0.6432114882506528, 0.767234375], "iscrowd": 0, "area": 16766.63015, "rle": {"size": [640, 383], "counts": "\\fW22lc04L4L4L4L4L4L4L4L4L4K5B>C=B>B=E:N2O0O2O1N2O1N1O2O1N2O0O10O0010O010O100O2O0O10000O101N100O101N2O1N3N1N2O1O1N3N1O1N2O2N1N200O101N100O101N10O0100000O10000O1000000O10000O1000000O01000O1000000O10000O1000000O10000O1O1O1O1N3N1O1O2N2N3L3N2N2N3M2N2M4M2N2N3M2M6K5K6B>ZOe0ZOg0ZO^Ye2"}, "label": "the baked chicken on the plate"}]} {"id": 339579, "image": "COCO_train2014_000000339579.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of meat with mushrooms on top\"."}], "task": "refering", "answer_template": "The \"a piece of meat with mushrooms on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [160, 161, 162, 173, 174, 175, 176, 186, 187, 188, 189, 190, 200, 201, 202, 203, 204, 214, 215, 216, 217, 218, 228, 229, 230, 231, 232, 242, 243], "bbox": [0.2973890339425588, 0.48267187500000003, 0.6432114882506528, 0.767234375], "iscrowd": 0, "area": 16766.63015, "rle": {"size": [640, 383], "counts": "\\fW22lc04L4L4L4L4L4L4L4L4L4K5B>C=B>B=E:N2O0O2O1N2O1N1O2O1N2O0O10O0010O010O100O2O0O10000O101N100O101N2O1N3N1N2O1O1N3N1O1N2O2N1N200O101N100O101N10O0100000O10000O1000000O10000O1000000O01000O1000000O10000O1000000O10000O1O1O1O1N3N1O1O2N2N3L3N2N2N3M2N2M4M2N2N3M2M6K5K6B>ZOe0ZOg0ZO^Ye2"}, "label": "a piece of meat with mushrooms on top"}]} {"id": 339579, "image": "COCO_train2014_000000339579.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue table cloth\"."}], "task": "refering", "answer_template": "The \"a blue table cloth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 212, 213, 214, 215, 216, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321], "bbox": [0.01877284595300261, 0.683140625, 1.0, 0.9887656249999999], "iscrowd": 0, "area": 68953.60224999998, "rle": {"size": [640, 383], "counts": "[o4>Vb0\\1cN]1dN\\1dN\\1K50000000000001O00000000000000000000000000000000000000000000001O0000000O100000000000000000000000000000000000001O000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000001O000000000000000000000000000000000000000O1000001O00000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000001O000000000000000000000O1000000000000000000000001O00000000000000000000000000000000000000000000001O00002N2N2N2N1O2N2N2N2N2N2N1O2N2N2N2N2N2N1O2N2N2N2N2N2N2N1OZ@"}, "label": "a blue table cloth"}]} {"id": 339579, "image": "COCO_train2014_000000339579.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the piece of sandwich that is to the left of two other sandwich pieces\"."}], "task": "refering", "answer_template": "The \"the piece of sandwich that is to the left of two other sandwich pieces\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 90, 91, 92, 93, 104, 105, 106, 118, 119, 120, 131, 132, 133, 134, 146, 147, 148, 161], "bbox": [0.4166318537859008, 0.241875, 0.6588250652741514, 0.482890625], "iscrowd": 0, "area": 8442.00715, "rle": {"size": [640, 383], "counts": "_XT3:[c0=J5L4L4L4K6K4L4L4K5J6J7I6I7J6J6K6J5L4L4K5L5J4M3N200O10000O10000O10000O10000O10001N10000O10000O10000O10000O10000O10000O_No@TNR?f1ZATNf>e1gATNY>f1UBSNk=f1bBSN_=f1nBTNQ=f1\\CSNe in this image.", "objects": [{"patches": [77, 78, 90, 91, 92, 93, 104, 105, 106, 118, 119, 120, 131, 132, 133, 134, 146, 147, 148, 161], "bbox": [0.4166318537859008, 0.241875, 0.6588250652741514, 0.482890625], "iscrowd": 0, "area": 8442.00715, "rle": {"size": [640, 383], "counts": "_XT3:[c0=J5L4L4L4K6K4L4L4K5J6J7I6I7J6J6K6J5L4L4K5L5J4M3N200O10000O10000O10000O10000O10001N10000O10000O10000O10000O10000O10000O_No@TNR?f1ZATNf>e1gATNY>f1UBSNk=f1bBSN_=f1nBTNQ=f1\\CSNe in this image.", "objects": [{"patches": [237, 238, 252, 253, 254, 267, 268, 269, 281, 282, 283, 284, 297, 298, 299, 313, 314, 328, 329, 343, 344], "bbox": [0.7882159624413145, 0.66125, 0.9994600938967135, 0.9977031249999999], "iscrowd": 0, "area": 12984.402750000003, "rle": {"size": [640, 426], "counts": "W`b65ec06J7H8I6PN]O]@j0^?AT@f0l?^Oj_Oh0U`0ZOd_Ok0]`0VO[_OQ1e`0T11K5I7H8I7H8H83M3M4L3M6J7I6J6J5J6K4L3M3M1O1O1O1Oa1_N9G2N2N2N2N2N2N2M3N5K8H7I0000000001O0000001O2N2N2N1O2N2N2N3M3M2N3M3M2N3M2N1O2N2N2N1O2N2N1O2N2N2N1O2O1N1O2Nn0ROn2RMn\\O"}, "label": "the bench with no one sitting on it"}]} {"id": 568962, "image": "COCO_train2014_000000568962.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"unoccupied bench\"."}], "task": "refering", "answer_template": "The \"unoccupied bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [237, 238, 252, 253, 254, 267, 268, 269, 281, 282, 283, 284, 297, 298, 299, 313, 314, 328, 329, 343, 344], "bbox": [0.7882159624413145, 0.66125, 0.9994600938967135, 0.9977031249999999], "iscrowd": 0, "area": 12984.402750000003, "rle": {"size": [640, 426], "counts": "W`b65ec06J7H8I6PN]O]@j0^?AT@f0l?^Oj_Oh0U`0ZOd_Ok0]`0VO[_OQ1e`0T11K5I7H8I7H8H83M3M4L3M6J7I6J6J5J6K4L3M3M1O1O1O1Oa1_N9G2N2N2N2N2N2N2M3N5K8H7I0000000001O0000001O2N2N2N1O2N2N2N3M3M2N3M3M2N3M2N1O2N2N2N1O2N2N1O2N2N2N1O2O1N1O2Nn0ROn2RMn\\O"}, "label": "unoccupied bench"}]} {"id": 364169, "image": "COCO_train2014_000000364169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person standing near to a girl only legs are visible\"."}], "task": "refering", "answer_template": "The \"a person standing near to a girl only legs are visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 40, 41, 42, 43, 64, 65, 88, 111], "bbox": [0.72115625, 0.0022482435597189696, 0.9040781250000001, 0.2876346604215457], "iscrowd": 0, "area": 6705.3569, "rle": {"size": [427, 640], "counts": "ldP6`0k<6J1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O001O001O1O001O001O1O001O001O001O1O001O1O001O1O001O001N2O1O2N3M3M2N3M2N2N2N2N3M2N3M4nESM^9P3YFVMg9X3O1O1O001O001O001O001OJ6E;J6N2O1N2N2O1N2K5J6J6J6J6K5J6J6J6J6J6J6L4N2N2N2N200O1O1O2O0O1OYki0"}, "label": "a person standing near to a girl only legs are visible"}]} {"id": 364169, "image": "COCO_train2014_000000364169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing jeans and a sweater standing near a child\"."}], "task": "refering", "answer_template": "The \"a person wearing jeans and a sweater standing near a child\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 20, 40, 41, 42, 43, 64, 65, 88, 111], "bbox": [0.72115625, 0.0022482435597189696, 0.9040781250000001, 0.2876346604215457], "iscrowd": 0, "area": 6705.3569, "rle": {"size": [427, 640], "counts": "ldP6`0k<6J1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O001O001O1O001O001O1O001O001O001O1O001O1O001O1O001O001N2O1O2N3M3M2N3M2N2N2N2N3M2N3M4nESM^9P3YFVMg9X3O1O1O001O001O001O001OJ6E;J6N2O1N2N2O1N2K5J6J6J6J6K5J6J6J6J6J6J6L4N2N2N2N200O1O1O2O0O1OYki0"}, "label": "a person wearing jeans and a sweater standing near a child"}]} {"id": 364169, "image": "COCO_train2014_000000364169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little girl on a bench playing on her laptop\"."}], "task": "refering", "answer_template": "The \"little girl on a bench playing on her laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 37, 38, 39, 40, 41, 60, 61, 62, 63, 64, 83, 84, 85, 86, 87, 88, 107, 108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 268, 269, 270, 291, 292], "bbox": [0.5884375000000001, 0.04199063231850117, 0.882234375, 0.8529742388758783], "iscrowd": 0, "area": 32165.420550000003, "rle": {"size": [427, 640], "counts": "][m41Z=00100O1O1O001O1O1O010O1O1SMHeH9Y7JeH6Y7MfH4X7OfH2X70hH0V72iHNU76iHKU77jHJT79jHHS7;lHFR7=lHCS7?mHAP7b0oH_Oo6d0oH]Oo6e0PI\\Om6h0QIXOn6j0QIWOm6k0RIVOm0\\Ol2a1VLSOm0^Oi2b1YLPOn0@g2b1YLoNo0Ae2b1[LnNn0Ce2a1[LmNo0Dc2a1]LlNo0Ea2a1_LjNP1G_2a1`LhNP1IfNEo1l1ZNgNo0KeNKj1e1`NfNP1LbN1g1_1fNdNQ1N^N1k1^1eNdNQ1O\\N0m1`1dNbNR10YNOR2`1bNbNQ12WNNU2a1aN`NR13UNMX2a1aN^NR16QNM[2b1`N\\NS16oMM_2b1^N\\NS17mMLa2c1^N[NR19kMKd2d1]NYNS1:hMKh2c1\\NXNT1f1`04[NWNT1e1a05[NVNS1e1a07[NUNR1d1d07ZNUNR1d1c09ZNTNR1c1d0:YNTNR1b1d0XNRNP1a1h0>WNRNo0a1i0?WNQNn0`1k0a0WNoMl0a1m0`0WNoMl0a1l0b0WNnMk0a1n0b0VNnMj0`1P1d0UNmMi0`1R1d0TNmMh0`1S1e0TNkMh0`1U1f0RNkMg0`1V1g0RNjMf0_1Y1h0PNjMf0^1[1h0nMjMd0a1^1f0mMjMb0a1c1e0kMjM>d1h1b0iMkM;f1l1a0gMjM:f1Q2a0cMiM9i1T2`0aMhM5l1\\2=]Mn1c2TN[Ml1f2VNWMi1k2XNTMg1l2[NRMd1P3]NoL\\OLbNU3S2nL[ONaNU3T2mLZO0_NT3X2lLWO2`NS3Y2jLWO5^NQ3\\2jLTO7^NQ3^2gLTO9]NQ3_2fLRO<\\No2d2dLoN>[NP3f2aLoNa0WNQ3j2^LmNc0VNQ3n2\\LkNd0UNR3V6QMfIR3Y6oLdIS3[6PMaIS3^6nL_IT3a6mL[IW3c6lLVIZ3i6gLRI]3n6eLkHa3T7`LgHe3X7W101O1O001O1O001O001N2O001O1O001O1O001O1O001O001O1O001O1O001N3N3M2N2N2M4M2N2N2M4M2N2N2M4M2N2N2M4M2N2M3N3M2N2M3N3M2N2M3N3M2N2M3N3K4L4L4K5L4L4L4L4K5L4L4LT1lNi0VO7J5K6JjQo0"}, "label": "little girl on a bench playing on her laptop"}]} {"id": 364169, "image": "COCO_train2014_000000364169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"young girl wearing a black vest sitting on a wooden box typing on a green laptop\"."}], "task": "refering", "answer_template": "The \"young girl wearing a black vest sitting on a wooden box typing on a green laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 37, 38, 39, 40, 41, 60, 61, 62, 63, 64, 83, 84, 85, 86, 87, 88, 107, 108, 109, 110, 111, 128, 129, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 268, 269, 270, 291, 292], "bbox": [0.5884375000000001, 0.04199063231850117, 0.882234375, 0.8529742388758783], "iscrowd": 0, "area": 32165.420550000003, "rle": {"size": [427, 640], "counts": "][m41Z=00100O1O1O001O1O1O010O1O1SMHeH9Y7JeH6Y7MfH4X7OfH2X70hH0V72iHNU76iHKU77jHJT79jHHS7;lHFR7=lHCS7?mHAP7b0oH_Oo6d0oH]Oo6e0PI\\Om6h0QIXOn6j0QIWOm6k0RIVOm0\\Ol2a1VLSOm0^Oi2b1YLPOn0@g2b1YLoNo0Ae2b1[LnNn0Ce2a1[LmNo0Dc2a1]LlNo0Ea2a1_LjNP1G_2a1`LhNP1IfNEo1l1ZNgNo0KeNKj1e1`NfNP1LbN1g1_1fNdNQ1N^N1k1^1eNdNQ1O\\N0m1`1dNbNR10YNOR2`1bNbNQ12WNNU2a1aN`NR13UNMX2a1aN^NR16QNM[2b1`N\\NS16oMM_2b1^N\\NS17mMLa2c1^N[NR19kMKd2d1]NYNS1:hMKh2c1\\NXNT1f1`04[NWNT1e1a05[NVNS1e1a07[NUNR1d1d07ZNUNR1d1c09ZNTNR1c1d0:YNTNR1b1d0XNRNP1a1h0>WNRNo0a1i0?WNQNn0`1k0a0WNoMl0a1m0`0WNoMl0a1l0b0WNnMk0a1n0b0VNnMj0`1P1d0UNmMi0`1R1d0TNmMh0`1S1e0TNkMh0`1U1f0RNkMg0`1V1g0RNjMf0_1Y1h0PNjMf0^1[1h0nMjMd0a1^1f0mMjMb0a1c1e0kMjM>d1h1b0iMkM;f1l1a0gMjM:f1Q2a0cMiM9i1T2`0aMhM5l1\\2=]Mn1c2TN[Ml1f2VNWMi1k2XNTMg1l2[NRMd1P3]NoL\\OLbNU3S2nL[ONaNU3T2mLZO0_NT3X2lLWO2`NS3Y2jLWO5^NQ3\\2jLTO7^NQ3^2gLTO9]NQ3_2fLRO<\\No2d2dLoN>[NP3f2aLoNa0WNQ3j2^LmNc0VNQ3n2\\LkNd0UNR3V6QMfIR3Y6oLdIS3[6PMaIS3^6nL_IT3a6mL[IW3c6lLVIZ3i6gLRI]3n6eLkHa3T7`LgHe3X7W101O1O001O1O001O001N2O001O1O001O1O001O1O001O001O1O001O1O001N3N3M2N2N2M4M2N2N2M4M2N2N2M4M2N2N2M4M2N2M3N3M2N2M3N3M2N2M3N3M2N2M3N3K4L4L4K5L4L4L4L4K5L4L4LT1lNi0VO7J5K6JjQo0"}, "label": "young girl wearing a black vest sitting on a wooden box typing on a green laptop"}]} {"id": 192137, "image": "COCO_train2014_000000192137.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plane with a red and white stripes on the tail\"."}], "task": "refering", "answer_template": "The \"a plane with a red and white stripes on the tail\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 193], "bbox": [0.142203125, 0.46106936416184974, 0.8262656250000001, 0.6765028901734104], "iscrowd": 0, "area": 13882.020449999996, "rle": {"size": [346, 640], "counts": "[Yo01i:000O1000000O100000001N100000000000000000001O000000000O10000000001O000000000000000000001O000000000O10000000001O000000000000000000001O0000000O1000000000001O00000000000cNC]H=c7C]H=c7C]H=c7C]H>b7B^H>b7B^H>b7B^H>b7B^H>b7B^H>b7C\\H>d7B\\H>d7B\\H>d7B\\H>d7B\\H>d7B\\H?c7A]H?c7A]H?c7A]H?c7B\\H>e7A[H?e7A[H?e7A[H?e7A[H?e7A[H?e7A[H?e7A[H`0d7@\\H`0d7A[H?e7A[H?e7A[H?e7AZH`0f7@ZH`0f7@ZH`0f7BXH>h7DVH in this image.", "objects": [{"patches": [121, 122, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 193], "bbox": [0.142203125, 0.46106936416184974, 0.8262656250000001, 0.6765028901734104], "iscrowd": 0, "area": 13882.020449999996, "rle": {"size": [346, 640], "counts": "[Yo01i:000O1000000O100000001N100000000000000000001O000000000O10000000001O000000000000000000001O000000000O10000000001O000000000000000000001O0000000O1000000000001O00000000000cNC]H=c7C]H=c7C]H=c7C]H>b7B^H>b7B^H>b7B^H>b7B^H>b7B^H>b7C\\H>d7B\\H>d7B\\H>d7B\\H>d7B\\H>d7B\\H?c7A]H?c7A]H?c7A]H?c7B\\H>e7A[H?e7A[H?e7A[H?e7A[H?e7A[H?e7A[H?e7A[H`0d7@\\H`0d7A[H?e7A[H?e7A[H?e7AZH`0f7@ZH`0f7@ZH`0f7BXH>h7DVH in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303], "bbox": [0.0026718749999999998, 0.028241469816272964, 0.23625000000000002, 0.9430446194225722], "iscrowd": 0, "area": 49161.19195, "rle": {"size": [381, 640], "counts": "nP1Z1i8k1UNl1TNk1UNk1\\Oe0L3M4L3M3M4K4M4L3M3M4L3O2M2O0000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O1000000000000000000O10000000000000000000000001O0000000000001O0000000000001O0000000000001O0O10000000001O000000000O1N3N1N2N4M2M3M4L3N2M4L3N2M4L3N2M4QNn1ZMf2YMfoe5"}, "label": "a part of a cellphone to the far left"}]} {"id": 118413, "image": "COCO_train2014_000000118413.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front portion of an old , grey cell phone\"."}], "task": "refering", "answer_template": "The \"the front portion of an old , grey cell phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 28, 29, 30, 31, 32, 33, 34, 51, 52, 53, 54, 55, 56, 57, 74, 75, 76, 77, 78, 79, 80, 97, 98, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309], "bbox": [0.2180625, 0.02246719160104987, 0.48159375000000004, 0.9775328083989501], "iscrowd": 0, "area": 55078.06025, "rle": {"size": [381, 640], "counts": "Q^d14e;4M3L4L4L5L3L4L4ZOg0fNY1hNX1gNY1gNY1hNY1fNY1gNY1L401O000O10001O0000001N100000001O0O10001O0000000O2O0000001O0O1000001O00000O10000000000O100000000000000000000000000001O00000000000000001O0000000000000000001O00000000000000001O00000O100000000O101O000O1000000O10001O000O1000000O1000001O0O1000000O1000001N100000000O1000000O2O0000000O101O001N101O001N101O001SJPfk3"}, "label": "the front portion of an old , grey cell phone"}]} {"id": 118413, "image": "COCO_train2014_000000118413.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"motorola c350 phone\"."}], "task": "refering", "answer_template": "The \"motorola c350 phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 28, 29, 30, 31, 32, 33, 34, 51, 52, 53, 54, 55, 56, 57, 74, 75, 76, 77, 78, 79, 80, 97, 98, 99, 100, 101, 102, 103, 120, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309], "bbox": [0.2180625, 0.02246719160104987, 0.48159375000000004, 0.9775328083989501], "iscrowd": 0, "area": 55078.06025, "rle": {"size": [381, 640], "counts": "Q^d14e;4M3L4L4L5L3L4L4ZOg0fNY1hNX1gNY1gNY1hNY1fNY1gNY1L401O000O10001O0000001N100000001O0O10001O0000000O2O0000001O0O1000001O00000O10000000000O100000000000000000000000000001O00000000000000001O0000000000000000001O00000000000000001O00000O100000000O101O000O1000000O10001O000O1000000O1000001O0O1000000O1000001N100000000O1000000O2O0000000O101O001N101O001N101O001SJPfk3"}, "label": "motorola c350 phone"}]} {"id": 102030, "image": "COCO_train2014_000000102030.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing blue jeans , a brown vest , a tan shirt and hat , holding the face of a sheep that has just been shorn\"."}], "task": "refering", "answer_template": "The \"a man wearing blue jeans , a brown vest , a tan shirt and hat , holding the face of a sheep that has just been shorn\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 59, 60, 61, 62, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 109, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 180, 181, 182, 183, 197, 198, 199, 200, 216, 217, 233, 234, 250, 251, 267, 268, 269, 284, 285, 286, 302], "bbox": [0.46039583333333334, 0.12465625, 0.8340833333333334, 0.7591875], "iscrowd": 0, "area": 35338.80509999999, "rle": {"size": [640, 480], "counts": "aXZ4S1jb0P2QN3M4L3M4L4K4M4L3M4L3L5L4L3M4M2N3M3M2N3M2N3M2N3M3M2N3M2N1O00010O01O0010O01O0010O01O00010O001O010O001O0100O100O100O10O0100O10000010O0001O010O000010O00QCZJd;g5ZD\\Jd;d5[D`Jc;_5[DcJe;^5YDcJg;^5WDcJi;^5VDbJk;]5UDdJj;]5UDcJk;^5TDbJl;_5SDaJn;_5QDaJo;`5PDaJo;_5QDaJo;`5PD`JQ<`5nC`JR in this image.", "objects": [{"patches": [44, 59, 60, 61, 62, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 109, 110, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 131, 132, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 180, 181, 182, 183, 197, 198, 199, 200, 216, 217, 233, 234, 250, 251, 267, 268, 269, 284, 285, 286, 302], "bbox": [0.46039583333333334, 0.12465625, 0.8340833333333334, 0.7591875], "iscrowd": 0, "area": 35338.80509999999, "rle": {"size": [640, 480], "counts": "aXZ4S1jb0P2QN3M4L3M4L4K4M4L3M4L3L5L4L3M4M2N3M3M2N3M2N3M2N3M3M2N3M2N1O00010O01O0010O01O0010O01O00010O001O010O001O0100O100O100O10O0100O10000010O0001O010O000010O00QCZJd;g5ZD\\Jd;d5[D`Jc;_5[DcJe;^5YDcJg;^5WDcJi;^5VDbJk;]5UDdJj;]5UDcJk;^5TDbJl;_5SDaJn;_5QDaJo;`5PDaJo;_5QDaJo;`5PD`JQ<`5nC`JR in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 259, 260, 261, 262, 276, 277, 278, 279, 282, 283, 284, 285, 300, 301, 302], "bbox": [0.0185, 0.2667083333333334, 0.728140625, 0.8182708333333333], "iscrowd": 0, "area": 57768.79220000001, "rle": {"size": [480, 640], "counts": "Pn5?n=f0ZOf0POo0SOn0I7I6J7I5K5K5M4M2N2M3N2_FoJP9b5N2M3N2N2N2M3N2N2N2N2N2O0O1O2N1O1O101N1O1O1O2N10O01O001O001O010O001O001O0100O010O010000O10000O100O01000O10000O100O10mMZHlLf7R3cHgL\\7Y3mH_LS7_3VIYLj6g3_IQLa6m3iIkKW6T4QJeKn5Z4\\J^Kd5a4eJWK[5g4_201N101O10O01000O10O01000000O01000O10O10O1000O010000O10O10O1000O010000O01000O1000O010000O01000O10O10O10000O01000O10O1000000O100000000O1000000O1000000O10oG[Kg5e4UJ`Kj5`4QJfKn5Z4mIlKR6S4jISLU6m3dI[L[6e3^IcLa6]3XIkLg6T3TISMk6m2nH[MP7f2hHcMW7]2`HmM_7j40000000001O00000000000000001O00000000000000001O00000000O1000000O10000O1O1O1O1O1O1O1O1O100O1O1O1001O001O00001O00001O001O000000000000RLnHNS71RIJn66UIFl6:WICi6=ZI@f6`0]I]Oc6c0`IZO`6f0cIWO]6i0fITOZ6l0iIPOX6P1kImNU6S1nIgNU6Y1oIiMk6W2XIkLc7U3T2000000000000000000000000000000000000001O00001O001O001O01O100O100O10O01O100O100O100O1O100O100O1N2N2M4M3N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N1O1O2N1O1O1O2N1O1O1O1N3N1O1O1O2N1O1O1O1O00000000000000000O10000000000000001O000000000000000O10000000000000000000000000000O0100000O1000000O1000O10O1000000O1000000O10O1000O1000000O100000O01000000O1000000O10O1000O100000O010000O100O10O010000O100O100O100O100O10000O100O1Lf[a2"}, "label": "an elephant standing at the edge of some water and tapping another elephant with it ' s trunk"}]} {"id": 298639, "image": "COCO_train2014_000000298639.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant standing next to water hole with trunk on baby elephant\"."}], "task": "refering", "answer_template": "The \"elephant standing next to water hole with trunk on baby elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 259, 260, 261, 262, 276, 277, 278, 279, 282, 283, 284, 285, 300, 301, 302], "bbox": [0.0185, 0.2667083333333334, 0.728140625, 0.8182708333333333], "iscrowd": 0, "area": 57768.79220000001, "rle": {"size": [480, 640], "counts": "Pn5?n=f0ZOf0POo0SOn0I7I6J7I5K5K5M4M2N2M3N2_FoJP9b5N2M3N2N2N2M3N2N2N2N2N2O0O1O2N1O1O101N1O1O1O2N10O01O001O001O010O001O001O0100O010O010000O10000O100O01000O10000O100O10mMZHlLf7R3cHgL\\7Y3mH_LS7_3VIYLj6g3_IQLa6m3iIkKW6T4QJeKn5Z4\\J^Kd5a4eJWK[5g4_201N101O10O01000O10O01000000O01000O10O10O1000O010000O10O10O1000O010000O01000O1000O010000O01000O10O10O10000O01000O10O1000000O100000000O1000000O1000000O10oG[Kg5e4UJ`Kj5`4QJfKn5Z4mIlKR6S4jISLU6m3dI[L[6e3^IcLa6]3XIkLg6T3TISMk6m2nH[MP7f2hHcMW7]2`HmM_7j40000000001O00000000000000001O00000000000000001O00000000O1000000O10000O1O1O1O1O1O1O1O1O100O1O1O1001O001O00001O00001O001O000000000000RLnHNS71RIJn66UIFl6:WICi6=ZI@f6`0]I]Oc6c0`IZO`6f0cIWO]6i0fITOZ6l0iIPOX6P1kImNU6S1nIgNU6Y1oIiMk6W2XIkLc7U3T2000000000000000000000000000000000000001O00001O001O001O01O100O100O10O01O100O100O100O1O100O100O1N2N2M4M3N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N1O1O2N1O1O1O2N1O1O1O1N3N1O1O1O2N1O1O1O1O00000000000000000O10000000000000001O000000000000000O10000000000000000000000000000O0100000O1000000O1000O10O1000000O1000000O10O1000O1000000O100000O01000000O1000000O10O1000O100000O010000O100O10O010000O100O100O100O100O10000O100O1Lf[a2"}, "label": "elephant standing next to water hole with trunk on baby elephant"}]} {"id": 306837, "image": "COCO_train2014_000000306837.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra standing on the left corner without eating grass\"."}], "task": "refering", "answer_template": "The \"zebra standing on the left corner without eating grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 53, 54, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 92, 93, 94, 95, 96, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 190, 191, 192, 193, 194, 207, 208, 209, 215, 216, 217, 230, 231, 232, 239, 253, 254, 255, 276, 277, 299, 300, 301, 322, 323, 324, 345], "bbox": [0.0, 0.04272916666666667, 0.456375, 0.9196875], "iscrowd": 0, "area": 59575.48624999999, "rle": {"size": [480, 640], "counts": "n0j in this image.", "objects": [{"patches": [2, 3, 23, 24, 25, 26, 27, 28, 29, 46, 47, 48, 49, 50, 51, 52, 53, 54, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 92, 93, 94, 95, 96, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 190, 191, 192, 193, 194, 207, 208, 209, 215, 216, 217, 230, 231, 232, 239, 253, 254, 255, 276, 277, 299, 300, 301, 322, 323, 324, 345], "bbox": [0.0, 0.04272916666666667, 0.456375, 0.9196875], "iscrowd": 0, "area": 59575.48624999999, "rle": {"size": [480, 640], "counts": "n0j in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 53, 54, 55, 56, 57, 58, 59, 60, 78, 79, 80, 81, 82, 83, 84, 85, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 124, 125, 126, 127, 128, 129, 130, 131, 132, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 165, 166, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 189, 190, 194, 195, 196, 198, 199, 200, 201, 202, 203, 212, 213, 215, 218, 219, 221, 222, 223, 224, 225, 226, 234, 235, 236, 238, 241, 242, 245, 246, 247, 248, 257, 258, 259, 260, 261, 264, 265, 268, 269, 270, 271, 281, 282, 283, 284, 287, 288, 291, 292, 293, 294, 295, 304, 305, 306, 310, 311, 315, 316, 317, 318, 327, 328, 329, 333, 334, 339, 340], "bbox": [0.19128125, 0.0046875, 0.855703125, 0.8771666666666668], "iscrowd": 0, "area": 79570.27899999998, "rle": {"size": [480, 640], "counts": "lei12m>100O1O2N1O1O1O1gFGk39TLKh37VLLg35YLMd34[L0a31^L2PNBT2=L3XM:h2DO5`LQ1\\3kN46iKh1n3SN89cKg1U4QN8:^Kg1Z4PN8;ZKf1^4QN7cIAY6>lICP68YJGc56dJKX51PKNm4M[K4`4IhK6S4ISL8d3N^L3Y32kLMl2:VMGa2>cMAU2e0mM\\Oj1i0ZNVOd1j0^NWOc1e0`NZOb1a0aN@`1g7FXH:d7LZH5d6m0[ISOa6S1]ImN^6Y1aIgN[6^1dIcNW6c1gI]NW6f1hIZNU6k1iIUNU6n1jISNS6Q2kIoMR6U2lIlMR6W2mIiMP6[2oIeMn5_2QJbMj5c2UJ]Mh5g2WJYMf5S6K5L3M4K4O20O01000O010O100O010O10O10O100O010O10O10O10O010O001O00100O001O1O100O010001O0O101O00001N2O2N2N1N3N2N2N2M2O001O0aHYJd5h5WJ\\Jh5e5TJ_Jl5b5nIcJR6^5iIfJW6Z5eIkJY6W5bImJ^6S5]IRKc6o4XIUKh6k4TIYKk6i4oH\\KQ7Z600O100O100O100O1O101N101N2O1O1O001O01O00000001O01O000000010O000000001O01O000001O0001O000001O000001O01O000000010O00000O1O2N1N2O1O1O2M2O2N1O1O2M2O2N1O1O2M2O2N1O2N1N2O2N1O2M2O1O2N1O2M2O1O2N2dKPHd1R8[NlHe0W7YOoHa0S7^OQI=Q7AUI9m6FWI5l6HZI1i6N[ILh62]IHf67_IBd6 in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 53, 54, 55, 56, 57, 58, 59, 60, 78, 79, 80, 81, 82, 83, 84, 85, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 124, 125, 126, 127, 128, 129, 130, 131, 132, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 165, 166, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 189, 190, 194, 195, 196, 198, 199, 200, 201, 202, 203, 212, 213, 215, 218, 219, 221, 222, 223, 224, 225, 226, 234, 235, 236, 238, 241, 242, 245, 246, 247, 248, 257, 258, 259, 260, 261, 264, 265, 268, 269, 270, 271, 281, 282, 283, 284, 287, 288, 291, 292, 293, 294, 295, 304, 305, 306, 310, 311, 315, 316, 317, 318, 327, 328, 329, 333, 334, 339, 340], "bbox": [0.19128125, 0.0046875, 0.855703125, 0.8771666666666668], "iscrowd": 0, "area": 79570.27899999998, "rle": {"size": [480, 640], "counts": "lei12m>100O1O2N1O1O1O1gFGk39TLKh37VLLg35YLMd34[L0a31^L2PNBT2=L3XM:h2DO5`LQ1\\3kN46iKh1n3SN89cKg1U4QN8:^Kg1Z4PN8;ZKf1^4QN7cIAY6>lICP68YJGc56dJKX51PKNm4M[K4`4IhK6S4ISL8d3N^L3Y32kLMl2:VMGa2>cMAU2e0mM\\Oj1i0ZNVOd1j0^NWOc1e0`NZOb1a0aN@`1g7FXH:d7LZH5d6m0[ISOa6S1]ImN^6Y1aIgN[6^1dIcNW6c1gI]NW6f1hIZNU6k1iIUNU6n1jISNS6Q2kIoMR6U2lIlMR6W2mIiMP6[2oIeMn5_2QJbMj5c2UJ]Mh5g2WJYMf5S6K5L3M4K4O20O01000O010O100O010O10O10O100O010O10O10O10O010O001O00100O001O1O100O010001O0O101O00001N2O2N2N1N3N2N2N2M2O001O0aHYJd5h5WJ\\Jh5e5TJ_Jl5b5nIcJR6^5iIfJW6Z5eIkJY6W5bImJ^6S5]IRKc6o4XIUKh6k4TIYKk6i4oH\\KQ7Z600O100O100O100O1O101N101N2O1O1O001O01O00000001O01O000000010O000000001O01O000001O0001O000001O000001O01O000000010O00000O1O2N1N2O1O1O2M2O2N1O1O2M2O2N1O1O2M2O2N1O2N1N2O2N1O2M2O1O2N1O2M2O1O2N2dKPHd1R8[NlHe0W7YOoHa0S7^OQI=Q7AUI9m6FWI5l6HZI1i6N[ILh62]IHf67_IBd6 in this image.", "objects": [{"patches": [78, 79, 80, 81, 82, 83, 100, 101, 102, 103, 104, 105, 106, 107, 123, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199], "bbox": [0.374734375, 0.21014018691588784, 0.6903750000000001, 0.5894392523364486], "iscrowd": 0, "area": 25211.786900000006, "rle": {"size": [428, 640], "counts": "a^T33U=4K6K4L4L5N1O1O2N100O2N100O2O0O1WFnNX7R1eHQO[7P1aHSO^7n0_HUOa7k0]HWOb7k0ZHXOf7h0WH[Oh7f0VH\\Oj7d0TH^Ok7d0RH]Oo7c0oG_Om7e0QH]Ok7i0QHYOl7k0QHWOk7o0QHSOk7R1RHPOk7U1QHmNo7T1lGPOT8R1gGQOY8P1cGSO]8o0]GUOc8l0YGWOg8g20000000000000001O0001O000000000000000001O00000000000000000000001O0000001O00001O0000001O01O01O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00010O00001O000000O10000O1000000O10000O1000000O1000000O100000000O1000001O0O1000002N1N2O1O2N1N2O2N1O1O1N3N1O1O1O0O2O1O1O001N2O1O1O0O2O1O1O001N2O1O1O001N2O1O001O1N2O1aMmEc1S:XN^FS1l9hN_2J6JTdb2"}, "label": "the poodle behind the the other dog next to the dark brown dog"}]} {"id": 331419, "image": "COCO_train2014_000000331419.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog back and in between two other dogs looking out side\"."}], "task": "refering", "answer_template": "The \"a dog back and in between two other dogs looking out side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 81, 82, 83, 100, 101, 102, 103, 104, 105, 106, 107, 123, 124, 125, 126, 127, 128, 129, 130, 147, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199], "bbox": [0.374734375, 0.21014018691588784, 0.6903750000000001, 0.5894392523364486], "iscrowd": 0, "area": 25211.786900000006, "rle": {"size": [428, 640], "counts": "a^T33U=4K6K4L4L5N1O1O2N100O2N100O2O0O1WFnNX7R1eHQO[7P1aHSO^7n0_HUOa7k0]HWOb7k0ZHXOf7h0WH[Oh7f0VH\\Oj7d0TH^Ok7d0RH]Oo7c0oG_Om7e0QH]Ok7i0QHYOl7k0QHWOk7o0QHSOk7R1RHPOk7U1QHmNo7T1lGPOT8R1gGQOY8P1cGSO]8o0]GUOc8l0YGWOg8g20000000000000001O0001O000000000000000001O00000000000000000000001O0000001O00001O0000001O01O01O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00010O00001O000000O10000O1000000O10000O1000000O1000000O100000000O1000001O0O1000002N1N2O1O2N1N2O2N1O1O1N3N1O1O1O0O2O1O1O001N2O1O1O0O2O1O1O001N2O1O1O001N2O1O001O1N2O1aMmEc1S:XN^FS1l9hN_2J6JTdb2"}, "label": "a dog back and in between two other dogs looking out side"}]} {"id": 331419, "image": "COCO_train2014_000000331419.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark brown dog beside two other dogs\"."}], "task": "refering", "answer_template": "The \"a dark brown dog beside two other dogs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 270, 271, 272, 273, 274], "bbox": [0.68225, 0.2816121495327103, 0.964140625, 0.7748831775700934], "iscrowd": 0, "area": 29801.807450000004, "rle": {"size": [428, 640], "counts": "Vjf56o07j:0hD=R;K^Dd0[;h0N1O2O0O1O2N1O1O2N1O101N1O1O1O1O1O1O1O100O1O1O1O1lFZMa7g2_H]M\\7d2dH`MV7b2iHbMS7_2mHeMn6\\2RIhMh6Z2WIkMd6V2\\InM_6S2aIQNY6T2cIQNX6S2eIRNU6R2hIPNT6U2hIlMW6e4O2O000O101O0O10O1000O10O1000O1000000O10000O1000000O10001O000000000000000000000000000000001O00000000000000000000000000000000001O000000000000000001O001O00001O001O010O001O00001O001O001O00001O001O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2N4L3M3QOnG^LV8R3XHlLk7e2cHYM`7a2eH]M_7]2dHbM_7X2fHfM]7U2gHiM]7Q2fHnM]7m1gHQN\\7j1hHTNY7i1jHVNW7g1Z2K4M3M3M3QOQ^9"}, "label": "a dark brown dog beside two other dogs"}]} {"id": 331419, "image": "COCO_train2014_000000331419.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dog on the far right\"."}], "task": "refering", "answer_template": "The \"the dog on the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 251, 270, 271, 272, 273, 274], "bbox": [0.68225, 0.2816121495327103, 0.964140625, 0.7748831775700934], "iscrowd": 0, "area": 29801.807450000004, "rle": {"size": [428, 640], "counts": "Vjf56o07j:0hD=R;K^Dd0[;h0N1O2O0O1O2N1O1O2N1O101N1O1O1O1O1O1O1O100O1O1O1O1lFZMa7g2_H]M\\7d2dH`MV7b2iHbMS7_2mHeMn6\\2RIhMh6Z2WIkMd6V2\\InM_6S2aIQNY6T2cIQNX6S2eIRNU6R2hIPNT6U2hIlMW6e4O2O000O101O0O10O1000O10O1000O1000000O10000O1000000O10001O000000000000000000000000000000001O00000000000000000000000000000000001O000000000000000001O001O00001O001O010O001O00001O001O001O00001O001O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2N4L3M3QOnG^LV8R3XHlLk7e2cHYM`7a2eH]M_7]2dHbM_7X2fHfM]7U2gHiM]7Q2fHnM]7m1gHQN\\7j1hHTNY7i1jHVNW7g1Z2K4M3M3M3QOQ^9"}, "label": "the dog on the far right"}]} {"id": 331419, "image": "COCO_train2014_000000331419.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog with a black band around it ' s snout\"."}], "task": "refering", "answer_template": "The \"a dog with a black band around it ' s snout\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 122, 123, 124, 139, 140, 141, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269], "bbox": [0.06459375, 0.3000934579439252, 0.746796875, 0.7689018691588786], "iscrowd": 0, "area": 53094.20264999996, "rle": {"size": [428, 640], "counts": "_Ya01W=4L3M4M3L12N2N39F in this image.", "objects": [{"patches": [94, 95, 96, 97, 98, 116, 117, 118, 119, 120, 121, 122, 123, 124, 139, 140, 141, 142, 143, 144, 145, 146, 147, 163, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269], "bbox": [0.06459375, 0.3000934579439252, 0.746796875, 0.7689018691588786], "iscrowd": 0, "area": 53094.20264999996, "rle": {"size": [428, 640], "counts": "_Ya01W=4L3M4M3L12N2N39F in this image.", "objects": [{"patches": [55, 69, 70, 84, 85, 96, 97, 98, 99, 100, 112, 113, 114, 126, 127, 128, 129, 141, 142, 143, 144, 145, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 204, 205, 207, 208, 220, 221, 235, 236], "bbox": [0.4204906542056075, 0.16112500000000002, 0.8828738317757009, 0.69340625], "iscrowd": 0, "area": 20566.963350000005, "rle": {"size": [640, 428], "counts": "Wh`34jc08H9G8H2N2N2M3N2N2N1O2N2N2N2N2^N\\Nf@e1X?`Nd@b1Y?cNc@_1Z?fNb@\\1[?iNb@X1[?mNa@U1_?mN]@T1d?nNX@T1h?nNT@T1l?nNQ@S1o?oNm_OS1T`0nNh_OT1X`0nNd_OS1]`0X1000010O001O0010O01O001O010O002N1O101N1O110O00001O00001O0000001O00001O000N3M2N3M3N2M3M3M2O2M3M3M3M3N2M2N3M3N2M3M3M2N3NiKmAc2P>\\MTBd2g=\\M]Be2_=[MeBe2W=[MlBf2o<[MTCe2ic0QBZOP>g0PBWOR>i0oATOS>l0nAQOS>P1nAmNT>S1mAjNU>V1lAgNU>Z1lAcNV>]1jAaNW>`1U200O0101O0O10001N10001N1000010O01O0010O00010O01O0010OO2N1O2N1N2O2N1N3N1O2M2O2N1N2O2N1N3N4L4K5L4L4K^ln0"}, "label": "a man in a blue shirt going up a skateboard bowl"}]} {"id": 290460, "image": "COCO_train2014_000000290460.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a bright blue t - shirt and cool white sunglasses rides the skateboard up the ramp\"."}], "task": "refering", "answer_template": "The \"a man in a bright blue t - shirt and cool white sunglasses rides the skateboard up the ramp\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 69, 70, 84, 85, 96, 97, 98, 99, 100, 112, 113, 114, 126, 127, 128, 129, 141, 142, 143, 144, 145, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 204, 205, 207, 208, 220, 221, 235, 236], "bbox": [0.4204906542056075, 0.16112500000000002, 0.8828738317757009, 0.69340625], "iscrowd": 0, "area": 20566.963350000005, "rle": {"size": [640, 428], "counts": "Wh`34jc08H9G8H2N2N2M3N2N2N1O2N2N2N2N2^N\\Nf@e1X?`Nd@b1Y?cNc@_1Z?fNb@\\1[?iNb@X1[?mNa@U1_?mN]@T1d?nNX@T1h?nNT@T1l?nNQ@S1o?oNm_OS1T`0nNh_OT1X`0nNd_OS1]`0X1000010O001O0010O01O001O010O002N1O101N1O110O00001O00001O0000001O00001O000N3M2N3M3N2M3M3M2O2M3M3M3M3N2M2N3M3N2M3M3M2N3NiKmAc2P>\\MTBd2g=\\M]Be2_=[MeBe2W=[MlBf2o<[MTCe2ic0QBZOP>g0PBWOR>i0oATOS>l0nAQOS>P1nAmNT>S1mAjNU>V1lAgNU>Z1lAcNV>]1jAaNW>`1U200O0101O0O10001N10001N1000010O01O0010O00010O01O0010OO2N1O2N1N2O2N1N3N1O2M2O2N1N2O2N1N3N4L4K5L4L4K^ln0"}, "label": "a man in a bright blue t - shirt and cool white sunglasses rides the skateboard up the ramp"}]} {"id": 478899, "image": "COCO_train2014_000000478899.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy going on stairs and one man doing ice skating\"."}], "task": "refering", "answer_template": "The \"a boy going on stairs and one man doing ice skating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 35, 36, 37, 38, 58, 59, 60, 61, 81, 82, 83, 104, 105, 106, 128, 129, 151, 152, 174], "bbox": [0.5198750000000001, 0.0016901408450704224, 0.6986875, 0.5186150234741784], "iscrowd": 0, "area": 11687.404400000001, "rle": {"size": [426, 640], "counts": "jcZ4>`<=ROn0J6M3N1O1O010O1O001O1TOVNWFl1h9VNUFl1j9VNSFl1m9i0O101N1O101N3N2M3M2L5K5iHSLi4R4RKQLm4U4kJnKT5U6O001O1O1O0O2O1O1O1O001N2O1O001O1O1OO1O1O1O1O1O1O1O1[O_I`Jb6_5bI]J_6b5dI[J]6e5fIVJ\\6i5c0`JjHk4X7PKlHo4U7lJQIS5c7O1`N`1H9K4N2N2N3N2WNlE9V:GlE5W:ImE2U:Mf1N2O0O1O010O1O00100O001O001O1O001O001O1O001O001O1O001O1O1O1O001O1O1O001O1O1LcW`2"}, "label": "a boy going on stairs and one man doing ice skating"}]} {"id": 478899, "image": "COCO_train2014_000000478899.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a blue and grey coat\"."}], "task": "refering", "answer_template": "The \"a person in a blue and grey coat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 35, 36, 37, 38, 58, 59, 60, 61, 81, 82, 83, 104, 105, 106, 128, 129, 151, 152, 174], "bbox": [0.5198750000000001, 0.0016901408450704224, 0.6986875, 0.5186150234741784], "iscrowd": 0, "area": 11687.404400000001, "rle": {"size": [426, 640], "counts": "jcZ4>`<=ROn0J6M3N1O1O010O1O001O1TOVNWFl1h9VNUFl1j9VNSFl1m9i0O101N1O101N3N2M3M2L5K5iHSLi4R4RKQLm4U4kJnKT5U6O001O1O1O0O2O1O1O1O001N2O1O001O1O1OO1O1O1O1O1O1O1O1[O_I`Jb6_5bI]J_6b5dI[J]6e5fIVJ\\6i5c0`JjHk4X7PKlHo4U7lJQIS5c7O1`N`1H9K4N2N2N3N2WNlE9V:GlE5W:ImE2U:Mf1N2O0O1O010O1O00100O001O001O1O001O001O1O001O001O1O001O1O1O1O001O1O1O001O1O1LcW`2"}, "label": "a person in a blue and grey coat"}]} {"id": 478899, "image": "COCO_train2014_000000478899.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a snowboarder with light blue pants\"."}], "task": "refering", "answer_template": "The \"a snowboarder with light blue pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 54, 55, 56, 77, 78, 79, 80, 100, 101, 102, 103, 123, 124, 125, 126, 146, 147, 148, 169, 170, 171, 193, 194, 195, 216, 217, 218, 239, 240], "bbox": [0.34976562499999997, 0.07490610328638497, 0.503390625, 0.6777230046948358], "iscrowd": 0, "area": 15049.272099999996, "rle": {"size": [426, 640], "counts": "ZWm29b24_70PHd0i7_OjGQ1P8ROdG]1W8eNQG^ONU2m8`NSG]OLW2n8^NTG^OJX2o8\\NSGAJW2P9f1M3N2M1O010000O10]GeKW8Z4iGiKU8V4jGoKS8P4mGSLQ8T3QHlLM5o7j2ZHPMF:n7a2bHQM@b0W1UOY5R3WJSMZOj0Q1SO_5R3UJmL[OR1m0ROc5Q3YJnM2TOe5P3UJnM3UOh5o2QJmM5WOj5l2oIoM6VOk5k2mIPN8UOl5l2iIQN:SOn5l2fIRNoI_OS6eJ[Oa5`0eJYOa5c0eJUOa5f0eJTO`5h0fJPO`5k0d3K5L3L5K5K5LbTT4"}, "label": "a snowboarder with light blue pants"}]} {"id": 478899, "image": "COCO_train2014_000000478899.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"snowboarder\"."}], "task": "refering", "answer_template": "The \"snowboarder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 54, 55, 56, 77, 78, 79, 80, 100, 101, 102, 103, 123, 124, 125, 126, 146, 147, 148, 169, 170, 171, 193, 194, 195, 216, 217, 218, 239, 240], "bbox": [0.34976562499999997, 0.07490610328638497, 0.503390625, 0.6777230046948358], "iscrowd": 0, "area": 15049.272099999996, "rle": {"size": [426, 640], "counts": "ZWm29b24_70PHd0i7_OjGQ1P8ROdG]1W8eNQG^ONU2m8`NSG]OLW2n8^NTG^OJX2o8\\NSGAJW2P9f1M3N2M1O010000O10]GeKW8Z4iGiKU8V4jGoKS8P4mGSLQ8T3QHlLM5o7j2ZHPMF:n7a2bHQM@b0W1UOY5R3WJSMZOj0Q1SO_5R3UJmL[OR1m0ROc5Q3YJnM2TOe5P3UJnM3UOh5o2QJmM5WOj5l2oIoM6VOk5k2mIPN8UOl5l2iIQN:SOn5l2fIRNoI_OS6eJ[Oa5`0eJYOa5c0eJUOa5f0eJTO`5h0fJPO`5k0d3K5L3L5K5K5LbTT4"}, "label": "snowboarder"}]} {"id": 536244, "image": "COCO_train2014_000000536244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sitting on a road taking a picture\"."}], "task": "refering", "answer_template": "The \"a man sitting on a road taking a picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 199, 200, 201, 222, 223, 224, 225, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 289, 290, 291, 293, 312, 313, 314], "bbox": [0.571390625, 0.5168544600938967, 0.78828125, 0.8921361502347418], "iscrowd": 0, "area": 10674.0704, "rle": {"size": [426, 640], "counts": "lbh43V=3M4L3M3M3M1O2N10000hC[Ok;e0TD_Oi;a0UDEf;O1O1O2O2N1O1O2N1O8F:F:EYlg1"}, "label": "a man sitting on a road taking a picture"}]} {"id": 536244, "image": "COCO_train2014_000000536244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white shirt with jeans taking a pic of a greyhound\"."}], "task": "refering", "answer_template": "The \"a man in a white shirt with jeans taking a pic of a greyhound\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 199, 200, 201, 222, 223, 224, 225, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 289, 290, 291, 293, 312, 313, 314], "bbox": [0.571390625, 0.5168544600938967, 0.78828125, 0.8921361502347418], "iscrowd": 0, "area": 10674.0704, "rle": {"size": [426, 640], "counts": "lbh43V=3M4L3M3M3M1O2N10000hC[Ok;e0TD_Oi;a0UDEf;O1O1O2O2N1O1O2N1O8F:F:EYlg1"}, "label": "a man in a white shirt with jeans taking a pic of a greyhound"}]} {"id": 536244, "image": "COCO_train2014_000000536244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a purple shirt is riding on a skateboard as his picture is being taken\"."}], "task": "refering", "answer_template": "The \"a man in a purple shirt is riding on a skateboard as his picture is being taken\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 186, 187, 188, 189, 209, 210, 211, 212, 232, 233, 234, 235, 254, 255, 256, 257, 258, 278, 280], "bbox": [0.07996875, 0.35065727699530513, 0.314875, 0.8285446009389672], "iscrowd": 0, "area": 15882.64795, "rle": {"size": [426, 640], "counts": "Zae02Q=9J6M3M2M4_MIjG?l7HYGYOCP1R9KUG[OFk0S9NRG\\OHg0T91PG[OJf0S92QG[OId0T95oF[OJa0U97oF[OJ>V99kFi0T9XOiFk0V9e1O2N1O1O100O1O2N1O1OAUGaLj8\\3[GcLd8^3^G`La8a3`G^L_8c3cGZL]8g3eGWLZ8j3hGTLV8m3mGQLR8P4c00O2O00000O2O000100O0010O01O010O001O010O1O010O1O010O1OZGXLk7h3UHZLj7f3VH[Lh7e3RHQLH in this image.", "objects": [{"patches": [120, 121, 122, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 186, 187, 188, 189, 209, 210, 211, 212, 232, 233, 234, 235, 254, 255, 256, 257, 258, 278, 280], "bbox": [0.07996875, 0.35065727699530513, 0.314875, 0.8285446009389672], "iscrowd": 0, "area": 15882.64795, "rle": {"size": [426, 640], "counts": "Zae02Q=9J6M3M2M4_MIjG?l7HYGYOCP1R9KUG[OFk0S9NRG\\OHg0T91PG[OJf0S92QG[OId0T95oF[OJa0U97oF[OJ>V99kFi0T9XOiFk0V9e1O2N1O1O100O1O2N1O1OAUGaLj8\\3[GcLd8^3^G`La8a3`G^L_8c3cGZL]8g3eGWLZ8j3hGTLV8m3mGQLR8P4c00O2O00000O2O000100O0010O01O010O001O010O1O010O1O010O1OZGXLk7h3UHZLj7f3VH[Lh7e3RHQLH in this image.", "objects": [{"patches": [102, 124, 125, 146, 147, 148, 149, 170, 171, 193, 194, 215, 216, 217, 238, 239, 240, 261, 262, 263], "bbox": [0.377234375, 0.2723887587822014, 0.48796875, 0.770304449648712], "iscrowd": 0, "area": 8874.53705, "rle": {"size": [427, 640], "counts": "aeT39P=4nFEn4?dJOX55^J5^5N[J;`5GYJ`0d5DUJc0h5_OQJh0m5[OnGROc1j1\\6DYId0e6@RIg0k6^OkHj0R7[OeHl0Y7YO\\Ho0a7^2ON3N3L4M2M4M3M2M4M3L3N3M3K5J6I7K5L4L4L3MfGgLj6U3SIVMi6o2kH\\Mo6o2aH]MY7l2XH]MZ7eNkHb5Q7eJiH^5T7?L4M1O1O2O0O100OhNjI^KU6^4QJ`Kn5^4WJ`Kj5[4\\JcKd5Z4bJcK^5Y4hJdKY5P4UKmKm4d3cKXL_4j0QId1o9[NSFT1^:jNdEm0d:RO^Em0b:RO`Em0a:QObEm0a:oNaEP1b:lNiEi0Z:ROSFd0P:VOUFh0X;M3L4L7IUcX4"}, "label": "a man wearing all black skis with a woman"}]} {"id": 577206, "image": "COCO_train2014_000000577206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old man skiing\"."}], "task": "refering", "answer_template": "The \"an old man skiing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 124, 125, 146, 147, 148, 149, 170, 171, 193, 194, 215, 216, 217, 238, 239, 240, 261, 262, 263], "bbox": [0.377234375, 0.2723887587822014, 0.48796875, 0.770304449648712], "iscrowd": 0, "area": 8874.53705, "rle": {"size": [427, 640], "counts": "aeT39P=4nFEn4?dJOX55^J5^5N[J;`5GYJ`0d5DUJc0h5_OQJh0m5[OnGROc1j1\\6DYId0e6@RIg0k6^OkHj0R7[OeHl0Y7YO\\Ho0a7^2ON3N3L4M2M4M3M2M4M3L3N3M3K5J6I7K5L4L4L3MfGgLj6U3SIVMi6o2kH\\Mo6o2aH]MY7l2XH]MZ7eNkHb5Q7eJiH^5T7?L4M1O1O2O0O100OhNjI^KU6^4QJ`Kn5^4WJ`Kj5[4\\JcKd5Z4bJcK^5Y4hJdKY5P4UKmKm4d3cKXL_4j0QId1o9[NSFT1^:jNdEm0d:RO^Em0b:RO`Em0a:QObEm0a:oNaEP1b:lNiEi0Z:ROSFd0P:VOUFh0X;M3L4L7IUcX4"}, "label": "an old man skiing"}]} {"id": 577206, "image": "COCO_train2014_000000577206.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in sunglasses and a pink shirt is skiing\"."}], "task": "refering", "answer_template": "The \"a woman in sunglasses and a pink shirt is skiing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 149, 150, 172, 173, 195, 196, 197, 218, 219, 220, 241, 242, 243, 264], "bbox": [0.475015625, 0.33864168618266977, 0.6036406249999999, 0.7556908665105386], "iscrowd": 0, "area": 6925.495299999999, "rle": {"size": [427, 640], "counts": "Smn34W=6]F3[56PJ8o5NiI5V6OPI^O\\Nf0b81VHYOTO;O>f82nG^1Q8eNgGa1X8aNcGc1]8^N^Ge1b8]NYGj1c8f1J1N2O1N1O2O1N13L4M2M4M3MO01^O\\G\\Lc8o2_GQM1O_8n2lGgLI9\\8n2[HQMd7Y2eGfMk0Oa7Y2kG`Mh05]7Y2ZIeMf6[2\\IbMf6_2ZI_Mf6c2P22O2M21O01O0000O2O00000PN^EU1b:fNdEY1]:aNjE]1V:^NPFa1P:[NVFc1l:N1OO1O001N2OO1O1O010O1O11N13N3L3N1N100O1VOQD5Q in this image.", "objects": [{"patches": [126, 149, 150, 172, 173, 195, 196, 197, 218, 219, 220, 241, 242, 243, 264], "bbox": [0.475015625, 0.33864168618266977, 0.6036406249999999, 0.7556908665105386], "iscrowd": 0, "area": 6925.495299999999, "rle": {"size": [427, 640], "counts": "Smn34W=6]F3[56PJ8o5NiI5V6OPI^O\\Nf0b81VHYOTO;O>f82nG^1Q8eNgGa1X8aNcGc1]8^N^Ge1b8]NYGj1c8f1J1N2O1N1O2O1N13L4M2M4M3MO01^O\\G\\Lc8o2_GQM1O_8n2lGgLI9\\8n2[HQMd7Y2eGfMk0Oa7Y2kG`Mh05]7Y2ZIeMf6[2\\IbMf6_2ZI_Mf6c2P22O2M21O01O0000O2O00000PN^EU1b:fNdEY1]:aNjE]1V:^NPFa1P:[NVFc1l:N1OO1O001N2OO1O1O010O1O11N13N3L3N1N100O1VOQD5Q in this image.", "objects": [{"patches": [245, 246, 267, 268, 269, 270, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 308, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 334, 335, 336], "bbox": [0.402734375, 0.6434934497816595, 0.7894375, 0.8901310043668123], "iscrowd": 0, "area": 14016.0078, "rle": {"size": [458, 640], "counts": "ihc33U>4L4L4L4L4L4M3L4L2O0O2O1N2O1N2O1N2O100O10000O100O10000O100O10000O100O10O10O100000000001O00000000000000000000000000001O000O1000000000000000000000001O00000000000000000000000O100000O100000000000000O1000000000000O100000000000O0100000000000000O1000000000000O100000000000000O10O1000N2N2O1N2N2N2O1N2N2O1N2N2O1N2O1N2O1O1O1O1O1O1O001O1O1N2O1O1O1O1O1O1O1O1O1O1O100O10000O1000000000O10000H8C=D in this image.", "objects": [{"patches": [245, 246, 267, 268, 269, 270, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 308, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 334, 335, 336], "bbox": [0.402734375, 0.6434934497816595, 0.7894375, 0.8901310043668123], "iscrowd": 0, "area": 14016.0078, "rle": {"size": [458, 640], "counts": "ihc33U>4L4L4L4L4L4M3L4L2O0O2O1N2O1N2O1N2O100O10000O100O10000O100O10000O100O10O10O100000000001O00000000000000000000000000001O000O1000000000000000000000001O00000000000000000000000O100000O100000000000000O1000000000000O100000000000O0100000000000000O1000000000000O100000000000000O10O1000N2N2O1N2N2N2O1N2N2O1N2N2O1N2O1N2O1O1O1O1O1O1O001O1O1N2O1O1O1O1O1O1O1O1O1O1O100O10000O1000000000O10000H8C=D in this image.", "objects": [{"patches": [164, 165, 166, 175, 176, 177, 178, 188, 189, 190, 200, 201, 212, 213], "bbox": [0.6611280487804878, 0.7191000000000001, 0.8700914634146341, 0.98426], "iscrowd": 0, "area": 6906.628099999997, "rle": {"size": [500, 328], "counts": "P[Z3`0h>f0\\Oe0C:ZCmMY;U2cDQNZ;Q2eDQNX;Q2gDQNW;P2hDRNU;P2jDSNS;m1mDUNQ;l1nDVNo:l1PEUNo:l1PETNo:n1PERNP;n1PERNP;o1oDQNP;Q2oDoMQ;R2nDnMR;R2nDnMQ;T2lDnMT;S2iDoMW;R2fDPNZ;Q2cDQN\\;P2bDRN^;o1_DSNa;P30000O100000000O100000000000mNaDQN_;o1eDmM[;R2jDjMV;V2nDfMR;Y2PEfMP;Z2PEfMP;Y2REfMo:X2REhMn:X2REhMn:W2SEiMm:W2TEhMl:W2UEiMk:W2UEiMk:V2VEjMj:V2WEhMj:W2WEhMj:X2VEgMl:V2YEfMh:X2`EaMa:]2_1J5L5K4L5J5L:F>A?BUcd0"}, "label": "an empty wine glass near the paper cup"}]} {"id": 569037, "image": "COCO_train2014_000000569037.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the clear glass cup next to the short white cup\"."}], "task": "refering", "answer_template": "The \"the clear glass cup next to the short white cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 166, 175, 176, 177, 178, 188, 189, 190, 200, 201, 212, 213], "bbox": [0.6611280487804878, 0.7191000000000001, 0.8700914634146341, 0.98426], "iscrowd": 0, "area": 6906.628099999997, "rle": {"size": [500, 328], "counts": "P[Z3`0h>f0\\Oe0C:ZCmMY;U2cDQNZ;Q2eDQNX;Q2gDQNW;P2hDRNU;P2jDSNS;m1mDUNQ;l1nDVNo:l1PEUNo:l1PETNo:n1PERNP;n1PERNP;o1oDQNP;Q2oDoMQ;R2nDnMR;R2nDnMQ;T2lDnMT;S2iDoMW;R2fDPNZ;Q2cDQN\\;P2bDRN^;o1_DSNa;P30000O100000000O100000000000mNaDQN_;o1eDmM[;R2jDjMV;V2nDfMR;Y2PEfMP;Z2PEfMP;Y2REfMo:X2REhMn:X2REhMn:W2SEiMm:W2TEhMl:W2UEiMk:W2UEiMk:V2VEjMj:V2WEhMj:W2WEhMj:X2VEgMl:V2YEfMh:X2`EaMa:]2_1J5L5K4L5J5L:F>A?BUcd0"}, "label": "the clear glass cup next to the short white cup"}]} {"id": 569037, "image": "COCO_train2014_000000569037.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"backside of a man\"."}], "task": "refering", "answer_template": "The \"backside of a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 26, 27, 38, 39, 40, 50, 51, 52, 62, 63, 74, 75, 76, 86, 87, 88], "bbox": [0.21887195121951222, 0.08532, 0.40728658536585366, 0.43774], "iscrowd": 0, "area": 5353.85865, "rle": {"size": [500, 328], "counts": "\\VS1:Z?>Ad0]OQ1oN?cC[Mm:e2SEfMc:Z2\\EPNZ:P2gEZNo9e1QFeNe9[1\\FlN]9S1cFmN_9R1`FnNa9Q1`FnNb9P1^FPOd9n0]FQOe9m0[FSOh9k0WFUOj9j0WFUOj9lN[E]1k0Gk9kN\\E\\1j0Hk9jN]E]1h0In9fN\\Ea1f0H^:8bEH_:7aEI_:7bEH_:8`EH`:8aEH_:7aEI_:7aEI_:7bEH_:8`EH`:8aEF`::`EFa:9`EF`::`EF`:;_EDb:<^EDc:;\\EFd:9]EGc:8^EHb:7_EAROoN_;`1_EAi:>XEAh:?YE@h:?YE@W;OkD1V;NiD2Y;LhD3Y;LhD3Z;KgD4Z;KREIn:8W2O1O00101N2N3M5L5J5Iecn2"}, "label": "backside of a man"}]} {"id": 569037, "image": "COCO_train2014_000000569037.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person in the background wearing a white hate\"."}], "task": "refering", "answer_template": "The \"the person in the background wearing a white hate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 26, 27, 38, 39, 40, 50, 51, 52, 62, 63, 74, 75, 76, 86, 87, 88], "bbox": [0.21887195121951222, 0.08532, 0.40728658536585366, 0.43774], "iscrowd": 0, "area": 5353.85865, "rle": {"size": [500, 328], "counts": "\\VS1:Z?>Ad0]OQ1oN?cC[Mm:e2SEfMc:Z2\\EPNZ:P2gEZNo9e1QFeNe9[1\\FlN]9S1cFmN_9R1`FnNa9Q1`FnNb9P1^FPOd9n0]FQOe9m0[FSOh9k0WFUOj9j0WFUOj9lN[E]1k0Gk9kN\\E\\1j0Hk9jN]E]1h0In9fN\\Ea1f0H^:8bEH_:7aEI_:7bEH_:8`EH`:8aEH_:7aEI_:7aEI_:7bEH_:8`EH`:8aEF`::`EFa:9`EF`::`EF`:;_EDb:<^EDc:;\\EFd:9]EGc:8^EHb:7_EAROoN_;`1_EAi:>XEAh:?YE@h:?YE@W;OkD1V;NiD2Y;LhD3Y;LhD3Z;KgD4Z;KREIn:8W2O1O00101N2N3M5L5J5Iecn2"}, "label": "the person in the background wearing a white hate"}]} {"id": 569037, "image": "COCO_train2014_000000569037.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a shaved head and a gold watch\"."}], "task": "refering", "answer_template": "The \"a man with a shaved head and a gold watch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 28, 29, 30, 31, 40, 41, 42, 43, 44, 48, 49, 50, 52, 53, 54, 55, 56, 57, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 77, 78, 79, 80, 81, 82, 90, 91, 92, 93, 103, 104], "bbox": [0.02350609756097561, 0.0571, 0.8822256097560975, 0.48178000000000004], "iscrowd": 0, "area": 19867.54895, "rle": {"size": [500, 328], "counts": "WQ4=U?101O100O1O1O0H_O_Aa0a>@]Aa0c>@[Aa0e>@YAa0g>60O101O0000000O10000000000O10001O00000O2O00001O001N10001O001O01O010O010O0010O01O010O010O0010O010O01O010O010O0010O01O001O010O001O001O010O001O0000010O00001O00010O00001O00010O00001O0^M[OVFd0g9BVF>f9IWF6g90VF0g96VFJi9:TFEl9?RF@m9f0mEZOS:T1^EnNb:R1]EoNb:R1]EnNc:S1\\EnNc:S1\\EnNd:R1[EoNd:R1[EoNd:R1[EoNd:R1[EPOd:P1ZEROe:P1YEQOf:P1YEQOf:Q1XEPOh:Q1VEQOh:P1WEQOh:Q1VEPOi:Q1VEPOj:R1SEoNl:T1PEnNP;T1mDnNR;U1jDlNW;U1fDlNZ;g2001O01O01O001O01O01O00100O1O100O10O01O100O1O100O00100O1O1O4L7I7I7I1O00000001O0000001O0O100iCnLl;S3SDPMj;P3UDUMg;k2XDZMd;f2[D_Ma;b2]DcM_;]2aDfM\\;Z2dDjMX;V2hDnMT;S2kDQNR;n1nDUNo:k1REXNj:h1VE\\Ng:d1XE`Nd:`1\\EdNa:[1`EbNb:_1]E]Ng:c1YEYNl:f1UETNP;l1PEPNU;P2jDlMZ;T2fDhM^;X2cDbMc;[2_D^Mh;`2ZDZMm;e2TDTMR in this image.", "objects": [{"patches": [16, 17, 18, 28, 29, 30, 31, 40, 41, 42, 43, 44, 48, 49, 50, 52, 53, 54, 55, 56, 57, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 77, 78, 79, 80, 81, 82, 90, 91, 92, 93, 103, 104], "bbox": [0.02350609756097561, 0.0571, 0.8822256097560975, 0.48178000000000004], "iscrowd": 0, "area": 19867.54895, "rle": {"size": [500, 328], "counts": "WQ4=U?101O100O1O1O0H_O_Aa0a>@]Aa0c>@[Aa0e>@YAa0g>60O101O0000000O10000000000O10001O00000O2O00001O001N10001O001O01O010O010O0010O01O010O010O0010O010O01O010O010O0010O01O001O010O001O001O010O001O0000010O00001O00010O00001O00010O00001O0^M[OVFd0g9BVF>f9IWF6g90VF0g96VFJi9:TFEl9?RF@m9f0mEZOS:T1^EnNb:R1]EoNb:R1]EnNc:S1\\EnNc:S1\\EnNd:R1[EoNd:R1[EoNd:R1[EoNd:R1[EPOd:P1ZEROe:P1YEQOf:P1YEQOf:Q1XEPOh:Q1VEQOh:P1WEQOh:Q1VEPOi:Q1VEPOj:R1SEoNl:T1PEnNP;T1mDnNR;U1jDlNW;U1fDlNZ;g2001O01O01O001O01O01O00100O1O100O10O01O100O1O100O00100O1O1O4L7I7I7I1O00000001O0000001O0O100iCnLl;S3SDPMj;P3UDUMg;k2XDZMd;f2[D_Ma;b2]DcM_;]2aDfM\\;Z2dDjMX;V2hDnMT;S2kDQNR;n1nDUNo:k1REXNj:h1VE\\Ng:d1XE`Nd:`1\\EdNa:[1`EbNb:_1]E]Ng:c1YEYNl:f1UETNP;l1PEPNU;P2jDlMZ;T2fDhM^;X2cDbMc;[2_D^Mh;`2ZDZMm;e2TDTMR in this image.", "objects": [{"patches": [121, 143, 144, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 257, 258, 259, 281, 282], "bbox": [0.187359375, 0.3814051522248244, 0.34790625, 0.8254332552693209], "iscrowd": 0, "area": 9124.31725, "rle": {"size": [427, 640], "counts": "f[b13T=4L5K400O2N1XND`F>\\9I`F6]92^FO_98\\FHa9`0[F@b9e0[F\\Oa9i0]FWOa9n0\\FSO`9S1[FoNc9V1SFROi9Y2N3M2N3M2N3O0O2UOXLRHi3m7ZLPHf3P8^LlGc3S8`LjGh3n7\\LmGQ4g7RLVHR4f7RLVHo3i7i001O1O001O1O00100O1O001O1O01000O010000O1iN`HcL`7Q3TIfLm6U3`IcL_6Y3l1K5L4L4M4K5L5J5F:H801N10001O00001O0WNPEV1P;bN\\EZ1^;M4L4L001O1O001O001N2001N2O1O1O1O1OO2N2M2M4L4L3M4K4M4L3Mmc]5"}, "label": "a young boy chasing a ball"}]} {"id": 413391, "image": "COCO_train2014_000000413391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little boy chasing a soccer ball\"."}], "task": "refering", "answer_template": "The \"little boy chasing a soccer ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 143, 144, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 257, 258, 259, 281, 282], "bbox": [0.187359375, 0.3814051522248244, 0.34790625, 0.8254332552693209], "iscrowd": 0, "area": 9124.31725, "rle": {"size": [427, 640], "counts": "f[b13T=4L5K400O2N1XND`F>\\9I`F6]92^FO_98\\FHa9`0[F@b9e0[F\\Oa9i0]FWOa9n0\\FSO`9S1[FoNc9V1SFROi9Y2N3M2N3M2N3O0O2UOXLRHi3m7ZLPHf3P8^LlGc3S8`LjGh3n7\\LmGQ4g7RLVHR4f7RLVHo3i7i001O1O001O1O00100O1O001O1O01000O010000O1iN`HcL`7Q3TIfLm6U3`IcL_6Y3l1K5L4L4M4K5L5J5F:H801N10001O00001O0WNPEV1P;bN\\EZ1^;M4L4L001O1O001O001N2001N2O1O1O1O1OO2N2M2M4L4L3M4K4M4L3Mmc]5"}, "label": "little boy chasing a soccer ball"}]} {"id": 413391, "image": "COCO_train2014_000000413391.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in a purple shirt\"."}], "task": "refering", "answer_template": "The \"a girl in a purple shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 99, 100, 101, 102, 122, 123, 124, 145, 146, 147, 148, 168, 169, 170, 171, 192, 193, 194, 195, 215, 216, 217, 218, 237, 238, 239, 240, 260, 261, 262, 263, 283, 284], "bbox": [0.31778125, 0.227423887587822, 0.50453125, 0.8165807962529275], "iscrowd": 0, "area": 15152.0008, "rle": {"size": [427, 640], "counts": "_nd21Y=:F9G2N2\\JZO`Nh0^1_O[Nc0d1DTN=k1GQN;n1EQN=m1DRN>m1BRN`0l1ASNm0`1SO_Nn0`1SO_Nn0a1RO^No0a1RO^No0b1QOPK:T1f0k3ROmJa0Q1>Q4POlIYOn0a1o07W4oNlI[Oi0g1n00\\4oNmI[Oe0n1l0Ib4nNmI\\O`0a2`0VOR5nNnI\\O in this image.", "objects": [{"patches": [77, 78, 79, 99, 100, 101, 102, 122, 123, 124, 145, 146, 147, 148, 168, 169, 170, 171, 192, 193, 194, 195, 215, 216, 217, 218, 237, 238, 239, 240, 260, 261, 262, 263, 283, 284], "bbox": [0.31778125, 0.227423887587822, 0.50453125, 0.8165807962529275], "iscrowd": 0, "area": 15152.0008, "rle": {"size": [427, 640], "counts": "_nd21Y=:F9G2N2\\JZO`Nh0^1_O[Nc0d1DTN=k1GQN;n1EQN=m1DRN>m1BRN`0l1ASNm0`1SO_Nn0`1SO_Nn0a1RO^No0a1RO^No0b1QOPK:T1f0k3ROmJa0Q1>Q4POlIYOn0a1o07W4oNlI[Oi0g1n00\\4oNmI[Oe0n1l0Ib4nNmI\\O`0a2`0VOR5nNnI\\O in this image.", "objects": [{"patches": [123, 124, 125, 126, 135, 136, 137, 138, 147, 148, 149, 150, 159, 160, 161, 162, 171, 172, 173, 174, 183, 184, 185, 186, 195, 196, 197, 198, 208, 209, 210], "bbox": [0.2606024096385542, 0.55886, 0.5821084337349397, 0.97684], "iscrowd": 0, "area": 15064.496200000001, "rle": {"size": [500, 332], "counts": "YhZ13a?3M3M3M3M3M3L4M3M3M3fEUOe5n0PJ[Oo5h0fIAY6b0\\IGc6l7EiGd0V8_O_Gj0`8YOUGP1j8SOkFV1T9mNaF\\1]9^201O1O1O1O1O001O1O1O1O1O1O001O1N100000O1000000O10000O1000000O10O10O1000000O1000000O10000O100000O010000O1000000O1iKeFU2[9dMPGX2P9aM[G[2e8^MfG]2[8[MQHa2o7XM\\Hd2d7UMgHf2Z7SMRIh2m6QM_Ik2i9K5K5K5K5K5K6J5K5K5J6K5K5K5K5K6J5K5Kh]S2"}, "label": "a glass model flower pot with flowers"}]} {"id": 388823, "image": "COCO_train2014_000000388823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glass vase with three flowers in it\"."}], "task": "refering", "answer_template": "The \"glass vase with three flowers in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 135, 136, 137, 138, 147, 148, 149, 150, 159, 160, 161, 162, 171, 172, 173, 174, 183, 184, 185, 186, 195, 196, 197, 198, 208, 209, 210], "bbox": [0.2606024096385542, 0.55886, 0.5821084337349397, 0.97684], "iscrowd": 0, "area": 15064.496200000001, "rle": {"size": [500, 332], "counts": "YhZ13a?3M3M3M3M3M3L4M3M3M3fEUOe5n0PJ[Oo5h0fIAY6b0\\IGc6l7EiGd0V8_O_Gj0`8YOUGP1j8SOkFV1T9mNaF\\1]9^201O1O1O1O1O001O1O1O1O1O1O001O1N100000O1000000O10000O1000000O10O10O1000000O1000000O10000O100000O010000O1000000O1iKeFU2[9dMPGX2P9aM[G[2e8^MfG]2[8[MQHa2o7XM\\Hd2d7UMgHf2Z7SMRIh2m6QM_Ik2i9K5K5K5K5K5K6J5K5K5J6K5K5K5K5K6J5K5Kh]S2"}, "label": "glass vase with three flowers in it"}]} {"id": 388823, "image": "COCO_train2014_000000388823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two empty vases near a vase of flowers ; one is clear glass and the other is opaque white\"."}], "task": "refering", "answer_template": "The \"two empty vases near a vase of flowers ; one is clear glass and the other is opaque white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 168, 169, 170, 180, 181, 182, 192, 193, 194, 195, 204, 205, 206], "bbox": [0.023704819277108435, 0.74988, 0.2571987951807229, 0.97684], "iscrowd": 0, "area": 6057.42355, "rle": {"size": [500, 332], "counts": "a[43_?3L5K4M3L5L3L4M4K4L4M4K4M3L5L3L3hN[N^Dh1b;gNmC[1R in this image.", "objects": [{"patches": [157, 158, 168, 169, 170, 180, 181, 182, 192, 193, 194, 195, 204, 205, 206], "bbox": [0.023704819277108435, 0.74988, 0.2571987951807229, 0.97684], "iscrowd": 0, "area": 6057.42355, "rle": {"size": [500, 332], "counts": "a[43_?3L5K4M3L5L3L4M4K4L4M4K4M3L5L3L3hN[N^Dh1b;gNmC[1R in this image.", "objects": [{"patches": [54, 55, 77, 78, 79, 99, 100, 101, 102, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 174, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 266, 267, 283, 284, 285, 286, 287, 288, 289, 290, 306, 307, 308, 309, 310, 311, 312, 313, 330, 331, 332, 333, 334, 335, 336], "bbox": [0.30196875, 0.15712264150943397, 0.624859375, 1.0], "iscrowd": 0, "area": 41399.37740000002, "rle": {"size": [424, 640], "counts": "UX`22U=4M1N2O4gIHYO:b0L\\O5`0O_O1=5AL;8CJ8;FG6=HD6?IB4a0JA3a0M@Oe0O[OOh0OBGa07_OHc07^OGe07[OHg06[OHh06XOIj06VOIl06TOJm05SOJo04ROKP14QOJQ15oNJS15mNPOhNmL]2R4kNPOnNhLX2W4lNoNoNgLV2Z4kNnNQOgLT2[4kNmNTOeLR2^4kNkNTOfLS2^4iNkNVOfLQ2^4jNkNWOeLS2]4fNmNY2Q1gMoN\\2n0dMQO`2l0`MSOc2k0^MSOe2k0[MUOg2i0XMWOk2g0UMYOk2g0UMXOl2h0TMXOk2i0UMWOk2i0UMWOk2i0UMWOk2i0UMWOk2i0TMXOl2h0TMXOm2g0SMYOm2g0RMPMlNn0S4R2oLhLWOV1j3R2nLeL\\OY1g3Q2lLcLA\\1c3Q2kLaLE^1a3Q2iL_LH`1_3Q2hL^LKb1\\3P2hL]LNc1[3o1fL]L1e1X3n1fL\\L4g1V3m1dL[L8i1S3l1dLZL;j1R3k1bLZL>l1o2j1cLXL`0o1l2j1bLVLd0Q2i2h1cLVLf0R2g2g1fLTLd0V2e2e1jLQLc0[2b2c1aN^N_1`1bN`N^1`1bNaN^1]1cNdN\\1[1eNfNZ1BWLZO`2T1Z1[O]L_OZ2W1X1WOcL_OV2[1W1SOgL_OS2`1T1eNWMHf1d1R1cNZMGe1g1o0aN`MEb1l1k0_NfMAa1Q2g0]NkM@_1U2d0ZNPN^O]1Y2a0XNVN\\OZ1]2>WNZNZOY1a2:TNaNWOW1f27RNdNVOV1i24oMkNUOR1n20jMSOVOn0Q3MfMh2[2YM_Mk2c2VMVMn2k2TMnLP3T3a31O1O3M3M3M3M3M4L4L3M4L4L9G001O1O001O1O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O1O2N1O1jHUJm6l5QIVJn6l5oHVJP7R6O001O1O1O1O1O1O1O2N2N1O2N2N1O2N2N1O2N2N2N1O2N2N3M3M3M3M3M3cK^H]1HPOR8VO_He1DQOY8iN]HP2_OSOZ9g0kFUOZ9e0kFWOY9d0lFXOY9>PG^Oa;L4L5K4L5KjnR3"}, "label": "a woman with earbuds in looking at her phone"}]} {"id": 577240, "image": "COCO_train2014_000000577240.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a halter dress with her hair pinned up and earbuds in her ears , looking downward\"."}], "task": "refering", "answer_template": "The \"a woman wearing a halter dress with her hair pinned up and earbuds in her ears , looking downward\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 77, 78, 79, 99, 100, 101, 102, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 174, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 261, 262, 263, 264, 265, 266, 267, 283, 284, 285, 286, 287, 288, 289, 290, 306, 307, 308, 309, 310, 311, 312, 313, 330, 331, 332, 333, 334, 335, 336], "bbox": [0.30196875, 0.15712264150943397, 0.624859375, 1.0], "iscrowd": 0, "area": 41399.37740000002, "rle": {"size": [424, 640], "counts": "UX`22U=4M1N2O4gIHYO:b0L\\O5`0O_O1=5AL;8CJ8;FG6=HD6?IB4a0JA3a0M@Oe0O[OOh0OBGa07_OHc07^OGe07[OHg06[OHh06XOIj06VOIl06TOJm05SOJo04ROKP14QOJQ15oNJS15mNPOhNmL]2R4kNPOnNhLX2W4lNoNoNgLV2Z4kNnNQOgLT2[4kNmNTOeLR2^4kNkNTOfLS2^4iNkNVOfLQ2^4jNkNWOeLS2]4fNmNY2Q1gMoN\\2n0dMQO`2l0`MSOc2k0^MSOe2k0[MUOg2i0XMWOk2g0UMYOk2g0UMXOl2h0TMXOk2i0UMWOk2i0UMWOk2i0UMWOk2i0UMWOk2i0TMXOl2h0TMXOm2g0SMYOm2g0RMPMlNn0S4R2oLhLWOV1j3R2nLeL\\OY1g3Q2lLcLA\\1c3Q2kLaLE^1a3Q2iL_LH`1_3Q2hL^LKb1\\3P2hL]LNc1[3o1fL]L1e1X3n1fL\\L4g1V3m1dL[L8i1S3l1dLZL;j1R3k1bLZL>l1o2j1cLXL`0o1l2j1bLVLd0Q2i2h1cLVLf0R2g2g1fLTLd0V2e2e1jLQLc0[2b2c1aN^N_1`1bN`N^1`1bNaN^1]1cNdN\\1[1eNfNZ1BWLZO`2T1Z1[O]L_OZ2W1X1WOcL_OV2[1W1SOgL_OS2`1T1eNWMHf1d1R1cNZMGe1g1o0aN`MEb1l1k0_NfMAa1Q2g0]NkM@_1U2d0ZNPN^O]1Y2a0XNVN\\OZ1]2>WNZNZOY1a2:TNaNWOW1f27RNdNVOV1i24oMkNUOR1n20jMSOVOn0Q3MfMh2[2YM_Mk2c2VMVMn2k2TMnLP3T3a31O1O3M3M3M3M3M4L4L3M4L4L9G001O1O001O1O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O1O2N1O1jHUJm6l5QIVJn6l5oHVJP7R6O001O1O1O1O1O1O1O2N2N1O2N2N1O2N2N1O2N2N2N1O2N2N3M3M3M3M3M3cK^H]1HPOR8VO_He1DQOY8iN]HP2_OSOZ9g0kFUOZ9e0kFWOY9d0lFXOY9>PG^Oa;L4L5K4L5KjnR3"}, "label": "a woman wearing a halter dress with her hair pinned up and earbuds in her ears , looking downward"}]} {"id": 20188, "image": "COCO_train2014_000000020188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a purple hat sitting on a skateboard\"."}], "task": "refering", "answer_template": "The \"a man holding a purple hat sitting on a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 23, 37, 38, 39, 40, 41, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 84, 85, 86, 87, 88, 89, 90, 91, 101, 102, 103, 104, 105, 106, 107, 108, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 241, 242, 243, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 261, 262, 263, 264, 265, 266, 267, 268, 269, 275, 276, 277, 278, 279, 280, 281, 282, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312], "bbox": [0.03175166297117517, 0.065203125, 1.0, 0.834765625], "iscrowd": 0, "area": 126162.31955000006, "rle": {"size": [640, 451], "counts": "cR91e10d`05o^O6m`00f^O in this image.", "objects": [{"patches": [21, 22, 23, 37, 38, 39, 40, 41, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 84, 85, 86, 87, 88, 89, 90, 91, 101, 102, 103, 104, 105, 106, 107, 108, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 241, 242, 243, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 261, 262, 263, 264, 265, 266, 267, 268, 269, 275, 276, 277, 278, 279, 280, 281, 282, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312], "bbox": [0.03175166297117517, 0.065203125, 1.0, 0.834765625], "iscrowd": 0, "area": 126162.31955000006, "rle": {"size": [640, 451], "counts": "cR91e10d`05o^O6m`00f^O in this image.", "objects": [{"patches": [0, 1, 16, 17, 18, 32, 33, 34, 35, 48, 49, 50, 51, 64, 65, 66, 67, 80, 81, 82, 83, 96, 97, 98, 99, 112, 113, 114, 128, 129, 130, 144, 145, 146, 160, 161, 162, 176, 177, 192], "bbox": [0.003702882483370288, 0.0026093749999999997, 0.2512860310421286, 0.5625], "iscrowd": 0, "area": 24298.70835, "rle": {"size": [640, 451], "counts": "SX1T;l8O1N2O1O1O1O1N2O1O1O1O1N2O1O1O001N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O3M3M2N3L4M2N3M2N3L4M2N3M3M2M4M3M2N3M2M4M3M2N3M1N3H7A`0@?B?@?A?B?L3O2O0O2N1O1O2N1O2N1O2N1O1O2N3M4L3M4L3M4L3M4L4L3M4L3M2N2O1N2N2N1O2N2N2N2N2N2N2N2N2N1O:F`0@a0_OkSc6"}, "label": "a man is standing"}]} {"id": 20188, "image": "COCO_train2014_000000020188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man is standing wearing a blue jeans\"."}], "task": "refering", "answer_template": "The \"a man is standing wearing a blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 16, 17, 18, 32, 33, 34, 35, 48, 49, 50, 51, 64, 65, 66, 67, 80, 81, 82, 83, 96, 97, 98, 99, 112, 113, 114, 128, 129, 130, 144, 145, 146, 160, 161, 162, 176, 177, 192], "bbox": [0.003702882483370288, 0.0026093749999999997, 0.2512860310421286, 0.5625], "iscrowd": 0, "area": 24298.70835, "rle": {"size": [640, 451], "counts": "SX1T;l8O1N2O1O1O1O1N2O1O1O1O1N2O1O1O001N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O3M3M2N3L4M2N3M2N3L4M2N3M3M2M4M3M2N3M2M4M3M2N3M1N3H7A`0@?B?@?A?B?L3O2O0O2N1O1O2N1O2N1O2N1O1O2N3M4L3M4L3M4L3M4L4L3M4L3M2N2O1N2N2N1O2N2N2N2N2N2N2N2N2N1O:F`0@a0_OkSc6"}, "label": "a man is standing wearing a blue jeans"}]} {"id": 20188, "image": "COCO_train2014_000000020188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue backpack on the skateboarder\"."}], "task": "refering", "answer_template": "The \"a blue backpack on the skateboarder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 141, 142, 143, 158, 159, 174, 175, 190, 191, 206, 207, 222, 223, 238, 239, 254, 255], "bbox": [0.8133924611973392, 0.31456249999999997, 1.0, 0.673890625], "iscrowd": 0, "area": 11235.783499999998, "rle": {"size": [640, 451], "counts": "ZbU71oc02O1O1O1N2O1O1N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2O1N2]@dN^<^1^CiN] in this image.", "objects": [{"patches": [125, 126, 127, 141, 142, 143, 158, 159, 174, 175, 190, 191, 206, 207, 222, 223, 238, 239, 254, 255], "bbox": [0.8133924611973392, 0.31456249999999997, 1.0, 0.673890625], "iscrowd": 0, "area": 11235.783499999998, "rle": {"size": [640, 451], "counts": "ZbU71oc02O1O1O1N2O1O1N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2O1N2]@dN^<^1^CiN] in this image.", "objects": [{"patches": [309, 310, 330, 331, 332, 352, 353, 355, 374, 375, 377, 378, 400], "bbox": [0.27046875, 0.731130604288499, 0.48981250000000004, 0.9705458089668616], "iscrowd": 0, "area": 2313.592950000003, "rle": {"size": [513, 640], "counts": "`Tg21o?1O101N1O1O1O2N001O1O001O001O1O001O001O1O00000000000000000001O000O100000000000000000000000000000000O1000000000O100000000000000O100000O100000O11O1O1O1O1O2N1O1O1O1O1O2Nhb1OW]N2N3N2M3M2N3N2M2N3N2M2N3M3N2M2hMmNaES1a:SOWEm0k:UOREk0P;TOnDl0T;TOjDl0X;TOfDm0[;SOdDl0_;SO_Dn0b;RO\\Dn0f;ROXDn0j;ROUDn0m;QOQDo0T in this image.", "objects": [{"patches": [309, 310, 330, 331, 332, 352, 353, 355, 374, 375, 377, 378, 400], "bbox": [0.27046875, 0.731130604288499, 0.48981250000000004, 0.9705458089668616], "iscrowd": 0, "area": 2313.592950000003, "rle": {"size": [513, 640], "counts": "`Tg21o?1O101N1O1O1O2N001O1O001O001O1O001O001O1O00000000000000000001O000O100000000000000000000000000000000O1000000000O100000000000000O100000O100000O11O1O1O1O1O2N1O1O1O1O1O2Nhb1OW]N2N3N2M3M2N3N2M2N3N2M2N3M3N2M2hMmNaES1a:SOWEm0k:UOREk0P;TOnDl0T;TOjDl0X;TOfDm0[;SOdDl0_;SO_Dn0b;RO\\Dn0f;ROXDn0j;ROUDn0m;QOQDo0T in this image.", "objects": [{"patches": [52, 75, 76, 97, 98, 99, 119, 120, 121, 122, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 257, 258, 259, 280, 281, 282, 303, 305, 326, 328], "bbox": [0.169703125, 0.15939571150097465, 0.3409375, 0.8341325536062377], "iscrowd": 0, "area": 18703.6817, "rle": {"size": [513, 640], "counts": "_Pg19f?2N2N2N2N2N2J7I6J6I700kNoNYCR1cbKP28Q3_6VLlHcNi0V5a6RLhHaNl0]5a6lKkIS4`7cJ`H]5R9O10O0100O100O100000O2M2N2N2ZOlJ]FW5U9Q1C=B>jNdHYIk7a6n0I6L5J6J6PKYF`3o9ZLSFU3b:fL`Eg2V;TMlDn1X in this image.", "objects": [{"patches": [52, 75, 76, 97, 98, 99, 119, 120, 121, 122, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 257, 258, 259, 280, 281, 282, 303, 305, 326, 328], "bbox": [0.169703125, 0.15939571150097465, 0.3409375, 0.8341325536062377], "iscrowd": 0, "area": 18703.6817, "rle": {"size": [513, 640], "counts": "_Pg19f?2N2N2N2N2N2J7I6J6I700kNoNYCR1cbKP28Q3_6VLlHcNi0V5a6RLhHaNl0]5a6lKkIS4`7cJ`H]5R9O10O0100O100O100000O2M2N2N2ZOlJ]FW5U9Q1C=B>jNdHYIk7a6n0I6L5J6J6PKYF`3o9ZLSFU3b:fL`Eg2V;TMlDn1X in this image.", "objects": [{"patches": [54, 55, 77, 78, 79, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 127, 146, 147, 148, 149, 169, 170, 171, 172, 192, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 241, 261, 262, 263, 264, 285, 286, 287, 308, 309, 310, 330, 331, 332], "bbox": [0.338640625, 0.10785575048732943, 0.554796875, 0.8157309941520468], "iscrowd": 0, "area": 25672.165450000004, "rle": {"size": [513, 640], "counts": "fj\\3;\\??E:K6Jj0WO4K4L4L4L4L2N2O1N2N2N3cImLGU38kLDZ3:gL^Ob3`0_LWOk3g0VLdN`4Z1aKfNa4W1`KiNa4U1`KkNa4S1`KmNd4n0]KROj4g0VKYOQ5?PKAZ54gJLV6VOkIj0X7RNiHn1h7`MYH`2o7WMRHi2V8nLkGR3^8dLcG\\3e8[L\\Ge3P9oKPGQ4\\9bKeF^4g9UKZFl4]:2N2N2N1O2N2N1O2N2N2N1O2O1N1O2N3N2M3M4MO00010O00010O00010O01O01O01O01O01O01O01O0cGlIm6U6RImIk6T6VInIf6S6ZIPJc6Q6]IQJ_6P6aIRJ\\6P6dIRJX6Q6fIRJV6P6jIRJR6P6mIRJQ6o5oISJm5o5RJTJj5n5VJTJf5n5YJTJa5Q6_JQJX5V6gJmIP5Z6PKhIh4^6WKeIn1jNZOh7gNaIb1]OUOY7YO\\IU11POj6JXIi0d0lNZ6;TI in this image.", "objects": [{"patches": [54, 55, 77, 78, 79, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 127, 146, 147, 148, 149, 169, 170, 171, 172, 192, 193, 194, 195, 215, 216, 217, 218, 238, 239, 240, 241, 261, 262, 263, 264, 285, 286, 287, 308, 309, 310, 330, 331, 332], "bbox": [0.338640625, 0.10785575048732943, 0.554796875, 0.8157309941520468], "iscrowd": 0, "area": 25672.165450000004, "rle": {"size": [513, 640], "counts": "fj\\3;\\??E:K6Jj0WO4K4L4L4L4L2N2O1N2N2N3cImLGU38kLDZ3:gL^Ob3`0_LWOk3g0VLdN`4Z1aKfNa4W1`KiNa4U1`KkNa4S1`KmNd4n0]KROj4g0VKYOQ5?PKAZ54gJLV6VOkIj0X7RNiHn1h7`MYH`2o7WMRHi2V8nLkGR3^8dLcG\\3e8[L\\Ge3P9oKPGQ4\\9bKeF^4g9UKZFl4]:2N2N2N1O2N2N1O2N2N2N1O2O1N1O2N3N2M3M4MO00010O00010O00010O01O01O01O01O01O01O01O0cGlIm6U6RImIk6T6VInIf6S6ZIPJc6Q6]IQJ_6P6aIRJ\\6P6dIRJX6Q6fIRJV6P6jIRJR6P6mIRJQ6o5oISJm5o5RJTJj5n5VJTJf5n5YJTJa5Q6_JQJX5V6gJmIP5Z6PKhIh4^6WKeIn1jNZOh7gNaIb1]OUOY7YO\\IU11POj6JXIi0d0lNZ6;TI in this image.", "objects": [{"patches": [172, 173, 174, 195, 196, 197, 218, 219, 220, 241, 242, 243, 244, 264, 265, 266, 267, 268, 288, 289, 290, 313], "bbox": [0.489453125, 0.45287499999999997, 0.658921875, 0.7855416666666667], "iscrowd": 0, "area": 8756.7464, "rle": {"size": [480, 640], "counts": "Snb41S?2_Oa0iB]Oe;i0TDYOk;n0mCSOSUD_Ol;?XD]Oi;b0\\1N010O001O010O4J8ISlU3"}, "label": "the wooden chair on the left in the row of three"}]} {"id": 36574, "image": "COCO_train2014_000000036574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wooden chair closest to the table\"."}], "task": "refering", "answer_template": "The \"the wooden chair closest to the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 195, 196, 197, 218, 219, 220, 241, 242, 243, 244, 264, 265, 266, 267, 268, 288, 289, 290, 313], "bbox": [0.489453125, 0.45287499999999997, 0.658921875, 0.7855416666666667], "iscrowd": 0, "area": 8756.7464, "rle": {"size": [480, 640], "counts": "Snb41S?2_Oa0iB]Oe;i0TDYOk;n0mCSOSUD_Ol;?XD]Oi;b0\\1N010O001O010O4J8ISlU3"}, "label": "the wooden chair closest to the table"}]} {"id": 36574, "image": "COCO_train2014_000000036574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden chair sitting between two other wooden chairs on a patio\"."}], "task": "refering", "answer_template": "The \"a wooden chair sitting between two other wooden chairs on a patio\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 198, 199, 221, 222, 223, 224, 244, 245, 246, 247, 267, 268, 269, 270, 291, 292, 314], "bbox": [0.6076875, 0.44502083333333337, 0.764375, 0.7774791666666668], "iscrowd": 0, "area": 8402.543700000002, "rle": {"size": [480, 640], "counts": "jaf56i>9H7I8H9G`0@3nBYNdMa_V2"}, "label": "a wooden chair sitting between two other wooden chairs on a patio"}]} {"id": 36574, "image": "COCO_train2014_000000036574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden chair at the middle of two other wooden chairs\"."}], "task": "refering", "answer_template": "The \"a wooden chair at the middle of two other wooden chairs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 198, 199, 221, 222, 223, 224, 244, 245, 246, 247, 267, 268, 269, 270, 291, 292, 314], "bbox": [0.6076875, 0.44502083333333337, 0.764375, 0.7774791666666668], "iscrowd": 0, "area": 8402.543700000002, "rle": {"size": [480, 640], "counts": "jaf56i>9H7I8H9G`0@3nBYNdMa_V2"}, "label": "a wooden chair at the middle of two other wooden chairs"}]} {"id": 69344, "image": "COCO_train2014_000000069344.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy wearing a blue shirt with ninja on it eating cake\"."}], "task": "refering", "answer_template": "The \"a young boy wearing a blue shirt with ninja on it eating cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 274, 290, 291, 292, 293, 294, 295, 296, 297, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 381, 382, 383, 384, 385, 386, 387, 388, 389, 404, 405, 406, 407, 408, 409, 411, 412], "bbox": [0.5640468750000001, 0.16890624999999998, 0.9593125, 0.7723437499999999], "iscrowd": 0, "area": 65194.34035000002, "rle": {"size": [640, 640], "counts": "mRR72jc05K4L5J5L5K4L5K4L5K4L5K4L5K4L4L5L3L5K4M4K4L5L3L5K4M4K4L5L3L5K4M3N2M3N2N2M3N2N2M3O1N2OO00001O000101N2O1N2N2O0O2O1N101N101N2N101N101N101N101O1O001O1N101O1O1O0RMYJdGh5V8iJ^GX5m7`KPH`4o7eKlG[4T8iKhGX4V8nKeGS4Z8QLcGn3]8VLZFjMf0Q6n8YLRFTNi0c5U9RMeFo2Z9cMUF]2i9m301O0O2O2N3M2M4L4L4M2M4L4L3M4M3N1N100O010O10000O010O10000O010O10000O010O10O10O10O1000O010000O01000O10O010000O10000O10000O100O10000O10000000000O100000O1000O1000000000000000O2O000001O0001O01O00001O01O01O00001O0cI`G[1_8]NPH\\1P8]NXHb1i7\\NXHd1h7\\NWHe1i7ZNWHg1j7XNVHh1j7XNUHj1j7UNWHk1i7UNVHl1k7SNTHn1l7QNUHo1k7PNUHQ2l7mMUHT2j7kMVHV2j7iMWHW2i7hMXHX2j7eMWH[2k7aMXH^2k7]MWHd2j7WMZHh2h7SM[Hm2g7oL]Ho2f7kL^HT3d7hL_HW3c7\\LiHd3Y7SLnHl3T7lKSIS4o6fKWIY4k6_K\\I`4g6WK`Ih4b6QKdIn4^6gJlIY5V6`JoI_5Z:O1O1O1O100O1O1O1O1O100O1O1O1O1O2O1N1Oe0^Nm@mL`?^2l@^MY?V2QAgMT?l1XAQNl>b1`AZNd>X1in?"}, "label": "a young boy wearing a blue shirt with ninja on it eating cake"}]} {"id": 69344, "image": "COCO_train2014_000000069344.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small boy in a blue shirt with a figure on the front that has a black and red hat\"."}], "task": "refering", "answer_template": "The \"a small boy in a blue shirt with a figure on the front that has a black and red hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [87, 88, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 250, 251, 267, 268, 269, 270, 271, 272, 273, 274, 290, 291, 292, 293, 294, 295, 296, 297, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 381, 382, 383, 384, 385, 386, 387, 388, 389, 404, 405, 406, 407, 408, 409, 411, 412], "bbox": [0.5640468750000001, 0.16890624999999998, 0.9593125, 0.7723437499999999], "iscrowd": 0, "area": 65194.34035000002, "rle": {"size": [640, 640], "counts": "mRR72jc05K4L5J5L5K4L5K4L5K4L5K4L5K4L4L5L3L5K4M4K4L5L3L5K4M4K4L5L3L5K4M3N2M3N2N2M3N2N2M3O1N2OO00001O000101N2O1N2N2O0O2O1N101N101N2N101N101N101N101O1O001O1N101O1O1O0RMYJdGh5V8iJ^GX5m7`KPH`4o7eKlG[4T8iKhGX4V8nKeGS4Z8QLcGn3]8VLZFjMf0Q6n8YLRFTNi0c5U9RMeFo2Z9cMUF]2i9m301O0O2O2N3M2M4L4L4M2M4L4L3M4M3N1N100O010O10000O010O10000O010O10000O010O10O10O10O1000O010000O01000O10O010000O10000O10000O100O10000O10000000000O100000O1000O1000000000000000O2O000001O0001O01O00001O01O01O00001O0cI`G[1_8]NPH\\1P8]NXHb1i7\\NXHd1h7\\NWHe1i7ZNWHg1j7XNVHh1j7XNUHj1j7UNWHk1i7UNVHl1k7SNTHn1l7QNUHo1k7PNUHQ2l7mMUHT2j7kMVHV2j7iMWHW2i7hMXHX2j7eMWH[2k7aMXH^2k7]MWHd2j7WMZHh2h7SM[Hm2g7oL]Ho2f7kL^HT3d7hL_HW3c7\\LiHd3Y7SLnHl3T7lKSIS4o6fKWIY4k6_K\\I`4g6WK`Ih4b6QKdIn4^6gJlIY5V6`JoI_5Z:O1O1O1O100O1O1O1O1O100O1O1O1O1O2O1N1Oe0^Nm@mL`?^2l@^MY?V2QAgMT?l1XAQNl>b1`AZNd>X1in?"}, "label": "a small boy in a blue shirt with a figure on the front that has a black and red hat"}]} {"id": 69344, "image": "COCO_train2014_000000069344.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cake with candles in it\"."}], "task": "refering", "answer_template": "The \"a cake with candles in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [379, 380, 381, 398, 399, 400, 401, 402, 403, 404, 405, 420, 421, 422, 423, 424, 425, 426, 427, 428, 443, 444, 445, 446, 447, 448, 449, 450, 451, 466, 467, 468, 469, 470, 471, 472, 473, 474, 490, 495, 496, 497, 519, 520], "bbox": [0.269109375, 0.700984375, 0.65165625, 0.9739374999999999], "iscrowd": 0, "area": 25249.7867, "rle": {"size": [640, 640], "counts": "`P\\33lc0=C=B>C:F1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N0O2O100O1O10O0100O11O2N1N2O2N1O2N1O2M2O1O2N1O2N1N2O2N1ON2E;F:L30100O1O1O10O01O1O100O1O00100O1=C8H1N2O1O1O1O1O1O1N2OO1O1O100O001O1O1O100O001O1O1K5I7J501000000O10O1000O2O000O1000000O1000000O101O0O1000000O100O100O1O1O2N100O1O1O1O101N1O2N1O2O0O2O001N101O0O101O001N101O0O2O001N101O001N101O0O2O001N101O2M2O1O1O1N2O1O2M2O1O1O10010O0001O0O1O2O0O1O2N100O2N1O1O2O0O2N1O101N1O1O2N100O2N1O100O1O1O1O100O1O1O010O1O1O1O100O1O1O0O2M3N2M3M3N3L in this image.", "objects": [{"patches": [379, 380, 381, 398, 399, 400, 401, 402, 403, 404, 405, 420, 421, 422, 423, 424, 425, 426, 427, 428, 443, 444, 445, 446, 447, 448, 449, 450, 451, 466, 467, 468, 469, 470, 471, 472, 473, 474, 490, 495, 496, 497, 519, 520], "bbox": [0.269109375, 0.700984375, 0.65165625, 0.9739374999999999], "iscrowd": 0, "area": 25249.7867, "rle": {"size": [640, 640], "counts": "`P\\33lc0=C=B>C:F1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N0O2O100O1O10O0100O11O2N1N2O2N1O2N1O2M2O1O2N1O2N1N2O2N1ON2E;F:L30100O1O1O10O01O1O100O1O00100O1=C8H1N2O1O1O1O1O1O1N2OO1O1O100O001O1O1O100O001O1O1K5I7J501000000O10O1000O2O000O1000000O1000000O101O0O1000000O100O100O1O1O2N100O1O1O1O101N1O2N1O2O0O2O001N101O0O101O001N101O0O2O001N101O001N101O0O2O001N101O2M2O1O1O1N2O1O2M2O1O1O10010O0001O0O1O2O0O1O2N100O2N1O1O2O0O2N1O101N1O1O2N100O2N1O100O1O1O1O100O1O1O010O1O1O1O100O1O1O0O2M3N2M3M3N3L in this image.", "objects": [{"patches": [7, 8, 9, 10, 29, 30, 31, 32, 33, 34, 52, 53, 54, 55, 56, 57, 75, 76, 77, 78, 79, 80, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 349, 350, 351, 352, 353, 354, 355, 356, 375, 377, 378, 379, 400, 401, 402], "bbox": [0.0494375, 0.031093749999999996, 0.512359375, 0.756921875], "iscrowd": 0, "area": 83555.82854999999, "rle": {"size": [640, 640], "counts": "mZd09`c0:F:F:F;F9L1N3N2N1O2N2N1O2N2N1O2N2N1O2N1O2M3N1O2N2N1O2N2N1O2M3N1N3N2N1N3N2M2O2M3N1O2M3N1N3N2N1N3N2M2O2M3N1O2M3N1N2O2M2O2N1N3N1N2O2M2O2N1N2O2M2O2M2O1O2M2O2M2O2N1N2O2M2O2M2O1O2N1O2O0O2N100O2N101N1O101N1O2O0O2N100O0010O01O010O0010O01O010O010O010O00100O010O010O0010O02O1N2O2M2N3nM`H`Ga7U8kHhGX7l7SISHn6i7WITHl6g7XIXHi6e7ZIYHh6c7]IZHf6b7]I]Hd6_7`I_Hc6\\7bIbH_6[7dIcH_6X7eIgH\\6V7hIgH[6T7iIjHY6R7lIkHV6R7mIkHU6R7RJiHP6T7WJfHj5W7^JcHd5Z7cJ`H^5^7c3N3L3N3M2M3N3M2N3L3N2N3O00001O001O00001O001O01O0O1O1O100O1O100O1O100O1O1O100O1O100O1O100000001O000000000000000000000001O00000000000000000eITCb5l in this image.", "objects": [{"patches": [7, 8, 9, 10, 29, 30, 31, 32, 33, 34, 52, 53, 54, 55, 56, 57, 75, 76, 77, 78, 79, 80, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 191, 192, 193, 194, 195, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 349, 350, 351, 352, 353, 354, 355, 356, 375, 377, 378, 379, 400, 401, 402], "bbox": [0.0494375, 0.031093749999999996, 0.512359375, 0.756921875], "iscrowd": 0, "area": 83555.82854999999, "rle": {"size": [640, 640], "counts": "mZd09`c0:F:F:F;F9L1N3N2N1O2N2N1O2N2N1O2N2N1O2N1O2M3N1O2N2N1O2N2N1O2M3N1N3N2N1N3N2M2O2M3N1O2M3N1N3N2N1N3N2M2O2M3N1O2M3N1N2O2M2O2N1N3N1N2O2M2O2N1N2O2M2O2M2O1O2M2O2M2O2N1N2O2M2O2M2O1O2N1O2O0O2N100O2N101N1O101N1O2O0O2N100O0010O01O010O0010O01O010O010O010O00100O010O010O0010O02O1N2O2M2N3nM`H`Ga7U8kHhGX7l7SISHn6i7WITHl6g7XIXHi6e7ZIYHh6c7]IZHf6b7]I]Hd6_7`I_Hc6\\7bIbH_6[7dIcH_6X7eIgH\\6V7hIgH[6T7iIjHY6R7lIkHV6R7mIkHU6R7RJiHP6T7WJfHj5W7^JcHd5Z7cJ`H^5^7c3N3L3N3M2M3N3M2N3L3N2N3O00001O001O00001O001O01O0O1O1O100O1O100O1O100O1O1O100O1O100O1O100000001O000000000000000000000001O00000000000000000eITCb5l in this image.", "objects": [{"patches": [93, 94, 95, 116, 117, 118, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 208, 209, 210, 231, 232, 233, 234, 254, 255, 256, 257, 277, 278, 279, 280], "bbox": [0.025968750000000002, 0.3119241192411924, 0.376578125, 0.9898644986449864], "iscrowd": 0, "area": 23817.61205, "rle": {"size": [369, 640], "counts": "[Z6l0b:4L5L3L5K4M4K4B>M4L3M4L3L6K6J7I6K5Je0\\Oe0ZO?Ad0\\Oa0^O5L4L3M4K5L4L4M1000000O10000O10000O10000000000O10000000000O100000000O1000000O1000000O10000O10000O10O010jI^L]4c3aK_L^4b3aK^L_4c3_K_LV4IcJi3V1^LV4n3hKSLX4n3gKSLX4m3cKiKnN:_5n3`KjKPO9`5n3]KlKRO7`5P4YK[Lf4h3UKZLk4i3PKZLo4U5O1O1O1O1O1N1eLbJc0_5[OhJ?Y5^OnJ=S5ATK9m4RNkIZ1^1?o4AUK in this image.", "objects": [{"patches": [93, 94, 95, 116, 117, 118, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 208, 209, 210, 231, 232, 233, 234, 254, 255, 256, 257, 277, 278, 279, 280], "bbox": [0.025968750000000002, 0.3119241192411924, 0.376578125, 0.9898644986449864], "iscrowd": 0, "area": 23817.61205, "rle": {"size": [369, 640], "counts": "[Z6l0b:4L5L3L5K4M4K4B>M4L3M4L3L6K6J7I6K5Je0\\Oe0ZO?Ad0\\Oa0^O5L4L3M4K5L4L4M1000000O10000O10000O10000000000O10000000000O100000000O1000000O1000000O10000O10000O10O010jI^L]4c3aK_L^4b3aK^L_4c3_K_LV4IcJi3V1^LV4n3hKSLX4n3gKSLX4m3cKiKnN:_5n3`KjKPO9`5n3]KlKRO7`5P4YK[Lf4h3UKZLk4i3PKZLo4U5O1O1O1O1O1N1eLbJc0_5[OhJ?Y5^OnJ=S5ATK9m4RNkIZ1^1?o4AUK in this image.", "objects": [{"patches": [74, 75, 96, 97, 98, 119, 120, 121, 122, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 187, 188, 189, 190, 210, 211, 212, 213], "bbox": [0.12464062499999999, 0.24009367681498828, 0.32353125, 0.6220140515222483], "iscrowd": 0, "area": 13346.460649999999, "rle": {"size": [427, 640], "counts": "ZbQ17l<:E:F:L5M2N2N2N3M2N101N2N2N2N1O2N2N2O1N1O2O1N2O4K6K5J4L100N2O1O001O1N2N2K4L5K5L4O001O1N2O1O1O001N2O1O1O1N2O1O1B>O100O010O1O100O100O1000000000000000001O00000000001O00000000001O04L4L4L4L3N3L3M1O0O2L4iNkFRNZ9l1mFkMW9T2nFcMW9[2PG\\MT9c2k0N1O2O1N101N2N3N2M4M2M4L3N2M3N2M3M3N3L3N2M3M3I`\\d5"}, "label": "the person sitting to the left of the scale"}]} {"id": 421618, "image": "COCO_train2014_000000421618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman selling produce in an outdoor marketplace\"."}], "task": "refering", "answer_template": "The \"woman selling produce in an outdoor marketplace\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 96, 97, 98, 119, 120, 121, 122, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 168, 187, 188, 189, 190, 210, 211, 212, 213], "bbox": [0.12464062499999999, 0.24009367681498828, 0.32353125, 0.6220140515222483], "iscrowd": 0, "area": 13346.460649999999, "rle": {"size": [427, 640], "counts": "ZbQ17l<:E:F:L5M2N2N2N3M2N101N2N2N2N1O2N2N2O1N1O2O1N2O4K6K5J4L100N2O1O001O1N2N2K4L5K5L4O001O1N2O1O1O001N2O1O1O1N2O1O1B>O100O010O1O100O100O1000000000000000001O00000000001O00000000001O04L4L4L4L3N3L3M1O0O2L4iNkFRNZ9l1mFkMW9T2nFcMW9[2PG\\MT9c2k0N1O2O1N101N2N3N2M4M2M4L3N2M3N2M3M3N3L3N2M3M3I`\\d5"}, "label": "woman selling produce in an outdoor marketplace"}]} {"id": 102144, "image": "COCO_train2014_000000102144.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the teddy bear on the end with a blue tie\"."}], "task": "refering", "answer_template": "The \"the teddy bear on the end with a blue tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 74, 93, 94, 95, 96, 97, 117, 118, 119, 120, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 234, 235], "bbox": [0.079703125, 0.2629179331306991, 0.365046875, 0.849452887537994], "iscrowd": 0, "area": 22423.121199999994, "rle": {"size": [329, 640], "counts": "Q``02T:4M4L4nIFn1>fKF?1i3=eKN1LW4:eK6CGf45fK>TODS52fKe1X4^NfKd1X4_NeKc1X4aNeKa1Y4aNfK_1X4eNeK]1Y4fNdK\\1Y4gNeK[1Y4iNcKX1[4nN`KT1_4oN^KR1a4ROZKP1e4Z201O1O1O010O1O1O001O0000000000001O0000000000001O0000000000001O000000000000001O0000000000001O000000000000001O0000000000001O0000000001O00000`LeKS1Z4eNoKZ1Q4^NWLb1j3UN_Lj1b3oMdLQ2^3kMeLT2\\3iMfLW2\\3dMhL[2Y3bMiL]2Y3`MjL^2Y3^MiLa2Y3[MkLd2W3XMkLg2W3VMlLh2V3UMlLj2W3QMmLm2U3PMmLo2X5N1O2N101N1O2N1O2N1010O01O010O0010O01O010O001O010O0010O01O010O001O010O0010O01O010O0010OO2N1O2O0O2N1O2O0O2N1O2N101N1O1O010O1TObGL_84dGH\\88gGFY8:iGCX8=jG@W8`0kG]OU8d0mGZOS8e0PHXOQ8h0RHTOo7l0g001N8I6J6JbaR4"}, "label": "the teddy bear on the end with a blue tie"}]} {"id": 102144, "image": "COCO_train2014_000000102144.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown teddy bear with a blue bow\"."}], "task": "refering", "answer_template": "The \"a brown teddy bear with a blue bow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 74, 93, 94, 95, 96, 97, 117, 118, 119, 120, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 234, 235], "bbox": [0.079703125, 0.2629179331306991, 0.365046875, 0.849452887537994], "iscrowd": 0, "area": 22423.121199999994, "rle": {"size": [329, 640], "counts": "Q``02T:4M4L4nIFn1>fKF?1i3=eKN1LW4:eK6CGf45fK>TODS52fKe1X4^NfKd1X4_NeKc1X4aNeKa1Y4aNfK_1X4eNeK]1Y4fNdK\\1Y4gNeK[1Y4iNcKX1[4nN`KT1_4oN^KR1a4ROZKP1e4Z201O1O1O010O1O1O001O0000000000001O0000000000001O0000000000001O000000000000001O0000000000001O000000000000001O0000000000001O0000000001O00000`LeKS1Z4eNoKZ1Q4^NWLb1j3UN_Lj1b3oMdLQ2^3kMeLT2\\3iMfLW2\\3dMhL[2Y3bMiL]2Y3`MjL^2Y3^MiLa2Y3[MkLd2W3XMkLg2W3VMlLh2V3UMlLj2W3QMmLm2U3PMmLo2X5N1O2N101N1O2N1O2N1010O01O010O0010O01O010O001O010O0010O01O010O001O010O0010O01O010O0010OO2N1O2O0O2N1O2O0O2N1O2N101N1O1O010O1TObGL_84dGH\\88gGFY8:iGCX8=jG@W8`0kG]OU8d0mGZOS8e0PHXOQ8h0RHTOo7l0g001N8I6J6JbaR4"}, "label": "a brown teddy bear with a blue bow"}]} {"id": 102144, "image": "COCO_train2014_000000102144.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bear furthest to the right\"."}], "task": "refering", "answer_template": "The \"the bear furthest to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 66, 86, 87, 88, 89, 110, 111, 112, 132, 133, 134, 135, 156, 157, 158, 159, 179, 180, 181, 182, 204, 205], "bbox": [0.772640625, 0.17948328267477204, 0.952375, 0.7485714285714284], "iscrowd": 0, "area": 11424.371050000005, "rle": {"size": [329, 640], "counts": "jSo41m15l59oIKj5 in this image.", "objects": [{"patches": [55, 56, 78, 79, 80, 81, 101, 102, 103, 104, 123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 192, 193, 194], "bbox": [0.353703125, 0.21267477203647417, 0.603796875, 0.7524012158054711], "iscrowd": 0, "area": 15693.7008, "rle": {"size": [329, 640], "counts": "ZTY22i01`83YG3e8OWG4h8NUG4j8ORG4k8ORG3m8OPG4n8d0O2M2O1O2M2O2N1O2[Od0\\Od0O2N1O2N1N2O2N1O1O1O1O001O0O2O1]NlLPLT3P4c1O1O001O001O1O001O010000O100O10000O10000O100O100N2N2M3N2N2N200O2N3N2M3M2O2M3M3N1N2N100000000O1000000O100000000O10001O001O0O10O1000000O10O1000001O001lMaI<_6CdI:^6CfI;[6CgI in this image.", "objects": [{"patches": [84, 85, 86, 87, 107, 108, 109, 110, 129, 130, 131, 132, 151, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 201, 202], "bbox": [0.596078125, 0.26066869300911855, 0.808640625, 0.7101215805471125], "iscrowd": 0, "area": 11458.550500000003, "rle": {"size": [329, 640], "counts": "^bj31X:3L4M3M3L4M3nF\\OX8h0^G_Ob8d0TGCl8m0O100O010O10000O100O10000O010O100O1001O1O001O001O1O001O1O001O001O1O00oMjNPKU1n4POoJP1o4VOnJi0Q5[OmJd0Q5BkJ>S5GkJ8S5NjJ1U53hJMV59gJGW5>eJB[5a0cJ^O]5c0cJ\\O]5f0aJZO_5g0aJXO_5j0_JWO`5j0]JXOc5j0ZJWOf5j0WJXOi5j0TJWOk5k0RJWOn5k0nIXOQ6i0mIXOS6j0iIXOW6i0gIXOY6\\201O000000000000000000000000000O10000000000000000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O]M[Jb0e5ZO`Je0`5XOdJh0[5UOjJi0W5ROnJm0S5oNQKP1Q5kNTKT1m4iNUKV1l4gNWKX1k4dNYK[1g4cN[K\\1g4`N\\K_1e4_N]Ka1d4[N`Kc1b4YNaKf1`4XNbKh1_4TNeKg1_4VNdKf1b6L5K4L9H:E;E;EnoV1"}, "label": "a darker brown teddy bear in a row of lighter teddy bears"}]} {"id": 102144, "image": "COCO_train2014_000000102144.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stuffed brown teddy bear with a red bow\"."}], "task": "refering", "answer_template": "The \"a stuffed brown teddy bear with a red bow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 87, 107, 108, 109, 110, 129, 130, 131, 132, 151, 152, 153, 154, 155, 156, 175, 176, 177, 178, 179, 201, 202], "bbox": [0.596078125, 0.26066869300911855, 0.808640625, 0.7101215805471125], "iscrowd": 0, "area": 11458.550500000003, "rle": {"size": [329, 640], "counts": "^bj31X:3L4M3M3L4M3nF\\OX8h0^G_Ob8d0TGCl8m0O100O010O10000O100O10000O010O100O1001O1O001O001O1O001O1O001O001O1O00oMjNPKU1n4POoJP1o4VOnJi0Q5[OmJd0Q5BkJ>S5GkJ8S5NjJ1U53hJMV59gJGW5>eJB[5a0cJ^O]5c0cJ\\O]5f0aJZO_5g0aJXO_5j0_JWO`5j0]JXOc5j0ZJWOf5j0WJXOi5j0TJWOk5k0RJWOn5k0nIXOQ6i0mIXOS6j0iIXOW6i0gIXOY6\\201O000000000000000000000000000O10000000000000000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O]M[Jb0e5ZO`Je0`5XOdJh0[5UOjJi0W5ROnJm0S5oNQKP1Q5kNTKT1m4iNUKV1l4gNWKX1k4dNYK[1g4cN[K\\1g4`N\\K_1e4_N]Ka1d4[N`Kc1b4YNaKf1`4XNbKh1_4TNeKg1_4VNdKf1b6L5K4L9H:E;E;EnoV1"}, "label": "a stuffed brown teddy bear with a red bow"}]} {"id": 331520, "image": "COCO_train2014_000000331520.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the thing being held\"."}], "task": "refering", "answer_template": "The \"the thing being held\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 105, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237, 238, 240, 241, 242, 243, 256, 257, 258, 259, 260, 261, 264, 265, 282, 283, 284], "bbox": [0.12303125, 0.2494375, 0.8477499999999999, 0.7191041666666667], "iscrowd": 0, "area": 51950.280549999996, "rle": {"size": [480, 640], "counts": "`ZU11n>1O010O0`A0Q>1lA3R>NjA6U>JhA:W>9N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O2N100O2N1O1O2N1O1O2N1O1O1O1O2N100O1O2N1O1O1O1O2N1O1O101N1O100O101N100O1O2O0O100O101N100O101N100O100O2N100O101N100O100O2O0O10001O0O10000O1000000O1L4K5J6K5M3O1N2O1N2O1N101N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1O1O1OmNkF]LU9a3oF]LP9a3UG\\Lk8b3[G[Ld8d3`GZL_8d3fGZLY8d3mGYLR8f3RHXLm7f3XHXLg7h3]HULb7k3aHSL^7n3dHPL[7P4iHmKV7S4nHjKQ7V4m1O0010O01O1O010000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O10000O100O101O0O101N101O0O2O000O2O0O2O001O001N10001O001O001O0O2O1O2N1O1O1O1N3N1O1O1O1017HO1N2N2N3M2N2N2O101N10000O101N100O10001NO2N1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1N2O2N1O1O1O1O2N1O1O1O1O2N1O1O1O3M4L3N1N1O010O1O100O1O010O1O100O1O010O100O1O100O002O0O2N101N1O101N1O2O0O2N101N1O101N1O2O0O2N100O2N101N101N1O2O0O1O2O0O2N101N1O2O0O1O2O001O0O2O00001O001O001N101O00001O0N3M2N3M2N2N3N101O001O010O00001O001O01O00000000001O00000O1000001O00000000001O000000000100O1O2N101N1O1O2O0O1O2N100O2N1O2O0O1O2N100O2N1O2O0O1O2N100N3L3M4M2M3M4L3M3NnW]1"}, "label": "the thing being held"}]} {"id": 331520, "image": "COCO_train2014_000000331520.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the object being wrapped with string\"."}], "task": "refering", "answer_template": "The \"the object being wrapped with string\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 104, 105, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237, 238, 240, 241, 242, 243, 256, 257, 258, 259, 260, 261, 264, 265, 282, 283, 284], "bbox": [0.12303125, 0.2494375, 0.8477499999999999, 0.7191041666666667], "iscrowd": 0, "area": 51950.280549999996, "rle": {"size": [480, 640], "counts": "`ZU11n>1O010O0`A0Q>1lA3R>NjA6U>JhA:W>9N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O2N100O2N1O1O2N1O1O2N1O1O1O1O2N100O1O2N1O1O1O1O2N1O1O101N1O100O101N100O1O2O0O100O101N100O101N100O100O2N100O101N100O100O2O0O10001O0O10000O1000000O1L4K5J6K5M3O1N2O1N2O1N101N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1O1O1OmNkF]LU9a3oF]LP9a3UG\\Lk8b3[G[Ld8d3`GZL_8d3fGZLY8d3mGYLR8f3RHXLm7f3XHXLg7h3]HULb7k3aHSL^7n3dHPL[7P4iHmKV7S4nHjKQ7V4m1O0010O01O1O010000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O10000O100O101O0O101N101O0O2O000O2O0O2O001O001N10001O001O001O0O2O1O2N1O1O1O1N3N1O1O1O1017HO1N2N2N3M2N2N2O101N10000O101N100O10001NO2N1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1N2O2N1O1O1O1O2N1O1O1O1O2N1O1O1O3M4L3N1N1O010O1O100O1O010O1O100O1O010O100O1O100O002O0O2N101N1O101N1O2O0O2N101N1O101N1O2O0O2N100O2N101N101N1O2O0O1O2O0O2N101N1O2O0O1O2O001O0O2O00001O001O001N101O00001O0N3M2N3M2N2N3N101O001O010O00001O001O01O00000000001O00000O1000001O00000000001O000000000100O1O2N101N1O1O2O0O1O2N100O2N1O2O0O1O2N100O2N1O2O0O1O2N100N3L3M4M2M3M4L3M3NnW]1"}, "label": "the object being wrapped with string"}]} {"id": 347908, "image": "COCO_train2014_000000347908.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in blue jeans standing behind the fence\"."}], "task": "refering", "answer_template": "The \"a man in blue jeans standing behind the fence\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 25, 26, 27, 28, 29, 44, 45, 46, 47, 62, 63, 64, 81, 82, 97, 99, 100, 115, 116, 118, 132, 133, 151], "bbox": [0.3743, 0.0, 0.64844, 0.6045012787723786], "iscrowd": 0, "area": 12243.653549999999, "rle": {"size": [391, 500], "counts": "hcW22Q<5K4K6K5J6aKUO\\Mo0^2WO^Mn0Z2YOaMm0W2ZOeMk0S2\\OiMh0Q2^OjMh0S2ZOiMk0V2UOfMP1X2QOcMT1\\2mN`MX1^2hNbMZ1]2fNdMZ1Z2gNhMW1W2jNjMV1U2jNlMV1R2kNPNS1o1nNRNn0Q2ROQNj0P2WOQNe0Q2\\OQN`0P2AQN;R2EPN7Q2JPN2R2OPNMR22PNJR27PNEQ2 in this image.", "objects": [{"patches": [7, 8, 9, 10, 25, 26, 27, 28, 29, 44, 45, 46, 47, 62, 63, 64, 81, 82, 97, 99, 100, 115, 116, 118, 132, 133, 151], "bbox": [0.3743, 0.0, 0.64844, 0.6045012787723786], "iscrowd": 0, "area": 12243.653549999999, "rle": {"size": [391, 500], "counts": "hcW22Q<5K4K6K5J6aKUO\\Mo0^2WO^Mn0Z2YOaMm0W2ZOeMk0S2\\OiMh0Q2^OjMh0S2ZOiMk0V2UOfMP1X2QOcMT1\\2mN`MX1^2hNbMZ1]2fNdMZ1Z2gNhMW1W2jNjMV1U2jNlMV1R2kNPNS1o1nNRNn0Q2ROQNj0P2WOQNe0Q2\\OQN`0P2AQN;R2EPN7Q2JPN2R2OPNMR22PNJR27PNEQ2 in this image.", "objects": [{"patches": [25, 42, 43, 44, 60, 61, 62, 77, 78, 79, 80, 81, 96, 97, 98, 99, 100, 116, 117, 118, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 169, 170, 171, 173, 174, 175, 176, 188, 189, 190, 193, 194, 195, 206, 207, 208, 212, 213], "bbox": [0.3005, 0.12997442455242966, 0.84526, 0.8131202046035806], "iscrowd": 0, "area": 18786.71145, "rle": {"size": [391, 500], "counts": "j\\i11U<2N2M3N2N2M3N2N2M3O1N2O1N2O1N100O100O100O10O0100O100O100O100[OoNSFR1h9TOWFl0f9XOXFi0f9YOZFg0d9\\OZFe0d9^O[Fa0d9A[F`0b9D\\F;c9I\\F2f92XFJk98TFIi9;VFFi9;UFGj9[1O2O1O001N101O1O0O2O1O2M3N2N2N2N2NMmFTMQ9l2PGUMl2NQ3n2TJTMe23V3i2VJUM^27[3e2XJTMX2;`3`2YJVMQ2?e3\\2[JUMk1c0i3X2]JVMe1f0m3U2_JVM^1i0R4Q2`JXMY1k0V4m1bJYMS1m0[4k1aJZMo0o0_4g1bJ\\Mj0Q1c4d1cJ[Mf0U1g4_1cJ]Mc0W1j4\\1cJ]Ma0Y1l4Y1cJ^M`0[1n4V1bJ_M>]1P5S1bJ`M=_1R5P1`JbM<`1T5m0aJbM:c1U5k0`JcM9d1X5g0`JdM7g1Y5d0`JfM5h1\\5a0_JfM4k1^5=^JiM2l1e56XJoM1m1m5MSJUNOP2S6dMlIe11hNNQ2b6VOaIhNLT2c6TO`IiNKU2f6PO`IjNIX2j0WMP5g1\\Jg1c0dLQ5c1\\Jj1c0dLP5a1^Jh0IdNh0TOQ5_1^Jd0OiN`0VOT5[1^Ja03nN9WOW5Y1]Ja05oN5XOZ5W1]J`05RO2YO]5S1\\Ja07ROO[O_5Q1[Jb08ROL\\Oc5n0ZJb09TOH]Of5l0YJb0:VOE]Oi5j0YJa0;XOA^Om5g0WJb0=YO]O_OP6e0WJa0>[OYO@T6b0UJb0?]OWO_OV6a0TJb0a0^OSO@Y6`0SJ?d05[5ZOQJ`0f05Z5ZOQJ>g09Y5XOoI=k0:W5XOnIT5WOkI8T1`0S5WOhI8W1`0R5WOgI7Z1a0Q5VOfI8Y1b0R5VOdI9Z1`0T5UOcI;Y1?U5UObI=Y1=W5UO_I?Y1 in this image.", "objects": [{"patches": [25, 42, 43, 44, 60, 61, 62, 77, 78, 79, 80, 81, 96, 97, 98, 99, 100, 116, 117, 118, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 169, 170, 171, 173, 174, 175, 176, 188, 189, 190, 193, 194, 195, 206, 207, 208, 212, 213], "bbox": [0.3005, 0.12997442455242966, 0.84526, 0.8131202046035806], "iscrowd": 0, "area": 18786.71145, "rle": {"size": [391, 500], "counts": "j\\i11U<2N2M3N2N2M3N2N2M3O1N2O1N2O1N100O100O100O10O0100O100O100O100[OoNSFR1h9TOWFl0f9XOXFi0f9YOZFg0d9\\OZFe0d9^O[Fa0d9A[F`0b9D\\F;c9I\\F2f92XFJk98TFIi9;VFFi9;UFGj9[1O2O1O001N101O1O0O2O1O2M3N2N2N2N2NMmFTMQ9l2PGUMl2NQ3n2TJTMe23V3i2VJUM^27[3e2XJTMX2;`3`2YJVMQ2?e3\\2[JUMk1c0i3X2]JVMe1f0m3U2_JVM^1i0R4Q2`JXMY1k0V4m1bJYMS1m0[4k1aJZMo0o0_4g1bJ\\Mj0Q1c4d1cJ[Mf0U1g4_1cJ]Mc0W1j4\\1cJ]Ma0Y1l4Y1cJ^M`0[1n4V1bJ_M>]1P5S1bJ`M=_1R5P1`JbM<`1T5m0aJbM:c1U5k0`JcM9d1X5g0`JdM7g1Y5d0`JfM5h1\\5a0_JfM4k1^5=^JiM2l1e56XJoM1m1m5MSJUNOP2S6dMlIe11hNNQ2b6VOaIhNLT2c6TO`IiNKU2f6PO`IjNIX2j0WMP5g1\\Jg1c0dLQ5c1\\Jj1c0dLP5a1^Jh0IdNh0TOQ5_1^Jd0OiN`0VOT5[1^Ja03nN9WOW5Y1]Ja05oN5XOZ5W1]J`05RO2YO]5S1\\Ja07ROO[O_5Q1[Jb08ROL\\Oc5n0ZJb09TOH]Of5l0YJb0:VOE]Oi5j0YJa0;XOA^Om5g0WJb0=YO]O_OP6e0WJa0>[OYO@T6b0UJb0?]OWO_OV6a0TJb0a0^OSO@Y6`0SJ?d05[5ZOQJ`0f05Z5ZOQJ>g09Y5XOoI=k0:W5XOnIT5WOkI8T1`0S5WOhI8W1`0R5WOgI7Z1a0Q5VOfI8Y1b0R5VOdI9Z1`0T5UOcI;Y1?U5UObI=Y1=W5UO_I?Y1 in this image.", "objects": [{"patches": [51, 52, 53, 66, 67, 68, 81, 82, 83, 84, 96, 97, 98, 110, 111, 112, 113, 124, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 143, 144, 145, 154, 155, 156, 157, 158, 159, 160, 169, 170, 171, 172, 173, 174, 175, 184, 185, 186, 187, 188, 189, 198, 199, 200, 201, 202, 203, 204, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 245, 246, 247, 248, 260, 261, 262, 263, 275, 276, 277, 278, 290, 291, 292, 293, 305, 306, 307, 308], "bbox": [0.25934426229508195, 0.1528125, 0.6770023419203747, 0.9078593749999999], "iscrowd": 0, "area": 48170.67435000001, "rle": {"size": [640, 427], "counts": "^gU28fc0:E;E;E;F9F2N2O2M2N3M2O1O2N1N2O]On^O\\NPa0e1d02WNXNXAi1f>^NTAc1k>cNo@_1o>hNj@Y1U?mNe@U1Y?QOa@P1^?WOZ@k0e?[OU@g0i?h1N3N1O1O2N1O1O2M2O2N1O1O2eAbKe3Jh4d4cGiK^3Em4c4cGPLX3@S5`4dGXLR3YOZ5^4cGaLk2TO`5\\4cGiLe2mNf5Z4dGQM_2gNl5X4cGYMR2iNY6n3dGaMb1lNj6b3cGjMR1POY7W3bGSNc0ROi7k2bG\\N4UOY8_2_GfNGWOh8S2_G`1_8`N_Gc1a8]N[Gg1d8YNZGj1e8h410O1O100O10OmM]EjJd:g4lEWKU:h4mEWKS:h4oEVKR:h4RFUKn9k4TFTKl9k4VFSKk9k4YFRKc9R5_FmJR9a5PG]Jb8P6bGnIo7`6SH^I_7P7cHnHo6`7TIXHj6n7c21000O10O2O001N2O1N2O1N2O1N2O1N2O1N12O1O2N1O1O1O1O1O1aITEh3n:bJmFW5UB=D=B>B>F9M4L4L4L3M4L4L4L3M4L4Ld1]Nnne2"}, "label": "all of the boy wearing a blue shirt , black shorts and dirty shoes"}]} {"id": 421643, "image": "COCO_train2014_000000421643.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy inside a house holding a wii controller in both hands\"."}], "task": "refering", "answer_template": "The \"a young boy inside a house holding a wii controller in both hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 66, 67, 68, 81, 82, 83, 84, 96, 97, 98, 110, 111, 112, 113, 124, 125, 126, 127, 128, 129, 130, 139, 140, 141, 142, 143, 144, 145, 154, 155, 156, 157, 158, 159, 160, 169, 170, 171, 172, 173, 174, 175, 184, 185, 186, 187, 188, 189, 198, 199, 200, 201, 202, 203, 204, 214, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 245, 246, 247, 248, 260, 261, 262, 263, 275, 276, 277, 278, 290, 291, 292, 293, 305, 306, 307, 308], "bbox": [0.25934426229508195, 0.1528125, 0.6770023419203747, 0.9078593749999999], "iscrowd": 0, "area": 48170.67435000001, "rle": {"size": [640, 427], "counts": "^gU28fc0:E;E;E;F9F2N2O2M2N3M2O1O2N1N2O]On^O\\NPa0e1d02WNXNXAi1f>^NTAc1k>cNo@_1o>hNj@Y1U?mNe@U1Y?QOa@P1^?WOZ@k0e?[OU@g0i?h1N3N1O1O2N1O1O2M2O2N1O1O2eAbKe3Jh4d4cGiK^3Em4c4cGPLX3@S5`4dGXLR3YOZ5^4cGaLk2TO`5\\4cGiLe2mNf5Z4dGQM_2gNl5X4cGYMR2iNY6n3dGaMb1lNj6b3cGjMR1POY7W3bGSNc0ROi7k2bG\\N4UOY8_2_GfNGWOh8S2_G`1_8`N_Gc1a8]N[Gg1d8YNZGj1e8h410O1O100O10OmM]EjJd:g4lEWKU:h4mEWKS:h4oEVKR:h4RFUKn9k4TFTKl9k4VFSKk9k4YFRKc9R5_FmJR9a5PG]Jb8P6bGnIo7`6SH^I_7P7cHnHo6`7TIXHj6n7c21000O10O2O001N2O1N2O1N2O1N2O1N2O1N12O1O2N1O1O1O1O1O1aITEh3n:bJmFW5UB=D=B>B>F9M4L4L4L3M4L4L4L3M4L4Ld1]Nnne2"}, "label": "a young boy inside a house holding a wii controller in both hands"}]} {"id": 306967, "image": "COCO_train2014_000000306967.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman standing in a dugout wearing a black jacket\"."}], "task": "refering", "answer_template": "The \"a woman standing in a dugout wearing a black jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 241, 242, 243, 264, 265, 266, 287, 288, 289, 290, 310, 311, 312, 313, 333, 334, 335, 336, 337], "bbox": [0.475703125, 0.6447058823529411, 0.676546875, 1.0], "iscrowd": 0, "area": 10515.034150000001, "rle": {"size": [425, 640], "counts": "cnn34S=3L4M4K4M3L4M3L4M3mNCUEa0i:Q1M3N2M4M2M3N2M3N2M5YOj0F;I6I0010O0100O1O100O1O100O1O10000001O0000000000001O00000000001O8H:F7I1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O2N4L4L4L4L4L4L4L2N100O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N1O2NP`e2"}, "label": "a woman standing in a dugout wearing a black jacket"}]} {"id": 306967, "image": "COCO_train2014_000000306967.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurried person standing in the dugout with beige shirt and a white top on\"."}], "task": "refering", "answer_template": "The \"a blurried person standing in the dugout with beige shirt and a white top on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 241, 242, 243, 264, 265, 266, 287, 288, 289, 290, 310, 311, 312, 313, 333, 334, 335, 336, 337], "bbox": [0.475703125, 0.6447058823529411, 0.676546875, 1.0], "iscrowd": 0, "area": 10515.034150000001, "rle": {"size": [425, 640], "counts": "cnn34S=3L4M4K4M3L4M3L4M3mNCUEa0i:Q1M3N2M4M2M3N2M3N2M5YOj0F;I6I0010O0100O1O100O1O100O1O10000001O0000000000001O00000000001O8H:F7I1O1O1O1O2N1O1O1O1O1O1O1O2N1O1O2N4L4L4L4L4L4L4L2N100O2N1O2N1O2N1O2N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N1O2NP`e2"}, "label": "a blurried person standing in the dugout with beige shirt and a white top on"}]} {"id": 528151, "image": "COCO_train2014_000000528151.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"liquor on a glass\"."}], "task": "refering", "answer_template": "The \"liquor on a glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 36, 37, 38, 39, 58, 59, 60, 61, 81, 82, 103, 125, 146, 147, 148, 168, 169, 170], "bbox": [0.631454248366013, 0.015049019607843139, 0.8247222222222222, 0.3611111111111111], "iscrowd": 0, "area": 10494.471400000008, "rle": {"size": [612, 612], "counts": "^_W74ob03L4M2M3N1N3O1O1O0O2O1`J^O\\Hc0c7GSH9m71iG0V8;^GGa8?YGBf8?YGBf8`0XG@h8b0VG_Oi8c0UG^Oj8d0SG]Om8e0QG[OP9e0oF[OQ9g0mFYOS9i0kFXOT9j0iFWOW9k0gFUOY9m0eFSO[9o0cFQO]9P1bFQO]9Q1`FPO`9Q1_FoNa9R1^FnNb9S1]FmNc9T1ZFoNe9R1SFUOm9l0kE[OU:f0dE@\\:a0\\EFd:;TEMk:5mD1S;0eD7[;J]D=c;DVDb0j;_OnCi0Q in this image.", "objects": [{"patches": [14, 15, 16, 17, 18, 36, 37, 38, 39, 58, 59, 60, 61, 81, 82, 103, 125, 146, 147, 148, 168, 169, 170], "bbox": [0.631454248366013, 0.015049019607843139, 0.8247222222222222, 0.3611111111111111], "iscrowd": 0, "area": 10494.471400000008, "rle": {"size": [612, 612], "counts": "^_W74ob03L4M2M3N1N3O1O1O0O2O1`J^O\\Hc0c7GSH9m71iG0V8;^GGa8?YGBf8?YGBf8`0XG@h8b0VG_Oi8c0UG^Oj8d0SG]Om8e0QG[OP9e0oF[OQ9g0mFYOS9i0kFXOT9j0iFWOW9k0gFUOY9m0eFSO[9o0cFQO]9P1bFQO]9Q1`FPO`9Q1_FoNa9R1^FnNb9S1]FmNc9T1ZFoNe9R1SFUOm9l0kE[OU:f0dE@\\:a0\\EFd:;TEMk:5mD1S;0eD7[;J]D=c;DVDb0j;_OnCi0Q in this image.", "objects": [{"patches": [77, 78, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 193, 194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 220, 221, 222, 223, 224, 239, 240, 242, 243, 244, 245, 246, 262, 263, 265, 266, 267, 268, 269, 285, 289, 290, 291, 292, 314, 315, 337, 338], "bbox": [0.292015625, 0.24730046948356807, 0.7529375, 0.9569953051643191], "iscrowd": 0, "area": 46854.65040000001, "rle": {"size": [426, 640], "counts": "em]23V=3L4M3L2O2N1O2O0O2N1O1O1O1O1O1N3N1O01O04M4L2M3N1O2N>B2N2M2O1O1O1O1O1N2O2N1N2O1O1N2O001O1N101O1O1N2O1O1O1N101O100O1O101N1O1O2N1O1PJiLk1Y3QNjLn1X3nMjLR2X3jMjLW2W3eMkL[2X3`MkL_2X3]MiLd2Y3XMhLh2Y3nLYL[M?g5Y3kLQMV3o2gLTMX3m2eLUM\\3k2aLWM_3j2^LXMb3i2[LZMd3g2YL[Mg3f2VL\\Mj3e2PL_MR4a2iKcMW4^2[KnMf4T2hJ[NY5f1cJ\\N^5e1^J^Nb5e1YJ\\Nh5f1TJ[Nm5h1nIZNR6^41O100O1O100O100O1O1O1O1N2hNX1K5L4L4M3N2N3L3O100O100O2O000O100O10001O00000000001O01O000000010O001O001O001O001O001O4L2N2N3M3M2SHaK6VOY6]6bIdI^6e6O00001O000001O00000O10000O2O0O10000eNZIQLg6o3\\InKe6Q4\\ImKe6R4]ImKc6S4_IjKc6U4^IjKb6U4aIhK`6X4bIfK_6Y4cIdK^6\\4dI`K^6a4dI[K]6e4T11O001O000010O01O1O10O01O100O1O0dIjJd4V5[KlJe4T5XKoJg4Q5VKRKj4o4SKSKn4l4oJWKQ5j4lJXKU5g4hJ\\KX5d4fJ^KZ5b4dJ`K]5`4aJaK_5_4`JbKa5]4^JdKb5\\4\\JfKe5Y4ZJhKf5Y4WJiKi5W4UJkKk5V4SJkKm5U4RJlKn5U4oImKQ6T4mImKT6S4jImKW6S4gIoKY6R4eIoK[6`50O1O1O2N1N200102M3jJZIb3i6YLZIf3i6VLYIj3m6nKTIR4o6jKSIU4P7gKQIY4R7`KRIa4l7O0001O00001N10001O00001N101O00001N1nMPKjKQ5S4SKjKn4U4VKhKk4V4XKgKi4X4[KeKf4Y4_KbKb4^4aK_K`4^4eK^K\\4`4iK]KX4_4nK]KS4b4PL[KQ4c4TLXKo3b4YLYKi3c4]LZKe3^4fL\\K\\3c4g2M2O2M2O2N2N2N2O1N2N2D in this image.", "objects": [{"patches": [77, 78, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 193, 194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 220, 221, 222, 223, 224, 239, 240, 242, 243, 244, 245, 246, 262, 263, 265, 266, 267, 268, 269, 285, 289, 290, 291, 292, 314, 315, 337, 338], "bbox": [0.292015625, 0.24730046948356807, 0.7529375, 0.9569953051643191], "iscrowd": 0, "area": 46854.65040000001, "rle": {"size": [426, 640], "counts": "em]23V=3L4M3L2O2N1O2O0O2N1O1O1O1O1O1N3N1O01O04M4L2M3N1O2N>B2N2M2O1O1O1O1O1N2O2N1N2O1O1N2O001O1N101O1O1N2O1O1O1N101O100O1O101N1O1O2N1O1PJiLk1Y3QNjLn1X3nMjLR2X3jMjLW2W3eMkL[2X3`MkL_2X3]MiLd2Y3XMhLh2Y3nLYL[M?g5Y3kLQMV3o2gLTMX3m2eLUM\\3k2aLWM_3j2^LXMb3i2[LZMd3g2YL[Mg3f2VL\\Mj3e2PL_MR4a2iKcMW4^2[KnMf4T2hJ[NY5f1cJ\\N^5e1^J^Nb5e1YJ\\Nh5f1TJ[Nm5h1nIZNR6^41O100O1O100O100O1O1O1O1N2hNX1K5L4L4M3N2N3L3O100O100O2O000O100O10001O00000000001O01O000000010O001O001O001O001O001O4L2N2N3M3M2SHaK6VOY6]6bIdI^6e6O00001O000001O00000O10000O2O0O10000eNZIQLg6o3\\InKe6Q4\\ImKe6R4]ImKc6S4_IjKc6U4^IjKb6U4aIhK`6X4bIfK_6Y4cIdK^6\\4dI`K^6a4dI[K]6e4T11O001O000010O01O1O10O01O100O1O0dIjJd4V5[KlJe4T5XKoJg4Q5VKRKj4o4SKSKn4l4oJWKQ5j4lJXKU5g4hJ\\KX5d4fJ^KZ5b4dJ`K]5`4aJaK_5_4`JbKa5]4^JdKb5\\4\\JfKe5Y4ZJhKf5Y4WJiKi5W4UJkKk5V4SJkKm5U4RJlKn5U4oImKQ6T4mImKT6S4jImKW6S4gIoKY6R4eIoK[6`50O1O1O2N1N200102M3jJZIb3i6YLZIf3i6VLYIj3m6nKTIR4o6jKSIU4P7gKQIY4R7`KRIa4l7O0001O00001N10001O00001N101O00001N1nMPKjKQ5S4SKjKn4U4VKhKk4V4XKgKi4X4[KeKf4Y4_KbKb4^4aK_K`4^4eK^K\\4`4iK]KX4_4nK]KS4b4PL[KQ4c4TLXKo3b4YLYKi3c4]LZKe3^4fL\\K\\3c4g2M2O2M2O2N2N2N2O1N2N2D in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 38, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 291, 292, 293, 294, 295, 296, 297, 314, 315, 316, 317, 318, 319, 337, 338, 339, 340, 341], "bbox": [0.51046875, 0.0, 0.9910781249999999, 0.9858313817330211], "iscrowd": 0, "area": 78146.13, "rle": {"size": [427, 640], "counts": "QeX43U=4M4K4L4M4L2O1O1O2M2O1O1N2O1iNQOQFQ1k9SOSFn0j9VOVFi0g9ZOYFf0e9]OZFc0e9_OZFa0f9_O[Fa0c9@]F`0b9A^F?b9A^F?b9A_F>`9C`F=`9D_F=`9C`F=`9CaF<_9DaF<^9EbF;^9EcF:]9FcF;\\9EeF:[9GdF9\\9GeF8[9HeF8[9HfF7Z9IfF7Y9JhF5X9JjF5V9KjF5V9KkF4U9LkF4U9LlF3T9MkF5U9JiF8W9HhF8X9IgF8Y9HfF9Z9GdF;\\9EcF<\\9DdF=\\9CcF>\\9CbF?\\9CcF>[9DdF=Z9EdF=[9DdF=\\9CcF>\\9CcF>\\9CbF?^9g1O100O100O100O100O1O2O0O2N1O1O1O2N100O1O2N1O1O1O2N1N2O1N2N2O0O2O1N2N2O1\\KaK:`4]N^KZN5X3^4[N`K\\N4V3^4\\N`K^N2T3_4]NaK^N1T3Y6aLiI;OQ3^7nLcHP3_7PMaHo2`7PMbHm2`7RMaHl2a7TM_Hk2b7TM_Hj2c7VM^Hh2c7WM^Hg2d7YM\\He2f7ZM\\Hd2e7\\M[Hb2g7]MZHa2h7_MYH_2h7`MYH^2i7aMXH]2j7cMVH\\2k7cMVH[2l7eMTHZ2m7eMTHZ2m7fMTHX2m7gMVHU2l7kMVHR2k7mMYHo1h7PN\\Hl1e7SN^Hi1e7mMXGFV1Z2c7PNgHm1Z7SNgHk1Z7UNgHh1[7XNfHf1[7ZNfHd1[7\\NeHc1\\7]NeH`1]7`NdH^1]7bNnHQ1T7oNnHn0S7ROnHl0S7TOnHi0T7WOnHf0S7ZOnHc0T7]OmHa0T7_OnH=T7CmH;T7DnH:S7FmH8U7HkH7V7IjH5X7KhH4Y7LfH4[7LeH2]7NcH1^7ObHO`71`HNa72^HNc72]HLe74[HKf75ZHJg76XHIj77VHHk78d2000000O100000000000000000000000000000000000000O10001O000000000000000000000000000000000O1000000000000000000001O000M4L3M4L3M3N3L3M5K5K6\\DdNQ;a1kD_NU;e1gD\\NX;g1eDYN[;k101N101O000O2O001O0O2O00001O0O2O001N2N2O2M2N2N2N2O1N3@?O1N2O1O1O1O2N1O1N2O1O1M4K4K6K6J6I7K5aHTKX6Q5eIQKW6T5gImJV6W5gIkJU6Z5iIfJT6_5iIcJS6b5kI_JQ6f5mI[Jo5j5nIWJP6m5nITJn5Q6oIQJm5T6QJPJh5U6UJoIe5P7J6K5J6J6J6DB>A?B?A>Ce0[Od0[OX\\2"}, "label": "man in black jacket getting a drink"}]} {"id": 438071, "image": "COCO_train2014_000000438071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the black sweatshirt\"."}], "task": "refering", "answer_template": "The \"the man in the black sweatshirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 38, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 291, 292, 293, 294, 295, 296, 297, 314, 315, 316, 317, 318, 319, 337, 338, 339, 340, 341], "bbox": [0.51046875, 0.0, 0.9910781249999999, 0.9858313817330211], "iscrowd": 0, "area": 78146.13, "rle": {"size": [427, 640], "counts": "QeX43U=4M4K4L4M4L2O1O1O2M2O1O1N2O1iNQOQFQ1k9SOSFn0j9VOVFi0g9ZOYFf0e9]OZFc0e9_OZFa0f9_O[Fa0c9@]F`0b9A^F?b9A^F?b9A_F>`9C`F=`9D_F=`9C`F=`9CaF<_9DaF<^9EbF;^9EcF:]9FcF;\\9EeF:[9GdF9\\9GeF8[9HeF8[9HfF7Z9IfF7Y9JhF5X9JjF5V9KjF5V9KkF4U9LkF4U9LlF3T9MkF5U9JiF8W9HhF8X9IgF8Y9HfF9Z9GdF;\\9EcF<\\9DdF=\\9CcF>\\9CbF?\\9CcF>[9DdF=Z9EdF=[9DdF=\\9CcF>\\9CcF>\\9CbF?^9g1O100O100O100O100O1O2O0O2N1O1O1O2N100O1O2N1O1O1O2N1N2O1N2N2O0O2O1N2N2O1\\KaK:`4]N^KZN5X3^4[N`K\\N4V3^4\\N`K^N2T3_4]NaK^N1T3Y6aLiI;OQ3^7nLcHP3_7PMaHo2`7PMbHm2`7RMaHl2a7TM_Hk2b7TM_Hj2c7VM^Hh2c7WM^Hg2d7YM\\He2f7ZM\\Hd2e7\\M[Hb2g7]MZHa2h7_MYH_2h7`MYH^2i7aMXH]2j7cMVH\\2k7cMVH[2l7eMTHZ2m7eMTHZ2m7fMTHX2m7gMVHU2l7kMVHR2k7mMYHo1h7PN\\Hl1e7SN^Hi1e7mMXGFV1Z2c7PNgHm1Z7SNgHk1Z7UNgHh1[7XNfHf1[7ZNfHd1[7\\NeHc1\\7]NeH`1]7`NdH^1]7bNnHQ1T7oNnHn0S7ROnHl0S7TOnHi0T7WOnHf0S7ZOnHc0T7]OmHa0T7_OnH=T7CmH;T7DnH:S7FmH8U7HkH7V7IjH5X7KhH4Y7LfH4[7LeH2]7NcH1^7ObHO`71`HNa72^HNc72]HLe74[HKf75ZHJg76XHIj77VHHk78d2000000O100000000000000000000000000000000000000O10001O000000000000000000000000000000000O1000000000000000000001O000M4L3M4L3M3N3L3M5K5K6\\DdNQ;a1kD_NU;e1gD\\NX;g1eDYN[;k101N101O000O2O001O0O2O00001O0O2O001N2N2O2M2N2N2N2O1N3@?O1N2O1O1O1O2N1O1N2O1O1M4K4K6K6J6I7K5aHTKX6Q5eIQKW6T5gImJV6W5gIkJU6Z5iIfJT6_5iIcJS6b5kI_JQ6f5mI[Jo5j5nIWJP6m5nITJn5Q6oIQJm5T6QJPJh5U6UJoIe5P7J6K5J6J6J6DB>A?B?A>Ce0[Od0[OX\\2"}, "label": "the man in the black sweatshirt"}]} {"id": 438071, "image": "COCO_train2014_000000438071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman pouring wine , a wine glass , and a basket are all within the red square . the woman is wearing an apron\"."}], "task": "refering", "answer_template": "The \"a woman pouring wine , a wine glass , and a basket are all within the red square . the woman is wearing an apron\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 74, 75, 76, 77, 97, 98, 99, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 258, 259], "bbox": [0.160421875, 0.15730679156908667, 0.503765625, 0.7730444964871195], "iscrowd": 0, "area": 25173.73975, "rle": {"size": [427, 640], "counts": "hV[1:o<:E;E;F9G5J6K2N2N2N2N2N2M3N2N1K6G9F:E;E;F:K5M3M3N2M3M3N2M3M3O1N2O1N2O1O1O10000O2O0O10000O2O0O2O1O1N2O1N2N2N2N2J6J6G:F9H8G7L4O2NfNbJjJ]5S5jJkJT5T5PKjJo4U5UKiJj4V5ZKiJe4U5_KiJa4S5dKlJ\\4P5jKoJV4l4PLRKP4m4SLQKn3m4ULRKj3m4YLQKh3m4[LQKe3n4^LQKb3o4_LoJb3Q5^LoJb3Q5_LmJb3S5`LjJa3V5`LiJ`3W5aLgJ`3Y5aLeJ`3[5aLdJ_3[5cLcJ^3]5X2O100O100O1001O00001O00001O1O1O1WJ`Je3a5XL`Ji3a5TLaJl3`5QLaJP4a5lK`JU4a5hKaJX4c5bK^J_4e5\\K\\Jf4f5UK[Jm4P2SJ[1j0gLT5h1WJg1=bL]5c1[Jo1IgLn5U1]JP3d5kLaJS3a5gLdJY3]5bLhJ\\3[5^LiJb3X5ZLkJe3X5WLkJh3V5ULlJj3W5QLmJn3U5nKmJR4W5hKlJP3`NfMe8Y2_GdMa8[2bGbM`8\\2cGbM]8]2fG`M\\8^2hG_MX8`2kG]MV8b2lG]MT8b2nG\\MS8c2oG\\MQ8c2QH[MP8S1bGA`0[On7h0oGK6ZOl7j0PHJ6[Oj7j0RHI6YOk7l0RHIZ96hFIX97iFGX98jFGV98jFIV97jFIV96jFKV94kFLU95iFLX93hFMX93hFMX93gFNZ91fFOZ91eF0[90eF0[90dF1]9ObF1^9OaF2_9NaF2`9M`F2a9N_F1b9O^F0c90^FNc93\\FKf95ZFJg96YFIh97XFHi98XFFi9:WFDk9SF@o9`0RF^Oo9b0QF]OP:c0PF[OR:e0nEZOS:f0mEXOU:i0V1000010OkCXOg;i0TD[Ol;e0oC@Q in this image.", "objects": [{"patches": [52, 53, 54, 74, 75, 76, 77, 97, 98, 99, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 258, 259], "bbox": [0.160421875, 0.15730679156908667, 0.503765625, 0.7730444964871195], "iscrowd": 0, "area": 25173.73975, "rle": {"size": [427, 640], "counts": "hV[1:o<:E;E;F9G5J6K2N2N2N2N2N2M3N2N1K6G9F:E;E;F:K5M3M3N2M3M3N2M3M3O1N2O1N2O1O1O10000O2O0O10000O2O0O2O1O1N2O1N2N2N2N2J6J6G:F9H8G7L4O2NfNbJjJ]5S5jJkJT5T5PKjJo4U5UKiJj4V5ZKiJe4U5_KiJa4S5dKlJ\\4P5jKoJV4l4PLRKP4m4SLQKn3m4ULRKj3m4YLQKh3m4[LQKe3n4^LQKb3o4_LoJb3Q5^LoJb3Q5_LmJb3S5`LjJa3V5`LiJ`3W5aLgJ`3Y5aLeJ`3[5aLdJ_3[5cLcJ^3]5X2O100O100O1001O00001O00001O1O1O1WJ`Je3a5XL`Ji3a5TLaJl3`5QLaJP4a5lK`JU4a5hKaJX4c5bK^J_4e5\\K\\Jf4f5UK[Jm4P2SJ[1j0gLT5h1WJg1=bL]5c1[Jo1IgLn5U1]JP3d5kLaJS3a5gLdJY3]5bLhJ\\3[5^LiJb3X5ZLkJe3X5WLkJh3V5ULlJj3W5QLmJn3U5nKmJR4W5hKlJP3`NfMe8Y2_GdMa8[2bGbM`8\\2cGbM]8]2fG`M\\8^2hG_MX8`2kG]MV8b2lG]MT8b2nG\\MS8c2oG\\MQ8c2QH[MP8S1bGA`0[On7h0oGK6ZOl7j0PHJ6[Oj7j0RHI6YOk7l0RHIZ96hFIX97iFGX98jFGV98jFIV97jFIV96jFKV94kFLU95iFLX93hFMX93hFMX93gFNZ91fFOZ91eF0[90eF0[90dF1]9ObF1^9OaF2_9NaF2`9M`F2a9N_F1b9O^F0c90^FNc93\\FKf95ZFJg96YFIh97XFHi98XFFi9:WFDk9SF@o9`0RF^Oo9b0QF]OP:c0PF[OR:e0nEZOS:f0mEXOU:i0V1000010OkCXOg;i0TD[Ol;e0oC@Q in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 26, 27, 28, 29, 30, 31, 32, 48, 49, 50, 51, 52, 53, 54, 55, 56, 71, 72, 73, 74, 75, 76, 77, 78, 79, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336], "bbox": [0.028671875000000003, 0.016103286384976528, 0.652046875, 0.9911502347417841], "iscrowd": 0, "area": 113011.38565, "rle": {"size": [426, 640], "counts": "Xj7>e<9G:G8L4kEdNV8a1aGeN]8_1\\GfNb8_1VGfNg8c1nFbNP9j1aF[N]9j2N2N2N2N2M3N2O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1`KlLiNU3T1QMhNP3T1VMiNk2S1[MjNf2S1`MiNa2S1eMkN[2Q1kMlNV2Q1oMmNQ2o0UNnNl1o0ZNnNf1o0_NnNb1o0cNoN]1n0hNoNY1o0kNoNU1n0POoNQ1o0WOkNi0R1@gNa0V1FfN:X1LdN4Y12dNN[17`NJ^1;_NE`1<`ND^1>bNB]1?cNA[1a0eN_OZ1a0gN_OW1c0hN^OW1c0hN^OV1d0iN]OV1d0iN]OU1e0jN\\OU1e0jN\\OT1f0kN[OT1f0iN]OV1d0hN^OW1c0fN@Y1Y600O1O1O1O1O100O1O1O1O1O1O100O1O100O1O1O100O1O1O100O1O100O10000O100O10000O100O10000O100O10000O100O100O100O100O100O100O10000O100O100O100O1000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000001O0000001O0000001O0000001O00001O1O001O001O001O1O001O00002N3M2N3M3M001O001O001O001O001O001O002N1O1O2N1O1O2N1O1O2N1O1O2N1O2N1O2N1O2N1O1O2N1O1O1O1O1O1O1O2N2N1O2N2N2N3M3M3M2N3M2N2N2N2N2N1O2UI\\Ll3f3PL\\LP4g3kK\\LT4P4]KSLc4Z5nIiJQ6\\5gIgJY6^5`IeJ_6S60001O001O001O001O001O001O00001O001O001O001O001O00001O001O001O001O001O001O00001O001O4L6J7I6J6J6J6J6J7I6J6J6J5K2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2Nb[l2"}, "label": "the blond boy"}]} {"id": 102208, "image": "COCO_train2014_000000102208.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy with blonde hair\"."}], "task": "refering", "answer_template": "The \"a boy with blonde hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 26, 27, 28, 29, 30, 31, 32, 48, 49, 50, 51, 52, 53, 54, 55, 56, 71, 72, 73, 74, 75, 76, 77, 78, 79, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336], "bbox": [0.028671875000000003, 0.016103286384976528, 0.652046875, 0.9911502347417841], "iscrowd": 0, "area": 113011.38565, "rle": {"size": [426, 640], "counts": "Xj7>e<9G:G8L4kEdNV8a1aGeN]8_1\\GfNb8_1VGfNg8c1nFbNP9j1aF[N]9j2N2N2N2N2M3N2O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1`KlLiNU3T1QMhNP3T1VMiNk2S1[MjNf2S1`MiNa2S1eMkN[2Q1kMlNV2Q1oMmNQ2o0UNnNl1o0ZNnNf1o0_NnNb1o0cNoN]1n0hNoNY1o0kNoNU1n0POoNQ1o0WOkNi0R1@gNa0V1FfN:X1LdN4Y12dNN[17`NJ^1;_NE`1<`ND^1>bNB]1?cNA[1a0eN_OZ1a0gN_OW1c0hN^OW1c0hN^OV1d0iN]OV1d0iN]OU1e0jN\\OU1e0jN\\OT1f0kN[OT1f0iN]OV1d0hN^OW1c0fN@Y1Y600O1O1O1O1O100O1O1O1O1O1O100O1O100O1O1O100O1O1O100O1O100O10000O100O10000O100O10000O100O10000O100O100O100O100O100O100O10000O100O100O100O1000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000001O0000001O0000001O0000001O00001O1O001O001O001O1O001O00002N3M2N3M3M001O001O001O001O001O001O002N1O1O2N1O1O2N1O1O2N1O1O2N1O2N1O2N1O2N1O1O2N1O1O1O1O1O1O1O2N2N1O2N2N2N3M3M3M2N3M2N2N2N2N2N1O2UI\\Ll3f3PL\\LP4g3kK\\LT4P4]KSLc4Z5nIiJQ6\\5gIgJY6^5`IeJ_6S60001O001O001O001O001O001O00001O001O001O001O001O00001O001O001O001O001O001O00001O001O4L6J7I6J6J6J6J6J7I6J6J6J5K2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2Nb[l2"}, "label": "a boy with blonde hair"}]} {"id": 102208, "image": "COCO_train2014_000000102208.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown haired girl with an orange in her mouth\"."}], "task": "refering", "answer_template": "The \"a brown haired girl with an orange in her mouth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 41, 59, 60, 61, 62, 63, 64, 65, 81, 82, 83, 84, 85, 86, 87, 88, 89, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.5325, 0.11011737089201877, 1.0, 0.9820187793427231], "iscrowd": 0, "area": 90157.92564999999, "rle": {"size": [426, 640], "counts": "VP^4k0T in this image.", "objects": [{"patches": [38, 39, 40, 41, 59, 60, 61, 62, 63, 64, 65, 81, 82, 83, 84, 85, 86, 87, 88, 89, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.5325, 0.11011737089201877, 1.0, 0.9820187793427231], "iscrowd": 0, "area": 90157.92564999999, "rle": {"size": [426, 640], "counts": "VP^4k0T in this image.", "objects": [{"patches": [245, 246, 247, 248, 260, 261, 262, 263, 275, 276, 277, 278, 289, 290, 291, 292, 304, 305, 306, 307, 320, 321, 322, 335, 336, 337], "bbox": [0.3153395784543326, 0.70909375, 0.5566978922716628, 0.9870156250000001], "iscrowd": 0, "area": 13125.2631, "rle": {"size": [640, 427], "counts": "[md27ec08G:G8H8H8H8H8H9F9H5K5K5J6K5K6I6K5K5J6K5K5M3L4M3O2M2O001O001O1O0O2O1O001O001N2O001O001O1N101O001OO1000O10000000000000O101O0000000N2N3L3M3M3M3M3M4L3M3M3M3N3L3M4L3N3L3M4L3M4L3L5J5L5K4K5L5K4K6K4F;^Oa0_Oa0A`0O0O]Yf3"}, "label": "the clear wine glass with the lower level of wine"}]} {"id": 20291, "image": "COCO_train2014_000000020291.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"central wineglass on table\"."}], "task": "refering", "answer_template": "The \"central wineglass on table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [245, 246, 247, 248, 260, 261, 262, 263, 275, 276, 277, 278, 289, 290, 291, 292, 304, 305, 306, 307, 320, 321, 322, 335, 336, 337], "bbox": [0.3153395784543326, 0.70909375, 0.5566978922716628, 0.9870156250000001], "iscrowd": 0, "area": 13125.2631, "rle": {"size": [640, 427], "counts": "[md27ec08G:G8H8H8H8H8H9F9H5K5K5J6K5K6I6K5K5J6K5K5M3L4M3O2M2O001O001O1O0O2O1O001O001N2O001O001O1N101O001OO1000O10000000000000O101O0000000N2N3L3M3M3M3M3M4L3M3M3M3N3L3M4L3N3L3M4L3M4L3L5J5L5K4K5L5K4K6K4F;^Oa0_Oa0A`0O0O]Yf3"}, "label": "central wineglass on table"}]} {"id": 347972, "image": "COCO_train2014_000000347972.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green and orange traincar standing to the left of the fully visible traincar\"."}], "task": "refering", "answer_template": "The \"a green and orange traincar standing to the left of the fully visible traincar\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 184, 185, 186, 201, 202, 203, 218, 219, 220, 236, 237, 253, 254, 270, 271, 287, 288, 305, 322], "bbox": [0.8629166666666667, 0.40225, 1.0, 0.8112343750000001], "iscrowd": 0, "area": 11879.468450000004, "rle": {"size": [640, 480], "counts": "YPS84jc0;E in this image.", "objects": [{"patches": [167, 168, 169, 184, 185, 186, 201, 202, 203, 218, 219, 220, 236, 237, 253, 254, 270, 271, 287, 288, 305, 322], "bbox": [0.8629166666666667, 0.40225, 1.0, 0.8112343750000001], "iscrowd": 0, "area": 11879.468450000004, "rle": {"size": [640, 480], "counts": "YPS84jc0;E in this image.", "objects": [{"patches": [187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259], "bbox": [0.001203125, 0.6748611111111111, 0.3169375, 0.9032222222222221], "iscrowd": 0, "area": 10370.452600000004, "rle": {"size": [360, 640], "counts": "kc0T1T:0O2O00000O100000000O1000000O2O0000000O1000001O0000000O1000000000002M2O1O2N1O1O001N101O00001O000O10O10000O1000O0100O10O01O001O001O0010O100O100O100O100O1O010O100O100O100O100O100O100O100O10000O1000000000O1000000000000000O100000000000000000000000000O1000O1000000000000000000000O1000000000000000000001O0O10000000001O0000000000001O00000000001O00000O1I7M302M2F:ROaa2Hd_M2N00000001O00000000000000000OaZi4"}, "label": "the grey car"}]} {"id": 388935, "image": "COCO_train2014_000000388935.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver mercedes benz next to a red sightseeing bus\"."}], "task": "refering", "answer_template": "The \"a silver mercedes benz next to a red sightseeing bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259], "bbox": [0.001203125, 0.6748611111111111, 0.3169375, 0.9032222222222221], "iscrowd": 0, "area": 10370.452600000004, "rle": {"size": [360, 640], "counts": "kc0T1T:0O2O00000O100000000O1000000O2O0000000O1000001O0000000O1000000000002M2O1O2N1O1O001N101O00001O000O10O10000O1000O0100O10O01O001O001O0010O100O100O100O100O1O010O100O100O100O100O100O100O100O10000O1000000000O1000000000000000O100000000000000000000000000O1000O1000000000000000000000O1000000000000000000001O0O10000000001O0000000000001O00000000001O00000O1I7M302M2F:ROaa2Hd_M2N00000001O00000000000000000OaZi4"}, "label": "a silver mercedes benz next to a red sightseeing bus"}]} {"id": 388935, "image": "COCO_train2014_000000388935.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and white double decker sight seeing bus\"."}], "task": "refering", "answer_template": "The \"a red and white double decker sight seeing bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 52, 53, 54, 55, 56, 57, 58, 59, 60, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 266, 267, 268], "bbox": [0.27809375, 0.11291666666666667, 0.86965625, 0.9129166666666666], "iscrowd": 0, "area": 82302.23644999998, "rle": {"size": [360, 640], "counts": "gkn19]:c0\\Oc0^Ob0]Oc0^Ob0^Ob0]Oc0^Ob0]Oc0^Ob0^Oc0\\Oc0G9O100O10000O100O2O000O100O100O10000O2O0O10000O100O100O10001N100O100O10000O100O2O000O100O100O10000O2O0O10001N2O1N101O1O1O001O001O000000001O000000001O00000000000000000000O100000000000000O100000000000000O1000000000000001O000000000000000000000001O001O001O001O01O01O001O001O001O001O00001O001O001O0010O01O00001O001O001O001O001O00001O0010O01O010O000010O01O010O001O010O000010O01O010O001O001M200O2N1O2O0O2N1O2O0O2N1O101N1O2N101N1O2O00001O001O0O2O001O001O00001N101O001O001O0O2O00001O001O001N101O001O00001N101O001O001O000O2O001O001O001N101O00001O001O0O2O001O1O3M4L3M3L5L3M3M4L3M4L3L4M3M001O001O001N101O0100O100O1O100O010ON3M2O2M2N3N1N3O010O001O01O01O0010O01O0010O01O0O2N1O2N1O2N1O101O0O2O001N101O0O01O1O001O1O1O1N101O1O1O18G=D in this image.", "objects": [{"patches": [31, 32, 33, 34, 52, 53, 54, 55, 56, 57, 58, 59, 60, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 266, 267, 268], "bbox": [0.27809375, 0.11291666666666667, 0.86965625, 0.9129166666666666], "iscrowd": 0, "area": 82302.23644999998, "rle": {"size": [360, 640], "counts": "gkn19]:c0\\Oc0^Ob0]Oc0^Ob0^Ob0]Oc0^Ob0]Oc0^Ob0^Oc0\\Oc0G9O100O10000O100O2O000O100O100O10000O2O0O10000O100O100O10001N100O100O10000O100O2O000O100O100O10000O2O0O10001N2O1N101O1O1O001O001O000000001O000000001O00000000000000000000O100000000000000O100000000000000O1000000000000001O000000000000000000000001O001O001O001O01O01O001O001O001O001O00001O001O001O0010O01O00001O001O001O001O001O00001O0010O01O010O000010O01O010O001O010O000010O01O010O001O001M200O2N1O2O0O2N1O2O0O2N1O101N1O2N101N1O2O00001O001O0O2O001O001O00001N101O001O001O0O2O00001O001O001N101O001O00001N101O001O001O000O2O001O001O001N101O00001O001O0O2O001O1O3M4L3M3L5L3M3M4L3M4L3L4M3M001O001O001N101O0100O100O1O100O010ON3M2O2M2N3N1N3O010O001O01O01O0010O01O0010O01O0O2N1O2N1O2N1O101O0O2O001N101O0O01O1O001O1O1O1N101O1O1O18G=D in this image.", "objects": [{"patches": [231, 232, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 326, 327, 328, 349, 350], "bbox": [0.04184375, 0.587125, 0.550421875, 0.9261874999999999], "iscrowd": 0, "area": 13097.747800000006, "rle": {"size": [480, 640], "counts": "cn<2m>0O2O100O100O10O10N2N2N2O10000PBGY=:bB0X=0eB5X=LgB6X=JgB9W=IfB:W=HgB:X=GfB]<_OgC`0Y<^OiCa0Y<]OgCd0[N2O100O1000000O2O000J6O100O2N100O1O10001O00000001O01O00000000001O01O01O0010O00010O01O00001O010O00001O001O00001O0K6O000K\\BnNe=P1^BnNb=Q1701O00001O010O000O2N1O1O2O0JkAAU>>mABS>=mACS>7ZBHG1o=7ZBHG1o=7[BGF2o=7[BGG1o=7ZBHG2n=6\\BGG2m=7\\BGH1l=8\\BGI1j=8dBH]=7dBI[=7fBHZ=8fBHZ=8gBHX=8i0O0000001O000000001O000001O00000000000000000000001O0000000000000000000000001O0001O001O1O001O1O001O1O001OYdV4"}, "label": "peacock standing in the grass"}]} {"id": 462664, "image": "COCO_train2014_000000462664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a peacock walking on grass\"."}], "task": "refering", "answer_template": "The \"a peacock walking on grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [231, 232, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 326, 327, 328, 349, 350], "bbox": [0.04184375, 0.587125, 0.550421875, 0.9261874999999999], "iscrowd": 0, "area": 13097.747800000006, "rle": {"size": [480, 640], "counts": "cn<2m>0O2O100O100O10O10N2N2N2O10000PBGY=:bB0X=0eB5X=LgB6X=JgB9W=IfB:W=HgB:X=GfB]<_OgC`0Y<^OiCa0Y<]OgCd0[N2O100O1000000O2O000J6O100O2N100O1O10001O00000001O01O00000000001O01O01O0010O00010O01O00001O010O00001O001O00001O0K6O000K\\BnNe=P1^BnNb=Q1701O00001O010O000O2N1O1O2O0JkAAU>>mABS>=mACS>7ZBHG1o=7ZBHG1o=7[BGF2o=7[BGG1o=7ZBHG2n=6\\BGG2m=7\\BGH1l=8\\BGI1j=8dBH]=7dBI[=7fBHZ=8fBHZ=8gBHX=8i0O0000001O000000001O000001O00000000000000000000001O0000000000000000000000001O0001O001O1O001O1O001O1O001OYdV4"}, "label": "a peacock walking on grass"}]} {"id": 257867, "image": "COCO_train2014_000000257867.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"car with a yellow license plate in traffic\"."}], "task": "refering", "answer_template": "The \"car with a yellow license plate in traffic\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305], "bbox": [0.0074687500000000006, 0.4630823529411765, 0.31189062500000003, 0.9170117647058824], "iscrowd": 0, "area": 30372.21845, "rle": {"size": [425, 640], "counts": "lX2a1f;R2PNo1QN6JO100O001O100O1O1O10O01O100O1O1O10O01O1O1001O000O1000000000000O1000001O00000000000000000O100000001O000000000000000O100000001O000000000000000O100000001O0000000000000000000000000010O000000000000000000000001O0001O00000000001O001O0010O01O001O001O001O00001O001O010O001O001O001O001O1O3M2N3N2M2N3M2N3M3M2N3M3M2N3M2O1N2N2N2N2N2N2N2N2N2O1N2N2N2N2M3M3M3M3L5L5K4L4K6I6J7I6J7I6K6K4M4K4M4K4M3LUff5"}, "label": "car with a yellow license plate in traffic"}]} {"id": 257867, "image": "COCO_train2014_000000257867.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the small grey car to the left of the tow truck\"."}], "task": "refering", "answer_template": "The \"the small grey car to the left of the tow truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305], "bbox": [0.0074687500000000006, 0.4630823529411765, 0.31189062500000003, 0.9170117647058824], "iscrowd": 0, "area": 30372.21845, "rle": {"size": [425, 640], "counts": "lX2a1f;R2PNo1QN6JO100O001O100O1O1O10O01O100O1O1O10O01O1O1001O000O1000000000000O1000001O00000000000000000O100000001O000000000000000O100000001O000000000000000O100000001O0000000000000000000000000010O000000000000000000000001O0001O00000000001O001O0010O01O001O001O001O00001O001O010O001O001O001O001O1O3M2N3N2M2N3M2N3M3M2N3M3M2N3M2O1N2N2N2N2N2N2N2N2N2O1N2N2N2N2M3M3M3M3L5L5K4L4K6I6J7I6J7I6K6K4M4K4M4K4M3LUff5"}, "label": "the small grey car to the left of the tow truck"}]} {"id": 257867, "image": "COCO_train2014_000000257867.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver sedan next to the right side of a white and red work truck\"."}], "task": "refering", "answer_template": "The \"a silver sedan next to the right side of a white and red work truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [204, 205, 206, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 293, 294], "bbox": [0.755296875, 0.5619294117647059, 1.0, 0.8254588235294118], "iscrowd": 0, "area": 10592.2767, "rle": {"size": [425, 640], "counts": "[hX6b0d<8G9G9H5N2N2M3NZN`De1];[NcDg1\\;XNeDi1\\;100O2N100O2N01O01O00010O0001O00000010O001O1O2N1O100O100O100O101N100O100O1000000O100000O0100000000O1000O01N2O1O1O1N2O1O001O1O1O1O1O1O010O1O1O1O100O10000O10000O10O1O1N2L4L4M3N2O1O1N2O1O1O100O101N10000O100O100O10000O1O100O100O1O100O1O100O1O1O1O2O0O1O1O1N2N2N2O1N2N2N2N2N2O1N2N2O1N2N2O1N2N2O1N2O`H"}, "label": "a silver sedan next to the right side of a white and red work truck"}]} {"id": 257867, "image": "COCO_train2014_000000257867.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white car following a tow truck\"."}], "task": "refering", "answer_template": "The \"a white car following a tow truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [229, 251, 252, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344], "bbox": [0.734015625, 0.5984941176470588, 1.0, 1.0], "iscrowd": 0, "area": 15362.141150000003, "rle": {"size": [425, 640], "counts": "X_S66P=4N1N2O1N2O1O1O1J6@`0L4L4L4N2N2N2N2O1O100O1O1O100O1O100O1O100000000000000000000O100000000000000000000000000001O0000000000000000000000000000N2O1O1N2O1O100O1O100O1O1O100O1O100O10000O1000000O10000O1000000O1000000O100000000E_MnEa2P:=N2N2O100O1O1O10000000000O1000000000000000000N2N2N2N2N2N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1O1N2O1N2O1"}, "label": "a white car following a tow truck"}]} {"id": 257867, "image": "COCO_train2014_000000257867.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is a white car behind a tow truck\"."}], "task": "refering", "answer_template": "The \"this is a white car behind a tow truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [229, 251, 252, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344], "bbox": [0.734015625, 0.5984941176470588, 1.0, 1.0], "iscrowd": 0, "area": 15362.141150000003, "rle": {"size": [425, 640], "counts": "X_S66P=4N1N2O1N2O1O1O1J6@`0L4L4L4N2N2N2N2O1O100O1O1O100O1O100O1O100000000000000000000O100000000000000000000000000001O0000000000000000000000000000N2O1O1N2O1O100O1O100O1O1O100O1O100O10000O1000000O10000O1000000O1000000O100000000E_MnEa2P:=N2N2O100O1O1O10000000000O1000000000000000000N2N2N2N2N2N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2O1N2N2O1N2O1N2O1N2O1O1N2O1N2O1"}, "label": "this is a white car behind a tow truck"}]} {"id": 257867, "image": "COCO_train2014_000000257867.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white tow truck with yellow and red stripes\"."}], "task": "refering", "answer_template": "The \"a white tow truck with yellow and red stripes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 56, 57, 75, 76, 77, 78, 79, 80, 97, 98, 99, 100, 101, 102, 103, 104, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293], "bbox": [0.18528125, 0.13611764705882354, 0.7631249999999999, 0.8646823529411765], "iscrowd": 0, "area": 75171.24735, "rle": {"size": [425, 640], "counts": "caa1>jk:R1L3N3L4L4M3L4M3L5L4K5K4M4L4M3L3M5K6J5L5J5L5K5J7J7I7H8H3M4L3M4M2N3N1N3N1O1O1O1N2O3M5K9G;E3_OdHoJ]7o4hHmJY7MaHn48TK_7c4YH[K92^7a4]H[K63^7^4l000O01O1D<001O00LSKmGl4R860000000000000000000000000000000000000000000000000000000000000000000000000000>B5K5K5K5K3M4L1O2N1O2N0000001O0000000000000000O10000O100O10000O100O10000O100O100O1000000O1000000000000000000O100000000000000000000O10000000001O2kFoKn8W4O2N0ZGhK[8X4dGkKY8V4fGmKX8T4fGoKX8Q4gGRLW8a4O2N2N2N2N2N2N2N2N1O2N2N1O2N2N2M2N3M3N2M2N3M3M3N1N3M3N2N1O2N2N2N1O2N1O2N2N1O1OO1O100O1O1O1O100O10000000000000000001O006J2N2]JnHk4h700000000000000001O3M3M2N3M2N001N2L3O2M3M2L5N1O1O12N3M2N001O1O2N1O1O001O001O00000O100aJkHj4U7VKkHj4U7VKkHj4U7VKkHj4U7VKkHj4U7VKkHj4U7VKjHk4V7VKiHj4W7VKiHj4W7VKiHj4W7VKiHj4W7VKiHj4W7VKiHj4W7VKiHj4W7VKiHj4k7000000000000000000000000O1O100O1O1O1O1O1O2M2O1O1N2O10001O0000011N3M4L4L3M6I in this image.", "objects": [{"patches": [59, 80, 81, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 264, 265, 266, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 295], "bbox": [0.291453125, 0.11910224438902743, 0.95040625, 0.9101246882793017], "iscrowd": 0, "area": 66860.04305000001, "rle": {"size": [401, 640], "counts": "U]Y21_<2O1N2N2O1N2N2O1N2N010O1O100O010O1000O1000O1000O1000O1000O1000O1000O1000O10O10bD_Ol:a0SE@m:`0SE@m:`0REAm:?TEBk:>TECl:=TECl:=SEEl:;TEEk:;UEFk::UEFk::TEHk:8UEHk:8TEIl:6UEJk:6TELk:4UELk:4TEMl:2UENk:2TE0j:1VEOj:1UE0k:0UE02A[:?bE2OB_:=aE1NE`::aE2MGa:8aE1LJb:5aEc0^:]ObEc0^:^O`Ed0_:\\OaEe0]:\\ObEd0_:]O`Eb0a:f01N2O11O2N1O1O1O1O1N3N1O1O5K7IO10000000000N1O2N2N2N2N2O1N2N2N2N2N101N2N2N22N2N3M2N3M2N2N3L3N2N2N1O1O1O1[N\\M[Ie2d6[M\\If2b6[M^If2`6[M`I_3f5aLZJc3a5^L^Jf3^5[LbJi3Z5WLfJm3U5TLkJo3Q5RLoJR4m4nKSKV4h4kKXKY4d4hK[K[4a4fK_K\\4^4eKbK\\4]4dKcK]4[4dKdK_4Y4bKgK_4X4aKhK`4W4`KiKb4U4^KkKc4U4\\KkKe4T4[KlKf4S4ZKmKh4Q4XKoKi4P4WKPLj4o3VKQLl4n3SKRLn4m3RKRLP5m3QKRLQ5l3oJTLR5k3SKPLn4o3WKlKi4U4\\KeKe4[4`K_K`4c4dKWK]4k4eKPK\\4R5gKhJY4[5jK_JW4c5Z12O2M3N2M3N1N3N2M3N2N2M2O2M3N2M3M2NK5I8I6I7J6I7I7J6I7J6I8I6I7J6I7I7J6I75K6K5J6J6J6J6J2N001O01O000001O0000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O01O000010O0001O01O01O000O2O000O2O0O2O0O101N100>C`0_O?A01N10001O000O2O00001O0O101O0000O01000000O010000000O01000000O010000000O010000000O0100000O10O100000O10O100C`IiJ`6W5<1000000O1000OO2N2N2N2N2M3N2N2N1N3N2N2N2N2N2N2N2N4L3N2M4L3M3M4L3N2M4L3M3M4L2N2O1N2N2N1O2N2N101N2N2N2N2N2N2O1N2N2N2N2N2K5L4K5L4K8I8G8I8H8G8I8G8I8G8I8Gd[<"}, "label": "a yellow boat with rx60 on it"}]} {"id": 372558, "image": "COCO_train2014_000000372558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large yellow boat with a group of black flags hanging over the side\"."}], "task": "refering", "answer_template": "The \"a large yellow boat with a group of black flags hanging over the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 80, 81, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 264, 265, 266, 267, 268, 269, 270, 271, 272, 289, 290, 291, 292, 293, 294, 295], "bbox": [0.291453125, 0.11910224438902743, 0.95040625, 0.9101246882793017], "iscrowd": 0, "area": 66860.04305000001, "rle": {"size": [401, 640], "counts": "U]Y21_<2O1N2N2O1N2N2O1N2N010O1O100O010O1000O1000O1000O1000O1000O1000O1000O1000O10O10bD_Ol:a0SE@m:`0SE@m:`0REAm:?TEBk:>TECl:=TECl:=SEEl:;TEEk:;UEFk::UEFk::TEHk:8UEHk:8TEIl:6UEJk:6TELk:4UELk:4TEMl:2UENk:2TE0j:1VEOj:1UE0k:0UE02A[:?bE2OB_:=aE1NE`::aE2MGa:8aE1LJb:5aEc0^:]ObEc0^:^O`Ed0_:\\OaEe0]:\\ObEd0_:]O`Eb0a:f01N2O11O2N1O1O1O1O1N3N1O1O5K7IO10000000000N1O2N2N2N2N2O1N2N2N2N2N101N2N2N22N2N3M2N3M2N2N3L3N2N2N1O1O1O1[N\\M[Ie2d6[M\\If2b6[M^If2`6[M`I_3f5aLZJc3a5^L^Jf3^5[LbJi3Z5WLfJm3U5TLkJo3Q5RLoJR4m4nKSKV4h4kKXKY4d4hK[K[4a4fK_K\\4^4eKbK\\4]4dKcK]4[4dKdK_4Y4bKgK_4X4aKhK`4W4`KiKb4U4^KkKc4U4\\KkKe4T4[KlKf4S4ZKmKh4Q4XKoKi4P4WKPLj4o3VKQLl4n3SKRLn4m3RKRLP5m3QKRLQ5l3oJTLR5k3SKPLn4o3WKlKi4U4\\KeKe4[4`K_K`4c4dKWK]4k4eKPK\\4R5gKhJY4[5jK_JW4c5Z12O2M3N2M3N1N3N2M3N2N2M2O2M3N2M3M2NK5I8I6I7J6I7I7J6I7J6I8I6I7J6I7I7J6I75K6K5J6J6J6J6J2N001O01O000001O0000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O01O000010O0001O01O01O000O2O000O2O0O2O0O101N100>C`0_O?A01N10001O000O2O00001O0O101O0000O01000000O010000000O01000000O010000000O010000000O0100000O10O100000O10O100C`IiJ`6W5<1000000O1000OO2N2N2N2N2M3N2N2N1N3N2N2N2N2N2N2N2N4L3N2M4L3M3M4L3N2M4L3M3M4L2N2O1N2N2N1O2N2N101N2N2N2N2N2N2O1N2N2N2N2N2K5L4K5L4K8I8G8I8H8G8I8G8I8G8I8Gd[<"}, "label": "a large yellow boat with a group of black flags hanging over the side"}]} {"id": 372558, "image": "COCO_train2014_000000372558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small brown colored boat is on the shore is shown in the image\"."}], "task": "refering", "answer_template": "The \"small brown colored boat is on the shore is shown in the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 166, 167, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 279, 280, 281], "bbox": [0.0895, 0.4902743142144638, 0.332265625, 0.8770573566084787], "iscrowd": 0, "area": 14882.190049999997, "rle": {"size": [401, 640], "counts": "iaf01c10X92fF1V92hF1V90iFOX92fFOZ92dFO\\92bFN_92aFN`92^FOV9JmE7l0OV9NlE3n0OV90kE1m00X9OkE1k01[9OiE1j01]9NjE0h03^9MjE0g0>U9CSF0f0a0T9@UFOf0f0R9JlF:Q9FnF?n8AQGd0k8]OSGg0k8ZORGk0j8XORGl0l8UOQGo0l8TOPGn0P9TOlFm0T9Y100O100O10O10O100O100O11N101O0O2N101N2N101NL4O2O0O101N11O0O5L1O00001O0O101O00001O00000000001O000000L4L4L4L4L4L3M4M3L4L4L4L4L4L4002N2N3L3N2N2N2N3M2N2N2N2N3M2M3N2N2MO2M2N3M3O010O010O10O10O10O10O01000O010O01000O0105J6K5K6K4L4K6L3L4L5EkEmMY:o15O1N2^N]EU1d:hNaEU1`:hNdEV1]:gNhEV1k:N3L3J7G9F9HhmV5"}, "label": "small brown colored boat is on the shore is shown in the image"}]} {"id": 372558, "image": "COCO_train2014_000000372558.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small red boat with the marking rx55\"."}], "task": "refering", "answer_template": "The \"a small red boat with the marking rx55\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 166, 167, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 255, 256, 257, 258, 259, 260, 279, 280, 281], "bbox": [0.0895, 0.4902743142144638, 0.332265625, 0.8770573566084787], "iscrowd": 0, "area": 14882.190049999997, "rle": {"size": [401, 640], "counts": "iaf01c10X92fF1V92hF1V90iFOX92fFOZ92dFO\\92bFN_92aFN`92^FOV9JmE7l0OV9NlE3n0OV90kE1m00X9OkE1k01[9OiE1j01]9NjE0h03^9MjE0g0>U9CSF0f0a0T9@UFOf0f0R9JlF:Q9FnF?n8AQGd0k8]OSGg0k8ZORGk0j8XORGl0l8UOQGo0l8TOPGn0P9TOlFm0T9Y100O100O10O10O100O100O11N101O0O2N101N2N101NL4O2O0O101N11O0O5L1O00001O0O101O00001O00000000001O000000L4L4L4L4L4L3M4M3L4L4L4L4L4L4002N2N3L3N2N2N2N3M2N2N2N2N3M2M3N2N2MO2M2N3M3O010O010O10O10O10O10O01000O010O01000O0105J6K5K6K4L4K6L3L4L5EkEmMY:o15O1N2^N]EU1d:hNaEU1`:hNdEV1]:gNhEV1k:N3L3J7G9F9HhmV5"}, "label": "a small red boat with the marking rx55"}]} {"id": 241491, "image": "COCO_train2014_000000241491.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in an olive shirt blows out candles to a cake\"."}], "task": "refering", "answer_template": "The \"a man in an olive shirt blows out candles to a cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 103, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327], "bbox": [0.0, 0.3145990566037736, 0.5523281250000001, 1.0], "iscrowd": 0, "area": 61678.77035, "rle": {"size": [424, 640], "counts": "i6=i in this image.", "objects": [{"patches": [98, 99, 103, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 322, 323, 324, 325, 326, 327], "bbox": [0.0, 0.3145990566037736, 0.5523281250000001, 1.0], "iscrowd": 0, "area": 61678.77035, "rle": {"size": [424, 640], "counts": "i6=i in this image.", "objects": [{"patches": [22, 45, 136, 137, 159, 160, 183, 206, 229, 252, 275, 298, 321, 344], "bbox": [0.9366875, 0.0026650943396226414, 1.0, 0.998183962264151], "iscrowd": 0, "area": 5680.719749999998, "rle": {"size": [424, 640], "counts": "cVh71S=5J6K5J6K5K5N2O2M3M3N2M3N2M3M3N1N2N2jHZOh1f0RNEh1=TNJh17UN0g10UN8f1HWN`0d1AYNf0c1ZOQLRNR1l2h2ROQL\\No0g2o2mNlKfNn0_2V3lNfKoNl0X2^3iNaKYOi0Q2g3fNZKBi0j16gL\\2n1_LIi0e17kL_2h3SM@9PMb2b3PM@9VMd2^3lL_O in this image.", "objects": [{"patches": [22, 45, 136, 137, 159, 160, 183, 206, 229, 252, 275, 298, 321, 344], "bbox": [0.9366875, 0.0026650943396226414, 1.0, 0.998183962264151], "iscrowd": 0, "area": 5680.719749999998, "rle": {"size": [424, 640], "counts": "cVh71S=5J6K5J6K5K5N2O2M3M3N2M3N2M3M3N1N2N2jHZOh1f0RNEh1=TNJh17UN0g10UN8f1HWN`0d1AYNf0c1ZOQLRNR1l2h2ROQL\\No0g2o2mNlKfNn0_2V3lNfKoNl0X2^3iNaKYOi0Q2g3fNZKBi0j16gL\\2n1_LIi0e17kL_2h3SM@9PMb2b3PM@9VMd2^3lL_O in this image.", "objects": [{"patches": [47, 70, 71, 93, 94, 95, 116, 117, 118, 139, 140, 141, 142, 161, 162, 163, 164, 184, 185, 186, 187, 208, 209, 210, 231, 232, 233, 234, 255, 256, 278, 279], "bbox": [0.030656250000000003, 0.1938407494145199, 0.182, 0.8447072599531615], "iscrowd": 0, "area": 16356.580300000001, "rle": {"size": [427, 640], "counts": "Sb85S=4K5K5BEgC?U<>K5L4K6J4L4PM\\NPJm1o5`NTH\\Oe0]2V7[NQHYO`0d2`7WNlGVO=i2f7UNjGRO9P3l7SNgGnN;o2o7ZNPHg1o7[NoGf1P8[NoGf1Q8ZNnG]1H[M[8Y1kG\\1NXMW8\\1jG]11UMU8_1eG`18oLS8a5]O3M3M3N2M5K:F1O1O1O1O1O1iJ_Ib3b6\\LkIX3V6fLmIY3S6eLnI[3S6dLmI]3S6aLnI_3S6_LlIiN_Oi4e6\\LmIf3X6TLiIm3]6lKcIT4d6dK]I]4d6`K]I`4e6]K\\Id4d6[K\\Ie4f6XK\\Ih4d6VK]Ij4e6SK\\In4`7001O3M2N2N3VHhJ[7f5M2O1O2M2O2N1N2O000O1lLXJeNh5Y1[JfNe5Y1]JeNc5\\1_JbNa5^1`J`Ne5[1]JdNf5Y1[JfNh5W1YJgNk5U1WJjNj5U1WJjNk5T1VJjNl5`0ZIjMk0d1m5=]ImMh0d1m5>lJ@W5>jJaNVNNR7`1iJaNWNLR7a1iJaNWNLR7b1hJaNXNJS7V1PIWNf1h0XNIY7o0PKVOjMH_7g0kJ_OiMGe7>hJI\\6LiI1k9M3N2ImVj6"}, "label": "the snow boarder in the blue jacket"}]} {"id": 544595, "image": "COCO_train2014_000000544595.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man snowboarding in a blue jacket\"."}], "task": "refering", "answer_template": "The \"a man snowboarding in a blue jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 70, 71, 93, 94, 95, 116, 117, 118, 139, 140, 141, 142, 161, 162, 163, 164, 184, 185, 186, 187, 208, 209, 210, 231, 232, 233, 234, 255, 256, 278, 279], "bbox": [0.030656250000000003, 0.1938407494145199, 0.182, 0.8447072599531615], "iscrowd": 0, "area": 16356.580300000001, "rle": {"size": [427, 640], "counts": "Sb85S=4K5K5BEgC?U<>K5L4K6J4L4PM\\NPJm1o5`NTH\\Oe0]2V7[NQHYO`0d2`7WNlGVO=i2f7UNjGRO9P3l7SNgGnN;o2o7ZNPHg1o7[NoGf1P8[NoGf1Q8ZNnG]1H[M[8Y1kG\\1NXMW8\\1jG]11UMU8_1eG`18oLS8a5]O3M3M3N2M5K:F1O1O1O1O1O1iJ_Ib3b6\\LkIX3V6fLmIY3S6eLnI[3S6dLmI]3S6aLnI_3S6_LlIiN_Oi4e6\\LmIf3X6TLiIm3]6lKcIT4d6dK]I]4d6`K]I`4e6]K\\Id4d6[K\\Ie4f6XK\\Ih4d6VK]Ij4e6SK\\In4`7001O3M2N2N3VHhJ[7f5M2O1O2M2O2N1N2O000O1lLXJeNh5Y1[JfNe5Y1]JeNc5\\1_JbNa5^1`J`Ne5[1]JdNf5Y1[JfNh5W1YJgNk5U1WJjNj5U1WJjNk5T1VJjNl5`0ZIjMk0d1m5=]ImMh0d1m5>lJ@W5>jJaNVNNR7`1iJaNWNLR7a1iJaNWNLR7b1hJaNXNJS7V1PIWNf1h0XNIY7o0PKVOjMH_7g0kJ_OiMGe7>hJI\\6LiI1k9M3N2ImVj6"}, "label": "a man snowboarding in a blue jacket"}]} {"id": 315221, "image": "COCO_train2014_000000315221.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bird on left in photo\"."}], "task": "refering", "answer_template": "The \"bird on left in photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 117, 118, 119, 120, 140, 141, 142, 162, 163, 164, 165, 185, 186, 187, 188, 189, 208, 209, 212, 213, 231], "bbox": [0.043984375, 0.3189518413597734, 0.28171875, 0.7887818696883853], "iscrowd": 0, "area": 8450.2477, "rle": {"size": [353, 640], "counts": "da:2n:2O1N3N1N2N3N1N2RG@Y7b0cH@\\7b0aH@^7b0_H@a7b0[H@d7b0YH@g7a0UHBj7a0RHAn7`0oGBP8`0mGBR8`0kG_OX8c0dG\\O^8f0^G[Ob8h0[GWOf8m0VGROl8`10O1O2O0O100O2c0\\O2N2O1N2N2N2O[OnG]NP8d1YHTNe7m1cHlM\\7U2fHiMX7X2h0001O00001O001O000010O0001O001O001O001O1O0L5H7K6M2O2M3N1N3O010000O100O10000O1000000O101O1N2O1O1N2O1O2N0]MoGU2R8gMSHX2m7fMUHZ2k7eMWHZ2i7fMWHk1H^NP8FZHj1I_Nl7G[Hh1M`Ng7G]Hh1O`Nc7G`Hf10cN_7GaHd14eNY7FeHc15fNg7X1]HgNc7V1aHiN^7S1hHlNW7P1nHPOS7i0SIXOm6b0XI^Oi6<\\IDe66`IJ`60gIOZ6KkI5a81O000000000010O00000 in this image.", "objects": [{"patches": [95, 96, 117, 118, 119, 120, 140, 141, 142, 162, 163, 164, 165, 185, 186, 187, 188, 189, 208, 209, 212, 213, 231], "bbox": [0.043984375, 0.3189518413597734, 0.28171875, 0.7887818696883853], "iscrowd": 0, "area": 8450.2477, "rle": {"size": [353, 640], "counts": "da:2n:2O1N3N1N2N3N1N2RG@Y7b0cH@\\7b0aH@^7b0_H@a7b0[H@d7b0YH@g7a0UHBj7a0RHAn7`0oGBP8`0mGBR8`0kG_OX8c0dG\\O^8f0^G[Ob8h0[GWOf8m0VGROl8`10O1O2O0O100O2c0\\O2N2O1N2N2N2O[OnG]NP8d1YHTNe7m1cHlM\\7U2fHiMX7X2h0001O00001O001O000010O0001O001O001O001O1O0L5H7K6M2O2M3N1N3O010000O100O10000O1000000O101O1N2O1O1N2O1O2N0]MoGU2R8gMSHX2m7fMUHZ2k7eMWHZ2i7fMWHk1H^NP8FZHj1I_Nl7G[Hh1M`Ng7G]Hh1O`Nc7G`Hf10cN_7GaHd14eNY7FeHc15fNg7X1]HgNc7V1aHiN^7S1hHlNW7P1nHPOS7i0SIXOm6b0XI^Oi6<\\IDe66`IJ`60gIOZ6KkI5a81O000000000010O00000 in this image.", "objects": [{"patches": [111, 112, 134, 135, 136, 156, 157, 158, 159, 179, 180, 181, 201, 202, 203, 204, 205, 226], "bbox": [0.759875, 0.3103116147308782, 0.9338749999999999, 0.7394900849858358], "iscrowd": 0, "area": 6963.454149999995, "rle": {"size": [353, 640], "counts": "ahW53l:2N1O2N1O2N1O2N101O001O001O001O2N2N2N2N1O00O10L4I7O1N2O1O1N2O1O5J5L2N1N2O2O01O01O001O1UHPNd6R2ZIoMf6R2XIoMg6S2WInMh6T2WIlMh6U2eH^NZ7e2eN\\MXJe2h5Z100000O100EfIoKZ6o3jIoKV6n3oIPLQ6m3UJPLk5m3ZJQLf5m3i0L3M3M3N3L3M3100O2OO0_OdH]M]7a2gH\\MY7c2hH]MX7b2jH]MW73[HU2?hMV73\\HS2?jMU73\\HR2`0kMT73\\HQ2a0lMS73\\HP2c0kMR75\\Hm1d0nMQ74[Hm1e0oMP74[Hl1f0PNo64\\Hj1f0RNn64\\Hi1g0SNm64]Hg1g0UNl64]He1i0WNk63\\He1j0XNj63]Hc1i0[Nk61\\Hc1j0\\Nk60[Hc1k0]Nl6NZHb1l0cNh6J]Ha1k0hNQ7U1PImNP7P1SIQOl6l0VIUOj6g0YI[Og6a0[I@g6;[IFg61`IOh801N2N2N2N2N1O2N2O1NSc>"}, "label": "the bird in front of the other bird"}]} {"id": 315221, "image": "COCO_train2014_000000315221.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray bird on the right of another gray bird\"."}], "task": "refering", "answer_template": "The \"a gray bird on the right of another gray bird\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [111, 112, 134, 135, 136, 156, 157, 158, 159, 179, 180, 181, 201, 202, 203, 204, 205, 226], "bbox": [0.759875, 0.3103116147308782, 0.9338749999999999, 0.7394900849858358], "iscrowd": 0, "area": 6963.454149999995, "rle": {"size": [353, 640], "counts": "ahW53l:2N1O2N1O2N1O2N101O001O001O001O2N2N2N2N1O00O10L4I7O1N2O1O1N2O1O5J5L2N1N2O2O01O01O001O1UHPNd6R2ZIoMf6R2XIoMg6S2WInMh6T2WIlMh6U2eH^NZ7e2eN\\MXJe2h5Z100000O100EfIoKZ6o3jIoKV6n3oIPLQ6m3UJPLk5m3ZJQLf5m3i0L3M3M3N3L3M3100O2OO0_OdH]M]7a2gH\\MY7c2hH]MX7b2jH]MW73[HU2?hMV73\\HS2?jMU73\\HR2`0kMT73\\HQ2a0lMS73\\HP2c0kMR75\\Hm1d0nMQ74[Hm1e0oMP74[Hl1f0PNo64\\Hj1f0RNn64\\Hi1g0SNm64]Hg1g0UNl64]He1i0WNk63\\He1j0XNj63]Hc1i0[Nk61\\Hc1j0\\Nk60[Hc1k0]Nl6NZHb1l0cNh6J]Ha1k0hNQ7U1PImNP7P1SIQOl6l0VIUOj6g0YI[Og6a0[I@g6;[IFg61`IOh801N2N2N2N2N1O2N2O1NSc>"}, "label": "a gray bird on the right of another gray bird"}]} {"id": 413556, "image": "COCO_train2014_000000413556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black and white cow on the right from the man\"."}], "task": "refering", "answer_template": "The \"black and white cow on the right from the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 219, 220, 221, 222, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 288, 289, 290, 312, 313], "bbox": [0.47596875, 0.5618561484918794, 0.699203125, 0.8846403712296983], "iscrowd": 0, "area": 8141.8339999999935, "rle": {"size": [431, 640], "counts": "feP41^=0O10000O2O000O100O101O1N4M3M1N3cC@f;c0PDFo;m0O1O1O1O1O1O1O10O0O1O00O2N1O2M3N2N1O2N0O02K52]E\\NX9h1dF\\NY9f1cF^N[9]2iEgMT:k2O00000001E;I7\\Od0N200001O0O10001N100hDXNh:\\2L2N`0_O10001O00RO\\FQNd9h1fFUNZ9j1jFTNU9l1oFPNQ9P2UGjMk8W2V1001O002gFfMY7\\2^HkMb7X2XHkMg7\\2QHfMo7a2gGaMZ8e2YGbMg8c30000000O101N1N2N2POmFbMU9Z2WG[Ml8X2fF`Mh00e8\\2]1N2N3L3N2N3M2N3M2M3K6J5NhDgN_:V1aElN`:Q1`ERO`:k0aEVO`:g0bEZO^:c0dE]O]:a0dE@\\:=fEC[::gEFZ:NfDMR15Y:NoE2R:MnE3S:MkE3W:MhE3X:NgE2Z:OdE1]:ObE1^:0aE0`:1^EOd:0[E0g:0WE0k:OTE1n:OoD2R;OlD1V;OgD1\\;OaD2a;N\\D2g;NVD2l;OQD1R in this image.", "objects": [{"patches": [196, 197, 198, 219, 220, 221, 222, 241, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 288, 289, 290, 312, 313], "bbox": [0.47596875, 0.5618561484918794, 0.699203125, 0.8846403712296983], "iscrowd": 0, "area": 8141.8339999999935, "rle": {"size": [431, 640], "counts": "feP41^=0O10000O2O000O100O101O1N4M3M1N3cC@f;c0PDFo;m0O1O1O1O1O1O1O10O0O1O00O2N1O2M3N2N1O2N0O02K52]E\\NX9h1dF\\NY9f1cF^N[9]2iEgMT:k2O00000001E;I7\\Od0N200001O0O10001N100hDXNh:\\2L2N`0_O10001O00RO\\FQNd9h1fFUNZ9j1jFTNU9l1oFPNQ9P2UGjMk8W2V1001O002gFfMY7\\2^HkMb7X2XHkMg7\\2QHfMo7a2gGaMZ8e2YGbMg8c30000000O101N1N2N2POmFbMU9Z2WG[Ml8X2fF`Mh00e8\\2]1N2N3L3N2N3M2N3M2M3K6J5NhDgN_:V1aElN`:Q1`ERO`:k0aEVO`:g0bEZO^:c0dE]O]:a0dE@\\:=fEC[::gEFZ:NfDMR15Y:NoE2R:MnE3S:MkE3W:MhE3X:NgE2Z:OdE1]:ObE1^:0aE0`:1^EOd:0[E0g:0WE0k:OTE1n:OoD2R;OlD1V;OgD1\\;OaD2a;N\\D2g;NVD2l;OQD1R in this image.", "objects": [{"patches": [210, 211, 212, 213, 217, 218, 231, 232, 233, 234, 235, 236, 240, 241, 242, 243, 253, 254, 255, 256, 257, 258, 259, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 285, 286, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332], "bbox": [0.03278125, 0.6320892018779343, 0.57725, 0.9961267605633803], "iscrowd": 0, "area": 28634.751300000004, "rle": {"size": [426, 640], "counts": "XR96j<;E;M3N2M3N2N2M2TDlNa;V1^DWOT;k0jDVOT;a1O1O10g0YOg0XO100O10000O10000O10000O100O1000O010000O10000O10O010000F:0000O100O01000O100O10000O01000O100O10000O01000O100O10000O01000O100O10000O01000001O0000001O0O1000001O0000001O000000001O0000001O000000001O000000001O0000000000O1000001O000O100000000O10000000000O10000000000O100000000O10000000000O100000000O10000000000OULgGg2Y8oLhHZ2X7\\M[I[2o8K4L4L4L4L6J9G9G9G6J6J5KS^a0>^a^O>A101O1O1O1O1O1O14K5L4L1O1O1O1O1O1N2O1O1O2N0000000O10lNDZE9IS:O_E: in this image.", "objects": [{"patches": [210, 211, 212, 213, 217, 218, 231, 232, 233, 234, 235, 236, 240, 241, 242, 243, 253, 254, 255, 256, 257, 258, 259, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 282, 285, 286, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332], "bbox": [0.03278125, 0.6320892018779343, 0.57725, 0.9961267605633803], "iscrowd": 0, "area": 28634.751300000004, "rle": {"size": [426, 640], "counts": "XR96j<;E;M3N2M3N2N2M2TDlNa;V1^DWOT;k0jDVOT;a1O1O10g0YOg0XO100O10000O10000O10000O100O1000O010000O10000O10O010000F:0000O100O01000O100O10000O01000O100O10000O01000O100O10000O01000O100O10000O01000001O0000001O0O1000001O0000001O000000001O0000001O000000001O000000001O0000000000O1000001O000O100000000O10000000000O10000000000O100000000O10000000000O100000000O10000000000OULgGg2Y8oLhHZ2X7\\M[I[2o8K4L4L4L4L6J9G9G9G6J6J5KS^a0>^a^O>A101O1O1O1O1O1O14K5L4L1O1O1O1O1O1N2O1O1O2N0000000O10lNDZE9IS:O_E: in this image.", "objects": [{"patches": [249, 250, 272, 273, 296, 298, 318, 319, 320, 321, 341, 342, 343, 344], "bbox": [0.8226718749999999, 0.6916901408450705, 1.0, 0.9982394366197183], "iscrowd": 0, "area": 6241.928399999999, "rle": {"size": [426, 640], "counts": "o`k6j0_aIG[6:[I3a6MTI`0i6ASIc0m6^OPId0P7\\OmHg0S7YOjHj0V7VOhHl0Y7TOcHo0^7PO`HR1`7nN]HU1d7kNYHW1g7iNWHY1j7fNTH\\1l7dNTH\\1m7dNRH\\1o7cNQH]1o7cNQH]1P8bNPH^1P8cNPH\\1Q8cNoG]1Q8cNoG]1R8cNmG]1T8bNlG^1_8WNaGi1P:01O000000001O00001O001O00001O00001O0000000000O100000000O1000000N2O1N2N2O1N2O1O1O1N2O1O1O1O1O1O100O100O100O100O1O1O1O1O1N23MY1gNU1kNmB"}, "label": "the seat the main in the background is sitting on in the gray shirt"}]} {"id": 487284, "image": "COCO_train2014_000000487284.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"striped barely visible couch with man sitting on it\"."}], "task": "refering", "answer_template": "The \"striped barely visible couch with man sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [249, 250, 272, 273, 296, 298, 318, 319, 320, 321, 341, 342, 343, 344], "bbox": [0.8226718749999999, 0.6916901408450705, 1.0, 0.9982394366197183], "iscrowd": 0, "area": 6241.928399999999, "rle": {"size": [426, 640], "counts": "o`k6j0_aIG[6:[I3a6MTI`0i6ASIc0m6^OPId0P7\\OmHg0S7YOjHj0V7VOhHl0Y7TOcHo0^7PO`HR1`7nN]HU1d7kNYHW1g7iNWHY1j7fNTH\\1l7dNTH\\1m7dNRH\\1o7cNQH]1o7cNQH]1P8bNPH^1P8cNPH\\1Q8cNoG]1Q8cNoG]1R8cNmG]1T8bNlG^1_8WNaGi1P:01O000000001O00001O001O00001O00001O0000000000O100000000O1000000N2O1N2N2O1N2O1O1O1N2O1O1O1O1O1O100O100O100O100O1O1O1O1O1N23MY1gNU1kNmB"}, "label": "striped barely visible couch with man sitting on it"}]} {"id": 36729, "image": "COCO_train2014_000000036729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair without the curtain on it\"."}], "task": "refering", "answer_template": "The \"chair without the curtain on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 253, 254, 276, 277, 278, 279, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326, 345, 346, 347, 348, 349, 368], "bbox": [0.01775, 0.6233744855967078, 0.19921875, 0.9740329218106996], "iscrowd": 0, "area": 11375.622250000002, "rle": {"size": [486, 640], "counts": "ee53X>k0UOk0UOl0TOk0UOl0UOj0M3010O000O1N3N1O1N2O2M2O1O1O2O0O100O2O0O100O2O000O6K:E;F:E;F4K10O10O10O100O010O10O010000O010O100O01000000O2O001N10001O0O2O00001N10001O0O2O000O2O001O0O101O;E7H100O1O2O0O1O1O2N100O1O2N1g0YO8HO2O000O101O000O101O3L in this image.", "objects": [{"patches": [230, 231, 253, 254, 276, 277, 278, 279, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326, 345, 346, 347, 348, 349, 368], "bbox": [0.01775, 0.6233744855967078, 0.19921875, 0.9740329218106996], "iscrowd": 0, "area": 11375.622250000002, "rle": {"size": [486, 640], "counts": "ee53X>k0UOk0UOl0TOk0UOl0UOj0M3010O000O1N3N1O1N2O2M2O1O1O2O0O100O2O0O100O2O000O6K:E;F:E;F4K10O10O10O100O010O10O010000O010O100O01000000O2O001N10001O0O2O00001N10001O0O2O000O2O001O0O101O;E7H100O1O2O0O1O1O2N100O1O2N1g0YO8HO2O000O101O000O101O3L in this image.", "objects": [{"patches": [115, 116, 138, 139, 161, 162, 184, 185, 207, 208, 230, 231, 232, 253, 254, 255, 276, 277, 278, 299, 300, 301, 322, 323, 324, 345, 346], "bbox": [0.005218749999999999, 0.31386313465783666, 0.126796875, 0.9518543046357616], "iscrowd": 0, "area": 13234.739950000001, "rle": {"size": [453, 640], "counts": "n^1 in this image.", "objects": [{"patches": [115, 116, 138, 139, 161, 162, 184, 185, 207, 208, 230, 231, 232, 253, 254, 255, 276, 277, 278, 299, 300, 301, 322, 323, 324, 345, 346], "bbox": [0.005218749999999999, 0.31386313465783666, 0.126796875, 0.9518543046357616], "iscrowd": 0, "area": 13234.739950000001, "rle": {"size": [453, 640], "counts": "n^1 in this image.", "objects": [{"patches": [218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.14103125, 0.6140559440559441, 1.0, 0.9875990675990676], "iscrowd": 0, "area": 62747.36355000001, "rle": {"size": [429, 640], "counts": "iSV11[=1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1O1O1O1O1O1O1O1O1O100O100O100O10000001O00000000000000000000001O00000000000000000000001O000000000000000000O10000O100O10000O10000O100O10000O10000O100O100O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O1M3L4K5K5L4K5L4L400O100O100O100O100O10000O100O11O1O1O1O1O1O1O1O00O100O100O1O100O100O100O1O100O100O100O1O100O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O1M3M3M3N2M3M3M3N2O100O100O10000O100O100O10000000000000000O10000000000000000O10000000000000000O11O0000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O0000001O0000001OO10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O11O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001OXF"}, "label": "a plate of cheese and crackers"}]} {"id": 569214, "image": "COCO_train2014_000000569214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a platter of cheese and crackers\"."}], "task": "refering", "answer_template": "The \"a platter of cheese and crackers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.14103125, 0.6140559440559441, 1.0, 0.9875990675990676], "iscrowd": 0, "area": 62747.36355000001, "rle": {"size": [429, 640], "counts": "iSV11[=1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1O1O1O1O1O1O1O1O1O100O100O100O10000001O00000000000000000000001O00000000000000000000001O000000000000000000O10000O100O10000O10000O100O10000O10000O100O100O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O1M3L4K5K5L4K5L4L400O100O100O100O100O10000O100O11O1O1O1O1O1O1O1O00O100O100O1O100O100O100O1O100O100O100O1O100O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O100O1O100O1O1O1M3M3M3N2M3M3M3N2O100O100O10000O100O100O10000000000000000O10000000000000000O10000000000000000O11O0000001O00001O00001O0000001O00001O0000001O00001O00001O0000001O00001O0000001O0000001O0000001OO10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O11O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001OXF"}, "label": "a platter of cheese and crackers"}]} {"id": 569214, "image": "COCO_train2014_000000569214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the edges of a table\"."}], "task": "refering", "answer_template": "The \"the edges of a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 179, 180, 181, 182, 183, 184, 185, 201, 202, 203, 204, 205, 206, 207, 208, 227, 228, 229, 230, 231], "bbox": [0.003484375, 0.44836829836829833, 1.0, 0.7087878787878787], "iscrowd": 0, "area": 9613.74385, "rle": {"size": [429, 640], "counts": "dR14Y=;D=D\\Oc0]OjI"}, "label": "the edges of a table"}]} {"id": 520063, "image": "COCO_train2014_000000520063.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dark bike parked in the street\"."}], "task": "refering", "answer_template": "The \"dark bike parked in the street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [228, 232, 233, 234, 245, 246, 248, 249, 250, 251, 262, 263, 264, 265, 267, 268, 279, 280, 281, 282, 284, 285, 286, 296, 297, 298, 299, 300, 301, 302, 303, 304, 313, 314, 315, 316, 317, 318, 319, 320, 321, 330, 331, 332, 333, 334, 335, 336, 337, 338, 353, 354], "bbox": [0.4172291666666667, 0.581609375, 0.9301875000000001, 0.885125], "iscrowd": 0, "area": 23997.104699999996, "rle": {"size": [640, 480], "counts": "m[m37ic0e0[O7H7J5e]OPOf04R?f1h@cNT?^1j@gNS?Y1k@kNS?U1l@oNR?f0VA^Oi>:]AIc>5[ANf>0XA4g>KWA8j>EUA?j>@TAd0l>ZOg@gNIS2`?YOb@iNHQ2g?XO\\@S1e?d1001O00010O001O00001N1O2N10O0000010O000000001O0000000010O2N2N2O1N2N2N1O2O1NH8K5O0100O1O100O010O1O100O1O010O100N2O1N3N1N2O1N2nL`_Oc2a`0\\Mc_Oa2]`0^Mh_O^2X`0aMm_OZ2U`0eMP@V2P`0iMU@S2k?lMZ@P2f?PN_@k1o`0L4L2N1O2N101N1012MO2NUM`NWC_1h3RAEY2j0lM]Oi>5PAD\\2i0lM_Og>4PAE]2g0nM_Oe>5o@F^2g0nM]Od>7o@F^2g0oM\\O`>;RAB`2h0nMZO]>`0SA_Ob2h0P=0PC1n<0RC1l<0TC0k<0UC2iROhAQ1W>UO`Ao0_>UO[Ao0c>TOWAP1h>SOSAQ1j>SOPAQ1o>Q2O2N1O2N1O1O10O01O1O001OnKcAa2]>YM]BT2b=eMeB\\2]=[MiBe2[=QMjBQ3R?1O001O001O1O0O2O1O001O1O00000O100002M4M3M1O001O0O101O00001O0O2O00001O0O101O001O001N2O001O1O001N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N2N3M4L4L3M4L3F;D in this image.", "objects": [{"patches": [228, 232, 233, 234, 245, 246, 248, 249, 250, 251, 262, 263, 264, 265, 267, 268, 279, 280, 281, 282, 284, 285, 286, 296, 297, 298, 299, 300, 301, 302, 303, 304, 313, 314, 315, 316, 317, 318, 319, 320, 321, 330, 331, 332, 333, 334, 335, 336, 337, 338, 353, 354], "bbox": [0.4172291666666667, 0.581609375, 0.9301875000000001, 0.885125], "iscrowd": 0, "area": 23997.104699999996, "rle": {"size": [640, 480], "counts": "m[m37ic0e0[O7H7J5e]OPOf04R?f1h@cNT?^1j@gNS?Y1k@kNS?U1l@oNR?f0VA^Oi>:]AIc>5[ANf>0XA4g>KWA8j>EUA?j>@TAd0l>ZOg@gNIS2`?YOb@iNHQ2g?XO\\@S1e?d1001O00010O001O00001N1O2N10O0000010O000000001O0000000010O2N2N2O1N2N2N1O2O1NH8K5O0100O1O100O010O1O100O1O010O100N2O1N3N1N2O1N2nL`_Oc2a`0\\Mc_Oa2]`0^Mh_O^2X`0aMm_OZ2U`0eMP@V2P`0iMU@S2k?lMZ@P2f?PN_@k1o`0L4L2N1O2N101N1012MO2NUM`NWC_1h3RAEY2j0lM]Oi>5PAD\\2i0lM_Og>4PAE]2g0nM_Oe>5o@F^2g0nM]Od>7o@F^2g0oM\\O`>;RAB`2h0nMZO]>`0SA_Ob2h0P=0PC1n<0RC1l<0TC0k<0UC2iROhAQ1W>UO`Ao0_>UO[Ao0c>TOWAP1h>SOSAQ1j>SOPAQ1o>Q2O2N1O2N1O1O10O01O1O001OnKcAa2]>YM]BT2b=eMeB\\2]=[MiBe2[=QMjBQ3R?1O001O001O1O0O2O1O001O1O00000O100002M4M3M1O001O0O101O00001O0O2O00001O0O101O001O001N2O001O1O001N2N2N1O2N2N1O2N2N2N1O2N2N1O2N2N2N3M4L4L3M4L3F;D in this image.", "objects": [{"patches": [190, 191, 207, 208, 209, 210, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 246, 259, 260, 261, 262, 263, 277, 278, 279, 280, 295, 296, 297, 312, 313, 314, 330, 331], "bbox": [0.17040983606557375, 0.48014062500000004, 0.49403688524590167, 0.8605625], "iscrowd": 0, "area": 17636.598799999992, "rle": {"size": [640, 488], "counts": "nWd11lc04M3L4L5K4L4M3L4L4L4M3L4L4L4M4K4L4L4M3L4L4L4L4M3L4L4L5O0O1000000O10000O1000001N10000O1000000O10000O101O000O10000O1000000O101O0O1000000O10000O1000001N10000O1000000O10000O101O0O1000000O10000O10001O0O10000O1000000O10000O2O00000O10000O1000000O101O0O1000000O10000O1000001N10000O1000000O10000O2O000O100000XMj@e0U?]NTBY1l=iM\\Cm1g?E:F:E;C>CTki4"}, "label": "a white remote control in a box with a kitten"}]} {"id": 470912, "image": "COCO_train2014_000000470912.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wire basket with a kitten in it\"."}], "task": "refering", "answer_template": "The \"a wire basket with a kitten in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 207, 208, 209, 210, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 246, 259, 260, 261, 262, 263, 277, 278, 279, 280, 295, 296, 297, 312, 313, 314, 330, 331], "bbox": [0.17040983606557375, 0.48014062500000004, 0.49403688524590167, 0.8605625], "iscrowd": 0, "area": 17636.598799999992, "rle": {"size": [640, 488], "counts": "nWd11lc04M3L4L5K4L4M3L4L4L4M3L4L4L4M4K4L4L4M3L4L4L4L4M3L4L4L5O0O1000000O10000O1000001N10000O1000000O10000O101O000O10000O1000000O101O0O1000000O10000O1000001N10000O1000000O10000O101O0O1000000O10000O10001O0O10000O1000000O10000O2O00000O10000O1000000O101O0O1000000O10000O1000001N10000O1000000O10000O2O000O100000XMj@e0U?]NTBY1l=iM\\Cm1g?E:F:E;C>CTki4"}, "label": "a wire basket with a kitten in it"}]} {"id": 470912, "image": "COCO_train2014_000000470912.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black remote next to the kitten\"."}], "task": "refering", "answer_template": "The \"the black remote next to the kitten\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 139, 140, 141, 142, 156, 157, 158, 159, 160, 174, 175, 176, 177, 192, 193, 194, 211], "bbox": [0.18327868852459017, 0.34028125, 0.4654098360655738, 0.547765625], "iscrowd": 0, "area": 8640.648200000001, "rle": {"size": [640, 488], "counts": "kkg11fc0:I7L4L4L4L4M3N2O0O2O1N101O1N2O001O1O1O100O010O1O1O1O2O0O1O1O1O000001O01O00000010O01O100O1O10O01O100O1O100O101N100O2O0O010O010O010O010O010O010O010O10O010O01O10O0100O010O1O10O0100O00100O010O1O010O1O100O1O100O100O2O0000O100O10O0100O010O100O010O1000OkMg^Oc1Za0YNm^Od1Ta0WNR_Og1fa0N2M5L0O2O0O6J4M2M3A?IQjR5"}, "label": "the black remote next to the kitten"}]} {"id": 470912, "image": "COCO_train2014_000000470912.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"remote controller with a green button\"."}], "task": "refering", "answer_template": "The \"remote controller with a green button\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 139, 140, 141, 142, 156, 157, 158, 159, 160, 174, 175, 176, 177, 192, 193, 194, 211], "bbox": [0.18327868852459017, 0.34028125, 0.4654098360655738, 0.547765625], "iscrowd": 0, "area": 8640.648200000001, "rle": {"size": [640, 488], "counts": "kkg11fc0:I7L4L4L4L4M3N2O0O2O1N101O1N2O001O1O1O100O010O1O1O1O2O0O1O1O1O000001O01O00000010O01O100O1O10O01O100O1O100O101N100O2O0O010O010O010O010O010O010O010O10O010O01O10O0100O010O1O10O0100O00100O010O1O010O1O100O1O100O100O2O0000O100O10O0100O010O100O010O1000OkMg^Oc1Za0YNm^Od1Ta0WNR_Og1fa0N2M5L0O2O0O6J4M2M3A?IQjR5"}, "label": "remote controller with a green button"}]} {"id": 577416, "image": "COCO_train2014_000000577416.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hand of a woman who is about to operate the mixie\"."}], "task": "refering", "answer_template": "The \"a hand of a woman who is about to operate the mixie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 184, 185, 199, 200, 207, 208, 209, 221, 222, 223, 230, 231, 232, 233, 243, 244, 245, 246, 247, 248, 249, 253, 254, 255, 256, 257, 266, 267, 268, 269, 270, 271, 272, 276, 277, 278, 279, 280, 281, 287, 288, 289, 290, 291, 292, 293, 294, 295, 299, 300, 301, 302, 303, 304, 305, 309, 310, 311, 312, 313, 314, 315, 316, 317, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.0, 0.40675, 0.835953125, 1.0], "iscrowd": 0, "area": 76398.28319999999, "rle": {"size": [480, 640], "counts": "T6l8U6O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O0000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O100N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O0000O1O100O010O1O100O10O01O100O100O1O101N101N1O2O0O2O001`KYGf2g8XM\\Gh2d8VM_Gi2`8UMcGk2_8QMdGm2_8nLeGQ3]8kLeGU3]8fLgGY3[8bLhG]3[8_LhG`3Z8[LjGd3X8XLjGh3X8SLlGl3V8PLlGo3d9O1O001N101O1O0O2O001O1N101O001N0100000O010000O10O1000O10O1000O10O10O1000O10O100O00100O0001O00001O00001O0O1O2N100O2N1O1O2O0O1N3M2M3N3M2M3N2N3M2O2N3M4M2M4L3M4L3E in this image.", "objects": [{"patches": [161, 184, 185, 199, 200, 207, 208, 209, 221, 222, 223, 230, 231, 232, 233, 243, 244, 245, 246, 247, 248, 249, 253, 254, 255, 256, 257, 266, 267, 268, 269, 270, 271, 272, 276, 277, 278, 279, 280, 281, 287, 288, 289, 290, 291, 292, 293, 294, 295, 299, 300, 301, 302, 303, 304, 305, 309, 310, 311, 312, 313, 314, 315, 316, 317, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.0, 0.40675, 0.835953125, 1.0], "iscrowd": 0, "area": 76398.28319999999, "rle": {"size": [480, 640], "counts": "T6l8U6O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O0000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O100N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O0000O1O100O010O1O100O10O01O100O100O1O101N101N1O2O0O2O001`KYGf2g8XM\\Gh2d8VM_Gi2`8UMcGk2_8QMdGm2_8nLeGQ3]8kLeGU3]8fLgGY3[8bLhG]3[8_LhG`3Z8[LjGd3X8XLjGh3X8SLlGl3V8PLlGo3d9O1O001N101O1O0O2O001O1N101O001N0100000O010000O10O1000O10O1000O10O10O1000O10O100O00100O0001O00001O00001O0O1O2N100O2N1O1O2O0O1N3M2M3N3M2M3N2N3M2O2N3M4M2M4L3M4L3E in this image.", "objects": [{"patches": [48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 185, 186, 187, 188, 189, 209, 210, 211, 212, 234, 235], "bbox": [0.0, 0.16704166666666667, 0.302015625, 0.6010416666666667], "iscrowd": 0, "area": 25847.0792, "rle": {"size": [480, 640], "counts": "]5h0h in this image.", "objects": [{"patches": [162, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240], "bbox": [0.045078125000000004, 0.48985948477751756, 0.4568125, 0.7240983606557377], "iscrowd": 0, "area": 17609.378049999992, "rle": {"size": [427, 640], "counts": "\\Z<2W=3M3M3M2N3M3O1O1O3M3M2N2N1O1O1O2N100O1O1O1O1O001O1O001O1O010O1O1O1O2N1O1O1O1O1O3M2O1N1O001O1O001O1O001O001O10O01O001O0000000000000001O00000000000000000000001O0001O00000000001O00000010O000001O000000001O0000010O000000000000000O1O1O1O1O1O1O1OO1O1O100O1O2N100O1O11O1O1O1O2O0000000O100000001O0000000O1000001O000000000000000O2O000001O00001O00001O00001N100000000000000002N2N2N2N2N0010O01O001O010O000001O0000001O1O001N101N2O1N2O1N101N2O1N2O1O11O1O1O1O0O01O010O001O001O001O01O01O001O001O0000000O101O00000O100O1M3M4L3N2N2O1O1O1O2N1O5Klk`4"}, "label": "car number five that is in the snow"}]} {"id": 307082, "image": "COCO_train2014_000000307082.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a truck number 14 on a snow bank\"."}], "task": "refering", "answer_template": "The \"a truck number 14 on a snow bank\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 242, 243, 244, 245, 246, 247, 248, 249, 266, 268, 269], "bbox": [0.55525, 0.44763466042154565, 0.8671875, 0.7899297423887587], "iscrowd": 0, "area": 17007.338249999993, "rle": {"size": [427, 640], "counts": "[gd44U=2M3M3N2000a0oNVO_Do0Z;ZO_Dk0[;g0WOSNjEQ2U:QNgEP2Z:SN`Eo1`:>1NB?O100O1O100O1O10O01O100O10K5G8M4000O1000O0100O10000O10000O10000O10000O01000O100O10000O10000O10000O10O10O10000O10000O2O4K4M3M3L4M4L1N3N1O2M3N000O10000O1NjKoFV4P93O1001O01QGiKh8]4001O0O1O1O2TOTGVMm8j2UGSMo8i2TGTMo8j2TGRMP9k2RGRMQ9k2SGQMQ9l2QGPMS9n2h0L4M2M4M2N3L10000O010O10O01N2O001N2O1O0O2O10O1000O01000O01000O010000O01000O01000O010000O010000O01000PEoMe:Q2XESNh:m1UEWNj:U2O101O0O101O0O101O0O1000O001]N]E>c:B^E in this image.", "objects": [{"patches": [154, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 242, 243, 244, 245, 246, 247, 248, 249, 266, 268, 269], "bbox": [0.55525, 0.44763466042154565, 0.8671875, 0.7899297423887587], "iscrowd": 0, "area": 17007.338249999993, "rle": {"size": [427, 640], "counts": "[gd44U=2M3M3N2000a0oNVO_Do0Z;ZO_Dk0[;g0WOSNjEQ2U:QNgEP2Z:SN`Eo1`:>1NB?O100O1O100O1O10O01O100O10K5G8M4000O1000O0100O10000O10000O10000O10000O01000O100O10000O10000O10000O10O10O10000O10000O2O4K4M3M3L4M4L1N3N1O2M3N000O10000O1NjKoFV4P93O1001O01QGiKh8]4001O0O1O1O2TOTGVMm8j2UGSMo8i2TGTMo8j2TGRMP9k2RGRMQ9k2SGQMQ9l2QGPMS9n2h0L4M2M4M2N3L10000O010O10O01N2O001N2O1O0O2O10O1000O01000O01000O010000O01000O01000O010000O010000O01000PEoMe:Q2XESNh:m1UEWNj:U2O101O0O101O0O101O0O1000O001]N]E>c:B^E in this image.", "objects": [{"patches": [152, 153, 175, 176, 197, 198, 199, 200, 220, 221, 222, 223, 243, 244, 245, 246, 267, 268, 269, 290, 291, 292, 313, 314, 315, 336, 338], "bbox": [0.5998593750000001, 0.3861987041036717, 0.7415156250000001, 0.848768898488121], "iscrowd": 0, "area": 10536.31265, "rle": {"size": [463, 640], "counts": "ik]5V1T=8H8H8H8I7H6O0O1O2O0O2O0O101N100O2N104hFQMg6]3gHkL^6I_H^5^7gJ_H[5]7jJaHV5^7lJ`HU5`7mJ^HS5b7oJ[HS5d7l0O10001O00M2O2N2N2K5[NTHaLR8Y3dHRLa7i3f1I6J6J6K55L4K6J5LXLfE^3V:^LYF[3d9`L`Fd3U:B9GO101N100O3N2M3M3N2M3N2M2O2QMPFY1S:^NjFk0X9mNWGk0j8SOZGi0h8WOZGf0g8YO[Gd0f8\\OaG=`8BRHKP85XHBi7=WHBk7=UHBl7>THBm7>RHAP8>oGCR8;oGDR8;oGCT8;mGDU89mGEV89jGGZ85gGI]83eGL^81cGMa80U]Z2"}, "label": "a man standing next to the tire"}]} {"id": 77709, "image": "COCO_train2014_000000077709.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with cote beside the bog\"."}], "task": "refering", "answer_template": "The \"a man with cote beside the bog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 175, 176, 197, 198, 199, 200, 220, 221, 222, 223, 243, 244, 245, 246, 267, 268, 269, 290, 291, 292, 313, 314, 315, 336, 338], "bbox": [0.5998593750000001, 0.3861987041036717, 0.7415156250000001, 0.848768898488121], "iscrowd": 0, "area": 10536.31265, "rle": {"size": [463, 640], "counts": "ik]5V1T=8H8H8H8I7H6O0O1O2O0O2O0O101N100O2N104hFQMg6]3gHkL^6I_H^5^7gJ_H[5]7jJaHV5^7lJ`HU5`7mJ^HS5b7oJ[HS5d7l0O10001O00M2O2N2N2K5[NTHaLR8Y3dHRLa7i3f1I6J6J6K55L4K6J5LXLfE^3V:^LYF[3d9`L`Fd3U:B9GO101N100O3N2M3M3N2M3N2M2O2QMPFY1S:^NjFk0X9mNWGk0j8SOZGi0h8WOZGf0g8YO[Gd0f8\\OaG=`8BRHKP85XHBi7=WHBk7=UHBl7>THBm7>RHAP8>oGCR8;oGDR8;oGCT8;mGDU89mGEV89jGGZ85gGI]83eGL^81cGMa80U]Z2"}, "label": "a man with cote beside the bog"}]} {"id": 380820, "image": "COCO_train2014_000000380820.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bicycle the zebra is on\"."}], "task": "refering", "answer_template": "The \"the bicycle the zebra is on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 194, 215, 216, 217, 220, 221, 222, 223, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 284, 285, 286, 289, 290, 291, 292, 313, 314], "bbox": [0.338734375, 0.42258333333333337, 0.7431093750000001, 0.784375], "iscrowd": 0, "area": 17401.22025, "rle": {"size": [480, 640], "counts": "^PV38g>9F;F3L4M2M3N3L3N2M4M2N2O1N3M2O1N3M100O2N101N1O101N101O0O101O0O2O00001O0O2O0000O10000O10000O100O10000O0100O0O2O1N1O2O1N1O2ZOXMSEi2l:XMREi2o:WMoDk2P;VMoDd0E[1[;SNnDa0K[1S;XNQE=M[1o:[NRE;1X1l:^NTE92X1i:`NUE82W1i:bNUE64V1f:fNVE36U1d:hNWE44S1e:iNXEGF4`:HaFIQO;`:K`FKRO3a:IiD1d13TOMb:NhD1d13TOHc:4eD2d13c:KjC1e13`:KmC1d14^:KnC2d14\\:JQD2d13Z:JTD2c14X:JVD2b15V:IXD3c13U:HZD4b14S:H\\D4a14R:H]D4c14n9GaD2d17j9GcD0e19g9GdDOh1:b9FQG:n8FSG:\\MHU;NaG:WMJW>6gALY>4fANZ><000010O0001N1O101N1O1O2N1O101O01O10M2M3M[k10eTN4L4K5L4L4L4L3N3L4M2M101N2O001N2O0O2O0O2L4O010O10O0100O5L5J6K5J6K4K3N1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O0O101N101O010O001O01O01O0010O01O010O000001O0000001N1000001O0O10001O000O2O001O001O1N2O1O0O2O1N2N2N2O0O2N2O1N2N101N2N2O1N6J9G2O2M2N3M3L3M4L3M4L3I8H7I8H\\Q]2"}, "label": "the bicycle the zebra is on"}]} {"id": 380820, "image": "COCO_train2014_000000380820.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bike the zebra is riding\"."}], "task": "refering", "answer_template": "The \"the bike the zebra is riding\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [171, 172, 194, 215, 216, 217, 220, 221, 222, 223, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 284, 285, 286, 289, 290, 291, 292, 313, 314], "bbox": [0.338734375, 0.42258333333333337, 0.7431093750000001, 0.784375], "iscrowd": 0, "area": 17401.22025, "rle": {"size": [480, 640], "counts": "^PV38g>9F;F3L4M2M3N3L3N2M4M2N2O1N3M2O1N3M100O2N101N1O101N101O0O101O0O2O00001O0O2O0000O10000O10000O100O10000O0100O0O2O1N1O2O1N1O2ZOXMSEi2l:XMREi2o:WMoDk2P;VMoDd0E[1[;SNnDa0K[1S;XNQE=M[1o:[NRE;1X1l:^NTE92X1i:`NUE82W1i:bNUE64V1f:fNVE36U1d:hNWE44S1e:iNXEGF4`:HaFIQO;`:K`FKRO3a:IiD1d13TOMb:NhD1d13TOHc:4eD2d13c:KjC1e13`:KmC1d14^:KnC2d14\\:JQD2d13Z:JTD2c14X:JVD2b15V:IXD3c13U:HZD4b14S:H\\D4a14R:H]D4c14n9GaD2d17j9GcD0e19g9GdDOh1:b9FQG:n8FSG:\\MHU;NaG:WMJW>6gALY>4fANZ><000010O0001N1O101N1O1O2N1O101O01O10M2M3M[k10eTN4L4K5L4L4L4L3N3L4M2M101N2O001N2O0O2O0O2L4O010O10O0100O5L5J6K5J6K4K3N1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O0O101N101O010O001O01O01O0010O01O010O000001O0000001N1000001O0O10001O000O2O001O001O1N2O1O0O2O1N2N2N2O0O2N2O1N2N101N2N2O1N6J9G2O2M2N3M3L3M4L3M4L3I8H7I8H\\Q]2"}, "label": "the bike the zebra is riding"}]} {"id": 380820, "image": "COCO_train2014_000000380820.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a van\"."}], "task": "refering", "answer_template": "The \"a van\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 110, 125, 126, 128, 129, 130, 131, 132, 133, 148, 149, 152, 153], "bbox": [0.46089062500000005, 0.22754166666666667, 0.7952031250000001, 0.4046875], "iscrowd": 0, "area": 9002.597600000001, "rle": {"size": [480, 640], "counts": "d]Z43a00f=6VBLf=;UBFj=>RBCm=h001N2O001O001O0O101O00001O0O101O002N1O2E;N100OZNkBb1T=^NmBb1R=^NnBc1Q=]NPCb1P=^NPC`1S=`NmB^1T=bNlB]1U=cNmBZ1T=gNlBW1S=kNnBR1S=POmBj0W=WOkBe0V=WOeBM:e0Q=^OhBK?>h\\CBd<>\\CBd<=^CAcoB_OM5S=0000001O2N3M2N2N2N2N1O0000O100000000000000000O10000000000000O1000000010O01O100O3M2N2N2N2N3M2N2N7IfXm1"}, "label": "a van"}]} {"id": 380820, "image": "COCO_train2014_000000380820.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver van with no windows\"."}], "task": "refering", "answer_template": "The \"a silver van with no windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 110, 125, 126, 128, 129, 130, 131, 132, 133, 148, 149, 152, 153], "bbox": [0.46089062500000005, 0.22754166666666667, 0.7952031250000001, 0.4046875], "iscrowd": 0, "area": 9002.597600000001, "rle": {"size": [480, 640], "counts": "d]Z43a00f=6VBLf=;UBFj=>RBCm=h001N2O001O001O0O101O00001O0O101O002N1O2E;N100OZNkBb1T=^NmBb1R=^NnBc1Q=]NPCb1P=^NPC`1S=`NmB^1T=bNlB]1U=cNmBZ1T=gNlBW1S=kNnBR1S=POmBj0W=WOkBe0V=WOeBM:e0Q=^OhBK?>h\\CBd<>\\CBd<=^CAcoB_OM5S=0000001O2N3M2N2N2N2N1O0000O100000000000000000O10000000000000O1000000010O01O100O3M2N2N2N2N3M2N2N7IfXm1"}, "label": "a silver van with no windows"}]} {"id": 184224, "image": "COCO_train2014_000000184224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a unbrella held my a woman\"."}], "task": "refering", "answer_template": "The \"a unbrella held my a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 12, 13, 14, 15, 16, 17, 18, 19, 22, 23, 25, 26, 27, 28, 33], "bbox": [0.21654804270462635, 8.968609865470852e-05, 0.9446975088967973, 0.1976681614349776], "iscrowd": 0, "area": 10482.651500000002, "rle": {"size": [446, 281], "counts": "edj01k=2N2N2N2L4K5K6K4N2N2N2N2N2N2O2M2N2N2O1O1O1O100O2N1O100O1O1O1O101N1O1O100O1O10000O2O001O1N2O00QOaD^O`;`0gD[OX;d0mDYOS;f0oDXOR;g0QEWOo:h0UEUOj:j0ZESOg:m0[EQOe:n0U110O10O010O010O010O0100O0100O100O2O000O2O0O101N101O1O2N1O001O1O1O1]D`Nf:c1REbNo:_1kDeNU;R200000000000000O10000000000O1000000000000O2O000000001OO001O10000O100O10000O10000O101O1N101O0O2O001N101N2O000O101O0O10001N10000O101N10000O101O0000001O0O2O1O001O001O1N101O1O001N1M4N2N1O2M2O2N2N1O2N2N1O2N1O2N2N1O2N2N2M3O1O1O1O0Olm6"}, "label": "a unbrella held my a woman"}]} {"id": 184224, "image": "COCO_train2014_000000184224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plaid style umbrella being held by a gray haired woman\"."}], "task": "refering", "answer_template": "The \"plaid style umbrella being held by a gray haired woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 12, 13, 14, 15, 16, 17, 18, 19, 22, 23, 25, 26, 27, 28, 33], "bbox": [0.21654804270462635, 8.968609865470852e-05, 0.9446975088967973, 0.1976681614349776], "iscrowd": 0, "area": 10482.651500000002, "rle": {"size": [446, 281], "counts": "edj01k=2N2N2N2L4K5K6K4N2N2N2N2N2N2O2M2N2N2O1O1O1O100O2N1O100O1O1O1O101N1O1O100O1O10000O2O001O1N2O00QOaD^O`;`0gD[OX;d0mDYOS;f0oDXOR;g0QEWOo:h0UEUOj:j0ZESOg:m0[EQOe:n0U110O10O010O010O010O0100O0100O100O2O000O2O0O101N101O1O2N1O001O1O1O1]D`Nf:c1REbNo:_1kDeNU;R200000000000000O10000000000O1000000000000O2O000000001OO001O10000O100O10000O10000O101O1N101O0O2O001N101N2O000O101O0O10001N10000O101N10000O101O0000001O0O2O1O001O001O1N101O1O001N1M4N2N1O2M2O2N2N1O2N2N1O2N1O2N2N1O2N2N2M3O1O1O1O0Olm6"}, "label": "plaid style umbrella being held by a gray haired woman"}]} {"id": 184224, "image": "COCO_train2014_000000184224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an umbrella over a dog\"."}], "task": "refering", "answer_template": "The \"an umbrella over a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 110, 111, 112, 113, 114, 115, 122, 123, 124], "bbox": [0.06576512455516015, 0.6298878923766816, 0.5454092526690392, 0.7611659192825112], "iscrowd": 0, "area": 4235.703, "rle": {"size": [446, 281], "counts": "ib81l=1O2O0O2O0O2N101N2O0O2N101N1O2O0O2O0O2N101N101N101N101O0O2O0O2O0O2O001N101N100O10000O100O100O10000O100O10000O100O1000000O1000000O1000000K5N21O2N1O1O1O1O1O001O001O1O001O1O001N2O1O1O1O001O1O1O1O1O1O001O1N2O1O1O001O1O1O1O1O001O1O1O1O1N101O1O1O1O1O001O1O1O1O1O001N2O1O1Ob]g1"}, "label": "an umbrella over a dog"}]} {"id": 184224, "image": "COCO_train2014_000000184224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umbrella covering the dog\"."}], "task": "refering", "answer_template": "The \"the umbrella covering the dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 103, 110, 111, 112, 113, 114, 115, 122, 123, 124], "bbox": [0.06576512455516015, 0.6298878923766816, 0.5454092526690392, 0.7611659192825112], "iscrowd": 0, "area": 4235.703, "rle": {"size": [446, 281], "counts": "ib81l=1O2O0O2O0O2N101N2O0O2N101N1O2O0O2O0O2N101N101N101N101O0O2O0O2O0O2O001N101N100O10000O100O100O10000O100O10000O100O1000000O1000000O1000000K5N21O2N1O1O1O1O1O001O001O1O001O1O001N2O1O1O1O001O1O1O1O1O1O001O1N2O1O1O001O1O1O1O1O001O1O1O1O1N101O1O1O1O1O001O1O1O1O1O001N2O1O1Ob]g1"}, "label": "the umbrella covering the dog"}]} {"id": 184224, "image": "COCO_train2014_000000184224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog with long hair\"."}], "task": "refering", "answer_template": "The \"a dog with long hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 132, 133, 134, 142, 143, 144, 152, 153, 154], "bbox": [0.1893594306049822, 0.7753363228699551, 0.4808896797153025, 0.9865246636771301], "iscrowd": 0, "area": 5841.519399999997, "rle": {"size": [446, 281], "counts": "h_g03e=;I7A;K6E:L3M6L2M2O1O1O1O1O1O1O1O100O001M3NM4O10O2O0K6N101O1O101O0O2O1O001O1O002O3MO1O00000001O0O10000O100O1001O001M2M4O001N101O1N2O001O1L4I7M2O3M3M3M3M3L5K4L5^OXC^OY=8VWo1"}, "label": "a dog with long hair"}]} {"id": 184224, "image": "COCO_train2014_000000184224.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small fluffy dog with long hair that is lightly colored but with some areas of grey hair\"."}], "task": "refering", "answer_template": "The \"a small fluffy dog with long hair that is lightly colored but with some areas of grey hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 132, 133, 134, 142, 143, 144, 152, 153, 154], "bbox": [0.1893594306049822, 0.7753363228699551, 0.4808896797153025, 0.9865246636771301], "iscrowd": 0, "area": 5841.519399999997, "rle": {"size": [446, 281], "counts": "h_g03e=;I7A;K6E:L3M6L2M2O1O1O1O1O1O1O1O100O001M3NM4O10O2O0K6N101O1O101O0O2O1O001O1O002O3MO1O00000001O0O10000O100O1001O001M2M4O001N101O1N2O001O1L4I7M2O3M3M3M3M3L5K4L5^OXC^OY=8VWo1"}, "label": "a small fluffy dog with long hair that is lightly colored but with some areas of grey hair"}]} {"id": 44960, "image": "COCO_train2014_000000044960.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bus that is almost out of view\"."}], "task": "refering", "answer_template": "The \"the bus that is almost out of view\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 48, 49, 69, 70, 71, 72, 73, 74, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212], "bbox": [0.0, 0.12583143507972666, 0.271296875, 0.6105694760820046], "iscrowd": 0, "area": 31044.031549999996, "rle": {"size": [439, 640], "counts": "g6e1h8Z3[Ne1O11O0O10001O0000001O00001O000O101O0000001O00001O000O101O0000001O00001O000O101O0000001O00001O000O101O0000001O0000001O0O101O0000001O0000001O0O101O0000001O0000001O000O2O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O000O2O0000001O00000O2O000O2O00000O2O00000O2O00001N10000O2O00000O2O00000O2O001O0O101O001N10001N101Oj0VOm0ROn0SOm0SOm0SOZaW6"}, "label": "the bus that is almost out of view"}]} {"id": 331680, "image": "COCO_train2014_000000331680.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse closest to the boy\"."}], "task": "refering", "answer_template": "The \"the horse closest to the boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 270, 271, 272, 273, 290, 292, 294, 295], "bbox": [0.5935625, 0.5075350701402805, 0.907125, 0.6989178356713427], "iscrowd": 0, "area": 11462.268400000003, "rle": {"size": [499, 640], "counts": "[_i53^?5K5K5L3L5K4L4M3M4L2N1O2N2N2N1O0O101O000010O1O010O1N101O1O1N1O2M3N1O2N20O100O100O1O100O2O0O2N101N1O2O0O2N100O100O1O2O0O1O100aBXNR=k1iB]NR=V2K6J100O1O100O100O2N10O010O010O100O100O100O100O100O100O100O100O2N1G901O0000001O000010O000001O000010O000001O00010O0000000001O00000000001O001O10O01O1O1O001O2N2N2O1N00O1O100O2N2N101N3M2N2O2M2N2N3N1N2N2N2O1N2N2J6B>000O11O1O001O1O0O2O1O001O00000O1O1O1O1N2O1O1N2O1O1O2M2O1O1O1Ncml0"}, "label": "the horse closest to the boy"}]} {"id": 331680, "image": "COCO_train2014_000000331680.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse on the right\"."}], "task": "refering", "answer_template": "The \"the horse on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 250, 266, 267, 268, 269, 270, 271, 272, 273, 290, 292, 294, 295], "bbox": [0.5935625, 0.5075350701402805, 0.907125, 0.6989178356713427], "iscrowd": 0, "area": 11462.268400000003, "rle": {"size": [499, 640], "counts": "[_i53^?5K5K5L3L5K4L4M3M4L2N1O2N2N2N1O0O101O000010O1O010O1N101O1O1N1O2M3N1O2N20O100O100O1O100O2O0O2N101N1O2O0O2N100O100O1O2O0O1O100aBXNR=k1iB]NR=V2K6J100O1O100O100O2N10O010O010O100O100O100O100O100O100O100O100O2N1G901O0000001O000010O000001O000010O000001O00010O0000000001O00000000001O001O10O01O1O1O001O2N2N2O1N00O1O100O2N2N101N3M2N2O2M2N2N3N1N2N2N2O1N2N2J6B>000O11O1O001O1O0O2O1O001O00000O1O1O1O1N2O1O1N2O1O1O2M2O1O1O1Ncml0"}, "label": "the horse on the right"}]} {"id": 520100, "image": "COCO_train2014_000000520100.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"avocado chunks and several carrots\"."}], "task": "refering", "answer_template": "The \"avocado chunks and several carrots\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 182, 183, 184, 185, 186, 187, 188, 189, 190], "bbox": [0.11716, 0.4688533333333333, 0.66092, 0.8120266666666667], "iscrowd": 0, "area": 31447.681599999996, "rle": {"size": [375, 500], "counts": "mje0k0S:m0_Ob0I6J4L3L4N2N2O1O1N2O1O1O1N2O1O2M2O1O1O1N10000O10001O00000O10001O000000001O0O100000001O0000000O101O0000O100000O10000000000000000O10000000000000000000O1000000000000000000000O100000000000000000000000O10000000000000000000000000O100000000000000000O1000000000000000000000000000000000O100000000000O10000000000000000000000000000000000000O100000O1000000000000000000000000000000000000000000000O010000000000000000000000000000000000000000000000O100000000000000000000000000000000000001O0O1O2O0O2O0O1O2O0O101N1O2O0O102M2N2O1N2N2O2M2O1N2N2N3L8G:G8G:Gf0YOfPn1"}, "label": "avocado chunks and several carrots"}]} {"id": 520100, "image": "COCO_train2014_000000520100.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lunch box with carrot on middle and other vegetables on side\"."}], "task": "refering", "answer_template": "The \"a lunch box with carrot on middle and other vegetables on side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 182, 183, 184, 185, 186, 187, 188, 189, 190], "bbox": [0.11716, 0.4688533333333333, 0.66092, 0.8120266666666667], "iscrowd": 0, "area": 31447.681599999996, "rle": {"size": [375, 500], "counts": "mje0k0S:m0_Ob0I6J4L3L4N2N2O1O1N2O1O1O1N2O1O2M2O1O1O1N10000O10001O00000O10001O000000001O0O100000001O0000000O101O0000O100000O10000000000000000O10000000000000000000O1000000000000000000000O100000000000000000000000O10000000000000000000000000O100000000000000000O1000000000000000000000000000000000O100000000000O10000000000000000000000000000000000000O100000O1000000000000000000000000000000000000000000000O010000000000000000000000000000000000000000000000O100000000000000000000000000000000000001O0O1O2O0O2O0O1O2O0O101N1O2O0O102M2N2O1N2N2O2M2O1N2N2N3L8G:G8G:Gf0YOfPn1"}, "label": "a lunch box with carrot on middle and other vegetables on side"}]} {"id": 520100, "image": "COCO_train2014_000000520100.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a clear glass bowl with slices of cucumber\"."}], "task": "refering", "answer_template": "The \"a clear glass bowl with slices of cucumber\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 67, 68, 69, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 175], "bbox": [0.6365599999999999, 0.23872, 0.95824, 0.7059733333333333], "iscrowd": 0, "area": 20207.385550000006, "rle": {"size": [375, 500], "counts": "mbd32`;9H7H9H7H8J7L3M4K4M3M4L3L5L3M4K4M3M4L3L5L3M3M4K4M4L3M4K4M3M4L3L9H00001N10001N10001O0O101O001N10001O0O101O000O2O00001N101O00O010000O0100000O0100000O010000O0100000O0100000O010000O010000O0100000O10001O1N101O001N101O001N101O0O2O001O0O2O001O1N101O001N101O3L4M3M3M4K4M3M4L3L4M3M4L3L4M4L3L4M3M4K4L4L5K4K5L4L5K4K5L5K4L4K5L5K4L4KZ_7"}, "label": "a clear glass bowl with slices of cucumber"}]} {"id": 520100, "image": "COCO_train2014_000000520100.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smallest bowl\"."}], "task": "refering", "answer_template": "The \"the smallest bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 67, 68, 69, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 175], "bbox": [0.6365599999999999, 0.23872, 0.95824, 0.7059733333333333], "iscrowd": 0, "area": 20207.385550000006, "rle": {"size": [375, 500], "counts": "mbd32`;9H7H9H7H8J7L3M4K4M3M4L3L5L3M4K4M3M4L3L5L3M3M4K4M4L3M4K4M3M4L3L9H00001N10001N10001O0O101O001N10001O0O101O000O2O00001N101O00O010000O0100000O0100000O010000O0100000O0100000O010000O010000O0100000O10001O1N101O001N101O001N101O0O2O001O0O2O001O1N101O001N101O3L4M3M3M4K4M3M4L3L4M3M4L3L4M4L3L4M3M4K4L4L5K4K5L4L5K4K5L5K4L4K5L5K4L4KZ_7"}, "label": "the smallest bowl"}]} {"id": 520112, "image": "COCO_train2014_000000520112.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a balding man with white hair wearing a gray checked shirt and holding a wine glass\"."}], "task": "refering", "answer_template": "The \"a balding man with white hair wearing a gray checked shirt and holding a wine glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 80, 81, 82, 83, 103, 104, 105, 106, 127, 128, 129, 130, 131, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 244, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295], "bbox": [0.38471875, 0.17283333333333334, 0.8527812499999999, 0.9827222222222222], "iscrowd": 0, "area": 41682.2148, "rle": {"size": [360, 640], "counts": "aSg22S;2O1O1N3N1O1N2O2N1O1N2O2N1N2O1O2N1N2O1O1O1N2O2N1O2N2N2N2N1N3N2N2N1O2N2N2M3N2N1O2O1O1O1O1N102N1O1N2O1O1O1O1N3N1O1O1O1000001N101O000O2O000O2O2N2N1N3N2N2`MYM[Li2_3\\M_Lg2^3[MaLg2[3[MeLe2Z3]MdLe2Z3]MeLd2X3_MgLb2W3`MhL`2W3aMhLa2V3aMiL`2T3cMkL]2T3fMjL[2T3iMhLY2U3mMgLR2Y3SNcLLLZO_3P1aLE`0lNn2c1^LBi0fNg2n1\\L[OS1bNa2W2VLYO^1ZN]2g4gMUKY2j4S200O10000O100O10000O10000001N101O00001O000010O01kJSKn3l4iK^KV4b4aKgK`4Y4WKoKi4Q4nJXLR5o41O001O001O0010O01O001O0000100O101N101O0O101N13N13NL3M4L4M2O2M2N3N2MlIfK[5X4`JPLb5l3WJ\\LQ6X4 in this image.", "objects": [{"patches": [57, 58, 59, 80, 81, 82, 83, 103, 104, 105, 106, 127, 128, 129, 130, 131, 148, 149, 150, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 241, 242, 244, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295], "bbox": [0.38471875, 0.17283333333333334, 0.8527812499999999, 0.9827222222222222], "iscrowd": 0, "area": 41682.2148, "rle": {"size": [360, 640], "counts": "aSg22S;2O1O1N3N1O1N2O2N1O1N2O2N1N2O1O2N1N2O1O1O1N2O2N1O2N2N2N2N1N3N2N2N1O2N2N2M3N2N1O2O1O1O1O1N102N1O1N2O1O1O1O1N3N1O1O1O1000001N101O000O2O000O2O2N2N1N3N2N2`MYM[Li2_3\\M_Lg2^3[MaLg2[3[MeLe2Z3]MdLe2Z3]MeLd2X3_MgLb2W3`MhL`2W3aMhLa2V3aMiL`2T3cMkL]2T3fMjL[2T3iMhLY2U3mMgLR2Y3SNcLLLZO_3P1aLE`0lNn2c1^LBi0fNg2n1\\L[OS1bNa2W2VLYO^1ZN]2g4gMUKY2j4S200O10000O100O10000O10000001N101O00001O000010O01kJSKn3l4iK^KV4b4aKgK`4Y4WKoKi4Q4nJXLR5o41O001O001O0010O01O001O0000100O101N101O0O101N13N13NL3M4L4M2O2M2N3N2MlIfK[5X4`JPLb5l3WJ\\LQ6X4 in this image.", "objects": [{"patches": [44, 45, 66, 67, 68, 89, 90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 182, 183, 205, 206, 228, 229, 251, 252, 272, 274, 275, 293, 294, 295, 296, 297, 298], "bbox": [0.7710625, 0.13258333333333333, 1.0, 0.9887777777777779], "iscrowd": 0, "area": 17094.628449999997, "rle": {"size": [360, 640], "counts": "Qe]5;m:00000000000O1000000000000O1000000000000O1000000000000O100000000000000O1000000000000O0100000000000O100M3L4O1000000O1000000O100UIVOX3j0bLBX3>cLMX32cL9W3GdLa0Z3^OaLi0^3VO^LQ1`3nN\\LX1c3gNYL`1e3_NYLf1d3ZN[Li1c3WN]Ll1`3TN_Lo1_3QN`LQ2_3oMaLS2]3mMbLV2\\3jMdLX2Y3iMfLY2Y3gMgL[2W3eMhL^2V3bMjL`2T3`MkLb2T3^MlLd2R3\\MmLg2Q3YMnLj2P3VMPMk2o2UMPMn2m2SMSMn2l2RMSMQ3k2oLUMR3j2nLUMU3i2kLWMV3h2jLWMY3g2gLYMZ3f2fLYM]3e2dLZM]3d2dL[M_3c2aL\\Ma3c2_L]Mc3a2]L_Md3`2\\L_Mg3_2YLaMh3^2XLbMi3]2WLcMj3\\2VLdMl3Z2TLfMm3EdKl0`0_Om3CQLa02Ln3A^L6D9P4]OkLMUOe0R4\\OWMBgNR1S4ZOfNf0[1WOgNi0[1TOfNl0Z1SOgNm0Z1QOgNo0Y1oNiNQ1X1mNiNR1X1mNiNS1X1jNjNV1V1iNkNW1V1gNkNY1U1eNmN[1T1cNmN]1S1bNnN^1S1_NoNa1Q1^NPOa1R1]NoNc1[500000000000000O1000000000_NbNXI^1a0@Y1Q1UN_OZOR66nH"}, "label": "a man with a bald head near a man in a striped shirt"}]} {"id": 520112, "image": "COCO_train2014_000000520112.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the younger bald man with the mole on his right cheek\"."}], "task": "refering", "answer_template": "The \"the younger bald man with the mole on his right cheek\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 66, 67, 68, 89, 90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 182, 183, 205, 206, 228, 229, 251, 252, 272, 274, 275, 293, 294, 295, 296, 297, 298], "bbox": [0.7710625, 0.13258333333333333, 1.0, 0.9887777777777779], "iscrowd": 0, "area": 17094.628449999997, "rle": {"size": [360, 640], "counts": "Qe]5;m:00000000000O1000000000000O1000000000000O1000000000000O100000000000000O1000000000000O0100000000000O100M3L4O1000000O1000000O100UIVOX3j0bLBX3>cLMX32cL9W3GdLa0Z3^OaLi0^3VO^LQ1`3nN\\LX1c3gNYL`1e3_NYLf1d3ZN[Li1c3WN]Ll1`3TN_Lo1_3QN`LQ2_3oMaLS2]3mMbLV2\\3jMdLX2Y3iMfLY2Y3gMgL[2W3eMhL^2V3bMjL`2T3`MkLb2T3^MlLd2R3\\MmLg2Q3YMnLj2P3VMPMk2o2UMPMn2m2SMSMn2l2RMSMQ3k2oLUMR3j2nLUMU3i2kLWMV3h2jLWMY3g2gLYMZ3f2fLYM]3e2dLZM]3d2dL[M_3c2aL\\Ma3c2_L]Mc3a2]L_Md3`2\\L_Mg3_2YLaMh3^2XLbMi3]2WLcMj3\\2VLdMl3Z2TLfMm3EdKl0`0_Om3CQLa02Ln3A^L6D9P4]OkLMUOe0R4\\OWMBgNR1S4ZOfNf0[1WOgNi0[1TOfNl0Z1SOgNm0Z1QOgNo0Y1oNiNQ1X1mNiNR1X1mNiNS1X1jNjNV1V1iNkNW1V1gNkNY1U1eNmN[1T1cNmN]1S1bNnN^1S1_NoNa1Q1^NPOa1R1]NoNc1[500000000000000O1000000000_NbNXI^1a0@Y1Q1UN_OZOR66nH"}, "label": "the younger bald man with the mole on his right cheek"}]} {"id": 511930, "image": "COCO_train2014_000000511930.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black smock\"."}], "task": "refering", "answer_template": "The \"a man wearing a black smock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 81, 82, 83, 84, 104, 105, 106, 107, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 176, 177, 178, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 240, 241, 242, 243, 263, 265, 266, 286], "bbox": [0.450609375, 0.1836065573770492, 0.77021875, 0.8179625292740046], "iscrowd": 0, "area": 23332.407449999988, "rle": {"size": [427, 640], "counts": "Skh34S=4L4M3N30OmNBZE=b:J]E4_:0bEOa6KnK;cMH]6=^K0UNB[6T1gJ_OPOZOW6a1[J\\O@oNU6T2hISO_7^3N2O2O0O100O101N100O100O20O0001N101O00001O0O2O2N1O1O1O1N3N001O1O1O1N101N101N1O2N1O2N1O2N1O2N1O2O0O001O1H8G9H8K5O1O1O1O100O1O1O1O1O100O001O1O1O1O10000O10000O10000O10lNfJWJZ5i1`JV1VHAj7>YH@h7?YH_Oh7`0ZH_Of7a0[H^Oe7b0]H\\Od7b0^H\\Oc7d0^H[Ob7d0`H[O`7c0cH\\O^7a0fH\\O[7b0hH]OX7`0lH_OT7?oH@R7>QI_OQ7?QI@Q7=QIBP7Y6^OjIa0X6[OjIe0]6QOfIo0[91O00001O0000001O00001O0000001O000O0100O10000O10000O100O1O001O1O1O1O001O1O1O1O001O100000000000000000001O0000000O10001O1O1O1O1O1O1N3M4M3LeRm1"}, "label": "a man wearing a black smock"}]} {"id": 511930, "image": "COCO_train2014_000000511930.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in black apron slicing meat off an animal ' s leg\"."}], "task": "refering", "answer_template": "The \"man in black apron slicing meat off an animal ' s leg\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 81, 82, 83, 84, 104, 105, 106, 107, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 176, 177, 178, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 240, 241, 242, 243, 263, 265, 266, 286], "bbox": [0.450609375, 0.1836065573770492, 0.77021875, 0.8179625292740046], "iscrowd": 0, "area": 23332.407449999988, "rle": {"size": [427, 640], "counts": "Skh34S=4L4M3N30OmNBZE=b:J]E4_:0bEOa6KnK;cMH]6=^K0UNB[6T1gJ_OPOZOW6a1[J\\O@oNU6T2hISO_7^3N2O2O0O100O101N100O100O20O0001N101O00001O0O2O2N1O1O1O1N3N001O1O1O1N101N101N1O2N1O2N1O2N1O2N1O2O0O001O1H8G9H8K5O1O1O1O100O1O1O1O1O100O001O1O1O1O10000O10000O10000O10lNfJWJZ5i1`JV1VHAj7>YH@h7?YH_Oh7`0ZH_Of7a0[H^Oe7b0]H\\Od7b0^H\\Oc7d0^H[Ob7d0`H[O`7c0cH\\O^7a0fH\\O[7b0hH]OX7`0lH_OT7?oH@R7>QI_OQ7?QI@Q7=QIBP7Y6^OjIa0X6[OjIe0]6QOfIo0[91O00001O0000001O00001O0000001O000O0100O10000O10000O100O1O001O1O1O1O001O1O1O1O001O100000000000000000001O0000000O10001O1O1O1O1O1O1N3M4M3LeRm1"}, "label": "man in black apron slicing meat off an animal ' s leg"}]} {"id": 511930, "image": "COCO_train2014_000000511930.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man serving soup\"."}], "task": "refering", "answer_template": "The \"a man serving soup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 53, 54, 55, 56, 75, 76, 77, 78, 79, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 278, 279, 280, 281, 282, 283, 302, 303, 304], "bbox": [0.11395312500000002, 0.10112412177985948, 0.52325, 0.9123653395784543], "iscrowd": 0, "area": 49652.386750000005, "rle": {"size": [427, 640], "counts": "dhn06i<=C=C=C=D=B=C=C=D\\701O0O2O0O101O0O2O0O2O000O2O001N100O2O001N100O2O001N1000[Xo3"}, "label": "a man serving soup"}]} {"id": 511930, "image": "COCO_train2014_000000511930.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a white shirt and navy apron preparing food\"."}], "task": "refering", "answer_template": "The \"a man with a white shirt and navy apron preparing food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 53, 54, 55, 56, 75, 76, 77, 78, 79, 96, 97, 98, 99, 100, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 278, 279, 280, 281, 282, 283, 302, 303, 304], "bbox": [0.11395312500000002, 0.10112412177985948, 0.52325, 0.9123653395784543], "iscrowd": 0, "area": 49652.386750000005, "rle": {"size": [427, 640], "counts": "dhn06i<=C=C=C=D=B=C=C=D\\701O0O2O0O101O0O2O0O2O000O2O001N100O2O001N100O2O001N1000[Xo3"}, "label": "a man with a white shirt and navy apron preparing food"}]} {"id": 544701, "image": "COCO_train2014_000000544701.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with spiky hair looking at a box of desserts\"."}], "task": "refering", "answer_template": "The \"a man with spiky hair looking at a box of desserts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 77, 78, 79, 80, 81, 82, 83, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 234, 235, 236, 237, 238, 239, 240, 241, 242, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 324, 325, 326, 327, 328, 329, 330, 331, 332, 347, 348, 349, 350, 351, 352, 353, 354, 355, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.07415625, 0.16629166666666664, 0.6471875, 1.0], "iscrowd": 0, "area": 78288.82629999999, "rle": {"size": [480, 640], "counts": "bnf03i>4M3M3M4K4M3M3M3L4M3M3O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1O1N2O1O1N2O2N1N2O1O1N2M3M3M3M3M3M3N2M3M3M4L3M3I7]Oc0@`0O1N2O1N2O1N2O1O1N3N1N2O1O1N2O1N2O1O1N2O1N2O1O2N1O100O1O1O1O1O1O1O1O1O1O1O1O2O0O1O1O1O1O1O1O1O1O1O1O1O101N1O1O1O1O1O1O1O1O1O1O1O100O2N10000O10000O100O10000O10001N100O10000O100O10000O10000O101N10000O100O10000O01000O100O10O10O10000O010O10000O10O010000O1K4L5K5K5K5K5K4L5K5K5K5K4L5kNWFnLn9h2[FTMj9d2^FWMg9b2_FZMf9d2\\FWMh9i2XFSMm9m2SFPMP:o2PFQMQ:o2oEPMR:P3mEPMS:P3nEoLS:Q3lEoLU:P3lEoLU:Q3jEoLW:Q3hEoLY:P3hEoLX:R3gEnLZ:Q3gEnLZ:R3j0O10O10O100O10O10O1M3N2000O100000000O10000000000O101O00000O100TNZN^Ff1`9fNVFZ1h9P2M3N2N2N2M3N2N2M3N2N2N2M3N2N3L3N2N2M3N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2O10000000000000000O100O1O100O1O2M2M3M3N2M3M3O100O1000000O2O1M3M3N2M3M3M3M3N2N2N2O1N2N2N2O1N2N2O1N2N1O2O0POQ1L3M3M4L3M4L3M4L3M4ZOe0O1O2N1O2N1O2O0O4L4L5J5K5J6K6I;FbiY3"}, "label": "a man with spiky hair looking at a box of desserts"}]} {"id": 544701, "image": "COCO_train2014_000000544701.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with blue shirt\"."}], "task": "refering", "answer_template": "The \"man with blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 77, 78, 79, 80, 81, 82, 83, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 234, 235, 236, 237, 238, 239, 240, 241, 242, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 324, 325, 326, 327, 328, 329, 330, 331, 332, 347, 348, 349, 350, 351, 352, 353, 354, 355, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.07415625, 0.16629166666666664, 0.6471875, 1.0], "iscrowd": 0, "area": 78288.82629999999, "rle": {"size": [480, 640], "counts": "bnf03i>4M3M3M4K4M3M3M3L4M3M3O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1O1N2O1O1N2O2N1N2O1O1N2M3M3M3M3M3M3N2M3M3M4L3M3I7]Oc0@`0O1N2O1N2O1N2O1O1N3N1N2O1O1N2O1N2O1O1N2O1N2O1O2N1O100O1O1O1O1O1O1O1O1O1O1O1O2O0O1O1O1O1O1O1O1O1O1O1O1O101N1O1O1O1O1O1O1O1O1O1O1O100O2N10000O10000O100O10000O10001N100O10000O100O10000O10000O101N10000O100O10000O01000O100O10O10O10000O010O10000O10O010000O1K4L5K5K5K5K5K4L5K5K5K5K4L5kNWFnLn9h2[FTMj9d2^FWMg9b2_FZMf9d2\\FWMh9i2XFSMm9m2SFPMP:o2PFQMQ:o2oEPMR:P3mEPMS:P3nEoLS:Q3lEoLU:P3lEoLU:Q3jEoLW:Q3hEoLY:P3hEoLX:R3gEnLZ:Q3gEnLZ:R3j0O10O10O100O10O10O1M3N2000O100000000O10000000000O101O00000O100TNZN^Ff1`9fNVFZ1h9P2M3N2N2N2M3N2N2M3N2N2N2M3N2N3L3N2N2M3N2N2O1N2N2N2O1N2N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2O10000000000000000O100O1O100O1O2M2M3M3N2M3M3O100O1000000O2O1M3M3N2M3M3M3M3N2N2N2O1N2N2N2O1N2N2O1N2N1O2O0POQ1L3M3M4L3M4L3M4L3M4ZOe0O1O2N1O2N1O2O0O4L4L5J5K5J6K6I;FbiY3"}, "label": "man with blue shirt"}]} {"id": 544701, "image": "COCO_train2014_000000544701.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two hands , one with a silver bracelet , grabbing white plastic - ware forks\"."}], "task": "refering", "answer_template": "The \"two hands , one with a silver bracelet , grabbing white plastic - ware forks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 342, 343, 344, 365, 366, 367, 388, 389, 390], "bbox": [0.83784375, 0.6813125, 1.0, 0.989875], "iscrowd": 0, "area": 8178.308200000002, "rle": {"size": [480, 640], "counts": "\\ck79c>5K5L5J5N2N2N3M2N2O2M2N2N2N3M2N2O2M2N2M2O1O1O1O1O1000N1O2N2J6O010O1O1OnCnNY:P1gEUOU:k0iE[OT:c0lE_OT:`0kECS:=lEET:9kEKS:5lEMT:2lE0R:OoE2Q:MoE5o9JRF6o9IQF7o9IQF8n9HRF8o9HPF8P:IoE7Q:JnE7Q:JnE6S:ImE7S:JlE6T:KkE5U:KkE6U:IkE7U:IkE7U:IkE7U:IkE7V:HjE9U:GkE9U:GjE:V:FjE:V:FjE:V:FjE:V:FjE;U:EkE;U:EkE;U:EkE;U:EkE;U:EjET:BlE>T:BlE>T:BmE=S:CmE=S:CmE=S:BnE?Q:@PF`0P:@PFa0o9^ORFb0n9]OTFc0j9]OWFc0i9]OWFc0i9\\OXFe0f9[O[Fe0e9[O[Ff0d9YO^Ff0a9ZO`Fg0_9XObFh0^9WOdFh0\\9XOdFi0Z9WOhFh0X9WOiFj0V9UOlFj0S9VOnFk0P9VOPGj0P9UORG>Y9BiFKh95YFYOO"}, "label": "two hands , one with a silver bracelet , grabbing white plastic - ware forks"}]} {"id": 544701, "image": "COCO_train2014_000000544701.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hands holding up two forks , on hand wearing a ring and a bracelet\"."}], "task": "refering", "answer_template": "The \"hands holding up two forks , on hand wearing a ring and a bracelet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 342, 343, 344, 365, 366, 367, 388, 389, 390], "bbox": [0.83784375, 0.6813125, 1.0, 0.989875], "iscrowd": 0, "area": 8178.308200000002, "rle": {"size": [480, 640], "counts": "\\ck79c>5K5L5J5N2N2N3M2N2O2M2N2N2N3M2N2O2M2N2M2O1O1O1O1O1000N1O2N2J6O010O1O1OnCnNY:P1gEUOU:k0iE[OT:c0lE_OT:`0kECS:=lEET:9kEKS:5lEMT:2lE0R:OoE2Q:MoE5o9JRF6o9IQF7o9IQF8n9HRF8o9HPF8P:IoE7Q:JnE7Q:JnE6S:ImE7S:JlE6T:KkE5U:KkE6U:IkE7U:IkE7U:IkE7U:IkE7V:HjE9U:GkE9U:GjE:V:FjE:V:FjE:V:FjE:V:FjE;U:EkE;U:EkE;U:EkE;U:EkE;U:EjET:BlE>T:BlE>T:BmE=S:CmE=S:CmE=S:BnE?Q:@PF`0P:@PFa0o9^ORFb0n9]OTFc0j9]OWFc0i9]OWFc0i9\\OXFe0f9[O[Fe0e9[O[Ff0d9YO^Ff0a9ZO`Fg0_9XObFh0^9WOdFh0\\9XOdFi0Z9WOhFh0X9WOiFj0V9UOlFj0S9VOnFk0P9VOPGj0P9UORG>Y9BiFKh95YFYOO"}, "label": "hands holding up two forks , on hand wearing a ring and a bracelet"}]} {"id": 307136, "image": "COCO_train2014_000000307136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red van approaching woman\"."}], "task": "refering", "answer_template": "The \"red van approaching woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 160, 181, 182, 183, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343, 344], "bbox": [0.7796406250000001, 0.43594847775175644, 0.9985312500000001, 0.9887587822014052], "iscrowd": 0, "area": 22796.746149999988, "rle": {"size": [427, 640], "counts": "\\[`63R=6J6J7M2000000O10001O0000000O101O00000000000O2O00000000001O0O10000[DA^:?]E3W:LcEg0R:XOhEZ1m9^1E:E in this image.", "objects": [{"patches": [158, 159, 160, 181, 182, 183, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343, 344], "bbox": [0.7796406250000001, 0.43594847775175644, 0.9985312500000001, 0.9887587822014052], "iscrowd": 0, "area": 22796.746149999988, "rle": {"size": [427, 640], "counts": "\\[`63R=6J6J7M2000000O10001O0000000O101O00000000000O2O00000000001O0O10000[DA^:?]E3W:LcEg0R:XOhEZ1m9^1E:E in this image.", "objects": [{"patches": [141, 142, 143, 163, 164, 165, 166, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 233, 234, 235], "bbox": [0.09796875000000001, 0.3771875, 0.31478124999999996, 0.6234375], "iscrowd": 0, "area": 9643.33, "rle": {"size": [480, 640], "counts": "ngm04g>6L3O2O0O2N1O2N101N1O2N101gB^OQ in this image.", "objects": [{"patches": [141, 142, 143, 163, 164, 165, 166, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 233, 234, 235], "bbox": [0.09796875000000001, 0.3771875, 0.31478124999999996, 0.6234375], "iscrowd": 0, "area": 9643.33, "rle": {"size": [480, 640], "counts": "ngm04g>6L3O2O0O2N1O2N101N1O2N101gB^OQ in this image.", "objects": [{"patches": [146, 147, 148, 168, 169, 170, 171, 191, 192, 193, 194, 213, 214, 215, 216, 235, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 265, 282, 283, 284, 285, 286, 287], "bbox": [0.25309375, 0.382875, 0.5545312499999999, 0.7485208333333333], "iscrowd": 0, "area": 17455.215700000004, "rle": {"size": [480, 640], "counts": "WW\\22l>5K5K5J6K5K4L5K5K5K5K4L3N2N2N2N2O1N2N2O1N1O2N2O1N2N2N2O1N2N2N2O1N2N2O0O2N2N2O1O0O0bNfMXFZ2g9fMYF[2g9dMZF\\2f9cM[F]2e9bM[F_2d9aM]F_2c9`M]Fa2c9^M^Fa2b9_M^Fb2b9]M_Fc2a9\\M`Fd2`9[M`Ff2_9ZMbFf2^9YMbFh2^9XMbFh2^9XMbFg2^9ZMaFg2_9YM_Fi2a9WM]Fk2b9VM\\Fl2d9TMZFn2f9RMXFP3h9PMWFQ3h9PMVFQ3k9oLSFS3m9mLQFU3o9kLoEW3P:jLnEX3R:k00000O010O100O10000O010O10000O100O10O0101O0OeMVFBj99\\FGb94dFL\\92gFMY93gFMY92iFNV92kFMU92lFNT91nFOQ91oFOQ90QGOo81QGOo80SG0l8OVG0j80VG0j8OXG1g8OYG1g8N[G1e8N\\G3d8L]G3c8L_G3a8M_G3a8LaG4^8KcG5]8KdG4\\8KeG6Z8IhG6X8JhG6X8IjG6V8IlG7S8ImG7S8HnG8R8GPH9o7GQH9o7GRH8Q8EoG;U8@lG`0X8\\OiGc0Z8ZOfGg0]8TOdGl0`8POaGo0a8oN_GQ1b8nN_GQ1c8lN^GT1d8jN\\GV1f8hN[GW1g8gNYGY1h8eNZGZ1h8dNXG\\1j8bNVG^1l8_NVG`1l8^NTGa1n8ZNVGc1V;M2M4L3M3M4N3N2M3N3L3N2N2M3N1N010O0100O010O01000O010O010O010O10O010O10O10O010O010O012N2M4M2M3N2NoXU4"}, "label": "the lamb in between the other lamb and the mother sheep"}]} {"id": 470977, "image": "COCO_train2014_000000470977.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby lamb sitting closer to the older lamb in the photo\"."}], "task": "refering", "answer_template": "The \"a baby lamb sitting closer to the older lamb in the photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 168, 169, 170, 171, 191, 192, 193, 194, 213, 214, 215, 216, 235, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 265, 282, 283, 284, 285, 286, 287], "bbox": [0.25309375, 0.382875, 0.5545312499999999, 0.7485208333333333], "iscrowd": 0, "area": 17455.215700000004, "rle": {"size": [480, 640], "counts": "WW\\22l>5K5K5J6K5K4L5K5K5K5K4L3N2N2N2N2O1N2N2O1N1O2N2O1N2N2N2O1N2N2N2O1N2N2O0O2N2N2O1O0O0bNfMXFZ2g9fMYF[2g9dMZF\\2f9cM[F]2e9bM[F_2d9aM]F_2c9`M]Fa2c9^M^Fa2b9_M^Fb2b9]M_Fc2a9\\M`Fd2`9[M`Ff2_9ZMbFf2^9YMbFh2^9XMbFh2^9XMbFg2^9ZMaFg2_9YM_Fi2a9WM]Fk2b9VM\\Fl2d9TMZFn2f9RMXFP3h9PMWFQ3h9PMVFQ3k9oLSFS3m9mLQFU3o9kLoEW3P:jLnEX3R:k00000O010O100O10000O010O10000O100O10O0101O0OeMVFBj99\\FGb94dFL\\92gFMY93gFMY92iFNV92kFMU92lFNT91nFOQ91oFOQ90QGOo81QGOo80SG0l8OVG0j80VG0j8OXG1g8OYG1g8N[G1e8N\\G3d8L]G3c8L_G3a8M_G3a8LaG4^8KcG5]8KdG4\\8KeG6Z8IhG6X8JhG6X8IjG6V8IlG7S8ImG7S8HnG8R8GPH9o7GQH9o7GRH8Q8EoG;U8@lG`0X8\\OiGc0Z8ZOfGg0]8TOdGl0`8POaGo0a8oN_GQ1b8nN_GQ1c8lN^GT1d8jN\\GV1f8hN[GW1g8gNYGY1h8eNZGZ1h8dNXG\\1j8bNVG^1l8_NVG`1l8^NTGa1n8ZNVGc1V;M2M4L3M3M4N3N2M3N3L3N2N2M3N1N010O0100O010O01000O010O010O010O10O010O10O10O010O010O012N2M4M2M3N2NoXU4"}, "label": "a baby lamb sitting closer to the older lamb in the photo"}]} {"id": 470977, "image": "COCO_train2014_000000470977.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mother sheep laying in hay watching over her two babies\"."}], "task": "refering", "answer_template": "The \"a mother sheep laying in hay watching over her two babies\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 61, 62, 63, 64, 65, 66, 83, 84, 85, 86, 87, 88, 89, 90, 106, 107, 108, 109, 110, 111, 112, 113, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.555390625, 0.1115625, 1.0, 0.9985625], "iscrowd": 0, "area": 89965.86405, "rle": {"size": [480, 640], "counts": "dcV51i>:F8H8L4L3M3M3M3M3M2N3M3M2O2M3N1O2M3N1N3N1N3^JoMRNS2l1SNmMo1R2RNmMo1R2RNlMP2R2RNmMo1S2QNlMP2T2PNjMR2V2nMiMS2V2nMhMT2X2lMgMU2Y2kMfMV2Z2jMeMV2\\2jMcMW2]2iMbMX2^2hM`MZ2`2fM_MZ2b2fM]M[2c2eM\\M\\2c2eM\\M[2e2eMZM\\2g2cMXM^2h2bM\\L\\O]MS3X6aM[L\\O\\MT3Y6`MZL^O[MS3[6_MYL@ZMR3]6^MXLBYMP3`6^MVLCYMo2b6^MTLEXMn2d6]MSLFXMm2f6]MQLGWMm2h6]MPLHVMk2`6hMZL]OUMl2U6SNeLROUMk2T5VOfMPNUMj2k4APNfMSMi2j4ERNdMRMg2i4IUN`MPMi2g4KXNi0VNRLW2Y3Cd0UNYLQ2V3Ia0UN^Lk1U30=SNbLg1U35:TNcLc1U397UNfL_1T3;7VNgL[1T3?5VNhLX1U3?6YNgLT1U3`07[NfLR1U3`07_NeLo0U3`08aNcLm0W3`08cNbLj0X3a08fN_Lg0[3a08hN^Ld0\\3a09lN[L`0^3c07oNZL=`3d06oN[L:a3g04POZL7d3i02QOZL3f3l00QO[L0g3n0NTOZLMi3o0MTO[LKi3Q1LUOZLIk3R1KVOZLFl3T1JVOZLEm3T1JXOYLBn3V1HZOZL^Oo3X1GZOZL\\OQ4Z1E[OZLYOR4\\1D[OZLXOS4\\1D]OYLUOT4^1B_OZLQOU4`1A_OZLPOV4a1@@ZLmNW4c1_OAYLkNY4c1_OBYLiNY4e1^OCYLfNZ4f1]OEYLcN\\4h1[OFYL`N]4i1[OGXL_N^4i1[OHXL]N^4j1[OJUL\\Na4j1ZOJUL[Nb4j1ZOLSLYNd4j1ZOMQLYNf4i1ZONPLXNg4j1YOOoKVNi4j1YO0mKVNk4i1YO1a1O_N2`1MaN3_1MaN3_1LbN4^1LcN3]1LeNjN\\K:o5l0fNfNhK4b5U1gNeNnK1[5Z1hNbNTLOT5_1iN`NXLMo4c1jN^Nh2b1YM\\Nh2d1YMZNh2f1Y500000000001O001O00001O00001O00001O00000000001O001O1O001O001O00000000000000O1001O001O001O00001O000000O100O1000000000000O1O1O1000000001O2N1O:F00O1O1O10000001O001O001O001O00001O001O00O10000O100000000003M4L:F00O1O17IO100O17I00000000O1002N2N5K6J001O00001O3M1O1O7I001O002N1O1O2N2fEcKm9^4PFdKP:g4O1O1O001O5K5K3M1O0000O1O1O100O1O1000000000000kJ"}, "label": "a mother sheep laying in hay watching over her two babies"}]} {"id": 470977, "image": "COCO_train2014_000000470977.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult sheep looking backwards\"."}], "task": "refering", "answer_template": "The \"an adult sheep looking backwards\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 61, 62, 63, 64, 65, 66, 83, 84, 85, 86, 87, 88, 89, 90, 106, 107, 108, 109, 110, 111, 112, 113, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344, 359, 360, 361, 362, 363, 364, 365, 366, 367, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.555390625, 0.1115625, 1.0, 0.9985625], "iscrowd": 0, "area": 89965.86405, "rle": {"size": [480, 640], "counts": "dcV51i>:F8H8L4L3M3M3M3M3M2N3M3M2O2M3N1O2M3N1N3N1N3^JoMRNS2l1SNmMo1R2RNmMo1R2RNlMP2R2RNmMo1S2QNlMP2T2PNjMR2V2nMiMS2V2nMhMT2X2lMgMU2Y2kMfMV2Z2jMeMV2\\2jMcMW2]2iMbMX2^2hM`MZ2`2fM_MZ2b2fM]M[2c2eM\\M\\2c2eM\\M[2e2eMZM\\2g2cMXM^2h2bM\\L\\O]MS3X6aM[L\\O\\MT3Y6`MZL^O[MS3[6_MYL@ZMR3]6^MXLBYMP3`6^MVLCYMo2b6^MTLEXMn2d6]MSLFXMm2f6]MQLGWMm2h6]MPLHVMk2`6hMZL]OUMl2U6SNeLROUMk2T5VOfMPNUMj2k4APNfMSMi2j4ERNdMRMg2i4IUN`MPMi2g4KXNi0VNRLW2Y3Cd0UNYLQ2V3Ia0UN^Lk1U30=SNbLg1U35:TNcLc1U397UNfL_1T3;7VNgL[1T3?5VNhLX1U3?6YNgLT1U3`07[NfLR1U3`07_NeLo0U3`08aNcLm0W3`08cNbLj0X3a08fN_Lg0[3a08hN^Ld0\\3a09lN[L`0^3c07oNZL=`3d06oN[L:a3g04POZL7d3i02QOZL3f3l00QO[L0g3n0NTOZLMi3o0MTO[LKi3Q1LUOZLIk3R1KVOZLFl3T1JVOZLEm3T1JXOYLBn3V1HZOZL^Oo3X1GZOZL\\OQ4Z1E[OZLYOR4\\1D[OZLXOS4\\1D]OYLUOT4^1B_OZLQOU4`1A_OZLPOV4a1@@ZLmNW4c1_OAYLkNY4c1_OBYLiNY4e1^OCYLfNZ4f1]OEYLcN\\4h1[OFYL`N]4i1[OGXL_N^4i1[OHXL]N^4j1[OJUL\\Na4j1ZOJUL[Nb4j1ZOLSLYNd4j1ZOMQLYNf4i1ZONPLXNg4j1YOOoKVNi4j1YO0mKVNk4i1YO1a1O_N2`1MaN3_1MaN3_1LbN4^1LcN3]1LeNjN\\K:o5l0fNfNhK4b5U1gNeNnK1[5Z1hNbNTLOT5_1iN`NXLMo4c1jN^Nh2b1YM\\Nh2d1YMZNh2f1Y500000000001O001O00001O00001O00001O00000000001O001O1O001O001O00000000000000O1001O001O001O00001O000000O100O1000000000000O1O1O1000000001O2N1O:F00O1O1O10000001O001O001O001O00001O001O00O10000O100000000003M4L:F00O1O17IO100O17I00000000O1002N2N5K6J001O00001O3M1O1O7I001O002N1O1O2N2fEcKm9^4PFdKP:g4O1O1O001O5K5K3M1O0000O1O1O100O1O1000000000000kJ"}, "label": "an adult sheep looking backwards"}]} {"id": 69579, "image": "COCO_train2014_000000069579.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white airplane facing a building\"."}], "task": "refering", "answer_template": "The \"a white airplane facing a building\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 159, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 272, 273, 274, 296], "bbox": [0.6064375, 0.3865727699530517, 0.99840625, 0.8347887323943662], "iscrowd": 0, "area": 10198.963249999992, "rle": {"size": [426, 640], "counts": "`TT51Y=000[=0eB00000010O000001O0000010O0000001O01O0001O00000010O000001O0001O01O0000001O01O0001O0000010O0000001O0001O01O00000010O000001O0000010O0000001O0001O01O000oCKm:5fD:X;GgD;X;DgD?F[OW;6SEa0A\\O\\;3SEi0m:XOQEi0o:WOQEi0o:WOPEk0o:UOQEl0o:TOoDo0o:QOQEP1n:POREQ1m:oNRER1n:oNQEP1P;POoDo0T;POlDo0U;QOjDn0X;SOgDl0Z;TOeDk0];UObDj0`;VO`Dh0b;YO\\Dg0f;XOYDg0i;YOVDg0k;YOUDe0m;;1N2O2M2O1N2O1O100010O0000000010O0000000001O01O000001O0iCQOSb:B_E;d:C^E:e:D]E:d:YOiD7d0=e:\\OhD6d0;e:@hD4d07h:EfD2bO10000000O1000O100000000000000O10O1000000[6"}, "label": "a white airplane facing a building"}]} {"id": 69579, "image": "COCO_train2014_000000069579.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pure white plane parked on a runway\"."}], "task": "refering", "answer_template": "The \"pure white plane parked on a runway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 159, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 272, 273, 274, 296], "bbox": [0.6064375, 0.3865727699530517, 0.99840625, 0.8347887323943662], "iscrowd": 0, "area": 10198.963249999992, "rle": {"size": [426, 640], "counts": "`TT51Y=000[=0eB00000010O000001O0000010O0000001O01O0001O00000010O000001O0001O01O0000001O01O0001O0000010O0000001O0001O01O00000010O000001O0000010O0000001O0001O01O000oCKm:5fD:X;GgD;X;DgD?F[OW;6SEa0A\\O\\;3SEi0m:XOQEi0o:WOQEi0o:WOPEk0o:UOQEl0o:TOoDo0o:QOQEP1n:POREQ1m:oNRER1n:oNQEP1P;POoDo0T;POlDo0U;QOjDn0X;SOgDl0Z;TOeDk0];UObDj0`;VO`Dh0b;YO\\Dg0f;XOYDg0i;YOVDg0k;YOUDe0m;;1N2O2M2O1N2O1O100010O0000000010O0000000001O01O000001O0iCQOSb:B_E;d:C^E:e:D]E:d:YOiD7d0=e:\\OhD6d0;e:@hD4d07h:EfD2bO10000000O1000O100000000000000O10O1000000[6"}, "label": "pure white plane parked on a runway"}]} {"id": 298956, "image": "COCO_train2014_000000298956.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the taller giraffe\"."}], "task": "refering", "answer_template": "The \"the taller giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 58, 59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 299, 300, 301, 322, 323], "bbox": [0.0, 0.10414117647058824, 0.75278125, 1.0], "iscrowd": 0, "area": 50205.91504999999, "rle": {"size": [425, 640], "counts": "g8b4g80O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1O100O1O10YOf001N10001O0O2O00001N10001O0O2O00001N10001O0O2O00001N10O1O1O001O1O1O010001O0O2O010O010O010O010O010O010O010O010O10O010O010O010O010O010O010O010O010O010O0100O010O010O010O010O010O010O100O10O010000O10O0100O100O010O100O10O0100O100O010O100O10O0100O100O010M3M3M3M2M4M3M3M3M2O2O100O10O01O100O10O0100O1O10O0100O100O00100O100O00100O100O010O1O010O010O1O010O010O010O0010O010O01O010O010O010O00100O010O010O0010O010O01O010O010O010O0010O0100O0010O010O010O01O010O010O0010O010O010O01O10O010O010O01O010O01O00001O001O010O0000000000000000O2O000000000000000O101O0000000000000O10001O00000000000O100000001O000000000O100000001O00XOeMcF[2]9gMaFY2`9iM]FV2d9mMYFS2h9nMVFR2j9QNSFo1n9SNoEm1R:TNlEl1T:WNjEh1W:d0001N101O001O001O001O0O20O010O0101N100O2O0O101N100O2O0O101N100O2O0O101N10001N100O2O0O101N100O2O0O101N100O100O100O100O100O2O000O100O100O100O100O2O0O100O100O100O100O100O2O0O100O100O101N101N100O2O0O2O0O101N101N101N10001N101N1O1L5J5K^`Q2"}, "label": "the taller giraffe"}]} {"id": 298956, "image": "COCO_train2014_000000298956.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall giraffe eating out of a metal basket\"."}], "task": "refering", "answer_template": "The \"a tall giraffe eating out of a metal basket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 58, 59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 85, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 142, 143, 144, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 299, 300, 301, 322, 323], "bbox": [0.0, 0.10414117647058824, 0.75278125, 1.0], "iscrowd": 0, "area": 50205.91504999999, "rle": {"size": [425, 640], "counts": "g8b4g80O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1O100O1O10YOf001N10001O0O2O00001N10001O0O2O00001N10001O0O2O00001N10O1O1O001O1O1O010001O0O2O010O010O010O010O010O010O010O010O10O010O010O010O010O010O010O010O010O010O0100O010O010O010O010O010O010O100O10O010000O10O0100O100O010O100O10O0100O100O010O100O10O0100O100O010M3M3M3M2M4M3M3M3M2O2O100O10O01O100O10O0100O1O10O0100O100O00100O100O00100O100O010O1O010O010O1O010O010O010O0010O010O01O010O010O010O00100O010O010O0010O010O01O010O010O010O0010O0100O0010O010O010O01O010O010O0010O010O010O01O10O010O010O01O010O01O00001O001O010O0000000000000000O2O000000000000000O101O0000000000000O10001O00000000000O100000001O000000000O100000001O00XOeMcF[2]9gMaFY2`9iM]FV2d9mMYFS2h9nMVFR2j9QNSFo1n9SNoEm1R:TNlEl1T:WNjEh1W:d0001N101O001O001O001O0O20O010O0101N100O2O0O101N100O2O0O101N100O2O0O101N10001N100O2O0O101N100O2O0O101N100O100O100O100O100O2O000O100O100O100O100O2O0O100O100O100O100O100O2O0O100O100O101N101N100O2O0O2O0O101N101N101N10001N101N1O1L5J5K^`Q2"}, "label": "a tall giraffe eating out of a metal basket"}]} {"id": 298956, "image": "COCO_train2014_000000298956.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe with a lowered head watching the higher giraffe eat\"."}], "task": "refering", "answer_template": "The \"a giraffe with a lowered head watching the higher giraffe eat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 309, 324, 325, 326, 327, 328, 329, 330, 331, 332], "bbox": [0.0835625, 0.4404470588235294, 0.87596875, 0.9887529411764706], "iscrowd": 0, "area": 40801.1527, "rle": {"size": [425, 640], "counts": "Ujf04P=5L4L4L4M3O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O100O100O1O100O100O100O100O100O100O100O100O100O100O101N100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1O100O1000000O1000000O10000O1000000O10000O1000000O1000000O10000O1000000O10000000000O1000000000000000000000000000000000000000000000000000000O1000000000000000QNhGmNX8R1jGmNV8Q1nGmNR8R1QHlNo7S1SHkNn7T1UHgNn7X1THeNn7Z1UHaNn7]1VH^Nm7a1UH[Nn7d1UHXNm7g1VHTNm7k1UHQNn7m1VHnMm7Q2UHlMm7S2VHhMm7X2b1O10000O010O10000O10000O10000O100O10000O10000O10O10O100O01000O10O10O1000O0100O10O10O10O10O1000O0100O10O10O1000O0100O01000O10O10O1000O0100O10O10O10O10O1000O0100O10O0100O1O010O1O10O01O10O01O100O00100O1O010O1O01O00001O01O01O00001O000010O000001O00001O01O01O00001O00010O000100O00100O00100O00100O0010O0BZN[Ee1e:\\N[Ed1d:^NZEb1g:_NXEa1g:aNXE^1h:dNVE]1i:eNVE[1j:fNUEY1k:iNSEX1n:hNQEX1o:iNPEW1P;`02O00001N10001N10001O0O101O000O2O000O2O00001N10001N10001N100O2O000O2O0O101O0O2O1O0O2O0O2O1O0O2O0O2O1O001O001O10O01O00100O001O010O1O001O10O01O1O010O1O001O010O1O00100O001N2N1N3N2M2O2M3N1OVoP1"}, "label": "a giraffe with a lowered head watching the higher giraffe eat"}]} {"id": 298956, "image": "COCO_train2014_000000298956.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the shorter giraffe with it ' s eyes closed standing in front of a taller giraffe\"."}], "task": "refering", "answer_template": "The \"the shorter giraffe with it ' s eyes closed standing in front of a taller giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 309, 324, 325, 326, 327, 328, 329, 330, 331, 332], "bbox": [0.0835625, 0.4404470588235294, 0.87596875, 0.9887529411764706], "iscrowd": 0, "area": 40801.1527, "rle": {"size": [425, 640], "counts": "Ujf04P=5L4L4L4M3O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O100O100O1O100O100O100O100O100O100O100O100O100O100O101N100O100O100O100O1O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O100O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1O100O1000000O1000000O10000O1000000O10000O1000000O1000000O10000O1000000O10000000000O1000000000000000000000000000000000000000000000000000000O1000000000000000QNhGmNX8R1jGmNV8Q1nGmNR8R1QHlNo7S1SHkNn7T1UHgNn7X1THeNn7Z1UHaNn7]1VH^Nm7a1UH[Nn7d1UHXNm7g1VHTNm7k1UHQNn7m1VHnMm7Q2UHlMm7S2VHhMm7X2b1O10000O010O10000O10000O10000O100O10000O10000O10O10O100O01000O10O10O1000O0100O10O10O10O10O1000O0100O10O10O1000O0100O01000O10O10O1000O0100O10O10O10O10O1000O0100O10O0100O1O010O1O10O01O10O01O100O00100O1O010O1O01O00001O01O01O00001O000010O000001O00001O01O01O00001O00010O000100O00100O00100O00100O0010O0BZN[Ee1e:\\N[Ed1d:^NZEb1g:_NXEa1g:aNXE^1h:dNVE]1i:eNVE[1j:fNUEY1k:iNSEX1n:hNQEX1o:iNPEW1P;`02O00001N10001N10001O0O101O000O2O000O2O00001N10001N10001N100O2O000O2O0O101O0O2O1O0O2O0O2O1O0O2O0O2O1O001O001O10O01O00100O001O010O1O001O10O01O1O010O1O001O010O1O00100O001N2N1N3N2M2O2M3N1OVoP1"}, "label": "the shorter giraffe with it ' s eyes closed standing in front of a taller giraffe"}]} {"id": 511967, "image": "COCO_train2014_000000511967.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man helping another man put on a tie ,\"."}], "task": "refering", "answer_template": "The \"a man helping another man put on a tie ,\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 124, 125, 126, 127, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 179, 180, 181, 182, 193, 194, 195, 196, 197, 198, 202, 203, 204, 205, 206, 213, 214, 215, 216, 217, 218, 219, 220, 221, 226, 227, 228, 229, 234, 235, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.15592187500000002, 0.2943793911007026, 0.9655468749999999, 0.9842622950819672], "iscrowd": 0, "area": 49183.988499999985, "rle": {"size": [427, 640], "counts": "RRZ18o<:E in this image.", "objects": [{"patches": [102, 103, 104, 124, 125, 126, 127, 147, 148, 149, 150, 151, 170, 171, 172, 173, 174, 179, 180, 181, 182, 193, 194, 195, 196, 197, 198, 202, 203, 204, 205, 206, 213, 214, 215, 216, 217, 218, 219, 220, 221, 226, 227, 228, 229, 234, 235, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.15592187500000002, 0.2943793911007026, 0.9655468749999999, 0.9842622950819672], "iscrowd": 0, "area": 49183.988499999985, "rle": {"size": [427, 640], "counts": "RRZ18o<:E in this image.", "objects": [{"patches": [14, 15, 16, 17, 36, 37, 38, 39, 40, 41, 42, 59, 60, 61, 62, 63, 64, 65, 81, 82, 83, 84, 85, 86, 87, 88, 104, 105, 106, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.40481249999999996, 0.03976580796252927, 1.0, 0.9835831381733021], "iscrowd": 0, "area": 89737.48425000001, "rle": {"size": [427, 640], "counts": "T]\\31X=2M3N2N2N2O1O1O1N101O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O001O1N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O001N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O2N1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1M3N2N2M3N2O100O100TL[LQNf3b1lLXNU3[1\\MaNd2R1mMiNT2P1WNkNj1S1]NhNd1V1aNeN`1N1NOO8LI1MB0a0O@Nd00]OMf03ZOJj04WOIl07TOFP19PODS1iNAY1>gNAZ1?fN@\\1?dN@]1`0cN_O_1`0aN_O`1a0`N^Ob1a0^N^Oc1b0]N\\Of1c0ZN\\Og1d0YN[Oi1d0WN[Oj1e0VNZOl1e0TNZOm1f0SNYOo1f0QNYOP2g0PNXOR2g0nMXOT2g0lMYOT2g0lMYOU2f0kMYOV2g0jMYOW2f0iMZOW2f0iMYOY2f0gMZOZ2e0fM[O`2?`M@f2;ZMEk25VMKo2ORM1T3GnL8X3AjL?[3[OfLe0_3TOcLk0o70000O10000O10000O1000000O1000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000001O001O001O001O001O001O001O001O001O001O001O001O001O001O0O2O001O001O001O001O001O001O001O001O001O001O001O0VJ^NKc13`NK`13cNL^12dNM\\11gNNZ10iNNW10lNOU1OmN0S1NQO0P1NRO2m0LVO3k0KWO4k0HYO6i0GYO8i0EYO:i0CYOi0@XO?j0^OXOa0j0]OWOb0l0ZOWOd0k0YOWOf0k0XOVOg0l0VOWOh0l0TOVOk0l0SOUOl0m0QOUOn0m0POUOn0n0nNSOR1o0kNSOT1o0hNSOX1T4_KnKa4^60000000000000001O000000000000000000001O000000000O10000000001O0000000O1000000000001O000000000000000000001O0000000000000000001O00000000000000000000000000000000O10000000O100Oh0YO^2aMYD"}, "label": "the man who is putting the tie around his neck"}]} {"id": 511967, "image": "COCO_train2014_000000511967.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man putting a tie on his own neck\"."}], "task": "refering", "answer_template": "The \"a man putting a tie on his own neck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 36, 37, 38, 39, 40, 41, 42, 59, 60, 61, 62, 63, 64, 65, 81, 82, 83, 84, 85, 86, 87, 88, 104, 105, 106, 107, 108, 109, 110, 111, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.40481249999999996, 0.03976580796252927, 1.0, 0.9835831381733021], "iscrowd": 0, "area": 89737.48425000001, "rle": {"size": [427, 640], "counts": "T]\\31X=2M3N2N2N2O1O1O1N101O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O001O1N2O1O1O1N2O1O1O1N2O1O1O1O1N2O1O001N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O2N1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1M3N2N2M3N2O100O100TL[LQNf3b1lLXNU3[1\\MaNd2R1mMiNT2P1WNkNj1S1]NhNd1V1aNeN`1N1NOO8LI1MB0a0O@Nd00]OMf03ZOJj04WOIl07TOFP19PODS1iNAY1>gNAZ1?fN@\\1?dN@]1`0cN_O_1`0aN_O`1a0`N^Ob1a0^N^Oc1b0]N\\Of1c0ZN\\Og1d0YN[Oi1d0WN[Oj1e0VNZOl1e0TNZOm1f0SNYOo1f0QNYOP2g0PNXOR2g0nMXOT2g0lMYOT2g0lMYOU2f0kMYOV2g0jMYOW2f0iMZOW2f0iMYOY2f0gMZOZ2e0fM[O`2?`M@f2;ZMEk25VMKo2ORM1T3GnL8X3AjL?[3[OfLe0_3TOcLk0o70000O10000O10000O1000000O1000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000001O001O001O001O001O001O001O001O001O001O001O001O001O001O0O2O001O001O001O001O001O001O001O001O001O001O001O0VJ^NKc13`NK`13cNL^12dNM\\11gNNZ10iNNW10lNOU1OmN0S1NQO0P1NRO2m0LVO3k0KWO4k0HYO6i0GYO8i0EYO:i0CYOi0@XO?j0^OXOa0j0]OWOb0l0ZOWOd0k0YOWOf0k0XOVOg0l0VOWOh0l0TOVOk0l0SOUOl0m0QOUOn0m0POUOn0n0nNSOR1o0kNSOT1o0hNSOX1T4_KnKa4^60000000000000001O000000000000000000001O000000000O10000000001O0000000O1000000000001O000000000000000000001O0000000000000000001O00000000000000000000000000000000O10000000O100Oh0YO^2aMYD"}, "label": "a man putting a tie on his own neck"}]} {"id": 143334, "image": "COCO_train2014_000000143334.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cup\"."}], "task": "refering", "answer_template": "The \"cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [284, 285, 286, 288, 289, 306, 307, 308, 309, 310, 311, 312, 313, 314, 329, 330, 331, 332, 333, 334, 335, 336, 337, 352, 353, 354, 355, 357, 358, 359, 360, 375, 376, 377, 378, 380, 381, 382, 383, 398, 399, 400, 401, 403, 404, 405, 406, 423, 427], "bbox": [0.30729687499999997, 0.524734375, 0.6822968749999999, 0.7955781249999999], "iscrowd": 0, "area": 26517.969450000004, "rle": {"size": [640, 640], "counts": "fak3;`c0 in this image.", "objects": [{"patches": [242, 260, 261, 262, 263, 264, 265, 266, 267, 282, 283, 284, 285, 288, 289, 290, 291, 305, 306, 307, 313, 314, 328, 329, 337, 351, 360, 361, 374, 383], "bbox": [0.25768749999999996, 0.470390625, 0.7058125000000001, 0.744921875], "iscrowd": 0, "area": 11056.607799999994, "rle": {"size": [640, 640], "counts": "iaW39Wc0a0_Oa0^Ob0I7N2M3N2M3N2M2O2M3M3N2M3N2M3N2M3N2M3N2M3M2OkNP@]Nm?b1\\@YNb?e1g@UNW?j1RAQNl>m1]AnMa>P2hAjMW>T2k1M3N2N3M2N2N2M3N2N3N10000000001O00000000001O000001O0001O00000000001O0000001O1O1O00100O1O1O1O001O1O1O1O1O001O1O1O10O1000O100000O0100000000O0100000O10O100000O1000O10O10000000O01000000O010000000O10O1000O1000hTb0WOnk]O3N2M4M2M3N3L3N3M2M3N3O01O010O00010O010O01O010O01O01O010O010O010O0010O010O00010O010O01O010O010O0010O0010O0010O010O01O010O010O01O01O100O101N1O100010O000000010O0000010O000001O01O0000010O0000010O000000010O0000010O09G:F:F9H9F20K8H8I7H8I6I8H8I7H8I7H7JXRe3"}, "label": "the second to largest measuring cup"}]} {"id": 143334, "image": "COCO_train2014_000000143334.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the second largest measuring cup\"."}], "task": "refering", "answer_template": "The \"the second largest measuring cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [242, 260, 261, 262, 263, 264, 265, 266, 267, 282, 283, 284, 285, 288, 289, 290, 291, 305, 306, 307, 313, 314, 328, 329, 337, 351, 360, 361, 374, 383], "bbox": [0.25768749999999996, 0.470390625, 0.7058125000000001, 0.744921875], "iscrowd": 0, "area": 11056.607799999994, "rle": {"size": [640, 640], "counts": "iaW39Wc0a0_Oa0^Ob0I7N2M3N2M3N2M2O2M3M3N2M3N2M3N2M3N2M3N2M3M2OkNP@]Nm?b1\\@YNb?e1g@UNW?j1RAQNl>m1]AnMa>P2hAjMW>T2k1M3N2N3M2N2N2M3N2N3N10000000001O00000000001O000001O0001O00000000001O0000001O1O1O00100O1O1O1O001O1O1O1O1O001O1O1O10O1000O100000O0100000000O0100000O10O100000O1000O10O10000000O01000000O010000000O10O1000O1000hTb0WOnk]O3N2M4M2M3N3L3N3M2M3N3O01O010O00010O010O01O010O01O01O010O010O010O0010O010O00010O010O01O010O010O0010O0010O0010O010O01O010O010O01O01O100O101N1O100010O000000010O0000010O000001O01O0000010O0000010O000000010O0000010O09G:F:F9H9F20K8H8I7H8I6I8H8I7H8I7H7JXRe3"}, "label": "the second largest measuring cup"}]} {"id": 53232, "image": "COCO_train2014_000000053232.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bed closest to the camera\"."}], "task": "refering", "answer_template": "The \"the bed closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 101, 102, 103, 104, 105, 106, 107, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215], "bbox": [0.22132, 0.39549253731343287, 1.0, 0.9887761194029852], "iscrowd": 0, "area": 58509.6256, "rle": {"size": [335, 500], "counts": "j`T1;S:=D in this image.", "objects": [{"patches": [89, 101, 102, 103, 104, 105, 106, 107, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215], "bbox": [0.22132, 0.39549253731343287, 1.0, 0.9887761194029852], "iscrowd": 0, "area": 58509.6256, "rle": {"size": [335, 500], "counts": "j`T1;S:=D in this image.", "objects": [{"patches": [96, 97, 98, 99, 109, 110, 111, 112, 113, 114, 115, 116, 117, 126, 127, 128, 129, 130, 131, 132, 133, 134, 144, 145, 146, 147, 165], "bbox": [0.03748, 0.44764179104477614, 0.50962, 0.798865671641791], "iscrowd": 0, "area": 12060.4824, "rle": {"size": [335, 500], "counts": "U]6g0h95J5K3M2O0O2O0O2O0O1O100O100O101O001N2O1O0O2O1O1N2O1O1N1000000000000O101O000000000O10001O000000000O1000000000000O1000002N2N2N2N2M3N2N2N2N00000O0100O10000O10O10O10000O100O10O01F:F:G9F:K50O10O100000000000O0100000000000O01000000000000O0100000000000O0100000000000O10O1000000000O10O1N2M3000O100000000000N1E<00000000000O10000000O1000O1000000000000O1000000000000O1000000000O10O1000000000000O1000000000000O1000000000O01000000000000O1000000000000O2O00000000001N1M3L4O1N2N2I8H7H8IYP`2"}, "label": "the bed closest to window"}]} {"id": 53232, "image": "COCO_train2014_000000053232.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bed with a brown cylindrical pillow near the window\"."}], "task": "refering", "answer_template": "The \"a bed with a brown cylindrical pillow near the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 109, 110, 111, 112, 113, 114, 115, 116, 117, 126, 127, 128, 129, 130, 131, 132, 133, 134, 144, 145, 146, 147, 165], "bbox": [0.03748, 0.44764179104477614, 0.50962, 0.798865671641791], "iscrowd": 0, "area": 12060.4824, "rle": {"size": [335, 500], "counts": "U]6g0h95J5K3M2O0O2O0O2O0O1O100O100O101O001N2O1O0O2O1O1N2O1O1N1000000000000O101O000000000O10001O000000000O1000000000000O1000002N2N2N2N2M3N2N2N2N00000O0100O10000O10O10O10000O100O10O01F:F:G9F:K50O10O100000000000O0100000000000O01000000000000O0100000000000O0100000000000O10O1000000000O10O1N2M3000O100000000000N1E<00000000000O10000000O1000O1000000000000O1000000000000O1000000000O10O1000000000000O1000000000000O1000000000O01000000000000O1000000000000O2O00000000001N1M3L4O1N2N2I8H7H8IYP`2"}, "label": "a bed with a brown cylindrical pillow near the window"}]} {"id": 266228, "image": "COCO_train2014_000000266228.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green and white boat\"."}], "task": "refering", "answer_template": "The \"a green and white boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [203, 206, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.474921875, 0.5263231850117096, 1.0, 1.0], "iscrowd": 0, "area": 36898.66205, "rle": {"size": [427, 640], "counts": "\\To31Y=1N3N1O1N3N1O2M2O20O010O010O01O010O10O010OO2O001N101O0O2O001O0O10000O102N2M3N2N3M2M3N2N3L3N2N2N2M2O1O1N10000O1000000O100O10000O100O10000O100O10000O100O10000O100O10000O100O100O10000O100O10000O100O10000O100O10000O100O10000O10000O100O10000O10000O100O10000O10000O10000O100O10000O100mNQMjGo2V8VMdGk2[8\\M_Gd2a8aMZG_2f8fMTG[2l8jMoFV2P9S10000O1000000O10000O1000000O1000000O100004L4L4L4L4L4L4L0000O100O100O100O100O100O100O100O100O1E_LlFa3S9B;EO10000O10000O10000O100O10000O10000O100O10000O10000O10000O1YO\\KgHd4Y7cK_H^4a7hKYHX4g7f06J in this image.", "objects": [{"patches": [203, 206, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.474921875, 0.5263231850117096, 1.0, 1.0], "iscrowd": 0, "area": 36898.66205, "rle": {"size": [427, 640], "counts": "\\To31Y=1N3N1O1N3N1O2M2O20O010O010O01O010O10O010OO2O001N101O0O2O001O0O10000O102N2M3N2N3M2M3N2N3L3N2N2N2M2O1O1N10000O1000000O100O10000O100O10000O100O10000O100O10000O100O10000O100O100O10000O100O10000O100O10000O100O10000O100O10000O10000O100O10000O10000O100O10000O10000O10000O100O10000O100mNQMjGo2V8VMdGk2[8\\M_Gd2a8aMZG_2f8fMTG[2l8jMoFV2P9S10000O1000000O10000O1000000O1000000O100004L4L4L4L4L4L4L0000O100O100O100O100O100O100O100O100O1E_LlFa3S9B;EO10000O10000O10000O100O10000O10000O100O10000O10000O10000O1YO\\KgHd4Y7cK_H^4a7hKYHX4g7f06J in this image.", "objects": [{"patches": [168, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 240, 241, 242, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288], "bbox": [0.159703125, 0.45882903981264633, 0.5403125, 0.8592740046838406], "iscrowd": 0, "area": 23477.404049999997, "rle": {"size": [427, 640], "counts": "nkZ12X=4K5L4J6I7J5J4M3L4H8G9F:G9K5K5K5L4K5L4K5K5L4K5N2000000000000000000000000000O1000O1000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000001O000000fN[LnHe3Q7cLhH]3X7jLaHV3_7[16J7I4L5K0O1000001aKZGW4f8iK\\GU4d8kK^GS4a8nKbGo3^8QLcGm3^8SLcGl3]8TLdGk3\\8ULeGj3[8VLeGj3Z8ULiGj3W8ULkGj3U8ULlGj3l8O1O001O1O001N2O1O001O1O001O1N101O1O1O1O1O0O2O1O1O1O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O001O001O001O1O001O0O2O001O001O001N2O001O001O001O0O2O0000001O000O2O000O101N10001N100O101N10001N10000O2O00001N10000O2O1O1A?K5K5K5K[Yj3"}, "label": "old boat in the middle"}]} {"id": 266228, "image": "COCO_train2014_000000266228.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boat in motion\"."}], "task": "refering", "answer_template": "The \"boat in motion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 240, 241, 242, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288], "bbox": [0.159703125, 0.45882903981264633, 0.5403125, 0.8592740046838406], "iscrowd": 0, "area": 23477.404049999997, "rle": {"size": [427, 640], "counts": "nkZ12X=4K5L4J6I7J5J4M3L4H8G9F:G9K5K5K5L4K5L4K5K5L4K5N2000000000000000000000000000O1000O1000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000001O000000fN[LnHe3Q7cLhH]3X7jLaHV3_7[16J7I4L5K0O1000001aKZGW4f8iK\\GU4d8kK^GS4a8nKbGo3^8QLcGm3^8SLcGl3]8TLdGk3\\8ULeGj3[8VLeGj3Z8ULiGj3W8ULkGj3U8ULlGj3l8O1O001O1O001N2O1O001O1O001O1N101O1O1O1O1O0O2O1O1O1O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O001O001O001O1O001O0O2O001O001O001N2O001O001O001O0O2O0000001O000O2O000O101N10001N100O101N10001N10000O2O00001N10000O2O1O1A?K5K5K5K[Yj3"}, "label": "boat in motion"}]} {"id": 495609, "image": "COCO_train2014_000000495609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant on the right is ahead of the other one\"."}], "task": "refering", "answer_template": "The \"the elephant on the right is ahead of the other one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 263, 264, 265, 266, 267, 268, 269, 270, 271, 286, 287, 288, 290, 291, 293, 294, 311, 313, 314, 316], "bbox": [0.467890625, 0.48831381733021073, 0.802359375, 0.9012880562060891], "iscrowd": 0, "area": 19661.814850000006, "rle": {"size": [427, 640], "counts": "hPm31V=5L5J6J6J6J5K6K2M4L3M3M4M2M4L3M4O0100O101N100O101N11O3N1N3M2N3N1N3M2O2M2N3M2OO00001O0000001O00001O0000002N1O2N1iN\\EEf:7hE]OY:a0SFTOo9i0[1N3M3M2N3Lid3Ah[L7I6J7J5J7I6J7I6J7J5J7I6J7I6K6I6J6J7I6K2M3M3M2N3M2O2M2N3M2N3N2M2N3M2N2OO01O000000010O000000010O000000010O01O001O010O001O0010O0001O0010OO2E:F;F9F;E:G:E:I80O01O01O0000001O00001O01O01O0000001O00001SG`Mm6`2QIbMo6^2PIdMo6\\2oHfMQ7[2mHfMT7Y2jHjMU7V2iHlMW7U2gHmMX7S2fHoMZ7Q2eHQNZ7o1dHSN\\7n1aHTNa7j1^HXNc7f1[HZNi7j1PHVNS8o1dGRN_8R2YGoMi8W2oFiMT9U33L4M3M3M3M3M3M2M4M3M3M3M4L4K5L3M4L4L3M4K5L4L3H9ZOf0[OZXd1"}, "label": "the elephant on the right is ahead of the other one"}]} {"id": 495609, "image": "COCO_train2014_000000495609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant walking in the front\"."}], "task": "refering", "answer_template": "The \"an elephant walking in the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 225, 241, 242, 243, 244, 245, 246, 247, 248, 263, 264, 265, 266, 267, 268, 269, 270, 271, 286, 287, 288, 290, 291, 293, 294, 311, 313, 314, 316], "bbox": [0.467890625, 0.48831381733021073, 0.802359375, 0.9012880562060891], "iscrowd": 0, "area": 19661.814850000006, "rle": {"size": [427, 640], "counts": "hPm31V=5L5J6J6J6J5K6K2M4L3M3M4M2M4L3M4O0100O101N100O101N11O3N1N3M2N3N1N3M2O2M2N3M2OO00001O0000001O00001O0000002N1O2N1iN\\EEf:7hE]OY:a0SFTOo9i0[1N3M3M2N3Lid3Ah[L7I6J7J5J7I6J7I6J7J5J7I6J7I6K6I6J6J7I6K2M3M3M2N3M2O2M2N3M2N3N2M2N3M2N2OO01O000000010O000000010O000000010O01O001O010O001O0010O0001O0010OO2E:F;F9F;E:G:E:I80O01O01O0000001O00001O01O01O0000001O00001SG`Mm6`2QIbMo6^2PIdMo6\\2oHfMQ7[2mHfMT7Y2jHjMU7V2iHlMW7U2gHmMX7S2fHoMZ7Q2eHQNZ7o1dHSN\\7n1aHTNa7j1^HXNc7f1[HZNi7j1PHVNS8o1dGRN_8R2YGoMi8W2oFiMT9U33L4M3M3M3M3M3M2M4M3M3M3M4L4K5L3M4L4L3M4K5L4L3H9ZOf0[OZXd1"}, "label": "an elephant walking in the front"}]} {"id": 495609, "image": "COCO_train2014_000000495609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant in front\"."}], "task": "refering", "answer_template": "The \"the elephant in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 256, 257, 258, 259, 260, 261, 262, 263, 264, 266, 279, 280, 281, 282, 283, 284, 285, 286, 289, 290, 302, 303, 304, 305, 312, 313], "bbox": [0.1424375, 0.4157377049180328, 0.6417031249999999, 0.9056206088992975], "iscrowd": 0, "area": 45506.500950000016, "rle": {"size": [427, 640], "counts": "_VV14U==C=C=C=C=BQ2PN2N2N2N3M2N2NVOlGaLQ8Q3bHmL\\7S2kGiM]12f6e2^IZM_6U3d18I7H5aGSLW7R4]HVLb7P4PHYLo7a401O100O1O1O1O1O1O1O010O1O1O1O100O100O1O010OBZHYKf7d4aHXK^7f4jHUKV7h4g0N3M2002O1N2O1N4M6I6K5J6K5K1N100O10000O100O10000O10000O100O10000O100O10000O100O10000O1000000O1000000O2O00N2M3L3N3M3M3L3N3000000O0100000000O10000000O100000000000000O10000O100O100O100O100O100O10000O100O2O0O100000001O00000000001O00000000001N1005K8H5K01O00000000001O00000000001O00000O10001O0O1000001N101O000O2O001O000O2O001O0O1CTH[Km7`4[H\\Ke7`4f0K4L5K4L5K4L5L3L5K4L5K4L5K401O010O001O01O00001O0O10001O000O2O0000001O00001O00000O2O00001O00001O2N2N1O2O1O1N100O10000O10O07J6K5K6J1O001O001003L4M4K7J6I6I8H7I8H8I4K2N3M3M3M3M2N3N8F9G3L4M2N3M3M2N2N2N1O1O1N2O1O1O1O1OSao2"}, "label": "the elephant in front"}]} {"id": 495609, "image": "COCO_train2014_000000495609.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant that is most visible\"."}], "task": "refering", "answer_template": "The \"the elephant that is most visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 145, 146, 147, 148, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 256, 257, 258, 259, 260, 261, 262, 263, 264, 266, 279, 280, 281, 282, 283, 284, 285, 286, 289, 290, 302, 303, 304, 305, 312, 313], "bbox": [0.1424375, 0.4157377049180328, 0.6417031249999999, 0.9056206088992975], "iscrowd": 0, "area": 45506.500950000016, "rle": {"size": [427, 640], "counts": "_VV14U==C=C=C=C=BQ2PN2N2N2N3M2N2NVOlGaLQ8Q3bHmL\\7S2kGiM]12f6e2^IZM_6U3d18I7H5aGSLW7R4]HVLb7P4PHYLo7a401O100O1O1O1O1O1O1O010O1O1O1O100O100O1O010OBZHYKf7d4aHXK^7f4jHUKV7h4g0N3M2002O1N2O1N4M6I6K5J6K5K1N100O10000O100O10000O10000O100O10000O100O10000O100O10000O1000000O1000000O2O00N2M3L3N3M3M3L3N3000000O0100000000O10000000O100000000000000O10000O100O100O100O100O100O10000O100O2O0O100000001O00000000001O00000000001N1005K8H5K01O00000000001O00000000001O00000O10001O0O1000001N101O000O2O001O000O2O001O0O1CTH[Km7`4[H\\Ke7`4f0K4L5K4L5K4L5L3L5K4L5K4L5K401O010O001O01O00001O0O10001O000O2O0000001O00001O00000O2O00001O00001O2N2N1O2O1O1N100O10000O10O07J6K5K6J1O001O001003L4M4K7J6I6I8H7I8H8I4K2N3M3M3M3M2N3N8F9G3L4M2N3M3M2N2N2N1O1O1N2O1O1O1O1OSao2"}, "label": "the elephant that is most visible"}]} {"id": 421887, "image": "COCO_train2014_000000421887.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in coveralls standing next to a the wing of a plane\"."}], "task": "refering", "answer_template": "The \"a man in coveralls standing next to a the wing of a plane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 42, 43, 44, 60, 61, 76, 77, 78, 79, 93, 94, 95, 96, 97, 109, 110, 111, 112, 113, 114, 115, 125, 126, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 147, 148, 149, 160, 161, 162, 163, 164, 165, 177, 178, 179, 180, 181, 182, 195, 196, 197, 198, 199, 212, 213, 214, 215, 216, 229, 230, 231, 232, 233, 246, 247, 248, 249, 250, 263, 264, 265, 266, 267, 280, 281, 282, 283, 284, 297, 298, 299, 300, 301, 314, 315, 316, 317, 318, 331, 332, 333, 334, 335, 347, 348, 349, 350, 351, 352, 367, 368], "bbox": [0.39565843621399177, 0.070859375, 0.797201646090535, 0.960984375], "iscrowd": 0, "area": 63472.837550000004, "rle": {"size": [640, 486], "counts": "^Vh3m0Rc02N2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N2N1O2mGjMYNX2f1iMXNY2g1hMWNZ2h1gMVNZ2i1hMUN]2g1dMWNe2a1\\M]Nl2\\1UMbNT3U1nLiNZ3P1gLnNa3k0`LROo3a0QL]Oa42`KLQ5EoJ9b5XO^Jf0T6iNmIV1c6[N]Ic1U7mMkHQ2f7`MZH^2W8RMjGl2h8dLXGZ3Y9WLgFg3g9mKYFQ4l9kKUFS4P:jKPFT4U:iKkEU4Z:hKfEV4_:TKTFk4P:QKQFo4R:mJoER5U:jJlEU5X:gJiEY5Z:cJgE\\5]:`JdE`5_:\\JbEc5b:YJ_Ef5e:VJ\\Ej5g:RJZEm5j:oIWEg3A^Mlm2ZAYMe>j2RA\\Mn>f2k@`MT?b2e@dMZ?_2^@gM`?^3O1O001N2O1O1O001N2O1O1O1N101O1O1O1N2O001O1O1N2O001O1N2O1O0I8E;@`0@`0@?@a0[Oe0\\Od0[Oe0[Od0\\Oe0XOh0ROn0eN[1eNZ1kNV1O1N2O3M3M2M4M2N3M3L3N3M3M2M4M3M2N3L3N3M3M2M4L4I6K6I6K6I7IWem1"}, "label": "a man in coveralls standing next to a the wing of a plane"}]} {"id": 421887, "image": "COCO_train2014_000000421887.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a pilot suit standing next to an airplane\"."}], "task": "refering", "answer_template": "The \"a man in a pilot suit standing next to an airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 42, 43, 44, 60, 61, 76, 77, 78, 79, 93, 94, 95, 96, 97, 109, 110, 111, 112, 113, 114, 115, 125, 126, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 147, 148, 149, 160, 161, 162, 163, 164, 165, 177, 178, 179, 180, 181, 182, 195, 196, 197, 198, 199, 212, 213, 214, 215, 216, 229, 230, 231, 232, 233, 246, 247, 248, 249, 250, 263, 264, 265, 266, 267, 280, 281, 282, 283, 284, 297, 298, 299, 300, 301, 314, 315, 316, 317, 318, 331, 332, 333, 334, 335, 347, 348, 349, 350, 351, 352, 367, 368], "bbox": [0.39565843621399177, 0.070859375, 0.797201646090535, 0.960984375], "iscrowd": 0, "area": 63472.837550000004, "rle": {"size": [640, 486], "counts": "^Vh3m0Rc02N2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N2N1O2mGjMYNX2f1iMXNY2g1hMWNZ2h1gMVNZ2i1hMUN]2g1dMWNe2a1\\M]Nl2\\1UMbNT3U1nLiNZ3P1gLnNa3k0`LROo3a0QL]Oa42`KLQ5EoJ9b5XO^Jf0T6iNmIV1c6[N]Ic1U7mMkHQ2f7`MZH^2W8RMjGl2h8dLXGZ3Y9WLgFg3g9mKYFQ4l9kKUFS4P:jKPFT4U:iKkEU4Z:hKfEV4_:TKTFk4P:QKQFo4R:mJoER5U:jJlEU5X:gJiEY5Z:cJgE\\5]:`JdE`5_:\\JbEc5b:YJ_Ef5e:VJ\\Ej5g:RJZEm5j:oIWEg3A^Mlm2ZAYMe>j2RA\\Mn>f2k@`MT?b2e@dMZ?_2^@gM`?^3O1O001N2O1O1O001N2O1O1O1N101O1O1O1N2O001O1O1N2O001O1N2O1O0I8E;@`0@`0@?@a0[Oe0\\Od0[Oe0[Od0\\Oe0XOh0ROn0eN[1eNZ1kNV1O1N2O3M3M2M4M2N3M3L3N3M3M2M4M3M2N3L3N3M3M2M4L4I6K6I6K6I7IWem1"}, "label": "a man in a pilot suit standing next to an airplane"}]} {"id": 61460, "image": "COCO_train2014_000000061460.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a brons shirt and stipped skirt standing next to luggage\"."}], "task": "refering", "answer_template": "The \"a person in a brons shirt and stipped skirt standing next to luggage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 212, 213, 214, 215, 218, 219, 220, 224, 225, 226, 227, 228, 229, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.029125, 0.5582311320754717, 1.0, 0.9912028301886792], "iscrowd": 0, "area": 79619.90549999998, "rle": {"size": [424, 640], "counts": "kX81V=1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O1O1MPOoCo0o;6M3N2M3N2M3M3N2M3N2M3N2M3N2M3M3N2M3N2M3N2M3N2M3O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1O100O1O100O1O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O1OVMkGc0U8ZOmGg0R8XOPHh0o7WORHj0m7UOUHk0k7SOVHn0i7QOXHP1g7oN[HQ1e7mN\\HT1c7kN_HU1`7lN_HU1a7kN^HV1a7kN_HU1`7kNaHU1_7kNaHU1^7lNbHT1]7mNcHS1]7lNdHT1[7mNeHS1Z7nNfHR1Y7oNfHR1Z7mNgHS1X7nNhHR1W7oNiHQ1W7oNiHQ1V7oNkHQ1T7POlHP1T7POlHP1S7QOmHo0R7QOoHo0Q7QOoHo0P7ROoHo0P7ROPIn0P7QOPIP1P7POoHQ1Q7oNnHR1R7mNnHT1R7kNnHV1R7jNmHW1S7hNmHY1S7fNmH[1S7dNmH]1S7bNmH_1S7`NmHa1S7^NmHc1S7\\NmHe1S7ZNmHg1S7XNmHi1T7UNlHl1U7RNkHo1V7oMjHR2W7lMiHU2W7jMiHW2X7hMhHX2Y7fMhHZ2Y7dMhH\\2Y7cMgH]2Z7aMgH_2Y7`MhH`2Y7^MhHb2Y7]MgHc2Z7[MhHd2Y7ZMhHf2X7ZMhHf2Y7XMhHh2Y7VMhHj2Y7TMgHm2Y7SMfHn2Z7QMfHP3Z7oLfHR3Z7nLeHS3[7lLdHV3]7hLcHY3]7gLbHZ3^7eLbH\\3^7cLbH^3^7aLbH`3^7`LaHa3`7]L_He3a7ZL_Hg3a7YL^Hh3b7WL^Hj3b7UL^Hl3b7SL^Hn3d7PL\\HP4f7mK[HS4h7iKYHW4]8O00001O0000001O00001O00001O00001O00001O00001O00001O00001O00001O0000001O00000000000000000000000000O100O10000O10000O100O10000O10000O10000O100O10000O1M3J6J6N200000000O10000000000O10000000000O10000000000O10000000000O1000000002N1O2N2N1O2N1O2N2N1O2N1O001O1O001O001O001O1O001O001O001O1O001O001O1O001O001O001O1O001O001O001O1O001O001O001O1O0000O1O1O1O1O1O1O1O1O1O1O1O1O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O10000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O100000000O1000000O1000000O1000000O1000000O100000000O1000000O11O00001O001O00001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O001O00001O00001O001O9Gk2UMYD"}, "label": "a person in a brons shirt and stipped skirt standing next to luggage"}]} {"id": 61460, "image": "COCO_train2014_000000061460.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hand , blue suitcase , and faded jeans\"."}], "task": "refering", "answer_template": "The \"a hand , blue suitcase , and faded jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 146, 147, 148, 149, 161, 162, 163, 164, 170, 171, 172, 184, 193, 194, 195, 216, 217, 218, 241], "bbox": [0.0, 0.0, 0.5217031249999999, 0.6760377358490566], "iscrowd": 0, "area": 62124.48285000001, "rle": {"size": [424, 640], "counts": "g3f3l5f30O100O10000O100O10000O100O10000O100O100O10000O100O10000O100O10000O100O100O10000O100O10000O100O100O10000O100O10000O100O10000O100O100O10000O100O10000O100O100O10000O100O10000O100O10000O100O100O10000O100O10000O100O100O10000O100O10000O100O10000O100O100O10000O100O10000O100O10000O100O100O10000O100O10000O100O100O10000O100O10000O010O10000O100O100O10000O100O10000O1jNlGXMT8e2QHXMP8f2THXMl7f2XHWMi7f2\\HXMd7f2`HXM`7f2dHWM]7f2hHXMX7f2lHWMU7g2oHWMQ7f2TIXMl6f2XIWMi6g2\\IVMd6g2aIVM`6h2dIVM\\6h2S2M3N2N2N2O12N3M2N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N1O000010O000000000001O02N7I6J7I6K6I6J6J7I6J7I6K6I6J7I6J7I6J7J5J7I6J7I6J7J5J7I6J7I6J7I6K6I6J7I6J7I6K6F9C>@b^n3"}, "label": "a hand , blue suitcase , and faded jeans"}]} {"id": 61460, "image": "COCO_train2014_000000061460.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing jeans\"."}], "task": "refering", "answer_template": "The \"a person wearing jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 146, 147, 148, 149, 161, 162, 163, 164, 170, 171, 172, 184, 193, 194, 195, 216, 217, 218, 241], "bbox": [0.0, 0.0, 0.5217031249999999, 0.6760377358490566], "iscrowd": 0, "area": 62124.48285000001, "rle": {"size": [424, 640], "counts": "g3f3l5f30O100O10000O100O10000O100O10000O100O100O10000O100O10000O100O10000O100O100O10000O100O10000O100O100O10000O100O10000O100O10000O100O100O10000O100O10000O100O100O10000O100O10000O100O10000O100O100O10000O100O10000O100O100O10000O100O10000O100O10000O100O100O10000O100O10000O100O10000O100O100O10000O100O10000O100O100O10000O100O10000O010O10000O100O100O10000O100O10000O1jNlGXMT8e2QHXMP8f2THXMl7f2XHWMi7f2\\HXMd7f2`HXM`7f2dHWM]7f2hHXMX7f2lHWMU7g2oHWMQ7f2TIXMl6f2XIWMi6g2\\IVMd6g2aIVM`6h2dIVM\\6h2S2M3N2N2N2O12N3M2N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N1O000010O000000000001O02N7I6J7I6K6I6J6J7I6J7I6K6I6J7I6J7I6J7J5J7I6J7I6J7J5J7I6J7I6J7I6K6I6J7I6J7I6K6F9C>@b^n3"}, "label": "a person wearing jeans"}]} {"id": 258071, "image": "COCO_train2014_000000258071.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"older woman wearing dark blue sweater\"."}], "task": "refering", "answer_template": "The \"older woman wearing dark blue sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 19, 35, 36, 37, 38, 39, 40, 41, 42, 58, 59, 60, 61, 62, 63, 64, 65, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 290, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.41685937500000003, 0.0015566037735849057, 1.0, 0.9858254716981132], "iscrowd": 0, "area": 99462.38254999998, "rle": {"size": [424, 640], "counts": "hk^32T=3M2O2O0O2N2OO010O01O010O0010O01O010O0O2O0O2O000O2O001N101O0O101N101O0O2O001N9H=C4L2O001O001O001O0O2O00001O001O001O00001N10000`HoMi3Q2fKUNTMOR7m1gKhNV4X1hKnNS4S1jKTOR4l0lKYOP4i0mK]Oo3c0oKBn3>QLGj3;ULIg3OhIUNa2n1e3LlIUN_2P2c3LPJRN]2S2b3JTJRNY2U2b3HWJRNW2W2`3GmL:R3FnL;Q3DPMl2AUM`0i2@XMa0g2_OYMb0f2VObMk0]2VObMl0\\2TOdMZ1n1gNQNf1b1[N]Ne1c1[N]Nf1b1[N]Ne1c1\\N\\Nd1d1\\N\\Nd1d1]N[Nc1e1^NZNc1e1]N[Nc1e1^NZNa1g1`NXN`1h1`NXN_1j1aNUN_1k1bNTN]1m1cNSN\\1n1eNQN]1m1dNRN^1m1aNSN`1l1aNSNa1k1`NTNb1j1^NVNb1k1^NTNc1k1^NTNc1k1]NUNc1k1^NTNc1l1]NSNd1l1\\NTNd1l1]NSNd1m1\\NRNe1m1[NSNe1m1\\NRNe1m1\\NRNe1n1ZNRNg1m1ZNRNg1m1ZNRNg1m1YNSNh1m1XNRNh1n1YNQNh1n1XNRNi1m1XNRNi1n1WNQNj1n1VNRNk1m1VNRNk1m1UNSNk1m1VNRNk1m1UNSNl1l1UNSNl1l1TNTNm1k1SNjMdMZNY4l3TNhMfMZNW4h1hK2[2LfMZNX4b1QLUOFL\\2Q1fM\\NX4\\1YLUOBOW2T1fM\\NY4V1bLTO]O3S2U1fM^NY4k0oLYOTO6n1W1gM_NX4?\\NHEZ1gM_NY4<`NEC_1eM`NY49dNABf1aM`NZ46^1Y1YJaNY44_1[1YJaNY42_1]1XJaNY40a1^1WJbNY4Mb1a1UJbNZ4Jc1c1TJcNc:\\1^EdNa:]1_EcNa:\\1`EdN`:\\1`EdN`:[1aEeN_:[1aEeN_:Z1bEfN^:Y1cEgN\\:Z1dEfN\\:Y1eEgN[:Y1eEgN[:X1fEhNZ:f0XFZOh9b0\\F^Oc9`0`F@`9=cFC]99gFGY96jFJV93mFMS9OQG1n8LVG4o:000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000lDKU95jFMU93jFOU91jF1U9OiF4V9LiF6V9JiF8V9HiF:V9FiF in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 19, 35, 36, 37, 38, 39, 40, 41, 42, 58, 59, 60, 61, 62, 63, 64, 65, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 290, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.41685937500000003, 0.0015566037735849057, 1.0, 0.9858254716981132], "iscrowd": 0, "area": 99462.38254999998, "rle": {"size": [424, 640], "counts": "hk^32T=3M2O2O0O2N2OO010O01O010O0010O01O010O0O2O0O2O000O2O001N101O0O101N101O0O2O001N9H=C4L2O001O001O001O0O2O00001O001O001O00001N10000`HoMi3Q2fKUNTMOR7m1gKhNV4X1hKnNS4S1jKTOR4l0lKYOP4i0mK]Oo3c0oKBn3>QLGj3;ULIg3OhIUNa2n1e3LlIUN_2P2c3LPJRN]2S2b3JTJRNY2U2b3HWJRNW2W2`3GmL:R3FnL;Q3DPMl2AUM`0i2@XMa0g2_OYMb0f2VObMk0]2VObMl0\\2TOdMZ1n1gNQNf1b1[N]Ne1c1[N]Nf1b1[N]Ne1c1\\N\\Nd1d1\\N\\Nd1d1]N[Nc1e1^NZNc1e1]N[Nc1e1^NZNa1g1`NXN`1h1`NXN_1j1aNUN_1k1bNTN]1m1cNSN\\1n1eNQN]1m1dNRN^1m1aNSN`1l1aNSNa1k1`NTNb1j1^NVNb1k1^NTNc1k1^NTNc1k1]NUNc1k1^NTNc1l1]NSNd1l1\\NTNd1l1]NSNd1m1\\NRNe1m1[NSNe1m1\\NRNe1m1\\NRNe1n1ZNRNg1m1ZNRNg1m1ZNRNg1m1YNSNh1m1XNRNh1n1YNQNh1n1XNRNi1m1XNRNi1n1WNQNj1n1VNRNk1m1VNRNk1m1UNSNk1m1VNRNk1m1UNSNl1l1UNSNl1l1TNTNm1k1SNjMdMZNY4l3TNhMfMZNW4h1hK2[2LfMZNX4b1QLUOFL\\2Q1fM\\NX4\\1YLUOBOW2T1fM\\NY4V1bLTO]O3S2U1fM^NY4k0oLYOTO6n1W1gM_NX4?\\NHEZ1gM_NY4<`NEC_1eM`NY49dNABf1aM`NZ46^1Y1YJaNY44_1[1YJaNY42_1]1XJaNY40a1^1WJbNY4Mb1a1UJbNZ4Jc1c1TJcNc:\\1^EdNa:]1_EcNa:\\1`EdN`:\\1`EdN`:[1aEeN_:[1aEeN_:Z1bEfN^:Y1cEgN\\:Z1dEfN\\:Y1eEgN[:Y1eEgN[:X1fEhNZ:f0XFZOh9b0\\F^Oc9`0`F@`9=cFC]99gFGY96jFJV93mFMS9OQG1n8LVG4o:000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000lDKU95jFMU93jFOU91jF1U9OiF4V9LiF6V9JiF8V9HiF:V9FiF in this image.", "objects": [{"patches": [72, 73, 74, 75, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286], "bbox": [0.044765624999999996, 0.21959905660377357, 0.45807812499999995, 0.8524764150943396], "iscrowd": 0, "area": 42201.58854999999, "rle": {"size": [424, 640], "counts": "mY<>i<>B3L2O1O2M2O1O1N3N1O1N2O1O2M2O1O1N3N1N2O1O2M2O1O1N3N1O1N2O2N1N2O1O2M2O1O1N3N1O1O1O2N1O1N011kEPMQ:Q3oEoLP:Q3oMoLRJR3i5YMPJh2XNmL^7e0TJ^2m5lMmIU2n5VNjIk1R6_NhIb1S6hNgIY1T6ROeIn0W6\\OcIe0W6FcI:Y61`I1Z6b3M3O1N2N101N2N2O1N2N100O00100O1O010O1O100O00100O10O01O100O1O010O1O10O01O100O1O010O1O10O1000000000000O1000000000000O1000000000000O1000000000000O12N1O1O1O2N1O1O2N1O1O1O2N1O1O1OO3N2N2N2N2N2N1O2N2N2M3N2N2N2N2N2N2N2N2M2O2N2N2N2N1O2N2N2N101N2N2N2iKRIQ2n6jMWIU2k6fMXIZ2j6aMYI_2i6[M\\Ie2e6VM^Ij2c6QMbIn2`6mLcIT3^6gLfIX3\\6bLhI^3Z6]LjIb3W6YLmIh3T6SLoIm3g7O0010O01O00001O001N101O001O0O101O001N101O002N1N3N2N1O2N2M2O2N2N1N3N2N1O2N2M2O2N2N1N3ROXEXOj:e0WE[Oj:d0VE\\Ol:a0UE_Om:>SEBo:=QECQ;:PEFR;7oDIR;6nDIU;4lDLV;1kDOV;0jD0X;MhD4Z;IgD6[;IeD7];FdD:^;DbD<`;AaD?`;_OaD`0b;^O^Db0d;[O\\Df0e;XO\\Dh0R in this image.", "objects": [{"patches": [72, 73, 74, 75, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 140, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 192, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286], "bbox": [0.044765624999999996, 0.21959905660377357, 0.45807812499999995, 0.8524764150943396], "iscrowd": 0, "area": 42201.58854999999, "rle": {"size": [424, 640], "counts": "mY<>i<>B3L2O1O2M2O1O1N3N1O1N2O1O2M2O1O1N3N1N2O1O2M2O1O1N3N1O1N2O2N1N2O1O2M2O1O1N3N1O1O1O2N1O1N011kEPMQ:Q3oEoLP:Q3oMoLRJR3i5YMPJh2XNmL^7e0TJ^2m5lMmIU2n5VNjIk1R6_NhIb1S6hNgIY1T6ROeIn0W6\\OcIe0W6FcI:Y61`I1Z6b3M3O1N2N101N2N2O1N2N100O00100O1O010O1O100O00100O10O01O100O1O010O1O10O01O100O1O010O1O10O1000000000000O1000000000000O1000000000000O1000000000000O12N1O1O1O2N1O1O2N1O1O1O2N1O1O1OO3N2N2N2N2N2N1O2N2N2M3N2N2N2N2N2N2N2N2M2O2N2N2N2N1O2N2N2N101N2N2N2iKRIQ2n6jMWIU2k6fMXIZ2j6aMYI_2i6[M\\Ie2e6VM^Ij2c6QMbIn2`6mLcIT3^6gLfIX3\\6bLhI^3Z6]LjIb3W6YLmIh3T6SLoIm3g7O0010O01O00001O001N101O001O0O101O001N101O002N1N3N2N1O2N2M2O2N2N1N3N2N1O2N2M2O2N2N1N3ROXEXOj:e0WE[Oj:d0VE\\Ol:a0UE_Om:>SEBo:=QECQ;:PEFR;7oDIR;6nDIU;4lDLV;1kDOV;0jD0X;MhD4Z;IgD6[;IeD7];FdD:^;DbD<`;AaD?`;_OaD`0b;^O^Db0d;[O\\Df0e;XO\\Dh0R in this image.", "objects": [{"patches": [50, 51, 52, 72, 73, 74, 75, 76, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 214, 215, 216, 231, 232, 233, 234, 235, 238, 239], "bbox": [0.07534375, 0.13014583333333332, 0.436640625, 0.6255624999999999], "iscrowd": 0, "area": 28581.010499999997, "rle": {"size": [480, 640], "counts": "nhf02k>6K5J6J6J3M3M3M3O010001O00000O2O000000001N100000001O0O1000001O000O10001O00000O101OgMi0iDWOV;Q1cDnN^;V1^DjNa;Z1]DeNb;^1\\DbNd;`1ZD_Nf;d1XD\\Nh;f1VDZNi;i1VDVNi;`210O100O0eMoLhHR3V7]M]Hc2a7kMRHV2k7UNmGj1R8_NfGb1X8hN_GY1_8QOYGo0e8[ORGf0l8g2O2N2N2O0O2N2N2N2N2N2N2O1N2N2N2O10000O100000000000000O10000000000000000O10000000000000000O11O00000\\N[HmKf7Q4YIQKg6o4_IkJa6U5e101O0000001O0O10001O00001O00000O2O00001O2N1O1O1O1O1N2O1O1O1O1O1O1O1O1N3N1O2N3M3M3M3M3gMoD=T;VOVEi0P;hN[EW1W<000000=C`0A>A5K1O1O1O100O1O1O1O1O100O1O1O1OO10001O00000000001O000000000O2O002N2N3M2N3L3N3M2N3M2N3L3N3M2N3M2N3L3N3M2N2N3M2Mi0XO5K4L5K4L4K6K4LWnX5"}, "label": "a woman in a red dress"}]} {"id": 159768, "image": "COCO_train2014_000000159768.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women wearing red colour dress\"."}], "task": "refering", "answer_template": "The \"a women wearing red colour dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 72, 73, 74, 75, 76, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 214, 215, 216, 231, 232, 233, 234, 235, 238, 239], "bbox": [0.07534375, 0.13014583333333332, 0.436640625, 0.6255624999999999], "iscrowd": 0, "area": 28581.010499999997, "rle": {"size": [480, 640], "counts": "nhf02k>6K5J6J6J3M3M3M3O010001O00000O2O000000001N100000001O0O1000001O000O10001O00000O101OgMi0iDWOV;Q1cDnN^;V1^DjNa;Z1]DeNb;^1\\DbNd;`1ZD_Nf;d1XD\\Nh;f1VDZNi;i1VDVNi;`210O100O0eMoLhHR3V7]M]Hc2a7kMRHV2k7UNmGj1R8_NfGb1X8hN_GY1_8QOYGo0e8[ORGf0l8g2O2N2N2O0O2N2N2N2N2N2N2O1N2N2N2O10000O100000000000000O10000000000000000O10000000000000000O11O00000\\N[HmKf7Q4YIQKg6o4_IkJa6U5e101O0000001O0O10001O00001O00000O2O00001O2N1O1O1O1O1N2O1O1O1O1O1O1O1O1N3N1O2N3M3M3M3M3gMoD=T;VOVEi0P;hN[EW1W<000000=C`0A>A5K1O1O1O100O1O1O1O1O100O1O1O1OO10001O00000000001O000000000O2O002N2N3M2N3L3N3M2N3M2N3L3N3M2N3M2N3L3N3M2N2N3M2Mi0XO5K4L5K4L4K6K4LWnX5"}, "label": "a women wearing red colour dress"}]} {"id": 159768, "image": "COCO_train2014_000000159768.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a green and blue striped shirt\"."}], "task": "refering", "answer_template": "The \"a woman in a green and blue striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 84, 85, 86, 87, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 247, 248, 249, 250, 251, 270, 271, 272, 273, 274, 296], "bbox": [0.586515625, 0.1415833333333333, 0.9539375, 0.7191041666666665], "iscrowd": 0, "area": 36425.12545, "rle": {"size": [480, 640], "counts": "dQ`51c>;E^FB_9b0`F^O_9c0aF]O^9d0bF\\O]9T3O1O100O1O1O1O1O1O100O1O1O1O2N1O2N100O2N1O2O001O001O001O000O2O001O001O1O9G9G9G=CO1O10O101O00001O0001O010O1O010O1O010O1O00100O1O010O1O00100O1O2O1N2N2O1N2N2N2O1N2N2O1N2N2O1N2N2N2OO00001O0000010O01O2O1N2N2O1N2N2N2O1N3M2O1N2N2N3N2M3M4M2M4L3N0O1O1O010O1O010O1O00100O00100O00100O1O00100O00100O001L4M4L3L5L3L4M4L3L4M4K4]MbD`1b;`N_20001O00001O001O00001N1L4K6K4Koj="}, "label": "a woman in a green and blue striped shirt"}]} {"id": 159768, "image": "COCO_train2014_000000159768.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a green shirt sitting at the table\"."}], "task": "refering", "answer_template": "The \"a woman in a green shirt sitting at the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 84, 85, 86, 87, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 247, 248, 249, 250, 251, 270, 271, 272, 273, 274, 296], "bbox": [0.586515625, 0.1415833333333333, 0.9539375, 0.7191041666666665], "iscrowd": 0, "area": 36425.12545, "rle": {"size": [480, 640], "counts": "dQ`51c>;E^FB_9b0`F^O_9c0aF]O^9d0bF\\O]9T3O1O100O1O1O1O1O1O100O1O1O1O2N1O2N100O2N1O2O001O001O001O000O2O001O001O1O9G9G9G=CO1O10O101O00001O0001O010O1O010O1O010O1O00100O1O010O1O00100O1O2O1N2N2O1N2N2N2O1N2N2O1N2N2O1N2N2N2OO00001O0000010O01O2O1N2N2O1N2N2N2O1N3M2O1N2N2N3N2M3M4M2M4L3N0O1O1O010O1O010O1O00100O00100O00100O1O00100O00100O001L4M4L3L5L3L4M4L3L4M4K4]MbD`1b;`N_20001O00001O001O00001N1L4K6K4Koj="}, "label": "a woman in a green shirt sitting at the table"}]} {"id": 159768, "image": "COCO_train2014_000000159768.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"clear goblet with ice in it but no wtaer\"."}], "task": "refering", "answer_template": "The \"clear goblet with ice in it but no wtaer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 222, 223, 243, 244, 245, 246, 266, 267, 268, 269, 270, 289, 290, 291, 292, 293, 312, 313, 314, 315, 336, 337, 338, 359, 360, 381, 382, 383, 384, 385], "bbox": [0.585171875, 0.5598124999999999, 0.7516875, 1.0], "iscrowd": 0, "area": 16412.117850000002, "rle": {"size": [480, 640], "counts": "\\R`5\\1W=P1POb0K5J5L5J2O2M2O2WGaLm5`3PJdLn5]3oIeLP6^3lIdLT6]3iIeLW6]3eIeL[6\\3cIeL]6\\3aIeL_6]3_IcLa6^3]IcLc6^3[IcLe6^3YIcLg6^3WIcLi6^3UIcLk6^3SIcLm6^3RIbLn6_3PIbLP7_3nHbLR7_3mHaLS7`3kHaLU7`3iHaLV7a3iH_LW7b3gH_LY7b3eH_L[7b3dH^L\\7c3bH^L^7b3aH_L_7b3`H^L`7b3`H^L`7c3^H^Lb7b3^H^Lb7c3]H]Lc7c3]H]Lc7c3]H]Lc7d3[H]Le7c3[H]Le7d3ZH\\Lf7d3ZH\\Lf7e3YH[Lg7_500000000000000000000000000000XNQH_Lo7a3^HRLb7m3_HSLa7l3`HTL`7k3bHTL^7k3cHUL]7k3cHUL]7j3dHVL\\7i3eHWL[7h3fHXLZ7g3gHYLZ7e3hHZLX7f3hHZLX7e3iH[LW7d3jH\\LV7c3kH]LU7b3lH^LT7b3mH]LS7b3nH^LS7`3nH`LR7_3PI`LP7^3RIbLo6\\3SIcLm6\\3TIdLm6Z3UIeLk6Y3WIgLj6W3WIiLi6V3YIiLh6U3YIkLg6S3\\IlLe6R3\\InLf6o2\\IPMf6j2_IUMe6c2`I\\Mi6W2]IgMl6l1YISNQ7_1UI_Ng;O2N1O1O2N1OURZ2"}, "label": "clear goblet with ice in it but no wtaer"}]} {"id": 159768, "image": "COCO_train2014_000000159768.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass with a lot of ice and a little bit of water to the right of the salt\"."}], "task": "refering", "answer_template": "The \"a glass with a lot of ice and a little bit of water to the right of the salt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 222, 223, 243, 244, 245, 246, 266, 267, 268, 269, 270, 289, 290, 291, 292, 293, 312, 313, 314, 315, 336, 337, 338, 359, 360, 381, 382, 383, 384, 385], "bbox": [0.585171875, 0.5598124999999999, 0.7516875, 1.0], "iscrowd": 0, "area": 16412.117850000002, "rle": {"size": [480, 640], "counts": "\\R`5\\1W=P1POb0K5J5L5J2O2M2O2WGaLm5`3PJdLn5]3oIeLP6^3lIdLT6]3iIeLW6]3eIeL[6\\3cIeL]6\\3aIeL_6]3_IcLa6^3]IcLc6^3[IcLe6^3YIcLg6^3WIcLi6^3UIcLk6^3SIcLm6^3RIbLn6_3PIbLP7_3nHbLR7_3mHaLS7`3kHaLU7`3iHaLV7a3iH_LW7b3gH_LY7b3eH_L[7b3dH^L\\7c3bH^L^7b3aH_L_7b3`H^L`7b3`H^L`7c3^H^Lb7b3^H^Lb7c3]H]Lc7c3]H]Lc7c3]H]Lc7d3[H]Le7c3[H]Le7d3ZH\\Lf7d3ZH\\Lf7e3YH[Lg7_500000000000000000000000000000XNQH_Lo7a3^HRLb7m3_HSLa7l3`HTL`7k3bHTL^7k3cHUL]7k3cHUL]7j3dHVL\\7i3eHWL[7h3fHXLZ7g3gHYLZ7e3hHZLX7f3hHZLX7e3iH[LW7d3jH\\LV7c3kH]LU7b3lH^LT7b3mH]LS7b3nH^LS7`3nH`LR7_3PI`LP7^3RIbLo6\\3SIcLm6\\3TIdLm6Z3UIeLk6Y3WIgLj6W3WIiLi6V3YIiLh6U3YIkLg6S3\\IlLe6R3\\InLf6o2\\IPMf6j2_IUMe6c2`I\\Mi6W2]IgMl6l1YISNQ7_1UI_Ng;O2N1O1O2N1OURZ2"}, "label": "a glass with a lot of ice and a little bit of water to the right of the salt"}]} {"id": 102432, "image": "COCO_train2014_000000102432.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with no shirt or shoes\"."}], "task": "refering", "answer_template": "The \"the man with no shirt or shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 42, 43, 59, 60, 61, 77, 78, 79, 95, 96, 97, 98, 114, 115, 116, 132, 134, 135, 150, 168], "bbox": [0.3211, 0.13029333333333334, 0.51928, 0.7524266666666667], "iscrowd": 0, "area": 8508.90555, "rle": {"size": [375, 500], "counts": "]Sk1;Q;>C=SOl0L5K5K4L5K5L3M4K5L3M4LQ1oN4L5J5L5K4K6XJkJW4K^K^53mJ\\4MZK[50nJd4\\6M2O2M2N2O010000ZNRLcKm3X4cL]K^3`3TLPLm5o3n0O1O1O1kM`HH4\\N\\7k1cHIf74^HLa71cHN]70fH0Y7IoH7Q7AWI>j6@XI`0h6]O[Ib0f6\\O]IS1T6jNnIW1Q6gNQJ[1m5cNUJ]1k5aNWJ^1j5`NXJ_1i5QNgJn1b7SO\\FEe9:^FCc9;aFB`9=cF@^9>fF@Z9`0hFXO^9i0h01O1O2N1O2N1O2N2O1O0O2O1N2O1O2M2O13J6J6J6J5K6J6J2N2N2O2M00001O01O01O003M2N2Obeg2"}, "label": "the man with no shirt or shoes"}]} {"id": 102432, "image": "COCO_train2014_000000102432.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man jumping on a beach\"."}], "task": "refering", "answer_template": "The \"a man jumping on a beach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 42, 43, 59, 60, 61, 77, 78, 79, 95, 96, 97, 98, 114, 115, 116, 132, 134, 135, 150, 168], "bbox": [0.3211, 0.13029333333333334, 0.51928, 0.7524266666666667], "iscrowd": 0, "area": 8508.90555, "rle": {"size": [375, 500], "counts": "]Sk1;Q;>C=SOl0L5K5K4L5K5L3M4K5L3M4LQ1oN4L5J5L5K4K6XJkJW4K^K^53mJ\\4MZK[50nJd4\\6M2O2M2N2O010000ZNRLcKm3X4cL]K^3`3TLPLm5o3n0O1O1O1kM`HH4\\N\\7k1cHIf74^HLa71cHN]70fH0Y7IoH7Q7AWI>j6@XI`0h6]O[Ib0f6\\O]IS1T6jNnIW1Q6gNQJ[1m5cNUJ]1k5aNWJ^1j5`NXJ_1i5QNgJn1b7SO\\FEe9:^FCc9;aFB`9=cF@^9>fF@Z9`0hFXO^9i0h01O1O2N1O2N1O2N2O1O0O2O1N2O1O2M2O13J6J6J6J5K6J6J2N2N2O2M00001O01O01O003M2N2Obeg2"}, "label": "a man jumping on a beach"}]} {"id": 389154, "image": "COCO_train2014_000000389154.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large sheep with lots of wool\"."}], "task": "refering", "answer_template": "The \"large sheep with lots of wool\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 175, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 280, 281, 282, 283, 284, 285, 286], "bbox": [0.177, 0.3343444730077121, 0.6403437500000001, 0.8996658097686376], "iscrowd": 0, "area": 41042.631, "rle": {"size": [389, 640], "counts": "gW[14i;:F;E:F;K4L5K4L5K4L4L2N2N2N101N2O1N2N2O0O2N2O1N2N2O1O001O1O1O100O1O00100O1O1O1O10O01O1O100O1O1O00100O1O1O10O01O1O1O100O1O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O1O1000000O1000O1000O100000000O100000O10O10000000O0100000000O010000000O10O100000O100O1O00100O1O100O00100`N_KVKg4i4[KUKf4i4]KUKd4i4_KVK`4k4aKSK`4m4aKQK`4o4aKoJ`4Q5aKmJ`4S5bKkJ^4U5cKiJ^4W5cKgJ^4Y5cKeJ^4[5dKbJ]4^5dK`J]4`5dK_J]4`5eK]J\\4d5dKZJ]4f5eKWJ\\4j5gKQJZ4o5n01O001O001O001O001OO1O2N00100O10O0100O10O0100O010O1000O010000O01000O010O1000O101O000O2O000O2O00001N10001O0O101O0O2O1O1O1N2N2N1O2N1O0O101O0O2O000O2O0002O1N2N2O1O001OUK_IS4`6mKaIS4]6mKeIS4Z6mKgIS4X6lKjIT4U6lKlIT4S6kKoI\\3FgLZ6MRJ[3EgLX6MUJ\\3CgLW6LXJ\\3CgLT6L[J\\3BgLS6L]J\\3BgLP6L`J\\3BfLo5LaJ^3@fLn5KeJ]3_OfLl5LiJ[3\\OiLj5JnJ[3YOjLj5IPK[3XOjLh5JTKY3UOmLf5HYKY3ROnLe5H\\KY3oNmLg5G_KZ3jNnLa7Q3_HnLc7P3^HnLe7Q3[HnLg7Q3YHnLi7P3XHnLj7^2dGSNa0_Om7\\2cGUN`0^Oo7Z2bGYN>]OQ8X2cGZNQ9c1PG^NQ9_1QGaNo8]1RGcNP9Z1RGfNo8W1RGjNn8S1UGlNm8o0VGROk8i0YGWOh8d0[G]Ok88[GFi83[GMe8O`G0`8LdG4]8GhG7Y8EkG;U8@QH?k901N102J6I[Rg2"}, "label": "large sheep with lots of wool"}]} {"id": 389154, "image": "COCO_train2014_000000389154.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the big sheep\"."}], "task": "refering", "answer_template": "The \"the big sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 174, 175, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 280, 281, 282, 283, 284, 285, 286], "bbox": [0.177, 0.3343444730077121, 0.6403437500000001, 0.8996658097686376], "iscrowd": 0, "area": 41042.631, "rle": {"size": [389, 640], "counts": "gW[14i;:F;E:F;K4L5K4L5K4L4L2N2N2N101N2O1N2N2O0O2N2O1N2N2O1O001O1O1O100O1O00100O1O1O1O10O01O1O100O1O1O00100O1O1O10O01O1O1O100O1O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O1O1000000O1000O1000O100000000O100000O10O10000000O0100000000O010000000O10O100000O100O1O00100O1O100O00100`N_KVKg4i4[KUKf4i4]KUKd4i4_KVK`4k4aKSK`4m4aKQK`4o4aKoJ`4Q5aKmJ`4S5bKkJ^4U5cKiJ^4W5cKgJ^4Y5cKeJ^4[5dKbJ]4^5dK`J]4`5dK_J]4`5eK]J\\4d5dKZJ]4f5eKWJ\\4j5gKQJZ4o5n01O001O001O001O001OO1O2N00100O10O0100O10O0100O010O1000O010000O01000O010O1000O101O000O2O000O2O00001N10001O0O101O0O2O1O1O1N2N2N1O2N1O0O101O0O2O000O2O0002O1N2N2O1O001OUK_IS4`6mKaIS4]6mKeIS4Z6mKgIS4X6lKjIT4U6lKlIT4S6kKoI\\3FgLZ6MRJ[3EgLX6MUJ\\3CgLW6LXJ\\3CgLT6L[J\\3BgLS6L]J\\3BgLP6L`J\\3BfLo5LaJ^3@fLn5KeJ]3_OfLl5LiJ[3\\OiLj5JnJ[3YOjLj5IPK[3XOjLh5JTKY3UOmLf5HYKY3ROnLe5H\\KY3oNmLg5G_KZ3jNnLa7Q3_HnLc7P3^HnLe7Q3[HnLg7Q3YHnLi7P3XHnLj7^2dGSNa0_Om7\\2cGUN`0^Oo7Z2bGYN>]OQ8X2cGZNQ9c1PG^NQ9_1QGaNo8]1RGcNP9Z1RGfNo8W1RGjNn8S1UGlNm8o0VGROk8i0YGWOh8d0[G]Ok88[GFi83[GMe8O`G0`8LdG4]8GhG7Y8EkG;U8@QH?k901N102J6I[Rg2"}, "label": "the big sheep"}]} {"id": 389154, "image": "COCO_train2014_000000389154.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young lamb resting in the grass next to an older sheep\"."}], "task": "refering", "answer_template": "The \"a young lamb resting in the grass next to an older sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [180, 182, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 267, 268, 269], "bbox": [0.631234375, 0.5205398457583548, 0.947515625, 0.8117737789203086], "iscrowd": 0, "area": 10880.158699999994, "rle": {"size": [389, 640], "counts": "[hi43o;8I5K5K2N2gD[OP;n0N1N3N2N2N2O100O010O100O100O10O0100O100O100O10O01O10000O1000O01000000O1000O01000000O1000O1000O10000000O0100000000O10O1000000000000O100000000000000O1000000000000O10O1000000000O100000000000O01000000000000O100000O100000O1000000000O1000OVO^NSGb1l8aNRG_1l8dNSG\\1l8gNRGY1m8iNRGW1l8mNQGT1n8nNQGQ1Q9POmFP1S9QOlFo0T9SOjFm0V9SOjFm0V9TOiFl0W9TOiFl0X9SOgFm0Z9TOeFl0[9TOdFm0\\9SOcFn0]9SObFm0^9SOaFn0`9RO_Fn0a9RO^Fn0d9QO[FP1f9g01O1O1O1O1O2N1O1N100000O10000O10000O10000O10000O10O10O10000O01N2M2O2N2M2O1N2O1O1N2DjEiNV:i0j0O100O100O1O1O2N1O2O1O3M5K6Jef<"}, "label": "a young lamb resting in the grass next to an older sheep"}]} {"id": 397362, "image": "COCO_train2014_000000397362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a leather love seat behind a table\"."}], "task": "refering", "answer_template": "The \"a leather love seat behind a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 141, 155, 157, 158, 159, 175, 176, 177], "bbox": [0.62982, 0.6124107142857144, 0.87186, 0.8317857142857144], "iscrowd": 0, "area": 3930.906250000001, "rle": {"size": [336, 500], "counts": "jbW33[:5K5M3M0000O100000000000O0OI62O001O1O001OGVFOj92900000001O010O10000O10001O0001O002N2N2N1OO101N11O1O001O1O00100O2N1O2N1O2N2`F[OR9e0jF^OV9d0fF^OZ9c0cF_O]9k0O0000001O0000001hFB\\8>bGD^8W1000000O100000000000000000000O100000000000002N2N1N1000O10001N1000000O2]Oc05L4M3LJ6J5K6K5J5K7I7J6Ijhd0"}, "label": "a leather love seat behind a table"}]} {"id": 397362, "image": "COCO_train2014_000000397362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch sitting in front of the window\"."}], "task": "refering", "answer_template": "The \"the couch sitting in front of the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 141, 155, 157, 158, 159, 175, 176, 177], "bbox": [0.62982, 0.6124107142857144, 0.87186, 0.8317857142857144], "iscrowd": 0, "area": 3930.906250000001, "rle": {"size": [336, 500], "counts": "jbW33[:5K5M3M0000O100000000000O0OI62O001O1O001OGVFOj92900000001O010O10000O10001O0001O002N2N2N1OO101N11O1O001O1O00100O2N1O2N1O2N2`F[OR9e0jF^OV9d0fF^OZ9c0cF_O]9k0O0000001O0000001hFB\\8>bGD^8W1000000O100000000000000000000O100000000000002N2N1N1000O10001N1000000O2]Oc05L4M3LJ6J5K6K5J5K7I7J6Ijhd0"}, "label": "the couch sitting in front of the window"}]} {"id": 176179, "image": "COCO_train2014_000000176179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant lying in the river water with her caretaker standing on it forelimb\"."}], "task": "refering", "answer_template": "The \"an elephant lying in the river water with her caretaker standing on it forelimb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 53, 54, 55, 56, 57, 58, 75, 76, 77, 78, 79, 80, 81, 82, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 302, 305, 306, 307, 308, 309, 310, 311, 312, 329, 330], "bbox": [0.13314062499999998, 0.08089583333333333, 0.864609375, 0.8539375], "iscrowd": 0, "area": 85544.94870000001, "rle": {"size": [480, 640], "counts": "PVX13h>6N101N2N2O0O2N2N2O0O2N2O1N1O2N2O1N1O2O1O1O001O10O0N2O2N1O1O1O2N1O101O0O10001N1000[OZOQCd0P=^OnBb0Q=@nB`0R=AmB>T=ClB in this image.", "objects": [{"patches": [31, 32, 33, 34, 53, 54, 55, 56, 57, 58, 75, 76, 77, 78, 79, 80, 81, 82, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 302, 305, 306, 307, 308, 309, 310, 311, 312, 329, 330], "bbox": [0.13314062499999998, 0.08089583333333333, 0.864609375, 0.8539375], "iscrowd": 0, "area": 85544.94870000001, "rle": {"size": [480, 640], "counts": "PVX13h>6N101N2N2O0O2N2N2O0O2N2O1N1O2N2O1N1O2O1O1O001O10O0N2O2N1O1O1O2N1O101O0O10001N1000[OZOQCd0P=^OnBb0Q=@nB`0R=AmB>T=ClB in this image.", "objects": [{"patches": [51, 52, 73, 74, 75, 96, 97, 98, 99, 119, 120, 121, 122, 123, 142, 143, 165, 166, 188, 189, 211, 234, 235, 256, 257], "bbox": [0.160875, 0.121375, 0.393203125, 0.71025], "iscrowd": 0, "area": 11063.73345, "rle": {"size": [480, 640], "counts": "Rc`15i>6J3L3N2N2M4oK7dHNk6o0]H]Ob7d0kGOS82kG2R80kG3T8NjG4T8NjG5T8KkG7S8LhG9T8JhG:V8HfG00000000000000000001O00000O1000000000000000000000000O2O00O1O1O1O1O100O10001N1004L00000FaA3_>MaA3`>KaA6^>JcA5^>IcA7c>O2O0O3NnUf5"}, "label": "a man in a red cloth washing an elephant"}]} {"id": 176179, "image": "COCO_train2014_000000176179.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in red shorts scrubbing an elephant ' s trunk\"."}], "task": "refering", "answer_template": "The \"a man in red shorts scrubbing an elephant ' s trunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 73, 74, 75, 96, 97, 98, 99, 119, 120, 121, 122, 123, 142, 143, 165, 166, 188, 189, 211, 234, 235, 256, 257], "bbox": [0.160875, 0.121375, 0.393203125, 0.71025], "iscrowd": 0, "area": 11063.73345, "rle": {"size": [480, 640], "counts": "Rc`15i>6J3L3N2N2M4oK7dHNk6o0]H]Ob7d0kGOS82kG2R80kG3T8NjG4T8NjG5T8KkG7S8LhG9T8JhG:V8HfG00000000000000000001O00000O1000000000000000000000000O2O00O1O1O1O1O100O10001N1004L00000FaA3_>MaA3`>KaA6^>JcA5^>IcA7c>O2O0O3NnUf5"}, "label": "a man in red shorts scrubbing an elephant ' s trunk"}]} {"id": 553021, "image": "COCO_train2014_000000553021.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wooden chair to the left of the other chair\"."}], "task": "refering", "answer_template": "The \"the wooden chair to the left of the other chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 253, 254, 255, 276, 277, 278, 299, 300, 301, 302, 322, 324, 325, 347, 348, 370, 371], "bbox": [0.001953125, 0.5898541666666667, 0.13671875, 0.9856874999999999], "iscrowd": 0, "area": 9621.875, "rle": {"size": [480, 640], "counts": "Qh0g2Y<\\1dN2NN2N2N2N2N2N2N2N1O2N2N2N2N2N2N2M3N2N2N2000O100000001O0000000000000O100000000000XOTD\\Nl;d1UD[Nk;e1VDZNj;e1XDYNi;g1XDVNj;j1WDRNl;m1WDnMk;S2VDiMm;W2>O10000O10000O10000O01000O1b1^NV2jM000000000000000O1000000 in this image.", "objects": [{"patches": [230, 231, 232, 253, 254, 255, 276, 277, 278, 299, 300, 301, 302, 322, 324, 325, 347, 348, 370, 371], "bbox": [0.001953125, 0.5898541666666667, 0.13671875, 0.9856874999999999], "iscrowd": 0, "area": 9621.875, "rle": {"size": [480, 640], "counts": "Qh0g2Y<\\1dN2NN2N2N2N2N2N2N2N1O2N2N2N2N2N2N2M3N2N2N2000O100000001O0000000000000O100000000000XOTD\\Nl;d1UD[Nk;e1VDZNj;e1XDYNi;g1XDVNj;j1WDRNl;m1WDnMk;S2VDiMm;W2>O10000O10000O10000O01000O1b1^NV2jM000000000000000O1000000 in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 53, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 169, 170, 171, 172, 173, 174, 179, 192, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 241, 242, 264, 265, 286, 287, 288, 289, 309, 310, 311, 312, 332, 333, 334, 335, 355, 356, 357, 358, 378, 379, 380, 381], "bbox": [0.309421875, 0.0, 0.8340781249999999, 0.9887916666666666], "iscrowd": 0, "area": 74344.6742, "rle": {"size": [480, 640], "counts": "Tnl24f>6I7J6I7J6lLmNTHY1m7hNlG^1T8bNfGd1Z8]N^Gj1b8WN\\Gj1d8WN[Gi1e8XN[Gg1e8YN[Gg1e8ZN[Ge1e8\\NZGd1f8]NZGb1f8^N[Ga1e8`NZG`1f8aNZG^1f8cNYG]1g8dNYG[1g8eNYG[1P1ZNc4<]JY1o0]Nc4;^JW1n0`Nc4:^JV1n0bNc49_JT1m0eNc47`JT1l0gNc46aJR1k0iNd46aJP1i0mNe44aJo0i0oNe42cJn0g0ROe41cJm0\\O`NMd0S60dJk0[OcNLd0T6OdJj0\\OdNJd0V6OdJh0[OgNHe0X6LfJg0YOkNFd0Z6KfJf0ZOlNDe0[6JgJd0YOoNCe0\\6IgJc0YOQOBe0]6HhJa0YOSO@f0^6FiJa0XOUO_Oe0`6FiJ?WOXO]Of0b6DjJ=WOZO[Og0c6CjJb01\\8^OUG`0`03Z8YOZGd0<3Z8VO]Gf0;3X8TO_Gj092X8POcGm063W8mNfGP133W8jNhGS132U8hNkGU113T8dNoGY1M4S8`NSH[1K5R8]NUH_1J3Q8ZNYHb1G4P8WN\\He1D4P8TN^Hh1C4\\9LeF3[9LfF4Z9LfF5X9KiF5W9JkF5U9KkF5T9KmF5S9KmF5S9JoF5P9LPG4P9KQG6n8JRG6n8ISG7l8JUG5k8JVG6j8JVG6i8JXG6h8IZG6f8JZG7e8H\\G8c8I]G7c8H_G7a8I_G7`8IaG7_8IaG7_8HcG7]8IcG8[8HfG8Z8HfG8Z8GhG8W8HjG8V8HjG8V8GkG9U8GlG8V8EkGX8_OiGa0Y8]OhGb0Z8[OgGe0[8YOeGg0]8VOdGj0^8SOdGj0`8TO`Gh0f8UO[Gg0k8WOUGe0Q9XOQGb0V9\\OjF`0\\9]OeF?a9_O_F=g9@[F;k9CUF8R:EoE7W:GiE5d in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 53, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 169, 170, 171, 172, 173, 174, 179, 192, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 241, 242, 264, 265, 286, 287, 288, 289, 309, 310, 311, 312, 332, 333, 334, 335, 355, 356, 357, 358, 378, 379, 380, 381], "bbox": [0.309421875, 0.0, 0.8340781249999999, 0.9887916666666666], "iscrowd": 0, "area": 74344.6742, "rle": {"size": [480, 640], "counts": "Tnl24f>6I7J6I7J6lLmNTHY1m7hNlG^1T8bNfGd1Z8]N^Gj1b8WN\\Gj1d8WN[Gi1e8XN[Gg1e8YN[Gg1e8ZN[Ge1e8\\NZGd1f8]NZGb1f8^N[Ga1e8`NZG`1f8aNZG^1f8cNYG]1g8dNYG[1g8eNYG[1P1ZNc4<]JY1o0]Nc4;^JW1n0`Nc4:^JV1n0bNc49_JT1m0eNc47`JT1l0gNc46aJR1k0iNd46aJP1i0mNe44aJo0i0oNe42cJn0g0ROe41cJm0\\O`NMd0S60dJk0[OcNLd0T6OdJj0\\OdNJd0V6OdJh0[OgNHe0X6LfJg0YOkNFd0Z6KfJf0ZOlNDe0[6JgJd0YOoNCe0\\6IgJc0YOQOBe0]6HhJa0YOSO@f0^6FiJa0XOUO_Oe0`6FiJ?WOXO]Of0b6DjJ=WOZO[Og0c6CjJb01\\8^OUG`0`03Z8YOZGd0<3Z8VO]Gf0;3X8TO_Gj092X8POcGm063W8mNfGP133W8jNhGS132U8hNkGU113T8dNoGY1M4S8`NSH[1K5R8]NUH_1J3Q8ZNYHb1G4P8WN\\He1D4P8TN^Hh1C4\\9LeF3[9LfF4Z9LfF5X9KiF5W9JkF5U9KkF5T9KmF5S9KmF5S9JoF5P9LPG4P9KQG6n8JRG6n8ISG7l8JUG5k8JVG6j8JVG6i8JXG6h8IZG6f8JZG7e8H\\G8c8I]G7c8H_G7a8I_G7`8IaG7_8IaG7_8HcG7]8IcG8[8HfG8Z8HfG8Z8GhG8W8HjG8V8HjG8V8GkG9U8GlG8V8EkGX8_OiGa0Y8]OhGb0Z8[OgGe0[8YOeGg0]8VOdGj0^8SOdGj0`8TO`Gh0f8UO[Gg0k8WOUGe0Q9XOQGb0V9\\OjF`0\\9]OeF?a9_O_F=g9@[F;k9CUF8R:EoE7W:GiE5d in this image.", "objects": [{"patches": [22, 23, 33, 34, 35, 45, 46, 47, 57, 58, 59, 68, 69, 70, 71, 80, 81, 82, 83, 92, 93, 94, 95, 104, 105, 106, 107, 117, 118, 119, 131, 143], "bbox": [0.6927327327327327, 0.06714, 1.0, 0.6666599999999999], "iscrowd": 0, "area": 19915.817049999998, "rle": {"size": [500, 333], "counts": "ePa37U?9G9G8I8G9G8H9H8G8I8L4L3L5L4K5L3L5L4K4M4K5L3L5L4K4M4K5L4K4M4L4K6K4K6L3N2N3M2N2N3M2N3M1O1O1O1O001O10000O0100000O10O10O1H8I7N11000O1000O100000O10O1000000O10O100000O1003M5L3L5L3L5L3M4K4M4L3L5L3M4K10N2N101NdN"}, "label": "a man playing the wii with his hands up"}]} {"id": 323664, "image": "COCO_train2014_000000323664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a wii remote above his head\"."}], "task": "refering", "answer_template": "The \"a man holding a wii remote above his head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 23, 33, 34, 35, 45, 46, 47, 57, 58, 59, 68, 69, 70, 71, 80, 81, 82, 83, 92, 93, 94, 95, 104, 105, 106, 107, 117, 118, 119, 131, 143], "bbox": [0.6927327327327327, 0.06714, 1.0, 0.6666599999999999], "iscrowd": 0, "area": 19915.817049999998, "rle": {"size": [500, 333], "counts": "ePa37U?9G9G8I8G9G8H9H8G8I8L4L3L5L4K5L3L5L4K4M4K5L3L5L4K4M4K5L4K4M4L4K6K4K6L3N2N3M2N2N3M2N3M1O1O1O1O001O10000O0100000O10O10O1H8I7N11000O1000O100000O10O1000000O10O100000O1003M5L3L5L3L5L3M4K4M4L3L5L3M4K10N2N101NdN"}, "label": "a man holding a wii remote above his head"}]} {"id": 323664, "image": "COCO_train2014_000000323664.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman in the tank top playing wii\"."}], "task": "refering", "answer_template": "The \"the woman in the tank top playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 112, 113, 114, 115, 124, 125, 126, 127, 135, 136, 137, 138, 146, 147, 148, 149, 150, 151, 158, 159, 160, 161, 162, 163, 164, 170, 171, 172, 173, 174, 175, 176, 182, 183, 184, 185, 186, 187, 194, 195, 196, 197, 198, 199, 205, 206, 207, 208, 209, 210, 211], "bbox": [0.14393393393393394, 0.48186, 0.6807207207207208, 0.98446], "iscrowd": 0, "area": 30920.87910000001, "rle": {"size": [500, 333], "counts": "Zmg02^?4L4L4L4K5L4L4L4L4L3M4L4L4K5L4L4L4K5L4L4K5L4H8^Ob0L3O2N2N2N2N2N2N2N2M3N2N2N2N2N2E:01O1O1O1O1O1N2O1O100O1O100O00100O10000000000000000000000WNgJWIX5i6oJQIQ5o6UKkHk4T7\\KfHd4Y7cKaH]4_7iKZHX4e7oKUHQ4k7TLPHl3o7ULPHl3o7ULQHk3o7TLRHl3m7ULRHl3n7TLRHl3m7[LmGe3R8o10O0100O10000O100O10000O100000000000O10000000O10000000000000000001O0O100000001O00000000001O00000000001N100fJ^Hb2b7]M_Hc2a7]M`Hb2`7]MaHc2`7\\M`He2_7ZMcHe2]7ZMdHf2\\7ZMdHf2]7XMeHg2]7WMcHj2]7TMdHl2^7QMcHo2_7oLbHP3_7nLbHR3`7lL`HU3`7iLbHV3`7gLaHY3a7eL_H[3b7cL`H\\3b7bL^H_3c7^L^Hb3c7]L]Hc3d7[L^Hd3d7YL]Hg3d7XL\\Hi3d7UL^Hj3d7nKbHR4_7eKiH[4X910O0XOWEeLi:V3aEeL`:U3kEeLU:U3VFgLi9T3_1J5K5J6KX1hN in this image.", "objects": [{"patches": [100, 101, 102, 112, 113, 114, 115, 124, 125, 126, 127, 135, 136, 137, 138, 146, 147, 148, 149, 150, 151, 158, 159, 160, 161, 162, 163, 164, 170, 171, 172, 173, 174, 175, 176, 182, 183, 184, 185, 186, 187, 194, 195, 196, 197, 198, 199, 205, 206, 207, 208, 209, 210, 211], "bbox": [0.14393393393393394, 0.48186, 0.6807207207207208, 0.98446], "iscrowd": 0, "area": 30920.87910000001, "rle": {"size": [500, 333], "counts": "Zmg02^?4L4L4L4K5L4L4L4L4L3M4L4L4K5L4L4L4K5L4L4K5L4H8^Ob0L3O2N2N2N2N2N2N2N2M3N2N2N2N2N2E:01O1O1O1O1O1N2O1O100O1O100O00100O10000000000000000000000WNgJWIX5i6oJQIQ5o6UKkHk4T7\\KfHd4Y7cKaH]4_7iKZHX4e7oKUHQ4k7TLPHl3o7ULPHl3o7ULQHk3o7TLRHl3m7ULRHl3n7TLRHl3m7[LmGe3R8o10O0100O10000O100O10000O100000000000O10000000O10000000000000000001O0O100000001O00000000001O00000000001N100fJ^Hb2b7]M_Hc2a7]M`Hb2`7]MaHc2`7\\M`He2_7ZMcHe2]7ZMdHf2\\7ZMdHf2]7XMeHg2]7WMcHj2]7TMdHl2^7QMcHo2_7oLbHP3_7nLbHR3`7lL`HU3`7iLbHV3`7gLaHY3a7eL_H[3b7cL`H\\3b7bL^H_3c7^L^Hb3c7]L]Hc3d7[L^Hd3d7YL]Hg3d7XL\\Hi3d7UL^Hj3d7nKbHR4_7eKiH[4X910O0XOWEeLi:V3aEeL`:U3kEeLU:U3VFgLi9T3_1J5K5J6KX1hN in this image.", "objects": [{"patches": [130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226], "bbox": [0.632859375, 0.3028958333333333, 0.8728281250000001, 0.5613125], "iscrowd": 0, "area": 16661.716750000007, "rle": {"size": [480, 640], "counts": "USn56f>4K5L4K5L4K5L5J5L4K5L4K5K5L4K5L4K5L4K5L4K5L4K5L4N200O1000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000O100000001O000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000A?[Oe0[Od0[Of0[Oe0ZOUkU1"}, "label": "a childs book on a bench"}]} {"id": 94300, "image": "COCO_train2014_000000094300.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tangerine bear hard back book\"."}], "task": "refering", "answer_template": "The \"a tangerine bear hard back book\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 221, 222, 223, 224, 225, 226], "bbox": [0.632859375, 0.3028958333333333, 0.8728281250000001, 0.5613125], "iscrowd": 0, "area": 16661.716750000007, "rle": {"size": [480, 640], "counts": "USn56f>4K5L4K5L4K5L5J5L4K5L4K5K5L4K5L4K5L4K5L4K5L4K5L4N200O1000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000O100000001O000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000A?[Oe0[Od0[Of0[Oe0ZOUkU1"}, "label": "a tangerine bear hard back book"}]} {"id": 94300, "image": "COCO_train2014_000000094300.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the small portion of a book visible underneath \" tangerine bear \"\"."}], "task": "refering", "answer_template": "The \"the small portion of a book visible underneath \" tangerine bear \"\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 180, 181, 203, 204, 222, 223, 224, 225, 226, 247, 248, 249], "bbox": [0.658625, 0.34270833333333334, 0.8803906250000001, 0.6019583333333334], "iscrowd": 0, "area": 2676.0699500000032, "rle": {"size": [480, 640], "counts": "eRV64g>5L4O1O100000000000000000001O0000000000000001O00000000000001O0000000000000000000000000000001O0001O000000000000000000000001O00000000000000000000000001O0001O0000000000000000000000000000001O00000000000001O0000000000000001O0000000000000000000000000000001O01O0000F:^Ob0^Oa0_O4K5L49G>B>C=B>BYnS1"}, "label": "the small portion of a book visible underneath \" tangerine bear \""}]} {"id": 94300, "image": "COCO_train2014_000000094300.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the book under the red book\"."}], "task": "refering", "answer_template": "The \"the book under the red book\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 180, 181, 203, 204, 222, 223, 224, 225, 226, 247, 248, 249], "bbox": [0.658625, 0.34270833333333334, 0.8803906250000001, 0.6019583333333334], "iscrowd": 0, "area": 2676.0699500000032, "rle": {"size": [480, 640], "counts": "eRV64g>5L4O1O100000000000000000001O0000000000000001O00000000000001O0000000000000000000000000000001O0001O000000000000000000000001O00000000000000000000000001O0001O0000000000000000000000000000001O00000000000001O0000000000000001O0000000000000000000000000000001O01O0000F:^Ob0^Oa0_O4K5L49G>B>C=B>BYnS1"}, "label": "the book under the red book"}]} {"id": 471136, "image": "COCO_train2014_000000471136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a blue hat\"."}], "task": "refering", "answer_template": "The \"a man wearing a blue hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344], "bbox": [0.758078125, 0.2067529411764706, 1.0, 0.9820235294117647], "iscrowd": 0, "area": 36855.39200000001, "rle": {"size": [425, 640], "counts": "g]Y61X=1O2M3N1O2M3N1O1O1N10001N10001N10001O0O100YD[Oj:e0UE\\Oj:e0TE]Ol:c0SE^Om:b0RE@l:a0SE@m:`0REAn:?QEBo:>PEDd6DSMi0XLDe4HbLMa2g0WLEb4O`LGf2f0WLE`45]LBk2e0WLE]4;\\L]Oo2d0WLFY4a0[LWOT3c0WLF6AY3W1PMQOY3b0WLF6ER3Z1RMkN_3`0VLJ3Eo2^1TMdNc3`0VLL1Gj2a1VM^Nh3?SL12Fe2d1XMXNm3>QL52F`2[2\\1VOQLdN[2Z2c19VNKi16TNMk14SNNk14RNOm12QN0n12nM1Q20mM2R2OkM4T2MjM5U2LhM7W2JgM8X2IfM9Y2HdM;[2FcM<\\2EaM>^2D_M>`2C]M`0b2A\\Ma0c2@ZMc0d2_O[Mb0d2_O[Mb0d2_O\\Ma0c2@\\Ma0d2_O[Mb0d2_O\\Ma0d2@ZMa0f2_OZMa0f2_OYMb0g2^OYMb0g2^OXMc0h2]OXMc0g2^OXMc0h2]OXMc0h2]OWMd0i2\\OVMe0j2\\OUMd0k2\\OTMe0l2\\OSMd0l2m4000000000000O1000000000000O100000000000000O1000000000000O100000000001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O]2cMl4TK\\E"}, "label": "a man wearing a blue hat"}]} {"id": 471136, "image": "COCO_train2014_000000471136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zoo employee in blue shirt and blue hat\"."}], "task": "refering", "answer_template": "The \"zoo employee in blue shirt and blue hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344], "bbox": [0.758078125, 0.2067529411764706, 1.0, 0.9820235294117647], "iscrowd": 0, "area": 36855.39200000001, "rle": {"size": [425, 640], "counts": "g]Y61X=1O2M3N1O2M3N1O1O1N10001N10001N10001O0O100YD[Oj:e0UE\\Oj:e0TE]Ol:c0SE^Om:b0RE@l:a0SE@m:`0REAn:?QEBo:>PEDd6DSMi0XLDe4HbLMa2g0WLEb4O`LGf2f0WLE`45]LBk2e0WLE]4;\\L]Oo2d0WLFY4a0[LWOT3c0WLF6AY3W1PMQOY3b0WLF6ER3Z1RMkN_3`0VLJ3Eo2^1TMdNc3`0VLL1Gj2a1VM^Nh3?SL12Fe2d1XMXNm3>QL52F`2[2\\1VOQLdN[2Z2c19VNKi16TNMk14SNNk14RNOm12QN0n12nM1Q20mM2R2OkM4T2MjM5U2LhM7W2JgM8X2IfM9Y2HdM;[2FcM<\\2EaM>^2D_M>`2C]M`0b2A\\Ma0c2@ZMc0d2_O[Mb0d2_O[Mb0d2_O\\Ma0c2@\\Ma0d2_O[Mb0d2_O\\Ma0d2@ZMa0f2_OZMa0f2_OYMb0g2^OYMb0g2^OXMc0h2]OXMc0g2^OXMc0h2]OXMc0h2]OWMd0i2\\OVMe0j2\\OUMd0k2\\OTMe0l2\\OSMd0l2m4000000000000O1000000000000O100000000000000O1000000000000O100000000001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O]2cMl4TK\\E"}, "label": "zoo employee in blue shirt and blue hat"}]} {"id": 471136, "image": "COCO_train2014_000000471136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white shirt and jeans smiling\"."}], "task": "refering", "answer_template": "The \"a man in a white shirt and jeans smiling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 186, 187, 208, 209, 210, 211, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 300, 301, 302, 303, 304, 323, 324, 325, 326, 327], "bbox": [0.031484374999999995, 0.5412, 0.251828125, 0.9858823529411764], "iscrowd": 0, "area": 17893.73635, "rle": {"size": [425, 640], "counts": "gd81U=3N2M3N2M4M2M4M2M4L4M2M4M2M7K4Ke1[N2O1N2N2O1N2O100O1O1YN_LeIb3Z6bLbI_3]6fL^I[3a6iL[IX3c6mLYIT3f6oLWIR3h6QMUIP3i6SMUIn2j6TMTIm2k6UMSIl2l6WMQIj2o6WMoHj2Q7WMmHj2R7XMlHi2T7YMiHh2W7YMgHh2X7ZMfHg2Z7[McHf2\\7\\MbHe2^7c100O10000O100O1000000001O000000001O00003M2N2N3M2N3M2N2N3M2N3M2N3M3M3M3M0000000000000000000000000000000000000000000000001O001O1O1O001O1O001O1O1O001O1O001O2N2N3M2N3M2ZOfFYM]9b0dFe09eNU9d0mF68ROn8e0TGI8^Of8g0]GZOR:c0\\1M4M2N3M2M3N3M^_V6"}, "label": "a man in a white shirt and jeans smiling"}]} {"id": 471136, "image": "COCO_train2014_000000471136.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smiling man in white\"."}], "task": "refering", "answer_template": "The \"the smiling man in white\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 186, 187, 208, 209, 210, 211, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 300, 301, 302, 303, 304, 323, 324, 325, 326, 327], "bbox": [0.031484374999999995, 0.5412, 0.251828125, 0.9858823529411764], "iscrowd": 0, "area": 17893.73635, "rle": {"size": [425, 640], "counts": "gd81U=3N2M3N2M4M2M4M2M4L4M2M4M2M7K4Ke1[N2O1N2N2O1N2O100O1O1YN_LeIb3Z6bLbI_3]6fL^I[3a6iL[IX3c6mLYIT3f6oLWIR3h6QMUIP3i6SMUIn2j6TMTIm2k6UMSIl2l6WMQIj2o6WMoHj2Q7WMmHj2R7XMlHi2T7YMiHh2W7YMgHh2X7ZMfHg2Z7[McHf2\\7\\MbHe2^7c100O10000O100O1000000001O000000001O00003M2N2N3M2N3M2N2N3M2N3M2N3M3M3M3M0000000000000000000000000000000000000000000000001O001O1O1O001O1O001O1O1O001O1O001O2N2N3M2N3M2ZOfFYM]9b0dFe09eNU9d0mF68ROn8e0TGI8^Of8g0]GZOR:c0\\1M4M2N3M2M3N3M^_V6"}, "label": "the smiling man in white"}]} {"id": 176229, "image": "COCO_train2014_000000176229.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wear gray jacket walking in the park\"."}], "task": "refering", "answer_template": "The \"a man wear gray jacket walking in the park\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 33, 34, 35, 56, 57, 58, 79, 80, 81, 102, 103, 104, 125, 126, 127, 148, 149, 150, 173, 195, 196, 197, 219, 220], "bbox": [0.46323437500000003, 0.02363849765258216, 0.59665625, 0.6407746478873239], "iscrowd": 0, "area": 9342.199299999995, "rle": {"size": [426, 640], "counts": "^bk34c0Ma11X85QFMa1=T8JTFMa1i0R8HhGg0o7ZOiGV1I[Ng6`0cHVO;a3R7YM_H[O=]3S7YM[H@`0X3U7hMhHZ2X7gMfHZ2Y7hMeHY2[7hMbHZ2]7hMaHY2_7gM`HY2a7hM\\HZ2c7hM[HY2e7f1O10000O100O10000O1004K4M3L4L5L3LkHRKS6j4mI[KQ6b4oIbKP6[4PJgKP6X4PJiKP6U4PJmKP6Q4PJQLP6m3PJULP6i3QJXLo5g3PJ[LP6c3PJ_LP6^3QJdLo5X3TJiLn5R3SJPMn5k2TJWMm5e2UJ\\Mm5_2WJ`Mj5[2[JdMf5Z2]JdMe5X2^JgMc5V2aJhM`5U2cJkM^5Q2fJmM[5P2hJoMY5n1jJQNX5k1jJUNW5h1kJXN_5Y1fJgNm5>[JBY6BoI>_901O00000000000000000N2O1O1N2O2M2O1N2O1O1N2Oe_[3"}, "label": "a man wear gray jacket walking in the park"}]} {"id": 176229, "image": "COCO_train2014_000000176229.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man and a woman are walking on the walkway\"."}], "task": "refering", "answer_template": "The \"a man and a woman are walking on the walkway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 33, 34, 35, 56, 57, 58, 79, 80, 81, 102, 103, 104, 125, 126, 127, 148, 149, 150, 173, 195, 196, 197, 219, 220], "bbox": [0.46323437500000003, 0.02363849765258216, 0.59665625, 0.6407746478873239], "iscrowd": 0, "area": 9342.199299999995, "rle": {"size": [426, 640], "counts": "^bk34c0Ma11X85QFMa1=T8JTFMa1i0R8HhGg0o7ZOiGV1I[Ng6`0cHVO;a3R7YM_H[O=]3S7YM[H@`0X3U7hMhHZ2X7gMfHZ2Y7hMeHY2[7hMbHZ2]7hMaHY2_7gM`HY2a7hM\\HZ2c7hM[HY2e7f1O10000O100O10000O1004K4M3L4L5L3LkHRKS6j4mI[KQ6b4oIbKP6[4PJgKP6X4PJiKP6U4PJmKP6Q4PJQLP6m3PJULP6i3QJXLo5g3PJ[LP6c3PJ_LP6^3QJdLo5X3TJiLn5R3SJPMn5k2TJWMm5e2UJ\\Mm5_2WJ`Mj5[2[JdMf5Z2]JdMe5X2^JgMc5V2aJhM`5U2cJkM^5Q2fJmM[5P2hJoMY5n1jJQNX5k1jJUNW5h1kJXN_5Y1fJgNm5>[JBY6BoI>_901O00000000000000000N2O1O1N2O2M2O1N2O1O1N2Oe_[3"}, "label": "a man and a woman are walking on the walkway"}]} {"id": 176229, "image": "COCO_train2014_000000176229.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on skateboard\"."}], "task": "refering", "answer_template": "The \"man on skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 85, 86, 87, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269], "bbox": [0.49548437500000003, 0.15694835680751174, 0.823828125, 0.775774647887324], "iscrowd": 0, "area": 25917.354099999997, "rle": {"size": [426, 640], "counts": "UbT43V=3N2N4L5K5K3L3N1O2N2N2M3N1O2N001O0O101O001O0O1mMjNRHW1i7TOPHl0l7ZORHf0l7^ORHb0k7CSH=j7HSH9k7KSH5k7OSH2i73UHMi76VHJh79VHHh7;WHEf7?YHAf7a0YH@e7b0ZHiNSOFa8d1ZHgNVOD_8h1YHdNYOC]8k1YHbN[OBZ8n1[H`N\\OAX8Q2[H_N]O@V8S2\\H]N_O_OT8V2\\H[NB]OQ8Z2\\HYND\\Oo7]2\\HWNF[Om7`2\\HUNHZOk7b2]HTNJYOg7e2^HSNMPOi7n2ZHRNU8P2jGPNU8R2jGnMU8S2kGmMT8U2kGkMT8W2kGiMT8X2lGhMS8Z2lGgMR8Z2nGeMR8]2mGcMR8_2mGaMS8_2mGaMR8a2mG_MR8b2nG^MQ8d2nG\\MQ8f2nGZMQ8g2oGYMQ8h2nGWMS8j2lGVMS8k2mGUMS8l2lGTMS8m2lGTMT8m2kGSMU8n2jGRMU8o2kGQMU8R40O010000O100O10001N101`HmJf6T5QIUKo6c5O10000O100O100O10000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O1000000000kN^IbKb6[4fI`KZ6^4nI\\KR6a4YJWKg5g4]JVKd5g4`JXK`5f4bJZK^5c4fJ\\KZ5^2XI^Oa1RNX5_2XI\\Od1TNT5_2ZIYOf1WNP5^2\\IYOg1XNm4]2_IYOa8f0`GXOb8f0aGWOa8g0bGWO_8g0cGWO^8h0eGUOe7PORHh1 in this image.", "objects": [{"patches": [63, 64, 85, 86, 87, 106, 107, 108, 109, 110, 128, 129, 130, 131, 132, 133, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 265, 266, 267, 268, 269], "bbox": [0.49548437500000003, 0.15694835680751174, 0.823828125, 0.775774647887324], "iscrowd": 0, "area": 25917.354099999997, "rle": {"size": [426, 640], "counts": "UbT43V=3N2N4L5K5K3L3N1O2N2N2M3N1O2N001O0O101O001O0O1mMjNRHW1i7TOPHl0l7ZORHf0l7^ORHb0k7CSH=j7HSH9k7KSH5k7OSH2i73UHMi76VHJh79VHHh7;WHEf7?YHAf7a0YH@e7b0ZHiNSOFa8d1ZHgNVOD_8h1YHdNYOC]8k1YHbN[OBZ8n1[H`N\\OAX8Q2[H_N]O@V8S2\\H]N_O_OT8V2\\H[NB]OQ8Z2\\HYND\\Oo7]2\\HWNF[Om7`2\\HUNHZOk7b2]HTNJYOg7e2^HSNMPOi7n2ZHRNU8P2jGPNU8R2jGnMU8S2kGmMT8U2kGkMT8W2kGiMT8X2lGhMS8Z2lGgMR8Z2nGeMR8]2mGcMR8_2mGaMS8_2mGaMR8a2mG_MR8b2nG^MQ8d2nG\\MQ8f2nGZMQ8g2oGYMQ8h2nGWMS8j2lGVMS8k2mGUMS8l2lGTMS8m2lGTMT8m2kGSMU8n2jGRMU8o2kGQMU8R40O010000O100O10001N101`HmJf6T5QIUKo6c5O10000O100O100O10000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O1000000000kN^IbKb6[4fI`KZ6^4nI\\KR6a4YJWKg5g4]JVKd5g4`JXK`5f4bJZK^5c4fJ\\KZ5^2XI^Oa1RNX5_2XI\\Od1TNT5_2ZIYOf1WNP5^2\\IYOg1XNm4]2_IYOa8f0`GXOb8f0aGWOa8g0bGWO_8g0cGWO^8h0eGUOe7PORHh1 in this image.", "objects": [{"patches": [95, 96, 118, 119, 120, 140, 141, 142, 143, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235], "bbox": [0.016453125, 0.3123708920187793, 0.272234375, 0.6898826291079811], "iscrowd": 0, "area": 14575.08955, "rle": {"size": [426, 640], "counts": "hi45T=6K5J6K0O2O1N2N2O0O2O1N2O1O1O00100O1O1O001]DkNo:V1nDlNS;T1jDoNU;R1hDPOX;P1hDQOW;P1hDPOX;Q1gDoNY;R1fDnN[;R1dDnN\\;R1dDnN\\;S1cDmN];T1bDlN^;`11O00000000001O00000000001O00001O0000O1M3L4L4M3L4POP1K5O1O1O2N1O1O1O1N20001N10000O10000O10000O10000O100000001O0O1000000O1]Oc0J6O1000000O10000O1000000O2O000001O00001O00001O00001O00000aNdGPN]8j1iGUNW8g1mGYNS8c1RH\\No7^1VHbNm7W1XHiNl7n0XHROm7d0YH[Om7;WHEP84QHKQ83oGMR83nGLT82lGNU82kGMW81iGOX80iGNZ81eGO\\80eGO]80cGO^80bG0`8O`G0a8O_G0c8O^G0d8O[G1f8NZG2h8MXG2i8MWG2l8MTG2m8MSG3n8LSG3o8LPG4Q9KPG4R9KmF4U9KkF5W9IjF6W9IiF4[9KfF2^9LbF1f9I]F2k9IVF3d;LWcQ6"}, "label": "a man sitting on the road taking a photo of a man skateboarding"}]} {"id": 176229, "image": "COCO_train2014_000000176229.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person holding the a camera , taking a photo\"."}], "task": "refering", "answer_template": "The \"the person holding the a camera , taking a photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 118, 119, 120, 140, 141, 142, 143, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235], "bbox": [0.016453125, 0.3123708920187793, 0.272234375, 0.6898826291079811], "iscrowd": 0, "area": 14575.08955, "rle": {"size": [426, 640], "counts": "hi45T=6K5J6K0O2O1N2N2O0O2O1N2O1O1O00100O1O1O001]DkNo:V1nDlNS;T1jDoNU;R1hDPOX;P1hDQOW;P1hDPOX;Q1gDoNY;R1fDnN[;R1dDnN\\;R1dDnN\\;S1cDmN];T1bDlN^;`11O00000000001O00000000001O00001O0000O1M3L4L4M3L4POP1K5O1O1O2N1O1O1O1N20001N10000O10000O10000O10000O100000001O0O1000000O1]Oc0J6O1000000O10000O1000000O2O000001O00001O00001O00001O00000aNdGPN]8j1iGUNW8g1mGYNS8c1RH\\No7^1VHbNm7W1XHiNl7n0XHROm7d0YH[Om7;WHEP84QHKQ83oGMR83nGLT82lGNU82kGMW81iGOX80iGNZ81eGO\\80eGO]80cGO^80bG0`8O`G0a8O_G0c8O^G0d8O[G1f8NZG2h8MXG2i8MWG2l8MTG2m8MSG3n8LSG3o8LPG4Q9KPG4R9KmF4U9KkF5W9IjF6W9IiF4[9KfF2^9LbF1f9I]F2k9IVF3d;LWcQ6"}, "label": "the person holding the a camera , taking a photo"}]} {"id": 110704, "image": "COCO_train2014_000000110704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in blue\"."}], "task": "refering", "answer_template": "The \"a woman in blue\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 51, 52, 53, 54, 74, 75, 76, 77, 96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 235, 236, 237, 238, 258, 259, 260, 261, 281, 282, 283, 284, 304, 305, 306, 307, 327, 328, 329, 330, 350, 351, 352, 353, 373, 374, 375, 376], "bbox": [0.187078125, 0.07122916666666666, 0.43989062500000004, 0.9858333333333333], "iscrowd": 0, "area": 41344.866599999994, "rle": {"size": [480, 640], "counts": "\\]h14i>6K5J6J7I:F9G:G8G9G5K6J5K6L3Mi0WOLhcW5"}, "label": "a woman in blue"}]} {"id": 110704, "image": "COCO_train2014_000000110704.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a nurse wearing blue scrubs\"."}], "task": "refering", "answer_template": "The \"a nurse wearing blue scrubs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 51, 52, 53, 54, 74, 75, 76, 77, 96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 235, 236, 237, 238, 258, 259, 260, 261, 281, 282, 283, 284, 304, 305, 306, 307, 327, 328, 329, 330, 350, 351, 352, 353, 373, 374, 375, 376], "bbox": [0.187078125, 0.07122916666666666, 0.43989062500000004, 0.9858333333333333], "iscrowd": 0, "area": 41344.866599999994, "rle": {"size": [480, 640], "counts": "\\]h14i>6K5J6J7I:F9G:G8G9G5K6J5K6L3Mi0WOLhcW5"}, "label": "a nurse wearing blue scrubs"}]} {"id": 274547, "image": "COCO_train2014_000000274547.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motorcycle on a brick street\"."}], "task": "refering", "answer_template": "The \"a motorcycle on a brick street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 216, 217, 218, 219, 220, 237, 238, 239, 240, 241, 242, 243, 244, 245, 260, 261, 262, 263, 264, 265, 267, 268, 283, 284, 285, 286, 287, 288, 290, 291, 308, 309], "bbox": [0.33546875, 0.4751836734693878, 0.69575, 0.7487142857142858], "iscrowd": 0, "area": 14217.481399999995, "rle": {"size": [490, 640], "counts": "iUW35T?8I1^ADS>=kAEU>WOe;_OmCW1d0WO_;BlCT1m0VOX;FjCP1U1WOQ;IiCm0^1VOi:MiCi0e1VOc:2gCd0X>\\OhA`0\\>_OfA=]>CcA=\\>DdA<[>DbA5O0`>3jAFU>8?O2M2N2Nim>2mQA in this image.", "objects": [{"patches": [195, 196, 197, 216, 217, 218, 219, 220, 237, 238, 239, 240, 241, 242, 243, 244, 245, 260, 261, 262, 263, 264, 265, 267, 268, 283, 284, 285, 286, 287, 288, 290, 291, 308, 309], "bbox": [0.33546875, 0.4751836734693878, 0.69575, 0.7487142857142858], "iscrowd": 0, "area": 14217.481399999995, "rle": {"size": [490, 640], "counts": "iUW35T?8I1^ADS>=kAEU>WOe;_OmCW1d0WO_;BlCT1m0VOX;FjCP1U1WOQ;IiCm0^1VOi:MiCi0e1VOc:2gCd0X>\\OhA`0\\>_OfA=]>CcA=\\>DdA<[>DbA5O0`>3jAFU>8?O2M2N2Nim>2mQA in this image.", "objects": [{"patches": [29, 30, 31, 46, 47, 48, 49, 64, 65, 66, 67, 82, 83, 84, 85, 86, 100, 101, 102, 103, 104, 119, 120, 121, 122, 123, 137, 138, 139, 140, 141, 152, 153, 154, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 176, 177, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212], "bbox": [0.49168, 0.12570996978851964, 0.86194, 0.99595166163142], "iscrowd": 0, "area": 31858.568600000002, "rle": {"size": [331, 500], "counts": "Th_22W:5L4M4L3M4L2N00100O1O001M3G9N101O1N2O001N2O1O0O2O1O0O2O1O1N101O1O1O0jHZNg5g1PJaNP6`1fIiN[6W1\\IQOd6o0SIZOm6R200kK]Mb0c2WOjMc0U2]OmMc0R2\\OoMd0Q2ZORNe0n1ZOTNe0l1ZOVNe0j1ZOXNe0h1ZOZNe0f1ZO[Ng0d1XO^Ng0b1XO`NoM]NZ2S3FbNmM^N\\2P3FdNkM_N_2l2DhNjM_Na2i2DXO;h0EXODD:=ED9>GC6?IC4?KCkNSM=[3g0DjNSM<[3j0CgNTM=[3k0BgNUMC=BZRf0"}, "label": "female standing outside on her phone"}]} {"id": 266369, "image": "COCO_train2014_000000266369.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman texting\"."}], "task": "refering", "answer_template": "The \"woman texting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 46, 47, 48, 49, 64, 65, 66, 67, 82, 83, 84, 85, 86, 100, 101, 102, 103, 104, 119, 120, 121, 122, 123, 137, 138, 139, 140, 141, 152, 153, 154, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 176, 177, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212], "bbox": [0.49168, 0.12570996978851964, 0.86194, 0.99595166163142], "iscrowd": 0, "area": 31858.568600000002, "rle": {"size": [331, 500], "counts": "Th_22W:5L4M4L3M4L2N00100O1O001M3G9N101O1N2O001N2O1O0O2O1O0O2O1O1N101O1O1O0jHZNg5g1PJaNP6`1fIiN[6W1\\IQOd6o0SIZOm6R200kK]Mb0c2WOjMc0U2]OmMc0R2\\OoMd0Q2ZORNe0n1ZOTNe0l1ZOVNe0j1ZOXNe0h1ZOZNe0f1ZO[Ng0d1XO^Ng0b1XO`NoM]NZ2S3FbNmM^N\\2P3FdNkM_N_2l2DhNjM_Na2i2DXO;h0EXODD:=ED9>GC6?IC4?KCkNSM=[3g0DjNSM<[3j0CgNTM=[3k0BgNUMC=BZRf0"}, "label": "woman texting"}]} {"id": 266369, "image": "COCO_train2014_000000266369.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing black watches wildlife\"."}], "task": "refering", "answer_template": "The \"a woman wearing black watches wildlife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 72, 73, 90, 91, 92, 108, 109, 127, 145], "bbox": [0.01578, 0.2460725075528701, 0.12544, 0.7252870090634441], "iscrowd": 0, "area": 4437.06685, "rle": {"size": [331, 500], "counts": "df26i9=N2[F_O]9i0N2O2M2O1N1O2oFnNf8_1L4\\O]NYHf1c7^NYHg1d7c0L4L3N3L4M3M3YImLo5X3dIVMX6b3L6J6K6I2OgK_Jg3`5WLdJh3^5TLeJk3^5oKdJR4j5001O01N1000000O4YOgIVM\\6V2`J_Mb5X2fJgM]5o1kJoMY5j1lJUNV5f1nJYNU5a1oJ^NS5]1Q2J6K6K4M6I7J5J7I]S]4"}, "label": "a woman wearing black watches wildlife"}]} {"id": 266369, "image": "COCO_train2014_000000266369.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a blue skirt standing next to a man\"."}], "task": "refering", "answer_template": "The \"a woman in a blue skirt standing next to a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 72, 73, 90, 91, 92, 108, 109, 127, 145], "bbox": [0.01578, 0.2460725075528701, 0.12544, 0.7252870090634441], "iscrowd": 0, "area": 4437.06685, "rle": {"size": [331, 500], "counts": "df26i9=N2[F_O]9i0N2O2M2O1N1O2oFnNf8_1L4\\O]NYHf1c7^NYHg1d7c0L4L3N3L4M3M3YImLo5X3dIVMX6b3L6J6K6I2OgK_Jg3`5WLdJh3^5TLeJk3^5oKdJR4j5001O01N1000000O4YOgIVM\\6V2`J_Mb5X2fJgM]5o1kJoMY5j1lJUNV5f1nJYNU5a1oJ^NS5]1Q2J6K6K4M6I7J5J7I]S]4"}, "label": "a woman in a blue skirt standing next to a man"}]} {"id": 323728, "image": "COCO_train2014_000000323728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an open silver laptop with the logo of a bitten apple which a woman is using\"."}], "task": "refering", "answer_template": "The \"an open silver laptop with the logo of a bitten apple which a woman is using\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [263, 264, 265, 266, 267, 287, 288, 289, 290, 291, 292, 293, 312, 313, 314, 315, 316, 335, 336, 337, 338, 339], "bbox": [0.46854687500000003, 0.6674166666666667, 0.792140625, 0.8876458333333334], "iscrowd": 0, "area": 11348.598449999998, "rle": {"size": [480, 640], "counts": "Pn\\43n>2M4L4L3M4L4L3N3L3M4L4L3M4LGhBmNV=Q1=O100O1000001O0000000000001O0000000000001O0000000000001O000000000000001O000001O000010O000001O00002O0O2N1XCPO^;R1ZDTOf;m0SDYOm;i0kC]OU in this image.", "objects": [{"patches": [93, 94, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 368, 369, 370], "bbox": [0.001953125, 0.2591041666666667, 0.302734375, 0.9856874999999999], "iscrowd": 0, "area": 41311.15, "rle": {"size": [480, 640], "counts": "cm04b>:E;^G[O^2P1_MWOS2S1kMROg1Y1VNmN\\1^1aNhNR1b1kNdNg0g1VO_N in this image.", "objects": [{"patches": [93, 94, 115, 116, 117, 118, 138, 139, 140, 141, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 368, 369, 370], "bbox": [0.001953125, 0.2591041666666667, 0.302734375, 0.9856874999999999], "iscrowd": 0, "area": 41311.15, "rle": {"size": [480, 640], "counts": "cm04b>:E;^G[O^2P1_MWOS2S1kMROg1Y1VNmN\\1^1aNhNR1b1kNdNg0g1VO_N in this image.", "objects": [{"patches": [106, 107, 108, 129, 130, 131, 132, 152, 153, 154, 155, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 293, 294, 316, 317, 339, 340, 361, 362, 363, 384, 385, 386], "bbox": [0.559265625, 0.25925, 0.8370312500000001, 0.9827083333333333], "iscrowd": 0, "area": 31727.190800000008, "rle": {"size": [480, 640], "counts": "RTX52k>3L4M3M3L4M4L3M3L4M3M3L4M4L3M3L4@`0_Oa0K6O0O100O10000O2O0nMmLVHT3i7TMPHl2o7[MkGe2S8dMeG]2Z8jM`GW2^8QN[Go1d8XNUGi1j8_NoFa1o8gNjFZ1U9nNdFR1[9Z2O2M2O1O1O1N2O1O2N1O1N2O1O1O2M2O1O1O1N2O1O20O00001O0001O01O0000001O01O0001O000010O000001O0000010O0000001O00010O0000001O01O0001O00000010O0002N3MTJPKj1n4UNSKn1k4QNUKQ2i4oMWKT2f4lMZKW2c4hM^KZ2`4fM`K]2]4cMcK_2]4_McKd2`4UMaKm2b4nL^KU3e4eL[K^3h4\\LXKf3k4TLVKo3m4kKSKW4Q5cKoJ`4S5[KmJg4W5RKjJQ5U5mJkJV5R5jJnJX5Q5gJoJ\\5n4cJSK^5l4bJTK_5l4`JTKa5k4^JVKc5j4ZJXKg5g4XJZKi5e4VJ\\Kk5d4RJ^Ko5a4PJ`KQ6_4nIbKS6^4jIdKW6[4hIfKY6Y4fIhK[6X4bIjK_6U4`IlKa6S4^InKc6R4ZIPLg6o3XIRLi6m3VITLk6_2RIlL0g0o6\\2UImNl6Q1YIkNh6T1[IiNf6U1_IgNb6X1bIdN_6[1eIaN\\6]1hI`NY6_1kI]NV6a1oI[NR6d1QJYNP6e1VJVNk5i1YJRNi5m1\\JkMh5S2]JgMf5X2_J`Mf5]2`J\\Mc5c2aJVMc5i2bJPMa5n2dJkL`5S3fJfL]5V3jJcLZ5Y3Y3L4L4L4L4LP2PNc^`1"}, "label": "a lady sitting outside on her laptop"}]} {"id": 323728, "image": "COCO_train2014_000000323728.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smiling blonde lady with her shades set against her hair\"."}], "task": "refering", "answer_template": "The \"a smiling blonde lady with her shades set against her hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 107, 108, 129, 130, 131, 132, 152, 153, 154, 155, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 243, 244, 245, 246, 247, 248, 249, 266, 267, 268, 269, 270, 271, 272, 293, 294, 316, 317, 339, 340, 361, 362, 363, 384, 385, 386], "bbox": [0.559265625, 0.25925, 0.8370312500000001, 0.9827083333333333], "iscrowd": 0, "area": 31727.190800000008, "rle": {"size": [480, 640], "counts": "RTX52k>3L4M3M3L4M4L3M3L4M3M3L4M4L3M3L4@`0_Oa0K6O0O100O10000O2O0nMmLVHT3i7TMPHl2o7[MkGe2S8dMeG]2Z8jM`GW2^8QN[Go1d8XNUGi1j8_NoFa1o8gNjFZ1U9nNdFR1[9Z2O2M2O1O1O1N2O1O2N1O1N2O1O1O2M2O1O1O1N2O1O20O00001O0001O01O0000001O01O0001O000010O000001O0000010O0000001O00010O0000001O01O0001O00000010O0002N3MTJPKj1n4UNSKn1k4QNUKQ2i4oMWKT2f4lMZKW2c4hM^KZ2`4fM`K]2]4cMcK_2]4_McKd2`4UMaKm2b4nL^KU3e4eL[K^3h4\\LXKf3k4TLVKo3m4kKSKW4Q5cKoJ`4S5[KmJg4W5RKjJQ5U5mJkJV5R5jJnJX5Q5gJoJ\\5n4cJSK^5l4bJTK_5l4`JTKa5k4^JVKc5j4ZJXKg5g4XJZKi5e4VJ\\Kk5d4RJ^Ko5a4PJ`KQ6_4nIbKS6^4jIdKW6[4hIfKY6Y4fIhK[6X4bIjK_6U4`IlKa6S4^InKc6R4ZIPLg6o3XIRLi6m3VITLk6_2RIlL0g0o6\\2UImNl6Q1YIkNh6T1[IiNf6U1_IgNb6X1bIdN_6[1eIaN\\6]1hI`NY6_1kI]NV6a1oI[NR6d1QJYNP6e1VJVNk5i1YJRNi5m1\\JkMh5S2]JgMf5X2_J`Mf5]2`J\\Mc5c2aJVMc5i2bJPMa5n2dJkL`5S3fJfL]5V3jJcLZ5Y3Y3L4L4L4L4LP2PNc^`1"}, "label": "a smiling blonde lady with her shades set against her hair"}]} {"id": 348315, "image": "COCO_train2014_000000348315.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two black and white zebras standing in a field of wheat\"."}], "task": "refering", "answer_template": "The \"two black and white zebras standing in a field of wheat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 190, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 244, 245, 246, 247, 248, 249, 263, 264, 267, 268, 269, 270, 291], "bbox": [0.25903125, 0.3109444444444444, 0.843109375, 0.9407499999999999], "iscrowd": 0, "area": 40326.06410000001, "rle": {"size": [360, 640], "counts": "Tbj16i::F9H9L4M3N1O2N2N1O2N2N2M2O00O1O1O1001O0O2O001O001O001O0O2jNVNmHj1T7ZNgHf1Z7\\NcHe1]7\\N`He1a7\\N]He1b7\\N]Hd1d7]NZHd1f7]NWHd1j7]NTHc1m7]NRHd1n7]NPHc1Q8^NlGc1U8^NiGc1W8^NgGb1Z8_NdGa1]8e0O01O100O1O100O1O100O00100O1O100O1O100O11O2M2O2N2N2N2N101N2O1O0O2O1O1N101O1N2O001N2O1O2M2O1O2M2O1O1N3N1O1N2O1O0O2O0O2O1O0O2N101N1O2O0O2N2N2O2M2N2O2M2N2N3N1N2N2O2M2N2O2M2N2N3N1N2N10O0O10001O0O101O0O101O000O101O000O2O000O101O0YNUKQMk4P3XKkLj4T3ZKhLf4Y3]KcLc4]3aK^La4b3b101O100O100O2O0O1O1O2N1O100O2N0000000000001O01O0000000000000000001O0001O00000000000001O0000000001O00000001O0000000000000001O01O00001O2N3M3M2N3M3M2N3L4M2N2N2N1O2N2N1O2N2N1O2N1N3N2N1O2N2N1O1O1O001O1O001N100000O100O100O100O100O10O0100O100O100O1000000O1000O1000O100000000O100000000O010O2O1O1N2O1O0bNaIRNa6k1mIiMT6T2RJhMo5V2WJdMk5Y2\\JbMe5[2f1M3N2M3M2N3N2M3M3M3N2M3M5K5K5L4K5K5K5M3O00001O000001O010O001N101N101N10O010O000010O01O001O0N2O2M2O1N2O1N2O2M2O1N3N4K4M[lR1"}, "label": "two black and white zebras standing in a field of wheat"}]} {"id": 348315, "image": "COCO_train2014_000000348315.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra standing in front of another , with his tail blowing in the wind\"."}], "task": "refering", "answer_template": "The \"a zebra standing in front of another , with his tail blowing in the wind\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 121, 122, 123, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 190, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 240, 241, 244, 245, 246, 247, 248, 249, 263, 264, 267, 268, 269, 270, 291], "bbox": [0.25903125, 0.3109444444444444, 0.843109375, 0.9407499999999999], "iscrowd": 0, "area": 40326.06410000001, "rle": {"size": [360, 640], "counts": "Tbj16i::F9H9L4M3N1O2N2N1O2N2N2M2O00O1O1O1001O0O2O001O001O001O0O2jNVNmHj1T7ZNgHf1Z7\\NcHe1]7\\N`He1a7\\N]He1b7\\N]Hd1d7]NZHd1f7]NWHd1j7]NTHc1m7]NRHd1n7]NPHc1Q8^NlGc1U8^NiGc1W8^NgGb1Z8_NdGa1]8e0O01O100O1O100O1O100O00100O1O100O1O100O11O2M2O2N2N2N2N101N2O1O0O2O1O1N101O1N2O001N2O1O2M2O1O2M2O1O1N3N1O1N2O1O0O2O0O2O1O0O2N101N1O2O0O2N2N2O2M2N2O2M2N2N3N1N2N2O2M2N2O2M2N2N3N1N2N10O0O10001O0O101O0O101O000O101O000O2O000O101O0YNUKQMk4P3XKkLj4T3ZKhLf4Y3]KcLc4]3aK^La4b3b101O100O100O2O0O1O1O2N1O100O2N0000000000001O01O0000000000000000001O0001O00000000000001O0000000001O00000001O0000000000000001O01O00001O2N3M3M2N3M3M2N3L4M2N2N2N1O2N2N1O2N2N1O2N1N3N2N1O2N2N1O1O1O001O1O001N100000O100O100O100O100O10O0100O100O100O1000000O1000O1000O100000000O100000000O010O2O1O1N2O1O0bNaIRNa6k1mIiMT6T2RJhMo5V2WJdMk5Y2\\JbMe5[2f1M3N2M3M2N3N2M3M3M3N2M3M5K5K5L4K5K5K5M3O00001O000001O010O001N101N101N10O010O000010O01O001O0N2O2M2O1N2O1N2O2M2O1N3N4K4M[lR1"}, "label": "a zebra standing in front of another , with his tail blowing in the wind"}]} {"id": 348315, "image": "COCO_train2014_000000348315.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra furthest away\"."}], "task": "refering", "answer_template": "The \"the zebra furthest away\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 189, 190, 191, 192, 193, 212, 213, 215, 216, 238, 239, 262], "bbox": [0.24270312500000002, 0.5325833333333333, 0.41965625000000006, 0.8786388888888889], "iscrowd": 0, "area": 4659.869549999999, "rle": {"size": [360, 640], "counts": "cgf11S;4K6K4K6K5J5L5K50O10O010O1O001O10O000O101N10000O100O1000O010O010O01J5L5J5N3O10O01O001O001O001O1O001O0100O1O1O100O002N2N2N2N3N1N2N2N2N2N2O1N3M3M3M4M3O0O2N2N101N2N101N101O1O001O001O1O002N2N3M2N2N2NN2N2N2N2N2N2N2N2N2N2N2F:C=D in this image.", "objects": [{"patches": [167, 168, 169, 189, 190, 191, 192, 193, 212, 213, 215, 216, 238, 239, 262], "bbox": [0.24270312500000002, 0.5325833333333333, 0.41965625000000006, 0.8786388888888889], "iscrowd": 0, "area": 4659.869549999999, "rle": {"size": [360, 640], "counts": "cgf11S;4K6K4K6K5J5L5K50O10O010O1O001O10O000O101N10000O100O1000O010O010O01J5L5J5N3O10O01O001O001O001O1O001O0100O1O1O100O002N2N2N2N3N1N2N2N2N2N2O1N3M3M3M4M3O0O2N2N101N2N101N101O1O001O001O1O002N2N3M2N2N2NN2N2N2N2N2N2N2N2N2N2N2F:C=D in this image.", "objects": [{"patches": [42, 43, 59, 60, 61, 62, 77, 78, 79, 80, 95, 96, 97, 98, 112, 113, 114, 115, 116, 129, 130, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 153, 154, 165, 166, 167, 168, 169, 170, 171, 172, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 200, 201, 202, 203, 204, 205, 206, 207, 208, 218, 219, 220, 221, 222, 223, 224, 225, 226], "bbox": [0.10578, 0.21056, 0.6164, 1.0], "iscrowd": 0, "area": 45368.05070000001, "rle": {"size": [375, 500], "counts": "Qhc01f;1N2O2N1O6I:G4L000O1000000O10000000000mNYOUGh0i8]OTGc0j8ATG?j8FSG:k8JSG6k8ORG1l83RGMk88QGJm8:PGGn8=nFEP9b1N2N2M3M3M3M3M3J6I7I7M3M3M3M3M3N2M3M3M3M3O1N2O1O1N2O1O1N2O1O1O1O1O1O100O1O1O1O1O100O100O10000O100O100O10000O100O10000O100_NZJQMf5i2cJTM]5e2lJYMT5`2UK^Mk4[2^KbMc4W2fKgMZ4T2nKiMR4V2QLhMo3V2ULhMk3W2XLfMi3X2[LfMe3Z2]LdMc3[2`LcM`3\\2cLbM]3^2eL_M\\3`2gL^MY3b2jLYMX3f2kLSMZ3l2i200O100O100O1O100O100O100O100O100O100O100O100O100O1000000O10000000000000000000000000000000000001O001O1O1O2N1O1O00000000000000000000001O1O1O1O1O1TJ^Ma2c2fLVNY3k1cLXN]3h1cLYN\\3i1cLVN]3l1aLUN^3m1`LSN`3P2WLWNh3l401O00001O001O001O001O001O0O1O2N1O2O0O2N2N2N1O2N2N2N1O3M3M2N3M3M2N3N2M2N3M3M3M2N3M3M2N3M3M3M3M3M3M3M3M2O2M3M3M4L3M3M4L3M3J7H7H9H8G8IoPV2"}, "label": "a woman in a blue blouse looks at her phone"}]} {"id": 495776, "image": "COCO_train2014_000000495776.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a light blue sleeveless shirt looking at her phone\"."}], "task": "refering", "answer_template": "The \"a woman in a light blue sleeveless shirt looking at her phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [42, 43, 59, 60, 61, 62, 77, 78, 79, 80, 95, 96, 97, 98, 112, 113, 114, 115, 116, 129, 130, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 153, 154, 165, 166, 167, 168, 169, 170, 171, 172, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 200, 201, 202, 203, 204, 205, 206, 207, 208, 218, 219, 220, 221, 222, 223, 224, 225, 226], "bbox": [0.10578, 0.21056, 0.6164, 1.0], "iscrowd": 0, "area": 45368.05070000001, "rle": {"size": [375, 500], "counts": "Qhc01f;1N2O2N1O6I:G4L000O1000000O10000000000mNYOUGh0i8]OTGc0j8ATG?j8FSG:k8JSG6k8ORG1l83RGMk88QGJm8:PGGn8=nFEP9b1N2N2M3M3M3M3M3J6I7I7M3M3M3M3M3N2M3M3M3M3O1N2O1O1N2O1O1N2O1O1O1O1O1O100O1O1O1O1O100O100O10000O100O100O10000O100O10000O100_NZJQMf5i2cJTM]5e2lJYMT5`2UK^Mk4[2^KbMc4W2fKgMZ4T2nKiMR4V2QLhMo3V2ULhMk3W2XLfMi3X2[LfMe3Z2]LdMc3[2`LcM`3\\2cLbM]3^2eL_M\\3`2gL^MY3b2jLYMX3f2kLSMZ3l2i200O100O100O1O100O100O100O100O100O100O100O100O100O1000000O10000000000000000000000000000000000001O001O1O1O2N1O1O00000000000000000000001O1O1O1O1O1TJ^Ma2c2fLVNY3k1cLXN]3h1cLYN\\3i1cLVN]3l1aLUN^3m1`LSN`3P2WLWNh3l401O00001O001O001O001O001O0O1O2N1O2O0O2N2N2N1O2N2N2N1O3M3M2N3M3M2N3N2M2N3M3M3M2N3M3M2N3M3M3M3M3M3M3M3M2O2M3M3M4L3M3M4L3M3J7H7H9H8G8IoPV2"}, "label": "a woman in a light blue sleeveless shirt looking at her phone"}]} {"id": 340129, "image": "COCO_train2014_000000340129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in red\"."}], "task": "refering", "answer_template": "The \"the man in red\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 60, 61, 62, 83, 84, 85, 86, 107, 108, 109, 110, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 337, 338, 339, 340, 341, 342], "bbox": [0.598375, 0.07836065573770493, 0.8990625, 0.981943793911007], "iscrowd": 0, "area": 44580.8641, "rle": {"size": [427, 640], "counts": "mlo47T=1N3N2N2M3N2N2M2O2N2N2M3N2gKH[K8b4L\\K5b4O\\K1a43^KM_48_KH_4;_KF^4?`KA_4a0`K@^4c0_K^O`4d0_K\\O`4f0_KZO`4i0]KXOc4i0[KXOd4k0YKVOf4l0WKVOh4l0VKVOi4k0TKWOk4l0RKUOm4m0QKTOn4n0oJTOQ5n0lJSOT5n0cIbM:b1S6m0WIe0h6o200000000000001O2N5K5K5K5K5K5K5K5K5K:Fh1XNb1^N4L4L3M4L4K3N0002N2N3M3M2N2N2N2N1O2N2N2N2N2N1O2N2N2N1O2N2N1O2N2N1O2N1O1O1O1O1O1O2N1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O4L6J7I7I5K4L5K4L4L5Kb0^Of0ZOP1PO5J6K5K4L5K5K5K4L5K5K5K4L5K5K5K4F;E;G9F9G:F:Fjgj0"}, "label": "the man in red"}]} {"id": 340129, "image": "COCO_train2014_000000340129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the red shirt carrying baseball bats\"."}], "task": "refering", "answer_template": "The \"the man in the red shirt carrying baseball bats\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 60, 61, 62, 83, 84, 85, 86, 107, 108, 109, 110, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 157, 174, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 314, 315, 316, 317, 318, 319, 337, 338, 339, 340, 341, 342], "bbox": [0.598375, 0.07836065573770493, 0.8990625, 0.981943793911007], "iscrowd": 0, "area": 44580.8641, "rle": {"size": [427, 640], "counts": "mlo47T=1N3N2N2M3N2N2M2O2N2N2M3N2gKH[K8b4L\\K5b4O\\K1a43^KM_48_KH_4;_KF^4?`KA_4a0`K@^4c0_K^O`4d0_K\\O`4f0_KZO`4i0]KXOc4i0[KXOd4k0YKVOf4l0WKVOh4l0VKVOi4k0TKWOk4l0RKUOm4m0QKTOn4n0oJTOQ5n0lJSOT5n0cIbM:b1S6m0WIe0h6o200000000000001O2N5K5K5K5K5K5K5K5K5K:Fh1XNb1^N4L4L3M4L4K3N0002N2N3M3M2N2N2N2N1O2N2N2N2N2N1O2N2N2N1O2N2N1O2N2N1O2N1O1O1O1O1O1O2N1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O4L6J7I7I5K4L5K4L4L5Kb0^Of0ZOP1PO5J6K5K4L5K5K5K4L5K5K5K4L5K5K5K4F;E;G9F9G:F:Fjgj0"}, "label": "the man in the red shirt carrying baseball bats"}]} {"id": 340129, "image": "COCO_train2014_000000340129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a man taking a pitcure of two other men\"."}], "task": "refering", "answer_template": "The \"the back of a man taking a pitcure of two other men\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 93, 94, 95, 96, 116, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328], "bbox": [0.0, 0.20899297423887586, 0.4063125, 0.9842622950819673], "iscrowd": 0, "area": 60227.10129999999, "rle": {"size": [427, 640], "counts": "k6Q3Y:3N1O1N3N1N3N1O1N3nH`LU4a3`KiL`4Y3TKQMl4P3iJZMW5g2]JdMb5_2RJkMn5V2gITNY6n1[I\\Ne6U40000O10000000000[NlIRLU6f3UJXLk5e3ZJYLf5d3_JZLa5c3dJ[L\\5b3iJ\\LW5a3nJ\\LS5i2kKTMU4l2mKRMS4m2oKRMQ4m2PLSMP4l2RLSMn3l2TLRMm3m2ULRMk3m2VLSMj3l2XLSMh3l2ZLSMf3l2[LTMe3k2]LSMd3l2^LSMb3m2_LPMc3o2^LnLe3R3\\LiLh3W3V3O10000O10000O1000000O10000O10000O10000O1000000O10000O1001O001O00001O001O00001O001O002N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O000000001O00000000001O001O1O001O001O1O001O001Ok0UO0000001O00000000000000000000000000000000000000000000000000000000M3000000000O100000001O001O00001O00001O001O00001O001RMSK\\Mm4e2kKbLU4^3cLjK^3U4[MRKe2n4m20O00010O01O011N1O2N101N1O2O1N1O2O0O2N1O4M5J6J6K2M2N2N101M3N2M2O2M3N2M2O2M3M2O2M3N2M2O2M3N2M2O2M3N1N3M3N2M2N3M3L3N3M3L4M2N3L9Hb0^O[Un4"}, "label": "the back of a man taking a pitcure of two other men"}]} {"id": 340129, "image": "COCO_train2014_000000340129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a white hat and holding a mobile\"."}], "task": "refering", "answer_template": "The \"a man wearing a white hat and holding a mobile\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 93, 94, 95, 96, 116, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328], "bbox": [0.0, 0.20899297423887586, 0.4063125, 0.9842622950819673], "iscrowd": 0, "area": 60227.10129999999, "rle": {"size": [427, 640], "counts": "k6Q3Y:3N1O1N3N1N3N1O1N3nH`LU4a3`KiL`4Y3TKQMl4P3iJZMW5g2]JdMb5_2RJkMn5V2gITNY6n1[I\\Ne6U40000O10000000000[NlIRLU6f3UJXLk5e3ZJYLf5d3_JZLa5c3dJ[L\\5b3iJ\\LW5a3nJ\\LS5i2kKTMU4l2mKRMS4m2oKRMQ4m2PLSMP4l2RLSMn3l2TLRMm3m2ULRMk3m2VLSMj3l2XLSMh3l2ZLSMf3l2[LTMe3k2]LSMd3l2^LSMb3m2_LPMc3o2^LnLe3R3\\LiLh3W3V3O10000O10000O1000000O10000O10000O10000O1000000O10000O1001O001O00001O001O00001O001O002N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O000000001O00000000001O001O1O001O001O1O001O001Ok0UO0000001O00000000000000000000000000000000000000000000000000000000M3000000000O100000001O001O00001O00001O001O00001O001RMSK\\Mm4e2kKbLU4^3cLjK^3U4[MRKe2n4m20O00010O01O011N1O2N101N1O2O1N1O2O0O2N1O4M5J6J6K2M2N2N101M3N2M2O2M3N2M2O2M3M2O2M3N2M2O2M3N2M2O2M3N1N3M3N2M2N3M3L3N3M3L4M2N3L9Hb0^O[Un4"}, "label": "a man wearing a white hat and holding a mobile"}]} {"id": 340129, "image": "COCO_train2014_000000340129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a short man stands next to a taller man\"."}], "task": "refering", "answer_template": "The \"a short man stands next to a taller man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 126, 127, 128, 149, 150, 151, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 332, 333, 334, 335, 336], "bbox": [0.41040625000000003, 0.2796018735362998, 0.65425, 0.9997423887587822], "iscrowd": 0, "area": 34078.40469999999, "rle": {"size": [427, 640], "counts": "Ro]3f0[;HSEm0^:\\O\\Eo0V:YOeER1o9]1H8XF`LW9n3K5QGmK`8V4]GRL[8`4N3M3M3M3M3nHgJn5\\5jIkJT6W5eIPKZ6R5^IUKb6j50000O1000000O10000O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O1O100gNkH[LV7d3lHWK4k0P7l3nHXK9e0j6f3jIYLV6a3QJ]LP6b3RJ\\Lo5c3RJ\\Lo5d3QJZLQ6e3QJQLX6n3c1O1O1O1O1O100O1O10000O10000O1000000O10000O11O00001O00001O00001O00001O000cH`Lo4`3nJeLP5\\3mJgLR5[3iJiLV5X3fJlLY5V3dJmLZ5T3eJlL[5U3cJlL]5Y3^JgLb5^3XJcLh5b3RJ_Ln5c3PJ]LP6d3ZI`K:n0[6m3bIUL^6]5O1O001O1O001O1O001O1O001O1O1O1O2N2K5K5N2oKdHo1c7iMaHU2f7cM]HZ2n7ZMVHc2P8dLkG27X3b8dLbGX3b8dLbGY3Z9Mj0WO6J6I7J6J6I6I8F:G9F9Gehk2"}, "label": "a short man stands next to a taller man"}]} {"id": 340129, "image": "COCO_train2014_000000340129.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a blue tshirt and blue baseball cap\"."}], "task": "refering", "answer_template": "The \"a man wearing a blue tshirt and blue baseball cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 105, 126, 127, 128, 149, 150, 151, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 332, 333, 334, 335, 336], "bbox": [0.41040625000000003, 0.2796018735362998, 0.65425, 0.9997423887587822], "iscrowd": 0, "area": 34078.40469999999, "rle": {"size": [427, 640], "counts": "Ro]3f0[;HSEm0^:\\O\\Eo0V:YOeER1o9]1H8XF`LW9n3K5QGmK`8V4]GRL[8`4N3M3M3M3M3nHgJn5\\5jIkJT6W5eIPKZ6R5^IUKb6j50000O1000000O10000O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O1O100gNkH[LV7d3lHWK4k0P7l3nHXK9e0j6f3jIYLV6a3QJ]LP6b3RJ\\Lo5c3RJ\\Lo5d3QJZLQ6e3QJQLX6n3c1O1O1O1O1O100O1O10000O10000O1000000O10000O11O00001O00001O00001O00001O000cH`Lo4`3nJeLP5\\3mJgLR5[3iJiLV5X3fJlLY5V3dJmLZ5T3eJlL[5U3cJlL]5Y3^JgLb5^3XJcLh5b3RJ_Ln5c3PJ]LP6d3ZI`K:n0[6m3bIUL^6]5O1O001O1O001O1O001O1O001O1O1O1O2N2K5K5N2oKdHo1c7iMaHU2f7cM]HZ2n7ZMVHc2P8dLkG27X3b8dLbGX3b8dLbGY3Z9Mj0WO6J6I7J6J6I6I8F:G9F9Gehk2"}, "label": "a man wearing a blue tshirt and blue baseball cap"}]} {"id": 479396, "image": "COCO_train2014_000000479396.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in black dress playing wii\"."}], "task": "refering", "answer_template": "The \"a woman in black dress playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 61, 62, 63, 80, 81, 82, 98, 99, 100, 114, 115, 116, 117, 118, 132, 133, 134, 135, 136, 151, 152, 153, 154, 170, 171, 172, 188, 189, 206, 207, 224, 225], "bbox": [0.3455, 0.2108533333333333, 0.58484, 0.9884], "iscrowd": 0, "area": 17375.871550000007, "rle": {"size": [375, 500], "counts": "\\ao12d;2N3M2O2L3M3N2N3N1O00000000000001O00001O00001O_EM^93`F1_9N`F3`9M_F5a9J]F8d9G[F:e9GYF:h9EWF;k9DTFZ5FcJ9Y5O`J5^50YJ6SO]1o2aNgM7lNi1Y3RNgM:iNl1\\3kMhMV4S2mKjMW4S2jKjM[4S2gKiM^4U2bKhMc4V2]KgMQ5m1PKPNS5P2mJlMW5S2jJjMY5V2gJgM\\5X2fJdM]5\\2gJ\\M]5c2l1O100O10000000000000000000000001O000000001O0000001O0010O01O1O2N1nNU1C=F:YOh0F9G9H;D in this image.", "objects": [{"patches": [44, 45, 61, 62, 63, 80, 81, 82, 98, 99, 100, 114, 115, 116, 117, 118, 132, 133, 134, 135, 136, 151, 152, 153, 154, 170, 171, 172, 188, 189, 206, 207, 224, 225], "bbox": [0.3455, 0.2108533333333333, 0.58484, 0.9884], "iscrowd": 0, "area": 17375.871550000007, "rle": {"size": [375, 500], "counts": "\\ao12d;2N3M2O2L3M3N2N3N1O00000000000001O00001O00001O_EM^93`F1_9N`F3`9M_F5a9J]F8d9G[F:e9GYF:h9EWF;k9DTFZ5FcJ9Y5O`J5^50YJ6SO]1o2aNgM7lNi1Y3RNgM:iNl1\\3kMhMV4S2mKjMW4S2jKjM[4S2gKiM^4U2bKhMc4V2]KgMQ5m1PKPNS5P2mJlMW5S2jJjMY5V2gJgM\\5X2fJdM]5\\2gJ\\M]5c2l1O100O10000000000000000000000001O000000001O0000001O0010O01O1O2N1nNU1C=F:YOh0F9G9H;D in this image.", "objects": [{"patches": [173, 174, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 225, 226, 227, 228, 229], "bbox": [0.54402, 0.6770133333333334, 0.78418, 1.0], "iscrowd": 0, "area": 9387.376400000003, "rle": {"size": [375, 500], "counts": "hnS33m91dG5W84`G0]8b0PGBm8m0cFVO[9c1000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000001O001O1O001O1OM3J6J6J6K5J6H8K5N2N2O13M2N3M2N1O001O1O00001O000000001O1O001O1O1O1O2N1O2N1O1O1O1O1O1O1O1O1N2M2O1N2O1N2O1N2N2M3N2N2N2O1O1N3N1`N_Fi0b9TOeFf0^9UOmFc0W:K5K3L4M2N[WW1"}, "label": "sofa"}]} {"id": 479396, "image": "COCO_train2014_000000479396.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty couch seat behind the woman holding the wii controller\"."}], "task": "refering", "answer_template": "The \"an empty couch seat behind the woman holding the wii controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 225, 226, 227, 228, 229], "bbox": [0.54402, 0.6770133333333334, 0.78418, 1.0], "iscrowd": 0, "area": 9387.376400000003, "rle": {"size": [375, 500], "counts": "hnS33m91dG5W84`G0]8b0PGBm8m0cFVO[9c1000000O1000000000000O100000000000000O1000000000000O1000000000000O100000000001O001O1O001O1OM3J6J6J6K5J6H8K5N2N2O13M2N3M2N1O001O1O00001O000000001O1O001O1O1O1O2N1O2N1O1O1O1O1O1O1O1O1N2M2O1N2O1N2O1N2N2M3N2N2N2O1O1N3N1`N_Fi0b9TOeFf0^9UOmFc0W:K5K3L4M2N[WW1"}, "label": "an empty couch seat behind the woman holding the wii controller"}]} {"id": 479396, "image": "COCO_train2014_000000479396.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a brown sweater and holding a game control\"."}], "task": "refering", "answer_template": "The \"a man wearing a brown sweater and holding a game control\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 32, 33, 34, 35, 50, 51, 52, 53, 68, 69, 70, 71, 86, 87, 88, 89, 104, 105, 106, 107, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 212, 213, 214, 215, 229, 230, 231, 232, 233], "bbox": [0.65604, 0.004906666666666667, 1.0, 0.9959466666666665], "iscrowd": 0, "area": 40902.612100000006, "rle": {"size": [375, 500], "counts": "kZh34a;2N110iDd0S:\\OiEh0X:WOdEm0\\:700000000000O100000AhN_FY1`9hN_FX1a9iN^FX1a9iN^FX1a9iN^FX1a9iN^FW1b9jN]FW1b9jN]FW1b9jN]FW1b9kN^FT1a9mN`FQ1`9POaFo0^9ROcFm0\\9TOeFk0[9UOdFl0[9UO`Fo0a9d000O1WInMl3R2RLoMo3P2PLQNQ4o1mKRNS4o1kKRNV4n1hKSNY4l1fKUNZ4l1dKUN]4k1aKVN`4j1]KXNc4i1[KXNf4g1YKZNh4f1VK[Nk4e1SK\\Nm4e1QK\\Nn4f1PK[Nn4g1QKZNm4j1PKWNo4n1kJTNS5Q2iJPNU5V2fJkMX5[2cJfM\\5^2`JcM^5c2]J^Ma5h2ZJYMd5l2XJUM\\1WOV2i3YLQM_1ZOT2k3WLmLc1\\OR2l3WLiLd1AP2l3WLdLe1FQ2j4lM\\K;`0XNV4Y1hK3n5I[J0g5M]JOf5M`JOb5NbJO`5MfJO\\5MjJOX5NlJOV5MnJ0U5MnJ1T5NmJ2S5NnJ0S50mJ0S5OoJOR51nJOR50oJOR51oJNQ52oJMR52oJNP53QKKP55PKKP54RKJo46QKJo45RKJo46RKIn47RKHo47Y300O10000O10000O100O10000000000001O0000000000001O00000000001O0000000000001O0000000000001O001O001O1O1N2O1O1PG@g6a0VICh6>UIGh6;TIJi67TINi64SI0k62RI0m62oH1P71lH2S71hH2W70eH3Z7O`H6_7LZH:e7m1O2N1O1O2N1O1O2N1O1O2N1O1@`0UKn0"}, "label": "a man wearing a brown sweater and holding a game control"}]} {"id": 479396, "image": "COCO_train2014_000000479396.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a sweater playing a video game with a woman\"."}], "task": "refering", "answer_template": "The \"a man in a sweater playing a video game with a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 32, 33, 34, 35, 50, 51, 52, 53, 68, 69, 70, 71, 86, 87, 88, 89, 104, 105, 106, 107, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 212, 213, 214, 215, 229, 230, 231, 232, 233], "bbox": [0.65604, 0.004906666666666667, 1.0, 0.9959466666666665], "iscrowd": 0, "area": 40902.612100000006, "rle": {"size": [375, 500], "counts": "kZh34a;2N110iDd0S:\\OiEh0X:WOdEm0\\:700000000000O100000AhN_FY1`9hN_FX1a9iN^FX1a9iN^FX1a9iN^FX1a9iN^FW1b9jN]FW1b9jN]FW1b9jN]FW1b9kN^FT1a9mN`FQ1`9POaFo0^9ROcFm0\\9TOeFk0[9UOdFl0[9UO`Fo0a9d000O1WInMl3R2RLoMo3P2PLQNQ4o1mKRNS4o1kKRNV4n1hKSNY4l1fKUNZ4l1dKUN]4k1aKVN`4j1]KXNc4i1[KXNf4g1YKZNh4f1VK[Nk4e1SK\\Nm4e1QK\\Nn4f1PK[Nn4g1QKZNm4j1PKWNo4n1kJTNS5Q2iJPNU5V2fJkMX5[2cJfM\\5^2`JcM^5c2]J^Ma5h2ZJYMd5l2XJUM\\1WOV2i3YLQM_1ZOT2k3WLmLc1\\OR2l3WLiLd1AP2l3WLdLe1FQ2j4lM\\K;`0XNV4Y1hK3n5I[J0g5M]JOf5M`JOb5NbJO`5MfJO\\5MjJOX5NlJOV5MnJ0U5MnJ1T5NmJ2S5NnJ0S50mJ0S5OoJOR51nJOR50oJOR51oJNQ52oJMR52oJNP53QKKP55PKKP54RKJo46QKJo45RKJo46RKIn47RKHo47Y300O10000O10000O100O10000000000001O0000000000001O00000000001O0000000000001O0000000000001O001O001O1O1N2O1O1PG@g6a0VICh6>UIGh6;TIJi67TINi64SI0k62RI0m62oH1P71lH2S71hH2W70eH3Z7O`H6_7LZH:e7m1O2N1O1O2N1O1O2N1O1O2N1O1@`0UKn0"}, "label": "a man in a sweater playing a video game with a woman"}]} {"id": 479396, "image": "COCO_train2014_000000479396.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a maroon velvet chair with matching pillow in a livingroom\"."}], "task": "refering", "answer_template": "The \"a maroon velvet chair with matching pillow in a livingroom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 144, 145, 146, 147, 162, 163, 164, 165, 180, 181, 182, 183, 184, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220], "bbox": [0.0, 0.6092, 0.2742, 0.9898666666666667], "iscrowd": 0, "area": 15681.146699999998, "rle": {"size": [375, 500], "counts": "X7f2Q9d1\\N00000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O1001O00000000001O000000001O00000000001O000000001O000000001O0010O0001O00001O001O00a0_O5K6J6J5K6J5K2N3M2N2N2N1O1O0000001O000000001O0000001O0000001O0000001O0000001O0000001O0000001O0000XdT4"}, "label": "a maroon velvet chair with matching pillow in a livingroom"}]} {"id": 479396, "image": "COCO_train2014_000000479396.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a maroon couch with a video game system box on top\"."}], "task": "refering", "answer_template": "The \"a maroon couch with a video game system box on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 144, 145, 146, 147, 162, 163, 164, 165, 180, 181, 182, 183, 184, 198, 199, 200, 201, 202, 216, 217, 218, 219, 220], "bbox": [0.0, 0.6092, 0.2742, 0.9898666666666667], "iscrowd": 0, "area": 15681.146699999998, "rle": {"size": [375, 500], "counts": "X7f2Q9d1\\N00000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O1001O00000000001O000000001O00000000001O000000001O000000001O0010O0001O00001O001O00a0_O5K6J6J5K6J5K2N3M2N2N2N1O1O0000001O000000001O0000001O0000001O0000001O0000001O0000001O0000001O0000XdT4"}, "label": "a maroon couch with a video game system box on top"}]} {"id": 45226, "image": "COCO_train2014_000000045226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman blowing out birthday candles\"."}], "task": "refering", "answer_template": "The \"a woman blowing out birthday candles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 76, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 128, 144, 145, 146, 147, 148, 149, 150, 151, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 304, 305, 306, 328, 329], "bbox": [0.2183125, 0.1485480093676815, 0.6510781250000001, 0.9686182669789227], "iscrowd": 0, "area": 56061.308650000006, "rle": {"size": [427, 640], "counts": "Rdj14U=5K5J6K5J6K5K6I6K5L4L7I7I8H8I7H8H8H8H8H8H8I7H7I8H8H8H8I6I1O1O2N1O100N2K5L4L4K5L4L4L4K5L4L4K5L4L4L4K5L4L5J5L4L4K5L4L4LnNXLiHc3\\7_LbH`3^7bLaH]3`7dL_HZ3c7hL[HW3e7kLZHT3g7mLXHR3i7PMUHo2k7SMTHl2m7UMRHi2o7ZMoGe2Q8]MnGb2R8`MmG_2S8dMkG[2U8gMjGX2V8jMiGT2X8oMfGP2Z8SNdGl1\\8XNaGg1a8ZN]Ge1d8]N[G`1g8bNWG]1j8fNSGY1n8a11O1O1N2O101O00000O10000000001O0O1000000000000O101O000000000O1000POSLbHm3]7]L[Hb3d7iLRHW3n7SMhGm2W8S10100O1O1O100O00100O1O100O1O011N1O2N101N1O2O0O2N101N1O2O0O2O001O001O0O2O001O001O001O001O001N101O001O001O001O1O001N101O001O001O001O001N101O001O001O001O001O0O2O002N2N2N3M2N2N2N2N3L3N2N2N2N3M2N2N2N3M2N2M3N2N3gKTHm2U901O1O001O001O0O2O1O001O001O001O001O1N101O001O001O1O001O001N101O001O2N1O2N1O2N1O2M2O2N1O2N1M4F9H9F9G:G8G:G8G`hl2"}, "label": "a woman blowing out birthday candles"}]} {"id": 45226, "image": "COCO_train2014_000000045226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing thick glasses blows out the candles on her birthday cake\"."}], "task": "refering", "answer_template": "The \"a woman wearing thick glasses blows out the candles on her birthday cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 76, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 128, 144, 145, 146, 147, 148, 149, 150, 151, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 258, 259, 260, 261, 262, 281, 282, 283, 284, 285, 304, 305, 306, 328, 329], "bbox": [0.2183125, 0.1485480093676815, 0.6510781250000001, 0.9686182669789227], "iscrowd": 0, "area": 56061.308650000006, "rle": {"size": [427, 640], "counts": "Rdj14U=5K5J6K5J6K5K6I6K5L4L7I7I8H8I7H8H8H8H8H8H8I7H7I8H8H8H8I6I1O1O2N1O100N2K5L4L4K5L4L4L4K5L4L4K5L4L4L4K5L4L5J5L4L4K5L4L4LnNXLiHc3\\7_LbH`3^7bLaH]3`7dL_HZ3c7hL[HW3e7kLZHT3g7mLXHR3i7PMUHo2k7SMTHl2m7UMRHi2o7ZMoGe2Q8]MnGb2R8`MmG_2S8dMkG[2U8gMjGX2V8jMiGT2X8oMfGP2Z8SNdGl1\\8XNaGg1a8ZN]Ge1d8]N[G`1g8bNWG]1j8fNSGY1n8a11O1O1N2O101O00000O10000000001O0O1000000000000O101O000000000O1000POSLbHm3]7]L[Hb3d7iLRHW3n7SMhGm2W8S10100O1O1O100O00100O1O100O1O011N1O2N101N1O2O0O2N101N1O2O0O2O001O001O0O2O001O001O001O001O001N101O001O001O001O1O001N101O001O001O001O001N101O001O001O001O001O0O2O002N2N2N3M2N2N2N2N3L3N2N2N2N3M2N2N2N3M2N2M3N2N3gKTHm2U901O1O001O001O0O2O1O001O001O001O001O1N101O001O001O1O001O001N101O001O2N1O2N1O2N1O2M2O2N1O2N1M4F9H9F9G:G8G:G8G`hl2"}, "label": "a woman wearing thick glasses blows out the candles on her birthday cake"}]} {"id": 45226, "image": "COCO_train2014_000000045226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a set of hands is clapping for the girl blowing out candles\"."}], "task": "refering", "answer_template": "The \"a set of hands is clapping for the girl blowing out candles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275], "bbox": [0.82346875, 0.41440281030444964, 1.0, 0.8018032786885246], "iscrowd": 0, "area": 14611.961100000002, "rle": {"size": [427, 640], "counts": "dRl64Q=7I6J6J6L4M3M3L4M3M3L4M3M3M3M3N2N2cNnM_GT2`8SNXGo1f8UNUGn1j8VNQGl1m8YNlFk1S9YNcFn1^9o0001O00O1N2N2N2O2M2N2N2N2N2O100O1O100O1O100O1O1O100O1O101N1O100O1O100O100O1O100O100O100O100O100O100O1000000O100000O10000000000O1000O100000O100000000O1000O10O100000000O2O1O001TNk1PN\\L"}, "label": "a set of hands is clapping for the girl blowing out candles"}]} {"id": 45226, "image": "COCO_train2014_000000045226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man who is about to clap\"."}], "task": "refering", "answer_template": "The \"the man who is about to clap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 160, 180, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275], "bbox": [0.82346875, 0.41440281030444964, 1.0, 0.8018032786885246], "iscrowd": 0, "area": 14611.961100000002, "rle": {"size": [427, 640], "counts": "dRl64Q=7I6J6J6L4M3M3L4M3M3L4M3M3M3M3N2N2cNnM_GT2`8SNXGo1f8UNUGn1j8VNQGl1m8YNlFk1S9YNcFn1^9o0001O00O1N2N2N2O2M2N2N2N2N2O100O1O100O1O100O1O1O100O1O101N1O100O1O100O100O1O100O100O100O100O100O100O1000000O100000O10000000000O1000O100000O100000000O1000O10O100000000O2O1O001TNk1PN\\L"}, "label": "the man who is about to clap"}]} {"id": 151729, "image": "COCO_train2014_000000151729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two little girls interacting with each other on a bench\"."}], "task": "refering", "answer_template": "The \"two little girls interacting with each other on a bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.013472803347280335, 0.998625, 0.988765690376569], "iscrowd": 0, "area": 275792.3718, "rle": {"size": [478, 640], "counts": "Q<7[>>_N\\OXDS1c;oNQD_1j;dNjCj1R in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.013472803347280335, 0.998625, 0.988765690376569], "iscrowd": 0, "area": 275792.3718, "rle": {"size": [478, 640], "counts": "Q<7[>>_N\\OXDS1c;oNQD_1j;dNjCj1R in this image.", "objects": [{"patches": [52, 74, 75, 76, 77, 96, 97, 98, 99, 100, 119, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 282, 283, 284, 286, 287, 288, 305, 306, 307, 308, 310, 311, 312, 329, 330, 331, 333, 334, 335, 351, 352, 353, 356, 357, 358, 374, 375], "bbox": [0.204296875, 0.16144351464435147, 0.60453125, 0.9618828451882846], "iscrowd": 0, "area": 46142.2304, "rle": {"size": [478, 640], "counts": "PYm14g>6J5K6J6J5K6dDkNf8[1TGhNi8^1QGeNl8`1PGbNm8d1mF`No8f1lF\\NQ9i1jFZNS9l1gFWNV9o1eFTNW9R2cFQN[9S2`FPN`9S2[FoMe9T2UFPNi9S2RFPNn9S2mEoMS:^3O2O0O7\\HQKm4T5kJoJU5V5cJmJ\\5Z5[JiJe5Y5XJhJg5[5VJfJj5]5RJdJn5^5oIcJP6`5mIaJS6a5jI`JU6c5hI^JX6e5dI\\J\\6f5aI\\J_O]Od6Z6jIZJ_OAe6W6iIYJ@Dd6V6iIWJAGd6T6hIhJV6[5fIhJj2eN7f6lLhJj2dN:e6iLjJi2cN>e6fLkJh2cNa0[8\\OfGe0[8WOhGh0Z8UOgGl0Z8POiGo0Y8nNhGS1Y8iNjGV1Y8eNiG\\1X8aNiG_1Y8^NiGb1X8]NgGc1[8]NdGc1]8YNeGg1]8UNfGk1[8QNgGo1[8nMgGQ2[8kMgGV2Z8fMiGY2Y8dMhG]2Y8_MjG`2X8\\MjGe2W8WMkGi2W8TMkGl2V8PMkGQ3W8kLkGV3V8gLkGY3l92O000O2O0O2O001N101O0O2O001N10001N101N101O0gLRLaKP4^4QLaKo3_4RL`Ko3^4TL_Km3a4TL^Km3a4UL]Kk3c4VL[Kk3e4VLZKk3d4XLYKk3e4VLYKn3d4SLYKQ4e4QLXKS4e4nKYKV4c4lK[KW4c4kKZKY4c4hK[K[4c4fK[K^4b4dK[K`4a4bK\\Kb4b4_K\\Ke4a4]K\\Kf4b4[K\\Ki4a4XK]Kl4_4WK^Kl4`4UK^Ko4_4RK_KQ5_4QK]KT5_4nJ_KV5^4lJ_KW5_4jJ_KZ5^4gJ`K]5]4eJ`K^5]4dJaK_5]4bJaKa5]4aJ_Kb5`4_J^Kc5a4^J]Kd5a4_J\\Kc5c4`JYKb5f4aJVKa5i4bJSKa5k4aJRKa5l4Z2O1O1O1O1N3N1O1O1N2O1O1O2N1N20000001O01O0000100O1O1O1O1O2O0O1O1O1O1O100O1O1ObFTLj7k3THeL_7Z3`HfLb7Z3\\HgLe7X3ZHhLh7W3VHjLl7U3SHlLn7S3QHmLQ8Q3oGoLR8n2oGTMR8i2oGWMS8e2oG[MS8b2mG`MT8\\2nGdMT8Y2mGhMT8U2mGkMU8Q2lGPNV8m1kGTNV8h1lGXNV8e1jG\\NX8`1TGfMBk0[9\\1SGlM@h0^9Y1TGPN]Og0a9V1RGVNZOf0e9P1SG[NWOe0h9m0QGaNUOb0l9j0oFgNSO`0o9e0PGlNoN`0S:a0nFROmN>V: in this image.", "objects": [{"patches": [52, 74, 75, 76, 77, 96, 97, 98, 99, 100, 119, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 195, 212, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 282, 283, 284, 286, 287, 288, 305, 306, 307, 308, 310, 311, 312, 329, 330, 331, 333, 334, 335, 351, 352, 353, 356, 357, 358, 374, 375], "bbox": [0.204296875, 0.16144351464435147, 0.60453125, 0.9618828451882846], "iscrowd": 0, "area": 46142.2304, "rle": {"size": [478, 640], "counts": "PYm14g>6J5K6J6J5K6dDkNf8[1TGhNi8^1QGeNl8`1PGbNm8d1mF`No8f1lF\\NQ9i1jFZNS9l1gFWNV9o1eFTNW9R2cFQN[9S2`FPN`9S2[FoMe9T2UFPNi9S2RFPNn9S2mEoMS:^3O2O0O7\\HQKm4T5kJoJU5V5cJmJ\\5Z5[JiJe5Y5XJhJg5[5VJfJj5]5RJdJn5^5oIcJP6`5mIaJS6a5jI`JU6c5hI^JX6e5dI\\J\\6f5aI\\J_O]Od6Z6jIZJ_OAe6W6iIYJ@Dd6V6iIWJAGd6T6hIhJV6[5fIhJj2eN7f6lLhJj2dN:e6iLjJi2cN>e6fLkJh2cNa0[8\\OfGe0[8WOhGh0Z8UOgGl0Z8POiGo0Y8nNhGS1Y8iNjGV1Y8eNiG\\1X8aNiG_1Y8^NiGb1X8]NgGc1[8]NdGc1]8YNeGg1]8UNfGk1[8QNgGo1[8nMgGQ2[8kMgGV2Z8fMiGY2Y8dMhG]2Y8_MjG`2X8\\MjGe2W8WMkGi2W8TMkGl2V8PMkGQ3W8kLkGV3V8gLkGY3l92O000O2O0O2O001N101O0O2O001N10001N101N101O0gLRLaKP4^4QLaKo3_4RL`Ko3^4TL_Km3a4TL^Km3a4UL]Kk3c4VL[Kk3e4VLZKk3d4XLYKk3e4VLYKn3d4SLYKQ4e4QLXKS4e4nKYKV4c4lK[KW4c4kKZKY4c4hK[K[4c4fK[K^4b4dK[K`4a4bK\\Kb4b4_K\\Ke4a4]K\\Kf4b4[K\\Ki4a4XK]Kl4_4WK^Kl4`4UK^Ko4_4RK_KQ5_4QK]KT5_4nJ_KV5^4lJ_KW5_4jJ_KZ5^4gJ`K]5]4eJ`K^5]4dJaK_5]4bJaKa5]4aJ_Kb5`4_J^Kc5a4^J]Kd5a4_J\\Kc5c4`JYKb5f4aJVKa5i4bJSKa5k4aJRKa5l4Z2O1O1O1O1N3N1O1O1N2O1O1O2N1N20000001O01O0000100O1O1O1O1O2O0O1O1O1O1O100O1O1ObFTLj7k3THeL_7Z3`HfLb7Z3\\HgLe7X3ZHhLh7W3VHjLl7U3SHlLn7S3QHmLQ8Q3oGoLR8n2oGTMR8i2oGWMS8e2oG[MS8b2mG`MT8\\2nGdMT8Y2mGhMT8U2mGkMU8Q2lGPNV8m1kGTNV8h1lGXNV8e1jG\\NX8`1TGfMBk0[9\\1SGlM@h0^9Y1TGPN]Og0a9V1RGVNZOf0e9P1SG[NWOe0h9m0QGaNUOb0l9j0oFgNSO`0o9e0PGlNoN`0S:a0nFROmN>V: in this image.", "objects": [{"patches": [0, 1, 23, 24, 46, 47, 69, 70, 71, 92, 93, 94, 115, 116, 117, 138, 139, 140, 161, 162, 163, 184, 185, 207, 208, 230, 253, 322, 345, 368], "bbox": [0.0, 0.002510460251046025, 0.11053125, 0.9890585774058577], "iscrowd": 0, "area": 16202.065949999998, "rle": {"size": [478, 640], "counts": "`<_1o6i2[KoMlMn7X2PHhMn7\\2PHdMn7a2nGaMo7c2oG]Mo7g2oGYMn7l2PHTMn7P3PHPMn7U3oGkLo7Y3oGgLo7]3oGcLo7a3oG_Lo7e3oG[Lo7j3mGXLP8l3nGTLP8P4nGPLo7U4oGkKo7Y4oGgKo7^4nGbKP8b4nG^KP8f4nGZKP8j4nGVKP8n4mGTKP8Q5mGoJQ8U5mGkJP8Z5nGfJP8^5nGbJP8b5nG^JP8g5mGYJQ8k5mGUJQ8o5mGQJQ8`6M2N2N2M3N2N2N2N2N6J8H7I8H7H9H7I8H7I8H7I8H7I8G8I8H7I8H7I8H7I8H7I8G8I8H7I8H7I]^Y8"}, "label": "the bench next to the one the kids are sitting on"}]} {"id": 151729, "image": "COCO_train2014_000000151729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bench with no one seen on it\"."}], "task": "refering", "answer_template": "The \"the bench with no one seen on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 46, 47, 69, 70, 71, 92, 93, 94, 115, 116, 117, 138, 139, 140, 161, 162, 163, 184, 185, 207, 208, 230, 253, 322, 345, 368], "bbox": [0.0, 0.002510460251046025, 0.11053125, 0.9890585774058577], "iscrowd": 0, "area": 16202.065949999998, "rle": {"size": [478, 640], "counts": "`<_1o6i2[KoMlMn7X2PHhMn7\\2PHdMn7a2nGaMo7c2oG]Mo7g2oGYMn7l2PHTMn7P3PHPMn7U3oGkLo7Y3oGgLo7]3oGcLo7a3oG_Lo7e3oG[Lo7j3mGXLP8l3nGTLP8P4nGPLo7U4oGkKo7Y4oGgKo7^4nGbKP8b4nG^KP8f4nGZKP8j4nGVKP8n4mGTKP8Q5mGoJQ8U5mGkJP8Z5nGfJP8^5nGbJP8b5nG^JP8g5mGYJQ8k5mGUJQ8o5mGQJQ8`6M2N2N2M3N2N2N2N2N6J8H7I8H7H9H7I8H7I8H7I8H7I8G8I8H7I8H7I8H7I8H7I8G8I8H7I8H7I]^Y8"}, "label": "the bench with no one seen on it"}]} {"id": 151729, "image": "COCO_train2014_000000151729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"little girl in green shorts on the right of the other girl\"."}], "task": "refering", "answer_template": "The \"little girl in green shorts on the right of the other girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 291, 292, 293, 294, 308, 309, 310, 311, 313, 314, 315, 316, 317, 331, 332, 333, 336, 337, 338, 339, 340, 358, 359, 360, 361, 362, 363, 382, 383, 384, 385, 386, 387], "bbox": [0.399515625, 0.1699581589958159, 0.8403749999999999, 1.0], "iscrowd": 0, "area": 54705.86034999999, "rle": {"size": [478, 640], "counts": "ekg3=a>:E2O001N2O1N2O1O1O001O1O1O1O001O1O000000000000000000000nIROTNm0e1^OXNb0`1J\\N6_12^NM`1:\\NFa1`0\\N@a1e0]NZOa1j0^NVO`1n0^NRO_1R1`NmN^1X1`NhN]1]1bNbN[1a1eN^NY1e1gN[NW1g1jNXNT1j1lNUNS1n1lNRNR1P2nNoMP1T2QOkMm0W2SOhMl0Z2TOfMj0]2UObMj0`2UOaMj0`2UOaMj0`2UO`Ml0`2SOaMl0a2SO^Mm0c2RO\\MP1d2oN\\MQ1e2nN[MR1f2mNYMT1i2jNWMW1i2hNWMX1j2fNWMZ1j2eNVM\\1j2bNWM^1k2aJPMj26d2k2ZJ[Mm2Jj2Y4oLhKQ3]4jLcKW3a4dL_K]3f4\\L\\Kc3i4XLWKi3l4RLUKo3n4mKRKT4Q5hKPKW4T5dKmJ]4[5ZKeJg4c5PK]JP5l5fJUJ[5S6\\JmIe5[6QJfIo5d71O1O1O0O2O1O001N2N1O2M3N2N2N2N101O1O1O1O1O1O0O2O1O1O1O1O10O2O1O1O1O2N1N2O1O1O1O1O2N1N2O1O1O1O1O2ULbGR1_8fNiGX1Y8_NPH`1Q8XNWHg1j7PN_Ho1b7iMfHV2[7^MRI`2g9O1N9Ha0_O0001O0000001O000000001O000`G\\Mf4d2XK_Mg4a2XKaMg4`2VKcMi4]2VKeMi4[2UKhMj4X2TKkMk4V2SKmMk4S2SKPNl4P2SKRNl4n1RKUNm4k1RKWNm4j1PKYNP5g1nJ[NQ5e1mJ^NR5c1lJ_NS5b1jJaNU5`1iJcNU5^1hJeNW5\\1gJfNX5[1eJhNZ5Y1dJiN[5X1bJkN]5V1aJlN^5U1_JnN`5R1_JPO`5Q1]JSOa5n0]JTOc4mMWJQ3R1UOf4RNPJn2T1ROk4XNiIl2T1oNR5]NbIj2T1lNY5Y2`JlM\\5Y2]JlM`5Z2YJjMe5[2SJkMi5[2PJiMm5f5N11O2N2N2N2N3M2N2M3N2NO1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O4L4L4L4]NmESNW:e1PFXNS:b1RF\\NR:]1TF`NP:Y1VFdNn9U1WFiNR=M2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1Oec_1"}, "label": "little girl in green shorts on the right of the other girl"}]} {"id": 151729, "image": "COCO_train2014_000000151729.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl with brown hair , green shorts and a white shirt leaning in to give a kiss\"."}], "task": "refering", "answer_template": "The \"a little girl with brown hair , green shorts and a white shirt leaning in to give a kiss\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 271, 287, 288, 289, 290, 291, 292, 293, 294, 308, 309, 310, 311, 313, 314, 315, 316, 317, 331, 332, 333, 336, 337, 338, 339, 340, 358, 359, 360, 361, 362, 363, 382, 383, 384, 385, 386, 387], "bbox": [0.399515625, 0.1699581589958159, 0.8403749999999999, 1.0], "iscrowd": 0, "area": 54705.86034999999, "rle": {"size": [478, 640], "counts": "ekg3=a>:E2O001N2O1N2O1O1O001O1O1O1O001O1O000000000000000000000nIROTNm0e1^OXNb0`1J\\N6_12^NM`1:\\NFa1`0\\N@a1e0]NZOa1j0^NVO`1n0^NRO_1R1`NmN^1X1`NhN]1]1bNbN[1a1eN^NY1e1gN[NW1g1jNXNT1j1lNUNS1n1lNRNR1P2nNoMP1T2QOkMm0W2SOhMl0Z2TOfMj0]2UObMj0`2UOaMj0`2UOaMj0`2UO`Ml0`2SOaMl0a2SO^Mm0c2RO\\MP1d2oN\\MQ1e2nN[MR1f2mNYMT1i2jNWMW1i2hNWMX1j2fNWMZ1j2eNVM\\1j2bNWM^1k2aJPMj26d2k2ZJ[Mm2Jj2Y4oLhKQ3]4jLcKW3a4dL_K]3f4\\L\\Kc3i4XLWKi3l4RLUKo3n4mKRKT4Q5hKPKW4T5dKmJ]4[5ZKeJg4c5PK]JP5l5fJUJ[5S6\\JmIe5[6QJfIo5d71O1O1O0O2O1O001N2N1O2M3N2N2N2N101O1O1O1O1O1O0O2O1O1O1O1O10O2O1O1O1O2N1N2O1O1O1O1O2N1N2O1O1O1O1O2ULbGR1_8fNiGX1Y8_NPH`1Q8XNWHg1j7PN_Ho1b7iMfHV2[7^MRI`2g9O1N9Ha0_O0001O0000001O000000001O000`G\\Mf4d2XK_Mg4a2XKaMg4`2VKcMi4]2VKeMi4[2UKhMj4X2TKkMk4V2SKmMk4S2SKPNl4P2SKRNl4n1RKUNm4k1RKWNm4j1PKYNP5g1nJ[NQ5e1mJ^NR5c1lJ_NS5b1jJaNU5`1iJcNU5^1hJeNW5\\1gJfNX5[1eJhNZ5Y1dJiN[5X1bJkN]5V1aJlN^5U1_JnN`5R1_JPO`5Q1]JSOa5n0]JTOc4mMWJQ3R1UOf4RNPJn2T1ROk4XNiIl2T1oNR5]NbIj2T1lNY5Y2`JlM\\5Y2]JlM`5Z2YJjMe5[2SJkMi5[2PJiMm5f5N11O2N2N2N2N3M2N2M3N2NO1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O4L4L4L4]NmESNW:e1PFXNS:b1RF\\NR:]1TF`NP:Y1VFdNn9U1WFiNR=M2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1Oec_1"}, "label": "a little girl with brown hair , green shorts and a white shirt leaning in to give a kiss"}]} {"id": 356535, "image": "COCO_train2014_000000356535.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small elephant\"."}], "task": "refering", "answer_template": "The \"a small elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 236, 237, 239, 259, 282], "bbox": [0.12775, 0.35962616822429905, 0.44234375000000004, 0.8051168224299065], "iscrowd": 0, "area": 20924.25545, "rle": {"size": [428, 640], "counts": "W^R17U=6I8I6J4K4M2N3L4M3M3L4M2N3L4M3L4M3M1N3N1O2M2O1O2M2O2M2O2N1N3N1N3N1O1N2O1N2O1O1N2O1O1N2O1O1O1O1N2OO1O10O01O100O00100O1O10O1000000O10000000000001O000000000O10001O000O10000002N2O0eGiL_6Y3PIZMl6h4K2NO10001O001O00001O001O001O00001O0O2_OZHZKf7c4^H\\Kc7`4bH^K^7`4eH_K[7^4iHaKX7\\4kHcKU7V4TIhKm6R4YImKg6P4^InKc6o3^1N3L3N2M3M4L3M3L4M4K4J6J7J5001O010O0010O01O0010O01O1O010O1O100O1O1O100O1O100O1O100O2N3N2M2OO01O100O1O100O10O00001O000O2O2N1O2N1O1O2M2O2M2O2M2O2M3N5J6K5J6I in this image.", "objects": [{"patches": [118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 146, 147, 164, 165, 166, 167, 168, 169, 170, 187, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 236, 237, 239, 259, 282], "bbox": [0.12775, 0.35962616822429905, 0.44234375000000004, 0.8051168224299065], "iscrowd": 0, "area": 20924.25545, "rle": {"size": [428, 640], "counts": "W^R17U=6I8I6J4K4M2N3L4M3M3L4M2N3L4M3L4M3M1N3N1O2M2O1O2M2O2M2O2N1N3N1N3N1O1N2O1N2O1O1N2O1O1N2O1O1O1O1N2OO1O10O01O100O00100O1O10O1000000O10000000000001O000000000O10001O000O10000002N2O0eGiL_6Y3PIZMl6h4K2NO10001O001O00001O001O001O00001O0O2_OZHZKf7c4^H\\Kc7`4bH^K^7`4eH_K[7^4iHaKX7\\4kHcKU7V4TIhKm6R4YImKg6P4^InKc6o3^1N3L3N2M3M4L3M3L4M4K4J6J7J5001O010O0010O01O0010O01O1O010O1O100O1O1O100O1O100O1O100O2N3N2M2OO01O100O1O100O10O00001O000O2O2N1O2N1O1O2M2O2M2O2M2O2M3N5J6K5J6I in this image.", "objects": [{"patches": [60, 61, 62, 79, 80, 81, 82, 83, 84, 85, 86, 87, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 170, 171, 174, 175, 176, 177, 178, 179, 180, 181, 194, 197, 198, 199, 200, 201, 202, 203, 204, 217, 220, 221, 225, 226, 227, 240, 243, 244, 248, 249, 250, 265, 266, 267, 270, 271, 272, 273], "bbox": [0.423796875, 0.1844158878504673, 0.907703125, 0.8046495327102804], "iscrowd": 0, "area": 50138.43385, "rle": {"size": [428, 640], "counts": "i^a32T=;F9G:F9F9H6Ja0_Oa0^O9H7I8H7H8XOi0A>M9G9H8G:F8I2MVM\\IQO`6o0gImNW6R1QJiNl5W1[JeNc5Z1dJ`N[5_1nJ[NQ5d1VKVNj4j1\\KQNc4n1eKSMS5m2SKbLQOHS5e3YM\\Lf2b3\\M^Ld2`3^MaLa2^3`MbL`2\\3bMeL]2Y3eMgL[2W3gMjLX2T3jMlLV2T3Q401O00000000000O10000O10000O100O2O000O100O10000O10000001O00000010O0001N2O001O1mIkLm1V3RNkLm1U3SNlLl1U3SNmLk1T3TNmLk1T3TNmLk1S3UNmLk1T3TNlLl1U3SNlLl1U3SNkLm1V3bKkLBL54g4V3XKWMED:0h4V3oJbMH_O=Jl4V3eJnMJYOa0DP5V5XJ\\Kc0AT5U6hJlIX5W6dJjI\\5X6`JjI_5Z6]JhIb5Q70001O0000001O00001O000000O1000000000O10000000\\N_J]K`5]4QKXKP5a4aKUK_4e4QLQKo3h4[LUKe3d4dLZK\\3a4lL\\KS3e4PMXKP3h4TMTKl2l4j2O10O10O1O1O1N101N2O1N2O1N110001O000O10001O000O10001O000000010O000000001O01O000001O000001O0001O000000010O000000001O01O000001O000001O01O0000000010O00000001O0001O0001O000dJUKW2k4hMVKX2j4gMWKZ2i4cMYK]2g4bMZK_2e4`M\\K`2d4^M^Kc2a4\\M`Kd2a4ZM_Kh2a4UMaKk2b4QM_KP3c4lL^KT3e4gL]KZ3d4cL]K]3d4aL]K`3c4]L_Kd3a4ZL`Kf3a4XL_Kj3a4SLaKm3`4kJ]K14U5_4cJmL]5T3[JSMf5Y5001O001N101O1O001O001O001N101O001O1O001O001N101O001O001O1O1O1N2O1O1O001O1O1O1O1N2O1O1O1O1O001O1O1N2O1O1O1aLjJfNW5W1WK]Nj4a1bKTN_4i1c3N4J7G8IcGZN`5`1aJbNb5W1\\JPOf5k0_I2d66THh0o7X23L4G9D in this image.", "objects": [{"patches": [60, 61, 62, 79, 80, 81, 82, 83, 84, 85, 86, 87, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 170, 171, 174, 175, 176, 177, 178, 179, 180, 181, 194, 197, 198, 199, 200, 201, 202, 203, 204, 217, 220, 221, 225, 226, 227, 240, 243, 244, 248, 249, 250, 265, 266, 267, 270, 271, 272, 273], "bbox": [0.423796875, 0.1844158878504673, 0.907703125, 0.8046495327102804], "iscrowd": 0, "area": 50138.43385, "rle": {"size": [428, 640], "counts": "i^a32T=;F9G:F9F9H6Ja0_Oa0^O9H7I8H7H8XOi0A>M9G9H8G:F8I2MVM\\IQO`6o0gImNW6R1QJiNl5W1[JeNc5Z1dJ`N[5_1nJ[NQ5d1VKVNj4j1\\KQNc4n1eKSMS5m2SKbLQOHS5e3YM\\Lf2b3\\M^Ld2`3^MaLa2^3`MbL`2\\3bMeL]2Y3eMgL[2W3gMjLX2T3jMlLV2T3Q401O00000000000O10000O10000O100O2O000O100O10000O10000001O00000010O0001N2O001O1mIkLm1V3RNkLm1U3SNlLl1U3SNmLk1T3TNmLk1T3TNmLk1S3UNmLk1T3TNlLl1U3SNlLl1U3SNkLm1V3bKkLBL54g4V3XKWMED:0h4V3oJbMH_O=Jl4V3eJnMJYOa0DP5V5XJ\\Kc0AT5U6hJlIX5W6dJjI\\5X6`JjI_5Z6]JhIb5Q70001O0000001O00001O000000O1000000000O10000000\\N_J]K`5]4QKXKP5a4aKUK_4e4QLQKo3h4[LUKe3d4dLZK\\3a4lL\\KS3e4PMXKP3h4TMTKl2l4j2O10O10O1O1O1N101N2O1N2O1N110001O000O10001O000O10001O000000010O000000001O01O000001O000001O0001O000000010O000000001O01O000001O000001O01O0000000010O00000001O0001O0001O000dJUKW2k4hMVKX2j4gMWKZ2i4cMYK]2g4bMZK_2e4`M\\K`2d4^M^Kc2a4\\M`Kd2a4ZM_Kh2a4UMaKk2b4QM_KP3c4lL^KT3e4gL]KZ3d4cL]K]3d4aL]K`3c4]L_Kd3a4ZL`Kf3a4XL_Kj3a4SLaKm3`4kJ]K14U5_4cJmL]5T3[JSMf5Y5001O001N101O1O001O001O001N101O001O1O001O001N101O001O001O1O1O1N2O1O1O001O1O1O1O1N2O1O1O1O1O001O1O1N2O1O1O1aLjJfNW5W1WK]Nj4a1bKTN_4i1c3N4J7G8IcGZN`5`1aJbNb5W1\\JPOf5k0_I2d66THh0o7X23L4G9D in this image.", "objects": [{"patches": [54, 55, 69, 70, 71, 72, 81, 82, 83, 84, 85, 86, 87, 97, 98, 99, 100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 120, 132, 133, 134, 135, 136, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 180, 181, 182, 183, 184, 185, 196, 197, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 228, 229, 233, 234, 235, 244, 245, 250, 251, 252, 259, 260, 267, 268, 275, 276, 284, 285, 291, 292, 300, 301], "bbox": [0.10716894977168949, 0.151078125, 0.8334246575342467, 0.8108281250000001], "iscrowd": 0, "area": 42023.30974999999, "rle": {"size": [640, 438], "counts": "Tam02mc02N3M2M4M3M2N3M1O1O1O101O000O1000000O10000O104L4L5J1000000000000001O00000000001O00000OWGiNkNW1>iNWG9Z8n0=CD=9FE<7GI95JI83KM51MO4NO02M130J35NH47MF79JD8G_O;a0G[O;e0FWO>g0EoNc0Q1^OeNl0Z1VO[NS1d1oNSNZ1l1hNQNY1o1hNoMX1Q2jNlMX1T2jNiMV1X2kNfMV1Z2lNcMT1]2nNaMS1_2BkL>V3?lKBT4g0bKZO^4g0`KYOa4g0^KZOb4g0TKXOaLgL[8S4jJ^OmL^LY8T4YHfKa1R4nMRLY8W4THWLf0P4nNaKX8Y4QHR2HcIX8\\4lGT2M_IW8]4jGV2^8kM_Gf1NXJc8T4[G`1 in this image.", "objects": [{"patches": [54, 55, 69, 70, 71, 72, 81, 82, 83, 84, 85, 86, 87, 97, 98, 99, 100, 101, 102, 103, 104, 115, 116, 117, 118, 119, 120, 132, 133, 134, 135, 136, 148, 149, 150, 151, 152, 164, 165, 166, 167, 168, 180, 181, 182, 183, 184, 185, 196, 197, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 228, 229, 233, 234, 235, 244, 245, 250, 251, 252, 259, 260, 267, 268, 275, 276, 284, 285, 291, 292, 300, 301], "bbox": [0.10716894977168949, 0.151078125, 0.8334246575342467, 0.8108281250000001], "iscrowd": 0, "area": 42023.30974999999, "rle": {"size": [640, 438], "counts": "Tam02mc02N3M2M4M3M2N3M1O1O1O101O000O1000000O10000O104L4L5J1000000000000001O00000000001O00000OWGiNkNW1>iNWG9Z8n0=CD=9FE<7GI95JI83KM51MO4NO02M130J35NH47MF79JD8G_O;a0G[O;e0FWO>g0EoNc0Q1^OeNl0Z1VO[NS1d1oNSNZ1l1hNQNY1o1hNoMX1Q2jNlMX1T2jNiMV1X2kNfMV1Z2lNcMT1]2nNaMS1_2BkL>V3?lKBT4g0bKZO^4g0`KYOa4g0^KZOb4g0TKXOaLgL[8S4jJ^OmL^LY8T4YHfKa1R4nMRLY8W4THWLf0P4nNaKX8Y4QHR2HcIX8\\4lGT2M_IW8]4jGV2^8kM_Gf1NXJc8T4[G`1 in this image.", "objects": [{"patches": [19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 88, 89, 90, 111, 112, 113, 133, 134, 135, 136, 156, 157, 158], "bbox": [0.809234375, 0.004309133489461358, 0.975671875, 0.44730679156908665], "iscrowd": 0, "area": 14089.067750000002, "rle": {"size": [427, 640], "counts": "VRh62U=5L3]FIn5kIFQ6?jIES6?iIDS6`0iIDS6a0hICW6?dIFZ6 in this image.", "objects": [{"patches": [19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 88, 89, 90, 111, 112, 113, 133, 134, 135, 136, 156, 157, 158], "bbox": [0.809234375, 0.004309133489461358, 0.975671875, 0.44730679156908665], "iscrowd": 0, "area": 14089.067750000002, "rle": {"size": [427, 640], "counts": "VRh62U=5L3]FIn5kIFQ6?jIES6?iIDS6`0iIDS6a0hICW6?dIFZ6 in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 62, 63, 64, 65, 66, 88, 111, 134, 157, 180, 181, 203, 204, 226, 227, 249, 250, 272, 273, 295, 296], "bbox": [0.359828125, 0.0038173302107728335, 1.0, 0.8510070257611241], "iscrowd": 0, "area": 23873.270899999996, "rle": {"size": [427, 640], "counts": "lgP31Z=001O001O001O001O001O001O001O0O2O001O001O001O001O001O00001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O0000001O000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000O10001O000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O00000000001O000000001O000000001O0000000000>B\\4dKm3SLNkMTEo1j:mM\\ES2n:000001O000000001O000000001OmFkNP6U1oINo42PKP1o3POPLR2o2nMPMT3P2kLoMW4P1iKoNY50gJO[6POeIo0\\9O1O1O1O1O1O1O100O10000O10000O10000O10000O100O10000O10000O10000O10000O100O10000O10000O10000O10000O100O100O1O1N2O1N2N2O1N2O1N2O1N2O1N2O1N2O1N2N2O1N2O1N2OZ="}, "label": "umbrella stand with the number 2"}]} {"id": 159957, "image": "COCO_train2014_000000159957.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umbrella that the two people are standing beneath\"."}], "task": "refering", "answer_template": "The \"the umbrella that the two people are standing beneath\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 29, 30, 31, 32, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 122, 123, 145, 146, 169, 192, 215, 237, 238, 239, 260, 261, 284, 307], "bbox": [0.060781249999999995, 0.0020843091334894616, 0.6408906249999999, 0.8957845433255269], "iscrowd": 0, "area": 28064.505, "rle": {"size": [427, 640], "counts": "S\\`03T=6K5J6N1N10001N10000O2O000O101O0O10001N100O101O0O10000O2O000O101O0O100O10000O10001N10000O10000O100O10000O10000O10000O101O0O100O10000O10000O10000O10000O10000O100O2O000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O10O1000O10000O1000000O1000O010000O1000000O10000O0100000O100N2M3N2N2N2N2M2O2N2N2N2N2M3N2N2N2N2M3N2gK]LhNe3W1]LfNe3n0YLXK6f3c3R1WLWK8d3c3U1ULVK:a3d3Y1RLUK<_3d3\\1PLTK>^3c3^1oKSK`0^3b3^1nKSKb0]3a3`1mKRKd0]3_3a1eL^N[3b1gL\\NY3d1hL[NX3e1iLZNW3e1kLYNV3g1kLXNU3h1lLWNT3i1mLVNS3j1nLUN_3^1cL`NW4f0jKYOP5MQK1j5UOWJj0c6\\N^Ic1U7jMlHU2W7hMPIQ2S7lMZIh1e6XNhI]1U6dNWJQ1\\1_N7?_NR1Y1WO@GWOT1W1MiNPO0T1V1d0RNXNh0V1T1Y1]MaM^1X1S1P2fLiLW2Y1Q1R3oNoLP1Q3POPMo0P3kMUKAm1b2n2hMQLnNS1Y3l2eMdNY2\\1bMjN]2V1bMlN\\2U1cMnN[2Q1dMRO[2n0dMUOZ2k0eMWOZ2i0dMZO[2f0dM]OZ2c0eM_OZ2a0oLTK?_4`2=QMWK;^4b2 in this image.", "objects": [{"patches": [7, 8, 29, 30, 31, 32, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 122, 123, 145, 146, 169, 192, 215, 237, 238, 239, 260, 261, 284, 307], "bbox": [0.060781249999999995, 0.0020843091334894616, 0.6408906249999999, 0.8957845433255269], "iscrowd": 0, "area": 28064.505, "rle": {"size": [427, 640], "counts": "S\\`03T=6K5J6N1N10001N10000O2O000O101O0O10001N100O101O0O10000O2O000O101O0O100O10000O10001N10000O10000O100O10000O10000O10000O101O0O100O10000O10000O10000O10000O10000O100O2O000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O10O1000O10000O1000000O1000O010000O1000000O10000O0100000O100N2M3N2N2N2N2M2O2N2N2N2N2M3N2N2N2N2M3N2gK]LhNe3W1]LfNe3n0YLXK6f3c3R1WLWK8d3c3U1ULVK:a3d3Y1RLUK<_3d3\\1PLTK>^3c3^1oKSK`0^3b3^1nKSKb0]3a3`1mKRKd0]3_3a1eL^N[3b1gL\\NY3d1hL[NX3e1iLZNW3e1kLYNV3g1kLXNU3h1lLWNT3i1mLVNS3j1nLUN_3^1cL`NW4f0jKYOP5MQK1j5UOWJj0c6\\N^Ic1U7jMlHU2W7hMPIQ2S7lMZIh1e6XNhI]1U6dNWJQ1\\1_N7?_NR1Y1WO@GWOT1W1MiNPO0T1V1d0RNXNh0V1T1Y1]MaM^1X1S1P2fLiLW2Y1Q1R3oNoLP1Q3POPMo0P3kMUKAm1b2n2hMQLnNS1Y3l2eMdNY2\\1bMjN]2V1bMlN\\2U1cMnN[2Q1dMRO[2n0dMUOZ2k0eMWOZ2i0dMZO[2f0dM]OZ2c0eM_OZ2a0oLTK?_4`2=QMWK;^4b2 in this image.", "objects": [{"patches": [142, 143, 165, 166, 167, 188, 189, 190, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 259, 280, 281, 282, 303, 304, 305], "bbox": [0.184859375, 0.4186651053864169, 0.312984375, 0.9085480093676817], "iscrowd": 0, "area": 9355.873050000002, "rle": {"size": [427, 640], "counts": "[]a12V=6K5L4M4mG^OT3g0fL\\OY3g0aJYOA1n5h0ZJ]O^N:NDY7i0TJ^ObNQ1Z7DSI:_O6g6_N_IT14 in this image.", "objects": [{"patches": [142, 143, 165, 166, 167, 188, 189, 190, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 259, 280, 281, 282, 303, 304, 305], "bbox": [0.184859375, 0.4186651053864169, 0.312984375, 0.9085480093676817], "iscrowd": 0, "area": 9355.873050000002, "rle": {"size": [427, 640], "counts": "[]a12V=6K5L4M4mG^OT3g0fL\\OY3g0aJYOA1n5h0ZJ]O^N:NDY7i0TJ^ObNQ1Z7DSI:_O6g6_N_IT14 in this image.", "objects": [{"patches": [151, 173, 174, 175, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 265, 266, 267, 268, 288, 289, 290, 291, 311, 312, 313], "bbox": [0.54425, 0.4248711943793911, 0.686671875, 0.8967681498829039], "iscrowd": 0, "area": 11425.044050000002, "rle": {"size": [427, 640], "counts": "Q]a42R<2aDb0W:ARF\\1\\9XO[Fo0`9WOYFn0;VN9LS7U1QHP13_Nc0]OY7^4bHcK^7b4\\H_Ke7e4UH\\Kk7i4oGXKP8R50OZOSKoHn4o6UKPIk4n6XKPIi4n6ZKQIf4n6\\KQIo2o0kLo58PIf2`1kL`5`0oH]2m1oLPNMQ7i3\\KWLd4a3cK`L\\4Z3jKgLW4[3eKfL[4^3aKbL_4a3]K`Lc4c3YK^Lg4f3UKZLk4i3QKXLo4k3mJVLS5n3iJRLW5Q4eJoK]5S4_JmKd5T4XJlKj5k4^ITKe6c0eHY4^7cKdH]4]7PKTIP5d7O1O001O10O2O000O10000O10N101N101N2O1N2O2O1N1O2O0O2N5L5J5K4M3L4L5L3L4L4M3L5K4L4M3L5K4M3L4L6K5I7SO\\DHj;1YDMm;LWD0o;JTD4R in this image.", "objects": [{"patches": [151, 173, 174, 175, 196, 197, 198, 199, 219, 220, 221, 222, 242, 243, 244, 245, 265, 266, 267, 268, 288, 289, 290, 291, 311, 312, 313], "bbox": [0.54425, 0.4248711943793911, 0.686671875, 0.8967681498829039], "iscrowd": 0, "area": 11425.044050000002, "rle": {"size": [427, 640], "counts": "Q]a42R<2aDb0W:ARF\\1\\9XO[Fo0`9WOYFn0;VN9LS7U1QHP13_Nc0]OY7^4bHcK^7b4\\H_Ke7e4UH\\Kk7i4oGXKP8R50OZOSKoHn4o6UKPIk4n6XKPIi4n6ZKQIf4n6\\KQIo2o0kLo58PIf2`1kL`5`0oH]2m1oLPNMQ7i3\\KWLd4a3cK`L\\4Z3jKgLW4[3eKfL[4^3aKbL_4a3]K`Lc4c3YK^Lg4f3UKZLk4i3QKXLo4k3mJVLS5n3iJRLW5Q4eJoK]5S4_JmKd5T4XJlKj5k4^ITKe6c0eHY4^7cKdH]4]7PKTIP5d7O1O001O10O2O000O10000O10N101N101N2O1N2O2O1N1O2O0O2N5L5J5K4M3L4L5L3L4L4M3L5K4L4M3L5K4M3L4L6K5I7SO\\DHj;1YDMm;LWD0o;JTD4R in this image.", "objects": [{"patches": [22, 44, 45, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114], "bbox": [0.74065625, 0.01953161592505855, 1.0, 0.29370023419203745], "iscrowd": 0, "area": 9208.783249999999, "rle": {"size": [427, 640], "counts": "WhU62V=4O0O2O001N2O0O2O001N101N101O0O2O0O2O001N101N10000O100O10000O100O10000O100O10000O2O0O10000O100O10000O100O10000O100O10000O2O0O10000O100O10000O100O10000O100O10000O101N10000O100O10000O100O10000O100O100O100O1O2N1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O2O0O1O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O2N1O100O1O1O1O1O100O1O1O1O100O1O1O1OH8ROo0QOn0SOe0"}, "label": "the rightmost sun blocker"}]} {"id": 159957, "image": "COCO_train2014_000000159957.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hat which is at the right side of the other two hats\"."}], "task": "refering", "answer_template": "The \"a hat which is at the right side of the other two hats\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 44, 45, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114], "bbox": [0.74065625, 0.01953161592505855, 1.0, 0.29370023419203745], "iscrowd": 0, "area": 9208.783249999999, "rle": {"size": [427, 640], "counts": "WhU62V=4O0O2O001N2O0O2O001N101N101O0O2O0O2O001N101N10000O100O10000O100O10000O100O10000O2O0O10000O100O10000O100O10000O100O10000O2O0O10000O100O10000O100O10000O100O10000O101N10000O100O10000O100O10000O100O100O100O1O2N1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O2O0O1O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O2N1O100O1O1O1O1O100O1O1O1O100O1O1O1OH8ROo0QOn0SOe0"}, "label": "a hat which is at the right side of the other two hats"}]} {"id": 176342, "image": "COCO_train2014_000000176342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a purple tie\"."}], "task": "refering", "answer_template": "The \"a man wearing a purple tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 55, 56, 57, 77, 78, 79, 80, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 190, 191, 192, 214, 239, 240, 262, 263, 264, 285, 286], "bbox": [0.282640625, 0.09424528301886792, 0.520421875, 0.8391745283018868], "iscrowd": 0, "area": 21603.936, "rle": {"size": [424, 640], "counts": "nT[24o<7J5K6I7J5J7J6J5J7J6I6K6J6I7J5J7J6J5J7K5M2O2M3M2N3M3M2N3M3M2N3M2NO1N2O1O1O2N2O001N2O1O1N2O1O0O2O1O1N2O1O10O01O100O1O100O00100O1O100O1O01\\HYLg5f3ZJ`L`5_3aJhLX5X3hJoLP5Q3QKVMg4j2ZK\\M_4e2aKbMW4^2jKiMn3W2SLjMk3W2ULjMi3V2XLkMf3U2[LlMc3U2]LlMa3T2aLlM]3T2dLlM[3U2eLlMY3T2hLmM[2n2fMSMW2o2iMRMU2n2lMSMQ2n2PNSMm1o2SNQMl1o2UNRMh1o2YNRMd1P3\\NQMb1o2_NRM^1o2cNRMZ1P3fNQMX1o2iNQMU1P3lNPMS1Q3mNoLS1P3nNPMR1o2oNQMP1P3POPMP1o2QOQMo0n2RORMm0o2SOQMm0n2TORMl0m2UOSMj0n2WOQMi0n2XORMh0m2YOSMf0n2ZORMf0m2[OSMf0k2[OUMe0k2[OUMe0j2\\OVMe0h2\\OXMd0h2\\OXMd0g2]OYMd0f2\\OZMd0e2^OZMb0f2^OZMc0d2DVM>h2IQM8m2OmL3Q33hLOV38_LOn0SLd0T4SN0U1oKf0X4jMO^1kKg0k8UOXGi0k8SOWGm0j8QOXGm0k8oNXGP1j8cNbG[1`8QNTHn1d9N2N101N2O1N2N2O5J6J7J5J6K6I6J7J5JZnn3"}, "label": "a man wearing a purple tie"}]} {"id": 176342, "image": "COCO_train2014_000000176342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black suit\"."}], "task": "refering", "answer_template": "The \"a man wearing a black suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 55, 56, 57, 77, 78, 79, 80, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 190, 191, 192, 214, 239, 240, 262, 263, 264, 285, 286], "bbox": [0.282640625, 0.09424528301886792, 0.520421875, 0.8391745283018868], "iscrowd": 0, "area": 21603.936, "rle": {"size": [424, 640], "counts": "nT[24o<7J5K6I7J5J7J6J5J7J6I6K6J6I7J5J7J6J5J7K5M2O2M3M2N3M3M2N3M3M2N3M2NO1N2O1O1O2N2O001N2O1O1N2O1O0O2O1O1N2O1O10O01O100O1O100O00100O1O100O1O01\\HYLg5f3ZJ`L`5_3aJhLX5X3hJoLP5Q3QKVMg4j2ZK\\M_4e2aKbMW4^2jKiMn3W2SLjMk3W2ULjMi3V2XLkMf3U2[LlMc3U2]LlMa3T2aLlM]3T2dLlM[3U2eLlMY3T2hLmM[2n2fMSMW2o2iMRMU2n2lMSMQ2n2PNSMm1o2SNQMl1o2UNRMh1o2YNRMd1P3\\NQMb1o2_NRM^1o2cNRMZ1P3fNQMX1o2iNQMU1P3lNPMS1Q3mNoLS1P3nNPMR1o2oNQMP1P3POPMP1o2QOQMo0n2RORMm0o2SOQMm0n2TORMl0m2UOSMj0n2WOQMi0n2XORMh0m2YOSMf0n2ZORMf0m2[OSMf0k2[OUMe0k2[OUMe0j2\\OVMe0h2\\OXMd0h2\\OXMd0g2]OYMd0f2\\OZMd0e2^OZMb0f2^OZMc0d2DVM>h2IQM8m2OmL3Q33hLOV38_LOn0SLd0T4SN0U1oKf0X4jMO^1kKg0k8UOXGi0k8SOWGm0j8QOXGm0k8oNXGP1j8cNbG[1`8QNTHn1d9N2N101N2O1N2N2O5J6J7J5J6K6I6J7J5JZnn3"}, "label": "a man wearing a black suit"}]} {"id": 176342, "image": "COCO_train2014_000000176342.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing cream color dress and cutting cake with a men\"."}], "task": "refering", "answer_template": "The \"a woman wearing cream color dress and cutting cake with a men\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 103, 104, 105, 106, 126, 127, 128, 129, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 287, 288, 289, 290], "bbox": [0.3394375, 0.22695754716981134, 0.647609375, 0.8359433962264151], "iscrowd": 0, "area": 26166.44815, "rle": {"size": [424, 640], "counts": "]Sj23S=8H6J3M2N2N2N3M2O100O2O000O101O000O2O001N101O001N101O000O10O10O100O100O1O001O1O100O1O001O100O100O1O010O100O1O100O100O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O001O1O1O1O1O1O1O100O10001N100O10001N10000QGXM]7i2jGZNl7f1oGhNg7Z1RHlNn7T1mGQOS8R300O10001jMoJSLQ5l3UKnKl4Q4YKkKg4S4`KhK`4W4eKeK[4Z4kK`KV4^4PL^KP4a4UL[Kk3d4[LWKf3g4^LUKc3k4`LRK`3m4eLoJ[3Q5hLiJ[3W5hLbJ\\3]5X200O2O0O10000O100O100O100O10001N1000000000000000000000001O00000001O001O00001O001O001O001O001O001O0]JkJS3V5gLPKX3Q5fLQKY3P5eLRKZ3o4dLSK[3n4cLTK\\3m4bLUK]3l4aLWK]3j4aLXK^3i4aLXK^3i4`LYKf1RO`Nf5HZKf1SO`Nd5H[Ke1VO`Na5G]Kf1UOaN_5F`Kf1TObN^5DbKg1TObN[5DeKg1SOcNe5WO\\KR2SOeNR8Q1RHlN^8a0gG]Ol8NYGOo8FVG8l8DXG:k8AZG in this image.", "objects": [{"patches": [80, 81, 82, 103, 104, 105, 106, 126, 127, 128, 129, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 287, 288, 289, 290], "bbox": [0.3394375, 0.22695754716981134, 0.647609375, 0.8359433962264151], "iscrowd": 0, "area": 26166.44815, "rle": {"size": [424, 640], "counts": "]Sj23S=8H6J3M2N2N2N3M2O100O2O000O101O000O2O001N101O001N101O000O10O10O100O100O1O001O1O100O1O001O100O100O1O010O100O1O100O100O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O001O1O1O1O1O1O1O100O10001N100O10001N10000QGXM]7i2jGZNl7f1oGhNg7Z1RHlNn7T1mGQOS8R300O10001jMoJSLQ5l3UKnKl4Q4YKkKg4S4`KhK`4W4eKeK[4Z4kK`KV4^4PL^KP4a4UL[Kk3d4[LWKf3g4^LUKc3k4`LRK`3m4eLoJ[3Q5hLiJ[3W5hLbJ\\3]5X200O2O0O10000O100O100O100O10001N1000000000000000000000001O00000001O001O00001O001O001O001O001O001O0]JkJS3V5gLPKX3Q5fLQKY3P5eLRKZ3o4dLSK[3n4cLTK\\3m4bLUK]3l4aLWK]3j4aLXK^3i4aLXK^3i4`LYKf1RO`Nf5HZKf1SO`Nd5H[Ke1VO`Na5G]Kf1UOaN_5F`Kf1TObN^5DbKg1TObN[5DeKg1SOcNe5WO\\KR2SOeNR8Q1RHlN^8a0gG]Ol8NYGOo8FVG8l8DXG:k8AZG in this image.", "objects": [{"patches": [146, 147, 148, 149, 161, 162, 163, 164, 176, 177, 178, 179, 192, 193, 194, 207, 208, 209, 222, 223, 224, 238, 239], "bbox": [0.7655269320843091, 0.40014062499999997, 1.0, 0.672796875], "iscrowd": 0, "area": 10996.798050000001, "rle": {"size": [640, 427], "counts": "Te\\67cc09J5L5K4L3M2N3M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1O1O1O1O001O1O1O1O1N2O1O1O1U@`MT>a2hAcMW>^2eAeM[>\\2aAgM_>Z2]AiMc>X2XAlMh>U2TAoMk>Q2RARNn>o1n@TNR?m1j@TNY?l1`@VNd?P33M3L4M4L3M4K4N31N101N2N1O7I:F9F;F9G9G0000000000000O1000000000000000000O100000001O001O00001N10001O00^Ob0RNn1RNmI"}, "label": "elephant in rear of procession"}]} {"id": 127214, "image": "COCO_train2014_000000127214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an african elephant walking behind the group of elephants\"."}], "task": "refering", "answer_template": "The \"an african elephant walking behind the group of elephants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 149, 161, 162, 163, 164, 176, 177, 178, 179, 192, 193, 194, 207, 208, 209, 222, 223, 224, 238, 239], "bbox": [0.7655269320843091, 0.40014062499999997, 1.0, 0.672796875], "iscrowd": 0, "area": 10996.798050000001, "rle": {"size": [640, 427], "counts": "Te\\67cc09J5L5K4L3M2N3M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2O1O1O1O1O001O1O1O1O1N2O1O1O1U@`MT>a2hAcMW>^2eAeM[>\\2aAgM_>Z2]AiMc>X2XAlMh>U2TAoMk>Q2RARNn>o1n@TNR?m1j@TNY?l1`@VNd?P33M3L4M4L3M4K4N31N101N2N1O7I:F9F;F9G9G0000000000000O1000000000000000000O100000001O001O00001N10001O00^Ob0RNn1RNmI"}, "label": "an african elephant walking behind the group of elephants"}]} {"id": 127214, "image": "COCO_train2014_000000127214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant that is most clearly visible in the picture\"."}], "task": "refering", "answer_template": "The \"the elephant that is most clearly visible in the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 127, 136, 137, 138, 139, 140, 141, 142, 143, 144, 151, 152, 153, 154, 155, 156, 157, 158, 159, 166, 167, 168, 169, 170, 171, 172, 173, 174, 181, 182, 183, 184, 185, 186, 187, 188, 189, 196, 197, 198, 199, 200, 201, 202, 203, 204, 211, 212, 213, 214, 215, 216, 217, 218, 219, 226, 227, 228, 229, 230, 231, 232, 233, 234, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 255, 256, 259, 260, 261, 262, 263, 264, 270, 271, 274, 275, 276, 277, 278, 285, 289, 290, 291, 292, 293, 300, 304, 305, 306, 307, 308, 319, 320, 322, 323], "bbox": [0.0, 0.358578125, 0.645456674473068, 0.9366093750000001], "iscrowd": 0, "area": 66416.16700000003, "rle": {"size": [640, 427], "counts": "Q`0h1Ub02N2N2O1N2N2N2N2N2O0O2N2N2N2N2O1N2N2N2N2N2O1N2N2N1O2O0O001O1O1K4L5K5K5K4L5K5K5K5K4L5K5J6K4L5K5oNQ11N101O1O0O2O1O1O001N2O1O001N2O001O1N2O002N1O1N3N1O1N3M2N2N3M2N2M3N3M2N2N3M2N2N3L300100O010000O1000O01005K7I7I8H6J5K6J6J6I6K6J6J6J5K6J6WGYHm4l7lJVHT5P8dJRH\\5T8\\JnGc5Y8UJiGk5\\8oIeGQ6a8gIaGY6e8_I]Ga6i8WIYGi6l8PIVGP7P9hHRGX7T9^HPGb7V9SHoFm7V9iGoFW8]:O100000000000000000000000000000000000000O10O1000000WLcGoK]8R4YHWKg7j4jHdJV7\\5PI]JQ7d5TIVJl6k5YIoIg6Q6_IiIa6X6dIbI\\6_6iI[IW6e6nIVIR6k6SJnHn5S7WJgHi5Y7]JaHc5`7bJZH^5g7gJSHX5n7]31O10O01O1O001O4L4L4L4L4L4M3L4L4L4L4L5K4L4L4L4L5L3L4L4L4L4L4L4L001O01O000001O0000000100O1O001O100O1O1O1O100O1O001O100O1O1O1O100O001O1O100M3N2N2N2N2N1N3N2N2N2N2J5J7I7I7I6^HoDh5_;UJeDn4W in this image.", "objects": [{"patches": [122, 123, 124, 125, 126, 127, 136, 137, 138, 139, 140, 141, 142, 143, 144, 151, 152, 153, 154, 155, 156, 157, 158, 159, 166, 167, 168, 169, 170, 171, 172, 173, 174, 181, 182, 183, 184, 185, 186, 187, 188, 189, 196, 197, 198, 199, 200, 201, 202, 203, 204, 211, 212, 213, 214, 215, 216, 217, 218, 219, 226, 227, 228, 229, 230, 231, 232, 233, 234, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 255, 256, 259, 260, 261, 262, 263, 264, 270, 271, 274, 275, 276, 277, 278, 285, 289, 290, 291, 292, 293, 300, 304, 305, 306, 307, 308, 319, 320, 322, 323], "bbox": [0.0, 0.358578125, 0.645456674473068, 0.9366093750000001], "iscrowd": 0, "area": 66416.16700000003, "rle": {"size": [640, 427], "counts": "Q`0h1Ub02N2N2O1N2N2N2N2N2O0O2N2N2N2N2O1N2N2N2N2N2O1N2N2N1O2O0O001O1O1K4L5K5K5K4L5K5K5K5K4L5K5J6K4L5K5oNQ11N101O1O0O2O1O1O001N2O1O001N2O001O1N2O002N1O1N3N1O1N3M2N2N3M2N2M3N3M2N2N3M2N2N3L300100O010000O1000O01005K7I7I8H6J5K6J6J6I6K6J6J6J5K6J6WGYHm4l7lJVHT5P8dJRH\\5T8\\JnGc5Y8UJiGk5\\8oIeGQ6a8gIaGY6e8_I]Ga6i8WIYGi6l8PIVGP7P9hHRGX7T9^HPGb7V9SHoFm7V9iGoFW8]:O100000000000000000000000000000000000000O10O1000000WLcGoK]8R4YHWKg7j4jHdJV7\\5PI]JQ7d5TIVJl6k5YIoIg6Q6_IiIa6X6dIbI\\6_6iI[IW6e6nIVIR6k6SJnHn5S7WJgHi5Y7]JaHc5`7bJZH^5g7gJSHX5n7]31O10O01O1O001O4L4L4L4L4L4M3L4L4L4L4L5K4L4L4L4L5L3L4L4L4L4L4L4L001O01O000001O0000000100O1O001O100O1O1O1O100O1O001O100O1O1O1O100O001O1O100M3N2N2N2N2N1N3N2N2N2N2J5J7I7I7I6^HoDh5_;UJeDn4W in this image.", "objects": [{"patches": [120, 135, 136, 150, 151, 165, 166, 180, 181, 195, 196, 210, 211, 225, 226, 240, 242, 256, 257, 271, 272, 273, 285, 286, 287, 300, 301, 302], "bbox": [0.0, 0.36710937499999996, 0.20789227166276344, 0.888359375], "iscrowd": 0, "area": 12520.662349999999, "rle": {"size": [640, 427], "counts": "k;`2P=a40001O0O10000QFnHo6S7nHnHS7R7jHPIW7P7gHQIY7P7dHRI]7o6_HSIb7j6_HXIa7e6_H]Ib7`6_HaIb7[6_HgIb7V6_HkIa7S6_HoIb7n5^HTJc7h5_HYJb7d5^H^Jc7_5^HcJb7Z5^HhJc7T5_HmJa7Q5_HQKb7l4_HUKb7g4_H[Kb7b4^H`Kc7]4^HdKc7Y4]HiKc7T4_HnKa7k3cHWL^7b3gH_LZ7Y3kHiLV7o2oHSMR7f2SI[Mn6]2WIeMi6U2\\IlMe6l1`IWN`6a1fI`N[6Y1iIiNX6l0QJUOP6?YJCg53aJO`5FiJ;X5ZOPKh0Q5lNYKU1h4`N`Kc1`4RNiKo1R;2N2O1N2M3M3M4M1N001O010O1O0102M4M3M211N3N1N3N1N3N1N3N1N3N1N3N1N3N1N3M2E in this image.", "objects": [{"patches": [120, 135, 136, 150, 151, 165, 166, 180, 181, 195, 196, 210, 211, 225, 226, 240, 242, 256, 257, 271, 272, 273, 285, 286, 287, 300, 301, 302], "bbox": [0.0, 0.36710937499999996, 0.20789227166276344, 0.888359375], "iscrowd": 0, "area": 12520.662349999999, "rle": {"size": [640, 427], "counts": "k;`2P=a40001O0O10000QFnHo6S7nHnHS7R7jHPIW7P7gHQIY7P7dHRI]7o6_HSIb7j6_HXIa7e6_H]Ib7`6_HaIb7[6_HgIb7V6_HkIa7S6_HoIb7n5^HTJc7h5_HYJb7d5^H^Jc7_5^HcJb7Z5^HhJc7T5_HmJa7Q5_HQKb7l4_HUKb7g4_H[Kb7b4^H`Kc7]4^HdKc7Y4]HiKc7T4_HnKa7k3cHWL^7b3gH_LZ7Y3kHiLV7o2oHSMR7f2SI[Mn6]2WIeMi6U2\\IlMe6l1`IWN`6a1fI`N[6Y1iIiNX6l0QJUOP6?YJCg53aJO`5FiJ;X5ZOPKh0Q5lNYKU1h4`N`Kc1`4RNiKo1R;2N2O1N2M3M3M4M1N001O010O1O0102M4M3M211N3N1N3N1N3N1N3N1N3N1N3N1N3N1N3M2E in this image.", "objects": [{"patches": [13, 14, 15, 24, 25, 26, 35, 36, 37, 38, 46, 47, 48, 49, 57, 58, 59, 60, 61, 68, 69, 70, 71, 72, 77, 78, 79, 80, 81, 82, 83, 88, 89, 90, 91, 92, 93, 94, 100, 101, 102, 103, 104, 105, 112, 113, 114, 115, 116, 123, 124, 125, 126, 127, 134, 135, 136, 137, 138, 146, 147, 148, 149, 157, 158, 159, 160, 168, 169, 170, 171, 180, 181, 182, 191, 192, 193, 201, 202, 203, 204, 211, 212, 213, 214, 215, 216, 223, 224, 225, 226, 227], "bbox": [0.04565079365079365, 0.044937500000000005, 0.6574603174603174, 0.9146093750000001], "iscrowd": 0, "area": 54836.0188, "rle": {"size": [640, 315], "counts": "ao81lc03M4L4L4L3M4L4L4L4L3M4N2O1N101O1O1N101O1O1N2O001O1N2O001O1N2O1O001N2O1O2N1ZKlM]GV2b8lM\\GU2b8nM\\GS2b8PN\\GR2a8QN\\GS2`8QN]GR2_8QN_GR2]8QNaGQ2\\8RNaGR2[8QN`GU2\\8nM_GX2^8jM^G[2^8hM]G]2`8fM[G`2a8cMZGc2c8_MXGg2d8\\MXGi2e8YMVGm2g8UMTGP3i8SMRGS3k8oLQGV3l8lLoFZ3m8iLnF]3o8eLkFa3R9bLiFd3T9^LfFi3W9YLcFn3_5lIUMY2WMQ4`5hIXMY2RMU4c5dIZMY2mLZ4e5_I\\MZ2iL^4g5[I^MY2fLb4h5WIaMY2aLg4j5RIcMi9Z2YFeMj9W2XFgMl9U2WFiMl9S2VFlMm9P2UFnMn9o1TFPNo9l1SFSNP:i1RFUNR:g1QFWNR:e1PFZNR:c1PF[NT:a1nE^NU:^1mEaNV:[1mEdNU:Y1lEfNW:V1jEkNW:R1jEnNY:o0gEROZ:k0gEUO[:h0fEYO[:e0eE[O^:a0cE@^:=cED^:9cEG_:7aEJ`:3aEMa:0`E1a:M`E2c:J^E9a:D`E?_:^ObEe0]:YOcEi0]:TOdEo0[:nNfEU1Y:iNgEZ1Y:bNhEa1W:\\NjEf1V:WNkEl1T:RNlEQ2S:lMnEW2Q:fMPF]2o9aMRFa2n9[MSFg2m9VMTFm2k9QMUFR3j9kLWFX3h9eLYF^3g9^LZFc3R:PLnEQ4]:bKdE_4n<1O1O1O1O100O1O1O1O1O1O1O1O001O1O1O1O100O1O001O1O1O1O1O1O1O001O1O1O102M3M4L4L4L3M4L4L3M4L4L3dFQJa4S6RH[Mi7c7J5L5J6K4L5J6K4K6K5J5LW2hMR1oN:F7H9H7H9H7H9H8G:G8H8GYnR2"}, "label": "a man wearing a hat skating"}]} {"id": 463101, "image": "COCO_train2014_000000463101.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older man in a reindeer sweater\"."}], "task": "refering", "answer_template": "The \"an older man in a reindeer sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 24, 25, 26, 35, 36, 37, 38, 46, 47, 48, 49, 57, 58, 59, 60, 61, 68, 69, 70, 71, 72, 77, 78, 79, 80, 81, 82, 83, 88, 89, 90, 91, 92, 93, 94, 100, 101, 102, 103, 104, 105, 112, 113, 114, 115, 116, 123, 124, 125, 126, 127, 134, 135, 136, 137, 138, 146, 147, 148, 149, 157, 158, 159, 160, 168, 169, 170, 171, 180, 181, 182, 191, 192, 193, 201, 202, 203, 204, 211, 212, 213, 214, 215, 216, 223, 224, 225, 226, 227], "bbox": [0.04565079365079365, 0.044937500000000005, 0.6574603174603174, 0.9146093750000001], "iscrowd": 0, "area": 54836.0188, "rle": {"size": [640, 315], "counts": "ao81lc03M4L4L4L3M4L4L4L4L3M4N2O1N101O1O1N101O1O1N2O001O1N2O001O1N2O1O001N2O1O2N1ZKlM]GV2b8lM\\GU2b8nM\\GS2b8PN\\GR2a8QN\\GS2`8QN]GR2_8QN_GR2]8QNaGQ2\\8RNaGR2[8QN`GU2\\8nM_GX2^8jM^G[2^8hM]G]2`8fM[G`2a8cMZGc2c8_MXGg2d8\\MXGi2e8YMVGm2g8UMTGP3i8SMRGS3k8oLQGV3l8lLoFZ3m8iLnF]3o8eLkFa3R9bLiFd3T9^LfFi3W9YLcFn3_5lIUMY2WMQ4`5hIXMY2RMU4c5dIZMY2mLZ4e5_I\\MZ2iL^4g5[I^MY2fLb4h5WIaMY2aLg4j5RIcMi9Z2YFeMj9W2XFgMl9U2WFiMl9S2VFlMm9P2UFnMn9o1TFPNo9l1SFSNP:i1RFUNR:g1QFWNR:e1PFZNR:c1PF[NT:a1nE^NU:^1mEaNV:[1mEdNU:Y1lEfNW:V1jEkNW:R1jEnNY:o0gEROZ:k0gEUO[:h0fEYO[:e0eE[O^:a0cE@^:=cED^:9cEG_:7aEJ`:3aEMa:0`E1a:M`E2c:J^E9a:D`E?_:^ObEe0]:YOcEi0]:TOdEo0[:nNfEU1Y:iNgEZ1Y:bNhEa1W:\\NjEf1V:WNkEl1T:RNlEQ2S:lMnEW2Q:fMPF]2o9aMRFa2n9[MSFg2m9VMTFm2k9QMUFR3j9kLWFX3h9eLYF^3g9^LZFc3R:PLnEQ4]:bKdE_4n<1O1O1O1O100O1O1O1O1O1O1O1O001O1O1O1O100O1O001O1O1O1O1O1O1O001O1O1O102M3M4L4L4L3M4L4L3M4L4L3dFQJa4S6RH[Mi7c7J5L5J6K4L5J6K4K6K5J5LW2hMR1oN:F7H9H7H9H7H9H8G:G8H8GYnR2"}, "label": "an older man in a reindeer sweater"}]} {"id": 37122, "image": "COCO_train2014_000000037122.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pigeon flying second position with another one\"."}], "task": "refering", "answer_template": "The \"a pigeon flying second position with another one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 166, 167, 168, 190, 191, 212, 213, 214, 215, 235, 236, 237, 238, 261], "bbox": [0.223046875, 0.5134357541899441, 0.38590625, 0.889441340782123], "iscrowd": 0, "area": 6092.925549999998, "rle": {"size": [358, 640], "counts": "TXb12S;3L4M3M2MC[E>b:4SN^OTIe0i6^OVIb0g6AYI`0d6CZI?e6BZI?d6C[I?c6DYI?d6EYI in this image.", "objects": [{"patches": [143, 144, 166, 167, 168, 190, 191, 212, 213, 214, 215, 235, 236, 237, 238, 261], "bbox": [0.223046875, 0.5134357541899441, 0.38590625, 0.889441340782123], "iscrowd": 0, "area": 6092.925549999998, "rle": {"size": [358, 640], "counts": "TXb12S;3L4M3M2MC[E>b:4SN^OTIe0i6^OVIb0g6AYI`0d6CZI?e6BZI?d6C[I?c6DYI?d6EYI in this image.", "objects": [{"patches": [38, 61, 62, 63, 84, 85, 86, 87, 108, 109, 110, 111, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 177, 178, 180, 181, 182, 183, 205, 206], "bbox": [0.6519999999999999, 0.12902234636871507, 0.981609375, 0.6645251396648045], "iscrowd": 0, "area": 13919.382649999996, "rle": {"size": [358, 640], "counts": "[Vb41T;1JOWE2i:NUE4j:MUE3k:MTE53JX:0oE8GJ[:0kE;U:HhE9X:HgE7Y:LdE52CU:g0kEZOT:g0kEZOT:f0kE\\OU:c0kE_OS:k000O10O0010O01O001O000101N5L0O2O0O2O0O1gHeNf4\\1XKfNf4[1XKiNe4Y1YKiNf4W1XKmNe4T1ZKnNe4S1XKQOe4P1ZKROe4o0XKUOe4l0YKWOf4i0YKZOd4h0YKZOg4f0XKZOh4h0UKYOl4g0SKZOl4h0QKYOo4h0PKXOQ5h0mJZOR5h0lJXOU5h0iJYOW5i0fJXO[5h0dJYO[5h0cJYO^5h0`JXO`5i0^JYOb5h0\\JXOd5i0ZJXOf5i0YJXOg5i0VJXOj5i0UJWOl5j0QJXOn5i0QJWOP6i0nIXOR6j0kIXOU6h0iIYOW6i0eIYO\\6g0bI[O]6f0aI[O_6f0^IZOe6f0YIYOi6_24M2N3M3M3L3N3M3M3M3L4M3M2N3L4N2N2O1N2O2M3N2N101O1N2O1O0O2O1O1O0O2O1O1ON101O1O001O1O1O001O000O2N1O1N3N1O1O2N101O00TNbGP1^8mNfGS1Y8kNiGU1W8hNmGW1S8gNoGZ1P8dNSH[1m7bNVH^1j7_NZHa1e7\\N^Hd1b7YNbHg1^801O01O01O010O010O0010O010O01O010O010O00100O0100O100O01000O100O02O1O1O10O2O00O1N2N101N2N2O1N1O2O1N2N2N101N2N2O1N1O2O1N2N2O0O2N2O1N2E:O2N2NTo3"}, "label": "a bird flying to the right of another bird"}]} {"id": 37122, "image": "COCO_train2014_000000037122.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bird on the right\"."}], "task": "refering", "answer_template": "The \"the bird on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 61, 62, 63, 84, 85, 86, 87, 108, 109, 110, 111, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 177, 178, 180, 181, 182, 183, 205, 206], "bbox": [0.6519999999999999, 0.12902234636871507, 0.981609375, 0.6645251396648045], "iscrowd": 0, "area": 13919.382649999996, "rle": {"size": [358, 640], "counts": "[Vb41T;1JOWE2i:NUE4j:MUE3k:MTE53JX:0oE8GJ[:0kE;U:HhE9X:HgE7Y:LdE52CU:g0kEZOT:g0kEZOT:f0kE\\OU:c0kE_OS:k000O10O0010O01O001O000101N5L0O2O0O2O0O1gHeNf4\\1XKfNf4[1XKiNe4Y1YKiNf4W1XKmNe4T1ZKnNe4S1XKQOe4P1ZKROe4o0XKUOe4l0YKWOf4i0YKZOd4h0YKZOg4f0XKZOh4h0UKYOl4g0SKZOl4h0QKYOo4h0PKXOQ5h0mJZOR5h0lJXOU5h0iJYOW5i0fJXO[5h0dJYO[5h0cJYO^5h0`JXO`5i0^JYOb5h0\\JXOd5i0ZJXOf5i0YJXOg5i0VJXOj5i0UJWOl5j0QJXOn5i0QJWOP6i0nIXOR6j0kIXOU6h0iIYOW6i0eIYO\\6g0bI[O]6f0aI[O_6f0^IZOe6f0YIYOi6_24M2N3M3M3L3N3M3M3M3L4M3M2N3L4N2N2O1N2O2M3N2N101O1N2O1O0O2O1O1O0O2O1O1ON101O1O001O1O1O001O000O2N1O1N3N1O1O2N101O00TNbGP1^8mNfGS1Y8kNiGU1W8hNmGW1S8gNoGZ1P8dNSH[1m7bNVH^1j7_NZHa1e7\\N^Hd1b7YNbHg1^801O01O01O010O010O0010O010O01O010O010O00100O0100O100O01000O100O02O1O1O10O2O00O1N2N101N2N2O1N1O2O1N2N2N101N2N2O1N1O2O1N2N2O0O2N2O1N2E:O2N2NTo3"}, "label": "the bird on the right"}]} {"id": 364803, "image": "COCO_train2014_000000364803.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small carrot that has a larger carrot to its left and onions to its right\"."}], "task": "refering", "answer_template": "The \"a small carrot that has a larger carrot to its left and onions to its right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 166, 167, 168, 169, 190, 191, 192, 193, 194, 215, 216, 217, 218, 239, 240, 241, 242, 264, 265], "bbox": [0.2383125, 0.3835205183585313, 0.5327968750000001, 0.6658747300215981], "iscrowd": 0, "area": 7054.825200000004, "rle": {"size": [463, 640], "counts": "Q\\U23Y>4L4K4M4L4L3M4M3O001O1O00100O001O1O001O10O01O1O001O1O010O1O010O10O01000O0100O01000O01000O010O10O10O10O01000O0100O01000O010O0100O010O010O10O010O0100O010O010O10O10O010O10O0100O010O010O10O010O0100O0100O0100O010O010O10O010O0100O010O010O10O010O10O10O0100O010O010O100O010O1O100O10O0100O100O10O0100O1O100O010O100O100O010O100O100O00100O100O10O0100O100O10O0100O1O100O010O100O100O010OTmV4"}, "label": "a small carrot that has a larger carrot to its left and onions to its right"}]} {"id": 364803, "image": "COCO_train2014_000000364803.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two orange carrots in the center of some other vegetables\"."}], "task": "refering", "answer_template": "The \"two orange carrots in the center of some other vegetables\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 166, 167, 168, 169, 190, 191, 192, 193, 194, 215, 216, 217, 218, 239, 240, 241, 242, 264, 265], "bbox": [0.2383125, 0.3835205183585313, 0.5327968750000001, 0.6658747300215981], "iscrowd": 0, "area": 7054.825200000004, "rle": {"size": [463, 640], "counts": "Q\\U23Y>4L4K4M4L4L3M4M3O001O1O00100O001O1O001O10O01O1O001O1O010O1O010O10O01000O0100O01000O01000O010O10O10O10O01000O0100O01000O010O0100O010O010O10O010O0100O010O010O10O10O010O10O0100O010O010O10O010O0100O0100O0100O010O010O10O010O0100O010O010O10O010O10O10O0100O010O010O100O010O1O100O10O0100O100O10O0100O1O100O010O100O100O010O100O100O00100O100O10O0100O100O10O0100O1O100O010O100O100O010OTmV4"}, "label": "two orange carrots in the center of some other vegetables"}]} {"id": 364803, "image": "COCO_train2014_000000364803.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"carrot\"."}], "task": "refering", "answer_template": "The \"carrot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 163, 164, 165, 166, 187, 188, 189, 190, 191, 212, 213, 214, 215, 216, 237, 238, 239, 240, 241, 261, 262, 263, 264, 265, 286, 287, 288, 289, 312], "bbox": [0.11915625, 0.4047084233261339, 0.592375, 0.7764794816414686], "iscrowd": 0, "area": 12809.181799999995, "rle": {"size": [463, 640], "counts": "laR11]>2O2M2N3M2N2O2M2N3N1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O0O2O1O0O20O01O010O010O1O010O0010O010O01O10O01O010O010O1O010O010O010O10O010O010O01000O0010O010O010O01O10O010O01O010O010O10O01O010O010O0010O0100O010O0010O010O01O010O10O010O01O010O010O0010O0100O010O0010O010O01O010O10O010O10O0100O010O10O010000O010O10O0100O010O10O10O100O010O10O0100O010O10O10O100O010O10O0100O010O10O10O10O0100O10O0100O01000O010O010O10O0100O010O10O10O10O010O0100O010O10O10O10O010O001O010O001O010O1O001O010O1O001O010O001O1O010O001O010O1O001O010O1O001O010O001O1O010O001O00100O001O010O1O010O0010O0100O010O0010O0100O010O00100O010O1O1O1N2N2N101N2N2Nkee3"}, "label": "carrot"}]} {"id": 364803, "image": "COCO_train2014_000000364803.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the carrot on top touching the green zucchini\"."}], "task": "refering", "answer_template": "The \"the carrot on top touching the green zucchini\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 163, 164, 165, 166, 187, 188, 189, 190, 191, 212, 213, 214, 215, 216, 237, 238, 239, 240, 241, 261, 262, 263, 264, 265, 286, 287, 288, 289, 312], "bbox": [0.11915625, 0.4047084233261339, 0.592375, 0.7764794816414686], "iscrowd": 0, "area": 12809.181799999995, "rle": {"size": [463, 640], "counts": "laR11]>2O2M2N3M2N2O2M2N3N1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O0O2O1O0O20O01O010O010O1O010O0010O010O01O10O01O010O010O1O010O010O010O10O010O010O01000O0010O010O010O01O10O010O01O010O010O10O01O010O010O0010O0100O010O0010O010O01O010O10O010O01O010O010O0010O0100O010O0010O010O01O010O10O010O10O0100O010O10O010000O010O10O0100O010O10O10O100O010O10O0100O010O10O10O100O010O10O0100O010O10O10O10O0100O10O0100O01000O010O010O10O0100O010O10O10O10O010O0100O010O10O10O10O010O001O010O001O010O1O001O010O1O001O010O001O1O010O001O010O1O001O010O1O001O010O001O1O010O001O00100O001O010O1O010O0010O0100O010O0010O0100O010O00100O010O1O1O1N2N2N101N2N2Nkee3"}, "label": "the carrot on top touching the green zucchini"}]} {"id": 102662, "image": "COCO_train2014_000000102662.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white vase that is located on the edge of the picture\"."}], "task": "refering", "answer_template": "The \"a white vase that is located on the edge of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 153, 170, 187, 204, 221, 222, 238, 239, 255, 256, 272, 273, 289, 290, 306, 307, 323, 324, 340, 341, 358], "bbox": [0.003, 0.368546875, 0.11985416666666666, 0.92584375], "iscrowd": 0, "area": 10993.885449999996, "rle": {"size": [640, 480], "counts": "al0o3m?T4lK[1fN4K5K4MUIPFm3k9ULYFj3b9XLbFg3Z9kIPGT1KP5P9oI[Gj0IU5l8QJ_Gd0IZ5g8QJeG>I`5b8QJjG8Hf5]8QJPH3Gj5Z8RJSHMGQ6U8QJXHIEW6R8PJ\\HDE\\6o7nI_HAEa6l7nISJR6m5lIUJT6j5lIVJU6i5jIYJU6h5jIXJW6h5gIZJY6e5gIZJZ6g5dIYJ^6g5aIYJ`6f5_IZJb6g5]IXJe6h5YIXJi6g5WIXJj6i5TIWJn6i5QIVJQ7i5nHWJS7[91N2N1U1jN?_O;F=Bi1XN>A7J5L4K4L3M3M4M2K6K4K8H8D=^OlWW8"}, "label": "a white vase that is located on the edge of the picture"}]} {"id": 102662, "image": "COCO_train2014_000000102662.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white cylinder to the left of the vase\"."}], "task": "refering", "answer_template": "The \"white cylinder to the left of the vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 153, 170, 187, 204, 221, 222, 238, 239, 255, 256, 272, 273, 289, 290, 306, 307, 323, 324, 340, 341, 358], "bbox": [0.003, 0.368546875, 0.11985416666666666, 0.92584375], "iscrowd": 0, "area": 10993.885449999996, "rle": {"size": [640, 480], "counts": "al0o3m?T4lK[1fN4K5K4MUIPFm3k9ULYFj3b9XLbFg3Z9kIPGT1KP5P9oI[Gj0IU5l8QJ_Gd0IZ5g8QJeG>I`5b8QJjG8Hf5]8QJPH3Gj5Z8RJSHMGQ6U8QJXHIEW6R8PJ\\HDE\\6o7nI_HAEa6l7nISJR6m5lIUJT6j5lIVJU6i5jIYJU6h5jIXJW6h5gIZJY6e5gIZJZ6g5dIYJ^6g5aIYJ`6f5_IZJb6g5]IXJe6h5YIXJi6g5WIXJj6i5TIWJn6i5QIVJQ7i5nHWJS7[91N2N1U1jN?_O;F=Bi1XN>A7J5L4K4L3M3M4M2K6K4K8H8D=^OlWW8"}, "label": "white cylinder to the left of the vase"}]} {"id": 102662, "image": "COCO_train2014_000000102662.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a vase with blue liquid containing yellow and purple flowers\"."}], "task": "refering", "answer_template": "The \"a vase with blue liquid containing yellow and purple flowers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 28, 29, 39, 44, 45, 46, 54, 55, 56, 57, 61, 62, 63, 71, 72, 73, 74, 76, 77, 78, 79, 89, 90, 91, 92, 93, 94, 95, 96, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 169, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 259, 260, 261, 262, 263, 264, 265, 266, 278, 279, 280, 281, 282, 283, 296, 297, 298, 299, 313, 314, 315, 316, 330, 331, 332, 333, 347, 348, 349, 350, 364, 365, 366, 367, 383], "bbox": [0.04502083333333333, 0.036359375, 1.0, 0.963640625], "iscrowd": 0, "area": 109531.23630000002, "rle": {"size": [640, 480], "counts": "ZR>0O00010O0001O01O00010O0001O01O01fHhMjLX2V3oMcLQ2^3TN]Ll1b3XNZLi1e3YNYLh1f3ZNbJ_OeLW2h8]N_JEaLo1P9]N[JFeLm1P9_NXJFhLk1o8aNUJGkLi1o8bNSJGmLh1o8cNPJHPMf1P9bNnIISMe1n8bNnIKSMd1n8bNlIMUMc1m8`NmIOVMb1k8_NmI2WMa1k8]NlI3YMb1i8[NlI6[M`1g8[NlI7\\M`1f8YNkI;^M]1f8YNhI>bMY1f8XNeIb0eMV1f8XNbIh0eMQ1h8XN^Io0gMi0i8YN]IU1fMb0l8[NZIZ1fMA`0A>A`0A;Da4`0O00001O00001O0000001O00O1000000O100000000O1000O1000O1000O10O1000O1000O1000O1000O1000O10O1000O1N2O001N2O1O0O2O1O1fNl@fMd>l2`AoLh=i3^BQLj in this image.", "objects": [{"patches": [11, 28, 29, 39, 44, 45, 46, 54, 55, 56, 57, 61, 62, 63, 71, 72, 73, 74, 76, 77, 78, 79, 89, 90, 91, 92, 93, 94, 95, 96, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 169, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 259, 260, 261, 262, 263, 264, 265, 266, 278, 279, 280, 281, 282, 283, 296, 297, 298, 299, 313, 314, 315, 316, 330, 331, 332, 333, 347, 348, 349, 350, 364, 365, 366, 367, 383], "bbox": [0.04502083333333333, 0.036359375, 1.0, 0.963640625], "iscrowd": 0, "area": 109531.23630000002, "rle": {"size": [640, 480], "counts": "ZR>0O00010O0001O01O00010O0001O01O01fHhMjLX2V3oMcLQ2^3TN]Ll1b3XNZLi1e3YNYLh1f3ZNbJ_OeLW2h8]N_JEaLo1P9]N[JFeLm1P9_NXJFhLk1o8aNUJGkLi1o8bNSJGmLh1o8cNPJHPMf1P9bNnIISMe1n8bNnIKSMd1n8bNlIMUMc1m8`NmIOVMb1k8_NmI2WMa1k8]NlI3YMb1i8[NlI6[M`1g8[NlI7\\M`1f8YNkI;^M]1f8YNhI>bMY1f8XNeIb0eMV1f8XNbIh0eMQ1h8XN^Io0gMi0i8YN]IU1fMb0l8[NZIZ1fMA`0A>A`0A;Da4`0O00001O00001O0000001O00O1000000O100000000O1000O1000O1000O10O1000O1000O1000O1000O1000O10O1000O1N2O001N2O1O0O2O1O1fNl@fMd>l2`AoLh=i3^BQLj in this image.", "objects": [{"patches": [98, 115, 116, 132, 133, 134, 135, 151, 152, 153, 168, 169, 170, 186, 187, 188, 204, 205], "bbox": [0.32522, 0.4625373134328358, 0.5555800000000001, 0.9838805970149255], "iscrowd": 0, "area": 8966.574250000001, "rle": {"size": [335, 500], "counts": "Qae15Z:1N2N3M2O1N1O100O1O100O1O00100O010O10O10O010O1[HBY5?bJ0R52hJ=n4DlJk0i4VOUKo0h4ROUKP1j4ROSKP1m4QOQKP1n4QOoJR1Q5oNhJW1W5Q200O100O100O2N1OG9K5K5M30O101O1O001O1O1N2O1O1O1N1O2O1N1OLmIPLQ6Q4QJmKo5S450011O1O1000000O10001O000000000O2N1N3N2M3N2UMVIm1m6nMZIT12gNe61^IV1\\8M4M2M3M3N3L3N1N3O11O1O1O1N11N1O1N2O1O1O1O1O2L3N?A2M3Ni\\X2"}, "label": "the man inside the glass , with his back turned"}]} {"id": 454923, "image": "COCO_train2014_000000454923.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with his back turned toward the camera , enjoying a conversation with a friend\"."}], "task": "refering", "answer_template": "The \"a man with his back turned toward the camera , enjoying a conversation with a friend\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 115, 116, 132, 133, 134, 135, 151, 152, 153, 168, 169, 170, 186, 187, 188, 204, 205], "bbox": [0.32522, 0.4625373134328358, 0.5555800000000001, 0.9838805970149255], "iscrowd": 0, "area": 8966.574250000001, "rle": {"size": [335, 500], "counts": "Qae15Z:1N2N3M2O1N1O100O1O100O1O00100O010O10O10O010O1[HBY5?bJ0R52hJ=n4DlJk0i4VOUKo0h4ROUKP1j4ROSKP1m4QOQKP1n4QOoJR1Q5oNhJW1W5Q200O100O100O2N1OG9K5K5M30O101O1O001O1O1N2O1O1O1N1O2O1N1OLmIPLQ6Q4QJmKo5S450011O1O1000000O10001O000000000O2N1N3N2M3N2UMVIm1m6nMZIT12gNe61^IV1\\8M4M2M3M3N3L3N1N3O11O1O1O1N11N1O1N2O1O1O1O1O2L3N?A2M3Ni\\X2"}, "label": "a man with his back turned toward the camera , enjoying a conversation with a friend"}]} {"id": 454923, "image": "COCO_train2014_000000454923.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with her hair in a bun talks on her cellphone\"."}], "task": "refering", "answer_template": "The \"a woman with her hair in a bun talks on her cellphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 73, 74, 75, 76, 92, 93, 94, 109, 110, 111, 112, 127, 128, 129, 130, 145, 146, 147, 148, 163, 164, 165, 166, 181, 182, 183, 184, 199, 200, 201, 202, 203], "bbox": [0.05872, 0.2966268656716418, 0.29058000000000006, 0.9887761194029852], "iscrowd": 0, "area": 19980.705000000005, "rle": {"size": [335, 500], "counts": "_i91Y:;F:E6J5L4VMnNnJ>UOi0g5JQJOJ;o5m2J6L4O1N2O1N2O1O1N2O100O1eNTKSMl4l2YKoLh4Q3\\KkLd4T3aKgL`4X3bKgL^4X3dKfL]4Y3fKeLZ4\\3fKcLZ4]3gKaLZ4_3gK`LY4`3hK^LY4c3gK\\LY4d3Y1001O00001O01H7L4N2O1N2O1N2O1N20000O10000O10000O10000O11O00000000001O00000000001O000000001O001O1O1O1O001O1O1O0cM_LlMb3YN_LH0m21oNa3WNcLIOP31lNV4oMmKU3OiNc4V1aKfN`4Y1cKdN^4\\1cKbN_4\\1cKbN_4\\1bKcN`4\\1aKaNb4]1`KaNb4fNQKW2>QOS5n0oJPOR5o0PKoNP5R1QKlNP5P1UKnNl4m0ZKQOg4j0`KSO`4i0fKUO[4f0lKWOU4d0QLZOP4<\\LAR4F[L8Q7M4L4L4L4LVjc3"}, "label": "a woman with her hair in a bun talks on her cellphone"}]} {"id": 454923, "image": "COCO_train2014_000000454923.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a businesswoman on a phone\"."}], "task": "refering", "answer_template": "The \"a businesswoman on a phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 73, 74, 75, 76, 92, 93, 94, 109, 110, 111, 112, 127, 128, 129, 130, 145, 146, 147, 148, 163, 164, 165, 166, 181, 182, 183, 184, 199, 200, 201, 202, 203], "bbox": [0.05872, 0.2966268656716418, 0.29058000000000006, 0.9887761194029852], "iscrowd": 0, "area": 19980.705000000005, "rle": {"size": [335, 500], "counts": "_i91Y:;F:E6J5L4VMnNnJ>UOi0g5JQJOJ;o5m2J6L4O1N2O1N2O1O1N2O100O1eNTKSMl4l2YKoLh4Q3\\KkLd4T3aKgL`4X3bKgL^4X3dKfL]4Y3fKeLZ4\\3fKcLZ4]3gKaLZ4_3gK`LY4`3hK^LY4c3gK\\LY4d3Y1001O00001O01H7L4N2O1N2O1N2O1N20000O10000O10000O10000O11O00000000001O00000000001O000000001O001O1O1O1O001O1O1O0cM_LlMb3YN_LH0m21oNa3WNcLIOP31lNV4oMmKU3OiNc4V1aKfN`4Y1cKdN^4\\1cKbN_4\\1cKbN_4\\1bKcN`4\\1aKaNb4]1`KaNb4fNQKW2>QOS5n0oJPOR5o0PKoNP5R1QKlNP5P1UKnNl4m0ZKQOg4j0`KSO`4i0fKUO[4f0lKWOU4d0QLZOP4<\\LAR4F[L8Q7M4L4L4L4LVjc3"}, "label": "a businesswoman on a phone"}]} {"id": 454923, "image": "COCO_train2014_000000454923.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in striped shirt wearing glasses\"."}], "task": "refering", "answer_template": "The \"man in striped shirt wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 120, 121, 137, 138, 139, 140, 155, 156, 157, 158, 173, 174, 175, 176], "bbox": [0.6144400000000001, 0.48161194029850746, 0.8132, 0.8447164179104478], "iscrowd": 0, "area": 7005.407800000001, "rle": {"size": [335, 500], "counts": "mdT32\\:4K2O2O0QGGg7;VHGj79SHJl77RHJo77PHIo78PHIn79RHGk7 in this image.", "objects": [{"patches": [102, 120, 121, 137, 138, 139, 140, 155, 156, 157, 158, 173, 174, 175, 176], "bbox": [0.6144400000000001, 0.48161194029850746, 0.8132, 0.8447164179104478], "iscrowd": 0, "area": 7005.407800000001, "rle": {"size": [335, 500], "counts": "mdT32\\:4K2O2O0QGGg7;VHGj79SHJl77RHJo77PHIo78PHIn79RHGk7 in this image.", "objects": [{"patches": [154, 155, 177, 178, 179, 200, 201, 202, 223, 224, 225, 246, 247, 248, 269, 270, 271, 292, 293, 294, 315, 316, 317], "bbox": [0.704265625, 0.4099287410926366, 0.819984375, 0.9295249406175773], "iscrowd": 0, "area": 9554.066800000006, "rle": {"size": [421, 640], "counts": "]di56n<8^CLe;k0N2N1mGjNV4X1eKmNZ4T1dKmN[4T1ZH]OY2@\\5U1WHQ1i0kMP7U1UHc19XMa7W1UHV2k7b1XOhJcIY5Z6jJdIW5Z6kJfIU5X6oJeIR5Y6QKfIo4X6UKeIl4Y6WKfIh4Z6[KdId4\\6^KbIb4_6eKZI[4e6S100O10hNcI^KICd6f4mI]K^OF<3X6c4XJaK\\6V4QJdKP6U4d1H82N3M2N2N3M2OZLaGf2b8TM`Gn2^8PMcGR3\\8kLfGW3Z8eLhG^3X8]LjGe3j82O1N3M2N7I7J6eG_Kd7g4SH^Kk7P5O0O9H0O10O001hL\\Hn0c7UOcHb0^7@gHWNGY1b7a0oHfMJe1X7e0dIhMnNP1`7U1TJiNn5U1UJhNm5U1VJoMRN`0j7OTHL73k1^OTNb0g7NPI0g1K[52PI2U:KWFIl94WFIl94d1N3MdQ_1"}, "label": "man with green shirt standing beside horse"}]} {"id": 405777, "image": "COCO_train2014_000000405777.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older man in brown shirt and tan slacks carries a brown paper shopping bag\"."}], "task": "refering", "answer_template": "The \"an older man in brown shirt and tan slacks carries a brown paper shopping bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 177, 178, 179, 200, 201, 202, 223, 224, 225, 246, 247, 248, 269, 270, 271, 292, 293, 294, 315, 316, 317], "bbox": [0.704265625, 0.4099287410926366, 0.819984375, 0.9295249406175773], "iscrowd": 0, "area": 9554.066800000006, "rle": {"size": [421, 640], "counts": "]di56n<8^CLe;k0N2N1mGjNV4X1eKmNZ4T1dKmN[4T1ZH]OY2@\\5U1WHQ1i0kMP7U1UHc19XMa7W1UHV2k7b1XOhJcIY5Z6jJdIW5Z6kJfIU5X6oJeIR5Y6QKfIo4X6UKeIl4Y6WKfIh4Z6[KdId4\\6^KbIb4_6eKZI[4e6S100O10hNcI^KICd6f4mI]K^OF<3X6c4XJaK\\6V4QJdKP6U4d1H82N3M2N2N3M2OZLaGf2b8TM`Gn2^8PMcGR3\\8kLfGW3Z8eLhG^3X8]LjGe3j82O1N3M2N7I7J6eG_Kd7g4SH^Kk7P5O0O9H0O10O001hL\\Hn0c7UOcHb0^7@gHWNGY1b7a0oHfMJe1X7e0dIhMnNP1`7U1TJiNn5U1UJhNm5U1VJoMRN`0j7OTHL73k1^OTNb0g7NPI0g1K[52PI2U:KWFIl94WFIl94d1N3MdQ_1"}, "label": "an older man in brown shirt and tan slacks carries a brown paper shopping bag"}]} {"id": 405777, "image": "COCO_train2014_000000405777.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman standing next to a horse head sculpture\"."}], "task": "refering", "answer_template": "The \"a woman standing next to a horse head sculpture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 175, 176, 198, 199, 200, 221, 222, 223, 244, 245, 246, 267, 268, 269, 290, 291, 292, 313, 314, 315], "bbox": [0.611984375, 0.4179809976247031, 0.7154687500000001, 0.9235866983372922], "iscrowd": 0, "area": 8736.493550000003, "rle": {"size": [421, 640], "counts": "V^Q5;^< in this image.", "objects": [{"patches": [152, 153, 175, 176, 198, 199, 200, 221, 222, 223, 244, 245, 246, 267, 268, 269, 290, 291, 292, 313, 314, 315], "bbox": [0.611984375, 0.4179809976247031, 0.7154687500000001, 0.9235866983372922], "iscrowd": 0, "area": 8736.493550000003, "rle": {"size": [421, 640], "counts": "V^Q5;^< in this image.", "objects": [{"patches": [12, 13, 35, 36, 55, 56, 57, 58, 59, 60, 78, 79, 80, 81, 82, 83, 102, 103, 104, 105, 106, 125, 126, 127, 128], "bbox": [0.405015625, 0.04442622950819672, 0.649921875, 0.3926697892271662], "iscrowd": 0, "area": 11727.705449999998, "rle": {"size": [427, 640], "counts": "_R\\32T=6O1O1O001N2O1O001O1O1N101O001O001]D^OZ:c0dE_O\\:b0aE@_:`0_EBa:?\\EDd:;ZEGf::XEGh:9VEJi:7TEKl:5REMn:4oDNQ;3lD0T;0jD1V;OjD1V;0hD1X;0fD2Y;OeD3Z;P100O1O1O1O1O1O1WEkM[:b201O1O1O1O10O01O1O1O1O1OO2O00000000001N10000000N2M2O2M3N2M3N2M3N2M3N2M3M3N2M3N2M3N3L3N2M3N2M30000000001OO1O1N3N1O1O1O1O1O1O1bMPG`0Q9[OTGd0n8WOUGi0m8QOXGn0k8lNXGT1j8fN[GY1h8aN\\G^1f8]N]Gc1e8WN`Gh1a8VN`Gj1a8TNaGk1`8SNaGm1`8QNbGn1_8PNcGP2]8nMdGR2k9O001O010O001O010O1N101N101N1O2O1N2O1N2N2O1N2O1O0O2O1O1O2M3N2N2M3N3K4K5K5K5KfZm2"}, "label": "broccoli stalk that is pointing up and is touching a sliced carrot"}]} {"id": 233746, "image": "COCO_train2014_000000233746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"2 pieces of broccoli closest to the left side of the plate\"."}], "task": "refering", "answer_template": "The \"2 pieces of broccoli closest to the left side of the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 77, 78, 100, 101, 119, 120, 121, 122, 123, 142, 143, 144, 145, 146, 165, 166, 167, 168, 190], "bbox": [0.18740625, 0.13482435597189696, 0.43629687500000003, 0.5483138173302108], "iscrowd": 0, "area": 7591.61915, "rle": {"size": [427, 640], "counts": "bVb1Q1X<2N3M2N2N2N2O100O100O10000O1000000000000001O00000O1O100O2N1O1O100DRDEl;;SDGm;9PDJo;6PDMn;4PDNo;2oC4m;MQD7l;g0M2NO100O2O0O100O100O1O3M2M\\OlNWEQ1i:POXEn0h:TOYEh0h:YOPELGi0Y;]OnDKKe0X;AkDKNa0X;EiDLO=X;IgDK2:W;MeDK54W;2cDK71V;6aDJ:MV;;^DJ]DJ>FU;P1mDmNS;T1nDlNQ;U1a01O1000001O0000000H9G8I7N2lNnMlFU2R9oMjFT2S9QNgFm0[OMl9ZOdFk0BJh9_ObFh0IGc9F^Fe01D`9J[Fb08DZ9Z1iFdNV9\\1lFbNT9^1oF`NP9`1RG_Nm8a1VG\\Nj8d1XG[Ng8d1]GZNb8f1`GXN`8h1cGVN\\8j1gGTNX8m1jGPNV8Q2lGmMS8T2oGjMo7X2RHgMm7Z2UHcMk7^2`11O01000006J8G5L5J5L3L3N2M3N2M3N2M2O2M2N3M3M3L7J6J\\]f4"}, "label": "2 pieces of broccoli closest to the left side of the plate"}]} {"id": 512282, "image": "COCO_train2014_000000512282.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man talking on a mobile phone wearing an open suit coat , jeans and a blue shirt carrying a bag\"."}], "task": "refering", "answer_template": "The \"a man talking on a mobile phone wearing an open suit coat , jeans and a blue shirt carrying a bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 89, 90, 91, 92, 110, 111, 112, 113, 131, 132, 133, 134, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 319, 320, 321, 322, 323, 324, 325, 326, 327, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 467, 468, 469, 470, 471, 472, 473, 474], "bbox": [0.11716521739130435, 0.15903125, 0.6634782608695652, 0.9985781249999999], "iscrowd": 0, "area": 113566.26360000002, "rle": {"size": [640, 575], "counts": "ZWZ12hc0?CZ5b0J7J6I6K6K5J6L3N3L4K4M5K7H9H7H9F9F:SN^^Oi0ma0jNY^OP1_b0K5L4L4M3L4M3L6K]Qh3"}, "label": "a man talking on a mobile phone wearing an open suit coat , jeans and a blue shirt carrying a bag"}]} {"id": 512282, "image": "COCO_train2014_000000512282.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue shirt and black jacket using a red mobile phone\"."}], "task": "refering", "answer_template": "The \"a man in a blue shirt and black jacket using a red mobile phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 89, 90, 91, 92, 110, 111, 112, 113, 131, 132, 133, 134, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 319, 320, 321, 322, 323, 324, 325, 326, 327, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 467, 468, 469, 470, 471, 472, 473, 474], "bbox": [0.11716521739130435, 0.15903125, 0.6634782608695652, 0.9985781249999999], "iscrowd": 0, "area": 113566.26360000002, "rle": {"size": [640, 575], "counts": "ZWZ12hc0?CZ5b0J7J6I6K6K5J6L3N3L4K4M5K7H9H7H9F9F:SN^^Oi0ma0jNY^OP1_b0K5L4L4M3L4M3L6K]Qh3"}, "label": "a man in a blue shirt and black jacket using a red mobile phone"}]} {"id": 45339, "image": "COCO_train2014_000000045339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady playing a throw plate game\"."}], "task": "refering", "answer_template": "The \"a lady playing a throw plate game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 149, 150, 151, 152, 153, 174, 175, 176, 177, 198, 199, 200, 201, 221, 222, 223, 224, 225, 244, 245, 246, 267, 268, 269, 290, 291, 292, 312, 313, 315], "bbox": [0.504390625, 0.3424355971896956, 0.79953125, 0.9201873536299765], "iscrowd": 0, "area": 12270.989800000005, "rle": {"size": [427, 640], "counts": "bkV44W=00000O1O1O2O000NIPC7oRIH^O9_15aN^O`7=RIG@6^1f0Q6]OQIFCFk1U1c5^OoHGk1j0Z5]OlHFl1k0[5]OoH@h1R1Z5^OgK`0Z4@jK:X4FkK6V4IoK2R4MTLMn31TLLm34ULJl35ULJl35VLIj38XLEi3;XLBi3?b4OO21O2N000O10000000001N1000000001OO1O13M1O1O3M100O3M2OO0N3M32MO1O2L6J6IaSe1"}, "label": "a lady playing a throw plate game"}]} {"id": 45339, "image": "COCO_train2014_000000045339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman throwing a frisbee\"."}], "task": "refering", "answer_template": "The \"a woman throwing a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 149, 150, 151, 152, 153, 174, 175, 176, 177, 198, 199, 200, 201, 221, 222, 223, 224, 225, 244, 245, 246, 267, 268, 269, 290, 291, 292, 312, 313, 315], "bbox": [0.504390625, 0.3424355971896956, 0.79953125, 0.9201873536299765], "iscrowd": 0, "area": 12270.989800000005, "rle": {"size": [427, 640], "counts": "bkV44W=00000O1O1O2O000NIPC7oRIH^O9_15aN^O`7=RIG@6^1f0Q6]OQIFCFk1U1c5^OoHGk1j0Z5]OlHFl1k0[5]OoH@h1R1Z5^OgK`0Z4@jK:X4FkK6V4IoK2R4MTLMn31TLLm34ULJl35ULJl35VLIj38XLEi3;XLBi3?b4OO21O2N000O10000000001N1000000001OO1O13M1O1O3M100O3M2OO0N3M32MO1O2L6J6IaSe1"}, "label": "a woman throwing a frisbee"}]} {"id": 381214, "image": "COCO_train2014_000000381214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white boat attached to a doc\"."}], "task": "refering", "answer_template": "The \"white boat attached to a doc\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245], "bbox": [0.439859375, 0.41233333333333333, 0.698171875, 0.6285000000000001], "iscrowd": 0, "area": 8863.140300000003, "rle": {"size": [480, 640], "counts": "d]T4P1P>1O001O001O1DmNiBS1W=oNgBR1W=POdBT1\\=9I61O01O000000001O00000000010O0000000TOYN_Dh1^;[NPDV2P<=1O1O010O0dMZD^1c<1000001O0O01000J6001O0000ZOUNWDk1i;WNgCV2Z<8O11O0aMgCW2Y\\OoA2OJS>3oA3NJS>1RB3MKQ>1TB3LJQ>2UB3JKQ>0XB4n=KSB4^>O1O1NRej2"}, "label": "white boat attached to a doc"}]} {"id": 381214, "image": "COCO_train2014_000000381214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the big boat\"."}], "task": "refering", "answer_template": "The \"the big boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 222, 241, 242, 243, 244, 245], "bbox": [0.439859375, 0.41233333333333333, 0.698171875, 0.6285000000000001], "iscrowd": 0, "area": 8863.140300000003, "rle": {"size": [480, 640], "counts": "d]T4P1P>1O001O001O1DmNiBS1W=oNgBR1W=POdBT1\\=9I61O01O000000001O00000000010O0000000TOYN_Dh1^;[NPDV2P<=1O1O010O0dMZD^1c<1000001O0O01000J6001O0000ZOUNWDk1i;WNgCV2Z<8O11O0aMgCW2Y\\OoA2OJS>3oA3NJS>1RB3MKQ>1TB3LJQ>2UB3JKQ>0XB4n=KSB4^>O1O1NRej2"}, "label": "the big boat"}]} {"id": 20769, "image": "COCO_train2014_000000020769.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white bus with no markings\"."}], "task": "refering", "answer_template": "The \"the white bus with no markings\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 53, 54, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192], "bbox": [0.0016875000000000002, 0.07425, 0.38340625000000006, 0.5226875], "iscrowd": 0, "area": 48302.0307, "rle": {"size": [480, 640], "counts": "U`0k2UVOj0WObPi5"}, "label": "the white bus with no markings"}]} {"id": 20769, "image": "COCO_train2014_000000020769.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bus behind the white suv\"."}], "task": "refering", "answer_template": "The \"a white bus behind the white suv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 53, 54, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192], "bbox": [0.0016875000000000002, 0.07425, 0.38340625000000006, 0.5226875], "iscrowd": 0, "area": 48302.0307, "rle": {"size": [480, 640], "counts": "U`0k2UVOj0WObPi5"}, "label": "a white bus behind the white suv"}]} {"id": 37160, "image": "COCO_train2014_000000037160.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"rainbow colored umbrella over a gent\"."}], "task": "refering", "answer_template": "The \"rainbow colored umbrella over a gent\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 69, 73, 74, 75, 76, 77, 92, 96, 97, 98, 99, 115, 120, 121, 122], "bbox": [0.0, 0.0, 0.461375, 0.4959452054794521], "iscrowd": 0, "area": 29157.27500000001, "rle": {"size": [365, 640], "counts": "28U;`0@f4ZKK5K5K5J5L5I7E;E;E;F:O1O1O1O1O1O1N2O1O1O1O1O1O1O1O100O1O1O100O1O100O1O100O1O100O1O1O100000000000000O100000000000000000000000000O010000000000000000000000000O10000000000000000000000001O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2Nk0UOO1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O100O1O1N2O1N2N200O10000O10000O10000O10000O10000O10000O10000N2N2O1N2N2N2N2N2N2O1N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2N2N2N3M2O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1N2ObZk3"}, "label": "rainbow colored umbrella over a gent"}]} {"id": 37160, "image": "COCO_train2014_000000037160.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a striped umbrella above a man ' s head\"."}], "task": "refering", "answer_template": "The \"a striped umbrella above a man ' s head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 69, 73, 74, 75, 76, 77, 92, 96, 97, 98, 99, 115, 120, 121, 122], "bbox": [0.0, 0.0, 0.461375, 0.4959452054794521], "iscrowd": 0, "area": 29157.27500000001, "rle": {"size": [365, 640], "counts": "28U;`0@f4ZKK5K5K5J5L5I7E;E;E;F:O1O1O1O1O1O1N2O1O1O1O1O1O1O1O100O1O1O100O1O100O1O100O1O100O1O1O100000000000000O100000000000000000000000000O010000000000000000000000000O10000000000000000000000001O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2Nk0UOO1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O100O1O1N2O1N2N200O10000O10000O10000O10000O10000O10000O10000N2N2O1N2N2N2N2N2N2O1N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2N2N2N3M2O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1N2ObZk3"}, "label": "a striped umbrella above a man ' s head"}]} {"id": 37160, "image": "COCO_train2014_000000037160.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"outer edge of a solid navy blue umbrella\"."}], "task": "refering", "answer_template": "The \"outer edge of a solid navy blue umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 42, 43, 44, 45, 67, 68], "bbox": [0.519046875, 0.00013698630136986303, 1.0, 0.20013698630136986], "iscrowd": 0, "area": 8016.207249999999, "rle": {"size": [365, 640], "counts": "[ff31[;3N1O2N1O1O2N1O1O2N1O1O2N1O1O1O00000000O10000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000O100000000000O100000000O1000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O1000000O100000000O1000000O100000000O1000000O1000000001O00001O001O00001O001O00001O00001O001O00001O001O00001O001O001O1O001O001O1O001O1O001O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O00100O001O001O1O001O1O001O001O1O001O1O001O001O1O001O001O00001O00001O00000010O0001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O000010O0001O0M"}, "label": "outer edge of a solid navy blue umbrella"}]} {"id": 37160, "image": "COCO_train2014_000000037160.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the edge of a blue umbrella\"."}], "task": "refering", "answer_template": "The \"the edge of a blue umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 42, 43, 44, 45, 67, 68], "bbox": [0.519046875, 0.00013698630136986303, 1.0, 0.20013698630136986], "iscrowd": 0, "area": 8016.207249999999, "rle": {"size": [365, 640], "counts": "[ff31[;3N1O2N1O1O2N1O1O2N1O1O2N1O1O1O00000000O10000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000O100000000000O100000000O1000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O1000000O100000000O1000000O100000000O1000000O1000000001O00001O001O00001O001O00001O00001O001O00001O001O00001O001O001O1O001O001O1O001O1O001O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O00100O001O001O1O001O1O001O001O1O001O1O001O001O1O001O001O00001O00001O00000010O0001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O000010O0001O0M"}, "label": "the edge of a blue umbrella"}]} {"id": 151854, "image": "COCO_train2014_000000151854.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a striped sweater\"."}], "task": "refering", "answer_template": "The \"a man wearing a striped sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 25, 46, 47, 48, 69, 70, 71, 92, 93, 115, 116, 138, 139, 161, 162, 163, 184, 185, 186, 207, 208, 209, 210, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 280, 299, 300, 301, 302, 303, 322, 323, 324, 325, 345, 346, 368, 369], "bbox": [0.0, 0.011229166666666667, 0.212359375, 0.9887708333333334], "iscrowd": 0, "area": 34214.2496, "rle": {"size": [480, 640], "counts": "o5;j2NP21[1e0PL\\1C`NbNMT3U3mM]2POkIc0]=^OR1000000001O000O10001O000O10001O0O10001O00001O001O001O001O001O001O001O10OXF@P5a0nJAR3G^Jh0^2Do2L_J`0a2Fk21aJ:b2Gh26cJ3c2Jf29dJNd2Ke2ZJGe5j41000O0100O0101O1N2O0O2N2N2N2M4M2N3M>A4M3M3L3N1O2N2N2M2O2N2N1O2N2M3N2N2N2N2N1O1O1O1O1O1O1N2O1N2N2O2M3N2M2O2M3N8G4L2N2N2N2N2N2O1N2N2N2N6J5Kf0ZOfl[7"}, "label": "a man wearing a striped sweater"}]} {"id": 151854, "image": "COCO_train2014_000000151854.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bald man in a white tee shirt looking at his phone\"."}], "task": "refering", "answer_template": "The \"a bald man in a white tee shirt looking at his phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 63, 64, 65, 66, 86, 87, 88, 89, 109, 110, 111, 112, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 252, 271, 272, 275], "bbox": [0.60778125, 0.09425, 1.0, 0.6788958333333334], "iscrowd": 0, "area": 38157.6511, "rle": {"size": [480, 640], "counts": "Ucf52m>3L3N3L4M3L3N3L4M3M2M4M3L4M2M4M3L4M2M4M3L4M1N3N1O2M3N1N3N1N3N2M2O2M2O2M3N1O2M101N2O1N2N2O0O2N2O1N2N2N101N2N2O1N2N2O0O2N2O1N2N2O1N2N2O1N2N2O1N2N20000O10001N10000O10000O10000O2O000O10000O10000O10001N10000O10000I7C=DCM3N2N2O1N2N2N2N2N100O1O2N1O1O1O100O1O1O1O1O100O1O1O1O1001O001O00001O000O2O0O1O2N100O2N100O2N1N3M2M3N3L3N3L5L3L4L5M2N2N3M2N3M2N2N3M2M4M2N2Ne1[N4L5K4L5K4N2M4L3M4L010O1O0010O01O1O0010O01O1O0010O01O1O0010O01O1O0010O01O1O010O001O1O010O001007H8I8G8I7I2M01O102M4L3N3L3M3M4M2M3M4M2D<\\Oe0[Od0\\Od0\\OSI"}, "label": "a bald man in a white tee shirt looking at his phone"}]} {"id": 151854, "image": "COCO_train2014_000000151854.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the black zippered shirt\"."}], "task": "refering", "answer_template": "The \"the man in the black zippered shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 63, 64, 65, 66, 86, 87, 88, 89, 109, 110, 111, 112, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 252, 271, 272, 275], "bbox": [0.60778125, 0.09425, 1.0, 0.6788958333333334], "iscrowd": 0, "area": 38157.6511, "rle": {"size": [480, 640], "counts": "Ucf52m>3L3N3L4M3L3N3L4M3M2M4M3L4M2M4M3L4M2M4M3L4M1N3N1O2M3N1N3N1N3N2M2O2M2O2M3N1O2M101N2O1N2N2O0O2N2O1N2N2N101N2N2O1N2N2O0O2N2O1N2N2O1N2N2O1N2N2O1N2N20000O10001N10000O10000O10000O2O000O10000O10000O10001N10000O10000I7C=DCM3N2N2O1N2N2N2N2N100O1O2N1O1O1O100O1O1O1O1O100O1O1O1O1001O001O00001O000O2O0O1O2N100O2N100O2N1N3M2M3N3L3N3L5L3L4L5M2N2N3M2N3M2N2N3M2M4M2N2Ne1[N4L5K4L5K4N2M4L3M4L010O1O0010O01O1O0010O01O1O0010O01O1O0010O01O1O0010O01O1O010O001O1O010O001007H8I8G8I7I2M01O102M4L3N3L3M3M4M2M3M4M2D<\\Oe0[Od0\\Od0\\OSI"}, "label": "the man in the black zippered shirt"}]} {"id": 151854, "image": "COCO_train2014_000000151854.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown table\"."}], "task": "refering", "answer_template": "The \"brown table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 173, 189, 190, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 226, 227, 228, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 249, 250, 251, 252, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.058859375000000005, 0.41954166666666665, 0.9972031249999999, 0.9845624999999999], "iscrowd": 0, "area": 102675.85559999995, "rle": {"size": [480, 640], "counts": "geb02n>1O001N101O001N101O001N101O1O0O2O001O0O2O001O001N2O001O001O001O001O001O001O1O001O0O2O001O001O001O1O001O001O001O001O0O2O001O1O001O001O001O001O001O1O0O2O001O001N101O001N101O1O0aMWNfGk1Y8ZNaGg1^8^N]Gd1b8aNWGa1h8eNQG^1n8fNlF\\1S9gNjF[1U9fNhF\\1\\7oMnHg0C]1[7TNmH`0F^1[7WNjH=H^1\\7\\NgH7J`1\\7`NdH2N_1]7eN_HN1`1_7gN[HJ3a1a7kNVHF5c1c7lNTHB4f1h7lNoGA2i1m7f301O001O1O001O0O2O001O0000000000000O1000000001O00000000001O000000001bIWHc5i7ZJYHg5g7WJZHj5f7SJ\\Hn5d7PJ^HP6c7lI_HU6a7iI`HX6`7eIcH[6]7cIdH^6n7100O1O10000000000000000000000001O00000000001O0000001O000000001O000000001O0000001O000000001O000000001O0000001O0000000000000000000000000000000000000000O10000oNPJRIP6f6]JUIc5c6hJZIX5a6PK\\IP5c6TKXIn4f6g1O1N2N2O1N2N21O00001O00001O0000001O00001O00001O00000000001O0000000000001O000000001O001O001O1O001O001O001O1O001O00001O00001O001O00001O00001O00001O001O002N2N2N3M2N2N2N3M2eI^HT5h8O001O001O001O001O00001O001O001O001O001O00001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O0000O100O1O100O100O100O1O100O100O100O1O100O100O1O1O1N2O1O1N2O1O1N200`0@001O00001O001O001O00001O001O001O1O001O001O001O1O001O001O001O1O001O000000000000O100000000O100000000O100000000O1000000O100000000O100000000O10000VOj0O1O100O100O1O100O100000000000000001O00000000000000000000001O000000000000000000001O000001O00000000000001O00000000000000000000001eKQGc2o8YMVGf2j8VMZGj2f8RM_Gm2a8PMbGP3^8lLgGS3Y8iLkGW3U8eLoG[3S8_LRH`3P8ZLTHf3n7TLWHk3k7PLXHP4]9O001O001O001O00001O001O001O001O0TOm0XNhd0"}, "label": "brown table"}]} {"id": 151854, "image": "COCO_train2014_000000151854.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table top all the drinks are on\"."}], "task": "refering", "answer_template": "The \"the table top all the drinks are on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 172, 173, 189, 190, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 226, 227, 228, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 249, 250, 251, 252, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.058859375000000005, 0.41954166666666665, 0.9972031249999999, 0.9845624999999999], "iscrowd": 0, "area": 102675.85559999995, "rle": {"size": [480, 640], "counts": "geb02n>1O001N101O001N101O001N101O1O0O2O001O0O2O001O001N2O001O001O001O001O001O001O1O001O0O2O001O001O001O1O001O001O001O001O0O2O001O1O001O001O001O001O001O1O0O2O001O001N101O001N101O1O0aMWNfGk1Y8ZNaGg1^8^N]Gd1b8aNWGa1h8eNQG^1n8fNlF\\1S9gNjF[1U9fNhF\\1\\7oMnHg0C]1[7TNmH`0F^1[7WNjH=H^1\\7\\NgH7J`1\\7`NdH2N_1]7eN_HN1`1_7gN[HJ3a1a7kNVHF5c1c7lNTHB4f1h7lNoGA2i1m7f301O001O1O001O0O2O001O0000000000000O1000000001O00000000001O000000001bIWHc5i7ZJYHg5g7WJZHj5f7SJ\\Hn5d7PJ^HP6c7lI_HU6a7iI`HX6`7eIcH[6]7cIdH^6n7100O1O10000000000000000000000001O00000000001O0000001O000000001O000000001O0000001O000000001O000000001O0000001O0000000000000000000000000000000000000000O10000oNPJRIP6f6]JUIc5c6hJZIX5a6PK\\IP5c6TKXIn4f6g1O1N2N2O1N2N21O00001O00001O0000001O00001O00001O00000000001O0000000000001O000000001O001O001O1O001O001O001O1O001O00001O00001O001O00001O00001O00001O001O002N2N2N3M2N2N2N3M2eI^HT5h8O001O001O001O001O00001O001O001O001O001O00001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O0000O100O1O100O100O100O1O100O100O100O1O100O100O1O1O1N2O1O1N2O1O1N200`0@001O00001O001O001O00001O001O001O1O001O001O001O1O001O001O001O1O001O000000000000O100000000O100000000O100000000O1000000O100000000O100000000O10000VOj0O1O100O100O1O100O100000000000000001O00000000000000000000001O000000000000000000001O000001O00000000000001O00000000000000000000001eKQGc2o8YMVGf2j8VMZGj2f8RM_Gm2a8PMbGP3^8lLgGS3Y8iLkGW3U8eLoG[3S8_LRH`3P8ZLTHf3n7TLWHk3k7PLXHP4]9O001O001O001O00001O001O001O001O0TOm0XNhd0"}, "label": "the table top all the drinks are on"}]} {"id": 37169, "image": "COCO_train2014_000000037169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red bus on a road\"."}], "task": "refering", "answer_template": "The \"red bus on a road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 263, 264, 265, 266, 267, 268], "bbox": [0.438296875, 0.4829166666666667, 0.7015625, 0.7097291666666667], "iscrowd": 0, "area": 15894.260750000001, "rle": {"size": [480, 640], "counts": "moS4f0m=i0YOf0^O9G7J6J6000001N100000000000001O0000000O10001O0000000000001O001O0O2O001O00001O000000000000O1000000O100000000O100000000000000000000000000000001O0000000000000000000O100000000000000001O000001O00000010O000001O0000010O0000001O00000O1O2N1O1O1O2N1O100O2N1000001N1000001N1000001N1000000O2O0000001O0000001O0O1O2N1O100O2N100O101N1O100O2O in this image.", "objects": [{"patches": [194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 263, 264, 265, 266, 267, 268], "bbox": [0.438296875, 0.4829166666666667, 0.7015625, 0.7097291666666667], "iscrowd": 0, "area": 15894.260750000001, "rle": {"size": [480, 640], "counts": "moS4f0m=i0YOf0^O9G7J6J6000001N100000000000001O0000000O10001O0000000000001O001O0O2O001O00001O000000000000O1000000O100000000O100000000000000000000000000000001O0000000000000000000O100000000000000001O000001O00000010O000001O0000010O0000001O00000O1O2N1O1O1O2N1O100O2N1000001N1000001N1000001N1000000O2O0000001O0000001O0O1O2N1O100O2N100O101N1O100O2O in this image.", "objects": [{"patches": [54, 55, 75, 76, 97, 98, 120, 121, 123, 124, 127, 128, 143, 144, 146, 147, 150, 151, 166, 167, 169, 170, 171, 173, 174], "bbox": [0.23365624999999998, 0.19663888888888892, 0.6018437499999999, 0.6190555555555556], "iscrowd": 0, "area": 6506.239599999999, "rle": {"size": [360, 640], "counts": "jkd1j0\\:2O1O1O1O1O1O3M5K5K6Jk0UO2N2N2N2N2N2N2N2N1N2O1OgMQH[1n7cNXHZ1g7dN_HY1`7hNcHU1\\7lNhHP1W7QOlHl0S7UOPIh0o6YOUIb0k6_OXI=h6D[I8e6I^I3b6NbIM^64eIG\\6:gIBY6?kIWOZ6i0R2O1N2O1N2O1O2M2O2N1N3N1N3N1O2N1010O01O0010O01O0010O01O00010O001O010O00XGm1]4SNdKl1[4UNeKk1[4TNfKk1Z4VNfKj1Z4VNfKj1Z4VNfKi1[4WNeKi1Z4WNgKh1Z4XNfKh1Z4XNfKb1`4^NaKo0Q5QOoJ=b5D^JJT65`200000O1000O1000O1000O10000000000000QHNh42SK6j4JQK>l4BnJg0o4YOlJo0Q5ROnJP1P5POSKP1j4POXKQ1e4oN^KQ1_4oNcKR1Z4nNiKR1T4nNoKQ1P4nNRLS1l3lNWLT1f3lN\\LS1d3lN_LP1c3oN_Lo0b3POaLm0_3SOcLj0_3UOdLh0P7N3L4M2N3L3N3M2MVdc01^R\\ONa76ZHLe78UHKj7:PHHo7^GDa8`0YGCf8b0TG@k8_10M2N2N3M4L4L4lN]F=h9DZF5i9LZFMj94YFDk9 in this image.", "objects": [{"patches": [54, 55, 75, 76, 97, 98, 120, 121, 123, 124, 127, 128, 143, 144, 146, 147, 150, 151, 166, 167, 169, 170, 171, 173, 174], "bbox": [0.23365624999999998, 0.19663888888888892, 0.6018437499999999, 0.6190555555555556], "iscrowd": 0, "area": 6506.239599999999, "rle": {"size": [360, 640], "counts": "jkd1j0\\:2O1O1O1O1O1O3M5K5K6Jk0UO2N2N2N2N2N2N2N2N1N2O1OgMQH[1n7cNXHZ1g7dN_HY1`7hNcHU1\\7lNhHP1W7QOlHl0S7UOPIh0o6YOUIb0k6_OXI=h6D[I8e6I^I3b6NbIM^64eIG\\6:gIBY6?kIWOZ6i0R2O1N2O1N2O1O2M2O2N1N3N1N3N1O2N1010O01O0010O01O0010O01O00010O001O010O00XGm1]4SNdKl1[4UNeKk1[4TNfKk1Z4VNfKj1Z4VNfKj1Z4VNfKi1[4WNeKi1Z4WNgKh1Z4XNfKh1Z4XNfKb1`4^NaKo0Q5QOoJ=b5D^JJT65`200000O1000O1000O1000O10000000000000QHNh42SK6j4JQK>l4BnJg0o4YOlJo0Q5ROnJP1P5POSKP1j4POXKQ1e4oN^KQ1_4oNcKR1Z4nNiKR1T4nNoKQ1P4nNRLS1l3lNWLT1f3lN\\LS1d3lN_LP1c3oN_Lo0b3POaLm0_3SOcLj0_3UOdLh0P7N3L4M2N3L3N3M2MVdc01^R\\ONa76ZHLe78UHKj7:PHHo7^GDa8`0YGCf8b0TG@k8_10M2N2N3M4L4L4lN]F=h9DZF5i9LZFMj94YFDk9 in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 59, 76, 77, 78, 79, 80, 81, 82, 83, 98, 99, 100, 101, 102, 103, 104, 105, 106, 121, 122, 123, 124, 125, 126, 127, 128, 129, 144, 145, 146, 147, 151, 152, 167, 168, 169, 170, 175], "bbox": [0.26954687499999996, 0.19677777777777777, 0.654734375, 0.62375], "iscrowd": 0, "area": 19912.591999999997, "rle": {"size": [360, 640], "counts": "`nl14R;5L4M2N3M3N2M2N3M2N3M2N3M5K4L4L5I6I7J6I7J6I7J7H8I1O1N101O1N110O0100O0O2N2aN]HXOd7j0hHgN[7Y1kH^NW7b1QITNR7l1T1010O010O010O10O0100O101N2O1N2O1O1N2O1gHiM_5W2PJWNS6\\32N3M3N2M2N3N2N2N2O1N21O9G9G4L1O100O1O1OmNS1C=C=L41O00000000001O0001O00000001O000001O000001O0000000001O01O0000000000010O000000000000000000000O100000000O1000000O101O00000O1000000O100000000O100000000O0100O1O001O1O1O10O01O1O1O1O010O1O1O001O100001O100O1O100O1O1O100O1O101N1O100O1O1O2O1N1O2O0O2N2O0O2N2N101N1O2O9F in this image.", "objects": [{"patches": [54, 55, 56, 57, 58, 59, 76, 77, 78, 79, 80, 81, 82, 83, 98, 99, 100, 101, 102, 103, 104, 105, 106, 121, 122, 123, 124, 125, 126, 127, 128, 129, 144, 145, 146, 147, 151, 152, 167, 168, 169, 170, 175], "bbox": [0.26954687499999996, 0.19677777777777777, 0.654734375, 0.62375], "iscrowd": 0, "area": 19912.591999999997, "rle": {"size": [360, 640], "counts": "`nl14R;5L4M2N3M3N2M2N3M2N3M2N3M5K4L4L5I6I7J6I7J6I7J7H8I1O1N101O1N110O0100O0O2N2aN]HXOd7j0hHgN[7Y1kH^NW7b1QITNR7l1T1010O010O010O10O0100O101N2O1N2O1O1N2O1gHiM_5W2PJWNS6\\32N3M3N2M2N3N2N2N2O1N21O9G9G4L1O100O1O1OmNS1C=C=L41O00000000001O0001O00000001O000001O000001O0000000001O01O0000000000010O000000000000000000000O100000000O1000000O101O00000O1000000O100000000O100000000O0100O1O001O1O1O10O01O1O1O1O010O1O1O001O100001O100O1O100O1O1O100O1O101N1O100O1O1O2O1N1O2O0O2N2O0O2N2N101N1O2O9F in this image.", "objects": [{"patches": [368, 391, 414, 415, 437, 438, 439, 440, 460, 461, 462, 463, 464, 465, 483, 484, 485, 486, 487, 488, 506, 507, 508, 509, 510, 511, 512], "bbox": [0.00225, 0.72359375, 0.292140625, 0.9842656250000001], "iscrowd": 0, "area": 14010.704499999994, "rle": {"size": [640, 640], "counts": "_R1U5m>N1O2N2N1O2N1O2N2N1O2N101N2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N1O1O2N1O2N1O001O001O001O001O00001O001O001O001O00001O001O001O001O001O00001O001O001O001O00001O001O001O001O00001O001O001O001O00001O001O001O001O00001O001O001O001O00001O001O001O0010O01O00001O001O001O001O001O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O[`j8"}, "label": "a blurry part of an animal that is not completely visible"}]} {"id": 28988, "image": "COCO_train2014_000000028988.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"ear of an animal not in the picture\"."}], "task": "refering", "answer_template": "The \"ear of an animal not in the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [368, 391, 414, 415, 437, 438, 439, 440, 460, 461, 462, 463, 464, 465, 483, 484, 485, 486, 487, 488, 506, 507, 508, 509, 510, 511, 512], "bbox": [0.00225, 0.72359375, 0.292140625, 0.9842656250000001], "iscrowd": 0, "area": 14010.704499999994, "rle": {"size": [640, 640], "counts": "_R1U5m>N1O2N2N1O2N1O2N2N1O2N101N2N1O2N1O2N2N1O2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N1O1O2N1O2N1O001O001O001O001O00001O001O001O001O00001O001O001O001O001O00001O001O001O001O00001O001O001O001O00001O001O001O001O00001O001O001O001O00001O001O001O001O00001O001O001O0010O01O00001O001O001O001O001O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O[`j8"}, "label": "ear of an animal not in the picture"}]} {"id": 28988, "image": "COCO_train2014_000000028988.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cow standing next to a tub\"."}], "task": "refering", "answer_template": "The \"a cow standing next to a tub\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 102, 103, 104, 105, 106, 107, 108, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 155, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 331, 332, 333, 334, 335, 336, 337, 338, 339, 354, 355, 356, 357, 358, 359, 360, 377, 378, 379, 380, 381, 382, 383, 400, 401, 402, 403, 404, 405, 406, 423, 424, 425, 426, 427, 447, 448, 449, 450, 470, 471, 472, 473, 493, 494, 495, 496, 516, 517, 518, 519], "bbox": [0.23421875, 0.1464375, 0.8067343750000001, 0.9887656250000001], "iscrowd": 0, "area": 101275.59214999995, "rle": {"size": [640, 640], "counts": "hPn21mc03M3M3M3M3N2M4L3M3M3N2M3M3N2M3N2N1O1O2M2O1O1O2N1O1000001O0O10001O0O10001O1O1N101O1O1O0001O000001O00000001O0000000000001O0O10000O1000001N10000O10000O2OO2M0Q^OTNma0P2H9F9N3Ne0[O8H5L3L5K5K4L4L4L4M3K5L4K8I6J7H7J7H:G?A9F9H8H7H9H6I8I5J6K5J6K5J4M4K5K4L5K5K4K6K4L5J6K;EM3M2N1O1N2O1O1O1000O10O1000000O1000000O1000000O1000000O100000`KfKZIZ4f6iKWIV4j6mKSIS4m6PLPIP4P7SLfFU1iNh2b:UL_FZ1kN`2g:YLWF_1nNX2k:\\LPFd1POQ2P;^LjEg1SOk1S;aLdEU6^:mI\\EV6d:mIVEV6j:mIPEV6P;mIjDU6W;Z10000O00fKiDMW;MoD3P;mLgEf0_O]2i:kLoE=_Og2b:jLVF5^OQ3\\:gL]F^OMk3e9fL^HZ3a7dLbH\\3^7aLeH^3[7`LhH`3W7^LlHb3T7[LoHe3P7YLSIf3m6XLVIh3j6VLWIk3h6SL[Im3d6QL_In3b6oKaIQ4^6nKdIP4]6nKfIP4V;N2N2N2N2N2N2N2N200000000000000O10000000000000_ObL_@_3i>^MRAb2V>[NeAd1c=i2WOh0N30O0001O01O01O0010O01O2N2O0O2N2O1N1O2N2O1N101N2O1N101N2O1N4M3L5L4L0O2O0O101N100fNZ1N3N1N2N3M2N2N2N3O0O101N100O1O2O0O101N1O100O2N100O1O2N2O1N2N2O1N1O2N2O1N2N2O1N2N2O0O2N3N3L3M4M2M4L3N3L3M4M2M3N2M3M3N2L3L5L4K5K5K5L3L5K5L4K5K5L3M4L4M3M4K5L4K5L4L4K4M4K5L4K5L4L:E;F:E in this image.", "objects": [{"patches": [82, 83, 102, 103, 104, 105, 106, 107, 108, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 155, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 331, 332, 333, 334, 335, 336, 337, 338, 339, 354, 355, 356, 357, 358, 359, 360, 377, 378, 379, 380, 381, 382, 383, 400, 401, 402, 403, 404, 405, 406, 423, 424, 425, 426, 427, 447, 448, 449, 450, 470, 471, 472, 473, 493, 494, 495, 496, 516, 517, 518, 519], "bbox": [0.23421875, 0.1464375, 0.8067343750000001, 0.9887656250000001], "iscrowd": 0, "area": 101275.59214999995, "rle": {"size": [640, 640], "counts": "hPn21mc03M3M3M3M3N2M4L3M3M3N2M3M3N2M3N2N1O1O2M2O1O1O2N1O1000001O0O10001O0O10001O1O1N101O1O1O0001O000001O00000001O0000000000001O0O10000O1000001N10000O10000O2OO2M0Q^OTNma0P2H9F9N3Ne0[O8H5L3L5K5K4L4L4L4M3K5L4K8I6J7H7J7H:G?A9F9H8H7H9H6I8I5J6K5J6K5J4M4K5K4L5K5K4K6K4L5J6K;EM3M2N1O1N2O1O1O1000O10O1000000O1000000O1000000O1000000O100000`KfKZIZ4f6iKWIV4j6mKSIS4m6PLPIP4P7SLfFU1iNh2b:UL_FZ1kN`2g:YLWF_1nNX2k:\\LPFd1POQ2P;^LjEg1SOk1S;aLdEU6^:mI\\EV6d:mIVEV6j:mIPEV6P;mIjDU6W;Z10000O00fKiDMW;MoD3P;mLgEf0_O]2i:kLoE=_Og2b:jLVF5^OQ3\\:gL]F^OMk3e9fL^HZ3a7dLbH\\3^7aLeH^3[7`LhH`3W7^LlHb3T7[LoHe3P7YLSIf3m6XLVIh3j6VLWIk3h6SL[Im3d6QL_In3b6oKaIQ4^6nKdIP4]6nKfIP4V;N2N2N2N2N2N2N2N200000000000000O10000000000000_ObL_@_3i>^MRAb2V>[NeAd1c=i2WOh0N30O0001O01O01O0010O01O2N2O0O2N2O1N1O2N2O1N101N2O1N101N2O1N4M3L5L4L0O2O0O101N100fNZ1N3N1N2N3M2N2N2N3O0O101N100O1O2O0O101N1O100O2N100O1O2N2O1N2N2O1N1O2N2O1N2N2O1N2N2O0O2N3N3L3M4M2M4L3N3L3M4M2M3N2M3M3N2L3L5L4K5K5K5L3L5K5L4K5K5L3M4L4M3M4K5L4K5L4L4K4M4K5L4K5L4L:E;F:E in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 31, 32, 33, 34, 35, 36, 37, 54, 55, 56, 57, 58, 59, 60, 78, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 128, 129], "bbox": [0.384859375, 0.00607981220657277, 0.6505625, 0.3455868544600939], "iscrowd": 0, "area": 21160.3112, "rle": {"size": [426, 640], "counts": "oZV33W=g0YOf0ZOg0YOg0YOg0YO3M000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O000000O1F:G9F:G9F:F:G9F:F:G9F:F:G9FWVm2"}, "label": "a white computer screen between two other computer screens"}]} {"id": 561479, "image": "COCO_train2014_000000561479.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a compute monitor that is off and is between two other monitors\"."}], "task": "refering", "answer_template": "The \"a compute monitor that is off and is between two other monitors\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 31, 32, 33, 34, 35, 36, 37, 54, 55, 56, 57, 58, 59, 60, 78, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 128, 129], "bbox": [0.384859375, 0.00607981220657277, 0.6505625, 0.3455868544600939], "iscrowd": 0, "area": 21160.3112, "rle": {"size": [426, 640], "counts": "oZV33W=g0YOf0ZOg0YOg0YOg0YO3M000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O00000000001O000000O1F:G9F:G9F:F:G9F:F:G9F:F:G9FWVm2"}, "label": "a compute monitor that is off and is between two other monitors"}]} {"id": 561479, "image": "COCO_train2014_000000561479.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small monitor on table next to package of cd disks and headphones\"."}], "task": "refering", "answer_template": "The \"small monitor on table next to package of cd disks and headphones\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 27, 28, 29, 30, 31, 50, 51, 52, 53, 54, 74, 75, 76, 77, 97], "bbox": [0.19693750000000002, 0.009507042253521126, 0.388734375, 0.2781455399061033], "iscrowd": 0, "area": 11635.96345, "rle": {"size": [426, 640], "counts": "j]d12W=7J5K5K6J5K5K6J5K5K6J5K5K6I6K5K6J5K5K6JN2000000000O10O10000000000000000000O01000000000000000000O1000O1000000000000000O100000O10000000000000O1000000000O10000000O10000000000000O100000O1000000000000000O1000O1000000000000000000O0100000000002N:F;E;dNQeR5"}, "label": "small monitor on table next to package of cd disks and headphones"}]} {"id": 561479, "image": "COCO_train2014_000000561479.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"first from the left computer monitor\"."}], "task": "refering", "answer_template": "The \"first from the left computer monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 27, 28, 29, 30, 31, 50, 51, 52, 53, 54, 74, 75, 76, 77, 97], "bbox": [0.19693750000000002, 0.009507042253521126, 0.388734375, 0.2781455399061033], "iscrowd": 0, "area": 11635.96345, "rle": {"size": [426, 640], "counts": "j]d12W=7J5K5K6J5K5K6J5K5K6J5K5K6I6K5K6J5K5K6JN2000000000O10O10000000000000000000O01000000000000000000O1000O1000000000000000O100000O10000000000000O1000000000O10000000O10000000000000O100000O1000000000000000O1000O1000000000000000000O0100000000002N:F;E;dNQeR5"}, "label": "first from the left computer monitor"}]} {"id": 561479, "image": "COCO_train2014_000000561479.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white apple monitor\"."}], "task": "refering", "answer_template": "The \"the white apple monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 60, 61, 62, 63, 64, 65, 66, 67, 68, 83, 84, 85, 86, 87, 88, 89, 90, 106, 107, 108, 109, 110, 111, 112, 113, 129, 130, 131, 132, 133, 134, 135, 136, 155, 156, 157, 158], "bbox": [0.637203125, 0.01812206572769953, 0.981234375, 0.46082159624413144], "iscrowd": 0, "area": 31458.820600000003, "rle": {"size": [426, 640], "counts": "TlY5:b<>B?@?B>B>B?A>B>A?D<01O0000001O000000010O0000001O0000001O000001O01O0000001O0000001O000001O01O0000001O0000001O000001O01O0000001O000000001O0001O01O0000001O000000001O0001O01O0000001O000000001O0001O01O0000001O000000001O0001O01O000000001O0000001O0001O01O000000001O0000001O0001O01O000000001O0000001O0001O01O000000001O0000001O0001O0001O000000M3M3M3L4M3M3M4L3M3M3L4M3M3M3M3M3L4M3M3M4L3M3M3L4M3M3M3M3M3M3L4M3M4L3M3M3L4M3M3M3M3M3M3L4M3M4L3M3M3M3L4MSo4"}, "label": "the white apple monitor"}]} {"id": 561479, "image": "COCO_train2014_000000561479.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white coloured monitor of a computer\"."}], "task": "refering", "answer_template": "The \"white coloured monitor of a computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 22, 37, 38, 39, 40, 41, 42, 43, 44, 45, 60, 61, 62, 63, 64, 65, 66, 67, 68, 83, 84, 85, 86, 87, 88, 89, 90, 106, 107, 108, 109, 110, 111, 112, 113, 129, 130, 131, 132, 133, 134, 135, 136, 155, 156, 157, 158], "bbox": [0.637203125, 0.01812206572769953, 0.981234375, 0.46082159624413144], "iscrowd": 0, "area": 31458.820600000003, "rle": {"size": [426, 640], "counts": "TlY5:b<>B?@?B>B>B?A>B>A?D<01O0000001O000000010O0000001O0000001O000001O01O0000001O0000001O000001O01O0000001O0000001O000001O01O0000001O000000001O0001O01O0000001O000000001O0001O01O0000001O000000001O0001O01O0000001O000000001O0001O01O000000001O0000001O0001O01O000000001O0000001O0001O01O000000001O0000001O0001O01O000000001O0000001O0001O0001O000000M3M3M3L4M3M3M4L3M3M3L4M3M3M3M3M3L4M3M3M4L3M3M3L4M3M3M3M3M3M3L4M3M4L3M3M3L4M3M3M3M3M3M3L4M3M4L3M3M3M3L4MSo4"}, "label": "white coloured monitor of a computer"}]} {"id": 561479, "image": "COCO_train2014_000000561479.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white keyboard beneath a monitor\"."}], "task": "refering", "answer_template": "The \"a white keyboard beneath a monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 197, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 248, 249, 250], "bbox": [0.5636875, 0.479718309859155, 0.914484375, 0.7274882629107982], "iscrowd": 0, "area": 10514.211049999998, "rle": {"size": [426, 640], "counts": "_]f42V=2N2N3M2O1N2N3M2N2N2N3M2N2N3M2N2O1N3M2N20010O00001O01O0001O01O0001O00010O0000010O00000010O0001O01O0001O0001O01O00010O00000010O0000010O0001O0001O01O0001O01O000010O0000010O000001O01O01O0001O01O0000010O00001O01O0001O01O0001O0001O01O000010O0000010O000001O01O01O0001O01O0000010O00001O01O0001O01O0001O01O01O00010O00010O000010O01O01O01O00010O00010O000010O00010O0001O01O01O01O01O00010O00010O000010O00010O0001O01O01O010O000010O00010O00010O0J6J6J7I6J6J7I6I\\df0"}, "label": "a white keyboard beneath a monitor"}]} {"id": 561479, "image": "COCO_train2014_000000561479.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a narrow , white keyboard in front of a white apple computer monitor\"."}], "task": "refering", "answer_template": "The \"a narrow , white keyboard in front of a white apple computer monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 177, 197, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223, 224, 225, 226, 227, 248, 249, 250], "bbox": [0.5636875, 0.479718309859155, 0.914484375, 0.7274882629107982], "iscrowd": 0, "area": 10514.211049999998, "rle": {"size": [426, 640], "counts": "_]f42V=2N2N3M2O1N2N3M2N2N2N3M2N2N3M2N2O1N3M2N20010O00001O01O0001O01O0001O00010O0000010O00000010O0001O01O0001O0001O01O00010O00000010O0000010O0001O0001O01O0001O01O000010O0000010O000001O01O01O0001O01O0000010O00001O01O0001O01O0001O0001O01O000010O0000010O000001O01O01O0001O01O0000010O00001O01O0001O01O0001O01O01O00010O00010O000010O01O01O01O00010O00010O000010O00010O0001O01O01O01O01O00010O00010O000010O00010O0001O01O01O010O000010O00010O00010O0J6J6J7I6J6J7I6I\\df0"}, "label": "a narrow , white keyboard in front of a white apple computer monitor"}]} {"id": 151893, "image": "COCO_train2014_000000151893.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman fixing a breakfast plate from a selection of fruit , yogurts , jams , and granola on a tray on which a green bird is perched\"."}], "task": "refering", "answer_template": "The \"a woman fixing a breakfast plate from a selection of fruit , yogurts , jams , and granola on a tray on which a green bird is perched\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 25, 26, 27, 28, 29, 30, 47, 48, 49, 50, 51, 52, 53, 70, 71, 72, 73, 74, 75, 76, 93, 94, 95, 96, 97, 98, 99, 116, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 209, 210, 211, 256, 279, 280], "bbox": [0.057906250000000006, 0.0011363636363636365, 0.3440625, 0.7691528925619836], "iscrowd": 0, "area": 35814.50629999999, "rle": {"size": [484, 640], "counts": "\\da0>Q=EgCW1S in this image.", "objects": [{"patches": [4, 5, 6, 7, 25, 26, 27, 28, 29, 30, 47, 48, 49, 50, 51, 52, 53, 70, 71, 72, 73, 74, 75, 76, 93, 94, 95, 96, 97, 98, 99, 116, 117, 118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 209, 210, 211, 256, 279, 280], "bbox": [0.057906250000000006, 0.0011363636363636365, 0.3440625, 0.7691528925619836], "iscrowd": 0, "area": 35814.50629999999, "rle": {"size": [484, 640], "counts": "\\da0>Q=EgCW1S in this image.", "objects": [{"patches": [146, 167, 168, 169, 170, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 282, 283, 284, 285, 305, 306, 307, 308], "bbox": [0.22246875, 0.3842626262626263, 0.50578125, 0.7662828282828282], "iscrowd": 0, "area": 17151.142949999994, "rle": {"size": [495, 640], "counts": "h\\U24Z?2M4L3M4L3M4M2M4L3M3M4M3L3M4L4L4K4M4K5L4K4M4K5L4M2N3L4M3L3N3M3L4M2M4M3M2SE\\Lk9e4O1N3N101N100O2O000O2O0O101N10001N100O1000O0100O01000O010O10O10O10O0100O01VOZFkKf9V4dF_K[9d4g01N2O1O1N2O1N2O1O1O1O1O1O1O1O101N1O00000000000000O10000000000oNSF`Lm9]3ZF_Lf9^3aF^L_9`3gF[LZ9b3mFZLS9g1TFCP1bNl8g1YFCR1bNe8g1_FCR1bN_8g1dFCR;9REDR;7QEGU;3nDKV;OnDNX;KlD2Y;HkD5[;DhD;\\;_OhD>^;[OfDb0U=O100O1000000000000O100000001O000O100000000000000O1000000000000O10001O0000000O10000000001O1N101O1O1O1O001N2O1Onnh4"}, "label": "a man in a red shirt skateboarding"}]} {"id": 332133, "image": "COCO_train2014_000000332133.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a red shirt performing a kick flip\"."}], "task": "refering", "answer_template": "The \"a man in a red shirt performing a kick flip\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 167, 168, 169, 170, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 282, 283, 284, 285, 305, 306, 307, 308], "bbox": [0.22246875, 0.3842626262626263, 0.50578125, 0.7662828282828282], "iscrowd": 0, "area": 17151.142949999994, "rle": {"size": [495, 640], "counts": "h\\U24Z?2M4L3M4L3M4M2M4L3M3M4M3L3M4L4L4K4M4K5L4K4M4K5L4M2N3L4M3L3N3M3L4M2M4M3M2SE\\Lk9e4O1N3N101N100O2O000O2O0O101N10001N100O1000O0100O01000O010O10O10O10O0100O01VOZFkKf9V4dF_K[9d4g01N2O1O1N2O1N2O1O1O1O1O1O1O1O101N1O00000000000000O10000000000oNSF`Lm9]3ZF_Lf9^3aF^L_9`3gF[LZ9b3mFZLS9g1TFCP1bNl8g1YFCR1bNe8g1_FCR1bN_8g1dFCR;9REDR;7QEGU;3nDKV;OnDNX;KlD2Y;HkD5[;DhD;\\;_OhD>^;[OfDb0U=O100O1000000000000O100000001O000O100000000000000O1000000000000O10001O0000000O10000000001O1N101O1O1O1O001N2O1Onnh4"}, "label": "a man in a red shirt performing a kick flip"}]} {"id": 332133, "image": "COCO_train2014_000000332133.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skateboarder who wears a hat\"."}], "task": "refering", "answer_template": "The \"the skateboarder who wears a hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 85, 86, 107, 108, 109, 110, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 174, 175, 176, 177, 198, 199, 200, 201, 221, 222, 223, 224, 225, 226, 244, 245, 247, 248, 249, 267, 268], "bbox": [0.5974999999999999, 0.14527272727272728, 0.857046875, 0.634], "iscrowd": 0, "area": 15541.893300000003, "rle": {"size": [495, 640], "counts": "jjh51[?7H8I7I7H8I7J4L3N2N1O2N2N1N3N1jDUNg8m1RGZNl8h1nF_No8b1PG`Nn8b1PG`Nm8c1QG_Nm8b1RG`Nl8b1QGaNm8`1RGbNl8`1RGbNl8`1RGbNk8`1TGbNj8`1TGbNj8`1TGbNk8^1TGdNj8^1TGdNj8]1UGeNj8\\1TGfNj8\\1TGfNj8[1UGgNj8Z1SGhNl8Z1QGhNn8Y1PGiNP9X1mFjNR9X1kFjNT9W1jFkNV9V1gFlNX9U1eFnNZ9T1cFnN]9S1`FoN_9R1_FPO`9`30WO`FcK_9\\4hF_KW9a4oFZKQ9d4UGXKj8h4_GPK`8o4Q1N100O2O0O101[MaE8`:GcE7]:HfE5[:JiE3X:LjE2V:NmENT:1oEMQ:3RFIn98TFFW9YN`FR2]GAb8?_GA_8`0cG_O\\8`0fG@X8a0jG^OU8b0lG^OR8b0QH]On7c0TH\\Oj7e0XHZOg7e0\\HZOd7a0bH^O]7?hH@W78SIGl67XIGi67ZIHe67\\IKb64`IL_63cIN\\61eIOZ60gI2W6MkI3T6MmI4R6LnI4Q6LoI6o5JRJ6m5JSJ8l5HSJ8m5HTJ8k5HUJ8k5HVJ7k5ITJ7l5ITJ7l5IUJ7k5ZOfFE_3Q1l5UOlFGY3S1l5POTGHR3W1k5lNoJT1g9O1N2N101N2N1O1O1O1O1O1O010O1O1O1O1O1O1O1O100O1O001O1O1O1O1O100N2Hdg[1"}, "label": "the skateboarder who wears a hat"}]} {"id": 332133, "image": "COCO_train2014_000000332133.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skateboarder wearing a hat performing trick\"."}], "task": "refering", "answer_template": "The \"skateboarder wearing a hat performing trick\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 85, 86, 107, 108, 109, 110, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 174, 175, 176, 177, 198, 199, 200, 201, 221, 222, 223, 224, 225, 226, 244, 245, 247, 248, 249, 267, 268], "bbox": [0.5974999999999999, 0.14527272727272728, 0.857046875, 0.634], "iscrowd": 0, "area": 15541.893300000003, "rle": {"size": [495, 640], "counts": "jjh51[?7H8I7I7H8I7J4L3N2N1O2N2N1N3N1jDUNg8m1RGZNl8h1nF_No8b1PG`Nn8b1PG`Nm8c1QG_Nm8b1RG`Nl8b1QGaNm8`1RGbNl8`1RGbNl8`1RGbNk8`1TGbNj8`1TGbNj8`1TGbNk8^1TGdNj8^1TGdNj8]1UGeNj8\\1TGfNj8\\1TGfNj8[1UGgNj8Z1SGhNl8Z1QGhNn8Y1PGiNP9X1mFjNR9X1kFjNT9W1jFkNV9V1gFlNX9U1eFnNZ9T1cFnN]9S1`FoN_9R1_FPO`9`30WO`FcK_9\\4hF_KW9a4oFZKQ9d4UGXKj8h4_GPK`8o4Q1N100O2O0O101[MaE8`:GcE7]:HfE5[:JiE3X:LjE2V:NmENT:1oEMQ:3RFIn98TFFW9YN`FR2]GAb8?_GA_8`0cG_O\\8`0fG@X8a0jG^OU8b0lG^OR8b0QH]On7c0TH\\Oj7e0XHZOg7e0\\HZOd7a0bH^O]7?hH@W78SIGl67XIGi67ZIHe67\\IKb64`IL_63cIN\\61eIOZ60gI2W6MkI3T6MmI4R6LnI4Q6LoI6o5JRJ6m5JSJ8l5HSJ8m5HTJ8k5HUJ8k5HVJ7k5ITJ7l5ITJ7l5IUJ7k5ZOfFE_3Q1l5UOlFGY3S1l5POTGHR3W1k5lNoJT1g9O1N2N101N2N1O1O1O1O1O1O010O1O1O1O1O1O1O1O100O1O001O1O1O1O1O100N2Hdg[1"}, "label": "skateboarder wearing a hat performing trick"}]} {"id": 422255, "image": "COCO_train2014_000000422255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the laptop is being used by the boy while the girl watches\"."}], "task": "refering", "answer_template": "The \"the laptop is being used by the boy while the girl watches\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 287, 307, 308, 309], "bbox": [0.303109375, 0.35752083333333334, 0.62953125, 0.8134791666666668], "iscrowd": 0, "area": 33325.335600000006, "rle": {"size": [480, 640], "counts": "dXk25g>4L5K5K4L5K4L5K5K4L5K4L5K5[Ld301O1O001O001O1O001O001O1O2N3M2N2N2O1N2N3M2N2N2N2N3M0000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000YOfGhJZ8X5iGeJW8\\5kGaJU8_5nG^JR8b5QH[Jo7e5THXJl7i5VHSJk7m5WHQJi7o5d0001O000000JRG[Jm8e5TGZJl8e5VGZJj8f5XGXJh8g5ZGYJe8g5\\GXJd8g5;O100O1]OiF]KW9c4mFXKT9i4oFSKQ9m4a0001O00001O00001O001O001O001O001O0000O1O100O1O1O1O100O1O1O1O100POVFiLk9U3XFiLi9T3[FjLf9T3]FkLc9S3aFjL`9T3cFjL^9T3eFjL\\9T3gFjLZ9V3gFiLY9V3iFhLX9X3jFdLX9[3jFbLX9^3U1000000O10000000000000000O10000000000000000O100000000000000O10000000:Eg0YOe0[O;E in this image.", "objects": [{"patches": [145, 146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 287, 307, 308, 309], "bbox": [0.303109375, 0.35752083333333334, 0.62953125, 0.8134791666666668], "iscrowd": 0, "area": 33325.335600000006, "rle": {"size": [480, 640], "counts": "dXk25g>4L5K5K4L5K4L5K5K4L5K4L5K5[Ld301O1O001O001O1O001O001O1O2N3M2N2N2O1N2N3M2N2N2N2N3M0000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000000000000000O10000000000YOfGhJZ8X5iGeJW8\\5kGaJU8_5nG^JR8b5QH[Jo7e5THXJl7i5VHSJk7m5WHQJi7o5d0001O000000JRG[Jm8e5TGZJl8e5VGZJj8f5XGXJh8g5ZGYJe8g5\\GXJd8g5;O100O1]OiF]KW9c4mFXKT9i4oFSKQ9m4a0001O00001O00001O001O001O001O001O0000O1O100O1O1O1O100O1O1O1O100POVFiLk9U3XFiLi9T3[FjLf9T3]FkLc9S3aFjL`9T3cFjL^9T3eFjL\\9T3gFjLZ9V3gFiLY9V3iFhLX9X3jFdLX9[3jFbLX9^3U1000000O10000000000000000O10000000000000000O100000000000000O10000000:Eg0YOe0[O;E in this image.", "objects": [{"patches": [116, 117, 118, 119, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236, 254, 255, 256, 257, 277, 278, 279, 280, 300, 301, 302, 303, 324, 325, 326], "bbox": [0.006562500000000001, 0.38023529411764706, 0.28529687499999995, 0.9777882352941176], "iscrowd": 0, "area": 24130.09475, "rle": {"size": [425, 640], "counts": "lj11X=3L3M4L3N3L3M4M2M3M2N100O1O100O1O1O101N1O100O1O1O2O001O001N10001O001O001O1O1O001O1O1N101O1O0hHZNm2g1SL_Of3b0ULId38XL2a3N[L=^3D^L`0a3AZLd0e3]OVLi0h3WOULm0i3TOSLQ1l3oNQLU1n3kNnKZ1Q4fNkK_1T4aNiKc1V4]NgKh1W4XNhKj1W4VNhKl1W4TNgKo1X4QNgKQ2X4oMfKT2Y4lMfKV2Y4jMeKZ2X4gMgK[2X4eMgK]2X4cMfK`2Y4`MfKb2Y4_MdKc2\\4]McKe2\\4\\MbKe2^4[M`Kg2`4ZM^Kg2b4YM\\Ki2d4XMZKi2f4XMXKi2h4WMWKj2i4WMUKj2k4WMRKk2n4VMPKk2P5WMmJj2S5WMkJj2U5XMhJj2W5WMfJk2Z5b21O0000000000000000000010OO1O1O1O1O1O1O1O1O2M2N2N2O1N3M2N2N3N1N3M2N3M2N2O2M2N3M2N3N1N2nLmHKW7KYICR7>UInNW7S1oH[N[7g1kHgM`7Z2fHTMd7n2[11N2O1N2002N1O1O2N1O1O2N1O1N01O1O1O10O01O1O10O2O1O1N2_O`EUNb:i1_EVNd:f1]EZNe:d1[E\\Ng:a1[E^Ng:`1YEaNh:]1XEcNk:Y1VEgNQ;Q1QEnNU;b0UE^OQ;3XEMS in this image.", "objects": [{"patches": [116, 117, 118, 119, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236, 254, 255, 256, 257, 277, 278, 279, 280, 300, 301, 302, 303, 324, 325, 326], "bbox": [0.006562500000000001, 0.38023529411764706, 0.28529687499999995, 0.9777882352941176], "iscrowd": 0, "area": 24130.09475, "rle": {"size": [425, 640], "counts": "lj11X=3L3M4L3N3L3M4M2M3M2N100O1O100O1O1O101N1O100O1O1O2O001O001N10001O001O001O1O1O001O1O1N101O1O0hHZNm2g1SL_Of3b0ULId38XL2a3N[L=^3D^L`0a3AZLd0e3]OVLi0h3WOULm0i3TOSLQ1l3oNQLU1n3kNnKZ1Q4fNkK_1T4aNiKc1V4]NgKh1W4XNhKj1W4VNhKl1W4TNgKo1X4QNgKQ2X4oMfKT2Y4lMfKV2Y4jMeKZ2X4gMgK[2X4eMgK]2X4cMfK`2Y4`MfKb2Y4_MdKc2\\4]McKe2\\4\\MbKe2^4[M`Kg2`4ZM^Kg2b4YM\\Ki2d4XMZKi2f4XMXKi2h4WMWKj2i4WMUKj2k4WMRKk2n4VMPKk2P5WMmJj2S5WMkJj2U5XMhJj2W5WMfJk2Z5b21O0000000000000000000010OO1O1O1O1O1O1O1O1O2M2N2N2O1N3M2N2N3N1N3M2N3M2N2O2M2N3M2N3N1N2nLmHKW7KYICR7>UInNW7S1oH[N[7g1kHgM`7Z2fHTMd7n2[11N2O1N2002N1O1O2N1O1O2N1O1N01O1O1O10O01O1O10O2O1O1N2_O`EUNb:i1_EVNd:f1]EZNe:d1[E\\Ng:a1[E^Ng:`1YEaNh:]1XEcNk:Y1VEgNQ;Q1QEnNU;b0UE^OQ;3XEMS in this image.", "objects": [{"patches": [8, 31, 54, 76, 77, 99, 100, 101, 122, 123, 124, 145, 146, 147, 148, 168, 169, 170, 171, 190, 191, 192, 213, 214, 215, 236, 237, 238, 239, 259, 260, 261, 262, 283, 284, 285, 306, 307, 308, 328, 329, 330, 351, 352, 353, 374, 375], "bbox": [0.27703125, 0.021395833333333333, 0.459453125, 0.9650833333333334], "iscrowd": 0, "area": 21187.196300000003, "rle": {"size": [480, 640], "counts": "V]c21m>2F;M3L4L4oJ4hJ1R5;bJKW5>bJGY5?aJFZ5a0^JD^5b0\\JB`5d0YJAc5f0VJ^Og5h0RJ\\Oj5l0mIXOP6o0hIVOU6P1cIUOZ6R1]ISO_6T1VI_N^OWNY7f3iHWN3QNk6b6ZI\\I[6k6jITIo5P7VJmHi5R7]JkHa5T7dJjHZ5U7kJiHS5U7\\1O1O1O_Oa0D`0UO\\O^8VOaG;f1n0eNVOZ8[8@gG>\\8^OgGa0Z8]OhGa0b1eNn2g0bKc0Z1POQ3;gKd0R1[OT3OkKg0j0EX3APLi0Kg0U4[NRLm0Cm0`4PNPLR1ZOQ1m4hMkKV1ROV1h7nNRHU1j7ROQHo0k7YORHf0j7BSH>i7JTH5h72`GaM8_2Q87dG]M:]2k7=iGWM;^2i7>jGWM:]2j7>jGWM;\\2h7`0kGVM;\\2R3WN7i2eLQOj2gN7[2mLoNl2nNNU2UMnNl2UOEP2[MoNn2YO^On4a0YKWOi4g0\\KSOg4l0^KmNe4Q1`KjMRM9a7l1bKfMXMN\\7Z2aKcM\\5[2iJ`MY5]2f3O2N2eNfCJ_<1eCN\\ in this image.", "objects": [{"patches": [8, 31, 54, 76, 77, 99, 100, 101, 122, 123, 124, 145, 146, 147, 148, 168, 169, 170, 171, 190, 191, 192, 213, 214, 215, 236, 237, 238, 239, 259, 260, 261, 262, 283, 284, 285, 306, 307, 308, 328, 329, 330, 351, 352, 353, 374, 375], "bbox": [0.27703125, 0.021395833333333333, 0.459453125, 0.9650833333333334], "iscrowd": 0, "area": 21187.196300000003, "rle": {"size": [480, 640], "counts": "V]c21m>2F;M3L4L4oJ4hJ1R5;bJKW5>bJGY5?aJFZ5a0^JD^5b0\\JB`5d0YJAc5f0VJ^Og5h0RJ\\Oj5l0mIXOP6o0hIVOU6P1cIUOZ6R1]ISO_6T1VI_N^OWNY7f3iHWN3QNk6b6ZI\\I[6k6jITIo5P7VJmHi5R7]JkHa5T7dJjHZ5U7kJiHS5U7\\1O1O1O_Oa0D`0UO\\O^8VOaG;f1n0eNVOZ8[8@gG>\\8^OgGa0Z8]OhGa0b1eNn2g0bKc0Z1POQ3;gKd0R1[OT3OkKg0j0EX3APLi0Kg0U4[NRLm0Cm0`4PNPLR1ZOQ1m4hMkKV1ROV1h7nNRHU1j7ROQHo0k7YORHf0j7BSH>i7JTH5h72`GaM8_2Q87dG]M:]2k7=iGWM;^2i7>jGWM:]2j7>jGWM;\\2h7`0kGVM;\\2R3WN7i2eLQOj2gN7[2mLoNl2nNNU2UMnNl2UOEP2[MoNn2YO^On4a0YKWOi4g0\\KSOg4l0^KmNe4Q1`KjMRM9a7l1bKfMXMN\\7Z2aKcM\\5[2iJ`MY5]2f3O2N2eNfCJ_<1eCN\\ in this image.", "objects": [{"patches": [165, 166, 186, 187, 188, 189, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 277, 278, 279, 280, 281, 282, 300, 301, 302, 303, 305, 324, 325, 347, 348, 370, 371], "bbox": [0.022593750000000003, 0.44020833333333337, 0.27468750000000003, 0.9591875000000001], "iscrowd": 0, "area": 19333.30234999999, "rle": {"size": [480, 640], "counts": "gk61k>8G9G9G9G9J6K3M2N3N1N3M2N3M2O2M2N3M2O2L4L3M4L3N3M2N3M2N3M2N3M2N3N1N1O1O00001O001O001O001O001O0YGTL[6l3dIXLY6i3eIZLY6i3lGQLX18l6h3jGXLR12R7i3iG\\Lo0KW7l3gG`Lk0F\\7m3fGdLh0_Oa7V5THoJj7X5mGlJQ8R601N101N101N2O0O2O001N101NYO[HUJe7h5_HWJa7h5bHVJ^7h5eHWJZ7i5iHUJW7i5lHVJT7i5oHUJQ7i5RIVJn6i5UIUJk6j5WIUJi6i5ZIVJf6i5[IWJe6h5\\IXJd6T2ZHV1U1eLa6T2\\HU1T1gL`6S2^HT1S1iL_6R2`H^NIn1[1aM\\6R2bH_NGl1\\1cM[6Q2dH_NGj1[1fMZ6o1gHaNDg1^1hMW6o1iHbNCd1^1kMV6n1kHcNAa1[9KVGcNA^1[9NUGeN@[1X9SNdFl1f0fN_OW1X9XNcFj1h0fN_OT1X9]N`Fh1k0gN]OR1X9`N`Ff1m0hN\\On0X9`0^GbNZOk0Y9c0_GaNZOi0X9e0`GbNXOf0Y9i0`GaNXOb0Y9m0aG`NWO`0Z9o0aGaNVO=Y9R1bGbNVO8Y9W1bGaNUO5Z9Z1cG`NUO3Y9\\1dGaNSO0Z9_1eGaNROL[9c1dGaNQOI]9e1dGbNPOF]9g1eGdNmNB_9j1fGdNP9^1oFcNo8^1QGcNm8_1QGcNl8`1SGaNk8`1UG`Nj8b1TG`Nj8b1UG^Nj8c1UG^Ni8e1VG[Nj8f1TG[Nk8g1TGYNk8h1UGXNk8i1SGWNm8k1RGUNm8l1RGTNo8m1PGSNo8o1PGPNP9T2oFhMS9]2iFaMW9d2eF[M\\9a0iEP1j0\\N_9b0lEP1d0[Nc9d0mEo0?[Nf9d0PFP18ZNj9e0RFo03YNm9g0_Fd0]:[OfEe0X:YOlEf0S:YOoEg0R:UOQFl0P:POSFQ1k9mNWFU1`NjNc:1PGe1P9ZNRGd1o8ZNUGc1l8\\NVGb1k8]NWG`1k8_NXGY1n8fNTGX1m8gNVGW1i8iN]GV1_8iNjGT1Q8kNXHMQMP1c:POfHk0k:M3L5K4I7IeTi6"}, "label": "man in purple"}]} {"id": 233841, "image": "COCO_train2014_000000233841.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a red shirt and black shorts and a hat is reaching down to pick up a white frisbee next to two women in green shirts\"."}], "task": "refering", "answer_template": "The \"a man with a red shirt and black shorts and a hat is reaching down to pick up a white frisbee next to two women in green shirts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 186, 187, 188, 189, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 277, 278, 279, 280, 281, 282, 300, 301, 302, 303, 305, 324, 325, 347, 348, 370, 371], "bbox": [0.022593750000000003, 0.44020833333333337, 0.27468750000000003, 0.9591875000000001], "iscrowd": 0, "area": 19333.30234999999, "rle": {"size": [480, 640], "counts": "gk61k>8G9G9G9G9J6K3M2N3N1N3M2N3M2O2M2N3M2O2L4L3M4L3N3M2N3M2N3M2N3M2N3N1N1O1O00001O001O001O001O001O0YGTL[6l3dIXLY6i3eIZLY6i3lGQLX18l6h3jGXLR12R7i3iG\\Lo0KW7l3gG`Lk0F\\7m3fGdLh0_Oa7V5THoJj7X5mGlJQ8R601N101N101N2O0O2O001N101NYO[HUJe7h5_HWJa7h5bHVJ^7h5eHWJZ7i5iHUJW7i5lHVJT7i5oHUJQ7i5RIVJn6i5UIUJk6j5WIUJi6i5ZIVJf6i5[IWJe6h5\\IXJd6T2ZHV1U1eLa6T2\\HU1T1gL`6S2^HT1S1iL_6R2`H^NIn1[1aM\\6R2bH_NGl1\\1cM[6Q2dH_NGj1[1fMZ6o1gHaNDg1^1hMW6o1iHbNCd1^1kMV6n1kHcNAa1[9KVGcNA^1[9NUGeN@[1X9SNdFl1f0fN_OW1X9XNcFj1h0fN_OT1X9]N`Fh1k0gN]OR1X9`N`Ff1m0hN\\On0X9`0^GbNZOk0Y9c0_GaNZOi0X9e0`GbNXOf0Y9i0`GaNXOb0Y9m0aG`NWO`0Z9o0aGaNVO=Y9R1bGbNVO8Y9W1bGaNUO5Z9Z1cG`NUO3Y9\\1dGaNSO0Z9_1eGaNROL[9c1dGaNQOI]9e1dGbNPOF]9g1eGdNmNB_9j1fGdNP9^1oFcNo8^1QGcNm8_1QGcNl8`1SGaNk8`1UG`Nj8b1TG`Nj8b1UG^Nj8c1UG^Ni8e1VG[Nj8f1TG[Nk8g1TGYNk8h1UGXNk8i1SGWNm8k1RGUNm8l1RGTNo8m1PGSNo8o1PGPNP9T2oFhMS9]2iFaMW9d2eF[M\\9a0iEP1j0\\N_9b0lEP1d0[Nc9d0mEo0?[Nf9d0PFP18ZNj9e0RFo03YNm9g0_Fd0]:[OfEe0X:YOlEf0S:YOoEg0R:UOQFl0P:POSFQ1k9mNWFU1`NjNc:1PGe1P9ZNRGd1o8ZNUGc1l8\\NVGb1k8]NWG`1k8_NXGY1n8fNTGX1m8gNVGW1i8iN]GV1_8iNjGT1Q8kNXHMQMP1c:POfHk0k:M3L5K4I7IeTi6"}, "label": "a man with a red shirt and black shorts and a hat is reaching down to pick up a white frisbee next to two women in green shirts"}]} {"id": 233841, "image": "COCO_train2014_000000233841.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the green shirt with the black sunglasses is picking up a yellow frisbee\"."}], "task": "refering", "answer_template": "The \"the man in the green shirt with the black sunglasses is picking up a yellow frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 225, 226, 227, 228, 247, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 339, 341, 342, 343, 363, 364, 366, 367], "bbox": [0.736828125, 0.5622708333333333, 0.984796875, 0.937375], "iscrowd": 0, "area": 14336.94195, "rle": {"size": [480, 640], "counts": "ddm66i>7YAD]>`000O2N1O101N1O1O2O0OO1N3N1N2OkL_OnG?P8IjG8T8NfG5W82bG0`82[G0e84TGNn84nFMR97iFJW97gFI[98bFI^9:^FFd9RFBo9?oEBR:>lECU:=jEBY:=eED\\:;cEEj0QOg7Z1]GFm0QOf7W1\\GIo0QOd7U1\\GJQ1SOc7R1ZGLT1SOa7P1ZGMW1TO^7o0XGO[1SO]7l0WG2\\1TO\\7j0VG2_1UO\\7g0TG5a1TO[7g0QG6e1TO[7^2eHcM[7\\2eHdM]7Y2dHhM]7V2cHkM^7R2dHnM]7P2cHQN\\7o1dHRN[7m1fHTNY7l1gHUNX7k1iHUNV7k1jHUNW7l1gHUNX7l1gHUNX7l1gHUNX7l1fHVNY7i1fHZNY7e1gH]NY7b1fH`NZ7^1mFPNP1e0S8Y1nF`N`08c8W1mFPO0KS9S1oFc0Q9\\OoFe0Q9YOQGg0o8XOQGi0o8WOPGj0P9WOoFi0Q9WOnFj0R9VOnF:HlMZ9j1mF0 in this image.", "objects": [{"patches": [224, 225, 226, 227, 228, 247, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 339, 341, 342, 343, 363, 364, 366, 367], "bbox": [0.736828125, 0.5622708333333333, 0.984796875, 0.937375], "iscrowd": 0, "area": 14336.94195, "rle": {"size": [480, 640], "counts": "ddm66i>7YAD]>`000O2N1O101N1O1O2O0OO1N3N1N2OkL_OnG?P8IjG8T8NfG5W82bG0`82[G0e84TGNn84nFMR97iFJW97gFI[98bFI^9:^FFd9RFBo9?oEBR:>lECU:=jEBY:=eED\\:;cEEj0QOg7Z1]GFm0QOf7W1\\GIo0QOd7U1\\GJQ1SOc7R1ZGLT1SOa7P1ZGMW1TO^7o0XGO[1SO]7l0WG2\\1TO\\7j0VG2_1UO\\7g0TG5a1TO[7g0QG6e1TO[7^2eHcM[7\\2eHdM]7Y2dHhM]7V2cHkM^7R2dHnM]7P2cHQN\\7o1dHRN[7m1fHTNY7l1gHUNX7k1iHUNV7k1jHUNW7l1gHUNX7l1gHUNX7l1gHUNX7l1fHVNY7i1fHZNY7e1gH]NY7b1fH`NZ7^1mFPNP1e0S8Y1nF`N`08c8W1mFPO0KS9S1oFc0Q9\\OoFe0Q9YOQGg0o8XOQGi0o8WOPGj0P9WOoFi0Q9WOnFj0R9VOnF:HlMZ9j1mF0 in this image.", "objects": [{"patches": [74, 75, 96, 97, 98, 119, 120, 121, 142, 143, 144, 145, 165, 166, 167, 189, 190, 212, 233, 234, 235, 236, 257, 258, 259, 281, 282, 283, 303, 304, 305, 325, 326, 327, 348, 349, 350, 371, 372], "bbox": [0.13634375, 0.21629166666666666, 0.31234375000000003, 0.9615416666666667], "iscrowd": 0, "area": 15576.943550000004, "rle": {"size": [480, 640], "counts": "fVY11n>2M4L3M3M3N3L3M3M3N2M4M2M3N2N2M3^LcN^I^1a6hNXIZ1g6mNQIU1m6POoHR1o6POnHR1P7QOmHQ1R7POlHd0]MPOf9>kH`0cMQO`9b0jH=jMPOZ9e0iH:QNPOU9h0hH6XNPOo8l0fH3_NPOQ6LPLQ1^O2fNPOi53nKm0@0lNoNc5;lKh0CNPOnN]5d0lKa0DNVOkNR5R1oK7GLZOjNh4[1WLOFLXNkNh6`1YKKEJZNlNe6b1\\KICI\\NnNa6b1`KIAG^NoN^6d1bKH@E`NPO[6f1fKE]OEbNROW6f1lKBZOFcNSOT6h1PL_OWOFeNUOQ6h1TL]OTOFgNVOn5i1ZLZOoNGiNWOk5k1^LWOlNGjNZOh5k1cLTOiNGlN[Od5n1hLoNgNHmN]O^5Q2PMiNcNIoN^OY5U2YMaN]NLQO_OT5Y2bMYNWNOSOAS5W2gMVNQN2UOBR5W2lMRNlM4VOEP5W2QNlMiM7VOHn4V2WNhMdM9WOKl4U2\\NeM`M:XONj4U2_NbM^M^1Q4Q1cN_M\\M:YO5g4T2dN]M[M9ZO8e4S2hN[MXM9[O;d4Q2kN[MTM7^O?b4P2lN[MQM6@b0a4n1eLhLP1a0WN8CJ10_4f2eLmLm0;WN8HJ4NZ4i2dLRMk06WN8KJ8LU4l2dLWMi0OWN9OI in this image.", "objects": [{"patches": [4, 5, 26, 27, 28, 49, 50, 51, 72, 73, 74, 75, 95, 96, 97, 98, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 214, 215, 230, 231, 232, 233], "bbox": [0.0016875000000000002, 0.0171875, 0.37164062500000006, 0.6382708333333333], "iscrowd": 0, "area": 29612.72215, "rle": {"size": [480, 640], "counts": "bg0P1b<^1O1O001O1O001O1O001O001O1O001O1O001O1O001O1O1O001O1O1O001O1N2O1O00100O10O0100O100O10O0100O100O10O0100O100O10O0100O100O10O0100O100O10O10O10000000O01000000O10OM4J6K5J6J6K4K6L4O100O1O010O1^MPLTJo3l5ZMkHg2T7b20000010O0001O0000001O0000001O00001O0000001O0000010O00001O0000001O0000001O000012M4L3M3N3L3M4L3N3L3M3M4M2M4L3N2M4L3M4M2M3M4L3N3L3M4M2M3M4L3N3L3M3M4M2M4L3N2O2O0O2N109F;E:F;F:E;E:GR1mNO1O101N1O10O00001N10001O0000010O01O010O00010O001O001N10001O001N101O001O000O2O001O001N10001O08I6I7J6I7IgUl5"}, "label": "the back of a mans head"}]} {"id": 135539, "image": "COCO_train2014_000000135539.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with a brown jacket on\"."}], "task": "refering", "answer_template": "The \"a person with a brown jacket on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 26, 27, 28, 49, 50, 51, 72, 73, 74, 75, 95, 96, 97, 98, 118, 119, 120, 121, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 191, 192, 207, 208, 209, 210, 211, 214, 215, 230, 231, 232, 233], "bbox": [0.0016875000000000002, 0.0171875, 0.37164062500000006, 0.6382708333333333], "iscrowd": 0, "area": 29612.72215, "rle": {"size": [480, 640], "counts": "bg0P1b<^1O1O001O1O001O1O001O001O1O001O1O001O1O001O1O1O001O1O1O001O1N2O1O00100O10O0100O100O10O0100O100O10O0100O100O10O0100O100O10O0100O100O10O10O10000000O01000000O10OM4J6K5J6J6K4K6L4O100O1O010O1^MPLTJo3l5ZMkHg2T7b20000010O0001O0000001O0000001O00001O0000001O0000010O00001O0000001O0000001O000012M4L3M3N3L3M4L3N3L3M3M4M2M4L3N2M4L3M4M2M3M4L3N3L3M4M2M3M4L3N3L3M3M4M2M4L3N2O2O0O2N109F;E:F;F:E;E:GR1mNO1O101N1O10O00001N10001O0000010O01O010O00010O001O001N10001O001N101O001O000O2O001O001N10001O08I6I7J6I7IgUl5"}, "label": "a person with a brown jacket on"}]} {"id": 135539, "image": "COCO_train2014_000000135539.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with facial hair wearing glasses eating a piece of bread\"."}], "task": "refering", "answer_template": "The \"a man with facial hair wearing glasses eating a piece of bread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.224078125, 0.0, 1.0, 0.9851875], "iscrowd": 0, "area": 184279.80469999992, "rle": {"size": [480, 640], "counts": "QQS2c0]>4K4M3M4L3M3M3M2N1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O3M2N3M2O2M2N3M2N3M2N3M2N3M2N3M2N3M2Nf1ZN2N3M3M3jHcJc4_5XKlJ`4W5[KSK]4P5_KYKY4j4bKYK]4j4_KWKa4k4\\KVKd4m4XKUKg4n4UKSKk4P5QKQKo4Q5nJQKQ5R5kJoJU5T5gJmJY5V5cJlJ\\5W5`JjJ`5X5]JiJc5Z5YJhJf5[5VJfJj5]5RJeJm5W70001O00001O0[LUGU1k8kNVGT1k8kNVGT1j8kNYGS1g8mNZGR1g8mNZGR1f8nN\\GP1d8PO]Go0d8oN^GP1b8PO`Gn0`8ROaGn0_8QObGn0^8QOeGm0[8TOeGk0\\8TOeGk0[8VOfGh0[8XOeGfNUOW1V93fGeNVOV1T96gGbNWOV1S98fGaNYOU1Q9;fG_N[OT1o8=hG]N[OT1n8?gG\\N]OS1l8b0gGZN_OR1j8d0iGXN^OS1j8e0hGWN@R1h8h0hGUNBQ1f8k0hGTNCo0f8l0iGSNCo0d8o0iGQNDo0c8Q1iGPNCo0e8P1jGoMBP1d8R1jGmMBQ1d8S1jGlMBP1e8S1kGkM@R1e8T1kGiMAR1d8V1kGgMAS1e8V1eHiN[7W1fHhN[7X1dHhN\\7Y1dHfN]7Y1cHgN^7Y1bHfNm7l0RHTOP8k0PHTOR8j0nGVOS8j0mGUOU8j0kGUOW8i0iGWOY8h0gGWO[8h0dGXO^8f0cGYO_8f0`GZOa8f0_GYOc8f0\\GZOf8d0[G[Og8d0YG[Oi8d0VG\\Ol8b0UG]Om8b0RG^Oo8b0QG]OQ9a0QG]OQ9b0PG\\OR9c0nF\\OT9c0mF\\OT9b0nF\\OS9d0nFZOT9e0kF[OW9c0iF]OY9b0fF^O\\9a0cF_O_9`0aF_O`9`0`F@b9?]FAe9>ZFBh9 in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.224078125, 0.0, 1.0, 0.9851875], "iscrowd": 0, "area": 184279.80469999992, "rle": {"size": [480, 640], "counts": "QQS2c0]>4K4M3M4L3M3M3M2N1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O3M2N3M2O2M2N3M2N3M2N3M2N3M2N3M2N3M2Nf1ZN2N3M3M3jHcJc4_5XKlJ`4W5[KSK]4P5_KYKY4j4bKYK]4j4_KWKa4k4\\KVKd4m4XKUKg4n4UKSKk4P5QKQKo4Q5nJQKQ5R5kJoJU5T5gJmJY5V5cJlJ\\5W5`JjJ`5X5]JiJc5Z5YJhJf5[5VJfJj5]5RJeJm5W70001O00001O0[LUGU1k8kNVGT1k8kNVGT1j8kNYGS1g8mNZGR1g8mNZGR1f8nN\\GP1d8PO]Go0d8oN^GP1b8PO`Gn0`8ROaGn0_8QObGn0^8QOeGm0[8TOeGk0\\8TOeGk0[8VOfGh0[8XOeGfNUOW1V93fGeNVOV1T96gGbNWOV1S98fGaNYOU1Q9;fG_N[OT1o8=hG]N[OT1n8?gG\\N]OS1l8b0gGZN_OR1j8d0iGXN^OS1j8e0hGWN@R1h8h0hGUNBQ1f8k0hGTNCo0f8l0iGSNCo0d8o0iGQNDo0c8Q1iGPNCo0e8P1jGoMBP1d8R1jGmMBQ1d8S1jGlMBP1e8S1kGkM@R1e8T1kGiMAR1d8V1kGgMAS1e8V1eHiN[7W1fHhN[7X1dHhN\\7Y1dHfN]7Y1cHgN^7Y1bHfNm7l0RHTOP8k0PHTOR8j0nGVOS8j0mGUOU8j0kGUOW8i0iGWOY8h0gGWO[8h0dGXO^8f0cGYO_8f0`GZOa8f0_GYOc8f0\\GZOf8d0[G[Og8d0YG[Oi8d0VG\\Ol8b0UG]Om8b0RG^Oo8b0QG]OQ9a0QG]OQ9b0PG\\OR9c0nF\\OT9c0mF\\OT9b0nF\\OS9d0nFZOT9e0kF[OW9c0iF]OY9b0fF^O\\9a0cF_O_9`0aF_O`9`0`F@b9?]FAe9>ZFBh9 in this image.", "objects": [{"patches": [82, 83, 104, 105, 106, 107, 127, 128, 129, 130, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 286, 287, 288, 289, 290, 291, 292, 293, 309, 310, 311, 312, 313, 314, 315, 316, 334, 335, 336, 337, 338, 339, 357, 358, 359, 360, 361, 362], "bbox": [0.4545, 0.21576036866359447, 0.771203125, 1.0], "iscrowd": 0, "area": 41957.7485, "rle": {"size": [434, 640], "counts": "Vfk32]=3M3M3N2M3M3N2N2N2N2N2N2N20001O0000000000000O10000000000000001N1O1N2O2M2O2M2O2M2N3L3]Od0\\Oc0]Od0\\Oc0]Od0]Ob0VIeJn4k5mJZJh4P6SKUJh4Q6SKSJh4R6WKoId4W6[KiI`4\\6`KdI[4a6fK^IU4h6jKXIR4l6nKTIm3R7SLmHk3U7ULkHi3X7VLhHh3Z7YLfHd3]7[LcHc3_7^L`H`3c7_L]H_3e7aL[H]3h7cLWH[3k7eLUHZ3m7eLSHY3o7hLPHW3R8hLnGV3T8jLlGU3V8kLiGS3Y8mLgGR3[8nLeGo2]8QMcGn2_8QMaGm2a8TM^Gk2d8TM\\Gj2f8VMZGi2h8WMWGg2k8YMUGf2m8YMSGg2m8ZMRGf2n8ZMRGe2P9ZMPGf2P9[MPGd2Q9[MoFd2R9]MmFc2T9\\MlFd2T9\\MlFd2U9\\MjFc2W9]MiFc2X9\\MhFd2X9]MgFb2[9]MeFc2[9]MeFc2\\9]McFb2^9^MbFb2_9^MaFa2_9_MaFa2_9_MaF`2a9`M^F`2b9`M^F`2c9_M]F`2d9`M\\F`2d9`M\\F`2e9_M[Fa2e9`MZF`2g9_MYFb2f9^MZFb2f9^M[Fb2e9]M[Fc2e9]M[Fd2d9\\M\\Fd2e9[M[Ff2d9ZM\\Ff2e9ZMZFg2e9YM[Fg2e9YM[Fh2e9WM[Fi2e9WM[Fj2d9VM\\Fj2e9UM[Fl2d9TM]Fk2d9UM[Fl2d9TM\\Fm2c9SM]Fn2c9QM]Fo2c9QM]FP3b9PM^FQ3b9nL^FS3a9mL_FS3b9lL^FU3a9lL^FU3a9kL_FV3a9iL`FW3_9iLaFW3`9hL`FY3_9gLaFZ3_9eLaF\\3_9cLaF]3_9cLaFa3\\9_LcFe3Y9[LgFi3V9VLjFo3R9PLnFT4n8lKRGX4k8gKUG^4f8bK[Ga4b8^K^Gg4^8XKbGl4Z8UKeGo4X8PKhGU5S8kJmGY5P8fJPH]5n7bJRHa5k7_JUHd5k73M3M4M2M3M3M3M3M3M3K5L4L4L4K5L4L4L4M3M2M4M2N2N3L3N2N3M2M3N3M2M3N3M2N2M6K9G:C=\\Oc0^Oc0\\Od0]Ob0]Od0]OXam1"}, "label": "blonde haired woman in a lace blouse"}]} {"id": 323960, "image": "COCO_train2014_000000323960.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"thin blonde haired woman who has a big smile\"."}], "task": "refering", "answer_template": "The \"thin blonde haired woman who has a big smile\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 104, 105, 106, 107, 127, 128, 129, 130, 150, 151, 152, 153, 154, 173, 174, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 224, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 286, 287, 288, 289, 290, 291, 292, 293, 309, 310, 311, 312, 313, 314, 315, 316, 334, 335, 336, 337, 338, 339, 357, 358, 359, 360, 361, 362], "bbox": [0.4545, 0.21576036866359447, 0.771203125, 1.0], "iscrowd": 0, "area": 41957.7485, "rle": {"size": [434, 640], "counts": "Vfk32]=3M3M3N2M3M3N2N2N2N2N2N2N20001O0000000000000O10000000000000001N1O1N2O2M2O2M2O2M2N3L3]Od0\\Oc0]Od0\\Oc0]Od0]Ob0VIeJn4k5mJZJh4P6SKUJh4Q6SKSJh4R6WKoId4W6[KiI`4\\6`KdI[4a6fK^IU4h6jKXIR4l6nKTIm3R7SLmHk3U7ULkHi3X7VLhHh3Z7YLfHd3]7[LcHc3_7^L`H`3c7_L]H_3e7aL[H]3h7cLWH[3k7eLUHZ3m7eLSHY3o7hLPHW3R8hLnGV3T8jLlGU3V8kLiGS3Y8mLgGR3[8nLeGo2]8QMcGn2_8QMaGm2a8TM^Gk2d8TM\\Gj2f8VMZGi2h8WMWGg2k8YMUGf2m8YMSGg2m8ZMRGf2n8ZMRGe2P9ZMPGf2P9[MPGd2Q9[MoFd2R9]MmFc2T9\\MlFd2T9\\MlFd2U9\\MjFc2W9]MiFc2X9\\MhFd2X9]MgFb2[9]MeFc2[9]MeFc2\\9]McFb2^9^MbFb2_9^MaFa2_9_MaFa2_9_MaF`2a9`M^F`2b9`M^F`2c9_M]F`2d9`M\\F`2d9`M\\F`2e9_M[Fa2e9`MZF`2g9_MYFb2f9^MZFb2f9^M[Fb2e9]M[Fc2e9]M[Fd2d9\\M\\Fd2e9[M[Ff2d9ZM\\Ff2e9ZMZFg2e9YM[Fg2e9YM[Fh2e9WM[Fi2e9WM[Fj2d9VM\\Fj2e9UM[Fl2d9TM]Fk2d9UM[Fl2d9TM\\Fm2c9SM]Fn2c9QM]Fo2c9QM]FP3b9PM^FQ3b9nL^FS3a9mL_FS3b9lL^FU3a9lL^FU3a9kL_FV3a9iL`FW3_9iLaFW3`9hL`FY3_9gLaFZ3_9eLaF\\3_9cLaF]3_9cLaFa3\\9_LcFe3Y9[LgFi3V9VLjFo3R9PLnFT4n8lKRGX4k8gKUG^4f8bK[Ga4b8^K^Gg4^8XKbGl4Z8UKeGo4X8PKhGU5S8kJmGY5P8fJPH]5n7bJRHa5k7_JUHd5k73M3M4M2M3M3M3M3M3M3K5L4L4L4K5L4L4L4M3M2M4M2N2N3L3N2N3M2M3N3M2M3N3M2N2M6K9G:C=\\Oc0^Oc0\\Od0]Ob0]Od0]OXam1"}, "label": "thin blonde haired woman who has a big smile"}]} {"id": 323960, "image": "COCO_train2014_000000323960.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a couple smiling for a picture\"."}], "task": "refering", "answer_template": "The \"a couple smiling for a picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 55, 56, 57, 58, 59, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 307, 308, 309, 310, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361], "bbox": [0.22756249999999997, 0.05854838709677419, 0.71171875, 0.997741935483871], "iscrowd": 0, "area": 65700.0881, "rle": {"size": [434, 640], "counts": "]Vn1k1i9Q2WOj0[Oa0_Ob0VIZJb5j5[JYJc5i5ZJZJc5j5YJYJe5i5XJZJe5i5XJZJf5i5VJZJg5i5VJZJh5h5UJ[Ji5g5TJ\\Ji5h6N2M3N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O100O100O10000O100O10000O100O100O10000O100O10000O100O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O1O2N1O1O1N2O1O1O1O1O1M3N2N2M3N2M3N2N2M3N2N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O100O1O1O1O1O1O1O1O1O1O1O[MdNVI[1h6UOkHk0R7ZOlHe0S7_OkH`0V7BhH=Y7DfH;[7FdH9]7HbH7_7IaH6`7J`H5`7L`H3a7M_H2b7N^H1c7O]H1c70\\HOe71[HOe71[HNf72ZHNf72ZHMY2_Oj0d0mLMR2Go0=oLKR2Jn0;PMKR2Km0;PMIS2Nl09QMIR20l07RMHS22i07TMFS26g04VMFR28f03XMDS2:d02YMDR2a00\\MBR2`0a0N]MAR2d0?K_MAR2d0>L`M@Q2f0>K`M_OQ2h0>IaM_OQ2i0 in this image.", "objects": [{"patches": [33, 34, 35, 36, 55, 56, 57, 58, 59, 78, 79, 80, 81, 82, 100, 101, 102, 103, 104, 105, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 307, 308, 309, 310, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361], "bbox": [0.22756249999999997, 0.05854838709677419, 0.71171875, 0.997741935483871], "iscrowd": 0, "area": 65700.0881, "rle": {"size": [434, 640], "counts": "]Vn1k1i9Q2WOj0[Oa0_Ob0VIZJb5j5[JYJc5i5ZJZJc5j5YJYJe5i5XJZJe5i5XJZJf5i5VJZJg5i5VJZJh5h5UJ[Ji5g5TJ\\Ji5h6N2M3N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O100O100O10000O100O10000O100O100O10000O100O10000O100O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O1O2N1O1O1N2O1O1O1O1O1M3N2N2M3N2M3N2N2M3N2N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O100O1O1O1O1O1O1O1O1O1O1O[MdNVI[1h6UOkHk0R7ZOlHe0S7_OkH`0V7BhH=Y7DfH;[7FdH9]7HbH7_7IaH6`7J`H5`7L`H3a7M_H2b7N^H1c7O]H1c70\\HOe71[HOe71[HNf72ZHNf72ZHMY2_Oj0d0mLMR2Go0=oLKR2Jn0;PMKR2Km0;PMIS2Nl09QMIR20l07RMHS22i07TMFS26g04VMFR28f03XMDS2:d02YMDR2a00\\MBR2`0a0N]MAR2d0?K_MAR2d0>L`M@Q2f0>K`M_OQ2h0>IaM_OQ2i0 in this image.", "objects": [{"patches": [9, 10, 11, 12, 32, 33, 34, 35, 55, 56, 57, 58, 59, 78, 79, 80, 81, 82, 83, 103, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 334, 335], "bbox": [0.397921875, 0.009227166276346604, 0.762921875, 0.98903981264637], "iscrowd": 0, "area": 47848.85605000002, "rle": {"size": [427, 640], "counts": "VdZ31Y=1N3nFMi44VKOg43WK0f41YK3c4N\\K8^4J`KmMX2f5hM\\JW2d5hM]JX2c5gM_JX2a5dMcJ\\2]5_MiJ`2Q801O1O1O1kNYEDh:9ZEGg:5\\EKe:2]EMe:O_E0b:LaE4`:IcE6^:FfE9[:DiE:X:@nE?R:[OVFc0[;N1O2N2N1O2N2N12M3M2N3N2M4L7I6J7I5K6J6J3M2N2N1O2N2N1O2N2N2N1O2N2N2N1O1O001O1O001O001O1L3I8I7H7J7H7J7H8N101O001N2O001N101O1O0O2O1O001N101O1N101O001N2O001QLTK]Ol4b0hKjNZ4T1ZLYNg3e1lLiMT3V2`MVMb2i2QNdLP2Z3Q4O2N2N101N2N1O2N2O0O2N2N2O0O2N2N1O2O1N2N1O2N2O0O2N2N1O2O1N2N1O2N2O0O2N2N2N101N2N1O2O1N1O2N2N2O0O2N2N3M2O2M2N3M2N3N1N3M2N3M2O2M4L4L4LeVo1"}, "label": "a man wearing a hat"}]} {"id": 545145, "image": "COCO_train2014_000000545145.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man cooking with a maroon shirt on and a black hat\"."}], "task": "refering", "answer_template": "The \"a man cooking with a maroon shirt on and a black hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 32, 33, 34, 35, 55, 56, 57, 58, 59, 78, 79, 80, 81, 82, 83, 103, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 310, 311, 312, 334, 335], "bbox": [0.397921875, 0.009227166276346604, 0.762921875, 0.98903981264637], "iscrowd": 0, "area": 47848.85605000002, "rle": {"size": [427, 640], "counts": "VdZ31Y=1N3nFMi44VKOg43WK0f41YK3c4N\\K8^4J`KmMX2f5hM\\JW2d5hM]JX2c5gM_JX2a5dMcJ\\2]5_MiJ`2Q801O1O1O1kNYEDh:9ZEGg:5\\EKe:2]EMe:O_E0b:LaE4`:IcE6^:FfE9[:DiE:X:@nE?R:[OVFc0[;N1O2N2N1O2N2N12M3M2N3N2M4L7I6J7I5K6J6J3M2N2N1O2N2N1O2N2N2N1O2N2N2N1O1O001O1O001O001O1L3I8I7H7J7H7J7H8N101O001N2O001N101O1O0O2O1O001N101O1N101O001N2O001QLTK]Ol4b0hKjNZ4T1ZLYNg3e1lLiMT3V2`MVMb2i2QNdLP2Z3Q4O2N2N101N2N1O2N2O0O2N2N2O0O2N2N1O2O1N2N1O2N2O0O2N2N1O2O1N2N1O2N2O0O2N2N2N101N2N1O2O1N1O2N2N2O0O2N2N3M2O2M2N3M2N3N1N3M2N3M2O2M4L4L4LeVo1"}, "label": "a man cooking with a maroon shirt on and a black hat"}]} {"id": 545145, "image": "COCO_train2014_000000545145.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the teen reaching for something on the table who is not wearing a hat\"."}], "task": "refering", "answer_template": "The \"the teen reaching for something on the table who is not wearing a hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 52, 73, 74, 75, 96, 97, 98, 118, 119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 191, 208, 209, 230, 231, 232, 253, 254, 276], "bbox": [0.003, 0.13175644028103045, 0.36665625, 0.816440281030445], "iscrowd": 0, "area": 22725.068199999998, "rle": {"size": [427, 640], "counts": "bU11jW15TUO7I7N2O1O1N2O1O1O0O2O1O1O1N2O1O1O1O001O100O1O1O1OO2N10O0O1O1O1O100O2N1O2N101N2N1O2N101N1O2N1O2O1L3M4L2002N2N1O2N110]OkDQOT;Q1mDmNS;T1nDjNR;V1PEhNo:Z1QEdNP;\\1REbNm:`1TE^Nl:b1VE\\Nj:e1VEZNi:g1YEWNg:j1;2N101N101N1H9B>C in this image.", "objects": [{"patches": [50, 51, 52, 73, 74, 75, 96, 97, 98, 118, 119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 185, 186, 187, 188, 189, 190, 191, 208, 209, 230, 231, 232, 253, 254, 276], "bbox": [0.003, 0.13175644028103045, 0.36665625, 0.816440281030445], "iscrowd": 0, "area": 22725.068199999998, "rle": {"size": [427, 640], "counts": "bU11jW15TUO7I7N2O1O1N2O1O1O0O2O1O1O1N2O1O1O1O001O100O1O1O1OO2N10O0O1O1O1O100O2N1O2N101N2N1O2N101N1O2N1O2O1L3M4L2002N2N1O2N110]OkDQOT;Q1mDmNS;T1nDjNR;V1PEhNo:Z1QEdNP;\\1REbNm:`1TE^Nl:b1VE\\Nj:e1VEZNi:g1YEWNg:j1;2N101N101N1H9B>C in this image.", "objects": [{"patches": [39, 40, 56, 57, 58, 73, 74, 75, 76, 90, 91, 92, 93, 107, 108, 109, 110, 123, 124, 125, 126, 127, 139, 140, 141, 143, 144, 156, 161], "bbox": [0.18877083333333333, 0.10525, 0.5273333333333333, 0.40635937499999997], "iscrowd": 0, "area": 12210.7657, "rle": {"size": [640, 480], "counts": "iSi12mc01N3N2N2N1O2N2M3N1O2N1O000O10O2O001O1N101O001O0O2O10O01O010O010O1O010O010O00100O010O0010O0100O0010O010O010O1O1O1O1O1O001O1O1O1RMA[B?eo0bAf0^>[O`Af0_>\\O^Af0b>\\OZAe0g>\\OWAe0i>\\OTAf0l>[OQAg0o>[On@f0R?[Ok@g0U?R20O1000000000000O100000O1001O2M3N2N2N2N2N2N2M3N2N2N3M2N3M3L3N3M3M2O11N1N2N2O0O1O00100O00100O1O00100O001O100O00100O001O1001N2O1O001N2O1O1N3N3UN\\_O3g`0I\\_O5g`0F]_O7g`0E\\_O9f`0C^_O;e`0A^_O in this image.", "objects": [{"patches": [39, 40, 56, 57, 58, 73, 74, 75, 76, 90, 91, 92, 93, 107, 108, 109, 110, 123, 124, 125, 126, 127, 139, 140, 141, 143, 144, 156, 161], "bbox": [0.18877083333333333, 0.10525, 0.5273333333333333, 0.40635937499999997], "iscrowd": 0, "area": 12210.7657, "rle": {"size": [640, 480], "counts": "iSi12mc01N3N2N2N1O2N2M3N1O2N1O000O10O2O001O1N101O001O0O2O10O01O010O010O1O010O010O00100O010O0010O0100O0010O010O010O1O1O1O1O1O001O1O1O1RMA[B?eo0bAf0^>[O`Af0_>\\O^Af0b>\\OZAe0g>\\OWAe0i>\\OTAf0l>[OQAg0o>[On@f0R?[Ok@g0U?R20O1000000000000O100000O1001O2M3N2N2N2N2N2N2M3N2N2N3M2N3M3L3N3M3M2O11N1N2N2O0O1O00100O00100O1O00100O001O100O00100O001O1001N2O1O001N2O1O1N3N3UN\\_O3g`0I\\_O5g`0F]_O7g`0E\\_O9f`0C^_O;e`0A^_O in this image.", "objects": [{"patches": [10, 26, 27, 28, 43, 44, 45, 46, 59, 60, 61, 62, 63, 76, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 111, 112, 113, 114, 128, 129, 130, 146, 147, 163, 164, 180, 181], "bbox": [0.5183541666666667, 0.033859375000000004, 0.7790208333333334, 0.44509375], "iscrowd": 0, "area": 18985.06880000001, "rle": {"size": [640, 480], "counts": "Ygk44jc06J6J6K5J6J6J6_^OfNR`0`1h_OfNR`0`1i_OfNP`0`1j_OfNP`0a1j_OeNo?]1o_OeNo?]1o_OeNP`0\\1o_OeNQ`0\\1l_OgNR`0Z1m_OgNS`0Y1k_OiNU`0W1j_OkNT`0V1j_OlNV`0`2N10001N101SOl0N3M2N2N3M2N3M2N3M6J6J7I6L4gCjI`:[6UElIj:X6lDnIT;W6aDoI_;T6XDSJg;S70001O0O101O00001O00001O000000O103M3M3M3M3M3M2M3N3M2N2N1O1O1O1O1N2O2cNUCaKl in this image.", "objects": [{"patches": [10, 26, 27, 28, 43, 44, 45, 46, 59, 60, 61, 62, 63, 76, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 111, 112, 113, 114, 128, 129, 130, 146, 147, 163, 164, 180, 181], "bbox": [0.5183541666666667, 0.033859375000000004, 0.7790208333333334, 0.44509375], "iscrowd": 0, "area": 18985.06880000001, "rle": {"size": [640, 480], "counts": "Ygk44jc06J6J6K5J6J6J6_^OfNR`0`1h_OfNR`0`1i_OfNP`0`1j_OfNP`0a1j_OeNo?]1o_OeNo?]1o_OeNP`0\\1o_OeNQ`0\\1l_OgNR`0Z1m_OgNS`0Y1k_OiNU`0W1j_OkNT`0V1j_OlNV`0`2N10001N101SOl0N3M2N2N3M2N3M2N3M6J6J7I6L4gCjI`:[6UElIj:X6lDnIT;W6aDoI_;T6XDSJg;S70001O0O101O00001O00001O000000O103M3M3M3M3M3M2M3N3M2N2N1O1O1O1O1N2O2cNUCaKl in this image.", "objects": [{"patches": [175, 176, 177, 178, 179, 198, 199, 200, 201, 202, 221, 222, 223, 224, 225, 244, 245, 246, 248, 268, 269], "bbox": [0.603234375, 0.4478125, 0.82725, 0.6977499999999999], "iscrowd": 0, "area": 9422.36735, "rle": {"size": [480, 640], "counts": "aUe51m>4L5J:G4K6K4L4M2O2M2O1N1O101O001N2UCbNR<^1kCiNQM3M3O1O1O1000001O00001O00001O0010O01O00010O0003M7I7Ih0XOO100O10O1O00105J9G1O00100O010OSOfDVNZ;j1kDQNU;n1TEcMS;]2h0O1N1O2N2O1N2O2N1O1K5N2N2O1N2N2O1M3M3N2N2N210O01O001O2O1N1O2N2N1O2O0O2N1O2N100O2N101N1O2N100O0001O010O1O100O001N2N2N2N2N6K4K4L3M8XOUBLS>KPB3T>ImA6W>ElA7mVd1"}, "label": "zebra in a cage"}]} {"id": 504187, "image": "COCO_train2014_000000504187.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra eating hay standing next to another zebra\"."}], "task": "refering", "answer_template": "The \"a zebra eating hay standing next to another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 286, 287, 288], "bbox": [0.378453125, 0.47375, 0.6405000000000001, 0.7245833333333334], "iscrowd": 0, "area": 10889.367749999992, "rle": {"size": [480, 640], "counts": "Sfa32l>6J7H6K2O19G8H2M01O100O1O1M2M4K5L5M4L4L4M3L5K4L4N2M7J6Jd0[O9H0O2O0O2O2M8I000O100000hNcDTOV;k0PEoNQ;P1UEjNl:U1[EdNf:[1`E_Na:`1eEZN\\:e1kETNU:l1`1O1O10000001O0000000000000O100001O000001O000000000001O000000bDSNi9e2ZE`Mf:d2QEaMo:c2hDbMX;S300001O00000\\Od0K5C=H8N2N1N3O100001O0000R1nN7I6J5K000001O1L4XOh0@`0M3M3M3M3M3F9L4001O00001O0000?A2O2M3N1N2O1N2O2M2O1N2O1N2O1N100O2O01N3M4L4L3L5L00O0010O01O^O]CdNg in this image.", "objects": [{"patches": [48, 49, 50, 66, 67, 68, 69, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 140, 141, 143], "bbox": [0.63882, 0.15293333333333334, 1.0, 0.61176], "iscrowd": 0, "area": 16458.5136, "rle": {"size": [375, 500], "counts": "Pod31d;4M2M4L3N3L4L3N3L4L3N3L4L3N3bN]1N3M3M2N3M2O0O2N1O2N101N1O1O2N1O2O0O2N1O2N1O2O0O1O2N1O2N101N1O2N110O00001O0010O01O001O00010O001O001O01O01O001O0010O01O00001O010O010O010O000100O100O2O0O100O1fLgHT2\\7hMfHX2\\7dMeH\\2]7aMeH_2]7\\MeHl1JVNc7KeHm1JXNc7GdHP2KXNd7DcHQ2K\\Nc7_OdHT2K\\Nd7\\OcHV2K^NS8`1oGaNQ8]1QHcNn7[1UHeNk7Z1VHfNj7Y1VHhNi7X1XHhNh7W1YHiNf7W1ZHjNf7U1[HdNk7\\1VH[NS8d1P110O001O001O001O010O001O001O1O0010O01O001O001O001O010O001O001O001O010O001O1O02O0O100O2O0O100O2O0O101N100O101N100O101O0O100O2O0O101N100QK"}, "label": "a blurry looking piece of broccoli sitting on a plate of other food"}]} {"id": 192891, "image": "COCO_train2014_000000192891.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"broccoli on a table plate to potatoes\"."}], "task": "refering", "answer_template": "The \"broccoli on a table plate to potatoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 49, 50, 66, 67, 68, 69, 83, 84, 85, 86, 87, 88, 101, 102, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 140, 141, 143], "bbox": [0.63882, 0.15293333333333334, 1.0, 0.61176], "iscrowd": 0, "area": 16458.5136, "rle": {"size": [375, 500], "counts": "Pod31d;4M2M4L3N3L4L3N3L4L3N3L4L3N3bN]1N3M3M2N3M2O0O2N1O2N101N1O1O2N1O2O0O2N1O2N1O2O0O1O2N1O2N101N1O2N110O00001O0010O01O001O00010O001O001O01O01O001O0010O01O00001O010O010O010O000100O100O2O0O100O1fLgHT2\\7hMfHX2\\7dMeH\\2]7aMeH_2]7\\MeHl1JVNc7KeHm1JXNc7GdHP2KXNd7DcHQ2K\\Nc7_OdHT2K\\Nd7\\OcHV2K^NS8`1oGaNQ8]1QHcNn7[1UHeNk7Z1VHfNj7Y1VHhNi7X1XHhNh7W1YHiNf7W1ZHjNf7U1[HdNk7\\1VH[NS8d1P110O001O001O001O010O001O001O1O0010O01O001O001O001O010O001O001O001O010O001O1O02O0O100O2O0O100O2O0O101N100O101N100O101O0O100O2O0O101N100QK"}, "label": "broccoli on a table plate to potatoes"}]} {"id": 192891, "image": "COCO_train2014_000000192891.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the broccoli under the other broccoli\"."}], "task": "refering", "answer_template": "The \"the broccoli under the other broccoli\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 141, 142, 143, 160, 161, 178, 179, 197], "bbox": [0.82268, 0.46152, 1.0, 0.8426400000000001], "iscrowd": 0, "area": 5489.088300000003, "rle": {"size": [375, 500], "counts": "lQg42e;1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O100O100O100O100O100O1`FgN`8Z1\\GlNb8T1YGTOc8m0XG[Oe8e0VGBg8h100O2O0O101N100O2N100O2O0O100O2O0O101N100O1O101N1001O2N3M2O1N2N2N3M2N2N2O1N3M2M3N2M3M4L3N2M3M3N3LgG"}, "label": "the broccoli under the other broccoli"}]} {"id": 192891, "image": "COCO_train2014_000000192891.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the small bit of broccoli that is pointing downwards , under the larger bit of broccoli\"."}], "task": "refering", "answer_template": "The \"the small bit of broccoli that is pointing downwards , under the larger bit of broccoli\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 141, 142, 143, 160, 161, 178, 179, 197], "bbox": [0.82268, 0.46152, 1.0, 0.8426400000000001], "iscrowd": 0, "area": 5489.088300000003, "rle": {"size": [375, 500], "counts": "lQg42e;1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O100O100O100O100O100O1`FgN`8Z1\\GlNb8T1YGTOc8m0XG[Oe8e0VGBg8h100O2O0O101N100O2N100O2O0O100O2O0O101N100O1O101N1001O2N3M2O1N2N2N3M2N2N2O1N3M2M3N2M3M4L3N2M3M3N3LgG"}, "label": "the small bit of broccoli that is pointing downwards , under the larger bit of broccoli"}]} {"id": 4477, "image": "COCO_train2014_000000004477.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant who has it ' s trunk curving upwards\"."}], "task": "refering", "answer_template": "The \"the elephant who has it ' s trunk curving upwards\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 203, 204, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 226, 227, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 249, 250, 251, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 272, 273, 274, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 299, 300, 301, 302, 303, 304, 305, 306, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 322, 323, 324, 325, 326, 327, 328, 329, 334, 335, 336, 337, 338, 339, 340, 341, 342, 345, 346, 347, 348, 349, 350, 351, 352, 358, 359, 363, 364, 368, 369, 370, 371, 372, 373, 374, 375, 391, 392, 393, 394, 395, 396, 397, 398], "bbox": [0.0, 0.03371541501976284, 0.93809375, 0.9887549407114624], "iscrowd": 0, "area": 190879.968, "rle": {"size": [506, 640], "counts": "V<^3a5k6VKj4001O00001O00001O0000001O00001O00001O0000001O00001O00000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001O0000000000001O0000000000001O00000000001O0000000000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000O10^KbNhJ^1W5T1XHlNh7c400000O10000000000O10O11O001O001N2O001O001O1O0O2O001O1O001N101O1O001O001N2O001O1O001O001O1O000000010O0000000000001O000000000000001O00000001O0001O0000000000001O0000000000001O000001O00000001O0000000000001O0000000000010O01O00001O0010O0001O001O000010O01O001O000100O100O1O2O0O100O100O2N100O100O1O2O0O100O1O2O0O100O100O2N100O100O1O100O100O00100O1001O001O1N101O1O0O2J6J6K4K6J6K4K6J6K4K6J6J501N2O0O2N2O0O2O1N101N2O1N101N4M3L3N3L4M3L4M2M4L4M3L4M2M4M3L4M2M4M3L4M3L3N4K4M4K5K4M4K5L3L5L4K4M4K5L3L3N1N101N2O0O2O1N2N101N2O1N1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N1O00010O001O00001O00001O00001O001O00001O00001O00001O001O00001O000010O0001O001OfLgMSIY2l6nMnHR2Q7UNjHj1U7\\NfHd1Y7cNaH]1^7jN]HU1b7ROXHn0g7YORHh0m7_OkGc0T8CdG`0[8G\\G in this image.", "objects": [{"patches": [0, 1, 2, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 203, 204, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 226, 227, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 249, 250, 251, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 272, 273, 274, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 299, 300, 301, 302, 303, 304, 305, 306, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 322, 323, 324, 325, 326, 327, 328, 329, 334, 335, 336, 337, 338, 339, 340, 341, 342, 345, 346, 347, 348, 349, 350, 351, 352, 358, 359, 363, 364, 368, 369, 370, 371, 372, 373, 374, 375, 391, 392, 393, 394, 395, 396, 397, 398], "bbox": [0.0, 0.03371541501976284, 0.93809375, 0.9887549407114624], "iscrowd": 0, "area": 190879.968, "rle": {"size": [506, 640], "counts": "V<^3a5k6VKj4001O00001O00001O0000001O00001O00001O0000001O00001O00000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001O0000000000001O0000000000001O00000000001O0000000000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000O10^KbNhJ^1W5T1XHlNh7c400000O10000000000O10O11O001O001N2O001O001O1O0O2O001O1O001N101O1O001O001N2O001O1O001O001O1O000000010O0000000000001O000000000000001O00000001O0001O0000000000001O0000000000001O000001O00000001O0000000000001O0000000000010O01O00001O0010O0001O001O000010O01O001O000100O100O1O2O0O100O100O2N100O100O1O2O0O100O1O2O0O100O100O2N100O100O1O100O100O00100O1001O001O1N101O1O0O2J6J6K4K6J6K4K6J6K4K6J6J501N2O0O2N2O0O2O1N101N2O1N101N4M3L3N3L4M3L4M2M4L4M3L4M2M4M3L4M2M4M3L4M3L3N4K4M4K5K4M4K5L3L5L4K4M4K5L3L3N1N101N2O0O2O1N2N101N2O1N1O2N2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N1O2N1O00010O001O00001O00001O00001O001O00001O00001O00001O001O00001O000010O0001O001OfLgMSIY2l6nMnHR2Q7UNjHj1U7\\NfHd1Y7cNaH]1^7jN]HU1b7ROXHn0g7YORHh0m7_OkGc0T8CdG`0[8G\\G in this image.", "objects": [{"patches": [156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 251, 252, 270, 271, 272, 274, 275, 294, 298], "bbox": [0.691125, 0.3707905138339921, 0.9984999999999999, 0.7033794466403162], "iscrowd": 0, "area": 19121.386199999994, "rle": {"size": [506, 640], "counts": "Xdj61g?6I7I6J7J6I7I7J6I7I6M4K5L3L4MIiBmMT=V2oBhMnB=C4L1O000000001O0000001O0000001O000000001O0000010O01O1O0010O01O00100O0012M5L3M4K5L0O01Ob9"}, "label": "the elephant behind the trunk"}]} {"id": 4477, "image": "COCO_train2014_000000004477.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant behind the front one\"."}], "task": "refering", "answer_template": "The \"the elephant behind the front one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 251, 252, 270, 271, 272, 274, 275, 294, 298], "bbox": [0.691125, 0.3707905138339921, 0.9984999999999999, 0.7033794466403162], "iscrowd": 0, "area": 19121.386199999994, "rle": {"size": [506, 640], "counts": "Xdj61g?6I7I6J7J6I7I7J6I7I6M4K5L3L4MIiBmMT=V2oBhMnB=C4L1O000000001O0000001O0000001O000000001O0000010O01O1O0010O01O00100O0012M5L3M4K5L0O01Ob9"}, "label": "the elephant behind the front one"}]} {"id": 332158, "image": "COCO_train2014_000000332158.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green couch with a person sitting on it\"."}], "task": "refering", "answer_template": "The \"a green couch with a person sitting on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 26, 27, 28, 49, 50, 51, 52, 72, 73, 74, 75, 95, 96, 97, 98], "bbox": [0.143171875, 0.0, 0.30209374999999994, 0.3343325526932084], "iscrowd": 0, "area": 8310.604699999998, "rle": {"size": [427, 640], "counts": "Y^V16d in this image.", "objects": [{"patches": [4, 5, 26, 27, 28, 49, 50, 51, 52, 72, 73, 74, 75, 95, 96, 97, 98], "bbox": [0.143171875, 0.0, 0.30209374999999994, 0.3343325526932084], "iscrowd": 0, "area": 8310.604699999998, "rle": {"size": [427, 640], "counts": "Y^V16d in this image.", "objects": [{"patches": [113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275], "bbox": [0.833234375, 0.204486301369863, 1.0, 0.5612328767123288], "iscrowd": 0, "area": 14751.599249999996, "rle": {"size": [584, 640], "counts": "kW`93Qb08I7J7I6J7I7J2M3M201N2O0YO[Nj@f1U?_Ng@a1Y?bNc@^1]?dNa@]1^?eNa@[1_?eN`@\\1b?cN]@]1f?a03L3N2N1ON2M2N3N2M4M2M3N2M4L3N2M4L3M3O101N100O2fNY1A?O101N001O1O100O1000O0100O10000O011O1N2O1N3N1O1N2O1O2M2O0O10000O2O000O100O101O0O10000O101N10000O101O0O10000O101N10000O101O^L`G\\Mh4"}, "label": "a yellow color truk parked near white car"}]} {"id": 438663, "image": "COCO_train2014_000000438663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the truck with no markings\"."}], "task": "refering", "answer_template": "The \"the truck with no markings\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 203, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275], "bbox": [0.833234375, 0.204486301369863, 1.0, 0.5612328767123288], "iscrowd": 0, "area": 14751.599249999996, "rle": {"size": [584, 640], "counts": "kW`93Qb08I7J7I6J7I7J2M3M201N2O0YO[Nj@f1U?_Ng@a1Y?bNc@^1]?dNa@]1^?eNa@[1_?eN`@\\1b?cN]@]1f?a03L3N2N1ON2M2N3N2M4M2M3N2M4L3N2M4L3M3O101N100O2fNY1A?O101N001O1O100O1000O0100O10000O011O1N2O1N3N1O1N2O1O2M2O0O10000O2O000O100O101O0O10000O101N10000O101O0O10000O101N10000O101O^L`G\\Mh4"}, "label": "the truck with no markings"}]} {"id": 438663, "image": "COCO_train2014_000000438663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white truck that has damages on the front\"."}], "task": "refering", "answer_template": "The \"a white truck that has damages on the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 375, 384], "bbox": [0.3000625, 0.11148972602739726, 0.8035781249999999, 0.7804109589041096], "iscrowd": 0, "area": 110073.32719999994, "rle": {"size": [584, 640], "counts": "ag]3l2d:VMnIW4\\4bLUKj4`3[KZLn4_3UK[LQ5b3RKYLR5e3QKPJQOd0T6Y5nJkIWOf0P6\\5lJgI\\Of0n5`5kJ`I@j0j5c5nJWI@P1g5g5[LRJj3k5ZLoIk3n5Z4N2N2N2N2N1O001O1O1O010O1O0000O1O100O1O100M3J6J6N2000000000000001O00000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000001O000000000001O000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000001O01O000000000000000000000000000000000000000000000000000000000000000001O0000000001O1O1O1O1O1O1O1O1O1O1O001O1O1O1O001O1O1O1O001O00000000O1O101N1O1O1O100N2N2M3N2M3N2N2D<0000O1000n2QM5L5K4K4M2M4L3N2M3N2M5K4M3L5K4L5J5K6J5KB>C>A>B>B>B?B=B>B>B7J5J7Im_W2"}, "label": "a white truck that has damages on the front"}]} {"id": 438663, "image": "COCO_train2014_000000438663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front of a staples truck that seems to have been in a small wreck\"."}], "task": "refering", "answer_template": "The \"the front of a staples truck that seems to have been in a small wreck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 375, 384], "bbox": [0.3000625, 0.11148972602739726, 0.8035781249999999, 0.7804109589041096], "iscrowd": 0, "area": 110073.32719999994, "rle": {"size": [584, 640], "counts": "ag]3l2d:VMnIW4\\4bLUKj4`3[KZLn4_3UK[LQ5b3RKYLR5e3QKPJQOd0T6Y5nJkIWOf0P6\\5lJgI\\Of0n5`5kJ`I@j0j5c5nJWI@P1g5g5[LRJj3k5ZLoIk3n5Z4N2N2N2N2N1O001O1O1O010O1O0000O1O100O1O100M3J6J6N2000000000000001O00000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000001O000000000001O000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000001O01O000000000000000000000000000000000000000000000000000000000000000001O0000000001O1O1O1O1O1O1O1O1O1O1O001O1O1O1O001O1O1O1O001O00000000O1O101N1O1O1O100N2N2M3N2M3N2N2D<0000O1000n2QM5L5K4K4M2M4L3N2M3N2M5K4M3L5K4L5J5K6J5KB>C>A>B>B>B?B=B>B>B7J5J7Im_W2"}, "label": "the front of a staples truck that seems to have been in a small wreck"}]} {"id": 438663, "image": "COCO_train2014_000000438663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a food truck parked next to a smashed truck\"."}], "task": "refering", "answer_template": "The \"a food truck parked next to a smashed truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 276, 277, 278, 279, 280, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326, 345, 346, 347, 348, 368, 369, 370, 391, 392, 393], "bbox": [0.0, 0.0024486301369863015, 0.30825, 0.8547945205479451], "iscrowd": 0, "area": 70653.88525, "rle": {"size": [584, 640], "counts": "4a9g8001O0O2O001O001O001O0O2O0lJfFZ1[9iMeGS2[8RMeHi2\\7oLoHm2Q7mLYIo2h6jLcIQ3]6jLlIQ3V6hLTJT3l5gL]JU3d5dLgJW3Y5dLPKX3T:L4L3MO1O2N1O1O1O1O12N4L4L4L1OO1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1M3L4M3M3L4M3L4M3L4O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1_LmJmIS5S6QKhIP5W6UKdIl4\\6YK^Ih4b6\\KYIf4e6_KVIb4j6bKRI^4n6gKlHZ4S7kKhHV4X7nKcHS4\\7RL]HQ4c7SLRHT4n7QLfGV4Y8oK\\GX4d8lKQG[4o8iKfF^4Y9Y20000O10000O10000O10000O101O2M2O2N2M3N2N1N3N2N2M2O2M3N2N2M2O2N2M3N1O2M3N2N2M2O2N2M3N1O2M3N2N2M2O2N2M3N1O2M3N2N2M2O2N2M3N1N3N2N2M2O2N2M3N2N1N3N2Nj1UNR2oMb]l7"}, "label": "a food truck parked next to a smashed truck"}]} {"id": 438663, "image": "COCO_train2014_000000438663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"truck on left hand side\"."}], "task": "refering", "answer_template": "The \"truck on left hand side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 23, 24, 25, 26, 27, 28, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 276, 277, 278, 279, 280, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326, 345, 346, 347, 348, 368, 369, 370, 391, 392, 393], "bbox": [0.0, 0.0024486301369863015, 0.30825, 0.8547945205479451], "iscrowd": 0, "area": 70653.88525, "rle": {"size": [584, 640], "counts": "4a9g8001O0O2O001O001O001O0O2O0lJfFZ1[9iMeGS2[8RMeHi2\\7oLoHm2Q7mLYIo2h6jLcIQ3]6jLlIQ3V6hLTJT3l5gL]JU3d5dLgJW3Y5dLPKX3T:L4L3MO1O2N1O1O1O1O12N4L4L4L1OO1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1M3L4M3M3L4M3L4M3L4O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1_LmJmIS5S6QKhIP5W6UKdIl4\\6YK^Ih4b6\\KYIf4e6_KVIb4j6bKRI^4n6gKlHZ4S7kKhHV4X7nKcHS4\\7RL]HQ4c7SLRHT4n7QLfGV4Y8oK\\GX4d8lKQG[4o8iKfF^4Y9Y20000O10000O10000O10000O101O2M2O2N2M3N2N1N3N2N2M2O2M3N2N2M2O2N2M3N1O2M3N2N2M2O2N2M3N1O2M3N2N2M2O2N2M3N1O2M3N2N2M2O2N2M3N1N3N2N2M2O2N2M3N2N1N3N2Nj1UNR2oMb]l7"}, "label": "truck on left hand side"}]} {"id": 258440, "image": "COCO_train2014_000000258440.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sink in a barber shop in front of the chair in the picture\"."}], "task": "refering", "answer_template": "The \"a sink in a barber shop in front of the chair in the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [185, 186, 200, 201, 202, 203, 204, 205, 206, 207, 216, 217, 218, 219, 220, 221, 233, 234, 235], "bbox": [0.3622041763341068, 0.532578125, 0.8489327146171693, 0.6890468750000001], "iscrowd": 0, "area": 7863.433149999998, "rle": {"size": [640, 431], "counts": "XkQ35`c0;O2O1O1O1O1N101O1O1O0O2O1O001O0OWOP]Oe0ob0[OQ]Of0nb0YOS]Og0mb0YOT]Og0kb0XOV]Oh0jb0XOV]Oi0hb0XOY]Og0gb0XOZ]Oi0eb0WO[]Oj0db0VO]]Oi0cb0VO^]Ok0`b0VO`]Ok0_b0TOc]Om0[b0SOe]On0Zb0ROf]Oo0Yb0POh]OQ1Vb0POk]OP1Tb0POl]OQ1Sb0nNn]OS1Qb0mNP^OS1^b01_OlNT^OU1ka0kNV^OU1Zb02N100O1O100O00010O001O010O00010O0010O00010O01O01O01O010O001O01O01O010O00010O00010O001O00001O00001O015J1O:F00001O00000010O000001O0000K5M300O2M2L40000000000000001O000O100000000000001O000000000000000000001O00000000000O10000000000000O10000000O10000O1Fb]ORO^b0l0 in this image.", "objects": [{"patches": [227, 228, 229, 231, 232, 243, 244, 245, 246, 248, 260, 261, 262, 263, 264, 265, 277, 278, 279, 280, 281, 282, 295, 296, 297, 298, 299, 312, 313, 314], "bbox": [0.324602510460251, 0.578546875, 0.6643514644351465, 0.813203125], "iscrowd": 0, "area": 11909.965850000008, "rle": {"size": [640, 478], "counts": "SYQ3=bc04K4M4L4L3M4N1O2M3N1N3N2N101O001O1O001N2L3N3M3M2N3M2N3M3M2N3L2O1_N]McAc2]>`M`Aa2_>bM^A^2b>dM\\A]2c>fMZAZ2g>hMVAY2i>jMTAW2k>kMSAU2m>nMo@T2P?oMm@Q2S?QNk@P2T?SNi@m1W?UNg@l1X?SNi@m1W?SNi@m1X?QNi@o1W?QNi@o1W?PNj@P2V?PNi@Q2W?nMj@R2V?nMi@S2W?lMi@U2W?kMh@V2X?iMh@X2X?hMg@Y2Y?fMg@[2Z?dMe@]2[?bMe@_2[?aMd@`2\\?_Md@b2\\?^Mc@c2]?\\Mc@e2]?[Mb@f2^?YMb@h2^?XMa@i2_?VMa@k2`?TM_@l2b?SM^@n2b?RM^@n2b?QM_@o2a?QM`@n2`?QMa@o2_?QMb@n2^?QMc@o2]?oLf@P3Z?oLg@Q3Y?nLh@R3Y?lLl@P3T?oLPAn2P?QMUAk2k>SM[Ai2e>VM`Af2`>YMeAc2[>\\MgAc2k?01O00010N1N2O1N3N1N2O2M2O1N3N1N2O2M2O1N2O1N2O1N2O1N2O101O000000000000O10001O00XOg]OIYb03R^OFma07Y^OGea0:^^ODk`0R1W_OmNX`0c1j_O\\NS`0f1R11N2N2N101N2N2O1001O0N3M3M3M3M3M3L4L4M2M4L4M3L4L4M3L4L3N3LbXT3"}, "label": "the bottom of the broken white vase"}]} {"id": 94600, "image": "COCO_train2014_000000094600.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a broken white pot that has been made into an art display and is now hanging by strings\"."}], "task": "refering", "answer_template": "The \"a broken white pot that has been made into an art display and is now hanging by strings\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [227, 228, 229, 231, 232, 243, 244, 245, 246, 248, 260, 261, 262, 263, 264, 265, 277, 278, 279, 280, 281, 282, 295, 296, 297, 298, 299, 312, 313, 314], "bbox": [0.324602510460251, 0.578546875, 0.6643514644351465, 0.813203125], "iscrowd": 0, "area": 11909.965850000008, "rle": {"size": [640, 478], "counts": "SYQ3=bc04K4M4L4L3M4N1O2M3N1N3N2N101O001O1O001N2L3N3M3M2N3M2N3M3M2N3L2O1_N]McAc2]>`M`Aa2_>bM^A^2b>dM\\A]2c>fMZAZ2g>hMVAY2i>jMTAW2k>kMSAU2m>nMo@T2P?oMm@Q2S?QNk@P2T?SNi@m1W?UNg@l1X?SNi@m1W?SNi@m1X?QNi@o1W?QNi@o1W?PNj@P2V?PNi@Q2W?nMj@R2V?nMi@S2W?lMi@U2W?kMh@V2X?iMh@X2X?hMg@Y2Y?fMg@[2Z?dMe@]2[?bMe@_2[?aMd@`2\\?_Md@b2\\?^Mc@c2]?\\Mc@e2]?[Mb@f2^?YMb@h2^?XMa@i2_?VMa@k2`?TM_@l2b?SM^@n2b?RM^@n2b?QM_@o2a?QM`@n2`?QMa@o2_?QMb@n2^?QMc@o2]?oLf@P3Z?oLg@Q3Y?nLh@R3Y?lLl@P3T?oLPAn2P?QMUAk2k>SM[Ai2e>VM`Af2`>YMeAc2[>\\MgAc2k?01O00010N1N2O1N3N1N2O2M2O1N3N1N2O2M2O1N2O1N2O1N2O1N2O101O000000000000O10001O00XOg]OIYb03R^OFma07Y^OGea0:^^ODk`0R1W_OmNX`0c1j_O\\NS`0f1R11N2N2N101N2N2O1001O0N3M3M3M3M3M3L4L4M2M4L4M3L4L4M3L4L3N3LbXT3"}, "label": "a broken white pot that has been made into an art display and is now hanging by strings"}]} {"id": 520590, "image": "COCO_train2014_000000520590.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a girl wearing a pink and white sweater\"."}], "task": "refering", "answer_template": "The \"the back of a girl wearing a pink and white sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 15, 16, 17, 18, 19, 20, 30, 31, 32, 33, 34, 45, 46, 47, 48, 60, 61, 62, 63, 75, 76, 77, 78, 90, 91, 92, 93, 105, 106, 107, 108, 120, 121, 122, 123, 135, 136, 137, 150, 151, 152, 165, 166, 180, 181, 195, 196, 210, 211, 225, 226, 240, 241, 255, 256, 257, 270, 271, 272, 285, 286, 287, 300, 301, 302, 315, 316, 317, 330, 331, 332], "bbox": [0.010070257611241218, 0.024921875, 0.36295081967213116, 0.9868125000000001], "iscrowd": 0, "area": 44157.02354999999, "rle": {"size": [640, 427], "counts": "ad2o>n44M2M3N2M3N2M3M3M3N3M2N2O1N2N2N2N2N2N3ROf_ORN\\`0Z1i@SNY?f1d1N2N2M3N2N2N3M2M3N2N2N2N2M3O2O000001O0000000001O0001O00000gE in this image.", "objects": [{"patches": [1, 2, 3, 15, 16, 17, 18, 19, 20, 30, 31, 32, 33, 34, 45, 46, 47, 48, 60, 61, 62, 63, 75, 76, 77, 78, 90, 91, 92, 93, 105, 106, 107, 108, 120, 121, 122, 123, 135, 136, 137, 150, 151, 152, 165, 166, 180, 181, 195, 196, 210, 211, 225, 226, 240, 241, 255, 256, 257, 270, 271, 272, 285, 286, 287, 300, 301, 302, 315, 316, 317, 330, 331, 332], "bbox": [0.010070257611241218, 0.024921875, 0.36295081967213116, 0.9868125000000001], "iscrowd": 0, "area": 44157.02354999999, "rle": {"size": [640, 427], "counts": "ad2o>n44M2M3N2M3N2M3M3M3N3M2N2O1N2N2N2N2N2N3ROf_ORN\\`0Z1i@SNY?f1d1N2N2M3N2N2N3M2M3N2N2N2N2M3O2O000001O0000000001O0001O00000gE in this image.", "objects": [{"patches": [34, 35, 36, 37, 48, 49, 50, 51, 52, 53, 63, 64, 65, 66, 67, 68, 78, 79, 80, 81, 82, 83, 86, 87, 93, 94, 95, 96, 97, 98, 102, 103, 108, 109, 110, 111, 112, 113, 117, 118, 123, 124, 125, 126, 127, 130, 131, 132, 133, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 272, 273, 274, 275, 276, 277, 278, 279, 280, 287, 288, 289, 290, 291, 292, 293, 294, 295, 303, 304, 305, 306, 307, 308, 309, 310, 318, 319, 320, 321, 322, 323, 324, 325, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.10440281030444964, 0.0985, 0.916135831381733, 0.986140625], "iscrowd": 0, "area": 130822.36760000001, "rle": {"size": [640, 427], "counts": "[]l0b0]c0c0]Od0\\Oc0]Oc0]Od0\\Oc0]Od0\\Oc0]Oc0]O6J3M4L4L4L3M4L4L3L5L4L3M4L4L3M4L4L4L3M4L4L3M4L4L3M4L4L3M4L4L4^G^F_6e9XIbFg6b9PIdFo6`9gHgFX7\\9`HjF_7Z9XHlFX7g9^H`FR7_;VNWCmKY=l3PCSLoA?@`0A?@a0@?AaUf0"}, "label": "a person in a blue and white outfit"}]} {"id": 520590, "image": "COCO_train2014_000000520590.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy looking at his cellphone\"."}], "task": "refering", "answer_template": "The \"a boy looking at his cellphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 35, 36, 37, 48, 49, 50, 51, 52, 53, 63, 64, 65, 66, 67, 68, 78, 79, 80, 81, 82, 83, 86, 87, 93, 94, 95, 96, 97, 98, 102, 103, 108, 109, 110, 111, 112, 113, 117, 118, 123, 124, 125, 126, 127, 130, 131, 132, 133, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 272, 273, 274, 275, 276, 277, 278, 279, 280, 287, 288, 289, 290, 291, 292, 293, 294, 295, 303, 304, 305, 306, 307, 308, 309, 310, 318, 319, 320, 321, 322, 323, 324, 325, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.10440281030444964, 0.0985, 0.916135831381733, 0.986140625], "iscrowd": 0, "area": 130822.36760000001, "rle": {"size": [640, 427], "counts": "[]l0b0]c0c0]Od0\\Oc0]Oc0]Od0\\Oc0]Od0\\Oc0]Oc0]O6J3M4L4L4L3M4L4L3L5L4L3M4L4L3M4L4L4L3M4L4L3M4L4L3M4L4L3M4L4L4^G^F_6e9XIbFg6b9PIdFo6`9gHgFX7\\9`HjF_7Z9XHlFX7g9^H`FR7_;VNWCmKY=l3PCSLoA?@`0A?@a0@?AaUf0"}, "label": "a boy looking at his cellphone"}]} {"id": 520590, "image": "COCO_train2014_000000520590.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"floral fabric\"."}], "task": "refering", "answer_template": "The \"floral fabric\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 87, 88, 89, 103, 104, 118, 119, 133, 134, 148, 149, 163, 164, 178, 179, 193, 194, 208, 209, 223, 224, 237, 238, 239, 252, 253, 254, 268], "bbox": [0.8408899297423887, 0.166234375, 0.9966276346604216, 0.787015625], "iscrowd": 0, "area": 17067.333149999995, "rle": {"size": [640, 427], "counts": "n`P72lc03L4M2M4L4lECa0`0SONg05nN8n0KfNc0W1^O]No0`1SOTNZ1j1gNjMg1R2ZNcMS2Z2oM\\M]2`2eM^M^2_2dM`M_2\\2cMcM_2Z2QNWMQ2g2`NhLc1T3oN[LS1b3_OmKb0Q40^K0a4b0nJ^OQ5T1^JlNa5f1mIZNT6W2[IiMd6j2jHVMU7\\3ZHdLe7k701O1O1O010O1O1O001OSOm0010O1O1O001O1O04M4L3M3L4M4L2M3N1O2N1N3N1O2M2O2VNi1PNQ2oMP2PNQ2oM[a0"}, "label": "floral fabric"}]} {"id": 520590, "image": "COCO_train2014_000000520590.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a floral suit\"."}], "task": "refering", "answer_template": "The \"a person in a floral suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 87, 88, 89, 103, 104, 118, 119, 133, 134, 148, 149, 163, 164, 178, 179, 193, 194, 208, 209, 223, 224, 237, 238, 239, 252, 253, 254, 268], "bbox": [0.8408899297423887, 0.166234375, 0.9966276346604216, 0.787015625], "iscrowd": 0, "area": 17067.333149999995, "rle": {"size": [640, 427], "counts": "n`P72lc03L4M2M4L4lECa0`0SONg05nN8n0KfNc0W1^O]No0`1SOTNZ1j1gNjMg1R2ZNcMS2Z2oM\\M]2`2eM^M^2_2dM`M_2\\2cMcM_2Z2QNWMQ2g2`NhLc1T3oN[LS1b3_OmKb0Q40^K0a4b0nJ^OQ5T1^JlNa5f1mIZNT6W2[IiMd6j2jHVMU7\\3ZHdLe7k701O1O1O010O1O1O001OSOm0010O1O1O001O1O04M4L3M3L4M4L2M3N1O2N1N3N1O2M2O2VNi1PNQ2oMP2PNQ2oM[a0"}, "label": "a person in a floral suit"}]} {"id": 561551, "image": "COCO_train2014_000000561551.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white boat with yellow and black bottom\"."}], "task": "refering", "answer_template": "The \"a white boat with yellow and black bottom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 164, 165, 167, 168, 179, 180, 181, 187, 188, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270, 290, 291, 292], "bbox": [0.15003125, 0.42061784897025173, 0.9139843750000001, 0.761441647597254], "iscrowd": 0, "area": 30947.352549999996, "rle": {"size": [437, 640], "counts": "fUY12`=4K4L5L4K4L5K4M4N110O010O010O010O010O010O010O010O010O010O010O010O010O0100O2M2J7I6I7Icb8Ob]G2O2N100O2N1O2O0L5K4L5K4L4M4K4L5K410O010O01O01O010O001O010O001O01O01O0010O01O0010O0001O010O001O010O00010O001O010O001O01O01O0010O0001O01O0001O01O0001O01O01O0001O01O0001O01O0001O01O00010O0000010O0000010O000010O0000010O0000010O0000010O0001O01O0001O01O0001O01O0001O00001N1000001O0000001O0000001O00001N1000001O0000001O0000001O00001O0O10001O0000001O00001O0000001O000O101O0000001O00001O0000001O000O101O0000001O00001O0000001O00000O2O0000001O00001O0000001O0000001N10001O0000001O0000001O0000001O0O101O0000001O0000001O0000001O0O101O0000000000O1000000000000000O10000000000000O10000000O10000000000000O10O01O100O1O100O10O01O100O100O1O100O00100O100O1O100O10O01O100O1O100O10O01O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1N2O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1N2O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O0O2O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1N2O1O1O001O1O1O1O1O1O1O1O00dXg0"}, "label": "a white boat with yellow and black bottom"}]} {"id": 561551, "image": "COCO_train2014_000000561551.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a long boat with a black base and white top\"."}], "task": "refering", "answer_template": "The \"a long boat with a black base and white top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 164, 165, 167, 168, 179, 180, 181, 187, 188, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 264, 265, 266, 267, 268, 269, 270, 290, 291, 292], "bbox": [0.15003125, 0.42061784897025173, 0.9139843750000001, 0.761441647597254], "iscrowd": 0, "area": 30947.352549999996, "rle": {"size": [437, 640], "counts": "fUY12`=4K4L5L4K4L5K4M4N110O010O010O010O010O010O010O010O010O010O010O010O010O0100O2M2J7I6I7Icb8Ob]G2O2N100O2N1O2O0L5K4L5K4L4M4K4L5K410O010O01O01O010O001O010O001O01O01O0010O01O0010O0001O010O001O010O00010O001O010O001O01O01O0010O0001O01O0001O01O0001O01O01O0001O01O0001O01O0001O01O00010O0000010O0000010O000010O0000010O0000010O0000010O0001O01O0001O01O0001O01O0001O00001N1000001O0000001O0000001O00001N1000001O0000001O0000001O00001O0O10001O0000001O00001O0000001O000O101O0000001O00001O0000001O000O101O0000001O00001O0000001O00000O2O0000001O00001O0000001O0000001N10001O0000001O0000001O0000001O0O101O0000001O0000001O0000001O0O101O0000000000O1000000000000000O10000000000000O10000000O10000000000000O10O01O100O1O100O10O01O100O100O1O100O00100O100O1O100O10O01O100O1O100O10O01O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1N2O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1N2O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O0O2O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1N2O1O1O001O1O1O1O1O1O1O1O00dXg0"}, "label": "a long boat with a black base and white top"}]} {"id": 94608, "image": "COCO_train2014_000000094608.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an airplane taking off\"."}], "task": "refering", "answer_template": "The \"an airplane taking off\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 149, 150, 151, 152, 155, 156, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204], "bbox": [0.3974375, 0.45138235294117646, 0.9735781250000001, 0.7582058823529411], "iscrowd": 0, "area": 12879.298850000005, "rle": {"size": [340, 640], "counts": "Rad21b:2N2N1O2N1006KO01O01O00001O000fE1o90mE6Q:9M1O000000001O01O00000001O000001O0001O00N2N2N2N3O01O00001O001O100O1O001O00001O01O01O00001O000010OE@bF`0[9=O1N3N1O1O10001O0000001O201N1O1O1O1O1O1O10J5N21O1O2O0O1O1O2O0O00O1O1O2N11O00100]G^NR8b1nG^NR8c1lG^NT8b1lG^NU8b1iG_NW8b1hG^NX8b1gG_NY8b1eG_N[8a1_GfNa8f10000001@eGeN[8[1eGeN[8\\1dGdN\\8\\1eGaN^8_1aG`N`8`1`G_Na8a1910O0001O000YG^N]8b1bG^N^8b1bG^N^8c1`G^N`8b1`G_N`8`1`G`N`8a1^G`Nb8a1]G_Nc8a1]G_Nd8a1ZG`Nf8g10000001F\\GbNd8^1\\GbNd8_1\\G`Nd8`1\\G`Ne8_1\\G`Nd8`1801O01O0VG^Nc8a1]G`Nb8`1^G`Nb8`1]GaNc8_1]GaNd8_1ZGbNf8^1ZGbNf8f110O0001O00001O01O01O0O101N100O101N100O2O00001O00001N10001O0000001O00001N10001O00001O00001N1000O1N2O1O1N2O1O1N101O1N2O1O1N2O1O1N2O001N2O1O1N2O1O1N2O001N2O1O1N2O1OUOlGXOR8h0PHXOo7f0THZOk7d0XH[Og7e0[H[Od7c0_H]O`7b0bH^O\\7a0gH_OX7`0jH@U7>nHBP7>RIBm6 in this image.", "objects": [{"patches": [133, 149, 150, 151, 152, 155, 156, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204], "bbox": [0.3974375, 0.45138235294117646, 0.9735781250000001, 0.7582058823529411], "iscrowd": 0, "area": 12879.298850000005, "rle": {"size": [340, 640], "counts": "Rad21b:2N2N1O2N1006KO01O01O00001O000fE1o90mE6Q:9M1O000000001O01O00000001O000001O0001O00N2N2N2N3O01O00001O001O100O1O001O00001O01O01O00001O000010OE@bF`0[9=O1N3N1O1O10001O0000001O201N1O1O1O1O1O1O10J5N21O1O2O0O1O1O2O0O00O1O1O2N11O00100]G^NR8b1nG^NR8c1lG^NT8b1lG^NU8b1iG_NW8b1hG^NX8b1gG_NY8b1eG_N[8a1_GfNa8f10000001@eGeN[8[1eGeN[8\\1dGdN\\8\\1eGaN^8_1aG`N`8`1`G_Na8a1910O0001O000YG^N]8b1bG^N^8b1bG^N^8c1`G^N`8b1`G_N`8`1`G`N`8a1^G`Nb8a1]G_Nc8a1]G_Nd8a1ZG`Nf8g10000001F\\GbNd8^1\\GbNd8_1\\G`Nd8`1\\G`Ne8_1\\G`Nd8`1801O01O0VG^Nc8a1]G`Nb8`1^G`Nb8`1]GaNc8_1]GaNd8_1ZGbNf8^1ZGbNf8f110O0001O00001O01O01O0O101N100O101N100O2O00001O00001N10001O0000001O00001N10001O00001O00001N1000O1N2O1O1N2O1O1N101O1N2O1O1N2O1O1N2O001N2O1O1N2O1O1N2O001N2O1O1N2O1OUOlGXOR8h0PHXOo7f0THZOk7d0XH[Og7e0[H[Od7c0_H]O`7b0bH^O\\7a0gH_OX7`0jH@U7>nHBP7>RIBm6 in this image.", "objects": [{"patches": [123, 124, 146, 147, 148, 169, 170, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199], "bbox": [0.34934375, 0.34392070484581494, 0.6829218750000001, 0.5405506607929514], "iscrowd": 0, "area": 6231.6801, "rle": {"size": [454, 640], "counts": "n_S3:l=gC@E2d`CANOb<`0`CA0Kb in this image.", "objects": [{"patches": [123, 124, 146, 147, 148, 169, 170, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199], "bbox": [0.34934375, 0.34392070484581494, 0.6829218750000001, 0.5405506607929514], "iscrowd": 0, "area": 6231.6801, "rle": {"size": [454, 640], "counts": "n_S3:l=gC@E2d`CANOb<`0`CA0Kb in this image.", "objects": [{"patches": [28, 46, 47, 65, 66, 67, 83, 84, 85, 86, 87, 102, 103, 105, 106, 120, 121, 122, 138, 139, 140, 156, 158, 174, 176, 177, 192, 195], "bbox": [0.5644600000000001, 0.09219219219219218, 0.89896, 0.8769369369369369], "iscrowd": 0, "area": 10339.421799999998, "rle": {"size": [333, 500], "counts": "hgk21[:4L3M4M3L2N2N1O2O03N1N2O1N10N1O3M3M3M3N2M3M2OYOnF5f90O100O1O100O1O100O100O1O10O01O100O00100O00100O1N101O3L9H9ZKeNGc1O^N1k1EUN=T2mLdN5ZOo2Z2ULQOd0fNW3d4_L^Kb3h4VLXKl3o4jKSKV4b51O1N101NM4L3M4OF:\\Oe0nN[KULk4i3m0O1N3M2N2O2M4L3N2M4O01O1O010O1O12N1N5L5K3MO100O1O2N101N`MPId1o6\\NSId1k6\\NXIb1h6]NZIc1d6]N_IFNQ1b6YOcIYO7^1U6XO_Jg0`5ZObJe0\\5[OfJd0Z5[OhJe0W5ZOkJe0U5ZOmJf0S5XOoJg0Q5XOQKh0n4XOSKg0m4XOVKg0i4XOZKf0g4YO[Kf0d4YO_Kf0`4YOcKg0[4YOgKh0V4WOmKi0R4UOQLk0m3UOULk0h3UO[Lk0c3UO_Ll0]3TOgLk0V3UOmLk0P3VORMj0l2UOXMh0g2XO\\Mf0c2[O_Md0`2[OcMc0\\2]OfMc0X2^OiMa0W2^OkMb0S2^OoMa0P2@QN`0n1_OTN`0k1AVN?h1AZN in this image.", "objects": [{"patches": [28, 46, 47, 65, 66, 67, 83, 84, 85, 86, 87, 102, 103, 105, 106, 120, 121, 122, 138, 139, 140, 156, 158, 174, 176, 177, 192, 195], "bbox": [0.5644600000000001, 0.09219219219219218, 0.89896, 0.8769369369369369], "iscrowd": 0, "area": 10339.421799999998, "rle": {"size": [333, 500], "counts": "hgk21[:4L3M4M3L2N2N1O2O03N1N2O1N10N1O3M3M3M3N2M3M2OYOnF5f90O100O1O100O1O100O100O1O10O01O100O00100O00100O1N101O3L9H9ZKeNGc1O^N1k1EUN=T2mLdN5ZOo2Z2ULQOd0fNW3d4_L^Kb3h4VLXKl3o4jKSKV4b51O1N101NM4L3M4OF:\\Oe0nN[KULk4i3m0O1N3M2N2O2M4L3N2M4O01O1O010O1O12N1N5L5K3MO100O1O2N101N`MPId1o6\\NSId1k6\\NXIb1h6]NZIc1d6]N_IFNQ1b6YOcIYO7^1U6XO_Jg0`5ZObJe0\\5[OfJd0Z5[OhJe0W5ZOkJe0U5ZOmJf0S5XOoJg0Q5XOQKh0n4XOSKg0m4XOVKg0i4XOZKf0g4YO[Kf0d4YO_Kf0`4YOcKg0[4YOgKh0V4WOmKi0R4UOQLk0m3UOULk0h3UO[Lk0c3UO_Ll0]3TOgLk0V3UOmLk0P3VORMj0l2UOXMh0g2XO\\Mf0c2[O_Md0`2[OcMc0\\2]OfMc0X2^OiMa0W2^OkMb0S2^OoMa0P2@QN`0n1_OTN`0k1AVN?h1AZN in this image.", "objects": [{"patches": [44, 45, 63, 64, 81, 82, 99, 100, 101, 117, 118, 134, 135, 136, 152, 153, 154, 170, 172, 190, 208], "bbox": [0.44, 0.1773273273273273, 0.62986, 0.9467267267267268], "iscrowd": 0, "area": 7822.224049999999, "rle": {"size": [333, 500], "counts": "fiW23Y:2N1O2N101N101N100O100O100O100001N1000O0A`0K4cJoN]1T1bNUOR1n0nNUOl0m0UOVOe0m0^OSO5ASL]1j3SONFQLU12fNQ4j0]Oe0o0FbN;b1GWN9m1JkM8X2JVKeNR2g1l2HnL:V3GfL8]3J^L6e3\\NTKl0S1g0l3[NXKl0g0h0T4[N[Kj0 in this image.", "objects": [{"patches": [44, 45, 63, 64, 81, 82, 99, 100, 101, 117, 118, 134, 135, 136, 152, 153, 154, 170, 172, 190, 208], "bbox": [0.44, 0.1773273273273273, 0.62986, 0.9467267267267268], "iscrowd": 0, "area": 7822.224049999999, "rle": {"size": [333, 500], "counts": "fiW23Y:2N1O2N101N101N100O100O100O100001N1000O0A`0K4cJoN]1T1bNUOR1n0nNUOl0m0UOVOe0m0^OSO5ASL]1j3SONFQLU12fNQ4j0]Oe0o0FbN;b1GWN9m1JkM8X2JVKeNR2g1l2HnL:V3GfL8]3J^L6e3\\NTKl0S1g0l3[NXKl0g0h0T4[N[Kj0 in this image.", "objects": [{"patches": [93, 94, 111, 112, 129, 130, 147, 148, 149, 165, 166, 167, 183], "bbox": [0.162, 0.42456456456456454, 0.294, 0.8524924924924925], "iscrowd": 0, "area": 4843.72875, "rle": {"size": [333, 500], "counts": "Raj06R:n0o5WOfIH6W1T6QOhIEO`1X6lNjIBHh1^6gNkI@Bn1S6RNSJ`08_OBP2R6VNmI?>ZOBR2S6TO`Jl0`5VO]Ji0e5YOYJf0h5\\OkIm0W6b12N3M3N2M2\\O[I_Mh6a2aISMb6m2>11N2O001N20O0100^MaHU2k70100O100O02O000001O000000002O4K5K3M2314f0QO1M3M2N1O1O1O1O1O<@]ab3"}, "label": "a blond male in a white t - shirt and gray pants running during a frisbee game"}]} {"id": 111000, "image": "COCO_train2014_000000111000.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy with a white shirt with teal sleeves\"."}], "task": "refering", "answer_template": "The \"a boy with a white shirt with teal sleeves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 111, 112, 129, 130, 147, 148, 149, 165, 166, 167, 183], "bbox": [0.162, 0.42456456456456454, 0.294, 0.8524924924924925], "iscrowd": 0, "area": 4843.72875, "rle": {"size": [333, 500], "counts": "Raj06R:n0o5WOfIH6W1T6QOhIEO`1X6lNjIBHh1^6gNkI@Bn1S6RNSJ`08_OBP2R6VNmI?>ZOBR2S6TO`Jl0`5VO]Ji0e5YOYJf0h5\\OkIm0W6b12N3M3N2M2\\O[I_Mh6a2aISMb6m2>11N2O001N20O0100^MaHU2k70100O100O02O000001O000000002O4K5K3M2314f0QO1M3M2N1O1O1O1O1O<@]ab3"}, "label": "a boy with a white shirt with teal sleeves"}]} {"id": 111000, "image": "COCO_train2014_000000111000.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with both arms high above his head\"."}], "task": "refering", "answer_template": "The \"man with both arms high above his head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 60, 61, 62, 78, 79, 80, 96, 97, 98, 114, 115, 116, 133, 134, 151, 152, 169, 170, 187, 188], "bbox": [0.36668, 0.17078078078078077, 0.49388, 0.9168468468468468], "iscrowd": 0, "area": 8598.925599999999, "rle": {"size": [333, 500], "counts": "Pnk16V:5K6J6J5K6UHoNo5W1aIXOZ6m0XIBb6d0oHJl6P2K5K5K5K5K6J5K5K5K5K5\\Od0L4L4L4L4L4L3M4L4L4M3LQN[M[L`2@cL^3T1PMT2ChL\\3X1SMj1BoLZ3Z1QMf1EPMZ3]1nLc1IPMY3`1jL`1NoLX3e1dL^1HZMd3[1_L\\1J\\Mg3[1ZL[1L_Mh3f0ULb20TMc3:cL[2FiM]4lNTLg4i3ZK]L^4d3cKZLb2AdNV4kNWLb23SNg3[OTLd28lMe3APLf2?cMb3IlKf2f0\\M^3OjKg2l0TM[36fKi2R1mLY3:cKk2X1eLa0b01OUNn2d2mMPOVO[NP3_2TNoNmN`NT3[2XNnNc2l0gMnN\\2k0k3F in this image.", "objects": [{"patches": [6, 7, 29, 30, 31, 32, 33, 51, 52, 53, 54, 55, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 256, 257, 258, 259, 260, 261, 266, 267, 268, 269, 270, 271, 272, 291, 292, 293, 294, 295, 315, 316, 317, 318, 338, 339, 340, 341, 361, 362, 363, 384, 385, 386], "bbox": [0.0, 0.025854166666666668, 0.858296875, 0.9719791666666667], "iscrowd": 0, "area": 87385.65370000004, "rle": {"size": [480, 640], "counts": "Q82k>3M4L3M3M4K4M4L3N3N101O00001O1O001O1N2O1OO100O1O1O100O00100O1O1O100O1O010O1O100000000000O2O000000001O00000O10001O00000000001O00000000001O00000000001N10000O101N1O1O1O1QDcNZ:^1dEdN[:^1cEcN\\:^1cEcN\\:^1bEdN]:]1bEdN]:]1bEdN]:]1bEdN^:]1_EeN`:\\1_EeN`:\\1_EeN`:\\1^EfNa:[1^EfNa:[1^EfNa:\\1]EeNb:\\1^EdNa:]1_EcN`:^1`EcN^:^1bEbN]:`1bE`N]:a1cE_N]:a1cE_N\\:b1dE^NY:e1gE[NU:i1kEWNR:m1mESNP:P2PFPNm9S2SFmMj9V2VFjMg9Y2YFgMd9\\2\\FdMb9_2]FaM`9b2`F^M^9d2bF\\M\\9f2dFZMY9i2gFWMW9k2iFUMU9n2jFRMS9Q3mFoLQ9S3oFmLo8U3QGkLl8X3TGhLi8\\3VGdLd8b3\\G^L`8f3`GZL]8i3cGWL[8k3eGULY8n3fGRLX8P4hGPLV8R4jGnKT8T4lGlKQ8X4nGhKP8Z4PHfKn7\\4RHdKl7_4SHaKk7a4UH_Ki7c4WH]Kf7g4YHYKe7i4[HWKc7k4]HUKa7m4_HSK_7P5`HPK^7R5bHnJ[7U5fHjJX7Y5gHgJW7CVHa5c0lJU7AZHe5a0jJS7_O^Hi5?hJP7^OcHm5 in this image.", "objects": [{"patches": [6, 7, 29, 30, 31, 32, 33, 51, 52, 53, 54, 55, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 256, 257, 258, 259, 260, 261, 266, 267, 268, 269, 270, 271, 272, 291, 292, 293, 294, 295, 315, 316, 317, 318, 338, 339, 340, 341, 361, 362, 363, 384, 385, 386], "bbox": [0.0, 0.025854166666666668, 0.858296875, 0.9719791666666667], "iscrowd": 0, "area": 87385.65370000004, "rle": {"size": [480, 640], "counts": "Q82k>3M4L3M3M4K4M4L3N3N101O00001O1O001O1N2O1OO100O1O1O100O00100O1O1O100O1O010O1O100000000000O2O000000001O00000O10001O00000000001O00000000001O00000000001N10000O101N1O1O1O1QDcNZ:^1dEdN[:^1cEcN\\:^1cEcN\\:^1bEdN]:]1bEdN]:]1bEdN]:]1bEdN^:]1_EeN`:\\1_EeN`:\\1_EeN`:\\1^EfNa:[1^EfNa:[1^EfNa:\\1]EeNb:\\1^EdNa:]1_EcN`:^1`EcN^:^1bEbN]:`1bE`N]:a1cE_N]:a1cE_N\\:b1dE^NY:e1gE[NU:i1kEWNR:m1mESNP:P2PFPNm9S2SFmMj9V2VFjMg9Y2YFgMd9\\2\\FdMb9_2]FaM`9b2`F^M^9d2bF\\M\\9f2dFZMY9i2gFWMW9k2iFUMU9n2jFRMS9Q3mFoLQ9S3oFmLo8U3QGkLl8X3TGhLi8\\3VGdLd8b3\\G^L`8f3`GZL]8i3cGWL[8k3eGULY8n3fGRLX8P4hGPLV8R4jGnKT8T4lGlKQ8X4nGhKP8Z4PHfKn7\\4RHdKl7_4SHaKk7a4UH_Ki7c4WH]Kf7g4YHYKe7i4[HWKc7k4]HUKa7m4_HSK_7P5`HPK^7R5bHnJ[7U5fHjJX7Y5gHgJW7CVHa5c0lJU7AZHe5a0jJS7_O^Hi5?hJP7^OcHm5 in this image.", "objects": [{"patches": [126, 127, 128, 129, 130, 131, 152, 153, 154, 155, 156, 157, 177, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 225, 226, 227, 228, 248, 249, 250, 272, 273, 295, 296, 297, 318, 319, 320, 341, 342, 343], "bbox": [0.46854687500000003, 0.294375, 0.95225, 0.8808958333333333], "iscrowd": 0, "area": 19090.282950000004, "rle": {"size": [480, 640], "counts": "]h\\41o>000010O000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O0001O01O00001O01O01O00001O0001O01O00001O00010O01O010O0010O01O0010O01O01O00010O00001O0000001O00001O00001O00001O0000001O000O2O00001O00001N10001O0000001N101O001O1O001O1O001O001O1O001O1O001O001O1O001O1O0010O01O1O001O1O010O00100O00100O0010O10000O10000O010O10O010O0010O010O01O010O010O003N3L4L4M4K3N0O0010O010O00010O01O01O010O01O010O00010O010O0010O0010O004M2M3M3N2M3N2M3M3N2M4M2M3M3N2M3cG`LT5c3hJ_LX5c3eJ^LZ5e3bJ\\L^5h3]JZLb5i3[JXLd5k3XJWLg5l3VJULi5n3SJSLm5P4mITLR6o3aI\\L^6g3VIcLi6`3jHkLU7X3^HSMa7Q3RHXMn7n4O1O1O1O02O001O1O0O2N1O2N2N0aLlG2T8NmG1T8MoG2P8NQH1P8NQH1o7NTH0l70UHOl7OWHOi71XHNi71XHNh71[HNd72]HMd71_HMa73`HId76]HGe78^HCe7:_HBe79aHCa79dHC_79fHE\\76jHGW76mHIT75oHIQ7lNRG?o1d0o6lNXG:k1i0n6lN]G4h1n0k6mNcGOd1R1j6nNhGIb1W1f6oNoGC]1]1d6oNUH^OY1b1b6oNiJo0W5QOjJn0V5QOmJl0U5ROmJm0S5ROoJm0Q5RORKl0n4SOTKl0l4TOVKj0j4UOXKi0i4VOYKi0g4VO\\Kh0e4VO]Ki0c4VO`Kh0`4WObKh0^4XOcKf0^4YOfKd0Z4[OkKa0U4^OPL>Q4@TL<\\9L4KeT>"}, "label": "a brown cat sleeping behind the black cat"}]} {"id": 274853, "image": "COCO_train2014_000000274853.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"light colored cat\"."}], "task": "refering", "answer_template": "The \"light colored cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 130, 131, 152, 153, 154, 155, 156, 157, 177, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 225, 226, 227, 228, 248, 249, 250, 272, 273, 295, 296, 297, 318, 319, 320, 341, 342, 343], "bbox": [0.46854687500000003, 0.294375, 0.95225, 0.8808958333333333], "iscrowd": 0, "area": 19090.282950000004, "rle": {"size": [480, 640], "counts": "]h\\41o>000010O000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O0001O01O00001O01O01O00001O0001O01O00001O00010O01O010O0010O01O0010O01O01O00010O00001O0000001O00001O00001O00001O0000001O000O2O00001O00001N10001O0000001N101O001O1O001O1O001O001O1O001O1O001O001O1O001O1O0010O01O1O001O1O010O00100O00100O0010O10000O10000O010O10O010O0010O010O01O010O010O003N3L4L4M4K3N0O0010O010O00010O01O01O010O01O010O00010O010O0010O0010O004M2M3M3N2M3N2M3M3N2M4M2M3M3N2M3cG`LT5c3hJ_LX5c3eJ^LZ5e3bJ\\L^5h3]JZLb5i3[JXLd5k3XJWLg5l3VJULi5n3SJSLm5P4mITLR6o3aI\\L^6g3VIcLi6`3jHkLU7X3^HSMa7Q3RHXMn7n4O1O1O1O02O001O1O0O2N1O2N2N0aLlG2T8NmG1T8MoG2P8NQH1P8NQH1o7NTH0l70UHOl7OWHOi71XHNi71XHNh71[HNd72]HMd71_HMa73`HId76]HGe78^HCe7:_HBe79aHCa79dHC_79fHE\\76jHGW76mHIT75oHIQ7lNRG?o1d0o6lNXG:k1i0n6lN]G4h1n0k6mNcGOd1R1j6nNhGIb1W1f6oNoGC]1]1d6oNUH^OY1b1b6oNiJo0W5QOjJn0V5QOmJl0U5ROmJm0S5ROoJm0Q5RORKl0n4SOTKl0l4TOVKj0j4UOXKi0i4VOYKi0g4VO\\Kh0e4VO]Ki0c4VO`Kh0`4WObKh0^4XOcKf0^4YOfKd0Z4[OkKa0U4^OPL>Q4@TL<\\9L4KeT>"}, "label": "light colored cat"}]} {"id": 242090, "image": "COCO_train2014_000000242090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black haired indian man facing a vendor in front of bananas\"."}], "task": "refering", "answer_template": "The \"a black haired indian man facing a vendor in front of bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 144, 145, 146, 167, 168, 169, 190, 191, 192, 212, 213, 214, 215, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 325, 326, 327, 328, 329, 330, 331, 348, 349, 350, 351, 352, 353, 354], "bbox": [0.12142187499999998, 0.3684090909090909, 0.48928125, 0.9865909090909092], "iscrowd": 0, "area": 36023.27020000001, "rle": {"size": [440, 640], "counts": "PlQ12d=4L3M3M3M3N1N3M3M3XDCQ:`0jEDT:?gEEW:>dEF[:<`EH^:;]EIa::YEKe:8VELh:c1N2O1N2N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2O1O100O100O1O2O0O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100000000O1000000YN^KXJb4P5YLmJg3Q5_LjJb3U5bLhJ^3W5fLdJ\\3Z5hLcJZ3[5iLaJY3]5kL`JV3_5mL]JU3b5nL[JS3d5a2O100O100O100O100O100O100O100O1O100O100O100O10000O1000000O100001O00000000001O0001O000001O00000000001O001O001O001[JcJm2^5kLjJT3V5iLoJU3R5gLRKX3n4fLUKY3l4cLYK[3h4aL\\K^3d4^LaKa3`4[LdKd3\\4YLiKe3X4WLlKh3U4TLoKk3T4oKQLo3T4iKPLV4T4cKPL\\4T4]KPLa4T4XKQLd4d6L3M3M3N3L3M3M3M3L5L3L4M3L4M4K4M3M3L4M4K4M3L4M4K4M300O101N1O101N2O1N3N1N2N3N1N3N1N2N3N1N3N1N2O2M2N3N1N2O2M2O2M2N2O2M2O2M2O2MSV\\4"}, "label": "a black haired indian man facing a vendor in front of bananas"}]} {"id": 242090, "image": "COCO_train2014_000000242090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white shirt standing by some bananas\"."}], "task": "refering", "answer_template": "The \"a man in a white shirt standing by some bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 144, 145, 146, 167, 168, 169, 190, 191, 192, 212, 213, 214, 215, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 301, 302, 303, 304, 305, 306, 307, 308, 325, 326, 327, 328, 329, 330, 331, 348, 349, 350, 351, 352, 353, 354], "bbox": [0.12142187499999998, 0.3684090909090909, 0.48928125, 0.9865909090909092], "iscrowd": 0, "area": 36023.27020000001, "rle": {"size": [440, 640], "counts": "PlQ12d=4L3M3M3M3N1N3M3M3XDCQ:`0jEDT:?gEEW:>dEF[:<`EH^:;]EIa::YEKe:8VELh:c1N2O1N2N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2O1O100O100O1O2O0O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100000000O1000000YN^KXJb4P5YLmJg3Q5_LjJb3U5bLhJ^3W5fLdJ\\3Z5hLcJZ3[5iLaJY3]5kL`JV3_5mL]JU3b5nL[JS3d5a2O100O100O100O100O100O100O100O1O100O100O100O10000O1000000O100001O00000000001O0001O000001O00000000001O001O001O001[JcJm2^5kLjJT3V5iLoJU3R5gLRKX3n4fLUKY3l4cLYK[3h4aL\\K^3d4^LaKa3`4[LdKd3\\4YLiKe3X4WLlKh3U4TLoKk3T4oKQLo3T4iKPLV4T4cKPL\\4T4]KPLa4T4XKQLd4d6L3M3M3N3L3M3M3M3L5L3L4M3L4M4K4M3M3L4M4K4M3L4M4K4M300O101N1O101N2O1N3N1N2N3N1N3N1N2N3N1N3N1N2O2M2N3N1N2O2M2O2M2N2O2M2O2M2O2MSV\\4"}, "label": "a man in a white shirt standing by some bananas"}]} {"id": 242090, "image": "COCO_train2014_000000242090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a striped shirt is standing oppsite and between two men\"."}], "task": "refering", "answer_template": "The \"a man in a striped shirt is standing oppsite and between two men\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 218, 219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 286, 287, 288, 289, 290, 291, 308, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 334, 335, 336, 337, 338, 354, 355, 356, 357, 358, 359, 360, 361], "bbox": [0.39175, 0.5213863636363636, 0.7447968749999999, 0.9898636363636363], "iscrowd": 0, "area": 24924.973500000007, "rle": {"size": [440, 640], "counts": "hX\\32d=2M3N2N2N2M3N2N2M3L4M3M3L4M3M3L4M3L4M3M3L4N2N2N2N2N2N2N2N2N2N2N2N2O100O1O1O1O1O100O1O1O1O1O100O1O1O10000O10000O10000O1000000O10000O10000O10000O10000O1000000mM]MPId2n6kMeHU2Y7TN`Hl1_7VN_Hk1_7WNaHi1]7ZNbHf1\\7]NbHd1\\7_NcHa1\\7`NdH`1[7_2O100O1O1O100O1O1O100O100O101O0O10000O100O10000O11O00001O00001O00001O00001O0000001O00001O00001O000]L_Ho0c7kNbHT1`7gNeHW1]7cNhH\\1Z7_NkH_1W7]NlHb1V7[NmHc1m7bMTH^2m7`MTH`2l7^MVHb2k7\\MVHd2j7[MWHe2j7XMXHh2h7SM]Hm2d7lLbHT3Q9000000000000000001O001O00001O001O00001O001O00001O001O00001O001O3M2N2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N3M2N2N2N3M2N2O2M2N2N2N3M2N2N2N3M2N2NecU2"}, "label": "a man in a striped shirt is standing oppsite and between two men"}]} {"id": 242090, "image": "COCO_train2014_000000242090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a stripped shirt looks angrily at the camera\"."}], "task": "refering", "answer_template": "The \"a man in a stripped shirt looks angrily at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 218, 219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 286, 287, 288, 289, 290, 291, 308, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 334, 335, 336, 337, 338, 354, 355, 356, 357, 358, 359, 360, 361], "bbox": [0.39175, 0.5213863636363636, 0.7447968749999999, 0.9898636363636363], "iscrowd": 0, "area": 24924.973500000007, "rle": {"size": [440, 640], "counts": "hX\\32d=2M3N2N2N2M3N2N2M3L4M3M3L4M3M3L4M3L4M3M3L4N2N2N2N2N2N2N2N2N2N2N2N2O100O1O1O1O1O100O1O1O1O1O100O1O1O10000O10000O10000O1000000O10000O10000O10000O10000O1000000mM]MPId2n6kMeHU2Y7TN`Hl1_7VN_Hk1_7WNaHi1]7ZNbHf1\\7]NbHd1\\7_NcHa1\\7`NdH`1[7_2O100O1O1O100O1O1O100O100O101O0O10000O100O10000O11O00001O00001O00001O00001O0000001O00001O00001O000]L_Ho0c7kNbHT1`7gNeHW1]7cNhH\\1Z7_NkH_1W7]NlHb1V7[NmHc1m7bMTH^2m7`MTH`2l7^MVHb2k7\\MVHd2j7[MWHe2j7XMXHh2h7SM]Hm2d7lLbHT3Q9000000000000000001O001O00001O001O00001O001O00001O001O00001O001O3M2N2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N3M2N2N2N3M2N2O2M2N2N2N3M2N2N2N3M2N2NecU2"}, "label": "a man in a stripped shirt looks angrily at the camera"}]} {"id": 242090, "image": "COCO_train2014_000000242090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy on the right closest to the pepsi cooler\"."}], "task": "refering", "answer_template": "The \"boy on the right closest to the pepsi cooler\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 179, 199, 200, 201, 202, 222, 223, 224, 225, 245, 246, 247, 248, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 315, 316, 317, 318, 319, 338, 339, 340, 341, 342, 361, 362, 363, 364, 365], "bbox": [0.6504218749999999, 0.44943181818181815, 0.9130625, 0.9887727272727272], "iscrowd": 0, "area": 23863.24945, "rle": {"size": [440, 640], "counts": "o_c53d=2M3N2M3N200O1O1O1O0O2RM]OeHd0W7BfH?V7GfH;W7JgH6U70SGZOh0g0R84RGXOj0e0S85oFZOl0b0S87nFZOm0`0S89lF\\On0gF_OP16W8Z1hGiNT8Y1jGjNT8W1jGmNR8U1lGnNQ8S1nGPOP8Q1nGSOn7o0PHTOm7n0RHUOk7k0THXOi7j0UHYOi7h0UH\\Oh7d0WH_Og7b0WHBe7?ZHDd7=ZHFd7:[HJb77]HK`76_HN^72aH1]7W3N3L3N3M2N2M4M2N3M2M3N3M2N1N3N2N00000001O000000001O000000001O0000001O000000001O0000000eJbI\\3_6aLgI[3\\6`LkI[3W6aLnI\\3U6_LnI`3T6\\LPJb3S6YLPJf3S6ULQJi3P8O1O2N1O1O001O1O1O001O1O1O001O1O001O1O1N101O1O001O001O001O001O001O001O00001O001O001O001O001O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O2N4hMVEZ1n:YNfE\\1g;D;E8H8H9Gedg0"}, "label": "boy on the right closest to the pepsi cooler"}]} {"id": 242090, "image": "COCO_train2014_000000242090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a solid white shirt with his back turned\"."}], "task": "refering", "answer_template": "The \"a man in a solid white shirt with his back turned\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 179, 199, 200, 201, 202, 222, 223, 224, 225, 245, 246, 247, 248, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 315, 316, 317, 318, 319, 338, 339, 340, 341, 342, 361, 362, 363, 364, 365], "bbox": [0.6504218749999999, 0.44943181818181815, 0.9130625, 0.9887727272727272], "iscrowd": 0, "area": 23863.24945, "rle": {"size": [440, 640], "counts": "o_c53d=2M3N2M3N200O1O1O1O0O2RM]OeHd0W7BfH?V7GfH;W7JgH6U70SGZOh0g0R84RGXOj0e0S85oFZOl0b0S87nFZOm0`0S89lF\\On0gF_OP16W8Z1hGiNT8Y1jGjNT8W1jGmNR8U1lGnNQ8S1nGPOP8Q1nGSOn7o0PHTOm7n0RHUOk7k0THXOi7j0UHYOi7h0UH\\Oh7d0WH_Og7b0WHBe7?ZHDd7=ZHFd7:[HJb77]HK`76_HN^72aH1]7W3N3L3N3M2N2M4M2N3M2M3N3M2N1N3N2N00000001O000000001O000000001O0000001O000000001O0000000eJbI\\3_6aLgI[3\\6`LkI[3W6aLnI\\3U6_LnI`3T6\\LPJb3S6YLPJf3S6ULQJi3P8O1O2N1O1O001O1O1O001O1O1O001O1O001O1O1N101O1O001O001O001O001O001O001O00001O001O001O001O001O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O2N4hMVEZ1n:YNfE\\1g;D;E8H8H9Gedg0"}, "label": "a man in a solid white shirt with his back turned"}]} {"id": 242090, "image": "COCO_train2014_000000242090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a merchant in a red plaid shirt\"."}], "task": "refering", "answer_template": "The \"a merchant in a red plaid shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 77, 78, 79, 99, 100, 101, 102, 122, 123, 124, 125, 146, 147, 148, 149, 170, 171, 172, 192, 193, 194, 195, 215, 216, 217, 218, 239, 240, 241, 263, 264], "bbox": [0.313765625, 0.13668181818181818, 0.49985937499999994, 0.7295227272727273], "iscrowd": 0, "area": 16370.060000000003, "rle": {"size": [440, 640], "counts": "j`f23_=6K4O2O1O001O1O001O1O001O1O001O1O001O1L4SOm0N3M3L4M2N3L4M2N3L4M3M2N3M3N2oGmL[5T3bJPM[5S3`JQM_5P3^JTM_5n2^JVM`5l2[JXMc5j2ZJZMd5h2WJ\\Mh5e2UJ_Mh5d2SJ`Ml5a2oIeMn5^2lIgMS6Z2iIiMV6Y2gIiMY6Y2cIjM[6X2bIjM^6X2^IjMa6X2\\IkMc6e4O100O2O00010O0000001O00010O0000001O01O01O0000001O01O0001O02N2O00100O100fKmHl1S7QNoHQ2Q7kMRIU2o6gMTIZ2l6aMWI`2j6\\MYIe2f6XM]Ih2d6TM^In2b6mLbIS3_6hLeIY3[6^LmIb3S6WLSJk3m5lK[JT4f5dKaJ]4\\7101O0O2O001O0OJ7A`0_Ob0_Ob0^Oa0_Oa0^Oc0^Oa0_O_ZY4"}, "label": "a merchant in a red plaid shirt"}]} {"id": 242090, "image": "COCO_train2014_000000242090.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a red plaid shirt behind a counter\"."}], "task": "refering", "answer_template": "The \"a man in a red plaid shirt behind a counter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 77, 78, 79, 99, 100, 101, 102, 122, 123, 124, 125, 146, 147, 148, 149, 170, 171, 172, 192, 193, 194, 195, 215, 216, 217, 218, 239, 240, 241, 263, 264], "bbox": [0.313765625, 0.13668181818181818, 0.49985937499999994, 0.7295227272727273], "iscrowd": 0, "area": 16370.060000000003, "rle": {"size": [440, 640], "counts": "j`f23_=6K4O2O1O001O1O001O1O001O1O001O1O001O1L4SOm0N3M3L4M2N3L4M2N3L4M3M2N3M3N2oGmL[5T3bJPM[5S3`JQM_5P3^JTM_5n2^JVM`5l2[JXMc5j2ZJZMd5h2WJ\\Mh5e2UJ_Mh5d2SJ`Ml5a2oIeMn5^2lIgMS6Z2iIiMV6Y2gIiMY6Y2cIjM[6X2bIjM^6X2^IjMa6X2\\IkMc6e4O100O2O00010O0000001O00010O0000001O01O01O0000001O01O0001O02N2O00100O100fKmHl1S7QNoHQ2Q7kMRIU2o6gMTIZ2l6aMWI`2j6\\MYIe2f6XM]Ih2d6TM^In2b6mLbIS3_6hLeIY3[6^LmIb3S6WLSJk3m5lK[JT4f5dKaJ]4\\7101O0O2O001O0OJ7A`0_Ob0_Ob0^Oa0_Oa0^Oc0^Oa0_O_ZY4"}, "label": "a man in a red plaid shirt behind a counter"}]} {"id": 332204, "image": "COCO_train2014_000000332204.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wood table\"."}], "task": "refering", "answer_template": "The \"wood table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 156, 157, 158, 159, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174], "bbox": [0.0, 0.8197530864197532, 0.6542592592592592, 0.9852160493827161], "iscrowd": 0, "area": 6576.313749999992, "rle": {"size": [324, 432], "counts": "h96j940000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O10001N1000000O11O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O00000000O10000000000000000000000000000O100000000000000000000000000O1000000000000000000000000O100000000007IO100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O100000000O1O1N2O1O1O1O1O1O1O1O1001O1O1O1O2N1O1O1ON2N2N2M3N2M3N2N2N2O1O1N2O1O1O1N2O1000000O1000000O10000O1000000O1000000O1000000O10000000000000000001O0000000000001O000000000000001O0000000000001O0000000000001O00000000000UOk0J7H_V_1"}, "label": "wood table"}]} {"id": 496053, "image": "COCO_train2014_000000496053.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sofa with a white cloth and a brown pillow on it\"."}], "task": "refering", "answer_template": "The \"a sofa with a white cloth and a brown pillow on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 351, 368, 369, 370, 371, 372, 373, 374], "bbox": [0.0016093750000000001, 0.6661666666666667, 0.2794375, 0.9894999999999999], "iscrowd": 0, "area": 20437.814549999996, "rle": {"size": [480, 640], "counts": "Xi0W1i=W2iMT1lN000O1000000000000000000O100000000000000000000O1000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O1000000000000000000O10000000000000010O01O1O1O00O1000000002N3M3M3M3M3M3M3M3M?Ae0[O;E0000001O000000001O00000000001O000000001O00000000001O000000001O000000001O0000001O00001O001O001O00001O001O001O00001O001O005K6I7J6I7J5J7J5J7I6K\\dg6"}, "label": "a sofa with a white cloth and a brown pillow on it"}]} {"id": 496053, "image": "COCO_train2014_000000496053.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown chair with a white blanket on the back and a pillow with a dog on it in the seat\"."}], "task": "refering", "answer_template": "The \"a brown chair with a white blanket on the back and a pillow with a dog on it in the seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325, 326, 327, 328, 345, 346, 347, 348, 349, 350, 351, 368, 369, 370, 371, 372, 373, 374], "bbox": [0.0016093750000000001, 0.6661666666666667, 0.2794375, 0.9894999999999999], "iscrowd": 0, "area": 20437.814549999996, "rle": {"size": [480, 640], "counts": "Xi0W1i=W2iMT1lN000O1000000000000000000O100000000000000000000O1000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O1000000000000000000O10000000000000010O01O1O1O00O1000000002N3M3M3M3M3M3M3M3M?Ae0[O;E0000001O000000001O00000000001O000000001O00000000001O000000001O000000001O0000001O00001O001O001O00001O001O001O00001O001O005K6I7J6I7J5J7J5J7I6K\\dg6"}, "label": "a brown chair with a white blanket on the back and a pillow with a dog on it in the seat"}]} {"id": 496053, "image": "COCO_train2014_000000496053.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing baseball cap\"."}], "task": "refering", "answer_template": "The \"man wearing baseball cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261, 280, 281, 282, 283, 303, 304, 305, 306, 327, 328, 329, 330, 350, 351, 352, 353, 374, 375, 376], "bbox": [0.17696875, 0.37135416666666665, 0.37584375000000003, 0.9915833333333333], "iscrowd": 0, "area": 22523.770249999998, "rle": {"size": [480, 640], "counts": "XXe13h>;F9G:E;F:F:F:E;dCcMe;n2L5L3N3M2O2M3M2O2M9G9H9G9HN12O1N2O1O0O2O0O100O10001N101ZGkJ^7W5`HQKY7P5fHTKU7m4jHUKU7l4iHUKV7m4hHUKV7l4iHUKU7n4jHRK\\6FYIZ59RK[6GZIY59QKZ6JZIW5;QKX6K[IV5;PKX6o5gIRJU6Q6jIPJT6S6jIoIR6U6lIlIS6U6lImIS6T6kInIT6S6hIQJW6]7O2N1O1O1O1O1O2N1O1O00cNWJdIh5X6_JeIa5V6gJhIX5T6oJjIP5R6WKkIj4o5^KnIc4m5dKmI_4n5jKkIZ4W6W23L4M3MUJVHb4g7\\K[Hf4b7XKaHk4\\7SKfHP5W7nJkHU5Q7jJQIY5Z83M3M8H7J7H7I2NO1O2N1O1N2O1O2N1O1O1O2N4L5K5K5hKbGc1d8UNRHX1T8`N[HS1j7eN]HW1g7bN`HZ1e7_NaH]1d7[NcHa1b7WNeHe1_7TNhHh1]7QNiHk1Z:L4L4K5L4L4L4L4L4L4K5H8Ggbj5"}, "label": "man wearing baseball cap"}]} {"id": 496053, "image": "COCO_train2014_000000496053.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing with light blue shirt and wearing cap\"."}], "task": "refering", "answer_template": "The \"a man standing with light blue shirt and wearing cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261, 280, 281, 282, 283, 303, 304, 305, 306, 327, 328, 329, 330, 350, 351, 352, 353, 374, 375, 376], "bbox": [0.17696875, 0.37135416666666665, 0.37584375000000003, 0.9915833333333333], "iscrowd": 0, "area": 22523.770249999998, "rle": {"size": [480, 640], "counts": "XXe13h>;F9G:E;F:F:F:E;dCcMe;n2L5L3N3M2O2M3M2O2M9G9H9G9HN12O1N2O1O0O2O0O100O10001N101ZGkJ^7W5`HQKY7P5fHTKU7m4jHUKU7l4iHUKV7m4hHUKV7l4iHUKU7n4jHRK\\6FYIZ59RK[6GZIY59QKZ6JZIW5;QKX6K[IV5;PKX6o5gIRJU6Q6jIPJT6S6jIoIR6U6lIlIS6U6lImIS6T6kInIT6S6hIQJW6]7O2N1O1O1O1O1O2N1O1O00cNWJdIh5X6_JeIa5V6gJhIX5T6oJjIP5R6WKkIj4o5^KnIc4m5dKmI_4n5jKkIZ4W6W23L4M3MUJVHb4g7\\K[Hf4b7XKaHk4\\7SKfHP5W7nJkHU5Q7jJQIY5Z83M3M8H7J7H7I2NO1O2N1O1N2O1O2N1O1O1O2N4L5K5K5hKbGc1d8UNRHX1T8`N[HS1j7eN]HW1g7bN`HZ1e7_NaH]1d7[NcHa1b7WNeHe1_7TNhHh1]7QNiHk1Z:L4L4K5L4L4L4L4L4L4K5H8Ggbj5"}, "label": "a man standing with light blue shirt and wearing cap"}]} {"id": 496053, "image": "COCO_train2014_000000496053.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a tv remote control\"."}], "task": "refering", "answer_template": "The \"a man holding a tv remote control\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 177, 178, 200, 201, 202, 223, 224, 225, 245, 246, 247, 248, 268, 269, 270, 271, 292, 293, 294, 315, 316, 317, 338, 339, 340, 361, 362, 363, 384, 385, 386], "bbox": [0.6505625, 0.36335416666666664, 0.8292187500000001, 0.9858333333333335], "iscrowd": 0, "area": 18351.706550000003, "rle": {"size": [480, 640], "counts": "TiS62m>2N2O1N2O1N1O100O100O1O100O100N2O1N2O1N101N2O001L4G8H9G8N3N2Me0\\Og0XOg1ZNg0XO4M4K4M4K4L5L3L5L3N3^N[IdJg6Z5dI[J_6a5gIZJ[6d5iIXJZ6IiH`5Q1cJX6KhHa5T1`JW6LfHb5W1_JU6MeHc5Z1\\JT6NcHe5U9000000000000001O1O1O100O2N1O1O1O1O8H9G8H4L001O1O1O001O1O1O001O1O1O001O1O1O001O1O1cLkHmNV7g0aImN`6g0WJmNj5g0mJmNT5g0cKmN^4P1oKeNR4X1\\L\\Ne3b1hLRNY3l1TMhMm2U2R5M4L8G9H8H7H9H8H8G8Ihib1"}, "label": "a man holding a tv remote control"}]} {"id": 496053, "image": "COCO_train2014_000000496053.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with khaki pants holding television remotes\"."}], "task": "refering", "answer_template": "The \"a man with khaki pants holding television remotes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 177, 178, 200, 201, 202, 223, 224, 225, 245, 246, 247, 248, 268, 269, 270, 271, 292, 293, 294, 315, 316, 317, 338, 339, 340, 361, 362, 363, 384, 385, 386], "bbox": [0.6505625, 0.36335416666666664, 0.8292187500000001, 0.9858333333333335], "iscrowd": 0, "area": 18351.706550000003, "rle": {"size": [480, 640], "counts": "TiS62m>2N2O1N2O1N1O100O100O1O100O100N2O1N2O1N101N2O001L4G8H9G8N3N2Me0\\Og0XOg1ZNg0XO4M4K4M4K4L5L3L5L3N3^N[IdJg6Z5dI[J_6a5gIZJ[6d5iIXJZ6IiH`5Q1cJX6KhHa5T1`JW6LfHb5W1_JU6MeHc5Z1\\JT6NcHe5U9000000000000001O1O1O100O2N1O1O1O1O8H9G8H4L001O1O1O001O1O1O001O1O1O001O1O1O001O1O1cLkHmNV7g0aImN`6g0WJmNj5g0mJmNT5g0cKmN^4P1oKeNR4X1\\L\\Ne3b1hLRNY3l1TMhMm2U2R5M4L8G9H8H7H9H8H8G8Ihib1"}, "label": "a man with khaki pants holding television remotes"}]} {"id": 20917, "image": "COCO_train2014_000000020917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of sesame\"."}], "task": "refering", "answer_template": "The \"a bowl of sesame\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 63, 64, 65, 83, 84, 85, 86, 87, 88, 89, 106, 107, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 153, 154, 155, 156], "bbox": [0.614125, 0.13019370460048427, 0.8881406250000001, 0.44583535108958844], "iscrowd": 0, "area": 16569.88835, "rle": {"size": [413, 640], "counts": "lbn45_< in this image.", "objects": [{"patches": [60, 61, 62, 63, 64, 65, 83, 84, 85, 86, 87, 88, 89, 106, 107, 108, 109, 110, 111, 112, 130, 131, 132, 133, 134, 153, 154, 155, 156], "bbox": [0.614125, 0.13019370460048427, 0.8881406250000001, 0.44583535108958844], "iscrowd": 0, "area": 16569.88835, "rle": {"size": [413, 640], "counts": "lbn45_< in this image.", "objects": [{"patches": [199, 200, 201, 221, 222, 223, 224, 225, 244, 245, 246, 247, 248, 269, 270, 292, 293, 315, 316, 317, 338, 339], "bbox": [0.6209687500000001, 0.5068526785714286, 0.801609375, 0.9061607142857142], "iscrowd": 0, "area": 8280.251449999996, "rle": {"size": [448, 640], "counts": "kn]52i=a0_O6K4O10000O010O100O10BVCFi<7]CGc<6bCH]<9eCD\\<;d010000O100O010O100O100O01000O100O10O0100O2O0O101N100DZO^Cg0`<\\O]Ck0[XHBi7=XHBi7XH@k7>m2N2M4M4L`Yg1"}, "label": "a guy riding a skateboard"}]} {"id": 160188, "image": "COCO_train2014_000000160188.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"guy with arms out skateboarding with headband on\"."}], "task": "refering", "answer_template": "The \"guy with arms out skateboarding with headband on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 221, 222, 223, 224, 225, 244, 245, 246, 247, 248, 269, 270, 292, 293, 315, 316, 317, 338, 339], "bbox": [0.6209687500000001, 0.5068526785714286, 0.801609375, 0.9061607142857142], "iscrowd": 0, "area": 8280.251449999996, "rle": {"size": [448, 640], "counts": "kn]52i=a0_O6K4O10000O010O100O10BVCFi<7]CGc<6bCH]<9eCD\\<;d010000O100O010O100O100O01000O100O10O0100O2O0O101N100DZO^Cg0`<\\O]Ck0[XHBi7=XHBi7XH@k7>m2N2M4M4L`Yg1"}, "label": "guy with arms out skateboarding with headband on"}]} {"id": 111040, "image": "COCO_train2014_000000111040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue cuchine\"."}], "task": "refering", "answer_template": "The \"a blue cuchine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 253, 254, 276, 277, 278, 281, 282, 283, 284, 285, 286, 287, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.6881690140845071, 0.9819687500000001, 1.0], "iscrowd": 0, "area": 38133.48635, "rle": {"size": [426, 640], "counts": "V9g1e;f1ZNO0O2N2O1N1O2O1N2O1N1O2O1N2N100O10O01O010O00100O010O00O2O00001N101O00001N10001O0O2O00001O0O2O00001OeMkEa1V:YNPFf1X:mMnES2e:O1O2N1O1O1O1O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O00O10000000000O10000000000O10000000000O10000000000O10000000000O1000000000000O10000000000O1000000000000O100000000000000000000001O0000000000001O00000000000000000000000000000000O100000000O100000000O1000000O100000000O100000000O1000000O100000000O10000O10000O1000000O10000O10000000000001O000000001O000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000001O001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O0000O10000O100O1000000001O00001O00010O00001O00001O001O00001O00001O00001O00001O001O00001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O100000000000000000000000000000000000000000000000000000000000000O100O10000O100O100O10000O100O100O1000000000000000000000000001O000000000000000000000000000000001O00O100000000000000000000O100000000000000001O001O1O0O2O1N1O2O0O2Oi0VO^c4"}, "label": "a blue cuchine"}]} {"id": 111040, "image": "COCO_train2014_000000111040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cat bed\"."}], "task": "refering", "answer_template": "The \"cat bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 253, 254, 276, 277, 278, 281, 282, 283, 284, 285, 286, 287, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.6881690140845071, 0.9819687500000001, 1.0], "iscrowd": 0, "area": 38133.48635, "rle": {"size": [426, 640], "counts": "V9g1e;f1ZNO0O2N2O1N1O2O1N2O1N1O2O1N2N100O10O01O010O00100O010O00O2O00001N101O00001N10001O0O2O00001O0O2O00001OeMkEa1V:YNPFf1X:mMnES2e:O1O2N1O1O1O1O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O00O10000000000O10000000000O10000000000O10000000000O10000000000O1000000000000O10000000000O1000000000000O100000000000000000000001O0000000000001O00000000000000000000000000000000O100000000O100000000O1000000O100000000O100000000O1000000O100000000O10000O10000O1000000O10000O10000000000001O000000001O000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000001O001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O0000O10000O100O1000000001O00001O00010O00001O00001O001O00001O00001O00001O00001O001O00001O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O100000000000000000000000000000000000000000000000000000000000000O100O10000O100O100O10000O100O100O1000000000000000000000000001O000000000000000000000000000000001O00O100000000000000000000O100000000000000001O001O1O0O2O1N1O2O0O2Oi0VO^c4"}, "label": "cat bed"}]} {"id": 111040, "image": "COCO_train2014_000000111040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"decorative pillow behind the cats\"."}], "task": "refering", "answer_template": "The \"decorative pillow behind the cats\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 99, 100, 101, 102, 107, 108, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 158, 159, 160, 183, 206], "bbox": [0.24162499999999998, 0.0028873239436619717, 0.998859375, 0.5640845070422534], "iscrowd": 0, "area": 53893.51, "rle": {"size": [426, 640], "counts": "ebP21Y=0O10000O2O0O10001N10000O101O0O100O101O0O10001N100O10001N10000O101O0O100O2O000O10001N10000O2O0O10000O2O000O100O2O000O101O0O10000O2O0O10001N10000O101O0O100O101O0O10001N100O10001N10000O101N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O1O2N1N2M3N2M4M2M3M3N3L3N2N2N3M2N3M3OnNoFkMQ9U2RGiMm8V2XGgMg8Y2\\GeMc8Z2bGbM^8^2eG`MZ8_2kG^MT8b2oG\\MP8c2UHYMk7g2XHWMg7h2^HUMa7j2`100O11O001O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O000000000000O1000000O100000000O100000000O1000000O101O00000O1000000O100000000O100000000O1000000O100000000O100000000O1000000O10000O1O1O1N2O1O1O1N2O1O1O1N2O1O1001O1O1O1O1O1O001O1O1O1O1O001O2N2N2N1O2N2N2N2N2N2N2N2N2N1O1O001O001O001O1O001O001O001O001O1O001O001O001O001O001O1O001O001O001O001O001O1O001O001O001O001O001O1O001O00001O00000000001O00000000001O00000000001O00000000001O00000000001O000000001O00000000001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O0010O01O00001O001O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N2N1O2N2N1O2N2N]KUKi0i4jLhMS3b9gLj5"}, "label": "decorative pillow behind the cats"}]} {"id": 111040, "image": "COCO_train2014_000000111040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"patterned pillows behind the cats\"."}], "task": "refering", "answer_template": "The \"patterned pillows behind the cats\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 99, 100, 101, 102, 107, 108, 109, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 158, 159, 160, 183, 206], "bbox": [0.24162499999999998, 0.0028873239436619717, 0.998859375, 0.5640845070422534], "iscrowd": 0, "area": 53893.51, "rle": {"size": [426, 640], "counts": "ebP21Y=0O10000O2O0O10001N10000O101O0O100O101O0O10001N100O10001N10000O101O0O100O2O000O10001N10000O2O0O10000O2O000O100O2O000O101O0O10000O2O0O10001N10000O101O0O100O101O0O10001N100O10001N10000O101N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O1O2N1N2M3N2M4M2M3M3N3L3N2N2N3M2N3M3OnNoFkMQ9U2RGiMm8V2XGgMg8Y2\\GeMc8Z2bGbM^8^2eG`MZ8_2kG^MT8b2oG\\MP8c2UHYMk7g2XHWMg7h2^HUMa7j2`100O11O001O1O1O1O001O1O1O1O1O001O1O1O1O1O001O1O1O1O000000000000O1000000O100000000O100000000O1000000O101O00000O1000000O100000000O100000000O1000000O100000000O100000000O1000000O10000O1O1O1N2O1O1O1N2O1O1O1N2O1O1001O1O1O1O1O1O001O1O1O1O1O001O2N2N2N1O2N2N2N2N2N2N2N2N2N1O1O001O001O001O1O001O001O001O001O1O001O001O001O001O001O1O001O001O001O001O001O1O001O001O001O001O001O1O001O00001O00000000001O00000000001O00000000001O00000000001O00000000001O000000001O00000000001O001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O001O001O0010O01O00001O001O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001O2N1O2N2N2N1O2N2N1O2N2N2N1O2N2N2N1O2N2N1O2N2N]KUKi0i4jLhMS3b9gLj5"}, "label": "patterned pillows behind the cats"}]} {"id": 111040, "image": "COCO_train2014_000000111040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black cat is sleeping near another cat on top of a bed\"."}], "task": "refering", "answer_template": "The \"a black cat is sleeping near another cat on top of a bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303], "bbox": [0.0, 0.25077464788732395, 0.612015625, 0.8794131455399061], "iscrowd": 0, "area": 79218.24940000002, "rle": {"size": [426, 640], "counts": "n7j0n:e1SOl0L4M3L4M4L3L4M3M2M4M3L3N3M3L3N3M3L3N3L2O1O2O0O2O0O101N101N100O2N100O2O0O2O0O101N2O0O10O01O10O0100O1SIUJ[6j5cI[J[6e5bI`J\\6`5aIdJ^6\\5`IhJ^6W5`InJ^6Q6N1O0O2O1O1O1O001O1O1O1N2N1O2N2O0O2N1O100O2O0O100O2O000O101N100O101O0O100O2O0O10000O100O100O101O0O10000O10000O10000O2O000O10000O1000000000001OO100O01000O100O10000O100O10000O10000O100O10000O100O10000O2O00000O100000000O10001O000O100000000O2O0000000O2O0000000O101O00000000001O000000001O000O1000001O00000000001O000000000O2O0000000000001O0O10000000001O0000000O101O0000000000001N10000000000000001O00000000000000000000000000001O000000000001O0000001O00000010O000001O00000010O0O101O00001O00001O000O2N1O1O2N1O1O3M3M3M3L4M3M3M3M3M3M3M3M3M2N3M2N3M3M2N3M2N3M2N3M2O2N2M2O2N1N3N2N2M2O2N2M3N1O2N2L4L3L5L4K5L4K4N3L4L3M4L3N3L4L3M4M2M4L3M1O2O1N1O2O1N1O2O1N1O2O1N1O2O1N1O2M3N1O2M4M3M4K5L3M[nV3"}, "label": "a black cat is sleeping near another cat on top of a bed"}]} {"id": 111040, "image": "COCO_train2014_000000111040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black cat curled up on cushion\"."}], "task": "refering", "answer_template": "The \"black cat curled up on cushion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303], "bbox": [0.0, 0.25077464788732395, 0.612015625, 0.8794131455399061], "iscrowd": 0, "area": 79218.24940000002, "rle": {"size": [426, 640], "counts": "n7j0n:e1SOl0L4M3L4M4L3L4M3M2M4M3L3N3M3L3N3M3L3N3L2O1O2O0O2O0O101N101N100O2N100O2O0O2O0O101N2O0O10O01O10O0100O1SIUJ[6j5cI[J[6e5bI`J\\6`5aIdJ^6\\5`IhJ^6W5`InJ^6Q6N1O0O2O1O1O1O001O1O1O1N2N1O2N2O0O2N1O100O2O0O100O2O000O101N100O101O0O100O2O0O10000O100O100O101O0O10000O10000O10000O2O000O10000O1000000000001OO100O01000O100O10000O100O10000O10000O100O10000O100O10000O2O00000O100000000O10001O000O100000000O2O0000000O2O0000000O101O00000000001O000000001O000O1000001O00000000001O000000000O2O0000000000001O0O10000000001O0000000O101O0000000000001N10000000000000001O00000000000000000000000000001O000000000001O0000001O00000010O000001O00000010O0O101O00001O00001O000O2N1O1O2N1O1O3M3M3M3L4M3M3M3M3M3M3M3M3M2N3M2N3M3M2N3M2N3M2N3M2O2N2M2O2N1N3N2N2M2O2N2M3N1O2N2L4L3L5L4K5L4K4N3L4L3M4L3N3L4L3M4M2M4L3M1O2O1N1O2O1N1O2O1N1O2O1N1O2O1N1O2M3N1O2M4M3M4K5L3M[nV3"}, "label": "black cat curled up on cushion"}]} {"id": 111040, "image": "COCO_train2014_000000111040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sleeping grey cat with black stripes\"."}], "task": "refering", "answer_template": "The \"sleeping grey cat with black stripes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 132, 133, 134, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316], "bbox": [0.46521875, 0.2107511737089202, 1.0, 0.8860328638497653], "iscrowd": 0, "area": 66855.4377, "rle": {"size": [426, 640], "counts": "WZl31X=1RH4]2MbM;V2FjM=R2EmM?n1BRNa0j1@UNe0f1\\OZNg0b1[O]NP1W1QOiNP1U1QOkNQ1R1POmNS1P1nNPOS1n0oNROR1k0oNVOQ1h0POXOR1e0oN\\OQ1b0QO^OP1?QOAQ1RLDn3E:L4K6K4L4L4L5KWH"}, "label": "sleeping grey cat with black stripes"}]} {"id": 111040, "image": "COCO_train2014_000000111040.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a multi - colored cat sleeping next to a black cat\"."}], "task": "refering", "answer_template": "The \"a multi - colored cat sleeping next to a black cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 132, 133, 134, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316], "bbox": [0.46521875, 0.2107511737089202, 1.0, 0.8860328638497653], "iscrowd": 0, "area": 66855.4377, "rle": {"size": [426, 640], "counts": "WZl31X=1RH4]2MbM;V2FjM=R2EmM?n1BRNa0j1@UNe0f1\\OZNg0b1[O]NP1W1QOiNP1U1QOkNQ1R1POmNS1P1nNPOS1n0oNROR1k0oNVOQ1h0POXOR1e0oN\\OQ1b0QO^OP1?QOAQ1RLDn3E:L4K6K4L4L4L5KWH"}, "label": "a multi - colored cat sleeping next to a black cat"}]} {"id": 397760, "image": "COCO_train2014_000000397760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black sheep lying down looking at the camera\"."}], "task": "refering", "answer_template": "The \"a black sheep lying down looking at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 317], "bbox": [0.593546875, 0.37635416666666666, 1.0, 0.7871041666666667], "iscrowd": 0, "area": 33814.671050000004, "rle": {"size": [480, 640], "counts": "g[b58_>;E;D in this image.", "objects": [{"patches": [152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 203, 204, 205, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 317], "bbox": [0.593546875, 0.37635416666666666, 1.0, 0.7871041666666667], "iscrowd": 0, "area": 33814.671050000004, "rle": {"size": [480, 640], "counts": "g[b58_>;E;D in this image.", "objects": [{"patches": [73, 74, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210], "bbox": [0.0, 0.226625, 0.25710937500000003, 0.5927083333333333], "iscrowd": 0, "area": 19017.23985000001, "rle": {"size": [480, 640], "counts": "S5c3];1O00001O001O00001O00001O0O10001O000000000000O1O1O100O1O100O1O100O10000O10000O0100000000000O100000O01000000O1000000O010O100O10000O100O10O0QOfLZFZ3e9hLZFW3f9kLYFT3g9nLXFQ3g9RMXFn2g9TMWFm2i9SMWFo2h9gLbE4f0V3l9jLTFW3k9jLSFW3m9iLSFV3o9iLQFV3P:jLPFX3n9iLPFX3P:hLPFX3o9iLQFW3o9iLQFV3o9lLoEU3Q:kLoEU3P:lLPFS3Q:mLoES3Q:nLmES3S:mLmER3n9cLfE;P3n9TMQFm2o9SMQFm2o9SMQFl2P:TMPFl2P:TMoEm2P:TMPFl2P:TMPFk2Q:TMmEo2S:RMiEQ3W:oLfES3[:nL`EV3`:e00O10000O10000O01O1O001O10O00010O100O1O100O10O01O100O100O1O100O100M3N2N2M3N2N2M3N2N3L3K6C<^Oc0M2AZC`Nk<^1YC\\NjK5M3Lcon6"}, "label": "the sheep on the far left"}]} {"id": 397760, "image": "COCO_train2014_000000397760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lamb on the left side looking away\"."}], "task": "refering", "answer_template": "The \"the lamb on the left side looking away\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210], "bbox": [0.0, 0.226625, 0.25710937500000003, 0.5927083333333333], "iscrowd": 0, "area": 19017.23985000001, "rle": {"size": [480, 640], "counts": "S5c3];1O00001O001O00001O00001O0O10001O000000000000O1O1O100O1O100O1O100O10000O10000O0100000000000O100000O01000000O1000000O010O100O10000O100O10O0QOfLZFZ3e9hLZFW3f9kLYFT3g9nLXFQ3g9RMXFn2g9TMWFm2i9SMWFo2h9gLbE4f0V3l9jLTFW3k9jLSFW3m9iLSFV3o9iLQFV3P:jLPFX3n9iLPFX3P:hLPFX3o9iLQFW3o9iLQFV3o9lLoEU3Q:kLoEU3P:lLPFS3Q:mLoES3Q:nLmES3S:mLmER3n9cLfE;P3n9TMQFm2o9SMQFm2o9SMQFl2P:TMPFl2P:TMoEm2P:TMPFl2P:TMPFk2Q:TMmEo2S:RMiEQ3W:oLfES3[:nL`EV3`:e00O10000O10000O01O1O001O10O00010O100O1O100O10O01O100O100O1O100O100M3N2N2M3N2N2M3N2N3L3K6C<^Oc0M2AZC`Nk<^1YC\\NjK5M3Lcon6"}, "label": "the lamb on the left side looking away"}]} {"id": 397760, "image": "COCO_train2014_000000397760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"beer in the middle\"."}], "task": "refering", "answer_template": "The \"beer in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 85, 86, 87, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 170, 171, 172, 173, 174, 175], "bbox": [0.366125, 0.17433333333333334, 0.8290312499999999, 0.4625], "iscrowd": 0, "area": 23966.288049999996, "rle": {"size": [480, 640], "counts": "jk]38f>3N1N3M2O2M2O2M2O2M2O2M2O2M2O1N2O1N2O1N2O0O2O1O1N2TO]NZDd1e;^NXDd1h;\\NWDe1h;\\NVDe1k;\\NSDe1l;\\NSDe1m;\\NPDg1n;ZNQDg1o;YNoCj1o;XNoCi1P100O1O001O1O1O001O1O1O001O1O001O1O1O001O000001O00000001O000001O0000000000001O0000000000001O000O100O10001N10000O101N10000O101N10000O100O2O000O2O0O101O0O10001N100O2O0O101N1O101N100O2O0O101N100O101N00O2N1O2N1O2N1O200O1O010O10O01O100O00100O1O010O100O001O1O1O0O2O1O1N101O1N1100000O010000O1000000O1000000O01000O1000000O10000O10O10O1O100O1O100O1O100000001O0O101O00001O00001O00001O00001O01O010O1O010O1O010O1O001O1O001O1O010O001O1O001O1O001O1O1O1O2N1O1O1O100O1O1O2N1O1O1O1O1N2N2M4M2N2N2M3M4M7H8H8H8H8HR]c1"}, "label": "beer in the middle"}]} {"id": 397760, "image": "COCO_train2014_000000397760.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sheep facing left\"."}], "task": "refering", "answer_template": "The \"sheep facing left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 85, 86, 87, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 170, 171, 172, 173, 174, 175], "bbox": [0.366125, 0.17433333333333334, 0.8290312499999999, 0.4625], "iscrowd": 0, "area": 23966.288049999996, "rle": {"size": [480, 640], "counts": "jk]38f>3N1N3M2O2M2O2M2O2M2O2M2O2M2O1N2O1N2O1N2O0O2O1O1N2TO]NZDd1e;^NXDd1h;\\NWDe1h;\\NVDe1k;\\NSDe1l;\\NSDe1m;\\NPDg1n;ZNQDg1o;YNoCj1o;XNoCi1P100O1O001O1O1O001O1O1O001O1O001O1O1O001O000001O00000001O000001O0000000000001O0000000000001O000O100O10001N10000O101N10000O101N10000O100O2O000O2O0O101O0O10001N100O2O0O101N1O101N100O2O0O101N100O101N00O2N1O2N1O2N1O200O1O010O10O01O100O00100O1O010O100O001O1O1O0O2O1O1N101O1N1100000O010000O1000000O1000000O01000O1000000O10000O10O10O1O100O1O100O1O100000001O0O101O00001O00001O00001O00001O01O010O1O010O1O010O1O001O1O001O1O010O001O1O001O1O001O1O1O1O2N1O1O1O100O1O1O2N1O1O1O1O1N2N2M4M2N2N2M3M4M7H8H8H8H8HR]c1"}, "label": "sheep facing left"}]} {"id": 78274, "image": "COCO_train2014_000000078274.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a futon in front of a tv , right next to a glass door\"."}], "task": "refering", "answer_template": "The \"a futon in front of a tv , right next to a glass door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283], "bbox": [0.0, 0.5907777777777778, 0.355875, 1.0], "iscrowd": 0, "area": 24394.7323, "rle": {"size": [360, 640], "counts": "e6c4f6O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1OO10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O1000000000mKeIY3[6fLgIY3\\6bLfI^3R70001O000000000000000000000000000000000000001O2N2N1O2N1O2N2N1O2N1O2N2N2N2N2N2N2N3M2N3M4L3M3M3M4L1O2N1O2N1O1O1O1O1O1O001O1O1O1O1O001O1O1O1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1Oi_`4"}, "label": "a futon in front of a tv , right next to a glass door"}]} {"id": 78274, "image": "COCO_train2014_000000078274.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the couch\"."}], "task": "refering", "answer_template": "The \"the couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283], "bbox": [0.0, 0.5907777777777778, 0.355875, 1.0], "iscrowd": 0, "area": 24394.7323, "rle": {"size": [360, 640], "counts": "e6c4f6O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1OO10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O1000000000mKeIY3[6fLgIY3\\6bLfI^3R70001O000000000000000000000000000000000000001O2N2N1O2N1O2N2N1O2N1O2N2N2N2N2N2N2N3M2N3M4L3M3M3M4L1O2N1O2N1O1O1O1O1O1O001O1O1O1O1O001O1O1O1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1Oi_`4"}, "label": "the couch"}]} {"id": 78274, "image": "COCO_train2014_000000078274.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the ottoman on front of the couch\"."}], "task": "refering", "answer_template": "The \"the ottoman on front of the couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 269, 284, 285, 286, 287, 288, 289, 290, 291], "bbox": [0.384171875, 0.7387777777777778, 0.701234375, 0.999111111111111], "iscrowd": 0, "area": 15769.503150000002, "rle": {"size": [360, 640], "counts": "ajf22S;3N2M3N2[NEZH>a7NVH4e78RHJl7:oGIQ87bG6^8[1N1000000O100000000O101O000O100000000O1000001O0O1000000000000000001O0000000O1000000000000000000000000O100000000000000000000000000O1000000000000000000000000O1000000000000000000000000O100000000000000000000000000O1000000000000001O1O001O001O001O001O001O1O001O001O001O001O001O1O001O001O001O001O001O1O010O001O001O001O001O1O001O001O001O001O1O001O001O001O001O001O1O001O001O001N101N101^OaFXOa9e0fFUO[9i0d0M2]OekR2"}, "label": "the ottoman on front of the couch"}]} {"id": 78274, "image": "COCO_train2014_000000078274.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a rectangular ottoman in front of the tv\"."}], "task": "refering", "answer_template": "The \"a rectangular ottoman in front of the tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 244, 245, 262, 263, 264, 265, 266, 267, 268, 269, 284, 285, 286, 287, 288, 289, 290, 291], "bbox": [0.384171875, 0.7387777777777778, 0.701234375, 0.999111111111111], "iscrowd": 0, "area": 15769.503150000002, "rle": {"size": [360, 640], "counts": "ajf22S;3N2M3N2[NEZH>a7NVH4e78RHJl7:oGIQ87bG6^8[1N1000000O100000000O101O000O100000000O1000001O0O1000000000000000001O0000000O1000000000000000000000000O100000000000000000000000000O1000000000000000000000000O1000000000000000000000000O100000000000000000000000000O1000000000000001O1O001O001O001O001O001O1O001O001O001O001O001O1O001O001O001O001O001O1O010O001O001O001O001O1O001O001O001O001O1O001O001O001O001O001O1O001O001O001N101N101^OaFXOa9e0fFUO[9i0d0M2]OekR2"}, "label": "a rectangular ottoman in front of the tv"}]} {"id": 258505, "image": "COCO_train2014_000000258505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a navy blue suv with dark windows and a silver grill\"."}], "task": "refering", "answer_template": "The \"a navy blue suv with dark windows and a silver grill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 74, 75, 76, 77, 78, 79, 80, 81, 92, 93, 94, 95, 97, 98, 99, 101, 102, 103, 104, 115, 116, 117, 118, 120, 121, 122, 124, 125, 126, 127, 128, 138, 139, 140, 144, 145, 146, 148, 149, 150, 151, 161, 162, 163, 168, 169, 171, 172, 173, 174, 184, 185, 186, 191, 192, 195, 196, 197, 207, 208, 214, 215, 218, 219, 220, 230, 237, 241, 242, 243, 253], "bbox": [0.004984375, 0.08298319327731092, 0.60565625, 0.666890756302521], "iscrowd": 0, "area": 48679.39850000003, "rle": {"size": [476, 640], "counts": "Sd1]2l6d5O01O001O0]Od0ZOd0@a0O01O1O10O01O1O10O01O10O01O1O100O2N2N2O1N2N101N1O1O1000000000O100000000000000O1N2N2O1N2N2N101N2N2N2O1N2bNcH[K_7U4l1O100O10000O10000O10000O10000O10000O10000O10000SMYEd1h:[N]Ea1c:^NbE^1^:aNhEY1Y:fNlEQ1Y:nNmEg0Y:XOlEa0W:^OoE;S:DQF8P:GSF5o9JUF1m9NVFNl92WFIk96XFGi98[FCg9<\\F@f9`0[2O1O100O1O100O1O100O100000O100000000000000000000000001O001O1O1O001O1O001O1[C^On:d0lDBR;`0iDDV;>dDG[;<_DH`;:ZDLd;6WDNh;4RD1m;2mC3Q<]1O1O2N1O1O1O2N1O1O2N1O2N2N1O2N1O5K;E7I00O1O100O1O100O1O100O0011O2N1O2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N3M2N3M2N2N`McFDZ99lFFf1gMW5a2XIH^1jMW5[2`IKV1nMW5U2hILn0PNZ5R2mINe0QN^5n1SJ0U9@mFa0R9^OnFd0Q9[OoFg0o8YOQGi0n8VORGl0m8SOSGo0l8POTGR1k8lNVGV1h8jNXGX1g8gNYG[1f8eNZG\\1e8cN[G_1d8aN[Ga1d8_N[Gc1c8]N]Ge1b8[N]Gg1b8XN^Gj1a8VN]Gm1b8SN[GR2b8nM]GU2b8kM[GY2d8fM[G]2d8cMYGa2f8_MXGd2f8\\MYGg2f8YMWGk2h8WMTGl2T1[Li5j0oHo2Q1\\LQ6g0jHP3o0]LW6f0eHQ3n0]L^6c0aHS3j0_Lf6`0[HU3i0_Lm6=WHY3c0_LW7i4aH[K`7i4VH\\Kj7k51O1O1O1O002N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2L4L4L6J6K5J6J6K5J6J6K5J6J6K5J6I7^Ob0^Ob0^Olmd3"}, "label": "a navy blue suv with dark windows and a silver grill"}]} {"id": 258505, "image": "COCO_train2014_000000258505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black suv behind the boys\"."}], "task": "refering", "answer_template": "The \"the black suv behind the boys\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 74, 75, 76, 77, 78, 79, 80, 81, 92, 93, 94, 95, 97, 98, 99, 101, 102, 103, 104, 115, 116, 117, 118, 120, 121, 122, 124, 125, 126, 127, 128, 138, 139, 140, 144, 145, 146, 148, 149, 150, 151, 161, 162, 163, 168, 169, 171, 172, 173, 174, 184, 185, 186, 191, 192, 195, 196, 197, 207, 208, 214, 215, 218, 219, 220, 230, 237, 241, 242, 243, 253], "bbox": [0.004984375, 0.08298319327731092, 0.60565625, 0.666890756302521], "iscrowd": 0, "area": 48679.39850000003, "rle": {"size": [476, 640], "counts": "Sd1]2l6d5O01O001O0]Od0ZOd0@a0O01O1O10O01O1O10O01O10O01O1O100O2N2N2O1N2N101N1O1O1000000000O100000000000000O1N2N2O1N2N2N101N2N2N2O1N2bNcH[K_7U4l1O100O10000O10000O10000O10000O10000O10000O10000SMYEd1h:[N]Ea1c:^NbE^1^:aNhEY1Y:fNlEQ1Y:nNmEg0Y:XOlEa0W:^OoE;S:DQF8P:GSF5o9JUF1m9NVFNl92WFIk96XFGi98[FCg9<\\F@f9`0[2O1O100O1O100O1O100O100000O100000000000000000000000001O001O1O1O001O1O001O1[C^On:d0lDBR;`0iDDV;>dDG[;<_DH`;:ZDLd;6WDNh;4RD1m;2mC3Q<]1O1O2N1O1O1O2N1O1O2N1O2N2N1O2N1O5K;E7I00O1O100O1O100O1O100O0011O2N1O2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N3M2N3M2N2N`McFDZ99lFFf1gMW5a2XIH^1jMW5[2`IKV1nMW5U2hILn0PNZ5R2mINe0QN^5n1SJ0U9@mFa0R9^OnFd0Q9[OoFg0o8YOQGi0n8VORGl0m8SOSGo0l8POTGR1k8lNVGV1h8jNXGX1g8gNYG[1f8eNZG\\1e8cN[G_1d8aN[Ga1d8_N[Gc1c8]N]Ge1b8[N]Gg1b8XN^Gj1a8VN]Gm1b8SN[GR2b8nM]GU2b8kM[GY2d8fM[G]2d8cMYGa2f8_MXGd2f8\\MYGg2f8YMWGk2h8WMTGl2T1[Li5j0oHo2Q1\\LQ6g0jHP3o0]LW6f0eHQ3n0]L^6c0aHS3j0_Lf6`0[HU3i0_Lm6=WHY3c0_LW7i4aH[K`7i4VH\\Kj7k51O1O1O1O002N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2L4L4L6J6K5J6J6K5J6J6K5J6J6K5J6I7^Ob0^Ob0^Olmd3"}, "label": "the black suv behind the boys"}]} {"id": 119244, "image": "COCO_train2014_000000119244.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a blue coat and black pants skiing downhill\"."}], "task": "refering", "answer_template": "The \"a person in a blue coat and black pants skiing downhill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 175, 176, 177, 198, 199, 200, 221, 222, 223, 224, 245, 246, 247, 248, 269, 270, 271, 272], "bbox": [0.61021875, 0.36711899791231734, 0.8327187500000001, 0.6936951983298538], "iscrowd": 0, "area": 9018.31295, "rle": {"size": [479, 640], "counts": "WTg5b0X>8J5K5K5K5K3M4K4M4A?M2O2M2O2M3N1N3N2M3N2N2O1N2O1O2N1N2O1O1O100O2O000O2O1O0O20O0100O01O001N101N2O1O0O2O1O01O1O0010O0O1O101fN]DTOg;h0[DWOg;g0ZDWOi;g0XDWOl;f0UDYOm;d0UDZOo;c0RD[OQ in this image.", "objects": [{"patches": [152, 153, 175, 176, 177, 198, 199, 200, 221, 222, 223, 224, 245, 246, 247, 248, 269, 270, 271, 272], "bbox": [0.61021875, 0.36711899791231734, 0.8327187500000001, 0.6936951983298538], "iscrowd": 0, "area": 9018.31295, "rle": {"size": [479, 640], "counts": "WTg5b0X>8J5K5K5K5K3M4K4M4A?M2O2M2O2M3N1N3N2M3N2N2O1N2O1O2N1N2O1O1O100O2O000O2O1O0O20O0100O01O001N101N2O1O0O2O1O01O1O0010O0O1O101fN]DTOg;h0[DWOg;g0ZDWOi;g0XDWOl;f0UDYOm;d0UDZOo;c0RD[OQ in this image.", "objects": [{"patches": [215, 216, 237, 238, 239, 240, 259, 260, 261, 262, 263, 283, 284, 285, 306, 307, 308, 329, 330], "bbox": [0.29590625, 0.5219132149901381, 0.48443749999999997, 0.7887179487179488], "iscrowd": 0, "area": 8994.770300000002, "rle": {"size": [507, 640], "counts": "Y\\n25d?5K5K5K5L4K5K6J7I7I=E;D in this image.", "objects": [{"patches": [215, 216, 237, 238, 239, 240, 259, 260, 261, 262, 263, 283, 284, 285, 306, 307, 308, 329, 330], "bbox": [0.29590625, 0.5219132149901381, 0.48443749999999997, 0.7887179487179488], "iscrowd": 0, "area": 8994.770300000002, "rle": {"size": [507, 640], "counts": "Y\\n25d?5K5K5K5L4K5K6J7I7I=E;D in this image.", "objects": [{"patches": [95, 96, 97, 117, 118, 119, 120, 121, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 214, 215, 216, 217, 218, 235, 237, 238, 239, 240, 241, 263], "bbox": [0.09415625, 0.2768541666666666, 0.500390625, 0.6590625], "iscrowd": 0, "area": 27125.864800000003, "rle": {"size": [480, 640], "counts": "PZl04e>;E;H9I4K4L4L4L3O2M4L5L3L5L2M3N2M3N2O2M2N2O1N2O1N1O2O1N2N2O001O1O1O0O2O1O1O001O1O1O1O1N101O1O1O1O001O1O1O1N101O1O1O1O001O1O1O10O01O1O1O1O1O0O2O1N2N2O1O001O100O1O010O1O1O100O001O16I1000O010O01000O010O010O10O0M4M2M4M2M4M3L4L4M3L4M3L4M3L4M4K4M3L5J5K5L5N1O2OO003N3L5K4M3L3Mh0YO2M01O00010O000010O01O00010O000010O01O00010O001O01O01O001O01O01O0010O01O00100O1O1O010O1O1O10O01O00010O000001O0000010O0001O01O0001O00010O0000010O000010O0000010O0001O01O0001O01O01O0001O01O00010O0000010O000010O000001N1L4M4L3M3M3L5L3M3M4L3L4M3M4M>D=B>Clge4"}, "label": "partially split sandwich to the left of two bags of chips"}]} {"id": 209356, "image": "COCO_train2014_000000209356.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwich on the left , that ' s partially bent apart\"."}], "task": "refering", "answer_template": "The \"the sandwich on the left , that ' s partially bent apart\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 117, 118, 119, 120, 121, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 214, 215, 216, 217, 218, 235, 237, 238, 239, 240, 241, 263], "bbox": [0.09415625, 0.2768541666666666, 0.500390625, 0.6590625], "iscrowd": 0, "area": 27125.864800000003, "rle": {"size": [480, 640], "counts": "PZl04e>;E;H9I4K4L4L4L3O2M4L5L3L5L2M3N2M3N2O2M2N2O1N2O1N1O2O1N2N2O001O1O1O0O2O1O1O001O1O1O1O1N101O1O1O1O001O1O1O1N101O1O1O1O001O1O1O10O01O1O1O1O1O0O2O1N2N2O1O001O100O1O010O1O1O100O001O16I1000O010O01000O010O010O10O0M4M2M4M2M4M3L4L4M3L4M3L4M3L4M4K4M3L5J5K5L5N1O2OO003N3L5K4M3L3Mh0YO2M01O00010O000010O01O00010O000010O01O00010O001O01O01O001O01O01O0010O01O00100O1O1O010O1O1O10O01O00010O000001O0000010O0001O01O0001O00010O0000010O000010O0000010O0001O01O0001O01O01O0001O01O00010O0000010O000010O000001N1L4M4L3M3M3L5L3M3M4L3L4M3M4M>D=B>Clge4"}, "label": "the sandwich on the left , that ' s partially bent apart"}]} {"id": 209356, "image": "COCO_train2014_000000209356.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sandwich to right of chips with white substance on it\"."}], "task": "refering", "answer_template": "The \"sandwich to right of chips with white substance on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 82, 83, 84, 85, 86, 105, 106, 107, 108, 109, 110, 129, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 225, 226, 227, 228, 249], "bbox": [0.59425, 0.1506875, 0.9548593750000001, 0.6224583333333332], "iscrowd": 0, "area": 22397.5039, "rle": {"size": [480, 640], "counts": "SWb51o>4L5K5K4L5K3M1O2M3M3M3M3M2N3M3M3M3N1N3M3M3M3M3M2N2N2N2N2N2N2N2O1N2N2N2N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O10000O100O10O10O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O100O10000O100O100O100O10000O100O100002N2N3NO0000000000000001O00000000000000000000001O00000000000000000000001O00000000000000000000001O0O1N2N2N2O1N2N2N2N2N2N3M2N2N2N2N2N2N2N2O1N2N2N2N3M2N2N2N2N2N2N2N2N2O1N2N3M2N2N2O1N2N2N2N2N2N2O1N2N3M2N2N2O1N2N2N2N2N2O1N2N2N2N2N2O1Nb[="}, "label": "sandwich to right of chips with white substance on it"}]} {"id": 209356, "image": "COCO_train2014_000000209356.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich that is to the right of the other food\"."}], "task": "refering", "answer_template": "The \"a sandwich that is to the right of the other food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 82, 83, 84, 85, 86, 105, 106, 107, 108, 109, 110, 129, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 225, 226, 227, 228, 249], "bbox": [0.59425, 0.1506875, 0.9548593750000001, 0.6224583333333332], "iscrowd": 0, "area": 22397.5039, "rle": {"size": [480, 640], "counts": "SWb51o>4L5K5K4L5K3M1O2M3M3M3M3M2N3M3M3M3N1N3M3M3M3M3M2N2N2N2N2N2N2N2O1N2N2N2N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O10000O100O10O10O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O100O10000O100O100O100O10000O100O100002N2N3NO0000000000000001O00000000000000000000001O00000000000000000000001O00000000000000000000001O0O1N2N2N2O1N2N2N2N2N2N3M2N2N2N2N2N2N2N2O1N2N2N2N3M2N2N2N2N2N2N2N2N2O1N2N3M2N2N2O1N2N2N2N2N2N2O1N2N3M2N2N2O1N2N2N2N2N2O1N2N2N2N2N2O1Nb[="}, "label": "a sandwich that is to the right of the other food"}]} {"id": 422354, "image": "COCO_train2014_000000422354.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with a pink cloth and a plastic cup in his right hand\"."}], "task": "refering", "answer_template": "The \"the man with a pink cloth and a plastic cup in his right hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 125, 126, 127, 133, 134, 135, 140, 141, 142, 148, 149, 150, 155, 156, 157, 158, 163, 164, 165, 171, 172, 173, 174, 177, 178, 179, 180, 186, 187, 191, 192, 193, 194, 195, 196, 199, 200, 201, 202, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 271, 272, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 294, 295, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 317, 318, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 340, 341], "bbox": [0.094671875, 0.3438317757009346, 0.88815625, 0.9887616822429905], "iscrowd": 0, "area": 76103.05485000006, "rle": {"size": [428, 640], "counts": "\\gi05U=6mB8X<`0SDZOP;c1M4L3M4oNQ1K4K6K4L5J6M2N3N2ZOiKRHY4l7kKQHW4m7kKPHX4m7lKPHV4m7nKoGT4o7g0M3M1O010O1O010O10O1000OlJ[Hb4f7^K[Ha4e7_K]H`4b7`K`H^4`7bKaH^4^7bKbH_4]7aKcHX1J^1b7YMeHX11X1Z7`MeHW19S1Q7fMfHW1=P1m6hMgHV1>Q1m6gMeHW1`0R1l6eMdHX1b0S1]7lNeHT1Z7lNgHT1Y7kNhHU1W7kNjHU1U7kNlHU1S7kNnHU1Q7kNPIU1o6kNSIT1m6kNUIT1j6lNXIS1g6mN\\IQ1c6oN_IP1`6POaIo0_6QObIo0^6PObIP1^6PObIP1^6oNcIQ1]6mNeIS1\\6gNiIX1X6aNoI_1Q6ZNVJf1k5RN\\Jn1d5nM`JR2`5kMcJU2^5gMeJY2[5eMhJZ2U8O100000000000000O10000000O1000000000000O100O100O100O100O100O100O100O100O100O100O1eG]Mm5c2lIdMS6]2fIjMZ6V2_IQN`6P2YIWNg6i1RI^Nn6b1lHdNT7[1fHlNY7U1`HRO`7n0YHYOg7g0RH@n7`0kGGU8h20000O10000000000000000O1000000000000O1N2O1N2O1O100001O000000000000000000000000000000O1000000O100O1000000O10000000000O100000000O100000000O1000000O1000000O100000000000000000000O1000000EYJQIg5n6]JoHc5o6aJoH_5P7eJmH[5R7`0F:C=O1O1O11O00I7H8H8H8O1O1O100O1O1O1O1O1O1O1O1O100000000O100000000000000O100000000000000O100000000000000O10000001O001O1O1O001O1O1O001O1O1O001O1[JQKk2P5mLYKQ3h4nLYKQ3g4nLZKR3g4lLZKT3g4kLZKT3g4jLZKV3f4iL[KW3f4gL[KY3V5kJcJg07^4l5_KVJ`4l5kJgJU5h600000001O000000001O00000jNnGRMR8j2WHQMj7i2aHQM_7k2jHPMV7l2VHmLi8g2]GiMY8g1QHiNe7S1YH\\O_7d0[HDb7RIWOQ7i0YIiNj6W1`I]Nc6c1gIPN\\6o1`210O01O001O0010O01O001O0010O01O00N2O1N3N1N2O1N2N2O0100O100O010O100O010O100O01000O100O100O2O0O100O2O0O100O101N10000O101N1GcMdE]2]:cMaE^2`:cM^E^2c:aM[E`2f:21O1O10^EaMW:P3J7I7I6J7I6K5J4L4L5K4M5M5K6G8G:F9dHbJY6[6M4L3M4L3M4LgIYJQ5d5PK`Jo4\\5QKiJm4T5TKQKg4o4XKVKc4j4^KZK^4e4bK`KY4`4hKeKT4Y4lKlKP4T4PLPLl3o3TLWLg3h3ZL\\La3d3_LaL]3_3bLfLZ3Y3gLlLT3S3lLRMAXNo1e4aNWMhNoN`2j3hNXOi0h0XO[Oe0d0\\O@`0?AE;9GJ65@PKPOS5X1LIQKoN\\5o0BVORK<1_Oa5e0_OAoJ;1_Oc;4]D<1_Od;4ZD>2^Od;3[D>2_Od;2ZD?2_Oe;0ZD5NM3Nf;OYD6NL4Of;NXD6NN4Ng;MVD8OL40g;LVD70M30h;KUD8OM50h;JTD80N40i;ISD90N40j;HQD;1L51i;HQD:1N6Nj;IoC;1M8Nj;4oCN7Mk;5mCN:Li;5nCO:Jh;8nCN;Ig;8oCNb in this image.", "objects": [{"patches": [117, 118, 125, 126, 127, 133, 134, 135, 140, 141, 142, 148, 149, 150, 155, 156, 157, 158, 163, 164, 165, 171, 172, 173, 174, 177, 178, 179, 180, 186, 187, 191, 192, 193, 194, 195, 196, 199, 200, 201, 202, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 271, 272, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 294, 295, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 317, 318, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 340, 341], "bbox": [0.094671875, 0.3438317757009346, 0.88815625, 0.9887616822429905], "iscrowd": 0, "area": 76103.05485000006, "rle": {"size": [428, 640], "counts": "\\gi05U=6mB8X<`0SDZOP;c1M4L3M4oNQ1K4K6K4L5J6M2N3N2ZOiKRHY4l7kKQHW4m7kKPHX4m7lKPHV4m7nKoGT4o7g0M3M1O010O1O010O10O1000OlJ[Hb4f7^K[Ha4e7_K]H`4b7`K`H^4`7bKaH^4^7bKbH_4]7aKcHX1J^1b7YMeHX11X1Z7`MeHW19S1Q7fMfHW1=P1m6hMgHV1>Q1m6gMeHW1`0R1l6eMdHX1b0S1]7lNeHT1Z7lNgHT1Y7kNhHU1W7kNjHU1U7kNlHU1S7kNnHU1Q7kNPIU1o6kNSIT1m6kNUIT1j6lNXIS1g6mN\\IQ1c6oN_IP1`6POaIo0_6QObIo0^6PObIP1^6PObIP1^6oNcIQ1]6mNeIS1\\6gNiIX1X6aNoI_1Q6ZNVJf1k5RN\\Jn1d5nM`JR2`5kMcJU2^5gMeJY2[5eMhJZ2U8O100000000000000O10000000O1000000000000O100O100O100O100O100O100O100O100O100O100O1eG]Mm5c2lIdMS6]2fIjMZ6V2_IQN`6P2YIWNg6i1RI^Nn6b1lHdNT7[1fHlNY7U1`HRO`7n0YHYOg7g0RH@n7`0kGGU8h20000O10000000000000000O1000000000000O1N2O1N2O1O100001O000000000000000000000000000000O1000000O100O1000000O10000000000O100000000O100000000O1000000O1000000O100000000000000000000O1000000EYJQIg5n6]JoHc5o6aJoH_5P7eJmH[5R7`0F:C=O1O1O11O00I7H8H8H8O1O1O100O1O1O1O1O1O1O1O1O100000000O100000000000000O100000000000000O100000000000000O10000001O001O1O1O001O1O1O001O1O1O001O1[JQKk2P5mLYKQ3h4nLYKQ3g4nLZKR3g4lLZKT3g4kLZKT3g4jLZKV3f4iL[KW3f4gL[KY3V5kJcJg07^4l5_KVJ`4l5kJgJU5h600000001O000000001O00000jNnGRMR8j2WHQMj7i2aHQM_7k2jHPMV7l2VHmLi8g2]GiMY8g1QHiNe7S1YH\\O_7d0[HDb7RIWOQ7i0YIiNj6W1`I]Nc6c1gIPN\\6o1`210O01O001O0010O01O001O0010O01O00N2O1N3N1N2O1N2N2O0100O100O010O100O010O100O01000O100O100O2O0O100O2O0O100O101N10000O101N1GcMdE]2]:cMaE^2`:cM^E^2c:aM[E`2f:21O1O10^EaMW:P3J7I7I6J7I6K5J4L4L5K4M5M5K6G8G:F9dHbJY6[6M4L3M4L3M4LgIYJQ5d5PK`Jo4\\5QKiJm4T5TKQKg4o4XKVKc4j4^KZK^4e4bK`KY4`4hKeKT4Y4lKlKP4T4PLPLl3o3TLWLg3h3ZL\\La3d3_LaL]3_3bLfLZ3Y3gLlLT3S3lLRMAXNo1e4aNWMhNoN`2j3hNXOi0h0XO[Oe0d0\\O@`0?AE;9GJ65@PKPOS5X1LIQKoN\\5o0BVORK<1_Oa5e0_OAoJ;1_Oc;4]D<1_Od;4ZD>2^Od;3[D>2_Od;2ZD?2_Oe;0ZD5NM3Nf;OYD6NL4Of;NXD6NN4Ng;MVD8OL40g;LVD70M30h;KUD8OM50h;JTD80N40i;ISD90N40j;HQD;1L51i;HQD:1N6Nj;IoC;1M8Nj;4oCN7Mk;5mCN:Li;5nCO:Jh;8nCN;Ig;8oCNb in this image.", "objects": [{"patches": [6, 29, 30, 52, 53, 54, 75, 76, 77, 78, 98, 99, 100, 101, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 167, 168, 169, 170, 191, 192, 214, 215, 237, 238, 239, 260, 261, 262, 283, 284, 285, 306], "bbox": [0.23025, 0.0400625, 0.43001562500000007, 0.7714375], "iscrowd": 0, "area": 21753.648099999995, "rle": {"size": [480, 640], "counts": "\\RU2:e>3M3N2B=M4N2N2M3N1O000O10O10O1000N1N3N`LoNUIn0i6ZOSIe0l6AQI=P7GmH6T7N\\H>e7EhGl0W8WOeGk0Z8XOaGj0`8XO\\Gj0c8YOXGi0h8ZOTGi0j8ZOmFR1n8POhF[1T9hNgF]1V9eNgF_1W9cNfF`1X9aNeFc1Y9]NfFf1Y9ZNeFi1Y9XNfFj1X9WNfFl1X9UNgFm1W9P2N?A6J5K4M101N2O1N1XKVIX1k6gMUI_N8d3d6hMVIcN7f3d6cMWIgN6g3d6^MXIiN5j3e6XMXInN3l3e6SMYIPO3n3e6oLYIRO3Q4d6kLZISO2S4e6gLZIUO3T4d6eLYIVO4W4c6`L[IWO3Z4c6]LZIXO4\\4c6YLZI[O3^4c6UL[I[O3a4c6QL[I]O3c4c6nKZI^O4f4b6iK\\I_O2k4`6eK_I_O2n4^6bK`I@2P5\\6_KcI@2S5Z6\\KdI@3V5W6YKgI@3Y5U6VKhI@4[5T6SKiIB2]5T6PKkIB2_5S6nJkIB3a5]7^JdHb5\\7]JeHd5Z7\\JfHe5Y82O1O1O1O1O1O1N2O0bKVJFk57hJXOY5e0YKkNh4R1]KkNd4R1`KlNa4R1cKjN_4U1cKiN_4U1dKhN^4V1dKhN^4V1eKgN^4V1dKhN^4V1eKgN]4W1eKfN_4W1dKfN^4W1dKhN_4U1bKjN`4T1bKjN`4S1bKlNa4P1bKmNa4P1aKoN`4o0cKoN_4n0cKQO_4l0dKRO^4j0eKUO\\4i0gKUO[4h0gKVO\\41\\LNf3M_L1b3KbL4`3FfL8\\3ChL in this image.", "objects": [{"patches": [6, 29, 30, 52, 53, 54, 75, 76, 77, 78, 98, 99, 100, 101, 120, 121, 122, 123, 124, 143, 144, 145, 146, 147, 167, 168, 169, 170, 191, 192, 214, 215, 237, 238, 239, 260, 261, 262, 283, 284, 285, 306], "bbox": [0.23025, 0.0400625, 0.43001562500000007, 0.7714375], "iscrowd": 0, "area": 21753.648099999995, "rle": {"size": [480, 640], "counts": "\\RU2:e>3M3N2B=M4N2N2M3N1O000O10O10O1000N1N3N`LoNUIn0i6ZOSIe0l6AQI=P7GmH6T7N\\H>e7EhGl0W8WOeGk0Z8XOaGj0`8XO\\Gj0c8YOXGi0h8ZOTGi0j8ZOmFR1n8POhF[1T9hNgF]1V9eNgF_1W9cNfF`1X9aNeFc1Y9]NfFf1Y9ZNeFi1Y9XNfFj1X9WNfFl1X9UNgFm1W9P2N?A6J5K4M101N2O1N1XKVIX1k6gMUI_N8d3d6hMVIcN7f3d6cMWIgN6g3d6^MXIiN5j3e6XMXInN3l3e6SMYIPO3n3e6oLYIRO3Q4d6kLZISO2S4e6gLZIUO3T4d6eLYIVO4W4c6`L[IWO3Z4c6]LZIXO4\\4c6YLZI[O3^4c6UL[I[O3a4c6QL[I]O3c4c6nKZI^O4f4b6iK\\I_O2k4`6eK_I_O2n4^6bK`I@2P5\\6_KcI@2S5Z6\\KdI@3V5W6YKgI@3Y5U6VKhI@4[5T6SKiIB2]5T6PKkIB2_5S6nJkIB3a5]7^JdHb5\\7]JeHd5Z7\\JfHe5Y82O1O1O1O1O1O1N2O0bKVJFk57hJXOY5e0YKkNh4R1]KkNd4R1`KlNa4R1cKjN_4U1cKiN_4U1dKhN^4V1dKhN^4V1eKgN^4V1dKhN^4V1eKgN]4W1eKfN_4W1dKfN^4W1dKhN_4U1bKjN`4T1bKjN`4S1bKlNa4P1bKmNa4P1aKoN`4o0cKoN_4n0cKQO_4l0dKRO^4j0eKUO\\4i0gKUO[4h0gKVO\\41\\LNf3M_L1b3KbL4`3FfL8\\3ChL in this image.", "objects": [{"patches": [335, 336, 337, 338, 339, 340, 341, 358, 359, 360, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.57303125, 0.8273333333333334, 0.9383437499999999, 0.97275], "iscrowd": 0, "area": 3108.409399999999, "rle": {"size": [480, 640], "counts": "k^\\52m=2nB1Q=0nB2P=OoB2P=OoB2P=OoB2P=OoB2P=OoB1Q=0oB0P=1oB0P=0PC1o in this image.", "objects": [{"patches": [13, 14, 15, 34, 35, 36, 37, 38, 39, 40, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114], "bbox": [0.267984375, 0.02022916666666667, 0.9977499999999999, 0.294375], "iscrowd": 0, "area": 32078.96784999999, "rle": {"size": [480, 640], "counts": "hg`21o>0O100O100O101O0O100O101N100O101N100O100O2O0O10001N100O100O2O0O100O2O0O100O101N100O101O0O100O101N100O100O100O100O100O10000O100O100O100O100O100O100O100O100O10000O100O100O101N100O100O100O100O10000O100O100O100O100O100O100O100O100O10000O100O1O100O100O100O100O2O0O1O100O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O101N100O100O100O1O100O100O100O100O100O1O100O100O10000O10000O100O10000O10000O100O10000O2O000O100O10000O100O10000O10000O100O100000000000000000000000000000000000000010O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O0010O01O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O010O001O001O1O001O001O001O001O1O001O001O001O001O1O001O001O001O001O1O001O010O001O001O001O1O001O001O001O001O1O001O001O001O001O1O001O001O001O001O010O1O001O001O001O001O1O001O001O001O001O1O001O001O001O001O1O001O001O001O0010O01O1O001O001O001O001O1O001O001O001O001O1O001O001O001O001O1O001O001O001O0M4M2Mo:"}, "label": "an umbrella in white over a table in an outdoor cafe area"}]} {"id": 258538, "image": "COCO_train2014_000000258538.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white bungalow top with wooden chairs in it\"."}], "task": "refering", "answer_template": "The \"white bungalow top with wooden chairs in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 34, 35, 36, 37, 38, 39, 40, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114], "bbox": [0.267984375, 0.02022916666666667, 0.9977499999999999, 0.294375], "iscrowd": 0, "area": 32078.96784999999, "rle": {"size": [480, 640], "counts": "hg`21o>0O100O100O101O0O100O101N100O101N100O100O2O0O10001N100O100O2O0O100O2O0O100O101N100O101O0O100O101N100O100O100O100O100O10000O100O100O100O100O100O100O100O100O10000O100O100O101N100O100O100O100O10000O100O100O100O100O100O100O100O100O10000O100O1O100O100O100O100O2O0O1O100O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O1O100O101N100O100O100O1O100O100O100O100O100O1O100O100O10000O10000O100O10000O10000O100O10000O2O000O100O10000O100O10000O10000O100O100000000000000000000000000000000000000010O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O1O001O0010O01O1O001O001O1O001O001O1O001O001O1O001O001O001O1O001O001O1O001O001O010O001O001O1O001O001O001O001O1O001O001O001O001O1O001O001O001O001O1O001O010O001O001O001O1O001O001O001O001O1O001O001O001O001O1O001O001O001O001O010O1O001O001O001O001O1O001O001O001O001O1O001O001O001O001O1O001O001O001O0010O01O1O001O001O001O001O1O001O001O001O001O1O001O001O001O001O1O001O001O001O0M4M2Mo:"}, "label": "white bungalow top with wooden chairs in it"}]} {"id": 94702, "image": "COCO_train2014_000000094702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table near a chair\"."}], "task": "refering", "answer_template": "The \"a table near a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 207, 208, 209, 223, 224, 225, 240, 241, 242, 257, 258, 259, 274, 275, 276, 292, 293, 309, 310, 326, 327, 343, 344, 360, 361, 377, 378], "bbox": [0.1433125, 0.526984375, 0.31666666666666665, 0.997296875], "iscrowd": 0, "area": 10911.752, "rle": {"size": [640, 480], "counts": "a^[16jc09G9G9G9G9G8H9G9G9G9G9G9G8H9G5K00001O00001O:F9G:F9G:F9G9G>B>C>A?A?A?A:F00000000O100000000aJhEV1X:gNVFn0j9nNeFf0\\9WORG>n8^OaG7_8FoGOQ8N^HFb76mH_OS7>[IWOe6e0jIoNW6n0WJgNi50lKDT46]L_Oc3;nLZOR3b0]MSOc2m0X8O100O10000O100O100O1O100O1O1O100O1O1O100O1Hfe\\6"}, "label": "a table near a chair"}]} {"id": 94702, "image": "COCO_train2014_000000094702.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden table and empty chair\"."}], "task": "refering", "answer_template": "The \"a wooden table and empty chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 207, 208, 209, 223, 224, 225, 240, 241, 242, 257, 258, 259, 274, 275, 276, 292, 293, 309, 310, 326, 327, 343, 344, 360, 361, 377, 378], "bbox": [0.1433125, 0.526984375, 0.31666666666666665, 0.997296875], "iscrowd": 0, "area": 10911.752, "rle": {"size": [640, 480], "counts": "a^[16jc09G9G9G9G9G8H9G9G9G9G9G9G8H9G5K00001O00001O:F9G:F9G:F9G9G>B>C>A?A?A?A:F00000000O100000000aJhEV1X:gNVFn0j9nNeFf0\\9WORG>n8^OaG7_8FoGOQ8N^HFb76mH_OS7>[IWOe6e0jIoNW6n0WJgNi50lKDT46]L_Oc3;nLZOR3b0]MSOc2m0X8O100O10000O100O100O1O100O1O1O100O1O1O100O1Hfe\\6"}, "label": "a wooden table and empty chair"}]} {"id": 479734, "image": "COCO_train2014_000000479734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in seashore\"."}], "task": "refering", "answer_template": "The \"man in seashore\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 120, 121, 122, 123, 142, 143, 144, 145, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 284, 285], "bbox": [0.19553125, 0.320327868852459, 0.47315625, 0.8303981264637004], "iscrowd": 0, "area": 21304.18214999999, "rle": {"size": [427, 640], "counts": "U\\d15g in this image.", "objects": [{"patches": [99, 100, 120, 121, 122, 123, 142, 143, 144, 145, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 234, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 284, 285], "bbox": [0.19553125, 0.320327868852459, 0.47315625, 0.8303981264637004], "iscrowd": 0, "area": 21304.18214999999, "rle": {"size": [427, 640], "counts": "U\\d15g in this image.", "objects": [{"patches": [164, 165, 166, 178, 181, 182, 183, 184, 185, 195, 196, 197, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 229, 230, 232, 233, 234, 235], "bbox": [0.49525104602510456, 0.4215, 0.9012552301255229, 0.6], "iscrowd": 0, "area": 13887.018000000002, "rle": {"size": [640, 478], "counts": "U]d4P1Pc0`0@2M3N3M2N2;FC>A?B=C=O1N1O8I7H8H7JVRm0"}, "label": "back of a sofa with a white blanket draped over it"}]} {"id": 520696, "image": "COCO_train2014_000000520696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown sofa with a towel draped over it\"."}], "task": "refering", "answer_template": "The \"a brown sofa with a towel draped over it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 166, 178, 181, 182, 183, 184, 185, 195, 196, 197, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 229, 230, 232, 233, 234, 235], "bbox": [0.49525104602510456, 0.4215, 0.9012552301255229, 0.6], "iscrowd": 0, "area": 13887.018000000002, "rle": {"size": [640, 478], "counts": "U]d4P1Pc0`0@2M3N3M2N2;FC>A?B=C=O1N1O8I7H8H7JVRm0"}, "label": "a brown sofa with a towel draped over it"}]} {"id": 61951, "image": "COCO_train2014_000000061951.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red moterscooter\"."}], "task": "refering", "answer_template": "The \"a red moterscooter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 315, 316, 317, 318, 338, 339, 340, 341, 342, 361, 362, 363, 364, 365, 384, 385, 386, 387], "bbox": [0.6683749999999999, 0.6407563025210085, 0.886140625, 0.989873949579832], "iscrowd": 0, "area": 14385.657950000006, "rle": {"size": [476, 640], "counts": "QXW63j>4K4L2O1N2N2O1N2N010eB\\O\\[CDd<<[CFd<;ZCFg<:WCHh<8WCJh<7UCLk<4TCLl<4SCNl<3RCOn<1QC0n<0QC2n in this image.", "objects": [{"patches": [268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 296, 315, 316, 317, 318, 338, 339, 340, 341, 342, 361, 362, 363, 364, 365, 384, 385, 386, 387], "bbox": [0.6683749999999999, 0.6407563025210085, 0.886140625, 0.989873949579832], "iscrowd": 0, "area": 14385.657950000006, "rle": {"size": [476, 640], "counts": "QXW63j>4K4L2O1N2N2O1N2N010eB\\O\\[CDd<<[CFd<;ZCFg<:WCHh<8WCJh<7UCLk<4TCLl<4SCNl<3RCOn<1QC0n<0QC2n in this image.", "objects": [{"patches": [126, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223], "bbox": [0.47228125, 0.33864168618266977, 0.907078125, 0.6262763466042154], "iscrowd": 0, "area": 23556.528399999996, "rle": {"size": [427, 640], "counts": "aSn31Y=3L4L4M3L3N3L4L4M3L3M4M3L4L4M2M4M3L4N1O1O1N2O1O1O001O1N2O1O1O001N2O1O1O1O0O2O1O1O1O1O0O2O1O1O1O1N101O1O1O1O1N101O1O1O1O000000O10000000000O1001O0000001O00001O00001O00001O0000001O00001O00001O00001O0000001O000O2O00001O00001O00001O0000001O000O2O00001O000000000000O1000000O10001O0O10001O0O10001O0O10001O0O10001O0O10001O0O10001O0O10001O000O101O000O101O000O101O000O2O00000O2O00000O2O00000O2O000O101O0O101O000O2O001N101O0O2O000O2O001O0O2O001N10001N101O001N10001N101O0O2O001O0O101O0O2O001N101O0O101O001N101O0O101O1N2O1O1O1N2O1O2M2O1O1N2O1O1N2O1O1O2M2N2M3M3N2M3M3M3Ne\\h0"}, "label": "a pizza closest to the hanging utencils"}]} {"id": 135694, "image": "COCO_train2014_000000135694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza in the right\"."}], "task": "refering", "answer_template": "The \"the pizza in the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 130, 131, 132, 133, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 221, 222, 223], "bbox": [0.47228125, 0.33864168618266977, 0.907078125, 0.6262763466042154], "iscrowd": 0, "area": 23556.528399999996, "rle": {"size": [427, 640], "counts": "aSn31Y=3L4L4M3L3N3L4L4M3L3M4M3L4L4M2M4M3L4N1O1O1N2O1O1O001O1N2O1O1O001N2O1O1O1O0O2O1O1O1O1O0O2O1O1O1O1N101O1O1O1O1N101O1O1O1O000000O10000000000O1001O0000001O00001O00001O00001O0000001O00001O00001O00001O0000001O000O2O00001O00001O00001O0000001O000O2O00001O000000000000O1000000O10001O0O10001O0O10001O0O10001O0O10001O0O10001O0O10001O0O10001O000O101O000O101O000O101O000O2O00000O2O00000O2O00000O2O000O101O0O101O000O2O001N101O0O2O000O2O001O0O2O001N10001N101O001N10001N101O0O2O001O0O101O0O2O001N101O0O101O001N101O0O101O1N2O1O1O1N2O1O2M2O1O1N2O1O1N2O1O1O2M2N2M3M3N2M3M3M3Ne\\h0"}, "label": "the pizza in the right"}]} {"id": 135694, "image": "COCO_train2014_000000135694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza to the far left\"."}], "task": "refering", "answer_template": "The \"the pizza to the far left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284], "bbox": [0.048671874999999996, 0.4701873536299766, 0.543171875, 0.8396018735362998], "iscrowd": 0, "area": 36743.888399999996, "rle": {"size": [427, 640], "counts": "TV=2R=8H8I7H7H9H8J6N2O001O1N2O001O1N2O001O1N2O1O001N2O1O1N2O1O2N1N3N1O1O2M2O2N1O1N3N1O2N1N2O2N1N3N1O2N1N2O2N1O2M2O1O2N1N3N1O1O2M2O2N1O1N101O0O1000000O1000000O101O000O10000000000000001O0O1000000000000000001O00000000001O000000001O000O10001O000000001O00000000001O000000001O000000001N0100000000O1000000O1000000O100000000O1000000O100000001N1000000O1000000O100000000O101O000O100000000O1000000O10001O000O1000000O1000000O100000001N1000000O10001O000O2O00001N1000001N10001O000O2O00000O2O00001O0O101O001N101O1O0O2O1O001O1N1O2O0O2N2N101N2N1O2O1N1O2N2O0O2N2N101N2N1O2N2O0O2N2N101N1O2N2O0O2N2N101N2N1O2N2O0O2N1O2O1N1O2N2N101N2N1O2N2N101N4L5K5K4M4K5KQmi3"}, "label": "the pizza to the far left"}]} {"id": 135694, "image": "COCO_train2014_000000135694.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza that has olives on it on the left burner\"."}], "task": "refering", "answer_template": "The \"the pizza that has olives on it on the left burner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284], "bbox": [0.048671874999999996, 0.4701873536299766, 0.543171875, 0.8396018735362998], "iscrowd": 0, "area": 36743.888399999996, "rle": {"size": [427, 640], "counts": "TV=2R=8H8I7H7H9H8J6N2O001O1N2O001O1N2O001O1N2O1O001N2O1O1N2O1O2N1N3N1O1O2M2O2N1O1N3N1O2N1N2O2N1N3N1O2N1N2O2N1O2M2O1O2N1N3N1O1O2M2O2N1O1N101O0O1000000O1000000O101O000O10000000000000001O0O1000000000000000001O00000000001O000000001O000O10001O000000001O00000000001O000000001O000000001N0100000000O1000000O1000000O100000000O1000000O100000001N1000000O1000000O100000000O101O000O100000000O1000000O10001O000O1000000O1000000O100000001N1000000O10001O000O2O00001N1000001N10001O000O2O00000O2O00001O0O101O001N101O1O0O2O1O001O1N1O2O0O2N2N101N2N1O2O1N1O2N2O0O2N2N101N2N1O2N2O0O2N2N101N1O2N2O0O2N2N101N2N1O2N2O0O2N1O2O1N1O2N2N101N2N1O2N2N101N4L5K5K4M4K5KQmi3"}, "label": "the pizza that has olives on it on the left burner"}]} {"id": 250387, "image": "COCO_train2014_000000250387.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra eating some grass next to an antelope\"."}], "task": "refering", "answer_template": "The \"zebra eating some grass next to an antelope\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 193, 194, 210, 211, 212, 213, 216, 217, 235, 240], "bbox": [0.145015625, 0.33023419203747073, 0.476875, 0.6829976580796253], "iscrowd": 0, "area": 18853.93404999999, "rle": {"size": [427, 640], "counts": "`QW15P=7H8cN^OiEi0U:\\O`El0^:YOWEo0g:VOnDR1o:f0O100O1O1O1O2N1O1N2N2N2N2N2N2N2N2O0OM4O010O100O100O100O100O1O010O1O1AcEQN]:n1kEjMW:T2`0N3N1O2M2010O01O01aFVN]7i1bHYN]7h1bHYN^7f1bH[N]7e1cH\\N]7d1bH]N]7c1cH^N]7b1bH_N]7a1aHbN_7]1`HeN_7\\1^HgNb7Z1ZHhNg7X1WHhNl7W1QHkNP8V1cGWNWOc0X9V1^GWOa8k0[GWOf8i0XGYOg8i0UGYOl8i0oFYOQ9n0fFUOZ9]21\\Oc001N101N101N1O2O0O2OlFUMi7S3mGQMR8Y3aGkL]8^3XGdLi8n3001O1O1O2O0000000SNhGgNX8W1VH]Nk7b1o10000000010O0000000001O00000000000000001O0000000000000000000000000000000O100000000O10000000000O1000ORGfMm6Y2SIiMl6V2TIkMk6U2UIlMk6S2VImMj6S2UInMk6U2oHnMQ7V2hHmMW7W2bHmM^7V2]HkMd7Y2VHiMj7Z2QHgMP8]2jGdMV8`2eGbM[8b2_G`Ma8d2YG^Mg8e2TG]Ml8a3O10M3I7I7N1O2N2N2O2M2N2N3jN`F]Nb9f1dFoM^9T2eFeM]9\\2h03M3N2M2N3N2M3M8dFQMm7V400001O001O00001O0nMXG_OLVNn8V2\\G@i8:]GfND2P9T1bG_NK7f8T1eG[N2;\\8U1lHeNV7V1RIdNP7W1f2K5J5L5Ko`[4"}, "label": "zebra eating some grass next to an antelope"}]} {"id": 250387, "image": "COCO_train2014_000000250387.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra grazing in a field\"."}], "task": "refering", "answer_template": "The \"a zebra grazing in a field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171, 187, 188, 189, 190, 193, 194, 210, 211, 212, 213, 216, 217, 235, 240], "bbox": [0.145015625, 0.33023419203747073, 0.476875, 0.6829976580796253], "iscrowd": 0, "area": 18853.93404999999, "rle": {"size": [427, 640], "counts": "`QW15P=7H8cN^OiEi0U:\\O`El0^:YOWEo0g:VOnDR1o:f0O100O1O1O1O2N1O1N2N2N2N2N2N2N2N2O0OM4O010O100O100O100O100O1O010O1O1AcEQN]:n1kEjMW:T2`0N3N1O2M2010O01O01aFVN]7i1bHYN]7h1bHYN^7f1bH[N]7e1cH\\N]7d1bH]N]7c1cH^N]7b1bH_N]7a1aHbN_7]1`HeN_7\\1^HgNb7Z1ZHhNg7X1WHhNl7W1QHkNP8V1cGWNWOc0X9V1^GWOa8k0[GWOf8i0XGYOg8i0UGYOl8i0oFYOQ9n0fFUOZ9]21\\Oc001N101N101N1O2O0O2OlFUMi7S3mGQMR8Y3aGkL]8^3XGdLi8n3001O1O1O2O0000000SNhGgNX8W1VH]Nk7b1o10000000010O0000000001O00000000000000001O0000000000000000000000000000000O100000000O10000000000O1000ORGfMm6Y2SIiMl6V2TIkMk6U2UIlMk6S2VImMj6S2UInMk6U2oHnMQ7V2hHmMW7W2bHmM^7V2]HkMd7Y2VHiMj7Z2QHgMP8]2jGdMV8`2eGbM[8b2_G`Ma8d2YG^Mg8e2TG]Ml8a3O10M3I7I7N1O2N2N2O2M2N2N3jN`F]Nb9f1dFoM^9T2eFeM]9\\2h03M3N2M2N3N2M3M8dFQMm7V400001O001O00001O0nMXG_OLVNn8V2\\G@i8:]GfND2P9T1bG_NK7f8T1eG[N2;\\8U1lHeNV7V1RIdNP7W1f2K5J5L5Ko`[4"}, "label": "a zebra grazing in a field"}]} {"id": 12824, "image": "COCO_train2014_000000012824.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sheep is running in the forest\"."}], "task": "refering", "answer_template": "The \"sheep is running in the forest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 172, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 263], "bbox": [0.44384375, 0.4411943793911007, 0.8060625, 0.7544964871194378], "iscrowd": 0, "area": 16430.233500000002, "rle": {"size": [427, 640], "counts": "egf31W=3L3N3M3M2N3M2N3M3M2N3M3M2N3M2N3M2N3SO`NlEb1T:fNbE]1\\:lNZEW1e:g001O001N101O001O0O2O001O001N2O00106J00000O010000N2O1O1O1O1O0O2O1O1O1O1O1N2O1O001O100000000001O0O10000000000000001N10000000000000000O101O0000000000000O10001O000000000O10000000001OO100O10O10O100O100O10000O010O100O100O10000O10O0100O10000O100O100O01000O100O10000O1000O10O10000O1000000O01000O1000000O100000O010000O1000000O1000O0100O1O1N2N2N2O1N2N1O2O1N2O1O1O100O1OO2M2O1O22N100O10001O000000001O0000000O2L3N2RORD in this image.", "objects": [{"patches": [155, 172, 173, 174, 175, 176, 177, 178, 179, 194, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 263], "bbox": [0.44384375, 0.4411943793911007, 0.8060625, 0.7544964871194378], "iscrowd": 0, "area": 16430.233500000002, "rle": {"size": [427, 640], "counts": "egf31W=3L3N3M3M2N3M2N3M3M2N3M3M2N3M2N3M2N3SO`NlEb1T:fNbE]1\\:lNZEW1e:g001O001N101O001O0O2O001O001N2O00106J00000O010000N2O1O1O1O1O0O2O1O1O1O1O1N2O1O001O100000000001O0O10000000000000001N10000000000000000O101O0000000000000O10001O000000000O10000000001OO100O10O10O100O100O10000O010O100O100O10000O10O0100O10000O100O100O01000O100O10000O1000O10O10000O1000000O01000O1000000O100000O010000O1000000O1000O0100O1O1N2N2N2O1N2N1O2O1N2O1O1O100O1OO2M2O1O22N100O10001O000000001O0000000O2L3N2RORD in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 29, 30, 31, 32, 33, 34, 35, 36, 52, 53, 54, 55, 56, 57, 58, 59, 60, 75, 76, 77, 78, 79, 98, 99, 100, 101, 121, 122, 123, 124], "bbox": [0.267984375, 0.02022916666666667, 0.638765625, 0.32583333333333336], "iscrowd": 0, "area": 20004.645, "rle": {"size": [480, 640], "counts": "ad`2^2b<8H8H7I8G9H8H8H5K0000000O10000000000000000O100000000000000O1000000000000O10000000000O10000000000000000O1000O10000000000000000000000001O000000000000000000000000000000000000001O00000000000001O0000003M4LMhKbET4_:PL`El3`:YL_Ec3a:bL^EZ3c:iL]ES3c:c02TNl1K6J5K5K6K41O0001O01O01O00001O0000001O000010O0001O0000001O00001O0001O01O00001O00001O01O0001O00001O000010O000001O00001O00010O0000001O00001O00001O00001O0000001O00001O001O001O001O010O001O0010O01O0010O01O001O01O0001N10000O2O000O2O001N2O1O1N101N2O1O1N2O2N1N3N1Oje\\3"}, "label": "white bus on a road"}]} {"id": 569889, "image": "COCO_train2014_000000569889.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white hj buss bus\"."}], "task": "refering", "answer_template": "The \"white hj buss bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 29, 30, 31, 32, 33, 34, 35, 36, 52, 53, 54, 55, 56, 57, 58, 59, 60, 75, 76, 77, 78, 79, 98, 99, 100, 101, 121, 122, 123, 124], "bbox": [0.267984375, 0.02022916666666667, 0.638765625, 0.32583333333333336], "iscrowd": 0, "area": 20004.645, "rle": {"size": [480, 640], "counts": "ad`2^2b<8H8H7I8G9H8H8H5K0000000O10000000000000000O100000000000000O1000000000000O10000000000O10000000000000000O1000O10000000000000000000000001O000000000000000000000000000000000000001O00000000000001O0000003M4LMhKbET4_:PL`El3`:YL_Ec3a:bL^EZ3c:iL]ES3c:c02TNl1K6J5K5K6K41O0001O01O01O00001O0000001O000010O0001O0000001O00001O0001O01O00001O00001O01O0001O00001O000010O000001O00001O00010O0000001O00001O00001O00001O0000001O00001O001O001O001O010O001O0010O01O0010O01O001O01O0001N10000O2O000O2O001N2O1O1N101N2O1O1N2O2N1N3N1Oje\\3"}, "label": "white hj buss bus"}]} {"id": 512561, "image": "COCO_train2014_000000512561.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white shirt and a lanyard , holding a wii remote control\"."}], "task": "refering", "answer_template": "The \"a man in a white shirt and a lanyard , holding a wii remote control\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 72, 84, 85, 86, 87, 98, 99, 100, 101, 102, 113, 114, 115, 116, 117, 129, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 159, 160, 161, 162, 163, 164, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 204, 205, 206, 207, 208, 209, 218, 219, 220, 221, 222, 223, 224, 232, 233, 234, 235, 236, 237, 238, 239, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 274, 275, 276, 278, 279, 280, 281, 282, 283, 284, 289, 290, 293, 294, 295, 296, 297, 298, 299, 309, 310, 311, 312, 313, 314], "bbox": [0.2436470588235294, 0.18426562500000002, 1.0, 0.8764062500000002], "iscrowd": 0, "area": 73438.45420000001, "rle": {"size": [640, 425], "counts": "m^Q21oc02M2O1N2O2N0O2O1O0O2O1N101O0e]O@Ua0a0g^OCYa0=c^OG\\a09c^OJ\\a06c^OK]a05a^OM^a03b^OO]a01a^O1_a0O`^O3^a0Ma^O5_a0K`^O6_a0K_^O8`a0G`^O:`a0F^^O_5R2jJQOm6eNeH in this image.", "objects": [{"patches": [69, 70, 71, 72, 84, 85, 86, 87, 98, 99, 100, 101, 102, 113, 114, 115, 116, 117, 129, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 159, 160, 161, 162, 163, 164, 174, 175, 176, 177, 178, 179, 189, 190, 191, 192, 193, 194, 204, 205, 206, 207, 208, 209, 218, 219, 220, 221, 222, 223, 224, 232, 233, 234, 235, 236, 237, 238, 239, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 274, 275, 276, 278, 279, 280, 281, 282, 283, 284, 289, 290, 293, 294, 295, 296, 297, 298, 299, 309, 310, 311, 312, 313, 314], "bbox": [0.2436470588235294, 0.18426562500000002, 1.0, 0.8764062500000002], "iscrowd": 0, "area": 73438.45420000001, "rle": {"size": [640, 425], "counts": "m^Q21oc02M2O1N2O2N0O2O1O0O2O1N101O0e]O@Ua0a0g^OCYa0=c^OG\\a09c^OJ\\a06c^OK]a05a^OM^a03b^OO]a01a^O1_a0O`^O3^a0Ma^O5_a0K`^O6_a0K_^O8`a0G`^O:`a0F^^O_5R2jJQOm6eNeH in this image.", "objects": [{"patches": [92, 93, 94, 107, 108, 109, 122, 123, 124, 137, 138, 139, 140, 153, 154, 155, 170], "bbox": [0.1395764705882353, 0.257859375, 0.40550588235294116, 0.49640625], "iscrowd": 0, "area": 8439.756699999998, "rle": {"size": [640, 425], "counts": "hgU12lc03N2N1TO3c]O0la0[1B>M2M4M3M3M2N3N2O0O2O1N2O0O2O1O1N101N2O1N2O0O2O1N2O10O01O100O1O010O100O1O010O100O1O010O010O00100O0010O010O01O10O010O0WNi_OJV`04Q@GP`06V@Gi?7Z@If?5[@Ke?3^@Ma?1a@O`?Nb@3c?E_@;h?\\OZ@e0l?ROW@m0P`0kNQ@V1U`0aNm_O`1Ya0001O010O001O11N2O1N2O1N2O1N2O1O1N2O0O2O1N2O1O1N2O1N2O1M3M3N2M3M3N2M3M[jm4"}, "label": "a man holding a steering of a video game sitting next to 3 other men holding steering of video games"}]} {"id": 512561, "image": "COCO_train2014_000000512561.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man all the way in the back with the blue shirt\"."}], "task": "refering", "answer_template": "The \"the man all the way in the back with the blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 107, 108, 109, 122, 123, 124, 137, 138, 139, 140, 153, 154, 155, 170], "bbox": [0.1395764705882353, 0.257859375, 0.40550588235294116, 0.49640625], "iscrowd": 0, "area": 8439.756699999998, "rle": {"size": [640, 425], "counts": "hgU12lc03N2N1TO3c]O0la0[1B>M2M4M3M3M2N3N2O0O2O1N2O0O2O1O1N101N2O1N2O0O2O1N2O10O01O100O1O010O100O1O010O100O1O010O010O00100O0010O010O01O10O010O0WNi_OJV`04Q@GP`06V@Gi?7Z@If?5[@Ke?3^@Ma?1a@O`?Nb@3c?E_@;h?\\OZ@e0l?ROW@m0P`0kNQ@V1U`0aNm_O`1Ya0001O010O001O11N2O1N2O1N2O1N2O1O1N2O0O2O1N2O1O1N2O1N2O1M3M3N2M3M3N2M3M[jm4"}, "label": "the man all the way in the back with the blue shirt"}]} {"id": 512561, "image": "COCO_train2014_000000512561.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the white shirt with the wrist watch\"."}], "task": "refering", "answer_template": "The \"the man in the white shirt with the wrist watch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 112, 113, 127, 128, 129, 137, 138, 142, 143, 144, 152, 153, 154, 156, 157, 158, 159, 168, 169, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 201, 202, 203, 204, 215, 216, 217, 218, 230, 231, 232, 233, 245, 246], "bbox": [0.15329411764705883, 0.28700000000000003, 0.6623294117647059, 0.7175], "iscrowd": 0, "area": 21495.90615, "rle": {"size": [640, 425], "counts": "_lX15ic04L3N3M2N3M2N3M3O2N1N2O0O10O10O01000O01000O01000O01000O01000O0100O010O10O0100O010O010O10O0100O010O10O0100O010O10O0100O010O10O0100O010O010O10O10O10S@WOak@mAa0R>AlA`0R>ClA=U>CjA>V>ChA=Y>DeA=[>DcA=\\>EaA<_>F_A;`>F_A;`>G^A9b>I[A9d>IYA8g>JVA8j>ISA9l>HQA:o>Hn@:Q?Jj@8U?Oc@2]?Y2O001O1O001O1O010O001O1O001O1O001O0O2O1O010O1O010O00100O010O1O010O0dMPLjDP4U;]L^Dd3a;hLTDW3l;nLoCS3P in this image.", "objects": [{"patches": [221, 222, 241, 242, 245, 246, 247, 248, 249, 258, 259, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 277, 278, 279, 280, 281, 282, 285, 286, 287, 288, 289, 290, 291, 292, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0, 0.5719375, 0.84021875, 0.9893958333333333], "iscrowd": 0, "area": 49213.97310000002, "rle": {"size": [480, 640], "counts": "j;k0U>U1kN1O0O2O001N101O0O2O001O0O2O001N101O0O2O1O0O2O001O0O2O001N101O0O2O001O0O2O001N101O0O2O001O1N101O0O2O001N101O001N101O0O2O000O1000000O10000O10000O101O000O10000O10000O1000000O10000O2O000O1000000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000QMkES1U:fNRFZ1m9cNWF]1i9bNYF]1g9bNZF^1e9bN\\F^1d9aN]F_1c9`N_F_1`9aNaF_1_9`NbF`1_9^NcFa1`9ZNbFf1a9VN`Fj1c9RN_Fm1c9PN^FP2e9lM\\FT2g9hM[FW2h9eMYF[2j9aMWF_2k9^MUFc2T;000O10000O10000O10000O10000O10000O10000O10O010O10O010O0100O010O010O1O010O010O10O010O010O010O010O010O010O0\\OWMREi2n:YMQEg2n:ZMQEf2P;ZMPEf2o:[MPEe2P;]MoDc2P;^MoDb2Q;_MoDa2P;aMnD_2R;eMiD]2W;eMfD[2Z;f0O001N1UOfLPF\\3Q:gLjEY3X:jLbEW3_:e01N2O0O2O0O2O0O101N100O101N100O1O010O100O010O10O0100O010O100O010O10O01O100O10O0100O100O010O01O001O001O0O110O001O001O001O001N10QMTEl1k:RNYEl1f:RN`El1^:TNeEj1\\:UNgEi1[:TNgEj1[:TNhEi1Z:UNhEj1Y:TNhEk1Z:SNgEm1Z:QNgEn1\\:oMeEQ2\\:mMeER2]:lMcEU2^:iMcEV2`:gMaEY2c;O10O0100O100O1O010O100O10O0100O10O0100O10O0100O100O1O10MPNVCo1m<12M3N2M3N2M3N1N2O1N101N2O0O2N2O0O2O1N101N2O0O2O1NhMVO\\Fj0d9VO\\Fi0e9WOYFk0f9WOWFj0j9VO[EO@l0U;UOWE=XO?a;UOVEf0nN6m;TOTEh1l:XNSEh1n:YNPEh1o:YNQEf1P;[NnDf1R;[NlDe1U;\\NjDd1V;\\NiDd1X;]NfDd1[;\\NdDc1];^NaDc1_;^N_Db1c;aNXD`1h;k0101N1O100O2O0O101N100O101N100O101N100O101N1O100O2O0O100O101N100O100O101N100O1O100O100O2O0O100O100O101N100O100O1O100O2O0O100O100O100O2O0O100O100O101N1O100O100O100O2O0O100O100O101N100O1O100O100O2O0O100O100O100O2O0O100O1O100O2O0O100Oi`_1"}, "label": "table with flowers on it"}]} {"id": 356916, "image": "COCO_train2014_000000356916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white table the man is standing next to\"."}], "task": "refering", "answer_template": "The \"the white table the man is standing next to\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 241, 242, 245, 246, 247, 248, 249, 258, 259, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 277, 278, 279, 280, 281, 282, 285, 286, 287, 288, 289, 290, 291, 292, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.0, 0.5719375, 0.84021875, 0.9893958333333333], "iscrowd": 0, "area": 49213.97310000002, "rle": {"size": [480, 640], "counts": "j;k0U>U1kN1O0O2O001N101O0O2O001O0O2O001N101O0O2O1O0O2O001O0O2O001N101O0O2O001O0O2O001N101O0O2O001O1N101O0O2O001N101O001N101O0O2O000O1000000O10000O10000O101O000O10000O10000O1000000O10000O2O000O1000000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000QMkES1U:fNRFZ1m9cNWF]1i9bNYF]1g9bNZF^1e9bN\\F^1d9aN]F_1c9`N_F_1`9aNaF_1_9`NbF`1_9^NcFa1`9ZNbFf1a9VN`Fj1c9RN_Fm1c9PN^FP2e9lM\\FT2g9hM[FW2h9eMYF[2j9aMWF_2k9^MUFc2T;000O10000O10000O10000O10000O10000O10000O10O010O10O010O0100O010O010O1O010O010O10O010O010O010O010O010O010O0\\OWMREi2n:YMQEg2n:ZMQEf2P;ZMPEf2o:[MPEe2P;]MoDc2P;^MoDb2Q;_MoDa2P;aMnD_2R;eMiD]2W;eMfD[2Z;f0O001N1UOfLPF\\3Q:gLjEY3X:jLbEW3_:e01N2O0O2O0O2O0O101N100O101N100O1O010O100O010O10O0100O010O100O010O10O01O100O10O0100O100O010O01O001O001O0O110O001O001O001O001N10QMTEl1k:RNYEl1f:RN`El1^:TNeEj1\\:UNgEi1[:TNgEj1[:TNhEi1Z:UNhEj1Y:TNhEk1Z:SNgEm1Z:QNgEn1\\:oMeEQ2\\:mMeER2]:lMcEU2^:iMcEV2`:gMaEY2c;O10O0100O100O1O010O100O10O0100O10O0100O10O0100O100O1O10MPNVCo1m<12M3N2M3N2M3N1N2O1N101N2O0O2N2O0O2O1N101N2O0O2O1NhMVO\\Fj0d9VO\\Fi0e9WOYFk0f9WOWFj0j9VO[EO@l0U;UOWE=XO?a;UOVEf0nN6m;TOTEh1l:XNSEh1n:YNPEh1o:YNQEf1P;[NnDf1R;[NlDe1U;\\NjDd1V;\\NiDd1X;]NfDd1[;\\NdDc1];^NaDc1_;^N_Db1c;aNXD`1h;k0101N1O100O2O0O101N100O101N100O101N100O101N1O100O2O0O100O101N100O100O101N100O1O100O100O2O0O100O100O101N100O100O1O100O2O0O100O100O100O2O0O100O100O101N1O100O100O100O2O0O100O100O101N100O1O100O100O2O0O100O100O100O2O0O100O1O100O2O0O100Oi`_1"}, "label": "the white table the man is standing next to"}]} {"id": 356916, "image": "COCO_train2014_000000356916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman that is standing and leaning over the table\"."}], "task": "refering", "answer_template": "The \"the woman that is standing and leaning over the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 92, 93, 94, 115, 138, 139, 161, 162, 184, 207, 230, 253, 276, 277, 299, 300], "bbox": [0.0033750000000000004, 0.1955, 0.09606250000000001, 0.7797708333333333], "iscrowd": 0, "area": 5581.292100000001, "rle": {"size": [480, 640], "counts": "dQ1S1m=f3YL;lEiMj6R2lI^MT6\\2aJUM^5f2XKjLg4Q3oK_LQ4[3WLcLh3X3^LhLa3S3fLkL[3o2kLQMT3j2SMUMl2f2ZMZMf2`2aM_M^2^2fMbMY2V1]IZOZ4_OY2T1bI[OV4@W2S1hI[OQ4BV2Q1nI[Om3BU2P1TJ]Of3CV2m0XJ_Oc3CT2l0^J_O^3ES2i0eJAX3ER2j0iJ_OU3GR2j0kJ^OS3GQ2l0nJ[OR3Ho1n0QKYOo2Io1o0TKVOn2In1Q1VKVOk2Io1Q1WKUOk2In1R1WKVOj2Ho1R1WKVOk2Gm1T1XKVOj2Eo1T1YKVOi2Do1V1XKWOi2AP2X1WKXOh2_OR2Y1VKXOi2^OQ2Z1WKXOg2]OS2Z1WKYOg2[OR2]1WKYOf2YOT2^1VKYOY9g0hFYOW9f0jFZOV9f0jF[OV9d0jF\\OX9b0iF^OW9a0iF_OX9?iFBX9 in this image.", "objects": [{"patches": [69, 70, 92, 93, 94, 115, 138, 139, 161, 162, 184, 207, 230, 253, 276, 277, 299, 300], "bbox": [0.0033750000000000004, 0.1955, 0.09606250000000001, 0.7797708333333333], "iscrowd": 0, "area": 5581.292100000001, "rle": {"size": [480, 640], "counts": "dQ1S1m=f3YL;lEiMj6R2lI^MT6\\2aJUM^5f2XKjLg4Q3oK_LQ4[3WLcLh3X3^LhLa3S3fLkL[3o2kLQMT3j2SMUMl2f2ZMZMf2`2aM_M^2^2fMbMY2V1]IZOZ4_OY2T1bI[OV4@W2S1hI[OQ4BV2Q1nI[Om3BU2P1TJ]Of3CV2m0XJ_Oc3CT2l0^J_O^3ES2i0eJAX3ER2j0iJ_OU3GR2j0kJ^OS3GQ2l0nJ[OR3Ho1n0QKYOo2Io1o0TKVOn2In1Q1VKVOk2Io1Q1WKUOk2In1R1WKVOj2Ho1R1WKVOk2Gm1T1XKVOj2Eo1T1YKVOi2Do1V1XKWOi2AP2X1WKXOh2_OR2Y1VKXOi2^OQ2Z1WKXOg2]OS2Z1WKYOg2[OR2]1WKYOf2YOT2^1VKYOY9g0hFYOW9f0jFZOV9f0jF[OV9d0jF\\OX9b0iF^OW9a0iF_OX9?iFBX9 in this image.", "objects": [{"patches": [142, 143, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 254, 255, 257, 277], "bbox": [0.022984375, 0.38966666666666666, 0.42115625, 0.7358333333333333], "iscrowd": 0, "area": 13924.149200000005, "rle": {"size": [480, 640], "counts": "nW7;e>0O1000001N10000K501O000000001O000002O0O2N2N00010O_D4n7LPH6Q8ImG9S8GkG;V8EfG>Z8BdG1l8NRG2_8B_EM3NoU]5"}, "label": "table woman in blue sweater is sitting at"}]} {"id": 356916, "image": "COCO_train2014_000000356916.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"table in front of seated woman\"."}], "task": "refering", "answer_template": "The \"table in front of seated woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 254, 255, 257, 277], "bbox": [0.022984375, 0.38966666666666666, 0.42115625, 0.7358333333333333], "iscrowd": 0, "area": 13924.149200000005, "rle": {"size": [480, 640], "counts": "nW7;e>0O1000001N10000K501O000000001O000002O0O2N2N00010O_D4n7LPH6Q8ImG9S8GkG;V8EfG>Z8BdG1l8NRG2_8B_EM3NoU]5"}, "label": "table in front of seated woman"}]} {"id": 340535, "image": "COCO_train2014_000000340535.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a striped shirt and glasses sitting at a table next to a female\"."}], "task": "refering", "answer_template": "The \"a man wearing a striped shirt and glasses sitting at a table next to a female\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 89, 90, 91, 92, 106, 107, 108, 109, 123, 124, 125, 126, 127, 140, 141, 142, 143, 144, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 195, 196, 206, 207, 208, 209, 210, 211, 212, 213, 230], "bbox": [0.160875, 0.193203125, 0.5762083333333334, 0.6018593750000001], "iscrowd": 0, "area": 21062.646200000003, "rle": {"size": [640, 480], "counts": "a^`11mc03M3N2M3M3M3N2M3N2N2N2N2M2O2N2N2N2N101N2O1N101N1O01O001O00O1N2N1O2N2N2M3M3M3M3M4L3M4N2M3N1O2M3N2N1N3N2N1N2\\MlM[CV2a`1QNU<[1_Bc0^1QNVk1oAUNQ>k1PBTNQ>j1PBVNP>j1PBVNP>i1QBVNQ>i1oAWNQ>h1PBXNP>h1QBVNQ>h1PBXNP>g1PBZNP>f1oAZNR>e1oA[NR>d1mA]NS>b1mA^NT>a1lA`NU>]1mAbNT>\\1mAeNS>Y1nAhNS>U1nAkNS>S1nAnNR>P1nAROS>k0nAUOS>i0mAYOS>e0nA\\OS>`0nACQ>;oAHP>6PBMo=1PB2Q>KoA8P>FoA>P>@PBc0P>[OnAi0Q>UOnAo0Q>POnAS1R>kNmAX1R>gNmA[1S>dNmA^1S>_NPBa1o=^NRBc1m=\\NTBe1S`01O1O1O001^^OaNa`0`1o^OoNQa0P20O105K8H7H8I8H4L3L4M3M3M4K4M3J6I7I7Jf`n3"}, "label": "a man wearing a striped shirt and glasses sitting at a table next to a female"}]} {"id": 340535, "image": "COCO_train2014_000000340535.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a blue and white shirt\"."}], "task": "refering", "answer_template": "The \"a man wearing a blue and white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 75, 89, 90, 91, 92, 106, 107, 108, 109, 123, 124, 125, 126, 127, 140, 141, 142, 143, 144, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 195, 196, 206, 207, 208, 209, 210, 211, 212, 213, 230], "bbox": [0.160875, 0.193203125, 0.5762083333333334, 0.6018593750000001], "iscrowd": 0, "area": 21062.646200000003, "rle": {"size": [640, 480], "counts": "a^`11mc03M3N2M3M3M3N2M3N2N2N2N2M2O2N2N2N2N101N2O1N101N1O01O001O00O1N2N1O2N2N2M3M3M3M3M4L3M4N2M3N1O2M3N2N1N3N2N1N2\\MlM[CV2a`1QNU<[1_Bc0^1QNVk1oAUNQ>k1PBTNQ>j1PBVNP>j1PBVNP>i1QBVNQ>i1oAWNQ>h1PBXNP>h1QBVNQ>h1PBXNP>g1PBZNP>f1oAZNR>e1oA[NR>d1mA]NS>b1mA^NT>a1lA`NU>]1mAbNT>\\1mAeNS>Y1nAhNS>U1nAkNS>S1nAnNR>P1nAROS>k0nAUOS>i0mAYOS>e0nA\\OS>`0nACQ>;oAHP>6PBMo=1PB2Q>KoA8P>FoA>P>@PBc0P>[OnAi0Q>UOnAo0Q>POnAS1R>kNmAX1R>gNmA[1S>dNmA^1S>_NPBa1o=^NRBc1m=\\NTBe1S`01O1O1O001^^OaNa`0`1o^OoNQa0P20O105K8H7H8I8H4L3L4M3M3M4K4M3J6I7I7Jf`n3"}, "label": "a man wearing a blue and white shirt"}]} {"id": 340535, "image": "COCO_train2014_000000340535.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man pouring wine in a white shirt\"."}], "task": "refering", "answer_template": "The \"a man pouring wine in a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 46, 47, 48, 49, 63, 64, 65, 66, 79, 80, 81, 82, 83, 84, 95, 96, 97, 98, 99, 100, 101, 112, 113, 114, 117, 118, 128, 129, 130, 131, 134, 135, 144, 145, 146, 147, 148, 151, 152, 160, 161, 162, 163, 164, 165, 166, 168, 169, 176, 177, 178, 179, 186, 195], "bbox": [0.39127083333333335, 0.05653125, 1.0, 0.51171875], "iscrowd": 0, "area": 27574.6005, "rle": {"size": [640, 480], "counts": "]ie35jc00010O01O00100O0JN_\\O3`c07O001O00100O00100O010O10O0100O1O100O100O100O2O0O2N2O1B]O\\]Oe0_b0a0S^O`NUa0d1e^ObNWa0b1c^OdNXa0U2N2N1O10O0000000O2O0G9M3N2M4M2N21O100O1O1O1O1O2N100O1O1O1O1O2N1O1O1O001O0O10N2N2N2N101N2N2000O100000O10O10000000000O1000O10000000L4O1O2N1O101N1N2N2O0O2N2O1O100O1O1000O1O2O1N1O2O1N1O2O1N1O27T_O`Mi?e2R@]Ml?g2k_O`MT`0W3N9G6K0O1O100O100O00100O1O100O00100O1O010O1O100O1O100OmMn@ZOQ?c0XAYOg>d0_AYO`>e0eAYOi=dNfAP2f0ZOc=jNcAl1m0XO[=TOeAb1V1VOkl_OCT`0:n_OGR`03S@Ol?E_@]3000010O000001O000000100O1O001O1O1O3L3N2N2N2N2N2M3M4L3M4L4C\\1dN`J"}, "label": "a man pouring wine in a white shirt"}]} {"id": 340535, "image": "COCO_train2014_000000340535.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with glasses pouring a bottle of wine\"."}], "task": "refering", "answer_template": "The \"man with glasses pouring a bottle of wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 46, 47, 48, 49, 63, 64, 65, 66, 79, 80, 81, 82, 83, 84, 95, 96, 97, 98, 99, 100, 101, 112, 113, 114, 117, 118, 128, 129, 130, 131, 134, 135, 144, 145, 146, 147, 148, 151, 152, 160, 161, 162, 163, 164, 165, 166, 168, 169, 176, 177, 178, 179, 186, 195], "bbox": [0.39127083333333335, 0.05653125, 1.0, 0.51171875], "iscrowd": 0, "area": 27574.6005, "rle": {"size": [640, 480], "counts": "]ie35jc00010O01O00100O0JN_\\O3`c07O001O00100O00100O010O10O0100O1O100O100O100O2O0O2N2O1B]O\\]Oe0_b0a0S^O`NUa0d1e^ObNWa0b1c^OdNXa0U2N2N1O10O0000000O2O0G9M3N2M4M2N21O100O1O1O1O1O2N100O1O1O1O1O2N1O1O1O001O0O10N2N2N2N101N2N2000O100000O10O10000000000O1000O10000000L4O1O2N1O101N1N2N2O0O2N2O1O100O1O1000O1O2O1N1O2O1N1O2O1N1O27T_O`Mi?e2R@]Ml?g2k_O`MT`0W3N9G6K0O1O100O100O00100O1O100O00100O1O010O1O100O1O100OmMn@ZOQ?c0XAYOg>d0_AYO`>e0eAYOi=dNfAP2f0ZOc=jNcAl1m0XO[=TOeAb1V1VOkl_OCT`0:n_OGR`03S@Ol?E_@]3000010O000001O000000100O1O001O1O1O3L3N2N2N2N2N2M3M4L3M4L4C\\1dN`J"}, "label": "man with glasses pouring a bottle of wine"}]} {"id": 487992, "image": "COCO_train2014_000000487992.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man wearing the long black wig with sunglasses on\"."}], "task": "refering", "answer_template": "The \"the man wearing the long black wig with sunglasses on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 23, 35, 36, 37, 38, 48, 49, 50, 51, 52, 53, 63, 64, 65, 66, 67, 68, 78, 79, 80, 81, 82, 83, 92, 93, 94, 95, 96, 97, 98, 107, 108, 109, 110, 111, 112, 113, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 153, 154, 155, 156, 157, 168, 169, 170, 171, 172, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 228, 229, 230, 231, 244, 245, 246, 247, 259, 261, 262, 276], "bbox": [0.16460048426150123, 0.05618181818181818, 0.5745762711864407, 0.9146], "iscrowd": 0, "area": 49355.945900000006, "rle": {"size": [550, 413], "counts": "`gT12Qa03M4L3M4L3M3M4K4M4L3^NmNnBW1P=lNkBW1T=mNfBW1W=mNdBV1[=mN`BV1^=oN\\BU1b=nNYBU1e=POUBT1i=oNRBT1l=U1N1N3N1O1N3N1N3N1hD_Li8d3hFZNg7g1kG6d6MmHP1c6QOoH^1n6_4N2N2N2N2O1N2O1N1O2O1N2`JcEU4^:eKiEX4Y:aKPF\\4Q:]KWF`4k9YK^Fd4c9UKeFh4o:M4M2M3N3L3N2M4M0ON2N3N1N3M2O2M2N3N1N3H7\\Od0\\Od0K4O1N2N2N2N2O1N2N2N2O1N2N2N2O1N3M2NbKaF[1]9^NPG^1o8[N^G`1`8ZNlGb1f8dMfGX2o8kL^GQ3^;M4M2N2N3M2N2N3M2N2M4M2N3M2N1OO100000O10O1000000O10000hM^MlEa2Q:iMiEW2S:SNfEn1V:[NeEe1W:eNcE[1Y:oNaEQ1[:YO^Eh0^:A]E`0^:J\\E6`:2\\EN`:9]EG_:`0]EA_:f0^E[O]:l0`ETOT:[1iEeNd9h4]Oc0]Oc0\\Oe0G8K5L4L4L5K4L4L4K5L4L5K5K6J7I7H8I7I7I7I7I7I7H8I7I6Jm0SOU1kNP]n2"}, "label": "the man wearing the long black wig with sunglasses on"}]} {"id": 487992, "image": "COCO_train2014_000000487992.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a lady holding glass\"."}], "task": "refering", "answer_template": "The \"a man with a lady holding glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 23, 35, 36, 37, 38, 48, 49, 50, 51, 52, 53, 63, 64, 65, 66, 67, 68, 78, 79, 80, 81, 82, 83, 92, 93, 94, 95, 96, 97, 98, 107, 108, 109, 110, 111, 112, 113, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 153, 154, 155, 156, 157, 168, 169, 170, 171, 172, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 228, 229, 230, 231, 244, 245, 246, 247, 259, 261, 262, 276], "bbox": [0.16460048426150123, 0.05618181818181818, 0.5745762711864407, 0.9146], "iscrowd": 0, "area": 49355.945900000006, "rle": {"size": [550, 413], "counts": "`gT12Qa03M4L3M4L3M3M4K4M4L3^NmNnBW1P=lNkBW1T=mNfBW1W=mNdBV1[=mN`BV1^=oN\\BU1b=nNYBU1e=POUBT1i=oNRBT1l=U1N1N3N1O1N3N1N3N1hD_Li8d3hFZNg7g1kG6d6MmHP1c6QOoH^1n6_4N2N2N2N2O1N2O1N1O2O1N2`JcEU4^:eKiEX4Y:aKPF\\4Q:]KWF`4k9YK^Fd4c9UKeFh4o:M4M2M3N3L3N2M4M0ON2N3N1N3M2O2M2N3N1N3H7\\Od0\\Od0K4O1N2N2N2N2O1N2N2N2O1N2N2N2O1N3M2NbKaF[1]9^NPG^1o8[N^G`1`8ZNlGb1f8dMfGX2o8kL^GQ3^;M4M2N2N3M2N2N3M2N2M4M2N3M2N1OO100000O10O1000000O10000hM^MlEa2Q:iMiEW2S:SNfEn1V:[NeEe1W:eNcE[1Y:oNaEQ1[:YO^Eh0^:A]E`0^:J\\E6`:2\\EN`:9]EG_:`0]EA_:f0^E[O]:l0`ETOT:[1iEeNd9h4]Oc0]Oc0\\Oe0G8K5L4L4L5K4L4L4K5L4L5K5K6J7I7H8I7I7I7I7I7I7H8I7I6Jm0SOU1kNP]n2"}, "label": "a man with a lady holding glass"}]} {"id": 487992, "image": "COCO_train2014_000000487992.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl with red tie\"."}], "task": "refering", "answer_template": "The \"the girl with red tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 53, 54, 55, 68, 69, 70, 82, 83, 84, 85, 97, 98, 99, 100, 101, 112, 113, 114, 115, 116, 117, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 147, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 203, 204, 205, 206, 218, 219, 220, 233, 234, 235, 249, 250, 264, 265, 278, 279, 280], "bbox": [0.5022518159806295, 0.12976363636363636, 0.8150847457627118, 0.9408181818181819], "iscrowd": 0, "area": 38143.47230000001, "rle": {"size": [550, 413], "counts": "mZ_3f3^=7I8G8I7H9H7H9Hb0]Ob0_O`0_O3N3L4M3M3J6K5J6K4L5J6K5J6K5L4M201N2O1N2O2M9H9F:G9F:G7nKPE_1X;UNnDj1X;jMnDU2Z;^MlDb2[;RMjDm2^;fLiDY3bC=C=C=BZcX1"}, "label": "the girl with red tie"}]} {"id": 487992, "image": "COCO_train2014_000000487992.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the lady on the right with the red necktie\"."}], "task": "refering", "answer_template": "The \"the lady on the right with the red necktie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 53, 54, 55, 68, 69, 70, 82, 83, 84, 85, 97, 98, 99, 100, 101, 112, 113, 114, 115, 116, 117, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 147, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 187, 188, 189, 190, 191, 203, 204, 205, 206, 218, 219, 220, 233, 234, 235, 249, 250, 264, 265, 278, 279, 280], "bbox": [0.5022518159806295, 0.12976363636363636, 0.8150847457627118, 0.9408181818181819], "iscrowd": 0, "area": 38143.47230000001, "rle": {"size": [550, 413], "counts": "mZ_3f3^=7I8G8I7H9H7H9Hb0]Ob0_O`0_O3N3L4M3M3J6K5J6K4L5J6K5J6K5L4M201N2O1N2O2M9H9F:G9F:G7nKPE_1X;UNnDj1X;jMnDU2Z;^MlDb2[;RMjDm2^;fLiDY3bC=C=C=BZcX1"}, "label": "the lady on the right with the red necktie"}]} {"id": 487992, "image": "COCO_train2014_000000487992.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wood chair in front\"."}], "task": "refering", "answer_template": "The \"the wood chair in front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [228, 242, 243, 244, 257, 258, 271, 272, 273, 285, 286, 287, 288], "bbox": [0.008958837772397095, 0.7665636363636363, 0.2750121065375303, 0.9985818181818182], "iscrowd": 0, "area": 6370.36675, "rle": {"size": [550, 413], "counts": "RW31Ta0100O1O1O100O1O100O001O100O1O100N2O1N2O1N2O1N2O1N2O0O2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2N2N2O1N2N2N3N1N2N2N2O1N2N2O1N2N2N2O1N2N3M2O1N2N2O1N2N2N2O1N000001O00001O00001O2N3M3M3M3N2N101N2N2N2O1N2N1O2N2K5A?B>B=B?BhUP5"}, "label": "the wood chair in front"}]} {"id": 487992, "image": "COCO_train2014_000000487992.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooden thing in the front of the picture\"."}], "task": "refering", "answer_template": "The \"wooden thing in the front of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [228, 242, 243, 244, 257, 258, 271, 272, 273, 285, 286, 287, 288], "bbox": [0.008958837772397095, 0.7665636363636363, 0.2750121065375303, 0.9985818181818182], "iscrowd": 0, "area": 6370.36675, "rle": {"size": [550, 413], "counts": "RW31Ta0100O1O1O100O1O100O001O100O1O100N2O1N2O1N2O1N2O1N2O0O2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2N2N2O1N2N2N3N1N2N2N2O1N2N2O1N2N2N2O1N2N3M2O1N2N2O1N2N2N2O1N000001O00001O00001O2N3M3M3M3N2N101N2N2N2O1N2N1O2N2K5A?B>B=B?BhUP5"}, "label": "wooden thing in the front of the picture"}]} {"id": 160313, "image": "COCO_train2014_000000160313.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red train car on the right sits at the platform\"."}], "task": "refering", "answer_template": "The \"the red train car on the right sits at the platform\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229], "bbox": [0.7966249999999999, 0.3122716627634661, 1.0, 0.6309367681498829], "iscrowd": 0, "area": 14426.733149999993, "rle": {"size": [427, 640], "counts": "eld6i0m:\\OUFS1j9QOfE]1W:iNYEc1f:b0M3N1N2O1O1O10JdE]M[:d2gEZMX:g2iEXMW:h251O10000O10000O2O000O100O100O1O100O1O101N1O100O100O100O100O10001O000O1000000000000O1O1O2N1O10000000000000000000001O000O100O100O100O100O101O0000000O10000000000O101O000O1000000O1000000O101O000O1000000O1000000O101O000O1000000O100000^L^H[1\\3"}, "label": "the red train car on the right sits at the platform"}]} {"id": 160313, "image": "COCO_train2014_000000160313.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red train car on train tracks\"."}], "task": "refering", "answer_template": "The \"red train car on train tracks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [114, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229], "bbox": [0.7966249999999999, 0.3122716627634661, 1.0, 0.6309367681498829], "iscrowd": 0, "area": 14426.733149999993, "rle": {"size": [427, 640], "counts": "eld6i0m:\\OUFS1j9QOfE]1W:iNYEc1f:b0M3N1N2O1O1O10JdE]M[:d2gEZMX:g2iEXMW:h251O10000O10000O2O000O100O100O1O100O1O101N1O100O100O100O100O10001O000O1000000000000O1O1O2N1O10000000000000000000001O000O100O100O100O100O101O0000000O10000000000O101O000O1000000O1000000O101O000O1000000O1000000O101O000O1000000O100000^L^H[1\\3"}, "label": "red train car on train tracks"}]} {"id": 209467, "image": "COCO_train2014_000000209467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black chair underneath a window\"."}], "task": "refering", "answer_template": "The \"a black chair underneath a window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 202, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 267, 268, 269, 270, 291, 292, 293, 315], "bbox": [0.624625, 0.5054583333333333, 0.8428437499999999, 0.7752291666666667], "iscrowd": 0, "area": 9254.755299999997, "rle": {"size": [480, 640], "counts": "nik5?@=E;M3L4N2N2N2N2O1N2O100000000001O0000000000000O10000000001O0000000000000000000OUDPNh:P2gDRNE5e;i1^DhNb;\\2O00001O00001O00001O00001N10000O0O2N2O1N1O2N2O1N2O0O2O1N20O10O1000O1000000O1000000O1000001N1cN]1lNT1O1N3N100O10001N10001O0O1N2O2M2O1O2O0O1O2O0O1O2O01O01O01O01O01O100O101N1O10O01O01O00010O000O2GUS_1"}, "label": "a black chair underneath a window"}]} {"id": 209467, "image": "COCO_train2014_000000209467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair on the right in the right hand picture\"."}], "task": "refering", "answer_template": "The \"the chair on the right in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 202, 221, 222, 223, 224, 225, 226, 244, 245, 246, 247, 267, 268, 269, 270, 291, 292, 293, 315], "bbox": [0.624625, 0.5054583333333333, 0.8428437499999999, 0.7752291666666667], "iscrowd": 0, "area": 9254.755299999997, "rle": {"size": [480, 640], "counts": "nik5?@=E;M3L4N2N2N2N2O1N2O100000000001O0000000000000O10000000001O0000000000000000000OUDPNh:P2gDRNE5e;i1^DhNb;\\2O00001O00001O00001O00001N10000O0O2N2O1N1O2N2O1N2O0O2O1N20O10O1000O1000000O1000000O1000001N1cN]1lNT1O1N3N100O10001N10001O0O1N2O2M2O1O2O0O1O2O0O1O2O01O01O01O01O01O100O101N1O10O01O01O00010O000O2GUS_1"}, "label": "the chair on the right in the right hand picture"}]} {"id": 209467, "image": "COCO_train2014_000000209467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the leather chair is facing the television\"."}], "task": "refering", "answer_template": "The \"the leather chair is facing the television\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [299, 300, 302, 303, 304, 322, 323, 324, 325, 326, 327, 345, 346, 347, 348, 349, 350, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.001484375, 0.7877500000000001, 0.4091875, 1.0], "iscrowd": 0, "area": 17176.264449999988, "rle": {"size": [480, 640], "counts": "_l0a1o;`1001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O0000001O0000O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O1001O000000001O0000001O00001O1O001O1O1O1O1O1O1O1O1O7I8H8H7I8H8H8H7I000000000000001O000000000000001O00000000000000001O0000000000000000001O00000000000000000000O1O1O1O1O1O1O1O100N2O1O1O1O1N2O1O1O1N2O1O1O1O100000000000000000000000000000000001O001O001O001O1O2N2N2N2N6J;E8H7ISg`5"}, "label": "the leather chair is facing the television"}]} {"id": 209467, "image": "COCO_train2014_000000209467.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black leather chair sitting beside a floral couch\"."}], "task": "refering", "answer_template": "The \"a black leather chair sitting beside a floral couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [299, 300, 302, 303, 304, 322, 323, 324, 325, 326, 327, 345, 346, 347, 348, 349, 350, 352, 353, 354, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377], "bbox": [0.001484375, 0.7877500000000001, 0.4091875, 1.0], "iscrowd": 0, "area": 17176.264449999988, "rle": {"size": [480, 640], "counts": "_l0a1o;`1001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O0000001O0000O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O1001O000000001O0000001O00001O1O001O1O1O1O1O1O1O1O1O7I8H8H7I8H8H8H7I000000000000001O000000000000001O00000000000000001O0000000000000000001O00000000000000000000O1O1O1O1O1O1O1O100N2O1O1O1O1N2O1O1O1N2O1O1O1O100000000000000000000000000000000001O001O001O001O1O2N2N2N2N6J;E8H7ISg`5"}, "label": "a black leather chair sitting beside a floral couch"}]} {"id": 266816, "image": "COCO_train2014_000000266816.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bird facing towards the cat in the picture\"."}], "task": "refering", "answer_template": "The \"a bird facing towards the cat in the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 131, 132, 133, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228], "bbox": [0.698859375, 0.3034953703703704, 0.978390625, 0.68125], "iscrowd": 0, "area": 11247.434250000004, "rle": {"size": [432, 640], "counts": "`gl51^=2M3N3M0O1000000O10001N11O01O10OUDFa:;^EH_:9aEI]:6PEDN:n:3PEIO7n:1oDN04o:MmD611n:h0oD[On:h0PEZOn:g1L3M3N1N2N3N1N20000000000000001O00022O4M4K:Da0@M0O2N2L3N3L4L3N3L4M2M4M2M4H8B=O2O1O010O00001O03M4L6K4K4hEWMa9m2VFYMj9V30001O01O000O101@ZFSMg9g2d0K6I6M30001O0000010O01O01O010O010O00100O010O1O010O10O01O010O10O01O1O001O1O0O2O001O1O001O1O001O1O001O002N3N2O1N2NO2OO1O1O1O1O1O1O1O001O1000000O10000YOiC4Y in this image.", "objects": [{"patches": [108, 109, 131, 132, 133, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 182, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228], "bbox": [0.698859375, 0.3034953703703704, 0.978390625, 0.68125], "iscrowd": 0, "area": 11247.434250000004, "rle": {"size": [432, 640], "counts": "`gl51^=2M3N3M0O1000000O10001N11O01O10OUDFa:;^EH_:9aEI]:6PEDN:n:3PEIO7n:1oDN04o:MmD611n:h0oD[On:h0PEZOn:g1L3M3N1N2N3N1N20000000000000001O00022O4M4K:Da0@M0O2N2L3N3L4L3N3L4M2M4M2M4H8B=O2O1O010O00001O03M4L6K4K4hEWMa9m2VFYMj9V30001O01O000O101@ZFSMg9g2d0K6I6M30001O0000010O01O01O010O010O00100O010O1O010O10O01O010O10O01O1O001O1O0O2O001O1O001O1O001O1O001O002N3N2O1N2NO2OO1O1O1O1O1O1O1O001O1000000O10000YOiC4Y in this image.", "objects": [{"patches": [107, 108, 129, 130, 131, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 218, 219, 221, 222], "bbox": [0.46306250000000004, 0.26666666666666666, 0.7475781250000001, 0.6623611111111111], "iscrowd": 0, "area": 12131.012100000004, "rle": {"size": [432, 640], "counts": "_am31^=1O1O1O1O1O1O1O1O1O1O1O1[CET<V:@lE?T:AlE>V:AkE>V:@kE`0V:_OkE`0V:_OkE`0U:@kE?W:@jE=Y:BgE=\\:AeE<^:CbEDC5P;5^EED0S;9[EFDD^;d0PEFX;9iDFY;7jDGX;7R1N2O1N2O1N2O1O1N2O1NoVT2"}, "label": "black bird toward left side"}]} {"id": 266816, "image": "COCO_train2014_000000266816.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bird on the left\"."}], "task": "refering", "answer_template": "The \"bird on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 129, 130, 131, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 177, 194, 195, 196, 197, 198, 199, 218, 219, 221, 222], "bbox": [0.46306250000000004, 0.26666666666666666, 0.7475781250000001, 0.6623611111111111], "iscrowd": 0, "area": 12131.012100000004, "rle": {"size": [432, 640], "counts": "_am31^=1O1O1O1O1O1O1O1O1O1O1O1[CET<V:@lE?T:AlE>V:AkE>V:@kE`0V:_OkE`0V:_OkE`0U:@kE?W:@jE=Y:BgE=\\:AeE<^:CbEDC5P;5^EED0S;9[EFDD^;d0PEFX;9iDFY;7jDGX;7R1N2O1N2O1N2O1O1N2O1NoVT2"}, "label": "bird on the left"}]} {"id": 332385, "image": "COCO_train2014_000000332385.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra in front of two others\"."}], "task": "refering", "answer_template": "The \"a zebra in front of two others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 141, 142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 282, 283, 284, 286, 287, 306, 307, 309, 310, 328, 329, 330, 332, 333, 351, 352, 353], "bbox": [0.16434375, 0.31312500000000004, 0.5264375, 0.9052708333333334], "iscrowd": 0, "area": 33308.202150000005, "rle": {"size": [480, 640], "counts": "l\\a11o>2N2N2M4M2N2N00000O101O000000000O10001O00 in this image.", "objects": [{"patches": [120, 121, 141, 142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 211, 212, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 239, 240, 241, 259, 260, 261, 262, 263, 264, 282, 283, 284, 286, 287, 306, 307, 309, 310, 328, 329, 330, 332, 333, 351, 352, 353], "bbox": [0.16434375, 0.31312500000000004, 0.5264375, 0.9052708333333334], "iscrowd": 0, "area": 33308.202150000005, "rle": {"size": [480, 640], "counts": "l\\a11o>2N2N2M4M2N2N00000O101O000000000O10001O00 in this image.", "objects": [{"patches": [123, 124, 130, 131, 145, 146, 147, 152, 153, 154, 155, 218, 219, 241, 242, 246, 247, 264, 265, 270, 287, 288, 292, 293, 310, 311, 315, 316], "bbox": [0.34045312499999997, 0.32416666666666666, 0.7803437499999999, 0.8140416666666667], "iscrowd": 0, "area": 7354.503550000001, "rle": {"size": [480, 640], "counts": "][V39`>9M2O1N3N1N2N100O100O1O100O100O100000000O1000000000000000000000000000O100O100O100O1O100O100O100O100F:N2O1O1N2O1O2M3N1OjPb00So]O5K5K5M4aN\\O\\Dg0];C]D`0[;L^D7[;3_D0Z;;_DHZ;c0_D@Z;j0`DYOY;[2H8M1101N101O0O2O001\\Oc0G:K4L5J6K5K5K5K5G9fNZ1O100O1O100O1O10d\\f00[cYO1N2O1O1N2O1O1O1N1000001N10000O101O000O10001N10000O101O0O10000O101O0O10000O10000O10000O0101O001O001O00001O001O0O2O0000UG]OZ3b0fL_OY3a0fLAZ3>\\J_ORO5a6<^J@oN4c6<_JBjN4g69`JEeN4j67bJFbN3m66`JK_N1P73bJOZN0T70bJ3WNNV7OcJ5TNMY7NbJ9QNK]7KbJ=nMI_7JcJX2]5gMcJY2^5fMaJ\\2_5cMaJ^2_5aMaJ_2_5aM`J`2a5_M_J`2c5^M^Ja2d5^M\\J`2f5`MYJ`2i5_MWJ`2k5_MUJ_2n5`MQJ`2Q6_MoI`2R6`MnI^2[9fNcCO^ in this image.", "objects": [{"patches": [123, 124, 130, 131, 145, 146, 147, 152, 153, 154, 155, 218, 219, 241, 242, 246, 247, 264, 265, 270, 287, 288, 292, 293, 310, 311, 315, 316], "bbox": [0.34045312499999997, 0.32416666666666666, 0.7803437499999999, 0.8140416666666667], "iscrowd": 0, "area": 7354.503550000001, "rle": {"size": [480, 640], "counts": "][V39`>9M2O1N3N1N2N100O100O1O100O100O100000000O1000000000000000000000000000O100O100O100O1O100O100O100O100F:N2O1O1N2O1O2M3N1OjPb00So]O5K5K5M4aN\\O\\Dg0];C]D`0[;L^D7[;3_D0Z;;_DHZ;c0_D@Z;j0`DYOY;[2H8M1101N101O0O2O001\\Oc0G:K4L5J6K5K5K5K5G9fNZ1O100O1O100O1O10d\\f00[cYO1N2O1O1N2O1O1O1N1000001N10000O101O000O10001N10000O101O0O10000O101O0O10000O10000O10000O0101O001O001O00001O001O0O2O0000UG]OZ3b0fL_OY3a0fLAZ3>\\J_ORO5a6<^J@oN4c6<_JBjN4g69`JEeN4j67bJFbN3m66`JK_N1P73bJOZN0T70bJ3WNNV7OcJ5TNMY7NbJ9QNK]7KbJ=nMI_7JcJX2]5gMcJY2^5fMaJ\\2_5cMaJ^2_5aMaJ_2_5aM`J`2a5_M_J`2c5^M^Ja2d5^M\\J`2f5`MYJ`2i5_MWJ`2k5_MUJ_2n5`MQJ`2Q6_MoI`2R6`MnI^2[9fNcCO^ in this image.", "objects": [{"patches": [124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 248, 249, 250, 266, 267, 272, 273, 289, 290, 295, 296, 311, 312, 313, 318, 319], "bbox": [0.41275000000000006, 0.3111875, 0.914421875, 0.8301875], "iscrowd": 0, "area": 34856.65464999999, "rle": {"size": [480, 640], "counts": "]mk31n>1O2M2O1N3N1N2O2N1O1O2O0O100O2O0O101O0O101N100O2O0O101N100O101N10100O001O1O001O001O1O0010O01O1O001O1O001O001O10O01O1O001O001O2N2N2N2O1N2N2N2N2N2PDUNg:S3O1O1O001O100O1O001O1O1O1O001O1O1YHWLV4h3hKYLY4h3eKXL\\4i3aKYL_4h3^KYLc4h3[KYLf4f3XK[Li4f3TK[Ln4i3lJXLT5P4bJQL`5U4WJmKRNE_7d4WJgKYNG`7i4nIaKaNHb7m4dI]KhNHd7S5ZIVKQOIf7V5RIRKVOJh7[5iHXKY7n4^HTKb7S5VHnJk7S61O001M2M4L3M4LVH_I[7^5\\HYK8^O\\7[5^IhJc6[5VIiJi6[5PIhJQ7[5hHiJW7Z5cHiJ^7Z5[HiJe7[5THhJm7U6001O1N101O001_MXHhMh7h1]IcMd6m1_30001O001O00010O00001O001O00001O01O01O001O00001O000010O01O00001O00001O0010O000000000000000O101O0000000000000O10000000000000000O1000000000000000000O100000O10000000O10000000000O10000000000O10O100O100O010O100O100O010O100O100O100O100O100O11O2N1O1O1O2N2N2N2N2N2]G_Lc5c3ZJ^Lf5d3WJ]Li5f3RJ\\Ln5f3oI[LQ6g3lIZLT6h3hIZLX6h3eIYL[6i3bIXL^6j3^IXLc6i3ZIXLf6j3WIXLh6k3SIWLm6n3mHSLT7Q4fHPLZ7U4_HmKb7V4RHRLo7_51O1O001O1N2M2O2N2M3N1O2M3NSGZJc8Y5hGjJW8S4jHoKU7U4eHnK[7U4`HnK_7U4[HnKe7U4VHlKj7X4PHiKR8Z4iGgKX8\\4bGeK`8\\52O1iKdGg1^8TNhGj1Z8QNkGl1W8PNnGm1T8oMQHo1Q8lMTHQ2n7SNoGk1R8YNlGc1V8`NiGEQ:>mE[OX:i0fEPO_:S1aEdNf:_1\\DhN3Bf;[2b0M3L5Kf0[O\\ai0"}, "label": "a zebra standing in between two zebras"}]} {"id": 332385, "image": "COCO_train2014_000000332385.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"middle of the three zebras\"."}], "task": "refering", "answer_template": "The \"middle of the three zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 127, 128, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 204, 219, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 248, 249, 250, 266, 267, 272, 273, 289, 290, 295, 296, 311, 312, 313, 318, 319], "bbox": [0.41275000000000006, 0.3111875, 0.914421875, 0.8301875], "iscrowd": 0, "area": 34856.65464999999, "rle": {"size": [480, 640], "counts": "]mk31n>1O2M2O1N3N1N2O2N1O1O2O0O100O2O0O101O0O101N100O2O0O101N100O101N10100O001O1O001O001O1O0010O01O1O001O1O001O001O10O01O1O001O001O2N2N2N2O1N2N2N2N2N2PDUNg:S3O1O1O001O100O1O001O1O1O1O001O1O1YHWLV4h3hKYLY4h3eKXL\\4i3aKYL_4h3^KYLc4h3[KYLf4f3XK[Li4f3TK[Ln4i3lJXLT5P4bJQL`5U4WJmKRNE_7d4WJgKYNG`7i4nIaKaNHb7m4dI]KhNHd7S5ZIVKQOIf7V5RIRKVOJh7[5iHXKY7n4^HTKb7S5VHnJk7S61O001M2M4L3M4LVH_I[7^5\\HYK8^O\\7[5^IhJc6[5VIiJi6[5PIhJQ7[5hHiJW7Z5cHiJ^7Z5[HiJe7[5THhJm7U6001O1N101O001_MXHhMh7h1]IcMd6m1_30001O001O00010O00001O001O00001O01O01O001O00001O000010O01O00001O00001O0010O000000000000000O101O0000000000000O10000000000000000O1000000000000000000O100000O10000000O10000000000O10000000000O10O100O100O010O100O100O010O100O100O100O100O100O11O2N1O1O1O2N2N2N2N2N2]G_Lc5c3ZJ^Lf5d3WJ]Li5f3RJ\\Ln5f3oI[LQ6g3lIZLT6h3hIZLX6h3eIYL[6i3bIXL^6j3^IXLc6i3ZIXLf6j3WIXLh6k3SIWLm6n3mHSLT7Q4fHPLZ7U4_HmKb7V4RHRLo7_51O1O001O1N2M2O2N2M3N1O2M3NSGZJc8Y5hGjJW8S4jHoKU7U4eHnK[7U4`HnK_7U4[HnKe7U4VHlKj7X4PHiKR8Z4iGgKX8\\4bGeK`8\\52O1iKdGg1^8TNhGj1Z8QNkGl1W8PNnGm1T8oMQHo1Q8lMTHQ2n7SNoGk1R8YNlGc1V8`NiGEQ:>mE[OX:i0fEPO_:S1aEdNf:_1\\DhN3Bf;[2b0M3L5Kf0[O\\ai0"}, "label": "middle of the three zebras"}]} {"id": 537196, "image": "COCO_train2014_000000537196.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"fingers holding the pizza box up\"."}], "task": "refering", "answer_template": "The \"fingers holding the pizza box up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [340, 341, 342, 343, 344, 345, 357, 358, 359, 360, 361, 363, 364, 365, 366, 367, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.0021458333333333334, 0.867375, 0.6615624999999999, 0.9987343750000001], "iscrowd": 0, "area": 12896.622850000009, "rle": {"size": [640, 480], "counts": "jV1U1fa0U10000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000nN^^OFba09`^OF`a09b^OF^a09c^OG]a08e^OG[a09f^OFZa09h^OFXa09i^OGVa09l^OFTa09n^OFRa09P_OFPa09R_OFn`09S_OGm`08U_OGk`08W_OGi`08Y_OGg`09Z_OFf`09[_OGe`08]_OGc`08__OGa`08a_OG_`08c_OG]`08d_OH\\`07f_OHZ`07h_OHX`07j_OHV`08k_OGU`08l_OGU`08m_OFT`09m_OFT`09m_OGS`08n_OGS`09m_OFT`09m_OGS`08m_OHT`08l_OGU`08l_OHT`07m_OHT`08l_OGT`09m_OGS`08m_OHT`08l_OGU`08l_OHT`07m_OHT`08j1O1OR^OJT`06l_OJT`05m_OKS`04m1O100OQ^ONR`01n_O0R`00m_O1R`0On_O2R`0Mo_O3P`0No_O2R`0Mn_O4Q`0Lo_O5Q`0Kn_O6Q`0Jo_O7la01TNHo_O9ka0001O100O1O1O1O001O1O1O1O1O1O010O1O1O1O1O00100000000O10O100000O1000O1000O10000000O0100000000O10O100000O1000O1000O10000000O11O01O000001O01O0001O0001O0001O0001O01O000000010O00000010O000000010O000000001O010O001O1O0010O01O1O001O0010O01O1O001O00100O001O001O010O1O001O1O1O010O1O1O1O1O10O01O1O1O1O1O10O01O1O[XU3"}, "label": "fingers holding the pizza box up"}]} {"id": 537196, "image": "COCO_train2014_000000537196.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left hands fingers of the person holding the box\"."}], "task": "refering", "answer_template": "The \"the left hands fingers of the person holding the box\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [340, 341, 342, 343, 344, 345, 357, 358, 359, 360, 361, 363, 364, 365, 366, 367, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.0021458333333333334, 0.867375, 0.6615624999999999, 0.9987343750000001], "iscrowd": 0, "area": 12896.622850000009, "rle": {"size": [640, 480], "counts": "jV1U1fa0U10000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000nN^^OFba09`^OF`a09b^OF^a09c^OG]a08e^OG[a09f^OFZa09h^OFXa09i^OGVa09l^OFTa09n^OFRa09P_OFPa09R_OFn`09S_OGm`08U_OGk`08W_OGi`08Y_OGg`09Z_OFf`09[_OGe`08]_OGc`08__OGa`08a_OG_`08c_OG]`08d_OH\\`07f_OHZ`07h_OHX`07j_OHV`08k_OGU`08l_OGU`08m_OFT`09m_OFT`09m_OGS`08n_OGS`09m_OFT`09m_OGS`08m_OHT`08l_OGU`08l_OHT`07m_OHT`08l_OGT`09m_OGS`08m_OHT`08l_OGU`08l_OHT`07m_OHT`08j1O1OR^OJT`06l_OJT`05m_OKS`04m1O100OQ^ONR`01n_O0R`00m_O1R`0On_O2R`0Mo_O3P`0No_O2R`0Mn_O4Q`0Lo_O5Q`0Kn_O6Q`0Jo_O7la01TNHo_O9ka0001O100O1O1O1O001O1O1O1O1O1O010O1O1O1O1O00100000000O10O100000O1000O1000O10000000O0100000000O10O100000O1000O1000O10000000O11O01O000001O01O0001O0001O0001O0001O01O000000010O00000010O000000010O000000001O010O001O1O0010O01O1O001O0010O01O1O001O00100O001O001O010O1O001O1O1O010O1O1O1O1O10O01O1O1O1O1O10O01O1O[XU3"}, "label": "the left hands fingers of the person holding the box"}]} {"id": 389743, "image": "COCO_train2014_000000389743.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man watching the tennis match\"."}], "task": "refering", "answer_template": "The \"the man watching the tennis match\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [23, 46, 47, 69, 70, 92, 93, 115, 116, 138, 139, 161, 162, 184, 185, 207, 208, 230, 231, 232, 253, 254, 255, 276, 277, 278, 300, 301, 323, 324], "bbox": [0.0016875000000000002, 0.06447916666666667, 0.12471874999999999, 0.8779583333333334], "iscrowd": 0, "area": 16693.661050000006, "rle": {"size": [480, 640], "counts": "Rd0P3m7X4L3M4L3M3M4M2M4L3M3M4L3M3M2N3M3M2N3M2N3M3M3M3M44K4LN2N7I=CO2N100O1O101 in this image.", "objects": [{"patches": [23, 46, 47, 69, 70, 92, 93, 115, 116, 138, 139, 161, 162, 184, 185, 207, 208, 230, 231, 232, 253, 254, 255, 276, 277, 278, 300, 301, 323, 324], "bbox": [0.0016875000000000002, 0.06447916666666667, 0.12471874999999999, 0.8779583333333334], "iscrowd": 0, "area": 16693.661050000006, "rle": {"size": [480, 640], "counts": "Rd0P3m7X4L3M4L3M3M4M2M4L3M3M4L3M3M2N3M3M2N3M2N3M3M3M3M44K4LN2N7I=CO2N100O1O101 in this image.", "objects": [{"patches": [34, 35, 36, 57, 58, 59, 60, 80, 81, 82, 103, 104, 105, 106, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 263, 264, 265, 266, 267, 268, 269, 270, 285, 286, 287, 288, 289, 290, 291, 292, 293, 308, 309, 310, 311, 312, 313, 314, 315, 316, 332, 333, 334, 335, 336, 337, 338, 339, 355, 356, 357, 358, 359, 360, 361, 362, 378, 379, 380, 382, 383, 384], "bbox": [0.41907812499999997, 0.08314583333333332, 0.77021875, 1.0], "iscrowd": 0, "area": 58327.0191, "rle": {"size": [480, 640], "counts": "Som31n>4M4K4M4K3N1N3N1N2O2M2nDROo7Q1lGUOQ8l0jGXOU8i0hGZOW8g0eG]OZ8d0cG_O\\8c0_GA`8`0]GCb8>[GEd8 in this image.", "objects": [{"patches": [34, 35, 36, 57, 58, 59, 60, 80, 81, 82, 103, 104, 105, 106, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 223, 241, 242, 243, 244, 245, 246, 263, 264, 265, 266, 267, 268, 269, 270, 285, 286, 287, 288, 289, 290, 291, 292, 293, 308, 309, 310, 311, 312, 313, 314, 315, 316, 332, 333, 334, 335, 336, 337, 338, 339, 355, 356, 357, 358, 359, 360, 361, 362, 378, 379, 380, 382, 383, 384], "bbox": [0.41907812499999997, 0.08314583333333332, 0.77021875, 1.0], "iscrowd": 0, "area": 58327.0191, "rle": {"size": [480, 640], "counts": "Som31n>4M4K4M4K3N1N3N1N2O2M2nDROo7Q1lGUOQ8l0jGXOU8i0hGZOW8g0eG]OZ8d0cG_O\\8c0_GA`8`0]GCb8>[GEd8 in this image.", "objects": [{"patches": [61, 62, 73, 74, 75, 86, 87, 88, 98, 99, 100, 111, 112, 123, 124, 125, 136, 137, 148, 149, 160, 161, 171, 172, 173, 184, 185, 195, 196, 197, 208, 209], "bbox": [0.02984, 0.24086000000000002, 0.8292266666666667, 0.94108], "iscrowd": 0, "area": 9813.192000000003, "rle": {"size": [500, 375], "counts": "ni52]?5O2O1N2O2M3N2M4M2M3N2M01O0O101O0O2O001N101O0O2O0010O010O010O10O010O010O01O1^O[OTBe0]>1O0@ZORBh0m=XOSBe0P>[OPBc0S>]OlAb0U>^OkAa0V>_OiAa0X>@gA`0Z>_OfA`0[>@eA`0[>AcA?^>AbA>_>BaA>`>B_A=b>C]A>c>B]A=d>D[AD[A;g>DYA;h>FVA;j>EVA;j>EVA;k>EUA:k>FUA:k>FUA:k>GUA8k>HUA8l>GUA8k>ITA7l>ITA7l>IUA6k>KTA5m>JSA5n>KSA4m>MRA3n>MSA2n>MRA3n>NQA2o>NRA2m>NSA2m>OSA0n>ORA1n>ORA1n>0RAOn>1RAOn>1RA0n>0RAOn>1RAOo>0RAOn>2QANP?1o@0Q?0m@2S?80O001O010O0010O01O01O01O010O0010O010O00010O00010O00010O00010O00010O000010O0001O01O01O0001O000000001O00000010O000001O000000001O0000001O0000001O00001O0001O01O00001O0000001O@N_A2_>2_AN^>8^AI_><_AD^>b0_A^O_>f0_AZO_>n0M3N2N2M3N2N1O2N2N1O2N2N2N1O2O1O0O2O1O001N2O001O1O0O2O100O010O10O10O10O01000O0100O01000O10O0100O010O1O010O10O01O10O2O00000O101O0O10000O2O000O100O2O0O100O101O0O100O101N100O2O0O2O001N1O2N1O2N101N1O2N1O2N1O200O010O00100O1O4L4M4K4L5KgSo0"}, "label": "a bright , multi - colored windsock hanging from a tent"}]} {"id": 45680, "image": "COCO_train2014_000000045680.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow , blue and pink fish kite\"."}], "task": "refering", "answer_template": "The \"a yellow , blue and pink fish kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 73, 74, 75, 86, 87, 88, 98, 99, 100, 111, 112, 123, 124, 125, 136, 137, 148, 149, 160, 161, 171, 172, 173, 184, 185, 195, 196, 197, 208, 209], "bbox": [0.02984, 0.24086000000000002, 0.8292266666666667, 0.94108], "iscrowd": 0, "area": 9813.192000000003, "rle": {"size": [500, 375], "counts": "ni52]?5O2O1N2O2M3N2M4M2M3N2M01O0O101O0O2O001N101O0O2O0010O010O010O10O010O010O01O1^O[OTBe0]>1O0@ZORBh0m=XOSBe0P>[OPBc0S>]OlAb0U>^OkAa0V>_OiAa0X>@gA`0Z>_OfA`0[>@eA`0[>AcA?^>AbA>_>BaA>`>B_A=b>C]A>c>B]A=d>D[AD[A;g>DYA;h>FVA;j>EVA;j>EVA;k>EUA:k>FUA:k>FUA:k>GUA8k>HUA8l>GUA8k>ITA7l>ITA7l>IUA6k>KTA5m>JSA5n>KSA4m>MRA3n>MSA2n>MRA3n>NQA2o>NRA2m>NSA2m>OSA0n>ORA1n>ORA1n>0RAOn>1RAOn>1RA0n>0RAOn>1RAOo>0RAOn>2QANP?1o@0Q?0m@2S?80O001O010O0010O01O01O01O010O0010O010O00010O00010O00010O00010O00010O000010O0001O01O01O0001O000000001O00000010O000001O000000001O0000001O0000001O00001O0001O01O00001O0000001O@N_A2_>2_AN^>8^AI_><_AD^>b0_A^O_>f0_AZO_>n0M3N2N2M3N2N1O2N2N1O2N2N2N1O2O1O0O2O1O001N2O001O1O0O2O100O010O10O10O10O01000O0100O01000O10O0100O010O1O010O10O01O10O2O00000O101O0O10000O2O000O100O2O0O100O101O0O100O101N100O2O0O2O001N1O2N1O2N101N1O2N1O2N1O200O010O00100O1O4L4M4K4L5KgSo0"}, "label": "a yellow , blue and pink fish kite"}]} {"id": 45680, "image": "COCO_train2014_000000045680.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man\"."}], "task": "refering", "answer_template": "The \"man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 92, 93, 104, 105, 106, 117, 118, 119, 130, 131, 132, 143, 144, 145, 156, 157, 169, 170, 182, 183, 195, 196], "bbox": [0.027093333333333334, 0.35074, 0.2339733333333333, 0.8641800000000001], "iscrowd": 0, "area": 9799.327149999997, "rle": {"size": [500, 375], "counts": "ZT5o0b>f0ZO9G9H6K3M4L3M4L3M3M6J5[DmL in this image.", "objects": [{"patches": [79, 80, 92, 93, 104, 105, 106, 117, 118, 119, 130, 131, 132, 143, 144, 145, 156, 157, 169, 170, 182, 183, 195, 196], "bbox": [0.027093333333333334, 0.35074, 0.2339733333333333, 0.8641800000000001], "iscrowd": 0, "area": 9799.327149999997, "rle": {"size": [500, 375], "counts": "ZT5o0b>f0ZO9G9H6K3M4L3M4L3M3M6J5[DmL in this image.", "objects": [{"patches": [72, 73, 85, 86, 98, 110, 111, 123, 136], "bbox": [0.48589333333333334, 0.31295999999999996, 0.6712266666666666, 0.59996], "iscrowd": 0, "area": 2357.6132, "rle": {"size": [500, 375], "counts": "PUi22\\?6I8I6J7I5N2O2O0O1O10M2N3M2O1N3M2N3M00J6J7I6K5J6K5J7L31O2N121N2N2O1N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2N2OQmk1"}, "label": "the pink dorsal fin part of the kite that looks like a fish"}]} {"id": 45680, "image": "COCO_train2014_000000045680.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pink back stripe of a flying fish windsock\"."}], "task": "refering", "answer_template": "The \"the pink back stripe of a flying fish windsock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 85, 86, 98, 110, 111, 123, 136], "bbox": [0.48589333333333334, 0.31295999999999996, 0.6712266666666666, 0.59996], "iscrowd": 0, "area": 2357.6132, "rle": {"size": [500, 375], "counts": "PUi22\\?6I8I6J7I5N2O2O0O1O10M2N3M2O1N3M2N3M00J6J7I6K5J6K5J7L31O2N121N2N2O1N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2N2O1N2O1N2N2O1N2N2O1N2N2OQmk1"}, "label": "the pink back stripe of a flying fish windsock"}]} {"id": 258679, "image": "COCO_train2014_000000258679.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a big dog near by the man\"."}], "task": "refering", "answer_template": "The \"a big dog near by the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 242, 243, 244, 266, 267, 268, 289, 290, 291, 312, 313, 314, 335, 336, 337, 359], "bbox": [0.51815625, 0.5576826722338204, 0.67875, 0.903089770354906], "iscrowd": 0, "area": 6814.014799999999, "rle": {"size": [479, 640], "counts": "fbk41m>2N2O1N2N2O1N2N2M3M2M3O1O1O00100O1O1O10O01O2N2O1O1O2OO01O00001[C[On:e0jDGR;9dD5X;l1M4L3N3L1O001O02jNZ1J3L002O100001O00001O0SDTNf:m1UE[Ng:e1SEdNLEP:g1lEIU:6eE0\\:Y21O1N2K5L4L4L4L4K5L4\\OaDbMd;[2b0N2M3N1N3M2O2M2N3N1O21N101N20O3N1N2O1N01^C[Nf;g1mCfNQ in this image.", "objects": [{"patches": [219, 220, 242, 243, 244, 266, 267, 268, 289, 290, 291, 312, 313, 314, 335, 336, 337, 359], "bbox": [0.51815625, 0.5576826722338204, 0.67875, 0.903089770354906], "iscrowd": 0, "area": 6814.014799999999, "rle": {"size": [479, 640], "counts": "fbk41m>2N2O1N2N2O1N2N2M3M2M3O1O1O00100O1O1O10O01O2N2O1O1O2OO01O00001[C[On:e0jDGR;9dD5X;l1M4L3N3L1O001O02jNZ1J3L002O100001O00001O0SDTNf:m1UE[Ng:e1SEdNLEP:g1lEIU:6eE0\\:Y21O1N2K5L4L4L4L4K5L4\\OaDbMd;[2b0N2M3N1N3M2O2M2N3N1O21N101N20O3N1N2O1N01^C[Nf;g1mCfNQ in this image.", "objects": [{"patches": [262, 263, 282, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309, 310, 328, 329, 333, 351, 352], "bbox": [0.25575000000000003, 0.6831666666666667, 0.5223906250000001, 0.9265625], "iscrowd": 0, "area": 8821.075899999998, "rle": {"size": [480, 640], "counts": "SW]24l>3M4L3L5L4L3ZCXOn:l0iDYOW;j0aD\\O];i0YD]Og;g0PD^OPB>BPR_4"}, "label": "the bench that is empty"}]} {"id": 160380, "image": "COCO_train2014_000000160380.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bench with no one on it\"."}], "task": "refering", "answer_template": "The \"the bench with no one on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [262, 263, 282, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309, 310, 328, 329, 333, 351, 352], "bbox": [0.25575000000000003, 0.6831666666666667, 0.5223906250000001, 0.9265625], "iscrowd": 0, "area": 8821.075899999998, "rle": {"size": [480, 640], "counts": "SW]24l>3M4L3L5L4L3ZCXOn:l0iDYOW;j0aD\\O];i0YD]Og;g0PD^OPB>BPR_4"}, "label": "the bench with no one on it"}]} {"id": 86654, "image": "COCO_train2014_000000086654.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an arm with dark blue sleeves\"."}], "task": "refering", "answer_template": "The \"an arm with dark blue sleeves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 138, 161, 162, 184, 185, 186, 187, 195, 196, 197, 198, 207, 208, 209, 210, 211, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.0016093750000000001, 0.33640845070422537, 0.6319999999999999, 1.0], "iscrowd": 0, "area": 49271.592949999984, "rle": {"size": [426, 640], "counts": "ja0\\1Q in this image.", "objects": [{"patches": [115, 138, 161, 162, 184, 185, 186, 187, 195, 196, 197, 198, 207, 208, 209, 210, 211, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 324, 325, 326, 327, 328, 329, 330], "bbox": [0.0016093750000000001, 0.33640845070422537, 0.6319999999999999, 1.0], "iscrowd": 0, "area": 49271.592949999984, "rle": {"size": [426, 640], "counts": "ja0\\1Q in this image.", "objects": [{"patches": [28, 29, 30, 50, 51, 52, 53, 54, 55, 73, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 106, 119, 120, 121, 122, 123, 124, 128, 129, 130, 142, 143, 144, 145, 146, 147, 150, 151, 152, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238], "bbox": [0.11646875000000001, 0.10483568075117371, 0.675046875, 0.7147887323943662], "iscrowd": 0, "area": 41657.78475, "rle": {"size": [426, 640], "counts": "k]o01X=2N2N2N2O1N2N2N2N101N2N2N2N2N2O1N2N2N1O2O0O1O2N1O1O2O0O2N1O1O2N100O2N1O1O2O0K5G:K4O1O2N1O2kMaM\\I`2n5XNoIi1e5fNVJ]1h5fNUJ[1i5iNTJX1k5lNQJV1m5mNPJT1n5POoIQ1P6ROmIP1Q6TOkIn0S6UOjIm0S6WOjIk0T6YOhIi0V6ZOgIh0W6[OfIg0W6]OfId0Y6@cIb0[6AbIa0\\6BaI`0^6C^I?`6W3O1N2O1N2N101N2O1N2O1N2N2O1N2O0O2N2O1N2O1N2O0O2N101N100O2O001O00001O001O001OO101O00000O10000000000O10001O000O100000000O100000001N10000000000O100000001N10001O1O0O2O001O1O0O2O1O001O1O0O2O1O0O2N101N2N1O2N2O0O2N2^MbIUN_6i1cIWN^6g1cIXN_6f1cIXN^6f1dIYN^6e1cI[N]6c1eI\\N]6b1dI]N^6`1dI_N]6`1eI^N]6`1dI`N]6]1eIbN\\6]1eIcN^6_OQI2c0>`6[OQI3a0b0a6VOSI1`0h0a6SORINc0n0^6oNSIMc0S1^6kN^I@9e1n9O100O2O0O100O100O101N100O100O101N100O100O101N100O100O100O1000000O1000000O1000000O1000000O1000000O100000O01O001O0010O01O001O001O010O00001O001O010O00001O001O01O01O0100O100O1O010O100O1O100O10O01O100O1O100O00100O2O0O2N1N3L3M4L3M4L3M4M2M4L3M3M4M2N3N1O2O\\Od0N2N1O2N2N2M3M3M3M3N2N2M3N2M3N2O1O100O100O1O100O100O1O100O101N101N2O0O2O0O2O1N105J6K6IP\\f2"}, "label": "boy with a strand of spaghetti in his mouth"}]} {"id": 86654, "image": "COCO_train2014_000000086654.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy is eating pizza\"."}], "task": "refering", "answer_template": "The \"a boy is eating pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 50, 51, 52, 53, 54, 55, 73, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 101, 106, 119, 120, 121, 122, 123, 124, 128, 129, 130, 142, 143, 144, 145, 146, 147, 150, 151, 152, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238], "bbox": [0.11646875000000001, 0.10483568075117371, 0.675046875, 0.7147887323943662], "iscrowd": 0, "area": 41657.78475, "rle": {"size": [426, 640], "counts": "k]o01X=2N2N2N2O1N2N2N2N101N2N2N2N2N2O1N2N2N1O2O0O1O2N1O1O2O0O2N1O1O2N100O2N1O1O2O0K5G:K4O1O2N1O2kMaM\\I`2n5XNoIi1e5fNVJ]1h5fNUJ[1i5iNTJX1k5lNQJV1m5mNPJT1n5POoIQ1P6ROmIP1Q6TOkIn0S6UOjIm0S6WOjIk0T6YOhIi0V6ZOgIh0W6[OfIg0W6]OfId0Y6@cIb0[6AbIa0\\6BaI`0^6C^I?`6W3O1N2O1N2N101N2O1N2O1N2N2O1N2O0O2N2O1N2O1N2O0O2N101N100O2O001O00001O001O001OO101O00000O10000000000O10001O000O100000000O100000001N10000000000O100000001N10001O1O0O2O001O1O0O2O1O001O1O0O2O1O0O2N101N2N1O2N2O0O2N2^MbIUN_6i1cIWN^6g1cIXN_6f1cIXN^6f1dIYN^6e1cI[N]6c1eI\\N]6b1dI]N^6`1dI_N]6`1eI^N]6`1dI`N]6]1eIbN\\6]1eIcN^6_OQI2c0>`6[OQI3a0b0a6VOSI1`0h0a6SORINc0n0^6oNSIMc0S1^6kN^I@9e1n9O100O2O0O100O100O101N100O100O101N100O100O101N100O100O100O1000000O1000000O1000000O1000000O1000000O100000O01O001O0010O01O001O001O010O00001O001O010O00001O001O01O01O0100O100O1O010O100O1O100O10O01O100O1O100O00100O2O0O2N1N3L3M4L3M4L3M4M2M4L3M3M4M2N3N1O2O\\Od0N2N1O2N2N2M3M3M3M3N2N2M3N2M3N2O1O100O100O1O100O100O1O100O101N101N2O0O2O0O2O1N105J6K6IP\\f2"}, "label": "a boy is eating pizza"}]} {"id": 86654, "image": "COCO_train2014_000000086654.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table cloth that is underneath the man and kids arms\"."}], "task": "refering", "answer_template": "The \"the table cloth that is underneath the man and kids arms\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [198, 199, 200, 201, 202, 222, 223, 224, 225, 266, 267, 268, 286, 287, 288, 289, 290, 291, 292, 307, 308, 309, 310, 311, 312, 313, 314, 329, 330, 331, 332, 333, 334, 335, 336], "bbox": [0.30803125, 0.587206572769953, 0.8047187499999999, 0.9890140845070422], "iscrowd": 0, "area": 15099.021149999999, "rle": {"size": [426, 640], "counts": "a[b26T=0O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O010O100O10000O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O010O100O10000O100O100O100O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O001O1O1O1O1O1O1O100O1000000001O0000001O0000001O0000001O000000001O0000001N10O100O100O10000O100O100O100O10000O100O100O10000O100O100O100O10000O100O10O0100O10000O1000000O1000000000000000M3M3N2O1O1O100O1O1O001O100OcMoMcIP2\\6RNdIm1]6SNcIl1^6TNbIl1^6TNbIk1^6VNbIi1_6WNaIh1`6YN_Ig1a6YN_If1b6ZN^Ie1c6[N]Ie1c6[N]Id1d6\\N\\Ic1e6]N[Ic1e6^NYIb1h6^NXIa1i6_NWIa1i6_NWI`1j6`NUI`1l6`NTI_1m6aNRI`1n6`NRI_1o6bNoH^1R7bNnH^1R7bNmH^1T7bNlH]1U7cNkH]1U7cNjH]1W7cNiH\\1X7dNgH]1Y7dNfH[1[7eNdH[1]7eNcHZ1^7fNaH[1_7eNcHX1^7hNgHR1Z7nNjHn0V7ROoHh0R7YOQIb0P7^OTI>l6BXI9i6HXI5j6KXI2h6NZIOg62[IJf67[IFf6:]ICc6>T31O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O001O1O1OO100O10000O10000O100O1O1O1O1O1N2O1O1O1O1N2O1O1O3M2M3N3M2N2NPhc1"}, "label": "the table cloth that is underneath the man and kids arms"}]} {"id": 209537, "image": "COCO_train2014_000000209537.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fat white man in a black shirt\"."}], "task": "refering", "answer_template": "The \"a fat white man in a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 129, 130, 131, 132, 152, 153, 154, 155, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 226, 242, 243, 244, 245, 246, 247, 248, 249, 270, 271], "bbox": [0.532578125, 0.2539375, 0.841015625, 0.6719166666666666], "iscrowd": 0, "area": 22233.680849999997, "rle": {"size": [480, 640], "counts": "WSP5=b>b0]O:G1N2O1O1N2O0O2O1N2O1O1N2O1N2O1O1N101N2O1N2O1O1N2O1N2O1O0O2O1N200O1O1O1O1O100O1O1O1O1O100O100O10000O01000O1000000O01000O10000O0100000WNRM_Go2`8UM]Gk2a8ZM\\Gf2c8^MWGd2i8`MlFh2S9\\McFk2\\9YMYFo2f9V1O1O001O1O1O1O1O001O2N1O10001O000000001O000000001O000000010O0000001O000000001O0001O1O100O001O1O100O1000000001NZL\\FQ2d9gMcFZ2]9^MkFb2U9UMSGk2n8mLZGR3g8fL`GZ3a8^LgGb3k901O001O010O001O001O010O001O0010O01O010O010O0010O010O010O01O010O010O010O010M2O2N1O3M4K4M4L4L4L3L5L4L4L3M4L7H:G9G9G8H9FPa_1"}, "label": "a fat white man in a black shirt"}]} {"id": 209537, "image": "COCO_train2014_000000209537.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in black shirt eating at an outside table\"."}], "task": "refering", "answer_template": "The \"man in black shirt eating at an outside table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 129, 130, 131, 132, 152, 153, 154, 155, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 226, 242, 243, 244, 245, 246, 247, 248, 249, 270, 271], "bbox": [0.532578125, 0.2539375, 0.841015625, 0.6719166666666666], "iscrowd": 0, "area": 22233.680849999997, "rle": {"size": [480, 640], "counts": "WSP5=b>b0]O:G1N2O1O1N2O0O2O1N2O1O1N2O1N2O1O1N101N2O1N2O1O1N2O1N2O1O0O2O1N200O1O1O1O1O100O1O1O1O1O100O100O10000O01000O1000000O01000O10000O0100000WNRM_Go2`8UM]Gk2a8ZM\\Gf2c8^MWGd2i8`MlFh2S9\\McFk2\\9YMYFo2f9V1O1O001O1O1O1O1O001O2N1O10001O000000001O000000001O000000010O0000001O000000001O0001O1O100O001O1O100O1000000001NZL\\FQ2d9gMcFZ2]9^MkFb2U9UMSGk2n8mLZGR3g8fL`GZ3a8^LgGb3k901O001O010O001O001O010O001O0010O01O010O010O0010O010O010O01O010O010O010O010M2O2N1O3M4K4M4L4L4L3L5L4L4L3M4L7H:G9G9G8H9FPa_1"}, "label": "man in black shirt eating at an outside table"}]} {"id": 209537, "image": "COCO_train2014_000000209537.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man weraing glasses , a gray sweatshort with writing on it , sitting at a table\"."}], "task": "refering", "answer_template": "The \"a man weraing glasses , a gray sweatshort with writing on it , sitting at a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 345, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372, 373], "bbox": [0.0, 0.305625, 0.34045312499999997, 0.9887708333333334], "iscrowd": 0, "area": 48572.50919999999, "rle": {"size": [480, 640], "counts": "Q9Y1f=b2^Mj1WN0O1O100O1O101N1O100O1O100O1O100O1O100O1O100O1B>M3N2M3M3M4M2M3M3O100O1O100O1O1O100O1O100O1O100O1O1O101N1O100O1O1O1O1N2N2N2N2N2N2N2N2N2N2N2M3N2N2N3M2N2N2N2N2N2N2N2M3N2O1O100O100O100O100O101N100O100O100O100O100O100O100O1O100O100O100O2O0000000001O000000001O000000001O000000001O01O00O1I8I6I7J6I7J6I8H7J6I7J7H8I7O1N1O2N2O1N2N101N3M3M4M8Gg0YO6K5J6J5K6K5WLjE]2\\:VMnEg2X:lLQFS3P;M2N3M2O2M2O2M2N3N1N3N1N3M2O2M2O1N3N2M2N3N1N3N1N3M2O2M3N1N3M2O2M2O2M2N3N2M7J8G8IU_U6"}, "label": "a man weraing glasses , a gray sweatshort with writing on it , sitting at a table"}]} {"id": 209537, "image": "COCO_train2014_000000209537.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a wisconsin university hoodie is sitting at a blue table with a nearly empty bowl in front of him\"."}], "task": "refering", "answer_template": "The \"a man in a wisconsin university hoodie is sitting at a blue table with a nearly empty bowl in front of him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 345, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372, 373], "bbox": [0.0, 0.305625, 0.34045312499999997, 0.9887708333333334], "iscrowd": 0, "area": 48572.50919999999, "rle": {"size": [480, 640], "counts": "Q9Y1f=b2^Mj1WN0O1O100O1O101N1O100O1O100O1O100O1O100O1O100O1B>M3N2M3M3M4M2M3M3O100O1O100O1O1O100O1O100O1O100O1O1O101N1O100O1O1O1O1N2N2N2N2N2N2N2N2N2N2N2M3N2N2N3M2N2N2N2N2N2N2N2M3N2O1O100O100O100O100O101N100O100O100O100O100O100O100O1O100O100O100O2O0000000001O000000001O000000001O000000001O01O00O1I8I6I7J6I7J6I8H7J6I7J7H8I7O1N1O2N2O1N2N101N3M3M4M8Gg0YO6K5J6J5K6K5WLjE]2\\:VMnEg2X:lLQFS3P;M2N3M2O2M2O2M2N3N1N3N1N3M2O2M2O1N3N2M2N3N1N3N1N3M2O2M3N1N3M2O2M2O2M2N3N2M7J8G8IU_U6"}, "label": "a man in a wisconsin university hoodie is sitting at a blue table with a nearly empty bowl in front of him"}]} {"id": 488073, "image": "COCO_train2014_000000488073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man sitting in the bleachers center\"."}], "task": "refering", "answer_template": "The \"the man sitting in the bleachers center\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 14, 15, 16, 17, 23, 24, 25, 26, 27, 33, 34, 35, 36, 37, 43, 44, 45, 53, 54], "bbox": [0.32093750000000004, 0.0045138888888888885, 0.7398611111111112, 0.3986574074074074], "iscrowd": 0, "area": 10377.626349999997, "rle": {"size": [432, 288], "counts": "klV11a=eE@V:l0eEROV:^1`EaN[:]2L4L4L4L4M3L4L3M4L4L4K5K5N2O1N1O0010N1N2O1O1N3N1O1N1O0000O2002N2M3N2N2N2M3N2O2M2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O2N1N2O1O1O1N3N3M2N2M3N3N1O1N101O0O2O0O101O2N1O1O2O0O1O2N101N2N2O1N2N3N1N2N2N100O2N100O1O1O>C9F9Ghao0"}, "label": "the man sitting in the bleachers center"}]} {"id": 488073, "image": "COCO_train2014_000000488073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a base ball bat\"."}], "task": "refering", "answer_template": "The \"a base ball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 14, 15, 16, 17, 23, 24, 25, 26, 27, 33, 34, 35, 36, 37, 43, 44, 45, 53, 54], "bbox": [0.32093750000000004, 0.0045138888888888885, 0.7398611111111112, 0.3986574074074074], "iscrowd": 0, "area": 10377.626349999997, "rle": {"size": [432, 288], "counts": "klV11a=eE@V:l0eEROV:^1`EaN[:]2L4L4L4L4M3L4L3M4L4L4K5K5N2O1N1O0010N1N2O1O1N3N1O1N1O0000O2002N2M3N2N2N2M3N2O2M2O1O1O1O1O1N2O1O1O1O1O1N2O1O1O2N1N2O1O1O1N3N3M2N2M3N3N1O1N101O0O2O0O101O2N1O1O2O0O1O2N101N2N2O1N2N3N1N2N2N100O2N100O1O1O>C9F9Ghao0"}, "label": "a base ball bat"}]} {"id": 488073, "image": "COCO_train2014_000000488073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in blue pant\"."}], "task": "refering", "answer_template": "The \"a man in blue pant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 10, 11, 12, 20, 21, 22, 30, 31, 41], "bbox": [0.0024652777777777776, 0.001412037037037037, 0.23149305555555552, 0.3217361111111111], "iscrowd": 0, "area": 6147.779899999999, "rle": {"size": [432, 288], "counts": "d?U1Z:T2M1O00001O000O100000000000000000001O0O10000000000000000000000001O1O1O1O14L3NdNPFZOk9d0VFI^95cFK^94bFL_92cFM^92bFN`9OaF1`9NaF1`9MaF3`9L`F4a9JeF1]9LjFNW91nFJS94SGGo87UGEm88YGCh8<\\G@f8=`G^Oa8a0Y2K5L3L5K5K4M4Kgcm2"}, "label": "a man in blue pant"}]} {"id": 488073, "image": "COCO_train2014_000000488073.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a spectator sitting in the stands , wearing blue jeans and a dark jacket\"."}], "task": "refering", "answer_template": "The \"a spectator sitting in the stands , wearing blue jeans and a dark jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 10, 11, 12, 20, 21, 22, 30, 31, 41], "bbox": [0.0024652777777777776, 0.001412037037037037, 0.23149305555555552, 0.3217361111111111], "iscrowd": 0, "area": 6147.779899999999, "rle": {"size": [432, 288], "counts": "d?U1Z:T2M1O00001O000O100000000000000000001O0O10000000000000000000000001O1O1O1O14L3NdNPFZOk9d0VFI^95cFK^94bFL_92cFM^92bFN`9OaF1`9NaF1`9MaF3`9L`F4a9JeF1]9LjFNW91nFJS94SGGo87UGEm88YGCh8<\\G@f8=`G^Oa8a0Y2K5L3L5K5K4M4Kgcm2"}, "label": "a spectator sitting in the stands , wearing blue jeans and a dark jacket"}]} {"id": 365205, "image": "COCO_train2014_000000365205.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat that looked shocked with big green eyes\"."}], "task": "refering", "answer_template": "The \"a cat that looked shocked with big green eyes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [37, 38, 39, 60, 61, 62, 83, 84, 85, 105, 106, 107, 108, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 328, 329, 330, 331, 332, 333, 334, 335, 336, 338, 339, 351, 352, 353, 354, 355, 356, 357, 358, 374, 375, 376, 377, 378, 379, 380, 381, 397, 398, 399, 400, 401, 402, 403, 420, 421, 422, 423, 424, 425, 426], "bbox": [0.29503124999999997, 0.06192883895131086, 0.9396093749999999, 0.989868913857678], "iscrowd": 0, "area": 96199.91385000004, "rle": {"size": [534, 640], "counts": "cQS3f0Z in this image.", "objects": [{"patches": [37, 38, 39, 60, 61, 62, 83, 84, 85, 105, 106, 107, 108, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 150, 151, 152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 218, 219, 220, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 328, 329, 330, 331, 332, 333, 334, 335, 336, 338, 339, 351, 352, 353, 354, 355, 356, 357, 358, 374, 375, 376, 377, 378, 379, 380, 381, 397, 398, 399, 400, 401, 402, 403, 420, 421, 422, 423, 424, 425, 426], "bbox": [0.29503124999999997, 0.06192883895131086, 0.9396093749999999, 0.989868913857678], "iscrowd": 0, "area": 96199.91385000004, "rle": {"size": [534, 640], "counts": "cQS3f0Z in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 110, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 192, 193, 194, 195, 196, 197, 198, 199, 200, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 269, 285, 286, 287, 288, 289, 290, 291, 308, 309, 310, 311, 312, 313, 332, 333, 334, 335, 336, 356, 357, 358, 359], "bbox": [0.33968750000000003, 0.23677083333333335, 0.8256718749999999, 0.9452916666666666], "iscrowd": 0, "area": 62189.412500000006, "rle": {"size": [480, 640], "counts": "i[V39e>8I8G8YC_Ob:h0kDIo:`0\\D2_;l1J5K8I9E:G:F9I001O1N2O1O001O1O1N2O001O1O1O1N101O1O1O1O0O2O1O001O001N101O1O4L4L4K5L4L3M1O1O2M2O1O2N1O1O2M2O1O2N1O1O2M2O1O2N1O1O1N3N1O1O2N1O1N3N1O1O2N1N2O2N1O1O2N1N2O1O001O000O2O02N5L4K5K5L4K5K5LiNcHkJ[7U5jHfJV7Z5PI`JQ7^5UI]Jk6c5[IXJe6g5aISJ_6m5fInIZ6R6[1O0O101O00001N10001O000O2O00001O0O101O001O0O101O00001O0O101O00001N10001O000O2O001O000O2O00001O0O101O00001O0O101O00001N10001O001N10001ON2K4M4L4K5L4L3M4K5L4L4K5M2O2M3N2M3N1N3M3N2M3N1N3N2M3M3N1N3N2M3N2M2O2M3M3N2M3N1YObGkJa8S5aGkJa8S5aGlJ`8R5bGnJ^8P5dGoJ\\8P5fGPKZ8n4hGRKX8l4jGTKV8j4lGVKT8h4oGWKP8h4RHWKo7g4SHYKm7d4VH\\Kj7b4XH^Kh7`4ZH`Ke7_4]H`Kd7^4^HbKb7\\4`HdK`7Z4bHfK^7X4dHhK\\7V4fHjKZ7S4iHlKW7R4lHnKT7n3QIQLo6l3TITLl6h3XIXLh6d3\\IYLg6d3\\IZLf6b3^I\\Lc6b3`I\\Lb6`3a2L4M3L4L4M3O0O2N2N2O1N2N2O0O2N2N2O1N2N2N101N2N2N2O1N2N101N2O1000000O0100000000000O0100000000000O1jNiBc0V=]OkBc0U=]OlBb0T=]OmBc0S=]OnBb0R=]OoB`0T=@lB=W=CjB:W=FjB7Y=IhB3[=LfB1]=OdBM_=2bBL`=4aBH]>Mf[d1"}, "label": "a half of a sandwich that has a piece of mushroom hanging out of the end"}]} {"id": 291493, "image": "COCO_train2014_000000291493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of sandwich closest to you\"."}], "task": "refering", "answer_template": "The \"a piece of sandwich closest to you\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 110, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 192, 193, 194, 195, 196, 197, 198, 199, 200, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 269, 285, 286, 287, 288, 289, 290, 291, 308, 309, 310, 311, 312, 313, 332, 333, 334, 335, 336, 356, 357, 358, 359], "bbox": [0.33968750000000003, 0.23677083333333335, 0.8256718749999999, 0.9452916666666666], "iscrowd": 0, "area": 62189.412500000006, "rle": {"size": [480, 640], "counts": "i[V39e>8I8G8YC_Ob:h0kDIo:`0\\D2_;l1J5K8I9E:G:F9I001O1N2O1O001O1O1N2O001O1O1O1N101O1O1O1O0O2O1O001O001N101O1O4L4L4K5L4L3M1O1O2M2O1O2N1O1O2M2O1O2N1O1O2M2O1O2N1O1O1N3N1O1O2N1O1N3N1O1O2N1N2O2N1O1O2N1N2O1O001O000O2O02N5L4K5K5L4K5K5LiNcHkJ[7U5jHfJV7Z5PI`JQ7^5UI]Jk6c5[IXJe6g5aISJ_6m5fInIZ6R6[1O0O101O00001N10001O000O2O00001O0O101O001O0O101O00001O0O101O00001N10001O000O2O001O000O2O00001O0O101O00001O0O101O00001N10001O001N10001ON2K4M4L4K5L4L3M4K5L4L4K5M2O2M3N2M3N1N3M3N2M3N1N3N2M3M3N1N3N2M3N2M2O2M3M3N2M3N1YObGkJa8S5aGkJa8S5aGlJ`8R5bGnJ^8P5dGoJ\\8P5fGPKZ8n4hGRKX8l4jGTKV8j4lGVKT8h4oGWKP8h4RHWKo7g4SHYKm7d4VH\\Kj7b4XH^Kh7`4ZH`Ke7_4]H`Kd7^4^HbKb7\\4`HdK`7Z4bHfK^7X4dHhK\\7V4fHjKZ7S4iHlKW7R4lHnKT7n3QIQLo6l3TITLl6h3XIXLh6d3\\IYLg6d3\\IZLf6b3^I\\Lc6b3`I\\Lb6`3a2L4M3L4L4M3O0O2N2N2O1N2N2O0O2N2N2O1N2N2N101N2N2N2O1N2N101N2O1000000O0100000000000O0100000000000O1jNiBc0V=]OkBc0U=]OlBb0T=]OmBc0S=]OnBb0R=]OoB`0T=@lB=W=CjB:W=FjB7Y=IhB3[=LfB1]=OdBM_=2bBL`=4aBH]>Mf[d1"}, "label": "a piece of sandwich closest to you"}]} {"id": 291493, "image": "COCO_train2014_000000291493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of a club sandwich with its filling obscured by the other half\"."}], "task": "refering", "answer_template": "The \"half of a club sandwich with its filling obscured by the other half\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 34, 35, 36, 37, 38, 39, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 126, 127], "bbox": [0.29493749999999996, 0.004229166666666667, 0.881453125, 0.3592916666666667], "iscrowd": 0, "area": 26814.044000000005, "rle": {"size": [480, 640], "counts": "\\gh25K5K5K6J4M3L3M3M2N2N2N2O1N2O001O100O1O1O010O1O100O1O010O1O100O1O00100O1O10O01O010O1O010O001O10O01O10O01O010O1O010O1O1O100O1O010O1O100O1O10O01O1O100O1O10O01O100O1O100O0010000O100O10O10O100O1000O0100O10000O10O010000O100O10O5L9G9G9G9G9G8G:G9GbME`F2`98_F_Oa9k0T21QBHR>7 in this image.", "objects": [{"patches": [13, 34, 35, 36, 37, 38, 39, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 126, 127], "bbox": [0.29493749999999996, 0.004229166666666667, 0.881453125, 0.3592916666666667], "iscrowd": 0, "area": 26814.044000000005, "rle": {"size": [480, 640], "counts": "\\gh25K5K5K6J4M3L3M3M2N2N2N2O1N2O001O100O1O1O010O1O100O1O010O1O100O1O00100O1O10O01O010O1O010O001O10O01O10O01O010O1O010O1O1O100O1O010O1O100O1O10O01O1O100O1O10O01O100O1O100O0010000O100O10O10O100O1000O0100O10000O10O010000O100O10O5L9G9G9G9G9G8G:G9GbME`F2`98_F_Oa9k0T21QBHR>7 in this image.", "objects": [{"patches": [299, 322, 323, 324, 345, 346, 347, 348, 349, 368, 369, 370, 371, 372, 373], "bbox": [0.0, 0.7793958333333334, 0.244375, 0.9906458333333333], "iscrowd": 0, "area": 8588.303, "rle": {"size": [480, 640], "counts": "f;`0a>n0ROn0QOe0\\OO001O001O1O010O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O10O01O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O00101N3M3M3M3MW]R7"}, "label": "part of table completely in shadows"}]} {"id": 291493, "image": "COCO_train2014_000000291493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the shaded area to the bottom left of the plate\"."}], "task": "refering", "answer_template": "The \"the shaded area to the bottom left of the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [299, 322, 323, 324, 345, 346, 347, 348, 349, 368, 369, 370, 371, 372, 373], "bbox": [0.0, 0.7793958333333334, 0.244375, 0.9906458333333333], "iscrowd": 0, "area": 8588.303, "rle": {"size": [480, 640], "counts": "f;`0a>n0ROn0QOe0\\OO001O001O1O010O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O10O01O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O00101N3M3M3M3MW]R7"}, "label": "the shaded area to the bottom left of the plate"}]} {"id": 291493, "image": "COCO_train2014_000000291493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tan colored table food is on\"."}], "task": "refering", "answer_template": "The \"tan colored table food is on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 65, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 135, 136, 137, 158, 159, 160, 182, 183, 206, 229, 252, 274, 275, 296, 297, 298, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.657296875, 0.10577083333333334, 1.0, 0.9799166666666665], "iscrowd": 0, "area": 29907.332749999998, "rle": {"size": [480, 640], "counts": "SiU61o>0O100O100O2N100O100O100O1O100O101N100O100O1O100O100O100O2N100O100O100O1O100O100O2O0O1O100O100O100O100O1O2O0O100O100O1O100O100O101N1O100O100O100O1O100O2O0O100O100O1O100O100O101N1O100O100O100O1O100O101N100O1O100O100O100O1O101N100]GfM_4[2aKfM^4Z2aKiM\\4X2dKiM[4W2dKkMZ4V2fKlMW4U2iKlMV4T2iKoMT4R2lKoMS4R2kKPNS4Q2mKQNQ4o1oKRNo3o1PLTNm3m1cJ^MROf0Z6l1dJbMmNd0]6k1fJeMiNb0_6i1gJjMfN>a6i1iJmMaN=d6f1kJQN]N:f6f1lJUNYN7i6e1nJXNUN5k6d1oJ[NQN3n6c1PK_NnM0P7a1RKcNjMMR7a1TKOk41TK2i4OWK2g4OYK3e4MZK5d4L\\K5c4K]K7`4J_K8`4H`K:]4HbK9]4GbK;\\4FdK in this image.", "objects": [{"patches": [43, 65, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 135, 136, 137, 158, 159, 160, 182, 183, 206, 229, 252, 274, 275, 296, 297, 298, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.657296875, 0.10577083333333334, 1.0, 0.9799166666666665], "iscrowd": 0, "area": 29907.332749999998, "rle": {"size": [480, 640], "counts": "SiU61o>0O100O100O2N100O100O100O1O100O101N100O100O1O100O100O100O2N100O100O100O1O100O100O2O0O1O100O100O100O100O1O2O0O100O100O1O100O100O101N1O100O100O100O1O100O2O0O100O100O1O100O100O101N1O100O100O100O1O100O101N100O1O100O100O100O1O101N100]GfM_4[2aKfM^4Z2aKiM\\4X2dKiM[4W2dKkMZ4V2fKlMW4U2iKlMV4T2iKoMT4R2lKoMS4R2kKPNS4Q2mKQNQ4o1oKRNo3o1PLTNm3m1cJ^MROf0Z6l1dJbMmNd0]6k1fJeMiNb0_6i1gJjMfN>a6i1iJmMaN=d6f1kJQN]N:f6f1lJUNYN7i6e1nJXNUN5k6d1oJ[NQN3n6c1PK_NnM0P7a1RKcNjMMR7a1TKOk41TK2i4OWK2g4OYK3e4MZK5d4L\\K5c4K]K7`4J_K8`4H`K:]4HbK9]4GbK;\\4FdK in this image.", "objects": [{"patches": [59, 61, 62, 76, 77, 78, 79, 80, 94, 95, 96, 97, 98, 112, 113, 114, 115, 129, 130, 131, 146, 147, 148, 163, 164, 165, 180, 181, 197, 198, 199, 214, 215], "bbox": [0.48533472803347283, 0.136015625, 0.8240585774058576, 0.5378281250000001], "iscrowd": 0, "area": 13488.92815, "rle": {"size": [640, 478], "counts": "TSa43hc07M3N1O2N1N3N1O2N2N1010O010O01000O010O0O2O0O2O0O2O0O1O010O01O010O0010O01O010O0010O00010O01O010O0010O01O010O003N2L4L4TOTO]^OP1^a0UO^^Oo0\\a0WO_^Om0\\a0ZO_^Oj0]a0P1TA_MR^OaAC_A7f>I\\A1g>O\\AKh>5YAFk>9XAAk>`0WAZOm>e0X2O010O100O010O1O010O12M3N1O2M3N2X]OQOZb0P1a]OVO\\b0Y1O2N1011O0N3K4L4L5K4L5L3L5K4L2N2O2M2N2NVZd1"}, "label": "a man skiing with one arm up in the air"}]} {"id": 332459, "image": "COCO_train2014_000000332459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in black overall with white buttons skiing\"."}], "task": "refering", "answer_template": "The \"a man in black overall with white buttons skiing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 61, 62, 76, 77, 78, 79, 80, 94, 95, 96, 97, 98, 112, 113, 114, 115, 129, 130, 131, 146, 147, 148, 163, 164, 165, 180, 181, 197, 198, 199, 214, 215], "bbox": [0.48533472803347283, 0.136015625, 0.8240585774058576, 0.5378281250000001], "iscrowd": 0, "area": 13488.92815, "rle": {"size": [640, 478], "counts": "TSa43hc07M3N1O2N1N3N1O2N2N1010O010O01000O010O0O2O0O2O0O2O0O1O010O01O010O0010O01O010O0010O00010O01O010O0010O01O010O003N2L4L4TOTO]^OP1^a0UO^^Oo0\\a0WO_^Om0\\a0ZO_^Oj0]a0P1TA_MR^OaAC_A7f>I\\A1g>O\\AKh>5YAFk>9XAAk>`0WAZOm>e0X2O010O100O010O1O010O12M3N1O2M3N2X]OQOZb0P1a]OVO\\b0Y1O2N1011O0N3K4L4L5K4L5L3L5K4L2N2O2M2N2NVZd1"}, "label": "a man in black overall with white buttons skiing"}]} {"id": 152237, "image": "COCO_train2014_000000152237.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top bunk\"."}], "task": "refering", "answer_template": "The \"the top bunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 143, 144, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 192, 193, 194, 195, 196, 216, 217, 218, 241], "bbox": [0.20684375, 0.19270833333333334, 0.6591875, 0.60975], "iscrowd": 0, "area": 34593.3004, "rle": {"size": [480, 640], "counts": "mnm13m>U1kNT1lN5K001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O00N2M3M3M3L4M3M4L3L4M3M3M3M3L4M3M3M3L4M3M3M3M3L4M3M3M3L4O1O1O1O2N2M2O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1Oo`V3"}, "label": "the top bunk"}]} {"id": 152237, "image": "COCO_train2014_000000152237.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bunk bed on top with purple unicorn not including the girl\"."}], "task": "refering", "answer_template": "The \"the bunk bed on top with purple unicorn not including the girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 143, 144, 145, 146, 147, 148, 149, 150, 151, 167, 168, 169, 170, 171, 172, 173, 192, 193, 194, 195, 196, 216, 217, 218, 241], "bbox": [0.20684375, 0.19270833333333334, 0.6591875, 0.60975], "iscrowd": 0, "area": 34593.3004, "rle": {"size": [480, 640], "counts": "mnm13m>U1kNT1lN5K001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O00N2M3M3M3L4M3M4L3L4M3M3M3M3L4M3M3M3L4M3M3M3M3L4M3M3M3L4O1O1O1O2N2M2O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1Oo`V3"}, "label": "the bunk bed on top with purple unicorn not including the girl"}]} {"id": 152237, "image": "COCO_train2014_000000152237.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange sheet covered mattress sitting on the lower bunk of the bed\"."}], "task": "refering", "answer_template": "The \"an orange sheet covered mattress sitting on the lower bunk of the bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 264, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 288, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 338, 339, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 361, 362, 363, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 383, 384, 385, 386, 387, 388], "bbox": [0.058859375000000005, 0.5828541666666667, 0.887890625, 0.9981041666666666], "iscrowd": 0, "area": 45966.19084999999, "rle": {"size": [480, 640], "counts": "aVb04i>5J6K4L5K5K5J5L5K5K4L5J6K5K4L5M3M2O2M3N1O2M2O1O1N2O2N1N2O1O2M2O1O1N2O2M2O1O1N2O2N1N2O1O1N3N1O1N2O1O2M2O1N10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O100O101O0O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O100O10000O10000O10N2O001O1O001O1O001O001O1N101N2M4M3M3L4M3L4M3M2M4M3M3L4M3M3L4M3LeFULh7h3XHZLi7c3WH_Lj7]3WHeLj7X3VHjLk7S3UHoLl7m2UHUMl7h2THZMm7c2SH_Mn7]2SHeMn7X2RHjMo7R2SHoMn7n1RHTNo7i1QHZNo7b1RH`No7]1QHeNo7Y1QHiNP8S1QHoNP8n0PHTOQ8i0oGYOR8c0oG_OR8>nGDS88nGJS83oGMR80QHOP8MTH2m7KVH4k7IXH6i7FZH:g7C\\Hc7^OaHa0`7\\OcHd0]7XOgHg0Z7VOiHi0X7TOkHk0V7QOnHn0S7oNoHQ1U9nNiDQ1T;SOlDl0R;WOnDh0P;[OPEd0m:@SE=m:FSE8l:KTE3j:1VEMi:6WEGi: in this image.", "objects": [{"patches": [236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 262, 263, 264, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 288, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 338, 339, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 361, 362, 363, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 383, 384, 385, 386, 387, 388], "bbox": [0.058859375000000005, 0.5828541666666667, 0.887890625, 0.9981041666666666], "iscrowd": 0, "area": 45966.19084999999, "rle": {"size": [480, 640], "counts": "aVb04i>5J6K4L5K5K5J5L5K5K4L5J6K5K4L5M3M2O2M3N1O2M2O1O1N2O2N1N2O1O2M2O1O1N2O2M2O1O1N2O2N1N2O1O1N3N1O1N2O1O2M2O1N10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O100O101O0O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O100O10000O10000O10N2O001O1O001O1O001O001O1N101N2M4M3M3L4M3L4M3M2M4M3M3L4M3M3L4M3LeFULh7h3XHZLi7c3WH_Lj7]3WHeLj7X3VHjLk7S3UHoLl7m2UHUMl7h2THZMm7c2SH_Mn7]2SHeMn7X2RHjMo7R2SHoMn7n1RHTNo7i1QHZNo7b1RH`No7]1QHeNo7Y1QHiNP8S1QHoNP8n0PHTOQ8i0oGYOR8c0oG_OR8>nGDS88nGJS83oGMR80QHOP8MTH2m7KVH4k7IXH6i7FZH:g7C\\Hc7^OaHa0`7\\OcHd0]7XOgHg0Z7VOiHi0X7TOkHk0V7QOnHn0S7oNoHQ1U9nNiDQ1T;SOlDl0R;WOnDh0P;[OPEd0m:@SE=m:FSE8l:KTE3j:1VEMi:6WEGi: in this image.", "objects": [{"patches": [152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 219, 220], "bbox": [0.5406875, 0.5462264150943396, 0.96490625, 0.8903144654088051], "iscrowd": 0, "area": 11634.521599999995, "rle": {"size": [318, 640], "counts": "le[3>Z9>_O;J6M4M4K2N2O1N1O2N2O1O0O2O1N2O001N100O10000O010O10000O1000O0010O0100O010N2N1O2N101N2N1O2N2O0O2N1N3N1O2M2O101O010O01O01O001O001O001O1N101O0011O0O100000000O1000000O2O00000000001O0000010O0000001O000000000000000001O0O100000000000O1000000O1000O10O100O10000O100O10000O100O01000O10000O10000O10000O10000O10000O1000000000000000000000000000O10000O1000000O1000001N1000000O1WOeGO[81eGO[82eGL\\84dGL\\84eGK[85eGK\\84eGI]87cGG_89bGD`8<`GBb8>_G^Od8b0\\G\\Of8d0[GYOg8g080001O000000000000000000001O01O0000000000000001O00000000001O00000000001O0000000001O01O00000000001O0000001O0000001O0000001O00001N101O001O001O003L3N3M\\n6"}, "label": "the right last bird in the right hand picture"}]} {"id": 365231, "image": "COCO_train2014_000000365231.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bird flying in the back of the pack\"."}], "task": "refering", "answer_template": "The \"the bird flying in the back of the pack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 219, 220], "bbox": [0.5406875, 0.5462264150943396, 0.96490625, 0.8903144654088051], "iscrowd": 0, "area": 11634.521599999995, "rle": {"size": [318, 640], "counts": "le[3>Z9>_O;J6M4M4K2N2O1N1O2N2O1O0O2O1N2O001N100O10000O010O10000O1000O0010O0100O010N2N1O2N101N2N1O2N2O0O2N1N3N1O2M2O101O010O01O01O001O001O001O1N101O0011O0O100000000O1000000O2O00000000001O0000010O0000001O000000000000000001O0O100000000000O1000000O1000O10O100O10000O100O10000O100O01000O10000O10000O10000O10000O10000O1000000000000000000000000000O10000O1000000O1000001N1000000O1WOeGO[81eGO[82eGL\\84dGL\\84eGK[85eGK\\84eGI]87cGG_89bGD`8<`GBb8>_G^Od8b0\\G\\Of8d0[GYOg8g080001O000000000000000000001O01O0000000000000001O00000000001O00000000001O0000000001O01O00000000001O0000001O0000001O0000001O00001N101O001O001O003L3N3M\\n6"}, "label": "the bird flying in the back of the pack"}]} {"id": 365231, "image": "COCO_train2014_000000365231.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bird is flying in front of two others\"."}], "task": "refering", "answer_template": "The \"a bird is flying in front of two others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 24, 25, 26, 27, 30, 31, 32, 48, 49, 50, 51, 52, 53, 54, 70, 71, 72, 73, 74, 75, 96, 97], "bbox": [0.03846875, 0.008616352201257862, 0.41139062499999995, 0.3763522012578616], "iscrowd": 0, "area": 9939.577850000001, "rle": {"size": [318, 640], "counts": "ak71m91O0O2O1N101O0O2O001N101N2O001N101UMEPLQKBP5?oJAR5>mJCS5>lJAU5`0jJ@V5a0hJ_OY5a0gJ_OZ5a0dJ_O]5b0bJ^O^5c0aJ]O_5c0`J]Oa5d0^J\\Oc5d0[J\\Og5d0XJ\\Oi5d0VJ[Ol5d0SJ]On5c0QJ]Oo5d0PJ[OR6e0lI\\OU6c0kI\\OW6d0gI]OZ6c0dI]O]6d0bI\\O_6c0`I^Oa6b0^I]Od6c0ZI^Og6b0XI]Oj6c0TI^Ol6b0TI_Ol6a0RI@o6`0PIAP7?nHCR7kHAU7?lH@T7`0lH@S7`0oH_OQ7a0PI^Oo6c0QI\\OP7c0U11000O0100O10O1000000O1000O10O1000O10O100000O010000O10O0100O1O10O0100O100O100O1O100O100O100O1O100O1O100O1O100O100O1O100OoZe3"}, "label": "a bird is flying in front of two others"}]} {"id": 365231, "image": "COCO_train2014_000000365231.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bird is flying in front of two others\"."}], "task": "refering", "answer_template": "The \"a bird is flying in front of two others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 24, 25, 26, 27, 30, 31, 32, 48, 49, 50, 51, 52, 53, 54, 70, 71, 72, 73, 74, 75, 96, 97], "bbox": [0.03846875, 0.008616352201257862, 0.41139062499999995, 0.3763522012578616], "iscrowd": 0, "area": 9939.577850000001, "rle": {"size": [318, 640], "counts": "ak71m91O0O2O1N101O0O2O001N101N2O001N101UMEPLQKBP5?oJAR5>mJCS5>lJAU5`0jJ@V5a0hJ_OY5a0gJ_OZ5a0dJ_O]5b0bJ^O^5c0aJ]O_5c0`J]Oa5d0^J\\Oc5d0[J\\Og5d0XJ\\Oi5d0VJ[Ol5d0SJ]On5c0QJ]Oo5d0PJ[OR6e0lI\\OU6c0kI\\OW6d0gI]OZ6c0dI]O]6d0bI\\O_6c0`I^Oa6b0^I]Od6c0ZI^Og6b0XI]Oj6c0TI^Ol6b0TI_Ol6a0RI@o6`0PIAP7?nHCR7kHAU7?lH@T7`0lH@S7`0oH_OQ7a0PI^Oo6c0QI\\OP7c0U11000O0100O10O1000000O1000O10O1000O10O100000O010000O10O0100O1O10O0100O100O100O1O100O100O100O1O100O1O100O1O100O100O1O100OoZe3"}, "label": "a bird is flying in front of two others"}]} {"id": 365231, "image": "COCO_train2014_000000365231.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle flying bird\"."}], "task": "refering", "answer_template": "The \"the middle flying bird\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196], "bbox": [0.311734375, 0.5157232704402516, 0.59528125, 0.7706603773584906], "iscrowd": 0, "area": 6336.5082, "rle": {"size": [318, 640], "counts": "hXn18n1Jh59TJGo5:nIFU6jHAU7a0jH@U7a0kH_OS7c0lH^OT7MRH>i0EV7KSH`0g0EV7HUHd0d0E]7;cHD^7 in this image.", "objects": [{"patches": [122, 145, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196], "bbox": [0.311734375, 0.5157232704402516, 0.59528125, 0.7706603773584906], "iscrowd": 0, "area": 6336.5082, "rle": {"size": [318, 640], "counts": "hXn18n1Jh59TJGo5:nIFU6jHAU7a0jH@U7a0kH_OS7c0lH^OT7MRH>i0EV7KSH`0g0EV7HUHd0d0E]7;cHD^7 in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 18, 19, 20, 21, 22, 23, 36, 37, 38, 39, 51, 52, 53, 54, 55, 56, 68, 69, 70, 71, 72, 73, 74, 85, 86, 87, 88, 89, 102, 103, 119, 120, 121, 136, 137], "bbox": [0.0, 0.0, 0.36131250000000004, 0.35578125], "iscrowd": 0, "area": 22054.396100000005, "rle": {"size": [640, 480], "counts": "e3b0^c0T1kN>C2M3N3L3N2M3N3L3N2M3N3L3M3N2M4M2M3N2M4M1N010O1hLVLZFj3f9YLVFh3j9[LSFd3n9^LoEc3Q:`LlE`3T:cLhE^3X:cLgE]3Y:eLdE\\3\\:fLbEY3_:iL^EX3b:iL]EW3c:kLZEV3f:lLWEU3i:mLUER3l:oLRER3n:PMPEP3P;QMnDP3R;RMlDn2T;SMjDm2W;UMgDk2Y;VMeDk2[;VMdDP1]N@o<_OdDR1_N]OmSNnAP2R>PNjAS2W>mMiAS2W>lMiAT2X>lMhAT2X>mMgAS2Y>mMfAT2Z>mMeAR2\\>nMcAS2]>nMbAR2^>nMbAR2^>oM`AQ2a>oM_AQ2a>PN^AP2b>PN]AQ2c>PN\\Ao1e>QNZAP2f>QNYAo1g>QNYAo1g>RNXAm1i>SNXAl1h>UNWAk1i>UNWAk1i>VNVAi1k>WNUAi1k>XNTAh1l>YNSAg1m>YNSAf1n>[NQAe1o>[NQAe1o>\\NPAd1P?\\NPAc1Q?^Nn@b1R?_Nm@a1S?_Nm@a1S?`Nl@_1U?aNk@_1U?dNh@\\1X?jNb@V1^?e11000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000006QNi@hN^?:bA^Oe>DYB5Pa0ITho5"}, "label": "a woman in a white shirt and blue jeans sits at a table with bad posture"}]} {"id": 209603, "image": "COCO_train2014_000000209603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man sits with a napkin in his lap\"."}], "task": "refering", "answer_template": "The \"a man sits with a napkin in his lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 18, 19, 20, 21, 22, 23, 36, 37, 38, 39, 51, 52, 53, 54, 55, 56, 68, 69, 70, 71, 72, 73, 74, 85, 86, 87, 88, 89, 102, 103, 119, 120, 121, 136, 137], "bbox": [0.0, 0.0, 0.36131250000000004, 0.35578125], "iscrowd": 0, "area": 22054.396100000005, "rle": {"size": [640, 480], "counts": "e3b0^c0T1kN>C2M3N3L3N2M3N3L3N2M3N3L3M3N2M4M2M3N2M4M1N010O1hLVLZFj3f9YLVFh3j9[LSFd3n9^LoEc3Q:`LlE`3T:cLhE^3X:cLgE]3Y:eLdE\\3\\:fLbEY3_:iL^EX3b:iL]EW3c:kLZEV3f:lLWEU3i:mLUER3l:oLRER3n:PMPEP3P;QMnDP3R;RMlDn2T;SMjDm2W;UMgDk2Y;VMeDk2[;VMdDP1]N@o<_OdDR1_N]OmSNnAP2R>PNjAS2W>mMiAS2W>lMiAT2X>lMhAT2X>mMgAS2Y>mMfAT2Z>mMeAR2\\>nMcAS2]>nMbAR2^>nMbAR2^>oM`AQ2a>oM_AQ2a>PN^AP2b>PN]AQ2c>PN\\Ao1e>QNZAP2f>QNYAo1g>QNYAo1g>RNXAm1i>SNXAl1h>UNWAk1i>UNWAk1i>VNVAi1k>WNUAi1k>XNTAh1l>YNSAg1m>YNSAf1n>[NQAe1o>[NQAe1o>\\NPAd1P?\\NPAc1Q?^Nn@b1R?_Nm@a1S?_Nm@a1S?`Nl@_1U?aNk@_1U?dNh@\\1X?jNb@V1^?e11000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000006QNi@hN^?:bA^Oe>DYB5Pa0ITho5"}, "label": "a man sits with a napkin in his lap"}]} {"id": 209603, "image": "COCO_train2014_000000209603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red jacket hanging on the back of a chair\"."}], "task": "refering", "answer_template": "The \"a red jacket hanging on the back of a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 24, 25, 26, 27, 28, 41, 42, 43, 44, 45, 58, 59, 60, 61, 62, 75, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97], "bbox": [0.4423125, 0.00284375, 0.7474791666666666, 0.24940625], "iscrowd": 0, "area": 16455.45379999999, "rle": {"size": [640, 480], "counts": "hTU43lc02M2N3W^OHg?:Q@5e?NS@`NkA]1^`0N1N2O2N1O2N1O1O1O1O1O0O2O1O0000000000001O0O1000000O2N1N2O1OV`[2"}, "label": "a red jacket hanging on the back of a chair"}]} {"id": 209603, "image": "COCO_train2014_000000209603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red jacket on a chair\"."}], "task": "refering", "answer_template": "The \"a red jacket on a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 9, 10, 11, 24, 25, 26, 27, 28, 41, 42, 43, 44, 45, 58, 59, 60, 61, 62, 75, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97], "bbox": [0.4423125, 0.00284375, 0.7474791666666666, 0.24940625], "iscrowd": 0, "area": 16455.45379999999, "rle": {"size": [640, 480], "counts": "hTU43lc02M2N3W^OHg?:Q@5e?NS@`NkA]1^`0N1N2O2N1O2N1O1O1O1O1O0O2O1O0000000000001O0O1000000O2N1N2O1OV`[2"}, "label": "a red jacket on a chair"}]} {"id": 209603, "image": "COCO_train2014_000000209603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden plank table with plates of food on it\"."}], "task": "refering", "answer_template": "The \"a wooden plank table with plates of food on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 187, 188, 189, 190, 191, 204, 205, 206, 207, 208, 209, 210, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 272, 273, 274, 275, 276, 277, 278, 279, 289, 290, 291, 292, 293, 294, 295, 306, 307, 308, 309, 310, 311, 312, 313, 323, 324, 325, 326, 327, 328, 329, 330, 340, 341, 342, 343, 344, 345, 346, 347, 348, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.0125, 0.45221875, 0.6120208333333333, 0.9744687500000001], "iscrowd": 0, "area": 66573.25749999998, "rle": {"size": [640, 480], "counts": "bZ4l0]a0g1XNh1YNg1YNg1XNh1]Oc0000000000000000000000000000000000000000000000000000000000000001O00001O00001O00001O00001O00001O0000001O00001O01O01O00001O00001O00001O00001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O0iKeDI\\;5fDJZ;5hDJX;6iDIX;5jDJV;5mDIT;5nDJR;6oDIQ;6QEIP;5REJn:5TEJm:5TEJl:5VEJj:5YEIh:5ZEJf:6[EIf:5\\EJd:5^EJc:4_EKa:5`EJ`:5bEJ_:4dEJ\\:5fEJ[:5fEJZ:5hEJX:5jEJW:4kEKU:5lEJU:4mEKS:4PFJP:5RFJo95RFJn95TFJm94UFKk94WFKi95XFJi94YFKg94\\FJe94]FKc95^FJc94_FKa94aFK_94cFK^94cFK]94eFK\\93gFKY94iFKW95jFJW94kFKU94mFKT93nFLR93PGLP94QGKP93SGKm84UGKl83VGLj84WGKi84YGKh83ZGLf83\\GLe83\\GLd83_GKb83`GL`83bGL^84cGK^83dGL\\83fGL[82gGMY83hGLX83kGKV83lGLT83nGLS83nGLR83PHLP83RHLo72SHMm73THLm72VHLj73XHLh73ZHLg73ZHLf73\\HLe72]HMc72_HMb72_HMa72bHL^73dHL]72eHM[73fHL[72gHMY72iHMW72kHMV72kHMU72nHLS72oHMQ72QIMo63RILo62SIMm62UIMl61VIOi61XINh61[IMf61\\INd61^INc61^INb61`INa60aIO_60cIO]61dIN]60fINZ61hINY60iIOW61jINV61lINU60mIOS60oIOR60oIOQ60RJNn51TJNm50R70ckc3"}, "label": "a wooden plank table with plates of food on it"}]} {"id": 209603, "image": "COCO_train2014_000000209603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table made of red wood lays beneath two black mats\"."}], "task": "refering", "answer_template": "The \"the table made of red wood lays beneath two black mats\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 187, 188, 189, 190, 191, 204, 205, 206, 207, 208, 209, 210, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 272, 273, 274, 275, 276, 277, 278, 279, 289, 290, 291, 292, 293, 294, 295, 306, 307, 308, 309, 310, 311, 312, 313, 323, 324, 325, 326, 327, 328, 329, 330, 340, 341, 342, 343, 344, 345, 346, 347, 348, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.0125, 0.45221875, 0.6120208333333333, 0.9744687500000001], "iscrowd": 0, "area": 66573.25749999998, "rle": {"size": [640, 480], "counts": "bZ4l0]a0g1XNh1YNg1YNg1XNh1]Oc0000000000000000000000000000000000000000000000000000000000000001O00001O00001O00001O00001O00001O0000001O00001O01O01O00001O00001O00001O00001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O001O00001O001O00001O001O00001O001O00001O001O00001O001O001O00001O0iKeDI\\;5fDJZ;5hDJX;6iDIX;5jDJV;5mDIT;5nDJR;6oDIQ;6QEIP;5REJn:5TEJm:5TEJl:5VEJj:5YEIh:5ZEJf:6[EIf:5\\EJd:5^EJc:4_EKa:5`EJ`:5bEJ_:4dEJ\\:5fEJ[:5fEJZ:5hEJX:5jEJW:4kEKU:5lEJU:4mEKS:4PFJP:5RFJo95RFJn95TFJm94UFKk94WFKi95XFJi94YFKg94\\FJe94]FKc95^FJc94_FKa94aFK_94cFK^94cFK]94eFK\\93gFKY94iFKW95jFJW94kFKU94mFKT93nFLR93PGLP94QGKP93SGKm84UGKl83VGLj84WGKi84YGKh83ZGLf83\\GLe83\\GLd83_GKb83`GL`83bGL^84cGK^83dGL\\83fGL[82gGMY83hGLX83kGKV83lGLT83nGLS83nGLR83PHLP83RHLo72SHMm73THLm72VHLj73XHLh73ZHLg73ZHLf73\\HLe72]HMc72_HMb72_HMa72bHL^73dHL]72eHM[73fHL[72gHMY72iHMW72kHMV72kHMU72nHLS72oHMQ72QIMo63RILo62SIMm62UIMl61VIOi61XINh61[IMf61\\INd61^INc61^INb61`INa60aIO_60cIO]61dIN]60fINZ61hINY60iIOW61jINV61lINU60mIOS60oIOR60oIOQ60RJNn51TJNm50R70ckc3"}, "label": "the table made of red wood lays beneath two black mats"}]} {"id": 209603, "image": "COCO_train2014_000000209603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the calzone in the middle of the picture\"."}], "task": "refering", "answer_template": "The \"the calzone in the middle of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 142, 143, 144, 145, 146, 147, 148, 158, 159, 160, 161, 162, 163, 164, 165, 166, 175, 176, 177, 178, 179, 180, 181, 182, 194, 195, 196, 197], "bbox": [0.3236041666666667, 0.341578125, 0.7760208333333334, 0.48764062500000005], "iscrowd": 0, "area": 15476.820399999997, "rle": {"size": [640, 480], "counts": "^TQ31mc08I6I6J3N1N3M2O2M2N3N1N3M2O1N3M2N3N1N3M2O2M2O2O001N101O00001O001O001O001O001O001O001O00001O00001O0000000000000O100000001O0000000000000000000000001O000O100000000000000000001O00000000000O1000000c^O_M[a0a2e^O_M[a0c20O100000000000000000O1000O100000000000000000000O1000O1000000000000000O1000000000O10000000000O0100000000000O10000000O10O10000000000O100000O100000O100000O1000O010000O100O101O1N2O1N2O1O1N2O1O1N2O2M2O1O1N3N1O1N2O2M2O1O1N2O2L3N3M5K5J6K5K5J6K^WS2"}, "label": "the calzone in the middle of the picture"}]} {"id": 209603, "image": "COCO_train2014_000000209603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza on the upper left looks hot\"."}], "task": "refering", "answer_template": "The \"the pizza on the upper left looks hot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 121, 122, 123, 124, 125, 126, 127, 137, 138, 139, 140, 141, 142, 143, 154, 155, 156, 157], "bbox": [0.06635416666666667, 0.273671875, 0.49729166666666663, 0.41600000000000004], "iscrowd": 0, "area": 10233.20575, "rle": {"size": [640, 480], "counts": "gWd07ec07J4K5M2N3L4M3O001N2O001N2O0O1000O010000O10O10O100O01000O1000O010000O100O10000O101O0O10000O100O2O000O10000O10000O10000O100O01000O10000O100O10O10O10000O100O10O10O10000O100O10000O100O10000O10000O100000O010O1O10O01O100O00100O1O010000000O01000000O1000000O01000O1O01O001O001N1N3N1N2O11O001O10O011O0O10001O000O2O00001O0O10001O00010O0000001O0000001O0000010O0000001O0000001O0O2N2N101N2N1O2O0O2N2N101N2O1N3N2M4M2M3N3L3N2M[]f4"}, "label": "the pizza on the upper left looks hot"}]} {"id": 209603, "image": "COCO_train2014_000000209603.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a calzone sitting on a pan to the left of another calzone and it is by the edge of the table\"."}], "task": "refering", "answer_template": "The \"a calzone sitting on a pan to the left of another calzone and it is by the edge of the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 121, 122, 123, 124, 125, 126, 127, 137, 138, 139, 140, 141, 142, 143, 154, 155, 156, 157], "bbox": [0.06635416666666667, 0.273671875, 0.49729166666666663, 0.41600000000000004], "iscrowd": 0, "area": 10233.20575, "rle": {"size": [640, 480], "counts": "gWd07ec07J4K5M2N3L4M3O001N2O001N2O0O1000O010000O10O10O100O01000O1000O010000O100O10000O101O0O10000O100O2O000O10000O10000O10000O100O01000O10000O100O10O10O10000O100O10O10O10000O100O10000O100O10000O10000O100000O010O1O10O01O100O00100O1O010000000O01000000O1000000O01000O1O01O001O001N1N3N1N2O11O001O10O011O0O10001O000O2O00001O0O10001O00010O0000001O0000001O0000010O0000001O0000001O0O2N2N101N2N1O2O0O2N2N101N2O1N3N2M4M2M3N3L3N2M[]f4"}, "label": "a calzone sitting on a pan to the left of another calzone and it is by the edge of the table"}]} {"id": 4830, "image": "COCO_train2014_000000004830.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sink on the left\"."}], "task": "refering", "answer_template": "The \"sink on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 116, 117, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194], "bbox": [0.072265625, 0.2758333333333333, 0.482578125, 0.5352916666666666], "iscrowd": 0, "area": 22606.717950000002, "rle": {"size": [480, 640], "counts": "aUf01o>2N1O2M2O2N1O1O2N1`AC[>d0M3M3M3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M3M3M3\\DVMl:_30000000O1000003M6J3M000O010000000000000O1000O100000000000O1000O1EnLhDR3X;PMfDP3Z;SMcDm2];UMaDk2_;:1000O100000000000O10O100000000000O2O3M3M3M2N0000000O100000000000000O10000000O10000000OESMeDm2[;VMbDj2^;XM`Dh2`;[M]De2c; in this image.", "objects": [{"patches": [98, 99, 100, 116, 117, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194], "bbox": [0.072265625, 0.2758333333333333, 0.482578125, 0.5352916666666666], "iscrowd": 0, "area": 22606.717950000002, "rle": {"size": [480, 640], "counts": "aUf01o>2N1O2M2O2N1O1O2N1`AC[>d0M3M3M3M3M3M3M3M3M3M4L3M3M3M3M3M3M3M3M3M3M3M3M3\\DVMl:_30000000O1000003M6J3M000O010000000000000O1000O100000000000O1000O1EnLhDR3X;PMfDP3Z;SMcDm2];UMaDk2_;:1000O100000000000O10O100000000000O2O3M3M3M2N0000000O100000000000000O10000000O10000000OESMeDm2[;VMbDj2^;XM`Dh2`;[M]De2c; in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 87, 88, 89, 90, 112, 113, 136], "bbox": [0.786640625, 0.0018266978922716628, 0.969703125, 0.3576580796252927], "iscrowd": 0, "area": 12574.571600000003, "rle": {"size": [427, 640], "counts": "Pka62m< in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 87, 88, 89, 90, 112, 113, 136], "bbox": [0.786640625, 0.0018266978922716628, 0.969703125, 0.3576580796252927], "iscrowd": 0, "area": 12574.571600000003, "rle": {"size": [427, 640], "counts": "Pka62m< in this image.", "objects": [{"patches": [189, 190, 191, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.50386, 0.79816, 0.97884, 1.0], "iscrowd": 0, "area": 10333.543900000002, "rle": {"size": [375, 500], "counts": "Qcl2?X;9G00001N1000001O0O10001O0O10001O000O101O000O10001O000O101O00000O2O00000O2O0000000O2O00000O2O0000000O2O0000001N1000000O2O0000000O2O00000O2O0000000O2O0O10001O01O0ZNYF\\1T:M2OO000001O01O0000010O000000010O0000010O000000010O000000010O00000010O00000010O00000010O000000010O000000010O000001O01O0000010O000000010O00000001O000000000O1000000000000O1O1O1O1O1N2H]NZFe1f9600000000001O0000000000001O0000000O100O2N100O1O100O1O101N1O100O1O100O2N100O1O100O100O2N100O1O100O1O2O0O1O100O1O101N1O100O1^OUE7l:H^EOd:N^E0\\Q4"}, "label": "the book under th phone"}]} {"id": 537337, "image": "COCO_train2014_000000537337.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a book holding a charging mobile near the cat\"."}], "task": "refering", "answer_template": "The \"a book holding a charging mobile near the cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [189, 190, 191, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.50386, 0.79816, 0.97884, 1.0], "iscrowd": 0, "area": 10333.543900000002, "rle": {"size": [375, 500], "counts": "Qcl2?X;9G00001N1000001O0O10001O0O10001O000O101O000O10001O000O101O00000O2O00000O2O0000000O2O00000O2O0000000O2O0000001N1000000O2O0000000O2O00000O2O0000000O2O0O10001O01O0ZNYF\\1T:M2OO000001O01O0000010O000000010O0000010O000000010O000000010O00000010O00000010O00000010O000000010O000000010O000001O01O0000010O000000010O00000001O000000000O1000000000000O1O1O1O1O1N2H]NZFe1f9600000000001O0000000000001O0000000O100O2N100O1O100O1O101N1O100O1O100O2N100O1O100O100O2N100O1O100O1O2O0O1O100O1O101N1O100O1^OUE7l:H^EOd:N^E0\\Q4"}, "label": "a book holding a charging mobile near the cat"}]} {"id": 226046, "image": "COCO_train2014_000000226046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bottle\"."}], "task": "refering", "answer_template": "The \"the bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [16, 17, 18, 19, 39, 40, 41, 42, 61, 62, 63, 64, 84, 85, 86, 107, 108, 110, 111, 130, 131, 132, 133, 134, 154, 155, 156, 178, 179], "bbox": [0.6663125, 0.0012916666666666667, 0.8610468749999999, 0.432625], "iscrowd": 0, "area": 12808.518649999998, "rle": {"size": [480, 640], "counts": "QkW6>[>6K6I6K6I5L5J6K5J5L5J6K5J5L5J6K5J5L5J6K5M21000000O10000O10001O0O1000000O1000hF[LV7e3jH\\LU7c3lH^LS7b3mH_LR7a3nHaLP7_3PIbLo6]3QIeLo6Z3QIgLn6Y3RIhLm6X3SIjLk6U3VIlLi6T3WImLi6R3WIoLi6P3WIRMh6l2XIVMh6i2XIXMh6g2XIZMh6e2XI]Mg6a2ZI`Mf6_2ZIbMf6]2ZIdMf6[2ZIfMf6X2ZIkMe6T2[ImMe6R2[IoMe6P2[IQNe6m1\\IUNc6j1]IWNc6h1]IYNc6e1]I]Nc6b1]I`Nb6_1^IbNb6]1^IdNb6Z1_IgNa6X1_IjN`6U1`IlN`6S1`InN`6P1`IRO`6m0`ITO`6k0`IWO_6h0aIYO_6e0bI\\O^6c0bI^O^6a0bIA]6>cIC]6;cIG]68cII]66cIL\\62eIO[60eI1[6MfI4Z6JgI7Y6HfI9[6EfI8^6GbI6b6H_I5e6J[I3i6KXI2l6LUI1o6NQIOS7OnHNV71jHLZ74eHI_77`HFd7:[HCi7=VH@n7?SH]OQ8c0nGZOV8f0iGWO[8i0dGTO`8l0_GQOe8o0ZGnNj8R1UGkNo8U1PGhNT9W1lFfNX9Z1gFcN]9]1T2O3M3M3M3M2N3L4M3M;EleY1"}, "label": "the bottle"}]} {"id": 226046, "image": "COCO_train2014_000000226046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a soup served with hamburg on atable\"."}], "task": "refering", "answer_template": "The \"a soup served with hamburg on atable\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 231, 232, 233, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260], "bbox": [0.0050625, 0.3303333333333333, 0.412921875, 0.6943750000000001], "iscrowd": 0, "area": 33871.7632, "rle": {"size": [480, 640], "counts": "bd13f>9G8G9H9L3N3N1N2N3M2N3M2N2N3M2N3M2N2N3M2N2N3M2N3M2N2O2M2N3M2O1N3N1O1N3N1N101O1N101N2O0O2O1O0O2O1N101O0O2O1N101N2O001N2O0O2O1O0O101N100O2O000O2O0O101O0O101N101N10001O00001O000O2O00001O00001O000O2O00001O00001O000O2O001O00001O00001N10001O00001O00001O0O100000000000O1000O100000O1000000000O010000000000O100000O100000O10000000000O10000000000O2O000000000O1000001O000O1000000000001N100000000O1O1O2N1O1O1O1O1O1N2O1O2N1O1O1O1O2N1O2N1O2M2O2N1O2N1O2N2N1O2N1O2N1N3N1O2N1O2N2N1O2N1O2N1O2N1N3N1O2N2N1O2N3M3L4K5K5K5K5J6K5K5K5K5K5J5L5K5K\\Q`5"}, "label": "a soup served with hamburg on atable"}]} {"id": 226046, "image": "COCO_train2014_000000226046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown bowl of soup\"."}], "task": "refering", "answer_template": "The \"brown bowl of soup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 122, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 231, 232, 233, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260], "bbox": [0.0050625, 0.3303333333333333, 0.412921875, 0.6943750000000001], "iscrowd": 0, "area": 33871.7632, "rle": {"size": [480, 640], "counts": "bd13f>9G8G9H9L3N3N1N2N3M2N3M2N2N3M2N3M2N2N3M2N2N3M2N3M2N2O2M2N3M2O1N3N1O1N3N1N101O1N101N2O0O2O1O0O2O1N101O0O2O1N101N2O001N2O0O2O1O0O101N100O2O000O2O0O101O0O101N101N10001O00001O000O2O00001O00001O000O2O00001O00001O000O2O001O00001O00001N10001O00001O00001O0O100000000000O1000O100000O1000000000O010000000000O100000O100000O10000000000O10000000000O2O000000000O1000001O000O1000000000001N100000000O1O1O2N1O1O1O1O1O1N2O1O2N1O1O1O1O2N1O2N1O2M2O2N1O2N1O2N2N1O2N1O2N1N3N1O2N1O2N2N1O2N1O2N1O2N1N3N1O2N2N1O2N3M3L4K5K5K5K5J6K5K5K5K5K5J5L5K5K\\Q`5"}, "label": "brown bowl of soup"}]} {"id": 226046, "image": "COCO_train2014_000000226046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich with a knife embedded on it\"."}], "task": "refering", "answer_template": "The \"a sandwich with a knife embedded on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 177, 178, 179, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 333, 334, 335, 336, 337, 338], "bbox": [0.3656875, 0.3718958333333333, 0.843109375, 0.8820625], "iscrowd": 0, "area": 52842.52445000002, "rle": {"size": [480, 640], "counts": "jo]33g>7J6K5N2N2N2N2M3N2N2N2N2J6H9G8J6K6J54MM3K4M4K4L5L8G;F5K4K6K5J6K3L4M3M3N2M3N2N2M3N2N2N2N2M3N2N1O1O100O1O1O100O1O2O0O1O100O2N2O1N101N2O1N102M2O1N3N1N2O1N3N1N2O2M2O1O1N3N1N2O1O2M2O0O10000O10000000000O100000000000000000000O10000000000000000000000O100000000000000000000000O1000O0100O10000O100O100O10000O100O10O10O100O100SJ\\H^4d7`K_H^4b7aK_H_4a7`KaH^4`7aKbH^4^7`KeH_4[7`KfH_4[7`KgH_4Y7`KiH^4X7aKiH_4W7_KlH`4T7_KnH_4S7`KoH_4Q7`KPI`4P7_KRI_4o6_KTI`4l6_KUI`4l6_KVI`4j6_KXI`4h6^K[Ia4e6^K\\Ib4e6\\K]Ic4c6\\K_Ib4b6]K_Ic4a6[KbId4^6[KdId4\\6[KfId4Z6[KgIe4Y6ZKiIe4W6YKlIf4T6YKmIg4S6XKoIg4Q6XKQJf4P6YKRJf4n5XKTJh4S8O1O001O1O1O001N2O001O1O001O1O1O0O2O\\NhKaHW4^7kKaHU4^7mKbHR4]7PLcHo3\\7SLdHl3[7VLdHj3[7YLdHf3[7\\LeHc3Z7_LeH`3[7bLeH]3Z7eLeH[3Z7hLcHY3\\7iLbHX3]7jLaHW3^7kL`HV3_7lL_HU3`7mL^HT3a7oL\\HQ3d7QMZHP3e7RMYHo2g7RMVHo2k7RMSHo2n7QMPHP3P8RMmGo2S8RMkGn2W8RMgGo2Y8RMeGo2\\8QMbGP3^8QM`Go2a8SM\\Gn2e8f10O101O0O2O000O2O001N10001N101O0O101O0O101N1N3M2N2N3M2N3L3N2N3M2cNXF`Mj9]2aFYMb9e2bFVMa9h2cFSM_9l2dFoL`9o2dFlL^9S3Y1M3N1O2N1O2N1N3M3M2N3N1N4L3M3M3M4M2M3M3M4L4L6J7I6K5JYd^1"}, "label": "a sandwich with a knife embedded on it"}]} {"id": 226046, "image": "COCO_train2014_000000226046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich with a knife in it\"."}], "task": "refering", "answer_template": "The \"a sandwich with a knife in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 177, 178, 179, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 333, 334, 335, 336, 337, 338], "bbox": [0.3656875, 0.3718958333333333, 0.843109375, 0.8820625], "iscrowd": 0, "area": 52842.52445000002, "rle": {"size": [480, 640], "counts": "jo]33g>7J6K5N2N2N2N2M3N2N2N2N2J6H9G8J6K6J54MM3K4M4K4L5L8G;F5K4K6K5J6K3L4M3M3N2M3N2N2M3N2N2N2N2M3N2N1O1O100O1O1O100O1O2O0O1O100O2N2O1N101N2O1N102M2O1N3N1N2O1N3N1N2O2M2O1O1N3N1N2O1O2M2O0O10000O10000000000O100000000000000000000O10000000000000000000000O100000000000000000000000O1000O0100O10000O100O100O10000O100O10O10O100O100SJ\\H^4d7`K_H^4b7aK_H_4a7`KaH^4`7aKbH^4^7`KeH_4[7`KfH_4[7`KgH_4Y7`KiH^4X7aKiH_4W7_KlH`4T7_KnH_4S7`KoH_4Q7`KPI`4P7_KRI_4o6_KTI`4l6_KUI`4l6_KVI`4j6_KXI`4h6^K[Ia4e6^K\\Ib4e6\\K]Ic4c6\\K_Ib4b6]K_Ic4a6[KbId4^6[KdId4\\6[KfId4Z6[KgIe4Y6ZKiIe4W6YKlIf4T6YKmIg4S6XKoIg4Q6XKQJf4P6YKRJf4n5XKTJh4S8O1O001O1O1O001N2O001O1O001O1O1O0O2O\\NhKaHW4^7kKaHU4^7mKbHR4]7PLcHo3\\7SLdHl3[7VLdHj3[7YLdHf3[7\\LeHc3Z7_LeH`3[7bLeH]3Z7eLeH[3Z7hLcHY3\\7iLbHX3]7jLaHW3^7kL`HV3_7lL_HU3`7mL^HT3a7oL\\HQ3d7QMZHP3e7RMYHo2g7RMVHo2k7RMSHo2n7QMPHP3P8RMmGo2S8RMkGn2W8RMgGo2Y8RMeGo2\\8QMbGP3^8QM`Go2a8SM\\Gn2e8f10O101O0O2O000O2O001N10001N101O0O101O0O101N1N3M2N2N3M2N3L3N2N3M2cNXF`Mj9]2aFYMb9e2bFVMa9h2cFSM_9l2dFoL`9o2dFlL^9S3Y1M3N1O2N1O2N1N3M3M2N3N1N4L3M3M3M4M2M3M3M4L4L6J7I6K5JYd^1"}, "label": "a sandwich with a knife in it"}]} {"id": 226046, "image": "COCO_train2014_000000226046.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person sitting at the table of food\"."}], "task": "refering", "answer_template": "The \"the person sitting at the table of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 12, 13, 14, 24, 25, 26, 27, 28, 29, 30, 31, 32, 34, 35, 36, 48, 49, 50, 53, 54, 55, 57, 58], "bbox": [0.04278125, 0.0, 0.631875, 0.169375], "iscrowd": 0, "area": 14797.5894, "rle": {"size": [480, 640], "counts": "Qe<1o>9G3M2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N2N2N2N1O2N2N2N2N1O6J0000000000000000000000000000001O2N1O1O1O1O000000O10000000000000000000000O1K5N2000000O1000000000000000000000000O1000000000000000000000000O12N2N3M3M2N3M2N0000XO[CUOeVCBj in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 12, 13, 14, 24, 25, 26, 27, 28, 29, 30, 31, 32, 34, 35, 36, 48, 49, 50, 53, 54, 55, 57, 58], "bbox": [0.04278125, 0.0, 0.631875, 0.169375], "iscrowd": 0, "area": 14797.5894, "rle": {"size": [480, 640], "counts": "Qe<1o>9G3M2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N2N2N2N1O2N2N2N2N1O6J0000000000000000000000000000001O2N1O1O1O1O000000O10000000000000000000000O1K5N2000000O1000000000000000000000000O1000000000000000000000000O12N2N3M3M2N3M2N0000XO[CUOeVCBj in this image.", "objects": [{"patches": [77, 78, 79, 94, 95, 96, 111, 112, 113, 128, 129, 130, 145, 146, 147, 162, 163, 164, 179, 180, 181, 196, 197, 213, 214], "bbox": [0.542112970711297, 0.1875, 0.7029707112970711, 0.567546875], "iscrowd": 0, "area": 10928.65845, "rle": {"size": [640, 478], "counts": "^SR53ic07kLNQB7n=JnA9S>GiA=V>DfA`0Z>@bAc0^>^O^Af0a>[O[Ai0e>WOWAl0i>UOSAo0Y1@j9AiDS1Y1@m9^OfDU1Z1Ao9[OcDX1Z1AR:XOaDZ1\\1_OS:WO`DZ1]1_OS:XO_DZ1]1_OS:XO^DZ1^1@S:CRDn0j1@S:EPDk0m1AS:GlCi0b0ZN^O7Ko0Y in this image.", "objects": [{"patches": [77, 78, 79, 94, 95, 96, 111, 112, 113, 128, 129, 130, 145, 146, 147, 162, 163, 164, 179, 180, 181, 196, 197, 213, 214], "bbox": [0.542112970711297, 0.1875, 0.7029707112970711, 0.567546875], "iscrowd": 0, "area": 10928.65845, "rle": {"size": [640, 478], "counts": "^SR53ic07kLNQB7n=JnA9S>GiA=V>DfA`0Z>@bAc0^>^O^Af0a>[O[Ai0e>WOWAl0i>UOSAo0Y1@j9AiDS1Y1@m9^OfDU1Z1Ao9[OcDX1Z1AR:XOaDZ1\\1_OS:WO`DZ1]1_OS:XO_DZ1]1_OS:XO^DZ1^1@S:CRDn0j1@S:EPDk0m1AS:GlCi0b0ZN^O7Ko0Y in this image.", "objects": [{"patches": [183, 202, 203, 221, 222, 223, 240, 241, 242, 243, 260, 261, 262, 279, 280, 281, 298, 299, 300, 317, 318, 319], "bbox": [0.6253119092627599, 0.413109375, 0.8427977315689982, 0.7456875000000001], "iscrowd": 0, "area": 9107.035349999996, "rle": {"size": [640, 529], "counts": "^T_63lc05L4L4K5L4K2O2O001N10001O1N2O2M3M2N3M2O2M2N3M2N3M2N3MYNR^O_1ka0[N\\^Of1oa04M2M4L4M2M4L3N2M3M3M3N2M3RAQMdf0iASOX>l0oAmNR>R1UBgNk=Y1\\B`Ne=_1dBXN]=g1_2O1O1O1O1O001O1O1O000013L4M3L4L3N2M2N3[_OjM[?W2^@PN`?S2Y@RNg?o1R@XNl?k1l_O[NS`0j20L3M3M4L3M4K4M4Lk_OXMY?\\2PAgMP?h1]A\\Nc>m1PAVNQ?R2d@QN\\?W2X@mMh?[2m_OhMS`0R32O00000L4H8H8H8G9Gm`c1"}, "label": "a small giraffe is standing side of the thorn tree"}]} {"id": 348935, "image": "COCO_train2014_000000348935.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"one giraffe standing alone\"."}], "task": "refering", "answer_template": "The \"one giraffe standing alone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [183, 202, 203, 221, 222, 223, 240, 241, 242, 243, 260, 261, 262, 279, 280, 281, 298, 299, 300, 317, 318, 319], "bbox": [0.6253119092627599, 0.413109375, 0.8427977315689982, 0.7456875000000001], "iscrowd": 0, "area": 9107.035349999996, "rle": {"size": [640, 529], "counts": "^T_63lc05L4L4K5L4K2O2O001N10001O1N2O2M3M2N3M2O2M2N3M2N3M2N3MYNR^O_1ka0[N\\^Of1oa04M2M4L4M2M4L3N2M3M3M3N2M3RAQMdf0iASOX>l0oAmNR>R1UBgNk=Y1\\B`Ne=_1dBXN]=g1_2O1O1O1O1O001O1O1O000013L4M3L4L3N2M2N3[_OjM[?W2^@PN`?S2Y@RNg?o1R@XNl?k1l_O[NS`0j20L3M3M4L3M4K4M4Lk_OXMY?\\2PAgMP?h1]A\\Nc>m1PAVNQ?R2d@QN\\?W2X@mMh?[2m_OhMS`0R32O00000L4H8H8H8G9Gm`c1"}, "label": "one giraffe standing alone"}]} {"id": 348935, "image": "COCO_train2014_000000348935.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"small giraffe on left\"."}], "task": "refering", "answer_template": "The \"small giraffe on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 175, 176, 195, 196, 214, 215, 216, 233, 234, 235, 236, 252, 253, 254, 255, 256, 271, 272, 273, 274, 275, 290, 291, 293, 294, 309, 310, 313, 329], "bbox": [0.25083175803402646, 0.38315625, 0.5073345935727788, 0.749359375], "iscrowd": 0, "area": 10546.606349999995, "rle": {"size": [640, 529], "counts": "h[c23lc03L4O0O2O0O2O0000213K2J3M3M2N31O13K8F:]_O`NS>j1cA[N[>m1[AWNb>Q2UATNh>T2n@PNY?n1^@VN`?R2V@SNf?W3L6IPLa@h3Y?ULm@n3m0TL_Q4fAlK\\>j3oARLR>g3XBVLi=a3bB\\L_=[3g1H9H7H9G8O20O01O1O001O1O3M2N3M3M3M2M2O001N10000O100O10001N100002N2N1O2N2O0O2N2N2N1O2N2N2O0O2N2N2N1O2N2N2O0O2N2N2O1O1O100O1O1O100O1O1OgNR_OBl`00c_O9R`0ZO\\@l0]?eNTA5bN=\\b0^Oh]O?\\b0[Oi]Oc0Yb0YOk]Od0kb0M3N2M3M3N2M3M3NO0`eR5"}, "label": "small giraffe on left"}]} {"id": 348935, "image": "COCO_train2014_000000348935.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe most close and facing the camera\"."}], "task": "refering", "answer_template": "The \"the giraffe most close and facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 175, 176, 195, 196, 214, 215, 216, 233, 234, 235, 236, 252, 253, 254, 255, 256, 271, 272, 273, 274, 275, 290, 291, 293, 294, 309, 310, 313, 329], "bbox": [0.25083175803402646, 0.38315625, 0.5073345935727788, 0.749359375], "iscrowd": 0, "area": 10546.606349999995, "rle": {"size": [640, 529], "counts": "h[c23lc03L4O0O2O0O2O0000213K2J3M3M2N31O13K8F:]_O`NS>j1cA[N[>m1[AWNb>Q2UATNh>T2n@PNY?n1^@VN`?R2V@SNf?W3L6IPLa@h3Y?ULm@n3m0TL_Q4fAlK\\>j3oARLR>g3XBVLi=a3bB\\L_=[3g1H9H7H9G8O20O01O1O001O1O3M2N3M3M3M2M2O001N10000O100O10001N100002N2N1O2N2O0O2N2N2N1O2N2N2O0O2N2N2N1O2N2N2O0O2N2N2O1O1O100O1O1O100O1O1OgNR_OBl`00c_O9R`0ZO\\@l0]?eNTA5bN=\\b0^Oh]O?\\b0[Oi]Oc0Yb0YOk]Od0kb0M3N2M3M3N2M3M3NO0`eR5"}, "label": "the giraffe most close and facing the camera"}]} {"id": 29473, "image": "COCO_train2014_000000029473.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bowl of rasberries\"."}], "task": "refering", "answer_template": "The \"the bowl of rasberries\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 36, 37, 38, 39, 40, 41, 42, 58, 59, 60, 61, 62, 63, 64, 65, 66, 81, 82, 83, 84, 85, 86, 87, 88, 89, 104, 105, 106, 107, 108, 109, 110, 111, 112, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157], "bbox": [0.53040625, 0.04504166666666667, 0.8800625, 0.4166666666666667], "iscrowd": 0, "area": 30470.631999999998, "rle": {"size": [480, 640], "counts": "ZPo42f>e0[O;J6J6J5K8H8I7J6J6I6K3M3M2N3M3M2M4M3M3M2N3M3M2O2M3M2O2M3N1N3M2O1N2O1N2N010O010O1O010O0010O010O01O010O01000O010O0100O010O10O010O01000O010O0100O010O010O10O10O0100000000O1000O100000O10000000O1000O1000000000O0100000000000O10000000O100000O100000000000000O100000000000000000001O00000010O00000001O0000001O0000001O0000001O0000000010O000001O0000001O0000001O1O001O010O10O010O01O10O010O01O10O0100O001O0O2M3N1O2M2O2M4M2M3N2M3N2N2M3N2M3M3K6J5K7H9H8H8H8G9D<_ObPT1"}, "label": "the bowl of rasberries"}]} {"id": 29473, "image": "COCO_train2014_000000029473.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bowl in which yoghurt and strawberries are kept\"."}], "task": "refering", "answer_template": "The \"the bowl in which yoghurt and strawberries are kept\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 36, 37, 38, 39, 40, 41, 42, 58, 59, 60, 61, 62, 63, 64, 65, 66, 81, 82, 83, 84, 85, 86, 87, 88, 89, 104, 105, 106, 107, 108, 109, 110, 111, 112, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157], "bbox": [0.53040625, 0.04504166666666667, 0.8800625, 0.4166666666666667], "iscrowd": 0, "area": 30470.631999999998, "rle": {"size": [480, 640], "counts": "ZPo42f>e0[O;J6J6J5K8H8I7J6J6I6K3M3M2N3M3M2M4M3M3M2N3M3M2O2M3M2O2M3N1N3M2O1N2O1N2N010O010O1O010O0010O010O01O010O01000O010O0100O010O10O010O01000O010O0100O010O010O10O10O0100000000O1000O100000O10000000O1000O1000000000O0100000000000O10000000O100000O100000000000000O100000000000000000001O00000010O00000001O0000001O0000001O0000001O0000000010O000001O0000001O0000001O1O001O010O10O010O01O10O010O01O10O0100O001O0O2M3N1O2M2O2M4M2M3N2M3N2N2M3N2M3M3K6J5K7H9H8H8H8G9D<_ObPT1"}, "label": "the bowl in which yoghurt and strawberries are kept"}]} {"id": 29473, "image": "COCO_train2014_000000029473.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bread that it vertical in the container\"."}], "task": "refering", "answer_template": "The \"the bread that it vertical in the container\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 72, 73, 74, 75, 76, 95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123], "bbox": [0.1383125, 0.15766666666666668, 0.39934375, 0.3576666666666667], "iscrowd": 0, "area": 10279.319400000002, "rle": {"size": [480, 640], "counts": "ojY12m>2O1N2N2N2N2\\OEaB<[=GdB;Y=HeB:W=JgB8V=KhB7T=MkB4R=NmB4o<0oB2m<2QC1k<2SC0i<3VCOh in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 27, 28, 29, 30, 31, 32, 33, 34, 51, 52, 53, 54, 55, 56, 57, 75, 76, 77, 78, 79, 100, 101], "bbox": [0.2095625, 0.0, 0.5143125, 0.28275], "iscrowd": 0, "area": 17366.9497, "rle": {"size": [480, 640], "counts": "Vkn1:e>3N2M3M3N2M3M3N1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O0O2N101N1O2O0O2N2O0O2N101N1O2O1N1O2O0O2N101N1O2O1N1O2O0O2O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O00100O001O001O001O1OO1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N4M2N2N2M3N2N2N3M2M3N2N2N2N1N2O2N2L4I7I7I7Haha4"}, "label": "the bowl with bread"}]} {"id": 29473, "image": "COCO_train2014_000000029473.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich with thick wheat bread in a white bowl\"."}], "task": "refering", "answer_template": "The \"a sandwich with thick wheat bread in a white bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 27, 28, 29, 30, 31, 32, 33, 34, 51, 52, 53, 54, 55, 56, 57, 75, 76, 77, 78, 79, 100, 101], "bbox": [0.2095625, 0.0, 0.5143125, 0.28275], "iscrowd": 0, "area": 17366.9497, "rle": {"size": [480, 640], "counts": "Vkn1:e>3N2M3M3N2M3M3N1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N2O0O2N101N1O2O0O2N2O0O2N101N1O2O1N1O2O0O2N101N1O2O1N1O2O0O2O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O001O1O001O001O001O1O001O001O001O00100O001O001O001O1OO1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N4M2N2N2M3N2N2N3M2M3N2N2N2N1N2O2N2L4I7I7I7Haha4"}, "label": "a sandwich with thick wheat bread in a white bowl"}]} {"id": 29473, "image": "COCO_train2014_000000029473.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white container with steamed carrots in it\"."}], "task": "refering", "answer_template": "The \"a white container with steamed carrots in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 323, 324, 325, 326, 327, 347, 348, 349, 350], "bbox": [0.03, 0.44, 0.41823437500000005, 0.9294166666666667], "iscrowd": 0, "area": 28838.0485, "rle": {"size": [480, 640], "counts": "PY92k>4L5K5K5K5K5J5L5K5K5K5K5M2M3N1O1O1O1O1O1N2O1O1O1O1B>kNU100O100O10004K9H6I10000O100J6G9J60000O1000000O1000000O1000004K6K3M000O2O00001N101O000O2O00001N101O0O1000000O1000000O10000N2O1O0O2O1N2N2N2N2N2N2N2N2N2N2N2NVMbHVN\\7l1nHjMR7V2WIaMi6_2`IXM`6g2bIXM^6h2dIVM\\6j2eIUM[6k2gIRMZ6m2hIRMX6n2jIPMV6P3kIoLU6Q3mImLS6R3oImLQ6S3QJkLo5U3RJjLn5U3UJiLk5W3WJgLi5Y3XJfLh5Z3ZJdLf5[3]JcLc5]3^JbLb5^3`J`L`5`3aJ_L_5`3dJ\\L^5d3dJWL_5i3cJSL_5m3eJlK^5T4g2O100000000O1000000O100000000O100000000O100O1O1O2N1O2N1O1O2N1I7F;D;K6M2O1O2N1N3N1O1N3N1O2N1O101O000O2O001N10001O0O2O000O2O001O00001O1O001O001O001O1O001O001O001O1O001O001O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O10O01O3M7I7I^b^5"}, "label": "a white container with steamed carrots in it"}]} {"id": 29473, "image": "COCO_train2014_000000029473.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white plastic bowl of carrots and hummus next to white bowls of other food\"."}], "task": "refering", "answer_template": "The \"white plastic bowl of carrots and hummus next to white bowls of other food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 212, 213, 214, 215, 216, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 323, 324, 325, 326, 327, 347, 348, 349, 350], "bbox": [0.03, 0.44, 0.41823437500000005, 0.9294166666666667], "iscrowd": 0, "area": 28838.0485, "rle": {"size": [480, 640], "counts": "PY92k>4L5K5K5K5K5J5L5K5K5K5K5M2M3N1O1O1O1O1O1N2O1O1O1O1B>kNU100O100O10004K9H6I10000O100J6G9J60000O1000000O1000000O1000004K6K3M000O2O00001N101O000O2O00001N101O0O1000000O1000000O10000N2O1O0O2O1N2N2N2N2N2N2N2N2N2N2N2NVMbHVN\\7l1nHjMR7V2WIaMi6_2`IXM`6g2bIXM^6h2dIVM\\6j2eIUM[6k2gIRMZ6m2hIRMX6n2jIPMV6P3kIoLU6Q3mImLS6R3oImLQ6S3QJkLo5U3RJjLn5U3UJiLk5W3WJgLi5Y3XJfLh5Z3ZJdLf5[3]JcLc5]3^JbLb5^3`J`L`5`3aJ_L_5`3dJ\\L^5d3dJWL_5i3cJSL_5m3eJlK^5T4g2O100000000O1000000O100000000O100000000O100O1O1O2N1O2N1O1O2N1I7F;D;K6M2O1O2N1N3N1O1N3N1O2N1O101O000O2O001N10001O0O2O000O2O001O00001O1O001O001O001O1O001O001O001O1O001O001O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O10O01O3M7I7I^b^5"}, "label": "white plastic bowl of carrots and hummus next to white bowls of other food"}]} {"id": 29473, "image": "COCO_train2014_000000029473.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl filled with fried greens\"."}], "task": "refering", "answer_template": "The \"a bowl filled with fried greens\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 357, 358, 359, 360, 361, 362, 363, 364, 365, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.48034375, 0.3656041666666667, 0.985953125, 0.9813333333333334], "iscrowd": 0, "area": 76838.58914999999, "rle": {"size": [480, 640], "counts": "fV`41h>a0@`0_Oa0@?A`0_O;F8J6M2M4M2M4M3L3N3M3L3N3L3N3L4M2N3L4M2M4M2M4M3M2M4M3L4M2M4M3M3L3N3L4M3L3N2N3N1N2O2M2N2O2M2O1N3M2O1N101N2N2O0O2N2O0O2O1N2N101N2O1N1O2O1N2O0O2N2O1N101N2N2O0O2O1N2N101O0000000O10001O0000000O100000001O000O10000000001O0O1000000000001N1000000000000O2O00000000000O10001O0000000O100000000000000O1000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000001O001O001O001O001O001O001N2N101N101N1O2O0O2O0O2N101N101N2N101N1O2O0O2O0O2N101N101N1O2O1N101N1O2O0O2N1N3N1N3N1N3M2O2M3N1N3N1N3N1N3N1N3N2M3N3L3N3L3N3L3N2M4L3N3L3N3L3N2M4M2M4M2M4M2M4M2M3N3J5L5J5L5K4K6K4K5L5J:G8G:G9F9H9G8G:G9F\\m3"}, "label": "a bowl filled with fried greens"}]} {"id": 29473, "image": "COCO_train2014_000000029473.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cooked zucchini\"."}], "task": "refering", "answer_template": "The \"cooked zucchini\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 357, 358, 359, 360, 361, 362, 363, 364, 365, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.48034375, 0.3656041666666667, 0.985953125, 0.9813333333333334], "iscrowd": 0, "area": 76838.58914999999, "rle": {"size": [480, 640], "counts": "fV`41h>a0@`0_Oa0@?A`0_O;F8J6M2M4M2M4M3L3N3M3L3N3L3N3L4M2N3L4M2M4M2M4M3M2M4M3L4M2M4M3M3L3N3L4M3L3N2N3N1N2O2M2N2O2M2O1N3M2O1N101N2N2O0O2N2O0O2O1N2N101N2O1N1O2O1N2O0O2N2O1N101N2N2O0O2O1N2N101O0000000O10001O0000000O100000001O000O10000000001O0O1000000000001N1000000000000O2O00000000000O10001O0000000O100000000000000O1000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000001O001O001O001O001O001O001N2N101N101N1O2O0O2O0O2N101N101N2N101N1O2O0O2O0O2N101N101N1O2O1N101N1O2O0O2N1N3N1N3N1N3M2O2M3N1N3N1N3N1N3N1N3N2M3N3L3N3L3N3L3N2M4L3N3L3N3L3N2M4M2M4M2M4M2M4M2M3N3J5L5J5L5K4K6K4K5L5J:G8G:G9F9H9G8G:G9F\\m3"}, "label": "cooked zucchini"}]} {"id": 217893, "image": "COCO_train2014_000000217893.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey monitor that is off\"."}], "task": "refering", "answer_template": "The \"a grey monitor that is off\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 210, 211, 212], "bbox": [0.13314062499999998, 0.24775, 0.38764062499999996, 0.5511250000000001], "iscrowd": 0, "area": 17789.1614, "rle": {"size": [480, 640], "counts": "[oW12n>7I6J7I7I6I8I6J7I6J7I6J7H8I6J7I6J7I6J1N2O001O00001O000O2O001OO100000O01000000O1000O10O1000000O10O1000O1000000O0100000O100000O0100000000O10O1000O1000000O0100000O100000O01000000O1000O10O1000000O10O1000O1000000O0100000O100000O10O1000000O10O1000O1000000O0100000O100000O01000000O1000O10O10005K6I7J6J5K6I7J6J6J6I7J5K6J6J6I7J6J5K6Il`g5"}, "label": "a grey monitor that is off"}]} {"id": 217893, "image": "COCO_train2014_000000217893.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a flat screen computer monitor with audio plugs inserted in the front panel\"."}], "task": "refering", "answer_template": "The \"a flat screen computer monitor with audio plugs inserted in the front panel\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 210, 211, 212], "bbox": [0.13314062499999998, 0.24775, 0.38764062499999996, 0.5511250000000001], "iscrowd": 0, "area": 17789.1614, "rle": {"size": [480, 640], "counts": "[oW12n>7I6J7I7I6I8I6J7I6J7I6J7H8I6J7I6J7I6J1N2O001O00001O000O2O001OO100000O01000000O1000O10O1000000O10O1000O1000000O0100000O100000O0100000000O10O1000O1000000O0100000O100000O01000000O1000O10O1000000O10O1000O1000000O0100000O100000O10O1000000O10O1000O1000000O0100000O100000O01000000O1000O10O10005K6I7J6J5K6I7J6J6J6I7J5K6J6J6I7J6J5K6Il`g5"}, "label": "a flat screen computer monitor with audio plugs inserted in the front panel"}]} {"id": 217893, "image": "COCO_train2014_000000217893.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a crt monitor with a black & white photo of bruce lee displayed\"."}], "task": "refering", "answer_template": "The \"a crt monitor with a black & white photo of bruce lee displayed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 129, 147, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245], "bbox": [0.42982812499999995, 0.347875, 0.6650937499999999, 0.6379791666666667], "iscrowd": 0, "area": 19197.131900000004, "rle": {"size": [480, 640], "counts": "]RQ49g>?@a0_Oa0@`0_Oa0_Oa0@`0@8H000000000000000000000000000O10000000001O000000000000000000000000000000000O100000001O000000000000jK]Eo3j:00000000000000000000000000001O000000000SEoKk:Q4UEoKk:Q4UEoKk:Q4UEoKk:R4100000000000001O00000000000O100000000000000000000000000000I70O1000000000000000000000000O1O100O1O1O100O1O2O001O001O001O001O001O001O1O001O00_TT3"}, "label": "a crt monitor with a black & white photo of bruce lee displayed"}]} {"id": 217893, "image": "COCO_train2014_000000217893.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white desktop monitor with brucelee photo on it\"."}], "task": "refering", "answer_template": "The \"a white desktop monitor with brucelee photo on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 129, 147, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 217, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245], "bbox": [0.42982812499999995, 0.347875, 0.6650937499999999, 0.6379791666666667], "iscrowd": 0, "area": 19197.131900000004, "rle": {"size": [480, 640], "counts": "]RQ49g>?@a0_Oa0@`0_Oa0_Oa0@`0@8H000000000000000000000000000O10000000001O000000000000000000000000000000000O100000001O000000000000jK]Eo3j:00000000000000000000000000001O000000000SEoKk:Q4UEoKk:Q4UEoKk:Q4UEoKk:R4100000000000001O00000000000O100000000000000000000000000000I70O1000000000000000000000000O1O100O1O1O100O1O2O001O001O001O001O001O001O1O001O00_TT3"}, "label": "a white desktop monitor with brucelee photo on it"}]} {"id": 365351, "image": "COCO_train2014_000000365351.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two hand towels hanging in the rod\"."}], "task": "refering", "answer_template": "The \"two hand towels hanging in the rod\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242], "bbox": [0.34953124999999996, 0.3299166666666667, 0.566984375, 0.6163958333333334], "iscrowd": 0, "area": 18283.283400000004, "rle": {"size": [480, 640], "counts": "cUY3P1o=^1bNY1gN1O1O1O1O1N2O1O1O1O1O1O1O001O1O1O1000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000O100000000000000000000000000000000001O2N3M2N3M3M2N3M2kN\\1aNZfQ4"}, "label": "two hand towels hanging in the rod"}]} {"id": 365351, "image": "COCO_train2014_000000365351.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"silver oven with two towels hanging on it\"."}], "task": "refering", "answer_template": "The \"silver oven with two towels hanging on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242], "bbox": [0.34953124999999996, 0.3299166666666667, 0.566984375, 0.6163958333333334], "iscrowd": 0, "area": 18283.283400000004, "rle": {"size": [480, 640], "counts": "cUY3P1o=^1bNY1gN1O1O1O1O1N2O1O1O1O1O1O1O001O1O1O1000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000O100000000000000000000000000000000001O2N3M2N3M3M2N3M2kN\\1aNZfQ4"}, "label": "silver oven with two towels hanging on it"}]} {"id": 95018, "image": "COCO_train2014_000000095018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child with a black shirt and blue helmet playing baseball\"."}], "task": "refering", "answer_template": "The \"a child with a black shirt and blue helmet playing baseball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 167, 168, 169, 170, 171, 191, 192, 193, 194, 215, 216, 217, 238, 239, 240, 261, 262, 263, 283, 284, 286, 287, 310], "bbox": [0.28875, 0.3804867256637168, 0.495875, 0.8286061946902654], "iscrowd": 0, "area": 9670.845449999999, "rle": {"size": [452, 640], "counts": "`ka21R>2O2M2O2M2O1O2N1O1O2N1O1O2N1O2N100O10000O10O010000O100O100O1YFZOo5e0PJ\\OP6e0nI\\OR6e0kI]OU6d0jI\\OV6e0iI[OW6f0hIZOX6f0hIZOX6f0hIZOX6e0hI\\OX6d0aHJ\\NBS9d0VH;aNQOY9c0THR1l7nNQHU1o7jNoGZ1o7gNmG]1S8bNkGa1U8_NhGd1X8[NfGh1Z8XNcGk1]8TNaGo1_8PN\\GV2d8iMRGb2n8aMlFb2T9W1O100OM2gNhF^Mb9c2n01L4F:00O001O100O01010O2O001oEoKm9R4RFoKm9T401O5K5J2O1O1O1O10O010000O1000000O01cLfFo1Z9aM_GV2`8fMnGQ2R8iM`Hk1`7oMmHj1n9I7Ic0\\O8I1O1O11N2O2M3N1N000100O001O1N101O1O01O2NO100O1O010O1O2O0N3M3N[W^4"}, "label": "a child with a black shirt and blue helmet playing baseball"}]} {"id": 95018, "image": "COCO_train2014_000000095018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child swinging a baseball bat\"."}], "task": "refering", "answer_template": "The \"a child swinging a baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [146, 147, 148, 167, 168, 169, 170, 171, 191, 192, 193, 194, 215, 216, 217, 238, 239, 240, 261, 262, 263, 283, 284, 286, 287, 310], "bbox": [0.28875, 0.3804867256637168, 0.495875, 0.8286061946902654], "iscrowd": 0, "area": 9670.845449999999, "rle": {"size": [452, 640], "counts": "`ka21R>2O2M2O2M2O1O2N1O1O2N1O1O2N1O2N100O10000O10O010000O100O100O1YFZOo5e0PJ\\OP6e0nI\\OR6e0kI]OU6d0jI\\OV6e0iI[OW6f0hIZOX6f0hIZOX6f0hIZOX6e0hI\\OX6d0aHJ\\NBS9d0VH;aNQOY9c0THR1l7nNQHU1o7jNoGZ1o7gNmG]1S8bNkGa1U8_NhGd1X8[NfGh1Z8XNcGk1]8TNaGo1_8PN\\GV2d8iMRGb2n8aMlFb2T9W1O100OM2gNhF^Mb9c2n01L4F:00O001O100O01010O2O001oEoKm9R4RFoKm9T401O5K5J2O1O1O1O10O010000O1000000O01cLfFo1Z9aM_GV2`8fMnGQ2R8iM`Hk1`7oMmHj1n9I7Ic0\\O8I1O1O11N2O2M3N1N000100O001O1N101O1O01O2NO100O1O010O1O2O0N3M3N[W^4"}, "label": "a child swinging a baseball bat"}]} {"id": 119604, "image": "COCO_train2014_000000119604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow food truck\"."}], "task": "refering", "answer_template": "The \"a yellow food truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 290, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366], "bbox": [0.47190624999999997, 0.47602083333333334, 1.0, 0.9254583333333334], "iscrowd": 0, "area": 48655.54280000001, "rle": {"size": [480, 640], "counts": "aj]41o>2M2N3M2O2M2N3N1N3M2N2L5L3O2M2O001O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N3N1O2N2N2N2M3N2N2N2N1O2M3N2N2N2N2M3N2N1O2N2N2M3N2N2N2NB>L3M4M3L4L4L4M2M4L4H8C=000O10000000000O10000000000O100000002N2M3N2N2N2N2N2M4M2N2N2N2N2M3N2N2N2N2N2M4M3M4L3M3M4K4M3M4L3M3M4K4M4L3M3M4L3L4M4L3M3M4L3L4M4L3M3M2N1N100000001O0O10001O00000O101O0000001N1000001O000O10001O00000O2O000000001N1000001O000O101O0000000O2O0000001O000000001O000O10000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O100000O10000000000001N2O2N2N1O2N2N1O2N1O2N2M2O2N2N1O2N2N1O2N1O2N2M2N3N2M2N3M3N1N3M2N3N2M2N2O1N2N2N101N2N2N2O0O2N2O1N2N1O2O1N2N2O0O2N2N2O1N1O2N2O^F"}, "label": "a yellow food truck"}]} {"id": 119604, "image": "COCO_train2014_000000119604.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a food truck on a city street with many customers\"."}], "task": "refering", "answer_template": "The \"a food truck on a city street with many customers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 288, 290, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 344, 360, 361, 362, 363, 364, 365, 366], "bbox": [0.47190624999999997, 0.47602083333333334, 1.0, 0.9254583333333334], "iscrowd": 0, "area": 48655.54280000001, "rle": {"size": [480, 640], "counts": "aj]41o>2M2N3M2O2M2N3N1N3M2N2L5L3O2M2O001O1O1O1N2O1O1O1O1O1N2O1O1O1O1O1N3N1O2N2N2N2M3N2N2N2N1O2M3N2N2N2N2M3N2N1O2N2N2M3N2N2N2NB>L3M4M3L4L4L4M2M4L4H8C=000O10000000000O10000000000O100000002N2M3N2N2N2N2N2M4M2N2N2N2N2M3N2N2N2N2N2M4M3M4L3M3M4K4M3M4L3M3M4K4M4L3M3M4L3L4M4L3M3M4L3L4M4L3M3M2N1N100000001O0O10001O00000O101O0000001N1000001O000O10001O00000O2O000000001N1000001O000O101O0000000O2O0000001O000000001O000O10000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O10000000000000000000000O100000O10000000000001N2O2N2N1O2N2N1O2N1O2N2M2O2N2N1O2N2N1O2N1O2N2M2N3N2M2N3M3N1N3M2N3N2M2N2O1N2N2N101N2N2N2O0O2N2O1N2N1O2O1N2N2O0O2N2N2O1N1O2N2O^F"}, "label": "a food truck on a city street with many customers"}]} {"id": 62263, "image": "COCO_train2014_000000062263.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra that is behind the front zebra and the elephant\"."}], "task": "refering", "answer_template": "The \"the zebra that is behind the front zebra and the elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 85, 86, 103, 104, 105, 106, 107, 108, 109, 121, 122, 123, 124, 126, 127, 128, 129, 130, 131, 132, 133, 144, 145, 146, 147, 154, 155, 156, 167, 168, 169, 170, 190, 191, 192, 193, 213, 214, 215, 216, 236, 237, 238, 239, 259, 260, 261, 262, 282, 283, 284, 306, 307], "bbox": [0.2714375, 0.2044131455399061, 0.7901093749999999, 0.9212910798122065], "iscrowd": 0, "area": 26886.47285, "rle": {"size": [426, 640], "counts": "kaX2_1k;7I7H7J7I7H7I4K5L4K6K4L4L4L4L4M4K=C=C^1bNg0YO7I2N1O1O1O100O1O1O1OhNX1O1O1O2N1O2XOmHWKU7e4j0L4L4L4L4L4L4L4L3M4L4L4L43XH_La5d3TJiLg5Z3oISMk5Q3kI[Mo5h2hIdMR6_2eIlMU6k4J6J6J5K1O00000000000O10fN^JlJb5R5fJhJZ5V5nJdJR5Z5WK_Ji4`5^KZJb4d5fKVJZ4h5oKQJQ4m5g1O0WKlH\\3V7VL]Ic3e6oKnIj3T6mKZJl3i7I6I7I7IS1nN2M2N3\\Oc003M4M2M3M3M3M3N2M3M3M3N2M3Moo=1ooA2O2M3M2N3M3N0O2N1O1O2O0O1O2N1O1O2O0O1O2N1O101N1O1O2N100O2N1O1O2N100O001O1O100O001O100000O010000O10O10O10000O0100000O01000O1000O01000000O01000O1000O01000000O01000O1000O010000O10O1000O1000O010000O10O1000O1000O010000O10O10O100000O010000000O10000001O00001O0O2O00001O001O00001O1O1O1O1O1O1N2O2N1O1O1O1O1O1O1O1O001O1O1O1N2O001O1O1O001O01O0010O01O1O010O0010O01O1O010O0010O01O5K6K6I7I6K6I7I6JQ1POUbg1"}, "label": "the zebra that is behind the front zebra and the elephant"}]} {"id": 62263, "image": "COCO_train2014_000000062263.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby zebra in back of the other two baby zebras\"."}], "task": "refering", "answer_template": "The \"baby zebra in back of the other two baby zebras\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 142, 143, 144, 165, 166, 167, 187, 188, 189, 190, 210, 211, 212, 213, 215, 233, 234, 237, 238, 256], "bbox": [0.1525625, 0.33314553990610324, 0.36346875, 0.7848356807511736], "iscrowd": 0, "area": 7040.998049999999, "rle": {"size": [426, 640], "counts": "[RY1=b<`0A?@9H7I7000000000000000000000001N1001O2N1O1O2N1O12N100O1O2O0O1O101N10001O00000010O000001O0DhMgEZ2W:gMhE\\2U:eMjE^2T:bMkE`2S:aMlEb2Q:_MnEc2P:^MoEe2n9\\MQFg2m9ZMQFh2m9YMRFj2k9;OPMVFc2i9]MYFb2e9_M\\Fa2b9`M_F_2`9aMcF^2[9cMfF]2Y9cMiF\\2U9dMmF[2R9fMPGY2n8hMSGX2k8iMVGW2h8kMYGT2e8mM\\GR2c8oM_GP2_8RNaGn1]8SNeGl1Y8UNlGf1S8\\NRH_1m7aNXH[1g7fN^H6UOZO\\8`0YJ[Oh5d0]JXOb5i0bJSO]5m0`300N2O1O1N3N1O1O1N2O2N1N2O1O2N1N2O2O000O101O0O101O0O1000Wl33bSL5K4M3N2N2M3N2002O1N2N2N2N1O2N[QY5"}, "label": "baby zebra in back of the other two baby zebras"}]} {"id": 62263, "image": "COCO_train2014_000000062263.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small baby zebra with black and white stripe\"."}], "task": "refering", "answer_template": "The \"a small baby zebra with black and white stripe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 178, 193, 194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 220, 221, 222, 223, 224, 240, 241, 242, 243, 244, 245, 246, 263, 264, 265, 267, 268, 269, 287, 288, 290, 291, 310, 311], "bbox": [0.39215625, 0.2748356807511737, 0.7858281249999999, 0.918169014084507], "iscrowd": 0, "area": 39742.2599, "rle": {"size": [426, 640], "counts": "ibX3f0_<:G9G9G9F:G9I7Ik0VO5J3M3E;N2M2O1O1N2O3M5L4L4L4M2M2N2N2O1O1O0010O00000O101N10000O1000000001O001O3M2N3M2N3M2N3M2Nb1^N00O1O1O001O10O01O001O001N101O0O2O001O1N101O001N101O0O2O001O0O2O001M2H8YOf0O2N2N2N2N1O2O1N2N1O2N2N2O1N1O2N2N2N2O0O2N200000O100000000O100000000O100000000O1000000O1000001O6I8I6J6J7I6I8I4L00O10O010000O10O010000O10O10O100O10O10O100eNbIiK^6Q4nIjKR6o3ZJlKf5o3dJkK]5T4iJgKW5Y4m1O101O0O10]HWLe5h3mIgLT6X3^IUMc6j2oHdMR7Z40O101N100O1O2O0O100O2O0O100O2N100O101N2O2M2N3N1N3N1N3N1N2gNXHmLk7m2[HQMf7k2_HRMd7h2bHVM_7e2gHXM\\7b2jH\\MW7^2PI_MS7[2SIcMo6V2XIgMk6T2X2L5K5K4N3O1O1O0O2O17H10O10O0100O010O010O10O0100O010O1L3F;D in this image.", "objects": [{"patches": [101, 102, 124, 125, 126, 127, 128, 129, 130, 131, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 178, 193, 194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 220, 221, 222, 223, 224, 240, 241, 242, 243, 244, 245, 246, 263, 264, 265, 267, 268, 269, 287, 288, 290, 291, 310, 311], "bbox": [0.39215625, 0.2748356807511737, 0.7858281249999999, 0.918169014084507], "iscrowd": 0, "area": 39742.2599, "rle": {"size": [426, 640], "counts": "ibX3f0_<:G9G9G9F:G9I7Ik0VO5J3M3E;N2M2O1O1N2O3M5L4L4L4M2M2N2N2O1O1O0010O00000O101N10000O1000000001O001O3M2N3M2N3M2N3M2Nb1^N00O1O1O001O10O01O001O001N101O0O2O001O1N101O001N101O0O2O001O0O2O001M2H8YOf0O2N2N2N2N1O2O1N2N1O2N2N2O1N1O2N2N2N2O0O2N200000O100000000O100000000O100000000O1000000O1000001O6I8I6J6J7I6I8I4L00O10O010000O10O010000O10O10O100O10O10O100eNbIiK^6Q4nIjKR6o3ZJlKf5o3dJkK]5T4iJgKW5Y4m1O101O0O10]HWLe5h3mIgLT6X3^IUMc6j2oHdMR7Z40O101N100O1O2O0O100O2O0O100O2N100O101N2O2M2N3N1N3N1N3N1N2gNXHmLk7m2[HQMf7k2_HRMd7h2bHVM_7e2gHXM\\7b2jH\\MW7^2PI_MS7[2SIcMo6V2XIgMk6T2X2L5K5K4N3O1O1O0O2O17H10O10O0100O010O010O10O0100O010O1L3F;D in this image.", "objects": [{"patches": [239, 261, 262, 263, 264, 265, 266, 267, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 330, 331, 332, 333, 337], "bbox": [0.336421875, 0.603375, 0.7209375, 0.8748333333333334], "iscrowd": 0, "area": 18169.623250000004, "rle": {"size": [480, 640], "counts": "mTU31b>?D in this image.", "objects": [{"patches": [239, 261, 262, 263, 264, 265, 266, 267, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 330, 331, 332, 333, 337], "bbox": [0.336421875, 0.603375, 0.7209375, 0.8748333333333334], "iscrowd": 0, "area": 18169.623250000004, "rle": {"size": [480, 640], "counts": "mTU31b>?D in this image.", "objects": [{"patches": [117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288], "bbox": [0.09990625, 0.3946045197740113, 0.9212031249999999, 0.9997457627118643], "iscrowd": 0, "area": 84671.66060000005, "rle": {"size": [354, 640], "counts": "a]f0:[:b0^Oc0]Oc0]Oa0^O?B>B>B>C=J5L5J6K5K4K5L4K6M2O1O1N3N1000001O00000O2O00000000000O1000000000000O1000000000000O1000000000000O10000000000000O1000000000000000000000000000000000000000000001O00000000001O001O001O001O1O001P6mIYg68bXIV1iNX1hN]2cM>B>C0O0000O2O00000O101O0O10001N1O1N200O100000000O2O00000O100000000O100000000O101O00000000000000000000000000000000000000000bNYKbLf4^3[KaLe4^3^K`Lb4`3_K_La4a3aK]L_4c3cK[L]4e3dKZL\\4e3eK[L[4e3eK[L[4e3eK[L[4e3eK[L[4e3eKZL]4d3dK\\L\\4d3dK\\L\\4d3cK]L]4c3cK]L]4c3cK]L]4c3cK]L]4d3bK[L_4e3aK[L_4e3aK[L_4e3aKULe4k3[KoKk4Q4S101O00000000`JmKQ4S4aK[L_4e3aK[L_4e3`K[La4e3_K[La4e3_K[La4e3_K[La4f3^KZLb4f3^KYLc4g3]KYLd4f3\\KZLd4f3\\KZLd4f3\\KZLd4f3\\KYLe4g3[KYLe4g3[KYLe4g3[KYLe4g3[KXLf4h3YKYLg4g3YKYLg4g3YKYLg4g3YKYLg4g3YKXLh4h3XKXLi4g3WKYLi4g3WKYLi4g3WKYLi4g3WKXLj4h3VKXLj4h3VKXLj4h3VKXLj4h3VKXLj4h3VKXLj4h3UKYLk4g3UKYLk4g3UKYLk4g3UKYLk4g3UKYLl4f3TKYLm4g3SKYLm4g3SKYLm4f3TKZLl4f3TKZLl4f3TKZLl4f3TKZLl4f3TKZLl4f3TKZLl4f3TKZLl4f3SK[Lm4e3SK[Lm4e3SK[Ln4d3RK\\Ln4d3RK\\Ln4d3RK\\Ln4d3RK\\Ln4d3RK\\Ln4d3RK[Lo4e3QK[Lo4e3QK[Lo4e3QK[Lo4e3QK[Lo4e3QK[Lo4e3QK[Lo4d3QK]Lo4c3QK]LP5b3PK^LP5b3PK^LP5b3PK^LP5b3PK^LP5b3PK^LP5b3PK]LQ5c3oJ]LQ5c3oJ]LQ5c3oJ]LQ5c3oJ]LQ5c3oJ]LQ5c3oJ\\LR5d3nJ\\LR5d3nJ\\LS5c3mJ]LS5c3mJ]LS5c3mJ]LS5c3mJ]LS5c3mJ\\LT5d3lJ\\LT5d3lJ\\LT5d3lJ\\LT5d3lJ\\LT5d3lJ\\LT5d3lJ\\LT5d3lJ[LU5e3kJ[LU5e3kJ[LU5e3kJ[LV5d3jJ\\LV5d3jJ\\LV5e3iJ[LW5e3iJ[LW5e3iJ[LW5e3iJZLX5f3hJZLX5f3hJZLX5f3R10000000VJiKi4W4nJRLS5m3iJWLW5i3iJWLW5j3hJVLX5j3hJVLX5j3hJVLX5j3hJVLX5j3hJVLX5j3gJWLY5i3gJWLY5i3gJWLY5i3gJWLY5i3gJWLY5i3gJWLY5i3gJWLZ5h3fJXLZ5h3fJXLZ5h3fJTL^5m3aJRL`5n3`JRL`5n3`JRL`5n3]JULc5e400000000001O000000000000000000001O000000000000000000001O0000000000000000O1O1O1O100O1O10000O100001O0000000000000000000000000000000O101O001O1O001O1O001O1O001O0O1000001O0000000000000O2O00000000000O10001O00000O10000000001O0O1000000000001O0000000O100000001O000000000000000O2O00000000000000001O00000O100000001O0000000000000O101O0000000000001O1O001O1N101O001O in this image.", "objects": [{"patches": [117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288], "bbox": [0.09990625, 0.3946045197740113, 0.9212031249999999, 0.9997457627118643], "iscrowd": 0, "area": 84671.66060000005, "rle": {"size": [354, 640], "counts": "a]f0:[:b0^Oc0]Oc0]Oa0^O?B>B>B>C=J5L5J6K5K4K5L4K6M2O1O1N3N1000001O00000O2O00000000000O1000000000000O1000000000000O1000000000000O10000000000000O1000000000000000000000000000000000000000000001O00000000001O001O001O001O1O001P6mIYg68bXIV1iNX1hN]2cM>B>C0O0000O2O00000O101O0O10001N1O1N200O100000000O2O00000O100000000O100000000O101O00000000000000000000000000000000000000000bNYKbLf4^3[KaLe4^3^K`Lb4`3_K_La4a3aK]L_4c3cK[L]4e3dKZL\\4e3eK[L[4e3eK[L[4e3eK[L[4e3eK[L[4e3eKZL]4d3dK\\L\\4d3dK\\L\\4d3cK]L]4c3cK]L]4c3cK]L]4c3cK]L]4d3bK[L_4e3aK[L_4e3aK[L_4e3aKULe4k3[KoKk4Q4S101O00000000`JmKQ4S4aK[L_4e3aK[L_4e3`K[La4e3_K[La4e3_K[La4e3_K[La4f3^KZLb4f3^KYLc4g3]KYLd4f3\\KZLd4f3\\KZLd4f3\\KZLd4f3\\KYLe4g3[KYLe4g3[KYLe4g3[KYLe4g3[KXLf4h3YKYLg4g3YKYLg4g3YKYLg4g3YKYLg4g3YKXLh4h3XKXLi4g3WKYLi4g3WKYLi4g3WKYLi4g3WKXLj4h3VKXLj4h3VKXLj4h3VKXLj4h3VKXLj4h3VKXLj4h3UKYLk4g3UKYLk4g3UKYLk4g3UKYLk4g3UKYLl4f3TKYLm4g3SKYLm4g3SKYLm4f3TKZLl4f3TKZLl4f3TKZLl4f3TKZLl4f3TKZLl4f3TKZLl4f3TKZLl4f3SK[Lm4e3SK[Lm4e3SK[Ln4d3RK\\Ln4d3RK\\Ln4d3RK\\Ln4d3RK\\Ln4d3RK\\Ln4d3RK[Lo4e3QK[Lo4e3QK[Lo4e3QK[Lo4e3QK[Lo4e3QK[Lo4e3QK[Lo4d3QK]Lo4c3QK]LP5b3PK^LP5b3PK^LP5b3PK^LP5b3PK^LP5b3PK^LP5b3PK]LQ5c3oJ]LQ5c3oJ]LQ5c3oJ]LQ5c3oJ]LQ5c3oJ]LQ5c3oJ\\LR5d3nJ\\LR5d3nJ\\LS5c3mJ]LS5c3mJ]LS5c3mJ]LS5c3mJ]LS5c3mJ\\LT5d3lJ\\LT5d3lJ\\LT5d3lJ\\LT5d3lJ\\LT5d3lJ\\LT5d3lJ\\LT5d3lJ[LU5e3kJ[LU5e3kJ[LU5e3kJ[LV5d3jJ\\LV5d3jJ\\LV5e3iJ[LW5e3iJ[LW5e3iJ[LW5e3iJZLX5f3hJZLX5f3hJZLX5f3R10000000VJiKi4W4nJRLS5m3iJWLW5i3iJWLW5j3hJVLX5j3hJVLX5j3hJVLX5j3hJVLX5j3hJVLX5j3gJWLY5i3gJWLY5i3gJWLY5i3gJWLY5i3gJWLY5i3gJWLY5i3gJWLZ5h3fJXLZ5h3fJXLZ5h3fJTL^5m3aJRL`5n3`JRL`5n3`JRL`5n3]JULc5e400000000001O000000000000000000001O000000000000000000001O0000000000000000O1O1O1O100O1O10000O100001O0000000000000000000000000000000O101O001O1O001O1O001O1O001O0O1000001O0000000000000O2O00000000000O10001O00000O10000000001O0O1000000000001O0000000O100000001O000000000000000O2O00000000000000001O00000O100000001O0000000000000O101O0000000000001O1O001O1N101O001O in this image.", "objects": [{"patches": [130, 131, 152, 153, 174, 175, 196, 197, 217, 218, 219, 239, 240, 241, 261, 262, 263, 283, 284, 285, 305, 306, 307, 327, 328, 329, 350, 351, 372, 373, 395], "bbox": [0.8804901960784314, 0.23423202614379085, 0.9978267973856211, 0.8102287581699347], "iscrowd": 0, "area": 16975.491199999993, "rle": {"size": [612, 612], "counts": "a]R:8db0"}, "label": "diner by himself on the right side of the table"}]} {"id": 578369, "image": "COCO_train2014_000000578369.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person partially visible on the right side\"."}], "task": "refering", "answer_template": "The \"the person partially visible on the right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 152, 153, 174, 175, 196, 197, 217, 218, 219, 239, 240, 241, 261, 262, 263, 283, 284, 285, 305, 306, 307, 327, 328, 329, 350, 351, 372, 373, 395], "bbox": [0.8804901960784314, 0.23423202614379085, 0.9978267973856211, 0.8102287581699347], "iscrowd": 0, "area": 16975.491199999993, "rle": {"size": [612, 612], "counts": "a]R:8db0"}, "label": "the person partially visible on the right side"}]} {"id": 578369, "image": "COCO_train2014_000000578369.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a hat\"."}], "task": "refering", "answer_template": "The \"a man wearing a hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 88, 110, 111, 112, 132, 133, 154, 155, 176, 177, 179, 180, 198, 199, 201, 202, 203, 205, 206, 207, 208, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 293, 308, 309, 310, 311, 312, 313, 314, 330, 331, 332, 333, 334, 335, 336, 352, 353, 354, 355, 356, 357, 374, 375, 376, 377, 396, 397, 398, 399, 418, 419, 420, 421, 440, 441, 442, 462, 463, 464], "bbox": [0.0, 0.14527777777777778, 0.477483660130719, 0.9876143790849673], "iscrowd": 0, "area": 59027.437450000005, "rle": {"size": [612, 612], "counts": "k2l2X`0g5XJh5XJh1YN02N2N2N2N2N2N2N2M3N2N2N1O2XFVLR4l3kKVLT4l3jKULU4m3iKTLV4n3gKTLX4n3fKSLY4o3eKRLY4Q4dKQL[4Q4cKPL\\4R4bKoK]4R4aKPL^4R4`KoK_4S4_KnK`4T4]KnKb4T4\\KmKc4U4[KlKd4V4XKmKg4U4VKmKh4T4VKoKi4P4VKSLi4m3VKULi4k3UKXLj4h3RK]Lm4c3oJbLP5^3lJgLS5Y3hJmLW5R3fJSMY5m2`J[M_5e2TJiMj5X2nIQNQ6o1hIYNW6g1fG^KV1T3S7^1gGbLKX2\\8V1jG?T8BlG?R8BnG?P8BPH?n7BRH?l7BTH?j7BVH?h7AYH`0e7A[H`0c7A]H`0a7A_H`0_7AaH`0]7AcH`0[7AeH`0Y7@hHa0V7@jHa0S7AmH`0P7BPI?n6BRI:P7HPI4R7NnHNT73mHIU79kHCW7?iH]OY7e0gHVO\\7l0dHPO^7R1bHjN`7X1`HdNc7\\1^H`Nd7b1\\HZNf7h1ZHTNh7o1WHmMk7`2jG[MY8e62O100O1O101N100O100O11G9D=B=D=L31O010O1O100O1O10O01O100O1O12N4K5L4L3M3M3M2NO01000O10000O10O01000nL_CA`<0100O010O01O010O001O010O0010O01O1O010O000001O01O0001O00010O001O1O2O0O2N1O2N101N1O2N1O2O0O2N1O2N101N1O2O0010O100O010O10O010O2O0O101O0O101N100O2O001N100O2O0O101O0O101N100O10000O100O1000O01O1N2N2N2N2O1N2N1O2N2O1N2N2N101O10O01O1O001O10OO2O1O0O2O1O0O5L8G9H7IUon5"}, "label": "a man wearing a hat"}]} {"id": 578369, "image": "COCO_train2014_000000578369.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man wearing a purple sweatshirt next to a woman\"."}], "task": "refering", "answer_template": "The \"the man wearing a purple sweatshirt next to a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 88, 110, 111, 112, 132, 133, 154, 155, 176, 177, 179, 180, 198, 199, 201, 202, 203, 205, 206, 207, 208, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 286, 287, 288, 289, 290, 291, 292, 293, 308, 309, 310, 311, 312, 313, 314, 330, 331, 332, 333, 334, 335, 336, 352, 353, 354, 355, 356, 357, 374, 375, 376, 377, 396, 397, 398, 399, 418, 419, 420, 421, 440, 441, 442, 462, 463, 464], "bbox": [0.0, 0.14527777777777778, 0.477483660130719, 0.9876143790849673], "iscrowd": 0, "area": 59027.437450000005, "rle": {"size": [612, 612], "counts": "k2l2X`0g5XJh5XJh1YN02N2N2N2N2N2N2N2M3N2N2N1O2XFVLR4l3kKVLT4l3jKULU4m3iKTLV4n3gKTLX4n3fKSLY4o3eKRLY4Q4dKQL[4Q4cKPL\\4R4bKoK]4R4aKPL^4R4`KoK_4S4_KnK`4T4]KnKb4T4\\KmKc4U4[KlKd4V4XKmKg4U4VKmKh4T4VKoKi4P4VKSLi4m3VKULi4k3UKXLj4h3RK]Lm4c3oJbLP5^3lJgLS5Y3hJmLW5R3fJSMY5m2`J[M_5e2TJiMj5X2nIQNQ6o1hIYNW6g1fG^KV1T3S7^1gGbLKX2\\8V1jG?T8BlG?R8BnG?P8BPH?n7BRH?l7BTH?j7BVH?h7AYH`0e7A[H`0c7A]H`0a7A_H`0_7AaH`0]7AcH`0[7AeH`0Y7@hHa0V7@jHa0S7AmH`0P7BPI?n6BRI:P7HPI4R7NnHNT73mHIU79kHCW7?iH]OY7e0gHVO\\7l0dHPO^7R1bHjN`7X1`HdNc7\\1^H`Nd7b1\\HZNf7h1ZHTNh7o1WHmMk7`2jG[MY8e62O100O1O101N100O100O11G9D=B=D=L31O010O1O100O1O10O01O100O1O12N4K5L4L3M3M3M2NO01000O10000O10O01000nL_CA`<0100O010O01O010O001O010O0010O01O1O010O000001O01O0001O00010O001O1O2O0O2N1O2N101N1O2N1O2O0O2N1O2N101N1O2O0010O100O010O10O010O2O0O101O0O101N100O2O001N100O2O0O101O0O101N100O10000O100O1000O01O1N2N2N2N2O1N2N1O2N2O1N2N2N101O10O01O1O001O10OO2O1O0O2O1O0O5L8G9H7IUon5"}, "label": "the man wearing a purple sweatshirt next to a woman"}]} {"id": 578369, "image": "COCO_train2014_000000578369.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a white shirt\"."}], "task": "refering", "answer_template": "The \"a woman in a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 90, 91, 111, 112, 113, 114, 133, 134, 135, 136, 155, 156, 157, 158, 177, 178, 179, 180, 199, 200, 201, 220, 221, 222, 225, 226, 230, 231, 243, 244, 247, 248, 251, 252, 253, 266, 269, 272, 273, 274], "bbox": [0.027401960784313725, 0.20545751633986928, 0.540359477124183, 0.5784150326797386], "iscrowd": 0, "area": 16823.392799999998, "rle": {"size": [612, 612], "counts": "Y^:1Rc03M2N3M2N3M2O2M2O1N2O1N2O2M2O1N2O1N2dKmNeFT1Y9mNdFW1Z9kNbFY1\\9hN`F^1]9cN`Fa1^9`NWDMb0h1T;^NWDNM[2j;hMWD0I^2n;cMWD3G]2PdLkA]3W>cLfA_3[>d02N1N3N2N1O2M2M4M2N3M2N3A>01O1N101N101O0O2O0010O100O100O2O0O2O00XOh@eMU?Z2RAbMm>\\2n0H:L4N2hNZ_ONh`0M__OOc`0Kd_O1ka0M2M3N3M2M3NPZ4OQfK0O1O1O1O001O1N2O1N2O1N2O0O010O010O0010O010OVOj02O1N2N2O1N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N3M2NSUi01njVO4K10000O101O0O1000000O2O000O1000000O10000O10000O1000000O10000O10000M3J6K5N2M3N2N2N1O200O1O1O1O10O010O010O0100OZO_^OJba05_^OJaa07`^OH`a07b^OG^a0:b^OE^a0;d^OD\\a0 in this image.", "objects": [{"patches": [89, 90, 91, 111, 112, 113, 114, 133, 134, 135, 136, 155, 156, 157, 158, 177, 178, 179, 180, 199, 200, 201, 220, 221, 222, 225, 226, 230, 231, 243, 244, 247, 248, 251, 252, 253, 266, 269, 272, 273, 274], "bbox": [0.027401960784313725, 0.20545751633986928, 0.540359477124183, 0.5784150326797386], "iscrowd": 0, "area": 16823.392799999998, "rle": {"size": [612, 612], "counts": "Y^:1Rc03M2N3M2N3M2O2M2O1N2O1N2O2M2O1N2O1N2dKmNeFT1Y9mNdFW1Z9kNbFY1\\9hN`F^1]9cN`Fa1^9`NWDMb0h1T;^NWDNM[2j;hMWD0I^2n;cMWD3G]2PdLkA]3W>cLfA_3[>d02N1N3N2N1O2M2M4M2N3M2N3A>01O1N101N101O0O2O0010O100O100O2O0O2O00XOh@eMU?Z2RAbMm>\\2n0H:L4N2hNZ_ONh`0M__OOc`0Kd_O1ka0M2M3N3M2M3NPZ4OQfK0O1O1O1O001O1N2O1N2O1N2O0O010O010O0010O010OVOj02O1N2N2O1N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2N3M2NSUi01njVO4K10000O101O0O1000000O2O000O1000000O10000O10000O1000000O10000O10000M3J6K5N2M3N2N2N1O200O1O1O1O10O010O010O0100OZO_^OJba05_^OJaa07`^OH`a07b^OG^a0:b^OE^a0;d^OD\\a0 in this image.", "objects": [{"patches": [34, 50, 51, 52, 53, 66, 67, 68, 69, 85], "bbox": [0.7013400000000001, 0.11431999999999999, 0.96812, 0.3201333333333334], "iscrowd": 0, "area": 3839.9155499999997, "rle": {"size": [375, 500], "counts": "edP41e;2M2O1O2M2O1O1N3N1O1O2O0O101N100O101N10001N100O101N100O2O0O100O2O0O101N10O001O0O2O001O001O001O1O0010O010O10O0100O010O010O10O0100O010O010O10O01000O010O0100O010O010O10O0100O010O010O10O0100O010O01000O010O10O010O0100O010O10O010O0100O010O10O2O3L3N2N2M4M2N2M3N3L3N2N2M4MVi5"}, "label": "a knife server underneath the shrimp pizza"}]} {"id": 332613, "image": "COCO_train2014_000000332613.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza slicer in a pizza\"."}], "task": "refering", "answer_template": "The \"a pizza slicer in a pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 50, 51, 52, 53, 66, 67, 68, 69, 85], "bbox": [0.7013400000000001, 0.11431999999999999, 0.96812, 0.3201333333333334], "iscrowd": 0, "area": 3839.9155499999997, "rle": {"size": [375, 500], "counts": "edP41e;2M2O1O2M2O1O1N3N1O1O2O0O101N100O101N10001N100O101N100O2O0O100O2O0O101N10O001O0O2O001O001O001O1O0010O010O10O0100O010O010O10O0100O010O010O10O01000O010O0100O010O010O10O0100O010O010O10O0100O010O01000O010O10O010O0100O010O10O010O0100O010O10O2O3L3N2N2M4M2N2M3N3L3N2N2M4MVi5"}, "label": "a pizza slicer in a pizza"}]} {"id": 291658, "image": "COCO_train2014_000000291658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black vehicle parked on side of road\"."}], "task": "refering", "answer_template": "The \"black vehicle parked on side of road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 187, 204, 205, 206, 221, 222, 223, 224, 238, 239, 240, 241, 255, 256, 257, 258, 272, 273, 274, 275, 289, 290, 291, 292, 306, 307, 308, 309, 323, 324, 325, 340, 341, 342, 357, 358], "bbox": [0.0007916666666666666, 0.43370312499999997, 0.20729166666666668, 0.9546718750000001], "iscrowd": 0, "area": 22439.951300000008, "rle": {"size": [640, 480], "counts": "Rb0Q1c9[91O2N2N1O2N2N2O0O2N2N2N1iGlEa6V:]IkEb6W:[IkEd6V:[IkEd6W:YIlEe6V:YIkEg6V:WIkEh6V:VIlEi6`:kHbES7S;WHoDh7e;004M0O000001O000001O01O0000001O0000001O000000001O00000O2M2O1O1N3N1O1O1N2O2N1O1O1O2M2O1O1O2N1N2O1N3N1O1M3M4L4K6K5I6K6I7J5J7I7J5J7I7E:F;A>B?I7Kg0YO7H8H8H8H8^OXS]7"}, "label": "black vehicle parked on side of road"}]} {"id": 291658, "image": "COCO_train2014_000000291658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black car parked by curb\"."}], "task": "refering", "answer_template": "The \"black car parked by curb\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 187, 204, 205, 206, 221, 222, 223, 224, 238, 239, 240, 241, 255, 256, 257, 258, 272, 273, 274, 275, 289, 290, 291, 292, 306, 307, 308, 309, 323, 324, 325, 340, 341, 342, 357, 358], "bbox": [0.0007916666666666666, 0.43370312499999997, 0.20729166666666668, 0.9546718750000001], "iscrowd": 0, "area": 22439.951300000008, "rle": {"size": [640, 480], "counts": "Rb0Q1c9[91O2N2N1O2N2N2O0O2N2N2N1iGlEa6V:]IkEb6W:[IkEd6V:[IkEd6W:YIlEe6V:YIkEg6V:WIkEh6V:VIlEi6`:kHbES7S;WHoDh7e;004M0O000001O000001O01O0000001O0000001O000000001O00000O2M2O1O1N3N1O1O1N2O2N1O1O1O2M2O1O1O2N1N2O1N3N1O1M3M4L4K6K5I6K6I7J5J7I7J5J7I7E:F;A>B?I7Kg0YO7H8H8H8H8^OXS]7"}, "label": "black car parked by curb"}]} {"id": 291658, "image": "COCO_train2014_000000291658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a white van\"."}], "task": "refering", "answer_template": "The \"the back of a white van\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 157, 158, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 204, 205, 206, 207, 208, 209, 224, 225, 226], "bbox": [0.0, 0.3140625, 0.34924999999999995, 0.607296875], "iscrowd": 0, "area": 22503.253150000004, "rle": {"size": [640, 480], "counts": "f6R2aa0=0001O00001O00001W@_MQ>c2mA]MS>d2kA]MU>d2jA\\MV>f2hA[MX>e2fA\\MZ>e2dA\\M\\>f2aA[M_>f2_A\\M`>f2]A[Mc>f2YA]Mg>d2UA_Mk>c2PA`MP?a2g@hMX?`3000001O0000001O0001O01O0000001O0000001O0000001O000000010O0000001O0000001O0000001O0000001O0001O01O0000001O1O1O2N1O2N2N1O2N2N1O2N1O2O1N1O2N1O2N2NO100000000O101O1O1O1O0O2O1O1O1O1N2O1O1O001O1N2O1O1O1O1N2O001O1O1O1N2O1O1O1O0O2001O00001O00001N10O1N2N2O1N2QOo0M3N4L3L5L4K4M4L3L5L4K4M4L3L5L4K4M4]Ob0\\O\\gR6"}, "label": "the back of a white van"}]} {"id": 291658, "image": "COCO_train2014_000000291658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white van under streetlamps is parked next to a sidewalk\"."}], "task": "refering", "answer_template": "The \"a white van under streetlamps is parked next to a sidewalk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 136, 137, 138, 139, 140, 141, 153, 154, 155, 156, 157, 158, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 204, 205, 206, 207, 208, 209, 224, 225, 226], "bbox": [0.0, 0.3140625, 0.34924999999999995, 0.607296875], "iscrowd": 0, "area": 22503.253150000004, "rle": {"size": [640, 480], "counts": "f6R2aa0=0001O00001O00001W@_MQ>c2mA]MS>d2kA]MU>d2jA\\MV>f2hA[MX>e2fA\\MZ>e2dA\\M\\>f2aA[M_>f2_A\\M`>f2]A[Mc>f2YA]Mg>d2UA_Mk>c2PA`MP?a2g@hMX?`3000001O0000001O0001O01O0000001O0000001O0000001O000000010O0000001O0000001O0000001O0000001O0001O01O0000001O1O1O2N1O2N2N1O2N2N1O2N1O2O1N1O2N1O2N2NO100000000O101O1O1O1O0O2O1O1O1O1N2O1O1O001O1N2O1O1O1O1N2O001O1O1O1N2O1O1O1O0O2001O00001O00001N10O1N2N2O1N2QOo0M3N4L3L5L4K4M4L3L5L4K4M4L3L5L4K4M4]Ob0\\O\\gR6"}, "label": "a white van under streetlamps is parked next to a sidewalk"}]} {"id": 62295, "image": "COCO_train2014_000000062295.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in red pants doing a trick with a snowboard\"."}], "task": "refering", "answer_template": "The \"a man in red pants doing a trick with a snowboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 66, 67, 68, 81, 82, 83, 84, 85, 86, 95, 96, 97, 98, 99, 100, 101, 102, 103, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 122, 123, 124, 125, 126, 127, 128, 129, 130, 132, 138, 139, 141, 142, 143, 144, 145, 155, 156, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 174, 175, 183, 184, 185, 186, 187, 188, 189, 198, 199, 202, 203, 216, 217, 218, 230, 231, 232, 233, 245, 246, 247, 260, 261], "bbox": [0.18777517564402812, 0.161296875, 0.925456674473068, 0.7653125000000001], "iscrowd": 0, "area": 45327.85305, "rle": {"size": [640, 427], "counts": "\\Wb11mc05L4L3L5L4L4K5M2M4M3L4M2M4h_O_Nk=d1QB^Nm=b1SB`Nl=_1TBdNi=\\1WBfNh=Y1XBiNf=V1[BlNd=S1\\BoNb=Q1^BRO`=m0`BUO^=j0cBXO\\=g0dB[OZ=e0eB^OZ=a0fBBW=>iBDU==jBEU=:kBHS=9lBIS=7lBLQ=5nBMQ=3nBOP=1PC1ohN]AQ1c>oNeAh0\\>XOlA`0T>@TB7m=I[BOe=1cBF^=:jB^OV=b0RCUOo in this image.", "objects": [{"patches": [52, 66, 67, 68, 81, 82, 83, 84, 85, 86, 95, 96, 97, 98, 99, 100, 101, 102, 103, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 122, 123, 124, 125, 126, 127, 128, 129, 130, 132, 138, 139, 141, 142, 143, 144, 145, 155, 156, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 174, 175, 183, 184, 185, 186, 187, 188, 189, 198, 199, 202, 203, 216, 217, 218, 230, 231, 232, 233, 245, 246, 247, 260, 261], "bbox": [0.18777517564402812, 0.161296875, 0.925456674473068, 0.7653125000000001], "iscrowd": 0, "area": 45327.85305, "rle": {"size": [640, 427], "counts": "\\Wb11mc05L4L3L5L4L4K5M2M4M3L4M2M4h_O_Nk=d1QB^Nm=b1SB`Nl=_1TBdNi=\\1WBfNh=Y1XBiNf=V1[BlNd=S1\\BoNb=Q1^BRO`=m0`BUO^=j0cBXO\\=g0dB[OZ=e0eB^OZ=a0fBBW=>iBDU==jBEU=:kBHS=9lBIS=7lBLQ=5nBMQ=3nBOP=1PC1ohN]AQ1c>oNeAh0\\>XOlA`0T>@TB7m=I[BOe=1cBF^=:jB^OV=b0RCUOo in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 47, 48, 49, 50, 54, 55, 56, 57, 58, 59, 62, 63, 64, 70, 71, 72, 73, 74, 77, 78, 86, 87, 88, 89, 92, 93, 101, 102, 103, 104, 160, 161, 175, 176, 190, 191, 205, 206, 207, 219, 220, 221, 222, 234, 235, 236, 237, 238, 250, 251, 252, 253], "bbox": [0.1697423887587822, 0.003046875, 0.9932786885245901, 0.71475], "iscrowd": 0, "area": 51820.29075000001, "rle": {"size": [640, 427], "counts": "aU]11ec0;F:E;E;E;E;E;E;E;E;E;F:E;E;E:I8I7I7I7N2O1O1O1OL4L4M3L4L4L4J6I7H8I7I7I7L4O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O100O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O10000O10000O100001O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1dIPMfKQ3V4RMjKn2R4WMmKj2n3ZMRLg2i3]MWLd2e3`MZL`2b3dM^L]2]3gMcLZ2X3kMgLV2U3mMjLT2R3PNnLQ2m2TNRMm1i2WNWMj1d2ZN\\Mf1a2^N^Mc1]2aNcM`1X2dNhM]1S2gNmMY1P2kNoMV1l1nNTNS1g1QOYNP1b1TO^Nm0^1WOaNi0[1[OeNf0V1^OjNc0Q1AoN`0m0DRO^LBb3`0\\L@d3b0ZL^Of3d0XL\\Og3g0WLYOi3i0ULWOk3k0SLUOm3m0QLSOo3o0oKQOQ4Q1mKoNS4S1kKmNU4U1iKkNW4W1gKiNY4Y1eKgNZ4\\1dKdN\\4^1aKcN_4_1_KaNa4a1\\K`Nd4b1ZK^Nf4d1WK]Ni4f1SK[Nm4g1QKYNo4j1mJWNS5k1kJUNT5o1hJRNX5P2fJPNZ5S2bJnM^5T2`JlM`5W2\\JjMd5Y800000000O10000000000000000O1000000000000000000O1000000000000000000O100000a0_OR1nNR1nNR1nNS1mNTg1"}, "label": "back legs and tail of adult elephant behind the baby"}]} {"id": 398172, "image": "COCO_train2014_000000398172.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tallest elephant in the back side\"."}], "task": "refering", "answer_template": "The \"tallest elephant in the back side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 47, 48, 49, 50, 54, 55, 56, 57, 58, 59, 62, 63, 64, 70, 71, 72, 73, 74, 77, 78, 86, 87, 88, 89, 92, 93, 101, 102, 103, 104, 160, 161, 175, 176, 190, 191, 205, 206, 207, 219, 220, 221, 222, 234, 235, 236, 237, 238, 250, 251, 252, 253], "bbox": [0.1697423887587822, 0.003046875, 0.9932786885245901, 0.71475], "iscrowd": 0, "area": 51820.29075000001, "rle": {"size": [640, 427], "counts": "aU]11ec0;F:E;E;E;E;E;E;E;E;E;F:E;E;E:I8I7I7I7N2O1O1O1OL4L4M3L4L4L4J6I7H8I7I7I7L4O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O100O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O10000O10000O100001O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O1O001O1dIPMfKQ3V4RMjKn2R4WMmKj2n3ZMRLg2i3]MWLd2e3`MZL`2b3dM^L]2]3gMcLZ2X3kMgLV2U3mMjLT2R3PNnLQ2m2TNRMm1i2WNWMj1d2ZN\\Mf1a2^N^Mc1]2aNcM`1X2dNhM]1S2gNmMY1P2kNoMV1l1nNTNS1g1QOYNP1b1TO^Nm0^1WOaNi0[1[OeNf0V1^OjNc0Q1AoN`0m0DRO^LBb3`0\\L@d3b0ZL^Of3d0XL\\Og3g0WLYOi3i0ULWOk3k0SLUOm3m0QLSOo3o0oKQOQ4Q1mKoNS4S1kKmNU4U1iKkNW4W1gKiNY4Y1eKgNZ4\\1dKdN\\4^1aKcN_4_1_KaNa4a1\\K`Nd4b1ZK^Nf4d1WK]Ni4f1SK[Nm4g1QKYNo4j1mJWNS5k1kJUNT5o1hJRNX5P2fJPNZ5S2bJnM^5T2`JlM`5W2\\JjMd5Y800000000O10000000000000000O1000000000000000000O1000000000000000000O100000a0_OR1nNR1nNR1nNS1mNTg1"}, "label": "tallest elephant in the back side"}]} {"id": 398172, "image": "COCO_train2014_000000398172.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby elephant walking along with his mother\"."}], "task": "refering", "answer_template": "The \"a baby elephant walking along with his mother\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 49, 50, 51, 52, 53, 54, 64, 65, 66, 67, 68, 69, 70, 78, 79, 80, 81, 82, 83, 84, 85, 86, 93, 94, 95, 96, 97, 98, 99, 100, 101, 108, 109, 110, 111, 112, 113, 114, 115, 116, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 153, 154, 155, 156, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 174, 175, 183, 184, 185, 186, 187, 188, 189, 190, 198, 199, 200, 201, 202, 203, 204, 205, 213, 214, 215, 216, 217, 218, 219, 220, 227, 228, 229, 230, 231, 232, 233, 234, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 275, 276, 277, 278, 279, 289, 290, 291, 292, 293], "bbox": [0.188711943793911, 0.10093749999999999, 0.7955737704918032, 0.86996875], "iscrowd": 0, "area": 85773.24434999998, "rle": {"size": [640, 427], "counts": "YQc15ic06J3M2N3M2N2N3M2M3N2N3L3N2N2N2M3N2eJdNPH^1`7SO\\HP1[7ZOaHj0V7_ObG^OiLV1\\;HeG_1S8mNgGU1P8WOjGl0R8\\OhGf0U8I\\G:a81oF7m80iF7T9OaF9\\9NRF`0k9FeEf0X:c4N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O001O1O100O1O1dIQFf2P:XMRFg2n9XMTFh2l9VMVFi2k9VMVFi2l9TMVFk2n9PMSFP3R:jLPFV3T:dLnE[3W:[LnEe3V:RLPFm3U:hKRFW4S:_KRFb4R:TKTFk4Y:cJmE\\5[<0100O001O10O01O100O100O2N100O101N1O100O101N1O100O100O2N100O100O2O0O10000O101N100O10000O2O0O100O100O2O000O100O101O0000000001O000000001O0B\\K_Ae4a>_K[Aa4e>bKXA^4h>eKUA\\4k><00010O01O01O01O010O0010O01O010O0O2O1N1O2N2O1N1O2O1N1O2O1N1O2N2EXJ\\Bi5b=\\JYBe5f= in this image.", "objects": [{"patches": [36, 37, 38, 49, 50, 51, 52, 53, 54, 64, 65, 66, 67, 68, 69, 70, 78, 79, 80, 81, 82, 83, 84, 85, 86, 93, 94, 95, 96, 97, 98, 99, 100, 101, 108, 109, 110, 111, 112, 113, 114, 115, 116, 123, 124, 125, 126, 127, 128, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 153, 154, 155, 156, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 174, 175, 183, 184, 185, 186, 187, 188, 189, 190, 198, 199, 200, 201, 202, 203, 204, 205, 213, 214, 215, 216, 217, 218, 219, 220, 227, 228, 229, 230, 231, 232, 233, 234, 245, 246, 247, 248, 249, 260, 261, 262, 263, 264, 275, 276, 277, 278, 279, 289, 290, 291, 292, 293], "bbox": [0.188711943793911, 0.10093749999999999, 0.7955737704918032, 0.86996875], "iscrowd": 0, "area": 85773.24434999998, "rle": {"size": [640, 427], "counts": "YQc15ic06J3M2N3M2N2N3M2M3N2N3L3N2N2N2M3N2eJdNPH^1`7SO\\HP1[7ZOaHj0V7_ObG^OiLV1\\;HeG_1S8mNgGU1P8WOjGl0R8\\OhGf0U8I\\G:a81oF7m80iF7T9OaF9\\9NRF`0k9FeEf0X:c4N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O001O1O100O1O1dIQFf2P:XMRFg2n9XMTFh2l9VMVFi2k9VMVFi2l9TMVFk2n9PMSFP3R:jLPFV3T:dLnE[3W:[LnEe3V:RLPFm3U:hKRFW4S:_KRFb4R:TKTFk4Y:cJmE\\5[<0100O001O10O01O100O100O2N100O101N1O100O101N1O100O100O2N100O100O2O0O10000O101N100O10000O2O0O100O100O2O000O100O101O0000000001O000000001O0B\\K_Ae4a>_K[Aa4e>bKXA^4h>eKUA\\4k><00010O01O01O01O010O0010O01O010O0O2O1N1O2N2O1N1O2O1N1O2O1N1O2N2EXJ\\Bi5b=\\JYBe5f= in this image.", "objects": [{"patches": [134, 135, 147, 148, 159, 160, 161, 172, 173, 174, 175, 186, 187, 198, 199, 200, 211, 212, 213], "bbox": [0.2862133333333333, 0.5931000000000001, 0.5114133333333334, 0.94266], "iscrowd": 0, "area": 6730.199100000002, "rle": {"size": [500, 375], "counts": "[fd17ZlF@V9?jF@X9?hF@Z9`0fF^O\\9b0cF]O_9b0k2000O0100O1N2N1N3M101N2N2O1NR_i2"}, "label": "a man with a yellow bandana , gray shirt , red shorts , and tennis shoes playing frisbee on the beach"}]} {"id": 275297, "image": "COCO_train2014_000000275297.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a green bandana\"."}], "task": "refering", "answer_template": "The \"a man with a green bandana\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 147, 148, 159, 160, 161, 172, 173, 174, 175, 186, 187, 198, 199, 200, 211, 212, 213], "bbox": [0.2862133333333333, 0.5931000000000001, 0.5114133333333334, 0.94266], "iscrowd": 0, "area": 6730.199100000002, "rle": {"size": [500, 375], "counts": "[fd17ZlF@V9?jF@X9?hF@Z9`0fF^O\\9b0cF]O_9b0k2000O0100O1N2N1N3M101N2N2O1NR_i2"}, "label": "a man with a green bandana"}]} {"id": 447349, "image": "COCO_train2014_000000447349.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair\"."}], "task": "refering", "answer_template": "The \"chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 18, 19, 21, 22, 31, 34, 35, 47, 48, 61, 62, 75, 88], "bbox": [0.39383999999999997, 5.9999999999999995e-05, 0.82248, 0.34472], "iscrowd": 0, "area": 7833.05235, "rle": {"size": [500, 375], "counts": "aXX23a?4L3M=hAHZ in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 18, 19, 21, 22, 31, 34, 35, 47, 48, 61, 62, 75, 88], "bbox": [0.39383999999999997, 5.9999999999999995e-05, 0.82248, 0.34472], "iscrowd": 0, "area": 7833.05235, "rle": {"size": [500, 375], "counts": "aXX23a?4L3M=hAHZ in this image.", "objects": [{"patches": [0, 1, 2, 13, 14, 15, 26, 27, 28, 29, 40, 41, 42, 54], "bbox": [0.0, 0.0, 0.2676266666666667, 0.23], "iscrowd": 0, "area": 7247.5962, "rle": {"size": [500, 375], "counts": "02b?3lBLb:7UEOk:5kD1U;2cD2^;2XD4h;OoC7Qn;BRD>n;ATD>l;AWD=i;BYD=g;B\\D in this image.", "objects": [{"patches": [64, 65, 86, 87, 88, 89, 109, 110, 111, 112, 132, 133, 134, 135, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 223, 224, 225, 226, 246, 247, 248, 249], "bbox": [0.6785, 0.17297169811320756, 0.915203125, 0.7331367924528303], "iscrowd": 0, "area": 20342.54399999999, "rle": {"size": [424, 640], "counts": "Ync53T=101N1000001O2N3M3M3M4L3M3D[OoCg0j;`0G9K5L4ARN_EQ2]:VN\\En1a:?L4M3M3L4M3L5L6I2O1N3N1N3N1O1N2O1O1O1O1O1O1O1M3L4N2N2N2M3M3N1O2N2O1O1N2dNQK^JP5^5VK_Jk4`5XK]Ji4c5YKZJh4f5ZKWJg4i5]KQJe4o5_KkIc4U6aKeIa4]6aK\\Ib4f6o02N2N2N2N2N3M4LiNVJSKg5k4\\JTKd5j4_JUKa5j4aJUK_5j4dJTK\\5l4VK`Jl4`5oJdJR5\\5nJcJS5]5oJaJQ5_5RK_Jm4a5UK]Jk4c5]10000O11O000G:M2N3M2N8G6K3M2N3M2N2M4LK[JoH`5S783M3L3M4L4O1000000O1E\\HQKe7o4:00O10000O2O2N1N3J6K5K5mKTGd3P9XLTGd3Z9N2M3hL[Fg2j9oL^Fm2Q:L301O00001B>POP1^Oc0CaCN` in this image.", "objects": [{"patches": [64, 65, 86, 87, 88, 89, 109, 110, 111, 112, 132, 133, 134, 135, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 223, 224, 225, 226, 246, 247, 248, 249], "bbox": [0.6785, 0.17297169811320756, 0.915203125, 0.7331367924528303], "iscrowd": 0, "area": 20342.54399999999, "rle": {"size": [424, 640], "counts": "Ync53T=101N1000001O2N3M3M3M4L3M3D[OoCg0j;`0G9K5L4ARN_EQ2]:VN\\En1a:?L4M3M3L4M3L5L6I2O1N3N1N3N1O1N2O1O1O1O1O1O1O1M3L4N2N2N2M3M3N1O2N2O1O1N2dNQK^JP5^5VK_Jk4`5XK]Ji4c5YKZJh4f5ZKWJg4i5]KQJe4o5_KkIc4U6aKeIa4]6aK\\Ib4f6o02N2N2N2N2N3M4LiNVJSKg5k4\\JTKd5j4_JUKa5j4aJUK_5j4dJTK\\5l4VK`Jl4`5oJdJR5\\5nJcJS5]5oJaJQ5_5RK_Jm4a5UK]Jk4c5]10000O11O000G:M2N3M2N8G6K3M2N3M2N2M4LK[JoH`5S783M3L3M4L4O1000000O1E\\HQKe7o4:00O10000O2O2N1N3J6K5K5mKTGd3P9XLTGd3Z9N2M3hL[Fg2j9oL^Fm2Q:L301O00001B>POP1^Oc0CaCN` in this image.", "objects": [{"patches": [148, 149, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 239, 240, 241, 243, 244, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289], "bbox": [0.42331250000000004, 0.43377358490566037, 0.6384062500000001, 0.8519575471698113], "iscrowd": 0, "area": 12107.1625, "rle": {"size": [424, 640], "counts": "Sa`3c0P:e2F:O1N2O1N2O1N2M3L4L4ZOf0JUNUHeNg7^1WIeMi6Z2`I^M`6a2iIWMW6h2RJPMn5n2W2O1O1O1O1O1O1O1O1O1O100001O001O00001O00001O0010O00100O0103L4M2M4M2M4M3L3NS1lNnHWOS7h0XImNj6Q1aIdN`6[1jI[NW6d1h2N1O1O2N1O1O1O1O1O1O100001O001O1O001O1O2N6J5K5K5K5mF`MT7e2_HcMa7c2QHfMn7_2eGjMZ8[2YGmMg8[3O1O001O1OWMoG;Q8GQH6n7KVH0j71YHFl7 in this image.", "objects": [{"patches": [148, 149, 171, 172, 173, 174, 175, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 239, 240, 241, 243, 244, 262, 263, 264, 265, 266, 267, 285, 286, 287, 288, 289], "bbox": [0.42331250000000004, 0.43377358490566037, 0.6384062500000001, 0.8519575471698113], "iscrowd": 0, "area": 12107.1625, "rle": {"size": [424, 640], "counts": "Sa`3c0P:e2F:O1N2O1N2O1N2M3L4L4ZOf0JUNUHeNg7^1WIeMi6Z2`I^M`6a2iIWMW6h2RJPMn5n2W2O1O1O1O1O1O1O1O1O1O100001O001O00001O00001O0010O00100O0103L4M2M4M2M4M3L3NS1lNnHWOS7h0XImNj6Q1aIdN`6[1jI[NW6d1h2N1O1O2N1O1O1O1O1O1O100001O001O1O001O1O2N6J5K5K5K5mF`MT7e2_HcMa7c2QHfMn7_2eGjMZ8[2YGmMg8[3O1O001O1OWMoG;Q8GQH6n7KVH0j71YHFl7 in this image.", "objects": [{"patches": [89, 111, 112, 113, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 180, 181, 182, 183, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 272, 273, 274, 275], "bbox": [0.61159375, 0.32206686930091183, 1.0, 0.9577811550151976], "iscrowd": 0, "area": 24186.801799999994, "rle": {"size": [329, 640], "counts": "Smm31X:6I8I2M101N100O2O000O101N100O2O0O100O100O01000O100000O10O10000000O01000000000O0100000O0100000O01000O10O1000O10O1000O10O1000O10O1000000O1000000O1000000O100O10000O10000O01000O10O0100000O010000O01000O01000O10000O01000O1000O01000O10O0100O010O01O1O001O001N101O0WMmN]LT1_3YOWLh0e3AUL?h3ISL8i30RL1j37QLIl3>oKDP4>nKBR4`0lKAS4a0kK@T4b0jK_OU4d0iK\\OU4g0iKYOW4i0gKXOX4j0gKVOX4m0eKTOZ4n0eKROZ4P1dKQO[4Q1cKQOZ4R1eKPOW4T1fKoNV4T1iKoNT4U1gKoNV4T1fKQOU4U1fKlNY4d3N1N3N001N10001N101O0O2O000O2O001O00001O001O001O00001O001O000000000001O00001N10000000000000001O00000O1000001O1O1O1O1O1O1O1O1N2O2N3M2N3M2N2N2N2mKWKb2k4ZMXKe2j4WMYKh2i4TMZKj2i4QMZKo2h4mL[KR3g4jL\\KU3f4hL\\KW3g4dL[K\\3h4_L[K`3h5O1O001O1O1N101O1O1A>[Of0ZOf0ZOf0ZO[J"}, "label": "a man in a white shirt and a black helmet flashing a peace sign"}]} {"id": 177019, "image": "COCO_train2014_000000177019.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with the black helmet on the right\"."}], "task": "refering", "answer_template": "The \"the man with the black helmet on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 111, 112, 113, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 180, 181, 182, 183, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 272, 273, 274, 275], "bbox": [0.61159375, 0.32206686930091183, 1.0, 0.9577811550151976], "iscrowd": 0, "area": 24186.801799999994, "rle": {"size": [329, 640], "counts": "Smm31X:6I8I2M101N100O2O000O101N100O2O0O100O100O01000O100000O10O10000000O01000000000O0100000O0100000O01000O10O1000O10O1000O10O1000O10O1000000O1000000O1000000O100O10000O10000O01000O10O0100000O010000O01000O01000O10000O01000O1000O01000O10O0100O010O01O1O001O001N101O0WMmN]LT1_3YOWLh0e3AUL?h3ISL8i30RL1j37QLIl3>oKDP4>nKBR4`0lKAS4a0kK@T4b0jK_OU4d0iK\\OU4g0iKYOW4i0gKXOX4j0gKVOX4m0eKTOZ4n0eKROZ4P1dKQO[4Q1cKQOZ4R1eKPOW4T1fKoNV4T1iKoNT4U1gKoNV4T1fKQOU4U1fKlNY4d3N1N3N001N10001N101O0O2O000O2O001O00001O001O001O00001O001O000000000001O00001N10000000000000001O00000O1000001O1O1O1O1O1O1O1O1N2O2N3M2N3M2N2N2N2mKWKb2k4ZMXKe2j4WMYKh2i4TMZKj2i4QMZKo2h4mL[KR3g4jL\\KU3f4hL\\KW3g4dL[K\\3h4_L[K`3h5O1O001O1O1N101O1O1A>[Of0ZOf0ZOf0ZO[J"}, "label": "the man with the black helmet on the right"}]} {"id": 177019, "image": "COCO_train2014_000000177019.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in green on a motorcycle\"."}], "task": "refering", "answer_template": "The \"a man in green on a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 23, 24, 25, 46, 47, 48, 69, 70, 71, 92, 93, 94, 115, 116, 138, 139, 140, 161, 162, 163, 184, 185, 186, 187, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256], "bbox": [0.00115625, 0.07191489361702127, 0.287640625, 0.9887537993920973], "iscrowd": 0, "area": 23610.22655000001, "rle": {"size": [329, 640], "counts": "d;j8^11O1O1O1O001O100O1O1O1O1O1O1O1O1O10000000000O100000000O1000000000eJXOU1g0cNC\\1=]NKb15]NMb13]NOb11]N1b1O]N3b1M]N5b1K]N7b1I]N9b1H\\N:c1F\\Nc1B\\N`0c1A[Na0d1_O\\Nb0c1^O\\Nd0c1]O[Ne0H\\NSOo0U1f0b1[O]Ng0HYNROP1V1h0a1YO]Ni0b1WO]Nk0b1UO^Nl0a1TO^Nn0JQNQOR1S1o0b1QO^NQ1JmMROR1S1S1`1mN`NT1JjMROS1R1U1`1kN_NX1KeMSOS1R1Y1^1hN`N_1G]MXOU1o0g1BTM^OV1n0P2gMoLj2S1]O[2QO]L2Y1k0[2SO\\L2Z1i0\\2TO[L2Z1h0\\2VOZL2[1f0]2WOYL2[1f0\\2XOZL1\\1d0\\2[OWL1^1a0]2^OVL0^1=a2BQL1_17e2HmK0_13i2LhK1d1Jh25eK1g1@j2>_K2m1VOh2h0\\K1R2mNg2Q1XK1X7OhH1Y7NhH1X7OhH1Y7NhH1X7OhH1Y7NhH1X7OhH1Y7NhH2W7NjH1W7NiH2W7OiH0W70jHOW70jHOV71kHNV71kHNU72lHMU72lHMT73mHLT73mHKT75mHIU77kHGV79lHDV7;kHCV7=kHAW7>iHAX7?iH_OX7a0iH\\OZ7d0fHZO[7f0fHXO\\7g0eHWO\\7i0dHVO^7i0cHUO^7k0P1010O0001O00001O01O01O00001O00001O00001O00001O0000010O00001O0000000000000O100000000O100000000O100000000O10O100O10O0O2O1O0O2O1O1M3M3N2M[Xb4"}, "label": "a man in green on a motorcycle"}]} {"id": 177019, "image": "COCO_train2014_000000177019.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"area under motorcycle\"."}], "task": "refering", "answer_template": "The \"area under motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 170, 171, 193, 194, 195, 211, 212, 213, 214, 216, 217, 218, 219, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 261, 262], "bbox": [0.092203125, 0.5470820668693009, 0.5469062499999999, 0.9819756838905777], "iscrowd": 0, "area": 13636.807000000004, "rle": {"size": [329, 640], "counts": "eWc04U:6I7J1N1O2O0O2N100O2N101N101N101O001N101O00001O0O2O001O0000000O10000000001N1000000000000O100000000000000000000O100000000O10000O1000001N10000O1000000O10000O1000000O1000000O10000O1000000O10000O2O00000O1000000O10000O1000000O10000O10000O1O100O100O100O101O000000000002N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N1O010O00001O0000001O00001O00001O00001O00000O2O00001OO10000O10O010000O10000O100O01OVM]OPLb0n3CoK=n3JoK5o31mK0P46mKIQ4 in this image.", "objects": [{"patches": [4, 26, 27, 28, 49, 50, 51, 52, 72, 73, 74, 75, 95, 96, 118, 119, 141, 164, 165], "bbox": [0.128515625, 0.054042553191489366, 0.30796875, 0.6283890577507599], "iscrowd": 0, "area": 7883.004400000001, "rle": {"size": [329, 640], "counts": "g`j02T:6J6I7J6J6YNISI=a5U1RJROj5j2K2N1O2N2N2N1O200O1000O010000O010@\\K[Kd4d4_KYKk3GeLn4DYKc33cLc4MXK\\3>aLY48UKU3f0bLT4 in this image.", "objects": [{"patches": [4, 26, 27, 28, 49, 50, 51, 52, 72, 73, 74, 75, 95, 96, 118, 119, 141, 164, 165], "bbox": [0.128515625, 0.054042553191489366, 0.30796875, 0.6283890577507599], "iscrowd": 0, "area": 7883.004400000001, "rle": {"size": [329, 640], "counts": "g`j02T:6J6I7J6J6YNISI=a5U1RJROj5j2K2N1O2N2N2N1O200O1000O010000O010@\\K[Kd4d4_KYKk3GeLn4DYKc33cLc4MXK\\3>aLY48UKU3f0bLT4 in this image.", "objects": [{"patches": [217, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267], "bbox": [0.35529687499999996, 0.8044072948328267, 0.62940625, 1.0], "iscrowd": 0, "area": 5740.281650000002, "rle": {"size": [329, 640], "counts": "YbY21X:0O101N100O10000O2O0O100O101N100O100O10000O10000O100O10000O10000O10000O100O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1O1N2O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O100O1O1O1002N2N2N2N2N1O1O1O001O1O001O001O001O001O001O001O001O2N3M2N2N1O001O001O001O001O00000000000000O1O1O1O1O1O1O100O100O100O100O100000000O10000000000001O1O2N1O1O1O1O2N1OO1000000O1O1N20N2N2N2N2N4L4J5K6KZk[2"}, "label": "the control panel of a motorcycle"}]} {"id": 226176, "image": "COCO_train2014_000000226176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a helmet\"."}], "task": "refering", "answer_template": "The \"a man in a helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 75, 76, 77, 78, 79, 99, 100, 101, 102, 122, 123, 124, 125, 126, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 234, 235, 236, 238, 239, 240, 241, 242, 243, 257, 260, 261, 262, 263, 264, 265, 282, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309, 328, 329, 330, 351, 352, 353], "bbox": [0.141171875, 0.13004566210045662, 0.5846406249999999, 0.986552511415525], "iscrowd": 0, "area": 47123.4866, "rle": {"size": [438, 640], "counts": "TUW12d=1O2N2N2N2N101N2N2N2N2N1O2N1O01O001O00lCJi:5VELi:5UEMk:3TENl:2SENn:2QEOo:1oD1Q;NoD3P;NoD2R;NmD3S;MkD5U;KjD6V;JiD7W;HjD7V;JjD6V;IjD8V;HjD8V;GjD9W;GiD9W;FiD;V;EkD;U;EkD;U;DkDV;BjD>U;BkD>V;BjD>V;BjD>V;BiD>X;AiD?W;AiD>W;ChD>X;BhD>X;BhD=Y;BgD?Y;AgD>Z;BfD>Y;CfD=[;BfD>Z;BfD>Z;BeD>\\;BdD>[;CeD<\\;CeD=Z;DeD<\\;DdD<[;EeD:[;GdD9];FdD:[;GeD8\\;HcD9\\;HdD7];HdD7\\;0]D1b;Q110O01O01O010O01O010OTFE^6<`INW62iI8n5GRJb0e5_OZJk0]5TOcJo0W1WMP1k1gMP1W1WMQ1i1hMQ1U1XMS1f1hMS1S1YMT1e1hMR1T1[MS1b1iMT1S1[MS1b1jMS1S1[MS1b1iMT1S1\\MR1a1kMS1R1]MR1`1mMR1R1^MQ1`1lMS1R1_MP1_1nMR1R1_MP1^1nMT1Q1_MP1^1oMS1P1`MP1^1oMS1Q1`Mo0\\1QNS1P1bMn0\\1RNR1P1bMn0\\1QNS1P1cMm0Z1TNS1n0dMm0Z1TNS1o0cMm0Z1TNR1o0eMl0Y1UNS1o0dMiNYOe0P2CS1n0eMfNE?d1LS1n0eMdNN;Z13S1n0eM`N2>V15S1l0fM]N5a0S15R1n0fMYN8d0P15R1m0gMVN;g0m05R1m0gMTNk0l0DfNB=k0m0CfNBj0Q1^O`NI?LA_Ma1l2\\NLa0JP2:]MNb0HR29[M1b0GS28XM3d0FU2i3iMYLX2f3fM[L[2e3cM]L]2c3aM^L`2a3_MaLa2_3]MbLd2^3ZMdLf2\\3WMfLj2Z3nLnLS3P3gLVMZ3c60O100O100O101N100O2O2M2O2M2O1N3N1N2O2M2O1N3N1N2O2M2O1N3N1N2O2M2O2M2O1N3N1N2O2M2O3L4M3L4M3M3L4M3L2O2M2O1N3N1N3N1N3N1N2O2M2O2M2O2M2N2M4M2N3L3N2N2M2O1O2M2O1M4M2M3N3M3L5L3M4K4M4M2O2M2O2M2O2M2O2M2O2M3N1N3N1N3N1N3N1N3N1N3N1N3M2N3L5L8H7H9H7I8H8^OZaa3"}, "label": "a man in a helmet"}]} {"id": 226176, "image": "COCO_train2014_000000226176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man driving a motorcycle with a helmet on\"."}], "task": "refering", "answer_template": "The \"a man driving a motorcycle with a helmet on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 75, 76, 77, 78, 79, 99, 100, 101, 102, 122, 123, 124, 125, 126, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 234, 235, 236, 238, 239, 240, 241, 242, 243, 257, 260, 261, 262, 263, 264, 265, 282, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309, 328, 329, 330, 351, 352, 353], "bbox": [0.141171875, 0.13004566210045662, 0.5846406249999999, 0.986552511415525], "iscrowd": 0, "area": 47123.4866, "rle": {"size": [438, 640], "counts": "TUW12d=1O2N2N2N2N101N2N2N2N2N1O2N1O01O001O00lCJi:5VELi:5UEMk:3TENl:2SENn:2QEOo:1oD1Q;NoD3P;NoD2R;NmD3S;MkD5U;KjD6V;JiD7W;HjD7V;JjD6V;IjD8V;HjD8V;GjD9W;GiD9W;FiD;V;EkD;U;EkD;U;DkDV;BjD>U;BkD>V;BjD>V;BjD>V;BiD>X;AiD?W;AiD>W;ChD>X;BhD>X;BhD=Y;BgD?Y;AgD>Z;BfD>Y;CfD=[;BfD>Z;BfD>Z;BeD>\\;BdD>[;CeD<\\;CeD=Z;DeD<\\;DdD<[;EeD:[;GdD9];FdD:[;GeD8\\;HcD9\\;HdD7];HdD7\\;0]D1b;Q110O01O01O010O01O010OTFE^6<`INW62iI8n5GRJb0e5_OZJk0]5TOcJo0W1WMP1k1gMP1W1WMQ1i1hMQ1U1XMS1f1hMS1S1YMT1e1hMR1T1[MS1b1iMT1S1[MS1b1jMS1S1[MS1b1iMT1S1\\MR1a1kMS1R1]MR1`1mMR1R1^MQ1`1lMS1R1_MP1_1nMR1R1_MP1^1nMT1Q1_MP1^1oMS1P1`MP1^1oMS1Q1`Mo0\\1QNS1P1bMn0\\1RNR1P1bMn0\\1QNS1P1cMm0Z1TNS1n0dMm0Z1TNS1o0cMm0Z1TNR1o0eMl0Y1UNS1o0dMiNYOe0P2CS1n0eMfNE?d1LS1n0eMdNN;Z13S1n0eM`N2>V15S1l0fM]N5a0S15R1n0fMYN8d0P15R1m0gMVN;g0m05R1m0gMTNk0l0DfNB=k0m0CfNBj0Q1^O`NI?LA_Ma1l2\\NLa0JP2:]MNb0HR29[M1b0GS28XM3d0FU2i3iMYLX2f3fM[L[2e3cM]L]2c3aM^L`2a3_MaLa2_3]MbLd2^3ZMdLf2\\3WMfLj2Z3nLnLS3P3gLVMZ3c60O100O100O101N100O2O2M2O2M2O1N3N1N2O2M2O1N3N1N2O2M2O1N3N1N2O2M2O2M2O1N3N1N2O2M2O3L4M3L4M3M3L4M3L2O2M2O1N3N1N3N1N3N1N2O2M2O2M2O2M2N2M4M2N3L3N2N2M2O1O2M2O1M4M2M3N3M3L5L3M4K4M4M2O2M2O2M2O2M2O2M2O2M3N1N3N1N3N1N3N1N3N1N3N1N3M2N3L5L8H7H9H7I8H8^OZaa3"}, "label": "a man driving a motorcycle with a helmet on"}]} {"id": 226176, "image": "COCO_train2014_000000226176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing blue shirt sitting in the back in the two weeler\"."}], "task": "refering", "answer_template": "The \"a man wearing blue shirt sitting in the back in the two weeler\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 59, 60, 61, 62, 82, 83, 84, 85, 105, 106, 107, 108, 129, 130, 131, 132, 152, 153, 154, 155, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 311, 312, 313, 334, 335, 336, 358, 359, 360], "bbox": [0.5153125000000001, 0.08257990867579909, 0.7859375, 0.9981963470319636], "iscrowd": 0, "area": 34374.79424999999, "rle": {"size": [438, 640], "counts": "a_]44_=5J7I6J6L4K5K5L4M3M3M4L3M3M3M3M3M3M3M3M4L3N2M3M3M3N2M3M3N2M4L3M3N2M3L4M3L4L4L4L3I7cN^1M2cKkIW1V6eNoIX1S6dNQJ[1P6`NUJ^1m5\\NYJb1i5YN\\Jf1d5WN`Jg1a5VNcJi1]5SNVK]1k4^NZKb1f4YN`Ke1a4WNcKh1^4WNcKi1]4UNfKi1[4VNfKj1Z4UNhKi1Y4VNhKj1X4UNjKi1W4VNjKi1W4WNiKg1Y4XNiKe1m1UMiMV1:c1l1\\MhMQ1=`1k1dMeMl0`0^1j1lMcMf0c0\\1i1TNaM`0g0X1j1\\N\\M;k0W1k1aNWM8o0T1k1hNSM4S1Q1l1nNnL1V1YO]MP1a4HjLOY1VOeMj0Y45fLK\\1TOoM in this image.", "objects": [{"patches": [36, 37, 38, 59, 60, 61, 62, 82, 83, 84, 85, 105, 106, 107, 108, 129, 130, 131, 132, 152, 153, 154, 155, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 287, 288, 289, 290, 291, 311, 312, 313, 334, 335, 336, 358, 359, 360], "bbox": [0.5153125000000001, 0.08257990867579909, 0.7859375, 0.9981963470319636], "iscrowd": 0, "area": 34374.79424999999, "rle": {"size": [438, 640], "counts": "a_]44_=5J7I6J6L4K5K5L4M3M3M4L3M3M3M3M3M3M3M3M4L3N2M3M3M3N2M3M3N2M4L3M3N2M3L4M3L4L4L4L3I7cN^1M2cKkIW1V6eNoIX1S6dNQJ[1P6`NUJ^1m5\\NYJb1i5YN\\Jf1d5WN`Jg1a5VNcJi1]5SNVK]1k4^NZKb1f4YN`Ke1a4WNcKh1^4WNcKi1]4UNfKi1[4VNfKj1Z4UNhKi1Y4VNhKj1X4UNjKi1W4VNjKi1W4WNiKg1Y4XNiKe1m1UMiMV1:c1l1\\MhMQ1=`1k1dMeMl0`0^1j1lMcMf0c0\\1i1TNaM`0g0X1j1\\N\\M;k0W1k1aNWM8o0T1k1hNSM4S1Q1l1nNnL1V1YO]MP1a4HjLOY1VOeMj0Y45fLK\\1TOoM in this image.", "objects": [{"patches": [105, 106, 128, 129, 151, 152, 153, 174, 175, 176, 196, 197, 198, 199, 200, 219, 222, 223], "bbox": [0.5307031249999999, 0.29427230046948355, 0.71721875, 0.6515962441314553], "iscrowd": 0, "area": 6300.58405, "rle": {"size": [426, 640], "counts": "ff]48k<7H8I7O100OO1O2N1N3N1O1O2N1O2O000100O010O010O00100O0`MHPH8n78cGH[8b0]G_Oa8g0[GYOc8l0YGTOf8R1VGnNh8X1SGhNk8^1QGcNl8d1PG[Nn8k1nFVNn8R2nFnMP9W2lFiMU9[2gFeMY9_2bFaM`9T30O10000O100O100O10000O100O1O2J6K5K4K6K5K5J5M4O100O10O10O100O10O10O1lMREj01MX;GRE;m:ZO]Ej0`;1O1O1O010O1O001O1O`0@O101N3N1N3M3N2M3N2M3N2M2O2M1O2O1N2O1N101N2N2N1O1O2M2O1O1O2N1O]Q[2"}, "label": "a baseball player wearing red and white swinging a bat with a red helmet on his head"}]} {"id": 119693, "image": "COCO_train2014_000000119693.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player swinging the bat\"."}], "task": "refering", "answer_template": "The \"a baseball player swinging the bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 128, 129, 151, 152, 153, 174, 175, 176, 196, 197, 198, 199, 200, 219, 222, 223], "bbox": [0.5307031249999999, 0.29427230046948355, 0.71721875, 0.6515962441314553], "iscrowd": 0, "area": 6300.58405, "rle": {"size": [426, 640], "counts": "ff]48k<7H8I7O100OO1O2N1N3N1O1O2N1O2O000100O010O010O00100O0`MHPH8n78cGH[8b0]G_Oa8g0[GYOc8l0YGTOf8R1VGnNh8X1SGhNk8^1QGcNl8d1PG[Nn8k1nFVNn8R2nFnMP9W2lFiMU9[2gFeMY9_2bFaM`9T30O10000O100O100O10000O100O1O2J6K5K4K6K5K5J5M4O100O10O10O100O10O10O1lMREj01MX;GRE;m:ZO]Ej0`;1O1O1O010O1O001O1O`0@O101N3N1N3M3N2M3N2M3N2M2O2M1O2O1N2O1N101N2N2N1O1O2M2O1O1O2N1O]Q[2"}, "label": "a baseball player swinging the bat"}]} {"id": 119693, "image": "COCO_train2014_000000119693.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball umpire in position in baseball match\"."}], "task": "refering", "answer_template": "The \"a baseball umpire in position in baseball match\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 166, 167, 168, 189, 190, 191, 192, 212, 213, 235, 236, 258, 259], "bbox": [0.214765625, 0.4443427230046948, 0.37103125, 0.7905164319248826], "iscrowd": 0, "area": 6046.761799999999, "rle": {"size": [426, 640], "counts": "_Wi11W=7H6J6K4L4M2N2N;E1N2O1O100O1O1O1O1O0hDcNe:^1WEfNg:[1WEgNh:[1TEiNn0Ai8g1WFjNj0Fm8U2QGnMn8R2oFQNP9P2nFRNQ9n1mFVNQ9k1mFWNS9h1kF[NT9f1iF]NV9d1hF_NW9`1aFiN^9b201N1O100O2N100O2O0O1O101OO1N2CnFaLS9[3RGcLo8j1hF@>cNj8m1jF\\Oj9c0WFYOm9f0UFSOQ:l0QFjNT:Y1mE^NZ:c1hESN`:l1>1001O11N101O@iDnNV;h0UEWOj:h0XEXOh:?kDXO`07f:`0mDUO`0:c:?eEAZ:`0gE^OZ:c0eE]O[:f0cEYO^:j0^EUOc:m0[ESOf:m0YESOg:m0YEROi:f0^EZOb:d0_E]Ob:?aE@a:=aECa:0bDMQ13_:LjE3f;1000O0100O010O010O10O10O10O2O000O10]]W5"}, "label": "a baseball umpire in position in baseball match"}]} {"id": 119693, "image": "COCO_train2014_000000119693.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umpire crouching behind the catcher and watching the batter\"."}], "task": "refering", "answer_template": "The \"umpire crouching behind the catcher and watching the batter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 166, 167, 168, 189, 190, 191, 192, 212, 213, 235, 236, 258, 259], "bbox": [0.214765625, 0.4443427230046948, 0.37103125, 0.7905164319248826], "iscrowd": 0, "area": 6046.761799999999, "rle": {"size": [426, 640], "counts": "_Wi11W=7H6J6K4L4M2N2N;E1N2O1O100O1O1O1O1O0hDcNe:^1WEfNg:[1WEgNh:[1TEiNn0Ai8g1WFjNj0Fm8U2QGnMn8R2oFQNP9P2nFRNQ9n1mFVNQ9k1mFWNS9h1kF[NT9f1iF]NV9d1hF_NW9`1aFiN^9b201N1O100O2N100O2O0O1O101OO1N2CnFaLS9[3RGcLo8j1hF@>cNj8m1jF\\Oj9c0WFYOm9f0UFSOQ:l0QFjNT:Y1mE^NZ:c1hESN`:l1>1001O11N101O@iDnNV;h0UEWOj:h0XEXOh:?kDXO`07f:`0mDUO`0:c:?eEAZ:`0gE^OZ:c0eE]O[:f0cEYO^:j0^EUOc:m0[ESOf:m0YESOg:m0YEROi:f0^EZOb:d0_E]Ob:?aE@a:=aECa:0bDMQ13_:LjE3f;1000O0100O010O010O10O10O10O2O000O10]]W5"}, "label": "umpire crouching behind the catcher and watching the batter"}]} {"id": 447374, "image": "COCO_train2014_000000447374.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bottle of wine below the sushi tray\"."}], "task": "refering", "answer_template": "The \"bottle of wine below the sushi tray\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 91, 114, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 296, 297], "bbox": [0.8473124999999999, 0.13454166666666667, 0.9983281249999999, 0.7162083333333332], "iscrowd": 0, "area": 14733.101049999988, "rle": {"size": [480, 640], "counts": "a[n72l>5J5K6J6J5L5J6I5L4L3M4K5L4L3M4K5K5K4L5K5J6K4L5J6K4K6K5J6J5K6N2N1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O001N2O001N2O001N2M3N1O2M3N1O2N2M2O2N2N1NJ7H8N110O1O10O01000DkIoGT6R8QJhGP6W8;100O100O100O10Q="}, "label": "bottle of wine below the sushi tray"}]} {"id": 447374, "image": "COCO_train2014_000000447374.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bottle of liquid in front of sushi\"."}], "task": "refering", "answer_template": "The \"bottle of liquid in front of sushi\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 91, 114, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 226, 227, 228, 229, 249, 250, 251, 252, 272, 273, 274, 275, 296, 297], "bbox": [0.8473124999999999, 0.13454166666666667, 0.9983281249999999, 0.7162083333333332], "iscrowd": 0, "area": 14733.101049999988, "rle": {"size": [480, 640], "counts": "a[n72l>5J5K6J6J5L5J6I5L4L3M4K5L4L3M4K5K5K4L5K5J6K4L5J6K4K6K5J6J5K6N2N1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O001N2O001N2O001N2M3N1O2M3N1O2N2M2O2N2N1NJ7H8N110O1O10O01000DkIoGT6R8QJhGP6W8;100O100O100O10Q="}, "label": "bottle of liquid in front of sushi"}]} {"id": 209809, "image": "COCO_train2014_000000209809.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown straw basket\"."}], "task": "refering", "answer_template": "The \"a brown straw basket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 34, 35, 36, 51, 52, 53, 54, 55, 56, 57, 58, 59, 75, 76, 77, 78, 79, 99], "bbox": [0.240671875, 0.03934272300469484, 0.601890625, 0.2874882629107981], "iscrowd": 0, "area": 6894.0319500000005, "rle": {"size": [426, 640], "counts": "WTP29Q=1O001O1O1000O10O10000O01000O10000O10000O10O10O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O001O1O1O1O1N2O1O000000000000O1O1O1O001O1O1O1O10000000000O10O1000000000O1000000000O10O10000000000000000000O1000O1000000000000000O10000000000000O10000000O1000001O0O10000O100O10001N10000O100O10N2N1O2N1O2N2O001O1O001O001O1O001O001O1O001O001O1N101N2H7L5O0O21O1N3N2M3N1O2M10O001O1O001O001O1N101O0O2O001001N100O2O1O2M3N1N3mC]O[;e0\\DEa;R1O01O0O2O001N2O001N1O2N2M2O2N1O2N00O1O1O0O21O3M3M3M3M3MRRZ3"}, "label": "a brown straw basket"}]} {"id": 209809, "image": "COCO_train2014_000000209809.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown bread basket behind and to the left of the cake\"."}], "task": "refering", "answer_template": "The \"a brown bread basket behind and to the left of the cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 34, 35, 36, 51, 52, 53, 54, 55, 56, 57, 58, 59, 75, 76, 77, 78, 79, 99], "bbox": [0.240671875, 0.03934272300469484, 0.601890625, 0.2874882629107981], "iscrowd": 0, "area": 6894.0319500000005, "rle": {"size": [426, 640], "counts": "WTP29Q=1O001O1O1000O10O10000O01000O10000O10000O10O10O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O001O1O1O1O1N2O1O000000000000O1O1O1O001O1O1O1O10000000000O10O1000000000O1000000000O10O10000000000000000000O1000O1000000000000000O10000000000000O10000000O1000001O0O10000O100O10001N10000O100O10N2N1O2N1O2N2O001O1O001O001O1O001O001O1O001O001O1N101N2H7L5O0O21O1N3N2M3N1O2M10O001O1O001O001O1N101O0O2O001001N100O2O1O2M3N1N3mC]O[;e0\\DEa;R1O01O0O2O001N2O001N1O2N2M2O2N1O2N00O1O1O0O21O3M3M3M3M3MRRZ3"}, "label": "a brown bread basket behind and to the left of the cake"}]} {"id": 136092, "image": "COCO_train2014_000000136092.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"book on brown pant leg\"."}], "task": "refering", "answer_template": "The \"book on brown pant leg\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 96, 97, 98, 114, 115, 116, 131, 132, 133, 134], "bbox": [0.2959, 0.31208, 0.4986, 0.6244533333333334], "iscrowd": 0, "area": 5275.224199999999, "rle": {"size": [375, 500], "counts": "]]f11d;3N1N2N3M2O2M2N2N3M2O1N3M2N3N1N2N3M2N2O2M2N1O100O1O1O1O100O1O1O100O1O1O100O1O1kNUN^Hk1e7VNWHk1j7WNSHi1o7XNmGi1U8XNgGi1[8XNaGi1`8YN]Gg1e8YNXGh1j8YNRGh1P9YNlFh1U9<2N2N2O010O100O010O100O010O100O00100O10O0100O10O0100O10O0100O100O00100O1gN[F?d9@^F?c9^O`Fb0`9\\OaFd0`9ZObFf0]9XOfFg0[9WOgFi0Y9TOjF3_O;h9_OlF3_O>d9]OPG3_O>]:AeE1L2_:KgE2ON[:NhE31Lg:1a0Onfk2"}, "label": "book on brown pant leg"}]} {"id": 136092, "image": "COCO_train2014_000000136092.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"second persons pamphlet\"."}], "task": "refering", "answer_template": "The \"second persons pamphlet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 96, 97, 98, 114, 115, 116, 131, 132, 133, 134], "bbox": [0.2959, 0.31208, 0.4986, 0.6244533333333334], "iscrowd": 0, "area": 5275.224199999999, "rle": {"size": [375, 500], "counts": "]]f11d;3N1N2N3M2O2M2N2N3M2O1N3M2N3N1N2N3M2N2O2M2N1O100O1O1O1O100O1O1O100O1O1O100O1O1kNUN^Hk1e7VNWHk1j7WNSHi1o7XNmGi1U8XNgGi1[8XNaGi1`8YN]Gg1e8YNXGh1j8YNRGh1P9YNlFh1U9<2N2N2O010O100O010O100O010O100O00100O10O0100O10O0100O10O0100O100O00100O1gN[F?d9@^F?c9^O`Fb0`9\\OaFd0`9ZObFf0]9XOfFg0[9WOgFi0Y9TOjF3_O;h9_OlF3_O>d9]OPG3_O>]:AeE1L2_:KgE2ON[:NhE31Lg:1a0Onfk2"}, "label": "second persons pamphlet"}]} {"id": 136092, "image": "COCO_train2014_000000136092.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a book on a lady ' s lap\"."}], "task": "refering", "answer_template": "The \"a book on a lady ' s lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 134, 135, 149, 150, 151, 152, 153, 154, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 192, 208, 209, 210, 211, 228, 229, 230], "bbox": [0.2893, 0.5404800000000001, 0.83774, 0.9825066666666666], "iscrowd": 0, "area": 14618.074249999998, "rle": {"size": [375, 500], "counts": "][e12d;1O2N1O2N1O2N1N2O2N1O2N1O2N1O1N3N1O2N1O2M2O1O2N1O2M2O2N1O2N1N2O2N1O2N1N3N1O1O2N1O2N1O2N1O1O2O0O2N1O2N1O1O2N101O010O010O00010O010O010O00010O010O010O00010O010O010O010O00010O010O010O00010O010O010O01O01O010O010O01O01O010O010O010O00010O010O010O01O01O010O010O01O01O`MVGS2k8jMXGV2U9O100O1O01O001O001O00001O010O000010O00101N101N1O2O0O2N2O0O2O0O2N101N2N101N10O00001O01O01O001O00010O001O01O010O00010O01O01O010O00010O01O01O010O00010O011N2N2O1N2N3N1N2O1N2N3N1N2O1N2N3N1N1O1O100O001O100O1O1O10O01O1O100O1O00100O1O1O100O001O100O001O010O001O10O01O0010O01O1O010O001O010OVQn0"}, "label": "a book on a lady ' s lap"}]} {"id": 201634, "image": "COCO_train2014_000000201634.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person wearing black shirt and blue jeans\"."}], "task": "refering", "answer_template": "The \"person wearing black shirt and blue jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 142, 143, 164, 165, 166, 167, 168, 187, 188, 189, 190, 210, 211, 212, 232, 233, 234, 235, 255, 256, 257, 258, 280, 281, 282, 303, 304, 326, 327, 349, 350, 351], "bbox": [0.1224375, 0.32515625, 0.31446874999999996, 0.87640625], "iscrowd": 0, "area": 15567.380000000006, "rle": {"size": [512, 640], "counts": "nYW11i?7H8B>B>CVIh1BiMY7=ZIDUNY1]1eNT7>^I\\O[NW1X1nNP7>]J1nN@e6`0ZM@g2?ZM@f2a0YM_Oh2`0XM@h2a0XM^Oi2a0WM_Oi2a0WM_Oj2?XM@i2>XMAi2>XMBi2`610O010N1O2O1N102O0O2O1N100O2O2MVYk6"}, "label": "person wearing black shirt and blue jeans"}]} {"id": 201634, "image": "COCO_train2014_000000201634.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing blue jeans and a dark colored t shirt is standing with a good shoe\"."}], "task": "refering", "answer_template": "The \"a person wearing blue jeans and a dark colored t shirt is standing with a good shoe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 142, 143, 164, 165, 166, 167, 168, 187, 188, 189, 190, 210, 211, 212, 232, 233, 234, 235, 255, 256, 257, 258, 280, 281, 282, 303, 304, 326, 327, 349, 350, 351], "bbox": [0.1224375, 0.32515625, 0.31446874999999996, 0.87640625], "iscrowd": 0, "area": 15567.380000000006, "rle": {"size": [512, 640], "counts": "nYW11i?7H8B>B>CVIh1BiMY7=ZIDUNY1]1eNT7>^I\\O[NW1X1nNP7>]J1nN@e6`0ZM@g2?ZM@f2a0YM_Oh2`0XM@h2a0XM^Oi2a0WM_Oi2a0WM_Oj2?XM@i2>XMAi2>XMBi2`610O010N1O2O1N102O0O2O1N100O2O2MVYk6"}, "label": "a person wearing blue jeans and a dark colored t shirt is standing with a good shoe"}]} {"id": 201634, "image": "COCO_train2014_000000201634.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in a light colored tee shirt standing next to two other boys and looking off into the distance\"."}], "task": "refering", "answer_template": "The \"a boy in a light colored tee shirt standing next to two other boys and looking off into the distance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 169, 170, 171, 192, 193, 194, 215, 216, 217, 238, 239, 240, 261, 262, 263, 284, 285, 286, 307, 308, 309, 330, 331, 332, 353, 354, 355, 356], "bbox": [0.35134375, 0.3276953125, 0.502703125, 0.875], "iscrowd": 0, "area": 14121.306100000002, "rle": {"size": [512, 640], "counts": "Wm`35f?8I7H9H7H8fLDfFc0m8NgF8n88hFLn8d0gFAo8P1fFSOR9\\1cFiNZ9a1[FcNc9n3M1O0010O01O0011N2O1N2O1N2M3M3M3N2M3MTO\\JcG`5a8eJZGX5i8mJlFU5W9RK]FP5g9a03L42M9H8H9G8I7I7I2NO101O0OTHRIh6n6WISIi6l6XITIh6l6WIUIi6k6VIVIj6i6VIXIk6a0fHm4>bJm6`0gHl4 in this image.", "objects": [{"patches": [147, 148, 169, 170, 171, 192, 193, 194, 215, 216, 217, 238, 239, 240, 261, 262, 263, 284, 285, 286, 307, 308, 309, 330, 331, 332, 353, 354, 355, 356], "bbox": [0.35134375, 0.3276953125, 0.502703125, 0.875], "iscrowd": 0, "area": 14121.306100000002, "rle": {"size": [512, 640], "counts": "Wm`35f?8I7H9H7H8fLDfFc0m8NgF8n88hFLn8d0gFAo8P1fFSOR9\\1cFiNZ9a1[FcNc9n3M1O0010O01O0011N2O1N2O1N2M3M3M3N2M3MTO\\JcG`5a8eJZGX5i8mJlFU5W9RK]FP5g9a03L42M9H8H9G8I7I7I2NO101O0OTHRIh6n6WISIi6l6XITIh6l6WIUIi6k6VIVIj6i6VIXIk6a0fHm4>bJm6`0gHl4 in this image.", "objects": [{"patches": [56, 57, 58, 59, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 122, 123, 124, 125, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 252, 253, 254, 255, 256, 257, 258, 259, 260, 275, 276, 277, 278, 279, 280, 281, 298, 299, 300, 301, 302, 303, 304, 320, 321, 322, 323, 324, 325, 342, 343, 344, 345, 365, 366, 367, 368, 369, 387, 388, 389, 390, 391, 392, 408, 409, 410, 411, 412, 413, 414, 431, 432, 433, 434, 435, 436, 458], "bbox": [0.49437908496732025, 0.09214052287581699, 0.8449509803921569, 0.9393300653594772], "iscrowd": 0, "area": 64754.373000000014, "rle": {"size": [612, 612], "counts": "X[e5l1Ta0?A=D3L4M3L3M4M3L4M3L4L4M2M3N2N110000OQMoKoFP4P9ULlFk3S9YLjFf3V9_LfFa3Y9dLcF\\3]9hL_FW3a9mL\\FS3c9RMYFn2f9WMVFh2j9\\MSFd2l9aMPF_2P:eMlEZ2T:jMiEV2V:oMfEQ2Y:SNeEl1Z:WNdEj1Z:YNdEg1\\:\\NaEe1]:^NaEc1]:`NaE`1^:cN`E^1^:eN`E\\1^:hN_EY1`:iN^EW1a:lN]EW1n1RLe5i2[HX1i1VLj5c2[HZ1c1[Lo5]2\\H[1]1`LU6W2[H\\1X1eL[6P2\\H^1P1jLb6i1\\H`1j0oLh6b1\\Hc1e0QMl6^1]He1`0SMQ7Z1\\Hg1QJoLi0d2U5>TJjLh0i2S5?VJeLg0m2S5>PMCP3eN]HKa4a1S3\\NZI\\Oc3Y2R3TNcIC\\3W2R3oMjIHW3V2R3hMPJOR3U2Z3ZMmI9R3V2b5cMgJV2X;J7G9ZOf0ZOaYh1"}, "label": "boy in a white abercrombie shirt with other boys"}]} {"id": 119714, "image": "COCO_train2014_000000119714.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy with the abercrombie shirt\"."}], "task": "refering", "answer_template": "The \"the boy with the abercrombie shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 58, 59, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 122, 123, 124, 125, 144, 145, 146, 147, 165, 166, 167, 168, 169, 170, 171, 172, 187, 188, 189, 190, 191, 192, 193, 194, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 252, 253, 254, 255, 256, 257, 258, 259, 260, 275, 276, 277, 278, 279, 280, 281, 298, 299, 300, 301, 302, 303, 304, 320, 321, 322, 323, 324, 325, 342, 343, 344, 345, 365, 366, 367, 368, 369, 387, 388, 389, 390, 391, 392, 408, 409, 410, 411, 412, 413, 414, 431, 432, 433, 434, 435, 436, 458], "bbox": [0.49437908496732025, 0.09214052287581699, 0.8449509803921569, 0.9393300653594772], "iscrowd": 0, "area": 64754.373000000014, "rle": {"size": [612, 612], "counts": "X[e5l1Ta0?A=D3L4M3L3M4M3L4M3L4L4M2M3N2N110000OQMoKoFP4P9ULlFk3S9YLjFf3V9_LfFa3Y9dLcF\\3]9hL_FW3a9mL\\FS3c9RMYFn2f9WMVFh2j9\\MSFd2l9aMPF_2P:eMlEZ2T:jMiEV2V:oMfEQ2Y:SNeEl1Z:WNdEj1Z:YNdEg1\\:\\NaEe1]:^NaEc1]:`NaE`1^:cN`E^1^:eN`E\\1^:hN_EY1`:iN^EW1a:lN]EW1n1RLe5i2[HX1i1VLj5c2[HZ1c1[Lo5]2\\H[1]1`LU6W2[H\\1X1eL[6P2\\H^1P1jLb6i1\\H`1j0oLh6b1\\Hc1e0QMl6^1]He1`0SMQ7Z1\\Hg1QJoLi0d2U5>TJjLh0i2S5?VJeLg0m2S5>PMCP3eN]HKa4a1S3\\NZI\\Oc3Y2R3TNcIC\\3W2R3oMjIHW3V2R3hMPJOR3U2Z3ZMmI9R3V2b5cMgJV2X;J7G9ZOf0ZOaYh1"}, "label": "the boy with the abercrombie shirt"}]} {"id": 119714, "image": "COCO_train2014_000000119714.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man with a grey t - shirt on\"."}], "task": "refering", "answer_template": "The \"a young man with a grey t - shirt on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 115, 134, 135, 136, 137, 138, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 418, 419, 420, 421, 422, 440, 441, 442, 443, 444, 462, 463, 464, 465, 466], "bbox": [0.01573529411764706, 0.24718954248366012, 0.4808986928104575, 0.9820261437908496], "iscrowd": 0, "area": 63721.641500000005, "rle": {"size": [612, 612], "counts": "m`6m0Va0Q1oNQ1oNQ1oNQ1POQ1nNQ1oNQ1oNQ1YOg0N2O1O1O1N2O1ROPFgHQ:o6ZFoHg9P7[FmHg9S7YFlHh9U7XFjHh9V7YFiHg9X7YFhHg9W7YFiHg9X7YFhHf9X7[FgHe9Y7\\FgHc9Y7]FhHb9U7bFjH^9S7RGaHm8\\7_1M3M3M3M3M3M3M3M3N2N3M2N2N2O1N2N2N2N2N2N2N2N2N2N2N2O101N100O100O100O10000O100O100O100O10001N10000O10000O10000O1000000O10000O2O000O100001O00000000001eGWKY2i4bM]K]2c4^MdK`2\\4\\MiKc2W4YMoKe2Q4YMSLf2m3WMVLh2j3VMZLh2f3WM\\Lh2d3VM`Lh2`3VMdLh2]3UMfLj2Z3TMjLj2V3VMkLl1eNiI`4Z4nLa1Y4^NjKV1`4jNaKl0i4ROiLSOa3l0`LkNj3T1VLdNS4Z1nK]N[4b1gK[N\\4c1eK]N[4b1fK^N[4`1fK`NZ4_1hKaNX4]1iKcNX4[1iKeNW4Z1jKfNW4X1kKgNU4X1lKiNT4U1mKkNS4T1oKkNR4S1oKmNR4Q1oKoNQ4P1PLPOQ4n0QLROn3l0TLTOm3i0ULWOk3g0XLXOi3f0XLZOi3c0YL^Of3`0]L_Od3>^LBb3=_LCb3:`LFa37bLG_37cLI^34dLL]32eLM[31gLOZ3NhL2Y3KiL5W3IlL6U3GmL9T3DnLP3AQM?P3^ORMb0o2[OSMd0n2ZOUMe0l2XOVMh0k2UOWMk0i2SOZMl0g2ROZMn0g2oN[MQ1f2lN\\MT1d2jN_MU1b2gNaMY1a2bNbM^1`2]NdMb1_2XNdMg1_2TNdMl1_2nMdMR2g9000000000000000000O0100O100O1O100O100O100O010O1O100O100O100O00100O100O100O1O010O10000001O001O0O101O00001O001O000010O0010O010O012M3N3L3N3M1N10O0010O01O010O0001O001O00001O00001O0O2O1N2N3M2N2N2N2N2N3M2N2N2O1N2N3M2NRcm5"}, "label": "a young man with a grey t - shirt on"}]} {"id": 119714, "image": "COCO_train2014_000000119714.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"left most person\"."}], "task": "refering", "answer_template": "The \"left most person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 115, 134, 135, 136, 137, 138, 155, 156, 157, 158, 159, 160, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 264, 265, 266, 267, 268, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 418, 419, 420, 421, 422, 440, 441, 442, 443, 444, 462, 463, 464, 465, 466], "bbox": [0.01573529411764706, 0.24718954248366012, 0.4808986928104575, 0.9820261437908496], "iscrowd": 0, "area": 63721.641500000005, "rle": {"size": [612, 612], "counts": "m`6m0Va0Q1oNQ1oNQ1oNQ1POQ1nNQ1oNQ1oNQ1YOg0N2O1O1O1N2O1ROPFgHQ:o6ZFoHg9P7[FmHg9S7YFlHh9U7XFjHh9V7YFiHg9X7YFhHg9W7YFiHg9X7YFhHf9X7[FgHe9Y7\\FgHc9Y7]FhHb9U7bFjH^9S7RGaHm8\\7_1M3M3M3M3M3M3M3M3N2N3M2N2N2O1N2N2N2N2N2N2N2N2N2N2N2O101N100O100O100O10000O100O100O100O10001N10000O10000O10000O1000000O10000O2O000O100001O00000000001eGWKY2i4bM]K]2c4^MdK`2\\4\\MiKc2W4YMoKe2Q4YMSLf2m3WMVLh2j3VMZLh2f3WM\\Lh2d3VM`Lh2`3VMdLh2]3UMfLj2Z3TMjLj2V3VMkLl1eNiI`4Z4nLa1Y4^NjKV1`4jNaKl0i4ROiLSOa3l0`LkNj3T1VLdNS4Z1nK]N[4b1gK[N\\4c1eK]N[4b1fK^N[4`1fK`NZ4_1hKaNX4]1iKcNX4[1iKeNW4Z1jKfNW4X1kKgNU4X1lKiNT4U1mKkNS4T1oKkNR4S1oKmNR4Q1oKoNQ4P1PLPOQ4n0QLROn3l0TLTOm3i0ULWOk3g0XLXOi3f0XLZOi3c0YL^Of3`0]L_Od3>^LBb3=_LCb3:`LFa37bLG_37cLI^34dLL]32eLM[31gLOZ3NhL2Y3KiL5W3IlL6U3GmL9T3DnLP3AQM?P3^ORMb0o2[OSMd0n2ZOUMe0l2XOVMh0k2UOWMk0i2SOZMl0g2ROZMn0g2oN[MQ1f2lN\\MT1d2jN_MU1b2gNaMY1a2bNbM^1`2]NdMb1_2XNdMg1_2TNdMl1_2nMdMR2g9000000000000000000O0100O100O1O100O100O100O010O1O100O100O100O00100O100O100O1O010O10000001O001O0O101O00001O001O000010O0010O010O012M3N3L3N3M1N10O0010O01O010O0001O001O00001O00001O0O2O1N2N3M2N2N2N2N2N3M2N2N2O1N2N3M2NRcm5"}, "label": "left most person"}]} {"id": 119714, "image": "COCO_train2014_000000119714.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on far right of photo in original shirt\"."}], "task": "refering", "answer_template": "The \"man on far right of photo in original shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 62, 63, 64, 65, 84, 85, 86, 87, 106, 107, 108, 109, 128, 129, 130, 131, 150, 151, 152, 153, 172, 173, 174, 175, 194, 195, 196, 197, 216, 217, 218, 219, 238, 239, 240, 241, 260, 261, 262, 263, 282, 283, 284, 285, 304, 305, 306, 307, 328, 329, 372, 373, 392, 393, 394, 395, 414, 415, 416, 417, 436, 437, 438, 439, 458, 459, 460, 461, 481, 482, 483], "bbox": [0.8174836601307189, 0.0625326797385621, 0.9687254901960785, 0.9785130718954249], "iscrowd": 0, "area": 36788.0929, "rle": {"size": [612, 612], "counts": "lmZ92`b0d0XOh0L4M3L5L5lB\\N_8i1TGfNd8a1mFnNm8V1jFUOS3^NHa2nL[OT3nNSOk1cMh0l1VNGW1UNV2e0\\M]N]2R2VM_O?ZN`2V2QM@`0WNa2Y2PM^Oa0VNa2\\2oL\\Oc0UN_2^2oL\\Od0TN^2`2oLZOe0TN]2b2oLXOf0TN\\2d2nLXOg0RN\\2f2nLVOi0PN[2j2lLUOk0oMZ2l2lLSO]7m0dHRO\\7n0dHRO\\7n0eHQO[7n0gHQOY7o0gHQOZ7n0gHQOY7o0hHPOX7P1hHPOX7P1iHoNX7P1iHoNW7Q1iHoNW7Q1jHmNW7S1iHmNW7R1kHmNV7R1kHmNU7S1kHmNU7S1lHlNT7S1nHlNS7S1mHmNS7R1nHoNQ7P1oHQOQ7n0PIROP7n0PISOP7k0QIUOo6j0RIWOm6i0SIWOm6h0TIXOm6f0TI[Ok6d0VI\\Oj6d0UI]Ol6a0UI@j6?WIAj6=WIDi6;WIEi6:XIFi68XIIg67YIIh65YILg62YIOg60ZI1f6NZI3e6L\\I5d6I]I9a6F`I;`6D`I=`6AaI`0^6_OcIb0]6]ObIf0\\6YOeIh0[6VOfIk0Z6SOgIn0X6ROhIo0X6oNiIR1V6mNkIU1V6hNjIY1X6cNiI^1X6`NgIh4U3TKlL[;hLUEe2l>iNV1jNW1iNVi:"}, "label": "man on far right of photo in original shirt"}]} {"id": 119714, "image": "COCO_train2014_000000119714.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy in the black t - shirt\"."}], "task": "refering", "answer_template": "The \"the boy in the black t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 95, 96, 97, 98, 117, 118, 119, 120, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 182, 183, 184, 185, 186, 187, 203, 204, 205, 206, 207, 208, 209, 224, 225, 226, 227, 228, 229, 230, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 319, 335, 336, 337, 338, 339, 340, 341, 400, 401, 402, 403, 404, 405, 406, 407, 408, 422, 423, 424, 425, 426, 427, 428, 429, 430, 444, 445, 446, 447, 448, 449, 450, 451, 452, 466, 467, 468, 469, 470, 471, 472], "bbox": [0.18877450980392158, 0.16890522875816993, 0.5956372549019607, 0.9826470588235294], "iscrowd": 0, "area": 64037.680049999995, "rle": {"size": [612, 612], "counts": "PeU25kb06J6I7J6J5K6J6J6J6M2N3M3X@XNo=k1j@ZOU?\\2oDaLZ6`3\\InL`6S3TI\\Mg6e2oHhMm6Y2iHUNR7m1fH^NU7c1kHaNP7a1PIbNk6_1UIeNf6\\1[IfNa6\\1^IdNa6]1_IcNa6]1`IbN_6`1`I`N_6a1bI^N]6d1bI\\N]6e1cI[N\\6f1dIZN[6h1dIXN[6i1eIWNZ6k1dIVN[6k1eIUN[6k1eIUNZ6m1eIRNZ6P2fIPNX6R2hInMW6T2hIlMV6V2jIjMU6X2jIhMT6Z2kIgMS6[2mIeMR6]2mIcMQ6_2oIaMP6`2PJ`Mn5c2QJ]Mn5d2RJ\\Ml5g2SJYMk5i2UJWMj5j2VJUMi5n2UJSMj5n2VJRMh5Q3WJoLg5S3YJmLf5T3ZJlLd5W3[JiLd5X3\\JhLb5Z3^JfL`5]3_JcL`5^3_JcL`5_3_JaL`5`3`J`L_5a3aJ_L^5c3aJ]L^5d3bJ[L^5g3aJYL^5h3bJXL]5i3cJWL]5i3cJWL\\5j3cJWL\\5i3eJWLm3UM[Kd6i0WLh3ZM]K`6k0VLd3_M_K[6n0VL`3cM`KW6Q1VL[3hMlJ\\O8g6a1ULX3lMeJH:W6j1ULS3jNlJQ5R2ULo2k5QMTJl2o5TMRJh2Q6YMoId2T6\\MlI`2W6aMiI\\2Y6eMgIW2]6iMcIT2_6mMaIo1b6RN^Ij1f6VNZIh1g6XNZIh1e6YN[If1f6ZNZIe1f6\\NZId1e6]N[Ib1f6^NYIc1f6^NZIa1f6`NZI_1g6aNYI_1f6bNZI]1f6cN[I\\1f6dNZI\\1e6eN[IZ1e6gN[IX1f6hNZIX1g6gNYIX1h6hNXIW1i6iNWIW1j6hNVIW1k6hNVIX1j6hNUIY1l6fNTIZ1l6fNTIZ1m6eNSI[1m6eNSI[1m6eNSIZ1o6eNQI[1o6eNQI[1P7dNPI\\1P7cNQI]1o6cNQI]1P7bNPI^1P7bNPI^1P7bNPI]1R7bNnH^1R7cNlH^1U7bNjH^1V7cNiH]1W7cNiH]1X7cNgH]1Y7dNfH\\1[7dNdH\\1\\7dNdH[1]7fNbH[1_7dN`H\\1a7dN^H\\1c7cN]H]1e7bNZH_1f7aNYH_1i7`NVH`1k7_NTHc1k7^NTHb1l7_NSHa1m7`NRH`1m7aNSH`1l7`NTH`1l7`NTH`1l7`NTHa1k7^NVHb1i7_NVHb1j7^NVHb1j7]NWHd1h7\\NXHd1h7\\NWHf1g7ZNZHg1e7YN[Hh1d7XN[Hj1d7UN]Hl1b7TN^Hm1`7TN`Hm1_7RNaHo1_7QNaHP29fK[4Z2\\KQ26gK^4X2[KS23hKb4U2[KT2OjKf4Q2\\KV2JkKk4o1[KW2FmKo4l1ZKY2DmKR5j1ZKZ2AnKU5h1ZK[2^OnKY5g1YK^2YOmK^5d1YKd2SOjKd5b1YKg2oNhKh5a1YKj2jNfKn5`1WKn2fNdKT6]1VKS3aNaKY6\\1VKZ3XN\\Kb6Z1VKk3eMlJV7X1UKo4k4QKUKo4k4QKUKo4k4QKTKP5l4PKTKo4X1dIVO]1Bo4?cJI>Ho4:nJH2NQ56WKFH4P52cKD]O:P5MnKCRO?Q5IXLBgNe0Q5DcLA\\Nk0Q5@mL_OQNR1Q5[OYM]OfMW1R5WOcM\\O[M]1R5ROoMZOoLd1R5mNZNYOdLi1S5jNcNWOZLo1R5fNoNUOoKU2R5aN^1`1bN[Nc1d1^NWNg1i1YNTNj1l1UNTNl1k1UNTNl1l1TNSNl1n1TNQNm1o1SNPNn1o1RNQNo1o1QNPNP2P2PNoMQ2Q2oMnMR2Q2oMnMQ2S2nMmMS2S2mMmMS2R2nMmMS2S2mMlMT2T2lMkMU2U2jMkM[2P2fMoM`2l1`MSNf2h1`7J6K5K5K2M2O1O1N2O1O2N1N2O1O1N2O2N1O1N2O1O2M2O1O1O1N2O]ec4"}, "label": "the boy in the black t - shirt"}]} {"id": 119714, "image": "COCO_train2014_000000119714.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy with long hair wearing a black shirt\"."}], "task": "refering", "answer_template": "The \"a boy with long hair wearing a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 95, 96, 97, 98, 117, 118, 119, 120, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 182, 183, 184, 185, 186, 187, 203, 204, 205, 206, 207, 208, 209, 224, 225, 226, 227, 228, 229, 230, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 319, 335, 336, 337, 338, 339, 340, 341, 400, 401, 402, 403, 404, 405, 406, 407, 408, 422, 423, 424, 425, 426, 427, 428, 429, 430, 444, 445, 446, 447, 448, 449, 450, 451, 452, 466, 467, 468, 469, 470, 471, 472], "bbox": [0.18877450980392158, 0.16890522875816993, 0.5956372549019607, 0.9826470588235294], "iscrowd": 0, "area": 64037.680049999995, "rle": {"size": [612, 612], "counts": "PeU25kb06J6I7J6J5K6J6J6J6M2N3M3X@XNo=k1j@ZOU?\\2oDaLZ6`3\\InL`6S3TI\\Mg6e2oHhMm6Y2iHUNR7m1fH^NU7c1kHaNP7a1PIbNk6_1UIeNf6\\1[IfNa6\\1^IdNa6]1_IcNa6]1`IbN_6`1`I`N_6a1bI^N]6d1bI\\N]6e1cI[N\\6f1dIZN[6h1dIXN[6i1eIWNZ6k1dIVN[6k1eIUN[6k1eIUNZ6m1eIRNZ6P2fIPNX6R2hInMW6T2hIlMV6V2jIjMU6X2jIhMT6Z2kIgMS6[2mIeMR6]2mIcMQ6_2oIaMP6`2PJ`Mn5c2QJ]Mn5d2RJ\\Ml5g2SJYMk5i2UJWMj5j2VJUMi5n2UJSMj5n2VJRMh5Q3WJoLg5S3YJmLf5T3ZJlLd5W3[JiLd5X3\\JhLb5Z3^JfL`5]3_JcL`5^3_JcL`5_3_JaL`5`3`J`L_5a3aJ_L^5c3aJ]L^5d3bJ[L^5g3aJYL^5h3bJXL]5i3cJWL]5i3cJWL\\5j3cJWL\\5i3eJWLm3UM[Kd6i0WLh3ZM]K`6k0VLd3_M_K[6n0VL`3cM`KW6Q1VL[3hMlJ\\O8g6a1ULX3lMeJH:W6j1ULS3jNlJQ5R2ULo2k5QMTJl2o5TMRJh2Q6YMoId2T6\\MlI`2W6aMiI\\2Y6eMgIW2]6iMcIT2_6mMaIo1b6RN^Ij1f6VNZIh1g6XNZIh1e6YN[If1f6ZNZIe1f6\\NZId1e6]N[Ib1f6^NYIc1f6^NZIa1f6`NZI_1g6aNYI_1f6bNZI]1f6cN[I\\1f6dNZI\\1e6eN[IZ1e6gN[IX1f6hNZIX1g6gNYIX1h6hNXIW1i6iNWIW1j6hNVIW1k6hNVIX1j6hNUIY1l6fNTIZ1l6fNTIZ1m6eNSI[1m6eNSI[1m6eNSIZ1o6eNQI[1o6eNQI[1P7dNPI\\1P7cNQI]1o6cNQI]1P7bNPI^1P7bNPI^1P7bNPI]1R7bNnH^1R7cNlH^1U7bNjH^1V7cNiH]1W7cNiH]1X7cNgH]1Y7dNfH\\1[7dNdH\\1\\7dNdH[1]7fNbH[1_7dN`H\\1a7dN^H\\1c7cN]H]1e7bNZH_1f7aNYH_1i7`NVH`1k7_NTHc1k7^NTHb1l7_NSHa1m7`NRH`1m7aNSH`1l7`NTH`1l7`NTH`1l7`NTHa1k7^NVHb1i7_NVHb1j7^NVHb1j7]NWHd1h7\\NXHd1h7\\NWHf1g7ZNZHg1e7YN[Hh1d7XN[Hj1d7UN]Hl1b7TN^Hm1`7TN`Hm1_7RNaHo1_7QNaHP29fK[4Z2\\KQ26gK^4X2[KS23hKb4U2[KT2OjKf4Q2\\KV2JkKk4o1[KW2FmKo4l1ZKY2DmKR5j1ZKZ2AnKU5h1ZK[2^OnKY5g1YK^2YOmK^5d1YKd2SOjKd5b1YKg2oNhKh5a1YKj2jNfKn5`1WKn2fNdKT6]1VKS3aNaKY6\\1VKZ3XN\\Kb6Z1VKk3eMlJV7X1UKo4k4QKUKo4k4QKUKo4k4QKTKP5l4PKTKo4X1dIVO]1Bo4?cJI>Ho4:nJH2NQ56WKFH4P52cKD]O:P5MnKCRO?Q5IXLBgNe0Q5DcLA\\Nk0Q5@mL_OQNR1Q5[OYM]OfMW1R5WOcM\\O[M]1R5ROoMZOoLd1R5mNZNYOdLi1S5jNcNWOZLo1R5fNoNUOoKU2R5aN^1`1bN[Nc1d1^NWNg1i1YNTNj1l1UNTNl1k1UNTNl1l1TNSNl1n1TNQNm1o1SNPNn1o1RNQNo1o1QNPNP2P2PNoMQ2Q2oMnMR2Q2oMnMQ2S2nMmMS2S2mMmMS2R2nMmMS2S2mMlMT2T2lMkMU2U2jMkM[2P2fMoM`2l1`MSNf2h1`7J6K5K5K2M2O1O1N2O1O2N1N2O1O1N2O2N1O1N2O1O2M2O1O1O1N2O]ec4"}, "label": "a boy with long hair wearing a black shirt"}]} {"id": 504744, "image": "COCO_train2014_000000504744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a orange fruit under the banana in the bowl of fruits\"."}], "task": "refering", "answer_template": "The \"a orange fruit under the banana in the bowl of fruits\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 268, 285, 286, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 314], "bbox": [0.42229687499999996, 0.6705854800936768, 0.6875937499999999, 0.9317564402810303], "iscrowd": 0, "area": 13572.566449999998, "rle": {"size": [427, 640], "counts": "nm`35g in this image.", "objects": [{"patches": [241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 268, 285, 286, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 314], "bbox": [0.42229687499999996, 0.6705854800936768, 0.6875937499999999, 0.9317564402810303], "iscrowd": 0, "area": 13572.566449999998, "rle": {"size": [427, 640], "counts": "nm`35g in this image.", "objects": [{"patches": [201, 202, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 314, 315, 316], "bbox": [0.6374375, 0.5636065573770491, 0.873921875, 0.9016393442622951], "iscrowd": 0, "area": 14356.875399999997, "rle": {"size": [427, 640], "counts": "a]Z51X=4K4M3M3M4L3M3M4L3O1N3N1N2O2M2O1N2O2M2O1N3N1N3N3L3N3M3L4M2M4M3M3dEfM`9V3M1N101O0O100O2O0O10001N100O010O10O0100O010O010O10O0100O010O10O010O10O10O01000O010O10O10O01000O0100O0100O0100O010000O010O10O10O10000O2O1O1N101N2O001N2O001N2O1N101N2O0O2O1N2O0O2N2O0O2O1N101N2O1N101N2O0O2N2O1N101N4M2M4M2M4M2M4M2M4M2M5Ka0_OR_Q1"}, "label": "an apple beneath a banana"}]} {"id": 504744, "image": "COCO_train2014_000000504744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"apple that is underneath a banana and right of an orange in a wooden bowl\"."}], "task": "refering", "answer_template": "The \"apple that is underneath a banana and right of an orange in a wooden bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 202, 222, 223, 224, 225, 226, 244, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 295, 314, 315, 316], "bbox": [0.6374375, 0.5636065573770491, 0.873921875, 0.9016393442622951], "iscrowd": 0, "area": 14356.875399999997, "rle": {"size": [427, 640], "counts": "a]Z51X=4K4M3M3M4L3M3M4L3O1N3N1N2O2M2O1N2O2M2O1N3N1N3N3L3N3M3L4M2M4M3M3dEfM`9V3M1N101O0O100O2O0O10001N100O010O10O0100O010O010O10O0100O010O10O010O10O10O01000O010O10O10O01000O0100O0100O0100O010000O010O10O10O10000O2O1O1N101N2O001N2O001N2O1N101N2O0O2O1N2O0O2N2O0O2O1N101N2O1N101N2O0O2N2O1N101N4M2M4M2M4M2M4M2M4M2M5Ka0_OR_Q1"}, "label": "apple that is underneath a banana and right of an orange in a wooden bowl"}]} {"id": 504744, "image": "COCO_train2014_000000504744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an orange placed behind an apple on the left side of a fruit bowl\"."}], "task": "refering", "answer_template": "The \"an orange placed behind an apple on the left side of a fruit bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 144, 145, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 210, 211, 212, 214, 233, 234], "bbox": [0.128734375, 0.43529274004683843, 0.35321875, 0.7505854800936768], "iscrowd": 0, "area": 9923.0267, "rle": {"size": [427, 640], "counts": "U^R11X=9G8H9F:G8H4L3M3M3M3M3L4L5K4L4L4M3N2N2N3N1N2N2N2N2N3N1N2N2N2N2N2ON1N2N2N3M2N3M2N3N1O2M2O2N1O2M2O2N2N1N3O010O001O10O01O1O100O00100O1000O010000O01000O10000O101N10000O2O000O10001N100O101O0O100O2O000O10010O0001O01O01O00010O001O00010O000010O00100O2N2O1N1O2O1N2N101N2N3M3M2N3N1N3M3M2N3M3M6J9GX\\\\5"}, "label": "an orange placed behind an apple on the left side of a fruit bowl"}]} {"id": 504744, "image": "COCO_train2014_000000504744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a orange in a bowl of fruit\"."}], "task": "refering", "answer_template": "The \"a orange in a bowl of fruit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 144, 145, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 210, 211, 212, 214, 233, 234], "bbox": [0.128734375, 0.43529274004683843, 0.35321875, 0.7505854800936768], "iscrowd": 0, "area": 9923.0267, "rle": {"size": [427, 640], "counts": "U^R11X=9G8H9F:G8H4L3M3M3M3M3L4L5K4L4L4M3N2N2N3N1N2N2N2N2N3N1N2N2N2N2N2ON1N2N2N3M2N3M2N3N1O2M2O2N1O2M2O2N2N1N3O010O001O10O01O1O100O00100O1000O010000O01000O10000O101N10000O2O000O10001N100O101O0O100O2O000O10010O0001O01O01O00010O001O00010O000010O00100O2N2O1N1O2O1N2N101N2N3M3M2N3N1N3M3M2N3M3M6J9GX\\\\5"}, "label": "a orange in a bowl of fruit"}]} {"id": 504744, "image": "COCO_train2014_000000504744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ripe apple between two oranges\"."}], "task": "refering", "answer_template": "The \"a ripe apple between two oranges\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 214, 215, 234, 235, 236, 237, 257, 258, 259, 260, 261, 262, 280, 281, 282, 283, 284, 285, 305, 306, 307], "bbox": [0.18367187499999998, 0.6070491803278688, 0.40815624999999994, 0.8988290398126463], "iscrowd": 0, "area": 12469.035950000001, "rle": {"size": [427, 640], "counts": "Taa15T=7J5J3K4M4K4M4K4M4L3L4M4K4M4K4M4L3M3M4L3M4L3M4O001N10001O001N101O000O2O001O0O2O00000O2O00001N1000001N1000001N10000010O000001O0000001O01O0001O0000001O0001O01O00001O0000010O0000001O0000010O0000001O01O00000PMdFQ2\\9eMRGW2n8dM\\GW2d8hMaGT2`8kMcGR2]8mMdGS2\\8mMdGS2\\8lMeGT2[8lMfGS2Z8lMgGT2Z8kMfGU2Z8jMhGU2X8kMhGU2g901O0000001O00001O0O1O2M2O1N2O2M2O2M2O1M4J5K5K6J5K6J5K5K[gm4"}, "label": "a ripe apple between two oranges"}]} {"id": 504744, "image": "COCO_train2014_000000504744.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red apple between two oranges on bottom left of bowl\"."}], "task": "refering", "answer_template": "The \"red apple between two oranges on bottom left of bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 214, 215, 234, 235, 236, 237, 257, 258, 259, 260, 261, 262, 280, 281, 282, 283, 284, 285, 305, 306, 307], "bbox": [0.18367187499999998, 0.6070491803278688, 0.40815624999999994, 0.8988290398126463], "iscrowd": 0, "area": 12469.035950000001, "rle": {"size": [427, 640], "counts": "Taa15T=7J5J3K4M4K4M4K4M4L3L4M4K4M4K4M4L3M3M4L3M4L3M4O001N10001O001N101O000O2O001O0O2O00000O2O00001N1000001N1000001N10000010O000001O0000001O01O0001O0000001O0001O01O00001O0000010O0000001O0000010O0000001O01O00000PMdFQ2\\9eMRGW2n8dM\\GW2d8hMaGT2`8kMcGR2]8mMdGS2\\8mMdGS2\\8lMeGT2[8lMfGS2Z8lMgGT2Z8kMfGU2Z8jMhGU2X8kMhGU2g901O0000001O00001O0O1O2M2O1N2O2M2O2M2O1M4J5K5K6J5K6J5K5K[gm4"}, "label": "red apple between two oranges on bottom left of bowl"}]} {"id": 562100, "image": "COCO_train2014_000000562100.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"apples to the right side of the bananas\"."}], "task": "refering", "answer_template": "The \"apples to the right side of the bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 177, 178, 179, 200, 201, 202, 217, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 358, 359, 360, 361, 362, 363, 364, 365, 366, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.44293750000000004, 0.36158004158004153, 0.972328125, 1.0], "iscrowd": 0, "area": 56435.850600000005, "rle": {"size": [481, 640], "counts": "XWU41n>7H7I7I8H7I8H7J6L5K4L5J5L5K6J6J6J6J3M2O2M2O1O2M10O1000O1000O0100000O010000O10O10O100O001O00000000000001O000001O00000000000001O00000010O01O0010O0010O010O01O010O001O001O0010O01O001O0O2O001O1O1O1O1O2N1O1O1N2O2N1O1O1O1O2N1O1O1O1N3N1O1O1O1O001O001O001N101O001O001O00001O0O2O001O001O0O2N1O2N1O2N1O2N1O2N1O2M2O2N1O1O2N1O2N1N2N3M2N2N2N2N2N2N2N2M3N3L3N2M3N2M3N2M3^NfH[K\\7e4VIfJm6Z5^1O00000000000000000O101O0000000000000000001O00000O1000000O10^H[Jh5e5WJ]Jh5c5VJ_Jj5a5UJaJj5^5UJdJk5\\5TJeJl5[5RJhJm5X5RJiJn5W5PJkJP6T5nIPKQ6P5mIRKS6n4jIUKV6k4hIXKW6g4hI[KX6e4eI^K[6a4dIbK[6^4`IgK`6X4[IoKd6P4WIVLi6i3QI^Lo6b3kHeLT7d50001O00001O001O00001O00001O001O001O1O1O1O001O1O1O1N101O1O1O001O1O1O1O1O1N2N1O2N2N2N2N2N2O1N1O2N2N2N2N2N2N2N1O2N2N2N2N2N3M2O2M2N3M3N1N3M2N3M3N1N3M2N3L4K4L5K5K4L3M3M2N3M2eMTD`1n;^NZDZ1i;dN]DU1e;kNaDm0a;TOkD>W;BXELj:5Q20O00100O1O010000O100O100O104K5L4K^T8"}, "label": "apples to the right side of the bananas"}]} {"id": 562100, "image": "COCO_train2014_000000562100.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and green apple with a brown stem\"."}], "task": "refering", "answer_template": "The \"a red and green apple with a brown stem\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 177, 178, 179, 200, 201, 202, 217, 222, 223, 224, 225, 226, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 358, 359, 360, 361, 362, 363, 364, 365, 366, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.44293750000000004, 0.36158004158004153, 0.972328125, 1.0], "iscrowd": 0, "area": 56435.850600000005, "rle": {"size": [481, 640], "counts": "XWU41n>7H7I7I8H7I8H7J6L5K4L5J5L5K6J6J6J6J3M2O2M2O1O2M10O1000O1000O0100000O010000O10O10O100O001O00000000000001O000001O00000000000001O00000010O01O0010O0010O010O01O010O001O001O0010O01O001O0O2O001O1O1O1O1O2N1O1O1N2O2N1O1O1O1O2N1O1O1O1N3N1O1O1O1O001O001O001N101O001O001O00001O0O2O001O001O0O2N1O2N1O2N1O2N1O2N1O2M2O2N1O1O2N1O2N1N2N3M2N2N2N2N2N2N2N2M3N3L3N2M3N2M3N2M3^NfH[K\\7e4VIfJm6Z5^1O00000000000000000O101O0000000000000000001O00000O1000000O10^H[Jh5e5WJ]Jh5c5VJ_Jj5a5UJaJj5^5UJdJk5\\5TJeJl5[5RJhJm5X5RJiJn5W5PJkJP6T5nIPKQ6P5mIRKS6n4jIUKV6k4hIXKW6g4hI[KX6e4eI^K[6a4dIbK[6^4`IgK`6X4[IoKd6P4WIVLi6i3QI^Lo6b3kHeLT7d50001O00001O001O00001O00001O001O001O1O1O1O001O1O1O1N101O1O1O001O1O1O1O1O1N2N1O2N2N2N2N2N2O1N1O2N2N2N2N2N2N2N1O2N2N2N2N2N3M2O2M2N3M3N1N3M2N3M3N1N3M2N3L4K4L5K5K4L3M3M2N3M2eMTD`1n;^NZDZ1i;dN]DU1e;kNaDm0a;TOkD>W;BXELj:5Q20O00100O1O010000O100O100O104K5L4K^T8"}, "label": "a red and green apple with a brown stem"}]} {"id": 562100, "image": "COCO_train2014_000000562100.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green apple next to a yellow banana\"."}], "task": "refering", "answer_template": "The \"green apple next to a yellow banana\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [255, 256, 277, 278, 279, 299, 300, 301, 302, 303, 306, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 353, 368, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.013218750000000001, 0.6925363825363826, 0.37065625, 1.0], "iscrowd": 0, "area": 21697.711349999998, "rle": {"size": [481, 640], "counts": "`U43j>o0QO9G9G:F8I4M2N2N2N2N3M2N2N2N2N2N3M2N2N2O1O2M2O1O1O1N2O1O1N2O1O1O100O1O1O101N1O1O100O1O1O100O1O1O100O1O1O1O10000O10000O10000O10000O1000000O10000O10000O1000000000000000000000000000000000000000YLbEk2]:SMhEk2X:RMoEj2Q:SMVFi2j9TM]Fh2c9TMeFh2[9UMlFg2T9VMRGg2n8VMWGh2b:O1O2N1O1O2N1O1O2N1O1O1O000000001O0000000000001O0000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000O1O1O1O100O1O2N100O1O1O1O100O1O1O100O100O10000O1002N2N1M4M3M3L4M2N3L4M3M3L3N3L4M3M3L4M2N3L4L4J6J5K6JUll5"}, "label": "green apple next to a yellow banana"}]} {"id": 562100, "image": "COCO_train2014_000000562100.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow green apple under a bunch of bananas\"."}], "task": "refering", "answer_template": "The \"a yellow green apple under a bunch of bananas\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [255, 256, 277, 278, 279, 299, 300, 301, 302, 303, 306, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 353, 368, 369, 370, 371, 372, 373, 374, 375, 376], "bbox": [0.013218750000000001, 0.6925363825363826, 0.37065625, 1.0], "iscrowd": 0, "area": 21697.711349999998, "rle": {"size": [481, 640], "counts": "`U43j>o0QO9G9G:F8I4M2N2N2N2N3M2N2N2N2N2N3M2N2N2O1O2M2O1O1O1N2O1O1N2O1O1O100O1O1O101N1O1O100O1O1O100O1O1O100O1O1O1O10000O10000O10000O10000O1000000O10000O10000O1000000000000000000000000000000000000000YLbEk2]:SMhEk2X:RMoEj2Q:SMVFi2j9TM]Fh2c9TMeFh2[9UMlFg2T9VMRGg2n8VMWGh2b:O1O2N1O1O2N1O1O2N1O1O1O000000001O0000000000001O0000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000000O1O1O1O100O1O2N100O1O1O1O100O1O1O100O100O10000O1002N2N1M4M3M3L4M2N3L4M3M3L3N3L4M3M3L4M2N3L4L4J6J5K6JUll5"}, "label": "a yellow green apple under a bunch of bananas"}]} {"id": 283573, "image": "COCO_train2014_000000283573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy keeping for the ruby game\"."}], "task": "refering", "answer_template": "The \"boy keeping for the ruby game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 151, 152, 168, 169, 185, 186, 201, 202, 203, 217, 218, 219, 220, 234, 235, 237, 251, 252, 253, 254, 270, 271, 288], "bbox": [0.7850208333333334, 0.32290625, 1.0, 0.716171875], "iscrowd": 0, "area": 12658.826000000001, "rle": {"size": [640, 480], "counts": "VP\\73kc02N3L3N3M3L3N3M2M4M3M2M4M3M2M4M2N3L4M2N3M2N3N2N1N3N2N1N3N1O2N2M2O2N1N3N2N0OL5L4L3M4L4K5L3M4LlN^NS@]1S`0ZOT_O`0R`0cNf@\\2a?fM^@S2h?oMV@j1Q`0XNn_Og1j1UN^;7fBf1j1XN[;4iBe1k1[NX;2lBd1i1`NV;NoBc1j1dNQ;LTC`1j1hNn:IWCa1i1kNk:F[C_1h1nNk:E[C_1h1nNj:F\\C]1i1nNj:G\\C[1i1POi:G\\C[1i1oNj:H\\CY1h1QOj:I\\CX1h1QOj:I]CV1h1ROj:J\\CU1i1SOi:J]CT1g1TOj:K]CR1h1UOi:\\2UEeMj:\\2UEfMj:[2TEfMl:Z2REhMn:X2QEjMn:W2PEjMP;V2oDlMP;U2mDmMS;S2hDSNW;m1eDWN[;j1_D\\N`;d1\\D`Nd;a1VDdNj;\\1QDjNn;V1nCnNR in this image.", "objects": [{"patches": [134, 135, 151, 152, 168, 169, 185, 186, 201, 202, 203, 217, 218, 219, 220, 234, 235, 237, 251, 252, 253, 254, 270, 271, 288], "bbox": [0.7850208333333334, 0.32290625, 1.0, 0.716171875], "iscrowd": 0, "area": 12658.826000000001, "rle": {"size": [640, 480], "counts": "VP\\73kc02N3L3N3M3L3N3M2M4M3M2M4M3M2M4M2N3L4M2N3M2N3N2N1N3N2N1N3N1O2N2M2O2N1N3N2N0OL5L4L3M4L4K5L3M4LlN^NS@]1S`0ZOT_O`0R`0cNf@\\2a?fM^@S2h?oMV@j1Q`0XNn_Og1j1UN^;7fBf1j1XN[;4iBe1k1[NX;2lBd1i1`NV;NoBc1j1dNQ;LTC`1j1hNn:IWCa1i1kNk:F[C_1h1nNk:E[C_1h1nNj:F\\C]1i1nNj:G\\C[1i1POi:G\\C[1i1oNj:H\\CY1h1QOj:I\\CX1h1QOj:I]CV1h1ROj:J\\CU1i1SOi:J]CT1g1TOj:K]CR1h1UOi:\\2UEeMj:\\2UEfMj:[2TEfMl:Z2REhMn:X2QEjMn:W2PEjMP;V2oDlMP;U2mDmMS;S2hDSNW;m1eDWN[;j1_D\\N`;d1\\D`Nd;a1VDdNj;\\1QDjNn;V1nCnNR in this image.", "objects": [{"patches": [59, 60, 61, 75, 76, 77, 78, 93, 94, 95, 110, 111, 112, 126, 127, 128, 129, 130, 142, 143, 144, 145, 146, 147, 158, 159, 160, 161, 162, 163, 164, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 211, 212, 213, 214, 215, 228, 229, 230, 231, 232, 245, 246, 247, 248, 249, 261, 262, 265, 266, 267, 277, 278, 279, 283, 284, 296, 301], "bbox": [0.30347916666666663, 0.14170312499999999, 0.7635208333333333, 0.7460625000000001], "iscrowd": 0, "area": 35142.69314999999, "rle": {"size": [640, 480], "counts": "Vfk22nc02M2O2N1O1N2O1O1O1N2O1O1O1O000000001PJLRH4m7NRH2n7OQH1n72PHNP83oGMP85oGKQ86nGJQ88nGIQ88mGIR89lGHT8:iGGV8:hGHX89fGHY89fGHZ88eGIZ89dGH\\88bGK\\87bGJ]87bGJ]88aGI^88`GJ_88_GI`88_GI`88_GI`88^GJa87^GK`86_GKa85^GLa85]GMb84]GMb84]GMb84]GMb84\\GNc83\\GNc83\\GOb82]GOb82\\G0c81QG;n8FZFR1e9nNoE_1P:aNnEb1Q:_NmEc1R:]NmEe1R:[NmEh1Q:XNmEk1R:UNmEm1S:SNkEo1T:QNjER2U:nMkEc1UNlLk8BWJo1hNa1`NhLa8IUJn1jN^1i7dL\\Io1iN\\1m7gLXIm1kNY1Q8kLRIl1lNX1U8mLmHk1nNU1X8RMhHi1oNT1\\8SMdHi1POR1_8VM_Hi1POQ1b8XM\\Hg1ROP1d8ZMYHe1ROQ1g8[MUHe1SOo0i8]MSHd1SOP1k8]MPHd1TOn0n8_MlGc1UOn0Q9`MhGb1WOm0R9bMSF_Om0S2Ll0V9cMiEET1l1Ml0W9dM\\E0]1b1Nj0Y9\\NfFk01g0Z9aNcFh02g0[9dN`Fg04c0]9lNYFb09a0_9POUFa0<=_9UORFa0=8b9ZOoEd091i9]OkEh07IP:AfEl05AX:E_En07[O[:J[Em08WO`:MUEo09SOc:1PEP19oNi:3kDP12UOS;S62O001N1100O10O010000O010O1000O1010O1O2O1N1O2O0O2N2O0O2O1N3M2O1N3M2O2M2N3N1N3N1N2N100O1O010OlGQFX6o9eIUFY6l9dIWF\\6h9bI\\F]6d9aI^F_6b9^IbF`6_9^IdFa6[9[IjFe6^9hHnFV7o:01O1N101O001N3N2hMgCdL[O2O0O2N100O2O1N2O0O2N2O1N2N1O2N2N2N1O2N2M3N1O2N2N2N2N2N2N2N3M2N2NViV2"}, "label": "a batter"}]} {"id": 283573, "image": "COCO_train2014_000000283573.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young boy swinging a baseball bat\"."}], "task": "refering", "answer_template": "The \"a young boy swinging a baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 75, 76, 77, 78, 93, 94, 95, 110, 111, 112, 126, 127, 128, 129, 130, 142, 143, 144, 145, 146, 147, 158, 159, 160, 161, 162, 163, 164, 176, 177, 178, 179, 180, 181, 195, 196, 197, 198, 211, 212, 213, 214, 215, 228, 229, 230, 231, 232, 245, 246, 247, 248, 249, 261, 262, 265, 266, 267, 277, 278, 279, 283, 284, 296, 301], "bbox": [0.30347916666666663, 0.14170312499999999, 0.7635208333333333, 0.7460625000000001], "iscrowd": 0, "area": 35142.69314999999, "rle": {"size": [640, 480], "counts": "Vfk22nc02M2O2N1O1N2O1O1O1N2O1O1O1O000000001PJLRH4m7NRH2n7OQH1n72PHNP83oGMP85oGKQ86nGJQ88nGIQ88mGIR89lGHT8:iGGV8:hGHX89fGHY89fGHZ88eGIZ89dGH\\88bGK\\87bGJ]87bGJ]88aGI^88`GJ_88_GI`88_GI`88_GI`88^GJa87^GK`86_GKa85^GLa85]GMb84]GMb84]GMb84]GMb84\\GNc83\\GNc83\\GOb82]GOb82\\G0c81QG;n8FZFR1e9nNoE_1P:aNnEb1Q:_NmEc1R:]NmEe1R:[NmEh1Q:XNmEk1R:UNmEm1S:SNkEo1T:QNjER2U:nMkEc1UNlLk8BWJo1hNa1`NhLa8IUJn1jN^1i7dL\\Io1iN\\1m7gLXIm1kNY1Q8kLRIl1lNX1U8mLmHk1nNU1X8RMhHi1oNT1\\8SMdHi1POR1_8VM_Hi1POQ1b8XM\\Hg1ROP1d8ZMYHe1ROQ1g8[MUHe1SOo0i8]MSHd1SOP1k8]MPHd1TOn0n8_MlGc1UOn0Q9`MhGb1WOm0R9bMSF_Om0S2Ll0V9cMiEET1l1Ml0W9dM\\E0]1b1Nj0Y9\\NfFk01g0Z9aNcFh02g0[9dN`Fg04c0]9lNYFb09a0_9POUFa0<=_9UORFa0=8b9ZOoEd091i9]OkEh07IP:AfEl05AX:E_En07[O[:J[Em08WO`:MUEo09SOc:1PEP19oNi:3kDP12UOS;S62O001N1100O10O010000O010O1000O1010O1O2O1N1O2O0O2N2O0O2O1N3M2O1N3M2O2M2N3N1N3N1N2N100O1O010OlGQFX6o9eIUFY6l9dIWF\\6h9bI\\F]6d9aI^F_6b9^IbF`6_9^IdFa6[9[IjFe6^9hHnFV7o:01O1N101O001N3N2hMgCdL[O2O0O2N100O2O1N2O0O2N2O1N2N1O2N2N2N1O2N2M3N1O2N2N2N2N2N2N2N3M2N2NViV2"}, "label": "a young boy swinging a baseball bat"}]} {"id": 95185, "image": "COCO_train2014_000000095185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wood rocking chair painted red\"."}], "task": "refering", "answer_template": "The \"a wood rocking chair painted red\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [299, 300, 301, 302, 303, 304, 305, 306, 307, 321, 322, 323, 324, 325, 326, 327, 328, 329, 344, 345, 346, 347, 348, 349, 350, 366, 367, 368, 369, 370, 371, 372, 388, 389, 390, 391, 392, 393, 394, 410, 411, 412, 413, 414, 415, 416, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 453, 454, 455, 456, 457, 458, 459, 460, 461, 475, 476, 477, 478, 479, 480, 481, 482], "bbox": [0.5655718954248365, 0.6088725490196079, 0.9791176470588235, 0.9800653594771241], "iscrowd": 0, "area": 44406.01529999999, "rle": {"size": [612, 612], "counts": "RZ_62Qc04L4K4M3L2O2M20010O001O2N4L5K5K5K4L5K5K5L3L2NN2000000000000000001OO100O1O100cKmMhGT2U8UNeGk1Y8\\NaGe1\\8_NbGb1]8aNaG_1^8cN`G^1`8cN^G^1a8dN\\G^1d8cNVGb1i8`NRGd1n8]NmFg1R9T1oCQOQ in this image.", "objects": [{"patches": [299, 300, 301, 302, 303, 304, 305, 306, 307, 321, 322, 323, 324, 325, 326, 327, 328, 329, 344, 345, 346, 347, 348, 349, 350, 366, 367, 368, 369, 370, 371, 372, 388, 389, 390, 391, 392, 393, 394, 410, 411, 412, 413, 414, 415, 416, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 453, 454, 455, 456, 457, 458, 459, 460, 461, 475, 476, 477, 478, 479, 480, 481, 482], "bbox": [0.5655718954248365, 0.6088725490196079, 0.9791176470588235, 0.9800653594771241], "iscrowd": 0, "area": 44406.01529999999, "rle": {"size": [612, 612], "counts": "RZ_62Qc04L4K4M3L2O2M20010O001O2N4L5K5K5K4L5K5K5L3L2NN2000000000000000001OO100O1O100cKmMhGT2U8UNeGk1Y8\\NaGe1\\8_NbGb1]8aNaG_1^8cN`G^1`8cN^G^1a8dN\\G^1d8cNVGb1i8`NRGd1n8]NmFg1R9T1oCQOQ in this image.", "objects": [{"patches": [20, 21, 22, 42, 43, 44, 45, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206], "bbox": [0.681046875, 0.021520833333333333, 1.0, 0.5125625], "iscrowd": 0, "area": 33124.37965, "rle": {"size": [480, 640], "counts": "ob\\63j>5K4K5L4L4L5K4M3L4M3L3N2M3N2M3N2M3N3L3N2M3N2N2M3N2M4M2M3N2N2M3N2N2N3L3N2N2N2M3N2N2N3M2M3N2N2N2M3N2N3M2O100O100O100O2O0O100O100O100O2O0O100O100O100O2O0O100O100O1000000O1000000O1000000O1000000O1000000O0100000O1000000O1000000O100XOh0L4L4L4L4M3N2O1O1N2O1O1N2O1O1O1O10000O100O10000O10000O010O10000O10000O100O11O001O00001O001O001O00001O001N10001O001N101O000O2O001O001XKkGb2X8ZMjGf2W8XMkGg2V8WMkGi2V8TMlGl2V8QMkGn2W8PMkGo2V8nLlGR3V8kLkGU3V8hLlGX3V8eLlGY3_8[LcGe3g8iKaGW4d901N1000001N10001O2N2MlL"}, "label": "the man with the mustache"}]} {"id": 578523, "image": "COCO_train2014_000000578523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing glasses and a pin striped shirt is smiling\"."}], "task": "refering", "answer_template": "The \"a man wearing glasses and a pin striped shirt is smiling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 42, 43, 44, 45, 64, 65, 66, 67, 68, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206], "bbox": [0.681046875, 0.021520833333333333, 1.0, 0.5125625], "iscrowd": 0, "area": 33124.37965, "rle": {"size": [480, 640], "counts": "ob\\63j>5K4K5L4L4L5K4M3L4M3L3N2M3N2M3N2M3N3L3N2M3N2N2M3N2M4M2M3N2N2M3N2N2N3L3N2N2N2M3N2N2N3M2M3N2N2N2M3N2N3M2O100O100O100O2O0O100O100O100O2O0O100O100O100O2O0O100O100O1000000O1000000O1000000O1000000O1000000O0100000O1000000O1000000O100XOh0L4L4L4L4M3N2O1O1N2O1O1N2O1O1O1O10000O100O10000O10000O010O10000O10000O100O11O001O00001O001O001O00001O001N10001O001N101O000O2O001O001XKkGb2X8ZMjGf2W8XMkGg2V8WMkGi2V8TMlGl2V8QMkGn2W8PMkGo2V8nLlGR3V8kLkGU3V8hLlGX3V8eLlGY3_8[LcGe3g8iKaGW4d901N1000001N10001O2N2MlL"}, "label": "a man wearing glasses and a pin striped shirt is smiling"}]} {"id": 578523, "image": "COCO_train2014_000000578523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person in the blue shirt\"."}], "task": "refering", "answer_template": "The \"the person in the blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 27, 28, 29, 30, 50, 51, 52, 53, 54, 73, 74, 75, 76, 77, 95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 284, 302, 303, 304, 305, 306, 307, 308, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 352, 353, 354, 355, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.087640625, 0.0516875, 0.483703125, 1.0], "iscrowd": 0, "area": 65392.12195000001, "rle": {"size": [480, 640], "counts": "S_j01n>3N1N3M2O2M2N3M3N1N3M2O2M2N3N1N4L5K5L4K5K5L4K5K5K5K5K5K5J6K5K5K6J5K5K5K4K5L4K6K4K5K5L4K6K4K5K5L4K5L5fNkI_IZ6Y6RJ^IS6Z6XJ_Il5Z6^J^Ig5Z6h1E>B?^Ob0^Oa0@9F:K5K5L4K5L5J5K5L4K5K5L4K5N2N2M3N3L3N2N2M2O000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O10000O100O10000O100O10000001O0`GYOg2g0YM[Oe2f0ZM[Oe2e0ZM]Oe2c0[M_Oc2b0\\M_Oc2a0]M@b2`0^MAa2`0^MB`2>_MD`2<`ME_2<`ME_2;aMG]2:bMG]29cMH\\28cMJ\\27cMJ\\26dMK[25eMMY24fMMY23fMOY22fMOY22fMOY21gM0X21gM1W20hM1W20gM2X2OgM3W2NhM3W2NhM3W2NhM4V2MhM5W2LhM6V2KiM6V2KiM7U2JhKcNWOe1o4JgKgNUOa1R5JeKlNTO\\1U5IdKROQOX1Y5HbKWOPOS1\\5H`K\\OoNn0_5H_K@mNj0b5H]KElNe0e5H[KJkN`0h5HYKOjN;k5HmJ>SOLn5GdJl0YO_OQ6G`JS1ZOXOT6G[Jk3c5WLWJo3g5SLRJR4n5_21000000000000O1NdK`G\\2_8[MmGc2R8TMZHj2d7SMcHk2\\7UMgHi2X7XMjHf2U7ZMnHd2P7^MRI`2m6aMUI]2j6cMXI\\2f6fM\\IX2c6hM_IW2`6jMbIT2\\6nMeIQ2[6nMhIP2W6QNjIn1V6QNmIm1R6TNPJj1P6VNQJi1n5WNUJg1k5YNWJe1h5[NZJd1f5\\N\\Jb1c5_N_J_1a5`NaJ_1^5bNdJ\\1\\5cNgJ[1X5fNiJY1W5gNkJW1U5hNnJV1R5jNoJU1Q5jNRKT1o4kNRKT1n4lNTKR1l4mNVKR1j4nNWKQ1i4nNYKQ1h4nNZKP1f4PO[Ko0e4PO]Ko0c4QO_Km0a4ROaKm0_4SObKl0_4SObKl0^4SOeKk0\\4TOeKk0^4ROcKm0b4mN`KR1m4aNUK]1X9O2N1O2N2N1O2N2N2N1O2N2N2N in this image.", "objects": [{"patches": [6, 27, 28, 29, 30, 50, 51, 52, 53, 54, 73, 74, 75, 76, 77, 95, 96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 284, 302, 303, 304, 305, 306, 307, 308, 326, 327, 328, 329, 330, 331, 332, 349, 350, 351, 352, 353, 354, 355, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.087640625, 0.0516875, 0.483703125, 1.0], "iscrowd": 0, "area": 65392.12195000001, "rle": {"size": [480, 640], "counts": "S_j01n>3N1N3M2O2M2N3M3N1N3M2O2M2N3N1N4L5K5L4K5K5L4K5K5K5K5K5K5J6K5K5K6J5K5K5K4K5L4K6K4K5K5L4K6K4K5K5L4K5L5fNkI_IZ6Y6RJ^IS6Z6XJ_Il5Z6^J^Ig5Z6h1E>B?^Ob0^Oa0@9F:K5K5L4K5L5J5K5L4K5K5L4K5N2N2M3N3L3N2N2M2O000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O10000O100O10000O100O10000001O0`GYOg2g0YM[Oe2f0ZM[Oe2e0ZM]Oe2c0[M_Oc2b0\\M_Oc2a0]M@b2`0^MAa2`0^MB`2>_MD`2<`ME_2<`ME_2;aMG]2:bMG]29cMH\\28cMJ\\27cMJ\\26dMK[25eMMY24fMMY23fMOY22fMOY22fMOY21gM0X21gM1W20hM1W20gM2X2OgM3W2NhM3W2NhM3W2NhM4V2MhM5W2LhM6V2KiM6V2KiM7U2JhKcNWOe1o4JgKgNUOa1R5JeKlNTO\\1U5IdKROQOX1Y5HbKWOPOS1\\5H`K\\OoNn0_5H_K@mNj0b5H]KElNe0e5H[KJkN`0h5HYKOjN;k5HmJ>SOLn5GdJl0YO_OQ6G`JS1ZOXOT6G[Jk3c5WLWJo3g5SLRJR4n5_21000000000000O1NdK`G\\2_8[MmGc2R8TMZHj2d7SMcHk2\\7UMgHi2X7XMjHf2U7ZMnHd2P7^MRI`2m6aMUI]2j6cMXI\\2f6fM\\IX2c6hM_IW2`6jMbIT2\\6nMeIQ2[6nMhIP2W6QNjIn1V6QNmIm1R6TNPJj1P6VNQJi1n5WNUJg1k5YNWJe1h5[NZJd1f5\\N\\Jb1c5_N_J_1a5`NaJ_1^5bNdJ\\1\\5cNgJ[1X5fNiJY1W5gNkJW1U5hNnJV1R5jNoJU1Q5jNRKT1o4kNRKT1n4lNTKR1l4mNVKR1j4nNWKQ1i4nNYKQ1h4nNZKP1f4PO[Ko0e4PO]Ko0c4QO_Km0a4ROaKm0_4SObKl0_4SObKl0^4SOeKk0\\4TOeKk0^4ROcKm0b4mN`KR1m4aNUK]1X9O2N1O2N2N1O2N2N2N1O2N2N2N in this image.", "objects": [{"patches": [163, 185, 186, 207, 208, 209, 230, 231, 232, 253, 254, 255, 276, 277, 278, 279, 300, 301, 302, 323, 324, 325, 346, 347, 348, 349, 370, 371, 372], "bbox": [0.027453125000000002, 0.46029166666666665, 0.21178124999999998, 1.0], "iscrowd": 0, "area": 14611.002300000006, "rle": {"size": [480, 640], "counts": "cg8?\\>?A`0@?A:E9I7I7J5K6J6I7J6J5J7L4L4M2M4L4L4L4L3M4L4L4L3N3L4L3M2O1N3M2O1N2N3N1N2N3N1N2N2O2M2N2O2M2N2O1O2M2O1N3NPJaH`4^7\\KlH_4R7^KXI]4g6_KeI[4^6\\KnI]4V6[KUJ^4n5ZK_J_4Y8H2N2N2O0O2N2O1N2N2N2M3M3M3M3M2N3M3M3M3N2M3M3M3M3M3M3M3M3M2N3M3M3M3M2N2N2N2N2N2N3M2N1O2N2N1O2N1O3M4L4L4L;E2N1O2NRX\\7"}, "label": "chair with woman in blue shirt sitting on it"}]} {"id": 578523, "image": "COCO_train2014_000000578523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"back of wooden chair being occupied by woman in blue\"."}], "task": "refering", "answer_template": "The \"back of wooden chair being occupied by woman in blue\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 185, 186, 207, 208, 209, 230, 231, 232, 253, 254, 255, 276, 277, 278, 279, 300, 301, 302, 323, 324, 325, 346, 347, 348, 349, 370, 371, 372], "bbox": [0.027453125000000002, 0.46029166666666665, 0.21178124999999998, 1.0], "iscrowd": 0, "area": 14611.002300000006, "rle": {"size": [480, 640], "counts": "cg8?\\>?A`0@?A:E9I7I7J5K6J6I7J6J5J7L4L4M2M4L4L4L4L3M4L4L4L3N3L4L3M2O1N3M2O1N2N3N1N2N3N1N2N2O2M2N2O2M2N2O1O2M2O1N3NPJaH`4^7\\KlH_4R7^KXI]4g6_KeI[4^6\\KnI]4V6[KUJ^4n5ZK_J_4Y8H2N2N2O0O2N2O1N2N2N2M3M3M3M3M2N3M3M3M3N2M3M3M3M3M3M3M3M3M2N3M3M3M3M2N2N2N2N2N2N3M2N1O2N2N1O2N1O3M4L4L4L;E2N1O2NRX\\7"}, "label": "back of wooden chair being occupied by woman in blue"}]} {"id": 373727, "image": "COCO_train2014_000000373727.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bottom oven\"."}], "task": "refering", "answer_template": "The \"a bottom oven\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273], "bbox": [0.683640625, 0.5450585480093677, 0.88559375, 0.7984309133489461], "iscrowd": 0, "area": 12163.923350000001, "rle": {"size": [427, 640], "counts": "anf59h<;E:F:G9F:F:F:F:I701O00000000000000001O0000000000000000001O00000000000000001O0000000000000000001O00000000000000001O0000000000000000001O00000000000000001O00000000000000001O00000000000001O0001O00000000000000001O0000000000000000001O000000M3F:F:F:E;F:F:F:F:F:FSWn0"}, "label": "a bottom oven"}]} {"id": 373727, "image": "COCO_train2014_000000373727.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver oven beneath another smaller oven\"."}], "task": "refering", "answer_template": "The \"a silver oven beneath another smaller oven\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 202, 203, 204, 222, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 273], "bbox": [0.683640625, 0.5450585480093677, 0.88559375, 0.7984309133489461], "iscrowd": 0, "area": 12163.923350000001, "rle": {"size": [427, 640], "counts": "anf59h<;E:F:G9F:F:F:F:I701O00000000000000001O0000000000000000001O00000000000000001O0000000000000000001O00000000000000001O0000000000000000001O00000000000000001O00000000000000001O00000000000001O0001O00000000000000001O0000000000000000001O000000M3F:F:F:E;F:F:F:F:F:FSWn0"}, "label": "a silver oven beneath another smaller oven"}]} {"id": 381923, "image": "COCO_train2014_000000381923.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a white shirt and black shorts holding a tennis racket on the ground\"."}], "task": "refering", "answer_template": "The \"a man with a white shirt and black shorts holding a tennis racket on the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [94, 95, 96, 97, 98, 99, 100, 108, 109, 110, 111, 112, 113, 114, 115, 123, 124, 125, 126, 127, 128, 129, 130, 138, 139, 140, 141, 142, 143, 153, 154, 155, 156, 157, 158, 168, 169, 170, 171, 172, 173, 182, 183, 184, 185, 186, 187, 188, 189, 196, 197, 198, 199, 200, 201, 202, 203, 204, 211, 212, 213, 214, 215, 216, 217, 218, 219, 226, 227, 230, 231, 233, 234, 241, 242, 244, 245, 249, 259, 260, 261, 275, 276], "bbox": [0.07105882352941176, 0.2555, 0.7208, 0.7948125000000001], "iscrowd": 0, "area": 45329.255200000014, "rle": {"size": [640, 425], "counts": "kTc04hc0;F;D in this image.", "objects": [{"patches": [94, 95, 96, 97, 98, 99, 100, 108, 109, 110, 111, 112, 113, 114, 115, 123, 124, 125, 126, 127, 128, 129, 130, 138, 139, 140, 141, 142, 143, 153, 154, 155, 156, 157, 158, 168, 169, 170, 171, 172, 173, 182, 183, 184, 185, 186, 187, 188, 189, 196, 197, 198, 199, 200, 201, 202, 203, 204, 211, 212, 213, 214, 215, 216, 217, 218, 219, 226, 227, 230, 231, 233, 234, 241, 242, 244, 245, 249, 259, 260, 261, 275, 276], "bbox": [0.07105882352941176, 0.2555, 0.7208, 0.7948125000000001], "iscrowd": 0, "area": 45329.255200000014, "rle": {"size": [640, 425], "counts": "kTc04hc0;F;D in this image.", "objects": [{"patches": [168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 254, 255, 256, 257, 258, 259, 260, 264, 265, 266, 277, 278, 279, 281, 282, 283, 304, 305, 306], "bbox": [0.054328125000000005, 0.4403958333333333, 0.608609375, 0.7922291666666667], "iscrowd": 0, "area": 29433.42095, "rle": {"size": [480, 640], "counts": "ag`05g>5K5O2N1N3N0O10000O2O000O100O100O2N100O100O1O101N1O100O1O100O100O1O100O1O100O1O1O1O1O1O1O1O1O1O1O1O001O1O001O010O1O001O001O1O00100O001O001O1O001L3L5K4L5N110O001O0010O01O001O010O001O0010L3K6J5O21N101N2O1N101O1N101N2O1N101N2O0O2O1N2O001bDbMW:`2XERNf:V3M3N2M3N1O000O2O00000O101O0O1000000O1000O10O10000O10O1M3N1N3N21O00001O00001OfNiEkMW:T2oEgMQ:X2QFgMo9Y2SFeMm9Z2UFeMk9Z2WFeMi9[2YFcMg9\\2[FcMe9\\2]FcMc9]2_FaMa9^2`FbM`9]2bFbM^9^2cFaM]9^2eFaM[9_2fF`MZ9_2hF`MX9`2iF_MW9`2kF_MU9a2lF]MU9b2mFYMW9g2a100O100O100O100O10000O10000O10000O01000O10000O1000O010000O10000O10O10O10000O100000O100000000000000O100000000000000O100000000000000O100000000000000O1000001O0000001N101O01O01O001O001O00001O001O001O00001O0O2O0O2O:E3N1N3N1N201O0000O100O1O10O01O01O01O01OC>N1O1N3N1O1SOSDfNo;X1RDgNo;Y1RDfNo;X1RDiNm;W1TDiNk;W1UDiNm;T1UDlNk;S1UDmNl;Q1VDoNj;P1VDQOk;l0WDSOj;l0VDUOj;i0XDVOj;h0VDYOj;e0XDZOi;e0WD\\Oj;b0WD^Oi;MkD3V;LkD4U;JlD6U;IlD6T;JlD6U;IlD6U;HlD9S;FoD9R;EoD;Q;CRE in this image.", "objects": [{"patches": [1, 2, 18, 19, 20, 36, 37, 38, 54, 55, 56, 72, 73, 74, 75, 90, 91, 92, 93, 108, 109, 110, 111, 126], "bbox": [0.00196, 0.052346666666666666, 0.20884, 0.5450666666666667], "iscrowd": 0, "area": 13100.7521, "rle": {"size": [375, 500], "counts": "U=o0g:7JOXGoNk6R1SI^O^6b0aILS64kI5m5MPJ6o5JoI8P6JmI8R6IjI in this image.", "objects": [{"patches": [1, 2, 18, 19, 20, 36, 37, 38, 54, 55, 56, 72, 73, 74, 75, 90, 91, 92, 93, 108, 109, 110, 111, 126], "bbox": [0.00196, 0.052346666666666666, 0.20884, 0.5450666666666667], "iscrowd": 0, "area": 13100.7521, "rle": {"size": [375, 500], "counts": "U=o0g:7JOXGoNk6R1SI^O^6b0aILS64kI5m5MPJ6o5JoI8P6JmI8R6IjI in this image.", "objects": [{"patches": [67, 68, 84, 85, 86, 87, 102, 103, 104, 105, 120, 121, 122, 123, 137, 138, 139, 140, 141, 157, 158], "bbox": [0.65, 0.2788, 0.87878, 0.6242399999999999], "iscrowd": 0, "area": 10678.979449999999, "rle": {"size": [375, 500], "counts": "hWg31c;3M3N3J5K5M3N2M3N2N2N2M3N2M3G9I7K5K5E;J6J6J6J6N2M3O1O2M200O100O100O100O100O100O100O100O10000O100O100O2O0O100O100O10000000000001O000000001O000000001O0001O01O00001O001O00001O000O2N1O2N1N2O2N3Mb0^O1O2M3N2M2N3L4L4M2gNaF4f9I_FMi9O]FHj95m0N3L4M3M2N[Vf0"}, "label": "a dog in a party hat"}]} {"id": 283624, "image": "COCO_train2014_000000283624.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small dog with a hat on its head\"."}], "task": "refering", "answer_template": "The \"a small dog with a hat on its head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 84, 85, 86, 87, 102, 103, 104, 105, 120, 121, 122, 123, 137, 138, 139, 140, 141, 157, 158], "bbox": [0.65, 0.2788, 0.87878, 0.6242399999999999], "iscrowd": 0, "area": 10678.979449999999, "rle": {"size": [375, 500], "counts": "hWg31c;3M3N3J5K5M3N2M3N2N2N2M3N2M3G9I7K5K5E;J6J6J6J6N2M3O1O2M200O100O100O100O100O100O100O100O10000O100O100O2O0O100O100O10000000000001O000000001O000000001O0001O01O00001O001O00001O000O2N1O2N1N2O2N3Mb0^O1O2M3N2M2N3L4L4M2gNaF4f9I_FMi9O]FHj95m0N3L4M3M2N[Vf0"}, "label": "a small dog with a hat on its head"}]} {"id": 480240, "image": "COCO_train2014_000000480240.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boy in blue jersey number 26 is playing with baseball bat\"."}], "task": "refering", "answer_template": "The \"a boy in blue jersey number 26 is playing with baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 68, 69, 70, 71, 84, 85, 86, 87, 100, 101, 102, 103, 104, 116, 117, 118, 119, 120, 132, 133, 134, 135, 136, 148, 149, 150, 151, 165, 166, 167, 180, 181, 182, 183, 184, 195, 196, 197, 198, 199, 200, 210, 211, 212, 213, 214, 215, 216, 226, 227, 228, 229, 230, 231, 232, 242, 243, 244, 246, 247, 248, 249, 259, 260, 263, 264, 265, 266, 275, 276, 277, 281, 282, 283, 284, 291, 292, 293, 298, 299, 300, 306, 307, 308, 314, 315, 316], "bbox": [0.14317391304347826, 0.153609375, 0.8030434782608696, 0.8672656249999999], "iscrowd": 0, "area": 45460.31075, "rle": {"size": [640, 460], "counts": "mhY11nc01O1O1O1O1O1O1O1O1O1O1O1dK2eDN[;a0VD@i;c0TD^Ok;e0RD\\Om;g0PDZOn;j0oCWOPP4O1O1O0O200O1PISJMm53UJJk56XJTKNY2k5c2YJPK1Z2f5e2\\JmJ3Z2b5h2_JiJ4\\2]5k2aJeJ7\\2b1PL[1o6oLaJ9]2\\1VLZ1k6TM^J;]2W1^LW1g6[MXJ<_2S1eLS1c6dMRJ:d2n0jLR1`6lMjI9h2i0RMn0\\6VNcI8l2c0XMm0X6_N\\I6P3?_Mi0U6gNUI6S3:gMf0P6o0VLZNmMe0m5T1RLWNUNa0i5\\1oKSNZN?f5b1lKoMbNa^OC_a0>_^OCba0<\\^OEea0[11N101N2O0O2O0O2O0O2O0O2O1N101N101N101O0O2O1N101N101N101N101N101N2O0O2O0O2O001N101N2O0O2O9FR\\h1"}, "label": "a boy in blue jersey number 26 is playing with baseball bat"}]} {"id": 480240, "image": "COCO_train2014_000000480240.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the batter\"."}], "task": "refering", "answer_template": "The \"the batter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 68, 69, 70, 71, 84, 85, 86, 87, 100, 101, 102, 103, 104, 116, 117, 118, 119, 120, 132, 133, 134, 135, 136, 148, 149, 150, 151, 165, 166, 167, 180, 181, 182, 183, 184, 195, 196, 197, 198, 199, 200, 210, 211, 212, 213, 214, 215, 216, 226, 227, 228, 229, 230, 231, 232, 242, 243, 244, 246, 247, 248, 249, 259, 260, 263, 264, 265, 266, 275, 276, 277, 281, 282, 283, 284, 291, 292, 293, 298, 299, 300, 306, 307, 308, 314, 315, 316], "bbox": [0.14317391304347826, 0.153609375, 0.8030434782608696, 0.8672656249999999], "iscrowd": 0, "area": 45460.31075, "rle": {"size": [640, 460], "counts": "mhY11nc01O1O1O1O1O1O1O1O1O1O1O1dK2eDN[;a0VD@i;c0TD^Ok;e0RD\\Om;g0PDZOn;j0oCWOPP4O1O1O0O200O1PISJMm53UJJk56XJTKNY2k5c2YJPK1Z2f5e2\\JmJ3Z2b5h2_JiJ4\\2]5k2aJeJ7\\2b1PL[1o6oLaJ9]2\\1VLZ1k6TM^J;]2W1^LW1g6[MXJ<_2S1eLS1c6dMRJ:d2n0jLR1`6lMjI9h2i0RMn0\\6VNcI8l2c0XMm0X6_N\\I6P3?_Mi0U6gNUI6S3:gMf0P6o0VLZNmMe0m5T1RLWNUNa0i5\\1oKSNZN?f5b1lKoMbNa^OC_a0>_^OCba0<\\^OEea0[11N101N2O0O2O0O2O0O2O0O2O1N101N101N101O0O2O1N101N101N101N101N101N2O0O2O0O2O001N101N2O0O2O9FR\\h1"}, "label": "the batter"}]} {"id": 480240, "image": "COCO_train2014_000000480240.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a catcher watching as a batter hits the ball\"."}], "task": "refering", "answer_template": "The \"a catcher watching as a batter hits the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 109, 110, 111, 125, 126, 127, 141, 142, 143, 157, 158, 159, 172, 173, 174, 175, 187, 188, 189, 190, 191, 204, 205, 206, 207, 222, 223, 237, 238, 239, 253, 254, 255, 269, 270, 271, 285, 286, 287, 302, 303, 318, 319], "bbox": [0.7159347826086956, 0.229203125, 1.0, 0.8337031249999999], "iscrowd": 0, "area": 24244.965899999996, "rle": {"size": [640, 460], "counts": "na^62mc03M2N2N2N3M2N2N2N3M2N2N3M3M3M4L3M1N3N1N2N2O@f]O\\OXb0e0k]OZOSb0g0o]OUORb0k0R^OoNQb0P1R^OlNPb0S1?O1O2O0O100000000001O2N101N1O2N1O2N1O2N1N2gAbNm9`1oEkNh9V1UFoNh9R1UFROi9o0TFUOj9l0SFXOj9k0SFXOk9i0SFZOk9g0TF[OU8oNiEg1P2]OT8TOeE`1V2^OR8ZOaEZ1Z2_OR8DRET1k2ZOd6`N]Gj3m1iMc6bN[Gg3Q2hMb6eNYGe3T2gMa6hNWGd3U2fMb6hNdFG3m3g2dM`6jNcFK2h3j2dM_6lNbFL1f3m2cM^6m3WHZJBj1U8n3XH\\J_Og1X8n3XH_J\\Od1[8n3YHaJXOb1_8m3XHeJVO^1a8n3XHiJRO[1e8l3XHmJoNX1h8l3XHPKlNU1l8k3WH_Mh7b2WH_Mi7a2VH`Mi7a2WH_Mi7a2UHaMj7`2UHaMk7_2THbMk7_2THbMl7^2RHdMm7]2RHdMn7\\2QHeMn7]2oGeMQ8[2nGfMQ8\\2mGeMS8[2lGfMJULa7W6dHdMJYL_7S6gHdMI]L]7P6iHcMJ`LZ7m5lHcMIcLY7k5mHbMJcLZ7j5lHcMIdL[7j5kHbMIeL\\7j5jHaMGhL`7g5hHaMFjLb7f5gH`MElLd7e5fH_MEmLf7e5cH^MFnLg7e5bH^MDoLj7d5aH_N_7b1`H^Na7b1^H^Nb7c1\\H^Nd7b1[H_Nf7a1XH`Nh7a1WH_Ni7a1VH`Nk7`1SHaNm7`1RH_No7b1PH^NP8b1PH^NP8c1oG]NQ8d1nG\\NR8d1nG\\NR8e1mG[NT8h6001O1O7In1RNn1QNo1RNn1RNo1QNVA"}, "label": "a catcher watching as a batter hits the ball"}]} {"id": 480240, "image": "COCO_train2014_000000480240.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"catcher behind the batter\"."}], "task": "refering", "answer_template": "The \"catcher behind the batter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 109, 110, 111, 125, 126, 127, 141, 142, 143, 157, 158, 159, 172, 173, 174, 175, 187, 188, 189, 190, 191, 204, 205, 206, 207, 222, 223, 237, 238, 239, 253, 254, 255, 269, 270, 271, 285, 286, 287, 302, 303, 318, 319], "bbox": [0.7159347826086956, 0.229203125, 1.0, 0.8337031249999999], "iscrowd": 0, "area": 24244.965899999996, "rle": {"size": [640, 460], "counts": "na^62mc03M2N2N2N3M2N2N2N3M2N2N3M3M3M4L3M1N3N1N2N2O@f]O\\OXb0e0k]OZOSb0g0o]OUORb0k0R^OoNQb0P1R^OlNPb0S1?O1O2O0O100000000001O2N101N1O2N1O2N1O2N1N2gAbNm9`1oEkNh9V1UFoNh9R1UFROi9o0TFUOj9l0SFXOj9k0SFXOk9i0SFZOk9g0TF[OU8oNiEg1P2]OT8TOeE`1V2^OR8ZOaEZ1Z2_OR8DRET1k2ZOd6`N]Gj3m1iMc6bN[Gg3Q2hMb6eNYGe3T2gMa6hNWGd3U2fMb6hNdFG3m3g2dM`6jNcFK2h3j2dM_6lNbFL1f3m2cM^6m3WHZJBj1U8n3XH\\J_Og1X8n3XH_J\\Od1[8n3YHaJXOb1_8m3XHeJVO^1a8n3XHiJRO[1e8l3XHmJoNX1h8l3XHPKlNU1l8k3WH_Mh7b2WH_Mi7a2VH`Mi7a2WH_Mi7a2UHaMj7`2UHaMk7_2THbMk7_2THbMl7^2RHdMm7]2RHdMn7\\2QHeMn7]2oGeMQ8[2nGfMQ8\\2mGeMS8[2lGfMJULa7W6dHdMJYL_7S6gHdMI]L]7P6iHcMJ`LZ7m5lHcMIcLY7k5mHbMJcLZ7j5lHcMIdL[7j5kHbMIeL\\7j5jHaMGhL`7g5hHaMFjLb7f5gH`MElLd7e5fH_MEmLf7e5cH^MFnLg7e5bH^MDoLj7d5aH_N_7b1`H^Na7b1^H^Nb7c1\\H^Nd7b1[H_Nf7a1XH`Nh7a1WH_Ni7a1VH`Nk7`1SHaNm7`1RH_No7b1PH^NP8b1PH^NP8c1oG]NQ8d1nG\\NR8d1nG\\NR8e1mG[NT8h6001O1O7In1RNn1QNo1RNn1RNo1QNVA"}, "label": "catcher behind the batter"}]} {"id": 357362, "image": "COCO_train2014_000000357362.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the railing that the people are standing on\"."}], "task": "refering", "answer_template": "The \"the railing that the people are standing on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 69, 92, 93, 94, 95, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 141, 142, 143, 144, 145, 146, 147, 161, 166, 167, 169, 184, 189, 192, 207, 208], "bbox": [0.0, 0.12406249999999999, 0.407734375, 0.550625], "iscrowd": 0, "area": 13325.373150000003, "rle": {"size": [480, 640], "counts": "l1c0k2o0U7h4fN1O0000001O000000000000000000000000O`2aM0lLiE_1W:aNoF9Q9GSG5m8KWG1i80ZGLf84\\GJd86]GIc87_GGa89`GF`8:bGD^8cGA]8`0cG^O^8b0bG]O_8c0bG\\O^8d0cGZO^8f0cGXO^8h0bGWO_8i0l2001O00000000000000001N100O100O100O101O000000000000001O0O1000000000001O00000000000000001O00000000000000001O0000000000000010O000nNWBm0i=SOWBm0i=SOWBm0i=ROXBn0m=00001O00000001O00000001O0001O001O001O01O0001O0000001O0001O0001O0000001O0001O0001O00oBKW;5oCf0P in this image.", "objects": [{"patches": [46, 69, 92, 93, 94, 95, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 141, 142, 143, 144, 145, 146, 147, 161, 166, 167, 169, 184, 189, 192, 207, 208], "bbox": [0.0, 0.12406249999999999, 0.407734375, 0.550625], "iscrowd": 0, "area": 13325.373150000003, "rle": {"size": [480, 640], "counts": "l1c0k2o0U7h4fN1O0000001O000000000000000000000000O`2aM0lLiE_1W:aNoF9Q9GSG5m8KWG1i80ZGLf84\\GJd86]GIc87_GGa89`GF`8:bGD^8cGA]8`0cG^O^8b0bG]O_8c0bG\\O^8d0cGZO^8f0cGXO^8h0bGWO_8i0l2001O00000000000000001N100O100O100O101O000000000000001O0O1000000000001O00000000000000001O00000000000000001O0000000000000010O000nNWBm0i=SOWBm0i=SOWBm0i=ROXBn0m=00001O00000001O00000001O0001O001O001O01O0001O0000001O0001O0001O0000001O0001O0001O00oBKW;5oCf0P in this image.", "objects": [{"patches": [154, 172, 173, 174, 176, 177, 194, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 225, 240, 241, 242, 243, 244, 246, 247, 248], "bbox": [0.45628124999999997, 0.5197067448680351, 0.810125, 0.9154252199413488], "iscrowd": 0, "area": 15699.183350000003, "rle": {"size": [341, 640], "counts": "n_Q3:n9?M3L4M2M4M3M3L2O2M3N2O001N2O1VOYNcHh1[7[NcHg1Y7\\NfHf1W7]NhHd1U7^NjHd1R7`NmHBAj1b7dNlHb1S7_NlHb1S7^NlHd1S7]NlHc1T7]NkHe1T7\\NkHe1T7\\NhHh1W7ZNeHg1\\7i0000000000000000O10000000000000000O101O0000000000000O2O0000000O1O1O2N1O1O1O1O1O10010O0000010O00001O01O01O00010O00001O00001O000O3N1O2N2N1N3N2N1O2N1N3N1O2N1O2M3N1O2N1O2M2O1O00000O100000O1O00100O1N2N2N1N3N001O000O20O01O010O010O00010OO2N1N2O2M2O2M2O2M2O2NWHkNW65XHh0_1YOU6O]Hf0]1@S6IbHf0Y1FR6DfHd0W1Mo5@kHb0U12m5\\OoHa0R18l5VOUI?n0a0i5POZI?k0e0h5lN^Ia0g0e0h5kNeI`0?g0j5hNkIa08h0n5eNmIc03i0P6bNPJf0Li0R7WOkHk0W7UOeHl0]7TOaHm0`7TO\\Hm0f7SOWHn0k7RORHP1P8i02O1N2O2M2N2OO001O0010O01O00010O0001N100O2O0O2O0O101N100O2O1N2O1N2O1N2O1N2O1N101N2O1N2O1L4M4K6J6J7I6Jd[X1"}, "label": "the bottom half of a man in red shorts on a bike"}]} {"id": 562176, "image": "COCO_train2014_000000562176.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver bike being riden by a man in rust colored shorts\"."}], "task": "refering", "answer_template": "The \"a silver bike being riden by a man in rust colored shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 172, 173, 174, 176, 177, 194, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 220, 221, 222, 223, 224, 225, 240, 241, 242, 243, 244, 246, 247, 248], "bbox": [0.45628124999999997, 0.5197067448680351, 0.810125, 0.9154252199413488], "iscrowd": 0, "area": 15699.183350000003, "rle": {"size": [341, 640], "counts": "n_Q3:n9?M3L4M2M4M3M3L2O2M3N2O001N2O1VOYNcHh1[7[NcHg1Y7\\NfHf1W7]NhHd1U7^NjHd1R7`NmHBAj1b7dNlHb1S7_NlHb1S7^NlHd1S7]NlHc1T7]NkHe1T7\\NkHe1T7\\NhHh1W7ZNeHg1\\7i0000000000000000O10000000000000000O101O0000000000000O2O0000000O1O1O2N1O1O1O1O1O10010O0000010O00001O01O01O00010O00001O00001O000O3N1O2N2N1N3N2N1O2N1N3N1O2N1O2M3N1O2N1O2M2O1O00000O100000O1O00100O1N2N2N1N3N001O000O20O01O010O010O00010OO2N1N2O2M2O2M2O2M2O2NWHkNW65XHh0_1YOU6O]Hf0]1@S6IbHf0Y1FR6DfHd0W1Mo5@kHb0U12m5\\OoHa0R18l5VOUI?n0a0i5POZI?k0e0h5lN^Ia0g0e0h5kNeI`0?g0j5hNkIa08h0n5eNmIc03i0P6bNPJf0Li0R7WOkHk0W7UOeHl0]7TOaHm0`7TO\\Hm0f7SOWHn0k7RORHP1P8i02O1N2O2M2N2OO001O0010O01O00010O0001N100O2O0O2O0O101N100O2O1N2O1N2O1N2O1N2O1N101N2O1N2O1L4M4K6J6J7I6Jd[X1"}, "label": "a silver bike being riden by a man in rust colored shorts"}]} {"id": 103430, "image": "COCO_train2014_000000103430.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the rightmost hotdog\"."}], "task": "refering", "answer_template": "The \"the rightmost hotdog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 106, 107, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.58984, 0.33984, 1.0, 0.9882933333333334], "iscrowd": 0, "area": 38271.3316, "rle": {"size": [375, 500], "counts": "PZ\\3b0X:Q1B?H8H8H8H8I7H8H7K6I7J6J6I3N1O1N2O1N2O1O1N2O1O1N200O1O100O1O100O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O100O100O10000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O100OjIeJR6`51O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000O10000O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O100O1O1OhNX1`M`2`MaN"}, "label": "the rightmost hotdog"}]} {"id": 103430, "image": "COCO_train2014_000000103430.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a roll in the bottom right corner\"."}], "task": "refering", "answer_template": "The \"a roll in the bottom right corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [89, 106, 107, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 159, 160, 161, 172, 173, 174, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 215, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.58984, 0.33984, 1.0, 0.9882933333333334], "iscrowd": 0, "area": 38271.3316, "rle": {"size": [375, 500], "counts": "PZ\\3b0X:Q1B?H8H8H8H8I7H8H7K6I7J6J6I3N1O1N2O1N2O1O1N2O1O1N200O1O100O1O100O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O100O100O10000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O100OjIeJR6`51O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000O10000O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O100O1O1OhNX1`M`2`MaN"}, "label": "a roll in the bottom right corner"}]} {"id": 103430, "image": "COCO_train2014_000000103430.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the only bagel dog that is completely visible\"."}], "task": "refering", "answer_template": "The \"the only bagel dog that is completely visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 18, 19, 20, 21, 22, 23, 24, 25, 36, 37, 38, 39, 40, 41, 42, 43, 54, 55, 56, 57, 58, 59, 60, 61, 62, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 127, 128, 129, 130, 131, 132, 133, 134, 135, 145, 146, 147, 148, 149, 150, 151, 152, 153, 164, 165, 166, 167, 168, 169, 170, 171, 182, 183, 184, 185, 186, 187, 188, 200, 201, 202, 203, 204, 205, 220, 221], "bbox": [0.0053, 0.009573333333333333, 0.55058, 0.9366399999999999], "iscrowd": 0, "area": 69991.69155, "rle": {"size": [375, 500], "counts": "oU1Y1X9[1I7I7I6J7I7I7I6J7I7I7I6J5K3M3M4L3M4M2M3M4L3M4L3M3O2M2O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1oL_Hj1c7PNiHh1X7RNUIf1m6TN_Id1b6WNjIa1W6ZNUJ^1T8N1O1O1O001O1O1O1O1O1O1O001O0000001O0000000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O0000000000001O00000000001O0000000000001O000001O1O010O1O001O1O1N101N2O0O2O1N101N2O0O2O1N101N2O0O2O1N2O1N2O1N2O1N2O1O2M2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N3N1N3N2M3N2M3N2N3L3N2M3N2M3N2Ma0@0O2O0O2N1N3N1O2N2M2O2N1N3N1O2N2M2O2N1O2M2O2N1N3N2N1O2M2O2N1O2M2O2N2M2O2N1O2M2O2N2N1N4M3M3L4M2N3M3L4M2M4K5J6K5K4L5J6K5K5Kj0UOm0TOaVb2"}, "label": "the only bagel dog that is completely visible"}]} {"id": 103430, "image": "COCO_train2014_000000103430.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pig in a blanket to the left of all the other ones\"."}], "task": "refering", "answer_template": "The \"a pig in a blanket to the left of all the other ones\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 18, 19, 20, 21, 22, 23, 24, 25, 36, 37, 38, 39, 40, 41, 42, 43, 54, 55, 56, 57, 58, 59, 60, 61, 62, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 127, 128, 129, 130, 131, 132, 133, 134, 135, 145, 146, 147, 148, 149, 150, 151, 152, 153, 164, 165, 166, 167, 168, 169, 170, 171, 182, 183, 184, 185, 186, 187, 188, 200, 201, 202, 203, 204, 205, 220, 221], "bbox": [0.0053, 0.009573333333333333, 0.55058, 0.9366399999999999], "iscrowd": 0, "area": 69991.69155, "rle": {"size": [375, 500], "counts": "oU1Y1X9[1I7I7I6J7I7I7I6J7I7I7I6J5K3M3M4L3M4M2M3M4L3M4L3M3O2M2O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1oL_Hj1c7PNiHh1X7RNUIf1m6TN_Id1b6WNjIa1W6ZNUJ^1T8N1O1O1O001O1O1O1O1O1O1O001O0000001O0000000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O0000000000001O00000000001O0000000000001O000001O1O010O1O001O1O1N101N2O0O2O1N101N2O0O2O1N101N2O0O2O1N2O1N2O1N2O1N2O1O2M2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N3N1N3N2M3N2M3N2N3L3N2M3N2M3N2Ma0@0O2O0O2N1N3N1O2N2M2O2N1N3N1O2N2M2O2N1O2M2O2N1N3N2N1O2M2O2N1O2M2O2N2M2O2N1O2M2O2N2N1N4M3M3L4M2N3M3L4M2M4K5J6K5K4L5J6K5K5Kj0UOm0TOaVb2"}, "label": "a pig in a blanket to the left of all the other ones"}]} {"id": 103430, "image": "COCO_train2014_000000103430.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top right food\"."}], "task": "refering", "answer_template": "The \"the top right food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 17, 27, 28, 29, 30, 31, 32, 33, 34, 35, 45, 46, 47, 48, 49, 50, 51, 52, 53, 63, 64, 65, 66, 67, 68, 69, 70, 71, 82, 83, 84, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106], "bbox": [0.50988, 0.0009333333333333333, 0.99622, 0.46298666666666666], "iscrowd": 0, "area": 32209.582899999994, "rle": {"size": [375, 500], "counts": "[^m2`0o: in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 25, 26, 27, 28, 29, 30, 31, 48, 49, 50, 51, 52, 53, 54, 71, 72, 73, 74, 75, 76, 77, 94, 95, 96, 97, 98, 99, 118, 119, 120, 121, 143], "bbox": [0.08545312499999999, 0.004637002341920374, 0.368828125, 0.43611241217798596], "iscrowd": 0, "area": 23976.3409, "rle": {"size": [427, 640], "counts": "nof0:i<;E;F:E;E in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 25, 26, 27, 28, 29, 30, 31, 48, 49, 50, 51, 52, 53, 54, 71, 72, 73, 74, 75, 76, 77, 94, 95, 96, 97, 98, 99, 118, 119, 120, 121, 143], "bbox": [0.08545312499999999, 0.004637002341920374, 0.368828125, 0.43611241217798596], "iscrowd": 0, "area": 23976.3409, "rle": {"size": [427, 640], "counts": "nof0:i<;E;F:E;E in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 36, 37, 38, 39, 40, 41, 60, 61, 62, 63, 84, 85, 107, 108, 130, 131, 153, 154, 175, 176, 177, 178, 198, 199, 200, 201, 222, 223], "bbox": [0.566734375, 0.0038173302107728335, 0.818609375, 0.6218032786885246], "iscrowd": 0, "area": 16714.410150000007, "rle": {"size": [427, 640], "counts": "k[g43X=5K5K4L5K5K1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1bIhMg1Z2WNiMf1X2XNlMe1U2ZNnMc1S2\\NPNa1Q2]NSN`1m1`NVN]1k1aNYN\\1g1dN\\NY1e1fN^NW1b1hNbNU1_1\\NPOc1P1YNTOg1m0UNVOk1j0QNZOo1g0mM\\OS2d0iM@W2a0eMCZ2=bMG^2:^MIb27ZMMf24VMOj21RM3n2OmL4S3MhL7X3JdL9\\3H_L^1D_N>`1D]N>b1D\\N=c1D[N>e1CXN?g1CWN>h1DUN>j1DSN>l1a500O1O1O1O100O1O1O1O100O1O1O1M3M3M3M3M3M3M3M3M3M3M3N7Hh[`1"}, "label": "glass of wine in table"}]} {"id": 431112, "image": "COCO_train2014_000000431112.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass of wine\"."}], "task": "refering", "answer_template": "The \"a glass of wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 36, 37, 38, 39, 40, 41, 60, 61, 62, 63, 84, 85, 107, 108, 130, 131, 153, 154, 175, 176, 177, 178, 198, 199, 200, 201, 222, 223], "bbox": [0.566734375, 0.0038173302107728335, 0.818609375, 0.6218032786885246], "iscrowd": 0, "area": 16714.410150000007, "rle": {"size": [427, 640], "counts": "k[g43X=5K5K4L5K5K1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1bIhMg1Z2WNiMf1X2XNlMe1U2ZNnMc1S2\\NPNa1Q2]NSN`1m1`NVN]1k1aNYN\\1g1dN\\NY1e1fN^NW1b1hNbNU1_1\\NPOc1P1YNTOg1m0UNVOk1j0QNZOo1g0mM\\OS2d0iM@W2a0eMCZ2=bMG^2:^MIb27ZMMf24VMOj21RM3n2OmL4S3MhL7X3JdL9\\3H_L^1D_N>`1D]N>b1D\\N=c1D[N>e1CXN?g1CWN>h1DUN>j1DSN>l1a500O1O1O1O100O1O1O1O100O1O1O1M3M3M3M3M3M3M3M3M3M3M3N7Hh[`1"}, "label": "a glass of wine"}]} {"id": 54282, "image": "COCO_train2014_000000054282.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a red shirt\"."}], "task": "refering", "answer_template": "The \"a man in a red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 143, 144, 145, 166, 167, 168, 169, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 240, 253, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 307, 324, 325, 327, 328, 329, 330], "bbox": [0.0, 0.35056470588235294, 0.44917187500000005, 1.0], "iscrowd": 0, "area": 33665.037000000004, "rle": {"size": [425, 640], "counts": "]:Y1P<000000000000000O100000000000000000000000O100002O1O1O1N2O1O1O1O1N2O00O1O2N1O1O2O0O1O2N000001O0000000O101O000000001O000O101O0000001O000O10001O0000001O0O101O001O001O001N101N101N101N101N100O10O0100O001I7I6J7I7K400O2O0O100O100O100O101O0O100O100O1O1O2O0nF`L[8a3cG_L^8b3`G^La8c3]G]Ld8d3ZG\\Lg8e3WG[Lj8e3UG\\Ll8b3TG^Lm8a3SG_Ln8a3QG_LP9`3PG`LQ9_3oFbLR9]3mFcLU9[3kFeLY1O[6Z3\\HgLY12Z6U3]HiLZ14X6R3]HkLZ16X6m2^HmLZ19W6h2_HoL[1;U6d2`HQM[1>T6`2`HRM]1`0R6\\2aHTM]1c0Q6W2bHWM]1d0o5T2dHXM]1g0n5o1eHZM^1i0l5k1fH\\M^1l0k5f1gH^M_1n0`4]OdJU2]OaM^1P1V4HkJf1@bM`1R1R4KkJ_1CdM`1U1P4MiJX1GfM`1X1m30hJP1KhMa1Z1i33hJi0NkM`1[1h37fJa02mMa1]1d3b2kJQLa1`1b3^2mJRLb1b1^3\\2PKRLb1e1[3Y2SKRLc1g1X3V2UKTLb1i1X3R2UKULd1k1W3m1UKXLd1n1V3h1VKZLe1P2U3c1VK]Le1S2U3]1VKaLd1U2U3X1WKcLe1W2T3S1WKfLe1Z2T3m0WKiLf1\\2R3i0XKkLf1_2R3c0XKoLf1`2R3>XKRMf1c2Q3:XKSMh1e2P35XKVMh1h2P3OXKYMi1j2n2KYK[Mi1m2n2EYK_Mi1n2m2AZKaMi1Q3m2[OZKdMi1T3m2UOZKgMj1V3k2QO[KiMj1Y3k2kN[KmMj1Z3k2fN[KPNj1]3j2bN[KQNl1_3i2]N[KTNl1b3i2WN[KWNm1d3g2SN\\KZNl1f3h2mM\\K]Nm1h3g2hM\\K`Nm1k3g2bM\\KcNn1m3f2]M\\KfNn1P4f2WM\\KiNn1S4g2PM[KnNn1T4g2kL[KQOn1W4g2fLZKSOP2X4g2bLYKVOP2[4g2\\LYKYOQ2]4f2WLYK]OP2_4g2QLYK@Q2a4f2lKYKCQ2d4f2fKYKFR2f4e2aKYKIR2i4f2ZKXKMS2k4e2UKXK1R2m4f2oJXK4S2o4e2[K[Mh4e2UK[Mn4e2oJ\\MS5e2jJZMY5f2dJ[M^5e2_J[Md5f2XJ[Mj5e2SJ[MP6f2mIZMU6j44M2M101N1O101N101N100O2O03N;D9G3N2M3M3N2M3N2M3M3N2M3N2M3N2M3M3N2M3N2M3N2M3N1N3M3N2M3N2M2O2M3M3N1N3N2M3N2M2O2M3M3L3N3L4M3L3M4M3L4L4M2M4M3L9@`0@WXb4"}, "label": "a man in a red shirt"}]} {"id": 54282, "image": "COCO_train2014_000000054282.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in red t - shirt\"."}], "task": "refering", "answer_template": "The \"man in red t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 143, 144, 145, 166, 167, 168, 169, 189, 190, 191, 192, 193, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 240, 253, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 307, 324, 325, 327, 328, 329, 330], "bbox": [0.0, 0.35056470588235294, 0.44917187500000005, 1.0], "iscrowd": 0, "area": 33665.037000000004, "rle": {"size": [425, 640], "counts": "]:Y1P<000000000000000O100000000000000000000000O100002O1O1O1N2O1O1O1O1N2O00O1O2N1O1O2O0O1O2N000001O0000000O101O000000001O000O101O0000001O000O10001O0000001O0O101O001O001O001N101N101N101N101N100O10O0100O001I7I6J7I7K400O2O0O100O100O100O101O0O100O100O1O1O2O0nF`L[8a3cG_L^8b3`G^La8c3]G]Ld8d3ZG\\Lg8e3WG[Lj8e3UG\\Ll8b3TG^Lm8a3SG_Ln8a3QG_LP9`3PG`LQ9_3oFbLR9]3mFcLU9[3kFeLY1O[6Z3\\HgLY12Z6U3]HiLZ14X6R3]HkLZ16X6m2^HmLZ19W6h2_HoL[1;U6d2`HQM[1>T6`2`HRM]1`0R6\\2aHTM]1c0Q6W2bHWM]1d0o5T2dHXM]1g0n5o1eHZM^1i0l5k1fH\\M^1l0k5f1gH^M_1n0`4]OdJU2]OaM^1P1V4HkJf1@bM`1R1R4KkJ_1CdM`1U1P4MiJX1GfM`1X1m30hJP1KhMa1Z1i33hJi0NkM`1[1h37fJa02mMa1]1d3b2kJQLa1`1b3^2mJRLb1b1^3\\2PKRLb1e1[3Y2SKRLc1g1X3V2UKTLb1i1X3R2UKULd1k1W3m1UKXLd1n1V3h1VKZLe1P2U3c1VK]Le1S2U3]1VKaLd1U2U3X1WKcLe1W2T3S1WKfLe1Z2T3m0WKiLf1\\2R3i0XKkLf1_2R3c0XKoLf1`2R3>XKRMf1c2Q3:XKSMh1e2P35XKVMh1h2P3OXKYMi1j2n2KYK[Mi1m2n2EYK_Mi1n2m2AZKaMi1Q3m2[OZKdMi1T3m2UOZKgMj1V3k2QO[KiMj1Y3k2kN[KmMj1Z3k2fN[KPNj1]3j2bN[KQNl1_3i2]N[KTNl1b3i2WN[KWNm1d3g2SN\\KZNl1f3h2mM\\K]Nm1h3g2hM\\K`Nm1k3g2bM\\KcNn1m3f2]M\\KfNn1P4f2WM\\KiNn1S4g2PM[KnNn1T4g2kL[KQOn1W4g2fLZKSOP2X4g2bLYKVOP2[4g2\\LYKYOQ2]4f2WLYK]OP2_4g2QLYK@Q2a4f2lKYKCQ2d4f2fKYKFR2f4e2aKYKIR2i4f2ZKXKMS2k4e2UKXK1R2m4f2oJXK4S2o4e2[K[Mh4e2UK[Mn4e2oJ\\MS5e2jJZMY5f2dJ[M^5e2_J[Md5f2XJ[Mj5e2SJ[MP6f2mIZMU6j44M2M101N1O101N101N100O2O03N;D9G3N2M3M3N2M3N2M3M3N2M3N2M3N2M3M3N2M3N2M3N2M3N1N3M3N2M3N2M2O2M3M3N1N3N2M3N2M2O2M3M3L3N3L4M3L3M4M3L4L4M2M4M3L9@`0@WXb4"}, "label": "man in red t - shirt"}]} {"id": 54282, "image": "COCO_train2014_000000054282.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skis sticking out behind the orange jacket\"."}], "task": "refering", "answer_template": "The \"the skis sticking out behind the orange jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 156, 178, 179, 200, 201, 222, 223, 245, 246, 290], "bbox": [0.602484375, 0.33157647058823525, 0.8372499999999998, 0.8696235294117647], "iscrowd": 0, "area": 3365.300000000001, "rle": {"size": [425, 640], "counts": "RbP51V=2O1N2O1N3N0O100O010O10O10O10O013M2M4Lfm40ZRK4M3M3L3N3M2NObC@U in this image.", "objects": [{"patches": [133, 134, 156, 178, 179, 200, 201, 222, 223, 245, 246, 290], "bbox": [0.602484375, 0.33157647058823525, 0.8372499999999998, 0.8696235294117647], "iscrowd": 0, "area": 3365.300000000001, "rle": {"size": [425, 640], "counts": "RbP51V=2O1N2O1N3N0O100O010O10O10O10O013M2M4Lfm40ZRK4M3M3L3N3M2NObC@U in this image.", "objects": [{"patches": [131, 154, 155, 177, 178, 199, 200, 201, 221, 222, 223, 224, 245, 246, 247, 268, 269, 270], "bbox": [0.6359375, 0.39237647058823527, 0.782375, 0.7973176470588236], "iscrowd": 0, "area": 7217.138599999997, "rle": {"size": [425, 640], "counts": "dUY51X=3M3M3M2N3N2N2OM2O2N1O2N1O0000O2O1N1O000000OUD5Z:LfE7W:IiE9T:HkE;S:E\\EC2k0^:BaEC1m0[:AdECOP1Z:]OgECOR1W:\\OiECOU1T:XOnEDMT1U:XOnEELT1T:XOoEELT1U:WOoEFKT1T:WOQFFJS1T:WOSFGGT1U:UOSFHHT1S:UOUFGHU1Q:UOWFFGW1Q:SOXFEHY1n9ROZFFHY1m9QO[FFGZ1e9`NZFa0:DH\\1b9YOfF[OH]1b9VOfF^OG^1c9SOfF^OH`1d8[NoGg0E^OHa1X8XNiG;<=K_OGc1U8_NgG6?8M@Id1Q8UOWHE0BH_1K]NT8Q1VHA3BGa1MZNR8Z1PHXO:BHc1MWNQ8V2YHQNIc1NUNo7W2ZHQNHe1S88VHRNHg1R86VHSNHh1Q85VHTNHi1R82VHTNIk1Q8OWHVNHl1Q8MWHVNHo1Q8JVHXNIo1Q8HVHYNIP2Q8FVHYNIS2Q8BWH[NHT2Q8@VH]NIS2R8_OUH]NIV2S8[OTH_NIV2T8YOTHaNHV2V8WOQHcNIX2V8TOQHdNIX2X8ROoGfNHZ2Y8oNoGfNI[2Z8lNmGjNIY2]8kNjGlNHZ2`8hNhGmNIZ2a8hNfGnNIZ2c8fNcGQOIX2h8dN`GSOIV2l8eN[GUOIT2P9eNWGWOHT2T9cNSGYOJS2V9bNQGZOIT2[9]NnF]OFT2e9XNnFd1_9PNdFk1^:12N3M3L4M2F?@?\\Oo^i1"}, "label": "a young man wearing black smiling at the camera"}]} {"id": 54282, "image": "COCO_train2014_000000054282.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in gray clothes\"."}], "task": "refering", "answer_template": "The \"a person in gray clothes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 154, 155, 177, 178, 199, 200, 201, 221, 222, 223, 224, 245, 246, 247, 268, 269, 270], "bbox": [0.6359375, 0.39237647058823527, 0.782375, 0.7973176470588236], "iscrowd": 0, "area": 7217.138599999997, "rle": {"size": [425, 640], "counts": "dUY51X=3M3M3M2N3N2N2OM2O2N1O2N1O0000O2O1N1O000000OUD5Z:LfE7W:IiE9T:HkE;S:E\\EC2k0^:BaEC1m0[:AdECOP1Z:]OgECOR1W:\\OiECOU1T:XOnEDMT1U:XOnEELT1T:XOoEELT1U:WOoEFKT1T:WOQFFJS1T:WOSFGGT1U:UOSFHHT1S:UOUFGHU1Q:UOWFFGW1Q:SOXFEHY1n9ROZFFHY1m9QO[FFGZ1e9`NZFa0:DH\\1b9YOfF[OH]1b9VOfF^OG^1c9SOfF^OH`1d8[NoGg0E^OHa1X8XNiG;<=K_OGc1U8_NgG6?8M@Id1Q8UOWHE0BH_1K]NT8Q1VHA3BGa1MZNR8Z1PHXO:BHc1MWNQ8V2YHQNIc1NUNo7W2ZHQNHe1S88VHRNHg1R86VHSNHh1Q85VHTNHi1R82VHTNIk1Q8OWHVNHl1Q8MWHVNHo1Q8JVHXNIo1Q8HVHYNIP2Q8FVHYNIS2Q8BWH[NHT2Q8@VH]NIS2R8_OUH]NIV2S8[OTH_NIV2T8YOTHaNHV2V8WOQHcNIX2V8TOQHdNIX2X8ROoGfNHZ2Y8oNoGfNI[2Z8lNmGjNIY2]8kNjGlNHZ2`8hNhGmNIZ2a8hNfGnNIZ2c8fNcGQOIX2h8dN`GSOIV2l8eN[GUOIT2P9eNWGWOHT2T9cNSGYOJS2V9bNQGZOIT2[9]NnF]OFT2e9XNnFd1_9PNdFk1^:12N3M3L4M2F?@?\\Oo^i1"}, "label": "a person in gray clothes"}]} {"id": 54282, "image": "COCO_train2014_000000054282.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"one ski\"."}], "task": "refering", "answer_template": "The \"one ski\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 54, 55, 77, 78, 99, 100, 101, 122, 123, 145, 146, 167, 168, 169, 190, 191, 213, 214, 235, 236, 258, 259, 280, 281, 282, 303, 304, 326, 327], "bbox": [0.1733125, 0.12169411764705883, 0.41353125, 1.0], "iscrowd": 0, "area": 10891.029050000001, "rle": {"size": [425, 640], "counts": "^_^12S=4L4L4L4L4L4L4L4N2N2N2N2O1N2N2O1N2N2N2O1N2N2N2O1N1O00O2O000O101O0O100O2O00O001O1O010O001O1O010O1O001O010O1O001O1O010O001O1O010O001O1O010O1O001O010O1O001O1O010O001O1O001O1O001N101O1O001O1O001N101O1O001O1O0O2O001O00001O001N11O01O010O00010K4K6K4K5L5J5N31N3M2O2M3N3L4L4M3L4M2M4L4M3L4M3L3N3L4L4M3L4M4K;E;F:E:G:EVak4"}, "label": "one ski"}]} {"id": 54282, "image": "COCO_train2014_000000054282.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black ski\"."}], "task": "refering", "answer_template": "The \"a black ski\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 54, 55, 77, 78, 99, 100, 101, 122, 123, 145, 146, 167, 168, 169, 190, 191, 213, 214, 235, 236, 258, 259, 280, 281, 282, 303, 304, 326, 327], "bbox": [0.1733125, 0.12169411764705883, 0.41353125, 1.0], "iscrowd": 0, "area": 10891.029050000001, "rle": {"size": [425, 640], "counts": "^_^12S=4L4L4L4L4L4L4L4N2N2N2N2O1N2N2O1N2N2N2O1N2N2N2O1N1O00O2O000O101O0O100O2O00O001O1O010O001O1O010O1O001O010O1O001O1O010O001O1O010O001O1O010O1O001O010O1O001O1O010O001O1O001O1O001N101O1O001O1O001N101O1O001O1O0O2O001O00001O001N11O01O010O00010K4K6K4K5L5J5N31N3M2O2M3N3L4L4M3L4M2M4L4M3L4M3L3N3L4L4M3L4M4K;E;F:E:G:EVak4"}, "label": "a black ski"}]} {"id": 300047, "image": "COCO_train2014_000000300047.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large elephant\"."}], "task": "refering", "answer_template": "The \"a large elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 170, 171, 172, 173, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 288], "bbox": [0.411984375, 0.45969555035128806, 0.5828593750000001, 0.816112412177986], "iscrowd": 0, "area": 11645.307149999995, "rle": {"size": [427, 640], "counts": "QY^3R1Y<2N2N2M3N2N2M3N3M2gD]Nf:e1SE`Nl:c1PE_NP;m110YE_Od8a0WGDh8=RGHo8Z200O01000001N1O1H8M31O1O1O101N1O1O1OfNmF\\NS9e1PGXNo8h1UGTNl8l1WGPNi8Q2ZGkMf8V2^GeMb8\\2aG`M_8b2cGYM^8h2eGTM[8n2gGnLY8T3n03M4L4L4L2N1O1O2N1O1O2N1O2N1O2N1O2N1O2N1O1O01J5L5I6K6I6K60O01O0010O01ORGhLi7X3THkLm7U3oGnLQ8V3hGmLY8V3`GmL`8X3WGlLj8i32O1O1O1O1O1O1N2O1L4F:\\Od0E;I7I7J7H8H8I7H7I8I7H8I_S_3"}, "label": "a large elephant"}]} {"id": 300047, "image": "COCO_train2014_000000300047.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant and with its child elephant\"."}], "task": "refering", "answer_template": "The \"elephant and with its child elephant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 170, 171, 172, 173, 193, 194, 195, 196, 197, 216, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 288], "bbox": [0.411984375, 0.45969555035128806, 0.5828593750000001, 0.816112412177986], "iscrowd": 0, "area": 11645.307149999995, "rle": {"size": [427, 640], "counts": "QY^3R1Y<2N2N2M3N2N2M3N3M2gD]Nf:e1SE`Nl:c1PE_NP;m110YE_Od8a0WGDh8=RGHo8Z200O01000001N1O1H8M31O1O1O101N1O1O1OfNmF\\NS9e1PGXNo8h1UGTNl8l1WGPNi8Q2ZGkMf8V2^GeMb8\\2aG`M_8b2cGYM^8h2eGTM[8n2gGnLY8T3n03M4L4L4L2N1O1O2N1O1O2N1O2N1O2N1O2N1O2N1O1O01J5L5I6K6I6K60O01O0010O01ORGhLi7X3THkLm7U3oGnLQ8V3hGmLY8V3`GmL`8X3WGlLj8i32O1O1O1O1O1O1N2O1L4F:\\Od0E;I7I7J7H8H8I7H7I8I7H8I_S_3"}, "label": "elephant and with its child elephant"}]} {"id": 234516, "image": "COCO_train2014_000000234516.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pickup truck directly beneath the shadow of the sign\"."}], "task": "refering", "answer_template": "The \"a pickup truck directly beneath the shadow of the sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 115, 116, 128, 129, 130, 131, 144, 145, 146, 147, 160, 161], "bbox": [0.5119711538461538, 0.5997714285714285, 0.826298076923077, 0.8480857142857142], "iscrowd": 0, "area": 6412.62905, "rle": {"size": [350, 416], "counts": "lPY24c:8G8K6O001O00001O001O001O1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O00000000000010O00000000000001O000000000000001O001O001O1O001O12M4M3M3M2ON2N2N2N2N2N2N1O100O1O1O001O00000000000001O00000000000000000O100O2O0O101N4M3L5L2M3N1N2O1O1O2N1O1O1O1O2N1O1O1M3J7K4K5L4LW[h0"}, "label": "a pickup truck directly beneath the shadow of the sign"}]} {"id": 234516, "image": "COCO_train2014_000000234516.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the truck parked inthe middle\"."}], "task": "refering", "answer_template": "The \"the truck parked inthe middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 115, 116, 128, 129, 130, 131, 144, 145, 146, 147, 160, 161], "bbox": [0.5119711538461538, 0.5997714285714285, 0.826298076923077, 0.8480857142857142], "iscrowd": 0, "area": 6412.62905, "rle": {"size": [350, 416], "counts": "lPY24c:8G8K6O001O00001O001O001O1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O2N1O1O00000000000010O00000000000001O000000000000001O001O001O1O001O12M4M3M3M2ON2N2N2N2N2N2N1O100O1O1O001O00000000000001O00000000000000000O100O2O0O101N4M3L5L2M3N1N2O1O1O2N1O1O1O1O2N1O1O1M3J7K4K5L4LW[h0"}, "label": "the truck parked inthe middle"}]} {"id": 234516, "image": "COCO_train2014_000000234516.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a truck right beneath the \" glass \" sign\"."}], "task": "refering", "answer_template": "The \"a truck right beneath the \" glass \" sign\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 131, 132, 133, 134, 147, 148, 149], "bbox": [0.6751682692307692, 0.5956857142857143, 1.0, 0.8225428571428571], "iscrowd": 0, "area": 5134.517350000003, "rle": {"size": [350, 416], "counts": "^XP31m:2N1O000000001O00000000000000001O00000O1O1N2N2O1N2N3M2N2O11O0000000001O0000101N3M2N3M1O1O2N101N1O1O2N1O1O2N1O2O0O1O2N1O3M2N3M01O01O00000000000001O000000000000001O010O1O001O1O001O3M4L4L2N000000001O01O0000100O2N1O10O01O001O01O0001O0001O01O000001O01OO101N1O1O1O101N1O1I7ZOUI"}, "label": "a truck right beneath the \" glass \" sign"}]} {"id": 234516, "image": "COCO_train2014_000000234516.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pick up truck in front of door with word glass on it\"."}], "task": "refering", "answer_template": "The \"pick up truck in front of door with word glass on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 131, 132, 133, 134, 147, 148, 149], "bbox": [0.6751682692307692, 0.5956857142857143, 1.0, 0.8225428571428571], "iscrowd": 0, "area": 5134.517350000003, "rle": {"size": [350, 416], "counts": "^XP31m:2N1O000000001O00000000000000001O00000O1O1N2N2O1N2N3M2N2O11O0000000001O0000101N3M2N3M1O1O2N101N1O1O2N1O1O2N1O2O0O1O2N1O3M2N3M01O01O00000000000001O000000000000001O010O1O001O1O001O3M4L4L2N000000001O01O0000100O2N1O10O01O001O01O0001O0001O01O000001O01OO101N1O1O1O101N1O1I7ZOUI"}, "label": "pick up truck in front of door with word glass on it"}]} {"id": 234516, "image": "COCO_train2014_000000234516.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old truck in front of other trucks and cars\"."}], "task": "refering", "answer_template": "The \"an old truck in front of other trucks and cars\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 110, 111, 112, 120, 121, 122, 123, 124, 125, 126, 127, 128, 135, 136, 137, 138, 139, 140, 141, 142, 143, 156, 157, 158], "bbox": [0.0, 0.5842571428571429, 0.5880048076923077, 0.8606857142857143], "iscrowd": 0, "area": 15298.885300000004, "rle": {"size": [350, 416], "counts": "a86^::F:E;F;E:O10001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O0000001OO1O1O1O100O1O1O1O1O1O100O1O1O10O10O10000O10000O10000O10000O1000000O100000000000000000000000000000000000N2K5J600000000000O10O100000000000000000000000000000001O00000000000000001N10000000000000001O000000000000001O000000000000000O2O00011N10001O0O101O001N10001O0O2OO1O00100O1O010O1O1O010O1O10O01O1O001O1O1O001O0O101O00001O00001O00001N2O1O1O1O1O1O1O1O1O1O1N2O2N1O1O1O1O1O1O1O1O1N2O1K5K5K5J6K5J6KTVj1"}, "label": "an old truck in front of other trucks and cars"}]} {"id": 234516, "image": "COCO_train2014_000000234516.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the truck furthest to the left\"."}], "task": "refering", "answer_template": "The \"the truck furthest to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 108, 109, 110, 111, 112, 120, 121, 122, 123, 124, 125, 126, 127, 128, 135, 136, 137, 138, 139, 140, 141, 142, 143, 156, 157, 158], "bbox": [0.0, 0.5842571428571429, 0.5880048076923077, 0.8606857142857143], "iscrowd": 0, "area": 15298.885300000004, "rle": {"size": [350, 416], "counts": "a86^::F:E;F;E:O10001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O0000001OO1O1O1O100O1O1O1O1O1O100O1O1O10O10O10000O10000O10000O10000O1000000O100000000000000000000000000000000000N2K5J600000000000O10O100000000000000000000000000000001O00000000000000001N10000000000000001O000000000000001O000000000000000O2O00011N10001O0O101O001N10001O0O2OO1O00100O1O010O1O1O010O1O10O01O1O001O1O1O001O0O101O00001O00001O00001N2O1O1O1O1O1O1O1O1O1O1N2O2N1O1O1O1O1O1O1O1O1N2O1K5K5K5J6K5J6KTVj1"}, "label": "the truck furthest to the left"}]} {"id": 439325, "image": "COCO_train2014_000000439325.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with glasses in a white shirt and black pants sitting on a wooden bench reading some papers\"."}], "task": "refering", "answer_template": "The \"a man with glasses in a white shirt and black pants sitting on a wooden bench reading some papers\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 99, 100, 101, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 234, 235, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 261, 262, 263, 264, 282, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332, 333, 351, 352, 353, 354, 355, 356, 357, 374, 375, 376, 379, 380], "bbox": [0.186296875, 0.214, 0.57665625, 1.0], "iscrowd": 0, "area": 45850.9611, "rle": {"size": [480, 640], "counts": "gQh12k>;F9G7I4L3L4M4L3M3M4K4L4L5K4L4L5K4L4L4L3N3M2N3M2M4M2N2N3L3N3M2N3M2M3N3M2N3M2N3M2N2M4M2O2N101N101O1N2O1N2O2M2O1N2O1N2O`0@k0TO7J5J6K5J6K5J6K5RJYHX4n7`KVH^4o7[KSHe4S8SKoGl4S900O10000O100O10000O10000O10000ZOQFZLP:d3UFYLk9d3[FXLf9f3_FWLa9f3eFWL[9g3jFULW9i3Q1M3N2M3N200O1O10ZOdLhE[3X:hLfEX3Z:jLdEU3\\:nLbER3^:PM`Eo2`:TM^El2b:VM\\Ei2d:ZMZEf2f:\\MXEc2h:`MVE`2W8nLSJl3h5^LSJc3h5gLSJY3g5QMTJP3g5ZMTJf2f5dMUJ]2j5hMQJY2P6kMkIU2V6nMeIR2]6RN]Io1d6TNXIl1f6[NTIf1j6`NQIa1m6fNnHZ1P7lNkHU1S7]3N2N2001O000000000010O1O1O001O1O1O1O1O0011OPKPHT3m7gLVH]3h7^L[Hg3b7TLaHQ4W2bJ`2Y1[KZ4j1bJo2n0[Kf4R1hJi3:XK\\5<^Jb4NUKQ6HUJW5CVK]6WOTJe5[OYKa7i4[H\\Kb7e4[H_Kc7R6N2M8H>Bm0SO7J0O00001O001O01O01O00001O010O01O010O00010O0010O010O0jNgFeLY9[3hFdLY9[3gFfLX9Z3iFeLW9[3jFdLW9[3jFeLU9Z3mFeLS94mFj11QNS93TGe1JXNQ91^G`1B^NQ9OfG\\1ZOdNo80oGU1SOjNn8OXHP1kNQOl8MbHk0cNWOk8OiHa0^N_Oi80PI9YNGf81WI1TNMe82^IIoM4b85eI_OjM in this image.", "objects": [{"patches": [77, 78, 99, 100, 101, 121, 122, 123, 124, 143, 144, 145, 146, 147, 166, 167, 168, 169, 170, 171, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 234, 235, 236, 237, 238, 239, 240, 241, 242, 257, 258, 259, 260, 261, 262, 263, 264, 282, 283, 284, 285, 286, 287, 305, 306, 307, 308, 309, 310, 328, 329, 330, 331, 332, 333, 351, 352, 353, 354, 355, 356, 357, 374, 375, 376, 379, 380], "bbox": [0.186296875, 0.214, 0.57665625, 1.0], "iscrowd": 0, "area": 45850.9611, "rle": {"size": [480, 640], "counts": "gQh12k>;F9G7I4L3L4M4L3M3M4K4L4L5K4L4L5K4L4L4L3N3M2N3M2M4M2N2N3L3N3M2N3M2M3N3M2N3M2N3M2N2M4M2O2N101N101O1N2O1N2O2M2O1N2O1N2O`0@k0TO7J5J6K5J6K5J6K5RJYHX4n7`KVH^4o7[KSHe4S8SKoGl4S900O10000O100O10000O10000O10000ZOQFZLP:d3UFYLk9d3[FXLf9f3_FWLa9f3eFWL[9g3jFULW9i3Q1M3N2M3N200O1O10ZOdLhE[3X:hLfEX3Z:jLdEU3\\:nLbER3^:PM`Eo2`:TM^El2b:VM\\Ei2d:ZMZEf2f:\\MXEc2h:`MVE`2W8nLSJl3h5^LSJc3h5gLSJY3g5QMTJP3g5ZMTJf2f5dMUJ]2j5hMQJY2P6kMkIU2V6nMeIR2]6RN]Io1d6TNXIl1f6[NTIf1j6`NQIa1m6fNnHZ1P7lNkHU1S7]3N2N2001O000000000010O1O1O001O1O1O1O1O0011OPKPHT3m7gLVH]3h7^L[Hg3b7TLaHQ4W2bJ`2Y1[KZ4j1bJo2n0[Kf4R1hJi3:XK\\5<^Jb4NUKQ6HUJW5CVK]6WOTJe5[OYKa7i4[H\\Kb7e4[H_Kc7R6N2M8H>Bm0SO7J0O00001O001O01O01O00001O010O01O010O00010O0010O010O0jNgFeLY9[3hFdLY9[3gFfLX9Z3iFeLW9[3jFdLW9[3jFeLU9Z3mFeLS94mFj11QNS93TGe1JXNQ91^G`1B^NQ9OfG\\1ZOdNo80oGU1SOjNn8OXHP1kNQOl8MbHk0cNWOk8OiHa0^N_Oi80PI9YNGf81WI1TNMe82^IIoM4b85eI_OjM in this image.", "objects": [{"patches": [241, 242, 243, 264, 265, 276, 277, 281, 282, 287, 288, 299, 300, 301, 302, 303, 304, 305, 310, 311, 322, 323, 324, 325, 326, 345, 346, 347, 348, 368, 369, 370, 371], "bbox": [0.0, 0.6013541666666666, 0.581078125, 1.0], "iscrowd": 0, "area": 16723.2177, "rle": {"size": [480, 640], "counts": "g:o0P>o1RN\\1dN000O1000000O1000000O1000000O100000000O1000lKcEe3]:VLjEh3U:TLRFj3d:N1O2N1O1O001O1O001O002N2N2N2N1O00000000000000000000000000000000000000000000000000O1000000000000O1000000000000O1000000000000O1000000000000O100000000000000O100000_Oa0nNR1XOh00O010000O100O10O0100O10000O10O0100O10000O10O0100O10O10O10O0100O01000O010O100O01000O010O10O01000O0100O010O1000O0100O010O10001N3N1N2O1N2O2NbYn1D`fQN in this image.", "objects": [{"patches": [241, 242, 243, 264, 265, 276, 277, 281, 282, 287, 288, 299, 300, 301, 302, 303, 304, 305, 310, 311, 322, 323, 324, 325, 326, 345, 346, 347, 348, 368, 369, 370, 371], "bbox": [0.0, 0.6013541666666666, 0.581078125, 1.0], "iscrowd": 0, "area": 16723.2177, "rle": {"size": [480, 640], "counts": "g:o0P>o1RN\\1dN000O1000000O1000000O1000000O100000000O1000lKcEe3]:VLjEh3U:TLRFj3d:N1O2N1O1O001O1O001O002N2N2N2N1O00000000000000000000000000000000000000000000000000O1000000000000O1000000000000O1000000000000O1000000000000O100000000000000O100000_Oa0nNR1XOh00O010000O100O10O0100O10000O10O0100O10000O10O0100O10O10O10O0100O01000O010O100O01000O010O10O01000O0100O010O1000O0100O010O10001N3N1N2O1N2O2NbYn1D`fQN in this image.", "objects": [{"patches": [226, 227, 249, 250, 251, 272, 273, 274, 275, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.6929531250000001, 0.5270309278350516, 0.9981406249999999, 0.9903092783505155], "iscrowd": 0, "area": 25087.72589999999, "rle": {"size": [485, 640], "counts": "bXb6c0a>1O100O1O100O1O1O101N10000O1N2O1O1N2O1N2O1O1N2O1N2O1L4J6N2N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N200O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1N2O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1M3M3M3N2M3M3eMcKkI`4T6WLPIo3P7S201O0001O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001YNf1cM]7"}, "label": "front grill of a car behind a parking meter"}]} {"id": 545832, "image": "COCO_train2014_000000545832.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the grill of a dodge truck sits behind a parking meter\"."}], "task": "refering", "answer_template": "The \"the grill of a dodge truck sits behind a parking meter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 249, 250, 251, 272, 273, 274, 275, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 361, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.6929531250000001, 0.5270309278350516, 0.9981406249999999, 0.9903092783505155], "iscrowd": 0, "area": 25087.72589999999, "rle": {"size": [485, 640], "counts": "bXb6c0a>1O100O1O100O1O1O101N10000O1N2O1O1N2O1N2O1O1N2O1N2O1L4J6N2N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N200O1O1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1N2O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1M3M3M3N2M3M3eMcKkI`4T6WLPIo3P7S201O0001O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O001O001O001O001O001O001O001O001O001O001O00001O001O001O001O001YNf1cM]7"}, "label": "the grill of a dodge truck sits behind a parking meter"}]} {"id": 54318, "image": "COCO_train2014_000000054318.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bed with red bedding next to a window\"."}], "task": "refering", "answer_template": "The \"a bed with red bedding next to a window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [106, 109, 110, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 260, 261, 262, 263, 284, 285, 286, 308], "bbox": [0.29221875, 0.30785547785547784, 0.804375, 0.8764102564102565], "iscrowd": 0, "area": 37546.96045, "rle": {"size": [429, 640], "counts": "Zc^22X=4L5L4K5fC@`;d0YDCd;a0TDFh;R1L4M3L3O2N2O1N2N4L5K4M4K4L3M3N1O2N1O1O001O1O1O1O001O1O1O1O001O1O1O1O003M`0@00000000000000001O1N101O1O1O1O1O001O1O1O1N1000001O0000000000001N100000O1O1N2O1O1N2O1N101O1N2O1O1N2O1N2O1O1N101N2O1N2O1N2N2O1N2O1N2O0O2O1N2O1N2N2O1N2O1N2000O1000O10000000000O10000000000O10000000000O1000000000O01000000000000O1000O100000O10000000000O01000000000OE[MPFe2o9]MPFc2o9_MoEb2Q:_MmEb2R:_MmEb2R:`MlE`2U:aMiE`2V: in this image.", "objects": [{"patches": [106, 109, 110, 129, 130, 131, 132, 133, 151, 152, 153, 154, 155, 156, 171, 172, 173, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 260, 261, 262, 263, 284, 285, 286, 308], "bbox": [0.29221875, 0.30785547785547784, 0.804375, 0.8764102564102565], "iscrowd": 0, "area": 37546.96045, "rle": {"size": [429, 640], "counts": "Zc^22X=4L5L4K5fC@`;d0YDCd;a0TDFh;R1L4M3L3O2N2O1N2N4L5K4M4K4L3M3N1O2N1O1O001O1O1O1O001O1O1O1O001O1O1O1O003M`0@00000000000000001O1N101O1O1O1O1O001O1O1O1N1000001O0000000000001N100000O1O1N2O1O1N2O1N101O1N2O1O1N2O1N2O1O1N101N2O1N2O1N2N2O1N2O1N2O0O2O1N2O1N2N2O1N2O1N2000O1000O10000000000O10000000000O10000000000O1000000000O01000000000000O1000O100000O10000000000O01000000000OE[MPFe2o9]MPFc2o9_MoEb2Q:_MmEb2R:_MmEb2R:`MlE`2U:aMiE`2V: in this image.", "objects": [{"patches": [113, 114, 136, 137, 159, 160, 182, 183, 204, 205, 206, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.41481250000000003, 0.2937296037296037, 1.0, 0.9820512820512821], "iscrowd": 0, "area": 52567.058700000016, "rle": {"size": [429, 640], "counts": "mi_31W11k:1RE1l:1SE0l:2QE0n:1PE1o:1nD1Q;FUD8h04S;NkD4S;OkD2T;OjD3T;0iD2U;0iD2U;1hD1W;0fD3X;ObD7\\;n0O1N2N2N2O1N2N2O1N2N2O1N2N2N2O1N2N2O1N2O1N2O1O1O1O1N2O1O1O1O1O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O1000000000000O1000000000000O100000000000000O1000000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O10000000000O10000000000O1000000000000O10000000000O10000000000O10000000000O1000000GaKdG_4o7bKRH2N\\4m7fKSH0N[4m7hKRH1OW4l7m0O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1N2O1O1O1O1N2O1L4C=D in this image.", "objects": [{"patches": [97, 98, 99, 100, 114, 115, 116, 117, 118, 132, 133, 134, 135, 136], "bbox": [0.32804, 0.37551999999999996, 0.62016, 0.5678933333333333], "iscrowd": 0, "area": 6702.687249999998, "rle": {"size": [375, 500], "counts": "`Xl11d;2M3N2N2M3N2N2N2M2O2N2N2M3N2N2O10000O10000000000000O10O100BmNXFS1h9mNXFS1g9nNYFR1g9nNYFR1W9QOfFN3Q1V9XO^FJ in this image.", "objects": [{"patches": [131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 203, 204, 205, 206, 207, 208, 209, 210, 211, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229], "bbox": [0.1987, 0.5454666666666667, 0.76754, 0.9844266666666668], "iscrowd": 0, "area": 25448.68565000001, "rle": {"size": [375, 500], "counts": "dcT11_;71O0000000000001O00000000001O000000000000O100O1O100O1O1O100O1O1N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2O100kMhMkJX2S5jMlJW2Q5lMoJT2o4nMQKR2m4PNSKP2k4RNTKo1k4RNUKn1k4QNVKo1j4QNVKo1j4QNUKP2k4PNUKP2k4PNUKP2k4PNUKP2k4PNTKQ2l4oMTKQ2l4oMUKP2k4PNUKP2k4PNUKP2k4PNVKo1j4QNVKo1i4RNWKn1i4QNYKn1g4RNYKn1g4RNZKn1e4RN[Kn1e4RN[Kn1e4RN\\Km1d4SN\\Km1d4TN\\Kk1d4VN[Kj1e4WNZKi1f4XNZKg1f4[NXKe1h4\\NWKd1i4]NWKb1i4`NUK`1j4bNVK]1j4dNUK\\1k4eNfIi2Z6YMaIj2_6WM\\Im2d6TMXIo2h6P1000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000dL^Ic1b6VNfIi1Z6UNiIj1W6UNkIj1U6VNlIi1T6VNnIi1R6WNoIh1Q6WNQJh1o5XNRJg1n5XNTJg1l5YNUJf1k5YNVJg1j5YNWJf1h5ZNZJe1f5ZN[Jf1e5ZN\\Je1d5ZN]Jf1c5YN_Jf1a5ZN_Jf1a5YNaJf1_5YNcJf1]5ZNcJf1]5YNeJf1[5YNfJg1Z5YNgJf1Y5ZNgJf1Y5YNiJf1W5ZNiJf1W5YNkJf1T5[NmJd1S5\\NmJd1S5[NoJd1Q5\\NoJd1Q5[NQKd1o4\\NQKd1o4\\NRKc1n4\\NSKd1m4\\NTKc1l4]NUKb1k4]NVKc1j4]NWKb1i4]NXKc1h4]NYKb1g4^NYKb1h4\\NZKc1g4\\NYKd1i4YNYKf1h4YNXKg1j4WNVKi1k4UNVKk1l4SNUKl1l4RNUKn1m4PNSKP2[7O0000001O00001O0000001O00001O0000001O0000001O00001O0000O100O100O1O100O100O1003M3M2N3M3M3M2N3M3M3M2N3M3M2N3M3M3M2N3M3M3M2N3M3M3M2NlSZ1"}, "label": "white table in front of other tables which the group of three women are resting their hands on"}]} {"id": 554031, "image": "COCO_train2014_000000554031.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white table between two women in pony tails\"."}], "task": "refering", "answer_template": "The \"the white table between two women in pony tails\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 169, 170, 171, 172, 185, 186, 187, 188, 189, 190, 191, 192, 203, 204, 205, 206, 207, 208, 209, 210, 211, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229], "bbox": [0.1987, 0.5454666666666667, 0.76754, 0.9844266666666668], "iscrowd": 0, "area": 25448.68565000001, "rle": {"size": [375, 500], "counts": "dcT11_;71O0000000000001O00000000001O000000000000O100O1O100O1O1O100O1O1N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2O100kMhMkJX2S5jMlJW2Q5lMoJT2o4nMQKR2m4PNSKP2k4RNTKo1k4RNUKn1k4QNVKo1j4QNVKo1j4QNUKP2k4PNUKP2k4PNUKP2k4PNUKP2k4PNTKQ2l4oMTKQ2l4oMUKP2k4PNUKP2k4PNUKP2k4PNVKo1j4QNVKo1i4RNWKn1i4QNYKn1g4RNYKn1g4RNZKn1e4RN[Kn1e4RN[Kn1e4RN\\Km1d4SN\\Km1d4TN\\Kk1d4VN[Kj1e4WNZKi1f4XNZKg1f4[NXKe1h4\\NWKd1i4]NWKb1i4`NUK`1j4bNVK]1j4dNUK\\1k4eNfIi2Z6YMaIj2_6WM\\Im2d6TMXIo2h6P1000000000000000000O100000000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O100000dL^Ic1b6VNfIi1Z6UNiIj1W6UNkIj1U6VNlIi1T6VNnIi1R6WNoIh1Q6WNQJh1o5XNRJg1n5XNTJg1l5YNUJf1k5YNVJg1j5YNWJf1h5ZNZJe1f5ZN[Jf1e5ZN\\Je1d5ZN]Jf1c5YN_Jf1a5ZN_Jf1a5YNaJf1_5YNcJf1]5ZNcJf1]5YNeJf1[5YNfJg1Z5YNgJf1Y5ZNgJf1Y5YNiJf1W5ZNiJf1W5YNkJf1T5[NmJd1S5\\NmJd1S5[NoJd1Q5\\NoJd1Q5[NQKd1o4\\NQKd1o4\\NRKc1n4\\NSKd1m4\\NTKc1l4]NUKb1k4]NVKc1j4]NWKb1i4]NXKc1h4]NYKb1g4^NYKb1h4\\NZKc1g4\\NYKd1i4YNYKf1h4YNXKg1j4WNVKi1k4UNVKk1l4SNUKl1l4RNUKn1m4PNSKP2[7O0000001O00001O0000001O00001O0000001O0000001O00001O0000O100O100O1O100O100O1003M3M2N3M3M3M2N3M3M3M2N3M3M2N3M3M3M2N3M3M3M2N3M3M3M2NlSZ1"}, "label": "the white table between two women in pony tails"}]} {"id": 259120, "image": "COCO_train2014_000000259120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"light big brown chair\"."}], "task": "refering", "answer_template": "The \"light big brown chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.12453125000000001, 0.6129107981220658, 0.528203125, 0.9892488262910798], "iscrowd": 0, "area": 33852.48720000001, "rle": {"size": [426, 640], "counts": "TbQ1=lA5L4K4M4K5L3M4K2O0O100O10000O100O10000O100O100O10000O100O10000O101O00000O100000000O1000000O100000000O100000000O100000000O100000000O100000000O100000000O1000000O100000000O100000000O100000000O100000000O100000000O1001O1O001O001O1O001O1O1O2N2N2N3M2N:F1O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O000000001O000000O100O1O100O100O1O100O100O100O1O100O100O100O1O100O100O1O100O100O10000001O000O100O1O1O1O2N100O1O1O1O2O0O1O3M2N2N2O2M2N2N2N3M4M8G8H9G9G9G8I8G9G9G8H9HYXm3"}, "label": "light big brown chair"}]} {"id": 259120, "image": "COCO_train2014_000000259120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a beige sofa sitting on a green carpet\"."}], "task": "refering", "answer_template": "The \"a beige sofa sitting on a green carpet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 212, 213, 214, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 325, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.12453125000000001, 0.6129107981220658, 0.528203125, 0.9892488262910798], "iscrowd": 0, "area": 33852.48720000001, "rle": {"size": [426, 640], "counts": "TbQ1=lA5L4K4M4K5L3M4K2O0O100O10000O100O10000O100O100O10000O100O10000O101O00000O100000000O1000000O100000000O100000000O100000000O100000000O100000000O100000000O1000000O100000000O100000000O100000000O100000000O100000000O1001O1O001O001O1O001O1O1O2N2N2N3M2N:F1O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O000000001O000000O100O1O100O100O1O100O100O100O1O100O100O100O1O100O100O1O100O100O10000001O000O100O1O1O1O2N100O1O1O1O2O0O1O3M2N2N2O2M2N2N2N3M4M8G8H9G9G9G8I8G9G9G8H9HYXm3"}, "label": "a beige sofa sitting on a green carpet"}]} {"id": 259120, "image": "COCO_train2014_000000259120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the desk with the green chair\"."}], "task": "refering", "answer_template": "The \"the desk with the green chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 153, 154, 155, 175, 176, 177, 178, 179, 180, 200, 202, 203, 223, 225, 246], "bbox": [0.5842031249999999, 0.4141549295774648, 0.834171875, 0.6866901408450704], "iscrowd": 0, "area": 5470.0160499999965, "rle": {"size": [426, 640], "counts": "SVl41X=101O001N101O001N101O001O001N101O001O001N101O001O001O0O2O001O001O0O2O001O001O0O2O001O001O0O2O00001O001N101O001O001N101O001O001N101O001O001N1010O010O01O010O010O0010O01O0jEiNZ8W1nF@R9`0VF8k9c1000O2O001QNn100O2O001O00001O00001O001O00001O001N10001O00001O001O00001O001O00001O0O101O001O00001O001O00001O00001O0O2OPEBT9>gFGY99bFK`94[F1e9OVF6k9IPF;Q:EjE`0V:@eEe0\\:ZO_Ei0c:WOWEd0U;[OfD?e;7 in this image.", "objects": [{"patches": [151, 152, 153, 154, 155, 175, 176, 177, 178, 179, 180, 200, 202, 203, 223, 225, 246], "bbox": [0.5842031249999999, 0.4141549295774648, 0.834171875, 0.6866901408450704], "iscrowd": 0, "area": 5470.0160499999965, "rle": {"size": [426, 640], "counts": "SVl41X=101O001N101O001N101O001O001N101O001O001N101O001O001O0O2O001O001O0O2O001O001O0O2O001O001O0O2O00001O001N101O001O001N101O001O001N101O001O001N1010O010O01O010O010O0010O01O0jEiNZ8W1nF@R9`0VF8k9c1000O2O001QNn100O2O001O00001O00001O001O00001O001N10001O00001O001O00001O001O00001O0O101O001O00001O001O00001O00001O0O2OPEBT9>gFGY99bFK`94[F1e9OVF6k9IPF;Q:EjE`0V:@eEe0\\:ZO_Ei0c:WOWEd0U;[OfD?e;7 in this image.", "objects": [{"patches": [228, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 317, 318, 319, 320, 321, 337, 341, 342, 343, 344], "bbox": [0.6625, 0.655962441314554, 0.99921875, 0.9885211267605634], "iscrowd": 0, "area": 16860.81895, "rle": {"size": [426, 640], "counts": "bf`51Y=1N2O0O2^E5c7LeGk0[8VOmFa1R9aNWFV2h9f0O100O1000M2J7I6K6I6J7J6H=Da0_Ob0^O_o4KmPK8I8G9H4K100O010O010O01O010O010O01g0XOXk7GQUH1O1XDa0c9@ZFd0d9]OZFf0c9]OYFg0e9ZOYFi0e9XOYFk0e9VO[Fk0b9WO]Fk0a9VO_Fj0`9WO_Fj0`9WO`Fh0_9[O`Fe0_9\\OaFc0_9^OaFb0^9_OaFa0_9@aF`0]9CbF<^9EbF;]9FcF9]9HbF9\\9HeF7[9JeF6Z9JgF5Y9LgFM_93aFEf9=ZFZOn9f0SFQOU:P1R10O2O001O0O1000001O0O100RFmNe7S1SHUOm7U3N2O001O0O2O1O001O1N101O1O0O2O1O001O0O2O00000O1000000O100000000O1000000O101O00000O1000000O1000000O100000000O1000000O1:F1O1O00000000000000000000001O000000000000000O100O10J60O101O0O101O0O2O000O2O00`3_L_C"}, "label": "a dinner table"}]} {"id": 259120, "image": "COCO_train2014_000000259120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red stained mahogany table and a matching chair sit in the corner of the room\"."}], "task": "refering", "answer_template": "The \"a red stained mahogany table and a matching chair sit in the corner of the room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [228, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 317, 318, 319, 320, 321, 337, 341, 342, 343, 344], "bbox": [0.6625, 0.655962441314554, 0.99921875, 0.9885211267605634], "iscrowd": 0, "area": 16860.81895, "rle": {"size": [426, 640], "counts": "bf`51Y=1N2O0O2^E5c7LeGk0[8VOmFa1R9aNWFV2h9f0O100O1000M2J7I6K6I6J7J6H=Da0_Ob0^O_o4KmPK8I8G9H4K100O010O010O01O010O010O01g0XOXk7GQUH1O1XDa0c9@ZFd0d9]OZFf0c9]OYFg0e9ZOYFi0e9XOYFk0e9VO[Fk0b9WO]Fk0a9VO_Fj0`9WO_Fj0`9WO`Fh0_9[O`Fe0_9\\OaFc0_9^OaFb0^9_OaFa0_9@aF`0]9CbF<^9EbF;]9FcF9]9HbF9\\9HeF7[9JeF6Z9JgF5Y9LgFM_93aFEf9=ZFZOn9f0SFQOU:P1R10O2O001O0O1000001O0O100RFmNe7S1SHUOm7U3N2O001O0O2O1O001O1N101O1O0O2O1O001O0O2O00000O1000000O100000000O1000000O101O00000O1000000O1000000O100000000O1000000O1:F1O1O00000000000000000000001O000000000000000O100O10J60O101O0O101O0O2O000O2O00`3_L_C"}, "label": "a red stained mahogany table and a matching chair sit in the corner of the room"}]} {"id": 291897, "image": "COCO_train2014_000000291897.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of bread with garnish\"."}], "task": "refering", "answer_template": "The \"a piece of bread with garnish\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211], "bbox": [0.001390625, 0.3443055555555556, 0.280546875, 0.7735833333333334], "iscrowd": 0, "area": 21595.46085, "rle": {"size": [360, 640], "counts": "ca0?b:e0ZO:G9F9H9F9H9F9H9F9O2O0O2O0O101N101N100O100O100O100O100O1O1O1N2O1N2OOTIhKj6X4XIgKf6Z440O101N10000O100O100O10000O100O100O101O000000000001O0000000000001O0001O0000000001O000000000000001O0001O000001O0000001O0000001O00000010O000001OO100000000000000O100000N2O1O1O001O1O1N2O1O1O1O001O1O1N2O1O1O001O1O1O1N2O001N2J6J6J6J5]Od0H8N2O1O1O0O2O1O1O1O1N101O1O1N2O1O001N2O1N2N2M2O2N2N7IQcQ5"}, "label": "a piece of bread with garnish"}]} {"id": 291897, "image": "COCO_train2014_000000291897.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of bread with greens on it to the left of a sandwich\"."}], "task": "refering", "answer_template": "The \"a piece of bread with greens on it to the left of a sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 121, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211], "bbox": [0.001390625, 0.3443055555555556, 0.280546875, 0.7735833333333334], "iscrowd": 0, "area": 21595.46085, "rle": {"size": [360, 640], "counts": "ca0?b:e0ZO:G9F9H9F9H9F9H9F9O2O0O2O0O101N101N100O100O100O100O100O1O1O1N2O1N2OOTIhKj6X4XIgKf6Z440O101N10000O100O100O10000O100O100O101O000000000001O0000000000001O0001O0000000001O000000000000001O0001O000001O0000001O0000001O00000010O000001OO100000000000000O100000N2O1O1O001O1O1N2O1O1O1O001O1O1N2O1O1O001O1O1O1N2O001N2J6J6J6J5]Od0H8N2O1O1O0O2O1O1O1O1N101O1O1N2O1O001N2O1N2N2M2O2N2N7IQcQ5"}, "label": "a piece of bread with greens on it to the left of a sandwich"}]} {"id": 291897, "image": "COCO_train2014_000000291897.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwich between the soup and one half of a sandwich top\"."}], "task": "refering", "answer_template": "The \"the sandwich between the soup and one half of a sandwich top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 52, 53, 54, 55, 56, 75, 76, 77, 78, 79, 80, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 213, 214, 215, 216, 237, 238, 239], "bbox": [0.231015625, 0.07597222222222223, 0.5215624999999999, 0.8358888888888889], "iscrowd": 0, "area": 32756.982050000013, "rle": {"size": [360, 640], "counts": "_Wd15P;8I7H8H6K3L3M3N3N1O1O2N1O1O2O1N1N3M3M2N3M3M2gMWNRKk1j4YNSKj1i4ZNTKo1c4TNZKQ2`4SN^Kn1`4VNeJTONi2Y5WNgJb2U5bMiJ`2T5cMjJ^2S5fMkJ^2P5eMmJ_2o4dMoJ_2m4eMPK_2l4cMRK`2j4R2M3L5L3L3N2M3N2M3N2N2N2N101N2N2O1N2N2O0O2N2O1N2N2O1N2O0O2N2O1N2N100O1O100O1O100000001O00000000000000O10000O10000O10000O10O1000O101O0O2O001N2N1O2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N1N3M3L3N3L3N3L4M2M4M3L4L4L4K5K5L4K4M4K5L3L5K5L3L5L3L5N4L4L2N2N3M2O1N3N7H101N1SMUH=OT1m7\\NVH`0OQ1l7^NWH`0OP1k7]NYHb0Og0MaNl74[Hb0Of0n7fNUHc00d0]8ZOfGd0[8YOiGd0Y8YOlGd0T8ZOQH0TO5k8IUHMSO:g8G[HJQO?d8DiH in this image.", "objects": [{"patches": [30, 31, 32, 33, 52, 53, 54, 55, 56, 75, 76, 77, 78, 79, 80, 98, 99, 100, 101, 102, 103, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 213, 214, 215, 216, 237, 238, 239], "bbox": [0.231015625, 0.07597222222222223, 0.5215624999999999, 0.8358888888888889], "iscrowd": 0, "area": 32756.982050000013, "rle": {"size": [360, 640], "counts": "_Wd15P;8I7H8H6K3L3M3N3N1O1O2N1O1O2O1N1N3M3M2N3M3M2gMWNRKk1j4YNSKj1i4ZNTKo1c4TNZKQ2`4SN^Kn1`4VNeJTONi2Y5WNgJb2U5bMiJ`2T5cMjJ^2S5fMkJ^2P5eMmJ_2o4dMoJ_2m4eMPK_2l4cMRK`2j4R2M3L5L3L3N2M3N2M3N2N2N2N101N2N2O1N2N2O0O2N2O1N2N2O1N2O0O2N2O1N2N100O1O100O1O100000001O00000000000000O10000O10000O10000O10O1000O101O0O2O001N2N1O2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N1N3M3L3N3L3N3L4M2M4M3L4L4L4K5K5L4K4M4K5L3L5K5L3L5L3L5N4L4L2N2N3M2O1N3N7H101N1SMUH=OT1m7\\NVH`0OQ1l7^NWH`0OP1k7]NYHb0Og0MaNl74[Hb0Of0n7fNUHc00d0]8ZOfGd0[8YOiGd0Y8YOlGd0T8ZOQH0TO5k8IUHMSO:g8G[HJQO?d8DiH in this image.", "objects": [{"patches": [9, 10, 11, 12, 29, 30, 31, 32, 33, 34, 35, 36, 37, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 402, 403, 404, 405], "bbox": [0.21334375, 0.024716599190283402, 0.917578125, 1.0], "iscrowd": 0, "area": 166185.89395, "rle": {"size": [494, 640], "counts": "b^R2f0V>h0XOh0oNo0mNR1YOg0D=C^Oc0XOg0jMW2F9M3N3M2N3N4K5K4L5L4K5K4M:Cd0[Of0YOnZi0"}, "label": "a red double decker bus"}]} {"id": 545850, "image": "COCO_train2014_000000545850.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red bus that is two stories\"."}], "task": "refering", "answer_template": "The \"a red bus that is two stories\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 29, 30, 31, 32, 33, 34, 35, 36, 37, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 402, 403, 404, 405], "bbox": [0.21334375, 0.024716599190283402, 0.917578125, 1.0], "iscrowd": 0, "area": 166185.89395, "rle": {"size": [494, 640], "counts": "b^R2f0V>h0XOh0oNo0mNR1YOg0D=C^Oc0XOg0jMW2F9M3N3M2N3N4K5K4L5L4K5K4M:Cd0[Of0YOnZi0"}, "label": "a red bus that is two stories"}]} {"id": 160852, "image": "COCO_train2014_000000160852.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cow that is the smallest\"."}], "task": "refering", "answer_template": "The \"the cow that is the smallest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [230, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 306, 324, 325, 326, 328, 329, 347, 348, 349, 351], "bbox": [0.0294375, 0.5835, 0.31790625, 0.945375], "iscrowd": 0, "area": 16784.952899999993, "rle": {"size": [480, 640], "counts": "UV94l>5K2OO000001O0001O00000000010O000000000010O000]Af0_=>O2N1N3N1O1O2N1O2M2O1O2O001O001O1O2N2N2O0O2N4L3M4L4L3M400O1L3M4L4mDfL9OP9X4iFhKX9^4aFcK_9d4YF]Kf9o40O1000000O10001M2_Ob0K5K5K4J7E;H701O1N101O0000XEUMZ9k2TFhMl9g3O1O001N101O001O0O2O001O001N_OPFTLP:^3dF]L[9^3oF]LQ9^3[1K4L5K5M3O1O00100O1O10O1000000O01000O10000O1000O010000O10000O01000O0100O010O10O0100O05L2N2M2O2N1N3N10001N5L4L5K4L5K4L2N1O2N1O1OO1M;F in this image.", "objects": [{"patches": [230, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 306, 324, 325, 326, 328, 329, 347, 348, 349, 351], "bbox": [0.0294375, 0.5835, 0.31790625, 0.945375], "iscrowd": 0, "area": 16784.952899999993, "rle": {"size": [480, 640], "counts": "UV94l>5K2OO000001O0001O00000000010O000000000010O000]Af0_=>O2N1N3N1O1O2N1O2M2O1O2O001O001O1O2N2N2O0O2N4L3M4L4L3M400O1L3M4L4mDfL9OP9X4iFhKX9^4aFcK_9d4YF]Kf9o40O1000000O10001M2_Ob0K5K5K4J7E;H701O1N101O0000XEUMZ9k2TFhMl9g3O1O001N101O001O0O2O001O001N_OPFTLP:^3dF]L[9^3oF]LQ9^3[1K4L5K5M3O1O00100O1O10O1000000O01000O10000O1000O010000O10000O01000O0100O010O10O0100O05L2N2M2O2N1N3N10001N5L4L5K4L5K4L2N1O2N1O1OO1M;F in this image.", "objects": [{"patches": [212, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 242, 259, 260, 261, 262, 263, 264, 265, 282, 283, 284, 285, 286, 287, 288, 306, 307, 308, 309, 310, 311, 329, 330, 331, 332, 333, 334, 353, 354, 355, 357, 376, 377, 380], "bbox": [0.246390625, 0.525875, 0.5641875, 0.9680416666666666], "iscrowd": 0, "area": 24727.2614, "rle": {"size": [480, 640], "counts": "aZZ21o>1N2O1O1O0O2O1O001N101O001O0O2O001O1N101O001O0O2O001O0dB_OXi in this image.", "objects": [{"patches": [212, 213, 214, 215, 216, 217, 218, 236, 237, 238, 239, 240, 241, 242, 259, 260, 261, 262, 263, 264, 265, 282, 283, 284, 285, 286, 287, 288, 306, 307, 308, 309, 310, 311, 329, 330, 331, 332, 333, 334, 353, 354, 355, 357, 376, 377, 380], "bbox": [0.246390625, 0.525875, 0.5641875, 0.9680416666666666], "iscrowd": 0, "area": 24727.2614, "rle": {"size": [480, 640], "counts": "aZZ21o>1N2O1O1O0O2O1O001N101O001O0O2O001O1N101O001O0O2O001O0dB_OXi in this image.", "objects": [{"patches": [197, 218, 219, 220, 221, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 342, 343, 344, 360, 365, 366, 367, 383, 388, 390], "bbox": [0.50965625, 0.519, 0.9712343750000001, 0.9760624999999999], "iscrowd": 0, "area": 35550.63284999997, "rle": {"size": [480, 640], "counts": "VRi43l>3M3N1O2N2N1O1O2N1N101O00001O01O001O001O000001O:F3M3M>B1N3OOM4M3M4M2N2O6I3M3M4M2N2N2N2M3N1O2N2N2N2N3M2O2M:F5L3L3N0010O1O100O1O01O0O1O2M2N2O1N2N100O2O00001N101O0000oKcE`3\\:_LfE`3Z:_LhEb3W:[LlEg3Q:XLQFi3n9TLUFn3h9PLZFR4e9lK\\FW4U:2N7Ic0]O>B001O000000001O00001O001O00001O000001O00000O10000PO]GcKc8Y4T1M3M4M2L4L4DGho7"}, "label": "the cow on the far right of the herd"}]} {"id": 160852, "image": "COCO_train2014_000000160852.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cow on the right\"."}], "task": "refering", "answer_template": "The \"the cow on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [197, 218, 219, 220, 221, 225, 226, 227, 228, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 336, 337, 338, 339, 340, 342, 343, 344, 360, 365, 366, 367, 383, 388, 390], "bbox": [0.50965625, 0.519, 0.9712343750000001, 0.9760624999999999], "iscrowd": 0, "area": 35550.63284999997, "rle": {"size": [480, 640], "counts": "VRi43l>3M3N1O2N2N1O1O2N1N101O00001O01O001O001O000001O:F3M3M>B1N3OOM4M3M4M2N2O6I3M3M4M2N2N2N2M3N1O2N2N2N2N3M2O2M:F5L3L3N0010O1O100O1O01O0O1O2M2N2O1N2N100O2O00001N101O0000oKcE`3\\:_LfE`3Z:_LhEb3W:[LlEg3Q:XLQFi3n9TLUFn3h9PLZFR4e9lK\\FW4U:2N7Ic0]O>B001O000000001O00001O001O00001O000001O00000O10000PO]GcKc8Y4T1M3M4M2L4L4DGho7"}, "label": "the cow on the right"}]} {"id": 447574, "image": "COCO_train2014_000000447574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of soup next to a sandwich\"."}], "task": "refering", "answer_template": "The \"a bowl of soup next to a sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 173, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 201, 202, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 291, 292], "bbox": [0.52928125, 0.4064583333333333, 0.822390625, 0.7166458333333333], "iscrowd": 0, "area": 21526.516250000004, "rle": {"size": [480, 640], "counts": "]Uo44h>5K6K4K6J5K6K4K6J5K6K4K5K5K5N2O1N2N2N2O1N2N2N2O1N2N2N3N1N2N2N2O1N2N2N2O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O001O0O2O001O001O001O001O001O001N101O001O001O0000000000000O100000001O000000000000000000000001O0000000000000001O000000000000000000000000O10000O1000000O101O0O2O00001N101O000O2O001N10001O0O2O0O1O2M2O2N1O1O2M2O2N1O2M4M2M4L3M4L3M4L3M3M3M2N3M3L4M3M2N3M3M3M3M2N3M3M3M8H8H8G8IQVe1"}, "label": "a bowl of soup next to a sandwich"}]} {"id": 447574, "image": "COCO_train2014_000000447574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bowl of soup next to a sandwich on a white plate\"."}], "task": "refering", "answer_template": "The \"a white bowl of soup next to a sandwich on a white plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 173, 174, 175, 176, 177, 178, 179, 196, 197, 198, 199, 200, 201, 202, 219, 220, 221, 222, 223, 224, 225, 242, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 291, 292], "bbox": [0.52928125, 0.4064583333333333, 0.822390625, 0.7166458333333333], "iscrowd": 0, "area": 21526.516250000004, "rle": {"size": [480, 640], "counts": "]Uo44h>5K6K4K6J5K6K4K6J5K6K4K5K5K5N2O1N2N2N2O1N2N2N2O1N2N2N3N1N2N2N2O1N2N2N2O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O001O0O2O001O001O001O001O001O001N101O001O001O0000000000000O100000001O000000000000000000000001O0000000000000001O000000000000000000000000O10000O1000000O101O0O2O00001N101O000O2O001N10001O0O2O0O1O2M2O2N1O1O2M2O2N1O2M4M2M4L3M4L3M4L3M3M3M2N3M3L4M3M2N3M3M3M3M2N3M3M3M8H8H8G8IQVe1"}, "label": "a white bowl of soup next to a sandwich on a white plate"}]} {"id": 447574, "image": "COCO_train2014_000000447574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"spoons\"."}], "task": "refering", "answer_template": "The \"spoons\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 180, 181, 182, 204, 205, 228, 229, 251, 252, 275, 298], "bbox": [0.855109375, 0.3769166666666666, 1.0, 0.7312708333333333], "iscrowd": 0, "area": 4032.390949999999, "rle": {"size": [480, 640], "counts": "_cP84c>e=1O001O1O001O1O001N2O001O1O1O1O0O2O1N2O0O2O1N2O0O100O100O10N2O1O001N2O001OTF"}, "label": "spoons"}]} {"id": 447574, "image": "COCO_train2014_000000447574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver spoon lying on the table next to a knife\"."}], "task": "refering", "answer_template": "The \"a silver spoon lying on the table next to a knife\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 180, 181, 182, 204, 205, 228, 229, 251, 252, 275, 298], "bbox": [0.855109375, 0.3769166666666666, 1.0, 0.7312708333333333], "iscrowd": 0, "area": 4032.390949999999, "rle": {"size": [480, 640], "counts": "_cP84c>e=1O001O1O001O1O001N2O001O1O1O1O0O2O1N2O0O2O1N2O0O100O100O10N2O1O001N2O001OTF"}, "label": "a silver spoon lying on the table next to a knife"}]} {"id": 447574, "image": "COCO_train2014_000000447574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of the sandwich on the back of the plate\"."}], "task": "refering", "answer_template": "The \"half of the sandwich on the back of the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 218, 219, 236, 237, 238, 239, 240, 241], "bbox": [0.25728125, 0.42364583333333333, 0.52803125, 0.6321666666666667], "iscrowd": 0, "area": 12644.08635, "rle": {"size": [480, 640], "counts": "mc]23k>3M3M3M2N3M3L4M3M2N3M3M3eNTO^Do0a;\\OQDh0m;CgC?Y in this image.", "objects": [{"patches": [167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 213, 214, 215, 216, 217, 218, 219, 236, 237, 238, 239, 240, 241], "bbox": [0.25728125, 0.42364583333333333, 0.52803125, 0.6321666666666667], "iscrowd": 0, "area": 12644.08635, "rle": {"size": [480, 640], "counts": "mc]23k>3M3M3M2N3M3L4M3M2N3M3M3eNTO^Do0a;\\OQDh0m;CgC?Y in this image.", "objects": [{"patches": [62, 63, 64, 65, 85, 86, 87, 88, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 153, 154, 155, 156, 178, 179], "bbox": [0.6893281250000001, 0.13483333333333333, 0.8545, 0.4404583333333333], "iscrowd": 0, "area": 11599.341699999997, "rle": {"size": [480, 640], "counts": "Rl^67[>f0ZOf0[Oe0D4K5L5J5L4K5L5J5L4L4O1O2N1O1O1O2N1N2O1O2O000000001O001O001O0O2O001O001O001O001O001O001O001O001O001O000000000001O00001N1N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N2O2M2N2N3M2N2N3M2M3M4K4M3M4L3M3L5L3M3M3M4K9H7I7I8G8In_[1"}, "label": "glass of water on a table"}]} {"id": 447574, "image": "COCO_train2014_000000447574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"clear glass full of water\"."}], "task": "refering", "answer_template": "The \"clear glass full of water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 65, 85, 86, 87, 88, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 153, 154, 155, 156, 178, 179], "bbox": [0.6893281250000001, 0.13483333333333333, 0.8545, 0.4404583333333333], "iscrowd": 0, "area": 11599.341699999997, "rle": {"size": [480, 640], "counts": "Rl^67[>f0ZOf0[Oe0D4K5L5J5L4K5L5J5L4L4O1O2N1O1O1O2N1N2O1O2O000000001O001O001O0O2O001O001O001O001O001O001O001O001O001O000000000001O00001N1N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N2O2M2N2N3M2N2N3M2M3M4K4M3M4L3M3L5L3M3M3M4K9H7I7I8G8In_[1"}, "label": "clear glass full of water"}]} {"id": 70745, "image": "COCO_train2014_000000070745.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant pushing the stone\"."}], "task": "refering", "answer_template": "The \"elephant pushing the stone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 50, 51, 52, 53, 54, 55, 56, 57, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 309, 310, 327], "bbox": [0.14070312499999998, 0.12850117096018734, 0.56734375, 0.9538875878220141], "iscrowd": 0, "area": 63194.476250000014, "rle": {"size": [427, 640], "counts": "ndU15R= in this image.", "objects": [{"patches": [30, 50, 51, 52, 53, 54, 55, 56, 57, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 188, 189, 190, 191, 192, 193, 194, 211, 212, 213, 214, 215, 216, 217, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 309, 310, 327], "bbox": [0.14070312499999998, 0.12850117096018734, 0.56734375, 0.9538875878220141], "iscrowd": 0, "area": 63194.476250000014, "rle": {"size": [427, 640], "counts": "ndU15R= in this image.", "objects": [{"patches": [113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 226, 227, 228, 229, 251, 252, 274, 275, 298, 321], "bbox": [0.85759375, 0.2763934426229508, 1.0, 0.9146135831381732], "iscrowd": 0, "area": 12325.4634, "rle": {"size": [427, 640], "counts": "UVU72W=4K4M3M4K4M001N10000O10001L3G9H8I7O1N2O2M2O1O1N2OL3K62M7L4L4L4L4L4L4L4L3L5L4L4L4L4L4L4L4L4L3M4L4L4L4L4L4K6K4L4L5K4L5L3L4L5L3L4M4K4L4M7H in this image.", "objects": [{"patches": [113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 226, 227, 228, 229, 251, 252, 274, 275, 298, 321], "bbox": [0.85759375, 0.2763934426229508, 1.0, 0.9146135831381732], "iscrowd": 0, "area": 12325.4634, "rle": {"size": [427, 640], "counts": "UVU72W=4K4M3M4K4M001N10000O10001L3G9H8I7O1N2O2M2O1O1N2OL3K62M7L4L4L4L4L4L4L4L3L5L4L4L4L4L4L4L4L4L3M4L4L4L4L4L4K6K4L4L5K4L5L3L4L5L3L4M4K4L4M7H in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 23, 24, 25, 26, 46, 47, 48, 49, 71, 72, 95, 118, 119, 120, 140, 141, 142, 143, 163, 164, 165], "bbox": [0.0, 0.0, 0.255140625, 0.4568200836820084], "iscrowd": 0, "area": 12547.76055, "rle": {"size": [478, 640], "counts": "0X1f=1O2N1O1O2N1O1O2N1O1O001O001O001O1O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O1O001O001O1O001O1O1O001O1O1O1O001O1O1O001TGWMa5j2[JZMd5g2XJ]Mg5e2TJ_Mk5c2PJaMo5b2jIcMU6_2fIeMY6]2bIgM]6[2`IgM_6\\2^IeMa6]2\\IeMc6]2[IdMd6^2YIdMf6_2VIcMi6_2UIbMj6`2SIbMl6`2RIaMm6b2oH`MP7b2mH`MR7b2lH^MT7d2iH]MW7e0jGIl0BZ7b0PHKc0E\\7>WHK:H_7:]HM2Ia77cH0HJe74iH0@Lg71oH2WOOi7MVI2oN1k7J\\I5eN2o7FbI7]N3Q8DhI7TN6T8@nIR2R6kMTJS2k5kM[JQ2e5lMaJQ2_5lMhJP2X5nMmJo1S5oMRKo1m4nMZKn1f4PN_Km1a4QNdKl1\\4RNjKj1V4SNPLi1Q4UNULf1l3XNYLd1h3YN^Lc1c3[N`Lc1a3[NaLd1`3ZNbLe1\\8O1O1O1O1O1O1O1N2O2N1O1O1O1O1O1O1O1O1O2N1O2N1O1N3N1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1OP`n6"}, "label": "a clear wine glass with liquid inside of it"}]} {"id": 21595, "image": "COCO_train2014_000000021595.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"empty clear glass margarita glass\"."}], "task": "refering", "answer_template": "The \"empty clear glass margarita glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 23, 24, 25, 26, 46, 47, 48, 49, 71, 72, 95, 118, 119, 120, 140, 141, 142, 143, 163, 164, 165], "bbox": [0.0, 0.0, 0.255140625, 0.4568200836820084], "iscrowd": 0, "area": 12547.76055, "rle": {"size": [478, 640], "counts": "0X1f=1O2N1O1O2N1O1O2N1O1O001O001O001O1O001O001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O001O001O1O001O001O1O001O1O1O001O1O1O1O001O1O1O001TGWMa5j2[JZMd5g2XJ]Mg5e2TJ_Mk5c2PJaMo5b2jIcMU6_2fIeMY6]2bIgM]6[2`IgM_6\\2^IeMa6]2\\IeMc6]2[IdMd6^2YIdMf6_2VIcMi6_2UIbMj6`2SIbMl6`2RIaMm6b2oH`MP7b2mH`MR7b2lH^MT7d2iH]MW7e0jGIl0BZ7b0PHKc0E\\7>WHK:H_7:]HM2Ia77cH0HJe74iH0@Lg71oH2WOOi7MVI2oN1k7J\\I5eN2o7FbI7]N3Q8DhI7TN6T8@nIR2R6kMTJS2k5kM[JQ2e5lMaJQ2_5lMhJP2X5nMmJo1S5oMRKo1m4nMZKn1f4PN_Km1a4QNdKl1\\4RNjKj1V4SNPLi1Q4UNULf1l3XNYLd1h3YN^Lc1c3[N`Lc1a3[NaLd1`3ZNbLe1\\8O1O1O1O1O1O1O1N2O2N1O1O1O1O1O1O1O1O1O2N1O2N1O1N3N1O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1OP`n6"}, "label": "empty clear glass margarita glass"}]} {"id": 496732, "image": "COCO_train2014_000000496732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in jean jacket with her back turned to man eating\"."}], "task": "refering", "answer_template": "The \"woman in jean jacket with her back turned to man eating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 84, 85, 86, 87, 88, 107, 108, 109, 110, 111, 131, 132, 133, 152, 153, 154, 155, 156, 157, 177, 178, 179, 180, 181, 201, 202, 203, 204, 225, 226, 227], "bbox": [0.626015625, 0.13854166666666667, 0.8862187499999999, 0.5803541666666666], "iscrowd": 0, "area": 17233.7064, "rle": {"size": [480, 640], "counts": "_Tl51o>001O001N10001O001O001N2O001O001O1N101O001O1O0O2O001O1O001O0O2O1O001O001N2O001eMUO_Fk0`9\\OZFe0d9CoDE;i0d:IfDEc0c0e:c0YE^Oe:e0YE[Of:h0XEYOf:o0SEROk:T1PElNo:W1oDjNn:Z1oDgNP;[1oDfNo:]1oDdNo:^1PEbNo:i2M3O0O2O1N101O1N2O0O2O1N101N2O1N101N2O001O1N2O001O1O001O1O1O001O1O001O1N2O001O1O001O1O1O001O1O001O1O10O01O10O0101N3MXMhFKV93mFNP90SG2k8KXG7e8H]G:a8CbG?[8_OhGc0U8\\OmGf0Q8WORHk0k7TOWHk0j7ROXHn0h7QOZHn0f7PO\\HP1e7nN\\HR1d7lN_HS1b7kN_HU1c7gN`HX1b7dN`H\\1a7bN`H]1c7_N`H`1b7\\N`Hd1b7XN`Hh1a7UNbHj1`7RNbHn1c7kM`HT2d7fM^HZ2g7`MZH_2l7ZMWHe2n7TMTHl2Q:O001O0O2O0O2N1O2N1O2O0O2N1O2N1O3K4M3L4M4K4M3`NcBX1b=fNjBk0Z=UOmQR1"}, "label": "woman in jean jacket with her back turned to man eating"}]} {"id": 496732, "image": "COCO_train2014_000000496732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a blue - jean jacket sitting behind a man eating a sandwich\"."}], "task": "refering", "answer_template": "The \"a woman wearing a blue - jean jacket sitting behind a man eating a sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 64, 84, 85, 86, 87, 88, 107, 108, 109, 110, 111, 131, 132, 133, 152, 153, 154, 155, 156, 157, 177, 178, 179, 180, 181, 201, 202, 203, 204, 225, 226, 227], "bbox": [0.626015625, 0.13854166666666667, 0.8862187499999999, 0.5803541666666666], "iscrowd": 0, "area": 17233.7064, "rle": {"size": [480, 640], "counts": "_Tl51o>001O001N10001O001O001N2O001O001O1N101O001O1O0O2O001O1O001O0O2O1O001O001N2O001eMUO_Fk0`9\\OZFe0d9CoDE;i0d:IfDEc0c0e:c0YE^Oe:e0YE[Of:h0XEYOf:o0SEROk:T1PElNo:W1oDjNn:Z1oDgNP;[1oDfNo:]1oDdNo:^1PEbNo:i2M3O0O2O1N101O1N2O0O2O1N101N2O1N101N2O001O1N2O001O1O001O1O1O001O1O001O1N2O001O1O001O1O1O001O1O001O1O10O01O10O0101N3MXMhFKV93mFNP90SG2k8KXG7e8H]G:a8CbG?[8_OhGc0U8\\OmGf0Q8WORHk0k7TOWHk0j7ROXHn0h7QOZHn0f7PO\\HP1e7nN\\HR1d7lN_HS1b7kN_HU1c7gN`HX1b7dN`H\\1a7bN`H]1c7_N`H`1b7\\N`Hd1b7XN`Hh1a7UNbHj1`7RNbHn1c7kM`HT2d7fM^HZ2g7`MZH_2l7ZMWHe2n7TMTHl2Q:O001O0O2O0O2N1O2N1O2O0O2N1O2N1O3K4M3L4M4K4M3`NcBX1b=fNjBk0Z=UOmQR1"}, "label": "a woman wearing a blue - jean jacket sitting behind a man eating a sandwich"}]} {"id": 496732, "image": "COCO_train2014_000000496732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man biting burger holding with two hands\"."}], "task": "refering", "answer_template": "The \"a man biting burger holding with two hands\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 123, 124, 125, 126, 127, 128, 129, 145, 146, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 352, 353, 354, 356, 357, 358, 359, 360, 362, 363, 364, 365, 383, 386, 387, 388], "bbox": [0.17021875, 0.2516875, 0.9084218749999999, 0.9752708333333333], "iscrowd": 0, "area": 106315.57669999998, "rle": {"size": [480, 640], "counts": "R^c13h>8H8I6I8I7H7L3L4L5K4L4L4L4M3L4L4O1N2N2N3M2N2N2N2N2N2N2N2M3N1O2M3M3oNP1N3N2N2N1N3N2N2M2O2N2M3N1O2M2O1O1N2O1O2N1N2O1O1N200O1O1O1O1O101N1O1O1O1O1O100O1O1O001O10O01O001O1O010O001O1O010O1O0010O01O1O0010O01O1O010lHTIS6m6^IaIb6_6QIoIn6R7O100O1O1N1O2N2O1N2N2O1N1O2N2O1N2N2O1N1O2O1N2N2N2O1N2O010O1O1O100O1O00100O1O1O100O00100000000000O10O1000000000O1000000000000O100000O100000O100000000000000O100000O1000001O00001O00001O00001O0001O01O000010O0001O000001O01O0000010O000000010O00000010O000001O0001O01O000010O0001O0010O1O010O1O010O1O10O01O10O01O10O01O10O01O100O00100O00100O00100O1O010O1O001O101N1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O00100O001O001O010O1O001O001N101N2O0O2O001N101N2O0L5J5J7J5L5O1N101O001N2O001O0O2O001O1N101O1O1N2O1O001O1N2O1O1O1N101O1O1O1O100O1O010O1O100O10O01O100O00100O00100O00100O10O01O10O0mGTJS7k5kHWJV7i5gHYJY7g5eH[J\\7d5bH^J_7b5^H`Jb7`5[HcJf7\\5XHfJi7Z5THiJl7V5RHlJn7U5oGmJR8R5lGPKU8o4hGTKX8j51O4M3L4L5K4L4L4L3M2O1N2N2N1O2N2N2N2O0O2N2N2N2N2N1O2O1N2N2N1O2N2N8H:G9F:F7I1O1O1N1O2M3N2N2N1O2N2N2N1N3N2N2N2N1O2N2N2M3N1O2N2N2N2N1K6A?@VYk0"}, "label": "a man biting burger holding with two hands"}]} {"id": 496732, "image": "COCO_train2014_000000496732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man eating\"."}], "task": "refering", "answer_template": "The \"man eating\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 123, 124, 125, 126, 127, 128, 129, 145, 146, 147, 148, 149, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 352, 353, 354, 356, 357, 358, 359, 360, 362, 363, 364, 365, 383, 386, 387, 388], "bbox": [0.17021875, 0.2516875, 0.9084218749999999, 0.9752708333333333], "iscrowd": 0, "area": 106315.57669999998, "rle": {"size": [480, 640], "counts": "R^c13h>8H8I6I8I7H7L3L4L5K4L4L4L4M3L4L4O1N2N2N3M2N2N2N2N2N2N2N2M3N1O2M3M3oNP1N3N2N2N1N3N2N2M2O2N2M3N1O2M2O1O1N2O1O2N1N2O1O1N200O1O1O1O1O101N1O1O1O1O1O100O1O1O001O10O01O001O1O010O001O1O010O1O0010O01O1O0010O01O1O010lHTIS6m6^IaIb6_6QIoIn6R7O100O1O1N1O2N2O1N2N2O1N1O2N2O1N2N2O1N1O2O1N2N2N2O1N2O010O1O1O100O1O00100O1O1O100O00100000000000O10O1000000000O1000000000000O100000O100000O100000000000000O100000O1000001O00001O00001O00001O0001O01O000010O0001O000001O01O0000010O000000010O00000010O000001O0001O01O000010O0001O0010O1O010O1O010O1O10O01O10O01O10O01O10O01O100O00100O00100O00100O1O010O1O001O101N1O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O1O1O1O00100O001O001O010O1O001O001N101N2O0O2O001N101N2O0L5J5J7J5L5O1N101O001N2O001O0O2O001O1N101O1O1N2O1O001O1N2O1O1O1N101O1O1O1O100O1O010O1O100O10O01O100O00100O00100O00100O10O01O10O0mGTJS7k5kHWJV7i5gHYJY7g5eH[J\\7d5bH^J_7b5^H`Jb7`5[HcJf7\\5XHfJi7Z5THiJl7V5RHlJn7U5oGmJR8R5lGPKU8o4hGTKX8j51O4M3L4L5K4L4L4L3M2O1N2N2N1O2N2N2N2O0O2N2N2N2N2N1O2O1N2N2N1O2N2N8H:G9F:F7I1O1O1N1O2M3N2N2N1O2N2N2N1N3N2N2N2N1O2N2N2M3N1O2N2N2N2N1K6A?@VYk0"}, "label": "man eating"}]} {"id": 496732, "image": "COCO_train2014_000000496732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the seat the man eating the sandwich is sitting on\"."}], "task": "refering", "answer_template": "The \"the seat the man eating the sandwich is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [225, 226, 227, 228, 248, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 319, 320, 321, 342, 343, 344, 365, 366, 367, 388, 389, 390], "bbox": [0.806640625, 0.5794374999999999, 0.998046875, 0.9830833333333333], "iscrowd": 0, "area": 13109.375, "rle": {"size": [480, 640], "counts": "fTb71o>1O1O2N1O2N1O2N1O2N1O1O2N1O1O1O1O001O1O001O001O1O001O001O1O1O4L4eBPO^A>C>A?A?A?Aa0_Ob0_Oa0^Ob0^Oc5"}, "label": "the seat the man eating the sandwich is sitting on"}]} {"id": 496732, "image": "COCO_train2014_000000496732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green bench with a older man sitting on it eaating a sandwich\"."}], "task": "refering", "answer_template": "The \"a green bench with a older man sitting on it eaating a sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [225, 226, 227, 228, 248, 249, 250, 251, 252, 272, 273, 274, 275, 295, 296, 297, 298, 319, 320, 321, 342, 343, 344, 365, 366, 367, 388, 389, 390], "bbox": [0.806640625, 0.5794374999999999, 0.998046875, 0.9830833333333333], "iscrowd": 0, "area": 13109.375, "rle": {"size": [480, 640], "counts": "fTb71o>1O1O2N1O2N1O2N1O2N1O1O2N1O1O1O1O001O1O001O001O1O001O001O1O1O4L4eBPO^A>C>A?A?A?Aa0_Ob0_Oa0^Ob0^Oc5"}, "label": "a green bench with a older man sitting on it eaating a sandwich"}]} {"id": 496732, "image": "COCO_train2014_000000496732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of the seat to the left of the man\"."}], "task": "refering", "answer_template": "The \"the back of the seat to the left of the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 210, 211, 212, 233, 234, 256, 257, 279, 280, 302, 303, 325, 326, 327, 328, 329, 349, 350, 351, 352], "bbox": [0.150390625, 0.5221250000000001, 0.330078125, 0.9023541666666667], "iscrowd": 0, "area": 6442.325, "rle": {"size": [480, 640], "counts": "R]]14^>>C=C=C=B>C=D=B=Da90001O00000001O0001O00000000001O0000000000001OO2O0O100O1O101N100O100O2O0O100O1O2O0O100O1O2O0O1O100O100O1O100O1cfX6"}, "label": "the back of the seat to the left of the man"}]} {"id": 496732, "image": "COCO_train2014_000000496732.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person outside\"."}], "task": "refering", "answer_template": "The \"person outside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 24, 25, 26, 47, 48, 49, 50, 93, 94, 117, 140, 162, 163, 185, 186], "bbox": [0.056874999999999995, 0.0, 0.18410937500000002, 0.4983541666666667], "iscrowd": 0, "area": 5622.68255, "rle": {"size": [480, 640], "counts": "Um`02c>;K5L4aG^O^2d0_M@_2a0`M@^2b0aM_O]2c0bM^O\\2d0cM^OZ2d0eM]OX2f0eHmNa4>g2h0hHkNV4h0Q3=hHlN7NW3k0j3;hHmN5NR1O>m0c58iHoN1OV1N8P1g54jHoN00W103o0l52jHPON0Y11Oo0P60jHPOL2[11KP1S6MkHQOJ2]13Fo0X6KkHROG3`13AP1]6HkHROC7d11]OQ1`6ElHDX1HVOR1f6BlHDX1JROR1j6@lHDX1MmNQ1o6^OlHDX1OhNS1S7ZOmHDX11dNS1W7XOcHNb1J_NR1\\7VO`H1e1IZNT1`7 in this image.", "objects": [{"patches": [199, 200, 201, 202, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 292, 293, 294], "bbox": [0.680015625, 0.5723167848699764, 0.8682187499999999, 0.8546099290780141], "iscrowd": 0, "area": 10962.8146, "rle": {"size": [423, 640], "counts": "Pnc51V=4cD6T9NUFg0j9]O_EW1`:j0M2O1N2O2N1N2O1N3N1O1N2O2M2O1O1N2O2M2O1O1O2N1O1O1O4L000000000000000000000000000000000000000001O000000010O00000000001O000O100O101N10000O101N100O10000O2O0O100M3N2M4L300O2O1O1N2O1kMcE^1^:^NfEa1[:[NiEc1Y:YNkEf1U:WNoEh1i:O0O2O1O1N2O1O1N2O0O2N2N2N2N2N2N1O3N2M4K4J7I6K5Jj\\R1"}, "label": "a cream color cup near a laptop"}]} {"id": 357470, "image": "COCO_train2014_000000357470.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the coffee mug next to the macbook\"."}], "task": "refering", "answer_template": "The \"the coffee mug next to the macbook\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 201, 202, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 268, 269, 270, 271, 272, 292, 293, 294], "bbox": [0.680015625, 0.5723167848699764, 0.8682187499999999, 0.8546099290780141], "iscrowd": 0, "area": 10962.8146, "rle": {"size": [423, 640], "counts": "Pnc51V=4cD6T9NUFg0j9]O_EW1`:j0M2O1N2O2N1N2O1N3N1O1N2O2M2O1O1N2O2M2O1O1O2N1O1O1O4L000000000000000000000000000000000000000001O000000010O00000000001O000O100O101N10000O101N100O10000O2O0O100M3N2M4L300O2O1O1N2O1kMcE^1^:^NfEa1[:[NiEc1Y:YNkEf1U:WNoEh1i:O0O2O1O1N2O1O1N2O0O2N2N2N2N2N2N1O3N2M4K4J7I6K5Jj\\R1"}, "label": "the coffee mug next to the macbook"}]} {"id": 160866, "image": "COCO_train2014_000000160866.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a chair\"."}], "task": "refering", "answer_template": "The \"a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 165, 166, 167, 168, 182, 183, 184, 185, 186, 200, 201, 202, 203, 217, 218, 219, 234, 235], "bbox": [0.7464375000000001, 0.373234375, 0.9740625, 0.600359375], "iscrowd": 0, "area": 9583.006099999999, "rle": {"size": [640, 480], "counts": "`PP72mc04K5L3N3L4L4L4L4M3L4M4L3L4M4K4M3M4K5L4L4K5L3L5L5K5J5L5__OkLl?d31O10O01O10O01O10O01O010O1O010O1O010O1O010O1O010O00100O00100O0010O01O10O01O10O01O010O1O010O1O010O1O100`Mc_O\\1^`0aNh_O[1Y`0cNi_O]1W`0`Nl_O_1U`0_Nl_Oa1U`0^Nl_Ob1T`0\\Nn_Oc1T`0[Nm_Od1U`0ZNl_Of1U`0XNk_Oh1W`0VNj_Oi1X`0TNj_Ol1[`0nMf_OQ2Ua0O101N1O100O2N100O4M4K6J5L5J6J6K4K;EoU7"}, "label": "a chair"}]} {"id": 160866, "image": "COCO_train2014_000000160866.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty chair at a desk\"."}], "task": "refering", "answer_template": "The \"an empty chair at a desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 165, 166, 167, 168, 182, 183, 184, 185, 186, 200, 201, 202, 203, 217, 218, 219, 234, 235], "bbox": [0.7464375000000001, 0.373234375, 0.9740625, 0.600359375], "iscrowd": 0, "area": 9583.006099999999, "rle": {"size": [640, 480], "counts": "`PP72mc04K5L3N3L4L4L4L4M3L4M4L3L4M4K4M3M4K5L4L4K5L3L5L5K5J5L5__OkLl?d31O10O01O10O01O10O01O010O1O010O1O010O1O010O1O010O00100O00100O0010O01O10O01O10O01O010O1O010O1O010O1O100`Mc_O\\1^`0aNh_O[1Y`0cNi_O]1W`0`Nl_O_1U`0_Nl_Oa1U`0^Nl_Ob1T`0\\Nn_Oc1T`0[Nm_Od1U`0ZNl_Of1U`0XNk_Oh1W`0VNj_Oi1X`0TNj_Ol1[`0nMf_OQ2Ua0O101N1O100O2N100O4M4K6J5L5J6J6K4K;EoU7"}, "label": "an empty chair at a desk"}]} {"id": 160866, "image": "COCO_train2014_000000160866.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a recliner on which ther is an orange cushion\"."}], "task": "refering", "answer_template": "The \"a recliner on which ther is an orange cushion\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 164, 165, 166, 180, 181, 182, 197, 198, 199, 200, 215, 216, 217], "bbox": [0.6110833333333333, 0.2859375, 1.0, 0.547015625], "iscrowd": 0, "area": 13609.417849999998, "rle": {"size": [640, 480], "counts": "c]g51kc04L5J5L5M200O2Oe0ZO2O0O100O2O0O1O1O1B?B=G900O2O0O100O101N10000O2O0O100O101N10000O2O0O100O101N100O101O0O100O101N100O101O0O100O101N100OjMk_Oc0U`0WOU@f0i?ZO[@c0a?@c@=Y?Gk@5Q?ORANj>5ZAIa>;cAAX>c0lAZOP>j0SBSOi=Q1[BlNa=V1cBgN]=Y1gBcNX=]1lB`NT=`1oB]NP=d1TCYNkB`J"}, "label": "a recliner on which ther is an orange cushion"}]} {"id": 160866, "image": "COCO_train2014_000000160866.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sole sofa chair\"."}], "task": "refering", "answer_template": "The \"the sole sofa chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 117, 118, 131, 132, 133, 134, 135, 147, 148, 149, 150, 151, 152, 164, 165, 166, 180, 181, 182, 197, 198, 199, 200, 215, 216, 217], "bbox": [0.6110833333333333, 0.2859375, 1.0, 0.547015625], "iscrowd": 0, "area": 13609.417849999998, "rle": {"size": [640, 480], "counts": "c]g51kc04L5J5L5M200O2Oe0ZO2O0O100O2O0O1O1O1B?B=G900O2O0O100O101N10000O2O0O100O101N10000O2O0O100O101N100O101O0O100O101N100O101O0O100O101N100OjMk_Oc0U`0WOU@f0i?ZO[@c0a?@c@=Y?Gk@5Q?ORANj>5ZAIa>;cAAX>c0lAZOP>j0SBSOi=Q1[BlNa=V1cBgN]=Y1gBcNX=]1lB`NT=`1oB]NP=d1TCYNkB`J"}, "label": "the sole sofa chair"}]} {"id": 70755, "image": "COCO_train2014_000000070755.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue toothbrush on a sink\"."}], "task": "refering", "answer_template": "The \"a blue toothbrush on a sink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 45, 67, 68, 89, 90, 112, 113, 134, 135], "bbox": [0.852171875, 0.0109375, 1.0, 0.3592708333333333], "iscrowd": 0, "area": 2845.089150000003, "rle": {"size": [480, 640], "counts": "Ydo71m>2N2N3M2N2N3M2N2N1O1O1O001O1O001O1O1O001O1O001O1O001O1O001O1O001O10O0100O010O10O01000O0100O010000O10000O1O100O100O1O2OO0O2O000O100N3M2O1N2O1O1O100O001001N2O1O0O2N2N2K5E;M3O001O1O1O1O1O2N1K1"}, "label": "a blue toothbrush on a sink"}]} {"id": 70755, "image": "COCO_train2014_000000070755.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue tooth brush on right\"."}], "task": "refering", "answer_template": "The \"blue tooth brush on right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [22, 45, 67, 68, 89, 90, 112, 113, 134, 135], "bbox": [0.852171875, 0.0109375, 1.0, 0.3592708333333333], "iscrowd": 0, "area": 2845.089150000003, "rle": {"size": [480, 640], "counts": "Ydo71m>2N2N3M2N2N3M2N2N1O1O1O001O1O001O1O1O001O1O001O1O001O1O001O1O001O10O0100O010O10O01000O0100O010000O10000O1O100O100O1O2OO0O2O000O100N3M2O1N2O1O1O100O001001N2O1O0O2N2N2K5E;M3O001O1O1O1O1O2N1K1"}, "label": "blue tooth brush on right"}]} {"id": 70755, "image": "COCO_train2014_000000070755.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the clear container on the right\"."}], "task": "refering", "answer_template": "The \"the clear container on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 68, 91, 114, 136, 137, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 251, 252, 275], "bbox": [0.885375, 0.04052083333333333, 1.0, 0.7057916666666666], "iscrowd": 0, "area": 11044.7844, "rle": {"size": [480, 640], "counts": "VQZ84i>5K4K5L4K5L4K5L4K5L4K5L4L5J5L4K5L4K5L4K5N2N2N2M3N2M4K4L4L4L4M3L4L4L4L4L4L4M3L5K4L4L4L4L4L4M3L4L4L4L4L5K4M3L4L4L4L4L4L4L4PO_H]Jd7_5`H]Jc7_5aH^Jc7^5aH^Jc7]5bH_Ja7^5bH_Jb7\\5cH`J`7]5dH`J_7[5V1L3L5m4RK]J"}, "label": "the clear container on the right"}]} {"id": 70755, "image": "COCO_train2014_000000070755.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty bottle off the screen\"."}], "task": "refering", "answer_template": "The \"an empty bottle off the screen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [45, 68, 91, 114, 136, 137, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 251, 252, 275], "bbox": [0.885375, 0.04052083333333333, 1.0, 0.7057916666666666], "iscrowd": 0, "area": 11044.7844, "rle": {"size": [480, 640], "counts": "VQZ84i>5K4K5L4K5L4K5L4K5L4K5L4L5J5L4K5L4K5L4K5N2N2N2M3N2M4K4L4L4L4M3L4L4L4L4L4L4M3L5K4L4L4L4L4L4M3L4L4L4L4L5K4M3L4L4L4L4L4L4L4PO_H]Jd7_5`H]Jc7_5aH^Jc7^5aH^Jc7]5bH_Ja7^5bH_Jb7\\5cH`J`7]5dH`J_7[5V1L3L5m4RK]J"}, "label": "an empty bottle off the screen"}]} {"id": 513124, "image": "COCO_train2014_000000513124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pitcher who just threw a baseball\"."}], "task": "refering", "answer_template": "The \"pitcher who just threw a baseball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 58, 59, 60, 61, 63, 64, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 116, 117, 118, 119, 120, 134, 135, 136, 137, 138, 152, 153, 154, 155, 171, 172, 173, 188, 189, 190, 191, 206, 207, 209, 210, 224, 225], "bbox": [0.23932, 0.17528000000000002, 0.68258, 0.9865066666666666], "iscrowd": 0, "area": 22402.133049999997, "rle": {"size": [375, 500], "counts": "iP\\12c;3N3M2N2M3N3M2N2M4M2N2N3ON101N2O001N2O001N2O1N101O1O01000000O01000O10000O1000O10O10000O100O010O00100O010O1O010O10O01O10O0100O0010O0100O010O10O0100O010O10O010O0100O01>A0100O001O001O0010O01O001O010OXLkNnLU1Q3lNPMS1o2nNQMR1n2oNRMR1k2POUMP1j2QOWMn0h2oN`IOh3R1g2oNaI0h3Q1e2QOcINh3R1c2POeIOi3P1a2ROfINi3P1`2ROhINh3P1^2SOjINh3o0]2TOkIMi3n0[2UOlINi3n0X2UOPJMh3n0W2VOPJMi3o0T2TOTJMi3T1l1PO[JLi3Z1d1kNcJLi3_1iLZNm3;bKLh3Z2`0lMgKKj3\\26mMQLGi3a2MkMZLEi3d2DlMcL_Oj3i2ZOlMmL[Oj3l2POnMUMWOk3T3bNiMdMSOj3^3SNcMSNPOj3f3eM_MaNkNj3o3WMZMPOgNj3W4hLVM^OdNj3_4ZLRML_Nj3f6VL[Ij3e6VL[IY3CRMP7E^IX3CSMm6G`IU3ESMj6IaIS3FTMg6JdIR3FTMc6LgIo2GUMa6LYI^38eL]6i3eIVLY6k3hIVLV6k3i01O1O101N1O101N1O2O1N2O10O1000000O10000000O01000000O11O001O001O001O001O0O2O3N101O0mIYKS5e52O1O0O2O1N101O0O01O0010O01dMdJ[N]5a1mJXNS5e1RKYNo4d1UKZNk4c1[KZNf4b1_K\\Nb4`1cK_N\\4_1hK_NY4]1mK`NS4]1RLaNo3\\1ULbNn3X1WLfNm3S1YLjNl3o0XLoNl3k0YLROk3i0XLVOl3e0VLYOn3c0ULZOo3a0TL]OP4?SL^OR4=PLAZ45iKHg4G\\K6U8LRmi1"}, "label": "pitcher who just threw a baseball"}]} {"id": 513124, "image": "COCO_train2014_000000513124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player pitching the ball\"."}], "task": "refering", "answer_template": "The \"a baseball player pitching the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 58, 59, 60, 61, 63, 64, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 116, 117, 118, 119, 120, 134, 135, 136, 137, 138, 152, 153, 154, 155, 171, 172, 173, 188, 189, 190, 191, 206, 207, 209, 210, 224, 225], "bbox": [0.23932, 0.17528000000000002, 0.68258, 0.9865066666666666], "iscrowd": 0, "area": 22402.133049999997, "rle": {"size": [375, 500], "counts": "iP\\12c;3N3M2N2M3N3M2N2M4M2N2N3ON101N2O001N2O001N2O1N101O1O01000000O01000O10000O1000O10O10000O100O010O00100O010O1O010O10O01O10O0100O0010O0100O010O10O0100O010O10O010O0100O01>A0100O001O001O0010O01O001O010OXLkNnLU1Q3lNPMS1o2nNQMR1n2oNRMR1k2POUMP1j2QOWMn0h2oN`IOh3R1g2oNaI0h3Q1e2QOcINh3R1c2POeIOi3P1a2ROfINi3P1`2ROhINh3P1^2SOjINh3o0]2TOkIMi3n0[2UOlINi3n0X2UOPJMh3n0W2VOPJMi3o0T2TOTJMi3T1l1PO[JLi3Z1d1kNcJLi3_1iLZNm3;bKLh3Z2`0lMgKKj3\\26mMQLGi3a2MkMZLEi3d2DlMcL_Oj3i2ZOlMmL[Oj3l2POnMUMWOk3T3bNiMdMSOj3^3SNcMSNPOj3f3eM_MaNkNj3o3WMZMPOgNj3W4hLVM^OdNj3_4ZLRML_Nj3f6VL[Ij3e6VL[IY3CRMP7E^IX3CSMm6G`IU3ESMj6IaIS3FTMg6JdIR3FTMc6LgIo2GUMa6LYI^38eL]6i3eIVLY6k3hIVLV6k3i01O1O101N1O101N1O2O1N2O10O1000000O10000000O01000000O11O001O001O001O001O0O2O3N101O0mIYKS5e52O1O0O2O1N101O0O01O0010O01dMdJ[N]5a1mJXNS5e1RKYNo4d1UKZNk4c1[KZNf4b1_K\\Nb4`1cK_N\\4_1hK_NY4]1mK`NS4]1RLaNo3\\1ULbNn3X1WLfNm3S1YLjNl3o0XLoNl3k0YLROk3i0XLVOl3e0VLYOn3c0ULZOo3a0TL]OP4?SL^OR4=PLAZ45iKHg4G\\K6U8LRmi1"}, "label": "a baseball player pitching the ball"}]} {"id": 513124, "image": "COCO_train2014_000000513124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"player bent over\"."}], "task": "refering", "answer_template": "The \"player bent over\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 68, 69, 70, 86, 87, 88, 104, 105, 106, 121, 122, 124], "bbox": [0.75506, 0.21509333333333333, 0.9493199999999999, 0.5484266666666666], "iscrowd": 0, "area": 4612.699650000002, "rle": {"size": [375, 500], "counts": "jcZ43b;2M4M3M3M3L3N3M001N2O1O1O0000000001O0000oNYOTGg0j8@RG?m8FoF:Q9JkF7V9LeF5[9NaF2a91[FOe9Q11O1N101O0O200O0100N2E:H900O2O1N101O1O2O1N3N0O2O001NO4L9H3K3N2N1O2M3N2N2M3N2O10O101O01O8I6I5K4L3N2N10O2O001N101N01M3N1O2M3N2M4L3M3L4M4L^Z9"}, "label": "player bent over"}]} {"id": 513124, "image": "COCO_train2014_000000513124.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player stands in the outfield\"."}], "task": "refering", "answer_template": "The \"a baseball player stands in the outfield\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 68, 69, 70, 86, 87, 88, 104, 105, 106, 121, 122, 124], "bbox": [0.75506, 0.21509333333333333, 0.9493199999999999, 0.5484266666666666], "iscrowd": 0, "area": 4612.699650000002, "rle": {"size": [375, 500], "counts": "jcZ43b;2M4M3M3M3L3N3M001N2O1O1O0000000001O0000oNYOTGg0j8@RG?m8FoF:Q9JkF7V9LeF5[9NaF2a91[FOe9Q11O1N101O0O200O0100N2E:H900O2O1N101O1O2O1N3N0O2O001NO4L9H3K3N2N1O2M3N2N2M3N2O10O101O01O8I6I5K4L3N2N10O2O001N101N01M3N1O2M3N2M4L3M3L4M4L^Z9"}, "label": "a baseball player stands in the outfield"}]} {"id": 521338, "image": "COCO_train2014_000000521338.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bear viewed in profile , sitting by the handle of wooden cart\"."}], "task": "refering", "answer_template": "The \"a bear viewed in profile , sitting by the handle of wooden cart\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 152, 153, 154, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 249, 268, 270, 271, 272, 293, 294, 295, 296, 317], "bbox": [0.575734375, 0.3910070257611241, 0.8830937500000001, 0.8876346604215456], "iscrowd": 0, "area": 17807.091649999995, "rle": {"size": [427, 640], "counts": "kfi42W= in this image.", "objects": [{"patches": [130, 152, 153, 154, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 249, 268, 270, 271, 272, 293, 294, 295, 296, 317], "bbox": [0.575734375, 0.3910070257611241, 0.8830937500000001, 0.8876346604215456], "iscrowd": 0, "area": 17807.091649999995, "rle": {"size": [427, 640], "counts": "kfi42W= in this image.", "objects": [{"patches": [11, 32, 33, 34, 35, 54, 55, 56, 57, 58, 78, 79, 80, 81, 101, 102, 103, 104, 105, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173], "bbox": [0.37043750000000003, 0.05782201405152225, 0.648859375, 0.5410070257611241], "iscrowd": 0, "area": 19917.946849999997, "rle": {"size": [427, 640], "counts": "^lR36T=5K4M1N1bGKe37ULOi33QL2o3OlK5S4LjK7U4KgK8X4JdK9\\4G`K=_4E]K=c4EYK=g4DVK?j4ARKb0n4_OnJd0R5\\OlJf0U5\\OeJg0\\5\\O^Je0c5^OXJd0i5_OQJc0o5AkIa0V6AeI`0\\6C_I?b6CYI?h6BTI`0l6CoH>S7CjH>V7EfH;\\7FaH:`7H^H8c7H\\H7f7JXH5i7MUH2m7OQH0P81PHMR84lGLU85iGIY89eGF]8;bGB`8`0^G^Oe8c0YG]Og8f0VG^Og8f2M3L3N3L4M3L4M2N2M3N2M3N1N3N2N2M3N2M2O1O0O101N10001O00O1O010O1O100O00100O1O100O00100O1O10O01O100O1O100O2N2O1N2N2O0UK\\Hn3f7PL^Hl3d7QL`Hm3a7QLcHk3_7RLfHj3[7ULiHh3X7ULmHa0UOX2P8oLVIb0nN]2\\9\\MgFd2]9UMgFi2^9PMfFn2m9O1N2N2O1N2N2O0O2N2O1N2N3N1N2N2O1N2N2O1^OmDiNV;V1`000001O00000001O0000000O1000001N0100000001O0O2O001OO1O1N101O1O1N2O1N2N2M3N2M3N2M3N2M3L4L4KQem2"}, "label": "teddy bear that you can see both eyes"}]} {"id": 521338, "image": "COCO_train2014_000000521338.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bear with a button on its chest\"."}], "task": "refering", "answer_template": "The \"a bear with a button on its chest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 32, 33, 34, 35, 54, 55, 56, 57, 58, 78, 79, 80, 81, 101, 102, 103, 104, 105, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173], "bbox": [0.37043750000000003, 0.05782201405152225, 0.648859375, 0.5410070257611241], "iscrowd": 0, "area": 19917.946849999997, "rle": {"size": [427, 640], "counts": "^lR36T=5K4M1N1bGKe37ULOi33QL2o3OlK5S4LjK7U4KgK8X4JdK9\\4G`K=_4E]K=c4EYK=g4DVK?j4ARKb0n4_OnJd0R5\\OlJf0U5\\OeJg0\\5\\O^Je0c5^OXJd0i5_OQJc0o5AkIa0V6AeI`0\\6C_I?b6CYI?h6BTI`0l6CoH>S7CjH>V7EfH;\\7FaH:`7H^H8c7H\\H7f7JXH5i7MUH2m7OQH0P81PHMR84lGLU85iGIY89eGF]8;bGB`8`0^G^Oe8c0YG]Og8f0VG^Og8f2M3L3N3L4M3L4M2N2M3N2M3N1N3N2N2M3N2M2O1O0O101N10001O00O1O010O1O100O00100O1O100O00100O1O10O01O100O1O100O2N2O1N2N2O0UK\\Hn3f7PL^Hl3d7QL`Hm3a7QLcHk3_7RLfHj3[7ULiHh3X7ULmHa0UOX2P8oLVIb0nN]2\\9\\MgFd2]9UMgFi2^9PMfFn2m9O1N2N2O1N2N2O0O2N2O1N2N3N1N2N2O1N2N2O1^OmDiNV;V1`000001O00000001O0000000O1000001N0100000001O0O2O001OO1O1N101O1O1N2O1N2N2M3N2M3N2M3N2M3L4L4KQem2"}, "label": "a bear with a button on its chest"}]} {"id": 103556, "image": "COCO_train2014_000000103556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white tee shirt and black shorts\"."}], "task": "refering", "answer_template": "The \"a man in a white tee shirt and black shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 118, 119, 120, 141, 142, 143, 164, 165, 166, 187, 188, 189, 210, 211, 212, 234, 235, 257, 258], "bbox": [0.134078125, 0.2833489461358314, 0.24567187500000004, 0.7847306791569086], "iscrowd": 0, "area": 9156.382300000001, "rle": {"size": [427, 640], "counts": "PRT11X=2N2N3M5L2M3M3M4K5K5K4cNZOoEl0m9YOgER1S:UOeER1d0RNi8T3RGTMh8o2PGZMl8e3M4K4M3M3N2N2O1N1O2N2O2N2dHQKX6o4cIWK[6[5nHlJP7h51N10001O0O1O101N3M2N2N2N2N2N1O1O001O1O001O1O1O1O1O1O1[MXHFh76_HFd74bHIa72dHLn7_OWH>n7[OXH`0o7XOVHd0P8UOUHg0Q8ROTHE]Ni0f9YOcH?c7YOfH?[:HVSY6"}, "label": "a man in a white tee shirt and black shorts"}]} {"id": 103556, "image": "COCO_train2014_000000103556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in white in the background\"."}], "task": "refering", "answer_template": "The \"a man in white in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 118, 119, 120, 141, 142, 143, 164, 165, 166, 187, 188, 189, 210, 211, 212, 234, 235, 257, 258], "bbox": [0.134078125, 0.2833489461358314, 0.24567187500000004, 0.7847306791569086], "iscrowd": 0, "area": 9156.382300000001, "rle": {"size": [427, 640], "counts": "PRT11X=2N2N3M5L2M3M3M4K5K5K4cNZOoEl0m9YOgER1S:UOeER1d0RNi8T3RGTMh8o2PGZMl8e3M4K4M3M3N2N2O1N1O2N2O2N2dHQKX6o4cIWK[6[5nHlJP7h51N10001O0O1O101N3M2N2N2N2N2N1O1O001O1O001O1O1O1O1O1O1[MXHFh76_HFd74bHIa72dHLn7_OWH>n7[OXH`0o7XOVHd0P8UOUHg0Q8ROTHE]Ni0f9YOcH?c7YOfH?[:HVSY6"}, "label": "a man in white in the background"}]} {"id": 103556, "image": "COCO_train2014_000000103556.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in a black sports top with a tennis racquet in her right hand\"."}], "task": "refering", "answer_template": "The \"a girl in a black sports top with a tennis racquet in her right hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 26, 27, 49, 50, 51, 73, 74, 75, 97, 98, 120, 121, 122, 144, 145, 146, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 197, 201, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 238, 239, 240, 241, 242, 243, 244, 245, 246, 260, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 306, 307, 308, 309, 310, 312, 313, 330, 331, 332, 333, 335, 336], "bbox": [0.125390625, 0.025550351288056206, 0.7664375, 0.9985714285714287], "iscrowd": 0, "area": 45961.744550000025, "rle": {"size": [427, 640], "counts": "ljQ13V=2N3N11O01O00010O000010O00010O01O1bNIfE7Y:NdE2Z:2cEO[:6aEJ^:;^EF`:?]EA`:e0\\E[Oc:j0YEWOf:l0WEUOi:n0RESOn:Q1mDQOS;c11000O01O1O1O3M3M2O2M3MI_NfD_1Z;gNcDV1\\;=1O1O4L3M3N2O1O1O1O1O1O1N2O1O1O00O10000O10O0100O10O01O1O1O10O01O1O10O01O1O100O001O100O00100O1O1O100O001O100O1O10O0001O0001O000001O0001O0001O1O010O001O1O010O001O1O0100O010O1000O0100O10O10O10O0fG\\Nm4d1mJPOe4o0VK>c3AYLg0c3YOXLP1e3POVLX1f3hNXL_1c3aN[Lb1d3^NYLd1h3\\NVLe1l3[NQLf1P4ZNnKg1S4YNkKh1V4XNhKi1Y4XNcKj1^4VN`Kk1b4TN\\Km1e4SNYKn1h4RNVKo1k4VNnJk1S5eNZJ]1g5ROhIo0Z6P31O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O1O1O1OO1K5O1N2O1O1N2O1O1O100E;00FSIUJn6k5TISJl6l5XIQJh6o5:O1000000O10000O10000O10000O10000O100O1O100O1O1O100O1O1O100O1O1O1O10^OfJSIY5d6RK[Im4`6ZK_Ie4a6]K^Ic4`6`K_I_4a6dK]I[4b6hK]IW4c6kK\\IU4c6mK\\IT4a6oK^IQ4a6^1N1N2O2M2O2N1N2O2N1N3N1N2O2M2O2N1N3N1N2O2N1N3N1N2O2002M3N2M3N3M2M3N2M3N2M4M2N2M3N2M3N3L3N2N2M01O001O001O1O0_JbIT4`6`KmI^4S6XKYJf4h5oJdJo4n6N3M2N2N2N2M3N5I7J6I7J6I8I6I7J6I7J2M101J5QOo00000000001O01O0000000001O000001O00000001O00000000O010000000000O10000000O0100000000O10000000O01000000O101O001O1N101O1O001N2O000010O1E;000O10001O0_OmCJS<5oCJQ<5QDJo;5RDKn;3UDKl;4VDKj;3YDLg;1]DNc;OaD0_;MeD2\\ in this image.", "objects": [{"patches": [3, 4, 26, 27, 49, 50, 51, 73, 74, 75, 97, 98, 120, 121, 122, 144, 145, 146, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 197, 201, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 238, 239, 240, 241, 242, 243, 244, 245, 246, 260, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 306, 307, 308, 309, 310, 312, 313, 330, 331, 332, 333, 335, 336], "bbox": [0.125390625, 0.025550351288056206, 0.7664375, 0.9985714285714287], "iscrowd": 0, "area": 45961.744550000025, "rle": {"size": [427, 640], "counts": "ljQ13V=2N3N11O01O00010O000010O00010O01O1bNIfE7Y:NdE2Z:2cEO[:6aEJ^:;^EF`:?]EA`:e0\\E[Oc:j0YEWOf:l0WEUOi:n0RESOn:Q1mDQOS;c11000O01O1O1O3M3M2O2M3MI_NfD_1Z;gNcDV1\\;=1O1O4L3M3N2O1O1O1O1O1O1N2O1O1O00O10000O10O0100O10O01O1O1O10O01O1O10O01O1O100O001O100O00100O1O1O100O001O100O1O10O0001O0001O000001O0001O0001O1O010O001O1O010O001O1O0100O010O1000O0100O10O10O10O0fG\\Nm4d1mJPOe4o0VK>c3AYLg0c3YOXLP1e3POVLX1f3hNXL_1c3aN[Lb1d3^NYLd1h3\\NVLe1l3[NQLf1P4ZNnKg1S4YNkKh1V4XNhKi1Y4XNcKj1^4VN`Kk1b4TN\\Km1e4SNYKn1h4RNVKo1k4VNnJk1S5eNZJ]1g5ROhIo0Z6P31O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O1O1O1OO1K5O1N2O1O1N2O1O1O100E;00FSIUJn6k5TISJl6l5XIQJh6o5:O1000000O10000O10000O10000O10000O100O1O100O1O1O100O1O1O100O1O1O1O10^OfJSIY5d6RK[Im4`6ZK_Ie4a6]K^Ic4`6`K_I_4a6dK]I[4b6hK]IW4c6kK\\IU4c6mK\\IT4a6oK^IQ4a6^1N1N2O2M2O2N1N2O2N1N3N1N2O2M2O2N1N3N1N2O2N1N3N1N2O2002M3N2M3N3M2M3N2M3N2M4M2N2M3N2M3N3L3N2N2M01O001O001O1O0_JbIT4`6`KmI^4S6XKYJf4h5oJdJo4n6N3M2N2N2N2M3N5I7J6I7J6I8I6I7J6I7J2M101J5QOo00000000001O01O0000000001O000001O00000001O00000000O010000000000O10000000O0100000000O10000000O01000000O101O001O1N101O1O001N2O000010O1E;000O10001O0_OmCJS<5oCJQ<5QDJo;5RDKn;3UDKl;4VDKj;3YDLg;1]DNc;OaD0_;MeD2\\ in this image.", "objects": [{"patches": [98, 99, 121, 122, 123, 144, 145, 146, 167, 168, 169, 170, 191, 192, 193, 214, 215, 216, 217, 238, 239, 240, 262, 263, 285, 286], "bbox": [0.266296875, 0.2468125, 0.4449375, 0.7479375], "iscrowd": 0, "area": 11524.088149999998, "rle": {"size": [480, 640], "counts": "]j_2?[>9K4K5L4L3M3M3N100O2N1O20O0001O000010O02N2nBkNU in this image.", "objects": [{"patches": [98, 99, 121, 122, 123, 144, 145, 146, 167, 168, 169, 170, 191, 192, 193, 214, 215, 216, 217, 238, 239, 240, 262, 263, 285, 286], "bbox": [0.266296875, 0.2468125, 0.4449375, 0.7479375], "iscrowd": 0, "area": 11524.088149999998, "rle": {"size": [480, 640], "counts": "]j_2?[>9K4K5L4L3M3M3N100O2N1O20O0001O000010O02N2nBkNU in this image.", "objects": [{"patches": [57, 58, 59, 80, 81, 82, 103, 104, 105, 106, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 314, 315, 333, 334, 335, 336, 337, 338, 357, 358, 359, 360, 361, 381, 382, 383, 384], "bbox": [0.439890625, 0.13670833333333335, 0.714609375, 1.0], "iscrowd": 0, "area": 46116.81260000001, "rle": {"size": [480, 640], "counts": "P_T4j0U>1000O1000O100000000O1000000O101O00001N100QDROl9o0mEXOR:h0hE@V:a0cEF[:;_ELW5VOfNn0mK4X5TOhNi0iK:\\5ROhNd0gK`0^5QOhN?dKh0`5nNhN in this image.", "objects": [{"patches": [57, 58, 59, 80, 81, 82, 103, 104, 105, 106, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 172, 173, 174, 175, 176, 195, 196, 197, 198, 199, 218, 219, 220, 221, 222, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 314, 315, 333, 334, 335, 336, 337, 338, 357, 358, 359, 360, 361, 381, 382, 383, 384], "bbox": [0.439890625, 0.13670833333333335, 0.714609375, 1.0], "iscrowd": 0, "area": 46116.81260000001, "rle": {"size": [480, 640], "counts": "P_T4j0U>1000O1000O100000000O1000000O101O00001N100QDROl9o0mEXOR:h0hE@V:a0cEF[:;_ELW5VOfNn0mK4X5TOhNi0iK:\\5ROhNd0gK`0^5QOhN?dKh0`5nNhN in this image.", "objects": [{"patches": [26, 27, 47, 48, 49, 50, 68, 69, 70, 71, 89, 90, 91, 92, 93, 111, 112, 113, 114, 115, 133, 134, 135, 136, 137, 138, 155, 156, 157, 158, 159, 160, 161, 177, 178, 179, 180, 181, 182, 183, 184, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 249, 250, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 331, 332, 333, 334, 335, 353, 354, 355, 356, 357, 375, 376, 377, 378, 379, 398, 399, 400, 401, 420, 421, 422, 423, 441, 442, 443, 444, 445, 446, 463, 464, 465, 466, 467], "bbox": [0.0476158940397351, 0.0561875, 0.40955298013245034, 0.9460625], "iscrowd": 0, "area": 67717.40995000002, "rle": {"size": [640, 604], "counts": "i\\b0i0Vc0?A4iNkN\\_OY1[`0oNb_OU1U`0ROi_OR1o?UOm_OP1j?XOS@l0d?[OZ@i0]?_O`@e0W?Ah@c0o6dMf0P2YH`0c1dMZ3Nc2T2_H>V1QNa3_Oc2X2eHEg2f5jLfJ8Km2`5jLhJM1Y3W5jLcNV3]1iLdNV3o0WL[Hb0f6X3o0WLZH`0h6X3o0YLWH?j6X3o0\\LUH;m6Y3n0]LTH9o6Y3m0`LSH6R7Y3k0cLPH4W7W3j0fLnG2Y7W3j0hLlG0\\7W3h0VMYOi2h0VMZOi2f0VM\\Oh2e0WM]Og2d0WM_Oh2a0UMCj2>QMGn29nLLQ34lL0S30iL5V3\\8O00000000001O000000000000001O000000000O101O0O01000O10O10O1000O010000O10O10O1N2M2O3L3M3M3N3L3M3M4M2M3M4J7I7J5L5K5K4L5J`kn6"}, "label": "the first man in line"}]} {"id": 316557, "image": "COCO_train2014_000000316557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a white coat to the left of some other people\"."}], "task": "refering", "answer_template": "The \"a person in a white coat to the left of some other people\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 47, 48, 49, 50, 68, 69, 70, 71, 89, 90, 91, 92, 93, 111, 112, 113, 114, 115, 133, 134, 135, 136, 137, 138, 155, 156, 157, 158, 159, 160, 161, 177, 178, 179, 180, 181, 182, 183, 184, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 249, 250, 265, 266, 267, 268, 269, 287, 288, 289, 290, 291, 309, 310, 311, 312, 313, 331, 332, 333, 334, 335, 353, 354, 355, 356, 357, 375, 376, 377, 378, 379, 398, 399, 400, 401, 420, 421, 422, 423, 441, 442, 443, 444, 445, 446, 463, 464, 465, 466, 467], "bbox": [0.0476158940397351, 0.0561875, 0.40955298013245034, 0.9460625], "iscrowd": 0, "area": 67717.40995000002, "rle": {"size": [640, 604], "counts": "i\\b0i0Vc0?A4iNkN\\_OY1[`0oNb_OU1U`0ROi_OR1o?UOm_OP1j?XOS@l0d?[OZ@i0]?_O`@e0W?Ah@c0o6dMf0P2YH`0c1dMZ3Nc2T2_H>V1QNa3_Oc2X2eHEg2f5jLfJ8Km2`5jLhJM1Y3W5jLcNV3]1iLdNV3o0WL[Hb0f6X3o0WLZH`0h6X3o0YLWH?j6X3o0\\LUH;m6Y3n0]LTH9o6Y3m0`LSH6R7Y3k0cLPH4W7W3j0fLnG2Y7W3j0hLlG0\\7W3h0VMYOi2h0VMZOi2f0VM\\Oh2e0WM]Og2d0WM_Oh2a0UMCj2>QMGn29nLLQ34lL0S30iL5V3\\8O00000000001O000000000000001O000000000O101O0O01000O10O10O1000O010000O10O10O1N2M2O3L3M3M3N3L3M3M4M2M3M4J7I7J5L5K5K4L5J`kn6"}, "label": "a person in a white coat to the left of some other people"}]} {"id": 316557, "image": "COCO_train2014_000000316557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with glasses staring down at an object in his hands while wearing skis\"."}], "task": "refering", "answer_template": "The \"a man with glasses staring down at an object in his hands while wearing skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 98, 99, 100, 101, 119, 120, 121, 122, 141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 188, 207, 208, 209, 210, 229, 230, 231, 232, 252, 253, 254, 274, 275, 276, 296, 297, 298, 318, 319, 320, 340, 341, 342, 362, 363, 364, 365, 384, 385, 386, 387], "bbox": [0.4266390728476821, 0.13082812500000002, 0.6144370860927152, 0.77934375], "iscrowd": 0, "area": 25549.42530000001, "rle": {"size": [640, 604], "counts": "k\\Q5>ac09G5K3M4M2M4L3M5K6\\_OFW=`0eB_OZ=f0cBZO\\=j0aBVO]=l0cBSO\\=n0eBROY=P1gBoNX=S1hBmNV=U1jBjNU=W1lBiNR=Y1nBfNR=[1nBeNP=\\1QCcNn<_1RCcNklIBT6`0jI@U6c0hI^OW6f0fI[OY6g0eIYO[6i0cIWO]6k0`IVO`6P1ZIQOe6[1oHeNQ7U701OXEmHc8S7XGSIg8m6XGVIf8j6XGYIh8g6WG[Ig8e6YG\\If8d6YG_Ie8b6ZG_If8`6ZGbIe6]O^IQ7MdIg5CPK0c4;?CPKIi4c0;AnJIi4d0?^OiJLi4e0a;M2O2N1O2N1N3L3M4MZUa4"}, "label": "a man with glasses staring down at an object in his hands while wearing skis"}]} {"id": 316557, "image": "COCO_train2014_000000316557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person that is wearing glasses\"."}], "task": "refering", "answer_template": "The \"the person that is wearing glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 98, 99, 100, 101, 119, 120, 121, 122, 141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 188, 207, 208, 209, 210, 229, 230, 231, 232, 252, 253, 254, 274, 275, 276, 296, 297, 298, 318, 319, 320, 340, 341, 342, 362, 363, 364, 365, 384, 385, 386, 387], "bbox": [0.4266390728476821, 0.13082812500000002, 0.6144370860927152, 0.77934375], "iscrowd": 0, "area": 25549.42530000001, "rle": {"size": [640, 604], "counts": "k\\Q5>ac09G5K3M4M2M4L3M5K6\\_OFW=`0eB_OZ=f0cBZO\\=j0aBVO]=l0cBSO\\=n0eBROY=P1gBoNX=S1hBmNV=U1jBjNU=W1lBiNR=Y1nBfNR=[1nBeNP=\\1QCcNn<_1RCcNklIBT6`0jI@U6c0hI^OW6f0fI[OY6g0eIYO[6i0cIWO]6k0`IVO`6P1ZIQOe6[1oHeNQ7U701OXEmHc8S7XGSIg8m6XGVIf8j6XGYIh8g6WG[Ig8e6YG\\If8d6YG_Ie8b6ZG_If8`6ZGbIe6]O^IQ7MdIg5CPK0c4;?CPKIi4c0;AnJIi4d0?^OiJLi4e0a;M2O2N1O2N1N3L3M4MZUa4"}, "label": "the person that is wearing glasses"}]} {"id": 316557, "image": "COCO_train2014_000000316557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person in light colored jacket and wearing a cap at the end of the line\"."}], "task": "refering", "answer_template": "The \"person in light colored jacket and wearing a cap at the end of the line\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 122, 123, 144, 145, 166, 167, 168, 188, 189, 190, 210, 211, 212, 232, 233, 234, 254, 255, 276, 277, 298, 299, 320, 321, 342, 343], "bbox": [0.5500993377483444, 0.2090625, 0.6806953642384106, 0.68053125], "iscrowd": 0, "area": 7038.086099999999, "rle": {"size": [640, 604], "counts": "dd_69gc01N2O1O1N2LAe\\O`0Zc0@f\\Ob0Wc0_Oh\\Oc0Wc0]Oi\\Oe0Tc0\\Ol\\Of0Rc05N3N2K4M4M3N4L3Q^ObNVa0b1\\^OiNca0j1o^OgM[`0Y2e_OPNU`0n1j_O[NR`0`1n_OY4^:bKaE^4d:SJlDj0?Q5l:oIgDo0=P5];oJbDP5`;QK^Dn4e;QKZDf4Qi0c0]Oc0\\Od0F:O1O001O1O1O1O1N101O1N2N2N2N1O2N2N2N2N2N3M2N2N3M2M4M2N2N3L3N2N3L3N2N3L3N3K4K5L7H9G9Hh\\h3"}, "label": "person in light colored jacket and wearing a cap at the end of the line"}]} {"id": 316557, "image": "COCO_train2014_000000316557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the farthest person in the picture\"."}], "task": "refering", "answer_template": "The \"the farthest person in the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 122, 123, 144, 145, 166, 167, 168, 188, 189, 190, 210, 211, 212, 232, 233, 234, 254, 255, 276, 277, 298, 299, 320, 321, 342, 343], "bbox": [0.5500993377483444, 0.2090625, 0.6806953642384106, 0.68053125], "iscrowd": 0, "area": 7038.086099999999, "rle": {"size": [640, 604], "counts": "dd_69gc01N2O1O1N2LAe\\O`0Zc0@f\\Ob0Wc0_Oh\\Oc0Wc0]Oi\\Oe0Tc0\\Ol\\Of0Rc05N3N2K4M4M3N4L3Q^ObNVa0b1\\^OiNca0j1o^OgM[`0Y2e_OPNU`0n1j_O[NR`0`1n_OY4^:bKaE^4d:SJlDj0?Q5l:oIgDo0=P5];oJbDP5`;QK^Dn4e;QKZDf4Qi0c0]Oc0\\Od0F:O1O001O1O1O1O1N101O1N2N2N2N1O2N2N2N2N2N3M2N2N3M2M4M2N2N3L3N2N3L3N2N3L3N3K4K5L7H9G9Hh\\h3"}, "label": "the farthest person in the picture"}]} {"id": 316557, "image": "COCO_train2014_000000316557.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the middle of the group of 3\"."}], "task": "refering", "answer_template": "The \"the man in the middle of the group of 3\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 73, 74, 75, 76, 95, 96, 97, 116, 117, 118, 119, 138, 139, 140, 141, 160, 161, 162, 163, 164, 183, 184, 185, 186, 206, 207, 208, 209, 225, 226, 230, 231, 247, 248, 249, 253, 270, 271, 272, 292, 293, 294, 314, 315, 316, 336, 337, 338, 358, 359, 360, 380, 381, 382, 401, 402, 403, 404, 405, 424, 425, 426, 427], "bbox": [0.2616721854304636, 0.0944375, 0.5461423841059603, 0.84765625], "iscrowd": 0, "area": 29434.91809999999, "rle": {"size": [640, 604], "counts": "UQS34lc07_BKi6:nHOm65lH3o60jH4V70cH3\\72\\H2d71VH2j72nG2R82fG2Z82^G2b81WG3i81oF3Q91gF3Z96WFOi9O3M4M0O00000000000000000001O00000000001O001O1O1O001O10O0O2O1O001N2O001O1O0O2N2N2N2N1O2N2N2N2N2N8G in this image.", "objects": [{"patches": [52, 53, 73, 74, 75, 76, 95, 96, 97, 116, 117, 118, 119, 138, 139, 140, 141, 160, 161, 162, 163, 164, 183, 184, 185, 186, 206, 207, 208, 209, 225, 226, 230, 231, 247, 248, 249, 253, 270, 271, 272, 292, 293, 294, 314, 315, 316, 336, 337, 338, 358, 359, 360, 380, 381, 382, 401, 402, 403, 404, 405, 424, 425, 426, 427], "bbox": [0.2616721854304636, 0.0944375, 0.5461423841059603, 0.84765625], "iscrowd": 0, "area": 29434.91809999999, "rle": {"size": [640, 604], "counts": "UQS34lc07_BKi6:nHOm65lH3o60jH4V70cH3\\72\\H2d71VH2j72nG2R82fG2Z82^G2b81WG3i81oF3Q91gF3Z96WFOi9O3M4M0O00000000000000000001O00000000001O001O1O1O001O10O0O2O1O001N2O001O1O0O2N2N2N2N1O2N2N2N2N2N8G in this image.", "objects": [{"patches": [78, 79, 93, 94, 95, 108, 109, 110, 111, 112, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 153, 154, 155, 156, 157, 158, 168, 169, 170, 171, 172, 173, 184, 185, 186, 187, 199, 200, 201, 214, 215], "bbox": [0.21231132075471698, 0.23748437500000003, 0.5508962264150943, 0.652328125], "iscrowd": 0, "area": 21844.36415, "rle": {"size": [640, 424], "counts": "g_h1c0\\c0V1jNS1mN001O0`NWMjAi2U>ZMiAf2U>^MiAc2U>aMhA_2W>dMgA\\2W>hMgAX2X>lMeAU2X>oMfAQ2Y>RNeAn1Z>VNcAj1[>ZNcAg1[>]NbAc1]>`NaA`1]>dNaA\\1^>hN_AY1^>kN`AV1^>mN`AU1]>oN`AS1^>oN`AS1`>W2O100O100O100O100O100O1000003M3M3M3M2N3mBTJl;P7000001O000000001O01O1O2N1O2K5K4L5K4L5K2N2O1O1O1N2O1O2N1O1O1N2O1O1O1O1N3N1O1O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1O2N1N2O1O1N2O1O1N3N2N1N3N2N2M3N2N1O2N2N2N2N2N1O2N4L4L5K5K5K5K5K5K4L5K5L4K5K5L4K4L5L6I7J6I6Bg^f3"}, "label": "net"}]} {"id": 193682, "image": "COCO_train2014_000000193682.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a male spectator in a blue shirt\"."}], "task": "refering", "answer_template": "The \"a male spectator in a blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [162, 163, 177, 178, 191, 192, 193, 206, 207, 208, 209, 220, 221, 222, 223, 224, 235, 236, 237, 238, 239, 251, 252, 253, 254, 266, 267, 268, 269, 281, 282, 283, 284, 296, 297, 298, 299, 311, 312, 313], "bbox": [0.7224999999999999, 0.45325, 1.0, 0.8802187499999998], "iscrowd": 0, "area": 22469.601800000004, "rle": {"size": [640, 424], "counts": "Udo52mc0d0\\Oc0\\Oe0\\Oc0]O6I5L5K4K6K4L4K6K4L5J5L5K4K5L5QBXK]BRAf?_O[@>k?^OW@?n?^OR@`0R`0]Oo_Oa0V`0[Ol_Ob0Y`0[Og_Oc0\\`0ROo_Oj0V`0iNU@Q1aa0C=D in this image.", "objects": [{"patches": [162, 163, 177, 178, 191, 192, 193, 206, 207, 208, 209, 220, 221, 222, 223, 224, 235, 236, 237, 238, 239, 251, 252, 253, 254, 266, 267, 268, 269, 281, 282, 283, 284, 296, 297, 298, 299, 311, 312, 313], "bbox": [0.7224999999999999, 0.45325, 1.0, 0.8802187499999998], "iscrowd": 0, "area": 22469.601800000004, "rle": {"size": [640, 424], "counts": "Udo52mc0d0\\Oc0\\Oe0\\Oc0]O6I5L5K4K6K4L4K6K4L5J5L5K4K5L5QBXK]BRAf?_O[@>k?^OW@?n?^OR@`0R`0]Oo_Oa0V`0[Ol_Ob0Y`0[Og_Oc0\\`0ROo_Oj0V`0iNU@Q1aa0C=D in this image.", "objects": [{"patches": [68, 82, 83, 97, 98, 112, 113, 127, 128, 142, 143, 144, 145, 146, 158, 159, 160, 161, 162, 172, 173, 174, 175, 176, 186, 187, 188, 189, 190, 201, 202, 203, 204, 205, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 244, 245, 246, 247, 248, 259, 260, 262, 263, 274, 275, 277, 278, 289, 290, 292, 293, 304, 305, 307, 308, 319, 320, 322, 323, 334, 335, 337, 338], "bbox": [0.28556603773584904, 0.19820312499999998, 0.8091037735849057, 1.0], "iscrowd": 0, "area": 36630.360250000005, "rle": {"size": [640, 424], "counts": "_S\\2`4\\?4K5L4K5L4L4M32N2N00M3M3M3L4M3M3M3M3L4M3M3M3L4M3M3MNoB[Ini3^ATLg>e3_AWLe>e3^AYLe>c3^AZLd>d3_AYLd>d3_AZLc>c3o0L4M3M3N2O1O100O1O1O1O10OO2N0001O01O000001O01O0001O000O101N10000O2O000O1010O001O000mN]_OkNc`0U1c_O^Nc`0b1n00O010O01O010O2O0O100O100O2O0O100O101N1O1O1O1O4J5K5K6K4K5J7H7HW[b1"}, "label": "the tennis player on the court"}]} {"id": 193682, "image": "COCO_train2014_000000193682.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman hitting a tennis ball with racket\"."}], "task": "refering", "answer_template": "The \"woman hitting a tennis ball with racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 82, 83, 97, 98, 112, 113, 127, 128, 142, 143, 144, 145, 146, 158, 159, 160, 161, 162, 172, 173, 174, 175, 176, 186, 187, 188, 189, 190, 201, 202, 203, 204, 205, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 244, 245, 246, 247, 248, 259, 260, 262, 263, 274, 275, 277, 278, 289, 290, 292, 293, 304, 305, 307, 308, 319, 320, 322, 323, 334, 335, 337, 338], "bbox": [0.28556603773584904, 0.19820312499999998, 0.8091037735849057, 1.0], "iscrowd": 0, "area": 36630.360250000005, "rle": {"size": [640, 424], "counts": "_S\\2`4\\?4K5L4K5L4L4M32N2N00M3M3M3L4M3M3M3M3L4M3M3M3L4M3M3MNoB[Ini3^ATLg>e3_AWLe>e3^AYLe>c3^AZLd>d3_AYLd>d3_AZLc>c3o0L4M3M3N2O1O100O1O1O1O10OO2N0001O01O000001O01O0001O000O101N10000O2O000O1010O001O000mN]_OkNc`0U1c_O^Nc`0b1n00O010O01O010O2O0O100O100O2O0O100O101N1O1O1O1O4J5K5K6K4K5J7H7HW[b1"}, "label": "woman hitting a tennis ball with racket"}]} {"id": 545948, "image": "COCO_train2014_000000545948.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person sitting in a chair\"."}], "task": "refering", "answer_template": "The \"a person sitting in a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 232, 233, 234, 254, 255, 256, 257, 277, 278, 279, 280, 301, 302, 303, 304, 326, 327], "bbox": [0.0703125, 0.6543091334894613, 0.262578125, 0.9876346604215457], "iscrowd": 0, "area": 7754.3715, "rle": {"size": [427, 640], "counts": "\\Sc02V=5K4L4M3L4L4L4M3M3M3M3N2M3M4L3M3M3N2N2O1N2O1N2N100O100O010O1O1O1O1O1O100O1O1O1O001O1O1O001N2O001N20O100000O10000000001O001N2O1O001O1N20O1000000O11OlMmES1o9jNYFU1c9hNdFX1X9dNoF\\1n8_NXG`1h8\\N^Gc1b8WNdGh1\\8TNjGj1m910O08IM2N3M2N1O1O001O1O1O0O2O1O1O1O001O001O001O001O001O1O1O2N1O1O1O2N1O1O1O2AeCJ\\ in this image.", "objects": [{"patches": [210, 232, 233, 234, 254, 255, 256, 257, 277, 278, 279, 280, 301, 302, 303, 304, 326, 327], "bbox": [0.0703125, 0.6543091334894613, 0.262578125, 0.9876346604215457], "iscrowd": 0, "area": 7754.3715, "rle": {"size": [427, 640], "counts": "\\Sc02V=5K4L4M3L4L4L4M3M3M3M3N2M3M4L3M3M3N2N2O1N2O1N2N100O100O010O1O1O1O1O1O100O1O1O1O001O1O1O001N2O001N20O100000O10000000001O001N2O1O001O1N20O1000000O11OlMmES1o9jNYFU1c9hNdFX1X9dNoF\\1n8_NXG`1h8\\N^Gc1b8WNdGh1\\8TNjGj1m910O08IM2N3M2N1O1O001O1O1O0O2O1O1O1O001O001O001O001O001O1O1O2N1O1O1O2N1O1O1O2AeCJ\\ in this image.", "objects": [{"patches": [81, 82, 83, 84, 85, 86, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.030343750000000003, 0.18358333333333335, 1.0, 0.9790833333333333], "iscrowd": 0, "area": 160989.97769999996, "rle": {"size": [480, 640], "counts": "PV91k> in this image.", "objects": [{"patches": [81, 82, 83, 84, 85, 86, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.030343750000000003, 0.18358333333333335, 1.0, 0.9790833333333333], "iscrowd": 0, "area": 160989.97769999996, "rle": {"size": [480, 640], "counts": "PV91k> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 23, 24, 25, 26, 27, 28, 29, 30, 46, 47, 48, 49, 50, 51, 52, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141], "bbox": [0.0, 0.0, 0.3471875, 0.4134791666666667], "iscrowd": 0, "area": 31680.472700000002, "rle": {"size": [480, 640], "counts": "4i1S3]Ok5l40000000O10000000000O10000000000O100000000O10000000000O1000000000O010000000000O100000000O10000000000O10000000000O10000000000O10000000000O10O100000O10000000000O10000000000O10000000000O100000000O10000O1O1O100O1O1O1O10O01O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O00100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O2N3M2N2O1N2N2N2N2O1N2L4J6K5K5K5K3M3M4L3M3L5L3M3M3M4L3N2O101N100O2N100O101N2O1N3N1N2N2O1N2O1N2O1N3M2O1N2O1N2O1N2NnmS6"}, "label": "a dish of broccoli"}]} {"id": 201897, "image": "COCO_train2014_000000201897.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman with the high white socks\"."}], "task": "refering", "answer_template": "The \"the woman with the high white socks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 117, 121, 122, 123, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 303, 304, 305, 326, 327, 328], "bbox": [0.07646875, 0.13574999999999998, 0.6373125, 0.8816458333333332], "iscrowd": 0, "area": 51286.1496, "rle": {"size": [480, 640], "counts": "^Tg03l>4L5K5K4L5K2N2N2N2N101N2O1N2O100O10000O01O010O0010O01O010O0010O000010O0000010O0001O01O01O1O100O1O100O1O1O100O0000010N100000000O2O0000000O10001PHbN\\2^1bMROP2n0oMSOQ2m0mMUOR2l0mMUOS2k0kMWOU2i0jMXOV2f0jM\\OV2a0lM@S2?mMCS2;nMFR28nMKQ23PNNP21PN0o11PN0P20PN0P21oMOQ21nM0R20nM0Q21nM0R20nM0R21lM0T20lM0T20lM0T20kM1T20lM0T20kM1U20jM0V20iM1W2OhM2W2OhM2X2NgM3Y2NeM3[2MdM4\\2LcM5\\2LcM4^2LaM4`2M^M2d2NVJROT1P1g4NjIKP16X5O^Id0j0]Oh50SIQ4o6o11N2O1O0O2O1N2O1N101N2O1N2O0O2O1O1N2O1N101N2O1N2O0O2N2O1N2N1O2N2O1N2N1O2N2N2O1N2N1O2N2O1N2O02O001O000O2O001O001N101O001O0O2O001O00001N101OO1O1O00100O1O10O01O100O10O01O100O10O10O10000O10O100N1O2O1N1O2O1N2O0O2N2O0O2N2O1N1O2O1000O10000O101O000O10000O1000001N1000D;0100O10O10O10O10O1000O010000O010O10O10O1O1O001O1O001O1O1O001O1O001O1O1O1L3N3MSOcH]JY7X5XIfJg6S5bIjJ`6T5cIjJ^6T5fIiJ[6U5hIiJY6T5lIiJV6T5mIjJT6T5oIjJR6T5RJiJo5U5TJiJm5U5VJiJl5S5R2N2N2N1O1N2O1O1O1O1O1N2O3M2N2N3L3N2N3N1O1O2N1O1O2N1O1O1O2M2O1O2N1O1O2N1O1O3M2N2N2N2N2N2M3N3M2O1N2N2N2N2N3M2N2N2N2M4L`0@?A`0@3M4L4Lbc\\3"}, "label": "the woman with the high white socks"}]} {"id": 201897, "image": "COCO_train2014_000000201897.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl wearing white knee socks\"."}], "task": "refering", "answer_template": "The \"a girl wearing white knee socks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 79, 80, 81, 82, 83, 101, 102, 103, 104, 105, 106, 117, 121, 122, 123, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 188, 189, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 303, 304, 305, 326, 327, 328], "bbox": [0.07646875, 0.13574999999999998, 0.6373125, 0.8816458333333332], "iscrowd": 0, "area": 51286.1496, "rle": {"size": [480, 640], "counts": "^Tg03l>4L5K5K4L5K2N2N2N2N101N2O1N2O100O10000O01O010O0010O01O010O0010O000010O0000010O0001O01O01O1O100O1O100O1O1O100O0000010N100000000O2O0000000O10001PHbN\\2^1bMROP2n0oMSOQ2m0mMUOR2l0mMUOS2k0kMWOU2i0jMXOV2f0jM\\OV2a0lM@S2?mMCS2;nMFR28nMKQ23PNNP21PN0o11PN0P20PN0P21oMOQ21nM0R20nM0Q21nM0R20nM0R21lM0T20lM0T20lM0T20kM1T20lM0T20kM1U20jM0V20iM1W2OhM2W2OhM2X2NgM3Y2NeM3[2MdM4\\2LcM5\\2LcM4^2LaM4`2M^M2d2NVJROT1P1g4NjIKP16X5O^Id0j0]Oh50SIQ4o6o11N2O1O0O2O1N2O1N101N2O1N2O0O2O1O1N2O1N101N2O1N2O0O2N2O1N2N1O2N2O1N2N1O2N2N2O1N2N1O2N2O1N2O02O001O000O2O001O001N101O001O0O2O001O00001N101OO1O1O00100O1O10O01O100O10O01O100O10O10O10000O10O100N1O2O1N1O2O1N2O0O2N2O0O2N2O1N1O2O1000O10000O101O000O10000O1000001N1000D;0100O10O10O10O10O1000O010000O010O10O10O1O1O001O1O001O1O1O001O1O001O1O1O1L3N3MSOcH]JY7X5XIfJg6S5bIjJ`6T5cIjJ^6T5fIiJ[6U5hIiJY6T5lIiJV6T5mIjJT6T5oIjJR6T5RJiJo5U5TJiJm5U5VJiJl5S5R2N2N2N1O1N2O1O1O1O1O1N2O3M2N2N3L3N2N3N1O1O2N1O1O2N1O1O1O2M2O1O2N1O1O2N1O1O3M2N2N2N2N2N2M3N3M2O1N2N2N2N2N3M2N2N2N2M4L`0@?A`0@3M4L4Lbc\\3"}, "label": "a girl wearing white knee socks"}]} {"id": 201897, "image": "COCO_train2014_000000201897.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"asian girl in long sleeve brown hoodie and holding wii remote\"."}], "task": "refering", "answer_template": "The \"asian girl in long sleeve brown hoodie and holding wii remote\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 63, 83, 84, 85, 86, 87, 106, 107, 108, 109, 110, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 226, 242, 243, 244, 245, 246, 247, 248, 249, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 371, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.1550625, 0.12808333333333333, 0.866296875, 0.9887708333333334], "iscrowd": 0, "area": 81383.932, "rle": {"size": [480, 640], "counts": "jk^11n>1O100O1O1O1O1O1O010O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O100O1O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1O1N2O001N2O1O1N2O1O1O1N2O1O1N200O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O100000000O10000000000O100000000O10000000000O10000000000O100000O10O100000000O10000000O010O1O100O010O100O100O00100O100O010O100O1O010O100O010O010O0010O0010O010O01dNkKmGT4Q8ULhGk3W8\\LcGe3Z8cLaG\\3]8kL^GV3_8oL_GP3`8SM^Gn2_8VM`Gi2^8[M`Gf2]8^MbGa2]8cM`G]2^8gM`GZ2]8jMbGU2\\8oMbGR2\\8QNcGn1[8]2M201N2O1O1N2O1N3N1N2O1N2O1M3N2N2N2N2O1O1N2O1O1N2M3N2N2M3N2M4M2O10000O10000O10000O10000O1O100O1O1O1O1O2N100O1O1O1BVGeJk8V5b0K5K5DI7K5VOZEUMk:j2f000O100O10O0010O01O001O1O1O1O1O1O1O100O2N1O1O1O1O1O1O1O100O1O2N1O1O1O1O1O1O100O1O1O1O2N1O1O1O1O101N1O1O1O1O2N1O2N1O2O0O2N1O2N1O2N1O2N1O2O0O2N1O2N1O2N1O2N2N1O2O1N3M3M3N1N3M3M3N0O1O2N100O2N1O100O2N1O101N1O1O2O0O1O101O005J;F:F9G9F3M3N2M2N3N2M3M2N3M2N2N2N1O2O1N2N2N2JC=C=C in this image.", "objects": [{"patches": [60, 61, 62, 63, 83, 84, 85, 86, 87, 106, 107, 108, 109, 110, 129, 130, 131, 132, 133, 134, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 226, 242, 243, 244, 245, 246, 247, 248, 249, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 371, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.1550625, 0.12808333333333333, 0.866296875, 0.9887708333333334], "iscrowd": 0, "area": 81383.932, "rle": {"size": [480, 640], "counts": "jk^11n>1O100O1O1O1O1O1O010O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O1O100O100O100O100O100O100O1O100O100O100O100O100O100O1O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1O1N2O001N2O1O1N2O1O1O1N2O1O1N200O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O100000000O10000000000O100000000O10000000000O10000000000O100000O10O100000000O10000000O010O1O100O010O100O100O00100O100O010O100O1O010O100O010O010O0010O0010O010O01dNkKmGT4Q8ULhGk3W8\\LcGe3Z8cLaG\\3]8kL^GV3_8oL_GP3`8SM^Gn2_8VM`Gi2^8[M`Gf2]8^MbGa2]8cM`G]2^8gM`GZ2]8jMbGU2\\8oMbGR2\\8QNcGn1[8]2M201N2O1O1N2O1N3N1N2O1N2O1M3N2N2N2N2O1O1N2O1O1N2M3N2N2M3N2M4M2O10000O10000O10000O10000O1O100O1O1O1O1O2N100O1O1O1BVGeJk8V5b0K5K5DI7K5VOZEUMk:j2f000O100O10O0010O01O001O1O1O1O1O1O1O100O2N1O1O1O1O1O1O1O100O1O2N1O1O1O1O1O1O100O1O1O1O2N1O1O1O1O101N1O1O1O1O2N1O2N1O2O0O2N1O2N1O2N1O2N1O2O0O2N1O2N1O2N1O2N2N1O2O1N3M3M3N1N3M3M3N0O1O2N100O2N1O100O2N1O101N1O1O2O0O1O101O005J;F:F9G9F3M3N2M2N3N2M3M2N3M2N2N2N1O2O1N2N2N2JC=C=C in this image.", "objects": [{"patches": [77, 78, 79, 98, 99, 100, 101, 102, 103, 104, 120, 121, 122, 123, 124, 125, 126, 127, 128, 143, 144, 145, 146, 147, 148, 149, 150, 151, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 235, 236, 237, 238, 239, 240, 241, 242, 243, 258, 259, 260, 261, 262, 263, 264, 265, 266, 281, 282, 283, 284, 285, 286, 287, 288, 289, 303, 304, 305, 306, 307, 308, 309, 310, 311, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.0016875000000000002, 0.22472916666666667, 0.62359375, 0.9865208333333334], "iscrowd": 0, "area": 84867.96489999998, "rle": {"size": [480, 640], "counts": "fm04e>7I7J6O1O100O100O100N2O1O1O1O100O100O100O100O10O0100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O10000O100O10000O10000O100O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1oJ[MfMf2X2aMcM_2Z2jM_MW2_2PN\\MP2b2XNWMi1f2_NTMb1j2eNQM[1l2nNmLS1Q3XOfLh0X3D]L=`30VL0h3=lKDQ4i0dKXOZ4R1^KnN`4Y1ZKhNc4_1YKaNe4f1UK[Nh4T2lJlMR5R6N2M3N2M3N2N2N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O100O1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000000000000000000000001O0000000000000O1000000000001O00000000000000000000000000001O0C[L_Ee3a:=O1O1O1O1O1O1O1O2N1O2N1O101N1O2O0O1O2N101N1O1O2O0O2N1O101N1O2N100O2O0O2O000O2O0O2O000O2N1O2N1O2N1N3N2N2L4L4L4L4L4M3M3M2N3N2M3M3N2L4J6K5J6J5L5K5K5K5K5K5L4L4L4L5L4K5K5L4K5K5\\Od0G9M3M3L4M3M3L4M3M3L4K5K9H8G9G9G:B=^OUh`3"}, "label": "a person in a black helmet and brown shirt smiling on a snowy mountaintop"}]} {"id": 87214, "image": "COCO_train2014_000000087214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person with glasses looking at the camera\"."}], "task": "refering", "answer_template": "The \"person with glasses looking at the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 98, 99, 100, 101, 102, 103, 104, 120, 121, 122, 123, 124, 125, 126, 127, 128, 143, 144, 145, 146, 147, 148, 149, 150, 151, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 235, 236, 237, 238, 239, 240, 241, 242, 243, 258, 259, 260, 261, 262, 263, 264, 265, 266, 281, 282, 283, 284, 285, 286, 287, 288, 289, 303, 304, 305, 306, 307, 308, 309, 310, 311, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378], "bbox": [0.0016875000000000002, 0.22472916666666667, 0.62359375, 0.9865208333333334], "iscrowd": 0, "area": 84867.96489999998, "rle": {"size": [480, 640], "counts": "fm04e>7I7J6O1O100O100O100N2O1O1O1O100O100O100O100O10O0100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O10000O10000O10000O10000O10000O10000O100O10000O10000O10000O10000O10000O10000O10000O100O10000O10000O100O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O100O1oJ[MfMf2X2aMcM_2Z2jM_MW2_2PN\\MP2b2XNWMi1f2_NTMb1j2eNQM[1l2nNmLS1Q3XOfLh0X3D]L=`30VL0h3=lKDQ4i0dKXOZ4R1^KnN`4Y1ZKhNc4_1YKaNe4f1UK[Nh4T2lJlMR5R6N2M3N2M3N2N2N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O100O1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000O10000O1000000O10000O10000000000000000000000001O0000000000000O1000000000001O00000000000000000000000000001O0C[L_Ee3a:=O1O1O1O1O1O1O1O2N1O2N1O101N1O2O0O1O2N101N1O1O2O0O2N1O101N1O2N100O2O0O2O000O2O0O2O000O2N1O2N1O2N1N3N2N2L4L4L4L4L4M3M3M2N3N2M3M3N2L4J6K5J6J5L5K5K5K5K5K5L4L4L4L5L4K5K5L4K5K5\\Od0G9M3M3L4M3M3L4M3M3L4K5K9H8G9G9G:B=^OUh`3"}, "label": "person with glasses looking at the camera"}]} {"id": 87214, "image": "COCO_train2014_000000087214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man with black helmet and blue jacket\"."}], "task": "refering", "answer_template": "The \"man with black helmet and blue jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 93, 94, 95, 96, 116, 117, 118, 119, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 322, 323, 324, 325, 345, 346], "bbox": [0.0, 0.20675, 0.23595312499999999, 0.9303333333333333], "iscrowd": 0, "area": 35473.15290000001, "rle": {"size": [480, 640], "counts": "Wk5o4SLbKl3_4RLcKl3_4QLdKm3^4PLeKo3[4PLiKl3Y4SLkKh3W4WLmKe3T4ZLPLa3R4^LRL]3P4RLPJWOT2c4m3RLUJWOR2c4m3PLWJZOo1b4o3oKkLm3X3oKlL]2eL`Nc6nNlL`2cL^Nf6mNkLb2cL]Ne6nNlLa2cL\\Nb:b1bEZN^:f1eEVN\\:j1gESNZ:l1iEPNX:P2kEjMY:U2jEeMY:[2W101N4L3M3M3M3M3_NmBo0T=nNoBQ1S=kNPCT1Q=iNRCW1\\=O1O1O1O2N1O2N2N1O`0A_lT7"}, "label": "man with black helmet and blue jacket"}]} {"id": 87214, "image": "COCO_train2014_000000087214.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man facing away in the helmet\"."}], "task": "refering", "answer_template": "The \"the man facing away in the helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 93, 94, 95, 96, 116, 117, 118, 119, 138, 139, 140, 141, 161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 322, 323, 324, 325, 345, 346], "bbox": [0.0, 0.20675, 0.23595312499999999, 0.9303333333333333], "iscrowd": 0, "area": 35473.15290000001, "rle": {"size": [480, 640], "counts": "Wk5o4SLbKl3_4RLcKl3_4QLdKm3^4PLeKo3[4PLiKl3Y4SLkKh3W4WLmKe3T4ZLPLa3R4^LRL]3P4RLPJWOT2c4m3RLUJWOR2c4m3PLWJZOo1b4o3oKkLm3X3oKlL]2eL`Nc6nNlL`2cL^Nf6mNkLb2cL]Ne6nNlLa2cL\\Nb:b1bEZN^:f1eEVN\\:j1gESNZ:l1iEPNX:P2kEjMY:U2jEeMY:[2W101N4L3M3M3M3M3_NmBo0T=nNoBQ1S=kNPCT1Q=iNRCW1\\=O1O1O1O2N1O2N2N1O`0A_lT7"}, "label": "the man facing away in the helmet"}]} {"id": 455874, "image": "COCO_train2014_000000455874.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in white with a black apron\"."}], "task": "refering", "answer_template": "The \"a man in white with a black apron\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 251, 269, 270, 271, 272, 293, 294, 316, 317], "bbox": [0.7224062499999999, 0.3, 0.9672031249999999, 0.8883606557377048], "iscrowd": 0, "area": 20630.16455, "rle": {"size": [427, 640], "counts": "XnP64n in this image.", "objects": [{"patches": [112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 251, 269, 270, 271, 272, 293, 294, 316, 317], "bbox": [0.7224062499999999, 0.3, 0.9672031249999999, 0.8883606557377048], "iscrowd": 0, "area": 20630.16455, "rle": {"size": [427, 640], "counts": "XnP64n in this image.", "objects": [{"patches": [85, 86, 98, 99, 100, 101, 102, 111, 112, 113, 114, 115, 116, 126, 127, 128, 129, 130, 140, 141, 142, 143, 144, 145, 155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 190, 199, 200, 201, 202, 203, 204, 205, 214, 215, 216, 217, 218, 219, 220, 229, 230, 231, 232, 233, 234, 243, 244, 245, 247, 248, 258, 259, 263, 264, 272, 273, 274, 279, 287, 288, 289, 290, 304, 305], "bbox": [0.1692, 0.240453125, 0.8256941176470587, 0.885390625], "iscrowd": 0, "area": 39430.78915, "rle": {"size": [640, 425], "counts": "d`]11nc02N2N2N3M2N2N2N3M2N2N2N3L2O2N1O2N1O2N1O1O2N1O2N1O2L3K6K4K6K4K5L5J5L5M2O2M2N3M2O2M2N2O2M2N3M2O1N2N2N2O1N`NZ@jNd?S1g@gNX?U1TAeNi>X1]AgNa>V1eAiNY>U1lAjNS>U1oAkNo=V1SBiNk=X1WBgNg=Z1ZBfNa=]1bBcNV=c1lB\\NndEmN=h0n9;fEmN9k0P:8hElN6n0R:5jEmN1Q1T:2lEmNM3LbNZ:^2nElNJ14eNS:^2PFlNFOnLS9[2`Ff0?oLP9Z2cFd0`0QMm8Z2eFb0a0TMi8Z2gF`0b0VMf8Y2jF>c0YMb8X2mF in this image.", "objects": [{"patches": [85, 86, 98, 99, 100, 101, 102, 111, 112, 113, 114, 115, 116, 126, 127, 128, 129, 130, 140, 141, 142, 143, 144, 145, 155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 190, 199, 200, 201, 202, 203, 204, 205, 214, 215, 216, 217, 218, 219, 220, 229, 230, 231, 232, 233, 234, 243, 244, 245, 247, 248, 258, 259, 263, 264, 272, 273, 274, 279, 287, 288, 289, 290, 304, 305], "bbox": [0.1692, 0.240453125, 0.8256941176470587, 0.885390625], "iscrowd": 0, "area": 39430.78915, "rle": {"size": [640, 425], "counts": "d`]11nc02N2N2N3M2N2N2N3M2N2N2N3L2O2N1O2N1O2N1O1O2N1O2N1O2L3K6K4K6K4K5L5J5L5M2O2M2N3M2O2M2N2O2M2N3M2O1N2N2N2O1N`NZ@jNd?S1g@gNX?U1TAeNi>X1]AgNa>V1eAiNY>U1lAjNS>U1oAkNo=V1SBiNk=X1WBgNg=Z1ZBfNa=]1bBcNV=c1lB\\NndEmN=h0n9;fEmN9k0P:8hElN6n0R:5jEmN1Q1T:2lEmNM3LbNZ:^2nElNJ14eNS:^2PFlNFOnLS9[2`Ff0?oLP9Z2cFd0`0QMm8Z2eFb0a0TMi8Z2gF`0b0VMf8Y2jF>c0YMb8X2mF in this image.", "objects": [{"patches": [232, 233, 234, 237, 238, 247, 248, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 276, 277, 278, 279, 280, 281, 282, 283], "bbox": [0.00428125, 0.6750233644859813, 0.828515625, 0.825443925233645], "iscrowd": 0, "area": 12588.617899999992, "rle": {"size": [428, 640], "counts": "]b1g0e2N2N00N2N2N2N2N2N2M3N2N21M2J7J6I7I6K6I`a23\\^M7H9G8K6O00O100000O100000O100000000000OK6H8H8HTf]1"}, "label": "a blue cloth on a table"}]} {"id": 513221, "image": "COCO_train2014_000000513221.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a long picnic table adorned with a blue and pink pastel cloth , and two sheet cakes\"."}], "task": "refering", "answer_template": "The \"a long picnic table adorned with a blue and pink pastel cloth , and two sheet cakes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [232, 233, 234, 237, 238, 247, 248, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 276, 277, 278, 279, 280, 281, 282, 283], "bbox": [0.00428125, 0.6750233644859813, 0.828515625, 0.825443925233645], "iscrowd": 0, "area": 12588.617899999992, "rle": {"size": [428, 640], "counts": "]b1g0e2N2N00N2N2N2N2N2N2M3N2N21M2J7J6I7I6K6I`a23\\^M7H9G8K6O00O100000O100000O100000000000OK6H8H8HTf]1"}, "label": "a long picnic table adorned with a blue and pink pastel cloth , and two sheet cakes"}]} {"id": 513221, "image": "COCO_train2014_000000513221.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table cloth that is draped over the edge of the table in front of the cake . it had paw - prints on it and blue / green / orange stripes\"."}], "task": "refering", "answer_template": "The \"the table cloth that is draped over the edge of the table in front of the cake . it had paw - prints on it and blue / green / orange stripes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [263, 264, 265, 266, 267, 268, 269, 270, 271, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.0, 0.7830607476635514, 0.8270625, 1.0], "iscrowd": 0, "area": 44381.5448, "rle": {"size": [428, 640], "counts": "T;X2S;100000000000000O100000000000000O100000000000000O100000000000000O1000000LdMXE\\2h:dMXE\\2h:eMWE[2i:eMWE[2i:400000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000O100000000000000000000000000O10000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000O1O1O1O1O1O1O001O100000000O100000000000000000000O100000000000000000000000000000O1000001O000000001O1O001O001O001O1O001O001O001O0000000000001O00000O10000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000001O00001O0000001O00O100000000000O1L4L4L4M3L4L4L4M3L4E;D in this image.", "objects": [{"patches": [263, 264, 265, 266, 267, 268, 269, 270, 271, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340], "bbox": [0.0, 0.7830607476635514, 0.8270625, 1.0], "iscrowd": 0, "area": 44381.5448, "rle": {"size": [428, 640], "counts": "T;X2S;100000000000000O100000000000000O100000000000000O100000000000000O1000000LdMXE\\2h:dMXE\\2h:eMWE[2i:eMWE[2i:400000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000O100000000000000000000000000O10000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000O1O1O1O1O1O1O001O100000000O100000000000000000000O100000000000000000000000000000O1000001O000000001O1O001O001O001O1O001O001O001O0000000000001O00000O10000000000000000000000000000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000001O00001O0000001O00O100000000000O1L4L4L4M3L4L4L4M3L4E;D in this image.", "objects": [{"patches": [215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 261, 262, 263, 264, 265, 266, 267, 268, 269], "bbox": [0.378515625, 0.6109579439252336, 0.7615000000000001, 0.751892523364486], "iscrowd": 0, "area": 12872.7928, "rle": {"size": [428, 640], "counts": "f^U34k<=C=B>J60000000000O100000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O10000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O1O2M2O1O2N1O2M2O1O2N1O1N3N1O2N1O1N3N1O1O2N1N3N1O1O2N1N2O2N1O2N1N2O2N1O1O2Mddo1"}, "label": "a large sheet cake with a black dog on it"}]} {"id": 513221, "image": "COCO_train2014_000000513221.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white cake with blue , white , and green frosting in front of a woman wearing a black and read flowered shirt\"."}], "task": "refering", "answer_template": "The \"the white cake with blue , white , and green frosting in front of a woman wearing a black and read flowered shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 261, 262, 263, 264, 265, 266, 267, 268, 269], "bbox": [0.378515625, 0.6109579439252336, 0.7615000000000001, 0.751892523364486], "iscrowd": 0, "area": 12872.7928, "rle": {"size": [428, 640], "counts": "f^U34k<=C=B>J60000000000O100000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O10000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O1O2M2O1O2N1O2M2O1O2N1O1N3N1O2N1O1N3N1O1O2N1N3N1O1O2N1N2O2N1O2N1N2O2N1O1O2Mddo1"}, "label": "the white cake with blue , white , and green frosting in front of a woman wearing a black and read flowered shirt"}]} {"id": 177353, "image": "COCO_train2014_000000177353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small giraffe leaning over a car behind a larger giraffe\"."}], "task": "refering", "answer_template": "The \"a small giraffe leaning over a car behind a larger giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [192, 193, 194, 195, 214, 215, 216, 217, 218, 234, 235, 237, 256, 257, 258, 259, 260, 279, 280, 281, 282, 302, 303, 325, 326, 348, 349, 371, 372], "bbox": [0.15326562500000002, 0.4657319587628866, 0.48704687500000005, 0.9645979381443298], "iscrowd": 0, "area": 9409.485049999996, "rle": {"size": [485, 640], "counts": "`i^14o>8H7H8IN2O1N1OQOEZC:\\1CQ9b0TEIf1NT9R1fFWOX9g0cFB[9=^FN_90\\F9[9[NdE[4[:eKeE\\4Z:dKeE^4Z:cKdE`4Z:`KeEb4Z:K9O12N2N2N2N3M2O1O1O1O2NL4\\Nd1YOg0O1N2O1N2O1N2O1N2N2O0O2O0O2O0O2O1N101N101N101O1O0010O01O00101M3G:F9H8G9GoT7NWkH4K4M4L3M4M2N3M2N3M2N3M2N2N2N3M2M3N2M3M3M3N200OO2M2M4WOi0001O000010O010O010O10O0100O010O10O01000O0100O010O10O0100O010O10O0100O01000O010O10O010O0100O010O10O01000O0100O010O10O0100O010O10O010O10O010O10O010O01001O100O1O100M3K5K5L4O10N101O001O001O?@2O1O2N1O1O1O2M[Sk4"}, "label": "a small giraffe leaning over a car behind a larger giraffe"}]} {"id": 177353, "image": "COCO_train2014_000000177353.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a graiffe on the right side of the jeep standing the closest\"."}], "task": "refering", "answer_template": "The \"a graiffe on the right side of the jeep standing the closest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 32, 33, 34, 55, 56, 57, 58, 78, 80, 81, 82, 83, 105, 106, 107, 108, 128, 129, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180], "bbox": [0.396703125, 0.025917525773195876, 0.8764843750000001, 0.47002061855670096], "iscrowd": 0, "area": 17566.607399999997, "rle": {"size": [485, 640], "counts": "b\\h33o>5K4L4K5L300O010O10000O0O2O1N2O0O2N1O1\\OQOoBS1Q=`01O1O100O1O1O1O001O100O101O00[OSORCl0nC3M1N3N001O0O2O001N1002N2N2N1O2O1O1N101N2N1O2N2N1O2M2OWNoB^1Q=aNRC^1m<`NVC_1j<`NXC_1h<_N[C`1d<`N^C_1b<_NaC`1R=01N2O1O1O1O1O1O1N2O1O100O1O1O101N1O2N1O2M2N2N2N2N2M3N2N2N1O2N20000O100000O0YA0Z>>0O10O2O1O0O2Og0YO0O2O1O1N101O1oC]N\\:e1`E`N^:`1_EfN^:[1^EkN^:W1^EnN`:R1]EQOc:P1YEROg:P1UEROl:n0QEUOn:m0nDUOS;l0iDVOW;k0gDWOY;j0cDXO];j0_DXOa;W20O0010O00010O0001O0O2O00001O0O101O001O0O101O00001N101O00001N101O00001N101O00001N101O00001N101O001O0O101O001O0O2O001O001N10001O001N101O001O0O2O00001O002M3N0000000O100000001N100000000O100000001N100000000O1000001O0O10001O000O2O001O001N10001O001N101O001O0O101O001O0O2O001O000O2O1O1O1O1O1O1N3N1O1O1O1O1O1O1O1O1N3N2N2N2N2N6JWWU1"}, "label": "a graiffe on the right side of the jeep standing the closest"}]} {"id": 275658, "image": "COCO_train2014_000000275658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the black horse pulling the carriage\"."}], "task": "refering", "answer_template": "The \"the black horse pulling the carriage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 92, 93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 141, 142, 143, 163, 164, 165, 166, 187, 188, 189, 191, 192, 210, 211, 214, 215, 233, 234, 238, 256, 257], "bbox": [0.0131875, 0.2687777777777778, 0.36887499999999995, 0.8829166666666667], "iscrowd": 0, "area": 10693.5028, "rle": {"size": [360, 640], "counts": "^o21P;9J6L3L5L4K5L3L4M2L4M3M3M3M3N2N1O2N2N1O2N2N1O2N2O001OO001O2O00010O00010O000O1O202N2O1N1O01N1O1O1019GO10000000001O0O2O1N2O1N3M4M2M4L100O1O100O1O100O1O108G2N2O1O1N20eM[I1g6H`I6c6C_I?c6ZOaIg0b6ROaIn0\\81SHPOa5Q1ZJROf5o0WJSOj5n0RJTOo5l0nIVOR6k0\\HnNR19c6h0[HROm08h6f0\\HTOg08n6c0[HXOb07S7a0[HZO>7W7?[H^O85]7=ZHF0Of7;YHMIJn79YHi0g7WOXHi0g7XOZHe0g7QOdHm0^7ROcHZO\\O^1R8ZObHWOCW14hNZ7m0mHUOKm0NTOX7l0oHRO1g0G^OW7m0nHoN7a0DEU7Q1mHmN97DNT7S1lHlNm04U6T1lHmNg02[6S1nHnN?4a6o0QIQO46i6k0SIROK9Q7f0TI8j6JVI6i6KWI5g6MYI2f60ZIOf62[IfN12e6Z1ZIaN9Ka6e1VI\\Nc0C\\6P2j1O2O2H9N2O1O1N2O001O1O1O1O1O1O1O2N2M3N2N2N3M2N2N2N3M2M3N2N2M4M2N2MSca02i\\^O5L4L4O2M2N5K8_FQOY8X1[GkNe8]1PGfNP9j1O00000000QN]GZ1b8dNdGX1Y9M3YOoE1h:M4Gog]4"}, "label": "the black horse pulling the carriage"}]} {"id": 275658, "image": "COCO_train2014_000000275658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black horse next to white horse pulling a carriage\"."}], "task": "refering", "answer_template": "The \"black horse next to white horse pulling a carriage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 92, 93, 94, 95, 115, 116, 117, 118, 138, 139, 140, 141, 142, 143, 163, 164, 165, 166, 187, 188, 189, 191, 192, 210, 211, 214, 215, 233, 234, 238, 256, 257], "bbox": [0.0131875, 0.2687777777777778, 0.36887499999999995, 0.8829166666666667], "iscrowd": 0, "area": 10693.5028, "rle": {"size": [360, 640], "counts": "^o21P;9J6L3L5L4K5L3L4M2L4M3M3M3M3N2N1O2N2N1O2N2N1O2N2O001OO001O2O00010O00010O000O1O202N2O1N1O01N1O1O1019GO10000000001O0O2O1N2O1N3M4M2M4L100O1O100O1O100O1O108G2N2O1O1N20eM[I1g6H`I6c6C_I?c6ZOaIg0b6ROaIn0\\81SHPOa5Q1ZJROf5o0WJSOj5n0RJTOo5l0nIVOR6k0\\HnNR19c6h0[HROm08h6f0\\HTOg08n6c0[HXOb07S7a0[HZO>7W7?[H^O85]7=ZHF0Of7;YHMIJn79YHi0g7WOXHi0g7XOZHe0g7QOdHm0^7ROcHZO\\O^1R8ZObHWOCW14hNZ7m0mHUOKm0NTOX7l0oHRO1g0G^OW7m0nHoN7a0DEU7Q1mHmN97DNT7S1lHlNm04U6T1lHmNg02[6S1nHnN?4a6o0QIQO46i6k0SIROK9Q7f0TI8j6JVI6i6KWI5g6MYI2f60ZIOf62[IfN12e6Z1ZIaN9Ka6e1VI\\Nc0C\\6P2j1O2O2H9N2O1O1N2O001O1O1O1O1O1O1O2N2M3N2N2N3M2N2N2N3M2M3N2N2M4M2N2MSca02i\\^O5L4L4O2M2N5K8_FQOY8X1[GkNe8]1PGfNP9j1O00000000QN]GZ1b8dNdGX1Y9M3YOoE1h:M4Gog]4"}, "label": "black horse next to white horse pulling a carriage"}]} {"id": 275658, "image": "COCO_train2014_000000275658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dappled gray horse pulling a red cart\"."}], "task": "refering", "answer_template": "The \"a dappled gray horse pulling a red cart\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 99, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 213, 214, 217, 218, 236, 237, 240, 241, 259, 260, 262, 263, 264], "bbox": [0.13760937499999998, 0.32102777777777775, 0.5142187500000001, 0.9197500000000001], "iscrowd": 0, "area": 24199.055050000003, "rle": {"size": [360, 640], "counts": "fSo07k:>QN]OSH0^OQ1Y8ROPHg1l7\\NiGk1X8b01O100O000001N1000000O100O10000OO2O1N101N101N10O0011N2N2O1N3N1N3N1N3N1N2O2M2O2N1N3N11O001O001O1O001O1O1O1O1O1O001O1O6J6J5K1O2N2O1N2N2N2N2O0O2N2N1O2O1N2N2N1O2O1N2N2N2N3M4L5K3nJQLe2R4RMTLn2o3lLTLT3o3eLUL[3n3_LULa3m3\\LTLd3o3WLSLi3P4SLQLm3R4nKPLR4S4iKPLV4S4dKPL\\4T4]KoKd4U4UKmKk4X51O001O001O0O2O0O2O001eMnJTNOZNT5`3QKSNM\\NS5^3TKUMOJMAQ5^3XKPMONW5Y2]JiMb0J01R5\\2_JdMi0DJ:o4^2aJ_Me1Ok3b2dJXMd7h2>01O00001O001O001O001O00001O00001O00000000001O0000000000O100O100O100O1O10O010O010O010O010O010O010O100O100N2O2N[GQNS8c2O2VJlL\\3U3cLkL]3W3`LjL`3W3^LkLa3W3]LiLc3Y3YLjLg3X3ULiLk3Z3PLhLP4\\3jKgLU4\\3gKeLY4^3bKdL^4_3^KbLb4a3fJTMZ5]41O000SM`J2`5KeJ2]5JhJ3Y5JlJ1X5LlJLZ52jJIZ56hJGY59iJDY5;iJnMZOU1n5n0`KgNb4X1fK_N\\4`1f201O1002N2N3M2N2N3M3M2N3N2cIYMR4j2fKZM\\4h2]K[Md4h2UK[Mm4e2mJ`MS5`2iJcMY5]2`JhMb5Z2VJjMk5_2fIfM\\6b32N1O2N2nLaIW1`6eMeJY2W7SN`Gn0b8iNjG;@Ki8ASH4C7[:L5J4MoP]3"}, "label": "a dappled gray horse pulling a red cart"}]} {"id": 275658, "image": "COCO_train2014_000000275658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the grey horse with its whole body visible\"."}], "task": "refering", "answer_template": "The \"the grey horse with its whole body visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 99, 101, 102, 118, 119, 120, 121, 122, 123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 190, 191, 192, 193, 194, 195, 213, 214, 217, 218, 236, 237, 240, 241, 259, 260, 262, 263, 264], "bbox": [0.13760937499999998, 0.32102777777777775, 0.5142187500000001, 0.9197500000000001], "iscrowd": 0, "area": 24199.055050000003, "rle": {"size": [360, 640], "counts": "fSo07k:>QN]OSH0^OQ1Y8ROPHg1l7\\NiGk1X8b01O100O000001N1000000O100O10000OO2O1N101N101N10O0011N2N2O1N3N1N3N1N3N1N2O2M2O2N1N3N11O001O001O1O001O1O1O1O1O1O001O1O6J6J5K1O2N2O1N2N2N2N2O0O2N2N1O2O1N2N2N1O2O1N2N2N2N3M4L5K3nJQLe2R4RMTLn2o3lLTLT3o3eLUL[3n3_LULa3m3\\LTLd3o3WLSLi3P4SLQLm3R4nKPLR4S4iKPLV4S4dKPL\\4T4]KoKd4U4UKmKk4X51O001O001O0O2O0O2O001eMnJTNOZNT5`3QKSNM\\NS5^3TKUMOJMAQ5^3XKPMONW5Y2]JiMb0J01R5\\2_JdMi0DJ:o4^2aJ_Me1Ok3b2dJXMd7h2>01O00001O001O001O001O00001O00001O00000000001O0000000000O100O100O100O1O10O010O010O010O010O010O010O100O100N2O2N[GQNS8c2O2VJlL\\3U3cLkL]3W3`LjL`3W3^LkLa3W3]LiLc3Y3YLjLg3X3ULiLk3Z3PLhLP4\\3jKgLU4\\3gKeLY4^3bKdL^4_3^KbLb4a3fJTMZ5]41O000SM`J2`5KeJ2]5JhJ3Y5JlJ1X5LlJLZ52jJIZ56hJGY59iJDY5;iJnMZOU1n5n0`KgNb4X1fK_N\\4`1f201O1002N2N3M2N2N3M3M2N3N2cIYMR4j2fKZM\\4h2]K[Md4h2UK[Mm4e2mJ`MS5`2iJcMY5]2`JhMb5Z2VJjMk5_2fIfM\\6b32N1O2N2nLaIW1`6eMeJY2W7SN`Gn0b8iNjG;@Ki8ASH4C7[:L5J4MoP]3"}, "label": "the grey horse with its whole body visible"}]} {"id": 234699, "image": "COCO_train2014_000000234699.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black grapes on a bread\"."}], "task": "refering", "answer_template": "The \"black grapes on a bread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 19, 20, 21, 22, 23, 36, 37, 38, 39, 40, 41, 54, 55, 56, 57, 58, 73, 74, 75], "bbox": [0.0381, 0.00936, 0.30482000000000004, 0.3622933333333333], "iscrowd": 0, "area": 13024.2737, "rle": {"size": [375, 500], "counts": "TQ73`;8I6I8H7J7H7I8J4L4L3M4M3L3M4M2M4L3M4M2M2O2O0O1O2N1O2O0O1O2N1O2O0O1O2N1O2O0O1O2N1O2O00001N100000O01000000O1000000O1000000O1000O10O1000000O1000000O10O1O00100O1O1O100O2N1O100O2N1O101N1O1O101N1O100O2N100O1O2O0O1O1O2O0O1O2O0O1O1O2M2N2N2N3M4L4L4L4L4L4L4L4L4L6J7I6J7H7J7I[\\o3"}, "label": "black grapes on a bread"}]} {"id": 234699, "image": "COCO_train2014_000000234699.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a powdered blackberry donut with real blackberries on top\"."}], "task": "refering", "answer_template": "The \"a powdered blackberry donut with real blackberries on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 19, 20, 21, 22, 23, 36, 37, 38, 39, 40, 41, 54, 55, 56, 57, 58, 73, 74, 75], "bbox": [0.0381, 0.00936, 0.30482000000000004, 0.3622933333333333], "iscrowd": 0, "area": 13024.2737, "rle": {"size": [375, 500], "counts": "TQ73`;8I6I8H7J7H7I8J4L4L3M4M3L3M4M2M4L3M4M2M2O2O0O1O2N1O2O0O1O2N1O2O0O1O2N1O2O0O1O2N1O2O00001N100000O01000000O1000000O1000000O1000O10O1000000O1000000O10O1O00100O1O1O100O2N1O100O2N1O101N1O1O101N1O100O2N100O1O2O0O1O1O2O0O1O2O0O1O1O2M2N2N2N3M4L4L4L4L4L4L4L4L4L6J7I6J7H7J7I[\\o3"}, "label": "a powdered blackberry donut with real blackberries on top"}]} {"id": 529624, "image": "COCO_train2014_000000529624.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurry male official wearing white pants with arms behind his back\"."}], "task": "refering", "answer_template": "The \"a blurry male official wearing white pants with arms behind his back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 176, 177, 178, 199, 200, 201, 202, 222, 223, 224, 225, 245, 246, 247, 268, 269, 270, 291, 292, 293, 313, 314, 315, 316, 317, 337, 339, 340], "bbox": [0.643734375, 0.2569375, 0.7986093750000001, 0.63253125], "iscrowd": 0, "area": 15217.077900000002, "rle": {"size": [640, 640], "counts": "akQ84ic06I7XLDbCc0VHdBIlNf0\\>E^Bk1]=ZNZBo1b=SNVBT2h=PNoAW2Q>m1N100O101ON1N3M2O2L4L3M4L3M4KESKmAi4f=VKXB60a4i=`KoA46Y4l=XLPBf3T>[LjAb3Y>o04K5L3M4L33N2N2M4M2N2M3O2M2N2O2M2O1N3M2O1O1O2O0O1O2OZJ^Bl4a=iJ]BL3]5X>1O2M2O1N3N1O1N2O2NO0O2O001O1O001O1O0`LPBP1Q>mNTBP1m=lNZBZN@W2V>]O`BWN_OX2R>_OcBVN_OW2o=AfBTN_OW2l=D\\C6h in this image.", "objects": [{"patches": [153, 154, 176, 177, 178, 199, 200, 201, 202, 222, 223, 224, 225, 245, 246, 247, 268, 269, 270, 291, 292, 293, 313, 314, 315, 316, 317, 337, 339, 340], "bbox": [0.643734375, 0.2569375, 0.7986093750000001, 0.63253125], "iscrowd": 0, "area": 15217.077900000002, "rle": {"size": [640, 640], "counts": "akQ84ic06I7XLDbCc0VHdBIlNf0\\>E^Bk1]=ZNZBo1b=SNVBT2h=PNoAW2Q>m1N100O101ON1N3M2O2L4L3M4L3M4KESKmAi4f=VKXB60a4i=`KoA46Y4l=XLPBf3T>[LjAb3Y>o04K5L3M4L33N2N2M4M2N2M3O2M2N2O2M2O1N3M2O1O1O2O0O1O2OZJ^Bl4a=iJ]BL3]5X>1O2M2O1N3N1O1N2O2NO0O2O001O1O001O1O0`LPBP1Q>mNTBP1m=lNZBZN@W2V>]O`BWN_OX2R>_OcBVN_OW2o=AfBTN_OW2l=D\\C6h in this image.", "objects": [{"patches": [194, 195, 196, 217, 218, 219, 240, 241, 242, 243, 244, 263, 264, 265, 266, 267, 286, 287, 288, 289, 290, 309, 310, 311, 312, 313, 332, 333, 334, 335, 336, 337], "bbox": [0.42921875, 0.3460625, 0.660671875, 0.633703125], "iscrowd": 0, "area": 16707.166, "rle": {"size": [640, 640], "counts": "fV\\53kc03M3L5L3M4L3K5K6J5K6J4L2O1R_OaN[?b1`@dN\\?^1`@hN\\?Z1_@mN]?V1^@PO]?S1_@SO]?o0^@XO\\?l0`@ZO[?j0`@\\O_?c0\\@Dd?\\2O100QOjKSBW4l=PLnAP4Q>VLiAk3V>\\L_AXO2\\4^>Q1O1O1O100O10000O100O10000O100O11O000000001O0000001O0000001O0000001O0000001O00001O1O2N2N2N2N1O4Lg0YO8H1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O0000000000000XOg_OiMY`0Q2Q@kMo?P2Y@mMg?m1c@oM]?l1k@QNU?i1d1J6005K5K4L5K5K5K4N3M3M3L3N3M1ON2M6K6J6J6J6J6J6J6J5K4L3M3M3M3M4L3M3M3L4M4L3M3M3M]WW4"}, "label": "the girl who is smiling"}]} {"id": 529624, "image": "COCO_train2014_000000529624.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis player facing the camera and shaking another player ' s hand\"."}], "task": "refering", "answer_template": "The \"a tennis player facing the camera and shaking another player ' s hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [194, 195, 196, 217, 218, 219, 240, 241, 242, 243, 244, 263, 264, 265, 266, 267, 286, 287, 288, 289, 290, 309, 310, 311, 312, 313, 332, 333, 334, 335, 336, 337], "bbox": [0.42921875, 0.3460625, 0.660671875, 0.633703125], "iscrowd": 0, "area": 16707.166, "rle": {"size": [640, 640], "counts": "fV\\53kc03M3L5L3M4L3K5K6J5K6J4L2O1R_OaN[?b1`@dN\\?^1`@hN\\?Z1_@mN]?V1^@PO]?S1_@SO]?o0^@XO\\?l0`@ZO[?j0`@\\O_?c0\\@Dd?\\2O100QOjKSBW4l=PLnAP4Q>VLiAk3V>\\L_AXO2\\4^>Q1O1O1O100O10000O100O10000O100O11O000000001O0000001O0000001O0000001O0000001O00001O1O2N2N2N2N1O4Lg0YO8H1O1O1O1O1O2N1O1O1O1O2N1O1O1O1O1O0000000000000XOg_OiMY`0Q2Q@kMo?P2Y@mMg?m1c@oM]?l1k@QNU?i1d1J6005K5K4L5K5K5K4N3M3M3L3N3M1ON2M6K6J6J6J6J6J6J6J5K4L3M3M3M3M4L3M3M3L4M4L3M3M3M]WW4"}, "label": "a tennis player facing the camera and shaking another player ' s hand"}]} {"id": 226527, "image": "COCO_train2014_000000226527.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small giraffe looking looking at a herd of horned animals\"."}], "task": "refering", "answer_template": "The \"a small giraffe looking looking at a herd of horned animals\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 223, 224, 245, 246, 267, 268, 269, 270, 290, 291, 292, 313, 314, 315, 336, 337, 338], "bbox": [0.6286562499999999, 0.5080208333333334, 0.77359375, 0.8743124999999999], "iscrowd": 0, "area": 6774.672050000005, "rle": {"size": [480, 640], "counts": "hYm55g>5]Ob0L5N2M3N1N3M3N0O2aN_1O1O1N2O1N2O001N2O1O1N2OAcD]M];\\2PE]MP;]2l0J7M2M4M2M4M2M3N3N1O2O0O1O2O02O0N3M2N2N3N1N2N3aDYNa9i1QFfNm9[1`EYO]:f2N3M3M2N3M3M3M2O2]OYKiFj4V9XKdFl4[9WK_Fn4`990N2O0O2ZNYGQMi8P1XF7_;GjDOY;NQEHR;5XEAk:=hEnN[:P1l1M3M3N2M3N200O3M3N1N2N2O0N3M3N2M2N3NRgS2"}, "label": "a small giraffe looking looking at a herd of horned animals"}]} {"id": 226527, "image": "COCO_train2014_000000226527.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"giraffee chaild was in forest\"."}], "task": "refering", "answer_template": "The \"giraffee chaild was in forest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [200, 201, 223, 224, 245, 246, 267, 268, 269, 270, 290, 291, 292, 313, 314, 315, 336, 337, 338], "bbox": [0.6286562499999999, 0.5080208333333334, 0.77359375, 0.8743124999999999], "iscrowd": 0, "area": 6774.672050000005, "rle": {"size": [480, 640], "counts": "hYm55g>5]Ob0L5N2M3N1N3M3N0O2aN_1O1O1N2O1N2O001N2O1O1N2OAcD]M];\\2PE]MP;]2l0J7M2M4M2M4M2M3N3N1O2O0O1O2O02O0N3M2N2N3N1N2N3aDYNa9i1QFfNm9[1`EYO]:f2N3M3M2N3M3M3M2O2]OYKiFj4V9XKdFl4[9WK_Fn4`990N2O0O2ZNYGQMi8P1XF7_;GjDOY;NQEHR;5XEAk:=hEnN[:P1l1M3M3N2M3N200O3M3N1N2N2O0N3M3N2M2N3NRgS2"}, "label": "giraffee chaild was in forest"}]} {"id": 226527, "image": "COCO_train2014_000000226527.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mother giraffe nursing her young\"."}], "task": "refering", "answer_template": "The \"a mother giraffe nursing her young\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 142, 143, 165, 188, 210, 211, 233, 234, 235, 256, 257, 279, 280, 302], "bbox": [0.1324375, 0.3379375, 0.238078125, 0.8271875000000001], "iscrowd": 0, "area": 5038.779100000002, "rle": {"size": [480, 640], "counts": "mSX17d>:F;E:H9J6I7J3L3N3M2M4M2M3N3Ld1]N0O100OD=I6M4K4M4K5K4N3M2O2M2O2M2O1N2O0O101N1001dMTNeGk1]8XN^Gi1c8WNZGj1h8VNnFS2S9mMcF\\2^9a13N1N2O100`MkE6T:B^F5`9EnF3Q9mNVHl0j7VO\\Ha0i7\\O]H=g7@`H8d7FaH3b7KeHL`71i3O001O00000000cbT7"}, "label": "a mother giraffe nursing her young"}]} {"id": 226527, "image": "COCO_train2014_000000226527.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"it is a giraffe in the forground facing forward\"."}], "task": "refering", "answer_template": "The \"it is a giraffe in the forground facing forward\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 36, 37, 38, 59, 60, 61, 82, 83, 84, 104, 105, 106, 126, 127, 128, 148, 149, 150, 151, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 261, 262, 263, 264, 265, 284, 285, 287, 288, 308, 310, 311, 330, 331, 333, 334, 353, 354, 356, 357, 377, 379, 380], "bbox": [0.35225, 0.035958333333333335, 0.6809062499999999, 0.9842708333333333], "iscrowd": 0, "area": 29402.818800000015, "rle": {"size": [480, 640], "counts": "fYZ32j>2N2N2N2N2QMIUG9f84RGLj8?oFA6CX6V1ZIXO?FT6W1VISOg0IP6X1TIoNn0Kl5Y1PInNT1Mh5Z1nHiN\\1Od5[1kHfNb10b5^1gHcNg10a5i1ZHXNW2O_5T2mGnMd2O^5R2lGQNh2L]5T2oGiMa28_5Q2THaMW2e0e5k1WHZMo1R1i5f1_JaNa5a1VJfNi5a1fIiNZ6^1hHXOX7W4O10O0N2M5L5J5K6J5K6J5O27H9H8G9H8G6K^MXH\\L1b0a7Q3eHVL1i0U7o2PIRL2o0h6o2\\IkK4U1[6n2gIhK4W1R6Q3VKjLj4T3]KfLc4Z3cK`L^4^3jK[LV4e3PLVLo3i3d300O101N100N2O1N2N2O2M2N2O1N21O0010O0001N1O2N1O2N1O1O2N5WIlK^2Y4VMRLf2S4nLXLn2m3\\KcK[Ol0U5e3]KfKoNP1`5_3\\KiKeNS1k5Y3[KdMb4`2[KcMa4a2[KcMa4a2[KcMa4a2[KcMa4a2\\KaMa4c2[K`Mc4c2YK_Mf4d2WK]Mh4f2TKgKh0jNT4a5PKcKW1fNg3j5nJ^Kc1dN^3b7jLZHU3c7SMYHm2k7Q2=C=C=aJ]Gi3m9O1O1O100O1O1OiKPL`MP4W2bLdKAFl3c4PMmJI in this image.", "objects": [{"patches": [14, 15, 36, 37, 38, 59, 60, 61, 82, 83, 84, 104, 105, 106, 126, 127, 128, 148, 149, 150, 151, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 261, 262, 263, 264, 265, 284, 285, 287, 288, 308, 310, 311, 330, 331, 333, 334, 353, 354, 356, 357, 377, 379, 380], "bbox": [0.35225, 0.035958333333333335, 0.6809062499999999, 0.9842708333333333], "iscrowd": 0, "area": 29402.818800000015, "rle": {"size": [480, 640], "counts": "fYZ32j>2N2N2N2N2QMIUG9f84RGLj8?oFA6CX6V1ZIXO?FT6W1VISOg0IP6X1TIoNn0Kl5Y1PInNT1Mh5Z1nHiN\\1Od5[1kHfNb10b5^1gHcNg10a5i1ZHXNW2O_5T2mGnMd2O^5R2lGQNh2L]5T2oGiMa28_5Q2THaMW2e0e5k1WHZMo1R1i5f1_JaNa5a1VJfNi5a1fIiNZ6^1hHXOX7W4O10O0N2M5L5J5K6J5K6J5O27H9H8G9H8G6K^MXH\\L1b0a7Q3eHVL1i0U7o2PIRL2o0h6o2\\IkK4U1[6n2gIhK4W1R6Q3VKjLj4T3]KfLc4Z3cK`L^4^3jK[LV4e3PLVLo3i3d300O101N100N2O1N2N2O2M2N2O1N21O0010O0001N1O2N1O2N1O1O2N5WIlK^2Y4VMRLf2S4nLXLn2m3\\KcK[Ol0U5e3]KfKoNP1`5_3\\KiKeNS1k5Y3[KdMb4`2[KcMa4a2[KcMa4a2[KcMa4a2[KcMa4a2\\KaMa4c2[K`Mc4c2YK_Mf4d2WK]Mh4f2TKgKh0jNT4a5PKcKW1fNg3j5nJ^Kc1dN^3b7jLZHU3c7SMYHm2k7Q2=C=C=aJ]Gi3m9O1O1O100O1O1OiKPL`MP4W2bLdKAFl3c4PMmJI in this image.", "objects": [{"patches": [123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219], "bbox": [0.32287499999999997, 0.36852459016393446, 0.58146875, 0.6704449648711943], "iscrowd": 0, "area": 12425.404149999995, "rle": {"size": [427, 640], "counts": "Xbf23U=3L4M3M3M3M3M3M3L4M4L3M3N2M3N2M3N2N2M3N2M1000O010O010O11N3N1N3N2N1N3N2M3N1O2N21N10000O1000000O10000O101O0O10000O1001O3M4L3M3M4M2M3M4L3M4L3M3M4LhN`FfN_9Y1dFeN\\9Z1hFcNX9\\1kFcNU9[1nFcNR9]1QG`No8_1TG_Nl8`1WG_Nh8`1\\G]Nd8b1_G\\Na8d1aGZN_8e1eGXN[8g1hGXNW8g1lGWNT8h1PHUNP8j1SHTNn7k1THTNk7n1VHoMj7S2VHkMj7W2d12N3M2N>B3M001O10O01O1O001O1O001O100O00HmEYMR:f2UFUMj9j2 in this image.", "objects": [{"patches": [123, 124, 125, 126, 127, 146, 147, 148, 149, 150, 151, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 214, 215, 216, 217, 218, 219], "bbox": [0.32287499999999997, 0.36852459016393446, 0.58146875, 0.6704449648711943], "iscrowd": 0, "area": 12425.404149999995, "rle": {"size": [427, 640], "counts": "Xbf23U=3L4M3M3M3M3M3M3L4M4L3M3N2M3N2M3N2N2M3N2M1000O010O010O11N3N1N3N2N1N3N2M3N1O2N21N10000O1000000O10000O101O0O10000O1001O3M4L3M3M4M2M3M4L3M4L3M3M4LhN`FfN_9Y1dFeN\\9Z1hFcNX9\\1kFcNU9[1nFcNR9]1QG`No8_1TG_Nl8`1WG_Nh8`1\\G]Nd8b1_G\\Na8d1aGZN_8e1eGXN[8g1hGXNW8g1lGWNT8h1PHUNP8j1SHTNn7k1THTNk7n1VHoMj7S2VHkMj7W2d12N3M2N>B3M001O10O01O1O001O1O001O100O00HmEYMR:f2UFUMj9j2 in this image.", "objects": [{"patches": [50, 53, 74, 75, 76, 77, 78, 79, 80, 81, 97, 98, 99, 100, 101, 102, 103, 104, 105, 120, 121, 122, 123, 124, 125, 126, 127, 128, 145, 146, 150, 151, 168, 169, 174, 196, 197], "bbox": [0.18848437499999998, 0.16278688524590165, 0.601359375, 0.6089695550351288], "iscrowd": 0, "area": 18874.332450000005, "rle": {"size": [427, 640], "counts": "c[c11[=000O1000000O10000O10000O10000O010[COo;1nC4P4Lo:FjDd0KN\\;^OfD`1Y;9100O01O0O100O010O100O100O1O100O1000O10000000001O000000000O10001O0000000001O2N2O2M2N2N2N2O1N2O2M2N2N2J6K5K5K6J5K56J7IN2N2N3M2M3N1O2N2M2OXLjF]3T9^LQGc3[92N1O2N1O00K5M3N2M3N2M3N2M3N2M3N2M3N2N2M3N2M3N2M3N2M3N1N3N2N200000000O10000000000O100000000O100000000000000000000000000001O0000000000000O100000001O000000000000000001O01O00000001O000000001O01O000001O00000000001O01O000001O000000001O01O001O10O01O1O001O10O01O1O00100O001O1O010O1O001O1O10O0101WFmMW8S2eGZNQ8h1kGaNo7`1lGfNR8Z1jGmNR8U1iGQOU8o0hGQO[8P1`GnNf8R1VGQOj8Q1RGQOo8f20O010O1000O0100O010O100^Oa0mNT1mNS1lNhWZ3"}, "label": "the large black cow"}]} {"id": 349408, "image": "COCO_train2014_000000349408.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a larger bull with a smaller brown one in front of it\"."}], "task": "refering", "answer_template": "The \"a larger bull with a smaller brown one in front of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 53, 74, 75, 76, 77, 78, 79, 80, 81, 97, 98, 99, 100, 101, 102, 103, 104, 105, 120, 121, 122, 123, 124, 125, 126, 127, 128, 145, 146, 150, 151, 168, 169, 174, 196, 197], "bbox": [0.18848437499999998, 0.16278688524590165, 0.601359375, 0.6089695550351288], "iscrowd": 0, "area": 18874.332450000005, "rle": {"size": [427, 640], "counts": "c[c11[=000O1000000O10000O10000O10000O010[COo;1nC4P4Lo:FjDd0KN\\;^OfD`1Y;9100O01O0O100O010O100O100O1O100O1000O10000000001O000000000O10001O0000000001O2N2O2M2N2N2N2O1N2O2M2N2N2J6K5K5K6J5K56J7IN2N2N3M2M3N1O2N2M2OXLjF]3T9^LQGc3[92N1O2N1O00K5M3N2M3N2M3N2M3N2M3N2M3N2N2M3N2M3N2M3N2M3N1N3N2N200000000O10000000000O100000000O100000000000000000000000000001O0000000000000O100000001O000000000000000001O01O00000001O000000001O01O000001O00000000001O01O000001O000000001O01O001O10O01O1O001O10O01O1O00100O001O1O010O1O001O1O10O0101WFmMW8S2eGZNQ8h1kGaNo7`1lGfNR8Z1jGmNR8U1iGQOU8o0hGQO[8P1`GnNf8R1VGQOj8Q1RGQOo8f20O010O1000O0100O010O100^Oa0mNT1mNS1lNhWZ3"}, "label": "a larger bull with a smaller brown one in front of it"}]} {"id": 316671, "image": "COCO_train2014_000000316671.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass of red wine being held by a woman\"."}], "task": "refering", "answer_template": "The \"a glass of red wine being held by a woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 212, 213, 214, 215, 229, 230, 231, 232, 246, 247, 248, 249, 264, 265, 281, 282, 298, 299, 314, 315, 331, 332], "bbox": [0.48539583333333336, 0.47640625, 0.6921250000000001, 0.8471875000000001], "iscrowd": 0, "area": 11712.237000000001, "rle": {"size": [640, 480], "counts": "]oa43jc06I8SAAn9e0nE\\Oo9j0lEYOo9n0mETOo9R1mEoNP:X1kEjNP:]1mEcNP:c1lE_NP:g1lEZNP:m1mETNo9S2lEnMQ:X2kEjMQ:\\2kEeMQ:`2nEaMn9b2QF`Mk9d2SF]Mi9h2VFXMj9i2TFYMk9h2SFYMm9g2SFYMm9h2QFYMo9h2oEYMQ:h2mEZMR:g2mEYMS:g2lEZMT:g2jEZMV:g2iEZMV:g2hEZMX:g2fEZMZ:f2fEZMZ:g2dEZM\\:g2bE[M]:f2aE[M_:f2cEWM]:i2fETMZ:m2hEQMW:P3kEmLU:T3mEiLS:X3oEeLQ:[3RFcLm9^3UF_Lk9b5VD\\Jk;c5XDZJh;f5[DWJe;i5^DUJa;k5bDRJ^;n5eDoI[;Q6Z10000E;@`0@`0@`0G900O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O3M7I7I6K6H8D;E in this image.", "objects": [{"patches": [195, 196, 197, 198, 212, 213, 214, 215, 229, 230, 231, 232, 246, 247, 248, 249, 264, 265, 281, 282, 298, 299, 314, 315, 331, 332], "bbox": [0.48539583333333336, 0.47640625, 0.6921250000000001, 0.8471875000000001], "iscrowd": 0, "area": 11712.237000000001, "rle": {"size": [640, 480], "counts": "]oa43jc06I8SAAn9e0nE\\Oo9j0lEYOo9n0mETOo9R1mEoNP:X1kEjNP:]1mEcNP:c1lE_NP:g1lEZNP:m1mETNo9S2lEnMQ:X2kEjMQ:\\2kEeMQ:`2nEaMn9b2QF`Mk9d2SF]Mi9h2VFXMj9i2TFYMk9h2SFYMm9g2SFYMm9h2QFYMo9h2oEYMQ:h2mEZMR:g2mEYMS:g2lEZMT:g2jEZMV:g2iEZMV:g2hEZMX:g2fEZMZ:f2fEZMZ:g2dEZM\\:g2bE[M]:f2aE[M_:f2cEWM]:i2fETMZ:m2hEQMW:P3kEmLU:T3mEiLS:X3oEeLQ:[3RFcLm9^3UF_Lk9b5VD\\Jk;c5XDZJh;f5[DWJe;i5^DUJa;k5bDRJ^;n5eDoI[;Q6Z10000E;@`0@`0@`0G900O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O3M7I7I6K6H8D;E in this image.", "objects": [{"patches": [60, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 148, 160, 161, 162, 163, 164, 165, 166, 176, 177, 178, 179, 180, 181, 182, 183, 184, 193, 194, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 217, 218, 227, 228, 229, 230, 231, 232, 233, 234, 244, 245, 246, 247, 248, 249, 250, 251, 261, 262, 263, 264, 265, 266, 267, 268, 269, 278, 279, 280, 281, 282, 283, 284], "bbox": [0.3565625, 0.17078125, 0.8569375000000001, 0.72584375], "iscrowd": 0, "area": 59848.249749999995, "rle": {"size": [640, 480], "counts": "SY[32lc07H8I6I6K5J6K5J7D9WN_Nh@j1g>]OY@n0b?j1M4M2M4L3M4M2M4L3N3K4K6J5L4K6K4K6J5L5J5XOi0L3N3L3N3L3N2M3M3J6J6J6J6J6J6J6J6nNR1O1O00100O010000O1000000O1000O01000000O10000O10O10O1000000O10000O10O1000O10000O10000O10O11O0O10000O1000001O000000000000001O0000000000001O0000000000001O00000000001O0000000000001O00000000002N1O2N2N1O2N1O3M4L3M4L3M2N3M2N3M2N3M2N3L3N3M2N3M2N3M2N3M2N3M4K5L4L5K4L4L5K4L4L2N1O2M3N1O2N1O2N1N3N1O001N2O1O001N2O1O001N2O001O1O1N101O1O0O2O1YNPB\\MQ>a2aBoL_=o2SC`Lo<\\3YC_Lh<^3YCcLj in this image.", "objects": [{"patches": [60, 75, 76, 77, 78, 79, 92, 93, 94, 95, 96, 109, 110, 111, 112, 113, 114, 126, 127, 128, 129, 130, 131, 143, 144, 145, 146, 147, 148, 160, 161, 162, 163, 164, 165, 166, 176, 177, 178, 179, 180, 181, 182, 183, 184, 193, 194, 195, 196, 197, 198, 199, 200, 201, 210, 211, 212, 213, 214, 215, 216, 217, 218, 227, 228, 229, 230, 231, 232, 233, 234, 244, 245, 246, 247, 248, 249, 250, 251, 261, 262, 263, 264, 265, 266, 267, 268, 269, 278, 279, 280, 281, 282, 283, 284], "bbox": [0.3565625, 0.17078125, 0.8569375000000001, 0.72584375], "iscrowd": 0, "area": 59848.249749999995, "rle": {"size": [640, 480], "counts": "SY[32lc07H8I6I6K5J6K5J7D9WN_Nh@j1g>]OY@n0b?j1M4M2M4L3M4M2M4L3N3K4K6J5L4K6K4K6J5L5J5XOi0L3N3L3N3L3N2M3M3J6J6J6J6J6J6J6J6nNR1O1O00100O010000O1000000O1000O01000000O10000O10O10O1000000O10000O10O1000O10000O10000O10O11O0O10000O1000001O000000000000001O0000000000001O0000000000001O00000000001O0000000000001O00000000002N1O2N2N1O2N1O3M4L3M4L3M2N3M2N3M2N3M2N3L3N3M2N3M2N3M2N3M2N3M4K5L4L5K4L4L5K4L4L2N1O2M3N1O2N1O2N1N3N1O001N2O1O001N2O1O001N2O001O1O1N101O1O0O2O1YNPB\\MQ>a2aBoL_=o2SC`Lo<\\3YC_Lh<^3YCcLj in this image.", "objects": [{"patches": [171, 172, 187, 188, 189, 190, 204, 205, 206, 207, 221, 222, 223, 224, 238, 239, 240, 241, 256, 257, 273, 274, 290, 291, 292, 306, 307, 308, 309], "bbox": [0.006333333333333333, 0.47484374999999995, 0.23162500000000003, 0.815984375], "iscrowd": 0, "area": 11843.8243, "rle": {"size": [640, 480], "counts": "lV2:Yc0a0D in this image.", "objects": [{"patches": [171, 172, 187, 188, 189, 190, 204, 205, 206, 207, 221, 222, 223, 224, 238, 239, 240, 241, 256, 257, 273, 274, 290, 291, 292, 306, 307, 308, 309], "bbox": [0.006333333333333333, 0.47484374999999995, 0.23162500000000003, 0.815984375], "iscrowd": 0, "area": 11843.8243, "rle": {"size": [640, 480], "counts": "lV2:Yc0a0D in this image.", "objects": [{"patches": [52, 53, 54, 55, 75, 76, 77, 78, 79, 100, 101, 102, 103, 126, 127, 128, 151, 152, 153, 176, 177, 178, 201], "bbox": [0.28609375, 0.12782201405152224, 0.7890312500000001, 0.5691100702576112], "iscrowd": 0, "area": 8526.207799999995, "rle": {"size": [427, 640], "counts": "o[\\25U=c0^O6I2O0O2O1N101N2O0O2O1N1O2O1N101N101N2O001O1O010O1O010O001O10O01O10O01O1O010O1O010O00100O001O010O0001O000010O000001O000010O01O1O1O1O1O100O1O1O1O1O100O001O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O10O0100O100O100O100O100O100O100O100O100O010O010O10O010O0100O010O010O10O01O010O10O0100O010O010O10O010O0100O010O010O10O010O0100O010O10O010O0100O010O010O10O010O0100O010O00100O010O10O010O0100O010O010O10O010O0100O010O010O10O0100O010O010O10O010O0100O010O010O10O010O01O10O010O010O010O010O010O0010O010O010O010O010O010O010O010O010O01O010O010O010O010O010O010O010O01O01O010O010O010O010O010O010O010O010O010O0010O010O010O010O0100O100O101N100O1OO2N1N3N1O2N1N3N1O2NXRh1"}, "label": "a small spatula which is above all spatula"}]} {"id": 79111, "image": "COCO_train2014_000000079111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"smallest wooden spoon of the group\"."}], "task": "refering", "answer_template": "The \"smallest wooden spoon of the group\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 54, 55, 75, 76, 77, 78, 79, 100, 101, 102, 103, 126, 127, 128, 151, 152, 153, 176, 177, 178, 201], "bbox": [0.28609375, 0.12782201405152224, 0.7890312500000001, 0.5691100702576112], "iscrowd": 0, "area": 8526.207799999995, "rle": {"size": [427, 640], "counts": "o[\\25U=c0^O6I2O0O2O1N101N2O0O2O1N1O2O1N101N101N2O001O1O010O1O010O001O10O01O10O01O1O010O1O010O00100O001O010O0001O000010O000001O000010O01O1O1O1O1O100O1O1O1O1O100O001O1O1O1O100O1O1O1O1O1O100O1O1O1O1O100O10O0100O100O100O100O100O100O100O100O100O010O010O10O010O0100O010O010O10O01O010O10O0100O010O010O10O010O0100O010O010O10O010O0100O010O10O010O0100O010O010O10O010O0100O010O00100O010O10O010O0100O010O010O10O010O0100O010O010O10O0100O010O010O10O010O0100O010O010O10O010O01O10O010O010O010O010O010O0010O010O010O010O010O010O010O010O010O01O010O010O010O010O010O010O010O01O01O010O010O010O010O010O010O010O010O010O0010O010O010O010O0100O100O101N100O1OO2N1N3N1O2N1N3N1O2NXRh1"}, "label": "smallest wooden spoon of the group"}]} {"id": 79111, "image": "COCO_train2014_000000079111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooden bowls holding apples and five wooden spoons\"."}], "task": "refering", "answer_template": "The \"wooden bowls holding apples and five wooden spoons\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 27, 28, 29, 30, 31, 32, 33, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 281, 282, 283, 284, 285, 286, 287, 288, 305, 306, 307, 308, 309], "bbox": [0.10495312500000001, 0.0539344262295082, 0.629703125, 0.8921311475409838], "iscrowd": 0, "area": 92562.51460000001, "rle": {"size": [427, 640], "counts": "^Sl0e0b in this image.", "objects": [{"patches": [7, 27, 28, 29, 30, 31, 32, 33, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 281, 282, 283, 284, 285, 286, 287, 288, 305, 306, 307, 308, 309], "bbox": [0.10495312500000001, 0.0539344262295082, 0.629703125, 0.8921311475409838], "iscrowd": 0, "area": 92562.51460000001, "rle": {"size": [427, 640], "counts": "^Sl0e0b in this image.", "objects": [{"patches": [149, 150, 173, 174, 197, 198, 199, 222, 223, 246, 247, 270, 271, 272, 294, 295, 296, 319, 320, 343], "bbox": [0.48410937499999995, 0.3957845433255269, 0.9605312500000001, 0.9637704918032787], "iscrowd": 0, "area": 5913.363349999997, "rle": {"size": [427, 640], "counts": "l]Q42Y=1N2O1O1O2O0O1O1O1O1O2O0O1O1O101N010O10000O10000O10O10O10000O010O1001N3N3M2N1O0O1000O10O1000O1000O0100000O01000000O0100000O01000O1000O10O1000O10O100000O010000O0100000O10O1000O1000O10O1000O01000000O0100000O0100000O10O10O1000O01000000O01000O10O10O10000O01000O10O10O10000O0100000O10O1000O1000000O10O1000O1000000O10O1000O1000000O1000O10O1000000O1000O10O100000O10O1000000O0100000O1000O10O1000000O0100000O10O1000O100000O01000000O10O1000O100000O0100000O10O1000O10O1000O10O1000O10O10O10O1000O0100O10O10O1000O010000O01000O10O01000O010000O0101O000O1000000O1000000O1000000O2O000O1000000O1000000O1000000N2O2N1O1O1O1O3M5KjP:"}, "label": "the handles of several wooden spoons"}]} {"id": 79111, "image": "COCO_train2014_000000079111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the stick sticking out the farthest\"."}], "task": "refering", "answer_template": "The \"the stick sticking out the farthest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 150, 173, 174, 197, 198, 199, 222, 223, 246, 247, 270, 271, 272, 294, 295, 296, 319, 320, 343], "bbox": [0.48410937499999995, 0.3957845433255269, 0.9605312500000001, 0.9637704918032787], "iscrowd": 0, "area": 5913.363349999997, "rle": {"size": [427, 640], "counts": "l]Q42Y=1N2O1O1O2O0O1O1O1O1O2O0O1O1O101N010O10000O10000O10O10O10000O010O1001N3N3M2N1O0O1000O10O1000O1000O0100000O01000000O0100000O01000O1000O10O1000O10O100000O010000O0100000O10O1000O1000O10O1000O01000000O0100000O0100000O10O10O1000O01000000O01000O10O10O10000O01000O10O10O10000O0100000O10O1000O1000000O10O1000O1000000O10O1000O1000000O1000O10O1000000O1000O10O100000O10O1000000O0100000O1000O10O1000000O0100000O10O1000O100000O01000000O10O1000O100000O0100000O10O1000O10O1000O10O1000O10O10O10O1000O0100O10O10O1000O010000O01000O10O01000O010000O0101O000O1000000O1000000O1000000O2O000O1000000O1000000O1000000N2O2N1O1O1O1O3M5KjP:"}, "label": "the stick sticking out the farthest"}]} {"id": 79111, "image": "COCO_train2014_000000079111.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden bowl containing another wooden bowl\"."}], "task": "refering", "answer_template": "The \"a wooden bowl containing another wooden bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 81, 82, 94, 105, 106, 116, 117, 129, 130, 139, 140, 152, 153, 162, 163, 175, 176, 185, 186, 199, 208, 209, 231, 232, 233, 243, 244, 255, 256, 257, 266, 267, 279, 280, 281, 282, 287, 288, 289, 303, 304, 305, 306, 307, 308, 309, 310, 311, 328, 329, 330, 331, 332, 333], "bbox": [0.058750000000000004, 0.1393208430913349, 0.689484375, 0.9629742388758782], "iscrowd": 0, "area": 21029.264249999997, "rle": {"size": [427, 640], "counts": "RR`0=fF;E9G6I8I6K6J5K6J5K6J5K6J5K6K4M4K4L5K5L3L4L4MlLdHZ4j0bIhNV2;[4k0`IjNU29]4m0]IkNV26`4m0[ImNU24b4o0XImNW22d4o0VIoNV2Og4R1RIPOW2Lj4S1oHQOW2Jm4S1lHTOW2Go4U1jHSOX2FQ5U1gHVOX2CS5W1eHVOX2AV5W1bHYOY2]OX5Y1_HYO]:f0cE[O^:d0aE]O_:b0bE^O_:a0`E@a:?_E@c:>]ECd:<\\EDe::[EGf:8ZEHg:6YEKh:4XELi:2WENl:0TE0o:MPE4S;HnD8U;EjD in this image.", "objects": [{"patches": [57, 58, 81, 82, 94, 105, 106, 116, 117, 129, 130, 139, 140, 152, 153, 162, 163, 175, 176, 185, 186, 199, 208, 209, 231, 232, 233, 243, 244, 255, 256, 257, 266, 267, 279, 280, 281, 282, 287, 288, 289, 303, 304, 305, 306, 307, 308, 309, 310, 311, 328, 329, 330, 331, 332, 333], "bbox": [0.058750000000000004, 0.1393208430913349, 0.689484375, 0.9629742388758782], "iscrowd": 0, "area": 21029.264249999997, "rle": {"size": [427, 640], "counts": "RR`0=fF;E9G6I8I6K6J5K6J5K6J5K6J5K6K4M4K4L5K5L3L4L4MlLdHZ4j0bIhNV2;[4k0`IjNU29]4m0]IkNV26`4m0[ImNU24b4o0XImNW22d4o0VIoNV2Og4R1RIPOW2Lj4S1oHQOW2Jm4S1lHTOW2Go4U1jHSOX2FQ5U1gHVOX2CS5W1eHVOX2AV5W1bHYOY2]OX5Y1_HYO]:f0cE[O^:d0aE]O_:b0bE^O_:a0`E@a:?_E@c:>]ECd:<\\EDe::[EGf:8ZEHg:6YEKh:4XELi:2WENl:0TE0o:MPE4S;HnD8U;EjD in this image.", "objects": [{"patches": [123, 147, 148, 149, 172, 173, 196, 197, 220, 221, 244, 245, 268, 269, 292], "bbox": [0.36603125, 0.3802576112412178, 0.73515625, 0.8299531615925059], "iscrowd": 0, "area": 3258.7957000000006, "rle": {"size": [427, 640], "counts": "`gQ31fj00dB010O1O010O1O010O10O01O10O01O10O01O10O01O010O0010O01O10O01O010O0010O01O10O01O01O00010O00010O00010O00010O0010O0010O01O100O1000000O0100000000000000O010000000000000O1000O1000000000O100000O10000000000000000O1000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000O1N2O1N2N2O1N2N2O1N2NeaV2"}, "label": "the bottom most wooden spoon closest to the apples"}]} {"id": 161031, "image": "COCO_train2014_000000161031.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"partially eaten pepperoni pizza in pizza box\"."}], "task": "refering", "answer_template": "The \"partially eaten pepperoni pizza in pizza box\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 24, 25, 26, 27, 28, 29, 30, 31, 43, 44, 45, 46, 47, 48, 49, 50, 61, 62, 63, 64, 65, 66, 67, 68, 69, 79, 80, 81, 82, 83, 84, 85, 86, 87, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 204, 205, 206, 207, 208, 209], "bbox": [0.12572, 0.053933933933933934, 0.89348, 0.9887687687687687], "iscrowd": 0, "area": 79038.05559999999, "rle": {"size": [333, 500], "counts": "jdd0>n99G8H8H4M3M3M3M3M3M3M3M3M2M4M3M3M1O2N1N2O2N1O1O2N1N2O2N1O1O2N1N2O2N1O1O1N3N1O1O2N1O1N2O001O1O001N2O001O001O1N101O1O001O1N101O1O001O0O2O001O1O001N101O1O001N101O001N2O001O0O2O1O000O2O00001N1000001N10001O0O10001O0O101O1ZLkJR2W5lMlJP2V5oMlJo1U5oMoJn1R5QNQKl1Q5RNRKj1P5UNSKh1m4XNVKe1j4[NXKc1h4\\NZKb1g4^NZKa1f4_N[K`1f4_N[K`1e4`N[K_1f4`N\\K_1d4aN]K^1c4bN^K]1b4cN_K[1c4dN^K[1b4eN_KZ1a4eNaKZ1`4eN`KZ1a4fN`KY1`4gNaKX1`4gNaKW1`4hNbKW1_4hNcKV1]4jNdKU1\\4kNeKS1\\4mNeKR1\\4mNfKQ1Z4oNgKP1Y4POhKn0Z4QOhKm0X4SOiKl0W4TOjKk0V4UOkKi0W4VOkKh0U4XOlKg0T4YOmKf0S4ZOnKe0S4ZOoKc0R4]OoKb0Q4^OPLa0Q4^OQL`0o3@RL>o3BRL=n3CSLdL@]3`0fL\\O[3d0hLYOX3g0W300000000O100000001N100000000O100000000O10000001O001O00001O00001O00001O001O00001O000O2O0O100O2O0O100O100O100O101N100O100O100O100O2O0O100O100O100O100O10000O2O0O100O100O100O100O100O2O0O101N101N100O2O0O2O0O101N101N100O2O0O2O0O101O1N101N2O1N101N2O0O2O1N101N2O0O2O1N2O1N2O1N2O1N2O0O2O1N2O0O2O0O2O0O2O0O2O001N101N101N2O0O2O0O2O2M2O1N2O1N2O1N3N1N2O1N2N2N3M2N3M3L4M2N3M3M3M3M3M2N3M3M3M4L3M3M3M3I8G8H8H8J6M4M4L4L3M4G;D>C=B>CnQa0"}, "label": "partially eaten pepperoni pizza in pizza box"}]} {"id": 161031, "image": "COCO_train2014_000000161031.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pack of pizza quarter eaten\"."}], "task": "refering", "answer_template": "The \"a pack of pizza quarter eaten\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 24, 25, 26, 27, 28, 29, 30, 31, 43, 44, 45, 46, 47, 48, 49, 50, 61, 62, 63, 64, 65, 66, 67, 68, 69, 79, 80, 81, 82, 83, 84, 85, 86, 87, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 204, 205, 206, 207, 208, 209], "bbox": [0.12572, 0.053933933933933934, 0.89348, 0.9887687687687687], "iscrowd": 0, "area": 79038.05559999999, "rle": {"size": [333, 500], "counts": "jdd0>n99G8H8H4M3M3M3M3M3M3M3M3M2M4M3M3M1O2N1N2O2N1O1O2N1N2O2N1O1O2N1N2O2N1O1O1N3N1O1O2N1O1N2O001O1O001N2O001O001O1N101O1O001O1N101O1O001O0O2O001O1O001N101O1O001N101O001N2O001O0O2O1O000O2O00001N1000001N10001O0O10001O0O101O1ZLkJR2W5lMlJP2V5oMlJo1U5oMoJn1R5QNQKl1Q5RNRKj1P5UNSKh1m4XNVKe1j4[NXKc1h4\\NZKb1g4^NZKa1f4_N[K`1f4_N[K`1e4`N[K_1f4`N\\K_1d4aN]K^1c4bN^K]1b4cN_K[1c4dN^K[1b4eN_KZ1a4eNaKZ1`4eN`KZ1a4fN`KY1`4gNaKX1`4gNaKW1`4hNbKW1_4hNcKV1]4jNdKU1\\4kNeKS1\\4mNeKR1\\4mNfKQ1Z4oNgKP1Y4POhKn0Z4QOhKm0X4SOiKl0W4TOjKk0V4UOkKi0W4VOkKh0U4XOlKg0T4YOmKf0S4ZOnKe0S4ZOoKc0R4]OoKb0Q4^OPLa0Q4^OQL`0o3@RL>o3BRL=n3CSLdL@]3`0fL\\O[3d0hLYOX3g0W300000000O100000001N100000000O100000000O10000001O001O00001O00001O00001O001O00001O000O2O0O100O2O0O100O100O100O101N100O100O100O100O2O0O100O100O100O100O10000O2O0O100O100O100O100O100O2O0O101N101N100O2O0O2O0O101N101N100O2O0O2O0O101O1N101N2O1N101N2O0O2O1N101N2O0O2O1N2O1N2O1N2O1N2O0O2O1N2O0O2O0O2O0O2O0O2O001N101N101N2O0O2O0O2O2M2O1N2O1N2O1N3N1N2O1N2N2N3M2N3M3L4M2N3M3M3M3M3M2N3M3M3M4L3M3M3M3I8G8H8H8J6M4M4L4L3M4G;D>C=B>CnQa0"}, "label": "a pack of pizza quarter eaten"}]} {"id": 29962, "image": "COCO_train2014_000000029962.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black umbrella , being held by a person in jeans\"."}], "task": "refering", "answer_template": "The \"a black umbrella , being held by a person in jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 66, 67, 68, 69, 79, 80, 81, 83, 93, 94, 108, 109, 123, 124], "bbox": [0.20296470588235296, 0.176328125, 0.6098588235294118, 0.368515625], "iscrowd": 0, "area": 5696.58455, "rle": {"size": [640, 425], "counts": "ene11lc03N3M3M3L4M2N3L4M3M3M2M4M3M3M3M4K5L4L4L4L101O0O2O00WOZ^OZOea0f0^^OWOba0i0`^OVO_a0j0d^OSO[a0n0h^OPOWa0o0m^OnNSa0R1P_OlNo`0T1T_OjNk`0V1X_OgNg`0Z1\\_OdNb`0\\1a_ObN_`0^1P11O100O010O100001N101O1O001N2O001O1O0O2O1O001OHlNf]OR1^b0oN^]OQ1fb023M3M0O11O00100O001O001O0100O10000O100O10O010000O100O100O10000O100O10O0100O10000O100O10O01000O010O0100O010O01000O100O101O000000010O000001O01O01O0000010O00000010O000001O0001O01O0000010O00000010O0001O0001O01O0000010O00AV]OHojX3"}, "label": "a black umbrella , being held by a person in jeans"}]} {"id": 29962, "image": "COCO_train2014_000000029962.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a skirt under an umbrella\"."}], "task": "refering", "answer_template": "The \"a woman in a skirt under an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 200, 201, 202, 203, 204, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 246, 247, 248, 249, 261, 262, 263, 264, 276, 277, 278, 279, 280, 291, 294], "bbox": [0.3207294117647059, 0.431171875, 0.696235294117647, 0.8467500000000001], "iscrowd": 0, "area": 19925.806699999994, "rle": {"size": [640, 425], "counts": "[\\e21mc02N2N3N1N2N3M2N2eMBRAa0k>BRA`0l>CQA`0k>DRA>l>EQA=m>FPA=l>GQA;m>HPA:n>Jn@9n>Ko@7o>Ln@6o>Nn@5o>No@3Q?Nl@4U?Lh@7X?Jd@8^?G_@e?BX@`0h?AU@b0j?_OS@c0m?^OP@d0P`0^Ol_Oe0S`0d10001O0aCVLh8j3nFaLQ9`3dFjL\\9V3ZFUMe9l2PF^MP:b2fEiMY:W2^ERNb:o1SE\\Nl:d1jDgNU;Y1aDQO_;P1VD[Oi;e0mCESC=B>B>B>B>C=B>B>L40000000000001O00000000000000000000000000001O2N6J6J6J6J6J6J6J6J6J6J6J6J5K6J6J6J6J6J6JgIaCW5Y in this image.", "objects": [{"patches": [155, 156, 157, 158, 159, 170, 171, 172, 173, 174, 185, 186, 187, 188, 189, 200, 201, 202, 203, 204, 214, 215, 216, 217, 218, 219, 231, 232, 233, 234, 246, 247, 248, 249, 261, 262, 263, 264, 276, 277, 278, 279, 280, 291, 294], "bbox": [0.3207294117647059, 0.431171875, 0.696235294117647, 0.8467500000000001], "iscrowd": 0, "area": 19925.806699999994, "rle": {"size": [640, 425], "counts": "[\\e21mc02N2N3N1N2N3M2N2eMBRAa0k>BRA`0l>CQA`0k>DRA>l>EQA=m>FPA=l>GQA;m>HPA:n>Jn@9n>Ko@7o>Ln@6o>Nn@5o>No@3Q?Nl@4U?Lh@7X?Jd@8^?G_@e?BX@`0h?AU@b0j?_OS@c0m?^OP@d0P`0^Ol_Oe0S`0d10001O0aCVLh8j3nFaLQ9`3dFjL\\9V3ZFUMe9l2PF^MP:b2fEiMY:W2^ERNb:o1SE\\Nl:d1jDgNU;Y1aDQO_;P1VD[Oi;e0mCESC=B>B>B>B>C=B>B>L40000000000001O00000000000000000000000000001O2N6J6J6J6J6J6J6J6J6J6J6J6J5K6J6J6J6J6J6JgIaCW5Y in this image.", "objects": [{"patches": [154, 155, 169, 170, 184, 185, 215, 216, 229, 230, 231, 245, 246, 260, 261, 262, 276], "bbox": [0.3190352941176471, 0.43256249999999996, 0.5080705882352942, 0.7914687499999999], "iscrowd": 0, "area": 4615.0054500000015, "rle": {"size": [640, 425], "counts": "fXe2d0\\c0f0ZOf0ZO?A2_@RNl0010O000010O01O000010O0o0QOj0VO8I7H8H7I8Imj;6mTD7H01O001O001N101O001O1O0O2M2O2MSUR4"}, "label": "the pants of the male"}]} {"id": 29962, "image": "COCO_train2014_000000029962.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"someone in black holding an umbrella\"."}], "task": "refering", "answer_template": "The \"someone in black holding an umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 169, 170, 184, 185, 215, 216, 229, 230, 231, 245, 246, 260, 261, 262, 276], "bbox": [0.3190352941176471, 0.43256249999999996, 0.5080705882352942, 0.7914687499999999], "iscrowd": 0, "area": 4615.0054500000015, "rle": {"size": [640, 425], "counts": "fXe2d0\\c0f0ZOf0ZO?A2_@RNl0010O000010O01O000010O0o0QOj0VO8I7H8H7I8Imj;6mTD7H01O001O001N101O001O1O0O2M2O2MSUR4"}, "label": "someone in black holding an umbrella"}]} {"id": 29962, "image": "COCO_train2014_000000029962.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pair of black umbrellas being held by a couple , shielding their faces\"."}], "task": "refering", "answer_template": "The \"a pair of black umbrellas being held by a couple , shielding their faces\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 80, 81, 82, 83, 84, 85, 94, 95, 96, 97, 98, 99, 100, 109, 110, 111, 112, 113, 114, 115, 116, 124, 125, 126, 127, 128, 129, 130, 131, 139, 140, 141, 142, 143, 144, 145, 156, 157, 158, 159], "bbox": [0.2688705882352941, 0.21028125000000003, 0.7746588235294118, 0.473625], "iscrowd": 0, "area": 24699.888649999994, "rle": {"size": [640, 425], "counts": "_RX23lc03M3M4K4M3M4K4M4L3M4K4M3M4L5J6K5K5J6K5K5K5J2O1O2M2O1O1O2M2O1O1O1N3N1O1N2O2N1O10001N10000O2O0O10001N10001N101N101O0O2O0O2O0O2O001N100O2O0O2O001N101N101N101O0O2O0O2O0O2O001N101N101N10001N101N101O001O01N100O101N100O2N100O2O000O2O000O101O0O101O0O101O0O10001N10001N10001N10001N10000O2O00001N10001N10000O2O000O2O000O2O000O2O000O101O0O101O0O101O0O10001N10001O1N2O1O2M2O2N1N2O2M2M4L3M3M4L3M4L3M4L4L3M4M3L4L3M4L4L3M4L4M2M4L4L3M4L4L3M4MYik1"}, "label": "a pair of black umbrellas being held by a couple , shielding their faces"}]} {"id": 29962, "image": "COCO_train2014_000000029962.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umbrella closest to the camera\"."}], "task": "refering", "answer_template": "The \"the umbrella closest to the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 80, 81, 82, 83, 84, 85, 94, 95, 96, 97, 98, 99, 100, 109, 110, 111, 112, 113, 114, 115, 116, 124, 125, 126, 127, 128, 129, 130, 131, 139, 140, 141, 142, 143, 144, 145, 156, 157, 158, 159], "bbox": [0.2688705882352941, 0.21028125000000003, 0.7746588235294118, 0.473625], "iscrowd": 0, "area": 24699.888649999994, "rle": {"size": [640, 425], "counts": "_RX23lc03M3M4K4M3M4K4M4L3M4K4M3M4L5J6K5K5J6K5K5K5J2O1O2M2O1O1O2M2O1O1O1N3N1O1N2O2N1O10001N10000O2O0O10001N10001N101N101O0O2O0O2O0O2O001N100O2O0O2O001N101N101N101O0O2O0O2O0O2O001N101N101N10001N101N101O001O01N100O101N100O2N100O2O000O2O000O101O0O101O0O101O0O10001N10001N10001N10001N10000O2O00001N10001N10000O2O000O2O000O2O000O2O000O101O0O101O0O101O0O10001N10001O1N2O1O2M2O2N1N2O2M2M4L3M3M4L3M4L3M4L4L3M4M3L4L3M4L4L3M4L4M2M4L4L3M4L4L3M4MYik1"}, "label": "the umbrella closest to the camera"}]} {"id": 210187, "image": "COCO_train2014_000000210187.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the table to the top left with the ketchup bottle on it\"."}], "task": "refering", "answer_template": "The \"the table to the top left with the ketchup bottle on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 24, 25, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 98, 99, 115, 116, 117, 118, 138, 139, 140], "bbox": [0.0017656249999999998, 0.044332603938730855, 0.33240625, 0.42365426695842445], "iscrowd": 0, "area": 14360.939049999999, "rle": {"size": [457, 640], "counts": "Za0=l=\\1dNX1hNO1000000O1000SORM_Fn2a9SM^Fl2b9VM]Fj2b9XM]Fh2c9ZM[Ff2d9\\M[Fc2e9_MZFa2e9aMZF_2f9bMYF^2g9cMXF\\2i9eMVF[2j9fMUFZ2k9gMTFX2m9iMRFW2n9jMQFV2o9lMoET2Q:nMmEQ2T:RNhEo1X:Q1O10O10[OZLgFf3Y9\\LeFd3Z9^LeFa3\\9aLbF_3]9cLbF]3^9dL`F]3^9fLaFY3Y9oLeFR3X9RMfFo2X9TMgFl2W9S1O2O11O1N2O001O1N105K4L1N2O1O1O0O2O1O2N4K5L4L3M0O1000000O1000000O106J6J6I7J3M000O100000O01000000O1000O10O10000[OaD`N_;]1dDbN\\;\\1gDdNY;X1n0M3M3O1O1N101O1O1O1O1O001O100000000O2O001O001O1N101O001O0O2O001O001O0O2O001O001N101O001O001N101O001O1O0O2O001O001O0O2O001O001O0010O010O010O010O010O010O1cBKe<4YC0f<0XC1h in this image.", "objects": [{"patches": [1, 24, 25, 46, 47, 48, 49, 50, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 98, 99, 115, 116, 117, 118, 138, 139, 140], "bbox": [0.0017656249999999998, 0.044332603938730855, 0.33240625, 0.42365426695842445], "iscrowd": 0, "area": 14360.939049999999, "rle": {"size": [457, 640], "counts": "Za0=l=\\1dNX1hNO1000000O1000SORM_Fn2a9SM^Fl2b9VM]Fj2b9XM]Fh2c9ZM[Ff2d9\\M[Fc2e9_MZFa2e9aMZF_2f9bMYF^2g9cMXF\\2i9eMVF[2j9fMUFZ2k9gMTFX2m9iMRFW2n9jMQFV2o9lMoET2Q:nMmEQ2T:RNhEo1X:Q1O10O10[OZLgFf3Y9\\LeFd3Z9^LeFa3\\9aLbF_3]9cLbF]3^9dL`F]3^9fLaFY3Y9oLeFR3X9RMfFo2X9TMgFl2W9S1O2O11O1N2O001O1N105K4L1N2O1O1O0O2O1O2N4K5L4L3M0O1000000O1000000O106J6J6I7J3M000O100000O01000000O1000O10O10000[OaD`N_;]1dDbN\\;\\1gDdNY;X1n0M3M3O1O1N101O1O1O1O1O001O100000000O2O001O001O1N101O001O0O2O001O001O0O2O001O001N101O001O001N101O001O1O0O2O001O001O0O2O001O001O0010O010O010O010O010O010O1cBKe<4YC0f<0XC1h in this image.", "objects": [{"patches": [281, 282, 287, 288, 289, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367], "bbox": [0.126765625, 0.7731728665207876, 0.9804375000000001, 0.9866739606126914], "iscrowd": 0, "area": 39542.57175000002, "rle": {"size": [457, 640], "counts": "ebT11W>1O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1N2O1O1N3N1O1N2O1O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1O1N2O2N1N2O1O1N2O1O1O1N2O1O1N2O1O1O1001O0000000000000000000000000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00000000000000O10000000000000000000000000000O100000001O0000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000000O1000000000000O100000000000000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001N100000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O100000000000000000000O1000000000000000000O100000000000O100000O100000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O10000000O1000000000O100000000000000000000O1000000000000001O001O001O001O001O001O001O001O001O001O003Mf0ZOg0YOY\\5"}, "label": "a book on table"}]} {"id": 210187, "image": "COCO_train2014_000000210187.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fine wooden table in front of a girl\"."}], "task": "refering", "answer_template": "The \"a fine wooden table in front of a girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [281, 282, 287, 288, 289, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367], "bbox": [0.126765625, 0.7731728665207876, 0.9804375000000001, 0.9866739606126914], "iscrowd": 0, "area": 39542.57175000002, "rle": {"size": [457, 640], "counts": "ebT11W>1O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1N2O1O1N3N1O1N2O1O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1O1N2O1O2M2O1O1N2O1O1N2O1O1N2O2N1N2O1O1N2O1O1O1N2O1O1N2O1O1O1001O0000000000000000000000000000001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00000000000000O10000000000000000000000000000O100000001O0000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O10000000000O1000000000000O1000000000000O100000000000000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001N100000000000000000O100000000000000000000O1000000000000000000O100000000000000000000O100000000000000000000O1000000000000000000O100000000000O100000O100000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O100000000000000000000O1000000000000000000O10000000O1000000000O100000000000000000000O1000000000000001O001O001O001O001O001O001O001O001O001O003Mf0ZOg0YOY\\5"}, "label": "a fine wooden table in front of a girl"}]} {"id": 562456, "image": "COCO_train2014_000000562456.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"half of sandwich closest to camera\"."}], "task": "refering", "answer_template": "The \"half of sandwich closest to camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [231, 232, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 280, 281, 282, 283, 284, 285, 286, 287, 296, 297, 298, 299, 300, 301, 302, 303, 304, 314, 315, 316, 317, 318, 319, 320, 331, 332, 333, 334, 335, 336, 350, 351, 352], "bbox": [0.4563333333333333, 0.604671875, 0.9311041666666665, 0.888234375], "iscrowd": 0, "area": 28307.453700000005, "rle": {"size": [640, 480], "counts": "X[Y4;ec01N3M2N2N3M2N2N3M2N2M4M2N3M2M3N3M2N2M4M2N2M4M2M2N3N2M3M3M2O2M3M3N1M4J6K5J6K4L5K5M3L3N3L4M3L4M201O1N2O0O2O1O0O2O1N101O0O2O1N101O1N101N2O001N2O0O2O1O0O2O1N10000000000000000000O100000000000000000000O100000000000000000000O1000000O10000O100O10000O10000O2O000O100O101O0O100O1O1O2N1O1O1O1O1O2N100O1N2O1N3N1O1N2O1O1N3N1O1N2O1O1O2N100O100O2N101N2O0O2O0O2O0O2O0O2N101N101N2O0O2O0O2N101N1O2O0O2N1O2O1N1O2N1O2O0O2N1O2N101N1O2N101N1O2K4M4K5K5L4K6K4L4L4L4L4L4L4L4L5K4L_Vd0"}, "label": "half of sandwich closest to camera"}]} {"id": 562456, "image": "COCO_train2014_000000562456.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the sandwich that has a bit out of it\"."}], "task": "refering", "answer_template": "The \"the sandwich that has a bit out of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [231, 232, 247, 248, 249, 250, 251, 252, 264, 265, 266, 267, 268, 269, 270, 280, 281, 282, 283, 284, 285, 286, 287, 296, 297, 298, 299, 300, 301, 302, 303, 304, 314, 315, 316, 317, 318, 319, 320, 331, 332, 333, 334, 335, 336, 350, 351, 352], "bbox": [0.4563333333333333, 0.604671875, 0.9311041666666665, 0.888234375], "iscrowd": 0, "area": 28307.453700000005, "rle": {"size": [640, 480], "counts": "X[Y4;ec01N3M2N2N3M2N2N3M2N2M4M2N3M2M3N3M2N2M4M2N2M4M2M2N3N2M3M3M2O2M3M3N1M4J6K5J6K4L5K5M3L3N3L4M3L4M201O1N2O0O2O1O0O2O1N101O0O2O1N101O1N101N2O001N2O0O2O1O0O2O1N10000000000000000000O100000000000000000000O100000000000000000000O1000000O10000O100O10000O10000O2O000O100O101O0O100O1O1O2N1O1O1O1O1O2N100O1N2O1N3N1O1N2O1O1N3N1O1N2O1O1O2N100O100O2N101N2O0O2O0O2O0O2O0O2N101N101N2O0O2O0O2N101N1O2O0O2N1O2O1N1O2N1O2O0O2N1O2N101N1O2N101N1O2K4M4K5K5L4K6K4L4L4L4L4L4L4L4L5K4L_Vd0"}, "label": "the sandwich that has a bit out of it"}]} {"id": 562456, "image": "COCO_train2014_000000562456.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"left half of sandwich on plate closest to silverware\"."}], "task": "refering", "answer_template": "The \"left half of sandwich on plate closest to silverware\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 214, 215, 216, 227, 228, 229, 230, 231, 232, 233, 234, 244, 245, 246, 247, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280], "bbox": [0.20302083333333334, 0.53190625, 0.778375, 0.733359375], "iscrowd": 0, "area": 12894.30625, "rle": {"size": [640, 480], "counts": "ZQm11oc02M3M3M3M20O001O1O010O1O0010O01O1O010O001O10O01O0001O01O000001O01O0001O0001O01O0001O01O0000010O000000010O000\\\\OMPc0MU]O6H2Qc0OP]O;nb0>N2N1N3N2N1O0O10O100000O10O10N2N2M3M3N1N3N2M3N2M2N3N2N2O10O01O1O100O001O100O1O001O100O1O1O010O1O1O10O01O1O100O001O100O1O010O1O1O10O01O1O100O00100O1O1O10O01O100O10O01O100O1O010O1O]Oi_O_MW`0`2d0O2O001O1N2O001O1N2O0O2N2O1N1O2O1N2N2O0O2N2N2O0O2O1000O0101O0O1000000O10000O10000O1000000O10001N1000000000000000000000000010O0001O0000001O00001O00001O0000001O00001O00001O0000001O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O0O2O1O001O1O1M2LXlQ2"}, "label": "left half of sandwich on plate closest to silverware"}]} {"id": 562456, "image": "COCO_train2014_000000562456.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the half of the sandwich closer to the silverware\"."}], "task": "refering", "answer_template": "The \"the half of the sandwich closer to the silverware\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 214, 215, 216, 227, 228, 229, 230, 231, 232, 233, 234, 244, 245, 246, 247, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280], "bbox": [0.20302083333333334, 0.53190625, 0.778375, 0.733359375], "iscrowd": 0, "area": 12894.30625, "rle": {"size": [640, 480], "counts": "ZQm11oc02M3M3M3M20O001O1O010O1O0010O01O1O010O001O10O01O0001O01O000001O01O0001O0001O01O0001O01O0000010O000000010O000\\\\OMPc0MU]O6H2Qc0OP]O;nb0>N2N1N3N2N1O0O10O100000O10O10N2N2M3M3N1N3N2M3N2M2N3N2N2O10O01O1O100O001O100O1O001O100O1O1O010O1O1O10O01O1O100O001O100O1O010O1O1O10O01O1O100O00100O1O1O10O01O100O10O01O100O1O010O1O]Oi_O_MW`0`2d0O2O001O1N2O001O1N2O0O2N2O1N1O2O1N2N2O0O2N2N2O0O2O1000O0101O0O1000000O10000O10000O1000000O10001N1000000000000000000000000010O0001O0000001O00001O00001O0000001O00001O00001O0000001O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O0O2O1O001O1O1M2LXlQ2"}, "label": "the half of the sandwich closer to the silverware"}]} {"id": 529689, "image": "COCO_train2014_000000529689.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of train that has red , green and white compartments\"."}], "task": "refering", "answer_template": "The \"the back of train that has red , green and white compartments\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 34, 45, 46, 47, 48, 49, 50, 60, 61, 62, 63, 64, 65, 75, 76, 77, 78, 79, 80, 91, 92, 93], "bbox": [0.05752941176470588, 0.094375, 0.39592941176470586, 0.287640625], "iscrowd": 0, "area": 13026.037, "rle": {"size": [640, 425], "counts": "lQ?3mc0k0UOl0TOg1YN00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00000000O100000000O10000000000O10000000000O1000000O1O100O1O1O1O1O1O1O1O1O1O1O1O2N2O1N2N2N2N101N2O1O1O1N2O1O1N2O001O1N2O1O1N2O1O1O0O2O1O1O1N2O1O1N2O001O1N2O1O1N2O1O1O1N101O1O1N2O1O1N2O1O001N2O1O1N2O1O1O1N2O001O1N2OX`P5"}, "label": "the back of train that has red , green and white compartments"}]} {"id": 529689, "image": "COCO_train2014_000000529689.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is the train on the far left\"."}], "task": "refering", "answer_template": "The \"this is the train on the far left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 34, 45, 46, 47, 48, 49, 50, 60, 61, 62, 63, 64, 65, 75, 76, 77, 78, 79, 80, 91, 92, 93], "bbox": [0.05752941176470588, 0.094375, 0.39592941176470586, 0.287640625], "iscrowd": 0, "area": 13026.037, "rle": {"size": [640, 425], "counts": "lQ?3mc0k0UOl0TOg1YN00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00000000O100000000O10000000000O10000000000O1000000O1O100O1O1O1O1O1O1O1O1O1O1O1O2N2O1N2N2N2N101N2O1O1O1N2O1O1N2O001O1N2O1O1N2O1O1O0O2O1O1O1N2O1O1N2O001O1N2O1O1N2O1O1O1N101O1O1N2O1O1N2O1O001N2O1O1N2O1O1O1N2O001O1N2OX`P5"}, "label": "this is the train on the far left"}]} {"id": 546078, "image": "COCO_train2014_000000546078.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl with green jacket\"."}], "task": "refering", "answer_template": "The \"girl with green jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 245, 246, 247, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 314, 315, 316, 317, 337, 338, 339, 340, 360, 361, 382, 383, 384], "bbox": [0.586890625, 0.5512083333333333, 0.79940625, 1.0], "iscrowd": 0, "area": 12841.847700000006, "rle": {"size": [480, 640], "counts": "Wc`53k>8H3M2N3N1O1O2NO1O1O02L31O010O00010O01O10O01O010O1O010O1O010O010O1O01O0000kEUOW6j0hIYOV6h0gI\\OW6d0fI@Z6?bIF]6:`IJ_66]INd61ZI2e6OWI5h6KVI8j6GSI=l6CRI`0m6@QIc0o6\\OoHg0P7ZOlHj0T7UOjHn0U7ROiHQ1V7oNhHT1X7kNfHX1Y7hNdHX1_7iNdG^OQOj1_9hN_GBkNf1j9hNYGGfNb1T:gNfF[O^O^2Q:VN`FU2J[Mm8a0WGQ2NaMk8?TGo11eMk8 in this image.", "objects": [{"patches": [223, 224, 245, 246, 247, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 314, 315, 316, 317, 337, 338, 339, 340, 360, 361, 382, 383, 384], "bbox": [0.586890625, 0.5512083333333333, 0.79940625, 1.0], "iscrowd": 0, "area": 12841.847700000006, "rle": {"size": [480, 640], "counts": "Wc`53k>8H3M2N3N1O1O2NO1O1O02L31O010O00010O01O10O01O010O1O010O1O010O010O1O01O0000kEUOW6j0hIYOV6h0gI\\OW6d0fI@Z6?bIF]6:`IJ_66]INd61ZI2e6OWI5h6KVI8j6GSI=l6CRI`0m6@QIc0o6\\OoHg0P7ZOlHj0T7UOjHn0U7ROiHQ1V7oNhHT1X7kNfHX1Y7hNdHX1_7iNdG^OQOj1_9hN_GBkNf1j9hNYGGfNb1T:gNfF[O^O^2Q:VN`FU2J[Mm8a0WGQ2NaMk8?TGo11eMk8 in this image.", "objects": [{"patches": [149, 166, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 209, 210, 211, 214, 215, 216, 217], "bbox": [0.099953125, 0.405125, 0.538078125, 0.5825416666666666], "iscrowd": 0, "area": 11611.661549999993, "rle": {"size": [480, 640], "counts": "_Xn02l>3N1N2N2O100O1O100O1O1O1O1O10O00100O001O01O01O01ON3O02N2O1N1O1O1O1O1O1O100O1O100O100O01O001O10O001O0O2O1O0100O10O10O10O10O10O10O10O1000000O100000000O10O0100O1O100O1O100O1O100O1N200O1000O10O10000O10001OO01000O1000O10000000000O1000000O10001O1O000000000000000O10000000000000000O1000001O000000001O000000000000000000000000000000000001O00001O0O2O00001O0000000001O5K2N1O1O00001O00O10000001O001O7I4L1O000000000000O100O1O1O1O1N2O1O1N2I70000000000O1000000O10000000N1M4N2N1O200O100001O0O2N10000O1O1O1O001O1O10O01N200O100O100O2O0O101N101N1O1O3L4L4N3L3N1N2O1N2O1N1M4N2N2NXaZ4"}, "label": "the light colored cat who is laying down"}]} {"id": 357663, "image": "COCO_train2014_000000357663.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a grey striped cat on his side\"."}], "task": "refering", "answer_template": "The \"a grey striped cat on his side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [149, 166, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 209, 210, 211, 214, 215, 216, 217], "bbox": [0.099953125, 0.405125, 0.538078125, 0.5825416666666666], "iscrowd": 0, "area": 11611.661549999993, "rle": {"size": [480, 640], "counts": "_Xn02l>3N1N2N2O100O1O100O1O1O1O1O10O00100O001O01O01O01ON3O02N2O1N1O1O1O1O1O1O100O1O100O100O01O001O10O001O0O2O1O0100O10O10O10O10O10O10O10O1000000O100000000O10O0100O1O100O1O100O1O100O1N200O1000O10O10000O10001OO01000O1000O10000000000O1000000O10001O1O000000000000000O10000000000000000O1000001O000000001O000000000000000000000000000000000001O00001O0O2O00001O0000000001O5K2N1O1O00001O00O10000001O001O7I4L1O000000000000O100O1O1O1O1N2O1O1N2I70000000000O1000000O10000000N1M4N2N1O200O100001O0O2N10000O1O1O1O001O1O10O01N200O100O100O2O0O101N101N1O1O3L4L4N3L3N1N2O1N2O1N1M4N2N2NXaZ4"}, "label": "a grey striped cat on his side"}]} {"id": 95518, "image": "COCO_train2014_000000095518.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on the frontmost bike\"."}], "task": "refering", "answer_template": "The \"man on the frontmost bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 214, 215, 216, 238, 239, 261, 262, 263, 285, 286, 308, 309, 331, 332, 354, 355, 377], "bbox": [0.33146875, 0.51075, 0.4536875, 0.9728125], "iscrowd": 0, "area": 5627.568399999999, "rle": {"size": [480, 640], "counts": "_dS37X>a0O2N1O2N1O2N1O2N2N101O001O001O001O1O001O001O001O1O0000000001O1O11Oa0_Oa0_O1O00O1O100O1O100O10VEjNi7U1UHVOc7i0[HA^7?_HL[71cH9a7[O]Hf0Q8nNmGR1W8kNgGV1\\8hNbGZ1_8fN^G[1e8cNYG^1j8aNTG_1o8_NPGa1R9^NmFc1U9\\NiFd1Z9ZNeFf1^9XNaFh1a9XN]Fh1f9XNWFh1l9YNPFh1R:XNkEh1W:YNfE0Fb0g:^O`EOO>d:DYEM8;c:GREMa08a:IkDNi05b:d0dEWOa:a0fE[O_:;jEA[:6nEEW:1RFLS:IWF2_ in this image.", "objects": [{"patches": [270, 271, 277, 278, 279, 280, 281, 285, 286, 287, 288, 292, 293, 294, 300, 301, 302, 303], "bbox": [0.0, 0.78834375, 0.7648941176470588, 0.9074375], "iscrowd": 0, "area": 4230.351549999999, "rle": {"size": [640, 425], "counts": "na07db0J]^O9_a0K_^O6]a00_^O3]a01a^O1_a00]^O3ba0N]^O3ba0N^^O2ba0N^^O1ba00^^O0aa00`^O0`a00`^O0_a01`^O0`a0Oa^O1^a00b^O0]a01c^ON^a01c^OO\\a02c^OO\\a02d^ON\\a01e^OOZa02f^ONYa03g^OMYa02h^OMXa04g^OMXa04h^OLXa03i^OMVa04j^OLUa05k^OKUa04k^OLUa05k^OKTa06l^OJTa05m^OKRa06n^OJQa07n^OJRa05o^OKRa04n^OKUa03k^OMWa00j^O0Xa0Nh^O2Za0Le^O5]a0Hd^O8^a0Fb^O9aa0E_^O;ca0B^^O>da0@[^Oa0ga0]OY^Oc0bb00000O10O100000000O010000000O10O10000000O1000O101O0000000O10001O0000000O101O000000000O101O0000000O1LUcV2IV]iM2N2N2M3N3M2N1OO1000000000000000O010000000000000000O10000000O10000000O10000000000000000O0100000000000000000O10000000O10000000O10000000000000000O01000000000000000O10000000O1000000000O10000000000000000O10000000000000000O10000O2M3M3Nkom1"}, "label": "the skis of the woman in white pants"}]} {"id": 62759, "image": "COCO_train2014_000000062759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"womans skiis\"."}], "task": "refering", "answer_template": "The \"womans skiis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [270, 271, 277, 278, 279, 280, 281, 285, 286, 287, 288, 292, 293, 294, 300, 301, 302, 303], "bbox": [0.0, 0.78834375, 0.7648941176470588, 0.9074375], "iscrowd": 0, "area": 4230.351549999999, "rle": {"size": [640, 425], "counts": "na07db0J]^O9_a0K_^O6]a00_^O3]a01a^O1_a00]^O3ba0N]^O3ba0N^^O2ba0N^^O1ba00^^O0aa00`^O0`a00`^O0_a01`^O0`a0Oa^O1^a00b^O0]a01c^ON^a01c^OO\\a02c^OO\\a02d^ON\\a01e^OOZa02f^ONYa03g^OMYa02h^OMXa04g^OMXa04h^OLXa03i^OMVa04j^OLUa05k^OKUa04k^OLUa05k^OKTa06l^OJTa05m^OKRa06n^OJQa07n^OJRa05o^OKRa04n^OKUa03k^OMWa00j^O0Xa0Nh^O2Za0Le^O5]a0Hd^O8^a0Fb^O9aa0E_^O;ca0B^^O>da0@[^Oa0ga0]OY^Oc0bb00000O10O100000000O010000000O10O10000000O1000O101O0000000O10001O0000000O101O000000000O101O0000000O1LUcV2IV]iM2N2N2M3N3M2N1OO1000000000000000O010000000000000000O10000000O10000000O10000000000000000O0100000000000000000O10000000O10000000O10000000000000000O01000000000000000O10000000O1000000000O10000000000000000O10000000000000000O10000O2M3M3Nkom1"}, "label": "womans skiis"}]} {"id": 62759, "image": "COCO_train2014_000000062759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in ski glasses and orange hat\"."}], "task": "refering", "answer_template": "The \"man in ski glasses and orange hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 24, 25, 26, 27, 39, 40, 41, 42, 54, 55, 56, 57, 58, 59, 69, 70, 71, 72, 73, 74, 84, 85, 86, 87, 88, 89, 98, 99, 100, 101, 102, 103, 104, 113, 114, 115, 116, 117, 118, 119, 127, 128, 129, 130, 131, 132, 133, 134, 142, 143, 144, 145, 146, 147, 148, 149, 157, 158, 160, 161, 162, 163, 164, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 205, 206, 207, 208, 209, 220, 221, 222, 223, 224, 235, 236, 237, 238, 239, 250, 251, 252, 253, 254, 266, 267, 268, 269, 281, 282, 283, 284, 295, 296, 297, 298, 299, 310, 311, 312, 313, 314, 326, 327, 328, 329, 341, 342, 343, 344], "bbox": [0.4776941176470588, 0.008890625000000001, 1.0, 0.9682812500000001], "iscrowd": 0, "area": 76111.25440000002, "rle": {"size": [640, 425], "counts": "]To32lc04M4K4M4K4M4K4M4K3N2M4M2M3M3M3M4K4M3M3M4L2M3N2N2N2N2M3N3M2N2N2M3N2N2N2N2N2N2N1O2NbNf_OZOX`0g0m_OUOQ`0m0T@nNj?S1[@hNd?Y1a@bN^?_1h@ZNX?g1m@TNR?m1TAmMk>T2ZAgMd>[2Y18H7I7L4O1O1O1O2N1O1O1O1O10001aNaL[B`3k<[MTCe2j<_MSCb2kB8G8lLnClMYlBBV=m<@UC`0k<]OWCd0h<[OYC]NiNi1n=G[C_NiNj1R>_OVCeNkNl1U>VORCmNkNd0H;`>LnBSOnNd0D>f>BjB[OnNa0Dd0j>XOeBAPO>Ej0l>nNaBIPO;Fo0o>eNUC7PNU1\\a0gNg^OZ1oa01O1O1O1O1O1O1O1O1O1O7Hd0]O7I001O1O1O1O1O1O001O1O1O1O1N2O1O001O1O1O1O1O1OO100000000000O01000000000000O100000000000000O0101O2N2N1O2N1O2N2N1\\I[LkLg3i2gLSM[3b2RMZMo2[2`M`Mb2U2kMgMW2S2RNhMo1V2VNeMm1X2XNdMj1Z2[NaMf1]2_N_Mc1^2cN]M^1a2gN[M[1b2jNZMX1c2nNXMS1f2ROVMP1g2VOTMl0i2YOSMh0k2]OQMe0m2@nLb0o2CmL>Q3HjL:S3KiL7T3NgL4m2 in this image.", "objects": [{"patches": [9, 10, 11, 24, 25, 26, 27, 39, 40, 41, 42, 54, 55, 56, 57, 58, 59, 69, 70, 71, 72, 73, 74, 84, 85, 86, 87, 88, 89, 98, 99, 100, 101, 102, 103, 104, 113, 114, 115, 116, 117, 118, 119, 127, 128, 129, 130, 131, 132, 133, 134, 142, 143, 144, 145, 146, 147, 148, 149, 157, 158, 160, 161, 162, 163, 164, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 205, 206, 207, 208, 209, 220, 221, 222, 223, 224, 235, 236, 237, 238, 239, 250, 251, 252, 253, 254, 266, 267, 268, 269, 281, 282, 283, 284, 295, 296, 297, 298, 299, 310, 311, 312, 313, 314, 326, 327, 328, 329, 341, 342, 343, 344], "bbox": [0.4776941176470588, 0.008890625000000001, 1.0, 0.9682812500000001], "iscrowd": 0, "area": 76111.25440000002, "rle": {"size": [640, 425], "counts": "]To32lc04M4K4M4K4M4K4M4K3N2M4M2M3M3M3M4K4M3M3M4L2M3N2N2N2N2M3N3M2N2N2M3N2N2N2N2N2N2N1O2NbNf_OZOX`0g0m_OUOQ`0m0T@nNj?S1[@hNd?Y1a@bN^?_1h@ZNX?g1m@TNR?m1TAmMk>T2ZAgMd>[2Y18H7I7L4O1O1O1O2N1O1O1O1O10001aNaL[B`3k<[MTCe2j<_MSCb2kB8G8lLnClMYlBBV=m<@UC`0k<]OWCd0h<[OYC]NiNi1n=G[C_NiNj1R>_OVCeNkNl1U>VORCmNkNd0H;`>LnBSOnNd0D>f>BjB[OnNa0Dd0j>XOeBAPO>Ej0l>nNaBIPO;Fo0o>eNUC7PNU1\\a0gNg^OZ1oa01O1O1O1O1O1O1O1O1O1O7Hd0]O7I001O1O1O1O1O1O001O1O1O1O1N2O1O001O1O1O1O1O1OO100000000000O01000000000000O100000000000000O0101O2N2N1O2N1O2N2N1\\I[LkLg3i2gLSM[3b2RMZMo2[2`M`Mb2U2kMgMW2S2RNhMo1V2VNeMm1X2XNdMj1Z2[NaMf1]2_N_Mc1^2cN]M^1a2gN[M[1b2jNZMX1c2nNXMS1f2ROVMP1g2VOTMl0i2YOSMh0k2]OQMe0m2@nLb0o2CmL>Q3HjL:S3KiL7T3NgL4m2 in this image.", "objects": [{"patches": [4, 5, 19, 20, 21, 33, 34, 35, 36, 48, 49, 50, 51, 62, 63, 64, 65, 66, 77, 78, 79, 80, 81, 92, 93, 94, 95, 96, 107, 108, 109, 110, 111, 122, 123, 124, 125, 137, 138, 139, 140, 152, 153, 154, 155, 168, 169, 170, 183, 184, 185, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 227, 228, 229, 230, 231, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 273, 274, 275, 276, 288, 289, 290, 291], "bbox": [0.07138823529411764, 0.017499999999999998, 0.4691058823529412, 0.8459375000000001], "iscrowd": 0, "area": 48794.480149999996, "rle": {"size": [640, 425], "counts": "]Wc01mc02N2N2N2N2M3N2N2N2N2N2M3N2N2N3M2N2M3N2N2N2N2M3N2N2N2N2N2M3N3M2N2N3M2dJiMlHZ2j6SNQIP2e6[NWIg1_6fN\\I]1Z6oNaIS1U6ZOfIh0Q6DiI`0l5MoI5h56TJLc5`0XJC]5j0^JXOY5T1aJoNU5^1fJeNS5d1hJ^NV5f1fJ]NW5h1dJZNZ5k1aJWN]5n1]JVN`5o1[JSNc5R2XJPNf5T2VJoMg5V2TJlMj5Y2QJiMm5\\2mIhMP6]2kIeMS6_2iIdMT6a2gIaMV6e2eI]MY6h2bI[M[6j2_IYM_6l2\\IVMb6n2ZIUMc6P3XIRMf6U8N3M2O2M2Oo0PO2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N101N2O1N2N2O1N2O1N2N2O1001O2N1O2N1O1O2N1O2N1THVL@l3:]LAd39eLC\\37mLEU35TMFm24\\MHf22bMJ_20jMLX2NQNMP2MYNOh1L`N0b1IhN2Y1IoNPOgLmK]4l4UOmNeLVLW4h4\\OiNcL^LS4b4DeN`LhLm3^4KaN^LPMh3Y43^N[LXMd3T4:YNYLbM^3o3b0VNVLjMZ3j3h0SNTLRNU3e3Q1nMQL\\NP3`3W1kMnKeNl2Z3_1gMlKnNf2V3f1cMjKVOb2c2[2mMYK_O]2P2Q3WNiJHX2\\1f3cNXJ0S2h0^4mNfI:n14_8Kl5O1O1O001O1O1O1O1O1O1O1O1OVg\\4"}, "label": "a woman in a green jacket"}]} {"id": 62759, "image": "COCO_train2014_000000062759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing green gloves and eye glass standing near a man\"."}], "task": "refering", "answer_template": "The \"a woman wearing green gloves and eye glass standing near a man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 19, 20, 21, 33, 34, 35, 36, 48, 49, 50, 51, 62, 63, 64, 65, 66, 77, 78, 79, 80, 81, 92, 93, 94, 95, 96, 107, 108, 109, 110, 111, 122, 123, 124, 125, 137, 138, 139, 140, 152, 153, 154, 155, 168, 169, 170, 183, 184, 185, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 227, 228, 229, 230, 231, 241, 242, 243, 244, 245, 246, 256, 257, 258, 259, 260, 261, 273, 274, 275, 276, 288, 289, 290, 291], "bbox": [0.07138823529411764, 0.017499999999999998, 0.4691058823529412, 0.8459375000000001], "iscrowd": 0, "area": 48794.480149999996, "rle": {"size": [640, 425], "counts": "]Wc01mc02N2N2N2N2M3N2N2N2N2N2M3N2N2N3M2N2M3N2N2N2N2M3N2N2N2N2N2M3N3M2N2N3M2dJiMlHZ2j6SNQIP2e6[NWIg1_6fN\\I]1Z6oNaIS1U6ZOfIh0Q6DiI`0l5MoI5h56TJLc5`0XJC]5j0^JXOY5T1aJoNU5^1fJeNS5d1hJ^NV5f1fJ]NW5h1dJZNZ5k1aJWN]5n1]JVN`5o1[JSNc5R2XJPNf5T2VJoMg5V2TJlMj5Y2QJiMm5\\2mIhMP6]2kIeMS6_2iIdMT6a2gIaMV6e2eI]MY6h2bI[M[6j2_IYM_6l2\\IVMb6n2ZIUMc6P3XIRMf6U8N3M2O2M2Oo0PO2N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N2O1N2O1N2N2O1N2O1N2O1N2N101N2O1N2N2O1N2O1N2N2O1001O2N1O2N1O1O2N1O2N1THVL@l3:]LAd39eLC\\37mLEU35TMFm24\\MHf22bMJ_20jMLX2NQNMP2MYNOh1L`N0b1IhN2Y1IoNPOgLmK]4l4UOmNeLVLW4h4\\OiNcL^LS4b4DeN`LhLm3^4KaN^LPMh3Y43^N[LXMd3T4:YNYLbM^3o3b0VNVLjMZ3j3h0SNTLRNU3e3Q1nMQL\\NP3`3W1kMnKeNl2Z3_1gMlKnNf2V3f1cMjKVOb2c2[2mMYK_O]2P2Q3WNiJHX2\\1f3cNXJ0S2h0^4mNfI:n14_8Kl5O1O1O001O1O1O1O1O1O1O1O1OVg\\4"}, "label": "a woman wearing green gloves and eye glass standing near a man"}]} {"id": 62759, "image": "COCO_train2014_000000062759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skis on feet of man\"."}], "task": "refering", "answer_template": "The \"skis on feet of man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [308, 309, 310, 318, 319, 320, 321, 322, 323, 325, 333, 334, 335, 337, 338, 339, 340], "bbox": [0.2301176470588235, 0.8764062499999999, 0.7309411764705882, 0.9887656249999999], "iscrowd": 0, "area": 3114.9494999999965, "rle": {"size": [640, 425], "counts": "\\km11oc01XO0_]O1_b0Oa]O2]b00c]O0[b02d]O0[b0Of]O1Yb00g]O0Xb00h]O0Yb00g]O0Xb00i]OOWb02i]OMXb02h]ONXb03h]OLXb05h]OKXb04h]OLXb05h]OJYb05h]OJXb07h00P]OIXb07h000100O0010O01O010O00100O010O0010O01O0O2O1O000O1000O10O10O1000O0100000O010000O01000O10O10O1000O10O1000O010000O01000O10O1000O10O10O1000O010000O0100000O01000O01000O10O1000O10O10O1000O010000O01Y]OEia0;W^OEia0;W^OEia0:W^OGha0:X^OFha0:X^OFha09Y^OHfa08Y^OIfa08Z^OHfa07[^OIea07[^OIea07Z^OKda05]^OKca05]^OKca05\\^OLda04\\^OLca04^^OLba04^^OLba04]^OMca02^^OMba04^^OLba04^^OLba03^^ONba02^^ONaa03_^OLba03^^ONba02^^ONba02^^ONaa03_^OLba03^^ONba02^^ONba02^^ONaa02`^ON`a02_^ONaa03_^OMaa02`^ON`a02`^ON_a03`^OMaa02`^ON`a02`^ON_a03`^ON`a02`^OM`a03a^OM_a03a^OM_a03`^ON_a02b^ON^a02b^OM_a03a^OM^a03b^ON^a02b^ON^a02b^OM^a03c^OM]a03b^ON]a03c^OM]a03c^OL^a03b^ON]a03c^OM]a03c^OM]a02d^ON[a03d^OM]a03c^OM\\a03e^OM[a03e^OM[a03d^OM\\a04d^OL\\a03e^OM[a03e^OMZa04e^OL\\a03e^OM[a03e^OMZa04e^OM[a02f^ONYa03g^OLZa04f^OLZa03f^ONYa03g^OMYa03g^OLZa04f^OLYa04g^OMYa03g^OMXa04h^OKYa04h^OL^a0Na^O3fb0000O0100000O0100000O10O102N7IldV2"}, "label": "skis on feet of man"}]} {"id": 62759, "image": "COCO_train2014_000000062759.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skis that the right hand side person is using\"."}], "task": "refering", "answer_template": "The \"the skis that the right hand side person is using\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [308, 309, 310, 318, 319, 320, 321, 322, 323, 325, 333, 334, 335, 337, 338, 339, 340], "bbox": [0.2301176470588235, 0.8764062499999999, 0.7309411764705882, 0.9887656249999999], "iscrowd": 0, "area": 3114.9494999999965, "rle": {"size": [640, 425], "counts": "\\km11oc01XO0_]O1_b0Oa]O2]b00c]O0[b02d]O0[b0Of]O1Yb00g]O0Xb00h]O0Yb00g]O0Xb00i]OOWb02i]OMXb02h]ONXb03h]OLXb05h]OKXb04h]OLXb05h]OJYb05h]OJXb07h00P]OIXb07h000100O0010O01O010O00100O010O0010O01O0O2O1O000O1000O10O10O1000O0100000O010000O01000O10O10O1000O10O1000O010000O01000O10O1000O10O10O1000O010000O0100000O01000O01000O10O1000O10O10O1000O010000O01Y]OEia0;W^OEia0;W^OEia0:W^OGha0:X^OFha0:X^OFha09Y^OHfa08Y^OIfa08Z^OHfa07[^OIea07[^OIea07Z^OKda05]^OKca05]^OKca05\\^OLda04\\^OLca04^^OLba04^^OLba04]^OMca02^^OMba04^^OLba04^^OLba03^^ONba02^^ONaa03_^OLba03^^ONba02^^ONba02^^ONaa03_^OLba03^^ONba02^^ONba02^^ONaa02`^ON`a02_^ONaa03_^OMaa02`^ON`a02`^ON_a03`^OMaa02`^ON`a02`^ON_a03`^ON`a02`^OM`a03a^OM_a03a^OM_a03`^ON_a02b^ON^a02b^OM_a03a^OM^a03b^ON^a02b^ON^a02b^OM^a03c^OM]a03b^ON]a03c^OM]a03c^OL^a03b^ON]a03c^OM]a03c^OM]a02d^ON[a03d^OM]a03c^OM\\a03e^OM[a03e^OM[a03d^OM\\a04d^OL\\a03e^OM[a03e^OMZa04e^OL\\a03e^OM[a03e^OMZa04e^OM[a02f^ONYa03g^OLZa04f^OLZa03f^ONYa03g^OMYa03g^OLZa04f^OLYa04g^OMYa03g^OMXa04h^OKYa04h^OL^a0Na^O3fb0000O0100000O0100000O10O102N7IldV2"}, "label": "the skis that the right hand side person is using"}]} {"id": 562474, "image": "COCO_train2014_000000562474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow book written by a . a . milne\"."}], "task": "refering", "answer_template": "The \"a yellow book written by a . a . milne\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [269, 270, 271, 291, 292, 293, 294, 295, 312, 313, 314, 315, 316, 317, 318, 319], "bbox": [0.577265625, 0.8105102040816328, 0.887703125, 0.9940561224489797], "iscrowd": 0, "area": 7834.87945, "rle": {"size": [392, 640], "counts": "]d]41W<0O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1N101N2N1O2O1N1O2O1N1O2O1N1O2N_Vk0"}, "label": "a yellow book written by a . a . milne"}]} {"id": 562474, "image": "COCO_train2014_000000562474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a book sitting on a table in front of a group of four bears\"."}], "task": "refering", "answer_template": "The \"a book sitting on a table in front of a group of four bears\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [269, 270, 271, 291, 292, 293, 294, 295, 312, 313, 314, 315, 316, 317, 318, 319], "bbox": [0.577265625, 0.8105102040816328, 0.887703125, 0.9940561224489797], "iscrowd": 0, "area": 7834.87945, "rle": {"size": [392, 640], "counts": "]d]41W<0O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100O100O1O100001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1O001O1N101N2N1O2O1N1O2O1N1O2O1N1O2N_Vk0"}, "label": "a book sitting on a table in front of a group of four bears"}]} {"id": 562474, "image": "COCO_train2014_000000562474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a book titled kakarhu\"."}], "task": "refering", "answer_template": "The \"a book titled kakarhu\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 119, 120, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189], "bbox": [0.012390624999999999, 0.4224744897959184, 0.28215625, 0.624719387755102], "iscrowd": 0, "area": 9446.699349999999, "rle": {"size": [392, 640], "counts": "ge32U<1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N101N2O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2000000O10000000000000O1000000000000000000000000000000000000000000000000000000001O000000000000001O0000000000001O000000000000001O000000000000001O000000000000001O00N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2Nlm_5"}, "label": "a book titled kakarhu"}]} {"id": 562474, "image": "COCO_train2014_000000562474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a story book with a bear on the cover on a table\"."}], "task": "refering", "answer_template": "The \"a story book with a bear on the cover on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 119, 120, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189], "bbox": [0.012390624999999999, 0.4224744897959184, 0.28215625, 0.624719387755102], "iscrowd": 0, "area": 9446.699349999999, "rle": {"size": [392, 640], "counts": "ge32U<1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N101N2O1O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N2O1N2000000O10000000000000O1000000000000000000000000000000000000000000000000000000001O000000000000001O0000000000001O000000000000001O000000000000001O000000000000001O00N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2Nlm_5"}, "label": "a story book with a bear on the cover on a table"}]} {"id": 54572, "image": "COCO_train2014_000000054572.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"polar bear cub on the left and leaning to the left\"."}], "task": "refering", "answer_template": "The \"polar bear cub on the left and leaning to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 280, 281, 282], "bbox": [0.130609375, 0.3545901639344262, 0.37778125, 0.8501170960187353], "iscrowd": 0, "area": 22719.693799999994, "rle": {"size": [427, 640], "counts": "_WS1B=B>C:F4L3M4N1O2O0O2N100O2Oj0UOk0UO3N2M3M3N1N3M3N2M3M2N3M3M3H8L3N3L4M3M3L4M2O2N2M3N001N100001O1N2O001O1N101O1O0O0100O1O010O1O10O0100O00100O1O010O1000O1000000O100K5O1000000001O0001O001O001O001O1O001O001O001O001O1N1O2O0O1O0100O10000O010O10hK`He2_7\\MbHn1S8QNoGl1S8SNoGk1S8SNoGk1R8TNQHi1Q8PNVHn1l7gM_HX2a7aMgH]2[7\\MlHc2U7XMTIb2m6YMbI]2o8G:F8H4L4L4L4L5K9G9G1O10O01O1O1O1O001O1O1O1O1O1O1O1O100OndU5"}, "label": "polar bear cub on the left and leaning to the left"}]} {"id": 54572, "image": "COCO_train2014_000000054572.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the polar cub on the left\"."}], "task": "refering", "answer_template": "The \"the polar cub on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 210, 211, 212, 213, 214, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 280, 281, 282], "bbox": [0.130609375, 0.3545901639344262, 0.37778125, 0.8501170960187353], "iscrowd": 0, "area": 22719.693799999994, "rle": {"size": [427, 640], "counts": "_WS1B=B>C:F4L3M4N1O2O0O2N100O2Oj0UOk0UO3N2M3M3N1N3M3N2M3M2N3M3M3H8L3N3L4M3M3L4M2O2N2M3N001N100001O1N2O001O1N101O1O0O0100O1O010O1O10O0100O00100O1O010O1000O1000000O100K5O1000000001O0001O001O001O001O1O001O001O001O001O1N1O2O0O1O0100O10000O010O10hK`He2_7\\MbHn1S8QNoGl1S8SNoGk1S8SNoGk1R8TNQHi1Q8PNVHn1l7gM_HX2a7aMgH]2[7\\MlHc2U7XMTIb2m6YMbI]2o8G:F8H4L4L4L4L5K9G9G1O10O01O1O1O1O001O1O1O1O1O1O1O1O100OndU5"}, "label": "the polar cub on the left"}]} {"id": 234819, "image": "COCO_train2014_000000234819.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motor bike parked amongst other motor bikes with its front wheel facing the right\"."}], "task": "refering", "answer_template": "The \"a motor bike parked amongst other motor bikes with its front wheel facing the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 54, 55, 56, 57, 58, 75, 76, 77, 78, 79, 80, 81, 82, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 112, 113, 114, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 256, 257, 258, 259, 260, 261, 262, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 279, 280, 281, 282, 283, 284, 285, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 302, 303, 304, 305, 306, 307, 314, 318, 319, 320, 321, 326, 327, 328, 329, 337, 360], "bbox": [0.134078125, 0.0979375, 1.0, 0.8930625], "iscrowd": 0, "area": 125155.81535, "rle": {"size": [480, 640], "counts": "]cX1=S>d0\\Oc0I8J5K5K6J5K5J7J5K6J5K5K6J5K5K4L3L4M3M2N3M3M3M3M3M3L4M3L4M3M3M3O1N2N2O1N2O1N2N2O1N2O1N2N101N1O2O0O2O1N1O2O001O001O001O0O2O001O001O001O0O100000000000000O2O0001O0000001O0000001O01O000]MTISMk6i2ZIVMf6f2`IXM`6c2fI]MY6a2jI^MV6b2lI\\MS6e2nIZMR6f2oIZMP6f2PJZMP6f2PJZMP6f2PJZMP6f2PJ[Mn5f2RJZMn5f2RJZMn5f2RJXMP6h2PJWMQ6i2oIUMR6l2nISMS6m2mIQMU6o2kIPMV6P3jInLX6R3hIlLZ6T3fIkLZ6V3fIhL\\6X3dIgL\\6Z3dIdL^6\\3bIcL^6^3bI`L_6a3aI^L_6d3`IZLa6h3aISL`6P4aIlK`6U4cIcK`6_4cIXKa6k4aIlJc6V5_IbJe6a5\\11N3M1O1O100O1O1O1O1O100O001O1O00ZN\\GRMb8P3bGkL]8W3gGeLW8]3mG^LS8c3QHYLm7j3THSLk7o3WHlKj7V4XHcKk7_4VH[Km7g4UHRKo7o4o02O1O100O1O101N1O1ULRKdMn4V2ZKgMg4Y2[KdMf4[2^KbMb4]2aK`M`4_2dK]M]4b2fK\\MZ4c2jKWMY4i2iKRMZ4m2iKnL[4P3jKjLX4U3lKeLW4[3mK^LV4b3nKWLU4h3oK_Kg4a4l20000O10001N10000O2O001N10001N101N1O101N1O2N2O2M2N2N2O1N3O03M4L4UG\\KQ7g4hH`KV7c4bHcK\\7a4\\HeKb7^4VHiKh7[4PHkKo7X4iGmKV8`5O1O1N2N1O2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N1O2N2N2N1O001O1O1O001O1O10O01O100O001O100O00100000O01000O1O001O10O01O001O1O010O001O001O010O001O001O010O1O1O1O1O100O1O2N2N2N2O1N1O00001O001O00001O001O00001O001O00001O001O00001O001O00001O0000SJjHQ4U7nKmHQ4S7oKnHQ4Q7oKQIo3o6PLSIo3m6QLTIo3k6PLWIo3i6oKYIR4f6lK\\IT4d6jK^IW4a6gKaIY4_6eKcI\\4\\6aKgI_4Y6^KjIc4U6YKoIg4R6TK_IWOLf5d6PK_IAJ_5g6lJ_ILGY5j6gJ^I5GT5k6cJ^I>FP5k6_J^If0Fk4m6ZJ^Io0Dh4m6VJ_IV1Cd4n6RJaI^7m64L3M4L3N3L3N3N1N3N1N3O001O0001O0O100000000000000000000O1000001O00001O001O001O0O2O001O001O000000O1O10O01O100O100O100O1O10000O01000000000O100000000O1000000001N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00001O0000001O00001O00001O0O101O0000001N10000000000O1000000000000O10000000000O100000000000000O1000O10000O10000O01000O10000O10000O0100000O10000O1000O010000O10000O100O1O100O1O1O2N1O100O1O1O1O101N2^KQJ0Q6^NQM0fN"}, "label": "a motor bike parked amongst other motor bikes with its front wheel facing the right"}]} {"id": 234819, "image": "COCO_train2014_000000234819.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red honda motorcycle on the grass in front of a couple of motorcycles and a few cars\"."}], "task": "refering", "answer_template": "The \"a red honda motorcycle on the grass in front of a couple of motorcycles and a few cars\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 54, 55, 56, 57, 58, 75, 76, 77, 78, 79, 80, 81, 82, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 112, 113, 114, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 256, 257, 258, 259, 260, 261, 262, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 279, 280, 281, 282, 283, 284, 285, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 302, 303, 304, 305, 306, 307, 314, 318, 319, 320, 321, 326, 327, 328, 329, 337, 360], "bbox": [0.134078125, 0.0979375, 1.0, 0.8930625], "iscrowd": 0, "area": 125155.81535, "rle": {"size": [480, 640], "counts": "]cX1=S>d0\\Oc0I8J5K5K6J5K5J7J5K6J5K5K6J5K5K4L3L4M3M2N3M3M3M3M3M3L4M3L4M3M3M3O1N2N2O1N2O1N2N2O1N2O1N2N101N1O2O0O2O1N1O2O001O001O001O0O2O001O001O001O0O100000000000000O2O0001O0000001O0000001O01O000]MTISMk6i2ZIVMf6f2`IXM`6c2fI]MY6a2jI^MV6b2lI\\MS6e2nIZMR6f2oIZMP6f2PJZMP6f2PJZMP6f2PJZMP6f2PJ[Mn5f2RJZMn5f2RJZMn5f2RJXMP6h2PJWMQ6i2oIUMR6l2nISMS6m2mIQMU6o2kIPMV6P3jInLX6R3hIlLZ6T3fIkLZ6V3fIhL\\6X3dIgL\\6Z3dIdL^6\\3bIcL^6^3bI`L_6a3aI^L_6d3`IZLa6h3aISL`6P4aIlK`6U4cIcK`6_4cIXKa6k4aIlJc6V5_IbJe6a5\\11N3M1O1O100O1O1O1O1O100O001O1O00ZN\\GRMb8P3bGkL]8W3gGeLW8]3mG^LS8c3QHYLm7j3THSLk7o3WHlKj7V4XHcKk7_4VH[Km7g4UHRKo7o4o02O1O100O1O101N1O1ULRKdMn4V2ZKgMg4Y2[KdMf4[2^KbMb4]2aK`M`4_2dK]M]4b2fK\\MZ4c2jKWMY4i2iKRMZ4m2iKnL[4P3jKjLX4U3lKeLW4[3mK^LV4b3nKWLU4h3oK_Kg4a4l20000O10001N10000O2O001N10001N101N1O101N1O2N2O2M2N2N2O1N3O03M4L4UG\\KQ7g4hH`KV7c4bHcK\\7a4\\HeKb7^4VHiKh7[4PHkKo7X4iGmKV8`5O1O1N2N1O2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N1O2N2N2N1O001O1O1O001O1O10O01O100O001O100O00100000O01000O1O001O10O01O001O1O010O001O001O010O001O001O010O1O1O1O1O100O1O2N2N2N2O1N1O00001O001O00001O001O00001O001O00001O001O00001O001O00001O0000SJjHQ4U7nKmHQ4S7oKnHQ4Q7oKQIo3o6PLSIo3m6QLTIo3k6PLWIo3i6oKYIR4f6lK\\IT4d6jK^IW4a6gKaIY4_6eKcI\\4\\6aKgI_4Y6^KjIc4U6YKoIg4R6TK_IWOLf5d6PK_IAJ_5g6lJ_ILGY5j6gJ^I5GT5k6cJ^I>FP5k6_J^If0Fk4m6ZJ^Io0Dh4m6VJ_IV1Cd4n6RJaI^7m64L3M4L3N3L3N3N1N3N1N3O001O0001O0O100000000000000000000O1000001O00001O001O001O0O2O001O001O000000O1O10O01O100O100O100O1O10000O01000000000O100000000O1000000001N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00001O0000001O00001O00001O0O101O0000001N10000000000O1000000000000O10000000000O100000000000000O1000O10000O10000O01000O10000O10000O0100000O10000O1000O010000O10000O100O1O100O1O1O2N1O100O1O1O1O101N2^KQJ0Q6^NQM0fN"}, "label": "a red honda motorcycle on the grass in front of a couple of motorcycles and a few cars"}]} {"id": 234819, "image": "COCO_train2014_000000234819.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the motorcycle that has a burgundy windshield\"."}], "task": "refering", "answer_template": "The \"the motorcycle that has a burgundy windshield\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 25, 26, 27, 28, 29, 30, 47, 48, 49, 50, 51, 52, 53, 54, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 253], "bbox": [0.0031406249999999998, 0.028874999999999998, 0.3739375, 0.6625833333333334], "iscrowd": 0, "area": 37790.5538, "rle": {"size": [480, 640], "counts": "bU1Y2^:Y201O000O101O00001O0O10O1000UN[K[Id4f6\\KZId4e6]KZId4e6^KZIb4e6_K[I`4f6`K]HNOb4R7]KcH4715^4Q7^KaH542;Z4P7bK^H335?V4P7cK]H217c0T4o6eKZH118g0Q4n6iKWH017k0P4m6jKUH008o0n3l6mLUIR3V6dKXIZ1c0R3V6aMkI_2U6`MlI_2U6aMkI_2U6`MlI`2T6_MmI`2T6_MmIa2S6^MnIb2R6]MnIc2S6cMgI]2Y6_MkIa2U6_MkIa2V6^MjIa2V6`MjIa2T6`MlI`2R6bMnI^2Q6cMoI]2P6fMnIZ2P6jMnIV2Q6mMmIT2R6nMlIR2U6oMiIQ2W6QNgIo1Y6SNeIm1\\6SNcIm1]6UNaIl1^6VN`Ij1a6WN]Ii1c6WN]Ii1c6XN\\Ih1c6YN]Ig1c6ZNZIh1e6YNZIh1e6ZNYIf1g6\\NWId1j6\\NTIf1k6\\NSId1m6]NRIc1o6^NoHc1P7V30010O000010O00010O00010O0N2K5L5J5K5L5M2M3N3M2N2M3N3M2M3N3M2N201O0O2O001O0O2O1O001N2O001O0O2O1O001N2O001O1O0O2O001O10O01000O10O010000O01000O10O0100001N101O1O1N2O1O001N2O1O1O1N101O1O1N2O00000O1001N10000000ULVFa2i9^MXFb2h9]MYFc2h9[MZFd2f9[M[Fe2f9YM[F2Df1Q:VN\\F3Fe1o9VN\\F3Hf1m9UN\\F4Je1k9UN]F3Kg1i9TN]F4Mg1f9SN^F5Ng1e9RN^F51g1b9lMdFXE]Ok:c0WEYOk:h0VETOl:m0UEnNn:S1_11O1O1O1O1O1O1O100O1O1O1O001O02N3M3M3M4L4L3M4Felk5"}, "label": "the motorcycle that has a burgundy windshield"}]} {"id": 234819, "image": "COCO_train2014_000000234819.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dark maroon motorcycle behind the red one\"."}], "task": "refering", "answer_template": "The \"the dark maroon motorcycle behind the red one\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 25, 26, 27, 28, 29, 30, 47, 48, 49, 50, 51, 52, 53, 54, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 253], "bbox": [0.0031406249999999998, 0.028874999999999998, 0.3739375, 0.6625833333333334], "iscrowd": 0, "area": 37790.5538, "rle": {"size": [480, 640], "counts": "bU1Y2^:Y201O000O101O00001O0O10O1000UN[K[Id4f6\\KZId4e6]KZId4e6^KZIb4e6_K[I`4f6`K]HNOb4R7]KcH4715^4Q7^KaH542;Z4P7bK^H335?V4P7cK]H217c0T4o6eKZH118g0Q4n6iKWH017k0P4m6jKUH008o0n3l6mLUIR3V6dKXIZ1c0R3V6aMkI_2U6`MlI_2U6aMkI_2U6`MlI`2T6_MmI`2T6_MmIa2S6^MnIb2R6]MnIc2S6cMgI]2Y6_MkIa2U6_MkIa2V6^MjIa2V6`MjIa2T6`MlI`2R6bMnI^2Q6cMoI]2P6fMnIZ2P6jMnIV2Q6mMmIT2R6nMlIR2U6oMiIQ2W6QNgIo1Y6SNeIm1\\6SNcIm1]6UNaIl1^6VN`Ij1a6WN]Ii1c6WN]Ii1c6XN\\Ih1c6YN]Ig1c6ZNZIh1e6YNZIh1e6ZNYIf1g6\\NWId1j6\\NTIf1k6\\NSId1m6]NRIc1o6^NoHc1P7V30010O000010O00010O00010O0N2K5L5J5K5L5M2M3N3M2N2M3N3M2M3N3M2N201O0O2O001O0O2O1O001N2O001O0O2O1O001N2O001O1O0O2O001O10O01000O10O010000O01000O10O0100001N101O1O1N2O1O001N2O1O1O1N101O1O1N2O00000O1001N10000000ULVFa2i9^MXFb2h9]MYFc2h9[MZFd2f9[M[Fe2f9YM[F2Df1Q:VN\\F3Fe1o9VN\\F3Hf1m9UN\\F4Je1k9UN]F3Kg1i9TN]F4Mg1f9SN^F5Ng1e9RN^F51g1b9lMdFXE]Ok:c0WEYOk:h0VETOl:m0UEnNn:S1_11O1O1O1O1O1O1O100O1O1O1O001O02N3M3M3M4L4L3M4Felk5"}, "label": "the dark maroon motorcycle behind the red one"}]} {"id": 390474, "image": "COCO_train2014_000000390474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue and black duffle bag\"."}], "task": "refering", "answer_template": "The \"a blue and black duffle bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 284, 285, 286, 287, 288, 289, 290, 291, 306, 307, 308, 309, 310, 311, 312, 313, 314, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.2051875, 0.5727777777777778, 0.721671875, 0.9971153846153845], "iscrowd": 0, "area": 44445.9024, "rle": {"size": [468, 640], "counts": "ohl11c>0O10000O100O100O100O1JMkA3T>OkA1S>9O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O1N2O1N2O1N2O1N2O1O1O1N2O1O1O1N2O1O1O1N2O2N1O1N2N2N2M3N2M3N2N2N2O1N2O1N2N2O1N2O1N2O1N2N2M3N2N2M3N2N2O1O1O1O1N2O1O1MRK_Fo4`9QK`FP5_9PKaFQ5]9PKcFQ5\\94O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1O1000000O1000000O1000000O100000000001O000000000000001O0000000000001O0000000000001O0000000000001O00000000010O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O0000000000001O0000000000001O00000000001O0000001O1O1O1O2N1O6J=C4L4L5K4L4L1O0000000000001O3M4L4lLUEW2o:eMUEW2n:eMWEW2m:eMWEW2m:dMWEY2m:cMSE]2c;0O01O1O001O001O1O001O0N3^NdCe0b in this image.", "objects": [{"patches": [217, 218, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 284, 285, 286, 287, 288, 289, 290, 291, 306, 307, 308, 309, 310, 311, 312, 313, 314, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384], "bbox": [0.2051875, 0.5727777777777778, 0.721671875, 0.9971153846153845], "iscrowd": 0, "area": 44445.9024, "rle": {"size": [468, 640], "counts": "ohl11c>0O10000O100O100O100O1JMkA3T>OkA1S>9O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1O100O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O100O1O1N2O1N2O1N2O1N2O1O1O1N2O1O1O1N2O1O1O1N2O2N1O1N2N2N2M3N2M3N2N2N2O1N2O1N2N2O1N2O1N2O1N2N2M3N2N2M3N2N2O1O1O1O1N2O1O1MRK_Fo4`9QK`FP5_9PKaFQ5]9PKcFQ5\\94O1O1N2O1O1O1N2O1O1O1N2O1O1N2O1O1O1N2O1O1O1O1000000O1000000O1000000O100000000001O000000000000001O0000000000001O0000000000001O0000000000001O00000000010O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O0000000000001O0000000000001O00000000001O0000001O1O1O1O2N1O6J=C4L4L5K4L4L1O0000000000001O3M4L4lLUEW2o:eMUEW2n:eMWEW2m:eMWEW2m:dMWEY2m:cMSE]2c;0O01O1O001O001O1O001O0N3^NdCe0b in this image.", "objects": [{"patches": [3, 24, 25, 26, 27, 28, 29, 47, 48, 49, 50, 51, 52, 53, 54, 70, 71, 72, 73, 74, 75, 76, 77, 93, 94, 95, 96, 97, 98, 99, 100, 116, 117, 118, 119, 120, 121, 122, 123, 143, 144, 145, 146], "bbox": [0.039890625, 0.042200854700854704, 0.381765625, 0.38245726495726495], "iscrowd": 0, "area": 24221.188349999997, "rle": {"size": [468, 640], "counts": "UP<3^>4L3L5L5K5K5J6K5K5K5J6K6J5K4K4M3M3M3L4M4L3M3M3L4M3M3M3O100O100O100O100O2O0O100O2O0O101N100O101N100O101N100O2O0O1O101N100O10000001O0000001O00001O0000001O00001O0000001O000010O000001O00001O0000001O00001O0000001O001O001O001O1O001O001O001O1O001O010O010O010O10O0010O01O1O010O001O010O001O010O001O010O001O10O01O0010O0000001O0000001O0000001O0000001O0000001O0000001O0000001N100O101N100O10001N100O100O2O0]NnDCS;;QECo:;TEDl:;WEBk:;YECg:;]EBd:<_EAc:=aE^Ob:?cE]O`:?eE]O]:`0gE\\O\\:a0iE[OY:b0lEYOW:d0nEXOU:d0o1M3M3M3M4L3Mamd5"}, "label": "a black duffle bag with green stripes and a small padlock"}]} {"id": 390474, "image": "COCO_train2014_000000390474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green duffel bag\"."}], "task": "refering", "answer_template": "The \"a green duffel bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 24, 25, 26, 27, 28, 29, 47, 48, 49, 50, 51, 52, 53, 54, 70, 71, 72, 73, 74, 75, 76, 77, 93, 94, 95, 96, 97, 98, 99, 100, 116, 117, 118, 119, 120, 121, 122, 123, 143, 144, 145, 146], "bbox": [0.039890625, 0.042200854700854704, 0.381765625, 0.38245726495726495], "iscrowd": 0, "area": 24221.188349999997, "rle": {"size": [468, 640], "counts": "UP<3^>4L3L5L5K5K5J6K5K5K5J6K6J5K4K4M3M3M3L4M4L3M3M3L4M3M3M3O100O100O100O100O2O0O100O2O0O101N100O101N100O101N100O2O0O1O101N100O10000001O0000001O00001O0000001O00001O0000001O000010O000001O00001O0000001O00001O0000001O001O001O001O1O001O001O001O1O001O010O010O010O10O0010O01O1O010O001O010O001O010O001O010O001O10O01O0010O0000001O0000001O0000001O0000001O0000001O0000001O0000001N100O101N100O10001N100O100O2O0]NnDCS;;QECo:;TEDl:;WEBk:;YECg:;]EBd:<_EAc:=aE^Ob:?cE]O`:?eE]O]:`0gE\\O\\:a0iE[OY:b0lEYOW:d0nEXOU:d0o1M3M3M3M4L3Mamd5"}, "label": "a green duffel bag"}]} {"id": 390474, "image": "COCO_train2014_000000390474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an ugly green suitcase\"."}], "task": "refering", "answer_template": "The \"an ugly green suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 24, 25, 26, 27, 28, 29, 30, 47, 48, 49, 50, 51, 52, 53, 54, 70, 71, 72, 73, 74, 75, 76, 77, 93, 94, 95, 96, 97, 98, 99, 116, 117, 118, 119, 120, 121, 122, 123, 143, 144, 145], "bbox": [0.043687500000000004, 0.049358974358974364, 0.379875, 0.39220085470085464], "iscrowd": 0, "area": 23877.2402, "rle": {"size": [468, 640], "counts": "_m<4X>9G9G8I8G8K6M2N3M2N3M4L7I6J7I2N2N2N2N1O2N2N2O1N2N2N2N1O2N2N2N2N2N2N2O100000000O2O00001O0000001O00001O00001O00001O0O10001O00001O00001O00001O0000001O000O2O00001N11O1O001O001N2O001O001O1O0O2O001O1O1O1O0O2O1O1O1O1O001N200O01N101O001N101O001O1N110O0010O010O01O10O01O010O000010O000001O0000001O01O0001O00001O001O010O00001O001O0010N100O100O10000O101N100O10001N100O2O0O2O000O2O0SNUE0m:MWE1i:MZE1g:N\\E0e:M^E1c:MaE1`:LcE3]:LfE1\\:LgE3Y:KkE2W:LkE3U:KoE2S:KPF4P:JTF4m9JUF4l9JXF4i9IZF5g9I]F5d9I_F4b9JaF5`9HdF6\\9HgF6[9HhF6X9HkF6W9GmF7S9GPG7i;N1O2M2OUYe5"}, "label": "an ugly green suitcase"}]} {"id": 390474, "image": "COCO_train2014_000000390474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue bag to the right of the black bag\"."}], "task": "refering", "answer_template": "The \"blue bag to the right of the black bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 291, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.186140625, 0.5758333333333333, 0.7065625, 0.981025641025641], "iscrowd": 0, "area": 39840.62215, "rle": {"size": [468, 640], "counts": "ZYg11b>1O1O1O100O1O1O1O1O1O100O1O1O1O001O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O001O100O1O1O1O1O1O1O100O1O1O1O1N2O1O1N2O1O1N2O1O0O2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O001N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O0O2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O001N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N21O0O10000000001O0000000000001O0000000000001O000000000O2O0000000000001O0000000000001O00000000001O00000O1000001O0000000000001O0000000000001O00000000001N1000000000001O0000000000001O00000000001O0000000O10001O0000000000001O0000000000001O00000000001O0O10000000001O0000000000001O0000001O4L5K5K5K5K5K5K5J4M2N2N2N3M2N2N2N2N3M2N2N2N2nNVDVOm;d0cDoN_;l0PEhNR;S1[1K5K5L4K4L5K5Kfae2"}, "label": "blue bag to the right of the black bag"}]} {"id": 390474, "image": "COCO_train2014_000000390474.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue duffel bag\"."}], "task": "refering", "answer_template": "The \"a blue duffel bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 239, 240, 241, 242, 243, 244, 245, 261, 262, 263, 264, 265, 266, 267, 268, 283, 284, 285, 286, 287, 288, 289, 290, 291, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.186140625, 0.5758333333333333, 0.7065625, 0.981025641025641], "iscrowd": 0, "area": 39840.62215, "rle": {"size": [468, 640], "counts": "ZYg11b>1O1O1O100O1O1O1O1O1O100O1O1O1O001O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O001O100O1O1O1O1O1O1O100O1O1O1O1N2O1O1N2O1O1N2O1O0O2O1O1N2O1O1N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O001N2O1O1N2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O0O2O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O001N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1N2O1O1N21O0O10000000001O0000000000001O0000000000001O000000000O2O0000000000001O0000000000001O00000000001O00000O1000001O0000000000001O0000000000001O00000000001N1000000000001O0000000000001O00000000001O0000000O10001O0000000000001O0000000000001O00000000001O0O10000000001O0000000000001O0000001O4L5K5K5K5K5K5K5J4M2N2N2N3M2N2N2N2N3M2N2N2N2nNVDVOm;d0cDoN_;l0PEhNR;S1[1K5K5L4K4L5K5Kfae2"}, "label": "a blue duffel bag"}]} {"id": 152922, "image": "COCO_train2014_000000152922.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tallest man in the image\"."}], "task": "refering", "answer_template": "The \"tallest man in the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 144, 145, 146, 167, 168, 169, 170, 190, 191, 192, 213, 214, 215, 236, 237, 238, 259, 260, 261, 262, 283, 284, 285, 306], "bbox": [0.2785, 0.32185011709601874, 0.40501562500000005, 0.8852927400468383], "iscrowd": 0, "area": 10900.3582, "rle": {"size": [427, 640], "counts": "i]Z22T=6J7J5L5L3L5K3N2M3N2M3kDdNY:^1dEROo9n0oE@e9`0XFOk0nNh6T1ZH<:oNV7e0]Hj0HPO1QOV7U1oHl2m6TMPIS3l6mLQIY3l6gLSIX3[6XKWJY6i5jIUJT6l5nISJP6o5QJPJn5P6TJoIj5R6YJlIf5U6[JjIc5W6_JhI`5X6bJgI\\5[6fJcIX5^6j00O2jM_IaMa6^2XKiKi4V4W2O0O1O20O3M3N2M4L3O1O1O2N1O1O1O2M4M4L5K5K5L37JN2N2N2K00O100O1O1VMcHD^7gNgHK3Z1W7jNQI]OOf1Q7lN\\InNKS2i6nNdJm0_5POhJi0\\5SOjJf0Z5VOnJb0W5ZOoJ?V5]OoJ in this image.", "objects": [{"patches": [122, 123, 144, 145, 146, 167, 168, 169, 170, 190, 191, 192, 213, 214, 215, 236, 237, 238, 259, 260, 261, 262, 283, 284, 285, 306], "bbox": [0.2785, 0.32185011709601874, 0.40501562500000005, 0.8852927400468383], "iscrowd": 0, "area": 10900.3582, "rle": {"size": [427, 640], "counts": "i]Z22T=6J7J5L5L3L5K3N2M3N2M3kDdNY:^1dEROo9n0oE@e9`0XFOk0nNh6T1ZH<:oNV7e0]Hj0HPO1QOV7U1oHl2m6TMPIS3l6mLQIY3l6gLSIX3[6XKWJY6i5jIUJT6l5nISJP6o5QJPJn5P6TJoIj5R6YJlIf5U6[JjIc5W6_JhI`5X6bJgI\\5[6fJcIX5^6j00O2jM_IaMa6^2XKiKi4V4W2O0O1O20O3M3N2M4L3O1O1O2N1O1O1O2M4M4L5K5K5L37JN2N2N2K00O100O1O1VMcHD^7gNgHK3Z1W7jNQI]OOf1Q7lN\\InNKS2i6nNdJm0_5POhJi0\\5SOjJf0Z5VOnJb0W5ZOoJ?V5]OoJ in this image.", "objects": [{"patches": [262, 263, 277, 278, 279, 292, 293, 294, 295, 307, 308, 309, 310, 322, 323, 324, 325, 338, 339], "bbox": [0.48823943661971836, 0.743109375, 0.7134741784037558, 0.970875], "iscrowd": 0, "area": 9002.026350000002, "rle": {"size": [640, 426], "counts": "[aR4g0^a0n1E:N2M3N2N2M3N1N3N1O1N2O2M2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O0011N1N3N2N2M3N2M3N2N2M2O2M3N2M3N2N2M3N1N3N1N2O2N1N3N1N3N1010O10N1O1O2N1O2O0O2M2O2N1CZ^O]Nha0a1Z^O^Nfa0a1]^OZNea0e1 in this image.", "objects": [{"patches": [262, 263, 277, 278, 279, 292, 293, 294, 295, 307, 308, 309, 310, 322, 323, 324, 325, 338, 339], "bbox": [0.48823943661971836, 0.743109375, 0.7134741784037558, 0.970875], "iscrowd": 0, "area": 9002.026350000002, "rle": {"size": [640, 426], "counts": "[aR4g0^a0n1E:N2M3N2N2M3N1N3N1O1N2O2M2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O0011N1N3N2N2M3N2M3N2N2M2O2M3N2M3N2N2M3N1N3N1N2O2N1N3N1N3N1010O10N1O1O2N1O2O0O2M2O2N1CZ^O]Nha0a1Z^O^Nfa0a1]^OZNea0e1 in this image.", "objects": [{"patches": [167, 168, 180, 181, 182, 183, 195, 196, 197, 198, 210, 211, 212, 213, 214, 215, 216, 225, 226, 227, 240, 241, 242], "bbox": [0.0, 0.5005625, 0.4340845070422535, 0.7420937500000001], "iscrowd": 0, "area": 10374.89825, "rle": {"size": [640, 426], "counts": "l;1lc03M3N2@?10000000000000000000O100h^OYOW?g0h@ZOX?g0e@\\OZ?d0d@^O\\?b0`@C_?=T@0l?S2O001O0O2O001O001O001O00001O0O2O001O001O001O001O001O0O2O001]Ob0O2M2N3I600O2O001O0O2O001N101O001O001O001O00XOYAXLf>e3_AYLa>d3dA[L[>c3iA[LW>b3nA]LQ>`3Z1L4M3N1O2N2M3N2QNW_Od0j`0ZOY_Od0i`0XO[_Oe0h`0XO[_Of0g`0VO]_Oh0oa0N2O1O1O2N1O1O001O1O1O001O1O001O1O001O001O1O001O001O001O001O1O1O1O1O1O1O0000000000000O10000000000000000000000O1000000000000000000000000O1000000000000000000O1000000O1000000O101O000O100000_Wf4"}, "label": "a man in gloves unloading luggage from a ship"}]} {"id": 202076, "image": "COCO_train2014_000000202076.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with black hair who is wearing a black shit and pulling on the rope\"."}], "task": "refering", "answer_template": "The \"a man with black hair who is wearing a black shit and pulling on the rope\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [167, 168, 180, 181, 182, 183, 195, 196, 197, 198, 210, 211, 212, 213, 214, 215, 216, 225, 226, 227, 240, 241, 242], "bbox": [0.0, 0.5005625, 0.4340845070422535, 0.7420937500000001], "iscrowd": 0, "area": 10374.89825, "rle": {"size": [640, 426], "counts": "l;1lc03M3N2@?10000000000000000000O100h^OYOW?g0h@ZOX?g0e@\\OZ?d0d@^O\\?b0`@C_?=T@0l?S2O001O0O2O001O001O001O00001O0O2O001O001O001O001O001O0O2O001]Ob0O2M2N3I600O2O001O0O2O001N101O001O001O001O00XOYAXLf>e3_AYLa>d3dA[L[>c3iA[LW>b3nA]LQ>`3Z1L4M3N1O2N2M3N2QNW_Od0j`0ZOY_Od0i`0XO[_Oe0h`0XO[_Of0g`0VO]_Oh0oa0N2O1O1O2N1O1O001O1O1O001O1O001O1O001O001O1O001O001O001O001O1O1O1O1O1O1O0000000000000O10000000000000000000000O1000000000000000000000000O1000000000000000000O1000000O1000000O101O000O100000_Wf4"}, "label": "a man with black hair who is wearing a black shit and pulling on the rope"}]} {"id": 390496, "image": "COCO_train2014_000000390496.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing in front of an airplane\"."}], "task": "refering", "answer_template": "The \"a man standing in front of an airplane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 132, 133, 134, 155, 156, 157, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 291, 292, 293, 294, 295, 296, 315, 316, 317, 318, 319, 338, 339, 340, 341, 342], "bbox": [0.65971875, 0.32236533957845437, 0.934359375, 0.9845667447306791], "iscrowd": 0, "area": 34129.01914999999, "rle": {"size": [427, 640], "counts": "XY`54o<;F:E;E;E in this image.", "objects": [{"patches": [110, 132, 133, 134, 155, 156, 157, 177, 178, 179, 180, 181, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 291, 292, 293, 294, 295, 296, 315, 316, 317, 318, 319, 338, 339, 340, 341, 342], "bbox": [0.65971875, 0.32236533957845437, 0.934359375, 0.9845667447306791], "iscrowd": 0, "area": 34129.01914999999, "rle": {"size": [427, 640], "counts": "XY`54o<;F:E;E;E in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 252, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 275, 290, 291, 298, 321], "bbox": [0.008953125000000001, 0.004800936768149883, 0.997046875, 0.8817564402810304], "iscrowd": 0, "area": 141376.51049999997, "rle": {"size": [427, 640], "counts": "if2P1[<^1bN4L001O000O10000000001O00000000000000001O0O1000000^IZM_2f2aMZM_2f2aM[M^2f2aMZM_2f2aM[M^2e2bM[M^2e2bM\\M]2d2cM\\M]2d2cM]M\\2c2dM]M\\2c2dM^M[2c2dM]M[2d2eM]MZ2c2fM]MZ2c2fM^MY2b2gM^MY2b2gM_MX2a2hM`MW2`2iM`MW2a2hM`MW2`2iM`MW2`2iMaMV2_2jMaMV2_2jMbMU2^2kMbMU2^2kMcMT2]2lMcMS2_2lMbMS2^2mMbMS2^2mMcMR2]2nMcMR2]2nMdMQ2\\2oMdMQ2\\2oMeMP2[2PNfMo1Z2QNfMo1[2PNfMo1Z2QNfMo1Z2QNgMn1Y2RNgMn1Y2RNhMm1X2SNhMm1X2SNiMk1X2UNhMk1Y2TNhMk1X2UNhMk1X2UNiMj1W2VNiMj1W2VNjMi1V2WNjMi1V2WNkMh1U2XNlMg1U2XNkMh1U2XNlMg1T2YNlMg1T2YNmMf1S2ZNmMf1S2ZNnMd1S2\\NmMd1S2\\NnMc1S2\\NmMd1S2\\NnMc1R2]NnMc1R2]NoMb1Q2^NoMb1Q2^NPNa1P2_NPNa1P2_NQN`1o1`NRN_1o1`NQN`1o1`NRN_1n1aNRN_1n1aNSN]1n1cNRN]1n1cNSN\\1m1dNSN\\1m1dNTN[1m1dNSN\\1m1dNTN[1l1eNTN[1l1eNUNZ1k1fNUNZ1k1fNVNY1j1gNVNY1j1gNWNX1j1gNWNX1i1hNWNX1i1hNXNW1h1iNXNV1i1jNXNU1h1kNXNU1h1kNYNT1g1lNYNT1h1kNYNT1g1lNYNS1h1mNYNR1g1nNYNR1g1nNZNQ1f1oNZNQ1f1oN[NP1e1PO[No0g1POZNo0f1QO[Nn0e1RO[Nn0e1RO\\Nm0d1SO\\Nl0e1TO\\Nk0d1UO\\Nk0d1UO]Nj0c1VO]Nj0d1UO]Nj0c1VO]Ni0d1WO]Nh0c1XO]Nh0c1XO^Ng0b1YO^Ng0b1YO_Nf0a1ZO_Ne0c1ZO^Ne0b1[O_Nd0a1\\O_Nd0a1\\O`Nc0`1]O`Nb0a1^O`Na0`1_O`Na0`1_OaN`0`1_O`Na0`1_OaN`0_1@aN?`1AaN>_1BaN>_1BbN=^1CbN=^1CcN;_1DaN<_1DbN;^1EcN:]1FcN:]1FdN9\\1GdN8]1HdN7\\1IdN7\\1IeN6\\1IdN7\\1IeN5\\1KdN5\\1KeN4[1LeN4[1LfN3Z1MfN3Z1MgN1[1NeN2[1NfN1Z1OgN0Y10gN0Y10hNOX11hNNY12hNMX13hNMY12hNMX13hNMX13iNKX15hNKX15iNJW16iNJW16jNIV17jNIW16jNHW18iNHW18jNGV19kNFU1:kNFU1:lNDU1 in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 252, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 275, 290, 291, 298, 321], "bbox": [0.008953125000000001, 0.004800936768149883, 0.997046875, 0.8817564402810304], "iscrowd": 0, "area": 141376.51049999997, "rle": {"size": [427, 640], "counts": "if2P1[<^1bN4L001O000O10000000001O00000000000000001O0O1000000^IZM_2f2aMZM_2f2aM[M^2f2aMZM_2f2aM[M^2e2bM[M^2e2bM\\M]2d2cM\\M]2d2cM]M\\2c2dM]M\\2c2dM^M[2c2dM]M[2d2eM]MZ2c2fM]MZ2c2fM^MY2b2gM^MY2b2gM_MX2a2hM`MW2`2iM`MW2a2hM`MW2`2iM`MW2`2iMaMV2_2jMaMV2_2jMbMU2^2kMbMU2^2kMcMT2]2lMcMS2_2lMbMS2^2mMbMS2^2mMcMR2]2nMcMR2]2nMdMQ2\\2oMdMQ2\\2oMeMP2[2PNfMo1Z2QNfMo1[2PNfMo1Z2QNfMo1Z2QNgMn1Y2RNgMn1Y2RNhMm1X2SNhMm1X2SNiMk1X2UNhMk1Y2TNhMk1X2UNhMk1X2UNiMj1W2VNiMj1W2VNjMi1V2WNjMi1V2WNkMh1U2XNlMg1U2XNkMh1U2XNlMg1T2YNlMg1T2YNmMf1S2ZNmMf1S2ZNnMd1S2\\NmMd1S2\\NnMc1S2\\NmMd1S2\\NnMc1R2]NnMc1R2]NoMb1Q2^NoMb1Q2^NPNa1P2_NPNa1P2_NQN`1o1`NRN_1o1`NQN`1o1`NRN_1n1aNRN_1n1aNSN]1n1cNRN]1n1cNSN\\1m1dNSN\\1m1dNTN[1m1dNSN\\1m1dNTN[1l1eNTN[1l1eNUNZ1k1fNUNZ1k1fNVNY1j1gNVNY1j1gNWNX1j1gNWNX1i1hNWNX1i1hNXNW1h1iNXNV1i1jNXNU1h1kNXNU1h1kNYNT1g1lNYNT1h1kNYNT1g1lNYNS1h1mNYNR1g1nNYNR1g1nNZNQ1f1oNZNQ1f1oN[NP1e1PO[No0g1POZNo0f1QO[Nn0e1RO[Nn0e1RO\\Nm0d1SO\\Nl0e1TO\\Nk0d1UO\\Nk0d1UO]Nj0c1VO]Nj0d1UO]Nj0c1VO]Ni0d1WO]Nh0c1XO]Nh0c1XO^Ng0b1YO^Ng0b1YO_Nf0a1ZO_Ne0c1ZO^Ne0b1[O_Nd0a1\\O_Nd0a1\\O`Nc0`1]O`Nb0a1^O`Na0`1_O`Na0`1_OaN`0`1_O`Na0`1_OaN`0_1@aN?`1AaN>_1BaN>_1BbN=^1CbN=^1CcN;_1DaN<_1DbN;^1EcN:]1FcN:]1FdN9\\1GdN8]1HdN7\\1IdN7\\1IeN6\\1IdN7\\1IeN5\\1KdN5\\1KeN4[1LeN4[1LfN3Z1MfN3Z1MgN1[1NeN2[1NfN1Z1OgN0Y10gN0Y10hNOX11hNNY12hNMX13hNMY12hNMX13hNMX13iNKX15hNKX15iNJW16iNJW16jNIV17jNIW16jNHW18iNHW18jNGV19kNFU1:kNFU1:lNDU1 in this image.", "objects": [{"patches": [115, 116, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164], "bbox": [0.0044843750000000005, 0.3281030444964871, 0.470171875, 0.5070725995316159], "iscrowd": 0, "area": 8381.242650000002, "rle": {"size": [427, 640], "counts": "g]1R1h;a0N2N2N2001O0O100000001O000O10001O0000000O101O0000000O101O000000001N10000000001O0O1000001O00000O10001O0000000O2O000000000O2O000000001O0O100000001O000O10001O0000000O101O0000000O101O000000001N10000000001O0O1000001O00000O10001O00000O101O000000000O2O000000001O0O100000001O000O10001O0000000O101O0000000O101O000000001N10000000001O0O0100O1O100O10O0100O100O10O0100O1O100O10O0100O100O100O010O100O1O100O010O100O100O010O100O1O100O010O100O100O10O0100O1O100O10O0100O100O10O0100O100O1O10O0100O100O100O010O100O1O100O01000001O00000O100000001O0O1000000000001N1000000000000O2O00000000000O1N3M2M3N2N2M3N2N2N2M4M2NQW]4"}, "label": "an airplane with three windows and white and black rudder"}]} {"id": 390496, "image": "COCO_train2014_000000390496.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"part of a small white airplane with three windows\"."}], "task": "refering", "answer_template": "The \"part of a small white airplane with three windows\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 116, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164], "bbox": [0.0044843750000000005, 0.3281030444964871, 0.470171875, 0.5070725995316159], "iscrowd": 0, "area": 8381.242650000002, "rle": {"size": [427, 640], "counts": "g]1R1h;a0N2N2N2001O0O100000001O000O10001O0000000O101O0000000O101O000000001N10000000001O0O1000001O00000O10001O0000000O2O000000000O2O000000001O0O100000001O000O10001O0000000O101O0000000O101O000000001N10000000001O0O1000001O00000O10001O00000O101O000000000O2O000000001O0O100000001O000O10001O0000000O101O0000000O101O000000001N10000000001O0O0100O1O100O10O0100O100O10O0100O1O100O10O0100O100O100O010O100O1O100O010O100O100O010O100O1O100O010O100O100O10O0100O1O100O10O0100O100O10O0100O100O1O10O0100O100O100O010O100O1O100O01000001O00000O100000001O0O1000000000001N1000000000000O2O00000000000O1N3M2M3N2N2M3N2N2N2M4M2NQW]4"}, "label": "part of a small white airplane with three windows"}]} {"id": 390496, "image": "COCO_train2014_000000390496.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person is boarding a small airplane in front of a man who is waiting to board\"."}], "task": "refering", "answer_template": "The \"a person is boarding a small airplane in front of a man who is waiting to board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 136, 157, 158, 159, 180, 181, 182, 205, 251, 274, 296, 297, 319, 320, 342, 343], "bbox": [0.8345312500000001, 0.3374941451990633, 0.958484375, 0.9968384074941453], "iscrowd": 0, "area": 6513.0388, "rle": {"size": [427, 640], "counts": "Vjn62Y=4bC5T;0\\D;d;e010000O001O1K5M32N1O2N000O1RDoNb;a1L4L001O1O00001O0O10001O1O1O1O001O1N2O001001O2M2OO1QKPNaNP2Z1VNeNk1U1ZNjNh1i0dNWO^1;nNES1[OhMbL\\1T4m0WOlMZL_1_4f0UOKl06QOLo05lNNU14^N7e1GWN;Q2_OlMc0[2XO`Mj0h2POUMR1R3hNkLY1]3bN_L`1h3\\NSLe1U4VNgKl1_4PN]KR2h4kMTKV2T5fMgJ\\2a5^M\\Jb2l5[MnIg2Z6UM`Im2]81O2O1N2N1n0SOe0[O:E4M3M3L4RObm:"}, "label": "a person is boarding a small airplane in front of a man who is waiting to board"}]} {"id": 390496, "image": "COCO_train2014_000000390496.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a yellow shirt inside the a plane\"."}], "task": "refering", "answer_template": "The \"a man in a yellow shirt inside the a plane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [134, 135, 136, 157, 158, 159, 180, 181, 182, 205, 251, 274, 296, 297, 319, 320, 342, 343], "bbox": [0.8345312500000001, 0.3374941451990633, 0.958484375, 0.9968384074941453], "iscrowd": 0, "area": 6513.0388, "rle": {"size": [427, 640], "counts": "Vjn62Y=4bC5T;0\\D;d;e010000O001O1K5M32N1O2N000O1RDoNb;a1L4L001O1O00001O0O10001O1O1O1O001O1N2O001001O2M2OO1QKPNaNP2Z1VNeNk1U1ZNjNh1i0dNWO^1;nNES1[OhMbL\\1T4m0WOlMZL_1_4f0UOKl06QOLo05lNNU14^N7e1GWN;Q2_OlMc0[2XO`Mj0h2POUMR1R3hNkLY1]3bN_L`1h3\\NSLe1U4VNgKl1_4PN]KR2h4kMTKV2T5fMgJ\\2a5^M\\Jb2l5[MnIg2Z6UM`Im2]81O2O1N2N1n0SOe0[O:E4M3M3L4RObm:"}, "label": "a man in a yellow shirt inside the a plane"}]} {"id": 103778, "image": "COCO_train2014_000000103778.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bowl of some sticky food possibly applesauce\"."}], "task": "refering", "answer_template": "The \"a bowl of some sticky food possibly applesauce\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 104, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 151, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 258, 259, 260, 261, 262, 263, 264, 265, 266, 282, 283, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310], "bbox": [0.205859375, 0.2770257611241218, 0.64615625, 0.9008899297423888], "iscrowd": 0, "area": 48005.01535000001, "rle": {"size": [427, 640], "counts": "fYg18k< in this image.", "objects": [{"patches": [103, 104, 124, 125, 126, 127, 144, 145, 146, 147, 148, 149, 150, 151, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 258, 259, 260, 261, 262, 263, 264, 265, 266, 282, 283, 284, 285, 286, 287, 288, 289, 307, 308, 309, 310], "bbox": [0.205859375, 0.2770257611241218, 0.64615625, 0.9008899297423888], "iscrowd": 0, "area": 48005.01535000001, "rle": {"size": [427, 640], "counts": "fYg18k< in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 145, 146, 147, 148, 149, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 214, 215, 217, 218, 237, 238], "bbox": [0.280375, 0.34763466042154567, 0.52475, 0.7004449648711945], "iscrowd": 0, "area": 11783.277900000001, "rle": {"size": [427, 640], "counts": "Pd[21[=0O100O100O1O1O1O00100O1O1O2N3M2N2N3YC1l;m0J7I1O1O1O100O1O1O1aEVN[9j1`FkNP9U1kF\\Oi8b0TGCk8;SGKk83TG2j8MTG7k8GSG?j8@UGe0i8ZOUGj0j8TOUGo0k8oNSGU1n8c11O1K5K6J5I7YOg0N2N3M2N2N2N2NnFQNm6l1YHPOg7n0mG_OT8?fGIZ85dGO\\8ObG4_8J_G:a8D]G`0c8BXGa0h8P210000O10000O1[Oe0A?B?M2N2N2N2N2N2M3N2N2N2N2N2O1N2O1O1O2N1O1O1O1O11O2N1O1O1O1O2N100O1O2N101N1O1O2O0Ob0^OP1PO4MO0O10001O00001N10002N2N1M4K4K6K5K4J7J:E>C=B:G9F:7H2O1O0O2N1OO2O0O5K;ENO2L3M4L4L3Mj_n3"}, "label": "a baby zebra getting help from his mother to help him stand up just moments after he was born"}]} {"id": 210279, "image": "COCO_train2014_000000210279.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby zeebra struggling to walk\"."}], "task": "refering", "answer_template": "The \"a baby zeebra struggling to walk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 145, 146, 147, 148, 149, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 214, 215, 217, 218, 237, 238], "bbox": [0.280375, 0.34763466042154567, 0.52475, 0.7004449648711945], "iscrowd": 0, "area": 11783.277900000001, "rle": {"size": [427, 640], "counts": "Pd[21[=0O100O100O1O1O1O00100O1O1O2N3M2N2N3YC1l;m0J7I1O1O1O100O1O1O1aEVN[9j1`FkNP9U1kF\\Oi8b0TGCk8;SGKk83TG2j8MTG7k8GSG?j8@UGe0i8ZOUGj0j8TOUGo0k8oNSGU1n8c11O1K5K6J5I7YOg0N2N3M2N2N2N2NnFQNm6l1YHPOg7n0mG_OT8?fGIZ85dGO\\8ObG4_8J_G:a8D]G`0c8BXGa0h8P210000O10000O1[Oe0A?B?M2N2N2N2N2N2M3N2N2N2N2N2O1N2O1O1O2N1O1O1O1O11O2N1O1O1O1O2N100O1O2N101N1O1O2O0Ob0^OP1PO4MO0O10001O00001N10002N2N1M4K4K6K5K4J7J:E>C=B:G9F:7H2O1O0O2N1OO2O0O5K;ENO2L3M4L4L3Mj_n3"}, "label": "a baby zeebra struggling to walk"}]} {"id": 210279, "image": "COCO_train2014_000000210279.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebras neck is bent down\"."}], "task": "refering", "answer_template": "The \"a zebras neck is bent down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 163, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.0, 0.48587822014051524, 0.42090625, 0.9981733021077284], "iscrowd": 0, "area": 41110.86120000001, "rle": {"size": [427, 640], "counts": "m9]3`6^31O0000001O000000001O0000001O000000001O0000001O000000001O000000001O0000001O000000001O000O101O0000001O001O001O00001O001O001O000N3L3N3N101N1O2O0O1O2O0O2N1O2O0O2N100O2N101N1O2N100O2N101N1O2O0O2N1O101N1O2O0O2N10001O0010O01O001O01O010O100O100O1O100O1O100O1O00100O1O100O1O100O1O100O1O10O07I01O00001O00001O0000001O00001O00001O00001O00001O00001O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N3M2N2N4L4L5K4L4L5K4L4L[Yj4"}, "label": "a zebras neck is bent down"}]} {"id": 210279, "image": "COCO_train2014_000000210279.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra grazing in the field\"."}], "task": "refering", "answer_template": "The \"a zebra grazing in the field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [161, 162, 163, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 261, 276, 277, 278, 279, 280, 281, 282, 283, 284, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.0, 0.48587822014051524, 0.42090625, 0.9981733021077284], "iscrowd": 0, "area": 41110.86120000001, "rle": {"size": [427, 640], "counts": "m9]3`6^31O0000001O000000001O0000001O000000001O0000001O000000001O000000001O0000001O000000001O000O101O0000001O001O001O00001O001O001O000N3L3N3N101N1O2O0O1O2O0O2N1O2O0O2N100O2N101N1O2N100O2N101N1O2O0O2N1O101N1O2O0O2N10001O0010O01O001O01O010O100O100O1O100O1O100O1O00100O1O100O1O100O1O100O1O10O07I01O00001O00001O0000001O00001O00001O00001O00001O00001O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N2N3M2N3M2N2N4L4L5K4L4L5K4L4L[Yj4"}, "label": "a zebra grazing in the field"}]} {"id": 210279, "image": "COCO_train2014_000000210279.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"female zebra\"."}], "task": "refering", "answer_template": "The \"female zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 57, 58, 59, 60, 61, 62, 80, 81, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 125, 126, 129, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 267, 268, 269], "bbox": [0.44926562499999995, 0.11878220140515222, 0.8657031249999999, 0.7675644028103045], "iscrowd": 0, "area": 34970.74874999999, "rle": {"size": [427, 640], "counts": "ZWh35T=4K5K6K4N2M101N1N3N1N2N3N1N3N1N2O1N2O0O2N1O2I6J7J510O10O010O10O010O010O01O010O10J5K6O010O0010O00010O01O010O010O00010O0010O01O0101N2O2M2N2O2M2O1N2O10O001O1O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001lFYMb7h2\\HYMd7g2[HZMe7g2YHZMg7f2XH[Mh7f2VH[Mj7e2UH\\Mk7i2oGXMR8P3dGQM\\8X3ZGiLg8_3oFbLl3FU1V4jNkKV1U4hNmKY1R4fNoKZ1Q4eNPL[1S4aNnK`1U4ZNmKf1X4TNiKm1Z4nMgKR2]4iMdKW2`4dMaK]2b4]M`Kc2d4XM]Ki2f4RM[Kn2n4hLSKX3`5SLbJn3Q6]KPJc4Z6RKgIn4\\6nJeIS5\\6iJfIW5\\6fJeI[5U700001O001N100O2N100O2N101N100O2N101N3N1N3M3L4L3L5L4K3N2N1N3N2M3N2M3N2M3N1N3N2M3N2M3N2N2M2O2M3N2M3N2N23M3M4K4M3M3M4L3M3M3M^NdHYMW7h2mHWMP7i2TIVMi6j2ZIVMc6j2aIVM[6j2iIUMT6k2PJTMm5l2VJTMg5k2^JQMa5P3cJkL^5T3fJhL[5X3iJdLW5[3nJ`LS5_3RK\\Lo4d3TKXLm4g3XKTLj4k3ZKPLg4o3]KnKc4R4[2O2O0O1O101N100O100O2`2_M:F2N2O1N1O0001O000O10001O2M2YOg0B?@?B>H9J5K6J5K5L5J5K5K6J5K5K6K4@a0fNohS1"}, "label": "female zebra"}]} {"id": 210279, "image": "COCO_train2014_000000210279.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult zebra is licking a young zebra\"."}], "task": "refering", "answer_template": "The \"an adult zebra is licking a young zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 57, 58, 59, 60, 61, 62, 80, 81, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 125, 126, 129, 130, 131, 132, 133, 134, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 245, 246, 247, 248, 249, 267, 268, 269], "bbox": [0.44926562499999995, 0.11878220140515222, 0.8657031249999999, 0.7675644028103045], "iscrowd": 0, "area": 34970.74874999999, "rle": {"size": [427, 640], "counts": "ZWh35T=4K5K6K4N2M101N1N3N1N2N3N1N3N1N2O1N2O0O2N1O2I6J7J510O10O010O10O010O010O01O010O10J5K6O010O0010O00010O01O010O010O00010O0010O01O0101N2O2M2N2O2M2O1N2O10O001O1O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001O1O001O001O001lFYMb7h2\\HYMd7g2[HZMe7g2YHZMg7f2XH[Mh7f2VH[Mj7e2UH\\Mk7i2oGXMR8P3dGQM\\8X3ZGiLg8_3oFbLl3FU1V4jNkKV1U4hNmKY1R4fNoKZ1Q4eNPL[1S4aNnK`1U4ZNmKf1X4TNiKm1Z4nMgKR2]4iMdKW2`4dMaK]2b4]M`Kc2d4XM]Ki2f4RM[Kn2n4hLSKX3`5SLbJn3Q6]KPJc4Z6RKgIn4\\6nJeIS5\\6iJfIW5\\6fJeI[5U700001O001N100O2N100O2N101N100O2N101N3N1N3M3L4L3L5L4K3N2N1N3N2M3N2M3N2M3N1N3N2M3N2M3N2N2M2O2M3N2M3N2N23M3M4K4M3M3M4L3M3M3M^NdHYMW7h2mHWMP7i2TIVMi6j2ZIVMc6j2aIVM[6j2iIUMT6k2PJTMm5l2VJTMg5k2^JQMa5P3cJkL^5T3fJhL[5X3iJdLW5[3nJ`LS5_3RK\\Lo4d3TKXLm4g3XKTLj4k3ZKPLg4o3]KnKc4R4[2O2O0O1O101N100O100O2`2_M:F2N2O1N1O0001O000O10001O2M2YOg0B?@?B>H9J5K6J5K5L5J5K5K6J5K5K6K4@a0fNohS1"}, "label": "an adult zebra is licking a young zebra"}]} {"id": 259443, "image": "COCO_train2014_000000259443.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on the left in the white cap\"."}], "task": "refering", "answer_template": "The \"man on the left in the white cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 53, 54, 55, 56, 76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 168, 169, 170, 171, 191, 192, 193, 194, 215, 216, 217, 238, 239, 240, 261, 262, 263, 284, 285, 286], "bbox": [0.28696875, 0.07619444444444444, 0.47615625, 0.9904166666666667], "iscrowd": 0, "area": 22362.018700000008, "rle": {"size": [360, 640], "counts": "]jP25Q;6J6I7J6J6J6J6J6J6I7J6J6J6J6J6J4L1N3N`NVHEi75aHI]72lHKS71UIMi64ZIIe69^IDa6C>A>B>7J?UN[HSOU80WH2X8ROTHo0Q92O1N3N1N2O2M2O1N2O2M2O1N3N1N2O3L6K5Jj0SNmLSJY3^5VM[JR3T5]MgJi2i4fMQKb2h4U2J7I6K6I6J7I7J6I8H8H8H7J7H8H`0@R1oN\\`e3"}, "label": "man on the left in the white cap"}]} {"id": 259443, "image": "COCO_train2014_000000259443.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a tattoo holding a tennis racket\"."}], "task": "refering", "answer_template": "The \"a man with a tattoo holding a tennis racket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 53, 54, 55, 56, 76, 77, 78, 79, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 168, 169, 170, 171, 191, 192, 193, 194, 215, 216, 217, 238, 239, 240, 261, 262, 263, 284, 285, 286], "bbox": [0.28696875, 0.07619444444444444, 0.47615625, 0.9904166666666667], "iscrowd": 0, "area": 22362.018700000008, "rle": {"size": [360, 640], "counts": "]jP25Q;6J6I7J6J6J6J6J6J6I7J6J6J6J6J6J4L1N3N`NVHEi75aHI]72lHKS71UIMi64ZIIe69^IDa6C>A>B>7J?UN[HSOU80WH2X8ROTHo0Q92O1N3N1N2O2M2O1N2O2M2O1N3N1N2O3L6K5Jj0SNmLSJY3^5VM[JR3T5]MgJi2i4fMQKb2h4U2J7I6K6I6J7I7J6I8H8H8H7J7H8H`0@R1oN\\`e3"}, "label": "a man with a tattoo holding a tennis racket"}]} {"id": 259443, "image": "COCO_train2014_000000259443.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with blue shirt\"."}], "task": "refering", "answer_template": "The \"a man with blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 34, 35, 54, 56, 57, 58, 59, 76, 77, 79, 80, 81, 82, 102, 103, 104, 105, 125, 126, 127, 128, 148, 149, 150, 151, 171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 218, 219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 287, 288, 289, 290], "bbox": [0.32665625, 0.05305555555555556, 0.6545625, 0.9880555555555557], "iscrowd": 0, "area": 22979.283399999993, "rle": {"size": [360, 640], "counts": "TbY21U;2N2N2N3N100O100O10O10O10O01000O101N1O1O2O0O2NP[c00nd\\O8J7J5J7IQ1oNi0WO:F9G:F9G9H9F9H9F9cNfJ\\L[5`3VKQLk4l3ZKRLf4k3_KRLb4k3cKSL]4l3fKQL[4o3hKnKX4Q4kKmKU4R4nKkKS4T4e1QMmH`1Y7UNQIg1T7nMWIm1m6jM\\IR2R8hNTGIS9KSG5S9_OTG`0k9O0O2O01O0O101O001O00001O0O2O000hNZOnGg0g7DUH?d7JWH:b7MZH6_73[H1a6jNiIW3U6QM\\IX3d0\\Lo4]4lJhKR5Z4hJkKW5W4dJmK[5U4_JoKa5R4[JRLd5j4O001O1O1O13M5K2N1O1N101O1N101N10000O10001N1002N2N2N3M2O100RKgJi3Y5ULlJh3T5ULRKh3n4ULXKh3h4VL^Kf3b4WLdKf30mKU3:QMk4Q3SKTMh4n2UKXMf4j2WK\\Mc4g2[K]Ma4f2[K`M`4h4M2M4M2M3cLoH^2S7\\MWI]2l6\\M_I^2c6ZMjI^2g7fNYGFm8EbG;f901N10000O2O0O1O1O2N1O1O1N3N1O1O2N1N2Ob[]2"}, "label": "a man with blue shirt"}]} {"id": 259443, "image": "COCO_train2014_000000259443.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with in blue shirt with tennis racket on his hand\"."}], "task": "refering", "answer_template": "The \"a man with in blue shirt with tennis racket on his hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 34, 35, 54, 56, 57, 58, 59, 76, 77, 79, 80, 81, 82, 102, 103, 104, 105, 125, 126, 127, 128, 148, 149, 150, 151, 171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 218, 219, 220, 221, 241, 242, 243, 244, 264, 265, 266, 267, 287, 288, 289, 290], "bbox": [0.32665625, 0.05305555555555556, 0.6545625, 0.9880555555555557], "iscrowd": 0, "area": 22979.283399999993, "rle": {"size": [360, 640], "counts": "TbY21U;2N2N2N3N100O100O10O10O10O01000O101N1O1O2O0O2NP[c00nd\\O8J7J5J7IQ1oNi0WO:F9G:F9G9H9F9H9F9cNfJ\\L[5`3VKQLk4l3ZKRLf4k3_KRLb4k3cKSL]4l3fKQL[4o3hKnKX4Q4kKmKU4R4nKkKS4T4e1QMmH`1Y7UNQIg1T7nMWIm1m6jM\\IR2R8hNTGIS9KSG5S9_OTG`0k9O0O2O01O0O101O001O00001O0O2O000hNZOnGg0g7DUH?d7JWH:b7MZH6_73[H1a6jNiIW3U6QM\\IX3d0\\Lo4]4lJhKR5Z4hJkKW5W4dJmK[5U4_JoKa5R4[JRLd5j4O001O1O1O13M5K2N1O1N101O1N101N10000O10001N1002N2N2N3M2O100RKgJi3Y5ULlJh3T5ULRKh3n4ULXKh3h4VL^Kf3b4WLdKf30mKU3:QMk4Q3SKTMh4n2UKXMf4j2WK\\Mc4g2[K]Ma4f2[K`M`4h4M2M4M2M3cLoH^2S7\\MWI]2l6\\M_I^2c6ZMjI^2g7fNYGFm8EbG;f901N10000O2O0O1O1O2N1O1O1N3N1O1O2N1N2Ob[]2"}, "label": "a man with in blue shirt with tennis racket on his hand"}]} {"id": 243066, "image": "COCO_train2014_000000243066.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue ski standing in the snow\"."}], "task": "refering", "answer_template": "The \"blue ski standing in the snow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 55, 56, 67, 68, 80, 92, 93, 104, 105, 116, 117, 128, 129, 140, 141, 153, 165, 177], "bbox": [0.6393993993993994, 0.19247999999999998, 0.827057057057057, 0.81736], "iscrowd": 0, "area": 6728.592399999997, "rle": {"size": [500, 333], "counts": "\\cX32a?2TANS>4jA3P>NnA9m=HnAa0l=AQBg0h=[OUBl0e=UOZBQ1a=PO]BV1]=f0J6J6J6K5J7I6J5L5J6J5K7I7J7H7I7I710iEPKe9a53N2G:G8I7J6K6O000102M3M01O1O0J7G9I6G9G0001O08H4L3M4L5L5PNTCl0Y=`NXCV1m=F:G9F:Fhmk0"}, "label": "blue ski standing in the snow"}]} {"id": 243066, "image": "COCO_train2014_000000243066.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue ski\"."}], "task": "refering", "answer_template": "The \"blue ski\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 55, 56, 67, 68, 80, 92, 93, 104, 105, 116, 117, 128, 129, 140, 141, 153, 165, 177], "bbox": [0.6393993993993994, 0.19247999999999998, 0.827057057057057, 0.81736], "iscrowd": 0, "area": 6728.592399999997, "rle": {"size": [500, 333], "counts": "\\cX32a?2TANS>4jA3P>NnA9m=HnAa0l=AQBg0h=[OUBl0e=UOZBQ1a=PO]BV1]=f0J6J6J6K5J7I6J5L5J6J5K7I7J7H7I7I710iEPKe9a53N2G:G8I7J6K6O000102M3M01O1O0J7G9I6G9G0001O08H4L3M4L5L5PNTCl0Y=`NXCV1m=F:G9F:Fhmk0"}, "label": "blue ski"}]} {"id": 243066, "image": "COCO_train2014_000000243066.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a set of blue skiis\"."}], "task": "refering", "answer_template": "The \"a set of blue skiis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 55, 56, 67, 68, 80, 92, 93, 104, 105, 116, 117, 128, 129, 140, 141, 153, 165, 177], "bbox": [0.6393993993993994, 0.19247999999999998, 0.827057057057057, 0.81736], "iscrowd": 0, "area": 6728.592399999997, "rle": {"size": [500, 333], "counts": "\\cX32a?2TANS>4jA3P>NnA9m=HnAa0l=AQBg0h=[OUBl0e=UOZBQ1a=PO]BV1]=f0J6J6J6K5J7I6J5L5J6J5K7I7J7H7I7I710iEPKe9a53N2G:G8I7J6K6O000102M3M01O1O0J7G9I6G9G0001O08H4L3M4L5L5PNTCl0Y=`NXCV1m=F:G9F:Fhmk0"}, "label": "a set of blue skiis"}]} {"id": 243066, "image": "COCO_train2014_000000243066.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the ski that is propped up on the right\"."}], "task": "refering", "answer_template": "The \"the ski that is propped up on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 46, 58, 70, 81, 82, 93, 94, 105, 106, 117, 118], "bbox": [0.78, 0.15584, 0.9087087087087088, 0.55064], "iscrowd": 0, "area": 3576.114599999999, "rle": {"size": [500, 333], "counts": "`Vo38V?"}, "label": "the ski that is propped up on the right"}]} {"id": 243066, "image": "COCO_train2014_000000243066.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"neon green skis\"."}], "task": "refering", "answer_template": "The \"neon green skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [34, 46, 58, 70, 81, 82, 93, 94, 105, 106, 117, 118], "bbox": [0.78, 0.15584, 0.9087087087087088, 0.55064], "iscrowd": 0, "area": 3576.114599999999, "rle": {"size": [500, 333], "counts": "`Vo38V?"}, "label": "neon green skis"}]} {"id": 128398, "image": "COCO_train2014_000000128398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"feta cheese coveres the foccacia on the table\"."}], "task": "refering", "answer_template": "The \"feta cheese coveres the foccacia on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 261, 262, 263, 264, 265, 266], "bbox": [0.14009375, 0.3258508158508158, 0.81040625, 0.7595571095571096], "iscrowd": 0, "area": 55653.71185000001, "rle": {"size": [429, 640], "counts": "knU12Z=1aN2]E1_:4^EM^:8_EJ^:;^EGb::\\EHb:;ZEGe: in this image.", "objects": [{"patches": [120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 261, 262, 263, 264, 265, 266], "bbox": [0.14009375, 0.3258508158508158, 0.81040625, 0.7595571095571096], "iscrowd": 0, "area": 55653.71185000001, "rle": {"size": [429, 640], "counts": "knU12Z=1aN2]E1_:4^EM^:8_EJ^:;^EGb::\\EHb:;ZEGe: in this image.", "objects": [{"patches": [100, 101, 123, 124, 145, 146, 166, 167, 168, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 239, 245, 246, 247, 248, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295], "bbox": [0.025374999999999998, 0.2876470588235294, 0.85059375, 0.8629176470588236], "iscrowd": 0, "area": 35986.79944999999, "rle": {"size": [425, 640], "counts": "lo61X=0O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O100O1000000O1000000O10000O1000000O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O1000000O10000M3F:F:F:J6O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1O1O100O100O10000O100O10000O100O100O10000O100O10000O100O100O10000O100O100O10000O100O10000O100O2O0O10000O100N2O1N2O1O1N2O1N[M`HE_7g6dNZIg00e0e6cNdIa0Hk0c6dNlI;BP1`6eNWJ3ZOX1^6eN_JMTO]1[6gNhJEoNc1X6gNkJFnNb1U6iNoJCmNd1S6hNRKDlNc1P6jNUKBlNd1n5iNXKCkNc1k5kN[KAkNd1i5jN^KBjNc1f5lNaK@jNd1d5kNdKAiNc1b5lNfK@iNd1_5mNiK_OiNc1]5mNlK_OhNd1Z5nNoK^OhNc1X5nNRL_OfNc1V5oN[LWOaNi1S5oNgLoNVNR2Q5POnMo0Q2PORNo0l1ROUNn0j1QOYNn0e1SO\\Nm0c1RO`Nm0_1SObNm0\\1TOfNk0Y1TOiNl0U1UOmNj0R1UOPOk0Q1SOQOl0o0ROSOn0m0QOUOn0k0POWOP1j0nNXOQ1h0mNZOS1f0lN\\OS1d0lN]OT1d0iN_OV1a0iN@W1`0gNCX1=gNDY1=dNF[1:dNG\\19bNI^17`NL_15^NMb13\\N0c1k501O001O001O001O001O00001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O0000000000O100000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O100ZOlNWET1i:oNTEQ1l:ROQEn0o:TOoDl0Q;WOlDi0T;ZOiDf0W;]OfDc0Y;@eD`0[;g000000000O100000000000000000001N100000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O1000000000000000000O100000000001O1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1N3L3N2M3M3M4M2M3M3M4L3N2M3M4L3N``W1"}, "label": "the front bike in the right hand picture"}]} {"id": 390565, "image": "COCO_train2014_000000390565.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange motorcycle\"."}], "task": "refering", "answer_template": "The \"the orange motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 123, 124, 145, 146, 166, 167, 168, 188, 189, 190, 191, 192, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 239, 245, 246, 247, 248, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295], "bbox": [0.025374999999999998, 0.2876470588235294, 0.85059375, 0.8629176470588236], "iscrowd": 0, "area": 35986.79944999999, "rle": {"size": [425, 640], "counts": "lo61X=0O1O1O100O1O1O100O1O1O1O100O1O1O100O1O1O100O1O1O1O100O100O1000000O1000000O10000O1000000O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O1000000O10000M3F:F:F:J6O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1N2N2N2O1O1O100O100O10000O100O10000O100O100O10000O100O10000O100O100O10000O100O100O10000O100O10000O100O2O0O10000O100N2O1N2O1O1N2O1N[M`HE_7g6dNZIg00e0e6cNdIa0Hk0c6dNlI;BP1`6eNWJ3ZOX1^6eN_JMTO]1[6gNhJEoNc1X6gNkJFnNb1U6iNoJCmNd1S6hNRKDlNc1P6jNUKBlNd1n5iNXKCkNc1k5kN[KAkNd1i5jN^KBjNc1f5lNaK@jNd1d5kNdKAiNc1b5lNfK@iNd1_5mNiK_OiNc1]5mNlK_OhNd1Z5nNoK^OhNc1X5nNRL_OfNc1V5oN[LWOaNi1S5oNgLoNVNR2Q5POnMo0Q2PORNo0l1ROUNn0j1QOYNn0e1SO\\Nm0c1RO`Nm0_1SObNm0\\1TOfNk0Y1TOiNl0U1UOmNj0R1UOPOk0Q1SOQOl0o0ROSOn0m0QOUOn0k0POWOP1j0nNXOQ1h0mNZOS1f0lN\\OS1d0lN]OT1d0iN_OV1a0iN@W1`0gNCX1=gNDY1=dNF[1:dNG\\19bNI^17`NL_15^NMb13\\N0c1k501O001O001O001O001O00001O001O001O001O001O001O001O00001O001O001O001O001O001O001O001O00001O001O001O001O0000000000O100000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O10000000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O100ZOlNWET1i:oNTEQ1l:ROQEn0o:TOoDl0Q;WOlDi0T;ZOiDf0W;]OfDc0Y;@eD`0[;g000000000O100000000000000000001N100000000000000000000O100000000000000000000O100000000000000000000O100000000000000000000O1000000000000000000O100000000001O1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1N3L3N2M3M3M4M2M3M3M4L3N2M3M4L3N``W1"}, "label": "the orange motorcycle"}]} {"id": 578990, "image": "COCO_train2014_000000578990.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person\"."}], "task": "refering", "answer_template": "The \"person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 123, 124, 125, 126, 127, 146, 147, 148, 149, 168, 169, 170, 171, 172, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 221, 243], "bbox": [0.34381249999999997, 0.27568750000000003, 0.637296875, 0.6140625], "iscrowd": 0, "area": 13554.742650000004, "rle": {"size": [480, 640], "counts": "eZW39a>`0A`0_O8VObNjCd1S<^NjCf1R<^NjCf1R<^NjCf1R<\\NlCl1m;UNRDP2h;h0K5L1N2O2M3N3L4M3O1N10OO1L41O0000N2N2N1O2O1O1O1O1O1O01000000000O10000000000001O001O1O1O1mLfE^1[:_NhE`1Y:]NjEb1W:]NkE`1Y:]NhEb1[:\\NeEb1`:eMQE<`0n1l:nMUEP2m:oMUEn1l:RNUEm1l:RNUEl1l:TNTEl1m:TNSEj1n:VNREi1o:WNREh1n:XNSEf1n:ZNREf1n:[NREc1o:]NREa1o:_NQEa1o:`NPE_1Q;aNPE^1P;bNPE]1Q;dNnD[1S;eNmD[1Q;iNmDV1S;lNlDT1S;nNmDP1S;ROlDn0T;ROlDm0U;ROlDn0T;ROlDm0T;SOmDl0T;TOmDk0S;TOnDk0R;VOnDj0Q;VOPEi0Q;WOPEh0o:XOREg0o:YOQEg0n:YOSEf0m:ZOUEe0k:[OUEd0k:\\OVEd0j:\\OWEc0h:]OYEb0h:^OYEb0e:^O\\Eb0c:^O_Eb0`:]ObEc0]:]OdEb0\\:]OeEd0Z:[OhEd0Y:ZOhEg0W:YOiEg0V:YOkEh0S:XOnEi0Q:WOoEi0P:WORFi0m9UOUFk0j9TOXFm0f9TOZFm0e9SO]Fk0b9UOaFj0]9WOeFg0[9XOhFg0V9YOmFe0S9ZOQGd0m8]OVGa0i8_OXG`0h8@YG`0e8A\\G>d8B\\G?b8B_G>`8B`G?_8@cG`0\\8_OeGb0Z8]OhGc0V8]OkGd0T8[OnGe0Q8QOYE3h2l0o7QOZENl2P1j7QO[Ho0f7PO[Ho0k:010O01O001O2O0O2O00N2N1O2N2N21O1O2O1N1O00O0100O010O01ON2N2O0O2O1O0O2O1O000O101N1O10O01O100N2O3L3M3NV`\\3"}, "label": "person"}]} {"id": 578990, "image": "COCO_train2014_000000578990.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a red jacket\"."}], "task": "refering", "answer_template": "The \"a man wearing a red jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 123, 124, 125, 126, 127, 146, 147, 148, 149, 168, 169, 170, 171, 172, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 220, 221, 243], "bbox": [0.34381249999999997, 0.27568750000000003, 0.637296875, 0.6140625], "iscrowd": 0, "area": 13554.742650000004, "rle": {"size": [480, 640], "counts": "eZW39a>`0A`0_O8VObNjCd1S<^NjCf1R<^NjCf1R<^NjCf1R<\\NlCl1m;UNRDP2h;h0K5L1N2O2M3N3L4M3O1N10OO1L41O0000N2N2N1O2O1O1O1O1O1O01000000000O10000000000001O001O1O1O1mLfE^1[:_NhE`1Y:]NjEb1W:]NkE`1Y:]NhEb1[:\\NeEb1`:eMQE<`0n1l:nMUEP2m:oMUEn1l:RNUEm1l:RNUEl1l:TNTEl1m:TNSEj1n:VNREi1o:WNREh1n:XNSEf1n:ZNREf1n:[NREc1o:]NREa1o:_NQEa1o:`NPE_1Q;aNPE^1P;bNPE]1Q;dNnD[1S;eNmD[1Q;iNmDV1S;lNlDT1S;nNmDP1S;ROlDn0T;ROlDm0U;ROlDn0T;ROlDm0T;SOmDl0T;TOmDk0S;TOnDk0R;VOnDj0Q;VOPEi0Q;WOPEh0o:XOREg0o:YOQEg0n:YOSEf0m:ZOUEe0k:[OUEd0k:\\OVEd0j:\\OWEc0h:]OYEb0h:^OYEb0e:^O\\Eb0c:^O_Eb0`:]ObEc0]:]OdEb0\\:]OeEd0Z:[OhEd0Y:ZOhEg0W:YOiEg0V:YOkEh0S:XOnEi0Q:WOoEi0P:WORFi0m9UOUFk0j9TOXFm0f9TOZFm0e9SO]Fk0b9UOaFj0]9WOeFg0[9XOhFg0V9YOmFe0S9ZOQGd0m8]OVGa0i8_OXG`0h8@YG`0e8A\\G>d8B\\G?b8B_G>`8B`G?_8@cG`0\\8_OeGb0Z8]OhGc0V8]OkGd0T8[OnGe0Q8QOYE3h2l0o7QOZENl2P1j7QO[Ho0f7PO[Ho0k:010O01O001O2O0O2O00N2N1O2N2N21O1O2O1N1O00O0100O010O01ON2N2O0O2O1O0O2O1O000O101N1O10O01O100N2O3L3M3NV`\\3"}, "label": "a man wearing a red jacket"}]} {"id": 292271, "image": "COCO_train2014_000000292271.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra eating grass on the right that is not next to any animals\"."}], "task": "refering", "answer_template": "The \"the zebra eating grass on the right that is not next to any animals\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 246, 247, 249, 251, 252, 269, 270, 272, 274, 275, 292, 293, 295, 297, 298], "bbox": [0.6994374999999999, 0.40449999999999997, 0.9876406249999998, 0.7640416666666667], "iscrowd": 0, "area": 14756.106750000006, "rle": {"size": [480, 640], "counts": "WXb69]>8G:M3O001O1O001O1O001O1ON2O1O2M3NmCk0a8QO`GS1^8kNaGZ1]8gN`G]1_8iNWG[1h8lNoFX1o8POfFU1Y9QO^FS1`9UOVFn0i9YOmEl0Q:S2O1N2N10O0O1N3nMRFPOo9o0ZFhNg9X1bF_N^9`1kFXNV9g1SGQNl8n1]GjMc8V2fGaM[8]2nG\\MQ8d2U2O2O2M2O2M2O2M2O1O2M2O2M2O1O20O01O001O01O01O0001N100000001O000O10001O0000001O0O1000001O0gEmMe7S2UHQNm7o1mGUNU8k1eGYN\\8i1]GZNf8f1TG\\NP9d1kF^NX9b1bF`Nb9a1WFbNl9^1nEdNV:\\1dEgN_:Y1[EiNh:X1REkNQ;V1iDkN[;Y24M3L4M3L4M2M4L4M3L5L3O1001O00001O00001O001O00001O00001O00001O001O3M4L4L4M2N3L?B8H3M3L5L3M2NO001O1O1O001O1O1O001N7WOPEVMc;X2bDbMQ in this image.", "objects": [{"patches": [177, 178, 179, 180, 181, 182, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 246, 247, 249, 251, 252, 269, 270, 272, 274, 275, 292, 293, 295, 297, 298], "bbox": [0.6994374999999999, 0.40449999999999997, 0.9876406249999998, 0.7640416666666667], "iscrowd": 0, "area": 14756.106750000006, "rle": {"size": [480, 640], "counts": "WXb69]>8G:M3O001O1O001O1O001O1ON2O1O2M3NmCk0a8QO`GS1^8kNaGZ1]8gN`G]1_8iNWG[1h8lNoFX1o8POfFU1Y9QO^FS1`9UOVFn0i9YOmEl0Q:S2O1N2N10O0O1N3nMRFPOo9o0ZFhNg9X1bF_N^9`1kFXNV9g1SGQNl8n1]GjMc8V2fGaM[8]2nG\\MQ8d2U2O2O2M2O2M2O2M2O1O2M2O2M2O1O20O01O001O01O01O0001N100000001O000O10001O0000001O0O1000001O0gEmMe7S2UHQNm7o1mGUNU8k1eGYN\\8i1]GZNf8f1TG\\NP9d1kF^NX9b1bF`Nb9a1WFbNl9^1nEdNV:\\1dEgN_:Y1[EiNh:X1REkNQ;V1iDkN[;Y24M3L4M3L4M2M4L4M3L5L3O1001O00001O00001O001O00001O00001O00001O001O3M4L4L4M2N3L?B8H3M3L5L3M2NO001O1O1O001O1O1O001N7WOPEVMc;X2bDbMQ in this image.", "objects": [{"patches": [161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 213, 233, 234, 235, 236, 258, 259, 260, 278, 280, 281, 282, 301, 302, 304, 305, 324, 325, 326, 327, 328], "bbox": [0.0, 0.4305833333333334, 0.33606250000000004, 0.8571666666666666], "iscrowd": 0, "area": 16356.747900000022, "rle": {"size": [480, 640], "counts": "Q8=[=X1000000O10000000000000000O100000O1000001O0000000000000000000000000000001O00000000000000000000000000001O00000O100000000000O10000oE_NVM0i8a1oIeNRMMn8^1nIkNnLKQ9[1oI]Oo5c0oI@P6`0jIGU68lIHT68mI]OUMVOl8^1oIXO[MYOe8`1PJhNkMFU8b1PJeNPNHo7c1QJcNTNIi7e1SJ_NU7b1jH[NY7e1Q3001N101O001O0O2O001O001O0O2O1O1O2M2O2M3M2N3N1eFiMk5Y2nIoMo5S2gIYNU6i1kIXNR6j1mIVNR6l1nIRNR6P2mIoMS6S2mIkMS6W2mIgMS6[2nIbMR6`2oI]MQ6d2QJYMo5g2SJWMm5i2VJTMj5l2YJQMh5m2\\JPMd5P3_JmLa5S3V3O1000000O2N1O1N2O1O11O100O1O1O2^EXMi8h2TG\\Mj8f2SG\\Mm8e2oF_Mo8d2mF_MQ9c2kFaMT9`2gFdMX9]2cFiM\\9X2^FmMa9T2\\FPNc9P2[FRNd9o1YFTNg9c30001kFjJ^8U5_GmJb8S5[GoJg8P5VGSKj8m4TGTKn8k4oFWKR9i4lFXKV9f4iF[KX9d4gF]K[9a4dF`K^9_4`FbKb9m42N2N2M3N2N2N2N2N3M2N2@jERLY:X3]FeLf9V3U1L5K5J5L6J8HY1gN2\\OWBOl=OWBLl=3a00001O001O00001OO100O100O2O0010O0001O0001O0O10000000000O1N2O1O1LmlV6"}, "label": "the zebra that is behind the goat , eating"}]} {"id": 292271, "image": "COCO_train2014_000000292271.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the big zebra in the biddle\"."}], "task": "refering", "answer_template": "The \"the big zebra in the biddle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 192, 193, 195, 197, 198, 215, 216, 218, 220, 221], "bbox": [0.36984374999999997, 0.3153125, 0.644828125, 0.5616666666666668], "iscrowd": 0, "area": 11793.668750000003, "rle": {"size": [480, 640], "counts": "XY_34g>5K5J6K6J5N2N2N2aCZOb:2XDOi0=l:HaDC=m0o:C`Ea0^:B^E`0b:B[E?d:DXE>h:DUEH]OC^;h0QEEd;>XDBj;`0oCCR in this image.", "objects": [{"patches": [124, 125, 126, 127, 128, 146, 147, 148, 149, 150, 151, 152, 169, 170, 171, 172, 173, 174, 175, 192, 193, 195, 197, 198, 215, 216, 218, 220, 221], "bbox": [0.36984374999999997, 0.3153125, 0.644828125, 0.5616666666666668], "iscrowd": 0, "area": 11793.668750000003, "rle": {"size": [480, 640], "counts": "XY_34g>5K5J6K6J5N2N2N2aCZOb:2XDOi0=l:HaDC=m0o:C`Ea0^:B^E`0b:B[E?d:DXE>h:DUEH]OC^;h0QEEd;>XDBj;`0oCCR in this image.", "objects": [{"patches": [47, 48, 64, 65, 66, 81, 82, 83, 84, 98, 99, 100, 101, 102, 116, 117, 118, 119], "bbox": [0.46536, 0.18296, 0.709, 0.5259466666666667], "iscrowd": 0, "area": 8507.527600000001, "rle": {"size": [375, 500], "counts": "V`e23b;2N2O1O1N2O2N1N2O1O1N2O2N1N2O1O1N3N1O1N2O1O1N3N1O1N2O1O1N3N1O1N2O1O1N3N1O1N2O1O1N2O1N3N1N2O1N2O1O1N2O1N3N1N2O1N2O1N2O1O2M2O1N2O1N2O1O1O2M2O1O1O1O1N2O1O1O2N1N2O1O1O1O1N2000O100000000O1O1L4L5K4L4L4L4M3L4L4L4L4L4K6K4L4L4K5K5L4K5L4K5L4K5K5L4L4Kl\\e1"}, "label": "an advertisement on a sheet of paper next to a laptop"}]} {"id": 423343, "image": "COCO_train2014_000000423343.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a magazine partially behind an open laptop\"."}], "task": "refering", "answer_template": "The \"a magazine partially behind an open laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 64, 65, 66, 81, 82, 83, 84, 98, 99, 100, 101, 102, 116, 117, 118, 119], "bbox": [0.46536, 0.18296, 0.709, 0.5259466666666667], "iscrowd": 0, "area": 8507.527600000001, "rle": {"size": [375, 500], "counts": "V`e23b;2N2O1O1N2O2N1N2O1O1N2O2N1N2O1O1N3N1O1N2O1O1N3N1O1N2O1O1N3N1O1N2O1O1N3N1O1N2O1O1N2O1N3N1N2O1N2O1O1N2O1N3N1N2O1N2O1N2O1O2M2O1N2O1N2O1O1O2M2O1O1O1O1N2O1O1O2N1N2O1O1O1O1N2000O100000000O1O1L4L5K4L4L4L4M3L4L4L4L4L4K6K4L4L4K5K5L4K5L4K5L4K5K5L4L4Kl\\e1"}, "label": "a magazine partially behind an open laptop"}]} {"id": 423343, "image": "COCO_train2014_000000423343.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the newspaper titled cityam\"."}], "task": "refering", "answer_template": "The \"the newspaper titled cityam\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 18, 19, 20, 21, 22, 36, 37, 38, 39, 40, 54, 55, 56, 57, 58], "bbox": [0.00714, 0.056613333333333335, 0.24286, 0.3105866666666667], "iscrowd": 0, "area": 8415.907599999999, "rle": {"size": [375, 500], "counts": "ea1k0n9n0A?O10O100000O10O10000000O10O100000O10O100000O10O100000O10O100000O10O100000O1000O1000O1000O1000O1000O1000O1000O1000O100000O10O100000O10O100000O10O100000O10O100000O10O10000000O010000000O010000000O010000000O010000000O10O10000000O01b0^OVhZ4"}, "label": "the newspaper titled cityam"}]} {"id": 423343, "image": "COCO_train2014_000000423343.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a newspaper\"."}], "task": "refering", "answer_template": "The \"a newspaper\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 18, 19, 20, 21, 22, 36, 37, 38, 39, 40, 54, 55, 56, 57, 58], "bbox": [0.00714, 0.056613333333333335, 0.24286, 0.3105866666666667], "iscrowd": 0, "area": 8415.907599999999, "rle": {"size": [375, 500], "counts": "ea1k0n9n0A?O10O100000O10O10000000O10O100000O10O100000O10O100000O10O100000O10O100000O1000O1000O1000O1000O1000O1000O1000O1000O100000O10O100000O10O100000O10O100000O10O100000O10O10000000O010000000O010000000O010000000O010000000O10O10000000O01b0^OVhZ4"}, "label": "a newspaper"}]} {"id": 259514, "image": "COCO_train2014_000000259514.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hp computer lid\"."}], "task": "refering", "answer_template": "The \"hp computer lid\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 266], "bbox": [0.31446874999999996, 0.4722045454545455, 0.624296875, 0.7547045454545455], "iscrowd": 0, "area": 20402.4433, "rle": {"size": [440, 640], "counts": "_ef22c=5N3M2N3M2N3M2N2M1M3C=D in this image.", "objects": [{"patches": [168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 243, 260, 261, 262, 263, 264, 265, 266], "bbox": [0.31446874999999996, 0.4722045454545455, 0.624296875, 0.7547045454545455], "iscrowd": 0, "area": 20402.4433, "rle": {"size": [440, 640], "counts": "_ef22c=5N3M2N3M2N3M2N2M1M3C=D in this image.", "objects": [{"patches": [165, 187, 188, 189, 210, 211, 212, 234, 235, 258, 259, 282, 283, 284, 285, 286, 287, 306, 307, 308, 309, 310, 330, 331, 332, 333], "bbox": [0.152953125, 0.4674090909090909, 0.50828125, 0.9190909090909091], "iscrowd": 0, "area": 10548.382950000003, "rle": {"size": [440, 640], "counts": "iZZ12e=3M3M4L3N2M3M4L3M3N3L3M3M3M4L3N2M3M4L3M3M3N3L3M3M3M4M2M3M2N2N2N1011O1O2M3N1O2M2O2N2N1N3N1O2N2M2O2N1O2M3N1O2M3N1O2N1N3N2N1O2M2O2N2M2O2N1O2M3N1O2N1N3N2N1N3N1O2N2M2O2N00O001O100O1O1O001O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O001O0001O000000000000000001O0001O000000000000000001O0001O000000000000000001O0001O000000000000000001O0001O000000000000000001O0001O0000000000005K4L5L4K4L5K4L5K4L5K5K4LhnV4"}, "label": "a black laptop with a blue screen"}]} {"id": 71099, "image": "COCO_train2014_000000071099.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large microwave with a container in it , it is sitting next to a samller microwave which is sitting next to a typewriter\"."}], "task": "refering", "answer_template": "The \"a large microwave with a container in it , it is sitting next to a samller microwave which is sitting next to a typewriter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280], "bbox": [0.0016875000000000002, 0.2763958333333333, 0.48034374999999996, 0.7550624999999999], "iscrowd": 0, "area": 57834.81855000001, "rle": {"size": [480, 640], "counts": "ac0d0\\>X1hNX1hNY1hNW1hNT1lN000001O0O10001O000O10001O000O10O100000O100000O1000O100000000O10O100000O100000000O1000O1000O10000000000O10O100000O10000000O0100000000O10O100000O10000000O010000000000O010000000O100000O10O100000000O100000000O100000000O01000000000O100000000O100000000O100000000O100000000O100000001N10000000000O100000000O101O00000O100000000O100000000O10O100000O100000O1000O100000000O10O100000O100000000O1000O1000O100000000O10000000O10O100000000O100000000O100000O10O100000000O10000000000O1000O1000O100000000O10001O1O1N101O1O1O0O2O1O1O001N2O001O1O1O0O2O001O001N2O001O001N101O1O001N101O001O0O2O1O00000O10000000001N10:FR1nNR1nNR1mN`jk4"}, "label": "a large microwave with a container in it , it is sitting next to a samller microwave which is sitting next to a typewriter"}]} {"id": 71099, "image": "COCO_train2014_000000071099.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the microwave on the left\"."}], "task": "refering", "answer_template": "The \"the microwave on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280], "bbox": [0.0016875000000000002, 0.2763958333333333, 0.48034374999999996, 0.7550624999999999], "iscrowd": 0, "area": 57834.81855000001, "rle": {"size": [480, 640], "counts": "ac0d0\\>X1hNX1hNY1hNW1hNT1lN000001O0O10001O000O10001O000O10O100000O100000O1000O100000000O10O100000O100000000O1000O1000O10000000000O10O100000O10000000O0100000000O10O100000O10000000O010000000000O010000000O100000O10O100000000O100000000O100000000O01000000000O100000000O100000000O100000000O100000000O100000001N10000000000O100000000O101O00000O100000000O100000000O10O100000O100000O1000O100000000O10O100000O100000000O1000O1000O100000000O10000000O10O100000000O100000000O100000O10O100000000O10000000000O1000O1000O100000000O10001O1O1N101O1O1O0O2O1O1O001N2O001O1O1O0O2O001O001N2O001O001N101O1O001N101O001O0O2O1O00000O10000000001N10:FR1nNR1nNR1mN`jk4"}, "label": "the microwave on the left"}]} {"id": 71099, "image": "COCO_train2014_000000071099.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"silver plated microwave\"."}], "task": "refering", "answer_template": "The \"silver plated microwave\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 241], "bbox": [0.434828125, 0.27416666666666667, 0.758421875, 0.60225], "iscrowd": 0, "area": 22818.96755, "rle": {"size": [480, 640], "counts": "amR41o>001O000O2O00001O000O101O00001O00001O000O2O00001O001O3M3M2N3M3LS1nNU1kNV1jN;E00O10000000O0100000000O100000000O010000000O10000000000O010000000O100000000O10O100000O100000000O10O100000O100000000O1000O1000O100000000O100000O10O100000000O10000000O0100000000O100000000O010000000O100000000O10001O001O0O101O001O0O101O001O0O101O001O0O101O001O0O101O001O0O101O00001N101O00001N101O00001N101O00001N101O00001N101O000H9C in this image.", "objects": [{"patches": [104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 131, 132, 149, 150, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 241], "bbox": [0.434828125, 0.27416666666666667, 0.758421875, 0.60225], "iscrowd": 0, "area": 22818.96755, "rle": {"size": [480, 640], "counts": "amR41o>001O000O2O00001O000O101O00001O00001O000O2O00001O001O3M3M2N3M3LS1nNU1kNV1jN;E00O10000000O0100000000O100000000O010000000O10000000000O010000000O100000000O10O100000O100000000O10O100000O100000000O1000O1000O100000000O100000O10O100000000O10000000O0100000000O100000000O010000000O100000000O10001O001O0O101O001O0O101O001O0O101O001O0O101O001O0O101O001O0O101O00001N101O00001N101O00001N101O00001N101O00001N101O000H9C in this image.", "objects": [{"patches": [139, 162, 163, 185, 186, 208, 209, 231, 232, 233, 254, 255, 256, 277, 278, 300, 322, 323, 345, 346, 368, 369, 370, 391, 392, 393, 414], "bbox": [0.002453125, 0.3451492537313433, 0.16295312499999998, 1.0], "iscrowd": 0, "area": 12026.9952, "rle": {"size": [536, 640], "counts": "`a17[`07I7J6I6J7J6I7I7J6I7I7J5J7J6J6K5J6K5K5K1O1O001O0N3M2N3M2N3M2N3oJlL^LW3f3n4O0100aNmHWIS7d6SI[In6_6XIaIh6Z6]IeIc6V6cIjI]6P6iIoIX6l5lIUJS6f5SJYJn5a5XJ^Jh5^5\\JcJd5W5bJhJ_5R5gJnJX5n4lJRKU5h4QKXKn4c4XK\\Ki4_4\\K`Kd4[4bKeK^4U4hKiKY4S4lKlKU4m3RLQLQ4c3ZL\\Li3W3aLhLb3m2hLRMZ3g2lLWMW3g2jLXMY3e2hLYM[3e2fLYM]3e2dLXM`3f2aLWMd3g2\\LVMk3e2WLWMQ4c2QLYMW4a2kK[M]4_2T5J6J6J7I5K1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1M3L4M3L4M3M3L4McXh8"}, "label": "skateboard positioned upright in hands of person on top of ramp"}]} {"id": 95676, "image": "COCO_train2014_000000095676.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baseball player with glove\"."}], "task": "refering", "answer_template": "The \"baseball player with glove\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 61, 62, 63, 64, 85, 86, 87, 88, 108, 109, 110, 111, 129, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 178, 179, 180, 181, 201, 202, 203, 204, 223, 224, 225, 226, 246, 247, 248, 249, 269, 270, 271, 272, 293, 294, 295, 296, 316, 317, 318, 319], "bbox": [0.61146875, 0.0739344262295082, 0.897609375, 0.9068852459016393], "iscrowd": 0, "area": 29280.25975, "rle": {"size": [427, 640], "counts": "`VS51Y=1N3N2M2O1N2O1O100O1000001O001O1O1O0O2O000001O1O001O1O1O100O01hLAcI>\\6FbI:\\6JaI7]6KbI5^6LaI5]6NaI3^6NaI2^60bI0]61bI0\\63aIN_63`IN^64`IM`65]IMa65]IMb65\\IKd65[IMd63\\INb63]INc62]INc62\\IOd61\\IOd61\\IOd61[I0d61\\IOd61[I0e60ZI1f6OWI4S1mN[2o0_L7S1nN]2k0XL?Y1jN\\2h0PLi0a1cN^2d0nKl0b1cN_2a0lKo0d1cN]2a0jKo0h1bN]2`0gKP1k1cN\\2`0cKQ1P2bN[2>_KT1U2aNZ2>]KR1Y2bNY2>[KP1[2fNW2<\\Kn0]2hNU2=ZKl0`2jNU2 in this image.", "objects": [{"patches": [38, 39, 40, 61, 62, 63, 64, 85, 86, 87, 88, 108, 109, 110, 111, 129, 130, 131, 132, 133, 134, 135, 153, 154, 155, 156, 157, 158, 178, 179, 180, 181, 201, 202, 203, 204, 223, 224, 225, 226, 246, 247, 248, 249, 269, 270, 271, 272, 293, 294, 295, 296, 316, 317, 318, 319], "bbox": [0.61146875, 0.0739344262295082, 0.897609375, 0.9068852459016393], "iscrowd": 0, "area": 29280.25975, "rle": {"size": [427, 640], "counts": "`VS51Y=1N3N2M2O1N2O1O100O1000001O001O1O1O0O2O000001O1O001O1O1O100O01hLAcI>\\6FbI:\\6JaI7]6KbI5^6LaI5]6NaI3^6NaI2^60bI0]61bI0\\63aIN_63`IN^64`IM`65]IMa65]IMb65\\IKd65[IMd63\\INb63]INc62]INc62\\IOd61\\IOd61\\IOd61[I0d61\\IOd61[I0e60ZI1f6OWI4S1mN[2o0_L7S1nN]2k0XL?Y1jN\\2h0PLi0a1cN^2d0nKl0b1cN_2a0lKo0d1cN]2a0jKo0h1bN]2`0gKP1k1cN\\2`0cKQ1P2bN[2>_KT1U2aNZ2>]KR1Y2bNY2>[KP1[2fNW2<\\Kn0]2hNU2=ZKl0`2jNU2 in this image.", "objects": [{"patches": [88, 89, 90, 111, 112, 113, 134, 135, 136, 157, 158, 159, 160, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 362, 363, 364, 365, 366, 367, 385, 386, 387, 388, 389, 390], "bbox": [0.5784687500000001, 0.20558333333333334, 0.9921562500000001, 0.9791458333333334], "iscrowd": 0, "area": 50608.03015, "rle": {"size": [480, 640], "counts": "Ye]59h>9G9H7HO1O2N00001O00001OM3G9J600001O000010O01O01O01O00010O0010O0001O010O1O100O1O100O1O1O10O000010O000001O01O000001O01O00000001O01O00000001O0001O000001O000000001O000001O0001O000000001O00000kCPOY:Q1eETOX:m0dEXOZ:i0cE\\OZ:d0cEAZ:a0cEDZ:=bEH\\:8bEL\\:5`E0Z2bNn4`1eH2Z2_NQ5_1cHa0o1PN^5`1_Hk0g1iMi5^1]HR1`1dMR6Z1[H[1Y1_M\\6W1XHb1R1[Mf6T1THj1m0UMn6R1SHa3m7i101O0O1000000O10000O1000000O1000000O100O100O100O100O100O10000O100O100O100O100O100O1O1O1O1O1]M\\HiMe7T2hHaMY7[2UIZMl6b2[IZMf6c2^I[Mc6a2cI\\M^6b2eI[M]6e2eIXM\\6g2fIWM[6h2hIUMY6j2jISMW6l2kIRMV6m2mIPMT6o2mInLV6Q3lIfL\\6Y3f200O1O100O1O100O1O1O100O1O100O0100000O1000000000000O1000000000000O1000001O000001O1O2N1O1O2O0O1O1O2N1O1O1O2N1O1O4L4L3L5L4L3M4L6J8H8H8H6J2N2N2N2N2N2N2N3N3L3M3M4L3M3M3M4L3M3M3M4nNQ1kNU1lNU1jNU1lNT1kNTT2"}, "label": "a woman in a bink sweater covering a white dress with pink flowers"}]} {"id": 136651, "image": "COCO_train2014_000000136651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing a backpack taking a picture on her phone\"."}], "task": "refering", "answer_template": "The \"a woman wearing a backpack taking a picture on her phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 111, 112, 113, 134, 135, 136, 157, 158, 159, 160, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 339, 340, 341, 342, 343, 344, 362, 363, 364, 365, 366, 367, 385, 386, 387, 388, 389, 390], "bbox": [0.5784687500000001, 0.20558333333333334, 0.9921562500000001, 0.9791458333333334], "iscrowd": 0, "area": 50608.03015, "rle": {"size": [480, 640], "counts": "Ye]59h>9G9H7HO1O2N00001O00001OM3G9J600001O000010O01O01O01O00010O0010O0001O010O1O100O1O100O1O1O10O000010O000001O01O000001O01O00000001O01O00000001O0001O000001O000000001O000001O0001O000000001O00000kCPOY:Q1eETOX:m0dEXOZ:i0cE\\OZ:d0cEAZ:a0cEDZ:=bEH\\:8bEL\\:5`E0Z2bNn4`1eH2Z2_NQ5_1cHa0o1PN^5`1_Hk0g1iMi5^1]HR1`1dMR6Z1[H[1Y1_M\\6W1XHb1R1[Mf6T1THj1m0UMn6R1SHa3m7i101O0O1000000O10000O1000000O1000000O100O100O100O100O100O10000O100O100O100O100O100O1O1O1O1O1]M\\HiMe7T2hHaMY7[2UIZMl6b2[IZMf6c2^I[Mc6a2cI\\M^6b2eI[M]6e2eIXM\\6g2fIWM[6h2hIUMY6j2jISMW6l2kIRMV6m2mIPMT6o2mInLV6Q3lIfL\\6Y3f200O1O100O1O100O1O1O100O1O100O0100000O1000000000000O1000000000000O1000001O000001O1O2N1O1O2O0O1O1O2N1O1O1O2N1O1O4L4L3L5L4L3M4L6J8H8H8H6J2N2N2N2N2N2N2N3N3L3M3M4L3M3M3M4L3M3M3M4nNQ1kNU1lNU1jNU1lNT1kNTT2"}, "label": "a woman wearing a backpack taking a picture on her phone"}]} {"id": 136651, "image": "COCO_train2014_000000136651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person wearing a yellow headband\"."}], "task": "refering", "answer_template": "The \"the person wearing a yellow headband\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [268, 269, 288, 289, 290, 291, 292, 311, 312, 313, 314, 315, 336, 337, 338, 359, 360, 361, 382, 383, 384], "bbox": [0.545609375, 0.6790625, 0.7364843750000001, 0.989875], "iscrowd": 0, "area": 11849.069699999998, "rle": {"size": [480, 640], "counts": "imS52n>7I6J6J1O1O1O1O1O1O1O1O1O001O10000O10000O10000O10000O10000O10000O10000O10000O2O0O100]DgN^9Y1YFoNh9Q1oEWOR:o0_EYOc:l0nD\\OS;j0^D^Oc;o12N1O1O[OoL]Eo2a:VM]Ei2b:[M\\Ed2b:bM\\E\\2b:iM\\EV2c:T1N2O1N2N2O100O1O100O1O100O1O1O100O1O2O0O1O100001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O1O1O1O1O1O1O001O1O1O1O1oNWm^2"}, "label": "the person wearing a yellow headband"}]} {"id": 136651, "image": "COCO_train2014_000000136651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman , kneeling , with a yellow headband looking at a display of a cow\"."}], "task": "refering", "answer_template": "The \"a woman , kneeling , with a yellow headband looking at a display of a cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [268, 269, 288, 289, 290, 291, 292, 311, 312, 313, 314, 315, 336, 337, 338, 359, 360, 361, 382, 383, 384], "bbox": [0.545609375, 0.6790625, 0.7364843750000001, 0.989875], "iscrowd": 0, "area": 11849.069699999998, "rle": {"size": [480, 640], "counts": "imS52n>7I6J6J1O1O1O1O1O1O1O1O1O001O10000O10000O10000O10000O10000O10000O10000O10000O2O0O100]DgN^9Y1YFoNh9Q1oEWOR:o0_EYOc:l0nD\\OS;j0^D^Oc;o12N1O1O[OoL]Eo2a:VM]Ei2b:[M\\Ed2b:bM\\E\\2b:iM\\EV2c:T1N2O1N2N2O100O1O100O1O100O1O1O100O1O2O0O1O100001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O1O1O1O1O1O1O001O1O1O1O1oNWm^2"}, "label": "a woman , kneeling , with a yellow headband looking at a display of a cow"}]} {"id": 136651, "image": "COCO_train2014_000000136651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large animal with long ears in an enclosure that people are looking at\"."}], "task": "refering", "answer_template": "The \"a large animal with long ears in an enclosure that people are looking at\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 368, 369, 370, 371, 372, 373], "bbox": [0.0, 0.5352708333333334, 0.471015625, 0.9997916666666666], "iscrowd": 0, "area": 43636.110349999995, "rle": {"size": [480, 640], "counts": "f9Z5f900000000000001O00000000001O00000O1000001O00000000001O0000000000001O00000000001O0000000000001O00000000001N1000000000001O00000000001O000000O10000O100O10000O100O10000O100O10000O010O10000O100O10000O100O10000O100O10000O100O10000O010O01000O010O0100O010O010O10O010O10OO2O0O2O0O2O0O2O0O2O1N101N101N101N101N101O00100O010O010O0010O010O010O10001N100O100O2O0O100O101O0O100O101N100O100O101N10000O101N100O100O2O0O10000O2O0O100O101N100O1M3M4K4M3L4M4K5L5L3O1N3N1O1N3N1N2O2N1N2O2M2O1O1N3N1N2O2N1N2O2M2O1N3N1O1N3N100O2N1O101N1O1O2O0O1O2N100O1O1O100O1O100O1O1O100O1O1O1N2N2N2N2N1O2N2N2N2N2M3N3M2N3M2N3M2N2N3M2N3M2M4M2N`an4"}, "label": "a large animal with long ears in an enclosure that people are looking at"}]} {"id": 136651, "image": "COCO_train2014_000000136651.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white animal fenced in\"."}], "task": "refering", "answer_template": "The \"white animal fenced in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 368, 369, 370, 371, 372, 373], "bbox": [0.0, 0.5352708333333334, 0.471015625, 0.9997916666666666], "iscrowd": 0, "area": 43636.110349999995, "rle": {"size": [480, 640], "counts": "f9Z5f900000000000001O00000000001O00000O1000001O00000000001O0000000000001O00000000001O0000000000001O00000000001N1000000000001O00000000001O000000O10000O100O10000O100O10000O100O10000O010O10000O100O10000O100O10000O100O10000O100O10000O010O01000O010O0100O010O010O10O010O10OO2O0O2O0O2O0O2O0O2O1N101N101N101N101N101O00100O010O010O0010O010O010O10001N100O100O2O0O100O101O0O100O101N100O100O101N10000O101N100O100O2O0O10000O2O0O100O101N100O1M3M4K4M3L4M4K5L5L3O1N3N1O1N3N1N2O2N1N2O2M2O1O1N3N1N2O2N1N2O2M2O1N3N1O1N3N100O2N1O101N1O1O2O0O1O2N100O1O1O100O1O100O1O1O100O1O1O1N2N2N2N2N1O2N2N2N2N2M3N3M2N3M2N3M2N2N3M2N3M2M4M2N`an4"}, "label": "white animal fenced in"}]} {"id": 275917, "image": "COCO_train2014_000000275917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing plaid shorts\"."}], "task": "refering", "answer_template": "The \"a man wearing plaid shorts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 93, 94, 95, 116, 117, 118, 139, 140, 141, 142, 163, 164, 165, 186, 187, 188, 210, 211, 212, 232, 233, 234, 235], "bbox": [0.044375, 0.18280602636534837, 0.264625, 0.5778907721280603], "iscrowd": 0, "area": 7773.027349999999, "rle": {"size": [531, 640], "counts": "Pf>1]`06O001N101N2YNHQC8nhBBW=`0iB@W=a0hB_OW=c0hB^OW=c0hB]OW=e0hB[OX=g0fBYOZ=g0fBYOY=i0fBWOZ=i0eBYO[=g0dBYO\\=g0dBYO]=g0bBYO^=g0bBYO_=f0aBZOX3FR6Q1eFZOW3FU6o0dF[OU3GX6o0nEWO44e3GZ6m0iE6i3mN`6W2]IjMd6W2YIjMh6W2UIjMl6W2QIjMP7W2mHjMT7X2hHiMY7X2eHhM\\7Y2aHhM`7Z2\\HgMe7Z2fEZMY25@m:P3QEaM<^OAd02^OQ;_2XE_MHb0?4i:j1WERN23f:k1XERNI^OKd0U;k1WESNI@Kd0iA_OX>^11N2O1O1N2O1O1Nj0WO6JiAQOoeBB\\=>bBC^=>aBB`=>_BBa=?]BBd=>[BBf=>YBBg=?WBBj=>UBBl=>RBCn=>QBBP>>oABR>>lACT>>jACW>=gADZ>=]AZOF:n>>XAGj>9SAHo>l03N3M2N3M2M2N2N2N2N2N3M2N2N2N2N2N2N2N2N2Nhnc7"}, "label": "a man wearing plaid shorts"}]} {"id": 275917, "image": "COCO_train2014_000000275917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in plaid shorts and a white tshirt to the left of the photo walking away\"."}], "task": "refering", "answer_template": "The \"a man in plaid shorts and a white tshirt to the left of the photo walking away\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 93, 94, 95, 116, 117, 118, 139, 140, 141, 142, 163, 164, 165, 186, 187, 188, 210, 211, 212, 232, 233, 234, 235], "bbox": [0.044375, 0.18280602636534837, 0.264625, 0.5778907721280603], "iscrowd": 0, "area": 7773.027349999999, "rle": {"size": [531, 640], "counts": "Pf>1]`06O001N101N2YNHQC8nhBBW=`0iB@W=a0hB_OW=c0hB^OW=c0hB]OW=e0hB[OX=g0fBYOZ=g0fBYOY=i0fBWOZ=i0eBYO[=g0dBYO\\=g0dBYO]=g0bBYO^=g0bBYO_=f0aBZOX3FR6Q1eFZOW3FU6o0dF[OU3GX6o0nEWO44e3GZ6m0iE6i3mN`6W2]IjMd6W2YIjMh6W2UIjMl6W2QIjMP7W2mHjMT7X2hHiMY7X2eHhM\\7Y2aHhM`7Z2\\HgMe7Z2fEZMY25@m:P3QEaM<^OAd02^OQ;_2XE_MHb0?4i:j1WERN23f:k1XERNI^OKd0U;k1WESNI@Kd0iA_OX>^11N2O1O1N2O1O1Nj0WO6JiAQOoeBB\\=>bBC^=>aBB`=>_BBa=?]BBd=>[BBf=>YBBg=?WBBj=>UBBl=>RBCn=>QBBP>>oABR>>lACT>>jACW>=gADZ>=]AZOF:n>>XAGj>9SAHo>l03N3M2N3M2M2N2N2N2N2N3M2N2N2N2N2N2N2N2N2Nhnc7"}, "label": "a man in plaid shorts and a white tshirt to the left of the photo walking away"}]} {"id": 275917, "image": "COCO_train2014_000000275917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboard laying upside down on the ground in front of the blue backpack\"."}], "task": "refering", "answer_template": "The \"a skateboard laying upside down on the ground in front of the blue backpack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [326, 327, 328, 347, 348, 349, 350, 351, 368, 369, 370, 371, 372, 373, 391, 392, 393, 394, 414, 415], "bbox": [0.0037656250000000003, 0.7414877589453861, 0.30478125, 0.9864030131826741], "iscrowd": 0, "area": 8975.456449999994, "rle": {"size": [531, 640], "counts": "l_1`1R?2O001O1N101O1O1N0100O100O100O100O100O100O1O100O10O010O0010O010O010O010O010O0010O010O010O0010O010O010O010O010O0010O010O010O0010O010O010O010O010O0010O010O010O0010O010O010O010O010O0010O010O010O010O0010O010O010O010O0010O010O010O010O0010O010O01@bNkA]1T>fNjAZ1U>iNjAW1T>lNjAT1W>mNhAS1X>oNeAQ1\\>POcAP1]>QOaAo0`>b00O100O2O3L3N0O2O001N100O2O0O2O0O2O0O101N101N01O001N101N1100O101O0O101N100O2O0O101N2O0O2O1N101N100O1O1O2N100O1O1O1O1O2O3L5K4L4LR[V7"}, "label": "a skateboard laying upside down on the ground in front of the blue backpack"}]} {"id": 275917, "image": "COCO_train2014_000000275917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skateboard laying next to a blue backpack\"."}], "task": "refering", "answer_template": "The \"a skateboard laying next to a blue backpack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [326, 327, 328, 347, 348, 349, 350, 351, 368, 369, 370, 371, 372, 373, 391, 392, 393, 394, 414, 415], "bbox": [0.0037656250000000003, 0.7414877589453861, 0.30478125, 0.9864030131826741], "iscrowd": 0, "area": 8975.456449999994, "rle": {"size": [531, 640], "counts": "l_1`1R?2O001O1N101O1O1N0100O100O100O100O100O100O1O100O10O010O0010O010O010O010O010O0010O010O010O0010O010O010O010O010O0010O010O010O0010O010O010O010O010O0010O010O010O0010O010O010O010O010O0010O010O010O010O0010O010O010O010O0010O010O010O010O0010O010O01@bNkA]1T>fNjAZ1U>iNjAW1T>lNjAT1W>mNhAS1X>oNeAQ1\\>POcAP1]>QOaAo0`>b00O100O2O3L3N0O2O001N100O2O0O2O0O2O0O101N101N01O001N101N1100O101O0O101N100O2O0O101N2O0O2O1N101N100O1O1O2N100O1O1O1O1O2O3L5K4L4LR[V7"}, "label": "a skateboard laying next to a blue backpack"}]} {"id": 275917, "image": "COCO_train2014_000000275917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skateboard being held by the man in the white shirt\"."}], "task": "refering", "answer_template": "The \"the skateboard being held by the man in the white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [289, 290, 291, 311, 312, 313, 314, 333, 334, 335, 336, 354, 355, 356, 357, 358, 359, 376, 377, 378, 379, 380, 381, 382, 398, 399, 400, 401, 402, 403, 404, 421, 422, 423, 424, 425, 426], "bbox": [0.299296875, 0.6331073446327684, 0.6932343750000001, 0.9885499058380416], "iscrowd": 0, "area": 20549.257199999996, "rle": {"size": [531, 640], "counts": "YRT31``02O1N2O1N2O1N2O1N2O1O1N2O1N3N1N2O1N2O1N2O1N2O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O2N1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O2N1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O01O00000000001O000000001O0000000O101O000000001O00000000001O000000001O000000000O2O000000001O00000000001O000000001O000000001N10000000001O000000001O001O1O1O1`NlB@U=?UCWOl in this image.", "objects": [{"patches": [289, 290, 291, 311, 312, 313, 314, 333, 334, 335, 336, 354, 355, 356, 357, 358, 359, 376, 377, 378, 379, 380, 381, 382, 398, 399, 400, 401, 402, 403, 404, 421, 422, 423, 424, 425, 426], "bbox": [0.299296875, 0.6331073446327684, 0.6932343750000001, 0.9885499058380416], "iscrowd": 0, "area": 20549.257199999996, "rle": {"size": [531, 640], "counts": "YRT31``02O1N2O1N2O1N2O1N2O1O1N2O1N3N1N2O1N2O1N2O1N2O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O2N1O1O100O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O2N1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O01O00000000001O000000001O0000000O101O000000001O00000000001O000000001O000000000O2O000000001O00000000001O000000001O000000001N10000000001O000000001O001O1O1O1`NlB@U=?UCWOl in this image.", "objects": [{"patches": [128, 129, 130, 151, 152, 153, 154, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 248, 249, 250, 265, 266, 267, 268, 269, 270, 271, 272, 273, 289, 290, 291, 292, 293, 294, 295, 296, 297, 314, 315, 316, 317, 318, 319, 320, 336, 337, 338, 339, 340, 341, 359, 360, 361, 362, 363, 381, 382, 383, 384, 385, 386, 403, 404, 405, 406, 407, 408, 426, 427, 428, 429, 430], "bbox": [0.5525, 0.2850094161958569, 0.9264218750000002, 0.9868173258003766], "iscrowd": 0, "area": 51477.913499999995, "rle": {"size": [531, 640], "counts": "Zlg53]`05L3dEIZ48^KOa43VK5j4LnJ;R5FeJb0a1ROYM?Q1b0c1UOYM:R1b0a1[OZM4S1b0`1AXMOV1a0_1FXMJW1a0^1KXMEX1a0]10XMAX1`0]15WM\\O[1`0[18YMXOZ1b0Z1;ZMSO[1c0Y1>YMPO]1c0Y1?YMnN]1d0Z1`0WMlN^1e0[1a0TMkN`1e0[1c0nLmNf1a0\\1c0bLWOQ27\\1e0VL_O]2M]1f0PLAb2J^1f0kKDf2G^1h0fKEk2D_1i0`KGo2Ba1i0YKJU3^Oa1j0UKLY3[Ob1k0oJN^3XOc1l0iJOd3VOb1f3^N[Lb1e3]N\\Lc1c3]N^Lb1b3^N_Lb1`3^NaLb1^3^NcLa1^3]NdLc1\\3\\NeLd1[3[NfLd1[3[NfLe1Z3ZNgLf1Y3YNhLf1Y3YNhLg1X3XNiLh1W3SM\\JlN^2P4W3RM^JlN\\2R4V3PMaJlNZ2T4U3nLeJkNW2V4U3mLgJjNV2Y4S3kLjJjNT2[4R3jLlJiNS2\\4R3iLnJiNQ2^4Q3hLPKhNP2`4P3gLRKgNo1a4P3gLTKeNm1d4P3eLUKeNl1f4o2eLVKcNl1h4o2cLWKcNk1j4o2aLZN`3f1^L[Nb3f1\\L[Ne3e1YL\\Ng3e1WL\\Nj3c1VL^Ni3c1UL^Nl3b1SL^Nm3c1QL^NP4b1oK^NQ4c1nK]NS4c1kK^NU4c1jK]NW4c1gK^NY4j1_KVNb4Q2VKoMj4Y2mJhMT5k6001O001O2N2N3M2N3M2N3M2N2N00001O001O00001O001O001O00001O001VFkIT9U6kFlIU9T6iFnIW9S6gFoIX9Q6gFPJW9S6fFoIX9S6gFnIW9T6gFnIX9T6cFQJZ9e6N2N2N3N1N2N2N3N1N2N2N3M2O1N2N3M2N2O1N3M2N2O1N3M2N2N2O2M2N2N2N3N1N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N3M2N2N201N100O101N100O1O2O0O100O2O0O100O2N100O103L4M3L3N3L4L4M3L4M2M4M3L4M3L4L4M2M5L3L4M3L4M3L5K4M3L4M3L4M4K4M3L4L4M3L4M4K4M3L4M3L4L5L3L4M[Ph0"}, "label": "the man in the green hat and white shirt using his phone and holding a skateboard"}]} {"id": 275917, "image": "COCO_train2014_000000275917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a green hat , white tshirt and jeans as he holds onto his skateboard by the wheels and a cell phone in the other hand\"."}], "task": "refering", "answer_template": "The \"a man wearing a green hat , white tshirt and jeans as he holds onto his skateboard by the wheels and a cell phone in the other hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 151, 152, 153, 154, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 224, 225, 226, 227, 243, 244, 245, 246, 247, 248, 249, 250, 265, 266, 267, 268, 269, 270, 271, 272, 273, 289, 290, 291, 292, 293, 294, 295, 296, 297, 314, 315, 316, 317, 318, 319, 320, 336, 337, 338, 339, 340, 341, 359, 360, 361, 362, 363, 381, 382, 383, 384, 385, 386, 403, 404, 405, 406, 407, 408, 426, 427, 428, 429, 430], "bbox": [0.5525, 0.2850094161958569, 0.9264218750000002, 0.9868173258003766], "iscrowd": 0, "area": 51477.913499999995, "rle": {"size": [531, 640], "counts": "Zlg53]`05L3dEIZ48^KOa43VK5j4LnJ;R5FeJb0a1ROYM?Q1b0c1UOYM:R1b0a1[OZM4S1b0`1AXMOV1a0_1FXMJW1a0^1KXMEX1a0]10XMAX1`0]15WM\\O[1`0[18YMXOZ1b0Z1;ZMSO[1c0Y1>YMPO]1c0Y1?YMnN]1d0Z1`0WMlN^1e0[1a0TMkN`1e0[1c0nLmNf1a0\\1c0bLWOQ27\\1e0VL_O]2M]1f0PLAb2J^1f0kKDf2G^1h0fKEk2D_1i0`KGo2Ba1i0YKJU3^Oa1j0UKLY3[Ob1k0oJN^3XOc1l0iJOd3VOb1f3^N[Lb1e3]N\\Lc1c3]N^Lb1b3^N_Lb1`3^NaLb1^3^NcLa1^3]NdLc1\\3\\NeLd1[3[NfLd1[3[NfLe1Z3ZNgLf1Y3YNhLf1Y3YNhLg1X3XNiLh1W3SM\\JlN^2P4W3RM^JlN\\2R4V3PMaJlNZ2T4U3nLeJkNW2V4U3mLgJjNV2Y4S3kLjJjNT2[4R3jLlJiNS2\\4R3iLnJiNQ2^4Q3hLPKhNP2`4P3gLRKgNo1a4P3gLTKeNm1d4P3eLUKeNl1f4o2eLVKcNl1h4o2cLWKcNk1j4o2aLZN`3f1^L[Nb3f1\\L[Ne3e1YL\\Ng3e1WL\\Nj3c1VL^Ni3c1UL^Nl3b1SL^Nm3c1QL^NP4b1oK^NQ4c1nK]NS4c1kK^NU4c1jK]NW4c1gK^NY4j1_KVNb4Q2VKoMj4Y2mJhMT5k6001O001O2N2N3M2N3M2N3M2N2N00001O001O00001O001O001O00001O001VFkIT9U6kFlIU9T6iFnIW9S6gFoIX9Q6gFPJW9S6fFoIX9S6gFnIW9T6gFnIX9T6cFQJZ9e6N2N2N3N1N2N2N3N1N2N2N3M2O1N2N3M2N2O1N3M2N2O1N3M2N2N2O2M2N2N2N3N1N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N2N3M2N2N3M2N2N201N100O101N100O1O2O0O100O2O0O100O2N100O103L4M3L3N3L4L4M3L4M2M4M3L4M3L4L4M2M5L3L4M3L4M3L5K4M3L4M3L4M4K4M3L4L4M3L4M4K4M3L4M3L4L5L3L4M[Ph0"}, "label": "a man wearing a green hat , white tshirt and jeans as he holds onto his skateboard by the wheels and a cell phone in the other hand"}]} {"id": 275917, "image": "COCO_train2014_000000275917.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy with glasses wear a dc hat\"."}], "task": "refering", "answer_template": "The \"a guy with glasses wear a dc hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 125, 126, 127, 128, 148, 149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 197, 217, 218, 219, 220, 240, 241, 242, 243, 263, 264, 265, 287, 288, 289, 310], "bbox": [0.435640625, 0.21001883239171373, 0.60503125, 0.6922410546139359], "iscrowd": 0, "area": 15699.764050000007, "rle": {"size": [531, 640], "counts": "ok`4:W`05L5K5K4[CVOY9k0eF^OT9c0gFCW9>bFI]98iDG26S;4aD81G];2XDh01YOm9nNaGb4\\8iK[GX4a8oK[Gg3gNoKk9?[Ge3mNgKf9i0YGd3SO_Kb9Q1XGb3l8`LRG`3n8cLoF^2VOZMj9:nFY2CVM_9b0lFU20RMS9l0jFo1?mLg8V1gFj1h:YNUEc1P;^NoD^1U;bNjD=2_NX;V1dDOc0eNm:]1^DJl0bNk:d1WDMn<4PC0m<1QC5OVNfVMk9;lE]2:WMj9:QF[27YMi9:WFX21]Mh99^FU2KaMg98cFT2g:jM\\EU2e:hM_EW2b:eMbEX2`:dMdE[2[ in this image.", "objects": [{"patches": [102, 103, 125, 126, 127, 128, 148, 149, 150, 151, 171, 172, 173, 174, 194, 195, 196, 197, 217, 218, 219, 220, 240, 241, 242, 243, 263, 264, 265, 287, 288, 289, 310], "bbox": [0.435640625, 0.21001883239171373, 0.60503125, 0.6922410546139359], "iscrowd": 0, "area": 15699.764050000007, "rle": {"size": [531, 640], "counts": "ok`4:W`05L5K5K4[CVOY9k0eF^OT9c0gFCW9>bFI]98iDG26S;4aD81G];2XDh01YOm9nNaGb4\\8iK[GX4a8oK[Gg3gNoKk9?[Ge3mNgKf9i0YGd3SO_Kb9Q1XGb3l8`LRG`3n8cLoF^2VOZMj9:nFY2CVM_9b0lFU20RMS9l0jFo1?mLg8V1gFj1h:YNUEc1P;^NoD^1U;bNjD=2_NX;V1dDOc0eNm:]1^DJl0bNk:d1WDMn<4PC0m<1QC5OVNfVMk9;lE]2:WMj9:QF[27YMi9:WFX21]Mh99^FU2KaMg98cFT2g:jM\\EU2e:hM_EW2b:eMbEX2`:dMdE[2[ in this image.", "objects": [{"patches": [407, 408, 409, 410, 411, 412, 413, 414, 415, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483], "bbox": [0.36098039215686273, 0.8340849673202614, 1.0, 1.0], "iscrowd": 0, "area": 33456.890250000004, "rle": {"size": [612, 612], "counts": "eeT43^b0c0O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O10000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000001O00000000001O001O001O001O001O001O001O001O001O00001O000000001O000000001O00000000001O000000001O00000000001O00001O00001O001O001O001O00001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O"}, "label": "a plate with four biscuits , sitting under two other plates"}]} {"id": 5587, "image": "COCO_train2014_000000005587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the muffins on the bottom of the photo\"."}], "task": "refering", "answer_template": "The \"the muffins on the bottom of the photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [407, 408, 409, 410, 411, 412, 413, 414, 415, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483], "bbox": [0.36098039215686273, 0.8340849673202614, 1.0, 1.0], "iscrowd": 0, "area": 33456.890250000004, "rle": {"size": [612, 612], "counts": "eeT43^b0c0O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O10000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000001O00000000001O001O001O001O001O001O001O001O001O00001O000000001O000000001O00000000001O000000001O00000000001O00001O00001O001O001O001O00001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O001O"}, "label": "the muffins on the bottom of the photo"}]} {"id": 5587, "image": "COCO_train2014_000000005587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the square table the people are sitting at\"."}], "task": "refering", "answer_template": "The \"the square table the people are sitting at\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 248, 249, 270, 271, 292, 293, 314, 315], "bbox": [0.14459150326797385, 0.42818627450980395, 0.3481535947712418, 0.6791339869281046], "iscrowd": 0, "area": 5496.9145499999995, "rle": {"size": [612, 612], "counts": "i^e15jb06N100O10000O100O1000001O000000000O1000000000001O00000000000000000000001O000000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001N10002N3M2N3M3L3N3L3N3L4UA[NQ<]4N101O1O1O1O10O01O1O000000O1N3M2N2N2M3O1eLbAn1`>QNoA_1S>_N^BQ1d=mNlBb0V=]OZC3g in this image.", "objects": [{"patches": [145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238], "bbox": [0.3009967320261438, 0.27034313725490194, 1.0, 0.4854411764705882], "iscrowd": 0, "area": 35047.751849999986, "rle": {"size": [612, 612], "counts": "[V^3d0`b000O10O10000O10000O10000O10000O0100000O101O0O2O001N101O0O2O001O0O2O001N101O0O2O001N101O001O001O001O001O001N101O001O001O001O001O001O001N101O001O001O001O001O001O0O2O001O001O001O001O001O001N101O001O001O0O2O001O0O2O001O0O10000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O10000O1000000O1000000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000M3N2M3M3N2M3M3N2M31O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O00001O0000001O00001O00001O0000001N10001O0O10001O0O101O00001N1000001N10001O0O10001O000O2O00001N1000001O0O101O000O101O000O101O00000O1000001N1000000O10001O000O1000000O101O000O100000001N1000000O10001O000O1000000O2O00000O1000001O0O1000000O101O000O100000001N1000000O10001O0O100000000O2O000O1O1O1O1N2O2N1O1O1O1O1N2O2N1O1O1O1N2O1O2N1O:F>AiH"}, "label": "top plate holding sandwiches"}]} {"id": 5587, "image": "COCO_train2014_000000005587.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the topmost plate of pastries on the pastry rack\"."}], "task": "refering", "answer_template": "The \"the topmost plate of pastries on the pastry rack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238], "bbox": [0.3009967320261438, 0.27034313725490194, 1.0, 0.4854411764705882], "iscrowd": 0, "area": 35047.751849999986, "rle": {"size": [612, 612], "counts": "[V^3d0`b000O10O10000O10000O10000O10000O0100000O101O0O2O001N101O0O2O001O0O2O001N101O0O2O001N101O001O001O001O001O001N101O001O001O001O001O001O001N101O001O001O001O001O001O0O2O001O001O001O001O001O001N101O001O001O0O2O001O0O2O001O0O10000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O10000O1000000O1000000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000M3N2M3M3N2M3M3N2M31O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O0000001O00001O0000001O00001O00001O0000001N10001O0O10001O0O101O00001N1000001N10001O0O10001O000O2O00001N1000001O0O101O000O101O000O101O00000O1000001N1000000O10001O000O1000000O101O000O100000001N1000000O10001O000O1000000O2O00000O1000001O0O1000000O101O000O100000001N1000000O10001O0O100000000O2O000O1O1O1O1N2O2N1O1O1O1O1N2O2N1O1O1O1N2O1O2N1O:F>AiH"}, "label": "the topmost plate of pastries on the pastry rack"}]} {"id": 439765, "image": "COCO_train2014_000000439765.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man on left\"."}], "task": "refering", "answer_template": "The \"man on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 29, 48, 49, 50, 51, 52, 53, 71, 72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 99, 100, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 307, 308, 309, 310, 322, 323, 324, 325, 326, 327, 330, 331, 332], "bbox": [0.0, 0.07648711943793911, 0.49365625, 0.9868149882903982], "iscrowd": 0, "area": 81567.67125, "rle": {"size": [427, 640], "counts": "_6d1h;Y3eLg1YN1O1O1N200O1O100O100O1K5O100000000O100000000O100000000O100000000O1000000O100000000O1000000O2O0000000O1000000iMeH]N\\7X1UIbNk6T1dIgN\\6W1lIcNT6[1QJaNP6]1VJ_Nj5_1[J^Ne5`1aJ[N`5c1fJYNZ5e1jJXNW5f1lJYNT5e1oJZNQ5d1RKZNo4d1TK[Nl4c1WK[Nj4c1YK\\Ng4c1[K[Nf4d1\\K[Nd4d1_KYNb4f1`KYN`4f1bKXN_4g1cKWN^4i1b3O1O1O1O100O1O1O1O100O100O100O100O100O100O2O0O100O100O100O100O100O100O100O1000000O10000O10000O10000O10000O1000000000000000000001O00000L4K5K5J6K6J5M3O1O1O100O1O1O2O0O1O2N2N1O2N1O2N1N3N1O2N2N1O2M2O2N2N1O2N2M2O2N2N1O2N2M3N2N2N2N2N2M3N2N2N2N2N2M3N2N2N2N4L3L4M4L3M4M2O1N3N1O2N2M201O1O001O1O00001\\IjIo5V6nIRJm5n5PJZJk5g5QJeJe5\\5XJSK\\5m4aJZK[5[610000O1eIeJd4[5WKhJk4X5PKjJR5W5iJlJY5T5bJnJ`5R5QJ\\KQ6d4fIcK\\6e51O2N1O2N2N1O2N2N1O2N1O2N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N3M2N3M2N3M2N3M2N3M2N3M2N2N3K4L5L3L5K4M4L3L5L3M4K4M4L3L5L3M5J:G9G9F\\gV4"}, "label": "man on left"}]} {"id": 439765, "image": "COCO_train2014_000000439765.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man leaning his head on his elbow\"."}], "task": "refering", "answer_template": "The \"a man leaning his head on his elbow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [26, 27, 28, 29, 48, 49, 50, 51, 52, 53, 71, 72, 73, 74, 75, 76, 94, 95, 96, 97, 98, 99, 100, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 284, 285, 286, 287, 299, 300, 301, 302, 303, 304, 307, 308, 309, 310, 322, 323, 324, 325, 326, 327, 330, 331, 332], "bbox": [0.0, 0.07648711943793911, 0.49365625, 0.9868149882903982], "iscrowd": 0, "area": 81567.67125, "rle": {"size": [427, 640], "counts": "_6d1h;Y3eLg1YN1O1O1N200O1O100O100O1K5O100000000O100000000O100000000O100000000O1000000O100000000O1000000O2O0000000O1000000iMeH]N\\7X1UIbNk6T1dIgN\\6W1lIcNT6[1QJaNP6]1VJ_Nj5_1[J^Ne5`1aJ[N`5c1fJYNZ5e1jJXNW5f1lJYNT5e1oJZNQ5d1RKZNo4d1TK[Nl4c1WK[Nj4c1YK\\Ng4c1[K[Nf4d1\\K[Nd4d1_KYNb4f1`KYN`4f1bKXN_4g1cKWN^4i1b3O1O1O1O100O1O1O1O100O100O100O100O100O100O2O0O100O100O100O100O100O100O100O1000000O10000O10000O10000O10000O1000000000000000000001O00000L4K5K5J6K6J5M3O1O1O100O1O1O2O0O1O2N2N1O2N1O2N1N3N1O2N2N1O2M2O2N2N1O2N2M2O2N2N1O2N2M3N2N2N2N2N2M3N2N2N2N2N2M3N2N2N2N4L3L4M4L3M4M2O1N3N1O2N2M201O1O001O1O00001\\IjIo5V6nIRJm5n5PJZJk5g5QJeJe5\\5XJSK\\5m4aJZK[5[610000O1eIeJd4[5WKhJk4X5PKjJR5W5iJlJY5T5bJnJ`5R5QJ\\KQ6d4fIcK\\6e51O2N1O2N2N1O2N2N1O2N1O2N2N2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N3M2N3M2N3M2N3M2N3M2N3M2N2N3K4L5L3L5K4M4L3L5L3M4K4M4L3L5L3M5J:G9G9F\\gV4"}, "label": "a man leaning his head on his elbow"}]} {"id": 439765, "image": "COCO_train2014_000000439765.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smiling man in a blue shirt\"."}], "task": "refering", "answer_template": "The \"a smiling man in a blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 33, 34, 35, 36, 37, 56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 102, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 260, 263, 264, 265, 266, 267, 268, 269, 282, 283, 284, 287, 288, 289, 290, 291, 292, 293, 294, 305, 306, 307, 310, 311, 312, 313, 314, 315, 316, 317, 337, 338, 339, 340], "bbox": [0.269296875, 0.01360655737704918, 0.826046875, 0.9886651053864169], "iscrowd": 0, "area": 63159.44775, "rle": {"size": [427, 640], "counts": "T`X29Q=8G7J1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O12N2N2N1O2N2N2N2N1O2N2N2N2N2N1O2NkJeN\\NY1d1hN]NV1b1lN_NQ1a1QO_Nn0a1SO`Nk0_1WObNg0]1[OeNb0[1_OfN?Y1ChN;W1GjN7V1JkN3U1OlNMU15lNGV1:lN@V1b0kNZOV1h0f52M3M4M2M3M3N3L3M3N2M4L3N2M3M3N3L3M3N2M4L3N2M3M4M2M3N2M3M4M2M3M3N3L3M3N2M4L3N2M4L3N3L3M4M2M4L3N3L3M4M2M4nM_JSLd5o2_KjLd4Q3fKgL\\4U3mKcLW4X3PLcLR4Y3QLfLR4U3PLkLR4Q3PLnLT4m2nKSMT4i2nKWMU4c2nK\\MU4`2nK_MT4]2nKcMT4Y2nKfMV4U2lKkMV4Q2lKoMV4m1PLoMR4m1VLlMn3o1ZLkMh3Q2`LiMb3S2n3N3L3N2N2N1O00000000000O100000000000001O000000000O1000000000000000001O000000000000000O100000000000001O00000000000001O00001O000010O0001O01O01O000010O0001O01O01O000PH`NV4`1`KkN^4U1YKVOe4j0QKBm4`0lJGS5:jJIT5:iJFW5=eJEZ5=cJE]5=_JE`5>\\JDc5>ZJDe5?XJAh5b0TJ@l5a0QJAn5b0nI@Q6f1hH\\NW7l310O1O1O10O01O1O2WLSJ1n5LRJ5o5JQJ7P6FQJS8^OQHb0o7\\OSHd0n7ZOSHf0o7WORHi0P8TOQHl0P8QOSHm0P8POQHP1Q8mNPHS1R8iNPHW1R8eNPH[1R8aNPH_1R8]NPHc1Q8ZNQHf1Q8VNRHi1P8SNRHm1f9000000000001O00000000000000001O000000000000001O00000O100000001O00000000001O1O001O1O001O1O001N2N1N3M3M5K7I7H8I7I7Ikl]1"}, "label": "a smiling man in a blue shirt"}]} {"id": 439765, "image": "COCO_train2014_000000439765.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man holding the other man ' s tie\"."}], "task": "refering", "answer_template": "The \"the man holding the other man ' s tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 33, 34, 35, 36, 37, 56, 57, 58, 59, 60, 79, 80, 81, 82, 83, 102, 103, 104, 105, 106, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 169, 170, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 223, 240, 241, 242, 243, 244, 245, 246, 260, 263, 264, 265, 266, 267, 268, 269, 282, 283, 284, 287, 288, 289, 290, 291, 292, 293, 294, 305, 306, 307, 310, 311, 312, 313, 314, 315, 316, 317, 337, 338, 339, 340], "bbox": [0.269296875, 0.01360655737704918, 0.826046875, 0.9886651053864169], "iscrowd": 0, "area": 63159.44775, "rle": {"size": [427, 640], "counts": "T`X29Q=8G7J1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O12N2N2N1O2N2N2N2N1O2N2N2N2N2N1O2NkJeN\\NY1d1hN]NV1b1lN_NQ1a1QO_Nn0a1SO`Nk0_1WObNg0]1[OeNb0[1_OfN?Y1ChN;W1GjN7V1JkN3U1OlNMU15lNGV1:lN@V1b0kNZOV1h0f52M3M4M2M3M3N3L3M3N2M4L3N2M3M3N3L3M3N2M4L3N2M3M4M2M3N2M3M4M2M3M3N3L3M3N2M4L3N2M4L3N3L3M4M2M4L3N3L3M4M2M4nM_JSLd5o2_KjLd4Q3fKgL\\4U3mKcLW4X3PLcLR4Y3QLfLR4U3PLkLR4Q3PLnLT4m2nKSMT4i2nKWMU4c2nK\\MU4`2nK_MT4]2nKcMT4Y2nKfMV4U2lKkMV4Q2lKoMV4m1PLoMR4m1VLlMn3o1ZLkMh3Q2`LiMb3S2n3N3L3N2N2N1O00000000000O100000000000001O000000000O1000000000000000001O000000000000000O100000000000001O00000000000001O00001O000010O0001O01O01O000010O0001O01O01O000PH`NV4`1`KkN^4U1YKVOe4j0QKBm4`0lJGS5:jJIT5:iJFW5=eJEZ5=cJE]5=_JE`5>\\JDc5>ZJDe5?XJAh5b0TJ@l5a0QJAn5b0nI@Q6f1hH\\NW7l310O1O1O10O01O1O2WLSJ1n5LRJ5o5JQJ7P6FQJS8^OQHb0o7\\OSHd0n7ZOSHf0o7WORHi0P8TOQHl0P8QOSHm0P8POQHP1Q8mNPHS1R8iNPHW1R8eNPH[1R8aNPH_1R8]NPHc1Q8ZNQHf1Q8VNRHi1P8SNRHm1f9000000000001O00000000000000001O000000000000001O00000O100000001O00000000001O1O001O1O001O1O001N2N1N3M3M5K7I7H8I7I7Ikl]1"}, "label": "the man holding the other man ' s tie"}]} {"id": 439765, "image": "COCO_train2014_000000439765.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue striped tie\"."}], "task": "refering", "answer_template": "The \"a blue striped tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 173, 174, 196, 197, 219, 220, 221, 243, 244, 245, 266, 267, 268, 269], "bbox": [0.514953125, 0.4218501170960187, 0.70975, 0.7833723653395784], "iscrowd": 0, "area": 4565.768999999998, "rle": {"size": [427, 640], "counts": "^kY42V=3M3M3N2M3M3M3N2M3M3N2O1O1N2O1O1O1O1O1O1O1O1VOSOUEn0j:j0O1O1O1O1O1O2N2O2N1O100O1O2N2O1OQOo0E;E;FdEVOl8=^GE^NNa90\\H3oM1T<1gC5V in this image.", "objects": [{"patches": [150, 151, 173, 174, 196, 197, 219, 220, 221, 243, 244, 245, 266, 267, 268, 269], "bbox": [0.514953125, 0.4218501170960187, 0.70975, 0.7833723653395784], "iscrowd": 0, "area": 4565.768999999998, "rle": {"size": [427, 640], "counts": "^kY42V=3M3M3N2M3M3M3N2M3M3N2O1O1N2O1O1O1O1O1O1O1O1VOSOUEn0j:j0O1O1O1O1O1O2N2O2N1O100O1O2N2O1OQOo0E;E;FdEVOl8=^GE^NNa90\\H3oM1T<1gC5V in this image.", "objects": [{"patches": [168, 169, 191, 192, 193, 214, 215, 216, 237, 238, 239, 260, 261, 262, 282, 283, 285, 286, 305, 306, 309], "bbox": [0.27303125, 0.42737499999999995, 0.45546875, 0.7898541666666666], "iscrowd": 0, "area": 7083.7986, "rle": {"size": [480, 640], "counts": "[\\b25j>3N0O1O100O1O1O100O2N100O1M3L4M3L4M3N2N2M3N2N]O^BLQ=LoB72LnJ\\<8VCM`0JY<:WCKf=6ZBIk;KdE=aNGi;4_Ee0Q:iNUEe0h0b0S:mNoDe0l0>V:mNlDh0l0;X:nNlDi0i0NQO]O[;LjDj0i0NROZO\\;OiDj0f0OUOWOX;7kDc0g00VOPO^;>cDc0h00g:^O`Db0h01b:f0]E[Oc:e0\\E\\Od:d0ZE^Of:b0YE_Of:b0YE_Og:a0XE@h:`0WE@j:`0UEAk:`0SEAm:d0lD^OT;W200001O00001O001N3N1O00000000aDjLXfDY10WNZ;?fD\\11SNY;`0fD^12QNY;`0dD_1k;`NTDa1m;^NSDP1b0cNQ;<]DR1b0oNe:MiDU1c0oNb:KkDV1d0PO`:IkDY1e0PO_:ElD\\1f0PO^:AlD`1f0PO^:^OlDc1f0QOe:U1XEkNh:T1XEnNg:n0\\ETOc:g0aEZO_:b0eE]O[:?iEBW::lEFU:H`DG`1a0P:HfD]O]1k0n9G]F8c9H^F6c9I^F6c9I_F:]9FcF:]9EeF;[9BhF=Y9@iFc0U9ZOnFh0P9UORGi0g;L3L5K4M1N2N2O1N3M2O^WS5"}, "label": "base ball player ready to swing"}]} {"id": 292315, "image": "COCO_train2014_000000292315.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man in white and red baseball uniform batting at plate\"."}], "task": "refering", "answer_template": "The \"man in white and red baseball uniform batting at plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 191, 192, 193, 214, 215, 216, 237, 238, 239, 260, 261, 262, 282, 283, 285, 286, 305, 306, 309], "bbox": [0.27303125, 0.42737499999999995, 0.45546875, 0.7898541666666666], "iscrowd": 0, "area": 7083.7986, "rle": {"size": [480, 640], "counts": "[\\b25j>3N0O1O100O1O1O100O2N100O1M3L4M3L4M3N2N2M3N2N]O^BLQ=LoB72LnJ\\<8VCM`0JY<:WCKf=6ZBIk;KdE=aNGi;4_Ee0Q:iNUEe0h0b0S:mNoDe0l0>V:mNlDh0l0;X:nNlDi0i0NQO]O[;LjDj0i0NROZO\\;OiDj0f0OUOWOX;7kDc0g00VOPO^;>cDc0h00g:^O`Db0h01b:f0]E[Oc:e0\\E\\Od:d0ZE^Of:b0YE_Of:b0YE_Og:a0XE@h:`0WE@j:`0UEAk:`0SEAm:d0lD^OT;W200001O00001O001N3N1O00000000aDjLXfDY10WNZ;?fD\\11SNY;`0fD^12QNY;`0dD_1k;`NTDa1m;^NSDP1b0cNQ;<]DR1b0oNe:MiDU1c0oNb:KkDV1d0PO`:IkDY1e0PO_:ElD\\1f0PO^:AlD`1f0PO^:^OlDc1f0QOe:U1XEkNh:T1XEnNg:n0\\ETOc:g0aEZO_:b0eE]O[:?iEBW::lEFU:H`DG`1a0P:HfD]O]1k0n9G]F8c9H^F6c9I^F6c9I_F:]9FcF:]9EeF;[9BhF=Y9@iFc0U9ZOnFh0P9UORGi0g;L3L5K4M1N2N2O1N3M2O^WS5"}, "label": "man in white and red baseball uniform batting at plate"}]} {"id": 292315, "image": "COCO_train2014_000000292315.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man squatting in a protective head gear between the batter and umpire\"."}], "task": "refering", "answer_template": "The \"man squatting in a protective head gear between the batter and umpire\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 219, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 286, 287, 288, 289, 309, 310, 311, 312], "bbox": [0.39596875, 0.5348958333333333, 0.612421875, 0.8279583333333332], "iscrowd": 0, "area": 10623.406799999997, "rle": {"size": [480, 640], "counts": "Q\\g34j>2M4L3N3L3M3O2O001O001O00001O0000[OPB:P>ATB`0V>000O10O1000O1fAKi=5WB1c=O]B6^=JbB6]=KdB5[=KeB6Z=JfB7Y=IgB8W=IiB8V=HjB9U=GkB9T=HlB9S=GnB9Q=GoB:P=FPC;noNc;e0PD<;QOc;d0RD<:POc;d0TD=7QOj:0ZEb0E>5QOe:7_E;G>4POa:>bE4I>3QO_:a0cE1K>1QO_:d0bENN>0PO_:f0bEL0>MQOa:g0_EK3>KQOc:h0]EI5?JPOd:i0[EI7?HPOe:i1cED^:eEAZ:`0gE_OY:a0iE]OW:c0jE[OW:e0jEZOV:f0kEYOU:g0lEXOT:h0mEWOS:i0nEVOR:m0lEROT:Q1hEPOX:R1fEnNZ:U1bEkN_:X1^EhNb:[1ZEfNf:j21O1O1O1O1O1N3N3B>M3M3M3M3M1O00001O0000001O0000001N1004L5L0O00010O0002N2N2O1N2N2N2L4M3L3M3M3M3M4J5M3O1O1O2N1N2N8I8G9G;E:Gelc3"}, "label": "man squatting in a protective head gear between the batter and umpire"}]} {"id": 292315, "image": "COCO_train2014_000000292315.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the catcher\"."}], "task": "refering", "answer_template": "The \"the catcher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 219, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 286, 287, 288, 289, 309, 310, 311, 312], "bbox": [0.39596875, 0.5348958333333333, 0.612421875, 0.8279583333333332], "iscrowd": 0, "area": 10623.406799999997, "rle": {"size": [480, 640], "counts": "Q\\g34j>2M4L3N3L3M3O2O001O001O00001O0000[OPB:P>ATB`0V>000O10O1000O1fAKi=5WB1c=O]B6^=JbB6]=KdB5[=KeB6Z=JfB7Y=IgB8W=IiB8V=HjB9U=GkB9T=HlB9S=GnB9Q=GoB:P=FPC;noNc;e0PD<;QOc;d0RD<:POc;d0TD=7QOj:0ZEb0E>5QOe:7_E;G>4POa:>bE4I>3QO_:a0cE1K>1QO_:d0bENN>0PO_:f0bEL0>MQOa:g0_EK3>KQOc:h0]EI5?JPOd:i0[EI7?HPOe:i1cED^:eEAZ:`0gE_OY:a0iE]OW:c0jE[OW:e0jEZOV:f0kEYOU:g0lEXOT:h0mEWOS:i0nEVOR:m0lEROT:Q1hEPOX:R1fEnNZ:U1bEkN_:X1^EhNb:[1ZEfNf:j21O1O1O1O1O1N3N3B>M3M3M3M3M1O00001O0000001O0000001N1004L5L0O00010O0002N2N2O1N2N2N2L4M3L3M3M3M3M4J5M3O1O1O2N1N2N8I8G9G;E:Gelc3"}, "label": "the catcher"}]} {"id": 292315, "image": "COCO_train2014_000000292315.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umpire in black behind catcher\"."}], "task": "refering", "answer_template": "The \"umpire in black behind catcher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 199, 200, 201, 221, 222, 223, 224, 243, 244, 245, 246, 247, 248, 268, 269, 270, 291, 292, 293, 294, 314, 315, 316, 317, 337, 338], "bbox": [0.586109375, 0.44554166666666667, 0.801484375, 0.8310625], "iscrowd": 0, "area": 10745.010449999996, "rle": {"size": [480, 640], "counts": "ZR`52k>4M0YAKe>90O01O100O100001O001O1O0O2O1O00O100O1O010O1N20001N100O1O1O10O01O1O001O001M2eN]OZDd0d;^O[Dc0c;^O^Da0^;DaD=^;E`D;a;G`CFa0d0[2]O[66T9a0\\FXOc06Q9b0[FXOl0]OkN6n9V1YFWOW1TOfN in this image.", "objects": [{"patches": [176, 199, 200, 201, 221, 222, 223, 224, 243, 244, 245, 246, 247, 248, 268, 269, 270, 291, 292, 293, 294, 314, 315, 316, 317, 337, 338], "bbox": [0.586109375, 0.44554166666666667, 0.801484375, 0.8310625], "iscrowd": 0, "area": 10745.010449999996, "rle": {"size": [480, 640], "counts": "ZR`52k>4M0YAKe>90O01O100O100001O001O1O0O2O1O00O100O1O010O1N20001N100O1O1O10O01O1O001O001M2eN]OZDd0d;^O[Dc0c;^O^Da0^;DaD=^;E`D;a;G`CFa0d0[2]O[66T9a0\\FXOc06Q9b0[FXOl0]OkN6n9V1YFWOW1TOfN in this image.", "objects": [{"patches": [164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332], "bbox": [0.025484374999999997, 0.5078688524590165, 0.47528125, 0.9865105386416863], "iscrowd": 0, "area": 50043.95695000001, "rle": {"size": [427, 640], "counts": "jo61V=`0@?A=C6J6J6J6K5J5K6J6L4M3M3M3L4M2N3M4L3M4L3L5L4L3M4L3M1N2O2M2N3N1O1O2N1N3N1O1O1O1O1O1N2O1O1O2N1O1O1O10000O10000O100O10000O10000O10000O100O10000O10000O100O100O100O100O100O100000000O100000000O10000000000O100000000O100000000O10000000000000000O10001O0000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O0001O0001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O00001O1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2M2O2N1O2N2N1N3N1O2N2N1N3N1N3M3M2N3L3N3M3M2N3M2N3M2N3M3M2M4C in this image.", "objects": [{"patches": [164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332], "bbox": [0.025484374999999997, 0.5078688524590165, 0.47528125, 0.9865105386416863], "iscrowd": 0, "area": 50043.95695000001, "rle": {"size": [427, 640], "counts": "jo61V=`0@?A=C6J6J6J6K5J5K6J6L4M3M3M3L4M2N3M4L3M4L3L5L4L3M4L3M1N2O2M2N3N1O1O2N1N3N1O1O1O1O1O1N2O1O1O2N1O1O1O10000O10000O100O10000O10000O10000O100O10000O10000O100O100O100O100O100O100000000O100000000O10000000000O100000000O100000000O10000000000000000O10001O0000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O0001O0001O0000001O0000001O000000001O0000001O0000001O000000001O0000001O0000001O00001O1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2M2O2N1O2N2N1N3N1O2N2N1N3N1N3M3M2N3L3N3M3M2N3M2N3M2N3M3M2M4C in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 32, 33, 34, 35, 36, 37, 38, 39, 40, 55, 56, 57, 58, 59, 60, 61, 62, 78, 79, 80, 101, 102, 103], "bbox": [0.4033125, 0.0022482435597189696, 0.74815625, 0.3236065573770492], "iscrowd": 0, "area": 16395.651749999994, "rle": {"size": [427, 640], "counts": "ed[32T=9F9J7J6K4K6J5K6K5J5K7I7J6I8H7I6K2N100O1O100O2O0O1O100O1O2O0O100O1O2O0O100O1O101N100O10000O2O000O10000O2O00fNXGQNh8n1nG]MQ8d2[1O10O0100O010O10O0100O100O100O10000O100O100O100O100O100O100O100O100O100O100O1000000O10000000000O10000000000O10000000000O10000000O10O10000000000O10000000000O1000000000000O10000000000O10000000000O10000000000O10000000000O100000000000000O10001O001O00001N101O001O001O000O2O001O001O001O0O2O00001O001O001O0O2O00002N1O1O1M4K4K5K6K4K5K5L`SS2"}, "label": "a container with black stuff in it"}]} {"id": 128475, "image": "COCO_train2014_000000128475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white glass bowl with a black dip\"."}], "task": "refering", "answer_template": "The \"white glass bowl with a black dip\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 32, 33, 34, 35, 36, 37, 38, 39, 40, 55, 56, 57, 58, 59, 60, 61, 62, 78, 79, 80, 101, 102, 103], "bbox": [0.4033125, 0.0022482435597189696, 0.74815625, 0.3236065573770492], "iscrowd": 0, "area": 16395.651749999994, "rle": {"size": [427, 640], "counts": "ed[32T=9F9J7J6K4K6J5K6K5J5K7I7J6I8H7I6K2N100O1O100O2O0O1O100O1O2O0O100O1O2O0O100O1O101N100O10000O2O000O10000O2O00fNXGQNh8n1nG]MQ8d2[1O10O0100O010O10O0100O100O100O10000O100O100O100O100O100O100O100O100O100O100O1000000O10000000000O10000000000O10000000000O10000000O10O10000000000O10000000000O1000000000000O10000000000O10000000000O10000000000O10000000000O100000000000000O10001O001O00001N101O001O001O000O2O001O001O001O0O2O00001O001O001O0O2O00002N1O1O1M4K4K5K6K4K5K5L`SS2"}, "label": "white glass bowl with a black dip"}]} {"id": 128475, "image": "COCO_train2014_000000128475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the middle donut\"."}], "task": "refering", "answer_template": "The \"the middle donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 192, 193, 194, 195, 216, 217, 218, 240, 241, 263, 264], "bbox": [0.07496875, 0.2876346604215457, 0.502265625, 0.7910070257611241], "iscrowd": 0, "area": 23257.003649999984, "rle": {"size": [427, 640], "counts": "^Xd01Y=0O100N3M2M3N2N3L3N2M4M2N2M3N3M2N2O101N100O2O000O100O2O0O100O2O0O100O2O1O1O1O1O1O1O1N2O1O001O1O1O1N2O1O1O1O1O1O001O100O1O100O100O100O010O100O1O10O0100O10O0100O100O00100O100O010O100O10O01O100O10O0100O100O010O10000000O1000O1000O100000O1000O10000000O10001O00000O10000000001O00000001O00000001O0001O000001O0001O01O00001O0001O01O00000010O0000010O000010O0000010O0001O01O010O001O010O0010O00010O010O010O10O010O010O01000O010O100O10O100000000O100001O0O2O00001O001O001O00001O001N101O002N1O1O1O1O1O101N1O1O1O1O1O2N1O3Mh0XOM3N2N2N3M3M3M3M3K5J6I7J6I8I9G9E:G:E>C[iT4"}, "label": "the middle donut"}]} {"id": 128475, "image": "COCO_train2014_000000128475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"second sweet in the line\"."}], "task": "refering", "answer_template": "The \"second sweet in the line\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 185, 186, 192, 193, 194, 195, 216, 217, 218, 240, 241, 263, 264], "bbox": [0.07496875, 0.2876346604215457, 0.502265625, 0.7910070257611241], "iscrowd": 0, "area": 23257.003649999984, "rle": {"size": [427, 640], "counts": "^Xd01Y=0O100N3M2M3N2N3L3N2M4M2N2M3N3M2N2O101N100O2O000O100O2O0O100O2O0O100O2O1O1O1O1O1O1O1N2O1O001O1O1O1N2O1O1O1O1O1O001O100O1O100O100O100O010O100O1O10O0100O10O0100O100O00100O100O010O100O10O01O100O10O0100O100O010O10000000O1000O1000O100000O1000O10000000O10001O00000O10000000001O00000001O00000001O0001O000001O0001O01O00001O0001O01O00000010O0000010O000010O0000010O0001O01O010O001O010O0010O00010O010O010O10O010O010O01000O010O100O10O100000000O100001O0O2O00001O001O001O00001O001N101O002N1O1O1O1O1O101N1O1O1O1O1O2N1O3Mh0XOM3N2N2N3M3M3M3M3K5J6I7J6I8I9G9E:G:E>C[iT4"}, "label": "second sweet in the line"}]} {"id": 128475, "image": "COCO_train2014_000000128475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an ice cream cup\"."}], "task": "refering", "answer_template": "The \"an ice cream cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 63, 64, 80, 81, 82, 83, 84, 85, 86, 87, 88, 103, 104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 133, 134, 149, 150, 155, 156, 179], "bbox": [0.48432812500000005, 0.14522248243559718, 0.83284375, 0.5138407494145198], "iscrowd": 0, "area": 20126.710049999998, "rle": {"size": [427, 640], "counts": "W\\Q4;i in this image.", "objects": [{"patches": [58, 59, 60, 61, 62, 63, 64, 80, 81, 82, 83, 84, 85, 86, 87, 88, 103, 104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 133, 134, 149, 150, 155, 156, 179], "bbox": [0.48432812500000005, 0.14522248243559718, 0.83284375, 0.5138407494145198], "iscrowd": 0, "area": 20126.710049999998, "rle": {"size": [427, 640], "counts": "W\\Q4;i in this image.", "objects": [{"patches": [40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160], "bbox": [0.764390625, 0.07199063231850117, 1.0, 0.42971896955503514], "iscrowd": 0, "area": 17019.741, "rle": {"size": [427, 640], "counts": "Po[61W=3M4K5L4N101O1N101O1O1O001O1N101O1O1O001O1N2O001O1O001N2O1O001O1O001N2O1ZFjNU7V1jHYOh6h0VIZOi6g0UIZOk6f0TI[Ol6f0RI[Om6g0QIZOo6k0kHVOU7[1YHfNf7l1hGUNX8\\2VGeMj8^3O10000O10000O10000O10000O1000000O10000O10O10O10000O100000000000000000000000000001O000000000000000000000000000O1000001O0000000000000000000000001O001O0O2O001O1O001O0O2O001O1O001N101O1O001O001N101O1O001O001N10]N"}, "label": "the cup of cream that is farthest back and right"}]} {"id": 128475, "image": "COCO_train2014_000000128475.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ramekin with yellow custard behind another ramekin\"."}], "task": "refering", "answer_template": "The \"a ramekin with yellow custard behind another ramekin\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 41, 42, 43, 44, 45, 63, 64, 65, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160], "bbox": [0.764390625, 0.07199063231850117, 1.0, 0.42971896955503514], "iscrowd": 0, "area": 17019.741, "rle": {"size": [427, 640], "counts": "Po[61W=3M4K5L4N101O1N101O1O1O001O1N101O1O1O001O1N2O001O1O001N2O1O001O1O001N2O1ZFjNU7V1jHYOh6h0VIZOi6g0UIZOk6f0TI[Ol6f0RI[Om6g0QIZOo6k0kHVOU7[1YHfNf7l1hGUNX8\\2VGeMj8^3O10000O10000O10000O10000O1000000O10000O10O10O10000O100000000000000000000000000001O000000000000000000000000000O1000001O0000000000000000000000001O001O0O2O001O1O001O0O2O001O1O001N101O1O001O001N101O1O001O001N10]N"}, "label": "a ramekin with yellow custard behind another ramekin"}]} {"id": 87522, "image": "COCO_train2014_000000087522.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bananas closest to the hand on top\"."}], "task": "refering", "answer_template": "The \"the bananas closest to the hand on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 73, 74, 75, 76, 77, 89, 90, 91, 92, 93, 106, 107, 108, 109, 110, 123, 124, 125, 126, 127, 140, 142, 143, 156, 157, 173, 174, 190, 191, 207, 208], "bbox": [0.1890625, 0.13984375, 0.5781458333333334, 0.548640625], "iscrowd": 0, "area": 16538.293100000003, "rle": {"size": [640, 480], "counts": "hTi1f0Vc09I7J6I8H8I7H8I2N2N2N2N2N2N2N2O1N1O2N01O001hNo^OAQa0;\\_OZOe`0c0h_OjNb`0S1W1L4M2N^OnN\\^Oo0fa0SOY^Oj0ja0XOS^Of0Wa0WOo^O5Ia0Za0]Oj^O4J=_a0@e^O5J9ca0E`^O5K3ha0I[^O6KOma0MT^O7NIQb01o]O9MEUb05j]O9OA[a0J^_O`1^`0lN\\_OS1``0VO[_Oi0d`0\\OZ_Ob0e`0CY_O;f`0IY_O5f`01V_ONj`06T_OHk`0>R_O@m`0g12I7N3N20O10000O10000O2O000O1000000O2O000O1000010O0010O01O1U_OaMT`0W3O2M3M2N3N2M2N3N2M2N2N100O1O100O100O1000000000000O10000000000O1000000000O10O0100O100O100O100O100O1O100O100O101M2C=M3M3M3M3M3N2N2N2N2N2N3M2N2M3M3M3M3M4L3M4L3N3L3M4L3M4L3M4L3M4M2O2N1O2O0O2N2N2O1N2N2N101N2N2O1N2N2N2O^hn3"}, "label": "the bananas closest to the hand on top"}]} {"id": 87522, "image": "COCO_train2014_000000087522.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"left most three banana on the head of the girl\"."}], "task": "refering", "answer_template": "The \"left most three banana on the head of the girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 60, 73, 74, 75, 76, 77, 89, 90, 91, 92, 93, 106, 107, 108, 109, 110, 123, 124, 125, 126, 127, 140, 142, 143, 156, 157, 173, 174, 190, 191, 207, 208], "bbox": [0.1890625, 0.13984375, 0.5781458333333334, 0.548640625], "iscrowd": 0, "area": 16538.293100000003, "rle": {"size": [640, 480], "counts": "hTi1f0Vc09I7J6I8H8I7H8I2N2N2N2N2N2N2N2O1N1O2N01O001hNo^OAQa0;\\_OZOe`0c0h_OjNb`0S1W1L4M2N^OnN\\^Oo0fa0SOY^Oj0ja0XOS^Of0Wa0WOo^O5Ia0Za0]Oj^O4J=_a0@e^O5J9ca0E`^O5K3ha0I[^O6KOma0MT^O7NIQb01o]O9MEUb05j]O9OA[a0J^_O`1^`0lN\\_OS1``0VO[_Oi0d`0\\OZ_Ob0e`0CY_O;f`0IY_O5f`01V_ONj`06T_OHk`0>R_O@m`0g12I7N3N20O10000O10000O2O000O1000000O2O000O1000010O0010O01O1U_OaMT`0W3O2M3M2N3N2M2N3N2M2N2N100O1O100O100O1000000000000O10000000000O1000000000O10O0100O100O100O100O100O1O100O100O101M2C=M3M3M3M3M3N2N2N2N2N2N3M2N2M3M3M3M3M4L3M4L3N3L3M4L3M4L3M4L3M4M2O2N1O2O0O2N2N2O1N2N2N101N2N2O1N2N2N2O^hn3"}, "label": "left most three banana on the head of the girl"}]} {"id": 87522, "image": "COCO_train2014_000000087522.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing bananas on her head\"."}], "task": "refering", "answer_template": "The \"a woman wearing bananas on her head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 162, 173, 174, 175, 176, 177, 178, 179, 180, 191, 192, 193, 194, 195, 196, 197, 198, 208, 209, 210, 211, 212, 213, 214, 215, 216, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 254, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 359, 360, 361, 362, 363, 364, 365, 366, 367, 375, 376, 377, 378, 379, 380, 381], "bbox": [0.09566666666666666, 0.3070625, 1.0, 0.988671875], "iscrowd": 0, "area": 94003.9675, "rle": {"size": [640, 480], "counts": "c[m01lc03N2M4M2M3N2M3N2M3N2M3M3N2M3N2M3N2M3O100O1O100O1O1O100O1O101N1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O2N100O1O100nFcM`0^2\\OdNE]17FjN:R1h0oMYOm1i1TMXNh2V2mLkMo2[2oLeMm2`2QMaMk2d2SM]Mj2h2TMXMk2j2SMWMk2m2RMTMm2n2RMRMm2P3QMQMn2R3oLoLP3S3nLnLQ3T3nLlLP3X3mLiLR3Y3lLhLS3Z3lLfLS3]3jLdLU3^3iLdLU3^3jLbLT3b3iL_LV3c3hL^LW3d3gL]LX3f3fLZLY3h3eLYLY3j3eLWLZ3l3dLTL[3n3cLSL\\3o3cLQL\\3R4aLoK^3S4aLmK_3T4`LlK`3V4^LjKb3W4\\LjKd3X4ZLhKg3X4XLhKh3Y4VLhKj3Z4TLfKl3[4SLeKm3\\4RLdKo3]4nKdKR4]4mKcKS4^4lKbKT4`4iKaKW4a4gK_KY4c4eK^K[4c4cK]K]4e4`K\\K`4f4^KZKb4i4[KWKe4k4YKUKh4l4UKUKk4m4SKSKm4o4QKQKo4Q5nJoJS5S5kJlJW5U5gJjJZ5Y5cJgJ]5[5`JeJb5\\5\\JcJh5\\5VJcJn5\\5oIeJT6Z5jIeJY6[5eIdJ_6[5_IdJe6\\5WIeJl6Z5RIeJR7Z5lHeJX7Z5eHfJ^7Z5`HfJa7[5]HdJd7^5ZHaJg7b5UH_Jl7j8001N101O001O0O2O001O001N10001O001N10000000001N100000000O1000001O0O2O1O001O1N101O1O001O1N2O001O1O0O2O1O001O1N2O001O1O001N2O001O1O0O2O1O1O001N2O001O1N101O1N2O001N2O001N2O001O1N2O1O1N2O2N1N2O1O2M2O1O1O2M2O1O1N3N1O1N2O2N1N2O1O2M2O1O1O2M101N1O2N1O2O0O2N1O2N1O101N1O2N1O2O0O2N1O2N101N1O2N1O101N1O2N1O2N1O2N1O2N1O2N1O2M2O2N1O2N2N3M2N2N2N3M2N2O2N1O1N3]MQBQOP>l0QBUOP>g0SBYOn=c0TB^On=>TBBm=:UBFm=6VBJl=2UBOl=MWB3j=IYB7h=FYB;i=@ZB?j=[OXBf0n=ROTBn0R>iNPBX1V>_NmAa1Y>WNhAj1^>mMeAS2b>cM`A^2Q`00O100O010O100O10000O100O1000O0100O10000O100O1000O0100O100O10000O100O10O10O100O10000O10O010000O10O010000O10O010000O010O10000O010O100O10O0100O1O10O01O100O1O010O1O100O0001O0O101O000O2L3L4M4K4M3M4K4M3L5K4C=De0ZO`C"}, "label": "a woman wearing bananas on her head"}]} {"id": 87522, "image": "COCO_train2014_000000087522.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young , caucasian woman wearing a bunch of bananas on her head\"."}], "task": "refering", "answer_template": "The \"a young , caucasian woman wearing a bunch of bananas on her head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 162, 173, 174, 175, 176, 177, 178, 179, 180, 191, 192, 193, 194, 195, 196, 197, 198, 208, 209, 210, 211, 212, 213, 214, 215, 216, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 254, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 359, 360, 361, 362, 363, 364, 365, 366, 367, 375, 376, 377, 378, 379, 380, 381], "bbox": [0.09566666666666666, 0.3070625, 1.0, 0.988671875], "iscrowd": 0, "area": 94003.9675, "rle": {"size": [640, 480], "counts": "c[m01lc03N2M4M2M3N2M3N2M3N2M3M3N2M3N2M3N2M3O100O1O100O1O1O100O1O101N1O100O1O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O2N100O1O100nFcM`0^2\\OdNE]17FjN:R1h0oMYOm1i1TMXNh2V2mLkMo2[2oLeMm2`2QMaMk2d2SM]Mj2h2TMXMk2j2SMWMk2m2RMTMm2n2RMRMm2P3QMQMn2R3oLoLP3S3nLnLQ3T3nLlLP3X3mLiLR3Y3lLhLS3Z3lLfLS3]3jLdLU3^3iLdLU3^3jLbLT3b3iL_LV3c3hL^LW3d3gL]LX3f3fLZLY3h3eLYLY3j3eLWLZ3l3dLTL[3n3cLSL\\3o3cLQL\\3R4aLoK^3S4aLmK_3T4`LlK`3V4^LjKb3W4\\LjKd3X4ZLhKg3X4XLhKh3Y4VLhKj3Z4TLfKl3[4SLeKm3\\4RLdKo3]4nKdKR4]4mKcKS4^4lKbKT4`4iKaKW4a4gK_KY4c4eK^K[4c4cK]K]4e4`K\\K`4f4^KZKb4i4[KWKe4k4YKUKh4l4UKUKk4m4SKSKm4o4QKQKo4Q5nJoJS5S5kJlJW5U5gJjJZ5Y5cJgJ]5[5`JeJb5\\5\\JcJh5\\5VJcJn5\\5oIeJT6Z5jIeJY6[5eIdJ_6[5_IdJe6\\5WIeJl6Z5RIeJR7Z5lHeJX7Z5eHfJ^7Z5`HfJa7[5]HdJd7^5ZHaJg7b5UH_Jl7j8001N101O001O0O2O001O001N10001O001N10000000001N100000000O1000001O0O2O1O001O1N101O1O001O1N2O001O1O0O2O1O001O1N2O001O1O001N2O001O1O0O2O1O1O001N2O001O1N101O1N2O001N2O001N2O001O1N2O1O1N2O2N1N2O1O2M2O1O1O2M2O1O1N3N1O1N2O2N1N2O1O2M2O1O1O2M101N1O2N1O2O0O2N1O2N1O101N1O2N1O2O0O2N1O2N101N1O2N1O101N1O2N1O2N1O2N1O2N1O2N1O2M2O2N1O2N2N3M2N2N2N3M2N2O2N1O1N3]MQBQOP>l0QBUOP>g0SBYOn=c0TB^On=>TBBm=:UBFm=6VBJl=2UBOl=MWB3j=IYB7h=FYB;i=@ZB?j=[OXBf0n=ROTBn0R>iNPBX1V>_NmAa1Y>WNhAj1^>mMeAS2b>cM`A^2Q`00O100O010O100O10000O100O1000O0100O10000O100O1000O0100O100O10000O100O10O10O100O10000O10O010000O10O010000O10O010000O010O10000O010O100O10O0100O1O10O01O100O1O010O1O100O0001O0O101O000O2L3L4M4K4M3M4K4M3L5K4C=De0ZO`C"}, "label": "a young , caucasian woman wearing a bunch of bananas on her head"}]} {"id": 87522, "image": "COCO_train2014_000000087522.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person ' s hand positioning bananas on another person ' s head\"."}], "task": "refering", "answer_template": "The \"a person ' s hand positioning bananas on another person ' s head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 120, 121, 122, 123, 136, 137, 138, 139, 140, 153, 154, 155, 156, 170, 171, 172, 187, 188, 204, 205, 221, 222, 223, 238, 239, 240, 255, 256, 257, 272, 273, 274, 289, 290, 291, 306, 307, 308, 323, 324, 325, 340, 341, 357, 358, 374, 375], "bbox": [0.003125, 0.27790625, 0.27479166666666666, 0.9711093749999999], "iscrowd": 0, "area": 28697.365299999998, "rle": {"size": [640, 480], "counts": "nc1`7[8T410O100O100O1O100O100O100O1O100O100O1O1O00100O1O1O1O100O1O1O00N2O2N1O1N2O1O2M2O1O1N2TGhGP6Y8PJmGi5T8WJQHb5Q8_JSH[5n7eJXHT5i7lJ\\Hn4e7RKaHf4a7[KcH_4^7aKgHY4Y7hKmHP4U7QLoHi3R7WLSIc3n6]LXIZ3k6fLZIR3i6oL[Ii2h6WM^I`2e6`M`IX2c6iMbIn1a6RNdIf1_6ZNfI^1]6bNiIT1Z6nNjIj0Y6VOlIb0W6^OoI9T6HPJ0S60SJGP69UJ_On5b0VJVOm5j0YJlNk5U1YJcNj5]1[J[Nh5f1]JQNf5o1_JiMd5X2aJ_Ma5b2dJVM_5k2eJmL^5S3g510O100O100O100O1O1N2N2O1N2N2N2O1N2N2N2O1N200O1O10O01O100O1N2M3M3N2M3N2N2M3N2N2M3N2N2M3N2N4L3L5B=K6I[Zi6"}, "label": "a person ' s hand positioning bananas on another person ' s head"}]} {"id": 87522, "image": "COCO_train2014_000000087522.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"right hand of person with banana hat on\"."}], "task": "refering", "answer_template": "The \"right hand of person with banana hat on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 120, 121, 122, 123, 136, 137, 138, 139, 140, 153, 154, 155, 156, 170, 171, 172, 187, 188, 204, 205, 221, 222, 223, 238, 239, 240, 255, 256, 257, 272, 273, 274, 289, 290, 291, 306, 307, 308, 323, 324, 325, 340, 341, 357, 358, 374, 375], "bbox": [0.003125, 0.27790625, 0.27479166666666666, 0.9711093749999999], "iscrowd": 0, "area": 28697.365299999998, "rle": {"size": [640, 480], "counts": "nc1`7[8T410O100O100O1O100O100O100O1O100O100O1O1O00100O1O1O1O100O1O1O00N2O2N1O1N2O1O2M2O1O1N2TGhGP6Y8PJmGi5T8WJQHb5Q8_JSH[5n7eJXHT5i7lJ\\Hn4e7RKaHf4a7[KcH_4^7aKgHY4Y7hKmHP4U7QLoHi3R7WLSIc3n6]LXIZ3k6fLZIR3i6oL[Ii2h6WM^I`2e6`M`IX2c6iMbIn1a6RNdIf1_6ZNfI^1]6bNiIT1Z6nNjIj0Y6VOlIb0W6^OoI9T6HPJ0S60SJGP69UJ_On5b0VJVOm5j0YJlNk5U1YJcNj5]1[J[Nh5f1]JQNf5o1_JiMd5X2aJ_Ma5b2dJVM_5k2eJmL^5S3g510O100O100O100O1O1N2N2O1N2N2N2O1N2N2N2O1N200O1O10O01O100O1N2M3M3N2M3N2N2M3N2N2M3N2N2M3N2N4L3L5B=K6I[Zi6"}, "label": "right hand of person with banana hat on"}]} {"id": 349686, "image": "COCO_train2014_000000349686.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a batter hitting a baseball\"."}], "task": "refering", "answer_template": "The \"a batter hitting a baseball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 173, 174, 175, 176, 196, 197, 198, 199, 200, 219, 220, 221, 242, 243, 244, 264, 265, 267, 287, 288, 310], "bbox": [0.47582812499999994, 0.3940208333333333, 0.709046875, 0.7823125], "iscrowd": 0, "area": 8776.4324, "rle": {"size": [480, 640], "counts": "\\Z_48g>4L1O1O1N2O001O1O1O1N101O1O1O1O0O2O1M3M3L3N3L4M2M3N2YL`NlIb1T6bNgI`1X6dNbI_1]6eN_I[1c6gNYIZ1h6hNWG_Om0k1m7hNPGEm0c1U8jNhFLk0\\1^8jNbF1j0V1e8kN[F7i0n0o87kFHW9=cFAa9c0XF]Ok9g0oEXOT:g23L4M4L3M4N1N3N1O2N2M9H9G2N1N101O1O1N2O001O1N2O1O1O0010000000O100000000O110OiMmCb1S<]NQD`1n;`NWD\\1h;dN\\DX1e;hN^DU1a;kNcDR1\\;nNhDo0X;POmDl0R;TOREh0o:WOUEf0j:ZOWEf0h:[OXEe0h:ZOZEf0d:ZO]Eg0b:XO_Ej0^:VOcEk0[:UOgEk0X:UOhEm0U:SOlEP1Q:oNQFU1i9kNXFZ1b9fN_F]1^9bNdF\\1\\9dNiFW1X9iNlFR1T9nNRGm0m8SOYGg0h8XO]Ga0e8_OaG:`8FcG6^8JcG3`8MaG0`80aGN`82aGKa85aGH`88aGFa8:R300000010O000001O01O0000010O000001O01O0001O0001O0001O0001O01O0000GXof2"}, "label": "a batter hitting a baseball"}]} {"id": 349686, "image": "COCO_train2014_000000349686.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player swinging at bat\"."}], "task": "refering", "answer_template": "The \"a baseball player swinging at bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 173, 174, 175, 176, 196, 197, 198, 199, 200, 219, 220, 221, 242, 243, 244, 264, 265, 267, 287, 288, 310], "bbox": [0.47582812499999994, 0.3940208333333333, 0.709046875, 0.7823125], "iscrowd": 0, "area": 8776.4324, "rle": {"size": [480, 640], "counts": "\\Z_48g>4L1O1O1N2O001O1O1O1N101O1O1O1O0O2O1M3M3L3N3L4M2M3N2YL`NlIb1T6bNgI`1X6dNbI_1]6eN_I[1c6gNYIZ1h6hNWG_Om0k1m7hNPGEm0c1U8jNhFLk0\\1^8jNbF1j0V1e8kN[F7i0n0o87kFHW9=cFAa9c0XF]Ok9g0oEXOT:g23L4M4L3M4N1N3N1O2N2M9H9G2N1N101O1O1N2O001O1N2O1O1O0010000000O100000000O110OiMmCb1S<]NQD`1n;`NWD\\1h;dN\\DX1e;hN^DU1a;kNcDR1\\;nNhDo0X;POmDl0R;TOREh0o:WOUEf0j:ZOWEf0h:[OXEe0h:ZOZEf0d:ZO]Eg0b:XO_Ej0^:VOcEk0[:UOgEk0X:UOhEm0U:SOlEP1Q:oNQFU1i9kNXFZ1b9fN_F]1^9bNdF\\1\\9dNiFW1X9iNlFR1T9nNRGm0m8SOYGg0h8XO]Ga0e8_OaG:`8FcG6^8JcG3`8MaG0`80aGN`82aGKa85aGH`88aGFa8:R300000010O000001O01O0000010O000001O01O0001O0001O0001O0001O01O0000GXof2"}, "label": "a baseball player swinging at bat"}]} {"id": 349686, "image": "COCO_train2014_000000349686.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the baseball umpire wearing black and gray\"."}], "task": "refering", "answer_template": "The \"the baseball umpire wearing black and gray\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 234, 235, 236, 257, 258, 259, 279, 280, 281, 282, 302, 303, 304, 305, 306, 325, 328, 329, 351, 352], "bbox": [0.1368125, 0.5560416666666667, 0.31812500000000005, 0.920875], "iscrowd": 0, "area": 8948.742199999999, "rle": {"size": [480, 640], "counts": "^dY13j>4L3M3N3L3M4L3K6I6K6L3M3N3L3M4L3NO2O1N2O2M2M10N2O1O1O1O3M4M4N100O2N101N1O100O2O0O1O100O2O0O1O101N1O100O2O0O1O100O2B=N2O1O101N1O100O1O101O000O1000001O0O1000fNiDkNW;BhDh03DU;CjDh03CR;EmDg03C8^Of97QFf04C1Lb9J\\Fe03CK6d9_O`Fg03BDa0i9QObFk01X1]9hMdFo00Z1_9aMcFT1N[1V:dNkE[1V:dNkE[1V:dNjE\\1W:bNkE]1V:bNjE^1X:`NiE_1[:]NeEc1`:XNaEg1d:TN\\El1i:oMXEP2m;O001O001O001O3L5K6]OkBPO[=g0jBWO\\=`0hB_O^=8fBFa=1cBNY>N1O1Od]\\6"}, "label": "the baseball umpire wearing black and gray"}]} {"id": 349686, "image": "COCO_train2014_000000349686.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the umpire in a black shirt standing behind the baseball players\"."}], "task": "refering", "answer_template": "The \"the umpire in a black shirt standing behind the baseball players\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [212, 213, 234, 235, 236, 257, 258, 259, 279, 280, 281, 282, 302, 303, 304, 305, 306, 325, 328, 329, 351, 352], "bbox": [0.1368125, 0.5560416666666667, 0.31812500000000005, 0.920875], "iscrowd": 0, "area": 8948.742199999999, "rle": {"size": [480, 640], "counts": "^dY13j>4L3M3N3L3M4L3K6I6K6L3M3N3L3M4L3NO2O1N2O2M2M10N2O1O1O1O3M4M4N100O2N101N1O100O2O0O1O100O2O0O1O101N1O100O2O0O1O100O2B=N2O1O101N1O100O1O101O000O1000001O0O1000fNiDkNW;BhDh03DU;CjDh03CR;EmDg03C8^Of97QFf04C1Lb9J\\Fe03CK6d9_O`Fg03BDa0i9QObFk01X1]9hMdFo00Z1_9aMcFT1N[1V:dNkE[1V:dNkE[1V:dNjE\\1W:bNkE]1V:bNjE^1X:`NiE_1[:]NeEc1`:XNaEg1d:TN\\El1i:oMXEP2m;O001O001O001O3L5K6]OkBPO[=g0jBWO\\=`0hB_O^=8fBFa=1cBNY>N1O1Od]\\6"}, "label": "the umpire in a black shirt standing behind the baseball players"}]} {"id": 333302, "image": "COCO_train2014_000000333302.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in flannel smiles\"."}], "task": "refering", "answer_template": "The \"a man in flannel smiles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 177, 178, 179, 180, 201, 202, 203, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 290, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 338, 339, 340, 341, 342], "bbox": [0.640328125, 0.3939578454332553, 0.973171875, 0.98903981264637], "iscrowd": 0, "area": 31472.948950000005, "rle": {"size": [427, 640], "counts": "jY[53W=3L4L4L6J6K6I6J2L4L4L4L4L4L4L4L4N2M3M3N2M3M3M3N2M2N2O1N2O1O100100N1O2N2M3K^G\\Mf6]2nHWNR7f1kH`NU7a1dHeN\\7]1\\HiNc7Y1XHkNh7W1SHlNm7U1oGnNQ8S1jGQOU8R1fGQOZ8P1bGSO^8m2O1O100O1O100O100RNPK`KR5]4TK^Km4a4XK[Kh4d4\\KXKe4f4aKUK`4j4eKRK[4m4jKnJW4Q5mKkJT4S5RLiJn3V5XLdJi3[5^L^Jc3a5`L]J`3a5cL]J^3b5eL[J\\3d5fL[JZ3d5hLZJY3e5W2O1O1O10000O2O00000000000O10000000000O100001O00001O00001O01O01O00001O00001O00001O001O1kIcJZ4_5cKdJ[4]5dKeJZ4\\5eKfJY4[5fKgJX4[5fKgJX4_5bKcJ\\4X7N100O00000000000000000000000000000000001O001O010O001O001O00001O001O001O001O001O1O1O001H8K5L4K4H9TOnFSM^9m2a000O1O1N2O1O1N2O1N3N1O1N2O1O1N2O1N3M2O1N3M2N2N3L3M4M2M3M4L3M3L4M3L4L4L4M3L4L4L4Jkg6"}, "label": "a man in flannel smiles"}]} {"id": 333302, "image": "COCO_train2014_000000333302.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man with a backpack\"."}], "task": "refering", "answer_template": "The \"the man with a backpack\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [154, 155, 156, 157, 177, 178, 179, 180, 201, 202, 203, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 290, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 338, 339, 340, 341, 342], "bbox": [0.640328125, 0.3939578454332553, 0.973171875, 0.98903981264637], "iscrowd": 0, "area": 31472.948950000005, "rle": {"size": [427, 640], "counts": "jY[53W=3L4L4L6J6K6I6J2L4L4L4L4L4L4L4L4N2M3M3N2M3M3M3N2M2N2O1N2O1O100100N1O2N2M3K^G\\Mf6]2nHWNR7f1kH`NU7a1dHeN\\7]1\\HiNc7Y1XHkNh7W1SHlNm7U1oGnNQ8S1jGQOU8R1fGQOZ8P1bGSO^8m2O1O100O1O100O100RNPK`KR5]4TK^Km4a4XK[Kh4d4\\KXKe4f4aKUK`4j4eKRK[4m4jKnJW4Q5mKkJT4S5RLiJn3V5XLdJi3[5^L^Jc3a5`L]J`3a5cL]J^3b5eL[J\\3d5fL[JZ3d5hLZJY3e5W2O1O1O10000O2O00000000000O10000000000O100001O00001O00001O01O01O00001O00001O00001O001O1kIcJZ4_5cKdJ[4]5dKeJZ4\\5eKfJY4[5fKgJX4[5fKgJX4_5bKcJ\\4X7N100O00000000000000000000000000000000001O001O010O001O001O00001O001O001O001O001O1O1O001H8K5L4K4H9TOnFSM^9m2a000O1O1N2O1O1N2O1N3N1O1N2O1O1N2O1N3M2O1N3M2N2N3L3M4M2M3M4L3M3L4M3L4L4L4M3L4L4L4Jkg6"}, "label": "the man with a backpack"}]} {"id": 112122, "image": "COCO_train2014_000000112122.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the side of the car visible in the mirror\"."}], "task": "refering", "answer_template": "The \"the side of the car visible in the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 394, 395, 396, 397, 398, 399, 400, 419], "bbox": [0.052593749999999995, 0.29411575562700965, 0.5716875, 0.8282315112540194], "iscrowd": 0, "area": 65990.40814999997, "rle": {"size": [622, 640], "counts": "and07Uc06I7J7H7I8I6I7J7H7I8I6I7I8I6I8I6I7I8I6I7J7H5K6K4K6J5I7I8I6I7J7H7I8I6I7J7H7M4L3N2N3M3M3L4M3M3M2M4M3M3M3M3L4M2N1O1N3N1O2N1O1N10001O0000001O0000001O000000001O0000001O000000001O0000001O0000001O000000001O0000001O000000001O0000001O0000000010O1O1O1O1N101O1N2O1O0O2O1O1O1N2O001N2O1XGZFf7g9WH`Fc7a9[HfF`7[9]HmF]7T9`HSG[7n8bHYGX7i8eH]GW7c8gH`GX7a8fHaGX7a8eHcGY7^8dHeG[7\\8bHhG[7W:O2N1N2O2N1O2M2O1O2M2O2N1N3N1O1O0O2O001N101O0O2O001O0O2O001N101O0O2O001N101O001N101O0O2O001N101O001N10O10O10O10O10O10O10O1000O0100O01000O01000O10O10O10O10000O1000000O10000O10000O10002M3N2N2N2M3N2N2M3N2N2M3N2N2N2M3N1O1N3N1O1N3N1O1N3N1O1O2M2O1O2M2O1O2M2O1O1O2M2O1O2M2O1O2M2O1O2N1N2O2N1N2O2N1N2O1O1N2O1O1O2M2O1O1N2O1O1N2O2N1O1N2O1O1N2O1O1N3N1O1O1N2O1O1N2O2N1N2O1O6I;F;EloU5"}, "label": "the side of the car visible in the mirror"}]} {"id": 112122, "image": "COCO_train2014_000000112122.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the reflection of a car with a dog hanging out of a window\"."}], "task": "refering", "answer_template": "The \"the reflection of a car with a dog hanging out of a window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [140, 141, 142, 143, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 394, 395, 396, 397, 398, 399, 400, 419], "bbox": [0.052593749999999995, 0.29411575562700965, 0.5716875, 0.8282315112540194], "iscrowd": 0, "area": 65990.40814999997, "rle": {"size": [622, 640], "counts": "and07Uc06I7J7H7I8I6I7J7H7I8I6I7I8I6I8I6I7I8I6I7J7H5K6K4K6J5I7I8I6I7J7H7I8I6I7J7H7M4L3N2N3M3M3L4M3M3M2M4M3M3M3M3L4M2N1O1N3N1O2N1O1N10001O0000001O0000001O000000001O0000001O000000001O0000001O0000001O000000001O0000001O000000001O0000001O0000000010O1O1O1O1N101O1N2O1O0O2O1O1O1N2O001N2O1XGZFf7g9WH`Fc7a9[HfF`7[9]HmF]7T9`HSG[7n8bHYGX7i8eH]GW7c8gH`GX7a8fHaGX7a8eHcGY7^8dHeG[7\\8bHhG[7W:O2N1N2O2N1O2M2O1O2M2O2N1N3N1O1O0O2O001N101O0O2O001O0O2O001N101O0O2O001N101O001N101O0O2O001N101O001N10O10O10O10O10O10O10O1000O0100O01000O01000O10O10O10O10000O1000000O10000O10000O10002M3N2N2N2M3N2N2M3N2N2M3N2N2N2M3N1O1N3N1O1N3N1O1N3N1O1O2M2O1O2M2O1O2M2O1O1O2M2O1O2M2O1O2M2O1O2N1N2O2N1N2O2N1N2O1O1N2O1O1O2M2O1O1N2O1O1N2O2N1O1N2O1O1N2O1O1N3N1O1O1N2O1O1N2O2N1N2O1O6I;F;EloU5"}, "label": "the reflection of a car with a dog hanging out of a window"}]} {"id": 538108, "image": "COCO_train2014_000000538108.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a flower pot bottle near by the keyboard and mug\"."}], "task": "refering", "answer_template": "The \"a flower pot bottle near by the keyboard and mug\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [224, 225, 226, 227, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 318, 319, 320, 321, 343, 344], "bbox": [0.7483906250000001, 0.5333333333333333, 1.0, 0.8602083333333332], "iscrowd": 0, "area": 16273.625649999996, "rle": {"size": [480, 640], "counts": "SjP72m>5K5K5K5K5@YO_Bl0\\=b0I6K1N3M3N2M3N1N3N2M3M2O2M3N1N3N2M2O0O2N2O1N2O0O2O1N2O0O2N2O1N101O1O1O0000000000000000000000000000000000002O0O101N100O0010O00100O010O1O10O0100O00100O1O010O10O01O100O00100O010O1O010O1O10O0100O0010O001O00001O00001O010O00001O00001O001O01O100O1O10O0100O1O10O0100O1O10O010000O100O01000O1000O0100O1O1O001O1O1O1H8@bF"}, "label": "a flower pot bottle near by the keyboard and mug"}]} {"id": 521726, "image": "COCO_train2014_000000521726.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the traffic signal has 3 colours of light\"."}], "task": "refering", "answer_template": "The \"the traffic signal has 3 colours of light\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 90, 91, 92, 93, 94, 106, 107, 108, 109, 110, 111, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 157, 158, 159, 160, 161, 162, 163, 174, 175, 176, 177, 178, 179, 180, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 225, 226, 227, 228, 229, 230, 231, 244, 245, 246, 247, 248, 261, 262, 263, 264, 265, 279, 280], "bbox": [0.26966666666666667, 0.191015625, 0.6412083333333333, 0.72809375], "iscrowd": 0, "area": 44008.49109999999, "rle": {"size": [640, 480], "counts": "Xl`21mc05j^ONV>7dAMY>8aALW in this image.", "objects": [{"patches": [73, 74, 75, 76, 77, 90, 91, 92, 93, 94, 106, 107, 108, 109, 110, 111, 124, 125, 126, 127, 128, 141, 142, 143, 144, 145, 157, 158, 159, 160, 161, 162, 163, 174, 175, 176, 177, 178, 179, 180, 192, 193, 194, 195, 196, 197, 208, 209, 210, 211, 212, 213, 214, 225, 226, 227, 228, 229, 230, 231, 244, 245, 246, 247, 248, 261, 262, 263, 264, 265, 279, 280], "bbox": [0.26966666666666667, 0.191015625, 0.6412083333333333, 0.72809375], "iscrowd": 0, "area": 44008.49109999999, "rle": {"size": [640, 480], "counts": "Xl`21mc05j^ONV>7dAMY>8aALW in this image.", "objects": [{"patches": [133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 318, 319, 320, 321, 341, 342, 343], "bbox": [0.62546875, 0.3490654205607477, 0.9997343750000001, 1.0], "iscrowd": 0, "area": 37635.499200000006, "rle": {"size": [428, 640], "counts": "g]W51V=6K6M2O2M3N1N3fGYOk3h0TLYOj3j0SLWOl3l0QLVOn3k0nKXOQ4k0SKOm42WJh0i5YO]I`1d6_NQIm1n6SNQIn1P7RNoHn1Q7RNnHo1R7QNnHP2R7PNlHQ2T7oMlHR2T7nMkHR2U7nMkHR2V7mMjHT2U7mMjHS2W7lMjHT2U7kMlHU2U7jMlHU2T7kMlHV2S7jMnHU2S7jMmHV2S7jMnHV2R7hMoHX2Q7fMRIZ2n6eMRI[2n6fMRIY2o6gMPIZ2o6fMRIY2o6hMoHY2P7hMPIW2P7jMPIU2Q7lMmHU2R7lMnHS2S7mMnHR2Q7oMQIn1P7SNQIj1o6XNRIf1n6YNUId1k6\\NXIb1h6^NZI_1f6aN\\I]1d6dN\\I\\1d6cN]I\\1c6dN^I[1c6eN^IY1b6gN_IX1b6hN^IX1a6hN`IW1a6hNaIV1_6kNaIT1_6lNbIS1^6mNdIR1\\6mNeIR1[6oNeIP1[6QOfIm0Z6TOfIk0Z6UOgIk0X6VOiIh0X6XOhIg0X6ZOhIe0X6]OhIa0X6AgI?X6BhI=X6EhI9Y6GgI8Y6IgI6Y6LfI4Y6MhI1X60hIOX63gILY65gIJY67gIIX69hIEX6hIAX6a0fI_OZ6b0fI^OY6c0gI\\OY6f0eIZO[6g0eIXO[6i0eIVO[6k0eIUOZ6m0dISO\\6n0dIRO[6o0eIQOZ6P1fIPOY6Q1gIoNX6R1hInNW6S1hInNW6S1iImNV6T1jIlNU6T1lIlNS6U1mIkNS6U1mIkNR6V1mIkNS6T1nImNQ6S1oImNP6T1PJlNP6S1QJmNn5T1RJlNn5T1QJmNn5S1SJmNm5S1SJnNk5S1UJmNk5R1VJnNj5R1VJnNi5S1VJnNj5Q1WJoNh5R1XJnNh5R1XJnNg5S1YJnNf5R1ZJnNf5Q1ZJPOe5Q1[JoNe5Q1[JoNd5R1\\JnNd5R1\\JnNc5S1]JmNc5S1\\JoNb5R1^JnNb5R1^JnNb5R1^JnNa5S1_JmNa5S1_JmNa5R1_JoN`5R1`JoN_5Q1aJoN_5Q1aJoN_5R1`JnN`5S1^JnNf0]N]2f2lLmN9kNk2Y2fLQO2SOW3o1`LSOK\\Oe3b1ZLXOCCS4V1SL]O\\OKa4j0lKm1U4VNcKc1i4`NoJ\\1Z5R38H9G9F:G9G9I^KQJT2o5bM\\J^2b5ZMhJg2V5oLUKQ3i4fLbKZ3]4]LmKc3Q4TLZLl3e3kKeLU4Y3bKRM^4n50000000000001O000000000000000O1O1O100O2N1O1O100O1O2N1O101N1O1O2N100O2N1O2N100O2N1O2N101N2N1O2N2O0O2N2N2N2M3N2N2N2hKcI]1`6`NTJl0o5ROdJ:_5DUKGn47eKTO^4k0UL^NQ4`1o3M3M3M3N2M4K6I7J5I8G in this image.", "objects": [{"patches": [133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 221, 222, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 318, 319, 320, 321, 341, 342, 343], "bbox": [0.62546875, 0.3490654205607477, 0.9997343750000001, 1.0], "iscrowd": 0, "area": 37635.499200000006, "rle": {"size": [428, 640], "counts": "g]W51V=6K6M2O2M3N1N3fGYOk3h0TLYOj3j0SLWOl3l0QLVOn3k0nKXOQ4k0SKOm42WJh0i5YO]I`1d6_NQIm1n6SNQIn1P7RNoHn1Q7RNnHo1R7QNnHP2R7PNlHQ2T7oMlHR2T7nMkHR2U7nMkHR2V7mMjHT2U7mMjHS2W7lMjHT2U7kMlHU2U7jMlHU2T7kMlHV2S7jMnHU2S7jMmHV2S7jMnHV2R7hMoHX2Q7fMRIZ2n6eMRI[2n6fMRIY2o6gMPIZ2o6fMRIY2o6hMoHY2P7hMPIW2P7jMPIU2Q7lMmHU2R7lMnHS2S7mMnHR2Q7oMQIn1P7SNQIj1o6XNRIf1n6YNUId1k6\\NXIb1h6^NZI_1f6aN\\I]1d6dN\\I\\1d6cN]I\\1c6dN^I[1c6eN^IY1b6gN_IX1b6hN^IX1a6hN`IW1a6hNaIV1_6kNaIT1_6lNbIS1^6mNdIR1\\6mNeIR1[6oNeIP1[6QOfIm0Z6TOfIk0Z6UOgIk0X6VOiIh0X6XOhIg0X6ZOhIe0X6]OhIa0X6AgI?X6BhI=X6EhI9Y6GgI8Y6IgI6Y6LfI4Y6MhI1X60hIOX63gILY65gIJY67gIIX69hIEX6hIAX6a0fI_OZ6b0fI^OY6c0gI\\OY6f0eIZO[6g0eIXO[6i0eIVO[6k0eIUOZ6m0dISO\\6n0dIRO[6o0eIQOZ6P1fIPOY6Q1gIoNX6R1hInNW6S1hInNW6S1iImNV6T1jIlNU6T1lIlNS6U1mIkNS6U1mIkNR6V1mIkNS6T1nImNQ6S1oImNP6T1PJlNP6S1QJmNn5T1RJlNn5T1QJmNn5S1SJmNm5S1SJnNk5S1UJmNk5R1VJnNj5R1VJnNi5S1VJnNj5Q1WJoNh5R1XJnNh5R1XJnNg5S1YJnNf5R1ZJnNf5Q1ZJPOe5Q1[JoNe5Q1[JoNd5R1\\JnNd5R1\\JnNc5S1]JmNc5S1\\JoNb5R1^JnNb5R1^JnNb5R1^JnNa5S1_JmNa5S1_JmNa5R1_JoN`5R1`JoN_5Q1aJoN_5Q1aJoN_5R1`JnN`5S1^JnNf0]N]2f2lLmN9kNk2Y2fLQO2SOW3o1`LSOK\\Oe3b1ZLXOCCS4V1SL]O\\OKa4j0lKm1U4VNcKc1i4`NoJ\\1Z5R38H9G9F:G9G9I^KQJT2o5bM\\J^2b5ZMhJg2V5oLUKQ3i4fLbKZ3]4]LmKc3Q4TLZLl3e3kKeLU4Y3bKRM^4n50000000000001O000000000000000O1O1O100O2N1O1O100O1O2N1O101N1O1O2N100O2N1O2N100O2N1O2N101N2N1O2N2O0O2N2N2N2M3N2N2N2hKcI]1`6`NTJl0o5ROdJ:_5DUKGn47eKTO^4k0UL^NQ4`1o3M3M3M3N2M4K6I7J5I8G in this image.", "objects": [{"patches": [58, 59, 60, 79, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 243, 244, 266, 267, 289, 290, 312, 313], "bbox": [0.39825, 0.1258411214953271, 0.7018125, 0.9325934579439253], "iscrowd": 0, "area": 30452.35279999998, "rle": {"size": [428, 640], "counts": "RjZ37n<:J6M3L3M4M3L4L3N3L4L4L3N3L4L100O1O100O1O1O100O1O100O2N1WOkMZFV2_9SN_Fn1Y9[NdFf1e8ROYGo0a8YO]Gh0_8\\O_Gd0^8B_G=`8G_G8^8\\2O1O1N2O1O1N2O3M2M4M2M4M2O2M2O2M2O2M201OJ5K6J60O101O1N101O0O2O1O0O2O001N01O1N1O2N101N2O010O10O10O10O0100O0100O001O0O2O0gH^Ji6c5TI`Jj6b5TIaJj6_5TIeJj6^5QIeJn6n5M3N1O2M3N1O2M101O0PLkIW1h5jN`Jo0_5QOiJg0V5ZOQK`0m4ASK?m4ASK?m4ASK`0m4_OSKa0o4\\ORKd0o4[OQKf0o4YOPKh0R5VOnJj0S5UOmJl0T5ROlJn0[5kNeJV1d5`N\\J`1k5YNUJg1R6PNPJQ2S6iMoIW2T6dMmI]2W6]MkId2W6WMkIi2Y6QMfIR3`6fL^I_3S84M4K4L4M4K4L4[MfKSL`4c3nKTLV4c3XLTLm3b3`LVLd3b3iLTL]3i3jLnKZ3P4lLhKY3U4nLaKX3\\4oL[KU3c4l2M3M2N3M3M3L3N3O1N101N2N102M3N2M2O2M3N1N3M3N2M2O2M3N2M3N2M3M3N2M3N6I7J7H8I6I4L5B=J7I6KTe_2"}, "label": "a man with a blue shirt with two dogs"}]} {"id": 46592, "image": "COCO_train2014_000000046592.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a blue shirt with his two dogs in front of him\"."}], "task": "refering", "answer_template": "The \"a man with a blue shirt with his two dogs in front of him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 79, 80, 81, 82, 83, 84, 102, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 243, 244, 266, 267, 289, 290, 312, 313], "bbox": [0.39825, 0.1258411214953271, 0.7018125, 0.9325934579439253], "iscrowd": 0, "area": 30452.35279999998, "rle": {"size": [428, 640], "counts": "RjZ37n<:J6M3L3M4M3L4L3N3L4L4L3N3L4L100O1O100O1O1O100O1O100O2N1WOkMZFV2_9SN_Fn1Y9[NdFf1e8ROYGo0a8YO]Gh0_8\\O_Gd0^8B_G=`8G_G8^8\\2O1O1N2O1O1N2O3M2M4M2M4M2O2M2O2M2O2M201OJ5K6J60O101O1N101O0O2O1O0O2O001N01O1N1O2N101N2O010O10O10O10O0100O0100O001O0O2O0gH^Ji6c5TI`Jj6b5TIaJj6_5TIeJj6^5QIeJn6n5M3N1O2M3N1O2M101O0PLkIW1h5jN`Jo0_5QOiJg0V5ZOQK`0m4ASK?m4ASK?m4ASK`0m4_OSKa0o4\\ORKd0o4[OQKf0o4YOPKh0R5VOnJj0S5UOmJl0T5ROlJn0[5kNeJV1d5`N\\J`1k5YNUJg1R6PNPJQ2S6iMoIW2T6dMmI]2W6]MkId2W6WMkIi2Y6QMfIR3`6fL^I_3S84M4K4L4M4K4L4[MfKSL`4c3nKTLV4c3XLTLm3b3`LVLd3b3iLTL]3i3jLnKZ3P4lLhKY3U4nLaKX3\\4oL[KU3c4l2M3M2N3M3M3L3N3O1N101N2N102M3N2M2O2M3N1N3M3N2M2O2M3N2M3N2M3M3N2M3N6I7J7H8I6I4L5B=J7I6KTe_2"}, "label": "a man with a blue shirt with his two dogs in front of him"}]} {"id": 46592, "image": "COCO_train2014_000000046592.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the corgi behind the other corgi\"."}], "task": "refering", "answer_template": "The \"the corgi behind the other corgi\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [214, 215, 216, 237, 238, 239, 240, 243, 266, 289], "bbox": [0.306015625, 0.6393691588785047, 0.6098281249999999, 0.8344626168224298], "iscrowd": 0, "area": 2893.8164500000007, "rle": {"size": [428, 640], "counts": "bVb22Y=2N3M2O2M3M1O2N10O01O000010O01O0100O100O1O010O100O1000000001N10001O00001O000O2O00001O0000001O000010O0010O00010O00010O00010O01O01O01O00010O00010O0010O00010O00010O01O00010O00001O01O01O0010O0001O101N2N2N2N2N2N2O1N2NgZj0n0YdUO3N3L4L3M4L3N20O1O1O1O100O1O1O=C6J2N2M3K6K4K5LTUX3"}, "label": "the corgi behind the other corgi"}]} {"id": 46592, "image": "COCO_train2014_000000046592.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the dog closest to the man\"."}], "task": "refering", "answer_template": "The \"the dog closest to the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [214, 215, 216, 237, 238, 239, 240, 243, 266, 289], "bbox": [0.306015625, 0.6393691588785047, 0.6098281249999999, 0.8344626168224298], "iscrowd": 0, "area": 2893.8164500000007, "rle": {"size": [428, 640], "counts": "bVb22Y=2N3M2O2M3M1O2N10O01O000010O01O0100O100O1O010O100O1000000001N10001O00001O000O2O00001O0000001O000010O0010O00010O00010O00010O01O01O01O00010O00010O0010O00010O00010O01O00010O00001O01O01O0010O0001O101N2N2N2N2N2N2O1N2NgZj0n0YdUO3N3L4L3M4L3N20O1O1O1O100O1O1O=C6J2N2M3K6K4K5LTUX3"}, "label": "the dog closest to the man"}]} {"id": 46592, "image": "COCO_train2014_000000046592.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty chair that is to the left of the main sitting down\"."}], "task": "refering", "answer_template": "The \"an empty chair that is to the left of the main sitting down\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 99, 100, 101, 102, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 212, 213, 214, 215, 235, 236, 237], "bbox": [0.214203125, 0.2513084112149533, 0.46498437500000006, 0.7382710280373832], "iscrowd": 0, "area": 19522.794150000005, "rle": {"size": [428, 640], "counts": "R^i1a0j<;F:E:F;F:E;F:E:G:F:F7H3N1O2N1N3N2N1O0O10O1O1O001O1O001O1O010O1O010O10O01O10O0100O010O10O0100O01000O010O1000O01000O01000O010N1N2N3N1N2O10000O10000000001N101O001O1O001N101O001O100O100O100O100O10000O100O101N100O100O10000OXOhG^LX8_3`HlK`7R4m0N3M3O10O010000O10O10O1N2N1O2O1N2N101N2N200O100O2O0O100O2O0O100O2O000O2O1SNQGZOQ9b0TG[On8a0WG[Ok8c0ZGYOh8c0]GYOf8c0T2M3L3M4M3LnW_4"}, "label": "an empty chair that is to the left of the main sitting down"}]} {"id": 46592, "image": "COCO_train2014_000000046592.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black chair to the left of the dog\"."}], "task": "refering", "answer_template": "The \"a black chair to the left of the dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 99, 100, 101, 102, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 212, 213, 214, 215, 235, 236, 237], "bbox": [0.214203125, 0.2513084112149533, 0.46498437500000006, 0.7382710280373832], "iscrowd": 0, "area": 19522.794150000005, "rle": {"size": [428, 640], "counts": "R^i1a0j<;F:E:F;F:E;F:E:G:F:F7H3N1O2N1N3N2N1O0O10O1O1O001O1O001O1O010O1O010O10O01O10O0100O010O10O0100O01000O010O1000O01000O01000O010N1N2N3N1N2O10000O10000000001N101O001O1O001N101O001O100O100O100O100O10000O100O101N100O100O10000OXOhG^LX8_3`HlK`7R4m0N3M3O10O010000O10O10O1N2N1O2O1N2N101N2N200O100O2O0O100O2O0O100O2O000O2O1SNQGZOQ9b0TG[On8a0WG[Ok8c0ZGYOh8c0]GYOf8c0T2M3L3M4M3LnW_4"}, "label": "a black chair to the left of the dog"}]} {"id": 226817, "image": "COCO_train2014_000000226817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an asian boy in a green long sleeved shirt playing wii\"."}], "task": "refering", "answer_template": "The \"an asian boy in a green long sleeved shirt playing wii\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 109, 110, 111, 112, 132, 133, 134, 155, 156, 157, 158, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 251, 270, 271, 272, 273, 274, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344], "bbox": [0.7095, 0.19545454545454544, 0.9920781249999999, 0.9993240093240094], "iscrowd": 0, "area": 33063.977699999996, "rle": {"size": [429, 640], "counts": "h^n53V=6H7J7H8I6]OlNeD\\1W;kN_D[1];>K4M4RElM\\:d2M4M3L3M4L3N3M3N101N2N1O2N1O2N2N1SGTLX8n3fGSLY8P4eGQLZ8P4dGRLZ8Q4cGQL[8Q4dGPLZ8S4cGnK\\8T4bGnKl6F_J_4bNmKg6MdJf4U5bKhJ_4Q5jKlJX4R5kKlJV4Q5nKlJT4R5oKkJR4S5RLjJP4T5SLiJo3T5ULiJm3T5VLjJl3T5VLkJk3R5WLmJk3Q5VLoJk3P5VLoJk3o4VLPKl3o4TLQKm3m4ULQKm3m4TLQKo3n4QLPKR4n4X2O1N2O1N1O2O1hKRHNAZ2^8fMVHKDW2Y8jMYHJ_O\\2Z8gM]HGZOb2[8dMmH\\2U7`MnH_2T7^MmHb2U7[MlHe2n800000000O10000000000000000001O1O001O001kG\\M_5e2^J^Ma5b2^J`Ma5a2]JaMb5`2\\JbMc5^2\\JcMd5^2ZJdMe5]2ZJdMe5\\2ZJfMe5[2YJfMg5Z2XJhMg5Y2WJiMh5Y2UJhMl5Z2PJhMo5Z2lIiMT6Z2gIiMX6Y2PITM_Og0`7W2cHhN]7e3O001O1O001J5K6I8eNZ101OkI^J^4b5_K`Jc4_5YKdJl4X5QKjJW5o4eJTKc5e4ZJ]Kn5[4oIhKY6Q4cIRLe6g3VI]LR7[3jHhL_7P3\\HUMl7b2RHaMS8Z2kGhM]8P2`GSNk8b1SG`NW9V1fFmN]9P1aFRO`9]22N1O2N1O2N1O2N1O2N1O2N1O1O1O1O1O1O1O1O1N3N1O1O1O1O1O1O1O1O1AYEVNh:i1\\ESNf:k1^EQNc:m1bEoM_:P2`0O1N2J6K5J6K5J6K6J5J6K7Hlg1"}, "label": "an asian boy in a green long sleeved shirt playing wii"}]} {"id": 226817, "image": "COCO_train2014_000000226817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"oriental boy in green shirt\"."}], "task": "refering", "answer_template": "The \"oriental boy in green shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 88, 89, 109, 110, 111, 112, 132, 133, 134, 155, 156, 157, 158, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 246, 247, 248, 249, 250, 251, 270, 271, 272, 273, 274, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344], "bbox": [0.7095, 0.19545454545454544, 0.9920781249999999, 0.9993240093240094], "iscrowd": 0, "area": 33063.977699999996, "rle": {"size": [429, 640], "counts": "h^n53V=6H7J7H8I6]OlNeD\\1W;kN_D[1];>K4M4RElM\\:d2M4M3L3M4L3N3M3N101N2N1O2N1O2N2N1SGTLX8n3fGSLY8P4eGQLZ8P4dGRLZ8Q4cGQL[8Q4dGPLZ8S4cGnK\\8T4bGnKl6F_J_4bNmKg6MdJf4U5bKhJ_4Q5jKlJX4R5kKlJV4Q5nKlJT4R5oKkJR4S5RLjJP4T5SLiJo3T5ULiJm3T5VLjJl3T5VLkJk3R5WLmJk3Q5VLoJk3P5VLoJk3o4VLPKl3o4TLQKm3m4ULQKm3m4TLQKo3n4QLPKR4n4X2O1N2O1N1O2O1hKRHNAZ2^8fMVHKDW2Y8jMYHJ_O\\2Z8gM]HGZOb2[8dMmH\\2U7`MnH_2T7^MmHb2U7[MlHe2n800000000O10000000000000000001O1O001O001kG\\M_5e2^J^Ma5b2^J`Ma5a2]JaMb5`2\\JbMc5^2\\JcMd5^2ZJdMe5]2ZJdMe5\\2ZJfMe5[2YJfMg5Z2XJhMg5Y2WJiMh5Y2UJhMl5Z2PJhMo5Z2lIiMT6Z2gIiMX6Y2PITM_Og0`7W2cHhN]7e3O001O1O001J5K6I8eNZ101OkI^J^4b5_K`Jc4_5YKdJl4X5QKjJW5o4eJTKc5e4ZJ]Kn5[4oIhKY6Q4cIRLe6g3VI]LR7[3jHhL_7P3\\HUMl7b2RHaMS8Z2kGhM]8P2`GSNk8b1SG`NW9V1fFmN]9P1aFRO`9]22N1O2N1O2N1O2N1O2N1O2N1O1O1O1O1O1O1O1O1N3N1O1O1O1O1O1O1O1O1AYEVNh:i1\\ESNf:k1^EQNc:m1bEoM_:P2`0O1N2J6K5J6K5J6K6J5J6K7Hlg1"}, "label": "oriental boy in green shirt"}]} {"id": 226817, "image": "COCO_train2014_000000226817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"male child in a short sleeved yellow shirt with a collar\"."}], "task": "refering", "answer_template": "The \"male child in a short sleeved yellow shirt with a collar\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 56, 57, 58, 78, 79, 80, 81, 102, 103, 104, 124, 125, 126, 127, 146, 147, 148, 149, 150, 151, 169, 170, 171, 172, 173, 174, 192, 193, 194, 195, 196, 197, 214, 215, 216, 217, 218, 219, 220, 238, 239, 240, 241, 242, 243, 244, 261, 262, 263, 264, 265, 266, 267, 268, 285, 286, 287, 288, 289, 290, 291, 308, 309, 310, 311, 312, 313, 314, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.332890625, 0.11841491841491841, 0.6672968750000001, 0.9835897435897437], "iscrowd": 0, "area": 44672.17555, "rle": {"size": [429, 640], "counts": "l_i2?l<2N2O0O2N2N2O0O2N2N2N1WOj0WOi0L4O1N101N2O1N103L4M3L4M2N3M3Ma0_O`0@2N2N1O2N2O1N2N1O2N2N2N2N1O2O1N2N1O2N2N2N2N1O5L;Db7FXH=d1hMd3o1dJ;c1jMh3P2`J8c1lMl3Q2\\J5d1mMP4i4kK[KT4f4hK^KX4c4cKbK^4]4]KgKe4X4WKkKl4X4jJlKX5]4YJgKi5n52N2N3M2N2N2N2N2N3M2N2N2N2N3M2M3N2N2N2N3M11M7I8H7I7I8H7I7I8H7I4L3M3M3M2N3M3M3M3M3M2N3M3M3M3M3M2N3M2N1O0000001O000O101O00002N2N3M2N2N2N3M2N2N2iN`E]Oc:D^F7a;K5K5K5K6Jajh2"}, "label": "male child in a short sleeved yellow shirt with a collar"}]} {"id": 226817, "image": "COCO_train2014_000000226817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby with brown hair and wearing all yellow and holding a white controller\"."}], "task": "refering", "answer_template": "The \"baby with brown hair and wearing all yellow and holding a white controller\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 71, 92, 93, 94, 95, 115, 116, 117, 138, 139, 140, 141, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 259, 260], "bbox": [0.003046875, 0.19955710955710956, 0.376953125, 0.7414918414918416], "iscrowd": 0, "area": 19877.411750000003, "rle": {"size": [429, 640], "counts": "`P1?m in this image.", "objects": [{"patches": [69, 70, 71, 92, 93, 94, 95, 115, 116, 117, 138, 139, 140, 141, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 209, 210, 211, 212, 213, 214, 215, 232, 233, 234, 235, 236, 237, 238, 259, 260], "bbox": [0.003046875, 0.19955710955710956, 0.376953125, 0.7414918414918416], "iscrowd": 0, "area": 19877.411750000003, "rle": {"size": [429, 640], "counts": "`P1?m in this image.", "objects": [{"patches": [50, 51, 73, 74, 96, 97, 98, 119, 120, 121, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 211, 212, 213, 234, 235, 236, 257, 258, 280], "bbox": [0.17459375, 0.11965973534971644, 0.31576562499999994, 0.6837051039697543], "iscrowd": 0, "area": 16581.841050000006, "rle": {"size": [529, 640], "counts": "WUj1`0^?g0XOh0XOh0YOg0XOh0YOg0QOo0oNQ1nNP1SOj0O1O001N2O001O1O02O2N3M2M3N3M2M3N2N3M2M3N3M2M3N2N3M2M3N3M2M3N2N3M2M3N3M2M3N2N3L3N2N3M2M3N2N3L3N2M3L5K4M3L5K4L4L4M4K4L4L5K4L4M3L5K4L4L5L3L4L4L5K9G:G:E;E;E;E;F9F]RR7"}, "label": "left most white umbrella"}]} {"id": 382469, "image": "COCO_train2014_000000382469.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white umbrella on the left\"."}], "task": "refering", "answer_template": "The \"the white umbrella on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [50, 51, 73, 74, 96, 97, 98, 119, 120, 121, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 211, 212, 213, 234, 235, 236, 257, 258, 280], "bbox": [0.17459375, 0.11965973534971644, 0.31576562499999994, 0.6837051039697543], "iscrowd": 0, "area": 16581.841050000006, "rle": {"size": [529, 640], "counts": "WUj1`0^?g0XOh0XOh0YOg0XOh0YOg0QOo0oNQ1nNP1SOj0O1O001N2O001O1O02O2N3M2M3N3M2M3N2N3M2M3N3M2M3N2N3M2M3N3M2M3N2N3M2M3N3M2M3N2N3L3N2N3M2M3N2N3L3N2M3L5K4M3L5K4L4L4M4K4L4L5K4L4M3L5K4L4L5L3L4L4L5K9G:G:E;E;E;E;F9F]RR7"}, "label": "the white umbrella on the left"}]} {"id": 382469, "image": "COCO_train2014_000000382469.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a rear facing chair with a blue towel on top\"."}], "task": "refering", "answer_template": "The \"a rear facing chair with a blue towel on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 243, 244, 266, 267, 289, 290, 291, 292, 293, 312, 313, 314, 315, 316, 336, 337, 338, 358, 361, 381, 384, 385, 408], "bbox": [0.5676875, 0.49463137996219286, 0.7614375, 0.91296786389414], "iscrowd": 0, "area": 6597.944699999998, "rle": {"size": [529, 640], "counts": "\\ik51_`03M4L4L4L4L4K4M4L4L4M3N2M2TBbNe20P6]1ZGgNBMm13\\7X1TGnNZOJn15j7Q1mFUOPOKU21R8o0hFDS1]O[8m0aFER1^Ob8m0ZFEQ1_Oi8k0VFEn0@R9j0nEHk0^O\\9i0hEJh0^Oe9f0cEMc0]Oo9f0ZE0c0[OX:c0QE7b0VOb:c0fDj0]AVOa>j0aAVO]>i0fAWOX>i0jAWOS>j0PBUOn=k0TBUOj=k0XBUOf=j0]BVOa=j0aBVO]=j0eBVOY=j0iBVOU=i0nBWOP=i0RCWOl[FBd9=^FCb9:bFE^99dFG[98gFE\\99fFD\\9:gFD[9:_3O2N1Nne^2"}, "label": "a rear facing chair with a blue towel on top"}]} {"id": 382469, "image": "COCO_train2014_000000382469.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair on the right side\"."}], "task": "refering", "answer_template": "The \"chair on the right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 243, 244, 266, 267, 289, 290, 291, 292, 293, 312, 313, 314, 315, 316, 336, 337, 338, 358, 361, 381, 384, 385, 408], "bbox": [0.5676875, 0.49463137996219286, 0.7614375, 0.91296786389414], "iscrowd": 0, "area": 6597.944699999998, "rle": {"size": [529, 640], "counts": "\\ik51_`03M4L4L4L4L4K4M4L4L4M3N2M2TBbNe20P6]1ZGgNBMm13\\7X1TGnNZOJn15j7Q1mFUOPOKU21R8o0hFDS1]O[8m0aFER1^Ob8m0ZFEQ1_Oi8k0VFEn0@R9j0nEHk0^O\\9i0hEJh0^Oe9f0cEMc0]Oo9f0ZE0c0[OX:c0QE7b0VOb:c0fDj0]AVOa>j0aAVO]>i0fAWOX>i0jAWOS>j0PBUOn=k0TBUOj=k0XBUOf=j0]BVOa=j0aBVO]=j0eBVOY=j0iBVOU=i0nBWOP=i0RCWOl[FBd9=^FCb9:bFE^99dFG[98gFE\\99fFD\\9:gFD[9:_3O2N1Nne^2"}, "label": "chair on the right side"}]} {"id": 382469, "image": "COCO_train2014_000000382469.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"umbrella to the right of glasses\"."}], "task": "refering", "answer_template": "The \"umbrella to the right of glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 86, 87, 88, 109, 110, 111, 112, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 247, 248, 249, 270, 271, 272, 294, 295], "bbox": [0.7076875, 0.11391304347826087, 0.8859999999999999, 0.6802079395085066], "iscrowd": 0, "area": 22707.752899999996, "rle": {"size": [529, 640], "counts": "VWZ75V`0:F:F9G:F:F:F9H9J6K5J6J5L5J6J6J5L5J6J5L2M4L4M2M4L4M2M4L4M2M4L4L3N3L4L3N3L4L3N3L4L4M2M4M3M2N3N2M2N3M3M2N3M3M2N3M3M2N3M3M2N3M3M3M2N001O1O1O001O1O001O1O1O1O1O1O001O1O1O1O1O001O12N3K5H8I7H8H8I7H8I7H8I7H8H8I7H8I7H8I`0_Od0]Od0[Od0\\Oe0\\Oc0\\Oe0\\Oc0\\Oe_U1"}, "label": "umbrella to the right of glasses"}]} {"id": 382469, "image": "COCO_train2014_000000382469.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right umbrella\"."}], "task": "refering", "answer_template": "The \"the right umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 86, 87, 88, 109, 110, 111, 112, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 223, 224, 225, 226, 227, 247, 248, 249, 270, 271, 272, 294, 295], "bbox": [0.7076875, 0.11391304347826087, 0.8859999999999999, 0.6802079395085066], "iscrowd": 0, "area": 22707.752899999996, "rle": {"size": [529, 640], "counts": "VWZ75V`0:F:F9G:F:F:F9H9J6K5J6J5L5J6J6J5L5J6J5L2M4L4M2M4L4M2M4L4M2M4L4L3N3L4L3N3L4L3N3L4L4M2M4M3M2N3N2M2N3M3M2N3M3M2N3M3M2N3M3M2N3M3M3M2N001O1O1O001O1O001O1O1O1O1O1O001O1O1O1O1O001O12N3K5H8I7H8H8I7H8I7H8I7H8H8I7H8I7H8I`0_Od0]Od0[Od0\\Oe0\\Oc0\\Oe0\\Oc0\\Oe_U1"}, "label": "the right umbrella"}]} {"id": 382469, "image": "COCO_train2014_000000382469.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white chair facing to the left\"."}], "task": "refering", "answer_template": "The \"a white chair facing to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 238, 239, 260, 261, 262, 281, 282, 283, 284, 304, 305, 306, 307, 328, 329, 330, 331, 351, 353, 354, 373, 374, 377, 400], "bbox": [0.22726562499999997, 0.4959357277882798, 0.41628124999999994, 0.9186389413988658], "iscrowd": 0, "area": 7022.4415500000005, "rle": {"size": [529, 640], "counts": "Wi[24Z`04K4L4M3L4M3N2O100O1O2SCTOS:l0iE]OR:c0jEDS:2mANU>2iANY>S1d0\\O6J2N2N2N2N2N2N2N2N2N2N2N2NO1N2O3M3M3M3L4F:M3N2M3N2M3N2N2M3N2N2N2N2M3N2N2N2N2N2N2NSN_Bn0^=TOgBj0U=WOQCh0h<[O[Cf0_<\\OdCe0V<]OmCd0l;_OWDa0c;BaD>X;DiD?P;EPE?j:CWE`0e:@[Ed0a:\\O`EKVOTOV;Q1eEF@VOf:T1lEAE\\O[:S1RF_OG]OT:T1VFBEZOR:T1YGfNeN4P:V1^GgNbN1o9W1bGjN]NNP:T1iGoNWNKP:Q1oGUOQNIn9n0XHYOiMHn9l0^H]OdMHj9g0hHB^MGh9b0PIFYMJc9<[IIRML`97eIKlL0\\9OPJ0dL0[9I\\J4ZL1Z9HbJ1ZL5Y?M4MbgP6"}, "label": "a white chair facing to the left"}]} {"id": 382469, "image": "COCO_train2014_000000382469.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"chair in the left\"."}], "task": "refering", "answer_template": "The \"chair in the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [215, 216, 238, 239, 260, 261, 262, 281, 282, 283, 284, 304, 305, 306, 307, 328, 329, 330, 331, 351, 353, 354, 373, 374, 377, 400], "bbox": [0.22726562499999997, 0.4959357277882798, 0.41628124999999994, 0.9186389413988658], "iscrowd": 0, "area": 7022.4415500000005, "rle": {"size": [529, 640], "counts": "Wi[24Z`04K4L4M3L4M3N2O100O1O2SCTOS:l0iE]OR:c0jEDS:2mANU>2iANY>S1d0\\O6J2N2N2N2N2N2N2N2N2N2N2N2NO1N2O3M3M3M3L4F:M3N2M3N2M3N2N2M3N2N2N2N2M3N2N2N2N2N2N2NSN_Bn0^=TOgBj0U=WOQCh0h<[O[Cf0_<\\OdCe0V<]OmCd0l;_OWDa0c;BaD>X;DiD?P;EPE?j:CWE`0e:@[Ed0a:\\O`EKVOTOV;Q1eEF@VOf:T1lEAE\\O[:S1RF_OG]OT:T1VFBEZOR:T1YGfNeN4P:V1^GgNbN1o9W1bGjN]NNP:T1iGoNWNKP:Q1oGUOQNIn9n0XHYOiMHn9l0^H]OdMHj9g0hHB^MGh9b0PIFYMJc9<[IIRML`97eIKlL0\\9OPJ0dL0[9I\\J4ZL1Z9HbJ1ZL5Y?M4MbgP6"}, "label": "chair in the left"}]} {"id": 529929, "image": "COCO_train2014_000000529929.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brownie with white top\"."}], "task": "refering", "answer_template": "The \"brownie with white top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 270, 271], "bbox": [0.691015625, 0.3595416666666667, 1.0, 0.6853958333333333], "iscrowd": 0, "area": 20654.031800000004, "rle": {"size": [480, 640], "counts": "U]_61o>3L3M3M3M4M2M3M3^B@_SCFk in this image.", "objects": [{"patches": [155, 156, 157, 158, 159, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 206, 223, 224, 225, 226, 227, 228, 229, 246, 247, 248, 249, 250, 251, 270, 271], "bbox": [0.691015625, 0.3595416666666667, 1.0, 0.6853958333333333], "iscrowd": 0, "area": 20654.031800000004, "rle": {"size": [480, 640], "counts": "U]_61o>3L3M3M3M4M2M3M3^B@_SCFk in this image.", "objects": [{"patches": [340, 341, 342, 343, 344, 345, 346, 347, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.87190625, 0.9647916666666667, 1.0], "iscrowd": 0, "area": 25668.882850000013, "rle": {"size": [640, 480], "counts": "^U12Z24T?MS_Ok1m`0d0001O000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O000000000M3N2N2M4M2N2N7HlP:"}, "label": "the table in front of the guy with the red sweatshirt"}]} {"id": 30231, "image": "COCO_train2014_000000030231.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the edge of a table . part of a pair of pants is visible\"."}], "task": "refering", "answer_template": "The \"the edge of a table . part of a pair of pants is visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [340, 341, 342, 343, 344, 345, 346, 347, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0, 0.87190625, 0.9647916666666667, 1.0], "iscrowd": 0, "area": 25668.882850000013, "rle": {"size": [640, 480], "counts": "^U12Z24T?MS_Ok1m`0d0001O000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O00000000000000001O00000000000000001O000000000M3N2N2M4M2N2N7HlP:"}, "label": "the edge of a table . part of a pair of pants is visible"}]} {"id": 480797, "image": "COCO_train2014_000000480797.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a very large ship called aurora\"."}], "task": "refering", "answer_template": "The \"a very large ship called aurora\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137], "bbox": [0.033625, 0.0013541666666666667, 0.998875, 0.33541666666666664], "iscrowd": 0, "area": 86680.32884999999, "rle": {"size": [480, 640], "counts": "[]:4i>5K4L4L4L4K6K4L4L4L4L5K4K5L4N2O2M2O1N2N2O2M2O1N2N3N1N2O1N2N3N1N100O1O100O100O1O100O100O100O10000O101O000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000O10O10O10O0100O10O0100O10O10000O1000000O1000000O10000O1000000O10000O101O001N10001N101O001N2O1O001N10000O1000000O10000O1000000O10001N1000000O10000O1000000O1000000O2O000O1000000O10000O1000000O10000O2O00000O100000000O10000000000000001O00000000000O1000000000000000001O000000000000000000000000000000001O000000000000000000000000000000001O0000000000000000000000000000001O0000000000000000O1O1N2O1O1O1N2O1O1O1O1000000O10000000000000000O10000000000000000O10000000000001O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O0000000000000000000000O10000000000000000000000000000000000000000000000000000Q>"}, "label": "a very large ship called aurora"}]} {"id": 480797, "image": "COCO_train2014_000000480797.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giant cruise ship\"."}], "task": "refering", "answer_template": "The \"the giant cruise ship\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137], "bbox": [0.033625, 0.0013541666666666667, 0.998875, 0.33541666666666664], "iscrowd": 0, "area": 86680.32884999999, "rle": {"size": [480, 640], "counts": "[]:4i>5K4L4L4L4K6K4L4L4L4L5K4K5L4N2O2M2O1N2N2O2M2O1N2N3N1N2O1N2N3N1N100O1O100O100O1O100O100O100O10000O101O000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000O10O10O10O0100O10O0100O10O10000O1000000O1000000O10000O1000000O10000O101O001N10001N101O001N2O1O001N10000O1000000O10000O1000000O10001N1000000O10000O1000000O1000000O2O000O1000000O10000O1000000O10000O2O00000O100000000O10000000000000001O00000000000O1000000000000000001O000000000000000000000000000000001O000000000000000000000000000000001O0000000000000000000000000000001O0000000000000000O1O1N2O1O1O1N2O1O1O1O1000000O10000000000000000O10000000000000000O10000000000001O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O001O1O1O1O001O1O1O1O0000000000000000000000O10000000000000000000000000000000000000000000000000000Q>"}, "label": "the giant cruise ship"}]} {"id": 136736, "image": "COCO_train2014_000000136736.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"train with green t on it\"."}], "task": "refering", "answer_template": "The \"train with green t on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 276, 277, 278, 279, 280, 281, 282, 283], "bbox": [0.02021875, 0.2561875, 0.559546875, 0.7528125], "iscrowd": 0, "area": 60806.16335, "rle": {"size": [480, 640], "counts": "U[6=a>k0UO]1cN2N2N3M2N2N2N2N3M2N2N2N3M2N2N2N2O2N1O1O1O1O2N1O1O1O1O2O0O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O100O1O1O100O1O1O100O1O010O1O1O100O1O1O100O1O1O100O1O100O1O1O100O001O100O1O100O1O1O100O1O1O100O1O1O100O1O1O010O1O1O100O1O2O1N2N2N2O2M2N2N2O1N2N2N2O1N10000O100O10000O01000O100O10000O10O10O10000O100O10000O01000O1000O0100O01000O10O10O10O01000O010000O01000O010000O10000O1000000O1000000O10O100000O1000000O100000000O1000000O100000000O10O1000O100000000O1000000O100000000O1000000O100000000O2O00000O100000000O1000000O100000000O1000000O100000000O1000001O0000000O100000000000000000000O100000000000001O00000O1gN\\GVLd8g3`GXLa8d3bG\\L^8a3fG^LZ8_3iGaLW8]3lGbLT8[3oGeLQ8X3SHgLn7U3UHkLk7R3YHmLg7P3\\HPMd7n2_HPMb7m2aHSM_7j2eHUM\\7g2gHYMY7d2kH[MU7b2nH^MR7`2QI_Mo6^2TIbMl6Y2ZIeMh6U2]IkMc6P2cInM^6m1gISNY6h1mIVNT6e1QJ[No5`1WJ^Nk5\\1ZJdNf5W1`JgNa5T1dJlN\\5n0kJPOV5k0oJUOQ5f0UKXOm4b0XK^Oh4=^KAc4:bKF^45hKIY42lKNT4OPLOR40nK0R4OPLOQ41oKNR41PLNP42PLMQ42QLMP42PLMQ42_500O1000j`T4"}, "label": "train with green t on it"}]} {"id": 136736, "image": "COCO_train2014_000000136736.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a train with red and blue markings\"."}], "task": "refering", "answer_template": "The \"a train with red and blue markings\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272], "bbox": [0.5078437499999999, 0.279375, 1.0, 0.707625], "iscrowd": 0, "area": 53326.88435000002, "rle": {"size": [480, 640], "counts": "Tch4Q1j=6J5K6K5J6J6J6K5J5K6J6J6K5J6J5K6K5J6J6N2N2M2O4K5L4L4K4M4K5L4M3L4MO010000O100O100O100O100O010O100O1000000O10000000O1000000000O1000000000O10000000O10000000000000O1000O100000000000000000O01000000000000000000O0100000000000000000O1000O100000000000O1000000000O10000000O100000000000O100000O10000000O100000000O010000000O2O0000000O2O0000000O2O00000O10001O000O10001O000O10001O000O10001O000O1000001N100000001N100000001N100000001N100000000O2O00000O101O00000O101O00000O10001O000O10001O000O10001O0O1000001O0O100000001N100000001N100000001N1000000O2O0000000O101O00000O101O00000O101O00000O10001O0O1000001O0O1000001O0O1000001O0O100000001N1000001NlJ"}, "label": "a train with red and blue markings"}]} {"id": 136736, "image": "COCO_train2014_000000136736.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train numbered 8343\"."}], "task": "refering", "answer_template": "The \"the train numbered 8343\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 268, 269, 270, 271, 272], "bbox": [0.5078437499999999, 0.279375, 1.0, 0.707625], "iscrowd": 0, "area": 53326.88435000002, "rle": {"size": [480, 640], "counts": "Tch4Q1j=6J5K6K5J6J6J6K5J5K6J6J6K5J6J5K6K5J6J6N2N2M2O4K5L4L4K4M4K5L4M3L4MO010000O100O100O100O100O010O100O1000000O10000000O1000000000O1000000000O10000000O10000000000000O1000O100000000000000000O01000000000000000000O0100000000000000000O1000O100000000000O1000000000O10000000O100000000000O100000O10000000O100000000O010000000O2O0000000O2O0000000O2O00000O10001O000O10001O000O10001O000O10001O000O1000001N100000001N100000001N100000001N100000000O2O00000O101O00000O101O00000O10001O000O10001O000O10001O0O1000001O0O100000001N100000001N100000001N1000000O2O0000000O101O00000O101O00000O101O00000O10001O0O1000001O0O1000001O0O1000001O0O100000001N1000001NlJ"}, "label": "the train numbered 8343"}]} {"id": 538155, "image": "COCO_train2014_000000538155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is broccoli on a plate of food\"."}], "task": "refering", "answer_template": "The \"there is broccoli on a plate of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 52, 53, 54, 55, 56, 57, 65, 66, 67, 68, 69, 70, 71, 72, 80, 81, 82, 83, 84, 95, 96, 97, 98, 99, 112, 113], "bbox": [0.34929577464788736, 0.10974999999999999, 0.846830985915493, 0.32442187499999997], "iscrowd": 0, "area": 14120.143199999999, "rle": {"size": [640, 426], "counts": "SYm24hc05K4M4K5K5L4O1N2N101N2N2N2O1N1O100O1O100O1O1O101N1O100O1O100O1O1O100O2N101N1O2O1N1O2N1N3N2N1O2N1O2N2N1O2N3L3N2N2N2O1N3M2N2O1N2N3M2O1N1O1O100O1O100000000O1000000000000O1000O10001N100O2O001N100O2O0O101O0O101N1M3M4K4M4L3N2M4M2N2N2N1O2N1O2M2O1O00O00010O01O11100O0100000O01000O10O1000O10O10O1000000O2O000O1000001N2O1O1N2O1O1O1N2O2N1N2O1O001N101O001O001O001O001O1O001O001O010O001O001O1O001O001O1O001O0O2N2N1O2M2O2N2N2N2M3N2N2NaTY1"}, "label": "there is broccoli on a plate of food"}]} {"id": 538155, "image": "COCO_train2014_000000538155.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a broccoli side dish\"."}], "task": "refering", "answer_template": "The \"a broccoli side dish\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 52, 53, 54, 55, 56, 57, 65, 66, 67, 68, 69, 70, 71, 72, 80, 81, 82, 83, 84, 95, 96, 97, 98, 99, 112, 113], "bbox": [0.34929577464788736, 0.10974999999999999, 0.846830985915493, 0.32442187499999997], "iscrowd": 0, "area": 14120.143199999999, "rle": {"size": [640, 426], "counts": "SYm24hc05K4M4K5K5L4O1N2N101N2N2N2O1N1O100O1O100O1O1O101N1O100O1O100O1O1O100O2N101N1O2O1N1O2N1N3N2N1O2N1O2N2N1O2N3L3N2N2N2O1N3M2N2O1N2N3M2O1N1O1O100O1O100000000O1000000000000O1000O10001N100O2O001N100O2O0O101O0O101N1M3M4K4M4L3N2M4M2N2N2N1O2N1O2M2O1O00O00010O01O11100O0100000O01000O10O1000O10O10O1000000O2O000O1000001N2O1O1N2O1O1O1N2O2N1N2O1O001N101O001O001O001O001O1O001O001O010O001O001O1O001O001O1O001O0O2N2N1O2M2O2N2N2N2M3N2N2NaTY1"}, "label": "a broccoli side dish"}]} {"id": 71229, "image": "COCO_train2014_000000071229.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the girl in the dress\"."}], "task": "refering", "answer_template": "The \"the girl in the dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 12, 13, 14, 15, 16, 24, 25, 26, 27, 28, 36, 37, 38, 39, 41, 48, 49, 50, 51, 52, 53, 54, 60, 61, 62, 63, 64, 65, 66, 72, 73, 74, 75, 76, 77, 78, 84, 85, 86, 87, 88, 89, 90, 96, 97, 98, 99, 100, 101, 102, 108, 109, 110, 111, 112, 113, 114, 120, 121, 122, 123, 124, 125, 132, 133, 134, 135, 136, 144, 145, 146, 147, 156, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 174, 180, 181, 182, 183, 184, 185, 186, 187, 192, 193, 194, 195, 196, 197, 198, 199, 200], "bbox": [0.0032926829268292686, 0.014645833333333334, 0.6954573170731708, 0.989875], "iscrowd": 0, "area": 64228.362550000005, "rle": {"size": [480, 328], "counts": "\\`0i5V9g7XH2O1N2O1N2O1O1N2O1O100O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1000000000000000000O1000000000O10000000000000O1000000000000000000000000O100000000000000001O0000001O0000001O00001O001^LJXH6c76VHKe79[HGa7=_HD\\7`0dH@X7d0hH]OU7e0kH]OQ7e0oH]Om6e0SI]Oj6d0VI_Oe6c0[I_Oa6c0_I_O]6c0cI@Y6a0gIAUN^OW6T1dK_ORNJl5j0RL]OoM7_5`0bLZOlMk0m4MWMXOlMl0k4MYMXOkMm0j4K[MYOjMm0i4K]MYOiMn0g4J`MXOiMo0f4IaMYOiMo0d4HbM[OiMn0d4GcM\\OiMn0b4FeM\\OjMn0`4FfM]OiMo0_4EgM^OiMm0_4FgM_OhMm0`4DgM@hMm0`4DgMAgMm0`4BiMBfMm0`4BiMCeMm0`4AjMCeMn0_4@kMDdMm0`4@kMEcMm0a4]OlMI`Ml0\\1gMn0e15L]Mi04TOW287Z5FgJ9\\5DdJ<^5BcJ=`5_ObJ`0j2eMWMh10b0]2WN_MV14c0U2cNeMh07d0n1nNiM;;f0l1ROhM6=g0k1VOgM1>i0k1XOdM0b0g0k1[O^MOi0e0i1DUMHS1c0h1MkLB]1a0d0W1hMXNf1`0a0V4AiK>W4CiK=V4EiK;V4GiK9W4HhK9V4HjK:S4HlK9R4HnK9P4IoK8o3JPL7n3JRL8k3JTL7j3JVL7g3LXL5f3MYL5d3L\\L5b3M]L4^30bL2R39mLHd0RMbNb3i0]O;o2ESM0W3OjLLZ34gLI]35eLF_39bLCb3<`L@c3?_L\\Oe3c0]LXOg3g0[LUOh3j0YLROl3l0VLoNn3P1R5O1O1O1O1N2O1O1O1O1O1O1O2N1O1O1O1O1O1O001O1O001O1O001O1O001O001O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1OY]^1"}, "label": "the girl in the dress"}]} {"id": 71229, "image": "COCO_train2014_000000071229.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in dress brushing her blonde hair\"."}], "task": "refering", "answer_template": "The \"woman in dress brushing her blonde hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 12, 13, 14, 15, 16, 24, 25, 26, 27, 28, 36, 37, 38, 39, 41, 48, 49, 50, 51, 52, 53, 54, 60, 61, 62, 63, 64, 65, 66, 72, 73, 74, 75, 76, 77, 78, 84, 85, 86, 87, 88, 89, 90, 96, 97, 98, 99, 100, 101, 102, 108, 109, 110, 111, 112, 113, 114, 120, 121, 122, 123, 124, 125, 132, 133, 134, 135, 136, 144, 145, 146, 147, 156, 157, 158, 159, 160, 168, 169, 170, 171, 172, 173, 174, 180, 181, 182, 183, 184, 185, 186, 187, 192, 193, 194, 195, 196, 197, 198, 199, 200], "bbox": [0.0032926829268292686, 0.014645833333333334, 0.6954573170731708, 0.989875], "iscrowd": 0, "area": 64228.362550000005, "rle": {"size": [480, 328], "counts": "\\`0i5V9g7XH2O1N2O1N2O1O1N2O1O100O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1000000000000000000O1000000000O10000000000000O1000000000000000000000000O100000000000000001O0000001O0000001O00001O001^LJXH6c76VHKe79[HGa7=_HD\\7`0dH@X7d0hH]OU7e0kH]OQ7e0oH]Om6e0SI]Oj6d0VI_Oe6c0[I_Oa6c0_I_O]6c0cI@Y6a0gIAUN^OW6T1dK_ORNJl5j0RL]OoM7_5`0bLZOlMk0m4MWMXOlMl0k4MYMXOkMm0j4K[MYOjMm0i4K]MYOiMn0g4J`MXOiMo0f4IaMYOiMo0d4HbM[OiMn0d4GcM\\OiMn0b4FeM\\OjMn0`4FfM]OiMo0_4EgM^OiMm0_4FgM_OhMm0`4DgM@hMm0`4DgMAgMm0`4BiMBfMm0`4BiMCeMm0`4AjMCeMn0_4@kMDdMm0`4@kMEcMm0a4]OlMI`Ml0\\1gMn0e15L]Mi04TOW287Z5FgJ9\\5DdJ<^5BcJ=`5_ObJ`0j2eMWMh10b0]2WN_MV14c0U2cNeMh07d0n1nNiM;;f0l1ROhM6=g0k1VOgM1>i0k1XOdM0b0g0k1[O^MOi0e0i1DUMHS1c0h1MkLB]1a0d0W1hMXNf1`0a0V4AiK>W4CiK=V4EiK;V4GiK9W4HhK9V4HjK:S4HlK9R4HnK9P4IoK8o3JPL7n3JRL8k3JTL7j3JVL7g3LXL5f3MYL5d3L\\L5b3M]L4^30bL2R39mLHd0RMbNb3i0]O;o2ESM0W3OjLLZ34gLI]35eLF_39bLCb3<`L@c3?_L\\Oe3c0]LXOg3g0[LUOh3j0YLROl3l0VLoNn3P1R5O1O1O1O1N2O1O1O1O1O1O1O2N1O1O1O1O1O1O001O1O001O1O001O1O001O001O1O001O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1OY]^1"}, "label": "woman in dress brushing her blonde hair"}]} {"id": 71229, "image": "COCO_train2014_000000071229.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman ' s reflection in the mirror\"."}], "task": "refering", "answer_template": "The \"a woman ' s reflection in the mirror\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 67, 68, 69, 78, 79, 80, 81, 82, 90, 91, 92, 93, 102, 103, 104, 105, 114, 115, 116, 117, 126, 127, 128], "bbox": [0.503140243902439, 0.2558125, 0.8484451219512196, 0.6176041666666666], "iscrowd": 0, "area": 14232.985749999998, "rle": {"size": [480, 328], "counts": "oc]25`> in this image.", "objects": [{"patches": [55, 56, 57, 67, 68, 69, 78, 79, 80, 81, 82, 90, 91, 92, 93, 102, 103, 104, 105, 114, 115, 116, 117, 126, 127, 128], "bbox": [0.503140243902439, 0.2558125, 0.8484451219512196, 0.6176041666666666], "iscrowd": 0, "area": 14232.985749999998, "rle": {"size": [480, 328], "counts": "oc]25`> in this image.", "objects": [{"patches": [189, 190, 203, 204, 205, 218, 219, 220, 233, 234, 235, 248, 249, 250, 263, 264, 265, 279, 280, 295], "bbox": [0.5342388758782202, 0.534828125, 0.7412177985948478, 0.8561562500000001], "iscrowd": 0, "area": 9289.376299999998, "rle": {"size": [640, 427], "counts": "Zn^43ic0;I6J6J6J7I6J4LO1O1N2O[NeNg@Z1U?oNg@o0V?ZOe@e0X?Dd@:X?0c@N[?;a@C\\?f0_@YO^?e21O1O10O0100O100O1O010O100O100O002O1N2O1N14M2N3M3M3M3\\AXLjS5O1O1O2M200O100O1O1O1O001O1O1OlKYBo1g=RN^Bi1a=XNcBd1]=oLYBe0=OMb0`=dNYBk0?[Ol0`0`iNe1d0f;]O[BU1R3WOf:^O[B\\1T3nNT;Q1SEeNQ;i1\\38J30J6K5J6J5K6I7I7I7J5J7IQZT2"}, "label": "person using skis and pole wearing a white jacket and a back pack"}]} {"id": 95809, "image": "COCO_train2014_000000095809.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with white overcoat slidding on the snow and two more people following him\"."}], "task": "refering", "answer_template": "The \"a person with white overcoat slidding on the snow and two more people following him\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [189, 190, 203, 204, 205, 218, 219, 220, 233, 234, 235, 248, 249, 250, 263, 264, 265, 279, 280, 295], "bbox": [0.5342388758782202, 0.534828125, 0.7412177985948478, 0.8561562500000001], "iscrowd": 0, "area": 9289.376299999998, "rle": {"size": [640, 427], "counts": "Zn^43ic0;I6J6J6J7I6J4LO1O1N2O[NeNg@Z1U?oNg@o0V?ZOe@e0X?Dd@:X?0c@N[?;a@C\\?f0_@YO^?e21O1O10O0100O100O1O010O100O100O002O1N2O1N14M2N3M3M3M3\\AXLjS5O1O1O2M200O100O1O1O1O001O1O1OlKYBo1g=RN^Bi1a=XNcBd1]=oLYBe0=OMb0`=dNYBk0?[Ol0`0`iNe1d0f;]O[BU1R3WOf:^O[B\\1T3nNT;Q1SEeNQ;i1\\38J30J6K5J6J5K6I7I7I7J5J7IQZT2"}, "label": "a person with white overcoat slidding on the snow and two more people following him"}]} {"id": 366148, "image": "COCO_train2014_000000366148.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the wooden bench on the left - hand side\"."}], "task": "refering", "answer_template": "The \"the wooden bench on the left - hand side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 238, 239], "bbox": [0.003359375, 0.36166666666666664, 0.43959374999999995, 0.6211874999999999], "iscrowd": 0, "area": 21294.6034, "rle": {"size": [480, 640], "counts": "jT1a0X=W1N2O1O001000000000000000000001O0O1000000000000000001O00000O100000000O100000000O100000000O101O00000O12N1O2N2N1O2N1O2N1O2N2N001O001O001O1O010O001O001O1O002N2N1O2N2N1O1O0000001O0000000010O00000001O00000000001O0000003M3M3M3M1O0000O101O0DYD_Mg;a2<0000001O0000000000000000000000000000001O000000000000000000000000000000001O000000000001O000000000000000001O0000000000000000000000000000001O000000000000000000000000000000001O000000000O1N2O1N200000000000001O002N101N25K7I7G8H9F5L000BQCjNoP2O1O001N3M5KTRX5"}, "label": "the wooden bench on the left - hand side"}]} {"id": 366148, "image": "COCO_train2014_000000366148.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden bench facing the ocean\"."}], "task": "refering", "answer_template": "The \"a wooden bench facing the ocean\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 238, 239], "bbox": [0.003359375, 0.36166666666666664, 0.43959374999999995, 0.6211874999999999], "iscrowd": 0, "area": 21294.6034, "rle": {"size": [480, 640], "counts": "jT1a0X=W1N2O1O001000000000000000000001O0O1000000000000000001O00000O100000000O100000000O100000000O101O00000O12N1O2N2N1O2N1O2N1O2N2N001O001O001O1O010O001O001O1O002N2N1O2N2N1O1O0000001O0000000010O00000001O00000000001O0000003M3M3M3M1O0000O101O0DYD_Mg;a2<0000001O0000000000000000000000000000001O000000000000000000000000000000001O000000000001O000000000000000001O0000000000000000000000000000001O000000000000000000000000000000001O000000000O1N2O1N200000000000001O002N101N25K7I7G8H9F5L000BQCjNoP2O1O001N3M5KTRX5"}, "label": "a wooden bench facing the ocean"}]} {"id": 366148, "image": "COCO_train2014_000000366148.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden bench to the right of another wooden bench\"."}], "task": "refering", "answer_template": "The \"a wooden bench to the right of another wooden bench\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 265, 266, 267, 270, 271, 272, 275, 298], "bbox": [0.542703125, 0.37527083333333333, 1.0, 0.7123541666666667], "iscrowd": 0, "area": 32188.459099999996, "rle": {"size": [480, 640], "counts": "djR51o>B>B=C>B>RDUNk9`300000O1000009G000000000000000hNaEoM_:Q2SF]Mm9c2X100O100000000000000000000002O9F8H0mD`Lf:`3ZE`Lf:`3YEbLf:^3ZEcLe:]3ZEdLf:\\3ZEeLe:k30O100O100O100O1YOPE]MQ;]2l00000000000001O00000000000000000001O01O0000000000000000000000001O0000000000000000000000001O00000000000001O00000001O0000000000000000000000001O000000000000000000001O in this image.", "objects": [{"patches": [377, 378, 379, 380, 381, 400, 401, 402, 403, 404, 422, 423, 424, 425, 426, 427, 445, 446, 447, 448, 449, 450, 468, 469, 470, 471, 472, 473, 491, 492, 493, 494, 495, 496], "bbox": [0.367953125, 0.69609375, 0.5959375, 0.937625], "iscrowd": 0, "area": 18851.135099999992, "rle": {"size": [640, 640], "counts": "\\^c41hc07J6I7J6I8H7J6I7J6I8I6I7J6I7I8I6I7J6I7J7H7L4000001O0000000000001O000001O00000000000000000000000000001O01O0000000000000000000000000001O000000000001O0000000000000001O000000000000000000000001O0001O00000000000000000000000000001O000001O000000000000000000000001O0000005K:F:F:F;F9F:F:F;E:F:F:F:F;EaYQ5"}, "label": "a tv underneath the clocl"}]} {"id": 538196, "image": "COCO_train2014_000000538196.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tv\"."}], "task": "refering", "answer_template": "The \"a tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [377, 378, 379, 380, 381, 400, 401, 402, 403, 404, 422, 423, 424, 425, 426, 427, 445, 446, 447, 448, 449, 450, 468, 469, 470, 471, 472, 473, 491, 492, 493, 494, 495, 496], "bbox": [0.367953125, 0.69609375, 0.5959375, 0.937625], "iscrowd": 0, "area": 18851.135099999992, "rle": {"size": [640, 640], "counts": "\\^c41hc07J6I7J6I8H7J6I7J6I8I6I7J6I7I8I6I7J6I7J7H7L4000001O0000000000001O000001O00000000000000000000000000001O01O0000000000000000000000000001O000000000001O0000000000000001O000000000000000000000001O0001O00000000000000000000000000001O000001O000000000000000000000001O0000005K:F:F:F;F9F:F:F;E:F:F:F:F;EaYQ5"}, "label": "a tv"}]} {"id": 538196, "image": "COCO_train2014_000000538196.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a television screen with an image of a man on it\"."}], "task": "refering", "answer_template": "The \"a television screen with an image of a man on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [358, 359, 360, 361, 362, 363, 383, 384, 385, 386, 387, 408, 409, 410, 431, 432, 433, 454, 455, 456, 477, 478, 479, 499, 500, 501, 502, 522, 523, 524, 525], "bbox": [0.5895937499999999, 0.660234375, 0.86934375, 0.99396875], "iscrowd": 0, "area": 15023.423649999997, "rle": {"size": [640, 640], "counts": "WQ\\7?ac000001O000001O0001O0000000010O0000000001O0001O0001O00000001O01O000000001O01O00000001O000001O01O0000000010O0000000001O0001O0001O00000000010O000gA^OT9b0lF^OT9b0kF@T9a0kF_OU9a0jF@W9?hFBX9>hFBX9>gFDX9XOi0WOh0WOj0WO_Rd1"}, "label": "a television screen with an image of a man on it"}]} {"id": 538196, "image": "COCO_train2014_000000538196.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"video displays in a hall\"."}], "task": "refering", "answer_template": "The \"video displays in a hall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [358, 359, 360, 361, 362, 363, 383, 384, 385, 386, 387, 408, 409, 410, 431, 432, 433, 454, 455, 456, 477, 478, 479, 499, 500, 501, 502, 522, 523, 524, 525], "bbox": [0.5895937499999999, 0.660234375, 0.86934375, 0.99396875], "iscrowd": 0, "area": 15023.423649999997, "rle": {"size": [640, 640], "counts": "WQ\\7?ac000001O000001O0001O0000000010O0000000001O0001O0001O00000001O01O000000001O01O00000001O000001O01O0000000010O0000000001O0001O0001O00000000010O000gA^OT9b0lF^OT9b0kF@T9a0kF_OU9a0jF@W9?hFBX9>hFBX9>gFDX9XOi0WOh0WOj0WO_Rd1"}, "label": "video displays in a hall"}]} {"id": 251493, "image": "COCO_train2014_000000251493.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis player in a black shirt\"."}], "task": "refering", "answer_template": "The \"a tennis player in a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 11, 12, 13, 28, 29, 30, 31, 32, 33, 34, 35, 36, 53, 54, 55, 56, 57, 58, 59, 78, 79, 80, 81, 82, 102, 103, 104, 105, 106, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 285, 286, 287, 288, 308, 309, 310, 311, 331, 332], "bbox": [0.19790625, 0.015362997658079624, 0.6926718749999999, 0.990632318501171], "iscrowd": 0, "area": 38472.1094, "rle": {"size": [427, 640], "counts": "Yod12Y=2M2N3N2M2N3M100O2N100O1O2O00001O00010O000010O0000010O01O00010O00010O000010O00010O0001O010O00010O000010O00010O0001O01O01O010O000010O00010O0001O01O01O01O01O0010O00010O0010O00010O01O01O01O01O010O01O01O01O010O00010O00010O010O01O001O010O001O001O00001O010O001O001O001O0_MInG8e75ZHKZ7a0gH_Ol6m0TISO`6Y1`IhNY6^1gIbNW6a1hI`NT6c1lI]NQ6f1oI[Nm5h1SJXNj5l1UJUNg5n1YJRNd5Q2]JPN^5S2bJPNX5S2hJPNR5T2mJnMn4U2RKnMh4U2XKoMV3[N_Li3;WN`2PO^Ll2Q1`Nk1^3UNnLT1]3lNoL>[3CoLH[38PMWOW3h0TMfNS3Z1W48]Oc0[Oe0G9L4K5L5J5L4L4K5L4L4M3M3N3L3M3N2M3M3N3L3M3N^K]Ha3_7`LgH\\3X7cLmHZ3Q7fLUIW3h6iLmId2R6[M[JZ2c5fMcJW2Z5jMkJj1\\5VNiJ]1^5dNdJT1a5mN_Jm0g5SOZJk0f5WOZJg0f5[OZJc0g5_OXJ?h5CXJ;k5DUJV6BiI=Y6BgI>Z6AfI?Z6@gI?[6]OfIe0[6WOgIj0Y6SOhIo0Y6mNhIU1Y6gNiIZ1W6cNjI_1W6^NiId1X6XNjIi1V6TNkIn1f83E;E;O1O1N3N1O2M2O1O2M2O2N1cGeLi6]3SIgLj6[3TIgLk6[3RIhLl6Y3RIiLl6[3QIgLm6[3PIgLo6\\3mHgL`1Ek2h3aKeL_1Lj2c3cKdL]11k2^3eKcL[18i2Z3fKcL[1 in this image.", "objects": [{"patches": [4, 5, 6, 11, 12, 13, 28, 29, 30, 31, 32, 33, 34, 35, 36, 53, 54, 55, 56, 57, 58, 59, 78, 79, 80, 81, 82, 102, 103, 104, 105, 106, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 217, 218, 219, 220, 239, 240, 241, 242, 243, 262, 263, 264, 265, 266, 285, 286, 287, 288, 308, 309, 310, 311, 331, 332], "bbox": [0.19790625, 0.015362997658079624, 0.6926718749999999, 0.990632318501171], "iscrowd": 0, "area": 38472.1094, "rle": {"size": [427, 640], "counts": "Yod12Y=2M2N3N2M2N3M100O2N100O1O2O00001O00010O000010O0000010O01O00010O00010O000010O00010O0001O010O00010O000010O00010O0001O01O01O010O000010O00010O0001O01O01O01O01O0010O00010O0010O00010O01O01O01O01O010O01O01O01O010O00010O00010O010O01O001O010O001O001O00001O010O001O001O001O0_MInG8e75ZHKZ7a0gH_Ol6m0TISO`6Y1`IhNY6^1gIbNW6a1hI`NT6c1lI]NQ6f1oI[Nm5h1SJXNj5l1UJUNg5n1YJRNd5Q2]JPN^5S2bJPNX5S2hJPNR5T2mJnMn4U2RKnMh4U2XKoMV3[N_Li3;WN`2PO^Ll2Q1`Nk1^3UNnLT1]3lNoL>[3CoLH[38PMWOW3h0TMfNS3Z1W48]Oc0[Oe0G9L4K5L5J5L4L4K5L4L4M3M3N3L3M3N2M3M3N3L3M3N^K]Ha3_7`LgH\\3X7cLmHZ3Q7fLUIW3h6iLmId2R6[M[JZ2c5fMcJW2Z5jMkJj1\\5VNiJ]1^5dNdJT1a5mN_Jm0g5SOZJk0f5WOZJg0f5[OZJc0g5_OXJ?h5CXJ;k5DUJV6BiI=Y6BgI>Z6AfI?Z6@gI?[6]OfIe0[6WOgIj0Y6SOhIo0Y6mNhIU1Y6gNiIZ1W6cNjI_1W6^NiId1X6XNjIi1V6TNkIn1f83E;E;O1O1N3N1O2M2O1O2M2O2N1cGeLi6]3SIgLj6[3TIgLk6[3RIhLl6Y3RIiLl6[3QIgLm6[3PIgLo6\\3mHgL`1Ek2h3aKeL_1Lj2c3cKdL]11k2^3eKcL[18i2Z3fKcL[1 in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 38, 39, 40, 41, 42, 43, 44, 61, 62, 63, 64, 65, 66, 67, 84, 85, 86, 87], "bbox": [0.674125, 0.0, 0.93353125, 0.24235416666666668], "iscrowd": 0, "area": 14134.631400000006, "rle": {"size": [480, 640], "counts": "aTZ63k=R1mNS1nNQ1G:00000O10000O10000O1000000O10000O10000O1000000O10000O10000O10000O1000000O10000O101O0O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O1GjChMV in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 21, 38, 39, 40, 41, 42, 43, 44, 61, 62, 63, 64, 65, 66, 67, 84, 85, 86, 87], "bbox": [0.674125, 0.0, 0.93353125, 0.24235416666666668], "iscrowd": 0, "area": 14134.631400000006, "rle": {"size": [480, 640], "counts": "aTZ63k=R1mNS1nNQ1G:00000O10000O10000O1000000O10000O10000O1000000O10000O10000O10000O1000000O10000O101O0O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O1GjChMV in this image.", "objects": [{"patches": [78, 79, 80, 81, 100, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.341046875, 0.24457746478873238, 0.665625, 0.9883802816901409], "iscrowd": 0, "area": 46890.258850000006, "rle": {"size": [426, 640], "counts": "ekj2l0\\<4K5J6K5K5J6K4L5J6K5K5J6L3M2O2M3M2N3N1N3M201O1N101O0O2O001N101O1N101N101O0O2_H`MV4`2hKkMo3V2nKUNCVNa2g3iM^NPOXNW3Z3gMKY26dML[25cMM]24`MN`22^M0a22\\M0d20ZM2f2OWM3h20TM2l2OQM3o2OmL3R3OkL3U3NhL4X3NdL4[3NbL4^3N^L4b3j4O10000O100O10000O101O0O10000000000000000000000000000000000000000000000O100000000000000000000001O00000000001O00000000001O000000001O00000000001O000000001O0000001O1O1O001O1O1O1O1O001O1O1O1O002ULWLQNk3m1WLQNj3m1XLRNj3l1XLRNi3m1XLRNj3k1YLSNh3l1YLSNi3j1ZLTNh3j1YLVNg3h1\\LVNf3h1[LWNf3h1\\LVNf3g1\\LXNe3g1]LWNe3g1\\LXNf3e1]LYNf3d1[L[Nj3fMUKT3T1TOk3ZM\\Ka3j0TO]4k0eKSO\\4k0fKTOZ4l0hKROY4m0hKROY4l0jKROW4m0jKROV4m0mKQOT4n0mKQOS4o0oKoNQ4P1QLoNP4P1RLnNn3R1ULkNU4j0nKTOT4j0oKSOS4j0QLSOQ4k0QLSOQ4YOhI9Z2;o3YOSJ1R2c0l3XO\\J_OZ2V1[3WOQNg0P2UOVNh0l1TOYNi0i1RO\\Nl0f1PO_Nm0R7M2N3M3M3M2N3M3M2N3M3M3Mhch2"}, "label": "a young boy leaning on a display of bananas"}]} {"id": 218734, "image": "COCO_train2014_000000218734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black boy standing beside a bunch of banannas at the market\"."}], "task": "refering", "answer_template": "The \"black boy standing beside a bunch of banannas at the market\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 81, 100, 101, 102, 103, 104, 105, 123, 124, 125, 126, 127, 128, 145, 146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 197, 198, 216, 217, 218, 219, 220, 221, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.341046875, 0.24457746478873238, 0.665625, 0.9883802816901409], "iscrowd": 0, "area": 46890.258850000006, "rle": {"size": [426, 640], "counts": "ekj2l0\\<4K5J6K5K5J6K4L5J6K5K5J6L3M2O2M3M2N3N1N3M201O1N101O0O2O001N101O1N101N101O0O2_H`MV4`2hKkMo3V2nKUNCVNa2g3iM^NPOXNW3Z3gMKY26dML[25cMM]24`MN`22^M0a22\\M0d20ZM2f2OWM3h20TM2l2OQM3o2OmL3R3OkL3U3NhL4X3NdL4[3NbL4^3N^L4b3j4O10000O100O10000O101O0O10000000000000000000000000000000000000000000000O100000000000000000000001O00000000001O00000000001O000000001O00000000001O000000001O0000001O1O1O001O1O1O1O1O001O1O1O1O002ULWLQNk3m1WLQNj3m1XLRNj3l1XLRNi3m1XLRNj3k1YLSNh3l1YLSNi3j1ZLTNh3j1YLVNg3h1\\LVNf3h1[LWNf3h1\\LVNf3g1\\LXNe3g1]LWNe3g1\\LXNf3e1]LYNf3d1[L[Nj3fMUKT3T1TOk3ZM\\Ka3j0TO]4k0eKSO\\4k0fKTOZ4l0hKROY4m0hKROY4l0jKROW4m0jKROV4m0mKQOT4n0mKQOS4o0oKoNQ4P1QLoNP4P1RLnNn3R1ULkNU4j0nKTOT4j0oKSOS4j0QLSOQ4k0QLSOQ4YOhI9Z2;o3YOSJ1R2c0l3XO\\J_OZ2V1[3WOQNg0P2UOVNh0l1TOYNi0i1RO\\Nl0f1PO_Nm0R7M2N3M3M3M2N3M3M2N3M3M3Mhch2"}, "label": "black boy standing beside a bunch of banannas at the market"}]} {"id": 218734, "image": "COCO_train2014_000000218734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an arm with bracelets on it\"."}], "task": "refering", "answer_template": "The \"an arm with bracelets on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 44, 45, 67, 68, 90, 91, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 250, 251, 252, 274, 275, 298, 320, 321, 343, 344], "bbox": [0.8851875, 0.0026056338028169015, 1.0, 0.9870187793427231], "iscrowd": 0, "area": 20191.086700000007, "rle": {"size": [426, 640], "counts": "gR\\7X2Y9dNUGl1j8[1O1N2O1N2O1N101N2O1O1N2O1N2O1N2O1O1N2O1N2O11O3M3M3L3N1O@?^K_KMP52bKmNn4R1eKlMk4S2gKlLh4U3g2O1O0FZFPMg9j2^FVMc9d2bF\\M_9^2fFbMZ9Y2lFfMU9T2PGlMQ9o1SGQNm8j1XGVNk8b1ZG^Ni8Z1\\GfNg8R1^GoNe8h0`GXOc8`0bG@a88dGH_80fG0^8HfG8f:000000000000001O0000000000000000000000L"}, "label": "an arm with bracelets on it"}]} {"id": 218734, "image": "COCO_train2014_000000218734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a womans arm\"."}], "task": "refering", "answer_template": "The \"a womans arm\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 44, 45, 67, 68, 90, 91, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 250, 251, 252, 274, 275, 298, 320, 321, 343, 344], "bbox": [0.8851875, 0.0026056338028169015, 1.0, 0.9870187793427231], "iscrowd": 0, "area": 20191.086700000007, "rle": {"size": [426, 640], "counts": "gR\\7X2Y9dNUGl1j8[1O1N2O1N2O1N101N2O1O1N2O1N2O1N2O1O1N2O1N2O11O3M3M3L3N1O@?^K_KMP52bKmNn4R1eKlMk4S2gKlLh4U3g2O1O0FZFPMg9j2^FVMc9d2bF\\M_9^2fFbMZ9Y2lFfMU9T2PGlMQ9o1SGQNm8j1XGVNk8b1ZG^Ni8Z1\\GfNg8R1^GoNe8h0`GXOc8`0bG@a88dGH_80fG0^8HfG8f:000000000000001O0000000000000000000000L"}, "label": "a womans arm"}]} {"id": 505486, "image": "COCO_train2014_000000505486.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra standing facing right with another zebra walking behind it\"."}], "task": "refering", "answer_template": "The \"zebra standing facing right with another zebra walking behind it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 264, 265, 267, 268, 287], "bbox": [0.477546875, 0.4342916666666667, 0.75590625, 0.7166875], "iscrowd": 0, "area": 10704.44565, "rle": {"size": [480, 640], "counts": "ce_4`0_>2N1N3N1O4hBZOo;i0mCJa;:ZD?n:DnD?P;DlD>T;EhD in this image.", "objects": [{"patches": [175, 176, 177, 195, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 241, 242, 243, 244, 245, 264, 265, 267, 268, 287], "bbox": [0.477546875, 0.4342916666666667, 0.75590625, 0.7166875], "iscrowd": 0, "area": 10704.44565, "rle": {"size": [480, 640], "counts": "ce_4`0_>2N1N3N1O4hBZOo;i0mCJa;:ZD?n:DnD?P;DlD>T;EhD in this image.", "objects": [{"patches": [278, 279, 301, 302, 303, 324, 325, 326, 327, 348, 349, 350, 351, 371, 372, 373, 374, 394, 395, 396, 397, 398, 418, 419, 420, 421, 422, 441, 442, 443, 444, 445, 446, 464, 465, 466, 467, 468, 469, 487, 488, 489, 490, 491, 492, 493, 511, 512, 513, 514, 515, 516], "bbox": [0.10112499999999999, 0.5619531249999999, 0.47640625, 0.984421875], "iscrowd": 0, "area": 28843.6333, "rle": {"size": [640, 640], "counts": "eoX15dc0:G8N3M3M3M3M2N3M3M3M3M2N3M3M3M3M2N3M3M3M3M2N3M3M3M3M2N3M3M3M4L3M3M3M3M3M3M3M4L3M3M3M3M3M3M4L3N2O1N2N2N2O2M2N2O1N2N2O2M2N2O1N2N2N3N1N2N2O1N2N3N1N2N2O1N2N2N3N1N2N2O1N2NO2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O2N1O1O101N1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2M2O1O2N1O1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N1O2N1O2N1O^h`6"}, "label": "a train going through the station"}]} {"id": 431761, "image": "COCO_train2014_000000431761.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a train is parked at the station\"."}], "task": "refering", "answer_template": "The \"a train is parked at the station\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [278, 279, 301, 302, 303, 324, 325, 326, 327, 348, 349, 350, 351, 371, 372, 373, 374, 394, 395, 396, 397, 398, 418, 419, 420, 421, 422, 441, 442, 443, 444, 445, 446, 464, 465, 466, 467, 468, 469, 487, 488, 489, 490, 491, 492, 493, 511, 512, 513, 514, 515, 516], "bbox": [0.10112499999999999, 0.5619531249999999, 0.47640625, 0.984421875], "iscrowd": 0, "area": 28843.6333, "rle": {"size": [640, 640], "counts": "eoX15dc0:G8N3M3M3M3M2N3M3M3M3M2N3M3M3M3M2N3M3M3M3M2N3M3M3M3M2N3M3M3M4L3M3M3M3M3M3M3M4L3M3M3M3M3M3M4L3N2O1N2N2N2O2M2N2O1N2N2O2M2N2O1N2N2N3N1N2N2O1N2N3N1N2N2O1N2N2N3N1N2N2O1N2NO2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O2N1O1O101N1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2M2O1O2N1O1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N1O2N1O2N1O^h`6"}, "label": "a train is parked at the station"}]} {"id": 513683, "image": "COCO_train2014_000000513683.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"plant inside the home\"."}], "task": "refering", "answer_template": "The \"plant inside the home\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 27, 28, 29, 30, 31, 32, 50, 51, 52, 53, 54, 55, 73, 74, 75, 76, 77, 97, 98, 99, 100, 120, 121, 122, 123, 143, 144, 145], "bbox": [0.172296875, 0.0011250000000000001, 0.45270312500000004, 0.42454166666666665], "iscrowd": 0, "area": 19082.550449999995, "rle": {"size": [480, 640], "counts": "adc11l>5K5K4L5L4M2M3N2M3N2L4K5L4O1O001N2O1O1O1O1O0O100010OaDYNe9g1[F^N`9c1^F`N`9`1_FbNa9]1^FfN`9[1VFPOh9P1VFROk9n0RFSOo9m0PFTOP:m0nETOR:n0kESOV:P1fEPOZ:R1cEoN]:R1aEoN`:P1_EQOa:o0_EQOa:o0^EROc:m0\\ETOd:l0[EUOe:l0ZETOf:l0YEUOg:k0XEVOh:k0VEVOj:j0UEWOk:i0UEWOk:j0oD[OQ;Z2O100O101N101N100O2O0O101N2O2M4YOeKcF^4W9hKeF\\4T9lKgFW4U9nKgFV4U9nKgFV4T9o0L1N010000O2O00010O01O10O01O10O01000O10000O10000O10000O10000O10000O10000]Oc0O1N2O1O1N3N1O10000O10000O1000000O100O1O1jNUFVMl9n1]GdMd8[2`GaMa8_2P2O1O100O100O1N2O1O2N1O1O100O100O100O100O1N2O1O1TOTCCm<;\\C]Oe in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 27, 28, 29, 30, 31, 32, 50, 51, 52, 53, 54, 55, 73, 74, 75, 76, 77, 97, 98, 99, 100, 120, 121, 122, 123, 143, 144, 145], "bbox": [0.172296875, 0.0011250000000000001, 0.45270312500000004, 0.42454166666666665], "iscrowd": 0, "area": 19082.550449999995, "rle": {"size": [480, 640], "counts": "adc11l>5K5K4L5L4M2M3N2M3N2L4K5L4O1O001N2O1O1O1O1O0O100010OaDYNe9g1[F^N`9c1^F`N`9`1_FbNa9]1^FfN`9[1VFPOh9P1VFROk9n0RFSOo9m0PFTOP:m0nETOR:n0kESOV:P1fEPOZ:R1cEoN]:R1aEoN`:P1_EQOa:o0_EQOa:o0^EROc:m0\\ETOd:l0[EUOe:l0ZETOf:l0YEUOg:k0XEVOh:k0VEVOj:j0UEWOk:i0UEWOk:j0oD[OQ;Z2O100O101N101N100O2O0O101N2O2M4YOeKcF^4W9hKeF\\4T9lKgFW4U9nKgFV4U9nKgFV4T9o0L1N010000O2O00010O01O10O01O10O01000O10000O10000O10000O10000O10000O10000]Oc0O1N2O1O1N3N1O10000O10000O1000000O100O1O1jNUFVMl9n1]GdMd8[2`GaMa8_2P2O1O100O100O1N2O1O2N1O1O100O100O100O100O1N2O1O1TOTCCm<;\\C]Oe in this image.", "objects": [{"patches": [105, 106, 107, 127, 128, 129, 130, 151, 152, 153, 154, 174, 175, 176, 177, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 311, 312, 313, 314, 315, 316, 317, 334, 335, 336, 337, 338, 339, 340, 357, 358, 359, 360, 361, 362, 363, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.52921875, 0.23820833333333333, 0.8005625000000001, 0.9865208333333333], "iscrowd": 0, "area": 44205.22485, "rle": {"size": [480, 640], "counts": "X[o4>]>9F6J6J6L4N2M3N2M3L4K5L4K5J5I8K5M3M3oHTMj1P3QN]Md1f2WNfM^1^2]NkM\\1X2cNlMV1X2iNmMP1V2nNoMf0Z2YOjM7c2HbMFl29XMZOR3e0SMbLjMb1\\5k1oLaLfMX1d5V2kL`LaMP1l5_2hL^L^Mh0R6i2eL]LZM?Y6S3cL[LUM3d6b3[LYLY4f3lKWLS4i3RLTLl3k3YLSLe3m3g3O100O100O100O1000000O1000000O1000000O10000000000000000000000000000000000001O00001O1O1O1O1O1O1O1O1O1O1O2N2N1O2N1O2N2N1O4L5K5K4L3M4L3M3M3M4L4L5K5K5K3M2N1O2N1O2N1O1O2N1O2N1O2N2N4L3M3M4L4L8H1O1O1O1O1O1O1O001O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O2N3M2N2N2J7I6J6J6J6J7J5`NXFaMm9W2c1I`0_O_fk1"}, "label": "the lady in white shirt"}]} {"id": 513683, "image": "COCO_train2014_000000513683.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman drinking wine\"."}], "task": "refering", "answer_template": "The \"woman drinking wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 107, 127, 128, 129, 130, 151, 152, 153, 154, 174, 175, 176, 177, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 311, 312, 313, 314, 315, 316, 317, 334, 335, 336, 337, 338, 339, 340, 357, 358, 359, 360, 361, 362, 363, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.52921875, 0.23820833333333333, 0.8005625000000001, 0.9865208333333333], "iscrowd": 0, "area": 44205.22485, "rle": {"size": [480, 640], "counts": "X[o4>]>9F6J6J6L4N2M3N2M3L4K5L4K5J5I8K5M3M3oHTMj1P3QN]Md1f2WNfM^1^2]NkM\\1X2cNlMV1X2iNmMP1V2nNoMf0Z2YOjM7c2HbMFl29XMZOR3e0SMbLjMb1\\5k1oLaLfMX1d5V2kL`LaMP1l5_2hL^L^Mh0R6i2eL]LZM?Y6S3cL[LUM3d6b3[LYLY4f3lKWLS4i3RLTLl3k3YLSLe3m3g3O100O100O100O1000000O1000000O1000000O10000000000000000000000000000000000001O00001O1O1O1O1O1O1O1O1O1O1O2N2N1O2N1O2N2N1O4L5K5K4L3M4L3M3M3M4L4L5K5K5K3M2N1O2N1O2N1O1O2N1O2N1O2N2N4L3M3M4L4L8H1O1O1O1O1O1O1O001O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O2N3M2N2N2J7I6J6J6J6J7J5`NXFaMm9W2c1I`0_O_fk1"}, "label": "woman drinking wine"}]} {"id": 513683, "image": "COCO_train2014_000000513683.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wine list at fine dining establishment or lodge of some sort\"."}], "task": "refering", "answer_template": "The \"a wine list at fine dining establishment or lodge of some sort\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.265203125, 0.884, 0.7888437500000001, 1.0], "iscrowd": 0, "area": 15306.10985, "rle": {"size": [480, 640], "counts": "oT`21n>1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O100000000000000000000000000001O00000000000000000000000000001O0000000000000000000000000000001O0000000000000000000000000000001O00000000000000000000000000001O0000000000000000000000000000001O00000000000000000000000000001O0000000000000000000000000000001O00000000000000000000000000000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O10000000000O10O1000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O3N5K6J6J5K6J5K6J5K6JSjn1"}, "label": "a wine list at fine dining establishment or lodge of some sort"}]} {"id": 513683, "image": "COCO_train2014_000000513683.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the menu in front of the man and woman\"."}], "task": "refering", "answer_template": "The \"the menu in front of the man and woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.265203125, 0.884, 0.7888437500000001, 1.0], "iscrowd": 0, "area": 15306.10985, "rle": {"size": [480, 640], "counts": "oT`21n>1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O100000000000000000000000000001O00000000000000000000000000001O0000000000000000000000000000001O0000000000000000000000000000001O00000000000000000000000000001O0000000000000000000000000000001O00000000000000000000000000001O0000000000000000000000000000001O00000000000000000000000000000000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O10000000000O10O1000000000O10000000000O1000000000000O1000000000000O10000000000O1000000000000O1000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O3N5K6J6J5K6J5K6J5K6JSjn1"}, "label": "the menu in front of the man and woman"}]} {"id": 513683, "image": "COCO_train2014_000000513683.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man drinking wine in blue tshirt with obama on it\"."}], "task": "refering", "answer_template": "The \"man drinking wine in blue tshirt with obama on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 77, 78, 79, 80, 81, 100, 101, 102, 103, 104, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 215, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 243, 258, 259, 260, 261, 262, 263, 264, 265, 266, 281, 282, 283, 284, 285, 286, 287, 288, 306, 307, 308, 309, 310, 311, 330, 331, 332, 333], "bbox": [0.23154687499999999, 0.14018750000000002, 0.58221875, 0.8404166666666667], "iscrowd": 0, "area": 50460.81304999999, "rle": {"size": [480, 640], "counts": "_fU25a>?A`0@:E;F:F9N3N2N2N2N2M3N2N2N2N2N2M2O2N2N2N2N2M3N2N2N2N2N2M2O2N2N2N2N2M3N2N2N2N2N2M3N101O10000O10000O10000O10O03N2N3L4M3M3L3N3M3L4M3M2M4M3M3L4M2N3L4M3L3N000O10000O10000O100oMoHSLQ7l3SIPLn6o3WImKi6R4[IkKe6T4_IhKb6X4aIeK^6[4fIbKZ6]4kI^KV6a4nI\\KR6c4RJYKo5f4UJWKk5h4YJUKg5j4]JRKd5m4aJoJ_5P5eJmJ[5S5hJiJY5V5kJgJU5X5oJeJQ5Z5[2O1O1O1O00100O1O1O1O1O1O1O1O1O100O1O1O1O1O001O1O1O100O1O1O1O1O1O11O001O1O001N2O1O001O1O001O1O001O1O001O1O001O1N2O001O1O001O1O001O1O001O1O001O1O1N101O1O001O1O001O1O001O1O001O1O7I6I7J6J7I6J6J6J7I5K4L4L4L4K6C in this image.", "objects": [{"patches": [55, 56, 77, 78, 79, 80, 81, 100, 101, 102, 103, 104, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 190, 191, 192, 193, 194, 195, 196, 197, 212, 213, 214, 215, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 243, 258, 259, 260, 261, 262, 263, 264, 265, 266, 281, 282, 283, 284, 285, 286, 287, 288, 306, 307, 308, 309, 310, 311, 330, 331, 332, 333], "bbox": [0.23154687499999999, 0.14018750000000002, 0.58221875, 0.8404166666666667], "iscrowd": 0, "area": 50460.81304999999, "rle": {"size": [480, 640], "counts": "_fU25a>?A`0@:E;F:F9N3N2N2N2N2M3N2N2N2N2N2M2O2N2N2N2N2M3N2N2N2N2N2M2O2N2N2N2N2M3N2N2N2N2N2M3N101O10000O10000O10000O10O03N2N3L4M3M3L3N3M3L4M3M2M4M3M3L4M2N3L4M3L3N000O10000O10000O100oMoHSLQ7l3SIPLn6o3WImKi6R4[IkKe6T4_IhKb6X4aIeK^6[4fIbKZ6]4kI^KV6a4nI\\KR6c4RJYKo5f4UJWKk5h4YJUKg5j4]JRKd5m4aJoJ_5P5eJmJ[5S5hJiJY5V5kJgJU5X5oJeJQ5Z5[2O1O1O1O00100O1O1O1O1O1O1O1O1O100O1O1O1O1O001O1O1O100O1O1O1O1O1O11O001O1O001N2O1O001O1O001O1O001O1O001O1O001O1N2O001O1O001O1O001O1O001O1O001O1O1N101O1O001O1O001O1O001O1O001O1O7I6I7J6J7I6J6J6J7I5K4L4L4L4K6C in this image.", "objects": [{"patches": [74, 96, 97, 98, 120, 121, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 235, 236, 237, 258, 259, 260, 281, 282, 283, 304, 305, 306, 327, 328, 329, 351, 352], "bbox": [0.17696875, 0.22247916666666667, 0.34045312499999997, 0.9011250000000001], "iscrowd": 0, "area": 19379.9919, "rle": {"size": [480, 640], "counts": "]Ve14`> in this image.", "objects": [{"patches": [74, 96, 97, 98, 120, 121, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 235, 236, 237, 258, 259, 260, 281, 282, 283, 304, 305, 306, 327, 328, 329, 351, 352], "bbox": [0.17696875, 0.22247916666666667, 0.34045312499999997, 0.9011250000000001], "iscrowd": 0, "area": 19379.9919, "rle": {"size": [480, 640], "counts": "]Ve14`> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215], "bbox": [0.00174, 0.0025970149253731344, 1.0, 0.9870149253731343], "iscrowd": 0, "area": 163375.9216, "rle": {"size": [335, 500], "counts": "b=U78R30000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000001O00000000000000000000000000000000000M"}, "label": "an open sandwich"}]} {"id": 226966, "image": "COCO_train2014_000000226966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"open sandwich on a blue plate\"."}], "task": "refering", "answer_template": "The \"open sandwich on a blue plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215], "bbox": [0.00174, 0.0025970149253731344, 1.0, 0.9870149253731343], "iscrowd": 0, "area": 163375.9216, "rle": {"size": [335, 500], "counts": "b=U78R30000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O000000000000000001O00000000000000000000000000000000000M"}, "label": "open sandwich on a blue plate"}]} {"id": 226966, "image": "COCO_train2014_000000226966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue plate\"."}], "task": "refering", "answer_template": "The \"the blue plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 19, 20, 21, 22, 27, 28, 29, 30, 31, 36, 37, 38, 39, 47, 48, 49, 50, 51, 52, 54, 55, 56, 57, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 85, 86, 87, 88, 89, 92, 93, 94, 104, 105, 106, 107, 123, 124, 125, 142, 143, 160, 161, 177, 178, 179, 194, 195, 196, 197, 211, 212, 213, 214, 215], "bbox": [0.00174, 0.06113432835820896, 1.0, 0.986], "iscrowd": 0, "area": 32284.68669999999, "rle": {"size": [335, 500], "counts": "eaNC_1;bNE_19bNG^18cNI]17bNI_16aNK_15`NKa14_NMa12_NO`12_NNb11^N0c1O\\N2d1M\\N3f1KZN6g1IXN8h1GXN9j1FUN;k1DUN=l1ATN?m1ARN`0o1^OQNc0o1]OPNc0R2[OmMg0T2WOlMj0T2VOkMj0W2fMZLe0_1f1W2bM^Lg0Z1g1Z2_M_Lh0W1j1Z2[McLi0S1m1[2WMeLj0P1o1]2SMgLk0m0S2\\2oLkLl0i0V2\\2lLnLl0f0X2[2kLSMk0b0[2[3dMeL]2Z3bMgL^2Z3aMfL`2Y3`MgLa2Y3^MgLb2Z3]MfLd2Y3[MhLf2X3YMhLg2X3YMhLh2X3WMhLk2W3TMiLm2V3RMkLo2U3PMkLR3S3nLmLS3S3lLmLV3Q3jLoLW3Q3gLPMZ3P3eLPM]3n2cLRM^3n2aLRM`3m2`LSMb3l2\\LUMe3j2[LVMg3i2lLcLU3]3SN"}, "label": "the blue plate"}]} {"id": 226966, "image": "COCO_train2014_000000226966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue serving plate is decorated with food\"."}], "task": "refering", "answer_template": "The \"a blue serving plate is decorated with food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 19, 20, 21, 22, 27, 28, 29, 30, 31, 36, 37, 38, 39, 47, 48, 49, 50, 51, 52, 54, 55, 56, 57, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 85, 86, 87, 88, 89, 92, 93, 94, 104, 105, 106, 107, 123, 124, 125, 142, 143, 160, 161, 177, 178, 179, 194, 195, 196, 197, 211, 212, 213, 214, 215], "bbox": [0.00174, 0.06113432835820896, 1.0, 0.986], "iscrowd": 0, "area": 32284.68669999999, "rle": {"size": [335, 500], "counts": "eaNC_1;bNE_19bNG^18cNI]17bNI_16aNK_15`NKa14_NMa12_NO`12_NNb11^N0c1O\\N2d1M\\N3f1KZN6g1IXN8h1GXN9j1FUN;k1DUN=l1ATN?m1ARN`0o1^OQNc0o1]OPNc0R2[OmMg0T2WOlMj0T2VOkMj0W2fMZLe0_1f1W2bM^Lg0Z1g1Z2_M_Lh0W1j1Z2[McLi0S1m1[2WMeLj0P1o1]2SMgLk0m0S2\\2oLkLl0i0V2\\2lLnLl0f0X2[2kLSMk0b0[2[3dMeL]2Z3bMgL^2Z3aMfL`2Y3`MgLa2Y3^MgLb2Z3]MfLd2Y3[MhLf2X3YMhLg2X3YMhLh2X3WMhLk2W3TMiLm2V3RMkLo2U3PMkLR3S3nLmLS3S3lLmLV3Q3jLoLW3Q3gLPMZ3P3eLPM]3n2cLRM^3n2aLRM`3m2`LSMb3l2\\LUMe3j2[LVMg3i2lLcLU3]3SN"}, "label": "a blue serving plate is decorated with food"}]} {"id": 38552, "image": "COCO_train2014_000000038552.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep in the middle of two other sheep\"."}], "task": "refering", "answer_template": "The \"a sheep in the middle of two other sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 222, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 307, 308, 309, 310, 311, 312, 330, 331, 332, 334, 335, 353, 354, 357, 358], "bbox": [0.357328125, 0.5262630480167015, 0.67959375, 0.941795407098121], "iscrowd": 0, "area": 19698.61415, "rle": {"size": [479, 640], "counts": "^a[38`>9G9G9H8H4M5L4N1YO\\NRDg1k;aNnC`1QPFCn9>QFCo9?oEBP:?nECR:>lECT:=jEDW:=gEDY:3;00000001O0000001O0000001N101O1N2O1O1N2OVdo2"}, "label": "a sheep in the middle of two other sheep"}]} {"id": 38552, "image": "COCO_train2014_000000038552.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small sheep standing between two other sheep\"."}], "task": "refering", "answer_template": "The \"a small sheep standing between two other sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 222, 239, 240, 241, 242, 243, 244, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 288, 289, 290, 307, 308, 309, 310, 311, 312, 330, 331, 332, 334, 335, 353, 354, 357, 358], "bbox": [0.357328125, 0.5262630480167015, 0.67959375, 0.941795407098121], "iscrowd": 0, "area": 19698.61415, "rle": {"size": [479, 640], "counts": "^a[38`>9G9G9H8H4M5L4N1YO\\NRDg1k;aNnC`1QPFCn9>QFCo9?oEBP:?nECR:>lECT:=jEDW:=gEDY:3;00000001O0000001O0000001N101O1N2O1O1N2OVdo2"}, "label": "a small sheep standing between two other sheep"}]} {"id": 38552, "image": "COCO_train2014_000000038552.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lamb in a meadow grazing next to two other lambs\"."}], "task": "refering", "answer_template": "The \"lamb in a meadow grazing next to two other lambs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 244, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 296, 297, 313, 314, 317, 318, 319, 320, 336, 337, 338, 340, 341, 343], "bbox": [0.62771875, 0.5827974947807933, 0.9447968750000001, 0.8645093945720251], "iscrowd": 0, "area": 14673.890050000002, "rle": {"size": [479, 640], "counts": "o]l55e>7I7I7I7H6K5K5TNcNSFS1]NTO];OQFm0eNSOW;7nEe0nNTOQ;=lE?VOSOk:e0iE7@SOd:l0gE0HTO_:R1cEJ1QO^:X1\\EF:PO[:^1UEAc0POX:c1PE]OP in this image.", "objects": [{"patches": [223, 244, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 296, 297, 313, 314, 317, 318, 319, 320, 336, 337, 338, 340, 341, 343], "bbox": [0.62771875, 0.5827974947807933, 0.9447968750000001, 0.8645093945720251], "iscrowd": 0, "area": 14673.890050000002, "rle": {"size": [479, 640], "counts": "o]l55e>7I7I7I7H6K5K5TNcNSFS1]NTO];OQFm0eNSOW;7nEe0nNTOQ;=lE?VOSOk:e0iE7@SOd:l0gE0HTO_:R1cEJ1QO^:X1\\EF:PO[:^1UEAc0POX:c1PE]OP in this image.", "objects": [{"patches": [72, 73, 74, 95, 96, 97, 118, 119, 120, 121, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236, 255, 256, 257, 258, 259, 279, 280, 281, 301, 302, 303, 304, 324, 325, 326, 327, 347, 348, 349, 350, 369, 370, 372, 373, 393], "bbox": [0.07640625, 0.1505625, 0.292140625, 0.757296875], "iscrowd": 0, "area": 29390.813699999995, "rle": {"size": [640, 640], "counts": "QRo03lc03M4K5L4L4K5L4L4M2O1OO10eKSOaEm0V:@eEb0Q:KeE;R:1dE6T:6cE0V::bEJY::cEKX:9eELT:8jELQ:8lELo98nEMk97RFNi96TFNX9e0eF]Oj8Q1WGPOf8j0bGWO\\8h0eGZOX8g0jGYOS8h0oGXOT8c0nG]OT8?nGAU8:nGEU87lGIV83lGMW8NkG2X8IkG5o6`MlFW2W29n6cMjFo1Y2?m6eMjFg1W2h0P7dMgF`1X2o0Q7eMfFo0_2`1l6cMeF9l2X2_6cMcF3k2]2c6bMbFNb2i2m6\\M`FHU2Z3]7oL]FJP2X3f7oLXFMk1U3P8nLTF0e1T3Y8nLPF1a1R3b8mLlE5[1n2m8nLfE7V1m2V9nLaE:R1i2`9mL]ER6f:oIXEo5j:a13M3M2N3M4L3M3M34L5K5K5K6J5K5K5K5K6J5K5K5K5K5K5L2M3M2N3MSFZGb8d8\\G`Ge8^8XGeGj8Y8SGjGl8X8PGlGo8T8oFnGP9S8mFPHR9Q8kFSHT9m7iFVHV9k7gFWHZ9W900010O00lLnEZLR:d3YFTLf9j3dFnK]9n3nFkKQ9S4YGeKg8Y4cG_K]8_4jG]KV8_4PH^KP8`4UH]Kk7a4ZH]Ke7a4aH[K_7b4gH\\KY7a4lHVKZ7h4kHoJ[7n4kHiJ[7U5jHcJ[7[5g3M2M3N2N3M3L5L4L4L4K5L4L4K5L4L4L4K5L4L4L4K5\\Oanj8"}, "label": "a boy wearing a white shirt and blue jeans standing on a skateboard"}]} {"id": 243354, "image": "COCO_train2014_000000243354.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing a white shirt who is riding a skateboard\"."}], "task": "refering", "answer_template": "The \"a person wearing a white shirt who is riding a skateboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 95, 96, 97, 118, 119, 120, 121, 140, 141, 142, 143, 144, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213, 232, 233, 234, 235, 236, 255, 256, 257, 258, 259, 279, 280, 281, 301, 302, 303, 304, 324, 325, 326, 327, 347, 348, 349, 350, 369, 370, 372, 373, 393], "bbox": [0.07640625, 0.1505625, 0.292140625, 0.757296875], "iscrowd": 0, "area": 29390.813699999995, "rle": {"size": [640, 640], "counts": "QRo03lc03M4K5L4L4K5L4L4M2O1OO10eKSOaEm0V:@eEb0Q:KeE;R:1dE6T:6cE0V::bEJY::cEKX:9eELT:8jELQ:8lELo98nEMk97RFNi96TFNX9e0eF]Oj8Q1WGPOf8j0bGWO\\8h0eGZOX8g0jGYOS8h0oGXOT8c0nG]OT8?nGAU8:nGEU87lGIV83lGMW8NkG2X8IkG5o6`MlFW2W29n6cMjFo1Y2?m6eMjFg1W2h0P7dMgF`1X2o0Q7eMfFo0_2`1l6cMeF9l2X2_6cMcF3k2]2c6bMbFNb2i2m6\\M`FHU2Z3]7oL]FJP2X3f7oLXFMk1U3P8nLTF0e1T3Y8nLPF1a1R3b8mLlE5[1n2m8nLfE7V1m2V9nLaE:R1i2`9mL]ER6f:oIXEo5j:a13M3M2N3M4L3M3M34L5K5K5K6J5K5K5K5K6J5K5K5K5K5K5L2M3M2N3MSFZGb8d8\\G`Ge8^8XGeGj8Y8SGjGl8X8PGlGo8T8oFnGP9S8mFPHR9Q8kFSHT9m7iFVHV9k7gFWHZ9W900010O00lLnEZLR:d3YFTLf9j3dFnK]9n3nFkKQ9S4YGeKg8Y4cG_K]8_4jG]KV8_4PH^KP8`4UH]Kk7a4ZH]Ke7a4aH[K_7b4gH\\KY7a4lHVKZ7h4kHoJ[7n4kHiJ[7U5jHcJ[7[5g3M2M3N2N3M3L5L4L4L4K5L4L4K5L4L4L4K5L4L4L4K5\\Oanj8"}, "label": "a person wearing a white shirt who is riding a skateboard"}]} {"id": 22195, "image": "COCO_train2014_000000022195.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train cart in front of and to the side of the blue train\"."}], "task": "refering", "answer_template": "The \"the train cart in front of and to the side of the blue train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230], "bbox": [0.0, 0.387018779342723, 0.18671875, 0.6753286384976526], "iscrowd": 0, "area": 12975.1764, "rle": {"size": [426, 640], "counts": "U5k3_9O1000001O0000000000001O0O10000000001O0000000000001N100000000000001O0000000O10001O0000000000001O000O100000001O0000000000001N1000000000001O000000000O101O000000000000001O00000O1000001O0000000000001O0O10000000001O00000000000O2O0000000000P3PMU_h6"}, "label": "the train cart in front of and to the side of the blue train"}]} {"id": 22195, "image": "COCO_train2014_000000022195.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the brown train next to the blue train\"."}], "task": "refering", "answer_template": "The \"the brown train next to the blue train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [115, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 230], "bbox": [0.0, 0.387018779342723, 0.18671875, 0.6753286384976526], "iscrowd": 0, "area": 12975.1764, "rle": {"size": [426, 640], "counts": "U5k3_9O1000001O0000000000001O0O10000000001O0000000000001N100000000000001O0000000O10001O0000000000001O000O100000001O0000000000001N1000000000001O000000000O101O000000000000001O00000O1000001O0000000000001O0O10000000001O00000000000O2O0000000000P3PMU_h6"}, "label": "the brown train next to the blue train"}]} {"id": 22195, "image": "COCO_train2014_000000022195.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue train\"."}], "task": "refering", "answer_template": "The \"blue train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 234, 235], "bbox": [0.182078125, 0.38340375586854464, 0.674578125, 0.6748826291079812], "iscrowd": 0, "area": 27271.346149999994, "rle": {"size": [426, 640], "counts": "[k`1X1j;Z1mNi0XO001N100O2O0O2O0O101N100O2OO0100O100O1O100O100O100O100000000000O100000000000000000000000001O0000001N100000001O0000001O0000001O0000000O2O00000000001O0O100000001O00000O10001O00000000001N10000000001O000O10001O000000000O2O00000000001O0O100000001O00000O10001O0000000O101O000000000O101O0000000O1000001O00000O1000001O000O10000000001N1000000000001N10000000000O101O000000000O101O0000000O1000001O000O10000000001O0O100000001O0O1000001O00000O2O0000001O0O1000001O000O101O000000001N1000001O000O10001O0000001N1000001O000O10001O00000O2O000000001O0O10001O00000O101O000000001N10000000000O1000000O2O000O10000O10000O10001N10000O10000O105K9F:G:F_Yf2"}, "label": "blue train"}]} {"id": 22195, "image": "COCO_train2014_000000022195.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train\"."}], "task": "refering", "answer_template": "The \"the train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 234, 235], "bbox": [0.182078125, 0.38340375586854464, 0.674578125, 0.6748826291079812], "iscrowd": 0, "area": 27271.346149999994, "rle": {"size": [426, 640], "counts": "[k`1X1j;Z1mNi0XO001N100O2O0O2O0O101N100O2OO0100O100O1O100O100O100O100000000000O100000000000000000000000001O0000001N100000001O0000001O0000001O0000000O2O00000000001O0O100000001O00000O10001O00000000001N10000000001O000O10001O000000000O2O00000000001O0O100000001O00000O10001O0000000O101O000000000O101O0000000O1000001O00000O1000001O000O10000000001N1000000000001N10000000000O101O000000000O101O0000000O1000001O000O10000000001O0O100000001O0O1000001O00000O2O0000001O0O1000001O000O101O000000001N1000001O000O10001O0000001N1000001O000O10001O00000O2O000000001O0O10001O00000O101O000000001N10000000000O1000000O2O000O10000O10000O10001N10000O10000O105K9F:G:F_Yf2"}, "label": "the train"}]} {"id": 169653, "image": "COCO_train2014_000000169653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball umpire calling a strikeout\"."}], "task": "refering", "answer_template": "The \"a baseball umpire calling a strikeout\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 102, 105, 106, 107, 125, 126, 127, 128, 129, 130, 131, 132, 133, 150, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 311, 312, 315, 316, 333, 334, 335, 339, 340, 356, 357, 362, 363, 379, 380, 385, 386], "bbox": [0.43162500000000004, 0.20603174603174604, 0.8227343749999999, 0.9196825396825397], "iscrowd": 0, "area": 35122.72864999998, "rle": {"size": [504, 640], "counts": "\\oW41d?3L4M300000000000000001O01O1O1O1O2N100O01O001O0010O100O100O010O100O100O010O010O01O0010O0UI_O`N`0\\1EXHKl5?i1KVHIQ6=e1NVHIU68b14THHY64`1e0`NZO]1j0bNWOZ1l0gNTOU1P1jNPOS1T1lNmNP1W1oNjNm0Z1SOfNi0]1WOcNf0a1YO`Nd0a1]O`N>d1C\\N9g1GYN5k1KVN0n10SNLP24QNGS2:lMBX2>iM_OY2a0hM]OX2d0iMYOY2g0gMWO[2i0fMTO\\2l0dMSOU2U1kMiNT2Z1mMcNS2_1mM_NS2c1mM\\NQ2g1oMWNQ2k1PNRNo1Q2QNmMo1U2QNjMn1X2SNeMl1]2UNaMk1a2UN^Mj1d2WNYMh1j2XNTMh1k2[NSMd1l2aNRM^1n2dNPM\\1P3fNnLY1T3hNjLX1W3jNoKmNoLX2S7mNgKQOTMR2V7PO^KUOYMk1Y7SOXKWO]Mf1]7UOQKYO`Ma1`7XOnJYO_M_1d7[OjJg1X5\\NdJd1^5aN\\J^1e5gNVJX1k5mNoIT1Q6QOjIo0W6VOcIj0]6[O^Ie0b6DUI6@M=4BN<3C093E081G26OH45NJ53MK52ML62KM71LL613F08j7M3NhUg1"}, "label": "a baseball umpire calling a strikeout"}]} {"id": 169653, "image": "COCO_train2014_000000169653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an umpire in a baseball game\"."}], "task": "refering", "answer_template": "The \"an umpire in a baseball game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 102, 105, 106, 107, 125, 126, 127, 128, 129, 130, 131, 132, 133, 150, 151, 152, 153, 154, 155, 156, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 220, 221, 222, 223, 224, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 292, 293, 311, 312, 315, 316, 333, 334, 335, 339, 340, 356, 357, 362, 363, 379, 380, 385, 386], "bbox": [0.43162500000000004, 0.20603174603174604, 0.8227343749999999, 0.9196825396825397], "iscrowd": 0, "area": 35122.72864999998, "rle": {"size": [504, 640], "counts": "\\oW41d?3L4M300000000000000001O01O1O1O1O2N100O01O001O0010O100O100O010O100O100O010O010O01O0010O0UI_O`N`0\\1EXHKl5?i1KVHIQ6=e1NVHIU68b14THHY64`1e0`NZO]1j0bNWOZ1l0gNTOU1P1jNPOS1T1lNmNP1W1oNjNm0Z1SOfNi0]1WOcNf0a1YO`Nd0a1]O`N>d1C\\N9g1GYN5k1KVN0n10SNLP24QNGS2:lMBX2>iM_OY2a0hM]OX2d0iMYOY2g0gMWO[2i0fMTO\\2l0dMSOU2U1kMiNT2Z1mMcNS2_1mM_NS2c1mM\\NQ2g1oMWNQ2k1PNRNo1Q2QNmMo1U2QNjMn1X2SNeMl1]2UNaMk1a2UN^Mj1d2WNYMh1j2XNTMh1k2[NSMd1l2aNRM^1n2dNPM\\1P3fNnLY1T3hNjLX1W3jNoKmNoLX2S7mNgKQOTMR2V7PO^KUOYMk1Y7SOXKWO]Mf1]7UOQKYO`Ma1`7XOnJYO_M_1d7[OjJg1X5\\NdJd1^5aN\\J^1e5gNVJX1k5mNoIT1Q6QOjIo0W6VOcIj0]6[O^Ie0b6DUI6@M=4BN<3C093E081G26OH45NJ53MK52ML62KM71LL613F08j7M3NhUg1"}, "label": "an umpire in a baseball game"}]} {"id": 169653, "image": "COCO_train2014_000000169653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player in the batter ' s box holding a bat and wearing a shin guard\"."}], "task": "refering", "answer_template": "The \"a baseball player in the batter ' s box holding a bat and wearing a shin guard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 56, 57, 58, 79, 80, 101, 102, 103, 124, 125, 126, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 193, 194, 216, 217, 239, 240, 262, 263, 285, 286], "bbox": [0.361734375, 0.07251984126984126, 0.541734375, 0.6791269841269841], "iscrowd": 0, "area": 15483.9067, "rle": {"size": [504, 640], "counts": "V\\b34c?3M4K5L3M3M0O2O1O1O1N101O1O1N2O001QEPOU7Q1hHTOU7m0gHYOV7h0gH^OU7b0iHAV7`0fHEX7 in this image.", "objects": [{"patches": [33, 34, 35, 56, 57, 58, 79, 80, 101, 102, 103, 124, 125, 126, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 193, 194, 216, 217, 239, 240, 262, 263, 285, 286], "bbox": [0.361734375, 0.07251984126984126, 0.541734375, 0.6791269841269841], "iscrowd": 0, "area": 15483.9067, "rle": {"size": [504, 640], "counts": "V\\b34c?3M4K5L3M3M0O2O1O1O1N101O1O1N2O001QEPOU7Q1hHTOU7m0gHYOV7h0gH^OU7b0iHAV7`0fHEX7 in this image.", "objects": [{"patches": [58, 59, 80, 81, 82, 83, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 237, 238, 239, 240, 241, 242, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 307, 308, 309, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.23737499999999997, 0.15504694835680752, 0.6805781249999999, 1.0], "iscrowd": 0, "area": 55026.114149999994, "rle": {"size": [426, 640], "counts": "]co18Q= in this image.", "objects": [{"patches": [58, 59, 80, 81, 82, 83, 100, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 146, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 237, 238, 239, 240, 241, 242, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 307, 308, 309, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.23737499999999997, 0.15504694835680752, 0.6805781249999999, 1.0], "iscrowd": 0, "area": 55026.114149999994, "rle": {"size": [426, 640], "counts": "]co18Q= in this image.", "objects": [{"patches": [83, 84, 85, 86, 106, 107, 108, 109, 129, 130, 131, 132, 152, 153, 154, 155, 156, 175, 176, 177, 178, 198, 199, 200, 201, 224], "bbox": [0.622203125, 0.23426229508196722, 0.796125, 0.6162997658079624], "iscrowd": 0, "area": 14471.292750000004, "rle": {"size": [427, 640], "counts": "]VV56c^Ob0_O`0_Ob0^Ob0^Oa0_Ob0^Ob0D6L2O2N1N2OT^f1"}, "label": "brown double bus"}]} {"id": 169656, "image": "COCO_train2014_000000169656.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red double decker bus with the number 113 on it\"."}], "task": "refering", "answer_template": "The \"a red double decker bus with the number 113 on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 106, 107, 108, 109, 129, 130, 131, 132, 152, 153, 154, 155, 156, 175, 176, 177, 178, 198, 199, 200, 201, 224], "bbox": [0.622203125, 0.23426229508196722, 0.796125, 0.6162997658079624], "iscrowd": 0, "area": 14471.292750000004, "rle": {"size": [427, 640], "counts": "]VV56c^Ob0_O`0_Ob0^Ob0^Oa0_Ob0^Ob0D6L2O2N1N2OT^f1"}, "label": "a red double decker bus with the number 113 on it"}]} {"id": 333498, "image": "COCO_train2014_000000333498.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black and purple piece of luggage outside of a train\"."}], "task": "refering", "answer_template": "The \"a black and purple piece of luggage outside of a train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367], "bbox": [0.5839218749999999, 0.5594784580498866, 1.0, 0.989047619047619], "iscrowd": 0, "area": 42295.80605000001, "rle": {"size": [441, 640], "counts": "l^Q5l0f;X1A>N2M3N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1000000O1000000O1000000O1000000O10000000000O1000000000000000000000000O101O000000000000000000000O10000000000000000000000O1000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O1O100000000001O000000001O00000000001O000000001O000000001O00000000001O00001O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O0YNn1kMRI"}, "label": "a black and purple piece of luggage outside of a train"}]} {"id": 333498, "image": "COCO_train2014_000000333498.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"gray luggage\"."}], "task": "refering", "answer_template": "The \"gray luggage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 252, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367], "bbox": [0.5839218749999999, 0.5594784580498866, 1.0, 0.989047619047619], "iscrowd": 0, "area": 42295.80605000001, "rle": {"size": [441, 640], "counts": "l^Q5l0f;X1A>N2M3N2M3M3N2M3N2M3M3N2M3N2M3M3N2M3N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1000000O1000000O1000000O1000000O10000000000O1000000000000000000000000O101O000000000000000000000O10000000000000000000000O1000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O1O100O1O100O1O100O1O100O1O100O1O1O100000000001O000000001O00000000001O000000001O000000001O00000000001O00001O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O0YNn1kMRI"}, "label": "gray luggage"}]} {"id": 333498, "image": "COCO_train2014_000000333498.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue suitcase\"."}], "task": "refering", "answer_template": "The \"a blue suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 242, 243, 244, 264, 265, 266, 267, 287, 288, 289, 310, 311, 312, 333, 334, 335, 355, 356, 357, 358], "bbox": [0.471796875, 0.6362585034013605, 0.6518124999999999, 0.9988662131519275], "iscrowd": 0, "area": 11345.353449999997, "rle": {"size": [441, 640], "counts": "`_R47T=>D;G:F:F:F:F:F:F:F:F:F:F:M3000000O100000000O100000O1000O100000000O100000000O100000000O100000000O010000000O10000000000O100000000O100000000OO2F:F:F:F:F:F:F:E;F:F:G801O1N2O1O1O1O1N2O1O1O001N2O1O1O1O1N2O1O1O1O0O2O1O1O1O1N2O1O1O1O2M_ho2"}, "label": "a blue suitcase"}]} {"id": 333498, "image": "COCO_train2014_000000333498.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue suitcase by a train\"."}], "task": "refering", "answer_template": "The \"a blue suitcase by a train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 242, 243, 244, 264, 265, 266, 267, 287, 288, 289, 310, 311, 312, 333, 334, 335, 355, 356, 357, 358], "bbox": [0.471796875, 0.6362585034013605, 0.6518124999999999, 0.9988662131519275], "iscrowd": 0, "area": 11345.353449999997, "rle": {"size": [441, 640], "counts": "`_R47T=>D;G:F:F:F:F:F:F:F:F:F:F:M3000000O100000000O100000O1000O100000000O100000000O100000000O100000000O010000000O10000000000O100000000O100000000OO2F:F:F:F:F:F:F:E;F:F:G801O1N2O1O1O1O1N2O1O1O001N2O1O1O1O1N2O1O1O1O0O2O1O1O1O1N2O1O1O1O2M_ho2"}, "label": "a blue suitcase by a train"}]} {"id": 136894, "image": "COCO_train2014_000000136894.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a big size mango in the plate\"."}], "task": "refering", "answer_template": "The \"a big size mango in the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 109, 110, 111, 129, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 204, 205, 221, 222, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 250, 269, 270, 271, 272], "bbox": [0.6193906250000001, 0.263986013986014, 0.9205781250000001, 0.7732167832167833], "iscrowd": 0, "area": 32220.715149999993, "rle": {"size": [429, 640], "counts": "TSV52Y=b0]Oc0^O>B4K5L4L4L4K5L4L4L4K5L4L4K5L4M3M3M2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N1O2N1O1O2O0O2O000O2O000O2O001O0O101O0O2O000O2O00001N101O0O101O0O101O001N10001N10001O0010O0001O001O00001O0000O2O00000000000O10000000001O000O1000000000000O101O000000000O10000000001O1O1N2O1N2N2N2N2N2N2N2N3L3N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2M3N2N2N2M5D;E in this image.", "objects": [{"patches": [107, 108, 109, 110, 111, 129, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 175, 176, 177, 178, 179, 180, 181, 182, 198, 199, 200, 201, 202, 203, 204, 205, 221, 222, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 250, 269, 270, 271, 272], "bbox": [0.6193906250000001, 0.263986013986014, 0.9205781250000001, 0.7732167832167833], "iscrowd": 0, "area": 32220.715149999993, "rle": {"size": [429, 640], "counts": "TSV52Y=b0]Oc0^O>B4K5L4L4L4K5L4L4L4K5L4L4K5L4M3M3M2N2N2N2N3M2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N1O2N1O1O2O0O2O000O2O000O2O001O0O101O0O2O000O2O00001N101O0O101O0O101O001N10001N10001O0010O0001O001O00001O0000O2O00000000000O10000000001O000O1000000000000O101O000000000O10000000001O1O1N2O1N2N2N2N2N2N2N2N3L3N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2M3N2N2N2M5D;E in this image.", "objects": [{"patches": [133, 134, 135, 156, 157, 158, 159, 175, 179, 180, 181, 182, 198, 199, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 249, 250, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344, 365, 366, 367, 388, 389, 390, 411, 412, 413, 435, 436, 458, 459], "bbox": [0.6045625, 0.24323076923076922, 1.0, 0.9407521367521368], "iscrowd": 0, "area": 28307.687550000002, "rle": {"size": [585, 640], "counts": "iYm66Pb09F6N3M2O2M2N101N1O1O100O1O1O100001O001O1O1O1O100O1O001O1OO0100O00100O010O010O10O0100O010O1O010O10O01O001O1O010O1O001O1O00100O001O001O1O010O00O100000001O0O10001O000O2O000000000O101O00000O1O1O100O1O1O1O101N001O00100O001O010O1O001O010O001O1O010O0PNPOhBQ1U=XOdBh0Z=A`B?^=D`B=]=GbB8]=L`B5^=NaB2]=1bBO[=6bBK\\=8cBH\\=:cBF\\=[2N1N2O2N1N2O2MG[BQLd=o3]BQLb=o3_BRL^=n3dBRL\\=m3eBSLZ=m3fBULX=j3jBVLV=i3kBWLT=i3lBYLR=g3oBYLQ=e3QC[LnWOb1R8mMfHa0XOc1U8hMcHe0XOd1X8cM`Hi0XOd1Z8aM^Hk0XOe1[8^M\\Ho0XOd1^8[MYHQ1YOe1_8XMXHS1YOf1`8UMWHU1YOg1b8QMUHY1XOf1e8oLRH\\1YOf1f8lLQH^1YOf1h8kLnG_1ZOg1i8hLQI=VNk2k8fLUI:oMP3m8eLUI:oMQ3m8cLVI:mMS3n8cLVI7nMU3n8bLVI7lMX3n8`LXI6kMY3o8_LXI6jMZ3T9oKQJ2lLP4c=O00100O001^N_D"}, "label": "a women in a pink shirt petting a horse"}]} {"id": 431817, "image": "COCO_train2014_000000431817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman touching the horse\"."}], "task": "refering", "answer_template": "The \"woman touching the horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 134, 135, 156, 157, 158, 159, 175, 179, 180, 181, 182, 198, 199, 202, 203, 204, 205, 206, 221, 222, 223, 224, 225, 226, 227, 228, 229, 246, 247, 249, 250, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344, 365, 366, 367, 388, 389, 390, 411, 412, 413, 435, 436, 458, 459], "bbox": [0.6045625, 0.24323076923076922, 1.0, 0.9407521367521368], "iscrowd": 0, "area": 28307.687550000002, "rle": {"size": [585, 640], "counts": "iYm66Pb09F6N3M2O2M2N101N1O1O100O1O1O100001O001O1O1O1O100O1O001O1OO0100O00100O010O010O10O0100O010O1O010O10O01O001O1O010O1O001O1O00100O001O001O1O010O00O100000001O0O10001O000O2O000000000O101O00000O1O1O100O1O1O1O101N001O00100O001O010O1O001O010O001O1O010O0PNPOhBQ1U=XOdBh0Z=A`B?^=D`B=]=GbB8]=L`B5^=NaB2]=1bBO[=6bBK\\=8cBH\\=:cBF\\=[2N1N2O2N1N2O2MG[BQLd=o3]BQLb=o3_BRL^=n3dBRL\\=m3eBSLZ=m3fBULX=j3jBVLV=i3kBWLT=i3lBYLR=g3oBYLQ=e3QC[LnWOb1R8mMfHa0XOc1U8hMcHe0XOd1X8cM`Hi0XOd1Z8aM^Hk0XOe1[8^M\\Ho0XOd1^8[MYHQ1YOe1_8XMXHS1YOf1`8UMWHU1YOg1b8QMUHY1XOf1e8oLRH\\1YOf1f8lLQH^1YOf1h8kLnG_1ZOg1i8hLQI=VNk2k8fLUI:oMP3m8eLUI:oMQ3m8cLVI:mMS3n8cLVI7nMU3n8bLVI7lMX3n8`LXI6kMY3o8_LXI6jMZ3T9oKQJ2lLP4c=O00100O001^N_D"}, "label": "woman touching the horse"}]} {"id": 505546, "image": "COCO_train2014_000000505546.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green fire hydrant with water coming out of it\"."}], "task": "refering", "answer_template": "The \"a green fire hydrant with water coming out of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 143, 144, 145, 146, 160, 161, 162, 163, 177, 178, 179, 180, 194, 195, 196, 197], "bbox": [0.3505625, 0.13273437500000002, 0.704125, 0.512515625], "iscrowd": 0, "area": 21443.2824, "rle": {"size": [640, 480], "counts": "SiY33lc03M3M3M3M3M2N3M3M3M1O1O1O1O1O1O1O2N1000000O100000000O2O0000000O100000000O2Oe0^@TNf;T4bA\\LP>R5G9O1O1N2O1O1O0O2O1O0O2O00000O1000000O2O000O1000000O101O0O1000000O10001O0H8C=000000O101O0000000O1000001O000O10000000001N10000?A00000010O0001O0000001O01O01VNmCeKS in this image.", "objects": [{"patches": [59, 60, 61, 75, 76, 77, 78, 79, 91, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 113, 126, 127, 128, 129, 143, 144, 145, 146, 160, 161, 162, 163, 177, 178, 179, 180, 194, 195, 196, 197], "bbox": [0.3505625, 0.13273437500000002, 0.704125, 0.512515625], "iscrowd": 0, "area": 21443.2824, "rle": {"size": [640, 480], "counts": "SiY33lc03M3M3M3M3M2N3M3M3M1O1O1O1O1O1O1O2N1000000O100000000O2O0000000O100000000O2Oe0^@TNf;T4bA\\LP>R5G9O1O1N2O1O1O0O2O1O0O2O00000O1000000O2O000O1000000O101O0O1000000O10001O0H8C=000000O101O0000000O1000001O000O10000000001N10000?A00000010O0001O0000001O01O01VNmCeKS in this image.", "objects": [{"patches": [46, 47, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 299, 301, 302, 324, 325, 326], "bbox": [0.003, 0.18918032786885247, 0.534953125, 0.9707259953161592], "iscrowd": 0, "area": 51813.732350000006, "rle": {"size": [427, 640], "counts": "Wm0f9e3O100O100O100O100O100O100O2O0O100O100O100O100O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1N2O2N1O1L4PNQL^JU4_5m1M3L4M4L300001O1O1Od0]O>A?ULcHj0l7bN[H\\1S8QNTHm1[8`MkG^2c9N01O00000001OO101O0000000000001O000000000000001O000000000000000000000000000lNSMjGm2o7]MnGc2l7fMQH[2h7oMTHQ2f7VNZHi1a7]N^Hc1\\7dNcH\\1W7jNhHW1R7POlHQ1n6VOQIj0j6\\OTIe0f6EUI in this image.", "objects": [{"patches": [46, 47, 69, 70, 71, 72, 73, 74, 75, 92, 93, 94, 95, 96, 97, 98, 99, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 161, 162, 163, 164, 165, 166, 167, 168, 169, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 207, 208, 209, 210, 211, 215, 216, 217, 218, 219, 230, 231, 232, 233, 234, 253, 254, 255, 256, 257, 276, 277, 278, 279, 299, 301, 302, 324, 325, 326], "bbox": [0.003, 0.18918032786885247, 0.534953125, 0.9707259953161592], "iscrowd": 0, "area": 51813.732350000006, "rle": {"size": [427, 640], "counts": "Wm0f9e3O100O100O100O100O100O100O2O0O100O100O100O100O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1N2O2N1O1L4PNQL^JU4_5m1M3L4M4L300001O1O1Od0]O>A?ULcHj0l7bN[H\\1S8QNTHm1[8`MkG^2c9N01O00000001OO101O0000000000001O000000000000001O000000000000000000000000000lNSMjGm2o7]MnGc2l7fMQH[2h7oMTHQ2f7VNZHi1a7]N^Hc1\\7dNcH\\1W7jNhHW1R7POlHQ1n6VOQIj0j6\\OTIe0f6EUI in this image.", "objects": [{"patches": [108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 245, 246, 247, 248, 249, 250, 251, 268, 269, 272, 273, 274, 290, 291, 292, 295, 296, 297, 298, 314, 315, 318, 319], "bbox": [0.47184375, 0.2826463700234192, 1.0, 0.8975175644028103], "iscrowd": 0, "area": 43932.189150000035, "rle": {"size": [427, 640], "counts": "UWn31X=2N2N2N2N2N2N2O1O1O100O00O1O1O1O1O2N1O2N2O1O10O01O1O100O00100O1O^O1jCOU<3gCOY<3cCO^<3]COc<=O00010O00010O000010O0001O01O01O01O01O00010O00010O000010O00010O0001O001O01O01O00001O00010O00001O00001O01O01O00001O000010O00O2M3N5J5L4K6K5J7J5K5L5K4L4QH\\MP5h2kJ\\MR5i2hJ[MU5i2fJ\\MV5h2eJ\\MX5i2bJ[M[5i2`J[M]5i2^JYMa5k2ZJXMd5m2WJTMh5P3UJQMi5S3SJoLk5U3QJlLn5Y3nIhLP6\\3lIeLS6_3jIbLT6[5O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O1O2N101N10001O001N1000O1O010O100O1O010O1VOVJRJk5l5cJhI\\5V6iJgIX5W6mJfIS5X6QKeIP5W6T1J5dM]ImMh6Q2Z2N2N1O2N2O1000O103M5K8G in this image.", "objects": [{"patches": [108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 218, 219, 221, 222, 223, 224, 225, 226, 227, 228, 229, 241, 245, 246, 247, 248, 249, 250, 251, 268, 269, 272, 273, 274, 290, 291, 292, 295, 296, 297, 298, 314, 315, 318, 319], "bbox": [0.47184375, 0.2826463700234192, 1.0, 0.8975175644028103], "iscrowd": 0, "area": 43932.189150000035, "rle": {"size": [427, 640], "counts": "UWn31X=2N2N2N2N2N2N2O1O1O100O00O1O1O1O1O2N1O2N2O1O10O01O1O100O00100O1O^O1jCOU<3gCOY<3cCO^<3]COc<=O00010O00010O000010O0001O01O01O01O01O00010O00010O000010O00010O0001O001O01O01O00001O00010O00001O00001O01O01O00001O000010O00O2M3N5J5L4K6K5J7J5K5L5K4L4QH\\MP5h2kJ\\MR5i2hJ[MU5i2fJ\\MV5h2eJ\\MX5i2bJ[M[5i2`J[M]5i2^JYMa5k2ZJXMd5m2WJTMh5P3UJQMi5S3SJoLk5U3QJlLn5Y3nIhLP6\\3lIeLS6_3jIbLT6[5O2N1O2N1O2N1O2N1O2N1O2N1O1O2N1O1O2N101N10001O001N1000O1O010O100O1O010O1VOVJRJk5l5cJhI\\5V6iJgIX5W6mJfIS5X6QKeIP5W6T1J5dM]ImMh6Q2Z2N2N1O2N2O1000O103M5K8G in this image.", "objects": [{"patches": [49, 66, 67, 68, 69, 70, 84, 85, 86, 87, 88, 103, 104, 105, 106, 107, 122, 123, 124, 141, 142], "bbox": [0.69438, 0.18650666666666665, 0.96236, 0.5550666666666667], "iscrowd": 0, "area": 9407.102899999998, "rle": {"size": [375, 500], "counts": "hUo33c;8H2N2N2N3M2N2N2N3N1M3L4L4N2N2N2N2M3N1O2N2N2N2N2N2N2O1N2O0O2N2O1N2O1N2N2O11O1N2O1O001N2O1O1O0O10O100O010O100O1O010O1O1O010O1O2N2O1N2N2O2M2N101N1O2O0O1O2O0O2N100O0001O01O01O01O0001O01O0001O01O00010O0000010O0003L3N3L3M3M4M2M3M3M2O2M2I8I7L3N3M2N3L3N3M2M4M2N3M2M4M2N3M2Lki6"}, "label": "the right - most hot dog on the far right of the lunch tray"}]} {"id": 554706, "image": "COCO_train2014_000000554706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small hot dog farthers from the drink\"."}], "task": "refering", "answer_template": "The \"a small hot dog farthers from the drink\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 66, 67, 68, 69, 70, 84, 85, 86, 87, 88, 103, 104, 105, 106, 107, 122, 123, 124, 141, 142], "bbox": [0.69438, 0.18650666666666665, 0.96236, 0.5550666666666667], "iscrowd": 0, "area": 9407.102899999998, "rle": {"size": [375, 500], "counts": "hUo33c;8H2N2N2N3M2N2N2N3N1M3L4L4N2N2N2N2M3N1O2N2N2N2N2N2N2O1N2O0O2N2O1N2O1N2N2O11O1N2O1O001N2O1O1O0O10O100O010O100O1O010O1O1O010O1O2N2O1N2N2O2M2N101N1O2O0O1O2O0O2N100O0001O01O01O01O0001O01O0001O01O00010O0000010O0003L3N3L3M3M4M2M3M3M2O2M2I8I7L3N3M2N3L3N3M2M4M2N3M2M4M2N3M2Lki6"}, "label": "a small hot dog farthers from the drink"}]} {"id": 554706, "image": "COCO_train2014_000000554706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hotdog with nothing on it\"."}], "task": "refering", "answer_template": "The \"a hotdog with nothing on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 66, 80, 81, 82, 83, 84, 85, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 138, 139], "bbox": [0.4747, 0.25176, 0.79412, 0.5623466666666667], "iscrowd": 0, "area": 11225.02965, "rle": {"size": [375, 500], "counts": "]mf21e;9G3M2N3N1O2N101N2N1O2N1O2N1O2O1N1O1O1O0010O0100O010O0L5I6O2O0O2O1N101N1FWNfFj1R9^NlFc1S9`NkFa1T9aNiF`1W9aNhF`1V9b0O0O2O001N101O1N101O0O20O01O010O1O010O0010O01O00100O0010O01O010O00010O001O01O01O010O00010O001O01O010O010O1O010O010O00100O010O0010O0100O0010OO2O001N101O1N101O0O2O001O1N101O1N2O1O1N2O1N2N2N2N2N200O1O100O1O100O1O2N2O1N1N3M3M3M3M3M3K5K5K5GmaU1"}, "label": "a hotdog with nothing on it"}]} {"id": 554706, "image": "COCO_train2014_000000554706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hot dog in a toasted bun , served without any toppings , laying on a yellow bag\"."}], "task": "refering", "answer_template": "The \"a hot dog in a toasted bun , served without any toppings , laying on a yellow bag\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 66, 80, 81, 82, 83, 84, 85, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 138, 139], "bbox": [0.4747, 0.25176, 0.79412, 0.5623466666666667], "iscrowd": 0, "area": 11225.02965, "rle": {"size": [375, 500], "counts": "]mf21e;9G3M2N3N1O2N101N2N1O2N1O2N1O2O1N1O1O1O0010O0100O010O0L5I6O2O0O2O1N101N1FWNfFj1R9^NlFc1S9`NkFa1T9aNiF`1W9aNhF`1V9b0O0O2O001N101O1N101O0O20O01O010O1O010O0010O01O00100O0010O01O010O00010O001O01O01O010O00010O001O01O010O010O1O010O010O00100O010O0010O0100O0010OO2O001N101O1N101O0O2O001O1N101O1N2O1O1N2O1N2N2N2N2N200O1O100O1O100O1O2N2O1N1N3M3M3M3M3M3K5K5K5GmaU1"}, "label": "a hot dog in a toasted bun , served without any toppings , laying on a yellow bag"}]} {"id": 554706, "image": "COCO_train2014_000000554706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fast food meal on a red tray\"."}], "task": "refering", "answer_template": "The \"a fast food meal on a red tray\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.00508, 0.010719999999999999, 1.0, 0.98816], "iscrowd": 0, "area": 173316.3197, "rle": {"size": [375, 500], "counts": "\\S1\\;;000000000000000000000000000O10000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000003M3M2N2N2N2N1O2N1O2N2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2Ni5WJk0UO`D"}, "label": "a fast food meal on a red tray"}]} {"id": 554706, "image": "COCO_train2014_000000554706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a meal of hot dogs , french fries , sauces and a beverage\"."}], "task": "refering", "answer_template": "The \"a meal of hot dogs , french fries , sauces and a beverage\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.00508, 0.010719999999999999, 1.0, 0.98816], "iscrowd": 0, "area": 173316.3197, "rle": {"size": [375, 500], "counts": "\\S1\\;;000000000000000000000000000O10000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000003M3M2N2N2N2N1O2N1O2N2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2Ni5WJk0UO`D"}, "label": "a meal of hot dogs , french fries , sauces and a beverage"}]} {"id": 554706, "image": "COCO_train2014_000000554706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right half of the table , which is divided by a line\"."}], "task": "refering", "answer_template": "The \"the right half of the table , which is divided by a line\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 33, 34, 35, 52, 53, 70, 71, 89], "bbox": [0.82598, 0.036373333333333334, 0.99936, 0.38442666666666675], "iscrowd": 0, "area": 5636.488299999997, "rle": {"size": [375, 500], "counts": "Qdg42e;1O2M2O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1l0UOY1aN^1_NWL"}, "label": "the right half of the table , which is divided by a line"}]} {"id": 554706, "image": "COCO_train2014_000000554706.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden table with nothing in site on it\"."}], "task": "refering", "answer_template": "The \"a wooden table with nothing in site on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 33, 34, 35, 52, 53, 70, 71, 89], "bbox": [0.82598, 0.036373333333333334, 0.99936, 0.38442666666666675], "iscrowd": 0, "area": 5636.488299999997, "rle": {"size": [375, 500], "counts": "Qdg42e;1O2M2O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1l0UOY1aN^1_NWL"}, "label": "a wooden table with nothing in site on it"}]} {"id": 513748, "image": "COCO_train2014_000000513748.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in black holding her surfboard\"."}], "task": "refering", "answer_template": "The \"a girl in black holding her surfboard\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 120, 121, 122, 143, 144, 145, 166, 167, 168, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 281, 282, 283, 303, 304, 305, 306, 326, 327, 328, 329], "bbox": [0.166734375, 0.3229009433962264, 0.37517187500000004, 0.9858254716981132], "iscrowd": 0, "area": 23041.60025, "rle": {"size": [424, 640], "counts": "bb\\1n0RB7H7J6O2M2O1N2O2N1N2O2M2O1O2M2O1N3N1N20001O0SHeLl5\\3oIjLP6W3jIoLU6Q3eIWMY6i2bIlMm5V2mIoMS6Q2kIQNX4aNbL^3TOSNX4dN`L[3UOSNZ4gN]LV3WOUNZ4kN[LS3VOTN^4nNXLU3QOoMe4QOVLX3lNiMm4TOSLS5k3RKRLn4m3UKQLk4m3YKPLh4o3\\KnKd4P4`KnK`4P4dKmK]4R4fKkKZ4T4kKhKV4W4lKgKU4W4nKgKS4X4PLeKQ4[4PLdKP4\\4QLbKP4^4]200000000001O00000000000000000000000000000000001N100000000000001O001O0PJ^KZ3b4dL`K\\3a4aLaK_3_4_LdK`3\\4]LgKc3Z4YLiKe3Y4YLjK`3]4]LeKY3e4gL[Kk2S5UMnJb2[5]MeJ`2^5`MbJ]2a5cM`JZ2c5dM^JY2e5gM[JV2m5eMTJX2T6bMlI[2c6YM]Id2g6[MZIa2j6^MVI`2m6VM\\Ig2h6gLVH3X1Y2jNRNl7_OhIm1WOYNT7GgIm1a9L3M3N210000000N2M4L3L4M3M3M3M3M3L4M3M3M4L3M4KR\\U5"}, "label": "a girl in black holding her surfboard"}]} {"id": 513748, "image": "COCO_train2014_000000513748.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the surfer who is standing , holding his board upright , looking to the right ,\"."}], "task": "refering", "answer_template": "The \"the surfer who is standing , holding his board upright , looking to the right ,\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 120, 121, 122, 143, 144, 145, 166, 167, 168, 188, 189, 190, 191, 192, 210, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 281, 282, 283, 303, 304, 305, 306, 326, 327, 328, 329], "bbox": [0.166734375, 0.3229009433962264, 0.37517187500000004, 0.9858254716981132], "iscrowd": 0, "area": 23041.60025, "rle": {"size": [424, 640], "counts": "bb\\1n0RB7H7J6O2M2O1N2O2N1N2O2M2O1O2M2O1N3N1N20001O0SHeLl5\\3oIjLP6W3jIoLU6Q3eIWMY6i2bIlMm5V2mIoMS6Q2kIQNX4aNbL^3TOSNX4dN`L[3UOSNZ4gN]LV3WOUNZ4kN[LS3VOTN^4nNXLU3QOoMe4QOVLX3lNiMm4TOSLS5k3RKRLn4m3UKQLk4m3YKPLh4o3\\KnKd4P4`KnK`4P4dKmK]4R4fKkKZ4T4kKhKV4W4lKgKU4W4nKgKS4X4PLeKQ4[4PLdKP4\\4QLbKP4^4]200000000001O00000000000000000000000000000000001N100000000000001O001O0PJ^KZ3b4dL`K\\3a4aLaK_3_4_LdK`3\\4]LgKc3Z4YLiKe3Y4YLjK`3]4]LeKY3e4gL[Kk2S5UMnJb2[5]MeJ`2^5`MbJ]2a5cM`JZ2c5dM^JY2e5gM[JV2m5eMTJX2T6bMlI[2c6YM]Id2g6[MZIa2j6^MVI`2m6VM\\Ig2h6gLVH3X1Y2jNRNl7_OhIm1WOYNT7GgIm1a9L3M3N210000000N2M4L3L4M3M3M3M3M3L4M3M3M4L3M4KR\\U5"}, "label": "the surfer who is standing , holding his board upright , looking to the right ,"}]} {"id": 513748, "image": "COCO_train2014_000000513748.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skate board which is held by a lady\"."}], "task": "refering", "answer_template": "The \"a skate board which is held by a lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 170, 171, 172, 192, 193, 194, 195, 196, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 287, 288, 306, 307, 308, 309, 310, 329, 330, 331, 332, 333], "bbox": [0.314125, 0.44719339622641513, 0.5448906250000001, 0.9865094339622641], "iscrowd": 0, "area": 24085.559249999995, "rle": {"size": [424, 640], "counts": "kcc2>Z<`0L4M3M3L4M3L4M3L4M3L4M3L4M3L5L3L4M3L4M3L4M3M3L4L4M3L4L4M3L4L4L4M3L4N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2N2N2N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1000000000000O1000000000000O100000000000000O100000000000000O10000000000002N2N1O2N2M3L4K5L4K5L4K5L4K5L3L5L4K5L4K5XOdG^L`8[3k0I7I7I7I7I7I7I8H7I7I8F9C>B=CnWh3"}, "label": "a skate board which is held by a lady"}]} {"id": 513748, "image": "COCO_train2014_000000513748.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"surfboard with a black and yellow design on it\"."}], "task": "refering", "answer_template": "The \"surfboard with a black and yellow design on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [148, 149, 170, 171, 172, 192, 193, 194, 195, 196, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 287, 288, 306, 307, 308, 309, 310, 329, 330, 331, 332, 333], "bbox": [0.314125, 0.44719339622641513, 0.5448906250000001, 0.9865094339622641], "iscrowd": 0, "area": 24085.559249999995, "rle": {"size": [424, 640], "counts": "kcc2>Z<`0L4M3M3L4M3L4M3L4M3L4M3L4M3L5L3L4M3L4M3L4M3M3L4L4M3L4L4M3L4L4L4M3L4N2N2N2N2M3N2N2N2N2N2N2M3N2N2N2N2N2M3N2N2N2N2N2N2N2O1O1N2O1N2O1O1N2O1O1N2O1N2O1O1N2O1000000000000O1000000000000O100000000000000O100000000000000O10000000000002N2N1O2N2M3L4K5L4K5L4K5L4K5L3L5L4K5L4K5XOdG^L`8[3k0I7I7I7I7I7I7I8H7I7I8F9C>B=CnWh3"}, "label": "surfboard with a black and yellow design on it"}]} {"id": 513748, "image": "COCO_train2014_000000513748.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person surfing\"."}], "task": "refering", "answer_template": "The \"the person surfing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [153, 154, 155, 175, 176, 177, 198, 199, 200, 201, 221, 222, 223, 244, 245, 267], "bbox": [0.600984375, 0.39422169811320756, 0.7600625, 0.7898584905660376], "iscrowd": 0, "area": 5370.963099999999, "rle": {"size": [424, 640], "counts": "^go41U=2O2N1N3N1O1N3N1O1N3N1]MHUH9j7JSH8k7LQH5n7NnG5P8NmG3R80kG1T83hGOV84gGMX86dGL[87bGK\\89`GG`8<]GCd8`0YG_Oh8d0TG\\Om8h0nFWOT9l0hFTOY9o0cFQO^9R1^FnNc9V1XFiNj9Z20000000000001O000O2O0VNSF:m9EXF7i9H[F4e9L_F0a9OeFL\\93QG@o8`0^GSOc8l0l1O101O0O101O0O1010O1O2O1N2N101N5L6I8I6J3M3MdNSE@1`0i:OWEA1b0e:L\\EA<9U:5aEAe02g9lE^Od03\\9a0RFZOd04X9c0VFXOc05S9d0]FWOa04o8f0bFUO`01Q9i0`FWO?MR9k0bFXOd:g0]EZOa:f0`EZO`:f0`E[O^:f0bE[O]:d0dE\\O[:e0fE[OY:e0gEZO[:e0eEZO\\:f0dEZO\\:e0gEXOZ:e0kEXOW:e0W1M3N2M3L5K4LZco1"}, "label": "the person surfing"}]} {"id": 415447, "image": "COCO_train2014_000000415447.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"oranges stacked\"."}], "task": "refering", "answer_template": "The \"oranges stacked\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 60, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0017656249999999998, 0.16470833333333335, 1.0, 0.9882291666666667], "iscrowd": 0, "area": 219395.25460000001, "rle": {"size": [480, 640], "counts": "bc0 in this image.", "objects": [{"patches": [57, 58, 60, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.0017656249999999998, 0.16470833333333335, 1.0, 0.9882291666666667], "iscrowd": 0, "area": 219395.25460000001, "rle": {"size": [480, 640], "counts": "bc0 in this image.", "objects": [{"patches": [174, 175, 195, 196, 197, 198, 217, 218], "bbox": [0.45725, 0.4176689976689977, 0.67796875, 0.6596969696969697], "iscrowd": 0, "area": 1072.3543499999982, "rle": {"size": [429, 640], "counts": "iPk31\\=0O2N1O101N1O1O01O0000010O00000010O0001O0001O01O0000010O00000010O000001O01O01O0000010O00001O01O0001O00010O00000010O0001O0001O01O000010O0000010O00011N1O1O101N1O100O1O1O2O0O1ODTC9l in this image.", "objects": [{"patches": [98, 99, 100, 121, 122, 123, 124, 145, 146, 147, 168, 169, 170, 171, 191, 192, 193, 194, 214, 215, 216, 217, 237, 238, 239, 240, 260, 261, 262, 263, 283, 284, 286, 287, 306, 307, 309, 310, 329, 330, 332, 333], "bbox": [0.28446875, 0.27032634032634034, 0.50840625, 0.9746153846153847], "iscrowd": 0, "area": 18518.112999999994, "rle": {"size": [429, 640], "counts": "j[\\21[=5L4K4M4K4M4K2O1N2O1N2O0O2O1N2O1O001O1]IgNP1Z1mNnNm0R1POVOk0k0RO]O^LYOb2[1l0D\\LSOg2Z1j0KWLmNo2P2O[OkLgNV3Q2hMhM2h1hNaN^3S2aMjM3j1gN[Nf3S2ZMnM4l1eNTNn3U2RMQN6m1cNoMV4V2jLSN8^2o2AcLWN9Z2V3A[LZN9X2]3ASL]N;T2c3CjK^N>R2i3CbKaN`0o1n3X1lKkNU4W1fKkN[4V1`KmNa4o41O1O1O1O3M3M3M2M4M0cMSKYLn4f3YKTLg4k3`KnKa4Q4gKgK[4W4lKcKT4[4UL]Kl3b4[LXKe3g4bLRK_3m4iLlJW3S5PMfJR3X5d2L3N2N2N2N2M3N3M2N2N2N1N2O2N1O1O2M2O2N1O1\\LkFo2W9oLkFn2W9QMkFm2W9QMlFl2[9mLgFQ3l9005K4M4K5K4L5K3N0O1O1O1O00100O1O1O1O100]HgKo5Z4nIiKQ6X4lIkKS6Z4eIjKZ6]4]IeKf6T2bHF?XNR7l1eH01UN]7d1hHP1[7iNkHT1X7eNoHX1T7bNQI]1R7\\NTIa1o6XNWIf1o6QNWIm1T7cMSIZ2R9N3N2M2N3M3N1N3M3N1N3M3L3M4L3M4M3L3M4L4L4L4M2M5KebS4"}, "label": "a man with a baseball mit throwing a baseball to someone"}]} {"id": 317149, "image": "COCO_train2014_000000317149.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black jacket begins to throw a ball\"."}], "task": "refering", "answer_template": "The \"a man wearing a black jacket begins to throw a ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 121, 122, 123, 124, 145, 146, 147, 168, 169, 170, 171, 191, 192, 193, 194, 214, 215, 216, 217, 237, 238, 239, 240, 260, 261, 262, 263, 283, 284, 286, 287, 306, 307, 309, 310, 329, 330, 332, 333], "bbox": [0.28446875, 0.27032634032634034, 0.50840625, 0.9746153846153847], "iscrowd": 0, "area": 18518.112999999994, "rle": {"size": [429, 640], "counts": "j[\\21[=5L4K4M4K4M4K2O1N2O1N2O0O2O1N2O1O001O1]IgNP1Z1mNnNm0R1POVOk0k0RO]O^LYOb2[1l0D\\LSOg2Z1j0KWLmNo2P2O[OkLgNV3Q2hMhM2h1hNaN^3S2aMjM3j1gN[Nf3S2ZMnM4l1eNTNn3U2RMQN6m1cNoMV4V2jLSN8^2o2AcLWN9Z2V3A[LZN9X2]3ASL]N;T2c3CjK^N>R2i3CbKaN`0o1n3X1lKkNU4W1fKkN[4V1`KmNa4o41O1O1O1O3M3M3M2M4M0cMSKYLn4f3YKTLg4k3`KnKa4Q4gKgK[4W4lKcKT4[4UL]Kl3b4[LXKe3g4bLRK_3m4iLlJW3S5PMfJR3X5d2L3N2N2N2N2M3N3M2N2N2N1N2O2N1O1O2M2O2N1O1\\LkFo2W9oLkFn2W9QMkFm2W9QMlFl2[9mLgFQ3l9005K4M4K5K4L5K3N0O1O1O1O00100O1O1O1O100]HgKo5Z4nIiKQ6X4lIkKS6Z4eIjKZ6]4]IeKf6T2bHF?XNR7l1eH01UN]7d1hHP1[7iNkHT1X7eNoHX1T7bNQI]1R7\\NTIa1o6XNWIf1o6QNWIm1T7cMSIZ2R9N3N2M2N3M3N1N3M3N1N3M3L3M4L3M4M3L3M4L4L4L4M2M5KebS4"}, "label": "a man wearing a black jacket begins to throw a ball"}]} {"id": 5862, "image": "COCO_train2014_000000005862.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog with brown fur , with its head up , laying on a gray sheet\"."}], "task": "refering", "answer_template": "The \"a dog with brown fur , with its head up , laying on a gray sheet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 151, 152, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 226, 227, 228, 250], "bbox": [0.5515625, 0.23902083333333335, 0.9383437499999999, 0.60225], "iscrowd": 0, "area": 21063.518799999998, "rle": {"size": [480, 640], "counts": "SdU51n>2N2N2N2N2O1O1N2O1O1O2N1O1O1N101O00000000000O101O001O001O0O101O001N101O0O2O000O2O001N101O000O10000O10000O10000O10000O10000O10000O10000O2O1O2M2O2M2O2N1N3N1O0O20O0001O01O01O000010O0001O100O1O1O100O1O1O100O1O1O1O101N1O2N100O001O010O00001O010O00001O010O00001O0001N1O11O00100O001O1O0WOZMWEh2h:\\MSEe2l:`MnDc2Q;aMiDb2U;e001N2N101N2O0O2O0O2O1N1O2O0O2O0O2O000010O0001O01O01O00010O000010O01O00010O000010O0001O000010O0001O00001O001O1O2O0O2N1O2N1O2N1O2N1O2N100O2N1O0000000000000001VNWF[Ni9a1]F]Nc9`1aF_N_9^1fF`NZ9\\1lFbNT9[1PGdNP9Y1UGeNk8W1[GfNf8KTF7Z1Lc8L[FNW14`8KaFFU1<\\8KhF_OP1d0Z8KnFVOn0m0U8KPI3R7KQI2Q7LRI2P7KRI3P7KSI2o6KUI3l6KVI3l6KWI2k6LWI3j6KYI2j6KYI3k:N__b0"}, "label": "a dog with brown fur , with its head up , laying on a gray sheet"}]} {"id": 5862, "image": "COCO_train2014_000000005862.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown dog\"."}], "task": "refering", "answer_template": "The \"brown dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 109, 110, 111, 112, 113, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 151, 152, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 200, 201, 202, 203, 204, 226, 227, 228, 250], "bbox": [0.5515625, 0.23902083333333335, 0.9383437499999999, 0.60225], "iscrowd": 0, "area": 21063.518799999998, "rle": {"size": [480, 640], "counts": "SdU51n>2N2N2N2N2O1O1N2O1O1O2N1O1O1N101O00000000000O101O001O001O0O101O001N101O0O2O000O2O001N101O000O10000O10000O10000O10000O10000O10000O10000O2O1O2M2O2M2O2N1N3N1O0O20O0001O01O01O000010O0001O100O1O1O100O1O1O100O1O1O1O101N1O2N100O001O010O00001O010O00001O010O00001O0001N1O11O00100O001O1O0WOZMWEh2h:\\MSEe2l:`MnDc2Q;aMiDb2U;e001N2N101N2O0O2O0O2O1N1O2O0O2O0O2O000010O0001O01O01O00010O000010O01O00010O000010O0001O000010O0001O00001O001O1O2O0O2N1O2N1O2N1O2N1O2N100O2N1O0000000000000001VNWF[Ni9a1]F]Nc9`1aF_N_9^1fF`NZ9\\1lFbNT9[1PGdNP9Y1UGeNk8W1[GfNf8KTF7Z1Lc8L[FNW14`8KaFFU1<\\8KhF_OP1d0Z8KnFVOn0m0U8KPI3R7KQI2Q7LRI2P7KRI3P7KSI2o6KUI3l6KVI3l6KWI2k6LWI3j6KYI2j6KYI3k:N__b0"}, "label": "brown dog"}]} {"id": 5862, "image": "COCO_train2014_000000005862.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white dog with a blue collar on a bed\"."}], "task": "refering", "answer_template": "The \"a white dog with a blue collar on a bed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 310, 326, 327, 333], "bbox": [0.01509375, 0.4791666666666667, 0.5221250000000001, 0.847375], "iscrowd": 0, "area": 36906.529200000004, "rle": {"size": [480, 640], "counts": "iP56c>7O1O10O10O10000O01000O1O10O01O100O00100O1O0010OM4K4K6H7J7J5J6O2N1O2O0O2O00O0O2O1O1N2O1O1011N1O2N1QDUNi:l1QEZNm:g1PE]Nn:e1nD_NP;b1aDnN^;[2N2N2N100O1O100O001O100O1O100O1O010O1O101N1O100O2O0O100O101N100O2O1N101N2O1O1O1N2O001O1O1O2N1O2M3N1O2N1O1O1O1O0O2O001O1O001O001O001O001O001O0O101O001O001O00001O001O0000000000O10000000000000000000000O1L4L4ZOaFgKc9m3n0M3N20000O10000O10000O100000000000000000eETLb9l3[FWLe9i3WF[Li9e3TF]Lm9d3oE_LQ:a3nE`LR:`3mEaLS:_3mEaLS:`3kEaLU:_3jEbLV:^3iEcLW:]3iEcLW:^3gEcLY:]3gEcLY:]3fEcL[:]3eEcL[:Q400000000000O10000O100O10000O100O10000O10000O10000O10000O10000O10000O1000000O100000001N101O1O1O0O2O1O1O0O2O1O001O1N1O2M3M2O2N2O0O2N2O0O2N2O0N3M3N1N3N2O001O001O1O010O001O1O001O001N2N101N2O0O2N2O0O2O1N101N2N101N2O0O2O1O1N2O1NiCPOa:o0PEFk:8gD9X;h1010O01ON2O2M4M2M4M2M4M2M4I6K6SNcDNb;GWV_4"}, "label": "a white dog with a blue collar on a bed"}]} {"id": 5862, "image": "COCO_train2014_000000005862.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white dog laying on couch\"."}], "task": "refering", "answer_template": "The \"white dog laying on couch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 302, 303, 304, 305, 306, 307, 310, 326, 327, 333], "bbox": [0.01509375, 0.4791666666666667, 0.5221250000000001, 0.847375], "iscrowd": 0, "area": 36906.529200000004, "rle": {"size": [480, 640], "counts": "iP56c>7O1O10O10O10000O01000O1O10O01O100O00100O1O0010OM4K4K6H7J7J5J6O2N1O2O0O2O00O0O2O1O1N2O1O1011N1O2N1QDUNi:l1QEZNm:g1PE]Nn:e1nD_NP;b1aDnN^;[2N2N2N100O1O100O001O100O1O100O1O010O1O101N1O100O2O0O100O101N100O2O1N101N2O1O1O1N2O001O1O1O2N1O2M3N1O2N1O1O1O1O0O2O001O1O001O001O001O001O001O0O101O001O001O00001O001O0000000000O10000000000000000000000O1L4L4ZOaFgKc9m3n0M3N20000O10000O10000O100000000000000000eETLb9l3[FWLe9i3WF[Li9e3TF]Lm9d3oE_LQ:a3nE`LR:`3mEaLS:_3mEaLS:`3kEaLU:_3jEbLV:^3iEcLW:]3iEcLW:^3gEcLY:]3gEcLY:]3fEcL[:]3eEcL[:Q400000000000O10000O100O10000O100O10000O10000O10000O10000O10000O10000O1000000O100000001N101O1O1O0O2O1O1O0O2O1O001O1N1O2M3M2O2N2O0O2N2O0O2N2O0N3M3N1N3N2O001O001O1O010O001O1O001O001N2N101N2O0O2N2O0O2O1N101N2N101N2O0O2O1O1N2O1NiCPOa:o0PEFk:8gD9X;h1010O01ON2O2M4M2M4M2M4M2M4I6K6SNcDNb;GWV_4"}, "label": "white dog laying on couch"}]} {"id": 333546, "image": "COCO_train2014_000000333546.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"right duplicate image of tennis player\"."}], "task": "refering", "answer_template": "The \"right duplicate image of tennis player\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 61, 62, 63, 64, 84, 85, 86, 87, 88, 107, 108, 109, 110, 111, 113, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 227, 245, 246, 249, 250, 268, 269, 272, 273, 274, 291, 292, 296, 297, 314, 315, 319, 320, 321, 336, 337, 338, 343], "bbox": [0.62946875, 0.1011520737327189, 0.974609375, 0.9060829493087557], "iscrowd": 0, "area": 33001.31135000001, "rle": {"size": [434, 640], "counts": "mU[51`=1O1O1O2N1O1O1O1O1O1O1N2O1O1N2iF_O\\5b0dJ1ZO]OE0Kd0U12XOEWO07:Z12UO0iNOb01`12QO1dN8e0Ge12PO1^Na0h0\\Ok14kN2[Nh0i0TOP24iN2\\Nl0f0oNU25eN3\\NP1d0jN[24cN4ZNU1c0dN_26`N4ZNY1a0_Ne25]N7YN^1=WNm26WN=XNg14hM\\36SN_5m1bJnMa5S2_JiMe5V2]JcMi5]2WJ^Mn5b2SJXMR6g2PJSMU6m2kIoLY6Q3hIkL[6T3gIiL[6W3eIfL]6Z3eIbL^6]3cI`L`6_3j100O1O1O100O00jLjLRLU3n3QMmKo2Q4WMkKh2T4_MgK`2X4gMcKY2[4mMaKR2^4UN]Kj1b4\\NZKc1e4cNVK^1i4fNTKY1k4lNPKU1o4QOlJP1R5UOiJl0V5ZOeJf0[5^O`Jc0_5B\\J`0c5EXJ;g5JTJ7l5MPJ4n52lIOT65gILX6h3O001N101N101N101010O1O010O1O010O000010O01O01O01O01O01O01O01O01O01O010O00010O00010OPJ`In4`6oJcIR5_71N1O2O1N101N1O2O0O2O1N1O2O0O2N10ZLXHb1f7^N[Hc1d7\\N^Hc1a7]NaHc1]7]NdHc1\\7\\NfHd1X7\\NjHd1Y7WNhHi1m7aMUH_2i7aMXH`2g7_M[H`2d7`M`H^2_7aMfH\\2X7dMnHX2Q7fMVIV2h6jM^IQ2b6nMfIl1X6TNRJd1m5[NXJb1f5^N^J_1a5aNcJ]1\\5bNgJ]1W5cNmJ[1R5dNRKZ1l4fNXKW1h4hN[KW1c4iN`KV1_4iNdKV1Z4iNkKU1T4jNoKT1P4lNSLS1l3lNWLS1g3mN\\LR1c3mN_LS1_3mNcLS1\\3lNfLT1X3lNiLV1U3iNmLW1Q3iNQMW1m2iNTMX1k2gNWMY1g2gN[MY1d2fN]M[1a2dNbM\\1[2eNgM[1W2eNjM[1T2fNnMY1Q2gNQNX1m1iNTNW1j1jNXNU1g1kN[NT1c1mN_NR1`1nNaNR1]1oNeNP1Z1POhNo0V1ROkNn0S1SOoNl0P1TOROk0l0UOVOk0i0UOYOj0e0WO]Oh0b0XO_Oh0?YOCe0<\\OFc09]OIb05_OM?3AO>OC3XOAj0>VOAm0=SOAP1=ROAP1>QO@R1=QO@Q1>RO^OQ1a0RO[OQ1b0RO[OP1d0P6N3N1M3M4L3M3La]6"}, "label": "right duplicate image of tennis player"}]} {"id": 333546, "image": "COCO_train2014_000000333546.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"lady who is playing in the right side of the image\"."}], "task": "refering", "answer_template": "The \"lady who is playing in the right side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [40, 61, 62, 63, 64, 84, 85, 86, 87, 88, 107, 108, 109, 110, 111, 113, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180, 199, 200, 201, 202, 203, 222, 223, 224, 225, 226, 227, 245, 246, 249, 250, 268, 269, 272, 273, 274, 291, 292, 296, 297, 314, 315, 319, 320, 321, 336, 337, 338, 343], "bbox": [0.62946875, 0.1011520737327189, 0.974609375, 0.9060829493087557], "iscrowd": 0, "area": 33001.31135000001, "rle": {"size": [434, 640], "counts": "mU[51`=1O1O1O2N1O1O1O1O1O1O1N2O1O1N2iF_O\\5b0dJ1ZO]OE0Kd0U12XOEWO07:Z12UO0iNOb01`12QO1dN8e0Ge12PO1^Na0h0\\Ok14kN2[Nh0i0TOP24iN2\\Nl0f0oNU25eN3\\NP1d0jN[24cN4ZNU1c0dN_26`N4ZNY1a0_Ne25]N7YN^1=WNm26WN=XNg14hM\\36SN_5m1bJnMa5S2_JiMe5V2]JcMi5]2WJ^Mn5b2SJXMR6g2PJSMU6m2kIoLY6Q3hIkL[6T3gIiL[6W3eIfL]6Z3eIbL^6]3cI`L`6_3j100O1O1O100O00jLjLRLU3n3QMmKo2Q4WMkKh2T4_MgK`2X4gMcKY2[4mMaKR2^4UN]Kj1b4\\NZKc1e4cNVK^1i4fNTKY1k4lNPKU1o4QOlJP1R5UOiJl0V5ZOeJf0[5^O`Jc0_5B\\J`0c5EXJ;g5JTJ7l5MPJ4n52lIOT65gILX6h3O001N101N101N101010O1O010O1O010O000010O01O01O01O01O01O01O01O01O01O010O00010O00010OPJ`In4`6oJcIR5_71N1O2O1N101N1O2O0O2O1N1O2O0O2N10ZLXHb1f7^N[Hc1d7\\N^Hc1a7]NaHc1]7]NdHc1\\7\\NfHd1X7\\NjHd1Y7WNhHi1m7aMUH_2i7aMXH`2g7_M[H`2d7`M`H^2_7aMfH\\2X7dMnHX2Q7fMVIV2h6jM^IQ2b6nMfIl1X6TNRJd1m5[NXJb1f5^N^J_1a5aNcJ]1\\5bNgJ]1W5cNmJ[1R5dNRKZ1l4fNXKW1h4hN[KW1c4iN`KV1_4iNdKV1Z4iNkKU1T4jNoKT1P4lNSLS1l3lNWLS1g3mN\\LR1c3mN_LS1_3mNcLS1\\3lNfLT1X3lNiLV1U3iNmLW1Q3iNQMW1m2iNTMX1k2gNWMY1g2gN[MY1d2fN]M[1a2dNbM\\1[2eNgM[1W2eNjM[1T2fNnMY1Q2gNQNX1m1iNTNW1j1jNXNU1g1kN[NT1c1mN_NR1`1nNaNR1]1oNeNP1Z1POhNo0V1ROkNn0S1SOoNl0P1TOROk0l0UOVOk0i0UOYOj0e0WO]Oh0b0XO_Oh0?YOCe0<\\OFc09]OIb05_OM?3AO>OC3XOAj0>VOAm0=SOAP1=ROAP1>QO@R1=QO@Q1>RO^OQ1a0RO[OQ1b0RO[OP1d0P6N3N1M3M4L3M3La]6"}, "label": "lady who is playing in the right side of the image"}]} {"id": 333546, "image": "COCO_train2014_000000333546.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman on the left\"."}], "task": "refering", "answer_template": "The \"the woman on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 30, 50, 51, 52, 53, 73, 74, 75, 76, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 211, 212, 214, 215, 234, 235, 237, 238, 239, 257, 258, 261, 262, 280, 281, 285, 286, 303, 304, 308, 309, 310, 325, 326, 327, 332], "bbox": [0.152046875, 0.09343317972350229, 0.4880625, 0.9058294930875576], "iscrowd": 0, "area": 31030.189500000008, "rle": {"size": [434, 640], "counts": "UmY12_=1O1O2N1O1O2N1O1O1O2O0eJCTJ5Y19U1BPO4^N>Q1N_1BoN6]Ng0i0Bj1ClN9]No0`0WOV2BkN<[NX18lN`2BjN>ZNb1O_Nm2BgNa0YNk1GSNX3CeNc0XNm1FoM[3CdNe0VNn1HkM]3CcNh0TNn1IhM_3DaNj0SNn1IfMc3C^Nj0UNQ2GbMf3EZNi0YNT2D_Mh3EYNh0[NX2A\\Mj3FWNf0^N]3[3oKTNd0aN_3[3nKQNd0cN_3\\3oKSMB]OP14a3[3nKQM]1Dg2Z3nKmL]1If2Z3oKiL]1Lf2[3nKdL^10f2\\3mKaL_12f2\\3mK]L`16e2\\3mKZL`19d2]3fNbL\\1^3cNaL^1_3bN`L`1_3`N`Lb1_3^N`LZNLl0d3j0_L[N9`0W3U1_L[Ng04j2a1^L\\NT1H]2l1]L^Na1[OR2W2\\L_Nn1oNe1R5gNcJW1]5QO\\Jo0d5TOZJk0f5WOXJi0h5ZOVJd0k5^OTJa0k5CRJ=n5FPJ9P6InI7R6LlI2U60jIOU65hIKX68fIGZ6;dIE\\6>bI@_6b0`I]O_6k300000001N2O1O100000000000000O10000000001O0000000O10000001O2N1O2N1OmJQIe3m6WLXIi3g6SL]In3f6kK]IW4e6aK_I`4d6XK`Ij4a6oJcIR5_71N1O2N1O2N101N1O2N1O2O0O2N1O1O20O100O2O0O2O0O2O0O101N101N1`KVHV3k7gL]HR3d7lLdHn2\\7PMmHh2U7UMoHh2R7UMTIh2j6XMZIe2e6[M_Ic2_6]MfI_2Y6aMkI]2T6aMQJ\\2n5dMWJY2g5gM]JV2b5jMbJT2\\5lMiJP2V5oMoJo1o4QNUKl1j4TN[Ki1c4WNaKf1^4ZNfKd1X4[NnKa1Q4_NSL_1k3aNXL]1g3cNZL^1d3bN]L^1b3aNaL^1_3aNbL`1\\3`NeL`1Z3`NgLa1W3_NkL`1T3`NmLa1Q3^NQMb1n2^NTMb1j2^NWMb1h2^NZMb1d2^N_M`1`2_NcMa1Z2`NhM_1U2cNnM\\1o1eNSNZ1j1hNYNW1d1iN_NV1^1lNeNR1Y1oNiNP1T1ROnNm0Q1SOROk0m0TOVOk0i0UOZOi0e0WO]Oh0b0XOAe0?[OCd0<[OGd07]OLa03_OO`00@3>LB6=IB;

;AEa0:^OFc0:\\OFe0:YOFi0:VOFk0:UOEl0;SOEn0:ROFo0:POER1;mNET1;kNEV1;iNDY1cN_O`1a0_N]Od1`0^N^Oe1?]N_Of1>\\N@g1=[NAh1;[NCh1:j5M3M3ModZ4"}, "label": "the woman on the left"}]} {"id": 497386, "image": "COCO_train2014_000000497386.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a train numbered 332010\"."}], "task": "refering", "answer_template": "The \"a train numbered 332010\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 51, 52, 53, 54, 55, 56, 57, 58, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 305, 306, 307, 325], "bbox": [0.131625, 0.10075117370892019, 0.76284375, 0.9501877934272301], "iscrowd": 0, "area": 83572.08395, "rle": {"size": [426, 640], "counts": "^fS12W=1TNN_F4_92[FOe96UFKk99PFIn9=lEDT:a0fEAY:c0bE^O]:h0]EYOc:l0WEVOh:o0REROm:S1nDoNQ;e1O101O0`FgMn7[2jGlMV8T2cGRN^8n1\\GXNc8j1UG]Nk8c1PGbNP9_1nFbNQ9_1oF_NS9a1lF\\NW9f1gFVN^9j1bFQNc9P2[FmMh9T2WFhMn9X2RFcMS:^2kEbMU:_2kEaMU:`2iE`MW:a2hE`MX:h210O10ObMRMbJn2X5_M`Jb2Z5kM_JT2]5WN]Ji1]5dN[J\\1`5QOYJo0b5]OVJd0g5FRJ9l51mIOQ6:gIGV6c0cI\\O[6k3N1N3M3N0O1O100O3M4M2M3M4M2M3N3L3M4M2M3M4M2M3M4M200O2N1O1O2N100O2N1O1O2N100O1O2N1O1O2N100O2N1O1O1O2O0O1O2N1O101O0O101O000O101O0O101O000O2O00000O2O000O2O00001N10001O0O10001N10001O0O101O000O1000O010O1O1O010O1O1O010O1O010O1O1O010O1O1O100O1O2O0O1O1O100O1O1O100O1O2N100O1O100O1O1O100O1O2N100O1O100O1O1O100O1O2N100O1O100O1O1O100O2N1O100O1O2N101N1O2O1N1O2N101N1O2N101N1O2O1N1O2N101N1O2N101N2N101N1O2N101N1O2N2O0O2N1O2O0O2N101N1O2N2O0O2N1O2O0O2N101N2N1O2O1N1O2N2O0O2N2O0O2N2N101N2N1O2O1N2N1O2O1N101O1O001O10O01O1O001O1O001O1O001O1O1N101O1O001O1O001O1\\Oc0O2O1N101N2N101N2N101N2O1N1O2O1N101N2N101N2N101N2O0O2N2O0O2N2O1N101N2N1mMQFk0P:QOW2L3N3N2O001N2O001N2O0O2O1O0O2O12N3L3N000O1000O010000fCf0b:Q11N2O0O2N2O0O2O1N2O0O2I7I6J7I7I6J7I7I]Ro1"}, "label": "a train numbered 332010"}]} {"id": 497386, "image": "COCO_train2014_000000497386.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bullet train with a japanese sun on the side\"."}], "task": "refering", "answer_template": "The \"a bullet train with a japanese sun on the side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 33, 51, 52, 53, 54, 55, 56, 57, 58, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 256, 257, 258, 259, 260, 261, 262, 263, 264, 279, 280, 281, 282, 283, 284, 285, 286, 302, 303, 305, 306, 307, 325], "bbox": [0.131625, 0.10075117370892019, 0.76284375, 0.9501877934272301], "iscrowd": 0, "area": 83572.08395, "rle": {"size": [426, 640], "counts": "^fS12W=1TNN_F4_92[FOe96UFKk99PFIn9=lEDT:a0fEAY:c0bE^O]:h0]EYOc:l0WEVOh:o0REROm:S1nDoNQ;e1O101O0`FgMn7[2jGlMV8T2cGRN^8n1\\GXNc8j1UG]Nk8c1PGbNP9_1nFbNQ9_1oF_NS9a1lF\\NW9f1gFVN^9j1bFQNc9P2[FmMh9T2WFhMn9X2RFcMS:^2kEbMU:_2kEaMU:`2iE`MW:a2hE`MX:h210O10ObMRMbJn2X5_M`Jb2Z5kM_JT2]5WN]Ji1]5dN[J\\1`5QOYJo0b5]OVJd0g5FRJ9l51mIOQ6:gIGV6c0cI\\O[6k3N1N3M3N0O1O100O3M4M2M3M4M2M3N3L3M4M2M3M4M2M3M4M200O2N1O1O2N100O2N1O1O2N100O1O2N1O1O2N100O2N1O1O1O2O0O1O2N1O101O0O101O000O101O0O101O000O2O00000O2O000O2O00001N10001O0O10001N10001O0O101O000O1000O010O1O1O010O1O1O010O1O010O1O1O010O1O1O100O1O2O0O1O1O100O1O1O100O1O2N100O1O100O1O1O100O1O2N100O1O100O1O1O100O1O2N100O1O100O1O1O100O2N1O100O1O2N101N1O2O1N1O2N101N1O2N101N1O2O1N1O2N101N1O2N101N2N101N1O2N101N1O2N2O0O2N1O2O0O2N101N1O2N2O0O2N1O2O0O2N101N2N1O2O1N1O2N2O0O2N2O0O2N2N101N2N1O2O1N2N1O2O1N101O1O001O10O01O1O001O1O001O1O001O1O1N101O1O001O1O001O1\\Oc0O2O1N101N2N101N2N101N2O1N1O2O1N101N2N101N2N101N2O0O2N2O0O2N2O1N101N2N1mMQFk0P:QOW2L3N3N2O001N2O001N2O0O2O1O0O2O12N3L3N000O1000O010000fCf0b:Q11N2O0O2N2O0O2O1N2O0O2I7I6J7I7I6J7I7I]Ro1"}, "label": "a bullet train with a japanese sun on the side"}]} {"id": 87792, "image": "COCO_train2014_000000087792.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark brown donut\"."}], "task": "refering", "answer_template": "The \"a dark brown donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 351, 352, 353, 354, 355, 356, 357, 358, 375, 376, 377, 378, 379, 380], "bbox": [0.236578125, 0.4700208333333334, 0.6526875000000001, 0.9890208333333333], "iscrowd": 0, "area": 51906.2389, "rle": {"size": [480, 640], "counts": "ZSW25k=]1B?A>A?B6J4L3L4M3M3L5L3M3L4M3M3M4K4M3M3L4M4L2N2M2O2N2M3N1O2M3N1O2N2M2O2N2O001N2O1O0O2O1O0O2O1O0O2O1O0O2O1O1N101O1N101O1N101O1N101O1N2O001N2O001N2O001N10000O101O0O10000O100000000O100000001N100000000O100000000O1000001O0O100000000O100000000O10001O000O100000000O100000000O10001O00000000000001O000000000010O0000000001O00000000001O0001O0001O0O100O1O101N100O1O100O2N100O1O101N2N101N2N101N2O0O2N2O0O2N2O0O2N2O0O2N2O0O2N2O0O2N101N2O0O2N2O0O2N2O0O2N2O1N3M3N2M2N3N2L4M2N3M3M3L3N3M3M3M2M4M3M3M2N4L4K5L4L4L4L4K5L4L4L5K4K5L4L5K6J6J7H7J6J7I6J6I7JmeW3"}, "label": "a dark brown donut"}]} {"id": 87792, "image": "COCO_train2014_000000087792.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dark brown glazed donut\"."}], "task": "refering", "answer_template": "The \"dark brown glazed donut\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [191, 192, 193, 194, 195, 196, 213, 214, 215, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 351, 352, 353, 354, 355, 356, 357, 358, 375, 376, 377, 378, 379, 380], "bbox": [0.236578125, 0.4700208333333334, 0.6526875000000001, 0.9890208333333333], "iscrowd": 0, "area": 51906.2389, "rle": {"size": [480, 640], "counts": "ZSW25k=]1B?A>A?B6J4L3L4M3M3L5L3M3L4M3M3M4K4M3M3L4M4L2N2M2O2N2M3N1O2M3N1O2N2M2O2N2O001N2O1O0O2O1O0O2O1O0O2O1O0O2O1O1N101O1N101O1N101O1N101O1N2O001N2O001N2O001N10000O101O0O10000O100000000O100000001N100000000O100000000O1000001O0O100000000O100000000O10001O000O100000000O100000000O10001O00000000000001O000000000010O0000000001O00000000001O0001O0001O0O100O1O101N100O1O100O2N100O1O101N2N101N2N101N2O0O2N2O0O2N2O0O2N2O0O2N2O0O2N2O0O2N101N2O0O2N2O0O2N2O0O2N2O1N3M3N2M2N3N2L4M2N3M3M3L3N3M3M3M2M4M3M3M2N4L4K5L4L4L4L4K5L4L4L5K4K5L4L5K6J6J7H7J6J7I6J6I7JmeW3"}, "label": "dark brown glazed donut"}]} {"id": 87792, "image": "COCO_train2014_000000087792.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a doughnut without a hole sitting below a light colored doughnut with a hole\"."}], "task": "refering", "answer_template": "The \"a doughnut without a hole sitting below a light colored doughnut with a hole\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 361, 362, 363, 364, 365, 366, 385, 386, 387, 388], "bbox": [0.6505625, 0.5393333333333333, 0.9741562500000001, 0.9595416666666666], "iscrowd": 0, "area": 33511.7129, "rle": {"size": [480, 640], "counts": "X[S61j>:G:F:F:F9G:F7K3L4M3L4M3L4L5L3L4M3L4M3L4L4M4K4N2N2O1N2N2O1N3N1N2N2O1N2N2O1N3M2O1N2O1N2N2O1N2O2M2O1N100O2O0O101N100O2N100O2O0O2O0O101N100O2O00001O000O2O00001O001O0O101O000000001O0O1000000000000O1000000000000000001O00000001O0000000000000000000000001O01O00000000000000000001O00000000000001O00O100O2N100O1O100O100O1O2O0O2O0O2N2O0O2N101N2O0O2N101N2N101N101N2N101N1O2O2M2O2M2M4K4L5K5K4L5K4L5K5K4L5K4M4K5K4L5J5J7I6J7I7IWe7"}, "label": "a doughnut without a hole sitting below a light colored doughnut with a hole"}]} {"id": 87792, "image": "COCO_train2014_000000087792.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small donut , with no whole and a lot of powdered sugar\"."}], "task": "refering", "answer_template": "The \"a small donut , with no whole and a lot of powdered sugar\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 268, 269, 270, 271, 272, 273, 274, 275, 291, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 361, 362, 363, 364, 365, 366, 385, 386, 387, 388], "bbox": [0.6505625, 0.5393333333333333, 0.9741562500000001, 0.9595416666666666], "iscrowd": 0, "area": 33511.7129, "rle": {"size": [480, 640], "counts": "X[S61j>:G:F:F:F9G:F7K3L4M3L4M3L4L5L3L4M3L4M3L4L4M4K4N2N2O1N2N2O1N3N1N2N2O1N2N2O1N3M2O1N2O1N2N2O1N2O2M2O1N100O2O0O101N100O2N100O2O0O2O0O101N100O2O00001O000O2O00001O001O0O101O000000001O0O1000000000000O1000000000000000001O00000001O0000000000000000000000001O01O00000000000000000001O00000000000001O00O100O2N100O1O100O100O1O2O0O2O0O2N2O0O2N101N2O0O2N101N2N101N101N2N101N1O2O2M2O2M2M4K4L5K5K4L5K4L5K5K4L5K4M4K5K4L5J5J7I6J7I7IWe7"}, "label": "a small donut , with no whole and a lot of powdered sugar"}]} {"id": 87792, "image": "COCO_train2014_000000087792.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the filled donut upper left\"."}], "task": "refering", "answer_template": "The \"the filled donut upper left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 25, 26, 27, 28, 29, 30, 31, 32, 48, 49, 50, 51, 52, 53, 54, 55, 56, 71, 72, 73, 74, 75, 76, 77, 78, 79, 94, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169], "bbox": [0.09628125, 0.0011250000000000001, 0.46115625, 0.44256249999999997], "iscrowd": 0, "area": 39275.185249999995, "rle": {"size": [480, 640], "counts": "aTm0b0Y>d0]Od0[O:G5K6J5L4M3L4M3L4M4K4M3M3O1N2N2N3N1N2N2N2N2O1N3M2N2O1N2N2N2N3N1N2N2N2O1N2N3M2O1N2N2N2N2O1N3M2N2O1N2N2N2N3N1N2O1O1O1O1O2N1O1O1O1O1O1O2N1O0000000000000001O000000000000000000000000000000000000000O1000001O000000000000000000000000000000000000000000001O000000000O10000000000000000000000000000000001O000O1000000000001O001O001O001O001O001N101O001O0O2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N2N2N3M3M2N3L3N3M3M2N3M2N3M3M2N3M2N3M3M2M4M2N3M3M2N3M2N3M3M2N3M2N4L8G9H8H8H8H8H8H8H8HgbQ5"}, "label": "the filled donut upper left"}]} {"id": 87792, "image": "COCO_train2014_000000087792.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the donut on the top left\"."}], "task": "refering", "answer_template": "The \"the donut on the top left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 25, 26, 27, 28, 29, 30, 31, 32, 48, 49, 50, 51, 52, 53, 54, 55, 56, 71, 72, 73, 74, 75, 76, 77, 78, 79, 94, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 141, 142, 143, 144, 145, 146, 147, 165, 166, 167, 168, 169], "bbox": [0.09628125, 0.0011250000000000001, 0.46115625, 0.44256249999999997], "iscrowd": 0, "area": 39275.185249999995, "rle": {"size": [480, 640], "counts": "aTm0b0Y>d0]Od0[O:G5K6J5L4M3L4M3L4M4K4M3M3O1N2N2N3N1N2N2N2N2O1N3M2N2O1N2N2N2N3N1N2N2N2O1N2N3M2O1N2N2N2N2O1N3M2N2O1N2N2N2N3N1N2O1O1O1O1O2N1O1O1O1O1O1O2N1O0000000000000001O000000000000000000000000000000000000000O1000001O000000000000000000000000000000000000000000001O000000000O10000000000000000000000000000000001O000O1000000000001O001O001O001O001O001N101O001O0O2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N2N2N3M3M2N3L3N3M3M2N3M2N3M3M2N3M2N3M3M2M4M2N3M3M2N3M2N3M3M2N3M2N4L8G9H8H8H8H8H8H8H8HgbQ5"}, "label": "the donut on the top left"}]} {"id": 87792, "image": "COCO_train2014_000000087792.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glazed cake donut on upper right\"."}], "task": "refering", "answer_template": "The \"glazed cake donut on upper right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 35, 36, 37, 38, 39, 40, 41, 57, 58, 59, 60, 61, 62, 63, 64, 65, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 126, 127, 128, 129, 130, 131, 132, 133, 134, 149, 150, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201], "bbox": [0.458421875, 0.006750000000000001, 0.8747187499999999, 0.5146041666666668], "iscrowd": 0, "area": 45378.457850000006, "rle": {"size": [480, 640], "counts": "_mY42l>2O2N1O2M2O1O2M2O2N1O2M2O2N1N3N1O2N1N2O2N1O2M2O2N1N3N2N;E;D;F;E5J4M3M3M3L4M3M3L3N3M3M3L4M3M3M2M4M3M3M3M3M3N2M2N3M2N2N1O2O1N1O2N2N1O2N2O0O2N2N1O2N2O0O2N2N1O2N2N101N2N1O2N2N101N1O2N1O1O2N100O101O000O2O000O2O00000O2O000O2O000O101O000O2O00001O00000000000000000000001O000000000001O000000000000010O000000000000001N100O2O000O101N10001N10001N100O101O0O101N10001N100O2O000O101N10001N100O2O1O0O2O001N2O0O2O001N2O0O2O001N2O0O2N101N2N2N3M2N2N3M2N2N2O2M2N2N3M2N2N3M2N2N3M2O1N3L3M3M3N3L3M3N3L3M3N3L3M3N3L3M3M3N3L3M3N3L3M3N3L3L4K6J5K5J6K6J;E;D in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 35, 36, 37, 38, 39, 40, 41, 57, 58, 59, 60, 61, 62, 63, 64, 65, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 126, 127, 128, 129, 130, 131, 132, 133, 134, 149, 150, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 197, 198, 199, 200, 201], "bbox": [0.458421875, 0.006750000000000001, 0.8747187499999999, 0.5146041666666668], "iscrowd": 0, "area": 45378.457850000006, "rle": {"size": [480, 640], "counts": "_mY42l>2O2N1O2M2O1O2M2O2N1O2M2O2N1N3N1O2N1N2O2N1O2M2O2N1N3N2N;E;D;F;E5J4M3M3M3L4M3M3L3N3M3M3L4M3M3M2M4M3M3M3M3M3N2M2N3M2N2N1O2O1N1O2N2N1O2N2O0O2N2N1O2N2O0O2N2N1O2N2N101N2N1O2N2N101N1O2N1O1O2N100O101O000O2O000O2O00000O2O000O2O000O101O000O2O00001O00000000000000000000001O000000000001O000000000000010O000000000000001N100O2O000O101N10001N10001N100O101O0O101N10001N100O2O000O101N10001N100O2O1O0O2O001N2O0O2O001N2O0O2O001N2O0O2N101N2N2N3M2N2N3M2N2N2O2M2N2N3M2N2N3M2N2N3M2O1N3L3M3M3N3L3M3N3L3M3N3L3M3N3L3M3M3N3L3M3N3L3M3N3L3L4K6J5K5J6K6J;E;D in this image.", "objects": [{"patches": [102, 103, 125, 126, 127, 129, 130, 149, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 311, 312, 313, 314, 334, 335, 336, 337, 338, 357, 358, 359, 360, 361, 381, 382, 383, 384, 385, 404, 405, 407, 408, 427, 428, 431, 432, 450, 451, 454, 455, 456, 473, 474, 478, 479], "bbox": [0.453421875, 0.18692967409948544, 0.837078125, 0.9797255574614064], "iscrowd": 0, "area": 36204.67454999998, "rle": {"size": [583, 640], "counts": "TWU52Sb08I7I4K3N3L3O2N1O2N1O2O0O2N1O1O2N1O2N1O2N1O2N12O2Ma0@O00001O001O0O2O001O001O001O0010O0010O0nCPOl6Q1jH\\OS7c0dHJX77^H7]7IZHf0a7ZOUHU1e7kNRHe1g7\\NoGT2k7kMlGe2m7\\MiGT3Q8kLfGd3T8\\LcGl3^8ULXGS4j8lKmF\\4T9eKbFd4_9[KXFm4i9TKmET5U:_10O2N100O1O0010O01O010O0:G>A>C6J6I8I6I7J:E;F:E;E;F3L00000000001O000000eJbImN^6o0jImNW6o0PJmNn5Q1ZJkNb5V1eJfNX5Y1QKbNk4e1VKdJZOR2\\5a3mLZLP3l3SMmKj2Z4YM]Kg2h4\\MoJd2X5UNhIl1^6SN]IP2f6nMUIU2o6jMlHX2X7gMdHZ2`7dM\\H^2h7aMSHa2Q8]MkGe2Y8b34L4M3M000001O0000001O00001O00000iKeFB\\99RG]On8?_GXOa8e0dGZO\\8a0jG^OU8>RH_On7=XHBg7`N`F3n1\\1b7^NcF5n1[1^7^NhF7j1Z1]7]NlF9h1X1]7\\NnFX8BjG=T8DoG:P8GQH7o7ISH6l7JVH5j7JYH4f7L\\H2e7cN^E:n2R1d8nN^GP1b8PO`Go0_8QOdGl0\\8TOfGk0Y8UOiGi0W8XOjGf0V8ZOlGe0S8[OPHb0P8^OQHb0n7^OSHa0m7@RH`0n7@SH`0m7_OSHa0n7^OSHb0l7^OTHb0m7^OSHb0m7]OSHc0m7]OTHb0m7]OSHd0m7[OSHe0m7\\ORHe0n7ZOSHe0n7ZORHg0l7ZOTHf0l7ZOTHf0k7[OVHe0i7ZOXHf0g7[OYHf0f7YO[Hg0d7YO^Hg0`7YOaHg0^7ZObHf0]7ZOeHf0Z7YOhHf0W7ZOkHf0S7[OnHd0Q7VOVIk0h6PO_Io0`6QObIn0]6ROeIn0Y6ROiIm0V6SOlIm0S6ROoIm0Q6ROQJn0n5QOTJo0k5POWJP1h5nNZJS1e5kN^JU1c5hN_JX1b5dNaJ]1U;00000000001N10O1N101N1O2O0O101N1O2O0O1O2N1O2N2O0O2N2N2M2N3M3L4Fjej1"}, "label": "a tennis player in blue t - shirt and black shorts just after serving"}]} {"id": 382707, "image": "COCO_train2014_000000382707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a player with blue teashirt and black shorts is looking somewhere\"."}], "task": "refering", "answer_template": "The \"a player with blue teashirt and black shorts is looking somewhere\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 125, 126, 127, 129, 130, 149, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 288, 289, 290, 291, 311, 312, 313, 314, 334, 335, 336, 337, 338, 357, 358, 359, 360, 361, 381, 382, 383, 384, 385, 404, 405, 407, 408, 427, 428, 431, 432, 450, 451, 454, 455, 456, 473, 474, 478, 479], "bbox": [0.453421875, 0.18692967409948544, 0.837078125, 0.9797255574614064], "iscrowd": 0, "area": 36204.67454999998, "rle": {"size": [583, 640], "counts": "TWU52Sb08I7I4K3N3L3O2N1O2N1O2O0O2N1O1O2N1O2N1O2N1O2N12O2Ma0@O00001O001O0O2O001O001O001O0010O0010O0nCPOl6Q1jH\\OS7c0dHJX77^H7]7IZHf0a7ZOUHU1e7kNRHe1g7\\NoGT2k7kMlGe2m7\\MiGT3Q8kLfGd3T8\\LcGl3^8ULXGS4j8lKmF\\4T9eKbFd4_9[KXFm4i9TKmET5U:_10O2N100O1O0010O01O010O0:G>A>C6J6I8I6I7J:E;F:E;E;F3L00000000001O000000eJbImN^6o0jImNW6o0PJmNn5Q1ZJkNb5V1eJfNX5Y1QKbNk4e1VKdJZOR2\\5a3mLZLP3l3SMmKj2Z4YM]Kg2h4\\MoJd2X5UNhIl1^6SN]IP2f6nMUIU2o6jMlHX2X7gMdHZ2`7dM\\H^2h7aMSHa2Q8]MkGe2Y8b34L4M3M000001O0000001O00001O00000iKeFB\\99RG]On8?_GXOa8e0dGZO\\8a0jG^OU8>RH_On7=XHBg7`N`F3n1\\1b7^NcF5n1[1^7^NhF7j1Z1]7]NlF9h1X1]7\\NnFX8BjG=T8DoG:P8GQH7o7ISH6l7JVH5j7JYH4f7L\\H2e7cN^E:n2R1d8nN^GP1b8PO`Go0_8QOdGl0\\8TOfGk0Y8UOiGi0W8XOjGf0V8ZOlGe0S8[OPHb0P8^OQHb0n7^OSHa0m7@RH`0n7@SH`0m7_OSHa0n7^OSHb0l7^OTHb0m7^OSHb0m7]OSHc0m7]OTHb0m7]OSHd0m7[OSHe0m7\\ORHe0n7ZOSHe0n7ZORHg0l7ZOTHf0l7ZOTHf0k7[OVHe0i7ZOXHf0g7[OYHf0f7YO[Hg0d7YO^Hg0`7YOaHg0^7ZObHf0]7ZOeHf0Z7YOhHf0W7ZOkHf0S7[OnHd0Q7VOVIk0h6PO_Io0`6QObIn0]6ROeIn0Y6ROiIm0V6SOlIm0S6ROoIm0Q6ROQJn0n5QOTJo0k5POWJP1h5nNZJS1e5kN^JU1c5hN_JX1b5dNaJ]1U;00000000001N10O1N101N1O2O0O101N1O2O0O1O2N1O2N2O0O2N2N2M2N3M3L4Fjej1"}, "label": "a player with blue teashirt and black shorts is looking somewhere"}]} {"id": 79611, "image": "COCO_train2014_000000079611.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"clock mounted to wall\"."}], "task": "refering", "answer_template": "The \"clock mounted to wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 286, 287, 288, 289, 290, 310, 311, 312, 313], "bbox": [0.41634374999999996, 0.33032941176470587, 0.69390625, 0.8988705882352941], "iscrowd": 0, "area": 33581.71324999999, "rle": {"size": [425, 640], "counts": "Se^32g in this image.", "objects": [{"patches": [125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 193, 194, 195, 196, 197, 198, 199, 216, 217, 218, 219, 220, 221, 222, 239, 240, 241, 242, 243, 244, 245, 263, 264, 265, 266, 267, 268, 286, 287, 288, 289, 290, 310, 311, 312, 313], "bbox": [0.41634374999999996, 0.33032941176470587, 0.69390625, 0.8988705882352941], "iscrowd": 0, "area": 33581.71324999999, "rle": {"size": [425, 640], "counts": "Se^32g in this image.", "objects": [{"patches": [96, 97, 117, 118, 119, 120, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 300, 301, 302, 303, 304], "bbox": [0.023875, 0.32583529411764706, 0.2790625, 0.9213411764705882], "iscrowd": 0, "area": 31720.704800000003, "rle": {"size": [425, 640], "counts": "k_63o<9G9G9G9G9H;D?D in this image.", "objects": [{"patches": [96, 97, 117, 118, 119, 120, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 300, 301, 302, 303, 304], "bbox": [0.023875, 0.32583529411764706, 0.2790625, 0.9213411764705882], "iscrowd": 0, "area": 31720.704800000003, "rle": {"size": [425, 640], "counts": "k_63o<9G9G9G9G9H;D?D in this image.", "objects": [{"patches": [91, 92, 108, 109, 110, 126, 127, 128, 129, 144, 145, 146, 147, 163, 164], "bbox": [0.0, 0.41, 0.19768, 0.7123200000000001], "iscrowd": 0, "area": 5923.655950000002, "rle": {"size": [375, 500], "counts": "e5l0l:0O100O101N10000O101N100O2O0O2N1O1N3N1N2O2N1N3N1N2O1N101N2O0^OTNXGn1f8TNWGn1h8UNUGm1i8UNTGn1k8SNRGo1m8TNPGn1o8>N2O1N2N2O2M2O1N2N3N1N2O1O2O000O01N2lLfGi2Z8WMhGf2Z8XMjGe2V8ZMmGd2T8[MoGa2S8]MPHa2P8^MRH`2P8_MRH_2e8M2O2N1N3N2N3L3N3M3M2N3M2N2O0O1O1O100O1O1O0N2N2O1N2N2N3N1N2N2N2O1N2N2N3Dddb4"}, "label": "a triangular finger sandwich , with only dark bread , pate and tomato"}]} {"id": 63238, "image": "COCO_train2014_000000063238.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"sandwich on white bread\"."}], "task": "refering", "answer_template": "The \"sandwich on white bread\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 10, 11, 12, 13, 14, 15, 27, 28, 29, 30, 31, 32, 33, 46, 47, 48, 49, 65], "bbox": [0.5, 0.0033866666666666667, 0.8598, 0.2736533333333333], "iscrowd": 0, "area": 10468.03575, "rle": {"size": [375, 500], "counts": "iak21f;1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1N3N1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O1O00O1O100O1O010O1O100O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O2O0O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O2O0O1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O100O1O100O8H3N2M3M4Mgci0"}, "label": "sandwich on white bread"}]} {"id": 63238, "image": "COCO_train2014_000000063238.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dark brown sandwich to left of white sandwich in middle row of food platters\"."}], "task": "refering", "answer_template": "The \"dark brown sandwich to left of white sandwich in middle row of food platters\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 22, 23, 24, 25, 26, 27, 28, 41, 42, 43, 44, 45, 46, 58, 59, 60, 61, 62, 63, 64, 76, 77, 78, 79, 80, 81, 82, 96, 97, 98], "bbox": [0.27133999999999997, 0.013493333333333333, 0.62022, 0.44944], "iscrowd": 0, "area": 19726.541250000006, "rle": {"size": [375, 500], "counts": "mia1f0Q;i0\\FEZ6l210O001O000010O01O001O0010O0001O0010O01O001O00010O001O001O010O00001O0010O01O001O00010O001O001O01O01O001O0010O01O00001O010O001O001O01O01O001O0010O01O00001O010O000000O1O2N100O1O100O2N1O100O1O1O2O0O1O100O1O2N100O1O1O101N1O100O1O2N100O1O100O2N1O100O1O1O2O0O1O102M3M3M2O2M3M2N3N2M2N3N2M3M2N3N2M2N3N2M3M2N3N2M2N3M3N1N3M3N2M2N3M3N1N3M3N1N3M3M3N1N3M3M^_U2"}, "label": "dark brown sandwich to left of white sandwich in middle row of food platters"}]} {"id": 464650, "image": "COCO_train2014_000000464650.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blonde woman using a stick blender\"."}], "task": "refering", "answer_template": "The \"the blonde woman using a stick blender\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 36, 37, 38, 58, 59, 60, 61, 81, 82, 83, 84, 105, 106, 107, 126, 127, 128, 129, 130, 131, 148, 149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 178, 193, 194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 220, 221, 222, 223, 224, 239, 240, 241, 242, 243, 244, 245, 246, 247, 263, 264, 265, 266, 267, 268, 269, 289, 290], "bbox": [0.411234375, 0.036104166666666666, 0.756734375, 0.7327291666666667], "iscrowd": 0, "area": 42046.19514999999, "rle": {"size": [480, 640], "counts": "hak32g>8I7I7M3L4M3M3L5L3M3M3L4M3M3L4M3M3M4K4M3M3L4M3M3M3L4M3M4L3L4M3M3L4M2N2N2M3N2N2M3N2N2O1N2O1O1N2O0O01000O010O10O10O2O100O10000OQMlF4T9LnF2R9OnF0Q91PGNP93PGLP94RGIn89RGFn8:SGCo8=RGAn8a0SG[Oo8e0VGUOk8l0YGmNh8T1]GfNd8[1`G^Nb8b1cGXN]8j1gGoM[8Q2jGiMW8X2X22QEjMP9Y2iFUNl8o1nF^Nh8e1QGhNf8\\1RGhNn8[1kFjNT9Y1cFnN[9V1\\FQOc9W3O1N2O001QNfJUJ\\5e5QKSJP5f5^KRJc4i5iKnIY4n5QLkIP4Q6YLiIh3T6aLdIa3X6iLaIY3Z6i2M3L4M4M2N2N2M4M2N2N2N3L3N2N3M2N3M2N3M2N001O1O1O1O0011O00000O101O00000O101O00000O101O00000O101O0O100XITK[3m4dLUK[3l4bLVK]3k4bLWK]3i4bLYK\\3i4bLYK]3g4aL\\K^3d4aL]K^3e4`L]K_3c4_L`K_3b4_L`K`3a4^LaK`3b4\\L`Kd3c4XL_Kf3d4WL^Kh3d4UL^Ki3f4RL]Km3e4PL\\KP4f4mK\\KQ4h4jK[KU4i4fKXKY4n4aKTK^4Q5RKZKm4l4^JeKa5m6O2O0O2O001N101N101N101N101N101O1N101N101N101N2O0O2O0O2O1O0O2O1N2N2N2M3N2N2N6J7H8I7I7\\Od0POP1POVnX2"}, "label": "the blonde woman using a stick blender"}]} {"id": 464650, "image": "COCO_train2014_000000464650.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde woman with a mixer\"."}], "task": "refering", "answer_template": "The \"a blonde woman with a mixer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 36, 37, 38, 58, 59, 60, 61, 81, 82, 83, 84, 105, 106, 107, 126, 127, 128, 129, 130, 131, 148, 149, 150, 151, 152, 153, 154, 155, 171, 172, 173, 174, 175, 176, 177, 178, 193, 194, 195, 196, 197, 198, 199, 200, 201, 216, 217, 218, 219, 220, 221, 222, 223, 224, 239, 240, 241, 242, 243, 244, 245, 246, 247, 263, 264, 265, 266, 267, 268, 269, 289, 290], "bbox": [0.411234375, 0.036104166666666666, 0.756734375, 0.7327291666666667], "iscrowd": 0, "area": 42046.19514999999, "rle": {"size": [480, 640], "counts": "hak32g>8I7I7M3L4M3M3L5L3M3M3L4M3M3L4M3M3M4K4M3M3L4M3M3M3L4M3M4L3L4M3M3L4M2N2N2M3N2N2M3N2N2O1N2O1O1N2O0O01000O010O10O10O2O100O10000OQMlF4T9LnF2R9OnF0Q91PGNP93PGLP94RGIn89RGFn8:SGCo8=RGAn8a0SG[Oo8e0VGUOk8l0YGmNh8T1]GfNd8[1`G^Nb8b1cGXN]8j1gGoM[8Q2jGiMW8X2X22QEjMP9Y2iFUNl8o1nF^Nh8e1QGhNf8\\1RGhNn8[1kFjNT9Y1cFnN[9V1\\FQOc9W3O1N2O001QNfJUJ\\5e5QKSJP5f5^KRJc4i5iKnIY4n5QLkIP4Q6YLiIh3T6aLdIa3X6iLaIY3Z6i2M3L4M4M2N2N2M4M2N2N2N3L3N2N3M2N3M2N3M2N001O1O1O1O0011O00000O101O00000O101O00000O101O00000O101O0O100XITK[3m4dLUK[3l4bLVK]3k4bLWK]3i4bLYK\\3i4bLYK]3g4aL\\K^3d4aL]K^3e4`L]K_3c4_L`K_3b4_L`K`3a4^LaK`3b4\\L`Kd3c4XL_Kf3d4WL^Kh3d4UL^Ki3f4RL]Km3e4PL\\KP4f4mK\\KQ4h4jK[KU4i4fKXKY4n4aKTK^4Q5RKZKm4l4^JeKa5m6O2O0O2O001N101N101N101N101N101O1N101N101N101N2O0O2O0O2O1O0O2O1N2N2N2M3N2N2N6J7H8I7I7\\Od0POP1POVnX2"}, "label": "a blonde woman with a mixer"}]} {"id": 22287, "image": "COCO_train2014_000000022287.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra which is near the giraffe\"."}], "task": "refering", "answer_template": "The \"a zebra which is near the giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [205, 206, 207, 220, 221, 222, 233, 235, 236, 237, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 291, 292, 293, 294, 295, 296, 297, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 337, 338, 339, 340, 341, 342], "bbox": [0.44230588235294116, 0.582953125, 0.8643529411764707, 0.9865468749999999], "iscrowd": 0, "area": 26427.34029999999, "rle": {"size": [640, 425], "counts": "UPf32nc06J5L5J5K6K4K6J5L5k^OhNV?]1]@mNa?Y1Q@QOn?T1c_OUOg>lNfB\\3Z=gLbBZ3]=kL^BV3a=nL[BS3d=QMXBP3h=SMTBo2j=UMRBl2m=XMoAi2P>[MlAf2S>_MhAb2X>aMdA`2[>dMaA]2^>gM^AZ2a>c100O1O1O01K4M3L4M3L5L3L4M3L5L3L4M4K5L3L5L4O1Q1oNX1hN00000O100000000000000O100000000000O2A>K5M3N2M3N2M4M2M3N2M3N210P1oN:F0001O001O001O001O00010O001O001UNgBmLX=T3i11N101N101O1N101N2K4H9G9F9G:F:L310O1O010O010O3M4M4K5L3M4K4M4L4L3M4L4L3M4L3M4L4L3M4L0O10000000000001O2N2M3_K`D>b;YOiECY:4fGTN]8c1U5G9G8H6J7I7I7K5O0O2O1N2O1O0O2O1O1N2O001N2O1O1KQlS1"}, "label": "a zebra which is near the giraffe"}]} {"id": 22287, "image": "COCO_train2014_000000022287.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra standing next to a giraffe\"."}], "task": "refering", "answer_template": "The \"zebra standing next to a giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [205, 206, 207, 220, 221, 222, 233, 235, 236, 237, 247, 248, 249, 250, 251, 252, 262, 263, 264, 265, 266, 276, 277, 278, 279, 280, 281, 291, 292, 293, 294, 295, 296, 297, 307, 308, 309, 310, 311, 312, 322, 323, 324, 325, 326, 327, 337, 338, 339, 340, 341, 342], "bbox": [0.44230588235294116, 0.582953125, 0.8643529411764707, 0.9865468749999999], "iscrowd": 0, "area": 26427.34029999999, "rle": {"size": [640, 425], "counts": "UPf32nc06J5L5J5K6K4K6J5L5k^OhNV?]1]@mNa?Y1Q@QOn?T1c_OUOg>lNfB\\3Z=gLbBZ3]=kL^BV3a=nL[BS3d=QMXBP3h=SMTBo2j=UMRBl2m=XMoAi2P>[MlAf2S>_MhAb2X>aMdA`2[>dMaA]2^>gM^AZ2a>c100O1O1O01K4M3L4M3L5L3L4M3L5L3L4M4K5L3L5L4O1Q1oNX1hN00000O100000000000000O100000000000O2A>K5M3N2M3N2M4M2M3N2M3N210P1oN:F0001O001O001O001O00010O001O001UNgBmLX=T3i11N101N101O1N101N2K4H9G9F9G:F:L310O1O010O010O3M4M4K5L3M4K4M4L4L3M4L4L3M4L3M4L4L3M4L0O10000000000001O2N2M3_K`D>b;YOiECY:4fGTN]8c1U5G9G8H6J7I7I7K5O0O2O1N2O1O0O2O1O1N2O001N2O1O1KQlS1"}, "label": "zebra standing next to a giraffe"}]} {"id": 22287, "image": "COCO_train2014_000000022287.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra sniffs the rear of another zebra\"."}], "task": "refering", "answer_template": "The \"a zebra sniffs the rear of another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 215, 227, 228, 229, 230, 231, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 270, 271, 272, 273, 274, 276, 277, 285, 286, 287, 288, 289, 300, 301, 302, 303, 315, 316, 317, 318, 330, 331, 332, 333], "bbox": [0.0033882352941176468, 0.606734375, 0.5042117647058824, 0.986515625], "iscrowd": 0, "area": 27892.33155, "rle": {"size": [640, 425], "counts": "eR19gc0P3oLn0SO001O001N10001O001O0O101O001O000O4M6J6J3MO01000O10000O01000O1000O010000O100O01000O1000O01K5L2N3L4M3N22N1N3N2N2N2N1O2N2M3N2N1O2N2N2N2M5L3M3M3M4K4M3M3M4L3M3L2O1O1O1O1N2O1O1OIkBeIT=Z6oBeIP=Z6SCeIk in this image.", "objects": [{"patches": [213, 214, 215, 227, 228, 229, 230, 231, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 270, 271, 272, 273, 274, 276, 277, 285, 286, 287, 288, 289, 300, 301, 302, 303, 315, 316, 317, 318, 330, 331, 332, 333], "bbox": [0.0033882352941176468, 0.606734375, 0.5042117647058824, 0.986515625], "iscrowd": 0, "area": 27892.33155, "rle": {"size": [640, 425], "counts": "eR19gc0P3oLn0SO001O001N10001O001O0O101O001O000O4M6J6J3MO01000O10000O01000O1000O010000O100O01000O1000O01K5L2N3L4M3N22N1N3N2N2N2N1O2N2M3N2N1O2N2N2N2M5L3M3M3M4K4M3M3M4L3M3L2O1O1O1O1N2O1O1OIkBeIT=Z6oBeIP=Z6SCeIk in this image.", "objects": [{"patches": [98, 99, 100, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 259, 260, 261], "bbox": [0.11646875000000001, 0.27058333333333334, 0.48706250000000006, 0.6611666666666667], "iscrowd": 0, "area": 32010.2417, "rle": {"size": [480, 640], "counts": "b\\S1h0T>5N2M2O2M3N2N2M2O2N2M3N2M2O2N2M3N1N3N2N2M3N1O2M3N2M3N1O2M3N2M3N1O2M3N2N1O2N1O2N1O1O1O2N1O1O2N1O1O2N100O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O1O2N1O101N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2O0O101O000O2O000O2O00001N10001N10001O0O10000O10000O1000000O10000O10000000000001O00000000001O000000001O00000000001O00000000001O00000O100O2O0O100O1O2O0O1O100O101N1O100O100O2N100O1O2O1N101N1O2O1N1O2O1N101N1O2O1N1O2O0O2O1N1N3N2M2N3N1N3N2M2O2M3N1N3N1N3M3N1N3N1N3N2M2O2L4M2M4M3L4M2M4L4M2M4J6H7H9G9H7H9Hfbi4"}, "label": "smaller pizza"}]} {"id": 513811, "image": "COCO_train2014_000000513811.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the smaller pizza\"."}], "task": "refering", "answer_template": "The \"the smaller pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 100, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 149, 164, 165, 166, 167, 168, 169, 170, 171, 172, 186, 187, 188, 189, 190, 191, 192, 193, 194, 209, 210, 211, 212, 213, 214, 215, 216, 217, 233, 234, 235, 236, 237, 238, 239, 259, 260, 261], "bbox": [0.11646875000000001, 0.27058333333333334, 0.48706250000000006, 0.6611666666666667], "iscrowd": 0, "area": 32010.2417, "rle": {"size": [480, 640], "counts": "b\\S1h0T>5N2M2O2M3N2N2M2O2N2M3N2M2O2N2M3N1N3N2N2M3N1O2M3N2M3N1O2M3N2M3N1O2M3N2N1O2N1O2N1O1O1O2N1O1O2N1O1O2N100O1O2N1O1O2N1O1O1O2N1O1O2N1O1O2N1O1O1O2N1O101N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2O0O101O000O2O000O2O00001N10001N10001O0O10000O10000O1000000O10000O10000000000001O00000000001O000000001O00000000001O00000000001O00000O100O2O0O100O1O2O0O1O100O101N1O100O100O2N100O1O2O1N101N1O2O1N1O2O1N101N1O2O1N1O2O0O2O1N1N3N2M2N3N1N3N2M2O2M3N1N3N1N3M3N1N3N1N3N2M2O2L4M2M4M3L4M2M4L4M2M4J6H7H9G9H7H9Hfbi4"}, "label": "the smaller pizza"}]} {"id": 513811, "image": "COCO_train2014_000000513811.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the larger pizza\"."}], "task": "refering", "answer_template": "The \"the larger pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 63, 82, 83, 84, 85, 86, 87, 88, 104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250], "bbox": [0.50225, 0.1528125, 0.9539375, 0.6494375], "iscrowd": 0, "area": 50337.72780000001, "rle": {"size": [480, 640], "counts": "Qdf42m>c0]O=D:E in this image.", "objects": [{"patches": [61, 62, 63, 82, 83, 84, 85, 86, 87, 88, 104, 105, 106, 107, 108, 109, 110, 111, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 243, 244, 245, 246, 247, 248, 249, 250], "bbox": [0.50225, 0.1528125, 0.9539375, 0.6494375], "iscrowd": 0, "area": 50337.72780000001, "rle": {"size": [480, 640], "counts": "Qdf42m>c0]O=D:E in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 26, 27, 28, 29, 30, 31, 32, 33, 34, 49, 50, 51, 52, 53, 54, 55, 56, 57, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.006734374999999999, 0.0045000000000000005, 0.87134375, 0.9865208333333334], "iscrowd": 0, "area": 164402.9496, "rle": {"size": [480, 640], "counts": "_i28a>7I7I7H8I7I7I7H8I7I7I7I7L4L4L4L4L4L4L4L4N2O100O1O2N1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1eNQKdHP5Z7_KYHa4f7lKmGU4R8]1O1O1O1O1N2O1O1SM\\I]Me6Z2hIaMY6W2RJeMo5R2]JjMd5n1hJlMZ5b1[KZNf4i0ZLSOg3i0aLRO`3j0gLROZ3j0mLROT3i0UMROl2j0[MQOg2k0`MQOa2j0hMQOY2k0nMQOS2k0TNQOm1k0[NPOf1k0bNQO_1m0fNnN\\1Q1Z6O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O001O001O001O01O01O001O001O001O00001O001O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O3M4L3M4L3M3M4L3M4L3M4L3M4L3M4L3M4L3M3M4L3M4L3M4L3M4L3oF_K]7e4VHeKi7^4kGkKU8Y4_GoKa8Y5O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O00001O001O001O2N2N2N1O2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N1O2N2N2N2N2N3M3M4L3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M2N3M3M2N3M001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O001O1O001O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O2N2O1N3M2N2N3M2N2N3M2N2N3M2N3M2N2N3M2N2N3M2N2N3M2N2NWoU1"}, "label": "a woman eating food inside"}]} {"id": 407318, "image": "COCO_train2014_000000407318.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"dark haired girl with big eyes eating a donut with her left hand\"."}], "task": "refering", "answer_template": "The \"dark haired girl with big eyes eating a donut with her left hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 26, 27, 28, 29, 30, 31, 32, 33, 34, 49, 50, 51, 52, 53, 54, 55, 56, 57, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.006734374999999999, 0.0045000000000000005, 0.87134375, 0.9865208333333334], "iscrowd": 0, "area": 164402.9496, "rle": {"size": [480, 640], "counts": "_i28a>7I7I7H8I7I7I7H8I7I7I7I7L4L4L4L4L4L4L4L4N2O100O1O2N1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O1O100O1O1O1O1O1O1eNQKdHP5Z7_KYHa4f7lKmGU4R8]1O1O1O1O1N2O1O1SM\\I]Me6Z2hIaMY6W2RJeMo5R2]JjMd5n1hJlMZ5b1[KZNf4i0ZLSOg3i0aLRO`3j0gLROZ3j0mLROT3i0UMROl2j0[MQOg2k0`MQOa2j0hMQOY2k0nMQOS2k0TNQOm1k0[NPOf1k0bNQO_1m0fNnN\\1Q1Z6O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O001O001O001O01O01O001O001O001O00001O001O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O1O1O2N1O1O1O2N1O1O1O3M4L3M4L3M3M4L3M4L3M4L3M4L3M4L3M4L3M3M4L3M4L3M4L3M4L3oF_K]7e4VHeKi7^4kGkKU8Y4_GoKa8Y5O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O00001O001O001O2N2N2N1O2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N1O2N2N2N2N2N3M3M4L3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M3M2N3M3M2N3M001O1O001O1O001O1O001O1O001O1O001O1O001O1O1O001O1O001O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O2N2O1N3M2N2N3M2N2N3M2N2N3M2N3M2N2N3M2N2N3M2N2N3M2N2NWoU1"}, "label": "dark haired girl with big eyes eating a donut with her left hand"}]} {"id": 46885, "image": "COCO_train2014_000000046885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skate boarder doing a trick off the ramp\"."}], "task": "refering", "answer_template": "The \"a skate boarder doing a trick off the ramp\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [18, 19, 30, 31, 41, 42, 43, 52, 53, 54, 55, 66, 67, 78, 89, 90, 101], "bbox": [0.39559171597633136, 0.08726, 0.6548816568047338, 0.4603], "iscrowd": 0, "area": 5465.885699999999, "rle": {"size": [500, 338], "counts": "YaQ23`?4L4L4L3M1O1O2N100O101O000000000000000000000000000VDYOg8g0XG[Og8e0WG]Oi8c0VG_Oi8a0UGAk8?TGAm8?QGCo8=PGCQ9=kFGV9:VF6m9K_Ef0c:[OSEk0o:VOhDQ1Z;PO]DU1f;lNQDZ1Q in this image.", "objects": [{"patches": [18, 19, 30, 31, 41, 42, 43, 52, 53, 54, 55, 66, 67, 78, 89, 90, 101], "bbox": [0.39559171597633136, 0.08726, 0.6548816568047338, 0.4603], "iscrowd": 0, "area": 5465.885699999999, "rle": {"size": [500, 338], "counts": "YaQ23`?4L4L4L3M1O1O2N100O101O000000000000000000000000000VDYOg8g0XG[Og8e0WG]Oi8c0VG_Oi8a0UGAk8?TGAm8?QGCo8=PGCQ9=kFGV9:VF6m9K_Ef0c:[OSEk0o:VOhDQ1Z;PO]DU1f;lNQDZ1Q in this image.", "objects": [{"patches": [124, 125, 147, 148, 149, 171, 172, 194, 195, 196, 197, 217, 218, 219, 220, 240, 241, 242, 243, 264, 266, 267, 287, 289, 290, 310, 312, 313], "bbox": [0.41042187500000005, 0.29125, 0.617796875, 0.8073541666666666], "iscrowd": 0, "area": 11438.925799999999, "rle": {"size": [480, 640], "counts": "Z^k36i>1N101N2O1J60O01O1O1O0N3M3L4N2O010003O03M3L5J6J5K6J4L3_DPNk9T2mESNQ:Q2fEVNW:n1`EZN^:i1XE`Ne:P3N2M4MULYE`3c:_LaEa3]:^LfEc3V:\\LmEe3Q:ZLRFg3d:3L4_GmKm5W4jIlKX6W4`IlKb6W4XIjKk6Y4mHiKV7Y4dHhK_7[4ZHfKi7g53M3M3M3M2N2N1O1N2O2G8A?A?A?[Of0oNP1O1O1O2N1O3M2N1O001O1O001O001O1O0013L4M0O00100O1O100O1O10O01O100O1O100O003N2M4L3[EcMb8_2ZGbMf8a2VGaMi8c2QG_MP9c2lF_MS9d2iF]MX9i2_FZMa9S3PFnLQ:_3`EcLa:n30O101M2O2N1O1O2N1O2N9G:Fa0_O9G1N2oMTELm:1XELk:OZENj:K\\E2jkb3"}, "label": "a giraffe next to zebra in a zoo"}]} {"id": 177959, "image": "COCO_train2014_000000177959.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe next to a zebra\"."}], "task": "refering", "answer_template": "The \"a giraffe next to a zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 147, 148, 149, 171, 172, 194, 195, 196, 197, 217, 218, 219, 220, 240, 241, 242, 243, 264, 266, 267, 287, 289, 290, 310, 312, 313], "bbox": [0.41042187500000005, 0.29125, 0.617796875, 0.8073541666666666], "iscrowd": 0, "area": 11438.925799999999, "rle": {"size": [480, 640], "counts": "Z^k36i>1N101N2O1J60O01O1O1O0N3M3L4N2O010003O03M3L5J6J5K6J4L3_DPNk9T2mESNQ:Q2fEVNW:n1`EZN^:i1XE`Ne:P3N2M4MULYE`3c:_LaEa3]:^LfEc3V:\\LmEe3Q:ZLRFg3d:3L4_GmKm5W4jIlKX6W4`IlKb6W4XIjKk6Y4mHiKV7Y4dHhK_7[4ZHfKi7g53M3M3M3M2N2N1O1N2O2G8A?A?A?[Of0oNP1O1O1O2N1O3M2N1O001O1O001O001O1O0013L4M0O00100O1O100O1O10O01O100O1O100O003N2M4L3[EcMb8_2ZGbMf8a2VGaMi8c2QG_MP9c2lF_MS9d2iF]MX9i2_FZMa9S3PFnLQ:_3`EcLa:n30O101M2O2N1O1O2N1O2N9G:Fa0_O9G1N2oMTELm:1XELk:OZENj:K\\E2jkb3"}, "label": "a giraffe next to a zebra"}]} {"id": 55092, "image": "COCO_train2014_000000055092.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large british airways jet next to water\"."}], "task": "refering", "answer_template": "The \"a large british airways jet next to water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [135, 156, 157, 158, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 178, 179, 180, 181, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 239, 240, 241], "bbox": [0.095515625, 0.391924882629108, 0.910390625, 0.6968544600938967], "iscrowd": 0, "area": 28865.250749999996, "rle": {"size": [426, 640], "counts": "Zci01Y=5K4K4L2N2O1N2N2N100O2N1O100O1O2N100O1O2N100O1O2N100O2N10000O2O0O100O2O0O101N100O101O2N1O2N2N2N8G5L00000000000000000O2OO1I7J6K500000000000O100000000001O0000000000000000001O0000000000000000001O0000000000000001O01O0000000000000000010O0000000000000000010O00000000000000000010O0000000000000000010O1O2N2N1O2N2N1O2N10O0000000000000000001O00000001O0000000001O0O100000001O000000001O000000001O000000001O0000001O000000lDUNi:l1REXNo:Q201O000001O0001O0001O00000010O00000001O01O00000000001O01O0000000000000000000O10000000000000001O0000000000000O1N2M3N2N2N2N200000001O000000000000000001O000000000000000000000001O0000000000000000000000000000000000000000000000000000000000001O0000000O100000000000000O10000000000000000O10000000000000000O2O000000000000000O100000000E^DnNb;R1;000000000000000000O1000000000000000000000001O000O1000000000000000000000000000000O1000O100O1O100O1O1O100O1O1O10O01O1O100O1O100O1O1O100O1O1O100O00100O1O1O101N1O1O100O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O01O0YO`EeN`:Z1bEdN_:[1dEcN]:[1fEcNZ:[1iEdNW:[1lEcNT:\\1oEbNR:\\1PFcNP:[1TFcNm9[1UFcNm9[1VFcNl9Z1WFdNj9Y1U1N2N3M2M3N2N2M4M2N2N2M4M2N2N2M4Mjng0"}, "label": "a large british airways jet next to water"}]} {"id": 55092, "image": "COCO_train2014_000000055092.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"british airways airplane sitting by a smaller gray plane\"."}], "task": "refering", "answer_template": "The \"british airways airplane sitting by a smaller gray plane\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [135, 156, 157, 158, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 178, 179, 180, 181, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 239, 240, 241], "bbox": [0.095515625, 0.391924882629108, 0.910390625, 0.6968544600938967], "iscrowd": 0, "area": 28865.250749999996, "rle": {"size": [426, 640], "counts": "Zci01Y=5K4K4L2N2O1N2N2N100O2N1O100O1O2N100O1O2N100O1O2N100O2N10000O2O0O100O2O0O101N100O101O2N1O2N2N2N8G5L00000000000000000O2OO1I7J6K500000000000O100000000001O0000000000000000001O0000000000000000001O0000000000000001O01O0000000000000000010O0000000000000000010O00000000000000000010O0000000000000000010O1O2N2N1O2N2N1O2N10O0000000000000000001O00000001O0000000001O0O100000001O000000001O000000001O000000001O0000001O000000lDUNi:l1REXNo:Q201O000001O0001O0001O00000010O00000001O01O00000000001O01O0000000000000000000O10000000000000001O0000000000000O1N2M3N2N2N2N200000001O000000000000000001O000000000000000000000001O0000000000000000000000000000000000000000000000000000000000001O0000000O100000000000000O10000000000000000O10000000000000000O2O000000000000000O100000000E^DnNb;R1;000000000000000000O1000000000000000000000001O000O1000000000000000000000000000000O1000O100O1O100O1O1O100O1O1O10O01O1O100O1O100O1O1O100O1O1O100O00100O1O1O101N1O1O100O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O1O100O1O100O1O1O100O1O1O100O1O01O0YO`EeN`:Z1bEdN_:[1dEcN]:[1fEcNZ:[1iEdNW:[1lEcNT:\\1oEbNR:\\1PFcNP:[1TFcNm9[1UFcNm9[1VFcNl9Z1WFdNj9Y1U1N2N3M2M3N2N2M4M2N2N2M4M2N2N2M4Mjng0"}, "label": "british airways airplane sitting by a smaller gray plane"}]} {"id": 30519, "image": "COCO_train2014_000000030519.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the orange truck and trailor closest to the orange box truck\"."}], "task": "refering", "answer_template": "The \"the orange truck and trailor closest to the orange box truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [195, 196, 197, 198, 199, 200, 216, 217, 218, 219, 220, 221, 222, 223, 238, 239, 240, 241, 242, 243, 244, 245, 246, 261, 262, 263, 264, 265, 266, 267, 268], "bbox": [0.37275, 0.579477434679335, 0.734953125, 0.80270783847981], "iscrowd": 0, "area": 13807.22435, "rle": {"size": [421, 640], "counts": "oaR33P=2M4L3M3N2M3M4L3N2M3M4L3N20001O0000:E2O1O1O1O001O0000001O0O10000000000000O100]OmDTOS;l0b0101L3M3M3M4N10000O2O2N7I9F2O001O1N2O001O1N1000000I_MiEb2V:`MiE`2W:`MiE`2W:71O4L0O100O1O2oN`EZOa:f0bEVO_:j0cESO^:m0dEPO]:P1eEmN\\:S1fEjN[:V1gEaN`:a1a03M2N3M3M3M1O000001O000000000000001O1O1O1O9G0000001G8000001O000TOaD2_;MbD3e;D]D in this image.", "objects": [{"patches": [178, 194, 195, 196, 197, 213, 214, 215, 231, 232, 233, 234, 248, 249, 250, 251, 265, 266, 267, 268, 282, 283, 284, 285, 301, 316, 318, 333], "bbox": [0.4305, 0.455375, 0.7982083333333333, 0.84103125], "iscrowd": 0, "area": 12538.8542, "rle": {"size": [640, 480], "counts": "RfQ41nc02N1O2N1O2N1N2O2N1O2N1O2N1O1O1O2N1O1O1O2O0O10001N100O10001N100O101O0O10001N100O10001N100O13N2M3M3M2OO01000O010O0100O0100O0100O0100O0100O0100O0100O0100O010O0100O01[^OPOY`0P1c_OCm?=Q@4a?L[@g0V?XOh@Y1i>hNTA^1P1mMa;e0[C`1m0SNh;=YCb1g0YNQ<5UCc1k0XNQ<4QCg1n0UNQ<5nBh1Q1SNR<5iBk1U1PNR<5gBm1W1nMS<5cBo1Z1lMT<4_BS2]1iMTh0UA\\Oh>c0YA]Oh>b0XA^Oi>`0XA@i>?WAAj>=VAEi>;WAEj>9WAEl>:TAWO\\?g0e@iNl?V1S@[N^`0c1c_OnMm`0Q2b0N101N2O1N2O1K5G9G8I8G9H8GQXl1"}, "label": "the giraffee eating leaves near zebra"}]} {"id": 30535, "image": "COCO_train2014_000000030535.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe who is behind the zebra\"."}], "task": "refering", "answer_template": "The \"the giraffe who is behind the zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 194, 195, 196, 197, 213, 214, 215, 231, 232, 233, 234, 248, 249, 250, 251, 265, 266, 267, 268, 282, 283, 284, 285, 301, 316, 318, 333], "bbox": [0.4305, 0.455375, 0.7982083333333333, 0.84103125], "iscrowd": 0, "area": 12538.8542, "rle": {"size": [640, 480], "counts": "RfQ41nc02N1O2N1O2N1N2O2N1O2N1O2N1O1O1O2N1O1O1O2O0O10001N100O10001N100O101O0O10001N100O10001N100O13N2M3M3M2OO01000O010O0100O0100O0100O0100O0100O0100O0100O0100O010O0100O01[^OPOY`0P1c_OCm?=Q@4a?L[@g0V?XOh@Y1i>hNTA^1P1mMa;e0[C`1m0SNh;=YCb1g0YNQ<5UCc1k0XNQ<4QCg1n0UNQ<5nBh1Q1SNR<5iBk1U1PNR<5gBm1W1nMS<5cBo1Z1lMT<4_BS2]1iMTh0UA\\Oh>c0YA]Oh>b0XA^Oi>`0XA@i>?WAAj>=VAEi>;WAEj>9WAEl>:TAWO\\?g0e@iNl?V1S@[N^`0c1c_OnMm`0Q2b0N101N2O1N2O1K5G9G8I8G9H8GQXl1"}, "label": "the giraffe who is behind the zebra"}]} {"id": 522062, "image": "COCO_train2014_000000522062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a striped shirt and sport coat holding a glass of wine\"."}], "task": "refering", "answer_template": "The \"a man with a striped shirt and sport coat holding a glass of wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 50, 51, 52, 53, 62, 63, 64, 65, 74, 75, 76, 77, 85, 86, 87, 88, 97, 98, 99, 100, 108, 109, 110, 111, 112, 120, 121, 122, 123, 124, 125, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 151, 156, 157, 158, 159, 160, 161, 162, 168, 169, 170, 171, 172, 173, 174, 180, 181, 182, 183, 184, 185, 186, 193, 194, 195, 196, 197, 198, 205, 206, 207, 208, 209, 210], "bbox": [0.03720720720720721, 0.19708, 0.6087087087087086, 0.99892], "iscrowd": 0, "area": 47757.35705, "rle": {"size": [500, 333], "counts": "UX61^?7I6I8_La3K5K4M4K5K4L5K5L4K4L5K5L4K8H8H8I6I8I7J5K6J6J6J5K1N2O1O1O2N1O1O1N2O1O1O1O1O1O2N1N2O1O1O1N2N2nNnF`KT9^4PG]KT9`4oF]KS9a3hFeL9FQ9c3RHZLP8e3RHXLP8f3THVLn7h3UHTLn7j3UHSLm7k3o1N3N1N2N2N2N2N2N2000001O00000000000000001O00000000000000001O0000000000000001O01O000000000001O0001O0001O000000000WGkLS5V3eJ[MR5d2lJhMj4X2TKTNb4l1\\KZN`4f1^K^N`4c1]KbN`4^1^KfN`4Z1^KjN`4W1]KnN`4T1\\KPOb4R1ZKSOd4m0YKWOe4k0VK[Og4g0TK^Oj4d0QKBl4`0oJEo4=kJIS59hJMU55fJ0X52cJ4[5M`J8^5J\\J=a5EYJa0e5@VJf0h5\\OQJl0l5`4N3M2N3N1N3M3M5K4L5K4L4LOfLYEf1i:[N[E]1g:dN]ET1g:mN\\El0g:UO\\Ed0g:\\O^E;f:F]E3f:N]EJf:8]EAf:?_EYOd:h0_EPOe:Q1^EhNf:W1^EaNg:_1]EYNh:f1c1O2O001O0O2O001O010O001N1D=L3N3L4M3L4M2M4L4L4LaTo1"}, "label": "a man with a striped shirt and sport coat holding a glass of wine"}]} {"id": 522062, "image": "COCO_train2014_000000522062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black suit and holding a champaign flute\"."}], "task": "refering", "answer_template": "The \"a man wearing a black suit and holding a champaign flute\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [38, 39, 40, 50, 51, 52, 53, 62, 63, 64, 65, 74, 75, 76, 77, 85, 86, 87, 88, 97, 98, 99, 100, 108, 109, 110, 111, 112, 120, 121, 122, 123, 124, 125, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 151, 156, 157, 158, 159, 160, 161, 162, 168, 169, 170, 171, 172, 173, 174, 180, 181, 182, 183, 184, 185, 186, 193, 194, 195, 196, 197, 198, 205, 206, 207, 208, 209, 210], "bbox": [0.03720720720720721, 0.19708, 0.6087087087087086, 0.99892], "iscrowd": 0, "area": 47757.35705, "rle": {"size": [500, 333], "counts": "UX61^?7I6I8_La3K5K4M4K5K4L5K5L4K4L5K5L4K8H8H8I6I8I7J5K6J6J6J5K1N2O1O1O2N1O1O1N2O1O1O1O1O1O2N1N2O1O1O1N2N2nNnF`KT9^4PG]KT9`4oF]KS9a3hFeL9FQ9c3RHZLP8e3RHXLP8f3THVLn7h3UHTLn7j3UHSLm7k3o1N3N1N2N2N2N2N2N2000001O00000000000000001O00000000000000001O0000000000000001O01O000000000001O0001O0001O000000000WGkLS5V3eJ[MR5d2lJhMj4X2TKTNb4l1\\KZN`4f1^K^N`4c1]KbN`4^1^KfN`4Z1^KjN`4W1]KnN`4T1\\KPOb4R1ZKSOd4m0YKWOe4k0VK[Og4g0TK^Oj4d0QKBl4`0oJEo4=kJIS59hJMU55fJ0X52cJ4[5M`J8^5J\\J=a5EYJa0e5@VJf0h5\\OQJl0l5`4N3M2N3N1N3M3M5K4L5K4L4LOfLYEf1i:[N[E]1g:dN]ET1g:mN\\El0g:UO\\Ed0g:\\O^E;f:F]E3f:N]EJf:8]EAf:?_EYOd:h0_EPOe:Q1^EhNf:W1^EaNg:_1]EYNh:f1c1O2O001O0O2O001O010O001N1D=L3N3L4M3L4M2M4L4L4LaTo1"}, "label": "a man wearing a black suit and holding a champaign flute"}]} {"id": 522062, "image": "COCO_train2014_000000522062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man ' s hand holding a wine glass in the foreground with a guy in a suit standing in background with a glass\"."}], "task": "refering", "answer_template": "The \"a man ' s hand holding a wine glass in the foreground with a guy in a suit standing in background with a glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 65, 66, 77, 78, 79, 88, 89, 90, 91, 100, 101, 102, 103, 104, 112, 113, 114, 115, 116, 125, 126, 127, 128, 137, 138, 139, 140, 150, 151, 152, 162, 163, 164, 174, 175, 176, 186, 187, 198, 199, 210, 211], "bbox": [0.3981381381381382, 0.26068, 0.6917117117117118, 0.98426], "iscrowd": 0, "area": 21457.734799999995, "rle": {"size": [500, 333], "counts": "lTQ2d1h=c0D3M2N2N3M2SO_MdDc2W;oMZDT2a;RN[DP2a;TN]Dn1_;WN^Dk1];Z1K4L4K6K4L4L5J5L4N3N100O101O0O101N100O4M6J5J7J6I5L4L4eJVIb1o6TNXIk1l6lMZIR2k6eM[IZ2i6]M^Ib2e6UMaIj2c6lLdIT3_6cLeI_3W91O2N2N1O2N1O2N2N1O1O2N1O1O1O2N1O1O2M2O1O1O2N1O1O1O2N1O1O2N1O1M3I7I8H7I7I7I7I:EB=C=D in this image.", "objects": [{"patches": [53, 65, 66, 77, 78, 79, 88, 89, 90, 91, 100, 101, 102, 103, 104, 112, 113, 114, 115, 116, 125, 126, 127, 128, 137, 138, 139, 140, 150, 151, 152, 162, 163, 164, 174, 175, 176, 186, 187, 198, 199, 210, 211], "bbox": [0.3981381381381382, 0.26068, 0.6917117117117118, 0.98426], "iscrowd": 0, "area": 21457.734799999995, "rle": {"size": [500, 333], "counts": "lTQ2d1h=c0D3M2N2N3M2SO_MdDc2W;oMZDT2a;RN[DP2a;TN]Dn1_;WN^Dk1];Z1K4L4K6K4L4L5J5L4N3N100O101O0O101N100O4M6J5J7J6I5L4L4eJVIb1o6TNXIk1l6lMZIR2k6eM[IZ2i6]M^Ib2e6UMaIj2c6lLdIT3_6cLeI_3W91O2N2N1O2N1O2N2N1O1O2N1O1O1O2N1O1O2M2O1O1O2N1O1O1O2N1O1O2N1O1M3I7I8H7I7I7I7I:EB=C=D in this image.", "objects": [{"patches": [82, 83, 94, 95, 105, 106, 107, 117, 118, 119, 129, 130, 131, 141, 142, 143, 153, 154, 155, 165, 166, 167, 177, 178, 179], "bbox": [0.7417717717717718, 0.33322, 1.0, 0.8281000000000001], "iscrowd": 0, "area": 15497.94515, "rle": {"size": [500, 333], "counts": "Vlh33]?6K5J5K4M3L4L5L3L4L4M3L4L4hCZN_:i1]EaNZ:b1bEiNW:W1eESOT:n0hEZOT:f0hEAT:`0cELZ:5ZE:a:`2bE\\Kf9g4nEdKo9S5N8G10000O10000O10000M3J6J6J6J6J6N2O1N2N2O1N2N2O1N2N2O10000O10000O1000000O10000001O001O1O001O1O001O1O3M3L4M3M6J8H4L3M3M4L3M3mITO"}, "label": "the woman with black hair and a black shirt"}]} {"id": 522062, "image": "COCO_train2014_000000522062.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"women in all black\"."}], "task": "refering", "answer_template": "The \"women in all black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 94, 95, 105, 106, 107, 117, 118, 119, 129, 130, 131, 141, 142, 143, 153, 154, 155, 165, 166, 167, 177, 178, 179], "bbox": [0.7417717717717718, 0.33322, 1.0, 0.8281000000000001], "iscrowd": 0, "area": 15497.94515, "rle": {"size": [500, 333], "counts": "Vlh33]?6K5J5K4M3L4L5L3L4L4M3L4L4hCZN_:i1]EaNZ:b1bEiNW:W1eESOT:n0hEZOT:f0hEAT:`0cELZ:5ZE:a:`2bE\\Kf9g4nEdKo9S5N8G10000O10000O10000M3J6J6J6J6J6N2O1N2N2O1N2N2O1N2N2O10000O10000O1000000O10000001O001O1O001O1O001O1O3M3L4M3M6J8H4L3M3M4L3M3mITO"}, "label": "women in all black"}]} {"id": 186198, "image": "COCO_train2014_000000186198.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white skiis on a snowcoved mountainside\"."}], "task": "refering", "answer_template": "The \"white skiis on a snowcoved mountainside\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 246, 247, 263, 264, 265], "bbox": [0.42781250000000004, 0.6438927738927739, 0.7676875000000001, 0.798088578088578], "iscrowd": 0, "area": 1324.9044000000006, "rle": {"size": [429, 640], "counts": "\\Sc33X=201O10000O10O01000O001O0001O01O00000010O000001O01O00000000010O00000000001O01O00000001O000001O01O00000000010O00000000001O01O0000000000001O000001O000001O0000000000001O0000000000001O00001OhZW13RehN5J7J5M4L1O1O1O1O0010000001N1000001N2O001O1O1N2O2N2N2N2M_Tn1"}, "label": "white skiis on a snowcoved mountainside"}]} {"id": 186198, "image": "COCO_train2014_000000186198.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing a black and yellow stripped jacket\"."}], "task": "refering", "answer_template": "The \"a person wearing a black and yellow stripped jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 216, 217, 218, 239, 240, 241, 262, 263, 264, 285, 286, 287], "bbox": [0.412734375, 0.3385314685314685, 0.55315625, 0.8500466200466199], "iscrowd": 0, "area": 10934.238949999999, "rle": {"size": [429, 640], "counts": "dm^39o:HhFo0k8TOQGW1e8lNWGY1T8gN`F0O4V1[1T8POeFIP1\\1V8@eGf0S8@iG`0X8CfG7`8K`G2`80aGUO[OJS9S1cGRO\\OHo8Z1eGnN]OEj8b1iGjN^OBf8P2aGbNI]Oe8`3\\G_Lc8c3]G\\Lb8e3_GZL`8h3_GXLa8h3`GVL`8l3_GTLa8l3_GTL`8n3_GRL_7IaH201O001N2O1O0nMPEf1R;UNWEf1[;M1O1N2XOQT29UlM81FnCXOQ in this image.", "objects": [{"patches": [160, 161, 162, 163, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 204, 205, 206, 207, 208, 209, 219, 220, 221, 222, 223, 224, 235, 236, 237, 238, 239, 250, 251, 252, 253, 254, 265, 266, 267, 268, 269, 281, 282, 283, 284, 296, 297, 298, 299, 312, 313, 327, 328, 342, 343], "bbox": [0.6212646370023418, 0.442484375, 1.0, 0.9942968750000001], "iscrowd": 0, "area": 33055.928499999995, "rle": {"size": [640, 427], "counts": "PPV51nc0102M3M3M3N1N3M3N2M3M3M3L4M2M4L4M3L4M3L4L4M3L3N3L4M3L4L4M4K6K4K6J5L5J5L4K6J5L5D;^Oc0G8L4L5J5L5K3M3M3L4M3N2M3N2N2M3N2M3N2M3N2N2M3N2M3N2M3N2N2M3N2N2N2N2N2aFdGV7^8dHfG[7]8_HhG`7Z8ZHjGf7X8THmGk7U8oGoGQ8S8hGSHW8o7cGUH]8m7]GXHb8j7XGZHh8h7RG]Hm8e7mF_HS9X9O001O001O001O000O01O100O1O102M2N3N1N3M2N3N1N3M2N3K4K6K4L3L5L4K4M4K5L3L5L4L3L5L4K4M4K5L3L5L4L3L5L3L5L4K4M4K4M4L4K4M4K4M4K5L3L5L3M4K5WLo@X2Z?eMn@j1_?SNg@]1e?aNa@o0l?mN\\@c0P`0ZOW@5V`0HV2I7J^E"}, "label": "the white puffy mattress"}]} {"id": 309084, "image": "COCO_train2014_000000309084.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the right side of a bed next to a man and a dog\"."}], "task": "refering", "answer_template": "The \"the right side of a bed next to a man and a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [160, 161, 162, 163, 175, 176, 177, 178, 179, 190, 191, 192, 193, 194, 204, 205, 206, 207, 208, 209, 219, 220, 221, 222, 223, 224, 235, 236, 237, 238, 239, 250, 251, 252, 253, 254, 265, 266, 267, 268, 269, 281, 282, 283, 284, 296, 297, 298, 299, 312, 313, 327, 328, 342, 343], "bbox": [0.6212646370023418, 0.442484375, 1.0, 0.9942968750000001], "iscrowd": 0, "area": 33055.928499999995, "rle": {"size": [640, 427], "counts": "PPV51nc0102M3M3M3N1N3M3N2M3M3M3L4M2M4L4M3L4M3L4L4M3L3N3L4M3L4L4M4K6K4K6J5L5J5L4K6J5L5D;^Oc0G8L4L5J5L5K3M3M3L4M3N2M3N2N2M3N2M3N2M3N2N2M3N2M3N2M3N2N2M3N2N2N2N2N2aFdGV7^8dHfG[7]8_HhG`7Z8ZHjGf7X8THmGk7U8oGoGQ8S8hGSHW8o7cGUH]8m7]GXHb8j7XGZHh8h7RG]Hm8e7mF_HS9X9O001O001O001O000O01O100O1O102M2N3N1N3M2N3N1N3M2N3K4K6K4L3L5L4K4M4K5L3L5L4L3L5L4K4M4K5L3L5L4L3L5L3L5L4K4M4K4M4L4K4M4K4M4K5L3L5L3M4K5WLo@X2Z?eMn@j1_?SNg@]1e?aNa@o0l?mN\\@c0P`0ZOW@5V`0HV2I7J^E"}, "label": "the right side of a bed next to a man and a dog"}]} {"id": 309084, "image": "COCO_train2014_000000309084.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large mattress with a blue flower on it\"."}], "task": "refering", "answer_template": "The \"a large mattress with a blue flower on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [261, 262, 263, 276, 277, 278, 279, 290, 291, 292, 293, 294, 304, 305, 306, 307, 308, 309, 310, 315, 316, 319, 320, 321, 322, 323, 324, 325, 326, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.0034192037470725995, 0.732875, 0.7802107728337236, 0.9885781250000001], "iscrowd": 0, "area": 21628.943049999998, "rle": {"size": [640, 427], "counts": "hW11fc09G9H81O1O0000000000000O100000000000001O000000000000000000000000000000000000001O001O1O1O1O001O1O1O001O1O1O001O1O1O1O00000000O10000000000000000O10000000000000000O100000000O10000O100O10000O10000O100O10000O100O10000O100O10000O100O10000O100N2N2M3N2N2M3N2N2N2O1N2O1N101N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1N2O1O1N2O1N2O1N2O1N2O1O11O1O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1O1O001O1O001O1O1O001O1O001O1O1O001O1N101O1O1O002N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N2N3M2N2N2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N2N001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O00gTj1"}, "label": "a large mattress with a blue flower on it"}]} {"id": 423776, "image": "COCO_train2014_000000423776.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bike of the girl in front wearing white t - shirt\"."}], "task": "refering", "answer_template": "The \"the bike of the girl in front wearing white t - shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 160, 169, 170, 171, 172, 181, 182, 183, 184, 194, 195, 196, 206, 207, 208], "bbox": [0.13342342342342342, 0.73576, 0.4292792792792792, 0.99888], "iscrowd": 0, "area": 6301.43955, "rle": {"size": [500, 333], "counts": "gke01c?5K4L5K5K4L5K5K5K4L00000SCXOn:i0kD_OS;a0gDES;POSD[19eNMU1a;QOYDU17kNHR1f;PO[DS15:_;dN\\DR13<`;bN^DR3a;PM^DP3a;RM^Dn2c;RM\\Dn2e;RMZDn2g;RMXDn2i;fLXDd3m;O1O1O1O1O0000000000001O00000000000000000M3J6K5K5M200001O01O000aM[Cf1oZDOb;9dC5\\<_1M34N4_NRCE1HnO1O1O1O1O001OR[l2"}, "label": "the bike of the girl in front wearing white t - shirt"}]} {"id": 423776, "image": "COCO_train2014_000000423776.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bicycle with the basket the woman ' s riding on\"."}], "task": "refering", "answer_template": "The \"the bicycle with the basket the woman ' s riding on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [158, 159, 160, 169, 170, 171, 172, 181, 182, 183, 184, 194, 195, 196, 206, 207, 208], "bbox": [0.13342342342342342, 0.73576, 0.4292792792792792, 0.99888], "iscrowd": 0, "area": 6301.43955, "rle": {"size": [500, 333], "counts": "gke01c?5K4L5K5K4L5K5K5K4L00000SCXOn:i0kD_OS;a0gDES;POSD[19eNMU1a;QOYDU17kNHR1f;PO[DS15:_;dN\\DR13<`;bN^DR3a;PM^DP3a;RM^Dn2c;RM\\Dn2e;RMZDn2g;RMXDn2i;fLXDd3m;O1O1O1O1O0000000000001O00000000000000000M3J6K5K5M200001O01O000aM[Cf1oZDOb;9dC5\\<_1M34N4_NRCE1HnO1O1O1O1O001OR[l2"}, "label": "the bicycle with the basket the woman ' s riding on"}]} {"id": 243574, "image": "COCO_train2014_000000243574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cat with its ears down , facing another cat\"."}], "task": "refering", "answer_template": "The \"a cat with its ears down , facing another cat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 77, 78, 79, 80, 81, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263], "bbox": [0.00222, 0.29899532710280374, 0.63152, 1.0], "iscrowd": 0, "area": 87715.55554999999, "rle": {"size": [428, 500], "counts": "Td0c6b4W2O10000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O10000O100O10000O100O100001O00001O000010O0001O00001O00001O00001O00001O00001O00001O00001O00001O001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00001O00000000O10000O10000O10000O10000O10000O1000000O10000O10000O10000O10000O10000O10000O10000O10000O10000O1000000000000000000000000001O0000000000000000000000000000001O000000000000000000000000000000001O000000000000000000000000000000001O000000000000000000000000000000001O0000000000000000000000000000001O00000000000000000000000000001O1O1O2N1O2N1O1O2N1_KWK=j4AXK>j4^OZK`0g4]O\\Kb0e4\\O]Kc0e4YO^Kf0c4WO`Kh0b4UO`Kj0a4SOcKk0^4QOfKn0\\4nNgKQ1Z4kNjKT1W4iNlKV1V4fNmKY1T4cNPL\\1Q4aNSL]1o3^NULa1l3\\NWLc1k3XNYLg1h3VN[Li1g3RN]Lm1h3lM\\LR2h3gM\\LX2i3aMZL^2j3[MZLd2k3UMXLj2l3PMWLS2j4eMe5IVf\\2"}, "label": "a cat with its ears down , facing another cat"}]} {"id": 243574, "image": "COCO_train2014_000000243574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tabby cat sitting outside a door\"."}], "task": "refering", "answer_template": "The \"a tabby cat sitting outside a door\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 87, 102, 103, 104, 105, 119, 120, 121, 122, 123, 124, 137, 138, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 160, 161, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 211, 212, 213, 214, 229, 230, 231, 232, 247, 248, 249, 250, 266, 267], "bbox": [0.6309400000000001, 0.28539719626168225, 1.0, 0.9505607476635515], "iscrowd": 0, "area": 29135.583550000003, "rle": {"size": [428, 500], "counts": "gYT4:P=9G3M2N2M4M2N3M2WOnNVEU1h:ROPEP1n:XOjDj0U;h0N1O2N1O2M2O1O2N1O100O2N1O100O2N1O101N1O1O100_OVM_Fj2`9ZM]Fg2`9^M]Fc2`9cM]F]2`9hM]FY2`9k0N2N200O10O0100O010O10O101N10001N100O2OgG^Lh6d3mHVMc6k2QIoM8iNi4Y3cJMY5W4LYHfJ^7W5`HjJ`7W5]HlJb7U5[HmJf7S5XHnJh7[5001O01O01O00001O01O01O0O1O2M2N3M2O2M3M2N3N2M2N2O10O01O1O100O1O1O100O1001O0000000O2O00001O1O1O2]H[K[6g4_I]Ka6d4[I`Kd6g4QI^Kn6k4eHYK[7]5O001O1O1N2O001O1O1O1O0J7ZOf0_Oa0K5K5J5L5K5J6K4L5J5N3M2N3M2N3M3M2N3M2N3M2N3M3M2N3M2N3M3M3M4L3M4L3M4M2N3M2N2N3L3N3M2N3M2N5K4L5J5L5KbH"}, "label": "a tabby cat sitting outside a door"}]} {"id": 243574, "image": "COCO_train2014_000000243574.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cat right in front of the window\"."}], "task": "refering", "answer_template": "The \"the cat right in front of the window\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 87, 102, 103, 104, 105, 119, 120, 121, 122, 123, 124, 137, 138, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 160, 161, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 211, 212, 213, 214, 229, 230, 231, 232, 247, 248, 249, 250, 266, 267], "bbox": [0.6309400000000001, 0.28539719626168225, 1.0, 0.9505607476635515], "iscrowd": 0, "area": 29135.583550000003, "rle": {"size": [428, 500], "counts": "gYT4:P=9G3M2N2M4M2N3M2WOnNVEU1h:ROPEP1n:XOjDj0U;h0N1O2N1O2M2O1O2N1O100O2N1O100O2N1O101N1O1O100_OVM_Fj2`9ZM]Fg2`9^M]Fc2`9cM]F]2`9hM]FY2`9k0N2N200O10O0100O010O10O101N10001N100O2OgG^Lh6d3mHVMc6k2QIoM8iNi4Y3cJMY5W4LYHfJ^7W5`HjJ`7W5]HlJb7U5[HmJf7S5XHnJh7[5001O01O01O00001O01O01O0O1O2M2N3M2O2M3M2N3N2M2N2O10O01O1O100O1O1O100O1001O0000000O2O00001O1O1O2]H[K[6g4_I]Ka6d4[I`Kd6g4QI^Kn6k4eHYK[7]5O001O1O1N2O001O1O1O1O0J7ZOf0_Oa0K5K5J5L5K5J6K4L5J5N3M2N3M2N3M3M2N3M2N3M2N3M3M2N3M2N3M3M3M4L3M4L3M4M2N3M2N2N3L3N3M2N3M2N5K4L5J5L5KbH"}, "label": "the cat right in front of the window"}]} {"id": 227198, "image": "COCO_train2014_000000227198.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"face of cow standing in front of sun\"."}], "task": "refering", "answer_template": "The \"face of cow standing in front of sun\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 80, 81, 82, 83, 94, 95, 96, 97, 98, 109, 110, 111, 112, 113, 123, 124, 125, 126, 127, 128, 139, 140, 141, 142, 143, 154, 155, 156, 157, 169, 170, 171, 172, 184, 185, 186, 187, 200, 202], "bbox": [0.2594626168224299, 0.238046875, 0.5817990654205607, 0.57775], "iscrowd": 0, "area": 19718.499500000005, "rle": {"size": [640, 428], "counts": "VcU24jc06J6I6K6I7J6I4M4L3L5L3L5L3N3M2N3M2O2oNaM\\@a2c?bMY@a2e?bMW@`2IgMS?Mo@_2D[Nn>YOZA^2F`Nj>m3K4L5Je0\\O1O001O1O001O1O001O1O00nNaBbK`=[4T1L4M3L4M3M36IV1kN001O1O0O2O1O001O000O100O1O1O1O2O0OE;L4N2M4M4K4I7H8H8O100O1O1O1O100O1O`A`Kb=_4UBSLd=k3XBbLa=]3]BhL`=W3^BnLa=P3^BSMa=l2]BXMc=f2\\B^Mb=b2\\BaMd=^401O000O10O1O1hNaBmK`=o3jBjKV=S4TCfKk in this image.", "objects": [{"patches": [79, 80, 81, 82, 83, 94, 95, 96, 97, 98, 109, 110, 111, 112, 113, 123, 124, 125, 126, 127, 128, 139, 140, 141, 142, 143, 154, 155, 156, 157, 169, 170, 171, 172, 184, 185, 186, 187, 200, 202], "bbox": [0.2594626168224299, 0.238046875, 0.5817990654205607, 0.57775], "iscrowd": 0, "area": 19718.499500000005, "rle": {"size": [640, 428], "counts": "VcU24jc06J6I6K6I7J6I4M4L3L5L3L5L3N3M2N3M2O2oNaM\\@a2c?bMY@a2e?bMW@`2IgMS?Mo@_2D[Nn>YOZA^2F`Nj>m3K4L5Je0\\O1O001O1O001O1O001O1O00nNaBbK`=[4T1L4M3L4M3M36IV1kN001O1O0O2O1O001O000O100O1O1O1O2O0OE;L4N2M4M4K4I7H8H8O100O1O1O1O100O1O`A`Kb=_4UBSLd=k3XBbLa=]3]BhL`=W3^BnLa=P3^BSMa=l2]BXMc=f2\\B^Mb=b2\\BaMd=^401O000O10O1O1hNaBmK`=o3jBjKV=S4TCfKk in this image.", "objects": [{"patches": [93, 94, 98, 99, 100, 101, 102, 107, 108, 109, 113, 114, 115, 116, 122, 123, 124, 128, 129, 130, 131, 137, 138, 139, 143, 144, 145, 152, 153, 154, 158, 159, 167, 168, 172, 173], "bbox": [0.17376168224299066, 0.28240625, 0.8133177570093458, 0.521484375], "iscrowd": 0, "area": 13848.04725, "rle": {"size": [640, 428], "counts": "n`^12ac0f0ZOf0ZOf0ZO=C=B?FWOl_OWNU`0f1S@PNQ`0n1V@iMn?h1]14K4j^OnMX`0U2b_OXNS`0m1f_OaNn?c1l_OeNo?e20010O010O00010O0100ROg@TMZ?o2j02N3M3M2N3M3M3Q@\\L\\?f3Z@`Lg?l31O1N2I7`Mk_Ol0^`0QOm_O<^`0@m_O:R`0CX@7g?F^@:`?Ce@_OUA>l>@XA=h>B\\A:f>E\\A9e>E_A8b>GaA6_>IeA4\\>KgA2[>LhA1Y>MkA0V>OmANT>1WZj1W1]gUN3N2N2N3M2N2VMZNXCi1goNiAP1Z>QOcAQ1]>SO\\AP1d>TOTAd1X>`NaAd1^>`NZAd1f>bNPAb1P?g10001O0000I7F:H8N2N2N2N2N20001O0000001O0000001O00000O2NN1N32N1N2O1O2M2O1O1N2O2N1N3N1O2M2O001N101O0O2O001N10001N101O0O2O001O101O001O001O001O001O001L3M3M3M2N3M3N2M3K5K5J6K5N20O010001N1000000O1N2N2N2M3NXja1"}, "label": "the side of a cow looking towards the camera behind another cow"}]} {"id": 227198, "image": "COCO_train2014_000000227198.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"all the cows except the first two from the front\"."}], "task": "refering", "answer_template": "The \"all the cows except the first two from the front\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 98, 99, 100, 101, 102, 107, 108, 109, 113, 114, 115, 116, 122, 123, 124, 128, 129, 130, 131, 137, 138, 139, 143, 144, 145, 152, 153, 154, 158, 159, 167, 168, 172, 173], "bbox": [0.17376168224299066, 0.28240625, 0.8133177570093458, 0.521484375], "iscrowd": 0, "area": 13848.04725, "rle": {"size": [640, 428], "counts": "n`^12ac0f0ZOf0ZOf0ZO=C=B?FWOl_OWNU`0f1S@PNQ`0n1V@iMn?h1]14K4j^OnMX`0U2b_OXNS`0m1f_OaNn?c1l_OeNo?e20010O010O00010O0100ROg@TMZ?o2j02N3M3M2N3M3M3Q@\\L\\?f3Z@`Lg?l31O1N2I7`Mk_Ol0^`0QOm_O<^`0@m_O:R`0CX@7g?F^@:`?Ce@_OUA>l>@XA=h>B\\A:f>E\\A9e>E_A8b>GaA6_>IeA4\\>KgA2[>LhA1Y>MkA0V>OmANT>1WZj1W1]gUN3N2N2N3M2N2VMZNXCi1goNiAP1Z>QOcAQ1]>SO\\AP1d>TOTAd1X>`NaAd1^>`NZAd1f>bNPAb1P?g10001O0000I7F:H8N2N2N2N2N20001O0000001O0000001O00000O2NN1N32N1N2O1O2M2O1O1N2O2N1N3N1O2M2O001N101O0O2O001N10001N101O0O2O001O101O001O001O001O001O001L3M3M3M2N3M3N2M3K5K5J6K5N20O010001N1000000O1N2N2N2M3NXja1"}, "label": "all the cows except the first two from the front"}]} {"id": 481165, "image": "COCO_train2014_000000481165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bottle with the name lumier on it\"."}], "task": "refering", "answer_template": "The \"bottle with the name lumier on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 154, 155, 156, 176, 177, 178, 179, 180, 200, 201, 202, 203, 223, 224, 225, 226, 246, 247, 248, 269, 270, 271], "bbox": [0.682125, 0.3463694267515923, 0.8514687500000001, 0.692611464968153], "iscrowd": 0, "area": 12418.956150000004, "rle": {"size": [471, 640], "counts": "dVY6:S>;M3N2O1N101N101N2N101N101N1O2O0O2[Cg0j9V2L100O2N1O100O1O1O100O1O1O2O0H8I7000000000001O00000000000000000000O10000000000O10000000000O100000000O10000000006I7J7I1O1O1O0IdERL^:k39N2M3M3M2N3M3L4L4L3M4K5L4L4K4M4L4K5L4L3L5I7H8H8H8G8HR`[1"}, "label": "bottle with the name lumier on it"}]} {"id": 481165, "image": "COCO_train2014_000000481165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"full perfume bottle\"."}], "task": "refering", "answer_template": "The \"full perfume bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [132, 133, 154, 155, 156, 176, 177, 178, 179, 180, 200, 201, 202, 203, 223, 224, 225, 226, 246, 247, 248, 269, 270, 271], "bbox": [0.682125, 0.3463694267515923, 0.8514687500000001, 0.692611464968153], "iscrowd": 0, "area": 12418.956150000004, "rle": {"size": [471, 640], "counts": "dVY6:S>;M3N2O1N101N101N2N101N101N1O2O0O2[Cg0j9V2L100O2N1O100O1O1O100O1O1O2O0H8I7000000000001O00000000000000000000O10000000000O10000000000O100000000O10000000006I7J7I1O1O1O0IdERL^:k39N2M3M3M2N3M3L4L4L3M4K5L4L4K4M4L4K5L4L3L5I7H8H8H8G8HR`[1"}, "label": "full perfume bottle"}]} {"id": 186255, "image": "COCO_train2014_000000186255.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra behind another zebra facing the camera\"."}], "task": "refering", "answer_template": "The \"a zebra behind another zebra facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 120, 121, 122, 123, 136, 137, 138, 152, 153, 154, 165, 166, 180, 181, 196, 197, 212, 213, 214, 228, 229, 230, 244, 245, 246, 260, 261, 262, 276, 277, 280, 292, 293, 296, 298, 308, 309, 312, 314, 324, 330, 331, 339, 340, 346, 355, 356, 362], "bbox": [0.21353711790393012, 0.25225, 0.7159606986899564, 0.991578125], "iscrowd": 0, "area": 20201.057399999998, "rle": {"size": [640, 458], "counts": "Zkm11mc03N2M3N2M3N1O1N3N1N3N1N2O2SIVO`Jl0[5[ObJg0Y5@bJb0Z5EbJ=X5JeJ7W50dJ2X55dJMW59eJIV5?fJCU5d0fJ^OV5h0gJYOX5k0cJXOZ5l0_JWOa5n0VJVOj5o0mITOU6o0cIUO]6V1TInNl6b1`HbNa7m1kGVNV8Z2WGiMi8h2bF\\M^9a3aEcL`:]60O10001L3K5K5K6J5J7J7I6J7O03N3L3NiNoCaJn;^5_DYJ_;f5mDQJP;n5]EjI`:T6f101N101N2N101N101N1O2O0O2O0O2XKYBU3h=fLfBQ3[=hLUCm2lD>CB=D=Bhoi00WPVO1O1m\\OOYb02b]O7Xb0Jc]O?Xb0Be]Of0Vb0[Oh]Oj0Sb0WOl]On0oa0SOo]OT1la0mNR^OX1ia0iNV^O\\1ia0aNV^Oe1ia0:010O00dFmMi0T2jNZNT1f1`NhN_1W1[NQOc1P1ZNTOd1l0[NWOc1j0[NYOc1g0]N[O_1g0UOgNf0]1l82N100O1O100O1O1O100O1O100O10O0100O1O100O10gMP_Ob1o`0^NT_O`1l`0`NU_O^1k`0bNX_O\\1i`0bNZ_O\\1g`0cNZ_O\\1g`0bNZ_O]1h`0aNY_O_1h`0_NY_Oa1h`0^NX_Oa1j`0]NW_Oc1j`0[NV_Of1k`0XNV_Oh1k`0WNU_Oh1m`0VNT_Oj1R8XNf0h1XO]Ne0b1[OcNa0]1^OgN?8TG2[8LaH9j5IdIHP8:aH6g5HTJ\\Od7j0_H4f5Fj2;^G2d5Cm2>^G1b5@P3b0]GLd5Bm2f0_GFd5Cm2j0^GBe5Dl2m0^G^Of5Dk2R1^GXOh5Fi2T1_GUOg5Gj2W1^GQOh5Hh2[1`GkNh5Ij2]1]GhNj5Kl2\\1YGhNk5Ln2\\1VGgNl5LR3\\1RGfNl5MV3\\1oFcNl50Y3\\1lFaNk52\\3[1lF`Ng55a3W1lFaNc57e3T1mFaN^5:i3Q1mFaN[5=k3o0]MPOg2m0YMROk2j0VMUOn2g0SMXOP3e0QMZOS3b0mL^OW3>jLAZ3 in this image.", "objects": [{"patches": [67, 68, 69, 70, 71, 72, 83, 84, 85, 86, 87, 88, 100, 101, 102, 103, 117, 118, 119, 134, 135, 136, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 181, 182, 183, 184, 185, 186, 187, 188, 198, 199, 200, 201, 202, 203, 204, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 246, 247, 248, 249, 250, 251, 252, 253, 262, 263, 264, 265, 266, 267, 268, 269, 270, 277, 278, 280, 281, 282, 283, 284, 285, 286, 293, 294, 296, 297, 298, 300, 301, 302, 309, 310, 312, 313, 314, 316, 317, 318, 325, 326, 328, 329, 330, 333, 334, 341, 342, 344, 345, 349, 350, 357, 360, 361, 366], "bbox": [0.20818777292576418, 0.17099999999999999, 0.9431441048034935, 0.9881562500000001], "iscrowd": 0, "area": 70952.94605000001, "rle": {"size": [640, 458], "counts": "gTl13lc01N3N2N1N3N2N2M2O2N2M3N1O2M3N1O2M3N2M210O1O1O00100O001O1O10O01O1O1O010O1O001O100O001O1O100O001O1L4N2N101N2QK^N^Gc1W8fNiG\\1l7nNTHS1a7VO_Hl0V7^OjHc0k6FVI;`6N_I5[5n0fJSOV5P1iJSOQ5P1PKQOl4R1SKQOh4Q1YKPOb4T1]KnN_4T1bKnNY4U1gKnNR4U1nKnNm3T1TLoNe3U1ZLnNa3T1`LoNhNmLT1W43oNcNSMQ1Q4=nN\\N[Mn0j3e0oNVN`Ml0d3o0nNPNgMh0^3X1nNhMnMh0V3`1oN]M[Nh0j2k1nNQMhNj0\\2U2oNfLTOl0P2^2oNZLAm0c1i2Y:0O010O101N2O1\\Mc2J7J6I7J6I7I6L5L4L4L4L4L3M4L4L4L4L4L3L5K5K4001O1O1O1O2N1O1O1N2N2N2N2N2N3M2N2N3McGeFX6Y9fIPGV6n8gIZGU6d8iIdGT6Y8jIPHQ6n7lI[HQ6b7mIgHo5V7oIRIm5l6QJYIo5d6nIaIS6\\6kIiIT6U6jIoIW6n5gIVJZ6h5cI]J\\6d5^IaJc6_5XIfJh6[5RIiJn6X5mHlJT7U5eHQK[7o4`HUK`7l4[HXKf7b8j1UN in this image.", "objects": [{"patches": [67, 68, 69, 70, 71, 72, 83, 84, 85, 86, 87, 88, 100, 101, 102, 103, 117, 118, 119, 134, 135, 136, 150, 151, 152, 153, 166, 167, 168, 169, 170, 171, 181, 182, 183, 184, 185, 186, 187, 188, 198, 199, 200, 201, 202, 203, 204, 214, 215, 216, 217, 218, 219, 220, 221, 230, 231, 232, 233, 234, 235, 236, 237, 246, 247, 248, 249, 250, 251, 252, 253, 262, 263, 264, 265, 266, 267, 268, 269, 270, 277, 278, 280, 281, 282, 283, 284, 285, 286, 293, 294, 296, 297, 298, 300, 301, 302, 309, 310, 312, 313, 314, 316, 317, 318, 325, 326, 328, 329, 330, 333, 334, 341, 342, 344, 345, 349, 350, 357, 360, 361, 366], "bbox": [0.20818777292576418, 0.17099999999999999, 0.9431441048034935, 0.9881562500000001], "iscrowd": 0, "area": 70952.94605000001, "rle": {"size": [640, 458], "counts": "gTl13lc01N3N2N1N3N2N2M2O2N2M3N1O2M3N1O2M3N2M210O1O1O00100O001O1O10O01O1O1O010O1O001O100O001O1O100O001O1L4N2N101N2QK^N^Gc1W8fNiG\\1l7nNTHS1a7VO_Hl0V7^OjHc0k6FVI;`6N_I5[5n0fJSOV5P1iJSOQ5P1PKQOl4R1SKQOh4Q1YKPOb4T1]KnN_4T1bKnNY4U1gKnNR4U1nKnNm3T1TLoNe3U1ZLnNa3T1`LoNhNmLT1W43oNcNSMQ1Q4=nN\\N[Mn0j3e0oNVN`Ml0d3o0nNPNgMh0^3X1nNhMnMh0V3`1oN]M[Nh0j2k1nNQMhNj0\\2U2oNfLTOl0P2^2oNZLAm0c1i2Y:0O010O101N2O1\\Mc2J7J6I7J6I7I6L5L4L4L4L4L3M4L4L4L4L4L3L5K5K4001O1O1O1O2N1O1O1N2N2N2N2N2N3M2N2N3McGeFX6Y9fIPGV6n8gIZGU6d8iIdGT6Y8jIPHQ6n7lI[HQ6b7mIgHo5V7oIRIm5l6QJYIo5d6nIaIS6\\6kIiIT6U6jIoIW6n5gIVJZ6h5cI]J\\6d5^IaJc6_5XIfJh6[5RIiJn6X5mHlJT7U5eHQK[7o4`HUK`7l4[HXKf7b8j1UN in this image.", "objects": [{"patches": [67, 68, 69, 70, 71, 83, 84, 85, 86, 87, 88, 89, 100, 101, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215], "bbox": [0.5657000000000001, 0.27531722054380664, 1.0, 0.9870090634441088], "iscrowd": 0, "area": 40696.29425, "rle": {"size": [331, 500], "counts": "onk22Y:0O2bF1^8O_G5_8M^G6a8J\\G;a8G[G=d8CZGa0c8AYGc0f8]OWGh0f8ZOWGi0h8?M2O2M2O2M2O2M2O2[Of0M3M4L3M3M4L3M3M4L3M2N1O1O2N100O1H9G8J6O101N100O101N10000O2N1L4L5K4L4L4O2N10000O2O000O1O2K4M3L4M4L3O100O2N100O2O0O100H9E9O11O1O001O1O1O001O1O001O1O001N2O1ON2M3M3M3M300001O1O001O1O1O001O1O001O1O001O1O1OO1O1O1O1O1N2O1O1O1O1O02O1O1O1O1O1O001O1O1O1O1O1O1O1O1O00O1O1O1O1O1O1O1O1O1O1O1N2O10000000000000O10000000000000000000001O00000000000000001O1O1O1O001O1O1O1O1O002N2nIdLQ5_3hJgLV5[3cJkL\\5W3^JnLa5o3N3N1O1O1O1O2N1O1O1O2NTK"}, "label": "a green stuffed animal on top of a baby"}]} {"id": 292751, "image": "COCO_train2014_000000292751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green stuffed animal\"."}], "task": "refering", "answer_template": "The \"green stuffed animal\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 69, 70, 71, 83, 84, 85, 86, 87, 88, 89, 100, 101, 102, 103, 104, 105, 106, 107, 118, 119, 120, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215], "bbox": [0.5657000000000001, 0.27531722054380664, 1.0, 0.9870090634441088], "iscrowd": 0, "area": 40696.29425, "rle": {"size": [331, 500], "counts": "onk22Y:0O2bF1^8O_G5_8M^G6a8J\\G;a8G[G=d8CZGa0c8AYGc0f8]OWGh0f8ZOWGi0h8?M2O2M2O2M2O2M2O2[Of0M3M4L3M3M4L3M3M4L3M2N1O1O2N100O1H9G8J6O101N100O101N10000O2N1L4L5K4L4L4O2N10000O2O000O1O2K4M3L4M4L3O100O2N100O2O0O100H9E9O11O1O001O1O1O001O1O001O1O001N2O1ON2M3M3M3M300001O1O001O1O1O001O1O001O1O001O1O1OO1O1O1O1O1N2O1O1O1O1O02O1O1O1O1O1O001O1O1O1O1O1O1O1O1O00O1O1O1O1O1O1O1O1O1O1O1N2O10000000000000O10000000000000000000001O00000000000000001O1O1O1O001O1O1O1O1O002N2nIdLQ5_3hJgLV5[3cJkL\\5W3^JnLa5o3N3N1O1O1O1O2N1O1O1O2NTK"}, "label": "green stuffed animal"}]} {"id": 292751, "image": "COCO_train2014_000000292751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a purple teddy bear\"."}], "task": "refering", "answer_template": "The \"a purple teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 18, 19, 20, 36, 37, 38, 39, 54, 55, 56, 57, 72, 73, 74, 75, 76, 90, 91, 92, 93, 94, 108, 109, 110, 111, 112, 113, 126, 129, 130, 131, 132, 144, 149, 150, 168], "bbox": [0.00148, 0.0038066465256797583, 0.38084, 0.8442598187311178], "iscrowd": 0, "area": 19768.4871, "rle": {"size": [331, 500], "counts": "RA1000000000000000O100000000O100O10O002N100O1O1O1O1O100O3M4eGaNb7b0dHE^O1m78lH3U7ISI2m6LZI0e6NbIM_60iIJX63R2MSPT3"}, "label": "a purple teddy bear"}]} {"id": 292751, "image": "COCO_train2014_000000292751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pruple bear next to the baby\"."}], "task": "refering", "answer_template": "The \"the pruple bear next to the baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 18, 19, 20, 36, 37, 38, 39, 54, 55, 56, 57, 72, 73, 74, 75, 76, 90, 91, 92, 93, 94, 108, 109, 110, 111, 112, 113, 126, 129, 130, 131, 132, 144, 149, 150, 168], "bbox": [0.00148, 0.0038066465256797583, 0.38084, 0.8442598187311178], "iscrowd": 0, "area": 19768.4871, "rle": {"size": [331, 500], "counts": "RA1000000000000000O100000000O100O10O002N100O1O1O1O1O100O3M4eGaNb7b0dHE^O1m78lH3U7ISI2m6LZI0e6NbIM_60iIJX63R2MSPT3"}, "label": "the pruple bear next to the baby"}]} {"id": 358289, "image": "COCO_train2014_000000358289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white and blue bus with tan fox banner\"."}], "task": "refering", "answer_template": "The \"white and blue bus with tan fox banner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 69, 70, 71, 72, 73, 74, 75, 76, 77, 92, 93, 94, 95, 96, 97, 98, 99, 100, 115, 116, 117, 118, 119, 120, 121, 122, 138, 139, 140, 141, 142, 143, 144, 145, 161, 162, 163, 164, 165, 166, 167, 168, 184, 185, 186, 187, 188, 189, 190, 191, 192, 208, 209, 211, 212], "bbox": [0.019359375, 0.18936768149882904, 0.36584374999999997, 0.6159484777517564], "iscrowd": 0, "area": 34766.288799999995, "rle": {"size": [427, 640], "counts": "RU51U=5K6J5SEO[87hFh0P9]O\\F[1[9`1I8G7I4L6J6J7I6N2O0O10000O10000O100O10000O10000O10000O100O10000O10000O10000O10000O1000000000000000000001O1O4L00001O0000000000000O100000000000000000000000000000000000001O000000000000000000000001O001O1O001O1O1O00000000000001O00000000000000000O100000000O100O100O10000O100O100O101N10000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000001O00000000000000000000000dLkIBU6:\\JYOd5b0e3K5L4K5L4K5K5L4K5LiVY5"}, "label": "white and blue bus with tan fox banner"}]} {"id": 358289, "image": "COCO_train2014_000000358289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a double - decker buss between a single bus and another double decker buss in a parking lot\"."}], "task": "refering", "answer_template": "The \"a double - decker buss between a single bus and another double decker buss in a parking lot\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 215, 216, 217, 218, 219, 220], "bbox": [0.343984375, 0.14784543325526933, 1.0, 0.6363700234192038], "iscrowd": 0, "area": 60191.90390000001, "rle": {"size": [427, 640], "counts": "mmk28dA?A?H800000001O000000000000000O100000001O00000000000000000O1000001O000000000000000000000O2O0000000000000000000000001O0O100000000000000000001O00000O1000000000000000001O000000000O1000000000001O0000000000000O10000000001O00000000000000000O10001O000000VM"}, "label": "a double - decker buss between a single bus and another double decker buss in a parking lot"}]} {"id": 358289, "image": "COCO_train2014_000000358289.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"passport application air bus running the road\"."}], "task": "refering", "answer_template": "The \"passport application air bus running the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 215, 216, 217, 218, 219, 220], "bbox": [0.343984375, 0.14784543325526933, 1.0, 0.6363700234192038], "iscrowd": 0, "area": 60191.90390000001, "rle": {"size": [427, 640], "counts": "mmk28dA?A?H800000001O000000000000000O100000001O00000000000000000O1000001O000000000000000000000O2O0000000000000000000000001O0O100000000000000000001O00000O1000000000000000001O000000000O1000000000001O0000000000000O10000000001O00000000000000000O10001O000000VM"}, "label": "passport application air bus running the road"}]} {"id": 104344, "image": "COCO_train2014_000000104344.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man over person in green ties right shoulder\"."}], "task": "refering", "answer_template": "The \"man over person in green ties right shoulder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 111, 112, 113, 133, 134, 135, 155, 156, 157, 177, 178, 179, 199, 200, 201, 221, 222, 223, 224, 243, 244, 245, 246, 266, 267, 268, 287, 288, 289, 290, 309, 310], "bbox": [0.06066993464052288, 0.20151960784313724, 0.1919281045751634, 0.6459150326797386], "iscrowd": 0, "area": 14342.746950000004, "rle": {"size": [612, 612], "counts": "n_f01ob04O1000000O100000000O10000\\H0SF1]4O]57nEi2j1PMU8?hES4d0_Ka9e0dEo3k0\\K^9U1VEd3[1WK\\9\\6aFgI_9Y6^FjIb9W6[FkIe9U6XFnIg9T6UFoIk9Q6QFSJo9m5`EdJ_:Q7000000O1000000O10000O11O001O1O1O0O2L4M2M4M3M2M4eHSE[6n;MmC]Ih;d6UD]Il;d6PD^IR in this image.", "objects": [{"patches": [90, 91, 111, 112, 113, 133, 134, 135, 155, 156, 157, 177, 178, 179, 199, 200, 201, 221, 222, 223, 224, 243, 244, 245, 246, 266, 267, 268, 287, 288, 289, 290, 309, 310], "bbox": [0.06066993464052288, 0.20151960784313724, 0.1919281045751634, 0.6459150326797386], "iscrowd": 0, "area": 14342.746950000004, "rle": {"size": [612, 612], "counts": "n_f01ob04O1000000O100000000O10000\\H0SF1]4O]57nEi2j1PMU8?hES4d0_Ka9e0dEo3k0\\K^9U1VEd3[1WK\\9\\6aFgI_9Y6^FjIb9W6[FkIe9U6XFnIg9T6UFoIk9Q6QFSJo9m5`EdJ_:Q7000000O1000000O10000O11O001O1O1O0O2L4M2M4M3M2M4eHSE[6n;MmC]Ih;d6UD]Il;d6PD^IR in this image.", "objects": [{"patches": [150, 151, 171, 172, 173, 174, 193, 194, 195, 196, 214, 215, 216, 217, 218, 236, 237, 238, 239, 258, 259, 260, 261, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 323, 324, 325, 326, 327, 328, 329, 347, 348, 349, 350, 351, 371, 372, 373, 394, 395], "bbox": [0.7031209150326797, 0.3097549019607843, 0.9667156862745098, 0.8163562091503268], "iscrowd": 0, "area": 24355.64295000001, "rle": {"size": [612, 612], "counts": "f[Q81Qc09G5K3M3L4M3M3M3L4M3L4F:F:EPOY:X2YE8:YN4XOY:T2\\E7>ZNL\\OY:R2^E_OO_Na0d0J\\OW:R2`EYO[1XOnN^OW:Q2`ESO`1^OiN_OV:P2bEmNd1EdN_OT:o1eEhNg1J`N@T:m1eEfNj1M]NAS:l1gEbNm11XNBS:l1hE\\NQ26UNBQ:k1mEUNR2>PNCQ:i1TIeNjLCQ:h1VIdNjLDP:g1WIdNiLFP:e1YIcNgLIP:b1\\IdNdLKo9`1bI`N_L1o9^1cI`N^L3o9\\1cIbN]L3R:W1bIfN]L4T:P1`IlN\\L5V:j0`IQOZL6Z:b0^IYOXL5^:WOYE`0W44RL7n:ZO^DMe4a0oK9m:XObDKc4e0nK8n:TO]Id0eK:P`0FP@;o?ER@;m?ES@=k?CV@?h?@X@d0d?]O\\@e0a?[O_@f0`?ZOa@c0a?]O_@b0b?^O_@`0b?@^@?c?A^@=c?C]@;e?E\\@9f?FZ@9g?E\\@7h?E[@8i?EZ@7[a0L5J5LX`;"}, "label": "a women sitting and taking notes"}]} {"id": 104344, "image": "COCO_train2014_000000104344.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a green shirt and ponytail is behind the man posing\"."}], "task": "refering", "answer_template": "The \"a woman in a green shirt and ponytail is behind the man posing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [150, 151, 171, 172, 173, 174, 193, 194, 195, 196, 214, 215, 216, 217, 218, 236, 237, 238, 239, 258, 259, 260, 261, 279, 280, 281, 282, 283, 301, 302, 303, 304, 305, 323, 324, 325, 326, 327, 328, 329, 347, 348, 349, 350, 351, 371, 372, 373, 394, 395], "bbox": [0.7031209150326797, 0.3097549019607843, 0.9667156862745098, 0.8163562091503268], "iscrowd": 0, "area": 24355.64295000001, "rle": {"size": [612, 612], "counts": "f[Q81Qc09G5K3M3L4M3M3M3L4M3L4F:F:EPOY:X2YE8:YN4XOY:T2\\E7>ZNL\\OY:R2^E_OO_Na0d0J\\OW:R2`EYO[1XOnN^OW:Q2`ESO`1^OiN_OV:P2bEmNd1EdN_OT:o1eEhNg1J`N@T:m1eEfNj1M]NAS:l1gEbNm11XNBS:l1hE\\NQ26UNBQ:k1mEUNR2>PNCQ:i1TIeNjLCQ:h1VIdNjLDP:g1WIdNiLFP:e1YIcNgLIP:b1\\IdNdLKo9`1bI`N_L1o9^1cI`N^L3o9\\1cIbN]L3R:W1bIfN]L4T:P1`IlN\\L5V:j0`IQOZL6Z:b0^IYOXL5^:WOYE`0W44RL7n:ZO^DMe4a0oK9m:XObDKc4e0nK8n:TO]Id0eK:P`0FP@;o?ER@;m?ES@=k?CV@?h?@X@d0d?]O\\@e0a?[O_@f0`?ZOa@c0a?]O_@b0b?^O_@`0b?@^@?c?A^@=c?C]@;e?E\\@9f?FZ@9g?E\\@7h?E[@8i?EZ@7[a0L5J5LX`;"}, "label": "a woman in a green shirt and ponytail is behind the man posing"}]} {"id": 104344, "image": "COCO_train2014_000000104344.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smiling man with brown hair wearing a checked shirt and green tie\"."}], "task": "refering", "answer_template": "The \"a smiling man with brown hair wearing a checked shirt and green tie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 33, 34, 51, 52, 53, 54, 55, 56, 57, 72, 73, 74, 75, 76, 77, 78, 79, 80, 94, 95, 96, 97, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 160, 161, 162, 163, 164, 165, 166, 167, 168, 182, 183, 184, 185, 186, 187, 188, 189, 190, 204, 205, 206, 207, 208, 209, 210, 211, 212, 226, 227, 228, 229, 230, 231, 232, 233, 234, 249, 250, 251, 252, 253, 254, 255, 271, 272, 273, 274, 275, 276, 277, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483], "bbox": [0.024722222222222222, 0.04718954248366013, 0.9865196078431373, 0.9865196078431373], "iscrowd": 0, "area": 182882.42964999998, "rle": {"size": [612, 612], "counts": "g[94Pc0?A?@?B?A?@?B?A?A>A`0A?A>B2M10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O101O000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10001N1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O10000O101O000O10000kJVIPNj6i1`ISNa6f1iIWNW6b1SJ[Nm5^1]J^Nd5[1gJaNY5X1QKeNo4T1[KiNe4P1eKlN\\4m0oKoNQ4j0YLSOg3f0cLWO]3b0mLZOT3?WM]Oi2QN_Oo1?WN]Oi1b0[NZOf1d0_NYOa1e0dNXO\\1g0iNUOW1i0mNTOT1j0oNUOQ1j0QOUOo0i0TOUOm0i0VOVOj0i0XOVOh0h0[OWOe0h0]OVOd0h0@VO`0h0CWO=h0EWO;g0HWO9g0JXO6g0LXO4f0OXO2f01YOOf03YOMe06ZOJd09ZOHe0:ZOFd0=[OCc0`0\\OAb0a0\\O@b0c0]O]Oa0f0^OZOa0h0]OYOa0j0^OVO`0m0_OSO`0o0_OQO?R1_OoN`0S1_OmNa0T1^OlNa0V1ZOnNf0S1TOROl0o0mNWOS1k0fNZOZ1g0_N_Oa1R9O100000000O10000000000O10000000000O100000000O10000000000O100000000O2O000000000O10000000000O100000000O10000000000001O00001O00001O0000001O00001O00001O00001O0000001O00010O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O0000001O000010O0001O1O2N2N1O2N1O2N2N1O2N2N1O2dG^M2c2JcM3_2GhM6Z2ElM8U2CQN;Q2@UN=m1]OZN`0g1\\O_Na0c1YOdNd0]1XOiNe0Y1UOnNh0T1SOROj0o0QOXOl0j0oN[Oo0g0kN^OT1c0hNAW1a0cNC]1>_NF`1kKJT4`0\\KFd4c0mJDR5e0^JBb5h0nI_OQ6j0`I\\O`6e6O00001O010O001O001O001O001O001O00001O001O001O001O001O001O00001O001O001O001O00001O001O00001O001O00001O001O01O01O001O00001O001O00001O001O00001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O000010O0001O001O00001O001O00001O001O00001O001O00001O001O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N100O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1N3N1O2N1N2O2N1N3N1O1O2M2O2N1O1N3N1O1N3N1H9F9G9G:F9H9F9GcX4"}, "label": "a smiling man with brown hair wearing a checked shirt and green tie"}]} {"id": 104344, "image": "COCO_train2014_000000104344.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an man wearing a white plaid shirt and a green tie , smiling\"."}], "task": "refering", "answer_template": "The \"an man wearing a white plaid shirt and a green tie , smiling\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 33, 34, 51, 52, 53, 54, 55, 56, 57, 72, 73, 74, 75, 76, 77, 78, 79, 80, 94, 95, 96, 97, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 144, 145, 146, 160, 161, 162, 163, 164, 165, 166, 167, 168, 182, 183, 184, 185, 186, 187, 188, 189, 190, 204, 205, 206, 207, 208, 209, 210, 211, 212, 226, 227, 228, 229, 230, 231, 232, 233, 234, 249, 250, 251, 252, 253, 254, 255, 271, 272, 273, 274, 275, 276, 277, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483], "bbox": [0.024722222222222222, 0.04718954248366013, 0.9865196078431373, 0.9865196078431373], "iscrowd": 0, "area": 182882.42964999998, "rle": {"size": [612, 612], "counts": "g[94Pc0?A?@?B?A?@?B?A?A>A`0A?A>B2M10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O101O000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O10001N1000000O1000000O10000O1000000O1000000O10000O1000000O1000000O1000000O10000O1000000O1000000O10000O101O000O10000kJVIPNj6i1`ISNa6f1iIWNW6b1SJ[Nm5^1]J^Nd5[1gJaNY5X1QKeNo4T1[KiNe4P1eKlN\\4m0oKoNQ4j0YLSOg3f0cLWO]3b0mLZOT3?WM]Oi2QN_Oo1?WN]Oi1b0[NZOf1d0_NYOa1e0dNXO\\1g0iNUOW1i0mNTOT1j0oNUOQ1j0QOUOo0i0TOUOm0i0VOVOj0i0XOVOh0h0[OWOe0h0]OVOd0h0@VO`0h0CWO=h0EWO;g0HWO9g0JXO6g0LXO4f0OXO2f01YOOf03YOMe06ZOJd09ZOHe0:ZOFd0=[OCc0`0\\OAb0a0\\O@b0c0]O]Oa0f0^OZOa0h0]OYOa0j0^OVO`0m0_OSO`0o0_OQO?R1_OoN`0S1_OmNa0T1^OlNa0V1ZOnNf0S1TOROl0o0mNWOS1k0fNZOZ1g0_N_Oa1R9O100000000O10000000000O10000000000O100000000O10000000000O100000000O2O000000000O10000000000O100000000O10000000000001O00001O00001O0000001O00001O00001O00001O0000001O00010O00001O00001O0000001O00001O00001O0000001O00001O00001O00001O0000001O00001O00001O00001O0000001O000010O0001O1O2N2N1O2N1O2N2N1O2N2N1O2dG^M2c2JcM3_2GhM6Z2ElM8U2CQN;Q2@UN=m1]OZN`0g1\\O_Na0c1YOdNd0]1XOiNe0Y1UOnNh0T1SOROj0o0QOXOl0j0oN[Oo0g0kN^OT1c0hNAW1a0cNC]1>_NF`1kKJT4`0\\KFd4c0mJDR5e0^JBb5h0nI_OQ6j0`I\\O`6e6O00001O010O001O001O001O001O001O00001O001O001O001O001O001O00001O001O001O001O00001O001O00001O001O00001O001O01O01O001O00001O001O00001O001O00001O00001O001O00001O001O00001O001O00001O001O00001O001O00001O001O000010O0001O001O00001O001O00001O001O00001O001O00001O001O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O2N100O2N1O2N1O1O2N1O2N1O1O2N1O2N1O1N3N1O2N1N2O2N1N3N1O1O2M2O2N1O1N3N1O1N3N1H9F9G9G:F9H9F9GcX4"}, "label": "an man wearing a white plaid shirt and a green tie , smiling"}]} {"id": 47001, "image": "COCO_train2014_000000047001.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece o pizza from the whole pizza in a plate\"."}], "task": "refering", "answer_template": "The \"a piece o pizza from the whole pizza in a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 270, 285, 286, 287, 288, 289, 290], "bbox": [0.424796875, 0.4898826291079812, 0.9109375, 0.8404460093896713], "iscrowd": 0, "area": 29230.841050000003, "rle": {"size": [426, 640], "counts": "m_a32U=4K4M3M3L4M3M3M3L4M4L3L4M3M3L4M3M3L4M4L3M3L4M3M3L4M3M3L4M4L3M3L4M3M3L4M3M3L5L3M3L4M3O100000001O01O0000000000001O01O0000000001O0O10000000001N1000000000001N10000000000O2O0000000000001N1000000000001O0O10000000001O000O100000001O00000O1000001O0000000O10001N10000O1000001N10000O10000O2O000O1000000O2O000O10000O101O0O1000000O101O0O10000O100O2O000O100O10001N10000O100O101O0O100O10000O2O0O10000O10001N100O10000O101N10000O100O101O0O10000O100O2O000O100O10001N100O10000O101N1O100O100O101N100O100O100O2O0O100O100O101N1O100O100O101N100O100O100O2O0O1O100O1O101N100O1O100O1O2O0O1O100O1O101N100O1O100O1O2O0O1O100O100O2N1O1N2O1O1O1O2N1O1IZ_g0"}, "label": "a piece o pizza from the whole pizza in a plate"}]} {"id": 268197, "image": "COCO_train2014_000000268197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with surfing board standing back of other man\"."}], "task": "refering", "answer_template": "The \"a man with surfing board standing back of other man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 125, 126, 127, 148, 171, 193, 194], "bbox": [0.40142187500000004, 0.3206837606837607, 0.581796875, 0.670911680911681], "iscrowd": 0, "area": 2418.28765, "rle": {"size": [351, 640], "counts": "ndh22m:1O2N00001O001O00001O00001O00001O0O2O000N1L3M3M3M3M3jMVOVJm0j5VORJk0o5WOoIh0R6ZOlIe0T6^OkI`0V6CgI;Z6JbI5_6N_IOc64ZIKh66WIGm69QIER7dHB^7=bHB_7?_HAc7`0ZHAf7a0XH^Oj7c0SH]Oo7d0nG[OU8c0lGWO[8h0dGTOa8k0j000O1O100O100O2O0O10O1000000O1000001O2M2O2NYOTF=j9BXF>g9A[F>e9A^F=b9B`F=`9CbF:`9EaF:_9GaF8_9HcF6]9JdF4]9LeF2[9OeF0[90fFOZ91hFMX93iFLW95jFIV97i001O01O0000001O0000001O001O1O001O001O01N2O001N3M3Megk2"}, "label": "a man with surfing board standing back of other man"}]} {"id": 268197, "image": "COCO_train2014_000000268197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an sea skater skating on the long waves\"."}], "task": "refering", "answer_template": "The \"an sea skater skating on the long waves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 104, 105, 125, 126, 127, 148, 171, 193, 194], "bbox": [0.40142187500000004, 0.3206837606837607, 0.581796875, 0.670911680911681], "iscrowd": 0, "area": 2418.28765, "rle": {"size": [351, 640], "counts": "ndh22m:1O2N00001O001O00001O00001O00001O0O2O000N1L3M3M3M3M3jMVOVJm0j5VORJk0o5WOoIh0R6ZOlIe0T6^OkI`0V6CgI;Z6JbI5_6N_IOc64ZIKh66WIGm69QIER7dHB^7=bHB_7?_HAc7`0ZHAf7a0XH^Oj7c0SH]Oo7d0nG[OU8c0lGWO[8h0dGTOa8k0j000O1O100O100O2O0O10O1000000O1000001O2M2O2NYOTF=j9BXF>g9A[F>e9A^F=b9B`F=`9CbF:`9EaF:_9GaF8_9HcF6]9JdF4]9LeF2[9OeF0[90fFOZ91hFMX93iFLW95jFIV97i001O01O0000001O0000001O001O1O001O001O01N2O001N3M3Megk2"}, "label": "an sea skater skating on the long waves"}]} {"id": 268197, "image": "COCO_train2014_000000268197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a surfer is leaning forward away from the wave , next to another surfer leaning into the wave\"."}], "task": "refering", "answer_template": "The \"a surfer is leaning forward away from the wave , next to another surfer leaning into the wave\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 149, 150, 151, 171, 172, 173, 194, 195, 196, 197, 218], "bbox": [0.45456250000000004, 0.41327635327635326, 0.625015625, 0.7286039886039887], "iscrowd": 0, "area": 3874.5688, "rle": {"size": [351, 640], "counts": "mnS31m:1N3M2O1N2N3M2O1N3M2N10OG8J7001O0OZOTO^Gk0^8]O^Gb0e8@WG`0j8EQG9R9g03M3M2N3N20O1RGLQ74kH6P7IkHk6BXI;h6E[I8e6H_I3b6MaI0_60dIM\\63gIIZ67W2O100000000O100000O10O0100O100O1000000000O1O1O100000002NfSb2"}, "label": "a surfer is leaning forward away from the wave , next to another surfer leaning into the wave"}]} {"id": 268197, "image": "COCO_train2014_000000268197.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the surfer who is displaying better balance\"."}], "task": "refering", "answer_template": "The \"the surfer who is displaying better balance\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 129, 149, 150, 151, 171, 172, 173, 194, 195, 196, 197, 218], "bbox": [0.45456250000000004, 0.41327635327635326, 0.625015625, 0.7286039886039887], "iscrowd": 0, "area": 3874.5688, "rle": {"size": [351, 640], "counts": "mnS31m:1N3M2O1N2N3M2O1N3M2N10OG8J7001O0OZOTO^Gk0^8]O^Gb0e8@WG`0j8EQG9R9g03M3M2N3N20O1RGLQ74kH6P7IkHk6BXI;h6E[I8e6H_I3b6MaI0_60dIM\\63gIIZ67W2O100000000O100000O10O0100O100O1000000000O1O1O100000002NfSb2"}, "label": "the surfer who is displaying better balance"}]} {"id": 563110, "image": "COCO_train2014_000000563110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an adult elephant which is partly not in the picture\"."}], "task": "refering", "answer_template": "The \"an adult elephant which is partly not in the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 65, 66, 68, 69, 70, 71, 73, 74, 84, 85, 88, 89, 99, 100, 103, 104, 115, 118, 119, 130, 133, 134, 145, 146, 148, 149, 163, 164, 178, 179, 193, 194, 208, 209, 222, 223, 224, 237, 238, 239, 254], "bbox": [0.16840749414519907, 0.0, 1.0, 0.710671875], "iscrowd": 0, "area": 54370.876849999986, "rle": {"size": [640, 427], "counts": "SP]12nc02N3M3M3M3M3M3M3M2N3M3M3M3M3M3M3M2N3M3M3M3M3M3M3M3M3M3M3M3M3M3M1O00001O00001O001O00001O001O00001O00001O001O00001O001O00010O001O00001O00001O001O00001O001O00001O001O00001O00001O001O00001O001O00001O001O00001O00001O001O00001O001O00001O00001O001O00001O001O0000ZOf0O1000000000000000000O100000000001O000000000000001O000000000000001O00000000001O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O2N3M2N3M3M2N3M3M2N3M2N3M3M2N3M2N3M3M2N3M3M2N3M2N3M3M2N3M5K:F:F]L\\C3[ in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 65, 66, 68, 69, 70, 71, 73, 74, 84, 85, 88, 89, 99, 100, 103, 104, 115, 118, 119, 130, 133, 134, 145, 146, 148, 149, 163, 164, 178, 179, 193, 194, 208, 209, 222, 223, 224, 237, 238, 239, 254], "bbox": [0.16840749414519907, 0.0, 1.0, 0.710671875], "iscrowd": 0, "area": 54370.876849999986, "rle": {"size": [640, 427], "counts": "SP]12nc02N3M3M3M3M3M3M3M2N3M3M3M3M3M3M3M2N3M3M3M3M3M3M3M3M3M3M3M3M3M3M1O00001O00001O001O00001O001O00001O00001O001O00001O001O00010O001O00001O00001O001O00001O001O00001O001O00001O00001O001O00001O001O00001O001O00001O00001O001O00001O001O00001O00001O001O00001O001O0000ZOf0O1000000000000000000O100000000001O000000000000001O000000000000001O00000000001O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O1O1O2N3M2N3M3M2N3M3M2N3M2N3M3M2N3M2N3M3M2N3M3M2N3M2N3M3M2N3M5K:F:F]L\\C3[ in this image.", "objects": [{"patches": [127, 128, 129, 149, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 220, 221, 222, 243, 244, 245, 246, 267, 268, 269, 290, 291, 292], "bbox": [0.511734375, 0.36230588235294114, 0.70896875, 0.8515529411764705], "iscrowd": 0, "area": 13846.603350000005, "rle": {"size": [425, 640], "counts": "WZX43U=3M3M3M4L3M3M3N2N3M2N2N2N2N2WOjN[EW1d:nNWET1f:QOVEQ1h:TOSEn0j:XOQEj0m:j0N3N2M3N2M2O2N2M3N2M2O2N2O1N2N1O2O1N2N1O2O1N2N1O2O1N22N3L5L3M2NN1O2N2O1N2N1N3J6K5K4K6L4M3M2N3M3M3M3N1O2N2N2N1O2N2N2O0O2N2N2N2O01000000O106I0001N100O11N4L3L5L9G8H9G4K5L4L4L4L5J5L4L4L4L4L4L4M4K4L4L4L4M6E;F:F:E;F:Fkk\\2"}, "label": "a pile of carrots peeled on a cutting board"}]} {"id": 325545, "image": "COCO_train2014_000000325545.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"peeled carrots laying on top of a cutting board\"."}], "task": "refering", "answer_template": "The \"peeled carrots laying on top of a cutting board\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [127, 128, 129, 149, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 220, 221, 222, 243, 244, 245, 246, 267, 268, 269, 290, 291, 292], "bbox": [0.511734375, 0.36230588235294114, 0.70896875, 0.8515529411764705], "iscrowd": 0, "area": 13846.603350000005, "rle": {"size": [425, 640], "counts": "WZX43U=3M3M3M4L3M3M3N2N3M2N2N2N2N2WOjN[EW1d:nNWET1f:QOVEQ1h:TOSEn0j:XOQEj0m:j0N3N2M3N2M2O2N2M3N2M2O2N2O1N2N1O2O1N2N1O2O1N2N1O2O1N22N3L5L3M2NN1O2N2O1N2N1N3J6K5K4K6L4M3M2N3M3M3M3N1O2N2N2N1O2N2N2O0O2N2N2N2O01000000O106I0001N100O11N4L3L5L9G8H9G4K5L4L4L4L5J5L4L4L4L4L4L4M4K4L4L4L4M6E;F:F:E;F:Fkk\\2"}, "label": "peeled carrots laying on top of a cutting board"}]} {"id": 325545, "image": "COCO_train2014_000000325545.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a carrot is to the left of the rest\"."}], "task": "refering", "answer_template": "The \"a carrot is to the left of the rest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 193, 194, 195, 217, 218, 219, 241, 242, 264, 265, 288, 289, 312], "bbox": [0.421875, 0.49411764705882355, 0.5890625, 0.8964705882352941], "iscrowd": 0, "area": 5080.5, "rle": {"size": [425, 640], "counts": "gX`37Q=:F2O1N3M2O1N3N1N2N3N1O1O2N1O1O2O0O2N2N2N1O2N2O1O001O1O001O100O00100O101N100O2O0O01O010O01O01O010O010O0010O010O01O01O010O01O010O010O00101N2O0O2N2O0O2O1N2N101N2O0O2N2O0O2O1N2N100O100O2N100O1O1O1O1O2M2O1O2N3M4K4K6JWi\\3"}, "label": "a carrot is to the left of the rest"}]} {"id": 325545, "image": "COCO_train2014_000000325545.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the first carrot nearer to ginger & garlic\"."}], "task": "refering", "answer_template": "The \"the first carrot nearer to ginger & garlic\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [170, 171, 193, 194, 195, 217, 218, 219, 241, 242, 264, 265, 288, 289, 312], "bbox": [0.421875, 0.49411764705882355, 0.5890625, 0.8964705882352941], "iscrowd": 0, "area": 5080.5, "rle": {"size": [425, 640], "counts": "gX`37Q=:F2O1N3M2O1N3N1N2N3N1O1O2N1O1O2O0O2N2N2N1O2N2O1O001O1O001O100O00100O101N100O2O0O01O010O01O01O010O010O0010O010O01O01O010O01O010O010O00101N2O0O2N2O0O2O1N2N101N2O0O2N2O0O2O1N2N100O100O2N100O1O1O1O1O2M2O1O2N3M4K4K6JWi\\3"}, "label": "the first carrot nearer to ginger & garlic"}]} {"id": 399276, "image": "COCO_train2014_000000399276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue car behind a parking meter\"."}], "task": "refering", "answer_template": "The \"a blue car behind a parking meter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 70, 71, 85, 86, 87, 88, 102, 103, 104, 105, 119, 120, 121, 122, 123, 136, 137, 138, 139, 140, 153, 154, 155, 156, 157, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 192, 204, 205, 206, 207, 208, 209, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 272, 273, 274, 275, 276, 277, 289, 290, 291, 292, 293, 294, 307, 308, 309, 310, 311, 324, 325, 326, 327, 328, 341, 342, 343, 344, 345], "bbox": [0.0, 0.1880625, 0.3453541666666667, 0.9155468749999999], "iscrowd": 0, "area": 60881.302550000015, "rle": {"size": [640, 480], "counts": "X4V;i8101N1O100O1O100O1O100O2N100O100O1O100O1O100O1O10000000000000000000000000000000000000001O000oIjDX3V;hJjFX5o;00000O10000000000000000O10000000000000000O10000000000000000O1000000000000000000O10000000000000000O100000000000000eB_JWC=C=C=C=B>C:F9G8HVfS6"}, "label": "a blue car behind a parking meter"}]} {"id": 399276, "image": "COCO_train2014_000000399276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blues car\"."}], "task": "refering", "answer_template": "The \"the blues car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [68, 69, 70, 71, 85, 86, 87, 88, 102, 103, 104, 105, 119, 120, 121, 122, 123, 136, 137, 138, 139, 140, 153, 154, 155, 156, 157, 170, 171, 172, 173, 174, 187, 188, 189, 190, 191, 192, 204, 205, 206, 207, 208, 209, 221, 222, 223, 224, 225, 226, 238, 239, 240, 241, 242, 243, 255, 256, 257, 258, 259, 260, 272, 273, 274, 275, 276, 277, 289, 290, 291, 292, 293, 294, 307, 308, 309, 310, 311, 324, 325, 326, 327, 328, 341, 342, 343, 344, 345], "bbox": [0.0, 0.1880625, 0.3453541666666667, 0.9155468749999999], "iscrowd": 0, "area": 60881.302550000015, "rle": {"size": [640, 480], "counts": "X4V;i8101N1O100O1O100O1O100O2N100O100O1O100O1O100O1O10000000000000000000000000000000000000001O000oIjDX3V;hJjFX5o;00000O10000000000000000O10000000000000000O10000000000000000O1000000000000000000O10000000000000000O100000000000000eB_JWC=C=C=C=B>C:F9G8HVfS6"}, "label": "the blues car"}]} {"id": 399276, "image": "COCO_train2014_000000399276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"awhite suv to the right of a parking meter\"."}], "task": "refering", "answer_template": "The \"awhite suv to the right of a parking meter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 83, 84, 99, 100, 101, 116, 117, 118, 133, 134, 135, 150, 151, 152, 167, 168, 169, 184, 185, 186, 201, 202, 203, 218, 219, 220, 235, 236, 237, 254], "bbox": [0.8327291666666666, 0.16990624999999998, 1.0, 0.629890625], "iscrowd": 0, "area": 18494.611500000003, "rle": {"size": [640, 480], "counts": "WWj7?Y1EU`0R1a_OOZ?>\\@e0i>]Om@T1j>oNl@^1n>o1J5K4L4L4L3L4M2N2N2N2N3M2N2M3N3M2N2O1N3N1N2O1N3N1O1N2O2M2O1O1N2O1N2O1N3N1O1N2O1O1O100O2O000O100O100O100O2N100O3N1N3N2M2N3N1N3N2M2O2N1O2N2N1O2N1O2NF:FmL"}, "label": "awhite suv to the right of a parking meter"}]} {"id": 399276, "image": "COCO_train2014_000000399276.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white vehicle\"."}], "task": "refering", "answer_template": "The \"a white vehicle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 83, 84, 99, 100, 101, 116, 117, 118, 133, 134, 135, 150, 151, 152, 167, 168, 169, 184, 185, 186, 201, 202, 203, 218, 219, 220, 235, 236, 237, 254], "bbox": [0.8327291666666666, 0.16990624999999998, 1.0, 0.629890625], "iscrowd": 0, "area": 18494.611500000003, "rle": {"size": [640, 480], "counts": "WWj7?Y1EU`0R1a_OOZ?>\\@e0i>]Om@T1j>oNl@^1n>o1J5K4L4L4L3L4M2N2N2N2N3M2N2M3N3M2N2O1N3N1N2O1N3N1O1N2O2M2O1O1N2O1N2O1N3N1O1N2O1O1O100O2O000O100O100O100O2N100O3N1N3N2M2N3N1N3N2M2O2N1O2N2N1O2N1O2NF:FmL"}, "label": "a white vehicle"}]} {"id": 513966, "image": "COCO_train2014_000000513966.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white four door car in a city intersection\"."}], "task": "refering", "answer_template": "The \"a white four door car in a city intersection\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [217, 218, 219, 220, 239, 240, 241, 242, 243, 244, 245, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 329, 330, 332, 333, 334], "bbox": [0.279078125, 0.6549884526558892, 0.7264375000000001, 0.9864434180138568], "iscrowd": 0, "area": 26408.589749999992, "rle": {"size": [433, 640], "counts": "lQ\\24]=000_O4\\CN_<9^CG`<>]CD` in this image.", "objects": [{"patches": [217, 218, 219, 220, 239, 240, 241, 242, 243, 244, 245, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 329, 330, 332, 333, 334], "bbox": [0.279078125, 0.6549884526558892, 0.7264375000000001, 0.9864434180138568], "iscrowd": 0, "area": 26408.589749999992, "rle": {"size": [433, 640], "counts": "lQ\\24]=000_O4\\CN_<9^CG`<>]CD` in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 17, 33, 34, 35, 36, 37, 38, 39, 40, 41, 57, 58, 59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 105, 106, 107, 128, 129, 130], "bbox": [0.44278125, 0.0072105263157894745, 0.8024687499999998, 0.4277368421052632], "iscrowd": 0, "area": 18440.429300000003, "rle": {"size": [380, 640], "counts": "c]Y32h;3N1N2O2M2N3N1O1O2O000O2O0O101N101O0O101N10001N100O2O000O2O0O1oE^Oh8b0VG1X80fG3Y8NdG4\\8LbG6]8KaG7_8I`G8`8H`G9^8I`G8`8H`G8`8H`G8_8IaG7_8IaG8^8I`G8_8IaG7_8IaG7_8IaG7_8IaG7^8JaG8^8IaG7_8IaG7^8JbG6^8JbG6^8JaG8]8JbG6^8JbG6^8JbG6]8KbG6^8JbG7]8JbG3`8N`GMe83[GHj88UGDP9 in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 15, 16, 17, 33, 34, 35, 36, 37, 38, 39, 40, 41, 57, 58, 59, 60, 61, 62, 63, 80, 81, 82, 83, 84, 105, 106, 107, 128, 129, 130], "bbox": [0.44278125, 0.0072105263157894745, 0.8024687499999998, 0.4277368421052632], "iscrowd": 0, "area": 18440.429300000003, "rle": {"size": [380, 640], "counts": "c]Y32h;3N1N2O2M2N3N1O1O2O000O2O0O101N101O0O101N10001N100O2O000O2O0O1oE^Oh8b0VG1X80fG3Y8NdG4\\8LbG6]8KaG7_8I`G8`8H`G9^8I`G8`8H`G8`8H`G8_8IaG7_8IaG8^8I`G8_8IaG7_8IaG7_8IaG7_8IaG7^8JaG8^8IaG7_8IaG7^8JbG6^8JbG6^8JaG8]8JbG6^8JbG6^8JbG6]8KbG6^8JbG7]8JbG3`8N`GMe83[GHj88UGDP9 in this image.", "objects": [{"patches": [9, 32, 54, 55, 56, 77, 78, 79, 100, 101, 102, 124, 125, 126, 147, 148, 149, 171, 172, 194, 195, 216, 217, 218, 219, 239, 240, 241, 242], "bbox": [0.353171875, 0.04676315789473684, 0.5366875, 0.7896052631578946], "iscrowd": 0, "area": 12851.26265, "rle": {"size": [380, 640], "counts": "omc22i;2N2O2M2N2N2OO1O010O1O10hEJj85QG3l8NQG6n8IQG:l8GTGc8D\\G`0a8A]Gb0a8^O`Gc0^8^OaGe0\\8[OeGf0X8[OhGh0U8D`G>]81UG0i8d1M4K5K5K5K4L5K4M4K4M4K4M4K4nK\\Kg0g4UO_Ke0f4VO]Kh0d4VO^Kj0c4ROaKl0a4QObKn0_4oNcKQ1]4mNfKT1Y4hNkKi1d3UN]Lm1b3QN`LP2^3oMcLT2[3jMgLW2X3fMkL\\2R3cMoL_2P3_MRMb2l2]MVMd2i2YMYMi2f2UM\\Mm2a2RM`MQ3^2mLdMU3Z2iLgMZ3V2dLlM_3R2_LoMe3o1XLRNn3P2iKQN\\4Q2]KPNd4U2TKlMn4TOdJR25kNY5nNeJ]4^5^KeJa4^5YKfJf4^5oJjJP5R6O1O1O1O100O1O2N1O2N0mMRIXOo6f0SIYOn6f0SIXOo6g0RIXOo6XOnH^OO45U1n6VORI_OL65T1n6TOTI@J84T1o6QOUIAJ94S1o6oNWICG;4R1o6mNYIDE=4Q1n6lN\\ICD`03P1n6jN]IECa03o0n6hN_IFAb04o0m6fNoI;UOm0n6`NTJc0oNl0b8TO_Gk0a8UO`Gj0`8UOaGj0`8VOaGi0b8mNoFF`0[1e8jNeGU1b9O1N2O1N2O1O1N1N5K5K5K5J7J[n]3"}, "label": "a plant in a white vase in the middle of a table"}]} {"id": 333748, "image": "COCO_train2014_000000333748.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a plant in a white vase in the middle of a table\"."}], "task": "refering", "answer_template": "The \"a plant in a white vase in the middle of a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [9, 32, 54, 55, 56, 77, 78, 79, 100, 101, 102, 124, 125, 126, 147, 148, 149, 171, 172, 194, 195, 216, 217, 218, 219, 239, 240, 241, 242], "bbox": [0.353171875, 0.04676315789473684, 0.5366875, 0.7896052631578946], "iscrowd": 0, "area": 12851.26265, "rle": {"size": [380, 640], "counts": "omc22i;2N2O2M2N2N2OO1O010O1O10hEJj85QG3l8NQG6n8IQG:l8GTGc8D\\G`0a8A]Gb0a8^O`Gc0^8^OaGe0\\8[OeGf0X8[OhGh0U8D`G>]81UG0i8d1M4K5K5K5K4L5K4M4K4M4K4M4K4nK\\Kg0g4UO_Ke0f4VO]Kh0d4VO^Kj0c4ROaKl0a4QObKn0_4oNcKQ1]4mNfKT1Y4hNkKi1d3UN]Lm1b3QN`LP2^3oMcLT2[3jMgLW2X3fMkL\\2R3cMoL_2P3_MRMb2l2]MVMd2i2YMYMi2f2UM\\Mm2a2RM`MQ3^2mLdMU3Z2iLgMZ3V2dLlM_3R2_LoMe3o1XLRNn3P2iKQN\\4Q2]KPNd4U2TKlMn4TOdJR25kNY5nNeJ]4^5^KeJa4^5YKfJf4^5oJjJP5R6O1O1O1O100O1O2N1O2N0mMRIXOo6f0SIYOn6f0SIXOo6g0RIXOo6XOnH^OO45U1n6VORI_OL65T1n6TOTI@J84T1o6QOUIAJ94S1o6oNWICG;4R1o6mNYIDE=4Q1n6lN\\ICD`03P1n6jN]IECa03o0n6hN_IFAb04o0m6fNoI;UOm0n6`NTJc0oNl0b8TO_Gk0a8UO`Gj0`8UOaGj0`8VOaGi0b8mNoFF`0[1e8jNeGU1b9O1N2O1N2O1O1N1N5K5K5K5J7J[n]3"}, "label": "a plant in a white vase in the middle of a table"}]} {"id": 6068, "image": "COCO_train2014_000000006068.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lone train car , soon to be bypassed by a locomotive in motion on the opposite track\"."}], "task": "refering", "answer_template": "The \"a lone train car , soon to be bypassed by a locomotive in motion on the opposite track\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191], "bbox": [0.177453125, 0.3645983379501385, 0.59828125, 0.6567313019390582], "iscrowd": 0, "area": 17470.3306, "rle": {"size": [361, 640], "counts": "bZX1l0]:n0RO?@1000001N1000000O2O000O10001O0O10000O101O003L3N2N2N2M3N1O0O1O1O1O10000000000001O000001O0001O00000000001O0000001O00000010O000001O0000001O000O101O00000O101O000O101O00000O2WMcGd2a801O0O100000001N1000000O101O000O10001M2N2N2O10010O00000000001O00000000001O000000001O000000000O2O00000000001O0O1000001O0000000O101O00000000001O0O1000001O0000000O101O00000000001N10000000001O000000010O00000000010O00000000010O000000010O00000000001O00000000001O000000001N10000000001N1000002N3M3M3L4M4L4L5K4L2M100000000O1000000O1000000O1000000O1000000O1000000O1000004Ki\\j2"}, "label": "a lone train car , soon to be bypassed by a locomotive in motion on the opposite track"}]} {"id": 6068, "image": "COCO_train2014_000000006068.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray and undetached train cart on the tracks\"."}], "task": "refering", "answer_template": "The \"a gray and undetached train cart on the tracks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 119, 120, 121, 122, 123, 124, 125, 142, 143, 144, 145, 146, 147, 148, 149, 150, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 189, 190, 191], "bbox": [0.177453125, 0.3645983379501385, 0.59828125, 0.6567313019390582], "iscrowd": 0, "area": 17470.3306, "rle": {"size": [361, 640], "counts": "bZX1l0]:n0RO?@1000001N1000000O2O000O10001O0O10000O101O003L3N2N2N2M3N1O0O1O1O1O10000000000001O000001O0001O00000000001O0000001O00000010O000001O0000001O000O101O00000O101O000O101O00000O2WMcGd2a801O0O100000001N1000000O101O000O10001M2N2N2O10010O00000000001O00000000001O000000001O000000000O2O00000000001O0O1000001O0000000O101O00000000001O0O1000001O0000000O101O00000000001N10000000001O000000010O00000000010O00000000010O000000010O00000000001O00000000001O000000001N10000000001N1000002N3M3M3L4M4L4L5K4L2M100000000O1000000O1000000O1000000O1000000O1000000O1000004Ki\\j2"}, "label": "a gray and undetached train cart on the tracks"}]} {"id": 55226, "image": "COCO_train2014_000000055226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the beige and brown duel recliner\"."}], "task": "refering", "answer_template": "The \"the beige and brown duel recliner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 133, 134, 135, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274], "bbox": [0.55803125, 0.3058125, 0.9978281250000001, 0.7133333333333333], "iscrowd": 0, "area": 37021.372350000005, "rle": {"size": [480, 640], "counts": "oaW52l>6J6J6K3N0aBZOc in this image.", "objects": [{"patches": [130, 131, 132, 133, 134, 135, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 244, 245, 246, 247, 248, 249, 250, 251, 269, 270, 271, 272, 273, 274], "bbox": [0.55803125, 0.3058125, 0.9978281250000001, 0.7133333333333333], "iscrowd": 0, "area": 37021.372350000005, "rle": {"size": [480, 640], "counts": "oaW52l>6J6J6K3N0aBZOc in this image.", "objects": [{"patches": [123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 210, 211, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236], "bbox": [0.11920312500000001, 0.33270833333333333, 0.572859375, 0.6408333333333334], "iscrowd": 0, "area": 28086.792250000002, "rle": {"size": [480, 640], "counts": "UkS13f>8J6J6K5K5K:H`0A2N2N2O1N2O001N2O1M3M3M3O0O2N2N2O1O1N2O001O1O1O1O1O1O1O1N2O1O001O1O1O1O1O=C1N3N2N00000000000O10O1N2O1N2N2O1J60O10000000O100000O10O100000000O1000O1000O10000000O10O100000000O010000000O1000O100000O10000000O0100000000O10O100000O100000O1000O10000000O0100000000O1000O1000O10000000000O010000000O10000000O010000000000O1000O1000O10000000000O010000000O10000000O10O100000000O1000O100000O100000000O01000000000O100000O10O100000000O1000O100000O100000000O100000000O101O001O2M4M00001O000O2O001O000O2O001O000O2O001O001N10001O0O2O00001N101O00001N101O000O2OVCoMe in this image.", "objects": [{"patches": [123, 124, 125, 126, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 210, 211, 212, 213, 214, 215, 216, 217, 218, 234, 235, 236], "bbox": [0.11920312500000001, 0.33270833333333333, 0.572859375, 0.6408333333333334], "iscrowd": 0, "area": 28086.792250000002, "rle": {"size": [480, 640], "counts": "UkS13f>8J6J6K5K5K:H`0A2N2N2O1N2O001N2O1M3M3M3O0O2N2N2O1O1N2O001O1O1O1O1O1O1O1N2O1O001O1O1O1O1O=C1N3N2N00000000000O10O1N2O1N2N2O1J60O10000000O100000O10O100000000O1000O1000O10000000O10O100000000O010000000O1000O100000O10000000O0100000000O10O100000O100000O1000O10000000O0100000000O1000O1000O10000000000O010000000O10000000O010000000000O1000O1000O10000000000O010000000O10000000O10O100000000O1000O100000O100000000O01000000000O100000O10O100000000O1000O100000O100000000O100000000O101O001O2M4M00001O000O2O001O000O2O001O000O2O001O001N10001O0O2O00001N101O00001N101O000O2OVCoMe in this image.", "objects": [{"patches": [238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 288, 304, 305, 306, 307, 308, 309, 310, 311, 327, 328, 329, 330, 331, 332, 333, 334, 351, 352, 353, 354, 356, 357, 375], "bbox": [0.21926562500000002, 0.5875416666666666, 0.540828125, 1.0], "iscrowd": 0, "area": 24460.438099999996, "rle": {"size": [480, 640], "counts": "jmQ22m>e0[Oe0[Oe0[O8I1N2N3M2O10000O101O0O1000000O101O0O1000000O2O000O1000001N10000O1000001N10000O100N3N1O1N2O1N2O2N1N2O1O1N2O1O4K6K5J6I7J6I8H2N2O2N1O10XOlFdKT9V4n0K50O0100000O010000O10O1000O10O1000O1000O0100000O010000O10O1000O1000O10O1000O01000000O01000O10O1000O10BTLhEl3U:XLiEi3W:WLiEh3W:ZLhEf3X:[LgEe3Y:[LfEe3Z:]LeEc3[:]LeEc3[:^LdEb3[:_LeE`3\\:b01O0O2O001O1N101O001N2O001N101O1O0O2O001O0O1000O01000000O01000O1000O10O100000O010000O0100000O10O10O101O1O1N30O5K5L4K08H6I0010O0100O0O2N1O2mNR1O2TOk0ROn`Y4"}, "label": "a white chair in a living room near a grey couch and two recliners"}]} {"id": 55226, "image": "COCO_train2014_000000055226.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a single - seat white sofa\"."}], "task": "refering", "answer_template": "The \"a single - seat white sofa\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 288, 304, 305, 306, 307, 308, 309, 310, 311, 327, 328, 329, 330, 331, 332, 333, 334, 351, 352, 353, 354, 356, 357, 375], "bbox": [0.21926562500000002, 0.5875416666666666, 0.540828125, 1.0], "iscrowd": 0, "area": 24460.438099999996, "rle": {"size": [480, 640], "counts": "jmQ22m>e0[Oe0[Oe0[O8I1N2N3M2O10000O101O0O1000000O101O0O1000000O2O000O1000001N10000O1000001N10000O100N3N1O1N2O1N2O2N1N2O1O1N2O1O4K6K5J6I7J6I8H2N2O2N1O10XOlFdKT9V4n0K50O0100000O010000O10O1000O10O1000O1000O0100000O010000O10O1000O1000O10O1000O01000000O01000O10O1000O10BTLhEl3U:XLiEi3W:WLiEh3W:ZLhEf3X:[LgEe3Y:[LfEe3Z:]LeEc3[:]LeEc3[:^LdEb3[:_LeE`3\\:b01O0O2O001O1N101O001N2O001N101O1O0O2O001O0O1000O01000000O01000O1000O10O100000O010000O0100000O10O10O101O1O1N30O5K5L4K08H6I0010O0100O0O2N1O2mNR1O2TOk0ROn`Y4"}, "label": "a single - seat white sofa"}]} {"id": 120782, "image": "COCO_train2014_000000120782.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue motorcycle next to a woman in jeans\"."}], "task": "refering", "answer_template": "The \"a blue motorcycle next to a woman in jeans\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 111, 127, 128, 129, 145, 146, 147, 163, 164, 165, 182], "bbox": [0.044520000000000004, 0.46781333333333336, 0.20176000000000002, 0.8067733333333333], "iscrowd": 0, "area": 6095.239750000001, "rle": {"size": [375, 500], "counts": "\\c84b;2N2N2N3N1dEC\\9>]FK`97YF2b91[F3b90ZF5b9M[F7b9LYF9c9P1N2N1O2N1N3N2N1O2N1N3N1O2N0O10001O000O1000000O10001O000O11O0`0@4L3M1O10O01O1O0010O01O00100O0001M2M3N3iLmGf2U8ZMnGa2T8_MoG\\2T8dMnGR2Y8nMiGe1b8[NaGX1j8gN[GQ1h8oN^Gh0e8XOaG`0b8_OlG0W80h101O000L5LRna4"}, "label": "a blue motorcycle next to a woman in jeans"}]} {"id": 120782, "image": "COCO_train2014_000000120782.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a byek on foot path\"."}], "task": "refering", "answer_template": "The \"a byek on foot path\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [108, 109, 110, 111, 127, 128, 129, 145, 146, 147, 163, 164, 165, 182], "bbox": [0.044520000000000004, 0.46781333333333336, 0.20176000000000002, 0.8067733333333333], "iscrowd": 0, "area": 6095.239750000001, "rle": {"size": [375, 500], "counts": "\\c84b;2N2N2N3N1dEC\\9>]FK`97YF2b91[F3b90ZF5b9M[F7b9LYF9c9P1N2N1O2N1N3N2N1O2N1N3N1O2N0O10001O000O1000000O10001O000O11O0`0@4L3M1O10O01O1O0010O01O00100O0001M2M3N3iLmGf2U8ZMnGa2T8_MoG\\2T8dMnGR2Y8nMiGe1b8[NaGX1j8gN[GQ1h8oN^Gh0e8XOaG`0b8_OlG0W80h101O000L5LRna4"}, "label": "a byek on foot path"}]} {"id": 120782, "image": "COCO_train2014_000000120782.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"truck\"."}], "task": "refering", "answer_template": "The \"truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 42, 43, 55, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 79, 91, 92, 93, 94, 96], "bbox": [0.06366, 0.20872, 0.41884000000000005, 0.4352266666666666], "iscrowd": 0, "area": 8929.98125, "rle": {"size": [375, 500], "counts": "Uk; in this image.", "objects": [{"patches": [39, 40, 41, 42, 43, 55, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 79, 91, 92, 93, 94, 96], "bbox": [0.06366, 0.20872, 0.41884000000000005, 0.4352266666666666], "iscrowd": 0, "area": 8929.98125, "rle": {"size": [375, 500], "counts": "Uk; in this image.", "objects": [{"patches": [162, 163, 164, 165, 184, 185, 186, 187, 188, 208, 209, 210, 211, 231, 232, 233, 234, 235, 236, 254, 255, 256, 257, 258, 259, 278, 279, 280], "bbox": [0.032015625, 0.43370833333333336, 0.306734375, 0.7393333333333333], "iscrowd": 0, "area": 13975.745700000001, "rle": {"size": [480, 640], "counts": "ob91P?4L4L5K4L4K6J5K5K5K6J5K5K1O001O1O001N1O2O>A`0A0O2N2O0O2N2O0O2N2O0O2O1N101N2O1O001O1N101O1O001O1O0O2O1O001O1O001O1N101O1O001O1O001N2O001O1O001O1N2OO100O100O10O10O10000O10000O01000O10000O10O10O2O4L4K5L4L4K5L4L4K5L4L4K5L4L4K5L4L4K5L2N1N2O1O0O2O001N2O001N101O0O2O1O0O2O001N101O1N101O0O2O1O0O2O001N101O1N101N101O0O2O1O0O2O001N2O001N101O0O2O1O0O2O001N101ORY`6"}, "label": "the bigger laptop on the left"}]} {"id": 456658, "image": "COCO_train2014_000000456658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a computer keyboard on a table\"."}], "task": "refering", "answer_template": "The \"a computer keyboard on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 243, 244, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 293, 294, 295], "bbox": [0.573953125, 0.56275, 0.869390625, 0.7337291666666667], "iscrowd": 0, "area": 7389.387049999992, "rle": {"size": [480, 640], "counts": "\\i\\51o>1N1O100O2N1O100O2N100O2N100O2N1O2O0O1O2O0O1O2O0O1O2N100O2N100N3N1N2O2M2O110O00001O01O01O000010O0001O010O000010O0001O01O01O00010O000010O0001O01O01O01O01O00010O000010O0001O01O01O00010O000010O0001O01O01O00010O00010O000010O0001O01O01O00010O000010O0001O01O01O0001O01O00010O00010O000010O0001O01O010O00010O00010O000O2N1O1O2O0O1O2N1O101N1O1O2N1O2O0O1O2N1O101N1O1O2N1O101N1M3M4M2M3M[RW1"}, "label": "a computer keyboard on a table"}]} {"id": 456658, "image": "COCO_train2014_000000456658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an white color external keyboard front of the monitor\"."}], "task": "refering", "answer_template": "The \"an white color external keyboard front of the monitor\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 243, 244, 245, 246, 247, 248, 249, 267, 268, 269, 270, 271, 272, 293, 294, 295], "bbox": [0.573953125, 0.56275, 0.869390625, 0.7337291666666667], "iscrowd": 0, "area": 7389.387049999992, "rle": {"size": [480, 640], "counts": "\\i\\51o>1N1O100O2N1O100O2N100O2N100O2N1O2O0O1O2O0O1O2O0O1O2N100O2N100N3N1N2O2M2O110O00001O01O01O000010O0001O010O000010O0001O01O01O00010O000010O0001O01O01O01O01O00010O000010O0001O01O01O00010O000010O0001O01O01O00010O00010O000010O0001O01O01O00010O000010O0001O01O01O0001O01O00010O00010O000010O0001O01O010O00010O00010O000O2N1O1O2O0O1O2N1O101N1O1O2N1O2O0O1O2N1O101N1O1O2N1O101N1M3M4M2M3M[RW1"}, "label": "an white color external keyboard front of the monitor"}]} {"id": 456658, "image": "COCO_train2014_000000456658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a computer screen turned on with a program on it\"."}], "task": "refering", "answer_template": "The \"a computer screen turned on with a program on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224], "bbox": [0.5813906249999999, 0.3140208333333333, 0.7842812499999999, 0.57525], "iscrowd": 0, "area": 13755.259200000004, "rle": {"size": [480, 640], "counts": "Zc^5?e=m0UOk0WOj0J5N2O1O2N1O000000010O00000000001O0000010O00001O0000001O0000010O00001O0000001O0001O01O000000001O00000001O000001O000000000001O0001O000000000000010O000000000000001O01O000000000010O00000001O000000010O000000001O0001O000001O00000O101O0O100O101O0O1hMX2K5L4K6K8Hi`P2"}, "label": "a computer screen turned on with a program on it"}]} {"id": 456658, "image": "COCO_train2014_000000456658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black color monitor near by cpu\"."}], "task": "refering", "answer_template": "The \"a black color monitor near by cpu\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [128, 129, 130, 131, 132, 151, 152, 153, 154, 155, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224], "bbox": [0.5813906249999999, 0.3140208333333333, 0.7842812499999999, 0.57525], "iscrowd": 0, "area": 13755.259200000004, "rle": {"size": [480, 640], "counts": "Zc^5?e=m0UOk0WOj0J5N2O1O2N1O000000010O00000000001O0000010O00001O0000001O0000010O00001O0000001O0001O01O000000001O00000001O000001O000000000001O0001O000000000000010O000000000000001O01O000000000010O00000001O000000010O000000001O0001O000001O00000O101O0O100O101O0O1hMX2K5L4K6K8Hi`P2"}, "label": "a black color monitor near by cpu"}]} {"id": 456658, "image": "COCO_train2014_000000456658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a screen shows the truck\"."}], "task": "refering", "answer_template": "The \"a screen shows the truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 249, 250, 251], "bbox": [0.7711718750000001, 0.3357291666666667, 0.9706250000000001, 0.63475], "iscrowd": 0, "area": 13664.315550000003, "rle": {"size": [480, 640], "counts": "^iW7f0o=g0YO;A?\\Oe0]Ob0O1O2O01O01O000010O0001O01O01O0000010O000010O0001O00010O000010O0001O0001O01O00010O00001O01O01O00010O00001O01O01O0000010O000010O0001O00010O000010O0001O0001O01O00010O00001O01O01O00010O00001O01O01O0001O0M3K5L4K6L3O1O1O2N1O1O1O1H9G8G9H8H9F9H8G9H8G^g8"}, "label": "a screen shows the truck"}]} {"id": 456658, "image": "COCO_train2014_000000456658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a computer monitor sitting on the end of a desk\"."}], "task": "refering", "answer_template": "The \"a computer monitor sitting on the end of a desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 249, 250, 251], "bbox": [0.7711718750000001, 0.3357291666666667, 0.9706250000000001, 0.63475], "iscrowd": 0, "area": 13664.315550000003, "rle": {"size": [480, 640], "counts": "^iW7f0o=g0YO;A?\\Oe0]Ob0O1O2O01O01O000010O0001O01O01O0000010O000010O0001O00010O000010O0001O0001O01O00010O00001O01O01O00010O00001O01O01O0000010O000010O0001O00010O000010O0001O0001O01O00010O00001O01O01O00010O00001O01O01O0001O0M3K5L4K6L3O1O1O2N1O1O1O1H9G8G9H8H9F9H8G9H8G^g8"}, "label": "a computer monitor sitting on the end of a desk"}]} {"id": 423890, "image": "COCO_train2014_000000423890.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown wooden table behind the man\"."}], "task": "refering", "answer_template": "The \"brown wooden table behind the man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [28, 29, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 84, 85, 86, 87, 88, 89, 90, 91, 105, 126, 127, 147, 148, 168, 169], "bbox": [0.0, 0.10562222222222223, 0.46853333333333336, 0.5213555555555556], "iscrowd": 0, "area": 17812.613250000006, "rle": {"size": [450, 600], "counts": "P`01Q>a0_OQ1oNQ1oNQ1oNa0_O1O1O1O10gMXF8f9H]F6b9JaF4^9LeF2Z9NiF2T9NnF7k8IXG;a8EbG`0V8@mGa0o7_OTH>l7BWH;i7EZH7h7HZH6f7J]H2d7N_HOa71bHK_75dHH\\78gHDZ7 in this image.", "objects": [{"patches": [210, 211, 212, 213, 214, 215, 216, 217, 218, 222, 223, 224, 231, 232, 233, 234, 235, 236, 237, 238, 239, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.0118, 0.6257777777777779, 0.9977500000000001, 0.9988222222222223], "iscrowd": 0, "area": 77329.88730000002, "rle": {"size": [450, 600], "counts": "Y\\33k=7J6I7I6K6J6J6K4K6J6L4M3M2N3M3M3N1N3M3M3M2N3N2M3M3M2N3M3M3N1N3M3M4L3M3N2M3M4L3M3M3M3N3L3M3M000001O0000000000001O00000000001O00000000001O0000000000001O00000000001O0000000000001O00000000001O000O1000001O000000O1000000000000O100000000000000O1000000000000O1000000000000O1000000000000O1000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O2N2N3M3M3M2N3M3M2N3M3M2N1O001O00000000001O0000000O101O00000000001O00000000001O0000000000O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O100O100O10000O100O10000O10000O10O010000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O100000000000000000000000000000000001O000000000O10000000000000000000000000000000001O000000000000000000000000000000000000000000001O000000000000000000000000000O10000000000000001O000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O000O10000000000000000000000000000000000000001O000000000000000000000000000000000000000000001O000000000000000000000O10000000000000000000001O000000000000000000007I4L3M3M4L3M3M3M4L3M3M4L3M3M3M4L3M3M4L3M3M3M4L3M3M4L3M3H8H9H7HS1"}, "label": "the pizza with mozerella cheese"}]} {"id": 423890, "image": "COCO_train2014_000000423890.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wooden table with two pizzas on top of it\"."}], "task": "refering", "answer_template": "The \"a wooden table with two pizzas on top of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [210, 211, 212, 213, 214, 215, 216, 217, 218, 222, 223, 224, 231, 232, 233, 234, 235, 236, 237, 238, 239, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335], "bbox": [0.0118, 0.6257777777777779, 0.9977500000000001, 0.9988222222222223], "iscrowd": 0, "area": 77329.88730000002, "rle": {"size": [450, 600], "counts": "Y\\33k=7J6I7I6K6J6J6K4K6J6L4M3M2N3M3M3N1N3M3M3M2N3N2M3M3M2N3M3M3N1N3M3M4L3M3N2M3M4L3M3M3M3N3L3M3M000001O0000000000001O00000000001O00000000001O0000000000001O00000000001O0000000000001O00000000001O000O1000001O000000O1000000000000O100000000000000O1000000000000O1000000000000O1000000000000O1000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O001O1O1O1O2N2N3M3M3M2N3M3M2N3M3M2N1O001O00000000001O0000000O101O00000000001O00000000001O0000000000O1O1N2O1O1O1O1O1O1N2O1O1O1O1O1N2O1O1O1O100O100O10000O100O10000O10000O10O010000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O100000000000000000000000000000000001O000000000O10000000000000000000000000000000001O000000000000000000000000000000000000000000001O000000000000000000000000000O10000000000000001O000000000000000000000000000000000000000000001O0000000000000000000000000000000000000000001O000O10000000000000000000000000000000000000001O000000000000000000000000000000000000000000001O000000000000000000000O10000000000000000000001O000000000000000000007I4L3M3M4L3M3M3M4L3M3M4L3M3M3M4L3M3M4L3M3M3M4L3M3M4L3M3H8H9H7HS1"}, "label": "a wooden table with two pizzas on top of it"}]} {"id": 423890, "image": "COCO_train2014_000000423890.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pizza with mozzarella slices\"."}], "task": "refering", "answer_template": "The \"pizza with mozzarella slices\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 275, 276, 277, 278, 279, 280, 281, 282, 297, 298, 299, 300, 301, 302], "bbox": [0.11080000000000001, 0.6515111111111112, 0.4473, 0.8983555555555556], "iscrowd": 0, "area": 17634.733299999993, "rle": {"size": [450, 600], "counts": "n[m03l=4M4K4M3L4M4L3N2M3M4M2M2N2O1N3M2O1N2N2O1N2N2O1N2O1N3M2O1N2N2O1N2N2O1N2O1N3N1N2N2O1N200O1000001O00000000000O100000001O0000000O1O100O100O1O2O0O100O1O100O1000000010O00000000000000000001O000001O000000000001O0000000000000001O000000000000000000000000001O0000000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O0000000O100hMTEY1m:fNTEY1m:eNVEY1l:eNUE[1k:cNXE[1i:`N\\E_1e:XNeEf1\\;01N1O2N1N3M2N2N3M2N3M2N3M2N3M2N3M2N3K4GYba4"}, "label": "pizza with mozzarella slices"}]} {"id": 423890, "image": "COCO_train2014_000000423890.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza with the white cheese on it\"."}], "task": "refering", "answer_template": "The \"the pizza with the white cheese on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 215, 216, 217, 218, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 275, 276, 277, 278, 279, 280, 281, 282, 297, 298, 299, 300, 301, 302], "bbox": [0.11080000000000001, 0.6515111111111112, 0.4473, 0.8983555555555556], "iscrowd": 0, "area": 17634.733299999993, "rle": {"size": [450, 600], "counts": "n[m03l=4M4K4M3L4M4L3N2M3M4M2M2N2O1N3M2O1N2N2O1N2N2O1N2O1N3M2O1N2N2O1N2N2O1N2O1N3N1N2N2O1N200O1000001O00000000000O100000001O0000000O1O100O100O1O2O0O100O1O100O1000000010O00000000000000000001O000001O000000000001O0000000000000001O000000000000000000000000001O0000000000000000001O00000000000000001O000000000000001O00000000000000001O00000000000000001O0000000O100hMTEY1m:fNTEY1m:eNVEY1l:eNUE[1k:cNXE[1i:`N\\E_1e:XNeEf1\\;01N1O2N1N3M2N2N3M2N3M2N3M2N3M2N3M2N3K4GYba4"}, "label": "the pizza with the white cheese on it"}]} {"id": 423890, "image": "COCO_train2014_000000423890.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza on the left without mozzarella cheese\"."}], "task": "refering", "answer_template": "The \"the pizza on the left without mozzarella cheese\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [242, 243, 244, 245, 246, 247, 248, 263, 264, 265, 266, 267, 268, 269, 270, 284, 285, 286, 287, 288, 289, 290, 291, 305, 306, 307, 308, 309, 310, 311, 312, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.5275333333333333, 0.6988666666666666, 0.8763999999999998, 0.9618], "iscrowd": 0, "area": 21240.054200000002, "rle": {"size": [450, 600], "counts": "_e[4:e=;D;F;E5J5[Oe0L4N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N101N101O00001O00001O0000001O0O101O0000001O00001O0000000O1000000000000000001O00000000000000000000000000000000000000000000000000000000O100000000000000000001O00000000001O000000010O000000000010O00000001O0001O000001O0001O0001O000000000010O01O00001O00001O0010O0001O0000000000000000001O000000000000000000001O000000000000000000001O1O1O1O1O1O1N3N1N2O1N2O1N4M3L4M3L4M4K4M3J6fNkC>[ in this image.", "objects": [{"patches": [242, 243, 244, 245, 246, 247, 248, 263, 264, 265, 266, 267, 268, 269, 270, 284, 285, 286, 287, 288, 289, 290, 291, 305, 306, 307, 308, 309, 310, 311, 312, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.5275333333333333, 0.6988666666666666, 0.8763999999999998, 0.9618], "iscrowd": 0, "area": 21240.054200000002, "rle": {"size": [450, 600], "counts": "_e[4:e=;D;F;E5J5[Oe0L4N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N101N101O00001O00001O0000001O0O101O0000001O00001O0000000O1000000000000000001O00000000000000000000000000000000000000000000000000000000O100000000000000000001O00000000001O000000010O000000000010O00000001O0001O000001O0001O0001O000000000010O01O00001O00001O0010O0001O0000000000000000001O000000000000000000001O000000000000000000001O1O1O1O1O1O1N3N1N2O1N2O1N4M3L4M3L4M4K4M3J6fNkC>[ in this image.", "objects": [{"patches": [10, 11, 26, 27, 28, 29, 44, 45, 46, 47, 62, 63, 64, 80, 81, 97, 98, 99], "bbox": [0.5303541666666667, 0.028359375, 0.8291666666666667, 0.264734375], "iscrowd": 0, "area": 7142.785299999997, "rle": {"size": [640, 480], "counts": "[]o4:fc01O01O0000001O000O01O1O100O1O1O1O1O1O1O100O1O1O1O1O100O100O100O10000O2O0O100O100O100O100O101N2O1N2O1N10100O1O1O100O1O1O100O1O100O1O2N100O1O2N100O2N1O100O2N101N3M4L3N3L3M4L3N3L3M2O2M2N3MZOY_OUNf`0i1]_OXN``0g1b_OZN\\`0e1f_O]NW`0c1k_O]NT`0a1n_O`NP`0_1S@`Nl?`1U@aNj?]1X@cNi?Z1Z@eNg?Y1Z@gNh?X1X@gNi?\\1S@eNn?]1n_OcNS`0_1n_O]NT`0d1U@PNl?P2T1N10001O1O2M3N2N2N2N2M10000000000O100000000N2O1N3N5J5L4L5J5L4K6H7HSSc1"}, "label": "a light brown horse directly behind the dark brown horse and girl"}]} {"id": 137173, "image": "COCO_train2014_000000137173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a horse that is behind the other horse\"."}], "task": "refering", "answer_template": "The \"a horse that is behind the other horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 26, 27, 28, 29, 44, 45, 46, 47, 62, 63, 64, 80, 81, 97, 98, 99], "bbox": [0.5303541666666667, 0.028359375, 0.8291666666666667, 0.264734375], "iscrowd": 0, "area": 7142.785299999997, "rle": {"size": [640, 480], "counts": "[]o4:fc01O01O0000001O000O01O1O100O1O1O1O1O1O1O100O1O1O1O1O100O100O100O10000O2O0O100O100O100O100O101N2O1N2O1N10100O1O1O100O1O1O100O1O100O1O2N100O1O2N100O2N1O100O2N101N3M4L3N3L3M4L3N3L3M2O2M2N3MZOY_OUNf`0i1]_OXN``0g1b_OZN\\`0e1f_O]NW`0c1k_O]NT`0a1n_O`NP`0_1S@`Nl?`1U@aNj?]1X@cNi?Z1Z@eNg?Y1Z@gNh?X1X@gNi?\\1S@eNn?]1n_OcNS`0_1n_O]NT`0d1U@PNl?P2T1N10001O1O2M3N2N2N2N2M10000000000O100000000N2O1N3N5J5L4L5J5L4K6H7HSSc1"}, "label": "a horse that is behind the other horse"}]} {"id": 137173, "image": "COCO_train2014_000000137173.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown horse with a saddle\"."}], "task": "refering", "answer_template": "The \"a brown horse with a saddle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [41, 42, 56, 57, 58, 59, 73, 74, 75, 76, 89, 90, 91, 92, 93, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 126, 127, 128, 137, 138, 139, 140, 141, 142, 143, 144, 145, 153, 154, 155, 156, 157, 158, 159, 160, 161, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 239, 240, 241, 242, 243, 244, 246, 247, 248, 257, 258, 259, 260, 261, 264, 265, 266, 275, 276, 277, 278, 279, 281, 282, 283, 292, 293, 294, 295, 296, 299, 300, 310, 311, 312, 313, 316, 317, 318, 327, 328, 329, 330, 334, 335], "bbox": [0.005729166666666666, 0.1, 0.7610833333333333, 0.8682343750000001], "iscrowd": 0, "area": 80502.8615, "rle": {"size": [640, 480], "counts": "gS2Y1jb0NO10000O10001N10000O10000O103M6I6K6J5K6I6K6J5J7J5K2M2M2N3N1N3M2N3N1N3M2N3N1N3M2N101N1O2N2O0O2ROjKoAX4P>l010O1O001O1O001O1O001O1O001O1O001O010N2O0O2N2N3M3M2N3M3N1N3M3M3M2N3M3M3M2O2M3M3M2N3M3M3M2O2M3M3M2N3M3M3M2O2M3M3M2N3M3M3M3N2M3M3M4L3M3M3M3M3N3L3M3M3M3M3M4L3N2M3M3M3M4L3M3POaDZIb;c6mDoHU;n6PEmHS;P7mDPIV;m6kDRIY;j6gDVI\\;i5PDkJe0[O^;g5PDjJc0_O`;d5QDiJ`0B`;e5SDeJ=G`;c5WDbJ:J`;d5XD_J8Ma;c5[D\\J50a;c5^DYJ14b;c5YE\\Ja01k7b5dG]Ja09e7Y5jG]Ja0QDb10kMTbNhNn\\Dk0RNfNh=9YDP1h in this image.", "objects": [{"patches": [41, 42, 56, 57, 58, 59, 73, 74, 75, 76, 89, 90, 91, 92, 93, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 125, 126, 127, 128, 137, 138, 139, 140, 141, 142, 143, 144, 145, 153, 154, 155, 156, 157, 158, 159, 160, 161, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 239, 240, 241, 242, 243, 244, 246, 247, 248, 257, 258, 259, 260, 261, 264, 265, 266, 275, 276, 277, 278, 279, 281, 282, 283, 292, 293, 294, 295, 296, 299, 300, 310, 311, 312, 313, 316, 317, 318, 327, 328, 329, 330, 334, 335], "bbox": [0.005729166666666666, 0.1, 0.7610833333333333, 0.8682343750000001], "iscrowd": 0, "area": 80502.8615, "rle": {"size": [640, 480], "counts": "gS2Y1jb0NO10000O10001N10000O10000O103M6I6K6J5K6I6K6J5J7J5K2M2M2N3N1N3M2N3N1N3M2N3N1N3M2N101N1O2N2O0O2ROjKoAX4P>l010O1O001O1O001O1O001O1O001O1O001O010N2O0O2N2N3M3M2N3M3N1N3M3M3M2N3M3M3M2O2M3M3M2N3M3M3M2O2M3M3M2N3M3M3M2O2M3M3M2N3M3M3M3N2M3M3M4L3M3M3M3M3N3L3M3M3M3M3M4L3N2M3M3M3M4L3M3POaDZIb;c6mDoHU;n6PEmHS;P7mDPIV;m6kDRIY;j6gDVI\\;i5PDkJe0[O^;g5PDjJc0_O`;d5QDiJ`0B`;e5SDeJ=G`;c5WDbJ:J`;d5XD_J8Ma;c5[D\\J50a;c5^DYJ14b;c5YE\\Ja01k7b5dG]Ja09e7Y5jG]Ja0QDb10kMTbNhNn\\Dk0RNfNh=9YDP1h in this image.", "objects": [{"patches": [63, 80, 81, 82, 95, 96, 97, 98, 99, 100, 101, 113, 114, 115, 116, 131, 132, 133, 149, 150, 151, 167, 168, 185, 186], "bbox": [0.28692, 0.32734328358208953, 0.6459600000000001, 0.8789253731343284], "iscrowd": 0, "area": 10242.747000000001, "rle": {"size": [335, 500], "counts": "oQ_11S:;D;F;D;F:F;E:I7H9H7I7I8J41L5K4L4L5K4M4K4L4L5K4L5K4O12O1N1O2O1mIlMh3U2hKYNZ4h1VKfNj4\\1fJROZ5P1UJ^Om5a20O3N1O1N2O1M3L4L4L4M3L4L4L4M3L4L4L4M3L4L4L4M3L4L4M31O1O11O00000000000M4L3M3M3M3M2N1O001_OmFFS9;PGAQ9?RG]On8d0UGXOl8g0>0O2O0O2O1N2O001N2N101N2O1N101N2N101N2K50O01O100O010O10O0100O10O10000O1000000O1000000O10000O1000000O1D in this image.", "objects": [{"patches": [63, 80, 81, 82, 95, 96, 97, 98, 99, 100, 101, 113, 114, 115, 116, 131, 132, 133, 149, 150, 151, 167, 168, 185, 186], "bbox": [0.28692, 0.32734328358208953, 0.6459600000000001, 0.8789253731343284], "iscrowd": 0, "area": 10242.747000000001, "rle": {"size": [335, 500], "counts": "oQ_11S:;D;F;D;F:F;E:I7H9H7I7I8J41L5K4L4L5K4M4K4L4L5K4L5K4O12O1N1O2O1mIlMh3U2hKYNZ4h1VKfNj4\\1fJROZ5P1UJ^Om5a20O3N1O1N2O1M3L4L4L4M3L4L4L4M3L4L4L4M3L4L4L4M3L4L4M31O1O11O00000000000M4L3M3M3M3M2N1O001_OmFFS9;PGAQ9?RG]On8d0UGXOl8g0>0O2O0O2O1N2O001N2N101N2O1N101N2N101N2K50O01O100O010O10O0100O10O10000O1000000O1000000O10000O1000000O1D in this image.", "objects": [{"patches": [99, 100, 116, 117, 118, 119, 133, 134, 135, 136, 137, 153, 154, 171, 172, 173, 191], "bbox": [0.41910000000000003, 0.46985074626865675, 0.6394000000000001, 0.85], "iscrowd": 0, "area": 6813.996349999999, "rle": {"size": [335, 500], "counts": "WmT27W:2M2N2N2N2N2M3N2]OYO\\Gi0d8WOYGm0d8UOZGm0b8WO]Gk0b8VO[Gm0c8TO\\Gm0d8SOZGn0f8SOYGm0h8<001O1O001O1O010O1O010O1O010O1O010O001O1O10O01O100001O000O101O001O001O001N5L4QI[Nl4j1cJbN]5P2]IcNd6d20O1O010O0000001O0010000K4H9H8H8L3M4L4N3N1N2O18H8H8H1O2N1O1O2jHVMd6Z31O[I_L]6a3_IbLa6f31O1lN]I[Nc6^1iI\\NX6\\1UJ^Nk5\\1aJ_N_5Z1mJ`NU5X1TKdNP5X1QKgNS5T1oJkNU5P1nJoNU5l0mJSOW5i0[2K6J5L4K5KWfj1"}, "label": "a baby zebra leaning to its mom seeking comfort"}]} {"id": 276444, "image": "COCO_train2014_000000276444.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child zebra with its parent in jungle\"."}], "task": "refering", "answer_template": "The \"a child zebra with its parent in jungle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 116, 117, 118, 119, 133, 134, 135, 136, 137, 153, 154, 171, 172, 173, 191], "bbox": [0.41910000000000003, 0.46985074626865675, 0.6394000000000001, 0.85], "iscrowd": 0, "area": 6813.996349999999, "rle": {"size": [335, 500], "counts": "WmT27W:2M2N2N2N2N2M3N2]OYO\\Gi0d8WOYGm0d8UOZGm0b8WO]Gk0b8VO[Gm0c8TO\\Gm0d8SOZGn0f8SOYGm0h8<001O1O001O1O010O1O010O1O010O1O010O001O1O10O01O100001O000O101O001O001O001N5L4QI[Nl4j1cJbN]5P2]IcNd6d20O1O010O0000001O0010000K4H9H8H8L3M4L4N3N1N2O18H8H8H1O2N1O1O2jHVMd6Z31O[I_L]6a3_IbLa6f31O1lN]I[Nc6^1iI\\NX6\\1UJ^Nk5\\1aJ_N_5Z1mJ`NU5X1TKdNP5X1QKgNS5T1oJkNU5P1nJoNU5l0mJSOW5i0[2K6J5L4K5KWfj1"}, "label": "a child zebra with its parent in jungle"}]} {"id": 186336, "image": "COCO_train2014_000000186336.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the gray stuffed animal in the red and white stripes\"."}], "task": "refering", "answer_template": "The \"the gray stuffed animal in the red and white stripes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 362, 363, 364, 365, 366, 367, 387, 388, 389, 390], "bbox": [0.5882031249999999, 0.17752083333333332, 1.0, 0.9887708333333333], "iscrowd": 0, "area": 77151.19544999997, "rle": {"size": [480, 640], "counts": "gb`53N2M3N2N2M3N3A>@a0G9ZOe0L5L4L3M4K5J5K6ZOf0M2N3N1N3M3N101O1_LXKhLh4U3[KkLf4Q3]KPMc4m2_KSMa4j2bKVM_4h2cKWM^4h2bKXM^4h2bKXM_4g2aKYM_4g2aKYM`4g2`KYM_4g2aKYM`4f2`KZMa4e2`KZM`4f2`KZMa4e2_K[Ma4e2`KZMa4e2_K[Ma4e2_KVMg4i2ZKTMi4k2WKUMi4k2WKUMj4j2WKUMi4l2VKTMk4k2UKUMl4k2TKTMl4l2TKTMm4k2SKUMm4l2SKSMn4l2RKTMn4m2QKSMP5l2QKSMZ5b2fJ^MY5d2fJ\\MZ5e2fJZMZ5f2fJYMZ5i2eJVM\\5j2dJUM]5l2cJRM_5n2`JQMa5o2_JPMc5P3]JnLe5Q3[JnLg5Q3YJnLi5Q3XJlLj5T3VJjLm5U3SJiLP6V3QJeLT6Z3lIcLW6]3iI`L[6_3fI_L\\6`3dI\\La6c3_IULi6k3XITLi6k3WIULj6j3VIVLk6i3VIULP7f3PIZLU7a3kH_LV7`3kH_LU7a3kH_LU7a3kH_LV7`3kH^LV7b3jH^LV7b3lH\\LU7c3lH\\LT7d3mH[LS7e3nHZLR7f3oHXLS7g3nHXLR7h3oHWLQ7i3PIVLP7j3QITLP7l3RIQLo6o3RIoKo6Q4RInKn6R4SIlKn6T4SIjKn6V4SIhKn6X4SIfKn6Z4SIeKm6Z4UIdKl6\\4m1O001O001O001O001O001O001O001O001O001O001O001N2O1N2O1N2O2M2O1N2O1N2O1N200O1O1O1O100O1O1O1O101N1O2N1O010O00001O01O01O000010O01O00001O000010O0001O00000001O0000001O000000001O000000000000O1000000000000O100000_IkLW1V3gNjLZ1V3eNkL[1V3cNjL^1V3aNkL_1V3_NjLb1V3]NkLc1V3[NkLe1U3ZNkLg1V3WNkLi1U3VNkLk1V3SNkLn1T3QNmLo1T3PNkLQ2U3nMlLR2T3mMmLS2T3lMlLT2T3kMnLT2R3kMoLU2R3iMoLV2R3jMnLV2S3hMnLX2R3gMPMX2P3hMPMX2Q3fMPMZ2P3eMmLjKCa6a3cMlLb2T3\\MnLd2R3[MnLf2S3XMmLi2S3VMmLk2S3TMnLl2S3RMmLo2S3PMmLQ3T3mLlLT3T3kLlLV3T3hLlLZ3U3dLkL]3U3bLkL_3U3`LkLa3V3]LiLe3W3ZLiLg3W3XLiLi3X3_J]LI;i5X3]JfL[O8Y6S3ZJoLiN8n6i2XJ\\Nj5e1TJ[Nm5e1QJ\\NP6e1nI[NS6e1lI[NU6T2[IlMf6PN"}, "label": "the gray stuffed animal in the red and white stripes"}]} {"id": 186336, "image": "COCO_train2014_000000186336.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a light brown teddy bear wearing a red and white striped shirt\"."}], "task": "refering", "answer_template": "The \"a light brown teddy bear wearing a red and white striped shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 175, 176, 177, 178, 179, 180, 181, 182, 183, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 313, 314, 315, 316, 317, 318, 319, 320, 321, 338, 339, 340, 341, 342, 343, 344, 362, 363, 364, 365, 366, 367, 387, 388, 389, 390], "bbox": [0.5882031249999999, 0.17752083333333332, 1.0, 0.9887708333333333], "iscrowd": 0, "area": 77151.19544999997, "rle": {"size": [480, 640], "counts": "gb`53N2M3N2N2M3N3A>@a0G9ZOe0L5L4L3M4K5J5K6ZOf0M2N3N1N3M3N101O1_LXKhLh4U3[KkLf4Q3]KPMc4m2_KSMa4j2bKVM_4h2cKWM^4h2bKXM^4h2bKXM_4g2aKYM_4g2aKYM`4g2`KYM_4g2aKYM`4f2`KZMa4e2`KZM`4f2`KZMa4e2_K[Ma4e2`KZMa4e2_K[Ma4e2_KVMg4i2ZKTMi4k2WKUMi4k2WKUMj4j2WKUMi4l2VKTMk4k2UKUMl4k2TKTMl4l2TKTMm4k2SKUMm4l2SKSMn4l2RKTMn4m2QKSMP5l2QKSMZ5b2fJ^MY5d2fJ\\MZ5e2fJZMZ5f2fJYMZ5i2eJVM\\5j2dJUM]5l2cJRM_5n2`JQMa5o2_JPMc5P3]JnLe5Q3[JnLg5Q3YJnLi5Q3XJlLj5T3VJjLm5U3SJiLP6V3QJeLT6Z3lIcLW6]3iI`L[6_3fI_L\\6`3dI\\La6c3_IULi6k3XITLi6k3WIULj6j3VIVLk6i3VIULP7f3PIZLU7a3kH_LV7`3kH_LU7a3kH_LU7a3kH_LV7`3kH^LV7b3jH^LV7b3lH\\LU7c3lH\\LT7d3mH[LS7e3nHZLR7f3oHXLS7g3nHXLR7h3oHWLQ7i3PIVLP7j3QITLP7l3RIQLo6o3RIoKo6Q4RInKn6R4SIlKn6T4SIjKn6V4SIhKn6X4SIfKn6Z4SIeKm6Z4UIdKl6\\4m1O001O001O001O001O001O001O001O001O001O001O001N2O1N2O1N2O2M2O1N2O1N2O1N200O1O1O1O100O1O1O1O101N1O2N1O010O00001O01O01O000010O01O00001O000010O0001O00000001O0000001O000000001O000000000000O1000000000000O100000_IkLW1V3gNjLZ1V3eNkL[1V3cNjL^1V3aNkL_1V3_NjLb1V3]NkLc1V3[NkLe1U3ZNkLg1V3WNkLi1U3VNkLk1V3SNkLn1T3QNmLo1T3PNkLQ2U3nMlLR2T3mMmLS2T3lMlLT2T3kMnLT2R3kMoLU2R3iMoLV2R3jMnLV2S3hMnLX2R3gMPMX2P3hMPMX2Q3fMPMZ2P3eMmLjKCa6a3cMlLb2T3\\MnLd2R3[MnLf2S3XMmLi2S3VMmLk2S3TMnLl2S3RMmLo2S3PMmLQ3T3mLlLT3T3kLlLV3T3hLlLZ3U3dLkL]3U3bLkL_3U3`LkLa3V3]LiLe3W3ZLiLg3W3XLiLi3X3_J]LI;i5X3]JfL[O8Y6S3ZJoLiN8n6i2XJ\\Nj5e1TJ[Nm5e1QJ\\NP6e1nI[NS6e1lI[NU6T2[IlMf6PN"}, "label": "a light brown teddy bear wearing a red and white striped shirt"}]} {"id": 522229, "image": "COCO_train2014_000000522229.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass of water on a table near a plate of food\"."}], "task": "refering", "answer_template": "The \"a glass of water on a table near a plate of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 43, 44, 45, 66, 67, 68, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 364, 365, 366, 367, 389, 390], "bbox": [0.77190625, 0.0022500000000000003, 0.9994375, 0.9595416666666666], "iscrowd": 0, "area": 47268.24985000001, "rle": {"size": [480, 640], "counts": "amW78c>a0@=B6K5J5L5J6K5J5L5J6K5J6K4K6K5J6K4K6K5J6K4K6K5K5K4L5K5K5J5L5K5K5K5K4L5J6K5K4L5K5K5K4K6K5K5K4L5K5K5J5L5K5K5K5K4L5J6K5K4L5K5K5K4L5J6K5K4L5K5K4L4K6K4L4L4L5K4L4K5L5K4L4L4L5K4L4000010O00000010O000001O01O0001O01O0001O0001O01O0000010O00O10001O000000001O000000001N100000001O0000001O0000000O2[Od0fIZ6eIn5"}, "label": "a glass of water on a table near a plate of food"}]} {"id": 522229, "image": "COCO_train2014_000000522229.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass full of honey - colored beer on a table\"."}], "task": "refering", "answer_template": "The \"a glass full of honey - colored beer on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 43, 44, 45, 66, 67, 68, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298, 316, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 364, 365, 366, 367, 389, 390], "bbox": [0.77190625, 0.0022500000000000003, 0.9994375, 0.9595416666666666], "iscrowd": 0, "area": 47268.24985000001, "rle": {"size": [480, 640], "counts": "amW78c>a0@=B6K5J5L5J6K5J5L5J6K5J6K4K6K5J6K4K6K5J6K4K6K5K5K4L5K5K5J5L5K5K5K5K4L5J6K5K4L5K5K5K4K6K5K5K4L5K5K5J5L5K5K5K5K4L5J6K5K4L5K5K5K4L5J6K5K4L5K5K4L4K6K4L4L4L5K4L4K5L5K4L4L4L5K4L4000010O00000010O000001O01O0001O01O0001O0001O01O0000010O00O10001O000000001O000000001N100000001O0000001O0000000O2[Od0fIZ6eIn5"}, "label": "a glass full of honey - colored beer on a table"}]} {"id": 432138, "image": "COCO_train2014_000000432138.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a kid playing baseball , squatting down because he is playing catcher\"."}], "task": "refering", "answer_template": "The \"a kid playing baseball , squatting down because he is playing catcher\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 166, 167, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 277, 278, 279, 280, 281], "bbox": [0.053171875, 0.4685680751173709, 0.275, 0.8633098591549295], "iscrowd": 0, "area": 12270.6422, "rle": {"size": [426, 640], "counts": "e]>5Q=7K4L4bCCi;a0RDHe;]3]8kLbGn2e8TMXG\\2Y9gMcFZ2]9iM_FX2b9g000000010O2N100O1O2N100O2N100001O001O0O2O0O2O0O3N1O2M2O2M2O2M2O2N1N3N1N3N1O2M2gNdEBd0UOi8R1jFF=YOj8k0QGI4^Ol8c0VGLMCm8a0WGKJFo8?WGKHIP9X9GZGJ^O`0X9FZGOXO<^9EZGY1e8fN\\GZ1d8fN\\GZ1d8fN\\GZ1d8eN]G\\1b8dN^G\\1b8dN]G]1c8bN_G]1a8cN_G^1`8bNaG]1_8cNbG\\1^8cNdG]1[8cNeG]1Y8eNhGZ1W8fNkGY1S8iNmGX1Q8iNPHV1n7kNTHT1k7mNUHS1j7nNWHP1h7QOZHn0e7SO\\HKjNNi89]HFSOK^8`0aHAYOIV8g0bH]O@Fl7o0ZJkNe5V1W31N2O010000O01000O100O1O1O100O001O100O1O1O1G9F9GljP6"}, "label": "a kid playing baseball , squatting down because he is playing catcher"}]} {"id": 432138, "image": "COCO_train2014_000000432138.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young man squatted down behind the batter acting as a catcher for a baseball game\"."}], "task": "refering", "answer_template": "The \"a young man squatted down behind the batter acting as a catcher for a baseball game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [163, 164, 166, 167, 185, 186, 187, 188, 189, 190, 208, 209, 210, 211, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 277, 278, 279, 280, 281], "bbox": [0.053171875, 0.4685680751173709, 0.275, 0.8633098591549295], "iscrowd": 0, "area": 12270.6422, "rle": {"size": [426, 640], "counts": "e]>5Q=7K4L4bCCi;a0RDHe;]3]8kLbGn2e8TMXG\\2Y9gMcFZ2]9iM_FX2b9g000000010O2N100O1O2N100O2N100001O001O0O2O0O2O0O3N1O2M2O2M2O2M2O2N1N3N1N3N1O2M2gNdEBd0UOi8R1jFF=YOj8k0QGI4^Ol8c0VGLMCm8a0WGKJFo8?WGKHIP9X9GZGJ^O`0X9FZGOXO<^9EZGY1e8fN\\GZ1d8fN\\GZ1d8fN\\GZ1d8eN]G\\1b8dN^G\\1b8dN]G]1c8bN_G]1a8cN_G^1`8bNaG]1_8cNbG\\1^8cNdG]1[8cNeG]1Y8eNhGZ1W8fNkGY1S8iNmGX1Q8iNPHV1n7kNTHT1k7mNUHS1j7nNWHP1h7QOZHn0e7SO\\HKjNNi89]HFSOK^8`0aHAYOIV8g0bH]O@Fl7o0ZJkNe5V1W31N2O010000O01000O100O1O1O100O001O100O1O1O1G9F9GljP6"}, "label": "a young man squatted down behind the batter acting as a catcher for a baseball game"}]} {"id": 79883, "image": "COCO_train2014_000000079883.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the hands . of a woman holding a puppy\"."}], "task": "refering", "answer_template": "The \"the hands . of a woman holding a puppy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 17, 18, 19, 20, 21, 22, 23, 24, 27, 31, 32, 33, 34, 35, 36, 37, 38, 41, 46, 47, 48, 49, 50, 51, 52, 55, 62, 63, 69, 94, 95, 96, 97, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 134, 135, 136, 137, 138], "bbox": [0.24535897435897436, 0.0011262798634812288, 1.0, 0.9876109215017065], "iscrowd": 0, "area": 31036.6098, "rle": {"size": [293, 390], "counts": "P_k02S92N2N2N2\\HHW6:cI2U60eI6Y65XI0g6X1O2N1O1O1O1O2N2N2N1O2N2N2N1O2N2N2N1O1O1O1O1O2N1OO10000000000O10000000000O10000O101O000O10000000000000000000000O100001O001O00001O001O002N3M3M4L2N1O1O2N1O1O2N1O1O2N1O00001O00001O00kNlJYNT5d1PK[NP5e1QK[Nn4e1SKZNm4f1UKXNk4i1UKWNj4i1WKUNj4k1WKTNi4m1WKRNi4n1YKoMh4Q2YKnMg4S2YKlMg4U2YKiMh4W2ZKgMf4Z2ZKeMf4\\2ZKbMg4_2YK`Mg4a2YK[Mj4f2WKSMn4n2d01O1O1O1O001O1O1O1O001O1O001O1O001O00001O001O001O3fNbLgL`3V3bLiL_3U3dLiL\\3V3gLhLY3V3iLjLW3U3kLjLU3U3lLkLT3T3mLlLS3S3oLlLP3S3RMmLl2T3UMfLZO_O`3k3XMdLYOA^3k3ZMcLYOB]3j3[MbLZOD[3i3]MaLYOFZ3g3_MbLXOGY3f3`MbLXOHY3d3`McLXOIX3c3aMcLXOJX3`3cMdLVOLW3_3dMdLVOMV3^3eMdLVONV3[3fMfLUOOU3Z3gMfLUO0T3X3iMgLTO1S3W3^NiLb1U3`NkL`1T3aNlL_1S3bNmL^1Q3dNoL\\1P3eNPMX1R3iNnLU1S3lNmLQ1U3POkLn0W3ROiLl0X3UOhLj0X3WOhLg0Z3YOfLf0Z3[OfLc0[3^OeLa0\\3_OdL`0\\3AdL>\\3CdL<\\3EdL:]3FcLiNh0YOd2o1dLSNa1Ik1U2dLdMS21[1[2bLdMU2NY1_2bLcMW2KW1c2bLbMm5^2SJbMl5_2TJaMk5`2UJ`Mk5`2UJ`Mk5`2UJ`Mj5a2VJ_Mj5a2VJ_Mi5b2WJ^Mi5b2WJ^Mi5b2WJ^Mh5c2XJ]Mh5c2XJ]Mg5d2YJ\\Mg5d2YJ[Mh5e2XJ[Mg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJYMh5g2XJYMh5g2XJYMh5g2XJYMh5g2XJYMh5g2XJYMh5g2XJYMh5g2XJYMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMj5e2VJ[Mj5e2VJ[Mi5f2WJZMh5m0\\K^OlNE[2Kk0S1dNRO`0Mj0R1fNQO`0Ng0R1jNPO?Ng0R1jNPO?Og0P1jNPO`01e0n0lNQO?2d0m0mNQO?3c0k0oNRO>4b0j0PORO=6b0h0QORO=7a0f0SOSO<9`0c0TOTO<;>`0WOUO;<=>YOVO:?:;\\OVO:h011FYO7j0MNMYO4k0ML0ZO2l0KK3ZOdNY2V1^N7j2FVM;j2DVM=k2@UMb0k2\\OVMe0j2YOWMh0i2WOWMj0j2SOWMm0"}, "label": "the hands . of a woman holding a puppy"}]} {"id": 79883, "image": "COCO_train2014_000000079883.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"hands placed on lap with a silver ring on the right hand\"."}], "task": "refering", "answer_template": "The \"hands placed on lap with a silver ring on the right hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 17, 18, 19, 20, 21, 22, 23, 24, 27, 31, 32, 33, 34, 35, 36, 37, 38, 41, 46, 47, 48, 49, 50, 51, 52, 55, 62, 63, 69, 94, 95, 96, 97, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 134, 135, 136, 137, 138], "bbox": [0.24535897435897436, 0.0011262798634812288, 1.0, 0.9876109215017065], "iscrowd": 0, "area": 31036.6098, "rle": {"size": [293, 390], "counts": "P_k02S92N2N2N2\\HHW6:cI2U60eI6Y65XI0g6X1O2N1O1O1O1O2N2N2N1O2N2N2N1O2N2N2N1O1O1O1O1O2N1OO10000000000O10000000000O10000O101O000O10000000000000000000000O100001O001O00001O001O002N3M3M4L2N1O1O2N1O1O2N1O1O2N1O00001O00001O00kNlJYNT5d1PK[NP5e1QK[Nn4e1SKZNm4f1UKXNk4i1UKWNj4i1WKUNj4k1WKTNi4m1WKRNi4n1YKoMh4Q2YKnMg4S2YKlMg4U2YKiMh4W2ZKgMf4Z2ZKeMf4\\2ZKbMg4_2YK`Mg4a2YK[Mj4f2WKSMn4n2d01O1O1O1O001O1O1O1O001O1O001O1O001O00001O001O001O3fNbLgL`3V3bLiL_3U3dLiL\\3V3gLhLY3V3iLjLW3U3kLjLU3U3lLkLT3T3mLlLS3S3oLlLP3S3RMmLl2T3UMfLZO_O`3k3XMdLYOA^3k3ZMcLYOB]3j3[MbLZOD[3i3]MaLYOFZ3g3_MbLXOGY3f3`MbLXOHY3d3`McLXOIX3c3aMcLXOJX3`3cMdLVOLW3_3dMdLVOMV3^3eMdLVONV3[3fMfLUOOU3Z3gMfLUO0T3X3iMgLTO1S3W3^NiLb1U3`NkL`1T3aNlL_1S3bNmL^1Q3dNoL\\1P3eNPMX1R3iNnLU1S3lNmLQ1U3POkLn0W3ROiLl0X3UOhLj0X3WOhLg0Z3YOfLf0Z3[OfLc0[3^OeLa0\\3_OdL`0\\3AdL>\\3CdL<\\3EdL:]3FcLiNh0YOd2o1dLSNa1Ik1U2dLdMS21[1[2bLdMU2NY1_2bLcMW2KW1c2bLbMm5^2SJbMl5_2TJaMk5`2UJ`Mk5`2UJ`Mk5`2UJ`Mj5a2VJ_Mj5a2VJ_Mi5b2WJ^Mi5b2WJ^Mi5b2WJ^Mh5c2XJ]Mh5c2XJ]Mg5d2YJ\\Mg5d2YJ[Mh5e2XJ[Mg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJZMg5f2YJYMh5g2XJYMh5g2XJYMh5g2XJYMh5g2XJYMh5g2XJYMh5g2XJYMh5g2XJYMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMi5f2WJZMj5e2VJ[Mj5e2VJ[Mi5f2WJZMh5m0\\K^OlNE[2Kk0S1dNRO`0Mj0R1fNQO`0Ng0R1jNPO?Ng0R1jNPO?Og0P1jNPO`01e0n0lNQO?2d0m0mNQO?3c0k0oNRO>4b0j0PORO=6b0h0QORO=7a0f0SOSO<9`0c0TOTO<;>`0WOUO;<=>YOVO:?:;\\OVO:h011FYO7j0MNMYO4k0ML0ZO2l0KK3ZOdNY2V1^N7j2FVM;j2DVM=k2@UMb0k2\\OVMe0j2YOWMh0i2WOWMj0j2SOWMm0"}, "label": "hands placed on lap with a silver ring on the right hand"}]} {"id": 333842, "image": "COCO_train2014_000000333842.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a samsung display on small table with many toys and next to another display of other brand with beautiful keyboard and mouse\"."}], "task": "refering", "answer_template": "The \"a samsung display on small table with many toys and next to another display of other brand with beautiful keyboard and mouse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 81, 82, 83, 84, 85, 86, 87, 88, 89, 109, 110, 111, 112, 133, 134, 135], "bbox": [0.5133593750000001, 0.0, 1.0, 0.38840749414519904], "iscrowd": 0, "area": 34163.28805, "rle": {"size": [427, 640], "counts": "SVY4?9HSB=CB1O001O1O1O1O001O3M2N00000000000hGaKd7_4XHeKh7[4SHjKm7V4PHmKP8S4nGoKR8Q4lGQLT8c400000000000000000000000000000000000000000000000000000000000000000O100O1O100O1O1O100O1O100O1O1O100POcGQM^8l2kGjLY8T3PH^LW8`3k0M3N2N2M3N2K5K5E;00O100000000000000O1000000000000O1000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O1"}, "label": "a samsung display on small table with many toys and next to another display of other brand with beautiful keyboard and mouse"}]} {"id": 333842, "image": "COCO_train2014_000000333842.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the monitor with mario in front of it\"."}], "task": "refering", "answer_template": "The \"the monitor with mario in front of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 122, 125, 126, 138, 140, 141, 142, 144, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190], "bbox": [0.0015, 0.0022482435597189696, 0.493265625, 0.5865105386416862], "iscrowd": 0, "area": 52805.887350000005, "rle": {"size": [427, 640], "counts": "Wa0`1l8o2TOl00001O000000001O000000001O0000000000O1O100O1O1O1O1O1O1O1O100O1O1O1O100001O00000000001O0000000000001O000000100O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2UI]Kj4d4QK`Ko4a4mJbKS5_4hJeKX5\\4cJhK]5S60000000000000000000000000000000000000000000000000\\Nd1\\Od00000000000000000000000000000IVHoJh7Q5\\HoJ`7P5dHQKX7o4kHSKP7m4TISKh6m4[IUK`6k4cIVKY6j4kIVKQ6j4RJXKi5h4XJ[Kd5d4]JaK^5_4bJeKZ5[4eJjKW5V4iJoKR5Q4nJSLn4m3RKSLn4m3RKRLo4n3QKRLo4n3QKRLo4m3RKSLn4m3RKRLo4n3QKRLo4n3QKRLo4n3QKRLo4n3QKRLo4n3QKQLP5o3PKQLP5n3QKRLo4n3QKRLo4n3QKQLP5o3PKQLP5o3PKQLP5o3PKQLP5o3PKQLP5o3PKPLQ5o3PKQLP5o3PKQLP5o3PKQLP5o3PKPLQ5P4oJPLQ5P4oJPLQ5P4oJPLQ5P4oJPLQ5o3oJQLR5o3nJQLR5o3nJQLS5n3mJRLS5n3mJQLT5o3lJQLT5o3lJQLT5o3lJQLT5n3mJRLS5n3mJQLT5o3lJQLT5o3lJQLT5o3lJQLT5o3lJPLU5P4kJPLU5P4kJPLU5o3lJQLT5o3lJQLT5o3lJPLU5P4kJPLU5P4kJPLU5P4kJPLU5P4kJoKV5Q4jJoKV5Q4jJoKV5P4kJPLU5P4kJPLU5P4kJoKV5Q4kJnKU5R4oJjKQ5V4RKgKn4Y4UKdKk4\\4YK_Kh4`4\\K]Kd4c4_KZKa4f4cKVK]4j4R200000000O10000N2M3N2N2M3N2N200O1O1O1O1O100O1O1@YG[Lh8e3?00000000000000000000000000000000000000001O001O1O1O1O001O1O1O001O1O1O1O001O1OO1L4M3O11O1O001O1O1O00100O1O=C000000000000000000000000001O000000000000000000000000000008Ha0_Oa0_Oa0@`0_O`0@a0_Oa0_Oa0_OinV4"}, "label": "the monitor with mario in front of it"}]} {"id": 333842, "image": "COCO_train2014_000000333842.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the monitor on the left\"."}], "task": "refering", "answer_template": "The \"the monitor on the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 115, 116, 117, 118, 119, 120, 121, 122, 125, 126, 138, 140, 141, 142, 144, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190], "bbox": [0.0015, 0.0022482435597189696, 0.493265625, 0.5865105386416862], "iscrowd": 0, "area": 52805.887350000005, "rle": {"size": [427, 640], "counts": "Wa0`1l8o2TOl00001O000000001O000000001O0000000000O1O100O1O1O1O1O1O1O1O100O1O1O1O100001O00000000001O0000000000001O000000100O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O1O2UI]Kj4d4QK`Ko4a4mJbKS5_4hJeKX5\\4cJhK]5S60000000000000000000000000000000000000000000000000\\Nd1\\Od00000000000000000000000000000IVHoJh7Q5\\HoJ`7P5dHQKX7o4kHSKP7m4TISKh6m4[IUK`6k4cIVKY6j4kIVKQ6j4RJXKi5h4XJ[Kd5d4]JaK^5_4bJeKZ5[4eJjKW5V4iJoKR5Q4nJSLn4m3RKSLn4m3RKRLo4n3QKRLo4n3QKRLo4m3RKSLn4m3RKRLo4n3QKRLo4n3QKRLo4n3QKRLo4n3QKRLo4n3QKQLP5o3PKQLP5n3QKRLo4n3QKRLo4n3QKQLP5o3PKQLP5o3PKQLP5o3PKQLP5o3PKQLP5o3PKPLQ5o3PKQLP5o3PKQLP5o3PKQLP5o3PKPLQ5P4oJPLQ5P4oJPLQ5P4oJPLQ5P4oJPLQ5o3oJQLR5o3nJQLR5o3nJQLS5n3mJRLS5n3mJQLT5o3lJQLT5o3lJQLT5o3lJQLT5n3mJRLS5n3mJQLT5o3lJQLT5o3lJQLT5o3lJQLT5o3lJPLU5P4kJPLU5P4kJPLU5o3lJQLT5o3lJQLT5o3lJPLU5P4kJPLU5P4kJPLU5P4kJPLU5P4kJoKV5Q4jJoKV5Q4jJoKV5P4kJPLU5P4kJPLU5P4kJoKV5Q4kJnKU5R4oJjKQ5V4RKgKn4Y4UKdKk4\\4YK_Kh4`4\\K]Kd4c4_KZKa4f4cKVK]4j4R200000000O10000N2M3N2N2M3N2N200O1O1O1O1O100O1O1@YG[Lh8e3?00000000000000000000000000000000000000001O001O1O1O1O001O1O1O001O1O1O1O001O1OO1L4M3O11O1O001O1O1O00100O1O=C000000000000000000000000001O000000000000000000000000000008Ha0_Oa0_Oa0@`0_O`0@a0_Oa0_Oa0_OinV4"}, "label": "the monitor on the left"}]} {"id": 505885, "image": "COCO_train2014_000000505885.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back of a girl\"."}], "task": "refering", "answer_template": "The \"the back of a girl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [136, 137, 150, 151, 152, 153, 166, 167, 168, 169, 170, 181, 182, 183, 184, 185, 187, 198, 199, 200, 202, 203, 204, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 245, 246, 247, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 268, 278, 279, 280, 281, 282, 283, 284, 294, 295, 296, 297, 298, 299, 300, 311, 312, 313, 314, 315, 316, 327, 328, 329, 330, 331, 332, 343, 344, 345, 346, 347, 348, 359, 360, 361, 362, 363, 364, 365], "bbox": [0.365296803652968, 0.371625, 0.8227397260273973, 0.9774843750000001], "iscrowd": 0, "area": 48087.28539999999, "rle": {"size": [640, 438], "counts": "WZT33g2?l=GPBc0e=_OZBl0Y=XOeBl0U=XOiBi0U=YOjBi0R=[OkBg0R=]OkBf0R=\\OlBf0Q=^OlBc0S=^OkBc0T=^OlBb0R=@nBa0o in this image.", "objects": [{"patches": [136, 137, 150, 151, 152, 153, 166, 167, 168, 169, 170, 181, 182, 183, 184, 185, 187, 198, 199, 200, 202, 203, 204, 217, 218, 219, 220, 231, 232, 233, 234, 235, 236, 245, 246, 247, 248, 249, 250, 251, 252, 261, 262, 263, 264, 265, 266, 267, 268, 278, 279, 280, 281, 282, 283, 284, 294, 295, 296, 297, 298, 299, 300, 311, 312, 313, 314, 315, 316, 327, 328, 329, 330, 331, 332, 343, 344, 345, 346, 347, 348, 359, 360, 361, 362, 363, 364, 365], "bbox": [0.365296803652968, 0.371625, 0.8227397260273973, 0.9774843750000001], "iscrowd": 0, "area": 48087.28539999999, "rle": {"size": [640, 438], "counts": "WZT33g2?l=GPBc0e=_OZBl0Y=XOeBl0U=XOiBi0U=YOjBi0R=[OkBg0R=]OkBf0R=\\OlBf0Q=^OlBc0S=^OkBc0T=^OlBb0R=@nBa0o in this image.", "objects": [{"patches": [147, 148, 149, 162, 163, 164, 165, 170, 171, 178, 179, 180, 181, 184, 185, 186, 187, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 240, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 272, 273, 274, 275, 276, 277, 278, 288, 289, 290, 291, 292, 293, 294, 295, 304, 305, 306, 307, 308, 309, 310, 311, 320, 321, 322, 323, 324, 325, 326, 327, 337, 338, 339, 340, 341, 342, 343, 353, 354, 355, 356, 357, 358, 359], "bbox": [0.04269406392694064, 0.410546875, 0.7486529680365296, 0.97909375], "iscrowd": 0, "area": 58573.27865000001, "rle": {"size": [640, 438], "counts": "kV in this image.", "objects": [{"patches": [147, 148, 149, 162, 163, 164, 165, 170, 171, 178, 179, 180, 181, 184, 185, 186, 187, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 240, 241, 242, 243, 244, 245, 256, 257, 258, 259, 260, 261, 272, 273, 274, 275, 276, 277, 278, 288, 289, 290, 291, 292, 293, 294, 295, 304, 305, 306, 307, 308, 309, 310, 311, 320, 321, 322, 323, 324, 325, 326, 327, 337, 338, 339, 340, 341, 342, 343, 353, 354, 355, 356, 357, 358, 359], "bbox": [0.04269406392694064, 0.410546875, 0.7486529680365296, 0.97909375], "iscrowd": 0, "area": 58573.27865000001, "rle": {"size": [640, 438], "counts": "kV in this image.", "objects": [{"patches": [48, 65, 66, 83, 84, 85, 101, 102, 103, 119, 120, 137, 138, 154, 155, 173, 174], "bbox": [0.58484, 0.19258666666666666, 0.78876, 0.7363999999999999], "iscrowd": 0, "area": 5283.476699999997, "rle": {"size": [375, 500], "counts": "]a[31g;0O2N1O101N1O1O2O0O2N1O101N1O2N100O2NM3K6J5KU;OQE0O1O1nM1\\HO_79]HHS64VI<]O^Od04`6?UI4E\\Oe01\\6i0PIOOXOc01\\6o0hHN8TOa00]6Y2RIhM`0O\\6[2SIQN6Ee6\\2UIPN4Ee6]2WIoM2Eb6a2[IlM2C]6h2`IfM3A\\6j2aIfM2@]6j2aIgM2_O[6k2bIiM1\\O\\6l2cIiM4WOY6P3cIjMn6X2QIiM^1ROf3V3kJjMY1UOl3Q3kJkMS1YOQ4m2lJkMQ1YOT4k2kJmMP1XOU4k2jJoMP1VOV4k2jJoMP1UOX4k2hJPNP1UOX4k2hJPNP1UOY4j2fJRNQ1TOY4j2fJRNQ1TOY4j2fJRNQ1TOZ4i2dJTNS1ROY4[2SKdNc0QOZ4Y2UKfNb0POY4X2_K`N8WOZ4P1bJI`1_OEGY4o0fJGn1SOTO6X4n0jJGU3[OR2k0nJGQ3]OU2e0nJMm2^OY27TK;d2]O\\21TKa0`2^Of3a0[L^Oe3b0\\L\\Of3c0ZL]Og3b0ZL]Of3c0ZL]Og3b0ZL]Og3b0YL^Oh3a0g310O1O001O1O1O001O1O0O2O00001O000O10001O0000002M3N`eV1"}, "label": "a woman on a bike and her dog"}]} {"id": 79912, "image": "COCO_train2014_000000079912.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in blue dragging her bike in a park\"."}], "task": "refering", "answer_template": "The \"a woman in blue dragging her bike in a park\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [48, 65, 66, 83, 84, 85, 101, 102, 103, 119, 120, 137, 138, 154, 155, 173, 174], "bbox": [0.58484, 0.19258666666666666, 0.78876, 0.7363999999999999], "iscrowd": 0, "area": 5283.476699999997, "rle": {"size": [375, 500], "counts": "]a[31g;0O2N1O101N1O1O2O0O2N1O101N1O2N100O2NM3K6J5KU;OQE0O1O1nM1\\HO_79]HHS64VI<]O^Od04`6?UI4E\\Oe01\\6i0PIOOXOc01\\6o0hHN8TOa00]6Y2RIhM`0O\\6[2SIQN6Ee6\\2UIPN4Ee6]2WIoM2Eb6a2[IlM2C]6h2`IfM3A\\6j2aIfM2@]6j2aIgM2_O[6k2bIiM1\\O\\6l2cIiM4WOY6P3cIjMn6X2QIiM^1ROf3V3kJjMY1UOl3Q3kJkMS1YOQ4m2lJkMQ1YOT4k2kJmMP1XOU4k2jJoMP1VOV4k2jJoMP1UOX4k2hJPNP1UOX4k2hJPNP1UOY4j2fJRNQ1TOY4j2fJRNQ1TOY4j2fJRNQ1TOZ4i2dJTNS1ROY4[2SKdNc0QOZ4Y2UKfNb0POY4X2_K`N8WOZ4P1bJI`1_OEGY4o0fJGn1SOTO6X4n0jJGU3[OR2k0nJGQ3]OU2e0nJMm2^OY27TK;d2]O\\21TKa0`2^Of3a0[L^Oe3b0\\L\\Of3c0ZL]Og3b0ZL]Of3c0ZL]Og3b0ZL]Og3b0YL^Oh3a0g310O1O001O1O1O001O1O0O2O00001O000O10001O0000002M3N`eV1"}, "label": "a woman in blue dragging her bike in a park"}]} {"id": 268334, "image": "COCO_train2014_000000268334.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man standing near the sheep touching a lying sheep with a rod\"."}], "task": "refering", "answer_template": "The \"a man standing near the sheep touching a lying sheep with a rod\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 52, 53, 54, 55, 56, 75, 76, 77, 78, 79, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 145, 146, 147, 148, 168, 169, 170, 171, 191, 192, 193, 194, 214, 215, 216, 217, 237, 238, 239, 240, 260, 261, 262, 263, 283, 284], "bbox": [0.260390625, 0.08313888888888889, 0.49803125000000004, 0.9730277777777778], "iscrowd": 0, "area": 23732.9001, "rle": {"size": [360, 640], "counts": "[ij16P;5K3M3M201O0O2O1N101O0O2O1O0O2O001N101O0O2O001N101O001N101O001O0O2O001O0TIgNi3Z1TLmNf3T1PKbNEa0V5n0RKjN^O`0[5f0UKSOWO>_5`0XK[OoN?e0mNe3P2nKmN_Od0>POg3f6oLfHZ1m7nMXHR2_80001N100000001N100000001N100000001N10HkM]GU2R8]NmGc1h7jNVHV1g7nNYHP1e7TOZHl0V6^NbJj0WOh0U6fN[Jk2b5^MUJe2i5]1O2N2NLVJnJg5S5\\JlJb5T580000000O2O1O05L5K5J5L5K5L4L3M^1bN4L3M3M4L3M3M1O2N2N1O1O0001O1N2O1O1M3K5H8H8H8H9H7IR1nN7I8H7K6J5ZL]Ja1i5UNcJd1a5ZNeJ]1`5eNaJS1e5mN^JXOY7j0hHoN]7R1dHfNa7[1bH[Ne7f1n02O2M4M2OM3L4M3M4K8I3M2M3N3BgEG[:3kEKW:OnEOg:O1OfQa3"}, "label": "a man standing near the sheep touching a lying sheep with a rod"}]} {"id": 268334, "image": "COCO_train2014_000000268334.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a older man wearing a gray t - shirt and gray sweat pants\"."}], "task": "refering", "answer_template": "The \"a older man wearing a gray t - shirt and gray sweat pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 52, 53, 54, 55, 56, 75, 76, 77, 78, 79, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 145, 146, 147, 148, 168, 169, 170, 171, 191, 192, 193, 194, 214, 215, 216, 217, 237, 238, 239, 240, 260, 261, 262, 263, 283, 284], "bbox": [0.260390625, 0.08313888888888889, 0.49803125000000004, 0.9730277777777778], "iscrowd": 0, "area": 23732.9001, "rle": {"size": [360, 640], "counts": "[ij16P;5K3M3M201O0O2O1N101O0O2O1O0O2O001N101O0O2O001N101O001N101O001O0O2O001O0TIgNi3Z1TLmNf3T1PKbNEa0V5n0RKjN^O`0[5f0UKSOWO>_5`0XK[OoN?e0mNe3P2nKmN_Od0>POg3f6oLfHZ1m7nMXHR2_80001N100000001N100000001N100000001N10HkM]GU2R8]NmGc1h7jNVHV1g7nNYHP1e7TOZHl0V6^NbJj0WOh0U6fN[Jk2b5^MUJe2i5]1O2N2NLVJnJg5S5\\JlJb5T580000000O2O1O05L5K5J5L5K5L4L3M^1bN4L3M3M4L3M3M1O2N2N1O1O0001O1N2O1O1M3K5H8H8H8H9H7IR1nN7I8H7K6J5ZL]Ja1i5UNcJd1a5ZNeJ]1`5eNaJS1e5mN^JXOY7j0hHoN]7R1dHfNa7[1bH[Ne7f1n02O2M4M2OM3L4M3M4K8I3M2M3N3BgEG[:3kEKW:OnEOg:O1OfQa3"}, "label": "a older man wearing a gray t - shirt and gray sweat pants"}]} {"id": 235582, "image": "COCO_train2014_000000235582.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty wine glass on the table with the birthday cake with candles\"."}], "task": "refering", "answer_template": "The \"an empty wine glass on the table with the birthday cake with candles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261, 280, 281, 282, 283, 304, 305, 328], "bbox": [0.19040625, 0.5505620608899298, 0.35982812500000005, 1.0], "iscrowd": 0, "area": 13940.494, "rle": {"size": [427, 640], "counts": "ZUc1k0U in this image.", "objects": [{"patches": [188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261, 280, 281, 282, 283, 304, 305, 328], "bbox": [0.19040625, 0.5505620608899298, 0.35982812500000005, 1.0], "iscrowd": 0, "area": 13940.494, "rle": {"size": [427, 640], "counts": "ZUc1k0U in this image.", "objects": [{"patches": [92, 115, 116, 138, 139, 140, 161, 162, 163, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 254, 255, 256], "bbox": [0.0, 0.24096018735362998, 0.16568750000000002, 0.7923419203747072], "iscrowd": 0, "area": 14246.801950000001, "rle": {"size": [427, 640], "counts": "Y3\\5T8f0YOL5K4L5K4L4L2O1N2N0000000001O0000000000001O00000000003N3M2N3M3M3M3M3NO0O2N1O2N1N3N1O2N1O2N1O2M2O2N1O2N1O2M2O2N1O2N1O2M2O2N1O2N1O2N1N3N1O2N1O2N1N3N1kNaF\\Na9a1aF^N`9a1aF^Na9_1bF_N_9`1bF^Na9_1aF`N`9_1aF`Na9]1bFaN`9]1aFbNa9[1aFdNa9Z1`FeNa9Y1bFdNa9Z1`FeNb9X1`FgNb9W1`FgNb9V1`FiNb9U1_FjNc9S1_FkNd9S1^FkNd9R1^FmNd9Q1]FnNd9P1^FoNd9o0^FoNd9n0^FPOe9m0]FROe9i0b1K5L4K5K5K]]n6"}, "label": "person with their hands clasped at the table"}]} {"id": 235582, "image": "COCO_train2014_000000235582.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person off screen with their hands balled up\"."}], "task": "refering", "answer_template": "The \"a person off screen with their hands balled up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 115, 116, 138, 139, 140, 161, 162, 163, 184, 185, 186, 187, 207, 208, 209, 210, 230, 231, 232, 233, 254, 255, 256], "bbox": [0.0, 0.24096018735362998, 0.16568750000000002, 0.7923419203747072], "iscrowd": 0, "area": 14246.801950000001, "rle": {"size": [427, 640], "counts": "Y3\\5T8f0YOL5K4L5K4L4L2O1N2N0000000001O0000000000001O00000000003N3M2N3M3M3M3M3NO0O2N1O2N1N3N1O2N1O2N1O2M2O2N1O2N1O2M2O2N1O2N1O2M2O2N1O2N1O2N1N3N1O2N1O2N1N3N1kNaF\\Na9a1aF^N`9a1aF^Na9_1bF_N_9`1bF^Na9_1aF`N`9_1aF`Na9]1bFaN`9]1aFbNa9[1aFdNa9Z1`FeNa9Y1bFdNa9Z1`FeNb9X1`FgNb9W1`FgNb9V1`FiNb9U1_FjNc9S1_FkNd9S1^FkNd9R1^FmNd9Q1]FnNd9P1^FoNd9o0^FoNd9n0^FPOe9m0]FROe9i0b1K5L4K5K5K]]n6"}, "label": "a person off screen with their hands balled up"}]} {"id": 235582, "image": "COCO_train2014_000000235582.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lady wearing a purple shirt with a birthday cake\"."}], "task": "refering", "answer_template": "The \"a lady wearing a purple shirt with a birthday cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 27, 28, 29, 30, 31, 49, 50, 51, 52, 53, 54, 72, 73, 74, 75, 76, 77, 95, 96, 97, 98, 99, 100, 101, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 193, 210, 211, 215, 233, 234, 256, 257], "bbox": [0.13646875, 0.0, 0.40604687500000003, 0.7712177985948477], "iscrowd": 0, "area": 35652.414450000004, "rle": {"size": [427, 640], "counts": "U\\T1a0g< in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 27, 28, 29, 30, 31, 49, 50, 51, 52, 53, 54, 72, 73, 74, 75, 76, 77, 95, 96, 97, 98, 99, 100, 101, 118, 119, 120, 121, 122, 123, 141, 142, 143, 144, 145, 146, 164, 165, 166, 167, 168, 169, 187, 188, 189, 190, 191, 192, 193, 210, 211, 215, 233, 234, 256, 257], "bbox": [0.13646875, 0.0, 0.40604687500000003, 0.7712177985948477], "iscrowd": 0, "area": 35652.414450000004, "rle": {"size": [427, 640], "counts": "U\\T1a0g< in this image.", "objects": [{"patches": [204, 222, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 268, 270, 271, 272, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343, 344], "bbox": [0.65159375, 0.5922014051522249, 0.9981874999999999, 0.9818266978922717], "iscrowd": 0, "area": 12399.263350000001, "rle": {"size": [427, 640], "counts": "oT^55V=:F:F:F:E01000000O10O1000O1000000O0100000O100000O01000000O10O100000O100000O1000000000001N10fNSDV1m;iNSDX1m;hNSDX1m;gNTDY1o;0LgNVDY1n;0O1000001O0O10000000000O1000001O0O10000000000O1000001O0O10000000000O1000001O0O10000000000O1000001O0O10000000000O1000001eE\\Nm8d1TG]Nj8c1WG^Nf8c1[G^Nc8b1]G`Na8`1`GaN^8_1cGbN[8^1fGcNW8^1iGdNU8\\1lGfNQ8Z1PHgNn7MWG9k0Kl7LaI6]6IeI8X6IiI8U6HlI9R6FoIm5BRJa0k5@UJb0i5]OXJf0e5ZOZJi0d5WO\\Jk0b5TO^Jo0`5QO`JQ1]5PObJQ1^5nNcJR1]5nNbJS1^5mNbJS1^5mNaJT1^5lNcJT1]5lNbJU1^5kNbJU1^5jNbJW1^5iNbJW1^5iNbJW1]5iNcJX1]5hNcJX1]5hNbJZ1]5eNdJ[1\\5eNcJ\\1\\5eNdJ[1\\5eNcJ\\1]5cNdJ]1\\5cNcJ^1]5bNcJ^1c5[N]Jf1h5UNXJk1d80O100O100O100O101N100O100O10000O100O100O100O100O100O101N10000O100O10000001O001O001O001O1O010O001O001O1O003M4L4L4L4L3L5E;E;E;E_1"}, "label": "table top between the woman in glasses and the seated man"}]} {"id": 235582, "image": "COCO_train2014_000000235582.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooden table with a cake\"."}], "task": "refering", "answer_template": "The \"wooden table with a cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [204, 222, 223, 224, 225, 226, 227, 245, 246, 247, 248, 249, 268, 270, 271, 272, 295, 296, 297, 298, 318, 319, 320, 321, 341, 342, 343, 344], "bbox": [0.65159375, 0.5922014051522249, 0.9981874999999999, 0.9818266978922717], "iscrowd": 0, "area": 12399.263350000001, "rle": {"size": [427, 640], "counts": "oT^55V=:F:F:F:E01000000O10O1000O1000000O0100000O100000O01000000O10O100000O100000O1000000000001N10fNSDV1m;iNSDX1m;hNSDX1m;gNTDY1o;0LgNVDY1n;0O1000001O0O10000000000O1000001O0O10000000000O1000001O0O10000000000O1000001O0O10000000000O1000001O0O10000000000O1000001eE\\Nm8d1TG]Nj8c1WG^Nf8c1[G^Nc8b1]G`Na8`1`GaN^8_1cGbN[8^1fGcNW8^1iGdNU8\\1lGfNQ8Z1PHgNn7MWG9k0Kl7LaI6]6IeI8X6IiI8U6HlI9R6FoIm5BRJa0k5@UJb0i5]OXJf0e5ZOZJi0d5WO\\Jk0b5TO^Jo0`5QO`JQ1]5PObJQ1^5nNcJR1]5nNbJS1^5mNbJS1^5mNaJT1^5lNcJT1]5lNbJU1^5kNbJU1^5jNbJW1^5iNbJW1^5iNbJW1]5iNcJX1]5hNcJX1]5hNbJZ1]5eNdJ[1\\5eNcJ\\1\\5eNdJ[1\\5eNcJ\\1]5cNdJ]1\\5cNcJ^1]5bNcJ^1c5[N]Jf1h5UNXJk1d80O100O100O100O101N100O100O10000O100O100O100O100O100O101N10000O100O10000001O001O001O001O1O010O001O001O1O003M4L4L4L4L3L5E;E;E;E_1"}, "label": "wooden table with a cake"}]} {"id": 235582, "image": "COCO_train2014_000000235582.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the birthday girl with glasses going to blow our her candles\"."}], "task": "refering", "answer_template": "The \"the birthday girl with glasses going to blow our her candles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 79, 80, 81, 82, 102, 103, 104, 105, 106, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 193, 197, 198, 199, 221, 222], "bbox": [0.420046875, 0.1459718969555035, 0.684796875, 0.6502576112412178], "iscrowd": 0, "area": 19423.2304, "rle": {"size": [427, 640], "counts": "W\\`3i0_<3M3M3M3L4M3L4K5L4LNdD]NZ;W1VEgNh:Z1d00O001001O01O100O00100G9G9F9E in this image.", "objects": [{"patches": [57, 58, 59, 79, 80, 81, 82, 102, 103, 104, 105, 106, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 176, 193, 197, 198, 199, 221, 222], "bbox": [0.420046875, 0.1459718969555035, 0.684796875, 0.6502576112412178], "iscrowd": 0, "area": 19423.2304, "rle": {"size": [427, 640], "counts": "W\\`3i0_<3M3M3M3L4M3L4K5L4LNdD]NZ;W1VEgNh:Z1d00O001001O01O100O00100G9G9F9E in this image.", "objects": [{"patches": [254, 255, 256, 257, 265, 277, 278, 279, 280, 288, 299, 300, 301, 302, 303, 304, 305, 310, 311, 322, 323, 324, 325, 326, 327, 328, 334], "bbox": [0.0, 0.7555503512880563, 0.56203125, 0.9988758782201405], "iscrowd": 0, "area": 13954.293050000004, "rle": {"size": [427, 640], "counts": "d in this image.", "objects": [{"patches": [66, 67, 68, 89, 90, 91, 111, 112, 113, 114, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 247, 248, 249, 271], "bbox": [0.7545468750000001, 0.14391100702576112, 0.995515625, 0.7421077283372365], "iscrowd": 0, "area": 23370.480799999998, "rle": {"size": [427, 640], "counts": "WdY6o0U<:F:G9J6L5J5L4L4K5L5K4K3N2N2M3N2N2M3N2N2M3N2N2M3N3M2M3N2M3N2N2M2O1N2O1N2O1N101N2O1N2O1O1O001O1O000000O100O2OOG:B>I6N3L40O10O10O10O10N2N2N1O2M3N3M3M3M3N2O1N2O1O1N2O1N2O1O1N2O1O1O1O1O2N1O1O1O1O1O100O100O1000000O1000000O10000O2N1O1O100O1O2N1001O0000010O00001O001O1N2O1O1N101N2O1O1N2O1N101N2O1N2N2O0O2N2J6J9Gb0XOXU1"}, "label": "man on right"}]} {"id": 235582, "image": "COCO_train2014_000000235582.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a black shirt at dinner\"."}], "task": "refering", "answer_template": "The \"a man wearing a black shirt at dinner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 68, 89, 90, 91, 111, 112, 113, 114, 133, 134, 135, 136, 137, 155, 156, 157, 158, 159, 160, 178, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 247, 248, 249, 271], "bbox": [0.7545468750000001, 0.14391100702576112, 0.995515625, 0.7421077283372365], "iscrowd": 0, "area": 23370.480799999998, "rle": {"size": [427, 640], "counts": "WdY6o0U<:F:G9J6L5J5L4L4K5L5K4K3N2N2M3N2N2M3N2N2M3N2N2M3N3M2M3N2M3N2N2M2O1N2O1N2O1N101N2O1N2O1O1O001O1O000000O100O2OOG:B>I6N3L40O10O10O10O10N2N2N1O2M3N3M3M3M3N2O1N2O1O1N2O1N2O1O1N2O1O1O1O1O2N1O1O1O1O1O100O100O1000000O1000000O10000O2N1O1O100O1O2N1001O0000010O00001O001O1N2O1O1N101N2O1O1N2O1N101N2O1N2N2O0O2N2J6J9Gb0XOXU1"}, "label": "a man wearing a black shirt at dinner"}]} {"id": 505924, "image": "COCO_train2014_000000505924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman who is marked number 14 and a pink hairband is running away from the goalie\"."}], "task": "refering", "answer_template": "The \"a woman who is marked number 14 and a pink hairband is running away from the goalie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 67, 68, 89, 90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275], "bbox": [0.8787499999999999, 0.10405152224824356, 1.0, 0.7879156908665106], "iscrowd": 0, "area": 14746.469199999994, "rle": {"size": [427, 640], "counts": "beZ71Z=1N2N2`IMVLNc05T32VLL`0MiL6]69ULGa0LlL5[6d0mK^Ok0ImL4Y6Q1eKTO`NIP2MbM6V6X1bKoN]N0V2Li3X1aKEmN\\O`0GQ5Y1aKGkN@=_OX5Z1[KNkNA=WO\\5[1WKf1VOPMb5[1TKl1UOkLf5[1[J]N0g3KbLi5[1XJl2JjKn5[1XJ]3g5dLXJ]3g5dLYJ\\3f5U200O1O1O1O1O001O10O01N10oMnKkJQ4m4\\LoJ^3R5kLiJb2h5eMTJY2l5b2O2N1O101N2NYKbJg1[5YNiJe1V5VNRKh1l4VNYKi1e4WN^Ki1`4[MWK]N>Y4Z4UMeLj2^3mLiLR3i6O00010O000010YOgKZHY4b7oKZHQ4d7TLYHm3c7YLXHj3d7\\LXHe3e7S1L4XIbJ^5b5aJ_J[5e5cJ]JY5h5dJYJY5k5dJWJ[4AoK_7l3iHnKY7n3PIiKU7T4R1M3M4M3L4L5GVM"}, "label": "a woman who is marked number 14 and a pink hairband is running away from the goalie"}]} {"id": 505924, "image": "COCO_train2014_000000505924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white female soccer player that is wearing white shorts and a white shirt that reads 14\"."}], "task": "refering", "answer_template": "The \"a white female soccer player that is wearing white shorts and a white shirt that reads 14\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 67, 68, 89, 90, 91, 112, 113, 114, 135, 136, 137, 158, 159, 160, 181, 182, 183, 204, 205, 206, 227, 228, 229, 250, 251, 252, 273, 274, 275], "bbox": [0.8787499999999999, 0.10405152224824356, 1.0, 0.7879156908665106], "iscrowd": 0, "area": 14746.469199999994, "rle": {"size": [427, 640], "counts": "beZ71Z=1N2N2`IMVLNc05T32VLL`0MiL6]69ULGa0LlL5[6d0mK^Ok0ImL4Y6Q1eKTO`NIP2MbM6V6X1bKoN]N0V2Li3X1aKEmN\\O`0GQ5Y1aKGkN@=_OX5Z1[KNkNA=WO\\5[1WKf1VOPMb5[1TKl1UOkLf5[1[J]N0g3KbLi5[1XJl2JjKn5[1XJ]3g5dLXJ]3g5dLYJ\\3f5U200O1O1O1O1O001O10O01N10oMnKkJQ4m4\\LoJ^3R5kLiJb2h5eMTJY2l5b2O2N1O101N2NYKbJg1[5YNiJe1V5VNRKh1l4VNYKi1e4WN^Ki1`4[MWK]N>Y4Z4UMeLj2^3mLiLR3i6O00010O000010YOgKZHY4b7oKZHQ4d7TLYHm3c7YLXHj3d7\\LXHe3e7S1L4XIbJ^5b5aJ_J[5e5cJ]JY5h5dJYJY5k5dJWJ[4AoK_7l3iHnKY7n3PIiKU7T4R1M3M4M3L4L5GVM"}, "label": "a white female soccer player that is wearing white shorts and a white shirt that reads 14"}]} {"id": 505924, "image": "COCO_train2014_000000505924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl playing soccer as a goalie and has the number 1\"."}], "task": "refering", "answer_template": "The \"a girl playing soccer as a goalie and has the number 1\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 80, 81, 102, 103, 104, 105, 125, 126, 127, 128, 149, 150, 151, 172, 173, 174, 197, 217, 218, 220, 240, 243], "bbox": [0.44321875000000005, 0.17960187353629975, 0.604453125, 0.6836768149882904], "iscrowd": 0, "area": 9297.4249, "rle": {"size": [427, 640], "counts": "Sbf32X=4L3N1N0YGH\\49cKH]48bKJ\\47cKJ]46bKL\\45cKL\\44dKM\\43cKOZ43eKNT49kKHn3?PLCi3d0bKnNXM>V7d0dKO[42fKMZ43gKLX45iKJV47kKHT49mKCU4 in this image.", "objects": [{"patches": [57, 80, 81, 102, 103, 104, 105, 125, 126, 127, 128, 149, 150, 151, 172, 173, 174, 197, 217, 218, 220, 240, 243], "bbox": [0.44321875000000005, 0.17960187353629975, 0.604453125, 0.6836768149882904], "iscrowd": 0, "area": 9297.4249, "rle": {"size": [427, 640], "counts": "Sbf32X=4L3N1N0YGH\\49cKH]48bKJ\\47cKJ]46bKL\\45cKL\\44dKM\\43cKOZ43eKNT49kKHn3?PLCi3d0bKnNXM>V7d0dKO[42fKMZ43gKLX45iKJV47kKHT49mKCU4 in this image.", "objects": [{"patches": [76, 77, 78, 99, 100, 101, 102, 123, 124, 125, 143, 144, 145, 146, 147, 148, 149, 169, 170, 171, 172, 173, 192, 193, 194, 195, 196, 214, 215, 218, 219, 237, 238, 242, 243, 260, 261, 266, 289], "bbox": [0.24889062499999998, 0.2176112412177986, 0.60721875, 0.8176112412177986], "iscrowd": 0, "area": 16521.608700000004, "rle": {"size": [427, 640], "counts": "`_R21U=5K5L5L3001O001O010O001O00100O001O001O010O001O0000010O00000001O00000000010O00000000001O0TGDi4f1h4lNfJ@a0e1i4nNbJ^Od0e1j4mN`J_Of0e1j4mN]J_Oh0f1k4kN[J@j0f1k4jNYJAl0f1k4jNVJAo0f1k4iNTJBP1f1m4iNPJBS1f1m4hNnICU1f1m41SK0l40SK2m4MSK4m4KSK6m4JRK7n4HQK:o4EQKoMXOV1h5i0PKRNGe0\\5W1lJTN85n4e1iJXNh0Ca4S2gJ[NX1QOT4a2dJ^Nh7`1XHaNj7]1THdNn7Z1RHgNo7W1QHjNP8V1nGjNT8W1hGkNX8V1fGkN[8U1cGkN^8V1`GkNa8V1[GkNf8V1XGkNh8V1VGkNk8U1SGkNn8W1nFkNS9U1kFlNU9U1iFkNX9d21N101O0O2O001N101N2O001N2O1O0mL\\F^2g9_M[F`2j9ZMYFe2k9VMVFj2W:O001O001O001O00001O001O001O001O010O001O0010O01O001O01O01O001O010O001O0010O01O001O0O2O00001N101O001N1nNmDKT;3mDMS;1oDOR;OoD1Q;MQE2Q;KQE5o:GUE9l:AYE?g:\\O^Ec0d:WOaEi0a;000000O1000000O1000000O10000000dDWOU:h0hE[OY:e0cE^O^:b0^EAc:?YEDh:;UEHl:=kDFV;?aDD`;S11O1O1O1O1O1O1O1O1O1O1O1N2K5L4L4L4L4L4L4Kf[X3"}, "label": "a young lady wearing dark blue playing soccer"}]} {"id": 505924, "image": "COCO_train2014_000000505924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman moving the ball to the goal while another shout\"."}], "task": "refering", "answer_template": "The \"a woman moving the ball to the goal while another shout\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 99, 100, 101, 102, 123, 124, 125, 143, 144, 145, 146, 147, 148, 149, 169, 170, 171, 172, 173, 192, 193, 194, 195, 196, 214, 215, 218, 219, 237, 238, 242, 243, 260, 261, 266, 289], "bbox": [0.24889062499999998, 0.2176112412177986, 0.60721875, 0.8176112412177986], "iscrowd": 0, "area": 16521.608700000004, "rle": {"size": [427, 640], "counts": "`_R21U=5K5L5L3001O001O010O001O00100O001O001O010O001O0000010O00000001O00000000010O00000000001O0TGDi4f1h4lNfJ@a0e1i4nNbJ^Od0e1j4mN`J_Of0e1j4mN]J_Oh0f1k4kN[J@j0f1k4jNYJAl0f1k4jNVJAo0f1k4iNTJBP1f1m4iNPJBS1f1m4hNnICU1f1m41SK0l40SK2m4MSK4m4KSK6m4JRK7n4HQK:o4EQKoMXOV1h5i0PKRNGe0\\5W1lJTN85n4e1iJXNh0Ca4S2gJ[NX1QOT4a2dJ^Nh7`1XHaNj7]1THdNn7Z1RHgNo7W1QHjNP8V1nGjNT8W1hGkNX8V1fGkN[8U1cGkN^8V1`GkNa8V1[GkNf8V1XGkNh8V1VGkNk8U1SGkNn8W1nFkNS9U1kFlNU9U1iFkNX9d21N101O0O2O001N101N2O001N2O1O0mL\\F^2g9_M[F`2j9ZMYFe2k9VMVFj2W:O001O001O001O00001O001O001O001O010O001O0010O01O001O01O01O001O010O001O0010O01O001O0O2O00001N101O001N1nNmDKT;3mDMS;1oDOR;OoD1Q;MQE2Q;KQE5o:GUE9l:AYE?g:\\O^Ec0d:WOaEi0a;000000O1000000O1000000O10000000dDWOU:h0hE[OY:e0cE^O^:b0^EAc:?YEDh:;UEHl:=kDFV;?aDD`;S11O1O1O1O1O1O1O1O1O1O1O1N2K5L4L4L4L4L4L4Kf[X3"}, "label": "a woman moving the ball to the goal while another shout"}]} {"id": 505924, "image": "COCO_train2014_000000505924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is a girl standing behind another girl on the right side of the picture\"."}], "task": "refering", "answer_template": "The \"there is a girl standing behind another girl on the right side of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 42, 43, 44, 65, 66, 67, 88, 89, 90, 111, 112, 134, 135, 158, 181, 182, 205, 228, 229, 251, 252, 274, 275], "bbox": [0.856640625, 0.05868852459016393, 1.0, 0.7635128805620609], "iscrowd": 0, "area": 5844.379600000005, "rle": {"size": [427, 640], "counts": "mdT71W=6H8]MBiF1Of0j8EPGMJj0k8BWGc1c8dNZG^1b8fN\\G\\1a8fN]G]1a8eN]G\\1b8fN\\G[1c8gN[GZ1d8gN[GZ1d8hNZGY1e8iNYGW1h8iNUGZ1k8P1UGfLk8X3XGgLh8V3]GhLd8e2oGYMR8d2SHZM?Ji5h2nI\\M40o5`2RJ^MJ7U6W2]KmMd4P2ZKSNi4j1TKZNm4f1mJ^NV5`1gJ_N^5a1W3N101N2O0O2O0O2O1N2O1N2N2N2M3N2NeJdNjNZ1h0AnN;?`0]O\\O9S1GYNe0l1Y55M3M3M2N3M3MO10000N2M4jNW1N3M2RO[h6_OnXI7I8H7_D[OT:n0cEXO[:l0^EYOd:g10O1O2N1O4IjG"}, "label": "there is a girl standing behind another girl on the right side of the picture"}]} {"id": 505924, "image": "COCO_train2014_000000505924.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white girl in blue uniform\"."}], "task": "refering", "answer_template": "The \"white girl in blue uniform\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 42, 43, 44, 65, 66, 67, 88, 89, 90, 111, 112, 134, 135, 158, 181, 182, 205, 228, 229, 251, 252, 274, 275], "bbox": [0.856640625, 0.05868852459016393, 1.0, 0.7635128805620609], "iscrowd": 0, "area": 5844.379600000005, "rle": {"size": [427, 640], "counts": "mdT71W=6H8]MBiF1Of0j8EPGMJj0k8BWGc1c8dNZG^1b8fN\\G\\1a8fN]G]1a8eN]G\\1b8fN\\G[1c8gN[GZ1d8gN[GZ1d8hNZGY1e8iNYGW1h8iNUGZ1k8P1UGfLk8X3XGgLh8V3]GhLd8e2oGYMR8d2SHZM?Ji5h2nI\\M40o5`2RJ^MJ7U6W2]KmMd4P2ZKSNi4j1TKZNm4f1mJ^NV5`1gJ_N^5a1W3N101N2O0O2O0O2O1N2O1N2N2N2M3N2NeJdNjNZ1h0AnN;?`0]O\\O9S1GYNe0l1Y55M3M3M2N3M3MO10000N2M4jNW1N3M2RO[h6_OnXI7I8H7_D[OT:n0cEXO[:l0^EYOd:g10O1O2N1O4IjG"}, "label": "white girl in blue uniform"}]} {"id": 350280, "image": "COCO_train2014_000000350280.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person far right bottom corner\"."}], "task": "refering", "answer_template": "The \"person far right bottom corner\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 49, 50, 51, 52, 67, 68, 69, 70, 86, 87, 88, 89, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196], "bbox": [0.6851799999999999, 0.14274143302180686, 0.99322, 0.9857943925233644], "iscrowd": 0, "area": 24618.82289999999, "rle": {"size": [321, 500], "counts": "[h[3>b92O1O1O1O1O0O2O1O1kMYOnJh0Q5[OlJf0R5^OjJc0V5@gJ`0X5DeJ\\2^OaM40b0^2]O`M0Og0`2\\O^ML1k0`2\\O\\MI1o0b2[OZMF2R1ZOkNP2a0bNA1W1XOkNU2`0_N^ODh1C_NY2>]N\\OEi1@aN^2=ZNYOGQ3n1iMYNVOGR3P2iMXNVOFR3Q2jMXNTOFT3Q2jMWNSOFT3S2jMWNRODV3T2iMWNROCW3U2iMVNPODX3V2jMTNoNDY3V2kMUNlNC[3W2kMTNT3l1nLSNR3l1PMRNQ3m1QMRNn2m1UMQNl2n1VMQNi2o1YMoMh2P2ZMoMf2P2f2000001O01O01O00000010O0001O0000010O000100O0010O01O1O010O1O010O1O0010O01O10O01O1O010O1O013L3M4M2M4L3M4M2M4L3L3M3L4M2M4M3L3N3L4M3L3N3L4I7iNV1_Ob0J6K4L5K5K5K4M4K5K5K4L5L4I6J7I8G>C>Blh0"}, "label": "person far right bottom corner"}]} {"id": 350280, "image": "COCO_train2014_000000350280.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in white pours a glass of wine\"."}], "task": "refering", "answer_template": "The \"a woman in white pours a glass of wine\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [32, 33, 49, 50, 51, 52, 67, 68, 69, 70, 86, 87, 88, 89, 103, 104, 105, 106, 107, 120, 121, 122, 123, 124, 125, 139, 140, 141, 142, 143, 156, 157, 158, 159, 160, 161, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196], "bbox": [0.6851799999999999, 0.14274143302180686, 0.99322, 0.9857943925233644], "iscrowd": 0, "area": 24618.82289999999, "rle": {"size": [321, 500], "counts": "[h[3>b92O1O1O1O1O0O2O1O1kMYOnJh0Q5[OlJf0R5^OjJc0V5@gJ`0X5DeJ\\2^OaM40b0^2]O`M0Og0`2\\O^ML1k0`2\\O\\MI1o0b2[OZMF2R1ZOkNP2a0bNA1W1XOkNU2`0_N^ODh1C_NY2>]N\\OEi1@aN^2=ZNYOGQ3n1iMYNVOGR3P2iMXNVOFR3Q2jMXNTOFT3Q2jMWNSOFT3S2jMWNRODV3T2iMWNROCW3U2iMVNPODX3V2jMTNoNDY3V2kMUNlNC[3W2kMTNT3l1nLSNR3l1PMRNQ3m1QMRNn2m1UMQNl2n1VMQNi2o1YMoMh2P2ZMoMf2P2f2000001O01O01O00000010O0001O0000010O000100O0010O01O1O010O1O010O1O0010O01O10O01O1O010O1O013L3M4M2M4L3M4M2M4L3L3M3L4M2M4M3L3N3L4M3L3N3L4I7iNV1_Ob0J6K4L5K5K5K4M4K5K5K4L5L4I6J7I8G>C>Blh0"}, "label": "a woman in white pours a glass of wine"}]} {"id": 350280, "image": "COCO_train2014_000000350280.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman with red hair\"."}], "task": "refering", "answer_template": "The \"woman with red hair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 63, 64, 65, 81, 82, 83, 84, 99, 100, 101, 102, 117, 118, 119, 120, 121, 135, 136, 137, 138, 139, 154, 155, 156, 172, 173], "bbox": [0.50868, 0.2590342679127726, 0.76228, 0.8820560747663551], "iscrowd": 0, "area": 13246.9212, "rle": {"size": [321, 500], "counts": "aT`23l92N4L3M3M3mL2nK2P41kK3R40jK4S4NjK6U4KeK;Y4G^Kc0`4^OVKl0h4YOnJn0P5SOoJn0Q5SOlJP1R5QOmJP1S5QOkJQ1S5POkJS1T5mNkJT1T5nNjJT1U5lNjJU1U5mNiJU1V5kNiJW1U5kNiJk0b5UO\\Ja0o5@PJ:V6HhI6[6JcI8\\6g1O5L1N2O1N2OS1mN010O00000001O0010O0_KXKj3`5O10O10O100O10O10O1000QOlKPLS4P4mKPLT4o3lKRLT4m3kKULU4j3kKWLU4h3kKYLU4f3kKXLX4f3hKTLa4j3_KTLf4i3ZKULk4h3UKWLo4f3PKYLU5d3kJZLd5X3]JgLd5Y3\\JfLf5Y3[JfLe5Z3\\JeLe5Y3\\JgLd5Y3]JfLd5Y3]JfLd5Y3\\JgLd5Y3]JfLd5X3]JhLc5X3^JgLc5X3^JgLc5X3]JhLc5W3^JiLc5V3^JiLb5W3^JiLd5U3\\JkLe5S3]JlLe5R3[JnLf5Q3[JnLf5Q3ZJoLh5n2YJRMh5m2YJQMj5m2VJSMl5k2TJUMn5i2f0O2lMgH\\1[7`NkH\\1W7aNmH[1V7aNPI[1S7aNQI\\1Q7`NUI\\1j7N00001O0N3N1O1O2N1O2N1N2O2N1O2N1O2N1N3N2N2N3MXSU1"}, "label": "woman with red hair"}]} {"id": 350280, "image": "COCO_train2014_000000350280.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman wearing a black shawl looking downward\"."}], "task": "refering", "answer_template": "The \"woman wearing a black shawl looking downward\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 63, 64, 65, 81, 82, 83, 84, 99, 100, 101, 102, 117, 118, 119, 120, 121, 135, 136, 137, 138, 139, 154, 155, 156, 172, 173], "bbox": [0.50868, 0.2590342679127726, 0.76228, 0.8820560747663551], "iscrowd": 0, "area": 13246.9212, "rle": {"size": [321, 500], "counts": "aT`23l92N4L3M3M3mL2nK2P41kK3R40jK4S4NjK6U4KeK;Y4G^Kc0`4^OVKl0h4YOnJn0P5SOoJn0Q5SOlJP1R5QOmJP1S5QOkJQ1S5POkJS1T5mNkJT1T5nNjJT1U5lNjJU1U5mNiJU1V5kNiJW1U5kNiJk0b5UO\\Ja0o5@PJ:V6HhI6[6JcI8\\6g1O5L1N2O1N2OS1mN010O00000001O0010O0_KXKj3`5O10O10O100O10O10O1000QOlKPLS4P4mKPLT4o3lKRLT4m3kKULU4j3kKWLU4h3kKYLU4f3kKXLX4f3hKTLa4j3_KTLf4i3ZKULk4h3UKWLo4f3PKYLU5d3kJZLd5X3]JgLd5Y3\\JfLf5Y3[JfLe5Z3\\JeLe5Y3\\JgLd5Y3]JfLd5Y3]JfLd5Y3\\JgLd5Y3]JfLd5X3]JhLc5X3^JgLc5X3^JgLc5X3]JhLc5W3^JiLc5V3^JiLb5W3^JiLd5U3\\JkLe5S3]JlLe5R3[JnLf5Q3[JnLf5Q3ZJoLh5n2YJRMh5m2YJQMj5m2VJSMl5k2TJUMn5i2f0O2lMgH\\1[7`NkH\\1W7aNmH[1V7aNPI[1S7aNQI\\1Q7`NUI\\1j7N00001O0N3N1O1O2N1O2N1N2O2N1O2N1O2N1N3N2N2N3MXSU1"}, "label": "woman wearing a black shawl looking downward"}]} {"id": 374873, "image": "COCO_train2014_000000374873.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red and gray skis sticking out of the snow\"."}], "task": "refering", "answer_template": "The \"red and gray skis sticking out of the snow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [218, 239, 240, 241, 242, 263, 264, 265, 286, 287, 288, 289, 310, 311, 312, 313, 334, 335, 336, 358, 359], "bbox": [0.41460937500000006, 0.51925, 0.6488750000000001, 0.926], "iscrowd": 0, "area": 4383.273499999989, "rle": {"size": [480, 640], "counts": "\\ol33k>3M4M2M3M3N2M4M29G1O0001O01O0000010O00000010O0001O0001O01O0000010O0000001O01O0001O0001O01OjMHcE8U:1jEOo98RFHh9>WFBf9b0YF^Od9f0[FZOb90dD9j1F`9JPE`0_1GU:9jEGW:9hEGY:9gEFY:;fEE[::eEG[:9dEG]:9cEF^::aEF`::_EFb::]EGc:9]EFc:;\\EEe:;ZEEg:;YEDh:;XEFh::WEFj::UEFl::TEEm:;REEo:;PEFo:;PEEQ;;oDDR; in this image.", "objects": [{"patches": [218, 239, 240, 241, 242, 263, 264, 265, 286, 287, 288, 289, 310, 311, 312, 313, 334, 335, 336, 358, 359], "bbox": [0.41460937500000006, 0.51925, 0.6488750000000001, 0.926], "iscrowd": 0, "area": 4383.273499999989, "rle": {"size": [480, 640], "counts": "\\ol33k>3M4M2M3M3N2M4M29G1O0001O01O0000010O00000010O0001O0001O01O0000010O0000001O01O0001O0001O01OjMHcE8U:1jEOo98RFHh9>WFBf9b0YF^Od9f0[FZOb90dD9j1F`9JPE`0_1GU:9jEGW:9hEGY:9gEFY:;fEE[::eEG[:9dEG]:9cEF^::aEF`::_EFb::]EGc:9]EFc:;\\EEe:;ZEEg:;YEDh:;XEFh::WEFj::UEFl::TEEm:;REEo:;PEFo:;PEEQ;;oDDR; in this image.", "objects": [{"patches": [156, 157, 158, 178, 179, 201, 205, 228, 251, 252, 274, 275, 297, 298, 320, 321, 343, 344], "bbox": [0.768640625, 0.3142857142857143, 0.9996562499999999, 0.7558458646616542], "iscrowd": 0, "area": 6527.017550000001, "rle": {"size": [532, 640], "counts": "Xjo7T1^?O1O1N2O1O1N3N1O1N2O1O1N2O1N3O100000001O00000000001O00000000001O1O1O001O1O1O10O10O100000O1000000000000O1000000000000O1000O10000000O1000001O000O100010O1O1O010O1O1O1O100O1O001ORFN[32bL0_3O_L3b3L[L7f3HXL:i3FSL=m3CQL`0n93N2N2H8kMTOZDT1f;TOiCU1W in this image.", "objects": [{"patches": [156, 157, 158, 178, 179, 201, 205, 228, 251, 252, 274, 275, 297, 298, 320, 321, 343, 344], "bbox": [0.768640625, 0.3142857142857143, 0.9996562499999999, 0.7558458646616542], "iscrowd": 0, "area": 6527.017550000001, "rle": {"size": [532, 640], "counts": "Xjo7T1^?O1O1N2O1O1N3N1O1N2O1O1N2O1N3O100000001O00000000001O00000000001O1O1O001O1O1O10O10O100000O1000000000000O1000000000000O1000O10000000O1000001O000O100010O1O1O010O1O1O1O100O1O001ORFN[32bL0_3O_L3b3L[L7f3HXL:i3FSL=m3CQL`0n93N2N2H8kMTOZDT1f;TOiCU1W in this image.", "objects": [{"patches": [84, 85, 86, 87, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180], "bbox": [0.66509375, 0.3422779922779923, 0.867359375, 0.8746332046332046], "iscrowd": 0, "area": 9829.070899999999, "rle": {"size": [259, 640], "counts": "Yk[32o0 in this image.", "objects": [{"patches": [84, 85, 86, 87, 107, 108, 109, 110, 111, 130, 131, 132, 133, 134, 153, 154, 155, 156, 157, 176, 177, 178, 179, 180], "bbox": [0.66509375, 0.3422779922779923, 0.867359375, 0.8746332046332046], "iscrowd": 0, "area": 9829.070899999999, "rle": {"size": [259, 640], "counts": "Yk[32o0 in this image.", "objects": [{"patches": [75, 76, 77, 78, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 167, 168, 169, 171, 172], "bbox": [0.273734375, 0.3498841698841699, 0.49381249999999993, 0.8847104247104247], "iscrowd": 0, "area": 10452.833500000004, "rle": {"size": [259, 640], "counts": "j[\\18e7?G6I;Ei0WO2N2O2M2N2O1N4M:F>A4M2M4M3L3N1N100O101OSOnKhMS4U2QLjMo3S2ULlMk3n1eLhM[3W2iLfMW3W2nLcMV3[2Z1J6O1O1O2O01O010O001O010O001O01O01O0010O00gKlMc2S2\\MPNd2o1YMUNf2k1XMWNh2i1TM\\Nl2c1QM`NP3_1lLfNT3Y1iLjNW3V1dLoN]3P1]LVOd3j0QL@P4?nKYOIQOY4k1gKPO:QOP4Q3VLkLj3T3d0N1L5G8J6L5M2N3M2L5L3O101N1000001O000100O00001O000001O01O0002O01O0O2M2N9G6L0010O00001O2N101N15K5LO000000O0000O1000000O2N2N1O2N1O2N1N2O1O001O0^N^Jb0c5]OXKEl49`1O2MYia2"}, "label": "the zebra in the middle"}]} {"id": 309338, "image": "COCO_train2014_000000309338.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra drinking water in the left side of the image\"."}], "task": "refering", "answer_template": "The \"zebra drinking water in the left side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 98, 99, 100, 101, 102, 121, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 167, 168, 169, 171, 172], "bbox": [0.273734375, 0.3498841698841699, 0.49381249999999993, 0.8847104247104247], "iscrowd": 0, "area": 10452.833500000004, "rle": {"size": [259, 640], "counts": "j[\\18e7?G6I;Ei0WO2N2O2M2N2O1N4M:F>A4M2M4M3L3N1N100O101OSOnKhMS4U2QLjMo3S2ULlMk3n1eLhM[3W2iLfMW3W2nLcMV3[2Z1J6O1O1O2O01O010O001O010O001O01O01O0010O00gKlMc2S2\\MPNd2o1YMUNf2k1XMWNh2i1TM\\Nl2c1QM`NP3_1lLfNT3Y1iLjNW3V1dLoN]3P1]LVOd3j0QL@P4?nKYOIQOY4k1gKPO:QOP4Q3VLkLj3T3d0N1L5G8J6L5M2N3M2L5L3O101N1000001O000100O00001O000001O01O0002O01O0O2M2N9G6L0010O00001O2N101N15K5LO000000O0000O1000000O2N2N1O2N1O2N1N2O1O001O0^N^Jb0c5]OXKEl49`1O2MYia2"}, "label": "zebra drinking water in the left side of the image"}]} {"id": 473182, "image": "COCO_train2014_000000473182.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bike in the forest\"."}], "task": "refering", "answer_template": "The \"a bike in the forest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 25, 26, 27, 28, 29, 30, 31, 32, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 70, 71, 72, 73, 74, 75, 76, 77, 78, 93, 94, 95, 96, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.0, 0.0, 0.455890625, 1.0], "iscrowd": 0, "area": 94869.7276, "rle": {"size": [428, 640], "counts": "P;\\2j:6N2O1N2O1N2O1N2O1N2O1N2N2O1N2N2O1N2N2N2O1N2N2O1lJjMPOX2a0\\NXOf1c0bNYO_1f0dNVO^1i0cNUO_1i0cNVO^1i0dNTO^1k0cNSO_1k0cNSO_1l0bNSO_1k0aNVO`1i0_NWOc1g0]NZOd1d0\\N]Oe1b0[N^Of1`0ZNAg1=YNDh1;WNFj17WNIk13XNMi10ZNOg1M]N2d1M]N2d1L^N3c1L^NXO`MXNS4^2`NWO]2h0dMWO]2h0dMVO^2i0cMVO^2i0bMVO`2i0aMVO`2h0aMXO`2g0aMXO`2g0aMYO_2f0bMYO_2f0bMZO^2e0cMZO^2e0cMZO^2d0eM[O[2a0iM^OX2>lMAU2;oMEQ26TNIm13WNLj12XNNh1O[N0f1N\\N2d1L_N_O_KIS6LjI>f4EfKOk5MnI8d4EkK5c5MRJ4_5M`0NUJ0_5O=0XJL^53:0=ND1>MC2l6N2000000000000000000000000000000000000000000000000000Pge11lXZN4L4oIHiNoL^5;aLS3oMdL_5:_LT3RNcL^5:]LU3TNcL\\4DWLe0X1V3SNcL\\4CYLe0U1W3VNbLZ4B\\Ld0T1Z3TNbLZ4A]Lc0T1\\3SNcLZ4_O_La0S1`3RNcLY4]ObL?S1b3QNfLV4ZOdL?T1d3oMiLT4UOgL?Q1P5S2aJjLa0o0V5o1^JPM=3F_Oi5S3[JXM80`6_2^I_M23e6S2`IgMM7g6h1bInMI:i6^1eIVNCP7i0gIfN[Oa0R7?jInNUOc0X73iIWOPOg0_;O1O00100N2O1O1N2O001N2O1O1O1N101O1N4MW\\a4"}, "label": "a bike in the forest"}]} {"id": 473182, "image": "COCO_train2014_000000473182.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"part of a bike on the ground surrounded by trees\"."}], "task": "refering", "answer_template": "The \"part of a bike on the ground surrounded by trees\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 4, 5, 6, 7, 8, 9, 25, 26, 27, 28, 29, 30, 31, 32, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 70, 71, 72, 73, 74, 75, 76, 77, 78, 93, 94, 95, 96, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 162, 163, 164, 165, 166, 167, 168, 169, 185, 186, 187, 188, 189, 190, 191, 192, 193, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331], "bbox": [0.0, 0.0, 0.455890625, 1.0], "iscrowd": 0, "area": 94869.7276, "rle": {"size": [428, 640], "counts": "P;\\2j:6N2O1N2O1N2O1N2O1N2O1N2N2O1N2N2O1N2N2N2O1N2N2O1lJjMPOX2a0\\NXOf1c0bNYO_1f0dNVO^1i0cNUO_1i0cNVO^1i0dNTO^1k0cNSO_1k0cNSO_1l0bNSO_1k0aNVO`1i0_NWOc1g0]NZOd1d0\\N]Oe1b0[N^Of1`0ZNAg1=YNDh1;WNFj17WNIk13XNMi10ZNOg1M]N2d1M]N2d1L^N3c1L^NXO`MXNS4^2`NWO]2h0dMWO]2h0dMVO^2i0cMVO^2i0bMVO`2i0aMVO`2h0aMXO`2g0aMXO`2g0aMYO_2f0bMYO_2f0bMZO^2e0cMZO^2e0cMZO^2d0eM[O[2a0iM^OX2>lMAU2;oMEQ26TNIm13WNLj12XNNh1O[N0f1N\\N2d1L_N_O_KIS6LjI>f4EfKOk5MnI8d4EkK5c5MRJ4_5M`0NUJ0_5O=0XJL^53:0=ND1>MC2l6N2000000000000000000000000000000000000000000000000000Pge11lXZN4L4oIHiNoL^5;aLS3oMdL_5:_LT3RNcL^5:]LU3TNcL\\4DWLe0X1V3SNcL\\4CYLe0U1W3VNbLZ4B\\Ld0T1Z3TNbLZ4A]Lc0T1\\3SNcLZ4_O_La0S1`3RNcLY4]ObL?S1b3QNfLV4ZOdL?T1d3oMiLT4UOgL?Q1P5S2aJjLa0o0V5o1^JPM=3F_Oi5S3[JXM80`6_2^I_M23e6S2`IgMM7g6h1bInMI:i6^1eIVNCP7i0gIfN[Oa0R7?jInNUOc0X73iIWOPOg0_;O1O00100N2O1O1N2O001N2O1O1O1N101O1N4MW\\a4"}, "label": "part of a bike on the ground surrounded by trees"}]} {"id": 145503, "image": "COCO_train2014_000000145503.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a blue tank top\"."}], "task": "refering", "answer_template": "The \"a woman in a blue tank top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 28, 29, 30, 46, 47, 48, 64, 65, 66, 67, 80, 81, 82, 83, 84, 85, 86, 99, 100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 139, 140, 157, 158, 176], "bbox": [0.46998, 0.05418666666666667, 0.83424, 0.7185866666666667], "iscrowd": 0, "area": 19279.032750000006, "rle": {"size": [375, 500], "counts": "PVf21e;2N2N2N2N2M4M2N2N2N2N2N1O2O00000010O00O10004L8H3M3M2N3M2N3M2N2N2O2N1O100O1O1O1O1O1O10N1O100O1O01O00O1O01K4SMjMWLY2h3kMQLY2n3kMkKY2T4lMeKW2Y4nMWJBc0b2T5PNVJAd0^2W5SNRJBd0[2[5VNlICg0V2^5YNhICi0V2]5YNgIDj0U2]5PO`JS1^5oN`JS1^5oN`JR1_5PO^JQ1b5QO\\Jo0d5RO[Jn0d5UOYJl0g5VOWJj0i5XOUJh0k5ZORJf0o5ZOQJf0o5[OPJe0P6[OoIf0Q6[OnIe0R6[OnIe0R6`200O10000000000000O10000000O11O1O1O1O001O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O2M5L4L4L4Lj0VO8H6I101O001O1O001O0011O2M3N1N3N2M2O2N2M3N1N3N2O0O2O1O0O3N4L4L4M2MN2N2N3M2N2N3L3N3M3M4L3M3M3M4K4D in this image.", "objects": [{"patches": [10, 11, 28, 29, 30, 46, 47, 48, 64, 65, 66, 67, 80, 81, 82, 83, 84, 85, 86, 99, 100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 139, 140, 157, 158, 176], "bbox": [0.46998, 0.05418666666666667, 0.83424, 0.7185866666666667], "iscrowd": 0, "area": 19279.032750000006, "rle": {"size": [375, 500], "counts": "PVf21e;2N2N2N2N2M4M2N2N2N2N2N1O2O00000010O00O10004L8H3M3M2N3M2N3M2N2N2O2N1O100O1O1O1O1O1O10N1O100O1O01O00O1O01K4SMjMWLY2h3kMQLY2n3kMkKY2T4lMeKW2Y4nMWJBc0b2T5PNVJAd0^2W5SNRJBd0[2[5VNlICg0V2^5YNhICi0V2]5YNgIDj0U2]5PO`JS1^5oN`JS1^5oN`JR1_5PO^JQ1b5QO\\Jo0d5RO[Jn0d5UOYJl0g5VOWJj0i5XOUJh0k5ZORJf0o5ZOQJf0o5[OPJe0P6[OoIf0Q6[OnIe0R6[OnIe0R6`200O10000000000000O10000000O11O1O1O1O001O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O2M5L4L4L4Lj0VO8H6I101O001O1O001O0011O2M3N1N3N2M2O2N2M3N1N3N2O0O2O1O0O3N4L4L4M2MN2N2N3M2N2N3L3N3M3M4L3M3M3M4K4D in this image.", "objects": [{"patches": [24, 25, 26, 27, 28, 47, 48, 49, 50, 51, 70, 71, 72, 73, 74, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143], "bbox": [0.06975, 0.08521172638436482, 0.24896875000000002, 0.6333224755700326], "iscrowd": 0, "area": 17362.24835, "rle": {"size": [307, 640], "counts": "ld=`0a8d0]Ob0^Ob0]Oc0^Ob0]Oc0^Ob0G9L4M3O1O1O0O10000000000O10000000000O10000000000O100000000001O0000000000000000000000001O000000000000000000000000001O0000000000000000000000001O000000000000000000000000001O0000000000001O001O1N2N2N3Mc0]Oc0]Oc0]Oc0]Od0\\Oc0]Oc0]O in this image.", "objects": [{"patches": [24, 25, 26, 27, 28, 47, 48, 49, 50, 51, 70, 71, 72, 73, 74, 93, 94, 95, 96, 97, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143], "bbox": [0.06975, 0.08521172638436482, 0.24896875000000002, 0.6333224755700326], "iscrowd": 0, "area": 17362.24835, "rle": {"size": [307, 640], "counts": "ld=`0a8d0]Ob0^Ob0]Oc0^Ob0]Oc0^Ob0G9L4M3O1O1O0O10000000000O10000000000O10000000000O100000000001O0000000000000000000000001O000000000000000000000000001O0000000000000000000000001O000000000000000000000000001O0000000000001O001O1N2N2N3Mc0]Oc0]Oc0]Oc0]Od0\\Oc0]Oc0]O in this image.", "objects": [{"patches": [17, 18, 19, 20, 39, 40, 41, 42, 43, 44, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 155, 156, 157, 158], "bbox": [0.687734375, 0.020228013029315962, 0.9787812499999999, 0.6247231270358307], "iscrowd": 0, "area": 26269.380550000005, "rle": {"size": [307, 640], "counts": "YPT43[9:G9F9H9F:F:G9H5L4M2M3M4M2M4L3N3L3M4M2M3M4M2M4L3O2M2N3M2O1N2N1O2O1N1O2N101N1O2N1O2O0O2N2N101N1O2O0O2O001O0O2O1O001N101O001N101O1O0O10001O0O1000000O2O0000000O101O000000000O2O00000000001O0O100000001O00O01000O101O000O2O000O2O000O2O000O2O000O101O0O101O0O101O000O2O000O2O0O100O2N2N2N1O2O1N2N1O2N2O1N1O2N2N2O1N1O2N2N2O0O3M2N2N3N1N3M2N2N3M2M3M4L3M3M4L3M3M4L5K5K5K6J5K5K5I7F;D;FhS4"}, "label": "a round white clock with black hands"}]} {"id": 243823, "image": "COCO_train2014_000000243823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white clock\"."}], "task": "refering", "answer_template": "The \"white clock\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [17, 18, 19, 20, 39, 40, 41, 42, 43, 44, 62, 63, 64, 65, 66, 67, 68, 84, 85, 86, 87, 88, 89, 90, 91, 108, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 155, 156, 157, 158], "bbox": [0.687734375, 0.020228013029315962, 0.9787812499999999, 0.6247231270358307], "iscrowd": 0, "area": 26269.380550000005, "rle": {"size": [307, 640], "counts": "YPT43[9:G9F9H9F:F:G9H5L4M2M3M4M2M4L3N3L3M4M2M3M4M2M4L3O2M2N3M2O1N2N1O2O1N1O2N101N1O2N1O2O0O2N2N101N1O2O0O2O001O0O2O1O001N101O001N101O1O0O10001O0O1000000O2O0000000O101O000000000O2O00000000001O0O100000001O00O01000O101O000O2O000O2O000O2O000O2O000O101O0O101O0O101O000O2O000O2O0O100O2N2N2N1O2O1N2N1O2N2O1N1O2N2N2O1N1O2N2N2O0O3M2N2N3N1N3M2N2N3M2M3M4L3M3M4L3M3M4L5K5K5K6J5K5K5I7F;D;FhS4"}, "label": "white clock"}]} {"id": 555120, "image": "COCO_train2014_000000555120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue car behind girls playing\"."}], "task": "refering", "answer_template": "The \"blue car behind girls playing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 148, 149, 150, 151, 152, 153, 154, 161, 162, 163, 164, 165, 166, 167, 172, 173, 174, 177, 178, 184, 185, 186, 187, 188, 189, 190, 194, 195, 196, 197, 200, 201, 207, 208, 209, 210, 211, 212, 213, 223, 224, 230, 231, 232, 233, 234, 235, 236, 247, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 324, 325, 326, 327, 328], "bbox": [0.0, 0.2995550351288056, 0.775375, 0.9797189695550352], "iscrowd": 0, "area": 58246.79009999999, "rle": {"size": [427, 640], "counts": "W6;P=e0[Of0ZOe0[Of0YOg0ZOe0[Of0ZOe0[O2N00000000000O10O1000000000000O10O100000000000O100000O100000O10000000000000O0100000O100O1O010O100O1O100O1O010O100O1O100O010O1O102M5K4M4K5L4K5K5L3L3N0O2N101N1O2O0O2O0O1O2O0O2O0O2N101N1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1000000O1000000000000000000000000000000000000000000000000O100000000000gHbKe5^4ZJcKf5]4YJdKg5\\4XJeKoN0V6[4jJfKoN4R6W4nJfKoN9m5R4SKfKoN=i5n3XKeKnNc0d5i3]KeKmNh0`5e3bKdKmNl0b1SOh1^4hMdKmNQ1X1UOP2W4jMdKmNU1o0YOU2P4nMcKmNZ1e0\\O\\2h3QNcKmNb18\\Of2`3UNbKlN\\3l2S1WNbKkN`3i2P1[NCa1>^NF^1;aNH[1:eNIW17iNMS14lNOQ12nN2m00RO4j0MVO61`0NCXOU1g0h5O1O100M3N2N2M3N2N200000000001O00000O10000000000000000000000000001O0000000000000000000O2O00001O00001O00001O00001O00001O00001O00001O00001O0O101O00001O00001O00001O00001O1O2N2N1O2N2N2N1O2N2QFbNm7`1oGbNP8`1lGdNS8^1iGdNW8^1dGfN[8\\1aGfN_8[1]GiNb8Y1YGjNg8X1UGkNj8W1QGlNo8U1nFmNR9U1iFmNX9U1dFlN]9V1^FlNc9]21O100O1O1O1O1O100O1O1O1O1O010O1O1O1O1O100O1O1O1F:000g0YO001O0000001O00001N203N101N2N2N2O1NO000000O1000000O10O10O10000N2L4L4K5L4K4Mn\\?Iob@=D=O0O2O0O2O000O2O0O2O001N101N10001N101N12O0O2O1N1O2O0O2O1N101O03N1O2M2O2M2O2N2M2O2M2O1O1N101N2O001N2O0O2O001N2O0O2O1O0O2O1O0O2OO0N3M3N1N3M2O2M8H8I7H8H8I1N2N2F;Edik1"}, "label": "blue car behind girls playing"}]} {"id": 555120, "image": "COCO_train2014_000000555120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the blue car in the back\"."}], "task": "refering", "answer_template": "The \"the blue car in the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 102, 103, 104, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 129, 130, 131, 138, 139, 140, 141, 142, 143, 144, 148, 149, 150, 151, 152, 153, 154, 161, 162, 163, 164, 165, 166, 167, 172, 173, 174, 177, 178, 184, 185, 186, 187, 188, 189, 190, 194, 195, 196, 197, 200, 201, 207, 208, 209, 210, 211, 212, 213, 223, 224, 230, 231, 232, 233, 234, 235, 236, 247, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 324, 325, 326, 327, 328], "bbox": [0.0, 0.2995550351288056, 0.775375, 0.9797189695550352], "iscrowd": 0, "area": 58246.79009999999, "rle": {"size": [427, 640], "counts": "W6;P=e0[Of0ZOe0[Of0YOg0ZOe0[Of0ZOe0[O2N00000000000O10O1000000000000O10O100000000000O100000O100000O10000000000000O0100000O100O1O010O100O1O100O1O010O100O1O100O010O1O102M5K4M4K5L4K5K5L3L3N0O2N101N1O2O0O2O0O1O2O0O2O0O2N101N1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1O100O100O1O100O1O100O100O1000000O1000000000000000000000000000000000000000000000000O100000000000gHbKe5^4ZJcKf5]4YJdKg5\\4XJeKoN0V6[4jJfKoN4R6W4nJfKoN9m5R4SKfKoN=i5n3XKeKnNc0d5i3]KeKmNh0`5e3bKdKmNl0b1SOh1^4hMdKmNQ1X1UOP2W4jMdKmNU1o0YOU2P4nMcKmNZ1e0\\O\\2h3QNcKmNb18\\Of2`3UNbKlN\\3l2S1WNbKkN`3i2P1[NCa1>^NF^1;aNH[1:eNIW17iNMS14lNOQ12nN2m00RO4j0MVO61`0NCXOU1g0h5O1O100M3N2N2M3N2N200000000001O00000O10000000000000000000000000001O0000000000000000000O2O00001O00001O00001O00001O00001O00001O00001O00001O0O101O00001O00001O00001O00001O1O2N2N1O2N2N2N1O2N2QFbNm7`1oGbNP8`1lGdNS8^1iGdNW8^1dGfN[8\\1aGfN_8[1]GiNb8Y1YGjNg8X1UGkNj8W1QGlNo8U1nFmNR9U1iFmNX9U1dFlN]9V1^FlNc9]21O100O1O1O1O1O100O1O1O1O1O010O1O1O1O1O100O1O1O1F:000g0YO001O0000001O00001N203N101N2N2N2O1NO000000O1000000O10O10O10000N2L4L4K5L4K4Mn\\?Iob@=D=O0O2O0O2O000O2O0O2O001N101N10001N101N12O0O2O1N1O2O0O2O1N101O03N1O2M2O2M2O2N2M2O2M2O1O1N101N2O001N2O0O2O001N2O0O2O1O0O2O1O0O2OO0N3M3N1N3M2O2M8H8I7H8H8I1N2N2F;Edik1"}, "label": "the blue car in the back"}]} {"id": 555120, "image": "COCO_train2014_000000555120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl with a pink shirt on\"."}], "task": "refering", "answer_template": "The \"a little girl with a pink shirt on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 127, 128, 129, 150, 151, 152, 153, 174, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 290, 291, 292, 293, 313, 314, 315, 316, 337, 338, 339], "bbox": [0.532546875, 0.26447306791569086, 0.765921875, 0.9864402810304449], "iscrowd": 0, "area": 25325.11639999999, "rle": {"size": [427, 640], "counts": "ZZ^4>R3Gc6`0VIHc69\\IJ`68^IK_66aIK]67aIK]67bIJ\\68cIJ[67cIK[67dIJZ68dIKY67fIJX6=aIE]6`0_I@a6b0\\I@b6c0]I\\Ob6g0\\IZOb6i0\\IWOd6l0ZITOe6n0YIROf6Q1XIPOg6Q1YInNg6T1WImNh6U1VIkNj6V1VIjNi6X1UIhNj6Z1VIfNi6Z1XIeNh6\\1XIdNg6\\1[IbNe6_1[IaNd6_1]I`Nb6b1^I^Na6c1\\I_Nd6a1[IaNd6`1ZIbNe6^1YIhNc6Y1[IkNb6U1[IRO`6P1XIZOe6P1jH[OR7b3M3M4L4M3L4L4L4L4L4M3L4L4L4L2N2O2M2N2N2O1N2N2O1N2N2O1N3M2O1N2O14iHTKe5P5PJVKQ6R5`ITKa6k51O1N2O2N1O1O1N2O1N2O00O001O001O002N2N2N1O2N2N2N2N2N2N2N2N1O2N3M3M2N3M3M3M3M3M3M2N3M3M3M3M3M3M4L4L3M4L3M2N2N2N2N2N3M2N2N2N2N2N2N2N3M4B=B?B=G:F>Cg0XOjVn1"}, "label": "a little girl with a pink shirt on"}]} {"id": 555120, "image": "COCO_train2014_000000555120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black girl in a bright pink shirt\"."}], "task": "refering", "answer_template": "The \"a black girl in a bright pink shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 127, 128, 129, 150, 151, 152, 153, 174, 175, 176, 177, 196, 197, 198, 199, 200, 219, 220, 221, 222, 223, 243, 244, 245, 246, 247, 266, 267, 268, 269, 270, 290, 291, 292, 293, 313, 314, 315, 316, 337, 338, 339], "bbox": [0.532546875, 0.26447306791569086, 0.765921875, 0.9864402810304449], "iscrowd": 0, "area": 25325.11639999999, "rle": {"size": [427, 640], "counts": "ZZ^4>R3Gc6`0VIHc69\\IJ`68^IK_66aIK]67aIK]67bIJ\\68cIJ[67cIK[67dIJZ68dIKY67fIJX6=aIE]6`0_I@a6b0\\I@b6c0]I\\Ob6g0\\IZOb6i0\\IWOd6l0ZITOe6n0YIROf6Q1XIPOg6Q1YInNg6T1WImNh6U1VIkNj6V1VIjNi6X1UIhNj6Z1VIfNi6Z1XIeNh6\\1XIdNg6\\1[IbNe6_1[IaNd6_1]I`Nb6b1^I^Na6c1\\I_Nd6a1[IaNd6`1ZIbNe6^1YIhNc6Y1[IkNb6U1[IRO`6P1XIZOe6P1jH[OR7b3M3M4L4M3L4L4L4L4L4M3L4L4L4L2N2O2M2N2N2O1N2N2O1N2N2O1N3M2O1N2O14iHTKe5P5PJVKQ6R5`ITKa6k51O1N2O2N1O1O1N2O1N2O00O001O001O002N2N2N1O2N2N2N2N2N2N2N2N1O2N3M3M2N3M3M3M3M3M3M2N3M3M3M3M3M3M4L4L3M4L3M2N2N2N2N2N3M2N2N2N2N2N2N2N3M4B=B?B=G:F>Cg0XOjVn1"}, "label": "a black girl in a bright pink shirt"}]} {"id": 538737, "image": "COCO_train2014_000000538737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the polar bear is fighting\"."}], "task": "refering", "answer_template": "The \"the polar bear is fighting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 79, 80, 81, 82, 83, 98, 99, 100, 101, 102, 103, 104, 105, 106, 121, 122, 123, 124, 125, 126, 127, 128, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 258, 259, 260, 261, 262, 263, 264, 265, 281, 282, 283, 284, 285, 286, 287, 288, 305, 306, 307, 309, 310, 311, 312, 332, 333, 334], "bbox": [0.23946874999999998, 0.20032786885245904, 0.6340468749999999, 0.9859016393442623], "iscrowd": 0, "area": 55216.15465000001, "rle": {"size": [427, 640], "counts": "STP21X=4M3L3N3M3bNFeE>P:2hE1m9?kEDk9i0PFYOe9U1UFnNb9]1\\FcN[9i1cFUNU9V2kFgMn8d2PG[Mh8o2WGoLb8[3^GbL[8[4I7I7L5L3L4L4M3L5K4M2M4M3L3M4M3L3K6H8H7I8H7I8L4M2N3M2N2N2O100O100O100O100O100O1O2O0O1000001N1000001N1000ZNZLgIf3Z6`L_I_3c6gLWIX3j6nLoHR3Q7YMdHg2]7c11N2O1O1O1O1N101O1O1N2O1O1O1O1N101O1O1O0O100000000000O100000000000000O0100000000000iNW12N2N1O2N2N2N2N2O1O1O1O001O1O2K8H9G;F in this image.", "objects": [{"patches": [83, 106, 107, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 174, 175, 176, 177, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 319, 320, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.546953125, 0.25166276346604216, 0.9570624999999999, 1.0], "iscrowd": 0, "area": 38428.07244999999, "rle": {"size": [427, 640], "counts": "hSb44U=2N3M2N2N3M2O10001N3N4L3M4K5L3M4L4L4L5J6K5K5K5K4H7M3M3M2N3M3M3M3M3M2M4L4Kh0YO4L4K5L4L3L5L4L4L4L4_KWJX1m5bNWJ\\1m5_NVJ_1m5\\NWJa1n5YNVJb1P6YNTJc1R6XNQJd1T6XNoIc1X6XNkId1Z6XNiIi1W6SNlIn1U6nMmIS2S6iMPJX2Q6cMRJ]2o5PMfHM^1S3n5lLjHJ[1[3k5iLdJW3^5fLcJZ3^5dLcJ\\3f7001O000000001O1O1O001O1O1O3M5K4L1O1O2N1O3M4L5K3M1O00000000001O1O1O001O1O1O1O001O=C3M4L3M3M3M3M4L3M3M3M3M4]I^IU6e6dI`I[6i6O1O1O1O2N1O1O1O1O1O;E6J2N2N2N2N3M2N2N1O1O00001O00001O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O6J2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O2N1O2N1O2N1O2N1O1O001O0000001O0000001O0000001O001O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O6J6J2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N2N\\X;"}, "label": "polar bear on right"}]} {"id": 538737, "image": "COCO_train2014_000000538737.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the polar bear on the right\"."}], "task": "refering", "answer_template": "The \"the polar bear on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 106, 107, 127, 128, 129, 130, 131, 150, 151, 152, 153, 154, 174, 175, 176, 177, 197, 198, 199, 200, 201, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 267, 268, 269, 270, 271, 272, 290, 291, 292, 293, 294, 295, 296, 313, 314, 315, 316, 317, 318, 319, 320, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.546953125, 0.25166276346604216, 0.9570624999999999, 1.0], "iscrowd": 0, "area": 38428.07244999999, "rle": {"size": [427, 640], "counts": "hSb44U=2N3M2N2N3M2O10001N3N4L3M4K5L3M4L4L4L5J6K5K5K5K4H7M3M3M2N3M3M3M3M3M2M4L4Kh0YO4L4K5L4L3L5L4L4L4L4_KWJX1m5bNWJ\\1m5_NVJ_1m5\\NWJa1n5YNVJb1P6YNTJc1R6XNQJd1T6XNoIc1X6XNkId1Z6XNiIi1W6SNlIn1U6nMmIS2S6iMPJX2Q6cMRJ]2o5PMfHM^1S3n5lLjHJ[1[3k5iLdJW3^5fLcJZ3^5dLcJ\\3f7001O000000001O1O1O001O1O1O3M5K4L1O1O2N1O3M4L5K3M1O00000000001O1O1O001O1O1O1O001O=C3M4L3M3M3M3M4L3M3M3M3M4]I^IU6e6dI`I[6i6O1O1O1O2N1O1O1O1O1O;E6J2N2N2N2N3M2N2N1O1O00001O00001O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O6J2N1O1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O2N1O2N1O2N1O2N1O1O001O0000001O0000001O0000001O001O2N1O1O2N1O1O2N1O2N1O1O2N1O2N1O1O2N1O1O2N1O2N1O1O6J6J2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N2N\\X;"}, "label": "the polar bear on the right"}]} {"id": 55412, "image": "COCO_train2014_000000055412.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person wearing a red helmet and riding a blue scooter\"."}], "task": "refering", "answer_template": "The \"a person wearing a red helmet and riding a blue scooter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 13, 14, 15, 25, 26, 27, 28, 37, 38, 39, 40, 50, 51, 62, 63, 74, 75, 87], "bbox": [0.11471641791044776, 0.017, 0.3918805970149254, 0.40144], "iscrowd": 0, "area": 6795.953449999997, "rle": {"size": [500, 335], "counts": "lcb01a?2N3M2fAIU=:hBIT=:kBHR=;lBFS=;lBGR=;kBHR=:mBHZVE[Oo:e0P20001O001O001O001O001N101N101N101N1O2O0O2M2O2N1N3N1N3Nl`S3"}, "label": "a person wearing a red helmet and riding a blue scooter"}]} {"id": 55412, "image": "COCO_train2014_000000055412.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing a red helmet\"."}], "task": "refering", "answer_template": "The \"man wearing a red helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 13, 14, 15, 25, 26, 27, 28, 37, 38, 39, 40, 50, 51, 62, 63, 74, 75, 87], "bbox": [0.11471641791044776, 0.017, 0.3918805970149254, 0.40144], "iscrowd": 0, "area": 6795.953449999997, "rle": {"size": [500, 335], "counts": "lcb01a?2N3M2fAIU=:hBIT=:kBHR=;lBFS=;lBGR=;kBHR=:mBHZVE[Oo:e0P20001O001O001O001O001N101N101N101N1O2O0O2M2O2N1N3N1N3Nl`S3"}, "label": "man wearing a red helmet"}]} {"id": 55412, "image": "COCO_train2014_000000055412.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blurry pick of a man with yellow on the back of a bike\"."}], "task": "refering", "answer_template": "The \"a blurry pick of a man with yellow on the back of a bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 12, 13, 24, 25, 36, 37, 48, 49, 50, 61, 62, 73, 74], "bbox": [0.0, 0.04496, 0.19701492537313434, 0.38051999999999997], "iscrowd": 0, "area": 4530.5699, "rle": {"size": [500, 335], "counts": "^1?T?n0cNU1ZO100O100O1O2O0O1O101N101O001O001O001O001O001O001O001O001OaMcCg1c=C4L4K10O00UOSB0l=0UBOl=OVB1i=NYB1g=OZB1e=N]B1c=N_B2a=M`B3_=LcB5[=JgB7W=HkB9S=FoB:g0^Of:6eDP2Z;nMiDQ2_ in this image.", "objects": [{"patches": [191, 192, 193, 194, 206, 207, 208, 209, 219, 220, 221, 222, 223, 224, 234, 235, 236, 237, 238, 239, 249, 250, 251, 252, 253, 254, 264, 265, 266, 267, 268, 269, 280, 281, 282, 283, 284, 295, 298, 299, 310, 313, 314], "bbox": [0.614929906542056, 0.53034375, 0.9778504672897196, 0.9078593749999999], "iscrowd": 0, "area": 23271.53704999999, "rle": {"size": [640, 428], "counts": "QjT5`0na0d1I7N2N2N2N2N2N1O2M3N2N2N2N2N2N2N2N2N2N2N2N=Ca0_Ob0^Ob0]O:G00000000000000000000000000bMRBjNn=V1cCYM\\ in this image.", "objects": [{"patches": [268, 289, 290, 291, 293, 294, 295, 311, 312, 313, 314, 315, 316, 317, 333, 334, 335, 336, 337, 338, 339, 357, 358], "bbox": [0.1343954248366013, 0.5769607843137255, 0.4452450980392157, 0.7435294117647059], "iscrowd": 0, "area": 10216.146750000002, "rle": {"size": [612, 612], "counts": "W\\a11Qc02N1O2O1N2O1O1O1O010O1j]OI^a08a^O1Va00i^O4Sa0Ml^O5Ra0Kn^O6Qa0Kn^O7Qa0Io^O7Pa0JP_O7o`0IQ_O8m`0HS_O9l`0HT_O9k`0GU_O9j`0HV_O9i`0GW_O:g`0GX_O:h`0EY_O\\`0Bd_O?Z`0Ce_O=[`0Ce_O>[`0Ae_O?\\`0@d_Oa0\\`0_Oc_Oa0^`0^Ob_Oc0^`0\\Oa_Of0_`0YOa_Og0``0YO__Oh0a`0WO^_Oj0c`0UO]_Ol0b`0UO]_Ol0b`0TO]_On0b`0SO]_On0c`0RO\\_On0d`0SOZ_Oo0e`0ROZ_Oo0e`0i0O1O1O00100O1O1O001O10O5L5J5L5J5LO00001N101O001O000O01000000O1000000O10O1000O1000000O101O000O101O000O10O1000O10O1000O10O10O10O01000O01000O0100O1000O0100O10000O01000O100O01000O1000O0100O10O10O10O01000O01000O0100O10O10O10O10O10O01000O0100O010O0N3M3M2N3L4M2N3M3L3N3L4GYkZ6"}, "label": "the blue letters on a colorful child ' s blanket on the bottom bunk"}]} {"id": 194685, "image": "COCO_train2014_000000194685.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person and their blanket on the bottom bunk\"."}], "task": "refering", "answer_template": "The \"the person and their blanket on the bottom bunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [268, 289, 290, 291, 293, 294, 295, 311, 312, 313, 314, 315, 316, 317, 333, 334, 335, 336, 337, 338, 339, 357, 358], "bbox": [0.1343954248366013, 0.5769607843137255, 0.4452450980392157, 0.7435294117647059], "iscrowd": 0, "area": 10216.146750000002, "rle": {"size": [612, 612], "counts": "W\\a11Qc02N1O2O1N2O1O1O1O010O1j]OI^a08a^O1Va00i^O4Sa0Ml^O5Ra0Kn^O6Qa0Kn^O7Qa0Io^O7Pa0JP_O7o`0IQ_O8m`0HS_O9l`0HT_O9k`0GU_O9j`0HV_O9i`0GW_O:g`0GX_O:h`0EY_O\\`0Bd_O?Z`0Ce_O=[`0Ce_O>[`0Ae_O?\\`0@d_Oa0\\`0_Oc_Oa0^`0^Ob_Oc0^`0\\Oa_Of0_`0YOa_Og0``0YO__Oh0a`0WO^_Oj0c`0UO]_Ol0b`0UO]_Ol0b`0TO]_On0b`0SO]_On0c`0RO\\_On0d`0SOZ_Oo0e`0ROZ_Oo0e`0i0O1O1O00100O1O1O001O10O5L5J5L5J5LO00001N101O001O000O01000000O1000000O10O1000O1000000O101O000O101O000O10O1000O10O1000O10O10O10O01000O01000O0100O1000O0100O10000O01000O100O01000O1000O0100O10O10O10O01000O01000O0100O10O10O10O10O10O01000O0100O010O0N3M3M2N3L4M2N3M3L3N3L4GYkZ6"}, "label": "the person and their blanket on the bottom bunk"}]} {"id": 268428, "image": "COCO_train2014_000000268428.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"front most motorcycle\"."}], "task": "refering", "answer_template": "The \"front most motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 74, 75, 76, 77, 96, 97, 98, 99, 100, 110, 111, 112, 119, 120, 121, 122, 123, 124, 125, 133, 134, 135, 136, 137, 142, 143, 144, 145, 146, 147, 148, 149, 155, 156, 157, 158, 159, 160, 165, 166, 167, 168, 169, 170, 171, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 283, 284, 285, 286, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 314, 315, 316, 317, 318, 319, 338, 339, 340, 341], "bbox": [0.00453125, 0.1687906976744186, 0.9934687499999999, 0.9732790697674418], "iscrowd": 0, "area": 109524.52640000002, "rle": {"size": [430, 640], "counts": "]c17iA?B>A?B>A?N2N2O1N101N2O1O1N2O1N2O1N2O1N10001N100O101N10000O2O0O100O101O0O100O2O0O10001N100O101N100O101O0O100O2O0000001O000O10001O0000001O0000001O0000001O0O10001O000000O100O100O100O100O100O010O100O100O100O100O010O100O010O100O010O100OcL^KoMc4Q2_K`L2[O^4U4dKVL7CT4W4iKlKX4@nK:T4CSL7g8J5L5J6J6J6J6K5JZ[1"}, "label": "front most motorcycle"}]} {"id": 268428, "image": "COCO_train2014_000000268428.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a motor cycle in front carrying a man and a lady\"."}], "task": "refering", "answer_template": "The \"a motor cycle in front carrying a man and a lady\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [52, 53, 74, 75, 76, 77, 96, 97, 98, 99, 100, 110, 111, 112, 119, 120, 121, 122, 123, 124, 125, 133, 134, 135, 136, 137, 142, 143, 144, 145, 146, 147, 148, 149, 155, 156, 157, 158, 159, 160, 165, 166, 167, 168, 169, 170, 171, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 283, 284, 285, 286, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 314, 315, 316, 317, 318, 319, 338, 339, 340, 341], "bbox": [0.00453125, 0.1687906976744186, 0.9934687499999999, 0.9732790697674418], "iscrowd": 0, "area": 109524.52640000002, "rle": {"size": [430, 640], "counts": "]c17iA?B>A?B>A?N2N2O1N101N2O1O1N2O1N2O1N2O1N10001N100O101N10000O2O0O100O101O0O100O2O0O10001N100O101N100O101O0O100O2O0000001O000O10001O0000001O0000001O0000001O0O10001O000000O100O100O100O100O100O010O100O100O100O100O010O100O010O100O010O100OcL^KoMc4Q2_K`L2[O^4U4dKVL7CT4W4iKlKX4@nK:T4CSL7g8J5L5J6J6J6J6K5JZ[1"}, "label": "a motor cycle in front carrying a man and a lady"}]} {"id": 268428, "image": "COCO_train2014_000000268428.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue motercycle\"."}], "task": "refering", "answer_template": "The \"a blue motercycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 23, 24, 25, 26, 27, 46, 47, 48, 49, 50, 69, 70, 71, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187], "bbox": [0.0, 0.03504651162790698, 0.19840625, 0.5965581395348837], "iscrowd": 0, "area": 21351.80455, "rle": {"size": [430, 640], "counts": "X3V1V<^2cMW1hN010_NYK[Jg4g5\\KTJe4l5\\KSJc4o5^KnIc4S6]KlIb4V6^KhIc4Y6]KfIb4\\6_KaIb4`6^K_Ib4`6`K^I`4b6cK[I^4d6dK[I[4f6fKXI[4X6UKdI`03Z4Z6XKbI?1Z4]6XKcI=OZ4_6[KaI;N[4a6\\K`I:MZ4c6^K_I8L\\4d6^K_I6L]4d6_K`I3J`4e6_K`I2Ha4h6^K_I1Hb4j6]K]IW5K\\Ja6=cI_5^6`JcI_5^6aJaI_5`6aJ_I^5c6bJ\\I^5e6bJYI_5h6aJWI_5j6a02O1O1eIUIQ6U7O1O1O1O2N1O1O1OO1O100O1O100O1O100O1O100O1O100O1O100O1O100O10000O1000cKTIX2l6hM[IQ2e6nM^IP2b6PN`In1`6RNaIl1`6TNbIj1^6UNeIi1[6WNeIi1[6WNfIh1Z6WNgIi1Y6WNhIh1X6XNhIh1X6XNhIh1X6WNjIh1V6XNjIh1W6WNjIh1V6XNjIh1W6VNjIj1V6VNjIj1W6UNiIk1W6UNiIk1X6SNiIm1W6SNiIm1X6RNhIn1X6RNiIm1X6RNhIn1X6QNiIo1X6PNhIP2X6PNhIP2X6PNhIP2Y6oMgIQ2Y6nMhIR2Y6mMgIS2Y6mMgIS2Z6lMfIT2Z6kMgIU2Z6jMfIV2Z6jMfIV2[6iMeI`0QOHZ7HeI=WOHU7IfI\\OnN7 in this image.", "objects": [{"patches": [0, 1, 23, 24, 25, 26, 27, 46, 47, 48, 49, 50, 69, 70, 71, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 138, 139, 140, 141, 142, 161, 162, 163, 164, 165, 184, 185, 186, 187], "bbox": [0.0, 0.03504651162790698, 0.19840625, 0.5965581395348837], "iscrowd": 0, "area": 21351.80455, "rle": {"size": [430, 640], "counts": "X3V1V<^2cMW1hN010_NYK[Jg4g5\\KTJe4l5\\KSJc4o5^KnIc4S6]KlIb4V6^KhIc4Y6]KfIb4\\6_KaIb4`6^K_Ib4`6`K^I`4b6cK[I^4d6dK[I[4f6fKXI[4X6UKdI`03Z4Z6XKbI?1Z4]6XKcI=OZ4_6[KaI;N[4a6\\K`I:MZ4c6^K_I8L\\4d6^K_I6L]4d6_K`I3J`4e6_K`I2Ha4h6^K_I1Hb4j6]K]IW5K\\Ja6=cI_5^6`JcI_5^6aJaI_5`6aJ_I^5c6bJ\\I^5e6bJYI_5h6aJWI_5j6a02O1O1eIUIQ6U7O1O1O1O2N1O1O1OO1O100O1O100O1O100O1O100O1O100O1O100O1O100O10000O1000cKTIX2l6hM[IQ2e6nM^IP2b6PN`In1`6RNaIl1`6TNbIj1^6UNeIi1[6WNeIi1[6WNfIh1Z6WNgIi1Y6WNhIh1X6XNhIh1X6XNhIh1X6WNjIh1V6XNjIh1W6WNjIh1V6XNjIh1W6VNjIj1V6VNjIj1W6UNiIk1W6UNiIk1X6SNiIm1W6SNiIm1X6RNhIn1X6RNiIm1X6RNhIn1X6QNiIo1X6PNhIP2X6PNhIP2X6PNhIP2Y6oMgIQ2Y6nMhIR2Y6mMgIS2Y6mMgIS2Z6lMfIT2Z6kMgIU2Z6jMfIV2Z6jMfIV2[6iMeI`0QOHZ7HeI=WOHU7IfI\\OnN7 in this image.", "objects": [{"patches": [93, 94, 95, 96, 97, 98, 99, 100, 116, 117, 118, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130], "bbox": [0.064671875, 0.2894172494172494, 0.66084375, 0.38587412587412584], "iscrowd": 0, "area": 7659.6254500000005, "rle": {"size": [429, 640], "counts": "SZa0`0m<000O100O1000O010000O100O1000O010O10O010O0100O010O10O010O1000O01000O0100O01000O10O10000O1000000000000000000000O10000000000000000000O1000000000000000000000O10000000000000000000000000O100001O000000001N10000000001O000000010O000000001O000000000010O00000001O000000001O0001O01O0000001O0000001O0001O01O0000001O0000010O0000010O0000010O000010O0000010O0001O01O0001O01O01O00010O0000010O000001O0000001O0000001O0000001O000000000000000O101O0000000000000000001O0000000000000000001O000000000O100000001O0000000000000000001O000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000O10000000000000000000000O10000000000000000000O1000O0100O10000O10000O10000O01000O100O10000O10O10O10000O10000N2L3Nlhj2"}, "label": "long streamer on kite in the air"}]} {"id": 497807, "image": "COCO_train2014_000000497807.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tail of a kite\"."}], "task": "refering", "answer_template": "The \"the tail of a kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [93, 94, 95, 96, 97, 98, 99, 100, 116, 117, 118, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130], "bbox": [0.064671875, 0.2894172494172494, 0.66084375, 0.38587412587412584], "iscrowd": 0, "area": 7659.6254500000005, "rle": {"size": [429, 640], "counts": "SZa0`0m<000O100O1000O010000O100O1000O010O10O010O0100O010O10O010O1000O01000O0100O01000O10O10000O1000000000000000000000O10000000000000000000O1000000000000000000000O10000000000000000000000000O100001O000000001N10000000001O000000010O000000001O000000000010O00000001O000000001O0001O01O0000001O0000001O0001O01O0000001O0000010O0000010O0000010O000010O0000010O0001O01O0001O01O01O00010O0000010O000001O0000001O0000001O0000001O000000000000000O101O0000000000000000001O0000000000000000001O000000000O100000001O0000000000000000001O000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000000000000O10000000000000000000000O10000000000000000000O1000O0100O10000O10000O10000O01000O100O10000O10O10O10000O10000N2L3Nlhj2"}, "label": "the tail of a kite"}]} {"id": 358543, "image": "COCO_train2014_000000358543.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman on a subway\"."}], "task": "refering", "answer_template": "The \"a woman on a subway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 30, 31, 32, 45, 46, 47, 48, 49, 62, 63, 64, 65, 66, 78, 79, 80, 81, 82, 83, 94, 95, 96, 97, 98, 99, 100, 111, 112, 113, 114, 115, 116, 117, 128, 129, 130, 131, 132, 133, 134, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 177, 178, 179, 180, 181, 182, 183, 184, 185, 194, 195, 196, 197, 198, 199, 200, 201, 202, 211, 212, 213, 214, 215, 216, 217, 218, 219, 229, 230, 231, 232, 233, 234, 235, 236, 237, 246, 247, 248, 249, 250, 251, 252, 253, 254, 263, 264, 265, 266, 267, 268, 269, 270, 281, 282, 283, 284, 285, 298, 299, 300, 301, 302, 315, 316, 317, 318, 319, 332, 333, 334, 335, 336, 350, 351, 352, 367, 368], "bbox": [0.43943750000000004, 0.014109374999999999, 0.9661458333333334, 0.935109375], "iscrowd": 0, "area": 85757.66515, "rle": {"size": [640, 480], "counts": "VVT43jc08I7I7I5K3M2N2M4M2N3L3N3L3N2M4L3N3L3N3L3M4M2M3N3L3N3Le0[O6K5J7J5J7J6I7I7J6I7J6I7Jj0UO3M4M3L3N3L3N3L3M4M3L3N3L3N3L3_FiHV6[7ZHSJe7P6UHTJi7o5RHUJm7n5mGVJQ8m5jGWJU8j8N2N2O1O100000000000000000001O000_IVDm4j;PK[Dn4e;oJ`Do4`;nJfDo4Z;PKiDn4W;RKkDl4U;TKmDj4S;VKPEg4P;YKREe4n:[KTEc4l:]KUEb4k:^KVEa4j:_KWE`4i:`KXE_4h:aKYE^4g:bKZE]4f:cK[E\\4U:hJhDl0T1[4S:VLnEi3Q:XLPFg3o9ZLRFe3m9\\LTFc3k9[LYFd3f9[L]Fd3b9[LaFd3^9[LeFb3]9\\LfF`3\\9^LgF`3Z9_LiF^3X9aLjF\\3X9cLjF[3X9bLlFZ3[9`LgF^3^9XLjFd3\\9PLmFn3X9gKPGW4^ in this image.", "objects": [{"patches": [13, 14, 30, 31, 32, 45, 46, 47, 48, 49, 62, 63, 64, 65, 66, 78, 79, 80, 81, 82, 83, 94, 95, 96, 97, 98, 99, 100, 111, 112, 113, 114, 115, 116, 117, 128, 129, 130, 131, 132, 133, 134, 144, 145, 146, 147, 148, 149, 150, 161, 162, 163, 164, 165, 166, 167, 168, 177, 178, 179, 180, 181, 182, 183, 184, 185, 194, 195, 196, 197, 198, 199, 200, 201, 202, 211, 212, 213, 214, 215, 216, 217, 218, 219, 229, 230, 231, 232, 233, 234, 235, 236, 237, 246, 247, 248, 249, 250, 251, 252, 253, 254, 263, 264, 265, 266, 267, 268, 269, 270, 281, 282, 283, 284, 285, 298, 299, 300, 301, 302, 315, 316, 317, 318, 319, 332, 333, 334, 335, 336, 350, 351, 352, 367, 368], "bbox": [0.43943750000000004, 0.014109374999999999, 0.9661458333333334, 0.935109375], "iscrowd": 0, "area": 85757.66515, "rle": {"size": [640, 480], "counts": "VVT43jc08I7I7I5K3M2N2M4M2N3L3N3L3N2M4L3N3L3N3L3M4M2M3N3L3N3Le0[O6K5J7J5J7J6I7I7J6I7J6I7Jj0UO3M4M3L3N3L3N3L3M4M3L3N3L3N3L3_FiHV6[7ZHSJe7P6UHTJi7o5RHUJm7n5mGVJQ8m5jGWJU8j8N2N2O1O100000000000000000001O000_IVDm4j;PK[Dn4e;oJ`Do4`;nJfDo4Z;PKiDn4W;RKkDl4U;TKmDj4S;VKPEg4P;YKREe4n:[KTEc4l:]KUEb4k:^KVEa4j:_KWE`4i:`KXE_4h:aKYE^4g:bKZE]4f:cK[E\\4U:hJhDl0T1[4S:VLnEi3Q:XLPFg3o9ZLRFe3m9\\LTFc3k9[LYFd3f9[L]Fd3b9[LaFd3^9[LeFb3]9\\LfF`3\\9^LgF`3Z9_LiF^3X9aLjF\\3X9cLjF[3X9bLlFZ3[9`LgF^3^9XLjFd3\\9PLmFn3X9gKPGW4^ in this image.", "objects": [{"patches": [4, 5, 6, 7, 20, 21, 22, 23, 24, 37, 38, 39, 40, 41, 53, 54, 55, 56, 57, 58, 69, 70, 71, 72, 73, 74, 75, 86, 87, 88, 89, 90, 91, 92, 103, 104, 105, 106, 107, 108, 109, 120, 121, 122, 123, 124, 125, 126, 137, 138, 139, 140, 141, 142, 143, 144, 154, 155, 156, 157, 158, 159, 160, 161, 171, 172, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 195, 206, 207, 208, 209, 210, 211, 212, 223, 224, 225, 226, 227, 228, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 275, 276, 277, 278, 279, 292, 293, 294, 295, 309, 310, 311, 312, 326, 327, 328, 329, 343, 344, 345, 346, 347, 360, 361, 362, 363, 364, 377, 378, 379, 380, 381, 382], "bbox": [0.06320833333333334, 0.003953125, 0.499625, 0.98815625], "iscrowd": 0, "area": 88039.0673, "rle": {"size": [640, 480], "counts": "Wmb01lc0:G9F:G8G5L3L4M3M3L4M3M8I;Em:;_DE];S1oCmNm;b1gC_NT in this image.", "objects": [{"patches": [4, 5, 6, 7, 20, 21, 22, 23, 24, 37, 38, 39, 40, 41, 53, 54, 55, 56, 57, 58, 69, 70, 71, 72, 73, 74, 75, 86, 87, 88, 89, 90, 91, 92, 103, 104, 105, 106, 107, 108, 109, 120, 121, 122, 123, 124, 125, 126, 137, 138, 139, 140, 141, 142, 143, 144, 154, 155, 156, 157, 158, 159, 160, 161, 171, 172, 173, 174, 175, 176, 177, 178, 188, 189, 190, 191, 192, 193, 194, 195, 206, 207, 208, 209, 210, 211, 212, 223, 224, 225, 226, 227, 228, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 275, 276, 277, 278, 279, 292, 293, 294, 295, 309, 310, 311, 312, 326, 327, 328, 329, 343, 344, 345, 346, 347, 360, 361, 362, 363, 364, 377, 378, 379, 380, 381, 382], "bbox": [0.06320833333333334, 0.003953125, 0.499625, 0.98815625], "iscrowd": 0, "area": 88039.0673, "rle": {"size": [640, 480], "counts": "Wmb01lc0:G9F:G8G5L3L4M3M3L4M3M8I;Em:;_DE];S1oCmNm;b1gC_NT in this image.", "objects": [{"patches": [66, 67, 68, 84, 85, 86, 87, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 137, 138, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 227, 232], "bbox": [0.6203, 0.27168367346938777, 0.9708, 0.8786479591836734], "iscrowd": 0, "area": 32041.480599999995, "rle": {"size": [392, 500], "counts": "Ulf31W<3L3M4M2M2N2O1N3MQ2PNc0\\O?A5J5J7J5K2M3N2N2M3N2SN[KTLh4k3^KnKd4P4cKiK^4W4iKbKX4]4nK]KR4b4ULXKl3g4ZLSKf3m4`LmJa3Q5fLjJY3V5l11O1O1O1N2O1O1O1O1O00100001O00000000000O100000000000000000000O1000000000001O00000O10000000000000000000000000000001O0000000001O01O0000000000001O0000000000001O0000000000001O1O2O0O1O1O1O2N1O1O1O2N1O1O1O15L2M1O1O1OO2N1O1O1O2N1O1O1O1O2gJlIZ4W6cKiI]4Y6aKgI^4]6_KdI_4_6_KaIa4b6\\K^Ic4e6[K\\Ic4h6XKZIg4i6UKYIk4S7N4L3L5J5K6K4K6M2N3L3N3M3L3N3L3Nj0VOl0SOn0SO4JX^5"}, "label": "a green vw van"}]} {"id": 14484, "image": "COCO_train2014_000000014484.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the front of a green and white vw van\"."}], "task": "refering", "answer_template": "The \"the front of a green and white vw van\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [66, 67, 68, 84, 85, 86, 87, 102, 103, 104, 105, 106, 119, 120, 121, 122, 123, 124, 137, 138, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 160, 161, 173, 174, 175, 176, 177, 178, 179, 191, 192, 193, 194, 195, 196, 197, 209, 210, 211, 212, 213, 214, 215, 227, 232], "bbox": [0.6203, 0.27168367346938777, 0.9708, 0.8786479591836734], "iscrowd": 0, "area": 32041.480599999995, "rle": {"size": [392, 500], "counts": "Ulf31W<3L3M4M2M2N2O1N3MQ2PNc0\\O?A5J5J7J5K2M3N2N2M3N2SN[KTLh4k3^KnKd4P4cKiK^4W4iKbKX4]4nK]KR4b4ULXKl3g4ZLSKf3m4`LmJa3Q5fLjJY3V5l11O1O1O1N2O1O1O1O1O00100001O00000000000O100000000000000000000O1000000000001O00000O10000000000000000000000000000001O0000000001O01O0000000000001O0000000000001O0000000000001O1O2O0O1O1O1O2N1O1O1O2N1O1O1O15L2M1O1O1OO2N1O1O1O2N1O1O1O1O2gJlIZ4W6cKiI]4Y6aKgI^4]6_KdI_4_6_KaIa4b6\\K^Ic4e6[K\\Ic4h6XKZIg4i6UKYIk4S7N4L3L5J5K6K4K6M2N3L3N3M3L3N3L3Nj0VOl0SOn0SO4JX^5"}, "label": "the front of a green and white vw van"}]} {"id": 14484, "image": "COCO_train2014_000000014484.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow volkswagen van\"."}], "task": "refering", "answer_template": "The \"a yellow volkswagen van\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 113, 114, 115, 127, 128, 129, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 150, 151, 152, 163, 164, 165, 166, 167, 168, 169, 170, 181, 182, 183, 184, 185, 186, 187, 188, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223], "bbox": [0.04276, 0.45375, 0.46021999999999996, 0.884923469387755], "iscrowd": 0, "area": 29512.79925, "rle": {"size": [392, 500], "counts": "d[82j;<1O0000M3oNQ1nNR1_Oa0M4M2M3M3M3M3M3M3M3M4L3M3M3O1O100O101N100O100O100O100O2O0O100O100000000O2O0000000000000O101O00000000000O10001O000000000O1000001O0000000O100000001O00000O100000O10000000O100000000000000O100000O10000000001O000000001O0000001O0O10001O0000001O000000001O0000001N1000001O000000001O0000001O00000O10000000001O00000000000O1O100O11O9G000010O000001O00O1N2ZOlHWLU7g3VIoKk6o3h0N2N2N2N2N3M2N2N2N2N2N2M3N2N2b0@N2N3M2N3M2N2N3L3L5K4VN^GHl:DaRW3"}, "label": "a yellow volkswagen van"}]} {"id": 14484, "image": "COCO_train2014_000000014484.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow vw bus with license plate lpl 3iip\"."}], "task": "refering", "answer_template": "The \"a yellow vw bus with license plate lpl 3iip\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [109, 110, 111, 112, 113, 114, 115, 127, 128, 129, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 150, 151, 152, 163, 164, 165, 166, 167, 168, 169, 170, 181, 182, 183, 184, 185, 186, 187, 188, 198, 199, 200, 201, 202, 203, 204, 205, 218, 219, 220, 221, 222, 223], "bbox": [0.04276, 0.45375, 0.46021999999999996, 0.884923469387755], "iscrowd": 0, "area": 29512.79925, "rle": {"size": [392, 500], "counts": "d[82j;<1O0000M3oNQ1nNR1_Oa0M4M2M3M3M3M3M3M3M3M4L3M3M3O1O100O101N100O100O100O100O2O0O100O100000000O2O0000000000000O101O00000000000O10001O000000000O1000001O0000000O100000001O00000O100000O10000000O100000000000000O100000O10000000001O000000001O0000001O0O10001O0000001O000000001O0000001N1000001O000000001O0000001O00000O10000000001O00000000000O1O100O11O9G000010O000001O00O1N2ZOlHWLU7g3VIoKk6o3h0N2N2N2N2N3M2N2N2N2N2N2M3N2N2b0@N2N3M2N3M2N2N3L3L5K4VN^GHl:DaRW3"}, "label": "a yellow vw bus with license plate lpl 3iip"}]} {"id": 153749, "image": "COCO_train2014_000000153749.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women standing wearing a green shirt and blue vest with her left hand on her hip\"."}], "task": "refering", "answer_template": "The \"a women standing wearing a green shirt and blue vest with her left hand on her hip\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 147, 162, 163, 164, 179, 180, 181, 196, 197, 198, 199, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 247, 248, 249, 250, 251, 252, 253, 265, 266, 267, 268, 269, 270, 281, 282, 283, 284, 285, 286, 298, 299, 300, 301, 302, 303, 314, 315, 316, 317, 318, 319, 320, 331, 332, 333, 334, 335, 336, 348, 349, 350, 351, 352, 353, 366, 367, 368, 369, 370, 383, 384, 385, 386], "bbox": [0.5190208333333334, 0.3449375, 0.8858958333333333, 1.0], "iscrowd": 0, "area": 48960.43015, "rle": {"size": [640, 480], "counts": "`Ul47Uc0j0WOi0_O`0I8H8H7eGTMVOT3c0\\MkNm2n0[MiNo2o0VMjNT3m0QMmNZ3i0kLcIjNU4^4W2lL^ImNT4Y4]2oLZInNR4U4c2RMVIoNo3R4k2TMoHQOo3m3Q3WMkHROm3i3W3ZMgHTOj3d3^3mN[LU1e3POTLR1k3TOnKn0Q4XOgKk0Y4ZOaKg0^4_O[Kc0d4BVK`0i4ERK in this image.", "objects": [{"patches": [145, 146, 147, 162, 163, 164, 179, 180, 181, 196, 197, 198, 199, 213, 214, 215, 216, 217, 230, 231, 232, 233, 234, 235, 247, 248, 249, 250, 251, 252, 253, 265, 266, 267, 268, 269, 270, 281, 282, 283, 284, 285, 286, 298, 299, 300, 301, 302, 303, 314, 315, 316, 317, 318, 319, 320, 331, 332, 333, 334, 335, 336, 348, 349, 350, 351, 352, 353, 366, 367, 368, 369, 370, 383, 384, 385, 386], "bbox": [0.5190208333333334, 0.3449375, 0.8858958333333333, 1.0], "iscrowd": 0, "area": 48960.43015, "rle": {"size": [640, 480], "counts": "`Ul47Uc0j0WOi0_O`0I8H8H7eGTMVOT3c0\\MkNm2n0[MiNo2o0VMjNT3m0QMmNZ3i0kLcIjNU4^4W2lL^ImNT4Y4]2oLZInNR4U4c2RMVIoNo3R4k2TMoHQOo3m3Q3WMkHROm3i3W3ZMgHTOj3d3^3mN[LU1e3POTLR1k3TOnKn0Q4XOgKk0Y4ZOaKg0^4_O[Kc0d4BVK`0i4ERK in this image.", "objects": [{"patches": [72, 73, 74, 89, 90, 91, 92, 105, 106, 107, 108, 109, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 290, 291, 292, 293, 294, 295, 296, 297, 298, 307, 308, 309, 310, 311, 312, 313, 314, 324, 325, 326, 327, 328, 329, 330, 331, 341, 342, 343, 344, 345, 346, 347, 348, 358, 359, 360, 361, 362, 363, 364, 365, 366, 375, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.0149375, 0.20392187499999997, 0.6128541666666667, 0.9864374999999999], "iscrowd": 0, "area": 98247.09205, "rle": {"size": [640, 480], "counts": "]g4;ac0U1lNT1lNd0[O5L3L4M3M4K4M3L5L3M3L4O2N1O1O1O2N1O1O2N1O4L3`ETKg5o4oIYKP6j4fI^KY6f4[IcKd6`4RIhKm6[4iHmKV7W4_HQL`7R4UHWLk7k3kG]LT8f3bGbL]8b3XGfLg8]3nFlLQ9W3eFQMZ9S3[FUMd9e6N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O100O100O1O100O100O100O100O101N100O100O1O100O100O100O100QN_D^Kb;`4dD\\K\\;b4kDXKV;e4QEWKo:g4WETKj:j4\\ERKd:k4cEPK_:m4hEmJY:Q5mEkJS:R5TFiJm9U5YFgJg9W5_FdJb9Y5fFbJZ9\\5f2N2M3N2N2O1O1O1O1O1O100O1O2N1O1O1O100O1O1O1O1O1O100O10000000000000001O0O1000000000000000000000000000000000001O01O1O1O001O1O1O1O001O1O1O1O1O001O1O1O10O01O1O1O1O1O001O1O1O1O1O2N2N3M2N2N3M2N2O2M2N2cCdJR:f7O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O101N1O1O1O1O001O1O001O001O1O001O1O001O00100O001O1VLPGbLP9]3^HVKc7g4fHRK\\7k4gHSK\\7h4iHUK[7f4hHXK[7c4iH[KZ7a4jH\\KY7_4kH_KX7]4kHaKX7Z4mHcKV7Y4mHeKV7V4nHhKU7T4oHiKT7R4PIhKW7T4lHgKZ7T4kHfK[7V4hHeK^7V4fHeK`7W4X4K5L4K7J7H8I7H8I7I7H8I7H5L5J5L5K4K6K4K6K4K6K4LTjc3"}, "label": "a man in hat holding a hotdog"}]} {"id": 153749, "image": "COCO_train2014_000000153749.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an old man eating a hot dog\"."}], "task": "refering", "answer_template": "The \"an old man eating a hot dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 89, 90, 91, 92, 105, 106, 107, 108, 109, 122, 123, 124, 125, 126, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 290, 291, 292, 293, 294, 295, 296, 297, 298, 307, 308, 309, 310, 311, 312, 313, 314, 324, 325, 326, 327, 328, 329, 330, 331, 341, 342, 343, 344, 345, 346, 347, 348, 358, 359, 360, 361, 362, 363, 364, 365, 366, 375, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.0149375, 0.20392187499999997, 0.6128541666666667, 0.9864374999999999], "iscrowd": 0, "area": 98247.09205, "rle": {"size": [640, 480], "counts": "]g4;ac0U1lNT1lNd0[O5L3L4M3M4K4M3L5L3M3L4O2N1O1O1O2N1O1O2N1O4L3`ETKg5o4oIYKP6j4fI^KY6f4[IcKd6`4RIhKm6[4iHmKV7W4_HQL`7R4UHWLk7k3kG]LT8f3bGbL]8b3XGfLg8]3nFlLQ9W3eFQMZ9S3[FUMd9e6N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O100O100O1O100O100O100O100O101N100O100O1O100O100O100O100QN_D^Kb;`4dD\\K\\;b4kDXKV;e4QEWKo:g4WETKj:j4\\ERKd:k4cEPK_:m4hEmJY:Q5mEkJS:R5TFiJm9U5YFgJg9W5_FdJb9Y5fFbJZ9\\5f2N2M3N2N2O1O1O1O1O1O100O1O2N1O1O1O100O1O1O1O1O1O100O10000000000000001O0O1000000000000000000000000000000000001O01O1O1O001O1O1O1O001O1O1O1O1O001O1O1O10O01O1O1O1O1O001O1O1O1O1O2N2N3M2N2N3M2N2O2M2N2cCdJR:f7O1O2N1O1O1O1O2N1O1O1O1O2N1O1O1O101N1O1O1O1O001O1O001O001O1O001O1O001O00100O001O1VLPGbLP9]3^HVKc7g4fHRK\\7k4gHSK\\7h4iHUK[7f4hHXK[7c4iH[KZ7a4jH\\KY7_4kH_KX7]4kHaKX7Z4mHcKV7Y4mHeKV7V4nHhKU7T4oHiKT7R4PIhKW7T4lHgKZ7T4kHfK[7V4hHeK^7V4fHeK`7W4X4K5L4K7J7H8I7H8I7I7H8I7H5L5J5L5K4K6K4K6K4K6K4LTjc3"}, "label": "an old man eating a hot dog"}]} {"id": 137377, "image": "COCO_train2014_000000137377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman on left with phone\"."}], "task": "refering", "answer_template": "The \"woman on left with phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 69, 70, 71, 92, 93, 94, 95, 115, 116, 117, 118, 119, 139, 140, 141, 142, 162, 163, 164, 165, 185, 186, 187, 188, 208, 209, 210, 211, 231, 232, 233, 234, 254, 255, 256, 257, 277, 278, 279, 280, 300, 301, 302, 303, 304, 323, 324, 325, 326, 327], "bbox": [0.0044843750000000005, 0.18536299765807965, 0.22140625000000003, 0.9880327868852459], "iscrowd": 0, "area": 28294.059500000003, "rle": {"size": [427, 640], "counts": "c[16Q=7I6J6J5J7J6N1N3N1O2N1N3N2N1O2N1N3N2N1O2M2O2N1O2M8I8H9G9G4K5L4L4L4K5L4^JPLa1T4WNYN@k17ZNIi1O\\N0i1G\\N9h1_O\\Na0h1VO]Nj0g1mN^NR1g1eN^N[1f1]N^Nc1f1TN_Nl1e1kM`NT2d1eM`N[2d1\\MaNd2S61O00001O0010O0001O00001O001O2N1O1O1O1O2N1O1gGlLY6U3cInL]6X3[IkLd6Z3UIhLk6^3lHfLS7`3cHdL]7f4O001O001O001O001O001O001O0010O01O001OTOnH_KR7_4n0N2O1N2N2N2O1N2OWJmK\\2S4eMSLT2m3lM[Ll1f3SN_Lh1a3SLRLg0>T3o4lLTKQ3m4oLTKo2l4QMWKl2i4TMYKj2h4VMZKg2g4XM[Kf2h4XMZKe2h4YMZKe2i4VM\\Kg2f4UM^Ki2e4RM`Kk2b4QMbKm2a4mLeKP3]4jLiKT3Z7M3M2N3M2O3L9G8H9G8H9G8H8H9G9GRY_6"}, "label": "woman on left with phone"}]} {"id": 137377, "image": "COCO_train2014_000000137377.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl with a cellphone\"."}], "task": "refering", "answer_template": "The \"a girl with a cellphone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [47, 48, 69, 70, 71, 92, 93, 94, 95, 115, 116, 117, 118, 119, 139, 140, 141, 142, 162, 163, 164, 165, 185, 186, 187, 188, 208, 209, 210, 211, 231, 232, 233, 234, 254, 255, 256, 257, 277, 278, 279, 280, 300, 301, 302, 303, 304, 323, 324, 325, 326, 327], "bbox": [0.0044843750000000005, 0.18536299765807965, 0.22140625000000003, 0.9880327868852459], "iscrowd": 0, "area": 28294.059500000003, "rle": {"size": [427, 640], "counts": "c[16Q=7I6J6J5J7J6N1N3N1O2N1N3N2N1O2N1N3N2N1O2M2O2N1O2M8I8H9G9G4K5L4L4L4K5L4^JPLa1T4WNYN@k17ZNIi1O\\N0i1G\\N9h1_O\\Na0h1VO]Nj0g1mN^NR1g1eN^N[1f1]N^Nc1f1TN_Nl1e1kM`NT2d1eM`N[2d1\\MaNd2S61O00001O0010O0001O00001O001O2N1O1O1O1O2N1O1gGlLY6U3cInL]6X3[IkLd6Z3UIhLk6^3lHfLS7`3cHdL]7f4O001O001O001O001O001O001O0010O01O001OTOnH_KR7_4n0N2O1N2N2N2O1N2OWJmK\\2S4eMSLT2m3lM[Ll1f3SN_Lh1a3SLRLg0>T3o4lLTKQ3m4oLTKo2l4QMWKl2i4TMYKj2h4VMZKg2g4XM[Kf2h4XMZKe2h4YMZKe2i4VM\\Kg2f4UM^Ki2e4RM`Kk2b4QMbKm2a4mLeKP3]4jLiKT3Z7M3M2N3M2O3L9G8H9G8H9G8H8H9G9GRY_6"}, "label": "a girl with a cellphone"}]} {"id": 301218, "image": "COCO_train2014_000000301218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver audi car\"."}], "task": "refering", "answer_template": "The \"a silver audi car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 188, 189, 190, 191, 192, 203, 204, 205, 206, 217, 218, 219, 220, 221, 233, 234, 235, 236], "bbox": [0.5228971962616823, 0.5018906249999999, 0.8083644859813085, 0.6973906249999999], "iscrowd": 0, "area": 11174.693850000001, "rle": {"size": [640, 428], "counts": "g\\\\47`c0:E:F;F9K6N101N101O0O2O:E8I2N1N2O1N2O0O2O1O1N100O100O1O100O1O1O1O100O1O1O100O1O1ON3N1O1N2O1O2O01O1O100O1O1O100O1O1O1O100O001O100O1000000O1000000O10000O101O000O10000001O0001O000001O00000001O01O000UM^_OX2b`0`Mg_O_2Y`0XMP@h2e`0O10O0L4L4K5L4M3000001O3M3MH8C=B>C=B>C=C=@Znb1"}, "label": "a silver audi car"}]} {"id": 301218, "image": "COCO_train2014_000000301218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white car next to truck with mendon truck leasing mud flap\"."}], "task": "refering", "answer_template": "The \"white car next to truck with mendon truck leasing mud flap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [175, 176, 177, 188, 189, 190, 191, 192, 203, 204, 205, 206, 217, 218, 219, 220, 221, 233, 234, 235, 236], "bbox": [0.5228971962616823, 0.5018906249999999, 0.8083644859813085, 0.6973906249999999], "iscrowd": 0, "area": 11174.693850000001, "rle": {"size": [640, 428], "counts": "g\\\\47`c0:E:F;F9K6N101N101O0O2O:E8I2N1N2O1N2O0O2O1O1N100O100O1O100O1O1O1O100O1O1O100O1O1ON3N1O1N2O1O2O01O1O100O1O1O100O1O1O1O100O001O100O1000000O1000000O10000O101O000O10000001O0001O000001O00000001O01O000UM^_OX2b`0`Mg_O_2Y`0XMP@h2e`0O10O0L4L4K5L4M3000001O3M3MH8C=B>C=B>C=C=@Znb1"}, "label": "white car next to truck with mendon truck leasing mud flap"}]} {"id": 301218, "image": "COCO_train2014_000000301218.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a front of a white delivery truck on the road\"."}], "task": "refering", "answer_template": "The \"a front of a white delivery truck on the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 15, 16, 30, 31, 45, 46, 60, 61, 62, 75, 76, 77, 78, 90, 91, 92, 93, 105, 106, 107, 108, 109, 120, 121, 122, 123, 124, 135, 136, 137, 138, 139, 150, 151, 152, 153, 154, 165, 166, 167, 168, 169, 170, 180, 181, 182, 183, 184, 185, 195, 196, 197, 198, 199, 200, 210, 211, 212, 213, 214, 215, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 255, 256, 257, 258, 270, 271, 272, 273, 285, 286, 287], "bbox": [0.0067289719626168224, 0.0045, 0.3494626168224299, 0.867421875], "iscrowd": 0, "area": 54398.355850000014, "rle": {"size": [640, 428], "counts": "Tl1R5n>g1YNh1XNh1XNg1YNh1XNh1XNl0TO1O1O1O1O001O1O1O1O1O1N101O1O1O1O1O1O1O001O1O1O1O1O1O001O1O1O0000000000bAkMQ;U2kCUOo;k0mB?m<`3J5K00001O0000000001O0001O000000000000001O00000000000N2M4L3M3M3M3M3M3M3N3L3M5K4L5K4L5K5K4J7ROm0A`0N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N2N2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N3I6mIgBW5`=dJiBP5^=lJlBh4Z=TKoBa4a>D=D;E;D in this image.", "objects": [{"patches": [0, 1, 15, 16, 30, 31, 45, 46, 60, 61, 62, 75, 76, 77, 78, 90, 91, 92, 93, 105, 106, 107, 108, 109, 120, 121, 122, 123, 124, 135, 136, 137, 138, 139, 150, 151, 152, 153, 154, 165, 166, 167, 168, 169, 170, 180, 181, 182, 183, 184, 185, 195, 196, 197, 198, 199, 200, 210, 211, 212, 213, 214, 215, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 255, 256, 257, 258, 270, 271, 272, 273, 285, 286, 287], "bbox": [0.0067289719626168224, 0.0045, 0.3494626168224299, 0.867421875], "iscrowd": 0, "area": 54398.355850000014, "rle": {"size": [640, 428], "counts": "Tl1R5n>g1YNh1XNh1XNg1YNh1XNh1XNl0TO1O1O1O1O001O1O1O1O1O1N101O1O1O1O1O1O1O001O1O1O1O1O1O001O1O1O0000000000bAkMQ;U2kCUOo;k0mB?m<`3J5K00001O0000000001O0001O000000000000001O00000000000N2M4L3M3M3M3M3M3M3N3L3M5K4L5K4L5K5K4J7ROm0A`0N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N2N2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N3M2N2N3M2N3I6mIgBW5`=dJiBP5^=lJlBh4Z=TKoBa4a>D=D;E;D in this image.", "objects": [{"patches": [151, 152, 174, 175, 197, 198, 220, 221, 240, 242, 243, 244, 263, 265, 266, 267, 289, 290, 312], "bbox": [0.4583125, 0.4245433255269321, 0.6311249999999999, 0.8952693208430913], "iscrowd": 0, "area": 4437.155749999998, "rle": {"size": [427, 640], "counts": "[`j31e;2aE:^:JSEc0k:@hDk0W;e0O2N1O2Si=]N^XBb0^O4L101N2N2NQOP10O2O0O10000O2O000O100O101O0O10000O2O0O10S2mM1N10O0@a0_Oa0@?10O1O010O00100O00M4kLAVIf0l63cGc0^8U20O1O2O0O1O1016I7I8H7JbMiGGV80_HE`72UIDi64lI@T66aJ_O^58TKAj46bKH]40nKNR4HZL3h3DdL7\\8L4KZRR3"}, "label": "a brown chair that the woman is sitting on"}]} {"id": 252093, "image": "COCO_train2014_000000252093.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair the woman in pink is sitting in\"."}], "task": "refering", "answer_template": "The \"the chair the woman in pink is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 152, 174, 175, 197, 198, 220, 221, 240, 242, 243, 244, 263, 265, 266, 267, 289, 290, 312], "bbox": [0.4583125, 0.4245433255269321, 0.6311249999999999, 0.8952693208430913], "iscrowd": 0, "area": 4437.155749999998, "rle": {"size": [427, 640], "counts": "[`j31e;2aE:^:JSEc0k:@hDk0W;e0O2N1O2Si=]N^XBb0^O4L101N2N2NQOP10O2O0O10000O2O000O100O101O0O10000O2O0O10S2mM1N10O0@a0_Oa0@?10O1O010O00100O00M4kLAVIf0l63cGc0^8U20O1O2O0O1O1016I7I8H7JbMiGGV80_HE`72UIDi64lI@T66aJ_O^58TKAj46bKH]40nKNR4HZL3h3DdL7\\8L4KZRR3"}, "label": "the chair the woman in pink is sitting in"}]} {"id": 252093, "image": "COCO_train2014_000000252093.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"unoccupied chair with visible seat\"."}], "task": "refering", "answer_template": "The \"unoccupied chair with visible seat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 209, 210, 211, 232, 233, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325], "bbox": [0.003, 0.5768384074941452, 0.18590625, 0.981334894613583], "iscrowd": 0, "area": 9151.019500000002, "rle": {"size": [427, 640], "counts": "aW15m<9F:F:YOPOjDY1V;hN_Db1a;5O000000000000000000000001O0000000001O000000000000000000000001O000000000001O000000000000000000OM4L3M4L3L5L3M4K4M4L4K4M4L3M4K4M4L3L5L3M4L4K4M4O010O0100O010O010O10O10O0101O1NVO]GnLc8m2eGQMY8j2PHTMo7g2YHWMf7d2bHZM\\7a2lH]MT7^2TI`Mj6\\2^IbMa6X2hIfMW6U2QJhMn5T2ZJjMe5P2m2L5K5J5L5K4K6K5K4K6K4L5J6KaPi6"}, "label": "unoccupied chair with visible seat"}]} {"id": 252093, "image": "COCO_train2014_000000252093.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an empty brown chair\"."}], "task": "refering", "answer_template": "The \"an empty brown chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [187, 188, 209, 210, 211, 232, 233, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325], "bbox": [0.003, 0.5768384074941452, 0.18590625, 0.981334894613583], "iscrowd": 0, "area": 9151.019500000002, "rle": {"size": [427, 640], "counts": "aW15m<9F:F:YOPOjDY1V;hN_Db1a;5O000000000000000000000001O0000000001O000000000000000000000001O000000000001O000000000000000000OM4L3M4L3L5L3M4K4M4L4K4M4L3M4K4M4L3L5L3M4L4K4M4O010O0100O010O010O10O10O0101O1NVO]GnLc8m2eGQMY8j2PHTMo7g2YHWMf7d2bHZM\\7a2lH]MT7^2TI`Mj6\\2^IbMa6X2hIfMW6U2QJhMn5T2ZJjMe5P2m2L5K5J5L5K4K6K5K4K6K4L5J6KaPi6"}, "label": "an empty brown chair"}]} {"id": 47294, "image": "COCO_train2014_000000047294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cop on a motorcycle behind another cop on a motorcycle\"."}], "task": "refering", "answer_template": "The \"a cop on a motorcycle behind another cop on a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 143, 162, 163, 164, 165, 166, 167, 185, 186, 187, 188, 189, 208, 209, 210, 211, 212, 232, 233, 234, 235, 254, 255, 256, 257, 258, 259, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 305, 322, 323, 324, 325, 326, 327, 328, 346, 348, 349, 350, 351, 372, 373, 374], "bbox": [0.034015625, 0.35904166666666665, 0.2993125, 0.9848750000000001], "iscrowd": 0, "area": 33483.0986, "rle": {"size": [480, 640], "counts": "be:8f>^3l7\\LfG6<`3o7YLeG7:b3Q8XLcG88c3V8TLbG96e3X8RLbG:3f3[8PLbG:1h3^8nK_G;0j3a8kK_G in this image.", "objects": [{"patches": [158, 181, 182, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 243, 244, 249, 250, 251, 252, 265, 266, 267, 268, 272, 273, 274, 275, 288, 289, 290, 291, 296, 297, 298, 311, 312, 313, 314, 315, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 363, 364, 365, 366, 367, 379, 380, 381, 382, 383, 384, 386, 387, 388, 389, 390], "bbox": [0.50953125, 0.394875, 0.995515625, 0.9845625000000001], "iscrowd": 0, "area": 44683.72660000001, "rle": {"size": [480, 640], "counts": "[Xi4;^=V1K6J5J7J6J6`NPNcE]2X:TNXEP2c:Y1K5L4K6L30000O10000O1000000O10000O10000O1000000O10000O1000000O10001O000000000000000000000000000000000000000000000000000000010O00000001O000000001O000000001O000000001O000000001O000000010O00000000001O00000oKWFk2i9jLfFR3h:L3M4L4L3M4L2N2N2N2N2N1O2N2N2N2N2N2N2O1N2M3K>C=B?VOXUa0J\\k^O4kGMT18fNKX17fNKX17\\N3d1OPN;P2GcMd0]2^OVMn0i2TORLP2m3RNkKU2T4mMjKT2U4nMjKR2V4oMiKR2V4nMiKS2V4nMjKR2V4nMjKR2V4oMiKQ2W4oMhKR2X4nMhKR2X4nMhKR2X4nMhKR2X4oMhKP2X4PNhKP2X4PNiKo1X4QNgKo1Y4QNhKn1X4RNiKm1X4SNgK;eL5d7@hK9fL6b7BhK6iL6`7CgK6kL6^7EgK3mL7\\7FiK0mL9Z7GjKOmL9Z7HeHROe2j0^M;X7IdHVOe2c0aM=V7KbHYOf2 in this image.", "objects": [{"patches": [158, 181, 182, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 243, 244, 249, 250, 251, 252, 265, 266, 267, 268, 272, 273, 274, 275, 288, 289, 290, 291, 296, 297, 298, 311, 312, 313, 314, 315, 317, 318, 319, 320, 321, 333, 334, 335, 336, 337, 338, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 363, 364, 365, 366, 367, 379, 380, 381, 382, 383, 384, 386, 387, 388, 389, 390], "bbox": [0.50953125, 0.394875, 0.995515625, 0.9845625000000001], "iscrowd": 0, "area": 44683.72660000001, "rle": {"size": [480, 640], "counts": "[Xi4;^=V1K6J5J7J6J6`NPNcE]2X:TNXEP2c:Y1K5L4K6L30000O10000O1000000O10000O10000O1000000O10000O1000000O10001O000000000000000000000000000000000000000000000000000000010O00000001O000000001O000000001O000000001O000000001O000000010O00000000001O00000oKWFk2i9jLfFR3h:L3M4L4L3M4L2N2N2N2N2N1O2N2N2N2N2N2N2O1N2M3K>C=B?VOXUa0J\\k^O4kGMT18fNKX17fNKX17\\N3d1OPN;P2GcMd0]2^OVMn0i2TORLP2m3RNkKU2T4mMjKT2U4nMjKR2V4oMiKR2V4nMiKS2V4nMjKR2V4nMjKR2V4oMiKQ2W4oMhKR2X4nMhKR2X4nMhKR2X4nMhKR2X4oMhKP2X4PNhKP2X4PNiKo1X4QNgKo1Y4QNhKn1X4RNiKm1X4SNgK;eL5d7@hK9fL6b7BhK6iL6`7CgK6kL6^7EgK3mL7\\7FiK0mL9Z7GjKOmL9Z7HeHROe2j0^M;X7IdHVOe2c0aM=V7KbHYOf2 in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 105, 106, 107, 108, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 281, 282, 283, 284, 285, 286, 287, 304, 305, 309, 310], "bbox": [0.00125, 0.23947916666666666, 0.8967031249999999, 0.7944583333333333], "iscrowd": 0, "area": 100068.41784999995, "rle": {"size": [480, 640], "counts": "Xh0W1Q in this image.", "objects": [{"patches": [100, 101, 102, 103, 104, 105, 106, 107, 108, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 281, 282, 283, 284, 285, 286, 287, 304, 305, 309, 310], "bbox": [0.00125, 0.23947916666666666, 0.8967031249999999, 0.7944583333333333], "iscrowd": 0, "area": 100068.41784999995, "rle": {"size": [480, 640], "counts": "Xh0W1Q in this image.", "objects": [{"patches": [151, 154, 161, 162, 163, 164, 165, 166, 171, 172, 173, 174, 175, 176, 181, 184, 185, 194, 195], "bbox": [0.08640138408304499, 0.67425, 0.6678546712802768, 0.8422031249999999], "iscrowd": 0, "area": 4233.9593, "rle": {"size": [640, 289], "counts": "oS`02kc04]NOR_O5k`01P_O0n`03Q_ONl`05R_OMk`06T_OKj`07T_OKi`08V_OIh`0:S_OJj`07T_OLk`05R_ONl`03R_OOn`01P_O2n`00o^O2Ra0Mo^O3Qa0Mo^O2Sa0Ln^O4Ra0Kn^O6Sa0Im^O6Ta0Im^O7Ta0En^O;Sa0_OS_Oa0n`0[OU_Od0l`0\\OT_Od0l`0\\OS_Od0o`0\\OP_Od0Pa0\\OP_O?Va0Ah^O;]a0Ec^O:_a0Ea^O:`a0G_^O8ca0G\\^O9ea0HZ^O7ha0HX^O7ia0IV^O7la0IS^O6na0JU^O2ma0NR100O2K4LnXe00UgZO2M2N2O1N2N2O1N2N2O1N2O1N2N3N2M4L3N2M3M3NA^]OC_b0?b]OA\\b0a0d]O_OYb0c0h]O]OUb0f0k]O[ORb0f0o]OZOna0h0S^OXOh0AW`0l0n^OZO39d0E^`0`0o^OB09?Id`06P_OIM8 in this image.", "objects": [{"patches": [151, 154, 161, 162, 163, 164, 165, 166, 171, 172, 173, 174, 175, 176, 181, 184, 185, 194, 195], "bbox": [0.08640138408304499, 0.67425, 0.6678546712802768, 0.8422031249999999], "iscrowd": 0, "area": 4233.9593, "rle": {"size": [640, 289], "counts": "oS`02kc04]NOR_O5k`01P_O0n`03Q_ONl`05R_OMk`06T_OKj`07T_OKi`08V_OIh`0:S_OJj`07T_OLk`05R_ONl`03R_OOn`01P_O2n`00o^O2Ra0Mo^O3Qa0Mo^O2Sa0Ln^O4Ra0Kn^O6Sa0Im^O6Ta0Im^O7Ta0En^O;Sa0_OS_Oa0n`0[OU_Od0l`0\\OT_Od0l`0\\OS_Od0o`0\\OP_Od0Pa0\\OP_O?Va0Ah^O;]a0Ec^O:_a0Ea^O:`a0G_^O8ca0G\\^O9ea0HZ^O7ha0HX^O7ia0IV^O7la0IS^O6na0JU^O2ma0NR100O2K4LnXe00UgZO2M2N2O1N2N2O1N2N2O1N2O1N2N3N2M4L3N2M3M3NA^]OC_b0?b]OA\\b0a0d]O_OYb0c0h]O]OUb0f0k]O[ORb0f0o]OZOna0h0S^OXOh0AW`0l0n^OZO39d0E^`0`0o^OB09?Id`06P_OIM8 in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 37, 38, 39, 40, 41, 42, 59, 60, 61, 81, 82, 83, 104, 105, 126, 127, 148, 149, 170, 171, 192, 193, 213, 214, 215, 216, 234, 235, 236, 237, 238], "bbox": [0.6705882352941176, 0.002352941176470588, 0.9341176470588235, 0.4941176470588235], "iscrowd": 0, "area": 16225.920000000006, "rle": {"size": [612, 612], "counts": "Ume72Qc01O101N1O2N1O1O2N101N1O2O0O101N101N100\\IJYJ8U5:jJFf4l0WKVOW4]1gKcNT4c1jK_NU4c1iK]NW4d1gK]NY4e1eK\\N[4d1cK]N]4e1aK\\N^4f1_K[Na4f1^KZNb4h1[KZNd4g1[KYNe4i1XKYNg4h1XKXNh4j1UKWNk4k1SKVNl4k1RKVNn4l1PKUNo4l1oJUNQ5l1nJUNQ5l1mJUNS5m1kJSNU5n1iJTNV5m1iJSNW5n1gJTNX5m1gJSNY5o1dJRN\\5o1\\JXNd5i1UJ]Nk5d1nIbNR6_1gIgNY6[1_IkNa6V1XIPOh6Q1PIVOP7k0iH[OW7f0bH@^7b0ZHDf7=SHIm78XGb0h8m40000000000000000000oHYFe4g9UKeFe4[9UKQGe4P9SK\\Gh4d8RKhGh4X8UKQHe4o7ZKVHS3TNlLf90YHj2YNVM^9O[Hb2`N^MU9O^HX2fNiMl8N`HP2mNQNc8NcHg1RO[N[8LfH_1YOdNQ8LiHV1^OnNi7KlHm0DWO`7LmHc0LAW7KPI:2Jn6LQI193f6LSIG`0<]6MUI\\Og0g0T6L\\IoNh0U1l5L[M3e2M[M3e2M\\M1e2O[M0f2O\\M0d20\\MOe21\\MMe23[MLf23\\MLd24\\MKe25\\MIe27[MHf27\\MHd28\\MGe29\\MEe2;[MDf2[M@f2a0YM^Oh2b0YM]Og2c0YM\\Oh2d0YMZOh2g0WMXOj2h0WMWOi2i0WMVOk2i0VMUOk2l0TMSOm2m0TMROl2n0TMQOm2o0d7001O0000001O000000001O0000001O000000001O0000001O0000001O000000001O0009Gm0TOekg0"}, "label": "wine glass nearest the woman ' s pizza"}]} {"id": 153814, "image": "COCO_train2014_000000153814.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wine glass that is closest to the lady in black\"."}], "task": "refering", "answer_template": "The \"wine glass that is closest to the lady in black\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 18, 19, 20, 37, 38, 39, 40, 41, 42, 59, 60, 61, 81, 82, 83, 104, 105, 126, 127, 148, 149, 170, 171, 192, 193, 213, 214, 215, 216, 234, 235, 236, 237, 238], "bbox": [0.6705882352941176, 0.002352941176470588, 0.9341176470588235, 0.4941176470588235], "iscrowd": 0, "area": 16225.920000000006, "rle": {"size": [612, 612], "counts": "Ume72Qc01O101N1O2N1O1O2N101N1O2O0O101N101N100\\IJYJ8U5:jJFf4l0WKVOW4]1gKcNT4c1jK_NU4c1iK]NW4d1gK]NY4e1eK\\N[4d1cK]N]4e1aK\\N^4f1_K[Na4f1^KZNb4h1[KZNd4g1[KYNe4i1XKYNg4h1XKXNh4j1UKWNk4k1SKVNl4k1RKVNn4l1PKUNo4l1oJUNQ5l1nJUNQ5l1mJUNS5m1kJSNU5n1iJTNV5m1iJSNW5n1gJTNX5m1gJSNY5o1dJRN\\5o1\\JXNd5i1UJ]Nk5d1nIbNR6_1gIgNY6[1_IkNa6V1XIPOh6Q1PIVOP7k0iH[OW7f0bH@^7b0ZHDf7=SHIm78XGb0h8m40000000000000000000oHYFe4g9UKeFe4[9UKQGe4P9SK\\Gh4d8RKhGh4X8UKQHe4o7ZKVHS3TNlLf90YHj2YNVM^9O[Hb2`N^MU9O^HX2fNiMl8N`HP2mNQNc8NcHg1RO[N[8LfH_1YOdNQ8LiHV1^OnNi7KlHm0DWO`7LmHc0LAW7KPI:2Jn6LQI193f6LSIG`0<]6MUI\\Og0g0T6L\\IoNh0U1l5L[M3e2M[M3e2M\\M1e2O[M0f2O\\M0d20\\MOe21\\MMe23[MLf23\\MLd24\\MKe25\\MIe27[MHf27\\MHd28\\MGe29\\MEe2;[MDf2[M@f2a0YM^Oh2b0YM]Og2c0YM\\Oh2d0YMZOh2g0WMXOj2h0WMWOi2i0WMVOk2i0VMUOk2l0TMSOm2m0TMROl2n0TMQOm2o0d7001O0000001O000000001O0000001O000000001O0000001O0000001O000000001O0009Gm0TOekg0"}, "label": "wine glass that is closest to the lady in black"}]} {"id": 153814, "image": "COCO_train2014_000000153814.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glass of juice in table\"."}], "task": "refering", "answer_template": "The \"glass of juice in table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 42, 43, 61, 62, 63, 64, 65, 83, 84, 85, 86, 87, 105, 106, 107, 108, 109, 128, 129, 130, 151, 173, 194, 195, 216, 217, 237, 238, 239, 258, 259, 260, 261, 262, 263, 282, 283, 284, 285], "bbox": [0.7505065359477124, 0.055343137254901954, 1.0, 0.5683496732026143], "iscrowd": 0, "area": 18547.520149999997, "rle": {"size": [612, 612], "counts": "Udb81Sc09F4M0O100O100O100O100O100O100O10000000000000000VHLcL5X3<\\LEa3j0RLVOl3X1hKhNV4`1dKaNY4c1eK]NY4f1fKZNY4h1fKYNW4j1hKVNV4m1iKSNV4o1iKRNV4o1iKQNW4P2hKPNX4R2fKoMY4R2fKnMZ4S2dKnM\\4S2cKmM]4T2bKmM]4T2bKlM^4U2aKkM_4V2`KkM_4V2`KjM`4X2^KhMb4Y2\\KiMc4X2\\KhMd4Y2[KgMe4Z2ZKfMf4[2YKfMf4[2YKeMg4\\2XKdMh4\\2XKeMg4\\2WKeMi4\\2TKfMl4[2oJiMQ5X2jJmMU5T2fJPNZ5Q2aJSN_5n1[JXNd5j1TJ\\Nl5f1jIbNV6a1_IgNb6Z1SIPOl6R1jHVOV7l0`H\\O`7g0THCk7?kGIU89mFc0T9f40_NdF[I[9]6nFbIR9V6WGiIi8o5`GPJ`8h5iGXJW8_5RH`Jn7X5[HgJe7R5cHmJ]7l4kHTKT7d4TI\\Kl6]4\\IbKe6V4bIjK^6n3kIRLT6g3SJYLm5`3[J_Le5Y3cJgL^5R3iJnLV5o2mJQMS5m2oJSMQ5k2RKTMn4j2TKWMk4h2VKXMk4f2WKYMi4g2WKYMi4f2YKYMg4f2ZK[Me4d2\\K\\Md4d2]K[Md4c2]K]Mc4b2^K^Mb4a2`K^M`4b2`K^M`4a2aK_M_4`2cK_M^4_2cKaM]4_2dK`M\\4_2eKaM[4^2gKbMX4]2iKcMX4\\2hKdMX4[2jKdMV4[2kKdMV4[2lKdMT4\\2lKcMU4\\2lKdMU4Z2mKdMT4[2mKeMS4[2nKcMS4\\2oKbMR4]2oKcMQ4\\2QLbMQ4]2oKcMQ4\\2QLbMP4]2QLcMo3\\2SLbMn3^2RLbMn3]2TLaMn3]2SLbMn3]2TLbMl3^2UL`Ml3_2VL`Mj3_2XL_Mi3`2ZL^Mg3a2^6O1O1O101N1O1O100O1O2N100N2K5Ki0WOPN"}, "label": "glass of juice in table"}]} {"id": 153814, "image": "COCO_train2014_000000153814.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman with the black tank top is sitting in front of a pizza that is on the table\"."}], "task": "refering", "answer_template": "The \"the woman with the black tank top is sitting in front of a pizza that is on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 67, 68, 69, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 89, 90, 91, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 169], "bbox": [0.06650326797385621, 0.007352941176470588, 0.7848529411764705, 0.33294117647058824], "iscrowd": 0, "area": 53344.40409999999, "rle": {"size": [612, 612], "counts": "bch04kb05J7M2N2O2N1N2O2M2O2O00001O00001O001O00001O00001O001O00001O000010O01O00001O00O1000001N2O1O1O1N2O1O1O0O2O00000O100N3N1N2N2N2N3N1N2N20022O2N2N2NfT`0[1ii_O3N2N2N2N2N2M3N2N25KI7J7I6J6J6J6J6DA?H8J6J6I70000000000000O10000O1000000O1000000O1000000O1000000O1O1N2N3M2N2N2M3N2N2N2N2N2N2N2N2N2M3N2N2N22NR1nNb0^O0000000000O10000O10000O101O000O10000O1000000O10000O10000O10000000000O100000000000000000000O2O0000000000000000000O10000000000000000000000000000000000001O000001O00000000000000000000000001O0000000000000000000000000000000000000010O00000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00ZNiB_MW=V2XCgMghNmAS1Q>kNXBo0g=oNcBk0[=SOnBh0P=VOSCk0k in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 67, 68, 69, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 89, 90, 91, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 169], "bbox": [0.06650326797385621, 0.007352941176470588, 0.7848529411764705, 0.33294117647058824], "iscrowd": 0, "area": 53344.40409999999, "rle": {"size": [612, 612], "counts": "bch04kb05J7M2N2O2N1N2O2M2O2O00001O00001O001O00001O00001O001O00001O000010O01O00001O00O1000001N2O1O1O1N2O1O1O0O2O00000O100N3N1N2N2N2N3N1N2N20022O2N2N2NfT`0[1ii_O3N2N2N2N2N2M3N2N25KI7J7I6J6J6J6J6DA?H8J6J6I70000000000000O10000O1000000O1000000O1000000O1000000O1O1N2N3M2N2N2M3N2N2N2N2N2N2N2N2N2M3N2N2N22NR1nNb0^O0000000000O10000O10000O101O000O10000O1000000O10000O10000O10000000000O100000000000000000000O2O0000000000000000000O10000000000000000000000000000000000001O000001O00000000000000000000000001O0000000000000000000000000000000000000010O00000000000000000000000000000000000000000000000000000000000000001O000000000000000000000000000001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00001O001O001O00ZNiB_MW=V2XCgMghNmAS1Q>kNXBo0g=oNcBk0[=SOnBh0P=VOSCk0k in this image.", "objects": [{"patches": [137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213], "bbox": [0.2247222222222222, 0.29437908496732024, 0.7595424836601308, 0.44269607843137254], "iscrowd": 0, "area": 25438.986, "rle": {"size": [612, 612], "counts": "Peb2d1`a000001O00001O001O00001O001O00001O001O00001O00001O001O0O101O001O00001O001O00001O000O2O001O00001O001O00001O001O0O101O00001O001O00001O00001O0O100000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O000000000000000000000000000000O100000000000000000000000001O00000000000O100000000000000000000000000000000000000O100000001O00000000000000000000000000000O1000000000000000000000000000001O000000000O10000000000000000000001O1O001O001O001O1N101O0O2N2O0O2N1O2O1N1O2N101N2N1O2O0O2N1O2O1N1O2O0O2N2N101N1O2N2O0O2N6J8I7H9GQdg2"}, "label": "a large pizza in front of a woman and behind another pizza"}]} {"id": 153814, "image": "COCO_train2014_000000153814.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza in front of a woman in a black tank top\"."}], "task": "refering", "answer_template": "The \"a pizza in front of a woman in a black tank top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213], "bbox": [0.2247222222222222, 0.29437908496732024, 0.7595424836601308, 0.44269607843137254], "iscrowd": 0, "area": 25438.986, "rle": {"size": [612, 612], "counts": "Peb2d1`a000001O00001O001O00001O001O00001O001O00001O00001O001O0O101O001O00001O001O00001O000O2O001O00001O001O00001O001O0O101O00001O001O00001O00001O0O100000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O000000000000000000001O00000000000000000000001O000000000000000000001O000000000000000000000000000000O100000000000000000000000001O00000000000O100000000000000000000000000000000000000O100000001O00000000000000000000000000000O1000000000000000000000000000001O000000000O10000000000000000000001O1O001O001O001O1N101O0O2N2O0O2N1O2O1N1O2N101N2N1O2O0O2N1O2O1N1O2O0O2N2N101N1O2N2O0O2N6J8I7H9GQdg2"}, "label": "a pizza in front of a woman in a black tank top"}]} {"id": 153814, "image": "COCO_train2014_000000153814.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the cheese and pepperoni pizza farthest away from the woman sitting at the table\"."}], "task": "refering", "answer_template": "The \"the cheese and pepperoni pizza farthest away from the woman sitting at the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483], "bbox": [0.0, 0.45593137254901955, 1.0, 1.0], "iscrowd": 0, "area": 183595.08435, "rle": {"size": [612, 612], "counts": "o98j7j0V20nJR2Q5iNRHZ3n7n300O100O10000O10000O10000O100O10000O10000O10000O100O10000O10000O10000O100O10000O10001N10000O1000000000000000000O1000000000000000000O10000000000000000O1000000000000000000O10000000000000O1000O10000000000000000O100000000000000O10000O10000O1000000O10000O10000O01000O10000O1000000O10000O10000O10000O10000O1000000O100000000000000000O1000000000000000O100000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000001O000O101O0000001O0000001O00001O0000001O0000001O0000001O0000001O0000001O00001O0000000000001O00000000000000001O000000000000001O000000000000000000001O000000000000000000001O000000000000000000001O000000000000000000001O000000000000000000001O000O10000000000000001O000000000000000000001O0000001O0000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O0000001O001O00001O001O00001O001O001O00001O001O001O00001O0000001O000000000000001O000000000000001O000000000000001O000000000000001O0000000000000000001O000000000000000010O0000000000000001O000000000000001O0000000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O0000000000010O00000000001O0000001O001O001O001O000010O01O001O001O0O101O001O0O2O001O000O2O001O001N101O001O0O2O1O1O1O1N2O001N2M3N2N2N2N2N1O2N2N2N2NU2kM=CnA"}, "label": "the cheese and pepperoni pizza farthest away from the woman sitting at the table"}]} {"id": 153814, "image": "COCO_train2014_000000153814.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza that is closer to the camera than the other . it is in front of the glasses and the water\"."}], "task": "refering", "answer_template": "The \"the pizza that is closer to the camera than the other . it is in front of the glasses and the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483], "bbox": [0.0, 0.45593137254901955, 1.0, 1.0], "iscrowd": 0, "area": 183595.08435, "rle": {"size": [612, 612], "counts": "o98j7j0V20nJR2Q5iNRHZ3n7n300O100O10000O10000O10000O100O10000O10000O10000O100O10000O10000O10000O100O10000O10001N10000O1000000000000000000O1000000000000000000O10000000000000000O1000000000000000000O10000000000000O1000O10000000000000000O100000000000000O10000O10000O1000000O10000O10000O01000O10000O1000000O10000O10000O10000O10000O1000000O100000000000000000O1000000000000000O100000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000001O000O101O0000001O0000001O00001O0000001O0000001O0000001O0000001O0000001O00001O0000000000001O00000000000000001O000000000000001O000000000000000000001O000000000000000000001O000000000000000000001O000000000000000000001O000000000000000000001O000O10000000000000001O000000000000000000001O0000001O0000001O0000001O0000001O0000001O0000001O00001O0000001O0000001O0000001O001O00001O001O00001O001O001O00001O001O001O00001O0000001O000000000000001O000000000000001O000000000000001O000000000000001O0000000000000000001O000000000000000010O0000000000000001O000000000000001O0000000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O0000000000010O00000000001O0000001O001O001O001O000010O01O001O001O0O101O001O0O2O001O000O2O001O001N101O001O0O2O1O1O1O1N2O001N2M3N2N2N2N2N1O2N2N2N2NU2kM=CnA"}, "label": "the pizza that is closer to the camera than the other . it is in front of the glasses and the water"}]} {"id": 489695, "image": "COCO_train2014_000000489695.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man drives motorcycle with a single headlight\"."}], "task": "refering", "answer_template": "The \"a man drives motorcycle with a single headlight\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 83, 84, 85, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 133, 154, 155, 177, 178, 200, 201, 202, 223, 224, 225, 248, 271, 294], "bbox": [0.542703125, 0.13539583333333333, 0.83090625, 0.7286458333333333], "iscrowd": 0, "area": 13359.30360000001, "rle": {"size": [480, 640], "counts": "liR51o>4K5L4K2O0O100O100O1O100O100O100O100O100O1N2N1N3M3N2M2N300O00100O00100O1O00100O00100O1O010O10000O01000O1RO`N\\D_1d;cN[D]1d;fNYD[1f;hNXDX1g;kNWDT1i;nNUDS1j;POSDQ1k;SORDn0m;UOPDk0Pm8ATG?k8AVG?j8@XG?g8AZG?f8@\\G>d8B]G>c8A_G>`8BaG>_8@dG?[8@gG`0Y8_OiG?W8@kG`0U8_OmG`0R8_OQH`0o7_ORHa0m7^OVHe0f7YO]Hl0\\7TOeHS1T7kNoHZ1j6fNXI`1_MXN[86XJ^2b5bM`J[2b5cMaJY2a5fMaJX2a5gMaJU2d5gM_JV2f5fM[JW2Z9M3M2M5L6I6K5J6K6I6K5J7J5J6KSZb1"}, "label": "a man drives motorcycle with a single headlight"}]} {"id": 489695, "image": "COCO_train2014_000000489695.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the biker behind the first biker\"."}], "task": "refering", "answer_template": "The \"the biker behind the first biker\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [60, 61, 62, 83, 84, 85, 105, 106, 107, 108, 109, 127, 128, 129, 130, 131, 132, 133, 154, 155, 177, 178, 200, 201, 202, 223, 224, 225, 248, 271, 294], "bbox": [0.542703125, 0.13539583333333333, 0.83090625, 0.7286458333333333], "iscrowd": 0, "area": 13359.30360000001, "rle": {"size": [480, 640], "counts": "liR51o>4K5L4K2O0O100O100O1O100O100O100O100O100O1N2N1N3M3N2M2N300O00100O00100O1O00100O00100O1O010O10000O01000O1RO`N\\D_1d;cN[D]1d;fNYD[1f;hNXDX1g;kNWDT1i;nNUDS1j;POSDQ1k;SORDn0m;UOPDk0Pm8ATG?k8AVG?j8@XG?g8AZG?f8@\\G>d8B]G>c8A_G>`8BaG>_8@dG?[8@gG`0Y8_OiG?W8@kG`0U8_OmG`0R8_OQH`0o7_ORHa0m7^OVHe0f7YO]Hl0\\7TOeHS1T7kNoHZ1j6fNXI`1_MXN[86XJ^2b5bM`J[2b5cMaJY2a5fMaJX2a5gMaJU2d5gM_JV2f5fM[JW2Z9M3M2M5L6I6K5J6K6I6K5J7J5J6KSZb1"}, "label": "the biker behind the first biker"}]} {"id": 489695, "image": "COCO_train2014_000000489695.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man riding a motorcycle ahead of another motorcycle\"."}], "task": "refering", "answer_template": "The \"a man riding a motorcycle ahead of another motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 28, 29, 30, 31, 51, 52, 53, 54, 74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 216, 217, 218, 239, 240, 241, 254, 262, 263, 264, 265, 276, 277, 285, 286, 287, 288, 299, 300, 301, 307, 308, 309, 310, 311, 312, 322, 323, 324, 333, 334, 335, 336, 345, 346, 347, 356, 357, 358, 359, 368, 369, 370], "bbox": [0.0032187500000000003, 0.0034375, 0.635734375, 0.9798333333333333], "iscrowd": 0, "area": 62076.75089999997, "rle": {"size": [480, 640], "counts": "V[1`1_=10000O10000O1000O0100O1I7I7H8I7I7I6I8I7I7I7I7H8I7N2000O100000O10000000bLPFl1P:SNRFl1n9TNRFk1o9TNSFk1m9UNSFk1m9TNTFl1l9TNUFk1k9TNVFl1i9UNXFj1h9UNYFk1i7XM[Im0lNk1d7^M_If0oNj1_7dMbIa0POk1Z7hMfI=POk1V7mMlI4POn1P7RNSJMmNQ2l6VNZJElNT2e6\\NcJ\\OhNX2a6aNiJSOgN\\2\\6eNPKlNeN^2W6jNWKdNcNb2R6nN_K\\N_Ne2o5TOeLi0W3[OmLa0P3BTM4o20SMDT3>PMUOV3n0nLfNX3\\1lLWNZ3m1iLfM^3\\2`43N2M4M2M3N3L3N2N2M4M2M3N2O2O0O1O2O0O1O101N1O100O1O10O010O01O010O0001O00001O01O01O00001O001O01O01O001O001O00010O001O001O001O0010O0001O0010O100O10O0100O1O10O0100O100QNZLcHf3Z7_LeHa3W7dLgH]3U7iLiHW3S7nLkHS3S7PMkHP3U7SMiHm2V7UMhHl2V7WMiHi2V7ZMgHg2X7[MgHd2X7_MaHg2^7\\MXHl2g7VMPHR3n7QMiGW3V8i102N1O100O2O0O100O1O2O0O100O2O0O100O2O0O10001O0O10001O0O101O0O2O001O0O2O00001O010O010O010O010O01O010O010O010O010O0010O01O1O1O1O1O1O1O1O2N1O1O1OnKmH8R7HnH:P7ERI;S2nKb2f3\\K>l1PLh2b3\\K?h1SLk2]3_Ka0`1VLQ3Y3_Kb0\\1YLT3T3bKc0U1]LY3P3bKe0P1^L^3l2dKf0i0cLb3f2fKi0;lLo3[2fKi2Z4UMhKj2Y4UMgKl2Y4RMhKn2Z4oLgKQ3Z4nLfKS3Z4kLgKU3Z4jLfKW3Z4gLgKY3Z4eLgK\\3Y4cLgK]3Z4aLgK_3Z4`LfKa3Z4]LgKc3\\4ZLdKg3]4VLdKj3_4RLbKo3_4oKbKP4a4lK`KT4b4jK^KW4d4eK]K[4e4bK\\K_4f4^KZKb4h4[KYKf4i4WKVKj4P5oJQKQ5W5gJhJ[5`5[J`Jf5i5PJWJR6Q6eIoI[6Z6[IfIg6X71O101N1O2N101ZKlH]1U7YNWId1k6YNYId1h6[N[Ic1e6\\N^Ia1d6]N_I`1b6_NaI^1a6`NbI^1^6aNeI\\1^6`NfI]1\\6aNgI`1X6]NkIc1T6[NoIf1P6WNRJk1l5SNWJn1h5oM[JR2c5kMaJV2^5gMdJZ2\\5cMfJ_2X5_MkJb2T5[MnJg2P5WMRKk2m4RMWIjN^1U4Z5nL[IVOS1m3a5jL]I@l0h3e5fL`IGg0e3h5aLbIOc0a3j5]LdI6?^3l5ZLfI9>_3k5ULhI<>a3h5QLkI?>a3f5lKnId0 in this image.", "objects": [{"patches": [5, 6, 7, 8, 28, 29, 30, 31, 51, 52, 53, 54, 74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 216, 217, 218, 239, 240, 241, 254, 262, 263, 264, 265, 276, 277, 285, 286, 287, 288, 299, 300, 301, 307, 308, 309, 310, 311, 312, 322, 323, 324, 333, 334, 335, 336, 345, 346, 347, 356, 357, 358, 359, 368, 369, 370], "bbox": [0.0032187500000000003, 0.0034375, 0.635734375, 0.9798333333333333], "iscrowd": 0, "area": 62076.75089999997, "rle": {"size": [480, 640], "counts": "V[1`1_=10000O10000O1000O0100O1I7I7H8I7I7I6I8I7I7I7I7H8I7N2000O100000O10000000bLPFl1P:SNRFl1n9TNRFk1o9TNSFk1m9UNSFk1m9TNTFl1l9TNUFk1k9TNVFl1i9UNXFj1h9UNYFk1i7XM[Im0lNk1d7^M_If0oNj1_7dMbIa0POk1Z7hMfI=POk1V7mMlI4POn1P7RNSJMmNQ2l6VNZJElNT2e6\\NcJ\\OhNX2a6aNiJSOgN\\2\\6eNPKlNeN^2W6jNWKdNcNb2R6nN_K\\N_Ne2o5TOeLi0W3[OmLa0P3BTM4o20SMDT3>PMUOV3n0nLfNX3\\1lLWNZ3m1iLfM^3\\2`43N2M4M2M3N3L3N2N2M4M2M3N2O2O0O1O2O0O1O101N1O100O1O10O010O01O010O0001O00001O01O01O00001O001O01O01O001O001O00010O001O001O001O0010O0001O0010O100O10O0100O1O10O0100O100QNZLcHf3Z7_LeHa3W7dLgH]3U7iLiHW3S7nLkHS3S7PMkHP3U7SMiHm2V7UMhHl2V7WMiHi2V7ZMgHg2X7[MgHd2X7_MaHg2^7\\MXHl2g7VMPHR3n7QMiGW3V8i102N1O100O2O0O100O1O2O0O100O2O0O100O2O0O10001O0O10001O0O101O0O2O001O0O2O00001O010O010O010O010O01O010O010O010O010O0010O01O1O1O1O1O1O1O1O2N1O1O1OnKmH8R7HnH:P7ERI;S2nKb2f3\\K>l1PLh2b3\\K?h1SLk2]3_Ka0`1VLQ3Y3_Kb0\\1YLT3T3bKc0U1]LY3P3bKe0P1^L^3l2dKf0i0cLb3f2fKi0;lLo3[2fKi2Z4UMhKj2Y4UMgKl2Y4RMhKn2Z4oLgKQ3Z4nLfKS3Z4kLgKU3Z4jLfKW3Z4gLgKY3Z4eLgK\\3Y4cLgK]3Z4aLgK_3Z4`LfKa3Z4]LgKc3\\4ZLdKg3]4VLdKj3_4RLbKo3_4oKbKP4a4lK`KT4b4jK^KW4d4eK]K[4e4bK\\K_4f4^KZKb4h4[KYKf4i4WKVKj4P5oJQKQ5W5gJhJ[5`5[J`Jf5i5PJWJR6Q6eIoI[6Z6[IfIg6X71O101N1O2N101ZKlH]1U7YNWId1k6YNYId1h6[N[Ic1e6\\N^Ia1d6]N_I`1b6_NaI^1a6`NbI^1^6aNeI\\1^6`NfI]1\\6aNgI`1X6]NkIc1T6[NoIf1P6WNRJk1l5SNWJn1h5oM[JR2c5kMaJV2^5gMdJZ2\\5cMfJ_2X5_MkJb2T5[MnJg2P5WMRKk2m4RMWIjN^1U4Z5nL[IVOS1m3a5jL]I@l0h3e5fL`IGg0e3h5aLbIOc0a3j5]LdI6?^3l5ZLfI9>_3k5ULhI<>a3h5QLkI?>a3f5lKnId0 in this image.", "objects": [{"patches": [51, 52, 72, 73, 74, 75, 76, 77, 94, 95, 96, 97, 98, 99, 100, 101, 117, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 262, 277, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.008421874999999999, 0.16854166666666667, 0.593265625, 0.9865208333333333], "iscrowd": 0, "area": 89106.53114999998, "rle": {"size": [480, 640], "counts": "Ub21h>:F;K4M4L1N2O1QIQOKP10TO0m0LVO3l0HXO8i0DZOAFW1l5SOQJ6=AEV1n5PORJ9:CET1o5nNTJ<7CDT1Q6kNUJ?6CCS1S6hNVJe00BER1U6eNWJk0L_OGQ1V6cNYJP1F]OJQ1W6aNZJT1BYONR1V6_N\\JZ1[OVO3Q1V6^N]J^1WORO6S1U6^N^Ja1SOnN:S1U6^N_Jc1nNmN=S1V6^N^Jf1kNhNb0T1T6_N_Jj2]OgNT6_N_Ji2]OiNT6_N_Jg2]OjNS6`N`Je2]OlNS6_N`Je2]OlN[5WOXKm1\\OmNX5[O\\Kf1]OoNT5^O_Kc1]OoNQ5AbK_1]OQOn4DdK[1^OQOj4HiKU1]OTOg4JlKR1]OTOd4MoKo0]OTOa41RLi0]OWO^43ULf0]OWO[46XLb0]OYOY48YL?^OYOV4;]L:^O[OS4=_L8]O\\OR4>aL6]O\\Oo3b0\\L8EWOm3c0WL=LPOj3f0RLa0HVOT4g7M3N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2O100O100O100O100O100O100O10000O100O100O100O100O100O100O100000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O10000000000001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O001O2N1O2N2N2N2N2N2N2N2N2N1O2]KQLQNQ4h1ULWNm3b1XL^Nj3[1[LeNg3W1[LiNg3T1[LkNg3R1ZLnNh3n0ZLROg3l0ZLTOh3i0ZLVOh3f0ZLkKXO\\2`4f1YLjKBY2W4j1YLhKHZ2Q4j1[LgKG_2P4g1\\LeKHc2n3e1]LbKIi2l3a1eM_N]2^1eMaN\\2\\1fMdN\\2X1gMgN^2R1dMnNa2k0aMUOd2d0_M[Of2>\\MBi27ZMHk21WMOT3BPM>Y3SOnLl0R3SNoMm1Q2oMTNP2l1nMVNR2k1lMVNT2j1lMWNS2i1mMWNS2i1mMXNR2h1nMXNR2h1nMXNR2h1nMYNQ2g1oMYNQ2g1oMZNP2f1PNZNP2f1PN[No1e1QN[No1e1RN[Nm1c1UN]Nk1a1WN`Nh1^1[NaNe1]1]NdNb1Z1aNeN_1Y1cNhN\\1V1gNiNY1U1iNlNV1R1lNoNS1o0XNmI3S5e1o0VNRJ4P5d1n0WNUJ5m4c1n0UNYJ7i4d1n0RN]J:e4c1n0PNaJfNZK3Z4T1;jN\\K2[4Q18nN^K1\\4n05RO`K0\\4l03VOaKNb4e0M]ObKMi4>DFdKLo46\\OOfKJW;8iDGW;9jDFU;;mDDR;=nDBR;>oDAQ;?PE@P;`0PEAo:`0QE_Oo:a0QE_On:b0SE]Om:c0SE^Ol:c0SE]Om:c0TE\\Ol:d0TE\\Ol:e0TE[Ok:e0UE[Oj:f0WEYOi:g0WEYOi:h0VEYOi:g0XEXOh:h0XEXOh:i0XEVOh:j0XEWOf:j0[EUOe:k0[EUOe:l0ZETOf:k0\\EUOc:k0]EUOc:k0^ETOb:l0^ETOb:l0_ETO_:m0aESO`:l0`ETO`:l0aETO^:l0bETO^:l0cESO]:m0cESO^:l0cETO\\:l0dETO\\:l0dETO\\:l0m1NiCVOZ:h0P2O1N2O2M6K6I6KWfi3"}, "label": "bike with three lights"}]} {"id": 489695, "image": "COCO_train2014_000000489695.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bike in front left side\"."}], "task": "refering", "answer_template": "The \"bike in front left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 72, 73, 74, 75, 76, 77, 94, 95, 96, 97, 98, 99, 100, 101, 117, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 208, 209, 210, 211, 212, 213, 214, 215, 216, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 254, 255, 256, 257, 258, 259, 260, 261, 262, 277, 278, 279, 280, 281, 282, 283, 284, 285, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379], "bbox": [0.008421874999999999, 0.16854166666666667, 0.593265625, 0.9865208333333333], "iscrowd": 0, "area": 89106.53114999998, "rle": {"size": [480, 640], "counts": "Ub21h>:F;K4M4L1N2O1QIQOKP10TO0m0LVO3l0HXO8i0DZOAFW1l5SOQJ6=AEV1n5PORJ9:CET1o5nNTJ<7CDT1Q6kNUJ?6CCS1S6hNVJe00BER1U6eNWJk0L_OGQ1V6cNYJP1F]OJQ1W6aNZJT1BYONR1V6_N\\JZ1[OVO3Q1V6^N]J^1WORO6S1U6^N^Ja1SOnN:S1U6^N_Jc1nNmN=S1V6^N^Jf1kNhNb0T1T6_N_Jj2]OgNT6_N_Ji2]OiNT6_N_Jg2]OjNS6`N`Je2]OlNS6_N`Je2]OlN[5WOXKm1\\OmNX5[O\\Kf1]OoNT5^O_Kc1]OoNQ5AbK_1]OQOn4DdK[1^OQOj4HiKU1]OTOg4JlKR1]OTOd4MoKo0]OTOa41RLi0]OWO^43ULf0]OWO[46XLb0]OYOY48YL?^OYOV4;]L:^O[OS4=_L8]O\\OR4>aL6]O\\Oo3b0\\L8EWOm3c0WL=LPOj3f0RLa0HVOT4g7M3N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2O100O100O100O100O100O100O10000O100O100O100O100O100O100O100000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O10000000000001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O00001O0000001O00001O001O2N1O2N2N2N2N2N2N2N2N2N1O2]KQLQNQ4h1ULWNm3b1XL^Nj3[1[LeNg3W1[LiNg3T1[LkNg3R1ZLnNh3n0ZLROg3l0ZLTOh3i0ZLVOh3f0ZLkKXO\\2`4f1YLjKBY2W4j1YLhKHZ2Q4j1[LgKG_2P4g1\\LeKHc2n3e1]LbKIi2l3a1eM_N]2^1eMaN\\2\\1fMdN\\2X1gMgN^2R1dMnNa2k0aMUOd2d0_M[Of2>\\MBi27ZMHk21WMOT3BPM>Y3SOnLl0R3SNoMm1Q2oMTNP2l1nMVNR2k1lMVNT2j1lMWNS2i1mMWNS2i1mMXNR2h1nMXNR2h1nMXNR2h1nMYNQ2g1oMYNQ2g1oMZNP2f1PNZNP2f1PN[No1e1QN[No1e1RN[Nm1c1UN]Nk1a1WN`Nh1^1[NaNe1]1]NdNb1Z1aNeN_1Y1cNhN\\1V1gNiNY1U1iNlNV1R1lNoNS1o0XNmI3S5e1o0VNRJ4P5d1n0WNUJ5m4c1n0UNYJ7i4d1n0RN]J:e4c1n0PNaJfNZK3Z4T1;jN\\K2[4Q18nN^K1\\4n05RO`K0\\4l03VOaKNb4e0M]ObKMi4>DFdKLo46\\OOfKJW;8iDGW;9jDFU;;mDDR;=nDBR;>oDAQ;?PE@P;`0PEAo:`0QE_Oo:a0QE_On:b0SE]Om:c0SE^Ol:c0SE]Om:c0TE\\Ol:d0TE\\Ol:e0TE[Ok:e0UE[Oj:f0WEYOi:g0WEYOi:h0VEYOi:g0XEXOh:h0XEXOh:i0XEVOh:j0XEWOf:j0[EUOe:k0[EUOe:l0ZETOf:k0\\EUOc:k0]EUOc:k0^ETOb:l0^ETOb:l0_ETO_:m0aESO`:l0`ETO`:l0aETO^:l0bETO^:l0cESO]:m0cESO^:l0cETO\\:l0dETO\\:l0dETO\\:l0m1NiCVOZ:h0P2O1N2O2M6K6I6KWfi3"}, "label": "bike in front left side"}]} {"id": 424165, "image": "COCO_train2014_000000424165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white shirt and black shorts on a tennis court is about to hit the ball\"."}], "task": "refering", "answer_template": "The \"a man in a white shirt and black shorts on a tennis court is about to hit the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [159, 176, 192, 193, 194, 208, 209, 210, 211, 226, 227, 228, 241, 242, 243, 244, 245, 246, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 293, 294, 295, 296, 297, 309, 310, 311, 312, 313, 314, 325, 326, 327, 328, 329, 330, 342, 345, 346], "bbox": [0.12827083333333333, 0.385953125, 0.48497916666666663, 0.91671875], "iscrowd": 0, "area": 20642.872899999995, "rle": {"size": [640, 480], "counts": "^YW13mc03L3M4L1O1O1O1O1O1O100O1O1O1O1O1O1O00O10001O00dLYO`Cg0`<^O[Ca0fRO\\;R2RDoNa0QO];P2oCQOd0oN^;P2iCUOh0mN^;o1_C^OS1dN^;Z4aDgK`;X4`DiK_;W4`DjKa;U4_DlKa;T4]DmKc;S4\\DoKd;P4\\DQLd;n3[DSLe;l3\\DULd;i3\\DXLd;g3]DZL];j3cDWLW;n3iDTL_8cNPI^5aNQL^8ROYHS5YOlK^8]6bGeI^8Y6bGhI^8W6bGiI_8V6aGZIP9e6PGWIV9g6jFYIW9f6iFZIX9e6iFZIX9e6hF[IZ9c6fF^IZ9a6fF_I[9`6eF`I\\9_6dFaI^9]6bFcI_9\\6aFdI`9]OWFh59lJS;f4mDZKa;X4_DhKo;_4\\CaKR=_5:F3N2M3M3M3M3M3N2M3M3M3M3M3N2M3M3M3M3M3M3N2M3M3M3M3M3M3M3M3M3M3NToi4"}, "label": "a man in a white shirt and black shorts on a tennis court is about to hit the ball"}]} {"id": 424165, "image": "COCO_train2014_000000424165.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man is jumping up to serve a tennis ball\"."}], "task": "refering", "answer_template": "The \"a man is jumping up to serve a tennis ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [159, 176, 192, 193, 194, 208, 209, 210, 211, 226, 227, 228, 241, 242, 243, 244, 245, 246, 258, 259, 260, 261, 262, 263, 277, 278, 279, 280, 293, 294, 295, 296, 297, 309, 310, 311, 312, 313, 314, 325, 326, 327, 328, 329, 330, 342, 345, 346], "bbox": [0.12827083333333333, 0.385953125, 0.48497916666666663, 0.91671875], "iscrowd": 0, "area": 20642.872899999995, "rle": {"size": [640, 480], "counts": "^YW13mc03L3M4L1O1O1O1O1O1O100O1O1O1O1O1O1O00O10001O00dLYO`Cg0`<^O[Ca0fRO\\;R2RDoNa0QO];P2oCQOd0oN^;P2iCUOh0mN^;o1_C^OS1dN^;Z4aDgK`;X4`DiK_;W4`DjKa;U4_DlKa;T4]DmKc;S4\\DoKd;P4\\DQLd;n3[DSLe;l3\\DULd;i3\\DXLd;g3]DZL];j3cDWLW;n3iDTL_8cNPI^5aNQL^8ROYHS5YOlK^8]6bGeI^8Y6bGhI^8W6bGiI_8V6aGZIP9e6PGWIV9g6jFYIW9f6iFZIX9e6iFZIX9e6hF[IZ9c6fF^IZ9a6fF_I[9`6eF`I\\9_6dFaI^9]6bFcI_9\\6aFdI`9]OWFh59lJS;f4mDZKa;X4_DhKo;_4\\CaKR=_5:F3N2M3M3M3M3M3N2M3M3M3M3M3N2M3M3M3M3M3M3N2M3M3M3M3M3M3M3M3M3M3NToi4"}, "label": "a man is jumping up to serve a tennis ball"}]} {"id": 547055, "image": "COCO_train2014_000000547055.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black and white cow behind mostly black cow\"."}], "task": "refering", "answer_template": "The \"black and white cow behind mostly black cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 158, 159, 160, 182, 183, 205, 206, 228, 229], "bbox": [0.82225, 0.1476235294117647, 1.0, 0.6397647058823529], "iscrowd": 0, "area": 12140.674499999996, "rle": {"size": [425, 640], "counts": "lij62W=10O00010O07I?A>C>A>B>C0OO1O1O2N1O2N1O2O0O2N1O20O10O0100O10O0100O0O2O1N101N2O1O1N101N2O1N2O1N2O001N2O1N2O1N2O1O0O2O1N2O1N2O1N101O1N2O1N2O1N2O1O1N2O1N3N1N2O1N2YHWLl5l3fI_LZ6Y5N101N2O0O2O001N2OO0N3N1O2M2O2N1N3N1M4M2QOP10O010O100O100O1001O1OoM"}, "label": "black and white cow behind mostly black cow"}]} {"id": 547055, "image": "COCO_train2014_000000547055.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"part of a white cow with a black spot around it ' s neck area\"."}], "task": "refering", "answer_template": "The \"part of a white cow with a black spot around it ' s neck area\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 158, 159, 160, 182, 183, 205, 206, 228, 229], "bbox": [0.82225, 0.1476235294117647, 1.0, 0.6397647058823529], "iscrowd": 0, "area": 12140.674499999996, "rle": {"size": [425, 640], "counts": "lij62W=10O00010O07I?A>C>A>B>C0OO1O1O2N1O2N1O2O0O2N1O20O10O0100O10O0100O0O2O1N101N2O1O1N101N2O1N2O1N2O001N2O1N2O1N2O1O0O2O1N2O1N2O1N101O1N2O1N2O1N2O1O1N2O1N3N1N2O1N2YHWLl5l3fI_LZ6Y5N101N2O0O2O001N2OO0N3N1O2M2O2N1N3N1M4M2QOP10O010O100O100O1001O1OoM"}, "label": "part of a white cow with a black spot around it ' s neck area"}]} {"id": 104692, "image": "COCO_train2014_000000104692.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an older bird with a yellow beak and a red patch of feathers sitting next to a grey bird on a tree branch\"."}], "task": "refering", "answer_template": "The \"an older bird with a yellow beak and a red patch of feathers sitting next to a grey bird on a tree branch\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 51, 52, 53, 54, 55, 68, 69, 70, 82, 83, 84, 85, 96, 97, 98, 99, 100, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 141, 142, 143, 144, 145, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 201, 202, 203], "bbox": [0.3964018691588785, 0.084046875, 0.7379906542056075, 0.6012968750000001], "iscrowd": 0, "area": 33754.10464999998, "rle": {"size": [640, 428], "counts": "[cZ35ic07J5J7J2M3N1N2YHSO^GOf03kMl0j9YOdGK;[=N3M2K5I8I6I\\fU2"}, "label": "an older bird with a yellow beak and a red patch of feathers sitting next to a grey bird on a tree branch"}]} {"id": 104692, "image": "COCO_train2014_000000104692.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bird with a rounded yellow beak\"."}], "task": "refering", "answer_template": "The \"bird with a rounded yellow beak\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 39, 40, 51, 52, 53, 54, 55, 68, 69, 70, 82, 83, 84, 85, 96, 97, 98, 99, 100, 111, 112, 113, 114, 115, 126, 127, 128, 129, 130, 141, 142, 143, 144, 145, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 186, 187, 188, 189, 190, 201, 202, 203], "bbox": [0.3964018691588785, 0.084046875, 0.7379906542056075, 0.6012968750000001], "iscrowd": 0, "area": 33754.10464999998, "rle": {"size": [640, 428], "counts": "[cZ35ic07J5J7J2M3N1N2YHSO^GOf03kMl0j9YOdGK;[=N3M2K5I8I6I\\fU2"}, "label": "bird with a rounded yellow beak"}]} {"id": 104692, "image": "COCO_train2014_000000104692.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby bird beside its mother\"."}], "task": "refering", "answer_template": "The \"a baby bird beside its mother\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 63, 64, 65, 66, 78, 79, 80, 81, 93, 94, 95, 96, 107, 108, 109, 110, 111, 122, 123, 124, 125, 126, 137, 138, 139, 140, 141, 152, 153, 154, 155, 156, 167, 168, 169, 170, 171, 183, 184, 185, 186, 198, 199, 200, 201, 214, 215, 216, 231, 246], "bbox": [0.1511915887850467, 0.14721875, 0.46411214953271024, 0.7445625], "iscrowd": 0, "area": 29889.28865, "rle": {"size": [640, 428], "counts": "TkX19ec0=D\\Oc0^O^i^4"}, "label": "a baby bird beside its mother"}]} {"id": 104692, "image": "COCO_train2014_000000104692.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"baby bird being fed by mother\"."}], "task": "refering", "answer_template": "The \"baby bird being fed by mother\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [49, 50, 51, 63, 64, 65, 66, 78, 79, 80, 81, 93, 94, 95, 96, 107, 108, 109, 110, 111, 122, 123, 124, 125, 126, 137, 138, 139, 140, 141, 152, 153, 154, 155, 156, 167, 168, 169, 170, 171, 183, 184, 185, 186, 198, 199, 200, 201, 214, 215, 216, 231, 246], "bbox": [0.1511915887850467, 0.14721875, 0.46411214953271024, 0.7445625], "iscrowd": 0, "area": 29889.28865, "rle": {"size": [640, 428], "counts": "TkX19ec0=D\\Oc0^O^i^4"}, "label": "baby bird being fed by mother"}]} {"id": 39159, "image": "COCO_train2014_000000039159.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a tank top on\"."}], "task": "refering", "answer_template": "The \"a woman with a tank top on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [143, 144, 158, 159, 172, 173, 174, 175, 187, 188, 189, 190, 202, 203, 204, 205, 217, 218, 219, 220, 232, 233, 234, 235, 247, 248, 249, 250, 262, 263, 264, 265, 277, 278, 279, 280, 292, 293, 294, 307, 308, 309, 323, 324, 337, 338, 339], "bbox": [0.4697183098591549, 0.39773437500000003, 0.7357042253521126, 0.981046875], "iscrowd": 0, "area": 24756.052699999993, "rle": {"size": [640, 426], "counts": "^_m33hc0RKZ9c7dF^HW9i7fFXHV9m7gFUHT9Q8iFQHU9R8hFPHW9S8eFoGZ9S8cFoG\\9S8aFoG^9S8_FoGa8EcG]8IPH_8JeGX8IPH]80gGQ8IQH\\85gGm7IPH]89gGi8X8[GdGf8[8^GaGc8_8_1O10000O10O100O100O10O0100O010O010O001N2O2M2O2M3N1N3N2M3aFkFP8W9nGkFP8W9nGlFn7W9oGlFo7V9oGlFm7X9QHjFg7^9WHdFb7c9[H`F^7j9]HXF`7n9ZHUFc7U:THmEi7W;M3M2O2I6@a0ZNYChKQ=S4ZC`Kl<\\4_CWKh in this image.", "objects": [{"patches": [143, 144, 158, 159, 172, 173, 174, 175, 187, 188, 189, 190, 202, 203, 204, 205, 217, 218, 219, 220, 232, 233, 234, 235, 247, 248, 249, 250, 262, 263, 264, 265, 277, 278, 279, 280, 292, 293, 294, 307, 308, 309, 323, 324, 337, 338, 339], "bbox": [0.4697183098591549, 0.39773437500000003, 0.7357042253521126, 0.981046875], "iscrowd": 0, "area": 24756.052699999993, "rle": {"size": [640, 426], "counts": "^_m33hc0RKZ9c7dF^HW9i7fFXHV9m7gFUHT9Q8iFQHU9R8hFPHW9S8eFoGZ9S8cFoG\\9S8aFoG^9S8_FoGa8EcG]8IPH_8JeGX8IPH]80gGQ8IQH\\85gGm7IPH]89gGi8X8[GdGf8[8^GaGc8_8_1O10000O10O100O100O10O0100O010O010O001N2O2M2O2M3N1N3N2M3aFkFP8W9nGkFP8W9nGlFn7W9oGlFo7V9oGlFm7X9QHjFg7^9WHdFb7c9[H`F^7j9]HXF`7n9ZHUFc7U:THmEi7W;M3M2O2I6@a0ZNYChKQ=S4ZC`Kl<\\4_CWKh in this image.", "objects": [{"patches": [125, 126, 140, 141, 142, 154, 155, 156, 157, 168, 169, 170, 171, 172, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202, 213, 214, 215, 216, 228, 229, 230, 231, 243, 244, 245, 246, 258, 259, 260, 261, 274, 275, 276, 289, 290, 291, 304, 305, 306, 319, 320, 334, 335, 336], "bbox": [0.19730046948356808, 0.357671875, 0.5057511737089202, 0.9996562499999999], "iscrowd": 0, "area": 31546.075150000004, "rle": {"size": [640, 426], "counts": "_nd15hc05K5K5K4L5J6K4L3M4L3M4L3L4UN]NVAg1a>dN[A]1b>fN]A\\1_>hN^AZ1_>jN^AX1_>kN_AX1\\>mN`AV1]>mN`AW1\\>mNaAU1\\>nNaAU1[>POaAT1X>ROeAQ1Y>ROcAQ1[>RO`AT1]>mN`AW1]>Y2M4L5SCoJa:V5XEVK_:o4YEXKb:m4VE\\Ke:i4kDfKo0lMl8n8jFYGT9k8cF\\GZ9i8^F]G`9h8ZFZGe9b9N2O1N2O1N2O1N101O1N2O1O1O0O10000O10000O2O0O1WOWE[Hi:`7fEUH[:g7kEWHU:d7RFYHo9c7WF[Hi9a7]1M3N200O1O1O1O1O1OcNnHnEQ7n9UIQFj6n9YIQFg6m9]IQFc6m9`IRF`6l9cISF]6k9fISF[6l9hIfESOGV7b:iIfEROFV7b:kIgEPOFU7a:mIjEnNCV7b:nIjEmNBV7b:PJkEW6T:kIlET6R:nInER6P:QJoEo5P:SJoEn5n9UJRFj5l9XJTFi5j9YJUFg5i9\\JVFe5g9^JXFb5f9aJYF`5d9dJYF]5e9fJYF[5e9iJXFY5e9jJZFV5d9f2L5K4L5nKWEcK;j1e:^2WEYKf0l1Z:h2bFlLe9m2n3[Ld@Y2d?aM`@Y2f?dM]@U2i?iMY@R2l?kMX@n1n?PNT@k1Ta0J7J5J7J7J5K6J6K3L4L4L4M3L]aS4"}, "label": "a woman wearing a white shirt and blue skirt holding a wine glass"}]} {"id": 39159, "image": "COCO_train2014_000000039159.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a white shirt\"."}], "task": "refering", "answer_template": "The \"a woman in a white shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 140, 141, 142, 154, 155, 156, 157, 168, 169, 170, 171, 172, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202, 213, 214, 215, 216, 228, 229, 230, 231, 243, 244, 245, 246, 258, 259, 260, 261, 274, 275, 276, 289, 290, 291, 304, 305, 306, 319, 320, 334, 335, 336], "bbox": [0.19730046948356808, 0.357671875, 0.5057511737089202, 0.9996562499999999], "iscrowd": 0, "area": 31546.075150000004, "rle": {"size": [640, 426], "counts": "_nd15hc05K5K5K4L5J6K4L3M4L3M4L3L4UN]NVAg1a>dN[A]1b>fN]A\\1_>hN^AZ1_>jN^AX1_>kN_AX1\\>mN`AV1]>mN`AW1\\>mNaAU1\\>nNaAU1[>POaAT1X>ROeAQ1Y>ROcAQ1[>RO`AT1]>mN`AW1]>Y2M4L5SCoJa:V5XEVK_:o4YEXKb:m4VE\\Ke:i4kDfKo0lMl8n8jFYGT9k8cF\\GZ9i8^F]G`9h8ZFZGe9b9N2O1N2O1N2O1N101O1N2O1O1O0O10000O10000O2O0O1WOWE[Hi:`7fEUH[:g7kEWHU:d7RFYHo9c7WF[Hi9a7]1M3N200O1O1O1O1O1OcNnHnEQ7n9UIQFj6n9YIQFg6m9]IQFc6m9`IRF`6l9cISF]6k9fISF[6l9hIfESOGV7b:iIfEROFV7b:kIgEPOFU7a:mIjEnNCV7b:nIjEmNBV7b:PJkEW6T:kIlET6R:nInER6P:QJoEo5P:SJoEn5n9UJRFj5l9XJTFi5j9YJUFg5i9\\JVFe5g9^JXFb5f9aJYF`5d9dJYF]5e9fJYF[5e9iJXFY5e9jJZFV5d9f2L5K4L5nKWEcK;j1e:^2WEYKf0l1Z:h2bFlLe9m2n3[Ld@Y2d?aM`@Y2f?dM]@U2i?iMY@R2l?kMX@n1n?PNT@k1Ta0J7J5J7J7J5K6J6K3L4L4L4M3L]aS4"}, "label": "a woman in a white shirt"}]} {"id": 30973, "image": "COCO_train2014_000000030973.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"girl in red shirt on top bunk bed staring into camera\"."}], "task": "refering", "answer_template": "The \"girl in red shirt on top bunk bed staring into camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 73, 74, 75, 76, 77, 90, 91, 92, 93, 94, 95, 106, 107, 108, 109, 110, 111, 112, 122, 123, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 145, 146, 147, 157, 158, 159, 160, 161, 162], "bbox": [0.22425, 0.143875, 0.6589791666666667, 0.41821875000000003], "iscrowd": 0, "area": 25232.453699999995, "rle": {"size": [640, 480], "counts": "`fS2R1lb02N3M3M2N3M3M2N3M3N2N1O2N2N1O2N2N2N1O2N2N101N1O2N1O2O0M4L4L3M4L3N3L3M4M2O2M2N3M2O1N2N1O2O1N2N1O2O1N2N2O2O0O1O100O2O0O100O2O0O1O101N01O001O001OM[Kl@e4T?\\Kk@d4V?4N1O2N1O2O0O2N1O1O100O1O2N1O1O100O1O10000000O2O000000000000000O101O00000000000000001O00000000000000001O000O101O1O1N2O001O1N2O1O1O1N2O1O001O1N2O1O1O1O1O1O0O2O1O1O1O1O1O1N2O002M2O2M3N2M3M3N1N3N2M3N2M3N1O2M3N2N2M2O2M3N2N:E2O001N2O001N2O1N101O1dNm]Oh0Tb0iN]^OU1Yb0M3M3L3N3M3M3M3L3N3M3M3MQYV3"}, "label": "girl in red shirt on top bunk bed staring into camera"}]} {"id": 30973, "image": "COCO_train2014_000000030973.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a red shirt\"."}], "task": "refering", "answer_template": "The \"a person in a red shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [57, 58, 59, 73, 74, 75, 76, 77, 90, 91, 92, 93, 94, 95, 106, 107, 108, 109, 110, 111, 112, 122, 123, 124, 125, 126, 127, 128, 129, 139, 140, 141, 142, 143, 144, 145, 146, 147, 157, 158, 159, 160, 161, 162], "bbox": [0.22425, 0.143875, 0.6589791666666667, 0.41821875000000003], "iscrowd": 0, "area": 25232.453699999995, "rle": {"size": [640, 480], "counts": "`fS2R1lb02N3M3M2N3M3M2N3M3N2N1O2N2N1O2N2N2N1O2N2N101N1O2N1O2O0M4L4L3M4L3N3L3M4M2O2M2N3M2O1N2N1O2O1N2N1O2O1N2N2O2O0O1O100O2O0O100O2O0O1O101N01O001O001OM[Kl@e4T?\\Kk@d4V?4N1O2N1O2O0O2N1O1O100O1O2N1O1O100O1O10000000O2O000000000000000O101O00000000000000001O00000000000000001O000O101O1O1N2O001O1N2O1O1O1N2O1O001O1N2O1O1O1O1O1O0O2O1O1O1O1O1O1N2O002M2O2M3N2M3M3N1N3N2M3N2M3N1O2M3N2N2M2O2M3N2N:E2O001N2O001N2O1N101O1dNm]Oh0Tb0iN]^OU1Yb0M3M3L3N3M3M3M3L3N3M3M3MQYV3"}, "label": "a person in a red shirt"}]} {"id": 30973, "image": "COCO_train2014_000000030973.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the gray folded comforter on the ground\"."}], "task": "refering", "answer_template": "The \"the gray folded comforter on the ground\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [204, 205, 206, 207, 221, 222, 223, 224, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.017354166666666667, 0.526046875, 0.9895833333333334, 0.9817656250000001], "iscrowd": 0, "area": 82407.79995000002, "rle": {"size": [640, 480], "counts": "^[5V8i;1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1000000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O004L5K4L5K5K1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O1O1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O0000000000000000000000000000000000000000000000O1N2N2N2N2N2N2N2N2N2N2N2N2O1001O001O001O1O001O001O001O001O1O001O001O001O1O001O000000O10000O100O100O100O10000O100O1QLdMZF\\2f9mMPFT2P:UNgEk1Y:^N^Eb1c:fNSE[1m:nNjDR1V;WO`Dj0`;_OWDa0i;HmC9T in this image.", "objects": [{"patches": [204, 205, 206, 207, 221, 222, 223, 224, 238, 239, 240, 241, 242, 255, 256, 257, 258, 259, 260, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.017354166666666667, 0.526046875, 0.9895833333333334, 0.9817656250000001], "iscrowd": 0, "area": 82407.79995000002, "rle": {"size": [640, 480], "counts": "^[5V8i;1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1O1O1O1O1N2O1O1O1000000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O004L5K4L5K5K1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O1O001O1O001O1O001O1O1O001O1O001O1O001O1O1O001O1O1O1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O2N1O1O0000000000000000000000000000000000000000000000O1N2N2N2N2N2N2N2N2N2N2N2N2O1001O001O001O1O001O001O001O001O1O001O001O001O1O001O000000O10000O100O100O100O10000O100O1QLdMZF\\2f9mMPFT2P:UNgEk1Y:^N^Eb1c:fNSE[1m:nNjDR1V;WO`Dj0`;_OWDa0i;HmC9T in this image.", "objects": [{"patches": [207, 208, 209, 224, 225, 226, 227, 228, 229, 230, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 278, 279, 280, 281, 282, 283, 284, 285, 296, 297, 298, 299, 300, 301, 302, 314, 315, 316, 317, 318, 319, 332, 333, 334, 335, 336], "bbox": [0.18195833333333333, 0.539359375, 0.7785208333333334, 0.8727031249999999], "iscrowd": 0, "area": 31520.7666, "rle": {"size": [640, 480], "counts": "iZg14mc02M3M4L3N2M2N1O2N2O1N2N2N2O1N2N1O2O1N2N1O1O100O1O1O2O0O1O1O100O1O1O1O100O1O1O2O0O1O1O100O1O1O1O100O1O1O2O0O1O1O100O1O1O1O010O1O1O010O1O1O10O01O1O1O10O01O1O100O001O100O1O001O100O1O00100O1O1O010O1O1O1O010O1O0010O01O1O00100O001O010O1O0010O01O1O0010O01O1O010O1O0010O01O1O0010O01O1O010O1O1O2O0O1O2N1O101N1O2N100O2N1O2O0O1O2N1O101N1O2N100O2N1O101N1O2N1O101N1O2N100O2N1O101N1O2N1O101N1O0001N10000O2O0O100O101O0O100O2O000O101N100O10001N100O101O0O100O2O0O1000001O0000001O0000001O0000000O101O000000001O00000000001O00000000001O00000O101O00000000001O00000000001O000000001O000O107Ck0hNojQ2"}, "label": "a girl wearing glasses and a gray shirt"}]} {"id": 30973, "image": "COCO_train2014_000000030973.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with eyeglasses\"."}], "task": "refering", "answer_template": "The \"a woman with eyeglasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [207, 208, 209, 224, 225, 226, 227, 228, 229, 230, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 278, 279, 280, 281, 282, 283, 284, 285, 296, 297, 298, 299, 300, 301, 302, 314, 315, 316, 317, 318, 319, 332, 333, 334, 335, 336], "bbox": [0.18195833333333333, 0.539359375, 0.7785208333333334, 0.8727031249999999], "iscrowd": 0, "area": 31520.7666, "rle": {"size": [640, 480], "counts": "iZg14mc02M3M4L3N2M2N1O2N2O1N2N2N2O1N2N1O2O1N2N1O1O100O1O1O2O0O1O1O100O1O1O1O100O1O1O2O0O1O1O100O1O1O1O100O1O1O2O0O1O1O100O1O1O1O010O1O1O010O1O1O10O01O1O1O10O01O1O100O001O100O1O001O100O1O00100O1O1O010O1O1O1O010O1O0010O01O1O00100O001O010O1O0010O01O1O0010O01O1O010O1O0010O01O1O0010O01O1O010O1O1O2O0O1O2N1O101N1O2N100O2N1O2O0O1O2N1O101N1O2N100O2N1O101N1O2N1O101N1O2N100O2N1O101N1O2N1O101N1O0001N10000O2O0O100O101O0O100O2O000O101N100O10001N100O101O0O100O2O0O1000001O0000001O0000001O0000000O101O000000001O00000000001O00000000001O00000O101O00000000001O00000000001O000000001O000O107Ck0hNojQ2"}, "label": "a woman with eyeglasses"}]} {"id": 334080, "image": "COCO_train2014_000000334080.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe that is standing behind the dried tree\"."}], "task": "refering", "answer_template": "The \"a giraffe that is standing behind the dried tree\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 142, 143, 168, 190, 191, 214], "bbox": [0.169421875, 0.32627634660421545, 0.33042187500000003, 0.6692271662763466], "iscrowd": 0, "area": 1991.8477999999996, "rle": {"size": [427, 640], "counts": "lb]16U=5K1OO10O100000002N2N2N1O2M_O]C>iQ1gBcNgNm0YAYN^1>eNl0\\AZNZ1>eNk0]AZNY1>dNh0cA^NR1?cN;h>e0gAaNo0\\1Z=3iAaNl0[1\\=2kAcNi0Y1\\=4mAbNg0Y1\\=5oAbNd0X1^=3RBeN`0V1^=4UBfNeBoN@b0k=>gBPO_O?l=6QAXOj13YO=m=5SC^OQO;m=5UC@nN9n=5VCBmN6n=3[CHfN3Q>1^CKbN2Q>1_CNaNMR>3`COm=OUB2i=MYB4i`01N2O1O1O0O10O10O10001N100001O1N101N1O2LnWb2"}, "label": "the man running behind the frisbee"}]} {"id": 97563, "image": "COCO_train2014_000000097563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy with adidas sockjs\"."}], "task": "refering", "answer_template": "The \"a guy with adidas sockjs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 98, 99, 113, 114, 128, 144, 145, 157, 158, 159, 173, 174, 187, 188, 202], "bbox": [0.5027803738317757, 0.2533125, 0.6927336448598131, 0.583546875], "iscrowd": 0, "area": 4054.4029499999997, "rle": {"size": [640, 428], "counts": "V[W43lc02N2M3N2N2M3[NBe_O`0Y`0Af_Oa0R>\\OcB6XO`0P>EaBL]Oa0n=HbBI^Oa0m=JcBG]Ob0l=LdBC^Oc0k=MfBB\\Om0hB_O\\O:\\OXO\\>R1kB^OZO8CWOU>U1nB\\OVO;Z><_BZOTO:\\>?`BWOSO:\\>b0aBSOQO<]>c0bBROoN:_>g0bBnNmN<_>i0dBlNkN;`>l0eBhNiN=a>n0eBfNhNQ1gBcNgNm0YAYN^1>eNl0\\AZNZ1>eNk0]AZNY1>dNh0cA^NR1?cN;h>e0gAaNo0\\1Z=3iAaNl0[1\\=2kAcNi0Y1\\=4mAbNg0Y1\\=5oAbNd0X1^=3RBeN`0V1^=4UBfNeBoN@b0k=>gBPO_O?l=6QAXOj13YO=m=5SC^OQO;m=5UC@nN9n=5VCBmN6n=3[CHfN3Q>1^CKbN2Q>1_CNaNMR>3`COm=OUB2i=MYB4i`01N2O1O1O0O10O10O10001N100001O1N101N1O2LnWb2"}, "label": "a guy with adidas sockjs"}]} {"id": 97563, "image": "COCO_train2014_000000097563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white shirt with a red design\"."}], "task": "refering", "answer_template": "The \"a man in a white shirt with a red design\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 69, 70, 77, 78, 79, 83, 84, 85, 90, 91, 92, 93, 94, 95, 96, 98, 105, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 136, 137, 138, 150, 151, 152, 153, 154, 165, 166, 167, 168, 169, 180, 181, 182, 183, 195, 196, 197, 198, 210, 211, 212], "bbox": [0.0, 0.1905, 0.7148598130841121, 0.652546875], "iscrowd": 0, "area": 26586.31355, "rle": {"size": [640, 428], "counts": "k53P5n1l7ZNkGf1V8bN`G`1`8gNXGY1j8fNUG[1k8fNRG[1o8eNPG[1Q9fNlF\\1U9cNjF]1W9dNgF]1Y9dNdF]1]9cNbF]1`9cN^F^1b9bN\\F_1e9bNYF_1g9aNWFa1j9_NTFb1l9^NSFc1m9^NPFd1P:]NlEf1T:ZNhEk1W:VNeEm1[:TNaEo1_:QN^ER2b:oMYEV2f:kMVEX2j:iMREZ2n:gMnD\\2R;eMiD`2V;aMfDb2Z;_MbDd2^;]M^Df2b;[MZDi2e;XMVDl2j;TMdC^3\\i0YATOj>k0XAQOk>n0VAPOl>o0VAmNm>R1UAjNn>U1SAiNo>U1TAgNo>X1RAfNo>Z1SAbNP?]1j1O1O1N2O1O1O1O1O1N2O001O1O1N2O1O1O1O1O1N2O1O1O1O001N2O1O1O100000000000000000O1000000000000000000000000O1N2O1N2O1001O1O001O1O2N3M2N3M2N3M2N`^k02^aTO3N2N3K4L4M2M000000010O00000001O01O0000000O1O2N1O1O1O2O1N2O0011O4L4L3L01O1N100O1001N10001O00001O1Ce\\O4[c0Kg\\O5Yc0Ih\\O7Xc0Gk\\O9^c0L\\\\OLdc03]\\OLcc04]\\OMcc02^\\OMcc02^\\ONac02`\\ON`c01a\\ON`c01a\\OO^c01b\\OO_c00b\\O0gc00g[]2"}, "label": "a man in a white shirt with a red design"}]} {"id": 97563, "image": "COCO_train2014_000000097563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in white trying to catch a frisbie\"."}], "task": "refering", "answer_template": "The \"a man in white trying to catch a frisbie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [62, 63, 69, 70, 77, 78, 79, 83, 84, 85, 90, 91, 92, 93, 94, 95, 96, 98, 105, 106, 107, 108, 109, 110, 111, 121, 122, 123, 124, 136, 137, 138, 150, 151, 152, 153, 154, 165, 166, 167, 168, 169, 180, 181, 182, 183, 195, 196, 197, 198, 210, 211, 212], "bbox": [0.0, 0.1905, 0.7148598130841121, 0.652546875], "iscrowd": 0, "area": 26586.31355, "rle": {"size": [640, 428], "counts": "k53P5n1l7ZNkGf1V8bN`G`1`8gNXGY1j8fNUG[1k8fNRG[1o8eNPG[1Q9fNlF\\1U9cNjF]1W9dNgF]1Y9dNdF]1]9cNbF]1`9cN^F^1b9bN\\F_1e9bNYF_1g9aNWFa1j9_NTFb1l9^NSFc1m9^NPFd1P:]NlEf1T:ZNhEk1W:VNeEm1[:TNaEo1_:QN^ER2b:oMYEV2f:kMVEX2j:iMREZ2n:gMnD\\2R;eMiD`2V;aMfDb2Z;_MbDd2^;]M^Df2b;[MZDi2e;XMVDl2j;TMdC^3\\i0YATOj>k0XAQOk>n0VAPOl>o0VAmNm>R1UAjNn>U1SAiNo>U1TAgNo>X1RAfNo>Z1SAbNP?]1j1O1O1N2O1O1O1O1O1N2O001O1O1N2O1O1O1O1O1N2O1O1O1O001N2O1O1O100000000000000000O1000000000000000000000000O1N2O1N2O1001O1O001O1O2N3M2N3M2N3M2N`^k02^aTO3N2N3K4L4M2M000000010O00000001O01O0000000O1O2N1O1O1O2O1N2O0011O4L4L3L01O1N100O1001N10001O00001O1Ce\\O4[c0Kg\\O5Yc0Ih\\O7Xc0Gk\\O9^c0L\\\\OLdc03]\\OLcc04]\\OMcc02^\\OMcc02^\\ONac02`\\ON`c01a\\ON`c01a\\OO^c01b\\OO_c00b\\O0gc00g[]2"}, "label": "a man in white trying to catch a frisbie"}]} {"id": 97563, "image": "COCO_train2014_000000097563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a gray shirt about to throw a frisbee\"."}], "task": "refering", "answer_template": "The \"a man wearing a gray shirt about to throw a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 96, 97, 98, 101, 110, 111, 112, 113, 115, 116, 124, 125, 126, 127, 128, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 153, 154, 155, 156, 157, 158, 159, 160, 161, 169, 170, 171, 172, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202, 203, 212, 213, 214, 215, 216, 217, 218, 219, 227, 228, 229, 230, 231, 232, 233, 234, 235, 241, 242, 243, 247, 248, 249, 250, 255, 256, 257, 264, 265, 270, 271, 279, 280, 285, 294, 295, 296, 309, 310, 311, 312], "bbox": [0.010070093457943925, 0.21573437499999998, 0.8434345794392524, 0.9033749999999999], "iscrowd": 0, "area": 47682.12394999994, "rle": {"size": [640, 428], "counts": "aP3P1nb0010O1O010O0010O0100O0010O01O10O01O010O10O01O10O01O100O1O010O100O1O010O1O100O00100O10O0100O100O010O100O10O0100O100O010O1O1O1O1N101O1O1N2O1O0O2O1O1O1N2O2N1N2O1O2N1N200O2O0O1O2O0O100O1O100O1O10O01O1O1UMZM\\Dh2c;]MVDe2j;_MQDc2m;cMlC`2Sn:BSE=n:BRE=P;CPEn:BSE?l:@TEc0j:]OVEd0i:[OXEf0f:[OZEf0e:YO[Ei0d:VO]Ek0b:UO^Em0_:SObEP1[:oNfES1X:mNgEV1W:iNjEY1S:hNmE[1P:dNQF^1m9aNSFb1k9^NUFd1h9\\NYFg1d9XN]Fj1a9VN^Fm1_9SNbFo1\\9QNdFQ2`NoLo8o0bHT2[NoLS9l0bHX2VNoLW9j0cHY2RNoL[9g0dH\\2mMoL_9d0eH_2gMPMd9a0dHT3RM]LY:?fHQ4Z7oKfHQ4Z7nKgHR4X7nKhHS4V7oKjHQ4U7oKlHQ4R7PLoHP4P7QLoHP4o6QLRIo3m6RLSIm3l6TLUIk3j6VLVIi3k6WLfIX3Y6iLPJn2P6RMYJe2f5\\McJZ2^5fMkJQ2T5PNUKg1k4YN^K]1a4eNhKR1V4PORLh0k3[O^L<`3FcL6\\3KgL3W3OjL0S33oLJP38RMFl2gMBV2c0iM\\OT2i0kMVOR2o0lMROP2T1nMkNo1Y1QNhNj1\\1VNgNc1^1[NeN_1_1aNdNX1`1hNcNQ1a1oNbNj0c1UO`Ne0c1[O`N>d1A_N9f1F]N3g1M\\NLh14[NEi1;o8O00001O0000001O00001O0000001O00001N2O001O1O1O\\nY1"}, "label": "a man wearing a gray shirt about to throw a frisbee"}]} {"id": 97563, "image": "COCO_train2014_000000097563.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a blue shirt with the numbers 01 printed on the back , preparing to pass a frisbee\"."}], "task": "refering", "answer_template": "The \"a man in a blue shirt with the numbers 01 printed on the back , preparing to pass a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [81, 82, 83, 96, 97, 98, 101, 110, 111, 112, 113, 115, 116, 124, 125, 126, 127, 128, 130, 131, 138, 139, 140, 141, 142, 143, 144, 145, 146, 153, 154, 155, 156, 157, 158, 159, 160, 161, 169, 170, 171, 172, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202, 203, 212, 213, 214, 215, 216, 217, 218, 219, 227, 228, 229, 230, 231, 232, 233, 234, 235, 241, 242, 243, 247, 248, 249, 250, 255, 256, 257, 264, 265, 270, 271, 279, 280, 285, 294, 295, 296, 309, 310, 311, 312], "bbox": [0.010070093457943925, 0.21573437499999998, 0.8434345794392524, 0.9033749999999999], "iscrowd": 0, "area": 47682.12394999994, "rle": {"size": [640, 428], "counts": "aP3P1nb0010O1O010O0010O0100O0010O01O10O01O010O10O01O10O01O100O1O010O100O1O010O1O100O00100O10O0100O100O010O100O10O0100O100O010O1O1O1O1N101O1O1N2O1O0O2O1O1O1N2O2N1N2O1O2N1N200O2O0O1O2O0O100O1O100O1O10O01O1O1UMZM\\Dh2c;]MVDe2j;_MQDc2m;cMlC`2Sn:BSE=n:BRE=P;CPEn:BSE?l:@TEc0j:]OVEd0i:[OXEf0f:[OZEf0e:YO[Ei0d:VO]Ek0b:UO^Em0_:SObEP1[:oNfES1X:mNgEV1W:iNjEY1S:hNmE[1P:dNQF^1m9aNSFb1k9^NUFd1h9\\NYFg1d9XN]Fj1a9VN^Fm1_9SNbFo1\\9QNdFQ2`NoLo8o0bHT2[NoLS9l0bHX2VNoLW9j0cHY2RNoL[9g0dH\\2mMoL_9d0eH_2gMPMd9a0dHT3RM]LY:?fHQ4Z7oKfHQ4Z7nKgHR4X7nKhHS4V7oKjHQ4U7oKlHQ4R7PLoHP4P7QLoHP4o6QLRIo3m6RLSIm3l6TLUIk3j6VLVIi3k6WLfIX3Y6iLPJn2P6RMYJe2f5\\McJZ2^5fMkJQ2T5PNUKg1k4YN^K]1a4eNhKR1V4PORLh0k3[O^L<`3FcL6\\3KgL3W3OjL0S33oLJP38RMFl2gMBV2c0iM\\OT2i0kMVOR2o0lMROP2T1nMkNo1Y1QNhNj1\\1VNgNc1^1[NeN_1_1aNdNX1`1hNcNQ1a1oNbNj0c1UO`Ne0c1[O`N>d1A_N9f1F]N3g1M\\NLh14[NEi1;o8O00001O0000001O00001O0000001O00001N2O001O1O1O\\nY1"}, "label": "a man in a blue shirt with the numbers 01 printed on the back , preparing to pass a frisbee"}]} {"id": 433435, "image": "COCO_train2014_000000433435.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the elephant which is in the centre of the picture\"."}], "task": "refering", "answer_template": "The \"the elephant which is in the centre of the picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [101, 102, 122, 123, 124, 125, 126, 127, 128, 129, 145, 146, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 237, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 263, 264, 266, 267, 286, 287, 289, 290], "bbox": [0.304375, 0.286875, 0.6591875, 0.7218541666666667], "iscrowd": 0, "area": 36497.83645000001, "rle": {"size": [480, 640], "counts": "]ck24i>5K5K6J5J6K5K5K6J5J6M3M6J6J6I7J6J6J7Ii1VN5L4L4L4M0O101N100O101N100O10O0100O100O010O100O10O0100O100O010O100000O01000000O010000000O_NRGQMo8o2eG[L]8d3YHeKi7[4^10O010000O0100000O10O10O10000O1000000O2O000O10000O10000O2O00000OQFmLS8U5K2N1O010O00000000000010O00000000000000001O000O10000O1O1O1O2N1N2L4D in this image.", "objects": [{"patches": [101, 102, 122, 123, 124, 125, 126, 127, 128, 129, 145, 146, 147, 148, 149, 150, 151, 152, 153, 168, 169, 170, 171, 172, 173, 174, 175, 176, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 237, 238, 239, 240, 241, 242, 243, 244, 260, 261, 262, 263, 264, 266, 267, 286, 287, 289, 290], "bbox": [0.304375, 0.286875, 0.6591875, 0.7218541666666667], "iscrowd": 0, "area": 36497.83645000001, "rle": {"size": [480, 640], "counts": "]ck24i>5K5K6J5J6K5K5K6J5J6M3M6J6J6I7J6J6J7Ii1VN5L4L4L4M0O101N100O101N100O10O0100O100O010O100O10O0100O100O010O100000O01000000O010000000O_NRGQMo8o2eG[L]8d3YHeKi7[4^10O010000O0100000O10O10O10000O1000000O2O000O10000O10000O2O00000OQFmLS8U5K2N1O010O00000000000010O00000000000000001O000O10000O1O1O1O2N1N2L4D in this image.", "objects": [{"patches": [31, 32, 33, 34, 35, 36, 37, 38, 54, 55, 56, 57, 58, 59, 60, 61, 62, 78, 79, 80, 81, 82, 83, 84, 85], "bbox": [0.359796875, 0.0709375, 0.7280468750000001, 0.21733333333333332], "iscrowd": 0, "area": 11968.81475, "rle": {"size": [480, 640], "counts": "mZ\\34k>2N2M3N2N2M3N2N2N2O1O1N2O2N1O1O1N2O1O1O1O1N2O1O1O000O2O0000000O100000001N100000000O1000001O000O100000000O10001O00000O10000000001N10000000000O10001O000O10000000000O2O000000000O100000001O0O100000000000001O01O00000000000001O0000000001O000001O000000000000000010O000000000000000001O0001O000000000001O0000000001O0000001O00000O101O00000O2O0000001O0O10001O000O101O00000O2O0000001O0O10001O000O101O0000000O2O0000001N100O101N10000O2O0O100O2O000O101N100O101O0O100O2O0O5L4L3L5L4K5L`_a2"}, "label": "a pizza in the background with ham on it"}]} {"id": 539941, "image": "COCO_train2014_000000539941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza covered in red onions on a wooden platter behind a pepperoni pizza\"."}], "task": "refering", "answer_template": "The \"a pizza covered in red onions on a wooden platter behind a pepperoni pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 33, 34, 35, 36, 37, 38, 54, 55, 56, 57, 58, 59, 60, 61, 62, 78, 79, 80, 81, 82, 83, 84, 85], "bbox": [0.359796875, 0.0709375, 0.7280468750000001, 0.21733333333333332], "iscrowd": 0, "area": 11968.81475, "rle": {"size": [480, 640], "counts": "mZ\\34k>2N2M3N2N2M3N2N2N2O1O1N2O2N1O1O1N2O1O1O1O1N2O1O1O000O2O0000000O100000001N100000000O1000001O000O100000000O10001O00000O10000000001N10000000000O10001O000O10000000000O2O000000000O100000001O0O100000000000001O01O00000000000001O0000000001O000001O000000000000000010O000000000000000001O0001O000000000001O0000000001O0000001O00000O101O00000O2O0000001O0O10001O000O101O00000O2O0000001O0O10001O000O101O0000000O2O0000001N100O101N10000O2O0O100O2O000O101N100O101O0O100O2O0O5L4L3L5L4K5L`_a2"}, "label": "a pizza covered in red onions on a wooden platter behind a pepperoni pizza"}]} {"id": 539941, "image": "COCO_train2014_000000539941.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pizza that has 10 pepperoni on it\"."}], "task": "refering", "answer_template": "The \"a pizza that has 10 pepperoni on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 375, 376, 377, 378, 379, 380, 381], "bbox": [0.0455, 0.2516875, 0.891578125, 0.9842708333333333], "iscrowd": 0, "area": 141484.84355, "rle": {"size": [480, 640], "counts": "Xl=9d>7J7I7H8I7H7J7Ic0\\O:G4L4N1N2N2N101N3M3M2M4L4M3L4M2M4L4M3L4L3M4M3L2N2O1N3M2N2O1N2N2N2M4M2M3M3M3N2M3M4L3L4M3OO`GdI_8]6aGcI^8^6bGbI\\8`6eG`IY8`6hG`IW8a6iG_IV8b622O1O1N20000O10000O1000000O10000O10000O10001N101O001N101O1N101O0O2O001N10001N101O001N100O2N1O100O2N100O1O2O0O1O100O2O0O1O100O1O2N1O100O1O2N100O101N1000001O1N2O001O1O1N2O1O1O1O1O1N2O1O1O1O1O1O1O2N1O1O1O1N3N1O1O1O1O001O1O1N2O1N2O1N2O1N2O1N3N1O001O001N2O001O001O001O001O0O101O001O001O00010O0000000001O01O0000000001O000000000000000000O10000000000000O010000000000000O0100000000O1000000O10000O1000O10O10000000000000000000O10000000000O1000000O1000000O10000O100000000O2O0000O1000000000O10000000O100000O010000O10000O1000O010000O1000000O1000O010000O10000O01000O10000001N2O001O1O1O001O1O0O2O1O1O001O1O0O2O1O1O1O1O1O1N101O1N2O1O1N101N10001O0O2O00001O001O0O101O001O00001N101O001O001N101O001O1O001N101O001O1O001O00001N101O0O101N100O2O000O2O0O101N101N10001N10001N100O2O001N101O0O2O001N101O0O2O001N101O0O2O0O2O001N101O0O2O0O2O001N101N101O0O2O0O2O001N100O2O001N100O2O001N101N101O0O101N101O0O2O0O101O0O2O1N1N3N2N2M3N2N2M3N2M2N4L3M3N3L3M3M3L5K7J5J6K6I6J6K6I;SMhD`1c;SNfDe1] in this image.", "objects": [{"patches": [99, 100, 101, 102, 103, 104, 105, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 375, 376, 377, 378, 379, 380, 381], "bbox": [0.0455, 0.2516875, 0.891578125, 0.9842708333333333], "iscrowd": 0, "area": 141484.84355, "rle": {"size": [480, 640], "counts": "Xl=9d>7J7I7H8I7H7J7Ic0\\O:G4L4N1N2N2N101N3M3M2M4L4M3L4M2M4L4M3L4L3M4M3L2N2O1N3M2N2O1N2N2N2M4M2M3M3M3N2M3M4L3L4M3OO`GdI_8]6aGcI^8^6bGbI\\8`6eG`IY8`6hG`IW8a6iG_IV8b622O1O1N20000O10000O1000000O10000O10000O10001N101O001N101O1N101O0O2O001N10001N101O001N100O2N1O100O2N100O1O2O0O1O100O2O0O1O100O1O2N1O100O1O2N100O101N1000001O1N2O001O1O1N2O1O1O1O1O1N2O1O1O1O1O1O1O2N1O1O1O1N3N1O1O1O1O001O1O1N2O1N2O1N2O1N2O1N3N1O001O001N2O001O001O001O001O0O101O001O001O00010O0000000001O01O0000000001O000000000000000000O10000000000000O010000000000000O0100000000O1000000O10000O1000O10O10000000000000000000O10000000000O1000000O1000000O10000O100000000O2O0000O1000000000O10000000O100000O010000O10000O1000O010000O1000000O1000O010000O10000O01000O10000001N2O001O1O1O001O1O0O2O1O1O001O1O0O2O1O1O1O1O1O1N101O1N2O1O1N101N10001O0O2O00001O001O0O101O001O00001N101O001O001N101O001O1O001N101O001O1O001O00001N101O0O101N100O2O000O2O0O101N101N10001N10001N100O2O001N101O0O2O001N101O0O2O001N101O0O2O0O2O001N101O0O2O0O2O001N101N101O0O2O0O2O001N100O2O001N100O2O001N101N101O0O101N101O0O2O0O101O0O2O1N1N3N2N2M3N2N2M3N2M2N4L3M3N3L3M3M3L5K7J5J6K6I6J6K6I;SMhD`1c;SNfDe1] in this image.", "objects": [{"patches": [73, 74, 94, 95, 96, 97, 115, 116, 117, 118, 119, 137, 138, 139, 140, 141, 160, 161, 162, 163, 181, 182, 183, 184, 203, 204, 205, 225, 226, 227], "bbox": [0.21500821018062397, 0.14825, 0.44594417077175696, 0.45620312500000004], "iscrowd": 0, "area": 13400.0331, "rle": {"size": [640, 609], "counts": "Web22nc00f_OOe<2XC0h<0WC1h<1UC2j\\=CbB>^=B`B`0_=B_B?a=B]B`0b=@]Ba0c=@[Ba0d=@[Ba0e=@XBc0g=]OXBd0h=]OVBc0j=_OTBa0l=@SB`0m=BQB>P>BoA>Q>DmAElA;U>DkADkAEjA;V>EjA;W>DhAEfA;Z>EfA;[>DeA<[>EdA;\\>EdA;\\>EdA;iNhNT?m0RB;iNjNT?k0SB;iNjNT?k0SB;hNlNU?i0RB;hNnNU?g0SB;gNPOU?e0SBWB7fN]OP?>YB5gN^On>`0VB6kN\\Ol>`0UB8nNZOj>a0SB9SOWOg>b0RB:WOVOd>c0PB;[OTOb>c0oA=^OQOa>e0kA?CnN^>g0jA>IlN]>e0fAc0LjN]>\\4OO101N1O101N1O101N1O101N1O101N1O101N3M2O2N2M2O2M2O2N2M2O2N2M2O2M3N1O2M3N1N3N2N1N3N2M3N2N2M4M2M3N2N2M3N2M4O01O1O1O1O100O1O1O11O0O1J6I8I6I7D in this image.", "objects": [{"patches": [73, 74, 94, 95, 96, 97, 115, 116, 117, 118, 119, 137, 138, 139, 140, 141, 160, 161, 162, 163, 181, 182, 183, 184, 203, 204, 205, 225, 226, 227], "bbox": [0.21500821018062397, 0.14825, 0.44594417077175696, 0.45620312500000004], "iscrowd": 0, "area": 13400.0331, "rle": {"size": [640, 609], "counts": "Web22nc00f_OOe<2XC0h<0WC1h<1UC2j\\=CbB>^=B`B`0_=B_B?a=B]B`0b=@]Ba0c=@[Ba0d=@[Ba0e=@XBc0g=]OXBd0h=]OVBc0j=_OTBa0l=@SB`0m=BQB>P>BoA>Q>DmAElA;U>DkADkAEjA;V>EjA;W>DhAEfA;Z>EfA;[>DeA<[>EdA;\\>EdA;\\>EdA;iNhNT?m0RB;iNjNT?k0SB;iNjNT?k0SB;hNlNU?i0RB;hNnNU?g0SB;gNPOU?e0SBWB7fN]OP?>YB5gN^On>`0VB6kN\\Ol>`0UB8nNZOj>a0SB9SOWOg>b0RB:WOVOd>c0PB;[OTOb>c0oA=^OQOa>e0kA?CnN^>g0jA>IlN]>e0fAc0LjN]>\\4OO101N1O101N1O101N1O101N1O101N1O101N3M2O2N2M2O2M2O2N2M2O2N2M2O2M3N1O2M3N1N3N2N1N3N2M3N2N2M4M2M3N2N2M3N2M4O01O1O1O1O100O1O1O11O0O1J6I8I6I7D in this image.", "objects": [{"patches": [269, 270, 291, 292, 293, 312, 313, 314, 315, 333, 334, 335, 336, 337, 355, 356, 357, 358, 359, 360, 380, 381, 382, 383, 402, 403, 404, 405], "bbox": [0.16755336617405583, 0.523578125, 0.44628899835796393, 0.8192812499999998], "iscrowd": 0, "area": 12666.687549999999, "rle": {"size": [640, 609], "counts": "PVP24hc07L4L1O2N1O1O000N3G8O2M2O1O1N2O1O1O1O1O10O11O0000001O00001O001O001N10O1O1O01QOdNZ_O]1c`0lNV_OT1h`0UOi^O]O2^1Ra0Bj^O:[a0Hc^O7^a0Y1K5m^O`M``0Q3O2N2N2N2N4L2N2N10001o_OULk?j3T@WLl?i3R@YLn?m31N10003M3L2O0001N2N2O100O1O010N2O0O2O0O3M1O101N1O1O1O1O10O0100O0100O100001O2N1N^Lb@b2^?]Md@a2]?_Md@`2\\?`Me@^2]?aMd@^2]?bMd@V1KQOa?Ie@V1MmN^?Mf@V1MkN]?0f@U1LkN^?0h@T1JkN_?0h@U1JjN]?2j@VOM_1NVO[?5\\Ab0ZOWO[?5`Aa0VOYOZ?5dA?SOZOY?6iA=oN[OY?6lA in this image.", "objects": [{"patches": [142, 143, 165, 166, 187, 188, 189, 210, 211, 212, 232, 233, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260, 278, 279, 280, 281, 301, 302, 303, 304, 325, 326, 327, 348, 349, 370, 371, 372], "bbox": [0.10785937500000001, 0.4, 0.391015625, 0.9887708333333334], "iscrowd": 0, "area": 19454.927999999996, "rle": {"size": [480, 640], "counts": "UeP1=a>d0\\O4gElNa6W1\\IjNb6Z1ZIhNd6\\1YIeNe6_1XIbNg6a1UI`Nj6c1TI^Nj6f1SI[Nk6i1QIYNn6j1oHWNo6m1mHUNP7P2mHQNo6U2nHlMm6Z2PIhMl6^2QIcMj6d2SI]Mi6i2SIYMi6m2TITMg6S3VInLf6X3VIjLe6]3XIdLd6d3VI^Lf6j3UIWLh6Q4RIPLm6U4mHmKQ7Y4jHhKU7^4eHcKY7c4aH_K^7g4\\HZKc7R6N2O1N2O1N2000000000000O10000000000VOj0N2O1OO001N101O1O0O2O2N2M3N2N2N2M3N1O2M3N2N2N2M3N2TNYH]Li7_3dHUL`7f3lHPLV7k3WIkKk6Q4bIdK`6X4lI]KX6^4W2K5L4L5K4L4L4010O0010nMQE2n:F]E7c:AhE=W:[OTFb0m9UO_Fh0`9QOjFl0V9lNUGR1_;N2M4M2M3N0O100O01000O010O10O0O2O0O2OO010O010O01O010O010O010O0010O010O010O010O01O010O010O0100O010O100O100O010O100O100O010O100O100O010O100O1N2N101N2N2O1N2N101N2N2O1NZPg5"}, "label": "a person in a black sweatshirt"}]} {"id": 15658, "image": "COCO_train2014_000000015658.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in black on a court\"."}], "task": "refering", "answer_template": "The \"a person in black on a court\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 165, 166, 187, 188, 189, 210, 211, 212, 232, 233, 234, 235, 236, 237, 238, 255, 256, 257, 258, 259, 260, 278, 279, 280, 281, 301, 302, 303, 304, 325, 326, 327, 348, 349, 370, 371, 372], "bbox": [0.10785937500000001, 0.4, 0.391015625, 0.9887708333333334], "iscrowd": 0, "area": 19454.927999999996, "rle": {"size": [480, 640], "counts": "UeP1=a>d0\\O4gElNa6W1\\IjNb6Z1ZIhNd6\\1YIeNe6_1XIbNg6a1UI`Nj6c1TI^Nj6f1SI[Nk6i1QIYNn6j1oHWNo6m1mHUNP7P2mHQNo6U2nHlMm6Z2PIhMl6^2QIcMj6d2SI]Mi6i2SIYMi6m2TITMg6S3VInLf6X3VIjLe6]3XIdLd6d3VI^Lf6j3UIWLh6Q4RIPLm6U4mHmKQ7Y4jHhKU7^4eHcKY7c4aH_K^7g4\\HZKc7R6N2O1N2O1N2000000000000O10000000000VOj0N2O1OO001N101O1O0O2O2N2M3N2N2N2M3N1O2M3N2N2N2M3N2TNYH]Li7_3dHUL`7f3lHPLV7k3WIkKk6Q4bIdK`6X4lI]KX6^4W2K5L4L5K4L4L4010O0010nMQE2n:F]E7c:AhE=W:[OTFb0m9UO_Fh0`9QOjFl0V9lNUGR1_;N2M4M2M3N0O100O01000O010O10O0O2O0O2OO010O010O01O010O010O010O0010O010O010O010O01O010O010O0100O010O100O100O010O100O100O010O100O100O010O100O1N2N101N2N2O1N2N101N2N2O1NZPg5"}, "label": "a person in black on a court"}]} {"id": 48432, "image": "COCO_train2014_000000048432.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bag with the nike swoosh on it\"."}], "task": "refering", "answer_template": "The \"the bag with the nike swoosh on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 189, 190, 191, 202, 203, 204, 205, 206, 207, 216, 217, 218, 219, 220, 221, 222, 223, 231, 232, 233, 234, 235, 236, 237, 238, 246, 247, 248, 249, 250, 251, 252, 253, 262, 263, 264, 265, 266, 267, 268], "bbox": [0.4559953161592506, 0.47490625, 0.925175644028103, 0.78478125], "iscrowd": 0, "area": 25284.952350000007, "rle": {"size": [640, 427], "counts": "lXj3a1[b05K5K5K5K5K4L5K5M3M3N2N1N3N2M3N2N2M201O0O10001O000000000000001O0000000000000O2O000000000000001O0O100000001O0000000O101O0O10000O2O0O1O1N2O1M4K4M3M3M3M4L3L4N2O1N2N3M2N2N2O1N3M2N2N2N21O105J9G7I8H7I8H7J2M0000gNQLaBo3^=TLaBk3^=WLaBi3]=[LaBe3^=^L`Bb3^=bLaB^3]=dLbB\\3]=gLaBY3^=jL`BV3b=jL]BU3d=kL[BU3f=QMSBo2n=\\MfAd2[>`1100O1O1O1O1O2N1O2N2O1N2N2N1O2N2N2N2N2O1N2N2N2N3M2N2N3N1N2N2N3M2N010O1O001O0010O01O1O001O00100O001O001O010O001O001O010O00001O001O010O001O00001O001O010O1O2N3M2N2H8B?]Ob0[Oecc0"}, "label": "the bag with the nike swoosh on it"}]} {"id": 48432, "image": "COCO_train2014_000000048432.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bule bag in man hand\"."}], "task": "refering", "answer_template": "The \"bule bag in man hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [174, 175, 176, 189, 190, 191, 202, 203, 204, 205, 206, 207, 216, 217, 218, 219, 220, 221, 222, 223, 231, 232, 233, 234, 235, 236, 237, 238, 246, 247, 248, 249, 250, 251, 252, 253, 262, 263, 264, 265, 266, 267, 268], "bbox": [0.4559953161592506, 0.47490625, 0.925175644028103, 0.78478125], "iscrowd": 0, "area": 25284.952350000007, "rle": {"size": [640, 427], "counts": "lXj3a1[b05K5K5K5K5K4L5K5M3M3N2N1N3N2M3N2N2M201O0O10001O000000000000001O0000000000000O2O000000000000001O0O100000001O0000000O101O0O10000O2O0O1O1N2O1M4K4M3M3M3M4L3L4N2O1N2N3M2N2N2O1N3M2N2N2N21O105J9G7I8H7I8H7J2M0000gNQLaBo3^=TLaBk3^=WLaBi3]=[LaBe3^=^L`Bb3^=bLaB^3]=dLbB\\3]=gLaBY3^=jL`BV3b=jL]BU3d=kL[BU3f=QMSBo2n=\\MfAd2[>`1100O1O1O1O1O2N1O2N2O1N2N2N1O2N2N2N2N2O1N2N2N2N3M2N2N3N1N2N2N3M2N010O1O001O0010O01O1O001O00100O001O001O010O001O001O010O00001O001O010O001O00001O001O010O1O2N3M2N2H8B?]Ob0[Oecc0"}, "label": "bule bag in man hand"}]} {"id": 253238, "image": "COCO_train2014_000000253238.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde - haired woman ' s head\"."}], "task": "refering", "answer_template": "The \"a blonde - haired woman ' s head\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 26, 27, 28, 29, 30, 31, 32, 33, 49, 50, 51, 52, 53, 54, 55, 56, 72, 73, 74, 75, 76, 77, 78, 79, 80, 95, 96, 97, 98, 99, 100, 101, 102, 103, 118, 119, 120, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 162, 163], "bbox": [0.066953125, 0.0, 0.50990625, 0.4469791666666667], "iscrowd": 0, "area": 40613.14314999999, "rle": {"size": [480, 640], "counts": "ajd02m>100O1O100O101N100O100O100O101N100O101N1N2M4M2M3N2M3O10O1N2O001N10O1000000010O0000bM1mEOb8_OlFl0=E`8P2]GPN]8X2bGhM[8]2bGcM]8`2bG`M]8c2`G^M^8e2aG[M^8h2_GYM]8m2aGSM[8T3aGmL[8Z3aGgL[8`3bG`L]8a3bG`L]8b3aG_L^8c3`G^L_8c3aG]L_8d3_G\\Lb8e3\\G\\Ld8f3YG[Lg8g3VGZLj8n400000O10000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000O10000000000O100000000O10000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000O1000000O10000O1000000O10000O10000000000O100000000O100000000O100000000000000000000000000000000000000000000VKjFj3V9ULlFj3T9ULoFi3Q9VLRGh3n8WLUGg3m8VLVGh3m8TLVGj3l8SLWGk3l8QL]Gg3e8VL^Gh3d8ULaGg3R:J3M3M3M2N3M2N2N1QOSDiNP in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 26, 27, 28, 29, 30, 31, 32, 33, 49, 50, 51, 52, 53, 54, 55, 56, 72, 73, 74, 75, 76, 77, 78, 79, 80, 95, 96, 97, 98, 99, 100, 101, 102, 103, 118, 119, 120, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 162, 163], "bbox": [0.066953125, 0.0, 0.50990625, 0.4469791666666667], "iscrowd": 0, "area": 40613.14314999999, "rle": {"size": [480, 640], "counts": "ajd02m>100O1O100O101N100O100O100O101N100O101N1N2M4M2M3N2M3O10O1N2O001N10O1000000010O0000bM1mEOb8_OlFl0=E`8P2]GPN]8X2bGhM[8]2bGcM]8`2bG`M]8c2`G^M^8e2aG[M^8h2_GYM]8m2aGSM[8T3aGmL[8Z3aGgL[8`3bG`L]8a3bG`L]8b3aG_L^8c3`G^L_8c3aG]L_8d3_G\\Lb8e3\\G\\Ld8f3YG[Lg8g3VGZLj8n400000O10000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000O10000000000O100000000O10000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000O1000000O10000O1000000O10000O10000000000O100000000O100000000O100000000000000000000000000000000000000000000VKjFj3V9ULlFj3T9ULoFi3Q9VLRGh3n8WLUGg3m8VLVGh3m8TLVGj3l8SLWGk3l8QL]Gg3e8VL^Gh3d8ULaGg3R:J3M3M3M2N3M2N2N1QOSDiNP in this image.", "objects": [{"patches": [59, 60, 61, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 103, 114, 115, 116, 117, 118, 119, 120, 121, 133, 134, 135, 136, 137, 138, 139, 151, 152, 153, 154, 155, 156, 157, 169, 170, 171, 174, 175, 187, 188, 189, 192, 193], "bbox": [0.31577999999999995, 0.3011111111111111, 0.77526, 0.9145945945945946], "iscrowd": 0, "area": 20332.254100000006, "rle": {"size": [333, 500], "counts": "e_c14T:6K5N200SGFf7:THLl74nG2R8NgG:X8FbG`0^8@\\Gg0c8b0002O1N2ON1M3M4M2M4L10M2M6J500100O2N100O1O1N3N1N2N2O2M3M3N2M3MTNZHU1d7jN]HX10eNW7DlH^2b70O010O10O01000O0100O01000O010TJ_Mj3b2QLcMo3\\2lKkMR4V2hKPNX4P2bKVN^4i1]K^Na4c1YKbNh4^1mJmNR5W31O10O01O010O010O1O010O0001O010O000000J6I7J7]NjIfN]6Y1gI^N_6b1dIoNh5Q1[JlNe5S1_JkN`5U1cJhN]5Y1eJeN[5Z1hJcNX5]1kJ`NU5`1nJ^NQ5b1TKYNl4g1\\KRNc4n1fKiMZ4W2jKfMU4Z2lKeMU4Z2mKeMR4[2oKdMQ4\\2PLdMo3\\2RLcMn3]2SLbMm3^2TLbMk3_2UL`Ml3_2VL`Mi3`2R20000000001O00000gMTHQ2l7mMVHS2j7kMXHU2h7jMYHV2o71O01O0000000O1000O001O1O1O001O1O001O1O1O010000O1000000O0100000O10000O1000000O10000O0100000O1000000O10000O100000O4M3M3M4L3M3M3M3M4M2M3N2M4M2N2M3N2M4M2M3N2M3N3L3NO010O10O010O10eMTJ3l5K\\JOe5McJJa54gJB^5;kJ[OX5b0^KgNg4V1Y2N3L3N3M3L3N3M2M4M7I8G8IhWT1"}, "label": "zebra in front"}]} {"id": 523577, "image": "COCO_train2014_000000523577.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra which is in front of ua and facing to us\"."}], "task": "refering", "answer_template": "The \"the zebra which is in front of ua and facing to us\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 103, 114, 115, 116, 117, 118, 119, 120, 121, 133, 134, 135, 136, 137, 138, 139, 151, 152, 153, 154, 155, 156, 157, 169, 170, 171, 174, 175, 187, 188, 189, 192, 193], "bbox": [0.31577999999999995, 0.3011111111111111, 0.77526, 0.9145945945945946], "iscrowd": 0, "area": 20332.254100000006, "rle": {"size": [333, 500], "counts": "e_c14T:6K5N200SGFf7:THLl74nG2R8NgG:X8FbG`0^8@\\Gg0c8b0002O1N2ON1M3M4M2M4L10M2M6J500100O2N100O1O1N3N1N2N2O2M3M3N2M3MTNZHU1d7jN]HX10eNW7DlH^2b70O010O10O01000O0100O01000O010TJ_Mj3b2QLcMo3\\2lKkMR4V2hKPNX4P2bKVN^4i1]K^Na4c1YKbNh4^1mJmNR5W31O10O01O010O010O1O010O0001O010O000000J6I7J7]NjIfN]6Y1gI^N_6b1dIoNh5Q1[JlNe5S1_JkN`5U1cJhN]5Y1eJeN[5Z1hJcNX5]1kJ`NU5`1nJ^NQ5b1TKYNl4g1\\KRNc4n1fKiMZ4W2jKfMU4Z2lKeMU4Z2mKeMR4[2oKdMQ4\\2PLdMo3\\2RLcMn3]2SLbMm3^2TLbMk3_2UL`Ml3_2VL`Mi3`2R20000000001O00000gMTHQ2l7mMVHS2j7kMXHU2h7jMYHV2o71O01O0000000O1000O001O1O1O001O1O001O1O1O010000O1000000O0100000O10000O1000000O10000O0100000O1000000O10000O100000O4M3M3M4L3M3M3M3M4M2M3N2M4M2N2M3N2M4M2M3N2M3N3L3NO010O10O010O10eMTJ3l5K\\JOe5McJJa54gJB^5;kJ[OX5b0^KgNg4V1Y2N3L3N3M3L3N3M2M4M7I8G8IhWT1"}, "label": "the zebra which is in front of ua and facing to us"}]} {"id": 523577, "image": "COCO_train2014_000000523577.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra resting its head on another zebra ' s back\"."}], "task": "refering", "answer_template": "The \"a zebra resting its head on another zebra ' s back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 80, 81, 82, 93, 94, 95, 98, 99, 100, 111, 112, 113, 114, 129, 130, 131, 132, 133, 147, 148, 150, 151, 165, 166, 167, 184], "bbox": [0.16240000000000002, 0.27315315315315314, 0.60342, 0.8793993993993994], "iscrowd": 0, "area": 11584.254649999999, "rle": {"size": [333, 500], "counts": "W`j01[:1O2M3N2N1O2M3N2N1O2M3O1N1O2N3M3N1N3M3M3M2N3N2fH_Nc5e1TJgNe5\\1SJoNg5T1SJVOg5m0SJ\\Of5g0UJAf5g0oIDi5c2LO2O1N2N1O2N2oNP13N1O2N2M3N1O2M3N1O2N2M6K6J6I6K2N1OYNiJZNU5_1VK^Ni4`1]K^Nb4_1dKWNb4g1dKoMc4n1cKhMc4V2l1M2N3O010O01O10O010O0100O010O1O010O010O13L5L2N2M3N2M2O00O001O00000000000O010000000001N2O2N1O1O2M2O1O3M4L4K5L4L2N2N2M3N2N2N2N2N2M3N2NS\\80lcG2M2O2M3M2O2M2O2M2N3M3M2N3L3N2O1O1O1O100O1O1O100O1O1O1O10O01O1O1O2N1O1O1O1O1O1O1O1O2O00004MO0001O001O02N2O1N2N2O1N2O1M3L5L5J5K5L4G9@jXP2"}, "label": "a zebra resting its head on another zebra ' s back"}]} {"id": 523577, "image": "COCO_train2014_000000523577.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a zebra with it ' s head draped over the back of another zebra\"."}], "task": "refering", "answer_template": "The \"a zebra with it ' s head draped over the back of another zebra\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 80, 81, 82, 93, 94, 95, 98, 99, 100, 111, 112, 113, 114, 129, 130, 131, 132, 133, 147, 148, 150, 151, 165, 166, 167, 184], "bbox": [0.16240000000000002, 0.27315315315315314, 0.60342, 0.8793993993993994], "iscrowd": 0, "area": 11584.254649999999, "rle": {"size": [333, 500], "counts": "W`j01[:1O2M3N2N1O2M3N2N1O2M3O1N1O2N3M3N1N3M3M3M2N3N2fH_Nc5e1TJgNe5\\1SJoNg5T1SJVOg5m0SJ\\Of5g0UJAf5g0oIDi5c2LO2O1N2N1O2N2oNP13N1O2N2M3N1O2M3N1O2N2M6K6J6I6K2N1OYNiJZNU5_1VK^Ni4`1]K^Nb4_1dKWNb4g1dKoMc4n1cKhMc4V2l1M2N3O010O01O10O010O0100O010O1O010O010O13L5L2N2M3N2M2O00O001O00000000000O010000000001N2O2N1O1O2M2O1O3M4L4K5L4L2N2N2M3N2N2N2N2N2M3N2NS\\80lcG2M2O2M3M2O2M2O2M2N3M3M2N3L3N2O1O1O1O100O1O1O100O1O1O1O10O01O1O1O2N1O1O1O1O1O1O1O1O2O00004MO0001O001O02N2O1N2N2O1N2O1M3L5L5J5K5L4G9@jXP2"}, "label": "a zebra with it ' s head draped over the back of another zebra"}]} {"id": 7504, "image": "COCO_train2014_000000007504.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a clock on a wall calling the time in bac kinh as 8 : 30\"."}], "task": "refering", "answer_template": "The \"a clock on a wall calling the time in bac kinh as 8 : 30\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 180, 181, 182, 183], "bbox": [0.8057343749999999, 0.23677725118483411, 0.987578125, 0.5170379146919432], "iscrowd": 0, "area": 10982.377500000004, "rle": {"size": [422, 640], "counts": "mhd6R1R<5L3M3M4K4M3M4K4M4L3M3L4M2N2M3O1N2O1N2O1N2O1N2O2M2O1N2O1N2N2O0O2O000O100000000O2O000000000O10001O000O100000000010O000000000000001O00000O1000001O0000001O1O001N101O1N1O2O1N101N1O2O1N1O2O1N101N2N101N2N1O2M3N1N3N2N1N5L4K5L4K6K4L4K5J<^OlT3"}, "label": "a clock on a wall calling the time in bac kinh as 8 : 30"}]} {"id": 7504, "image": "COCO_train2014_000000007504.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the clock for bac kinh\"."}], "task": "refering", "answer_template": "The \"the clock for bac kinh\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [88, 89, 90, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 180, 181, 182, 183], "bbox": [0.8057343749999999, 0.23677725118483411, 0.987578125, 0.5170379146919432], "iscrowd": 0, "area": 10982.377500000004, "rle": {"size": [422, 640], "counts": "mhd6R1R<5L3M3M4K4M3M4K4M4L3M3L4M2N2M3O1N2O1N2O1N2O1N2O2M2O1N2O1N2N2O0O2O000O100000000O2O000000000O10001O000O100000000010O000000000000001O00000O1000001O0000001O1O001N101O1N1O2O1N101N1O2O1N1O2O1N101N2N101N2N1O2M3N1N3N2N1N5L4K5L4K6K4L4K5J<^OlT3"}, "label": "the clock for bac kinh"}]} {"id": 7504, "image": "COCO_train2014_000000007504.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a round wall clock showing london time next to two other clocks showing times for other cities\"."}], "task": "refering", "answer_template": "The \"a round wall clock showing london time next to two other clocks showing times for other cities\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 166, 167], "bbox": [0.13490625, 0.22708530805687202, 0.334265625, 0.5269194312796209], "iscrowd": 0, "area": 12565.039000000002, "rle": {"size": [422, 640], "counts": "PcS14iAX]_5"}, "label": "a round wall clock showing london time next to two other clocks showing times for other cities"}]} {"id": 7504, "image": "COCO_train2014_000000007504.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"clock on wall showing 12 : 21\"."}], "task": "refering", "answer_template": "The \"clock on wall showing 12 : 21\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [73, 74, 75, 95, 96, 97, 98, 99, 118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 166, 167], "bbox": [0.13490625, 0.22708530805687202, 0.334265625, 0.5269194312796209], "iscrowd": 0, "area": 12565.039000000002, "rle": {"size": [422, 640], "counts": "PcS14iAX]_5"}, "label": "clock on wall showing 12 : 21"}]} {"id": 7504, "image": "COCO_train2014_000000007504.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black , brown , gold , and white clock that says it ' s 7 : 30 in ha noi\"."}], "task": "refering", "answer_template": "The \"a black , brown , gold , and white clock that says it ' s 7 : 30 in ha noi\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 172, 173, 174, 175], "bbox": [0.47307812499999996, 0.22623222748815167, 0.66796875, 0.5133649289099526], "iscrowd": 0, "area": 11949.1892, "rle": {"size": [422, 640], "counts": "[Pm3d0^<9F8I7J5N3M3M2M3M3M3M2M4M3M3M2M4M2N2N1O2M3N1O2N2N1O2N2O001N2O000O2O0O10001N10000O2O0O10001N100O10000O10000O100O1001O0000000000001O000000000001O01O00000000000O2O0O10000O2O0O2O1N101O0O2O0O2O1N101N101N1O2N1O2O1N1O2N3M2O2M3M2N3M2N3L3L5L3M5K5K5K5J6K5E;^OaWg2"}, "label": "a black , brown , gold , and white clock that says it ' s 7 : 30 in ha noi"}]} {"id": 7504, "image": "COCO_train2014_000000007504.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a clock on a pink wall with the characters ' ha noi ' printed beneath it\"."}], "task": "refering", "answer_template": "The \"a clock on a pink wall with the characters ' ha noi ' printed beneath it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 82, 83, 103, 104, 105, 106, 107, 125, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 172, 173, 174, 175], "bbox": [0.47307812499999996, 0.22623222748815167, 0.66796875, 0.5133649289099526], "iscrowd": 0, "area": 11949.1892, "rle": {"size": [422, 640], "counts": "[Pm3d0^<9F8I7J5N3M3M2M3M3M3M2M4M3M3M2M4M2N2N1O2M3N1O2N2N1O2N2O001N2O000O2O0O10001N10000O2O0O10001N100O10000O10000O100O1001O0000000000001O000000000001O01O00000000000O2O0O10000O2O0O2O1N101O0O2O0O2O1N101N101N1O2N1O2O1N1O2N3M2O2M3M2N3M2N3L3L5L3M5K5K5K5J6K5E;^OaWg2"}, "label": "a clock on a pink wall with the characters ' ha noi ' printed beneath it"}]} {"id": 220504, "image": "COCO_train2014_000000220504.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"female child riding between parents on motorcycle\"."}], "task": "refering", "answer_template": "The \"female child riding between parents on motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 183, 203, 204, 205, 225, 226, 227, 228, 247, 248, 249, 250, 269, 270, 271, 272, 273, 293, 294, 295, 296, 315, 316, 317, 318, 339, 340], "bbox": [0.24017973856209152, 0.38099673202614376, 0.48754901960784314, 0.7025653594771242], "iscrowd": 0, "area": 11201.2731, "rle": {"size": [612, 612], "counts": "[Uh24nb09kMDRAe0f>CRA>m>Ki@7V?2a@0]?8Z@Jf?l0RATOn>l0SATOl>l0TATOl>l0TAUOk>6Z@Cl07i>6\\@Bk08k>3[@Ej09m>OZ@Gj0:n>KY@Ki0:Q?HW@Mh0U?_OU@3g0=W?]OS@5f0?Y?XOS@7e0b0Z?TOQ@:f0a0h?^OX@c0g?]OY@c0h?\\OY@d0g?[OY@f0f?ZOZ@f0g?YOY@h0g?VO[@i0e?WO[@j0e?UOZ@l0g?SOX@o0g?ROX@n0i?ROU@P1j?POU@R1k?nNT@R1m?nNQ@T1n?mNP@T1Q`0lNm_OV1S`0jNl_OV1T`0jNk_OX1U`0hNi_OY1X`0gNg_OZ1X`0gNf_OZ1[`0fNc_O\\1\\`0eNb_O\\1_`0cN`_O_1``0aN^_O`1c`0`N[_Ob1d`0_NZ_Ob1g`0^NW_Od1i`0\\NU_Oe1l`0:O0100O10O0100O100N1N3N2O1N101O1N2O001N2O1N101O1N2O002M7J5J6K6J5J6K5K\\Xk5"}, "label": "female child riding between parents on motorcycle"}]} {"id": 220504, "image": "COCO_train2014_000000220504.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baby between two adults on a motorcycle\"."}], "task": "refering", "answer_template": "The \"a baby between two adults on a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [181, 182, 183, 203, 204, 205, 225, 226, 227, 228, 247, 248, 249, 250, 269, 270, 271, 272, 273, 293, 294, 295, 296, 315, 316, 317, 318, 339, 340], "bbox": [0.24017973856209152, 0.38099673202614376, 0.48754901960784314, 0.7025653594771242], "iscrowd": 0, "area": 11201.2731, "rle": {"size": [612, 612], "counts": "[Uh24nb09kMDRAe0f>CRA>m>Ki@7V?2a@0]?8Z@Jf?l0RATOn>l0SATOl>l0TATOl>l0TAUOk>6Z@Cl07i>6\\@Bk08k>3[@Ej09m>OZ@Gj0:n>KY@Ki0:Q?HW@Mh0U?_OU@3g0=W?]OS@5f0?Y?XOS@7e0b0Z?TOQ@:f0a0h?^OX@c0g?]OY@c0h?\\OY@d0g?[OY@f0f?ZOZ@f0g?YOY@h0g?VO[@i0e?WO[@j0e?UOZ@l0g?SOX@o0g?ROX@n0i?ROU@P1j?POU@R1k?nNT@R1m?nNQ@T1n?mNP@T1Q`0lNm_OV1S`0jNl_OV1T`0jNk_OX1U`0hNi_OY1X`0gNg_OZ1X`0gNf_OZ1[`0fNc_O\\1\\`0eNb_O\\1_`0cN`_O_1``0aN^_O`1c`0`N[_Ob1d`0_NZ_Ob1g`0^NW_Od1i`0\\NU_Oe1l`0:O0100O10O0100O100N1N3N2O1N101O1N2O001N2O1N101O1N2O002M7J5J6K6J5J6K5K\\Xk5"}, "label": "a baby between two adults on a motorcycle"}]} {"id": 220504, "image": "COCO_train2014_000000220504.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman on the motorcycle behind the man and holding a baby\"."}], "task": "refering", "answer_template": "The \"the woman on the motorcycle behind the man and holding a baby\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [91, 92, 93, 113, 114, 115, 135, 136, 137, 157, 158, 159, 179, 180, 181, 200, 201, 202, 203, 222, 223, 224, 225, 244, 245, 246, 247, 266, 267, 268, 269, 270, 271, 272, 273, 288, 289, 290, 291, 292, 293, 294, 295, 310, 311, 312, 313, 314, 315, 316, 332, 333, 334, 335, 336, 337, 338, 356, 357, 358, 359, 360, 380, 381, 382, 402, 403, 404, 424, 425, 426, 446, 447, 448, 468, 469, 470], "bbox": [0.11035947712418302, 0.1948202614379085, 0.4256699346405229, 0.9898692810457516], "iscrowd": 0, "area": 40352.7736, "rle": {"size": [612, 612], "counts": "glX1h0Zb0P1QOV1jNj1UN8I2M3N1O2M2O2N2M2O2N1N3nL]K`Gd4_8^K^Ge4`8]K^Ge4`8^K]Gc4b8_K]Gb4a8`K]Ga4b8aK\\Ga4c8`K\\G`4c8bK[G`4c8bK[G`4c8bK\\G^4c8dK[G^4c8dK[G]4d8eK[G\\4c8gKZGZ4e8hKYGZ4e8kKWGU4h8oKTGS4k8PLQGQ4n8SLoFn3o8WLlFj3S9ZLiFh3U9\\LhFd3W9V3N1O2N1O2O0O2O001O001O001O00001O001O001O001O001O001O001O00001O000011N2N3M2O1N3M2O1N3ULkFaMW9[2kFeMW9W2kFjMW9CVGaNEl1W9@cGTNYO[2X9]ObGRNYOb2d9kNWG^NWOg2R:[NjFhNWOm2V;SMQGm2Q9PMQGP3o8oLQGQ3P9mLQGS3Q9kLoFV3Q9hLQGW3P9hLPGX3R9eLoF\\3Q9cLoF]3R9aLPG^3R9`LnF`3S9^LnFc3R9\\LnFd3Zb0RAZOS?d0h1N2N2N2N2N2N2N2N1O2O1N2O1N2OfPb6"}, "label": "the woman on the motorcycle behind the man and holding a baby"}]} {"id": 220504, "image": "COCO_train2014_000000220504.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman on the back of a motorcycle\"."}], "task": "refering", "answer_template": "The \"a woman on the back of a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [91, 92, 93, 113, 114, 115, 135, 136, 137, 157, 158, 159, 179, 180, 181, 200, 201, 202, 203, 222, 223, 224, 225, 244, 245, 246, 247, 266, 267, 268, 269, 270, 271, 272, 273, 288, 289, 290, 291, 292, 293, 294, 295, 310, 311, 312, 313, 314, 315, 316, 332, 333, 334, 335, 336, 337, 338, 356, 357, 358, 359, 360, 380, 381, 382, 402, 403, 404, 424, 425, 426, 446, 447, 448, 468, 469, 470], "bbox": [0.11035947712418302, 0.1948202614379085, 0.4256699346405229, 0.9898692810457516], "iscrowd": 0, "area": 40352.7736, "rle": {"size": [612, 612], "counts": "glX1h0Zb0P1QOV1jNj1UN8I2M3N1O2M2O2N2M2O2N1N3nL]K`Gd4_8^K^Ge4`8]K^Ge4`8^K]Gc4b8_K]Gb4a8`K]Ga4b8aK\\Ga4c8`K\\G`4c8bK[G`4c8bK[G`4c8bK\\G^4c8dK[G^4c8dK[G]4d8eK[G\\4c8gKZGZ4e8hKYGZ4e8kKWGU4h8oKTGS4k8PLQGQ4n8SLoFn3o8WLlFj3S9ZLiFh3U9\\LhFd3W9V3N1O2N1O2O0O2O001O001O001O00001O001O001O001O001O001O001O00001O000011N2N3M2O1N3M2O1N3ULkFaMW9[2kFeMW9W2kFjMW9CVGaNEl1W9@cGTNYO[2X9]ObGRNYOb2d9kNWG^NWOg2R:[NjFhNWOm2V;SMQGm2Q9PMQGP3o8oLQGQ3P9mLQGS3Q9kLoFV3Q9hLQGW3P9hLPGX3R9eLoF\\3Q9cLoF]3R9aLPG^3R9`LnF`3S9^LnFc3R9\\LnFd3Zb0RAZOS?d0h1N2N2N2N2N2N2N2N1O2O1N2O1N2OfPb6"}, "label": "a woman on the back of a motorcycle"}]} {"id": 220504, "image": "COCO_train2014_000000220504.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elderly man in a blue cap\"."}], "task": "refering", "answer_template": "The \"an elderly man in a blue cap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 119, 120, 121, 122, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 183, 184, 185, 186, 187, 188, 189, 205, 206, 207, 208, 209, 210, 211, 212, 213, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 296, 297, 298, 299, 300, 301, 302, 318, 319, 320, 321, 322, 323, 324, 325, 339, 340, 341, 342, 343, 344, 345, 346, 347, 361, 362, 363, 364, 365, 366, 367, 368, 369, 383, 384, 385, 386, 387, 388, 389, 390, 391, 407, 408, 409, 410, 411, 412, 413, 429, 430, 431, 432, 433, 434, 435, 451, 452, 453, 454, 472, 473, 474, 475], "bbox": [0.3280882352941176, 0.11011437908496732, 0.8853921568627451, 0.9842647058823529], "iscrowd": 0, "area": 84663.52125000002, "rle": {"size": [612, 612], "counts": "X\\h3=`b0b0^Oc0]O8H7J501N2O1N101N2O1N101O1N2O0O2O1e@ZMV>h2jAXMU>h2lAYMQ>i2nAXMQ>i2oAXMn=j2RBVMm=l2QBVMm=n2PBSMn=P3PBPMo=S3oAnLn=V3oAkLP>X3nAiLP>[3mAeLR>^3kAdLS>_3kAbLS>a3kA_LT>d3jA]LT>g3hAZLW>i3gAXLW>[4N3N2lMTKmEm4P:ZKkEh4Q:`KiEa4T:gKfE[4W:lKcEU4Z:SL`Eo3\\:YL^Ei3^:_L\\Eb3\\2eJT5P2YH^3a2eJQ5T2YHX3e2gJm4Y2XHR3j2gJi4o7VKSHg4P8WKSHg4n7XKTHe4m7ZKVHd4k7[KWHc4i7]KXHb4i7\\KZHb4g7]K[H`4f7_K]H_4d7`K^H^4c7`K`H^4a7aK`H^4`7aKcH\\4_7cKcH[4^7cKeH[4[7eKgHY4Z7fKgHY4Z7eKiHX4Y7gKiHW4W7hKlHV4U7iKmHU4T7iKnHV4R7jKPIS4R7kKQIS4P7lKPIT4Q7kKoHU4Q7jKPIV4Q7iKoHV4S7hKmHY4S7gKmHY4T7eKmH[4T7dKlH\\4U7bKlH]4U7cKkH]4V7aKkH_4V7`KjH`4V7`KjH`4W7^KiHc4X7\\KhHc4Z7[KgHe4i7kJWHU5l:0000O100]MPCROodCB[<>fCA[<>fCCY<W4[:_LXEWO?Z4Z:_LWESOc0\\4W:aLWEPOd0^4V:bLWEmNf0`4T:cLbF[3_9eLbFZ3^9fLdFW3^9iLbFV3^9jLcFT3_9lLbFR3^9nLcFP3^9PMcFo2^9QMcFl2^9TMcFk2]9UMdFi2^9WMcFg2]9YMdFe2]9[MeFc2\\9]MdF`2^9`MeFY2`9gMbFR2b9nMaFk1c9UNaF^1i9bNYFh0Z:XOiE3j:LYE^OZ;c0d3O010O1O010O0010O01O010O10O01O010O0010O010O01O10O01O010O0010O0100O0010OO2N1N3M2O2M3M2N3N1N3M2N3N1Nb`Y1"}, "label": "an elderly man in a blue cap"}]} {"id": 220504, "image": "COCO_train2014_000000220504.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a blue hat riding a motorcycle\"."}], "task": "refering", "answer_template": "The \"a man wearing a blue hat riding a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [53, 54, 55, 74, 75, 76, 77, 78, 96, 97, 98, 99, 100, 119, 120, 121, 122, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 183, 184, 185, 186, 187, 188, 189, 205, 206, 207, 208, 209, 210, 211, 212, 213, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 296, 297, 298, 299, 300, 301, 302, 318, 319, 320, 321, 322, 323, 324, 325, 339, 340, 341, 342, 343, 344, 345, 346, 347, 361, 362, 363, 364, 365, 366, 367, 368, 369, 383, 384, 385, 386, 387, 388, 389, 390, 391, 407, 408, 409, 410, 411, 412, 413, 429, 430, 431, 432, 433, 434, 435, 451, 452, 453, 454, 472, 473, 474, 475], "bbox": [0.3280882352941176, 0.11011437908496732, 0.8853921568627451, 0.9842647058823529], "iscrowd": 0, "area": 84663.52125000002, "rle": {"size": [612, 612], "counts": "X\\h3=`b0b0^Oc0]O8H7J501N2O1N101N2O1N101O1N2O0O2O1e@ZMV>h2jAXMU>h2lAYMQ>i2nAXMQ>i2oAXMn=j2RBVMm=l2QBVMm=n2PBSMn=P3PBPMo=S3oAnLn=V3oAkLP>X3nAiLP>[3mAeLR>^3kAdLS>_3kAbLS>a3kA_LT>d3jA]LT>g3hAZLW>i3gAXLW>[4N3N2lMTKmEm4P:ZKkEh4Q:`KiEa4T:gKfE[4W:lKcEU4Z:SL`Eo3\\:YL^Ei3^:_L\\Eb3\\2eJT5P2YH^3a2eJQ5T2YHX3e2gJm4Y2XHR3j2gJi4o7VKSHg4P8WKSHg4n7XKTHe4m7ZKVHd4k7[KWHc4i7]KXHb4i7\\KZHb4g7]K[H`4f7_K]H_4d7`K^H^4c7`K`H^4a7aK`H^4`7aKcH\\4_7cKcH[4^7cKeH[4[7eKgHY4Z7fKgHY4Z7eKiHX4Y7gKiHW4W7hKlHV4U7iKmHU4T7iKnHV4R7jKPIS4R7kKQIS4P7lKPIT4Q7kKoHU4Q7jKPIV4Q7iKoHV4S7hKmHY4S7gKmHY4T7eKmH[4T7dKlH\\4U7bKlH]4U7cKkH]4V7aKkH_4V7`KjH`4V7`KjH`4W7^KiHc4X7\\KhHc4Z7[KgHe4i7kJWHU5l:0000O100]MPCROodCB[<>fCA[<>fCCY<W4[:_LXEWO?Z4Z:_LWESOc0\\4W:aLWEPOd0^4V:bLWEmNf0`4T:cLbF[3_9eLbFZ3^9fLdFW3^9iLbFV3^9jLcFT3_9lLbFR3^9nLcFP3^9PMcFo2^9QMcFl2^9TMcFk2]9UMdFi2^9WMcFg2]9YMdFe2]9[MeFc2\\9]MdF`2^9`MeFY2`9gMbFR2b9nMaFk1c9UNaF^1i9bNYFh0Z:XOiE3j:LYE^OZ;c0d3O010O1O010O0010O01O010O10O01O010O0010O010O01O10O01O010O0010O0100O0010OO2N1N3M2O2M3M2N3N1N3M2N3N1Nb`Y1"}, "label": "a man wearing a blue hat riding a motorcycle"}]} {"id": 97632, "image": "COCO_train2014_000000097632.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tennis player with blue t - shirt & shorts in the court\"."}], "task": "refering", "answer_template": "The \"a tennis player with blue t - shirt & shorts in the court\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 146, 147, 148, 169, 170, 171, 192, 193, 194, 195, 216, 217, 218, 239, 240, 241, 262, 263, 264, 265, 286, 288, 309, 310], "bbox": [0.34471875, 0.3669086651053864, 0.556125, 0.9257845433255268], "iscrowd": 0, "area": 12252.942749999998, "rle": {"size": [427, 640], "counts": "c[l27o<8cNBcEd0Z:C^Ea0_:FYE>d:ITE;i:HSE[<\\OjC>e in this image.", "objects": [{"patches": [123, 146, 147, 148, 169, 170, 171, 192, 193, 194, 195, 216, 217, 218, 239, 240, 241, 262, 263, 264, 265, 286, 288, 309, 310], "bbox": [0.34471875, 0.3669086651053864, 0.556125, 0.9257845433255268], "iscrowd": 0, "area": 12252.942749999998, "rle": {"size": [427, 640], "counts": "c[l27o<8cNBcEd0Z:C^Ea0_:FYE>d:ITE;i:HSE[<\\OjC>e in this image.", "objects": [{"patches": [69, 70, 71, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171], "bbox": [0.0044843750000000005, 0.2185411764705882, 0.474546875, 0.5308941176470588], "iscrowd": 0, "area": 22047.03465, "rle": {"size": [425, 640], "counts": "hZ1k0^ in this image.", "objects": [{"patches": [69, 70, 71, 92, 93, 94, 95, 96, 97, 115, 116, 117, 118, 119, 120, 121, 122, 123, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171], "bbox": [0.0044843750000000005, 0.2185411764705882, 0.474546875, 0.5308941176470588], "iscrowd": 0, "area": 22047.03465, "rle": {"size": [425, 640], "counts": "hZ1k0^ in this image.", "objects": [{"patches": [30, 31, 32, 53, 54, 55, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 144, 145, 146, 147, 167, 168, 169, 190, 191, 192, 214, 237], "bbox": [0.22409374999999998, 0.07428934010152284, 0.45735937499999996, 0.7221573604060914], "iscrowd": 0, "area": 16433.5698, "rle": {"size": [394, 640], "counts": "cag11Y<0F0ZD1e;O[D1d;0\\D1b;0]D1b;0^D1a;O_D1`;0`D0_;1aDO_;1aD0^;0bD0^;0bD0^;0bD0^;0bD0^;0bD1];OcD1];OcD1];OcD2\\;NdD2\\;NdD2\\;NdD3Z;NfD2Z;NfD3Y;MgD3X;NhD2X;NhD3W;MiD3V;NjD1R;4nDLP;6PEIn::REFl:lJn3_1cKb3c0nJi3b1cK^3h0oJc3d1eK\\3j0oJa3i6cLUI\\3j6gLUIY3i6U1O1O1O001O1O2O0O1O1O1O1O1O2N2oNkJjJW5n4W1J6O1O1O00001O001O2cL_Hm1b7PN`HP2k7cMWH]2k7_MXH`2i7^MXHb2i7[MYHe2f801O0000001O0000001O0000001cNUGYOl8h0VGTOk8m0WGoNj8R1WGkNj8V1UGiNm8X1SGeNn8\\1TG`Nm8b1TGZNm8g1TGVNm8j1VGQNl8l1k0M4K5L4L4L4M3POlDe0V;YOmDd0V;YOnDc0b;K5I6JQ\\U4"}, "label": "this is a woman wearing a green shirt and black pants"}]} {"id": 56677, "image": "COCO_train2014_000000056677.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a lday wearing goggles , blue tsheir , black trouser standing with old lady and 2 kids\"."}], "task": "refering", "answer_template": "The \"a lday wearing goggles , blue tsheir , black trouser standing with old lady and 2 kids\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 32, 53, 54, 55, 75, 76, 77, 78, 79, 97, 98, 99, 100, 101, 102, 120, 121, 122, 123, 124, 125, 144, 145, 146, 147, 167, 168, 169, 190, 191, 192, 214, 237], "bbox": [0.22409374999999998, 0.07428934010152284, 0.45735937499999996, 0.7221573604060914], "iscrowd": 0, "area": 16433.5698, "rle": {"size": [394, 640], "counts": "cag11Y<0F0ZD1e;O[D1d;0\\D1b;0]D1b;0^D1a;O_D1`;0`D0_;1aDO_;1aD0^;0bD0^;0bD0^;0bD0^;0bD0^;0bD1];OcD1];OcD1];OcD2\\;NdD2\\;NdD2\\;NdD3Z;NfD2Z;NfD3Y;MgD3X;NhD2X;NhD3W;MiD3V;NjD1R;4nDLP;6PEIn::REFl:lJn3_1cKb3c0nJi3b1cK^3h0oJc3d1eK\\3j0oJa3i6cLUI\\3j6gLUIY3i6U1O1O1O001O1O2O0O1O1O1O1O1O2N2oNkJjJW5n4W1J6O1O1O00001O001O2cL_Hm1b7PN`HP2k7cMWH]2k7_MXH`2i7^MXHb2i7[MYHe2f801O0000001O0000001O0000001cNUGYOl8h0VGTOk8m0WGoNj8R1WGkNj8V1UGiNm8X1SGeNn8\\1TG`Nm8b1TGZNm8g1TGVNm8j1VGQNl8l1k0M4K5L4L4L4M3POlDe0V;YOmDd0V;YOnDc0b;K5I6JQ\\U4"}, "label": "a lday wearing goggles , blue tsheir , black trouser standing with old lady and 2 kids"}]} {"id": 56677, "image": "COCO_train2014_000000056677.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child looks up at an older female while sitting on a chair\"."}], "task": "refering", "answer_template": "The \"a child looks up at an older female while sitting on a chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [211, 212, 213, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 308], "bbox": [0.1328125, 0.6808883248730964, 0.44546875, 0.988756345177665], "iscrowd": 0, "area": 15311.930399999996, "rle": {"size": [394, 640], "counts": "mQQ13V<7H7I2O2M2O1N3M2O1N3N1N2O1N3M2O1N3N1N2N3N1N2O1N2O1N101^OWNbFj1\\9YNcFg1[9\\NdFc1[9aNcF_1\\9cNcF]1[9fNdFY1[9kNcFU1[9nNdFR1Z9QOeFo0Y9UOeFj0Z9T1N2O10O01000O0100O10O01000O0100O10O10O100O3N3L3N3M3L3N1N10000O100O100001O00000O10000000001O000000000000001O0000003M4L4L5K4L4L1OM3N20000001O0000001O00001O0000001O00001O00001O0000001O001O001O00001O001dNaFK_95eFG\\99gFCY9=kF_OV9`0nF\\OR9d0RGXOn8h0RGXOo8g0PGZOP9f0PGZOQ9e0nF\\OR9e0mF[OT9d0kF]OU9c0kF]OU9c0jF^OW9a0iF_OW9a0hF@Y9?gFAY9?gFAZ9>eFC[9=eFC\\9 in this image.", "objects": [{"patches": [211, 212, 213, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 262, 263, 279, 280, 281, 282, 283, 284, 285, 302, 303, 304, 305, 306, 307, 308], "bbox": [0.1328125, 0.6808883248730964, 0.44546875, 0.988756345177665], "iscrowd": 0, "area": 15311.930399999996, "rle": {"size": [394, 640], "counts": "mQQ13V<7H7I2O2M2O1N3M2O1N3N1N2O1N3M2O1N3N1N2N3N1N2O1N2O1N101^OWNbFj1\\9YNcFg1[9\\NdFc1[9aNcF_1\\9cNcF]1[9fNdFY1[9kNcFU1[9nNdFR1Z9QOeFo0Y9UOeFj0Z9T1N2O10O01000O0100O10O01000O0100O10O10O100O3N3L3N3M3L3N1N10000O100O100001O00000O10000000001O000000000000001O0000003M4L4L5K4L4L1OM3N20000001O0000001O00001O0000001O00001O00001O0000001O001O001O00001O001dNaFK_95eFG\\99gFCY9=kF_OV9`0nF\\OR9d0RGXOn8h0RGXOo8g0PGZOP9f0PGZOQ9e0nF\\OR9e0mF[OT9d0kF]OU9c0kF]OU9c0jF^OW9a0iF_OW9a0hF@Y9?gFAY9?gFAZ9>eFC[9=eFC\\9 in this image.", "objects": [{"patches": [146, 147, 148, 169, 170, 176, 192, 199, 200, 222, 223, 245, 246, 268, 269, 286, 287, 289, 290, 312, 313, 314], "bbox": [0.37353125, 0.48314720812182743, 0.7249218749999999, 0.9775380710659898], "iscrowd": 0, "area": 5051.948450000002, "rle": {"size": [394, 640], "counts": "iTl2=m;h0WO7J0O00010O01O001O2N1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1N2O1O1O1O2N1O1O1O1O1O1N2O1O1O1O1O1O1O1Ogm5:PRJ000010O00001O1O1O010O1O1O001O1O10O01O1O1O0010fWc04og\\O5L5M3001OXDF`;:`DF`;:`DF`;:`DFa;9_DFb;:^DFb;9^DHb;8^DHc;7]DIc;7]DHd;8\\DHe;6\\DJd;6[DKe;5[DJf;6ZDJg;5YDKg;4ZDLf;4YDMg;3YDLi;3WDMi;2XDNh;2XDNh;2XDMj;2UDOk;1UDOk;0VD0k;OUD0l;0TD0l;0TD0l;OTD2m;MSD3m;MSD2n;NRD2n;NRD2o;LRD4n;LQD4o7NmKNTL4m75jKGYL4k7 in this image.", "objects": [{"patches": [146, 147, 148, 169, 170, 176, 192, 199, 200, 222, 223, 245, 246, 268, 269, 286, 287, 289, 290, 312, 313, 314], "bbox": [0.37353125, 0.48314720812182743, 0.7249218749999999, 0.9775380710659898], "iscrowd": 0, "area": 5051.948450000002, "rle": {"size": [394, 640], "counts": "iTl2=m;h0WO7J0O00010O01O001O2N1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1N2O1O1O1O2N1O1O1O1O1O1N2O1O1O1O1O1O1O1Ogm5:PRJ000010O00001O1O1O010O1O1O001O1O10O01O1O1O0010fWc04og\\O5L5M3001OXDF`;:`DF`;:`DF`;:`DFa;9_DFb;:^DFb;9^DHb;8^DHc;7]DIc;7]DHd;8\\DHe;6\\DJd;6[DKe;5[DJf;6ZDJg;5YDKg;4ZDLf;4YDMg;3YDLi;3WDMi;2XDNh;2XDNh;2XDMj;2UDOk;1UDOk;0VD0k;OUD0l;0TD0l;0TD0l;OTD2m;MSD3m;MSD2n;NRD2n;NRD2o;LRD4n;LQD4o7NmKNTL4m75jKGYL4k7 in this image.", "objects": [{"patches": [236, 237, 238, 239, 257, 261, 262, 280, 284, 285, 303, 304, 307], "bbox": [0.1739375, 0.7216497461928933, 0.415796875, 0.9991116751269035], "iscrowd": 0, "area": 2732.12775, "rle": {"size": [394, 640], "counts": "kP[11U<3N3L4L31O2O0O1OoDGS:6nELU;3N2M3N2O1PEOh90XF0h9OYF1h9MYF3g9LZF4f9LZF4g9JZF6h9GXF:l:00000O100000000000001O02N2M3IZh=0kWB4L1O000000000000001O000000000000000000000000000000001O0000001O00001O00001O001cE2\\8OcG6W8KhG7W8JiG5W8KiG5W8LhG3Y8NfG2Z8OeG1[80dG0\\80eGN\\83cGL^84bGK_86`GIa87_GIa88^GHb88_GGa8:^GFb8:^GFb8;]GEc8;]GEc8<\\GDd8<\\GDd8=[GCe8=[GCe8>ZGBf8>ZGBf8?YGAh8>XGBh8?WGAi8?WGAj8?UGAk8?UGBj8>VGBk8=UGCk8>TGBl8>TGCl8P9]OkF25b0j9\\O_F<]l_4"}, "label": "the chair in which the child is sitting"}]} {"id": 56677, "image": "COCO_train2014_000000056677.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue and white lawn chair\"."}], "task": "refering", "answer_template": "The \"a blue and white lawn chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [236, 237, 238, 239, 257, 261, 262, 280, 284, 285, 303, 304, 307], "bbox": [0.1739375, 0.7216497461928933, 0.415796875, 0.9991116751269035], "iscrowd": 0, "area": 2732.12775, "rle": {"size": [394, 640], "counts": "kP[11U<3N3L4L31O2O0O1OoDGS:6nELU;3N2M3N2O1PEOh90XF0h9OYF1h9MYF3g9LZF4f9LZF4g9JZF6h9GXF:l:00000O100000000000001O02N2M3IZh=0kWB4L1O000000000000001O000000000000000000000000000000001O0000001O00001O00001O001cE2\\8OcG6W8KhG7W8JiG5W8KiG5W8LhG3Y8NfG2Z8OeG1[80dG0\\80eGN\\83cGL^84bGK_86`GIa87_GIa88^GHb88_GGa8:^GFb8:^GFb8;]GEc8;]GEc8<\\GDd8<\\GDd8=[GCe8=[GCe8>ZGBf8>ZGBf8?YGAh8>XGBh8?WGAi8?WGAj8?UGAk8?UGBj8>VGBk8=UGCk8>TGBl8>TGCl8P9]OkF25b0j9\\O_F<]l_4"}, "label": "a blue and white lawn chair"}]} {"id": 32105, "image": "COCO_train2014_000000032105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue bag with a d logo\"."}], "task": "refering", "answer_template": "The \"blue bag with a d logo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 94, 95, 96, 97, 98, 117, 118, 119, 120, 121, 144, 167, 168, 189, 190, 191, 192, 212, 213, 214, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 323, 324, 325, 326], "bbox": [0.001234375, 0.21717622080679408, 0.3549375, 0.8587685774946922], "iscrowd": 0, "area": 23881.93290000003, "rle": {"size": [471, 640], "counts": "Sh0[2]jNAV1?jNAV1?kN@U1`0kN@V1`0iN@W1`0iN@W1`0iN@W1`0iN@W1a0iN^OW1b0iN^OW1b0iN]OX1c0hN]OX1d0gN[OZ1e0fN[OZ1e0fNZO[1f0fNYOZ1h0eNXOZ1i0fNVO[1j0eNVOZ1k0gNTOY1m0fNSOS1T1mNkNl0]1TOcNk0^1VOaNj0`1UO`Nj0a1VO^Nj0c1WO\\Ni0d1WO\\Nh0f1WOZNi0f1WOZNh0g1YOYNf0g1ZOYNe0i1ZOWNf0i1ZOWNe0j1\\OnMcNgKi12UNZ6_1jMd0_2\\OfLaNXMT2U6[O`LeNkL1NP2i6ZO\\LiNjLW2l6POWLnNjLR2P7POTL_1o3`NlKd1U4\\NgKg1Z4YN`Kk1b4VNYKm1h4SNWKl1l4SNSKm1n4SNPKn1Q5RNnJm1T5SNjJm1Y5RNfJm1\\5SNbJm1`5TN^Jk1d5UNZJk1i5TNVJi1n5WNRJf1Q6[NmIc1V6^NcIc1b6]N\\Ic1g6]NWIb1k6_NSI_1Q7`NnH_1T7cNiH\\1Z7dNdHZ1_7gN_HX1d7iNYHU1k7kNSHP1S8ROjGl0Z8TOdGl00`M[7e1bHk02bM]7c1_Hk03eM_7_1\\Hm02gMc7]1XHm03iMe7Z1UHn05jMh7X1PHo06lMk7T1mGQ15nMR8n0gGV14oMW8j0bGY12QN^8e0^Gc2d8]MYGc2k8[MSGf2R9WMlFh2Z9X15J7J>B5J5L5J4L2N2N2N2O1N2N2UOk0N2CWDiMk;R2`0L4L4L4M3M4M2N3M3M3M2N3M3M3M2M4M3L3N3L4M3LWgm5"}, "label": "blue bag with a d logo"}]} {"id": 32105, "image": "COCO_train2014_000000032105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a duffel bag with a green hat on top\"."}], "task": "refering", "answer_template": "The \"a duffel bag with a green hat on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 94, 95, 96, 97, 98, 117, 118, 119, 120, 121, 144, 167, 168, 189, 190, 191, 192, 212, 213, 214, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 304, 323, 324, 325, 326], "bbox": [0.001234375, 0.21717622080679408, 0.3549375, 0.8587685774946922], "iscrowd": 0, "area": 23881.93290000003, "rle": {"size": [471, 640], "counts": "Sh0[2]jNAV1?jNAV1?kN@U1`0kN@V1`0iN@W1`0iN@W1`0iN@W1`0iN@W1a0iN^OW1b0iN^OW1b0iN]OX1c0hN]OX1d0gN[OZ1e0fN[OZ1e0fNZO[1f0fNYOZ1h0eNXOZ1i0fNVO[1j0eNVOZ1k0gNTOY1m0fNSOS1T1mNkNl0]1TOcNk0^1VOaNj0`1UO`Nj0a1VO^Nj0c1WO\\Ni0d1WO\\Nh0f1WOZNi0f1WOZNh0g1YOYNf0g1ZOYNe0i1ZOWNf0i1ZOWNe0j1\\OnMcNgKi12UNZ6_1jMd0_2\\OfLaNXMT2U6[O`LeNkL1NP2i6ZO\\LiNjLW2l6POWLnNjLR2P7POTL_1o3`NlKd1U4\\NgKg1Z4YN`Kk1b4VNYKm1h4SNWKl1l4SNSKm1n4SNPKn1Q5RNnJm1T5SNjJm1Y5RNfJm1\\5SNbJm1`5TN^Jk1d5UNZJk1i5TNVJi1n5WNRJf1Q6[NmIc1V6^NcIc1b6]N\\Ic1g6]NWIb1k6_NSI_1Q7`NnH_1T7cNiH\\1Z7dNdHZ1_7gN_HX1d7iNYHU1k7kNSHP1S8ROjGl0Z8TOdGl00`M[7e1bHk02bM]7c1_Hk03eM_7_1\\Hm02gMc7]1XHm03iMe7Z1UHn05jMh7X1PHo06lMk7T1mGQ15nMR8n0gGV14oMW8j0bGY12QN^8e0^Gc2d8]MYGc2k8[MSGf2R9WMlFh2Z9X15J7J>B5J5L5J4L2N2N2N2O1N2N2UOk0N2CWDiMk;R2`0L4L4L4M3M4M2N3M3M3M2N3M3M3M2M4M3L3N3L4M3LWgm5"}, "label": "a duffel bag with a green hat on top"}]} {"id": 32105, "image": "COCO_train2014_000000032105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black suitcase with a red tie on the zipper stands beside an identical black suitcase\"."}], "task": "refering", "answer_template": "The \"a black suitcase with a red tie on the zipper stands beside an identical black suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 98, 99, 100, 101, 102, 103, 104, 105, 106, 121, 122, 123, 124, 125, 126, 127, 128, 129, 144, 145, 146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 238, 239, 240], "bbox": [0.25193750000000004, 0.0, 0.667984375, 0.6418895966029724], "iscrowd": 0, "area": 58696.33265, "rle": {"size": [471, 640], "counts": "gQZ27`>f0ZOf0ZOf0ZOf0ZOf0ZO4L1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O5K7I6J7J6I7I6J6J1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N2N2N3M3M2N3M2N3M2cJTGk4o8PKSGP5o8kJTGU5W91O000000000000000000000O1000000000000O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O1O100O100O100O2O0O100O100O100O100O1O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O1O101N100L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L6J6J6J6J7I6J6J6J7I6J6J6J6J`_Q3"}, "label": "a black suitcase with a red tie on the zipper stands beside an identical black suitcase"}]} {"id": 32105, "image": "COCO_train2014_000000032105.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black suitcase touching a blue and black bag with a green hat on top\"."}], "task": "refering", "answer_template": "The \"a black suitcase touching a blue and black bag with a green hat on top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 98, 99, 100, 101, 102, 103, 104, 105, 106, 121, 122, 123, 124, 125, 126, 127, 128, 129, 144, 145, 146, 147, 148, 149, 150, 151, 152, 168, 169, 170, 171, 172, 173, 174, 191, 192, 193, 194, 195, 196, 197, 215, 216, 217, 218, 219, 238, 239, 240], "bbox": [0.25193750000000004, 0.0, 0.667984375, 0.6418895966029724], "iscrowd": 0, "area": 58696.33265, "rle": {"size": [471, 640], "counts": "gQZ27`>f0ZOf0ZOf0ZOf0ZOf0ZO4L1O1O2N1O1O1O2N1O1O2N1O1O1O2N1O1O1O2N1O5K7I6J7J6I7I6J6J1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N2N2N3M3M2N3M2N3M2cJTGk4o8PKSGP5o8kJTGU5W91O000000000000000000000O1000000000000O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O1O100O100O100O2O0O100O100O100O100O1O100O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O1O100O100O100O100O100O100O100O100O1O101N100L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L4L6J6J6J6J7I6J6J6J7I6J6J6J6J`_Q3"}, "label": "a black suitcase touching a blue and black bag with a green hat on top"}]} {"id": 580979, "image": "COCO_train2014_000000580979.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bench next to a green suitcase\"."}], "task": "refering", "answer_template": "The \"a bench next to a green suitcase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [261, 262, 263, 283, 284, 285, 286, 287, 288, 289, 306, 307, 308, 309, 310, 311, 312, 331, 332, 333, 334, 335, 355, 356, 357, 358, 359, 378, 379, 380, 381], "bbox": [0.323359375, 0.672125, 0.6145625, 0.9858958333333334], "iscrowd": 0, "area": 15693.75395, "rle": {"size": [480, 640], "counts": "\\\\Q31o>1N2N2O0O2N2O1N2N101N2N2O1N1O2O1N2N2O0O2N2O1N2N101N2N2O1M2N3L4L4M3L4L3N3N2O1O1N101O1O1O00001O001N101O00001O001O00001O010O00010O001O01O01O010O00010O0010O00010O01O01O01O010XE_Mo8a2mFcMS9]2iFgMX9Y2cFjM^9V2^FnMb9S2YFQNg9o1UFUNl9j1PFYNQ:h1jE\\NV:d1fE`N[:`1aEcN_:]1_EeNa:\\1\\EeNf:Z1XEhNh:h2001N10001O000O2O00001N10001O000O101O00001N10001O0O101O00001N10001O0O101O00001N1000001N10001O000O2O00001O0O101O000O2O00001O0O101O000O2O0000001N102N4L5K4K6K4L5K4J7I6J9G9G:F9Gd[c3"}, "label": "a bench next to a green suitcase"}]} {"id": 122231, "image": "COCO_train2014_000000122231.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"person on right with hat\"."}], "task": "refering", "answer_template": "The \"person on right with hat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 126, 148, 149, 150, 171, 172, 173, 174, 194, 195, 196, 197, 218, 219, 220, 240, 241, 242, 243, 263, 264, 266, 286, 287, 289, 290, 309, 310, 312, 313], "bbox": [0.438265625, 0.33182669789227165, 0.615984375, 0.9091334894613583], "iscrowd": 0, "area": 12468.051399999998, "rle": {"size": [427, 640], "counts": "_ae32X=2N1O2N2N2N2N1O1aJ?SMAj2e0SM\\Oj2j0RMWOR2B[J\\1`3SOQ2HYJZ1c3oNQ2LYJX1c3mNP22YJU1c3jNQ27XJR1d3hNQ2;WJQ1`2JV3]1WLTOi3n0SLTOm3o0mKTOS4n0iKTOW4n0eKTO[4n0aKTO_4o0\\KSOd4o0WKTOi4n0SKTOm4l0QKVOo4j0oJXOQ5h0lJ[OT5e0jJ]OW5b0gJ@Y5`0fJlNElMe5X3eJgN0mM[5\\3eJbN:mMQ5a3eJ]Nd0mMh4e3dJYNf6g1ZITNk6l1UInMQ7R2oHiMV7W2iHeM^7Y2bHbMd7]2\\H`Mh7_2XHbMi7\\2WHdMj7[2VHeMl7Y2THgMo7V2PHlMQ8R2oGnMS8P2mGPNV8m1jGQNZ8m1fGQN^8m1bGQNa8n1_GPNe8n1ZGQNi8n1WGoMn8o1RGoMP9Q2PGmMS9R2mFlMV9S2jFkMY9T2fFkM]9T2cFiM`9V2aFhMb9W2^FgMe9X2[FgMg9X2YFhMh9W2XFiMh9W2WFjMj9U2VFkMk9T2UFlMl9S2TFmMm9R2SFPNm9n1SFTNo9h1QFZNP:c1oE`NR:]1nEeNT:W1lEkNU:R1kEPOV:m0jEUOX:g0hE[OY:b0gE@Y:>fEEY::gEHY:6gELY:2gE0X:NiE5V:IjE:U:ClE`0R:_OmEd0S:YOnEj0Q:SOPFP1n9mNTFV1k9eNXF]1h9_N\\Fb1b9[NeFc1;RNg76UHh1NXNk7L^Hm1^O[NT8EeHQ2nN]NR:\\25K5L3L4L4L4L4L4L4L4L4K6E>C=C]WV3"}, "label": "person on right with hat"}]} {"id": 122231, "image": "COCO_train2014_000000122231.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the asian woman in blue\"."}], "task": "refering", "answer_template": "The \"the asian woman in blue\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 145, 146, 147, 169, 170, 171, 192, 193, 194, 215, 216, 217, 238, 239, 240, 260, 261, 262, 263], "bbox": [0.3315625, 0.32894613583138177, 0.45931249999999996, 0.7937939110070259], "iscrowd": 0, "area": 8660.67565, "rle": {"size": [427, 640], "counts": "Vgh27R=2N2dKIbK9\\4L_K6_40\\K1b45XKMf46WKLg48TKKk47QKLm48nJKP59jJK_4m0]KVOb4m0oJ@Q5a0YJ4f5OSJ6l5MmI8R6JhI;W6HbI;`6F[I=e6FTI>l6EmH=U7EeH<`7d23L4M3L5L3L4L5K4L4L4L5K4L5L5J5K5L5L3M3M4M3L3M4L4M2O2N2N2N16Kf0ZO7H6K7I8G:G3M1N2O00M2N3N1N3M4M4K5cM^GHg8IfG4`8]OlGa0X8POUHm0W:L4K5L4K5K5L4K5K]o_4"}, "label": "the asian woman in blue"}]} {"id": 122231, "image": "COCO_train2014_000000122231.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a blue shirt\"."}], "task": "refering", "answer_template": "The \"a person in a blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 145, 146, 147, 169, 170, 171, 192, 193, 194, 215, 216, 217, 238, 239, 240, 260, 261, 262, 263], "bbox": [0.3315625, 0.32894613583138177, 0.45931249999999996, 0.7937939110070259], "iscrowd": 0, "area": 8660.67565, "rle": {"size": [427, 640], "counts": "Vgh27R=2N2dKIbK9\\4L_K6_40\\K1b45XKMf46WKLg48TKKk47QKLm48nJKP59jJK_4m0]KVOb4m0oJ@Q5a0YJ4f5OSJ6l5MmI8R6JhI;W6HbI;`6F[I=e6FTI>l6EmH=U7EeH<`7d23L4M3L5L3L4L5K4L4L4L5K4L5L5J5K5L5L3M3M4M3L3M4L4M2O2N2N2N16Kf0ZO7H6K7I8G:G3M1N2O00M2N3N1N3M4M4K5cM^GHg8IfG4`8]OlGa0X8POUHm0W:L4K5L4K5K5L4K5K]o_4"}, "label": "a person in a blue shirt"}]} {"id": 408954, "image": "COCO_train2014_000000408954.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in dark blue jacket with red and black scarf\"."}], "task": "refering", "answer_template": "The \"woman in dark blue jacket with red and black scarf\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 364, 365, 366, 367, 387, 388, 389, 390], "bbox": [0.797296875, 0.17004166666666667, 1.0, 0.9876041666666667], "iscrowd": 0, "area": 41578.25885, "rle": {"size": [480, 640], "counts": "cZ_76[>`0dB[OX14e8d0jEC7Q1m9_ObEJ in this image.", "objects": [{"patches": [67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 134, 135, 136, 137, 157, 158, 159, 160, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229, 248, 249, 250, 251, 252, 271, 272, 273, 274, 275, 294, 295, 296, 297, 298, 317, 318, 319, 320, 321, 340, 341, 342, 343, 344, 364, 365, 366, 367, 387, 388, 389, 390], "bbox": [0.797296875, 0.17004166666666667, 1.0, 0.9876041666666667], "iscrowd": 0, "area": 41578.25885, "rle": {"size": [480, 640], "counts": "cZ_76[>`0dB[OX14e8d0jEC7Q1m9_ObEJ in this image.", "objects": [{"patches": [81, 82, 83, 84, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 240, 241, 242, 243, 244, 245, 246, 247, 248, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.338203125, 0.20466666666666666, 0.8078749999999999, 1.0], "iscrowd": 0, "area": 74227.06509999999, "rle": {"size": [480, 640], "counts": "SfU31m>6I7J6J6J5J4M2N3N2M3N2N2M3N2N1N3N1N2O1O1N2O1O1N2N2M3N2M3M3M3M3N2M3M3N2O1N2N2N2N2N2N2O1N2M3N2N2M3N2N2N2M3N2N2N2M3N2N2M3N2N2N2M3N2N2M3N2N2N2O1N2N2O1N2N2O1N2O1N2N2O1O1O100O1O1O100O1O1O100O1O1O1O1N2O1N2O1N2O100O100O100O1O1UMTJbLl5\\3bJWL_5f3oJoKQ5n3SKPLn4n3UKQLk4l3XKSLi4k3ZKSLg4i3^KVLb4f3cKYL]4b3hK^LX4^3lKbLT4\\3oKbLR4[3QLeLo3Y3TLfLl3X3VLhLj3T3[LkLe3Q3_LnLb3m2dLRM\\3l2fLTMZ3k2hLTMX3k2iLUMW3i2kLWMU3h2mLVMT3i2mLWMS3h2oLWMQ3h2PMXMP3g2RMWMo2h2SMWMm2i2TMVMl2j2UMUMk2k2VMSMk2l2WMSMi2m2XMRMh2n2ZMoLg2Q3ZMnLf2Q3\\MmLe2S3\\MlLd2T3]MjLd2U3^MjLb2V3_MhLb2X3c400O1000000O10000O100000000O10000000000O100000000O10000000000O10000000000000000001O00000000000000000000000000001O001O001O1O001O001O001O1O001O1O1O001O1O1O1O1O1O1WH\\LT4e3kK]LS4d3mK]LQ4c3oK_Lo3b3PL`Ln3a3QLbLl3_3TLaLk3a3SL`Ll3b3RL^Ln3d3PL\\LP4f3oKZLP4h3nKXLR4j3lKWLS4k3kKULU4l3kKSLU4P4hKQLW4T4dKlK\\4Y4_KhK`4]4\\KbKd4c4WK^Kh4g4SKYKm4l4nJTKR5n4mJRKR5P5lJPKT5S5hJoJW5S5fJoJY5S5dJoJ[5S5bJoJ]5S5`JoJ_5R5_JPK`5R5]JQKa5Q5\\JQKc5P5[JRKd5Q5WJRKh5R5SJQKk5R5QJRKl5Q5PJRKn5Q5nIQKQ6\\5aIeJ_6o6O001O001O001O001O001I6M4M2N3M2M4M2N3M3L4C=C=D in this image.", "objects": [{"patches": [81, 82, 83, 84, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 196, 197, 198, 199, 200, 201, 218, 219, 220, 221, 222, 223, 224, 240, 241, 242, 243, 244, 245, 246, 247, 248, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385], "bbox": [0.338203125, 0.20466666666666666, 0.8078749999999999, 1.0], "iscrowd": 0, "area": 74227.06509999999, "rle": {"size": [480, 640], "counts": "SfU31m>6I7J6J6J5J4M2N3N2M3N2N2M3N2N1N3N1N2O1O1N2O1O1N2N2M3N2M3M3M3M3N2M3M3N2O1N2N2N2N2N2N2O1N2M3N2N2M3N2N2N2M3N2N2N2M3N2N2M3N2N2N2M3N2N2M3N2N2N2O1N2N2O1N2N2O1N2O1N2N2O1O1O100O1O1O100O1O1O100O1O1O1O1N2O1N2O1N2O100O100O100O1O1UMTJbLl5\\3bJWL_5f3oJoKQ5n3SKPLn4n3UKQLk4l3XKSLi4k3ZKSLg4i3^KVLb4f3cKYL]4b3hK^LX4^3lKbLT4\\3oKbLR4[3QLeLo3Y3TLfLl3X3VLhLj3T3[LkLe3Q3_LnLb3m2dLRM\\3l2fLTMZ3k2hLTMX3k2iLUMW3i2kLWMU3h2mLVMT3i2mLWMS3h2oLWMQ3h2PMXMP3g2RMWMo2h2SMWMm2i2TMVMl2j2UMUMk2k2VMSMk2l2WMSMi2m2XMRMh2n2ZMoLg2Q3ZMnLf2Q3\\MmLe2S3\\MlLd2T3]MjLd2U3^MjLb2V3_MhLb2X3c400O1000000O10000O100000000O10000000000O100000000O10000000000O10000000000000000001O00000000000000000000000000001O001O001O1O001O001O001O1O001O1O1O001O1O1O1O1O1O1WH\\LT4e3kK]LS4d3mK]LQ4c3oK_Lo3b3PL`Ln3a3QLbLl3_3TLaLk3a3SL`Ll3b3RL^Ln3d3PL\\LP4f3oKZLP4h3nKXLR4j3lKWLS4k3kKULU4l3kKSLU4P4hKQLW4T4dKlK\\4Y4_KhK`4]4\\KbKd4c4WK^Kh4g4SKYKm4l4nJTKR5n4mJRKR5P5lJPKT5S5hJoJW5S5fJoJY5S5dJoJ[5S5bJoJ]5S5`JoJ_5R5_JPK`5R5]JQKa5Q5\\JQKc5P5[JRKd5Q5WJRKh5R5SJQKk5R5QJRKl5Q5PJRKn5Q5nIQKQ6\\5aIeJ_6o6O001O001O001O001O001I6M4M2N3M2M4M2N3M3L4C=C=D in this image.", "objects": [{"patches": [72, 73, 74, 94, 95, 96, 97, 117, 118, 119, 120, 121, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 368, 369, 370, 371, 372, 373, 374, 375], "bbox": [0.0016875000000000002, 0.21572916666666667, 0.355625, 0.9865208333333334], "iscrowd": 0, "area": 63275.050449999995, "rle": {"size": [480, 640], "counts": "Pf0U1k=Y2fM[2fMR2nM0O100O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O100O1O1O1O1O1O1O1O1O1O100O1O100O100O1O100O100O1O100O100O1O100O1ROeGWK[8]4Z1C=I7O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O10000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000001jGcLg4_3VKbLj4_3SKdLl4]3QKeLo4\\3oJfLP5[3mJgLS5Z3jJhLV5Y3hJiLW5X3fJjLZ5W3cJlL\\5U3bJlL^5U3_JnL`5S3]JoLc5R3[JoLe5R3XJQMg5P3VJRMj5o2TJSMk5n2RJTMn5m2PJUMo5l2nIVMR6k2lIVMT6n2gITMX6W3[IkLe6_3PIcLo6_3nHbLR7`3jHcLU7`3gHaLY7a3dH`L\\7b3`HaL_7a3^H`Lb7b3[H`Ld7^501O00001O3M3M2N3M2N3M3M2N3M2N3M2N3M3M2N3M2N3M3M3M5K5K4L5K4L5K4L5K4L5K4L5K4L5K4L5SOPE`MT;V2[EaMj:T2X1G:E:F;E:F;E:FjgP6"}, "label": "a gray haired man wearing glasses and a gray coat"}]} {"id": 408954, "image": "COCO_train2014_000000408954.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a balding man with a brown coat and glasses\"."}], "task": "refering", "answer_template": "The \"a balding man with a brown coat and glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 74, 94, 95, 96, 97, 117, 118, 119, 120, 121, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 190, 207, 208, 209, 210, 211, 212, 213, 214, 230, 231, 232, 233, 234, 235, 236, 237, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 282, 283, 299, 300, 301, 302, 303, 304, 305, 306, 307, 322, 323, 324, 325, 326, 327, 328, 329, 330, 345, 346, 347, 348, 349, 350, 351, 352, 368, 369, 370, 371, 372, 373, 374, 375], "bbox": [0.0016875000000000002, 0.21572916666666667, 0.355625, 0.9865208333333334], "iscrowd": 0, "area": 63275.050449999995, "rle": {"size": [480, 640], "counts": "Pf0U1k=Y2fM[2fMR2nM0O100O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O100O10000O10000O100O100O1O1O1O1O1O1O1O1O1O100O1O100O100O1O100O100O1O100O100O1O100O1ROeGWK[8]4Z1C=I7O1O1O1O1O1O1O1O1O1O1O1O1N2O1O1O10000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000001jGcLg4_3VKbLj4_3SKdLl4]3QKeLo4\\3oJfLP5[3mJgLS5Z3jJhLV5Y3hJiLW5X3fJjLZ5W3cJlL\\5U3bJlL^5U3_JnL`5S3]JoLc5R3[JoLe5R3XJQMg5P3VJRMj5o2TJSMk5n2RJTMn5m2PJUMo5l2nIVMR6k2lIVMT6n2gITMX6W3[IkLe6_3PIcLo6_3nHbLR7`3jHcLU7`3gHaLY7a3dH`L\\7b3`HaL_7a3^H`Lb7b3[H`Ld7^501O00001O3M3M2N3M2N3M3M2N3M2N3M2N3M3M2N3M2N3M3M3M5K5K4L5K4L5K4L5K4L5K4L5K4L5K4L5SOPE`MT;V2[EaMj:T2X1G:E:F;E:F;E:FjgP6"}, "label": "a balding man with a brown coat and glasses"}]} {"id": 56699, "image": "COCO_train2014_000000056699.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red bus with only the back end visible , in front of the bus labeled j322bsh\"."}], "task": "refering", "answer_template": "The \"the red bus with only the back end visible , in front of the bus labeled j322bsh\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 23, 24, 25, 26, 46, 47, 48, 69, 70, 71, 92, 93, 94, 115, 116, 117, 138, 139, 140, 161, 162, 163, 184, 185, 186, 207, 208, 209, 230, 231], "bbox": [0.003453125, 0.002240325865580448, 0.13274999999999998, 0.6179837067209776], "iscrowd": 0, "area": 22518.8476, "rle": {"size": [491, 640], "counts": "jU1[2m5S70O1O100O100O100O100O1O2O0O100O100O100O1O100O100O100O1O100O100O100O100O1O100O100O100O100O1O100O101N100O1O100O100O100O100O1O100O100O100O1O100O100O100O100O1O100O1gLY3lLjVZ8"}, "label": "the red bus with only the back end visible , in front of the bus labeled j322bsh"}]} {"id": 56699, "image": "COCO_train2014_000000056699.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"front of the big red tram in the other lane on the highway\"."}], "task": "refering", "answer_template": "The \"front of the big red tram in the other lane on the highway\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 26, 27, 28, 29, 30, 31, 32, 33, 34, 49, 50, 51, 52, 53, 54, 55, 56, 72, 73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171], "bbox": [0.12757812500000001, 0.013482688391038698, 0.48100000000000004, 0.41124236252545826], "iscrowd": 0, "area": 41973.00675, "rle": {"size": [491, 640], "counts": "e^W1a1S in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 26, 27, 28, 29, 30, 31, 32, 33, 34, 49, 50, 51, 52, 53, 54, 55, 56, 72, 73, 74, 75, 76, 77, 78, 79, 95, 96, 97, 98, 99, 100, 101, 102, 117, 118, 119, 120, 121, 122, 123, 124, 125, 140, 141, 142, 143, 144, 145, 146, 147, 148, 164, 165, 166, 167, 168, 169, 170, 171], "bbox": [0.12757812500000001, 0.013482688391038698, 0.48100000000000004, 0.41124236252545826], "iscrowd": 0, "area": 41973.00675, "rle": {"size": [491, 640], "counts": "e^W1a1S in this image.", "objects": [{"patches": [76, 77, 78, 79, 99, 100, 101, 102, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 188, 190, 191, 192, 193, 194, 214, 215, 216, 217, 236, 237, 238, 239, 240, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 287, 288, 304, 305, 306, 309, 310, 311, 312, 327, 328, 334, 335, 336, 350, 351, 358, 359, 372, 373, 374], "bbox": [0.166515625, 0.18683333333333335, 0.645171875, 1.0], "iscrowd": 0, "area": 38473.65630000002, "rle": {"size": [480, 640], "counts": "S\\b13l>2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2OO00hIaNPO^1Q1cNnN]1R1dNmN\\1S1eNlN[1T1fNlNX1V1hNiNX1W1iNhNW1X1iNhNU1Z1lNeNQ1^1POaNf0QJWOb74]Ne0QJUOd76ZNe0TJTOa78ZNe0UJSO`79ZNd0WJSO^7:ZNd0XJRO]63ZJ8P5c0ZJROW6R6DjI in this image.", "objects": [{"patches": [76, 77, 78, 79, 99, 100, 101, 102, 121, 122, 123, 124, 125, 126, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 188, 190, 191, 192, 193, 194, 214, 215, 216, 217, 236, 237, 238, 239, 240, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 287, 288, 304, 305, 306, 309, 310, 311, 312, 327, 328, 334, 335, 336, 350, 351, 358, 359, 372, 373, 374], "bbox": [0.166515625, 0.18683333333333335, 0.645171875, 1.0], "iscrowd": 0, "area": 38473.65630000002, "rle": {"size": [480, 640], "counts": "S\\b13l>2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2M3N2N2OO00hIaNPO^1Q1cNnN]1R1dNmN\\1S1eNlN[1T1fNlNX1V1hNiNX1W1iNhNW1X1iNhNU1Z1lNeNQ1^1POaNf0QJWOb74]Ne0QJUOd76ZNe0TJTOa78ZNe0UJSO`79ZNd0WJSO^7:ZNd0XJRO]63ZJ8P5c0ZJROW6R6DjI in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 95, 96, 97, 98, 99, 101, 102, 104, 105, 118, 119, 120, 121, 125, 127, 128, 141, 142, 143, 144, 150, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 191, 210, 211, 212, 213, 214], "bbox": [0.10904687500000002, 0.0022950819672131148, 0.635625, 0.6517564402810305], "iscrowd": 0, "area": 49611.32170000001, "rle": {"size": [427, 640], "counts": "Y]m01Y=1O1N1ZCMU<4iCNV<3hCNW<3gCOY<2eCOZ<3dCO[<1dC0\\<1bC0g5Gj0h0WOZOf0g0ZO\\Ob0d0_O^O=c0D@8`0aJUOa4>i0?dJTOd4?kND`Mk0oNTOg4`0ZN0fMc0WOoNj4`0iM=mM;^OkNl4`0YMi0SN4FdNo4b0gLV1[NJM`NR5b0VLd1aNA5[NU5c0eKS4TO]KW5b0UK_4BPKZ5d0dJh40fJ]5W7TKXHW4h7fKZH[4e7cK]H^4b7_KaHb4^7\\KdHe4\\7XKfHi4Y7TKjHm4U7QKmHP5f71O1O1O1O100O1O1O1OUKPH]4o7bKQH`4n7^KSHd4l7ZKUHh4j7WKVHk4i7SKXHo4g7PKYHR5f7lJ[HV5d7iJ\\HY5i72O1N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2I7I7I7H8I7I7H8004L3M3UK^Jg1e5kLRKPOZOU4h5mKhKLbNW4l7hKUHX4l7fKUHZ4k7dKWH\\4i7cKXH]4i7\\K]Hd4U8000O1000000000000000000000000O1000000\\LiKTNW4j1oKRNQ4k1ULRNk3l1[LPNe3m1aLPN_3n1gLnMY3P2lLmMT3P2SMkMn2S2WMjMi2T2[MjMe2T2_MjMa2T2cMjM]2T2gMjMY2T2kMjMU2T2oMjMQ2T2SNjMm1T2WNjMi1U2ZNhMg1V2]NhMc1V2aNhM_1V2eNhM[1V2iNhMW1V2mNhMS1U2ROiMn0o1]OnMc0k1GRN9l1LPN5P2OlM1T22iMNX24eML[27bMI^2;^MEb2b5001O000000001O000000001O000000001O000000001O000000001O000000001O0O1000O10000O10000O10000O10000O10000O10000008H7I8H8H4L2N2N2N1O2N2N2N2N1O2N1O1O1O2N1O1O1O2N1O1OO1O1O100O1O1G9G9F:G9I7O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1001O1O001O1O1O1O1O5K5K6`GoLZ6V3XI^M]6g2UIbMk6d2fHfMY7V40000000001O0000000000001O00000000001O00J6D in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 95, 96, 97, 98, 99, 101, 102, 104, 105, 118, 119, 120, 121, 125, 127, 128, 141, 142, 143, 144, 150, 163, 164, 165, 166, 167, 186, 187, 188, 189, 190, 191, 210, 211, 212, 213, 214], "bbox": [0.10904687500000002, 0.0022950819672131148, 0.635625, 0.6517564402810305], "iscrowd": 0, "area": 49611.32170000001, "rle": {"size": [427, 640], "counts": "Y]m01Y=1O1N1ZCMU<4iCNV<3hCNW<3gCOY<2eCOZ<3dCO[<1dC0\\<1bC0g5Gj0h0WOZOf0g0ZO\\Ob0d0_O^O=c0D@8`0aJUOa4>i0?dJTOd4?kND`Mk0oNTOg4`0ZN0fMc0WOoNj4`0iM=mM;^OkNl4`0YMi0SN4FdNo4b0gLV1[NJM`NR5b0VLd1aNA5[NU5c0eKS4TO]KW5b0UK_4BPKZ5d0dJh40fJ]5W7TKXHW4h7fKZH[4e7cK]H^4b7_KaHb4^7\\KdHe4\\7XKfHi4Y7TKjHm4U7QKmHP5f71O1O1O1O100O1O1O1OUKPH]4o7bKQH`4n7^KSHd4l7ZKUHh4j7WKVHk4i7SKXHo4g7PKYHR5f7lJ[HV5d7iJ\\HY5i72O1N2N2N2N2N2N2N2N2N2O1N2N2N2N2N2N2N2I7I7I7H8I7I7H8004L3M3UK^Jg1e5kLRKPOZOU4h5mKhKLbNW4l7hKUHX4l7fKUHZ4k7dKWH\\4i7cKXH]4i7\\K]Hd4U8000O1000000000000000000000000O1000000\\LiKTNW4j1oKRNQ4k1ULRNk3l1[LPNe3m1aLPN_3n1gLnMY3P2lLmMT3P2SMkMn2S2WMjMi2T2[MjMe2T2_MjMa2T2cMjM]2T2gMjMY2T2kMjMU2T2oMjMQ2T2SNjMm1T2WNjMi1U2ZNhMg1V2]NhMc1V2aNhM_1V2eNhM[1V2iNhMW1V2mNhMS1U2ROiMn0o1]OnMc0k1GRN9l1LPN5P2OlM1T22iMNX24eML[27bMI^2;^MEb2b5001O000000001O000000001O000000001O000000001O000000001O000000001O0O1000O10000O10000O10000O10000O10000O10000008H7I8H8H4L2N2N2N1O2N2N2N2N1O2N1O1O1O2N1O1O1O2N1O1OO1O1O100O1O1G9G9F:G9I7O1O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1001O1O001O1O1O1O1O5K5K6`GoLZ6V3XI^M]6g2UIbMk6d2fHfMY7V40000000001O0000000000001O00000000001O00J6D in this image.", "objects": [{"patches": [100, 101, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 239], "bbox": [0.31046874999999996, 0.3204449648711944, 0.757015625, 0.6742857142857144], "iscrowd": 0, "area": 23272.2618, "rle": {"size": [427, 640], "counts": "dTc23Y2Oc85YGOc83ZG3a8O\\G6`8L^G9]8J_G;]8GaG>Z8EbG`0Z8CbG;_8H`G:\\8JaG9[8JcG8[8KcG7\\8IbG:\\8GbG;]8GaG;^8E`G>^8@cGb0\\8\\OfGf0X8XOiGk0V8ROkGP1U8mNmGU1Q8kNmGY1R8hNiG\\1W8eNeG_1Z8cNaGa1]8aN_Gb1a8_N[Ge1d8]NWGf1h8\\NTGh1k8YNQGj1o8Z1O0O101O0O2O001O0000001O00000O100000001O00000000001O000000010O000000001O01O001O1O1O1O1O4L3M2M3L5L2N2N2M2O2N1O1O001XMXFm1h9QNZFP2d9oM^FQ2b9mM`FT2_9kMbFU2]9jMeFV2[9iMfFX2Y9gMhFY2W9gMjFZ2U9eMlF[2T9dMmF]2Q9cMPG]2P9aMRG_2n8`MSGa2k8_MVGa2j8^MWGc2h8\\MYGd2f8\\M[Ge2d8YM^Gg2b8XM_Gh2`8XMaGi2^8VMcGj2]8UMdGl2Z8TMgGl2Y8RMiGo2V8PMkGo2U8QMlGo2T8PMmGo2X900O100O101N100O100O101N100O100O101O00000000001O00000000001O00000O10001O000000001O0000000010O00000001O00000010O00000001O0000000010O00000001O0000001O00001O001O1O1O1O001O1O1O1O00100O1O1O00100O1O1O00100O1O1O001O100O1O1O00100O1O1O00101N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O10O01O001O001O001O0O2O001O001O1O0O2O001O001O001N101O001O001O0O2O001O00lhP2"}, "label": "a small horse lays on the beach with the ocean in the background"}]} {"id": 228734, "image": "COCO_train2014_000000228734.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray horse laying on the sand on the beach\"."}], "task": "refering", "answer_template": "The \"a gray horse laying on the sand on the beach\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [100, 101, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 199, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 239], "bbox": [0.31046874999999996, 0.3204449648711944, 0.757015625, 0.6742857142857144], "iscrowd": 0, "area": 23272.2618, "rle": {"size": [427, 640], "counts": "dTc23Y2Oc85YGOc83ZG3a8O\\G6`8L^G9]8J_G;]8GaG>Z8EbG`0Z8CbG;_8H`G:\\8JaG9[8JcG8[8KcG7\\8IbG:\\8GbG;]8GaG;^8E`G>^8@cGb0\\8\\OfGf0X8XOiGk0V8ROkGP1U8mNmGU1Q8kNmGY1R8hNiG\\1W8eNeG_1Z8cNaGa1]8aN_Gb1a8_N[Ge1d8]NWGf1h8\\NTGh1k8YNQGj1o8Z1O0O101O0O2O001O0000001O00000O100000001O00000000001O000000010O000000001O01O001O1O1O1O1O4L3M2M3L5L2N2N2M2O2N1O1O001XMXFm1h9QNZFP2d9oM^FQ2b9mM`FT2_9kMbFU2]9jMeFV2[9iMfFX2Y9gMhFY2W9gMjFZ2U9eMlF[2T9dMmF]2Q9cMPG]2P9aMRG_2n8`MSGa2k8_MVGa2j8^MWGc2h8\\MYGd2f8\\M[Ge2d8YM^Gg2b8XM_Gh2`8XMaGi2^8VMcGj2]8UMdGl2Z8TMgGl2Y8RMiGo2V8PMkGo2U8QMlGo2T8PMmGo2X900O100O101N100O100O101N100O100O101O00000000001O00000000001O00000O10001O000000001O0000000010O00000001O00000010O00000001O0000000010O00000001O0000001O00001O001O1O1O1O001O1O1O1O00100O1O1O00100O1O1O00100O1O1O001O100O1O1O00100O1O1O00101N1O2N1O2N2N1O2N1O2N1O2N1O2N1O2N1O2N1O10O01O001O001O001O0O2O001O001O1O0O2O001O001O001N101O001O001O0O2O001O00lhP2"}, "label": "a gray horse laying on the sand on the beach"}]} {"id": 490887, "image": "COCO_train2014_000000490887.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bike to the right of the frame\"."}], "task": "refering", "answer_template": "The \"the bike to the right of the frame\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 203, 204, 205, 206, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 360, 361, 362, 363, 364, 365, 366, 383, 384, 385, 386, 387, 388, 389, 406, 407, 408, 409, 410, 411, 429, 430, 431, 432, 433, 434, 452, 453, 454, 455, 456, 457, 475, 476, 477, 478, 479, 499, 500, 501], "bbox": [0.652359375, 0.219140625, 1.0, 0.9508906249999999], "iscrowd": 0, "area": 64947.231500000016, "rle": {"size": [640, 640], "counts": "g]U83jc03M3M3M3WDk1d1XNYNT3;oLBU3:nLDV37mLFW36lLGY33kLJY32jLK[30hLM\\30fLM^33aLJd34\\LIh37WLFn38RLER4:nKCW4`K_Od4a0[K\\Oi4c0WKZOn4d0RKYOR5g0mJVOX5h0hJUO\\5j0dJSOa5l0^JQOf5n0ZJoNk5o0UJnNo5R1PJkNT6T1lIjNX6U1gIhN]6W1cIfNb6X1^IeNf6[1YIdNj6Z1VIeNl6Z1TIeNo6Z1oHeNT7Z1lHeNW7Y1hHgNZ7Y1eHfN]7Y1bHfN`7Z1`HeNa7\\1]HdNd7\\1[HdNg7[1YHdNh7]1VHbNl7^1THaNm7`1QH`NP8`1oG`NS8_1mG_NU8b1iG^NX8b1hG]NY8c1fG]N[8d1dG[N^8d1cGYN_8g1aGXN`8i1`GUNa8k1_GTNb8l1_GQNd8o1[GPNf8P2ZGoMg8Q2ZGmMg8T2XGjMj8V2WGhMk8X2TGgMm8Y2TGeMm8[2SGdMn8]2RG`MP9`2PG_MR9`2nF_MS9b2mF\\MT9d2lFZMV9f2kFXMV9i2iFVMX9j2iFTMX9l2hFRMY9P3gFnLY9S3gFlLY9U3gFjLZ9W3fFgLZ9Z3fFdL[9^3eF`L[9a3eF^L[9c3fF[L[9f3dFYL\\9h3eFVL[9k3eFTL[9n3dFQL]9o3dFoK\\9R4dFmK\\9U4dFiK\\9X4dFgK\\9Z4eFdK\\9]4cFbK]9_4dF_K\\9c4cF\\K]9e4dFYK]9g4cFXK]9j4bFUK^9l4cFRK]9o4cFPK^9Q5bFlJ_9U5aFjJ_9W5bFgJ^9Z5bFeJ^9[5dFcJ]9]5cFbJ]9^5dFaJ\\9_5oFVJQ9k5\\GgIe8X6iGZIW8f6WHlHi7U7`20010O0001O010O00001N1O1O2O0O2N1O1O2N1O2N1O1O2O0O1O2N1O2N1O1O2N101N1O1O2N1O1O2N1_E`Ga9c8^F]G_9f8aFZG\\9i8cFXGZ9l8eFTGY9n8gFRGV9R9hFoFU9T9kFmFQ9V9oFjFn8Z9PGgFm8\\9SGdFk8^9UGbFh8b9VG_Fg8e9XG[Fe8i9ZGWFa8o9]GSF]8S:bGmEY8Y:fGgEU8b:fG_EU8k:fGUEU8];K6J6J6J5L5O1O010O100O00100O10O01O100O010O1O010O100O00100O10O01O100O010O1O010O100O00W1jNZ1eN[1fNZ1eN[1iNkD"}, "label": "the bike to the right of the frame"}]} {"id": 490887, "image": "COCO_train2014_000000490887.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue bicycle leaning against the pole\"."}], "task": "refering", "answer_template": "The \"blue bicycle leaning against the pole\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [130, 131, 132, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 203, 204, 205, 206, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 269, 270, 271, 272, 273, 274, 275, 292, 293, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 319, 320, 321, 337, 338, 339, 340, 341, 342, 343, 360, 361, 362, 363, 364, 365, 366, 383, 384, 385, 386, 387, 388, 389, 406, 407, 408, 409, 410, 411, 429, 430, 431, 432, 433, 434, 452, 453, 454, 455, 456, 457, 475, 476, 477, 478, 479, 499, 500, 501], "bbox": [0.652359375, 0.219140625, 1.0, 0.9508906249999999], "iscrowd": 0, "area": 64947.231500000016, "rle": {"size": [640, 640], "counts": "g]U83jc03M3M3M3WDk1d1XNYNT3;oLBU3:nLDV37mLFW36lLGY33kLJY32jLK[30hLM\\30fLM^33aLJd34\\LIh37WLFn38RLER4:nKCW4`K_Od4a0[K\\Oi4c0WKZOn4d0RKYOR5g0mJVOX5h0hJUO\\5j0dJSOa5l0^JQOf5n0ZJoNk5o0UJnNo5R1PJkNT6T1lIjNX6U1gIhN]6W1cIfNb6X1^IeNf6[1YIdNj6Z1VIeNl6Z1TIeNo6Z1oHeNT7Z1lHeNW7Y1hHgNZ7Y1eHfN]7Y1bHfN`7Z1`HeNa7\\1]HdNd7\\1[HdNg7[1YHdNh7]1VHbNl7^1THaNm7`1QH`NP8`1oG`NS8_1mG_NU8b1iG^NX8b1hG]NY8c1fG]N[8d1dG[N^8d1cGYN_8g1aGXN`8i1`GUNa8k1_GTNb8l1_GQNd8o1[GPNf8P2ZGoMg8Q2ZGmMg8T2XGjMj8V2WGhMk8X2TGgMm8Y2TGeMm8[2SGdMn8]2RG`MP9`2PG_MR9`2nF_MS9b2mF\\MT9d2lFZMV9f2kFXMV9i2iFVMX9j2iFTMX9l2hFRMY9P3gFnLY9S3gFlLY9U3gFjLZ9W3fFgLZ9Z3fFdL[9^3eF`L[9a3eF^L[9c3fF[L[9f3dFYL\\9h3eFVL[9k3eFTL[9n3dFQL]9o3dFoK\\9R4dFmK\\9U4dFiK\\9X4dFgK\\9Z4eFdK\\9]4cFbK]9_4dF_K\\9c4cF\\K]9e4dFYK]9g4cFXK]9j4bFUK^9l4cFRK]9o4cFPK^9Q5bFlJ_9U5aFjJ_9W5bFgJ^9Z5bFeJ^9[5dFcJ]9]5cFbJ]9^5dFaJ\\9_5oFVJQ9k5\\GgIe8X6iGZIW8f6WHlHi7U7`20010O0001O010O00001N1O1O2O0O2N1O1O2N1O2N1O1O2O0O1O2N1O2N1O1O2N101N1O1O2N1O1O2N1_E`Ga9c8^F]G_9f8aFZG\\9i8cFXGZ9l8eFTGY9n8gFRGV9R9hFoFU9T9kFmFQ9V9oFjFn8Z9PGgFm8\\9SGdFk8^9UGbFh8b9VG_Fg8e9XG[Fe8i9ZGWFa8o9]GSF]8S:bGmEY8Y:fGgEU8b:fG_EU8k:fGUEU8];K6J6J6J5L5O1O010O100O00100O10O01O100O010O1O010O100O00100O10O01O100O010O1O010O100O00W1jNZ1eN[1fNZ1eN[1iNkD"}, "label": "blue bicycle leaning against the pole"}]} {"id": 490887, "image": "COCO_train2014_000000490887.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bicycle on the left side of a pole\"."}], "task": "refering", "answer_template": "The \"a white bicycle on the left side of a pole\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 245, 246, 247, 248, 267, 268, 269, 270, 291, 292, 313, 314, 335, 336, 337, 358, 359, 360, 381, 382, 383, 404, 405, 406, 427, 428, 429], "bbox": [0.5626562500000001, 0.4214375, 0.80978125, 0.825171875], "iscrowd": 0, "area": 12799.712350000005, "rle": {"size": [640, 640], "counts": "l]Q72gc07J7N1O100O2N1O1O2O0\\^OZOm?g0b_OL[`06S_O=j`0DR_Ob0l`0^OR_Of0k`0\\OQ_Oi0l`0XOR_Ol0k`0UOS_On0k`0V1L4M3L3N1O1N3N1N2O2M2O1O1N3N1N2O2M2O1O2M2O1O1O0O2O001O001O0O2SN^K[Dc4e;_KXDa4h;cKTD^4k;eKRD[4n;hKoCY4PPMQAk2T?VMk@d2[?j06K6I6K5N2N3M2N2M3N2N3M2N2M3N3M2N2N2N2M4M2N2N2N3N100O101N100O10001N100O101N100O10001N101N3N1N3N1O1N3N1N3N1N3N101N101N1O2O0O2N101N101N1O101N1O2O0O2N101N101N1O2O0O2N101N101N1O2O0O1O2O0Omn[2"}, "label": "a white bicycle on the left side of a pole"}]} {"id": 490887, "image": "COCO_train2014_000000490887.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the back wheel of a bicycle that is next to a blue bicycle\"."}], "task": "refering", "answer_template": "The \"the back wheel of a bicycle that is next to a blue bicycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [223, 224, 245, 246, 247, 248, 267, 268, 269, 270, 291, 292, 313, 314, 335, 336, 337, 358, 359, 360, 381, 382, 383, 404, 405, 406, 427, 428, 429], "bbox": [0.5626562500000001, 0.4214375, 0.80978125, 0.825171875], "iscrowd": 0, "area": 12799.712350000005, "rle": {"size": [640, 640], "counts": "l]Q72gc07J7N1O100O2N1O1O2O0\\^OZOm?g0b_OL[`06S_O=j`0DR_Ob0l`0^OR_Of0k`0\\OQ_Oi0l`0XOR_Ol0k`0UOS_On0k`0V1L4M3L3N1O1N3N1N2O2M2O1O1N3N1N2O2M2O1O2M2O1O1O0O2O001O001O0O2SN^K[Dc4e;_KXDa4h;cKTD^4k;eKRD[4n;hKoCY4PPMQAk2T?VMk@d2[?j06K6I6K5N2N3M2N2M3N2N3M2N2M3N3M2N2N2N2M4M2N2N2N3N100O101N100O10001N100O101N100O10001N101N3N1N3N1O1N3N1N3N1N3N101N101N1O2O0O2N101N101N1O101N1O2O0O2N101N101N1O2O0O2N101N101N1O2O0O1O2O0Omn[2"}, "label": "the back wheel of a bicycle that is next to a blue bicycle"}]} {"id": 122259, "image": "COCO_train2014_000000122259.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wine glass to the back right of another glass\"."}], "task": "refering", "answer_template": "The \"a wine glass to the back right of another glass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 101, 102, 103, 119, 120, 121, 137, 138, 139, 156, 173, 174, 191, 192, 193], "bbox": [0.6263200000000001, 0.36568, 0.74374, 0.8136533333333333], "iscrowd": 0, "area": 5735.157700000002, "rle": {"size": [375, 500], "counts": "Pmb35_;301N2fKNmL3h2c0kL_OS3k0dLUO\\3S1PJcNb1;]4Z1dIaNn16]4m1aKTN_4m1`KSN`4n1_KRNa4o1^KRNa4o1]KRNc4o1\\KQNd4P2[KQNd4P2[KPNe4Q2YKPNg4Q2XKPNg4Q2XKoMh4R2WKnMi4S2UKoMj4R2UKnMk4W2PKiMP5]2bJkM_5[2jIUNV6f301O000000VObI[L^6W3QJhLo5i2`JWM`5[2oJdMQ5T2WKlMi4R2YKnMg4Q2ZKoMf4o1\\KQNd4n1]KRNc4m1^KSNb4m1_KRNa4m1`KSN`4m1`KRNa4n1_KRNa4m1`KSN`4m1`KSN`4l1aKTN_4l1bKRN_4l1dKSN\\4k1gKSNZ4k1hKUNX4i1kKVNU4h1nKUNT4i1oKTNS4j1PLRNV4i1Y3BP1PO[f^1"}, "label": "a wine glass to the back right of another glass"}]} {"id": 122259, "image": "COCO_train2014_000000122259.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glass on right closet to person\"."}], "task": "refering", "answer_template": "The \"glass on right closet to person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [83, 84, 85, 101, 102, 103, 119, 120, 121, 137, 138, 139, 156, 173, 174, 191, 192, 193], "bbox": [0.6263200000000001, 0.36568, 0.74374, 0.8136533333333333], "iscrowd": 0, "area": 5735.157700000002, "rle": {"size": [375, 500], "counts": "Pmb35_;301N2fKNmL3h2c0kL_OS3k0dLUO\\3S1PJcNb1;]4Z1dIaNn16]4m1aKTN_4m1`KSN`4n1_KRNa4o1^KRNa4o1]KRNc4o1\\KQNd4P2[KQNd4P2[KPNe4Q2YKPNg4Q2XKPNg4Q2XKoMh4R2WKnMi4S2UKoMj4R2UKnMk4W2PKiMP5]2bJkM_5[2jIUNV6f301O000000VObI[L^6W3QJhLo5i2`JWM`5[2oJdMQ5T2WKlMi4R2YKnMg4Q2ZKoMf4o1\\KQNd4n1]KRNc4m1^KSNb4m1_KRNa4m1`KSN`4m1`KRNa4n1_KRNa4m1`KSN`4m1`KSN`4l1aKTN_4l1bKRN_4l1dKSN\\4k1gKSNZ4k1hKUNX4i1kKVNU4h1nKUNT4i1oKTNS4j1PLRNV4i1Y3BP1PO[f^1"}, "label": "glass on right closet to person"}]} {"id": 122259, "image": "COCO_train2014_000000122259.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man wearing a brown striped sweater sitting on the right laughing\"."}], "task": "refering", "answer_template": "The \"man wearing a brown striped sweater sitting on the right laughing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [14, 15, 16, 17, 32, 33, 34, 35, 50, 51, 52, 53, 67, 68, 69, 70, 71, 84, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 160, 161], "bbox": [0.65224, 0.006373333333333333, 0.9994400000000001, 0.6715466666666666], "iscrowd": 0, "area": 31136.491250000003, "rle": {"size": [375, 500], "counts": "Ycg37U;f0YO?B;D in this image.", "objects": [{"patches": [14, 15, 16, 17, 32, 33, 34, 35, 50, 51, 52, 53, 67, 68, 69, 70, 71, 84, 85, 86, 87, 88, 89, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 137, 138, 139, 140, 141, 142, 143, 155, 156, 157, 158, 159, 160, 161], "bbox": [0.65224, 0.006373333333333333, 0.9994400000000001, 0.6715466666666666], "iscrowd": 0, "area": 31136.491250000003, "rle": {"size": [375, 500], "counts": "Ycg37U;f0YO?B;D in this image.", "objects": [{"patches": [6, 7, 8, 24, 25, 26, 42, 43, 44, 59, 60, 61, 62, 63, 76, 77, 78, 79, 80, 81, 93, 94, 95, 96, 97, 98, 111, 112, 113, 114, 115, 128, 129, 130, 131, 132, 133, 147, 148, 149, 150, 151], "bbox": [0.15744, 0.026533333333333332, 0.5587000000000001, 0.63376], "iscrowd": 0, "area": 21934.2272, "rle": {"size": [375, 500], "counts": "lTm03a;3M4L3M4K5L3M4L3M3M3M3M3N2N2O1N2N2N2N2O1N2N2N2O2M2N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O100O1O1O2O0O1O100O1O100O1O100O10000O10000O1WOSLhIm3l5`LRJa3c5kL\\JU3Y5VMgJj2V5ZMhJg2V5[MiJf2T5^MgJf2V5^MfJe2X5n1M31O01N1N2O1N2N2N200O10000O10000O100O100O100O1O100O100O100O100O2O01O0O101O000O2O00001O0O101nN[KPKf4e4lKSKT4c4_LSKc3b4oLVKR3_4Y2E;H7N3N2M3N5J5K6jN`GlNe8m0^GQOg8h0]GWOh8b0[G\\Ok8:\\GEV9CPG=U:00000000O10001O00000O1000000O2O0000000O2O00000O2O00000O2O00000O2O1O1O1N2O1O1O2M2OTj`2"}, "label": "a blurry mans face wearing a diamond patterned sweater"}]} {"id": 122259, "image": "COCO_train2014_000000122259.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy with blue on his sweater\"."}], "task": "refering", "answer_template": "The \"a guy with blue on his sweater\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 24, 25, 26, 42, 43, 44, 59, 60, 61, 62, 63, 76, 77, 78, 79, 80, 81, 93, 94, 95, 96, 97, 98, 111, 112, 113, 114, 115, 128, 129, 130, 131, 132, 133, 147, 148, 149, 150, 151], "bbox": [0.15744, 0.026533333333333332, 0.5587000000000001, 0.63376], "iscrowd": 0, "area": 21934.2272, "rle": {"size": [375, 500], "counts": "lTm03a;3M4L3M4K5L3M4L3M3M3M3M3N2N2O1N2N2N2N2O1N2N2N2O2M2N2N2N2O1N2N2N2O1N2N2N2N2O1N2N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O100O1O1O2O0O1O100O1O100O1O100O10000O10000O1WOSLhIm3l5`LRJa3c5kL\\JU3Y5VMgJj2V5ZMhJg2V5[MiJf2T5^MgJf2V5^MfJe2X5n1M31O01N1N2O1N2N2N200O10000O10000O100O100O100O1O100O100O100O100O2O01O0O101O000O2O00001O0O101nN[KPKf4e4lKSKT4c4_LSKc3b4oLVKR3_4Y2E;H7N3N2M3N5J5K6jN`GlNe8m0^GQOg8h0]GWOh8b0[G\\Ok8:\\GEV9CPG=U:00000000O10001O00000O1000000O2O0000000O2O00000O2O00000O2O00000O2O1O1O1N2O1O1O2M2OTj`2"}, "label": "a guy with blue on his sweater"}]} {"id": 253335, "image": "COCO_train2014_000000253335.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse behind the woman\"."}], "task": "refering", "answer_template": "The \"the horse behind the woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 73, 74, 96, 97, 98, 102, 119, 120, 121, 122, 123, 124, 125, 126, 142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 214, 215, 216, 237], "bbox": [0.175484375, 0.16796252927400468, 0.48475, 0.7070257611241217], "iscrowd": 0, "area": 21980.7005, "rle": {"size": [427, 640], "counts": "ij^11b;2jE8U:f0cDF\\;U1O1O100O100O100O100O100O1O1TFWNU8j1ZGgNd8Q3O1O101N1O1O1O10O01O001O0N2O2M2N3M2N2N3N1O1O2N1O26I8H7J4L0O2O000O2O000O2O000O2O001N10001O0O101O00001OO100000000000O10O1000000000000O100000O11O1O2N1N3N1O1N3M2N3M2N2N3N1N3M2NZOf000O1O100O2O0O100O1O100O101N10000000000O10001O00000000000000O100O10000O1000O010000O10000O10000O10000O1000M2J7J6I601O1O0O2O1O001N101O1O1O0O2O1O1O100O010O100O10O0100O100O10O0100O100jNZEFf:8`EEX;@nD5[ in this image.", "objects": [{"patches": [51, 73, 74, 96, 97, 98, 102, 119, 120, 121, 122, 123, 124, 125, 126, 142, 143, 144, 145, 146, 147, 148, 149, 165, 166, 167, 168, 169, 170, 171, 172, 189, 190, 191, 192, 193, 194, 214, 215, 216, 237], "bbox": [0.175484375, 0.16796252927400468, 0.48475, 0.7070257611241217], "iscrowd": 0, "area": 21980.7005, "rle": {"size": [427, 640], "counts": "ij^11b;2jE8U:f0cDF\\;U1O1O100O100O100O100O100O1O1TFWNU8j1ZGgNd8Q3O1O101N1O1O1O10O01O001O0N2O2M2N3M2N2N3N1O1O2N1O26I8H7J4L0O2O000O2O000O2O000O2O001N10001O0O101O00001OO100000000000O10O1000000000000O100000O11O1O2N1N3N1O1N3M2N3M2N2N3N1N3M2NZOf000O1O100O2O0O100O1O100O101N10000000000O10001O00000000000000O100O10000O1000O010000O10000O10000O10000O1000M2J7J6I601O1O0O2O1O001N101O1O1O0O2O1O1O100O010O100O10O0100O100O10O0100O100jNZEFf:8`EEX;@nD5[ in this image.", "objects": [{"patches": [2, 3, 25, 26, 47, 48, 49, 69, 70, 71, 72, 92, 93, 94, 95, 115, 116, 117, 138, 139, 140, 161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326], "bbox": [0.0, 0.044192037470725996, 0.26745312499999996, 0.990304449648712], "iscrowd": 0, "area": 40812.91910000001, "rle": {"size": [427, 640], "counts": "n7Y5j2X5001O001O00001O00001O000O2O00001O00001O000000000000000000O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1M3M3L4N2N2N2N2N2N1O2N2N2N2N2O1O1O1O1O100O1O1O1O1O1O1O1OaHWO^2h0^M_O`2`0[MGd29WMNg21UM6i2JRM in this image.", "objects": [{"patches": [2, 3, 25, 26, 47, 48, 49, 69, 70, 71, 72, 92, 93, 94, 95, 115, 116, 117, 138, 139, 140, 161, 162, 163, 164, 184, 185, 186, 187, 188, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 303, 322, 323, 324, 325, 326], "bbox": [0.0, 0.044192037470725996, 0.26745312499999996, 0.990304449648712], "iscrowd": 0, "area": 40812.91910000001, "rle": {"size": [427, 640], "counts": "n7Y5j2X5001O001O00001O00001O000O2O00001O00001O000000000000000000O1O1O1N2O1O1O1O1O1N2O1O1O1O1N2O1O1O1O1M3M3L4N2N2N2N2N2N1O2N2N2N2N2O1O1O1O1O100O1O1O1O1O1O1O1OaHWO^2h0^M_O`2`0[MGd29WMNg21UM6i2JRM in this image.", "objects": [{"patches": [76, 77, 78, 79, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 170, 171, 172], "bbox": [0.31840625, 0.24872222222222223, 0.51409375, 0.5698611111111112], "iscrowd": 0, "area": 11065.544, "rle": {"size": [360, 640], "counts": "PkW2b0]: in this image.", "objects": [{"patches": [76, 77, 78, 79, 99, 100, 101, 102, 103, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 170, 171, 172], "bbox": [0.31840625, 0.24872222222222223, 0.51409375, 0.5698611111111112], "iscrowd": 0, "area": 11065.544, "rle": {"size": [360, 640], "counts": "PkW2b0]: in this image.", "objects": [{"patches": [17, 18, 19, 40, 41, 42, 63, 64, 65, 66, 85, 86, 87, 88, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 252, 274], "bbox": [0.6507343750000001, 0.05518779342723005, 0.95896875, 0.753169014084507], "iscrowd": 0, "area": 40792.6866, "rle": {"size": [426, 640], "counts": "VZ]53a in this image.", "objects": [{"patches": [17, 18, 19, 40, 41, 42, 63, 64, 65, 66, 85, 86, 87, 88, 108, 109, 110, 111, 112, 113, 130, 131, 132, 133, 134, 135, 136, 153, 154, 155, 156, 157, 158, 159, 176, 177, 178, 179, 180, 181, 182, 199, 200, 201, 202, 203, 204, 205, 222, 223, 224, 225, 226, 227, 228, 245, 246, 247, 248, 249, 250, 251, 252, 274], "bbox": [0.6507343750000001, 0.05518779342723005, 0.95896875, 0.753169014084507], "iscrowd": 0, "area": 40792.6866, "rle": {"size": [426, 640], "counts": "VZ]53a in this image.", "objects": [{"patches": [26, 27, 28, 29, 49, 50, 51, 52, 53, 54, 72, 73, 74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 101, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 259, 260, 261, 281, 282, 283, 284], "bbox": [0.056843750000000005, 0.07190140845070422, 0.42181250000000003, 0.8561737089201878], "iscrowd": 0, "area": 40343.72435000001, "rle": {"size": [426, 640], "counts": "Uf?5R=4K6J5L4K5L4hCPOPW5EiJ in this image.", "objects": [{"patches": [26, 27, 28, 29, 49, 50, 51, 52, 53, 54, 72, 73, 74, 75, 76, 77, 78, 95, 96, 97, 98, 99, 100, 101, 118, 119, 120, 121, 122, 123, 124, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 186, 187, 188, 189, 190, 208, 209, 210, 211, 212, 213, 231, 232, 233, 234, 235, 254, 255, 256, 257, 258, 259, 260, 261, 281, 282, 283, 284], "bbox": [0.056843750000000005, 0.07190140845070422, 0.42181250000000003, 0.8561737089201878], "iscrowd": 0, "area": 40343.72435000001, "rle": {"size": [426, 640], "counts": "Uf?5R=4K6J5L4K5L4hCPOPW5EiJ in this image.", "objects": [{"patches": [8, 9, 10, 24, 25, 26, 27, 28, 29, 41, 42, 44, 45, 46, 47, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 77, 78, 79, 80, 81, 82, 83, 84, 95, 96, 97, 98, 99, 100, 101, 115, 116, 118], "bbox": [0.4369166666666667, 0.004671875000000001, 1.0, 0.2705625], "iscrowd": 0, "area": 21714.77970000001, "rle": {"size": [640, 480], "counts": "eYS44jc05K5L4K5K6J5K5L4K6K4L4L4L1O2N1O2N1O2M3N1O2N2N2N1OiNh^OIVa05P_OIn`06W_OHg`08\\_OGb`0:__OF_`0:d_OD[`0=g_OBW`0>l_O@R`0b0o_O[OR`0e0Q@WOP`0j0R@ROo?n0T@mNn?S1U@iNl?X1U@eNl?\\1[1100O11O0000__OoNT>P1kAROT>n0jATOW>l0gAVOX>j0fAXO[>g0eAZOZ>f0fAZO[>e0dA]O[>c0eA]O\\>b0cA@\\>`0dA@\\>`0dAA\\>?bAB^>>bAB_>>_AD`><`ADa>;_AF`>;^AFc>9]AHb>9]AGd>8\\AIc>8\\AHe>7[AJd>7[AIe>7[AJe>5[AKe>6ZAKf>4ZALf>5YALg>3YAMg>4XAMh>2XANi>2VAOi>1WAOj>1VAOj>0VA0j>0VA0k>0TA0l>1SAOn>1QAOP?1o@OQ?2n@NS?2l@MV?2j@NV?3i@MX?3g@MZ?3e@M[?4e@K\\?5c@K^?5a@J`?7_@Ib?6^@Jc?6\\@Jd?7\\@He?8Z@Hf?9Y@Fh?;W@Ej?;U@Ek?R@Bo?>P@AQ`0`0n_OAQ`0`0n_O@S`0`0m_O@R`0a0k_OAU`0`0f_OEY`0;c_OJ\\`07__OMb`0i1O001O1O0010O1O2O0O100O0010O00010O0010O0001O01O010O2O0O100O2N100O10O0010O0010O010O00010O010O010O01O010O10O010O010O01O010O010O1O001O001N101O001O001N2O001O001O0O2O00001O000O10001N1O100O2O0O1O100O2N100O101O0001O0001O0001O01O000001O00000000001O000000001O0001O000001O00000000001O00000001O01O000001O01O000001O0001O0PM"}, "label": "a right hand grasping a handle at the top of a blender"}]} {"id": 179618, "image": "COCO_train2014_000000179618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a mans hand reaching for the handle\"."}], "task": "refering", "answer_template": "The \"a mans hand reaching for the handle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 24, 25, 26, 27, 28, 29, 41, 42, 44, 45, 46, 47, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 77, 78, 79, 80, 81, 82, 83, 84, 95, 96, 97, 98, 99, 100, 101, 115, 116, 118], "bbox": [0.4369166666666667, 0.004671875000000001, 1.0, 0.2705625], "iscrowd": 0, "area": 21714.77970000001, "rle": {"size": [640, 480], "counts": "eYS44jc05K5L4K5K6J5K5L4K6K4L4L4L1O2N1O2N1O2M3N1O2N2N2N1OiNh^OIVa05P_OIn`06W_OHg`08\\_OGb`0:__OF_`0:d_OD[`0=g_OBW`0>l_O@R`0b0o_O[OR`0e0Q@WOP`0j0R@ROo?n0T@mNn?S1U@iNl?X1U@eNl?\\1[1100O11O0000__OoNT>P1kAROT>n0jATOW>l0gAVOX>j0fAXO[>g0eAZOZ>f0fAZO[>e0dA]O[>c0eA]O\\>b0cA@\\>`0dA@\\>`0dAA\\>?bAB^>>bAB_>>_AD`><`ADa>;_AF`>;^AFc>9]AHb>9]AGd>8\\AIc>8\\AHe>7[AJd>7[AIe>7[AJe>5[AKe>6ZAKf>4ZALf>5YALg>3YAMg>4XAMh>2XANi>2VAOi>1WAOj>1VAOj>0VA0j>0VA0k>0TA0l>1SAOn>1QAOP?1o@OQ?2n@NS?2l@MV?2j@NV?3i@MX?3g@MZ?3e@M[?4e@K\\?5c@K^?5a@J`?7_@Ib?6^@Jc?6\\@Jd?7\\@He?8Z@Hf?9Y@Fh?;W@Ej?;U@Ek?R@Bo?>P@AQ`0`0n_OAQ`0`0n_O@S`0`0m_O@R`0a0k_OAU`0`0f_OEY`0;c_OJ\\`07__OMb`0i1O001O1O0010O1O2O0O100O0010O00010O0010O0001O01O010O2O0O100O2N100O10O0010O0010O010O00010O010O010O01O010O10O010O010O01O010O010O1O001O001N101O001O001N2O001O001O0O2O00001O000O10001N1O100O2O0O1O100O2N100O101O0001O0001O0001O01O000001O00000000001O000000001O0001O000001O00000000001O00000001O01O000001O01O000001O0001O0PM"}, "label": "a mans hand reaching for the handle"}]} {"id": 179618, "image": "COCO_train2014_000000179618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"someone holding the base of a blender\"."}], "task": "refering", "answer_template": "The \"someone holding the base of a blender\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [311, 316, 317, 327, 328, 329, 333, 334, 335, 343, 344, 345, 346, 347, 351, 352, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.18060416666666665, 0.789890625, 0.7260416666666667, 0.998828125], "iscrowd": 0, "area": 16484.4729, "rle": {"size": [640, 480], "counts": "Sof12mc03M3N2jNH_^O;]a0Ia^O9[a0Kb^O8Za0Ke^O7Wa0Mf^O6Va0Mh^O6Ta0Nj^O4Sa0Oj^O4Ua0Lj^OM]a05`^OM_a05_^OKaa06]^OLca05Z^OLfa06X^OKga06W^OKia0W1O010O001O010O001O1O010O001O0001O0000001O00nNkMY@U2f?PNV@P2j?SNS@m1l?WNS@g1m?]NP@b1o?aNP@^1o?dNQ@[1o?gNQ@W1n?kNR@T1m?nNS@Q1m?POS@o0l?SOU@k0k?WOT@h0k?ZOU@e0j?]OV@b0j?_OW@?h?CX@JgA7X>McA4^>O]A3c>1XA0g>4TANl>5PALP?8j@IX?9c@I]?:_@Gb?CUXa2"}, "label": "someone holding the base of a blender"}]} {"id": 179618, "image": "COCO_train2014_000000179618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person holding blender with left hand\"."}], "task": "refering", "answer_template": "The \"a person holding blender with left hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [311, 316, 317, 327, 328, 329, 333, 334, 335, 343, 344, 345, 346, 347, 351, 352, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386], "bbox": [0.18060416666666665, 0.789890625, 0.7260416666666667, 0.998828125], "iscrowd": 0, "area": 16484.4729, "rle": {"size": [640, 480], "counts": "Sof12mc03M3N2jNH_^O;]a0Ia^O9[a0Kb^O8Za0Ke^O7Wa0Mf^O6Va0Mh^O6Ta0Nj^O4Sa0Oj^O4Ua0Lj^OM]a05`^OM_a05_^OKaa06]^OLca05Z^OLfa06X^OKga06W^OKia0W1O010O001O010O001O1O010O001O0001O0000001O00nNkMY@U2f?PNV@P2j?SNS@m1l?WNS@g1m?]NP@b1o?aNP@^1o?dNQ@[1o?gNQ@W1n?kNR@T1m?nNS@Q1m?POS@o0l?SOU@k0k?WOT@h0k?ZOU@e0j?]OV@b0j?_OW@?h?CX@JgA7X>McA4^>O]A3c>1XA0g>4TANl>5PALP?8j@IX?9c@I]?:_@Gb?CUXa2"}, "label": "a person holding blender with left hand"}]} {"id": 482731, "image": "COCO_train2014_000000482731.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"player in the jersey labeled 49\"."}], "task": "refering", "answer_template": "The \"player in the jersey labeled 49\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 83, 84, 103, 104, 105, 106, 107, 108, 128, 129, 130, 131, 152, 153, 154, 174, 175, 176, 177, 197, 198, 199, 200, 220, 221, 222, 223, 243, 244, 245, 266, 267, 268], "bbox": [0.49996875, 0.20829573934837092, 0.731828125, 0.8098245614035088], "iscrowd": 0, "area": 13978.684100000006, "rle": {"size": [399, 640], "counts": "Yil33Z<5L3N1O0O2M3O001O2O001N10O101N2N2O1N100O011O00000O0100O10O10O100O1O1O100O2N001O01O01O0000010O00010O001O010O01O001O00YI@`1`0UNMh13oM9n1GlM?T2B_Mj0`2WOQMX1n2jNoLX1Q3hNnLY1Q3iNhKAfNh1a5iNcKGgNa1f5iN`KKfN^1j5ROQKEQO[1n5WOnIfN2h3Q6RNjIP2W6RNdIP2]6o12N1O2M3N2N2M3FGoJgIk4]6\\K_Ia4^6fKbIV4_6mK`IP4b6QL^Il31]KY6i0fIg31hKS6c0jI`37RLk5k4ZJWKd5f4_J]K]5a4eJcKW5\\4kJPLi4o3XKVLb4i3_KZL_4e3bK]L\\4b3fK^LY4b3gK_LW4a3iKbLU4b3fK`LY4e3aK\\L_4g3\\K[Le4f3WK]Lh4e3VK[Lj4d51N2O1L4YOmJlIY5R6e0L4M3M3M3M2cJcIf4_6lJ_I35Q5P7O100000O02O0O1GhHZKY7f4jHVKX7h48O1cMZH0h7JcH0]7MPIHR7OWH[Nk0c1o6L`IKg61^IGh67QJmNU6o0g2N3M2O10001O1O1O7I2M2N2N3J]lR2"}, "label": "player in the jersey labeled 49"}]} {"id": 482731, "image": "COCO_train2014_000000482731.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a player with no 49\"."}], "task": "refering", "answer_template": "The \"a player with no 49\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [80, 81, 83, 84, 103, 104, 105, 106, 107, 108, 128, 129, 130, 131, 152, 153, 154, 174, 175, 176, 177, 197, 198, 199, 200, 220, 221, 222, 223, 243, 244, 245, 266, 267, 268], "bbox": [0.49996875, 0.20829573934837092, 0.731828125, 0.8098245614035088], "iscrowd": 0, "area": 13978.684100000006, "rle": {"size": [399, 640], "counts": "Yil33Z<5L3N1O0O2M3O001O2O001N10O101N2N2O1N100O011O00000O0100O10O10O100O1O1O100O2N001O01O01O0000010O00010O001O010O01O001O00YI@`1`0UNMh13oM9n1GlM?T2B_Mj0`2WOQMX1n2jNoLX1Q3hNnLY1Q3iNhKAfNh1a5iNcKGgNa1f5iN`KKfN^1j5ROQKEQO[1n5WOnIfN2h3Q6RNjIP2W6RNdIP2]6o12N1O2M3N2N2M3FGoJgIk4]6\\K_Ia4^6fKbIV4_6mK`IP4b6QL^Il31]KY6i0fIg31hKS6c0jI`37RLk5k4ZJWKd5f4_J]K]5a4eJcKW5\\4kJPLi4o3XKVLb4i3_KZL_4e3bK]L\\4b3fK^LY4b3gK_LW4a3iKbLU4b3fK`LY4e3aK\\L_4g3\\K[Le4f3WK]Lh4e3VK[Lj4d51N2O1L4YOmJlIY5R6e0L4M3M3M3M2cJcIf4_6lJ_I35Q5P7O100000O02O0O1GhHZKY7f4jHVKX7h48O1cMZH0h7JcH0]7MPIHR7OWH[Nk0c1o6L`IKg61^IGh67QJmNU6o0g2N3M2O10001O1O1O7I2M2N2N3J]lR2"}, "label": "a player with no 49"}]} {"id": 482731, "image": "COCO_train2014_000000482731.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in baseball uniform straddled with hand ready to catch a ball\"."}], "task": "refering", "answer_template": "The \"a man in baseball uniform straddled with hand ready to catch a ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 121, 122, 123, 124, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 217, 218, 219, 234, 235, 241, 242, 243, 257], "bbox": [0.168125, 0.28538847117794486, 0.57440625, 0.826967418546366], "iscrowd": 0, "area": 17778.583599999998, "rle": {"size": [399, 640], "counts": "h\\Z13Y<3M3M4L3M4L3M3M4L30O01O010O0010O01O010O1O010O1O10L2M13L4L4N2O100oNeNnF\\1Q9gNmFY1R9iNlFX1S9kNjFV1U9lNjFT1U9oNhFR1X9oNfFR1Z9oNeFR1Y9QOdFP1\\9QOcFo0]9SO`Fn0_9TO_Fm0a9TO^Fl0b9UO\\Fl0d9UOZFl0e9VOZFj0f9o000L4L4L400OO2O1N2N2O1N101N2N2O1N2N101N21O001O0O2O001O1O001N101O1O001O0O20O0100O0100O010O0100O010O010O10OCVMVGk2i8WMVGh2k8XMTGi2k8YMTGf2l8[MSGf2m8[MRGe2m8]MQGc2o8^MPGc2o8_MPGb2o8^MPGc2n8`0O2N1O2M2O1O2O0O2O0O101O001O0O101O001O000100O010O1O011N2N2O0O2O>A0010O010O000100O10O01O100O100O1O100O100O1O100O100O2N100O101N1O100O2O0O1O102M3QORFXOP:a0UF_On99WFFl93XFNj9J\\F6f9C_F=c9\\ObFd0`9UOeFk0`:0000000O10000000O100000O100000000000000O10O1000000000O1000001O00001O2M2O1OO1O1O100O001O1O1O1O1O1O1O00100O1O1O1O1O001O1O1O1O100O1O0N3L4LeRZ3"}, "label": "a man in baseball uniform straddled with hand ready to catch a ball"}]} {"id": 482731, "image": "COCO_train2014_000000482731.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"first baseman for the seattle mariners holding his foot on first base and trying to catch the ball\"."}], "task": "refering", "answer_template": "The \"first baseman for the seattle mariners holding his foot on first base and trying to catch the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [99, 100, 121, 122, 123, 124, 143, 144, 145, 146, 147, 148, 166, 167, 168, 169, 170, 171, 189, 190, 191, 192, 193, 194, 195, 211, 212, 213, 217, 218, 219, 234, 235, 241, 242, 243, 257], "bbox": [0.168125, 0.28538847117794486, 0.57440625, 0.826967418546366], "iscrowd": 0, "area": 17778.583599999998, "rle": {"size": [399, 640], "counts": "h\\Z13Y<3M3M4L3M4L3M3M4L30O01O010O0010O01O010O1O010O1O10L2M13L4L4N2O100oNeNnF\\1Q9gNmFY1R9iNlFX1S9kNjFV1U9lNjFT1U9oNhFR1X9oNfFR1Z9oNeFR1Y9QOdFP1\\9QOcFo0]9SO`Fn0_9TO_Fm0a9TO^Fl0b9UO\\Fl0d9UOZFl0e9VOZFj0f9o000L4L4L400OO2O1N2N2O1N101N2N2O1N2N101N21O001O0O2O001O1O001N101O1O001O0O20O0100O0100O010O0100O010O010O10OCVMVGk2i8WMVGh2k8XMTGi2k8YMTGf2l8[MSGf2m8[MRGe2m8]MQGc2o8^MPGc2o8_MPGb2o8^MPGc2n8`0O2N1O2M2O1O2O0O2O0O101O001O0O101O001O000100O010O1O011N2N2O0O2O>A0010O010O000100O10O01O100O100O1O100O100O1O100O100O2N100O101N1O100O2O0O1O102M3QORFXOP:a0UF_On99WFFl93XFNj9J\\F6f9C_F=c9\\ObFd0`9UOeFk0`:0000000O10000000O100000O100000000000000O10O1000000000O1000001O00001O2M2O1OO1O1O100O001O1O1O1O1O1O1O00100O1O1O1O1O001O1O1O1O100O1O0N3L4LeRZ3"}, "label": "first baseman for the seattle mariners holding his foot on first base and trying to catch the ball"}]} {"id": 376241, "image": "COCO_train2014_000000376241.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the arm and the brown shirt of the person standing on far right\"."}], "task": "refering", "answer_template": "The \"the arm and the brown shirt of the person standing on far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 136, 137, 159, 160, 181, 182, 183, 205, 206, 228, 229, 251, 252, 275, 298, 321, 343, 344], "bbox": [0.9028750000000001, 0.29011709601873537, 1.0, 0.9836299765807962], "iscrowd": 0, "area": 11117.998650000012, "rle": {"size": [427, 640], "counts": "^Wa72V=3M3L4M3M3M3M3M3M3L4M?Aj0VOj0VOd0\\O3L5L4L3M4L4L3M4K5L3M4L4L3O1bKhJi0X5TOkJl0U5QOnJo0R5nNQKR1o4jNUKV1k4gNXKX1i4eNZK[1f4bN]K^1b4`NaK`1_4]NdKc1\\4ZNgKf1o7O1000000000000O1000000O2O00000O1000000O101OcJa1"}, "label": "the arm and the brown shirt of the person standing on far right"}]} {"id": 376241, "image": "COCO_train2014_000000376241.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"partial view of a man in brown shirt\"."}], "task": "refering", "answer_template": "The \"partial view of a man in brown shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 114, 136, 137, 159, 160, 181, 182, 183, 205, 206, 228, 229, 251, 252, 275, 298, 321, 343, 344], "bbox": [0.9028750000000001, 0.29011709601873537, 1.0, 0.9836299765807962], "iscrowd": 0, "area": 11117.998650000012, "rle": {"size": [427, 640], "counts": "^Wa72V=3M3L4M3M3M3M3M3M3L4M?Aj0VOj0VOd0\\O3L5L4L3M4L4L3M4K5L3M4L4L3O1bKhJi0X5TOkJl0U5QOnJo0R5nNQKR1o4jNUKV1k4gNXKX1i4eNZK[1f4bN]K^1b4`NaK`1_4]NdKc1\\4ZNgKf1o7O1000000000000O1000000O2O00000O1000000O101OcJa1"}, "label": "partial view of a man in brown shirt"}]} {"id": 376241, "image": "COCO_train2014_000000376241.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a green shirt brushing his teeth\"."}], "task": "refering", "answer_template": "The \"a man in a green shirt brushing his teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 30, 31, 32, 33, 53, 54, 55, 56, 76, 77, 78, 79, 99, 100, 101, 102, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 307, 308, 309, 310, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.21315625, 0.03637002341920374, 0.545875, 0.9870023419203747], "iscrowd": 0, "area": 57014.218149999986, "rle": {"size": [427, 640], "counts": "lSi11i in this image.", "objects": [{"patches": [8, 9, 30, 31, 32, 33, 53, 54, 55, 56, 76, 77, 78, 79, 99, 100, 101, 102, 121, 122, 123, 124, 125, 126, 127, 143, 144, 145, 146, 147, 148, 149, 150, 166, 167, 168, 169, 170, 171, 172, 173, 189, 190, 191, 192, 193, 194, 195, 196, 212, 213, 214, 215, 216, 217, 218, 235, 236, 237, 238, 239, 240, 241, 258, 259, 260, 261, 262, 263, 264, 281, 282, 283, 284, 285, 286, 287, 304, 305, 306, 307, 308, 309, 310, 326, 327, 328, 329, 330, 331, 332, 333], "bbox": [0.21315625, 0.03637002341920374, 0.545875, 0.9870023419203747], "iscrowd": 0, "area": 57014.218149999986, "rle": {"size": [427, 640], "counts": "lSi11i in this image.", "objects": [{"patches": [129, 130, 151, 152, 153, 154, 173, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 249, 265, 266, 267, 268, 269, 270, 271, 272, 273, 288, 289, 290, 291, 292, 293, 294, 295, 296, 311, 312, 313, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.534046875, 0.38775175644028104, 0.8956406250000001, 0.9886651053864168], "iscrowd": 0, "area": 41735.77125000001, "rle": {"size": [427, 640], "counts": "]n^42<2]<0^C5`ZKCf4=YKDf4TKCk4c0nJ_OR5e0cJB\\5`0cJ@\\5b0bJ_O^5a0aJ@^5a0aJ_O`5a0`J_O_5c0_J^Oa5b0^J_Oa5V4O100O100O100O100O10000000000O10000000000O100000000000000001O00001O00001O00001O00001O00001O00001O000O2O00001O0000001O00001O0000001O00001O00001O00007I8H8H001O000000000000001O001O001O1O001O1O001O001O1O2N4L3M3M3M4K4M3M2N2N2N1O2N2N2N2N1O2N2N2N1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2N00001O00001O00001O00001^LlFk2T9nLTGQ3l8gL]GX3\\9O1O1O1O1O6J7I1O1O0O2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1E;kNZdk0"}, "label": "a girl in a grey shirt brushing her teeth with two other people"}]} {"id": 376241, "image": "COCO_train2014_000000376241.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a grey t shirt brushing her teeth\"."}], "task": "refering", "answer_template": "The \"a woman in a grey t shirt brushing her teeth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 151, 152, 153, 154, 173, 174, 175, 176, 177, 178, 197, 198, 199, 200, 201, 202, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 249, 265, 266, 267, 268, 269, 270, 271, 272, 273, 288, 289, 290, 291, 292, 293, 294, 295, 296, 311, 312, 313, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.534046875, 0.38775175644028104, 0.8956406250000001, 0.9886651053864168], "iscrowd": 0, "area": 41735.77125000001, "rle": {"size": [427, 640], "counts": "]n^42<2]<0^C5`ZKCf4=YKDf4TKCk4c0nJ_OR5e0cJB\\5`0cJ@\\5b0bJ_O^5a0aJ@^5a0aJ_O`5a0`J_O_5c0_J^Oa5b0^J_Oa5V4O100O100O100O100O10000000000O10000000000O100000000000000001O00001O00001O00001O00001O00001O00001O000O2O00001O0000001O00001O0000001O00001O00001O00007I8H8H001O000000000000001O001O001O1O001O1O001O001O1O2N4L3M3M3M4K4M3M2N2N2N1O2N2N2N2N1O2N2N2N1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2N00001O00001O00001O00001^LlFk2T9nLTGQ3l8gL]GX3\\9O1O1O1O1O6J7I1O1O0O2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1E;kNZdk0"}, "label": "a woman in a grey t shirt brushing her teeth"}]} {"id": 7601, "image": "COCO_train2014_000000007601.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black cow that is only half visible standing close to a fully visible cow\"."}], "task": "refering", "answer_template": "The \"a black cow that is only half visible standing close to a fully visible cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 89, 90, 108, 109, 110, 111, 112, 113, 131, 132, 133, 134, 135, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 247, 248], "bbox": [0.7139375, 0.22736470588235294, 0.9533437499999999, 0.7380941176470588], "iscrowd": 0, "area": 19248.920999999995, "rle": {"size": [425, 640], "counts": "Zim57Q==B6J3N2M3N3L3M3N2M2PEPOe9S1ZFWO\\9j0cFXOj0]Og6\\1]HZOe0Ff6R1dH[O?Oe6g0kH\\O<6b6>RI^O95e6=RI@74f6e5@cJ:]5DlJ6S5ITKNo40^KAf4>S4N2O1N2O1O1N2O001O1N2O1O1O1O1N2O1O1O1N2O1Ogh<"}, "label": "a black cow that is only half visible standing close to a fully visible cow"}]} {"id": 7601, "image": "COCO_train2014_000000007601.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the far right black cow\"."}], "task": "refering", "answer_template": "The \"the far right black cow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 113, 114, 135, 136, 137, 158, 159, 160, 182, 183, 205, 206, 228, 229, 251, 252, 274], "bbox": [0.8986875, 0.22717647058823529, 1.0, 0.7989176470588236], "iscrowd": 0, "area": 11457.14575, "rle": {"size": [425, 640], "counts": "Zi^73U=8G:G8H8G:G8H9F9H8G:G9G9F;F:F`0_Oc0^Oc0\\Oc0^O>B1N2O1O1N2O1O2M2O1N2O1O1N2O1O1NSOiJlIW5h5aKmI^4h5g11O2N1O2N1N3N1O2N1O1O2N1O2N1O2N1O1N3N1O2N1O2N00L4M3L4MRM"}, "label": "the far right black cow"}]} {"id": 7601, "image": "COCO_train2014_000000007601.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cow on the far right who is barely visible\"."}], "task": "refering", "answer_template": "The \"cow on the far right who is barely visible\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 113, 114, 135, 136, 137, 158, 159, 160, 182, 183, 205, 206, 228, 229, 251, 252, 274], "bbox": [0.8986875, 0.22717647058823529, 1.0, 0.7989176470588236], "iscrowd": 0, "area": 11457.14575, "rle": {"size": [425, 640], "counts": "Zi^73U=8G:G8H8G:G8H9F9H8G:G9G9F;F:F`0_Oc0^Oc0\\Oc0^O>B1N2O1O1N2O1O2M2O1N2O1O1N2O1O1NSOiJlIW5h5aKmI^4h5g11O2N1O2N1N3N1O2N1O1O2N1O2N1O2N1O1N3N1O2N1O2N00L4M3L4MRM"}, "label": "cow on the far right who is barely visible"}]} {"id": 7601, "image": "COCO_train2014_000000007601.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cow standing in high grass wearing an ear tag with the number 342\"."}], "task": "refering", "answer_template": "The \"a cow standing in high grass wearing an ear tag with the number 342\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 256, 257, 258, 259, 260, 279, 280, 282], "bbox": [0.159953125, 0.30967058823529414, 0.79115625, 0.853764705882353], "iscrowd": 0, "area": 59812.068650000016, "rle": {"size": [425, 640], "counts": "aQ[15`:0mG6b7 in this image.", "objects": [{"patches": [96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 256, 257, 258, 259, 260, 279, 280, 282], "bbox": [0.159953125, 0.30967058823529414, 0.79115625, 0.853764705882353], "iscrowd": 0, "area": 59812.068650000016, "rle": {"size": [425, 640], "counts": "aQ[15`:0mG6b7 in this image.", "objects": [{"patches": [142, 143, 144, 156, 157, 158, 159, 171, 172, 173, 174, 186, 187, 188, 189, 201, 202, 203, 204], "bbox": [0.3938084112149533, 0.41475, 0.6687149532710281, 0.59859375], "iscrowd": 0, "area": 10909.352, "rle": {"size": [640, 428], "counts": "UnY36fc07I7J6I7I7I7J6I7L2N2N2N2N1O2N2N2N2N2M3N2N2N2N2N2N2N2N2N101O1N2O1O1N2O1O1N2O1O0O10000O10000O10000O100000000O100000000000000000000000000000000O1001O00001O000O2O0O101N100O2O0O101O0O101N101N2O1N3N1N2O1N2O1N3N1N2O1N2O2L3L4K5K5K5K:F in this image.", "objects": [{"patches": [142, 143, 144, 156, 157, 158, 159, 171, 172, 173, 174, 186, 187, 188, 189, 201, 202, 203, 204], "bbox": [0.3938084112149533, 0.41475, 0.6687149532710281, 0.59859375], "iscrowd": 0, "area": 10909.352, "rle": {"size": [640, 428], "counts": "UnY36fc07I7J6I7I7I7J6I7L2N2N2N2N1O2N2N2N2N2M3N2N2N2N2N2N2N2N2N101O1N2O1O1N2O1O1N2O1O0O10000O10000O10000O100000000O100000000000000000000000000000000O1001O00001O000O2O0O101N100O2O0O101O0O101N101N2O1N3N1N2O1N2O1N3N1N2O1N2O2L3L4K5K5K5K:F in this image.", "objects": [{"patches": [162, 166, 175, 176, 178, 179, 180, 182, 183, 192, 193, 194, 195, 196, 197, 199, 200, 210, 211, 212, 213, 214, 215, 216, 217, 227, 228, 229, 230, 231, 232, 233, 246, 247, 248, 249, 263, 264, 265, 266, 280, 281, 282, 283, 297, 298, 299, 300, 314, 315, 316, 317, 331, 332, 333, 334, 348, 349, 350, 351, 365, 366, 367, 368, 382, 383, 384, 385], "bbox": [0.3333333333333333, 0.43131250000000004, 0.8078125, 0.9876093749999999], "iscrowd": 0, "area": 32767.68475, "rle": {"size": [640, 480], "counts": "^YT31oc02M3N2N2M3N2N2N2M3N2N2M3N2N2N2M3N2N2M3N2N2N2N2N2N2N2N2N2N2N2N20000000000001N1N2O1O1O2M2O2N1O2M2O2N1O2M2O2N1O2M2M01O1O010O00100O001O010O1RDjNQ5V1oJlNP5T1nJnNQ5S1_E@^3_OQ7Q1nD in this image.", "objects": [{"patches": [162, 166, 175, 176, 178, 179, 180, 182, 183, 192, 193, 194, 195, 196, 197, 199, 200, 210, 211, 212, 213, 214, 215, 216, 217, 227, 228, 229, 230, 231, 232, 233, 246, 247, 248, 249, 263, 264, 265, 266, 280, 281, 282, 283, 297, 298, 299, 300, 314, 315, 316, 317, 331, 332, 333, 334, 348, 349, 350, 351, 365, 366, 367, 368, 382, 383, 384, 385], "bbox": [0.3333333333333333, 0.43131250000000004, 0.8078125, 0.9876093749999999], "iscrowd": 0, "area": 32767.68475, "rle": {"size": [640, 480], "counts": "^YT31oc02M3N2N2M3N2N2N2M3N2N2M3N2N2N2M3N2N2M3N2N2N2N2N2N2N2N2N2N2N2N20000000000001N1N2O1O1O2M2O2N1O2M2O2N1O2M2O2N1O2M2M01O1O010O00100O001O010O1RDjNQ5V1oJlNP5T1nJnNQ5S1_E@^3_OQ7Q1nD in this image.", "objects": [{"patches": [159, 166, 176, 217, 228, 234, 245, 251, 262, 268, 279, 280, 285, 296, 297, 302, 314, 319, 331, 336, 353], "bbox": [0.3752083333333333, 0.38190625, 0.8274791666666667, 0.89446875], "iscrowd": 0, "area": 5574.801000000003, "rle": {"size": [640, 480], "counts": "[h`3k0Uc010O01O0010O7I8I7H4M_b21_]M4M4K4L4L4L5K4L4M4K4M3O2N1N2O1N3N1N2O2M21O100O1O1O10O01O100O1O1O1N1O2N2N01O000O10000Va^2^NXTbM0f[O;E:N3M3L3N3O02M3NZ_Og1k>a2`Mb1^Nb0^O4L4L4L49G?A?A?A?A?XKPBb2[>RMdBP2]=nMbCT1_ in this image.", "objects": [{"patches": [60, 76, 77, 93, 94, 95, 110, 111, 112, 127, 128, 129, 144, 145, 146, 161, 163, 178, 195, 212, 228, 229, 245, 246, 262, 263], "bbox": [0.4555625, 0.129734375, 0.6194583333333333, 0.692109375], "iscrowd": 0, "area": 5928.264050000004, "rle": {"size": [640, 480], "counts": "aXY4a0Qc0b0E;K5L4L4L4aNYN\\@f1^?bNb@Y1[?oNe@l0X?\\Oh@?U?Ji@3R?6PACm>e0VASOg>U1[AdNb>e1`ATN\\>T2fAeMW>c2kAVMR>R3V19G8H8G8I3M3M4L3M3M4LM3M3M2N3M3M3M3M3M7I7I9G=C=C\\OkA7T>C`B1_=ITCJn in this image.", "objects": [{"patches": [72, 73, 89, 90, 104, 105, 106, 107, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 207, 208, 209, 224, 225, 226, 241, 242, 243, 258, 259, 260, 275, 276, 277, 292, 293], "bbox": [0.1055, 0.188859375, 0.46170833333333333, 0.7535937500000001], "iscrowd": 0, "area": 24109.639899999995, "rle": {"size": [640, 480], "counts": "YUP1B>BO28H8H8G8I6oBRKd:S5REZKf:l4nDbKj:d4kDiKl:l6J3M2N5J7J5K5K5J5L3M001N1N2N2M3N2N2N2LoM`FjI\\9i5XGTJc8j5fGTJU8o5oGoIm7R6YHkIb7X6bHgI\\7W6jHfIU7Y6oHeIQ7X6TIfIk6R6aIkI^6m5mIQJR6l5TJRJl5S6[HWIOIgNk0n8[6RHlI[OVOH`0j8d6iGaJ_OiNg8m6_G_JLbNe8T7VG^J7\\Nb8b9`G[F`8f9cGVF]8k9eGRF\\8n9fGoEZ8R:hGkEX8V:jGgEV8Z:h001O1O2`GYFc6h9ZIZFf6g9XIZFh6h9TI[Fl5JaIl9a0[Fg52fId9`0\\Fd57kI_9>\\F_5=RJX9<]F[5d0WJP91hFZ5g0dJc8_OXG[5g0UKc9`4jF^KV9Z4SGeKn8R4\\GlKd8]1TJbNl5_1TJ`Nm5`1TJ^Nl5c1SJ]Nm5d1SJ[Nn5e1QJ[NP6e1oI[NS6c1nI\\NU6b1jI^NY6`1gI_N\\6_1cIaN_6]1aIcNb6[1^IdNe6Z1ZIfNi6W1XIfNm6X1RIfNR7X1nHfNW7W1l5M3M00000O2O0000001O0O101O000O1O2O0O1O101N1O1O100O1O100O100000O0100000000O100000O10N2N2N2N2M3N2NldQ5"}, "label": "a man in a blue coat standing in the snow"}]} {"id": 286132, "image": "COCO_train2014_000000286132.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with blue shirt and black pants holding skis\"."}], "task": "refering", "answer_template": "The \"a man with blue shirt and black pants holding skis\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 89, 90, 104, 105, 106, 107, 121, 122, 123, 124, 138, 139, 140, 141, 142, 143, 154, 155, 156, 157, 158, 159, 171, 172, 173, 174, 175, 188, 189, 190, 191, 192, 207, 208, 209, 224, 225, 226, 241, 242, 243, 258, 259, 260, 275, 276, 277, 292, 293], "bbox": [0.1055, 0.188859375, 0.46170833333333333, 0.7535937500000001], "iscrowd": 0, "area": 24109.639899999995, "rle": {"size": [640, 480], "counts": "YUP1B>BO28H8H8G8I6oBRKd:S5REZKf:l4nDbKj:d4kDiKl:l6J3M2N5J7J5K5K5J5L3M001N1N2N2M3N2N2N2LoM`FjI\\9i5XGTJc8j5fGTJU8o5oGoIm7R6YHkIb7X6bHgI\\7W6jHfIU7Y6oHeIQ7X6TIfIk6R6aIkI^6m5mIQJR6l5TJRJl5S6[HWIOIgNk0n8[6RHlI[OVOH`0j8d6iGaJ_OiNg8m6_G_JLbNe8T7VG^J7\\Nb8b9`G[F`8f9cGVF]8k9eGRF\\8n9fGoEZ8R:hGkEX8V:jGgEV8Z:h001O1O2`GYFc6h9ZIZFf6g9XIZFh6h9TI[Fl5JaIl9a0[Fg52fId9`0\\Fd57kI_9>\\F_5=RJX9<]F[5d0WJP91hFZ5g0dJc8_OXG[5g0UKc9`4jF^KV9Z4SGeKn8R4\\GlKd8]1TJbNl5_1TJ`Nm5`1TJ^Nl5c1SJ]Nm5d1SJ[Nn5e1QJ[NP6e1oI[NS6c1nI\\NU6b1jI^NY6`1gI_N\\6_1cIaN_6]1aIcNb6[1^IdNe6Z1ZIfNi6W1XIfNm6X1RIfNR7X1nHfNW7W1l5M3M00000O2O0000001O0O101O000O1O2O0O1O101N1O1O100O1O100O100000O0100000000O100000O10N2N2N2N2M3N2NldQ5"}, "label": "a man with blue shirt and black pants holding skis"}]} {"id": 425398, "image": "COCO_train2014_000000425398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the larger of two sheep\"."}], "task": "refering", "answer_template": "The \"the larger of two sheep\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 302, 303, 305], "bbox": [0.15090625, 0.5699568965517241, 0.41384375, 0.8296120689655172], "iscrowd": 0, "area": 9226.37105, "rle": {"size": [464, 640], "counts": "VY\\14[>5L4K5L5J5K5L4K5L1N100O1O1N2O1O1N2O1N2O1N2O1O1N10L3L5K5K4M40O2O0O101O0O2O0O101O0O2O0O101N101O2M200O2O000O101N100O101N100O2O000O101N100O101N1lDgMQ:Z2jEjMV:X2dElM[:X2^ElMb:V2XEnMg:U2TEnMl:k2O2O000OlNVEdNj:\\1bEXN]:i1oEkMQ:U2T12O3M3L4M@[DYN`;h1cDXNZ;h1iDXNS;i1PEWNm:i1VEWNg:i1\\EWN`:j1cEWNZ:g1jEYNS:g1PFXNn9h1VFTNj9l1YFQNh9n1[FnMf9R2]FkMc9U2`FhM`9X2]1O0O1000001O0000001O000O10001O0N2O1O2N1O1N2N3M2M3N2N2N3L3N2N2N3M2M3O101O0O10000O2O00000O2O000O1000000O10000O1000000O10000O1000O01DXeY5"}, "label": "the larger of two sheep"}]} {"id": 425398, "image": "COCO_train2014_000000425398.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sheep grazing with a bunch of fur on its back\"."}], "task": "refering", "answer_template": "The \"a sheep grazing with a bunch of fur on its back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [213, 214, 235, 236, 237, 238, 239, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 302, 303, 305], "bbox": [0.15090625, 0.5699568965517241, 0.41384375, 0.8296120689655172], "iscrowd": 0, "area": 9226.37105, "rle": {"size": [464, 640], "counts": "VY\\14[>5L4K5L5J5K5L4K5L1N100O1O1N2O1O1N2O1N2O1N2O1O1N10L3L5K5K4M40O2O0O101O0O2O0O101O0O2O0O101N101O2M200O2O000O101N100O101N100O2O000O101N100O101N1lDgMQ:Z2jEjMV:X2dElM[:X2^ElMb:V2XEnMg:U2TEnMl:k2O2O000OlNVEdNj:\\1bEXN]:i1oEkMQ:U2T12O3M3L4M@[DYN`;h1cDXNZ;h1iDXNS;i1PEWNm:i1VEWNg:i1\\EWN`:j1cEWNZ:g1jEYNS:g1PFXNn9h1VFTNj9l1YFQNh9n1[FnMf9R2]FkMc9U2`FhM`9X2]1O0O1000001O0000001O000O10001O0N2O1O2N1O1N2N3M2M3N2N2N3L3N2N2N3M2M3O101O0O10000O2O00000O2O000O1000000O10000O1000000O10000O1000O01DXeY5"}, "label": "a sheep grazing with a bunch of fur on its back"}]} {"id": 204216, "image": "COCO_train2014_000000204216.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the train that is on the tracks\"."}], "task": "refering", "answer_template": "The \"the train that is on the tracks\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 258, 259, 260, 261, 262, 263, 264, 265], "bbox": [0.23757812500000003, 0.5349069767441861, 0.8201718750000001, 0.800674418604651], "iscrowd": 0, "area": 25466.249699999997, "rle": {"size": [430, 640], "counts": "YTP22[=3M3L4M3M2N3M3WOi0@a0H7J5K4M3M3O1O2N1O1O1O1O1O2N101N2O1O0O2O1O1N10000O100O10000O01000O10000O10000O1000O01000000001O0000000000001O00000O10000000001O00000000000O10001O000000000O1000001O00000O1000000000001O0O1000000000000O2O00000000000O10001O0000000O100000000O2O0000000O1000000O2O0000000O100000000O2O00000O100000000O2O0000000O1000000O2O0000000O100000000O2O0000000O1000000O2O0000000O100000001N1000000O100000000O2O00000O1000000O101O000O100000001N1000000O1000001N1000000O10001O0O1000000O101O000O100000001N1000000O1000001N1000000O10001O0O1000000O101O000O100000001N1000000O1000001N1000000O10001O0O1000000O101O0000000O10001O000000000O101O000000000O101O00000000000O2O00000000000O2O0000000000001N1000000000001N1000000000001O0O10000000001O0O10000000001O000O10lP`1"}, "label": "the train that is on the tracks"}]} {"id": 204216, "image": "COCO_train2014_000000204216.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow grey and blue train\"."}], "task": "refering", "answer_template": "The \"the yellow grey and blue train\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 258, 259, 260, 261, 262, 263, 264, 265], "bbox": [0.23757812500000003, 0.5349069767441861, 0.8201718750000001, 0.800674418604651], "iscrowd": 0, "area": 25466.249699999997, "rle": {"size": [430, 640], "counts": "YTP22[=3M3L4M3M2N3M3WOi0@a0H7J5K4M3M3O1O2N1O1O1O1O1O2N101N2O1O0O2O1O1N10000O100O10000O01000O10000O10000O1000O01000000001O0000000000001O00000O10000000001O00000000000O10001O000000000O1000001O00000O1000000000001O0O1000000000000O2O00000000000O10001O0000000O100000000O2O0000000O1000000O2O0000000O100000000O2O00000O100000000O2O0000000O1000000O2O0000000O100000000O2O0000000O1000000O2O0000000O100000001N1000000O100000000O2O00000O1000000O101O000O100000001N1000000O1000001N1000000O10001O0O1000000O101O000O100000001N1000000O1000001N1000000O10001O0O1000000O101O000O100000001N1000000O1000001N1000000O10001O0O1000000O101O0000000O10001O000000000O101O000000000O101O00000000000O2O00000000000O2O0000000000001N1000000000001N1000000000001O0O10000000001O0O10000000001O000O10lP`1"}, "label": "the yellow grey and blue train"}]} {"id": 523711, "image": "COCO_train2014_000000523711.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"cow painted limo\"."}], "task": "refering", "answer_template": "The \"cow painted limo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 170, 171, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 245, 246, 247, 248], "bbox": [0.264515625, 0.4276041666666667, 0.838703125, 0.6103958333333334], "iscrowd": 0, "area": 20959.757950000003, "rle": {"size": [480, 640], "counts": "i^_21m>2O2M2N3N1N3N1N3M201N101O0O2O001N101O1N2N2N3M2M3N3M2N2M2O1O1N2O1O1N2O1000O1O1O1O001O1O1O1O1000O1000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001N1000000000000000001O0000000000000000001O0000000000000000001O00000000000000001O0000000000000000001N1000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O000000000000000O101O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000O10001O001O1O001O1O1O1O001O1O1000O010O1O001O1O010O1O1O001O1O001O1O0O2O1O001O1O001O1N101O1O1O0O2N2O0O2N2N1O2M2O10001O00000000000010O00001O00000010O0001O00001O01O0000000000001O0000000000001O000000000000001M2N2N2N2N2O101N10000O100O101O0O100O10000O3M3M4K4M3L4M3M3LPQ`1"}, "label": "cow painted limo"}]} {"id": 523711, "image": "COCO_train2014_000000523711.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a jersey cow painted stretch limousine with cow , pig , and rooster roof ornaments\"."}], "task": "refering", "answer_template": "The \"a jersey cow painted stretch limousine with cow , pig , and rooster roof ornaments\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 170, 171, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 245, 246, 247, 248], "bbox": [0.264515625, 0.4276041666666667, 0.838703125, 0.6103958333333334], "iscrowd": 0, "area": 20959.757950000003, "rle": {"size": [480, 640], "counts": "i^_21m>2O2M2N3N1N3N1N3M201N101O0O2O001N101O1N2N2N3M2M3N3M2N2M2O1O1N2O1O1N2O1000O1O1O1O001O1O1O1O1000O1000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001N1000000000000000001O0000000000000000001O0000000000000000001O00000000000000001O0000000000000000001N1000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O000000000000000O101O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000000001O0000000000000O10001O001O1O001O1O1O1O001O1O1000O010O1O001O1O010O1O1O001O1O001O1O0O2O1O001O1O001O1N101O1O1O0O2N2O0O2N2N1O2M2O10001O00000000000010O00001O00000010O0001O00001O01O0000000000001O0000000000001O000000000000001M2N2N2N2N2O101N10000O100O101O0O100O10000O3M3M4K4M3L4M3M3LPQ`1"}, "label": "a jersey cow painted stretch limousine with cow , pig , and rooster roof ornaments"}]} {"id": 15809, "image": "COCO_train2014_000000015809.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue yarn bear on the right\"."}], "task": "refering", "answer_template": "The \"blue yarn bear on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 35, 36, 37, 38, 39, 40, 41, 42, 58, 59, 60, 61, 62, 63, 64, 81, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 128, 129, 130, 131, 132, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 226, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 294, 295, 296, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342, 357, 358, 359, 363, 364, 380, 381, 382], "bbox": [0.433375, 0.0035208333333333333, 0.923203125, 0.9980625], "iscrowd": 0, "area": 76544.19944999999, "rle": {"size": [480, 640], "counts": "\\RR44f=X1^Oa0O1N2O1O00001O00000000001O0000000000001O0000000000001O0O10000000001O0000000000001O0000000000001O000000000gJhMjMX2WOmMYNO\\2V2TOXNXNFa2R2POfNXNZOf2P2lNROWNROj2l1iN_OVNhNn2i1eNLVN_NR3e1bN9TNUNX3b1^Nf0SNkM\\3`1ZNQ1SNcM`3\\1VN]1TN[Mc3X1SNd1YNVMb3V1oMj1^NTM`3R1kMR2dNmL`3Q1fMY2hNiL`3n0bM_2nNdL_3n0[Mf2UO]L_3m0VMm2ZOWL_3l0QMT3_OQL_3k0lLZ3EmK]3i0lL]3FkK]3h0kL`3GiK]3g0jLb3IhK\\3g0iLc3JgK\\3f0hLf3KeK[1bNOS2iNi3LdKR1kN6h1jNk3NcKh0WO<[1lNn3OaK`0Aa0Q1nNP40_K8Lf0e0POR42^KN7m0:QOS43mMh0PNSOV44mMf0mMTOY46mMb0d4^O_K?b4@bK<_4CdK6`4J\\5OO1O100O100O1O100O1O100O101N1O100O1O100O100O1O100O1O100O100O001O1O1O1O001O1O1O1O>B2N1O2N2N1O2H7I8I6I8H8I6I8H7J7H7I8K41O001O1O1O1O001O1O100O001O1O1O1O1O001O100O1O001O1O1O1O001O1O100O001M]I`Lk1^3UNcLl1[3TNfLm1W3SNkLn1S3RNnLn1P3SNRMm1l2RNVMo1h2QNYMP2e2oM]MR2a2nM`McM2c1^2i0bMcM5`1W2l0fMdM6^1S2m0hMeM8\\1o1m0kMgM9Z1j1o0nMgM;W1g1P1QNhM;V1b1R1TNhM=T1^1S1VNiM?R1Y1T1ZNjM`0P1U1U1\\NkMb0m0Q1W1_NlMc0k0m0W1cNmMa0l0k0V1eNmMa0n0h0T1iNnM`0n0f0R1lNPN?n0c0Q1POQN=n0c0P1ROQNm0YOTN9P1KbNLo09SO`NlMo1Q2XOQ17SOZ1L`NQ15SO[1L`NR13RO^1L`NR11RO_1LaNQ1OSOb1K_NS1OQOb1L`NS1MPOd1M`NS1KPOe1M`NT1KnNe1NaNS1JnNg1N`NT1HnNh1N`NU1GlNj1O`NU1FkNj10aNU1DjNl11`NV1CiNn10`NV1CiNm11aNV1AhNo11aNX1_OfNQ22aNX1]OeNS23`NY1]ObNU24_NY1\\ObNV25_NY1[O`NW27^NZ1[O]NX29^NZ1YO\\NZ2:^NZ1XOZN[2<]NZ1XOYN]2<\\N[1WOWN^2>\\N[1UOVN`2?[N\\1UORNb2b0ZN\\1SOQNd2c0ZN[1SOPNe2d0XN]1ROnMg2e0XN]1ROkMg2h0WN^1ROgMi2k0VN^1QOdMk2n0UN]1ROaMl2Q1RN]1TO_Ml2T1QN[1UO^Ml2W1PNY1VO]Ml2Z1nMX1YOXMm2`1kMV1ZOUMP3d1gMV1ZORMR3h1dMU1c4jN_KT1b4lN_KR1b4nN^KQ1d4nN]KP1d4PO]Kn0d4RO\\Kn0d4RO]Kl0d4TO\\Kk0f4TO[Kj0g4UOZKi0h4VOXKi0k4UOVKi0l4VOUKi0m4UOTKi0o4UOQKj0S5SOnJk0W5QOjJm0[5oNfJo0_5mNaJR1d5jN]JU1g5gNZJW1k5eNVJY1T6^NlIa1k9N2N2N101N2N2N1O2N2N2O1N2N2N2N2N2N2O1N2N2N1O3M2N2O1N3M2N2N2N3M2O1Nibf0"}, "label": "blue yarn bear on the right"}]} {"id": 15809, "image": "COCO_train2014_000000015809.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a knit bear figure to the right of a duplicate\"."}], "task": "refering", "answer_template": "The \"a knit bear figure to the right of a duplicate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 14, 15, 16, 17, 18, 35, 36, 37, 38, 39, 40, 41, 42, 58, 59, 60, 61, 62, 63, 64, 81, 82, 83, 84, 85, 86, 104, 105, 106, 107, 108, 109, 128, 129, 130, 131, 132, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 219, 220, 221, 222, 223, 224, 225, 226, 242, 243, 244, 245, 246, 247, 265, 266, 267, 268, 269, 270, 271, 288, 289, 290, 291, 292, 293, 294, 295, 296, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 334, 335, 336, 337, 338, 339, 340, 341, 342, 357, 358, 359, 363, 364, 380, 381, 382], "bbox": [0.433375, 0.0035208333333333333, 0.923203125, 0.9980625], "iscrowd": 0, "area": 76544.19944999999, "rle": {"size": [480, 640], "counts": "\\RR44f=X1^Oa0O1N2O1O00001O00000000001O0000000000001O0000000000001O0O10000000001O0000000000001O0000000000001O000000000gJhMjMX2WOmMYNO\\2V2TOXNXNFa2R2POfNXNZOf2P2lNROWNROj2l1iN_OVNhNn2i1eNLVN_NR3e1bN9TNUNX3b1^Nf0SNkM\\3`1ZNQ1SNcM`3\\1VN]1TN[Mc3X1SNd1YNVMb3V1oMj1^NTM`3R1kMR2dNmL`3Q1fMY2hNiL`3n0bM_2nNdL_3n0[Mf2UO]L_3m0VMm2ZOWL_3l0QMT3_OQL_3k0lLZ3EmK]3i0lL]3FkK]3h0kL`3GiK]3g0jLb3IhK\\3g0iLc3JgK\\3f0hLf3KeK[1bNOS2iNi3LdKR1kN6h1jNk3NcKh0WO<[1lNn3OaK`0Aa0Q1nNP40_K8Lf0e0POR42^KN7m0:QOS43mMh0PNSOV44mMf0mMTOY46mMb0d4^O_K?b4@bK<_4CdK6`4J\\5OO1O100O100O1O100O1O100O101N1O100O1O100O100O1O100O1O100O100O001O1O1O1O001O1O1O1O>B2N1O2N2N1O2H7I8I6I8H8I6I8H7J7H7I8K41O001O1O1O1O001O1O100O001O1O1O1O1O001O100O1O001O1O1O1O001O1O100O001M]I`Lk1^3UNcLl1[3TNfLm1W3SNkLn1S3RNnLn1P3SNRMm1l2RNVMo1h2QNYMP2e2oM]MR2a2nM`McM2c1^2i0bMcM5`1W2l0fMdM6^1S2m0hMeM8\\1o1m0kMgM9Z1j1o0nMgM;W1g1P1QNhM;V1b1R1TNhM=T1^1S1VNiM?R1Y1T1ZNjM`0P1U1U1\\NkMb0m0Q1W1_NlMc0k0m0W1cNmMa0l0k0V1eNmMa0n0h0T1iNnM`0n0f0R1lNPN?n0c0Q1POQN=n0c0P1ROQNm0YOTN9P1KbNLo09SO`NlMo1Q2XOQ17SOZ1L`NQ15SO[1L`NR13RO^1L`NR11RO_1LaNQ1OSOb1K_NS1OQOb1L`NS1MPOd1M`NS1KPOe1M`NT1KnNe1NaNS1JnNg1N`NT1HnNh1N`NU1GlNj1O`NU1FkNj10aNU1DjNl11`NV1CiNn10`NV1CiNm11aNV1AhNo11aNX1_OfNQ22aNX1]OeNS23`NY1]ObNU24_NY1\\ObNV25_NY1[O`NW27^NZ1[O]NX29^NZ1YO\\NZ2:^NZ1XOZN[2<]NZ1XOYN]2<\\N[1WOWN^2>\\N[1UOVN`2?[N\\1UORNb2b0ZN\\1SOQNd2c0ZN[1SOPNe2d0XN]1ROnMg2e0XN]1ROkMg2h0WN^1ROgMi2k0VN^1QOdMk2n0UN]1ROaMl2Q1RN]1TO_Ml2T1QN[1UO^Ml2W1PNY1VO]Ml2Z1nMX1YOXMm2`1kMV1ZOUMP3d1gMV1ZORMR3h1dMU1c4jN_KT1b4lN_KR1b4nN^KQ1d4nN]KP1d4PO]Kn0d4RO\\Kn0d4RO]Kl0d4TO\\Kk0f4TO[Kj0g4UOZKi0h4VOXKi0k4UOVKi0l4VOUKi0m4UOTKi0o4UOQKj0S5SOnJk0W5QOjJm0[5oNfJo0_5mNaJR1d5jN]JU1g5gNZJW1k5eNVJY1T6^NlIa1k9N2N2N101N2N2N1O2N2N2O1N2N2N2N2N2N2O1N2N2N1O3M2N2O1N3M2N2N2N3M2O1Nibf0"}, "label": "a knit bear figure to the right of a duplicate"}]} {"id": 540110, "image": "COCO_train2014_000000540110.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"grey lifted truk with red and white signal signs on the tailgate\"."}], "task": "refering", "answer_template": "The \"grey lifted truk with red and white signal signs on the tailgate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [263, 265, 266, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 331, 332, 333, 334, 335, 336, 354, 355, 356, 357, 358, 359, 378, 382], "bbox": [0.41903125, 0.6835655737704918, 0.64228125, 0.9650819672131147], "iscrowd": 0, "area": 14200.8341, "rle": {"size": [488, 640], "counts": "nRP45P?b0_O`0_Ob0^Oa0@`0_O5K4L3N2M4L3N3L3M3M4O001O00001O0O2O000011N100O2N100N2N3jL`D_2b;XMfDf2o;O1O1N2O1C=00O0100000000000O1000000000000001O2N1N2O2N1O000000O1000nN]M_Eb2b:hMTEX2l:SNiDm1V;_N_D`1b;S1000O10000O10000O10000O10000O10000O100O100O100O1O3O0100O10O01OO2N1O2M2O2N1O1O2N1N3N1O21N2M3M2N3M3L4M2N0000O1000O100000O1000O100000O4M4L4RNRDlN1Q1P in this image.", "objects": [{"patches": [263, 265, 266, 285, 286, 287, 288, 289, 290, 308, 309, 310, 311, 312, 313, 331, 332, 333, 334, 335, 336, 354, 355, 356, 357, 358, 359, 378, 382], "bbox": [0.41903125, 0.6835655737704918, 0.64228125, 0.9650819672131147], "iscrowd": 0, "area": 14200.8341, "rle": {"size": [488, 640], "counts": "nRP45P?b0_O`0_Ob0^Oa0@`0_O5K4L3N2M4L3N3L3M3M4O001O00001O0O2O000011N100O2N100N2N3jL`D_2b;XMfDf2o;O1O1N2O1C=00O0100000000000O1000000000000001O2N1N2O2N1O000000O1000nN]M_Eb2b:hMTEX2l:SNiDm1V;_N_D`1b;S1000O10000O10000O10000O10000O10000O100O100O100O1O3O0100O10O01OO2N1O2M2O2N1O1O2N1N3N1O21N2M3M2N3M3L4M2N0000O1000O100000O1000O100000O4M4L4RNRDlN1Q1P in this image.", "objects": [{"patches": [123, 124, 146, 147, 148, 168, 169, 170, 191, 192, 193, 194, 213, 214, 215, 216, 235, 236, 237, 238, 239, 258, 259, 260, 261, 282], "bbox": [0.23746874999999998, 0.3620941176470588, 0.4479218749999999, 0.8310352941176471], "iscrowd": 0, "area": 12741.16945, "rle": {"size": [425, 640], "counts": "_\\o11W=2M2O1N3M2O1N3N1N2N2O2M2O1N3M2O1N3N1N2N3N1iDmNV:U1^EWOa:k0SEAj:`1N3N1N3N1N00100O10O01000O1000O01000000O01000O1000O10O100002M4M3M3K6J4L3M3M3M3M2N3M3M_OkFTMR9o2oFPMn8R3TGmLi8V3XGiLe8Z3\\GdLb8^3_GbL^8a3cG_LY8d3hG[LU8g3mGXLQ8j3oGVLn7m3SHSLi7P4WHPLf7R4[HnKe7R4[HnKe7R4[HoKc7Q4^HoKb7Q4^HoKb7Q4^HoKa7Q4`HPL_7P4aHPL_7P4bHoK]7Q4dHoK\\7P4eHPL[7o3fHRLX7o3hHQLX7n3iHRLW7m3jHSLU7m3nHRLQ7n3RIoKn6P4VImKi6S4[IjKe6V4^IhKa6W4X101O10O01O1O010O1O00100O0_MnFl0S9TOoFi0R9VOQGh0n8YOTGd0m8\\OUG`0m8@UGmF^OZ9 in this image.", "objects": [{"patches": [123, 124, 146, 147, 148, 168, 169, 170, 191, 192, 193, 194, 213, 214, 215, 216, 235, 236, 237, 238, 239, 258, 259, 260, 261, 282], "bbox": [0.23746874999999998, 0.3620941176470588, 0.4479218749999999, 0.8310352941176471], "iscrowd": 0, "area": 12741.16945, "rle": {"size": [425, 640], "counts": "_\\o11W=2M2O1N3M2O1N3N1N2N2O2M2O1N3M2O1N3N1N2N3N1iDmNV:U1^EWOa:k0SEAj:`1N3N1N3N1N00100O10O01000O1000O01000000O01000O1000O10O100002M4M3M3K6J4L3M3M3M3M2N3M3M_OkFTMR9o2oFPMn8R3TGmLi8V3XGiLe8Z3\\GdLb8^3_GbL^8a3cG_LY8d3hG[LU8g3mGXLQ8j3oGVLn7m3SHSLi7P4WHPLf7R4[HnKe7R4[HnKe7R4[HoKc7Q4^HoKb7Q4^HoKb7Q4^HoKa7Q4`HPL_7P4aHPL_7P4bHoK]7Q4dHoK\\7P4eHPL[7o3fHRLX7o3hHQLX7n3iHRLW7m3jHSLU7m3nHRLQ7n3RIoKn6P4VImKi6S4[IjKe6V4^IhKa6W4X101O10O01O1O010O1O00100O0_MnFl0S9TOoFi0R9VOQGh0n8YOTGd0m8\\OUG`0m8@UGmF^OZ9 in this image.", "objects": [{"patches": [105, 128, 129, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 220, 221, 222, 243, 244, 245], "bbox": [0.546953125, 0.32527058823529414, 0.68690625, 0.7317176470588236], "iscrowd": 0, "area": 10677.45315, "rle": {"size": [425, 640], "counts": "Z_a42U=5L4K6K4SOGcD=X;n0K5L5J5K5K6H7G9G9H9mFfLj7b3RH_Lm7d3PH]Lo7f3nG[LQ8g3lGZLT8i3iGXLV8k3gGULY8m3eGTLZ8^4O01O01O001O000100O1OnGVKh7i4VHYKj7f4UH\\Kk7d4RH_Kn7`4QHbKo7]4PHeKP8j4000000001O1O1O001O1N2N101N2N4M2M3M3N2M6J:F3N1O2N2O0O2O1N101O1N101N2O1O1N2O1N2O1O1N2O1N2O3M4eNkEXOY:<`FTOc9`0k1D in this image.", "objects": [{"patches": [105, 128, 129, 150, 151, 152, 153, 173, 174, 175, 176, 196, 197, 198, 199, 220, 221, 222, 243, 244, 245], "bbox": [0.546953125, 0.32527058823529414, 0.68690625, 0.7317176470588236], "iscrowd": 0, "area": 10677.45315, "rle": {"size": [425, 640], "counts": "Z_a42U=5L4K6K4SOGcD=X;n0K5L5J5K5K6H7G9G9H9mFfLj7b3RH_Lm7d3PH]Lo7f3nG[LQ8g3lGZLT8i3iGXLV8k3gGULY8m3eGTLZ8^4O01O01O001O000100O1OnGVKh7i4VHYKj7f4UH\\Kk7d4RH_Kn7`4QHbKo7]4PHeKP8j4000000001O1O1O001O1N2N101N2N4M2M3M3N2M6J:F3N1O2N2O0O2O1N101O1N101N2O1O1N2O1N2O1O1N2O1N2O3M4eNkEXOY:<`FTOc9`0k1D in this image.", "objects": [{"patches": [68, 91, 109, 110, 111, 112, 113, 114, 134, 135, 136, 137, 158, 159, 160, 180, 181, 182], "bbox": [0.758640625, 0.16011709601873536, 1.0, 0.5106791569086652], "iscrowd": 0, "area": 7565.546450000006, "rle": {"size": [427, 640], "counts": "RiZ61Y=2O001N10001O001O0O101O001O001O000O2O001O00001O001N10001O000000000O101O00000000000O2O000000000YE[Oh8e0XG\\Og8e0XG\\Of8e0ZG\\Oe8d0[G]Od8c0\\G^Oc8b0]G_Ob8a0^G@a8`0_GA`8`0_G@a8`0_G@`8a0`G@_8`0aG@_8`0aG@_8`0aG@_8`0aG@_8a0`G_O`8a0`G_O`8a0`G_Oa8`0_G@a8`0_G@a8`0]GCb8=\\GEd8 in this image.", "objects": [{"patches": [68, 91, 109, 110, 111, 112, 113, 114, 134, 135, 136, 137, 158, 159, 160, 180, 181, 182], "bbox": [0.758640625, 0.16011709601873536, 1.0, 0.5106791569086652], "iscrowd": 0, "area": 7565.546450000006, "rle": {"size": [427, 640], "counts": "RiZ61Y=2O001N10001O001O0O101O001O001O000O2O001O00001O001N10001O000000000O101O00000000000O2O000000000YE[Oh8e0XG\\Og8e0XG\\Of8e0ZG\\Oe8d0[G]Od8c0\\G^Oc8b0]G_Ob8a0^G@a8`0_GA`8`0_G@a8`0_G@`8a0`G@_8`0aG@_8`0aG@_8`0aG@_8`0aG@_8a0`G_O`8a0`G_O`8a0`G_Oa8`0_G@a8`0_G@a8`0]GCb8=\\GEd8 in this image.", "objects": [{"patches": [292, 293, 311, 312, 313, 314, 315, 316, 317, 318, 333, 334, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.46792187500000004, 0.8317330210772833, 0.8439687499999999, 0.9992037470725994], "iscrowd": 0, "area": 8114.819999999999, "rle": {"size": [427, 640], "counts": "h_m32X=200O0O2O1O001O1O001N10001O000000001N1000001O0000001N100000001O00000O2O000000001O0O10001O0000001O0O1N2N201O0000001O0000001O000000001O0O100000000000000000000000000000000000000001O000000000000000000001O00000000000000001O0000000000000000001O0000000000000000XOSOQEn0o:ROPEo0P;ROnDo0Q;ROmDP1S;POlDQ1T;oNkDR1U;nNjDS1W;mNgDT1Y;lNfDU1Z;kNeDV1[;jNdDW1\\;iNcDX1];iNaDX1Q;fNUE2HZ1R;eNWE0F\\1S;eNVEOF]1S;eNWE`1h:bNWE^1i:bNWE^1j:aNUE`1k:aNTE_1l:aNSE`1m:aNQE`1o:`NPEa1Q;_NmDb1S;^NlDc1T;]NkDd1U;900001O000000001O000000001O0\\OiDTOW;i0lDWOT;g0mDZOS;d0nD]OS;?oDBQ; in this image.", "objects": [{"patches": [292, 293, 311, 312, 313, 314, 315, 316, 317, 318, 333, 334, 335, 336, 337, 338, 339, 340, 341], "bbox": [0.46792187500000004, 0.8317330210772833, 0.8439687499999999, 0.9992037470725994], "iscrowd": 0, "area": 8114.819999999999, "rle": {"size": [427, 640], "counts": "h_m32X=200O0O2O1O001O1O001N10001O000000001N1000001O0000001N100000001O00000O2O000000001O0O10001O0000001O0O1N2N201O0000001O0000001O000000001O0O100000000000000000000000000000000000000001O000000000000000000001O00000000000000001O0000000000000000001O0000000000000000XOSOQEn0o:ROPEo0P;ROnDo0Q;ROmDP1S;POlDQ1T;oNkDR1U;nNjDS1W;mNgDT1Y;lNfDU1Z;kNeDV1[;jNdDW1\\;iNcDX1];iNaDX1Q;fNUE2HZ1R;eNWE0F\\1S;eNVEOF]1S;eNWE`1h:bNWE^1i:bNWE^1j:aNUE`1k:aNTE_1l:aNSE`1m:aNQE`1o:`NPEa1Q;_NmDb1S;^NlDc1T;]NkDd1U;900001O000000001O000000001O0\\OiDTOW;i0lDWOT;g0mDZOS;d0nD]OS;?oDBQ; in this image.", "objects": [{"patches": [51, 52, 73, 74, 75, 76, 92, 93, 94, 95, 96, 97, 98, 99, 100, 117, 118, 119, 120, 121, 122, 123, 140, 141, 142, 145, 163, 164], "bbox": [0.0075, 0.16461358313817331, 0.3868125, 0.4949414519906324], "iscrowd": 0, "area": 13313.633649999996, "rle": {"size": [427, 640], "counts": "\\V22Y=3M2N3M3M2N1O0000001O000000001O000000001O000000001O000000001O000000001O000000001O0000001O000000001O0000000O2O0VEVOS9j0lFXOS9h0mFYOR9g0nFZOQ9g0mF[OR9e0nF\\OQ9d0oF]OP9c0eFUOQO;X:`0gFWOoN:Y:`0gFWOoN:Y:?hFYOlN:[:=iFZOkN9\\:=iFZOkN9\\:=iFZOjN;\\: in this image.", "objects": [{"patches": [129, 130, 151, 152, 153, 154, 155, 170, 171, 172, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 218, 219, 220, 242], "bbox": [0.3891875, 0.3704918032786885, 0.7663593750000001, 0.6722950819672131], "iscrowd": 0, "area": 11636.21075, "rle": {"size": [427, 640], "counts": "]^X32X=1O2O001O0O2O001O0O2O1O001N101O001N2O001O001O00001O000000001O0000001O00000000000000001O000000001O00001O000SEWOX9i0hFWOX9j0fFXOY9h0gFXOY9h0fFZOY9f0gFZOY9g0eFZO[9f0eFZO[9f0eF[OZ9e0fF[OZ9f0dF[O\\9e0dF[O\\9e0dF[O\\9e0dF\\O[9e0cF\\O]9d0cF\\O]9d0bF]O^9c0bF]O_9b0aF^O_9b0`F@_9a0_F@a9`0^FAb9?]FBc9>\\FCd9=ZFEf9;YFGf99YFHg98XFIh97WFJi96UFMj93UFNk92TFOl91SF0m90RF1n9OPF3P:MnE5R:KlE7T:HkE:U:FiEY:BeE`0[:B`Ea0`:U1000O1M3N2N2O1O1O1O1O1O1N2O1O1O1O1O100O0010O010O0100O010O010O010O10O0100O010O1O00100O00100O00100OQOZN[Fe1f9]NWFd1i9\\NWF^1XO`Nb:3UF]1ZO^Nb:5SF^1Q:cNmE^1S:cNhEa1X:`NcEd1]:\\N_Eh1a:`01000000000000000000000O10000000000000gNZEMf:NaEO`:MfE1[:KgE6Y:GjE9V:EkE in this image.", "objects": [{"patches": [129, 130, 151, 152, 153, 154, 155, 170, 171, 172, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 218, 219, 220, 242], "bbox": [0.3891875, 0.3704918032786885, 0.7663593750000001, 0.6722950819672131], "iscrowd": 0, "area": 11636.21075, "rle": {"size": [427, 640], "counts": "]^X32X=1O2O001O0O2O001O0O2O1O001N101O001N2O001O001O00001O000000001O0000001O00000000000000001O000000001O00001O000SEWOX9i0hFWOX9j0fFXOY9h0gFXOY9h0fFZOY9f0gFZOY9g0eFZO[9f0eFZO[9f0eF[OZ9e0fF[OZ9f0dF[O\\9e0dF[O\\9e0dF[O\\9e0dF\\O[9e0cF\\O]9d0cF\\O]9d0bF]O^9c0bF]O_9b0aF^O_9b0`F@_9a0_F@a9`0^FAb9?]FBc9>\\FCd9=ZFEf9;YFGf99YFHg98XFIh97WFJi96UFMj93UFNk92TFOl91SF0m90RF1n9OPF3P:MnE5R:KlE7T:HkE:U:FiEY:BeE`0[:B`Ea0`:U1000O1M3N2N2O1O1O1O1O1O1N2O1O1O1O1O100O0010O010O0100O010O010O010O10O0100O010O1O00100O00100O00100OQOZN[Fe1f9]NWFd1i9\\NWF^1XO`Nb:3UF]1ZO^Nb:5SF^1Q:cNmE^1S:cNhEa1X:`NcEd1]:\\N_Eh1a:`01000000000000000000000O10000000000000gNZEMf:NaEO`:MfE1[:KgE6Y:GjE9V:EkE in this image.", "objects": [{"patches": [70, 92, 93, 94, 115, 116, 117, 118, 140, 141, 142, 164, 165, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 234, 235, 258, 281], "bbox": [0.01490625, 0.26966321243523317, 0.336125, 0.8741450777202073], "iscrowd": 0, "area": 9605.331950000002, "rle": {"size": [386, 640], "counts": "kl3 in this image.", "objects": [{"patches": [70, 92, 93, 94, 115, 116, 117, 118, 140, 141, 142, 164, 165, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 234, 235, 258, 281], "bbox": [0.01490625, 0.26966321243523317, 0.336125, 0.8741450777202073], "iscrowd": 0, "area": 9605.331950000002, "rle": {"size": [386, 640], "counts": "kl3 in this image.", "objects": [{"patches": [122, 123, 145, 146, 168, 169, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 281, 282, 283, 284], "bbox": [0.239828125, 0.37326424870466324, 0.44195312500000006, 0.896658031088083], "iscrowd": 0, "area": 8639.549699999998, "rle": {"size": [386, 640], "counts": "aoi11j;=D1O2SL_JS1d5gNbJh0Q6QOUJb0X6XOnI;]6hNaJl0c5[NfHj0i1>l5jNXHh0o12R64oI@\\6`0fIUOh6g0YIYOj6d0UI^Oh6e0WI[Of6e0\\I\\Oa6e0aI\\O\\6c0gI]OV6c0lI_OQ6a0QJ_Ol5MSHLV27e5JRK5m4HXK7[8N2O1O000O01O10O0100O00100O10O01O010O010O0010O010O01O010O010O0010O01[YV4"}, "label": "the giraffe looking at the wall"}]} {"id": 171484, "image": "COCO_train2014_000000171484.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a young giraffe in a zoo , facing away from the camera toward an adult giraffe\"."}], "task": "refering", "answer_template": "The \"a young giraffe in a zoo , facing away from the camera toward an adult giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 145, 146, 168, 169, 191, 192, 212, 213, 214, 215, 235, 236, 237, 238, 239, 240, 258, 259, 260, 261, 281, 282, 283, 284], "bbox": [0.239828125, 0.37326424870466324, 0.44195312500000006, 0.896658031088083], "iscrowd": 0, "area": 8639.549699999998, "rle": {"size": [386, 640], "counts": "aoi11j;=D1O2SL_JS1d5gNbJh0Q6QOUJb0X6XOnI;]6hNaJl0c5[NfHj0i1>l5jNXHh0o12R64oI@\\6`0fIUOh6g0YIYOj6d0UI^Oh6e0WI[Of6e0\\I\\Oa6e0aI\\O\\6c0gI]OV6c0lI_OQ6a0QJ_Ol5MSHLV27e5JRK5m4HXK7[8N2O1O000O01O10O0100O00100O10O01O010O010O0010O010O01O010O010O0010O01[YV4"}, "label": "a young giraffe in a zoo , facing away from the camera toward an adult giraffe"}]} {"id": 171484, "image": "COCO_train2014_000000171484.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the giraffe on the far right\"."}], "task": "refering", "answer_template": "The \"the giraffe on the far right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [113, 135, 136, 137, 157, 158, 159, 160, 179, 180, 181, 201, 202, 203, 204, 224, 225, 226, 227, 247, 248, 249, 250, 270, 271, 272, 293, 295], "bbox": [0.7496562499999999, 0.3325906735751295, 0.97990625, 0.8719170984455958], "iscrowd": 0, "area": 9905.386800000002, "rle": {"size": [386, 640], "counts": "lVe54i;7J6I7]OAZEf0b:`0QFnNg8X2N3M2M3N2M3N2O1N3NiN_GoN_8Q1RH`Nm7_1dHRNZ7n1]12Mk0VOc0\\O2O1N1O200O01WNmGCR8 in this image.", "objects": [{"patches": [113, 135, 136, 137, 157, 158, 159, 160, 179, 180, 181, 201, 202, 203, 204, 224, 225, 226, 227, 247, 248, 249, 250, 270, 271, 272, 293, 295], "bbox": [0.7496562499999999, 0.3325906735751295, 0.97990625, 0.8719170984455958], "iscrowd": 0, "area": 9905.386800000002, "rle": {"size": [386, 640], "counts": "lVe54i;7J6I7]OAZEf0b:`0QFnNg8X2N3M2M3N2M3N2O1N3NiN_GoN_8Q1RH`Nm7_1dHRNZ7n1]12Mk0VOc0\\O2O1N1O200O01WNmGCR8 in this image.", "objects": [{"patches": [61, 62, 84, 85, 107, 108, 131, 154, 155, 177, 178, 179, 199, 200, 201, 223, 224, 246, 247, 269, 270, 292], "bbox": [0.6448437499999999, 0.171139896373057, 0.78703125, 0.8933678756476683], "iscrowd": 0, "area": 8283.401200000004, "rle": {"size": [386, 640], "counts": "Xik44k;3O1O1O1O1O1O1O1O001O1O1N2O1O0O2O1N2O1N2O1O1N2N20000001O00^H_Oc3a0QL0e35WL4a3O\\L9^3G`L?EYNV1Z1QOe0ZOaN_1k0lNR1VObNj1=hN]1oN`NZ22]NP2gNYNm2FmMe4Y2UKUM\\5k2hJlL^5S3fJeL_5Z3d10TOk0O3L5kH_Kc6f4YI_KW52cKe4RO\\KW5>WKi4d4]1M4N4L5L3kLdH^1[2SMc24ZLi2KYN_5g1XJ`Nj5`1mIhNT6X1dInN]6S1[ISOg6T31N2N2O1N2ROWI_Lj6[3iIWLY6a3X1[Of0E;G8I7J6J5O2N1N3N2M2O2N2M2O2N1N3N4K5L4L3L5ITRc1"}, "label": "the giraffe standing to the right of two giraffes and to the left of one giraffe"}]} {"id": 171484, "image": "COCO_train2014_000000171484.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"tallest giraffe\"."}], "task": "refering", "answer_template": "The \"tallest giraffe\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [61, 62, 84, 85, 107, 108, 131, 154, 155, 177, 178, 179, 199, 200, 201, 223, 224, 246, 247, 269, 270, 292], "bbox": [0.6448437499999999, 0.171139896373057, 0.78703125, 0.8933678756476683], "iscrowd": 0, "area": 8283.401200000004, "rle": {"size": [386, 640], "counts": "Xik44k;3O1O1O1O1O1O1O1O001O1O1N2O1O0O2O1N2O1N2O1O1N2N20000001O00^H_Oc3a0QL0e35WL4a3O\\L9^3G`L?EYNV1Z1QOe0ZOaN_1k0lNR1VObNj1=hN]1oN`NZ22]NP2gNYNm2FmMe4Y2UKUM\\5k2hJlL^5S3fJeL_5Z3d10TOk0O3L5kH_Kc6f4YI_KW52cKe4RO\\KW5>WKi4d4]1M4N4L5L3kLdH^1[2SMc24ZLi2KYN_5g1XJ`Nj5`1mIhNT6X1dInN]6S1[ISOg6T31N2N2O1N2ROWI_Lj6[3iIWLY6a3X1[Of0E;G8I7J6J5O2N1N3N2M2O2N2M2O2N1N3N4K5L4L3L5ITRc1"}, "label": "tallest giraffe"}]} {"id": 81372, "image": "COCO_train2014_000000081372.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small blonde teddy bear with a pink bow sitting in the lap of a larger teddy bear\"."}], "task": "refering", "answer_template": "The \"a small blonde teddy bear with a pink bow sitting in the lap of a larger teddy bear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 171, 191, 192, 193, 194, 195, 215, 216, 217, 218, 239, 240], "bbox": [0.33046875, 0.3857611241217799, 0.48864062500000005, 0.7190866510538642], "iscrowd": 0, "area": 9768.123399999997, "rle": {"size": [427, 640], "counts": "[ch2;l<A?B8J5L4L4K2O1O1N3N1O1O1O1O1O1O2N1O100Og0ZO0O2O0O101O000O101O000000O1000O10O1000000O10000L\\FaLd9n2]F]Mg9b2^FYMd9f2_FWMb9h2a0N2O^F^M]8c2XGeMj8]2TGaMn8`2RGaMl8`2TGbMi8`2VGbMg8b2WG_Mf8f2VGYMj8i2UGWMj8j2VGUMj8l2VGSMj8n2WGPMi8Q3WGoLh8R3XGmLh8T3XGkLh8V3XGjLg8W3YGhLg8Y3ZGeLf8\\3ZGdLe8]3[GbLe8_3[G`Le8a3[G_Ld8b3]G\\Lc8e3]GZLc8f3^GYLb8h3^GXLa8h3`GWL`8j3`GUL`8k3bGTL]8m3`0001O00`MjFl0U9PORG:2_Nl8S1ZG9OaNf8R1cG9R9[O]Ga0f8WOdGd0a8UOdGi0`8QOeGk0`:N2M3M4M2M3M5L5JWSX4"}, "label": "a small blonde teddy bear with a pink bow sitting in the lap of a larger teddy bear"}]} {"id": 81372, "image": "COCO_train2014_000000081372.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a teddy bear with a pink ribbon around its neck sitting on a bigger teddy bear ' s lap\"."}], "task": "refering", "answer_template": "The \"a teddy bear with a pink ribbon around its neck sitting on a bigger teddy bear ' s lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 145, 146, 147, 148, 168, 169, 170, 171, 191, 192, 193, 194, 195, 215, 216, 217, 218, 239, 240], "bbox": [0.33046875, 0.3857611241217799, 0.48864062500000005, 0.7190866510538642], "iscrowd": 0, "area": 9768.123399999997, "rle": {"size": [427, 640], "counts": "[ch2;l<A?B8J5L4L4K2O1O1N3N1O1O1O1O1O1O2N1O100Og0ZO0O2O0O101O000O101O000000O1000O10O1000000O10000L\\FaLd9n2]F]Mg9b2^FYMd9f2_FWMb9h2a0N2O^F^M]8c2XGeMj8]2TGaMn8`2RGaMl8`2TGbMi8`2VGbMg8b2WG_Mf8f2VGYMj8i2UGWMj8j2VGUMj8l2VGSMj8n2WGPMi8Q3WGoLh8R3XGmLh8T3XGkLh8V3XGjLg8W3YGhLg8Y3ZGeLf8\\3ZGdLe8]3[GbLe8_3[G`Le8a3[G_Ld8b3]G\\Lc8e3]GZLc8f3^GYLb8h3^GXLa8h3`GWL`8j3`GUL`8k3bGTL]8m3`0001O00`MjFl0U9PORG:2_Nl8S1ZG9OaNf8R1cG9R9[O]Ga0f8WOdGd0a8UOdGi0`8QOeGk0`:N2M3M4M2M3M5L5JWSX4"}, "label": "a teddy bear with a pink ribbon around its neck sitting on a bigger teddy bear ' s lap"}]} {"id": 81372, "image": "COCO_train2014_000000081372.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large teddy bear with a smaller one in its lap\"."}], "task": "refering", "answer_template": "The \"a large teddy bear with a smaller one in its lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 76, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 287, 288], "bbox": [0.2583125, 0.14892271662763468, 0.618765625, 0.8488056206088994], "iscrowd": 0, "area": 41865.930149999986, "rle": {"size": [427, 640], "counts": "]PU22U=`0@`0@6J5L4K5K4L5L4N1O2M3N2M2O2M3N1O2M3N1N2O1O1N101N200O100O100O10O0:G:E9JS3GfL?8HS3HeLa09ER3IgLb08BS3KfLc09_OQ3MhLd08]OQ3NgLf09YOR3NhLi07UOS30iLl05nNV34gLn04iNX37gLP12dN[3:fLQ10`N_3;dLU1N[Ne37dL]1IVNh6j1^2O0105J0001O01O01O0000001O01O01O0O1O2N1O1O2N1O1N3N1O1O001O1O1N101O1N2O0O2O1O2M2O1O2M2N2M4L5Jo_U3"}, "label": "a large teddy bear with a smaller one in its lap"}]} {"id": 81372, "image": "COCO_train2014_000000081372.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large stuffed mama bear seated on bench with baby bear in her lap\"."}], "task": "refering", "answer_template": "The \"large stuffed mama bear seated on bench with baby bear in her lap\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 57, 76, 77, 78, 79, 80, 81, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 144, 145, 146, 147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 197, 198, 214, 215, 216, 217, 218, 219, 220, 221, 237, 238, 239, 240, 241, 242, 260, 261, 262, 263, 264, 265, 283, 284, 285, 286, 287, 288], "bbox": [0.2583125, 0.14892271662763468, 0.618765625, 0.8488056206088994], "iscrowd": 0, "area": 41865.930149999986, "rle": {"size": [427, 640], "counts": "]PU22U=`0@`0@6J5L4K5K4L5L4N1O2M3N2M2O2M3N1O2M3N1N2O1O1N101N200O100O100O10O0:G:E9JS3GfL?8HS3HeLa09ER3IgLb08BS3KfLc09_OQ3MhLd08]OQ3NgLf09YOR3NhLi07UOS30iLl05nNV34gLn04iNX37gLP12dN[3:fLQ10`N_3;dLU1N[Ne37dL]1IVNh6j1^2O0105J0001O01O01O0000001O01O01O0O1O2N1O1O2N1O1N3N1O1O001O1O1N101O1N2O0O2O1O2M2O1O2M2N2M4L5Jo_U3"}, "label": "large stuffed mama bear seated on bench with baby bear in her lap"}]} {"id": 155107, "image": "COCO_train2014_000000155107.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white computer monitor that is in use\"."}], "task": "refering", "answer_template": "The \"a white computer monitor that is in use\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 165, 166, 167, 168, 169, 170, 171], "bbox": [0.16459375, 0.0, 0.6459531249999999, 0.4407531380753138], "iscrowd": 0, "area": 59138.82460000001, "rle": {"size": [478, 640], "counts": "`Pa19d>o2RMd2[M5L3M4L3M4L3M0000001OO100000000000000000000000000O100000000000000000000000000O1000000000000000000000001O0O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O10001O000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O1000000000001O0000000000000O100000000000000000000000000O1000000000000000000G9G9F:G9F:G9F:G9F:G9F:G9F:G9F:G9F:G9FQoY3"}, "label": "a white computer monitor that is in use"}]} {"id": 155107, "image": "COCO_train2014_000000155107.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a flat screen computer screen on a desk\"."}], "task": "refering", "answer_template": "The \"a flat screen computer screen on a desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 165, 166, 167, 168, 169, 170, 171], "bbox": [0.16459375, 0.0, 0.6459531249999999, 0.4407531380753138], "iscrowd": 0, "area": 59138.82460000001, "rle": {"size": [478, 640], "counts": "`Pa19d>o2RMd2[M5L3M4L3M4L3M0000001OO100000000000000000000000000O100000000000000000000000000O1000000000000000000000001O0O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O10001O000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O100000000000000000000000000O1000000000001O0000000000000O100000000000000000000000000O1000000000000000000G9G9F:G9F:G9F:G9F:G9F:G9F:G9F:G9F:G9FQoY3"}, "label": "a flat screen computer screen on a desk"}]} {"id": 155107, "image": "COCO_train2014_000000155107.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white and gray wireless keyboard for a desktop computer\"."}], "task": "refering", "answer_template": "The \"a white and gray wireless keyboard for a desktop computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 246, 261, 262, 263, 264, 265, 266, 267, 268, 269, 286, 287, 288, 289], "bbox": [0.35750000000000004, 0.559560669456067, 0.750234375, 0.759560669456067], "iscrowd": 0, "area": 12719.843049999994, "rle": {"size": [478, 640], "counts": "ZV[33j>2M4M3L2O1O0010O100O10O0100001O1O1N2O1O1O1O1N101O1O1O1O1N2O1O1O1O1O1N101O1O1O1N2O1O1O1O1O1N2O1O001O1O1N2O1O1O1O1N2O1O1O1O001N2O1O1O1O1N2O1O00O1000O10O1000000O01000O1000O10O10000O10O1000O1000O10O10000O10O1000O10000O0100000O10O1000O10000O0100000O10O10O1000000O0100000O1000O01000000O01000O100000O01000000O01000O100000O010000O1000O10O100000O010000O1000O10O10000O10O1000O1000O11N2O001N2O001O1N101O1N2O001O1N101O1O0O2O1O1N101O1O0O2O1O0O2O1O1O0O2O1O001N2O001N2O1O001N2O001N2O1O001N2O001O1N101O1N2O001O1N101OX\\Z2"}, "label": "a white and gray wireless keyboard for a desktop computer"}]} {"id": 155107, "image": "COCO_train2014_000000155107.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white keyboard sitting in front of an imac computer on a desk\"."}], "task": "refering", "answer_template": "The \"white keyboard sitting in front of an imac computer on a desk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [219, 220, 221, 222, 238, 239, 240, 241, 242, 243, 244, 245, 246, 261, 262, 263, 264, 265, 266, 267, 268, 269, 286, 287, 288, 289], "bbox": [0.35750000000000004, 0.559560669456067, 0.750234375, 0.759560669456067], "iscrowd": 0, "area": 12719.843049999994, "rle": {"size": [478, 640], "counts": "ZV[33j>2M4M3L2O1O0010O100O10O0100001O1O1N2O1O1O1O1N101O1O1O1O1N2O1O1O1O1O1N101O1O1O1N2O1O1O1O1O1N2O1O001O1O1N2O1O1O1O1N2O1O1O1O001N2O1O1O1O1N2O1O00O1000O10O1000000O01000O1000O10O10000O10O1000O1000O10O10000O10O1000O10000O0100000O10O1000O10000O0100000O10O10O1000000O0100000O1000O01000000O01000O100000O01000000O01000O100000O010000O1000O10O100000O010000O1000O10O10000O10O1000O1000O11N2O001N2O001O1N101O1N2O001O1N101O1O0O2O1O1N101O1O0O2O1O0O2O1O1O0O2O1O001N2O001N2O1O001N2O001N2O1O001N2O001O1N101O1N2O001O1N101OX\\Z2"}, "label": "white keyboard sitting in front of an imac computer on a desk"}]} {"id": 155107, "image": "COCO_train2014_000000155107.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white color laptop is on a table\"."}], "task": "refering", "answer_template": "The \"a white color laptop is on a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229], "bbox": [0.77540625, 0.2378242677824268, 1.0, 0.5726569037656905], "iscrowd": 0, "area": 19463.193400000004, "rle": {"size": [478, 640], "counts": "WfW73d>8H8G9H8G9H8H8^DZNh9i1PF`Nn9a1iEhNV:Y1bEPO\\:Q1[EYOc:h0UE@j:\\2N2N2N101N2N1O2N1O0000000000010O00000000001O000001O0001O0000000000001O01O00000001O0000000001O01O00000000001O000001O0001O00000000000010O0000000001O00000000010O0000000000001O0001O000001O00000000001O01O0000000001O00000001O01O00000000001O000001O000001O000000000010OQL"}, "label": "a white color laptop is on a table"}]} {"id": 155107, "image": "COCO_train2014_000000155107.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a small white laptop\"."}], "task": "refering", "answer_template": "The \"a small white laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 225, 226, 227, 228, 229], "bbox": [0.77540625, 0.2378242677824268, 1.0, 0.5726569037656905], "iscrowd": 0, "area": 19463.193400000004, "rle": {"size": [478, 640], "counts": "WfW73d>8H8G9H8G9H8H8^DZNh9i1PF`Nn9a1iEhNV:Y1bEPO\\:Q1[EYOc:h0UE@j:\\2N2N2N101N2N1O2N1O0000000000010O00000000001O000001O0001O0000000000001O01O00000001O0000000001O01O00000000001O000001O0001O00000000000010O0000000001O00000000010O0000000000001O0001O000001O00000000001O01O0000000001O00000001O01O00000000001O000001O000001O000000000010OQL"}, "label": "a small white laptop"}]} {"id": 155107, "image": "COCO_train2014_000000155107.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a gray laptop next a white mac computer\"."}], "task": "refering", "answer_template": "The \"a gray laptop next a white mac computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 69, 70, 92, 93, 115, 116, 138, 139, 161, 162, 184, 185, 186, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 322, 323, 324, 345], "bbox": [0.0084375, 0.1297907949790795, 0.362484375, 0.9085774058577406], "iscrowd": 0, "area": 30338.856949999998, "rle": {"size": [478, 640], "counts": "e\\2U1i=^:bE00O1O100O100O1O100O1O010O100O1O100O1O100O4M5J7I7J5J7I7J5J7J6I6J7J6I6J7J6I6K6I7I6K6I7J5J7I7J5J7I7J5J7J6I6J6K1N1O100O101N1O100O2N100O100O2N100O1O2O0O100O1O2O0O1O101N100O1O101N100O1O101N1O100O2O0O1O100O2N100O101N1O100O1O2O0O100O2N100O1O101N100O1O101N1O100O2O0O1O100O2O0O1O101N1O100O101N1O100O2N100O100O2N100O1O101N100O1O2O0O1O100O2O0O1O101N1O2O0O101N1O2O0O2N100O2O0O2N100O2O0O2N100O2N101N100O2N101N1O2O0O101N1O2O0O1O2O0O2O0O1O2O0O2N100O2O0O2N101N1O10Ydn5"}, "label": "a gray laptop next a white mac computer"}]} {"id": 155107, "image": "COCO_train2014_000000155107.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"silver laptop computer\"."}], "task": "refering", "answer_template": "The \"silver laptop computer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 47, 69, 70, 92, 93, 115, 116, 138, 139, 161, 162, 184, 185, 186, 207, 208, 209, 210, 211, 212, 213, 230, 231, 232, 233, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 259, 260, 276, 277, 278, 279, 280, 281, 299, 300, 301, 302, 322, 323, 324, 345], "bbox": [0.0084375, 0.1297907949790795, 0.362484375, 0.9085774058577406], "iscrowd": 0, "area": 30338.856949999998, "rle": {"size": [478, 640], "counts": "e\\2U1i=^:bE00O1O100O100O1O100O1O010O100O1O100O1O100O4M5J7I7J5J7I7J5J7J6I6J7J6I6J7J6I6K6I7I6K6I7J5J7I7J5J7I7J5J7J6I6J6K1N1O100O101N1O100O2N100O100O2N100O1O2O0O100O1O2O0O1O101N100O1O101N100O1O101N1O100O2O0O1O100O2N100O101N1O100O1O2O0O100O2N100O1O101N100O1O101N1O100O2O0O1O100O2O0O1O101N1O100O101N1O100O2N100O100O2N100O1O101N100O1O2O0O1O100O2O0O1O101N1O2O0O101N1O2O0O2N100O2O0O2N100O2O0O2N100O2N101N100O2N101N1O2O0O101N1O2O0O1O2O0O2O0O1O2O0O2N100O2O0O2N101N1O10Ydn5"}, "label": "silver laptop computer"}]} {"id": 7653, "image": "COCO_train2014_000000007653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the speedboat in the water\"."}], "task": "refering", "answer_template": "The \"the speedboat in the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293], "bbox": [0.193015625, 0.548467583497053, 0.773859375, 0.6855402750491159], "iscrowd": 0, "area": 16983.217099999994, "rle": {"size": [509, 640], "counts": "_mm12k?3M3M4L3M3M3M3M3M3M3M3M3M4L2N1O001O0000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000001O0001O0000000000000000000000000000001O00000000000000000000000001O000001O0000000000000000000000000M3M3M4K4N20000000000001O0000000000000001O0000000000000000000000000000010O0000000000000000000000000000001O000000000001O00000000000000000001O000000000000000000000001O00003N2M2N1O0000001O000000001O00000000010O000000001O000000001O00000000010O000000001O00000000001O000000001O000000000O2O000000001O00000000001O000000001O00000000001O0000000O2O00000000001O000000001O00000000001O000000001O000000000O2O000000001O00000000001O000000001O00000000001O0000000O10000^OXAOh>N_ANa>ObA1_>LcA4]>LcA4]>KdA5\\>KdA5\\>KdA5R?000O1O1O1O1O1M3L4L4L4N20000O1000O108HYhW2"}, "label": "the speedboat in the water"}]} {"id": 7653, "image": "COCO_train2014_000000007653.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a boat\"."}], "task": "refering", "answer_template": "The \"a boat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [216, 217, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293], "bbox": [0.193015625, 0.548467583497053, 0.773859375, 0.6855402750491159], "iscrowd": 0, "area": 16983.217099999994, "rle": {"size": [509, 640], "counts": "_mm12k?3M3M4L3M3M3M3M3M3M3M3M3M4L2N1O001O0000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000000000000000001O0001O0000000000000000000000000000001O00000000000000000000000001O000001O0000000000000000000000000M3M3M4K4N20000000000001O0000000000000001O0000000000000000000000000000010O0000000000000000000000000000001O000000000001O00000000000000000001O000000000000000000000001O00003N2M2N1O0000001O000000001O00000000010O000000001O000000001O00000000010O000000001O00000000001O000000001O000000000O2O000000001O00000000001O000000001O00000000001O0000000O2O00000000001O000000001O00000000001O000000001O000000000O2O000000001O00000000001O000000001O00000000001O0000000O10000^OXAOh>N_ANa>ObA1_>LcA4]>LcA4]>KdA5\\>KdA5\\>KdA5R?000O1O1O1O1O1M3L4L4L4N20000O1000O108HYhW2"}, "label": "a boat"}]} {"id": 40433, "image": "COCO_train2014_000000040433.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"smaller gray elephant in sunlight\"."}], "task": "refering", "answer_template": "The \"smaller gray elephant in sunlight\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 215, 216, 217, 239, 240, 263], "bbox": [0.27365625, 0.3051875, 0.5033749999999999, 0.6565416666666667], "iscrowd": 0, "area": 15054.707099999996, "rle": {"size": [480, 640], "counts": "gVb2;`>7J5K6J5K6J5J7M2N3N1N3N1O0O2O1O0O2O001N2O0O200O010O0100O010O010O0010O012N1N2O2M2N3L3N2M4M2N9F;F9G0O10O1000O10O10000O001O1O1O001O10000O010O10000O100O1001N100000001O1O2N1O1N2O1O1O1O1O1O1O1O1O2N3M2N2O1O2N1N3N3M2N1O1OO00100O010O0100O010O010O01O001O1O001O1O1mNWFmLj9i2bFTM_9g2hFVMY9d2oFXMS9c2TGYMn8b2k1M3L4L4L4L4M3L5K6J6K5J8H;E:Fbkd4"}, "label": "smaller gray elephant in sunlight"}]} {"id": 40433, "image": "COCO_train2014_000000040433.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant out from under the shade with two short husks showing\"."}], "task": "refering", "answer_template": "The \"an elephant out from under the shade with two short husks showing\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [121, 122, 123, 124, 125, 144, 145, 146, 147, 148, 149, 167, 168, 169, 170, 171, 172, 191, 192, 193, 194, 195, 215, 216, 217, 239, 240, 263], "bbox": [0.27365625, 0.3051875, 0.5033749999999999, 0.6565416666666667], "iscrowd": 0, "area": 15054.707099999996, "rle": {"size": [480, 640], "counts": "gVb2;`>7J5K6J5K6J5J7M2N3N1N3N1O0O2O1O0O2O001N2O0O200O010O0100O010O010O0010O012N1N2O2M2N3L3N2M4M2N9F;F9G0O10O1000O10O10000O001O1O1O001O10000O010O10000O100O1001N100000001O1O2N1O1N2O1O1O1O1O1O1O1O1O2N3M2N2O1O2N1N3N3M2N1O1OO00100O010O0100O010O010O01O001O1O001O1O1mNWFmLj9i2bFTM_9g2hFVMY9d2oFXMS9c2TGYMn8b2k1M3L4L4L4L4M3L5K6J6K5J8H;E:Fbkd4"}, "label": "an elephant out from under the shade with two short husks showing"}]} {"id": 40433, "image": "COCO_train2014_000000040433.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant which is standing on the right side of the two elephants\"."}], "task": "refering", "answer_template": "The \"an elephant which is standing on the right side of the two elephants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [82, 83, 84, 104, 105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 225, 226, 227, 243, 244, 245, 246, 248, 249, 271, 272], "bbox": [0.512890625, 0.21995833333333334, 0.8820625, 0.6543541666666666], "iscrowd": 0, "area": 31400.000099999994, "rle": {"size": [480, 640], "counts": "S]j41n>1O1O1O100O1O1O1O1O100O1N2M6K4fN_OnCf0o;@iCf0R in this image.", "objects": [{"patches": [82, 83, 84, 104, 105, 106, 107, 108, 109, 110, 127, 128, 129, 130, 131, 132, 133, 134, 150, 151, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 178, 179, 180, 196, 197, 198, 199, 200, 201, 202, 203, 220, 221, 222, 223, 225, 226, 227, 243, 244, 245, 246, 248, 249, 271, 272], "bbox": [0.512890625, 0.21995833333333334, 0.8820625, 0.6543541666666666], "iscrowd": 0, "area": 31400.000099999994, "rle": {"size": [480, 640], "counts": "S]j41n>1O1O1O100O1O1O1O1O100O1N2M6K4fN_OnCf0o;@iCf0R in this image.", "objects": [{"patches": [74, 75, 89, 90, 91, 105, 106, 107, 120, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 141, 142, 152, 153, 154, 155, 156, 157, 168, 169, 170, 171, 172, 173, 185, 186, 187, 188, 189, 201, 202, 203, 204, 205, 217, 218, 219, 220, 221, 232, 233, 234, 235, 236, 237, 238, 248, 249, 250, 251, 252, 253, 254, 265, 266, 268, 269, 281, 282, 284, 285, 286, 297, 298, 300, 301, 302, 312, 313, 317, 318, 328, 329, 333, 334, 335, 344, 350, 351], "bbox": [0.48954441913439634, 0.193515625, 1.0, 0.944140625], "iscrowd": 0, "area": 49010.42950000001, "rle": {"size": [640, 439], "counts": "VnV41nc02N2N1O2N2N2cF2eNOl0b0QO@`0o0^OSO3\\1KeN4]1JcN6`1GaN8a1F`N9b1WKkM>d0Y4e1WKgM?e0Y4f1VKgM?d0Z4g1VKeM?e0Z4i1TKcMa0e0Z4j1TKaMa0f0Y4k1UK`M`0g0Z4j1UKaM?f0[4j1SKcMa0d0[4j1QKeMc0b0[4j1PKfM3CPMn0k7k1oJhM4APMm0l7j1nJkM5@oLl0m7j1mJlM5AoLj0n7j1kJnM7@nLi0o7j1jJoM8@mLg0P8l1hJQNROQOVO>mNe0R8l1fJSNoNAnNS1\\7Y1WIF7bNgN?k7Z1YHf0l0Fk6EVHk0g0EQ7BTHP1b0CZ7^OQHU1nKb2Q8nLTL in this image.", "objects": [{"patches": [74, 75, 89, 90, 91, 105, 106, 107, 120, 121, 122, 123, 124, 125, 136, 137, 138, 139, 140, 141, 142, 152, 153, 154, 155, 156, 157, 168, 169, 170, 171, 172, 173, 185, 186, 187, 188, 189, 201, 202, 203, 204, 205, 217, 218, 219, 220, 221, 232, 233, 234, 235, 236, 237, 238, 248, 249, 250, 251, 252, 253, 254, 265, 266, 268, 269, 281, 282, 284, 285, 286, 297, 298, 300, 301, 302, 312, 313, 317, 318, 328, 329, 333, 334, 335, 344, 350, 351], "bbox": [0.48954441913439634, 0.193515625, 1.0, 0.944140625], "iscrowd": 0, "area": 49010.42950000001, "rle": {"size": [640, 439], "counts": "VnV41nc02N2N1O2N2N2cF2eNOl0b0QO@`0o0^OSO3\\1KeN4]1JcN6`1GaN8a1F`N9b1WKkM>d0Y4e1WKgM?e0Y4f1VKgM?d0Z4g1VKeM?e0Z4i1TKcMa0e0Z4j1TKaMa0f0Y4k1UK`M`0g0Z4j1UKaM?f0[4j1SKcMa0d0[4j1QKeMc0b0[4j1PKfM3CPMn0k7k1oJhM4APMm0l7j1nJkM5@oLl0m7j1mJlM5AoLj0n7j1kJnM7@nLi0o7j1jJoM8@mLg0P8l1hJQNROQOVO>mNe0R8l1fJSNoNAnNS1\\7Y1WIF7bNgN?k7Z1YHf0l0Fk6EVHk0g0EQ7BTHP1b0CZ7^OQHU1nKb2Q8nLTL in this image.", "objects": [{"patches": [39, 54, 55, 56, 68, 69, 70, 71, 72, 82, 83, 84, 85, 86, 87, 98, 99, 100, 101, 102, 103, 113, 114, 115, 116, 117, 118, 119, 120, 130, 131, 132, 133, 134, 135, 136, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 211, 212, 213, 214, 228, 229, 230, 244, 245, 246, 260, 261, 262, 276, 277, 278, 291, 292, 293, 307, 308, 309, 310, 323, 324, 325, 326, 339, 340, 341, 342, 356, 357, 358], "bbox": [0.09968109339407745, 0.1165, 0.5364920273348519, 0.9875468749999999], "iscrowd": 0, "area": 53119.635600000016, "rle": {"size": [640, 439], "counts": "dfk02lc03M3M4M2M3M3M4M2M3M4M2M3M3M4M2\\_OeN_>^1]AfN`>^1ZAgNc>\\1YAiNd>Y1WAlNf>W1VAnNf>V1TAoNi>T1SAPOk>R1PATOl>o0PAUOm>o0m@VOP?m0l@XOQ?j0k@ZOR?m2L3M3N3L3M3M4M2M3M4M2M4L3N2RIQJIQ65TJFm56ZJFg56`JFb55fJF[55mJFV55QKGP55UKIl43WKMk4NXK2i4JZK6i4DZK;k4^OXKb0l4WOWKi0m4POVKP1n4iNUKW1o4bNTK^1o4\\NTKc1Q5UNSKk1Q5nMRKR2R5gMQKY2S5`MPK`2T5YMoJg2U5RMnJm2W5lLlJT3[5bLiJ]3^5YLeJg3c5nK`JR4g5dK\\J\\4l5YKWJf4Q6PKRJP5V6dJnI\\5Y6ZJjIf5X:01N6K4L4L5K4L4K6K4L4L1O000O10000000000O10000000000O10000000000O10000000000O100000mKmKYHS4e7TLVHk3h7]LSHc3k7eLoG[3n7mLmGS3Q8TMjGh2W8aMcG\\2^8kM]GQ2b8YNnEnMOf3P:hNjEiM0\\3R:WOhEdM0Q3V:FdE`M0g2X:4cE]MN\\2[:?eE\\MJQ2_:i0gE\\MDj1_:Q1lE?o9HoE8k9OTF2f95XFKb9<]FD]9d0aF\\OZ9j0eFVOU9R1iFoNQ9W1nFiNl8X6K5J7I6K500010O1O0010O01OXJZDW3f;hL^DV3b;iL`DV3_;jLcDT3];lLeDS3Z;mLiDP3X;oLjDP3V;oLlDP3U;mLoDP3S;nLoDQ3Q;nLQEP3Q;nLREP3o:nLSEQ3m:nLUEP3m:nLVEP3k:mLXEj2P;UMRE_2Z;_MhDU2d;iM_Dk1l;SNVDa1U<^NmCV1_ in this image.", "objects": [{"patches": [39, 54, 55, 56, 68, 69, 70, 71, 72, 82, 83, 84, 85, 86, 87, 98, 99, 100, 101, 102, 103, 113, 114, 115, 116, 117, 118, 119, 120, 130, 131, 132, 133, 134, 135, 136, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 178, 179, 180, 181, 182, 183, 195, 196, 197, 198, 199, 211, 212, 213, 214, 228, 229, 230, 244, 245, 246, 260, 261, 262, 276, 277, 278, 291, 292, 293, 307, 308, 309, 310, 323, 324, 325, 326, 339, 340, 341, 342, 356, 357, 358], "bbox": [0.09968109339407745, 0.1165, 0.5364920273348519, 0.9875468749999999], "iscrowd": 0, "area": 53119.635600000016, "rle": {"size": [640, 439], "counts": "dfk02lc03M3M4M2M3M3M4M2M3M4M2M3M3M4M2\\_OeN_>^1]AfN`>^1ZAgNc>\\1YAiNd>Y1WAlNf>W1VAnNf>V1TAoNi>T1SAPOk>R1PATOl>o0PAUOm>o0m@VOP?m0l@XOQ?j0k@ZOR?m2L3M3N3L3M3M4M2M3M4M2M4L3N2RIQJIQ65TJFm56ZJFg56`JFb55fJF[55mJFV55QKGP55UKIl43WKMk4NXK2i4JZK6i4DZK;k4^OXKb0l4WOWKi0m4POVKP1n4iNUKW1o4bNTK^1o4\\NTKc1Q5UNSKk1Q5nMRKR2R5gMQKY2S5`MPK`2T5YMoJg2U5RMnJm2W5lLlJT3[5bLiJ]3^5YLeJg3c5nK`JR4g5dK\\J\\4l5YKWJf4Q6PKRJP5V6dJnI\\5Y6ZJjIf5X:01N6K4L4L5K4L4K6K4L4L1O000O10000000000O10000000000O10000000000O10000000000O100000mKmKYHS4e7TLVHk3h7]LSHc3k7eLoG[3n7mLmGS3Q8TMjGh2W8aMcG\\2^8kM]GQ2b8YNnEnMOf3P:hNjEiM0\\3R:WOhEdM0Q3V:FdE`M0g2X:4cE]MN\\2[:?eE\\MJQ2_:i0gE\\MDj1_:Q1lE?o9HoE8k9OTF2f95XFKb9<]FD]9d0aF\\OZ9j0eFVOU9R1iFoNQ9W1nFiNl8X6K5J7I6K500010O1O0010O01OXJZDW3f;hL^DV3b;iL`DV3_;jLcDT3];lLeDS3Z;mLiDP3X;oLjDP3V;oLlDP3U;mLoDP3S;nLoDQ3Q;nLQEP3Q;nLREP3o:nLSEQ3m:nLUEP3m:nLVEP3k:mLXEj2P;UMRE_2Z;_MhDU2d;iM_Dk1l;SNVDa1U<^NmCV1_ in this image.", "objects": [{"patches": [265, 266, 278, 279, 280, 281, 282, 283, 284, 293, 294, 295, 296, 297, 298, 299, 308, 309, 310, 311, 312, 313, 314, 323, 325, 326, 327, 328, 329, 338, 341, 342, 344], "bbox": [0.5311058823529412, 0.759140625, 1.0, 1.0], "iscrowd": 0, "area": 17884.4883, "rle": {"size": [640, 425], "counts": "nj]4:_c0>B?B=B5K001O1O1O1O0c0^O:FXOh05J7J5M3100O100O2N100O100O1O100O2O1N1O2O1N101N2N101N2O0O2N101N2O0O2N2O0O2O1N1O2O1N101N2N101N2O0O2N2O0O2O1N101N2O0O2O1N101N2O001N2O0010O01O10O01O10O01O10O01O10O01O10O01O10O01O10O01O10O01O?B9F0001O00001O00001O00001O00001O00001N1N2N3M2M4M2N3M2N3M2N3M2N3N2M2O2N1O2N1O2N1O2N1O2N1O2M2O2N1O2N2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N2N1O2N1O00000O1000`_OgNZ>Y1`AnN_>R1\\ASOd>m0VAZOi>f0RA@m>`0m@FS?:h@LW?4c@2]?N^@8a?HY@?f?AT@e0l?ZOo_Om0P`0a101O1O001O001O1Y@"}, "label": "chair with visible back"}]} {"id": 204294, "image": "COCO_train2014_000000204294.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blue striped tie worn by man with microphone and notepad\"."}], "task": "refering", "answer_template": "The \"blue striped tie worn by man with microphone and notepad\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 146, 160, 175, 189, 190, 204, 219, 248, 249, 263, 264], "bbox": [0.5633021077283372, 0.384625, 0.7511007025761125, 0.77371875], "iscrowd": 0, "area": 1402.0934500000017, "rle": {"size": [640, 427], "counts": "`Rg4>_c0;I6I2N01N1O2N1O2N1O2O0O1O1O001O^N@f_O?Y`0Cg_O;X`0Hg_O8W`0Jj_O5T`0Ko_O3Q`0LR@1n?NU@0k?0X@Mh?2\\@Kd?5S2O0O100O2O000OO2O0O012N4M3M3M4MO0O1O1M3N2N2M3N3N3Oa\\O0Pc01g\\O5[c06N1M3L5L5O0100O10O0100O10O0100O101O1O1N101O1N2O8I5K4LmoQ2"}, "label": "blue striped tie worn by man with microphone and notepad"}]} {"id": 335376, "image": "COCO_train2014_000000335376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the two slices of pizza to the right\"."}], "task": "refering", "answer_template": "The \"the two slices of pizza to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 45, 46, 47, 61, 62, 63, 64, 65, 66, 67, 80, 81, 82, 83, 84, 85, 86, 98, 99, 100, 101, 102, 103, 104], "bbox": [0.42498, 0.18386666666666668, 0.8416600000000001, 0.46656000000000003], "iscrowd": 0, "area": 14162.043099999999, "rle": {"size": [375, 500], "counts": "mg]21b;4N3N101nN1ZF0b95\\FKc98[FId9:YFHe9WFCh9c0RF^Om9W1O1O2N1O1O1O00000000001O001O1O4L5L1N2N2TG`M]8b2`G`M`8b2\\G`Mc8b2ZG`Me8b2ZG_Md8k2O101N000001O0000001O0000010O00000000001O000000000000000000010O0000000000000000000O2O000000001O0O1000001O00000O101O0000000O2O000000001O0O101O001O001N101O001N101O001N10000O10001O0O10000O1000000O2O000O2O00001N10001N101O00001O001N10001O001O0O2O001O1O001O1N101O1O001O0O2O1O001O1O0O2O1O001O1O001N101O1N101N101O0O2O0O2O1O0O2O1O1N2O1N2O1O1N2O1O1N3N1N2Kbil0"}, "label": "the two slices of pizza to the right"}]} {"id": 335376, "image": "COCO_train2014_000000335376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"topmost two slices of pizza\"."}], "task": "refering", "answer_template": "The \"topmost two slices of pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [43, 44, 45, 46, 47, 61, 62, 63, 64, 65, 66, 67, 80, 81, 82, 83, 84, 85, 86, 98, 99, 100, 101, 102, 103, 104], "bbox": [0.42498, 0.18386666666666668, 0.8416600000000001, 0.46656000000000003], "iscrowd": 0, "area": 14162.043099999999, "rle": {"size": [375, 500], "counts": "mg]21b;4N3N101nN1ZF0b95\\FKc98[FId9:YFHe9WFCh9c0RF^Om9W1O1O2N1O1O1O00000000001O001O1O4L5L1N2N2TG`M]8b2`G`M`8b2\\G`Mc8b2ZG`Me8b2ZG_Md8k2O101N000001O0000001O0000010O00000000001O000000000000000000010O0000000000000000000O2O000000001O0O1000001O00000O101O0000000O2O000000001O0O101O001O001N101O001N101O001N10000O10001O0O10000O1000000O2O000O2O00001N10001N101O00001O001N10001O001O0O2O001O1O001O1N101O1O001O0O2O1O001O1O0O2O1O001O1O001N101O1N101N101O0O2O0O2O1O0O2O1O1N2O1N2O1O1N2O1O1N3N1N2Kbil0"}, "label": "topmost two slices of pizza"}]} {"id": 335376, "image": "COCO_train2014_000000335376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the metal pizza knife with the wooden handle\"."}], "task": "refering", "answer_template": "The \"the metal pizza knife with the wooden handle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 119, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 140, 141, 142, 143, 157, 158, 159, 160], "bbox": [0.5013, 0.46527999999999997, 1.0, 0.6699466666666667], "iscrowd": 0, "area": 10903.99365, "rle": {"size": [375, 500], "counts": "aSl26[;6I7J7N11O01O00001O00001O00001O01O01O0000001O00001O000010O0001O0000001O0000001O00010O0000001O0000001O000010O000001O0000001O00000010O0001O0000001O0000001O00010O0000001O0000001O000010O000001O0000001O00001O01O0001O0000001O0000001O01O01O0000001O0000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O00000O2O0000001O0000001O0000001O000000O1O1O1O1O2N1O1O1O2N1O1N3N1O1O1O2N1O1O2N1O1O001O001O001O001O0010O010O0100O10O0100O010000O10000O1000000O10000O10000O10000O1000N2WOZK"}, "label": "the metal pizza knife with the wooden handle"}]} {"id": 335376, "image": "COCO_train2014_000000335376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the pizza serving utensil\"."}], "task": "refering", "answer_template": "The \"the pizza serving utensil\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [117, 118, 119, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 140, 141, 142, 143, 157, 158, 159, 160], "bbox": [0.5013, 0.46527999999999997, 1.0, 0.6699466666666667], "iscrowd": 0, "area": 10903.99365, "rle": {"size": [375, 500], "counts": "aSl26[;6I7J7N11O01O00001O00001O00001O01O01O0000001O00001O000010O0001O0000001O0000001O00010O0000001O0000001O000010O000001O0000001O00000010O0001O0000001O0000001O00010O0000001O0000001O000010O000001O0000001O00001O01O0001O0000001O0000001O01O01O0000001O0000001O0000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O0000001O0000001O00000O2O0000001O0000001O0000001O000000O1O1O1O1O2N1O1O1O2N1O1N3N1O1O1O2N1O1O2N1O1O001O001O001O001O0010O010O0100O10O0100O010000O10000O1000000O10000O10000O10000O1000N2WOZK"}, "label": "the pizza serving utensil"}]} {"id": 335376, "image": "COCO_train2014_000000335376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the two slices of pizza that are more towards the left side of the photo , they are also closer to the camera than the other two slices\"."}], "task": "refering", "answer_template": "The \"the two slices of pizza that are more towards the left side of the photo , they are also closer to the camera than the other two slices\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 96, 97, 98, 109, 110, 111, 112, 113, 114, 115, 116, 128, 129, 130, 131, 132, 133, 134, 135, 146, 147, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 204, 205, 206], "bbox": [0.10786, 0.39549333333333336, 0.53764, 0.9123466666666666], "iscrowd": 0, "area": 27174.6357, "rle": {"size": [375, 500], "counts": "Znc0b0U;4K5K3M3M3M3N3M2N2N3L4M3M3M1O1O1O1O1N2O002N1O2N2N1O2N2N2N2N1O2N1O100O1O1O2M2O1N2N2O1N3N0O2N2O0O2N2O1N1O2O1N2N2O1N2N101N2N2O0O2N2N101N101O0O2O001O001N24K10O1O1O1O1O1O1O1O1O1O1O1O100O1O2N1O1O100O010O010O0100O010O01O001O001O00001O010O001O001O00001O001O001N101O00001O00001O001N10001O00001N10001O001N10001O0O101O00001O00001O00000000001O0001O000001O000001O01O0000[NoIdMQ6]2PJaMQ6^2RJ_Mn5a2UJ]Mj5c2YJZMg5f2\\JWMe5h2_JTMa5l2hJkLX5U3nJfLR5Y3SKbLn4^3UK^Lm4`3WK\\Lj4c3ZKXLi4f3f1lLUHX2m7^M_H^2`8M2O3L5K3N2M3M;E2O1N3M2O2M2N2XOkEJW:JVF4i:M4L3N2MkYd2"}, "label": "the two slices of pizza that are more towards the left side of the photo , they are also closer to the camera than the other two slices"}]} {"id": 335376, "image": "COCO_train2014_000000335376.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two closest pizza slices\"."}], "task": "refering", "answer_template": "The \"two closest pizza slices\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 95, 96, 97, 98, 109, 110, 111, 112, 113, 114, 115, 116, 128, 129, 130, 131, 132, 133, 134, 135, 146, 147, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 171, 184, 185, 186, 187, 188, 189, 204, 205, 206], "bbox": [0.10786, 0.39549333333333336, 0.53764, 0.9123466666666666], "iscrowd": 0, "area": 27174.6357, "rle": {"size": [375, 500], "counts": "Znc0b0U;4K5K3M3M3M3N3M2N2N3L4M3M3M1O1O1O1O1N2O002N1O2N2N1O2N2N2N2N1O2N1O100O1O1O2M2O1N2N2O1N3N0O2N2O0O2N2O1N1O2O1N2N2O1N2N101N2N2O0O2N2N101N101O0O2O001O001N24K10O1O1O1O1O1O1O1O1O1O1O1O100O1O2N1O1O100O010O010O0100O010O01O001O001O00001O010O001O001O00001O001O001N101O00001O00001O001N10001O00001N10001O001N10001O0O101O00001O00001O00000000001O0001O000001O000001O01O0000[NoIdMQ6]2PJaMQ6^2RJ_Mn5a2UJ]Mj5c2YJZMg5f2\\JWMe5h2_JTMa5l2hJkLX5U3nJfLR5Y3SKbLn4^3UK^Lm4`3WK\\Lj4c3ZKXLi4f3f1lLUHX2m7^M_H^2`8M2O3L5K3N2M3M;E2O1N3M2O2M2N2XOkEJW:JVF4i:M4L3N2MkYd2"}, "label": "two closest pizza slices"}]} {"id": 187939, "image": "COCO_train2014_000000187939.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver laptop\"."}], "task": "refering", "answer_template": "The \"a silver laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [30, 31, 51, 52, 53, 54, 55, 72, 73, 74, 75, 76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.0, 0.060812500000000005, 0.77534375, 0.9842291666666667], "iscrowd": 0, "area": 129981.7601, "rle": {"size": [480, 640], "counts": "h8P1Q in this image.", "objects": [{"patches": [30, 31, 51, 52, 53, 54, 55, 72, 73, 74, 75, 76, 77, 78, 94, 95, 96, 97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383], "bbox": [0.0, 0.060812500000000005, 0.77534375, 0.9842291666666667], "iscrowd": 0, "area": 129981.7601, "rle": {"size": [480, 640], "counts": "h8P1Q in this image.", "objects": [{"patches": [64, 65, 85, 86, 87, 88, 89, 90, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 151, 152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 221, 222, 223, 224, 245, 246], "bbox": [0.537640625, 0.13258333333333333, 0.9741562500000001, 0.60675], "iscrowd": 0, "area": 28651.34730000001, "rle": {"size": [480, 640], "counts": "a^Q51n>1O2N2N2O1N1O2N2N2O1N2N1O2N2O1N2N2N1O2O1N2N2N2N101N2N2N2N1O2O1N2N2N2N101N2N2N2N2O0O2N2N2N2O0O2N2N2N2O1N1O2N2N2N2O1N1O2N2N2O1N2N1O2N2O1N2N1O2N2O1N2N2N1O2N2O1N2N2N1O2O1N2N2N1O2O1N2N2N2N101N2N2N2N2O0O2N000000010O00000001O000000010O000000001O0001O0001O000000000010O00000001O000000010O000000001O000001O01O00000000001O01O000001O000000010O000000001O000001O01O0000001O1O2N2O0O2N2N2N1O2N2N1O2O1N1O2N2N1O2N2N1O2O1N2N1O2N2N1O2N2N101N2N1O2N2N1O2N2O0O2N2N2N1O2N2N1O2O1N1O2N2N1O2N2N1O2O1N1O2N2N2N1O2N2N101N2N1O2N2N1O2N2N101N2N2N1O2N2N1O2N2O0O2N2N1O2N2N1O2O1N1O2N2N2N1O2Nnj7"}, "label": "a black laptop set next to a white one"}]} {"id": 187939, "image": "COCO_train2014_000000187939.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black laptop\"."}], "task": "refering", "answer_template": "The \"a black laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 85, 86, 87, 88, 89, 90, 107, 108, 109, 110, 111, 112, 113, 114, 129, 130, 131, 132, 133, 134, 135, 136, 151, 152, 153, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 180, 197, 198, 199, 200, 201, 202, 221, 222, 223, 224, 245, 246], "bbox": [0.537640625, 0.13258333333333333, 0.9741562500000001, 0.60675], "iscrowd": 0, "area": 28651.34730000001, "rle": {"size": [480, 640], "counts": "a^Q51n>1O2N2N2O1N1O2N2N2O1N2N1O2N2O1N2N2N1O2O1N2N2N2N101N2N2N2N1O2O1N2N2N2N101N2N2N2N2O0O2N2N2N2O0O2N2N2N2O1N1O2N2N2N2O1N1O2N2N2O1N2N1O2N2O1N2N1O2N2O1N2N2N1O2N2O1N2N2N1O2O1N2N2N1O2O1N2N2N2N101N2N2N2N2O0O2N000000010O00000001O000000010O000000001O0001O0001O000000000010O00000001O000000010O000000001O000001O01O00000000001O01O000001O000000010O000000001O000001O01O0000001O1O2N2O0O2N2N2N1O2N2N1O2O1N1O2N2N1O2N2N1O2O1N2N1O2N2N1O2N2N101N2N1O2N2N1O2N2O0O2N2N2N1O2N2N1O2O1N1O2N2N1O2N2N1O2O1N1O2N2N2N1O2N2N101N2N1O2N2N1O2N2N101N2N2N1O2N2N1O2N2O0O2N2N1O2N2N1O2O1N1O2N2N2N1O2Nnj7"}, "label": "a black laptop"}]} {"id": 187939, "image": "COCO_train2014_000000187939.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black laptop on a desk sitting beside a silver laptop\"."}], "task": "refering", "answer_template": "The \"a black laptop on a desk sitting beside a silver laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 292, 293, 294], "bbox": [0.3471875, 0.0022500000000000003, 1.0, 0.7730416666666666], "iscrowd": 0, "area": 94146.3691, "rle": {"size": [480, 640], "counts": "RRX31o>2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1N2O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1OO1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O5K:F:F:F:F:F1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1N3N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2M2O2N2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2NS2mMTJ"}, "label": "a black laptop on a desk sitting beside a silver laptop"}]} {"id": 187939, "image": "COCO_train2014_000000187939.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"this is the black laptop next to a white laptop\"."}], "task": "refering", "answer_template": "The \"this is the black laptop next to a white laptop\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 244, 245, 246, 247, 248, 249, 250, 268, 269, 270, 271, 272, 292, 293, 294], "bbox": [0.3471875, 0.0022500000000000003, 1.0, 0.7730416666666666], "iscrowd": 0, "area": 94146.3691, "rle": {"size": [480, 640], "counts": "RRX31o>2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N3M2N2N2N2N2N2N2N1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1N2O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O001O1OO1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O5K:F:F:F:F:F1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1N3N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2M2O2N2N1O2N2N1O2N2N1O2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2N2N1O2NS2mMTJ"}, "label": "this is the black laptop next to a white laptop"}]} {"id": 327209, "image": "COCO_train2014_000000327209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blonde young man\"."}], "task": "refering", "answer_template": "The \"a blonde young man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 28, 29, 30, 51, 52, 53, 54, 74, 75, 76, 96, 97, 98, 99, 118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 303, 304, 305, 306, 326, 327, 328, 329], "bbox": [0.139390625, 0.06233490566037736, 0.412984375, 0.9870047169811321], "iscrowd": 0, "area": 40241.63050000001, "rle": {"size": [424, 640], "counts": "lRU14R==DbFUOa9k0_1O0002N2N3M2N10O0000000000001O1O1O1N3L3N2N2N3M2M3N3M2N2N3L3N2N3Mo]k4"}, "label": "a blonde young man"}]} {"id": 327209, "image": "COCO_train2014_000000327209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"boy in a purple and black jacket and khaki pants\"."}], "task": "refering", "answer_template": "The \"boy in a purple and black jacket and khaki pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 28, 29, 30, 51, 52, 53, 54, 74, 75, 76, 96, 97, 98, 99, 118, 119, 120, 121, 122, 141, 142, 143, 144, 145, 164, 165, 166, 167, 168, 187, 188, 189, 190, 191, 192, 193, 210, 211, 212, 213, 214, 215, 216, 233, 234, 235, 236, 237, 238, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 283, 303, 304, 305, 306, 326, 327, 328, 329], "bbox": [0.139390625, 0.06233490566037736, 0.412984375, 0.9870047169811321], "iscrowd": 0, "area": 40241.63050000001, "rle": {"size": [424, 640], "counts": "lRU14R==DbFUOa9k0_1O0002N2N3M2N10O0000000000001O1O1O1N3L3N2N2N3M2M3N3M2N2N3L3N2N3Mo]k4"}, "label": "boy in a purple and black jacket and khaki pants"}]} {"id": 327209, "image": "COCO_train2014_000000327209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black table with kids around it\"."}], "task": "refering", "answer_template": "The \"a black table with kids around it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [286, 287, 288, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.24714062499999997, 0.8314622641509435, 0.93940625, 1.0], "iscrowd": 0, "area": 21601.40215, "rle": {"size": [424, 640], "counts": "nWR22U=1O1O1O1O1O1N2O1O1O1O1O1O1O1O100O100O1O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O100O100O10000O10000O10000O100O10000O10000O100O10000O10000O100O10000O01000O1000000O100000000O100000000O100000000O100000000O100000000O100000000O1000000O100000000O100000000O100000000O10000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O1000000000000000000000000000000000000000000000000000000001O000O100000000000000000000000000000000000000000001O000000000000000000000000000000000000000000001O0000001O000000001O0000001O000000001O0000001O000000001O00000bNhDl0Y;ROkDk0U;TOlDl0T;TOlDl0T;SOmDm0S;ROnDn0S;QOmDo0S;POnDP1R;oNoDQ1Q;oNoDQ1R;mNoDS1Q;mNoDS1Q;lNPET1P;kNPEV1P;jNPEV1a;00000000000000000000000O1000000O10000O1000000O10000O1000000O10000O10000000000001O001O001O00001O001O001O00001O001O001O00001O001O001O00001N101O001O00001O001O001O00001O001O001O00001O001O001O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1Ofg?"}, "label": "a black table with kids around it"}]} {"id": 327209, "image": "COCO_train2014_000000327209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table that has three people around it\"."}], "task": "refering", "answer_template": "The \"a table that has three people around it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [286, 287, 288, 289, 290, 291, 292, 293, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.24714062499999997, 0.8314622641509435, 0.93940625, 1.0], "iscrowd": 0, "area": 21601.40215, "rle": {"size": [424, 640], "counts": "nWR22U=1O1O1O1O1O1N2O1O1O1O1O1O1O1O100O100O1O100O1O100O1O100O100O1O100O1O100O1O100O100O1O100O100O100O10000O10000O10000O100O10000O10000O100O10000O10000O100O10000O01000O1000000O100000000O100000000O100000000O100000000O100000000O100000000O1000000O100000000O100000000O100000000O10000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O1000000000000O100000000000000O1000000000000O1000000000000O1000000000000000000000000000000000000000000000000000000001O000O100000000000000000000000000000000000000000001O000000000000000000000000000000000000000000001O0000001O000000001O0000001O000000001O0000001O000000001O00000bNhDl0Y;ROkDk0U;TOlDl0T;TOlDl0T;SOmDm0S;ROnDn0S;QOmDo0S;POnDP1R;oNoDQ1Q;oNoDQ1R;mNoDS1Q;mNoDS1Q;lNPET1P;kNPEV1P;jNPEV1a;00000000000000000000000O1000000O10000O1000000O10000O1000000O10000O10000000000001O001O001O00001O001O001O00001O001O001O00001O001O001O00001N101O001O00001O001O001O00001O001O001O00001O001O001O1O1O1O1O001O1O1O1O1O1O1O1O001O1O1O1O1O1O1Ofg?"}, "label": "a table that has three people around it"}]} {"id": 327209, "image": "COCO_train2014_000000327209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person in the middle\"."}], "task": "refering", "answer_template": "The \"the person in the middle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [54, 55, 56, 77, 78, 79, 100, 101, 102, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 261, 262, 263, 264, 265, 284, 285, 286, 287, 288, 307, 308], "bbox": [0.32178125, 0.13766509433962262, 0.567859375, 0.8935141509433963], "iscrowd": 0, "area": 30603.780249999993, "rle": {"size": [424, 640], "counts": "c_e23R=8H8I7H8H8I6I4L4M2M4L3N2N3N1O1VHiMc4W2ZKmMd4U2XKoMg4Q2`IjM;;R6l1dIlM3 in this image.", "objects": [{"patches": [54, 55, 56, 77, 78, 79, 100, 101, 102, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 150, 168, 169, 170, 171, 172, 173, 191, 192, 193, 194, 195, 196, 215, 216, 217, 218, 219, 238, 239, 240, 241, 242, 261, 262, 263, 264, 265, 284, 285, 286, 287, 288, 307, 308], "bbox": [0.32178125, 0.13766509433962262, 0.567859375, 0.8935141509433963], "iscrowd": 0, "area": 30603.780249999993, "rle": {"size": [424, 640], "counts": "c_e23R=8H8I7H8H8I6I4L4M2M4L3N2N3N1O1VHiMc4W2ZKmMd4U2XKoMg4Q2`IjM;;R6l1dIlM3 in this image.", "objects": [{"patches": [178, 179, 201, 202, 203, 223, 224, 225, 226, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 296, 297, 316, 317, 318, 319, 320], "bbox": [0.63846875, 0.5207311320754717, 0.9199531249999999, 0.9326415094339623], "iscrowd": 0, "area": 18600.15165, "rle": {"size": [424, 640], "counts": "PfY53S=2N3M2N3M3M2N3M3M2N3O1N101O1N101N101O1N101O1N101N2O001L4J5O2M2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1O100O10000O10000O10000O10000O10000O1WFYMQ9g2mF[MQ9g2mF[MP9h2nFZMP9h2nFZMb7F^IR3mN[Me7EYIS3POZMg7EUIS3ROZMi7EQIS3TOZMk7ElHT3WOYMm7EhHl3X7VLdHl3\\7VL_Hm3a7UL[Hm3e7ULWHm3i7ULRHm3o7f000000000000000000000000000000000000000000O100O100O100O100O100O100O100O100O2O0O101N1O2O0O101N100O2O`0_O3N1N3N2M3O1O1N101O1O1O1O1O001O1N2O1O1O1O001O1O1O1N2O001O1O1O1O1N1O2N2M3N2N2N2M2O3M5J5L4L4L5J5L4]Oc0[OYid0"}, "label": "a young girl wearing brown holding pink phone"}]} {"id": 327209, "image": "COCO_train2014_000000327209.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in a black jacket sitting at a table\"."}], "task": "refering", "answer_template": "The \"a girl in a black jacket sitting at a table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [178, 179, 201, 202, 203, 223, 224, 225, 226, 245, 246, 247, 248, 249, 250, 267, 268, 269, 270, 271, 272, 273, 290, 291, 292, 293, 294, 295, 296, 297, 316, 317, 318, 319, 320], "bbox": [0.63846875, 0.5207311320754717, 0.9199531249999999, 0.9326415094339623], "iscrowd": 0, "area": 18600.15165, "rle": {"size": [424, 640], "counts": "PfY53S=2N3M2N3M3M2N3M3M2N3O1N101O1N101N101O1N101O1N101N2O001L4J5O2M2N2O1N2O1N2N2O1N2N2O1N2O1N2N2O1O100O10000O10000O10000O10000O10000O1WFYMQ9g2mF[MQ9g2mF[MP9h2nFZMP9h2nFZMb7F^IR3mN[Me7EYIS3POZMg7EUIS3ROZMi7EQIS3TOZMk7ElHT3WOYMm7EhHl3X7VLdHl3\\7VL_Hm3a7UL[Hm3e7ULWHm3i7ULRHm3o7f000000000000000000000000000000000000000000O100O100O100O100O100O100O100O100O2O0O101N1O2O0O101N100O2O`0_O3N1N3N2M3O1O1N101O1O1O1O1O001O1N2O1O1O1O001O1O1O1N2O001O1O1O1O1N1O2N2M3N2N2N2M2O3M5J5L4L4L5J5L4]Oc0[OYid0"}, "label": "a girl in a black jacket sitting at a table"}]} {"id": 204339, "image": "COCO_train2014_000000204339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown antique wooden chair in a living room\"."}], "task": "refering", "answer_template": "The \"a brown antique wooden chair in a living room\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 147, 148, 149, 150, 170, 171, 172, 173, 174, 193, 194, 195, 196, 217, 218, 219, 220, 240], "bbox": [0.42554687500000005, 0.40969135802469137, 0.5769843750000001, 0.8794135802469136], "iscrowd": 0, "area": 8082.957300000002, "rle": {"size": [324, 640], "counts": "`Vf21R:W1jNZ1fN9F9H8G7JO0O100WOi03M2N2N3N1N2_JlLP4U3fKTMZ4n2TKbMl4_2PKeMo4\\2nJfMS5Z2jJhMV5Y2hJiMW5X2fJjMZ5`300100O0M4K5UOaJlLb5S3i0N2M2O2N1O10O100O14K9H>B1O001O001O18G10O010O100AlHlMV7R2?O1O1O1O00001N10000000000O1O1O1O1O001N20O01O100OlG`Nf7_1VHlNd7T1VHYOb7g1N0001O1N2gN]HOd7OhHFY7DcHK<:R7FlHI:9l6ITJ2l5KXJ4n7N1O1Nm_e2"}, "label": "a brown antique wooden chair in a living room"}]} {"id": 204339, "image": "COCO_train2014_000000204339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"wooden carved chair with a brown cat laying in it\"."}], "task": "refering", "answer_template": "The \"wooden carved chair with a brown cat laying in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [124, 125, 126, 147, 148, 149, 150, 170, 171, 172, 173, 174, 193, 194, 195, 196, 217, 218, 219, 220, 240], "bbox": [0.42554687500000005, 0.40969135802469137, 0.5769843750000001, 0.8794135802469136], "iscrowd": 0, "area": 8082.957300000002, "rle": {"size": [324, 640], "counts": "`Vf21R:W1jNZ1fN9F9H8G7JO0O100WOi03M2N2N3N1N2_JlLP4U3fKTMZ4n2TKbMl4_2PKeMo4\\2nJfMS5Z2jJhMV5Y2hJiMW5X2fJjMZ5`300100O0M4K5UOaJlLb5S3i0N2M2O2N1O10O100O14K9H>B1O001O001O18G10O010O100AlHlMV7R2?O1O1O1O00001N10000000000O1O1O1O1O001N20O01O100OlG`Nf7_1VHlNd7T1VHYOb7g1N0001O1N2gN]HOd7OhHFY7DcHK<:R7FlHI:9l6ITJ2l5KXJ4n7N1O1Nm_e2"}, "label": "wooden carved chair with a brown cat laying in it"}]} {"id": 204339, "image": "COCO_train2014_000000204339.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black chair that is to the left of the tv\"."}], "task": "refering", "answer_template": "The \"black chair that is to the left of the tv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 201, 202, 203, 224], "bbox": [0.7295625, 0.4114506172839506, 0.88025, 0.7968827160493827], "iscrowd": 0, "area": 7429.148499999999, "rle": {"size": [324, 640], "counts": "`mc4f0X9a0AO1O010O100000O10000000O01000O10O11N101O002M3N2N2M4M4L2N2N2N2N2N3M3M3M00001O001O001O00001O001O00001O001O000103L4M1NO101O001O00001O000O2O0O2]Ob0D in this image.", "objects": [{"patches": [131, 132, 133, 154, 155, 156, 157, 158, 177, 178, 179, 180, 181, 201, 202, 203, 224], "bbox": [0.7295625, 0.4114506172839506, 0.88025, 0.7968827160493827], "iscrowd": 0, "area": 7429.148499999999, "rle": {"size": [324, 640], "counts": "`mc4f0X9a0AO1O010O100000O10000000O01000O10O11N101O002M3N2N2M4M4L2N2N2N2N2N3M3M3M00001O001O001O00001O001O00001O001O000103L4M1NO101O001O00001O000O2O0O2]Ob0D in this image.", "objects": [{"patches": [64, 65, 81, 82, 83, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 136, 137, 138], "bbox": [0.4883, 0.29057057057057056, 0.71976, 0.6762462462462462], "iscrowd": 0, "area": 8154.627899999998, "rle": {"size": [333, 500], "counts": "``_22V:9J4K3_OKhF7Q9g0N1O2O0O2O0O101N101O001O0010O01O01O01O0010O010O100O100O2N01N1O2O0O2N101N101N1O101N1M4D;F;G8O2O0O2O0O1O2O0O2O0O2O001O00001O0010O01O01O010O010O1O010O1O010O003N2Mf0ZO010O1O100O1O101N2N101N2N102M2N3M3N2M2N3N2M3M5L3L4L4M5J9G8I8GmY]1"}, "label": "older fat man"}]} {"id": 540211, "image": "COCO_train2014_000000540211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"old fat man\"."}], "task": "refering", "answer_template": "The \"old fat man\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [64, 65, 81, 82, 83, 98, 99, 100, 101, 102, 116, 117, 118, 119, 120, 136, 137, 138], "bbox": [0.4883, 0.29057057057057056, 0.71976, 0.6762462462462462], "iscrowd": 0, "area": 8154.627899999998, "rle": {"size": [333, 500], "counts": "``_22V:9J4K3_OKhF7Q9g0N1O2O0O2O0O101N101O001O0010O01O01O01O0010O010O100O100O2N01N1O2O0O2N101N101N1O101N1M4D;F;G8O2O0O2O0O1O2O0O2O0O2O001O00001O0010O01O01O010O010O1O010O1O010O003N2Mf0ZO010O1O100O1O101N2N101N2N102M2N3M3N2M2N3N2M3M5L3L4L4M5J9G8I8GmY]1"}, "label": "old fat man"}]} {"id": 540211, "image": "COCO_train2014_000000540211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the boy in the jersey\"."}], "task": "refering", "answer_template": "The \"the boy in the jersey\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 92, 93, 94, 109, 110, 111, 112, 127, 128, 129, 130, 131, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 181, 182, 183, 184, 185, 199, 200, 201, 202, 203], "bbox": [0.07034, 0.3533633633633634, 0.33524, 0.9893393393393393], "iscrowd": 0, "area": 18313.4684, "rle": {"size": [333, 500], "counts": "bc;4R:=B?B=E;G9H8G:G7H6K5J5L5L3M4K5L3M4L3L5L4L3M4K4M3M1N2O1O1N2N2O1N2N2CcJB_5;bJE`59aJFa58_JHc56]JJe54[JKh53YJLi52WJNk50UJ0m5NTJ1m5NSJ1P6MPJ3R6KnI5S6JnI5T6IlI7V6GjI9W6FjI9X6EiI:Y6DiI:X6EiI:Y6CiI;Z6BhI=Y6AiI>Y6_OiI`0Y6]OiIb0[6YOgIf0^6TOeIj0_6QOcIn0o7O1O1N2O1O2N1N3M2N3M2N3CQf[3"}, "label": "the boy in the jersey"}]} {"id": 540211, "image": "COCO_train2014_000000540211.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a white shirt holding a phone to his ear\"."}], "task": "refering", "answer_template": "The \"a man in a white shirt holding a phone to his ear\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [74, 75, 76, 92, 93, 94, 109, 110, 111, 112, 127, 128, 129, 130, 131, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 181, 182, 183, 184, 185, 199, 200, 201, 202, 203], "bbox": [0.07034, 0.3533633633633634, 0.33524, 0.9893393393393393], "iscrowd": 0, "area": 18313.4684, "rle": {"size": [333, 500], "counts": "bc;4R:=B?B=E;G9H8G:G7H6K5J5L5L3M4K5L3M4L3L5L4L3M4K4M3M1N2O1O1N2N2O1N2N2CcJB_5;bJE`59aJFa58_JHc56]JJe54[JKh53YJLi52WJNk50UJ0m5NTJ1m5NSJ1P6MPJ3R6KnI5S6JnI5T6IlI7V6GjI9W6FjI9X6EiI:Y6DiI:X6EiI:Y6CiI;Z6BhI=Y6AiI>Y6_OiI`0Y6]OiIb0[6YOgIf0^6TOeIj0_6QOcIn0o7O1O1N2O1O2N1N3M2N3M2N3CQf[3"}, "label": "a man in a white shirt holding a phone to his ear"}]} {"id": 433723, "image": "COCO_train2014_000000433723.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skateboard of the boy wearing red and white tennis shoes\"."}], "task": "refering", "answer_template": "The \"the skateboard of the boy wearing red and white tennis shoes\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [190, 213, 214, 236, 237, 238, 261, 262, 285, 286, 309, 310, 311, 332, 333, 334], "bbox": [0.282921875, 0.5684507042253522, 0.5688281249999999, 0.9621830985915493], "iscrowd": 0, "area": 4759.5045, "rle": {"size": [426, 640], "counts": "Ua[2h0c<000010O0001O0WD]Ok:b0SEBk:>SEEl: in this image.", "objects": [{"patches": [190, 213, 214, 236, 237, 238, 261, 262, 285, 286, 309, 310, 311, 332, 333, 334], "bbox": [0.282921875, 0.5684507042253522, 0.5688281249999999, 0.9621830985915493], "iscrowd": 0, "area": 4759.5045, "rle": {"size": [426, 640], "counts": "Ua[2h0c<000010O0001O0WD]Ok:b0SEBk:>SEEl: in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 77, 78, 79, 80, 81, 82, 83, 99, 100, 101, 102, 103, 104, 105, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 193, 194, 195, 196, 215, 216, 217, 218, 239, 240, 241, 262, 263, 264, 265], "bbox": [0.322359375, 0.13889671361502348, 0.625328125, 0.789131455399061], "iscrowd": 0, "area": 28137.523850000005, "rle": {"size": [426, 640], "counts": "Yke21V=7I6J7I7I6J7I6J7I7J3L4L4L3O2N101N101N2O0O2O0O2O0O2N2O0O2O0O2O0cGRM]6P3aIRM^6o2`ISM^6n2aITM^6m2`IUM^6l2`IVM`6k2^IWM`6j2_IXM_6j2_IXM`6i2^IYM`6h2_IZM`6g2^I[M`6f2_I\\M`6e2^I]M`6d2_I]Ma6d2]I^Ma6d2]I^Mb6b2]I`Ma6b2]I`Mb6`2]IbMa6`2]IbMb6^2]IdMa6^2]IdMa6^2]IcMc6]2\\IeMb6]2\\IeMb6\\2]IfMa6\\2]IfMa6[2^IlMZ6W2dIQNT6P2kIWNn5k1PJ\\Ni5f1UJ\\Nh5f1WJ[Nh5g1VJ[Nh5f1WJ[Nh5g1VJ[Ng5g1XJZNg5h1WJZNg5e4O2M2O2N1O2M2O2N1O2M2O2O0O2O0O2O0O2O001N101N101O001O001O001O001O0000000000O10000000000^N]K]Jc4`5bK]J_4\\1ZKZ1=XMY4]1\\KW1`0ZMT4^1^KT1c0\\Mo3^1`KR1g0^Mi3_1bKo0k0_Mc3`1eKm0m0aM^3a1gKj0P1bMZ3c1gKg0U1dMU3b1iKg0V1eMQ3c1kKe0X1fMm2d1mKc0Z1gMi2d1oKc0\\1gMe2e1QLa0^1hMb2e1RL?`1jM_2e1SL?b1jM[2f1UL=d1kMW2g1WL;f1lMS2i1XL8h1nMP2j1XL6l1nMm1k1XL4o1oMi1m1YL1Q2PNg1o1YLNP73RIGQ79QICR7 in this image.", "objects": [{"patches": [55, 56, 57, 58, 59, 77, 78, 79, 80, 81, 82, 83, 99, 100, 101, 102, 103, 104, 105, 122, 123, 124, 125, 126, 127, 145, 146, 147, 148, 149, 150, 169, 170, 171, 172, 173, 193, 194, 195, 196, 215, 216, 217, 218, 239, 240, 241, 262, 263, 264, 265], "bbox": [0.322359375, 0.13889671361502348, 0.625328125, 0.789131455399061], "iscrowd": 0, "area": 28137.523850000005, "rle": {"size": [426, 640], "counts": "Yke21V=7I6J7I7I6J7I6J7I7J3L4L4L3O2N101N101N2O0O2O0O2O0O2N2O0O2O0O2O0cGRM]6P3aIRM^6o2`ISM^6n2aITM^6m2`IUM^6l2`IVM`6k2^IWM`6j2_IXM_6j2_IXM`6i2^IYM`6h2_IZM`6g2^I[M`6f2_I\\M`6e2^I]M`6d2_I]Ma6d2]I^Ma6d2]I^Mb6b2]I`Ma6b2]I`Mb6`2]IbMa6`2]IbMb6^2]IdMa6^2]IdMa6^2]IcMc6]2\\IeMb6]2\\IeMb6\\2]IfMa6\\2]IfMa6[2^IlMZ6W2dIQNT6P2kIWNn5k1PJ\\Ni5f1UJ\\Nh5f1WJ[Nh5g1VJ[Nh5f1WJ[Nh5g1VJ[Ng5g1XJZNg5h1WJZNg5e4O2M2O2N1O2M2O2N1O2M2O2O0O2O0O2O0O2O001N101N101O001O001O001O001O0000000000O10000000000^N]K]Jc4`5bK]J_4\\1ZKZ1=XMY4]1\\KW1`0ZMT4^1^KT1c0\\Mo3^1`KR1g0^Mi3_1bKo0k0_Mc3`1eKm0m0aM^3a1gKj0P1bMZ3c1gKg0U1dMU3b1iKg0V1eMQ3c1kKe0X1fMm2d1mKc0Z1gMi2d1oKc0\\1gMe2e1QLa0^1hMb2e1RL?`1jM_2e1SL?b1jM[2f1UL=d1kMW2g1WL;f1lMS2i1XL8h1nMP2j1XL6l1nMm1k1XL4o1oMi1m1YL1Q2PNg1o1YLNP73RIGQ79QICR7 in this image.", "objects": [{"patches": [229, 251, 252, 274, 275, 296, 297, 298, 319, 320, 321, 341, 342, 343, 344, 362, 363, 364, 365, 366, 367, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.6824375, 0.5720624999999999, 1.0, 0.9842291666666666], "iscrowd": 0, "area": 14961.006550000002, "rle": {"size": [480, 640], "counts": "ch]61o>000O100O10000O100O101O0O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10001N100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O2O000O100O10000O100O10000O2O0O100N2N2N2N2N2O1N2N2N2N2N2N2O1N2N2N2N2N2N101N2N2N2N2N2O1N2N2N1O2N2N2O1N2N2N2N2N2N2O0O2N2N2N2N2N2O1N2N2N1O2N2N2O1N2N2N2N2N2N2O0O2N2N2N2N2N2O1N2N2N2N2N2O1N2N2N2N2N2N2O1N21O1O1O1O1O1O1O1O1O2N1O1O1A?@`0@`0@`0hNY1]NZH"}, "label": "this is a pretty colored flower tablecloth"}]} {"id": 196156, "image": "COCO_train2014_000000196156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"floral tablecloth with pink flowers and green leaves\"."}], "task": "refering", "answer_template": "The \"floral tablecloth with pink flowers and green leaves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 87, 88, 89, 90, 91, 92, 93, 111, 112, 113, 114, 115, 116, 136, 137, 159, 160, 182, 183, 206, 253, 276, 277, 299, 300, 322, 323, 324, 345, 346, 347, 348, 349, 368, 369, 370, 371, 372], "bbox": [0.0, 0.0022500000000000003, 0.994375, 0.9865208333333333], "iscrowd": 0, "area": 61012.17649999999, "rle": {"size": [480, 640], "counts": "S4V1^;\\2YIoLJe48\\LfNd3[1^MbMa2`2aN]L_1e3^O\\Kb0e4]O[Kb0g4]OYKc0i4[OWKd0k4[OUKe0l4ZOTKf0m4YOSKf0P5XOPKh0Q5WOoJh0S5XOlJh0U5WOkJh0X5VOhJj0Y5UOgJk0Z5TOfJk0\\5TOdJl0^5RObJm0`5RO`Jn0a5QO_Jo0b5PO^Jo0e5oN[JQ1f5nNZJQ1h5nNXJP1k5oNUJP1n5nNRJQ1P6nNPJP1S6oNmIP1U6oNkIP1X6nNhIQ1Z6nNfIP1]6oNcIP1_6oNaIP1b6nN^IP1e6oN[IP1g6oNYIP1i6oNWIP1l6nNTIP1o6oNQIP1Q7oNoHP1T7nNlHP1W7oNiHP1Y7oNgHP1[7oNeHP1^7nNbHP1a7oN_HP1c7oN]HP1e7oN[Ho0i7oNWHP1k7oNUHP1m7oNSHP1o7oNQHo0S8oNmGP1U8oNkGP1W8oNiGo0Z8POfGo0]8oNcGP1_8oNaGP1a8oN_Go0d8PO\\GP1f8nNZGQ1h8nNXGQ1j8nNVGR1k8mNUGR1n8lNRGT1o8kNQGT1Q9kNoFU1R9jNnFU1U9iNkFW1V9hNjFW1X9hNhFW1Z9hNfFX1\\9fNdFY1^9fNbFZ1_9eNaFZ1a9eN_F[1c9cN]F\\1e9cN[F\\1g9cNYF]1i9aNWF^1k9aNUF_1l9`NTF_1n9`NRF`1o9_NQF`1P:`NPF_1Q:aNoE_1R:`NnE_1S:aNmE_1T:`NlE_1U:aNkE_1V:`NjE_1W:aNiE^1X:bNhE^1Y:aNgE^1Z:bNfE^1[:aNeE^1\\:bNdE^1]:aNcE^1^:bNbE]1`:cN_E]1a:cN_E\\1b:dN^E\\1c:cN]E\\1d:dN\\E\\1e:cN[E\\1f:dNZE[1h:dNXE\\1h:dNXE[1i:dNXE\\1i:aNYE^1h:aNYE_1h:^NZEa1g:^NZEb1g:[N[Ed1f:ZN\\Ee1f:YN[Eg1e:WN]Eh1d:WN]Ei1d:TN^Ek1c:SN_Em1b:QN_En1b:PN`Eo1b:oM_EQ2a:mMaER2a:kMaEU2_:jMbEU2e;00O1O100O100O100O1O100O100O100O1O100O100O100O100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O1000000000000000000000000000000000000O10000000000000000000000000000000000000000O1000O1000000000000000000000000000000000O10000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O1O001O001O001O001O001O1O001O001O1O001O1O001O001O1O001O001O1O001O1O001O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O001O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O1O2N3M3M2N3M3M2N3M3M3M2N3M3M2N3M3M2N3M3M3M2N3M3M2N3M3M2N3M3M2N3M3M3M2N3M2N001O001O001O1O001O001O001O001O1O001OWOhl1"}, "label": "floral tablecloth with pink flowers and green leaves"}]} {"id": 196156, "image": "COCO_train2014_000000196156.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a table covered in a floral tablecloth\"."}], "task": "refering", "answer_template": "The \"a table covered in a floral tablecloth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 87, 88, 89, 90, 91, 92, 93, 111, 112, 113, 114, 115, 116, 136, 137, 159, 160, 182, 183, 206, 253, 276, 277, 299, 300, 322, 323, 324, 345, 346, 347, 348, 349, 368, 369, 370, 371, 372], "bbox": [0.0, 0.0022500000000000003, 0.994375, 0.9865208333333333], "iscrowd": 0, "area": 61012.17649999999, "rle": {"size": [480, 640], "counts": "S4V1^;\\2YIoLJe48\\LfNd3[1^MbMa2`2aN]L_1e3^O\\Kb0e4]O[Kb0g4]OYKc0i4[OWKd0k4[OUKe0l4ZOTKf0m4YOSKf0P5XOPKh0Q5WOoJh0S5XOlJh0U5WOkJh0X5VOhJj0Y5UOgJk0Z5TOfJk0\\5TOdJl0^5RObJm0`5RO`Jn0a5QO_Jo0b5PO^Jo0e5oN[JQ1f5nNZJQ1h5nNXJP1k5oNUJP1n5nNRJQ1P6nNPJP1S6oNmIP1U6oNkIP1X6nNhIQ1Z6nNfIP1]6oNcIP1_6oNaIP1b6nN^IP1e6oN[IP1g6oNYIP1i6oNWIP1l6nNTIP1o6oNQIP1Q7oNoHP1T7nNlHP1W7oNiHP1Y7oNgHP1[7oNeHP1^7nNbHP1a7oN_HP1c7oN]HP1e7oN[Ho0i7oNWHP1k7oNUHP1m7oNSHP1o7oNQHo0S8oNmGP1U8oNkGP1W8oNiGo0Z8POfGo0]8oNcGP1_8oNaGP1a8oN_Go0d8PO\\GP1f8nNZGQ1h8nNXGQ1j8nNVGR1k8mNUGR1n8lNRGT1o8kNQGT1Q9kNoFU1R9jNnFU1U9iNkFW1V9hNjFW1X9hNhFW1Z9hNfFX1\\9fNdFY1^9fNbFZ1_9eNaFZ1a9eN_F[1c9cN]F\\1e9cN[F\\1g9cNYF]1i9aNWF^1k9aNUF_1l9`NTF_1n9`NRF`1o9_NQF`1P:`NPF_1Q:aNoE_1R:`NnE_1S:aNmE_1T:`NlE_1U:aNkE_1V:`NjE_1W:aNiE^1X:bNhE^1Y:aNgE^1Z:bNfE^1[:aNeE^1\\:bNdE^1]:aNcE^1^:bNbE]1`:cN_E]1a:cN_E\\1b:dN^E\\1c:cN]E\\1d:dN\\E\\1e:cN[E\\1f:dNZE[1h:dNXE\\1h:dNXE[1i:dNXE\\1i:aNYE^1h:aNYE_1h:^NZEa1g:^NZEb1g:[N[Ed1f:ZN\\Ee1f:YN[Eg1e:WN]Eh1d:WN]Ei1d:TN^Ek1c:SN_Em1b:QN_En1b:PN`Eo1b:oM_EQ2a:mMaER2a:kMaEU2_:jMbEU2e;00O1O100O100O100O1O100O100O100O1O100O100O100O100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O1000000000000000000000000000000000000O10000000000000000000000000000000000000000O1000O1000000000000000000000000000000000O10000000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O1O001O001O001O001O001O001O001O001O1O001O001O001O001O001O1O001O001O1O001O1O001O001O1O001O001O1O001O1O001O001O1O001O001O1O001O1O001O001O1O001O1O001O001O1O001O001O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O001O1O1O1O1O001O1O1O1O1O2N3M3M2N3M3M2N3M3M3M2N3M3M2N3M3M2N3M3M3M2N3M3M2N3M3M2N3M3M2N3M3M3M2N3M2N001O001O001O1O001O001O001O001O1O001OWOhl1"}, "label": "a table covered in a floral tablecloth"}]} {"id": 171581, "image": "COCO_train2014_000000171581.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man riding a horse\"."}], "task": "refering", "answer_template": "The \"a man riding a horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 120, 121, 143, 144, 145, 165, 166, 168, 188, 189, 190, 211, 212, 234, 235, 236, 258], "bbox": [0.1676875, 0.31820093457943927, 0.32253125, 0.7625934579439252], "iscrowd": 0, "area": 5007.80375, "rle": {"size": [428, 640], "counts": "X[]12Z=1O2M2O1O1N2O1OO010O010OTDHf:7YENc:3\\E1a:N^E6`:J_E:^:EaE?\\:BcEb0Z:^OeEf0X:ZOfEj0X:VOgEl0X:TOgEo0W:QOhEQ1V:oNiEU1U:kNjEW1U:iNjEZ1T:fNjE]1T:dNkE_1R:bNmE`1R:`NmEc1P:^NoEd1P:\\NoEg1n9YNRFi1l9XNSFk1k9UNTFm1j9TNVFa0F;Q:UOXF`0J9l9XOZF`0K7j9ZOZF?N5g9^OZF=13c9CYF=50^9E[F=92l8LiF3==[8BWG2`0:_7dNcHY20SOZ7jNbHR26TOT7oNcHn1:ROQ7ROeHk1h6^O`HXO3n0:L>=f6A^HYO4m0;K>Ec06a6=^HFa0Db04c6b0[HCd0D`03d6g0WH]Ok0H;0f6l0jIROBMj6o0fIQOBLk6T1cIoNR8o0PHPOP8o0W2O1O2N2O2M5L7I1HTCNm<17O2O1N1Og=9RB100O2N1O2N10001O001O000OO001M3DUWe5"}, "label": "a man riding a horse"}]} {"id": 171581, "image": "COCO_train2014_000000171581.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy in a red shirt and jeans on top of a brown horse\"."}], "task": "refering", "answer_template": "The \"a guy in a red shirt and jeans on top of a brown horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 120, 121, 143, 144, 145, 165, 166, 168, 188, 189, 190, 211, 212, 234, 235, 236, 258], "bbox": [0.1676875, 0.31820093457943927, 0.32253125, 0.7625934579439252], "iscrowd": 0, "area": 5007.80375, "rle": {"size": [428, 640], "counts": "X[]12Z=1O2M2O1O1N2O1OO010O010OTDHf:7YENc:3\\E1a:N^E6`:J_E:^:EaE?\\:BcEb0Z:^OeEf0X:ZOfEj0X:VOgEl0X:TOgEo0W:QOhEQ1V:oNiEU1U:kNjEW1U:iNjEZ1T:fNjE]1T:dNkE_1R:bNmE`1R:`NmEc1P:^NoEd1P:\\NoEg1n9YNRFi1l9XNSFk1k9UNTFm1j9TNVFa0F;Q:UOXF`0J9l9XOZF`0K7j9ZOZF?N5g9^OZF=13c9CYF=50^9E[F=92l8LiF3==[8BWG2`0:_7dNcHY20SOZ7jNbHR26TOT7oNcHn1:ROQ7ROeHk1h6^O`HXO3n0:L>=f6A^HYO4m0;K>Ec06a6=^HFa0Db04c6b0[HCd0D`03d6g0WH]Ok0H;0f6l0jIROBMj6o0fIQOBLk6T1cIoNR8o0PHPOP8o0W2O1O2N2O2M5L7I1HTCNm<17O2O1N1Og=9RB100O2N1O2N10001O001O000OO001M3DUWe5"}, "label": "a guy in a red shirt and jeans on top of a brown horse"}]} {"id": 171581, "image": "COCO_train2014_000000171581.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a white dress riding a brown horse\"."}], "task": "refering", "answer_template": "The \"a woman with a white dress riding a brown horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 149, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 196, 197, 198, 200, 201, 202, 219, 221, 223], "bbox": [0.48754687499999994, 0.36343457943925234, 0.7953125, 0.6430140186915889], "iscrowd": 0, "area": 8145.944149999999, "rle": {"size": [428, 640], "counts": "kbR43W=6J6I4M2M2N2N3M2N3N1O101O11N10O1O0100O101O1N101O1N2O1N2O1O10000001N1O1mEVOb7k0\\HVOd7k0[HUOe7f0QF\\OX2Og7f0UFXOQ23j7g0[HYOe7i0UH[Ok7g0oG]OQ8d0jG@V8b0dGB\\8`0^GDb8`0VGDj8`0nFDR9?gFB]9a0[F@h9c0QF]OS:g0eEYO_:f15L3M3N2N3L3N2M4M2N3L3N2N3L30O01O02O3M2N3M2N3M2N3M2N3MSNQE_1m:]NZEc1d:XNcEg1[:UNkEl1l:mNaD6_;GfD6[;GiD8X;ZOcDM8h0V<2MM1O2N1O6IlMI[24M3L4L5K4M3L4L4L10O01O1O010O1O0KgNXDZ1h;51O0O10000O10001N11O1O1O1O100O1O1O1O1O1O100O1O1O1O1oD\\NX:e1fEhNo9X1oEiNQ:X1mEjNR:W1lEjNT:W1jEjNW:V1gElNX:U1fElNZ:U1dE[N>Gn9W2e0 in this image.", "objects": [{"patches": [126, 127, 149, 150, 151, 152, 153, 154, 155, 173, 174, 175, 176, 177, 178, 196, 197, 198, 200, 201, 202, 219, 221, 223], "bbox": [0.48754687499999994, 0.36343457943925234, 0.7953125, 0.6430140186915889], "iscrowd": 0, "area": 8145.944149999999, "rle": {"size": [428, 640], "counts": "kbR43W=6J6I4M2M2N2N3M2N3N1O101O11N10O1O0100O101O1N101O1N2O1N2O1O10000001N1O1mEVOb7k0\\HVOd7k0[HUOe7f0QF\\OX2Og7f0UFXOQ23j7g0[HYOe7i0UH[Ok7g0oG]OQ8d0jG@V8b0dGB\\8`0^GDb8`0VGDj8`0nFDR9?gFB]9a0[F@h9c0QF]OS:g0eEYO_:f15L3M3N2N3L3N2M4M2N3L3N2N3L30O01O02O3M2N3M2N3M2N3M2N3MSNQE_1m:]NZEc1d:XNcEg1[:UNkEl1l:mNaD6_;GfD6[;GiD8X;ZOcDM8h0V<2MM1O2N1O6IlMI[24M3L4L5K4M3L4L4L10O01O1O010O1O0KgNXDZ1h;51O0O10000O10001N11O1O1O1O100O1O1O1O1O1O100O1O1O1O1oD\\NX:e1fEhNo9X1oEiNQ:X1mEjNR:W1lEjNT:W1jEjNW:V1gElNX:U1fElNZ:U1dE[N>Gn9W2e0 in this image.", "objects": [{"patches": [131, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205], "bbox": [0.71275, 0.34859813084112146, 0.99353125, 0.602126168224299], "iscrowd": 0, "area": 7556.488899999997, "rle": {"size": [428, 640], "counts": "`hn51Y=2N3L3M3N2N2N3L3N2N2N2O2O000O10O2N1O1O2N1O110O01O010O10O01O2O0O101N1O101N2cEoN[8W1^GjNc8V1\\GjNd8X1YGiNg8X1WGiNj8W1TGjNl8X1SGgNm8Z1RGfNo8[1oFeNQ9\\1nFdNR9`1kF_NV9f1dFZN\\9T2g05L0O1O001OG9L53L8H1O1O1O1O0O2NB\\EWNb:f1cEZN\\:_1mE`NQ:[1VFeNh9Z1\\FeNc9[1^FeN`9\\1aFcN`9\\1aFdN^9\\1bFeN]9MTFf0`0]O[9H]Fh08@Y:=hECZ::fEG\\:5eEL]:NgE1\\:HhE9f;O2MkMLXG0f84W22M4M:E;bCoNn;X1O1O1ZEhNW9Y1iFgNV9Y1jFhNU9Y1jFhNU9Y1jFhNU9Y1jFhNT9Z1lFfNT9Z1jFhNU9X1jFiNW9W1hFjNY9U1eFmN\\9Q1dFnN^9Q1aFPO`9P1^FQOc9n0]FQOe9n0ZFROh9n0WFROj9m0UFSOm9m0QFTOP:l0mEUOU:o0dESO]:l13N1N2O1N2O2N1O1N2I8J5N2M4O01O2N101N1O101N1O2N101ZEWNe9k1XFVNi9n0dED`0@l9h0lEG3CR:b0oE5CZO^:=UFe0k9nN^EOl0R1g9nNdFP1\\9SOcFj0_9KTFLl94YFGh97]F@h9`0]FWOh9h0^110001O01O100O2O01O2N100O1NO2N100O2N1O1O2N1O2N1O3M3M2N3M3M2N[^1"}, "label": "a white horse on the beach being ridden by a young woman"}]} {"id": 171581, "image": "COCO_train2014_000000171581.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman riding a white horse\"."}], "task": "refering", "answer_template": "The \"a woman riding a white horse\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [131, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183, 202, 203, 204, 205], "bbox": [0.71275, 0.34859813084112146, 0.99353125, 0.602126168224299], "iscrowd": 0, "area": 7556.488899999997, "rle": {"size": [428, 640], "counts": "`hn51Y=2N3L3M3N2N2N3L3N2N2N2O2O000O10O2N1O1O2N1O110O01O010O10O01O2O0O101N1O101N2cEoN[8W1^GjNc8V1\\GjNd8X1YGiNg8X1WGiNj8W1TGjNl8X1SGgNm8Z1RGfNo8[1oFeNQ9\\1nFdNR9`1kF_NV9f1dFZN\\9T2g05L0O1O001OG9L53L8H1O1O1O1O0O2NB\\EWNb:f1cEZN\\:_1mE`NQ:[1VFeNh9Z1\\FeNc9[1^FeN`9\\1aFcN`9\\1aFdN^9\\1bFeN]9MTFf0`0]O[9H]Fh08@Y:=hECZ::fEG\\:5eEL]:NgE1\\:HhE9f;O2MkMLXG0f84W22M4M:E;bCoNn;X1O1O1ZEhNW9Y1iFgNV9Y1jFhNU9Y1jFhNU9Y1jFhNU9Y1jFhNT9Z1lFfNT9Z1jFhNU9X1jFiNW9W1hFjNY9U1eFmN\\9Q1dFnN^9Q1aFPO`9P1^FQOc9n0]FQOe9n0ZFROh9n0WFROj9m0UFSOm9m0QFTOP:l0mEUOU:o0dESO]:l13N1N2O1N2O2N1O1N2I8J5N2M4O01O2N101N1O101N1O2N101ZEWNe9k1XFVNi9n0dED`0@l9h0lEG3CR:b0oE5CZO^:=UFe0k9nN^EOl0R1g9nNdFP1\\9SOcFj0_9KTFLl94YFGh97]F@h9`0]FWOh9h0^110001O01O100O2O01O2N100O1NO2N100O2N1O1O2N1O2N1O3M3M2N3M3M2N[^1"}, "label": "a woman riding a white horse"}]} {"id": 548416, "image": "COCO_train2014_000000548416.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white bowl full of soup\"."}], "task": "refering", "answer_template": "The \"white bowl full of soup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 33, 34, 35, 53, 71, 89, 143, 161, 162, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.0019399999999999999, 0.03117333333333333, 1.0, 0.9844266666666668], "iscrowd": 0, "area": 20084.484600000014, "rle": {"size": [375, 500], "counts": "Rd0i0n:c1]Ni0XOO1O1O2N1O2N1O2N1O2N2N1O2N1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00000010O0001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O00000000000000000000001O00000000000000000000001O00000000000000000000001O00000000000000000000001O000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000O1000000O1000000O10000O1000000O10000O1000000O1000000O10000O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1eFjNT8W1kGmNQ8T1nGPOn7Q1QHTOj7m0UHWOg7j0XHZOd7g0[H]Oa7d0^H\\Oa7f0]H\\Ob7e0]H\\Oa7f0^H[Oa7f0^H[Oa7f0^HZOa7h0_HXO`7i0_HaNC4l7]1aH`NMHa7i1bH_NR8c1nG]NQ8d1oG\\No7f1QHZNn7f1SHZNk7h1UHXNj7i1VHWNh7k1XHUNg7l1YHUNe7l1[HTNc7n1\\HSNc7n1]HRNa7o1`HQN_7P2aHPN]7R2cHoM[7R2eHPNW7R2iHoMU7R2kHPNQ7R2oHoMo6R2QIPNk6Q2VIPNh6Q2XIQNd6Q2\\IPN?TOn4m2dJPN;[Oi4f2lJPN9Ca4_2WKoM5J[4Y2`KnM3LZ4W2dKnM0MX4V2iKnMMNW4V2mKmMI0W4T2PLmMG1V4T2SLmME1T4T2XLkMB3T4S2ZLkMA4Q4S2_LjM]O5R4R2aLjM[O7P4P2gLjMWO7o3Q2jLiMUO9n3o1mLiMSO:m3o1QM_Ol2b0UM_Oh2b0ZM]Ob2f0_M[O\\2h0fMXOU2k0lMVOo1m0RNTOi1o0YNPOc1T1]NhNf1X1\\NaNk1]1VNbLk3]3X2OWI"}, "label": "white bowl full of soup"}]} {"id": 548416, "image": "COCO_train2014_000000548416.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the white bowl\"."}], "task": "refering", "answer_template": "The \"the white bowl\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [15, 16, 17, 33, 34, 35, 53, 71, 89, 143, 161, 162, 178, 179, 180, 181, 196, 197, 198, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.0019399999999999999, 0.03117333333333333, 1.0, 0.9844266666666668], "iscrowd": 0, "area": 20084.484600000014, "rle": {"size": [375, 500], "counts": "Rd0i0n:c1]Ni0XOO1O1O2N1O2N1O2N1O2N2N1O2N1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O001O1O1O1O001O1O001O1O001O001O1O001O1O001O1O001O001O1O001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O0000001O00001O00000010O0001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O000000001O0000001O00000000000000000000001O00000000000000000000001O00000000000000000000001O00000000000000000000001O000000000000000000000000001O000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000000000000000000O1000000000000O1000000O1000000O10000O1000000O10000O1000000O1000000O10000O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O100O1O1eFjNT8W1kGmNQ8T1nGPOn7Q1QHTOj7m0UHWOg7j0XHZOd7g0[H]Oa7d0^H\\Oa7f0]H\\Ob7e0]H\\Oa7f0^H[Oa7f0^H[Oa7f0^HZOa7h0_HXO`7i0_HaNC4l7]1aH`NMHa7i1bH_NR8c1nG]NQ8d1oG\\No7f1QHZNn7f1SHZNk7h1UHXNj7i1VHWNh7k1XHUNg7l1YHUNe7l1[HTNc7n1\\HSNc7n1]HRNa7o1`HQN_7P2aHPN]7R2cHoM[7R2eHPNW7R2iHoMU7R2kHPNQ7R2oHoMo6R2QIPNk6Q2VIPNh6Q2XIQNd6Q2\\IPN?TOn4m2dJPN;[Oi4f2lJPN9Ca4_2WKoM5J[4Y2`KnM3LZ4W2dKnM0MX4V2iKnMMNW4V2mKmMI0W4T2PLmMG1V4T2SLmME1T4T2XLkMB3T4S2ZLkMA4Q4S2_LjM]O5R4R2aLjM[O7P4P2gLjMWO7o3Q2jLiMUO9n3o1mLiMSO:m3o1QM_Ol2b0UM_Oh2b0ZM]Ob2f0_M[O\\2h0fMXOU2k0lMVOo1m0RNTOi1o0YNPOc1T1]NhNf1X1\\NaNk1]1VNbLk3]3X2OWI"}, "label": "the white bowl"}]} {"id": 548416, "image": "COCO_train2014_000000548416.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"soup with chicken and carrots and yellow broth\"."}], "task": "refering", "answer_template": "The \"soup with chicken and carrots and yellow broth\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.0, 0.0, 1.0, 1.0], "iscrowd": 0, "area": 184284.46810000003, "rle": {"size": [375, 500], "counts": "0VoS31`\\lL000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000O100000000000000WOi0^Mb2"}, "label": "soup with chicken and carrots and yellow broth"}]} {"id": 548416, "image": "COCO_train2014_000000548416.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bowl of soup\"."}], "task": "refering", "answer_template": "The \"bowl of soup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233], "bbox": [0.0, 0.0, 1.0, 1.0], "iscrowd": 0, "area": 184284.46810000003, "rle": {"size": [375, 500], "counts": "0VoS31`\\lL000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000O100000000000000WOi0^Mb2"}, "label": "bowl of soup"}]} {"id": 261696, "image": "COCO_train2014_000000261696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white round vase with a long neck\"."}], "task": "refering", "answer_template": "The \"a white round vase with a long neck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 111, 112, 126, 127, 128, 141, 142, 143, 144, 145, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 187, 188, 189, 190, 202, 203, 204, 205], "bbox": [0.40191588785046733, 0.26975, 0.7195327102803739, 0.600984375], "iscrowd": 0, "area": 15638.2204, "rle": {"size": [640, 428], "counts": "ge[35fc08L4K4M4L4N101O1M2M4M3M2N3M3Y_OcNd>`1WAeNg>^1SAhNj>[1QAiNm>Y1o@lNn>W1n@lNQ?V1k@mNW?Q1f@SO\\?j0a@YOa?f0[@^Ob?d0[@_Ob?e0Z@_Oc?c0Z@@X?aNd@S2OAZ?`Nb@R21A[?i2N3N1N3N1O2M2O2N1N2O2K4M4L3M40O10O01O010O10O01N1O2M3M2O2M2N3N1N3M3N1N3M2O2L4M2N3M2N1O001O000O1000001O0000000O10001O0000000O2O0O2O001N101O0O2O0O2O001N101O0O2O0O2N1O3M2N2N2N2N2N3M2N2O1N2N2M4M2M3N2M5K5L4L4L4J6J7I6Jc0]OQfZ2"}, "label": "a white round vase with a long neck"}]} {"id": 261696, "image": "COCO_train2014_000000261696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a flower vase that is round at the bottom and a narrow cylinder at the top\"."}], "task": "refering", "answer_template": "The \"a flower vase that is round at the bottom and a narrow cylinder at the top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [96, 97, 111, 112, 126, 127, 128, 141, 142, 143, 144, 145, 156, 157, 158, 159, 160, 171, 172, 173, 174, 175, 187, 188, 189, 190, 202, 203, 204, 205], "bbox": [0.40191588785046733, 0.26975, 0.7195327102803739, 0.600984375], "iscrowd": 0, "area": 15638.2204, "rle": {"size": [640, 428], "counts": "ge[35fc08L4K4M4L4N101O1M2M4M3M2N3M3Y_OcNd>`1WAeNg>^1SAhNj>[1QAiNm>Y1o@lNn>W1n@lNQ?V1k@mNW?Q1f@SO\\?j0a@YOa?f0[@^Ob?d0[@_Ob?e0Z@_Oc?c0Z@@X?aNd@S2OAZ?`Nb@R21A[?i2N3N1N3N1O2M2O2N1N2O2K4M4L3M40O10O01O010O10O01N1O2M3M2O2M2N3N1N3M3N1N3M2O2L4M2N3M2N1O001O000O1000001O0000000O10001O0000000O2O0O2O001N101O0O2O0O2O001N101O0O2O0O2N1O3M2N2N2N2N2N3M2N2O1N2N2M4M2M3N2M5K5L4L4L4J6J7I6Jc0]OQfZ2"}, "label": "a flower vase that is round at the bottom and a narrow cylinder at the top"}]} {"id": 261696, "image": "COCO_train2014_000000261696.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tall white cylindrical vase\"."}], "task": "refering", "answer_template": "The \"a tall white cylindrical vase\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [92, 93, 94, 109, 110, 120, 121, 123, 124, 125, 135, 136, 137, 138, 139, 140, 141, 151, 152, 153, 154, 155, 156, 167, 168, 169, 170, 171, 182, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 228, 229, 230, 231, 232, 233, 244, 245, 246, 247, 248, 249, 259, 260, 261, 262, 263, 264, 275, 276, 277, 278, 279, 291, 292, 293], "bbox": [0.05712616822429906, 0.26742187500000003, 0.6350934579439252, 0.8516875], "iscrowd": 0, "area": 39087.62325000001, "rle": {"size": [640, 428], "counts": "\\W?1oc0a0@5K0O2O0O101N101N101O0O101N101N10001N101N101N10001jMfNbA[1]>eNdA\\1Z>eNeA\\1Z>dNgA^1U>cNkA_1S>aNnA`1P>`NQBa1m=_NSBc1j=_NVBb1h=^NXBe1e=[N\\Bf1b=ZN^Bh1_=YNbBh1\\=XNeBi1Y=WNgBl1V=UNjBl1S=UNmBm1Q=SNoBo1o in this image.", "objects": [{"patches": [92, 93, 94, 109, 110, 120, 121, 123, 124, 125, 135, 136, 137, 138, 139, 140, 141, 151, 152, 153, 154, 155, 156, 167, 168, 169, 170, 171, 182, 183, 184, 185, 186, 187, 198, 199, 200, 201, 202, 213, 214, 215, 216, 217, 218, 228, 229, 230, 231, 232, 233, 244, 245, 246, 247, 248, 249, 259, 260, 261, 262, 263, 264, 275, 276, 277, 278, 279, 291, 292, 293], "bbox": [0.05712616822429906, 0.26742187500000003, 0.6350934579439252, 0.8516875], "iscrowd": 0, "area": 39087.62325000001, "rle": {"size": [640, 428], "counts": "\\W?1oc0a0@5K0O2O0O101N101N101O0O101N101N10001N101N101N10001jMfNbA[1]>eNdA\\1Z>eNeA\\1Z>dNgA^1U>cNkA_1S>aNnA`1P>`NQBa1m=_NSBc1j=_NVBb1h=^NXBe1e=[N\\Bf1b=ZN^Bh1_=YNbBh1\\=XNeBi1Y=WNgBl1V=UNjBl1S=UNmBm1Q=SNoBo1o in this image.", "objects": [{"patches": [133, 149, 150, 151, 165, 166, 167, 168, 181, 182, 183, 184, 199, 200], "bbox": [0.05568, 0.5848533333333333, 0.42607999999999996, 0.9068266666666667], "iscrowd": 0, "area": 4182.163450000002, "rle": {"size": [375, 500], "counts": "Zb:;[;2N2N1O2N1O001O001N101O001O001O1O001O0010O01O010O010O0010O010O010O00010O001O001O01O01O001O001O01O01O001O00010O001O001O00010O001O001O10O01O010O010000O010O10000O100O101N10001N100O101N10000O100O100O1O2O0O01O001O010O001O01O01O001O010O001O010O00001O010O001O010O001O00010O001O010O001O00010O001O010O001O00100O001O001O0O2O001O001O0O2O001O001O0O2O001O001O0O2O001O001O001N10kcY3"}, "label": "knife on a white plate"}]} {"id": 7746, "image": "COCO_train2014_000000007746.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a knife in the plate\"."}], "task": "refering", "answer_template": "The \"a knife in the plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [133, 149, 150, 151, 165, 166, 167, 168, 181, 182, 183, 184, 199, 200], "bbox": [0.05568, 0.5848533333333333, 0.42607999999999996, 0.9068266666666667], "iscrowd": 0, "area": 4182.163450000002, "rle": {"size": [375, 500], "counts": "Zb:;[;2N2N1O2N1O001O001N101O001O001O1O001O0010O01O010O010O0010O010O010O00010O001O001O01O01O001O001O01O01O001O00010O001O001O00010O001O001O10O01O010O010000O010O10000O100O101N10001N100O101N10000O100O100O1O2O0O01O001O010O001O01O01O001O010O001O010O00001O010O001O010O001O00010O001O010O001O00010O001O010O001O00100O001O001O0O2O001O001O0O2O001O001O0O2O001O001O0O2O001O001O001N10kcY3"}, "label": "a knife in the plate"}]} {"id": 48707, "image": "COCO_train2014_000000048707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the flower on the table\"."}], "task": "refering", "answer_template": "The \"the flower on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 24, 25, 26, 27, 28, 29, 30, 47, 48, 49, 50, 51, 52, 53, 70, 71, 72, 73, 74, 75, 76, 93, 94, 95, 96], "bbox": [0.060250000000000005, 0.0022377622377622378, 0.33440625, 0.31909090909090915], "iscrowd": 0, "area": 19507.7431, "rle": {"size": [429, 640], "counts": "lZ`07V=9G8H8H8H9G8H8H9G8H8H8H9G8H8H9G4LO1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O101O000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000000000O100000000000000O10000000000000000O1000000O10000O10000O10000_Na`b5"}, "label": "the flower on the table"}]} {"id": 48707, "image": "COCO_train2014_000000048707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green blurry object that resembles a tree behind the main course\"."}], "task": "refering", "answer_template": "The \"a green blurry object that resembles a tree behind the main course\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 24, 25, 26, 27, 28, 29, 30, 47, 48, 49, 50, 51, 52, 53, 70, 71, 72, 73, 74, 75, 76, 93, 94, 95, 96], "bbox": [0.060250000000000005, 0.0022377622377622378, 0.33440625, 0.31909090909090915], "iscrowd": 0, "area": 19507.7431, "rle": {"size": [429, 640], "counts": "lZ`07V=9G8H8H8H9G8H8H9G8H8H8H9G8H8H9G4LO1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O101O000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000O1000000000000O100000000000000O10000000000000000O1000000O10000O10000O10000_Na`b5"}, "label": "a green blurry object that resembles a tree behind the main course"}]} {"id": 48707, "image": "COCO_train2014_000000048707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a tea cup with blue pattern on saucer next to a bowl full of edibles\"."}], "task": "refering", "answer_template": "The \"a tea cup with blue pattern on saucer next to a bowl full of edibles\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 21, 36, 37, 38, 39, 40, 41, 42, 43, 44, 59, 60, 61, 62, 63, 64, 65, 66, 67, 82, 83, 84, 85, 86, 87, 88], "bbox": [0.5641093749999999, 0.007785547785547785, 0.9454062499999999, 0.2649417249417249], "iscrowd": 0, "area": 21271.824900000003, "rle": {"size": [429, 640], "counts": "^Yg4V1a:f1000000000000000001O01O00000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001O000001O000001O00000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001O000000000000J6[Oe000O100000000O1000000O100000000O100000000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O2O0O1N2M3N2N2N2M3N2N2M5L4L4L4K5L4L4L4Kdd>"}, "label": "a tea cup with blue pattern on saucer next to a bowl full of edibles"}]} {"id": 48707, "image": "COCO_train2014_000000048707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blurred white coffee cup\"."}], "task": "refering", "answer_template": "The \"blurred white coffee cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [13, 14, 15, 16, 17, 18, 19, 20, 21, 36, 37, 38, 39, 40, 41, 42, 43, 44, 59, 60, 61, 62, 63, 64, 65, 66, 67, 82, 83, 84, 85, 86, 87, 88], "bbox": [0.5641093749999999, 0.007785547785547785, 0.9454062499999999, 0.2649417249417249], "iscrowd": 0, "area": 21271.824900000003, "rle": {"size": [429, 640], "counts": "^Yg4V1a:f1000000000000000001O01O00000000000000000000000001O000000000000000000000000001O00000000000000000000000000001O000000000000000000000000001O00000000000000000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001O000001O000001O00000000001O0000000000001O0000000000001O00000000001O0000000000001O0000000000001O000000000000J6[Oe000O100000000O1000000O100000000O100000000O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O100O2O0O1N2M3N2N2N2M3N2N2M5L4L4L4K5L4L4L4Kdd>"}, "label": "blurred white coffee cup"}]} {"id": 360002, "image": "COCO_train2014_000000360002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red sedan\"."}], "task": "refering", "answer_template": "The \"a red sedan\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 185, 186, 188, 189, 190, 191, 192, 200, 201, 203, 204, 205, 206, 218], "bbox": [0.3511241217798595, 0.5018125, 0.8745901639344263, 0.614609375], "iscrowd": 0, "area": 8291.813900000001, "rle": {"size": [640, 427], "counts": "bSn21mc03N1N3M2N3M2O2M2N3M2N3M2O2M2O2N1O2O0O2O001N101N101O0O2O0O2O001N101N101O0O2O0O2O0O2O001NM4cN\\fd0FX[[O7J7I6J6I8I6J4L2M3N2N2O1O1O1O1O1N101O1O1000000O100000O100000000000000000O0100000000000000000000O0100000000000000000O1000O10000000000000O10000000O1000000000O100O1O1O1N2O001O1O1O1O2M3N2N1O2N2N1O2N2M2O2O1O1O001O1N101O1O001O1O1O001O1O001O1O001O1N2O001O1O001O1O001O1O1O001O1O0O2O1O001O1O1O00i\\Q1"}, "label": "a red sedan"}]} {"id": 360002, "image": "COCO_train2014_000000360002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red car\"."}], "task": "refering", "answer_template": "The \"a red car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [173, 174, 175, 176, 185, 186, 188, 189, 190, 191, 192, 200, 201, 203, 204, 205, 206, 218], "bbox": [0.3511241217798595, 0.5018125, 0.8745901639344263, 0.614609375], "iscrowd": 0, "area": 8291.813900000001, "rle": {"size": [640, 427], "counts": "bSn21mc03N1N3M2N3M2O2M2N3M2N3M2O2M2O2N1O2O0O2O001N101N101O0O2O0O2O001N101N101O0O2O0O2O0O2O001NM4cN\\fd0FX[[O7J7I6J6I8I6J4L2M3N2N2O1O1O1O1O1N101O1O1000000O100000O100000000000000000O0100000000000000000000O0100000000000000000O1000O10000000000000O10000000O1000000000O100O1O1O1N2O001O1O1O1O2M3N2N1O2N2N1O2N2M2O2O1O1O001O1N101O1O001O1O1O001O1O001O1O001O1N2O001O1O001O1O001O1O1O001O1O0O2O1O001O1O1O00i\\Q1"}, "label": "a red car"}]} {"id": 360002, "image": "COCO_train2014_000000360002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dog coming towards camera\"."}], "task": "refering", "answer_template": "The \"a dog coming towards camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [234, 235, 249, 250, 251, 252, 264, 265, 266, 267, 278, 279, 280, 281, 296], "bbox": [0.5769320843091335, 0.649203125, 0.8564168618266979, 0.838875], "iscrowd": 0, "area": 6373.708899999997, "rle": {"size": [640, 427], "counts": "SXj41mc04N1N2O0O100\\MH_A9]>MaA3^>OaA1_>0`A1`>N`A2`>O_A0b>1\\AOe>1ZAOf>3YAMg>4VAMj>6RAKo>8V_OA[17a?d0\\@]Oe?e0W@\\Ok?d0P@_OQ`0b0k_OAV`0`0e_OB\\`0`0__OCa`0`0Y_OCh`0a0h^OJXa0^1101O000O10OO2O1O0O2O1N1101N1O2O1N2Q_O^M^`0d2^_O^Mb`0c2[_O_Me`0b2Y_O`Mf`0`2Y_OaMg`0_2X_ObMh`0^2W_OdMh`0j20ON3O10\\M[_Om1d`0RN__Ol1c`0RN__Om1a`0RNb_Oj1``0SNd_Oi1\\a0M5L2N1000001OO2N1Oa0_O2O00000000UOm^OjNRa0U1\\_O^Nd`0a1n0O1O1O1O1O2004L4^^OYNb`0l1S_O[Nl`0j1j^O[NVa0V22O1O1L4K5L4K5K5`No]Og0Yb0ZOj]O=\\b0CU^OFQb09l00000O10000O10000O1000000O2O001N2O1O00jXV1"}, "label": "a dog coming towards camera"}]} {"id": 360002, "image": "COCO_train2014_000000360002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black dog jumping to catch frisbee\"."}], "task": "refering", "answer_template": "The \"a black dog jumping to catch frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [234, 235, 249, 250, 251, 252, 264, 265, 266, 267, 278, 279, 280, 281, 296], "bbox": [0.5769320843091335, 0.649203125, 0.8564168618266979, 0.838875], "iscrowd": 0, "area": 6373.708899999997, "rle": {"size": [640, 427], "counts": "SXj41mc04N1N2O0O100\\MH_A9]>MaA3^>OaA1_>0`A1`>N`A2`>O_A0b>1\\AOe>1ZAOf>3YAMg>4VAMj>6RAKo>8V_OA[17a?d0\\@]Oe?e0W@\\Ok?d0P@_OQ`0b0k_OAV`0`0e_OB\\`0`0__OCa`0`0Y_OCh`0a0h^OJXa0^1101O000O10OO2O1O0O2O1N1101N1O2O1N2Q_O^M^`0d2^_O^Mb`0c2[_O_Me`0b2Y_O`Mf`0`2Y_OaMg`0_2X_ObMh`0^2W_OdMh`0j20ON3O10\\M[_Om1d`0RN__Ol1c`0RN__Om1a`0RNb_Oj1``0SNd_Oi1\\a0M5L2N1000001OO2N1Oa0_O2O00000000UOm^OjNRa0U1\\_O^Nd`0a1n0O1O1O1O1O2004L4^^OYNb`0l1S_O[Nl`0j1j^O[NVa0V22O1O1L4K5L4K5K5`No]Og0Yb0ZOj]O=\\b0CU^OFQb09l00000O10000O10000O1000000O2O001N2O1O00jXV1"}, "label": "a black dog jumping to catch frisbee"}]} {"id": 572998, "image": "COCO_train2014_000000572998.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"referee dressed in a black shirt\"."}], "task": "refering", "answer_template": "The \"referee dressed in a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 163, 164, 165, 185, 186, 187, 188, 207, 208, 209, 210, 230, 231, 232, 233, 234, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325], "bbox": [0.0015, 0.45393442622950825, 0.20540625, 0.9842622950819672], "iscrowd": 0, "area": 15375.392600000001, "rle": {"size": [427, 640], "counts": "mf01W=4L4M3L4L4YE3_7ZOkFg0c1:Y7oNRGi0c1:Z7mNQGm0b18\\7lNoFP1b16^7kNnFR1a15`72]H0b72[H0d72YH0f71XH2g7OVH3i7OSH4l7NQH4n7NoG4P8NmG4R8NkG4T8MjG5U8MhG5W8MfG5Z8LcG7[8KbG7]8K`G7_8K^G7a8Z2O1O1O1O100O1O2N00M3M3M3M3M4L3M3M3M3M3M3M3MZHZLn5c3RJaLn5[3QJjLn5V3oIoLo5R3mIRMR6o2jIUMU6l2fIYMY6h2cI\\M\\6d2aI`M^6a2]IdMc6\\2YIhMf6Y2VIkMi6V2RIoMm6R2oHRNP7?gHTO5a0S7;jHROOh0V77lHW1S7iNPIU1o6kNSIT1m6kNUIT1j6mNWIR1h6nNZIQ1e6oN]IP1b6QO_In0a6RO`Im0`6SOaIl0_6TObIk0_6TObIk0^6UOcIj0]6VOeIh0[6XOfIg0[6XOfIg0Z6YOgIf0V6]OkIb0R6AoI?m5DTJ;h5XOZIZNo0]2d5\\O^IVNP1\\2_5@bITNT1W2W5GfIRNX1R2o4NjIPN\\1m1h45lInMa1h1c49nInMd1d1^4=oIoMh1_1Y4b0oIoMm1Z1U4e0oIQNQ2U1P4i0PJRNU2P1k3m0QJSNY2k0f3R1RJRN]2g0b3U1RJTNa2b0]3Y1SJUNe2=X3^1SJUNj28S3b1TJVNn23n2k0lIYO:HQ3Oj2P1mIUO9LV3Id2V1oISO7Ng9o0UFoN52f9P1WFjN46e9P1YFhN37e9P1QGPOo8Q1PGoNP9Q1PGoNP9Q1PGoNP9R1oFnNR9Q1oFlNS9T1mFgNX9Z1\\100010O0000O2M2N2N2N2N3L:G;EWQd6"}, "label": "referee dressed in a black shirt"}]} {"id": 572998, "image": "COCO_train2014_000000572998.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an umpire\"."}], "task": "refering", "answer_template": "The \"an umpire\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 163, 164, 165, 185, 186, 187, 188, 207, 208, 209, 210, 230, 231, 232, 233, 234, 253, 254, 255, 256, 276, 277, 278, 279, 299, 300, 301, 302, 322, 323, 324, 325], "bbox": [0.0015, 0.45393442622950825, 0.20540625, 0.9842622950819672], "iscrowd": 0, "area": 15375.392600000001, "rle": {"size": [427, 640], "counts": "mf01W=4L4M3L4L4YE3_7ZOkFg0c1:Y7oNRGi0c1:Z7mNQGm0b18\\7lNoFP1b16^7kNnFR1a15`72]H0b72[H0d72YH0f71XH2g7OVH3i7OSH4l7NQH4n7NoG4P8NmG4R8NkG4T8MjG5U8MhG5W8MfG5Z8LcG7[8KbG7]8K`G7_8K^G7a8Z2O1O1O1O100O1O2N00M3M3M3M3M4L3M3M3M3M3M3M3MZHZLn5c3RJaLn5[3QJjLn5V3oIoLo5R3mIRMR6o2jIUMU6l2fIYMY6h2cI\\M\\6d2aI`M^6a2]IdMc6\\2YIhMf6Y2VIkMi6V2RIoMm6R2oHRNP7?gHTO5a0S7;jHROOh0V77lHW1S7iNPIU1o6kNSIT1m6kNUIT1j6mNWIR1h6nNZIQ1e6oN]IP1b6QO_In0a6RO`Im0`6SOaIl0_6TObIk0_6TObIk0^6UOcIj0]6VOeIh0[6XOfIg0[6XOfIg0Z6YOgIf0V6]OkIb0R6AoI?m5DTJ;h5XOZIZNo0]2d5\\O^IVNP1\\2_5@bITNT1W2W5GfIRNX1R2o4NjIPN\\1m1h45lInMa1h1c49nInMd1d1^4=oIoMh1_1Y4b0oIoMm1Z1U4e0oIQNQ2U1P4i0PJRNU2P1k3m0QJSNY2k0f3R1RJRN]2g0b3U1RJTNa2b0]3Y1SJUNe2=X3^1SJUNj28S3b1TJVNn23n2k0lIYO:HQ3Oj2P1mIUO9LV3Id2V1oISO7Ng9o0UFoN52f9P1WFjN46e9P1YFhN37e9P1QGPOo8Q1PGoNP9Q1PGoNP9Q1PGoNP9R1oFnNR9Q1oFlNS9T1mFgNX9Z1\\100010O0000O2M2N2N2N2N3L:G;EWQd6"}, "label": "an umpire"}]} {"id": 572998, "image": "COCO_train2014_000000572998.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball bats man hitting the ball\"."}], "task": "refering", "answer_template": "The \"a baseball bats man hitting the ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 216, 217, 218, 219, 239, 240, 241, 242, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 289, 290, 307, 308], "bbox": [0.382546875, 0.3560889929742389, 0.656640625, 0.8955971896955504], "iscrowd": 0, "area": 20495.375899999988, "rle": {"size": [427, 640], "counts": "U`V3g0b<;E100O1O2N1O1N20010O001O00010O00001O010O00001O01O01O0I7G:E:jNW1K5J6K4L5J6K5K5K4L5J6L4M3N1O2N2N2N2M2O2N2N2O010O10O010000O010O10O01000O0100O10O01000O0100O010O100O01000O010O10N2M3M3N2M3M3PNo1U1lN3M3M4L3M4K3N3M3MQNVH`Ng7`1\\H_Nb7`1cH]N\\7c1hHZNV7f1nHXNQ7h1SIUNl6j1YISNf6m1]IQNb6n1cInM]6R2iIiMU6W2RJdMm5[2ZJ_Mf5a2_J[M`5d2fJWMZ5i2fJXMY5g2hJZMV5g2kJXMU5g2lJZMS5f2mJZMS5e2oJ[MP5d2QK]Mo4b2QK^Mo4a2SK_Mm4`2SK`Mn4^2TKbMm4\\2SKeMo4W2RKiMP5U2QKkMP5R2QKoMP5o1PKRNQ5l1PKTNQ5j1oJXNQ5f1oJ[NR5b1PK^NQ5`1oJaNR5]1nJdNS5Z1nJfNS5X1mJiNbMUOZ7o1VKmN]MUO^7m1UKSOn4k0RKUOo4j0RKUOP5i0PKWOQ5g0PKYOR5e0oJZOS5d0mJ\\OT5c0lJ\\OW5b0jJ]OV5c0jJ]OW5a0jJ_OW5`0jJ_OV5a0kJ^OV5`0mJ^OT5a0mJ]OU5b0kJ^OU5b0lJ]OU5a0lJ_OU5`0kJ@Y5 in this image.", "objects": [{"patches": [126, 127, 128, 148, 149, 150, 151, 152, 171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 216, 217, 218, 219, 239, 240, 241, 242, 262, 263, 264, 265, 266, 267, 284, 285, 286, 287, 289, 290, 307, 308], "bbox": [0.382546875, 0.3560889929742389, 0.656640625, 0.8955971896955504], "iscrowd": 0, "area": 20495.375899999988, "rle": {"size": [427, 640], "counts": "U`V3g0b<;E100O1O2N1O1N20010O001O00010O00001O010O00001O01O01O0I7G:E:jNW1K5J6K4L5J6K5K5K4L5J6L4M3N1O2N2N2N2M2O2N2N2O010O10O010000O010O10O01000O0100O10O01000O0100O010O100O01000O010O10N2M3M3N2M3M3PNo1U1lN3M3M4L3M4K3N3M3MQNVH`Ng7`1\\H_Nb7`1cH]N\\7c1hHZNV7f1nHXNQ7h1SIUNl6j1YISNf6m1]IQNb6n1cInM]6R2iIiMU6W2RJdMm5[2ZJ_Mf5a2_J[M`5d2fJWMZ5i2fJXMY5g2hJZMV5g2kJXMU5g2lJZMS5f2mJZMS5e2oJ[MP5d2QK]Mo4b2QK^Mo4a2SK_Mm4`2SK`Mn4^2TKbMm4\\2SKeMo4W2RKiMP5U2QKkMP5R2QKoMP5o1PKRNQ5l1PKTNQ5j1oJXNQ5f1oJ[NR5b1PK^NQ5`1oJaNR5]1nJdNS5Z1nJfNS5X1mJiNbMUOZ7o1VKmN]MUO^7m1UKSOn4k0RKUOo4j0RKUOP5i0PKWOQ5g0PKYOR5e0oJZOS5d0mJ\\OT5c0lJ\\OW5b0jJ]OV5c0jJ]OW5a0jJ_OW5`0jJ_OV5a0kJ^OV5`0mJ^OT5a0mJ]OU5b0kJ^OU5b0lJ]OU5a0lJ_OU5`0kJ@Y5 in this image.", "objects": [{"patches": [188, 189, 211, 212, 233, 234, 235, 236, 238, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 284, 302, 303, 304, 305, 325, 326, 327, 328], "bbox": [0.14254687500000002, 0.5799765807962529, 0.386421875, 0.9782201405152225], "iscrowd": 0, "area": 12995.527049999997, "rle": {"size": [427, 640], "counts": "[YV19XI6K6J5J7J5J6K2M3N2N2M3N2M3N2M3N1N3N2N2M3O1N2N100O2N1K6E:G:N1N3N1O2M2O1O1N2O1N200000000000000000000000dMPH]OP8a0QH@o7?RHAn7=SHDm7:UHFk79UHHk76WHJi74XHMh72YHNg70[H0e7O[H2e7L]H4c7J^H7c7H]H8e7FZH;g7DYH in this image.", "objects": [{"patches": [188, 189, 211, 212, 233, 234, 235, 236, 238, 256, 257, 258, 259, 260, 261, 279, 280, 281, 282, 284, 302, 303, 304, 305, 325, 326, 327, 328], "bbox": [0.14254687500000002, 0.5799765807962529, 0.386421875, 0.9782201405152225], "iscrowd": 0, "area": 12995.527049999997, "rle": {"size": [427, 640], "counts": "[YV19XI6K6J5J7J5J6K2M3N2N2M3N2M3N2M3N1N3N2N2M3O1N2N100O2N1K6E:G:N1N3N1O2M2O1O1N2O1N200000000000000000000000dMPH]OP8a0QH@o7?RHAn7=SHDm7:UHFk79UHHk76WHJi74XHMh72YHNg70[H0e7O[H2e7L]H4c7J^H7c7H]H8e7FZH;g7DYH in this image.", "objects": [{"patches": [19, 20, 22, 23, 35, 36, 38, 39, 51, 52, 53, 54, 55, 67, 68, 69, 70, 71, 73, 74, 75, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 130, 131, 132, 133, 134, 135, 136, 147, 148, 149, 150, 151, 152, 153, 154, 163, 164, 165, 166, 167, 168, 169, 170, 171, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 227, 228, 229, 231, 232, 233, 234, 235, 236, 237, 238, 239, 243, 244, 245, 248, 249, 250, 251, 252, 253, 254, 255, 265, 266, 267, 268, 269, 270, 271, 282, 283, 284, 285, 286, 287, 298, 299, 300, 301, 302, 303, 315, 316, 317, 318, 332, 333, 334, 349, 364], "bbox": [0.009930875576036866, 0.062921875, 1.0, 0.986515625], "iscrowd": 0, "area": 97197.81735000006, "rle": {"size": [640, 434], "counts": "od21ec0;M4N1O2N1O2N1N3N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N1O01O01O0001O01O0000010O00000010O0001O01O0001O0001O01^^OoMSa0Q2l^OPNTa0P2j^ORNVa0n1i^OSNS4Ki8R2RCTN`2HiN:d;j1RCVN^21]N6RWECi:>TEEl::SEHl:9QEJo:6\\C\\MX1`2];3WCaMY1_2`;;^DGc;7[DLe;4YDNGVMP;k2VE1GYMR;f2UE3E\\Ml7GcJi2iM7D^Mm7J`Ja2gM?IYMo7M]JZ2WMT19iLQ81YJU6bMnIS85VJl5cMSJV87RJk6l5]IoIb6o5fIVIo6i6U3N2N2O1N2N100O001O100O1O10O01O100O1O10O01O11O2N2M3N3M3M3M3M3L4M3M3M4L3M3L4M3L4L5L4K5K5L4K5K5L4K3M1O100mNjE[HW:d7XFmGi9R8]FiGd9U8T1O1O1O1N2O00001O0001O1O001N2O1N2N2O0O2N2N2N2N2N2M2O2N2^Ob0J5K5M3N2M3001O100O1O1O1O1O1O1O1003M2N3M2N3M2N3M2O1N2N1O2OkF]Gl6b8oHfGP7X8lHoGS7P8^HaH`7^7^HSIU7k6iHZIU7f6gH`IX7^6fHfIY7Y6eHlIZ7^900O100O100OO2N1O2N2N1O2N1N3N1N3M3M_IbDc4\\;]KiDa4T;_KQE^4m:bKVE^4h:aKYE`4e:aK\\E`4`:bK`E`4[:cKfE]4W:eKiE]4S:eKnE[4o9gKRFZ4j9hKVFY4h9hKYFY4c9iK]FX4a9iK`FW4]9kKcFW4Z9jKgFV4W9kKjFU4T9lKlFU4S9kKnFU4P9lKPGV4m8kKTGU4j8kKWGV4g8kKZGU4e8kK[GV4c8kK^GV4`8jKaGV4^8iKdGW4[8iKfGW4X8jKjGU4U8jKmGW4Q8iKPHW4o7iKRHW4l7iKWHV4h7jK\\HS4c7mK`HR4]7nKfHQ4Y7oKgHR4X7nKhHS4W7lKjHU4T7lKmHU4Q7kKPIU4o6jKRIW4m6iKTIW4k6iKVIW4i6hKYIX4f6hKZIZ4c6fK_IZ4`6fKaIZ4^6eKdI[4[6eKeI\\4Z6cKhI^4V6bKkI^4T6aKnI_4Q6aKoI`4o5`KSJ`4l5`KUJa4i5^KYJb4f5]K[Jd4d5\\K]Jd4b5[K`Je4_5ZKcJf4[5ZKgJf4X5YKkJf4T5XKPKg4o4XKSKh4l4WKWKh4h4WK[Kh4d4UK`Kk4_4SKeKl4Z4RKjKm4U4PKoKP5P4nJVLo4i3nJYLT5f3iJ\\LY5c3eJ_L\\5`3bJbL_5]3_JfLa5Y3]JiLd5V3[JlLe5S3YJoLh5P3WJRMi5m2UJUMl5i2TJYMl5f2RJ^Mm5a2RJcMl5\\2RJiMl5V2SJmMl5R2RJQNn5n1PJUNP6j1kI[NV6i1`I]N`6a71O100O1O100O1O100O1O100O1O100O1O1001O00WERI_8n6\\GUIg8j6UGWIn8i6nFYIU9f6fF\\I]9d6_F^Ic9b6XF`Il9_6PFcIS:\\6hEfI\\:Y6`EiIb:W6ZEjIj:U6QEnIR;Q6jDPJY;P6bDSJ`;m5\\DTJg;l5TDWJn;i5nCXJT in this image.", "objects": [{"patches": [19, 20, 22, 23, 35, 36, 38, 39, 51, 52, 53, 54, 55, 67, 68, 69, 70, 71, 73, 74, 75, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 130, 131, 132, 133, 134, 135, 136, 147, 148, 149, 150, 151, 152, 153, 154, 163, 164, 165, 166, 167, 168, 169, 170, 171, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 227, 228, 229, 231, 232, 233, 234, 235, 236, 237, 238, 239, 243, 244, 245, 248, 249, 250, 251, 252, 253, 254, 255, 265, 266, 267, 268, 269, 270, 271, 282, 283, 284, 285, 286, 287, 298, 299, 300, 301, 302, 303, 315, 316, 317, 318, 332, 333, 334, 349, 364], "bbox": [0.009930875576036866, 0.062921875, 1.0, 0.986515625], "iscrowd": 0, "area": 97197.81735000006, "rle": {"size": [640, 434], "counts": "od21ec0;M4N1O2N1O2N1N3N1O2N1O1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N1O01O01O0001O01O0000010O00000010O0001O01O0001O0001O01^^OoMSa0Q2l^OPNTa0P2j^ORNVa0n1i^OSNS4Ki8R2RCTN`2HiN:d;j1RCVN^21]N6RWECi:>TEEl::SEHl:9QEJo:6\\C\\MX1`2];3WCaMY1_2`;;^DGc;7[DLe;4YDNGVMP;k2VE1GYMR;f2UE3E\\Ml7GcJi2iM7D^Mm7J`Ja2gM?IYMo7M]JZ2WMT19iLQ81YJU6bMnIS85VJl5cMSJV87RJk6l5]IoIb6o5fIVIo6i6U3N2N2O1N2N100O001O100O1O10O01O100O1O10O01O11O2N2M3N3M3M3M3M3L4M3M3M4L3M3L4M3L4L5L4K5K5L4K5K5L4K3M1O100mNjE[HW:d7XFmGi9R8]FiGd9U8T1O1O1O1N2O00001O0001O1O001N2O1N2N2O0O2N2N2N2N2N2M2O2N2^Ob0J5K5M3N2M3001O100O1O1O1O1O1O1O1003M2N3M2N3M2N3M2O1N2N1O2OkF]Gl6b8oHfGP7X8lHoGS7P8^HaH`7^7^HSIU7k6iHZIU7f6gH`IX7^6fHfIY7Y6eHlIZ7^900O100O100OO2N1O2N2N1O2N1N3N1N3M3M_IbDc4\\;]KiDa4T;_KQE^4m:bKVE^4h:aKYE`4e:aK\\E`4`:bK`E`4[:cKfE]4W:eKiE]4S:eKnE[4o9gKRFZ4j9hKVFY4h9hKYFY4c9iK]FX4a9iK`FW4]9kKcFW4Z9jKgFV4W9kKjFU4T9lKlFU4S9kKnFU4P9lKPGV4m8kKTGU4j8kKWGV4g8kKZGU4e8kK[GV4c8kK^GV4`8jKaGV4^8iKdGW4[8iKfGW4X8jKjGU4U8jKmGW4Q8iKPHW4o7iKRHW4l7iKWHV4h7jK\\HS4c7mK`HR4]7nKfHQ4Y7oKgHR4X7nKhHS4W7lKjHU4T7lKmHU4Q7kKPIU4o6jKRIW4m6iKTIW4k6iKVIW4i6hKYIX4f6hKZIZ4c6fK_IZ4`6fKaIZ4^6eKdI[4[6eKeI\\4Z6cKhI^4V6bKkI^4T6aKnI_4Q6aKoI`4o5`KSJ`4l5`KUJa4i5^KYJb4f5]K[Jd4d5\\K]Jd4b5[K`Je4_5ZKcJf4[5ZKgJf4X5YKkJf4T5XKPKg4o4XKSKh4l4WKWKh4h4WK[Kh4d4UK`Kk4_4SKeKl4Z4RKjKm4U4PKoKP5P4nJVLo4i3nJYLT5f3iJ\\LY5c3eJ_L\\5`3bJbL_5]3_JfLa5Y3]JiLd5V3[JlLe5S3YJoLh5P3WJRMi5m2UJUMl5i2TJYMl5f2RJ^Mm5a2RJcMl5\\2RJiMl5V2SJmMl5R2RJQNn5n1PJUNP6j1kI[NV6i1`I]N`6a71O100O1O100O1O100O1O100O1O100O1O1001O00WERI_8n6\\GUIg8j6UGWIn8i6nFYIU9f6fF\\I]9d6_F^Ic9b6XF`Il9_6PFcIS:\\6hEfI\\:Y6`EiIb:W6ZEjIj:U6QEnIR;Q6jDPJY;P6bDSJ`;m5\\DTJg;l5TDWJn;i5nCXJT in this image.", "objects": [{"patches": [168, 169, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 302, 303, 304, 326], "bbox": [0.08303125, 0.5012206572769953, 0.463796875, 0.9700469483568075], "iscrowd": 0, "area": 26011.20915, "rle": {"size": [426, 640], "counts": "^Zf01X=3L3N2M4dCFb;=ZDFf;;XDGg;;WDGf;;XDGg;>TDDj;P1M2O2M3N1O2M3N1N3N2N2M3N3L5L3M4K4M2M4M2N3L2O1O0O100O2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N3N4K3N2N1N100O3N2N3L010O0100_OjGoKT8R4mGmKS8R4SHjKk7V4XHhKh7X4]HbKd7]4d010O10O01O010O10O010O01O10O010O0100O0010O0100O010O00100O010O1O010O010O10O01O010O10O010O01O10O010O0100O00100O010O010O1O010O010O10O01O010O10O010O01O10O0100O0010O0100O010O1O100O100O105J1O2O0O2O1N3N1N2N2O0O100O100O1O2O1N2O1N2O1N2N3N3L5L2M2N101N101N101N1O101N100O2O0O1O100O2O0O2O0O1O2O0O2O0O2O0O1O2O0O20O2N5LN1O2N1O1mNbD9`;FdD5];KR1O1N1M4MP^^4"}, "label": "a tan modern looking couch being view from a higher angle"}]} {"id": 523863, "image": "COCO_train2014_000000523863.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"couch and pillows on top which is behind ottoman\"."}], "task": "refering", "answer_template": "The \"couch and pillows on top which is behind ottoman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [168, 169, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 217, 232, 233, 234, 235, 236, 237, 238, 239, 240, 255, 256, 257, 258, 259, 260, 261, 262, 278, 279, 280, 281, 282, 283, 302, 303, 304, 326], "bbox": [0.08303125, 0.5012206572769953, 0.463796875, 0.9700469483568075], "iscrowd": 0, "area": 26011.20915, "rle": {"size": [426, 640], "counts": "^Zf01X=3L3N2M4dCFb;=ZDFf;;XDGg;;WDGf;;XDGg;>TDDj;P1M2O2M3N1O2M3N1N3N2N2M3N3L5L3M4K4M2M4M2N3L2O1O0O100O2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1O1N3N4K3N2N1N100O3N2N3L010O0100_OjGoKT8R4mGmKS8R4SHjKk7V4XHhKh7X4]HbKd7]4d010O10O01O010O10O010O01O10O010O0100O0010O0100O010O00100O010O1O010O010O10O01O010O10O010O01O10O010O0100O00100O010O010O1O010O010O10O01O010O10O010O01O10O0100O0010O0100O010O1O100O100O105J1O2O0O2O1N3N1N2N2O0O100O100O1O2O1N2O1N2O1N2N3N3L5L2M2N101N101N101N1O101N100O2O0O1O100O2O0O2O0O1O2O0O2O0O2O0O1O2O0O20O2N5LN1O2N1O1mNbD9`;FdD5];KR1O1N1M4MP^^4"}, "label": "couch and pillows on top which is behind ottoman"}]} {"id": 564823, "image": "COCO_train2014_000000564823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of bread on a plate\"."}], "task": "refering", "answer_template": "The \"a piece of bread on a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98, 110, 111, 112, 113, 114, 115, 116, 128, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 168, 180, 181, 182, 183, 184, 185, 197, 198, 199, 200, 201, 202, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 249, 250, 251, 252, 253, 266, 267, 268, 269, 270, 284, 285, 286, 287, 302, 303], "bbox": [0.4834791666666666, 0.13999999999999999, 0.9281041666666667, 0.767703125], "iscrowd": 0, "area": 50615.165550000005, "rle": {"size": [640, 480], "counts": "cSa43ic07J7I6J7I6J7H7K6L4L3M4L3M4L3M4L3M4L3M4L3M4L3M4L3M4L3N3L3M4L3M4L3M4L3M4L4L3M4L3M4L3M4L3M4L3M3M2O2M3M3M3M3M2N3M3M3M3M3M3M2N3M3M3M3M3M3N1N3N2N2M3N2N1O2M3N2N2M3N2N1O2M3N2N2M3N2N1N3N2N2N2M3N1O2M3N2N2M3N2N1O1N10001N1000001N10000O2O000O101O000O101O0O101O0O10001O0000001O0001O01O0000001O01O0001OO1O101N1O1O1O1O1O1O2O0O1O1O1O1O1O2O0O1O1O1O1O2N1O100O1O1O1O2N3M4L3M4L3L4M4L3M4K4M4L3L5L3M3M4K4L5I6J7I6J7H7J6J7I6J7I6J7H7J9G=C>B=C>A>YOh0YOg0YOh`e0"}, "label": "a piece of bread on a plate"}]} {"id": 564823, "image": "COCO_train2014_000000564823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top piece of bread for a sandwich\"."}], "task": "refering", "answer_template": "The \"the top piece of bread for a sandwich\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [59, 60, 61, 62, 76, 77, 78, 79, 80, 93, 94, 95, 96, 97, 98, 110, 111, 112, 113, 114, 115, 116, 128, 129, 130, 131, 132, 133, 145, 146, 147, 148, 149, 150, 162, 163, 164, 165, 166, 167, 168, 180, 181, 182, 183, 184, 185, 197, 198, 199, 200, 201, 202, 214, 215, 216, 217, 218, 219, 232, 233, 234, 235, 236, 249, 250, 251, 252, 253, 266, 267, 268, 269, 270, 284, 285, 286, 287, 302, 303], "bbox": [0.4834791666666666, 0.13999999999999999, 0.9281041666666667, 0.767703125], "iscrowd": 0, "area": 50615.165550000005, "rle": {"size": [640, 480], "counts": "cSa43ic07J7I6J7I6J7H7K6L4L3M4L3M4L3M4L3M4L3M4L3M4L3M4L3M4L3N3L3M4L3M4L3M4L3M4L4L3M4L3M4L3M4L3M4L3M3M2O2M3M3M3M3M2N3M3M3M3M3M3M2N3M3M3M3M3M3N1N3N2N2M3N2N1O2M3N2N2M3N2N1O2M3N2N2M3N2N1N3N2N2N2M3N1O2M3N2N2M3N2N1O1N10001N1000001N10000O2O000O101O000O101O0O101O0O10001O0000001O0001O01O0000001O01O0001OO1O101N1O1O1O1O1O1O2O0O1O1O1O1O1O2O0O1O1O1O1O2N1O100O1O1O1O2N3M4L3M4L3L4M4L3M4K4M4L3L5L3M3M4K4L5I6J7I6J7H7J6J7I6J7I6J7H7J9G=C>B=C>A>YOh0YOg0YOh`e0"}, "label": "the top piece of bread for a sandwich"}]} {"id": 564823, "image": "COCO_train2014_000000564823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sandwich filled with lettuce , tomato , and crumbled cheese\"."}], "task": "refering", "answer_template": "The \"a sandwich filled with lettuce , tomato , and crumbled cheese\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 74, 75, 87, 88, 89, 90, 91, 92, 93, 103, 104, 105, 106, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125, 126, 127, 128, 137, 138, 139, 140, 141, 142, 143, 144, 145, 154, 155, 156, 157, 158, 159, 160, 161, 162, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 309, 310, 311, 312, 313, 314, 315, 316, 317, 327, 328, 329, 330, 332, 333, 334], "bbox": [0.04608333333333334, 0.170375, 0.7572083333333333, 0.8617343749999999], "iscrowd": 0, "area": 97025.60615, "rle": {"size": [640, 480], "counts": "aR>4bc0l2oARMU>k2kASMZ>k2eATM_>j2aATMc>j2\\AUMh>h2XAWMl>g2TAWMP?g2o@XMU?f2j@YM[?c2f@[M^?a2c@]Mb?`2^@_Mf?]2R1L4M3L4L4M3L3M2O1N3M2O2M2N3N1N2N3N1N3M2O1N3N1N3M2O2M2N2O2M2N3N1NReX2"}, "label": "a sandwich filled with lettuce , tomato , and crumbled cheese"}]} {"id": 564823, "image": "COCO_train2014_000000564823.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a huge sandwich with lots of veggies piled on a blue plate\"."}], "task": "refering", "answer_template": "The \"a huge sandwich with lots of veggies piled on a blue plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 72, 73, 74, 75, 87, 88, 89, 90, 91, 92, 93, 103, 104, 105, 106, 107, 108, 109, 110, 111, 120, 121, 122, 123, 124, 125, 126, 127, 128, 137, 138, 139, 140, 141, 142, 143, 144, 145, 154, 155, 156, 157, 158, 159, 160, 161, 162, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 309, 310, 311, 312, 313, 314, 315, 316, 317, 327, 328, 329, 330, 332, 333, 334], "bbox": [0.04608333333333334, 0.170375, 0.7572083333333333, 0.8617343749999999], "iscrowd": 0, "area": 97025.60615, "rle": {"size": [640, 480], "counts": "aR>4bc0l2oARMU>k2kASMZ>k2eATM_>j2aATMc>j2\\AUMh>h2XAWMl>g2TAWMP?g2o@XMU?f2j@YM[?c2f@[M^?a2c@]Mb?`2^@_Mf?]2R1L4M3L4L4M3L3M2O1N3M2O2M2N3N1N2N3N1N3M2O1N3N1N3M2O2M2N2O2M2N3N1NReX2"}, "label": "a huge sandwich with lots of veggies piled on a blue plate"}]} {"id": 466523, "image": "COCO_train2014_000000466523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"vehicle with hood up\"."}], "task": "refering", "answer_template": "The \"vehicle with hood up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213], "bbox": [0.008375, 0.4853958944281525, 0.28137500000000004, 0.7887683284457478], "iscrowd": 0, "area": 14607.3916, "rle": {"size": [341, 640], "counts": "nl13i9i0@`0K5J6M4N1O1O1O2O0O1O1O2N1O1000001O00000000001O00000001O000001O000000000000001O000000000001O01O000000001O1N101N2N2O1N2N101O1N2O001O1O0O2O00000O100000O0100000000O1000000O1000000O10O100000O10000O100O100O100O10O100000000000000000000000000O10000000000000001O000000000O100000000000001O001O001O001N101O001O001O00001O001O1N3N1O2N1N3N2N1O3L3N2N3L3M3M4nNZGNi8KhGHfWi4"}, "label": "vehicle with hood up"}]} {"id": 466523, "image": "COCO_train2014_000000466523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red truck has it ' s hood and doors open\"."}], "task": "refering", "answer_template": "The \"a red truck has it ' s hood and doors open\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [119, 120, 138, 139, 140, 141, 142, 143, 144, 161, 162, 163, 164, 165, 166, 167, 184, 185, 186, 187, 188, 189, 190, 209, 210, 211, 212, 213], "bbox": [0.008375, 0.4853958944281525, 0.28137500000000004, 0.7887683284457478], "iscrowd": 0, "area": 14607.3916, "rle": {"size": [341, 640], "counts": "nl13i9i0@`0K5J6M4N1O1O1O2O0O1O1O2N1O1000001O00000000001O00000001O000001O000000000000001O000000000001O01O000000001O1N101N2N2O1N2N101O1N2O001O1O0O2O00000O100000O0100000000O1000000O1000000O10O100000O10000O100O100O100O10O100000000000000000000000000O10000000000000001O000000000O100000000000001O001O001O001N101O001O001O00001O001O1N3N1O2N1N3N2N1O3L3N2N3L3M3M4nNZGNi8KhGHfWi4"}, "label": "a red truck has it ' s hood and doors open"}]} {"id": 466523, "image": "COCO_train2014_000000466523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"silver camper towed by red car\"."}], "task": "refering", "answer_template": "The \"silver camper towed by red car\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 159, 174, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206], "bbox": [0.59965625, 0.4559824046920821, 0.96715625, 0.7549560117302053], "iscrowd": 0, "area": 18195.142050000002, "rle": {"size": [341, 640], "counts": "jRP49X:8K4L4L2M2O2N2N1B>N2M3M3N2O1O1N3N1O1O1O1N2O1O1O1O10000O10000O1000000O10000O10001N1000000O10000O2O00000O1001O0000000000000001O00000000000000000001O0000000000000000000000000000000000001O00O1000000000000000000O1000000O101N3M3N2M2O0000000O100000O100O1O100O1000000000000000000002N1O1O2N1O001O0O101O00000009F3N000O10000O10000O10000O100O1000000000000O100000000000000000001O0000000000001N1000000000001O000000001O001O001O1O001O001O1O001O001O1O1O2N2N2N1O2N2N1O5K9G9F7J4L5K5J6K4LUh6"}, "label": "silver camper towed by red car"}]} {"id": 466523, "image": "COCO_train2014_000000466523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver van in the road\"."}], "task": "refering", "answer_template": "The \"a silver van in the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 158, 159, 174, 175, 176, 177, 178, 179, 180, 181, 182, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206], "bbox": [0.59965625, 0.4559824046920821, 0.96715625, 0.7549560117302053], "iscrowd": 0, "area": 18195.142050000002, "rle": {"size": [341, 640], "counts": "jRP49X:8K4L4L2M2O2N2N1B>N2M3M3N2O1O1N3N1O1O1O1N2O1O1O1O10000O10000O1000000O10000O10001N1000000O10000O2O00000O1001O0000000000000001O00000000000000000001O0000000000000000000000000000000000001O00O1000000000000000000O1000000O101N3M3N2M2O0000000O100000O100O1O100O1000000000000000000002N1O1O2N1O001O0O101O00000009F3N000O10000O10000O10000O100O1000000000000O100000000000000000001O0000000000001N1000000000001O000000001O001O001O1O001O001O1O001O001O1O1O2N2N2N1O2N2N1O5K9G9F7J4L5K5J6K4LUh6"}, "label": "a silver van in the road"}]} {"id": 466523, "image": "COCO_train2014_000000466523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red car with a white roof inbetween two other cars\"."}], "task": "refering", "answer_template": "The \"a red car with a white roof inbetween two other cars\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 197], "bbox": [0.286390625, 0.5623460410557184, 0.5944843750000001, 0.7489149560117302], "iscrowd": 0, "area": 8673.867949999996, "rle": {"size": [341, 640], "counts": "_Um12a:4M2O0000001]Ob0N2O10000000000O10001O001O2N1O1O1O1O2N001O0O101O001O0000O100000000000000O100O100O1N2O1N2H8O11O1N2N2N2N2M3N2N2N1O1O1O1O1N2O10000O10001O0000000000000000000000000000O010000000000000000000000000000000000000000000000000000000000001O000000000000000000000O1000000001O002N1O2N001O1O10O00000001O000001O000000001O000O2N1O2N1O2L3HnFkNU9T19M1O1O11OO2O0O1O1O1O1O1O0000000000000000000000000O2O0CVFOj9O[FNg9O^FL]ef2"}, "label": "a red car with a white roof inbetween two other cars"}]} {"id": 466523, "image": "COCO_train2014_000000466523.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a classic car attached to a silver trailer\"."}], "task": "refering", "answer_template": "The \"a classic car attached to a silver trailer\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [147, 148, 149, 150, 167, 168, 169, 170, 171, 172, 173, 174, 190, 191, 192, 193, 194, 195, 196, 197], "bbox": [0.286390625, 0.5623460410557184, 0.5944843750000001, 0.7489149560117302], "iscrowd": 0, "area": 8673.867949999996, "rle": {"size": [341, 640], "counts": "_Um12a:4M2O0000001]Ob0N2O10000000000O10001O001O2N1O1O1O1O2N001O0O101O001O0000O100000000000000O100O100O1N2O1N2H8O11O1N2N2N2N2M3N2N2N1O1O1O1O1N2O10000O10001O0000000000000000000000000000O010000000000000000000000000000000000000000000000000000000000001O000000000000000000000O1000000001O002N1O2N001O1O10O00000001O000001O000000001O000O2N1O2N1O2L3HnFkNU9T19M1O1O11OO2O0O1O1O1O1O1O0000000000000000000000000O2O0CVFOj9O[FNg9O^FL]ef2"}, "label": "a classic car attached to a silver trailer"}]} {"id": 196198, "image": "COCO_train2014_000000196198.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a ceramic top to a toilet tank\"."}], "task": "refering", "answer_template": "The \"a ceramic top to a toilet tank\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 28, 29, 32, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 132, 137, 138, 139, 140, 141], "bbox": [0.28614, 0.10807980049875313, 0.93776, 0.5586034912718204], "iscrowd": 0, "area": 46740.50025, "rle": {"size": [401, 500], "counts": "UTh1<^;i0^Ob0L3N3L3M4M3L3M4M2M4]FhMg8\\2WGdMh8^2WGbMg8b2WG_Mf8e2XG[Mf8h2YGXMf8k2XGVMe8m2ZGSMd8^3O2M2N2O2M2N2O2M2N2O2M2N2O2M20010O00001O2N3M3M3M1O0000O2N100O1O1O100O1O100O1O1O2O0O1O100O10000000000000001O00000000000000000000001O0000000O100000000000001O00000000000000000000001O0000000O10000000000000001O000000000000000000000010O0000001O0000001O0000010O0000001O0000001O000000010O0000001O0000001O0000010O0000001O0000001O000001O01O0000001O0000001O0000010O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O000000001O00000O2O000000001O0000001O0000001O000000001O00O100O1O1O1O1O1O1O1O1O100O1O100O100O100O100O100O100O100O100O100O100O100O2N100O1O1O101N1O1O100O1O2N100O1O1O101N1O2N;F:E in this image.", "objects": [{"patches": [24, 25, 26, 27, 28, 29, 32, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 132, 137, 138, 139, 140, 141], "bbox": [0.28614, 0.10807980049875313, 0.93776, 0.5586034912718204], "iscrowd": 0, "area": 46740.50025, "rle": {"size": [401, 500], "counts": "UTh1<^;i0^Ob0L3N3L3M4M3L3M4M2M4]FhMg8\\2WGdMh8^2WGbMg8b2WG_Mf8e2XG[Mf8h2YGXMf8k2XGVMe8m2ZGSMd8^3O2M2N2O2M2N2O2M2N2O2M2N2O2M20010O00001O2N3M3M3M1O0000O2N100O1O1O100O1O100O1O1O2O0O1O100O10000000000000001O00000000000000000000001O0000000O100000000000001O00000000000000000000001O0000000O10000000000000001O000000000000000000000010O0000001O0000001O0000010O0000001O0000001O000000010O0000001O0000001O0000010O0000001O0000001O000001O01O0000001O0000001O0000010O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O000000001O00000O2O000000001O0000001O0000001O000000001O00O100O1O1O1O1O1O1O1O1O100O1O100O100O100O100O100O100O100O100O100O100O100O2N100O1O1O101N1O1O100O1O2N100O1O1O101N1O2N;F:E in this image.", "objects": [{"patches": [116, 132, 133, 134, 135, 136, 137, 138, 139, 150, 151, 152, 153, 154, 155, 156, 157, 158, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194, 204, 205, 206, 207, 208, 209, 210, 211, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247], "bbox": [0.3397, 0.4951870324189526, 0.84256, 1.0], "iscrowd": 0, "area": 40049.7802, "rle": {"size": [401, 500], "counts": "giR27W<6L4K5K5K5`HZOn2j0hL]OW3g0_L@`3c0WLDh3?oKHQ4:hKKW47fKKY47dKK\\45bKM]44aKN^43`KO`42]K0b41\\K1c40\\K1d40YK2f4OXK3g4NWK4i4MTK5k4LSK6l4KRK7n4IPK9o4InJ9Q5HmJ:S5FkJZ5CeJ>[5BdJ?\\5BbJ?]5BbJ?^5A`Ja0_5A_J`0a5@^Ja0b5_O]Jb0b5@\\Ja0d5_O[Jb0e5^OZJc0e5^OZJc0f5^OXJc0h5]OVJe0i5\\OVJe0j5\\OTJe0l5[OSJf0l5[OSJf0m5[OPJg0P6YOnIi0Q6XOmIj0S6VOkIl0U6UOhIm0W6g20000O10000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000001O000000000000000000000000000000001O0000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O000000001O0000001O000000001O0RM\\IOd61]INc62^IMb62`IMa62aIL_64bIK_64bIK^65cIJ^65dII\\67eIH\\67eIH[68fIG[68gIFY6:hIEY6:hIEX6:jIEW6:jIEV6;lICU6oI@R6?RJ]On5c0VJYOk5f0XJWOh5i0\\JSOe5l0_JPOa5o0cJnN^5Q1fJkNZ5U1jJgNW5X1Q301O001O001N1O2O0O2N101N101N1O2O1N3M2O2M3M2O2M3M2O2M3M2O2M3L4@Xfn0"}, "label": "a toilet with a wooden seat"}]} {"id": 196198, "image": "COCO_train2014_000000196198.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white and brown uropian closet\"."}], "task": "refering", "answer_template": "The \"white and brown uropian closet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [116, 132, 133, 134, 135, 136, 137, 138, 139, 150, 151, 152, 153, 154, 155, 156, 157, 158, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 187, 188, 189, 190, 191, 192, 193, 194, 204, 205, 206, 207, 208, 209, 210, 211, 222, 223, 224, 225, 226, 227, 228, 229, 240, 241, 242, 243, 244, 245, 246, 247], "bbox": [0.3397, 0.4951870324189526, 0.84256, 1.0], "iscrowd": 0, "area": 40049.7802, "rle": {"size": [401, 500], "counts": "giR27W<6L4K5K5K5`HZOn2j0hL]OW3g0_L@`3c0WLDh3?oKHQ4:hKKW47fKKY47dKK\\45bKM]44aKN^43`KO`42]K0b41\\K1c40\\K1d40YK2f4OXK3g4NWK4i4MTK5k4LSK6l4KRK7n4IPK9o4InJ9Q5HmJ:S5FkJZ5CeJ>[5BdJ?\\5BbJ?]5BbJ?^5A`Ja0_5A_J`0a5@^Ja0b5_O]Jb0b5@\\Ja0d5_O[Jb0e5^OZJc0e5^OZJc0f5^OXJc0h5]OVJe0i5\\OVJe0j5\\OTJe0l5[OSJf0l5[OSJf0m5[OPJg0P6YOnIi0Q6XOmIj0S6VOkIl0U6UOhIm0W6g20000O10000000000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000000000001O000000000000000000000000000000001O0000001O000000001O000000001O0000001O000000001O000000001O000000001O0000001O000000001O000000001O0000001O000000001O0RM\\IOd61]INc62^IMb62`IMa62aIL_64bIK_64bIK^65cIJ^65dII\\67eIH\\67eIH[68fIG[68gIFY6:hIEY6:hIEX6:jIEW6:jIEV6;lICU6oI@R6?RJ]On5c0VJYOk5f0XJWOh5i0\\JSOe5l0_JPOa5o0cJnN^5Q1fJkNZ5U1jJgNW5X1Q301O001O001N1O2O0O2N101N101N1O2O1N3M2O2M3M2O2M3M2O2M3M2O2M3L4@Xfn0"}, "label": "white and brown uropian closet"}]} {"id": 327271, "image": "COCO_train2014_000000327271.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"white pickup truck parked on the side of the road\"."}], "task": "refering", "answer_template": "The \"white pickup truck parked on the side of the road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 122, 123, 124, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 301, 302, 303, 304, 309, 310, 325, 326, 327], "bbox": [0.044859375, 0.29583333333333334, 0.5087499999999999, 0.8522916666666667], "iscrowd": 0, "area": 58500.848500000015, "rle": {"size": [480, 640], "counts": "Yj=S1`=d0I6J7I7H8I6J7I7I7I5K5K4K6K5K4L5K3M001O0O101O00001O00001N10001O0000001O000O2O00N2L4L4M3L4O01000000N2L4L4L4L4L4L4L4M3L4L4L4L4L4L4L4L8H8L4L4L3M4L4L4L4K5L4L4L1O1O1O001O001O001O1O001O0O10000000O1000000000000O100000O100000O1000000L4K5J6K5J6K5K5M300000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000001O000000001O000000001O000000001O00000000001O000000001O000000001O0000001O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N1O22N1N201N10M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M3L3N3M2M4K4M4K4M4L7H9G8B>B?A>B?A>BVlb4"}, "label": "white pickup truck parked on the side of the road"}]} {"id": 327271, "image": "COCO_train2014_000000327271.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pickup vehicle\"."}], "task": "refering", "answer_template": "The \"a pickup vehicle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 122, 123, 124, 141, 142, 143, 144, 145, 146, 147, 148, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 301, 302, 303, 304, 309, 310, 325, 326, 327], "bbox": [0.044859375, 0.29583333333333334, 0.5087499999999999, 0.8522916666666667], "iscrowd": 0, "area": 58500.848500000015, "rle": {"size": [480, 640], "counts": "Yj=S1`=d0I6J7I7H8I6J7I7I7I5K5K4K6K5K4L5K3M001O0O101O00001O00001N10001O0000001O000O2O00N2L4L4M3L4O01000000N2L4L4L4L4L4L4L4M3L4L4L4L4L4L4L4L8H8L4L4L3M4L4L4L4K5L4L4L1O1O1O001O001O001O1O001O0O10000000O1000000000000O100000O100000O1000000L4K5J6K5J6K5K5M300000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000001O000000001O000000001O000000001O00000000001O000000001O000000001O0000001O2N1O2N2N1O2N1O2N2N1O2N1O2N2N1O2N1O2N1O22N1N201N10M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M2N3M3L3N3M2M4K4M4K4M4L7H9G8B>B?A>B?A>BVlb4"}, "label": "a pickup vehicle"}]} {"id": 441972, "image": "COCO_train2014_000000441972.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the horse in the left side\"."}], "task": "refering", "answer_template": "The \"the horse in the left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [33, 34, 35, 36, 37, 56, 57, 58, 59, 60, 79, 80, 81, 82, 101, 102, 103, 104, 105, 106, 123, 124, 125, 126, 127, 128, 129, 143, 144, 145, 146, 147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 258, 259, 260, 261, 262, 263, 264, 265, 266, 281, 282, 283, 284, 285, 286, 287, 288, 289, 304, 305, 306, 307, 308, 309, 310, 311, 327, 328, 329, 330, 331, 332, 333, 334], "bbox": [0.208984375, 0.05375586854460094, 0.6212500000000001, 0.9892488262910797], "iscrowd": 0, "area": 67241.49215000003, "rle": {"size": [426, 640], "counts": "kog1X1h;e0jNR1C=Db8YOiG`0\\8WOoGb0a:K5J6J6K_aT3"}, "label": "the horse in the left side"}]} {"id": 441972, "image": "COCO_train2014_000000441972.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"horse closest to yellow line on road\"."}], "task": "refering", "answer_template": "The \"horse closest to yellow line on road\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 68, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 289, 290, 291, 292, 293, 294, 295, 296, 311, 312, 313, 314, 315, 316, 317, 318], "bbox": [0.5539375, 0.13220657276995304, 0.9991875, 0.9265962441314555], "iscrowd": 0, "area": 60814.945999999996, "rle": {"size": [426, 640], "counts": "PRd43T=3M3M3M4K4M3M3M3M3M3M3L5L3M3SNBkFa0S93XF0f9f13L4L4L]NPMVIl2m6aMgH[2\\7QNYHk1j7aNkG\\1W8QO]GR1_8o1L4L4L4K5L3M4L4L4aI]JR5g5kJ[JR5i5kJYJU5h5iJYJW5h5fJYJ[5h5bJZJ^5f5aJ[J`5e5]J]Jc5d5ZJ^Jf5c5XJ^Jh5c5UJ_Jk5b5RJ`Jn5a5PJ`JQ6`5lIbJT6_5iIcJW6]5hIdJX6W6001N10000000001O00000000001O00000000001O000001O01O000000000000O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O100O100O100O100O100O100O100O100O100O1O1N2O1N2N2O1N2N2O1N2N2N2O1N2N2O1N2N2O1N2N2O1N2N2N20000O10000O10000O10000O10000O10000O100O10000QO_FTN`9P2lFbMR9a2i03M4L3M3M4L3M3M4L3M3M4K4M3M3M4L3M3M4K4K4M3L4L4L4L4L4L4E;C>E:N21O1O1O1O1O1O001O1O1O1O1O100O1O1O1O1O1O001O0001O000001O00000001O0001O0001O00O00O100O1O1O1^J\\I_4d6aK_I[4b6eKbIV4_6jKdIR4]6nKfIo3Z6RLhIj3Y6VLjIi1^OPOi6WOlI\\1KVOZ6^OoIn07^Oj5DRJb0c0C\\5KTJ2S1Kk44TJYOl1;S4;cM[Oa2d0eMSO_2l0P5O100O1O1O1N2M3MV;"}, "label": "horse closest to yellow line on road"}]} {"id": 441972, "image": "COCO_train2014_000000441972.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"second horse to the person left\"."}], "task": "refering", "answer_template": "The \"second horse to the person left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [65, 66, 68, 87, 88, 89, 90, 91, 109, 110, 111, 112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 266, 267, 268, 269, 270, 271, 272, 273, 289, 290, 291, 292, 293, 294, 295, 296, 311, 312, 313, 314, 315, 316, 317, 318], "bbox": [0.5539375, 0.13220657276995304, 0.9991875, 0.9265962441314555], "iscrowd": 0, "area": 60814.945999999996, "rle": {"size": [426, 640], "counts": "PRd43T=3M3M3M4K4M3M3M3M3M3M3L5L3M3SNBkFa0S93XF0f9f13L4L4L]NPMVIl2m6aMgH[2\\7QNYHk1j7aNkG\\1W8QO]GR1_8o1L4L4L4K5L3M4L4L4aI]JR5g5kJ[JR5i5kJYJU5h5iJYJW5h5fJYJ[5h5bJZJ^5f5aJ[J`5e5]J]Jc5d5ZJ^Jf5c5XJ^Jh5c5UJ_Jk5b5RJ`Jn5a5PJ`JQ6`5lIbJT6_5iIcJW6]5hIdJX6W6001N10000000001O00000000001O00000000001O000001O01O000000000000O1O1O100O1O1O1O100O1O1O100O1O1O1O100O1O1O1O1O100O1O1O1O100O1O1O1O1O100O1O100O100O100O100O100O100O100O100O100O1O1N2O1N2N2O1N2N2O1N2N2N2O1N2N2O1N2N2O1N2N2O1N2N2N20000O10000O10000O10000O10000O10000O100O10000QO_FTN`9P2lFbMR9a2i03M4L3M3M4L3M3M4L3M3M4K4M3M3M4L3M3M4K4K4M3L4L4L4L4L4L4E;C>E:N21O1O1O1O1O1O001O1O1O1O1O100O1O1O1O1O1O001O0001O000001O00000001O0001O0001O00O00O100O1O1O1^J\\I_4d6aK_I[4b6eKbIV4_6jKdIR4]6nKfIo3Z6RLhIj3Y6VLjIi1^OPOi6WOlI\\1KVOZ6^OoIn07^Oj5DRJb0c0C\\5KTJ2S1Kk44TJYOl1;S4;cM[Oa2d0eMSO_2l0P5O100O1O1O1N2M3MV;"}, "label": "second horse to the person left"}]} {"id": 441972, "image": "COCO_train2014_000000441972.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the two adults that are sitting behind the driver and the child in the carriage . one is facing away from the camera and has brown hair , the other one is facing the camera\"."}], "task": "refering", "answer_template": "The \"the two adults that are sitting behind the driver and the child in the carriage . one is facing away from the camera and has brown hair , the other one is facing the camera\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [72, 73, 94, 95, 96, 99, 100, 117, 118, 119, 122, 123, 140, 141, 142, 163], "bbox": [0.1008125, 0.22788732394366196, 0.38168749999999996, 0.5237793427230046], "iscrowd": 0, "area": 5142.921100000001, "rle": {"size": [426, 640], "counts": "kVk05R=5K6K7I7I7I;Ej9OkEBc:b0ZEUOn:d1O1O1O1O10O1000O10000O1000O0100000000001O0O100000000000001N1SOSE]Oo:a0RE^Oo:`0RE@o:>REBP;3Aa;N_D?4A];NaD?4CZ;MeD<3HX;JgD<2JP in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183], "bbox": [0.786265625, 0.004485776805251641, 1.0, 0.4831509846827134], "iscrowd": 0, "area": 24644.78310000001, "rle": {"size": [457, 640], "counts": "QeP77g=b0^O=B=DC in this image.", "objects": [{"patches": [18, 19, 20, 21, 22, 41, 42, 43, 44, 45, 64, 65, 66, 67, 68, 87, 88, 89, 90, 91, 110, 111, 112, 113, 114, 133, 134, 135, 136, 137, 156, 157, 158, 159, 160, 179, 180, 181, 182, 183], "bbox": [0.786265625, 0.004485776805251641, 1.0, 0.4831509846827134], "iscrowd": 0, "area": 24644.78310000001, "rle": {"size": [457, 640], "counts": "QeP77g=b0^O=B=DC in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 148, 149, 150, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 247], "bbox": [0.47254687500000003, 0.26459016393442625, 0.7869531249999999, 0.6760421545667448], "iscrowd": 0, "area": 23555.627049999992, "rle": {"size": [427, 640], "counts": "]Sn31W=4L5K4M4L;F9H8H001O0O10000000001VEdNc9\\1PFSOn9m0eEBY:j1M3N2N2N1O0000000O1000000000000O1000I7G9F9010000000O1000`0@6J3L5TOl0L3L1O1O1O1O1O1O1O1O1000000O100000000000000000N2M2N3L4M3003M4L3M3M4L3L5L1O000000001O0000000O10001O00L4K5LL36K4100000O001O1N2O001O000O10000000000O2N1K5M300000000000000002O2M3M3M3M3M1O1O00001O0010O01O0000000000O100000001O0000000O100000000O1SM[GP1i8kN]GP1h8jN^GR1e8iN`GS1d8hNbGT1a8gNeGX1\\8cNiG\\1W8`NnG_1S8]NWH_1g7\\NjH]1P7_N`I\\1Z6_NQJa1i86J4L000001O000000000001ON3L3N2N4K6K5K5J[_h1"}, "label": "a tractor with a pull card attached has two men sitting in it"}]} {"id": 507520, "image": "COCO_train2014_000000507520.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"fully loaded tractor with two persons on it\"."}], "task": "refering", "answer_template": "The \"fully loaded tractor with two persons on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 107, 108, 109, 126, 127, 128, 129, 130, 131, 132, 148, 149, 150, 151, 152, 153, 154, 155, 172, 173, 174, 175, 176, 177, 178, 195, 196, 197, 198, 199, 200, 201, 202, 218, 219, 220, 221, 222, 223, 224, 225, 247], "bbox": [0.47254687500000003, 0.26459016393442625, 0.7869531249999999, 0.6760421545667448], "iscrowd": 0, "area": 23555.627049999992, "rle": {"size": [427, 640], "counts": "]Sn31W=4L5K4M4L;F9H8H001O0O10000000001VEdNc9\\1PFSOn9m0eEBY:j1M3N2N2N1O0000000O1000000000000O1000I7G9F9010000000O1000`0@6J3L5TOl0L3L1O1O1O1O1O1O1O1O1000000O100000000000000000N2M2N3L4M3003M4L3M3M4L3L5L1O000000001O0000000O10001O00L4K5LL36K4100000O001O1N2O001O000O10000000000O2N1K5M300000000000000002O2M3M3M3M3M1O1O00001O0010O01O0000000000O100000001O0000000O100000000O1SM[GP1i8kN]GP1h8jN^GR1e8iN`GS1d8hNbGT1a8gNeGX1\\8cNiG\\1W8`NnG_1S8]NWH_1g7\\NjH]1P7_N`I\\1Z6_NQJa1i86J4L000001O000000000001ON3L3N2N4K6K5K5J[_h1"}, "label": "fully loaded tractor with two persons on it"}]} {"id": 155268, "image": "COCO_train2014_000000155268.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a green color sofa where opposite side of woman\"."}], "task": "refering", "answer_template": "The \"a green color sofa where opposite side of woman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 12, 27, 28, 29, 30, 31, 32, 33, 34, 35, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 163, 164, 165, 166, 167, 168, 169, 170, 171, 186, 187, 188, 189, 190, 191, 192, 193, 194], "bbox": [0.099203125, 0.003779342723004695, 0.566421875, 0.5847652582159625], "iscrowd": 0, "area": 58543.26659999999, "rle": {"size": [426, 640], "counts": "RZj0c1f;7I7J6I7I7I7I7I7J6I7I7J6I7J4K3M2O2M2O2M100O1O100O100O1O010O100O100O1O100O1O1O1O1O001O1O1O1O1O1O100O1O1O1O001O1O1fNeJdJ\\5\\5dJcJ]5\\5dJcJ]5]5dJaJ]5_5cJ_J_5a5aJ^J`5b5`J]Ja5c5`JZJb5f5^JYJb5h5`JUJa5k5`JQJc5o5^JnId5`5nIfJ?Ge5c5nIcJ?Fe5h5W100000O1000000O100003M3M000000000000O10000000000000000O1000000000O100000O10000000000000000O10000000000000000O10000000000000000O10000000000000000O100000000000000O1000O1000O100001O1O1O0000000000000000000000000000O100000000000000000000000000000000000000000000000000000O10000000000000O1O100000000000000000000000000000000001O0000000000QO_JUJa5e5U1N2M3N2N2N2N2N2N2N2N2N2O1N3M2N3M3M2N3M3M2N3L4M3M2N3M3L3N3M3J6K6I6J7L301O0O2O0O2O0O2N1O2M2O2N1O2N1O2N2M4M3L3N3L4L3N3K5K4L5KS1lN[Tc3"}, "label": "a green color sofa where opposite side of woman"}]} {"id": 155268, "image": "COCO_train2014_000000155268.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman sits in a hall\"."}], "task": "refering", "answer_template": "The \"woman sits in a hall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 105, 106, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 236, 238, 239, 240, 241, 242, 243, 244, 245, 246, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.22834374999999998, 0.304037558685446, 0.7014374999999999, 1.0], "iscrowd": 0, "area": 56157.12569999999, "rle": {"size": [426, 640], "counts": "dRm1;\\ZFG_9;aFJY96gF0S91mF0R90nF0Q90PG0o81PG0P90PG0o80RG0m81kFaNH^1]91iFcNI]1]90iFeNJ[1\\91iFdNL[1[91hFeNMZ1[91gFfNMZ1]9OfFgNMZ1]90dFgNOY1]90cFgN1Y1\\90bFhN1Y1]9ObFhN1Y1]9OaFiN2X1]9O`FjN3W1]9O_FjN4X1^9:bFF^9:bFF^9:bFF^9:aFG_98bFH^98bFH^98bFH_97aFI_97aFI_96aFK_95aFK_95aFK_95aFK_95aFK`94`FL`94_FMa93_FMa92`FN`92`FN`91aFO_90bF0_9O`F0b9O_FOc90^FOc91]FMe92\\FLf94YFLh93YFKj93WFKk95UFJl95UFIm96TFHn98QFH0SO]9S1dFH0VO\\9Q1cFJ2UO[9Q1_FL7TO[9n0[F0 in this image.", "objects": [{"patches": [104, 105, 106, 126, 127, 128, 129, 130, 149, 150, 151, 152, 153, 171, 172, 173, 174, 175, 176, 194, 195, 196, 197, 198, 199, 200, 217, 218, 219, 220, 221, 222, 223, 236, 238, 239, 240, 241, 242, 243, 244, 245, 246, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337], "bbox": [0.22834374999999998, 0.304037558685446, 0.7014374999999999, 1.0], "iscrowd": 0, "area": 56157.12569999999, "rle": {"size": [426, 640], "counts": "dRm1;\\ZFG_9;aFJY96gF0S91mF0R90nF0Q90PG0o81PG0P90PG0o80RG0m81kFaNH^1]91iFcNI]1]90iFeNJ[1\\91iFdNL[1[91hFeNMZ1[91gFfNMZ1]9OfFgNMZ1]90dFgNOY1]90cFgN1Y1\\90bFhN1Y1]9ObFhN1Y1]9OaFiN2X1]9O`FjN3W1]9O_FjN4X1^9:bFF^9:bFF^9:bFF^9:aFG_98bFH^98bFH^98bFH_97aFI_97aFI_96aFK_95aFK_95aFK_95aFK_95aFK`94`FL`94_FMa93_FMa92`FN`92`FN`91aFO_90bF0_9O`F0b9O_FOc90^FOc91]FMe92\\FLf94YFLh93YFKj93WFKk95UFJl95UFIm96TFHn98QFH0SO]9S1dFH0VO\\9Q1cFJ2UO[9Q1_FL7TO[9n0[F0 in this image.", "objects": [{"patches": [21, 22, 39, 40, 41, 42, 43, 44, 45, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 130, 131, 132, 133, 134, 135, 136, 137, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 200, 201, 202, 203, 204, 205, 223, 224, 225, 226, 227, 228, 246, 247, 248, 249, 250, 268, 269, 270, 271, 291, 292, 314, 315, 316, 336, 337, 338, 339, 340, 341], "bbox": [0.56828125, 0.0036619718309859155, 1.0, 0.9885915492957746], "iscrowd": 0, "area": 60349.13815000001, "rle": {"size": [426, 640], "counts": "S`g4h1b;4LO0011N100O10000O2O0O100O10001N100O10000O2O0O100O10001N100O100O101O0O100000001O000O1000001jM_MaIa2]6aMcI_2\\6bMdI`2X6bMhI`2T6cMkI`2Q6aMnIb2n5`MRJb2k5_MUJc2g5_MYJc2c5_M]Jc2`5^M`Jd2\\5^MdJd2X5^MhJd2U5]MkJe2Q5]MnJf2o4[MQKh2j4ZMVKh2f4ZMZKh2c4YM]Ki2_4YMaKi2\\4XMdKj2X4XMhKj2T4XMlKj2P4XMPLj2i3[MVLh2b3^M^Ld2[3aMeLa2S3eMmL^2j2hMVMZ2c2kM]MW2[2oMeMS2T2RNlMP2l1VNTNl1d1ZN[Ni1U1eNkNg1T8CkG=T8DlG in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 23, 24, 25, 26, 27, 46, 47, 48, 49, 50, 69, 70, 71, 72], "bbox": [0.0, 0.0, 0.20846874999999998, 0.24854460093896713], "iscrowd": 0, "area": 11208.628800000002, "rle": {"size": [426, 640], "counts": "h08U14P:;`E3X:6dEL[:8_EL_:a1N1O2O001O001O00001O00001O00001O00001O00001O001O00001O00001O00001O0000001O0000000000000000000000000000001O001O1O001O1O1O001O1N2ON200O100O100O100O100O10000O100O10000O100O1O100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O100fNPE9Q;FPE9Q;ERE9o:FRE9o:ETE9m:FUE8l:HTE7m:ITE5m:JTE5m:KTE3m:MSE2n:NRE1o:OQE0S;MmD2Y;HhD7];EeD7b;DaD8d;D_D7] in this image.", "objects": [{"patches": [0, 1, 2, 3, 4, 23, 24, 25, 26, 27, 46, 47, 48, 49, 50, 69, 70, 71, 72], "bbox": [0.0, 0.0, 0.20846874999999998, 0.24854460093896713], "iscrowd": 0, "area": 11208.628800000002, "rle": {"size": [426, 640], "counts": "h08U14P:;`E3X:6dEL[:8_EL_:a1N1O2O001O001O00001O00001O00001O00001O00001O001O00001O00001O00001O0000001O0000000000000000000000000000001O001O1O001O1O1O001O1N2ON200O100O100O100O100O10000O100O10000O100O1O100O1O1O1O1O1O1O1O1O1O100O1O1O1O1O1O1O1O1O1O1O100fNPE9Q;FPE9Q;ERE9o:FRE9o:ETE9m:FUE8l:HTE7m:ITE5m:JTE5m:KTE3m:MSE2n:NRE1o:OQE0S;MmD2Y;HhD7];EeD7b;DaD8d;D_D7] in this image.", "objects": [{"patches": [79, 97, 114, 115, 116, 117, 132, 133, 134, 135, 150, 151, 152, 153], "bbox": [0.35246, 0.38323353293413176, 0.52264, 0.7546407185628743], "iscrowd": 0, "area": 4155.8511, "rle": {"size": [334, 500], "counts": "aci11W:6VFJZ98fFIX98hFIV98jFIS99mFHQ99nFIP98PGIn88RGIk89UGHi89WGIf88ZGId89[GHd88\\GIc87]GJb86^GJb86^GJb86^GKa85_GKa85^GLb84^GMa83_GMa83_GMa83_GM`84`GM_83aGM_83aGM_83aGN^82bGN_7]OTIe0]ON]7ATIa0_OM]7U1bHlNS7_1mHaNo6c1QI]Nk6g1UIYNg6k1YITNg6m1YISNg6m1YISNf6n1ZIRNf6n1ZIRNe6LVI_15eNj6FTIc12fNQ7_O]I]1BTOc83]GMk8KTG6l8JTG6Q8_O_H;@5Q8E\\H5C5R8GZH4D3T8IWH5E1U8KjG?1EV8k0kGSOW8m0iGROX8n0hGTOV8l0jGVOT8j0lGYOQ8EmGd01JP8BnGe02Lm7^OQHf03Nk7[OQHh040h7XOSHi051f7UOUHk052d7SOVHl061e7QOVHn051f7POUHo051k7NVH2j7NVH3j7LVH4j7LVH4k7KTH6l7JTH5n7JRH5P8IQH6Q8IoG6S8ImG5V8JjG5X8JhG3[8MeGO`80`GKj8OWG0o8KQG5T9FnF7W9EjF:f9O2Mdg]2"}, "label": "a bicycle is standing without a person"}]} {"id": 204420, "image": "COCO_train2014_000000204420.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a yellow bicycle standing in the grass\"."}], "task": "refering", "answer_template": "The \"a yellow bicycle standing in the grass\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [79, 97, 114, 115, 116, 117, 132, 133, 134, 135, 150, 151, 152, 153], "bbox": [0.35246, 0.38323353293413176, 0.52264, 0.7546407185628743], "iscrowd": 0, "area": 4155.8511, "rle": {"size": [334, 500], "counts": "aci11W:6VFJZ98fFIX98hFIV98jFIS99mFHQ99nFIP98PGIn88RGIk89UGHi89WGIf88ZGId89[GHd88\\GIc87]GJb86^GJb86^GJb86^GKa85_GKa85^GLb84^GMa83_GMa83_GMa83_GM`84`GM_83aGM_83aGM_83aGN^82bGN_7]OTIe0]ON]7ATIa0_OM]7U1bHlNS7_1mHaNo6c1QI]Nk6g1UIYNg6k1YITNg6m1YISNg6m1YISNf6n1ZIRNf6n1ZIRNe6LVI_15eNj6FTIc12fNQ7_O]I]1BTOc83]GMk8KTG6l8JTG6Q8_O_H;@5Q8E\\H5C5R8GZH4D3T8IWH5E1U8KjG?1EV8k0kGSOW8m0iGROX8n0hGTOV8l0jGVOT8j0lGYOQ8EmGd01JP8BnGe02Lm7^OQHf03Nk7[OQHh040h7XOSHi051f7UOUHk052d7SOVHl061e7QOVHn051f7POUHo051k7NVH2j7NVH3j7LVH4j7LVH4k7KTH6l7JTH5n7JRH5P8IQH6Q8IoG6S8ImG5V8JjG5X8JhG3[8MeGO`80`GKj8OWG0o8KQG5T9FnF7W9EjF:f9O2Mdg]2"}, "label": "a yellow bicycle standing in the grass"}]} {"id": 581256, "image": "COCO_train2014_000000581256.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black chair with a stuffed animal on it\"."}], "task": "refering", "answer_template": "The \"black chair with a stuffed animal on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 68, 69, 70, 71, 73, 74, 85, 86, 87, 88, 89, 90, 91, 102, 103, 104, 105, 106, 107, 108, 109, 120, 121, 122, 123, 124, 125, 137, 140, 141, 158], "bbox": [0.009083333333333334, 0.129109375, 0.4182916666666667, 0.427109375], "iscrowd": 0, "area": 16843.481499999998, "rle": {"size": [640, 480], "counts": "\\W33mc02M3N2M3M3N2M3N2M2N3N2M3N2M3c]OlNja0V1T^OmNha0V1V^OlNha0V1V^OmNga0U1W^OnNea0U1Y^OmNea0T1Z^OoNca0S1[^OQO`a0P2H9F9H8H9F9H8H9F9H8G10N2N101N101L4F9G:I70O100O2O000O10000O100O10000O101O0O100002N1O1O2N1O1O2N1O2O0O1O2N1O1O2N1O2N1O1O2N100O2N1O2N1O1O2_M__O_1b`0VNh_Oj1Wa0000000001O00000000000001O00000001O00000000000000000000001O00000001O0003M3MkM_^Om1_a0oMe^OT2aa04M3L4M3L4M3L4L4M3L4MfNa_OVO[`0j0n_OQOo?n0Y@jNg?V1a@aN_>AbAm18YNW>M]Aj1T`0XNj_Og1W`0\\Ne_Od1\\`0^Na_Ob1``0aN]_O^1d`0dNY_O\\1h`0fNV_OY1k`0jNQ_OV1Pa0lNn^OS1Sa0POi^OP1Xa0k01O1O1O1O1O1O1O1O1O1O1O1O1O1O1O11O5LRN^^Oa1]a0]Nh^Oe1Ta0XNR_Oi1j`0TN[_On1a`0PNc_Om1]`0QNh_Oj1Za0L2O1N2M3@a0N1O1N2O1N3N1O1N2O1N8I9GaV^5"}, "label": "black chair with a stuffed animal on it"}]} {"id": 581256, "image": "COCO_train2014_000000581256.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair with the stuffed animal owl sitting in it\"."}], "task": "refering", "answer_template": "The \"the chair with the stuffed animal owl sitting in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [51, 52, 53, 68, 69, 70, 71, 73, 74, 85, 86, 87, 88, 89, 90, 91, 102, 103, 104, 105, 106, 107, 108, 109, 120, 121, 122, 123, 124, 125, 137, 140, 141, 158], "bbox": [0.009083333333333334, 0.129109375, 0.4182916666666667, 0.427109375], "iscrowd": 0, "area": 16843.481499999998, "rle": {"size": [640, 480], "counts": "\\W33mc02M3N2M3M3N2M3N2M2N3N2M3N2M3c]OlNja0V1T^OmNha0V1V^OlNha0V1V^OmNga0U1W^OnNea0U1Y^OmNea0T1Z^OoNca0S1[^OQO`a0P2H9F9H8H9F9H8H9F9H8G10N2N101N101L4F9G:I70O100O2O000O10000O100O10000O101O0O100002N1O1O2N1O1O2N1O2O0O1O2N1O1O2N1O2N1O1O2N100O2N1O2N1O1O2_M__O_1b`0VNh_Oj1Wa0000000001O00000000000001O00000001O00000000000000000000001O00000001O0003M3MkM_^Om1_a0oMe^OT2aa04M3L4M3L4M3L4L4M3L4MfNa_OVO[`0j0n_OQOo?n0Y@jNg?V1a@aN_>AbAm18YNW>M]Aj1T`0XNj_Og1W`0\\Ne_Od1\\`0^Na_Ob1``0aN]_O^1d`0dNY_O\\1h`0fNV_OY1k`0jNQ_OV1Pa0lNn^OS1Sa0POi^OP1Xa0k01O1O1O1O1O1O1O1O1O1O1O1O1O1O1O11O5LRN^^Oa1]a0]Nh^Oe1Ta0XNR_Oi1j`0TN[_On1a`0PNc_Om1]`0QNh_Oj1Za0L2O1N2M3@a0N1O1N2O1N3N1O1N2O1N8I9GaV^5"}, "label": "the chair with the stuffed animal owl sitting in it"}]} {"id": 229002, "image": "COCO_train2014_000000229002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"green broccoli on a plate of food\"."}], "task": "refering", "answer_template": "The \"green broccoli on a plate of food\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [123, 124, 125, 140, 141, 142, 143, 158, 159, 160, 161, 176, 177, 178, 179, 194, 195, 196, 197, 211, 212, 213, 214, 215, 229, 230, 231, 232, 233], "bbox": [0.73588, 0.5011733333333334, 0.9988199999999999, 0.9882400000000001], "iscrowd": 0, "area": 16490.830049999997, "rle": {"size": [375, 500], "counts": "aSW41e;3M2O1N2N3M2N2N3M2O1N2O2M2O0O10O0100O100O100O100O1O1O100O1O1O1O1O1O1fLhNTLY1j3mNRLS1m3ROnKo0P4WOkKj0T4d0^J]O`5h0\\JYOc5h0\\JXOc5j0\\JWOb5l0]JTOb5m0XJYOf5j0PJ_Oo5b0gIHW6;_IN`6e2N2O1N2O100O1O100O1O100O1O10000O100000O10000000000000O100000000000000000000O10000000000001N10001O00001O00001O001O00001O00001O00001O001O00001N10001O001O3M4L4J6D;E in this image.", "objects": [{"patches": [123, 124, 125, 140, 141, 142, 143, 158, 159, 160, 161, 176, 177, 178, 179, 194, 195, 196, 197, 211, 212, 213, 214, 215, 229, 230, 231, 232, 233], "bbox": [0.73588, 0.5011733333333334, 0.9988199999999999, 0.9882400000000001], "iscrowd": 0, "area": 16490.830049999997, "rle": {"size": [375, 500], "counts": "aSW41e;3M2O1N2N3M2N2N3M2O1N2O2M2O0O10O0100O100O100O100O1O1O100O1O1O1O1O1O1fLhNTLY1j3mNRLS1m3ROnKo0P4WOkKj0T4d0^J]O`5h0\\JYOc5h0\\JXOc5j0\\JWOb5l0]JTOb5m0XJYOf5j0PJ_Oo5b0gIHW6;_IN`6e2N2O1N2O100O1O100O1O100O1O10000O100000O10000000000000O100000000000000000000O10000000000001N10001O00001O00001O001O00001O00001O00001O001O00001N10001O001O3M4L4J6D;E in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 18, 19, 20, 21, 22, 23, 24, 36, 37, 38, 39, 40, 41, 42, 57, 58], "bbox": [0.010199999999999999, 0.0, 0.40046, 0.26576], "iscrowd": 0, "area": 14462.306400000001, "rle": {"size": [375, 500], "counts": "Yl17X;:M3L4L3N3L4L4M2M3N101N2O1N2O1N2O1N2O0O2O0O100O10000O1000O0100[OXN[Gh1e8ZNYGf1g8\\NVGe1j8^NSGb1m8`NQG`1n8cNPG]1P9eNmF\\1S9d000000O100000000O100000000O100000000O100000000O1O1000000001O000000001O0000001O1O1O1O1O1O001O00001O0O101O001O00001O00001O00001O001O1O1O1O1O1O0000O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O10000O100O10000O100O100O10000O100O10000O100O10000O100O1O1N2M3N2N2M3N2I8E:H8M3M3Khk]3"}, "label": "a rose cup with a handle containing some black tea"}]} {"id": 229002, "image": "COCO_train2014_000000229002.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a pink coffee cup with someting in it\"."}], "task": "refering", "answer_template": "The \"a pink coffee cup with someting in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [1, 2, 3, 4, 5, 6, 7, 18, 19, 20, 21, 22, 23, 24, 36, 37, 38, 39, 40, 41, 42, 57, 58], "bbox": [0.010199999999999999, 0.0, 0.40046, 0.26576], "iscrowd": 0, "area": 14462.306400000001, "rle": {"size": [375, 500], "counts": "Yl17X;:M3L4L3N3L4L4M2M3N101N2O1N2O1N2O1N2O0O2O0O100O10000O1000O0100[OXN[Gh1e8ZNYGf1g8\\NVGe1j8^NSGb1m8`NQG`1n8cNPG]1P9eNmF\\1S9d000000O100000000O100000000O100000000O100000000O1O1000000001O000000001O0000001O1O1O1O1O1O001O00001O0O101O001O00001O00001O00001O001O1O1O1O1O1O0000O100O100O100O100O100O100O100O100O100O100O100O100O100O10000O100O100O100O100O100O100O100O10000O100O10000O100O100O10000O100O10000O100O10000O100O1O1N2M3N2N2M3N2I8E:H8M3M3Khk]3"}, "label": "a pink coffee cup with someting in it"}]} {"id": 97936, "image": "COCO_train2014_000000097936.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the woman holding the buckle of a dog\"."}], "task": "refering", "answer_template": "The \"the woman holding the buckle of a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 34, 35, 57, 58, 80, 81, 102, 103, 104, 124, 125, 126, 147, 148, 149, 196, 217, 218, 219, 220, 240, 243, 262, 263, 265, 288], "bbox": [0.397203125, 0.03166666666666667, 0.5891093749999999, 0.8138732394366198], "iscrowd": 0, "area": 6285.413950000002, "rle": {"size": [426, 640], "counts": "YoY33V=2O0WK0XL0g35ULLi36VLJj36VLKh36XLJg37YLJe37[LId38[LId38\\LIc36^LJa37_LJ_37aLI_37aLJ^36bLJ^35cLL[35eLK[34fLLZ34fLMY32hLNX32hLOW31iLOV31kL0T30lL0T3OmL1R30nL1Q3NPM2o2OQM2n2MRM4m2MSM4l2LTM4k2LVM4a2]OQI?^44`2_OPI=a44_2@oH<_47a2^OoH;]4;c2\\OoH;Y4OhKGP71mH=U4JSLGj63nH`0n3E^LHf6Q1fLYOfLEc6W1`LVOoLBa6[1ZLUOWM_O_6_1SLUO_M\\O^6c1lKTOgMXO]6g1fKTOnMTO\\6j1aKTOTNRO[6i1_KXOWNnNZ6h1^K]O[5a0dJB\\5c0F[O8e0KnKNb06_30kKKh02]36iKIj0O]3]GDe8:[GDj8:VGZOW9d0k1O010O010O010O010O0M4L3M4K4MaS]3"}, "label": "the woman holding the buckle of a dog"}]} {"id": 97936, "image": "COCO_train2014_000000097936.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman in a grey bucket hat walking a dog\"."}], "task": "refering", "answer_template": "The \"woman in a grey bucket hat walking a dog\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 34, 35, 57, 58, 80, 81, 102, 103, 104, 124, 125, 126, 147, 148, 149, 196, 217, 218, 219, 220, 240, 243, 262, 263, 265, 288], "bbox": [0.397203125, 0.03166666666666667, 0.5891093749999999, 0.8138732394366198], "iscrowd": 0, "area": 6285.413950000002, "rle": {"size": [426, 640], "counts": "YoY33V=2O0WK0XL0g35ULLi36VLJj36VLKh36XLJg37YLJe37[LId38[LId38\\LIc36^LJa37_LJ_37aLI_37aLJ^36bLJ^35cLL[35eLK[34fLLZ34fLMY32hLNX32hLOW31iLOV31kL0T30lL0T3OmL1R30nL1Q3NPM2o2OQM2n2MRM4m2MSM4l2LTM4k2LVM4a2]OQI?^44`2_OPI=a44_2@oH<_47a2^OoH;]4;c2\\OoH;Y4OhKGP71mH=U4JSLGj63nH`0n3E^LHf6Q1fLYOfLEc6W1`LVOoLBa6[1ZLUOWM_O_6_1SLUO_M\\O^6c1lKTOgMXO]6g1fKTOnMTO\\6j1aKTOTNRO[6i1_KXOWNnNZ6h1^K]O[5a0dJB\\5c0F[O8e0KnKNb06_30kKKh02]36iKIj0O]3]GDe8:[GDj8:VGZOW9d0k1O010O010O010O010O0M4L3M4K4MaS]3"}, "label": "woman in a grey bucket hat walking a dog"}]} {"id": 97936, "image": "COCO_train2014_000000097936.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman wearing white gloves\"."}], "task": "refering", "answer_template": "The \"a woman wearing white gloves\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 58, 59, 81, 82, 83, 104, 105, 106, 126, 127, 128, 129, 149, 150, 151, 152, 172, 173, 174, 175, 194, 195, 196, 197, 198, 220, 221, 243, 244], "bbox": [0.471171875, 0.06929577464788732, 0.62971875, 0.7277230046948356], "iscrowd": 0, "area": 12591.768349999997, "rle": {"size": [426, 640], "counts": "kkm35Q=4L4L4M3L4L4M3L4L4L4M3L4L4M3L4M3N2N2N2N2N2N2N2NkNYEKe:^OTFb0j9@VF`0h9BXF=f9F[F9a9K_F5\\90dF0W95jFJR9:nFEn8`0RG@i8e0WG[Oe8i0\\GVOS6ZOgJo1VOfNS6O[IT2b0mMR6^3nI_LT6b3mIZLU6g3kIULY6k3gIRL[6o3fIlK]6U4cIhK_6Y4aIcKc6]4]I`Ke6a4Q11O100O1O10001O1O1O1O1jGYKg7h4WHaKa7_4^HkKY7V4fHlK8YO`6l4VIlK9]O]6h4XIlK;_OZ6b5fIbJV6_5iIdJU6[5kIiJQ6W5oImJm5S5SJPKj5P5VJTKf5l4ZJWKc5i4]JWKc5i4]JXKb5h4^JXKb5h4]JZKb5f4^JZKb5FZJm34^La5SOPK\\4_OaLk6[3UIfL8mNn5Z4iIiL8YOf5l3RJmL5]Of5d3UJPM3]Oj5a3SJSM1]On5^3QJWMN\\OR6\\3PJYMM[OU6Z3nI\\MKZOZ6X3mI^MF[O_6m1bI^NgI\\N:Q1R6`0iI\\N8P1Q6a0kI]N7n0P6b0PJZN3P1U6?_K]Oh4;]KAk46ZKEn43^KBi47aK_Of49fK[Ob4=]4H9Hl]R3"}, "label": "a woman wearing white gloves"}]} {"id": 97936, "image": "COCO_train2014_000000097936.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two lovely ladies conversing while walking a dog , behind a bicycle\"."}], "task": "refering", "answer_template": "The \"two lovely ladies conversing while walking a dog , behind a bicycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [35, 36, 58, 59, 81, 82, 83, 104, 105, 106, 126, 127, 128, 129, 149, 150, 151, 152, 172, 173, 174, 175, 194, 195, 196, 197, 198, 220, 221, 243, 244], "bbox": [0.471171875, 0.06929577464788732, 0.62971875, 0.7277230046948356], "iscrowd": 0, "area": 12591.768349999997, "rle": {"size": [426, 640], "counts": "kkm35Q=4L4L4M3L4L4M3L4L4L4M3L4L4M3L4M3N2N2N2N2N2N2N2NkNYEKe:^OTFb0j9@VF`0h9BXF=f9F[F9a9K_F5\\90dF0W95jFJR9:nFEn8`0RG@i8e0WG[Oe8i0\\GVOS6ZOgJo1VOfNS6O[IT2b0mMR6^3nI_LT6b3mIZLU6g3kIULY6k3gIRL[6o3fIlK]6U4cIhK_6Y4aIcKc6]4]I`Ke6a4Q11O100O1O10001O1O1O1O1jGYKg7h4WHaKa7_4^HkKY7V4fHlK8YO`6l4VIlK9]O]6h4XIlK;_OZ6b5fIbJV6_5iIdJU6[5kIiJQ6W5oImJm5S5SJPKj5P5VJTKf5l4ZJWKc5i4]JWKc5i4]JXKb5h4^JXKb5h4]JZKb5f4^JZKb5FZJm34^La5SOPK\\4_OaLk6[3UIfL8mNn5Z4iIiL8YOf5l3RJmL5]Of5d3UJPM3]Oj5a3SJSM1]On5^3QJWMN\\OR6\\3PJYMM[OU6Z3nI\\MKZOZ6X3mI^MF[O_6m1bI^NgI\\N:Q1R6`0iI\\N8P1Q6a0kI]N7n0P6b0PJZN3P1U6?_K]Oh4;]KAk46ZKEn43^KBi47aK_Of49fK[Ob4=]4H9Hl]R3"}, "label": "two lovely ladies conversing while walking a dog , behind a bicycle"}]} {"id": 212635, "image": "COCO_train2014_000000212635.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"brown chair left side\"."}], "task": "refering", "answer_template": "The \"brown chair left side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [276, 277, 278, 299, 300, 301, 302, 322, 323, 324, 325, 326, 345, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372, 373], "bbox": [0.000609375, 0.7047708333333333, 0.23459375, 1.0], "iscrowd": 0, "area": 16500.9785, "rle": {"size": [480, 640], "counts": "f:Z4f:000000000O100000000000000O100000000000000O100000000000000O1000000000000000000000000001O0000000000000000001O000000000000000000000000000000001O0000001O1O1O3M3M3M3M2N1O2N2N3M2N3M2N3QMZD^2j;]M\\D^2U in this image.", "objects": [{"patches": [276, 277, 278, 299, 300, 301, 302, 322, 323, 324, 325, 326, 345, 346, 347, 348, 349, 350, 368, 369, 370, 371, 372, 373], "bbox": [0.000609375, 0.7047708333333333, 0.23459375, 1.0], "iscrowd": 0, "area": 16500.9785, "rle": {"size": [480, 640], "counts": "f:Z4f:000000000O100000000000000O100000000000000O100000000000000O1000000000000000000000000001O0000000000000000001O000000000000000000000000000000001O0000001O1O1O3M3M3M3M2N1O2N2N3M2N3M2N3QMZD^2j;]M\\D^2U in this image.", "objects": [{"patches": [259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.251015625, 0.6540625, 1.0, 1.0], "iscrowd": 0, "area": 72847.11490000002, "rle": {"size": [480, 640], "counts": "Tj[24j>6I6cABP>l0J5K5K5J4M3M2N200O101O0QCWNg in this image.", "objects": [{"patches": [259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.251015625, 0.6540625, 1.0, 1.0], "iscrowd": 0, "area": 72847.11490000002, "rle": {"size": [480, 640], "counts": "Tj[24j>6I6cABP>l0J5K5K5J4M3M2N200O101O0QCWNg in this image.", "objects": [{"patches": [102, 103, 125, 126, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 217, 218, 219, 240, 241, 242, 263, 264, 265, 286, 287, 288, 289, 309, 310, 311, 312, 332, 333, 334, 335], "bbox": [0.412640625, 0.2538125, 0.587140625, 0.8802083333333334], "iscrowd": 0, "area": 14836.651800000001, "rle": {"size": [480, 640], "counts": "ank39f>>B2N2N3M2N3M2M4M1O2M2N3L3N3UCgNk;]2J7I6KRN[Db0c;BXEBe:c0]EXOa:l0cEoNZ:U1iEfNU:^1oE]Nm1Id4n1cISNf11g4P2eIkMa17h4^2[LcMd3R3gKoLX4h3QKYLn4V4cJkK]5W4`JkK^4CYJe4V1iK_4GWJc4X1fK^4MWJ^4`1_KV49VJ\\4g1VKo3d0VJb4k7cKSHi4a7ZK[Hl4i0dJS5g0mIk4j0_JY5Q7`JoH`5W7ZJjHf5[7TJeHl5a7oI^HR6X8O010O01N1ZMjIcLW6W3UJbLl5X3`JaLb5X3jJaLW5_3PKZLQ5g3SKRLn4P4VKjKk4W4[K`Kg4V2cInNm1cN`4a2jIfNl1_N\\4m2oI\\Nl1^NU4[3SJPNa8U2cGcM[8HlFg1m0ZNX8ERGV2i0oMV8FSG`2j0dMS8HVGi2h0[MR8GXGS3i0PMP8IYG[3j0fLm7K[Gd3j0\\L^8m3W19F:G9F;F6J1N2O2M2O1N2O1O1N2OXLPH`09`MS6P2YJMAUNV6l1`JKVO\\NZ6g1fJJlNbN]6d1mJFbNiNa6_1TKDXNoNf6Y1XKEnMUOl6S1XKHhMXOS7k0YKM_M[O\\7b0XK5WM[Oe7;WK;nL]Oo72VKc1n4WNVKe1P5VNSKe1T5UNoJg1X5SNlJh1[5RNhJj1[9K4M4K5K5K5K6J5K5K5K5L4Km[k3"}, "label": "woman in the pink jacket"}]} {"id": 335524, "image": "COCO_train2014_000000335524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a red jacket\"."}], "task": "refering", "answer_template": "The \"a woman in a red jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 125, 126, 148, 149, 150, 170, 171, 172, 173, 193, 194, 195, 196, 217, 218, 219, 240, 241, 242, 263, 264, 265, 286, 287, 288, 289, 309, 310, 311, 312, 332, 333, 334, 335], "bbox": [0.412640625, 0.2538125, 0.587140625, 0.8802083333333334], "iscrowd": 0, "area": 14836.651800000001, "rle": {"size": [480, 640], "counts": "ank39f>>B2N2N3M2N3M2M4M1O2M2N3L3N3UCgNk;]2J7I6KRN[Db0c;BXEBe:c0]EXOa:l0cEoNZ:U1iEfNU:^1oE]Nm1Id4n1cISNf11g4P2eIkMa17h4^2[LcMd3R3gKoLX4h3QKYLn4V4cJkK]5W4`JkK^4CYJe4V1iK_4GWJc4X1fK^4MWJ^4`1_KV49VJ\\4g1VKo3d0VJb4k7cKSHi4a7ZK[Hl4i0dJS5g0mIk4j0_JY5Q7`JoH`5W7ZJjHf5[7TJeHl5a7oI^HR6X8O010O01N1ZMjIcLW6W3UJbLl5X3`JaLb5X3jJaLW5_3PKZLQ5g3SKRLn4P4VKjKk4W4[K`Kg4V2cInNm1cN`4a2jIfNl1_N\\4m2oI\\Nl1^NU4[3SJPNa8U2cGcM[8HlFg1m0ZNX8ERGV2i0oMV8FSG`2j0dMS8HVGi2h0[MR8GXGS3i0PMP8IYG[3j0fLm7K[Gd3j0\\L^8m3W19F:G9F;F6J1N2O2M2O1N2O1O1N2OXLPH`09`MS6P2YJMAUNV6l1`JKVO\\NZ6g1fJJlNbN]6d1mJFbNiNa6_1TKDXNoNf6Y1XKEnMUOl6S1XKHhMXOS7k0YKM_M[O\\7b0XK5WM[Oe7;WK;nL]Oo72VKc1n4WNVKe1P5VNSKe1T5UNoJg1X5SNlJh1[5RNhJj1[9K4M4K5K5K5K6J5K5K5K5L4Km[k3"}, "label": "a woman in a red jacket"}]} {"id": 335524, "image": "COCO_train2014_000000335524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in awhite jacket\"."}], "task": "refering", "answer_template": "The \"a woman in awhite jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 107, 108, 129, 130, 131, 132, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 221, 222, 223, 224, 244, 245, 246, 247, 268, 269, 270, 291, 292, 293, 314, 315, 316, 337, 338, 339, 362], "bbox": [0.613484375, 0.20675, 0.8612343750000001, 0.894375], "iscrowd": 0, "area": 20290.007999999994, "rle": {"size": [480, 640], "counts": "a]h5:b>:F7I4M3L4L5K4L4L4M4K4gCRN_;o1\\DYNa;g1YDbNc;_1XDjNc;X1VDQOg;W2L5L3L4L5L3L4L4M4K4L4M4K4L4iHSKU4Q5eKSKX4Q5dKRKY4T5`KPK^4W5ZKlJc4]5TKfJi3EUKl5i0cJo3f6iK]IT4k6cKYIY4Q7^KRI_4V7YKjHg4^7kJgHU5b800O1000O010000O0101ULmJhMT5\\2TLWLm3m3fLZK\\3j4WM^Jk2e5[MPJf2T6_MbIb2a6m24L4L4L4L4L3M3M3M3M2N3MRIYIf5i6eIiI\\6^71O0O2VI]IQ5f6hJ_IX5c6bJaI]5b6\\JcIc5`6VJdIk5]6PJgIo5\\6jIiIU6Z6eIjIZ6_700O1O10iJ\\JQ1c5dNcKb0\\4SO_L=b3WOmLc0U3ROZMi0g2lNhMn0Z2hNSNT1o1mNSNm0o1TOTNg0m1ZOVN`0n1_OUN in this image.", "objects": [{"patches": [84, 85, 107, 108, 129, 130, 131, 132, 152, 153, 154, 155, 156, 157, 175, 176, 177, 178, 179, 180, 198, 199, 200, 201, 221, 222, 223, 224, 244, 245, 246, 247, 268, 269, 270, 291, 292, 293, 314, 315, 316, 337, 338, 339, 362], "bbox": [0.613484375, 0.20675, 0.8612343750000001, 0.894375], "iscrowd": 0, "area": 20290.007999999994, "rle": {"size": [480, 640], "counts": "a]h5:b>:F7I4M3L4L5K4L4L4M4K4gCRN_;o1\\DYNa;g1YDbNc;_1XDjNc;X1VDQOg;W2L5L3L4L5L3L4L4M4K4L4M4K4L4iHSKU4Q5eKSKX4Q5dKRKY4T5`KPK^4W5ZKlJc4]5TKfJi3EUKl5i0cJo3f6iK]IT4k6cKYIY4Q7^KRI_4V7YKjHg4^7kJgHU5b800O1000O010000O0101ULmJhMT5\\2TLWLm3m3fLZK\\3j4WM^Jk2e5[MPJf2T6_MbIb2a6m24L4L4L4L4L3M3M3M3M2N3MRIYIf5i6eIiI\\6^71O0O2VI]IQ5f6hJ_IX5c6bJaI]5b6\\JcIc5`6VJdIk5]6PJgIo5\\6jIiIU6Z6eIjIZ6_700O1O10iJ\\JQ1c5dNcKb0\\4SO_L=b3WOmLc0U3ROZMi0g2lNhMn0Z2hNSNT1o1mNSNm0o1TOTNg0m1ZOVN`0n1_OUN in this image.", "objects": [{"patches": [331, 332, 334, 335, 352, 353, 354, 355, 356, 357, 373, 374, 375, 376, 377, 378], "bbox": [0.25165625, 0.8241875000000001, 0.608484375, 0.9941041666666667], "iscrowd": 0, "area": 2921.2129499999983, "rle": {"size": [480, 640], "counts": "Qm[25k>2N1O2O0O2N101N1O1000O2N1N3N1O1N2O00O010O10O0100O010O10O0100O010O10O0100O00100O010O10O0100O010O10O0100O010O010O010O010O010O0010O010O010O010OPBF[=;dBHZ=7fBLW=5iBMU=2lB0Q=1PC0n in this image.", "objects": [{"patches": [331, 332, 334, 335, 352, 353, 354, 355, 356, 357, 373, 374, 375, 376, 377, 378], "bbox": [0.25165625, 0.8241875000000001, 0.608484375, 0.9941041666666667], "iscrowd": 0, "area": 2921.2129499999983, "rle": {"size": [480, 640], "counts": "Qm[25k>2N1O2O0O2N101N1O1000O2N1N3N1O1N2O00O010O10O0100O010O10O0100O010O10O0100O00100O010O10O0100O010O10O0100O010O010O010O010O010O0010O010O010O010OPBF[=;dBHZ=7fBLW=5iBMU=2lB0Q=1PC0n in this image.", "objects": [{"patches": [98, 99, 120, 121, 122, 143, 144, 145, 146, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 212, 213, 214, 235, 236, 237, 259, 260, 261, 282, 283, 284, 304, 305, 306, 307], "bbox": [0.194125, 0.2535, 0.38690625, 0.8290833333333333], "iscrowd": 0, "area": 17866.48755000001, "rle": {"size": [480, 640], "counts": "P[j18f>5K6K5J5K1O2N2M3M2N3L4M2N3L4N101N2O1O0O2O1O001N2O00>B4L4K5L4L3M4L4nGfLW4^3dKeLY4a3aKaL^4e3\\K\\Lc4j3VKXLi4m3QKULo4P4jJRLU5X4`JjK`5`4TJbKk5h4jIYKP6T5hInJU6Z5dIgJZ6a5_IaJ^6h5ZIYJe6n5UISJj6S6oHnIQ7Q7N101O0O2O001O0010O01O001O0100O1fM`ITL`6\\3\\JWLf5n3_JgKa5^4oJlJR5Z5W25K5K6K4K6SI_Io4i6eJ[I\\5l6XJYIh5n6lIVIV6o6_IVIa6d72N3N1N3N2M2N3N8G2O0O101N1O2O0N2N3PLlHNV7CcI1`6LkIJW64TJ@o5>\\JWOf5g0dJnN_5P1oJaNT5]1aKnMa4P2Y4N1N3N2N2N2N2N2N2^NhBY1Z=fNiBU1Y=kNjBQ1X=nNlBm0V=ROmBj0U=UOoBf0i=L5K4K5LTPh5"}, "label": "a skier in yellow pants"}]} {"id": 335524, "image": "COCO_train2014_000000335524.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a women with green jacket\"."}], "task": "refering", "answer_template": "The \"a women with green jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [98, 99, 120, 121, 122, 143, 144, 145, 146, 165, 166, 167, 168, 169, 188, 189, 190, 191, 192, 212, 213, 214, 235, 236, 237, 259, 260, 261, 282, 283, 284, 304, 305, 306, 307], "bbox": [0.194125, 0.2535, 0.38690625, 0.8290833333333333], "iscrowd": 0, "area": 17866.48755000001, "rle": {"size": [480, 640], "counts": "P[j18f>5K6K5J5K1O2N2M3M2N3L4M2N3L4N101N2O1O0O2O1O001N2O00>B4L4K5L4L3M4L4nGfLW4^3dKeLY4a3aKaL^4e3\\K\\Lc4j3VKXLi4m3QKULo4P4jJRLU5X4`JjK`5`4TJbKk5h4jIYKP6T5hInJU6Z5dIgJZ6a5_IaJ^6h5ZIYJe6n5UISJj6S6oHnIQ7Q7N101O0O2O001O0010O01O001O0100O1fM`ITL`6\\3\\JWLf5n3_JgKa5^4oJlJR5Z5W25K5K6K4K6SI_Io4i6eJ[I\\5l6XJYIh5n6lIVIV6o6_IVIa6d72N3N1N3N2M2N3N8G2O0O101N1O2O0N2N3PLlHNV7CcI1`6LkIJW64TJ@o5>\\JWOf5g0dJnN_5P1oJaNT5]1aKnMa4P2Y4N1N3N2N2N2N2N2N2^NhBY1Z=fNiBU1Y=kNjBQ1X=nNlBm0V=ROmBj0U=UOoBf0i=L5K4K5LTPh5"}, "label": "a women with green jacket"}]} {"id": 261800, "image": "COCO_train2014_000000261800.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red umbrella opened for protection during a rain storm\"."}], "task": "refering", "answer_template": "The \"a red umbrella opened for protection during a rain storm\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 49, 50, 51, 52, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147], "bbox": [0.0022549019607843133, 0.06573529411764706, 0.6966339869281045, 0.319656862745098], "iscrowd": 0, "area": 36762.11134999999, "rle": {"size": [612, 612], "counts": "_[12Qc01O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2O0O2N1O2N1O1O2N1O2N1O2N1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2O000O10000O2O0O10000O101O0O100O101O0O10000O101N10000O10001N100O10000O10000O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O10000O100000000O10000000000000O01000000000000O100000000000000O100000000000000O1000000000000O100000000000000O10000N2M3N2N2M3N2M3N2N2M3N2N21O2N2N2N2N2N3M2N2N2N2N3M2N2N2N1O1O00001O00001O0000001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O000010O000001O00001O00001O00001O00001O00001O0000001O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O101N1O1OVo^3"}, "label": "a red umbrella opened for protection during a rain storm"}]} {"id": 261800, "image": "COCO_train2014_000000261800.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red umbrella\"."}], "task": "refering", "answer_template": "The \"a red umbrella\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 49, 50, 51, 52, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147], "bbox": [0.0022549019607843133, 0.06573529411764706, 0.6966339869281045, 0.319656862745098], "iscrowd": 0, "area": 36762.11134999999, "rle": {"size": [612, 612], "counts": "_[12Qc01O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2N1O2N1O2N1O1O2O0O2N1O2N1O1O2N1O2N1O2N1O1O2N1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2N1O1O1O1O1O2O000O10000O2O0O10000O101O0O100O101O0O10000O101N10000O10001N100O10000O10000O100O10000O10000O100O10000O10000O100O10000O10000O100O10000O100O10000O10000O100O10000O10000O100000000O10000000000000O01000000000000O100000000000000O100000000000000O1000000000000O100000000000000O10000N2M3N2N2M3N2M3N2N2M3N2N21O2N2N2N2N2N3M2N2N2N2N3M2N2N2N1O1O00001O00001O0000001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O0000001O00001O00001O00001O000010O000001O00001O00001O00001O00001O00001O0000001O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O100O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1O101N1O1OVo^3"}, "label": "a red umbrella"}]} {"id": 220842, "image": "COCO_train2014_000000220842.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"four burners on top of a large stove\"."}], "task": "refering", "answer_template": "The \"four burners on top of a large stove\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [290, 291, 292, 293, 294, 295, 296, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.479984375, 0.7342916666666667, 0.9493124999999999, 0.8788750000000001], "iscrowd": 0, "area": 12065.794950000005, "rle": {"size": [480, 640], "counts": "lX`43k>3O0O1O100O100O2O0O100O100O100O2O0O100O100O101N1O100O1O12N1O010O001O001O0001O01O0000001O01O0001O0000001O01N1000000000001O00000O100000001O000000000O101O0000000000001O000O10000O100O2O0O100O1O1O1O2N1O1N200O1O1O101O0000000000001O00000O1000001O000000000000001O0O100000000O2O01O0000000001O0000000001O0001O0000000000001N10000000000O1010O001O00001O010O001O00000001O000000000O10001N1O100O11O2N01O000000000000001O0000000000001O00000001O0001O000002N1O2N1O01O0000000000001O000000000001O01O00000O1O100010O000000000001O000003M00000001O0000000000001O0000000000001O000000000000001O00O1N2O1N2O1N2O1N2N2N3M4L4L5K7I8H8Hkc>"}, "label": "four burners on top of a large stove"}]} {"id": 220842, "image": "COCO_train2014_000000220842.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a stainless steel gas stove top with a pie\"."}], "task": "refering", "answer_template": "The \"a stainless steel gas stove top with a pie\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [290, 291, 292, 293, 294, 295, 296, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.479984375, 0.7342916666666667, 0.9493124999999999, 0.8788750000000001], "iscrowd": 0, "area": 12065.794950000005, "rle": {"size": [480, 640], "counts": "lX`43k>3O0O1O100O100O2O0O100O100O100O2O0O100O100O101N1O100O1O12N1O010O001O001O0001O01O0000001O01O0001O0000001O01N1000000000001O00000O100000001O000000000O101O0000000000001O000O10000O100O2O0O100O1O1O1O2N1O1N200O1O1O101O0000000000001O00000O1000001O000000000000001O0O100000000O2O01O0000000001O0000000001O0001O0000000000001N10000000000O1010O001O00001O010O001O00000001O000000000O10001N1O100O11O2N01O000000000000001O0000000000001O00000001O0001O000002N1O2N1O01O0000000000001O000000000001O01O00000O1O100010O000000000001O000003M00000001O0000000000001O0000000000001O000000000000001O00O1N2O1N2O1N2O1N2N2N3M4L4L5K7I8H8Hkc>"}, "label": "a stainless steel gas stove top with a pie"}]} {"id": 220842, "image": "COCO_train2014_000000220842.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the face of the oven\"."}], "task": "refering", "answer_template": "The \"the face of the oven\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [332, 333, 334, 335, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389], "bbox": [0.46115625, 0.8502291666666667, 0.9256718750000001, 0.989875], "iscrowd": 0, "area": 12867.970050000002, "rle": {"size": [480, 640], "counts": "bgZ49j=m0UOk0O100O100001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O0000000000001O000O1000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O0000000000001O00000000001O00000000001O00000000001O0000000000001O00000000002NhQf0"}, "label": "the face of the oven"}]} {"id": 73387, "image": "COCO_train2014_000000073387.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra with birds on its back\"."}], "task": "refering", "answer_template": "The \"zebra with birds on its back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [103, 106, 107, 108, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 226, 227, 228, 234, 235, 236, 239, 240, 241, 242, 243, 244, 245, 246, 247, 257, 258, 259, 262, 263, 264, 265, 266, 267, 268, 269, 280, 281, 285, 286, 287, 288, 291, 292, 303, 304, 308, 311, 314, 315, 331, 334, 337, 338], "bbox": [0.153046875, 0.31440281030444966, 0.9433906249999999, 0.9799063231850117], "iscrowd": 0, "area": 66534.56845, "rle": {"size": [427, 640], "counts": "fmY12Y=00001O001O00001`FLg54^JN[53eJ0W50fJ4Y5LdJ8[5K_J:_5H]J;b5GYJ>e5EUJ?j5BRJa0n5@oIb0P6@lIc0T6]OiIf0W6[OeIg0\\6ZO`Ii0`6WO^Ik0b6VOZIm0e6UOaGLf0R1i7RO^Gb03?_8POZGi2f8n000O10O100000O100M2N3M3N2N101N2O001N2O001NO2M210O1O00100O001O100O10000O100O10000O100O100000000000000O1O1O1O1L4H8\\Od000O1O1O100O2O1O100O1O1O100O2N1O100O1O1O100000000O10O100O10O10O10000O10O10O10000O01000O100O01000O10O10O1000O010000O010O10000O2O000O10000O10c0]O2N3M3M2cGkK_7Y4\\HXLU7k3eHfLm6]3nHTMc6n2XIcMZ6T5C7I8H7I1O001O1O0O2O1O00M3D in this image.", "objects": [{"patches": [103, 106, 107, 108, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 226, 227, 228, 234, 235, 236, 239, 240, 241, 242, 243, 244, 245, 246, 247, 257, 258, 259, 262, 263, 264, 265, 266, 267, 268, 269, 280, 281, 285, 286, 287, 288, 291, 292, 303, 304, 308, 311, 314, 315, 331, 334, 337, 338], "bbox": [0.153046875, 0.31440281030444966, 0.9433906249999999, 0.9799063231850117], "iscrowd": 0, "area": 66534.56845, "rle": {"size": [427, 640], "counts": "fmY12Y=00001O001O00001`FLg54^JN[53eJ0W50fJ4Y5LdJ8[5K_J:_5H]J;b5GYJ>e5EUJ?j5BRJa0n5@oIb0P6@lIc0T6]OiIf0W6[OeIg0\\6ZO`Ii0`6WO^Ik0b6VOZIm0e6UOaGLf0R1i7RO^Gb03?_8POZGi2f8n000O10O100000O100M2N3M3N2N101N2O001N2O001NO2M210O1O00100O001O100O10000O100O10000O100O100000000000000O1O1O1O1L4H8\\Od000O1O1O100O2O1O100O1O1O100O2N1O100O1O1O100000000O10O100O10O10O10000O10O10O10000O01000O100O01000O10O10O1000O010000O010O10000O2O000O10000O10c0]O2N3M3M2cGkK_7Y4\\HXLU7k3eHfLm6]3nHTMc6n2XIcMZ6T5C7I8H7I1O001O1O0O2O1O00M3D in this image.", "objects": [{"patches": [77, 80, 81, 82, 83, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 120, 121, 122, 123, 124, 125, 142, 143, 144, 164, 165, 166, 187, 188, 210, 211, 214, 215, 216, 233, 234, 238, 239, 257, 261, 262, 266, 267, 284, 285, 289, 290], "bbox": [0.1499375, 0.2532552693208431, 0.7121718749999999, 0.8712177985948477], "iscrowd": 0, "area": 16150.909450000005, "rle": {"size": [427, 640], "counts": "YXX11Y=1N2O1O1N2O1gCNX;3fD7P;KmD?k:BSEh0d:YOYER1]:POaEZ1V:gNgEc1P:_NnEi1k9XNVFm1c9TN^Fk1`9WNaFh1^9YNcFf1\\9[NeFd1Y9^NhFa1W9`NjF_1U9bNlF]1R9eNoFZ1P9gNQGX1n8iNSGV1k8lNVGS1i8nNXGQ1g8POZGYOAQ1S9H]GTOJl0h81_GoN4h0\\8:aGkNg1FYN9f1H[N8c1J^N5a1L_N3`1O`N1^11bNO]12cNM\\15dNK[16eNIZ19fNFY1gN@Y1b0gN\\OZ1e0fNYOZ1i0fNWOX1k0hNWOU1i0lNXOQ1h0QOYOl0g0VO[Og0d0[O]Ob0c0@^O>a0DA8?JB3>OCO<3EJ;8GH5:LF2;NFO<1EL=3DK>5CH?8AFa0:_ODc0<^OAd0?\\O_Of0a0[O\\Og0d0YOZOi0f0WOXOk0h0VOVOk0i0VOUOm0j0j501O001O001O001O001N101O001O001O001O001O001O001O001O001O0O2O001O001O001O001O001O001O001N1OUc\\2"}, "label": "the zebra behind another zebra"}]} {"id": 73387, "image": "COCO_train2014_000000073387.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"two zeebras having their food on the field\"."}], "task": "refering", "answer_template": "The \"two zeebras having their food on the field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 80, 81, 82, 83, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 120, 121, 122, 123, 124, 125, 142, 143, 144, 164, 165, 166, 187, 188, 210, 211, 214, 215, 216, 233, 234, 238, 239, 257, 261, 262, 266, 267, 284, 285, 289, 290], "bbox": [0.1499375, 0.2532552693208431, 0.7121718749999999, 0.8712177985948477], "iscrowd": 0, "area": 16150.909450000005, "rle": {"size": [427, 640], "counts": "YXX11Y=1N2O1O1N2O1gCNX;3fD7P;KmD?k:BSEh0d:YOYER1]:POaEZ1V:gNgEc1P:_NnEi1k9XNVFm1c9TN^Fk1`9WNaFh1^9YNcFf1\\9[NeFd1Y9^NhFa1W9`NjF_1U9bNlF]1R9eNoFZ1P9gNQGX1n8iNSGV1k8lNVGS1i8nNXGQ1g8POZGYOAQ1S9H]GTOJl0h81_GoN4h0\\8:aGkNg1FYN9f1H[N8c1J^N5a1L_N3`1O`N1^11bNO]12cNM\\15dNK[16eNIZ19fNFY1gN@Y1b0gN\\OZ1e0fNYOZ1i0fNWOX1k0hNWOU1i0lNXOQ1h0QOYOl0g0VO[Og0d0[O]Ob0c0@^O>a0DA8?JB3>OCO<3EJ;8GH5:LF2;NFO<1EL=3DK>5CH?8AFa0:_ODc0<^OAd0?\\O_Of0a0[O\\Og0d0YOZOi0f0WOXOk0h0VOVOk0i0VOUOm0j0j501O001O001O001O001N101O001O001O001O001O001O001O001O001O0O2O001O001O001O001O001O001O001N1OUc\\2"}, "label": "two zeebras having their food on the field"}]} {"id": 7852, "image": "COCO_train2014_000000007852.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the old school fire truck with a blue engine top\"."}], "task": "refering", "answer_template": "The \"the old school fire truck with a blue engine top\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [36, 37, 38, 56, 57, 58, 59, 60, 61, 76, 77, 78, 79, 80, 81, 82, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 257, 258, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 311, 312, 313, 314, 315, 335, 336, 337], "bbox": [0.17359375, 0.09625, 0.857859375, 0.8670416666666667], "iscrowd": 0, "area": 83824.80470000002, "rle": {"size": [480, 640], "counts": "ZYd13k>7H7I8H8H5jLmNXHX1c7kNZH[1`7hN\\H^1_7eN]Ha1`7`N\\Hf1b7ZNZHl1c7UNYHR2d7nMXHX2f7hMVH^2h7bMUHc2h7^MVHf2h7ZMXHf2h7ZMXHg2f7ZM[Hf2d7ZM\\Hf2d7ZM\\Hg2c7YM]Hg2b7ZM^Hg2a7YM_Hg2a7YM_Hh2`7XM`Hi2^7XMbHh2^7XMbHi2]7WMcHi2\\7XMdHi2[7WMeHj2Z7VMfHj2Y7WMgHj2X7UMiHk2W7UMiHl2U7UMkHk2U7UMkHk2T7UMmHk2S7UMmHk2S7UMmHj2S7VM_HY3a7gLjGn3V8RLhGo3X8RLfGP4Z8oKeGR4\\8nKaGU4^8lK`GV4`8jK^GW4b8iK]GY4c8gK[G[4e8eKYG\\4g8eKVG[4m8dKRGY4S9gKkFU4Z9lKdFS4_9g001N101O0O101N101O0O101O0O101O0O2O000O2O0O2O000O2O000O2O001N100O2O000O2O001N10001N101O0O101N10001N101O0O101O0O2O000O2O0O101O0O2O000O2O001N100O2O000O2O001N10001N101O0O101N10001N101O0O101O0O101O0O2O0O101O0O2O000O2O000O2O0O2O000O2O001N10001N10001N101N10001N101OO001\\HcIe6]6[IhI_6X6bImIY6S6gISJR6n5mI[Jk5d5VJhJ]5Y5cJSKQ5l4PK\\Kg4e4XK\\Kg4e4YK[Kg4d4ZK]Ke4c4[K]Kd4d4[K]Ke4b4\\K^Kc4c4]K]Kc4c4\\K_Kc4`4^K`Ka4a4_K_Ka4`4`K`K_4a4`K`K`4`4`KaK_4^4bKbK]4_4cKaK]4_4bKbK]4^4dKbK\\4^4dKcK[4]4eKcKZ4]4fKdKZ4\\4fKdKY4]4gKcKY4\\4hKeKW4[4hKfKW4Z4jKfKV4Z4jKfKV4Z4iKgKV4Y4kKhKT4X4lKhKS4Y4mKgKS4X4mKiKS4W4mKiKR4X4nKiKS4T4nKlKS4S4lKnKV4P4jKPLX4m3iKSLX4l3hKULY4h3gKYL[4e3eK[L]4c3cK]L^4a3cK_L_4_3`KbLb4\\3^KdLc4Z3^KfLd4X3\\KhLf4V3YKkLn4m2SKSMU5e2kJ[M]5\\2cJeMd5T2\\JlMi5o1WJQNi5n1XJRNi5m1VJUNi5k1WJUNj5i1WJWNi5i1WJWNi5i1VJXNk5f1VJZNj5f1VJZNk5e1UJ[Nk5d1UJ]Nl5b1TJ^Nl5a1UJ_Nl5`1TJ`Nl5`1SJaNn5]1SJcNm5]1SJcNn5\\1RJdNn5[1RJfNo5Y1QJgNo5Y1QJgNP6W1PJkNo5U1QJkNP6T1PJlNP6S1QJmNP6R1oIoNQ6P1PJPOQ6o0oIQOQ6o0oIQOR6m0nITOR6l0nITOR6l0nITOS6j0nIWOQ6i0nIXOS6g0mIYOS6f0nIZOS6e0mI[OS6e0lI\\OU6c0kI]OU6c0kI]OV6c0hI^OX6b0hI_OX6`0hI@X6`0hI@Y6?fIBZ6>fIB[6=eIC[6=eIC\\6o5CjId0U6T4O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1M3K5J6J7I6N2N2O1N2N2N3N11O1O001O1O10O01O1O00000010O00000001O01O0001O000000010O0000001O0001O00010O000010O0001O0O2O0O1O2O0O101N1O101N1O2O0O101N1O101mLoDT2R;iMSEU2m:hMWEU2l:gMXEV2j:gMZEV2i:fM[EW2g:fM^EW2c:fMaEW2f;WNZCT1i in this image.", "objects": [{"patches": [36, 37, 38, 56, 57, 58, 59, 60, 61, 76, 77, 78, 79, 80, 81, 82, 96, 97, 98, 99, 100, 101, 102, 119, 120, 121, 122, 123, 124, 125, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 257, 258, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 311, 312, 313, 314, 315, 335, 336, 337], "bbox": [0.17359375, 0.09625, 0.857859375, 0.8670416666666667], "iscrowd": 0, "area": 83824.80470000002, "rle": {"size": [480, 640], "counts": "ZYd13k>7H7I8H8H5jLmNXHX1c7kNZH[1`7hN\\H^1_7eN]Ha1`7`N\\Hf1b7ZNZHl1c7UNYHR2d7nMXHX2f7hMVH^2h7bMUHc2h7^MVHf2h7ZMXHf2h7ZMXHg2f7ZM[Hf2d7ZM\\Hf2d7ZM\\Hg2c7YM]Hg2b7ZM^Hg2a7YM_Hg2a7YM_Hh2`7XM`Hi2^7XMbHh2^7XMbHi2]7WMcHi2\\7XMdHi2[7WMeHj2Z7VMfHj2Y7WMgHj2X7UMiHk2W7UMiHl2U7UMkHk2U7UMkHk2T7UMmHk2S7UMmHk2S7UMmHj2S7VM_HY3a7gLjGn3V8RLhGo3X8RLfGP4Z8oKeGR4\\8nKaGU4^8lK`GV4`8jK^GW4b8iK]GY4c8gK[G[4e8eKYG\\4g8eKVG[4m8dKRGY4S9gKkFU4Z9lKdFS4_9g001N101O0O101N101O0O101O0O101O0O2O000O2O0O2O000O2O000O2O001N100O2O000O2O001N10001N101O0O101N10001N101O0O101O0O2O000O2O0O101O0O2O000O2O001N100O2O000O2O001N10001N101O0O101N10001N101O0O101O0O101O0O2O0O101O0O2O000O2O000O2O0O2O000O2O001N10001N10001N101N10001N101OO001\\HcIe6]6[IhI_6X6bImIY6S6gISJR6n5mI[Jk5d5VJhJ]5Y5cJSKQ5l4PK\\Kg4e4XK\\Kg4e4YK[Kg4d4ZK]Ke4c4[K]Kd4d4[K]Ke4b4\\K^Kc4c4]K]Kc4c4\\K_Kc4`4^K`Ka4a4_K_Ka4`4`K`K_4a4`K`K`4`4`KaK_4^4bKbK]4_4cKaK]4_4bKbK]4^4dKbK\\4^4dKcK[4]4eKcKZ4]4fKdKZ4\\4fKdKY4]4gKcKY4\\4hKeKW4[4hKfKW4Z4jKfKV4Z4jKfKV4Z4iKgKV4Y4kKhKT4X4lKhKS4Y4mKgKS4X4mKiKS4W4mKiKR4X4nKiKS4T4nKlKS4S4lKnKV4P4jKPLX4m3iKSLX4l3hKULY4h3gKYL[4e3eK[L]4c3cK]L^4a3cK_L_4_3`KbLb4\\3^KdLc4Z3^KfLd4X3\\KhLf4V3YKkLn4m2SKSMU5e2kJ[M]5\\2cJeMd5T2\\JlMi5o1WJQNi5n1XJRNi5m1VJUNi5k1WJUNj5i1WJWNi5i1WJWNi5i1VJXNk5f1VJZNj5f1VJZNk5e1UJ[Nk5d1UJ]Nl5b1TJ^Nl5a1UJ_Nl5`1TJ`Nl5`1SJaNn5]1SJcNm5]1SJcNn5\\1RJdNn5[1RJfNo5Y1QJgNo5Y1QJgNP6W1PJkNo5U1QJkNP6T1PJlNP6S1QJmNP6R1oIoNQ6P1PJPOQ6o0oIQOQ6o0oIQOR6m0nITOR6l0nITOR6l0nITOS6j0nIWOQ6i0nIXOS6g0mIYOS6f0nIZOS6e0mI[OS6e0lI\\OU6c0kI]OU6c0kI]OV6c0hI^OX6b0hI_OX6`0hI@X6`0hI@Y6?fIBZ6>fIB[6=eIC[6=eIC\\6o5CjId0U6T4O1O1O1O2N1O1O1O1O1O1O2N1O1O1O1O1O1O2N1O1M3K5J6J7I6N2N2O1N2N2N3N11O1O001O1O10O01O1O00000010O00000001O01O0001O000000010O0000001O0001O00010O000010O0001O0O2O0O1O2O0O101N1O101N1O2O0O101N1O101mLoDT2R;iMSEU2m:hMWEU2l:gMXEV2j:gMZEV2i:fM[EW2g:fM^EW2c:fMaEW2f;WNZCT1i in this image.", "objects": [{"patches": [85, 86, 87, 88, 89, 90, 91, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 226, 227, 228, 229], "bbox": [0.47190624999999997, 0.17752083333333332, 1.0, 0.5910208333333333], "iscrowd": 0, "area": 34236.63854999999, "rle": {"size": [480, 640], "counts": "Wf]44l>9G:F6J1O00001O001O001O001O001O000000000O1000000000000000O100000000000000000O100000O1000000001O001O00001O001O0O2O00001O001O001O00001O001O0000000O10000000000000000000000000000000000O100000000000000000000000000000000O10000O11O6J6JM3M3M3M3M3[C^Ni;e1TD^Ni;e1UD]Nh;m1oCVNn;_20000000001N100000001O0O1000001O00000O10001O0000000O2O000000001N100000001O000O1000001O00000O101O0000000O2ON2O1N2N2O1N1011O001O1O1O0O2O1O1O001O1N2O001O1O1O1O1N2O1O1O1O001N2O1O1O1O1O1O1O0O0100000000O10O100000O10000000gETL]9m3_FXL`9g3]F^Lb9b3[FbLd9^3XFgLg9Y3UFmLi9S3TFQMk9S4N2O1O1O1O1O4L5K5J3NO1O1N2O1O1O001O1O1NN3M3N21O1O1010O00000000000000000001O0000000000001O000001O0fFbJR9f5001O0000000000001O000000000000006I2O0O2O1N101N101N2O0O2O0O2O3L;F8G4M3L3N3L3N3L4M2M4M2M4M3L3N3L3N3L4M2MeJ"}, "label": "firetruck in the right hand picture"}]} {"id": 7852, "image": "COCO_train2014_000000007852.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red fire truck furthest to the right\"."}], "task": "refering", "answer_template": "The \"the red fire truck furthest to the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 87, 88, 89, 90, 91, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 202, 203, 204, 205, 206, 226, 227, 228, 229], "bbox": [0.47190624999999997, 0.17752083333333332, 1.0, 0.5910208333333333], "iscrowd": 0, "area": 34236.63854999999, "rle": {"size": [480, 640], "counts": "Wf]44l>9G:F6J1O00001O001O001O001O001O000000000O1000000000000000O100000000000000000O100000O1000000001O001O00001O001O0O2O00001O001O001O00001O001O0000000O10000000000000000000000000000000000O100000000000000000000000000000000O10000O11O6J6JM3M3M3M3M3[C^Ni;e1TD^Ni;e1UD]Nh;m1oCVNn;_20000000001N100000001O0O1000001O00000O10001O0000000O2O000000001N100000001O000O1000001O00000O101O0000000O2ON2O1N2N2O1N1011O001O1O1O0O2O1O1O001O1N2O001O1O1O1O1N2O1O1O1O001N2O1O1O1O1O1O1O0O0100000000O10O100000O10000000gETL]9m3_FXL`9g3]F^Lb9b3[FbLd9^3XFgLg9Y3UFmLi9S3TFQMk9S4N2O1O1O1O1O4L5K5J3NO1O1N2O1O1O001O1O1NN3M3N21O1O1010O00000000000000000001O0000000000001O000001O0fFbJR9f5001O0000000000001O000000000000006I2O0O2O1N101N101N2O0O2O0O2O3L;F8G4M3L3N3L3N3L4M2M4M2M4M3L3N3L3N3L4M2MeJ"}, "label": "the red fire truck furthest to the right"}]} {"id": 89773, "image": "COCO_train2014_000000089773.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a picture featuring a girl and three panda bears\"."}], "task": "refering", "answer_template": "The \"a picture featuring a girl and three panda bears\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 182, 183, 184, 185, 186, 187, 188, 201], "bbox": [0.10714, 0.06491978609625669, 0.9057799999999999, 0.8644117647058822], "iscrowd": 0, "area": 93864.8106, "rle": {"size": [374, 500], "counts": "Vhc09];9G:F9G:F9G:F:F9G:F9G:F9G:F:F9G:E:G:F9G:F:F9G:F9G:F9G:F:F9G3M0000O100000000O10000000000O100000000O1000O1000O100000000O100000000O100000000O100000000O100000000O10000000000O10O100000O100000000O100000000O100000000O100000000O100000000O100000000O1000O100000O100000000O100000000O100000000O100000000O100000000O100000000O10O100000O10000000000O100000000O100000000O100000000O100000000O100000000O010000000O100000000O10000000000O100000000O100000000O100000000O10000000O0100000000O100000000O100000000O10000000000O100000000O100000000O100000O10O100000000O100000000O100000000O100000000O10000000000O100000000O100000O10O100000000O100000000O100000000O100000000O100000000O10000000000O1000O1000O100000000O10001O4L4K5L4L4L5K4K5L4L4L4L5J5L4L4L4L4K6K4L4L4L4K5L5K4L4L4L4K6K4L4L4L4K5L5K4L4L4K5L4L5K4L4K5L4L4L5K4K5L4L4L5K4K5L`m`0"}, "label": "a picture featuring a girl and three panda bears"}]} {"id": 89773, "image": "COCO_train2014_000000089773.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a laptop cover\"."}], "task": "refering", "answer_template": "The \"a laptop cover\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [10, 11, 12, 13, 14, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 182, 183, 184, 185, 186, 187, 188, 201], "bbox": [0.10714, 0.06491978609625669, 0.9057799999999999, 0.8644117647058822], "iscrowd": 0, "area": 93864.8106, "rle": {"size": [374, 500], "counts": "Vhc09];9G:F9G:F9G:F:F9G:F9G:F9G:F:F9G:E:G:F9G:F:F9G:F9G:F9G:F:F9G3M0000O100000000O10000000000O100000000O1000O1000O100000000O100000000O100000000O100000000O100000000O10000000000O10O100000O100000000O100000000O100000000O100000000O100000000O100000000O1000O100000O100000000O100000000O100000000O100000000O100000000O100000000O10O100000O10000000000O100000000O100000000O100000000O100000000O100000000O010000000O100000000O10000000000O100000000O100000000O100000000O10000000O0100000000O100000000O100000000O10000000000O100000000O100000000O100000O10O100000000O100000000O100000000O100000000O10000000000O100000000O100000O10O100000000O100000000O100000000O100000000O100000000O10000000000O1000O1000O100000000O10001O4L4K5L4L4L5K4K5L4L4L4L5J5L4L4L4L4K6K4L4L4L4K5L5K4L4L4L4K6K4L4L4L4K5L5K4L4L4K5L4L5K4L4K5L4L4L5K4K5L4L4L5K4K5L`m`0"}, "label": "a laptop cover"}]} {"id": 376505, "image": "COCO_train2014_000000376505.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the armchair with a bag in it\"."}], "task": "refering", "answer_template": "The \"the armchair with a bag in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [97, 98, 99, 100, 120, 121, 123, 124, 144, 145, 147, 148, 167, 168, 169, 170, 171, 190, 191], "bbox": [0.24625, 0.23210416666666667, 0.459953125, 0.4917708333333334], "iscrowd": 0, "area": 6535.1932, "rle": {"size": [480, 640], "counts": "lUZ2;c>7J6J6K4K010kBoNZb:A_E?a:_O`Ec0^:]OcEc0]:\\OdEe0[:ZOeEg0[:YOeEh0Z:XOfEh0Z:WOfEk0Y:UOgEk0Y:TOhEl0X:TOgEn0W:SOjEl0V:TOlEj0T:WOlEh0T:XOnEe0S:[OnEd0R:]OoEa0Q:_OQF?o9ASF=m9CVF9k9GXF6g9K[F3e9M]F1c9O]F1c9O]F1c9O]F0d90\\F0d90\\F0d90\\F0d90]FOc91]FNc94\\FLd94\\FLd95[FKe95[FKe95[FJf96ZFJf96ZFJf96ZFJf96ZFJe97\\FGe99[FGe99[FGe99[FGe99[FGe99[FFf9:YFGg99YFGg99YFGg99YFGf9:ZFEh9:XFFh9:XFFh9:XFFh9:XFFi99VFGk99UFGl99SFGn98RFHo97QFIQ:5oEJT:4lELX:0hE0]:KcE5b:F^E:g:AYE?d01O0O2O001O002M3NeoQ5"}, "label": "the armchair with a bag in it"}]} {"id": 89788, "image": "COCO_train2014_000000089788.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman laying down with a coffee cup in her hand\"."}], "task": "refering", "answer_template": "The \"a woman laying down with a coffee cup in her hand\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [63, 64, 65, 78, 79, 80, 81, 82, 83, 95, 96, 97, 98, 99, 100, 109, 110, 111, 112, 113, 114, 115, 116, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 139, 140, 141, 143, 144, 145, 146, 147, 148, 149, 150, 154, 155, 156, 157, 161, 162, 163, 164, 165, 166, 167, 171, 172, 173, 178, 179, 180, 181, 182, 183, 184, 189, 190, 195, 196, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 217, 218, 228, 229, 230, 231, 232, 233, 234, 235, 244, 245, 246, 247, 248, 249, 250, 251, 260, 261, 262, 263, 264, 265, 266, 267, 268, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 311, 312, 313, 320, 321], "bbox": [0.051229166666666666, 0.14837499999999998, 0.9243958333333333, 0.807578125], "iscrowd": 0, "area": 76913.74185000002, "rle": {"size": [640, 480], "counts": "Qm?1nc01O2N2N1O2N1O2O0O2N2N1O2N1O2O0O2N1O2N2O001N101O001O1O001O001O001O1O001O001O001O0O2O1O001O001O001O1O001O0O1000O100O10OO1000000O1000000O100000000O1000000O1000000YO]]O1db00^]OMbb03a]OJ_b06d]OG\\b09g]OC[b0 in this image.", "objects": [{"patches": [63, 64, 65, 78, 79, 80, 81, 82, 83, 95, 96, 97, 98, 99, 100, 109, 110, 111, 112, 113, 114, 115, 116, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 139, 140, 141, 143, 144, 145, 146, 147, 148, 149, 150, 154, 155, 156, 157, 161, 162, 163, 164, 165, 166, 167, 171, 172, 173, 178, 179, 180, 181, 182, 183, 184, 189, 190, 195, 196, 197, 198, 199, 200, 201, 212, 213, 214, 215, 216, 217, 218, 228, 229, 230, 231, 232, 233, 234, 235, 244, 245, 246, 247, 248, 249, 250, 251, 260, 261, 262, 263, 264, 265, 266, 267, 268, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 311, 312, 313, 320, 321], "bbox": [0.051229166666666666, 0.14837499999999998, 0.9243958333333333, 0.807578125], "iscrowd": 0, "area": 76913.74185000002, "rle": {"size": [640, 480], "counts": "Qm?1nc01O2N2N1O2N1O2O0O2N2N1O2N1O2O0O2N1O2N2O001N101O001O1O001O001O001O1O001O001O001O0O2O1O001O001O001O1O001O0O1000O100O10OO1000000O1000000O100000000O1000000O1000000YO]]O1db00^]OMbb03a]OJ_b06d]OG\\b09g]OC[b0 in this image.", "objects": [{"patches": [125, 140, 141, 142, 143, 144, 156, 157, 158, 159, 160, 161, 170, 171, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 204, 205, 206, 207, 208, 209, 210, 211, 212, 221, 222, 223, 224, 225, 226, 227, 228, 238, 239, 240, 241, 242, 243, 244, 255, 256, 257, 258, 259, 260, 272, 273, 274, 275, 276, 277, 289, 290, 291, 292, 293, 294, 306, 307, 308, 309, 310, 311, 323, 324, 325, 326, 327, 328, 340, 341, 342, 343, 344, 345, 357, 358, 359, 360, 361, 362, 374, 375, 376, 377, 378, 379], "bbox": [0.005, 0.342390625, 0.5240208333333334, 0.997859375], "iscrowd": 0, "area": 69747.85715, "rle": {"size": [640, 480], "counts": "Sa1l:T90001O0000001O0000001O0O10001O0000001O0000001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O001O001O001O00001O001O001O001O001O001O001O001O00N2O1O1dNmEjHT:V7nEgHS:X7PFeHQ:Z7RFcHo9\\7UF`Hl9_7WF^Hj9a7YF\\Hh9c7[FZHf9f7\\FWHe9h7_FTHb9k7aFQHa9n7bFoG_9P8dFmG]9R8Y1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O100O10000O100O10000O100O10000O100O010O1000lKcHeK\\7W4kHgKU7T4RIjKm6Q4\\IlKd6o3cIoK]6l3kIQLT6k3SJSLm5h3[JULd5g3cJWL]5d3jJZLV5a3RK\\Lm4c3WK[Li4d3[KYLd4f3`KXL`4f3eKWL[4h3hKVLW4i3mKULS4i3RLTLm3l3VLRLj3l3[LQLe3n3^LPLa3o3cLoK]3o3hLnKW3R4lLlKT3R4QMkKo2T4TMjKk2U4ZMhKf2V4^MhKa2X4nMZKR2d4_NmJa1Q5PO`Jo0`5AQJ?m5j601N2N101N1O2O1N3N100O1O1O1O2N100O1O1O1O1O2N100O1O1O1O2N100O1O1O1O1O2O0O1O1O1O1O2O0O1O1N2O1O1O1O2M2O1O1O1N2O1O1O1O2M2O1O1O1N2O1O1O1O2M2O1O2N3L4M3M3J6POP1@`0F:G:E:F:G[X^4"}, "label": "the person in red"}]} {"id": 89788, "image": "COCO_train2014_000000089788.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a red robe looking up at the person she is resting on\"."}], "task": "refering", "answer_template": "The \"a woman in a red robe looking up at the person she is resting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [125, 140, 141, 142, 143, 144, 156, 157, 158, 159, 160, 161, 170, 171, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 204, 205, 206, 207, 208, 209, 210, 211, 212, 221, 222, 223, 224, 225, 226, 227, 228, 238, 239, 240, 241, 242, 243, 244, 255, 256, 257, 258, 259, 260, 272, 273, 274, 275, 276, 277, 289, 290, 291, 292, 293, 294, 306, 307, 308, 309, 310, 311, 323, 324, 325, 326, 327, 328, 340, 341, 342, 343, 344, 345, 357, 358, 359, 360, 361, 362, 374, 375, 376, 377, 378, 379], "bbox": [0.005, 0.342390625, 0.5240208333333334, 0.997859375], "iscrowd": 0, "area": 69747.85715, "rle": {"size": [640, 480], "counts": "Sa1l:T90001O0000001O0000001O0O10001O0000001O0000001O0000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O001O001O001O00001O001O001O001O001O001O001O001O00N2O1O1dNmEjHT:V7nEgHS:X7PFeHQ:Z7RFcHo9\\7UF`Hl9_7WF^Hj9a7YF\\Hh9c7[FZHf9f7\\FWHe9h7_FTHb9k7aFQHa9n7bFoG_9P8dFmG]9R8Y1O100O1O1O1O1O1O1O100O1O1O1O1O1O100O1O100O10000O100O10000O100O10000O100O010O1000lKcHeK\\7W4kHgKU7T4RIjKm6Q4\\IlKd6o3cIoK]6l3kIQLT6k3SJSLm5h3[JULd5g3cJWL]5d3jJZLV5a3RK\\Lm4c3WK[Li4d3[KYLd4f3`KXL`4f3eKWL[4h3hKVLW4i3mKULS4i3RLTLm3l3VLRLj3l3[LQLe3n3^LPLa3o3cLoK]3o3hLnKW3R4lLlKT3R4QMkKo2T4TMjKk2U4ZMhKf2V4^MhKa2X4nMZKR2d4_NmJa1Q5PO`Jo0`5AQJ?m5j601N2N101N1O2O1N3N100O1O1O1O2N100O1O1O1O1O2N100O1O1O1O2N100O1O1O1O1O2O0O1O1O1O1O2O0O1O1N2O1O1O1O2M2O1O1O1N2O1O1O1O2M2O1O1O1N2O1O1O1O2M2O1O2N3L4M3M3J6POP1@`0F:G:E:F:G[X^4"}, "label": "a woman in a red robe looking up at the person she is resting on"}]} {"id": 16069, "image": "COCO_train2014_000000016069.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person in the rain slicker crossing the bridge\"."}], "task": "refering", "answer_template": "The \"the person in the rain slicker crossing the bridge\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 127, 128, 150, 151, 173, 174, 196, 197, 219, 220, 242, 243, 244, 264, 265, 267, 287, 288], "bbox": [0.49800000000000005, 0.3086792452830189, 0.647546875, 0.8256132075471698], "iscrowd": 0, "area": 8145.275550000002, "rle": {"size": [424, 640], "counts": "^]T41V=3N1N2N101N1O2M3L3N2M3N2M3VK0ZL2e32QL5n3OdK<\\4GYKa0f4CRKb0n4BiJc0W4[NSJV1^1d0^4[NoIV1Z1c0g4ZNlIZ1Q1b0S5UNjI_1i0`0^5RNgIT5Z6nJcIR5^6k01O100O1O1O1O1O101N1VO]InJe6k4gInJZ6n4S1K4M3M4Lb0A4M2N0oNPGiMP9T2UGjMj8T2ZGkMf8S2]GlMb8R2bGmM^8Q2dGoM[8o1iGPNW8n1kGRNT8l1PHSNP8k1RHUNm7i1WHVNi7h1YHXNi7c1ZH]Nh7_1ZHaNi7Y1ZHfNk7T1WHlNl7n0WHROn7g0UHXOQ8?RHAT86oGJW8NkG2f:0000000O100O100O2O009F100O100OO1O2M2O1O1O2M2O1O1O4L3L5ImXm2"}, "label": "the person in the rain slicker crossing the bridge"}]} {"id": 16069, "image": "COCO_train2014_000000016069.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a raincoat walking by the water\"."}], "task": "refering", "answer_template": "The \"a woman in a raincoat walking by the water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [104, 127, 128, 150, 151, 173, 174, 196, 197, 219, 220, 242, 243, 244, 264, 265, 267, 287, 288], "bbox": [0.49800000000000005, 0.3086792452830189, 0.647546875, 0.8256132075471698], "iscrowd": 0, "area": 8145.275550000002, "rle": {"size": [424, 640], "counts": "^]T41V=3N1N2N101N1O2M3L3N2M3N2M3VK0ZL2e32QL5n3OdK<\\4GYKa0f4CRKb0n4BiJc0W4[NSJV1^1d0^4[NoIV1Z1c0g4ZNlIZ1Q1b0S5UNjI_1i0`0^5RNgIT5Z6nJcIR5^6k01O100O1O1O1O1O101N1VO]InJe6k4gInJZ6n4S1K4M3M4Lb0A4M2N0oNPGiMP9T2UGjMj8T2ZGkMf8S2]GlMb8R2bGmM^8Q2dGoM[8o1iGPNW8n1kGRNT8l1PHSNP8k1RHUNm7i1WHVNi7h1YHXNi7c1ZH]Nh7_1ZHaNi7Y1ZHfNk7T1WHlNl7n0WHROn7g0UHXOQ8?RHAT86oGJW8NkG2f:0000000O100O100O2O009F100O100OO1O2M2O1O1O2M2O1O1O4L3L5ImXm2"}, "label": "a woman in a raincoat walking by the water"}]} {"id": 212679, "image": "COCO_train2014_000000212679.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the fork in the side of green plate\"."}], "task": "refering", "answer_template": "The \"the fork in the side of green plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [284, 285, 306, 307, 308, 328, 329, 350, 351, 373], "bbox": [0.21271874999999998, 0.703125, 0.44526562500000005, 0.9592291666666667], "iscrowd": 0, "area": 2544.165649999999, "rle": {"size": [480, 640], "counts": "mUP22m>7I2O1N2N2N101N2N2N10O0000001O0001O01O00001O0000010O00001O00001O01O01O00001O01O0001O00010O00001O00010O00001O01O0001O00010O00001O00010O00001O01O0001O000010O0001O00010O00001O01O01O00000010O0001O0001N1O1O2M3N2O1O1O1O1O1O1O1O1O1O1O1O1O100O01O000O2O001O0]OoA8Q>GPB9Q>FoA:Q>FoAOO3S>MnA0O2T>NmA0O2T>MnA0OOONU>2nA0N0ONU>2nAOO0OOU>1mA0O0OOU>1mAO01NOU>0nA0O0O0U>OmA1O0OOV>0lA001NOW>OkA1000OU>0RB1IOb>0_A0a>0_A0a>0_A0h>0omW5"}, "label": "the fork in the side of green plate"}]} {"id": 212679, "image": "COCO_train2014_000000212679.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a steel fork on the right side of a green plate\"."}], "task": "refering", "answer_template": "The \"a steel fork on the right side of a green plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [284, 285, 306, 307, 308, 328, 329, 350, 351, 373], "bbox": [0.21271874999999998, 0.703125, 0.44526562500000005, 0.9592291666666667], "iscrowd": 0, "area": 2544.165649999999, "rle": {"size": [480, 640], "counts": "mUP22m>7I2O1N2N2N101N2N2N10O0000001O0001O01O00001O0000010O00001O00001O01O01O00001O01O0001O00010O00001O00010O00001O01O0001O00010O00001O00010O00001O01O0001O000010O0001O00010O00001O01O01O00000010O0001O0001N1O1O2M3N2O1O1O1O1O1O1O1O1O1O1O1O1O100O01O000O2O001O0]OoA8Q>GPB9Q>FoA:Q>FoAOO3S>MnA0O2T>NmA0O2T>MnA0OOONU>2nA0N0ONU>2nAOO0OOU>1mA0O0OOU>1mAO01NOU>0nA0O0O0U>OmA1O0OOV>0lA001NOW>OkA1000OU>0RB1IOb>0_A0a>0_A0a>0_A0h>0omW5"}, "label": "a steel fork on the right side of a green plate"}]} {"id": 532175, "image": "COCO_train2014_000000532175.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue bicycle with a wire basket\"."}], "task": "refering", "answer_template": "The \"a blue bicycle with a wire basket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [263, 265, 266, 286, 287, 288, 289, 290, 305, 309, 310, 311, 312, 313, 323, 324, 325, 326, 327, 328, 330, 331, 332, 333, 334, 335, 336, 346, 347, 348, 349, 350, 351, 353, 354, 355, 356, 357, 358, 359, 360, 369, 370, 371, 372, 373, 374, 375, 376, 378, 379, 380, 381, 382, 383], "bbox": [0.043546875, 0.6774166666666667, 0.667734375, 1.0], "iscrowd": 0, "area": 31843.5602, "rle": {"size": [480, 640], "counts": "cb==W>o in this image.", "objects": [{"patches": [263, 265, 266, 286, 287, 288, 289, 290, 305, 309, 310, 311, 312, 313, 323, 324, 325, 326, 327, 328, 330, 331, 332, 333, 334, 335, 336, 346, 347, 348, 349, 350, 351, 353, 354, 355, 356, 357, 358, 359, 360, 369, 370, 371, 372, 373, 374, 375, 376, 378, 379, 380, 381, 382, 383], "bbox": [0.043546875, 0.6774166666666667, 0.667734375, 1.0], "iscrowd": 0, "area": 31843.5602, "rle": {"size": [480, 640], "counts": "cb==W>o in this image.", "objects": [{"patches": [78, 79, 80, 101, 102, 103, 124, 125, 126, 146, 147, 148, 149, 150, 170, 171, 172, 193, 194, 195, 216, 217, 218, 240, 241, 263, 264], "bbox": [0.379359375, 0.19608333333333333, 0.5379531249999999, 0.6926875], "iscrowd": 0, "area": 13453.825099999998, "rle": {"size": [480, 640], "counts": "kRb39f>2N2^OGZB;d=KUB7j=a0N2N1O2N2O0O2N2O0O2UOk0J6L4M3]D\\Mf:f2XE]Me:f2YE\\Md:f2[E]Ma:f2\\E]MKGW:o2lE]MGKY:k2lEnMP:U2nElMQ:b3O2N100O1O100O1jGVKV6i4fI_KV6b4hI`KW6a4hI`KX6`4gIaKX6`4fIbKY6^4gIcKX6_4fIbKZ6j4XIXKg6m4TITKk6P5PIQKP7S5kHoJU7X5bHjJ]7\\600100O1O1O1N2O1O1N31N1O2O0O1O2N100O1O100OlM]HlLc7Q3aHlL`7T3bHjL^7U3fHfL]7X3fHeL\\7Y3jH`LX7_3oHXLU7a2THYMm0OR7e2SH[MP1IQ7j2PH\\MQ:b2QF\\MQ:a2QFSM\\:k2gEoL_:m2l0O2M3M4L3M5L6I7IA3N1N2N4M3YOQB6_>M2O1L3M4MVbZ4"}, "label": "a huge fish kite that is mostly blue and purple"}]} {"id": 188120, "image": "COCO_train2014_000000188120.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giant fish kite\"."}], "task": "refering", "answer_template": "The \"a giant fish kite\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [78, 79, 80, 101, 102, 103, 124, 125, 126, 146, 147, 148, 149, 150, 170, 171, 172, 193, 194, 195, 216, 217, 218, 240, 241, 263, 264], "bbox": [0.379359375, 0.19608333333333333, 0.5379531249999999, 0.6926875], "iscrowd": 0, "area": 13453.825099999998, "rle": {"size": [480, 640], "counts": "kRb39f>2N2^OGZB;d=KUB7j=a0N2N1O2N2O0O2N2O0O2UOk0J6L4M3]D\\Mf:f2XE]Me:f2YE\\Md:f2[E]Ma:f2\\E]MKGW:o2lE]MGKY:k2lEnMP:U2nElMQ:b3O2N100O1O100O1jGVKV6i4fI_KV6b4hI`KW6a4hI`KX6`4gIaKX6`4fIbKY6^4gIcKX6_4fIbKZ6j4XIXKg6m4TITKk6P5PIQKP7S5kHoJU7X5bHjJ]7\\600100O1O1O1N2O1O1N31N1O2O0O1O2N100O1O100OlM]HlLc7Q3aHlL`7T3bHjL^7U3fHfL]7X3fHeL\\7Y3jH`LX7_3oHXLU7a2THYMm0OR7e2SH[MP1IQ7j2PH\\MQ:b2QF\\MQ:a2QFSM\\:k2gEoL_:m2l0O2M3M4L3M5L6I7IA3N1N2N4M3YOQB6_>M2O1L3M4MVbZ4"}, "label": "a giant fish kite"}]} {"id": 319192, "image": "COCO_train2014_000000319192.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hot dog on the left side of a plate\"."}], "task": "refering", "answer_template": "The \"a hot dog on the left side of a plate\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 52, 53, 54, 55, 56, 58, 74, 75, 76, 77, 78, 79, 80, 81, 82, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261, 281, 282, 283], "bbox": [0.0698125, 0.09080078125, 0.583015625, 0.6827734375], "iscrowd": 0, "area": 55141.1004, "rle": {"size": [512, 640], "counts": "jff05i?5J7J5J6K5J7J5J6K6I6K5K5L5K4L4L4K6K2N2N2N2M2O2N2N2N2M2O2N2N2N2M2O2N2N2N2M2O2N2N2N2N1N3O1N2N2N1O2N2N2N2N2N1O2O1N2N2N1O2N2N2N2N1O2N2O1N2N1O2N2N2N1O2N2N101N2N1O2N2N2N1O2O1N1O2N2N101N2N2N1O2O1N1O2N2O1N1O2N2N101N2N1O1O100O1O1O1O2O0O1O1O100O1O1O1O100O1O1O1O101O00000000000000000000000000001O00000000000O100O1O1O1O1O1O2O0O0001O01O01O00002O0O2N1O2O0O2N1O2N100O2N1O1O01O01O00010O001O1O010O001M3L3M4M3L3M4MD;4M2M4L4M2M4L4M2M4M3L3M4M3M2O3N1O1O1N2O2N1O2M3N2N2N2M3N2N2N2M3N2N2N2M3N3M2N2M4M3M3M3L3N3M3M3L4M2N3M3L4M3M2N5J2O001O1N2O001O1N101O1O1N101O1O0O2O1O001N2O001O001O001O001O001O001O001O001O001O0010O01O001O2N1O1O2N1O1O2N3M2N3M2N3M2N3M3M2N3M2N3M2N3M3M2N3M2N3M2N3M3M\\\\U4"}, "label": "a hot dog on the left side of a plate"}]} {"id": 319192, "image": "COCO_train2014_000000319192.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a hot dog with ketchup next to a stove\"."}], "task": "refering", "answer_template": "The \"a hot dog with ketchup next to a stove\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [31, 32, 52, 53, 54, 55, 56, 58, 74, 75, 76, 77, 78, 79, 80, 81, 82, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 185, 186, 187, 188, 189, 190, 191, 192, 193, 209, 210, 211, 212, 213, 214, 215, 216, 232, 233, 234, 235, 236, 237, 238, 257, 258, 259, 260, 261, 281, 282, 283], "bbox": [0.0698125, 0.09080078125, 0.583015625, 0.6827734375], "iscrowd": 0, "area": 55141.1004, "rle": {"size": [512, 640], "counts": "jff05i?5J7J5J6K5J7J5J6K6I6K5K5L5K4L4L4K6K2N2N2N2M2O2N2N2N2M2O2N2N2N2M2O2N2N2N2M2O2N2N2N2N1N3O1N2N2N1O2N2N2N2N2N1O2O1N2N2N1O2N2N2N2N1O2N2O1N2N1O2N2N2N1O2N2N101N2N1O2N2N2N1O2O1N1O2N2N101N2N2N1O2O1N1O2N2O1N1O2N2N101N2N1O1O100O1O1O1O2O0O1O1O100O1O1O1O100O1O1O1O101O00000000000000000000000000001O00000000000O100O1O1O1O1O1O2O0O0001O01O01O00002O0O2N1O2O0O2N1O2N100O2N1O1O01O01O00010O001O1O010O001M3L3M4M3L3M4MD;4M2M4L4M2M4L4M2M4M3L3M4M3M2O3N1O1O1N2O2N1O2M3N2N2N2M3N2N2N2M3N2N2N2M3N3M2N2M4M3M3M3L3N3M3M3L4M2N3M3L4M3M2N5J2O001O1N2O001O1N101O1O1N101O1O0O2O1O001N2O001O001O001O001O001O001O001O001O001O0010O01O001O2N1O1O2N1O1O2N3M2N3M2N3M2N3M3M2N3M2N3M2N3M3M2N3M2N3M2N3M3M\\\\U4"}, "label": "a hot dog with ketchup next to a stove"}]} {"id": 319192, "image": "COCO_train2014_000000319192.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the piece with sauce in the right side of the image\"."}], "task": "refering", "answer_template": "The \"the piece with sauce in the right side of the image\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 85, 86, 88, 105, 106, 107, 108, 109, 110, 111, 112, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 149, 150, 151, 152, 153, 154, 155, 156, 157, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 284, 285, 286, 287, 288, 289, 290, 291, 292, 307, 308, 309, 310, 311, 312, 313, 314, 315, 331, 332, 333, 334, 335, 336, 337, 355, 356, 357, 358, 359, 360], "bbox": [0.370265625, 0.175234375, 0.8803124999999999, 0.89646484375], "iscrowd": 0, "area": 74401.15144999999, "rle": {"size": [512, 640], "counts": "Vif3B5K4L5L3L5K5K4L5K4L5K4L5L3L5K4L5K5K4L3M2O2M2N3M3N1N3N1O2N1O2M3N1O2N1O2M3N1O2N1O2M3N1O2N1O2M2O2N2N1O2M2O2N2N1O2M2O2N2N1O2M2O2N1O2N2M2O2N1O2N2M2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N2O1O1O1O1N200O100O2N101N101N101N1O2O0O101N101N1O2O0O2O0O2O0O2N101N101N101N1O101N101N101N1O2O0O2O0O2O0O1O010O100O010O1O10O0100O10O0100000O1000000O10O1000O10O1OO1O100O1O100O2N100O1O100O1O100O101N1O2O1N1O2N1O2N2N1O2N2N101N2N1O2N2N1O2N2N1O2N2N2O1N2N3M2N2N2N2N3M2N2M3N2N3M2N2N2N2N3M2N6J7I7I8G5L2N3M2N3M2N3M2N3M2N3M2N3L3N2N2N3M2N2N2N3M2N2N3M2N2N2N3M2N2N3M2N2N2N1O1O1O1O1O1O10O01O1O1O1O1O1O1O1O1O001O1O2N3M3L4M2N3M3M3M3M2M4M3M3M2N3M3M3L3N3M3M3M2N3M3L4M2N3M3M3JV\\V1"}, "label": "the piece with sauce in the right side of the image"}]} {"id": 327404, "image": "COCO_train2014_000000327404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass of juice\"."}], "task": "refering", "answer_template": "The \"a glass of juice\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [285, 286, 287, 302, 303, 304, 305, 319, 320, 321, 322, 336, 337, 338, 353, 354, 355, 370, 371, 372, 387, 388, 389], "bbox": [0.765375, 0.71446875, 0.9595208333333333, 0.9797812499999999], "iscrowd": 0, "area": 12347.277999999998, "rle": {"size": [640, 480], "counts": "jjU72mc0f0ZOe0ZOg0ZOe0[Of0ZOe0[O in this image.", "objects": [{"patches": [285, 286, 287, 302, 303, 304, 305, 319, 320, 321, 322, 336, 337, 338, 353, 354, 355, 370, 371, 372, 387, 388, 389], "bbox": [0.765375, 0.71446875, 0.9595208333333333, 0.9797812499999999], "iscrowd": 0, "area": 12347.277999999998, "rle": {"size": [640, 480], "counts": "jjU72mc0f0ZOe0ZOg0ZOe0[Of0ZOe0[O in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 21, 22, 23, 24, 25, 26, 27, 28, 29, 39, 40, 41, 42, 43, 44, 45, 56, 57, 58, 59, 60, 61, 62, 63, 73, 74, 75, 76, 77, 78, 79, 80, 89, 90, 91, 92, 93, 94, 95, 96, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 296, 297, 298, 299, 300, 301, 305], "bbox": [0.0, 0.0, 1.0, 0.77778125], "iscrowd": 0, "area": 160883.4144499999, "rle": {"size": [640, 480], "counts": "P9b6]=1N2O1O0O2O1O1N2O1O1N2O1O1N2O1O1N2O001N2O1O1N2O1N2O1O1N2O1O1N2O001N2O1O1N2O1O1N2O1O1N2O1O1N101O1N2O1O1N2O1O1O1O10KjD`GU;]8PEbGo:[8UEeGk:X8?N2N3N11O01O01O01O010O00010O00010O00010O00010O00010O00010O00010O01O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O100kKYGPMh8P3XGPMh8P3YGoLg8Q3YGnLh8R3YGmLg8T3XGlLh8T3YGjLi8U3WGkLi8U3XGiLj8V3VGjLj8W3VGhLk8W3UGhLl8X3UGgLl8U2eG_J_O[3n8U2VHjMj7V2VHTKcNi1X9R3VHQKQO`1j8_3[HhJZO]1\\8j3SJjKn5V4^J^Kc5a4aJ[K_5e4aJZKa5e4_J[Ka5e4_JZKc5e4]J[Kc5e4]J[Kd5d4\\J[Ke5e4[J[Kf5d4[JZKf5f4ZJZKe5g4[JYKe5g4[JXKf5h4ZJXKe5i4aJoJa5Q5fJgJ[5Y5f4000O1000000O100001O0000001O0000001O00002N2N2N1O2N2N1O2N2N1O0000000000O100000000000000O1000000000000O100000000000000O100000000000000O100000000000001N1001O000000000000000000001O000000000000000000001O000000000000000000001O000000000000000000001O000000000000000000001O000000000000eEoJg5Q5XJQKg5o4YJRKf5n4YJTKf5l4ZJVKe5i4ZJYKe5g4[JZKd5f4[J\\Kd5d4\\J]Kc5c4^J]Ka5c4_J_K_5a4aJ`K^5`4bJaK]5_4cJbK\\5^4eJbKZ5^4fJdKY5[4gJeKY5[4gJfKX5Z4hJfKX5Z4iJfKV5Z4jJfKV5Z4jJgKU5Y4iJiKW5W4gJlKX5T4gJmKY5S4eJPLZ5P4dJRL\\5n3cJSL]5m3aJVL^5j3`JXL`5h3_JZL`5f3lHWKmNT1W8e3jHZKmNS1X8c3iH^KmNo0Z8c3eHdKnNk0\\8a3SHXL_O8^8`3dGiLKI`8^3ZGVM3]Oc8]3TG]M8WOc8\\3nF4R9LgFm4dAVK^>i4^AZKd>d4ZA^Kh>l42N2N2N2N2N2N2N2N2N2N2N2N2N2N2N6J=C_D"}, "label": "a man with glasses and a striped shirt"}]} {"id": 327404, "image": "COCO_train2014_000000327404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the guy with his arms on the table\"."}], "task": "refering", "answer_template": "The \"the guy with his arms on the table\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 8, 9, 10, 11, 12, 21, 22, 23, 24, 25, 26, 27, 28, 29, 39, 40, 41, 42, 43, 44, 45, 56, 57, 58, 59, 60, 61, 62, 63, 73, 74, 75, 76, 77, 78, 79, 80, 89, 90, 91, 92, 93, 94, 95, 96, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 296, 297, 298, 299, 300, 301, 305], "bbox": [0.0, 0.0, 1.0, 0.77778125], "iscrowd": 0, "area": 160883.4144499999, "rle": {"size": [640, 480], "counts": "P9b6]=1N2O1O0O2O1O1N2O1O1N2O1O1N2O1O1N2O001N2O1O1N2O1N2O1O1N2O1O1N2O001N2O1O1N2O1O1N2O1O1N2O1O1N101O1N2O1O1N2O1O1O1O10KjD`GU;]8PEbGo:[8UEeGk:X8?N2N3N11O01O01O01O010O00010O00010O00010O00010O00010O00010O00010O01O100O1O100O1O100O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O100O100O100kKYGPMh8P3XGPMh8P3YGoLg8Q3YGnLh8R3YGmLg8T3XGlLh8T3YGjLi8U3WGkLi8U3XGiLj8V3VGjLj8W3VGhLk8W3UGhLl8X3UGgLl8U2eG_J_O[3n8U2VHjMj7V2VHTKcNi1X9R3VHQKQO`1j8_3[HhJZO]1\\8j3SJjKn5V4^J^Kc5a4aJ[K_5e4aJZKa5e4_J[Ka5e4_JZKc5e4]J[Kc5e4]J[Kd5d4\\J[Ke5e4[J[Kf5d4[JZKf5f4ZJZKe5g4[JYKe5g4[JXKf5h4ZJXKe5i4aJoJa5Q5fJgJ[5Y5f4000O1000000O100001O0000001O0000001O00002N2N2N1O2N2N1O2N2N1O0000000000O100000000000000O1000000000000O100000000000000O100000000000000O100000000000001N1001O000000000000000000001O000000000000000000001O000000000000000000001O000000000000000000001O000000000000000000001O000000000000eEoJg5Q5XJQKg5o4YJRKf5n4YJTKf5l4ZJVKe5i4ZJYKe5g4[JZKd5f4[J\\Kd5d4\\J]Kc5c4^J]Ka5c4_J_K_5a4aJ`K^5`4bJaK]5_4cJbK\\5^4eJbKZ5^4fJdKY5[4gJeKY5[4gJfKX5Z4hJfKX5Z4iJfKV5Z4jJfKV5Z4jJgKU5Y4iJiKW5W4gJlKX5T4gJmKY5S4eJPLZ5P4dJRL\\5n3cJSL]5m3aJVL^5j3`JXL`5h3_JZL`5f3lHWKmNT1W8e3jHZKmNS1X8c3iH^KmNo0Z8c3eHdKnNk0\\8a3SHXL_O8^8`3dGiLKI`8^3ZGVM3]Oc8]3TG]M8WOc8\\3nF4R9LgFm4dAVK^>i4^AZKd>d4ZA^Kh>l42N2N2N2N2N2N2N2N2N2N2N2N2N2N2N6J=C_D"}, "label": "the guy with his arms on the table"}]} {"id": 327404, "image": "COCO_train2014_000000327404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white cup with coffee that says blue and white restaurant\"."}], "task": "refering", "answer_template": "The \"a white cup with coffee that says blue and white restaurant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [274, 275, 276, 277, 278, 291, 292, 293, 294, 295, 296, 308, 309, 310, 311, 312, 313, 325, 326, 327, 328, 329, 342, 343, 344, 345, 360, 361], "bbox": [0.1135625, 0.705515625, 0.4280833333333333, 0.922046875], "iscrowd": 0, "area": 15177.234650000004, "rle": {"size": [640, 480], "counts": "ojR18fc09G8H9G9G8H9G4N2M3N2N2M3N2M3N2M3N1O2M3N1N3N1O2M2O2M2O2N1O2N1O2N1N3N1O2N2N1O1P@VLh?k3W@ULi?l3U@VLj?P401N10001O000000000000000O1000000000000000000O100000000000000000000O1000000000O10000O1O1O101N1O1O2O0O1O2N100O2N1O101N1O1O2N1L4L5J5L4K6K4O101N100O2O000O2O0O101N100O101N100O2O000O101N100O101N100O101N101O0O2O0H9G8G:G8H9F_][5"}, "label": "a white cup with coffee that says blue and white restaurant"}]} {"id": 327404, "image": "COCO_train2014_000000327404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white mug that says blue & white restaurant\"."}], "task": "refering", "answer_template": "The \"a white mug that says blue & white restaurant\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [274, 275, 276, 277, 278, 291, 292, 293, 294, 295, 296, 308, 309, 310, 311, 312, 313, 325, 326, 327, 328, 329, 342, 343, 344, 345, 360, 361], "bbox": [0.1135625, 0.705515625, 0.4280833333333333, 0.922046875], "iscrowd": 0, "area": 15177.234650000004, "rle": {"size": [640, 480], "counts": "ojR18fc09G8H9G9G8H9G4N2M3N2N2M3N2M3N2M3N1O2M3N1N3N1O2M2O2M2O2N1O2N1O2N1N3N1O2N2N1O1P@VLh?k3W@ULi?l3U@VLj?P401N10001O000000000000000O1000000000000000000O100000000000000000000O1000000000O10000O1O1O101N1O1O2O0O1O2N100O2N1O101N1O1O2N1L4L5J5L4K6K4O101N100O2O000O2O0O101N100O101N100O2O000O101N100O101N100O101N101O0O2O0H9G8G:G8H9F_][5"}, "label": "a white mug that says blue & white restaurant"}]} {"id": 204529, "image": "COCO_train2014_000000204529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bottle of chardonnay\"."}], "task": "refering", "answer_template": "The \"a bottle of chardonnay\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [2, 3, 25, 26, 47, 48, 49, 50, 70, 71, 72, 73, 93, 94, 95, 96, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 185, 186, 187, 188, 209, 210, 211, 232, 233, 234, 255, 256, 257, 278, 279, 280, 281, 301, 302, 303, 304], "bbox": [0.064375, 0.005583333333333333, 0.247859375, 0.8017083333333334], "iscrowd": 0, "area": 28259.636300000002, "rle": {"size": [480, 640], "counts": "d\\c05b>?@?B?@`0A?A?@`0A?@`0A?@`0A>B?@`0A?@`0D in this image.", "objects": [{"patches": [2, 3, 25, 26, 47, 48, 49, 50, 70, 71, 72, 73, 93, 94, 95, 96, 116, 117, 118, 119, 120, 139, 140, 141, 142, 143, 162, 163, 164, 165, 166, 185, 186, 187, 188, 209, 210, 211, 232, 233, 234, 255, 256, 257, 278, 279, 280, 281, 301, 302, 303, 304], "bbox": [0.064375, 0.005583333333333333, 0.247859375, 0.8017083333333334], "iscrowd": 0, "area": 28259.636300000002, "rle": {"size": [480, 640], "counts": "d\\c05b>?@?B?@`0A?A?@`0A?@`0A?@`0A>B?@`0A?@`0D in this image.", "objects": [{"patches": [126, 127, 128, 149, 150, 151, 171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 240, 241, 242, 243, 265, 287, 288, 289, 310, 311, 312, 334], "bbox": [0.466859375, 0.3208125, 0.617515625, 0.8333333333333334], "iscrowd": 0, "area": 16936.011449999995, "rle": {"size": [480, 640], "counts": "S\\\\4=_>b0]Oc0]Oc0]O?B5J6J5L5aFnLi6X3TInLc6W3\\IoL[6V3bIQMh5WOaHn3e1lLj5ZOXHo3l1jLk5^3TJbLk5`3SJaLm5_3SJaLm5`3QJbLn5_3QJaLo5_3QJaLn5a3PJ`LP6a3oI`LP6a3oI_LQ6a3nI`LQ6b3nI^LR6c3lI_LS6a3mI_LS6b3lI^LT6c3jI^LU6c3kI^LT6c3jI^LV6c3iI]LW6c3iI^LU6d3jI\\LV6e3iI[LW6e3iI[LW6f3gI\\LX6e3gI[LY6g3eIYL[6j3bIVL^6l3`IUL_6n3^IRLb6P4[IQLe6R4nHXLR7l5O0000001O0001O01O000000O1[MfIdLZ6[3gIeLY6Z3hIfLX6Z3iIeLW6Z3jIeLW6[3iIeLX6Y3iIgLW6Y3iIgLW6X3jIhLV6X3kIfLV6Y3kIgLU6X3mIgLS6Y3mIgLT6W3nIgLS6Y3mIgLS6X3oIgLQ6Y3oIgLR6W3oIiLQ6V3QJhLP6X3PJhLP6W3RJhLo5W3RJhLn5W3TJhLl5X3UJeLm5Z3VJcLl5\\3UJaLm5^3UJ`Ll5^3WJ`Lk5^3XJ^Lo5[3TJcLS6V3oIhLX6P3kImL\\6m2gIPM`6h2cIUMd6e2^IYMi6`2ZI]Ml6\\2S3I6IZdb3"}, "label": "glass in front of the pinot bottle"}]} {"id": 204529, "image": "COCO_train2014_000000204529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"glass in front of pinot blanco bottle\"."}], "task": "refering", "answer_template": "The \"glass in front of pinot blanco bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 128, 149, 150, 151, 171, 172, 173, 174, 175, 194, 195, 196, 197, 198, 217, 218, 219, 220, 221, 240, 241, 242, 243, 265, 287, 288, 289, 310, 311, 312, 334], "bbox": [0.466859375, 0.3208125, 0.617515625, 0.8333333333333334], "iscrowd": 0, "area": 16936.011449999995, "rle": {"size": [480, 640], "counts": "S\\\\4=_>b0]Oc0]Oc0]O?B5J6J5L5aFnLi6X3TInLc6W3\\IoL[6V3bIQMh5WOaHn3e1lLj5ZOXHo3l1jLk5^3TJbLk5`3SJaLm5_3SJaLm5`3QJbLn5_3QJaLo5_3QJaLn5a3PJ`LP6a3oI`LP6a3oI_LQ6a3nI`LQ6b3nI^LR6c3lI_LS6a3mI_LS6b3lI^LT6c3jI^LU6c3kI^LT6c3jI^LV6c3iI]LW6c3iI^LU6d3jI\\LV6e3iI[LW6e3iI[LW6f3gI\\LX6e3gI[LY6g3eIYL[6j3bIVL^6l3`IUL_6n3^IRLb6P4[IQLe6R4nHXLR7l5O0000001O0001O01O000000O1[MfIdLZ6[3gIeLY6Z3hIfLX6Z3iIeLW6Z3jIeLW6[3iIeLX6Y3iIgLW6Y3iIgLW6X3jIhLV6X3kIfLV6Y3kIgLU6X3mIgLS6Y3mIgLT6W3nIgLS6Y3mIgLS6X3oIgLQ6Y3oIgLR6W3oIiLQ6V3QJhLP6X3PJhLP6W3RJhLo5W3RJhLn5W3TJhLl5X3UJeLm5Z3VJcLl5\\3UJaLm5^3UJ`Ll5^3WJ`Lk5^3XJ^Lo5[3TJcLS6V3oIhLX6P3kImL\\6m2gIPM`6h2cIUMd6e2^IYMi6`2ZI]Ml6\\2S3I6IZdb3"}, "label": "glass in front of pinot blanco bottle"}]} {"id": 204529, "image": "COCO_train2014_000000204529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a wine bottle with a blue label\"."}], "task": "refering", "answer_template": "The \"a wine bottle with a blue label\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 31, 32, 53, 54, 55, 56, 76, 77, 78, 79, 99, 100, 101, 102, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 168, 169, 170, 171, 172, 192, 193, 194, 215, 216, 217, 238, 239, 240, 241, 261, 262, 263, 264, 284, 285, 286], "bbox": [0.313171875, 0.004479166666666667, 0.489828125, 0.7325416666666666], "iscrowd": 0, "area": 25696.240299999994, "rle": {"size": [480, 640], "counts": "[mm22[>g0[Oe0H8H8H8H5K4L4L4L4L5L3M3L4M5K4K6K5K5K4K6K5K5J5L6^NZKaHl4\\7`1L3M3M4L3L5L3M4L3M4fKYHS1j7eNZH\\1j7ZN\\Hf1l7mMYHR2o7aMWH_2Q8TMTHl2T8gLQHY3W8YLoGg3Y8lKlGT4_91000O10000000O100000O10000000O2O00000000000O3N5K5K5K6J5K5K6I6K5K5K6J5K5K5J7J2N2N2N1O2N2N2M3N1O2N2N2N2N1N3N2L4M3nL^H`Ne7Y1aIfMb6S2cIlM`6l1gIQN]6g1iIXN[6`1kI]Nc6Q1cInNk6?\\I_OQ70UIMZ7@lH?m:L?B[oh4"}, "label": "a wine bottle with a blue label"}]} {"id": 204529, "image": "COCO_train2014_000000204529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"pinot blanc wine bottle\"."}], "task": "refering", "answer_template": "The \"pinot blanc wine bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 31, 32, 53, 54, 55, 56, 76, 77, 78, 79, 99, 100, 101, 102, 122, 123, 124, 125, 126, 145, 146, 147, 148, 149, 168, 169, 170, 171, 172, 192, 193, 194, 215, 216, 217, 238, 239, 240, 241, 261, 262, 263, 264, 284, 285, 286], "bbox": [0.313171875, 0.004479166666666667, 0.489828125, 0.7325416666666666], "iscrowd": 0, "area": 25696.240299999994, "rle": {"size": [480, 640], "counts": "[mm22[>g0[Oe0H8H8H8H5K4L4L4L4L5L3M3L4M5K4K6K5K5K4K6K5K5J5L6^NZKaHl4\\7`1L3M3M4L3L5L3M4L3M4fKYHS1j7eNZH\\1j7ZN\\Hf1l7mMYHR2o7aMWH_2Q8TMTHl2T8gLQHY3W8YLoGg3Y8lKlGT4_91000O10000000O100000O10000000O2O00000000000O3N5K5K5K6J5K5K6I6K5K5K6J5K5K5J7J2N2N2N1O2N2N2M3N1O2N2N2N2N1N3N2L4M3nL^H`Ne7Y1aIfMb6S2cIlM`6l1gIQN]6g1iIXN[6`1kI]Nc6Q1cInNk6?\\I_OQ70UIMZ7@lH?m:L?B[oh4"}, "label": "pinot blanc wine bottle"}]} {"id": 204529, "image": "COCO_train2014_000000204529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white strewn wine bottle\"."}], "task": "refering", "answer_template": "The \"a white strewn wine bottle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [11, 12, 13, 34, 35, 36, 56, 57, 58, 59, 79, 80, 81, 82, 83, 102, 103, 104, 105, 106, 126, 127, 128], "bbox": [0.468265625, 0.002895833333333333, 0.612046875, 0.3837291666666667], "iscrowd": 0, "area": 11501.81795, "rle": {"size": [480, 640], "counts": "if\\4n0]=i0B=L5L3L5K5K4L5L3L5K5K4L5LO01O1O00100O1L3K6K5J6ZOe01000000O1000000O1000000O1000000000000000000000000000000000000000000000000iKTFZ3l9^L]Fa3c9XLfFf3Z9SLnFl3U:O1N1O2O0O2N101N3M2O2M3M2O2M2N3N1N3M3O12O1N2ND=VOi0WOj0VObUd3"}, "label": "a white strewn wine bottle"}]} {"id": 204529, "image": "COCO_train2014_000000204529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a glass on white wine second from the right\"."}], "task": "refering", "answer_template": "The \"a glass on white wine second from the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [107, 108, 129, 130, 131, 132, 152, 153, 154, 155, 175, 176, 177, 178, 198, 199, 200, 201, 221, 222, 223, 224, 245, 246, 267, 268, 269, 290, 291, 292], "bbox": [0.626234375, 0.2747083333333334, 0.764125, 0.728375], "iscrowd": 0, "area": 13295.1295, "rle": {"size": [480, 640], "counts": "gUl5e0`2D^8R1_GVOU8Q1iGSOo7S1oGnNl7X1RHiNh7^1VHcNe7d1WH^Nd7h1ZHXNb7n1\\HSN^7T2`HmM[7Y2cHhMX7^2eHdMV7a2iH`MQ7f2nH[Mo6g2QIZMm6h2RIXMm6i2SIXMk6j2SIWMl6j2TIVMl6k2SIUMl6l2TITMk6n2SISMl6n2TIRMl6o2SIQMm6o2SIQMm6P3QIQMo6o2QIQMo6P3PIPMo6Q3PIPMP7Q3oHPMP7P3PIPMP7Q3nHPMR7_5000000000000O10001O00O1000000eMmHhLT7X3lHhLT7W3nHhLR7W3PIhLP7X3PIhLP7W3RIgLP7X3PIhLP7W3RIhLn6X3SIgLm6X3TIgLm6Y3TIfLl6Y3UIgLk6Y3VIfLj6Y3XIfLh6Z3XIeLi6Z3YIeLh6Z3XIfLh6Y3ZIfLf6Y3\\IeLe6[3[IdLf6[3\\IcLf6\\3ZIcLg6\\3[IbLg6]3ZI`Li6^3XIaLj6^3WI`Lj6_3ZI]Lh6b3[IZLg6d3[201N2M3M4K;E:G:E:G:E:F;AQcV2"}, "label": "a glass on white wine second from the right"}]} {"id": 204529, "image": "COCO_train2014_000000204529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a guy who is wearing orange color shirt\"."}], "task": "refering", "answer_template": "The \"a guy who is wearing orange color shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 29, 30, 31, 32, 33, 34, 36, 52, 53, 54, 56, 59, 60, 75, 76, 79, 83, 98, 99, 106, 121, 122, 144, 145], "bbox": [0.26110937500000003, 0.0024791666666666664, 0.6351875, 0.40741666666666665], "iscrowd": 0, "area": 10699.51935, "rle": {"size": [480, 640], "counts": "d]^22i>7H8I7I7mBWOf;P1RDVOi;Q1oCUOl;R1kCUOo;Q2K5K5K5K5J6K5K5K5K5K5J6K5K5K5K5K5J6K5K5K5O10000000L4ROn0QOo0K5M3M3N2M3N2M3N2M3M3N2M3N2M3N2M2O2M3M3N2M3N2]OcSe0]NcnZO9G9G9G8H8H3M3M3M4L3M3M3M4M2M3M3M4L3M3M3M4L3M3M3M4LN2`N`1K5O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O`fk0YOXZTO:G:E:G:E:G:E:G9F:GL3N6I9H8H7H9H7H9H8G8I8H7HaS]3"}, "label": "a guy who is wearing orange color shirt"}]} {"id": 204529, "image": "COCO_train2014_000000204529.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the left most person\"."}], "task": "refering", "answer_template": "The \"the left most person\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [6, 7, 8, 9, 10, 11, 29, 30, 31, 32, 33, 34, 36, 52, 53, 54, 56, 59, 60, 75, 76, 79, 83, 98, 99, 106, 121, 122, 144, 145], "bbox": [0.26110937500000003, 0.0024791666666666664, 0.6351875, 0.40741666666666665], "iscrowd": 0, "area": 10699.51935, "rle": {"size": [480, 640], "counts": "d]^22i>7H8I7I7mBWOf;P1RDVOi;Q1oCUOl;R1kCUOo;Q2K5K5K5K5J6K5K5K5K5K5J6K5K5K5K5K5J6K5K5K5O10000000L4ROn0QOo0K5M3M3N2M3N2M3N2M3M3N2M3N2M3N2M2O2M3M3N2M3N2]OcSe0]NcnZO9G9G9G8H8H3M3M3M4L3M3M3M4M2M3M3M4L3M3M3M4L3M3M3M4LN2`N`1K5O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O`fk0YOXZTO:G:E:G:E:G:E:G9F:GL3N6I9H8H7H9H7H9H8G8I8H7HaS]3"}, "label": "the left most person"}]} {"id": 491249, "image": "COCO_train2014_000000491249.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a girl in the green dress located in the top right corner of the photo\"."}], "task": "refering", "answer_template": "The \"a girl in the green dress located in the top right corner of the photo\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 135, 136, 137, 158], "bbox": [0.847296875, 0.0, 1.0, 0.4163145539906103], "iscrowd": 0, "area": 11916.858699999997, "rle": {"size": [426, 640], "counts": "lbQ72W=6F5L3L5K4M4K5K4M4K4L5L3N3M3M2N3M2N3N1N3M2N3XFbMa8`2_GcM\\8`2dGcMW8^2iGfMQ8]2oGeMl7]2THgMf7\\2ZHfMa7\\2_HhM]7Y2cHjMZ7U2fHnMX7R2hHQNU7n1kHUNR7l1nHWNo6h1QIYNo6g1QIYNo6g1PIYNP7i1oHWNQ7i1nHWNS7i1mHWNS7i1lHXNS7j1kHVNV7j1jHVNU7k1jHUNW7k1iHUNV7m1hHTNX7l1hHSNX7n1gHSNY7m1gHRNY7P2eHQN[7o1eHQNZ7P2eHPN\\7P2dHPN[7R2cHoM]7Q2cHnM]7S2bHnM^7R2bHmM_7T2_HmMa7S2^HnMb7R2XHSNi7m1RHXNn7a310000O10000O10000O10000O10001O0O10000O10000O1000000O10000O10000O1000S3mLlL"}, "label": "a girl in the green dress located in the top right corner of the photo"}]} {"id": 491249, "image": "COCO_train2014_000000491249.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in a sleeveless top with dark hair and part of a pair of tongs\"."}], "task": "refering", "answer_template": "The \"a woman in a sleeveless top with dark hair and part of a pair of tongs\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [20, 21, 22, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 111, 112, 113, 114, 135, 136, 137, 158], "bbox": [0.847296875, 0.0, 1.0, 0.4163145539906103], "iscrowd": 0, "area": 11916.858699999997, "rle": {"size": [426, 640], "counts": "lbQ72W=6F5L3L5K4M4K5K4M4K4L5L3N3M3M2N3M2N3N1N3M2N3XFbMa8`2_GcM\\8`2dGcMW8^2iGfMQ8]2oGeMl7]2THgMf7\\2ZHfMa7\\2_HhM]7Y2cHjMZ7U2fHnMX7R2hHQNU7n1kHUNR7l1nHWNo6h1QIYNo6g1QIYNo6g1PIYNP7i1oHWNQ7i1nHWNS7i1mHWNS7i1lHXNS7j1kHVNV7j1jHVNU7k1jHUNW7k1iHUNV7m1hHTNX7l1hHSNX7n1gHSNY7m1gHRNY7P2eHQN[7o1eHQNZ7P2eHPN\\7P2dHPN[7R2cHoM]7Q2cHnM]7S2bHnM^7R2bHmM_7T2_HmMa7S2^HnMb7R2XHSNi7m1RHXNn7a310000O10000O10000O10000O10001O0O10000O10000O1000000O10000O10000O1000S3mLlL"}, "label": "a woman in a sleeveless top with dark hair and part of a pair of tongs"}]} {"id": 491249, "image": "COCO_train2014_000000491249.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"blurry person sitting behind pizza\"."}], "task": "refering", "answer_template": "The \"blurry person sitting behind pizza\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 53, 54, 55, 56, 57, 58, 60, 61, 62, 83, 84, 85], "bbox": [0.20746875, 0.007793427230046948, 0.7174843750000001, 0.23896713615023474], "iscrowd": 0, "area": 21441.06235, "rle": {"size": [426, 640], "counts": "[[g1;m<`0@9F3N2N1O2M2O2N2O0100O10000O100O10000O100O10000O10000O10000O1000000O1000000O10000O1000000O1000000O100000000000000000000O100000000000000000000000000001O001O1O001O1VDkN];U1cDPOX;Q1gDTOT;m0kDVOR;j0nDWOQ;j0nDVOR;j0nDWOQ;j0mDWOS;j0lDWOS;i0mDWOS;j0lDWOS;`101O1O1O2N1O2N1O2N1O00000000000000000000000000000000O1000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000D in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 53, 54, 55, 56, 57, 58, 60, 61, 62, 83, 84, 85], "bbox": [0.20746875, 0.007793427230046948, 0.7174843750000001, 0.23896713615023474], "iscrowd": 0, "area": 21441.06235, "rle": {"size": [426, 640], "counts": "[[g1;m<`0@9F3N2N1O2M2O2N2O0100O10000O100O10000O100O10000O10000O10000O1000000O1000000O10000O1000000O1000000O100000000000000000000O100000000000000000000000000001O001O1O001O1VDkN];U1cDPOX;Q1gDTOT;m0kDVOR;j0nDWOQ;j0nDVOR;j0nDWOQ;j0mDWOS;j0lDWOS;i0mDWOS;j0lDWOS;`101O1O1O2N1O2N1O2N1O00000000000000000000000000000000O1000000000000000000000000000000000000O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001O00000000000000000000000000001O00000000000000000000000000D in this image.", "objects": [{"patches": [206, 229, 250, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344, 365, 366, 367, 388, 389, 390], "bbox": [0.871875, 0.4887708333333334, 1.0, 1.0], "iscrowd": 0, "area": 14431.563999999995, "rle": {"size": [480, 640], "counts": "\\kU88d>4M4K4M301O0aDEY8;eGW1k6iNkHU2a6lMTI^2l6bMiHi2W7WMeHm2[7SMcHo2]7QMaHQ3_7PM^HR3b7nL\\HT3d7lL[HU3e7kLYHW3g7jLVHX3j7hLUHY3k7gLQH]3o7cLPH^3P8cLmG_3S8aLkGa3U8_LjGb3V8^LhGd3X8\\LfGf3Z8[LdGf3]8YLaGi3_8WL`Gj3`8WL]Gk3c8UL[Gm3e8TLYGm3g8SLWGo3i8RLTGP4m8Q13M3M3M3M1O001O0000M3N2N2N2M3N2N2M3N2N2M3N2N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2O1O1N2O1O1N2O1N2O1"}, "label": "a silver four door car is closely behind a large truck"}]} {"id": 98038, "image": "COCO_train2014_000000098038.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a silver car stuck in traffic behind an orange and yellow truck , next to a man on a motorcycle and a herd of oxen\"."}], "task": "refering", "answer_template": "The \"a silver car stuck in traffic behind an orange and yellow truck , next to a man on a motorcycle and a herd of oxen\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [206, 229, 250, 251, 252, 273, 274, 275, 296, 297, 298, 319, 320, 321, 342, 343, 344, 365, 366, 367, 388, 389, 390], "bbox": [0.871875, 0.4887708333333334, 1.0, 1.0], "iscrowd": 0, "area": 14431.563999999995, "rle": {"size": [480, 640], "counts": "\\kU88d>4M4K4M301O0aDEY8;eGW1k6iNkHU2a6lMTI^2l6bMiHi2W7WMeHm2[7SMcHo2]7QMaHQ3_7PM^HR3b7nL\\HT3d7lL[HU3e7kLYHW3g7jLVHX3j7hLUHY3k7gLQH]3o7cLPH^3P8cLmG_3S8aLkGa3U8_LjGb3V8^LhGd3X8\\LfGf3Z8[LdGf3]8YLaGi3_8WL`Gj3`8WL]Gk3c8UL[Gm3e8TLYGm3g8SLWGo3i8RLTGP4m8Q13M3M3M3M1O001O0000M3N2N2N2M3N2N2M3N2N2M3N2N2N2O1N2N2O1N2N2O1N2N2O1N2N2O1N2O1O1N2O1O1N2O1N2O1"}, "label": "a silver car stuck in traffic behind an orange and yellow truck , next to a man on a motorcycle and a herd of oxen"}]} {"id": 286483, "image": "COCO_train2014_000000286483.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black snowboard with a woman ' s feet attached to it\"."}], "task": "refering", "answer_template": "The \"a black snowboard with a woman ' s feet attached to it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [245, 246, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 309, 310, 311, 312, 313, 332, 333, 334, 335, 355, 356], "bbox": [0.44662499999999994, 0.5910208333333333, 0.741578125, 0.9280833333333334], "iscrowd": 0, "area": 11067.988, "rle": {"size": [480, 640], "counts": "d^V4a0]>=C2O2M2N3M3M2N3M2N3M3N1N3M2N3M3M2O20O0100O0N3N1N2O1N101O1N101N2O0O2O0001O1O1N101O1O1O001O1O1N101O1O1O001O1O1N101O1O1O001O1O010O1O010O010O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O00010O010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O010O0010O010O01O010N1N3M2N3M2M4M2N3M2N3M2M4M2N3M2N3M2M4M2NWa]2"}, "label": "a black snowboard with a woman ' s feet attached to it"}]} {"id": 286483, "image": "COCO_train2014_000000286483.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black snowboard in the snow\"."}], "task": "refering", "answer_template": "The \"black snowboard in the snow\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [245, 246, 266, 267, 268, 269, 287, 288, 289, 290, 291, 292, 309, 310, 311, 312, 313, 332, 333, 334, 335, 355, 356], "bbox": [0.44662499999999994, 0.5910208333333333, 0.741578125, 0.9280833333333334], "iscrowd": 0, "area": 11067.988, "rle": {"size": [480, 640], "counts": "d^V4a0]>=C2O2M2N3M3M2N3M2N3M3N1N3M2N3M3M2O20O0100O0N3N1N2O1N101O1N101N2O0O2O0001O1O1N101O1O1O001O1O1N101O1O1O001O1O1N101O1O1O001O1O010O1O010O010O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O00010O010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O0010O010O01O010O010O010O0010O010O01O010N1N3M2N3M2M4M2N3M2N3M2M4M2N3M2N3M2M4M2NWa]2"}, "label": "black snowboard in the snow"}]} {"id": 286483, "image": "COCO_train2014_000000286483.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman snowboarder\"."}], "task": "refering", "answer_template": "The \"woman snowboarder\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 282, 283, 284, 285, 286, 287, 288, 289, 309], "bbox": [0.2543125, 0.3107083333333333, 0.68165625, 0.7749166666666666], "iscrowd": 0, "area": 28327.6665, "rle": {"size": [480, 640], "counts": "Wf\\2;b>a0_O`0@`0@3N3L4L3M3N3L3M3M3N2M3M4L3N2M3M3M3M3N3L3M3M3M3N2M4L3M3M3N2M3M4L3M3N2M3K5L5J5K5M300O1O2N1O1O010O1O1O1O010O1O1O1O010O1O1O00100O1001N100000000O10001O0O100000000O200O1O2N100O2N1O1O2O0O1O2N100O2N1O102M6J6J6K5J6J6J2O0O1O100O1O010O1O100O1O010O1O100O1O010O10010O0000000000010O0O1N2O1O1O1N3N1O1N2O1O1O1N3N1O1N2O1O1O0O2O1O1N101O1N2O0O2N2O1N101N2O0O2O1N2O0O2O1N2O0O2O1N2O0O2N2O1N101N2O1N101N2O1N101N2O1N101N2O1N1O2O1N100O100O100O2O0O100O100O100O100O1O100O100O100O100O101N100O100O100O100O1O100O100O100O100O100O100O100O100O010O100O1O100O100O010O101N4M_Zo2"}, "label": "woman snowboarder"}]} {"id": 286483, "image": "COCO_train2014_000000286483.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman with a snowboard sitting down on the hill\"."}], "task": "refering", "answer_template": "The \"a woman with a snowboard sitting down on the hill\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [122, 123, 124, 145, 146, 147, 167, 168, 169, 170, 190, 191, 192, 193, 194, 213, 214, 215, 216, 217, 218, 219, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 282, 283, 284, 285, 286, 287, 288, 289, 309], "bbox": [0.2543125, 0.3107083333333333, 0.68165625, 0.7749166666666666], "iscrowd": 0, "area": 28327.6665, "rle": {"size": [480, 640], "counts": "Wf\\2;b>a0_O`0@`0@3N3L4L3M3N3L3M3M3N2M3M4L3N2M3M3M3M3N3L3M3M3M3N2M4L3M3M3N2M3M4L3M3N2M3K5L5J5K5M300O1O2N1O1O010O1O1O1O010O1O1O1O010O1O1O00100O1001N100000000O10001O0O100000000O200O1O2N100O2N1O1O2O0O1O2N100O2N1O102M6J6J6K5J6J6J2O0O1O100O1O010O1O100O1O010O1O100O1O010O10010O0000000000010O0O1N2O1O1O1N3N1O1N2O1O1O1N3N1O1N2O1O1O0O2O1O1N101O1N2O0O2N2O1N101N2O0O2O1N2O0O2O1N2O0O2O1N2O0O2N2O1N101N2O1N101N2O1N101N2O1N101N2O1N1O2O1N100O100O100O2O0O100O100O100O100O1O100O100O100O100O101N100O100O100O100O1O100O100O100O100O100O100O100O100O010O100O1O100O100O010O101N4M_Zo2"}, "label": "a woman with a snowboard sitting down on the hill"}]} {"id": 212757, "image": "COCO_train2014_000000212757.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a black shirt with one prosthetic leg holding a golf club\"."}], "task": "refering", "answer_template": "The \"a man in a black shirt with one prosthetic leg holding a golf club\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [7, 8, 30, 31, 32, 52, 53, 54, 75, 76, 77, 98, 99, 100, 120, 121, 122, 123, 143, 144, 145, 146, 166, 167, 168, 169, 190, 191, 192, 213, 214, 215, 236, 237, 259, 260, 280, 281, 282, 283, 284, 303, 304, 305], "bbox": [0.20193750000000002, 0.052253521126760565, 0.40834375000000006, 0.9376291079812207], "iscrowd": 0, "area": 20769.89585, "rle": {"size": [426, 640], "counts": "nPf12W=6J4K3N1O2N1O1N3N1O2N1N201N1O2N100O2N1O2O0O1O2N101N1O2N100O2N1lJhNUNZ1e1VOnMl0n1]OkMf0o1AmMa0n1GmM in this image.", "objects": [{"patches": [7, 8, 30, 31, 32, 52, 53, 54, 75, 76, 77, 98, 99, 100, 120, 121, 122, 123, 143, 144, 145, 146, 166, 167, 168, 169, 190, 191, 192, 213, 214, 215, 236, 237, 259, 260, 280, 281, 282, 283, 284, 303, 304, 305], "bbox": [0.20193750000000002, 0.052253521126760565, 0.40834375000000006, 0.9376291079812207], "iscrowd": 0, "area": 20769.89585, "rle": {"size": [426, 640], "counts": "nPf12W=6J4K3N1O2N1O1N3N1O2N1N201N1O2N100O2N1O2O0O1O2N101N1O2N100O2N1lJhNUNZ1e1VOnMl0n1]OkMf0o1AmMa0n1GmM in this image.", "objects": [{"patches": [61, 62, 84, 85, 107, 108, 109, 130, 131, 132, 133, 153, 154, 155, 176, 177, 178, 199, 200, 201, 222, 223, 224, 245, 246, 247, 268, 269, 270], "bbox": [0.648078125, 0.1751643192488263, 0.792640625, 0.7651877934272301], "iscrowd": 0, "area": 11157.694900000002, "rle": {"size": [426, 640], "counts": "[i\\53T=5J6K5K5K5K5XHROm2S1PMQOj2R1UMoNg2V1WMjNi2W1mJbN58m4X1`JXO7AY5W1[JF2SOc5Y1UJO1iNj5X1PJ;O]NY5HWJb1 in this image.", "objects": [{"patches": [61, 62, 84, 85, 107, 108, 109, 130, 131, 132, 133, 153, 154, 155, 176, 177, 178, 199, 200, 201, 222, 223, 224, 245, 246, 247, 268, 269, 270], "bbox": [0.648078125, 0.1751643192488263, 0.792640625, 0.7651877934272301], "iscrowd": 0, "area": 11157.694900000002, "rle": {"size": [426, 640], "counts": "[i\\53T=5J6K5K5K5K5XHROm2S1PMQOj2R1UMoNg2V1WMjNi2W1mJbN58m4X1`JXO7AY5W1[JF2SOc5Y1UJO1iNj5X1PJ;O]NY5HWJb1 in this image.", "objects": [{"patches": [198, 199, 200, 221, 222, 223, 245, 246, 247, 267, 268, 269, 270, 290, 291, 292, 293, 314, 315, 316, 317, 337, 338, 339, 340, 361, 362, 384, 385, 407, 408, 431, 432, 454, 455, 477], "bbox": [0.6360725552050472, 0.367734375, 0.8140851735015773, 0.8752656249999999], "iscrowd": 0, "area": 14379.887549999996, "rle": {"size": [640, 634], "counts": "STl72lc04L5J5l^OB`>b0ZAAd>e0UA_Oi>e0PA_Oo>e0k@_OR?g0g@\\OX?h0b@[O^?h0\\@\\Od?g0T@]Om?e0m_O_OS`0c0g_OBX`0a0a_OE]`0Q2M3N1O2N2N2N1O2N2N2N2N1O2N2N2N1O2N6J9G9G9G9F:G9I2OO1O0O2O1O1N1\\JTBR5m=iJVBX5k=cJXB]5j=^JYBc5Q>2O1O0O2O1O001N2O1O001N2\\DPJU9Q6hFQJX9Q6dFPJ\\9S6`FoI`9S6\\FnIe9U6VFlIk9V6QFlIn9k4_DUL_1QOS:g4bDYLW1ROX:b4dD^Lo0QOf:W4^DiLh0QOS;k3YDUM`0POb;`3QDaM9POo;U3jCmM3oN] in this image.", "objects": [{"patches": [198, 199, 200, 221, 222, 223, 245, 246, 247, 267, 268, 269, 270, 290, 291, 292, 293, 314, 315, 316, 317, 337, 338, 339, 340, 361, 362, 384, 385, 407, 408, 431, 432, 454, 455, 477], "bbox": [0.6360725552050472, 0.367734375, 0.8140851735015773, 0.8752656249999999], "iscrowd": 0, "area": 14379.887549999996, "rle": {"size": [640, 634], "counts": "STl72lc04L5J5l^OB`>b0ZAAd>e0UA_Oi>e0PA_Oo>e0k@_OR?g0g@\\OX?h0b@[O^?h0\\@\\Od?g0T@]Om?e0m_O_OS`0c0g_OBX`0a0a_OE]`0Q2M3N1O2N2N2N1O2N2N2N2N1O2N2N2N1O2N6J9G9G9G9F:G9I2OO1O0O2O1O1N1\\JTBR5m=iJVBX5k=cJXB]5j=^JYBc5Q>2O1O0O2O1O001N2O1O001N2\\DPJU9Q6hFQJX9Q6dFPJ\\9S6`FoI`9S6\\FnIe9U6VFlIk9V6QFlIn9k4_DUL_1QOS:g4bDYLW1ROX:b4dD^Lo0QOf:W4^DiLh0QOS;k3YDUM`0POb;`3QDaM9POo;U3jCmM3oN] in this image.", "objects": [{"patches": [141, 142, 157, 158, 172, 173, 187, 188, 189, 202, 203, 204, 205, 206, 217, 218, 219, 232, 233, 234, 247, 248, 249, 262, 263, 264], "bbox": [0.45158273381294967, 0.5091, 0.7567865707434053, 0.98702], "iscrowd": 0, "area": 9426.8745, "rle": {"size": [500, 417], "counts": "bQl21c?2M4M3M3L4M3L4M2N33M000000O1O001N3N4K5L3L5L4L1N2O1N2O2M2O1N2OR4nKN1O2N1O00O10000O1000000O100001F>[Od0O1O100OM3N2M3N2M01O1O001O001O0M4J6I6K6J5J7K5M4M3M2M4M2N3`MUDR1n;lNoD5T;KgEXO\\:f0Z2O2N2O2M4L2N0001O0000001O0001O01O0000001O0000000010O000001O0000001O0001O01O000000O1M3N2M3N2M\\^a1"}, "label": "a guy in orange and white jersey reaching for a frisbee"}]} {"id": 565018, "image": "COCO_train2014_000000565018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the man in the white and red shirt who caught the frisbee\"."}], "task": "refering", "answer_template": "The \"the man in the white and red shirt who caught the frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [141, 142, 157, 158, 172, 173, 187, 188, 189, 202, 203, 204, 205, 206, 217, 218, 219, 232, 233, 234, 247, 248, 249, 262, 263, 264], "bbox": [0.45158273381294967, 0.5091, 0.7567865707434053, 0.98702], "iscrowd": 0, "area": 9426.8745, "rle": {"size": [500, 417], "counts": "bQl21c?2M4M3M3L4M3L4M2N33M000000O1O001N3N4K5L3L5L4L1N2O1N2O2M2O1N2OR4nKN1O2N1O00O10000O1000000O100001F>[Od0O1O100OM3N2M3N2M01O1O001O001O0M4J6I6K6J5J7K5M4M3M2M4M2N3`MUDR1n;lNoD5T;KgEXO\\:f0Z2O2N2O2M4L2N0001O0000001O0001O01O0000001O0000000010O000001O0000001O0001O01O000000O1M3N2M3N2M\\^a1"}, "label": "the man in the white and red shirt who caught the frisbee"}]} {"id": 565018, "image": "COCO_train2014_000000565018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the shorter guy in yellow shirt trying to catch frisbee\"."}], "task": "refering", "answer_template": "The \"the shorter guy in yellow shirt trying to catch frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 170, 171, 184, 185, 199, 200, 201, 202, 215, 216, 230, 231, 245, 246, 259, 260, 261], "bbox": [0.3144124700239809, 0.5939800000000001, 0.5249640287769785, 0.98524], "iscrowd": 0, "area": 6637.206749999999, "rle": {"size": [500, 417], "counts": "X^P21b?1N101WL9dGHh75UE7o2Fj77SE7P3Ck7S1RHnNm7T1PHnNo7V1YGcNZN8]:[1nFhN^NNc:a1dFlNaNEj:c1]F]Ob9h0VF\\Oj9g0QF[Oo9h0kE[OU:h0fEZOZ:P300KYK]Ef4b:7K5O1O1M3O1O1O1]Oc000O100OZKkFiNGn3^9TMPGlNEo3[9eLeF_Oi0k3b8dLoGY3U8cLlG[3^8[LcGe3^8YLdGf3[8[LeGe3Z8[LhGd3X8[LiGe3V8\\LkGc3U8\\LmGc3S8\\LnGd3R8[LPHd3Q:O0YNRDHo;nNYD60j0h;oN_DN4j0^;WOgEe0Z:ZOlE`0T:@Q3O1O000001O00000O010000O010O10O010O001000O100O100O001O0000000000100O1OYZP3"}, "label": "the shorter guy in yellow shirt trying to catch frisbee"}]} {"id": 565018, "image": "COCO_train2014_000000565018.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in a yellow t - shirt reaching for a frisbee\"."}], "task": "refering", "answer_template": "The \"a man in a yellow t - shirt reaching for a frisbee\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [155, 170, 171, 184, 185, 199, 200, 201, 202, 215, 216, 230, 231, 245, 246, 259, 260, 261], "bbox": [0.3144124700239809, 0.5939800000000001, 0.5249640287769785, 0.98524], "iscrowd": 0, "area": 6637.206749999999, "rle": {"size": [500, 417], "counts": "X^P21b?1N101WL9dGHh75UE7o2Fj77SE7P3Ck7S1RHnNm7T1PHnNo7V1YGcNZN8]:[1nFhN^NNc:a1dFlNaNEj:c1]F]Ob9h0VF\\Oj9g0QF[Oo9h0kE[OU:h0fEZOZ:P300KYK]Ef4b:7K5O1O1M3O1O1O1]Oc000O100OZKkFiNGn3^9TMPGlNEo3[9eLeF_Oi0k3b8dLoGY3U8cLlG[3^8[LcGe3^8YLdGf3[8[LeGe3Z8[LhGd3X8[LiGe3V8\\LkGc3U8\\LmGc3S8\\LnGd3R8[LPHd3Q:O0YNRDHo;nNYD60j0h;oN_DN4j0^;WOgEe0Z:ZOlE`0T:@Q3O1O000001O00000O010000O010O10O010O001000O100O100O001O0000000000100O1OYZP3"}, "label": "a man in a yellow t - shirt reaching for a frisbee"}]} {"id": 114459, "image": "COCO_train2014_000000114459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"bus with orange paint on the back\"."}], "task": "refering", "answer_template": "The \"bus with orange paint on the back\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 256, 257, 258, 260, 261, 262, 263, 264, 265, 266, 279, 280, 281, 283, 284, 285, 286, 287, 288, 302, 303, 304, 306, 307, 308], "bbox": [0.12978125000000001, 0.43370833333333336, 0.56965625, 0.8022500000000001], "iscrowd": 0, "area": 40439.639100000015, "rle": {"size": [480, 640], "counts": "]XW19T>c0^Ob0^Oc0\\Oc0^Ob0]Oc0^Ob0^Ob0N20000O101O000000000000000000001N1000000000000000001O00000O1000000000001O00000000000O100000001O000000000000000O10VOdFQL[9m3mFmKS9R4TGhKl8V4P1O1N2O1N2O1N2O1N101N2O1N2O1M3M3L4M3L4L4M3L4M3L4N2O1N2O1N2O1N2O1O12N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N3M2iDWMS:l3O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00O100000000O100000000O10000000000O100000000O100000000O11O2N1O1O1O2N1O00O1O1N2O1O1N2O1O1N2O1O1O1O1000000O01000O10000O10000O10000O1000000O10000O10000O10000O10000O1000000O10000O10000O10000O10000001O001O00001O001N101O00O1N2O1N2O1O10000000000O10003M2N3L3N3M2N3M2N3L3N3M2N3M2M4RMUhP4"}, "label": "bus with orange paint on the back"}]} {"id": 114459, "image": "COCO_train2014_000000114459.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bus that is white , blue , orange and red\"."}], "task": "refering", "answer_template": "The \"the bus that is white , blue , orange and red\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 256, 257, 258, 260, 261, 262, 263, 264, 265, 266, 279, 280, 281, 283, 284, 285, 286, 287, 288, 302, 303, 304, 306, 307, 308], "bbox": [0.12978125000000001, 0.43370833333333336, 0.56965625, 0.8022500000000001], "iscrowd": 0, "area": 40439.639100000015, "rle": {"size": [480, 640], "counts": "]XW19T>c0^Ob0^Oc0\\Oc0^Ob0]Oc0^Ob0^Ob0N20000O101O000000000000000000001N1000000000000000001O00000O1000000000001O00000000000O100000001O000000000000000O10VOdFQL[9m3mFmKS9R4TGhKl8V4P1O1N2O1N2O1N2O1N101N2O1N2O1M3M3L4M3L4L4M3L4M3L4N2O1N2O1N2O1N2O1O12N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N3M2iDWMS:l3O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O1O00O100000000O100000000O10000000000O100000000O100000000O11O2N1O1O1O2N1O00O1O1N2O1O1N2O1O1N2O1O1O1O1000000O01000O10000O10000O10000O1000000O10000O10000O10000O10000O1000000O10000O10000O10000O10000001O001O00001O001N101O00O1N2O1N2O1O10000000000O10003M2N3L3N3M2N3M2N3L3N3M2N3M2M4RMUhP4"}, "label": "the bus that is white , blue , orange and red"}]} {"id": 171808, "image": "COCO_train2014_000000171808.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the long brown ottoman\"."}], "task": "refering", "answer_template": "The \"the long brown ottoman\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [182, 183, 184, 185, 186, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 248, 249, 250, 251, 252, 253, 254, 268, 269, 270, 271], "bbox": [0.5221458333333333, 0.44851562500000003, 1.0, 0.6950468750000001], "iscrowd": 0, "area": 21330.55215, "rle": {"size": [640, 480], "counts": "QWm45kc05K6J5K5K5K00O1000000O100001O00001O00001O001O00001O00001O00001O00001O00001O00001O01O01O00001O01O01O00001O01O01O001O000010O0001O000010O0001O00001O01O01O00001O01O01O0000001O01O0001O0000001O00001O0`N`N_@a1_?bN`@^1_?dN`@]1^?fN`@Z1_?hNW@HVO`1a`0kNV@GYO^1``0mNT@G[O^1_`0nNR@G^O[1_`0POQ@F@Z1]`0SOP@FBW1]`0UOn_OFEV1[`0VOm_OFHT1Z`0YOj_OFKQ1Y`0\\Oi_OEMQ1X`0\\Oh_OE0o0W`0P2N2O2N1000001N1000001N10001O000O101O0000001O0000001O00001O0000001O0000001O0001O01O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O0000001O00001O01O0001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001nF"}, "label": "the long brown ottoman"}]} {"id": 171808, "image": "COCO_train2014_000000171808.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a dark brown couch in front of a fireplace\"."}], "task": "refering", "answer_template": "The \"a dark brown couch in front of a fireplace\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [182, 183, 184, 185, 186, 199, 200, 201, 202, 203, 213, 214, 215, 216, 217, 218, 219, 220, 230, 231, 232, 233, 234, 235, 236, 237, 248, 249, 250, 251, 252, 253, 254, 268, 269, 270, 271], "bbox": [0.5221458333333333, 0.44851562500000003, 1.0, 0.6950468750000001], "iscrowd": 0, "area": 21330.55215, "rle": {"size": [640, 480], "counts": "QWm45kc05K6J5K5K5K00O1000000O100001O00001O00001O001O00001O00001O00001O00001O00001O00001O01O01O00001O01O01O00001O01O01O001O000010O0001O000010O0001O00001O01O01O00001O01O01O0000001O01O0001O0000001O00001O0`N`N_@a1_?bN`@^1_?dN`@]1^?fN`@Z1_?hNW@HVO`1a`0kNV@GYO^1``0mNT@G[O^1_`0nNR@G^O[1_`0POQ@F@Z1]`0SOP@FBW1]`0UOn_OFEV1[`0VOm_OFHT1Z`0YOj_OFKQ1Y`0\\Oi_OEMQ1X`0\\Oh_OE0o0W`0P2N2O2N1000001N1000001N10001O000O101O0000001O0000001O00001O0000001O0000001O0001O01O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001O0000001O00001O01O0001O0000001O0000001O00001O0000001O0000001O0000001O00001O0000001O0000001nF"}, "label": "a dark brown couch in front of a fireplace"}]} {"id": 343847, "image": "COCO_train2014_000000343847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"there is an elephant standing first from right side\"."}], "task": "refering", "answer_template": "The \"there is an elephant standing first from right side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [112, 113, 114, 131, 132, 133, 134, 135, 136, 137, 154, 155, 156, 157, 158, 159, 160, 176, 177, 178, 179, 180, 181, 182, 183, 199, 200, 201, 202, 203, 204, 205, 206, 222, 223, 224, 225, 226, 227, 228, 229, 245, 246, 247, 248, 249, 250, 251, 252, 268, 269, 270, 271, 272, 273, 291, 292, 293, 294, 314, 315, 316, 317, 337, 338, 339, 340, 360, 361, 362], "bbox": [0.66328125, 0.26741721854304634, 0.9988906249999999, 0.9842604856512142], "iscrowd": 0, "area": 42954.286049999995, "rle": {"size": [453, 640], "counts": "S^l54l=5L4L4K5L4K6K4L4SKXOnKW1m3[OVKU1e4]O]JT1]5_OfIQ1U6BmHn0n6W3K6J5L4M3N2M3M3M3M3M4L3M3M3M3M3M3M3N2O2N1O1O1O1O1O1O1O1O1O1O001O10000O1000000000000000000000000000000O10M3N2M3N2M3M3N2M3N2M3N2M3M3N2M2O2M3M3N2M3N2M3M3N2M3M3L4L4K5dNkHdKY7T4WI_Km6b4[IRKi6n4W1000000001O000000000000001O00000000000O101O000000000000001O000000000000001N10000O100O10001N010O10O10O010O10O010O10O10O010O10O10O010N2B=100O10000O100O10000O10O1000O10000000O01000O10000O0100000O10O10O10000O01000O10000O011O5K5J5^Oc0^Na1_NQ;"}, "label": "there is an elephant standing first from right side"}]} {"id": 343847, "image": "COCO_train2014_000000343847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an elephant reaching out its trunk in thick brush\"."}], "task": "refering", "answer_template": "The \"an elephant reaching out its trunk in thick brush\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 168, 169, 170, 171, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 232, 233, 234, 235, 236, 237, 238, 242, 243, 244, 245, 255, 256, 257, 267, 268, 290, 291, 313, 314, 337, 360], "bbox": [0.084484375, 0.20834437086092714, 0.9836093750000001, 0.9493156732891833], "iscrowd": 0, "area": 57374.83055, "rle": {"size": [453, 640], "counts": "]dh02S>0O2O0O2O0O2O1O0O2O0O2O0O2O0O2O0O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O1O10000O10000O10000O1000O10O10O10O01000O0100O01000O010O0100O0010O0100O0010O01O10O01O010O1O010O010O1O010O00100O00100O00100O00100O0010O01O10O01O10O01O10O01O10O01O10O01O001O1O001O1O0O2O1O001O1O001O1N101O001N2O001N2O001O1N101O1N1TOm0K5L3M4O0O2O1O0O2O1O0O2O1N101O1O001N2O001O001O1N1N3L4L3M4Lf0[O4M2M3O1O1O1O1O0O100000000000O2O1O001O001O10O100O01000O11O00001O0000001O0000001O0000001O0000001O00000000000000O1000000O1000000000000000000000000O10000000M3I700000000000O10XOZGQLg8l3]GRLd8k3`GSL`8k3dGSL]8l3dGSL\\8l3gGRLZ8m3hGQLY8m3jGQLV8o3lGoKU8o3mGPLT8[4bGcK^8^4e0O01eFhKh8W4XGmKe8S4ZGlKg8U4UGnKl8Q4PGRLQ9`42O1N101N2N10O010000OmGXKk6h4SI\\Kk6d4TI_Ki6b4UIaKj6_4UIbKk6]4VIcKj6]4VIcKj6]4VIcKj6]4gHXKJ;^7^4fHZKK8_7]4eH]KK7`7[4dH`KL5`7[4aHcKO2`7Z4aHfKN0a7Z4`HgKOOa7Y4`HiKONa7Y4_HkKOLb7Y4^HlKOLc7X4]HmKOKe7X4[HnK0Je7X4ZHPLOIg7W4YHQLOIh7V4XHRLOHj7V4VHSLOHk7U4UHULOFl7U4UHULNGm7T4THiLl7W3SHjLm7g401O00001O00001O1O1O1O1O00001O0000001O001O1RJaIT4`6iJmIMIV5[6kJnILKW5X6jJQJLIY5W6iJRJLH[5W6fJTJ0CNDS5f6mJVJ1_OMJR5b6mJXJ4\\OKLS5a6lJYJ5ZOKMT5e7QK^HJNU5d7QKjHo4V7QKjHo4V7QKjHo4V7QKjHo4V7QKjHo4V7QKjHo4V7QKjHo4V7QKjHo4V7QKkHn4U7RKkHf0WOZ3n7PLlHe0VO[3n7PLlHb0ZO]3j7QLmH?[O`3l7mKiHa0]Ob3j8[LXGe3h8ZLYGf3h8XLZGg3f8WL\\Gi3e8UL\\Gk3d8TL[Gn3\\92N2N2O1\\JnKS1T4cNZLc0\\4YOlK:]4CiK2_4KhKCf4:`KUOm4i0ZKeNT5X1RKWN\\5f1kJmM^5P2hJfM_5V2iJ`M]5\\2jJZM]5a2U3L5L4K5L4K5M201O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O10O100000O1000O1000O1000O100000O10O100000O10O100000O10O100000000O100000000O100000000O100000000O100000000O20O00000000001O0001O000001O00000001O01O00000000001O000O10000O2O000O10000O2O000O101O0O10001N100O2O00000O2O000O101O0O10001O0O10001N10000O2O001O0O2O000O2O001N101O0O101O001N10001N101O0O2O00001O001O001O00001O001O0000kg4"}, "label": "an elephant reaching out its trunk in thick brush"}]} {"id": 343847, "image": "COCO_train2014_000000343847.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"elephant with outstretched trunk\"."}], "task": "refering", "answer_template": "The \"elephant with outstretched trunk\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 168, 169, 170, 171, 172, 173, 174, 175, 176, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 232, 233, 234, 235, 236, 237, 238, 242, 243, 244, 245, 255, 256, 257, 267, 268, 290, 291, 313, 314, 337, 360], "bbox": [0.084484375, 0.20834437086092714, 0.9836093750000001, 0.9493156732891833], "iscrowd": 0, "area": 57374.83055, "rle": {"size": [453, 640], "counts": "]dh02S>0O2O0O2O0O2O1O0O2O0O2O0O2O0O2O0O1O100O1O100O1O100O1O100O1O100O1O100O1O100O1O1O10000O10000O10000O1000O10O10O10O01000O0100O01000O010O0100O0010O0100O0010O01O10O01O010O1O010O010O1O010O00100O00100O00100O00100O0010O01O10O01O10O01O10O01O10O01O10O01O001O1O001O1O0O2O1O001O1O001O1N101O001N2O001N2O001O1N101O1N1TOm0K5L3M4O0O2O1O0O2O1O0O2O1N101O1O001N2O001O001O1N1N3L4L3M4Lf0[O4M2M3O1O1O1O1O0O100000000000O2O1O001O001O10O100O01000O11O00001O0000001O0000001O0000001O0000001O00000000000000O1000000O1000000000000000000000000O10000000M3I700000000000O10XOZGQLg8l3]GRLd8k3`GSL`8k3dGSL]8l3dGSL\\8l3gGRLZ8m3hGQLY8m3jGQLV8o3lGoKU8o3mGPLT8[4bGcK^8^4e0O01eFhKh8W4XGmKe8S4ZGlKg8U4UGnKl8Q4PGRLQ9`42O1N101N2N10O010000OmGXKk6h4SI\\Kk6d4TI_Ki6b4UIaKj6_4UIbKk6]4VIcKj6]4VIcKj6]4VIcKj6]4gHXKJ;^7^4fHZKK8_7]4eH]KK7`7[4dH`KL5`7[4aHcKO2`7Z4aHfKN0a7Z4`HgKOOa7Y4`HiKONa7Y4_HkKOLb7Y4^HlKOLc7X4]HmKOKe7X4[HnK0Je7X4ZHPLOIg7W4YHQLOIh7V4XHRLOHj7V4VHSLOHk7U4UHULOFl7U4UHULNGm7T4THiLl7W3SHjLm7g401O00001O00001O1O1O1O1O00001O0000001O001O1RJaIT4`6iJmIMIV5[6kJnILKW5X6jJQJLIY5W6iJRJLH[5W6fJTJ0CNDS5f6mJVJ1_OMJR5b6mJXJ4\\OKLS5a6lJYJ5ZOKMT5e7QK^HJNU5d7QKjHo4V7QKjHo4V7QKjHo4V7QKjHo4V7QKjHo4V7QKjHo4V7QKjHo4V7QKjHo4V7QKkHn4U7RKkHf0WOZ3n7PLlHe0VO[3n7PLlHb0ZO]3j7QLmH?[O`3l7mKiHa0]Ob3j8[LXGe3h8ZLYGf3h8XLZGg3f8WL\\Gi3e8UL\\Gk3d8TL[Gn3\\92N2N2O1\\JnKS1T4cNZLc0\\4YOlK:]4CiK2_4KhKCf4:`KUOm4i0ZKeNT5X1RKWN\\5f1kJmM^5P2hJfM_5V2iJ`M]5\\2jJZM]5a2U3L5L4K5L4K5M201O1O1O1O1O1O001O1O1O1O1O001O1O1O1O1O10O100000O1000O1000O1000O100000O10O100000O10O100000O10O100000000O100000000O100000000O100000000O100000000O20O00000000001O0001O000001O00000001O01O00000000001O000O10000O2O000O10000O2O000O101O0O10001N100O2O00000O2O000O101O0O10001O0O10001N10000O2O001O0O2O000O2O001N101O0O101O001N10001N101O0O2O00001O001O001O00001O001O0000kg4"}, "label": "elephant with outstretched trunk"}]} {"id": 32575, "image": "COCO_train2014_000000032575.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier wearing green pants and blue jacket\"."}], "task": "refering", "answer_template": "The \"a skier wearing green pants and blue jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 128, 129, 150, 151, 152, 174, 175, 176, 197, 198, 199, 220, 221, 222, 243, 244, 245], "bbox": [0.51996875, 0.28469483568075116, 0.66875, 0.7200000000000001], "iscrowd": 0, "area": 7754.94285, "rle": {"size": [426, 640], "counts": "PgZ42W=2N101O1O001O1O001O1O0O2O00000O01000O001O1O001O1N2O001O1O0O2O1N1O2O1N1`D^OY:h0\\EFY:d0ZEK[:i1G8H8I7WOdLWGd3c8f0K6J6I7J5K3M1O111N1O2OO0O2O0UN^HmMc7P2dHkM^7P2jHkMW7R2PIiMR7U2SIgMn6Y2UIbMn6]2[IXMg6h2eIjL]6W3l1gLTFQ3S:00001O2N2N2O2M3M5K5L4K4M4L4K21M4L6SLZGi2i9K5K5K6J5I8F;F:E;E;@mmg2"}, "label": "a skier wearing green pants and blue jacket"}]} {"id": 32575, "image": "COCO_train2014_000000032575.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the skier wearing green pants and a blue shirt\"."}], "task": "refering", "answer_template": "The \"the skier wearing green pants and a blue shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [105, 106, 128, 129, 150, 151, 152, 174, 175, 176, 197, 198, 199, 220, 221, 222, 243, 244, 245], "bbox": [0.51996875, 0.28469483568075116, 0.66875, 0.7200000000000001], "iscrowd": 0, "area": 7754.94285, "rle": {"size": [426, 640], "counts": "PgZ42W=2N101O1O001O1O001O1O0O2O00000O01000O001O1O001O1N2O001O1O0O2O1N1O2O1N1`D^OY:h0\\EFY:d0ZEK[:i1G8H8I7WOdLWGd3c8f0K6J6I7J5K3M1O111N1O2OO0O2O0UN^HmMc7P2dHkM^7P2jHkMW7R2PIiMR7U2SIgMn6Y2UIbMn6]2[IXMg6h2eIjL]6W3l1gLTFQ3S:00001O2N2N2O2M3M5K5L4K4M4L4K21M4L6SLZGi2i9K5K5K6J5I8F;F:E;E;@mmg2"}, "label": "the skier wearing green pants and a blue shirt"}]} {"id": 32575, "image": "COCO_train2014_000000032575.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier in a yellow jacket\"."}], "task": "refering", "answer_template": "The \"a skier in a yellow jacket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 99, 100, 101, 123, 124, 125, 146, 147, 169, 170, 191, 192, 193, 214, 215, 237, 238, 239], "bbox": [0.323984375, 0.2024413145539906, 0.4416875, 0.6909624413145541], "iscrowd": 0, "area": 6865.12935, "rle": {"size": [426, 640], "counts": "eif29n<8G5K4M3L4N2O2M2O1N2ULkN`KW1\\4YOUKi0i4AnJ`0o4HlHlNS1^1n51]HmNa1T1P6g0mI[OP6j0lIXOn5P1nIROR6o0kISOT6P1hIQOX6Q1eIQO[6P1bIRO]6o0aISO_6n0VIUMAo1U7mMjHQ3>[Oc6n0ZITOc6P1[IQOd6T1VInNi6j3O1OVOZIRKe6j4cISK]6i4jITKV6g4PJXKQ6g4PJXKP6g4QJZKn5e4TJZKl5f4TJZKl5e4VJZKj5e4WJ[Ki5d4YJ[Kg5e0XIm2Q1^Lg5e0\\Ih2o0bLf5d0`Ie2k0gLe5d0dIZ2o0QM]5>WI]Oa0o2m0UM]5 in this image.", "objects": [{"patches": [77, 78, 99, 100, 101, 123, 124, 125, 146, 147, 169, 170, 191, 192, 193, 214, 215, 237, 238, 239], "bbox": [0.323984375, 0.2024413145539906, 0.4416875, 0.6909624413145541], "iscrowd": 0, "area": 6865.12935, "rle": {"size": [426, 640], "counts": "eif29n<8G5K4M3L4N2O2M2O1N2ULkN`KW1\\4YOUKi0i4AnJ`0o4HlHlNS1^1n51]HmNa1T1P6g0mI[OP6j0lIXOn5P1nIROR6o0kISOT6P1hIQOX6Q1eIQO[6P1bIRO]6o0aISO_6n0VIUMAo1U7mMjHQ3>[Oc6n0ZITOc6P1[IQOd6T1VInNi6j3O1OVOZIRKe6j4cISK]6i4jITKV6g4PJXKQ6g4PJXKP6g4QJZKn5e4TJZKl5f4TJZKl5e4VJZKj5e4WJ[Ki5d4YJ[Kg5e0XIm2Q1^Lg5e0\\Ih2o0bLf5d0`Ie2k0gLe5d0dIZ2o0QM]5>WI]Oa0o2m0UM]5 in this image.", "objects": [{"patches": [267, 268, 269, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316, 333, 334, 335, 336, 337, 338, 339, 356, 357, 358, 359, 360, 361, 381, 382], "bbox": [0.497046875, 0.6768958333333334, 0.7570781249999999, 0.9528333333333333], "iscrowd": 0, "area": 15146.786049999997, "rle": {"size": [480, 640], "counts": "d^e41o>1O1N3N1O2N1N3N1jADf=>VBEi= in this image.", "objects": [{"patches": [267, 268, 269, 289, 290, 291, 292, 293, 311, 312, 313, 314, 315, 316, 333, 334, 335, 336, 337, 338, 339, 356, 357, 358, 359, 360, 361, 381, 382], "bbox": [0.497046875, 0.6768958333333334, 0.7570781249999999, 0.9528333333333333], "iscrowd": 0, "area": 15146.786049999997, "rle": {"size": [480, 640], "counts": "d^e41o>1O1N3N1O2N1N3N1jADf=>VBEi= in this image.", "objects": [{"patches": [264, 265, 280, 281, 282, 283, 297, 298, 299, 300, 314, 315, 316, 317, 331, 332, 333, 334, 347, 348, 349, 350, 351, 364, 365, 366, 367, 368, 369, 381, 382, 383, 384, 385, 386], "bbox": [0.42641666666666667, 0.69121875, 0.7739375000000001, 0.997484375], "iscrowd": 0, "area": 20048.395799999995, "rle": {"size": [640, 480], "counts": "WfP4hMiAY2Q>PNkAQ2o=XNmAi1n=_NnAb1l=fNQB[1i=nNSBS1g=VOVBj0d=_OXBb0e=DWB=i=FSB;m=GPB:P>k2O10000O10000O1000000O10000O10000O1000000O100000000000000000000000000000000000000000000000000000000001O001O1O1O1O001O1O1O1O001jLkA?V>@kA?V>_OmA?S>AnA>S>AoA=R>BoA=R>AQB=P>BQB=o=CRBW1n=nMkAg09Z1m=lMoAe08]1j=lMRB;>h1`=kMWB2b0R2X=iMZBKh0Z2oO2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N1O2N2N2N1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1OS`S2"}, "label": "a woman in a white shirt talking on a phone"}]} {"id": 565070, "image": "COCO_train2014_000000565070.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"woman talking on phone\"."}], "task": "refering", "answer_template": "The \"woman talking on phone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [264, 265, 280, 281, 282, 283, 297, 298, 299, 300, 314, 315, 316, 317, 331, 332, 333, 334, 347, 348, 349, 350, 351, 364, 365, 366, 367, 368, 369, 381, 382, 383, 384, 385, 386], "bbox": [0.42641666666666667, 0.69121875, 0.7739375000000001, 0.997484375], "iscrowd": 0, "area": 20048.395799999995, "rle": {"size": [640, 480], "counts": "WfP4hMiAY2Q>PNkAQ2o=XNmAi1n=_NnAb1l=fNQB[1i=nNSBS1g=VOVBj0d=_OXBb0e=DWB=i=FSB;m=GPB:P>k2O10000O10000O1000000O10000O10000O1000000O100000000000000000000000000000000000000000000000000000000001O001O1O1O1O001O1O1O1O001jLkA?V>@kA?V>_OmA?S>AnA>S>AoA=R>BoA=R>AQB=P>BQB=o=CRBW1n=nMkAg09Z1m=lMoAe08]1j=lMRB;>h1`=kMWB2b0R2X=iMZBKh0Z2oO2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N1O2N2N2N1O2N1O2N1O1O2N1O1O2N1O2N1O1O2N1O1O2N1O1O2N1O1O2N1O1O2N1OS`S2"}, "label": "woman talking on phone"}]} {"id": 24404, "image": "COCO_train2014_000000024404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra with no greenery covering part of him or her\"."}], "task": "refering", "answer_template": "The \"the zebra with no greenery covering part of him or her\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 22, 23, 24, 40, 41, 42, 43, 58, 59, 60, 61, 76, 77, 78, 79, 80, 95, 96, 97, 98, 114, 115, 116, 117, 132, 133, 134, 135, 150, 151, 152, 153, 169, 170, 171, 188], "bbox": [0.2427, 0.04493333333333334, 0.52922, 0.7820266666666668], "iscrowd": 0, "area": 19657.092049999996, "rle": {"size": [375, 500], "counts": "`Z\\13d;l0TOl0TOl0UOl0SO3M1O1O100O1O1O1O100O1O1O1000000001O00000000000000000000000O100000000000000000000000000000000000000000000c0]Oh0XO001O0000001O0000001O00000010O0001O0000001O0000001O00001O0000001O0000001O0000001O01O100O1O1O010N2N2N2N2N1O2O1N2N2N2N2N1O2N2N2O1N2N1O2N2N2M3]LaHh2c7kLnHn2S8I8I6I8H8I6I8I7H7I8I7H7I8IaUf2"}, "label": "the zebra with no greenery covering part of him or her"}]} {"id": 24404, "image": "COCO_train2014_000000024404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"larger zebra on left\"."}], "task": "refering", "answer_template": "The \"larger zebra on left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 22, 23, 24, 40, 41, 42, 43, 58, 59, 60, 61, 76, 77, 78, 79, 80, 95, 96, 97, 98, 114, 115, 116, 117, 132, 133, 134, 135, 150, 151, 152, 153, 169, 170, 171, 188], "bbox": [0.2427, 0.04493333333333334, 0.52922, 0.7820266666666668], "iscrowd": 0, "area": 19657.092049999996, "rle": {"size": [375, 500], "counts": "`Z\\13d;l0TOl0TOl0UOl0SO3M1O1O100O1O1O1O100O1O1O1000000001O00000000000000000000000O100000000000000000000000000000000000000000000c0]Oh0XO001O0000001O0000001O00000010O0001O0000001O0000001O00001O0000001O0000001O0000001O01O100O1O1O010N2N2N2N2N1O2O1N2N2N2N2N1O2N2N2O1N2N1O2N2N2M3]LaHh2c7kLnHn2S8I8I6I8H8I6I8I7H7I8I7H7I8IaUf2"}, "label": "larger zebra on left"}]} {"id": 24404, "image": "COCO_train2014_000000024404.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra in the back side\"."}], "task": "refering", "answer_template": "The \"zebra in the back side\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 24, 25, 26, 27, 28, 29, 30, 31, 32, 44, 45, 46, 47, 48, 49, 50, 62, 63, 64, 66, 67, 68, 82, 84, 85, 86, 100, 103, 104, 118, 121, 122], "bbox": [0.24438, 0.00224, 0.82752, 0.5078666666666667], "iscrowd": 0, "area": 23699.380200000007, "rle": {"size": [375, 500], "counts": "ne\\12_;7N101O001O00001O001O001O001O00001O001O001O00001O001O001O001O001O001O001O1O001O001O1O001O1O001O010O001O00001O001O001O00001O001O001O00001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O1O001aGYNl6g1PI`Nm6a1nHgNn6Y1mHoNP7R1kHROU7n0fHVO[7k0`HXOa7h0[H[Of7f0TH]On7c0nG@S8Q21O1O1N3M2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2O1N2N21O0000001O0000001O0000001O0000001O6J8H9G9[HXMi5b4I6J3M001O1O1O1O1O001O1O1O1OgMQJjNo5R1]JgNb5U1jJdNU5W1WKcNh4Y1dK`N[4[1QL^No3^1]LRNk3j1`LhMg3S2Q3L4L4N2O1O2M2O1O1O1O1O1N21O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O2N3M3M3M3M3M3M3M3M3M3M3M in this image.", "objects": [{"patches": [4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 24, 25, 26, 27, 28, 29, 30, 31, 32, 44, 45, 46, 47, 48, 49, 50, 62, 63, 64, 66, 67, 68, 82, 84, 85, 86, 100, 103, 104, 118, 121, 122], "bbox": [0.24438, 0.00224, 0.82752, 0.5078666666666667], "iscrowd": 0, "area": 23699.380200000007, "rle": {"size": [375, 500], "counts": "ne\\12_;7N101O001O00001O001O001O001O00001O001O001O00001O001O001O001O001O001O001O1O001O001O1O001O1O001O010O001O00001O001O001O00001O001O001O00001O001O001O001O00001O001O001O001O001O001O001O001O001O001O001O001O1O001aGYNl6g1PI`Nm6a1nHgNn6Y1mHoNP7R1kHROU7n0fHVO[7k0`HXOa7h0[H[Of7f0TH]On7c0nG@S8Q21O1O1N3M2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2O1N2N21O0000001O0000001O0000001O0000001O6J8H9G9[HXMi5b4I6J3M001O1O1O1O1O001O1O1O1OgMQJjNo5R1]JgNb5U1jJdNU5W1WKcNh4Y1dK`N[4[1QL^No3^1]LRNk3j1`LhMg3S2Q3L4L4N2O1O2M2O1O1O1O1O1N21O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O001O1O1O1O1O1O2N3M3M3M3M3M3M3M3M3M3M3M in this image.", "objects": [{"patches": [86, 87, 109, 110, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 164, 165, 166, 167, 168, 169, 170, 171, 174, 175, 176, 177, 178, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 239, 259, 260], "bbox": [0.10471875, 0.18969555035128804, 0.9244843749999999, 0.7973067915690866], "iscrowd": 0, "area": 29814.816049999987, "rle": {"size": [427, 640], "counts": "XVl02X=5K4L4L3M1O2M2O2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N1O1N2O2N1O1O1O1O2N1O1O1N3O0O1O1O2N1O1O101N1O1O1O2N1O100O1O2N1O1O1O2O0O1O1O2N1O100O2N1O1O1O1O2O0O1O1O2N1O100O2N100O101N100O100O2O0O10001O0hGgK^7Y4[HnKe7R4UHTLj7n3nGZLQ8`4000001O0000001O0O1000001O0000001O00XOUHPLj7l3\\HSLd7i3`HXL_7d3fH[LZ7e3fH[LZ7e3gHZLY7e3hHZLY7f3hHYLX7g3hHXLY7h3hHWLX7i3iHULW7k3jHSLX7m3iHPLY7P4gHnK[7R4fHkK\\7T4n000WMQGV1o8jNZGm0f8SOdGc0\\8]OmG9S8GWH0i70aHF_7:iH^OW7b0jH\\OW7d0iH\\OW7c0jH]OV7c0jH]OV7c0kH\\OU7d0kH[OV7e0jH[OU7e0lH[OT7e0mHZOS7f0mHYOT7g0lHYOT7g0lHYOT7g0mHXOS7g0nHYOR7g0nHXOS7h0mHXOS7h0nHWOQ7j0oHVOQ7i0PIVOQ7j0oHVOQ7j0PIUOP7k0PIUOP7k0PIUOP7k0PITOQ7k0QITOo6l0QITOo6l0QITOn6m0RIROo6n0RIQOn6n0SIROm6n0SIROm6n0SIROm6n0TIPOm6P1SIPOm6P1SIPOm6o0TIQOl6o0UIoNk6R1UInNk6R1UInNk6R1UInNk6Q1WInNi6R1WImNj6S1VImNj6S1VImNj6S1WIlNi6T1WIkNj6T1WIlNi6T1WIlNh6U1YIjNg6V1YIjNg6V1YIiNh6V1YIjNg6V1g200000000O01000003M4L3M4L00000O10000000000000O100000000O1M3N3M2N21O2N1O2N2OO0G900O1000001O0O100000001N10000000000O2O0000000O10001O000O1000001O0O100000000O2O000000000O101O00000O1000001O0O1003M5K4M0O0000000000O10O1000000000O1000O10O1N2O1N2O1N2O1N101N21O1O1O1O1O1N2O1O001O1O1O1O1O0O2O1O1O001O1O1O001N2O1O001GdNaD]1^;cNbD^1\\;dNcD[1^;eNbDY1`;gN_DX1c;iN\\DU1f;kNZDS1h;nNWDP1k;POUDn0m;SORDk0PlF_OR9>VG_Oh8?_G^Oa8>cGB^88gGGZ85jGKV81nGOR8LSH4n7GVH9^:000O01000000000000O1000000000O10O1000000000000O100000O100000O1000000000000O1000O10000000O1000000000000O0100000000000O100000000000O01000000000000O100O1N2O0OmXd0"}, "label": "a large airplane being boarded"}]} {"id": 368479, "image": "COCO_train2014_000000368479.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"large white plane that is unloading\"."}], "task": "refering", "answer_template": "The \"large white plane that is unloading\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [86, 87, 109, 110, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 164, 165, 166, 167, 168, 169, 170, 171, 174, 175, 176, 177, 178, 186, 187, 188, 189, 190, 191, 209, 210, 211, 212, 213, 214, 215, 233, 234, 235, 236, 237, 238, 239, 259, 260], "bbox": [0.10471875, 0.18969555035128804, 0.9244843749999999, 0.7973067915690866], "iscrowd": 0, "area": 29814.816049999987, "rle": {"size": [427, 640], "counts": "XVl02X=5K4L4L3M1O2M2O2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N1O2M2O2N1O2N1O2N1O2N1O1N2O2N1O1O1O1O2N1O1O1N3O0O1O1O2N1O1O101N1O1O1O2N1O100O1O2N1O1O1O2O0O1O1O2N1O100O2N1O1O1O1O2O0O1O1O2N1O100O2N100O101N100O100O2O0O10001O0hGgK^7Y4[HnKe7R4UHTLj7n3nGZLQ8`4000001O0000001O0O1000001O0000001O00XOUHPLj7l3\\HSLd7i3`HXL_7d3fH[LZ7e3fH[LZ7e3gHZLY7e3hHZLY7f3hHYLX7g3hHXLY7h3hHWLX7i3iHULW7k3jHSLX7m3iHPLY7P4gHnK[7R4fHkK\\7T4n000WMQGV1o8jNZGm0f8SOdGc0\\8]OmG9S8GWH0i70aHF_7:iH^OW7b0jH\\OW7d0iH\\OW7c0jH]OV7c0jH]OV7c0kH\\OU7d0kH[OV7e0jH[OU7e0lH[OT7e0mHZOS7f0mHYOT7g0lHYOT7g0lHYOT7g0mHXOS7g0nHYOR7g0nHXOS7h0mHXOS7h0nHWOQ7j0oHVOQ7i0PIVOQ7j0oHVOQ7j0PIUOP7k0PIUOP7k0PIUOP7k0PITOQ7k0QITOo6l0QITOo6l0QITOn6m0RIROo6n0RIQOn6n0SIROm6n0SIROm6n0SIROm6n0TIPOm6P1SIPOm6P1SIPOm6o0TIQOl6o0UIoNk6R1UInNk6R1UInNk6R1UInNk6Q1WInNi6R1WImNj6S1VImNj6S1VImNj6S1WIlNi6T1WIkNj6T1WIlNi6T1WIlNh6U1YIjNg6V1YIjNg6V1YIiNh6V1YIjNg6V1g200000000O01000003M4L3M4L00000O10000000000000O100000000O1M3N3M2N21O2N1O2N2OO0G900O1000001O0O100000001N10000000000O2O0000000O10001O000O1000001O0O100000000O2O000000000O101O00000O1000001O0O1003M5K4M0O0000000000O10O1000000000O1000O10O1N2O1N2O1N2O1N101N21O1O1O1O1O1N2O1O001O1O1O1O1O0O2O1O1O001O1O1O001N2O1O001GdNaD]1^;cNbD^1\\;dNcD[1^;eNbDY1`;gN_DX1c;iN\\DU1f;kNZDS1h;nNWDP1k;POUDn0m;SORDk0PlF_OR9>VG_Oh8?_G^Oa8>cGB^88gGGZ85jGKV81nGOR8LSH4n7GVH9^:000O01000000000000O1000000000O10O1000000000000O100000O100000O1000000000000O1000O10000000O1000000000000O0100000000000O100000000000O01000000000000O100O1N2O0OmXd0"}, "label": "large white plane that is unloading"}]} {"id": 466790, "image": "COCO_train2014_000000466790.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a giraffe looking to the left\"."}], "task": "refering", "answer_template": "The \"a giraffe looking to the left\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [71, 88, 89, 104, 105, 106, 123, 140, 141, 157, 158, 173, 174, 175, 189, 190, 191, 192, 193, 206, 207, 208, 209, 210, 223, 224, 225, 226, 227, 240, 241, 242, 243, 244, 257, 258, 259, 260, 261, 274, 275, 276, 277, 278, 291, 292, 293, 294, 295, 308, 309, 310, 311, 312, 325, 326, 327, 329, 342, 343, 344, 346, 359, 360, 361, 363, 364, 378], "bbox": [0.11335416666666666, 0.188140625, 0.4265416666666667, 0.968890625], "iscrowd": 0, "area": 33204.28355000001, "rle": {"size": [640, 480], "counts": "SUR11kc0:G:E:VBSOR8W1`GPO\\8Z1WGlNf8^1lFiNR9a1`FfN\\9d1WFbNf9h1lE_NR:k1`E\\N\\:n1WEXNf:R2mDUNP;U2aDRN\\;X2WDnMg;[2lClMP in this image.", "objects": [{"patches": [71, 88, 89, 104, 105, 106, 123, 140, 141, 157, 158, 173, 174, 175, 189, 190, 191, 192, 193, 206, 207, 208, 209, 210, 223, 224, 225, 226, 227, 240, 241, 242, 243, 244, 257, 258, 259, 260, 261, 274, 275, 276, 277, 278, 291, 292, 293, 294, 295, 308, 309, 310, 311, 312, 325, 326, 327, 329, 342, 343, 344, 346, 359, 360, 361, 363, 364, 378], "bbox": [0.11335416666666666, 0.188140625, 0.4265416666666667, 0.968890625], "iscrowd": 0, "area": 33204.28355000001, "rle": {"size": [640, 480], "counts": "SUR11kc0:G:E:VBSOR8W1`GPO\\8Z1WGlNf8^1lFiNR9a1`FfN\\9d1WFbNf9h1lE_NR:k1`E\\N\\:n1WEXNf:R2mDUNP;U2aDRN\\;X2WDnMg;[2lClMP in this image.", "objects": [{"patches": [132, 133, 149, 166, 182, 183, 198, 199, 200, 215, 216, 217, 232, 233, 234, 249, 250, 251, 266, 267, 268, 283, 284, 285], "bbox": [0.6802291666666667, 0.305796875, 0.8570208333333333, 0.727359375], "iscrowd": 0, "area": 8135.529800000001, "rle": {"size": [640, 480], "counts": "of\\6d0Uc0=D8I8L3k_OaNd=c1SBcNl=f1a@QNd0b0i>f1W@oMe0c0S?l1a@]N^?S3O10L3I7\\NP@mN]`0P1S@TN[`0k1m0KY@TNf=_1QA:k>FUA>g>EUA>j>DPA`0o>Fg@?X?U2O1N1000\\OQASLS?h3e0J5N2VOg_OmM[`0R2m_OdMW`0Z2R@\\MQ`0b2f0O1O10O2O1g_OaMT=JgCf2mNXNVX1^BbNf=P1jBjNZ=i0TCSOo<9gCA] in this image.", "objects": [{"patches": [132, 133, 149, 166, 182, 183, 198, 199, 200, 215, 216, 217, 232, 233, 234, 249, 250, 251, 266, 267, 268, 283, 284, 285], "bbox": [0.6802291666666667, 0.305796875, 0.8570208333333333, 0.727359375], "iscrowd": 0, "area": 8135.529800000001, "rle": {"size": [640, 480], "counts": "of\\6d0Uc0=D8I8L3k_OaNd=c1SBcNl=f1a@QNd0b0i>f1W@oMe0c0S?l1a@]N^?S3O10L3I7\\NP@mN]`0P1S@TN[`0k1m0KY@TNf=_1QA:k>FUA>g>EUA>j>DPA`0o>Fg@?X?U2O1N1000\\OQASLS?h3e0J5N2VOg_OmM[`0R2m_OdMW`0Z2R@\\MQ`0b2f0O1O10O2O1g_OaMT=JgCf2mNXNVX1^BbNf=P1jBjNZ=i0TCSOo<9gCA] in this image.", "objects": [{"patches": [194, 195, 196, 197, 198, 199, 200, 201, 202, 217, 218, 219, 240, 241, 242], "bbox": [0.46434375, 0.6221388888888889, 0.8091406250000001, 0.7943888888888889], "iscrowd": 0, "area": 4773.800200000001, "rle": {"size": [360, 640], "counts": "geX36`:c0]Oc0M3O1O1O1O1O1O1O0O10000000000000000000000000000000000000000O1000000000000000iNdFb0\\9^OW100000000000000000000000O1000O1000000000000000000000000000000O1000O1000000000000000000000000000O10000000000000O100000000000000000O100000000000000000000000O10000000O10000000000000000000000000000000O0100000000000000000000000000000000O10000000O100000000000000000000000O1000000000000000O1000000000000000O10000000000000000000001O000000000O100000000000000000001O0IeeZ1"}, "label": "the second bus in from the bottom of the picture"}]} {"id": 147303, "image": "COCO_train2014_000000147303.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"almost hidden bus that is the third one in counting from the bottom up\"."}], "task": "refering", "answer_template": "The \"almost hidden bus that is the third one in counting from the bottom up\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 194, 195, 196, 197, 198, 199, 200, 217], "bbox": [0.43770312499999997, 0.6016666666666667, 0.75565625, 0.7663055555555556], "iscrowd": 0, "area": 1872.2170500000018, "rle": {"size": [360, 640], "counts": "ZfR32o9X1O1O1O1O1O1O1O1OO1000000000O010000gNY100000000O10000000000000000000O10000000000000O1000000000000000000000O1000000000O10000000000000000000000000O100000O100000000000000000000000000000O10O100000000000000000000000000000000O10O100000000000000000000000000000O100000O10000000000000000000000000O1000000000O100000000000000000000000O100000000000O10000000000000000000O1000000000000000O1000000000000000O101O00001O001O\\of1"}, "label": "almost hidden bus that is the third one in counting from the bottom up"}]} {"id": 147303, "image": "COCO_train2014_000000147303.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"red white orange and blue bus that is the second from the building\"."}], "task": "refering", "answer_template": "The \"red white orange and blue bus that is the second from the building\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [176, 177, 178, 194, 195, 196, 197, 198, 199, 200, 217], "bbox": [0.43770312499999997, 0.6016666666666667, 0.75565625, 0.7663055555555556], "iscrowd": 0, "area": 1872.2170500000018, "rle": {"size": [360, 640], "counts": "ZfR32o9X1O1O1O1O1O1O1O1OO1000000000O010000gNY100000000O10000000000000000000O10000000000000O1000000000000000000000O1000000000O10000000000000000000000000O100000O100000000000000000000000000000O10O100000000000000000000000000000000O10O100000000000000000000000000000O100000O10000000000000000000000000O1000000000O100000000000000000000000O100000000000O10000000000000000000O1000000000000000O1000000000000000O101O00001O001O\\of1"}, "label": "red white orange and blue bus that is the second from the building"}]} {"id": 147303, "image": "COCO_train2014_000000147303.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bus in front of others\"."}], "task": "refering", "answer_template": "The \"a bus in front of others\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 270, 271, 272, 273], "bbox": [0.5245, 0.6430277777777778, 0.96284375, 0.88625], "iscrowd": 0, "area": 19974.00385, "rle": {"size": [360, 640], "counts": "U]f3;b: in this image.", "objects": [{"patches": [196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 265, 266, 267, 268, 269, 270, 271, 272, 273], "bbox": [0.5245, 0.6430277777777778, 0.96284375, 0.88625], "iscrowd": 0, "area": 19974.00385, "rle": {"size": [360, 640], "counts": "U]f3;b: in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 169, 170, 171, 172, 173, 188, 189, 190, 191], "bbox": [0.3978, 0.4112275449101796, 0.64098, 0.862934131736527], "iscrowd": 0, "area": 16640.833, "rle": {"size": [334, 500], "counts": "nQQ2 in this image.", "objects": [{"patches": [97, 98, 99, 100, 101, 115, 116, 117, 118, 119, 133, 134, 135, 136, 137, 151, 152, 153, 154, 155, 169, 170, 171, 172, 173, 188, 189, 190, 191], "bbox": [0.3978, 0.4112275449101796, 0.64098, 0.862934131736527], "iscrowd": 0, "area": 16640.833, "rle": {"size": [334, 500], "counts": "nQQ2 in this image.", "objects": [{"patches": [93, 94, 95, 96, 97, 111, 112, 113, 114, 115, 129, 130, 131, 132, 133, 147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 183, 184, 185, 186, 187], "bbox": [0.21015999999999999, 0.41278443113772456, 0.40079999999999993, 0.8779640718562874], "iscrowd": 0, "area": 13906.733050000003, "rle": {"size": [334, 500], "counts": "h]R1j0Y9c1]N`1`N in this image.", "objects": [{"patches": [93, 94, 95, 96, 97, 111, 112, 113, 114, 115, 129, 130, 131, 132, 133, 147, 148, 149, 150, 151, 165, 166, 167, 168, 169, 183, 184, 185, 186, 187], "bbox": [0.21015999999999999, 0.41278443113772456, 0.40079999999999993, 0.8779640718562874], "iscrowd": 0, "area": 13906.733050000003, "rle": {"size": [334, 500], "counts": "h]R1j0Y9c1]N`1`N in this image.", "objects": [{"patches": [145, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216, 235, 236, 237, 238], "bbox": [0.198734375, 0.4597668997668998, 0.419625, 0.7250116550116551], "iscrowd": 0, "area": 12448.562399999999, "rle": {"size": [429, 640], "counts": "m]e14S=>D6M3N1N3M2O1N3M3N3L4L4M4K4L4M3L4L2O1N3M2N2O1O1N2O1O1N3N001N2O1O1N101O1N2O1O0O2O1O1N2O001O1N10000000000O10000000000O1000000000000O100000O1000O1000000000O01000000001O000O1000000000001O0O10000O100O10001N10000O10000O100O2O000O2O001N2N1O2M2O2N1O2M3N1O2N1N3L3M4K5L3M4K4M4L3M4K5L5K4J6J7InVk4"}, "label": "a clear mixing bowl full of chocolate mix"}]} {"id": 16243, "image": "COCO_train2014_000000016243.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the bowl with a spoon sticking out of it with brown frosting in it\"."}], "task": "refering", "answer_template": "The \"the bowl with a spoon sticking out of it with brown frosting in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [145, 165, 166, 167, 168, 169, 170, 188, 189, 190, 191, 192, 193, 212, 213, 214, 215, 216, 235, 236, 237, 238], "bbox": [0.198734375, 0.4597668997668998, 0.419625, 0.7250116550116551], "iscrowd": 0, "area": 12448.562399999999, "rle": {"size": [429, 640], "counts": "m]e14S=>D6M3N1N3M2O1N3M3N3L4L4M4K4L4M3L4L2O1N3M2N2O1O1N2O1O1N3N001N2O1O1N101O1N2O1O0O2O1O1N2O001O1N10000000000O10000000000O1000000000000O100000O1000O1000000000O01000000001O000O1000000000001O0O10000O100O10001N10000O10000O100O2O000O2O001N2N1O2M2O2N1O2M3N1O2N1N3L3M4K5L3M4K4M4L3M4K5L5K4J6J7InVk4"}, "label": "the bowl with a spoon sticking out of it with brown frosting in it"}]} {"id": 434039, "image": "COCO_train2014_000000434039.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"stack of red apples\"."}], "task": "refering", "answer_template": "The \"stack of red apples\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 242, 243, 244, 245, 246, 248, 249, 250, 251, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.43689062500000003, 0.666285046728972, 1.0, 1.0], "iscrowd": 0, "area": 41246.12185, "rle": {"size": [428, 640], "counts": "[^e31Z=1O100O1O1O1O1hLKXI6h6LfGOJ5`8MbG2L2b8L`G6KOe8K]G:LLg8KYG=NIh8LWG?NFk8KTGc00Bl8LPGf00AP9IjFn01\\OU9G`FW17UOY9DWFb1;mN^9a1^FbNb9_1ZFdNe9_1VFdNj9_2O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O100O100O10000O100O10000O100O10000O100O1000000000000000000000000000000000000O1000000000000001O1O001O1O001O1O001O1O001O1O001O1O001O7I8H7I8H2NN2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1O10000000000000000000000000000000000O10000000000000000000000000000000000000000O100001O00001O00001O00001O00001O00001O0000001O001O2N2N2N3M2N2N3M2N2N3M3M3M3M3M3M3M2N000000O100O100O10000O100O10000O100O100\\Od0N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2O1O100O100O100O100O100O100O1000000000000000000000000000000000000000000000000001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O002N3M2N2N2N2N3M2N2N2N3M2N1O1O1O001O1O1O001O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O"}, "label": "stack of red apples"}]} {"id": 434039, "image": "COCO_train2014_000000434039.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bunch of red apples\"."}], "task": "refering", "answer_template": "The \"a bunch of red apples\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [241, 242, 243, 244, 245, 246, 248, 249, 250, 251, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.43689062500000003, 0.666285046728972, 1.0, 1.0], "iscrowd": 0, "area": 41246.12185, "rle": {"size": [428, 640], "counts": "[^e31Z=1O100O1O1O1O1hLKXI6h6LfGOJ5`8MbG2L2b8L`G6KOe8K]G:LLg8KYG=NIh8LWG?NFk8KTGc00Bl8LPGf00AP9IjFn01\\OU9G`FW17UOY9DWFb1;mN^9a1^FbNb9_1ZFdNe9_1VFdNj9_2O1O1O1O1O1O100O1O1O1O1O1O1O100O1O1O1O1O1O1O100O100O100O10000O100O10000O100O10000O100O1000000000000000000000000000000000000O1000000000000001O1O001O1O001O1O001O1O001O1O001O1O001O7I8H7I8H2NN2O1O1O1N2O1O1N2O1O1O1N2O1O1O1N2O1O1O10000000000000000000000000000000000O10000000000000000000000000000000000000000O100001O00001O00001O00001O00001O00001O0000001O001O2N2N2N3M2N2N3M2N2N3M3M3M3M3M3M3M2N000000O100O100O10000O100O10000O100O100\\Od0N2N2N2N2N2N2O1N2N2N2N2N2N2N2N2N2N2O1O100O100O100O100O100O100O1000000000000000000000000000000000000000000000000001O001O001O001O1O001O001O001O001O001O001O1O001O001O001O002N3M2N2N2N2N3M2N2N2N3M2N1O1O1O001O1O1O001O1O1O1O1O1O2N1O1O1O1O1O1O1O1O1O1O1O1O1O"}, "label": "a bunch of red apples"}]} {"id": 384888, "image": "COCO_train2014_000000384888.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a large glass with an amber liquid in it\"."}], "task": "refering", "answer_template": "The \"a large glass with an amber liquid in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [193, 194, 208, 209, 223, 224, 238, 239, 253, 254, 268, 269, 283, 284, 298, 299, 312, 313, 314, 327, 328, 329, 342, 343, 344], "bbox": [0.8442018779342723, 0.534359375, 1.0, 1.0], "iscrowd": 0, "area": 14332.149799999997, "rle": {"size": [640, 426], "counts": "^bQ7m0mb07K5L3M4L4L4L4L4L4L3M4L4M3L4fLcMlDa2n:iMjD[2Q;nMgDV2T;TNcDo1X;[N]Dl1^;]NnCS2l;XN`CW2[ in this image.", "objects": [{"patches": [193, 194, 208, 209, 223, 224, 238, 239, 253, 254, 268, 269, 283, 284, 298, 299, 312, 313, 314, 327, 328, 329, 342, 343, 344], "bbox": [0.8442018779342723, 0.534359375, 1.0, 1.0], "iscrowd": 0, "area": 14332.149799999997, "rle": {"size": [640, 426], "counts": "^bQ7m0mb07K5L3M4L4L4L4L4L4L3M4L4M3L4fLcMlDa2n:iMjD[2Q;nMgDV2T;TNcDo1X;[N]Dl1^;]NnCS2l;XN`CW2[ in this image.", "objects": [{"patches": [32, 33, 54, 55, 56, 57, 58, 77, 78, 79, 80, 100, 101, 102, 103, 104, 123, 124, 125], "bbox": [0.3666875, 0.11235981308411215, 0.560546875, 0.37303738317757007], "iscrowd": 0, "area": 8174.50495, "rle": {"size": [428, 640], "counts": "[ZR3C in this image.", "objects": [{"patches": [32, 33, 54, 55, 56, 57, 58, 77, 78, 79, 80, 100, 101, 102, 103, 104, 123, 124, 125], "bbox": [0.3666875, 0.11235981308411215, 0.560546875, 0.37303738317757007], "iscrowd": 0, "area": 8174.50495, "rle": {"size": [428, 640], "counts": "[ZR3C in this image.", "objects": [{"patches": [245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.521109375, 0.6779205607476635, 0.9240937500000002, 1.0], "iscrowd": 0, "area": 26301.13845, "rle": {"size": [428, 640], "counts": "cP\\41Z=1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1N2O1O1O1O1O1O1O1O1O1O1O1lNkNPFV1o9kNPFV1o9lNoEU1o9nNoES1P:oNmES1R:nNmES1Q:POmEQ1R:QOlEP1S:ROkEo0S:SOlEn0S:TOkEm0T:UOjEl0T:WOjEj0U:XOiEi0U:YOjEh0U:ZOiEg0V:[OhEf0V:]OhEd0W:]OhEd0W:Y1N2O1O1O100O10000O010O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O1000O010000O100O10000O100000000O100000000000000O1000000000000O1000000000000O10000000000000O01000000000000O1000000000000O100000000000000O1000000000000O1000000000000O10000000000001N10000000000000000000000000001O00000000000000000000000000001O000000000000000000000000000O2O00000L4F:F:G9F:G9F:G9F:G9F:G:E_Vd0"}, "label": "black jacket on the bag"}]} {"id": 8063, "image": "COCO_train2014_000000008063.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a piece of luggage with a piece of black clothing folded on top of it\"."}], "task": "refering", "answer_template": "The \"a piece of luggage with a piece of black clothing folded on top of it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [245, 246, 247, 248, 249, 250, 251, 266, 267, 268, 269, 270, 271, 272, 273, 274, 289, 290, 291, 292, 293, 294, 295, 296, 312, 313, 314, 315, 316, 317, 318, 319, 334, 335, 336, 337, 338, 339, 340, 341, 342], "bbox": [0.521109375, 0.6779205607476635, 0.9240937500000002, 1.0], "iscrowd": 0, "area": 26301.13845, "rle": {"size": [428, 640], "counts": "cP\\41Z=1O1O1O1O1O1O1N2O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1N2O1O1O1O1O1O1O1O1O1O1O1lNkNPFV1o9kNPFV1o9lNoEU1o9nNoES1P:oNmES1R:nNmES1Q:POmEQ1R:QOlEP1S:ROkEo0S:SOlEn0S:TOkEm0T:UOjEl0T:WOjEj0U:XOiEi0U:YOjEh0U:ZOiEg0V:[OhEf0V:]OhEd0W:]OhEd0W:Y1N2O1O1O100O10000O010O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O10000O100O1000O010000O100O10000O100000000O100000000000000O1000000000000O1000000000000O10000000000000O01000000000000O1000000000000O100000000000000O1000000000000O1000000000000O10000000000001N10000000000000000000000000001O00000000000000000000000000001O000000000000000000000000000O2O00000L4F:F:G9F:G9F:G9F:G9F:G:E_Vd0"}, "label": "a piece of luggage with a piece of black clothing folded on top of it"}]} {"id": 65407, "image": "COCO_train2014_000000065407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the chair where the woman in a spotted shirt is sitting\"."}], "task": "refering", "answer_template": "The \"the chair where the woman in a spotted shirt is sitting\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 63, 64, 65, 80, 81, 82, 83, 97, 98, 99, 100, 101, 115, 116, 117, 118], "bbox": [0.38689999999999997, 0.2227932960893855, 0.6434, 0.510586592178771], "iscrowd": 0, "area": 5081.61615, "rle": {"size": [358, 500], "counts": "TdS2?g:2M2O1O000000000000000000000O10000000O1O1O6J2N1O2M3M2Ne^29TaM0O001O1O1M3N2XOh0O00100O1O100O00100000000000O1000000O100O1O100O100O1O10000O1000000O[NQOTIP1l6POTIP1l6QOSIo0m6RORIm0o6SOQIm0o6TOPIl0P7TOPIl0P7UOnHl0R7UOlHl0T7VOiHk0W7WO]HS1c7POXHR1h7QOSHQ1m7TOlGn0T8WOeGk0[8P100000000000000000YOg0010O001O0000O10000O1000000N2L4M4K4L4L4K5K6J5K5J6K5K6JmTn1"}, "label": "the chair where the woman in a spotted shirt is sitting"}]} {"id": 65407, "image": "COCO_train2014_000000065407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the red chair the woman in blue jeans is sitting in\"."}], "task": "refering", "answer_template": "The \"the red chair the woman in blue jeans is sitting in\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [46, 63, 64, 65, 80, 81, 82, 83, 97, 98, 99, 100, 101, 115, 116, 117, 118], "bbox": [0.38689999999999997, 0.2227932960893855, 0.6434, 0.510586592178771], "iscrowd": 0, "area": 5081.61615, "rle": {"size": [358, 500], "counts": "TdS2?g:2M2O1O000000000000000000000O10000000O1O1O6J2N1O2M3M2Ne^29TaM0O001O1O1M3N2XOh0O00100O1O100O00100000000000O1000000O100O1O100O100O1O10000O1000000O[NQOTIP1l6POTIP1l6QOSIo0m6RORIm0o6SOQIm0o6TOPIl0P7TOPIl0P7UOnHl0R7UOlHl0T7VOiHk0W7WO]HS1c7POXHR1h7QOSHQ1m7TOlGn0T8WOeGk0[8P100000000000000000YOg0010O001O0000O10000O1000000N2L4M4K4L4L4K5K6J5K5J6K5K6JmTn1"}, "label": "the red chair the woman in blue jeans is sitting in"}]} {"id": 65407, "image": "COCO_train2014_000000065407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red chair that rick walton is sitting on\"."}], "task": "refering", "answer_template": "The \"a red chair that rick walton is sitting on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 141, 142, 143, 159, 160, 161, 170, 177, 178, 179, 186, 187, 188, 189, 190, 191, 195, 196, 197, 203, 204, 205, 206, 207, 208, 209, 210, 211, 214, 215, 222, 223, 224, 225, 226, 227, 228, 229], "bbox": [0.32545999999999997, 0.5481564245810057, 1.0, 1.0], "iscrowd": 0, "area": 19071.51905, "rle": {"size": [358, 500], "counts": "bYi13o:5L3L5K4[F@`8e0SGGi8=kFOQ9U1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O11O001O001O001O001N1000O1N2O1O1O1O1O100O100O100O100O100O1O100O100O10000001O00001O001O00001O00001O001O00001O001O00001O001O00001O00001O001O00001O001O00001O00001O0O2O00001O001O000O2M2O2M2O1O2M2O1N3N1O2O00001O001O0O101O001O00001N10001O001O00001N10001O000000010O01O000010O01nLjNRLV1n3lNPLU1o3lNQLS1o3nNPLS1o3oNoKQ1Q4POnKR1P4oNoKQ1Q4POoKn0Q4TOnKk0S4WOkKi0U4XOjKg0W4XOkKg0U4YOkKf0V4YOkKg0U4YOkKg0U4XOlKh0T4XOmKg0S4XOnKh0R4XOnKh0S4VOnKj0R4VOnKj0R4UOPLj0P4VOPLj0P4VOPLj0P4VOPLj0Q4UOPLj0P4VOPLj0P4VOPLj0Q4UOoKk0Q4TOPLl0Q4SOPLm0o3SOQLm0o3SOQLm0P4ROPLn0P4ROQLn0o3QOQLo0o3QOQLo0P4POPLP1P4POPL?PMOT:NoE2P:LSF3m9LUF3k9KXF4i9IZF6f9I\\F4f9J^FOhf`00ab_O2TGOi83WGMg85YGLd87[GIb8:^GG_8;aGE]8=cGDZ8>eGDW8?iGAU8a0kG@R8b0nG^OP8d0PH]Ol7f0THZOj7h0VHYOg7i0YHWOe7k0[HVOa7m0^HTOa7m0_HTO`7l0`HTO`7l0`HTO_7m0aHTO^7l0bHTO]7m0cHSO]7m0cHTO[7m0dHTO\\7l0dHTO[7m0eHTOZ7P1bHPO]7]201N100O2O000O2O0O101N100O2O000O2O00001O000O2O00001O000O2O00001O00001N10001O00001N1000O1ZOf0ZOf0[Od0[Of0[OaJ"}, "label": "a red chair that rick walton is sitting on"}]} {"id": 65407, "image": "COCO_train2014_000000065407.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"rick walton sitting in a red chair with legs crossed\"."}], "task": "refering", "answer_template": "The \"rick walton sitting in a red chair with legs crossed\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [138, 141, 142, 143, 159, 160, 161, 170, 177, 178, 179, 186, 187, 188, 189, 190, 191, 195, 196, 197, 203, 204, 205, 206, 207, 208, 209, 210, 211, 214, 215, 222, 223, 224, 225, 226, 227, 228, 229], "bbox": [0.32545999999999997, 0.5481564245810057, 1.0, 1.0], "iscrowd": 0, "area": 19071.51905, "rle": {"size": [358, 500], "counts": "bYi13o:5L3L5K4[F@`8e0SGGi8=kFOQ9U1O100O1O100O100O1O100O1O100O100O1O100O1O100O100O11O001O001O001O001N1000O1N2O1O1O1O1O100O100O100O100O100O1O100O100O10000001O00001O001O00001O00001O001O00001O001O00001O001O00001O00001O001O00001O001O00001O00001O0O2O00001O001O000O2M2O2M2O1O2M2O1N3N1O2O00001O001O0O101O001O00001N10001O001O00001N10001O000000010O01O000010O01nLjNRLV1n3lNPLU1o3lNQLS1o3nNPLS1o3oNoKQ1Q4POnKR1P4oNoKQ1Q4POoKn0Q4TOnKk0S4WOkKi0U4XOjKg0W4XOkKg0U4YOkKf0V4YOkKg0U4YOkKg0U4XOlKh0T4XOmKg0S4XOnKh0R4XOnKh0S4VOnKj0R4VOnKj0R4UOPLj0P4VOPLj0P4VOPLj0P4VOPLj0Q4UOPLj0P4VOPLj0P4VOPLj0Q4UOoKk0Q4TOPLl0Q4SOPLm0o3SOQLm0o3SOQLm0P4ROPLn0P4ROQLn0o3QOQLo0o3QOQLo0P4POPLP1P4POPL?PMOT:NoE2P:LSF3m9LUF3k9KXF4i9IZF6f9I\\F4f9J^FOhf`00ab_O2TGOi83WGMg85YGLd87[GIb8:^GG_8;aGE]8=cGDZ8>eGDW8?iGAU8a0kG@R8b0nG^OP8d0PH]Ol7f0THZOj7h0VHYOg7i0YHWOe7k0[HVOa7m0^HTOa7m0_HTO`7l0`HTO`7l0`HTO_7m0aHTO^7l0bHTO]7m0cHSO]7m0cHTO[7m0dHTO\\7l0dHTO[7m0eHTOZ7P1bHPO]7]201N100O2O000O2O0O101N100O2O000O2O00001O000O2O00001O000O2O00001O00001N10001O00001N1000O1ZOf0ZOf0[Od0[Of0[OaJ"}, "label": "rick walton sitting in a red chair with legs crossed"}]} {"id": 466825, "image": "COCO_train2014_000000466825.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the portion of the table largely empty\"."}], "task": "refering", "answer_template": "The \"the portion of the table largely empty\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [227, 228, 235, 240, 241, 242, 243, 250, 251, 252, 253, 254, 258, 265, 266, 267, 268, 269, 273, 274, 280, 281, 282, 283, 284, 288, 295, 296, 297, 298, 299, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.680515625, 1.0, 1.0], "iscrowd": 0, "area": 39898.01525000001, "rle": {"size": [640, 428], "counts": "nc02T>ObG4^8L^G8b8HZGi8CSGa0m8@oFc0Q9]OkFg0U9YOhFj0X9WOgFi0Y9WOgFi0X9YOgFg0Y9YOgFg0Y9ZOfFf0Z9ZOfFf0Z9[OeFe0[9[OeFe0[9[OeFe0Z9]OeFc0[9]OeFc0[9^OdFb0\\9^OdFb0\\9_OcFa0]9_OcFa0\\9AcF?]9AbF`0^9@bF`0^9AaF?_9AaF?_9B`F>`9B`F>_9D`F<`9D`F<`9E_F;a9E_F;a9E_F;a9F^F:b9F^F:a9H^F8b9H^F8b9I]F7c9I]F7c9J\\F6d9J\\F6c9K]F5c9L\\F4d9L\\F4d9M[F3e9M[F3e9NZF2f9NZF2e9O[F1e9N\\F2d9N\\F2d9N\\F2d9N\\F2d9N\\F2c9O]F1c9O\\F2d9N\\F2d9M]F3c9M]F3c9M\\F4d9L\\F4c9M\\F4d9L[F5e9KZF6f9JYF7g9HYF9g9GXF:h9FWF;h9FWF;i9EVFl9BSF?m9@SFa0l9ARF`0n9APF`0P:@oEa0Q:@nE`0R:AlE`0T:@kEa0U:@iEa0V:@iEa0W:@gEa0Y:@eEa0[:_OdEb0\\:_OcEa0]:_OcEa0]:@cE?\\:CdE<\\:DdE<\\:EdE:\\:GdE8\\:HdE8\\:IdE6[:KfE4Z:2`EN`:=VEBj:i0lDVOT;n1hCRNX_OfA`0m`001O1O001O1O001O00001O00001O00001O0000001O00001O0000001O00001O0000001O0000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O100000000000000000000L4E;E;D in this image.", "objects": [{"patches": [227, 228, 235, 240, 241, 242, 243, 250, 251, 252, 253, 254, 258, 265, 266, 267, 268, 269, 273, 274, 280, 281, 282, 283, 284, 288, 295, 296, 297, 298, 299, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344], "bbox": [0.0, 0.680515625, 1.0, 1.0], "iscrowd": 0, "area": 39898.01525000001, "rle": {"size": [640, 428], "counts": "nc02T>ObG4^8L^G8b8HZGi8CSGa0m8@oFc0Q9]OkFg0U9YOhFj0X9WOgFi0Y9WOgFi0X9YOgFg0Y9YOgFg0Y9ZOfFf0Z9ZOfFf0Z9[OeFe0[9[OeFe0[9[OeFe0Z9]OeFc0[9]OeFc0[9^OdFb0\\9^OdFb0\\9_OcFa0]9_OcFa0\\9AcF?]9AbF`0^9@bF`0^9AaF?_9AaF?_9B`F>`9B`F>_9D`F<`9D`F<`9E_F;a9E_F;a9E_F;a9F^F:b9F^F:a9H^F8b9H^F8b9I]F7c9I]F7c9J\\F6d9J\\F6c9K]F5c9L\\F4d9L\\F4d9M[F3e9M[F3e9NZF2f9NZF2e9O[F1e9N\\F2d9N\\F2d9N\\F2d9N\\F2d9N\\F2c9O]F1c9O\\F2d9N\\F2d9M]F3c9M]F3c9M\\F4d9L\\F4c9M\\F4d9L[F5e9KZF6f9JYF7g9HYF9g9GXF:h9FWF;h9FWF;i9EVFl9BSF?m9@SFa0l9ARF`0n9APF`0P:@oEa0Q:@nE`0R:AlE`0T:@kEa0U:@iEa0V:@iEa0W:@gEa0Y:@eEa0[:_OdEb0\\:_OcEa0]:_OcEa0]:@cE?\\:CdE<\\:DdE<\\:EdE:\\:GdE8\\:HdE8\\:IdE6[:KfE4Z:2`EN`:=VEBj:i0lDVOT;n1hCRNX_OfA`0m`001O1O001O1O001O00001O00001O00001O0000001O00001O0000001O00001O0000001O0000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O1000000000000000000000000000000O1000000000000000000000000000000O10000000000000000000000000000O100000000000000000000L4E;E;D in this image.", "objects": [{"patches": [75, 76, 77, 78, 79, 80, 90, 91, 92, 93, 94, 95, 105, 106, 107, 108, 109, 110, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 151, 152, 153, 154], "bbox": [0.0055841121495327104, 0.21240625, 0.3660046728971962, 0.46096875], "iscrowd": 0, "area": 20292.972899999997, "rle": {"size": [640, 428], "counts": "X\\18hc0f2ZM[1fN0O000001O01O000001O01O00000000010O0000000010O00000001O01O00000001O01O000001O0001O0001O00000001O01O000000010O00000000010O0000000010O00000001O0001O000001O0001O0001O00000N2N3M2O1N2N2O101O000000001O000000001O00000000001O000000001O0000N2M3N3L6K4L5J5L4K6K4L4M1O1N2O1O1O1O2M2O1O1O1O1O1N2O2N1O1O1N2O1O1O1]Od0[Od0[Oe0\\OcWY5"}, "label": "the table behind the cake holding food dishes"}]} {"id": 466825, "image": "COCO_train2014_000000466825.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a buffet table behind the wedding cake\"."}], "task": "refering", "answer_template": "The \"a buffet table behind the wedding cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [75, 76, 77, 78, 79, 80, 90, 91, 92, 93, 94, 95, 105, 106, 107, 108, 109, 110, 120, 121, 122, 123, 124, 125, 135, 136, 137, 138, 139, 151, 152, 153, 154], "bbox": [0.0055841121495327104, 0.21240625, 0.3660046728971962, 0.46096875], "iscrowd": 0, "area": 20292.972899999997, "rle": {"size": [640, 428], "counts": "X\\18hc0f2ZM[1fN0O000001O01O000001O01O00000000010O0000000010O00000001O01O00000001O01O000001O0001O0001O00000001O01O000000010O00000000010O0000000010O00000001O0001O000001O0001O0001O00000N2N3M2O1N2N2O101O000000001O000000001O00000000001O000000001O0000N2M3N3L6K4L5J5L4K6K4L4M1O1N2O1O1O1O2M2O1O1O1O1O1N2O2N1O1O1N2O1O1O1]Od0[Od0[Oe0\\OcWY5"}, "label": "a buffet table behind the wedding cake"}]} {"id": 253834, "image": "COCO_train2014_000000253834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the glass of ice water\"."}], "task": "refering", "answer_template": "The \"the glass of ice water\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 46, 47, 66, 67, 68, 69, 88, 89, 90, 91, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 199, 200, 201, 202, 203, 221, 222, 223, 224, 245], "bbox": [0.0, 0.11684640522875818, 0.2337091503267974, 0.5056209150326797], "iscrowd": 0, "area": 24692.77365, "rle": {"size": [612, 612], "counts": "b2W3l?3M4L3N2M3M3M3M3M3M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2NL41O00000000001O00000000001O0000000010O00001O0010O01O001O0010O01O001O00010O2N5K4L4M3L5K4L4L4L5J5L4K6J5K5L4K6J5K5K6K4K5K5K6K4K5K6J5L4K5K6J5K5L5J5K5K5L5J5K5K6K4K5K5KZQh8"}, "label": "the glass of ice water"}]} {"id": 253834, "image": "COCO_train2014_000000253834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a cup of ice water in a clear cup\"."}], "task": "refering", "answer_template": "The \"a cup of ice water in a clear cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [44, 45, 46, 47, 66, 67, 68, 69, 88, 89, 90, 91, 110, 111, 112, 113, 114, 132, 133, 134, 135, 136, 154, 155, 156, 157, 158, 176, 177, 178, 179, 180, 199, 200, 201, 202, 203, 221, 222, 223, 224, 245], "bbox": [0.0, 0.11684640522875818, 0.2337091503267974, 0.5056209150326797], "iscrowd": 0, "area": 24692.77365, "rle": {"size": [612, 612], "counts": "b2W3l?3M4L3N2M3M3M3M3M3M3N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2M3N2N2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2NL41O00000000001O00000000001O0000000010O00001O0010O01O001O0010O01O001O00010O2N5K4L4M3L5K4L4L4L5J5L4K6J5K5L4K6J5K5K6K4K5K5K6K4K5K6J5L4K5K6J5K5L5J5K5K5L5J5K5K6K4K5K5KZQh8"}, "label": "a cup of ice water in a clear cup"}]} {"id": 253834, "image": "COCO_train2014_000000253834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"coffee in a white cup\"."}], "task": "refering", "answer_template": "The \"coffee in a white cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 214, 215, 216, 217, 237, 238, 239], "bbox": [0.7315359477124183, 0.17715686274509804, 0.9910457516339869, 0.4858823529411765], "iscrowd": 0, "area": 20497.181500000002, "rle": {"size": [612, 612], "counts": "nl[82Pc04L4L4L4M3L4L4L4L4M3L4L4L6Jc0^Oa0^Oc0]Oc0]Oc0_O5J2O1N2O001N2O1O0O2O1O1N2O0O2O1O1N101O0O10001N100O10001N10000O101O0O100O101O0O10001N10000O1010O000000000O100O100O100O1O101N100O100O1O1N2N2N2O1N2N2O1N3M2N2O1N2N2O1N2N2N2O1N2N2O1N3M2N2O1N2N2O1N2O1O1O1O1O1N2O2N1O3M2N2N3L3N2N3M2N3M2M3N3M2N3M2N2M4M2N2N3M2N3M2M3N3K4K6K4K5L5K4K5L5J5Lk]3"}, "label": "coffee in a white cup"}]} {"id": 253834, "image": "COCO_train2014_000000253834.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"to go coffee cup\"."}], "task": "refering", "answer_template": "The \"to go coffee cup\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [84, 104, 105, 106, 107, 108, 126, 127, 128, 129, 130, 131, 148, 149, 150, 151, 152, 153, 170, 171, 172, 173, 174, 175, 192, 193, 194, 195, 196, 214, 215, 216, 217, 237, 238, 239], "bbox": [0.7315359477124183, 0.17715686274509804, 0.9910457516339869, 0.4858823529411765], "iscrowd": 0, "area": 20497.181500000002, "rle": {"size": [612, 612], "counts": "nl[82Pc04L4L4L4M3L4L4L4L4M3L4L4L6Jc0^Oa0^Oc0]Oc0]Oc0_O5J2O1N2O001N2O1O0O2O1O1N2O0O2O1O1N101O0O10001N100O10001N10000O101O0O100O101O0O10001N10000O1010O000000000O100O100O100O1O101N100O100O1O1N2N2N2O1N2N2O1N3M2N2O1N2N2O1N2N2N2O1N2N2O1N3M2N2O1N2N2O1N2O1O1O1O1O1N2O2N1O3M2N2N3L3N2N3M2N3M2M3N3M2N3M2N2M4M2N2N3M2N3M2M3N3K4K6K4K5L5K4K5L5J5Lk]3"}, "label": "to go coffee cup"}]} {"id": 106397, "image": "COCO_train2014_000000106397.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"parking meter closest to the orange ticket\"."}], "task": "refering", "answer_template": "The \"parking meter closest to the orange ticket\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 79, 80, 98, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 128, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 282, 283, 284, 285, 286, 287, 288, 289, 290, 305, 306, 307, 308, 309, 310, 311, 312, 313, 328, 329, 330, 331, 332, 333, 334, 335, 336, 352, 353, 354, 355, 356, 357], "bbox": [0.2283125, 0.17078125, 0.64359375, 0.894375], "iscrowd": 0, "area": 75901.07615, "rle": {"size": [512, 640], "counts": "jWY22l?6I7J6J6J6I7J6J6J6J7J5J6F:E;E;E;E;E;E;E;H8I7J6I7J7H7I7J6I7J6I7I7J6I7I7K5L4K5K5K5L5J5K5]IaF]6b9N1O1O100O2N100O1O101N1O1O101N1O100O1O2N100O1O101N1O100O2N1O100O1O2O0O1O1O101N1O100O1O2N100O1O2O0O1O10010O000000001O0000001O01O000001O000000001O0001O0001O0000001O000000010O000000001O0000001O0001O0001O000001O01O01O001O010O001O010O001O00010O001O0010O01O0010O01O000010O01O0O2N1O2O0O2N1O2O0O2N1O1O2O1N2N2N2O2M2N2N2N2O1N2N2N2N3N1N2N2N2O1N2N2N3M2O1N2N2N2O1N2N2N3M2O1N2N2N2N2O1N2N3M2O1N2N2N2N2O1N2N3M2O1N3M4L4L4M3L4L4L4L4M3C=ZOg0ZOe0ZOf0ZOf0B>F:E;E;F:E;E;F:Eiha3"}, "label": "parking meter closest to the orange ticket"}]} {"id": 106397, "image": "COCO_train2014_000000106397.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the closer parking meter\"."}], "task": "refering", "answer_template": "The \"the closer parking meter\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [76, 77, 78, 79, 80, 98, 99, 100, 101, 102, 103, 104, 121, 122, 123, 124, 125, 126, 127, 128, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 282, 283, 284, 285, 286, 287, 288, 289, 290, 305, 306, 307, 308, 309, 310, 311, 312, 313, 328, 329, 330, 331, 332, 333, 334, 335, 336, 352, 353, 354, 355, 356, 357], "bbox": [0.2283125, 0.17078125, 0.64359375, 0.894375], "iscrowd": 0, "area": 75901.07615, "rle": {"size": [512, 640], "counts": "jWY22l?6I7J6J6J6I7J6J6J6J7J5J6F:E;E;E;E;E;E;E;H8I7J6I7J7H7I7J6I7J6I7I7J6I7I7K5L4K5K5K5L5J5K5]IaF]6b9N1O1O100O2N100O1O101N1O1O101N1O100O1O2N100O1O101N1O100O2N1O100O1O2O0O1O1O101N1O100O1O2N100O1O2O0O1O10010O000000001O0000001O01O000001O000000001O0001O0001O0000001O000000010O000000001O0000001O0001O0001O000001O01O01O001O010O001O010O001O00010O001O0010O01O0010O01O000010O01O0O2N1O2O0O2N1O2O0O2N1O1O2O1N2N2N2O2M2N2N2N2O1N2N2N2N3N1N2N2N2O1N2N2N3M2O1N2N2N2O1N2N2N3M2O1N2N2N2N2O1N2N3M2O1N2N2N2N2O1N2N3M2O1N3M4L4L4M3L4L4L4L4M3C=ZOg0ZOe0ZOf0ZOf0B>F:E;E;F:E;E;F:Eiha3"}, "label": "the closer parking meter"}]} {"id": 106397, "image": "COCO_train2014_000000106397.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"left side of parking meter in front of passenger side of truck\"."}], "task": "refering", "answer_template": "The \"left side of parking meter in front of passenger side of truck\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 323, 324, 325], "bbox": [0.005375, 0.19533203125, 0.281609375, 0.8051953125], "iscrowd": 0, "area": 43300.87015, "rle": {"size": [512, 640], "counts": "\\d16h?h1XNh1XNh1XNh1XNn0RO2N2N2N3M2N2N2O2N100O2O0O1O101N100O2O0O100O2O0O100O2O0O101N1000001O0O101O0000001N10001O000O101O00000010O0001O01O0001O01O01O0001O01O000010O000001O0001O01O000010O00000O101N10000O100O2O0O10000O2O0O100O101O0O100O100O2O000O100O2O1N3N1O2M3N1N3N2M2O2M3N1O2M3N1N3N2M3N2N3L3N3L3N2M4M2N3L3N2M4M2M4Mk1UNc0\\O?B;D9H8G9H7I8G5L2M3N2M3N2L4M3M3L4M3M3L4M3L4M3M3L4LmdU7"}, "label": "left side of parking meter in front of passenger side of truck"}]} {"id": 106397, "image": "COCO_train2014_000000106397.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"meter in the left of the other\"."}], "task": "refering", "answer_template": "The \"meter in the left of the other\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [70, 71, 72, 92, 93, 94, 95, 96, 115, 116, 117, 118, 119, 120, 138, 139, 140, 141, 142, 143, 161, 162, 163, 164, 165, 166, 184, 185, 186, 187, 188, 189, 207, 208, 209, 210, 211, 212, 230, 231, 232, 233, 234, 235, 253, 254, 255, 256, 257, 258, 276, 277, 278, 279, 280, 281, 282, 299, 300, 301, 302, 303, 304, 323, 324, 325], "bbox": [0.005375, 0.19533203125, 0.281609375, 0.8051953125], "iscrowd": 0, "area": 43300.87015, "rle": {"size": [512, 640], "counts": "\\d16h?h1XNh1XNh1XNh1XNn0RO2N2N2N3M2N2N2O2N100O2O0O1O101N100O2O0O100O2O0O100O2O0O101N1000001O0O101O0000001N10001O000O101O00000010O0001O01O0001O01O01O0001O01O000010O000001O0001O01O000010O00000O101N10000O100O2O0O10000O2O0O100O101O0O100O100O2O000O100O2O1N3N1O2M3N1N3N2M2O2M3N1O2M3N1N3N2M3N2N3L3N3L3N2M4M2N3L3N2M4M2M4Mk1UNc0\\O?B;D9H8G9H7I8G5L2M3N2M3N2L4M3M3L4M3M3L4M3L4M3M3L4LmdU7"}, "label": "meter in the left of the other"}]} {"id": 155549, "image": "COCO_train2014_000000155549.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a brown sofa in front of the television beside the wall\"."}], "task": "refering", "answer_template": "The \"a brown sofa in front of the television beside the wall\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [199, 200, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 312, 313, 335, 336], "bbox": [0.579984375, 0.5515058823529412, 0.8279375, 0.9806823529411763], "iscrowd": 0, "area": 15157.965300000003, "rle": {"size": [425, 640], "counts": "`Yj46j in this image.", "objects": [{"patches": [199, 200, 220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 312, 313, 335, 336], "bbox": [0.579984375, 0.5515058823529412, 0.8279375, 0.9806823529411763], "iscrowd": 0, "area": 15157.965300000003, "rle": {"size": [425, 640], "counts": "`Yj46j in this image.", "objects": [{"patches": [257, 258, 280, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 328, 329, 330, 331, 332], "bbox": [0.191625, 0.7533882352941177, 0.47339062500000006, 1.0], "iscrowd": 0, "area": 10119.796750000003, "rle": {"size": [425, 640], "counts": "Y\\c12V=1O2N1O2N4L3M3M3L4M3M3M3M3M3M3M3M3M3M3L4O1N0100000O0100000O010000O0100000O0100000O10O10O1000O10O10000O2O1O001N0100O1O001O10O01O1O000010O0001O0010O0001O0010O0001O00010O001O00010O00001O0010O0001O0010O000000001O000000001O000000001O00000000001O000000001O00000000O100O100O10000O100O100O10000O1001O00000000000000000000000000001O000000000000000M3K5K6J6K5J6J6I7J6I6K_`[4"}, "label": "a red recliner viewed through a fisheye lense"}]} {"id": 155549, "image": "COCO_train2014_000000155549.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a single person arm chair in a living room setting with no one sitting in it\"."}], "task": "refering", "answer_template": "The \"a single person arm chair in a living room setting with no one sitting in it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [257, 258, 280, 281, 282, 283, 284, 285, 286, 304, 305, 306, 307, 308, 309, 328, 329, 330, 331, 332], "bbox": [0.191625, 0.7533882352941177, 0.47339062500000006, 1.0], "iscrowd": 0, "area": 10119.796750000003, "rle": {"size": [425, 640], "counts": "Y\\c12V=1O2N1O2N4L3M3M3L4M3M3M3M3M3M3M3M3M3M3L4O1N0100000O0100000O010000O0100000O0100000O10O10O1000O10O10000O2O1O001N0100O1O001O10O01O1O000010O0001O0010O0001O0010O0001O00010O001O00010O00001O0010O0001O0010O000000001O000000001O000000001O00000000001O000000001O00000000O100O100O10000O100O100O10000O1001O00000000000000000000000000001O000000000000000M3K5K6J6K5J6J6I7J6I6K_`[4"}, "label": "a single person arm chair in a living room setting with no one sitting in it"}]} {"id": 319396, "image": "COCO_train2014_000000319396.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a white bidet sits next a toilet in a dark bathroom\"."}], "task": "refering", "answer_template": "The \"a white bidet sits next a toilet in a dark bathroom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 293, 294, 295, 296, 297, 298, 299, 310, 311, 312, 313, 314, 315, 316, 327, 328, 329, 330, 331, 332, 333, 345, 346, 347, 348, 349, 363, 364, 365, 366, 380, 381, 382, 383], "bbox": [0.2576875, 0.6494375, 0.59925, 0.9887656249999999], "iscrowd": 0, "area": 27049.989749999997, "rle": {"size": [640, 480], "counts": "kn]2]1bb04L4L4L3M4L4L3M2N2O2M2N2N2N3M2N2N3M2N2N3M2N1O2N1O2N2N1O2N101O1N101O0O2O001N2O001N101N2O001N101YAVLV=k3bB]L\\=d3^BbLb=_3VBiLi=W3QBoLn=S3jAUMU>\\4O0O2O1O0O2O001N2O001N2O000O1000001N100000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O101O000O100000000O1000000O10O1O1O1O001O1N2O001O1O1O1O0O2O1O1O1O001N2O1O001O1O1O2M6K4D=B>B>I7I7H8I7Ib0^O\\1dNW`g3"}, "label": "a white bidet sits next a toilet in a dark bathroom"}]} {"id": 319396, "image": "COCO_train2014_000000319396.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the yellow toilet\"."}], "task": "refering", "answer_template": "The \"the yellow toilet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [261, 262, 263, 264, 276, 277, 278, 279, 280, 281, 282, 293, 294, 295, 296, 297, 298, 299, 310, 311, 312, 313, 314, 315, 316, 327, 328, 329, 330, 331, 332, 333, 345, 346, 347, 348, 349, 363, 364, 365, 366, 380, 381, 382, 383], "bbox": [0.2576875, 0.6494375, 0.59925, 0.9887656249999999], "iscrowd": 0, "area": 27049.989749999997, "rle": {"size": [640, 480], "counts": "kn]2]1bb04L4L4L3M4L4L3M2N2O2M2N2N2N3M2N2N3M2N2N3M2N1O2N1O2N2N1O2N101O1N101O0O2O001N2O001N101N2O001N101YAVLV=k3bB]L\\=d3^BbLb=_3VBiLi=W3QBoLn=S3jAUMU>\\4O0O2O1O0O2O001N2O001N2O000O1000001N100000000O1000000O100000000O1000000O100000000O1000000O100000000O1000000O100000000O101O000O100000000O1000000O10O1O1O1O001O1N2O001O1O1O1O0O2O1O1O1O001N2O1O001O1O1O2M6K4D=B>B>I7I7H8I7Ib0^O\\1dNW`g3"}, "label": "the yellow toilet"}]} {"id": 319396, "image": "COCO_train2014_000000319396.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"western toilet in bathroom\"."}], "task": "refering", "answer_template": "The \"western toilet in bathroom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 143, 144, 145, 146, 147, 160, 161, 162, 163, 164, 177, 178, 179, 180, 181, 193, 194, 195, 196, 197, 198, 207, 208, 209, 210, 211, 212, 213, 214, 215, 223, 224, 225, 226, 227, 228, 229, 230, 240, 241, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 276, 277], "bbox": [0.16441666666666668, 0.323703125, 0.6926458333333334, 0.7286249999999999], "iscrowd": 0, "area": 30851.43695000001, "rle": {"size": [640, 480], "counts": "`ga1b0]c0g310O01O00100O001O10O01O1O0010O01O1O010O1O10000O1000O10O10000O10000O10O1000O10000O1000O10O10000O1000000O01000O10000O1000O10O10000O1000N2O0O2O1O0O2@`0ZOf0YOf0[Of0I7O001N2O001O1O1O001O1O001O1O1O001O1O001O1O0011O1O1N101O1N2O1O001N2O1O0O2O1O1O1N101O1N2O001N2O00O10O0100O0012N1O2M3N1M4I6I8I6I8I6G:F9F;E:M4N1N3N000O010000O01000O10O10O10O10O1000O010000O01000O01000O10O1000001N10000O2O00000O2O0J6I7H9G8H8I7H=C`0@VXl2"}, "label": "western toilet in bathroom"}]} {"id": 319396, "image": "COCO_train2014_000000319396.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue toilet in a bathroom\"."}], "task": "refering", "answer_template": "The \"a blue toilet in a bathroom\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 127, 143, 144, 145, 146, 147, 160, 161, 162, 163, 164, 177, 178, 179, 180, 181, 193, 194, 195, 196, 197, 198, 207, 208, 209, 210, 211, 212, 213, 214, 215, 223, 224, 225, 226, 227, 228, 229, 230, 240, 241, 242, 243, 244, 245, 246, 247, 259, 260, 261, 262, 263, 276, 277], "bbox": [0.16441666666666668, 0.323703125, 0.6926458333333334, 0.7286249999999999], "iscrowd": 0, "area": 30851.43695000001, "rle": {"size": [640, 480], "counts": "`ga1b0]c0g310O01O00100O001O10O01O1O0010O01O1O010O1O10000O1000O10O10000O10000O10O1000O10000O1000O10O10000O1000000O01000O10000O1000O10O10000O1000N2O0O2O1O0O2@`0ZOf0YOf0[Of0I7O001N2O001O1O1O001O1O001O1O1O001O1O001O1O0011O1O1N101O1N2O1O001N2O1O0O2O1O1O1N101O1N2O001N2O00O10O0100O0012N1O2M3N1M4I6I8I6I8I6G:F9F;E:M4N1N3N000O010000O01000O10O10O10O10O1000O010000O01000O01000O10O1000001N10000O2O00000O2O0J6I7H9G8H8I7H=C`0@VXl2"}, "label": "a blue toilet in a bathroom"}]} {"id": 270248, "image": "COCO_train2014_000000270248.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man in orange jersey and blue helmet\"."}], "task": "refering", "answer_template": "The \"a man in orange jersey and blue helmet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 295, 312, 313, 314, 315, 316, 317, 318, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.546515625, 0.6016745283018868, 0.959296875, 1.0], "iscrowd": 0, "area": 25601.5823, "rle": {"size": [424, 640], "counts": "gZa41W=0O1O100O1O1O100O1O100O1O1O100O1O100bMMfG4S8n0QGSOg8Y1UGgNd8b1YG_N`8k1\\GVN`8P2]GQN`8T2]GmM`8X2^GhM_8]2^GdM_8b2]G_Ma8e2\\G\\Ma8i2\\GXMb8l2\\GTMc8o2ZGRMe8i300O1O1O1O1O1O1O100O1O100O1O100O1O100O100O1000000O10000O1000000O10000O10000O1000000O1000000O1000000000000000000O100000000000000000000O10000001O0000001O0000001O0000001O0000001O0000001O00001O001O001O001O001O001O001O001O001O001O1O001O1O001O1]N]G_Nc8_1_GaNb8]1_GcNb8[1`GdN`8Z1bGfN_8Y1aGgN`8X1`GhNb8V1_GiNb8V1^GjNd8T1\\GlNe8S1\\GlNf8R1ZGnNg8Q1YGoNi8o0WGQOk8m0VGROl8l0TGTOn8j0RGVOP9_OSGOMb0R9\\OSG1Lb0R9\\OSG1Kc0T9YOTG2He0V9VOTG4Ge0b9jNhF`0Ff0`:XO`Eh0a:WO_Ei0c:UO^Ej0c:UO]Ek0d:SO]Em0d:QO^En0c:PO^EP1c:nN^ER1d:kN^ET1c:jN^EV1[;O00001O001O001O00001O001O00001O002N2N1O2N2N1O2N0000001O0000001O000000001O000000001O0000001O000000001O0000001O000000001O0000001O0000001O1O1O001O1O1O1O1O1O1O1OY[:"}, "label": "a man in orange jersey and blue helmet"}]} {"id": 270248, "image": "COCO_train2014_000000270248.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"catcher of baseball game\"."}], "task": "refering", "answer_template": "The \"catcher of baseball game\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [220, 221, 222, 223, 224, 225, 243, 244, 245, 246, 247, 248, 266, 267, 268, 269, 270, 271, 289, 290, 291, 292, 293, 294, 295, 312, 313, 314, 315, 316, 317, 318, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343], "bbox": [0.546515625, 0.6016745283018868, 0.959296875, 1.0], "iscrowd": 0, "area": 25601.5823, "rle": {"size": [424, 640], "counts": "gZa41W=0O1O100O1O1O100O1O100O1O1O100O1O100bMMfG4S8n0QGSOg8Y1UGgNd8b1YG_N`8k1\\GVN`8P2]GQN`8T2]GmM`8X2^GhM_8]2^GdM_8b2]G_Ma8e2\\G\\Ma8i2\\GXMb8l2\\GTMc8o2ZGRMe8i300O1O1O1O1O1O1O100O1O100O1O100O1O100O100O1000000O10000O1000000O10000O10000O1000000O1000000O1000000000000000000O100000000000000000000O10000001O0000001O0000001O0000001O0000001O0000001O00001O001O001O001O001O001O001O001O001O001O1O001O1O001O1]N]G_Nc8_1_GaNb8]1_GcNb8[1`GdN`8Z1bGfN_8Y1aGgN`8X1`GhNb8V1_GiNb8V1^GjNd8T1\\GlNe8S1\\GlNf8R1ZGnNg8Q1YGoNi8o0WGQOk8m0VGROl8l0TGTOn8j0RGVOP9_OSGOMb0R9\\OSG1Lb0R9\\OSG1Kc0T9YOTG2He0V9VOTG4Ge0b9jNhF`0Ff0`:XO`Eh0a:WO_Ei0c:UO^Ej0c:UO]Ek0d:SO]Em0d:QO^En0c:PO^EP1c:nN^ER1d:kN^ET1c:jN^EV1[;O00001O001O001O00001O001O00001O002N2N1O2N2N1O2N0000001O0000001O000000001O000000001O0000001O000000001O0000001O000000001O0000001O0000001O1O1O001O1O1O1O1O1O1O1OY[:"}, "label": "catcher of baseball game"}]} {"id": 270248, "image": "COCO_train2014_000000270248.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a baseball player in an orange shirt\"."}], "task": "refering", "answer_template": "The \"a baseball player in an orange shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [21, 22, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 112, 113, 114, 135, 136, 137, 157, 158, 159, 160, 181, 183], "bbox": [0.85821875, 0.006721698113207547, 0.99975, 0.4871698113207547], "iscrowd": 0, "area": 11547.842899999994, "rle": {"size": [424, 640], "counts": "W`S73Q=6J7I7mL1dH2j6c0PIAj6g0PI\\Om6j0kG;R8KhG8W8LbG9[8Z2O2O1N2N2O0O00100O1OO3J5K6K5M2N3M2N3M2N3M2N3M2N3L3MCcLUG[3o8lLiFR3Y9>00000000000012N4L4L5K4L5K5K4L5L3Le0\\O2M3N2M3N2N00O1^JkHo4V7oJlHQ5T7lJoHS5T7hJnHY5a7O10OO2N2N101PNaHTNa7^1RI[No6\\1_I^Nc6a1cIXN_6f1hIRNZ6m1b2O1N103L5K5K5K5I7E;D in this image.", "objects": [{"patches": [21, 22, 43, 44, 45, 65, 66, 67, 68, 88, 89, 90, 91, 112, 113, 114, 135, 136, 137, 157, 158, 159, 160, 181, 183], "bbox": [0.85821875, 0.006721698113207547, 0.99975, 0.4871698113207547], "iscrowd": 0, "area": 11547.842899999994, "rle": {"size": [424, 640], "counts": "W`S73Q=6J7I7mL1dH2j6c0PIAj6g0PI\\Om6j0kG;R8KhG8W8LbG9[8Z2O2O1N2N2O0O00100O1OO3J5K6K5M2N3M2N3M2N3M2N3M2N3L3MCcLUG[3o8lLiFR3Y9>00000000000012N4L4L5K4L5K5K4L5L3Le0\\O2M3N2M3N2N00O1^JkHo4V7oJlHQ5T7lJoHS5T7hJnHY5a7O10OO2N2N101PNaHTNa7^1RI[No6\\1_I^Nc6a1cIXN_6f1hIRNZ6m1b2O1N103L5K5K5K5I7E;D in this image.", "objects": [{"patches": [77, 78, 79, 80, 81, 100, 101, 102, 103, 104, 122, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 178, 188, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 224, 225, 226, 233, 234, 235, 236, 239, 240, 241, 242, 243, 248, 249, 250, 258, 259, 260, 261, 262, 263, 264, 265, 266, 271, 272, 273, 274, 282, 283, 284, 285, 286, 287, 288, 295, 296, 297, 306, 307, 308, 309, 310, 318, 319, 320, 342, 343], "bbox": [0.1393125, 0.19485849056603774, 0.95446875, 0.9733726415094339], "iscrowd": 0, "area": 59992.34240000001, "rle": {"size": [424, 640], "counts": "ZSU12V=6J6J7I6I7J2N1O00000O101O000000001N100000001O0O1000001O00000O101O0000000O2O0000001O0O101O1O1O1O1O1001O1O001O1O001O1O1O001O1O00000000000000000000000000000000O1ROoNdEQ1Y:QOhEo0V:ROkEn0R:UOnEk0o9WORFi0k9YOVFg0g9[OZFe0c9]O^Fc0_9_ObFa0[9AfF?W9CjF=T9DmFi4eN\\Lm0kN?g4fN]Lk0mN?d4gN`Li0mN`0a4iNaLf0oNb0_4hNbLf0POb0\\4jNcLd0ROb0Y4kNeLb0SOd0V4lNfL`0UOd0S4mNhL?VOd0Q4nNhL>WOe0o3nNjLU1i2h1iLSMTOhe;"}, "label": "baseball player ` gibson \u00b4."}]} {"id": 270248, "image": "COCO_train2014_000000270248.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man leans in to bunt a ball\"."}], "task": "refering", "answer_template": "The \"a man leans in to bunt a ball\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [77, 78, 79, 80, 81, 100, 101, 102, 103, 104, 122, 123, 124, 125, 126, 127, 128, 129, 130, 146, 147, 148, 149, 150, 151, 152, 153, 154, 170, 171, 172, 173, 174, 175, 176, 177, 178, 188, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 224, 225, 226, 233, 234, 235, 236, 239, 240, 241, 242, 243, 248, 249, 250, 258, 259, 260, 261, 262, 263, 264, 265, 266, 271, 272, 273, 274, 282, 283, 284, 285, 286, 287, 288, 295, 296, 297, 306, 307, 308, 309, 310, 318, 319, 320, 342, 343], "bbox": [0.1393125, 0.19485849056603774, 0.95446875, 0.9733726415094339], "iscrowd": 0, "area": 59992.34240000001, "rle": {"size": [424, 640], "counts": "ZSU12V=6J6J7I6I7J2N1O00000O101O000000001N100000001O0O1000001O00000O101O0000000O2O0000001O0O101O1O1O1O1O1001O1O001O1O001O1O1O001O1O00000000000000000000000000000000O1ROoNdEQ1Y:QOhEo0V:ROkEn0R:UOnEk0o9WORFi0k9YOVFg0g9[OZFe0c9]O^Fc0_9_ObFa0[9AfF?W9CjF=T9DmFi4eN\\Lm0kN?g4fN]Lk0mN?d4gN`Li0mN`0a4iNaLf0oNb0_4hNbLf0POb0\\4jNcLd0ROb0Y4kNeLb0SOd0V4lNfL`0UOd0S4mNhL?VOd0Q4nNhL>WOe0o3nNjLU1i2h1iLSMTOhe;"}, "label": "a man leans in to bunt a ball"}]} {"id": 573360, "image": "COCO_train2014_000000573360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man with a baseball bat in a field\"."}], "task": "refering", "answer_template": "The \"a man with a baseball bat in a field\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 124, 125, 126, 127, 139, 140, 141, 142, 153, 154, 155, 156, 168, 169, 170, 183, 184, 185, 186, 198, 199, 200, 201, 213, 214, 215, 216, 228, 230, 231, 245, 246], "bbox": [0.19983606557377048, 0.30060937499999996, 0.5437704918032786, 0.72003125], "iscrowd": 0, "area": 15713.40835, "rle": {"size": [640, 427], "counts": "_Uf14kc03M2N1O2N2O0O2N2N1J7F:G8H9bM]NTBl1i=ZNQBj1l=[NoAg1n=`NlAb1Q>dNiA`1S>fNhA\\1V>iNeAY1X>nNaAU1\\>QO_AR1]>TO]Ao0a>VOYAm0d>ZOUAi0i>Y2M2O[OdAhKZ>Q4RBlKm=P4[BmKc=S4bBkK\\=T4W110O0010O01N101N1O2N101N1O2N101N1O21N3N2M4M2M3N3M2N2N3M6J6J7QBdKm;b4mC`KQbD\\1S2XOkNmN_:a0cD;IF_2;jNQO]:c0aD6LK_27jNTOZ:d0aD2OOd2NfN[OV:g0aDN01n2C_NFR:i0aDJ14W3XOXN0P:j0`DG36j`04T_OB39k`05T_O^O2=k`06T_OYO2a0l`06S_OUO2e0l`05d_OJ^`04c_OL^`02c_ON^`00c_OO``0Na_O2a`0IZ_OGUO?ea0BY_Og0Pb000001N2N2O11O1O2N1O1N2N3N1N2O1N2N101N2N101O010O10O01O1O1N4M3Mmci3"}, "label": "a man with a baseball bat in a field"}]} {"id": 573360, "image": "COCO_train2014_000000573360.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man holding a baseball bat\"."}], "task": "refering", "answer_template": "The \"a man holding a baseball bat\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [110, 111, 124, 125, 126, 127, 139, 140, 141, 142, 153, 154, 155, 156, 168, 169, 170, 183, 184, 185, 186, 198, 199, 200, 201, 213, 214, 215, 216, 228, 230, 231, 245, 246], "bbox": [0.19983606557377048, 0.30060937499999996, 0.5437704918032786, 0.72003125], "iscrowd": 0, "area": 15713.40835, "rle": {"size": [640, 427], "counts": "_Uf14kc03M2N1O2N2O0O2N2N1J7F:G8H9bM]NTBl1i=ZNQBj1l=[NoAg1n=`NlAb1Q>dNiA`1S>fNhA\\1V>iNeAY1X>nNaAU1\\>QO_AR1]>TO]Ao0a>VOYAm0d>ZOUAi0i>Y2M2O[OdAhKZ>Q4RBlKm=P4[BmKc=S4bBkK\\=T4W110O0010O01N101N1O2N101N1O2N101N1O21N3N2M4M2M3N3M2N2N3M6J6J7QBdKm;b4mC`KQbD\\1S2XOkNmN_:a0cD;IF_2;jNQO]:c0aD6LK_27jNTOZ:d0aD2OOd2NfN[OV:g0aDN01n2C_NFR:i0aDJ14W3XOXN0P:j0`DG36j`04T_OB39k`05T_O^O2=k`06T_OYO2a0l`06S_OUO2e0l`05d_OJ^`04c_OL^`02c_ON^`00c_OO``0Na_O2a`0IZ_OGUO?ea0BY_Og0Pb000001N2N2O11O1O2N1O1N2N3N1N2O1N2N101N2N101O010O10O01O1O1N4M3Mmci3"}, "label": "a man holding a baseball bat"}]} {"id": 163764, "image": "COCO_train2014_000000163764.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man drinking through a straw , and holding a pink box of pink , sprinkled doughnuts\"."}], "task": "refering", "answer_template": "The \"a man drinking through a straw , and holding a pink box of pink , sprinkled doughnuts\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [39, 40, 41, 42, 43, 44, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 90, 91, 92, 93, 94, 95, 108, 109, 110, 111, 112, 113, 125, 126, 127, 128, 129, 130, 131, 132, 133, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.030660980810234544, 0.083140625, 0.9812793176972281, 0.9887656250000001], "iscrowd": 0, "area": 178415.14834999994, "rle": {"size": [640, 469], "counts": "PX91dc0=B>C=C=B>CC=C=B>CC=B>CC=B>L3N3L4M3L4M2N3M3N2N2N2N1O2N2N2N2N1O1O1O000O2O001O00001O00001O001O00001O002N2N1O2N2M2O001O00000000001O0000000000001O0000000O2O001N1O1O2N1O2N1O101N1O2N1O1O2N1O2N100O2N1O001O1O1O1O10O01RHgDT7Z;hHjDW7W;eHmD[7S;bHPE]7P;`HTE`7l:]HWEb7j:ZHZEe7a;00O1O100O1O100O100O1O100O1O100O1O1bLdCXO\\ in this image.", "objects": [{"patches": [39, 40, 41, 42, 43, 44, 56, 57, 58, 59, 60, 61, 73, 74, 75, 76, 77, 78, 90, 91, 92, 93, 94, 95, 108, 109, 110, 111, 112, 113, 125, 126, 127, 128, 129, 130, 131, 132, 133, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388], "bbox": [0.030660980810234544, 0.083140625, 0.9812793176972281, 0.9887656250000001], "iscrowd": 0, "area": 178415.14834999994, "rle": {"size": [640, 469], "counts": "PX91dc0=B>C=C=B>CC=C=B>CC=B>CC=B>L3N3L4M3L4M2N3M3N2N2N2N1O2N2N2N2N1O1O1O000O2O001O00001O00001O001O00001O002N2N1O2N2M2O001O00000000001O0000000000001O0000000O2O001N1O1O2N1O2N1O101N1O2N1O1O2N1O2N100O2N1O001O1O1O1O10O01RHgDT7Z;hHjDW7W;eHmD[7S;bHPE]7P;`HTE`7l:]HWEb7j:ZHZEe7a;00O1O100O1O100O100O1O100O1O100O1O1bLdCXO\\ in this image.", "objects": [{"patches": [120, 121, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 259, 260, 280, 282, 283, 302, 303, 306, 326], "bbox": [0.15, 0.298875, 0.34045312499999997, 0.8292083333333333], "iscrowd": 0, "area": 13303.6771, "rle": {"size": [480, 640], "counts": "Z\\]11n>1O2N1N2O2M2O1N2O1O1O1M3K5L4L4L4K6K4L4L4L4K5L4J6bL\\NZIj1\\6eNXIb1e6fNPI`1m6eNjHW1^7lN[HX1e7kNUHY1j7jNPHY1Q8jNjGW1X8jNcGX1`8hN\\GZ1f8fNUG\\1m8eNnF]1U9cNfF_1]9o12N3M2O2M3M2N3M3M2N3M3M2N3M2N3N2M2NI700O1O21O000O2O00001O0O2011O00000000000000001O0001O1O001O1O1O2N8I6J6J6K0OO2O0O2OT1kNk0VOb0]OO2SOoFhKS9R1oFn08nMj8R1SGi07SNh8S1TGc0:WNd8U1UG=<\\Na8V1VG7?`N]8X1WG1b0dNY8Z1XGKd0jNU8Z1ZGFf0mNR8\\1\\G_Og0SOo7]1]GYOj0WOk7_1^GSOd:l0_EmNf:R1]EhNg:W1b1O9G8F5L4L3I8GRcU6"}, "label": "a man on a bicycle"}]} {"id": 352185, "image": "COCO_train2014_000000352185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man on a bike\"."}], "task": "refering", "answer_template": "The \"a man on a bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [120, 121, 142, 143, 144, 145, 165, 166, 167, 168, 188, 189, 190, 191, 211, 212, 213, 214, 234, 235, 236, 237, 257, 258, 259, 260, 280, 282, 283, 302, 303, 306, 326], "bbox": [0.15, 0.298875, 0.34045312499999997, 0.8292083333333333], "iscrowd": 0, "area": 13303.6771, "rle": {"size": [480, 640], "counts": "Z\\]11n>1O2N1N2O2M2O1N2O1O1O1M3K5L4L4L4K6K4L4L4L4K5L4J6bL\\NZIj1\\6eNXIb1e6fNPI`1m6eNjHW1^7lN[HX1e7kNUHY1j7jNPHY1Q8jNjGW1X8jNcGX1`8hN\\GZ1f8fNUG\\1m8eNnF]1U9cNfF_1]9o12N3M2O2M3M2N3M3M2N3M3M2N3M2N3N2M2NI700O1O21O000O2O00001O0O2011O00000000000000001O0001O1O001O1O1O2N8I6J6J6K0OO2O0O2OT1kNk0VOb0]OO2SOoFhKS9R1oFn08nMj8R1SGi07SNh8S1TGc0:WNd8U1UG=<\\Na8V1VG7?`N]8X1WG1b0dNY8Z1XGKd0jNU8Z1ZGFf0mNR8\\1\\G_Og0SOo7]1]GYOj0WOk7_1^GSOd:l0_EmNf:R1]EhNg:W1b1O9G8F5L4L3I8GRcU6"}, "label": "a man on a bike"}]} {"id": 352185, "image": "COCO_train2014_000000352185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a bus with a red front is in behind a man on a bike\"."}], "task": "refering", "answer_template": "The \"a bus with a red front is in behind a man on a bike\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 139, 140, 141, 142, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 162, 163, 164, 165, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 185, 186, 187, 188, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 208, 209, 210, 211, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 231, 232, 233, 234, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 254, 255, 256, 257, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 290, 291, 292], "bbox": [0.042140625, 0.06741666666666667, 0.758421875, 0.7730416666666666], "iscrowd": 0, "area": 106891.64335, "rle": {"size": [480, 640], "counts": "Sib3]4]KdKb01Q4]4ZKQL5Da4\\4WKeLCQOV5\\4SKQNm4P2PKRNP5P2mJQNS5P2jJRNV5P2fJRNZ5o1dJRN\\5P2`JRN_5P2^JRNb5P2ZJRNf5d500000000000000O1000000000000000000O10000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1001O000000001O000000001O00000000001O000000001O000000010O00000000001O000000001O000000001O000000001O000kF]Jm8k500000000000000000000000000000000000000000000000000001O00000000001O004L4L4M3L2N1O1O001O1O1O1O1O1O1O1O1O1O00O100O100O10000O100O100O10000O100O100O1_OhJZGX5f8lJVGT5j8oJRGR5n8QKoFo4P9a001O0000000O100000000O1000006J7I8G8I7I6J6J6J6I6K6mLnGlNX8KQIOU7iNSJQ1S6gMUKS2S9H7J6J6J7I6I7J6J7I6JdZX2"}, "label": "a bus with a red front is in behind a man on a bike"}]} {"id": 352185, "image": "COCO_train2014_000000352185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the big red bus\"."}], "task": "refering", "answer_template": "The \"the big red bus\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 139, 140, 141, 142, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 162, 163, 164, 165, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 185, 186, 187, 188, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 208, 209, 210, 211, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 231, 232, 233, 234, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 254, 255, 256, 257, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 290, 291, 292], "bbox": [0.042140625, 0.06741666666666667, 0.758421875, 0.7730416666666666], "iscrowd": 0, "area": 106891.64335, "rle": {"size": [480, 640], "counts": "Sib3]4]KdKb01Q4]4ZKQL5Da4\\4WKeLCQOV5\\4SKQNm4P2PKRNP5P2mJQNS5P2jJRNV5P2fJRNZ5o1dJRN\\5P2`JRN_5P2^JRNb5P2ZJRNf5d500000000000000O1000000000000000000O10000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1000000000000000000O1001O000000001O000000001O00000000001O000000001O000000010O00000000001O000000001O000000001O000000001O000kF]Jm8k500000000000000000000000000000000000000000000000000001O00000000001O004L4L4M3L2N1O1O001O1O1O1O1O1O1O1O1O1O00O100O100O10000O100O100O10000O100O100O1_OhJZGX5f8lJVGT5j8oJRGR5n8QKoFo4P9a001O0000000O100000000O1000006J7I8G8I7I6J6J6J6I6K6mLnGlNX8KQIOU7iNSJQ1S6gMUKS2S9H7J6J6J7I6I7J6J7I6JdZX2"}, "label": "the big red bus"}]} {"id": 352185, "image": "COCO_train2014_000000352185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black suv stopped in traffic\"."}], "task": "refering", "answer_template": "The \"a black suv stopped in traffic\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 160, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298], "bbox": [0.7483906250000001, 0.37904166666666667, 1.0, 0.7424791666666667], "iscrowd": 0, "area": 23265.834749999998, "rle": {"size": [480, 640], "counts": "`iP7>A6K9F?B>A;F4K4M4K4M4K4M2N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2OM3H8001O1O1O1O1O1O1N2O001O1O1000000O10O100000O10000000i0WO00000000000000000O0100000000000000000000O1000000000000000000000O01000000N2O1O1N2O1O1O1N2O1O1000000000000000O1000000000000000000000000000000000000000000O100000000000000000000001O1O2N1O1O\\2dMhG"}, "label": "a black suv stopped in traffic"}]} {"id": 352185, "image": "COCO_train2014_000000352185.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"black suv\"."}], "task": "refering", "answer_template": "The \"black suv\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [157, 158, 159, 160, 179, 180, 181, 182, 183, 201, 202, 203, 204, 205, 206, 224, 225, 226, 227, 228, 229, 247, 248, 249, 250, 251, 252, 270, 271, 272, 273, 274, 275, 293, 294, 295, 296, 297, 298], "bbox": [0.7483906250000001, 0.37904166666666667, 1.0, 0.7424791666666667], "iscrowd": 0, "area": 23265.834749999998, "rle": {"size": [480, 640], "counts": "`iP7>A6K9F?B>A;F4K4M4K4M4K4M2N2O1N2O1O1N2O1N2O1O1N2O1N2O1O1N2O1N2O1N2O1O1N2O1N2OM3H8001O1O1O1O1O1O1N2O001O1O1000000O10O100000O10000000i0WO00000000000000000O0100000000000000000000O1000000000000000000000O01000000N2O1O1N2O1O1O1N2O1O1000000000000000O1000000000000000000000000000000000000000000O100000000000000000000001O1O2N1O1O\\2dMhG"}, "label": "black suv"}]} {"id": 32707, "image": "COCO_train2014_000000032707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skier in orange coat in the right hand picture\"."}], "task": "refering", "answer_template": "The \"skier in orange coat in the right hand picture\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 140, 141, 154, 155, 156, 168, 169, 170, 183, 184, 185, 198, 199, 200, 212, 213, 214, 215], "bbox": [0.18082159624413147, 0.38615625, 0.45176056338028164, 0.655796875], "iscrowd": 0, "area": 8236.101349999999, "rle": {"size": [640, 426], "counts": "c``12ec0628H1N3N1O2M2O2N1N3N2N1N3N1O2M2O1O1N101O1N101N2M2O2NSOZ^ODb`0Gi_Og0H_O_`02`_Oa03[O\\`0=W_O9>ROZO6Pa0g1e_OWN[OOo`0k1e_OZN\\OIn`0m1e_O]N\\`0c1c_O`N[`0`1c_OcN[`0_1c_OdN[`0\\1d_OfN[`0Z1d_OhN[`0X1d_OkNY`0V1e_OmNZ`0S1e_OoNZ`0R1d_OQOZ`0o0e_OSOY`0n0f_OTOY`0l0f_OWOX`0i0f_OZOX`0g0g_O[OX`0X2N1O1N3N1O2N1N2O2N1O1N@Z@kLd?T3^@lLa?S3a@mL^?R3d@oLY?Q3j@mLV?R3l@nLS?Q3o@nLP?R3SAlLm>S3UAmLj>R3XAmLh>R3ZAnLd>R3_AlLb>R3`AmL`>e1f@\\Ol0oN]>e1j@XOl0QOZ>f1n@UOj0UOW>f1RAQOi0XOV>e1UAoNg0[OT>f1XAkNg0^OP>g1\\AfNg0Bm=g1`AcNe0Ek=h1cA_Ne0Hh=h1fA\\Nd0Ke=i1kAXNb0Ob=i1l2100O101N11O100O1O100O1O1O100O1O100Fi]OjNXb0U1l]OgNUb0W1;N1N3N2N2N2N2N2N5K7I7Jmoa4"}, "label": "skier in orange coat in the right hand picture"}]} {"id": 32707, "image": "COCO_train2014_000000032707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skier with black pants and orange jacket facing another skier in a green ski suit\"."}], "task": "refering", "answer_template": "The \"skier with black pants and orange jacket facing another skier in a green ski suit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [126, 140, 141, 154, 155, 156, 168, 169, 170, 183, 184, 185, 198, 199, 200, 212, 213, 214, 215], "bbox": [0.18082159624413147, 0.38615625, 0.45176056338028164, 0.655796875], "iscrowd": 0, "area": 8236.101349999999, "rle": {"size": [640, 426], "counts": "c``12ec0628H1N3N1O2M2O2N1N3N2N1N3N1O2M2O1O1N101O1N101N2M2O2NSOZ^ODb`0Gi_Og0H_O_`02`_Oa03[O\\`0=W_O9>ROZO6Pa0g1e_OWN[OOo`0k1e_OZN\\OIn`0m1e_O]N\\`0c1c_O`N[`0`1c_OcN[`0_1c_OdN[`0\\1d_OfN[`0Z1d_OhN[`0X1d_OkNY`0V1e_OmNZ`0S1e_OoNZ`0R1d_OQOZ`0o0e_OSOY`0n0f_OTOY`0l0f_OWOX`0i0f_OZOX`0g0g_O[OX`0X2N1O1N3N1O2N1N2O2N1O1N@Z@kLd?T3^@lLa?S3a@mL^?R3d@oLY?Q3j@mLV?R3l@nLS?Q3o@nLP?R3SAlLm>S3UAmLj>R3XAmLh>R3ZAnLd>R3_AlLb>R3`AmL`>e1f@\\Ol0oN]>e1j@XOl0QOZ>f1n@UOj0UOW>f1RAQOi0XOV>e1UAoNg0[OT>f1XAkNg0^OP>g1\\AfNg0Bm=g1`AcNe0Ek=h1cA_Ne0Hh=h1fA\\Nd0Ke=i1kAXNb0Ob=i1l2100O101N11O100O1O100O1O1O100O1O100Fi]OjNXb0U1l]OgNUb0W1;N1N3N2N2N2N2N2N5K7I7Jmoa4"}, "label": "skier with black pants and orange jacket facing another skier in a green ski suit"}]} {"id": 32707, "image": "COCO_train2014_000000032707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a skier in a green outfit\"."}], "task": "refering", "answer_template": "The \"a skier in a green outfit\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 144, 145, 146, 158, 159, 160, 161, 174, 175, 189, 190, 205, 206, 220, 221], "bbox": [0.5619718309859155, 0.354265625, 0.7809154929577465, 0.6477031249999999], "iscrowd": 0, "area": 6623.915749999999, "rle": {"size": [640, 426], "counts": "Wee41mc05K5K3O10O00000001O000001O01O0000001O00000000000O100000000^NCa_O=[`0Hd_O8X`0Nf_O2Y`00e_O1[`00d_O0[`02d_ON\\`03c_OM]`04b_OLM@\\?g0e@IC1c?7X@g1b?[NY@n1b?TNY@T2b?mMZ@Z2a?V1M3M4YA[Kk=h4RBXKn=k4nAWKQ>k4mAUKS>l4kAVKT>j4gASKL4\\>j4gATKK2^>k4eA\\KZ>e4eA[K\\>d4dA]K\\>c4cA^K]>R510O0100O10001N2O2M3`MeAVO]>b0jA]OW>`0mA]OV>?nA_OS>=RB@P>=SBAP>;TBAo=Bn@@X1k0l=\\OYABP1n0j=WObADi0Q1g=ROhCj0k?L4L3M1O001O1O001O1O001M3L4M2N3Mb[j1"}, "label": "a skier in a green outfit"}]} {"id": 32707, "image": "COCO_train2014_000000032707.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"someone in a green ski waiting to ski\"."}], "task": "refering", "answer_template": "The \"someone in a green ski waiting to ski\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [129, 130, 144, 145, 146, 158, 159, 160, 161, 174, 175, 189, 190, 205, 206, 220, 221], "bbox": [0.5619718309859155, 0.354265625, 0.7809154929577465, 0.6477031249999999], "iscrowd": 0, "area": 6623.915749999999, "rle": {"size": [640, 426], "counts": "Wee41mc05K5K3O10O00000001O000001O01O0000001O00000000000O100000000^NCa_O=[`0Hd_O8X`0Nf_O2Y`00e_O1[`00d_O0[`02d_ON\\`03c_OM]`04b_OLM@\\?g0e@IC1c?7X@g1b?[NY@n1b?TNY@T2b?mMZ@Z2a?V1M3M4YA[Kk=h4RBXKn=k4nAWKQ>k4mAUKS>l4kAVKT>j4gASKL4\\>j4gATKK2^>k4eA\\KZ>e4eA[K\\>d4dA]K\\>c4cA^K]>R510O0100O10001N2O2M3`MeAVO]>b0jA]OW>`0mA]OV>?nA_OS>=RB@P>=SBAP>;TBAo=Bn@@X1k0l=\\OYABP1n0j=WObADi0Q1g=ROhCj0k?L4L3M1O001O1O001O1O001M3L4M2N3Mb[j1"}, "label": "someone in a green ski waiting to ski"}]} {"id": 524227, "image": "COCO_train2014_000000524227.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a sofa behind an eating boy\"."}], "task": "refering", "answer_template": "The \"a sofa behind an eating boy\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [102, 103, 124, 125, 126, 127, 129, 130, 131, 132, 133, 134, 135, 146, 147, 150, 152, 153, 154, 155, 156, 157, 158, 169, 170], "bbox": [0.36529687499999997, 0.28125, 0.8776875000000001, 0.42995833333333333], "iscrowd": 0, "area": 6986.471700000002, "rle": {"size": [480, 640], "counts": "ok]32m>2M3N2M3N2M3N2N2M3N2M3N3M1N2O1N2O1O0O2O1N2O1O0O2O1N2O1OTOjBOU=MSC1l\\110O00010O01O100O1000000001O0000000000000000001N100O100O100O100O101O0O100O100O100O101N1001O0000001O0000010O0000001O1O1O1O2N100O1O1O2N1O1O1O1O1O2O0O1O1O1O2N1O1O1O101N1O1O1O001OCYgc0XObY\\O:F:H8O1001O0001O000001O00000000001O00000000001O00000000001O0001O000001O0000000000001O00000000001O0000000ZOiAc0W>\\OnA`0\\>M3G\\A0g>3N2M4M2M3N2M3N3M2N2O1O110O000000010O000001O01O000001O01O0001O0001O0001O0001O01O000001O01O00000010O000000010O000001O01O000001O01O0001O0001O0001O0001O01O000001O01O0000010O00000010O0000N2M4M2N2N2N3M2N2N9Gh\\T1"}, "label": "a sofa behind an eating boy"}]} {"id": 524227, "image": "COCO_train2014_000000524227.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the kid with his arm up in the air\"."}], "task": "refering", "answer_template": "The \"the kid with his arm up in the air\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 35, 36, 58, 59, 82, 83, 105, 106, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 213, 214, 215, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 243, 258, 259, 260, 261, 262, 263, 264, 265, 266, 281, 282, 283, 284, 285, 286, 287, 288, 304, 305], "bbox": [0.22009375000000003, 0.03329166666666667, 0.628109375, 0.78725], "iscrowd": 0, "area": 39530.43915, "rle": {"size": [480, 640], "counts": "n]R25h>9G8H7I4L3M3M3M4L3N2N3M2O1O1N3N1N2O1N3N1N2O2M2O1N2O2M2O1N3N0O2O00000000001O000000001O00000000001O00000000001O00000000001O0000O1N2N2N2N2M3N2N3M2N2N201N1O1O2N11O3N2M3M3N2M2N3M3N2M3M2O2M3jDQMV:R3cESM]:o2\\EUMd:n2UEWMj:`3O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O010O1O100O1O100O1O100O1O100O1M3^Ob0N2M3N2N2N2N2N2M3N1O2N2N2N3L3N2O1000000O10000O1000000O1000000O10000O2O00000O1000000O100001O00001O00001O00001O01O01OWLbIPO^6n0fIPOZ6n0jIQOU6m0oIQOQ6l0UJROj5l0ZJSOe5k0_JSOa5m0aJRO^5n0eJPOZ5P1hJnNX5R1jJmNU5S1mJlNR5T1PKjNP5V1SKhNl4X1VKgNi4Y1XKgNg4Y1ZKfNf4Z1[KfNd4Z1\\KgNc4Y1^KfNc4Y1]KhNb4X1_KhN`4X1aKgN_4Y1aK^Nh4b1`KlMj4T2^KZMl4f2]KfLe4c3cKkK[4b4e2>C>A>C>CB>B>B>^Ob0]Oo^_3"}, "label": "the kid with his arm up in the air"}]} {"id": 524227, "image": "COCO_train2014_000000524227.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a child with grey dress\"."}], "task": "refering", "answer_template": "The \"a child with grey dress\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 13, 35, 36, 58, 59, 82, 83, 105, 106, 125, 126, 127, 128, 129, 148, 149, 150, 151, 152, 170, 171, 172, 173, 174, 175, 191, 192, 193, 194, 195, 196, 197, 198, 213, 214, 215, 216, 217, 218, 219, 220, 235, 236, 237, 238, 239, 240, 241, 242, 243, 258, 259, 260, 261, 262, 263, 264, 265, 266, 281, 282, 283, 284, 285, 286, 287, 288, 304, 305], "bbox": [0.22009375000000003, 0.03329166666666667, 0.628109375, 0.78725], "iscrowd": 0, "area": 39530.43915, "rle": {"size": [480, 640], "counts": "n]R25h>9G8H7I4L3M3M3M4L3N2N3M2O1O1N3N1N2O1N3N1N2O2M2O1N2O2M2O1N3N0O2O00000000001O000000001O00000000001O00000000001O00000000001O0000O1N2N2N2N2M3N2N3M2N2N201N1O1O2N11O3N2M3M3N2M2N3M3N2M3M2O2M3jDQMV:R3cESM]:o2\\EUMd:n2UEWMj:`3O100O1O100O1O100O1O100O100O1O100O1O100O1O100O1O010O1O100O1O100O1O100O1O100O1M3^Ob0N2M3N2N2N2N2N2M3N1O2N2N2N3L3N2O1000000O10000O1000000O1000000O10000O2O00000O1000000O100001O00001O00001O00001O01O01OWLbIPO^6n0fIPOZ6n0jIQOU6m0oIQOQ6l0UJROj5l0ZJSOe5k0_JSOa5m0aJRO^5n0eJPOZ5P1hJnNX5R1jJmNU5S1mJlNR5T1PKjNP5V1SKhNl4X1VKgNi4Y1XKgNg4Y1ZKfNf4Z1[KfNd4Z1\\KgNc4Y1^KfNc4Y1]KhNb4X1_KhN`4X1aKgN_4Y1aK^Nh4b1`KlMj4T2^KZMl4f2]KfLe4c3cKkK[4b4e2>C>A>C>CB>B>B>^Ob0]Oo^_3"}, "label": "a child with grey dress"}]} {"id": 253907, "image": "COCO_train2014_000000253907.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing a grey hooded sweatshirt\"."}], "task": "refering", "answer_template": "The \"a man wearing a grey hooded sweatshirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 30, 31, 47, 48, 49, 50, 65, 66, 67, 68, 83, 84, 85, 86, 101, 119], "bbox": [0.6100399999999999, 0.06502631578947368, 0.8298599999999999, 0.4439473684210527], "iscrowd": 0, "area": 7185.362400000002, "rle": {"size": [380, 500], "counts": "nYa3i0b:c0K4L4L4L4K6K4L4L3M4K4O1O10000O2O00UOnFlNQ9U1PGjNP9U1SGiNm8V1UGiNj8W1XGiNg8W1ZGhNe8X1^GfNb8Y1`GfN`8X1cGgN\\8X1gGhNX8U1lGjNT8T1Z1OG821O1O1O1O1O1XOg0O2O1N2N200000000000000000O2OO10000000000001O1O1O001O1N3N`0@6J1O1O2N1O1O2N1O1O1O5K3M010O000001O0001O01O0000010O01O3M2O1N3M2N2O2M2N2N3L3M3L4L5K4Ln]o0"}, "label": "a man wearing a grey hooded sweatshirt"}]} {"id": 253907, "image": "COCO_train2014_000000253907.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a smiling man with brown hair wearing a gray sweatshirt\"."}], "task": "refering", "answer_template": "The \"a smiling man with brown hair wearing a gray sweatshirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [12, 30, 31, 47, 48, 49, 50, 65, 66, 67, 68, 83, 84, 85, 86, 101, 119], "bbox": [0.6100399999999999, 0.06502631578947368, 0.8298599999999999, 0.4439473684210527], "iscrowd": 0, "area": 7185.362400000002, "rle": {"size": [380, 500], "counts": "nYa3i0b:c0K4L4L4L4K6K4L4L3M4K4O1O10000O2O00UOnFlNQ9U1PGjNP9U1SGiNm8V1UGiNj8W1XGiNg8W1ZGhNe8X1^GfNb8Y1`GfN`8X1cGgN\\8X1gGhNX8U1lGjNT8T1Z1OG821O1O1O1O1O1XOg0O2O1N2N200000000000000000O2OO10000000000001O1O1O001O1N3N`0@6J1O1O2N1O1O2N1O1O1O5K3M010O000001O0001O01O0000010O01O3M2O1N3M2N2O2M2N2N3L3M3L4L5K4Ln]o0"}, "label": "a smiling man with brown hair wearing a gray sweatshirt"}]} {"id": 253907, "image": "COCO_train2014_000000253907.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the person wearing the blue , white , and black shirt\"."}], "task": "refering", "answer_template": "The \"the person wearing the blue , white , and black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 26, 27, 44, 45, 46, 62, 63, 64, 80, 81, 82, 98, 99], "bbox": [0.43566000000000005, 0.022263157894736846, 0.60652, 0.41010526315789475], "iscrowd": 0, "area": 7271.955400000001, "rle": {"size": [380, 500], "counts": "dZa21k;1N100O3N4gDN_:7UE3h:b0O2M2O2M2O4K8aFdNP8d1`GdN_8^2O0O100RORMmHP3k6YMQIi2m6T1O1O1O100O10001N1000O1O001N2O001O1F911N1O1O1O2N2N2O0O2N2O2M2O2N2N2N3`M]Ga1f8ZN]Ge1f8VN]Gi1f8QN_Gm1d8nM_GP2X9O1O1O2N1O1O1000001N1000002M3N2M3M3N2M3N2M3M3N2MQPY2"}, "label": "the person wearing the blue , white , and black shirt"}]} {"id": 253907, "image": "COCO_train2014_000000253907.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person with white and blue shirt on sitting on a motorcycle\"."}], "task": "refering", "answer_template": "The \"a person with white and blue shirt on sitting on a motorcycle\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [8, 9, 26, 27, 44, 45, 46, 62, 63, 64, 80, 81, 82, 98, 99], "bbox": [0.43566000000000005, 0.022263157894736846, 0.60652, 0.41010526315789475], "iscrowd": 0, "area": 7271.955400000001, "rle": {"size": [380, 500], "counts": "dZa21k;1N100O3N4gDN_:7UE3h:b0O2M2O2M2O4K8aFdNP8d1`GdN_8^2O0O100RORMmHP3k6YMQIi2m6T1O1O1O100O10001N1000O1O001N2O001O1F911N1O1O1O2N2N2O0O2N2O2M2O2N2N2N3`M]Ga1f8ZN]Ge1f8VN]Gi1f8QN_Gm1d8nM_GP2X9O1O1O2N1O1O1000001N1000002M3N2M3M3N2M3N2M3M3N2MQPY2"}, "label": "a person with white and blue shirt on sitting on a motorcycle"}]} {"id": 548822, "image": "COCO_train2014_000000548822.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a man wearing red\"."}], "task": "refering", "answer_template": "The \"a man wearing red\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [5, 6, 7, 19, 20, 21, 22, 32, 33, 34, 35, 36, 47, 48, 49, 50, 51, 62, 63, 64, 65, 66, 77, 78, 79, 80, 81, 92, 93, 94, 95, 96, 107, 108, 109, 110, 111, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 153, 154, 155, 156, 157, 167, 168, 169, 171, 172, 182, 183, 184, 186, 187, 188, 197, 198, 199, 202, 203, 212, 213, 217, 218, 219, 227, 228, 229, 232, 233, 234, 248], "bbox": [0.1549414519906323, 0.0045, 0.6534192037470726, 0.701125], "iscrowd": 0, "area": 41688.90054999999, "rle": {"size": [640, 427], "counts": "l[Y1i0Rc08H8H8I`1aN5L5J6K4K6J6K5J5bFRLd2S4WJQOd5U1WJnNg5V1SJnNj5V1RJnNj5X1PJkNn5Y1mIiNQ6\\1jIfNU6^1fIdNX6a1cIaN\\6c1_I_Na6d1YI_Ng6d1TI^Nk6f1PI\\NP7g1hH^NX7e1\\HdNc7a1PHhNP8d600O10000O10000O10000O10000O100O1OoNXHlEh7c9jH\\FU7U9[IkFd6Q9bInF^6n8gIQGX6k8nITGQ6i8SJWGm5e8XJZGg5c8^J\\Ga5a8dJ^G[5]8lJbGT5U8UKkGj4n7_KQH`4i7hKVHW4h7nKVHQ4h7TLVHk3i7YLUHf3i7^LVHa3k7aLSH^3m7eLoG\\3Q8Y41O010O1O1O1000O10O1000000O0100000N2O1O1N101O0O2O001N101O0O2O001O0O2O001N101O001N110001N1PLSEmNm:Q1^EeNd:W1iE_NW:^1UFYNk9e1`FQNa9l1kFkMV9R2UGdMl8Y2]GjMZ8S2kGoMR8d0PFeMR2i1j7>fIDV67QJKl50ZJ2b5IeJ:V5BPKb0k4[OYKi0a4TOdKo0W4oNmKU1n3gNWL]1c3aNaL`1\\3]NjLc1T3ZNQMf1l2XNZMg1c2VNcMj1[2SNkMl1S2PNSNo1m1mMYNR2e1SNXNm1g1oM_NP2_1nMeNR2Z1jMkNV2S1gMSOX2l0eMXO[2X9001O7I in this image.", "objects": [{"patches": [5, 6, 7, 19, 20, 21, 22, 32, 33, 34, 35, 36, 47, 48, 49, 50, 51, 62, 63, 64, 65, 66, 77, 78, 79, 80, 81, 92, 93, 94, 95, 96, 107, 108, 109, 110, 111, 122, 123, 124, 125, 126, 138, 139, 140, 141, 142, 153, 154, 155, 156, 157, 167, 168, 169, 171, 172, 182, 183, 184, 186, 187, 188, 197, 198, 199, 202, 203, 212, 213, 217, 218, 219, 227, 228, 229, 232, 233, 234, 248], "bbox": [0.1549414519906323, 0.0045, 0.6534192037470726, 0.701125], "iscrowd": 0, "area": 41688.90054999999, "rle": {"size": [640, 427], "counts": "l[Y1i0Rc08H8H8I`1aN5L5J6K4K6J6K5J5bFRLd2S4WJQOd5U1WJnNg5V1SJnNj5V1RJnNj5X1PJkNn5Y1mIiNQ6\\1jIfNU6^1fIdNX6a1cIaN\\6c1_I_Na6d1YI_Ng6d1TI^Nk6f1PI\\NP7g1hH^NX7e1\\HdNc7a1PHhNP8d600O10000O10000O10000O10000O100O1OoNXHlEh7c9jH\\FU7U9[IkFd6Q9bInF^6n8gIQGX6k8nITGQ6i8SJWGm5e8XJZGg5c8^J\\Ga5a8dJ^G[5]8lJbGT5U8UKkGj4n7_KQH`4i7hKVHW4h7nKVHQ4h7TLVHk3i7YLUHf3i7^LVHa3k7aLSH^3m7eLoG\\3Q8Y41O010O1O1O1000O10O1000000O0100000N2O1O1N101O0O2O001N101O0O2O001O0O2O001N101O001N110001N1PLSEmNm:Q1^EeNd:W1iE_NW:^1UFYNk9e1`FQNa9l1kFkMV9R2UGdMl8Y2]GjMZ8S2kGoMR8d0PFeMR2i1j7>fIDV67QJKl50ZJ2b5IeJ:V5BPKb0k4[OYKi0a4TOdKo0W4oNmKU1n3gNWL]1c3aNaL`1\\3]NjLc1T3ZNQMf1l2XNZMg1c2VNcMj1[2SNkMl1S2PNSNo1m1mMYNR2e1SNXNm1g1oM_NP2_1nMeNR2Z1jMkNV2S1gMSOX2l0eMXO[2X9001O7I in this image.", "objects": [{"patches": [289, 303, 304, 305, 306, 307, 308, 309, 320, 321, 322, 323, 324, 325, 338, 339, 340, 341, 356, 357, 358, 374, 375, 376], "bbox": [0.0003125, 0.74590625, 1.0, 1.0], "iscrowd": 0, "area": 11047.860950000004, "rle": {"size": [640, 480], "counts": "l?`0`c0d1m]O\\Na>T2_@]Na?c1\\@aNb?_1_@aNa?^1_@cNa?\\1_@eN`?\\1`@dN`?[1`@fNj>kN]A_2IgNj>nNYA[2MgNk>oNUAZ20hNk>QORAV24iNk>ROo@U25jNl>SOl@S28kNm>SOi@Q2;lNl>c1TA^Nm>a1RA`No>^1RAbNn>]1RAdNo>[1PAfNP?Y1QAgNP?W1PAjNP?U1PAlNQ?S1o@mNQ?R1o@oNR?o0n@SOQ?m0o@SOR?k0n@VOc?8]@I[`0_Oe_Oa0Sa0fNm^O[1la00O1O100O1O1aMUNjBj1W=\\NeBa1[=gN_BW1a=PO[Bm0f=ZOTBd0l=CPB:Q>MiA1W>6eAG\\>?_A_Oa>c0`AZOa>d0bAZO^>e0fAYOZ>f0hAXOX>g0lAVOU>h0nAVOR>i0RBTOo=j0UBSOk=l0XBROi=l0[BQOe=o0]BoNc=P1aBmN_=R1eBkN\\=S1iBiNW=V1j2O101N1O1O1O2N1O100O2N1O1O1O2O0O1O1O2N1O1O1O2N1O1O1N3N1O1O1O2N1O1O\\gl51cXSJ2N2N2N1O2N2N2O1N1O2N2N1O2N2N1100O1000O010000O10O1O001O1N101O1O001N2O001O1O001N2O2N2N2N2N2M3N2N2N2N2N3L3N2N2N2N2M3N2N2N2N2N2M3N3M2N2N2N2M1000O1O1O1O001O001O0001O01OSA"}, "label": "table underneath the platters of pizza in front of the smiling guy with glasses"}]} {"id": 131030, "image": "COCO_train2014_000000131030.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black man sits at a table alone\"."}], "task": "refering", "answer_template": "The \"a black man sits at a table alone\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 73, 74, 90, 91, 106, 107, 108, 122, 123, 124, 125, 139, 140, 141, 156, 157], "bbox": [0.212375, 0.12646875, 0.4040416666666667, 0.43928124999999996], "iscrowd": 0, "area": 7140.06085, "rle": {"size": [640, 480], "counts": "ino19\\c0c0F?@i0XO5K4M2N2O1N2N2N2O1NO1N3J6I7J6M2O2M3N2N110O1O1O010O1OO1O2O0O1O101N1N2M3M22O0O2O001N101O001O001O001OeNdNR@]1i?oNo_OR1n?VOl_Ol0Q`0\\Og_Of0V`0c1O2O0O1O2O01O10O01O010O0010O01N2N101N101O001O001cM`_OY1b`0cNd_OZ1]`0cNi_OX1Z`0eNj_OX1W`0gNl_OU1\\`0cNf_O[1g`0YNZ_O0LP1cb0F5L4I9Hbbb5"}, "label": "a black man sits at a table alone"}]} {"id": 131030, "image": "COCO_train2014_000000131030.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"man next to wall in dark shirt and dark hair facing away\"."}], "task": "refering", "answer_template": "The \"man next to wall in dark shirt and dark hair facing away\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [56, 57, 73, 74, 90, 91, 106, 107, 108, 122, 123, 124, 125, 139, 140, 141, 156, 157], "bbox": [0.212375, 0.12646875, 0.4040416666666667, 0.43928124999999996], "iscrowd": 0, "area": 7140.06085, "rle": {"size": [640, 480], "counts": "ino19\\c0c0F?@i0XO5K4M2N2O1N2N2N2O1NO1N3J6I7J6M2O2M3N2N110O1O1O010O1OO1O2O0O1O101N1N2M3M22O0O2O001N101O001O001O001OeNdNR@]1i?oNo_OR1n?VOl_Ol0Q`0\\Og_Of0V`0c1O2O0O1O2O01O10O01O010O0010O01N2N101N101O001O001cM`_OY1b`0cNd_OZ1]`0cNi_OX1Z`0eNj_OX1W`0gNl_OU1\\`0cNf_O[1g`0YNZ_O0LP1cb0F5L4I9Hbbb5"}, "label": "man next to wall in dark shirt and dark hair facing away"}]} {"id": 131030, "image": "COCO_train2014_000000131030.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a male smiling with spikey hair and glasses\"."}], "task": "refering", "answer_template": "The \"a male smiling with spikey hair and glasses\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [24, 25, 26, 27, 40, 41, 42, 43, 44, 57, 58, 59, 60, 61, 74, 75, 76, 77, 78, 91, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 113, 114, 124, 125, 126, 127, 128, 129, 130, 131, 132, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201], "bbox": [0.2456875, 0.047343750000000004, 0.8629166666666667, 0.51925], "iscrowd": 0, "area": 55310.168999999994, "rle": {"size": [640, 480], "counts": "ooY2j1Ub07H2N2O2M2O1N2N2O1N2O1N2N2O1O1O1O1O1O1O1O1O1O100O1O2N1O1O1O1O1O100O100O100O100O10000O100O1O1O1O1O2N1O1O1O1O1O1O1O1O1WLiLhFX3V9mLfFT3Y9QMbFP3]9TM_Fm2_9XM]Fi2b9\\MYFe2f9_MVFb2h9cMSF_2l9fMoE[2P:k3N2O1O1N3N1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1O1O1O1O1O1O2N1O1000000000000000000000000000000000000000001O00000000000000000000000001O0000001O0000001O01O0001O0000001O0000001O00001O001O001O00001O0010O0001O001O001O1O1O1O1O1O1O1O1O1O1QHXEd6[ in this image.", "objects": [{"patches": [24, 25, 26, 27, 40, 41, 42, 43, 44, 57, 58, 59, 60, 61, 74, 75, 76, 77, 78, 91, 92, 93, 94, 95, 96, 108, 109, 110, 111, 112, 113, 114, 124, 125, 126, 127, 128, 129, 130, 131, 132, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201], "bbox": [0.2456875, 0.047343750000000004, 0.8629166666666667, 0.51925], "iscrowd": 0, "area": 55310.168999999994, "rle": {"size": [640, 480], "counts": "ooY2j1Ub07H2N2O2M2O1N2N2O1N2O1N2N2O1O1O1O1O1O1O1O1O1O100O1O2N1O1O1O1O1O100O100O100O100O10000O100O1O1O1O1O2N1O1O1O1O1O1O1O1O1WLiLhFX3V9mLfFT3Y9QMbFP3]9TM_Fm2_9XM]Fi2b9\\MYFe2f9_MVFb2h9cMSF_2l9fMoE[2P:k3N2O1O1N3N1O1N2O1O1N2O1O1O1N2O1O1N2O1O1O1O1O1O1O1O1O2N1O1000000000000000000000000000000000000000001O00000000000000000000000001O0000001O0000001O01O0001O0000001O0000001O00001O001O001O00001O0010O0001O001O001O1O1O1O1O1O1O1O1O1O1QHXEd6[ in this image.", "objects": [{"patches": [85, 86, 87, 102, 103, 104, 105, 119, 120, 121, 122, 136, 137, 138, 139, 153, 154, 155, 156, 170, 171, 172, 173, 187, 188, 189, 190, 204, 205, 206, 221], "bbox": [0.000125, 0.219046875, 0.23535416666666667, 0.58821875], "iscrowd": 0, "area": 18651.614049999996, "rle": {"size": [640, 480], "counts": "a4a0_c0R1PBaNZ7U2iGTNU8a5PDeJPm0mAROU>m0lAROT>m0mAROU>l0mASOT>l0lASOV>k0kAUOU>j0mATOU>k0kAUOV>i0kAVOW>h0kAWOU>i0kAVOW>h0jAXOW>f0lAWOV>f0mAYOT>d0oAZOR>d0RBZOo=c0TB[On=c0UB[Ol=c0VB[Ok=d0WB[Oj=c0YB[Oh=OVAlN]1l0^=6^AeNQ29a in this image.", "objects": [{"patches": [85, 86, 87, 102, 103, 104, 105, 119, 120, 121, 122, 136, 137, 138, 139, 153, 154, 155, 156, 170, 171, 172, 173, 187, 188, 189, 190, 204, 205, 206, 221], "bbox": [0.000125, 0.219046875, 0.23535416666666667, 0.58821875], "iscrowd": 0, "area": 18651.614049999996, "rle": {"size": [640, 480], "counts": "a4a0_c0R1PBaNZ7U2iGTNU8a5PDeJPm0mAROU>m0lAROT>m0mAROU>l0mASOT>l0lASOV>k0kAUOU>j0mATOU>k0kAUOV>i0kAVOW>h0kAWOU>i0kAVOW>h0jAXOW>f0lAWOV>f0mAYOT>d0oAZOR>d0RBZOo=c0TB[On=c0UB[Ol=c0VB[Ok=d0WB[Oj=c0YB[Oh=OVAlN]1l0^=6^AeNQ29a in this image.", "objects": [{"patches": [77, 94, 95, 96, 112, 113, 114, 115, 130, 131, 147, 148, 149, 165, 166, 167, 168, 185, 186], "bbox": [0.18209999999999998, 0.40138036809815947, 0.41264, 0.8585889570552147], "iscrowd": 0, "area": 6594.041549999999, "rle": {"size": [326, 500], "counts": "TWm0=b98N2N2N100O001O100O1O00100O1O010O0001O010O0010O0001O010O0010O01O0001fMWOVKi0k4CgI1H=a6m0\\ITOc6o0YIROg6Q1VIoNj6T1SIlNm6W2O00010O00001O001O00010O001O00f0ZO;E6G4000000001O000000000jMdJB\\5=S5lN`Je0b0 in this image.", "objects": [{"patches": [77, 94, 95, 96, 112, 113, 114, 115, 130, 131, 147, 148, 149, 165, 166, 167, 168, 185, 186], "bbox": [0.18209999999999998, 0.40138036809815947, 0.41264, 0.8585889570552147], "iscrowd": 0, "area": 6594.041549999999, "rle": {"size": [326, 500], "counts": "TWm0=b98N2N2N100O001O100O1O00100O1O010O0001O010O0010O0001O010O0010O01O0001fMWOVKi0k4CgI1H=a6m0\\ITOc6o0YIROg6Q1VIoNj6T1SIlNm6W2O00010O00001O001O00010O001O00f0ZO;E6G4000000001O000000000jMdJB\\5=S5lN`Je0b0 in this image.", "objects": [{"patches": [151, 163, 164, 165, 175, 176, 177, 178, 189, 190, 202, 203], "bbox": [0.5189066666666666, 0.6395599999999999, 0.7804266666666666, 0.8966399999999999], "iscrowd": 0, "area": 4879.570799999999, "rle": {"size": [500, 375], "counts": "ibo22`?2N3M2O2M2O00010O000O1N3N1O1000001aB2W;NeD7Z;IfD8Y;HfD:Y;FgDa0=000010O2N1O101N1O2N1O2N1O0O10000O100O10O0100O100M3MPUX1"}, "label": "a man in black skiing down a hill"}]} {"id": 417761, "image": "COCO_train2014_000000417761.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"skier in white and black going down a slope\"."}], "task": "refering", "answer_template": "The \"skier in white and black going down a slope\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [151, 163, 164, 165, 175, 176, 177, 178, 189, 190, 202, 203], "bbox": [0.5189066666666666, 0.6395599999999999, 0.7804266666666666, 0.8966399999999999], "iscrowd": 0, "area": 4879.570799999999, "rle": {"size": [500, 375], "counts": "ibo22`?2N3M2O2M2O00010O000O1N3N1O1000001aB2W;NeD7Z;IfD8Y;HfD:Y;FgDa0=000010O2N1O101N1O2N1O2N1O0O10000O100O10O0100O100M3MPUX1"}, "label": "skier in white and black going down a slope"}]} {"id": 155618, "image": "COCO_train2014_000000155618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the tank of the toilet\"."}], "task": "refering", "answer_template": "The \"the tank of the toilet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [276, 277, 292, 293, 294, 295, 296, 297, 309, 310, 311, 312, 313, 326, 327, 328, 329, 330, 343, 344, 345, 346, 347, 360, 361, 362, 363, 364, 379, 380], "bbox": [0.1960041841004184, 0.7173437500000001, 0.4885146443514644, 0.9876093750000001], "iscrowd": 0, "area": 14379.818050000002, "rle": {"size": [640, 478], "counts": "`Wk14hc05L5J5K5L4N2O1N2O2M;F>A?B>B?@?B2M2O1N2O1N2O1N2O1O1N2O0O2O1N101N2O0O2O1N101O0O2O1N101N2O0O2O10O0100O0010O0100O00100O00100O010O1O010O00100O010OO101N1O2O0O1O2O0O101N1O2O0O1O2O0O2O0O101O0O101N101N100O2O0N3nL_@a1c?ZNc@c1_?]Nb@`1a?^N`@`1b?`N_@1o`0NR_OE\\a0:e^OYOga0f0k0O0O101O0O2O0O101N100O2O0O2O000O2O0O2O0O101N100O2O0O2O000O2Oadh4"}, "label": "the tank of the toilet"}]} {"id": 155618, "image": "COCO_train2014_000000155618.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the top of the toilet\"."}], "task": "refering", "answer_template": "The \"the top of the toilet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [276, 277, 292, 293, 294, 295, 296, 297, 309, 310, 311, 312, 313, 326, 327, 328, 329, 330, 343, 344, 345, 346, 347, 360, 361, 362, 363, 364, 379, 380], "bbox": [0.1960041841004184, 0.7173437500000001, 0.4885146443514644, 0.9876093750000001], "iscrowd": 0, "area": 14379.818050000002, "rle": {"size": [640, 478], "counts": "`Wk14hc05L5J5K5L4N2O1N2O2M;F>A?B>B?@?B2M2O1N2O1N2O1N2O1O1N2O0O2O1N101N2O0O2O1N101O0O2O1N101N2O0O2O10O0100O0010O0100O00100O00100O010O1O010O00100O010OO101N1O2O0O1O2O0O101N1O2O0O1O2O0O2O0O101O0O101N101N100O2O0N3nL_@a1c?ZNc@c1_?]Nb@`1a?^N`@`1b?`N_@1o`0NR_OE\\a0:e^OYOga0f0k0O0O101O0O2O0O101N100O2O0O2O000O2O0O2O0O101N100O2O0O2O000O2Oadh4"}, "label": "the top of the toilet"}]} {"id": 286690, "image": "COCO_train2014_000000286690.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a black office chair with a coat hanging on it\"."}], "task": "refering", "answer_template": "The \"a black office chair with a coat hanging on it\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 108, 109, 110, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 170, 180, 181, 182, 183, 184, 185, 186, 187, 188, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224], "bbox": [0.01456, 0.3827733333333333, 0.48598, 0.9957599999999999], "iscrowd": 0, "area": 34038.674100000004, "rle": {"size": [375, 500], "counts": "kf24b;`0@`0@`0_Ob0_O`0@a0_O`0@a0_O`0A?A`0@?A`0@8H0000000000000000000000000000000000000000001O0000000001O1O1O001O1O1O001O1O1O2N2N2N1O2N2N2N2N2O0O2N2N2N2N1O2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N1O2N2O1N2N2N1O2N2N2N00000000000000000000O100000XLkHn2U7QMmHn2S7RMoHl2Q7TMQIj2o6VMSIi2l6WMSIj2m6UMTIk2l6UMTIk2k6VMUIj2k6VMTIk2l6UMTIk2l6TMUIl2k6TMUIl2k6TMUIl2k6TMTIm2l6RMUIn2k6RMUIn2k6RMUIn2k6RMUIn2j6SMUIn2k6QMVIo2j6QMVIo2j6QMVIo2j6QMUIP3k6PMUIP3k6oLVIQ3j6oLVIQ3j6oLVIQ3j6oLUIR3k6nLUIS3i6mLXIS3h6mLXIS3h6mLWIT3i6lLWIT3i6lLWIT3i6kLXIU3h6kLXIU3h6kLWIV3i6jLWIV3i6jLWIV3i6iLXIW3h6iLXIW3g6jLXIW3h6iLXIW3h6iLXIW3h6hLYIX3g6hLXIY3h6gLXIY3h6gLXIY3h6gLXIY3h6fLYIZ3g6fLXI[3h6eLXI[3g6fLYI[3i6aLXI_3n6[LQIf3T7ULlHk3`700000001O00001O00001O00001O001O00001O000010O0001O00001O00001O00001O00001O001O00001O000O2M2O1N3M2O1N5K4L5L3L5K4M4K4L5K4M4K5K4M4K4L5L3L5K4L5L3L5K4Mhim2"}, "label": "a black office chair with a coat hanging on it"}]} {"id": 286690, "image": "COCO_train2014_000000286690.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"computer chair\"."}], "task": "refering", "answer_template": "The \"computer chair\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 108, 109, 110, 126, 127, 128, 129, 130, 131, 144, 145, 146, 147, 148, 149, 150, 151, 162, 163, 164, 165, 166, 167, 168, 169, 170, 180, 181, 182, 183, 184, 185, 186, 187, 188, 198, 199, 200, 201, 202, 203, 204, 205, 206, 216, 217, 218, 219, 220, 221, 222, 223, 224], "bbox": [0.01456, 0.3827733333333333, 0.48598, 0.9957599999999999], "iscrowd": 0, "area": 34038.674100000004, "rle": {"size": [375, 500], "counts": "kf24b;`0@`0@`0_Ob0_O`0@a0_O`0@a0_O`0A?A`0@?A`0@8H0000000000000000000000000000000000000000001O0000000001O1O1O001O1O1O001O1O1O2N2N2N1O2N2N2N2N2O0O2N2N2N2N1O2N2N2N2N1O2N2N2N2N2N1O2N2N2N2N1O2N2O1N2N2N1O2N2N2N00000000000000000000O100000XLkHn2U7QMmHn2S7RMoHl2Q7TMQIj2o6VMSIi2l6WMSIj2m6UMTIk2l6UMTIk2k6VMUIj2k6VMTIk2l6UMTIk2l6TMUIl2k6TMUIl2k6TMUIl2k6TMTIm2l6RMUIn2k6RMUIn2k6RMUIn2k6RMUIn2j6SMUIn2k6QMVIo2j6QMVIo2j6QMVIo2j6QMUIP3k6PMUIP3k6oLVIQ3j6oLVIQ3j6oLVIQ3j6oLUIR3k6nLUIS3i6mLXIS3h6mLXIS3h6mLWIT3i6lLWIT3i6lLWIT3i6kLXIU3h6kLXIU3h6kLWIV3i6jLWIV3i6jLWIV3i6iLXIW3h6iLXIW3g6jLXIW3h6iLXIW3h6iLXIW3h6hLYIX3g6hLXIY3h6gLXIY3h6gLXIY3h6gLXIY3h6fLYIZ3g6fLXI[3h6eLXI[3g6fLYI[3i6aLXI_3n6[LQIf3T7ULlHk3`700000001O00001O00001O00001O001O00001O000010O0001O00001O00001O00001O00001O001O00001O000O2M2O1N3M2O1N5K4L5L3L5K4M4K4L5K4M4K5K4M4K4L5L3L5K4L5L3L5K4Mhim2"}, "label": "computer chair"}]} {"id": 311273, "image": "COCO_train2014_000000311273.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a person in a black shirt in the background\"."}], "task": "refering", "answer_template": "The \"a person in a black shirt in the background\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 17, 18, 19, 20, 34, 35, 36, 37, 51, 52, 53, 54, 68, 69, 70, 71, 85, 86, 87, 88, 104, 105], "bbox": [0.0, 0.0021718749999999998, 0.23826446280991734, 0.30751562499999996], "iscrowd": 0, "area": 17110.665, "rle": {"size": [640, 484], "counts": "1Q1ob0R2nMk0UO:F7I001O001O00001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O001O00001O001O2N1O1O1O1O1O1O101N100O100O10001N100O100O10000O2O0O100O10000O100OO2N2O1N2N3N1N2N3M2O1N2N3N1N2N3M2M3M3M4L3L4M4L3M7IT1kN;F:F:F:F:E;F:Fg_V7"}, "label": "a person in a black shirt in the background"}]} {"id": 311273, "image": "COCO_train2014_000000311273.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a woman in the background in a black shirt\"."}], "task": "refering", "answer_template": "The \"a woman in the background in a black shirt\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [0, 1, 2, 3, 17, 18, 19, 20, 34, 35, 36, 37, 51, 52, 53, 54, 68, 69, 70, 71, 85, 86, 87, 88, 104, 105], "bbox": [0.0, 0.0021718749999999998, 0.23826446280991734, 0.30751562499999996], "iscrowd": 0, "area": 17110.665, "rle": {"size": [640, 484], "counts": "1Q1ob0R2nMk0UO:F7I001O001O00001O00001O00001O00001O001O00001O00001O00001O001O00001O00001O00001O001O00001O001O2N1O1O1O1O1O1O101N100O100O10001N100O100O10000O2O0O100O10000O100OO2N2O1N2N3N1N2N3M2O1N2N3N1N2N3M2M3M3M4L3L4M4L3M7IT1kN;F:F:F:F:E;F:Fg_V7"}, "label": "a woman in the background in a black shirt"}]} {"id": 311273, "image": "COCO_train2014_000000311273.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl smiling as a woman cuts a cake\"."}], "task": "refering", "answer_template": "The \"a little girl smiling as a woman cuts a cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 105, 106, 107, 108, 109, 110, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 154, 155, 156, 157, 158, 159, 160, 161, 170, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 238, 239, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 263, 272, 273, 274, 275, 276, 277, 278, 279, 280, 289, 290, 291, 292, 293, 294, 295, 296, 297, 306, 307, 308, 309], "bbox": [0.0029752066115702478, 0.24284374999999997, 0.5437809917355372, 0.8091406250000001], "iscrowd": 0, "area": 71396.585, "rle": {"size": [640, 484], "counts": "Sn02nc0a0^Ob0^Oa0_Ob0^Ob0^Ob0^Oa0@a0^Ob0^Ob0^Oa0_O:F10O01O1O1O1O001O1N2N2N2N1O2N2N2N2N2O0O2N2O1N2N2O0O2N2O1N2N101N2O1O100O010O100O100O00100O100O010O100O1O10O0100O100O10O010000O1000O010000O10O10M3M3L4L4M2M4M3L4M3L4N1O2N2N2O1N1O2O1N2N2O1O001O1O1O100O001O1O1O1O1O00100O1O1O001O1O1O1O1O010O1O100O10O1000O1000O10O1000000O01000O100000O01000000O10O1000O1000O1000O100000O10O10000000O0100000000O02O00001O0O10001O00000O2O00001O000O101O00000O2O00001O000O101O00001O0O101O001O000O2O001O2N3L4M3M2N3M3M3L4M2N3M2N2N2M3N2N1O2N2PN]D^Kf;W4fDdK^;Q4nDjKV;o3RElKQ;R4SEiKQ;T4SEgKQ;U4TEfKP;W4TEdKP;Y4TEbKP;Z4UEaKn:]4VE^Kn:_4VE\\K[;S4j2\\Mm_O:Tb0H8G:G8H9F9HchY4"}, "label": "a little girl smiling as a woman cuts a cake"}]} {"id": 311273, "image": "COCO_train2014_000000311273.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a little girl wearing a maroon shirt , visibly happy to get a piece of the cake\"."}], "task": "refering", "answer_template": "The \"a little girl wearing a maroon shirt , visibly happy to get a piece of the cake\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [90, 91, 92, 105, 106, 107, 108, 109, 110, 122, 123, 124, 125, 126, 127, 138, 139, 140, 141, 142, 143, 144, 154, 155, 156, 157, 158, 159, 160, 161, 170, 171, 172, 173, 174, 175, 176, 177, 178, 187, 188, 189, 190, 191, 192, 193, 194, 195, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 238, 239, 240, 241, 242, 243, 244, 245, 246, 255, 256, 257, 258, 259, 260, 261, 262, 263, 272, 273, 274, 275, 276, 277, 278, 279, 280, 289, 290, 291, 292, 293, 294, 295, 296, 297, 306, 307, 308, 309], "bbox": [0.0029752066115702478, 0.24284374999999997, 0.5437809917355372, 0.8091406250000001], "iscrowd": 0, "area": 71396.585, "rle": {"size": [640, 484], "counts": "Sn02nc0a0^Ob0^Oa0_Ob0^Ob0^Ob0^Oa0@a0^Ob0^Ob0^Oa0_O:F10O01O1O1O1O001O1N2N2N2N1O2N2N2N2N2O0O2N2O1N2N2O0O2N2O1N2N101N2O1O100O010O100O100O00100O100O010O100O1O10O0100O100O10O010000O1000O010000O10O10M3M3L4L4M2M4M3L4M3L4N1O2N2N2O1N1O2O1N2N2O1O001O1O1O100O001O1O1O1O1O00100O1O1O001O1O1O1O1O010O1O100O10O1000O1000O10O1000000O01000O100000O01000000O10O1000O1000O1000O100000O10O10000000O0100000000O02O00001O0O10001O00000O2O00001O000O101O00000O2O00001O000O101O00001O0O101O001O000O2O001O2N3L4M3M2N3M3M3L4M2N3M2N2N2M3N2N1O2N2PN]D^Kf;W4fDdK^;Q4nDjKV;o3RElKQ;R4SEiKQ;T4SEgKQ;U4TEfKP;W4TEdKP;Y4TEbKP;Z4UEaKn:]4VE^Kn:_4VE\\K[;S4j2\\Mm_O:Tb0H8G:G8H9F9HchY4"}, "label": "a little girl wearing a maroon shirt , visibly happy to get a piece of the cake"}]} {"id": 221169, "image": "COCO_train2014_000000221169.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a fat man in grey pants\"."}], "task": "refering", "answer_template": "The \"a fat man in grey pants\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [95, 96, 97, 115, 116, 117, 118, 135, 136, 137, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 198, 199, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 273, 274, 275, 276, 277, 278, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 335, 336, 337, 338, 355, 356, 357, 358, 375, 376, 377, 378, 395, 396, 397, 398, 415, 416, 417, 418, 434, 435, 436, 437, 438, 439], "bbox": [0.653884007029877, 0.18175, 1.0, 0.949609375], "iscrowd": 0, "area": 52680.27399999999, "rle": {"size": [640, 569], "counts": "ghX79cc0;E6K3M4M2N3M2N3M2O2M2N2N3e_OaNQ>a1kAcNS>`1gAfNV>\\1eAjNX>Y1cAlNZ>V1dAlN[>U1cAnNZ>T1dAoNY>T1dAoNY>S1fAnNX>T1fAoNW>S1gAPOV>R1hAQOU>R1iAoNU>S1jAoNS>S1mAnNP>T1QBkNl=X1TBhNj=Z1VBfNh=\\1XBeNe=]1[BdNb=^1^BcN_=`1`B`N^=d1`B]N_=d1_B^N_=e1_B\\N`=f1^B^N]=d1bBhNQ=[1mBQOg in this image.", "objects": [{"patches": [95, 96, 97, 115, 116, 117, 118, 135, 136, 137, 154, 155, 156, 157, 158, 173, 174, 175, 176, 177, 178, 179, 193, 194, 195, 196, 197, 198, 199, 213, 214, 215, 216, 217, 218, 219, 234, 235, 236, 237, 238, 253, 254, 255, 256, 257, 258, 273, 274, 275, 276, 277, 278, 294, 295, 296, 297, 298, 314, 315, 316, 317, 318, 335, 336, 337, 338, 355, 356, 357, 358, 375, 376, 377, 378, 395, 396, 397, 398, 415, 416, 417, 418, 434, 435, 436, 437, 438, 439], "bbox": [0.653884007029877, 0.18175, 1.0, 0.949609375], "iscrowd": 0, "area": 52680.27399999999, "rle": {"size": [640, 569], "counts": "ghX79cc0;E6K3M4M2N3M2N3M2O2M2N2N3e_OaNQ>a1kAcNS>`1gAfNV>\\1eAjNX>Y1cAlNZ>V1dAlN[>U1cAnNZ>T1dAoNY>T1dAoNY>S1fAnNX>T1fAoNW>S1gAPOV>R1hAQOU>R1iAoNU>S1jAoNS>S1mAnNP>T1QBkNl=X1TBhNj=Z1VBfNh=\\1XBeNe=]1[BdNb=^1^BcN_=`1`B`N^=d1`B]N_=d1_B^N_=e1_B\\N`=f1^B^N]=d1bBhNQ=[1mBQOg in this image.", "objects": [{"patches": [47, 66, 67, 68, 86, 87, 88, 106, 107, 108, 109, 125, 126, 127, 128, 129, 130, 144, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 204, 205, 206, 207, 208, 209, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 285, 286, 287, 288, 289, 305, 306, 307, 308, 309, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349, 365, 366, 367, 368, 369, 385, 386, 387, 388, 389, 405, 406, 407, 408, 409, 424, 425, 426, 427, 428, 429], "bbox": [0.16175746924428824, 0.12584375, 0.5535500878734623, 0.9460625], "iscrowd": 0, "area": 64665.97124999997, "rle": {"size": [640, 569], "counts": "Uki13lc04K6J6K4K3M3N2M3M3N2M3N0O1O100O00100O1O1M2O2M3N2M3N2M2O2M01L3N3O010O00VFWNb1h1aNZNZ1g1iNYNP1l1QOUNg0R2YOoM`0V2AlM7Y2JhM1^2_HaMV62X1a2\\HfMX6IX1g2YHjMY6AY1j2XHnMZ6YOZ1o2UHRN[6PO]1R3RHWN\\6hN^1Q5\\NPK`1Y5[NhJ`1`5[NaJb1g5WN[Jg1l5RNVJl1P6nMRJQ2T6hMnIV2Y6dMhI[2^6^MdIa2a6YMaIe2h6RMYIn2o6iLSIV3U7aLmH]3]7ZLdHe3d7RL^Hm3j7jKXHT4Q8cKQH\\4j9hIXFV6j;O001N2O1O1N2O1O1N2O1N101O1O1O100O1O1O1O10O01O1O1O100O1O1_NVBkLj=Q3dBfL\\=V3QC`LP=\\3XC_Li<^3\\C_Le in this image.", "objects": [{"patches": [47, 66, 67, 68, 86, 87, 88, 106, 107, 108, 109, 125, 126, 127, 128, 129, 130, 144, 145, 146, 147, 148, 149, 150, 164, 165, 166, 167, 168, 169, 170, 184, 185, 186, 187, 188, 189, 190, 204, 205, 206, 207, 208, 209, 223, 224, 225, 226, 227, 228, 229, 243, 244, 245, 246, 247, 248, 249, 263, 264, 265, 266, 267, 268, 269, 285, 286, 287, 288, 289, 305, 306, 307, 308, 309, 325, 326, 327, 328, 329, 345, 346, 347, 348, 349, 365, 366, 367, 368, 369, 385, 386, 387, 388, 389, 405, 406, 407, 408, 409, 424, 425, 426, 427, 428, 429], "bbox": [0.16175746924428824, 0.12584375, 0.5535500878734623, 0.9460625], "iscrowd": 0, "area": 64665.97124999997, "rle": {"size": [640, 569], "counts": "Uki13lc04K6J6K4K3M3N2M3M3N2M3N0O1O100O00100O1O1M2O2M3N2M3N2M2O2M01L3N3O010O00VFWNb1h1aNZNZ1g1iNYNP1l1QOUNg0R2YOoM`0V2AlM7Y2JhM1^2_HaMV62X1a2\\HfMX6IX1g2YHjMY6AY1j2XHnMZ6YOZ1o2UHRN[6PO]1R3RHWN\\6hN^1Q5\\NPK`1Y5[NhJ`1`5[NaJb1g5WN[Jg1l5RNVJl1P6nMRJQ2T6hMnIV2Y6dMhI[2^6^MdIa2a6YMaIe2h6RMYIn2o6iLSIV3U7aLmH]3]7ZLdHe3d7RL^Hm3j7jKXHT4Q8cKQH\\4j9hIXFV6j;O001N2O1O1N2O1O1N2O1N101O1O1O100O1O1O1O10O01O1O1O100O1O1_NVBkLj=Q3dBfL\\=V3QC`LP=\\3XC_Li<^3\\C_Le in this image.", "objects": [{"patches": [58, 59, 60, 80, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.35384375, 0.1467291666666667, 0.997171875, 1.0], "iscrowd": 0, "area": 105923.31354999999, "rle": {"size": [480, 640], "counts": "kXZ32l>a0^Ob0_O4K4M3L4M3L4M3M3N2M3M3M3M3M3M3M3M3M3N2N2N2N1O1O2N1O1O100O2N1O1O1O1O1O2N1O1O1O1O2O0O1O1O1O1O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1O1O0O10O100O10000O01000O10000O01000O100T3lL1hK[HT1f7kNhHg0Y7XOhHg0Y7YOhHe0Y7ZOiHd0X7[OjHc0W7\\OkHb0V7^OkH`0V7_OlH?U7@mH]ODYN`7Y2nH\\OT8c0nG[OS8e0mGZOT8e0nGYOS8f0oGXOR8g0PHWOQ8h0QHVOP8j0QHTOP8k0RHTOn7k0THSOm7l0UHSOk7l0XHROh7n0\\HmNe7R1_HkNa7T1cHiN]7V1gHfNZ7Y1kHcNU7]1nH`NR7_1RI^Nn6a1WIZNj6e1ZIXNf6g1]IWNc6j1_IRNb6n1`IPN`6P2cImM]6S2eIkM[6V2W3000000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O00000000001O0000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000001O0000002N1O1O1SE[MLMY8i2eGbMLI^8g2^GiMMEd8c2XGQNN@i8`2SGXNN\\On8]2mF`NOWOS9Z2gFDX9S3O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O001O00001O00001O00001O001O00001O00001O00001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O2N2N1O2N2N1O2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N2N2N6J7I7I7I7I7I7I7I7I7I8H7I7I7I7I7I7I7I7I7IU?"}, "label": "man in white shirt and blue and blue checked vest"}]} {"id": 376817, "image": "COCO_train2014_000000376817.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"an officer in a blue vest\"."}], "task": "refering", "answer_template": "The \"an officer in a blue vest\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [58, 59, 60, 80, 81, 82, 83, 84, 85, 103, 104, 105, 106, 107, 108, 127, 128, 129, 130, 131, 149, 150, 151, 152, 153, 154, 172, 173, 174, 175, 176, 177, 178, 179, 180, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390], "bbox": [0.35384375, 0.1467291666666667, 0.997171875, 1.0], "iscrowd": 0, "area": 105923.31354999999, "rle": {"size": [480, 640], "counts": "kXZ32l>a0^Ob0_O4K4M3L4M3L4M3M3N2M3M3M3M3M3M3M3M3M3N2N2N2N1O1O2N1O1O100O2N1O1O1O1O1O2N1O1O1O1O2O0O1O1O1O1O1O1O1N2O1O1N2O1O1N2O1O1N2O1O1N2O1O1O1O1O0O10O100O10000O01000O10000O01000O100T3lL1hK[HT1f7kNhHg0Y7XOhHg0Y7YOhHe0Y7ZOiHd0X7[OjHc0W7\\OkHb0V7^OkH`0V7_OlH?U7@mH]ODYN`7Y2nH\\OT8c0nG[OS8e0mGZOT8e0nGYOS8f0oGXOR8g0PHWOQ8h0QHVOP8j0QHTOP8k0RHTOn7k0THSOm7l0UHSOk7l0XHROh7n0\\HmNe7R1_HkNa7T1cHiN]7V1gHfNZ7Y1kHcNU7]1nH`NR7_1RI^Nn6a1WIZNj6e1ZIXNf6g1]IWNc6j1_IRNb6n1`IPN`6P2cImM]6S2eIkM[6V2W3000000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O00000000001O0000000000000000000000000000000000000000000000001O00000000000000000000000000000000000000000000001O0000002N1O1O1SE[MLMY8i2eGbMLI^8g2^GiMMEd8c2XGQNN@i8`2SGXNN\\On8]2mF`NOWOS9Z2gFDX9S3O001O1O1O1O1O1O1O1O1O1O1O1O1O001O1O1O1O1O1O1O1O1O1O001O00001O00001O00001O001O00001O00001O00001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O00001O0000001O0000001O2N2N1O2N2N1O2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N2N2N1O2N2N2N1O2N2N2N2N1O2N2N2N2N6J7I7I7I7I7I7I7I7I7I8H7I7I7I7I7I7I7I7I7IU?"}, "label": "an officer in a blue vest"}]} {"id": 442356, "image": "COCO_train2014_000000442356.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a multi colored kite lays on the carpet\"."}], "task": "refering", "answer_template": "The \"a multi colored kite lays on the carpet\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [142, 143, 159, 160, 176, 177, 192, 193, 194, 209, 210, 211, 226, 227, 228, 242, 243, 244, 245, 259, 260, 261, 262, 276, 277, 278, 279, 293, 294, 295, 296, 309, 310, 311, 312, 313, 314, 315, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.17079166666666667, 0.341578125, 0.8539375000000001, 0.9887656249999999], "iscrowd": 0, "area": 50020.83964999999, "rle": {"size": [640, 480], "counts": "gkc12ic05L4K5L4L4K5L4K5L4L4K5L4K5L4K5L4K5K5K5L4K5K5K5L4K5K5K5L4K5K5K5L4K5K5K5L4K5K5K5L4K5K5L4M3M3M3M3M3M3N2M3M3M3M3M3M3M3M3M3M3M2N3M3N2M3M3M3M3M3M3M3M3M3M3M3M3M3N2M3M3M3M3M3M3M3M3M3M3M3M3M3N2M3M3M3M3M3M3M3M3M3M3M3M3M3M3N2M3M3M3M3M3M3M3M3N2:F9G9G:F9G9G9G:F9G9G:kIZF\\1o9jMgFQ2b9UMUGe2U9`LaG[3h8kKnGP4W in this image.", "objects": [{"patches": [142, 143, 159, 160, 176, 177, 192, 193, 194, 209, 210, 211, 226, 227, 228, 242, 243, 244, 245, 259, 260, 261, 262, 276, 277, 278, 279, 293, 294, 295, 296, 309, 310, 311, 312, 313, 314, 315, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387], "bbox": [0.17079166666666667, 0.341578125, 0.8539375000000001, 0.9887656249999999], "iscrowd": 0, "area": 50020.83964999999, "rle": {"size": [640, 480], "counts": "gkc12ic05L4K5L4L4K5L4K5L4L4K5L4K5L4K5L4K5K5K5L4K5K5K5L4K5K5K5L4K5K5K5L4K5K5K5L4K5K5K5L4K5K5L4M3M3M3M3M3M3N2M3M3M3M3M3M3M3M3M3M3M2N3M3N2M3M3M3M3M3M3M3M3M3M3M3M3M3N2M3M3M3M3M3M3M3M3M3M3M3M3M3N2M3M3M3M3M3M3M3M3M3M3M3M3M3M3N2M3M3M3M3M3M3M3M3N2:F9G9G:F9G9G9G:F9G9G:kIZF\\1o9jMgFQ2b9UMUGe2U9`LaG[3h8kKnGP4W in this image.", "objects": [{"patches": [145, 146, 147, 148, 162, 163, 164, 165, 166, 179, 180, 181, 182, 183, 197, 198, 199, 214, 215, 216, 232, 249], "bbox": [0.5574583333333333, 0.35921875000000003, 0.7813333333333332, 0.6340625], "iscrowd": 0, "area": 10695.418800000003, "rle": {"size": [640, 480], "counts": "[hW5m0lb07L4L5K4L4M4N2N2N3M2N2N3N1Q_OkMP`0Y2k_OiMU`0Z2]_OQNc`0g2N101N10001N101O0O101O001O00001O001O001O00001O001O00001O001O0b@\\L`>d3\\A`Le>`3VAeLi>[3SAiLm>X3n@mLQ?S3k@RMT?n2h@VMX?Y3kAmK]=j3hBZLX=g3cB]L]=c3_BaLa=`3ZBdLf=]3UBgLk=]3mAgLT>\\3dAfL^>[43L5L3L5L3M3L5L3M3L5L3L5L3M4K4M4K5L3M4H7N3M3N1O2M2O2M2O2N2M2O2M2O2N1N3N3L4M2N3L3N3L4M2L5G8EP\\Q2"}, "label": "a chair containing white color seat"}]} {"id": 253942, "image": "COCO_train2014_000000253942.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"zebra who is first in the group\"."}], "task": "refering", "answer_template": "The \"zebra who is first in the group\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [118, 119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 186, 188, 189, 190, 191, 192, 211, 214, 215], "bbox": [0.09721874999999999, 0.3296478873239437, 0.393390625, 0.6262910798122066], "iscrowd": 0, "area": 13321.573549999996, "rle": {"size": [426, 640], "counts": "aPj02k;O]E7a:KYE:e:JTE in this image.", "objects": [{"patches": [118, 119, 120, 121, 122, 123, 140, 141, 142, 143, 144, 145, 146, 163, 164, 165, 166, 167, 168, 169, 186, 188, 189, 190, 191, 192, 211, 214, 215], "bbox": [0.09721874999999999, 0.3296478873239437, 0.393390625, 0.6262910798122066], "iscrowd": 0, "area": 13321.573549999996, "rle": {"size": [426, 640], "counts": "aPj02k;O]E7a:KYE:e:JTE in this image.", "objects": [{"patches": [85, 86, 108, 109, 110, 111, 112, 131, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 179, 181, 182, 202, 204, 205], "bbox": [0.709, 0.24269953051643192, 0.9543125, 0.5685446009389671], "iscrowd": 0, "area": 9784.834099999998, "rle": {"size": [426, 640], "counts": "ePm53T=4L4L4L4L5oN\\OUEh0i:]OjDk0V;d01O001O1O1O2N2N2N2N2N101N2O11O0N3L4M2M4M3O1O1O1N3N1O1O1O1N2O1O1O1O1N2O1O1O1O5J7J7I7I000O10O1YOg04L4K5L4L3M00000000cN\\1010O1O010O0010O01N2O00000O1000000O1000000O1000000O10000O0100000O1000000O10O1000O010000O010O1000O012N2O0O2N2O0O2N2O1Nj0VO010O1O10`NnEFS:;PFAP:`0RFTOW:m0lE\\Ng:e1`01N1O1O100O2N1M3MlNTEN9ROn9n0oE?P:_OVF?h9_O_F=a9AfF;Z9ARG6o8F[G4e8I`G4`8LcG2]8MfG0[80fGOY80kGMV8OPHOP8NUHOk7OYH0g7M^H0oT<"}, "label": "the zebra on the extreme right hand side looking other zebras"}]} {"id": 253942, "image": "COCO_train2014_000000253942.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"the zebra on the right\"."}], "task": "refering", "answer_template": "The \"the zebra on the right\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [85, 86, 108, 109, 110, 111, 112, 131, 132, 133, 134, 135, 136, 155, 156, 157, 158, 159, 179, 181, 182, 202, 204, 205], "bbox": [0.709, 0.24269953051643192, 0.9543125, 0.5685446009389671], "iscrowd": 0, "area": 9784.834099999998, "rle": {"size": [426, 640], "counts": "ePm53T=4L4L4L4L5oN\\OUEh0i:]OjDk0V;d01O001O1O1O2N2N2N2N2N101N2O11O0N3L4M2M4M3O1O1O1N3N1O1O1O1N2O1O1O1O1N2O1O1O1O5J7J7I7I000O10O1YOg04L4K5L4L3M00000000cN\\1010O1O010O0010O01N2O00000O1000000O1000000O1000000O10000O0100000O1000000O10O1000O010000O010O1000O012N2O0O2N2O0O2N2O1Nj0VO010O1O10`NnEFS:;PFAP:`0RFTOW:m0lE\\Ng:e1`01N1O1O100O2N1M3MlNTEN9ROn9n0oE?P:_OVF?h9_O_F=a9AfF;Z9ARG6o8F[G4e8I`G4`8LcG2]8MfG0[80fGOY80kGMV8OPHOP8NUHOk7OYH0g7M^H0oT<"}, "label": "the zebra on the right"}]} {"id": 458751, "image": "COCO_train2014_000000458751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue fire hydrant on a street\"."}], "task": "refering", "answer_template": "The \"a blue fire hydrant on a street\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 90, 91, 92, 110, 111, 112, 113, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 194, 195, 196, 197, 215, 216, 217, 218], "bbox": [0.23699324324324325, 0.1648784722222222, 0.48116554054054056, 0.4936284722222223], "iscrowd": 0, "area": 16238.20485, "rle": {"size": [576, 592], "counts": "hl^21ka05L301O1O1O001O1O10O01O1f_OGk>:n@MQ?3k@3S?Nj@5U?Li@6V?Kh@8V?Hi@:V?4[@Nd?Y3nM01O00000000000000YOkKcCU4ZmNUAQ1m>nNTAQ1n>nNRAQ1o>oNQAT1l>lNTAV1k>hNVA[1g>eNYA^1d>bN\\A_1d>`N\\A`1d>`N\\Aa1c>^N^Ab1c>]N]Ac1i>WNWAi1P?PNPAP2`?000000000000000000000000000000001O00N2O1O1N2O1N3N3L5L3L4M4L3L4M9F;F9F`_\\5"}, "label": "a blue fire hydrant on a street"}]} {"id": 458751, "image": "COCO_train2014_000000458751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a blue fire hydrant being stood on\"."}], "task": "refering", "answer_template": "The \"a blue fire hydrant being stood on\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [69, 70, 90, 91, 92, 110, 111, 112, 113, 131, 132, 133, 134, 135, 152, 153, 154, 155, 156, 157, 173, 174, 175, 176, 177, 194, 195, 196, 197, 215, 216, 217, 218], "bbox": [0.23699324324324325, 0.1648784722222222, 0.48116554054054056, 0.4936284722222223], "iscrowd": 0, "area": 16238.20485, "rle": {"size": [576, 592], "counts": "hl^21ka05L301O1O1O001O1O10O01O1f_OGk>:n@MQ?3k@3S?Nj@5U?Li@6V?Kh@8V?Hi@:V?4[@Nd?Y3nM01O00000000000000YOkKcCU4ZmNUAQ1m>nNTAQ1n>nNRAQ1o>oNQAT1l>lNTAV1k>hNVA[1g>eNYA^1d>bN\\A_1d>`N\\A`1d>`N\\Aa1c>^N^Ab1c>]N]Ac1i>WNWAi1P?PNPAP2`?000000000000000000000000000000001O00N2O1O1N2O1N3N3L5L3L4M4L3L4M9F;F9F`_\\5"}, "label": "a blue fire hydrant being stood on"}]} {"id": 458751, "image": "COCO_train2014_000000458751.jpg", "conversations": [{"from": "human", "value": "Please carefully check the image and detect the object this sentence describes: \"a red and gray fire hydrant filtered in a coloring book style\"."}], "task": "refering", "answer_template": "The \"a red and gray fire hydrant filtered in a coloring book style\" refers to <|Obj_0|> in this image.", "objects": [{"patches": [279, 280, 300, 301, 302, 303, 320, 321, 322, 323, 324, 341, 342, 343, 344, 345, 346, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 404, 405, 406, 407, 425, 426, 427, 428], "bbox": [0.24434121621621624, 0.6336284722222223, 0.4843074324324325, 0.9744444444444444], "iscrowd": 0, "area": 17611.85165, "rle": {"size": [576, 592], "counts": "Uoa29ga0001O001\\_O6P?Jn@9Q?Hm@;Q?En@>P?Cn@>R?Bn@>R?Cm@>R?Bn@>R?Cm@=S?Cm@>S?Bl@>T?X3oM0000000O1000001O0nNjKZDV4d;SLTDn3k;VLRDj3l;YLSDg3k;\\LTDd3k;^LTDb3j;aLTD`3k;bLTD^3k;dLTD]3j;eLUD[3j;gLUDY3j;iLTDX3k;jLTDV3k;lLTDT3k;nLRDT3m;nLQDS3n;oLPDR3o;`1O101O000001O000000000000010O000000000000001O1O2O0O1O1O2N1O in this image.", "objects": [{"patches": [279, 280, 300, 301, 302, 303, 320, 321, 322, 323, 324, 341, 342, 343, 344, 345, 346, 362, 363, 364, 365, 366, 367, 383, 384, 385, 386, 387, 404, 405, 406, 407, 425, 426, 427, 428], "bbox": [0.24434121621621624, 0.6336284722222223, 0.4843074324324325, 0.9744444444444444], "iscrowd": 0, "area": 17611.85165, "rle": {"size": [576, 592], "counts": "Uoa29ga0001O001\\_O6P?Jn@9Q?Hm@;Q?En@>P?Cn@>R?Bn@>R?Cm@>R?Bn@>R?Cm@=S?Cm@>S?Bl@>T?X3oM0000000O1000001O0nNjKZDV4d;SLTDn3k;VLRDj3l;YLSDg3k;\\LTDd3k;^LTDb3j;aLTD`3k;bLTD^3k;dLTD]3j;eLUD[3j;gLUDY3j;iLTDX3k;jLTDV3k;lLTDT3k;nLRDT3m;nLQDS3n;oLPDR3o;`1O101O000001O000000000000010O000000000000001O1O2O0O1O1O2N1O in this image.", "objects": [{"patches": [77, 78, 97, 98, 99, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 159, 160, 161, 162, 163, 164, 181, 182, 183, 184, 185, 202, 203, 204, 223, 224, 225], "bbox": [0.6029391891891892, 0.15904513888888888, 0.844543918918919, 0.4856770833333333], "iscrowd": 0, "area": 16687.808200000003, "rle": {"size": [576, 592], "counts": "`nX62`10l>2QA1m>2o@1o>1o@1o>2m@2P?0m@3P?1l@2R?0k@3S?1h@2V?2e@2X?2c@1[?2a@1\\?4_@O_?d1E?A?A?A2N0O1000000000000O1000000000OYOgKhCZ4V in this image.", "objects": [{"patches": [77, 78, 97, 98, 99, 117, 118, 119, 120, 121, 139, 140, 141, 142, 143, 159, 160, 161, 162, 163, 164, 181, 182, 183, 184, 185, 202, 203, 204, 223, 224, 225], "bbox": [0.6029391891891892, 0.15904513888888888, 0.844543918918919, 0.4856770833333333], "iscrowd": 0, "area": 16687.808200000003, "rle": {"size": [576, 592], "counts": "`nX62`10l>2QA1m>2o@1o>1o@1o>2m@2P?0m@3P?1l@2R?0k@3S?1h@2V?2e@2X?2c@1[?2a@1\\?4_@O_?d1E?A?A?A2N0O1000000000000O1000000000OYOgKhCZ4V