split stringclasses 1 value | image_id stringlengths 12 25 | file_name stringlengths 16 29 | image_info dict | caption_info dict | mask_annotations listlengths 2 52 | categories listlengths 1 1 |
|---|---|---|---|---|---|---|
train | 000000578736 | 000000578736.jpg | {
"data_source": "COCONut",
"file_name": "000000578736.jpg",
"height": 439,
"id": "000000578736",
"width": 640
} | {
"caption": "The image shows a person walking barefoot on the sandy beach, carrying a long white surfboard with blue floral patterns overhead. The woman wearing a bikini is positioned slightly to the right, while few green palm leaves appear in the lower-left corner of the tropical scene.",
"caption_ann": "The image shows a <3:person walking barefoot> on the <0:sandy beach>, carrying a <2:long white surfboard with blue floral patterns> overhead. The <3:woman wearing a bikini> is positioned slightly to the right, while few <1:green palm leaves> appear in the lower-left corner of the tropical scene.",
"id": 1300,
"image_id": "000000578736",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "person walking barefoot"
},
{
"mask_ids": [
0
],
"txt_desc": "sandy beach"
},
{
"mask_ids": [
2
],
"txt_desc": "long white surfboard with blue floral patterns"
},
{
"mask_ids": [
3
],
"txt_desc": "woman wearing a bikini"
},
{
"mask_ids": [
1
],
"txt_desc": "green palm leaves"
}
],
"labels": [
"sand",
"tree-merged",
"surfboard",
"person"
]
} | [
{
"area": 223071,
"bbox": [
0,
0,
640,
439
],
"category_id": 154,
"id": 14936,
"image_id": "000000578736",
"iscrowd": 0,
"segmentation": {
"counts": "0Q9f400O11O1O0000000000000000O100000000<[KeFV4_9O001O00001O000RM`Fe1`9[NdFa1]9]NfFa1Z9^NhFa1b9TN_Fl1f9nM... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000579341 | 000000579341.jpg | {
"data_source": "COCONut",
"file_name": "000000579341.jpg",
"height": 394,
"id": "000000579341",
"width": 640
} | {
"caption": "The image shows two aircraft performing a synchronized aerobatic maneuver in the sky. The plane on the left is flying upright, while the plane on the right is inverted beneath it, perfectly aligned in a precise formation. Their paths intersect closely, creating the illusion of a near-miss. The background features a clear blue sky with faint clouds and trails of white smoke.",
"caption_ann": "The image shows two <1,2:aircraft> performing a synchronized aerobatic maneuver in the <0:sky>. The <2:plane on the left> is flying upright, while the <1:plane on the right> is inverted beneath it, perfectly aligned in a precise formation. Their paths intersect closely, creating the illusion of a near-miss. The background features a <0:clear blue sky with faint clouds and trails of white smoke>.",
"id": 1301,
"image_id": "000000579341",
"label_matched": [
{
"mask_ids": [
1,
2
],
"txt_desc": "aircraft"
},
{
"mask_ids": [
0
],
"txt_desc": "sky"
},
{
"mask_ids": [
2
],
"txt_desc": "plane on the left"
},
{
"mask_ids": [
1
],
"txt_desc": "plane on the right"
},
{
"mask_ids": [
0
],
"txt_desc": "clear blue sky with faint clouds and trails of white smoke"
}
],
"labels": [
"sky-other-merged",
"airplane",
"airplane"
]
} | [
{
"area": 244196,
"bbox": [
0,
0,
640,
394
],
"category_id": 187,
"id": 14940,
"image_id": "000000579341",
"iscrowd": 0,
"segmentation": {
"counts": "0lgf21\\dYM2O1O001O1N101O001O00001O001O001O01O01O001O00001O00001O1O2N2N1O1O1O1O001O00001O01O01BWOYEi0f:Y... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000579453 | 000000579453.jpg | {
"data_source": "COCONut",
"file_name": "000000579453.jpg",
"height": 500,
"id": "000000579453",
"width": 336
} | {
"caption": "The image shows a bouquet of vibrant yellow daffodils with orange centers arranged in a simple gray vase. The flowers are placed on a smooth, light-colored surface, close to a dark black background on the upper part of the image. ",
"caption_ann": "The image shows a <2:bouquet of vibrant yellow daffodils with orange centers> arranged in a <3:simple gray vase>. The <2:flowers> are placed on a <1:smooth, light-colored surface>, close to a <0:dark black background> on the upper part of the image. ",
"id": 1302,
"image_id": "000000579453",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "bouquet of vibrant yellow daffodils with orange centers"
},
{
"mask_ids": [
3
],
"txt_desc": "simple gray vase"
},
{
"mask_ids": [
2
],
"txt_desc": "flowers"
},
{
"mask_ids": [
1
],
"txt_desc": "smooth, light-colored surface"
},
{
"mask_ids": [
0
],
"txt_desc": "dark black background"
}
],
"labels": [
"wall-other-merged",
"table-merged",
"flower",
"vase"
]
} | [
{
"area": 16354,
"bbox": [
14,
0,
322,
80
],
"category_id": 199,
"id": 14943,
"image_id": "000000579453",
"iscrowd": 0,
"segmentation": {
"counts": "_l6:20a>h0L5L2N2O00O0HiAVOW>l0nAmNR>T161XOPOlBn0T=VOiBj0V=YOhBb0]=@aB?`=B_B=c=EZB;g=h01O1O1O1ROSB4n=KVB2j... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000579466 | 000000579466.jpg | {
"data_source": "COCONut",
"file_name": "000000579466.jpg",
"height": 480,
"id": "000000579466",
"width": 640
} | {
"caption": "The image shows a baby wearing a pink long-sleeve shirt standing next to an Acer white laptop placed on a grey couch. The baby is reaching out with both hands, pressing the keyboard, and looking up with a curious and slightly surprised expression. The floor beneath is carpeted, and the overall setting appears to be indoors.",
"caption_ann": "The image shows a <1:baby wearing a pink long-sleeve shirt> standing next to an <2:Acer white laptop> placed on a <3:grey couch>. The <1:baby> is reaching out with both hands, pressing the <4:keyboard>, and looking up with a curious and slightly surprised expression. The <0:floor> beneath is carpeted, and the overall setting appears to be indoors.",
"id": 1303,
"image_id": "000000579466",
"label_matched": [
{
"mask_ids": [
1
],
"txt_desc": "baby wearing a pink long-sleeve shirt"
},
{
"mask_ids": [
2
],
"txt_desc": "Acer white laptop"
},
{
"mask_ids": [
3
],
"txt_desc": "grey couch"
},
{
"mask_ids": [
1
],
"txt_desc": "baby"
},
{
"mask_ids": [
4
],
"txt_desc": "keyboard"
},
{
"mask_ids": [
0
],
"txt_desc": "floor"
}
],
"labels": [
"rug-merged",
"person",
"laptop",
"couch",
"keyboard"
]
} | [
{
"area": 40341,
"bbox": [
243,
0,
397,
480
],
"category_id": 200,
"id": 14947,
"image_id": "000000579466",
"iscrowd": 0,
"segmentation": {
"counts": "o[b31n>1O1O1O1N2O1N2O1O1O1O1O1O1N2O1O1O1O1O1N2O1N2O1O1O1O1O1O1O1N2O1O1O1O1N2N2O100O1O1N2O1O1O1N2O1N2O1O... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000579785 | 000000579785.jpg | {
"data_source": "COCONut",
"file_name": "000000579785.jpg",
"height": 427,
"id": "000000579785",
"width": 640
} | {
"caption": "The image shows three people closely gathered, focusing on a smartphone. The person in the middle, dressed in a white and black polka-dot top, appears to be showing something on the phone. On the right, a person wearing a light blue graphic shirt and glasses leans in with curiosity, while the individual on the left, dressed in a purple and white striped hoodie, looks closely as well. The setting appears to be indoors, with beige-colored walls and a white door and frames visible in the background.",
"caption_ann": "The image shows three <3,4,5:people> closely gathered, focusing on a <2:smartphone>. The <4:person in the middle, dressed in a white and black polka-dot top>, appears to be showing something on the <2:phone>. On the right, a <3:person wearing a light blue graphic shirt and glasses> leans in with curiosity, while the <5:individual on the left, dressed in a purple and white striped hoodie>, looks closely as well. The setting appears to be indoors, with <0:beige-colored walls> and a <1:white door and frames> visible in the background.",
"id": 1304,
"image_id": "000000579785",
"label_matched": [
{
"mask_ids": [
3,
4,
5
],
"txt_desc": "people"
},
{
"mask_ids": [
2
],
"txt_desc": "smartphone"
},
{
"mask_ids": [
4
],
"txt_desc": "person in the middle, dressed in a white and black polka-dot top"
},
{
"mask_ids": [
2
],
"txt_desc": "phone"
},
{
"mask_ids": [
3
],
"txt_desc": "person wearing a light blue graphic shirt and glasses"
},
{
"mask_ids": [
5
],
"txt_desc": "individual on the left, dressed in a purple and white striped hoodie"
},
{
"mask_ids": [
0
],
"txt_desc": "beige-colored walls"
},
{
"mask_ids": [
1
],
"txt_desc": "white door and frames"
}
],
"labels": [
"wall-other-merged",
"door-stuff",
"cell phone",
"person",
"person",
"person"
]
} | [
{
"area": 45633,
"bbox": [
0,
0,
640,
427
],
"category_id": 199,
"id": 14952,
"image_id": "000000579785",
"iscrowd": 0,
"segmentation": {
"counts": "0\\2o:O100O1O100O100O100O1001O1O1O001O1O1O1O00O1O1O100O1O1O1O10000O100O100O100O10000O10000O10000O10000O10... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000580104 | 000000580104.jpg | {
"data_source": "COCONut",
"file_name": "000000580104.jpg",
"height": 500,
"id": "000000580104",
"width": 375
} | {
"caption": "The image shows a man with a beard and glasses wearing a grey t-shirt taking a selfie using an Apple white smartphone, with a serious expression on his face. In the lower-left corner of the mirror, a black and white cat is partially visible, sitting in the background. The setting features blue walls, and a shower area with light-colored tiles is visible in the background.",
"caption_ann": "The image shows a <3:man with a beard and glasses wearing a grey t-shirt> taking a selfie using an <4:Apple white smartphone>, with a serious expression on his face. In the lower-left corner of the mirror, a <2:black and white cat> is partially visible, sitting in the background. The setting features <0:blue walls>, and a <1:shower area with light-colored tiles> is visible in the background.",
"id": 1305,
"image_id": "000000580104",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "man with a beard and glasses wearing a grey t-shirt"
},
{
"mask_ids": [
4
],
"txt_desc": "Apple white smartphone"
},
{
"mask_ids": [
2
],
"txt_desc": "black and white cat"
},
{
"mask_ids": [
0
],
"txt_desc": "blue walls"
},
{
"mask_ids": [
1
],
"txt_desc": "shower area with light-colored tiles"
}
],
"labels": [
"wall-other-merged",
"wall-tile",
"cat",
"person",
"cell phone"
]
} | [
{
"area": 77910,
"bbox": [
0,
0,
375,
500
],
"category_id": 199,
"id": 14958,
"image_id": "000000580104",
"iscrowd": 0,
"segmentation": {
"counts": "0gR`08b\\@<D4L5J5cN\\ORC[1_<TO[CR12SNR<\\3B>E;B>B>M3E;[Oe0CXKPGf3f8i1B>D<E;C=B>C=E;XOeFjJh9o4b0K5C\\KcKkN... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000580162 | 000000580162.jpg | {
"data_source": "COCONut",
"file_name": "000000580162.jpg",
"height": 480,
"id": "000000580162",
"width": 640
} | {
"caption": "The image shows a relaxed gray and white cat with green eyes, lying stretched out on top of a gray microwave. The cat rests elegantly on the metallic object with an alert expression. Behind, a gray wall and a white cabinets are partially visible.",
"caption_ann": "The image shows a relaxed <2:gray and white cat with green eyes>, lying stretched out on top of a <3:gray microwave>. The <2:cat> rests elegantly on the <3:metallic object> with an alert expression. Behind, a <0:gray wall> and a <1:white cabinets> are partially visible.",
"id": 1306,
"image_id": "000000580162",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "gray and white cat with green eyes"
},
{
"mask_ids": [
3
],
"txt_desc": "gray microwave"
},
{
"mask_ids": [
2
],
"txt_desc": "cat"
},
{
"mask_ids": [
3
],
"txt_desc": "metallic object"
},
{
"mask_ids": [
0
],
"txt_desc": "gray wall"
},
{
"mask_ids": [
1
],
"txt_desc": "white cabinets"
}
],
"labels": [
"wall-other-merged",
"cabinet-merged",
"cat",
"microwave"
]
} | [
{
"area": 102065,
"bbox": [
0,
56,
640,
292
],
"category_id": 199,
"id": 14963,
"image_id": "000000580162",
"iscrowd": 0,
"segmentation": {
"counts": "n1n8S6N10000O10000O100hNkIeIU6[6lIcIU6]6mIaIS6_6nI`IR6`6oI^IR6a6QJ]Io5c6RJ\\In5d6TJYIm5g6TJXIl5h6UJWIk5... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000580363 | 000000580363.jpg | {
"data_source": "COCONut",
"file_name": "000000580363.jpg",
"height": 480,
"id": "000000580363",
"width": 640
} | {
"caption": "The image captures an indoor scene with sunlight streaming through a frosted glass window with a light wood frame. Below the window, a light-colored wall is visible. In the foreground, a blue and white fabric surface possibly a blanket or bed cover, is gently illuminated by the warm, diffused sunlight. The out-of-focus background and soft shadows create a calm and cozy atmosphere.",
"caption_ann": "The image captures an indoor scene with sunlight streaming through a <2:frosted glass window> with a <1:light wood frame>. Below the window, a <0:light-colored wall> is visible. In the foreground, a <3:blue and white fabric surface possibly a blanket or bed cover>, is gently illuminated by the warm, diffused sunlight. The out-of-focus background and soft shadows create a calm and cozy atmosphere.",
"id": 1307,
"image_id": "000000580363",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "frosted glass window"
},
{
"mask_ids": [
1
],
"txt_desc": "light wood frame"
},
{
"mask_ids": [
0
],
"txt_desc": "light-colored wall"
},
{
"mask_ids": [
3
],
"txt_desc": "blue and white fabric surface possibly a blanket or bed cover"
}
],
"labels": [
"wall-other-merged",
"wall-wood",
"window-other",
"bed"
]
} | [
{
"area": 58531,
"bbox": [
0,
194,
640,
146
],
"category_id": 199,
"id": 14967,
"image_id": "000000580363",
"iscrowd": 0,
"segmentation": {
"counts": "a6S4m:00O10002N3L8I1O1O0O1000000O1000000O0100000O1000000O1000000O1000000O100000O0100000000O100000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000580549 | 000000580549.jpg | {
"data_source": "COCONut",
"file_name": "000000580549.jpg",
"height": 480,
"id": "000000580549",
"width": 640
} | {
"caption": "The image shows two orange and black buses parked side by side in a bus station or parking area. The bus in the foreground displays the destination \\\"Giggleswick 580\\\" on its signboard. Behind the buses, there is a pavemented area with a fence running along its edge. A car can be seen parked further back near the buildings, which are constructed with classic stone architecture. In the background, there are several trees adding greenery to the urban setting and the sky above is bright.",
"caption_ann": "The image shows two <6,7:orange and black buses> parked side by side in a <1:bus station or parking area>. The <6:bus in the foreground> displays the destination \\\"Giggleswick 580\\\" on its signboard. Behind the <6,7:buses>, there is a <5:pavemented area> with a <4:fence> running along its edge. A <8:car> can be seen parked further back near the <2:buildings>, which are constructed with classic stone architecture. In the background, there are several <3:trees> adding greenery to the urban setting and the <0:sky> above is bright.",
"id": 1308,
"image_id": "000000580549",
"label_matched": [
{
"mask_ids": [
6,
7
],
"txt_desc": "orange and black buses"
},
{
"mask_ids": [
1
],
"txt_desc": "bus station or parking area"
},
{
"mask_ids": [
6
],
"txt_desc": "bus in the foreground"
},
{
"mask_ids": [
6,
7
],
"txt_desc": "buses"
},
{
"mask_ids": [
5
],
"txt_desc": "pavemented area"
},
{
"mask_ids": [
4
],
"txt_desc": "fence"
},
{
"mask_ids": [
8
],
"txt_desc": "car"
},
{
"mask_ids": [
2
],
"txt_desc": "buildings"
},
{
"mask_ids": [
3
],
"txt_desc": "trees"
},
{
"mask_ids": [
0
],
"txt_desc": "sky"
}
],
"labels": [
"sky-other-merged",
"road",
"house",
"tree-merged",
"fence-merged",
"pavement-merged",
"bus",
"bus",
"car"
]
} | [
{
"area": 42740,
"bbox": [
0,
0,
502,
126
],
"category_id": 187,
"id": 14971,
"image_id": "000000580549",
"iscrowd": 0,
"segmentation": {
"counts": "0o2Q<000000O1bNoCoN1:O[Oo;\\1QDoN25V<k0kClN1b0k;d0QDMo;b10JRDZMn;_2=O1O100000000O10000N2BeCRN`<n190000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000580625 | 000000580625.jpg | {
"data_source": "COCONut",
"file_name": "000000580625.jpg",
"height": 480,
"id": "000000580625",
"width": 640
} | {
"caption": "The image shows a modern living room featuring a red large couch and a single-seater armchair. In front of the seating area is a glass-top black coffee table with some paper on top. To the right, there is a brick fireplace with a decorative orange vase placed on it. A wooden cabinet stands beside the fireplace, and a window with brown curtains is visible above the tiled floor. The room has a clean and minimalistic design with neutral-colored walls and flooring.",
"caption_ann": "The image shows a modern living room featuring a <7:red large couch> and a <8:single-seater armchair>. In front of the seating area is a <3:glass-top black coffee table> with some <6:paper> on top. To the right, there is a <2:brick fireplace> with a <9:decorative orange vase> placed on it. A <4:wooden cabinet> stands beside the fireplace, and a <5:window with brown curtains> is visible above the tiled floor. The room has a clean and minimalistic design with <1:neutral-colored walls> and <0:flooring>.",
"id": 1309,
"image_id": "000000580625",
"label_matched": [
{
"mask_ids": [
7
],
"txt_desc": "red large couch"
},
{
"mask_ids": [
8
],
"txt_desc": "single-seater armchair"
},
{
"mask_ids": [
3
],
"txt_desc": "glass-top black coffee table"
},
{
"mask_ids": [
6
],
"txt_desc": "paper"
},
{
"mask_ids": [
2
],
"txt_desc": "brick fireplace"
},
{
"mask_ids": [
9
],
"txt_desc": "decorative orange vase"
},
{
"mask_ids": [
4
],
"txt_desc": "wooden cabinet"
},
{
"mask_ids": [
5
],
"txt_desc": "window with brown curtains"
},
{
"mask_ids": [
1
],
"txt_desc": "neutral-colored walls"
},
{
"mask_ids": [
0
],
"txt_desc": "flooring"
}
],
"labels": [
"floor-other-merged",
"wall-other-merged",
"wall-brick",
"table-merged",
"cabinet-merged",
"curtain",
"paper-merged",
"couch",
"chair",
"vase"
]
} | [
{
"area": 38542,
"bbox": [
160,
194,
414,
286
],
"category_id": 190,
"id": 14980,
"image_id": "000000580625",
"iscrowd": 0,
"segmentation": {
"counts": "XV[24k>4L4M2N2M3N1O001O0O100000001O0000000000000004LN21O0O10000000O100001O000000001OO10000N21O1O00000... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000580829 | 000000580829.jpg | {
"data_source": "COCONut",
"file_name": "000000580829.jpg",
"height": 640,
"id": "000000580829",
"width": 480
} | {
"caption": "The image shows a winding mountain road bordered by lush green grass and dense bushes. A yellow car is seen driving along the curved road. The surrounding mountains are covered in patches of greenery and exposed rock.",
"caption_ann": "The image shows a <0:winding mountain road> bordered by <3:lush green grass> and <1:dense bushes>. A <4:yellow car> is seen driving along the <0:curved road>. The surrounding <2:mountains> are covered in patches of greenery and exposed rock.",
"id": 1310,
"image_id": "000000580829",
"label_matched": [
{
"mask_ids": [
0
],
"txt_desc": "winding mountain road"
},
{
"mask_ids": [
3
],
"txt_desc": "lush green grass"
},
{
"mask_ids": [
1
],
"txt_desc": "dense bushes"
},
{
"mask_ids": [
4
],
"txt_desc": "yellow car"
},
{
"mask_ids": [
0
],
"txt_desc": "curved road"
},
{
"mask_ids": [
2
],
"txt_desc": "mountains"
}
],
"labels": [
"road",
"tree-merged",
"mountain-merged",
"grass-merged",
"car"
]
} | [
{
"area": 87611,
"bbox": [
0,
379,
480,
261
],
"category_id": 149,
"id": 14990,
"image_id": "000000580829",
"iscrowd": 0,
"segmentation": {
"counts": "P<P8P<0000000000000000000000000O100000000000000000000000000001O00000000000000000000000000O100000000O100... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000580851 | 000000580851.jpg | {
"data_source": "COCONut",
"file_name": "000000580851.jpg",
"height": 640,
"id": "000000580851",
"width": 480
} | {
"caption": "The image shows a metallic cup holding toothbrushes with white and blue handles. The toothbrushes have bristles in different colors. The cup is placed on a wooden surface against a tiled wall background.",
"caption_ann": "The image shows a <7:metallic cup> holding <2,3,4,5,6:toothbrushes with white and blue handles>. The <2,3,4,5,6:toothbrushes>have bristles in different colors. The cup is placed on a <1:wooden surface> against a <0:tiled wall background>.",
"id": 1311,
"image_id": "000000580851",
"label_matched": [
{
"mask_ids": [
7
],
"txt_desc": "metallic cup"
},
{
"mask_ids": [
2,
3,
4,
5,
6
],
"txt_desc": "toothbrushes with white and blue handles"
},
{
"mask_ids": [
2,
3,
4,
5,
6
],
"txt_desc": "toothbrushes"
},
{
"mask_ids": [
1
],
"txt_desc": "wooden surface"
},
{
"mask_ids": [
0
],
"txt_desc": "tiled wall background"
}
],
"labels": [
"wall-tile",
"table-merged",
"toothbrush",
"toothbrush",
"toothbrush",
"toothbrush",
"toothbrush",
"cup"
]
} | [
{
"area": 125902,
"bbox": [
0,
0,
480,
441
],
"category_id": 176,
"id": 14995,
"image_id": "000000580851",
"iscrowd": 0,
"segmentation": {
"counts": "0`=`600fFaIS5_6lJlIj4T6VKnIh4R6XKPJf4P6YKSJe4m5[KVJb4j5^KWJa4i5^KZJ`4f5`K\\J^4d5bK_J[4a5dKcJY4]5gKhJT4X5... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000580905 | 000000580905.jpg | {
"data_source": "COCONut",
"file_name": "000000580905.jpg",
"height": 481,
"id": "000000580905",
"width": 640
} | {
"caption": "A woman in a black vest, a green top and jeans wearing a black bag is holding a green umbrella with a colorful patterns and pointing it forward. A child in a pink floral jacket is climbing a stack of hay bales that extends across the ground. In the background, there are two cars, lush trees, a person in a purple-pink coat, and the cloudy sky overhead.",
"caption_ann": "A <4:woman in a black vest, a green top and jeans> wearing a <7:black bag> is holding a <3:green umbrella with a colorful patterns> and pointing it forward. A <5:child in a pink floral jacket> is climbing a <1:stack of hay bales> that extends across the ground. In the background, there are two <8,9:cars>, <2:lush trees>, a <6:person in a purple-pink coat>, and the <0:cloudy sky> overhead.",
"id": 1312,
"image_id": "000000580905",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "woman in a black vest, a green top and jeans"
},
{
"mask_ids": [
7
],
"txt_desc": "black bag"
},
{
"mask_ids": [
3
],
"txt_desc": "green umbrella with a colorful patterns"
},
{
"mask_ids": [
5
],
"txt_desc": "child in a pink floral jacket"
},
{
"mask_ids": [
1
],
"txt_desc": "stack of hay bales"
},
{
"mask_ids": [
8,
9
],
"txt_desc": "cars"
},
{
"mask_ids": [
2
],
"txt_desc": "lush trees"
},
{
"mask_ids": [
6
],
"txt_desc": "person in a purple-pink coat"
},
{
"mask_ids": [
0
],
"txt_desc": "cloudy sky"
}
],
"labels": [
"sky-other-merged",
"grass-merged",
"tree-merged",
"umbrella",
"person",
"person",
"person",
"handbag",
"car",
"car"
]
} | [
{
"area": 61240,
"bbox": [
0,
0,
640,
165
],
"category_id": 187,
"id": 15003,
"image_id": "000000580905",
"iscrowd": 0,
"segmentation": {
"counts": "0h1Y=001O0000O11O00000000000000000000001O000000001O00001O00001O1O00001O1O00001O001O001O1O001O0000001O0000... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000580951 | 000000580951.jpg | {
"data_source": "COCONut",
"file_name": "000000580951.jpg",
"height": 640,
"id": "000000580951",
"width": 425
} | {
"caption": "The image shows an elephant standing amidst tall green grass near a body of water. The water’s surface is calm and reflective and the sky above is clear, creating a calm and natural atmosphere.",
"caption_ann": "The image shows an <3:elephant> standing amidst <2:tall green grass> near a <1:body of water>. The <1:water>’s surface is calm and reflective and the <0:sky> above is clear, creating a calm and natural atmosphere.",
"id": 1313,
"image_id": "000000580951",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "elephant"
},
{
"mask_ids": [
2
],
"txt_desc": "tall green grass"
},
{
"mask_ids": [
1
],
"txt_desc": "body of water"
},
{
"mask_ids": [
1
],
"txt_desc": "water"
},
{
"mask_ids": [
0
],
"txt_desc": "sky"
}
],
"labels": [
"sky-other-merged",
"water-other",
"grass-merged",
"elephant"
]
} | [
{
"area": 33307,
"bbox": [
0,
0,
425,
190
],
"category_id": 187,
"id": 15013,
"image_id": "000000580951",
"iscrowd": 0,
"segmentation": {
"counts": "0i3W`0O100O1O1O1Mf_O_L[`0\\3f_OfL\\`0\\332N2N2N001O1O00001OO10000O100O1N2L4Gb_OnL``0R370000MW_OnL13c`0W3]... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000581249 | 000000581249.jpg | {
"data_source": "COCONut",
"file_name": "000000581249.jpg",
"height": 640,
"id": "000000581249",
"width": 480
} | {
"caption": "n a room with a beige rug, a grey tabby cat walks in the foreground while a black and white cat sits on top of a black suitcase placed in front of a wooden cabinet, which holds a black television and two potted plants against a light yellow wall next to a wooden door.",
"caption_ann": "n a room with a <0:beige rug>, a <7:grey tabby cat> walks in the foreground while a <9:black and white cat> sits on top of a <8:black suitcase> placed in front of a <2:wooden cabinet>, which holds a <6:black television> and <4,5:two potted plants> against a <1:light yellow wall> next to a <3:wooden door>.",
"id": 1314,
"image_id": "000000581249",
"label_matched": [
{
"mask_ids": [
0
],
"txt_desc": "beige rug"
},
{
"mask_ids": [
7
],
"txt_desc": "grey tabby cat"
},
{
"mask_ids": [
9
],
"txt_desc": "black and white cat"
},
{
"mask_ids": [
8
],
"txt_desc": "black suitcase"
},
{
"mask_ids": [
2
],
"txt_desc": "wooden cabinet"
},
{
"mask_ids": [
6
],
"txt_desc": "black television"
},
{
"mask_ids": [
4,
5
],
"txt_desc": "two potted plants"
},
{
"mask_ids": [
1
],
"txt_desc": "light yellow wall"
},
{
"mask_ids": [
3
],
"txt_desc": "wooden door"
}
],
"labels": [
"rug-merged",
"wall-other-merged",
"cabinet-merged",
"door-stuff",
"potted plant",
"potted plant",
"tv",
"cat",
"suitcase",
"cat"
]
} | [
{
"area": 89532,
"bbox": [
0,
296,
480,
344
],
"category_id": 200,
"id": 15017,
"image_id": "000000581249",
"iscrowd": 0,
"segmentation": {
"counts": "j`0V3c`07K5M3O10000O1L4M3L4L4L4L4M3L4M3L4M3M3L4L4L4L4N23M2N2VKhAh3\\>QLoAg3U>QLUBi3S?N3M3M_IcLcLY3X3XM_... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000581297 | 000000581297.jpg | {
"data_source": "COCONut",
"file_name": "000000581297.jpg",
"height": 427,
"id": "000000581297",
"width": 640
} | {
"caption": "A young soccer player in a blue jersey and red shorts is dribbling a white and blue soccer ball on a grass field. In the background, there's a goalkeeper in colorful gear standing near the goalpost, slightly out of focus, giving the impression of an active soccer match.",
"caption_ann": "A <3:young soccer player in a blue jersey and red shorts> is dribbling a <2:white and blue soccer ball> on a <1:grass field>. In the background, there's a <4:goalkeeper in colorful gear> standing near the <0:goalpost>, slightly out of focus, giving the impression of an active soccer match.",
"id": 1315,
"image_id": "000000581297",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "young soccer player in a blue jersey and red shorts"
},
{
"mask_ids": [
2
],
"txt_desc": "white and blue soccer ball"
},
{
"mask_ids": [
1
],
"txt_desc": "grass field"
},
{
"mask_ids": [
4
],
"txt_desc": "goalkeeper in colorful gear"
},
{
"mask_ids": [
0
],
"txt_desc": "goalpost"
}
],
"labels": [
"net",
"playingfield",
"sports ball",
"person",
"person"
]
} | [
{
"area": 122639,
"bbox": [
0,
0,
640,
247
],
"category_id": 138,
"id": 15027,
"image_id": "000000581297",
"iscrowd": 0,
"segmentation": {
"counts": "0g7d500000000000000000000000000O100000000000000000000000000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000581338 | 000000581338.jpg | {
"data_source": "COCONut",
"file_name": "000000581338.jpg",
"height": 427,
"id": "000000581338",
"width": 640
} | {
"caption": "The image shows a close-up of a black toothbrush with white bristles placed on a wooden surface. The focus is on the toothbrush while the background surface is blurred. ",
"caption_ann": "The image shows a close-up of a <1:black toothbrush with white bristles> placed on a <0:wooden surface>. The focus is on the <1:toothbrush> while the <0:background surface> is blurred. ",
"id": 1316,
"image_id": "000000581338",
"label_matched": [
{
"mask_ids": [
1
],
"txt_desc": "black toothbrush with white bristles"
},
{
"mask_ids": [
0
],
"txt_desc": "wooden surface"
},
{
"mask_ids": [
1
],
"txt_desc": "toothbrush"
},
{
"mask_ids": [
0
],
"txt_desc": "background surface"
}
],
"labels": [
"table-merged",
"toothbrush"
]
} | [
{
"area": 230048,
"bbox": [
0,
0,
640,
427
],
"category_id": 189,
"id": 15032,
"image_id": "000000581338",
"iscrowd": 0,
"segmentation": {
"counts": "1Y=2mZ8OTeG000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000581362 | 000000581362.jpg | {
"data_source": "COCONut",
"file_name": "000000581362.jpg",
"height": 480,
"id": "000000581362",
"width": 640
} | {
"caption": "The image shows a close-up of a freshly baked bread roll placed on a textured white paper towel. In the background, there's a glazed donut with a shiny golden-brown surface, slightly out of focus.",
"caption_ann": "The image shows a close-up of a <1:freshly baked bread roll> placed on a <0:textured white paper towel>. In the background, there's a <2:glazed donut with a shiny golden-brown surface>, slightly out of focus.",
"id": 1317,
"image_id": "000000581362",
"label_matched": [
{
"mask_ids": [
1
],
"txt_desc": "freshly baked bread roll"
},
{
"mask_ids": [
0
],
"txt_desc": "textured white paper towel"
},
{
"mask_ids": [
2
],
"txt_desc": "glazed donut with a shiny golden-brown surface"
}
],
"labels": [
"paper-merged",
"food-other-merged",
"donut"
]
} | [
{
"area": 135468,
"bbox": [
0,
0,
640,
480
],
"category_id": 195,
"id": 15034,
"image_id": "000000581362",
"iscrowd": 0,
"segmentation": {
"counts": "0S[d0>\\S\\O7J4K5K4K6K4N3L3M4M2N3M2N2M4M2N3M2N2N2O1N2N1O2N2O0O2N2O0O2N1O2O1N1O2O0O2O1N100O2O0O2O0O2O0O2O... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000581419 | 000000581419.jpg | {
"data_source": "COCONut",
"file_name": "000000581419.jpg",
"height": 612,
"id": "000000581419",
"width": 612
} | {
"caption": "The image shows four black ceramic clock faces, each featuring white numbers around the edges but missing clock hands. A small white sticker labeled “6-” is placed at the center of each plate. The plates are arranged on a patterned brown-tiled floor. At the bottom of the image, a person wearing black open-toe sandals with red-painted toenails is visible, standing near the clocks.",
"caption_ann": "The image shows four <1,3,4,5:black ceramic clock faces>, each featuring white numbers around the edges but missing clock hands. A small white sticker labeled “6-” is placed at the center of each plate. The plates are arranged on a <0:patterned brown-tiled floor>. At the bottom of the image, a <2:person wearing black open-toe sandals with red-painted toenails> is visible, standing near the <1,3,4,5:clocks>.",
"id": 1318,
"image_id": "000000581419",
"label_matched": [
{
"mask_ids": [
1,
3,
4,
5
],
"txt_desc": "black ceramic clock faces"
},
{
"mask_ids": [
0
],
"txt_desc": "patterned brown-tiled floor"
},
{
"mask_ids": [
2
],
"txt_desc": "person wearing black open-toe sandals with red-painted toenails"
},
{
"mask_ids": [
1,
3,
4,
5
],
"txt_desc": "clocks"
}
],
"labels": [
"floor-other-merged",
"clock",
"person",
"clock",
"clock",
"clock"
]
} | [
{
"area": 170593,
"bbox": [
0,
0,
612,
612
],
"category_id": 190,
"id": 15037,
"image_id": "000000581419",
"iscrowd": 0,
"segmentation": {
"counts": "0YY7a0nXIh0^O=D;C:J7H9H7J5K4J7K6K4L2M4M3M4K3N3M2N3L4M2N2N2M3N3N1N2N1O2O1N2O1N2N101N2O1N2O1N100O2O1N101O1... | [
{
"id": 1,
"name": "object"
}
] |
train | 000000581422 | 000000581422.jpg | {
"data_source": "COCONut",
"file_name": "000000581422.jpg",
"height": 498,
"id": "000000581422",
"width": 640
} | {
"caption": "The black-and-white image shows two men sitting on a wooden crates outdoors, in a field under the shade of a tree. The man on the left with a wide-brimmed hat, a shirt, and a tie is writing in a notebook, while the man on the right in a buttoned-up work shirt, looks at him attentively. Beside them, a calf with a black-and-white coat stands calmly, facing the camera. In the background, a wooden barn or shed is visible along with the sky giving the scene a rural, farm-like setting.",
"caption_ann": "The black-and-white image shows two <4,5:men> sitting on a <6:wooden crates> outdoors, in a <3:field> under the shade of a <2:tree>. The <4:man on the left with a wide-brimmed hat, a shirt, and a tie> is writing in a <8:notebook>, while the <5:man on the right in a buttoned-up work shirt>, looks at him attentively. Beside them, a <7:calf with a black-and-white coat> stands calmly, facing the camera. In the background, a <1:wooden barn or shed> is visible along with the <0:sky> giving the scene a rural, farm-like setting.",
"id": 1319,
"image_id": "000000581422",
"label_matched": [
{
"mask_ids": [
4,
5
],
"txt_desc": "men"
},
{
"mask_ids": [
6
],
"txt_desc": "wooden crates"
},
{
"mask_ids": [
3
],
"txt_desc": "field"
},
{
"mask_ids": [
2
],
"txt_desc": "tree"
},
{
"mask_ids": [
4
],
"txt_desc": "man on the left with a wide-brimmed hat, a shirt, and a tie"
},
{
"mask_ids": [
8
],
"txt_desc": "notebook"
},
{
"mask_ids": [
5
],
"txt_desc": "man on the right in a buttoned-up work shirt"
},
{
"mask_ids": [
7
],
"txt_desc": "calf with a black-and-white coat"
},
{
"mask_ids": [
1
],
"txt_desc": "wooden barn or shed"
},
{
"mask_ids": [
0
],
"txt_desc": "sky"
}
],
"labels": [
"sky-other-merged",
"building-other-merged",
"tree-merged",
"dirt-merged",
"person",
"person",
"bench",
"cow",
"book"
]
} | [
{
"area": 14771,
"bbox": [
0,
0,
582,
95
],
"category_id": 187,
"id": 15043,
"image_id": "000000581422",
"iscrowd": 0,
"segmentation": {
"counts": "0>?Z1^<eNdCX1^<fNeCW1]O_Nn<:eCW1]<kNaCT1`<oN^CP1b<RO[Co0e<ROYCo0g<ROWCn0j<SOUCn0j<ROWCl0j<TOVCl0j<UOVCk0i<... | [
{
"id": 1,
"name": "object"
}
] |
train | 100_9OFxODps2Uo_00000610 | 100_9OFxODps2Uo_00000610.jpg | {
"data_source": "VIPSeg",
"file_name": "100_9OFxODps2Uo_00000610.jpg",
"height": 720,
"id": "100_9OFxODps2Uo_00000610",
"width": 1280
} | {
"caption": "Two people walk on a green and white labyrinth design on a carpet in a room with a concrete floor and grey walls. A brown door on the left and two large glass windows with dark muntins looking out onto trees on the right.",
"caption_ann": "Two <6,7:people> walk on a <2:green and white labyrinth design on a carpet> in a room with a <1:concrete floor> and <0:grey walls>. A <3:brown door> on the left and <4,5:two large glass windows with dark muntins> looking out onto trees on the right.",
"id": 1320,
"image_id": "100_9OFxODps2Uo_00000610",
"label_matched": [
{
"mask_ids": [
6,
7
],
"txt_desc": "people"
},
{
"mask_ids": [
2
],
"txt_desc": "green and white labyrinth design on a carpet"
},
{
"mask_ids": [
1
],
"txt_desc": "concrete floor"
},
{
"mask_ids": [
0
],
"txt_desc": "grey walls"
},
{
"mask_ids": [
3
],
"txt_desc": "brown door"
},
{
"mask_ids": [
4,
5
],
"txt_desc": "two large glass windows with dark muntins"
}
],
"labels": [
"wall",
"floor",
"cushion_or_carpet",
"door",
"window",
"window",
"person",
"person"
]
} | [
{
"area": 190448,
"bbox": [
0,
0,
1280,
379
],
"category_id": 0,
"id": 15052,
"image_id": "100_9OFxODps2Uo_00000610",
"iscrowd": 0,
"segmentation": {
"counts": "0k;e:O10000O1000000000000O1000000O100000000000000O10000O1000000000000O1000000O100000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1015_-KqXcm-I2zY_00000107 | 1015_-KqXcm-I2zY_00000107.jpg | {
"data_source": "VIPSeg",
"file_name": "1015_-KqXcm-I2zY_00000107.jpg",
"height": 720,
"id": "1015_-KqXcm-I2zY_00000107",
"width": 1280
} | {
"caption": "\"A court-level view of a junior tennis match where a girl in a red t-shirt and light blue shorts is holding a tennis racket and in motion to play. The green and red court features a white net across the center. On the top left corner a girl in a dark pink tank top and light blue skirt, a boy in a white t-shirt and shorts, a girl in a white tank top and black skirt, and a girl in light pink tank top and skirt can be seen while watching the match. They stand near a green and white scoreboard and a black door. In the background, a girl in a pink t-shirt is standing in playing stance near a green wall.\"",
"caption_ann": "\"A court-level view of a junior tennis match where a <5:girl in a red t-shirt and light blue shorts> is holding a tennis racket and in motion to play. The <1:green and red court> features a <3:white net> across the center. On the top left corner a <9:girl in a dark pink tank top and light blue skirt>, a <10:boy in a white t-shirt and shorts>, a <7:girl in a white tank top and black skirt>, and a <8:girl in light pink tank top and skirt> can be seen while watching the match. They stand near a <2:green and white scoreboard> and a <4:black door>. In the background, a <6:girl in a pink t-shirt> is standing in playing stance near a <0:green wall>.\"",
"id": 1321,
"image_id": "1015_-KqXcm-I2zY_00000107",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "girl in a red t-shirt and light blue shorts"
},
{
"mask_ids": [
1
],
"txt_desc": "green and red court"
},
{
"mask_ids": [
3
],
"txt_desc": "white net"
},
{
"mask_ids": [
9
],
"txt_desc": "girl in a dark pink tank top and light blue skirt"
},
{
"mask_ids": [
10
],
"txt_desc": "boy in a white t-shirt and shorts"
},
{
"mask_ids": [
7
],
"txt_desc": "girl in a white tank top and black skirt"
},
{
"mask_ids": [
8
],
"txt_desc": "girl in light pink tank top and skirt"
},
{
"mask_ids": [
2
],
"txt_desc": "green and white scoreboard"
},
{
"mask_ids": [
4
],
"txt_desc": "black door"
},
{
"mask_ids": [
6
],
"txt_desc": "girl in a pink t-shirt"
},
{
"mask_ids": [
0
],
"txt_desc": "green wall"
}
],
"labels": [
"wall",
"floor",
"billboard_or_Bulletin_Board",
"ball_net",
"door",
"person",
"person",
"person",
"person",
"person",
"person"
]
} | [
{
"area": 138376,
"bbox": [
0,
0,
1280,
161
],
"category_id": 0,
"id": 15060,
"image_id": "1015_-KqXcm-I2zY_00000107",
"iscrowd": 0,
"segmentation": {
"counts": "3i3Uc0B6J=C7I<D7I=C6J>B6J=C7IWmc11hR\\N7I:F0000001O000T[O[Omc0f0R\\O\\Olc0d0T\\O@hc0`0X\\OBf... | [
{
"id": 1,
"name": "object"
}
] |
train | 1018_MPQxpQfaoV0_00005958 | 1018_MPQxpQfaoV0_00005958.jpg | {
"data_source": "VIPSeg",
"file_name": "1018_MPQxpQfaoV0_00005958.jpg",
"height": 720,
"id": "1018_MPQxpQfaoV0_00005958",
"width": 1280
} | {
"caption": "A room with a tan wall where a mixed martial arts cage can be seen. On the left, a short blonde-haired man in red and black shorts, with white gloves, in a fighting stance, while on the right, a man with a faded haircut and a heavy beard, wearing a black tank top, also in a fighting stance. They stand on a blue mat inside an octagonal cage that has a black fence with a red rail. In the background, on the left, a black kick shield leans against the fence.",
"caption_ann": "A room with a <0:tan wall> where a mixed martial arts cage can be seen. On the left, a <3:short blonde-haired man in red and black shorts, with white gloves>, in a fighting stance, while on the right, a <4:man with a faded haircut and a heavy beard, wearing a black tank top>, also in a fighting stance. They stand on a <2:blue mat> inside an octagonal cage that has a <1:black fence with a red rail>. In the background, on the left, a <5:black kick shield> leans against the <1:fence>.",
"id": 1322,
"image_id": "1018_MPQxpQfaoV0_00005958",
"label_matched": [
{
"mask_ids": [
0
],
"txt_desc": "tan wall"
},
{
"mask_ids": [
3
],
"txt_desc": "short blonde-haired man in red and black shorts, with white gloves"
},
{
"mask_ids": [
4
],
"txt_desc": "man with a faded haircut and a heavy beard, wearing a black tank top"
},
{
"mask_ids": [
2
],
"txt_desc": "blue mat"
},
{
"mask_ids": [
1
],
"txt_desc": "black fence with a red rail"
},
{
"mask_ids": [
5
],
"txt_desc": "black kick shield"
},
{
"mask_ids": [
1
],
"txt_desc": "fence"
}
],
"labels": [
"wall",
"handrail_or_fence",
"floor",
"person",
"person",
"traveling_case_or_trolley_case"
]
} | [
{
"area": 63630,
"bbox": [
0,
0,
1226,
69
],
"category_id": 0,
"id": 15071,
"image_id": "1018_MPQxpQfaoV0_00005958",
"iscrowd": 0,
"segmentation": {
"counts": "0c1md00000000000000000001O0000000000000000001O0000000000000000000000000000000000001O0000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1019_OfA5_-Seg_c_00007923 | 1019_OfA5_-Seg_c_00007923.jpg | {
"data_source": "VIPSeg",
"file_name": "1019_OfA5_-Seg_c_00007923.jpg",
"height": 720,
"id": "1019_OfA5_-Seg_c_00007923",
"width": 1280
} | {
"caption": "In a room two men are practising martial arts on a green mat. A man in a white t-shirt and printed shorts> is on his back, with his legs raised and wrapped around the waist of another man in a black t-shirt and shorts who is punching him. The room has a white wall with a large mirror on the left side and a smaller mirror on the right.",
"caption_ann": "In a room <2,3:two men> are practising martial arts on a <1:green mat>. A <2:man in a white t-shirt and printed shorts>> is on his back, with his legs raised and wrapped around the waist of another <3:man in a black t-shirt and shorts> who is punching him. The room has a <0:white wall> with a <5:large mirror> on the left side and a <4:smaller mirror> on the right.",
"id": 1323,
"image_id": "1019_OfA5_-Seg_c_00007923",
"label_matched": [
{
"mask_ids": [
2,
3
],
"txt_desc": "two men"
},
{
"mask_ids": [
1
],
"txt_desc": "green mat"
},
{
"mask_ids": [
2
],
"txt_desc": "man in a white t-shirt and printed shorts"
},
{
"mask_ids": [
3
],
"txt_desc": "man in a black t-shirt and shorts"
},
{
"mask_ids": [
0
],
"txt_desc": "white wall"
},
{
"mask_ids": [
5
],
"txt_desc": "large mirror"
},
{
"mask_ids": [
4
],
"txt_desc": "smaller mirror"
}
],
"labels": [
"wall",
"floor",
"person",
"person",
"mirror",
"mirror"
]
} | [
{
"area": 211155,
"bbox": [
0,
0,
1280,
446
],
"category_id": 0,
"id": 15077,
"image_id": "1019_OfA5_-Seg_c_00007923",
"iscrowd": 0,
"segmentation": {
"counts": "0c0j9S1c0nN[OS1e0mN[OS1e0mN[OS1e0mN[OS1e0mN[OS1e0mN[OS1e0mN[OS1e0mN[OS1e0mN[OS1e0mN[OS1e0mN[... | [
{
"id": 1,
"name": "object"
}
] |
train | 102_aME6mBuWEAc_00003790 | 102_aME6mBuWEAc_00003790.jpg | {
"data_source": "VIPSeg",
"file_name": "102_aME6mBuWEAc_00003790.jpg",
"height": 720,
"id": "102_aME6mBuWEAc_00003790",
"width": 1280
} | {
"caption": "A man in an orange polo shirt and a woman in a blue button-down shirt stand facing each other and talking in a large grassy field, with a line of trees and a small patch of sky visible in the background.",
"caption_ann": "A <3:man in an orange polo shirt> and a <4:woman in a blue button-down shirt> stand facing each other and talking in a large <0:grassy field>, with a line of <2:trees> and a small patch of <1:sky> visible in the background.",
"id": 1324,
"image_id": "102_aME6mBuWEAc_00003790",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "man in an orange polo shirt"
},
{
"mask_ids": [
4
],
"txt_desc": "woman in a blue button-down shirt"
},
{
"mask_ids": [
0
],
"txt_desc": "grassy field"
},
{
"mask_ids": [
2
],
"txt_desc": "trees"
},
{
"mask_ids": [
1
],
"txt_desc": "sky"
}
],
"labels": [
"grass",
"sky",
"tree",
"person",
"person"
]
} | [
{
"area": 492799,
"bbox": [
0,
156,
1280,
564
],
"category_id": 15,
"id": 15083,
"image_id": "102_aME6mBuWEAc_00003790",
"iscrowd": 0,
"segmentation": {
"counts": "f5j`0g5O0000000000000000000000000000000000000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1030_J_m6vc8gtvQ_00008373 | 1030_J_m6vc8gtvQ_00008373.jpg | {
"data_source": "VIPSeg",
"file_name": "1030_J_m6vc8gtvQ_00008373.jpg",
"height": 720,
"id": "1030_J_m6vc8gtvQ_00008373",
"width": 1280
} | {
"caption": "In a department store a toddler in a black outfit is looking into a black miniature shopping basket that contains various food items while standing on the orange floor. In the background, there is a large storage_rack, filled with different products.",
"caption_ann": "In a department store a <2:toddler in a black outfit> is looking into a <3:black miniature shopping basket that contains various food items> while standing on the <0:orange floor>. In the background, there is a <1:large storage_rack>, filled with different products.",
"id": 1325,
"image_id": "1030_J_m6vc8gtvQ_00008373",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "toddler in a black outfit"
},
{
"mask_ids": [
3
],
"txt_desc": "black miniature shopping basket that contains various food items"
},
{
"mask_ids": [
0
],
"txt_desc": "orange floor"
},
{
"mask_ids": [
1
],
"txt_desc": "large storage_rack"
}
],
"labels": [
"floor",
"cupboard_or_showcase_or_storage_rack",
"person",
"basket"
]
} | [
{
"area": 232488,
"bbox": [
0,
0,
1280,
720
],
"category_id": 13,
"id": 15088,
"image_id": "1030_J_m6vc8gtvQ_00008373",
"iscrowd": 0,
"segmentation": {
"counts": "0bVT24j_lM6J6J7I6J7I6J6J7I6J7I6J6J7I6J7I6J6J7I6J7I6J6J7I6J7I6J6J7I6J7I6J6J7I6J7I6J6J7I6J7I6... | [
{
"id": 1,
"name": "object"
}
] |
train | 1033_79Whs_1_nWs_00000452 | 1033_79Whs_1_nWs_00000452.jpg | {
"data_source": "VIPSeg",
"file_name": "1033_79Whs_1_nWs_00000452.jpg",
"height": 720,
"id": "1033_79Whs_1_nWs_00000452",
"width": 1280
} | {
"caption": "A woman wearing a green and black shirt sits on the carpeted floor next to a white window bench that has a yellow patterned cushion on it. The bench is positioned in front of a large window that overlooks a wooded area, and to the left is a light-colored wall with an electrical outlet.",
"caption_ann": "A <3:woman wearing a green and black shirt> sits on the <1:carpeted floor> next to a <4:white window bench that has a yellow patterned cushion on it>. The bench is positioned in front of a large <2:window> that overlooks a wooded area, and to the left is a <0:light-colored wall> with an electrical outlet.",
"id": 1326,
"image_id": "1033_79Whs_1_nWs_00000452",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "woman wearing a green and black shirt"
},
{
"mask_ids": [
1
],
"txt_desc": "carpeted floor"
},
{
"mask_ids": [
4
],
"txt_desc": "white window bench that has a yellow patterned cushion on it"
},
{
"mask_ids": [
2
],
"txt_desc": "window"
},
{
"mask_ids": [
0
],
"txt_desc": "light-colored wall"
}
],
"labels": [
"wall",
"floor",
"window",
"person",
"table_or_desk"
]
} | [
{
"area": 221573,
"bbox": [
0,
0,
1280,
693
],
"category_id": 0,
"id": 15092,
"image_id": "1033_79Whs_1_nWs_00000452",
"iscrowd": 0,
"segmentation": {
"counts": "0ee0k00000O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1000000O10000O10000O1... | [
{
"id": 1,
"name": "object"
}
] |
train | 1033_sh81AwYuihg_00003348 | 1033_sh81AwYuihg_00003348.jpg | {
"data_source": "VIPSeg",
"file_name": "1033_sh81AwYuihg_00003348.jpg",
"height": 720,
"id": "1033_sh81AwYuihg_00003348",
"width": 1280
} | {
"caption": "In a play area with a purple floor and light-colored walls, a young girl in a denim jacket interacts with a large red machine while another girl in a white t-shirt uses a wall-mounted tool. To the left, a person in a grey shirt is partially visible walking past.",
"caption_ann": "In a play area with a <1:purple floor> and <0:light-colored walls>, a <5:young girl in a denim jacket> interacts with a <2:large red machine> while another <4:girl in a white t-shirt> uses a <3:wall-mounted tool>. To the left, a <6:person in a grey shirt> is partially visible walking past.",
"id": 1327,
"image_id": "1033_sh81AwYuihg_00003348",
"label_matched": [
{
"mask_ids": [
1
],
"txt_desc": "purple floor"
},
{
"mask_ids": [
0
],
"txt_desc": "light-colored walls"
},
{
"mask_ids": [
5
],
"txt_desc": "young girl in a denim jacket"
},
{
"mask_ids": [
2
],
"txt_desc": "large red machine"
},
{
"mask_ids": [
4
],
"txt_desc": "girl in a white t-shirt"
},
{
"mask_ids": [
3
],
"txt_desc": "wall-mounted tool"
},
{
"mask_ids": [
6
],
"txt_desc": "person in a grey shirt"
}
],
"labels": [
"wall",
"floor",
"other_machine",
"tool",
"person",
"person",
"person"
]
} | [
{
"area": 216444,
"bbox": [
39,
0,
1241,
579
],
"category_id": 0,
"id": 15097,
"image_id": "1033_sh81AwYuihg_00003348",
"iscrowd": 0,
"segmentation": {
"counts": "`]k01_f000000000000000000000000000000000000000000000000XZO0Pe00oZO1Qe0OnZO3Qe0MnZO4Re0LnZO5... | [
{
"id": 1,
"name": "object"
}
] |
train | 1034_-JxWPcDaJM0_00002268 | 1034_-JxWPcDaJM0_00002268.jpg | {
"data_source": "VIPSeg",
"file_name": "1034_-JxWPcDaJM0_00002268.jpg",
"height": 720,
"id": "1034_-JxWPcDaJM0_00002268",
"width": 1280
} | {
"caption": "A ghost-like sculpture draped in a white cloth stands with a vintage-style bicycle on a wooden platform. The platform is set on stony ground with a patch of grass behind it. In the background, a short, wooden fence runs in front of distant rocky mountains under a clear blue sky.",
"caption_ann": "A <5:ghost-like sculpture> draped in a white cloth stands with a <6:vintage-style bicycle> on a wooden platform. The platform is set on <1:stony ground> with a patch of <2:grass> behind it. In the background, a <0:short, wooden fence> runs in front of distant <4:rocky mountains> under a <3:clear blue sky>.",
"id": 1328,
"image_id": "1034_-JxWPcDaJM0_00002268",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "ghost-like sculpture"
},
{
"mask_ids": [
6
],
"txt_desc": "vintage-style bicycle"
},
{
"mask_ids": [
1
],
"txt_desc": "stony ground"
},
{
"mask_ids": [
2
],
"txt_desc": "grass"
},
{
"mask_ids": [
0
],
"txt_desc": "short, wooden fence"
},
{
"mask_ids": [
4
],
"txt_desc": "rocky mountains"
},
{
"mask_ids": [
3
],
"txt_desc": "clear blue sky"
}
],
"labels": [
"handrail_or_fence",
"ground",
"grass",
"sky",
"mountain",
"sculpture",
"bicycle"
]
} | [
{
"area": 122225,
"bbox": [
643,
208,
637,
319
],
"category_id": 7,
"id": 15104,
"image_id": "1034_-JxWPcDaJM0_00002268",
"iscrowd": 0,
"segmentation": {
"counts": "o\\T>1Rf0>B>B?G9N2N2N2N2N3M2N2N2N2N3M2N2N2N2N2M4M2N1O1O00001O00001O00001O000O2O00001O0000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1035_P1c2o2aupEc_00000152 | 1035_P1c2o2aupEc_00000152.jpg | {
"data_source": "VIPSeg",
"file_name": "1035_P1c2o2aupEc_00000152.jpg",
"height": 720,
"id": "1035_P1c2o2aupEc_00000152",
"width": 1280
} | {
"caption": "In a yoga studio with a wood-plank floor and purple walls, a large class of women in colourful workout attire practice various advanced poses on their yoga mats. The scene is reflected in a large mirror on the left wall, above which a wall-mounted display is visible. The room is furnished with shelves, brown curtains, and a glass door that has a sign on it. Personal items including clothes, bags, and mobile phones are visible on the floor around the participants.",
"caption_ann": "In a yoga studio with a <1:wood-plank floor> and <0:purple walls>, a large class of <7,8,9,10,11,12,13,14,15,16:women in colourful workout attire> practice various advanced poses on their <2:yoga mats>. The scene is reflected in a large <19:mirror> on the left wall, above which a <20:wall-mounted display> is visible. The room is furnished with <3:shelves>, <4:brown curtains>, and a <6:glass door> that has a sign on it. Personal items including <5:clothes>, <17,18:bags>, and <21,22:mobile phones> are visible on the floor around the participants.",
"id": 1329,
"image_id": "1035_P1c2o2aupEc_00000152",
"label_matched": [
{
"mask_ids": [
1
],
"txt_desc": "wood-plank floor"
},
{
"mask_ids": [
0
],
"txt_desc": "purple walls"
},
{
"mask_ids": [
7,
8,
9,
10,
11,
12,
13,
14,
15,
16
],
"txt_desc": "women in colourful workout attire"
},
{
"mask_ids": [
2
],
"txt_desc": "yoga mats"
},
{
"mask_ids": [
19
],
"txt_desc": "mirror"
},
{
"mask_ids": [
20
],
"txt_desc": "wall-mounted display"
},
{
"mask_ids": [
3
],
"txt_desc": "shelves"
},
{
"mask_ids": [
4
],
"txt_desc": "brown curtains"
},
{
"mask_ids": [
6
],
"txt_desc": "glass door"
},
{
"mask_ids": [
5
],
"txt_desc": "clothes"
},
{
"mask_ids": [
17,
18
],
"txt_desc": "bags"
},
{
"mask_ids": [
21,
22
],
"txt_desc": "mobile phones"
}
],
"labels": [
"wall",
"floor",
"cushion_or_carpet",
"shelf",
"curtain",
"clothes",
"door",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"bag_or_package",
"bag_or_package",
"mirror",
"clock",
"Mobile_phone",
"Mobile_phone"
]
} | [
{
"area": 120819,
"bbox": [
0,
0,
1279,
720
],
"category_id": 0,
"id": 15111,
"image_id": "1035_P1c2o2aupEc_00000152",
"iscrowd": 0,
"segmentation": {
"counts": "0\\1Te01N100000000000000000000000000000000000000000fNU[Oo0kd0QOZ[Oj0fd0VOd[O`0\\d0@h[O<Xd0DR... | [
{
"id": 1,
"name": "object"
}
] |
train | 1036_RsM6XEE7P3g_00000152 | 1036_RsM6XEE7P3g_00000152.jpg | {
"data_source": "VIPSeg",
"file_name": "1036_RsM6XEE7P3g_00000152.jpg",
"height": 720,
"id": "1036_RsM6XEE7P3g_00000152",
"width": 1280
} | {
"caption": "In a yoga studio with a wooden floor and light blue walls, a class is in session on yellow yoga mats. A male instructor kneels to assist a woman in a blue outfit with a stretch against the wall. They are surrounded by other students in various yoga attire who are either watching or practicing poses. In the background, a man observes near a cabinet. A green curtain hangs on the left wall, and in the foreground, personal items including several bags, an mobile phone, and clothes are on the floor.",
"caption_ann": "In a yoga studio with a <1:wooden floor> and <0:light blue walls>, a class is in session on <2:yellow yoga mats>. A <7:male instructor> kneels to assist a <6:woman in a blue outfit> with a stretch against the wall. They are surrounded by other <8,9,10,11,12,13:students in various yoga attire> who are either watching or practicing poses. In the background, a <14:man> observes near a <3:cabinet>. A <4:green curtain> hangs on the left wall, and in the foreground, personal items including <15,16,17:several bags>, an <18:mobile phone>, and <5:clothes> are on the floor.",
"id": 1330,
"image_id": "1036_RsM6XEE7P3g_00000152",
"label_matched": [
{
"mask_ids": [
1
],
"txt_desc": "wooden floor"
},
{
"mask_ids": [
0
],
"txt_desc": "light blue walls"
},
{
"mask_ids": [
2
],
"txt_desc": "yellow yoga mats"
},
{
"mask_ids": [
7
],
"txt_desc": "male instructor"
},
{
"mask_ids": [
6
],
"txt_desc": "woman in a blue outfit"
},
{
"mask_ids": [
8,
9,
10,
11,
12,
13
],
"txt_desc": "students in various yoga attire"
},
{
"mask_ids": [
14
],
"txt_desc": "man"
},
{
"mask_ids": [
3
],
"txt_desc": "cabinet"
},
{
"mask_ids": [
4
],
"txt_desc": "green curtain"
},
{
"mask_ids": [
15,
16,
17
],
"txt_desc": "several bags"
},
{
"mask_ids": [
18
],
"txt_desc": "mobile phone"
},
{
"mask_ids": [
5
],
"txt_desc": "clothes"
}
],
"labels": [
"wall",
"floor",
"cushion_or_carpet",
"cupboard_or_showcase_or_storage_rack",
"curtain",
"clothes",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"bag_or_package",
"bag_or_package",
"bag_or_package",
"Mobile_phone"
]
} | [
{
"area": 318718,
"bbox": [
1,
0,
1279,
720
],
"category_id": 0,
"id": 15134,
"image_id": "1036_RsM6XEE7P3g_00000152",
"iscrowd": 0,
"segmentation": {
"counts": "om0Q?_702N1O1O2N1O1O2N002SNl1[Lf3fHZUZ1lNRSfNl2TM]1cNV2jM3M6J4L6J3M6J3M6J4L6J3M6J3M6J4L6J3M6... | [
{
"id": 1,
"name": "object"
}
] |
train | 103_5ZujNM_MouM_00000158 | 103_5ZujNM_MouM_00000158.jpg | {
"data_source": "VIPSeg",
"file_name": "103_5ZujNM_MouM_00000158.jpg",
"height": 720,
"id": "103_5ZujNM_MouM_00000158",
"width": 1280
} | {
"caption": "Two young boxers spar in a ring with white roped barriers over a light gray floor. One boy wears a blue headguard and boxing gloves, while the other child wears a red headguard. Standing over the children is a man in a light green shirt and dark trousers. In the background, behind the ropes and a brown wall, several people are visible, including a person in a red jacket, and three spectators standing near a dark wooden chair.",
"caption_ann": "Two young boxers spar in a ring with white <0:roped barriers> over a <1:light gray floor>. One <6:boy wears a blue headguard and boxing gloves>, while the other <7:child wears a red headguard>. Standing over the children is a <8:man in a light green shirt and dark trousers>. In the background, behind the ropes and a <2:brown wall>, several people are visible, including a <3:person in a red jacket>, and <4,5,9:three spectators> standing near a <10,11:dark wooden chair>.",
"id": 1331,
"image_id": "103_5ZujNM_MouM_00000158",
"label_matched": [
{
"mask_ids": [
0
],
"txt_desc": "roped barriers"
},
{
"mask_ids": [
1
],
"txt_desc": "light gray floor"
},
{
"mask_ids": [
6
],
"txt_desc": "boy wears a blue headguard and boxing gloves"
},
{
"mask_ids": [
7
],
"txt_desc": "child wears a red headguard"
},
{
"mask_ids": [
8
],
"txt_desc": "man in a light green shirt and dark trousers"
},
{
"mask_ids": [
2
],
"txt_desc": "brown wall"
},
{
"mask_ids": [
3
],
"txt_desc": "person in a red jacket"
},
{
"mask_ids": [
4,
5,
9
],
"txt_desc": "three spectators"
},
{
"mask_ids": [
10,
11
],
"txt_desc": "dark wooden chair"
}
],
"labels": [
"handrail_or_fence",
"floor",
"grandstand",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"chair_or_seat",
"chair_or_seat"
]
} | [
{
"area": 117726,
"bbox": [
0,
0,
1280,
628
],
"category_id": 7,
"id": 15153,
"image_id": "103_5ZujNM_MouM_00000158",
"iscrowd": 0,
"segmentation": {
"counts": "0_8j4UHIOjNl7]1UHIOjNk7^1VHHOjNk7^1VHINiNl7^1VHINiNk7_1WHHNiNl7^1VHINiNl7]1WHJMiNl7]1WHJLjNm7... | [
{
"id": 1,
"name": "object"
}
] |
train | 1050_6EN3gJoc4-U_00002763 | 1050_6EN3gJoc4-U_00002763.jpg | {
"data_source": "VIPSeg",
"file_name": "1050_6EN3gJoc4-U_00002763.jpg",
"height": 720,
"id": "1050_6EN3gJoc4-U_00002763",
"width": 1280
} | {
"caption": "Three men are gathered in an office with pink walls and a white ceiling. A man in a striped shirt is seated in a black chair on the left, looking up at the other two men wearing ties. In the center, a man in a dark blue jacket and light pants with a green tie stands next to a large black television screen. The television and another electronic grey item rest on a dark wood cabinet, which appears to be a separate piece from the main desk in the foreground. A third man in a dark gray suit stands on the right, also looking at the man seated. On the left, a tall potted plant sits on the floor next to a dark pot, and two circular lamps are embedded in the ceiling.",
"caption_ann": "<5,6,7:Three men> are gathered in an office with pink <0:walls> and a white <1:ceiling>. A <7:man in a striped shirt> is seated in a <10:black chair> on the left, looking up at the other <5,6:two men wearing ties>. In the center, a <5:man in a dark blue jacket and light pants> with a green tie stands next to a <12:large black television screen>. The <12:television> and another <4:electronic grey item> rest on a dark wood <9:cabinet>, which appears to be a separate piece from the <8:main desk> in the foreground. A <6:third man in a dark gray suit> stands on the right, also looking at the <5:man seated>. On the left, a tall <3:potted plant> sits on the floor next to a <11:dark pot>, and two circular <2:lamps> are embedded in the <1:ceiling>.",
"id": 1332,
"image_id": "1050_6EN3gJoc4-U_00002763",
"label_matched": [
{
"mask_ids": [
5,
6,
7
],
"txt_desc": "Three men"
},
{
"mask_ids": [
0
],
"txt_desc": "walls"
},
{
"mask_ids": [
1
],
"txt_desc": "ceiling"
},
{
"mask_ids": [
7
],
"txt_desc": "man in a striped shirt"
},
{
"mask_ids": [
10
],
"txt_desc": "black chair"
},
{
"mask_ids": [
5,
6
],
"txt_desc": "two men wearing ties"
},
{
"mask_ids": [
5
],
"txt_desc": "man in a dark blue jacket and light pants"
},
{
"mask_ids": [
12
],
"txt_desc": "large black television screen"
},
{
"mask_ids": [
12
],
"txt_desc": "television"
},
{
"mask_ids": [
4
],
"txt_desc": "electronic grey item"
},
{
"mask_ids": [
9
],
"txt_desc": "cabinet"
},
{
"mask_ids": [
8
],
"txt_desc": "main desk"
},
{
"mask_ids": [
6
],
"txt_desc": "third man in a dark gray suit"
},
{
"mask_ids": [
5
],
"txt_desc": "man seated"
},
{
"mask_ids": [
3
],
"txt_desc": "potted plant"
},
{
"mask_ids": [
11
],
"txt_desc": "dark pot"
},
{
"mask_ids": [
2
],
"txt_desc": "lamps"
},
{
"mask_ids": [
1
],
"txt_desc": "ceiling"
}
],
"labels": [
"wall",
"ceiling",
"lamp",
"other_plant",
"other_electronic_product",
"person",
"person",
"person",
"table_or_desk",
"table_or_desk",
"chair_or_seat",
"flower_pot_or_vase",
"screen_or_television"
]
} | [
{
"area": 257768,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 15165,
"image_id": "1050_6EN3gJoc4-U_00002763",
"iscrowd": 0,
"segmentation": {
"counts": "a0V2Zd02N002N002N002OO02N002OO02N002N001O100O1O1O100O1O1O1O1O1O1O100O1O1O100O1O1O1O1O1O1O100... | [
{
"id": 1,
"name": "object"
}
] |
train | 1053_MLWQ-qvra68_00000152 | 1053_MLWQ-qvra68_00000152.jpg | {
"data_source": "VIPSeg",
"file_name": "1053_MLWQ-qvra68_00000152.jpg",
"height": 720,
"id": "1053_MLWQ-qvra68_00000152",
"width": 1280
} | {
"caption": "Two dogs are captivated by a television screen in a dimly lit room with a light green and pink wall on the left side of the frame. The dogs are watching a television screen that displays an abstract blue image with a white circle in the middle. The television screen is mounted on a brown wooden cabinet. A blue blanket is in the bottom right corner of the frame. A piece of equipment is on the left side, close to the wall. The floor appears to be a light brown color.",
"caption_ann": "Two <5,6:dogs> are captivated by a television screen in a dimly lit room with a light green and pink <0:wall> on the left side of the frame. The <5,6:dogs> are watching a <7:television screen> that displays an abstract blue image with a white circle in the middle. The <7:television screen> is mounted on a brown <2:wooden cabinet>. A <4:blue blanket> is in the bottom right corner of the frame. A <3:piece of equipment> is on the left side, close to the <0:wall>. The <1:floor> appears to be a light brown color.",
"id": 1333,
"image_id": "1053_MLWQ-qvra68_00000152",
"label_matched": [
{
"mask_ids": [
5,
6
],
"txt_desc": "dogs"
},
{
"mask_ids": [
0
],
"txt_desc": "wall"
},
{
"mask_ids": [
5,
6
],
"txt_desc": "dogs"
},
{
"mask_ids": [
7
],
"txt_desc": "television screen"
},
{
"mask_ids": [
7
],
"txt_desc": "television screen"
},
{
"mask_ids": [
2
],
"txt_desc": "wooden cabinet"
},
{
"mask_ids": [
4
],
"txt_desc": "blue blanket"
},
{
"mask_ids": [
3
],
"txt_desc": "piece of equipment"
},
{
"mask_ids": [
0
],
"txt_desc": "wall"
},
{
"mask_ids": [
1
],
"txt_desc": "floor"
}
],
"labels": [
"wall",
"floor",
"cupboard_or_showcase_or_storage_rack",
"other_machine",
"textiles",
"dog",
"dog",
"screen_or_television"
]
} | [
{
"area": 38921,
"bbox": [
0,
0,
127,
534
],
"category_id": 0,
"id": 15178,
"image_id": "1053_MLWQ-qvra68_00000152",
"iscrowd": 0,
"segmentation": {
"counts": "[5[;T;0010000O10O0100O010O10000O001000O0100O1000O01O10O10O100]NeDPH\\;o7iDmGV;T8QEdGP;[8UEaGk:... | [
{
"id": 1,
"name": "object"
}
] |
train | 1076_iF7wrp19ssY_00000423 | 1076_iF7wrp19ssY_00000423.jpg | {
"data_source": "VIPSeg",
"file_name": "1076_iF7wrp19ssY_00000423.jpg",
"height": 720,
"id": "1076_iF7wrp19ssY_00000423",
"width": 1280
} | {
"caption": "A classic white Mercedes-Benz SL convertible is driving down a long, gray road with a white painted line. In the background on the left is a long gray building with multiple garage doors. The building has a long, elevated concrete barrier in front of it. In the distance, a tall, brown pole is seen on the left side of the road, and another thin pole stands on the right side of the road. The sky is overcast and gray.",
"caption_ann": "A classic <4:white Mercedes-Benz SL convertible> is driving down a long, gray <2:road> with a white painted line. In the background on the left is a long <0:gray building> with multiple garage doors. The building has a long, elevated concrete barrier in front of it. In the distance, a tall, brown <1:pole> is seen on the left side of the <2:road>, and another thin pole stands on the right side of the <2:road>. The <3:sky> is overcast and gray.",
"id": 1334,
"image_id": "1076_iF7wrp19ssY_00000423",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "white Mercedes-Benz SL convertible"
},
{
"mask_ids": [
2
],
"txt_desc": "road"
},
{
"mask_ids": [
0
],
"txt_desc": "gray building"
},
{
"mask_ids": [
1
],
"txt_desc": "pole"
},
{
"mask_ids": [
2
],
"txt_desc": "road"
},
{
"mask_ids": [
2
],
"txt_desc": "road"
},
{
"mask_ids": [
3
],
"txt_desc": "sky"
}
],
"labels": [
"wall",
"pole",
"road",
"sky",
"car"
]
} | [
{
"area": 260582,
"bbox": [
0,
0,
1280,
310
],
"category_id": 0,
"id": 15186,
"image_id": "1076_iF7wrp19ssY_00000423",
"iscrowd": 0,
"segmentation": {
"counts": "0f9j<00000000000000000000000000000000O100000000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1079_5G7dnQAl-GI_00000802 | 1079_5G7dnQAl-GI_00000802.jpg | {
"data_source": "VIPSeg",
"file_name": "1079_5G7dnQAl-GI_00000802.jpg",
"height": 720,
"id": "1079_5G7dnQAl-GI_00000802",
"width": 1280
} | {
"caption": "A man wearing a blue patterned t-shirt is squatting down on the dirt ground to fish in a body of water that has a long fishing pole in it. The man is next to a young boy wearing a pink shirt who is also sitting on the dirt ground. Next to the young boy is a larger red bucket and a small red container. The body of water is next to a patch of green grass.",
"caption_ann": "A <5:man wearing a blue patterned t-shirt> is squatting down on the <1:dirt ground> to fish in a <3:body of water> that has a long <0:fishing pole> in it. The <5:man> is next to a <4:young boy wearing a pink shirt> who is also sitting on the <1:dirt ground>. Next to the <4:young boy> is a larger <6:red bucket> and a small <7:red container>. The <3:body of water> is next to a patch of <2:green grass>.",
"id": 1335,
"image_id": "1079_5G7dnQAl-GI_00000802",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "man wearing a blue patterned t-shirt"
},
{
"mask_ids": [
1
],
"txt_desc": "dirt ground"
},
{
"mask_ids": [
3
],
"txt_desc": "body of water"
},
{
"mask_ids": [
0
],
"txt_desc": "fishing pole"
},
{
"mask_ids": [
5
],
"txt_desc": "man"
},
{
"mask_ids": [
4
],
"txt_desc": "young boy wearing a pink shirt"
},
{
"mask_ids": [
1
],
"txt_desc": "dirt ground"
},
{
"mask_ids": [
4
],
"txt_desc": "young boy"
},
{
"mask_ids": [
6
],
"txt_desc": "red bucket"
},
{
"mask_ids": [
7
],
"txt_desc": "red container"
},
{
"mask_ids": [
3
],
"txt_desc": "body of water"
},
{
"mask_ids": [
2
],
"txt_desc": "green grass"
}
],
"labels": [
"pole",
"ground",
"grass",
"lake",
"person",
"person",
"tub_or_bowl_or_pot",
"barrel"
]
} | [
{
"area": 4571,
"bbox": [
637,
410,
638,
239
],
"category_id": 12,
"id": 15191,
"image_id": "1079_5G7dnQAl-GI_00000802",
"iscrowd": 0,
"segmentation": {
"counts": "^YP>3\\f03N0O1O2O0000001O000000001O001N100100O100O1O010O01O0001O01O0010O0010O010O01O0010O0... | [
{
"id": 1,
"name": "object"
}
] |
train | 1084_YGqNFSzNW1Y_00005133 | 1084_YGqNFSzNW1Y_00005133.jpg | {
"data_source": "VIPSeg",
"file_name": "1084_YGqNFSzNW1Y_00005133.jpg",
"height": 720,
"id": "1084_YGqNFSzNW1Y_00005133",
"width": 1280
} | {
"caption": "A person is crouching down in a field on a sunny day. A person wearing a red and black shirt and a woven hat is kneeling in the mud, with their hands in a small ditch in the brown dirt ground. To the left of the person there is a blue plastic basket on the ground. In the background, there is a tree with lush green leaves and a vast field of green grass. The sky is clear and blue.",
"caption_ann": "A person is crouching down in a field on a sunny day. A <4:person wearing a red and black shirt and a woven hat> is kneeling in the mud, with their hands in a small <0:ditch in the brown dirt ground>. To the left of the <4:person> there is a <5:blue plastic basket> on the <0:ground>. In the background, there is a <3:tree> with lush green leaves and a vast field of <1:green grass>. The <2:sky> is clear and blue.",
"id": 1336,
"image_id": "1084_YGqNFSzNW1Y_00005133",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "person wearing a red and black shirt and a woven hat"
},
{
"mask_ids": [
0
],
"txt_desc": "ditch in the brown dirt ground"
},
{
"mask_ids": [
4
],
"txt_desc": "person"
},
{
"mask_ids": [
5
],
"txt_desc": "blue plastic basket"
},
{
"mask_ids": [
0
],
"txt_desc": "ground"
},
{
"mask_ids": [
3
],
"txt_desc": "tree"
},
{
"mask_ids": [
1
],
"txt_desc": "green grass"
},
{
"mask_ids": [
2
],
"txt_desc": "sky"
}
],
"labels": [
"ground",
"grass",
"sky",
"tree",
"person",
"basket"
]
} | [
{
"area": 493530,
"bbox": [
0,
263,
1280,
457
],
"category_id": 14,
"id": 15199,
"image_id": "1084_YGqNFSzNW1Y_00005133",
"iscrowd": 0,
"segmentation": {
"counts": "l9d<k910000000000000000000000000000000000000000000000O10000000000000000000000O10000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1086_me8xX6AivQ0_00000498 | 1086_me8xX6AivQ0_00000498.jpg | {
"data_source": "VIPSeg",
"file_name": "1086_me8xX6AivQ0_00000498.jpg",
"height": 720,
"id": "1086_me8xX6AivQ0_00000498",
"width": 1280
} | {
"caption": "Three people are shown working in a muddy stream with some tall green grass on the banks. The person in a yellow and red shirt on the left is bent over the stream and seems to be looking into it. Next to the person in a yellow and red shirt is a blue plastic basket that is partially submerged in the stream. Behind him is another person in a blue shirt bent over the stream. In the foreground, the back of a third person in a black shirt is visible as they also bend over the stream. All three people appear to be focused on the water.",
"caption_ann": "Three <2,3,4:people> are shown working in a muddy <1:stream> with some <0:tall green grass> on the banks. The <3:person in a yellow and red shirt> on the left is bent over the <1:stream> and seems to be looking into it. Next to the <3:person in a yellow and red shirt> is a <5:blue plastic basket> that is partially submerged in the <1:stream>. Behind him is another <2:person in a blue shirt> bent over the <1:stream>. In the foreground, the back of a <4:third person in a black shirt> is visible as they also bend over the <1:stream>. All three <2,3,4:people> appear to be focused on the water.",
"id": 1337,
"image_id": "1086_me8xX6AivQ0_00000498",
"label_matched": [
{
"mask_ids": [
2,
3,
4
],
"txt_desc": "people"
},
{
"mask_ids": [
1
],
"txt_desc": "stream"
},
{
"mask_ids": [
0
],
"txt_desc": "tall green grass"
},
{
"mask_ids": [
3
],
"txt_desc": "person in a yellow and red shirt"
},
{
"mask_ids": [
1
],
"txt_desc": "stream"
},
{
"mask_ids": [
3
],
"txt_desc": "person in a yellow and red shirt"
},
{
"mask_ids": [
5
],
"txt_desc": "blue plastic basket"
},
{
"mask_ids": [
1
],
"txt_desc": "stream"
},
{
"mask_ids": [
2
],
"txt_desc": "person in a blue shirt"
},
{
"mask_ids": [
1
],
"txt_desc": "stream"
},
{
"mask_ids": [
4
],
"txt_desc": "third person in a black shirt"
},
{
"mask_ids": [
1
],
"txt_desc": "stream"
},
{
"mask_ids": [
2,
3,
4
],
"txt_desc": "people"
}
],
"labels": [
"grass",
"water",
"person",
"person",
"person",
"basket"
]
} | [
{
"area": 636205,
"bbox": [
0,
0,
1280,
720
],
"category_id": 15,
"id": 15205,
"image_id": "1086_me8xX6AivQ0_00000498",
"iscrowd": 0,
"segmentation": {
"counts": "0[7U?00001O00001O000000001O001O00000000001O001O000000001O00001O000000001O00001O0000001O0000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1090_B2m0ExqAGCo_00002493 | 1090_B2m0ExqAGCo_00002493.jpg | {
"data_source": "VIPSeg",
"file_name": "1090_B2m0ExqAGCo_00002493.jpg",
"height": 720,
"id": "1090_B2m0ExqAGCo_00002493",
"width": 1280
} | {
"caption": "A person is walking away from the camera in a concrete parking garage with a low ceiling and gray walls. The person is wearing a dark jacket with a white 'X' on the back, and is facing a white and yellow security car that has \"SECURITE\" written on its side. There is a person standing beside the door of the white and yellow security car The white and yellow security car is parked next to a white sedan and a white sedan is parked on the left side of the frame, next to an coffee-colored car with a brown front bumper. Several lamps are mounted on the ceiling.",
"caption_ann": "A <9:person> is walking away from the camera in a <2:concrete parking garage> with a <1:low ceiling> and gray <0:walls>. The <9:person> is wearing a dark jacket with a white 'X' on the back, and is facing a <6:white and yellow security car> that has \"SECURITE\" written on its side. There is a <10:person> standing beside the <4:door> of the <6:white and yellow security car> The <6:white and yellow security car> is parked next to a <7:white sedan> and a <5:white sedan> is parked on the left side of the frame, next to an <8:coffee-colored car> with a brown front bumper. Several <3:lamps> are mounted on the <1:ceiling>.",
"id": 1338,
"image_id": "1090_B2m0ExqAGCo_00002493",
"label_matched": [
{
"mask_ids": [
9
],
"txt_desc": "person"
},
{
"mask_ids": [
2
],
"txt_desc": "concrete parking garage"
},
{
"mask_ids": [
1
],
"txt_desc": "low ceiling"
},
{
"mask_ids": [
0
],
"txt_desc": "walls"
},
{
"mask_ids": [
9
],
"txt_desc": "person"
},
{
"mask_ids": [
6
],
"txt_desc": "white and yellow security car"
},
{
"mask_ids": [
10
],
"txt_desc": "person"
},
{
"mask_ids": [
4
],
"txt_desc": "door"
},
{
"mask_ids": [
6
],
"txt_desc": "white and yellow security car"
},
{
"mask_ids": [
6
],
"txt_desc": "white and yellow security car"
},
{
"mask_ids": [
7
],
"txt_desc": "white sedan"
},
{
"mask_ids": [
5
],
"txt_desc": "white sedan"
},
{
"mask_ids": [
8
],
"txt_desc": "coffee-colored car"
},
{
"mask_ids": [
3
],
"txt_desc": "lamps"
},
{
"mask_ids": [
1
],
"txt_desc": "ceiling"
}
],
"labels": [
"wall",
"ceiling",
"floor",
"lamp",
"door",
"car",
"car",
"car",
"car",
"person",
"person"
]
} | [
{
"area": 108650,
"bbox": [
0,
194,
1055,
271
],
"category_id": 0,
"id": 15211,
"image_id": "1090_B2m0ExqAGCo_00002493",
"iscrowd": 0,
"segmentation": {
"counts": "R6h4ha01O000000001O000000001O000000001O00000a0_Oc0^OO0000001O0000000010O00000001O000000001... | [
{
"id": 1,
"name": "object"
}
] |
train | 1096_sTtQpCDAZaQ_00000017 | 1096_sTtQpCDAZaQ_00000017.jpg | {
"data_source": "VIPSeg",
"file_name": "1096_sTtQpCDAZaQ_00000017.jpg",
"height": 720,
"id": "1096_sTtQpCDAZaQ_00000017",
"width": 1280
} | {
"caption": "Two large green combine harvesters are working in a large wheat field with trees visible in the background. The combine harvesters are positioned side by side and are moving through the wheat field with their large cutting heads extended, kicking up dust and chaff. The trees form a dense line at the edge of the wheat field, beneath a hazy, light-colored sky.",
"caption_ann": "Two <1:large green combine harvesters> are working in a large <3:wheat field> with <2:trees> visible in the background. The <1:combine harvesters> are positioned side by side and are moving through the <3:wheat field> with their large cutting heads extended, kicking up dust and chaff. The <2:trees> form a dense line at the edge of the <3:wheat field>, beneath a hazy, light-colored <0:sky>.",
"id": 1339,
"image_id": "1096_sTtQpCDAZaQ_00000017",
"label_matched": [
{
"mask_ids": [
1
],
"txt_desc": "large green combine harvesters"
},
{
"mask_ids": [
3
],
"txt_desc": "wheat field"
},
{
"mask_ids": [
2
],
"txt_desc": "trees"
},
{
"mask_ids": [
1
],
"txt_desc": "combine harvesters"
},
{
"mask_ids": [
3
],
"txt_desc": "wheat field"
},
{
"mask_ids": [
2
],
"txt_desc": "trees"
},
{
"mask_ids": [
3
],
"txt_desc": "wheat field"
},
{
"mask_ids": [
0
],
"txt_desc": "sky"
}
],
"labels": [
"sky",
"wheeled_machine",
"tree",
"other_plant"
]
} | [
{
"area": 188020,
"bbox": [
0,
0,
1280,
304
],
"category_id": 28,
"id": 15222,
"image_id": "1096_sTtQpCDAZaQ_00000017",
"iscrowd": 0,
"segmentation": {
"counts": "0Y9W=1O00000000000000000000000000000000000000000000001O000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 109_pAfVWACPtw8_00009148 | 109_pAfVWACPtw8_00009148.jpg | {
"data_source": "VIPSeg",
"file_name": "109_pAfVWACPtw8_00009148.jpg",
"height": 720,
"id": "109_pAfVWACPtw8_00009148",
"width": 1280
} | {
"caption": "A man in a blue and green patterned rash guard stands in a clear, turquoise lake holding a two small dark-furred otters close to his chest. In the background are lush green trees and a small waterfall visible in the background.",
"caption_ann": "A <2:man in a blue and green patterned rash guard> stands in a <0:clear, turquoise lake> holding a <3,4:two small dark-furred otters> close to his chest. In the background are lush <1:green trees> and a small waterfall visible in the background.",
"id": 1340,
"image_id": "109_pAfVWACPtw8_00009148",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "man in a blue and green patterned rash guard"
},
{
"mask_ids": [
0
],
"txt_desc": "clear, turquoise lake"
},
{
"mask_ids": [
3,
4
],
"txt_desc": "two small dark-furred otters"
},
{
"mask_ids": [
1
],
"txt_desc": "green trees"
}
],
"labels": [
"lake",
"tree",
"person",
"other_animal",
"other_animal"
]
} | [
{
"area": 604730,
"bbox": [
0,
133,
1280,
587
],
"category_id": 37,
"id": 15226,
"image_id": "109_pAfVWACPtw8_00009148",
"iscrowd": 0,
"segmentation": {
"counts": "Z4Vb0Z40000000O10000000000000000000000000000O10000O10000001O00001O000000000000000000001O00... | [
{
"id": 1,
"name": "object"
}
] |
train | 10_-xsT-ViO4e8_00000640 | 10_-xsT-ViO4e8_00000640.jpg | {
"data_source": "VIPSeg",
"file_name": "10_-xsT-ViO4e8_00000640.jpg",
"height": 720,
"id": "10_-xsT-ViO4e8_00000640",
"width": 1280
} | {
"caption": "In an outdoor setting, a woman with blonde hair wearing a blue t-shirt and a cream-colored, pumpkin-patterned dress sits atop a large, black horse equipped with a dark brown bridle. A second person in red skirt and black shoes trying to stand on a bright green barrel to assist the rider. The scene is set on a patch of grass and dirt outlined by grey concrete boundary in front of a weathered wooden barn, with a stand of tall, dry grass to the left and a dense forest trees in the background.cross-shaped fence visible behind the tall, dry grass.",
"caption_ann": "In an outdoor setting, a <6:woman with blonde hair wearing a blue t-shirt and a cream-colored, pumpkin-patterned dress> sits atop a <8:large, black horse equipped with a dark brown bridle>. A <7:second person in red skirt and black shoes> trying to stand on a <9:bright green barrel> to assist the rider. The scene is set on a patch of <2:grass and dirt> outlined by <1:grey concrete boundary> in front of a <3:weathered wooden barn>, with a stand of <5:tall, dry grass> to the left and a <4:dense forest trees> in the background.<0:cross-shaped fence> visible behind the <5:tall, dry grass>.",
"id": 1341,
"image_id": "10_-xsT-ViO4e8_00000640",
"label_matched": [
{
"mask_ids": [
6
],
"txt_desc": "woman with blonde hair wearing a blue t-shirt and a cream-colored, pumpkin-patterned dress"
},
{
"mask_ids": [
8
],
"txt_desc": "large, black horse equipped with a dark brown bridle"
},
{
"mask_ids": [
7
],
"txt_desc": "second person in red skirt and black shoes"
},
{
"mask_ids": [
9
],
"txt_desc": "bright green barrel"
},
{
"mask_ids": [
2
],
"txt_desc": "grass and dirt"
},
{
"mask_ids": [
1
],
"txt_desc": "grey concrete boundary"
},
{
"mask_ids": [
3
],
"txt_desc": "weathered wooden barn"
},
{
"mask_ids": [
5
],
"txt_desc": "tall, dry grass"
},
{
"mask_ids": [
4
],
"txt_desc": "dense forest trees"
},
{
"mask_ids": [
0
],
"txt_desc": "cross-shaped fence"
},
{
"mask_ids": [
5
],
"txt_desc": "tall, dry grass"
}
],
"labels": [
"handrail_or_fence",
"ground",
"grass",
"house",
"tree",
"other_plant",
"person",
"person",
"horse",
"barrel"
]
} | [
{
"area": 3752,
"bbox": [
398,
337,
124,
94
],
"category_id": 7,
"id": 15231,
"image_id": "10_-xsT-ViO4e8_00000640",
"iscrowd": 0,
"segmentation": {
"counts": "VWh8:le0910000000000000000000O100000000000000000000000000000000000000000000O100000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 111_-j2svV5ufH0_00000482 | 111_-j2svV5ufH0_00000482.jpg | {
"data_source": "VIPSeg",
"file_name": "111_-j2svV5ufH0_00000482.jpg",
"height": 720,
"id": "111_-j2svV5ufH0_00000482",
"width": 1280
} | {
"caption": "On an outdoor basketball court, a person wearing a black winter jacket and shorts over trousers dribbles a basketball towards a hoop with a white backboard on a grey pole, with a large brick building and bare trees in the background.",
"caption_ann": "On an <1:outdoor basketball court>, a <5:person wearing a black winter jacket and shorts over trousers> dribbles a <6:basketball> towards a <4:hoop with a white backboard> on a <0:grey pole>, with a <2:large brick building> and <3:bare trees> in the background.",
"id": 1342,
"image_id": "111_-j2svV5ufH0_00000482",
"label_matched": [
{
"mask_ids": [
1
],
"txt_desc": "outdoor basketball court"
},
{
"mask_ids": [
5
],
"txt_desc": "person wearing a black winter jacket and shorts over trousers"
},
{
"mask_ids": [
6
],
"txt_desc": "basketball"
},
{
"mask_ids": [
4
],
"txt_desc": "hoop with a white backboard"
},
{
"mask_ids": [
0
],
"txt_desc": "grey pole"
},
{
"mask_ids": [
2
],
"txt_desc": "large brick building"
},
{
"mask_ids": [
3
],
"txt_desc": "bare trees"
}
],
"labels": [
"pole",
"ground",
"house",
"tree",
"backboard",
"person",
"ball"
]
} | [
{
"area": 7072,
"bbox": [
628,
188,
28,
268
],
"category_id": 12,
"id": 15241,
"image_id": "111_-j2svV5ufH0_00000482",
"iscrowd": 0,
"segmentation": {
"counts": "lgi=\\8T>000000000000000001O00000000000000000000000000000gMY2_Na1oLTef=",
"size": [
... | [
{
"id": 1,
"name": "object"
}
] |
train | 1121_4utXb3auOew_00006663 | 1121_4utXb3auOew_00006663.jpg | {
"data_source": "VIPSeg",
"file_name": "1121_4utXb3auOew_00006663.jpg",
"height": 720,
"id": "1121_4utXb3auOew_00006663",
"width": 1280
} | {
"caption": "In a formal ceremony, three soldiers in dark blue uniforms guard the large, white stone tomb. A soldier stands at attention on the right holding a rifle, while two soldiers march on the left with a rifle. The ceremony takes place on a large paved plaza, where a white stand is placed in front of the monument. The tomb is flanked by large green hedges, with a distant mountain range in the background and patches of grass visible under a blue sky.",
"caption_ann": "In a formal ceremony, <7,8,9:three soldiers in dark blue uniforms> guard the <2:large, white stone tomb>. A <7:soldier> stands at attention on the right holding a <11:rifle>, while two <8,9:soldiers> march on the left with a <10:rifle>. The ceremony takes place on a large <0:paved plaza>, where a <6:white stand> is placed in front of the monument. The tomb is flanked by large <5:green hedges>, with a distant <4:mountain range> in the background and patches of <1:grass> visible under a <3:blue sky>.",
"id": 1343,
"image_id": "1121_4utXb3auOew_00006663",
"label_matched": [
{
"mask_ids": [
7,
8,
9
],
"txt_desc": "three soldiers in dark blue uniforms"
},
{
"mask_ids": [
2
],
"txt_desc": "large, white stone tomb"
},
{
"mask_ids": [
7
],
"txt_desc": "soldier"
},
{
"mask_ids": [
11
],
"txt_desc": "rifle"
},
{
"mask_ids": [
8,
9
],
"txt_desc": "soldiers"
},
{
"mask_ids": [
10
],
"txt_desc": "rifle"
},
{
"mask_ids": [
0
],
"txt_desc": "paved plaza"
},
{
"mask_ids": [
6
],
"txt_desc": "white stand"
},
{
"mask_ids": [
5
],
"txt_desc": "green hedges"
},
{
"mask_ids": [
4
],
"txt_desc": "mountain range"
},
{
"mask_ids": [
1
],
"txt_desc": "grass"
},
{
"mask_ids": [
3
],
"txt_desc": "blue sky"
}
],
"labels": [
"ground",
"grass",
"other_construction",
"sky",
"mountain",
"tree",
"shelf",
"person",
"person",
"person",
"gun",
"gun"
]
} | [
{
"area": 320398,
"bbox": [
0,
307,
1280,
413
],
"category_id": 14,
"id": 15248,
"image_id": "1121_4utXb3auOew_00006663",
"iscrowd": 0,
"segmentation": {
"counts": "Y=W9Y=00000000000000000000000000000000000000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1126_kVcxUWjFEiQ_00000032 | 1126_kVcxUWjFEiQ_00000032.jpg | {
"data_source": "VIPSeg",
"file_name": "1126_kVcxUWjFEiQ_00000032.jpg",
"height": 720,
"id": "1126_kVcxUWjFEiQ_00000032",
"width": 1280
} | {
"caption": "The view over the handlebars of a mountain bike shows the rider's gloved hands. The rider is following a person with a backpack on a red mountain bike along a narrow dirt path. Further ahead, another biker rides a bicycle. The path, lined with patches of green grass, runs along the edge of a steep canyon with a river visible at the bottom, all under a hazy sky.",
"caption_ann": "The view over the <6:handlebars> of a mountain bike shows the <10:rider's gloved hands>. The rider is following a <9:person with a backpack> on a <5:red mountain bike> along a narrow <1:dirt path>. Further ahead, another <8:biker> rides a <7:bicycle>. The path, lined with patches of <0:green grass>, runs along the edge of a steep <3:canyon> with a <4:river> visible at the bottom, all under a hazy <2:sky>.",
"id": 1344,
"image_id": "1126_kVcxUWjFEiQ_00000032",
"label_matched": [
{
"mask_ids": [
6
],
"txt_desc": "handlebars"
},
{
"mask_ids": [
10
],
"txt_desc": "rider's gloved hands"
},
{
"mask_ids": [
9
],
"txt_desc": "person with a backpack"
},
{
"mask_ids": [
5
],
"txt_desc": "red mountain bike"
},
{
"mask_ids": [
1
],
"txt_desc": "dirt path"
},
{
"mask_ids": [
8
],
"txt_desc": "biker"
},
{
"mask_ids": [
7
],
"txt_desc": "bicycle"
},
{
"mask_ids": [
0
],
"txt_desc": "green grass"
},
{
"mask_ids": [
3
],
"txt_desc": "canyon"
},
{
"mask_ids": [
4
],
"txt_desc": "river"
},
{
"mask_ids": [
2
],
"txt_desc": "sky"
}
],
"labels": [
"grass",
"path",
"sky",
"mountain",
"river",
"bicycle",
"bicycle",
"bicycle",
"person",
"person",
"person"
]
} | [
{
"area": 226256,
"bbox": [
0,
274,
1159,
446
],
"category_id": 15,
"id": 15260,
"image_id": "1126_kVcxUWjFEiQ_00000032",
"iscrowd": 0,
"segmentation": {
"counts": "a<P4g4XMj7i2THXMl7h2THXMl7i2RHXMn7h2RHXMn7i2QHWMn7j2QHWMo7j2PHVMP8j2oGWMQ8j2nGVMR8j2mGWMS... | [
{
"id": 1,
"name": "object"
}
] |
train | 1128_b-ECRS4R1Po_00000993 | 1128_b-ECRS4R1Po_00000993.jpg | {
"data_source": "VIPSeg",
"file_name": "1128_b-ECRS4R1Po_00000993.jpg",
"height": 720,
"id": "1128_b-ECRS4R1Po_00000993",
"width": 1280
} | {
"caption": "During a solemn ceremony, eight service members are gathered on a paved path next to a green lawn. A group of six Marines in two lines carry a casket draped with an American flag. Nearby, a Marine with his hand raised salutes while an officer in a white uniform stands at attention. A large tree trunk stands in the foreground, and a green wall or fence is visible in the background.",
"caption_ann": "During a solemn ceremony, <6,7,8,9,10,11,12,13:eight service members> are gathered on a <2:paved path> next to a <1:green lawn>. A group of <7,8,9,10,11,12:six Marines in two lines> carry a <5:casket draped with an American flag>. Nearby, a <6:Marine with his hand raised> salutes while an <13:officer in a white uniform> stands at attention. A <4:large tree trunk> stands in the foreground, and a <0:green wall or fence> is visible in the background.",
"id": 1345,
"image_id": "1128_b-ECRS4R1Po_00000993",
"label_matched": [
{
"mask_ids": [
6,
7,
8,
9,
10,
11,
12,
13
],
"txt_desc": "eight service members"
},
{
"mask_ids": [
2
],
"txt_desc": "paved path"
},
{
"mask_ids": [
1
],
"txt_desc": "green lawn"
},
{
"mask_ids": [
7,
8,
9,
10,
11,
12
],
"txt_desc": "six Marines in two lines"
},
{
"mask_ids": [
5
],
"txt_desc": "casket draped with an American flag"
},
{
"mask_ids": [
6
],
"txt_desc": "Marine with his hand raised"
},
{
"mask_ids": [
13
],
"txt_desc": "officer in a white uniform"
},
{
"mask_ids": [
4
],
"txt_desc": "large tree trunk"
},
{
"mask_ids": [
0
],
"txt_desc": "green wall or fence"
}
],
"labels": [
"wall",
"grass",
"path",
"wood",
"tree",
"flag",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person"
]
} | [
{
"area": 138153,
"bbox": [
0,
0,
1280,
244
],
"category_id": 0,
"id": 15271,
"image_id": "1128_b-ECRS4R1Po_00000993",
"iscrowd": 0,
"segmentation": {
"counts": "0W6Y`00000000000001O0000000000001O00000000000000000000000000001O00000000001O0000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 112_KfcWKl62A5Y_00000865 | 112_KfcWKl62A5Y_00000865.jpg | {
"data_source": "VIPSeg",
"file_name": "112_KfcWKl62A5Y_00000865.jpg",
"height": 720,
"id": "112_KfcWKl62A5Y_00000865",
"width": 1280
} | {
"caption": "A man wearing a black leather jacket and a light blue cap gestures towards the camera while sitting at a long wooden table with a silver laptop and a black microphone, in front of a row of beige, front-loading washing machines set against a cinder block wall that has a white sign and a vertical pipe. There is a small partially visible window in the right corner of the image.",
"caption_ann": "A <5:man wearing a black leather jacket and a light blue cap> gestures towards the camera while sitting at a <6:long wooden table> with a <12:silver laptop> and a <3:black microphone>, in front of a row of <7,8,9,10,11:beige, front-loading washing machines> set against a <0:cinder block wall> that has a <1:white sign> and a <2:vertical pipe>. There is a small partially visible <4:window> in the right corner of the image.",
"id": 1346,
"image_id": "112_KfcWKl62A5Y_00000865",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "man wearing a black leather jacket and a light blue cap"
},
{
"mask_ids": [
6
],
"txt_desc": "long wooden table"
},
{
"mask_ids": [
12
],
"txt_desc": "silver laptop"
},
{
"mask_ids": [
3
],
"txt_desc": "black microphone"
},
{
"mask_ids": [
7,
8,
9,
10,
11
],
"txt_desc": "beige, front-loading washing machines"
},
{
"mask_ids": [
0
],
"txt_desc": "cinder block wall"
},
{
"mask_ids": [
1
],
"txt_desc": "white sign"
},
{
"mask_ids": [
2
],
"txt_desc": "vertical pipe"
},
{
"mask_ids": [
4
],
"txt_desc": "window"
}
],
"labels": [
"wall",
"billboard_or_Bulletin_Board",
"pipeline",
"other_electronic_product",
"window",
"person",
"table_or_desk",
"washing_machine",
"washing_machine",
"washing_machine",
"washing_machine",
"washing_machine",
"computer"
]
} | [
{
"area": 118027,
"bbox": [
383,
0,
897,
530
],
"category_id": 0,
"id": 15285,
"image_id": "112_KfcWKl62A5Y_00000865",
"iscrowd": 0,
"segmentation": {
"counts": "`Y]81_f000000000000000000000000000001O0000000000001O0000000000000000000000000000001O00000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1135_Pvi7mPwBX2Q_00000377 | 1135_Pvi7mPwBX2Q_00000377.jpg | {
"data_source": "VIPSeg",
"file_name": "1135_Pvi7mPwBX2Q_00000377.jpg",
"height": 720,
"id": "1135_Pvi7mPwBX2Q_00000377",
"width": 1280
} | {
"caption": "In a covered stable with a corrugated metal roof, three brown horses stand on the sandy ground. They eat from two metal feeding tubs. The stable is constructed with four wooden support pillars and a wooden fence, and is set in front of a dry, grassy hill.",
"caption_ann": "In a covered stable with a <0:corrugated metal roof>, <5,6,7:three brown horses> stand on the <3:sandy ground>. They eat from <8,9:two metal feeding tubs>. The stable is constructed with <2:four wooden support pillars> and a <1:wooden fence>, and is set in front of a <4:dry, grassy hill>.",
"id": 1347,
"image_id": "1135_Pvi7mPwBX2Q_00000377",
"label_matched": [
{
"mask_ids": [
0
],
"txt_desc": "corrugated metal roof"
},
{
"mask_ids": [
5,
6,
7
],
"txt_desc": "three brown horses"
},
{
"mask_ids": [
3
],
"txt_desc": "sandy ground"
},
{
"mask_ids": [
8,
9
],
"txt_desc": "two metal feeding tubs"
},
{
"mask_ids": [
2
],
"txt_desc": "four wooden support pillars"
},
{
"mask_ids": [
1
],
"txt_desc": "wooden fence"
},
{
"mask_ids": [
4
],
"txt_desc": "dry, grassy hill"
}
],
"labels": [
"ceiling",
"handrail_or_fence",
"pillar",
"ground",
"mountain",
"horse",
"horse",
"horse",
"tub_or_bowl_or_pot",
"tub_or_bowl_or_pot"
]
} | [
{
"area": 165777,
"bbox": [
0,
0,
1280,
210
],
"category_id": 1,
"id": 15298,
"image_id": "1135_Pvi7mPwBX2Q_00000377",
"iscrowd": 0,
"segmentation": {
"counts": "0R5^a0000000000000000000000000000000000000000000000000001O0000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1135__VZ0gDib37o_00001098 | 1135__VZ0gDib37o_00001098.jpg | {
"data_source": "VIPSeg",
"file_name": "1135__VZ0gDib37o_00001098.jpg",
"height": 720,
"id": "1135__VZ0gDib37o_00001098",
"width": 1280
} | {
"caption": "In a well-furnished bedroom, a smiling woman in a long-sleeved pink dress stands between a large dark brown wardrobe and a bed with a large beige headboard. The bed is decorated with four green and gold pillows. The room also features a dark wood nightstand with a table lamp with a beige shade, a sheer white curtain, and an off-white wall.",
"caption_ann": "In a well-furnished bedroom, a <4:smiling woman in a long-sleeved pink dress> stands between a large <2:dark brown wardrobe> and a <5:bed with a large beige headboard>. The <5:bed> is decorated with <6,7,8,9:four green and gold pillows>. The room also features a <10:dark wood nightstand> with a <1:table lamp with a beige shade>, a <3:sheer white curtain>, and an <0:off-white wall>.",
"id": 1348,
"image_id": "1135__VZ0gDib37o_00001098",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "smiling woman in a long-sleeved pink dress"
},
{
"mask_ids": [
2
],
"txt_desc": "dark brown wardrobe"
},
{
"mask_ids": [
5
],
"txt_desc": "bed with a large beige headboard"
},
{
"mask_ids": [
5
],
"txt_desc": "bed"
},
{
"mask_ids": [
6,
7,
8,
9
],
"txt_desc": "four green and gold pillows"
},
{
"mask_ids": [
10
],
"txt_desc": "dark wood nightstand"
},
{
"mask_ids": [
1
],
"txt_desc": "table lamp with a beige shade"
},
{
"mask_ids": [
3
],
"txt_desc": "sheer white curtain"
},
{
"mask_ids": [
0
],
"txt_desc": "off-white wall"
}
],
"labels": [
"wall",
"lamp",
"cupboard_or_showcase_or_storage_rack",
"curtain",
"person",
"bed",
"pillow",
"pillow",
"pillow",
"pillow",
"table_or_desk"
]
} | [
{
"area": 178536,
"bbox": [
1,
0,
1279,
567
],
"category_id": 0,
"id": 15308,
"image_id": "1135__VZ0gDib37o_00001098",
"iscrowd": 0,
"segmentation": {
"counts": "`f0[;U;O1O100N200O1O1O1O1O100O1O1O1O1O100N200N2O1O1O1O1O1N2O1O1O1O1O1N200N200N2O1O1O1N2O1O1O... | [
{
"id": 1,
"name": "object"
}
] |
train | 1183_X7DEl4QgWT4_00000272 | 1183_X7DEl4QgWT4_00000272.jpg | {
"data_source": "VIPSeg",
"file_name": "1183_X7DEl4QgWT4_00000272.jpg",
"height": 720,
"id": "1183_X7DEl4QgWT4_00000272",
"width": 1280
} | {
"caption": "A man in a black fedora plays a large black accordion while sitting on a silver bench, with his open black instrument case on the grey pavement in front of him. A group of eight pedestrians walk past him against a tiled wall. The group includes a man in a brown jacket and cap carrying two white and red plastic bags, a person in a green hooded jacket, an woman with a ponytail carrying a blue handbag, and a man in a dark suit holding a black briefcase.",
"caption_ann": "A <4:man in a black fedora> plays a <17:large black accordion> while sitting on a <16:silver bench>, with his <11:open black instrument case> on the <1:grey pavement> in front of him. A group of <2,3,5,6,7,8,9,10:eight pedestrians> walk past him against a <0:tiled wall>. The group includes a <5:man in a brown jacket and cap> carrying <14,15:two white and red plastic bags>, a <6:person in a green hooded jacket>, an <8:woman with a ponytail> carrying a <13:blue handbag>, and a <7:man in a dark suit> holding a <12:black briefcase>.",
"id": 1349,
"image_id": "1183_X7DEl4QgWT4_00000272",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "man in a black fedora"
},
{
"mask_ids": [
17
],
"txt_desc": "large black accordion"
},
{
"mask_ids": [
16
],
"txt_desc": "silver bench"
},
{
"mask_ids": [
11
],
"txt_desc": "open black instrument case"
},
{
"mask_ids": [
1
],
"txt_desc": "grey pavement"
},
{
"mask_ids": [
2,
3,
5,
6,
7,
8,
9,
10
],
"txt_desc": "eight pedestrians"
},
{
"mask_ids": [
0
],
"txt_desc": "tiled wall"
},
{
"mask_ids": [
5
],
"txt_desc": "man in a brown jacket and cap"
},
{
"mask_ids": [
14,
15
],
"txt_desc": "two white and red plastic bags"
},
{
"mask_ids": [
6
],
"txt_desc": "person in a green hooded jacket"
},
{
"mask_ids": [
8
],
"txt_desc": "woman with a ponytail"
},
{
"mask_ids": [
13
],
"txt_desc": "blue handbag"
},
{
"mask_ids": [
7
],
"txt_desc": "man in a dark suit"
},
{
"mask_ids": [
12
],
"txt_desc": "black briefcase"
}
],
"labels": [
"wall",
"ground",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"box",
"bag_or_package",
"bag_or_package",
"bag_or_package",
"bag_or_package",
"chair_or_seat",
"instrument"
]
} | [
{
"area": 248292,
"bbox": [
4,
0,
1276,
566
],
"category_id": 0,
"id": 15319,
"image_id": "1183_X7DEl4QgWT4_00000272",
"iscrowd": 0,
"segmentation": {
"counts": "fj22\\f03L3N3@`0O1O001O1O1O001O00O1M3L4L4L4L4L4M3O100O1O100OQ^80PbG2N1O2N1O2N2N4L4L4S^O]Oe=f... | [
{
"id": 1,
"name": "object"
}
] |
train | 1185_ifHRDBx-ctw_00000212 | 1185_ifHRDBx-ctw_00000212.jpg | {
"data_source": "VIPSeg",
"file_name": "1185_ifHRDBx-ctw_00000212.jpg",
"height": 720,
"id": "1185_ifHRDBx-ctw_00000212",
"width": 1280
} | {
"caption": "Three young men sit on two black chairs in a grassy field with white flowers, playing their accordions in front of a stone wall. The man in the center wearing a red polo shirt plays a red accordion, the man on the left in a grey and blue shirt plays a black accordion, and the man on the right in a white t-shirt plays a black Weltmeister accordion.",
"caption_ann": "<2,3,4:Three young men> sit on <5,6:two black chairs> in a <1:grassy field with white flowers>, playing their accordions in front of a <0:stone wall>. The <3:man in the center wearing a red polo shirt> plays a <7:red accordion>, the <2:man on the left in a grey and blue shirt> plays a <8:black accordion>, and the <4:man on the right in a white t-shirt> plays a <9:black Weltmeister accordion>.",
"id": 1350,
"image_id": "1185_ifHRDBx-ctw_00000212",
"label_matched": [
{
"mask_ids": [
2,
3,
4
],
"txt_desc": "Three young men"
},
{
"mask_ids": [
5,
6
],
"txt_desc": "two black chairs"
},
{
"mask_ids": [
1
],
"txt_desc": "grassy field with white flowers"
},
{
"mask_ids": [
0
],
"txt_desc": "stone wall"
},
{
"mask_ids": [
3
],
"txt_desc": "man in the center wearing a red polo shirt"
},
{
"mask_ids": [
7
],
"txt_desc": "red accordion"
},
{
"mask_ids": [
2
],
"txt_desc": "man on the left in a grey and blue shirt"
},
{
"mask_ids": [
8
],
"txt_desc": "black accordion"
},
{
"mask_ids": [
4
],
"txt_desc": "man on the right in a white t-shirt"
},
{
"mask_ids": [
9
],
"txt_desc": "black Weltmeister accordion"
}
],
"labels": [
"wall",
"grass",
"person",
"person",
"person",
"chair_or_seat",
"chair_or_seat",
"instrument",
"instrument",
"instrument"
]
} | [
{
"area": 175650,
"bbox": [
0,
0,
1280,
353
],
"category_id": 0,
"id": 15337,
"image_id": "1185_ifHRDBx-ctw_00000212",
"iscrowd": 0,
"segmentation": {
"counts": "0W1Ye0001O00001O0000001O001O000000001O00001O0000001O001O000000001O00001O0000001O001O00000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1192_4k7fcPh3jCU_00000137 | 1192_4k7fcPh3jCU_00000137.jpg | {
"data_source": "VIPSeg",
"file_name": "1192_4k7fcPh3jCU_00000137.jpg",
"height": 720,
"id": "1192_4k7fcPh3jCU_00000137",
"width": 1280
} | {
"caption": "In a backyard, a young child in a striped hoodie and a cap aims an orange toy bow and a blue-fletched arrow at a target. The child stands on a concrete path next to a patch of dirt ground and the house with light-colored siding. In the background, a white vinyl fence stands in front of green trees under a clear blue sky.",
"caption_ann": "In a backyard, a <7:young child in a striped hoodie and a cap> aims an <6:orange toy bow and a blue-fletched arrow> at a target. The child stands on a <2:concrete path> next to a patch of <1:dirt ground> and the <3:house with light-colored siding>. In the background, a <0:white vinyl fence> stands in front of <5:green trees> under a <4:clear blue sky>.",
"id": 1351,
"image_id": "1192_4k7fcPh3jCU_00000137",
"label_matched": [
{
"mask_ids": [
7
],
"txt_desc": "young child in a striped hoodie and a cap"
},
{
"mask_ids": [
6
],
"txt_desc": "orange toy bow and a blue-fletched arrow"
},
{
"mask_ids": [
2
],
"txt_desc": "concrete path"
},
{
"mask_ids": [
1
],
"txt_desc": "dirt ground"
},
{
"mask_ids": [
3
],
"txt_desc": "house with light-colored siding"
},
{
"mask_ids": [
0
],
"txt_desc": "white vinyl fence"
},
{
"mask_ids": [
5
],
"txt_desc": "green trees"
},
{
"mask_ids": [
4
],
"txt_desc": "clear blue sky"
}
],
"labels": [
"handrail_or_fence",
"ground",
"path",
"house",
"sky",
"tree",
"toy",
"person"
]
} | [
{
"area": 93971,
"bbox": [
865,
231,
415,
297
],
"category_id": 7,
"id": 15347,
"image_id": "1192_4k7fcPh3jCU_00000137",
"iscrowd": 0,
"segmentation": {
"counts": "]dPc0Q1[c0T2lMT2mNS10000000000000000000000000000000000K5H8G9G9H8N200O100O100001O0O16J:F;Ef... | [
{
"id": 1,
"name": "object"
}
] |
train | 1192_P77M15bgQWs_00002005 | 1192_P77M15bgQWs_00002005.jpg | {
"data_source": "VIPSeg",
"file_name": "1192_P77M15bgQWs_00002005.jpg",
"height": 720,
"id": "1192_P77M15bgQWs_00002005",
"width": 1280
} | {
"caption": "A dynamic, low-angle shot captures four four female runners in mid-stride during a race on an outdoor athletic field. A runner with blond hair and an LSU-branded uniform is in the foreground, running alongside a runner in a black uniform. In the background, a runner wearing a black uniform with a white graphic on it and a partially visible runner are also competing. The background features a vibrant green grassy area and a sign for ESPN NCAA.",
"caption_ann": "A dynamic, low-angle shot captures four <2,3,4,5:four female runners> in mid-stride during a race on an outdoor <1:athletic field>. A <5:runner with blond hair and an LSU-branded uniform> is in the foreground, running alongside a <3:runner in a black uniform>. In the background, a <4:runner wearing a black uniform with a white graphic on it> and a <2:partially visible runner> are also competing. The background features a vibrant <0:green grassy area> and a sign for ESPN NCAA.",
"id": 1352,
"image_id": "1192_P77M15bgQWs_00002005",
"label_matched": [
{
"mask_ids": [
2,
3,
4,
5
],
"txt_desc": "four female runners"
},
{
"mask_ids": [
1
],
"txt_desc": "athletic field"
},
{
"mask_ids": [
5
],
"txt_desc": "runner with blond hair and an LSU-branded uniform"
},
{
"mask_ids": [
3
],
"txt_desc": "runner in a black uniform"
},
{
"mask_ids": [
4
],
"txt_desc": "runner wearing a black uniform with a white graphic on it"
},
{
"mask_ids": [
2
],
"txt_desc": "partially visible runner"
},
{
"mask_ids": [
0
],
"txt_desc": "green grassy area"
}
],
"labels": [
"grass",
"athletic_field",
"person",
"person",
"person",
"person"
]
} | [
{
"area": 97269,
"bbox": [
328,
0,
952,
222
],
"category_id": 15,
"id": 15355,
"image_id": "1192_P77M15bgQWs_00002005",
"iscrowd": 0,
"segmentation": {
"counts": "PdV71_f00000001O0000001O000000001O0000001O0000001O0000001O000000001O0000001O0000001O0000001... | [
{
"id": 1,
"name": "object"
}
] |
train | 1193_6cwPvrMW7tE_00000272 | 1193_6cwPvrMW7tE_00000272.jpg | {
"data_source": "VIPSeg",
"file_name": "1193_6cwPvrMW7tE_00000272.jpg",
"height": 720,
"id": "1193_6cwPvrMW7tE_00000272",
"width": 1280
} | {
"caption": "A man in a grey t-shirt, glasses, and a camouflage cap stands outside holding up a packaged green toy bow and arrow set. He is standing on a patch of dirt and grass in front of a wooden house or deck and a background of dense green trees.",
"caption_ann": "A <4:man in a grey t-shirt, glasses, and a camouflage cap> stands outside holding up a <3:packaged green toy bow and arrow set>. He is standing on a <0:patch of dirt and grass> in front of a <1:wooden house or deck> and a background of <2:dense green trees>.",
"id": 1353,
"image_id": "1193_6cwPvrMW7tE_00000272",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "man in a grey t-shirt, glasses, and a camouflage cap"
},
{
"mask_ids": [
3
],
"txt_desc": "packaged green toy bow and arrow set"
},
{
"mask_ids": [
0
],
"txt_desc": "patch of dirt and grass"
},
{
"mask_ids": [
1
],
"txt_desc": "wooden house or deck"
},
{
"mask_ids": [
2
],
"txt_desc": "dense green trees"
}
],
"labels": [
"grass",
"house",
"tree",
"toy",
"person"
]
} | [
{
"area": 67127,
"bbox": [
0,
460,
527,
259
],
"category_id": 15,
"id": 15361,
"image_id": "1193_6cwPvrMW7tE_00000272",
"iscrowd": 0,
"segmentation": {
"counts": "\\>S8^>O001O1O001O1O002N001O001O002N001O1O001O1O001O1O001O1O001O1O1O001O002N001O001O1O1O001... | [
{
"id": 1,
"name": "object"
}
] |
train | 1194_ASFK3OCpctI_00001037 | 1194_ASFK3OCpctI_00001037.jpg | {
"data_source": "VIPSeg",
"file_name": "1194_ASFK3OCpctI_00001037.jpg",
"height": 720,
"id": "1194_ASFK3OCpctI_00001037",
"width": 1280
} | {
"caption": "At an indoor archery range with a beige wall with blue targets and a white ceiling, two children are practicing on a smooth grey floor. The young girl with a ponytail in a dark t-shirt picks up an arrow from the quiver. The young boy in a blue and white t-shirt stands beside a quiver of colorful arrows.",
"caption_ann": "At an indoor archery range with a <0:beige wall with blue targets> and a <1:white ceiling>, <4,5:two children> are practicing on a <2:smooth grey floor>. The <4:young girl with a ponytail in a dark t-shirt> picks up an arrow from the <3:quiver>. The <5:young boy in a blue and white t-shirt> stands beside a <3:quiver of colorful arrows>.",
"id": 1354,
"image_id": "1194_ASFK3OCpctI_00001037",
"label_matched": [
{
"mask_ids": [
0
],
"txt_desc": "beige wall with blue targets"
},
{
"mask_ids": [
1
],
"txt_desc": "white ceiling"
},
{
"mask_ids": [
4,
5
],
"txt_desc": "two children"
},
{
"mask_ids": [
2
],
"txt_desc": "smooth grey floor"
},
{
"mask_ids": [
4
],
"txt_desc": "young girl with a ponytail in a dark t-shirt"
},
{
"mask_ids": [
3
],
"txt_desc": "quiver"
},
{
"mask_ids": [
5
],
"txt_desc": "young boy in a blue and white t-shirt"
},
{
"mask_ids": [
3
],
"txt_desc": "quiver of colorful arrows"
}
],
"labels": [
"wall",
"ceiling",
"floor",
"toy",
"person",
"person"
]
} | [
{
"area": 402115,
"bbox": [
0,
0,
1280,
476
],
"category_id": 0,
"id": 15366,
"image_id": "1194_ASFK3OCpctI_00001037",
"iscrowd": 0,
"segmentation": {
"counts": "1k>e7O100O100O10000O1O1000000O1O10000O1O10000O100O100O100O10000O1O1000000O1O10000O1O10000O10... | [
{
"id": 1,
"name": "object"
}
] |
train | 1195_g20Acgu1Lgo_00003595 | 1195_g20Acgu1Lgo_00003595.jpg | {
"data_source": "VIPSeg",
"file_name": "1195_g20Acgu1Lgo_00003595.jpg",
"height": 720,
"id": "1195_g20Acgu1Lgo_00003595",
"width": 1280
} | {
"caption": "Five runners are competing in a race on a track and field surface with numbered lanes, in front of a patch of green grass. A runner in a maroon and white uniform is in the lead, passing a large digital timer displaying the number time 19.6. Behind him, a runner in a red uniform, a runner in a yellow and black uniform, and a runner in a white and orange uniform follow closely. A referee in white and a person standing are in the middle of the grassy area looking at the competition. ",
"caption_ann": "<3,4,5,6,9:Five runners> are competing in a race on a <1:track and field surface> with numbered lanes, in front of a patch of <0:green grass>. A <6:runner in a maroon and white uniform> is in the lead, passing a <2:large digital timer displaying the number time 19.6>. Behind him, a <5:runner in a red uniform>, a <3:runner in a yellow and black uniform>, and a <4:runner in a white and orange uniform> follow closely. A <7:referee in white> and a <8:person standing> are in the middle of the <0:grassy area> looking at the competition. ",
"id": 1355,
"image_id": "1195_g20Acgu1Lgo_00003595",
"label_matched": [
{
"mask_ids": [
3,
4,
5,
6,
9
],
"txt_desc": "Five runners"
},
{
"mask_ids": [
1
],
"txt_desc": "track and field surface"
},
{
"mask_ids": [
0
],
"txt_desc": "green grass"
},
{
"mask_ids": [
6
],
"txt_desc": "runner in a maroon and white uniform"
},
{
"mask_ids": [
2
],
"txt_desc": "large digital timer displaying the number time 19.6"
},
{
"mask_ids": [
5
],
"txt_desc": "runner in a red uniform"
},
{
"mask_ids": [
3
],
"txt_desc": "runner in a yellow and black uniform"
},
{
"mask_ids": [
4
],
"txt_desc": "runner in a white and orange uniform"
},
{
"mask_ids": [
7
],
"txt_desc": "referee in white"
},
{
"mask_ids": [
8
],
"txt_desc": "person standing"
},
{
"mask_ids": [
0
],
"txt_desc": "grassy area"
}
],
"labels": [
"grass",
"athletic_field",
"billboard_or_Bulletin_Board",
"person",
"person",
"person",
"person",
"person",
"person",
"person"
]
} | [
{
"area": 215804,
"bbox": [
260,
17,
1020,
606
],
"category_id": 15,
"id": 15372,
"image_id": "1195_g20Acgu1Lgo_00003595",
"iscrowd": 0,
"segmentation": {
"counts": "kkf52_f02M4L3M1O00000000O100000001O0O100000000O11O0000000000010O0000000000001O0000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1198_xPxvRxZRb_Y_00002830 | 1198_xPxvRxZRb_Y_00002830.jpg | {
"data_source": "VIPSeg",
"file_name": "1198_xPxvRxZRb_Y_00002830.jpg",
"height": 720,
"id": "1198_xPxvRxZRb_Y_00002830",
"width": 1280
} | {
"caption": "During a race, three runners are competing on a red running track with white lanes next to the green infield grass. A runner in a maroon and white jersey is in the forground, followed by a runner in maroon jersey and a runner in a yellow jersey. A sign for the Track & Field Championships is placed nearby, next to a a person in white full sleeved shirt and dark athletic pants standing and cheering. In the background, we see three partially visible people passing by, a bright red conical roadblock and a dark article of clothing lying on the grass.",
"caption_ann": "During a race, <5,6,7:three runners> are competing on a <1:red running track with white lanes> next to the <0:green infield grass>. A <7:runner in a maroon and white jersey> is in the forground, followed by a <6:runner in maroon jersey> and a <5:runner in a yellow jersey>. A <2:sign for the Track & Field Championships> is placed nearby, next to a <8:a person in white full sleeved shirt and dark athletic pants> standing and cheering. In the background, we see <9,10,11:three partially visible people> passing by, a <4:bright red conical roadblock> and a <3:dark article of clothing> lying on the <0:grass>.",
"id": 1356,
"image_id": "1198_xPxvRxZRb_Y_00002830",
"label_matched": [
{
"mask_ids": [
5,
6,
7
],
"txt_desc": "three runners"
},
{
"mask_ids": [
1
],
"txt_desc": "red running track with white lanes"
},
{
"mask_ids": [
0
],
"txt_desc": "green infield grass"
},
{
"mask_ids": [
7
],
"txt_desc": "runner in a maroon and white jersey"
},
{
"mask_ids": [
6
],
"txt_desc": "runner in maroon jersey"
},
{
"mask_ids": [
5
],
"txt_desc": "runner in a yellow jersey"
},
{
"mask_ids": [
2
],
"txt_desc": "sign for the Track & Field Championships"
},
{
"mask_ids": [
8
],
"txt_desc": "a person in white full sleeved shirt and dark athletic pants"
},
{
"mask_ids": [
9,
10,
11
],
"txt_desc": "three partially visible people"
},
{
"mask_ids": [
4
],
"txt_desc": "bright red conical roadblock"
},
{
"mask_ids": [
3
],
"txt_desc": "dark article of clothing"
},
{
"mask_ids": [
0
],
"txt_desc": "grass"
}
],
"labels": [
"grass",
"athletic_field",
"billboard_or_Bulletin_Board",
"clothes",
"roadblock",
"person",
"person",
"person",
"person",
"person",
"person",
"person"
]
} | [
{
"area": 220385,
"bbox": [
0,
0,
1280,
479
],
"category_id": 15,
"id": 15382,
"image_id": "1198_xPxvRxZRb_Y_00002830",
"iscrowd": 0,
"segmentation": {
"counts": "0Q2_d000001O00001O00001O00001O00001O0000001O00001O00001O00001O00001O00001O00001O00001O00001... | [
{
"id": 1,
"name": "object"
}
] |
train | 11_0jBjg717TJc_00000325 | 11_0jBjg717TJc_00000325.jpg | {
"data_source": "VIPSeg",
"file_name": "11_0jBjg717TJc_00000325.jpg",
"height": 720,
"id": "11_0jBjg717TJc_00000325",
"width": 1280
} | {
"caption": "A person wearing a dark helmet and a white top rides a white horse on the left side of a vast, open green field. The landscape stretches out under a heavily clouded sky, with a line of dark green trees visible on the distant horizon.",
"caption_ann": "A <3:person wearing a dark helmet and a white top> rides a <4:white horse> on the left side of a vast, <0:open green field>. The landscape stretches out under a <1:heavily clouded sky>, with a line of <2:dark green trees> visible on the distant horizon.",
"id": 1357,
"image_id": "11_0jBjg717TJc_00000325",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "person wearing a dark helmet and a white top"
},
{
"mask_ids": [
4
],
"txt_desc": "white horse"
},
{
"mask_ids": [
0
],
"txt_desc": "open green field"
},
{
"mask_ids": [
1
],
"txt_desc": "heavily clouded sky"
},
{
"mask_ids": [
2
],
"txt_desc": "dark green trees"
}
],
"labels": [
"grass",
"sky",
"tree",
"person",
"horse"
]
} | [
{
"area": 497661,
"bbox": [
0,
306,
1280,
414
],
"category_id": 15,
"id": 15394,
"image_id": "11_0jBjg717TJc_00000325",
"iscrowd": 0,
"segmentation": {
"counts": "b9n<b9O10000000000000000000001O0000000000000000000000001O000000000000000000000000001O000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 11_OiKZma1JIUc_00000436 | 11_OiKZma1JIUc_00000436.jpg | {
"data_source": "VIPSeg",
"file_name": "11_OiKZma1JIUc_00000436.jpg",
"height": 720,
"id": "11_OiKZma1JIUc_00000436",
"width": 1280
} | {
"caption": "A young boy in a gray hoodie and black pants rides a light blue snowboard down a white artificial ski slope. Further up the slope, other people are enjoying the activity: a person in a blue top sits on a white sled with red lettering, a boy in a dark top is on a white sled with blue lettering, and a small child wearing a helmet slides down on a blue sled. The slope is bordered by a grassy bank with some dry brown ground, where a person dressed in black stands near a gray pole.",
"caption_ann": "A <4:young boy in a gray hoodie and black pants> rides a <12:light blue snowboard> down a <3:white artificial ski slope>. Further up the slope, other people are enjoying the activity: a <5:person in a blue top> sits on a <11:white sled with red lettering>, a <6:boy in a dark top> is on a <10:white sled with blue lettering>, and a <7:small child wearing a helmet> slides down on a <9:blue sled>. The slope is bordered by a <2:grassy bank> with some <1:dry brown ground>, where a <8:person dressed in black> stands near a <0:gray pole>.",
"id": 1358,
"image_id": "11_OiKZma1JIUc_00000436",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "young boy in a gray hoodie and black pants"
},
{
"mask_ids": [
12
],
"txt_desc": "light blue snowboard"
},
{
"mask_ids": [
3
],
"txt_desc": "white artificial ski slope"
},
{
"mask_ids": [
5
],
"txt_desc": "person in a blue top"
},
{
"mask_ids": [
11
],
"txt_desc": "white sled with red lettering"
},
{
"mask_ids": [
6
],
"txt_desc": "boy in a dark top"
},
{
"mask_ids": [
10
],
"txt_desc": "white sled with blue lettering"
},
{
"mask_ids": [
7
],
"txt_desc": "small child wearing a helmet"
},
{
"mask_ids": [
9
],
"txt_desc": "blue sled"
},
{
"mask_ids": [
2
],
"txt_desc": "grassy bank"
},
{
"mask_ids": [
1
],
"txt_desc": "dry brown ground"
},
{
"mask_ids": [
8
],
"txt_desc": "person dressed in black"
},
{
"mask_ids": [
0
],
"txt_desc": "gray pole"
}
],
"labels": [
"pole",
"ground",
"grass",
"snowfield",
"person",
"person",
"person",
"person",
"person",
"skateboard",
"skateboard",
"skateboard",
"skateboard"
]
} | [
{
"area": 6971,
"bbox": [
639,
0,
129,
177
],
"category_id": 12,
"id": 15399,
"image_id": "11_OiKZma1JIUc_00000436",
"iscrowd": 0,
"segmentation": {
"counts": "Y\\Q>h2o`0i20000000000000000000000000000000O11O0000000000000000000000000000000000O1N2YMWoZ1[MU... | [
{
"id": 1,
"name": "object"
}
] |
train | 120_GbVSlpEAg9o_00000271 | 120_GbVSlpEAg9o_00000271.jpg | {
"data_source": "VIPSeg",
"file_name": "120_GbVSlpEAg9o_00000271.jpg",
"height": 720,
"id": "120_GbVSlpEAg9o_00000271",
"width": 1280
} | {
"caption": "A man in a black t-shirt and jeans pushes a red and black canal boat away from a concrete dock into the water, with modern buildings and a white sculptural bridge in the background.",
"caption_ann": "A <5:man in a black t-shirt and jeans> pushes a <4:red and black canal boat> away from a <0:concrete dock> into the <3:water>, with <1:modern buildings> and a <2:white sculptural bridge> in the background.",
"id": 1359,
"image_id": "120_GbVSlpEAg9o_00000271",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "man in a black t-shirt and jeans"
},
{
"mask_ids": [
4
],
"txt_desc": "red and black canal boat"
},
{
"mask_ids": [
0
],
"txt_desc": "concrete dock"
},
{
"mask_ids": [
3
],
"txt_desc": "water"
},
{
"mask_ids": [
1
],
"txt_desc": "modern buildings"
},
{
"mask_ids": [
2
],
"txt_desc": "white sculptural bridge"
}
],
"labels": [
"ground",
"building",
"other_construction",
"water",
"ship_or_boat",
"person"
]
} | [
{
"area": 161795,
"bbox": [
0,
340,
695,
380
],
"category_id": 14,
"id": 15412,
"image_id": "120_GbVSlpEAg9o_00000271",
"iscrowd": 0,
"segmentation": {
"counts": "d:l;d:000000000000000001O0000000000000000001O000000000000000000e0[Of0ZO00000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1220_eITb-5R76TU_00002500 | 1220_eITb-5R76TU_00002500.jpg | {
"data_source": "VIPSeg",
"file_name": "1220_eITb-5R76TU_00002500.jpg",
"height": 720,
"id": "1220_eITb-5R76TU_00002500",
"width": 1280
} | {
"caption": "A vintage white car with roof lights speeds down a sandy dirt road, kicking up dust next to a sparse, dry grassy field and a barbed-wire fence with wooden posts, all under a clear blue sky.",
"caption_ann": "A <4:vintage white car with roof lights> speeds down a <2:sandy dirt road>, kicking up dust next to a <1:sparse, dry grassy field> and a <0:barbed-wire fence with wooden posts>, all under a <3:clear blue sky>.",
"id": 1360,
"image_id": "1220_eITb-5R76TU_00002500",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "vintage white car with roof lights"
},
{
"mask_ids": [
2
],
"txt_desc": "sandy dirt road"
},
{
"mask_ids": [
1
],
"txt_desc": "sparse, dry grassy field"
},
{
"mask_ids": [
0
],
"txt_desc": "barbed-wire fence with wooden posts"
},
{
"mask_ids": [
3
],
"txt_desc": "clear blue sky"
}
],
"labels": [
"handrail_or_fence",
"grass",
"sand",
"sky",
"car"
]
} | [
{
"area": 16623,
"bbox": [
0,
213,
217,
312
],
"category_id": 7,
"id": 15418,
"image_id": "1220_eITb-5R76TU_00002500",
"iscrowd": 0,
"segmentation": {
"counts": "e6=o2OkNOk>1YB0lNOk>2YBNlN0m>1WBOkN0n>2VBOkN0P?0TB0mNOo>2TBNmN0P?2SBNlN0R?2QBOlNOT?2oAOnNNT?... | [
{
"id": 1,
"name": "object"
}
] |
train | 1230_AGY-gQ_3O8Y_00000205 | 1230_AGY-gQ_3O8Y_00000205.jpg | {
"data_source": "VIPSeg",
"file_name": "1230_AGY-gQ_3O8Y_00000205.jpg",
"height": 720,
"id": "1230_AGY-gQ_3O8Y_00000205",
"width": 1280
} | {
"caption": "A person in an orange shirt and black shorts plays badminton on a green badminton court with white lines next to a badminton net. The court is in a large hall with a pale green paneled wall decorated with four advertising banners and a round wall clock. Overhead is a high arched wooden ceiling with a long fluorescent light fixture, and on the side of the court are a wooden bench and a court cleaning machine.",
"caption_ann": "A <6:person in an orange shirt and black shorts> plays badminton on a <2:green badminton court with white lines> next to a <4:badminton net>. The court is in a large hall with a <0:pale green paneled wall> decorated with <8,9,10,11:four advertising banners> and a <12:round wall clock>. Overhead is a <1:high arched wooden ceiling> with a <3:long fluorescent light fixture>, and on the side of the court are a <7:wooden bench> and a <5:court cleaning machine>.",
"id": 1361,
"image_id": "1230_AGY-gQ_3O8Y_00000205",
"label_matched": [
{
"mask_ids": [
6
],
"txt_desc": "person in an orange shirt and black shorts"
},
{
"mask_ids": [
2
],
"txt_desc": "green badminton court with white lines"
},
{
"mask_ids": [
4
],
"txt_desc": "badminton net"
},
{
"mask_ids": [
0
],
"txt_desc": "pale green paneled wall"
},
{
"mask_ids": [
8,
9,
10,
11
],
"txt_desc": "four advertising banners"
},
{
"mask_ids": [
12
],
"txt_desc": "round wall clock"
},
{
"mask_ids": [
1
],
"txt_desc": "high arched wooden ceiling"
},
{
"mask_ids": [
3
],
"txt_desc": "long fluorescent light fixture"
},
{
"mask_ids": [
7
],
"txt_desc": "wooden bench"
},
{
"mask_ids": [
5
],
"txt_desc": "court cleaning machine"
}
],
"labels": [
"wall",
"ceiling",
"floor",
"lamp",
"ball_net",
"other_machine",
"person",
"bench",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"clock"
]
} | [
{
"area": 466103,
"bbox": [
0,
0,
1280,
557
],
"category_id": 0,
"id": 15423,
"image_id": "1230_AGY-gQ_3O8Y_00000205",
"iscrowd": 0,
"segmentation": {
"counts": "`4h9S2ZG`7f8`HYGa7g8_HYGb7f8^HZGb7f8^HZGb7f8^HYGd7f8[H[Ge7e8[H[Ge7e8[H[Gf7d8ZH[Gg7d8ZH\\Gf7d... | [
{
"id": 1,
"name": "object"
}
] |
train | 1233_HDAaojBqRl8_00000145 | 1233_HDAaojBqRl8_00000145.jpg | {
"data_source": "VIPSeg",
"file_name": "1233_HDAaojBqRl8_00000145.jpg",
"height": 720,
"id": "1233_HDAaojBqRl8_00000145",
"width": 1280
} | {
"caption": "In a large gymnasium with a blue wall, four people are playing a doubles badminton match on an orange court with white lines. The man in a grey shirt and the woman in a blue shirt are on one side of the badminton net, while the woman in an orange shirt and the man in a light blue shirt are on the other. The wall has three high windows and a yellow door. A white bench is visible in the background.",
"caption_ann": "In a large gymnasium with a <0:blue wall>, <7,8,9,10:four people> are playing a doubles badminton match on an <1:orange court with white lines>. The <7:man in a grey shirt> and the <8:woman in a blue shirt> are on one side of the <2:badminton net>, while the <9:woman in an orange shirt> and the <10:man in a light blue shirt> are on the other. The <0:wall> has <4,5,6:three high windows> and a <3:yellow door>. A <11:white bench> is visible in the background.",
"id": 1362,
"image_id": "1233_HDAaojBqRl8_00000145",
"label_matched": [
{
"mask_ids": [
0
],
"txt_desc": "blue wall"
},
{
"mask_ids": [
7,
8,
9,
10
],
"txt_desc": "four people"
},
{
"mask_ids": [
1
],
"txt_desc": "orange court with white lines"
},
{
"mask_ids": [
7
],
"txt_desc": "man in a grey shirt"
},
{
"mask_ids": [
8
],
"txt_desc": "woman in a blue shirt"
},
{
"mask_ids": [
2
],
"txt_desc": "badminton net"
},
{
"mask_ids": [
9
],
"txt_desc": "woman in an orange shirt"
},
{
"mask_ids": [
10
],
"txt_desc": "man in a light blue shirt"
},
{
"mask_ids": [
0
],
"txt_desc": "wall"
},
{
"mask_ids": [
4,
5,
6
],
"txt_desc": "three high windows"
},
{
"mask_ids": [
3
],
"txt_desc": "yellow door"
},
{
"mask_ids": [
11
],
"txt_desc": "white bench"
}
],
"labels": [
"wall",
"floor",
"ball_net",
"door",
"window",
"window",
"window",
"person",
"person",
"person",
"person",
"bench"
]
} | [
{
"area": 429630,
"bbox": [
0,
0,
1280,
453
],
"category_id": 0,
"id": 15436,
"image_id": "1233_HDAaojBqRl8_00000145",
"iscrowd": 0,
"segmentation": {
"counts": "0h1hd0P7PIQ1oN0aEUHY6k7gIUHY6k7gIUHY6k7gIUHY6k7gIUHY6k7gIUHY6k7gIUHY6k7gIUHY6k7hITHX6l7hITHX... | [
{
"id": 1,
"name": "object"
}
] |
train | 1237_h7cthCHYtYM_00000317 | 1237_h7cthCHYtYM_00000317.jpg | {
"data_source": "VIPSeg",
"file_name": "1237_h7cthCHYtYM_00000317.jpg",
"height": 720,
"id": "1237_h7cthCHYtYM_00000317",
"width": 1280
} | {
"caption": "A woman with glasses and a purple shirt operates a large office printer while holding a stack of papers, standing in front of a beige wall where a framed picture is hanging.",
"caption_ann": "A <2:woman with glasses and a purple shirt> operates a <4:large office printer> while holding a <1:stack of papers>, standing in front of a <0:beige wall> where a <3:framed picture> is hanging.",
"id": 1363,
"image_id": "1237_h7cthCHYtYM_00000317",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "woman with glasses and a purple shirt"
},
{
"mask_ids": [
4
],
"txt_desc": "large office printer"
},
{
"mask_ids": [
1
],
"txt_desc": "stack of papers"
},
{
"mask_ids": [
0
],
"txt_desc": "beige wall"
},
{
"mask_ids": [
3
],
"txt_desc": "framed picture"
}
],
"labels": [
"wall",
"book",
"person",
"painting_or_poster",
"printer"
]
} | [
{
"area": 376173,
"bbox": [
0,
0,
1116,
718
],
"category_id": 0,
"id": 15448,
"image_id": "1237_h7cthCHYtYM_00000317",
"iscrowd": 0,
"segmentation": {
"counts": "c8k2n=g5TMl2000O1000000O1000000O1000000O1000000O10000O1000000O1000000O1000000O1000000O100000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1238_vVirreWp5IY_00000377 | 1238_vVirreWp5IY_00000377.jpg | {
"data_source": "VIPSeg",
"file_name": "1238_vVirreWp5IY_00000377.jpg",
"height": 720,
"id": "1238_vVirreWp5IY_00000377",
"width": 1280
} | {
"caption": "A man in a blue polo shirt and khaki shorts stands in a showroom with three large office printers on a grey carpeted floor. He is using the printer, which is positioned between a printer wrapped in plastic and a dark printer. The scene is set in front of a light blue wall with a Copier1.com sign on it.",
"caption_ann": "A <2:man in a blue polo shirt and khaki shorts> stands in a showroom with <4,5,6:three large office printers> on a <1:grey carpeted floor>. He is using the <5:printer>, which is positioned between a <4:printer wrapped in plastic> and a <6:dark printer>. The scene is set in front of a <0:light blue wall> with a <3:Copier1.com sign> on it.",
"id": 1364,
"image_id": "1238_vVirreWp5IY_00000377",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "man in a blue polo shirt and khaki shorts"
},
{
"mask_ids": [
4,
5,
6
],
"txt_desc": "three large office printers"
},
{
"mask_ids": [
1
],
"txt_desc": "grey carpeted floor"
},
{
"mask_ids": [
5
],
"txt_desc": "printer"
},
{
"mask_ids": [
4
],
"txt_desc": "printer wrapped in plastic"
},
{
"mask_ids": [
6
],
"txt_desc": "dark printer"
},
{
"mask_ids": [
0
],
"txt_desc": "light blue wall"
},
{
"mask_ids": [
3
],
"txt_desc": "Copier1.com sign"
}
],
"labels": [
"wall",
"floor",
"person",
"painting_or_poster",
"printer",
"printer",
"printer"
]
} | [
{
"area": 400150,
"bbox": [
2,
0,
1278,
596
],
"category_id": 0,
"id": 15453,
"image_id": "1238_vVirreWp5IY_00000377",
"iscrowd": 0,
"segmentation": {
"counts": "P]1db0l3000000000000000000O100000000000000000000000000O10000000000000000000000000000O1000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1241_qYvEuwrSiXc_00000775 | 1241_qYvEuwrSiXc_00000775.jpg | {
"data_source": "VIPSeg",
"file_name": "1241_qYvEuwrSiXc_00000775.jpg",
"height": 720,
"id": "1241_qYvEuwrSiXc_00000775",
"width": 1280
} | {
"caption": "In an indoor badminton court, two people are on a green badminton court. A young man in a red and yellow shirt is in the foreground, while a woman in a black outfit stands on the other side of the badminton net. The hall has a dark green wall with five advertising banners on it, a high, arched ceiling, and long light fixtures. Near the man an opened cardboard box is visible surrounded by many shuttles on the court floor. A beige metal cupboard is present on the left side. A bench is visible in the background.",
"caption_ann": "In an indoor badminton court, <6,7:two people> are on a <2:green badminton court>. A <6:young man in a red and yellow shirt> is in the foreground, while a <7:woman in a black outfit> stands on the other side of the <4:badminton net>. The hall has a <0:dark green wall> with <10,11,12,13,14:five advertising banners> on it, a <1:high, arched ceiling>, and <3:long light fixtures>. Near the <6:man> an opened <8:cardboard box> is visible surrounded by many shuttles on the <2:court floor>. A <5:beige metal cupboard> is present on the left side. A <9:bench> is visible in the background.",
"id": 1365,
"image_id": "1241_qYvEuwrSiXc_00000775",
"label_matched": [
{
"mask_ids": [
6,
7
],
"txt_desc": "two people"
},
{
"mask_ids": [
2
],
"txt_desc": "green badminton court"
},
{
"mask_ids": [
6
],
"txt_desc": "young man in a red and yellow shirt"
},
{
"mask_ids": [
7
],
"txt_desc": "woman in a black outfit"
},
{
"mask_ids": [
4
],
"txt_desc": "badminton net"
},
{
"mask_ids": [
0
],
"txt_desc": "dark green wall"
},
{
"mask_ids": [
10,
11,
12,
13,
14
],
"txt_desc": "five advertising banners"
},
{
"mask_ids": [
1
],
"txt_desc": "high, arched ceiling"
},
{
"mask_ids": [
3
],
"txt_desc": "long light fixtures"
},
{
"mask_ids": [
6
],
"txt_desc": "man"
},
{
"mask_ids": [
8
],
"txt_desc": "cardboard box"
},
{
"mask_ids": [
2
],
"txt_desc": "court floor"
},
{
"mask_ids": [
5
],
"txt_desc": "beige metal cupboard"
},
{
"mask_ids": [
9
],
"txt_desc": "bench"
}
],
"labels": [
"wall",
"ceiling",
"floor",
"lamp",
"ball_net",
"cupboard_or_showcase_or_storage_rack",
"person",
"person",
"box",
"bench",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster"
]
} | [
{
"area": 145784,
"bbox": [
26,
1,
1254,
439
],
"category_id": 0,
"id": 15460,
"image_id": "1241_qYvEuwrSiXc_00000775",
"iscrowd": 0,
"segmentation": {
"counts": "geb0k0[a0S2Q@mMP>R4PBnKR;P7nDPIR;P7nDPIR;P7nDPIS;o6mDQIS;o6mDPIT;P7lDPIT;P7lDPIT;P7lDPIT;P7... | [
{
"id": 1,
"name": "object"
}
] |
train | 1242_JTEr1piJRN8_00001742 | 1242_JTEr1piJRN8_00001742.jpg | {
"data_source": "VIPSeg",
"file_name": "1242_JTEr1piJRN8_00001742.jpg",
"height": 720,
"id": "1242_JTEr1piJRN8_00001742",
"width": 1280
} | {
"caption": "A young girl with long dark hair wearing a pink patterned swimsuit lies on a large, black and white inflatable orca toy. They are floating in the bright blue water of an above-ground swimming pool with a blue liner. In the background, there is a plain tan wall.",
"caption_ann": "A <4:young girl with long dark hair wearing a pink patterned swimsuit> lies on a <2:large, black and white inflatable orca toy>. They are floating in the <1:bright blue water> of an <3:above-ground swimming pool with a blue liner>. In the background, there is a plain <0:tan wall>.",
"id": 1366,
"image_id": "1242_JTEr1piJRN8_00001742",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "young girl with long dark hair wearing a pink patterned swimsuit"
},
{
"mask_ids": [
2
],
"txt_desc": "large, black and white inflatable orca toy"
},
{
"mask_ids": [
1
],
"txt_desc": "bright blue water"
},
{
"mask_ids": [
3
],
"txt_desc": "above-ground swimming pool with a blue liner"
},
{
"mask_ids": [
0
],
"txt_desc": "tan wall"
}
],
"labels": [
"wall",
"water",
"toy",
"bathtub",
"person"
]
} | [
{
"area": 307697,
"bbox": [
0,
0,
1280,
278
],
"category_id": 0,
"id": 15475,
"image_id": "1242_JTEr1piJRN8_00001742",
"iscrowd": 0,
"segmentation": {
"counts": "0o7a>1O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1243_0w6XTqaZ6Ps_00002065 | 1243_0w6XTqaZ6Ps_00002065.jpg | {
"data_source": "VIPSeg",
"file_name": "1243_0w6XTqaZ6Ps_00002065.jpg",
"height": 720,
"id": "1243_0w6XTqaZ6Ps_00002065",
"width": 1280
} | {
"caption": "A man wearing a red long-sleeved shirt and jeans stands on a balcony while opening a set of large, dark, louvred folding doors. The balcony has a glass railing and overlooks another modern apartment building under the bright sky. Inside the room, a black LG box and a large item wrapped in plastic are visible next to a mirror.",
"caption_ann": "A <5:man wearing a red long-sleeved shirt and jeans> stands on a <1:balcony> while opening a set of <4:large, dark, louvred folding doors>. The <1:balcony> has a <0:glass railing> and overlooks another <2:modern apartment building> under the <3:bright sky>. Inside the room, a <6:black LG box> and a <7:large item wrapped in plastic> are visible next to a <8:mirror>.",
"id": 1367,
"image_id": "1243_0w6XTqaZ6Ps_00002065",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "man wearing a red long-sleeved shirt and jeans"
},
{
"mask_ids": [
1
],
"txt_desc": "balcony"
},
{
"mask_ids": [
4
],
"txt_desc": "large, dark, louvred folding doors"
},
{
"mask_ids": [
1
],
"txt_desc": "balcony"
},
{
"mask_ids": [
0
],
"txt_desc": "glass railing"
},
{
"mask_ids": [
2
],
"txt_desc": "modern apartment building"
},
{
"mask_ids": [
3
],
"txt_desc": "bright sky"
},
{
"mask_ids": [
6
],
"txt_desc": "black LG box"
},
{
"mask_ids": [
7
],
"txt_desc": "large item wrapped in plastic"
},
{
"mask_ids": [
8
],
"txt_desc": "mirror"
}
],
"labels": [
"handrail_or_fence",
"floor",
"building",
"sky",
"door",
"person",
"box",
"bag_or_package",
"mirror"
]
} | [
{
"area": 25840,
"bbox": [
603,
405,
677,
315
],
"category_id": 7,
"id": 15480,
"image_id": "1243_0w6XTqaZ6Ps_00002065",
"iscrowd": 0,
"segmentation": {
"counts": "U\\X=>Sf0N10000000000000000000000000kAHR68nIOl50TJ?]5AcJc0Y5]OhJb0X5^OhJb0X5^OiJb0V5EcJ;]5... | [
{
"id": 1,
"name": "object"
}
] |
train | 1245_jyBWK57N950_00002440 | 1245_jyBWK57N950_00002440.jpg | {
"data_source": "VIPSeg",
"file_name": "1245_jyBWK57N950_00002440.jpg",
"height": 720,
"id": "1245_jyBWK57N950_00002440",
"width": 1280
} | {
"caption": "From a bedroom, a woman in a white dress is seen through a large glass sliding door. She is on a cluttered balcony with a railing, surrounded by pieces of wood and a black plant stand. The balcony is filled with empty pots, including a terracotta pot, a brown pot, a green pot, a dark pot, and a red rectangular planter. In the foreground is a bed with a pink comforter. A tall green room plant stands next to a wall with a brown pattern, and a wire photo grid hangs on the right. The sky outside is very bright.",
"caption_ann": "From a bedroom, a <7:woman in a white dress> is seen through a <6:large glass sliding door>. She is on a <1:cluttered balcony with a railing>, surrounded by <3:pieces of wood> and a <5:black plant stand>. The balcony is filled with empty pots, including a <10:terracotta pot>, a <11:brown pot>, a <12:green pot>, a <13:dark pot>, and a <14:red rectangular planter>. In the foreground is a <8:bed with a pink comforter>. A <4:tall green room plant> stands next to a <0:wall with a brown pattern>, and a <9:wire photo grid> hangs on the right. The <2:sky> outside is very bright.",
"id": 1368,
"image_id": "1245_jyBWK57N950_00002440",
"label_matched": [
{
"mask_ids": [
7
],
"txt_desc": "woman in a white dress"
},
{
"mask_ids": [
6
],
"txt_desc": "large glass sliding door"
},
{
"mask_ids": [
1
],
"txt_desc": "cluttered balcony with a railing"
},
{
"mask_ids": [
3
],
"txt_desc": "pieces of wood"
},
{
"mask_ids": [
5
],
"txt_desc": "black plant stand"
},
{
"mask_ids": [
10
],
"txt_desc": "terracotta pot"
},
{
"mask_ids": [
11
],
"txt_desc": "brown pot"
},
{
"mask_ids": [
12
],
"txt_desc": "green pot"
},
{
"mask_ids": [
13
],
"txt_desc": "dark pot"
},
{
"mask_ids": [
14
],
"txt_desc": "red rectangular planter"
},
{
"mask_ids": [
8
],
"txt_desc": "bed with a pink comforter"
},
{
"mask_ids": [
4
],
"txt_desc": "tall green room plant"
},
{
"mask_ids": [
0
],
"txt_desc": "wall with a brown pattern"
},
{
"mask_ids": [
9
],
"txt_desc": "wire photo grid"
},
{
"mask_ids": [
2
],
"txt_desc": "sky"
}
],
"labels": [
"wall",
"handrail_or_fence",
"sky",
"wood",
"other_plant",
"shelf",
"door",
"person",
"bed",
"painting_or_poster",
"flower_pot_or_vase",
"flower_pot_or_vase",
"flower_pot_or_vase",
"flower_pot_or_vase",
"flower_pot_or_vase"
]
} | [
{
"area": 278245,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 15489,
"image_id": "1245_jyBWK57N950_00002440",
"iscrowd": 0,
"segmentation": {
"counts": "0mb0o0b_OVO`=g3QBiLo=W3iAQMV:TORHk3bMWM[:oNSHj3bMWM[:oNTHg2QLWN_14[:oNUHe2RLXN]15[:oNWH`2TL\... | [
{
"id": 1,
"name": "object"
}
] |
train | 1248_cbVuF98g-dE_00000077 | 1248_cbVuF98g-dE_00000077.jpg | {
"data_source": "VIPSeg",
"file_name": "1248_cbVuF98g-dE_00000077.jpg",
"height": 720,
"id": "1248_cbVuF98g-dE_00000077",
"width": 1280
} | {
"caption": "A person's hand holds a black toilet auger over the bowl of a ceramic toilet with the seat up. The bathroom has a white tiled floor and a yellow coloured wall. A small piece of white tissue roll is visible on the top right corner of the image.",
"caption_ann": "A <4:person's hand> holds a <2:black toilet auger> over the bowl of a <5:ceramic toilet with the seat up>. The bathroom has a <1:white tiled floor> and a <0:yellow coloured wall>. A small <3:piece of white tissue roll> is visible on the top right corner of the image.",
"id": 1369,
"image_id": "1248_cbVuF98g-dE_00000077",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "person's hand"
},
{
"mask_ids": [
2
],
"txt_desc": "black toilet auger"
},
{
"mask_ids": [
5
],
"txt_desc": "ceramic toilet with the seat up"
},
{
"mask_ids": [
1
],
"txt_desc": "white tiled floor"
},
{
"mask_ids": [
0
],
"txt_desc": "yellow coloured wall"
},
{
"mask_ids": [
3
],
"txt_desc": "piece of white tissue roll"
}
],
"labels": [
"wall",
"floor",
"tool",
"tissue",
"person",
"commode"
]
} | [
{
"area": 240262,
"bbox": [
0,
0,
1280,
514
],
"category_id": 0,
"id": 15504,
"image_id": "1248_cbVuF98g-dE_00000077",
"iscrowd": 0,
"segmentation": {
"counts": "0S9]=0000000000000000000000000000000000000000000000000000O100000000000000000000000000O100000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1249_lGoUlhtY3lM_00000032 | 1249_lGoUlhtY3lM_00000032.jpg | {
"data_source": "VIPSeg",
"file_name": "1249_lGoUlhtY3lM_00000032.jpg",
"height": 720,
"id": "1249_lGoUlhtY3lM_00000032",
"width": 1280
} | {
"caption": "This image shows two green Gumby toy figures placed on the back of a white toilet with its seat and lid raised. The toilet is in a bathroom with tan walls and a light-coloured marble tiled floor. A white trash can is visible in the top right corner of the image.",
"caption_ann": "This image shows <3:two green Gumby toy figures> placed on the back of a <4:white toilet with its seat and lid raised>. The toilet is in a bathroom with <0:tan walls> and a <2:light-coloured marble tiled floor>. A <1:white trash can> is visible in the top right corner of the image.",
"id": 1370,
"image_id": "1249_lGoUlhtY3lM_00000032",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "two green Gumby toy figures"
},
{
"mask_ids": [
4
],
"txt_desc": "white toilet with its seat and lid raised"
},
{
"mask_ids": [
0
],
"txt_desc": "tan walls"
},
{
"mask_ids": [
2
],
"txt_desc": "light-coloured marble tiled floor"
},
{
"mask_ids": [
1
],
"txt_desc": "white trash can"
}
],
"labels": [
"wall",
"pillar",
"floor",
"toy",
"commode"
]
} | [
{
"area": 294349,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 15510,
"image_id": "1249_lGoUlhtY3lM_00000032",
"iscrowd": 0,
"segmentation": {
"counts": "0`f01O00000000000000000000000000000000000000000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 124_XdJDi0upTx4_00000145 | 124_XdJDi0upTx4_00000145.jpg | {
"data_source": "VIPSeg",
"file_name": "124_XdJDi0upTx4_00000145.jpg",
"height": 720,
"id": "124_XdJDi0upTx4_00000145",
"width": 1280
} | {
"caption": "A large cargo ship with \"GREAT REPUBLIC\" written on its side sails on a body of water under a cloudy sky. A large cargo ship passing a tall black lighthouse-like tower on a green grassy embankment with a short retaining wall, a grassy area and trees. In the distance, houses lining a sandy beach. A massive metal lift bridge spans the canal overhead. A small ship visible in the far distance and buildings on the opposite shore.",
"caption_ann": "A <11:large cargo ship> with \"GREAT REPUBLIC\" written on its side sails on a <9:body of water> under a <8:cloudy sky>. A <11:large cargo ship> passing a <7:tall black lighthouse-like tower> on a <3:green grassy embankment> with a <0:short retaining wall>, a <1:grassy area> and <10:trees>. In the distance, <5:houses> lining a <2:sandy beach>. A massive <6:metal lift bridge> spans the canal overhead. A <12:small ship> visible in the far distance and <4:buildings> on the opposite shore.",
"id": 1371,
"image_id": "124_XdJDi0upTx4_00000145",
"label_matched": [
{
"mask_ids": [
11
],
"txt_desc": "large cargo ship"
},
{
"mask_ids": [
9
],
"txt_desc": "body of water"
},
{
"mask_ids": [
8
],
"txt_desc": "cloudy sky"
},
{
"mask_ids": [
11
],
"txt_desc": "large cargo ship"
},
{
"mask_ids": [
7
],
"txt_desc": "tall black lighthouse-like tower"
},
{
"mask_ids": [
3
],
"txt_desc": "green grassy embankment"
},
{
"mask_ids": [
0
],
"txt_desc": "short retaining wall"
},
{
"mask_ids": [
1
],
"txt_desc": "grassy area"
},
{
"mask_ids": [
10
],
"txt_desc": "trees"
},
{
"mask_ids": [
5
],
"txt_desc": "houses"
},
{
"mask_ids": [
2
],
"txt_desc": "sandy beach"
},
{
"mask_ids": [
6
],
"txt_desc": "metal lift bridge"
},
{
"mask_ids": [
12
],
"txt_desc": "small ship"
},
{
"mask_ids": [
4
],
"txt_desc": "buildings"
}
],
"labels": [
"handrail_or_fence",
"grass",
"sand",
"path",
"building",
"house",
"bridge",
"tower",
"sky",
"sea",
"tree",
"ship_or_boat",
"ship_or_boat"
]
} | [
{
"area": 21743,
"bbox": [
0,
449,
1013,
183
],
"category_id": 7,
"id": 15515,
"image_id": "124_XdJDi0upTx4_00000145",
"iscrowd": 0,
"segmentation": {
"counts": "kb0m0ce0000000000O1000O1000000000O1000O100000000000O010000000000000O0100000000000000O0100000... | [
{
"id": 1,
"name": "object"
}
] |
train | 124_mrA1OawpeNk_00000662 | 124_mrA1OawpeNk_00000662.jpg | {
"data_source": "VIPSeg",
"file_name": "124_mrA1OawpeNk_00000662.jpg",
"height": 720,
"id": "124_mrA1OawpeNk_00000662",
"width": 1280
} | {
"caption": "Inside a laboratory fume hood, a person's hand, wearing blue gloves is stretched towards a black countertop where a corning hot plate sits, with a metal support structure and a white wall in the background.",
"caption_ann": "Inside a laboratory fume hood, a <3:person's hand, wearing blue gloves> is stretched towards a <4:black countertop> where a <2:corning hot plate> sits, with a <1:metal support structure> and a <0:white wall> in the background.",
"id": 1372,
"image_id": "124_mrA1OawpeNk_00000662",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "person's hand, wearing blue gloves"
},
{
"mask_ids": [
4
],
"txt_desc": "black countertop"
},
{
"mask_ids": [
2
],
"txt_desc": "corning hot plate"
},
{
"mask_ids": [
1
],
"txt_desc": "metal support structure"
},
{
"mask_ids": [
0
],
"txt_desc": "white wall"
}
],
"labels": [
"wall",
"pole",
"other_electronic_product",
"person",
"table_or_desk"
]
} | [
{
"area": 432983,
"bbox": [
0,
0,
1280,
495
],
"category_id": 0,
"id": 15528,
"image_id": "124_mrA1OawpeNk_00000662",
"iscrowd": 0,
"segmentation": {
"counts": "0i:g;001O1O001O1O00002N00001O1O001O001O001O1O001O1O001O001O001O1O001O1O001O001O001O1O001O1O00... | [
{
"id": 1,
"name": "object"
}
] |
train | 1251__MCY-QkVaZw_00000047 | 1251__MCY-QkVaZw_00000047.jpg | {
"data_source": "VIPSeg",
"file_name": "1251__MCY-QkVaZw_00000047.jpg",
"height": 720,
"id": "1251__MCY-QkVaZw_00000047",
"width": 1280
} | {
"caption": "A woman with dark hair wearing a green bikini stands in the shallow, clear ocean water while holding a large green towel up in the air. In the background, a white boat in the centre and a white motorboat in the top right corner are visible on the horizon under a clear blue sky.",
"caption_ann": "A <5:woman with dark hair wearing a green bikini> stands in the <1:shallow, clear ocean water> while holding a <2:large green towel> up in the air. In the background, a <3:white boat> in the centre and a <4:white motorboat> in the top right corner are visible on the horizon under a <0:clear blue sky>.",
"id": 1373,
"image_id": "1251__MCY-QkVaZw_00000047",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "woman with dark hair wearing a green bikini"
},
{
"mask_ids": [
1
],
"txt_desc": "shallow, clear ocean water"
},
{
"mask_ids": [
2
],
"txt_desc": "large green towel"
},
{
"mask_ids": [
3
],
"txt_desc": "white boat"
},
{
"mask_ids": [
4
],
"txt_desc": "white motorboat"
},
{
"mask_ids": [
0
],
"txt_desc": "clear blue sky"
}
],
"labels": [
"sky",
"sea",
"textiles",
"ship_or_boat",
"ship_or_boat",
"person"
]
} | [
{
"area": 55525,
"bbox": [
0,
0,
1280,
46
],
"category_id": 28,
"id": 15533,
"image_id": "1251__MCY-QkVaZw_00000047",
"iscrowd": 0,
"segmentation": {
"counts": "0^1Re0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1252_gbunBa7DHmo_00000167 | 1252_gbunBa7DHmo_00000167.jpg | {
"data_source": "VIPSeg",
"file_name": "1252_gbunBa7DHmo_00000167.jpg",
"height": 720,
"id": "1252_gbunBa7DHmo_00000167",
"width": 1280
} | {
"caption": "A woman with dark hair wearing a dark blue sweater looks at the camera while speaking. She is holding a white slatted basket that is filled with folded white towels. In the background, a white tiled wall with built-in shelves where a potted plant with green leaves in a woven pot is visible. To the left of the image, a blurry green houseplant is visible in front of a glass door.",
"caption_ann": "A <4:woman with dark hair wearing a dark blue sweater> looks at the camera while speaking. She is holding a <5:white slatted basket> that is filled with <2:folded white towels>. In the background, a <0:white tiled wall with built-in shelves> where a <6:potted plant with green leaves in a woven pot> is visible. To the left of the image, a <1:blurry green houseplant> is visible in front of a <3:glass door>.",
"id": 1374,
"image_id": "1252_gbunBa7DHmo_00000167",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "woman with dark hair wearing a dark blue sweater"
},
{
"mask_ids": [
5
],
"txt_desc": "white slatted basket"
},
{
"mask_ids": [
2
],
"txt_desc": "folded white towels"
},
{
"mask_ids": [
0
],
"txt_desc": "white tiled wall with built-in shelves"
},
{
"mask_ids": [
6
],
"txt_desc": "potted plant with green leaves in a woven pot"
},
{
"mask_ids": [
1
],
"txt_desc": "blurry green houseplant"
},
{
"mask_ids": [
3
],
"txt_desc": "glass door"
}
],
"labels": [
"wall",
"other_plant",
"textiles",
"door",
"person",
"basket",
"flower_pot_or_vase"
]
} | [
{
"area": 362194,
"bbox": [
0,
0,
1280,
719
],
"category_id": 0,
"id": 15539,
"image_id": "1252_gbunBa7DHmo_00000167",
"iscrowd": 0,
"segmentation": {
"counts": "0_f01g@b8@^G?c8A]G<g8CYG<i8CWG:l8FTG9n8FRG7R9HnF7S9ImF4W9KiF4Y9KgF2\\9NdF1^9NbFOb90^FOc91]FK... | [
{
"id": 1,
"name": "object"
}
] |
train | 1253_NokM1l-YyNg_00000497 | 1253_NokM1l-YyNg_00000497.jpg | {
"data_source": "VIPSeg",
"file_name": "1253_NokM1l-YyNg_00000497.jpg",
"height": 720,
"id": "1253_NokM1l-YyNg_00000497",
"width": 960
} | {
"caption": "In this image three electric fans are lined up on a dark-topped cabinet against a light-coloured wall. On the left is a fan with a blue base and blue blades. In the middle is a smaller fan with a red base and red blades. On the right is a white fan with a white base.",
"caption_ann": "In this image <2,3,4:three electric fans> are lined up on a <1:dark-topped cabinet> against a <0:light-coloured wall>. On the left is a <2:fan with a blue base and blue blades>. In the middle is a <3:smaller fan with a red base and red blades>. On the right is a <4:white fan with a white base>.",
"id": 1375,
"image_id": "1253_NokM1l-YyNg_00000497",
"label_matched": [
{
"mask_ids": [
2,
3,
4
],
"txt_desc": "three electric fans"
},
{
"mask_ids": [
1
],
"txt_desc": "dark-topped cabinet"
},
{
"mask_ids": [
0
],
"txt_desc": "light-coloured wall"
},
{
"mask_ids": [
2
],
"txt_desc": "fan with a blue base and blue blades"
},
{
"mask_ids": [
3
],
"txt_desc": "smaller fan with a red base and red blades"
},
{
"mask_ids": [
4
],
"txt_desc": "white fan with a white base"
}
],
"labels": [
"wall",
"table_or_desk",
"fan",
"fan",
"fan"
]
} | [
{
"area": 382000,
"bbox": [
0,
0,
960,
720
],
"category_id": 0,
"id": 15546,
"image_id": "1253_NokM1l-YyNg_00000497",
"iscrowd": 0,
"segmentation": {
"counts": "0]aP2:iTPN0Ng0ZOf0[O00e0YOg0[O0Of0YOg0[O0O<E000N2O100N2O100N20000O1N200O1N20000O100N2O100N2O1... | [
{
"id": 1,
"name": "object"
}
] |
train | 1255_megGoHWei58_00000152 | 1255_megGoHWei58_00000152.jpg | {
"data_source": "VIPSeg",
"file_name": "1255_megGoHWei58_00000152.jpg",
"height": 720,
"id": "1255_megGoHWei58_00000152",
"width": 1280
} | {
"caption": "A black wall-mounted fan with a metal cage is spinning. It is fixed on a white wall near the white ceiling. To the left of the fan, a round white analogue clock is also hanging on the wall.",
"caption_ann": "A <2:black wall-mounted fan with a metal cage> is spinning. It is fixed on a <0:white wall> near the <1:white ceiling>. To the left of the fan, a <3:round white analogue clock> is also hanging on the wall.",
"id": 1376,
"image_id": "1255_megGoHWei58_00000152",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "black wall-mounted fan with a metal cage"
},
{
"mask_ids": [
0
],
"txt_desc": "white wall"
},
{
"mask_ids": [
1
],
"txt_desc": "white ceiling"
},
{
"mask_ids": [
3
],
"txt_desc": "round white analogue clock"
}
],
"labels": [
"wall",
"ceiling",
"fan",
"clock"
]
} | [
{
"area": 419357,
"bbox": [
0,
81,
1280,
639
],
"category_id": 0,
"id": 15551,
"image_id": "1255_megGoHWei58_00000152",
"iscrowd": 0,
"segmentation": {
"counts": "a2o6V4k1[NUNl5dMTJ[2m5eMSJ[2n5dMSJZ2n5fMRJY2o5gMQJY2P6fMPJY2Q6gMPJX2Q6gMoIX2R6hMnIX2S6gMnIW... | [
{
"id": 1,
"name": "object"
}
] |
train | 1256_Y_19xu4yTms_00000302 | 1256_Y_19xu4yTms_00000302.jpg | {
"data_source": "VIPSeg",
"file_name": "1256_Y_19xu4yTms_00000302.jpg",
"height": 720,
"id": "1256_Y_19xu4yTms_00000302",
"width": 1280
} | {
"caption": "A group of six people are riding in a rigid inflatable boat on the sea. The woman in the foreground wears a grey jacket and a striped hat while sitting next to a man in a green beanie. In the middle, a woman with a long ponytail stands near a person in a black hooded jacket and a person holding a camera. On the right, a man with short hair and wearing a black jacket is standing.",
"caption_ann": "A group of <2,3,4,5,6,7:six people> are riding in a <1:rigid inflatable boat> on the <0:sea>. The <2:woman in the foreground wears a grey jacket and a striped hat> while sitting next to a <6:man in a green beanie>. In the middle, a <3:woman with a long ponytail> stands near a <4:person in a black hooded jacket> and a <5:person holding a camera>. On the right, a <7:man with short hair and wearing a black jacket> is standing.",
"id": 1377,
"image_id": "1256_Y_19xu4yTms_00000302",
"label_matched": [
{
"mask_ids": [
2,
3,
4,
5,
6,
7
],
"txt_desc": "six people"
},
{
"mask_ids": [
1
],
"txt_desc": "rigid inflatable boat"
},
{
"mask_ids": [
0
],
"txt_desc": "sea"
},
{
"mask_ids": [
2
],
"txt_desc": "woman in the foreground wears a grey jacket and a striped hat"
},
{
"mask_ids": [
6
],
"txt_desc": "man in a green beanie"
},
{
"mask_ids": [
3
],
"txt_desc": "woman with a long ponytail"
},
{
"mask_ids": [
4
],
"txt_desc": "person in a black hooded jacket"
},
{
"mask_ids": [
5
],
"txt_desc": "person holding a camera"
},
{
"mask_ids": [
7
],
"txt_desc": "man with short hair and wearing a black jacket"
}
],
"labels": [
"sea",
"ship_or_boat",
"person",
"person",
"person",
"person",
"person",
"person"
]
} | [
{
"area": 424038,
"bbox": [
0,
1,
1280,
719
],
"category_id": 35,
"id": 15555,
"image_id": "1256_Y_19xu4yTms_00000302",
"iscrowd": 0,
"segmentation": {
"counts": "1_f0100000000000000000000000000000000000000000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1258_n1qRa2K8i3s_00000077 | 1258_n1qRa2K8i3s_00000077.jpg | {
"data_source": "VIPSeg",
"file_name": "1258_n1qRa2K8i3s_00000077.jpg",
"height": 720,
"id": "1258_n1qRa2K8i3s_00000077",
"width": 1280
} | {
"caption": "A bamboo log raft floating on a calm river. Ahead, a wooden weir spans the water, with green trees lining the banks. In the background, a range of hazy, pointed mountains is visible under a pale, overcast sky.",
"caption_ann": "A <5:bamboo log raft> floating on a <3:calm river>. Ahead, a <0:wooden weir> spans the water, with <4:green trees> lining the banks. In the background, a range of <2:hazy, pointed mountains> is visible under a <1:pale, overcast sky>.",
"id": 1378,
"image_id": "1258_n1qRa2K8i3s_00000077",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "bamboo log raft"
},
{
"mask_ids": [
3
],
"txt_desc": "calm river"
},
{
"mask_ids": [
0
],
"txt_desc": "wooden weir"
},
{
"mask_ids": [
4
],
"txt_desc": "green trees"
},
{
"mask_ids": [
2
],
"txt_desc": "hazy, pointed mountains"
},
{
"mask_ids": [
1
],
"txt_desc": "pale, overcast sky"
}
],
"labels": [
"bridge",
"sky",
"mountain",
"lake",
"tree",
"raft"
]
} | [
{
"area": 11556,
"bbox": [
592,
224,
534,
83
],
"category_id": 23,
"id": 15563,
"image_id": "1258_n1qRa2K8i3s_00000077",
"iscrowd": 0,
"segmentation": {
"counts": "Z`P=:Tf0f0ZO<D2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2N2O101OSOm000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1259_nY3cWyZ3Z3U_00000092 | 1259_nY3cWyZ3Z3U_00000092.jpg | {
"data_source": "VIPSeg",
"file_name": "1259_nY3cWyZ3Z3U_00000092.jpg",
"height": 720,
"id": "1259_nY3cWyZ3Z3U_00000092",
"width": 1280
} | {
"caption": "This image shows two men are on a small raft in the middle of a lake. Nearby, a larger floating platform with a small hut and another long raft are visible. In the background, a large mountain rises above a shoreline of trees under a pale, cloudy sky.",
"caption_ann": "This image shows <7,8:two men> are on a small <6:raft> in the middle of a <2:lake>. Nearby, a larger <5:floating platform with a small hut> and another long <4:raft> are visible. In the background, a large <1:mountain> rises above a shoreline of <3:trees> under a <0:pale, cloudy sky>.",
"id": 1379,
"image_id": "1259_nY3cWyZ3Z3U_00000092",
"label_matched": [
{
"mask_ids": [
7,
8
],
"txt_desc": "two men"
},
{
"mask_ids": [
6
],
"txt_desc": "raft"
},
{
"mask_ids": [
2
],
"txt_desc": "lake"
},
{
"mask_ids": [
5
],
"txt_desc": "floating platform with a small hut"
},
{
"mask_ids": [
4
],
"txt_desc": "raft"
},
{
"mask_ids": [
1
],
"txt_desc": "mountain"
},
{
"mask_ids": [
3
],
"txt_desc": "trees"
},
{
"mask_ids": [
0
],
"txt_desc": "pale, cloudy sky"
}
],
"labels": [
"sky",
"mountain",
"lake",
"tree",
"raft",
"raft",
"raft",
"person",
"person"
]
} | [
{
"area": 146781,
"bbox": [
0,
0,
1280,
187
],
"category_id": 28,
"id": 15569,
"image_id": "1259_nY3cWyZ3Z3U_00000092",
"iscrowd": 0,
"segmentation": {
"counts": "0]4Sb01O0000000000000000000000001O000000000000000000000000001O0000000000000000000000001O000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1260_rK2vcm94i_Q_00000218 | 1260_rK2vcm94i_Q_00000218.jpg | {
"data_source": "VIPSeg",
"file_name": "1260_rK2vcm94i_Q_00000218.jpg",
"height": 720,
"id": "1260_rK2vcm94i_Q_00000218",
"width": 1280
} | {
"caption": "An outdoor image of a forest shows three people floating down a calm river in a grey inflatable raft. A man in the middle sits and rows, while the woman at the back and the man at the front are standing. The riverbanks are lined with green shrubs and a dense forest of trees under a pale blue sky. A piece of driftwood is visible in the right side of the image.",
"caption_ann": "An outdoor image of a forest shows <6,7,8:three people> floating down a <3:calm river> in a <5:grey inflatable raft>. A <7:man in the middle> sits and rows, while the <6:woman at the back> and the <8:man at the front> are standing. The riverbanks are lined with <0:green shrubs> and a dense forest of <4:trees> under a <1:pale blue sky>. A <2:piece of driftwood> is visible in the right side of the image.",
"id": 1380,
"image_id": "1260_rK2vcm94i_Q_00000218",
"label_matched": [
{
"mask_ids": [
6,
7,
8
],
"txt_desc": "three people"
},
{
"mask_ids": [
3
],
"txt_desc": "calm river"
},
{
"mask_ids": [
5
],
"txt_desc": "grey inflatable raft"
},
{
"mask_ids": [
7
],
"txt_desc": "man in the middle"
},
{
"mask_ids": [
6
],
"txt_desc": "woman at the back"
},
{
"mask_ids": [
8
],
"txt_desc": "man at the front"
},
{
"mask_ids": [
0
],
"txt_desc": "green shrubs"
},
{
"mask_ids": [
4
],
"txt_desc": "trees"
},
{
"mask_ids": [
1
],
"txt_desc": "pale blue sky"
},
{
"mask_ids": [
2
],
"txt_desc": "piece of driftwood"
}
],
"labels": [
"grass",
"sky",
"wood",
"lake",
"tree",
"raft",
"person",
"person",
"person"
]
} | [
{
"area": 26027,
"bbox": [
0,
311,
919,
85
],
"category_id": 15,
"id": 15578,
"image_id": "1260_rK2vcm94i_Q_00000218",
"iscrowd": 0,
"segmentation": {
"counts": "Q;[1Ue0O101N1O1001N2N1O1O1O1000000O100O10001O00O10O1000000000000000N200O1O1O1O001O100O10000O... | [
{
"id": 1,
"name": "object"
}
] |
train | 1262_2gZoLJyRyJA_00000017 | 1262_2gZoLJyRyJA_00000017.jpg | {
"data_source": "VIPSeg",
"file_name": "1262_2gZoLJyRyJA_00000017.jpg",
"height": 720,
"id": "1262_2gZoLJyRyJA_00000017",
"width": 1280
} | {
"caption": "A scenic view of a person with their arms outstretched standing on a metal bridge next to the powerful waterfall. The waterfall cascades over a large moss-covered rock adjacent to lush green trees on the left side of the image.",
"caption_ann": "A scenic view of a <4:person with their arms outstretched> standing on a <0:metal bridge> next to the <2:powerful waterfall>. The <2:waterfall> cascades over a <1:large moss-covered rock> adjacent to <3:lush green trees> on the left side of the image.",
"id": 1381,
"image_id": "1262_2gZoLJyRyJA_00000017",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "person with their arms outstretched"
},
{
"mask_ids": [
0
],
"txt_desc": "metal bridge"
},
{
"mask_ids": [
2
],
"txt_desc": "powerful waterfall"
},
{
"mask_ids": [
2
],
"txt_desc": "waterfall"
},
{
"mask_ids": [
1
],
"txt_desc": "large moss-covered rock"
},
{
"mask_ids": [
3
],
"txt_desc": "lush green trees"
}
],
"labels": [
"bridge",
"stone",
"waterfall",
"tree",
"person"
]
} | [
{
"area": 140216,
"bbox": [
0,
402,
1280,
318
],
"category_id": 23,
"id": 15587,
"image_id": "1262_2gZoLJyRyJA_00000017",
"iscrowd": 0,
"segmentation": {
"counts": "b<Z2Wd0O00000000000000000000000001O0000000000000001O000000000001O000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1274_g1o31fsZxCc_00000100 | 1274_g1o31fsZxCc_00000100.jpg | {
"data_source": "VIPSeg",
"file_name": "1274_g1o31fsZxCc_00000100.jpg",
"height": 720,
"id": "1274_g1o31fsZxCc_00000100",
"width": 1286
} | {
"caption": "The image shows two pairs of four dancers on a large dance floor with a wall behind. In the centre of the image, a man in a white shirt and black tuxedo jacket and a woman in a shimmering white gown with a white float perform a ballroom dance. To the right side of the image, a woman in a black and cream gown with a black float and a man whose head is partially visible dance together.",
"caption_ann": "The image shows two pairs of <2,3,4,5:four dancers> on a <1:large dance floor> with a <0:wall> behind. In the centre of the image, a <2:man in a white shirt and black tuxedo jacket> and a <3:woman in a shimmering white gown with a white float> perform a ballroom dance. To the right side of the image, a <4:woman in a black and cream gown with a black float> and a <5:man> whose head is partially visible dance together.",
"id": 1382,
"image_id": "1274_g1o31fsZxCc_00000100",
"label_matched": [
{
"mask_ids": [
2,
3,
4,
5
],
"txt_desc": "four dancers"
},
{
"mask_ids": [
1
],
"txt_desc": "large dance floor"
},
{
"mask_ids": [
0
],
"txt_desc": "wall"
},
{
"mask_ids": [
2
],
"txt_desc": "man in a white shirt and black tuxedo jacket"
},
{
"mask_ids": [
3
],
"txt_desc": "woman in a shimmering white gown with a white float"
},
{
"mask_ids": [
4
],
"txt_desc": "woman in a black and cream gown with a black float"
},
{
"mask_ids": [
5
],
"txt_desc": "man"
}
],
"labels": [
"wall",
"floor",
"person",
"person",
"person",
"person"
]
} | [
{
"area": 40854,
"bbox": [
0,
0,
988,
77
],
"category_id": 0,
"id": 15592,
"image_id": "1274_g1o31fsZxCc_00000100",
"iscrowd": 0,
"segmentation": {
"counts": "0]2Sd000N20000000000000000000000000000000000000000O1000000000000000000000000000000000000000000N... | [
{
"id": 1,
"name": "object"
}
] |
train | 1278_NyRFJsHs_lc_00001000 | 1278_NyRFJsHs_lc_00001000.jpg | {
"data_source": "VIPSeg",
"file_name": "1278_NyRFJsHs_lc_00001000.jpg",
"height": 720,
"id": "1278_NyRFJsHs_lc_00001000",
"width": 1280
} | {
"caption": "A smiling woman with dark hair in an updo, wearing a white robe with lace sleeves and a silver tiara stands beside an ornate metal handrail adorned with white flowers. In the background, a landscape painting hangs on the beige wall next to a green curtain.",
"caption_ann": "A <3:smiling woman with dark hair in an updo, wearing a white robe with lace sleeves and a silver tiara> stands beside an <1:ornate metal handrail adorned with white flowers>. In the background, a <4:landscape painting> hangs on the <0:beige wall> next to a <2:green curtain>.",
"id": 1383,
"image_id": "1278_NyRFJsHs_lc_00001000",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "smiling woman with dark hair in an updo, wearing a white robe with lace sleeves and a silver tiara"
},
{
"mask_ids": [
1
],
"txt_desc": "ornate metal handrail adorned with white flowers"
},
{
"mask_ids": [
4
],
"txt_desc": "landscape painting"
},
{
"mask_ids": [
0
],
"txt_desc": "beige wall"
},
{
"mask_ids": [
2
],
"txt_desc": "green curtain"
}
],
"labels": [
"wall",
"handrail_or_fence",
"curtain",
"person",
"painting_or_poster"
]
} | [
{
"area": 465458,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 15598,
"image_id": "1278_NyRFJsHs_lc_00001000",
"iscrowd": 0,
"segmentation": {
"counts": "0aYY2d0mYgM^2hNX1UNk1000000000000000000000000000001O00000000000000001O0000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1279_CFPRtlTbIXU_00000452 | 1279_CFPRtlTbIXU_00000452.jpg | {
"data_source": "VIPSeg",
"file_name": "1279_CFPRtlTbIXU_00000452.jpg",
"height": 720,
"id": "1279_CFPRtlTbIXU_00000452",
"width": 1280
} | {
"caption": "An outdoor setting of a lawn shows a cat with long, dark fur trapped inside a metal cage on the grass. A small white bowl is visible on the right side of the cat inside a cage. In the background, there is a metal chain-link fence attached to a metal pole, and dirt ground is visible next to the paved path.",
"caption_ann": "An outdoor setting of a lawn shows a <6:cat with long, dark fur> trapped inside a <5:metal cage> on the <3:grass>. A <7:small white bowl> is visible on the right side of the <6:cat> inside a <5:cage>. In the background, there is a <0:metal chain-link fence> attached to a <1:metal pole>, and <2:dirt ground> is visible next to the <4:paved path>.",
"id": 1384,
"image_id": "1279_CFPRtlTbIXU_00000452",
"label_matched": [
{
"mask_ids": [
6
],
"txt_desc": "cat with long, dark fur"
},
{
"mask_ids": [
5
],
"txt_desc": "metal cage"
},
{
"mask_ids": [
3
],
"txt_desc": "grass"
},
{
"mask_ids": [
7
],
"txt_desc": "small white bowl"
},
{
"mask_ids": [
6
],
"txt_desc": "cat"
},
{
"mask_ids": [
5
],
"txt_desc": "cage"
},
{
"mask_ids": [
0
],
"txt_desc": "metal chain-link fence"
},
{
"mask_ids": [
1
],
"txt_desc": "metal pole"
},
{
"mask_ids": [
2
],
"txt_desc": "dirt ground"
},
{
"mask_ids": [
4
],
"txt_desc": "paved path"
}
],
"labels": [
"handrail_or_fence",
"pole",
"ground",
"grass",
"path",
"cage",
"cat",
"tub_or_bowl_or_pot"
]
} | [
{
"area": 131557,
"bbox": [
371,
0,
909,
224
],
"category_id": 7,
"id": 15603,
"image_id": "1279_CFPRtlTbIXU_00000452",
"iscrowd": 0,
"segmentation": {
"counts": "YnT8l0`d0T1[Ne1000000000000000000000000000000000001O00000000000000000000000000000000001O000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1287_7k3YMJeriqo_00000017 | 1287_7k3YMJeriqo_00000017.jpg | {
"data_source": "VIPSeg",
"file_name": "1287_7k3YMJeriqo_00000017.jpg",
"height": 720,
"id": "1287_7k3YMJeriqo_00000017",
"width": 1280
} | {
"caption": "From a first-person perspective, a person's hand holds the reins of a dark brown horse as it navigates sandy terrain. Ahead, another light-colored horse carries a rider in a pink top and blue jeans across the sandy landscape. The area is dotted with green shrubs and small bushes, all under a clear blue sky.",
"caption_ann": "From a first-person perspective, a <3:person's hand> holds the reins of a <7:dark brown horse> as it navigates sandy terrain. Ahead, another <6:light-colored horse> carries a <4:rider in a pink top and blue jeans> across the <0:sandy landscape>. The area is dotted with <2:green shrubs> and <5:small bushes>, all under a <1:clear blue sky>.",
"id": 1385,
"image_id": "1287_7k3YMJeriqo_00000017",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "person's hand"
},
{
"mask_ids": [
7
],
"txt_desc": "dark brown horse"
},
{
"mask_ids": [
6
],
"txt_desc": "light-colored horse"
},
{
"mask_ids": [
4
],
"txt_desc": "rider in a pink top and blue jeans"
},
{
"mask_ids": [
0
],
"txt_desc": "sandy landscape"
},
{
"mask_ids": [
2
],
"txt_desc": "green shrubs"
},
{
"mask_ids": [
5
],
"txt_desc": "small bushes"
},
{
"mask_ids": [
1
],
"txt_desc": "clear blue sky"
}
],
"labels": [
"sand",
"sky",
"other_plant",
"person",
"person",
"person",
"horse",
"horse"
]
} | [
{
"area": 259861,
"bbox": [
204,
195,
1076,
525
],
"category_id": 16,
"id": 15611,
"image_id": "1287_7k3YMJeriqo_00000017",
"iscrowd": 0,
"segmentation": {
"counts": "_T`41_f00000000O10000O100000000O10000O100000000O10000O100000000O10000O100000000O10000O1... | [
{
"id": 1,
"name": "object"
}
] |
train | 1290_Dd-mg5yf8Sw_00000670 | 1290_Dd-mg5yf8Sw_00000670.jpg | {
"data_source": "VIPSeg",
"file_name": "1290_Dd-mg5yf8Sw_00000670.jpg",
"height": 720,
"id": "1290_Dd-mg5yf8Sw_00000670",
"width": 1280
} | {
"caption": "The image captures a man wearing a white t-shirt that has grey sleeves and a grey beanie in mid-air while dunking a basketball. He is holding the rim of the clear glass backboard, which is attached to a metal pole that stands next to a black metal wall. A lemon-colored ceiling with a bright tubelight is visible in the top right corner of the image.",
"caption_ann": "The image captures a <5:man wearing a white t-shirt that has grey sleeves and a grey beanie> in mid-air while dunking a <6:basketball>. He is holding the rim of the <4:clear glass backboard>, which is attached to a <2:metal pole> that stands next to a <0:black metal wall>. A <1:lemon-colored ceiling> with a <3:bright tubelight> is visible in the top right corner of the image.",
"id": 1386,
"image_id": "1290_Dd-mg5yf8Sw_00000670",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "man wearing a white t-shirt that has grey sleeves and a grey beanie"
},
{
"mask_ids": [
6
],
"txt_desc": "basketball"
},
{
"mask_ids": [
4
],
"txt_desc": "clear glass backboard"
},
{
"mask_ids": [
2
],
"txt_desc": "metal pole"
},
{
"mask_ids": [
0
],
"txt_desc": "black metal wall"
},
{
"mask_ids": [
1
],
"txt_desc": "lemon-colored ceiling"
},
{
"mask_ids": [
3
],
"txt_desc": "bright tubelight"
}
],
"labels": [
"wall",
"ceiling",
"pole",
"lamp",
"backboard",
"person",
"ball"
]
} | [
{
"area": 665060,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 15619,
"image_id": "1290_Dd-mg5yf8Sw_00000670",
"iscrowd": 0,
"segmentation": {
"counts": "0Xnc6c0dg\\I10000N20000O1O10000O1O100ZB:a4GZJ_1e5aN[J_1d5bN\\J^1d5bN\\J^1c5h700O100O1O10000O... | [
{
"id": 1,
"name": "object"
}
] |
train | 1291_XwQZzmd_qmg_00000032 | 1291_XwQZzmd_qmg_00000032.jpg | {
"data_source": "VIPSeg",
"file_name": "1291_XwQZzmd_qmg_00000032.jpg",
"height": 720,
"id": "1291_XwQZzmd_qmg_00000032",
"width": 1280
} | {
"caption": "From the first-person perspective atop a light-colored horse, and a partially visible head of a horse are visible in the foreground, a person wearing a blue helmet and an tan t-shirt follows two other riders. Ahead of them, a person in a dark blue shirt and a person in a light-colored shirt ride two tan horses with black manes and tails. The group is proceeding up a long, straight dirt path flanked by grassy verges and rows of tall pine trees under a hazy sky.",
"caption_ann": "From the first-person perspective atop a <10:light-colored horse>, and a <9:partially visible head of a horse> are visible in the foreground, a <6:person wearing a blue helmet and an tan t-shirt> follows <4,5:two other riders>. Ahead of them, a <5:person in a dark blue shirt> and a <4:person in a light-colored shirt> ride <7,8:two tan horses with black manes and tails>. The group is proceeding up a <1:long, straight dirt path> flanked by <0:grassy verges> and rows of <3:tall pine trees> under a <2:hazy sky>.",
"id": 1387,
"image_id": "1291_XwQZzmd_qmg_00000032",
"label_matched": [
{
"mask_ids": [
10
],
"txt_desc": "light-colored horse"
},
{
"mask_ids": [
9
],
"txt_desc": "partially visible head of a horse"
},
{
"mask_ids": [
6
],
"txt_desc": "person wearing a blue helmet and an tan t-shirt"
},
{
"mask_ids": [
4,
5
],
"txt_desc": "two other riders"
},
{
"mask_ids": [
5
],
"txt_desc": "person in a dark blue shirt"
},
{
"mask_ids": [
4
],
"txt_desc": "person in a light-colored shirt"
},
{
"mask_ids": [
7,
8
],
"txt_desc": "two tan horses with black manes and tails"
},
{
"mask_ids": [
1
],
"txt_desc": "long, straight dirt path"
},
{
"mask_ids": [
0
],
"txt_desc": "grassy verges"
},
{
"mask_ids": [
3
],
"txt_desc": "tall pine trees"
},
{
"mask_ids": [
2
],
"txt_desc": "hazy sky"
}
],
"labels": [
"grass",
"path",
"sky",
"tree",
"person",
"person",
"person",
"horse",
"horse",
"horse",
"horse"
]
} | [
{
"area": 192857,
"bbox": [
0,
194,
815,
526
],
"category_id": 15,
"id": 15626,
"image_id": "1291_XwQZzmd_qmg_00000032",
"iscrowd": 0,
"segmentation": {
"counts": "Rb0^4Qb0100N200O100O1O1O100O1O1O100O100N200O1O100O1O100N200O1O100O1O100N200O1O100O1O100O1O... | [
{
"id": 1,
"name": "object"
}
] |
train | 1292_J5b7h72Lb6Y_00002395 | 1292_J5b7h72Lb6Y_00002395.jpg | {
"data_source": "VIPSeg",
"file_name": "1292_J5b7h72Lb6Y_00002395.jpg",
"height": 720,
"id": "1292_J5b7h72Lb6Y_00002395",
"width": 1280
} | {
"caption": "In a spacious indoor court with dark grey walls and a tiled floor, a man in black t-shirt and shorts dribbles a white basketball towards a man in blue jersey. A basketball hoop on a black pole stands nearby. On the left, a man in a grey hoddie and another person in black pants sit on a white ornate fireplace. In the background two white double doors are visible.",
"caption_ann": "In a spacious indoor court with <0:dark grey walls> and a <2:tiled floor>, a <9:man in black t-shirt and shorts> dribbles a <11:white basketball> towards a <10:man in blue jersey>. A <3:basketball hoop> on a <1:black pole> stands nearby. On the left, a <7:man in a grey hoddie> and another <8:person in black pants> sit on a <6:white ornate fireplace>. In the background <4,5:two white double doors> are visible.",
"id": 1388,
"image_id": "1292_J5b7h72Lb6Y_00002395",
"label_matched": [
{
"mask_ids": [
0
],
"txt_desc": "dark grey walls"
},
{
"mask_ids": [
2
],
"txt_desc": "tiled floor"
},
{
"mask_ids": [
9
],
"txt_desc": "man in black t-shirt and shorts"
},
{
"mask_ids": [
11
],
"txt_desc": "white basketball"
},
{
"mask_ids": [
10
],
"txt_desc": "man in blue jersey"
},
{
"mask_ids": [
3
],
"txt_desc": "basketball hoop"
},
{
"mask_ids": [
1
],
"txt_desc": "black pole"
},
{
"mask_ids": [
7
],
"txt_desc": "man in a grey hoddie"
},
{
"mask_ids": [
8
],
"txt_desc": "person in black pants"
},
{
"mask_ids": [
6
],
"txt_desc": "white ornate fireplace"
},
{
"mask_ids": [
4,
5
],
"txt_desc": "two white double doors"
}
],
"labels": [
"wall",
"pole",
"floor",
"backboard",
"door",
"door",
"sculpture",
"person",
"person",
"person",
"person",
"ball"
]
} | [
{
"area": 385682,
"bbox": [
0,
0,
1280,
612
],
"category_id": 0,
"id": 15637,
"image_id": "1292_J5b7h72Lb6Y_00002395",
"iscrowd": 0,
"segmentation": {
"counts": "0e3kb01O1O2N1O1O1O2N1O1OO1L4O1M3N2M3N2M3O1O1O1O100N200N200O1O1O1O1O100O1O1O1O1O100N200O100O1... | [
{
"id": 1,
"name": "object"
}
] |
train | 1294_t6mCYJJq9tQ_00000167 | 1294_t6mCYJJq9tQ_00000167.jpg | {
"data_source": "VIPSeg",
"file_name": "1294_t6mCYJJq9tQ_00000167.jpg",
"height": 720,
"id": "1294_t6mCYJJq9tQ_00000167",
"width": 1280
} | {
"caption": "The image shows a man in a black t-shirt and dark pants riding a white horse on a ground of sand and dirt. A black fence is visible in the background, separating the riding area from a grassy area and a line of trees. A few patches of light blue sky can be seen through the trees.",
"caption_ann": "The image shows a <5:man in a black t-shirt and dark pants> riding a <6:white horse> on a <1:ground of sand and dirt>. A <0:black fence> is visible in the background, separating the <1:riding area> from a <2:grassy area> and a line of <4:trees>. A <3:few patches of light blue sky> can be seen through the <4:trees>.",
"id": 1389,
"image_id": "1294_t6mCYJJq9tQ_00000167",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "man in a black t-shirt and dark pants"
},
{
"mask_ids": [
6
],
"txt_desc": "white horse"
},
{
"mask_ids": [
1
],
"txt_desc": "ground of sand and dirt"
},
{
"mask_ids": [
0
],
"txt_desc": "black fence"
},
{
"mask_ids": [
1
],
"txt_desc": "riding area"
},
{
"mask_ids": [
2
],
"txt_desc": "grassy area"
},
{
"mask_ids": [
4
],
"txt_desc": "trees"
},
{
"mask_ids": [
3
],
"txt_desc": "few patches of light blue sky"
},
{
"mask_ids": [
4
],
"txt_desc": "trees"
}
],
"labels": [
"handrail_or_fence",
"ground",
"grass",
"sky",
"tree",
"person",
"horse"
]
} | [
{
"area": 60192,
"bbox": [
0,
411,
1280,
114
],
"category_id": 7,
"id": 15649,
"image_id": "1294_t6mCYJJq9tQ_00000167",
"iscrowd": 0,
"segmentation": {
"counts": "k<b1od0O00000000000000000000000000000000000000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1297_X7S2nETEm7I_00000617 | 1297_X7S2nETEm7I_00000617.jpg | {
"data_source": "VIPSeg",
"file_name": "1297_X7S2nETEm7I_00000617.jpg",
"height": 720,
"id": "1297_X7S2nETEm7I_00000617",
"width": 1280
} | {
"caption": "A man wearing a red hard hat, a bright orange safety vest, and blue jeans is working on a bitumen road with yellow dividing lines on a sunny day. He is kneeling next to a manhole cover and is using a yellow and grey tool. The scene is set under a blue sky with fluffy white clouds, and a white building on the grassy ground is visible on the left side of the image.",
"caption_ann": "A <6:man wearing a red hard hat, a bright orange safety vest, and blue jeans> is working on a <1:bitumen road with yellow dividing lines> on a sunny day. He is kneeling next to a <3:manhole cover> and is using a <5:yellow and grey tool>. The scene is set under a <4:blue sky with fluffy white clouds>, and a <2:white building> on the <0:grassy ground> is visible on the left side of the image.",
"id": 1390,
"image_id": "1297_X7S2nETEm7I_00000617",
"label_matched": [
{
"mask_ids": [
6
],
"txt_desc": "man wearing a red hard hat, a bright orange safety vest, and blue jeans"
},
{
"mask_ids": [
1
],
"txt_desc": "bitumen road with yellow dividing lines"
},
{
"mask_ids": [
3
],
"txt_desc": "manhole cover"
},
{
"mask_ids": [
5
],
"txt_desc": "yellow and grey tool"
},
{
"mask_ids": [
4
],
"txt_desc": "blue sky with fluffy white clouds"
},
{
"mask_ids": [
2
],
"txt_desc": "white building"
},
{
"mask_ids": [
0
],
"txt_desc": "grassy ground"
}
],
"labels": [
"grass",
"road",
"house",
"well_or_well_lid",
"sky",
"tool",
"person"
]
} | [
{
"area": 63507,
"bbox": [
0,
265,
1225,
174
],
"category_id": 15,
"id": 15656,
"image_id": "1297_X7S2nETEm7I_00000617",
"iscrowd": 0,
"segmentation": {
"counts": "Y:^3Rc0000000O100000000O10000000O100000O1000000O100000000000O1000O100000000O10000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 1298_2tsd2aEL5XA_00000475 | 1298_2tsd2aEL5XA_00000475.jpg | {
"data_source": "VIPSeg",
"file_name": "1298_2tsd2aEL5XA_00000475.jpg",
"height": 720,
"id": "1298_2tsd2aEL5XA_00000475",
"width": 1280
} | {
"caption": "An outdoor basketball court surrounded by a black fence features a concrete surface and a basketball backboard with a red rim and white net attached to a black pole. Near the pole, a guy in a light blue shirt is standing. To the right of the court, there is a tall light pole, a tan house with a tile roof and a wooden ladder leaning against it. On the bottom left corner of the image, two cardboard boxes are visible. In the background are large trees under a light blue sky.",
"caption_ann": "An outdoor basketball court surrounded by a <0:black fence> features a <2:concrete surface> and a <6:basketball backboard with a red rim and white net> attached to a black pole. Near the pole, a <8:guy in a light blue shirt> is standing. To the right of the court, there is a <1:tall light pole>, a <3:tan house with a tile roof> and a <7:wooden ladder> leaning against it. On the bottom left corner of the image, <9,10:two cardboard boxes> are visible. In the background are <5:large trees> under a <4:light blue sky>.",
"id": 1391,
"image_id": "1298_2tsd2aEL5XA_00000475",
"label_matched": [
{
"mask_ids": [
0
],
"txt_desc": "black fence"
},
{
"mask_ids": [
2
],
"txt_desc": "concrete surface"
},
{
"mask_ids": [
6
],
"txt_desc": "basketball backboard with a red rim and white net"
},
{
"mask_ids": [
8
],
"txt_desc": "guy in a light blue shirt"
},
{
"mask_ids": [
1
],
"txt_desc": "tall light pole"
},
{
"mask_ids": [
3
],
"txt_desc": "tan house with a tile roof"
},
{
"mask_ids": [
7
],
"txt_desc": "wooden ladder"
},
{
"mask_ids": [
9,
10
],
"txt_desc": "two cardboard boxes"
},
{
"mask_ids": [
5
],
"txt_desc": "large trees"
},
{
"mask_ids": [
4
],
"txt_desc": "light blue sky"
}
],
"labels": [
"handrail_or_fence",
"pole",
"ground",
"house",
"sky",
"tree",
"backboard",
"shelf",
"person",
"box",
"box"
]
} | [
{
"area": 206712,
"bbox": [
0,
0,
989,
567
],
"category_id": 7,
"id": 15663,
"image_id": "1298_2tsd2aEL5XA_00000475",
"iscrowd": 0,
"segmentation": {
"counts": "0ga0i4O1000000000000O10000000000O1000000000000O10000000000O1000000000000O10001O1O1O1N2O1O2N1O... | [
{
"id": 1,
"name": "object"
}
] |
train | 1298_d7DtRPgYwME_00000047 | 1298_d7DtRPgYwME_00000047.jpg | {
"data_source": "VIPSeg",
"file_name": "1298_d7DtRPgYwME_00000047.jpg",
"height": 720,
"id": "1298_d7DtRPgYwME_00000047",
"width": 1280
} | {
"caption": "A close-up shot shows a man in dark pants and brown boots kneeling on the dirt ground to work on an open well. The man is using his gloved hands to grip a yellow tool inside the well.",
"caption_ann": "A close-up shot shows a <3:man in dark pants and brown boots> kneeling on the <0:dirt ground> to work on an <1:open well>. The <3:man> is using his gloved hands to grip a <2:yellow tool> inside the <1:well>.",
"id": 1392,
"image_id": "1298_d7DtRPgYwME_00000047",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "man in dark pants and brown boots"
},
{
"mask_ids": [
0
],
"txt_desc": "dirt ground"
},
{
"mask_ids": [
1
],
"txt_desc": "open well"
},
{
"mask_ids": [
3
],
"txt_desc": "man"
},
{
"mask_ids": [
2
],
"txt_desc": "yellow tool"
},
{
"mask_ids": [
1
],
"txt_desc": "well"
}
],
"labels": [
"ground",
"well_or_well_lid",
"tool",
"person"
]
} | [
{
"area": 570930,
"bbox": [
0,
0,
1280,
720
],
"category_id": 14,
"id": 15674,
"image_id": "1298_d7DtRPgYwME_00000047",
"iscrowd": 0,
"segmentation": {
"counts": "0`ie06jlZO;E8H003M1O2N1O2N1O2N2N2N002N2N2N1O2N1O2N1O2N1O2N2N2N1O2N1O2N1O2N1O1O1O1O1O1O1O1O1... | [
{
"id": 1,
"name": "object"
}
] |
train | 1299_DSxbx2TLpcY_00007900 | 1299_DSxbx2TLpcY_00007900.jpg | {
"data_source": "VIPSeg",
"file_name": "1299_DSxbx2TLpcY_00007900.jpg",
"height": 720,
"id": "1299_DSxbx2TLpcY_00007900",
"width": 1280
} | {
"caption": "On an outdoor basketball court, two people are playing basketball. A young boy with blond hair, light-colored t-shirt and multicoloured shorts holds a basketball. Next to him, a man in a grey t-shirt and jeans stands near a black pole that holds up a basketball backboard with a net and rim. In the background is a line of pine trees, and a basketball rests on the grassy ground.",
"caption_ann": "On an outdoor <1:basketball court>, <5,6:two people> are playing basketball. A <5:young boy with blond hair, light-colored t-shirt and multicoloured shorts> holds a <8:basketball>. Next to him, a <6:man in a grey t-shirt and jeans> stands near a <0:black pole> that holds up a <4:basketball backboard with a net and rim>. In the background is a line of <3:pine trees>, and a <7:basketball> rests on the <2:grassy ground>.",
"id": 1393,
"image_id": "1299_DSxbx2TLpcY_00007900",
"label_matched": [
{
"mask_ids": [
1
],
"txt_desc": "basketball court"
},
{
"mask_ids": [
5,
6
],
"txt_desc": "two people"
},
{
"mask_ids": [
5
],
"txt_desc": "young boy with blond hair, light-colored t-shirt and multicoloured shorts"
},
{
"mask_ids": [
8
],
"txt_desc": "basketball"
},
{
"mask_ids": [
6
],
"txt_desc": "man in a grey t-shirt and jeans"
},
{
"mask_ids": [
0
],
"txt_desc": "black pole"
},
{
"mask_ids": [
4
],
"txt_desc": "basketball backboard with a net and rim"
},
{
"mask_ids": [
3
],
"txt_desc": "pine trees"
},
{
"mask_ids": [
7
],
"txt_desc": "basketball"
},
{
"mask_ids": [
2
],
"txt_desc": "grassy ground"
}
],
"labels": [
"pole",
"ground",
"grass",
"tree",
"backboard",
"person",
"person",
"ball",
"ball"
]
} | [
{
"area": 49772,
"bbox": [
911,
31,
140,
581
],
"category_id": 12,
"id": 15678,
"image_id": "1299_DSxbx2TLpcY_00007900",
"iscrowd": 0,
"segmentation": {
"counts": "PTQd0a0ne01M3O1L4O1M3O1M3N2N2N2N2N2M3O1M3O1M3N2M3O1M3O1[Me2dNhLc_OU3]`0hLe_OY3g:fL`I0jKZ3j... | [
{
"id": 1,
"name": "object"
}
] |
train | 1303_m-6Al08LvDM_00002305 | 1303_m-6Al08LvDM_00002305.jpg | {
"data_source": "VIPSeg",
"file_name": "1303_m-6Al08LvDM_00002305.jpg",
"height": 720,
"id": "1303_m-6Al08LvDM_00002305",
"width": 1280
} | {
"caption": "The image shows a group of six people playing basketball on an outdoor court with a multicolored geometric surface under a light grey sky. A bearded man in a white tank top and black Nike shorts is standing next to a guy in a black and grey tank top with red sneakers. A guy with tattoos on his chest and wearing black pants stands nearby. On the left, a guy in a white t-shirt and shorts and an guy in grey pants and multicolored shoes are standing near a black pole that has a basketball backboard with net attached to it. In the mid-left corner, a bare-chested guy in dark red shorts is looking up. The court is enclosed by a black fence, with a large tree and a house in the background.",
"caption_ann": "The image shows a group of <7,8,9,10,11,12:six people> playing basketball on an <2:outdoor court with a multicolored geometric surface> under a <4:light grey sky>. A <12:bearded man in a white tank top and black Nike shorts> is standing next to a <11:guy in a black and grey tank top with red sneakers>. A <10:guy with tattoos on his chest and wearing black pants> stands nearby. On the left, a <9:guy in a white t-shirt and shorts> and an <8:guy in grey pants and multicolored shoes> are standing near a <1:black pole> that has a <6:basketball backboard with net> attached to it. In the mid-left corner, a <7:bare-chested guy in dark red shorts> is looking up. The court is enclosed by a <0:black fence>, with a <5:large tree> and a <3:house> in the background.",
"id": 1394,
"image_id": "1303_m-6Al08LvDM_00002305",
"label_matched": [
{
"mask_ids": [
7,
8,
9,
10,
11,
12
],
"txt_desc": "six people"
},
{
"mask_ids": [
2
],
"txt_desc": "outdoor court with a multicolored geometric surface"
},
{
"mask_ids": [
4
],
"txt_desc": "light grey sky"
},
{
"mask_ids": [
12
],
"txt_desc": "bearded man in a white tank top and black Nike shorts"
},
{
"mask_ids": [
11
],
"txt_desc": "guy in a black and grey tank top with red sneakers"
},
{
"mask_ids": [
10
],
"txt_desc": "guy with tattoos on his chest and wearing black pants"
},
{
"mask_ids": [
9
],
"txt_desc": "guy in a white t-shirt and shorts"
},
{
"mask_ids": [
8
],
"txt_desc": "guy in grey pants and multicolored shoes"
},
{
"mask_ids": [
1
],
"txt_desc": "black pole"
},
{
"mask_ids": [
6
],
"txt_desc": "basketball backboard with net"
},
{
"mask_ids": [
7
],
"txt_desc": "bare-chested guy in dark red shorts"
},
{
"mask_ids": [
0
],
"txt_desc": "black fence"
},
{
"mask_ids": [
5
],
"txt_desc": "large tree"
},
{
"mask_ids": [
3
],
"txt_desc": "house"
}
],
"labels": [
"handrail_or_fence",
"pole",
"ground",
"house",
"sky",
"tree",
"backboard",
"person",
"person",
"person",
"person",
"person",
"person"
]
} | [
{
"area": 114488,
"bbox": [
0,
236,
1280,
211
],
"category_id": 7,
"id": 15687,
"image_id": "1303_m-6Al08LvDM_00002305",
"iscrowd": 0,
"segmentation": {
"counts": "a7k1cd0QOc[ONKWOad0f0_[O[O9g0Yd0Mn[OORd01P\\ONoc02T\\OKlc06U\\OHmc06U\\OBXOK00cd0d0U\\O]OX... | [
{
"id": 1,
"name": "object"
}
] |
train | 1305_qNxcPINHSC4_00002656 | 1305_qNxcPINHSC4_00002656.jpg | {
"data_source": "VIPSeg",
"file_name": "1305_qNxcPINHSC4_00002656.jpg",
"height": 720,
"id": "1305_qNxcPINHSC4_00002656",
"width": 1280
} | {
"caption": "Image displays an outdoor basketball court with markings, where three men are playing basketball in front of a chain-link fence and under a blue sky with white clouds. A bare-chested man in grey shorts and a blue t-shirt wrapped around his head facing away from the camera dribbles a basketball. Facing him towards the right of the bare-chested man, another man in a white t-shirt and grey headband is visible in a defensive stance, while a partially visible guy in a black t-shirt and shorts approaches the bare-chested man. A basketball backboard is attached to a black pole on the right. In the background, a building behind the fence along with some trees are visible.",
"caption_ann": "Image displays an outdoor <2:basketball court with markings>, where <7,8,9:three men> are playing basketball in front of a <0:chain-link fence> and under a <4:blue sky with white clouds>. A <7:bare-chested man in grey shorts and a blue t-shirt wrapped around his head> facing away from the camera dribbles a <10:basketball>. Facing him towards the right of the <7:bare-chested man>, another <9:man in a white t-shirt and grey headband> is visible in a defensive stance, while a partially visible <8:guy in a black t-shirt and shorts> approaches the <7:bare-chested man>. A <6:basketball backboard> is attached to a <1:black pole> on the right. In the background, a <3:building> behind the <0:fence> along with some <5:trees> are visible.",
"id": 1395,
"image_id": "1305_qNxcPINHSC4_00002656",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "basketball court with markings"
},
{
"mask_ids": [
7,
8,
9
],
"txt_desc": "three men"
},
{
"mask_ids": [
0
],
"txt_desc": "chain-link fence"
},
{
"mask_ids": [
4
],
"txt_desc": "blue sky with white clouds"
},
{
"mask_ids": [
7
],
"txt_desc": "bare-chested man in grey shorts and a blue t-shirt wrapped around his head"
},
{
"mask_ids": [
10
],
"txt_desc": "basketball"
},
{
"mask_ids": [
7
],
"txt_desc": "bare-chested man"
},
{
"mask_ids": [
9
],
"txt_desc": "man in a white t-shirt and grey headband"
},
{
"mask_ids": [
8
],
"txt_desc": "guy in a black t-shirt and shorts"
},
{
"mask_ids": [
7
],
"txt_desc": "bare-chested man"
},
{
"mask_ids": [
6
],
"txt_desc": "basketball backboard"
},
{
"mask_ids": [
1
],
"txt_desc": "black pole"
},
{
"mask_ids": [
3
],
"txt_desc": "building"
},
{
"mask_ids": [
0
],
"txt_desc": "fence"
},
{
"mask_ids": [
5
],
"txt_desc": "trees"
}
],
"labels": [
"handrail_or_fence",
"pole",
"ground",
"building",
"sky",
"tree",
"backboard",
"person",
"person",
"person",
"ball"
]
} | [
{
"area": 117693,
"bbox": [
0,
415,
1280,
216
],
"category_id": 7,
"id": 15700,
"image_id": "1305_qNxcPINHSC4_00002656",
"iscrowd": 0,
"segmentation": {
"counts": "P=e3kb00000000000LW]OaLib0_34O1O1O1O1O1O100N110N2N2L4M3M3N2L4N2L4N2M3O1L4N2M3O1M3O1MX]OXNn... | [
{
"id": 1,
"name": "object"
}
] |
train | 1307_SU4Ys9lCL5Q_00001540 | 1307_SU4Ys9lCL5Q_00001540.jpg | {
"data_source": "VIPSeg",
"file_name": "1307_SU4Ys9lCL5Q_00001540.jpg",
"height": 720,
"id": "1307_SU4Ys9lCL5Q_00001540",
"width": 1280
} | {
"caption": "A young girl wearing a white t-shirt and pink checkered pants is kneeling on a tiled floor. She is holding a light grey pastic fan, seemingly in the process of fixing it. An electronic extension board, a small monkey is standing and looking at her. In the background, a brown patterned cushion is visible leaning against a light green wall.",
"caption_ann": "A <4:young girl wearing a white t-shirt and pink checkered pants> is kneeling on a <1:tiled floor>. She is holding a <6:light grey pastic fan>, seemingly in the process of fixing it. An <3:electronic extension board>, a <5:small monkey> is standing and looking at her. In the background, a <2:brown patterned cushion> is visible leaning against a <0:light green wall>.",
"id": 1396,
"image_id": "1307_SU4Ys9lCL5Q_00001540",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "young girl wearing a white t-shirt and pink checkered pants"
},
{
"mask_ids": [
1
],
"txt_desc": "tiled floor"
},
{
"mask_ids": [
6
],
"txt_desc": "light grey pastic fan"
},
{
"mask_ids": [
3
],
"txt_desc": "electronic extension board"
},
{
"mask_ids": [
5
],
"txt_desc": "small monkey"
},
{
"mask_ids": [
2
],
"txt_desc": "brown patterned cushion"
},
{
"mask_ids": [
0
],
"txt_desc": "light green wall"
}
],
"labels": [
"wall",
"floor",
"cushion_or_carpet",
"other_electronic_product",
"person",
"other_animal",
"fan"
]
} | [
{
"area": 239872,
"bbox": [
574,
0,
706,
560
],
"category_id": 0,
"id": 15711,
"image_id": "1307_SU4Ys9lCL5Q_00001540",
"iscrowd": 0,
"segmentation": {
"counts": "_fc<6le0;M5C;M5C:N5C;M5C;M4G92N2N2O1O100N200N200O1O1O100O1O1O100N200N200O1O1O1O1N2O1O1N2O1O... | [
{
"id": 1,
"name": "object"
}
] |
train | 1311_Br7du9NYVEY_00000077 | 1311_Br7du9NYVEY_00000077.jpg | {
"data_source": "VIPSeg",
"file_name": "1311_Br7du9NYVEY_00000077.jpg",
"height": 720,
"id": "1311_Br7du9NYVEY_00000077",
"width": 1280
} | {
"caption": "At a wooden dining table, two people are eating a meal. A young boy with a black and white long-sleeved shirt eats chicken nuggets from his white plate with a fork. On the table are two drinking glasses, a bottle of ketchup, a small red toy car, and three other white plates and bowls. The boy is sitting on a wooden chair in a room with a beige wall and a large window with white blinds.",
"caption_ann": "At a <16:wooden dining table>, <7,8:two people> are eating a meal. A <8:young boy with a black and white long-sleeved shirt> eats <4:chicken nuggets> from his <11:white plate> with a <3:fork>. On the <16:table> are <13,15:two drinking glasses>, a <14:bottle of ketchup>, a <1:small red toy car>, and <9,10,12:three other white plates and bowls>. The <8:boy> is sitting on a <17:wooden chair> in a room with a <0:beige wall> and a <2,5,6:large window with white blinds>.",
"id": 1397,
"image_id": "1311_Br7du9NYVEY_00000077",
"label_matched": [
{
"mask_ids": [
16
],
"txt_desc": "wooden dining table"
},
{
"mask_ids": [
7,
8
],
"txt_desc": "two people"
},
{
"mask_ids": [
8
],
"txt_desc": "young boy with a black and white long-sleeved shirt"
},
{
"mask_ids": [
4
],
"txt_desc": "chicken nuggets"
},
{
"mask_ids": [
11
],
"txt_desc": "white plate"
},
{
"mask_ids": [
3
],
"txt_desc": "fork"
},
{
"mask_ids": [
16
],
"txt_desc": "table"
},
{
"mask_ids": [
13,
15
],
"txt_desc": "two drinking glasses"
},
{
"mask_ids": [
14
],
"txt_desc": "bottle of ketchup"
},
{
"mask_ids": [
1
],
"txt_desc": "small red toy car"
},
{
"mask_ids": [
9,
10,
12
],
"txt_desc": "three other white plates and bowls"
},
{
"mask_ids": [
8
],
"txt_desc": "boy"
},
{
"mask_ids": [
17
],
"txt_desc": "wooden chair"
},
{
"mask_ids": [
0
],
"txt_desc": "beige wall"
},
{
"mask_ids": [
2,
5,
6
],
"txt_desc": "large window with white blinds"
}
],
"labels": [
"wall",
"toy",
"curtain",
"tool",
"food",
"window",
"window",
"person",
"person",
"plate",
"plate",
"plate",
"plate",
"bottle_or_cup",
"bottle_or_cup",
"bottle_or_cup",
"table_or_desk",
"chair_or_seat"
]
} | [
{
"area": 73186,
"bbox": [
747,
0,
533,
438
],
"category_id": 0,
"id": 15718,
"image_id": "1311_Br7du9NYVEY_00000077",
"iscrowd": 0,
"segmentation": {
"counts": "kd]`04Pf0<XOi0C<XOh0D<XOh0G90000O10000000001O00000O100000000O1000000K5F:]Oc0F:\\Od0F:\\Od0F:... | [
{
"id": 1,
"name": "object"
}
] |
train | 1311_sFPdOYM8g4U_00004435 | 1311_sFPdOYM8g4U_00004435.jpg | {
"data_source": "VIPSeg",
"file_name": "1311_sFPdOYM8g4U_00004435.jpg",
"height": 720,
"id": "1311_sFPdOYM8g4U_00004435",
"width": 1280
} | {
"caption": "In an indoor setting, a man wearing a gray t-shirt, dark shorts, and black gloves is shown in the process of tiling a wall. He is seated on a surface covered with a drop cloth on the floor, holding a white tile with fresh mortar on the back. The wall behind him has a layer of grooved mortar applied, with a row of white tiles already installed at the bottom. To the right, a red pole, likely part of a ladder, and a wooden table are visible.Four bags or packages containing materials are scattered on the floor around him.",
"caption_ann": "In an indoor setting, a <3:man wearing a gray t-shirt, dark shorts, and black gloves> is shown in the process of tiling a <0:wall>. He is seated on a surface covered with a <4:drop cloth> on the <2:floor>, holding a white tile with fresh mortar on the back. The <0:wall> behind him has a layer of grooved mortar applied, with a row of white tiles already installed at the bottom. To the right, a <1:red pole>, likely part of a ladder, and a <9:wooden table> are visible.<5,6,7,8:Four bags or packages> containing materials are scattered on the <2:floor> around him.",
"id": 1398,
"image_id": "1311_sFPdOYM8g4U_00004435",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "man wearing a gray t-shirt, dark shorts, and black gloves"
},
{
"mask_ids": [
0
],
"txt_desc": "wall"
},
{
"mask_ids": [
4
],
"txt_desc": "drop cloth"
},
{
"mask_ids": [
2
],
"txt_desc": "floor"
},
{
"mask_ids": [
0
],
"txt_desc": "wall"
},
{
"mask_ids": [
1
],
"txt_desc": "red pole"
},
{
"mask_ids": [
9
],
"txt_desc": "wooden table"
},
{
"mask_ids": [
5,
6,
7,
8
],
"txt_desc": "Four bags or packages"
},
{
"mask_ids": [
2
],
"txt_desc": "floor"
}
],
"labels": [
"wall",
"pole",
"floor",
"person",
"bag_or_package",
"bag_or_package",
"bag_or_package",
"bag_or_package",
"bag_or_package",
"table_or_desk"
]
} | [
{
"area": 511359,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 15736,
"image_id": "1311_sFPdOYM8g4U_00004435",
"iscrowd": 0,
"segmentation": {
"counts": "0X;X;d:\\E00O100O100O100O1O10000O1O10000O1O10000O1O100O1O10000O1O10000O1O10000O1O100O100O100... | [
{
"id": 1,
"name": "object"
}
] |
train | 1312_-CF5nP71J_I_00002013 | 1312_-CF5nP71J_I_00002013.jpg | {
"data_source": "VIPSeg",
"file_name": "1312_-CF5nP71J_I_00002013.jpg",
"height": 720,
"id": "1312_-CF5nP71J_I_00002013",
"width": 1280
} | {
"caption": "On a baseball field, a man wearing a light blue t-shirt over a long-sleeved sweatshirt and dark pants stands on the dirt baseline while holding a long metal tool, appearing to groom the field. The field is composed of the dirt baseline and surrounding green grass. In the background, a tall chain-link fence encloses the area, with a wheeled field-marking machine resting on the grass nearby. Beyond the fence is a line of dense green trees. On the field, a black pole with a square base is placed into the ground, and a small cardboard box sits on the grass.",
"caption_ann": "On a baseball field, a <7:man wearing a light blue t-shirt over a long-sleeved sweatshirt and dark pants> stands on the <2:dirt baseline> while holding a <6:long metal tool>, appearing to groom the field. The field is composed of the <2:dirt baseline> and surrounding <3:green grass>. In the background, a <0:tall chain-link fence> encloses the area, with a <4:wheeled field-marking machine> resting on the <3:grass> nearby. Beyond the fence is a line of <5:dense green trees>. On the field, a <1:black pole with a square base> is placed into the ground, and a small <8:cardboard box> sits on the <3:grass>.",
"id": 1399,
"image_id": "1312_-CF5nP71J_I_00002013",
"label_matched": [
{
"mask_ids": [
7
],
"txt_desc": "man wearing a light blue t-shirt over a long-sleeved sweatshirt and dark pants"
},
{
"mask_ids": [
2
],
"txt_desc": "dirt baseline"
},
{
"mask_ids": [
6
],
"txt_desc": "long metal tool"
},
{
"mask_ids": [
2
],
"txt_desc": "dirt baseline"
},
{
"mask_ids": [
3
],
"txt_desc": "green grass"
},
{
"mask_ids": [
0
],
"txt_desc": "tall chain-link fence"
},
{
"mask_ids": [
4
],
"txt_desc": "wheeled field-marking machine"
},
{
"mask_ids": [
3
],
"txt_desc": "grass"
},
{
"mask_ids": [
5
],
"txt_desc": "dense green trees"
},
{
"mask_ids": [
1
],
"txt_desc": "black pole with a square base"
},
{
"mask_ids": [
8
],
"txt_desc": "cardboard box"
},
{
"mask_ids": [
3
],
"txt_desc": "grass"
}
],
"labels": [
"handrail_or_fence",
"pole",
"ground",
"grass",
"wheeled_machine",
"tree",
"tool",
"person",
"box"
]
} | [
{
"area": 182688,
"bbox": [
0,
0,
1280,
419
],
"category_id": 7,
"id": 15746,
"image_id": "1312_-CF5nP71J_I_00002013",
"iscrowd": 0,
"segmentation": {
"counts": "0U7=\\NP9e1QGXNP9j1oFUNQ9l1PGQNQ9Q2oFmMQ9U2nFjMR9W2oFfMR9\\2mFcMS9^2nF_MS9c2mF[MS9f2mFXMT9j2... | [
{
"id": 1,
"name": "object"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.