images images listlengths 1 1 | data_source stringclasses 1 value | reward_model dict | extra_info dict | prompt listlengths 2 2 | ability stringclasses 1 value | env_name stringclasses 1 value |
|---|---|---|---|---|---|---|
sealvqa_spatial | {
"ground_truth": "The remote is above the tv."
} | {
"answer": "The remote is above the tv.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000111922.jpg",
"index": "sealvqa_spatial_5",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the remote on the left side of the tv?"
},
{
"from": "gpt",
"value": "No, the remote is not on the left side of the tv."
},
{
"from": "human",
"value": "Is the remote above or below the tv?"
},
{
"from": "gpt",
"value": "The remote is above the tv."
}
],
"question": "Is the remote above or below the tv?",
"question_id": "sealvqa_spatial_5",
"search": 1,
"target_instances": [
{
"bbox": [
345.77,
172.05,
12.82,
17.13
],
"instance_id": "",
"name": "remote"
},
{
"bbox": [
185.33,
305.15,
79.27,
68.91
],
"instance_id": "",
"name": "tv"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the sandwich is not above the keyboard."
} | {
"answer": "No, the sandwich is not above the keyboard.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000054943.jpg",
"index": "sealvqa_spatial_14",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the sandwich to the right of the keyboard?"
},
{
"from": "gpt",
"value": "Yes, the sandwich is to the right of the keyboard."
},
{
"from": "human",
"value": "Is the sandwich above the keyboard?"
},
{
"from": "gpt",
"value": "No, the sandwich is not above the keyboard."
}
],
"question": "Is the sandwich above the keyboard?",
"question_id": "sealvqa_spatial_14",
"search": 1,
"target_instances": [
{
"bbox": [
369.93,
236.18,
28.61,
14.81
],
"instance_id": "",
"name": "sandwich"
},
{
"bbox": [
212.5,
140.02,
19.79,
3.32
],
"instance_id": "",
"name": "keyboard"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The clock is below the oven."
} | {
"answer": "The clock is below the oven.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000178872.jpg",
"index": "sealvqa_spatial_22",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the oven to the right of the clock?"
},
{
"from": "gpt",
"value": "No, the oven is not to the right of the clock."
},
{
"from": "human",
"value": "Is the clock above or below the oven?"
},
{
"from": "gpt",
"value": "The clock is below the oven."
}
],
"question": "Is the clock above or below the oven?",
"question_id": "sealvqa_spatial_22",
"search": 1,
"target_instances": [
{
"bbox": [
524.98,
306.29,
9.05,
20.22
],
"instance_id": "",
"name": "clock"
},
{
"bbox": [
248.91,
193.45,
75.76,
58.38
],
"instance_id": "",
"name": "oven"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the sink is on the right side of the microwave."
} | {
"answer": "Yes, the sink is on the right side of the microwave.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000304391.jpg",
"index": "sealvqa_spatial_41",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the sink on the right side of the microwave?"
},
{
"from": "gpt",
"value": "Yes, the sink is on the right side of the microwave."
}
],
"question": "Is the sink on the right side of the microwave?",
"question_id": "sealvqa_spatial_41",
"search": 1,
"target_instances": [
{
"bbox": [
127.86,
351.06,
69.24,
11.19
],
"instance_id": "",
"name": "sink"
},
{
"bbox": [
89.6,
333.51,
18.59,
48.47
],
"instance_id": "",
"name": "microwave"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the snowboard is not below the backpack."
} | {
"answer": "No, the snowboard is not below the backpack.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000417742.jpg",
"index": "sealvqa_spatial_57",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the backpack on the left or right side of the snowboard?"
},
{
"from": "gpt",
"value": "The backpack is on the right side of the snowboard."
},
{
"from": "human",
"value": "Is the snowboard below the backpack?"
},
{
"from": "gpt",
"value": "No, the snowboard is not below the backpack."
}
],
"question": "Is the snowboard below the backpack?",
"question_id": "sealvqa_spatial_57",
"search": 1,
"target_instances": [
{
"bbox": [
49.87,
137.85,
8.79,
29.5
],
"instance_id": "",
"name": "snowboard"
},
{
"bbox": [
229.61,
176.03,
60.82,
104.91
],
"instance_id": "",
"name": "backpack"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The handbag is above the backpack."
} | {
"answer": "The handbag is above the backpack.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000217136.jpg",
"index": "sealvqa_spatial_91",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the handbag on the left or right side of the backpack?"
},
{
"from": "gpt",
"value": "The handbag is on the left side of the backpack."
},
{
"from": "human",
"value": "Is the handbag above or below the backpack?"
},
{
"from": "gpt",
"value": "The handbag is above the backpack."
}
],
"question": "Is the handbag above or below the backpack?",
"question_id": "sealvqa_spatial_91",
"search": 1,
"target_instances": [
{
"bbox": [
120.72,
171.42,
9.81,
9.36
],
"instance_id": "",
"name": "handbag"
},
{
"bbox": [
458.46,
198.28,
29.49,
11.98
],
"instance_id": "",
"name": "backpack"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the microwave is not above the orange."
} | {
"answer": "No, the microwave is not above the orange.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000130555.jpg",
"index": "sealvqa_spatial_100",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the microwave on the left or right side of the orange?"
},
{
"from": "gpt",
"value": "The microwave is on the left side of the orange."
},
{
"from": "human",
"value": "Is the microwave above the orange?"
},
{
"from": "gpt",
"value": "No, the microwave is not above the orange."
}
],
"question": "Is the microwave above the orange?",
"question_id": "sealvqa_spatial_100",
"search": 1,
"target_instances": [
{
"bbox": [
13.06,
252.9,
81.95,
67
],
"instance_id": "",
"name": "microwave"
},
{
"bbox": [
165.31,
188.04,
14.92,
7.05
],
"instance_id": "",
"name": "orange"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the sink is on the right side of the cup."
} | {
"answer": "Yes, the sink is on the right side of the cup.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000475546.jpg",
"index": "sealvqa_spatial_111",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the sink on the right side of the cup?"
},
{
"from": "gpt",
"value": "Yes, the sink is on the right side of the cup."
}
],
"question": "Is the sink on the right side of the cup?",
"question_id": "sealvqa_spatial_111",
"search": 1,
"target_instances": [
{
"bbox": [
195.73,
267.76,
13.14,
25.15
],
"instance_id": "",
"name": "cup"
},
{
"bbox": [
270.32,
260.22,
114.92,
67.4
],
"instance_id": "",
"name": "sink"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the cup is on the right side of the chair."
} | {
"answer": "Yes, the cup is on the right side of the chair.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000453680.jpg",
"index": "sealvqa_spatial_119",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the cup on the right side of the chair?"
},
{
"from": "gpt",
"value": "Yes, the cup is on the right side of the chair."
}
],
"question": "Is the cup on the right side of the chair?",
"question_id": "sealvqa_spatial_119",
"search": 1,
"target_instances": [
{
"bbox": [
467.35,
174.49,
11.38,
24.26
],
"instance_id": "",
"name": "cup"
},
{
"bbox": [
403.48,
157.7,
63.12,
79.85
],
"instance_id": "",
"name": "chair"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the sink is not above the spoon."
} | {
"answer": "No, the sink is not above the spoon.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000090255.jpg",
"index": "sealvqa_spatial_122",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the spoon on the left or right side of the sink?"
},
{
"from": "gpt",
"value": "The spoon is on the right side of the sink."
},
{
"from": "human",
"value": "Is the sink above the spoon?"
},
{
"from": "gpt",
"value": "No, the sink is not above the spoon."
}
],
"question": "Is the sink above the spoon?",
"question_id": "sealvqa_spatial_122",
"search": 1,
"target_instances": [
{
"bbox": [
487.76,
115.31,
34.56,
95.6
],
"instance_id": "",
"name": "spoon"
},
{
"bbox": [
358.33,
462.54,
120.79,
33.37
],
"instance_id": "",
"name": "sink"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the sports ball is on the right side of the dog."
} | {
"answer": "Yes, the sports ball is on the right side of the dog.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000170292.jpg",
"index": "sealvqa_spatial_127",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the sports ball on the right side of the dog?"
},
{
"from": "gpt",
"value": "Yes, the sports ball is on the right side of the dog."
}
],
"question": "Is the sports ball on the right side of the dog?",
"question_id": "sealvqa_spatial_127",
"search": 1,
"target_instances": [
{
"bbox": [
160.98,
174.27,
48.13,
36.32
],
"instance_id": "",
"name": "sports ball"
},
{
"bbox": [
55.97,
136.57,
63.91,
40.89
],
"instance_id": "",
"name": "dog"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the baseball glove is below the sports ball."
} | {
"answer": "Yes, the baseball glove is below the sports ball.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000554664.jpg",
"index": "sealvqa_spatial_135",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the sports ball on the right side of the baseball glove?"
},
{
"from": "gpt",
"value": "Yes, the sports ball is on the right side of the baseball glove."
},
{
"from": "human",
"value": "Is the baseball glove below the sports ball?"
},
{
"from": "gpt",
"value": "Yes, the baseball glove is below the sports ball."
}
],
"question": "Is the baseball glove below the sports ball?",
"question_id": "sealvqa_spatial_135",
"search": 1,
"target_instances": [
{
"bbox": [
324.54,
125.45,
18.95,
31.09
],
"instance_id": "",
"name": "baseball glove"
},
{
"bbox": [
585.09,
46.96,
8.83,
9.58
],
"instance_id": "",
"name": "sports ball"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the backpack is not below the teddy bear."
} | {
"answer": "No, the backpack is not below the teddy bear.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000136503.jpg",
"index": "sealvqa_spatial_139",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the backpack on the right side of the teddy bear?"
},
{
"from": "gpt",
"value": "No, the backpack is not on the right side of the teddy bear."
},
{
"from": "human",
"value": "Is the backpack below the teddy bear?"
},
{
"from": "gpt",
"value": "No, the backpack is not below the teddy bear."
}
],
"question": "Is the backpack below the teddy bear?",
"question_id": "sealvqa_spatial_139",
"search": 1,
"target_instances": [
{
"bbox": [
256.65,
178.24,
5.74,
15.27
],
"instance_id": "",
"name": "backpack"
},
{
"bbox": [
342.33,
298.99,
30.43,
76.59
],
"instance_id": "",
"name": "teddy bear"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The cell phone is below the keyboard."
} | {
"answer": "The cell phone is below the keyboard.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000385150.jpg",
"index": "sealvqa_spatial_141",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the cell phone to the left or right of the keyboard?"
},
{
"from": "gpt",
"value": "The cell phone is to the left of the keyboard."
},
{
"from": "human",
"value": "Is the cell phone above or below the keyboard?"
},
{
"from": "gpt",
"value": "The cell phone is below the keyboard."
}
],
"question": "Is the cell phone above or below the keyboard?",
"question_id": "sealvqa_spatial_141",
"search": 1,
"target_instances": [
{
"bbox": [
200.5,
352.62,
44.54,
16.99
],
"instance_id": "",
"name": "cell phone"
},
{
"bbox": [
326.57,
298.92,
68.92,
36.1
],
"instance_id": "",
"name": "keyboard"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the backpack is not below the handbag."
} | {
"answer": "No, the backpack is not below the handbag.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000543263.jpg",
"index": "sealvqa_spatial_148",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the handbag on the right side of the backpack?"
},
{
"from": "gpt",
"value": "No, the handbag is not on the right side of the backpack."
},
{
"from": "human",
"value": "Is the backpack below the handbag?"
},
{
"from": "gpt",
"value": "No, the backpack is not below the handbag."
}
],
"question": "Is the backpack below the handbag?",
"question_id": "sealvqa_spatial_148",
"search": 1,
"target_instances": [
{
"bbox": [
564.99,
101.55,
35.91,
56.72
],
"instance_id": "",
"name": "backpack"
},
{
"bbox": [
197.58,
315.81,
92.35,
71.94
],
"instance_id": "",
"name": "handbag"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the mouse is below the person."
} | {
"answer": "Yes, the mouse is below the person.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000386768.jpg",
"index": "sealvqa_spatial_153",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the mouse on the left or right side of the person?"
},
{
"from": "gpt",
"value": "The mouse is on the left side of the person."
},
{
"from": "human",
"value": "Is the mouse below the person?"
},
{
"from": "gpt",
"value": "Yes, the mouse is below the person."
}
],
"question": "Is the mouse below the person?",
"question_id": "sealvqa_spatial_153",
"search": 1,
"target_instances": [
{
"bbox": [
276.03,
148.99,
23.78,
14.09
],
"instance_id": "",
"name": "mouse"
},
{
"bbox": [
341.28,
83.83,
33.72,
20.14
],
"instance_id": "",
"name": "person"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the wine glass is to the right of the mouse."
} | {
"answer": "Yes, the wine glass is to the right of the mouse.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000354059.jpg",
"index": "sealvqa_spatial_162",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the wine glass to the right of the mouse?"
},
{
"from": "gpt",
"value": "Yes, the wine glass is to the right of the mouse."
}
],
"question": "Is the wine glass to the right of the mouse?",
"question_id": "sealvqa_spatial_162",
"search": 1,
"target_instances": [
{
"bbox": [
234.34,
196.52,
35.94,
29.11
],
"instance_id": "",
"name": "mouse"
},
{
"bbox": [
497.61,
175.51,
21.32,
24.88
],
"instance_id": "",
"name": "wine glass"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the bowl is not above the chair."
} | {
"answer": "No, the bowl is not above the chair.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000124279.jpg",
"index": "sealvqa_spatial_169",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the bowl on the right side of the chair?"
},
{
"from": "gpt",
"value": "Yes, the bowl is on the right side of the chair."
},
{
"from": "human",
"value": "Is the bowl above the chair?"
},
{
"from": "gpt",
"value": "No, the bowl is not above the chair."
}
],
"question": "Is the bowl above the chair?",
"question_id": "sealvqa_spatial_169",
"search": 1,
"target_instances": [
{
"bbox": [
176.77,
262.58,
67.84,
47.1
],
"instance_id": "",
"name": "bowl"
},
{
"bbox": [
0,
168.63,
15.28,
58.25
],
"instance_id": "",
"name": "chair"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The remote is below the mouse."
} | {
"answer": "The remote is below the mouse.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000460484.jpg",
"index": "sealvqa_spatial_171",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the mouse to the right of the remote?"
},
{
"from": "gpt",
"value": "Yes, the mouse is to the right of the remote."
},
{
"from": "human",
"value": "Is the remote above or below the mouse?"
},
{
"from": "gpt",
"value": "The remote is below the mouse."
}
],
"question": "Is the remote above or below the mouse?",
"question_id": "sealvqa_spatial_171",
"search": 1,
"target_instances": [
{
"bbox": [
458.68,
117.34,
35.48,
28.72
],
"instance_id": "",
"name": "mouse"
},
{
"bbox": [
289.31,
169.23,
44.08,
42.85
],
"instance_id": "",
"name": "remote"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the handbag is to the left of the vase."
} | {
"answer": "Yes, the handbag is to the left of the vase.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000044054.jpg",
"index": "sealvqa_spatial_176",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the handbag to the left of the vase?"
},
{
"from": "gpt",
"value": "Yes, the handbag is to the left of the vase."
}
],
"question": "Is the handbag to the left of the vase?",
"question_id": "sealvqa_spatial_176",
"search": 1,
"target_instances": [
{
"bbox": [
225.21,
322.62,
60.06,
70.85
],
"instance_id": "",
"name": "handbag"
},
{
"bbox": [
380.08,
326.88,
53.84,
87.81
],
"instance_id": "",
"name": "vase"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the bowl is not above the bottle."
} | {
"answer": "No, the bowl is not above the bottle.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000124279.jpg",
"index": "sealvqa_spatial_178",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the bowl to the right of the bottle?"
},
{
"from": "gpt",
"value": "No, the bowl is not to the right of the bottle."
},
{
"from": "human",
"value": "Is the bowl above the bottle?"
},
{
"from": "gpt",
"value": "No, the bowl is not above the bottle."
}
],
"question": "Is the bowl above the bottle?",
"question_id": "sealvqa_spatial_178",
"search": 1,
"target_instances": [
{
"bbox": [
247.89,
151.31,
15.01,
43.52
],
"instance_id": "",
"name": "bottle"
},
{
"bbox": [
176.77,
262.58,
67.84,
47.1
],
"instance_id": "",
"name": "bowl"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The umbrella is above the bicycle."
} | {
"answer": "The umbrella is above the bicycle.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000051985.jpg",
"index": "sealvqa_spatial_186",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the bicycle to the left of the umbrella?"
},
{
"from": "gpt",
"value": "Yes, the bicycle is to the left of the umbrella."
},
{
"from": "human",
"value": "Is the umbrella above or below the bicycle?"
},
{
"from": "gpt",
"value": "The umbrella is above the bicycle."
}
],
"question": "Is the umbrella above or below the bicycle?",
"question_id": "sealvqa_spatial_186",
"search": 1,
"target_instances": [
{
"bbox": [
54.67,
240.8,
64.14,
46.25
],
"instance_id": "",
"name": "bicycle"
},
{
"bbox": [
443.8,
182.35,
34.72,
13.27
],
"instance_id": "",
"name": "umbrella"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the frisbee is not below the bottle."
} | {
"answer": "No, the frisbee is not below the bottle.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000098721.jpg",
"index": "sealvqa_spatial_188",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the bottle to the right of the frisbee?"
},
{
"from": "gpt",
"value": "Yes, the bottle is to the right of the frisbee."
},
{
"from": "human",
"value": "Is the frisbee below the bottle?"
},
{
"from": "gpt",
"value": "No, the frisbee is not below the bottle."
}
],
"question": "Is the frisbee below the bottle?",
"question_id": "sealvqa_spatial_188",
"search": 1,
"target_instances": [
{
"bbox": [
85.85,
283.29,
16.26,
6.23
],
"instance_id": "",
"name": "frisbee"
},
{
"bbox": [
577.4,
327.19,
6.52,
17.94
],
"instance_id": "",
"name": "bottle"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the donut is not to the right of the backpack."
} | {
"answer": "No, the donut is not to the right of the backpack.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000544568.jpg",
"index": "sealvqa_spatial_203",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the donut to the right of the backpack?"
},
{
"from": "gpt",
"value": "No, the donut is not to the right of the backpack."
}
],
"question": "Is the donut to the right of the backpack?",
"question_id": "sealvqa_spatial_203",
"search": 1,
"target_instances": [
{
"bbox": [
254.06,
120.61,
57.13,
35.28
],
"instance_id": "",
"name": "donut"
},
{
"bbox": [
484.91,
54.55,
85.95,
93.01
],
"instance_id": "",
"name": "backpack"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the handbag is not below the baseball bat."
} | {
"answer": "No, the handbag is not below the baseball bat.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000183620.jpg",
"index": "sealvqa_spatial_207",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the baseball bat to the left of the handbag?"
},
{
"from": "gpt",
"value": "Yes, the baseball bat is to the left of the handbag."
},
{
"from": "human",
"value": "Is the handbag below the baseball bat?"
},
{
"from": "gpt",
"value": "No, the handbag is not below the baseball bat."
}
],
"question": "Is the handbag below the baseball bat?",
"question_id": "sealvqa_spatial_207",
"search": 1,
"target_instances": [
{
"bbox": [
232.48,
275.01,
56.89,
74.17
],
"instance_id": "",
"name": "baseball bat"
},
{
"bbox": [
399.34,
126.53,
46.64,
43.53
],
"instance_id": "",
"name": "handbag"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The couch is on the right side of the cat."
} | {
"answer": "The couch is on the right side of the cat.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000464339.jpg",
"index": "sealvqa_spatial_220",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the couch on the left or right side of the cat?"
},
{
"from": "gpt",
"value": "The couch is on the right side of the cat."
}
],
"question": "Is the couch on the left or right side of the cat?",
"question_id": "sealvqa_spatial_220",
"search": 1,
"target_instances": [
{
"bbox": [
423.83,
423.51,
32.7,
51.89
],
"instance_id": "",
"name": "cat"
},
{
"bbox": [
536.11,
388.3,
103.26,
90.74
],
"instance_id": "",
"name": "couch"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The vase is on the right side of the couch."
} | {
"answer": "The vase is on the right side of the couch.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000489622.jpg",
"index": "sealvqa_spatial_229",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the vase on the left or right side of the couch?"
},
{
"from": "gpt",
"value": "The vase is on the right side of the couch."
}
],
"question": "Is the vase on the left or right side of the couch?",
"question_id": "sealvqa_spatial_229",
"search": 1,
"target_instances": [
{
"bbox": [
201.52,
182.85,
91.15,
96.88
],
"instance_id": "",
"name": "couch"
},
{
"bbox": [
500.42,
203.47,
25.8,
38.54
],
"instance_id": "",
"name": "vase"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the book is below the dining table."
} | {
"answer": "Yes, the book is below the dining table.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000037456.jpg",
"index": "sealvqa_spatial_230",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the dining table on the left or right side of the book?"
},
{
"from": "gpt",
"value": "The dining table is on the right side of the book."
},
{
"from": "human",
"value": "Is the book below the dining table?"
},
{
"from": "gpt",
"value": "Yes, the book is below the dining table."
}
],
"question": "Is the book below the dining table?",
"question_id": "sealvqa_spatial_230",
"search": 1,
"target_instances": [
{
"bbox": [
461.75,
113.07,
37.45,
19.4
],
"instance_id": "",
"name": "dining table"
},
{
"bbox": [
356.13,
305.36,
35.16,
16.62
],
"instance_id": "",
"name": "book"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the bus is to the left of the stop sign."
} | {
"answer": "Yes, the bus is to the left of the stop sign.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000300001.jpg",
"index": "sealvqa_spatial_244",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the bus to the left of the stop sign?"
},
{
"from": "gpt",
"value": "Yes, the bus is to the left of the stop sign."
}
],
"question": "Is the bus to the left of the stop sign?",
"question_id": "sealvqa_spatial_244",
"search": 1,
"target_instances": [
{
"bbox": [
271.12,
233.4,
130.33,
34.53
],
"instance_id": "",
"name": "bus"
},
{
"bbox": [
413.49,
237.93,
10.42,
10.68
],
"instance_id": "",
"name": "stop sign"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the couch is above the spoon."
} | {
"answer": "Yes, the couch is above the spoon.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000142767.jpg",
"index": "sealvqa_spatial_263",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the spoon to the left of the couch?"
},
{
"from": "gpt",
"value": "No, the spoon is not to the left of the couch."
},
{
"from": "human",
"value": "Is the couch above the spoon?"
},
{
"from": "gpt",
"value": "Yes, the couch is above the spoon."
}
],
"question": "Is the couch above the spoon?",
"question_id": "sealvqa_spatial_263",
"search": 1,
"target_instances": [
{
"bbox": [
106.09,
81.99,
135.97,
69.01
],
"instance_id": "",
"name": "couch"
},
{
"bbox": [
265.68,
246.13,
48.8,
36.21
],
"instance_id": "",
"name": "spoon"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the clock is not to the left of the car."
} | {
"answer": "No, the clock is not to the left of the car.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000413291.jpg",
"index": "sealvqa_spatial_267",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the clock to the left of the car?"
},
{
"from": "gpt",
"value": "No, the clock is not to the left of the car."
}
],
"question": "Is the clock to the left of the car?",
"question_id": "sealvqa_spatial_267",
"search": 1,
"target_instances": [
{
"bbox": [
256.33,
97.74,
26.89,
27.97
],
"instance_id": "",
"name": "clock"
},
{
"bbox": [
0.82,
119.28,
40.33,
25.37
],
"instance_id": "",
"name": "car"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the bottle is above the fork."
} | {
"answer": "Yes, the bottle is above the fork.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000089367.jpg",
"index": "sealvqa_spatial_282",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the fork on the left side of the bottle?"
},
{
"from": "gpt",
"value": "No, the fork is not on the left side of the bottle."
},
{
"from": "human",
"value": "Is the bottle above the fork?"
},
{
"from": "gpt",
"value": "Yes, the bottle is above the fork."
}
],
"question": "Is the bottle above the fork?",
"question_id": "sealvqa_spatial_282",
"search": 1,
"target_instances": [
{
"bbox": [
87.54,
181.14,
41.24,
101.37
],
"instance_id": "",
"name": "bottle"
},
{
"bbox": [
249.68,
454.39,
65.85,
65.12
],
"instance_id": "",
"name": "fork"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The handbag is below the cup."
} | {
"answer": "The handbag is below the cup.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000269174.jpg",
"index": "sealvqa_spatial_283",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the cup to the right of the handbag?"
},
{
"from": "gpt",
"value": "No, the cup is not to the right of the handbag."
},
{
"from": "human",
"value": "Is the handbag above or below the cup?"
},
{
"from": "gpt",
"value": "The handbag is below the cup."
}
],
"question": "Is the handbag above or below the cup?",
"question_id": "sealvqa_spatial_283",
"search": 1,
"target_instances": [
{
"bbox": [
0.15,
261.98,
31.12,
21.57
],
"instance_id": "",
"name": "cup"
},
{
"bbox": [
518.82,
365.25,
121.18,
57.31
],
"instance_id": "",
"name": "handbag"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The bench is on the right side of the bottle."
} | {
"answer": "The bench is on the right side of the bottle.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000569543.jpg",
"index": "sealvqa_spatial_295",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the bench on the left or right side of the bottle?"
},
{
"from": "gpt",
"value": "The bench is on the right side of the bottle."
}
],
"question": "Is the bench on the left or right side of the bottle?",
"question_id": "sealvqa_spatial_295",
"search": 1,
"target_instances": [
{
"bbox": [
564.42,
206.62,
8.72,
30.16
],
"instance_id": "",
"name": "bottle"
},
{
"bbox": [
592.68,
174.96,
47.32,
62.14
],
"instance_id": "",
"name": "bench"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The umbrella is to the left of the chair."
} | {
"answer": "The umbrella is to the left of the chair.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000290282.jpg",
"index": "sealvqa_spatial_296",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the umbrella to the left or right of the chair?"
},
{
"from": "gpt",
"value": "The umbrella is to the left of the chair."
}
],
"question": "Is the umbrella to the left or right of the chair?",
"question_id": "sealvqa_spatial_296",
"search": 1,
"target_instances": [
{
"bbox": [
18.01,
270.28,
18.43,
38.34
],
"instance_id": "",
"name": "umbrella"
},
{
"bbox": [
122.47,
287.26,
19.77,
28.81
],
"instance_id": "",
"name": "chair"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The dining table is above the bottle."
} | {
"answer": "The dining table is above the bottle.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000230537.jpg",
"index": "sealvqa_spatial_301",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the dining table on the left or right side of the bottle?"
},
{
"from": "gpt",
"value": "The dining table is on the right side of the bottle."
},
{
"from": "human",
"value": "Is the dining table above or below the bottle?"
},
{
"from": "gpt",
"value": "The dining table is above the bottle."
}
],
"question": "Is the dining table above or below the bottle?",
"question_id": "sealvqa_spatial_301",
"search": 1,
"target_instances": [
{
"bbox": [
281.76,
267.42,
28.57,
87.82
],
"instance_id": "",
"name": "bottle"
},
{
"bbox": [
356.28,
202.79,
48.79,
45.81
],
"instance_id": "",
"name": "dining table"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The backpack is above the bowl."
} | {
"answer": "The backpack is above the bowl.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000258611.jpg",
"index": "sealvqa_spatial_312",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the backpack on the left side of the bowl?"
},
{
"from": "gpt",
"value": "Yes, the backpack is on the left side of the bowl."
},
{
"from": "human",
"value": "Is the backpack above or below the bowl?"
},
{
"from": "gpt",
"value": "The backpack is above the bowl."
}
],
"question": "Is the backpack above or below the bowl?",
"question_id": "sealvqa_spatial_312",
"search": 1,
"target_instances": [
{
"bbox": [
42.53,
253.99,
40.92,
52.18
],
"instance_id": "",
"name": "backpack"
},
{
"bbox": [
428.2,
348.42,
61.95,
26.55
],
"instance_id": "",
"name": "bowl"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the vase is not below the handbag."
} | {
"answer": "No, the vase is not below the handbag.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000481355.jpg",
"index": "sealvqa_spatial_313",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the vase to the right of the handbag?"
},
{
"from": "gpt",
"value": "No, the vase is not to the right of the handbag."
},
{
"from": "human",
"value": "Is the vase below the handbag?"
},
{
"from": "gpt",
"value": "No, the vase is not below the handbag."
}
],
"question": "Is the vase below the handbag?",
"question_id": "sealvqa_spatial_313",
"search": 1,
"target_instances": [
{
"bbox": [
271.83,
156.2,
49.8,
49.48
],
"instance_id": "",
"name": "vase"
},
{
"bbox": [
581.4,
348.42,
58.1,
46.93
],
"instance_id": "",
"name": "handbag"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the knife is not above the banana."
} | {
"answer": "No, the knife is not above the banana.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000031537.jpg",
"index": "sealvqa_spatial_326",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the knife on the right side of the banana?"
},
{
"from": "gpt",
"value": "No, the knife is not on the right side of the banana."
},
{
"from": "human",
"value": "Is the knife above the banana?"
},
{
"from": "gpt",
"value": "No, the knife is not above the banana."
}
],
"question": "Is the knife above the banana?",
"question_id": "sealvqa_spatial_326",
"search": 1,
"target_instances": [
{
"bbox": [
95.81,
371.3,
42.04,
18.31
],
"instance_id": "",
"name": "knife"
},
{
"bbox": [
266.48,
270.52,
11.56,
13.17
],
"instance_id": "",
"name": "banana"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The couch is on the right side of the chair."
} | {
"answer": "The couch is on the right side of the chair.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000060350.jpg",
"index": "sealvqa_spatial_329",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the couch on the left or right side of the chair?"
},
{
"from": "gpt",
"value": "The couch is on the right side of the chair."
}
],
"question": "Is the couch on the left or right side of the chair?",
"question_id": "sealvqa_spatial_329",
"search": 1,
"target_instances": [
{
"bbox": [
602.75,
212.03,
37.25,
79.93
],
"instance_id": "",
"name": "couch"
},
{
"bbox": [
505.43,
196.15,
28.29,
78.16
],
"instance_id": "",
"name": "chair"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The tie is above the knife."
} | {
"answer": "The tie is above the knife.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000572233.jpg",
"index": "sealvqa_spatial_332",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the knife on the right side of the tie?"
},
{
"from": "gpt",
"value": "No, the knife is not on the right side of the tie."
},
{
"from": "human",
"value": "Is the tie above or below the knife?"
},
{
"from": "gpt",
"value": "The tie is above the knife."
}
],
"question": "Is the tie above or below the knife?",
"question_id": "sealvqa_spatial_332",
"search": 1,
"target_instances": [
{
"bbox": [
291.26,
204.88,
47.27,
41.5
],
"instance_id": "",
"name": "tie"
},
{
"bbox": [
80.38,
311.48,
83.84,
14.08
],
"instance_id": "",
"name": "knife"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the mouse is not below the cup."
} | {
"answer": "No, the mouse is not below the cup.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000116556.jpg",
"index": "sealvqa_spatial_335",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the mouse on the left or right side of the cup?"
},
{
"from": "gpt",
"value": "The mouse is on the right side of the cup."
},
{
"from": "human",
"value": "Is the mouse below the cup?"
},
{
"from": "gpt",
"value": "No, the mouse is not below the cup."
}
],
"question": "Is the mouse below the cup?",
"question_id": "sealvqa_spatial_335",
"search": 1,
"target_instances": [
{
"bbox": [
279.19,
153.5,
15.69,
7.78
],
"instance_id": "",
"name": "mouse"
},
{
"bbox": [
141.78,
167.31,
17.19,
23.23
],
"instance_id": "",
"name": "cup"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the bottle is not above the cup."
} | {
"answer": "No, the bottle is not above the cup.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000344031.jpg",
"index": "sealvqa_spatial_336",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the cup on the left or right side of the bottle?"
},
{
"from": "gpt",
"value": "The cup is on the right side of the bottle."
},
{
"from": "human",
"value": "Is the bottle above the cup?"
},
{
"from": "gpt",
"value": "No, the bottle is not above the cup."
}
],
"question": "Is the bottle above the cup?",
"question_id": "sealvqa_spatial_336",
"search": 1,
"target_instances": [
{
"bbox": [
78.07,
344.61,
16.19,
29.79
],
"instance_id": "",
"name": "bottle"
},
{
"bbox": [
306.38,
172.54,
9.26,
18.83
],
"instance_id": "",
"name": "cup"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the sports ball is not above the baseball bat."
} | {
"answer": "No, the sports ball is not above the baseball bat.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000243660.jpg",
"index": "sealvqa_spatial_340",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the baseball bat to the right of the sports ball?"
},
{
"from": "gpt",
"value": "No, the baseball bat is not to the right of the sports ball."
},
{
"from": "human",
"value": "Is the sports ball above the baseball bat?"
},
{
"from": "gpt",
"value": "No, the sports ball is not above the baseball bat."
}
],
"question": "Is the sports ball above the baseball bat?",
"question_id": "sealvqa_spatial_340",
"search": 1,
"target_instances": [
{
"bbox": [
436.9,
225.5,
11.15,
4.64
],
"instance_id": "",
"name": "sports ball"
},
{
"bbox": [
256.13,
196.58,
36.54,
27.1
],
"instance_id": "",
"name": "baseball bat"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The remote is to the left of the chair."
} | {
"answer": "The remote is to the left of the chair.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000298657.jpg",
"index": "sealvqa_spatial_346",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the remote to the left or right of the chair?"
},
{
"from": "gpt",
"value": "The remote is to the left of the chair."
}
],
"question": "Is the remote to the left or right of the chair?",
"question_id": "sealvqa_spatial_346",
"search": 1,
"target_instances": [
{
"bbox": [
99.69,
215.72,
62.55,
127.99
],
"instance_id": "",
"name": "chair"
},
{
"bbox": [
0,
279.64,
41.55,
22.07
],
"instance_id": "",
"name": "remote"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the tennis racket is below the bottle."
} | {
"answer": "Yes, the tennis racket is below the bottle.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000276197.jpg",
"index": "sealvqa_spatial_348",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the bottle on the left or right side of the tennis racket?"
},
{
"from": "gpt",
"value": "The bottle is on the right side of the tennis racket."
},
{
"from": "human",
"value": "Is the tennis racket below the bottle?"
},
{
"from": "gpt",
"value": "Yes, the tennis racket is below the bottle."
}
],
"question": "Is the tennis racket below the bottle?",
"question_id": "sealvqa_spatial_348",
"search": 1,
"target_instances": [
{
"bbox": [
211,
277.33,
62.19,
148.67
],
"instance_id": "",
"name": "tennis racket"
},
{
"bbox": [
538.11,
57.96,
16.06,
16.88
],
"instance_id": "",
"name": "bottle"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The tie is below the sports ball."
} | {
"answer": "The tie is below the sports ball.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000231748.jpg",
"index": "sealvqa_spatial_355",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the tie on the left or right side of the sports ball?"
},
{
"from": "gpt",
"value": "The tie is on the left side of the sports ball."
},
{
"from": "human",
"value": "Is the tie above or below the sports ball?"
},
{
"from": "gpt",
"value": "The tie is below the sports ball."
}
],
"question": "Is the tie above or below the sports ball?",
"question_id": "sealvqa_spatial_355",
"search": 1,
"target_instances": [
{
"bbox": [
318.33,
72.75,
14.5,
13.96
],
"instance_id": "",
"name": "sports ball"
},
{
"bbox": [
236.72,
173.19,
10.07,
44.12
],
"instance_id": "",
"name": "tie"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the knife is not below the spoon."
} | {
"answer": "No, the knife is not below the spoon.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000041482.jpg",
"index": "sealvqa_spatial_358",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the knife to the right of the spoon?"
},
{
"from": "gpt",
"value": "Yes, the knife is to the right of the spoon."
},
{
"from": "human",
"value": "Is the knife below the spoon?"
},
{
"from": "gpt",
"value": "No, the knife is not below the spoon."
}
],
"question": "Is the knife below the spoon?",
"question_id": "sealvqa_spatial_358",
"search": 1,
"target_instances": [
{
"bbox": [
297.36,
358.84,
87.75,
43.87
],
"instance_id": "",
"name": "knife"
},
{
"bbox": [
0,
427.43,
42.23,
36.72
],
"instance_id": "",
"name": "spoon"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The remote is on the right side of the person."
} | {
"answer": "The remote is on the right side of the person.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000480594.jpg",
"index": "sealvqa_spatial_363",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the remote on the left or right side of the person?"
},
{
"from": "gpt",
"value": "The remote is on the right side of the person."
}
],
"question": "Is the remote on the left or right side of the person?",
"question_id": "sealvqa_spatial_363",
"search": 1,
"target_instances": [
{
"bbox": [
324.22,
183.94,
42.29,
194.55
],
"instance_id": "",
"name": "person"
},
{
"bbox": [
373.2,
193.12,
43.41,
9.4
],
"instance_id": "",
"name": "remote"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the truck is not on the left side of the umbrella."
} | {
"answer": "No, the truck is not on the left side of the umbrella.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000384027.jpg",
"index": "sealvqa_spatial_368",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the truck on the left side of the umbrella?"
},
{
"from": "gpt",
"value": "No, the truck is not on the left side of the umbrella."
}
],
"question": "Is the truck on the left side of the umbrella?",
"question_id": "sealvqa_spatial_368",
"search": 1,
"target_instances": [
{
"bbox": [
115.74,
317.14,
22.56,
39.72
],
"instance_id": "",
"name": "umbrella"
},
{
"bbox": [
576,
239.95,
64,
90.24
],
"instance_id": "",
"name": "truck"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the car is on the right side of the snowboard."
} | {
"answer": "Yes, the car is on the right side of the snowboard.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000030371.jpg",
"index": "sealvqa_spatial_375",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the car on the right side of the snowboard?"
},
{
"from": "gpt",
"value": "Yes, the car is on the right side of the snowboard."
}
],
"question": "Is the car on the right side of the snowboard?",
"question_id": "sealvqa_spatial_375",
"search": 1,
"target_instances": [
{
"bbox": [
133.03,
226.95,
158.36,
9.94
],
"instance_id": "",
"name": "snowboard"
},
{
"bbox": [
323.51,
230.72,
88.05,
32.09
],
"instance_id": "",
"name": "car"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the toilet is below the sink."
} | {
"answer": "Yes, the toilet is below the sink.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000050096.jpg",
"index": "sealvqa_spatial_382",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the toilet on the left side of the sink?"
},
{
"from": "gpt",
"value": "No, the toilet is not on the left side of the sink."
},
{
"from": "human",
"value": "Is the toilet below the sink?"
},
{
"from": "gpt",
"value": "Yes, the toilet is below the sink."
}
],
"question": "Is the toilet below the sink?",
"question_id": "sealvqa_spatial_382",
"search": 1,
"target_instances": [
{
"bbox": [
403,
417.8,
21,
85.61
],
"instance_id": "",
"name": "toilet"
},
{
"bbox": [
0,
366.06,
88.41,
27.54
],
"instance_id": "",
"name": "sink"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the cow is on the left side of the horse."
} | {
"answer": "Yes, the cow is on the left side of the horse.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000416960.jpg",
"index": "sealvqa_spatial_383",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the cow on the left side of the horse?"
},
{
"from": "gpt",
"value": "Yes, the cow is on the left side of the horse."
}
],
"question": "Is the cow on the left side of the horse?",
"question_id": "sealvqa_spatial_383",
"search": 1,
"target_instances": [
{
"bbox": [
474.08,
121.74,
71.11,
75.88
],
"instance_id": "",
"name": "horse"
},
{
"bbox": [
275.53,
73.56,
76.05,
54.86
],
"instance_id": "",
"name": "cow"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The sink is on the right side of the vase."
} | {
"answer": "The sink is on the right side of the vase.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000256189.jpg",
"index": "sealvqa_spatial_388",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the sink on the left or right side of the vase?"
},
{
"from": "gpt",
"value": "The sink is on the right side of the vase."
}
],
"question": "Is the sink on the left or right side of the vase?",
"question_id": "sealvqa_spatial_388",
"search": 1,
"target_instances": [
{
"bbox": [
251.02,
210.23,
104.74,
68.99
],
"instance_id": "",
"name": "sink"
},
{
"bbox": [
123.33,
216.9,
39.81,
52.26
],
"instance_id": "",
"name": "vase"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the cup is not to the right of the backpack."
} | {
"answer": "No, the cup is not to the right of the backpack.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000332836.jpg",
"index": "sealvqa_spatial_406",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the cup to the right of the backpack?"
},
{
"from": "gpt",
"value": "No, the cup is not to the right of the backpack."
}
],
"question": "Is the cup to the right of the backpack?",
"question_id": "sealvqa_spatial_406",
"search": 1,
"target_instances": [
{
"bbox": [
17.68,
230.2,
16.68,
21.73
],
"instance_id": "",
"name": "cup"
},
{
"bbox": [
342.76,
201.24,
56.6,
71.72
],
"instance_id": "",
"name": "backpack"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the book is to the right of the remote."
} | {
"answer": "Yes, the book is to the right of the remote.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000214848.jpg",
"index": "sealvqa_spatial_423",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the book to the right of the remote?"
},
{
"from": "gpt",
"value": "Yes, the book is to the right of the remote."
}
],
"question": "Is the book to the right of the remote?",
"question_id": "sealvqa_spatial_423",
"search": 1,
"target_instances": [
{
"bbox": [
67.8,
287.18,
26.06,
8.86
],
"instance_id": "",
"name": "remote"
},
{
"bbox": [
241.14,
276.3,
45.55,
18.3
],
"instance_id": "",
"name": "book"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the sandwich is not above the sink."
} | {
"answer": "No, the sandwich is not above the sink.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000147545.jpg",
"index": "sealvqa_spatial_425",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the sink on the left side of the sandwich?"
},
{
"from": "gpt",
"value": "Yes, the sink is on the left side of the sandwich."
},
{
"from": "human",
"value": "Is the sandwich above the sink?"
},
{
"from": "gpt",
"value": "No, the sandwich is not above the sink."
}
],
"question": "Is the sandwich above the sink?",
"question_id": "sealvqa_spatial_425",
"search": 1,
"target_instances": [
{
"bbox": [
555.87,
294.61,
25.14,
15.39
],
"instance_id": "",
"name": "sandwich"
},
{
"bbox": [
105.89,
272.48,
39.37,
5.3
],
"instance_id": "",
"name": "sink"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the knife is not to the right of the spoon."
} | {
"answer": "No, the knife is not to the right of the spoon.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000165803.jpg",
"index": "sealvqa_spatial_432",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the knife to the right of the spoon?"
},
{
"from": "gpt",
"value": "No, the knife is not to the right of the spoon."
}
],
"question": "Is the knife to the right of the spoon?",
"question_id": "sealvqa_spatial_432",
"search": 1,
"target_instances": [
{
"bbox": [
110.76,
409.91,
33.75,
91.39
],
"instance_id": "",
"name": "knife"
},
{
"bbox": [
280.82,
491.81,
27.23,
33.99
],
"instance_id": "",
"name": "spoon"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "The fork is above the microwave."
} | {
"answer": "The fork is above the microwave.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000302077.jpg",
"index": "sealvqa_spatial_434",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the fork on the left side of the microwave?"
},
{
"from": "gpt",
"value": "Yes, the fork is on the left side of the microwave."
},
{
"from": "human",
"value": "Is the fork above or below the microwave?"
},
{
"from": "gpt",
"value": "The fork is above the microwave."
}
],
"question": "Is the fork above or below the microwave?",
"question_id": "sealvqa_spatial_434",
"search": 1,
"target_instances": [
{
"bbox": [
439.76,
113.82,
2.72,
10.02
],
"instance_id": "",
"name": "fork"
},
{
"bbox": [
469,
124.26,
31,
32.99
],
"instance_id": "",
"name": "microwave"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the car is not below the sports ball."
} | {
"answer": "No, the car is not below the sports ball.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000474953.jpg",
"index": "sealvqa_spatial_439",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the sports ball on the right side of the car?"
},
{
"from": "gpt",
"value": "No, the sports ball is not on the right side of the car."
},
{
"from": "human",
"value": "Is the car below the sports ball?"
},
{
"from": "gpt",
"value": "No, the car is not below the sports ball."
}
],
"question": "Is the car below the sports ball?",
"question_id": "sealvqa_spatial_439",
"search": 1,
"target_instances": [
{
"bbox": [
253.84,
124.22,
8.26,
8.49
],
"instance_id": "",
"name": "sports ball"
},
{
"bbox": [
621.2,
107.27,
18.8,
8.32
],
"instance_id": "",
"name": "car"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the backpack is not to the left of the cup."
} | {
"answer": "No, the backpack is not to the left of the cup.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000536609.jpg",
"index": "sealvqa_spatial_445",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the backpack to the left of the cup?"
},
{
"from": "gpt",
"value": "No, the backpack is not to the left of the cup."
}
],
"question": "Is the backpack to the left of the cup?",
"question_id": "sealvqa_spatial_445",
"search": 1,
"target_instances": [
{
"bbox": [
143.48,
192.65,
7.47,
24.38
],
"instance_id": "",
"name": "backpack"
},
{
"bbox": [
7,
199.05,
14.39,
22.54
],
"instance_id": "",
"name": "cup"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the bench is not below the baseball bat."
} | {
"answer": "No, the bench is not below the baseball bat.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000363831.jpg",
"index": "sealvqa_spatial_448",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the baseball bat to the left of the bench?"
},
{
"from": "gpt",
"value": "No, the baseball bat is not to the left of the bench."
},
{
"from": "human",
"value": "Is the bench below the baseball bat?"
},
{
"from": "gpt",
"value": "No, the bench is not below the baseball bat."
}
],
"question": "Is the bench below the baseball bat?",
"question_id": "sealvqa_spatial_448",
"search": 1,
"target_instances": [
{
"bbox": [
391.46,
235.21,
72.99,
13.66
],
"instance_id": "",
"name": "baseball bat"
},
{
"bbox": [
57.63,
8.54,
88.16,
49.59
],
"instance_id": "",
"name": "bench"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the skis is not to the right of the tv."
} | {
"answer": "No, the skis is not to the right of the tv.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000227226.jpg",
"index": "sealvqa_spatial_449",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the skis to the right of the tv?"
},
{
"from": "gpt",
"value": "No, the skis is not to the right of the tv."
}
],
"question": "Is the skis to the right of the tv?",
"question_id": "sealvqa_spatial_449",
"search": 1,
"target_instances": [
{
"bbox": [
285.9,
163.31,
47.31,
42.93
],
"instance_id": "",
"name": "tv"
},
{
"bbox": [
126.08,
187.07,
15.04,
135.34
],
"instance_id": "",
"name": "skis"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the truck is above the bench."
} | {
"answer": "Yes, the truck is above the bench.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000250608.jpg",
"index": "sealvqa_spatial_452",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the truck on the left or right side of the bench?"
},
{
"from": "gpt",
"value": "The truck is on the left side of the bench."
},
{
"from": "human",
"value": "Is the truck above the bench?"
},
{
"from": "gpt",
"value": "Yes, the truck is above the bench."
}
],
"question": "Is the truck above the bench?",
"question_id": "sealvqa_spatial_452",
"search": 1,
"target_instances": [
{
"bbox": [
572.8,
283.56,
27.06,
9.68
],
"instance_id": "",
"name": "bench"
},
{
"bbox": [
39.72,
241.83,
52.77,
33.06
],
"instance_id": "",
"name": "truck"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "No, the bus is not on the right side of the car."
} | {
"answer": "No, the bus is not on the right side of the car.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 2,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000338880.jpg",
"index": "sealvqa_spatial_454",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the bus on the right side of the car?"
},
{
"from": "gpt",
"value": "No, the bus is not on the right side of the car."
}
],
"question": "Is the bus on the right side of the car?",
"question_id": "sealvqa_spatial_454",
"search": 1,
"target_instances": [
{
"bbox": [
221.81,
448.05,
61.38,
23.63
],
"instance_id": "",
"name": "car"
},
{
"bbox": [
66.34,
426.16,
86.93,
43.05
],
"instance_id": "",
"name": "bus"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the motorcycle is above the book."
} | {
"answer": "Yes, the motorcycle is above the book.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000080441.jpg",
"index": "sealvqa_spatial_458",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the book to the right of the motorcycle?"
},
{
"from": "gpt",
"value": "Yes, the book is to the right of the motorcycle."
},
{
"from": "human",
"value": "Is the motorcycle above the book?"
},
{
"from": "gpt",
"value": "Yes, the motorcycle is above the book."
}
],
"question": "Is the motorcycle above the book?",
"question_id": "sealvqa_spatial_458",
"search": 1,
"target_instances": [
{
"bbox": [
545.6,
339.23,
6.22,
11.84
],
"instance_id": "",
"name": "book"
},
{
"bbox": [
316.46,
305.92,
9.28,
17.35
],
"instance_id": "",
"name": "motorcycle"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 | |
sealvqa_spatial | {
"ground_truth": "Yes, the bowl is below the sandwich."
} | {
"answer": "Yes, the bowl is below the sandwich.",
"bbox_format": "absolute_xywh",
"bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format",
"conversation_turns": 4,
"data_split": "train",
"dataset_type": "spatial_relation",
"has_targets": true,
"image_path": "coco2017/train2017/000000079277.jpg",
"index": "sealvqa_spatial_464",
"original_conversations": [
{
"from": "human",
"value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the bowl to the left of the sandwich?"
},
{
"from": "gpt",
"value": "No, the bowl is not to the left of the sandwich."
},
{
"from": "human",
"value": "Is the bowl below the sandwich?"
},
{
"from": "gpt",
"value": "Yes, the bowl is below the sandwich."
}
],
"question": "Is the bowl below the sandwich?",
"question_id": "sealvqa_spatial_464",
"search": 1,
"target_instances": [
{
"bbox": [
113.36,
377.33,
116.24,
84.66
],
"instance_id": "",
"name": "sandwich"
},
{
"bbox": [
330.21,
620.12,
96.79,
19.88
],
"instance_id": "",
"name": "bowl"
}
]
} | [
{
"content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci... | sealvqa_spatial_spatial | visual_toolbox_v2 |
End of preview. Expand
in Data Studio
README.md exists but content is empty.
- Downloads last month
- 11