Dataset Viewer
Auto-converted to Parquet Duplicate
images
images listlengths
1
1
data_source
stringclasses
1 value
reward_model
dict
extra_info
dict
prompt
listlengths
2
2
ability
stringclasses
1 value
env_name
stringclasses
1 value
sealvqa_spatial
{ "ground_truth": "The remote is above the tv." }
{ "answer": "The remote is above the tv.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000111922.jpg", "index": "sealvqa_spatial_5", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the remote on the left side of the tv?" }, { "from": "gpt", "value": "No, the remote is not on the left side of the tv." }, { "from": "human", "value": "Is the remote above or below the tv?" }, { "from": "gpt", "value": "The remote is above the tv." } ], "question": "Is the remote above or below the tv?", "question_id": "sealvqa_spatial_5", "search": 1, "target_instances": [ { "bbox": [ 345.77, 172.05, 12.82, 17.13 ], "instance_id": "", "name": "remote" }, { "bbox": [ 185.33, 305.15, 79.27, 68.91 ], "instance_id": "", "name": "tv" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the sandwich is not above the keyboard." }
{ "answer": "No, the sandwich is not above the keyboard.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000054943.jpg", "index": "sealvqa_spatial_14", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the sandwich to the right of the keyboard?" }, { "from": "gpt", "value": "Yes, the sandwich is to the right of the keyboard." }, { "from": "human", "value": "Is the sandwich above the keyboard?" }, { "from": "gpt", "value": "No, the sandwich is not above the keyboard." } ], "question": "Is the sandwich above the keyboard?", "question_id": "sealvqa_spatial_14", "search": 1, "target_instances": [ { "bbox": [ 369.93, 236.18, 28.61, 14.81 ], "instance_id": "", "name": "sandwich" }, { "bbox": [ 212.5, 140.02, 19.79, 3.32 ], "instance_id": "", "name": "keyboard" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The clock is below the oven." }
{ "answer": "The clock is below the oven.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000178872.jpg", "index": "sealvqa_spatial_22", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the oven to the right of the clock?" }, { "from": "gpt", "value": "No, the oven is not to the right of the clock." }, { "from": "human", "value": "Is the clock above or below the oven?" }, { "from": "gpt", "value": "The clock is below the oven." } ], "question": "Is the clock above or below the oven?", "question_id": "sealvqa_spatial_22", "search": 1, "target_instances": [ { "bbox": [ 524.98, 306.29, 9.05, 20.22 ], "instance_id": "", "name": "clock" }, { "bbox": [ 248.91, 193.45, 75.76, 58.38 ], "instance_id": "", "name": "oven" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the sink is on the right side of the microwave." }
{ "answer": "Yes, the sink is on the right side of the microwave.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000304391.jpg", "index": "sealvqa_spatial_41", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the sink on the right side of the microwave?" }, { "from": "gpt", "value": "Yes, the sink is on the right side of the microwave." } ], "question": "Is the sink on the right side of the microwave?", "question_id": "sealvqa_spatial_41", "search": 1, "target_instances": [ { "bbox": [ 127.86, 351.06, 69.24, 11.19 ], "instance_id": "", "name": "sink" }, { "bbox": [ 89.6, 333.51, 18.59, 48.47 ], "instance_id": "", "name": "microwave" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the snowboard is not below the backpack." }
{ "answer": "No, the snowboard is not below the backpack.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000417742.jpg", "index": "sealvqa_spatial_57", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the backpack on the left or right side of the snowboard?" }, { "from": "gpt", "value": "The backpack is on the right side of the snowboard." }, { "from": "human", "value": "Is the snowboard below the backpack?" }, { "from": "gpt", "value": "No, the snowboard is not below the backpack." } ], "question": "Is the snowboard below the backpack?", "question_id": "sealvqa_spatial_57", "search": 1, "target_instances": [ { "bbox": [ 49.87, 137.85, 8.79, 29.5 ], "instance_id": "", "name": "snowboard" }, { "bbox": [ 229.61, 176.03, 60.82, 104.91 ], "instance_id": "", "name": "backpack" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The handbag is above the backpack." }
{ "answer": "The handbag is above the backpack.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000217136.jpg", "index": "sealvqa_spatial_91", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the handbag on the left or right side of the backpack?" }, { "from": "gpt", "value": "The handbag is on the left side of the backpack." }, { "from": "human", "value": "Is the handbag above or below the backpack?" }, { "from": "gpt", "value": "The handbag is above the backpack." } ], "question": "Is the handbag above or below the backpack?", "question_id": "sealvqa_spatial_91", "search": 1, "target_instances": [ { "bbox": [ 120.72, 171.42, 9.81, 9.36 ], "instance_id": "", "name": "handbag" }, { "bbox": [ 458.46, 198.28, 29.49, 11.98 ], "instance_id": "", "name": "backpack" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the microwave is not above the orange." }
{ "answer": "No, the microwave is not above the orange.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000130555.jpg", "index": "sealvqa_spatial_100", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the microwave on the left or right side of the orange?" }, { "from": "gpt", "value": "The microwave is on the left side of the orange." }, { "from": "human", "value": "Is the microwave above the orange?" }, { "from": "gpt", "value": "No, the microwave is not above the orange." } ], "question": "Is the microwave above the orange?", "question_id": "sealvqa_spatial_100", "search": 1, "target_instances": [ { "bbox": [ 13.06, 252.9, 81.95, 67 ], "instance_id": "", "name": "microwave" }, { "bbox": [ 165.31, 188.04, 14.92, 7.05 ], "instance_id": "", "name": "orange" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the sink is on the right side of the cup." }
{ "answer": "Yes, the sink is on the right side of the cup.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000475546.jpg", "index": "sealvqa_spatial_111", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the sink on the right side of the cup?" }, { "from": "gpt", "value": "Yes, the sink is on the right side of the cup." } ], "question": "Is the sink on the right side of the cup?", "question_id": "sealvqa_spatial_111", "search": 1, "target_instances": [ { "bbox": [ 195.73, 267.76, 13.14, 25.15 ], "instance_id": "", "name": "cup" }, { "bbox": [ 270.32, 260.22, 114.92, 67.4 ], "instance_id": "", "name": "sink" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the cup is on the right side of the chair." }
{ "answer": "Yes, the cup is on the right side of the chair.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000453680.jpg", "index": "sealvqa_spatial_119", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the cup on the right side of the chair?" }, { "from": "gpt", "value": "Yes, the cup is on the right side of the chair." } ], "question": "Is the cup on the right side of the chair?", "question_id": "sealvqa_spatial_119", "search": 1, "target_instances": [ { "bbox": [ 467.35, 174.49, 11.38, 24.26 ], "instance_id": "", "name": "cup" }, { "bbox": [ 403.48, 157.7, 63.12, 79.85 ], "instance_id": "", "name": "chair" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the sink is not above the spoon." }
{ "answer": "No, the sink is not above the spoon.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000090255.jpg", "index": "sealvqa_spatial_122", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the spoon on the left or right side of the sink?" }, { "from": "gpt", "value": "The spoon is on the right side of the sink." }, { "from": "human", "value": "Is the sink above the spoon?" }, { "from": "gpt", "value": "No, the sink is not above the spoon." } ], "question": "Is the sink above the spoon?", "question_id": "sealvqa_spatial_122", "search": 1, "target_instances": [ { "bbox": [ 487.76, 115.31, 34.56, 95.6 ], "instance_id": "", "name": "spoon" }, { "bbox": [ 358.33, 462.54, 120.79, 33.37 ], "instance_id": "", "name": "sink" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the sports ball is on the right side of the dog." }
{ "answer": "Yes, the sports ball is on the right side of the dog.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000170292.jpg", "index": "sealvqa_spatial_127", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the sports ball on the right side of the dog?" }, { "from": "gpt", "value": "Yes, the sports ball is on the right side of the dog." } ], "question": "Is the sports ball on the right side of the dog?", "question_id": "sealvqa_spatial_127", "search": 1, "target_instances": [ { "bbox": [ 160.98, 174.27, 48.13, 36.32 ], "instance_id": "", "name": "sports ball" }, { "bbox": [ 55.97, 136.57, 63.91, 40.89 ], "instance_id": "", "name": "dog" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the baseball glove is below the sports ball." }
{ "answer": "Yes, the baseball glove is below the sports ball.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000554664.jpg", "index": "sealvqa_spatial_135", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the sports ball on the right side of the baseball glove?" }, { "from": "gpt", "value": "Yes, the sports ball is on the right side of the baseball glove." }, { "from": "human", "value": "Is the baseball glove below the sports ball?" }, { "from": "gpt", "value": "Yes, the baseball glove is below the sports ball." } ], "question": "Is the baseball glove below the sports ball?", "question_id": "sealvqa_spatial_135", "search": 1, "target_instances": [ { "bbox": [ 324.54, 125.45, 18.95, 31.09 ], "instance_id": "", "name": "baseball glove" }, { "bbox": [ 585.09, 46.96, 8.83, 9.58 ], "instance_id": "", "name": "sports ball" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the backpack is not below the teddy bear." }
{ "answer": "No, the backpack is not below the teddy bear.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000136503.jpg", "index": "sealvqa_spatial_139", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the backpack on the right side of the teddy bear?" }, { "from": "gpt", "value": "No, the backpack is not on the right side of the teddy bear." }, { "from": "human", "value": "Is the backpack below the teddy bear?" }, { "from": "gpt", "value": "No, the backpack is not below the teddy bear." } ], "question": "Is the backpack below the teddy bear?", "question_id": "sealvqa_spatial_139", "search": 1, "target_instances": [ { "bbox": [ 256.65, 178.24, 5.74, 15.27 ], "instance_id": "", "name": "backpack" }, { "bbox": [ 342.33, 298.99, 30.43, 76.59 ], "instance_id": "", "name": "teddy bear" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The cell phone is below the keyboard." }
{ "answer": "The cell phone is below the keyboard.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000385150.jpg", "index": "sealvqa_spatial_141", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the cell phone to the left or right of the keyboard?" }, { "from": "gpt", "value": "The cell phone is to the left of the keyboard." }, { "from": "human", "value": "Is the cell phone above or below the keyboard?" }, { "from": "gpt", "value": "The cell phone is below the keyboard." } ], "question": "Is the cell phone above or below the keyboard?", "question_id": "sealvqa_spatial_141", "search": 1, "target_instances": [ { "bbox": [ 200.5, 352.62, 44.54, 16.99 ], "instance_id": "", "name": "cell phone" }, { "bbox": [ 326.57, 298.92, 68.92, 36.1 ], "instance_id": "", "name": "keyboard" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the backpack is not below the handbag." }
{ "answer": "No, the backpack is not below the handbag.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000543263.jpg", "index": "sealvqa_spatial_148", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the handbag on the right side of the backpack?" }, { "from": "gpt", "value": "No, the handbag is not on the right side of the backpack." }, { "from": "human", "value": "Is the backpack below the handbag?" }, { "from": "gpt", "value": "No, the backpack is not below the handbag." } ], "question": "Is the backpack below the handbag?", "question_id": "sealvqa_spatial_148", "search": 1, "target_instances": [ { "bbox": [ 564.99, 101.55, 35.91, 56.72 ], "instance_id": "", "name": "backpack" }, { "bbox": [ 197.58, 315.81, 92.35, 71.94 ], "instance_id": "", "name": "handbag" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the mouse is below the person." }
{ "answer": "Yes, the mouse is below the person.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000386768.jpg", "index": "sealvqa_spatial_153", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the mouse on the left or right side of the person?" }, { "from": "gpt", "value": "The mouse is on the left side of the person." }, { "from": "human", "value": "Is the mouse below the person?" }, { "from": "gpt", "value": "Yes, the mouse is below the person." } ], "question": "Is the mouse below the person?", "question_id": "sealvqa_spatial_153", "search": 1, "target_instances": [ { "bbox": [ 276.03, 148.99, 23.78, 14.09 ], "instance_id": "", "name": "mouse" }, { "bbox": [ 341.28, 83.83, 33.72, 20.14 ], "instance_id": "", "name": "person" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the wine glass is to the right of the mouse." }
{ "answer": "Yes, the wine glass is to the right of the mouse.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000354059.jpg", "index": "sealvqa_spatial_162", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the wine glass to the right of the mouse?" }, { "from": "gpt", "value": "Yes, the wine glass is to the right of the mouse." } ], "question": "Is the wine glass to the right of the mouse?", "question_id": "sealvqa_spatial_162", "search": 1, "target_instances": [ { "bbox": [ 234.34, 196.52, 35.94, 29.11 ], "instance_id": "", "name": "mouse" }, { "bbox": [ 497.61, 175.51, 21.32, 24.88 ], "instance_id": "", "name": "wine glass" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the bowl is not above the chair." }
{ "answer": "No, the bowl is not above the chair.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000124279.jpg", "index": "sealvqa_spatial_169", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the bowl on the right side of the chair?" }, { "from": "gpt", "value": "Yes, the bowl is on the right side of the chair." }, { "from": "human", "value": "Is the bowl above the chair?" }, { "from": "gpt", "value": "No, the bowl is not above the chair." } ], "question": "Is the bowl above the chair?", "question_id": "sealvqa_spatial_169", "search": 1, "target_instances": [ { "bbox": [ 176.77, 262.58, 67.84, 47.1 ], "instance_id": "", "name": "bowl" }, { "bbox": [ 0, 168.63, 15.28, 58.25 ], "instance_id": "", "name": "chair" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The remote is below the mouse." }
{ "answer": "The remote is below the mouse.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000460484.jpg", "index": "sealvqa_spatial_171", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the mouse to the right of the remote?" }, { "from": "gpt", "value": "Yes, the mouse is to the right of the remote." }, { "from": "human", "value": "Is the remote above or below the mouse?" }, { "from": "gpt", "value": "The remote is below the mouse." } ], "question": "Is the remote above or below the mouse?", "question_id": "sealvqa_spatial_171", "search": 1, "target_instances": [ { "bbox": [ 458.68, 117.34, 35.48, 28.72 ], "instance_id": "", "name": "mouse" }, { "bbox": [ 289.31, 169.23, 44.08, 42.85 ], "instance_id": "", "name": "remote" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the handbag is to the left of the vase." }
{ "answer": "Yes, the handbag is to the left of the vase.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000044054.jpg", "index": "sealvqa_spatial_176", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the handbag to the left of the vase?" }, { "from": "gpt", "value": "Yes, the handbag is to the left of the vase." } ], "question": "Is the handbag to the left of the vase?", "question_id": "sealvqa_spatial_176", "search": 1, "target_instances": [ { "bbox": [ 225.21, 322.62, 60.06, 70.85 ], "instance_id": "", "name": "handbag" }, { "bbox": [ 380.08, 326.88, 53.84, 87.81 ], "instance_id": "", "name": "vase" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the bowl is not above the bottle." }
{ "answer": "No, the bowl is not above the bottle.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000124279.jpg", "index": "sealvqa_spatial_178", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the bowl to the right of the bottle?" }, { "from": "gpt", "value": "No, the bowl is not to the right of the bottle." }, { "from": "human", "value": "Is the bowl above the bottle?" }, { "from": "gpt", "value": "No, the bowl is not above the bottle." } ], "question": "Is the bowl above the bottle?", "question_id": "sealvqa_spatial_178", "search": 1, "target_instances": [ { "bbox": [ 247.89, 151.31, 15.01, 43.52 ], "instance_id": "", "name": "bottle" }, { "bbox": [ 176.77, 262.58, 67.84, 47.1 ], "instance_id": "", "name": "bowl" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The umbrella is above the bicycle." }
{ "answer": "The umbrella is above the bicycle.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000051985.jpg", "index": "sealvqa_spatial_186", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the bicycle to the left of the umbrella?" }, { "from": "gpt", "value": "Yes, the bicycle is to the left of the umbrella." }, { "from": "human", "value": "Is the umbrella above or below the bicycle?" }, { "from": "gpt", "value": "The umbrella is above the bicycle." } ], "question": "Is the umbrella above or below the bicycle?", "question_id": "sealvqa_spatial_186", "search": 1, "target_instances": [ { "bbox": [ 54.67, 240.8, 64.14, 46.25 ], "instance_id": "", "name": "bicycle" }, { "bbox": [ 443.8, 182.35, 34.72, 13.27 ], "instance_id": "", "name": "umbrella" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the frisbee is not below the bottle." }
{ "answer": "No, the frisbee is not below the bottle.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000098721.jpg", "index": "sealvqa_spatial_188", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the bottle to the right of the frisbee?" }, { "from": "gpt", "value": "Yes, the bottle is to the right of the frisbee." }, { "from": "human", "value": "Is the frisbee below the bottle?" }, { "from": "gpt", "value": "No, the frisbee is not below the bottle." } ], "question": "Is the frisbee below the bottle?", "question_id": "sealvqa_spatial_188", "search": 1, "target_instances": [ { "bbox": [ 85.85, 283.29, 16.26, 6.23 ], "instance_id": "", "name": "frisbee" }, { "bbox": [ 577.4, 327.19, 6.52, 17.94 ], "instance_id": "", "name": "bottle" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the donut is not to the right of the backpack." }
{ "answer": "No, the donut is not to the right of the backpack.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000544568.jpg", "index": "sealvqa_spatial_203", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the donut to the right of the backpack?" }, { "from": "gpt", "value": "No, the donut is not to the right of the backpack." } ], "question": "Is the donut to the right of the backpack?", "question_id": "sealvqa_spatial_203", "search": 1, "target_instances": [ { "bbox": [ 254.06, 120.61, 57.13, 35.28 ], "instance_id": "", "name": "donut" }, { "bbox": [ 484.91, 54.55, 85.95, 93.01 ], "instance_id": "", "name": "backpack" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the handbag is not below the baseball bat." }
{ "answer": "No, the handbag is not below the baseball bat.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000183620.jpg", "index": "sealvqa_spatial_207", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the baseball bat to the left of the handbag?" }, { "from": "gpt", "value": "Yes, the baseball bat is to the left of the handbag." }, { "from": "human", "value": "Is the handbag below the baseball bat?" }, { "from": "gpt", "value": "No, the handbag is not below the baseball bat." } ], "question": "Is the handbag below the baseball bat?", "question_id": "sealvqa_spatial_207", "search": 1, "target_instances": [ { "bbox": [ 232.48, 275.01, 56.89, 74.17 ], "instance_id": "", "name": "baseball bat" }, { "bbox": [ 399.34, 126.53, 46.64, 43.53 ], "instance_id": "", "name": "handbag" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The couch is on the right side of the cat." }
{ "answer": "The couch is on the right side of the cat.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000464339.jpg", "index": "sealvqa_spatial_220", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the couch on the left or right side of the cat?" }, { "from": "gpt", "value": "The couch is on the right side of the cat." } ], "question": "Is the couch on the left or right side of the cat?", "question_id": "sealvqa_spatial_220", "search": 1, "target_instances": [ { "bbox": [ 423.83, 423.51, 32.7, 51.89 ], "instance_id": "", "name": "cat" }, { "bbox": [ 536.11, 388.3, 103.26, 90.74 ], "instance_id": "", "name": "couch" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The vase is on the right side of the couch." }
{ "answer": "The vase is on the right side of the couch.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000489622.jpg", "index": "sealvqa_spatial_229", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the vase on the left or right side of the couch?" }, { "from": "gpt", "value": "The vase is on the right side of the couch." } ], "question": "Is the vase on the left or right side of the couch?", "question_id": "sealvqa_spatial_229", "search": 1, "target_instances": [ { "bbox": [ 201.52, 182.85, 91.15, 96.88 ], "instance_id": "", "name": "couch" }, { "bbox": [ 500.42, 203.47, 25.8, 38.54 ], "instance_id": "", "name": "vase" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the book is below the dining table." }
{ "answer": "Yes, the book is below the dining table.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000037456.jpg", "index": "sealvqa_spatial_230", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the dining table on the left or right side of the book?" }, { "from": "gpt", "value": "The dining table is on the right side of the book." }, { "from": "human", "value": "Is the book below the dining table?" }, { "from": "gpt", "value": "Yes, the book is below the dining table." } ], "question": "Is the book below the dining table?", "question_id": "sealvqa_spatial_230", "search": 1, "target_instances": [ { "bbox": [ 461.75, 113.07, 37.45, 19.4 ], "instance_id": "", "name": "dining table" }, { "bbox": [ 356.13, 305.36, 35.16, 16.62 ], "instance_id": "", "name": "book" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the bus is to the left of the stop sign." }
{ "answer": "Yes, the bus is to the left of the stop sign.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000300001.jpg", "index": "sealvqa_spatial_244", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the bus to the left of the stop sign?" }, { "from": "gpt", "value": "Yes, the bus is to the left of the stop sign." } ], "question": "Is the bus to the left of the stop sign?", "question_id": "sealvqa_spatial_244", "search": 1, "target_instances": [ { "bbox": [ 271.12, 233.4, 130.33, 34.53 ], "instance_id": "", "name": "bus" }, { "bbox": [ 413.49, 237.93, 10.42, 10.68 ], "instance_id": "", "name": "stop sign" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the couch is above the spoon." }
{ "answer": "Yes, the couch is above the spoon.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000142767.jpg", "index": "sealvqa_spatial_263", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the spoon to the left of the couch?" }, { "from": "gpt", "value": "No, the spoon is not to the left of the couch." }, { "from": "human", "value": "Is the couch above the spoon?" }, { "from": "gpt", "value": "Yes, the couch is above the spoon." } ], "question": "Is the couch above the spoon?", "question_id": "sealvqa_spatial_263", "search": 1, "target_instances": [ { "bbox": [ 106.09, 81.99, 135.97, 69.01 ], "instance_id": "", "name": "couch" }, { "bbox": [ 265.68, 246.13, 48.8, 36.21 ], "instance_id": "", "name": "spoon" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the clock is not to the left of the car." }
{ "answer": "No, the clock is not to the left of the car.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000413291.jpg", "index": "sealvqa_spatial_267", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the clock to the left of the car?" }, { "from": "gpt", "value": "No, the clock is not to the left of the car." } ], "question": "Is the clock to the left of the car?", "question_id": "sealvqa_spatial_267", "search": 1, "target_instances": [ { "bbox": [ 256.33, 97.74, 26.89, 27.97 ], "instance_id": "", "name": "clock" }, { "bbox": [ 0.82, 119.28, 40.33, 25.37 ], "instance_id": "", "name": "car" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the bottle is above the fork." }
{ "answer": "Yes, the bottle is above the fork.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000089367.jpg", "index": "sealvqa_spatial_282", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the fork on the left side of the bottle?" }, { "from": "gpt", "value": "No, the fork is not on the left side of the bottle." }, { "from": "human", "value": "Is the bottle above the fork?" }, { "from": "gpt", "value": "Yes, the bottle is above the fork." } ], "question": "Is the bottle above the fork?", "question_id": "sealvqa_spatial_282", "search": 1, "target_instances": [ { "bbox": [ 87.54, 181.14, 41.24, 101.37 ], "instance_id": "", "name": "bottle" }, { "bbox": [ 249.68, 454.39, 65.85, 65.12 ], "instance_id": "", "name": "fork" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The handbag is below the cup." }
{ "answer": "The handbag is below the cup.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000269174.jpg", "index": "sealvqa_spatial_283", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the cup to the right of the handbag?" }, { "from": "gpt", "value": "No, the cup is not to the right of the handbag." }, { "from": "human", "value": "Is the handbag above or below the cup?" }, { "from": "gpt", "value": "The handbag is below the cup." } ], "question": "Is the handbag above or below the cup?", "question_id": "sealvqa_spatial_283", "search": 1, "target_instances": [ { "bbox": [ 0.15, 261.98, 31.12, 21.57 ], "instance_id": "", "name": "cup" }, { "bbox": [ 518.82, 365.25, 121.18, 57.31 ], "instance_id": "", "name": "handbag" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The bench is on the right side of the bottle." }
{ "answer": "The bench is on the right side of the bottle.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000569543.jpg", "index": "sealvqa_spatial_295", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the bench on the left or right side of the bottle?" }, { "from": "gpt", "value": "The bench is on the right side of the bottle." } ], "question": "Is the bench on the left or right side of the bottle?", "question_id": "sealvqa_spatial_295", "search": 1, "target_instances": [ { "bbox": [ 564.42, 206.62, 8.72, 30.16 ], "instance_id": "", "name": "bottle" }, { "bbox": [ 592.68, 174.96, 47.32, 62.14 ], "instance_id": "", "name": "bench" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The umbrella is to the left of the chair." }
{ "answer": "The umbrella is to the left of the chair.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000290282.jpg", "index": "sealvqa_spatial_296", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the umbrella to the left or right of the chair?" }, { "from": "gpt", "value": "The umbrella is to the left of the chair." } ], "question": "Is the umbrella to the left or right of the chair?", "question_id": "sealvqa_spatial_296", "search": 1, "target_instances": [ { "bbox": [ 18.01, 270.28, 18.43, 38.34 ], "instance_id": "", "name": "umbrella" }, { "bbox": [ 122.47, 287.26, 19.77, 28.81 ], "instance_id": "", "name": "chair" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The dining table is above the bottle." }
{ "answer": "The dining table is above the bottle.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000230537.jpg", "index": "sealvqa_spatial_301", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the dining table on the left or right side of the bottle?" }, { "from": "gpt", "value": "The dining table is on the right side of the bottle." }, { "from": "human", "value": "Is the dining table above or below the bottle?" }, { "from": "gpt", "value": "The dining table is above the bottle." } ], "question": "Is the dining table above or below the bottle?", "question_id": "sealvqa_spatial_301", "search": 1, "target_instances": [ { "bbox": [ 281.76, 267.42, 28.57, 87.82 ], "instance_id": "", "name": "bottle" }, { "bbox": [ 356.28, 202.79, 48.79, 45.81 ], "instance_id": "", "name": "dining table" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The backpack is above the bowl." }
{ "answer": "The backpack is above the bowl.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000258611.jpg", "index": "sealvqa_spatial_312", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the backpack on the left side of the bowl?" }, { "from": "gpt", "value": "Yes, the backpack is on the left side of the bowl." }, { "from": "human", "value": "Is the backpack above or below the bowl?" }, { "from": "gpt", "value": "The backpack is above the bowl." } ], "question": "Is the backpack above or below the bowl?", "question_id": "sealvqa_spatial_312", "search": 1, "target_instances": [ { "bbox": [ 42.53, 253.99, 40.92, 52.18 ], "instance_id": "", "name": "backpack" }, { "bbox": [ 428.2, 348.42, 61.95, 26.55 ], "instance_id": "", "name": "bowl" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the vase is not below the handbag." }
{ "answer": "No, the vase is not below the handbag.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000481355.jpg", "index": "sealvqa_spatial_313", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the vase to the right of the handbag?" }, { "from": "gpt", "value": "No, the vase is not to the right of the handbag." }, { "from": "human", "value": "Is the vase below the handbag?" }, { "from": "gpt", "value": "No, the vase is not below the handbag." } ], "question": "Is the vase below the handbag?", "question_id": "sealvqa_spatial_313", "search": 1, "target_instances": [ { "bbox": [ 271.83, 156.2, 49.8, 49.48 ], "instance_id": "", "name": "vase" }, { "bbox": [ 581.4, 348.42, 58.1, 46.93 ], "instance_id": "", "name": "handbag" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the knife is not above the banana." }
{ "answer": "No, the knife is not above the banana.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000031537.jpg", "index": "sealvqa_spatial_326", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the knife on the right side of the banana?" }, { "from": "gpt", "value": "No, the knife is not on the right side of the banana." }, { "from": "human", "value": "Is the knife above the banana?" }, { "from": "gpt", "value": "No, the knife is not above the banana." } ], "question": "Is the knife above the banana?", "question_id": "sealvqa_spatial_326", "search": 1, "target_instances": [ { "bbox": [ 95.81, 371.3, 42.04, 18.31 ], "instance_id": "", "name": "knife" }, { "bbox": [ 266.48, 270.52, 11.56, 13.17 ], "instance_id": "", "name": "banana" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The couch is on the right side of the chair." }
{ "answer": "The couch is on the right side of the chair.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000060350.jpg", "index": "sealvqa_spatial_329", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the couch on the left or right side of the chair?" }, { "from": "gpt", "value": "The couch is on the right side of the chair." } ], "question": "Is the couch on the left or right side of the chair?", "question_id": "sealvqa_spatial_329", "search": 1, "target_instances": [ { "bbox": [ 602.75, 212.03, 37.25, 79.93 ], "instance_id": "", "name": "couch" }, { "bbox": [ 505.43, 196.15, 28.29, 78.16 ], "instance_id": "", "name": "chair" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The tie is above the knife." }
{ "answer": "The tie is above the knife.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000572233.jpg", "index": "sealvqa_spatial_332", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the knife on the right side of the tie?" }, { "from": "gpt", "value": "No, the knife is not on the right side of the tie." }, { "from": "human", "value": "Is the tie above or below the knife?" }, { "from": "gpt", "value": "The tie is above the knife." } ], "question": "Is the tie above or below the knife?", "question_id": "sealvqa_spatial_332", "search": 1, "target_instances": [ { "bbox": [ 291.26, 204.88, 47.27, 41.5 ], "instance_id": "", "name": "tie" }, { "bbox": [ 80.38, 311.48, 83.84, 14.08 ], "instance_id": "", "name": "knife" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the mouse is not below the cup." }
{ "answer": "No, the mouse is not below the cup.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000116556.jpg", "index": "sealvqa_spatial_335", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the mouse on the left or right side of the cup?" }, { "from": "gpt", "value": "The mouse is on the right side of the cup." }, { "from": "human", "value": "Is the mouse below the cup?" }, { "from": "gpt", "value": "No, the mouse is not below the cup." } ], "question": "Is the mouse below the cup?", "question_id": "sealvqa_spatial_335", "search": 1, "target_instances": [ { "bbox": [ 279.19, 153.5, 15.69, 7.78 ], "instance_id": "", "name": "mouse" }, { "bbox": [ 141.78, 167.31, 17.19, 23.23 ], "instance_id": "", "name": "cup" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the bottle is not above the cup." }
{ "answer": "No, the bottle is not above the cup.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000344031.jpg", "index": "sealvqa_spatial_336", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the cup on the left or right side of the bottle?" }, { "from": "gpt", "value": "The cup is on the right side of the bottle." }, { "from": "human", "value": "Is the bottle above the cup?" }, { "from": "gpt", "value": "No, the bottle is not above the cup." } ], "question": "Is the bottle above the cup?", "question_id": "sealvqa_spatial_336", "search": 1, "target_instances": [ { "bbox": [ 78.07, 344.61, 16.19, 29.79 ], "instance_id": "", "name": "bottle" }, { "bbox": [ 306.38, 172.54, 9.26, 18.83 ], "instance_id": "", "name": "cup" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the sports ball is not above the baseball bat." }
{ "answer": "No, the sports ball is not above the baseball bat.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000243660.jpg", "index": "sealvqa_spatial_340", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the baseball bat to the right of the sports ball?" }, { "from": "gpt", "value": "No, the baseball bat is not to the right of the sports ball." }, { "from": "human", "value": "Is the sports ball above the baseball bat?" }, { "from": "gpt", "value": "No, the sports ball is not above the baseball bat." } ], "question": "Is the sports ball above the baseball bat?", "question_id": "sealvqa_spatial_340", "search": 1, "target_instances": [ { "bbox": [ 436.9, 225.5, 11.15, 4.64 ], "instance_id": "", "name": "sports ball" }, { "bbox": [ 256.13, 196.58, 36.54, 27.1 ], "instance_id": "", "name": "baseball bat" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The remote is to the left of the chair." }
{ "answer": "The remote is to the left of the chair.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000298657.jpg", "index": "sealvqa_spatial_346", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the remote to the left or right of the chair?" }, { "from": "gpt", "value": "The remote is to the left of the chair." } ], "question": "Is the remote to the left or right of the chair?", "question_id": "sealvqa_spatial_346", "search": 1, "target_instances": [ { "bbox": [ 99.69, 215.72, 62.55, 127.99 ], "instance_id": "", "name": "chair" }, { "bbox": [ 0, 279.64, 41.55, 22.07 ], "instance_id": "", "name": "remote" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the tennis racket is below the bottle." }
{ "answer": "Yes, the tennis racket is below the bottle.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000276197.jpg", "index": "sealvqa_spatial_348", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the bottle on the left or right side of the tennis racket?" }, { "from": "gpt", "value": "The bottle is on the right side of the tennis racket." }, { "from": "human", "value": "Is the tennis racket below the bottle?" }, { "from": "gpt", "value": "Yes, the tennis racket is below the bottle." } ], "question": "Is the tennis racket below the bottle?", "question_id": "sealvqa_spatial_348", "search": 1, "target_instances": [ { "bbox": [ 211, 277.33, 62.19, 148.67 ], "instance_id": "", "name": "tennis racket" }, { "bbox": [ 538.11, 57.96, 16.06, 16.88 ], "instance_id": "", "name": "bottle" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The tie is below the sports ball." }
{ "answer": "The tie is below the sports ball.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000231748.jpg", "index": "sealvqa_spatial_355", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the tie on the left or right side of the sports ball?" }, { "from": "gpt", "value": "The tie is on the left side of the sports ball." }, { "from": "human", "value": "Is the tie above or below the sports ball?" }, { "from": "gpt", "value": "The tie is below the sports ball." } ], "question": "Is the tie above or below the sports ball?", "question_id": "sealvqa_spatial_355", "search": 1, "target_instances": [ { "bbox": [ 318.33, 72.75, 14.5, 13.96 ], "instance_id": "", "name": "sports ball" }, { "bbox": [ 236.72, 173.19, 10.07, 44.12 ], "instance_id": "", "name": "tie" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the knife is not below the spoon." }
{ "answer": "No, the knife is not below the spoon.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000041482.jpg", "index": "sealvqa_spatial_358", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the knife to the right of the spoon?" }, { "from": "gpt", "value": "Yes, the knife is to the right of the spoon." }, { "from": "human", "value": "Is the knife below the spoon?" }, { "from": "gpt", "value": "No, the knife is not below the spoon." } ], "question": "Is the knife below the spoon?", "question_id": "sealvqa_spatial_358", "search": 1, "target_instances": [ { "bbox": [ 297.36, 358.84, 87.75, 43.87 ], "instance_id": "", "name": "knife" }, { "bbox": [ 0, 427.43, 42.23, 36.72 ], "instance_id": "", "name": "spoon" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The remote is on the right side of the person." }
{ "answer": "The remote is on the right side of the person.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000480594.jpg", "index": "sealvqa_spatial_363", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the remote on the left or right side of the person?" }, { "from": "gpt", "value": "The remote is on the right side of the person." } ], "question": "Is the remote on the left or right side of the person?", "question_id": "sealvqa_spatial_363", "search": 1, "target_instances": [ { "bbox": [ 324.22, 183.94, 42.29, 194.55 ], "instance_id": "", "name": "person" }, { "bbox": [ 373.2, 193.12, 43.41, 9.4 ], "instance_id": "", "name": "remote" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the truck is not on the left side of the umbrella." }
{ "answer": "No, the truck is not on the left side of the umbrella.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000384027.jpg", "index": "sealvqa_spatial_368", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the truck on the left side of the umbrella?" }, { "from": "gpt", "value": "No, the truck is not on the left side of the umbrella." } ], "question": "Is the truck on the left side of the umbrella?", "question_id": "sealvqa_spatial_368", "search": 1, "target_instances": [ { "bbox": [ 115.74, 317.14, 22.56, 39.72 ], "instance_id": "", "name": "umbrella" }, { "bbox": [ 576, 239.95, 64, 90.24 ], "instance_id": "", "name": "truck" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the car is on the right side of the snowboard." }
{ "answer": "Yes, the car is on the right side of the snowboard.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000030371.jpg", "index": "sealvqa_spatial_375", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the car on the right side of the snowboard?" }, { "from": "gpt", "value": "Yes, the car is on the right side of the snowboard." } ], "question": "Is the car on the right side of the snowboard?", "question_id": "sealvqa_spatial_375", "search": 1, "target_instances": [ { "bbox": [ 133.03, 226.95, 158.36, 9.94 ], "instance_id": "", "name": "snowboard" }, { "bbox": [ 323.51, 230.72, 88.05, 32.09 ], "instance_id": "", "name": "car" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the toilet is below the sink." }
{ "answer": "Yes, the toilet is below the sink.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000050096.jpg", "index": "sealvqa_spatial_382", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the toilet on the left side of the sink?" }, { "from": "gpt", "value": "No, the toilet is not on the left side of the sink." }, { "from": "human", "value": "Is the toilet below the sink?" }, { "from": "gpt", "value": "Yes, the toilet is below the sink." } ], "question": "Is the toilet below the sink?", "question_id": "sealvqa_spatial_382", "search": 1, "target_instances": [ { "bbox": [ 403, 417.8, 21, 85.61 ], "instance_id": "", "name": "toilet" }, { "bbox": [ 0, 366.06, 88.41, 27.54 ], "instance_id": "", "name": "sink" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the cow is on the left side of the horse." }
{ "answer": "Yes, the cow is on the left side of the horse.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000416960.jpg", "index": "sealvqa_spatial_383", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the cow on the left side of the horse?" }, { "from": "gpt", "value": "Yes, the cow is on the left side of the horse." } ], "question": "Is the cow on the left side of the horse?", "question_id": "sealvqa_spatial_383", "search": 1, "target_instances": [ { "bbox": [ 474.08, 121.74, 71.11, 75.88 ], "instance_id": "", "name": "horse" }, { "bbox": [ 275.53, 73.56, 76.05, 54.86 ], "instance_id": "", "name": "cow" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The sink is on the right side of the vase." }
{ "answer": "The sink is on the right side of the vase.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000256189.jpg", "index": "sealvqa_spatial_388", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the sink on the left or right side of the vase?" }, { "from": "gpt", "value": "The sink is on the right side of the vase." } ], "question": "Is the sink on the left or right side of the vase?", "question_id": "sealvqa_spatial_388", "search": 1, "target_instances": [ { "bbox": [ 251.02, 210.23, 104.74, 68.99 ], "instance_id": "", "name": "sink" }, { "bbox": [ 123.33, 216.9, 39.81, 52.26 ], "instance_id": "", "name": "vase" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the cup is not to the right of the backpack." }
{ "answer": "No, the cup is not to the right of the backpack.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000332836.jpg", "index": "sealvqa_spatial_406", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the cup to the right of the backpack?" }, { "from": "gpt", "value": "No, the cup is not to the right of the backpack." } ], "question": "Is the cup to the right of the backpack?", "question_id": "sealvqa_spatial_406", "search": 1, "target_instances": [ { "bbox": [ 17.68, 230.2, 16.68, 21.73 ], "instance_id": "", "name": "cup" }, { "bbox": [ 342.76, 201.24, 56.6, 71.72 ], "instance_id": "", "name": "backpack" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the book is to the right of the remote." }
{ "answer": "Yes, the book is to the right of the remote.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000214848.jpg", "index": "sealvqa_spatial_423", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the book to the right of the remote?" }, { "from": "gpt", "value": "Yes, the book is to the right of the remote." } ], "question": "Is the book to the right of the remote?", "question_id": "sealvqa_spatial_423", "search": 1, "target_instances": [ { "bbox": [ 67.8, 287.18, 26.06, 8.86 ], "instance_id": "", "name": "remote" }, { "bbox": [ 241.14, 276.3, 45.55, 18.3 ], "instance_id": "", "name": "book" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the sandwich is not above the sink." }
{ "answer": "No, the sandwich is not above the sink.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000147545.jpg", "index": "sealvqa_spatial_425", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the sink on the left side of the sandwich?" }, { "from": "gpt", "value": "Yes, the sink is on the left side of the sandwich." }, { "from": "human", "value": "Is the sandwich above the sink?" }, { "from": "gpt", "value": "No, the sandwich is not above the sink." } ], "question": "Is the sandwich above the sink?", "question_id": "sealvqa_spatial_425", "search": 1, "target_instances": [ { "bbox": [ 555.87, 294.61, 25.14, 15.39 ], "instance_id": "", "name": "sandwich" }, { "bbox": [ 105.89, 272.48, 39.37, 5.3 ], "instance_id": "", "name": "sink" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the knife is not to the right of the spoon." }
{ "answer": "No, the knife is not to the right of the spoon.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000165803.jpg", "index": "sealvqa_spatial_432", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the knife to the right of the spoon?" }, { "from": "gpt", "value": "No, the knife is not to the right of the spoon." } ], "question": "Is the knife to the right of the spoon?", "question_id": "sealvqa_spatial_432", "search": 1, "target_instances": [ { "bbox": [ 110.76, 409.91, 33.75, 91.39 ], "instance_id": "", "name": "knife" }, { "bbox": [ 280.82, 491.81, 27.23, 33.99 ], "instance_id": "", "name": "spoon" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "The fork is above the microwave." }
{ "answer": "The fork is above the microwave.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000302077.jpg", "index": "sealvqa_spatial_434", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the fork on the left side of the microwave?" }, { "from": "gpt", "value": "Yes, the fork is on the left side of the microwave." }, { "from": "human", "value": "Is the fork above or below the microwave?" }, { "from": "gpt", "value": "The fork is above the microwave." } ], "question": "Is the fork above or below the microwave?", "question_id": "sealvqa_spatial_434", "search": 1, "target_instances": [ { "bbox": [ 439.76, 113.82, 2.72, 10.02 ], "instance_id": "", "name": "fork" }, { "bbox": [ 469, 124.26, 31, 32.99 ], "instance_id": "", "name": "microwave" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the car is not below the sports ball." }
{ "answer": "No, the car is not below the sports ball.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000474953.jpg", "index": "sealvqa_spatial_439", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the sports ball on the right side of the car?" }, { "from": "gpt", "value": "No, the sports ball is not on the right side of the car." }, { "from": "human", "value": "Is the car below the sports ball?" }, { "from": "gpt", "value": "No, the car is not below the sports ball." } ], "question": "Is the car below the sports ball?", "question_id": "sealvqa_spatial_439", "search": 1, "target_instances": [ { "bbox": [ 253.84, 124.22, 8.26, 8.49 ], "instance_id": "", "name": "sports ball" }, { "bbox": [ 621.2, 107.27, 18.8, 8.32 ], "instance_id": "", "name": "car" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the backpack is not to the left of the cup." }
{ "answer": "No, the backpack is not to the left of the cup.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000536609.jpg", "index": "sealvqa_spatial_445", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the backpack to the left of the cup?" }, { "from": "gpt", "value": "No, the backpack is not to the left of the cup." } ], "question": "Is the backpack to the left of the cup?", "question_id": "sealvqa_spatial_445", "search": 1, "target_instances": [ { "bbox": [ 143.48, 192.65, 7.47, 24.38 ], "instance_id": "", "name": "backpack" }, { "bbox": [ 7, 199.05, 14.39, 22.54 ], "instance_id": "", "name": "cup" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the bench is not below the baseball bat." }
{ "answer": "No, the bench is not below the baseball bat.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000363831.jpg", "index": "sealvqa_spatial_448", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the baseball bat to the left of the bench?" }, { "from": "gpt", "value": "No, the baseball bat is not to the left of the bench." }, { "from": "human", "value": "Is the bench below the baseball bat?" }, { "from": "gpt", "value": "No, the bench is not below the baseball bat." } ], "question": "Is the bench below the baseball bat?", "question_id": "sealvqa_spatial_448", "search": 1, "target_instances": [ { "bbox": [ 391.46, 235.21, 72.99, 13.66 ], "instance_id": "", "name": "baseball bat" }, { "bbox": [ 57.63, 8.54, 88.16, 49.59 ], "instance_id": "", "name": "bench" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the skis is not to the right of the tv." }
{ "answer": "No, the skis is not to the right of the tv.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000227226.jpg", "index": "sealvqa_spatial_449", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the skis to the right of the tv?" }, { "from": "gpt", "value": "No, the skis is not to the right of the tv." } ], "question": "Is the skis to the right of the tv?", "question_id": "sealvqa_spatial_449", "search": 1, "target_instances": [ { "bbox": [ 285.9, 163.31, 47.31, 42.93 ], "instance_id": "", "name": "tv" }, { "bbox": [ 126.08, 187.07, 15.04, 135.34 ], "instance_id": "", "name": "skis" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the truck is above the bench." }
{ "answer": "Yes, the truck is above the bench.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000250608.jpg", "index": "sealvqa_spatial_452", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the truck on the left or right side of the bench?" }, { "from": "gpt", "value": "The truck is on the left side of the bench." }, { "from": "human", "value": "Is the truck above the bench?" }, { "from": "gpt", "value": "Yes, the truck is above the bench." } ], "question": "Is the truck above the bench?", "question_id": "sealvqa_spatial_452", "search": 1, "target_instances": [ { "bbox": [ 572.8, 283.56, 27.06, 9.68 ], "instance_id": "", "name": "bench" }, { "bbox": [ 39.72, 241.83, 52.77, 33.06 ], "instance_id": "", "name": "truck" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "No, the bus is not on the right side of the car." }
{ "answer": "No, the bus is not on the right side of the car.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 2, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000338880.jpg", "index": "sealvqa_spatial_454", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the bus on the right side of the car?" }, { "from": "gpt", "value": "No, the bus is not on the right side of the car." } ], "question": "Is the bus on the right side of the car?", "question_id": "sealvqa_spatial_454", "search": 1, "target_instances": [ { "bbox": [ 221.81, 448.05, 61.38, 23.63 ], "instance_id": "", "name": "car" }, { "bbox": [ 66.34, 426.16, 86.93, 43.05 ], "instance_id": "", "name": "bus" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the motorcycle is above the book." }
{ "answer": "Yes, the motorcycle is above the book.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000080441.jpg", "index": "sealvqa_spatial_458", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the book to the right of the motorcycle?" }, { "from": "gpt", "value": "Yes, the book is to the right of the motorcycle." }, { "from": "human", "value": "Is the motorcycle above the book?" }, { "from": "gpt", "value": "Yes, the motorcycle is above the book." } ], "question": "Is the motorcycle above the book?", "question_id": "sealvqa_spatial_458", "search": 1, "target_instances": [ { "bbox": [ 545.6, 339.23, 6.22, 11.84 ], "instance_id": "", "name": "book" }, { "bbox": [ 316.46, 305.92, 9.28, 17.35 ], "instance_id": "", "name": "motorcycle" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
sealvqa_spatial
{ "ground_truth": "Yes, the bowl is below the sandwich." }
{ "answer": "Yes, the bowl is below the sandwich.", "bbox_format": "absolute_xywh", "bbox_format_description": "Bounding boxes use absolute pixel coordinates in [x, y, width, height] format", "conversation_turns": 4, "data_split": "train", "dataset_type": "spatial_relation", "has_targets": true, "image_path": "coco2017/train2017/000000079277.jpg", "index": "sealvqa_spatial_464", "original_conversations": [ { "from": "human", "value": "<image>\nAdditional visual information to focus on: <object>; <object>.\nIs the bowl to the left of the sandwich?" }, { "from": "gpt", "value": "No, the bowl is not to the left of the sandwich." }, { "from": "human", "value": "Is the bowl below the sandwich?" }, { "from": "gpt", "value": "Yes, the bowl is below the sandwich." } ], "question": "Is the bowl below the sandwich?", "question_id": "sealvqa_spatial_464", "search": 1, "target_instances": [ { "bbox": [ 113.36, 377.33, 116.24, 84.66 ], "instance_id": "", "name": "sandwich" }, { "bbox": [ 330.21, 620.12, 96.79, 19.88 ], "instance_id": "", "name": "bowl" } ] }
[ { "content": "You are a helpful assistant.\n\n# Tools\nYou may call one or more functions to assist with the user query.\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>\n{\"type\":\"function\",\"function\":{\"name\":\"image_zoom_in_tool\",\"description\":\"Zoom in on a speci...
sealvqa_spatial_spatial
visual_toolbox_v2
End of preview. Expand in Data Studio
README.md exists but content is empty.
Downloads last month
11