{"id": 0, "image": ["rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_2/frame_0.png", "rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_9/frame_0.png"], "height": [128, 128], "width": [128, 128], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the cloth."}, {"from": "gpt", "value": "[[784, 312], [688, 344], [575, 352], [496, 320]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the grapes."}, {"from": "gpt", "value": "[[900, 353], [771, 344], [632, 278], [500, 195]]"}]}
{"id": 2, "image": ["rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_17/frame_0.png", "rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_17/frame_15.png"], "height": [128, 128], "width": [128, 128], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the croissant."}, {"from": "gpt", "value": "[[824, 261], [765, 315], [723, 391], [697, 466]]"}, {"from": "human", "value": "Image-2:\n You need to move the croissant on the plate."}, {"from": "gpt", "value": "[[655, 357], [513, 332], [412, 332], [319, 357]]"}]}
{"id": 4, "image": ["rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_20/frame_0.png", "rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_29/frame_0.png"], "height": [128, 128], "width": [128, 128], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the cloth."}, {"from": "gpt", "value": "[[789, 312], [737, 352], [643, 374], [532, 365]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the grapes."}, {"from": "gpt", "value": "[[797, 327], [717, 313], [650, 333], [602, 354]]"}]}
{"id": 6, "image": ["rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_29/frame_15.png", "rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_34/frame_0.png"], "height": [128, 128], "width": [128, 128], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move the grapes towards the plate."}, {"from": "gpt", "value": "[[437, 276], [352, 313], [297, 378], [250, 472]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the edge of the cloth."}, {"from": "gpt", "value": "[[769, 288], [669, 353], [579, 487], [523, 665]]"}]}
{"id": 8, "image": ["rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_36/frame_0.png", "rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_37/frame_0.png"], "height": [128, 128], "width": [128, 128], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the cloth."}, {"from": "gpt", "value": "[[774, 308], [712, 353], [622, 389], [498, 400]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the cloth."}, {"from": "gpt", "value": "[[776, 305], [718, 367], [647, 404], [523, 436]]"}]}
{"id": 10, "image": ["rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_56/frame_0.png", "rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_56/frame_15.png"], "height": [128, 128], "width": [128, 128], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the croissant."}, {"from": "gpt", "value": "[[748, 321], [695, 331], [656, 352], [618, 404]]"}, {"from": "human", "value": "Image-2:\n You need to place the croissant on the plate."}, {"from": "gpt", "value": "[[581, 346], [500, 323], [423, 346], [394, 372]]"}]}
{"id": 12, "image": ["rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_70/frame_0.png", "rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_71/frame_0.png"], "height": [128, 128], "width": [128, 128], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the grapes."}, {"from": "gpt", "value": "[[917, 353], [793, 372], [669, 421], [553, 515]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the grapes."}, {"from": "gpt", "value": "[[828, 319], [739, 408], [672, 475], [588, 559]]"}]}
{"id": 14, "image": ["rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_71/frame_15.png", "rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_75/frame_0.png"], "height": [128, 128], "width": [128, 128], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move the grapes towards the plate."}, {"from": "gpt", "value": "[[429, 265], [370, 357], [319, 399], [290, 458]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the corner of the cloth."}, {"from": "gpt", "value": "[[789, 308], [761, 350], [707, 383], [639, 391]]"}]}
{"id": 16, "image": ["rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_76/frame_15.png", "rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_85/frame_0.png"], "height": [128, 128], "width": [128, 128], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to place the croissant on the plate."}, {"from": "gpt", "value": "[[709, 216], [633, 274], [530, 308], [402, 329]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the corner of the cloth."}, {"from": "gpt", "value": "[[793, 316], [692, 291], [613, 244], [524, 162]]"}]}
{"id": 18, "image": ["rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_93/frame_0.png", "rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_93/frame_15.png"], "height": [128, 128], "width": [128, 128], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the croissant."}, {"from": "gpt", "value": "[[835, 344], [797, 365], [759, 365], [714, 338]]"}, {"from": "human", "value": "Image-2:\n You need to place the croissant on the plate."}, {"from": "gpt", "value": "[[650, 195], [573, 212], [509, 258], [434, 357]]"}]}
{"id": 20, "image": ["rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_106/frame_0.png", "rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_106/frame_15.png"], "height": [128, 128], "width": [128, 128], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the grapes."}, {"from": "gpt", "value": "[[807, 332], [739, 357], [693, 399], [664, 462]]"}, {"from": "human", "value": "Image-2:\n You need to place the grapes on the plate."}, {"from": "gpt", "value": "[[445, 492], [353, 487], [286, 504], [269, 550]]"}]}
{"id": 22, "image": ["rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_107/frame_0.png", "rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_107/frame_15.png"], "height": [128, 128], "width": [128, 128], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the grapes."}, {"from": "gpt", "value": "[[798, 307], [693, 218], [580, 193], [462, 210]]"}, {"from": "human", "value": "Image-2:\n You need to move the grapes towards the plate."}, {"from": "gpt", "value": "[[454, 139], [382, 223], [319, 328], [303, 479]]"}]}
{"id": 24, "image": ["rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_110/frame_0.png", "rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_114/frame_0.png"], "height": [128, 128], "width": [128, 128], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the corner of the cloth."}, {"from": "gpt", "value": "[[705, 275], [659, 303], [627, 350], [604, 425]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the croissant."}, {"from": "gpt", "value": "[[798, 248], [714, 307], [664, 391], [618, 500]]"}]}
{"id": 26, "image": ["rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_114/frame_15.png", "rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_141/frame_15.png"], "height": [128, 128], "width": [128, 128], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to place the croissant on the plate."}, {"from": "gpt", "value": "[[588, 349], [521, 332], [454, 332], [395, 361]]"}, {"from": "human", "value": "Image-2:\n You need to place the grapes on the plate."}, {"from": "gpt", "value": "[[811, 286], [640, 304], [487, 367], [340, 484]]"}]}
{"id": 28, "image": ["rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_145/frame_0.png", "rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_145/frame_15.png"], "height": [128, 128], "width": [128, 128], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the croissant."}, {"from": "gpt", "value": "[[798, 277], [748, 319], [697, 366], [605, 429]]"}, {"from": "human", "value": "Image-2:\n You need to place the croissant on the plate."}, {"from": "gpt", "value": "[[555, 303], [445, 340], [361, 374], [298, 433]]"}]}
{"id": 30, "image": ["rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_158/frame_0.png", "rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_174/frame_0.png"], "height": [128, 128], "width": [128, 128], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the cloth."}, {"from": "gpt", "value": "[[765, 239], [622, 336], [538, 399], [445, 483]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the grapes."}, {"from": "gpt", "value": "[[808, 345], [773, 359], [740, 408], [718, 510]]"}]}
{"id": 32, "image": ["rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_174/frame_15.png", "rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_180/frame_0.png"], "height": [128, 128], "width": [128, 128], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move the grapes towards the plate."}, {"from": "gpt", "value": "[[504, 433], [408, 419], [337, 452], [318, 499]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the cloth."}, {"from": "gpt", "value": "[[769, 239], [630, 324], [542, 382], [496, 450]]"}]}
{"id": 34, "image": ["rtx_frames_success_0/10_utokyo_pr2_tabletop_manipulation_converted_externally_to_rlds#episode_189/frame_0.png", "rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_0/frame_0.png"], "height": [128, 224], "width": [128, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the grapes."}, {"from": "gpt", "value": "[[805, 323], [748, 327], [671, 356], [634, 380]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the white plate."}, {"from": "gpt", "value": "[[598, 362], [608, 411], [608, 472], [593, 522], [577, 569], [555, 608], [504, 667]]"}]}
{"id": 36, "image": ["rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_0/frame_15.png", "rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_7/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to place the white plate on the red plate."}, {"from": "gpt", "value": "[[341, 602], [264, 679], [163, 738], [79, 764]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the white plate."}, {"from": "gpt", "value": "[[483, 370], [560, 487], [594, 624], [596, 784]]"}]}
{"id": 38, "image": ["rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_7/frame_15.png", "rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_10/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to place the white plate on the red plate."}, {"from": "gpt", "value": "[[630, 650], [538, 707], [408, 748], [269, 759]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the white plate."}, {"from": "gpt", "value": "[[392, 352], [439, 441], [478, 535], [510, 624], [514, 691], [516, 752], [508, 827]]"}]}
{"id": 40, "image": ["rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_10/frame_15.png", "rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_11/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to place the white plate on the red plate."}, {"from": "gpt", "value": "[[522, 852], [472, 858], [415, 862], [358, 856], [264, 829], [219, 817], [179, 785]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the white plate."}, {"from": "gpt", "value": "[[432, 342], [541, 421], [624, 523], [703, 688]]"}]}
{"id": 42, "image": ["rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_11/frame_15.png", "rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_12/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to place the white plate on the red plate."}, {"from": "gpt", "value": "[[553, 579], [449, 647], [316, 688], [182, 695]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the white plate."}, {"from": "gpt", "value": "[[459, 391], [551, 502], [622, 624], [669, 786]]"}]}
{"id": 44, "image": ["rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_19/frame_0.png", "rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_25/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the white plate."}, {"from": "gpt", "value": "[[424, 366], [424, 466], [429, 584], [387, 676]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the white bowl."}, {"from": "gpt", "value": "[[388, 398], [445, 472], [492, 565], [522, 624], [555, 679], [571, 740], [563, 815], [557, 854]]"}]}
{"id": 46, "image": ["rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_30/frame_0.png", "rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_30/frame_15.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the white plate."}, {"from": "gpt", "value": "[[622, 366], [632, 431], [632, 486], [624, 549], [608, 585], [591, 628], [563, 657], [528, 681]]"}, {"from": "human", "value": "Image-2:\n You need to place the white plate on the red plate."}, {"from": "gpt", "value": "[[490, 524], [449, 573], [421, 602], [394, 640], [364, 663], [317, 695], [256, 738]]"}]}
{"id": 48, "image": ["rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_31/frame_0.png", "rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_34/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the white plate."}, {"from": "gpt", "value": "[[312, 460], [428, 514], [526, 590], [587, 737]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the white plate."}, {"from": "gpt", "value": "[[391, 340], [445, 462], [483, 609], [508, 786]]"}]}
{"id": 50, "image": ["rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_34/frame_15.png", "rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_36/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to place the white plate on the red plate."}, {"from": "gpt", "value": "[[294, 592], [206, 664], [118, 735], [42, 807]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the white plate."}, {"from": "gpt", "value": "[[420, 357], [424, 479], [445, 576], [454, 702]]"}]}
{"id": 52, "image": ["rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_36/frame_15.png", "rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_38/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to place the white plate on the red plate."}, {"from": "gpt", "value": "[[471, 521], [328, 559], [210, 643], [105, 739]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the white plate."}, {"from": "gpt", "value": "[[433, 353], [437, 445], [471, 559], [513, 702]]"}]}
{"id": 54, "image": ["rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_38/frame_15.png", "rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_39/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to place the white plate on the red plate."}, {"from": "gpt", "value": "[[408, 508], [277, 567], [202, 626], [113, 786]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the white plate."}, {"from": "gpt", "value": "[[374, 340], [391, 462], [416, 592], [437, 752]]"}]}
{"id": 56, "image": ["rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_39/frame_15.png", "rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_42/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to place the white plate on the red plate."}, {"from": "gpt", "value": "[[555, 718], [433, 752], [298, 782], [101, 811]]"}, {"from": "human", "value": "Image-2:\n You need to approach the white plate."}, {"from": "gpt", "value": "[[466, 294], [437, 466], [424, 643], [412, 794]]"}]}
{"id": 58, "image": ["rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_42/frame_15.png", "rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_61/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to place the white plate on the red plate."}, {"from": "gpt", "value": "[[235, 508], [118, 605], [67, 651], [34, 752]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the white plate."}, {"from": "gpt", "value": "[[357, 286], [424, 412], [487, 538], [521, 668]]"}]}
{"id": 60, "image": ["rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_61/frame_15.png", "rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_62/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to place the white plate on the red plate."}, {"from": "gpt", "value": "[[584, 601], [429, 660], [277, 718], [109, 777]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the white plate."}, {"from": "gpt", "value": "[[346, 329], [434, 427], [487, 549], [513, 722]]"}]}
{"id": 62, "image": ["rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_62/frame_15.png", "rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_66/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to place the white plate on the red plate."}, {"from": "gpt", "value": "[[402, 673], [331, 714], [216, 756], [100, 776]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the white plate."}, {"from": "gpt", "value": "[[378, 340], [466, 487], [559, 622], [630, 769]]"}]}
{"id": 64, "image": ["rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_66/frame_15.png", "rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_70/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to place the white plate on the red plate."}, {"from": "gpt", "value": "[[546, 727], [429, 769], [303, 819], [155, 853]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the white plate."}, {"from": "gpt", "value": "[[380, 417], [421, 498], [444, 598], [455, 737]]"}]}
{"id": 66, "image": ["rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_70/frame_15.png", "rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_72/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to place the white plate on the red plate."}, {"from": "gpt", "value": "[[368, 588], [333, 662], [246, 729], [130, 782]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the white plate."}, {"from": "gpt", "value": "[[412, 332], [445, 450], [454, 534], [462, 639]]"}]}
{"id": 68, "image": ["rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_73/frame_0.png", "rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_78/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the white plate."}, {"from": "gpt", "value": "[[353, 408], [366, 584], [370, 723], [374, 836]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the white plate."}, {"from": "gpt", "value": "[[487, 336], [450, 559], [399, 685], [336, 853]]"}]}
{"id": 70, "image": ["rtx_frames_success_0/11_utokyo_xarm_pick_and_place_converted_externally_to_rlds#episode_80/frame_0.png", "rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_1/frame_0.png"], "height": [224, 480], "width": [224, 640], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the white plate."}, {"from": "gpt", "value": "[[521, 445], [479, 559], [437, 630], [387, 718]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the bowl."}, {"from": "gpt", "value": "[[342, 697], [447, 642], [581, 612], [699, 627]]"}]}
{"id": 72, "image": ["rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_15/frame_0.png", "rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_15/frame_15.png"], "height": [480, 480], "width": [640, 640], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the microwave handle."}, {"from": "gpt", "value": "[[336, 697], [464, 599], [596, 454], [695, 273]]"}, {"from": "human", "value": "Image-2:\n You need to open the cabinet door."}, {"from": "gpt", "value": "[[703, 276], [673, 303], [630, 313], [596, 296]]"}]}
{"id": 74, "image": ["rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_16/frame_0.png", "rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_24/frame_0.png"], "height": [480, 480], "width": [640, 640], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the cabinet door handle."}, {"from": "gpt", "value": "[[350, 699], [436, 762], [581, 822], [739, 852]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the bowl."}, {"from": "gpt", "value": "[[338, 652], [370, 679], [425, 694], [461, 682]]"}]}
{"id": 76, "image": ["rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_24/frame_15.png", "rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_33/frame_0.png"], "height": [480, 480], "width": [640, 640], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to place the bowl in the below cabinet."}, {"from": "gpt", "value": "[[461, 684], [506, 767], [570, 842], [656, 912]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the microwave handle."}, {"from": "gpt", "value": "[[346, 714], [468, 617], [590, 446], [673, 278]]"}]}
{"id": 78, "image": ["rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_39/frame_0.png", "rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_49/frame_0.png"], "height": [480, 480], "width": [640, 640], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the cabinet door handle."}, {"from": "gpt", "value": "[[336, 712], [427, 774], [538, 817], [686, 855]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the teapot."}, {"from": "gpt", "value": "[[333, 612], [481, 566], [628, 549], [789, 551]]"}]}
{"id": 80, "image": ["rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_54/frame_0.png", "rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_54/frame_15.png"], "height": [480, 480], "width": [640, 640], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the bowl."}, {"from": "gpt", "value": "[[325, 697], [380, 689], [423, 652], [462, 607]]"}, {"from": "human", "value": "Image-2:\n You need to place the bowl in the below cabinet."}, {"from": "gpt", "value": "[[517, 622], [564, 682], [600, 810], [598, 937]]"}]}
{"id": 82, "image": ["rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_61/frame_0.png", "rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_61/frame_15.png"], "height": [480, 480], "width": [640, 640], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the bowl."}, {"from": "gpt", "value": "[[346, 684], [451, 622], [603, 596], [786, 619]]"}, {"from": "human", "value": "Image-2:\n You need to place the bowl in the below cabinet."}, {"from": "gpt", "value": "[[788, 617], [716, 672], [664, 779], [639, 915]]"}]}
{"id": 84, "image": ["rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_79/frame_0.png", "rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_80/frame_0.png"], "height": [480, 480], "width": [640, 640], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the microwave handle."}, {"from": "gpt", "value": "[[342, 697], [491, 554], [566, 456], [673, 278]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the bowl."}, {"from": "gpt", "value": "[[336, 697], [402, 689], [477, 654], [541, 609]]"}]}
{"id": 86, "image": ["rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_86/frame_0.png", "rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_114/frame_0.png"], "height": [480, 480], "width": [640, 640], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the microwave handle."}, {"from": "gpt", "value": "[[327, 682], [477, 589], [613, 426], [694, 281]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the bowl."}, {"from": "gpt", "value": "[[338, 702], [387, 649], [462, 617], [562, 612]]"}]}
{"id": 88, "image": ["rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_114/frame_15.png", "rtx_frames_success_0/12_ucsd_kitchen_dataset_converted_externally_to_rlds#episode_132/frame_0.png"], "height": [480, 480], "width": [640, 640], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to place the bowl in the below cabinet."}, {"from": "gpt", "value": "[[549, 551], [543, 662], [575, 817], [620, 930]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the cabinet door handle."}, {"from": "gpt", "value": "[[346, 694], [417, 744], [555, 815], [690, 865]]"}]}
{"id": 90, "image": ["rtx_frames_success_0/15_dlr_sara_grid_clamp_converted_externally_to_rlds#episode_64/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_15/frame_0.png"], "height": [480, 224], "width": [640, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to hover over the grid clamp."}, {"from": "gpt", "value": "[[277, 746], [271, 757], [270, 774], [275, 786]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[193, 399], [231, 433], [261, 466], [269, 546]]"}]}
{"id": 92, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_18/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_19/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[203, 437], [327, 459], [443, 474], [531, 502], [563, 549]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[219, 435], [278, 465], [344, 492], [402, 553]]"}]}
{"id": 94, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_41/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_49/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[214, 378], [315, 483], [395, 571], [454, 702]]"}, {"from": "human", "value": "Image-2:\n You need to hover over the red object."}, {"from": "gpt", "value": "[[222, 421], [254, 463], [272, 522], [276, 575]]"}]}
{"id": 96, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_56/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_75/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[210, 382], [282, 458], [353, 538], [454, 706]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[210, 374], [235, 466], [252, 542], [273, 660]]"}]}
{"id": 98, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_76/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_135/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[197, 370], [202, 458], [210, 567], [210, 685]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[197, 429], [246, 455], [313, 488], [415, 531], [476, 565], [522, 620], [537, 663]]"}]}
{"id": 100, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_145/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_151/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[219, 425], [325, 463], [455, 547], [514, 620]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[256, 395], [340, 441], [416, 500], [504, 609]]"}]}
{"id": 102, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_169/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_180/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red hexagonal prism."}, {"from": "gpt", "value": "[[218, 403], [324, 487], [433, 576], [538, 693]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[215, 437], [232, 500], [252, 583], [246, 636]]"}]}
{"id": 104, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_186/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_187/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[211, 437], [278, 455], [348, 488], [411, 522], [478, 559], [522, 602], [565, 656]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[217, 429], [317, 486], [388, 531], [425, 583], [449, 634]]"}]}
{"id": 106, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_195/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_251/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[209, 447], [244, 494], [278, 567], [283, 638]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[211, 408], [368, 447], [492, 509], [624, 605]]"}]}
{"id": 108, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_257/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_264/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[195, 414], [338, 470], [464, 541], [575, 658]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red block."}, {"from": "gpt", "value": "[[214, 414], [274, 453], [297, 511], [297, 590]]"}]}
{"id": 110, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_268/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_272/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[207, 414], [244, 459], [259, 508], [254, 583]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[201, 423], [195, 487], [211, 545], [244, 586]]"}]}
{"id": 112, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_278/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_286/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[197, 409], [272, 470], [317, 531], [348, 583], [364, 648]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[207, 443], [287, 490], [339, 541], [372, 581], [392, 634], [396, 691]]"}]}
{"id": 114, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_296/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_301/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[191, 421], [228, 478], [246, 528], [268, 589], [270, 650]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[217, 427], [327, 437], [445, 478], [516, 541], [569, 610]]"}]}
{"id": 116, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_304/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_316/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[201, 435], [380, 445], [490, 480], [579, 549]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[197, 427], [335, 453], [459, 478], [579, 553]]"}]}
{"id": 118, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_325/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_337/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[199, 415], [236, 484], [238, 571], [220, 634]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[211, 415], [238, 506], [264, 587], [281, 634]]"}]}
{"id": 120, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_340/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_349/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[215, 433], [281, 461], [348, 516], [382, 565], [390, 638]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[164, 399], [336, 454], [458, 534], [588, 693]]"}]}
{"id": 122, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_371/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_394/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[197, 374], [328, 408], [420, 441], [529, 517]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red block."}, {"from": "gpt", "value": "[[202, 387], [315, 458], [395, 550], [399, 685]]"}]}
{"id": 124, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_424/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_442/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[222, 425], [344, 470], [463, 530], [557, 591], [612, 691]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[207, 419], [215, 457], [220, 514], [207, 585]]"}]}
{"id": 126, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_470/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_471/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[207, 447], [358, 498], [474, 553], [518, 628]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[215, 439], [246, 492], [274, 547], [281, 602]]"}]}
{"id": 128, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_490/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_509/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[205, 439], [281, 472], [396, 533], [431, 596]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[202, 366], [227, 445], [235, 559], [261, 655]]"}]}
{"id": 130, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_511/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_516/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to hover over the red block."}, {"from": "gpt", "value": "[[231, 349], [252, 408], [277, 441], [277, 521]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[217, 421], [281, 453], [321, 478], [364, 518]]"}]}
{"id": 132, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_526/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_537/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[227, 374], [298, 408], [370, 450], [479, 525]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[222, 441], [333, 453], [421, 478], [516, 514], [539, 567]]"}]}
{"id": 134, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_540/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_549/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[217, 421], [278, 447], [325, 476], [382, 512], [435, 553], [457, 602]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[205, 404], [408, 461], [564, 534], [726, 658]]"}]}
{"id": 136, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_551/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_553/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[199, 417], [274, 464], [335, 541], [370, 643]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[227, 382], [353, 420], [492, 466], [622, 559]]"}]}
{"id": 138, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_555/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_573/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the pot."}, {"from": "gpt", "value": "[[162, 164], [199, 227], [235, 267], [301, 305]]"}, {"from": "human", "value": "Image-2:\n You need to hover over the red object."}, {"from": "gpt", "value": "[[206, 378], [303, 433], [403, 492], [525, 613]]"}]}
{"id": 140, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_582/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_602/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[218, 400], [188, 459], [205, 508], [244, 553]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[212, 419], [256, 442], [282, 474], [282, 545]]"}]}
{"id": 142, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_611/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_614/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[209, 427], [311, 467], [374, 537], [374, 569]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[201, 437], [305, 467], [372, 498], [433, 531], [449, 594]]"}]}
{"id": 144, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_615/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_619/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[213, 431], [299, 482], [384, 533], [482, 587], [541, 656]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[207, 415], [301, 437], [376, 461], [449, 508]]"}]}
{"id": 146, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_634/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_643/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[203, 439], [281, 476], [339, 528], [398, 626]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[207, 421], [331, 486], [380, 569], [407, 659]]"}]}
{"id": 148, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_674/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_674/frame_15.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the pot on the stove."}, {"from": "gpt", "value": "[[126, 177], [209, 211], [265, 263], [318, 344]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the pot on the stove."}, {"from": "gpt", "value": "[[348, 333], [427, 314], [526, 318], [618, 342]]"}]}
{"id": 150, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_678/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_689/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[223, 395], [319, 424], [361, 458], [399, 546]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[206, 391], [328, 450], [429, 517], [546, 609]]"}]}
{"id": 152, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_726/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_766/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[210, 370], [277, 424], [307, 496], [315, 571]]"}, {"from": "human", "value": "Image-2:\n You need to hover over the red object."}, {"from": "gpt", "value": "[[211, 414], [376, 466], [528, 551], [694, 724]]"}]}
{"id": 154, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_775/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_780/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the red object."}, {"from": "gpt", "value": "[[192, 414], [218, 519], [297, 613], [432, 688]]"}, {"from": "human", "value": "Image-2:\n You need to hover over the red object."}, {"from": "gpt", "value": "[[218, 432], [299, 427], [378, 449], [453, 496]]"}]}
{"id": 156, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_780/frame_15.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_785/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to lower towards the red object."}, {"from": "gpt", "value": "[[472, 528], [549, 547], [600, 585], [635, 648]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[220, 410], [382, 432], [523, 491], [707, 615]]"}]}
{"id": 158, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_788/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_800/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[218, 425], [323, 432], [449, 468], [560, 524]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[234, 448], [379, 477], [506, 523], [645, 639]]"}]}
{"id": 160, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_814/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_833/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[202, 387], [244, 466], [277, 550], [319, 676]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[201, 437], [262, 494], [299, 553], [325, 587]]"}]}
{"id": 162, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_835/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_861/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[209, 427], [313, 514], [364, 557], [427, 610], [480, 679]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[215, 429], [232, 482], [256, 581], [246, 654]]"}]}
{"id": 164, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_881/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_883/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[218, 425], [301, 468], [359, 551], [376, 675]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[199, 414], [323, 481], [415, 556], [504, 667]]"}]}
{"id": 166, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_919/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_922/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[195, 419], [289, 444], [352, 483], [414, 555]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[207, 425], [305, 459], [376, 521], [444, 609]]"}]}
{"id": 168, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_926/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_942/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the pot on the stove."}, {"from": "gpt", "value": "[[139, 139], [202, 189], [244, 235], [298, 298]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[193, 423], [262, 476], [305, 530], [343, 600], [368, 683]]"}]}
{"id": 170, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_946/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_950/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[201, 429], [325, 453], [437, 510], [520, 581], [585, 665]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[195, 427], [323, 459], [407, 492], [474, 563]]"}]}
{"id": 172, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_962/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_972/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[197, 439], [319, 445], [409, 472], [455, 506]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[197, 439], [323, 467], [461, 516], [600, 565], [693, 622]]"}]}
{"id": 174, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_974/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_983/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[222, 433], [376, 474], [455, 502], [526, 531], [581, 565], [616, 612]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[220, 421], [281, 486], [309, 573], [333, 632]]"}]}
{"id": 176, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1003/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1008/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[189, 415], [246, 465], [303, 508], [333, 585]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[205, 417], [335, 492], [421, 559], [470, 687]]"}]}
{"id": 178, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1044/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1056/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[185, 378], [248, 458], [282, 517], [298, 609]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[193, 366], [315, 424], [403, 483], [496, 601]]"}]}
{"id": 180, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1071/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1103/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[202, 374], [340, 458], [471, 542], [580, 702]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[213, 435], [321, 465], [394, 514], [457, 596]]"}]}
{"id": 182, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1117/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1122/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[193, 366], [223, 462], [256, 525], [294, 626]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the pot on the stove."}, {"from": "gpt", "value": "[[134, 147], [189, 206], [210, 235], [244, 303]]"}]}
{"id": 184, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1126/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1134/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[209, 410], [289, 451], [359, 521], [398, 618]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red block."}, {"from": "gpt", "value": "[[219, 437], [319, 437], [411, 467], [486, 514], [577, 593]]"}]}
{"id": 186, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1140/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1152/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[172, 408], [303, 483], [437, 559], [643, 676]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[214, 374], [214, 437], [214, 496], [231, 550]]"}]}
{"id": 188, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1168/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1168/frame_15.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the pot on the stove."}, {"from": "gpt", "value": "[[124, 175], [199, 230], [246, 276], [260, 305]]"}, {"from": "human", "value": "Image-2:\n You need to place the pot in the sink."}, {"from": "gpt", "value": "[[524, 289], [594, 311], [642, 352], [693, 415]]"}]}
{"id": 190, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1173/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1196/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[213, 427], [289, 449], [360, 504], [388, 591]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[199, 425], [307, 472], [406, 539], [476, 598], [516, 661]]"}]}
{"id": 192, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1202/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1209/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[203, 423], [352, 467], [449, 528], [502, 628]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[193, 419], [268, 463], [319, 531], [352, 632]]"}]}
{"id": 194, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1220/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1254/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red block."}, {"from": "gpt", "value": "[[206, 387], [244, 450], [265, 500], [265, 571]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[214, 387], [311, 458], [370, 550], [403, 714]]"}]}
{"id": 196, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1263/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1284/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[235, 382], [315, 420], [378, 500], [382, 584]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[197, 361], [332, 416], [437, 483], [534, 592]]"}]}
{"id": 198, "image": ["rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1344/frame_0.png", "rtx_frames_success_0/19_ucsd_pick_and_place_dataset_converted_externally_to_rlds#episode_1348/frame_0.png"], "height": [224, 224], "width": [224, 224], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[220, 429], [201, 459], [197, 510], [199, 608]]"}, {"from": "human", "value": "Image-2:\n You need to move towards the red object."}, {"from": "gpt", "value": "[[193, 433], [285, 484], [388, 547], [439, 612], [453, 679]]"}]}
{"id": 200, "image": ["rtx_frames_success_0/25_nyu_door_opening_surprising_effectiveness#episode_1/frame_0.png", "rtx_frames_success_0/25_nyu_door_opening_surprising_effectiveness#episode_13/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door handles."}, {"from": "gpt", "value": "[[427, 982], [412, 772], [447, 388], [500, 173]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door handles."}, {"from": "gpt", "value": "[[427, 982], [357, 797], [316, 584], [327, 421]]"}]}
{"id": 202, "image": ["rtx_frames_success_0/25_nyu_door_opening_surprising_effectiveness#episode_103/frame_0.png", "rtx_frames_success_0/25_nyu_door_opening_surprising_effectiveness#episode_107/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[232, 702], [211, 629], [223, 579], [261, 551]]"}, {"from": "human", "value": "Image-2:\n You need to approach the trash can door."}, {"from": "gpt", "value": "[[261, 657], [354, 500], [455, 421], [539, 382]]"}]}
{"id": 204, "image": ["rtx_frames_success_0/25_nyu_door_opening_surprising_effectiveness#episode_115/frame_0.png", "rtx_frames_success_0/25_nyu_door_opening_surprising_effectiveness#episode_116/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door handles."}, {"from": "gpt", "value": "[[408, 992], [389, 878], [379, 732], [387, 628], [418, 534]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door handles."}, {"from": "gpt", "value": "[[436, 992], [430, 927], [462, 837], [506, 782]]"}]}
{"id": 206, "image": ["rtx_frames_success_0/25_nyu_door_opening_surprising_effectiveness#episode_127/frame_0.png", "rtx_frames_success_0/25_nyu_door_opening_surprising_effectiveness#episode_128/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet handle."}, {"from": "gpt", "value": "[[442, 992], [414, 912], [414, 812], [436, 717]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[429, 990], [453, 887], [536, 744], [620, 639]]"}]}
{"id": 208, "image": ["rtx_frames_success_0/25_nyu_door_opening_surprising_effectiveness#episode_136/frame_0.png", "rtx_frames_success_0/25_nyu_door_opening_surprising_effectiveness#episode_144/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[427, 987], [461, 850], [532, 732], [626, 659]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[388, 955], [388, 899], [421, 820], [430, 747]]"}]}
{"id": 210, "image": ["rtx_frames_success_0/25_nyu_door_opening_surprising_effectiveness#episode_148/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_25/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[434, 949], [383, 758], [333, 567], [299, 343]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet handles."}, {"from": "gpt", "value": "[[472, 780], [463, 730], [443, 653], [423, 598]]"}]}
{"id": 212, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_28/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_34/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door handle."}, {"from": "gpt", "value": "[[449, 826], [492, 680], [573, 541], [695, 405]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[460, 780], [449, 649], [420, 525]]"}]}
{"id": 214, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_42/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_44/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[452, 749], [408, 602], [347, 490], [301, 398], [266, 197]]"}, {"from": "human", "value": "Image-2:\n You need to reach for the handles."}, {"from": "gpt", "value": "[[397, 969], [371, 876], [356, 776], [374, 680], [414, 510]]"}]}
{"id": 216, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_49/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_60/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[463, 749], [437, 676], [405, 595], [365, 514]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[446, 784], [420, 718], [388, 656], [347, 602]]"}]}
{"id": 218, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_61/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_64/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[403, 977], [376, 865], [379, 737], [397, 660], [434, 564], [475, 514]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door handles."}, {"from": "gpt", "value": "[[472, 741], [446, 595], [403, 471], [397, 317]]"}]}
{"id": 220, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_67/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_70/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door handle."}, {"from": "gpt", "value": "[[469, 680], [463, 579], [463, 486], [414, 324]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door handles."}, {"from": "gpt", "value": "[[365, 981], [342, 819], [350, 726], [417, 579]]"}]}
{"id": 222, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_72/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_79/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[423, 819], [423, 683], [423, 571], [414, 467]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[449, 799], [504, 753], [565, 714], [614, 664]]"}]}
{"id": 224, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_156/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_164/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[434, 896], [452, 822], [446, 737], [411, 699]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[484, 710], [458, 602], [426, 479], [350, 255]]"}]}
{"id": 226, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_165/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_166/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[443, 846], [492, 726], [553, 687]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door handles."}, {"from": "gpt", "value": "[[397, 988], [403, 772], [414, 633], [501, 440], [611, 305]]"}]}
{"id": 228, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_167/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_169/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet handle."}, {"from": "gpt", "value": "[[434, 784], [458, 668], [417, 606]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[437, 718], [437, 606], [455, 448], [489, 305], [515, 216]]"}]}
{"id": 230, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_172/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_174/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[452, 826], [394, 722], [417, 622], [507, 541]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[478, 757], [515, 660], [553, 575], [602, 521]]"}]}
{"id": 232, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_175/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_192/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door handles."}, {"from": "gpt", "value": "[[411, 977], [362, 811], [336, 637], [333, 483], [397, 170]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[403, 799], [408, 645], [382, 517], [290, 340]]"}]}
{"id": 234, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_196/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_197/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door handles."}, {"from": "gpt", "value": "[[414, 985], [426, 927], [458, 880], [495, 834], [536, 799], [597, 749]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door handle."}, {"from": "gpt", "value": "[[408, 861], [388, 753], [434, 645], [510, 602], [585, 645]]"}]}
{"id": 236, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_201/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_202/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[460, 761], [475, 618], [507, 463], [542, 328]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[463, 734], [492, 591], [524, 479], [573, 359]]"}]}
{"id": 238, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_213/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_223/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door handles."}, {"from": "gpt", "value": "[[408, 977], [388, 903], [374, 865], [362, 741]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door handle."}, {"from": "gpt", "value": "[[446, 853], [458, 753], [434, 683]]"}]}
{"id": 240, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_227/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_229/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[431, 803], [452, 718], [486, 664], [507, 656]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door handle."}, {"from": "gpt", "value": "[[411, 965], [376, 819], [379, 653], [403, 602]]"}]}
{"id": 242, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_230/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_234/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door handle."}, {"from": "gpt", "value": "[[449, 861], [489, 772], [547, 691], [588, 587]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet handle."}, {"from": "gpt", "value": "[[420, 822], [388, 761], [388, 699], [437, 614]]"}]}
{"id": 244, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_236/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_253/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door handles."}, {"from": "gpt", "value": "[[397, 915], [434, 826], [507, 757], [582, 753]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door handles."}, {"from": "gpt", "value": "[[388, 965], [359, 865], [345, 726], [353, 583], [466, 278]]"}]}
{"id": 246, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_257/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_260/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet handle."}, {"from": "gpt", "value": "[[437, 911], [460, 857], [492, 784], [486, 691]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet handle."}, {"from": "gpt", "value": "[[423, 822], [397, 726], [397, 614]]"}]}
{"id": 248, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_263/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_264/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door handles."}, {"from": "gpt", "value": "[[446, 811], [431, 633], [397, 548], [356, 467]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet handle."}, {"from": "gpt", "value": "[[449, 826], [481, 741], [455, 683]]"}]}
{"id": 250, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_278/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_280/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[434, 807], [394, 718], [362, 672], [319, 637]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "[[458, 861], [420, 776], [359, 664], [321, 618]]"}]}
{"id": 252, "image": ["rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_282/frame_0.png", "rtx_frames_success_1/25_nyu_door_opening_surprising_effectiveness#episode_283/frame_0.png"], "height": [720, 720], "width": [960, 960], "conversations": [{"from": "system", "value": "You are InternVL, a robotic entity agent.\n\nYou are given one observation of the scene and a task to perform. Based on the task, You should return a set of points as the trajectory to execute the task. The point trajectory format should be \"[[x1, y1], [x2, y2], [x3, y3],...]\". The point coordinates are normalized to integers between 0 and 1000. Please return the point trajectory directly."}, {"from": "human", "value": "Image-1:\n You need to reach for the handle."}, {"from": "gpt", "value": "[[405, 973], [400, 826], [420, 672], [498, 475]]"}, {"from": "human", "value": "Image-2:\n You need to approach the cabinet door."}, {"from": "gpt", "value": "