| { | |
| "featuresDict": { | |
| "features": { | |
| "steps": { | |
| "pythonClassName": "tensorflow_datasets.core.features.dataset_feature.Dataset", | |
| "sequence": { | |
| "feature": { | |
| "featuresDict": { | |
| "features": { | |
| "action": { | |
| "featuresDict": { | |
| "features": { | |
| "gripper_closedness_action": { | |
| "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", | |
| "tensor": { | |
| "dtype": "float32", | |
| "encoding": "none", | |
| "shape": {} | |
| } | |
| }, | |
| "rotation_delta": { | |
| "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", | |
| "tensor": { | |
| "dtype": "float32", | |
| "encoding": "none", | |
| "shape": { | |
| "dimensions": [ | |
| "3" | |
| ] | |
| } | |
| } | |
| }, | |
| "terminate_episode": { | |
| "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", | |
| "tensor": { | |
| "dtype": "float32", | |
| "encoding": "none", | |
| "shape": {} | |
| } | |
| }, | |
| "world_vector": { | |
| "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", | |
| "tensor": { | |
| "dtype": "float32", | |
| "encoding": "none", | |
| "shape": { | |
| "dimensions": [ | |
| "3" | |
| ] | |
| } | |
| } | |
| } | |
| } | |
| }, | |
| "pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict" | |
| }, | |
| "is_first": { | |
| "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", | |
| "tensor": { | |
| "dtype": "bool", | |
| "encoding": "none", | |
| "shape": {} | |
| } | |
| }, | |
| "is_last": { | |
| "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", | |
| "tensor": { | |
| "dtype": "bool", | |
| "encoding": "none", | |
| "shape": {} | |
| } | |
| }, | |
| "is_terminal": { | |
| "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", | |
| "tensor": { | |
| "dtype": "bool", | |
| "encoding": "none", | |
| "shape": {} | |
| } | |
| }, | |
| "observation": { | |
| "featuresDict": { | |
| "features": { | |
| "agentview_rgb": { | |
| "description": "RGB captured by workspace camera", | |
| "image": { | |
| "dtype": "uint8", | |
| "shape": { | |
| "dimensions": [ | |
| "224", | |
| "224", | |
| "3" | |
| ] | |
| } | |
| }, | |
| "pythonClassName": "tensorflow_datasets.core.features.image_feature.Image" | |
| }, | |
| "ee_states": { | |
| "description": "Pose of the end effector specified as a homogenous matrix.", | |
| "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", | |
| "tensor": { | |
| "dtype": "float32", | |
| "encoding": "none", | |
| "shape": { | |
| "dimensions": [ | |
| "16" | |
| ] | |
| } | |
| } | |
| }, | |
| "eye_in_hand_rgb": { | |
| "description": "RGB captured by in hand camera", | |
| "image": { | |
| "dtype": "uint8", | |
| "shape": { | |
| "dimensions": [ | |
| "224", | |
| "224", | |
| "3" | |
| ] | |
| } | |
| }, | |
| "pythonClassName": "tensorflow_datasets.core.features.image_feature.Image" | |
| }, | |
| "gripper_states": { | |
| "description": "gripper_states = 0 means the gripper is fully closed. The value represents the gripper width of Franka Panda Gripper.", | |
| "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", | |
| "tensor": { | |
| "dtype": "float32", | |
| "encoding": "none", | |
| "shape": { | |
| "dimensions": [ | |
| "1" | |
| ] | |
| } | |
| } | |
| }, | |
| "joint_states": { | |
| "description": "joint values", | |
| "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", | |
| "tensor": { | |
| "dtype": "float32", | |
| "encoding": "none", | |
| "shape": { | |
| "dimensions": [ | |
| "7" | |
| ] | |
| } | |
| } | |
| }, | |
| "natural_language_embedding": { | |
| "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", | |
| "tensor": { | |
| "dtype": "float32", | |
| "encoding": "none", | |
| "shape": { | |
| "dimensions": [ | |
| "512" | |
| ] | |
| } | |
| } | |
| }, | |
| "natural_language_instruction": { | |
| "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor", | |
| "tensor": { | |
| "dtype": "string", | |
| "encoding": "none", | |
| "shape": {} | |
| } | |
| } | |
| } | |
| }, | |
| "pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict" | |
| }, | |
| "reward": { | |
| "pythonClassName": "tensorflow_datasets.core.features.scalar.Scalar", | |
| "tensor": { | |
| "dtype": "float32", | |
| "encoding": "none", | |
| "shape": {} | |
| } | |
| } | |
| } | |
| }, | |
| "pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict" | |
| }, | |
| "length": "-1" | |
| } | |
| } | |
| } | |
| }, | |
| "pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict" | |
| } |