|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import numpy as np
|
|
|
import torch
|
|
|
|
|
|
from lerobot.envs.utils import preprocess_observation
|
|
|
from lerobot.processor.env_processor import LiberoProcessorStep
|
|
|
from lerobot.processor.pipeline import PolicyProcessorPipeline
|
|
|
|
|
|
seed = 42
|
|
|
np.random.seed(seed)
|
|
|
|
|
|
B = 5
|
|
|
obs1 = {
|
|
|
"pixels": {
|
|
|
"image": (np.random.rand(B, 256, 256, 3) * 255).astype(np.uint8),
|
|
|
"image2": (np.random.rand(B, 256, 256, 3) * 255).astype(np.uint8),
|
|
|
},
|
|
|
"robot_state": {
|
|
|
"eef": {
|
|
|
"pos": np.random.randn(B, 3),
|
|
|
"quat": np.random.randn(B, 4),
|
|
|
"mat": np.random.randn(B, 3, 3),
|
|
|
},
|
|
|
"gripper": {
|
|
|
"qpos": np.random.randn(B, 2),
|
|
|
"qvel": np.random.randn(B, 2),
|
|
|
},
|
|
|
"joints": {
|
|
|
"pos": np.random.randn(B, 7),
|
|
|
"vel": np.random.randn(B, 7),
|
|
|
},
|
|
|
},
|
|
|
}
|
|
|
|
|
|
observation = preprocess_observation(obs1)
|
|
|
libero_preprocessor = PolicyProcessorPipeline(
|
|
|
steps=[
|
|
|
LiberoProcessorStep(),
|
|
|
]
|
|
|
)
|
|
|
processed_obs = libero_preprocessor(observation)
|
|
|
assert "observation.state" in processed_obs
|
|
|
state = processed_obs["observation.state"]
|
|
|
assert isinstance(state, torch.Tensor)
|
|
|
assert state.dtype == torch.float32
|
|
|
|
|
|
assert state.shape[0] == B
|
|
|
assert state.shape[1] == 8
|
|
|
|
|
|
assert "observation.images.image" in processed_obs
|
|
|
assert "observation.images.image2" in processed_obs
|
|
|
|
|
|
assert isinstance(processed_obs["observation.images.image"], torch.Tensor)
|
|
|
assert isinstance(processed_obs["observation.images.image2"], torch.Tensor)
|
|
|
|
|
|
assert processed_obs["observation.images.image"].shape == (B, 3, 256, 256)
|
|
|
assert processed_obs["observation.images.image2"].shape == (B, 3, 256, 256)
|
|
|
|