| import datasets |
| import json |
| import numpy |
| import torch |
|
|
| _DESCRIPTION = """\ |
| Dataset of pre-processed samples from a small portion of the \ |
| Waymo Open Motion Data for our risk-biased prediction task. |
| """ |
|
|
| _CITATION = """\ |
| @InProceedings{NiMe:2022, |
| author = {Haruki Nishimura, Jean Mercat, Blake Wulfe, Rowan McAllister}, |
| title = {RAP: Risk-Aware Prediction for Robust Planning}, |
| booktitle = {Proceedings of the 2022 IEEE International Conference on Robot Learning (CoRL)}, |
| month = {December}, |
| year = {2022}, |
| address = {Grafton Road, Auckland CBD, Auckland 1010}, |
| url = {}, |
| } |
| """ |
|
|
| _URL = "https://huggingface.co/datasets/jmercat/risk_biased_dataset/resolve/main/" |
| _URLS = { |
| "test": _URL + "data.json", |
| } |
|
|
| class RiskBiasedDataset(datasets.GeneratorBasedBuilder): |
| """Dataset of pre-processed samples from a portion of the |
| Waymo Open Motion Data for the risk-biased prediction task.""" |
| |
| VERSION = datasets.Version("0.0.0") |
| |
| BUILDER_CONFIGS = [ |
| datasets.BuilderConfig(name="risk_biased_dataset", version=VERSION, description="Dataset of pre-processed samples from a portion of the Waymo Open Motion Data for the risk-biased prediction task."), |
| ] |
| |
| DEFAULT_CONFIG_NAME = "risk_biased_dataset" |
| |
| def _info(self): |
| return datasets.DatasetInfo( |
| description= _DESCRIPTION, |
| features=datasets.Features( |
| {"x": datasets.Sequence(datasets.Sequence(datasets.Sequence(datasets.Sequence(datasets.Value("float32"))))), |
| "mask_x": datasets.Sequence(datasets.Sequence(datasets.Sequence(datasets.Value("bool")))), |
| "y": datasets.Sequence(datasets.Sequence(datasets.Sequence(datasets.Sequence(datasets.Value("float32"))))), |
| "mask_y": datasets.Sequence(datasets.Sequence(datasets.Sequence(datasets.Value("bool")))), |
| "mask_loss": datasets.Sequence(datasets.Sequence(datasets.Sequence(datasets.Value("bool")))), |
| "map_data": datasets.Sequence(datasets.Sequence(datasets.Sequence(datasets.Sequence(datasets.Value("float32"))))), |
| "mask_map": datasets.Sequence(datasets.Sequence(datasets.Sequence(datasets.Value("bool")))), |
| "offset": datasets.Sequence(datasets.Sequence(datasets.Sequence(datasets.Value("float32")))), |
| "x_ego": datasets.Sequence(datasets.Sequence(datasets.Sequence(datasets.Sequence(datasets.Value("float32"))))), |
| "y_ego": datasets.Sequence(datasets.Sequence(datasets.Sequence(datasets.Sequence(datasets.Value("float32"))))), |
| } |
| ), |
| supervised_keys=None, |
| homepage="https://sites.google.com/view/corl-risk/home", |
| citation=_CITATION, |
| ) |
| |
| def _split_generators(self, dl_manager): |
| urls_to_download = _URLS |
| downloaded_files = dl_manager.download_and_extract(urls_to_download) |
| |
| return [datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"filepath": downloaded_files["test"], "split": "test"}),] |
| |
| def _generate_examples(self, filepath, split): |
| """Yields examples.""" |
| assert split == "test" |
| with open(filepath, "r") as f: |
| data = json.load(f) |
| |
| x = torch.from_numpy(numpy.array(data["x"]).astype(numpy.float32)) |
| mask_x = torch.from_numpy(numpy.array(data["mask_x"]).astype(numpy.bool_)) |
| y = torch.from_numpy(numpy.array(data["y"]).astype(numpy.float32)) |
| mask_y = torch.from_numpy(numpy.array(data["mask_y"]).astype(numpy.bool_)) |
| mask_loss = torch.from_numpy( numpy.array(data["mask_loss"]).astype(numpy.bool_)) |
| map_data = torch.from_numpy(numpy.array(data["map_data"]).astype(numpy.float32)) |
| mask_map = torch.from_numpy(numpy.array(data["mask_map"]).astype(numpy.bool_)) |
| offset = torch.from_numpy(numpy.array(data["offset"]).astype(numpy.float32)) |
| x_ego = torch.from_numpy(numpy.array(data["x_ego"]).astype(numpy.float32)) |
| y_ego = torch.from_numpy(numpy.array(data["y_ego"]).astype(numpy.float32)) |
| |
| batch_size = x.shape[0] |
| |
| for i in range(batch_size): |
| |
| |
| |
| |
| |
| |
| yield i, {"x": x[i:i+1], "mask_x": mask_x[i:i+1], |
| "y": y[i:i+1], "mask_y": mask_y[i:i+1], "mask_loss": mask_loss[i:i+1], |
| "map_data": map_data[i:i+1], "mask_map": mask_map[i:i+1], |
| "offset": offset[i:i+1], |
| "x_ego": x_ego[i:i+1], |
| "y_ego": y_ego[i:i+1]} |
| |
| |
|
|