| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | from dataclasses import dataclass, field |
| | from typing import Iterator, Optional |
| |
|
| | import datasets |
| | import pandas as pd |
| | import pyarrow as pa |
| | import pyarrow.parquet as pq |
| | from gluonts.dataset.field_names import FieldName |
| |
|
| | _CITATION = """\ |
| | @article{woo2023pushing, |
| | title={Pushing the Limits of Pre-training for Time Series Forecasting in the CloudOps Domain}, |
| | author={Woo, Gerald and Liu, Chenghao and Kumar, Akshat and Sahoo, Doyen}, |
| | journal={arXiv preprint arXiv:2310.05063}, |
| | year={2023} |
| | } |
| | """ |
| |
|
| |
|
| | _CONFIGS = { |
| | "azure_vm_traces_2017": { |
| | "optional_fields": ( |
| | FieldName.FEAT_STATIC_CAT, |
| | FieldName.FEAT_STATIC_REAL, |
| | FieldName.PAST_FEAT_DYNAMIC_REAL, |
| | ), |
| | "prediction_length": 48, |
| | "freq": "5T", |
| | "stride": 48, |
| | "univariate": True, |
| | "multivariate": False, |
| | "rolling_evaluations": 12, |
| | "test_split_date": pd.Period( |
| | year=2016, month=12, day=13, hour=15, minute=55, freq="5T" |
| | ), |
| | "_feat_static_cat_cardinalities": { |
| | "train_test": ( |
| | ("vm_id", 17568), |
| | ("subscription_id", 2713), |
| | ("deployment_id", 3255), |
| | ("vm_category", 3), |
| | ), |
| | "pretrain": ( |
| | ("vm_id", 177040), |
| | ("subscription_id", 5514), |
| | ("deployment_id", 15208), |
| | ("vm_category", 3), |
| | ), |
| | }, |
| | "target_dim": 1, |
| | "feat_static_real_dim": 3, |
| | "past_feat_dynamic_real_dim": 2, |
| | }, |
| | "borg_cluster_data_2011": { |
| | "optional_fields": ( |
| | FieldName.FEAT_STATIC_CAT, |
| | FieldName.PAST_FEAT_DYNAMIC_REAL, |
| | ), |
| | "prediction_length": 48, |
| | "freq": "5T", |
| | "stride": 48, |
| | "univariate": False, |
| | "multivariate": True, |
| | "rolling_evaluations": 12, |
| | "test_split_date": pd.Period( |
| | year=2011, month=5, day=28, hour=18, minute=55, freq="5T" |
| | ), |
| | "_feat_static_cat_cardinalities": { |
| | "train_test": ( |
| | ("job_id", 850), |
| | ("task_id", 11117), |
| | ("user", 282), |
| | ("scheduling_class", 4), |
| | ("logical_job_name", 718), |
| | ), |
| | "pretrain": ( |
| | ("job_id", 6072), |
| | ("task_id", 154503), |
| | ("user", 518), |
| | ("scheduling_class", 4), |
| | ("logical_job_name", 3899), |
| | ), |
| | }, |
| | "target_dim": 2, |
| | "past_feat_dynamic_real_dim": 5, |
| | }, |
| | "alibaba_cluster_trace_2018": { |
| | "optional_fields": ( |
| | FieldName.FEAT_STATIC_CAT, |
| | FieldName.PAST_FEAT_DYNAMIC_REAL, |
| | ), |
| | "prediction_length": 48, |
| | "freq": "5T", |
| | "stride": 48, |
| | "univariate": False, |
| | "multivariate": True, |
| | "rolling_evaluations": 12, |
| | "test_split_date": pd.Period( |
| | year=2018, month=1, day=8, hour=11, minute=55, freq="5T" |
| | ), |
| | "_feat_static_cat_cardinalities": { |
| | "train_test": ( |
| | ("container_id", 6048), |
| | ("app_du", 1292), |
| | ), |
| | "pretrain": ( |
| | ("container_id", 64457), |
| | ("app_du", 9484), |
| | ), |
| | }, |
| | "target_dim": 2, |
| | "past_feat_dynamic_real_dim": 6, |
| | }, |
| | } |
| |
|
| | PRETRAIN = datasets.splits.NamedSplit("pretrain") |
| | TRAIN_TEST = datasets.splits.NamedSplit("train_test") |
| |
|
| | Cardinalities = tuple[tuple[str, int], ...] |
| |
|
| |
|
| | @dataclass |
| | class CloudOpsTSFConfig(datasets.BuilderConfig): |
| | """BuilderConfig for CloudOpsTSF.""" |
| |
|
| | |
| | train_test: bool = field(default=True, init=False) |
| | pretrain: bool = field(default=False, init=False) |
| |
|
| | |
| | prediction_length: int = field(default=None) |
| | freq: str = field(default=None) |
| | stride: int = field(default=None) |
| | univariate: bool = field(default=None) |
| | multivariate: bool = field(default=None) |
| | optional_fields: tuple[str, ...] = field(default=None) |
| | rolling_evaluations: int = field(default=None) |
| | test_split_date: pd.Period = field(default=None) |
| | _feat_static_cat_cardinalities: dict[str, Cardinalities] = field( |
| | default_factory=dict |
| | ) |
| | target_dim: int = field(default=1) |
| | feat_static_real_dim: int = field(default=0) |
| | past_feat_dynamic_real_dim: int = field(default=0) |
| |
|
| | def feat_static_cat_cardinalities( |
| | self, split: str = "train_test" |
| | ) -> Optional[list[int]]: |
| | if FieldName.FEAT_STATIC_CAT not in self.optional_fields: |
| | return None |
| |
|
| | return [c[1] for c in self._feat_static_cat_cardinalities[split]] |
| |
|
| |
|
| | class CloudOpsTSF(datasets.ArrowBasedBuilder): |
| | VERSION = datasets.Version("1.0.0") |
| |
|
| | BUILDER_CONFIGS = [] |
| | for dataset, config in _CONFIGS.items(): |
| | BUILDER_CONFIGS.append( |
| | CloudOpsTSFConfig(name=dataset, version=VERSION, description="", **config) |
| | ) |
| |
|
| | def _info(self) -> datasets.DatasetInfo: |
| | def sequence_feature(dtype: str, univar: bool) -> datasets.Sequence: |
| | if univar: |
| | return datasets.Sequence(datasets.Value(dtype)) |
| | return datasets.Sequence(datasets.Sequence(datasets.Value(dtype))) |
| |
|
| | features = { |
| | FieldName.START: datasets.Value("timestamp[s]"), |
| | FieldName.TARGET: sequence_feature("float32", self.config.univariate), |
| | FieldName.ITEM_ID: datasets.Value("string"), |
| | } |
| |
|
| | CAT_FEATS = ( |
| | FieldName.FEAT_STATIC_CAT, |
| | FieldName.FEAT_DYNAMIC_CAT, |
| | FieldName.PAST_FEAT_DYNAMIC, |
| | ) |
| | REAL_FEATS = ( |
| | FieldName.FEAT_STATIC_REAL, |
| | FieldName.FEAT_DYNAMIC_REAL, |
| | FieldName.PAST_FEAT_DYNAMIC_REAL, |
| | ) |
| | STATIC_FEATS = (FieldName.FEAT_STATIC_CAT, FieldName.FEAT_STATIC_REAL) |
| | DYNAMIC_FEATS = ( |
| | FieldName.FEAT_DYNAMIC_CAT, |
| | FieldName.FEAT_DYNAMIC_REAL, |
| | FieldName.PAST_FEAT_DYNAMIC, |
| | FieldName.PAST_FEAT_DYNAMIC_REAL, |
| | ) |
| |
|
| | for ts_field in self.config.optional_fields: |
| | |
| | if ts_field in CAT_FEATS: |
| | dtype = "int32" |
| | elif ts_field in REAL_FEATS: |
| | dtype = "float32" |
| | else: |
| | raise ValueError(f"Invalid field: {ts_field}") |
| |
|
| | |
| | if ts_field in STATIC_FEATS: |
| | univar = True |
| | elif ts_field in DYNAMIC_FEATS: |
| | univar = False |
| | else: |
| | raise ValueError(f"Invalid field: {ts_field}") |
| |
|
| | features[ts_field] = sequence_feature(dtype, univar) |
| |
|
| | features = datasets.Features(features) |
| |
|
| | return datasets.DatasetInfo( |
| | features=features, |
| | citation=_CITATION, |
| | ) |
| |
|
| | def _split_generators(self, dl_manager) -> list[datasets.SplitGenerator]: |
| | downloaded_files = dl_manager.download_and_extract( |
| | [ |
| | f"{self.config.name}/train_test.zip", |
| | f"{self.config.name}/pretrain.zip", |
| | ] |
| | ) |
| |
|
| | generators = [ |
| | datasets.SplitGenerator( |
| | name=TRAIN_TEST, |
| | gen_kwargs={"filepath": downloaded_files[0]}, |
| | ), |
| | datasets.SplitGenerator( |
| | name=PRETRAIN, |
| | gen_kwargs={"filepath": downloaded_files[1]}, |
| | ), |
| | ] |
| |
|
| | return generators |
| |
|
| | def _generate_tables(self, filepath: str) -> Iterator[pa.Table]: |
| | table = pq.read_table(filepath) |
| |
|
| | for batch in table.to_batches(): |
| | columns = batch.columns |
| | schema = batch.schema |
| |
|
| | yield batch[FieldName.ITEM_ID].to_pylist(), pa.Table.from_arrays( |
| | columns, schema=schema |
| | ) |
| |
|