# general information - the purpose of this the repository is to create a dataset loading script for a milling related dataset - further information about the dataset are in https://data.mendeley.com/datasets/zpxs87bjt8/3 - the dataset size is 10 Gigabit. hugging face doesn't allow uploading that amount of data. This rep contain a small section of the dataset (see *data* folder) only to prove the concept. loading script for the original dataset and the small dataset # how to load dataset - first download the dataset from https://data.mendeley.com/datasets/zpxs87bjt8/3 - loading script (milling_LUH_data.py) should be at the same directory as data directory (the folder where all h5 files are) - load data as shown in notebook.ipynb - define train, test, validation splits in the loading script (milling_LUH_data.py)