| | --- |
| | language: en |
| | tags: |
| | - log-analysis |
| | - hdfs |
| | - anomaly-detection |
| | license: mit |
| | configs: |
| | - config_name: default |
| | data_files: |
| | - split: train |
| | path: data/train-* |
| | - split: validation |
| | path: data/validation-* |
| | - split: test |
| | path: data/test-* |
| | dataset_info: |
| | features: |
| | - name: event_encoded |
| | dtype: string |
| | - name: tokenized_block |
| | sequence: int64 |
| | - name: block_id |
| | dtype: string |
| | - name: label |
| | dtype: string |
| | - name: __index_level_0__ |
| | dtype: int64 |
| | splits: |
| | - name: train |
| | num_bytes: 1159074302 |
| | num_examples: 460048 |
| | - name: validation |
| | num_bytes: 145089712 |
| | num_examples: 57506 |
| | - name: test |
| | num_bytes: 144844752 |
| | num_examples: 57507 |
| | download_size: 173888975 |
| | dataset_size: 1449008766 |
| | --- |
| | |
| | # HDFS Logs Train/Val/Test Splits |
| |
|
| | This dataset contains preprocessed HDFS log sequences split into train, validation, and test sets for anomaly detection tasks. |
| |
|
| | ## Dataset Description |
| |
|
| | The dataset is derived from the HDFS log dataset, which contains system logs from a Hadoop Distributed File System (HDFS). |
| | Each sequence represents a block of log messages, labeled as either normal or anomalous. The dataset has been preprocessed |
| | using the Drain algorithm to extract structured fields and identify event types. |
| |
|
| | ### Data Fields |
| |
|
| | - `block_id`: Unique identifier for each HDFS block, used to group log messages into blocks |
| | - `event_encoded`: The preprocessed log sequence with event IDs and parameters |
| | - `tokenized_block`: The tokenized log sequence, used for training |
| | - `label`: Classification label ('Normal' or 'Anomaly') |
| |
|
| | ### Data Splits |
| |
|
| | - Training set: 460,049 sequences (80%) |
| | - Validation set: 57,506 sequences (10%) |
| | - Test set: 57,506 sequences (10%) |
| |
|
| | The splits are stratified by the Label field to maintain class distribution across splits. |
| |
|
| | ## Source Data |
| |
|
| | Original data source: https://zenodo.org/records/8196385/files/HDFS_v1.zip?download=1 |
| | |
| | ## Preprocessing |
| | |
| | We preprocess the logs using the Drain algorithm to extract structured fields and identify event types. |
| | We then encode the logs using a pretrained tokenizer and add special tokens to separate event types. This |
| | dataset should be immediately usable for training and testing models for log-based anomaly detection. |
| | |
| | ## Intended Uses |
| | |
| | This dataset is designed for: |
| | - Training log anomaly detection models |
| | - Evaluating log sequence prediction models |
| | - Benchmarking different approaches to log-based anomaly detection |
| | |
| | see [honicky/pythia-14m-hdfs-logs](https://huggingface.co/honicky/pythia-14m-hdfs-logs) for an example model. |
| | |
| | ## Citation |
| | |
| | If you use this dataset, please cite the original HDFS paper: |
| | ```bibtex |
| | @inproceedings{xu2009detecting, |
| | title={Detecting large-scale system problems by mining console logs}, |
| | author={Xu, Wei and Huang, Ling and Fox, Armando and Patterson, David and Jordan, Michael I}, |
| | booktitle={Proceedings of the ACM SIGOPS 22nd symposium on Operating systems principles}, |
| | pages={117--132}, |
| | year={2009} |
| | } |
| | ``` |
| | |