File size: 4,399 Bytes
33206dc |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 |
# UO Processed Dataset
Documentation for the **University of Ottawa (UO) bearing dataset** after running the updated preprocessing pipeline. Along with KAIST and PU, this is one of the three datasets that were refreshed most recently.
---
## Folder Layout
```
UO/
βββ train.pt
βββ val.pt
βββ test.pt
βββ args.json
βββ additional_features.pt
βββ before_sliding_window/
βββ train.pt
βββ val.pt
βββ test.pt
```
- `train.pt`, `val.pt`, `test.pt` β windowed tensors post sliding-window subsampling.
- `before_sliding_window/*.pt` β the same splits before windowing (full-length sequences).
- `args.json` β preprocessing arguments (window size, stride, split ratios, etc.).
- `additional_features.pt` β Torch-serialized metadata collected from each MAT file (including channel names and auxiliary sensor info).
---
## Saved Tensor Structure
Each windowed `.pt` file is a dictionary with:
- `samples`: `torch.Tensor` of shape `[num_windows, num_channels, window_length]`
- `labels`: `torch.Tensor` with class ids (`0: healthy`, `1: inner race fault`, `2: outer race fault`)
- `sequence_ids`: indices that point back to the original MAT files
- `sliding_window_sequence_ids`: mapping from windows to their source sequences
- `size`: split ratio used when generating the split (for traceability)
The `before_sliding_window` tensors use the same keys but keep the un-windowed signals (length β 2,000,000 samples per channel).
---
## Features
Each MAT file contains several arrays; the preprocessing script keeps only the first two vibration channels (`channel_1`, `channel_2`) that provide the full 2β―000β―000 samples used in prior work. Their names are captured in `additional_features.pt` under each MAT file:
```python
import torch
meta = torch.load("additional_features.pt")
print(meta["healthy"]["H-A-1.mat"]["name_features"])
```
Use this metadata to understand the physical meaning of each channel or to filter specific features.
---
## Usage Example
```python
import torch
train = torch.load("train.pt")
windows = train["samples"] # [N, num_channels, window_length]
labels = train["labels"] # bearing condition ids
original_indices = train["sliding_window_sequence_ids"]
print(windows.shape)
print(labels.unique())
```
Access the un-windowed signals:
```python
raw = torch.load("before_sliding_window/train.pt")
full_sequences = raw["samples"] # [num_sequences, num_channels, original_length]
```
---
## Processing Pipeline
1. **Raw input** β MAT files organised into three lists: healthy (`H-*`), inner faults (`I-*`), outer faults (`O-*`).
2. **Channel selection** β For every file, the script extracts `channel_1` and `channel_2`, each of length 2β―000β―000 samples. Additional metadata (speed, load, etc.) is preserved in `additional_features.pt`.
3. **Class-wise sequence split** β Using `train_size`, sequences are randomly assigned to train vs. (val+test); the remaining sequences are divided into validation and test according to `val_size`/`test_size`.
4. **Save before-window tensors** β Full-length tensors are written to `before_sliding_window/{train,val,test}.pt` for troubleshooting.
5. **Sliding-window sampling** β Windows of length `window_size` are generated every `step = window_size * stride` samples from each sequence.
6. **Persist final datasets** β Windowed tensors and labels are stored in the root `.pt` files together with the mapping fields (`sequence_ids`, `sliding_window_sequence_ids`, `size`).
---
## Input / Output Cheat Sheet
| Stage | Shape | Description |
|-------|-------|-------------|
| Raw MAT arrays | `(2β―000β―000,)` per channel | Original vibration signals |
| After loading | `(1, 2, 2β―000β―000)` | Tensor for a single MAT file (two channels) |
| Before sliding window (train) | `[num_sequences, 2, original_length]` | Randomly selected sequences saved to `before_sliding_window/train.pt` |
| After sliding window (train) | `[num_windows, 2, window_size]` | Final training dataset in `train.pt` |
---
## Notes
- This README applies to the refreshed implementation; other datasets still rely on the legacy processing approach.
- The supplied splits rely on random shuffling with the configured ratios. Re-run the pipeline to regenerate different splits if required.
|