Upload 4 files
Browse files- README.md +14 -14
- config.json +7 -7
- pytorch_model.bin +1 -1
README.md
CHANGED
|
@@ -22,14 +22,14 @@ model-index:
|
|
| 22 |
name: Video Classification
|
| 23 |
metrics:
|
| 24 |
- type: accuracy
|
| 25 |
-
value: 0.
|
| 26 |
name: Accuracy
|
| 27 |
- type: f1
|
| 28 |
-
value: 0.
|
| 29 |
name: Macro F1
|
| 30 |
---
|
| 31 |
|
| 32 |
-
# Driver Behavior Detection Model (Epoch
|
| 33 |
|
| 34 |
운전자 이상행동 감지를 위한 Video Swin Transformer 기반 모델입니다.
|
| 35 |
|
|
@@ -44,18 +44,18 @@ model-index:
|
|
| 44 |
|
| 45 |
| Label | Class | F1-Score |
|
| 46 |
|:-----:|-------|:--------:|
|
| 47 |
-
| 0 | 정상 (Normal) | 0.
|
| 48 |
-
| 1 | 졸음운전 (Drowsy Driving) | 0.
|
| 49 |
-
| 2 | 물건찾기 (Reaching/Searching) | 0.
|
| 50 |
-
| 3 | 휴대폰 사용 (Phone Usage) | 0.
|
| 51 |
| 4 | 운전자 폭행 (Driver Assault) | 1.00 |
|
| 52 |
|
| 53 |
-
## Performance (Epoch
|
| 54 |
|
| 55 |
| Metric | Value |
|
| 56 |
|--------|-------|
|
| 57 |
-
| **Accuracy** |
|
| 58 |
-
| **Macro F1** | 0.
|
| 59 |
| **Validation Samples** | 1,371,062 |
|
| 60 |
|
| 61 |
## Training Configuration
|
|
@@ -64,14 +64,14 @@ model-index:
|
|
| 64 |
|-----------|-------|
|
| 65 |
| Hardware | 2x NVIDIA RTX A6000 (48GB) |
|
| 66 |
| Distributed | DDP (DistributedDataParallel) |
|
| 67 |
-
| Batch Size | 32 (16
|
| 68 |
| Gradient Accumulation | 4 |
|
| 69 |
| Effective Batch | 128 |
|
| 70 |
| Optimizer | AdamW (lr=1e-3, wd=0.05) |
|
| 71 |
| Scheduler | OneCycleLR |
|
| 72 |
| Mixed Precision | FP16 |
|
| 73 |
| Loss | CrossEntropy + Label Smoothing (0.1) |
|
| 74 |
-
| Regularization | Mixup (
|
| 75 |
|
| 76 |
## Usage
|
| 77 |
|
|
@@ -98,7 +98,7 @@ with torch.no_grad():
|
|
| 98 |
- **Total Samples (windows)**: 1,371,062
|
| 99 |
- **Window Size**: 30 frames
|
| 100 |
- **Stride**: 15 frames
|
| 101 |
-
- **Resolution**:
|
| 102 |
|
| 103 |
## Augmentation (Training)
|
| 104 |
|
|
@@ -106,7 +106,7 @@ with torch.no_grad():
|
|
| 106 |
- HorizontalFlip (p=0.5)
|
| 107 |
- ColorJitter, HueSaturationValue
|
| 108 |
- Temporal Augmentation (speed change, frame drop)
|
| 109 |
-
- Mixup (
|
| 110 |
- CoarseDropout
|
| 111 |
|
| 112 |
## License
|
|
|
|
| 22 |
name: Video Classification
|
| 23 |
metrics:
|
| 24 |
- type: accuracy
|
| 25 |
+
value: 0.9683
|
| 26 |
name: Accuracy
|
| 27 |
- type: f1
|
| 28 |
+
value: 0.9600
|
| 29 |
name: Macro F1
|
| 30 |
---
|
| 31 |
|
| 32 |
+
# Driver Behavior Detection Model (Epoch 4)
|
| 33 |
|
| 34 |
운전자 이상행동 감지를 위한 Video Swin Transformer 기반 모델입니다.
|
| 35 |
|
|
|
|
| 44 |
|
| 45 |
| Label | Class | F1-Score |
|
| 46 |
|:-----:|-------|:--------:|
|
| 47 |
+
| 0 | 정상 (Normal) | 0.95 |
|
| 48 |
+
| 1 | 졸음운전 (Drowsy Driving) | 0.99 |
|
| 49 |
+
| 2 | 물건찾기 (Reaching/Searching) | 0.94 |
|
| 50 |
+
| 3 | 휴대폰 사용 (Phone Usage) | 0.93 |
|
| 51 |
| 4 | 운전자 폭행 (Driver Assault) | 1.00 |
|
| 52 |
|
| 53 |
+
## Performance (Epoch 4)
|
| 54 |
|
| 55 |
| Metric | Value |
|
| 56 |
|--------|-------|
|
| 57 |
+
| **Accuracy** | 96.83% |
|
| 58 |
+
| **Macro F1** | 0.9600 |
|
| 59 |
| **Validation Samples** | 1,371,062 |
|
| 60 |
|
| 61 |
## Training Configuration
|
|
|
|
| 64 |
|-----------|-------|
|
| 65 |
| Hardware | 2x NVIDIA RTX A6000 (48GB) |
|
| 66 |
| Distributed | DDP (DistributedDataParallel) |
|
| 67 |
+
| Batch Size | 32 (16 x 2 GPU) |
|
| 68 |
| Gradient Accumulation | 4 |
|
| 69 |
| Effective Batch | 128 |
|
| 70 |
| Optimizer | AdamW (lr=1e-3, wd=0.05) |
|
| 71 |
| Scheduler | OneCycleLR |
|
| 72 |
| Mixed Precision | FP16 |
|
| 73 |
| Loss | CrossEntropy + Label Smoothing (0.1) |
|
| 74 |
+
| Regularization | Mixup (a=0.4), Dropout (0.3) |
|
| 75 |
|
| 76 |
## Usage
|
| 77 |
|
|
|
|
| 98 |
- **Total Samples (windows)**: 1,371,062
|
| 99 |
- **Window Size**: 30 frames
|
| 100 |
- **Stride**: 15 frames
|
| 101 |
+
- **Resolution**: 224x224
|
| 102 |
|
| 103 |
## Augmentation (Training)
|
| 104 |
|
|
|
|
| 106 |
- HorizontalFlip (p=0.5)
|
| 107 |
- ColorJitter, HueSaturationValue
|
| 108 |
- Temporal Augmentation (speed change, frame drop)
|
| 109 |
+
- Mixup (a=0.4)
|
| 110 |
- CoarseDropout
|
| 111 |
|
| 112 |
## License
|
config.json
CHANGED
|
@@ -11,9 +11,9 @@
|
|
| 11 |
"layers": ["LayerNorm(768)", "Dropout(0.3)", "Linear(768, 5)"]
|
| 12 |
},
|
| 13 |
"training": {
|
| 14 |
-
"epoch":
|
| 15 |
-
"accuracy": 0.
|
| 16 |
-
"macro_f1": 0.
|
| 17 |
"batch_size": 32,
|
| 18 |
"optimizer": "AdamW",
|
| 19 |
"learning_rate": 1e-3,
|
|
@@ -23,10 +23,10 @@
|
|
| 23 |
"augmentation": ["Mixup(0.4)", "RandomResizedCrop", "HorizontalFlip", "ColorJitter", "TemporalAugmentation"]
|
| 24 |
},
|
| 25 |
"performance": {
|
| 26 |
-
"정상": {"precision": 0.
|
| 27 |
-
"졸음운전": {"precision": 0.99, "recall": 0.
|
| 28 |
-
"물건찾기": {"precision": 0.
|
| 29 |
-
"휴대폰 사용": {"precision": 0.
|
| 30 |
"운전자 폭행": {"precision": 1.00, "recall": 1.00, "f1": 1.00}
|
| 31 |
}
|
| 32 |
}
|
|
|
|
| 11 |
"layers": ["LayerNorm(768)", "Dropout(0.3)", "Linear(768, 5)"]
|
| 12 |
},
|
| 13 |
"training": {
|
| 14 |
+
"epoch": 4,
|
| 15 |
+
"accuracy": 0.9683,
|
| 16 |
+
"macro_f1": 0.9600,
|
| 17 |
"batch_size": 32,
|
| 18 |
"optimizer": "AdamW",
|
| 19 |
"learning_rate": 1e-3,
|
|
|
|
| 23 |
"augmentation": ["Mixup(0.4)", "RandomResizedCrop", "HorizontalFlip", "ColorJitter", "TemporalAugmentation"]
|
| 24 |
},
|
| 25 |
"performance": {
|
| 26 |
+
"정상": {"precision": 0.95, "recall": 0.94, "f1": 0.95},
|
| 27 |
+
"졸음운전": {"precision": 0.99, "recall": 0.98, "f1": 0.99},
|
| 28 |
+
"물건찾기": {"precision": 0.93, "recall": 0.95, "f1": 0.94},
|
| 29 |
+
"휴대폰 사용": {"precision": 0.92, "recall": 0.93, "f1": 0.93},
|
| 30 |
"운전자 폭행": {"precision": 1.00, "recall": 1.00, "f1": 1.00}
|
| 31 |
}
|
| 32 |
}
|
pytorch_model.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 126244047
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:346b59e7df4bf056e98b43f7e4f6a43f495516429899f7301679cea380fac98a
|
| 3 |
size 126244047
|