Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,4 @@
|
|
|
|
|
| 1 |
language:
|
| 2 |
- en
|
| 3 |
library_name: timm
|
|
@@ -80,7 +81,7 @@ The model achieves strong discriminative performance with an AUC ROC of 0.9561 o
|
|
| 80 |
Source code and training pipeline are available at:
|
| 81 |
https://github.com/ArchitRastogi20/vit-spoof-detection-pda
|
| 82 |
|
| 83 |
-
|
| 84 |
|
| 85 |
## Model Summary
|
| 86 |
|
|
@@ -92,7 +93,7 @@ https://github.com/ArchitRastogi20/vit-spoof-detection-pda
|
|
| 92 |
- Framework: PyTorch with timm
|
| 93 |
- Pretraining: ImageNet
|
| 94 |
|
| 95 |
-
|
| 96 |
|
| 97 |
## Intended Use
|
| 98 |
|
|
@@ -101,7 +102,7 @@ Potential application domains include biometric authentication systems, access c
|
|
| 101 |
|
| 102 |
The model is not intended for deployment in high risk security environments without additional validation, calibration, and fairness analysis.
|
| 103 |
|
| 104 |
-
|
| 105 |
|
| 106 |
## Dataset
|
| 107 |
|
|
@@ -115,7 +116,7 @@ CelebA Spoof is a large scale face anti spoofing dataset containing diverse spoo
|
|
| 115 |
Dataset reference:
|
| 116 |
https://github.com/Davidzhangyuanhan/CelebA-Spoof
|
| 117 |
|
| 118 |
-
|
| 119 |
|
| 120 |
## Data Augmentation
|
| 121 |
|
|
@@ -139,7 +140,7 @@ Applied transformations include:
|
|
| 139 |
|
| 140 |
Normalization follows ImageNet statistics used by ViT models.
|
| 141 |
|
| 142 |
-
|
| 143 |
|
| 144 |
## Model Architecture
|
| 145 |
|
|
@@ -159,7 +160,7 @@ Key configuration details:
|
|
| 159 |
* Dropout: 0.1
|
| 160 |
* Mixed precision training enabled
|
| 161 |
|
| 162 |
-
|
| 163 |
|
| 164 |
## Training Procedure
|
| 165 |
|
|
@@ -180,7 +181,7 @@ Training configuration:
|
|
| 180 |
| Early stopping | Enabled |
|
| 181 |
| Device | NVIDIA RTX A5000 |
|
| 182 |
|
| 183 |
-
|
| 184 |
|
| 185 |
## Evaluation
|
| 186 |
|
|
@@ -189,7 +190,7 @@ Threshold optimization was applied to balance false acceptance and false rejecti
|
|
| 189 |
|
| 190 |
Reported metrics include accuracy, F1 score, AUC ROC, precision, recall, specificity, FAR, FRR, and EER.
|
| 191 |
|
| 192 |
-
|
| 193 |
|
| 194 |
## Results
|
| 195 |
|
|
@@ -218,7 +219,7 @@ Reported metrics include accuracy, F1 score, AUC ROC, precision, recall, specifi
|
|
| 218 |
| FRR | 0.0232 |
|
| 219 |
| EER | 0.1083 |
|
| 220 |
|
| 221 |
-
|
| 222 |
|
| 223 |
## Confusion Matrix
|
| 224 |
|
|
@@ -227,7 +228,7 @@ Reported metrics include accuracy, F1 score, AUC ROC, precision, recall, specifi
|
|
| 227 |
| Actual Spoof | 404 | 267 |
|
| 228 |
| Actual Live | 25 | 1051 |
|
| 229 |
|
| 230 |
-
|
| 231 |
|
| 232 |
## Limitations
|
| 233 |
|
|
@@ -235,7 +236,7 @@ Reported metrics include accuracy, F1 score, AUC ROC, precision, recall, specifi
|
|
| 235 |
* High FAR indicates sensitivity to certain spoof patterns
|
| 236 |
* No cross dataset evaluation included
|
| 237 |
|
| 238 |
-
|
| 239 |
|
| 240 |
## Citation
|
| 241 |
|
|
|
|
| 1 |
+
---
|
| 2 |
language:
|
| 3 |
- en
|
| 4 |
library_name: timm
|
|
|
|
| 81 |
Source code and training pipeline are available at:
|
| 82 |
https://github.com/ArchitRastogi20/vit-spoof-detection-pda
|
| 83 |
|
| 84 |
+
|
| 85 |
|
| 86 |
## Model Summary
|
| 87 |
|
|
|
|
| 93 |
- Framework: PyTorch with timm
|
| 94 |
- Pretraining: ImageNet
|
| 95 |
|
| 96 |
+
|
| 97 |
|
| 98 |
## Intended Use
|
| 99 |
|
|
|
|
| 102 |
|
| 103 |
The model is not intended for deployment in high risk security environments without additional validation, calibration, and fairness analysis.
|
| 104 |
|
| 105 |
+
|
| 106 |
|
| 107 |
## Dataset
|
| 108 |
|
|
|
|
| 116 |
Dataset reference:
|
| 117 |
https://github.com/Davidzhangyuanhan/CelebA-Spoof
|
| 118 |
|
| 119 |
+
|
| 120 |
|
| 121 |
## Data Augmentation
|
| 122 |
|
|
|
|
| 140 |
|
| 141 |
Normalization follows ImageNet statistics used by ViT models.
|
| 142 |
|
| 143 |
+
|
| 144 |
|
| 145 |
## Model Architecture
|
| 146 |
|
|
|
|
| 160 |
* Dropout: 0.1
|
| 161 |
* Mixed precision training enabled
|
| 162 |
|
| 163 |
+
|
| 164 |
|
| 165 |
## Training Procedure
|
| 166 |
|
|
|
|
| 181 |
| Early stopping | Enabled |
|
| 182 |
| Device | NVIDIA RTX A5000 |
|
| 183 |
|
| 184 |
+
|
| 185 |
|
| 186 |
## Evaluation
|
| 187 |
|
|
|
|
| 190 |
|
| 191 |
Reported metrics include accuracy, F1 score, AUC ROC, precision, recall, specificity, FAR, FRR, and EER.
|
| 192 |
|
| 193 |
+
|
| 194 |
|
| 195 |
## Results
|
| 196 |
|
|
|
|
| 219 |
| FRR | 0.0232 |
|
| 220 |
| EER | 0.1083 |
|
| 221 |
|
| 222 |
+
|
| 223 |
|
| 224 |
## Confusion Matrix
|
| 225 |
|
|
|
|
| 228 |
| Actual Spoof | 404 | 267 |
|
| 229 |
| Actual Live | 25 | 1051 |
|
| 230 |
|
| 231 |
+
|
| 232 |
|
| 233 |
## Limitations
|
| 234 |
|
|
|
|
| 236 |
* High FAR indicates sensitivity to certain spoof patterns
|
| 237 |
* No cross dataset evaluation included
|
| 238 |
|
| 239 |
+
|
| 240 |
|
| 241 |
## Citation
|
| 242 |
|