File size: 6,971 Bytes
66780fc f6bb345 66780fc f6bb345 66780fc f6bb345 66780fc |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 |
# π΅ Deepfake Audio Detection Model
A machine learning model to detect deepfake/synthetic audio using Wav2Vec2 embeddings and classical ML classifiers.
[](https://huggingface.co/hjsgfd/deepfake_audio_classifier)
[](https://www.python.org/downloads/)
[](https://opensource.org/licenses/MIT)
## π Model Performance
| Model | Accuracy | Precision | Recall | F1-Score |
|-------|----------|-----------|--------|----------|
| **Logistic Regression** | **92.86%** | 0.95 | 0.93 | 0.93 |
| SVM | 85.71% | 0.89 | 0.86 | 0.85 |
| Random Forest | 78.57% | 0.85 | 0.79 | 0.76 |
**Best Model: Logistic Regression with 92.86% accuracy**
## π― Approach
### 1. Dataset
- **Source**: [Real vs Fake Human Voice Deepfake Audio Dataset](https://huggingface.co/datasets/ud-nlp/real-vs-fake-human-voice-deepfake-audio)
- **Size**: 70 audio samples
- **Classes**: 5 classes (0, 1, 2, 3, 4)
- **Distribution**: Perfectly balanced (14 samples per class)
### 2. Feature Extraction
We use **Wav2Vec2** (facebook/wav2vec2-base-960h) to extract deep audio embeddings:
- Pre-trained self-supervised model
- Extracts 768-dimensional feature vectors
- Captures semantic audio information
- Handles variable-length audio automatically
**Pipeline:**
```
Audio File β Wav2Vec2 β 768-dim Embedding β Classifier β Prediction
```
### 3. Model Training
Three classifiers were trained and compared:
#### Logistic Regression (Best)
- **Accuracy**: 92.86%
- Multi-class classification with OvR strategy
- Max iterations: 1000
- Features: StandardScaler normalized
#### SVM
- **Accuracy**: 85.71%
- RBF kernel
- Probability estimates enabled
#### Random Forest
- **Accuracy**: 78.57%
- 200 estimators
- Parallel processing enabled
### 4. Preprocessing
- **Audio Loading**: Support for both URLs and local files
- **Resampling**: All audio converted to 16kHz
- **Stereo to Mono**: Averaged across channels
- **Normalization**: StandardScaler on embeddings
## π Quick Start
### Installation
```bash
pip install transformers torch librosa soundfile scikit-learn huggingface-hub requests numpy
```
### Usage
#### Simple Prediction
```python
from predict_from_hf import AudioDeepfakeDetectorFromHF
# Initialize detector (downloads model automatically)
detector = AudioDeepfakeDetectorFromHF("hjsgfd/deepfake_audio_classifier")
# Predict from URL
result = detector.predict("https://your-audio-file.wav", is_url=True)
print(f"Prediction: {result['label']} ({result['confidence']:.1%})")
```
#### Batch Prediction
```python
from predict_from_hf import AudioDeepfakeDetectorFromHF
detector = AudioDeepfakeDetectorFromHF("hjsgfd/deepfake_audio_classifier")
# Multiple URLs
audio_urls = [
"https://example.com/audio1.wav",
"https://example.com/audio2.wav",
"https://example.com/audio3.wav",
]
results = detector.predict_batch(audio_urls, are_urls=True)
# Print results
for result in results:
if 'prediction' in result:
print(f"{result['audio_source']}: {result['label']} ({result['confidence']:.1%})")
```
#### Local Files
```python
# Single file
result = detector.predict("path/to/audio.wav", is_url=False)
# Multiple files
local_files = ["audio1.wav", "audio2.wav", "audio3.wav"]
results = detector.predict_batch(local_files, are_urls=False)
```
## π Model Files
The model consists of three files hosted on Hugging Face:
1. **deepfake_audio_classifier.pkl** - Trained Logistic Regression classifier
2. **audio_scaler.pkl** - StandardScaler for feature normalization
3. **model_metadata.json** - Model configuration and metadata
```json
{
"model_type": "LogisticRegression",
"accuracy": 0.9286,
"feature_extractor": "facebook/wav2vec2-base-960h",
"embedding_dim": 768,
"num_classes": 5,
"class_labels": {
"0": "class_0",
"1": "class_1",
"2": "class_2",
"3": "class_3",
"4": "class_4"
}
}
```
## π Detailed Results
### Training Configuration
- **Training Samples**: 56 (80%)
- **Testing Samples**: 14 (20%)
- **Feature Dimension**: 768
- **Stratified Split**: Maintains class distribution
### Logistic Regression Performance (Best Model)
```
precision recall f1-score support
class_0 1.00 0.67 0.80 3
class_1 1.00 1.00 1.00 2
class_2 1.00 1.00 1.00 3
class_3 0.75 1.00 0.86 3
class_4 1.00 1.00 1.00 3
accuracy 0.93 14
macro avg 0.95 0.93 0.93 14
weighted avg 0.95 0.93 0.93 14
```
### Key Metrics
- **Macro Average Precision**: 0.95
- **Macro Average Recall**: 0.93
- **Macro Average F1-Score**: 0.93
- **Overall Accuracy**: 92.86%
## π§ Technical Details
### Dependencies
```
transformers>=4.30.0
torch>=2.0.0
librosa>=0.10.0
soundfile>=0.12.0
scikit-learn>=1.3.0
huggingface-hub>=0.16.0
requests>=2.31.0
numpy>=1.24.0
```
### Model Architecture
```
Input: Audio File (any format supported by soundfile)
β
Preprocessing (16kHz, Mono)
β
Wav2Vec2 Feature Extractor
β
768-dimensional Embedding
β
StandardScaler Normalization
β
Logistic Regression Classifier
β
Output: Class Prediction + Confidence Scores
```
### Supported Audio Formats
- WAV
- MP3
- FLAC
- OGG
- M4A
## π Training Process
1. **Data Loading**: Load dataset with disabled auto-decoding
2. **Feature Extraction**: Extract Wav2Vec2 embeddings (768-dim vectors)
3. **Train-Test Split**: 80-20 stratified split
4. **Normalization**: StandardScaler on training data
5. **Model Training**: Train 3 classifiers (LR, RF, SVM)
6. **Evaluation**: Compare performance on test set
7. **Selection**: Choose best model (Logistic Regression)
8. **Export**: Save model, scaler, and metadata
## π― Use Cases
- Deepfake audio detection
- Voice authentication systems
- Media verification tools
- Forensic audio analysis
- Content moderation platforms
## π€ Contributing
Contributions are welcome! Please feel free to submit a Pull Request.
## π Citation
If you use this model, please cite:
```bibtex
@misc{deepfake_audio_classifier_2024,
author = {Your Name},
title = {Deepfake Audio Detection Model},
year = {2024},
publisher = {Hugging Face},
howpublished = {\url{https://huggingface.co/hjsgfd/deepfake_audio_classifier}}
}
```
## π Acknowledgments
- **Dataset**: [ud-nlp/real-vs-fake-human-voice-deepfake-audio](https://huggingface.co/datasets/ud-nlp/real-vs-fake-human-voice-deepfake-audio)
- **Feature Extractor**: [facebook/wav2vec2-base-960h](https://huggingface.co/facebook/wav2vec2-base-960h)
- **Transformers Library**: Hugging Face
## π§ Contact
For questions or feedback, please open an issue on the repository.
---
|