ash12321's picture
Upload folder using huggingface_hub
378602c verified
# Deepfake Detector V12 - RAM Optimized (2 Hour Runtime)
## 🎯 Production-Grade Fine-tuned Ensemble (16K Samples, 2 Epochs)
### Built on V11, RAM-Safe Training for 2 Hour Runtime
This is **V12 RAM Optimized**, a fine-tuned version of the V11 ensemble detector with 30 real datasets, minimal synthetic generation, and 2-epoch high-quality fine-tuning optimized for RAM safety and 2 hour runtime.
## πŸ“Š Performance
### V12 Ensemble Performance (Test Set - NEVER SEEN):
- **Test Accuracy**: 97.94%
- **Test Precision**: 0.9957
- **Test Recall**: 0.9486
- **Test F1 Score**: 0.9715
### Individual Models:
- **Model 1**: 95.95% val accuracy βœ“ from V11
- **Model 2**: 97.40% val accuracy βœ“ from V11
- **Model 3**: 96.25% val accuracy βœ“ from V11
**Successfully loaded 3/3 models from V11!**
## ⚑ RAM Optimizations
### Training Configuration:
- **Epochs**: 2 (high-quality fine-tuning)
- **Batch Size**: 32 (RAM safe)
- **Target Samples**: 16K (reduced for RAM)
- **Pin Memory**: Enabled
- **Num Workers**: 2 (parallel loading)
- **Device**: GPU (CUDA) or CPU
- **Expected RAM**: ~5-6GB during training
- **Training Time**: ~1.5 hours
### RAM Safety Strategy:
- Reduced samples: 16K vs 30K (47% less data)
- Smaller batches: 32 vs 64 (50% less per batch)
- Same dataset diversity: All 50 datasets still used
- Per-dataset targets unchanged
- Should stay well under 12GB RAM
## πŸ“¦ Dataset Strategy
### Real Images (30 Datasets) - UNCHANGED:
- Core datasets: beans, cats_vs_dogs, tiny-imagenet, flowers, oxford-pets
- Classification: cifar10, mnist, fashion_mnist, caltech101, food101
- Specialized: stanford_dogs, gtsrb, eurosat, aircraft, sun397
- Medical/Scientific: patch_camelyon, NIH chest x-rays
- Target: ~8K real images with minimal synthetic (<1.5K if needed)
### Fake Images (20 Datasets) - UNCHANGED:
- GAN datasets: AFHQ, pokemon, wikiart, metfaces, celeba
- Style transfer: winter2summer, horse2zebra, watercolor2photo
- Diffusion models: pokemon-gpt4-captions, few-shot-universe
- Target: ~8K fake images with minimal synthetic (<1.5K if needed)
## 🎯 Key Features
1. **2 Epochs**: High-quality fine-tuning from V11 base
2. **RAM Safe**: 16K samples, batch 32
3. **Same Datasets**: All 50 datasets still used (30 real + 20 fake)
4. **Minimal Synthetic**: Only if <70% of target reached
5. **GPU Accelerated**: Optimized for both GPU and CPU
6. **Fine-tuned from V11**: Transfer learning from proven V11 architecture
## πŸ’Ύ Training Details
- **Training Time**: 23.0 minutes (~0.4h)
- **Epochs per Model**: 2
- **Batch Size**: 16 (RAM optimized)
- **Target Samples**: 10,000
- **Models Loaded from V11**: 3/3
- **Real Datasets**: 31 (unchanged)
- **Fake Datasets**: 20 (unchanged)
- **Synthetic Used**: Minimal (only if needed)
## πŸ›‘οΈ Anti-Memorization
### 80/10/10 Split (STRICT)
- **Training**: 80% (10,470 samples)
- **Validation**: 10% (1,308 samples)
- **Test**: 10% (1,310 samples) - **NEVER SEEN**
## πŸ“„ License
MIT License
---
**Model Version**: V12 RAM Optimized (16K Dataset, 2 Epochs)
**Base Model**: ash12321/deepfake-detector-v11
**Release Date**: 2025-11-06
**Training Time**: ~1.5 hours
**Status**: Production Ready βœ… (RAM Safe + High-Quality Fine-tuning)