🌿 BDA - Botanical Dormancy Architecture
Model Description
BDA is a novel neural network architecture where each neuron independently learns when to enter a "dormant" state, inspired by selective plant cell dormancy during winter. This per-neuron adaptive sparsity mechanism achieves 55-84% neuron dormancy with minimal inference overhead (2.5-8%).
Key Features
- 🧠Per-neuron learnable thresholds
- âš¡ Minimal overhead (2.5-8%)
- 💾 96% cache hit rate
- 🔧 Hardware agnostic (P100, T4, A100)
Performance Results
A100 GPU (FP16)
| Batch Size | Standard (ms) | BDA (ms) | Overhead | Dormancy |
|---|---|---|---|---|
| 1 | 0.594 | 0.701 | +18.0% | 55% |
| 8 | 0.847 | 0.917 | +8.3% | 55% |
| 32 | 2.906 | 3.252 | +11.9% | 55% |
Other GPUs (Batch=8)
| GPU | Standard (ms) | BDA (ms) | Overhead | Dormancy |
|---|---|---|---|---|
| T4 | 7.23 | 7.89 | +9.1% | 82% |
| P100 | 9.43 | 9.67 | +2.5% | 84% |
How It Works
Each BDA layer has a learnable threshold θ = sigmoid(φ) × 0.5. During inference:
- Neuron enters dormant state when activation < threshold
- Dormant neurons output zero, reducing computation
- Cache mechanism reuses previous decisions (96% hit rate)
Installation
pip install torch torchvision numpy
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support