🌿 BDA - Botanical Dormancy Architecture

Model Description

BDA is a novel neural network architecture where each neuron independently learns when to enter a "dormant" state, inspired by selective plant cell dormancy during winter. This per-neuron adaptive sparsity mechanism achieves 55-84% neuron dormancy with minimal inference overhead (2.5-8%).

Key Features

  • 🧠 Per-neuron learnable thresholds
  • âš¡ Minimal overhead (2.5-8%)
  • 💾 96% cache hit rate
  • 🔧 Hardware agnostic (P100, T4, A100)

Performance Results

A100 GPU (FP16)

Batch Size Standard (ms) BDA (ms) Overhead Dormancy
1 0.594 0.701 +18.0% 55%
8 0.847 0.917 +8.3% 55%
32 2.906 3.252 +11.9% 55%

Other GPUs (Batch=8)

GPU Standard (ms) BDA (ms) Overhead Dormancy
T4 7.23 7.89 +9.1% 82%
P100 9.43 9.67 +2.5% 84%

How It Works

Each BDA layer has a learnable threshold θ = sigmoid(φ) × 0.5. During inference:

  • Neuron enters dormant state when activation < threshold
  • Dormant neurons output zero, reducing computation
  • Cache mechanism reuses previous decisions (96% hit rate)

Installation

pip install torch torchvision numpy
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Datasets used to train ay933/BDA-Botanical-Dormancy