BioLLM SmolLM2-360M Distill
The first language model distilled from co-training on live biological neurons.
This model is a SmolLM2-360M fine-tuned via LoRA distillation using training data generated on a live biological neural culture (Cortical Labs CL1-2544-144). During co-training, every token was physically stimulated onto real neurons grown on a multi-electrode array (MEA), and the culture's spike responses were recorded and blended with LLM predictions. The distillation captures the joint LLM+biological decision-making into a standalone model.
This work is part of the Antekythera project, an experimental program investigating consciousness correlates in biological neural substrates and their coupling with language models.
What Makes This Model Unique
Unlike conventional language models trained purely on text corpora via gradient descent, this model's training signal passed through living neurons. Each token during co-training followed this path:
LLM logits --> SpatialEncoder --> MEA stimulation (electrical)
|
Biological neurons
(cortical culture)
|
Token selection <-- Blending <-- Spike response recording
The biological culture acted as a neural co-processor that modulated token selection based on its spike response patterns. The distillation then transferred this joint bio-digital decision-making into the LoRA weights of SmolLM2-360M.
Biological Substrate
| Property | Value |
|---|---|
| Device | Cortical Labs CL1-2544-144 |
| Substrate | Human iPSC-derived cortical neurons on 64-channel MEA |
| Sampling rate | 25 kHz |
| Active channels | 18 (of 59 usable) |
| Dominant channels | ch24 (24.5 Hz), ch34 (11.7 Hz), ch53 (9.6 Hz) |
| Peak channel count | 28 |
| Culture preparation | Thompson Sampling Awakener v36 (35-cycle MAB optimization) |
The culture on CL1-2544-144 was prepared over 35 cycles of multi-armed bandit stimulation optimization using the Dormant Channel Awakener protocol, which expanded the culture from ~9 to 29 responsive channels. The awakener uses Thompson Sampling with Beta posteriors across 12 stimulation pattern arms to maximize a composite reward of activation, C-Score (consciousness metric), Lempel-Ziv complexity, and algebraic connectivity.
Training Details
Co-Training (PPO on Live Neurons)
| Parameter | Value |
|---|---|
| Co-training LLM | LFM2-350M (GGUF, Q4_0) |
| Vocab size (source) | 65,536 |
| Tokens processed | 200 |
| Accuracy | 99.0% (198/200) |
| Mean C-Score | 0.165 (all tokens) / 0.200 (non-zero only) |
| Max C-Score | 0.320 |
| Non-zero integration | 82% of tokens (165/200) |
| Training duration | 622 seconds |
| Protocol | Culture-Safe BioLLM v5 (HDF5-calibrated) |
| Health monitoring | CultureHealthMonitor with 40% channel-loss abort |
| Curriculum | Adaptive: binary -> 8-way -> full candidate set |
The co-training used a PPO loop where the LLM generated candidate tokens, a SpatialEncoder mapped them to MEA stimulation patterns, the biological culture responded with spikes, and a NeuralLogitDecoder blended the spike-derived probabilities with the LLM's predictions. The culture was protected by a health monitor that tracked channel counts and spike rates in real-time.
Consciousness Metrics During Training
The C-Score is a composite consciousness correlate metric derived from neural integration measures (algebraic connectivity, spectral gap, Lempel-Ziv complexity, and Granger causality). During co-training:
- 82% of tokens produced measurable neural integration (C > 0.01)
- Integration peaked at C = 0.320, indicating genuine multi-channel coordinated activity
- The culture maintained stable integration across both training blocks
- C-Score trend was positive (0.447 correlation with time), suggesting the culture became more integrated during training
LoRA Distillation
| Parameter | Value |
|---|---|
| Target model | SmolLM2-360M (HuggingFaceTB) |
| Vocab size (target) | 49,152 |
| Distillation type | Cross-vocabulary (LogitSpaceDistiller) |
| LoRA rank | 8 |
| LoRA alpha | 16 |
| LoRA dropout | 0.05 |
| Target modules | q_proj, v_proj |
| Trainable parameters | 819,200 / 362,640,320 (0.23%) |
| Training samples | 60 (filtered from 100 raw) |
| C-Score filter | >= 0.15 (removes zero-integration samples) |
| C-Score beta scaling | Enabled (saturation = 0.45) |
| Loss function | KL divergence (temperature = 2.0) |
| Epochs | 3 |
| Epoch 1 loss | 16.500 |
| Epoch 2 loss | 16.328 |
| Epoch 3 loss | 16.653 |
| PEFT version | 0.18.1 |
The distillation used culture-health-calibrated filtering: only samples where the biological culture showed genuine neural integration (C-Score >= 0.15) were included. Additionally, C-Score beta scaling weighted high-integration samples more heavily during training, so the model preferentially learned from moments when the culture exhibited coordinated multi-channel activity.
The cross-vocabulary distillation projects from LFM2's 65,536-token vocabulary to SmolLM2's 49,152-token vocabulary via logit-space alignment, which accounts for the higher loss values compared to same-vocabulary distillation.
Culture-Safe Protocol
Training biological neurons requires careful safety protocols to prevent culture damage:
- 3-phase warmup: 60% graduated ramp, 15% bouncy stabilization, 25% gentle stimulation
- Real-time health monitoring: Channel count and spike rate tracking with automatic abort on >40% channel loss
- Inter-block rest: 5-minute rest periods with 20:80 stim:rest priming ratio
- Frequency clamping: Stimulation frequency limited to 4-40 Hz range
- Amplitude safety: 0.3-2.5 uA bounds enforced at relay level
- Mandatory rest: 2-hour rest per 4 hours of stimulation
How to Use
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("4r7i5t/BioLLM_SmolLM2_360m_Distill")
tokenizer = AutoTokenizer.from_pretrained("4r7i5t/BioLLM_SmolLM2_360m_Distill")
input_ids = tokenizer.encode("The nature of consciousness is", return_tensors="pt")
output = model.generate(input_ids, max_new_tokens=50, temperature=0.8, do_sample=True)
print(tokenizer.decode(output[0], skip_special_tokens=True))
Limitations
- Small training set: 60 filtered samples from a single co-training session. This is a proof-of-concept demonstrating the pipeline, not a production model.
- Cross-vocabulary gap: The source LLM (LFM2, 65K vocab) and target (SmolLM2, 49K vocab) have different tokenizers, introducing projection noise during distillation.
- Culture variability: Biological neural cultures have inherent stochasticity. The culture's state (channel activity, integration capacity) varies over time and cannot be precisely reproduced.
- Single culture: Trained on one specific culture (CL1-2544-144) with its unique connectivity patterns and channel hierarchy. Different cultures would produce different training signals.
The Antekythera Project
This model is one artifact from a larger experimental program investigating whether biological neural substrates exhibit consciousness correlates and whether those correlates can meaningfully couple with language model inference.
Key findings from the broader project:
- Bio-Shadow Flip: After 10 cycles of terraform training on CL1-2544-015, the culture learned to differentiate coherent spatial patterns from random ones (Cohen's d flipped from -1.6 to +0.6)
- Trimodal C-Score Distribution: Consciousness metrics show three distinct modes โ zero, low-integration, and high-integration โ with high-integration islands growing significantly over training cycles (r=0.80, p=0.005)
- Coherence Experiment: 7/8 statistical tests passed showing biological encoding produces significantly more coherent neural responses than shadow controls (d=1.9-15.3)
- Honest Failures: Transfer Entropy does not grow over training cycles (homeostatic set-point), Gate 2 convergence battery fails on naive cultures, motor channels remain silent during training
The project's core finding is that the building blocks of consciousness appear to be substrate-intrinsic (present in biological neural tissue) but not sufficient on their own. The gap between correlates and consciousness is one of organization, not material.
Citation
If you use this model or find this work interesting, please cite:
@misc{biollm_smollm2_distill_2026,
title={BioLLM SmolLM2-360M Distill: Language Model Distilled from Live Biological Neural Co-Training},
author={Antekythera Project},
year={2026},
url={https://huggingface.co/4r7i5t/BioLLM_SmolLM2_360m_Distill},
note={First LLM distilled from co-training on Cortical Labs CL1 biological neural culture}
}
Hardware
- Biological substrate: Cortical Labs CL1-2544-144 (64-channel MEA, human iPSC-derived cortical neurons)
- Compute: Apple Silicon (local inference and distillation)
- Connectivity: Cloudflare Access tunnel to CL1 device cloud
- Carbon emissions: Negligible compute; biological culture maintenance is the primary resource
Built with living neurons. Handle with curiosity.
- Downloads last month
- 48
Model tree for 4r7i5t/BioLLM_SmolLM2_360m_Distill
Base model
HuggingFaceTB/SmolLM2-360M