lightbrain-100m
Model Description
LIGHTBRAIN is a novel neural architecture based on Hybrid Field Transformer paradigm.
Key Features
- Sparse Activation: Only ~0.1-10% of field regions active during inference
- Field Dynamics: Pattern resonance for knowledge retrieval
- Transformer Integration: Self-attention for sequence modeling (hybrid)
- OpenAI-Compatible API: Drop-in replacement for chat completions
Architecture
| Component | Value |
|---|---|
| Hidden Size | 768 |
| Layers | 12 |
| Attention Heads | 12 |
| Field Regions | 128 |
| Field Size | 128 |
| Field Depth | 64 |
βββββββββββββββββββββββββββββββββββββββ
β TRANSFORMER ENCODER LAYERS β
β (Self-Attention + FFN) β
βββββββββββββββββββββββββββββββββββββββ
β
βββββββββββββββββββββββββββββββββββββββ
β FIELD DYNAMICS CORE β
β (Sparse Activation + Evolution) β
βββββββββββββββββββββββββββββββββββββββ
β
βββββββββββββββββββββββββββββββββββββββ
β OUTPUT PROJECTION β
β (Pattern β Token Logits) β
βββββββββββββββββββββββββββββββββββββββ
Model Files
| File | Description |
|---|---|
Model-001.safetensors |
Model weights (721.30 MB) |
config.json |
Model configuration |
tokenizer.json |
Tokenizer vocabulary |
tokenizer_config.json |
Tokenizer configuration |
generation_config.json |
Generation parameters |
params.json |
LIGHTBRAIN parameters |
Model Stats
- Original Size: 721.28 MB
- File Size: 721.30 MB
- Compression Ratio: 1.00x
- Number of Tensors: 200
Usage
With LIGHTBRAIN Library
from lightbrain.model import HybridFieldTransformer
from lightbrain.inference import InferenceEngine
# Load model
model = HybridFieldTransformer.load("path/to/model")
engine = InferenceEngine(model=model)
# Generate
result = engine.generate("Hello, how are you?")
print(result.text)
Loading from Safetensors
from safetensors.numpy import load_file
import json
# Load weights
weights = load_file("Model-001.safetensors")
# Load config
with open("config.json") as f:
config = json.load(f)
# Reconstruct model from weights
In Google Colab
# Install
!pip install safetensors
# Download
from huggingface_hub import snapshot_download
model_path = snapshot_download(repo_id="lightbrain-100m")
# Load and use
from safetensors.numpy import load_file
weights = load_file(f"{model_path}/Model-001.safetensors")
Training
Trained using LIGHTBRAIN framework with:
- Resonance Alignment (Hebbian learning)
- Gradient-based fine-tuning for transformer layers
- Field topology optimization
License
MIT License
Citation
@misc{lightbrain2024,
title={LIGHTBRAIN: Hybrid Field Dynamics for Efficient LLMs},
year={2024},
publisher={HuggingFace}
}
- Downloads last month
- 22