lightbrain-100m

Model Description

LIGHTBRAIN is a novel neural architecture based on Hybrid Field Transformer paradigm.

Key Features

  • Sparse Activation: Only ~0.1-10% of field regions active during inference
  • Field Dynamics: Pattern resonance for knowledge retrieval
  • Transformer Integration: Self-attention for sequence modeling (hybrid)
  • OpenAI-Compatible API: Drop-in replacement for chat completions

Architecture

Component Value
Hidden Size 768
Layers 12
Attention Heads 12
Field Regions 128
Field Size 128
Field Depth 64
β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚   TRANSFORMER ENCODER LAYERS        β”‚
β”‚   (Self-Attention + FFN)            β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
        ↓
β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚   FIELD DYNAMICS CORE               β”‚
β”‚   (Sparse Activation + Evolution)   β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
        ↓
β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚   OUTPUT PROJECTION                 β”‚
β”‚   (Pattern β†’ Token Logits)          β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜

Model Files

File Description
Model-001.safetensors Model weights (721.30 MB)
config.json Model configuration
tokenizer.json Tokenizer vocabulary
tokenizer_config.json Tokenizer configuration
generation_config.json Generation parameters
params.json LIGHTBRAIN parameters

Model Stats

  • Original Size: 721.28 MB
  • File Size: 721.30 MB
  • Compression Ratio: 1.00x
  • Number of Tensors: 200

Usage

With LIGHTBRAIN Library

from lightbrain.model import HybridFieldTransformer
from lightbrain.inference import InferenceEngine

# Load model
model = HybridFieldTransformer.load("path/to/model")
engine = InferenceEngine(model=model)

# Generate
result = engine.generate("Hello, how are you?")
print(result.text)

Loading from Safetensors

from safetensors.numpy import load_file
import json

# Load weights
weights = load_file("Model-001.safetensors")

# Load config
with open("config.json") as f:
    config = json.load(f)

# Reconstruct model from weights

In Google Colab

# Install
!pip install safetensors

# Download
from huggingface_hub import snapshot_download
model_path = snapshot_download(repo_id="lightbrain-100m")

# Load and use
from safetensors.numpy import load_file
weights = load_file(f"{model_path}/Model-001.safetensors")

Training

Trained using LIGHTBRAIN framework with:

  • Resonance Alignment (Hebbian learning)
  • Gradient-based fine-tuning for transformer layers
  • Field topology optimization

License

MIT License

Citation

@misc{lightbrain2024,
  title={LIGHTBRAIN: Hybrid Field Dynamics for Efficient LLMs},
  year={2024},
  publisher={HuggingFace}
}
Downloads last month
22
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Evaluation results