braille256-v6: Lattice-Aware Multimodal Braille Model

The first LLM with explicit dot-lattice structure in its architecture.

Model Description

braille256-v6 builds on the multimodal foundation of v5, integrating formal lattice theory into the training pipeline. This is not just a Braille-native model—it's a lattice-native model that understands the mathematical structure of Braille at the architectural level.

Key Innovations

Feature Description
Lattice Attention Attention scores incorporate Hamming-based similarity on Braille cells
Lattice Embeddings Token embeddings initialized to respect Boolean lattice structure
Morphological Regularization Training loss includes equivariance under erosion/dilation
Haptic Evaluation New metrics for tactile quality of outputs

Architecture

Parameters: ~12M
Layers: 4
Heads: 4
Hidden: 256
Vocab: 32,000 (SentencePiece)
Context: 512

Lattice Attention

Standard transformer attention computes:

Attention(Q, K, V) = softmax(QK^T / √d) V

Lattice attention blends this with Braille-aware similarity:

LatticeAttn = (1-λ) * StandardAttn + λ * HammingAttn

where HammingAttn[i,j] = 8 - popcount(token[i] XOR token[j])

This gives the model an inductive bias toward understanding Braille structure.

Lattice Embeddings

For the first 256 tokens (corresponding to Braille cells), embeddings are initialized as:

embedding[i] = Σ basis[b] for each raised dot b in cell i

This means similar Braille cells (low Hamming distance) start with similar embeddings.

Morphological Regularization

Training includes a regularization term:

L_morph = ReLU(||emb - erode(emb)|| - ||emb - dilate(emb)||)

This encourages embeddings to respect the lattice ordering: erode(x) ≤ x ≤ dilate(x).

Theoretical Foundation

This model implements the formal theory from:

"Theoretical Foundations for 8-Dot Braille-Native LLMs"

Key theoretical components:

  1. Braille Lattice: Boolean algebra (B⁸, ∧, ∨, ¬) with 256 elements
  2. Morphological Operators: Erosion, dilation, opening, closing
  3. Modality-Invariant Representation: (modality, sequence, embedding) triple
  4. Lattice Metrics: Hamming distance, Jaccard similarity

See: braille_lattice_theory.py for full implementation.

Modality Support

Modality Header Status
TEXT ⣿⠁ ✅ Trained
IMAGE ⣿⠃ ✅ Trained
AUDIO ⣿⠇ ✅ Trained
BINARY ⣿⠏ ✅ Trained
VIDEO ⣿⠗ 🔄 Framework ready

Haptic Evaluation Metrics

v6 introduces new evaluation metrics for tactile quality:

Metric Description Target Achieved
Lattice Coherence Adjacent tokens have low Hamming distance > 0.7 0.743
Morphological Stability Outputs stable under erosion/dilation > 0.5 0.453
Haptic Score Combined tactile quality metric > 0.5 0.598

Training Results

Metric Value
Final Loss 1.23
Training Steps 10,000
Training Time 2h 7m
Corpus Balanced multimodal (25% each: text, image, audio, binary)
Corpus Size 164M chars

Usage

import torch
from train_lattice_v6 import Braille256LatticeModel, LatticeConfig

# Load model
config = LatticeConfig.from_dict(json.load(open("config.json")))
model = Braille256LatticeModel(config)
model.load_state_dict(torch.load("pytorch_model.bin"))

# Generate
input_ids = torch.tensor([[0x28, 0x29, 0x2A]])  # Some Braille tokens
output = model.generate(input_ids, max_length=100)

Training

python train_lattice_v6.py \
    --corpus corpus/braille_multimodal_corpus.txt \
    --tokenizer tokenizers/braille_8dot_32k/braille_8dot_32k.model \
    --output models/braille256_v6_lattice \
    --steps 10000

Training Options

Flag Description
--no-lattice-attention Disable lattice attention (ablation)
--no-lattice-embeddings Disable lattice embeddings (ablation)
--no-morph-regularization Disable morphological regularization (ablation)

Model Family

Version Focus Parameters Key Feature
v1-v3 6-dot Braille ~10M Basic Braille LM
v4 8-dot Braille 29.9M Full byte encoding
v5 Multimodal 11.5M TEXT/IMAGE/AUDIO/BINARY
v6 Lattice-aware 11.5M Hamming attention, morphological regularization, balanced multimodal corpus

Why Lattice-Aware?

Standard LLMs treat tokens as arbitrary symbols. braille256-v6 knows that:

  1. Braille cells form a lattice: 256 elements with meet (∧) and join (∨)
  2. Similar cells should have similar representations: Hamming distance matters
  3. Morphological operations preserve meaning: Erosion/dilation are semantic
  4. Tactile quality is measurable: Haptic metrics evaluate output quality

This makes v6 the first LLM designed for tactile-first AI.

Citation

@misc{braille256v6,
  author = {Barrett, Ryan},
  title = {braille256-v6: Lattice-Aware Multimodal Braille Model},
  year = {2025},
  publisher = {HuggingFace},
  url = {https://huggingface.co/ryanscottbarrett/braille256-v6}
}

License

MIT

Links


The first LLM where Braille is not just the output format, but the computational substrate.

Downloads last month
35
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support