Spaces:
Sleeping
A newer version of the Gradio SDK is available:
6.9.0
title: MoireFormer Chat
emoji: 🌊
colorFrom: blue
colorTo: indigo
sdk: gradio
app_file: app.py
pinned: false
license: mit
MoireFormer (104.9M Proof-of-Concept)
This repository hosts the PyTorch weights moire_phase2_weights_final.pt for MoireFormer, a neural network architecture that replaces standard dot-product attention with Moiré phase-interference wave mechanics.
Instead of computing attention via Q · K^T, this model splits token embeddings into amplitude and phase and computes attention through geometric wave resonance.
GitHub Code: https://github.com/anttiluode/MoireFormer
Theory: https://github.com/anttiluode/Geometric-Neuron
Model Details
Architecture: MoireGPT (custom transformer)
Parameters: 104.9M
Structure:
- 8 layers
- 8 heads
- 768 embedding dimension
Capabilities:
- English / Spanish syntax
- conversational structure
- instruction following
Note: This is a proof-of-substrate model, not a factual knowledge model.
How To Run
This model cannot be loaded with AutoModel.
It must run through the custom architecture.
1 Clone repo
git clone https://github.com/anttiluode/MoireFormer.git
cd MoireFormer
2 Install dependencies
pip install torch transformers datasets
3 Download weights
Download:
https://huggingface.co/Aluode/MoireFormer/blob/main/moire_phase2_weights_final.pt
Place the file inside the repo folder.
4 Run chat interface
python moire_chat.py --weights moire_phase2_weights_final.pt --size large
Training Curriculum
Phase 1
15 epochs on Dolly-15k, WikiText-2, OpenAssistant.
Phase 2
5 epochs on Guanaco dataset.
The experiment demonstrates that wave-field attention can learn discrete language syntax via phase geometry.
Disclaimer
This is an experimental architecture exploring biological wave-field computation in neural networks.
At 100M parameters it will hallucinate factual information.