--- title: MoireFormer Chat emoji: 🌊 colorFrom: blue colorTo: indigo sdk: gradio app_file: app.py pinned: false license: mit --- # MoireFormer (104.9M Proof-of-Concept) This repository hosts the PyTorch weights **moire_phase2_weights_final.pt** for MoireFormer, a neural network architecture that replaces standard dot-product attention with **Moiré phase-interference wave mechanics**. Instead of computing attention via Q · K^T, this model splits token embeddings into amplitude and phase and computes attention through geometric wave resonance. GitHub Code: https://github.com/anttiluode/MoireFormer Theory: https://github.com/anttiluode/Geometric-Neuron --- ## Model Details Architecture: MoireGPT (custom transformer) Parameters: 104.9M Structure: - 8 layers - 8 heads - 768 embedding dimension Capabilities: - English / Spanish syntax - conversational structure - instruction following Note: This is a **proof-of-substrate model**, not a factual knowledge model. --- ## How To Run This model cannot be loaded with `AutoModel`. It must run through the custom architecture. ### 1 Clone repo git clone https://github.com/anttiluode/MoireFormer.git cd MoireFormer ### 2 Install dependencies pip install torch transformers datasets ### 3 Download weights Download: https://huggingface.co/Aluode/MoireFormer/blob/main/moire_phase2_weights_final.pt Place the file inside the repo folder. ### 4 Run chat interface python moire_chat.py --weights moire_phase2_weights_final.pt --size large --- ## Training Curriculum Phase 1 15 epochs on Dolly-15k, WikiText-2, OpenAssistant. Phase 2 5 epochs on Guanaco dataset. The experiment demonstrates that **wave-field attention can learn discrete language syntax via phase geometry**. --- ## Disclaimer This is an experimental architecture exploring biological wave-field computation in neural networks. At 100M parameters it will hallucinate factual information.