MoireFormer-Chat / README.md
Aluode's picture
Update README.md
0c8ec8c verified
---
title: MoireFormer Chat
emoji: 🌊
colorFrom: blue
colorTo: indigo
sdk: gradio
app_file: app.py
pinned: false
license: mit
---
# MoireFormer (104.9M Proof-of-Concept)
This repository hosts the PyTorch weights **moire_phase2_weights_final.pt** for MoireFormer, a neural network architecture that replaces standard dot-product attention with **Moiré phase-interference wave mechanics**.
Instead of computing attention via Q · K^T, this model splits token embeddings into amplitude and phase and computes attention through geometric wave resonance.
GitHub Code:
https://github.com/anttiluode/MoireFormer
Theory:
https://github.com/anttiluode/Geometric-Neuron
---
## Model Details
Architecture: MoireGPT (custom transformer)
Parameters: 104.9M
Structure:
- 8 layers
- 8 heads
- 768 embedding dimension
Capabilities:
- English / Spanish syntax
- conversational structure
- instruction following
Note: This is a **proof-of-substrate model**, not a factual knowledge model.
---
## How To Run
This model cannot be loaded with `AutoModel`.
It must run through the custom architecture.
### 1 Clone repo
git clone https://github.com/anttiluode/MoireFormer.git
cd MoireFormer
### 2 Install dependencies
pip install torch transformers datasets
### 3 Download weights
Download:
https://huggingface.co/Aluode/MoireFormer/blob/main/moire_phase2_weights_final.pt
Place the file inside the repo folder.
### 4 Run chat interface
python moire_chat.py --weights moire_phase2_weights_final.pt --size large
---
## Training Curriculum
Phase 1
15 epochs on Dolly-15k, WikiText-2, OpenAssistant.
Phase 2
5 epochs on Guanaco dataset.
The experiment demonstrates that **wave-field attention can learn discrete language syntax via phase geometry**.
---
## Disclaimer
This is an experimental architecture exploring biological wave-field computation in neural networks.
At 100M parameters it will hallucinate factual information.