--- license: apache-2.0 datasets: - Salesforce/wikitext language: - en metrics: - perplexity pipeline_tag: text-generation tags: - wavelet - attention-free - no-attention - attentionfree - FWHT - walsh-hadamard - PKM - product-key-memory - causal-lm - text-generation - sub-quadratic --- # WaveletLM WaveletLM is a fully causal, attention-free language model that mixes tokens through learned lifting wavelet decomposition, a Fast Walsh-Hadamard Transform, per-scale gated spectral mixing with SwiGLU activation, an inverse FWHT, and wavelet reconstruction. Combined with expanded MLPs and sparse product-key memory, this yields an architecture with no attention and O(n log n) scaling in sequence length. Full code, training details, ablations, and documentation: [github.com/ramongougis/WaveletLM](https://github.com/ramongougis/WaveletLM) ## Results | Dataset | Params | Perplexity | BPB | |---------|--------|------------|-----| | WikiText-103 | 883M | 23.8 | 1.0140 | | PG-19 (1 epoch) | 808M | 27.4 | 1.0853 | ## How to Use ```python import torch from huggingface_hub import hf_hub_download # Download the checkpoint ckpt_path = hf_hub_download(repo_id="ragou19/WaveletLM", filename="best_model.pt") ``` Then follow the instructions in the GitHub repo to load and run: https://github.com/ramongougis/WaveletLM ## Architecture ![WaveletLM architecture](https://raw.githubusercontent.com/ramongougis/WaveletLM/main/assets/waveletlm-architecture.svg) ## Training - Trained on a single RTX 5090 for 5 epochs - WikiText-103: best PPL of 23.749 with mean PPL of 23.818 across 3 seeds. - PG-19: PPL of 27.40 (single seed). - VRAM required: 18.3 GB. - Time to train: 16 hours 15 minutes. ## Generation - VRAM: 5.0 GB by default, 4.5 GB with `--ptq8` enabled. - Can set `compile:false` to save 0.5-1 GB, but it's slower. - 28.8 tokens/s. on a 5090 by default. - Future enhancements expected to increase speed by up to 120%. ## Logs See runs.md for the full training history. ## License Apache 2.0. See LICENSE.