--- language: - en --- ## ChatGCLM-330M ChatGCLM Hero A high-performance language model architecture. --- ## Overview **ChatGCLM** is a generative language model that deviates from the traditional Transformer architecture by utilizing a hybrid approach of **Local** and **Global Convolutions**. By leveraging Fast Fourier Transforms (FFT) for global context, ChatGCLM achieves a massive receptive field with a fraction of the computational overhead associated with standard attention mechanisms. The architecture is designed for efficiency, speed, and high-quality generation, featuring a custom vocabulary reduction system that optimizes the embedding space for specific datasets. ## 📦 Installation Download this repository and extract it. --- ## Usage ### 1. Training the Model Place your `.txt` data files in the `data/` directory and run: ```bash python train_chatgclm.py ``` This script will build the vocabulary and train the foundation model ### 2. Sample from the model Run sample.py to generate text with the model ```bash python sample.py ``` --- ## Fine-tuning You may fine-tune the model by resuming training from a checkpoint, you may use a different dataset, you may also change parameters such as the learning rate, batch size, etc.

Built with ❤️ by AG