metadata
license: apache-2.0
library_name: pytorch
tags:
- language-model
- causal-lm
- gpt
- from-scratch
- educational
pipeline_tag: text-generation
framework: pytorch
SimBot GPT (Level 1)
SimBot GPT is a from-scratch GPT-style language model implemented in PyTorch. This project is focused on learning LLM internals, not on instruction tuning or production use.
Model Overview
- Architecture: Decoder-only Transformer (GPT-like)
- Training Objective: Causal Language Modeling
- Dataset: Domain-specific text (Simdega / regional data)
- Purpose: Educational (understanding how LLMs work internally)
⚠️ This is a base language model, not instruction-tuned and not grounded with RAG.
Repository Contents
simbot.safetensors— model weights (safe & HF-recommended format)tokenizer.json— BPE tokenizerconfig.json— model hyperparametersmodel/simbot.py— model architecture (PyTorch)
Requirements (Inference Only)
The following packages are required to load and run the model:
torch==2.9.1
tokenizers==0.22.1
safetensors
Usage Example
import json
from safetensors.torch import load_file
from tokenizers import Tokenizer
from model.simbot import SIMGPT
# Load tokenizer
tokenizer = Tokenizer.from_file("tokenizer.json")
# Load config
with open("config.json") as f:
cfg = json.load(f)
# Build model
model = SIMGPT(
vocab_size=cfg["vocab_size"],
block_size=cfg["block_size"],
n_layers=cfg["n_layers"],
n_heads=cfg["n_heads"],
d_model=cfg["d_model"]
)
# Load weights
state_dict = load_file("simbot.safetensors")
model.load_state_dict(state_dict)
model.eval()
Prompting the Model
This model is a custom PyTorch implementation and does not support the Hugging Face inference widget.
Interactive Usage (Recommended)
python inference.py