π§ Fine-Tuned Nano LLM
This repository contains a fine-tuned version of a small LLM trained using LoRA / QLoRA.
π Model Overview
- Base Model: Nano LLM
- Fine-tuning Method: LoRA / QLoRA
- Dataset: Custom synthetic + curated dataset
- Task: Text generation
- Framework: PyTorch + Hugging Face Transformers
- Training Environment: Google Colab Free Tier
π§ͺ Training Details
- Mixed precision (fp16 / nf4)
- Gradient checkpointing enabled
- LoRA rank & parameters carefully optimized for small hardware
- Trained on DatasetDict format with proper splits
π Files Included
This repo contains:
config.jsonadapter_config.jsonadapter_model.bintokenizer.jsontokenizer.modelspecial_tokens_map.jsongeneration_config.jsontraining_args.bin
How to load model
from modeling_tinygpt import TinyGPT import torch import json
with open("config.json") as f: cfg = json.load(f)
model = TinyGPT( vocab_size=cfg["vocab_size"], d_model=cfg["d_model"], n_heads=cfg["n_heads"], n_layers=cfg["n_layers"], d_ff=cfg["d_ff"], max_seq_len=cfg["max_seq_len"] )
state = torch.load("pytorch_model.bin", map_location="cpu") model.load_state_dict(state) model.eval()
π Benchmark & Notes
This model is designed for experimentation, education, and small-scale inference.
This model uses a custom architecture. Please load using trust_remote_code=True.
- Downloads last month
- 2
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support