SLIM-1
Collection
SLIM-1 from scratch models. Trained by me • 2 items • Updated
This repository contains intermediate checkpoints for the SLIM-1-base-chkp project. This is a Small Language Model (SLM) trained from scratch on high-quality educational and code data.
Since this is a Base model, it is designed for text completion rather than instruction-following. It is best used for further fine-tuning.
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("lumasik/SLIM-1-base-chkp2500")
tokenizer = AutoTokenizer.from_pretrained("lumasik/SLIM-1-base-chkp2500")
prompt = "The relationship between large language models and tokenization is"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=50)
print(tokenizer.decode(outputs[0]))