Toy Models to Study
Collection
9 items • Updated • 2
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("nilq/lua-mistral-1L-mini")
model = AutoModelForCausalLM.from_pretrained("nilq/lua-mistral-1L-mini")This model is a mini single-layer Mistral model pre-trained on on the nilq/small-lua-stack dataset.
It achieves the following results on the evaluation set:
This model might contain some very simple model of Lua.
Let's see if we can find some interesting stuff inside this model.
Trained on the Lua subset of The Stack.
The following hyperparameters were used during training:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="nilq/lua-mistral-1L-mini")