tiny-llama / MODEL_CARD.md
manojredhat's picture
Correct Tiny LLaMA model metadata
756e974 verified
metadata
library_name: transformers
license: apache-2.0

Tiny LLaMA

A 6.27M parameter LLaMA-style causal language model trained on TinyStories.

Model Specifications

Property Value
Parameters 6,270,624
Layers 6
Attention Heads 6
Key/Value Heads 6
Head Dimension 48
Hidden Size 288
Intermediate Size 768
Vocabulary Size 512
Training Sequence Length 256
Data Type float32

Intended Use

  • TinyStories-style text generation
  • Educational examples
  • Small-model research
  • ASHA backend inference testing

Out-of-Scope Uses

  • Production deployments
  • Knowledge-intensive tasks
  • Long-form generation
  • Multilingual generation

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

tokenizer = AutoTokenizer.from_pretrained("manojredhat/tiny-llama")
model = AutoModelForCausalLM.from_pretrained("manojredhat/tiny-llama")

inputs = tokenizer("Once upon a time", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=40, do_sample=False)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))