You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

I am still learning

Model Description

This is a toy language model trained from scratch for educational purposes.

Model Type: Causal Language Model Architecture: Transformer decoder-only Parameters: 30,044,544 Context Length: 256 tokens Vocabulary Size: 50257

Training Details

Architecture

  • Layers: 6
  • Hidden Size: 384
  • Attention Heads: 6
  • Dropout: 0.1

Training Configuration

  • Batch Size: 8
  • Learning Rate: 0.0003
  • Optimizer: AdamW
  • Final Loss: 8.232668980683012

Usage

Quick Start

import torch
from transformers import AutoModel, AutoTokenizer

model = AutoModel.from_pretrained("Sak2004/I-am-still-learning", trust_remote_code=True)
tokenizer = AutoTokenizer.from_pretrained("Sak2004/I-am-still-learning")

inputs = tokenizer("The benefits of renewable energy include", return_tensors="pt")
with torch.no_grad():
    outputs = model.generate(inputs.input_ids, max_length=100, temperature=0.8)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Limitations

  • This is an educational/toy model and is not intended for production use.
  • Generated text may be incoherent or factually incorrect.
  • Trained on a limited dataset; performance is expected to be modest.

License

This model is released under the MIT License.

Citation

If you use this model, please cite:

@misc{toyllm_2026,
  title={I am still learning - A Toy Language Model},
  author={Sak2004},
  year={2026},
  url={https://huggingface.co/Sak2004/I-am-still-learning}
}
Downloads last month
12
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train Sak2004/I-am-still-learning