i3-tiny / README.md
FlameF0X's picture
Update README.md
f9723b3 verified
|
raw
history blame
2.75 kB
metadata
license: apache-2.0
language:
  - en
pipeline_tag: text-generation
library_name: transformers
tags:
  - i3-arhitecture

i3-tiny

i3-tiny is a compact, efficient character-level language model designed for experimentation and exploration in text generation. Despite its small size, it packs a surprising punch for creative and research-oriented tasks, generating sequences that are quirky, unpredictable, and full of “human-like” character-level errors.


Model Overview

i3-tiny is trained to predict the next character in a sequence, making it ideal for character-level language modeling, creative text generation, and research on lightweight, efficient models. Its small footprint allows rapid experimentation, even on modest hardware, and it provides a playground for studying how models learn patterns in sequences of characters.

The model is intentionally experimental — it’s not aligned, fact-checked, or polished. Instead, it showcases how a compact architecture can capture patterns in text, learn from repetition, and generate outputs that are sometimes surprisingly coherent, sometimes hilariously garbled.


Training Details

  • Dataset: ~45,830 characters (a curated text corpus repeated to improve exposure)
  • Vocabulary: 34 characters (all lowercased)
  • Sequence length: 128
  • Training iterations: 2,000
  • Batch size: 2
  • Optimizer: AdamW, learning rate 3e-4
  • Model parameters: 711,106
  • Performance notes: Each iteration takes roughly 400–500 ms; 100 iterations take ~45 s on average. Loss steadily decreased from 3.53 to 2.15 over training.

Example generation (iteration 1200):

Prompt: "The quick"
Generated: the quick efehn. dethe cans the fice the fpeens antary of eathetint, an thadat hitimes the and cow thig, and

These outputs capture the chaotic creativity of a character-level model: a mixture of readable words, invented forms, and surprising sequences.


Intended Uses

  • Character-level text generation experiments
  • Research and education: studying lightweight language models, sequence learning, and text modeling
  • Creative exploration: generating quirky text or procedural content for games, demos, or artistic projects

⚠️ i3-tiny is experimental and not intended for production or high-stakes applications. Text may be repetitive, nonsensical, or inconsistent.


Limitations

  • Small vocabulary and character-level modeling limit natural language fluency
  • Outputs are highly experimental and not fact-checked
  • Generated sequences can be repetitive or unexpectedly garbled
  • Not aligned or safety-checked

Model Weights

  • Stored in model.bin
  • Compatible with PyTorch