My Local LLM
Model Details
- Architecture: BERT / GPT / Custom
- Hidden Size: 256
- Layers: 6
- Heads: 4
- Vocabulary Size: 30,522
- Tokenizer: Word-level / BPE
Training Data
- Description of dataset (e.g., cricket commentary, Wikipedia subset)
- Preprocessing steps
Intended Use
- Text generation
- Question answering
- Classification
Limitations
- May hallucinate
- Biases present in training data
How to use
from transformers import AutoModel, AutoTokenizer
tok = AutoTokenizer.from_pretrained("username/my-llm")
model = AutoModel.from_pretrained("username/my-llm")
inputs = tok("Hello world!", return_tensors="pt")
outputs = model(**inputs)