rogersam's picture
Update README.md
b4ce18e verified
---
base_model: unsloth/tinyllama-bnb-4bit
library_name: peft
pipeline_tag: text-generation
tags:
- base_model:adapter:unsloth/tinyllama-bnb-4bit
- lora
- transformers
- unsloth
- instruction-tuned
- tamil
- english
---
# TinyLlama Instruct Lite v1
## πŸ“Œ Model Summary
`rogersam/tinyllama-instruct-lite-v1` is a **LoRA fine-tuned TinyLlama model** using [Unsloth](https://github.com/unslothai/unsloth).
It is designed for **instruction-following tasks** in **English + Tamil**, such as:
- General Q&A
- Summarization
- Basic math & reasoning
- English ↔ Tamil translation
This project demonstrates how a **lightweight 1B model** can be adapted for multiple domains with limited resources.
---
## πŸ”Ž Model Details
- **Developed by:** Roger Samuel J ([Hugging Face Profile](https://huggingface.co/rogersam))
- **Model type:** Causal LM (decoder-only)
- **Languages:** English, Tamil
- **License:** Same as base model (TinyLlama)
- **Fine-tuned from:** `unsloth/tinyllama-bnb-4bit`
- **Method:** LoRA via PEFT + Unsloth
---
## πŸ“‚ Model Sources
- **Model Repo:** [rogersam/tinyllama-instruct-lite-v1](https://huggingface.co/rogersam/tinyllama-instruct-lite-v1)
- **Base Model:** [unsloth/tinyllama-bnb-4bit](https://huggingface.co/unsloth/tinyllama-bnb-4bit)
---
## πŸ’‘ Uses
### Direct Use
- Running lightweight instruction tasks on CPU/GPU
- Translating English ↔ Tamil sentences
- Answering short questions and reasoning queries
- Summarizing small texts
### Out-of-Scope
- Sensitive decision-making (finance, healthcare, law)
- Long context generation (>512 tokens)
- Production-grade chatbots
---
## ⚠️ Bias, Risks & Limitations
- Small dataset β†’ may hallucinate facts
- Not aligned for safety or toxicity filtering
- Limited Tamil coverage (basic sentences only)
**Recommendation:** Use for demo & educational purposes only.
---
## πŸš€ How to Get Started
```python
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
model_id = "rogersam/tinyllama-instruct-lite-v1"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto")
pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
prompt = "Translate English to Tamil: How are you?"
print(pipe(prompt, max_new_tokens=50)[0]["generated_text"])