File size: 2,381 Bytes
b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e b0704e1 b4ce18e |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 |
---
base_model: unsloth/tinyllama-bnb-4bit
library_name: peft
pipeline_tag: text-generation
tags:
- base_model:adapter:unsloth/tinyllama-bnb-4bit
- lora
- transformers
- unsloth
- instruction-tuned
- tamil
- english
---
# TinyLlama Instruct Lite v1
## π Model Summary
`rogersam/tinyllama-instruct-lite-v1` is a **LoRA fine-tuned TinyLlama model** using [Unsloth](https://github.com/unslothai/unsloth).
It is designed for **instruction-following tasks** in **English + Tamil**, such as:
- General Q&A
- Summarization
- Basic math & reasoning
- English β Tamil translation
This project demonstrates how a **lightweight 1B model** can be adapted for multiple domains with limited resources.
---
## π Model Details
- **Developed by:** Roger Samuel J ([Hugging Face Profile](https://huggingface.co/rogersam))
- **Model type:** Causal LM (decoder-only)
- **Languages:** English, Tamil
- **License:** Same as base model (TinyLlama)
- **Fine-tuned from:** `unsloth/tinyllama-bnb-4bit`
- **Method:** LoRA via PEFT + Unsloth
---
## π Model Sources
- **Model Repo:** [rogersam/tinyllama-instruct-lite-v1](https://huggingface.co/rogersam/tinyllama-instruct-lite-v1)
- **Base Model:** [unsloth/tinyllama-bnb-4bit](https://huggingface.co/unsloth/tinyllama-bnb-4bit)
---
## π‘ Uses
### Direct Use
- Running lightweight instruction tasks on CPU/GPU
- Translating English β Tamil sentences
- Answering short questions and reasoning queries
- Summarizing small texts
### Out-of-Scope
- Sensitive decision-making (finance, healthcare, law)
- Long context generation (>512 tokens)
- Production-grade chatbots
---
## β οΈ Bias, Risks & Limitations
- Small dataset β may hallucinate facts
- Not aligned for safety or toxicity filtering
- Limited Tamil coverage (basic sentences only)
**Recommendation:** Use for demo & educational purposes only.
---
## π How to Get Started
```python
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
model_id = "rogersam/tinyllama-instruct-lite-v1"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto")
pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
prompt = "Translate English to Tamil: How are you?"
print(pipe(prompt, max_new_tokens=50)[0]["generated_text"])
|