| | --- |
| | library_name: transformers |
| | base_model: unsloth/tinyllama-bnb-4bit |
| | license: mit |
| | datasets: |
| | - yahma/alpaca-cleaned |
| | language: |
| | - en |
| | pipeline_tag: text-generation |
| | tags: |
| | - Instruct |
| | - TinyLlama |
| | --- |
| | |
| | # Steps to try the model: |
| |
|
| | ### prompt Template |
| | ```python |
| | alpaca_prompt = """Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request. |
| | |
| | ### Instruction: |
| | {} |
| | |
| | ### Input: |
| | {} |
| | |
| | ### Response: |
| | {}""" |
| | ``` |
| | ### load the model |
| |
|
| | ```python |
| | from transformers import AutoTokenizer, AutoModelForCausalLM |
| | |
| | tokenizer = AutoTokenizer.from_pretrained("damerajee/tinyllama-sft-small-v2") |
| | model = AutoModelForCausalLM.from_pretrained("damerajee/tinyllama-sft-small-v2") |
| | ``` |
| | ### Inference |
| |
|
| | ```python |
| | inputs = tokenizer( |
| | [ |
| | alpaca_prompt.format( |
| | "best places to visit in india", # instruction |
| | "", # input |
| | "", # output |
| | ) |
| | ]*1, return_tensors = "pt") |
| | |
| | outputs = model.generate(**inputs, max_new_tokens = 128, use_cache = True) |
| | tokenizer.batch_decode(outputs) |
| | ``` |
| |
|
| |
|
| |
|
| | # Model Information |
| | The base model [unsloth/tinyllama-bnb-4bit](https://huggingface.co/unsloth/tinyllama-bnb-4bit) was Instruct finetuned using [Unsloth](https://github.com/unslothai/unsloth) |
| |
|
| | # Model Limitations |
| | The model was trained on a very small dataset so it might not be as good ,will be training on larger dataset soon |
| | # Training Details |
| | The model was trained for 1 epoch on a free goggle colab which took about 1 hour and 30 mins approximately |