| --- |
| base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0 |
| library_name: llama.cpp |
| pipeline_tag: text-generation |
| tags: |
| - tinyllama |
| - email-reply |
| - gguf |
| - ollama |
| - local-ai |
| license: mit |
| language: |
| - en |
| --- |
| |
| # TinyLlama Email Reply Generator |
|
|
| A **fine-tuned TinyLlama model** for generating professional email replies from incoming emails. |
|
|
| The adapter was trained on the **Enron Email Reply Dataset** to learn professional communication patterns. |
|
|
| --- |
|
|
| ## Model Overview |
|
|
| * **Base Model:** TinyLlama/TinyLlama-1.1B-Chat-v1.0 |
| * **Format:** GGUF (Q4_K_M quantization) |
| * **Size:** ~667 MB |
| * **Task:** Email reply generation |
| * **Language:** English |
|
|
| --- |
|
|
| ## Quick Start with GGUF |
|
|
| ### Using Ollama (Recommended) |
|
|
| ```bash |
| # Pull the model from Hugging Face |
| huggingface-cli download AshankGupta/tinyllama-email-reply tinyllama-chat.Q4_K_M.gguf --local-dir ./model |
| |
| # Or download directly |
| curl -L -o model.gguf "https://huggingface.co/AshankGupta/tinyllama-email-reply/resolve/main/tinyllama-chat.Q4_K_M.gguf" |
| |
| # Create Ollama model |
| ollama create tinyllama-email-reply -f Modelfile |
| |
| # Run |
| ollama run tinyllama-email-reply |
| ``` |
|
|
| ### Using llama.cpp |
|
|
| ```bash |
| # Download GGUF file from Hugging Face |
| curl -L -o tinyllama-email-reply.Q4_K_M.gguf \ |
| "https://huggingface.co/AshankGupta/tinyllama-email-reply/resolve/main/tinyllama-chat.Q4_K_M.gguf" |
| |
| # Run inference |
| ./llama.cpp/llama-cli -m tinyllama-email-reply.Q4_K_M.gguf -p "Write a professional email reply to: Can you send the invoice by tomorrow?" |
| ``` |
|
|
| ### Using Python |
|
|
| ```python |
| from llama_cpp import Llama |
| |
| model = Llama( |
| model_path="tinyllama-chat.Q4_K_M.gguf", |
| n_ctx=1024, |
| ) |
| |
| prompt = """You are an AI assistant that writes professional email replies. |
| |
| Email: |
| Can you send the invoice by tomorrow? |
| |
| Reply:""" |
| |
| output = model(prompt, max_tokens=120) |
| print(output["choices"][0]["text"]) |
| ``` |
|
|
| --- |
|
|
| ## Intended Use |
|
|
| This model is designed for: |
|
|
| * Email reply suggestion systems |
| * AI productivity tools |
| * Email assistants |
| * Local AI workflows |
| * Research on small language models |
|
|
| --- |
|
|
| ## Training Dataset |
|
|
| The model was trained using the **Enron Email Reply Dataset**, which contains real-world corporate email conversations. |
|
|
| Dataset characteristics: |
|
|
| * ~15,000 email-reply pairs |
| * Business and professional communication |
| * Cleaned and formatted into instruction-style prompts |
|
|
| Training format example: |
|
|
| ``` |
| Instruction: |
| Generate a professional email reply. |
| |
| Email: |
| Can you send the project report by tomorrow? |
| |
| Reply: |
| Sure, I will send the report by tomorrow. |
| ``` |
|
|
| --- |
|
|
| ## Training Details |
|
|
| * **Fine-tuning technique:** LoRA |
| * **Training framework:** Unsloth |
| * **Sequence length:** 512 tokens |
| * **Optimizer:** AdamW |
| * **Base architecture:** TinyLlama 1.1B |
| * **Quantization:** Q4_K_M |
|
|
| --- |
|
|
| ## Example |
|
|
| **Input Email** |
|
|
| ``` |
| Hi, |
| |
| Can you send the invoice by tomorrow? |
| ``` |
|
|
| **Generated Reply** |
|
|
| ``` |
| Sure, I will send the invoice by tomorrow. |
| ``` |
|
|
| --- |
|
|
| ## Limitations |
|
|
| * The model may produce generic replies. |
| * Performance is limited by the small size of the base model. |
| * It may occasionally generate repetitive outputs. |
| * Not suitable for sensitive or confidential communications. |
|
|
| --- |
|
|
| ## License |
|
|
| This model follows the license of the base model: |
|
|
| TinyLlama License |
|
|
| Please review the base model license before commercial usage. |
|
|
| --- |
|
|
| ## Acknowledgements |
|
|
| * TinyLlama team for the base model |
| * Unsloth for efficient LoRA training |
| * Enron Email Dataset for training data |