AshankGupta commited on
Commit
fb6106e
·
verified ·
1 Parent(s): a17e56d

Updated README.md File

Browse files
Files changed (1) hide show
  1. README.md +108 -137
README.md CHANGED
@@ -9,202 +9,173 @@ tags:
9
  - transformers
10
  - trl
11
  - unsloth
 
 
 
12
  ---
 
13
 
14
- # Model Card for Model ID
15
 
16
- <!-- Provide a quick summary of what the model is/does. -->
17
 
 
18
 
 
19
 
20
- ## Model Details
21
-
22
- ### Model Description
23
-
24
- <!-- Provide a longer summary of what this model is. -->
25
-
26
-
27
-
28
- - **Developed by:** [More Information Needed]
29
- - **Funded by [optional]:** [More Information Needed]
30
- - **Shared by [optional]:** [More Information Needed]
31
- - **Model type:** [More Information Needed]
32
- - **Language(s) (NLP):** [More Information Needed]
33
- - **License:** [More Information Needed]
34
- - **Finetuned from model [optional]:** [More Information Needed]
35
-
36
- ### Model Sources [optional]
37
-
38
- <!-- Provide the basic links for the model. -->
39
-
40
- - **Repository:** [More Information Needed]
41
- - **Paper [optional]:** [More Information Needed]
42
- - **Demo [optional]:** [More Information Needed]
43
-
44
- ## Uses
45
-
46
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
47
-
48
- ### Direct Use
49
 
50
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
51
 
52
- [More Information Needed]
53
 
54
- ### Downstream Use [optional]
55
 
56
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
57
 
58
- [More Information Needed]
 
 
 
 
59
 
60
- ### Out-of-Scope Use
61
 
62
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
 
 
63
 
64
- [More Information Needed]
65
 
66
- ## Bias, Risks, and Limitations
67
 
68
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
69
 
70
- [More Information Needed]
71
 
72
- ### Recommendations
 
 
73
 
74
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
75
 
76
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
 
 
77
 
78
- ## How to Get Started with the Model
 
79
 
80
- Use the code below to get started with the model.
 
 
81
 
82
- [More Information Needed]
83
 
84
  ## Training Details
85
 
86
- ### Training Data
87
-
88
- <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
89
-
90
- [More Information Needed]
91
-
92
- ### Training Procedure
93
-
94
- <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
95
-
96
- #### Preprocessing [optional]
97
-
98
- [More Information Needed]
99
-
100
-
101
- #### Training Hyperparameters
102
 
103
- - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
104
 
105
- #### Speeds, Sizes, Times [optional]
106
-
107
- <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
108
-
109
- [More Information Needed]
110
-
111
- ## Evaluation
112
-
113
- <!-- This section describes the evaluation protocols and provides the results. -->
114
-
115
- ### Testing Data, Factors & Metrics
116
-
117
- #### Testing Data
118
-
119
- <!-- This should link to a Dataset Card if possible. -->
120
-
121
- [More Information Needed]
122
-
123
- #### Factors
124
-
125
- <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
126
-
127
- [More Information Needed]
128
-
129
- #### Metrics
130
-
131
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
132
-
133
- [More Information Needed]
134
-
135
- ### Results
136
-
137
- [More Information Needed]
138
-
139
- #### Summary
140
 
 
141
 
 
142
 
143
- ## Model Examination [optional]
 
 
144
 
145
- <!-- Relevant interpretability work for the model goes here -->
146
 
147
- [More Information Needed]
 
148
 
149
- ## Environmental Impact
 
 
 
150
 
151
- <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
 
152
 
153
- Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
154
 
155
- - **Hardware Type:** [More Information Needed]
156
- - **Hours used:** [More Information Needed]
157
- - **Cloud Provider:** [More Information Needed]
158
- - **Compute Region:** [More Information Needed]
159
- - **Carbon Emitted:** [More Information Needed]
160
 
161
- ## Technical Specifications [optional]
 
162
 
163
- ### Model Architecture and Objective
 
164
 
165
- [More Information Needed]
 
166
 
167
- ### Compute Infrastructure
 
168
 
169
- [More Information Needed]
170
 
171
- #### Hardware
172
 
173
- [More Information Needed]
 
174
 
175
- #### Software
176
 
177
- [More Information Needed]
178
 
179
- ## Citation [optional]
180
 
181
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
 
182
 
183
- **BibTeX:**
 
184
 
185
- [More Information Needed]
186
 
187
- **APA:**
 
 
188
 
189
- [More Information Needed]
190
 
191
- ## Glossary [optional]
192
 
193
- <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
 
 
 
194
 
195
- [More Information Needed]
196
 
197
- ## More Information [optional]
198
 
199
- [More Information Needed]
200
 
201
- ## Model Card Authors [optional]
202
 
203
- [More Information Needed]
204
 
205
- ## Model Card Contact
206
 
207
- [More Information Needed]
208
- ### Framework versions
209
 
210
- - PEFT 0.18.1
 
 
 
9
  - transformers
10
  - trl
11
  - unsloth
12
+ license: mit
13
+ language:
14
+ - en
15
  ---
16
+ # TinyLlama Email Reply Generator (LoRA)
17
 
18
+ A **LoRA fine-tuned TinyLlama model** for generating professional email replies from incoming emails.
19
 
20
+ The adapter was trained on the **Enron Email Reply Dataset** to learn professional communication patterns.
21
 
22
+ ---
23
 
24
+ ## Model Overview
25
 
26
+ * **Base Model:** TinyLlama/TinyLlama-1.1B-Chat-v1.0
27
+ * **Fine-tuning Method:** LoRA (Low Rank Adaptation)
28
+ * **Task:** Email reply generation
29
+ * **Language:** English
30
+ * **Framework:** Unsloth + Transformers
31
+ * **Deployment:** Designed for local inference with FastAPI or Ollama
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32
 
33
+ This adapter allows users to generate contextual email replies without relying on large commercial APIs.
34
 
35
+ ---
36
 
37
+ ## Intended Use
38
 
39
+ This model is designed for:
40
 
41
+ * Email reply suggestion systems
42
+ * AI productivity tools
43
+ * Email assistants
44
+ * Local AI workflows
45
+ * Research on small language models
46
 
47
+ Example applications include:
48
 
49
+ * Gmail Smart Reply style systems
50
+ * Chrome extensions for automated responses
51
+ * Offline AI assistants
52
 
53
+ ---
54
 
55
+ ## Training Dataset
56
 
57
+ The model was trained using the **Enron Email Reply Dataset**, which contains real-world corporate email conversations.
58
 
59
+ Dataset characteristics:
60
 
61
+ * ~15,000 email–reply pairs
62
+ * Business and professional communication
63
+ * Cleaned and formatted into instruction-style prompts
64
 
65
+ Training format example:
66
 
67
+ ```
68
+ Instruction:
69
+ Generate a professional email reply.
70
 
71
+ Email:
72
+ Can you send the project report by tomorrow?
73
 
74
+ Reply:
75
+ Sure, I will send the report by tomorrow.
76
+ ```
77
 
78
+ ---
79
 
80
  ## Training Details
81
 
82
+ * **Fine-tuning technique:** LoRA
83
+ * **Training framework:** Unsloth
84
+ * **Sequence length:** 512 tokens
85
+ * **Optimizer:** AdamW
86
+ * **Base architecture:** TinyLlama 1.1B
 
 
 
 
 
 
 
 
 
 
 
87
 
88
+ Only a small percentage of parameters were trained via LoRA adapters, enabling efficient training on consumer GPUs.
89
 
90
+ ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
91
 
92
+ ## Usage
93
 
94
+ Install dependencies:
95
 
96
+ ```
97
+ pip install unsloth transformers accelerate bitsandbytes
98
+ ```
99
 
100
+ Load the model:
101
 
102
+ ```python
103
+ from unsloth import FastLanguageModel
104
 
105
+ model, tokenizer = FastLanguageModel.from_pretrained(
106
+ "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
107
+ load_in_4bit=True
108
+ )
109
 
110
+ model.load_adapter("ashankgupta/tinyllama-email-reply")
111
+ ```
112
 
113
+ Example inference:
114
 
115
+ ```python
116
+ email = """
117
+ Hi,
 
 
118
 
119
+ Can you send the invoice by tomorrow?
120
+ """
121
 
122
+ prompt = f"""
123
+ You are an AI assistant that writes professional email replies.
124
 
125
+ Email:
126
+ {email}
127
 
128
+ Reply:
129
+ """
130
 
131
+ inputs = tokenizer(prompt, return_tensors="pt")
132
 
133
+ output = model.generate(**inputs, max_new_tokens=120)
134
 
135
+ print(tokenizer.decode(output[0], skip_special_tokens=True))
136
+ ```
137
 
138
+ ---
139
 
140
+ ## Example
141
 
142
+ **Input Email**
143
 
144
+ ```
145
+ Hi,
146
 
147
+ Can you send the invoice by tomorrow?
148
+ ```
149
 
150
+ **Generated Reply**
151
 
152
+ ```
153
+ Sure, I will send the invoice by tomorrow.
154
+ ```
155
 
156
+ ---
157
 
158
+ ## Limitations
159
 
160
+ * The model may produce generic replies.
161
+ * Performance is limited by the small size of the base model.
162
+ * It may occasionally generate repetitive outputs.
163
+ * Not suitable for sensitive or confidential communications.
164
 
165
+ ---
166
 
167
+ ## License
168
 
169
+ This model follows the license of the base model:
170
 
171
+ TinyLlama License
172
 
173
+ Please review the base model license before commercial usage.
174
 
175
+ ---
176
 
177
+ ## Acknowledgements
 
178
 
179
+ * TinyLlama team for the base model
180
+ * Unsloth for efficient LoRA training
181
+ * Enron Email Dataset for training data