am-om commited on
Commit
500663f
·
verified ·
1 Parent(s): fcce9ea

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +130 -130
README.md CHANGED
@@ -1,199 +1,199 @@
1
  ---
2
  library_name: transformers
3
- tags: []
 
 
 
 
 
 
 
 
 
4
  ---
5
 
6
- # Model Card for Model ID
7
-
8
- <!-- Provide a quick summary of what the model is/does. -->
9
 
 
 
10
 
 
11
 
12
  ## Model Details
13
 
14
  ### Model Description
15
 
16
- <!-- Provide a longer summary of what this model is. -->
 
17
 
18
- This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
 
 
19
 
20
- - **Developed by:** [More Information Needed]
21
- - **Funded by [optional]:** [More Information Needed]
22
- - **Shared by [optional]:** [More Information Needed]
23
- - **Model type:** [More Information Needed]
24
- - **Language(s) (NLP):** [More Information Needed]
25
- - **License:** [More Information Needed]
26
- - **Finetuned from model [optional]:** [More Information Needed]
27
 
28
- ### Model Sources [optional]
29
 
30
- <!-- Provide the basic links for the model. -->
31
 
32
- - **Repository:** [More Information Needed]
33
- - **Paper [optional]:** [More Information Needed]
34
- - **Demo [optional]:** [More Information Needed]
35
 
36
- ## Uses
37
 
38
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
 
40
  ### Direct Use
41
 
42
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
-
44
- [More Information Needed]
45
 
46
- ### Downstream Use [optional]
 
47
 
48
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
 
50
- [More Information Needed]
51
 
52
- ### Out-of-Scope Use
53
-
54
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
 
56
- [More Information Needed]
57
 
58
- ## Bias, Risks, and Limitations
 
 
 
59
 
60
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
 
62
- [More Information Needed]
 
 
 
63
 
64
- ### Recommendations
65
-
66
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
 
68
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
 
 
 
69
 
70
  ## How to Get Started with the Model
71
 
72
  Use the code below to get started with the model.
73
 
74
- [More Information Needed]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
75
 
76
  ## Training Details
77
 
78
  ### Training Data
79
 
80
- <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
-
82
- [More Information Needed]
83
 
84
  ### Training Procedure
85
 
86
- <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
-
88
- #### Preprocessing [optional]
89
-
90
- [More Information Needed]
91
-
92
 
93
  #### Training Hyperparameters
94
 
95
- - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
-
97
- #### Speeds, Sizes, Times [optional]
98
-
99
- <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
-
101
- [More Information Needed]
102
-
103
- ## Evaluation
104
-
105
- <!-- This section describes the evaluation protocols and provides the results. -->
106
-
107
- ### Testing Data, Factors & Metrics
108
-
109
- #### Testing Data
110
-
111
- <!-- This should link to a Dataset Card if possible. -->
112
-
113
- [More Information Needed]
114
-
115
- #### Factors
116
-
117
- <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
-
119
- [More Information Needed]
120
-
121
- #### Metrics
122
-
123
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
-
125
- [More Information Needed]
126
-
127
- ### Results
128
-
129
- [More Information Needed]
130
-
131
- #### Summary
132
-
133
-
134
-
135
- ## Model Examination [optional]
136
-
137
- <!-- Relevant interpretability work for the model goes here -->
138
-
139
- [More Information Needed]
140
 
141
  ## Environmental Impact
142
 
143
- <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
-
145
  Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
 
147
- - **Hardware Type:** [More Information Needed]
148
- - **Hours used:** [More Information Needed]
149
- - **Cloud Provider:** [More Information Needed]
150
- - **Compute Region:** [More Information Needed]
151
- - **Carbon Emitted:** [More Information Needed]
152
 
153
  ## Technical Specifications [optional]
154
 
155
  ### Model Architecture and Objective
156
 
157
- [More Information Needed]
158
 
159
  ### Compute Infrastructure
160
 
161
- [More Information Needed]
162
-
163
  #### Hardware
164
-
165
- [More Information Needed]
166
 
167
  #### Software
168
-
169
- [More Information Needed]
170
-
171
- ## Citation [optional]
172
-
173
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
-
175
- **BibTeX:**
176
-
177
- [More Information Needed]
178
-
179
- **APA:**
180
-
181
- [More Information Needed]
182
-
183
- ## Glossary [optional]
184
-
185
- <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
-
187
- [More Information Needed]
188
-
189
- ## More Information [optional]
190
-
191
- [More Information Needed]
192
 
193
  ## Model Card Authors [optional]
194
 
195
- [More Information Needed]
196
 
197
  ## Model Card Contact
198
 
199
- [More Information Needed]
 
 
 
1
  ---
2
  library_name: transformers
3
+ license: apache-2.0
4
+ language:
5
+ - en
6
+ base_model: google/gemma-3-1b-it
7
+ tags:
8
+ - gemma
9
+ - finetune
10
+ - qlora
11
+ - chatbot
12
+ - tars
13
  ---
14
 
15
+ # Model Card for TARS (Gemma 3 1B Fine-tune)
 
 
16
 
17
+ This is a fine-tuned version of `google/gemma-3-1b-it` trained to act as the **TARS astronaut assistant** from *Interstellar*.
18
+ It is designed to be professional for tasks but witty for off-topic chat, and its responses are guided by a simulated user emotion tag.
19
 
20
+ ---
21
 
22
  ## Model Details
23
 
24
  ### Model Description
25
 
26
+ This model is a QLoRA fine-tune of `google/gemma-3-1b-it` on a custom synthetic dataset.
27
+ The goal was to create a chatbot that embodies the **TARS persona**:
28
 
29
+ - **Task-Oriented:** Professional, direct, and helpful for mission-related queries.
30
+ - **Persona-Driven:** Witty, empathetic, or humorous for off-topic or personal chat.
31
+ - **Emotion-Aware:** The model's response style is influenced by a `[Detected Emotion: ...]` tag.
32
 
33
+ **Developed by:** (huggingface.co/am-om)
34
+ **Shared by:** (Om Singh)
35
+ **Model type:** Causal Language Model
36
+ **Language(s):** English (`en`)
37
+ **License:** apache-2.0
38
+ **Finetuned from model:** `google/gemma-3-1b-it`
 
39
 
40
+ ---
41
 
42
+ ## Model Sources (optional)
43
 
44
+ - **Repository:** [https://huggingface.co/am-om/tars_ai]
 
 
45
 
46
+ ---
47
 
48
+ ## Uses
49
 
50
  ### Direct Use
51
 
52
+ This model is intended for **direct use as a chatbot**, following a specific prompt format.
 
 
53
 
54
+ ⚠️ **Important:** This model requires a specific prompt format that includes a detected emotion.
55
+ Do **not** send raw text as the user query.
56
 
 
57
 
 
58
 
59
+ #### Prompt Format
 
 
60
 
61
+ The user turn *must* follow this structure:
62
 
63
+ ```
64
+ [Detected Emotion: {emotion}]
65
+ [User Query: {your_text_here}]
66
+ ```
67
 
68
+ **Example:**
69
 
70
+ ```
71
+ [Detected Emotion: anxious]
72
+ [User Query: Are we going to make it?]
73
+ ```
74
 
75
+ ### Out-of-Scope Use
 
 
76
 
77
+ This model is not intended for:
78
+ * Any use without the required `[Detected Emotion: ...]` and `[User Query: ...]` tags.
79
+ * Use as a base model for further fine-tuning.
80
+ * Any critical decision-making without human oversight.
81
 
82
  ## How to Get Started with the Model
83
 
84
  Use the code below to get started with the model.
85
 
86
+ ```python
87
+ from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
88
+ import torch
89
+
90
+ # Load the model from the Hub
91
+ model_id = "am-om/tars_ai"
92
+
93
+ model = AutoModelForCausalLM.from_pretrained(
94
+ model_id,
95
+ device_map="auto",
96
+ torch_dtype=torch.bfloat16
97
+ )
98
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
99
+
100
+ pipe = pipeline(
101
+ "text-generation",
102
+ model=model,
103
+ tokenizer=tokenizer
104
+ )
105
+
106
+ # --- Define your chat history ---
107
+ # The system prompt is automatically loaded from the tokenizer's chat template.
108
+ messages = []
109
+
110
+ # Example query
111
+ user_query = "I'm feeling a bit lonely out here."
112
+ emotion = "sad"
113
+
114
+ # Format the input correctly!
115
+ formatted_input = f"[Detected Emotion: {emotion}]\n[User Query: {user_query}]"
116
+
117
+ messages.append({"role": "user", "content": formatted_input})
118
+
119
+ # --- Generate the response ---
120
+ prompt = pipe.tokenizer.apply_chat_template(
121
+ messages,
122
+ tokenize=False,
123
+ add_generation_prompt=True
124
+ )
125
+
126
+ outputs = pipe(
127
+ prompt,
128
+ max_new_tokens=256,
129
+ do_sample=True,
130
+ temperature=0.7,
131
+ top_p=0.95,
132
+ pad_token_id=pipe.tokenizer.eos_token_id
133
+ )
134
+
135
+ # Extract and print just the new response
136
+ response = outputs[0]["generated_text"][len(prompt):].strip()
137
+ print(f"TARS: {response}")
138
+ ```
139
 
140
  ## Training Details
141
 
142
  ### Training Data
143
 
144
+ This model was fine-tuned on a custom, synthetically-generated dataset of 344 prompt/response pairs. The dataset was designed to teach the model to differentiate between task-oriented and persona-driven queries based on the emotion tag.
 
 
145
 
146
  ### Training Procedure
147
 
148
+ The model was fine-tuned using QLoRA for 3 epochs. The adapter (from checkpoint-156, the best-performing epoch) was then merged with the base model.
 
 
 
 
 
149
 
150
  #### Training Hyperparameters
151
 
152
+ * **Framework:** TRL (Transformer Reinforcement Learning)
153
+ * **Quantization:** 4-bit (bnb_4bit_quant_type="nf4")
154
+ * **LoRA `r`:** 16
155
+ * **LoRA `alpha`:** 32
156
+ * **LoRA `dropout`:** 0.05
157
+ * **Optimizer:** paged_adamw_8bit
158
+ * **Learning Rate:** 5e-5
159
+ * **LR Scheduler:** constant
160
+ * **Epochs:** 3
161
+ * **Batch Size:** 4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
162
 
163
  ## Environmental Impact
164
 
 
 
165
  Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
166
 
167
+ * **Hardware Type:** NVIDIA T4
168
+ * **Hours used:** ~4hours
169
+ * **Cloud Provider:** Google Colab
170
+ * **Compute Region:** (e.g., us-central1 - *check your Colab instance*)
171
+ * **Carbon Emitted:** ~5.5 g CO2eq (Estimated)
172
 
173
  ## Technical Specifications [optional]
174
 
175
  ### Model Architecture and Objective
176
 
177
+ This is a standard decoder-only Transformer (Gemma 3) fine-tuned with a Causal Language Modeling objective.
178
 
179
  ### Compute Infrastructure
180
 
 
 
181
  #### Hardware
182
+ * NVIDIA T4 16GB (Google Collab )
 
183
 
184
  #### Software
185
+ * `transformers`
186
+ * `trl`
187
+ * `bitsandbytes`
188
+ * `accelerate`
189
+ * `peft`
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
190
 
191
  ## Model Card Authors [optional]
192
 
193
+ (Om Singh)(huggingface.co/am-om)
194
 
195
  ## Model Card Contact
196
 
197
+ (huggingface.co/am-om)
198
+
199
+