SallySims commited on
Commit
8ba3175
·
verified ·
1 Parent(s): 62bd775

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +77 -23
README.md CHANGED
@@ -8,8 +8,9 @@ base_model:
8
 
9
  # Model Card for AnthroBot (Llama-3.2-1B-Instruct Fine-tuned)
10
 
11
- <!-- This model is a fine-tuned version of meta-llama/Llama-3.2-1B-Instruct, adapted for reasoning and generating contextual insights from anthropometric data (e.g., age, sex, weight, height, waist circumference). It can summarize or comment on health-related metrics conversationally. -->
12
-
 
13
 
14
 
15
  ## Model Details
@@ -20,19 +21,19 @@ base_model:
20
 
21
  This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
22
 
23
- - **Developed by:** [More Information Needed]
24
- - **Funded by [optional]:** [More Information Needed]
25
- - **Shared by [optional]:** [More Information Needed]
26
- - **Model type:** [More Information Needed]
27
- - **Language(s) (NLP):** [More Information Needed]
28
- - **License:** [More Information Needed]
29
- - **Finetuned from model [optional]:** [More Information Needed]
30
 
31
  ### Model Sources [optional]
32
 
33
  <!-- Provide the basic links for the model. -->
34
 
35
- - **Repository:** [More Information Needed]
36
  - **Paper [optional]:** [More Information Needed]
37
  - **Demo [optional]:** [More Information Needed]
38
 
@@ -44,23 +45,34 @@ This is the model card of a 🤗 transformers model that has been pushed on the
44
 
45
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
46
 
 
 
 
47
  [More Information Needed]
48
 
49
  ### Downstream Use [optional]
50
 
51
  <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
52
-
 
53
  [More Information Needed]
54
 
55
  ### Out-of-Scope Use
56
 
57
  <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
 
 
 
 
 
58
 
59
  [More Information Needed]
60
 
61
  ## Bias, Risks, and Limitations
62
 
63
  <!-- This section is meant to convey both technical and sociotechnical limitations. -->
 
 
64
 
65
  [More Information Needed]
66
 
@@ -68,12 +80,26 @@ This is the model card of a 🤗 transformers model that has been pushed on the
68
 
69
  <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
70
 
71
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
72
 
73
  ## How to Get Started with the Model
74
 
75
  Use the code below to get started with the model.
76
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
77
  [More Information Needed]
78
 
79
  ## Training Details
@@ -81,6 +107,8 @@ Use the code below to get started with the model.
81
  ### Training Data
82
 
83
  <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
 
 
84
 
85
  [More Information Needed]
86
 
@@ -90,12 +118,28 @@ Use the code below to get started with the model.
90
 
91
  #### Preprocessing [optional]
92
 
93
- [More Information Needed]
94
 
95
 
96
  #### Training Hyperparameters
97
 
98
  - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
99
 
100
  #### Speeds, Sizes, Times [optional]
101
 
@@ -112,6 +156,8 @@ Use the code below to get started with the model.
112
  #### Testing Data
113
 
114
  <!-- This should link to a Dataset Card if possible. -->
 
 
115
 
116
  [More Information Needed]
117
 
@@ -124,12 +170,14 @@ Use the code below to get started with the model.
124
  #### Metrics
125
 
126
  <!-- These are the evaluation metrics being used, ideally with a description of why. -->
 
127
 
128
  [More Information Needed]
129
 
130
  ### Results
131
 
132
- [More Information Needed]
 
133
 
134
  #### Summary
135
 
@@ -147,17 +195,17 @@ Use the code below to get started with the model.
147
 
148
  Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
149
 
150
- - **Hardware Type:** [More Information Needed]
151
- - **Hours used:** [More Information Needed]
152
- - **Cloud Provider:** [More Information Needed]
153
- - **Compute Region:** [More Information Needed]
154
- - **Carbon Emitted:** [More Information Needed]
155
 
156
  ## Technical Specifications [optional]
157
 
158
  ### Model Architecture and Objective
159
 
160
- [More Information Needed]
161
 
162
  ### Compute Infrastructure
163
 
@@ -165,15 +213,21 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
165
 
166
  #### Hardware
167
 
168
- [More Information Needed]
169
 
170
  #### Software
171
 
172
- [More Information Needed]
173
 
174
  ## Citation [optional]
175
 
176
  <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
 
 
 
 
 
 
177
 
178
  **BibTeX:**
179
 
@@ -199,4 +253,4 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
199
 
200
  ## Model Card Contact
201
 
202
- [More Information Needed]
 
8
 
9
  # Model Card for AnthroBot (Llama-3.2-1B-Instruct Fine-tuned)
10
 
11
+ <!-- Provide a longer summary of what this model is. -->
12
+ This model is a fine-tuned version of meta-llama/Llama-3.2-1B-Instruct, adapted for reasoning and generating contextual insights from anthropometric data (e.g., age, sex, weight, height, waist circumference).
13
+ It can summarise or comment on health-related metrics conversationally.
14
 
15
 
16
  ## Model Details
 
21
 
22
  This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
23
 
24
+ - **Developed by:** Sally S. Simmons
25
+ - **Funded by [optional]:** NA
26
+ - **Shared by [optional]:** https://huggingface.co/SallySims
27
+ - **Model type:** Causal Language Model (LLM) with Instruction Tuning
28
+ - **Language(s) (NLP):** English
29
+ - **License:** Apache 2.0 (or specify if different)
30
+ - **Finetuned from model [optional]:** meta-llama/Llama-3.2-1B-Instruct
31
 
32
  ### Model Sources [optional]
33
 
34
  <!-- Provide the basic links for the model. -->
35
 
36
+ - **Repository:** https://huggingface.co/SallySims/AnthroBot_Model_Lora
37
  - **Paper [optional]:** [More Information Needed]
38
  - **Demo [optional]:** [More Information Needed]
39
 
 
45
 
46
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
47
 
48
+ The model is intended to analyze structured health-related user inputs and return conversational,
49
+ personalized feedback.It is designed for educational, wellness, or research purposes.
50
+
51
  [More Information Needed]
52
 
53
  ### Downstream Use [optional]
54
 
55
  <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
56
+ This model can be incorporated into chatbot systems or mobile health platforms that require
57
+ health-data-aware natural language interaction.
58
  [More Information Needed]
59
 
60
  ### Out-of-Scope Use
61
 
62
  <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
63
+ *Medical diagnosis or treatment
64
+
65
+ *Critical healthcare decision-making
66
+
67
+ *Inputs in languages other than English
68
 
69
  [More Information Needed]
70
 
71
  ## Bias, Risks, and Limitations
72
 
73
  <!-- This section is meant to convey both technical and sociotechnical limitations. -->
74
+ The model is trained on 20000 observations based on anthropometric data collected during the WHO STEPS survey and not in clinical settings.
75
+ Outputs may reflect biases present in the training prompts or may misinterpret edge cases.
76
 
77
  [More Information Needed]
78
 
 
80
 
81
  <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
82
 
83
+ Seek professional guidance in addition to the outcomes produced by the model
84
 
85
  ## How to Get Started with the Model
86
 
87
  Use the code below to get started with the model.
88
 
89
+ from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
90
+
91
+ model_id = "your-username/AnthroBot"
92
+
93
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
94
+ model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto")
95
+
96
+ pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
97
+
98
+ input_text = "Age: 30, Sex: female, Height: 150.5 cm, Weight: 75.3 kg, WC: 68.0 cm"
99
+ output = pipe(input_text, max_new_tokens=150, do_sample=True)
100
+ print(output[0]['generated_text'])
101
+
102
+
103
  [More Information Needed]
104
 
105
  ## Training Details
 
107
  ### Training Data
108
 
109
  <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
110
+ Custom curated structured anthropometric prompts designed to simulate
111
+ health-focused instruction-following behavior.
112
 
113
  [More Information Needed]
114
 
 
118
 
119
  #### Preprocessing [optional]
120
 
121
+ Prompts were normalised for consistent numerical formats and tokenization performance.
122
 
123
 
124
  #### Training Hyperparameters
125
 
126
  - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
127
+ Epochs: 5
128
+
129
+ Batch size: 2 (accumulation: 4 steps)
130
+
131
+ Learning rate: 2e-4
132
+
133
+ Precision: Mixed precision (fp16 / bf16)
134
+
135
+ LoRA Parameters:
136
+
137
+ r=16, alpha=32, dropout=0.05
138
+
139
+ Quantization
140
+ 4-bit quantization using BitsAndBytesConfig
141
+
142
+ Enabled llm_int8_enable_fp32_cpu_offload
143
 
144
  #### Speeds, Sizes, Times [optional]
145
 
 
156
  #### Testing Data
157
 
158
  <!-- This should link to a Dataset Card if possible. -->
159
+ Evaluation performed on held-out anthropometricindices and recommendations prompts
160
+ with expected interpretive outputs.
161
 
162
  [More Information Needed]
163
 
 
170
  #### Metrics
171
 
172
  <!-- These are the evaluation metrics being used, ideally with a description of why. -->
173
+ Human-judged relevance, clarity, and accuracy.
174
 
175
  [More Information Needed]
176
 
177
  ### Results
178
 
179
+ Manual inspection shows clear, concise, and useful summaries in the majority of cases.
180
+ Some rare edge cases may produce vague or overly generic responses.
181
 
182
  #### Summary
183
 
 
195
 
196
  Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
197
 
198
+ - **Hardware Type:** NVIDIA T4 GPU
199
+ - **Hours used:** ~ 2 hours
200
+ - **Cloud Provider:** Google Colab
201
+ - **Compute Region:** USA
202
+ - **Carbon Emitted:** ~1.2 kg CO₂eq (approx.)
203
 
204
  ## Technical Specifications [optional]
205
 
206
  ### Model Architecture and Objective
207
 
208
+ Decoder-only transformer based on the LLaMA 3.2B architecture.
209
 
210
  ### Compute Infrastructure
211
 
 
213
 
214
  #### Hardware
215
 
216
+ Google Colab (A100)
217
 
218
  #### Software
219
 
220
+ PyTorch, Hugging Face Transformers, PEFT, BitsAndBytes
221
 
222
  ## Citation [optional]
223
 
224
  <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
225
+ @misc{AnthroBot2025,
226
+ author = {Sally Sonia Simmons},
227
+ title = {AnthroBot: Instruction-Tuned LLaMA-3.2-1B for Anthropometric Reasoning},
228
+ year = {2025},
229
+ url = {https://huggingface.co/SallySimmons/AnthroBot_Model_Lora}
230
+ }
231
 
232
  **BibTeX:**
233
 
 
253
 
254
  ## Model Card Contact
255
 
256
+ simmonssallysonia@gmail.com