Medical_Homework2 / README.md
Abeersherif's picture
Update README.md
172bd30 verified
---
library_name: transformers
pipeline_tag: text-generation
base_model: SmolAI/SmolLM2-1.7B
license: apache-2.0
language:
- en
tags:
- smolllm2
- finetuned
- medical
- homework
model_type: causal-lm
---
# Medical_Homework2 — Fine-Tuned SmolLM2-1.7B for Medical Reasoning
Medical_Homework2 is a fine-tuned version of SmolAI/SmolLM2-1.7B, trained specifically on structured medical question-answer data and short reasoning tasks.
The model aims to provide concise, accurate, and educational medical explanations suitable for students and basic learning purposes.
---
## Model Overview
This model is optimized for medical comprehension tasks such as:
- Short medical answers
- Step-by-step reasoning
- Explanations of conditions, symptoms, and basic physiology
- Educational or homework-style responses
It is not designed for professional medical diagnosis or treatment decisions.
---
## Intended Use
### Recommended Use Cases
- Medical homework and assignment assistance
- Explanation of medical concepts in simple language
- Introductory physiology and pathology topics
- Basic reasoning about medical questions
### Not Recommended
- Real-world clinical decision-making
- Emergency or diagnostic use
- Any situation requiring professional medical judgement
---
## Training Data
The model was fine-tuned using:
- Synthetic medical question-answer pairs
- Simplified educational medical explanations
- Instruction-answer examples
- Homework-style reasoning data
No real patient data or clinical records were used.
---
## Training Details
- Base model: SmolAI/SmolLM2-1.7B
- Fine-tuning objective: Causal language modeling
- Method: Full or LoRA fine-tuning (depending on your actual setup)
- Optimizer: AdamW
- Typical epochs: 1–3
If you want, a full training script section can be added.
---
## Usage Example
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
model_name = "Abeersherif/Medical_Homework2"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)
prompt = "Explain what type 2 diabetes is in simple terms."
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(
**inputs,
max_new_tokens=150,
temperature=0.7,
top_p=0.9,
)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))