Oswestry-Instruct / README.md
DrAleML's picture
Update README.md
a95adca verified
---
library_name: transformers
tags:
- medical
- spanish
- llm
- fine-tuning
- lora
- qlora
- oswestry
- functional-assessment
- kaggle
- mistral
---
# Model Card for oswestry-mistral-finetuned
This is a fine-tuned version of the `Mistral-7B-Instruct-v0.2` model specialized in scoring functional disability interviews using the Oswestry Disability Index (ODI) scale in Spanish. The model is designed to transform clinical-style interview transcripts into structured scores, demonstrating improved performance compared to the base model.
---
## Model Details
### Model Description
- **Developed by:** [Alejandro M.L]
- **Model type:** Causal Decoder-Only Transformer (LLM)
- **Language(s):** Spanish (with clinical vocabulary)
- **License:** apache-2.0
- **Fine-tuned from model:** mistral-7b-instruct-v0.2
### Model Sources
- **Repository:** [Github](https://github.com/ALM-Bloom/TFG_Alejandro)
- **Paper [Transformaci贸n de Informes M茅dicos en escalas funcionales]:**
---
## Uses
### Direct Use
The model takes as input a clinical interview transcript in Spanish (following a structured instruction format) and returns a text output containing the scores of each item in the Oswestry scale.
### Downstream Use
Can be integrated into tools that support:
- Preliminary functional assessment in telemedicine
- Research pipelines for NLP in healthcare
- Spanish-language LLM benchmarking on medical tasks
### Out-of-Scope Use
- Not suitable for general-purpose chat applications
- Should not be used for real clinical decisions without expert supervision
- Not intended for languages other than Spanish
---
## Bias, Risks, and Limitations
- The model was fine-tuned on synthetic data, which may limit generalizability.
- Outputs might include hallucinations if the input format is not followed.
- It reflects the biases of the base model and the prompt structure used.
### Recommendations
- Use only for research and non-critical applications.
- Always validate outputs against clinical judgment.
- Further training with real anonymized clinical data is highly recommended.
---
## How to Get Started with the Model
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
model = AutoModelForCausalLM.from_pretrained("DrAleML/Oswestry-Instruct")
tokenizer = AutoTokenizer.from_pretrained("DrAleML/Oswestry-Instruct")
prompt = "Entrevista:\nPaciente refiere dolor lumbar que aumenta al estar de pie...\n\nResponde con puntuaci贸n Oswestry:"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=300)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))