|
|
--- |
|
|
library_name: transformers |
|
|
tags: |
|
|
- medical |
|
|
- spanish |
|
|
- llm |
|
|
- fine-tuning |
|
|
- lora |
|
|
- qlora |
|
|
- oswestry |
|
|
- functional-assessment |
|
|
- kaggle |
|
|
- mistral |
|
|
--- |
|
|
|
|
|
# Model Card for oswestry-mistral-finetuned |
|
|
|
|
|
This is a fine-tuned version of the `Mistral-7B-Instruct-v0.2` model specialized in scoring functional disability interviews using the Oswestry Disability Index (ODI) scale in Spanish. The model is designed to transform clinical-style interview transcripts into structured scores, demonstrating improved performance compared to the base model. |
|
|
|
|
|
--- |
|
|
|
|
|
## Model Details |
|
|
|
|
|
### Model Description |
|
|
|
|
|
- **Developed by:** [Alejandro M.L] |
|
|
- **Model type:** Causal Decoder-Only Transformer (LLM) |
|
|
- **Language(s):** Spanish (with clinical vocabulary) |
|
|
- **License:** apache-2.0 |
|
|
- **Fine-tuned from model:** mistral-7b-instruct-v0.2 |
|
|
|
|
|
### Model Sources |
|
|
|
|
|
- **Repository:** [Github](https://github.com/ALM-Bloom/TFG_Alejandro) |
|
|
- **Paper [Transformaci贸n de Informes M茅dicos en escalas funcionales]:** |
|
|
|
|
|
--- |
|
|
|
|
|
## Uses |
|
|
|
|
|
### Direct Use |
|
|
|
|
|
The model takes as input a clinical interview transcript in Spanish (following a structured instruction format) and returns a text output containing the scores of each item in the Oswestry scale. |
|
|
|
|
|
### Downstream Use |
|
|
|
|
|
Can be integrated into tools that support: |
|
|
- Preliminary functional assessment in telemedicine |
|
|
- Research pipelines for NLP in healthcare |
|
|
- Spanish-language LLM benchmarking on medical tasks |
|
|
|
|
|
### Out-of-Scope Use |
|
|
|
|
|
- Not suitable for general-purpose chat applications |
|
|
- Should not be used for real clinical decisions without expert supervision |
|
|
- Not intended for languages other than Spanish |
|
|
|
|
|
--- |
|
|
|
|
|
## Bias, Risks, and Limitations |
|
|
|
|
|
- The model was fine-tuned on synthetic data, which may limit generalizability. |
|
|
- Outputs might include hallucinations if the input format is not followed. |
|
|
- It reflects the biases of the base model and the prompt structure used. |
|
|
|
|
|
### Recommendations |
|
|
|
|
|
- Use only for research and non-critical applications. |
|
|
- Always validate outputs against clinical judgment. |
|
|
- Further training with real anonymized clinical data is highly recommended. |
|
|
|
|
|
--- |
|
|
|
|
|
## How to Get Started with the Model |
|
|
|
|
|
```python |
|
|
from transformers import AutoTokenizer, AutoModelForCausalLM |
|
|
|
|
|
model = AutoModelForCausalLM.from_pretrained("DrAleML/Oswestry-Instruct") |
|
|
tokenizer = AutoTokenizer.from_pretrained("DrAleML/Oswestry-Instruct") |
|
|
|
|
|
prompt = "Entrevista:\nPaciente refiere dolor lumbar que aumenta al estar de pie...\n\nResponde con puntuaci贸n Oswestry:" |
|
|
inputs = tokenizer(prompt, return_tensors="pt") |
|
|
outputs = model.generate(**inputs, max_new_tokens=300) |
|
|
print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
|
|
|