File size: 2,666 Bytes
bfa613b
 
a95adca
 
 
 
 
 
 
 
 
 
 
bfa613b
 
a95adca
bfa613b
a95adca
bfa613b
a95adca
bfa613b
 
 
 
 
a95adca
 
 
 
 
bfa613b
a95adca
bfa613b
a95adca
 
bfa613b
a95adca
bfa613b
 
 
 
 
a95adca
bfa613b
a95adca
bfa613b
a95adca
 
 
 
bfa613b
 
 
a95adca
 
 
bfa613b
a95adca
bfa613b
 
 
a95adca
 
 
bfa613b
 
 
a95adca
 
 
bfa613b
a95adca
bfa613b
 
 
a95adca
 
bfa613b
a95adca
 
bfa613b
a95adca
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
---
library_name: transformers
tags:
  - medical
  - spanish
  - llm
  - fine-tuning
  - lora
  - qlora
  - oswestry
  - functional-assessment
  - kaggle
  - mistral
---

# Model Card for oswestry-mistral-finetuned

This is a fine-tuned version of the `Mistral-7B-Instruct-v0.2` model specialized in scoring functional disability interviews using the Oswestry Disability Index (ODI) scale in Spanish. The model is designed to transform clinical-style interview transcripts into structured scores, demonstrating improved performance compared to the base model.

---

## Model Details

### Model Description

- **Developed by:** [Alejandro M.L]
- **Model type:** Causal Decoder-Only Transformer (LLM)
- **Language(s):** Spanish (with clinical vocabulary)
- **License:** apache-2.0 
- **Fine-tuned from model:** mistral-7b-instruct-v0.2

### Model Sources

- **Repository:** [Github](https://github.com/ALM-Bloom/TFG_Alejandro)
- **Paper [Transformación de Informes Médicos en escalas funcionales]:** 

---

## Uses

### Direct Use

The model takes as input a clinical interview transcript in Spanish (following a structured instruction format) and returns a text output containing the scores of each item in the Oswestry scale.

### Downstream Use

Can be integrated into tools that support:
- Preliminary functional assessment in telemedicine
- Research pipelines for NLP in healthcare
- Spanish-language LLM benchmarking on medical tasks

### Out-of-Scope Use

- Not suitable for general-purpose chat applications
- Should not be used for real clinical decisions without expert supervision
- Not intended for languages other than Spanish

---

## Bias, Risks, and Limitations

- The model was fine-tuned on synthetic data, which may limit generalizability.
- Outputs might include hallucinations if the input format is not followed.
- It reflects the biases of the base model and the prompt structure used.

### Recommendations

- Use only for research and non-critical applications.
- Always validate outputs against clinical judgment.
- Further training with real anonymized clinical data is highly recommended.

---

## How to Get Started with the Model

```python
from transformers import AutoTokenizer, AutoModelForCausalLM

model = AutoModelForCausalLM.from_pretrained("DrAleML/Oswestry-Instruct")
tokenizer = AutoTokenizer.from_pretrained("DrAleML/Oswestry-Instruct")

prompt = "Entrevista:\nPaciente refiere dolor lumbar que aumenta al estar de pie...\n\nResponde con puntuación Oswestry:"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=300)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))