wesley7137/qa_dataset
Viewer • Updated • 4.45k • 4
How to use sanjithrj/T5-medi with Transformers:
# Load model directly
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
tokenizer = AutoTokenizer.from_pretrained("sanjithrj/T5-medi")
model = AutoModelForSeq2SeqLM.from_pretrained("sanjithrj/T5-medi")This model is a fine-tuned version of google/flan-t5-small on Medical QA dataset (wesley7137/qa_dataset). It is just for educational purpose and does not provide accurate results. It achieves the following results on the evaluation set:
This is a fine-tuned T5 model for Question-Answering tasks in Medical Field
May struggle with creative or subjective content. Requires fine-tuning for different tasks
More information needed
The following hyperparameters were used during training:
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum |
|---|---|---|---|---|---|---|---|
| No log | 1.0 | 223 | 0.8782 | 0.3148 | 0.2630 | 0.3010 | 0.3098 |
| No log | 2.0 | 446 | 0.7820 | 0.3148 | 0.2650 | 0.3026 | 0.3111 |
| 1.1386 | 3.0 | 669 | 0.7456 | 0.3170 | 0.2681 | 0.3049 | 0.3131 |
| 1.1386 | 4.0 | 892 | 0.7259 | 0.3198 | 0.2699 | 0.3072 | 0.3156 |
| 0.7884 | 5.0 | 1115 | 0.7130 | 0.3189 | 0.2700 | 0.3068 | 0.3149 |
Base model
google/flan-t5-small