| --- |
| license: apache-2.0 |
| tags: |
| - generated_from_trainer |
| metrics: |
| - rouge |
| model-index: |
| - name: DaMedSum-large |
| results: [] |
| pipeline_tag: summarization |
| language: |
| - da |
| --- |
| |
| ``` |
| _____ ______ __ __ ______ _____ ______ __ __ __ __ |
| /\ __-. /\ __ \ /\ "-./ \ /\ ___\ /\ __-. /\ ___\ /\ \/\ \ /\ "-./ \ |
| \ \ \/\ \\ \ __ \\ \ \-./\ \\ \ __\ \ \ \/\ \\ \___ \\ \ \_\ \\ \ \-./\ \ |
| \ \____- \ \_\ \_\\ \_\ \ \_\\ \_____\\ \____- \/\_____\\ \_____\\ \_\ \ \_\ |
| \/____/ \/_/\/_/ \/_/ \/_/ \/_____/ \/____/ \/_____/ \/_____/ \/_/ \/_/ |
| |
| ``` |
|
|
| ## Model description |
|
|
| This repository contains a model for Danish abstractive summarisation of medical text. |
|
|
| This model is a fine-tuned version of DanSumT5-large trained on a danish medical text dataset. |
|
|
| The model was trained on LUMI using 1 AMD MI250X GPU. |
|
|
| ## Authors |
| Nicolaj Larsen |
| Mikkel Kildeberg |
| Emil Schledermann |
|
|
| ### Framework versions |
|
|
| - Transformers 4.30.2 |
| - Pytorch 1.12.1+git7548e2f |
| - Datasets 2.13.2 |
| - Tokenizers 0.13.3 |