Commit ·
061b42f
1
Parent(s): b174615
Update README.md
Browse files
README.md
CHANGED
|
@@ -8,4 +8,41 @@ metrics:
|
|
| 8 |
- bleu
|
| 9 |
library_name: transformers
|
| 10 |
pipeline_tag: summarization
|
| 11 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 8 |
- bleu
|
| 9 |
library_name: transformers
|
| 10 |
pipeline_tag: summarization
|
| 11 |
+
---
|
| 12 |
+
|
| 13 |
+
# t5-small-finetuned
|
| 14 |
+
|
| 15 |
+
## Model Description
|
| 16 |
+
- **Purpose and Use**: This model is designed for abstractive text summarization with a focus on the SAMSum Dialogue Dataset.
|
| 17 |
+
- **Model Architecture**: The architecture is based on a fine-tuned T5-small model, which consists of 60 million parameters.
|
| 18 |
+
- **Training Data**: Trained on the SAMSum Dialogue Dataset, which comprises approximately 15,000 dialogue-summary pairs.
|
| 19 |
+
|
| 20 |
+
## Training Procedure
|
| 21 |
+
- **Preprocessing**: Data preprocessing involved the removal of irrelevant tags and tokenization to ensure data consistency.
|
| 22 |
+
- **Training Details**: The model was fine-tuned over 4 epochs with a learning rate of 2e-5 and a batch size of 2, utilizing gradient accumulation for optimization.
|
| 23 |
+
- **Infrastructure**: Training was conducted using GPU acceleration and the Hugging Face Trainer API, with progress monitored via TensorBoard.
|
| 24 |
+
|
| 25 |
+
## Evaluation Results
|
| 26 |
+
- **Metrics Used**: Evaluation metrics included ROUGE-1, ROUGE-2, ROUGE-L, BLEU, and Cosine Similarity.
|
| 27 |
+
- **Performance**: The fine-tuned T5-small model demonstrated superior efficiency and effectiveness in summarization tasks, outperforming its larger counterparts.
|
| 28 |
+
|
| 29 |
+
## Validation and Test Set Performance
|
| 30 |
+
|
| 31 |
+
| Metric | Validation Set | Test Set |
|
| 32 |
+
|----------|--------------------|--------------|
|
| 33 |
+
| ROUGE-1 | 0.5667 | 0.5536 |
|
| 34 |
+
| ROUGE-2 | 0.2923 | 0.2718 |
|
| 35 |
+
| ROUGE-L | 0.5306 | 0.5210 |
|
| 36 |
+
|
| 37 |
+
The table above presents the performance of the model on both the validation and test sets, indicating the quality of content overlap and structural fluency in the summaries generated.
|
| 38 |
+
|
| 39 |
+
## Performance Metrics Comparison Across Models
|
| 40 |
+
|
| 41 |
+
| Model | ROUGE-1 | ROUGE-2 | ROUGE-L | BLEU | Cosine Similarity |
|
| 42 |
+
|----------|---------|---------|---------|------|-------------------|
|
| 43 |
+
| My Model | 0.3767 | 0.1596 | 0.2896 | 9.52 | 0.7698 |
|
| 44 |
+
| T5 Large | 0.3045 | 0.0960 | 0.2315 | 4.82 | 0.6745 |
|
| 45 |
+
| Bart | 0.3189 | 0.0989 | 0.2352 | 6.28 | 0.6961 |
|
| 46 |
+
| Pegasus | 0.2702 | 0.0703 | 0.2093 | 3.88 | 0.6432 |
|
| 47 |
+
|
| 48 |
+
In the table above shows results on 50 samples for the test set that is being compared across various models.
|