Utkarsh124 commited on
Commit
061b42f
·
1 Parent(s): b174615

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +38 -1
README.md CHANGED
@@ -8,4 +8,41 @@ metrics:
8
  - bleu
9
  library_name: transformers
10
  pipeline_tag: summarization
11
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  - bleu
9
  library_name: transformers
10
  pipeline_tag: summarization
11
+ ---
12
+
13
+ # t5-small-finetuned
14
+
15
+ ## Model Description
16
+ - **Purpose and Use**: This model is designed for abstractive text summarization with a focus on the SAMSum Dialogue Dataset.
17
+ - **Model Architecture**: The architecture is based on a fine-tuned T5-small model, which consists of 60 million parameters.
18
+ - **Training Data**: Trained on the SAMSum Dialogue Dataset, which comprises approximately 15,000 dialogue-summary pairs.
19
+
20
+ ## Training Procedure
21
+ - **Preprocessing**: Data preprocessing involved the removal of irrelevant tags and tokenization to ensure data consistency.
22
+ - **Training Details**: The model was fine-tuned over 4 epochs with a learning rate of 2e-5 and a batch size of 2, utilizing gradient accumulation for optimization.
23
+ - **Infrastructure**: Training was conducted using GPU acceleration and the Hugging Face Trainer API, with progress monitored via TensorBoard.
24
+
25
+ ## Evaluation Results
26
+ - **Metrics Used**: Evaluation metrics included ROUGE-1, ROUGE-2, ROUGE-L, BLEU, and Cosine Similarity.
27
+ - **Performance**: The fine-tuned T5-small model demonstrated superior efficiency and effectiveness in summarization tasks, outperforming its larger counterparts.
28
+
29
+ ## Validation and Test Set Performance
30
+
31
+ | Metric | Validation Set | Test Set |
32
+ |----------|--------------------|--------------|
33
+ | ROUGE-1 | 0.5667 | 0.5536 |
34
+ | ROUGE-2 | 0.2923 | 0.2718 |
35
+ | ROUGE-L | 0.5306 | 0.5210 |
36
+
37
+ The table above presents the performance of the model on both the validation and test sets, indicating the quality of content overlap and structural fluency in the summaries generated.
38
+
39
+ ## Performance Metrics Comparison Across Models
40
+
41
+ | Model | ROUGE-1 | ROUGE-2 | ROUGE-L | BLEU | Cosine Similarity |
42
+ |----------|---------|---------|---------|------|-------------------|
43
+ | My Model | 0.3767 | 0.1596 | 0.2896 | 9.52 | 0.7698 |
44
+ | T5 Large | 0.3045 | 0.0960 | 0.2315 | 4.82 | 0.6745 |
45
+ | Bart | 0.3189 | 0.0989 | 0.2352 | 6.28 | 0.6961 |
46
+ | Pegasus | 0.2702 | 0.0703 | 0.2093 | 3.88 | 0.6432 |
47
+
48
+ In the table above shows results on 50 samples for the test set that is being compared across various models.