Update README.md
Browse files
README.md
CHANGED
|
@@ -19,21 +19,8 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 19 |
# BART-CNN-Convosumm
|
| 20 |
|
| 21 |
This model is a fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the arg-filtered reddit part of [Convosumm](https://github.com/Yale-LILY/ConvoSumm) dataset.
|
| 22 |
-
|
| 23 |
-
- Loss: 3.8797
|
| 24 |
-
- Rouge1: 38.6252
|
| 25 |
-
- Rouge2: 12.2556
|
| 26 |
-
- Rougel: 23.902
|
| 27 |
-
- Rougelsum: 34.6324
|
| 28 |
-
- Gen Len: 81.28
|
| 29 |
|
| 30 |
-
It achieves the following results on the test set (250 data points):
|
| 31 |
-
- Loss: 3.8343
|
| 32 |
-
- Rouge1: 38.3642
|
| 33 |
-
- Rouge2: 12.2056
|
| 34 |
-
- Rougel: 23.7782
|
| 35 |
-
- Rougelsum: 34.3959
|
| 36 |
-
- Gen Len: 84.132
|
| 37 |
|
| 38 |
## Model description
|
| 39 |
|
|
@@ -76,6 +63,21 @@ The following hyperparameters were used during training:
|
|
| 76 |
| 2.9933 | 5.97 | 60 | 3.8809 | 38.66 | 12.3337 | 23.4394 | 35.1976 | 83.26 |
|
| 77 |
| 2.834 | 6.97 | 70 | 3.8797 | 38.6252 | 12.2556 | 23.902 | 34.6324 | 81.28 |
|
| 78 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 79 |
|
| 80 |
### Framework versions
|
| 81 |
|
|
|
|
| 19 |
# BART-CNN-Convosumm
|
| 20 |
|
| 21 |
This model is a fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the arg-filtered reddit part of [Convosumm](https://github.com/Yale-LILY/ConvoSumm) dataset.
|
| 22 |
+
Model is trained for [multilanguage telegram-bot summarizer](https://github.com/akaRemeris/XLConvosumm-bot).
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 23 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 24 |
|
| 25 |
## Model description
|
| 26 |
|
|
|
|
| 63 |
| 2.9933 | 5.97 | 60 | 3.8809 | 38.66 | 12.3337 | 23.4394 | 35.1976 | 83.26 |
|
| 64 |
| 2.834 | 6.97 | 70 | 3.8797 | 38.6252 | 12.2556 | 23.902 | 34.6324 | 81.28 |
|
| 65 |
|
| 66 |
+
It achieves the following results on the evaluation set (50 data points):
|
| 67 |
+
- Loss: 3.8797
|
| 68 |
+
- Rouge1: 38.6252
|
| 69 |
+
- Rouge2: 12.2556
|
| 70 |
+
- Rougel: 23.902
|
| 71 |
+
- Rougelsum: 34.6324
|
| 72 |
+
- Gen Len: 81.28
|
| 73 |
+
|
| 74 |
+
It achieves the following results on the test set (250 data points):
|
| 75 |
+
- Loss: 3.8343
|
| 76 |
+
- Rouge1: 38.3642
|
| 77 |
+
- Rouge2: 12.2056
|
| 78 |
+
- Rougel: 23.7782
|
| 79 |
+
- Rougelsum: 34.3959
|
| 80 |
+
- Gen Len: 84.132
|
| 81 |
|
| 82 |
### Framework versions
|
| 83 |
|