EuriskoMobility commited on
Commit
7816fdd
·
1 Parent(s): aec6720

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +58 -0
README.md ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+
3
+ language: en
4
+
5
+ license: apache-2.0
6
+
7
+ ---
8
+
9
+
10
+
11
+
12
+ HF-version model for PRIMERA: Pyramid-based Masked Sentence Pre-training for Multi-document Summarization (ACL 2022).
13
+
14
+
15
+
16
+ The original code can be found [here](https://github.com/allenai/PRIMER). You can find the script and notebook to train/evaluate the model in the original github repo.
17
+
18
+
19
+
20
+ * Note: due to the difference between the implementations of the original Longformer and the Huggingface LED model, the results of converted models are slightly different. We run a sanity check on both fine-tuned and non fine-tuned models on the **MultiNews dataset**, and show the results below:
21
+
22
+
23
+
24
+ | Model | Rouge-1 | Rouge-2 | Rouge-L |
25
+
26
+ | --- | ----------- |----------- |----------- |
27
+
28
+ | PRIMERA | 42.0 | 13.6 | 20.8|
29
+
30
+ | PRIMERA-hf | 41.7 |13.6 | 20.5|
31
+
32
+ | PRIMERA(finetuned) | 49.9 | 21.1 | 25.9|
33
+
34
+ | PRIMERA-hf(finetuned) | 49.9 | 20.9 | 25.8|
35
+
36
+
37
+
38
+ You can use it by
39
+
40
+ ```
41
+
42
+ from transformers import (
43
+
44
+ AutoTokenizer,
45
+
46
+ LEDConfig,
47
+
48
+ LEDForConditionalGeneration,
49
+
50
+ )
51
+
52
+ tokenizer = AutoTokenizer.from_pretrained('allenai/PRIMERA')
53
+
54
+ config=LEDConfig.from_pretrained('allenai/PRIMERA')
55
+
56
+ model = LEDForConditionalGeneration.from_pretrained('allenai/PRIMERA')
57
+
58
+ ```