Update README.md
#4
by
cherry0328 - opened
README.md
CHANGED
|
@@ -6,42 +6,66 @@ tags:
|
|
| 6 |
datasets:
|
| 7 |
- midas/inspec
|
| 8 |
widget:
|
| 9 |
-
- text:
|
| 10 |
-
|
| 11 |
-
|
| 12 |
-
|
| 13 |
-
|
| 14 |
-
|
| 15 |
-
|
| 16 |
-
|
| 17 |
-
|
| 18 |
-
|
| 19 |
-
|
| 20 |
-
|
| 21 |
-
|
| 22 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 23 |
model-index:
|
| 24 |
- name: DeDeckerThomas/keyphrase-generation-t5-small-inspec
|
| 25 |
results:
|
| 26 |
-
- task:
|
| 27 |
type: keyphrase-generation
|
| 28 |
name: Keyphrase Generation
|
| 29 |
dataset:
|
| 30 |
type: midas/inspec
|
| 31 |
name: inspec
|
| 32 |
metrics:
|
| 33 |
-
|
| 34 |
-
|
| 35 |
-
|
| 36 |
-
|
| 37 |
-
|
| 38 |
-
|
| 39 |
-
|
| 40 |
-
|
| 41 |
-
|
| 42 |
-
|
| 43 |
-
|
| 44 |
-
|
|
|
|
|
|
|
| 45 |
---
|
| 46 |
|
| 47 |
# 🔑 Keyphrase Generation Model: T5-small-inspec
|
|
|
|
| 6 |
datasets:
|
| 7 |
- midas/inspec
|
| 8 |
widget:
|
| 9 |
+
- text: >-
|
| 10 |
+
Keyphrase extraction is a technique in text analysis where you extract the
|
| 11 |
+
important keyphrases from a document. Thanks to these keyphrases humans can
|
| 12 |
+
understand the content of a text very quickly and easily without reading it
|
| 13 |
+
completely. Keyphrase extraction was first done primarily by human
|
| 14 |
+
annotators, who read the text in detail and then wrote down the most
|
| 15 |
+
important keyphrases. The disadvantage is that if you work with a lot of
|
| 16 |
+
documents, this process can take a lot of time.
|
| 17 |
+
|
| 18 |
+
Here is where Artificial Intelligence comes in. Currently, classical machine
|
| 19 |
+
learning methods, that use statistical and linguistic features, are widely
|
| 20 |
+
used for the extraction process. Now with deep learning, it is possible to
|
| 21 |
+
capture the semantic meaning of a text even better than these classical
|
| 22 |
+
methods. Classical methods look at the frequency, occurrence and order of
|
| 23 |
+
words in the text, whereas these neural approaches can capture long-term
|
| 24 |
+
semantic dependencies and context of words in a text.
|
| 25 |
+
example_title: Example 1
|
| 26 |
+
- text: >-
|
| 27 |
+
In this work, we explore how to learn task specific language models aimed
|
| 28 |
+
towards learning rich representation of keyphrases from text documents. We
|
| 29 |
+
experiment with different masking strategies for pre-training transformer
|
| 30 |
+
language models (LMs) in discriminative as well as generative settings. In
|
| 31 |
+
the discriminative setting, we introduce a new pre-training objective -
|
| 32 |
+
Keyphrase Boundary Infilling with Replacement (KBIR), showing large gains in
|
| 33 |
+
performance (up to 9.26 points in F1) over SOTA, when LM pre-trained using
|
| 34 |
+
KBIR is fine-tuned for the task of keyphrase extraction. In the generative
|
| 35 |
+
setting, we introduce a new pre-training setup for BART - KeyBART, that
|
| 36 |
+
reproduces the keyphrases related to the input text in the CatSeq format,
|
| 37 |
+
instead of the denoised original input. This also led to gains in
|
| 38 |
+
performance (up to 4.33 points inF1@M) over SOTA for keyphrase generation.
|
| 39 |
+
Additionally, we also fine-tune the pre-trained language models on named
|
| 40 |
+
entity recognition(NER), question answering (QA), relation extraction (RE),
|
| 41 |
+
abstractive summarization and achieve comparable performance with that of
|
| 42 |
+
the SOTA, showing that learning rich representation of keyphrases is indeed
|
| 43 |
+
beneficial for many other fundamental NLP tasks.
|
| 44 |
+
example_title: Example 2
|
| 45 |
model-index:
|
| 46 |
- name: DeDeckerThomas/keyphrase-generation-t5-small-inspec
|
| 47 |
results:
|
| 48 |
+
- task:
|
| 49 |
type: keyphrase-generation
|
| 50 |
name: Keyphrase Generation
|
| 51 |
dataset:
|
| 52 |
type: midas/inspec
|
| 53 |
name: inspec
|
| 54 |
metrics:
|
| 55 |
+
- type: F1@M (Present)
|
| 56 |
+
value: 0.317
|
| 57 |
+
name: F1@M (Present)
|
| 58 |
+
- type: F1@O (Present)
|
| 59 |
+
value: 0.279
|
| 60 |
+
name: F1@O (Present)
|
| 61 |
+
- type: F1@M (Absent)
|
| 62 |
+
value: 0.073
|
| 63 |
+
name: F1@M (Absent)
|
| 64 |
+
- type: F1@O (Absent)
|
| 65 |
+
value: 0.065
|
| 66 |
+
name: F1@O (Absent)
|
| 67 |
+
base_model:
|
| 68 |
+
- google-t5/t5-small
|
| 69 |
---
|
| 70 |
|
| 71 |
# 🔑 Keyphrase Generation Model: T5-small-inspec
|