|
|
|
|
|
import numpy as np |
|
|
import pandas as pd |
|
|
|
|
|
import spacy |
|
|
from spacy.matcher import Matcher |
|
|
from spacy.language import Language |
|
|
|
|
|
from packaging import version |
|
|
from scipy.sparse import csr_matrix |
|
|
from typing import List, Mapping, Tuple, Union |
|
|
from sklearn import __version__ as sklearn_version |
|
|
from bertopic.representation._base import BaseRepresentation |
|
|
|
|
|
|
|
|
class PartOfSpeech(BaseRepresentation): |
|
|
""" Extract Topic Keywords based on their Part-of-Speech |
|
|
|
|
|
DEFAULT_PATTERNS = [ |
|
|
[{'POS': 'ADJ'}, {'POS': 'NOUN'}], |
|
|
[{'POS': 'NOUN'}], |
|
|
[{'POS': 'ADJ'}] |
|
|
] |
|
|
|
|
|
From candidate topics, as extracted with c-TF-IDF, |
|
|
find documents that contain keywords found in the |
|
|
candidate topics. These candidate documents then |
|
|
serve as the representative set of documents from |
|
|
which the Spacy model can extract a set of candidate |
|
|
keywords for each topic. |
|
|
|
|
|
These candidate keywords are first judged by whether |
|
|
they fall within the DEFAULT_PATTERNS or the user-defined |
|
|
pattern. Then, the resulting keywords are sorted by |
|
|
their respective c-TF-IDF values. |
|
|
|
|
|
Arguments: |
|
|
model: The Spacy model to use |
|
|
top_n_words: The top n words to extract |
|
|
pos_patterns: Patterns for Spacy to use. |
|
|
See https://spacy.io/usage/rule-based-matching |
|
|
|
|
|
Usage: |
|
|
|
|
|
```python |
|
|
from bertopic.representation import PartOfSpeech |
|
|
from bertopic import BERTopic |
|
|
|
|
|
# Create your representation model |
|
|
representation_model = PartOfSpeech("en_core_web_sm") |
|
|
|
|
|
# Use the representation model in BERTopic on top of the default pipeline |
|
|
topic_model = BERTopic(representation_model=representation_model) |
|
|
``` |
|
|
|
|
|
You can define custom POS patterns to be extracted: |
|
|
|
|
|
```python |
|
|
pos_patterns = [ |
|
|
[{'POS': 'ADJ'}, {'POS': 'NOUN'}], |
|
|
[{'POS': 'NOUN'}], [{'POS': 'ADJ'}] |
|
|
] |
|
|
representation_model = PartOfSpeech("en_core_web_sm", pos_patterns=pos_patterns) |
|
|
``` |
|
|
""" |
|
|
def __init__(self, |
|
|
model: Union[str, Language] = "en_core_web_sm", |
|
|
top_n_words: int = 10, |
|
|
pos_patterns: List[str] = None): |
|
|
if isinstance(model, str): |
|
|
self.model = spacy.load(model) |
|
|
elif isinstance(model, Language): |
|
|
self.model = model |
|
|
else: |
|
|
raise ValueError("Make sure that the Spacy model that you" |
|
|
"pass is either a string referring to a" |
|
|
"Spacy model or a Spacy nlp object.") |
|
|
|
|
|
self.top_n_words = top_n_words |
|
|
|
|
|
if pos_patterns is None: |
|
|
self.pos_patterns = [ |
|
|
[{'POS': 'ADJ'}, {'POS': 'NOUN'}], |
|
|
[{'POS': 'NOUN'}], [{'POS': 'ADJ'}] |
|
|
] |
|
|
else: |
|
|
self.pos_patterns = pos_patterns |
|
|
|
|
|
def extract_topics(self, |
|
|
topic_model, |
|
|
documents: pd.DataFrame, |
|
|
c_tf_idf: csr_matrix, |
|
|
topics: Mapping[str, List[Tuple[str, float]]] |
|
|
) -> Mapping[str, List[Tuple[str, float]]]: |
|
|
""" Extract topics |
|
|
|
|
|
Arguments: |
|
|
topic_model: A BERTopic model |
|
|
documents: All input documents |
|
|
c_tf_idf: Not used |
|
|
topics: The candidate topics as calculated with c-TF-IDF |
|
|
|
|
|
Returns: |
|
|
updated_topics: Updated topic representations |
|
|
""" |
|
|
matcher = Matcher(self.model.vocab) |
|
|
matcher.add("Pattern", self.pos_patterns) |
|
|
|
|
|
candidate_topics = {} |
|
|
for topic, values in topics.items(): |
|
|
keywords = list(zip(*values))[0] |
|
|
|
|
|
|
|
|
candidate_documents = [] |
|
|
for keyword in keywords: |
|
|
selection = documents.loc[documents.Topic == topic, :] |
|
|
selection = selection.loc[selection.Document.str.contains(keyword), "Document"] |
|
|
if len(selection) > 0: |
|
|
for document in selection[:2]: |
|
|
candidate_documents.append(document) |
|
|
candidate_documents = list(set(candidate_documents)) |
|
|
|
|
|
|
|
|
docs_pipeline = self.model.pipe(candidate_documents) |
|
|
updated_keywords = [] |
|
|
for doc in docs_pipeline: |
|
|
matches = matcher(doc) |
|
|
for _, start, end in matches: |
|
|
updated_keywords.append(doc[start:end].text) |
|
|
candidate_topics[topic] = list(set(updated_keywords)) |
|
|
|
|
|
|
|
|
|
|
|
if version.parse(sklearn_version) >= version.parse("1.0.0"): |
|
|
words = list(topic_model.vectorizer_model.get_feature_names_out()) |
|
|
else: |
|
|
words = list(topic_model.vectorizer_model.get_feature_names()) |
|
|
|
|
|
|
|
|
words_lookup = dict(zip(words, range(len(words)))) |
|
|
updated_topics = {topic: [] for topic in topics.keys()} |
|
|
|
|
|
for topic, candidate_keywords in candidate_topics.items(): |
|
|
word_indices = [words_lookup.get(keyword) for keyword in candidate_keywords if words_lookup.get(keyword)] |
|
|
vals = topic_model.c_tf_idf_[:, np.array(word_indices)][topic + topic_model._outliers] |
|
|
indices = np.argsort(np.array(vals.todense().reshape(1, -1))[0])[-self.top_n_words:][::-1] |
|
|
vals = np.sort(np.array(vals.todense().reshape(1, -1))[0])[-self.top_n_words:][::-1] |
|
|
topic_words = [(words[word_indices[index]], val) for index, val in zip(indices, vals)] |
|
|
updated_topics[topic] = topic_words |
|
|
if len(updated_topics[topic]) < self.top_n_words: |
|
|
updated_topics[topic] += [("", 0) for _ in range(self.top_n_words-len(updated_topics[topic]))] |
|
|
|
|
|
return updated_topics |
|
|
|