| import numpy as np |
| from itertools import chain |
| from typing import List |
|
|
| from scipy import sparse |
| from scipy.sparse import csr_matrix |
|
|
| from sklearn.feature_extraction.text import CountVectorizer |
|
|
|
|
| class OnlineCountVectorizer(CountVectorizer): |
| """ An online variant of the CountVectorizer with updating vocabulary. |
| |
| At each `.partial_fit`, its vocabulary is updated based on any OOV words |
| it might find. Then, `.update_bow` can be used to track and update |
| the Bag-of-Words representation. These functions are separated such that |
| the vectorizer can be used in iteration without updating the Bag-of-Words |
| representation can might speed up the fitting process. However, the |
| `.update_bow` function is used in BERTopic to track changes in the |
| topic representations and allow for decay. |
| |
| This class inherits its parameters and attributes from: |
| `sklearn.feature_extraction.text.CountVectorizer` |
| |
| Arguments: |
| decay: A value between [0, 1] to weight the percentage of frequencies |
| the previous bag-of-words should be decreased. For example, |
| a value of `.1` will decrease the frequencies in the bag-of-words |
| matrix with 10% at each iteration. |
| delete_min_df: Delete words at each iteration from its vocabulary |
| that are below a minimum frequency. |
| This will keep the resulting bag-of-words matrix small |
| such that it does not explode in size with increasing |
| vocabulary. If `decay` is None then this equals `min_df`. |
| **kwargs: Set of parameters inherited from: |
| `sklearn.feature_extraction.text.CountVectorizer` |
| In practice, this means that you can still use parameters |
| from the original CountVectorizer, like `stop_words` and |
| `ngram_range`. |
| |
| Attributes: |
| X_ (scipy.sparse.csr_matrix) : The Bag-of-Words representation |
| |
| Examples: |
| |
| ```python |
| from bertopic.vectorizers import OnlineCountVectorizer |
| vectorizer = OnlineCountVectorizer(stop_words="english") |
| |
| for index, doc in enumerate(my_docs): |
| vectorizer.partial_fit(doc) |
| |
| # Update and clean the bow every 100 iterations: |
| if index % 100 == 0: |
| X = vectorizer.update_bow() |
| ``` |
| |
| To use the model in BERTopic: |
| |
| ```python |
| from bertopic import BERTopic |
| from bertopic.vectorizers import OnlineCountVectorizer |
| |
| vectorizer_model = OnlineCountVectorizer(stop_words="english") |
| topic_model = BERTopic(vectorizer_model=vectorizer_model) |
| ``` |
| |
| References: |
| Adapted from: https://github.com/idoshlomo/online_vectorizers |
| """ |
| def __init__(self, |
| decay: float = None, |
| delete_min_df: float = None, |
| **kwargs): |
| self.decay = decay |
| self.delete_min_df = delete_min_df |
| super(OnlineCountVectorizer, self).__init__(**kwargs) |
|
|
| def partial_fit(self, raw_documents: List[str]) -> None: |
| """ Perform a partial fit and update vocabulary with OOV tokens |
| |
| Arguments: |
| raw_documents: A list of documents |
| """ |
| if not hasattr(self, 'vocabulary_'): |
| return self.fit(raw_documents) |
|
|
| analyzer = self.build_analyzer() |
| analyzed_documents = [analyzer(doc) for doc in raw_documents] |
| new_tokens = set(chain.from_iterable(analyzed_documents)) |
| oov_tokens = new_tokens.difference(set(self.vocabulary_.keys())) |
|
|
| if oov_tokens: |
| max_index = max(self.vocabulary_.values()) |
| oov_vocabulary = dict(zip(oov_tokens, list(range(max_index + 1, max_index + 1 + len(oov_tokens), 1)))) |
| self.vocabulary_.update(oov_vocabulary) |
|
|
| return self |
|
|
| def update_bow(self, raw_documents: List[str]) -> csr_matrix: |
| """ Create or update the bag-of-words matrix |
| |
| Update the bag-of-words matrix by adding the newly transformed |
| documents. This may add empty columns if new words are found and/or |
| add empty rows if new topics are found. |
| |
| During this process, the previous bag-of-words matrix might be |
| decayed if `self.decay` has been set during init. Similarly, words |
| that do not exceed `self.delete_min_df` are removed from its |
| vocabulary and bag-of-words matrix. |
| |
| Arguments: |
| raw_documents: A list of documents |
| |
| Returns: |
| X_: Bag-of-words matrix |
| """ |
| if hasattr(self, "X_"): |
| X = self.transform(raw_documents) |
|
|
| |
| columns = csr_matrix((self.X_.shape[0], X.shape[1] - self.X_.shape[1]), dtype=int) |
| self.X_ = sparse.hstack([self.X_, columns]) |
|
|
| |
| rows = csr_matrix((X.shape[0] - self.X_.shape[0], self.X_.shape[1]), dtype=int) |
| self.X_ = sparse.vstack([self.X_, rows]) |
|
|
| |
| if self.decay is not None: |
| self.X_ = self.X_ * (1 - self.decay) |
|
|
| self.X_ += X |
| else: |
| self.X_ = self.transform(raw_documents) |
|
|
| if self.delete_min_df is not None: |
| self._clean_bow() |
|
|
| return self.X_ |
|
|
| def _clean_bow(self) -> None: |
| """ Remove words that do not exceed `self.delete_min_df` """ |
| |
| indices = np.where(self.X_.sum(0) >= self.delete_min_df)[1] |
| indices_dict = {index: index for index in indices} |
| self.X_ = self.X_[:, indices] |
|
|
| |
| new_vocab = {} |
| vocabulary_dict = {v: k for k, v in self.vocabulary_.items()} |
| for i, index in enumerate(indices): |
| if indices_dict.get(index) is not None: |
| new_vocab[vocabulary_dict[index]] = i |
|
|
| self.vocabulary_ = new_vocab |
|
|