|
|
import numpy as np |
|
|
from tqdm import tqdm |
|
|
from typing import List |
|
|
from bertopic.backend import BaseEmbedder |
|
|
from gensim.models.keyedvectors import Word2VecKeyedVectors |
|
|
|
|
|
|
|
|
class GensimBackend(BaseEmbedder): |
|
|
""" Gensim Embedding Model |
|
|
|
|
|
The Gensim embedding model is typically used for word embeddings with |
|
|
GloVe, Word2Vec or FastText. |
|
|
|
|
|
Arguments: |
|
|
embedding_model: A Gensim embedding model |
|
|
|
|
|
Examples: |
|
|
|
|
|
```python |
|
|
from bertopic.backend import GensimBackend |
|
|
import gensim.downloader as api |
|
|
|
|
|
ft = api.load('fasttext-wiki-news-subwords-300') |
|
|
ft_embedder = GensimBackend(ft) |
|
|
``` |
|
|
""" |
|
|
def __init__(self, embedding_model: Word2VecKeyedVectors): |
|
|
super().__init__() |
|
|
|
|
|
if isinstance(embedding_model, Word2VecKeyedVectors): |
|
|
self.embedding_model = embedding_model |
|
|
else: |
|
|
raise ValueError("Please select a correct Gensim model: \n" |
|
|
"`import gensim.downloader as api` \n" |
|
|
"`ft = api.load('fasttext-wiki-news-subwords-300')`") |
|
|
|
|
|
def embed(self, |
|
|
documents: List[str], |
|
|
verbose: bool = False) -> np.ndarray: |
|
|
""" Embed a list of n documents/words into an n-dimensional |
|
|
matrix of embeddings |
|
|
|
|
|
Arguments: |
|
|
documents: A list of documents or words to be embedded |
|
|
verbose: Controls the verbosity of the process |
|
|
|
|
|
Returns: |
|
|
Document/words embeddings with shape (n, m) with `n` documents/words |
|
|
that each have an embeddings size of `m` |
|
|
""" |
|
|
vector_shape = self.embedding_model.get_vector(list(self.embedding_model.index_to_key)[0]).shape[0] |
|
|
empty_vector = np.zeros(vector_shape) |
|
|
|
|
|
|
|
|
embeddings = [] |
|
|
for doc in tqdm(documents, disable=not verbose, position=0, leave=True): |
|
|
embedding = [self.embedding_model.get_vector(word) for word in doc.split() |
|
|
if word in self.embedding_model.key_to_index] |
|
|
|
|
|
if len(embedding) > 0: |
|
|
embeddings.append(np.mean(embedding, axis=0)) |
|
|
else: |
|
|
embeddings.append(empty_vector) |
|
|
|
|
|
embeddings = np.array(embeddings) |
|
|
return embeddings |
|
|
|