import copy import pytest import numpy as np import pandas as pd from sklearn.feature_extraction.text import CountVectorizer @pytest.mark.parametrize('model', [('kmeans_pca_topic_model'), ('base_topic_model'), ('custom_topic_model'), ('merged_topic_model'), ('reduced_topic_model')]) def test_update_topics(model, documents, request): topic_model = copy.deepcopy(request.getfixturevalue(model)) old_ctfidf = topic_model.c_tf_idf_ old_topics = topic_model.topics_ topic_model.update_topics(documents, n_gram_range=(1, 3)) assert old_ctfidf.shape[1] < topic_model.c_tf_idf_.shape[1] assert old_topics == topic_model.topics_ updated_topics = [topic if topic != 1 else 0 for topic in old_topics] topic_model.update_topics(documents, topics=updated_topics, n_gram_range=(1, 3)) assert len(set(old_topics)) - 1 == len(set(topic_model.topics_)) old_topics = topic_model.topics_ updated_topics = [topic if topic != 2 else 0 for topic in old_topics] topic_model.update_topics(documents, topics=updated_topics, n_gram_range=(1, 3)) assert len(set(old_topics)) - 1 == len(set(topic_model.topics_)) @pytest.mark.parametrize('model', [('kmeans_pca_topic_model'), ('base_topic_model'), ('custom_topic_model'), ('merged_topic_model'), ('reduced_topic_model'), ('online_topic_model')]) def test_extract_topics(model, documents, request): topic_model = copy.deepcopy(request.getfixturevalue(model)) nr_topics = 5 documents = pd.DataFrame({"Document": documents, "ID": range(len(documents)), "Topic": np.random.randint(-1, nr_topics-1, len(documents))}) topic_model._update_topic_size(documents) topic_model._extract_topics(documents) freq = topic_model.get_topic_freq() assert topic_model.c_tf_idf_.shape[0] == 5 assert topic_model.c_tf_idf_.shape[1] > 100 assert isinstance(freq, pd.DataFrame) assert nr_topics == len(freq.Topic.unique()) assert freq.Count.sum() == len(documents) assert len(freq.Topic.unique()) == len(freq) @pytest.mark.parametrize('model', [('kmeans_pca_topic_model'), ('base_topic_model'), ('custom_topic_model'), ('merged_topic_model'), ('reduced_topic_model'), ('online_topic_model')]) def test_extract_topics_custom_cv(model, documents, request): topic_model = copy.deepcopy(request.getfixturevalue(model)) nr_topics = 5 documents = pd.DataFrame({"Document": documents, "ID": range(len(documents)), "Topic": np.random.randint(-1, nr_topics-1, len(documents))}) cv = CountVectorizer(ngram_range=(1, 2)) topic_model.vectorizer_model = cv topic_model._update_topic_size(documents) topic_model._extract_topics(documents) freq = topic_model.get_topic_freq() assert topic_model.c_tf_idf_.shape[0] == 5 assert topic_model.c_tf_idf_.shape[1] > 100 assert isinstance(freq, pd.DataFrame) assert nr_topics == len(freq.Topic.unique()) assert freq.Count.sum() == len(documents) assert len(freq.Topic.unique()) == len(freq) @pytest.mark.parametrize('model', [('kmeans_pca_topic_model'), ('base_topic_model'), ('custom_topic_model'), ('merged_topic_model'), ('reduced_topic_model'), ('online_topic_model')]) @pytest.mark.parametrize("reduced_topics", [2, 4, 10]) def test_topic_reduction(model, reduced_topics, documents, request): topic_model = copy.deepcopy(request.getfixturevalue(model)) old_topics = copy.deepcopy(topic_model.topics_) old_freq = topic_model.get_topic_freq() topic_model.reduce_topics(documents, nr_topics=reduced_topics) new_freq = topic_model.get_topic_freq() if model != "online_topic_model": assert old_freq.Count.sum() == new_freq.Count.sum() assert len(old_freq.Topic.unique()) == len(old_freq) assert len(new_freq.Topic.unique()) == len(new_freq) assert len(topic_model.topics_) == len(old_topics) assert topic_model.topics_ != old_topics @pytest.mark.parametrize('model', [('kmeans_pca_topic_model'), ('base_topic_model'), ('custom_topic_model'), ('merged_topic_model'), ('reduced_topic_model'), ('online_topic_model')]) def test_topic_reduction_edge_cases(model, documents, request): topic_model = copy.deepcopy(request.getfixturevalue(model)) topic_model.nr_topics = 100 nr_topics = 5 topics = np.random.randint(-1, nr_topics - 1, len(documents)) old_documents = pd.DataFrame({"Document": documents, "ID": range(len(documents)), "Topic": topics}) topic_model._update_topic_size(old_documents) topic_model._extract_topics(old_documents) old_freq = topic_model.get_topic_freq() new_documents = topic_model._reduce_topics(old_documents) new_freq = topic_model.get_topic_freq() assert not set(old_documents.Topic).difference(set(new_documents.Topic)) pd.testing.assert_frame_equal(old_documents, new_documents) pd.testing.assert_frame_equal(old_freq, new_freq) @pytest.mark.parametrize('model', [('kmeans_pca_topic_model'), ('base_topic_model'), ('custom_topic_model'), ('merged_topic_model'), ('reduced_topic_model'), ('online_topic_model')]) def test_find_topics(model, request): topic_model = copy.deepcopy(request.getfixturevalue(model)) similar_topics, similarity = topic_model.find_topics("car") assert np.mean(similarity) > 0.1 assert len(similar_topics) > 0