| { |
| "paper_id": "2021", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T10:41:17.941799Z" |
| }, |
| "title": "OCTIS: Comparing and Optimizing Topic Models is Simple!", |
| "authors": [ |
| { |
| "first": "Silvia", |
| "middle": [], |
| "last": "Terragni", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "s.terragni4@campus.unimib.it" |
| }, |
| { |
| "first": "Elisabetta", |
| "middle": [], |
| "last": "Fersini", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "elisabetta.fersini@unimib.it" |
| }, |
| { |
| "first": "Bruno", |
| "middle": [], |
| "last": "Galuzzi", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "bruno.galuzzi@unimib.it" |
| }, |
| { |
| "first": "Pietro", |
| "middle": [], |
| "last": "Tropeano", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "p.tropeano1@campus.unimib.it" |
| }, |
| { |
| "first": "Antonio", |
| "middle": [], |
| "last": "Candelieri", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "antonio.candelieri@unimib.it" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "In this paper, we present OCTIS, a framework for training, analyzing, and comparing Topic Models, whose optimal hyper-parameters are estimated using a Bayesian Optimization approach. The proposed solution integrates several state-of-the-art topic models and evaluation metrics. These metrics can be targeted as objective by the underlying optimization procedure to determine the best hyper-parameter configuration. OCTIS allows researchers and practitioners to have a fair comparison between topic models of interest, using several benchmark datasets and well-known evaluation metrics, to integrate novel algorithms, and to have an interactive visualization of the results for understanding the behavior of each model. The code is available at the following link: https://github.com/ MIND-Lab/OCTIS.", |
| "pdf_parse": { |
| "paper_id": "2021", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "In this paper, we present OCTIS, a framework for training, analyzing, and comparing Topic Models, whose optimal hyper-parameters are estimated using a Bayesian Optimization approach. The proposed solution integrates several state-of-the-art topic models and evaluation metrics. These metrics can be targeted as objective by the underlying optimization procedure to determine the best hyper-parameter configuration. OCTIS allows researchers and practitioners to have a fair comparison between topic models of interest, using several benchmark datasets and well-known evaluation metrics, to integrate novel algorithms, and to have an interactive visualization of the results for understanding the behavior of each model. The code is available at the following link: https://github.com/ MIND-Lab/OCTIS.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Topic models are promising statistical methods that aim to extract the hidden topics underlying a collection of documents. Although researchers have proposed several models across the years (Blei, 2012; Vayansky and Kumar, 2020) , their evaluation and comparison is still a hard task. The evaluation of a topic model usually involves different datasets (with non-standard pre-processing) (Schofield and Mimno, 2016; Schofield et al., 2017) and several evaluation metrics (Lau et al., 2014; Wallach et al., 2009; Terragni et al., 2020a) . Furthermore, topic models are usually compared by fixing their hyperparameters. However, choosing the optimal hyperparameter configuration for a given dataset and a given evaluation metric is fundamental to induce each model at the best of its capabilities, and therefore to guarantee a fair comparison with other models.", |
| "cite_spans": [ |
| { |
| "start": 190, |
| "end": 202, |
| "text": "(Blei, 2012;", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 203, |
| "end": 228, |
| "text": "Vayansky and Kumar, 2020)", |
| "ref_id": "BIBREF33" |
| }, |
| { |
| "start": 388, |
| "end": 415, |
| "text": "(Schofield and Mimno, 2016;", |
| "ref_id": "BIBREF26" |
| }, |
| { |
| "start": 416, |
| "end": 439, |
| "text": "Schofield et al., 2017)", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 471, |
| "end": 489, |
| "text": "(Lau et al., 2014;", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 490, |
| "end": 511, |
| "text": "Wallach et al., 2009;", |
| "ref_id": "BIBREF34" |
| }, |
| { |
| "start": 512, |
| "end": 535, |
| "text": "Terragni et al., 2020a)", |
| "ref_id": "BIBREF31" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Current topic modeling frameworks (McCallum et al., 2005; Qiang et al., 2018; Lisena et al., 2020) typically focus on the release of topic modeling algorithms while ignoring one or more critical aspects of the topic modeling pipeline, such as preprocessing, evaluation, comparison of the models, and visualization. Most importantly, they disregard the hyper-parameter selection.", |
| "cite_spans": [ |
| { |
| "start": 34, |
| "end": 57, |
| "text": "(McCallum et al., 2005;", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 58, |
| "end": 77, |
| "text": "Qiang et al., 2018;", |
| "ref_id": "BIBREF23" |
| }, |
| { |
| "start": 78, |
| "end": 98, |
| "text": "Lisena et al., 2020)", |
| "ref_id": "BIBREF18" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this paper, we present OCTIS (Optimizing and Comparing Topic models Is Simple) 1 , a unified and open-source evaluation framework for training, analyzing, and comparing Topic Models, over several datasets and evaluation metrics. Their optimal hyper-parameter configuration is determined according to a Bayesian Optimization (BO) strategy Snoek et al., 2012; Galuzzi et al., 2020) .", |
| "cite_spans": [ |
| { |
| "start": 341, |
| "end": 360, |
| "text": "Snoek et al., 2012;", |
| "ref_id": "BIBREF28" |
| }, |
| { |
| "start": 361, |
| "end": 382, |
| "text": "Galuzzi et al., 2020)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In the following, we summarize the main contributions of the proposed framework:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 several open-source topic models have been integrated into a unified framework, providing a common interface that allows the users to easily experiment with topic models;", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 a single-objective BO approach has been integrated to determine the optimal hyperparameter values of each model, for a given dataset and a specific evaluation metric of interest;", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 an interactive visualization of the results for inspecting the details of the models, providing insights about the optimization strategy, word and topic distributions, and robustness of the estimated configuration;", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 a python library for advanced exploitation of the framework for integrating novel algorithms, with their training and inference algorithms.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "2 System design and architecture", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "OCTIS is an open-source evaluation framework for the comparison of a set of state-of-the-art topic models, that allows the user to optimize the models' hyper-parameters for a fair experimental comparison. The proposed framework follows an objectoriented paradigm, providing all the tools for running a topic modeling pipeline. The main functionalities of the proposed OC-TIS are related to dataset pre-processing, training topic models, estimating evaluation metrics, hyperparameter optimization, and interactive web dashboard visualization. Figure 1 summarizes the workflow involving the first four modules (the dashboard interacts with all of them). The framework can be used both as a python library and as a dashboard. The python library offers more advanced functionalities than the ones available in the dashboard. The modules that comprise the OCTIS framework are detailed in the following sections.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 542, |
| "end": 550, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The first step of the topic modeling pipeline is the pre-processing of the input dataset. OCTIS includes the following pre-processing utilities:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Datasets and Pre-processing", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "\u2022 reducing the text to lowercase;", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Datasets and Pre-processing", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "\u2022 punctuation removal;", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Datasets and Pre-processing", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "\u2022 lemmatization;", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Datasets and Pre-processing", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "\u2022 stop-words removal;", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Datasets and Pre-processing", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "\u2022 removal of unfrequent and most frequent words (according to a specified frequency threshold);", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Datasets and Pre-processing", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "\u2022 removal of documents with few words (according to a specified frequency threshold).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Datasets and Pre-processing", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "These utilities include the most common techniques for pre-processing text for topic modeling. However, some of these features may not be appropriate for specific domains and languages, e.g. requiring language-specific or domain-specific stop-words. OCTIS currently provides 4 pre-processed datasets, i.e. 20 NewsGroups 2 , M10 (Lim and Buntine, 2014), DBLP 3 and BBC News (Greene and Cunningham, 2006) , different in nature and length.", |
| "cite_spans": [ |
| { |
| "start": 373, |
| "end": 402, |
| "text": "(Greene and Cunningham, 2006)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Datasets and Pre-processing", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "The datasets already available in OCTIS, and accessible through the web dashboard, have been preprocessed according to the length of the documents. In particular, we removed the punctuation, we lemmatized the text, filtered out the stop-words (using the English stop-words list provided by MALLET), and removed the words that have a word frequency less than 0.5% for 20 Newsgroups and BBC News and less than 0.05% for DBLP and M10. Subsequently, we removed the documents with less than 5 words for 20 Newsgroups and BBC News and less than 3 words for the other datasets. Although OCTIS already provides some datasets, a user can upload and pre-process any dataset (using the python library) according to its needs.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Datasets and Pre-processing", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "OCTIS integrates both classical topic models and neural topic models. In particular, the following traditional and neural approaches are available to be trained, optimized, analyzed, and compared (the models that are available in the web dashboard are marked with ):", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic Modeling", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Latent Dirichlet Allocation (Blei et al., 2003, LDA) ; 4", |
| "cite_spans": [ |
| { |
| "start": 30, |
| "end": 54, |
| "text": "(Blei et al., 2003, LDA)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic Modeling", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Non-negative Matrix Factorization (Lee and Seung, 2000, NMF); 4", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic Modeling", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Latent Semantic Analysis (Hofmann, 1999, LSI) ; 4", |
| "cite_spans": [ |
| { |
| "start": 27, |
| "end": 47, |
| "text": "(Hofmann, 1999, LSI)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic Modeling", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Hierarchical Dirichlet Process (Teh et al., 2004 , HDP); 4", |
| "cite_spans": [ |
| { |
| "start": 33, |
| "end": 50, |
| "text": "(Teh et al., 2004", |
| "ref_id": "BIBREF30" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic Modeling", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Neural LDA (Srivastava and Sutton, 2017); 5", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic Modeling", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Product-of-Experts LDA (Srivastava and Sutton, 2017, ProdLDA); 5", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic Modeling", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Embedded Topic Models (Dieng et al., 2019, ETM); 6", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic Modeling", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Contextualized Topic Models (Bianchi et al., 2021, CTM) . 7", |
| "cite_spans": [ |
| { |
| "start": 30, |
| "end": 57, |
| "text": "(Bianchi et al., 2021, CTM)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic Modeling", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Moreover, we defined a standard interface for allowing a user to integrate their topic model's implementation. A topic model is indeed a black-box, a system solely viewed in terms of its inputs and outputs and whose internal workings are invisible. This black-box topic model takes as input a dataset and a set of hyperparameters values and returns the top-t topic words, the document-topic distributions, and the topic-word distribution in a specified format.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic Modeling", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "4 https://radimrehurek.com/gensim/ 5 https://github.com/estebandito22/ PyTorchAVITM 6 https://github.com/adjidieng/ETM 7 https://github.com/MilaNLProc/ contextualized-topic-models", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Topic Modeling", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "The proposed framework provides several evaluation metrics. A metric can be used as the objective targeted by a Bayesian Optimization strategy, or to monitor the behavior of a topic model while the model is optimized on a different objective. The performance of a topic model can be evaluated by investigating different aspects, according to the following evaluation metrics:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation Metrics", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "\u2022 Topic coherence metrics (Lau et al., 2014; R\u00f6der et al., 2015) that compute how the top-k words of a topic are related to each other;", |
| "cite_spans": [ |
| { |
| "start": 26, |
| "end": 44, |
| "text": "(Lau et al., 2014;", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 45, |
| "end": 64, |
| "text": "R\u00f6der et al., 2015)", |
| "ref_id": "BIBREF24" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation Metrics", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "\u2022 Topic significance metrics (AlSumait et al., 2009; Terragni et al., 2020b ) that focus on the document-topic and topic-word distributions to discover high-quality and junk topics;", |
| "cite_spans": [ |
| { |
| "start": 29, |
| "end": 52, |
| "text": "(AlSumait et al., 2009;", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 53, |
| "end": 75, |
| "text": "Terragni et al., 2020b", |
| "ref_id": "BIBREF32" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation Metrics", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "\u2022 Diversity metrics (Dieng et al., 2019; Bianchi et al., 2020) that measure how diverse the topk words of a topic are to each other;", |
| "cite_spans": [ |
| { |
| "start": 20, |
| "end": 40, |
| "text": "(Dieng et al., 2019;", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 41, |
| "end": 62, |
| "text": "Bianchi et al., 2020)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation Metrics", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "\u2022 Classification metrics (Phan et al., 2008; Terragni et al., 2020a) where the document-topic distribution of each document is used as the K-dimensional representation to train a classifier that predicts the document's class.", |
| "cite_spans": [ |
| { |
| "start": 25, |
| "end": 44, |
| "text": "(Phan et al., 2008;", |
| "ref_id": "BIBREF22" |
| }, |
| { |
| "start": 45, |
| "end": 68, |
| "text": "Terragni et al., 2020a)", |
| "ref_id": "BIBREF31" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation Metrics", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "OCTIS provides 10 evaluation metrics directly available in the web dashboard, and 13 accessible through the python library.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation Metrics", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "The proposed framework uses Bayesian Optimization (Snoek et al., 2012; Shahriari et al., 2015) to tune the hyper-parameters of the topic models. If any of the available hyper-parameters is selected to be optimized for a given evaluation metric, BO explores the search space to determine the optimal settings. Since the performance estimated by the evaluation metrics can be affected by noise, the objective function is computed as the median of a given number of model runs (i.e., topic models run with the same hyperparameter configuration) computed for the selected evaluation metric. BO is a sequential model-based optimization strategy for expensive and noisy black-box functions (e.g. topic models). The basic idea consists of using all the model's configurations evaluated so far to approximate the value of the performance metric and then selects a new promising configuration to evaluate. The approximation is provided by a probabilistic surrogate model, which describes the prior belief over the objective function using the observed configurations. The next configuration to evaluate is selected through the optimization of an acquisition function, which leverages the uncertainty in the posterior to guide the exploration.", |
| "cite_spans": [ |
| { |
| "start": 50, |
| "end": 70, |
| "text": "(Snoek et al., 2012;", |
| "ref_id": "BIBREF28" |
| }, |
| { |
| "start": 71, |
| "end": 94, |
| "text": "Shahriari et al., 2015)", |
| "ref_id": "BIBREF27" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Hyper-parameter Optimization", |
| "sec_num": "2.4" |
| }, |
| { |
| "text": "We integrated into OCTIS most of the BO algorithms of the Scikit-Optimize library (Head et al., 2018) to provide a robust and efficient BO implementation. We integrated Gaussian Process and Random Forest as surrogate models, while we included Probability of Improvement, Expected Improvement, and Upper Confidence Bound as acquisition functions. See (Snoek et al., 2012; for more details about the use of BO for hyper-parameter optimization.", |
| "cite_spans": [ |
| { |
| "start": 82, |
| "end": 101, |
| "text": "(Head et al., 2018)", |
| "ref_id": null |
| }, |
| { |
| "start": 350, |
| "end": 370, |
| "text": "(Snoek et al., 2012;", |
| "ref_id": "BIBREF28" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Hyper-parameter Optimization", |
| "sec_num": "2.4" |
| }, |
| { |
| "text": "Instead of performing BO, a user can also use a random search technique to find the best hyperparameter configuration. Since the Bayesian Optimization requires some initial configurations to fit the surrogate model, the user can provide the initial configurations, according to their domain knowledge. Alternatively, a user can perform a pure exploration of the search space using a random sampling strategy. Different algorithms are available (e.g. Uniform Random Sampling or Latin Hypercube sequence) for sampling the initial configurations.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Hyper-parameter Optimization", |
| "sec_num": "2.4" |
| }, |
| { |
| "text": "The existing topic modeling frameworks usually provide topic modeling algorithms, while disregarding other essential aspects of the whole topic modeling pipeline: pre-processing, evaluation, comparison, and visualization of the results and, most importantly, the hyper-parameter selection. In the fol-lowing, we outline the existing frameworks, highlighting their advantages and limitations.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Existing frameworks", |
| "sec_num": "3" |
| }, |
| { |
| "text": "MALLET (McCallum, 2002) and gensim 4 are the most known topic modeling libraries and include several classical topic models. They provide pre-processing methods and the estimation of the hyper-parameters using maximum likelihood estimation (MLE) techniques. These libraries do not include the recently proposed neural topic models, and they just provide topic coherence metrics.", |
| "cite_spans": [ |
| { |
| "start": 7, |
| "end": 23, |
| "text": "(McCallum, 2002)", |
| "ref_id": "BIBREF20" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Existing frameworks", |
| "sec_num": "3" |
| }, |
| { |
| "text": "STTM (Qiang et al., 2018 ) is a java library that provides a set of topic models that are specifically designed for short texts, providing several evaluation metrics.", |
| "cite_spans": [ |
| { |
| "start": 5, |
| "end": 24, |
| "text": "(Qiang et al., 2018", |
| "ref_id": "BIBREF23" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Existing frameworks", |
| "sec_num": "3" |
| }, |
| { |
| "text": "ToModAPI (Lisena et al., 2020 ) is a python API that allows for training, inference, and evaluating different topic models, also including some of the most recent. However, it does not provide a method for finding the best hyper-parameter configuration of topic models. Instead, a tool that allows for optimizing the hyper-parameter of a machine learning model is PyCARET (Ali, 2020) . However, it employs a grid-search technique to tune the hyper-parameters. This approach can be very time-consuming if the number of hyperparameters is high and the search space is huge (Bergstra and Bengio, 2012) .", |
| "cite_spans": [ |
| { |
| "start": 9, |
| "end": 29, |
| "text": "(Lisena et al., 2020", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 372, |
| "end": 383, |
| "text": "(Ali, 2020)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 571, |
| "end": 598, |
| "text": "(Bergstra and Bengio, 2012)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Existing frameworks", |
| "sec_num": "3" |
| }, |
| { |
| "text": "OCTIS stands at the union of the features of the existing frameworks: we integrated both classical and recent neural topic models, providing preprocessing methods, evaluation metrics, and the possibility of optimizing the hyper-parameters. Finally, a user-friendly graphical interface to launch one or more hyper-parameter optimization experiments on a given topic model and on a specific dataset has been provided. 4 System usage OCTIS has been designed to be used as a python library by advanced users, as well as a simple web dashboard by anyone.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Existing frameworks", |
| "sec_num": "3" |
| }, |
| { |
| "text": "4.1 Example of use case for the python library The above lines of code will execute an optimization experiment that will provide an optimal configuration of the hyperparameters \u03b1 and \u03b2 for LDA with 25 topics by maximizing the diversity of the topics.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Existing frameworks", |
| "sec_num": "3" |
| }, |
| { |
| "text": "The dashboard includes a set of simple but useful operations to conduct an experimental campaign on different topic models. Here we briefly explain the four main functionalities of the dashboard.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Web-based dashboard", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Experiment creation. First, a user can define an optimization experiment by selecting the dataset, the topic model, the corresponding hyperparameter to optimize, the evaluation metric to be considered by the BO (possibly other extra metrics to evaluate), and the settings of the optimization process.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Web-based dashboard", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Management of the experiments' queue. The user can monitor the queue of the experiments and see the corresponding progress. The user can also pause, restart, or delete an experiment that has been launched before. Additionally, the user can easily change the order of the queue of the experiments, by allowing a given run to be executed before others. Comparison of the Topic Models. The user can select the models to be analyzed and compared. At the first stage, one can observe the progress of the BO iterations, observing a plot that contains at each iteration the best-seen evaluation, i.e. the median at each iteration of the metric that has been optimized (see Figure 2 ). Alternatively, a user can visualize a box plot at each iteration (see Figure 3) to understand if a given hyper-parameter configuration is noisy (high variance) or not.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 666, |
| "end": 674, |
| "text": "Figure 2", |
| "ref_id": "FIGREF2" |
| }, |
| { |
| "start": 748, |
| "end": 757, |
| "text": "Figure 3)", |
| "ref_id": "FIGREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Web-based dashboard", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Analysis of a single experiment. A user can further inspect the results of a specific topic model on a given dataset with respect to the considered metrics, by analyzing a single experiment.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Web-based dashboard", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Here, a user can visualize all the information and statistics related to the experiment, including the best hyper-parameter configuration and the best value of the optimized metric. They can also have an outline of the statistics of the other extra metrics that they had chosen to evaluate. We provide three different plots for inspecting the output of a single run of a topic model. Figure 4 shows the word cloud obtained from the most relevant words of a given topic, scaled by their probability. Focusing on the distributions inferred by a topic model, Figure 5 shows the topic distribution of a document, and Figure 6 represents an example of the weight of a selected word of the vocabulary for each topic. ", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 384, |
| "end": 392, |
| "text": "Figure 4", |
| "ref_id": "FIGREF4" |
| }, |
| { |
| "start": 556, |
| "end": 564, |
| "text": "Figure 5", |
| "ref_id": "FIGREF5" |
| }, |
| { |
| "start": 613, |
| "end": 621, |
| "text": "Figure 6", |
| "ref_id": "FIGREF6" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Web-based dashboard", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "In this paper, we presented the framework OCTIS for training, analyzing, and comparing Topic Models. The proposed framework is composed of a python library and a web dashboard and integrates several state-of-the-art topic models (both traditional and neural). These models can be trained by searching for their optimal hyperparameter configuration, for a given metric and dataset, exploiting a BO strategy. OCTIS allows researchers to train existing models, integrate new training and inference algorithms, and fairly compare the topic models of interest. On the other hand, practitioners could use OCTIS to boost the performance of Topic Models for their preferred downstream task or a wide range of practical applications, such as data exploratory analysis (Boyd-Graber et al., 2017) .", |
| "cite_spans": [ |
| { |
| "start": 759, |
| "end": 785, |
| "text": "(Boyd-Graber et al., 2017)", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Regarding future work, OCTIS could integrate a multi-objective optimization strategy to optimize multiple metrics in the same BO procedure (Paria et al., 2020) . For example, this could allow a user to find an optimal hyper-parameter configuration for both topic coherence and document classification.", |
| "cite_spans": [ |
| { |
| "start": 139, |
| "end": 159, |
| "text": "(Paria et al., 2020)", |
| "ref_id": "BIBREF21" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5" |
| }, |
| { |
| "text": "The video demonstration is available at https:// youtu.be/nPmiWBFFJ8E.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "http://people.csail.mit.edu/jrennie/ 20Newsgroups/ 3 https://github.com/shiruipan/TriDNR/ tree/master/data", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "PyCaret: An open source, low-code machine learning library in Python", |
| "authors": [ |
| { |
| "first": "Moez", |
| "middle": [], |
| "last": "Ali", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Moez Ali. 2020. PyCaret: An open source, low-code machine learning library in Python. PyCaret ver- sion 2.3.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Topic significance ranking of LDA generative models", |
| "authors": [ |
| { |
| "first": "Loulwah", |
| "middle": [], |
| "last": "Alsumait", |
| "suffix": "" |
| }, |
| { |
| "first": "Daniel", |
| "middle": [], |
| "last": "Barbar\u00e1", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Gentle", |
| "suffix": "" |
| }, |
| { |
| "first": "Carlotta", |
| "middle": [], |
| "last": "Domeniconi", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Machine Learning and Knowledge Discovery in Databases", |
| "volume": "5781", |
| "issue": "", |
| "pages": "67--82", |
| "other_ids": { |
| "DOI": [ |
| "10.1007/978-3-642-04180-8_22" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Loulwah AlSumait, Daniel Barbar\u00e1, James Gentle, and Carlotta Domeniconi. 2009. Topic significance rank- ing of LDA generative models. In Machine Learn- ing and Knowledge Discovery in Databases, Euro- pean Conference, ECML PKDD 2009, volume 5781 of Lecture Notes in Computer Science, pages 67-82. Springer.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Bayesian Optimization and Data Science", |
| "authors": [ |
| { |
| "first": "Francesco", |
| "middle": [], |
| "last": "Archetti", |
| "suffix": "" |
| }, |
| { |
| "first": "Antonio", |
| "middle": [], |
| "last": "Candelieri", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Francesco Archetti and Antonio Candelieri. 2019. Bayesian Optimization and Data Science. Springer International Publishing.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Random search for hyper-parameter optimization", |
| "authors": [ |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Bergstra", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Journal of Machine Learning Research", |
| "volume": "13", |
| "issue": "", |
| "pages": "281--305", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "James Bergstra and Yoshua Bengio. 2012. Random search for hyper-parameter optimization. Journal of Machine Learning Research, 13:281-305.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Pre-training is a hot topic: Contextualized document embeddings improve topic coherence", |
| "authors": [ |
| { |
| "first": "Federico", |
| "middle": [], |
| "last": "Bianchi", |
| "suffix": "" |
| }, |
| { |
| "first": "Silvia", |
| "middle": [], |
| "last": "Terragni", |
| "suffix": "" |
| }, |
| { |
| "first": "Dirk", |
| "middle": [], |
| "last": "Hovy", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:2004.03974" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Federico Bianchi, Silvia Terragni, and Dirk Hovy. 2020. Pre-training is a hot topic: Contextual- ized document embeddings improve topic coher- ence. arXiv preprint arXiv:2004.03974.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Cross-lingual Contextualized Topic Models with Zero-shot Learning", |
| "authors": [ |
| { |
| "first": "Federico", |
| "middle": [], |
| "last": "Bianchi", |
| "suffix": "" |
| }, |
| { |
| "first": "Silvia", |
| "middle": [], |
| "last": "Terragni", |
| "suffix": "" |
| }, |
| { |
| "first": "Dirk", |
| "middle": [], |
| "last": "Hovy", |
| "suffix": "" |
| }, |
| { |
| "first": "Debora", |
| "middle": [], |
| "last": "Nozza", |
| "suffix": "" |
| }, |
| { |
| "first": "Elisabetta", |
| "middle": [], |
| "last": "Fersini", |
| "suffix": "" |
| } |
| ], |
| "year": 2021, |
| "venue": "Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Federico Bianchi, Silvia Terragni, Dirk Hovy, Debora Nozza, and Elisabetta Fersini. 2021. Cross-lingual Contextualized Topic Models with Zero-shot Learn- ing. In Proceedings of the 16th Conference of the European Chapter of the Association for Computa- tional Linguistics.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Probabilistic topic models. Commun", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [ |
| "M" |
| ], |
| "last": "Blei", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "", |
| "volume": "55", |
| "issue": "", |
| "pages": "77--84", |
| "other_ids": { |
| "DOI": [ |
| "10.1145/2133806.2133826" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "David M. Blei. 2012. Probabilistic topic models. Com- mun. ACM, 55(4):77-84.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Latent dirichlet allocation", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [ |
| "M" |
| ], |
| "last": "Blei", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [ |
| "Y" |
| ], |
| "last": "Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [ |
| "I" |
| ], |
| "last": "Jordan", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Journal of Machine Learning Research", |
| "volume": "3", |
| "issue": "", |
| "pages": "993--1022", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David M. Blei, Andrew Y. Ng, and Michael I. Jordan. 2003. Latent dirichlet allocation. Journal of Ma- chine Learning Research, 3:993-1022.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Applications of topic models", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Jordan", |
| "suffix": "" |
| }, |
| { |
| "first": "Yuening", |
| "middle": [], |
| "last": "Boyd-Graber", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [ |
| "M" |
| ], |
| "last": "Hu", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Mimno", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Found. Trends Inf. Retr", |
| "volume": "11", |
| "issue": "2-3", |
| "pages": "143--296", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jordan L. Boyd-Graber, Yuening Hu, and David M. Mimno. 2017. Applications of topic models. Found. Trends Inf. Retr., 11(2-3):143-296.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Global optimization in machine learning: the design of a predictive analytics application", |
| "authors": [ |
| { |
| "first": "Antonio", |
| "middle": [], |
| "last": "Candelieri", |
| "suffix": "" |
| }, |
| { |
| "first": "Francesco", |
| "middle": [], |
| "last": "Archetti", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Soft Comput", |
| "volume": "23", |
| "issue": "9", |
| "pages": "2969--2977", |
| "other_ids": { |
| "DOI": [ |
| "10.1007/s00500-018-3597-8" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Antonio Candelieri and Francesco Archetti. 2019. Global optimization in machine learning: the design of a predictive analytics application. Soft Comput., 23(9):2969-2977.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Topic modeling in embedding spaces", |
| "authors": [ |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Adji", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Dieng", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [ |
| "R" |
| ], |
| "last": "Francisco", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [ |
| "M" |
| ], |
| "last": "Ruiz", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Blei", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "CoRR", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Adji B. Dieng, Francisco J. R. Ruiz, and David M. Blei. 2019. Topic modeling in embedding spaces. CoRR, abs/1907.04907.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Hyperparameter optimization for recommender systems through bayesian optimization", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Bg Galuzzi", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Giordani", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Candelieri", |
| "suffix": "" |
| }, |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Perego", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Archetti", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Computational Management Science", |
| "volume": "", |
| "issue": "", |
| "pages": "1--21", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "BG Galuzzi, I Giordani, A Candelieri, R Perego, and F Archetti. 2020. Hyperparameter optimization for recommender systems through bayesian optimiza- tion. Computational Management Science, pages 1- 21.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Practical solutions to the problem of diagonal dominance in kernel document clustering", |
| "authors": [ |
| { |
| "first": "Derek", |
| "middle": [], |
| "last": "Greene", |
| "suffix": "" |
| }, |
| { |
| "first": "P\u00e1draig", |
| "middle": [], |
| "last": "Cunningham", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of the 23rd International Conference on Machine learning (ICML'06)", |
| "volume": "", |
| "issue": "", |
| "pages": "377--384", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Derek Greene and P\u00e1draig Cunningham. 2006. Practi- cal solutions to the problem of diagonal dominance in kernel document clustering. In Proceedings of the 23rd International Conference on Machine learning (ICML'06), pages 377-384. ACM Press.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Probabilistic latent semantic indexing", |
| "authors": [ |
| { |
| "first": "Thomas", |
| "middle": [], |
| "last": "Hofmann", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "SIGIR '99: Proceedings of the 22nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval", |
| "volume": "", |
| "issue": "", |
| "pages": "50--57", |
| "other_ids": { |
| "DOI": [ |
| "10.1145/312624.312649" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Thomas Hofmann. 1999. Probabilistic latent semantic indexing. In SIGIR '99: Proceedings of the 22nd Annual International ACM SIGIR Conference on Re- search and Development in Information Retrieval, pages 50-57. ACM.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Machine reading tea leaves: Automatically evaluating topic coherence and topic model quality", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Jey Han Lau", |
| "suffix": "" |
| }, |
| { |
| "first": "Timothy", |
| "middle": [], |
| "last": "Newman", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Baldwin", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 14th Conference of the European Chapter of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "530--539", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jey Han Lau, David Newman, and Timothy Baldwin. 2014. Machine reading tea leaves: Automatically evaluating topic coherence and topic model quality. In Proceedings of the 14th Conference of the Euro- pean Chapter of the Association for Computational Linguistics, EACL 2014, pages 530-539.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Algorithms for non-negative matrix factorization", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Daniel", |
| "suffix": "" |
| }, |
| { |
| "first": "H", |
| "middle": [ |
| "Sebastian" |
| ], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Seung", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "Advances in Neural Information Processing Systems 13, Papers from Neural Information Processing Systems (NIPS) 2000", |
| "volume": "", |
| "issue": "", |
| "pages": "556--562", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Daniel D. Lee and H. Sebastian Seung. 2000. Algo- rithms for non-negative matrix factorization. In Ad- vances in Neural Information Processing Systems 13, Papers from Neural Information Processing Sys- tems (NIPS) 2000, pages 556-562. MIT Press.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Bibliographic analysis with the citation network topic model", |
| "authors": [ |
| { |
| "first": "Wai", |
| "middle": [], |
| "last": "Kar", |
| "suffix": "" |
| }, |
| { |
| "first": "Wray", |
| "middle": [ |
| "L" |
| ], |
| "last": "Lim", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Buntine", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the Sixth Asian Conference on Machine Learning", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kar Wai Lim and Wray L. Buntine. 2014. Bibli- ographic analysis with the citation network topic model. In Proceedings of the Sixth Asian Confer- ence on Machine Learning, ACML 2014.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "ToModAPI: A Topic Modeling API to Train, Use and Compare Topic Models", |
| "authors": [ |
| { |
| "first": "Pasquale", |
| "middle": [], |
| "last": "Lisena", |
| "suffix": "" |
| }, |
| { |
| "first": "Ismail", |
| "middle": [], |
| "last": "Harrando", |
| "suffix": "" |
| }, |
| { |
| "first": "Oussama", |
| "middle": [], |
| "last": "Kandakji", |
| "suffix": "" |
| }, |
| { |
| "first": "Raphael", |
| "middle": [], |
| "last": "Troncy", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "2 nd International Workshop for Natural Language Processing Open Source Software (NLP-OSS)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Pasquale Lisena, Ismail Harrando, Oussama Kandakji, and Raphael Troncy. 2020. ToModAPI: A Topic Modeling API to Train, Use and Compare Topic Models. In 2 nd International Workshop for Natural Language Processing Open Source Software (NLP- OSS).", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Topic and role discovery in social networks", |
| "authors": [ |
| { |
| "first": "Andrew", |
| "middle": [], |
| "last": "Mccallum", |
| "suffix": "" |
| }, |
| { |
| "first": "Andr\u00e9s", |
| "middle": [], |
| "last": "Corrada-Emmanuel", |
| "suffix": "" |
| }, |
| { |
| "first": "Xuerui", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "IJCAI-05, Proceedings of the Nineteenth International Joint Conference on Artificial Intelligence", |
| "volume": "", |
| "issue": "", |
| "pages": "786--791", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Andrew McCallum, Andr\u00e9s Corrada-Emmanuel, and Xuerui Wang. 2005. Topic and role discovery in social networks. In IJCAI-05, Proceedings of the Nineteenth International Joint Conference on Artifi- cial Intelligence, pages 786-791.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Mallet: A machine learning for language toolkit", |
| "authors": [ |
| { |
| "first": "Andrew Kachites", |
| "middle": [], |
| "last": "Mccallum", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Andrew Kachites McCallum. 2002. Mallet: A ma- chine learning for language toolkit. http://mallet. cs. umass. edu.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "A flexible framework for multiobjective bayesian optimization using random scalarizations", |
| "authors": [ |
| { |
| "first": "Biswajit", |
| "middle": [], |
| "last": "Paria", |
| "suffix": "" |
| }, |
| { |
| "first": "Kirthevasan", |
| "middle": [], |
| "last": "Kandasamy", |
| "suffix": "" |
| }, |
| { |
| "first": "Barnab\u00e1s", |
| "middle": [], |
| "last": "P\u00f3czos", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Uncertainty in Artificial Intelligence", |
| "volume": "", |
| "issue": "", |
| "pages": "766--776", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Biswajit Paria, Kirthevasan Kandasamy, and Barnab\u00e1s P\u00f3czos. 2020. A flexible framework for multi- objective bayesian optimization using random scalarizations. In Uncertainty in Artificial Intelli- gence, pages 766-776. PMLR.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Learning to classify short and sparse text & web with hidden topics from largescale data collections", |
| "authors": [ |
| { |
| "first": "Xuan", |
| "middle": [], |
| "last": "Hieu Phan", |
| "suffix": "" |
| }, |
| { |
| "first": "Minh", |
| "middle": [ |
| "Le" |
| ], |
| "last": "Nguyen", |
| "suffix": "" |
| }, |
| { |
| "first": "Susumu", |
| "middle": [], |
| "last": "Horiguchi", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proceedings of the 17th International Conference on World Wide Web", |
| "volume": "", |
| "issue": "", |
| "pages": "91--100", |
| "other_ids": { |
| "DOI": [ |
| "10.1145/1367497.1367510" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Xuan Hieu Phan, Minh Le Nguyen, and Susumu Horiguchi. 2008. Learning to classify short and sparse text & web with hidden topics from large- scale data collections. In Proceedings of the 17th International Conference on World Wide Web, WWW 2008, pages 91-100. ACM.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Sttm: A tool for short text topic modeling", |
| "authors": [ |
| { |
| "first": "Jipeng", |
| "middle": [], |
| "last": "Qiang", |
| "suffix": "" |
| }, |
| { |
| "first": "Yun", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Yunhao", |
| "middle": [], |
| "last": "Yuan", |
| "suffix": "" |
| }, |
| { |
| "first": "Wei", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Xindong", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1808.02215" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jipeng Qiang, Yun Li, Yunhao Yuan, Wei Liu, and Xin- dong Wu. 2018. Sttm: A tool for short text topic modeling. arXiv preprint arXiv:1808.02215.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Exploring the space of topic coherence measures", |
| "authors": [ |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "R\u00f6der", |
| "suffix": "" |
| }, |
| { |
| "first": "Andreas", |
| "middle": [], |
| "last": "Both", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexander", |
| "middle": [], |
| "last": "Hinneburg", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the Eighth ACM International Conference on Web Search and Data Mining, WSDM 2015", |
| "volume": "", |
| "issue": "", |
| "pages": "399--408", |
| "other_ids": { |
| "DOI": [ |
| "10.1145/2684822.2685324" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Michael R\u00f6der, Andreas Both, and Alexander Hinneb- urg. 2015. Exploring the space of topic coherence measures. In Proceedings of the Eighth ACM Inter- national Conference on Web Search and Data Min- ing, WSDM 2015, Shanghai, China, February 2-6, 2015, pages 399-408. ACM.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "Pulling out the stops: Rethinking stopword removal for topic models", |
| "authors": [ |
| { |
| "first": "Alexandra", |
| "middle": [], |
| "last": "Schofield", |
| "suffix": "" |
| }, |
| { |
| "first": "M\u00e5ns", |
| "middle": [], |
| "last": "Magnusson", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Mimno", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics", |
| "volume": "2", |
| "issue": "", |
| "pages": "432--436", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alexandra Schofield, M\u00e5ns Magnusson, and David Mimno. 2017. Pulling out the stops: Rethinking stopword removal for topic models. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Vol- ume 2, Short Papers, pages 432-436.", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "Comparing apples to apple: The effects of stemmers on topic models", |
| "authors": [ |
| { |
| "first": "Alexandra", |
| "middle": [], |
| "last": "Schofield", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Mimno", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Transactions of the Association for Computational Linguistics", |
| "volume": "4", |
| "issue": "", |
| "pages": "287--300", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alexandra Schofield and David Mimno. 2016. Com- paring apples to apple: The effects of stemmers on topic models. Transactions of the Association for Computational Linguistics, 4:287-300.", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "Taking the human out of the loop: A review of bayesian optimization", |
| "authors": [ |
| { |
| "first": "Bobak", |
| "middle": [], |
| "last": "Shahriari", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Swersky", |
| "suffix": "" |
| }, |
| { |
| "first": "Ziyu", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Ryan", |
| "suffix": "" |
| }, |
| { |
| "first": "Nando De", |
| "middle": [], |
| "last": "Adams", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Freitas", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the IEEE", |
| "volume": "104", |
| "issue": "1", |
| "pages": "148--175", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bobak Shahriari, Kevin Swersky, Ziyu Wang, Ryan P Adams, and Nando De Freitas. 2015. Taking the hu- man out of the loop: A review of bayesian optimiza- tion. Proceedings of the IEEE, 104(1):148-175.", |
| "links": null |
| }, |
| "BIBREF28": { |
| "ref_id": "b28", |
| "title": "Practical bayesian optimization of machine learning algorithms", |
| "authors": [ |
| { |
| "first": "Jasper", |
| "middle": [], |
| "last": "Snoek", |
| "suffix": "" |
| }, |
| { |
| "first": "Hugo", |
| "middle": [], |
| "last": "Larochelle", |
| "suffix": "" |
| }, |
| { |
| "first": "Ryan", |
| "middle": [ |
| "P" |
| ], |
| "last": "Adams", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Advances in Neural Information Processing Systems 25: 26th Annual Conference on Neural Information Processing Systems", |
| "volume": "", |
| "issue": "", |
| "pages": "2960--2968", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jasper Snoek, Hugo Larochelle, and Ryan P. Adams. 2012. Practical bayesian optimization of machine learning algorithms. In Advances in Neural Infor- mation Processing Systems 25: 26th Annual Con- ference on Neural Information Processing Systems 2012, pages 2960-2968.", |
| "links": null |
| }, |
| "BIBREF29": { |
| "ref_id": "b29", |
| "title": "Autoencoding variational inference for topic models", |
| "authors": [ |
| { |
| "first": "Akash", |
| "middle": [], |
| "last": "Srivastava", |
| "suffix": "" |
| }, |
| { |
| "first": "Charles", |
| "middle": [], |
| "last": "Sutton", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "5th International Conference on Learning Representations", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Akash Srivastava and Charles Sutton. 2017. Autoen- coding variational inference for topic models. In 5th International Conference on Learning Represen- tations, ICLR 2017.", |
| "links": null |
| }, |
| "BIBREF30": { |
| "ref_id": "b30", |
| "title": "Sharing clusters among related groups: Hierarchical dirichlet processes", |
| "authors": [ |
| { |
| "first": "Yee Whye", |
| "middle": [], |
| "last": "Teh", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [ |
| "I" |
| ], |
| "last": "Jordan", |
| "suffix": "" |
| }, |
| { |
| "first": "Matthew", |
| "middle": [ |
| "J" |
| ], |
| "last": "Beal", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [ |
| "M" |
| ], |
| "last": "Blei", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Advances in Neural Information Processing Systems", |
| "volume": "17", |
| "issue": "", |
| "pages": "1385--1392", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yee Whye Teh, Michael I. Jordan, Matthew J. Beal, and David M. Blei. 2004. Sharing clusters among related groups: Hierarchical dirichlet processes. In Advances in Neural Information Processing Systems, 17, pages 1385-1392.", |
| "links": null |
| }, |
| "BIBREF31": { |
| "ref_id": "b31", |
| "title": "Constrained relational topic models. Information Sciences", |
| "authors": [ |
| { |
| "first": "Silvia", |
| "middle": [], |
| "last": "Terragni", |
| "suffix": "" |
| }, |
| { |
| "first": "Elisabetta", |
| "middle": [], |
| "last": "Fersini", |
| "suffix": "" |
| }, |
| { |
| "first": "Enza", |
| "middle": [], |
| "last": "Messina", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "", |
| "volume": "512", |
| "issue": "", |
| "pages": "581--594", |
| "other_ids": { |
| "DOI": [ |
| "10.1016/j.ins.2019.09.039" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Silvia Terragni, Elisabetta Fersini, and Enza Messina. 2020a. Constrained relational topic models. Infor- mation Sciences, 512:581 -594.", |
| "links": null |
| }, |
| "BIBREF32": { |
| "ref_id": "b32", |
| "title": "Which matters most? comparing the impact of concept and document relationships in topic models", |
| "authors": [ |
| { |
| "first": "Silvia", |
| "middle": [], |
| "last": "Terragni", |
| "suffix": "" |
| }, |
| { |
| "first": "Debora", |
| "middle": [], |
| "last": "Nozza", |
| "suffix": "" |
| }, |
| { |
| "first": "Elisabetta", |
| "middle": [], |
| "last": "Fersini", |
| "suffix": "" |
| }, |
| { |
| "first": "Messina", |
| "middle": [], |
| "last": "Enza", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Proceedings of the First Workshop on Insights from Negative Results in NLP", |
| "volume": "", |
| "issue": "", |
| "pages": "32--40", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/2020.insights-1.5" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Silvia Terragni, Debora Nozza, Elisabetta Fersini, and Messina Enza. 2020b. Which matters most? com- paring the impact of concept and document relation- ships in topic models. In Proceedings of the First Workshop on Insights from Negative Results in NLP, pages 32-40.", |
| "links": null |
| }, |
| "BIBREF33": { |
| "ref_id": "b33", |
| "title": "A review of topic modeling methods", |
| "authors": [ |
| { |
| "first": "Ike", |
| "middle": [], |
| "last": "Vayansky", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [ |
| "P" |
| ], |
| "last": "Sathish", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Kumar", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "", |
| "volume": "94", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "DOI": [ |
| "10.1016/j.is.2020.101582" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ike Vayansky and Sathish A. P. Kumar. 2020. A re- view of topic modeling methods. Information Sys- tems, 94:101582.", |
| "links": null |
| }, |
| "BIBREF34": { |
| "ref_id": "b34", |
| "title": "Evaluation methods for topic models", |
| "authors": [ |
| { |
| "first": "Iain", |
| "middle": [], |
| "last": "Hanna M Wallach", |
| "suffix": "" |
| }, |
| { |
| "first": "Ruslan", |
| "middle": [], |
| "last": "Murray", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Salakhutdinov", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Mimno", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Proceedings of the 26th annual international conference on machine learning", |
| "volume": "", |
| "issue": "", |
| "pages": "1105--1112", |
| "other_ids": { |
| "DOI": [ |
| "10.1145/1553374.1553515" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hanna M Wallach, Iain Murray, Ruslan Salakhutdinov, and David Mimno. 2009. Evaluation methods for topic models. In Proceedings of the 26th annual in- ternational conference on machine learning, pages 1105-1112.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "Workflow of the OCTIS framework", |
| "num": null |
| }, |
| "FIGREF1": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "loading of a pre-processed dataset dataset = Dataset() dataset.load(\"path/to/dataset\") #model instantiation lda = LDA(num_topics=25) #definition of the metric td = TopicDiversity() #definition of the search space search_space = { \"eta\": Real(low=0.01, high=5.0), \"alpha\": Real(low=0.01, high=5.0) } #define and launch optimization optimizer=Optimizer() opt_result = optimizer.optimize(model, dataset, td, search_space)", |
| "num": null |
| }, |
| "FIGREF2": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "Example of the best-seen evolution for an optimization experiment.", |
| "num": null |
| }, |
| "FIGREF3": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "Example of box plot of an optimization experiment.", |
| "num": null |
| }, |
| "FIGREF4": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "Example of word cloud of a topic.", |
| "num": null |
| }, |
| "FIGREF5": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "Example of distribution of the topics in a selected document.", |
| "num": null |
| }, |
| "FIGREF6": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "Example of the weight of the word \"network\" for each document.", |
| "num": null |
| }, |
| "TABREF0": { |
| "num": null, |
| "text": "reports some statistics about the currently available pre-processed.", |
| "html": null, |
| "content": "<table><tr><td>Dataset</td><td>Domain</td><td># Docs</td><td>Avg # words in docs</td><td># Unique words</td></tr><tr><td>20 News-groups</td><td>Forum posts</td><td>16309</td><td>48</td><td>1612</td></tr><tr><td>BBC News</td><td>News</td><td>2225</td><td>150</td><td>3106</td></tr><tr><td>M10</td><td>Scientific papers</td><td>8355</td><td>6</td><td>1696</td></tr><tr><td>DBLP</td><td>Scientific papers</td><td>54595</td><td>5</td><td>1513</td></tr></table>", |
| "type_str": "table" |
| }, |
| "TABREF1": { |
| "num": null, |
| "text": "", |
| "html": null, |
| "content": "<table/>", |
| "type_str": "table" |
| }, |
| "TABREF3": { |
| "num": null, |
| "text": "Comparison between OCTIS and the most well-known topic modeling libraries.", |
| "html": null, |
| "content": "<table/>", |
| "type_str": "table" |
| }, |
| "TABREF4": { |
| "num": null, |
| "text": "", |
| "html": null, |
| "content": "<table><tr><td>summarizes the main features of the</td></tr></table>", |
| "type_str": "table" |
| } |
| } |
| } |
| } |