How Many Topics? Stability Analysis for Topic Models
|Title:||How Many Topics? Stability Analysis for Topic Models||Authors:||Greene, Derek
|Permanent link:||http://hdl.handle.net/10197/6617||Date:||19-Sep-2014||Abstract:||Topic modeling refers to the task of discovering the underlyingthematic structure in a text corpus, where the output is commonlypresented as a report of the top terms appearing in each topic. Despitethe diversity of topic modeling algorithms that have been proposed, acommon challenge in successfully applying these techniques is the selectionof an appropriate number of topics for a given corpus. Choosingtoo few topics will produce results that are overly broad, while choosingtoo many will result in theover-clustering of a corpus into many small,highly-similar topics. In this paper, we propose a term-centric stabilityanalysis strategy to address this issue, the idea being that a model withan appropriate number of topics will be more robust to perturbations inthe data. Using a topic modeling approach based on matrix factorization,evaluations performed on a range of corpora show that this strategy cansuccessfully guide the model selection process.||Type of material:||Conference Publication||Copyright (published version):||2014 Springer||Keywords:||Statistics;Machine learning;Latent Dirichlet Allocation (LDA);Non-negative Matrix Factorization (NMF);Topic modeling;Corpora||DOI:||10.1007/978-3-662-44848-9_32||Language:||en||Status of Item:||Peer reviewed||Conference Details:||European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML '14), 15-19 September, Nancy, France|
|Appears in Collections:||Computer Science Research Collection|
Insight Research Collection
Show full item record
This item is available under the Attribution-NonCommercial-NoDerivs 3.0 Ireland. No item may be reproduced for commercial purposes. For other possible restrictions on use please refer to the publisher's URL where this is made available, or to notes contained in the item itself. Other terms may apply.