Improving document clustering in a learned concept space
Most document clustering algorithms operate in a high dimensional bag-of-words space. The inherent presence of noise in such representation obviously degrades the performance of most of these approaches. In this paper we investigate an unsupervised dimensionality reduction technique for document clu...
Gespeichert in:
Veröffentlicht in: | Information processing & management 2010-03, Vol.46 (2), p.180-192 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Most document clustering algorithms operate in a high dimensional bag-of-words space. The inherent presence of noise in such representation obviously degrades the performance of most of these approaches. In this paper we investigate an unsupervised dimensionality reduction technique for document clustering. This technique is based upon the assumption that terms co-occurring in the same context with the same frequencies are semantically related. On the basis of this assumption we first find term clusters using a classification version of the
EM algorithm. Documents are then represented in the space of these term clusters and a multinomial mixture model (
MM) is used to build document clusters. We empirically show on four document collections,
Reuters-21578,
Reuters RCV2-French,
20Newsgroups and
WebKB, that this new text representation noticeably increases the performance of the
MM model. By relating the proposed approach to the Probabilistic Latent Semantic Analysis (
PLSA) model we further propose an extension of the latter in which an extra latent variable allows the model to co-cluster documents and terms simultaneously. We show on these four datasets that the proposed extended version of the
PLSA model produces statistically significant improvements with respect to two clustering measures over all variants of the original
PLSA and the
MM models. |
---|---|
ISSN: | 0306-4573 1873-5371 |
DOI: | 10.1016/j.ipm.2009.09.007 |