An information theoretic approach to finding word groups for text classification

Abstract : This thesis concerns finding the 'optimal' number of (non-overlapping) word groups for text classification. We present a method to select which words to cluster in word groups and how many such word groups to use on the basis of a set of pre-classified texts. The method involves a greedy search through the space of possible word groups. The criterion on which is navigated through this space is based on 'mutual information' and is known as 'Jensen Shannon divergence'. The criterion to decide which number of word groups to use is based on Rissanen's MDL Principle. We present empirical results that indicate that the proposed method performs well at its task. The prediction model used is based on the Naive Bayes model and the data set used for the experiments is a subset of the 20 Newsgroup data set.
Document type :
Master thesis
Complete list of metadatas


https://hal.inria.fr/inria-00321519
Contributor : Jakob Verbeek <>
Submitted on : Wednesday, February 16, 2011 - 5:00:56 PM
Last modification on : Monday, September 25, 2017 - 10:08:04 AM
Long-term archiving on : Tuesday, May 17, 2011 - 2:38:35 AM

Files

verbeek00msc.pdf
Files produced by the author(s)

Identifiers

  • HAL Id : inria-00321519, version 1

Citation

Jakob Verbeek. An information theoretic approach to finding word groups for text classification. Machine Learning [cs.LG]. 2000. ⟨inria-00321519⟩

Share

Metrics

Record views

194

Files downloads

535