Welcome to Journal of Beijing Institute of Technology
Volume 14Issue 1
.
Turn off MathJax
Article Contents
LIU Li-zhen, SONG Han-tao, LU Yu-chang. Dimensionality Reduction by Mutual Information for Text Classification[J]. JOURNAL OF BEIJING INSTITUTE OF TECHNOLOGY, 2005, 14(1): 32-36.
Citation: LIU Li-zhen, SONG Han-tao, LU Yu-chang. Dimensionality Reduction by Mutual Information for Text Classification[J].JOURNAL OF BEIJING INSTITUTE OF TECHNOLOGY, 2005, 14(1): 32-36.

Dimensionality Reduction by Mutual Information for Text Classification

Funds:theNational“973”ProgramProjects(G1998030414)
  • Received Date:2003-08-29
  • The frame of text classification system was presented. The high dimensionality in feature space for text classification was studied. The mutual information is a widely used information theoretic measure, in a descriptive way, to measure the stochastic dependency of discrete random variables. The measure method was used as a criterion to reduce high dimensionality of feature vectors in text classification on Web. Feature selections or conversions were performed by using maximum mutual information including linear and non-linear feature conversions. Entropy was used and extended to find right features commendably in pattern recognition systems. Favorable foundation would be established for text classification mining.
  • loading
  • [1]
    Hutter M. Distributio n of mutual informat ion[Z]. 14thInternational Conference on Neural I nformation P rocessing Systems, V ancouver, 2001.
    [2]
    Yang Y, Pedersen J O. A comparative study on featureselectio n in tex t classification[Z]. KDD 2000 Sixth A CMSIGK DD Inter national Conference on Knowledg e Discover y and Data M ining, Boston, M A, 2000.
    [3]
    Torkkola K. On feature ex traction by mutual infor mationmaximization[EB/OL]. http:(citeseer. nj. nec. com/torkkola02feature. htm.
    [4]
    Jongzhi S. Knowledg e discover y[M]. Beijing:T singhuaU niversity P ress, 2002. 338-348.
    [5]
    G alav otti L, Sebastiani F, Simi M. F eatur e selection andnegative ev idence in automated t ex t classificat ion[Z]. K DD 2000 Six th ACM SIG KDD Inter nation Conferenceon K now ledge Discov er y and Data M ining, Boston, M A, 2000.
    [6]
    Zaffalon M, Hutter M. Robust featur e selection by mutual infor matio n distributio ns[EB/OL]. http:(ww w. idsia. ch/marcus/ai/feature. htm.
    [7]
    T o rkkola K, W illiam M. Campbell mutual infor mation inlearning feature transformat ions[Z]. 17th InternationalConf on M achine L ear ning, San Francisco, CA, 2000.
    [8]
    T o rkkola K. Nonlinear feature transforms using max imum mutual information[Z]. Conf of the IJCN N, Washington D. C. , 2001.
    [9]
    K arypis G, Han E H. Concept indexing:A fast dimensio nality reduction algor ithm w ith applications to document retrieval and classification[Z]. 9th Intl Conf Information and Knowledg e M anagement CIK M, Washington D. C. , 2000.
  • 加载中

Catalog

    通讯作者:陈斌, bchen63@163.com
    • 1.

      沈阳化工大学材料科学与工程学院 沈阳 110142

    1. 本站搜索
    2. 百度学术搜索
    3. 万方数据库搜索
    4. CNKI搜索

    Article Metrics

    Article views (208) PDF downloads(0) Cited by()
    Proportional views
    Related

    /

      Return
      Return
        Baidu
        map