Negation Naive Bayes for Text Classification

  • Komiya Kanako
    Institute of Engineering, Tokyo University of Agriculture and Technology
  • Ito Yusuke
    Faculty of Engineering, Tokyo University of Agriculture and Technology
  • Sato Naoto
    Graduate School of Engineering, Tokyo University of Agriculture and Technology
  • Kotani Yoshiyuki
    Institute of Engineering, Tokyo University of Agriculture and Technology

Bibliographic Information

Other Title
  • 文書分類のための Negation Naive Bayes
  • ブンショ ブンルイ ノ タメ ノ Negation Naive Bayes

Search this article

Abstract

In this study, we proposed negation naive Bayes (NNB), a new method for text classification. Similar to complement naive Bayes (CNB), NNB uses the complement class. However, unlike CNB, NNB properly considers the prior in a mathematical way because NNB is derivable from the same equation (the maximum a posteriori equation) from which naive Bayes (NB) is derived. We carried out classification experiments on products offered on an internet auction site and on the 20 Newsgroups data set. For the latter, we carried out experiments in the following two settings and discussed the properties of NNB: (1) settings in which the number of words in each document decreases and (2) settings in which the distribution of documents over classes is skewed. We compared NNB with NB, CNB, and support vector machine (SVM). Our experiments showed that NNB outperforms other Bayesian approaches when the number of words in each document decreases and when texts are distributed non-uniformly over classes. Our experiments also showed that NNB sometimes provides the best accuracy and significantly outperforms SVM.

Journal

References(20)*help

See more

Related Projects

See more

Details 詳細情報について

Report a problem

Back to top