The Naïve Bayes model for unsupervised word sense disambiguation : aspects concerning feature selection
Author(s)
Bibliographic Information
The Naïve Bayes model for unsupervised word sense disambiguation : aspects concerning feature selection
Springer, c2013
Available at 2 libraries
  Aomori
  Iwate
  Miyagi
  Akita
  Yamagata
  Fukushima
  Ibaraki
  Tochigi
  Gunma
  Saitama
  Chiba
  Tokyo
  Kanagawa
  Niigata
  Toyama
  Ishikawa
  Fukui
  Yamanashi
  Nagano
  Gifu
  Shizuoka
  Aichi
  Mie
  Shiga
  Kyoto
  Osaka
  Hyogo
  Nara
  Wakayama
  Tottori
  Shimane
  Okayama
  Hiroshima
  Yamaguchi
  Tokushima
  Kagawa
  Ehime
  Kochi
  Fukuoka
  Saga
  Nagasaki
  Kumamoto
  Oita
  Miyazaki
  Kagoshima
  Okinawa
  Korea
  China
  Thailand
  United Kingdom
  Germany
  Switzerland
  France
  Belgium
  Netherlands
  Sweden
  Norway
  United States of America
Note
Includes bibliographical references and index
Description and Table of Contents
Description
This book presents recent advances (from 2008 to 2012) concerning use of the Naive Bayes model in unsupervised word sense disambiguation (WSD).
While WSD, in general, has a number of important applications in various fields of artificial intelligence (information retrieval, text processing, machine translation, message understanding, man-machine communication etc.), unsupervised WSD is considered important because it is language-independent and does not require previously annotated corpora. The Naive Bayes model has been widely used in supervised WSD, but its use in unsupervised WSD has led to more modest disambiguation results and has been less frequent. It seems that the potential of this statistical model with respect to unsupervised WSD continues to remain insufficiently explored.
The present book contends that the Naive Bayes model needs to be fed knowledge in order to perform well as a clustering technique for unsupervised WSD and examines three entirely different sources of such knowledge for feature selection: WordNet, dependency relations and web N-grams. WSD with an underlying Naive Bayes model is ultimately positioned on the border between unsupervised and knowledge-based techniques. The benefits of feeding knowledge (of various natures) to a knowledge-lean algorithm for unsupervised WSD that uses the Naive Bayes model as clustering technique are clearly highlighted. The discussion shows that the Naive Bayes model still holds promise for the open problem of unsupervised WSD.
Table of Contents
1.Preliminaries.- 2.The Naive Bayes Model in the Context of Word Sense Disambiguation.- 3.Semantic WordNet-based Feature Selection.- 4.Syntactic Dependency-based Feature Selection.- 5.N-Gram Features for Unsupervised WSD with an Underlying Naive Bayes Model References.- Index.
by "Nielsen BookData"