Ayuda
Ir al contenido

Dialnet


Alternative prior assumptions for improving the performance of naïve Bayesian classifiers

  • Autores: Tzu-Tsung Wong
  • Localización: Data mining and knowledge discovery, ISSN 1384-5810, Vol. 18, Nº 2, 2009, págs. 183-213
  • Idioma: inglés
  • Texto completo no disponible (Saber más ...)
  • Resumen
    • Abstract: The prior distribution of an attribute in a naïve Bayesian classifier is typically assumed to be a Dirichlet distribution, and this is called the Dirichlet assumption. The variables in a Dirichlet random vector can never be positively correlated and must have the same confidence level as measured by normalized variance. Both the generalized Dirichlet and the Liouville distributions include the Dirichlet distribution as a special case. These two multivariate distributions, also defined on the unit simplex, are employed to investigate the impact of the Dirichlet assumption in naïve Bayesian classifiers. We propose methods to construct appropriate generalized Dirichlet and Liouville priors for naïve Bayesian classifiers. Our experimental results on 18 data sets reveal that the generalized Dirichlet distribution has the best performance among the three distribution families. Not only is the Dirichlet assumption inappropriate, but also forcing the variables in a prior to be all positively correlated can deteriorate the performance of the naïve Bayesian classifier.


Fundación Dialnet

Dialnet Plus

  • Más información sobre Dialnet Plus

Opciones de compartir

Opciones de entorno