Naive Bayes classifiers

Mau Rua
1 min readSep 30, 2020

--

In statistics, Naive Bayes classifiers are a family of simple “probabilistic classifiers” based on applying Bayes’ theorem with strong (naïve) independence assumptions between the features.

They are among the simplest Bayesian network models.

But they could be coupled with Kernel density estimation and achieve higher accuracy levels.

Naïve Bayes classifiers are highly scalable, requiring a number of parameters linear in the number of variables (features/predictors) in a learning problem.

Maximum-likelihood training can be done by evaluating a closed-form expression, which takes linear time, rather than by expensive iterative approximation as used for many other types of classifiers.

In the statistics and computer science literature, naive Bayes models are known under a variety of names, including simple Bayes and independence Bayes.

All these names reference the use of Bayes’ theorem in the classifier’s decision rule, but naïve Bayes is not (necessarily) a Bayesian method.

--

--

Mau Rua
Mau Rua

Written by Mau Rua

Welcome to my software engineering blog. Please visit my career portfolio at https://mruanova.com 🚀🌎

No responses yet