Gaussian Naive Bayes - Jaikrishna 3
Gaussian Naive Bayes - Jaikrishna 3
UNIT-III S4 SLO-1
INTRODUCTION TO NAIVE BAYES
CLASSIFIER
Definition: Naive Bayes is a probabilistic classification algorithm based on Bayes’ Theorem
Bayes’ Theorem: Provides a way to update the probability estimate for a hypothesis as more
evidence is acquired.
Naivety Assumption: Assumes that the features are independent given the class label.
Gaussian Naive Bayes: A variant of Naive Bayes where the features are assumed to follow a
Feature Distribution: For each class, the features are modeled using a Gaussian distribution:
the mean and σi2\sigma_i^2σi2is the variance of feature iii in class CCC.
Objective: Calculate the posterior probability of a class given the features and classify based
Training:
Calculate the mean (μ\muμ) and variance (σ2\sigma^2σ2) of each feature for each class from the
training data.
Prediction:
For a new data point, compute the posterior probability for each class using:
Applications: Used in text classification (spam detection), medical diagnosis, and any scenario
Advantages:
Simplicity and Efficiency: Fast to train and predict.
Good Performance with Large Datasets: Performs well even with relatively simple assumptions.
Limitations:
Independence Assumption: Assumes features are independent, which may not hold in real-world data.
Gaussian Assumption: Assumes features follow a Gaussian distribution, which might not always be true.