Naive Bayes Classification

Harvard EPS-210 | Interactive tutorial — Click to add data points and explore probabilistic classification

Classification Space

Class 0
Class 1
Probe Point

Class-Conditional Distributions P(x|C)

P(x|C) = P(x₁|C) × P(x₂|C)
"Naive" independence assumption

Feature Distributions (1D Marginals)

Posterior Probability P(C|x)

P(C|x) = P(x|C) × P(C) / P(x)
Bayes' Theorem

Add Data Points

Click on the classification canvas to add points

Prior Probabilities P(C)

P(C₀) = 0.50
P(C₁) = 0.50

Computed from class frequencies

Model Statistics

Accuracy
--
Total Points
0
Class 0
0
Class 1
0
Hover over classification space to probe
Position: --
P(C₀|x): --
P(C₁|x): --
Prediction: --

Quick Examples

How it works: Naive Bayes assumes features are conditionally independent given the class. It estimates P(x|class) from data, then uses Bayes' theorem to compute P(class|x) for classification.