Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Naive bayes #68

Merged
merged 10 commits into from
Jun 23, 2021
128 changes: 128 additions & 0 deletions numpy_ml/naive_bayes/naive_bayes.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,128 @@
import numpy as np

class GaussianNB():
"""
Gaussian Naive Bayes

Assume each class conditional feature distribution is
independent and estimate the mean and variance from the
training data

Parameters
----------
epsilon: float
a value that add to variance to prevent numerical error

Attributes
----------
num_class : ndarray of shape (n_classes,)
count of each class in the training sample

mean: ndarray of shape (n_classes,)
mean of each variance

sigma: ndarray of shape (n_classes,)
variance of each class

prior : ndarray of shape (n_classes,)
probability of each class

"""
def __init__(self,eps=1e-6):
self.eps = eps

def fit(self,X,y):
"""
Train the model with X,y

Parameters
----------
X: ndarray of shape (n_samples, n_features)
Input data
y: ndarray of shape (n_samples,)
Target

returns
--------
self: object
"""

self.n_sample, self.n_features = X.shape
self.labels = np.unique(y)
self.n_classes = len(self.labels)

self.mean = np.zeros((self.n_classes,self.n_features))
self.sigma = np.zeros((self.n_classes,self.n_features))
self.prior = np.zeros((self.n_classes,))

for i in range(self.n_classes):
X_c = X[y==i,:]

self.mean[i,:] = np.mean(X_c,axis=0)
self.sigma[i,:] = np.var(X_c,axis=0) + self.eps
self.prior[i] = X_c.shape[0]/self.n_sample

return self

def predict(self,X):
"""
used the trained model to generate prediction

Parameters
---------
X: ndarray of shape (n_samples, n_features)
Input data

returns
-------
probs : ndarray of shape (n_samples, n_classes)
The model predictions for each items in X to be in each class
"""

probs = np.zeros((X.shape[0],self.n_classes))
for i in range(self.n_classes):
probs[:,i] = self.prob(X,self.mean[i,:],self.sigma[i,:],self.prior[i])


return probs

def prob(self,X,mean,sigma,prior):
"""
compute the joint log likelihood of data based on gaussian distribution
Copy link
Owner

@ddbourgin ddbourgin May 30, 2021

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Nomenclature: I'm not sure it's right to call this the joint log likelihood? That is, this function computes the unnormalized quantity P(y = c | X, mean_c, sigma_c), which I'd think is the class posterior.


X: ndarray of shape (n_samples, n_features)
Input data

mean: ndarray of shape (n_classes,)
mean of each variance

sigma: ndarray of shape (n_classes,)
variance of each class

prior : ndarray of shape (n_classes,)
probability of each class

returns
-------
joint_log_likelihood : ndarry of shape (n_samples,)
joint log likelihood of data

"""

prob = -self.n_features / 2 * np.log(2 * np.pi) - 0.5 * np.sum(
np.log(sigma )
)
prob -= 0.5 * np.sum(np.power(X -mean, 2) / (sigma), 1)

joint_log_likelihood = prior + prob
return joint_log_likelihood
Copy link
Owner

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think there are a few errors here.

  • The log Gaussian likelihood calc in prob isn't quite right (see later commit for details)
  • In the joint_log_likelihood = prior + prob line, you're adding a log-transformed vector (prob) to raw probabilities (prior), which doesn't make sense. I think what you want is np.log(prior) + prob. See my later commit for details.
  • This isn't a joint likelihood, right? You're computing the the joint class likelihood in prob, but when you add in the prior, this will be proportional to the log class posterior.