from sklearn import naive_bayes
from sklearn import metrics
from sklearn.model_selection import train_test_split
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
def readFileThroughPandas(filename):
# Reads the entire data file
data = pd.read_csv(filename)
att = data[["Marital Status","Kids","Annual Household Salary","Loan Amount","Car owner", "Education Level"]]
lab = data["Loan Granted"]
# Let us do a normalization of our dataset because the attributes are of significantly different orders of magnitude
# Standard deviation based normalization
att=(att-att.mean())/att.std()
# Zero-to-One normalization
# att=(att-att.min())/(att.max()-att.min())
return(att,lab)
(att,lab) = readFileThroughPandas("loanacceptance.csv")
# shape of the variables
print(att.shape)
print(lab.shape)
# Use the first 400 rows for training and the remaining rows for testing
x_train = att.iloc[0:400]
y_train = lab.iloc[0:400]
x_test = att.iloc[400:]
y_test = lab.iloc[400:]
# Alternatively use the following code to choose the rows randomly
# x_train, x_test, y_train, y_test = train_test_split(att, lab, test_size = 0.20)
(500, 6) (500,)
# Note that only 80% of the dataset is being used for training
clf = naive_bayes.GaussianNB()
clf.fit(X=x_train,y=y_train)
GaussianNB()
# returns accuracy
print("Training accuracy",clf.score(X=x_train,y=y_train))
print("Testing accuracy",clf.score(X=x_test,y=y_test))
y_predicted = clf.predict(x_test)
# Following is an alternative way to get the accuracy scores
# print("Testing accuracy",metrics.accuracy_score(y_test,y_predicted))
Training accuracy 0.8802992518703242 Testing accuracy 0.84
m = metrics.confusion_matrix(y_test,y_predicted,labels=clf.classes_)
print(m)
[[21 9] [ 7 63]]
# Better visualization of a confusion matrix
disp = metrics.ConfusionMatrixDisplay(confusion_matrix=m,display_labels=clf.classes_)
disp.plot()
plt.show()