Model Performance Assessment
Model Performance Assessment
Model Performance
Model evaluation is the process that uses some metrics which help us
to analyze the performance of the model.
There are many types of Cross-Validation out of which K Fold Cross Validation is mostly used.
In K Fold Cross Validation the original dataset is divided into k subsets. The subsets are known
as folds.
Rest k-1 folds are used for training the model. It is seen that this technique generalizes the
model well and reduces the error rate.
Holdout
In this technique, the dataset is divided into train and test datasets.
Normally a large percentage of data is used for training the model and a
small portion of the dataset is used for testing the model.
Accuracy
Precision = TP/(TP+FP)
Recall = TP/(TP+FN)
True Positives: It is also known as TP. It is the output in which the actual and the
predicted values are YES.
True Negatives: It is also known as TN. It is the output in which the actual and
the predicted values are NO.
False Positives: It is also known as FP. It is the output in which the actual value is
NO but the predicted value is YES.
False Negatives: It is also known as FN. It is the output in which the actual value
Confusion Matrix
Area Under Curve (AUC) /The Receiver Operating Characteristic(ROC) curve
FPR: It stands for False Positive rate. It is defined as the ratio of False positives to the
summation of false positives and True negatives.
This curve is useful as it helps us to determine the model’s capacity to distinguish between
different classes.
A model is considered good if the AUC score is greater than 0.5 and approaches 1. A poor
model has an AUC score of 0.
AUC or ROC
Area under the ROC Curve (AUC) measures how much better a machine learning model predicts classification
versus a random luck model.