Classifier evaluation methods are used to measure the quality of predictions from a classification algorithm. We have multiple aproaches, will try to cover some of the methods here.
This is the most common method to evaluate a classifier. In this method, the given data set is divided into two parts as a test and train set 20% and 80% respectively. The train set is used to train the data and the unseen test set is used to test its predictive power.
In this method, the data set is randamly partitioned into k mutually exclusive subsets, each of which is of the same size. Out of these, one is kept for testing and others are used to train the model. The same process takes place for all k folds.
cross validation is the best solution for overfitting problem. We can't know how well our model will perform on new data until we thoroughly tested it. To cross this, we can split initial dataset into separate training and test subsets.
Classification report is to find out how many predictions are True and how many are False. More specifically, True Positives, False Positives, True negatives and False Negatives are used to predict the metrics of a classification report.
Positive and negative are generic names for the predicted classes. There are four ways to check if the predictions are right or wrong:
Receiver Operating Characteristics or ROC Curve is used for visual comparision of classification models, which shows the relationship between the true positive rate and the false positive rate. The area under the ROC curve is the measure of the accuracy oh the model.
The gree line denotes the change of TPR(True Positive Rate) with different FPR(False Positive Rate) for a model. More the ratio of the area under the curve and the total area (100 x 100 in this case) defines more the accuracy of the model. If it becomes 1, the model is overfit and if it is equal below 0.5 (i.e when the curve is along the dotted diagonal line), the model is inaccurate to use.