Which metrics should the data scientist use to optimize the classifier?
(Choose two.)
Specificity
False positive rate
Accuracy
F1 score
True positive rate
Explanations:
Specificity measures the proportion of actual negatives that are correctly identified as such. In fraud detection, where the positive class (fraudulent transactions) is rare, optimizing for specificity may lead to missing many fraudulent transactions, which is not desirable.
False positive rate measures the proportion of actual negatives incorrectly identified as positives. In the context of fraud detection, minimizing false positives is important, but it does not help in capturing fraudulent transactions effectively, which is the primary goal.
Accuracy measures the overall correctness of the model (true positives + true negatives) divided by total cases. Given the low prevalence of fraud (2%), a model could achieve high accuracy by predicting most transactions as non-fraudulent, but this would fail to capture fraudulent cases effectively.
The F1 score is the harmonic mean of precision and recall, providing a balance between the two. It is especially useful in imbalanced classes like fraud detection, where maximizing the identification of fraudulent transactions (recall) while maintaining reasonable precision is crucial.
True positive rate (sensitivity or recall) measures the proportion of actual positives (fraudulent transactions) that are correctly identified. In this scenario, optimizing for the true positive rate is critical to capturing as many fraudulent transactions as possible.