Evaluation
The evaluation metric for this competition is the macro F1 score (i.e., the unweighted mean F1). The F1 score, commonly used in information retrieval, measures accuracy using the statistics precision and recall .
Precision is the ratio of true positives to all predicted positives . Recall is the ratio of true positives to all actual positives . The F1 score is given by:
The F1 metric weighs recall and precision equally. Moderately good performance on both will be favored over extremely good performance on one and poor performance on the other.
Submission Format
You must produce a single submission file based on test.csv containing exactly two columns: ID and LABEL.
The file should contain a header and have the following format:
1ID,LABEL218742,0314108,1