pipeline:window:cryolo:evaluate_results

This version (2020/03/16 15:27) is a *draft*.

Approvals: 0/1

Approvals: 0/1

The evaluation tool allows you, based on your validation micrographs, to get statistics about the success of your training.

To understand the outcome, you have to know what precision and recall is. Here is good figure from wikipedia:

Another important measure is the F1 (β=1) and F2 (β=2) score:

If your validation micrographs are not labeled to completion the **precision** value will be misleading. crYOLO will start picking the remaining 'unlabeled' particles, but for statistics they are counted as false-positive (as the software takes your labeled data as ground truth).

If you followed the tutorial, the validation data are selected randomly. A run file for each training is created and saved into the folder runfiles/ in your project directory. These runfiles are .json files containing information about what micrographs were selected for validation. To calculate evaluation metrics select the *evaluation* action.

Fill out the fields in the *“Required arguments”* tab:

Click to display ⇲

Click to hide ⇱

cryolo_evaluation.py -c config.json -w model.h5 -r runfiles/run_YearMonthDay-HourMinuteSecond.json -g 0

The html file you specified as output looks like this:

The table contains several statistics:

**AUC:**Area under curve of the precision-recall curve. Overall summary statistics. Perfect classifier = 1, Worst classifier = 0**Topt:**Optimal confidence threshold with respect to the F1 score. It might not be ideal for your picking, as the F1 score weighs recall and precision equally. In single particle analysis, recall is often more important than the precision.**R (Topt):**Recall using the optimal confidence threshold.**R (0.3):**Recall using a confidence threshold of 0.3.**R (0.2):**Recall using a confidence threshold of 0.2.**P (Topt):**Precision using the optimal confidence threshold.**P (0.3):**Precision using a confidence threshold of 0.3.**P (0.2):**Precision using a confidence threshold of 0.2.**F1 (Topt):**Harmonic mean of precision and recall using the optimal confidence threshold.**F1 (0.3):**Harmonic mean of precision and recall using a confidence threshold of 0.3.**F1 (0.2):**Harmonic mean of precision and recall using a confidence threshold of 0.2.**IOU (Topt):**Intersection over union of the auto-picked particles and the corresponding ground-truth boxes. The higher, the better – evaluated with the optimal confidence threshold.**IOU (0.3):**Intersection over union of the auto-picked particles and the corresponding ground-truth boxes. The higher, the better – evaluated with a confidence threshold of 0.3.**IOU (0.2):**Intersection over union of the auto-picked particles and the corresponding ground-truth boxes. The higher, the better – evaluated with a confidence threshold of 0.2.

If the training data consist of multiple folders, then evaluation will be done for each folder separately. Furthermore, crYOLO estimates the optimal picking threshold regarding the F1 Score and F2 Score. Both are basically average values of the recall and prediction, whereas the F2 score puts more weights on the recall, which is in cryo-EM often more important.

pipeline/window/cryolo/evaluate_results.txt · Last modified: 2020/03/16 15:27 by twagner

Except where otherwise noted, content on this wiki is licensed under the following license: GNU Free Documentation License 1.3