Overview
This application allows you to compare annotations made by different users, generate a report with metrics and perform actions with images based on the comparison results.
How To Run
Step 1: Run the application from the ecosystem.
Step 2: Wait until the app starts.
Once the app is started, a new task appear in workspace tasks. Wait for the message Application is started ...
and then press Open
button.
Step 3: Open the app.
Step 4: Select projects, datasets and annotators to compare.
You can select multiple annotators and datasets. You can see added annotations in the table to the right. To remove an annotation from comparison, select it in the table and press the Remove
button.

Step 5: Run the comparison.
Press the Calculate consensus
button to start the comparison. The comparison may take a long time, depending on the number of images and annotations. After the calculation is finished, you will see the results in the comparison matrix.

Step 6: View the detailed report.
Click on a cell of the comparison matrix with a score to see a detailed report for the given pair of annotations.

What Report metrics mean
Table Objects counts per class
- GT Objects - Total number of objects of the given class on all images in the benchmark dataset
- Labeled Objects - Total number of objects of the given class on all images in the exam dataset
- Recall(Matched objects) - Number of correctly labeled objects of the given class divided by the total number of objects of the given class on all images in the benchmark dataset.
- Precision - Number of correctly labeled objects of the given class divided by the total number of objects of the given class on all images in the exam dataset
- F-measure - Harmonic mean of Recall and Precision
- Objects Score (average F-measure) - Average of F-measures of all classes
Table Geometry quality
- Pixel Accuracy - Number of correctly labeled or unlabeled pixels on all images divided by the total number of pixels
- IOU - Intersection over union. Sum of intersection areas of all objects on all images divided by the sum of union areas of all objects on all images for the given class
- Geometry Score (average IoU) - Average of IOUs of all classes
- GT Tags - Total number of tags of all objects on all images in the benchmark dataset
- Labeled Tags - Total number of tags of all objects on all images in the exam dataset
- Precision - Number of correctly labeled tags divided by the total number of tags of all objects on all images in the exam dataset
- Recall - Number of correctly labeled tags divided by the total number of tags of all objects on all images in the benchmark dataset
- F-measure - Harmonic mean of Recall and Precision
- Tags Score - Average of F-measures of all tags
Table Report per image
- Objects Score - F-measure of objects of any class on the given image
- Objects Missing - Number of objects of any class on the given image that are not labeled
- Objects False Postitve - Number of objects of any class on the given image that are labeled but not present in the benchmark dataset
- Tags Score - F-measure of tags of any object on the given image
- Tags Missing - Number of tags of any object on the given image that are not labeled
- Tags False Positive - Number of tags of any object on the given image that are labeled but not present in the benchmark dataset
- Geometry Score - Intersection over union for all objects of any class on the given image
- Overall Score - Average of Objects Score, Tags Score and Geometry Score