Table of Contents
Benchmarks is a Labelbox QA tool that automatically compares all labels on a data row to a “gold standard” label you set. Once an asset with a Benchmark label gets a human- or computer-generated label, the Benchmark agreement score is automatically calculated.
Benchmarks is only supported for Bounding boxes, Polygons, and Masks on images.
The following is not supported:
- Points on images
- Polylines on images
- Video labeling
- Text labeling
Labelbox follows a similar methodology for calculating the agreement scores for both Benchmarks and Consensus. The only difference in the calculations is the entity to which the Labels are compared.
Benchmarks works by interspersing data to be labeled, for which there is a Benchmark label, to each person labeling. These labeled data are compared against their respective Benchmark and an accuracy score between 0 and 100 percent is calculated.
When a Label is created or updated, the Benchmarks score will be recalculated as long as there is one Label on the Data Row. if a Label gets deleted, no benchmark score will appear for that Data Row.
Bounding boxes, polygons, and masks
Generally speaking, calculating agreement for the polygons of a Label involves Intersection-over-Union and a series of averages to calculate the final agreement between two Labels on an image.
There are three global classification types supported in Benchmarks: radio, checklist, and dropdown. The calculation method for each classification type is different. One commonality, however, is that if two classifications of the same type are compared and there are no corresponding selections between the two classifications at all, the agreement will be 0%.
A Radio classification can only have one selected answer. Therefore, the agreement between two radio classifications will either be 0% or 100%. 0% means no agreement and 100% means agreement.
A Checklist classification can have more than one selected answer, which makes the agreement calculation a little more complex. The agreement between two checklist classifications is generated by dividing the number of overlapping answers by the number of selected answers.
A Dropdown classification can have only one selected answer, however the answer choices can be nested. The calculation for dropdown is similar to that of checklist classification, except that the agreement calculation divides the number of overlapping answers by the total depth of the selection (how many levels). Answers nested under different top-level classifications can still have overlap if the classifications at the next level match. On the flip side, answers that do not match exactly can still have overlap if they are under the same top-level classification.
Either Benchmarks or Consensus can be turned on for a project at any given time, but it is not possible to have both on at the same time.
- Create a project.
- Navigate to Settings > Quality and select Benchmarks to turn this QA feature on for your project.
- To mark a label as a Benchmark, select a data row from the Activity table to enter Open review mode. In the Label browser, click on the context menu (three dots) on a label and mark it as a benchmark label. When you go back the Activity table, that label you selected as the benchmark will be marked with a gold star.
Benchmark labels are marked with a gold star in the Activity table under the Labels tab.
Under the Labels tab, there is also a Benchmarks table where you’ll see a list of all the Benchmarks labels for that project. Click on “View Results” to see all labels associated with that benchmark label.
When the Benchmarks tool is active for your project, the Individual performance section under the Performance tab will display a Benchmarks column that indicates the average Benchmark score for that labeler.