LogoLogo
RedBrick AIGuides
  • Quick Start
    • Walkthrough Guides
    • Get Started with a Project
    • Get Started with Workspace
      • Cohort Creation
      • Datapoint Classification
      • Configuring Metadata Schema
    • Creating a RedBrick AI Account
  • Organizations
    • Organization and Project Roles
    • Inviting Your Team
      • Single Sign-on
  • Dashboard
    • Account Settings
    • User Preferences
    • Worklist
    • Preview Tool
    • Integrations
    • Taxonomies
    • Boost
      • Auto Annotator
    • Home
      • Sections
  • Importing Data
    • Uploading Data to RedBrick
    • Import Cloud Data
      • Configuring AWS s3
      • Configuring Azure Blob
      • Configuring GCS
      • Configuring AltaDB
      • Creating an Items List
    • Troubleshooting
  • Projects
    • Tasks & Assignment
    • Comments & Raise Issue
    • Reference Standards
    • Project & Task Analytics
    • Labeler Evaluation
  • Project Pages
    • Data Page
    • Settings Page
      • Custom Label Validation
      • Custom Hanging Protocol
      • Webhooks
    • Multiple Labeling
      • Consensus
        • Agreement calculation
      • Task duplication
  • Annotation & viewer
    • Viewer Basics
      • Document Viewer
      • Multiple Modalities
      • Intellisync
      • Annotation Mirroring
    • Creating, Editing and Deleting Annotations
    • Visualization and Masking
    • Segmentation
      • Segmentation Tools
      • Instance vs. Semantic
      • Overlapping Segmentations
    • Heat maps
  • Python SDK & CLI
    • Full documentation
    • Installation & API Keys
    • SDK Overview
      • Importing Data & Annotations
      • Programmatic Label & Review
      • Assigning & Querying Tasks
      • Exporting Annotations
    • CLI Overview
      • Creating & Cloning Projects
      • Import Data & Annotations
      • Exporting Annotations
    • Importing Annotations Guide
    • Formats
      • Full Format Reference
      • Export Structure
  • Useful Links
    • Privacy Policy
Powered by GitBook
On this page

Was this helpful?

  1. Project Pages
  2. Multiple Labeling
  3. Consensus

Agreement calculation

This section of the documentation will cover how RedBrick AI calculates inter-annotator agreement between two users.

For two sets of labels, annotation instances are matched up by category. For the same category, instances are matched up by selecting pairs that maximize the overall agreement score. For two instances of the same category, RedBrick AI uses the following similarity functions

Bounding box, Polygon, and Pixel Segmentation

RedBrick AI uses IOU for these annotation types. For two annotations A and B IOU is defined by:

IOU=A∪BA∩BIOU = \frac{A\cup B}{A\cap B}IOU=A∩BA∪B​

​Landmarks

For landmarks/keypoints, RedBrick AI uses a normalized Root Mean Squared Error (RMSE) to compute similarity, where similarity is Similarity=1−RMSESimilarity = 1 - RMSESimilarity=1−RMSE.

MSE=1n∑in(Pi−P^i)2RMSE=MSEMSE = \frac{1}{n}\sum_{i}^{n}(P_{i} - \hat P_{i})^2 \\ RMSE = \sqrt{MSE}MSE=n1​i∑n​(Pi​−P^i​)2RMSE=MSE​

Where nnn​ is the number of components of the point (2 for 2D, 3 for 3D), and Pi,Pi^P_i, \hat{P_i} Pi​,Pi​^​​ are normalized components (by width, height, depth of the image) of the two points.

Length Measurements

Comparisons of length measurements are done by comparing the two sets of points (using the technique covered above) that define the length line.

Angle Measurements

For angle measurements, the vectors between each arm of the angle measurement are compared. The two angles comparing both sets of measurement arms are computed. The similarity score is then defined by:

Similarity=1−θ1+θ22πSimilarity = 1 - \frac{\theta_1 + \theta_2}{2\pi}Similarity=1−2πθ1​+θ2​​

​Where θ1,θ2\theta_1, \theta_2θ1​,θ2​​ are the angles between the two sets of measurement arms.

Classification

For classification labels, the agreement is binary. If the chosen category and attributes match, the consensus score will be 100%, otherwise, it will be 0%.

Generating a single score

To generate a single score between two sets of labels, a series of averages are computed.

  1. Scores of matching annotations instances of the same category are averaged, to generate a single score per category.

  2. Scores are then averaged per category.

  3. Scores are then averaged per label type to generate a single score per label type.

  4. For videos, scores are calculated per frame and averaged to generate a single score per sequence.

  5. For multi-series studies, scores are averaged by volume to generate a single score per study. ​

PreviousConsensusNextTask duplication

Last updated 2 years ago

Was this helpful?