Skip to main content
Consensus Projects have the following types of reviewers:
  • Determine consensus: Reviewers in these blocks are responsible for ensuring that labels made by one or more annotators are correct across an entire data unit. This allows for bulk label selection across a data unit.
  • Review and refine: Reviewers can select and edit the best specific label on an object or frame, from one or more annotators.

Determine Consensus

Reviewers on Consensus Projects have a different goal from Reviewers on other types of Encord Projects. Reviewers on Consensus Projects review ALL the labels on a data unit, made by a specified number of annotators. The reviewers then make a decision on whether there is agreement/consensus on ALL the labels on a task (data unit). If the reviewer decides that ALL the labels on a data unit are correct/in agreement, the reviewer selects the annotators that are in agreement and clicks Consensus reached. If the reviewer decides that there is no agreement on the labels or that the labels are incorrect, the reviewer clicks No Consensus. Currently, Object labels and Classifications are supported for consensus review on all data unit types. Data units are images, image groups, image sequences, videos, and DICOM series. To review labels on a Consensus Project:
  1. Log in to Encord. Your review task queue appears for the latest Project you are assigned for review.
  2. Go to Annotate > Projects > Annotate projects. The Annotate projects page appears displaying a list of Projects you are assigned to review.
  3. Click a Consensus Project. The Project opens with the Review populated with review tasks.
  4. Click Initiate on a task. The Reviewing page appears.
  5. Review the labels from each annotator.
Determine Consensus in Encord is based on Consensus for ALL labels on a data unit. A data unit being an image, an image group, an image sequence, a video, or a DICOM series. Encord Consensus is not granular to the label itself (yet). This means reviewers must ensure ALL labels on a data unit are in consensus before deciding consensus is reached.
The Reviewing page is different for single images and all other data units. When reviewing single images, expand each user to display their labels and decide on consensus. When reviewing videos, image groups, image sequences, and DICOM series a timeline appears under the data unit. Use the timeline to speed up reviewing all the labels on the data unit.
Use the show-hide labels buttons, workspace settings, and the drag and select in the workspace to speed up your approved label selection and review process.
Single Image, Image Group
Video, Image Sequence, DICOM Series
  1. If the labels are correct:
a. Select the annotators that have done the best job of labeling the data unit. The Consensus reached button becomes clickable once the required number of annotators required are selected. b. Click the Consensus reached button. c. Select the annotator that did the best job. That annotator’s work on the data unit is used if an Admin or Team Manager export the labels for the data unit. d. Click Submit. Single Image, Image Group
Video, Image Sequence, DICOM Series
  1. Click No consensus if the labels are not correct OR if not enough of the annotators have labeled the data unit correctly.

Review and Refine Consensus Labels

Reviewers working in the Review and Refine block can see all labels from all annotators from a Consensus block. The reviewer is able to select the best labels from each object on a data unit and edit those labels or they can create new labels for the data unit. The Review and Refine page is different for single images and all other data units. When reviewing and refining single images, expand each user to display their labels and decide on the best labels. When reviewing and refining videos, image groups, image sequences, and DICOM series, a timeline appears under the data unit. Use the timeline to speed up reviewing and refining all the labels on the data unit.

Enhanced By-Label View

When using the By label view in Review and Refine mode, you’ll notice several improvements designed to streamline your workflow:
  • Organized display: Classifications appear before objects in the label list, making it easier to review different types of annotations in a logical order.
  • Hover-based visibility controls: Eye icons for showing and hiding labels appear only when you hover over label rows, reducing visual clutter while keeping controls accessible.
  • Group-level visibility: Toggle visibility for entire groups of related labels at once, or control individual annotations within each group.
Use the show-hide labels buttons, workspace settings, and the drag and select in the workspace to speed up your approved label selection and review process.Optimize Review
To review and refine labels from a Consensus Project: Only data units for which consensus is not reached can be reviewed and refined.
  1. Log in to Encord. Your review and refine task queue appears for the latest Project you are assigned for review and refinement.
  2. Go to Annotate > Projects > Annotate projects. The Annotate projects page appears displaying a list of Projects you are assigned to review and refine.
  3. Click a Consensus Project. The Project opens with Review and Refine populated with tasks.
  4. Click Initiate on a task. The Review and refine page appears.
  5. Review and refine the labels on each data unit.
Single Image, Image Group
Video, Image Sequence, DICOM Series
  1. Click Edit approved labels to add new labels to the data unit or modify existing approved labels.
When editing approved labels, any changes you make are preserved even if you later revert the approval of those labels. This allows you to refine labels and maintain your edits throughout the review process.
  1. Select the best labels for each object on the data unit.
  2. Click Approve labels. The labels appear in the Approved labels list.
  3. Select a label from the Approved labels list.
  4. Click Edit approved labels if the labels need to be adjusted.
  5. Click Approve task after selecting, refining, and creating the labels needed for the data unit.
  6. Click Reject task if the labels on the data unit cannot be added or corrected so the labels on the data unit meet your needs.

Frame and Tile Navigation

When reviewing consensus Projects with videos, image sequences, or Data Groups containing multiple tiles, the left sidebar displays frame numbers and tile names for each annotation. This helps you quickly identify where annotations are located and navigate to specific frames or tiles.
Frame numbers and tile names are displayed for annotations on videos, image sequences, DICOM series, and Data Groups with multiple tiles. Single images and image groups show only the annotation details.
Frame Navigation
  • Click on any frame number to jump directly to that frame in the corresponding tile
  • Frame ranges are displayed as individual numbers (for single frames) or ranges (for multi-frame annotations)
  • Time stamps are shown instead of frame numbers when enabled in Project settings
Tile Navigation
  • Tile names appear as badges next to frame information
  • Clicking a frame number automatically switches to the correct tile if needed
  • Each annotation shows which tile it belongs to, making it easier to review multi-tile data units
Use frame and tile navigation to quickly jump between different parts of your data unit during review. This is especially helpful when comparing annotations from different annotators across multiple frames or tiles.

Group-Level Classifications

When reviewing tasks that contain Data Groups and certain types of data units (image groups, image sequences, videos, or DICOM series), you may encounter group-level classifications. Group-level classifications are classifications that apply to the entire Data Group instead of specific frames or tiles. Group-level classifications are displayed with a Global icon to distinguish them from frame-specific or tile-specific classifications.
Group-level classifications are handled differently during consensus review. When approving a group-level classification from one annotator, any conflicting group-level classifications from other annotators for the same feature are automatically disabled to prevent conflicts.

Advanced - Allow task reassessment

This feature is available in both the ANNOTATE and REVIEW blocks. It determines whether users can act on tasks they have already annotated or reviewed as those tasks progress through the workflow. Users with the Annotator & Reviewer role can access both blocks in a Consensus.
  • In the ANNOTATE block, they cannot see other annotators’ labels.
  • In the REVIEW block, they can view all labels (including their own) because this feature is enabled by default.
Behavior by block: ANNOTATE block: When enabled, annotators can see and review the same task again later in the workflow. Default: Disabled. REVIEW block: When enabled, reviewers can review the same task multiple times as it moves through the workflow. This also allows a user to review tasks they previously annotated. Default: Enabled.

Advanced - Hide unassigned tasks

Users with the Annotator, Reviewer, or Annotator & Reviewer roles can only see tasks assigned to them. Admins and Team Managers can view all tasks, including unassigned ones.