Assignment Analytics

Prev Next

Assignment Analytics Overview

The Assignment Analytics page provides performance and progress insights for a specific assignment. It helps project managers monitor item completion, annotation activity, review status, and quality issues at the assignment level. Also, it allows you to analyze data across Progress, Performance, QA, User Stats, and Scores. The example below shows the Progress view of an assignment. You can explore similar insights in the other tabs. Learn more about Analytics here.

Why These Metrics Matter

Task analytics help you:

  • Monitor individual assignment progress

  • Identify quality or review bottlenecks

  • Analyze annotation distribution

  • Track annotator performance

  • Make allocation adjustments if needed

High-Level Metrics (Top Summary)

These metrics give a quick snapshot of the task status:

  • Total Items: The total number of items included in this assignment.

  • Raw Items (%): The percentage of items that remain untouched or unprocessed.

  • Annotations: The total number of annotations created within this assignment.


Item Status Metrics

These counters break down the current state of items in the task:

  • Completed: Number of items successfully completed by annotators.

  • Secondary Status: Displays additional status information if secondary workflows are enabled.

  • Discarded: Number of items removed from the workflow due to irrelevance or quality issues.

  • With Issues: Number of items that have reported issues associated with them.

  • Corrected: Number of items that were modified or fixed during review or QA.

  • Reviewed: Number of items that have passed the QA or review process.

Each metric includes both a count and a percentage relative to the total number of items.


Status Distribution Graph

This chart visualizes how items are distributed across different statuses, such as:

  • Complete

  • Discarded

It provides a fast way to assess overall task health and identify bottlenecks.


Annotation Metrics

Annotations by Labels

This table and bar chart show how annotations are distributed by label:

  • Label name

  • Number of uses

  • Percentage of total annotations

This helps identify:

  • Label imbalance

  • Overused or underused labels

  • Annotation trends within the task

Annotations by Tools

This view groups annotations by the annotation tools used (for example, bounding box, polygon, or classification), helping teams understand how data is being annotated.


User Filter

You can filter analytics by user to:

  • View individual annotator contributions

  • Compare performance across users

  • Identify training or quality gaps

Issues / For Review Section

This table lists items within the assignment that:

  • Have reported issues

  • Are marked for review

It includes:

  • Item thumbnail

  • Item name

  • Assignee

  • Number of issues or review flags

This section allows managers to:

  • Identify problematic items

  • Follow up with annotators

  • Track review workload