Task Analytics Overview
The Analytics tab provides visibility into task progress, annotation activity, and data quality. It helps project managers track completion, identify issues, and monitor annotation distribution. Also, it allows you to analyze data across Progress, Performance, QA, User Stats, and Scores. The example below shows the Progress view of a task. You can explore similar insights in the other tabs. Learn more about Analytics here.

Why These Metrics Matter
Task analytics help you:
Track progress in real time
Detect annotation or review bottlenecks
Monitor data quality and consistency
Make informed decisions about task allocation and QA
High-Level Metrics (Top Summary)
These metrics give a quick snapshot of the task status:
Progress (%): Shows the percentage of completed items out of the total items in the task.
Total Items: The total number of data items included in the task.
Remaining: The number of items that still require work.
Annotations: The total number of annotations created for the task so far.
Item Status Metrics
These counters break down the current state of items in the task:
Completed: Number of items successfully completed by annotators.
Secondary Status: Displays additional status information if secondary workflows are enabled (for example, rework or custom states).
Discarded: Number of items removed from the workflow due to irrelevance or quality issues.
With Issues: Number of items that have reported issues associated with them.
Corrected: Number of items that were modified or fixed during review or QA.
Reviewed: Number of items that have passed the QA or review process.
Each metric includes both a count and a percentage relative to the total number of items.
Status Distribution Graph
This chart visualizes how items are distributed across different statuses, such as:
Complete
Discarded
It provides a fast way to assess overall task health and identify bottlenecks.
Annotation Metrics
Annotations by Labels
This table and bar chart show how annotations are distributed by label:
Label name
Number of uses
Percentage of total annotations
This helps identify:
Label imbalance
Overused or underused labels
Annotation trends within the task
Annotations by Tools
This view groups annotations by the annotation tools used (for example, bounding box, polygon, or classification), helping teams understand how data is being annotated.
User Filter
You can filter analytics by user to:
View individual annotator contributions
Compare performance across users
Identify training or quality gaps
Issues / For Review Section
This table lists items within the task that:
Have reported issues
Are marked for review
It includes:
Item thumbnail
Item name
Assignee
Number of issues or review flags
This section allows managers to:
Identify problematic items
Follow up with annotators
Track review workload