Define the rubric
The label schema, the guidance, and the examples are versioned. Every annotator works against the same definition.
→Labeling alone does not tell you whether a release should ship. Annotation keeps the label, the reviewer note, and the verdict together.
The label schema is reviewable and the same on every batch.
Reviewer notes follow the label out of the workspace.
IAA, reopen rate, and verdict sealed with the export.
Define the rubric. Label with review. Sign the dataset before it leaves.
The label schema, the guidance, and the examples are versioned. Every annotator works against the same definition.
→Annotation and review run in the same workspace. IAA, reopen rate, and reviewer notes stay attached to the batch.
→Export only when the gate clears. Rubric version, reviewer record, and verdict travel with the dataset.
Purpose-built tools for each modality, on the same quality path. Masks, timelines, waveforms, cuboids, and spans all attach to the same review record.
Bounding boxes, polygons, and pixel-mask segmentation with SAM2 assist. Every stroke and every correction stays on the same review record.
Frame-by-frame tracking with timelines and action-recognition segments that reviewers can scrub together. Spans survive review.
Voice comments on segments, diarization, transcripts, and biosignal overlays kept together. The reviewer's voice note lives on the row that earned it.
Named entity recognition, sentiment, classification, span annotation, and multilingual prompts in one governed workflow.
Hierarchical taxonomies, metadata validation, and bulk edits for tables and schema-driven labels. One source of truth for the schema.
LiDAR and depth-sensor annotation with cuboids, measurement tools, and calibration evidence that travels with the sensor frame.
Every number below is a snapshot from the platform's live telemetry pipeline — not a designed claim. The same numbers move the rubric, the queue, and the export gate.
Target 0.75 · trend up
Ceiling 2.0% · trend down
3 blocked · 1 in review
6 auto-escalated · 38m median wait
The Studio is where the work happens. Review is where it gets cleared. The Quality Hub is where it gets measured. Datasets are where it leaves.
Teams work across image, video, text, audio, and LiDAR without relearning the workflow. Autosave and offline queueing keep work intact when a connection drops.
Disagreements turn into clear decisions with shared context. Voice and video comments stay attached to the task, not scattered in side channels.
IAA, annotator drift, gold-set gaps, and the export gate live in one hub before data leaves the workflow. Threshold rules auto-escalate when metrics drift.
Projects, datasets, queue routing, and exports stay tied to the same production workflow. Dataset management and export flows share one source of truth.
Every batch shows where the rubric was hit, where review caught the disagreement, and where policy held the line. One readout, three readings.
Every batch leaves something the training team can act on — and something the next reviewer can read.
The label schema and guidance the batch was annotated against, pinned and immutable.
Annotations tied to the rubric version, the annotator, and the time they were made.
The disagreement, the reasoning, and the resolution stay with the row that triggered them.
Export only after the gate clears. Rubric version and verdict travel inside the package.
Inter-annotator agreement, reopen rate, and pass rate for every batch — readable, not buried.
Seven shipped export writers. Every format carries the manifest with it — rubric version, reviewer coverage, gate state, and the checksum on the dataset.
Instance + keypoint + panoptic JSON for image and video workflows.
Bounding-box TXT files per image, class-index manifest included.
Polygon + mask coordinates for YOLO segmentation training.
Pascal VOC XML per image for long-tail legacy pipelines.
LabelMe JSON with shapes, groups, and flags preserved.
JSON Lines for streaming and incremental dataset updates.
Columnar dataset with schema enforced by the taxonomy editor.
mask · span · waveform · cuboid
final label + reason retained
reopen + drift checks pass
“Adjudication moved from spreadsheets to one review record. IAA is visible before export, and the quality lead can reopen a case without losing context.”
Six things change as soon as the workflow runs end-to-end. None of them are about labelling speed — they are about the record the labels leave behind.
Disputed labels go through adjudication instead of spreadsheets and side threads. Every disagreement turns into a row with a resolution.
Inter-annotator agreement is not a quarterly report. It runs against every batch, with reviewer-level and project-level views.
Reopen rate per annotator and per project with SLA alerts. A drift in either one fires before the dataset leaves.
Four checks block a release when the dataset is weak. Gate state and the evidence live on the job — not on a dashboard nobody reads.
Complete audit trail for every label decision. Who labelled it, who reviewed it, what changed, and why the cleared dataset was allowed to leave.
Autosave and CRDT replay keep work intact when the connection drops. Robotics sessions in the field arrive attached to the same clip record.
IndexedDB plus CRDT replay keeps the workspace running when the network goes away. Mask edits, span boundaries, and reviewer notes queue locally and reconcile on reconnect.
Robotics teams labelling in the field, clinical reviewers in a basement reading room, and audio teams in a studio booth — none of them lose work. The conflict resolver merges with the reason on the record.
Autosave committed
29 mask edits queued locally
Two reviewers edited the same span
CRDT merge accepted · ops reconciled
Six questions the program lead asks before they sign the export. The same six come up across labs, programs, and regulated workflows.
Rubrics are versioned. A change creates a new version, and the program lead approves the move. Existing batches stay on the version they were labelled against — the export carries the version with it.
Adjudication is a defined role with its own queue. Disagreements flow into a single record with both labels, both reasons, and the resolver's call. The decision lives with the row.
Per batch and rolling 30-day, per annotator and per project. The Quality Hub surfaces the trend, the drift, and the threshold against the gate. Reopens and IAA share one source of truth.
Four checks. IAA below the threshold. Reopen rate above the ceiling. Coverage below the required percentage. Gold-set agreement below the floor. Any one of them blocks the release with the reason attached.
Through the export manifest. Seven formats are shipped. The manifest carries the rubric version, reviewer coverage, gate state, and the dataset checksum so the training side can verify what it received.
Image, video, audio, text, structured, 3D point cloud, and biosignal are all on the same workflow. New modalities ship behind the same review, quality, and export gates as the existing ones.
Annotation breaks when the label, the disagreement, the quality reading, and the export gate live in different systems. The labels arrive. The disputes get emailed. The quality report comes out a week later. And the dataset leaves before anyone has actually looked at it.
On one record, the rubric version is pinned to the batch. The disagreement is a row in adjudication, not a thread in chat. The IAA reads against the gate before the export is even staged. And the dataset only leaves when the gate has cleared — with the reason, the reviewer, and the version still attached.
Test the run. Review the hard cases. Recruit the right specialist. Remember the misses. Approve what's right.
Specialists routed to the rows the rubric flagged.
See the page →Reads the rubric, the review, and the dataset alongside you.
See the page →The same rubric that grades a release grades the dataset.
See the page →Bring the rubric your reviewers already trust. We'll keep it attached to every label, every review, and every dataset that leaves.