Concept
When autotag is enabled for a project, ML models analyze your data and produce confidence-scored predictions. These predictions appear as suggested annotations that annotators can accept, modify, or reject. Autotag works with both image and object-level predictions.Image Prefix Queries
Filter items by image-level autotag predictions using theimage_prefix: syntax:
Object Prefix Queries
Filter items by object-level autotag predictions using theobject_prefix: syntax:
Score Ranges
Autotag predictions include a confidence score ranging from -1 to 1:| Score Range | Meaning |
|---|---|
0.8 to 1.0 | High confidence — model is very certain |
0.5 to 0.8 | Medium confidence — likely correct but should be verified |
0.0 to 0.5 | Low confidence — uncertain prediction |
-1.0 to 0.0 | Negative confidence — model predicts the tag does not apply |
Items with low confidence scores are good candidates for manual review, as they represent cases where the model is uncertain.
Training Set Queries
Filter items by the training set used to generate autotag predictions:Usage in the Annotation Workflow
Enabling Autotag
- Navigate to your project in Mission Control.
- Go to Settings → Autotag.
- Select the ML model to use for predictions.
- Choose the data to run autotag on (full dataset or specific slices).
- Click Run Autotag to start the prediction job.
Reviewing Autotag Predictions
- Open the annotation editor for an item with autotag predictions.
- Suggested annotations appear with a distinct visual indicator.
- For each suggestion:
- Accept — Confirm the prediction as correct.
- Modify — Adjust the annotation (resize, relabel, etc.).
- Reject — Remove the incorrect prediction.
- Save your review to finalize the annotations.
Filtering by Autotag Status
Use the query language to find items based on their autotag review status:Best Practices
- Start with high-confidence predictions — Review items with scores above 0.8 first for quick wins.
- Use low-confidence items for model improvement — These edge cases are valuable for retraining.
- Run autotag on new data incrementally — Process new uploads as they arrive rather than waiting for large batches.
- Compare model versions — Use training set queries to evaluate whether a newer model performs better.