Autotag uses machine learning models to automatically generate suggested tags and labels for your data. These auto-generated tags help accelerate annotation by pre-labeling items that can then be reviewed and corrected by annotators.Documentation Index
Fetch the complete documentation index at: https://avala.ai/docs/llms.txt
Use this file to discover all available pages before exploring further.
Concept
When autotag is enabled for a project, ML models analyze your data and produce confidence-scored predictions. These predictions appear as suggested annotations that annotators can accept, modify, or reject. Autotag works with both image and object-level predictions.Image Prefix Queries
Filter items by image-level autotag predictions using theimage_prefix: syntax:
Object Prefix Queries
Filter items by object-level autotag predictions using theobject_prefix: syntax:
Score Ranges
Autotag predictions include a confidence score ranging from -1 to 1:| Score Range | Meaning |
|---|---|
0.8 to 1.0 | High confidence — model is very certain |
0.5 to 0.8 | Medium confidence — likely correct but should be verified |
0.0 to 0.5 | Low confidence — uncertain prediction |
-1.0 to 0.0 | Negative confidence — model predicts the tag does not apply |
Training Set Queries
Filter items by the training set used to generate autotag predictions:Usage in the Annotation Workflow
Enabling Autotag
- Navigate to your project in Avala.
- Go to Settings → Autotag.
- Select the ML model to use for predictions.
- Choose the data to run autotag on (full dataset or specific slices).
- Click Run Autotag to start the prediction job.
Reviewing Autotag Predictions
- Open the annotation editor for an item with autotag predictions.
- Suggested annotations appear with a distinct visual indicator.
- For each suggestion:
- Accept — Confirm the prediction as correct.
- Modify — Adjust the annotation (resize, relabel, etc.).
- Reject — Remove the incorrect prediction.
- Save your review to finalize the annotations.
Filtering by Autotag Status
Use the query language to find items based on their autotag review status:Best Practices
- Start with high-confidence predictions — Review items with scores above 0.8 first for quick wins.
- Use low-confidence items for model improvement — These edge cases are valuable for retraining.
- Run autotag on new data incrementally — Process new uploads as they arrive rather than waiting for large batches.
- Compare model versions — Use training set queries to evaluate whether a newer model performs better.