Export Formats
Avala supports exporting annotations in standard ML formats. Choose the format that matches your training framework.| Format | Annotation Types | Frameworks |
|---|---|---|
| Avala JSON | All types | Custom pipelines, Avala SDK |
| COCO | Bounding boxes, polygons, keypoints, segmentation | Detectron2, MMDetection, PyTorch |
| YOLO | Bounding boxes | Ultralytics, YOLOv5/v8 |
| Pascal VOC | Bounding boxes | TensorFlow, older pipelines |
| Segmentation masks | Semantic/instance segmentation | Any framework (PNG masks) |
| KITTI | 3D cuboids, 2D boxes | Autonomous driving pipelines |
Creating an Export
Export Filtering
Control exactly which annotations are included in your export:| Filter | Description |
|---|---|
include_approved_only | Only include annotations that passed QC review |
dataset_uids | Limit export to specific datasets within the project |
slice_uids | Export only items in specific slices |
label_filter | Include only specific object classes |
PyTorch Integration
Loading COCO Exports with torchvision
Loading with Detectron2
Hugging Face Integration
Loading with Hugging Face Datasets
Training Loop Automation
End-to-End Pipeline
Combine Avala exports with webhooks to trigger training automatically when new annotations are approved.Webhook-Triggered Training
Scheduling Periodic Exports
For pipelines that do not need real-time triggers, schedule periodic exports:Active Learning Loop
Use model predictions to prioritize which data gets annotated next, creating a feedback loop between your model and your annotation team.How It Works
- Train an initial model on a small labeled dataset
- Run inference on unlabeled data
- Score uncertainty — identify items where the model is least confident
- Import predictions into Avala as pre-annotations
- Prioritize uncertain items for human annotation using work batches
- Annotators review and correct the model predictions (faster than labeling from scratch)
- Export the corrected annotations and retrain
Importing Model Predictions
Use batch auto-labeling to import model predictions as pre-annotations:Measuring Improvement
Track these metrics across active learning iterations:| Metric | Description | Goal |
|---|---|---|
| Model mAP | Mean average precision on held-out test set | Increasing each iteration |
| Annotation time per item | Average time annotators spend per item | Decreasing (pre-annotations save time) |
| Correction rate | % of pre-annotations that need human correction | Decreasing each iteration |
| Items labeled per iteration | Number of new items added to training set | Depends on budget |
Dataset Versioning
Keep track of which data was used to train which model.Using Slices for Versioning
Slices let you create named subsets of a dataset without duplicating data:Versioning Best Practices
| Practice | Benefit |
|---|---|
| Create a new slice for each training run | Reproducible experiments |
| Include the model version in the slice name | Easy cross-reference |
Export with include_approved_only=True | Only train on reviewed data |
| Keep a held-out test slice | Consistent evaluation across versions |