Detection Transformer (DETR) for Object Detection
Details
Object detection is still one of the highest-ROI models in applied ML, powering quality inspection, inventory systems, safety monitoring, medical imaging, and autonomous systems. Fine-tuned detectors consistently outperform general-purpose vision models on domain-specific tasks, and they do it with a fraction of the compute, latency, and cost.
RSVP on Luma to get access to a Flyte cluster for the workshop: https://luma.com/jghz6s3v
In this hands-on workshop, we'll fine-tune a DETR (DEtection TRansformer) model on a custom dataset and deploy it behind a simple UI. By the end, you'll have a working detector, a walkthrough of how to build your own datasets, and a reusable pipeline you can point at your next detection problem. A full end-to-end computer vision workflow built on infrastructure that scales from your laptop to a production cluster.
What we'll cover
- A practical intro to DETR and why transformers changed object detection
- How to build your own dataset: collection, labeling workflows, and common pitfalls
- Fine-tuning DETR with Hugging Face Transformers
- Orchestrating the pipeline with Flyte 2: cached data prep, GPU-aware training, and reproducible runs
- Deploying the model with aUI, with a path to scaled inference
- Patterns for extending to your own detection problem
What you'll leave with
- A fine-tuned DETR model trained on a custom dataset
- A reusable training and deployment pipeline you can adapt to your own data
- The knowledge to build and label your own datasets for future projects
- A portfolio-ready project and a certificate of participation
Who it's for
ML engineers and practitioners who want to move past pretrained demos and train detectors on their own data. Whether you're prototyping at work, evaluating infrastructure for a production CV use case, or building a portfolio project, you'll leave with code you can keep extending.
Hosted by Sage Elliott, AI Engineer at Union.ai.
