
What weâre about
đ This virtual group is for data scientists, machine learning engineers, and open source enthusiasts.
Every month weâll bring you diverse speakers working at the cutting edge of AI, machine learning, and computer vision.
- Are you interested in speaking at a future Meetup?
- Is your company interested in sponsoring a Meetup?
This Meetup is sponsored by Voxel51, the lead maintainers of the open source FiftyOne computer vision toolset. To learn more, visit the FiftyOne project page on GitHub.
Upcoming events (4+)
See all- Network event79 attendees from 37 groups hostingMay 21 - Advanced Computer Vision Data Curation and Model EvaluationLink visible for attendees
When and Where
May 21, 2025 | 9:00 â 10:30 AM Pacific
About the Workshop
Are you looking for simpler and more accurate ways to perform common data curation and model evaluation tasks for your computer vision workflows?Then this workshop with Harpreet Sahota is for you! In this 90 min hands-on workshop, weâll show you how to make use of FiftyOneâs panel and plugin framework to learn how to:
- Customize the FiftyOne App to work the way you want to work
- Quickly integrate FiftyOne with new models, datasets, and MLOps tools
- Automate common data curation and model evaluation tasks
- Streamline your computer vision workflows with less code and more clicks
Whether you are a beginner or advanced user of FiftyOne, looking for how to get started with customizing the dozens of existing plugins or interested in creating your own, there will be something for you in this workshop!
Prerequisites
A working knowledge of Python and basic familiarity with FiftyOne. All attendees will get access to the tutorials, videos, and code examples used in the workshop.
- Network event262 attendees from 36 groups hostingMay 22 - AI, ML and Computer Vision MeetupLink visible for attendees
When and Where
- May 22, 2025 | 10:00 AM Pacific
- Virtual - Register for the Zoom
CountGD: Multi-Modal Open-World Counting
We propose CountGD, the first open-world counting model that can count any object specified by text only, visual examples only, or both together. CountGD extends the Grounding DINO architecture and adds components to enable specifying the object with visual examples. This new capability â being able to specify the target object by multi-modalites (text and exemplars) â lead to an improvement in counting accuracy. CountGD is powering multiple products and has been applied to problems across different domains including counting large populations of penguins to monitor the influence of climate change, counting buildings from satellite images, and counting seals for conservation.
About the Speaker
Niki Amini-Naieni is a DPhil student focusing on developing foundation model capabilities for visual understanding of the open world at the Visual Geometry Group (VGG), Oxford supervised by Andrew Zisserman. In the past, Niki has consulted with Amazon and other companies in robotics and computer vision, interned at SpaceX, and studied computer science and engineering at Cornell.
GorillaWatch: Advancing Gorilla Re-Identification and Population Monitoring with AI
Accurate monitoring of endangered gorilla populations is critical for conservation efforts in the field, where scientists currently rely on labor-intensive manual video labeling methods. The GorillaWatch project applies visual AI to provide robust re-identification of individual gorillas and generate local population estimates in wildlife encounters.
About the Speaker
Maximilian von Klinski is a Computer Science student at the Hasso-Plattner-Institut and is currently working on the GorillaWatch project alongside seven fellow students.
This Gets Under Your Skin â The Art of Skin Type Classification
Skin analysis is deceptively hard: inconsistent portrait quality, lighting variations, and the presence of sunscreen or makeup often obscure whatâs truly âunder the skin.â In this talk, Iâll share how we built an AI pipeline for skin type classification that tackles these real-world challenges with a combination of vision models. The architecture includes image quality control, facial segmentation, and a final classifier trained on curated dermatological features.
About the Speaker
Markus Hinsche is the co-founder and CTO of Thea Care, where he builds AI-powered skincare solutions at the intersection of health, beauty, and longevity. He holds a Masterâs in Software Engineering from the Hasso Plattner Institute and brings a deep background in AI and product development.
A Spot Pattern Is like a Fingerprint: Jaguar Identification Project
The Jaguar Identification Project is a citizen science initiative actively engaging the public in conservation efforts in Porto Jofre, Brazil. This project increases awareness and provides an interesting and challenging dataset that requires the use of fine-grained visual classification algorithms. We use this rich dataset for dual purposes: teaching data-centric visual AI and directly contributing to conservation efforts for this vulnerable species.
About the Speaker
Antonio Rueda-Toicen, an AI Engineer in Berlin, has extensive experience in deploying machine learning models and has taught over 300 professionals. He is currently a Research Scientist at the Hasso Plattner Institute. Since 2019, he has organized the Berlin Computer Vision Group and taught at Berlinâs Data Science Retreat. He specializes in computer vision, cloud technologies, and machine learning. Antonio is also a certified instructor of deep learning and diffusion models in NVIDIAâs Deep Learning Institute.
- Network event133 attendees from 36 groups hostingMay 29 - Best of WACV 2025Link visible for attendees
This is a virtual event taking place on May 29, 2025 at 9 AM Pacific.
Welcome to the Best of WACV 2025 virtual series that highlights some of the groundbreaking research, insights, and innovations that defined this yearâs conference. Live streaming from the authors to you. IEEE/CVF Winter Conference on Applications of Computer Vision (WACV) is the premier international computer vision event comprising the main conference and several co-located workshops and tutorials.
DreamBlend: Advancing Personalized Fine-tuning of Text-to-Image Diffusion Models
Given a small number of images of a subject, personalized image generation techniques can fine-tune large pre-trained text-to-image diffusion models to generate images of the subject in novel contexts, conditioned on text prompts. In doing so, a trade-off is made between prompt fidelity, subject fidelity and diversity. As the pre-trained model is fine-tuned, earlier checkpoints synthesize images with low subject fidelity but high prompt fidelity and diversity. In contrast, later checkpoints generate images with low prompt fidelity and diversity but high subject fidelity. This inherent trade-off limits the prompt fidelity, subject fidelity and diversity of generated images. In this work, we propose DreamBlend to combine the prompt fidelity from earlier checkpoints and the subject fidelity from later checkpoints during inference. We perform a cross attention guided image synthesis from a later checkpoint, guided by an image generated by an earlier checkpoint, for the same prompt. This enables generation of images with better subject fidelity, prompt fidelity and diversity on challenging prompts, outperforming state-of-the-art fine-tuning methods.
Paper: DreamBlend: Advancing Personalized Fine-tuning of Text-to-Image Diffusion Models
About the Speaker
Shwetha Ram is an Applied Scientist at Amazon, where she focuses on advancing multimodal capabilities for Rufus, Amazonâs generative AI-powered conversational shopping assistant. Her work has contributed to a range of innovative initiatives across Amazon, including Lab126, Scout (the autonomous sidewalk delivery robot), and M5 (Amazonâs foundation models). Prior to joining Amazon, Shwetha was part of the Image Technology Incubation team at Dolby Laboratories, where she explored emerging opportunities for Dolby in AR/ VR and immersive media technologies.
Robust Multi-Class Anomaly Detection under Domain Shift
Robust multi-class anomaly detection under domain shift is a fundamental challenge in real-world scenarios, where detectors should distinguish different types of anomalies despite significant distribution shifts. Traditional approaches often struggle to generalize across domains and handle inter-class interference. ROADS addresses these limitations through a prompt-driven framework that combines a hierarchical class-aware prompt mechanism with a domain adapter to jointly encode discriminative, class-specific prompts and learn domain-invariant representations. Extensive evaluations on the MVTec-AD and VISA datasets show that ROADS achieves superior performance in both anomaly detection and localization, particularly in out-of-distribution settings.
Paper: ROADS: Robust Prompt-driven Multi-Class Anomaly Detection under Domain Shift
About the Speaker
Hossein Kashiani is a fourth-year Ph.D. student at Clemson University. His research focuses on developing generalizable and trustworthy AI systems, with publications in top venues such as CVPR (2025), WACV (2025), ICIP, and TBIOM. His work spans diverse applications, including anomaly detection, media forensics, biometrics, healthcare, and visual perception.
What Remains Unsolved in Computer Vision? Rethinking the Boundaries of State-of-the-Art
Despite rapid progress and increasingly powerful models, computer vision still struggles with a range of foundational challenges. This talk revisits the âblind spotsâ of state-of-the-art vision systems, focusing on problems that remain difficult in real-world applications. I will share insights from recent work on multi-object trackingâspecifically cases involving prolonged occlusions, identity switches, and visually indistinguishable subjects such as identical triplets in motion. Through examples from DragonTrack and other mehtods, Iâll explore why these problems persist and what they reveal about the current limits of our models. Ultimately, this talk invites us to look beyond benchmark scores and rethink how we define progress in visual perception.
About the Speaker
Bishoy Galoaa is an incoming PhD student in Electrical and Computer Engineering at Northeastern University, under the supervision of Prof. Sarah Ostadabbas. His research centers on multi-object tracking and scene understanding in complex environments, with a focus on problems that challenge the assumptions of current deep learning models.
LLAVIDAL: A Large LAnguage VIsion Model for Daily Activities of Living
Current Large Language Vision Models trained on web videos perform well in general video understanding but struggle with fine-grained details, complex human-object interactions (HOI), and view-invariant representation learning essential for Activities of Daily Living (ADL). In this talk, I will introduce a foundation model: LLAVIDAL catered towards understanding ADL and the tricks to train such models.
Paper: LLAVIDAL: A Large LAnguage VIsion Model for Daily Activities of Living
About the Speaker
Srijan Das is an Assistant Professor in the Department of Computer Science at the University of North Carolina at Charlotte. At UNC Charlotte, he is working on Video Representation Learning, and Robotic Vision. He is a member of the AI4Health Center and one of the founding members of the Charlotte Machine Learning Lab (CharMLab) at UNC Charlotte.
- Network event117 attendees from 36 groups hostingMay 30 - Best of WACV 2025Link visible for attendees
This is a virtual event taking place on May 29, 2025 at 9 AM Pacific.
Welcome to the Best of WACV 2025 virtual series that highlights some of the groundbreaking research, insights, and innovations that defined this yearâs conference. Live streaming from the authors to you. IEEE/CVF Winter Conference on Applications of Computer Vision (WACV) is the premier international computer vision event comprising the main conference and several co-located workshops and tutorials.
Iris Recognition for Infants
Non-invasive, efficient, physical token-less, accurate, and stable identification methods for newborns may prevent baby swapping at birth, limit baby abductions, and improve post-natal health monitoring across geographies, within both formal (e.g., hospitals) and informal (e.g., humanitarian and fragile settings) health sectors. This talk explores the feasibility of applying iris recognition as a biometric identifier for 4-6 week old infants.
About the Speaker
Rasel Ahmed Bhuiyan is a fourth-year PhD student at the University of Notre Dame, supervised by Adam Czajka. His research focuses on iris recognition at life extremes, specifically infants and post-mortem cases.
Advancing Autonomous Simulation with Generative AI
Autonomous vehicle (AV) technology, including self-driving systems, is rapidly advancing but is hindered by the limited availability of diverse and realistic driving data. Traditional data collection methods, which deploy sensor-equipped vehicles to capture real-world scenarios, are costly, time-consuming, and risk-prone, especially for rare but critical edge cases.
We introduce the Autonomous Temporal Diffusion Model (AutoTDM), a foundation model that generates realistic, physics-consistent driving videos. By leveraging natural language prompts and integrating semantic sensory data inputs like depth maps, edge detection, segmentation maps, and camera positions, AutoTDM produces high-quality, consistent driving scenes that are controllable and adaptable to various simulation needs. This capability is crucial for developing robust autonomous navigation systems, as it allows for the simulation of long-duration driving scenarios under diverse conditions.
AutoTDM offers a scalable, cost-effective solution for training and validating autonomous systems, enhancing safety and accelerating industry advancements by simulating comprehensive driving scenarios in a controlled virtual environment, which marks a significant leap forward in autonomous vehicle development.
About the Speaker
Xiangyu Bai is a second-year PhD candidate at ACLab, Northeastern University, specializing in generative AI and computer vision, with a focus on autonomous simulation. His research centers on developing innovative, physics-aware generative vision frameworks that enhance simulation systems to provide realistic, scalable solutions for autonomous navigation. He has authored six papers in top-tier conferences and journals, including three as first author, highlighting his significant contributions to the field.
Classification of Infant SleepâWake States from Natural Overnight In-Crib Sleep Videosâ
Infant sleep plays a vital role in brain development, but conventional monitoring techniques are often intrusive or require extensive manual annotation, limiting their practicality. To address this, we develop a deep learning model that classifies infant sleepâwake states from 90-second video segments using a two-stream spatiotemporal architecture that fuses RGB frames with optical flow features. The model achieves over 80% precision and recall on clips dominated by a single state and demonstrates robust performance on more heterogeneous clips, supporting future applications in sleep segmentation and sleep quality assessment from full overnight recordings.
About the Speaker
Shayda Moezzi is pursuing a PhD in Computer Engineering at Northeastern University in the Augmented Cognition Lab, under the guidance of Professor Sarah Ostadabbas. Her current research focuses on computer vision techniques for video segmentation.
Leveraging Vision Language Models for Specialized Agricultural Tasks
Traditional plant stress phenotyping requires experts to annotate thousands of samples per task â a resource-intensive process limiting agricultural applications. We demonstrate that state-of-the-art Vision Language Models (VLMs) can achieve F1 scores of 73.37% across 12 diverse plant stress tasks using just a handful of annotated examples.
This work establishes how general-purpose VLMs with strategic few-shot learning can dramatically reduce annotation burden while maintaining accuracy, transforming specialized agricultural visual tasks.About the Speaker
Muhammad Arbab Arshad is a Ph.D. candidate in Computer Science at Iowa State University, affiliated with AIIRA. His research focuses on Generative AI and Large Language Models, developing methodologies to leverage state-of-the-art AI models with limited annotated data for specialized tasks.
Past events (33)
See all- Network event203 attendees from 36 groups hostingMay 20 - Image Generation: Diffusion Models & U-Net WorkshopThis event has passed