Skip to content

About us

🖖 This virtual group is for data scientists, machine learning engineers, and open source enthusiasts who want to expand their knowledge of computer vision and complementary technologies. Every month we’ll bring you two diverse speakers working at the cutting edge of computer vision.

  • Are you interested in speaking at a future Meetup?
  • Is your company interested in sponsoring a Meetup?

Contact the Meetup organizers!

This Meetup is sponsored by Voxel51, the lead maintainers of the open source FiftyOne computer vision toolset. To learn more about FiftyOne, visit the project page on GitHub: https://github.com/voxel51/fiftyone

📣 Past Speakers

* Sage Elliott at Union.ai
* Michael Wornow at Microsoft
* Argo Saakyan at Veryfi
* Justin Trugman at Softwaretesting.ai
* Johannes Flotzinger at Universität der Bundeswehr Mßnchen
* Harpreet Sahota at Deci,ai
* Nora Gourmelon at Friedrich-Alexander-Universität Erlangen-Nßrnberg
* Reid Pryzant at Microsoft
* David Mezzetti at NeuML
* Chaitanya Mitash at Amazon Robotics
* Fan Wang at Amazon Robotics
* Mani Nambi at Amazon Robotics
* Joy Timmermans at Secury360
* Eduardo Alvarez at Intel
* Minye Wu at KU Leuven
* Jizhizi Li at University of Sydney
* Raz Petel at SightX
* Karttikeya Mangalam at UC Berkeley
* Dolev Ofri-Amar at Weizmann Institute of Science
* Roushanak Rahmat, PhD
* Folefac Martins
* Zhixi Cai at Monash University
* Filip Haltmayer at Zilliz
* Stephanie Fu at MIT
* Shobhita Sundaram at MIT
* Netanel Tamir at Weizmann Institute of Science
* Glenn Jocher at Ultralytics
* Michal Geyer at Weizmann Institute of Science
* Narek Tumanya at Weizmann Institute of Science
* Jerome Pasquero at Sama
* Eric Zimmermann at Sama
* Victor Anton at Wildlife.ai
* Shashwat Srivastava at Opendoor
* Eugene Khvedchenia at Deci.ai
* Hila Chefer at Tel-Aviv University
* Zhuo Wu at Intel
* Chuan Guo at University of Alberta
* Dhruv Batra Meta & Georgia Tech
* Benjamin Lahner at MIT
* Jiajing Chen at Syracuse University
* Soumik Rakshit at Weights & Biases
* Jiajing Chen at Syracuse University
* Paula Ramos, PhD at Intel
* Vishal Rajput at Skybase
* Cameron Wolfe at Alegion/Rice University
* Julien Simon at Hugging Face
* Kris Kitani at Carnegie Mellon University
* Anna Kogan at OpenCV.ai
* Kacper Łukawski at Qdrant
* Sri Anumakonda
* Tarik Hammadou at NVIDIA
* Zain Hasan at Weaviate
* Jai Chopra at LanceDB
* Sven Dickinson at University of Toronto & Samsung
* Nalini Singh at MIT

📚 Resources

* YouTube Playlist of previous Meetups
* Recap blogs including Q&A and speaker resource links

Sponsors

Voxel51

Voxel51

Administration, promotion, giveaways and charitable contributions.

Voxel51

Voxel51

Administration, promotion, giveaways and charitable contributions.

Upcoming events

11

See all
  • April 3 - Video Understanding AI Hackathon at Northeastern University

    April 3 - Video Understanding AI Hackathon at Northeastern University

    Northeastern University, Raytheon Amphitheater (Egan 240), 120 Forsyth St, Boston, MA, US

    Join our in-person AI Hackathon at Northeastern University on April 3 inspired by the CVPR 2026 CV4Smalls workshop and challenge.

    Pre-registration is mandatory as seats are limited

    The event will focus on video understanding using the FiftyOne open-source ecosystem and TwelveLabs’ models and APIs.

    Date, Time and Location

    • April 3
    • 9 AM - 4:30 PM - Raytheon Auditorium
    • 5-7 PM - Second Floor Suites (demos, prizes)

    Northeastern University
    360 Huntington Ave, 240 Egan
    Boston, MA

    Additional details, schedule of events, pre-requisites checklist and more can be found here.

    • Photo of the user
    • Photo of the user
    • Photo of the user
    4 attendees
  • Network event
    April 8 - Getting Started with FiftyOne

    April 8 - Getting Started with FiftyOne

    ¡
    Online
    Online
    16 attendees from 16 groups

    This workshop provides a technical foundation for managing large scale computer vision datasets. You will learn to curate, visualize, and evaluate models using the open source FiftyOne app.

    Date, Time and Location

    Apr 8, 2026
    10 AM PST - 11 AM Pacific
    Online. Register for the Zoom!

    The session covers data ingestion, embedding visualization, and model failure analysis. You will build workflows to identify dataset bias, find annotation errors, and select informative samples for training. Attendees leave with a framework for data centric AI for research and production pipelines, prioritizing data quality over pure model iteration.

    What you'll learn

    • Structure unstructured data. Map data and metadata into a queryable schema for images, videos, and point clouds.
    • Query datasets with the FiftyOne SDK. Create complex views based on model predictions, labels, and custom tags. Use the FiftyOne to filter data based on logical conditions and confidence scores.
    • Visualize high dimensional embeddings. Project features into lower dimensions to find clusters of similar samples. Identify data gaps and outliers using FiftyOne Brain.
    • Automate data curation. Implement algorithmic measures to select diverse subsets for training. Reduce labeling costs by prioritizing high entropy samples.
    • Debug model performance. Run evaluation routines to generate confusion matrices and precision recall curves. Visualize false positives and false negatives directly in the App to understand model failures.
    • Customize FiftyOne. Build custom dashboards and interactive panels. Create specialized views for domain specific tasks.

    Prerequisites:

    • Working knowledge of Python and machine learning and/or computer vision fundamentals.
    • All attendees will get access to the tutorials and code examples used in the workshop.
    • Photo of the user
    2 attendees from this group
  • Network event
    April 9 - Workshop: Build a Visual Agent that can Navigate GUIs like Humans

    April 9 - Workshop: Build a Visual Agent that can Navigate GUIs like Humans

    ¡
    Online
    Online
    36 attendees from 16 groups

    This hands-on workshop provides a comprehensive introduction to building and evaluating visual agents for GUI automation using modern tools and techniques.

    Date, Time and Location

    April 9, 2026 at 9 AM Pacific
    Online.
    Register for the Zoom

    Visual agents that can understand and interact with graphical user interfaces represent a transformative frontier in AI automation. These systems combine computer vision, natural language understanding, and spatial reasoning to enable machines to navigate complex interfaces—from web applications to desktop software—just as humans do. However, building robust GUI agents requires careful attention to dataset curation, model evaluation, and iterative improvement workflows.

    Participants will learn how to leverage FiftyOne, an open-source toolkit for dataset curation and computer vision workflows, to build production-ready GUI agent systems.

    What You'll Learn:

    • Dataset Creation & Management: How to structure, annotate, and load GUI interaction datasets using the COCO4GUI standardized format
    • Data Exploration & Analysis: Using FiftyOne's interactive interface to visualize datasets, analyze action distributions, and understand annotation patterns
    • Multimodal Embeddings: Computing embeddings for screenshots and UI element patches to enable similarity search and retrieval
    • Model Inference: Running state-of-the-art models like Microsoft's GUI-Actor to predict interaction points from natural language instructions
    • Performance Evaluation: Measuring model accuracy using standard metrics and normalized click distance to assess localization precision
    • Failure Analysis: Investigating model failures through attention maps, error pattern analysis, and systematic debugging workflows
    • Data-Driven Improvement: Tagging samples based on error types (attention misalignment vs. localization errors) to prioritize fine-tuning efforts
    • Synthetic Data Generation: Using FiftyOne plugins to augment training data with synthetic task descriptions and variations

    About the Speaker

    Harpreet Sahota is a hacker-in-residence and machine learning engineer with a passion for deep learning and generative AI. He’s got a deep interest in RAG, Agents, and Multimodal AI.

    1 attendee from this group
  • Network event
    April 23 - Advances in AI at Johns Hopkins University

    April 23 - Advances in AI at Johns Hopkins University

    ¡
    Online
    Online
    28 attendees from 16 groups

    Join our virtual Meetup to hear talks from researchers at Johns Hopkins University on cutting-edge AI topics.

    Date, Time and Location

    Apr 23, 2026
    9AM PST
    Online.
    Register for the Zoom!

    Recent Advancements in Image Generation and Understanding

    In this talk, I will provide an overview of my research and then take a closer look at three recent works. Image generation has progressed rapidly in the past decade-evolving from Gaussian Mixture Models (GMMs) to Variational Autoencoders (VAEs), GANs, and more recently diffusion models, which have set new standards for quality. I will begin with DiffNat (TMLR’25), which draws inspiration from a simple yet powerful observation: the kurtosis concentration property of natural images. By incorporating a kurtosis concentration loss together with a perceptual guidance strategy, DiffNat can be plugged directly into existing diffusion pipelines, leading to sharper and more faithful generations across tasks such as personalization, super-resolution, and unconditional synthesis.

    Continuing the theme of improving quality under constraints, I will then discuss DuoLoRA (ICCV’25), which tackles the challenge of content–style personalization from just a few examples. DuoLoRA introduces adaptive-rank LoRA merging with cycle-consistency, allowing the model to better disentangle style from content. This not only improves personalization quality but also achieves it with 19× fewer trainable parameters, making it far more efficient than conventional merging strategies.

    Finally, I will turn to Cap2Aug (WACV’25), which directly addresses data scarcity. This approach uses captions as a bridge for semantic augmentation, applying cross-modal backtranslation (image → text → image) to generate diverse synthetic samples. By aligning real and synthetic distributions, Cap2Aug boosts both few-shot and long-tail classification performance on multiple benchmarks.

    About the Speaker

    Aniket Roy is currently a Research Scientist at NEC Labs America. He recently earned a PhD from the Computer Science department at Johns Hopkins University under the guidance of Bloomberg Distinguished Professor Prof. Rama Chellappa.

    From Representation Analysis to Data Refinement: Understanding Correlations in Deep Models

    This talk examines how deep learning models encode information beyond their intended objectives and how such dependencies influence reliability, fairness, and generalization. Representation-level analysis using mutual information–based expressivity estimation is introduced to quantify the extent to which attributes such as demographics or anatomical structural factors are implicitly captured in learned embeddings, even when they are not explicitly used for supervision. These analyses reveal hierarchical patterns of attribute encoding and highlight how correlated factors emerge across layers. Data attribution techniques are then discussed to identify influential training samples that contribute to model errors and reinforce dependencies that reduce robustness. By auditing the training data through influence estimation, harmful instances can be identified and removed to improve model behavior. Together, these components highlight a unified, data-centric perspective for analyzing and refining correlations in deep models.

    About the Speaker

    Basudha Pal is a recent PhD graduate from the Electrical and Computer Engineering Department at Johns Hopkins University. Her research lies at the intersection of computer vision and representation learning, focusing on understanding and refining correlations in deep neural network representations for biometric and medical imaging using mutual information analysis, data attribution, and generative modeling to improve robustness, fairness, and reliability in high-stakes AI systems.

    Scalable & Precise Histopathology: Next-Gen Deep Learning for Digital Histopathology

    Whole slide images (WSIs) present a unique computational challenge in digital pathology, with single images reaching gigapixel resolution, equivalent to 500+ photos stitched together. This talk presents two complementary deep learning solutions for scalable and accurate WSI analysis. First, I introduce a Task-Specific Self-Supervised Learning (TS-SSL) framework that uses spatial-channel attention to learn domain-optimized feature representations, outperforming existing foundation models across multiple cancer classification benchmarks. Second, I present CEMIL, a contextual attention-based MIL framework that leverages instructor-learner knowledge distillation to classify cancer subtypes using only a fraction of WSI patches, achieving state-of-the-art accuracy with significantly reduced computational cost. Together, these methods address critical bottlenecks in generalization and efficiency for clinical-grade computational pathology.

    About the Speaker

    Tawsifur Rahman is a Ph.D. candidate in Biomedical Engineering at Johns Hopkins University, advised by Prof. Rama Chellappa and Dr. Alex Baras, with research focused on weakly supervised and self-supervised deep learning for computational pathology. He has completed two clinical data science internships at Johnson & Johnson MedTech and has published extensively in venues including Nature Modern Pathology, Nature Digital Medicine, MIDL, and IEEE WACV, accumulating over 8,500 citations and recognition in Stanford's Top 2% Scientists ranking.

    Towards trustworthy AI under real world data challenges

    The current paradigm of training AI models relies on fundamental assumptions that the data we have is clean, properly annotated, and sufficiently diverse across domains. However, this is not always true for the real world. In practice, data is may be physically corrupt, incompletely annotated, and specific to certain domains. As me move towards large scale general purpose models like LLMs and foundation models, it is even more important to address these data challenges so that we can train trustworthy AI models even with noisy real world data. In this presentation, we discuss some methods to tackle these potential issues.

    About the Speaker

    Ayush Gupta is a Ph.D. student at the AIEM lab, Johns Hopkins University in the department of Computer Science. He is advised by Prof. Rama Chellappa and is working on problems in Computer Vision and Deep Learning. His research has two focus points - general-purpose vision language models, where he works on multimodal LLMs on tasks like VQA, Video Grounding and LLM interpretability; and on fine-grained computer vision problems, where he works on person re-identification and gait recognition.

    • Photo of the user
    1 attendee from this group

Group links

Members

752
See all