Hacking Generative AI and Language Models with AI Red Teaming and Beyond


Details
Abstract:
Large language models have revolutionized natural language processing, but their expanding capabilities also raise concerns about vulnerabilities and potential attacks. In this presentation, we embark on a journey to explore the frontiers of large language models, unveiling attack strategies and discussing effective safeguards. We showcase real-world examples of adversarial attacks, highlighting their impact on model integrity and reliability. Moreover, we delve into state-of-the-art research and best practices for fortifying models against attacks. Ethical considerations and responsible AI practices are also addressed. Join us to gain valuable insights into the evolving landscape of large language models and ensure their responsible and secure use.
Speaker:
Gaspard Baye is a PhD candidate and a security AI scientist with over 5+ years of experience developing AI-driven defensive security applications. He has been recognized with several research publications in prestigious conferences such as NeurIPS, HASP and IEEE Access, accumulating 47+ citations. He is a recognized CVE holder with certifications, including OSCP, PNPT, Scrum, NSE1, NSE2 and CEH Practical. His work has been showcased at cybersecurity conferences such as DEFCON, BSides, and The Diana Initiative. Through a dedicated vulnerability disclosure program, he's identified and helped remediate over 20+ critical security vulnerabilities, earning Hall of Fame recognitions from Nokia and Ford Motors.
COVID-19 safety measures

Every 3rd Thursday of the month
Hacking Generative AI and Language Models with AI Red Teaming and Beyond