Skip to content

๐ŸŒŸ Breaking and Securing LLMs๐ŸŒŸ

Photo of Yury Michurin
Hosted By
Yury M. and Yanai E.
๐ŸŒŸ Breaking and Securing LLMs๐ŸŒŸ

Details

๐ŸŒŸ First 2025 Meetup - 3 Special lectures ๐ŸŒŸ

๐Ÿ• 18:00 - Pizza and mingling
๐Ÿ‡ฎ๐Ÿ‡ฑ 18:15 - Main event
๐ŸŽŸ๏ธ Raffles - ReactNext and NodeTLV tickes!!! ๐ŸŽŸ๏ธ

*** To get all the events that happen in Gav-Yam - register here: ***

๐Ÿ—ฃ๏ธ Ran Bar-zik
Senior Software architect @ Cyberark
Dancer | Poet | Artist | Seรฑor Senior soup maker

๐Ÿคฉ Practical Attacks on Artificial Intelligence ๐Ÿคฉ
ื”ืชืงืคื•ืช ืžืขืฉื™ื•ืช ืขืœ ื‘ื™ื ื” ืžืœืื›ื•ืชื™ืช
ื‘ืขื•ืœื ืฉื‘ื• ื‘ื™ื ื” ืžืœืื›ื•ืชื™ืช ื ื›ื ืกืช ืœื™ื•ืชืจ ื•ื™ื•ืชืจ ืžื•ืฆืจื™ื, ื™ืฉ ื’ื ื”-ืจ-ื‘-ื” ื™ื•ืชืจ ืžืชืงืคื•ืช ืืคืฉืจื™ื•ืช. ื‘ืกืฉืŸ ื”ื–ื” ืžืจืื” ืžืชืงืคื•ืช ืžื”ืขื•ืœื ื”ืืžื™ืชื™ ืฉืขื‘ื“ื• ืขืœ ืžื•ืฆืจื™ื ืืžื™ืชื™ื™ื ื•ื ืœืžื“ ืื™ืš ื”ืืงืจื™ื ืขื•ื‘ื“ื™ื ื‘ืขื™ื“ืŸ ื”ื—ื“ืฉ ืฉืœ ื” LLM

๐Ÿ—ฃ๏ธ Niv Rabin
Principal Software Architect @ Cyberark
Niv Rabin is a Principal Software Architect at CyberArk with over 15 years of experience in software development and architecture. In recent years, he has focused on AI security, specializing in LLM attack methodologies and detection techniques. His work combines hands-on research and engineering expertise to mitigate risks in AI-driven security.

๐Ÿคฉ Evolving Jailbreaks and Mitigation Strategies ๐Ÿคฉ
As large language models (LLMs) become more integrated into applications, understanding and preventing jailbreak attacks is critical. This talk explores cutting-edge techniques for bypassing LLM safeguards and the strategies to defend against them. Weโ€™ll start with semantic fuzzing, showcasing how category-based and language-disruptive paraphrasing can evolve to defeat alignment. Then, weโ€™ll delve into iterative refinement mechanisms, where multiple LLMs collaborate to create increasingly effective jailbreak prompts.
The session will also cover evaluation methods, including how to numerically distinguish compliance from rejection in LLM outputs. Finally, weโ€™ll present mitigation strategies, highlighting the strengths and limitations of model alignment, external safeguards, LLMs as judges, and hybrid defenses.
Attendees will gain practical insights into both attacking and securing LLMs, leaving equipped to build safer, more resilient AI systems.
Key Takeaways:

  1. Learn how semantic fuzzing generates prompt variations to bypass LLM defenses.
  2. Understand the role of iterative feedback loops in evolving jailbreak prompts.
  3. Discover effective methods for evaluating LLM responses numerically.
  4. Explore multi-layered mitigation strategies to prevent harmful content generation.
Photo of Negev Web Developers group
Negev Web Developers
See more events
Wix Be'er-Sheva
Torat HaYahasut St 11 ยท Be'er Sheva