Wed, Mar 18 · 6:00 PM CDT
Testing AI Agents: Context, Evals, and Observability 101 for Software Testers
Traditional software testing assumes determinism.
Given the same input, you expect the same output.
AI agents don’t work that way.
When systems become non-deterministic, context-aware, and probabilistic, QA must evolve. Manual prompt checking and “vibe-based validation” are not enough. Testing AI requires instrumentation, evaluation strategies, and observability at a level most teams haven’t yet formalized.
At the 4th Chicago Meetup, we move from theory to engineering discipline-exploring what a professionalized testing lifecycle looks like for AI agents.
Event Details
Date: 18th March 2026
Time: 6:00 PM – 8:00 PM CST
Venue: Association Forum, Chicago, Illinois
Session: Testing AI Agents - Context, Evals, and Observability 101 for Software Testers
Speaker: Michael Kessler – Technology Leader / Business Developer
About the Session
In this technical session, we shift from manual experimentation to structured verification.
Michael Kessler will demonstrate a live, pre-built AI agent constructed with LangChain, then walk through how to instrument it for reliability in real-world environments.
This session focuses on:
Designing evaluation frameworks for probabilistic systems
Instrumenting agents for observability and traceability
Treating orchestration logic as a testable surface
Moving beyond intuition toward measurable quality
You’ll gain a practical understanding of how to apply structured QA practices to AI-driven systems-bridging the gap between experimentation and production readiness.
Key Takeaways
Moving from “Vibes” to Verification – replacing intuition with measurable evaluation strategies
Debugging the “Black Box” with Observability – understanding agent reasoning and trace flows
The “Unit Test” for Intelligence – designing meaningful evals for non-deterministic behavior
Orchestration as a Test Surface – testing how agents coordinate tools, memory, and context
About the Speaker
Michael Kessler is a seasoned technology executive and entrepreneur with over 25 years of experience in the tech industry.
Throughout his career, he has led technology initiatives at organizations including Lucent Technologies, FactSet, and Vivid Seats, and has also built businesses in the fitness and dining sectors.
A specialist in Quality Assurance and Technology modernization, Mike is passionate about evolving traditional testing mindsets to address the challenges of AI-driven systems. He holds a degree in Computer Information Systems from DeVry University and has completed graduate-level coursework at DePaul University and MIT xPRO, focusing specifically on designing and building AI products.
Why You Should Attend
Learn how to test non-deterministic AI systems with engineering rigor
Understand evaluation strategies beyond prompt tinkering
Gain insight into observability patterns for AI agents
See a live demonstration of agent instrumentation in practice
Connect with Chicago’s AI and QA community
Be part of The Test Tribe’s 4th Chicago Meetup
About The Test Tribe
The Test Tribe is the world’s largest software testing community, empowering testers globally since 2018. With 700+ events and 150K+ members across 130+ countries, we bring testers together to learn, grow, and connect through meetups, conferences, cohorts, masterclasses, and more.
By RSVPing, you agree to our Terms and Conditions and Privacy Policy and consent to be contacted by The Test Tribe and our event partners