Skip to content

Details

Network, learn, ask a question, meet other folks - these are all things that happen at user group events. These events are a really great opportunity to socialise in an informal learning experience.

Remember to tell your friends and the people you work with; make sure you register as soon as you can.

In-Person only event. Not being recorded.

Please complete the registration form with your full name and organisation you are from - we only collect this information to give to building security to let you into the event.

17:45 – 18:00 Intro and updates
18:00 – 19:00 Shubhangi Goyal - Evaluation of LLMs using prompt engineering.
n this session, I will discuss how prompt engineering can be leveraged to evaluate the performance, reliability, and limitations of Large Language Models (LLMs). By systematically varying prompts, we uncover model behavior across various tasks, including reasoning, summarization, and code generation. We’ll discuss prompt-based benchmarking methods, prompt sensitivity, and how to design effective evaluation frameworks.
Takeaways:
- Understand the role of prompt design in LLM evaluation.
- Learn techniques to test reasoning, bias, and robustness via prompts.
- Explore limitations of prompt-based metrics and how to mitigate them.
- Gain practical examples of evaluation setups across NLP tasks.

19:00 – 19:15 Break & Pizza
19:15 – 20:15 TBC

Related topics

Events in Manchester, GB
Data Analytics
Advanced SQL Server
SQL Server
SQL
Microsoft

You may also like