What you will Learn
- Identify where evals add more value than traditional tests: Learn how to decide which parts of a coding agent should be validated with unit and integration tests, and where evals are the right tool to measure code quality, correctness, and safety.
- Design and implement practical evals for code generation: See concrete examples of rule-based and model-graded evals for a coding agent, including how they were created, what they measure, and how to balance determinism, cost, and signal quality.
- Run and interpret evals to improve agent behavior. Understand how eval results are analyzed, what a 'good' score actually means, and how findings are used to refine prompts, workflows, and constraints.
Session Details
- Intermediate
- 30 minutes
- 15mins Q&A
- Emerging Landscapes (Vibe Testing & Agentic AI)
Session Speaker

Damián Pereira
Head of Testing - Endava, Uruguay
Damián Pereira is Head of Testing at Endava, leading a team of 50 testers and driving innovation in software quality and automation. With over 15 years of experience across testing, automation engineering, and leadership, he is passionate about combining technology and testing to empower teams. He has created open-source tools such as the API Automation Agent and TestCraft extension, and is an active international speaker. Damián has presented at QA or the Highway, TestingUY, and QualitySense Conf, covering AI in testing, automation strategies, and the evolving role of quality.





