Overview

This workshop by SallyAnn DeLucia and Fuad Ali from Arize demonstrates how to build a prompt learning optimization loop for AI agents. Prompt learning uses textual feedback and explanations to iteratively improve system prompts, going beyond traditional optimization methods that only focus on scores. The session covers why agents fail, introduces prompt learning methodology, and walks through a hands-on coding workshop implementing an optimization loop.

Key Takeaways

  • Most agent failures stem from weak environments and instructions rather than weak models - focus on improving system prompts before reaching for fine-tuning or architecture changes
  • Prompt learning leverages rich textual feedback explaining WHY outputs failed, not just binary correct/incorrect labels - use human annotations and LLM-as-judge explanations to guide prompt optimization
  • Simple rule additions to system prompts can achieve dramatic improvements - adding engineering best practices as rules improved coding agent performance by 15% with no other changes
  • Continuous optimization beats static prompts - treat prompt optimization as an ongoing process that adapts to new failure patterns over time
  • Evaluator quality determines optimization success - invest equal effort in optimizing your evaluation prompts as your agent prompts since they provide the learning signal

Topics Covered