Comet reposted this
Most eval methods stop at single responses. But real reliability comes from evaluating whole conversations. That is why "Building Conversational AI Agents with Thread Level Eval Metrics" was one of the hands-on workshops selected by our volunteer Steering Committee for the 6th Annual MLOps World | GenAI Summit (Oct 8-9, Austin) Co-hosted by Tony Kipkemboi, Head of Developer Relations, CrewAI, and Claire Longo, Lead AI Researcher, Comet, this session will show you how to: ✅ Use CrewAI to define multi agent workflows and tool integrations ✅ Apply Comet Opik to design thread level eval metrics that capture full conversations ✅ Combine orchestration and evaluation into a repeatable workflow that improves agent quality If you're a serious AI technologist this session will help you and your bot projects reach the next level. DYK? MLOps World is the only major AI tech event programmed by practitioners, for practitioners. You’ll find hands-on workshops, technical deep dives, and real-world case studies hosted by some of the world’s most esteemed AI teams, including JFrog, TikTok, DICK'S Sporting Goods, Google DeepMind, Outerbounds, and Fujitsu. Check out the full agenda and get tickets (few still available): https://lnkd.in/gVbbyAR2