DeepLearning.AI’s Post

At AI Dev 25 x NYC, Hatice Ozen, Head of Developer Relations at Groq, showcased how to build a deep research agent with one API call. AI agents today require complex orchestration: managing state, routing tools, handling retries, and coordinating multiple LLM calls, all while fighting latency at every step. In this hands-on workshop, Hatice guided the audience to built a sophisticated deep research agent using Groq's compound system: web search, code execution, and multi-step reasoning in a single API call with zero orchestration code. Attendees saw how instant inference enables intelligent server-side orchestration and walk away knowing when to use direct APIs vs. frameworks. Watch the full session: https://lnkd.in/efhVsrEJ

The 'thinking' pause is a huge pain point for user experience, specifically in live chatbots where expectations for speed are high. We can try to mask the latency with streaming or animations, but when the 'band-aid' eventually falls off - users notice. Reducing the orchestration overhead is the only way to truly solve it. Excited to watch the session.

To view or add a comment, sign in

Explore content categories