AI Engineer World's Fair 2025: Takeaways & Spec
AI Engineer World's Fair 2025 Insights

AI Engineer World's Fair 2025: Takeaways & Spec

San Francisco, June 2025

Your AI can write 10,000 lines of code per hour. So can your competitor's. The models aren't the differentiator anymore--everyone has access to the same frontier capabilities. What happens when code becomes a commodity?

At this year's AI Engineer World's Fair, we found our answer: engineering excellence = articulation excellence. The bottleneck is no longer implementation--it's imagination, specification, and the ability to articulate exactly what should exist. This isn't just about engineers. Product managers, lawmakers, architects, designers--anyone who shapes systems now faces the same truth: clear thinking beats fast typing. And the teams winning have fundamentally reimagined their development process around this principle.


Foundation: Three Core Ideas

1. SPEC (Ideal State)

Key shift: writing specifications that fully capture intent and values becomes the bottleneck, not coding.

This thinking was heavily influenced by Sean Grove's keynote "The New Code" at the AI Engineer World's Fair. Sean (from OpenAI) articulated how specifications are becoming the primary development artifact--an insight that crystallized this framework.

What it looks like:

  • Markdown files in version control
  • Google Docs / Notion for collaboration
  • GitHub Wiki for team access

2. EXPOSURE (Reality)

What customers actually experience. Code is temporary; the spec is permanent.

Key insight: As foundational models improve, the same spec generates better implementations over time--across languages, frameworks, and paradigms.

Spec becomes everything:

  • Source of truth for development
  • Marketing materials generator
  • Onboarding documentation
  • Demo and content foundation

3. TASK DELTA (Work)

Continuous loop: evaluate SPEC ↔ PRODUCT, identify gaps, decompose into actionable tasks.

Success = Educatability: Teams win by constantly asking better questions and learning from every iteration. Learn from your LLMs, don't just use them. Use the tools, don't let the tools use you.


What Matters Most: Context & Quality

First-Class Citizens:

  • SPECS (markdown, version-controlled)
  • Rules/Values (company principles)
  • Prompt Banks (curated, effective prompts)

Context Sources:

  • Customer feedback
  • Slack conversations
  • Granola transcripts
  • Sketches, mockups, screenshots
  • Bug reports and logs

Critical: Share, sync, and consolidate context. One source of truth > scattered documents.

Building Toward This Vision

Parallelization Strategy

Project-Level:

  • Multiple repo clones on different branches
  • Git worktrees for multiple workspace instances
  • Containers (reliable environments)
  • 1Password secrets management

Task-Level:

  • One agent generates code
  • One agent works on tests
  • One agent updates docs
  • Start by coordinating through shared markdown files


Tool Stack Reality

Terminal-First:

IDEs - The Good:

  • VS Code - tried and true, massive extension ecosystem
  • Zed - Rust-powered, blazing fast, native vim mode

IDEs - The Fancy (VS Code forks): Cursor, Windsurf

Extensions:

Quality Philosophy

Multi-Layer Detection:

  1. Static Analysis - lints, type checking
  2. Dynamic Testing - sandboxed unit/integration tests
  3. AI Review - LLM validation against specs
  4. Production Monitoring - automated bug/log analysis
  5. User Feedback Loops - rapid iteration based on actual usage

Bug prevention > Bug squashing

Getting Started

Foundation:

  • Pick ONE AI tool and master it - terminal-based recommended
  • Write specs before code
  • Source control mastery and parallel development setup: with git worktrees, multiple repos... or maybe even learn Jujutsu (jj)

Growth:

  • Build prompt banks and templates
  • Add memory systems (start with shared markdown)
  • Create your first autonomous workflows

Scale:

  • Context ingestion pipelines
  • Team-wide knowledge sharing
  • Container isolation for AI agents


Critical Anti-Patterns to Avoid

As teams adopt AI-driven development, new failure modes emerge:

"AI Whisperer" Trap

  • Problem: One person becomes the go-to for all AI tools
  • Solution: Mandatory team rotation on AI experiments
  • Goal: Distributed AI literacy across entire team

Documentation Debt Explosion

  • Problem: Code generation accelerates, documentation lags
  • Solution: Documentation IS the spec - write it first
  • Reality: In a spec-first world, clear written communication becomes the core skill

Review Bottleneck

  • Problem: Human review becomes the constraint when AI generates code 10x faster
  • Solution: Tiered review - AI for patterns, humans for logic

Key practices:

  • Batch similar changes for efficient review sessions
  • Focus human review on architecture decisions and business logic
  • Use AI to summarize large PRs before human review

Warning: Don't let unreviewed code accumulate--technical debt compounds faster with AI

Conference Hot Topics

The ecosystem is exploding with tools, but quality varies wildly:

WiFi Reality Check

What happens when conference WiFi crashes? Workshop presenters learned the hard way--always have fallbacks. Local models aren't just nice-to-have anymore. Ollama, LM Studio, and MLX models running on your MacBook Pro are getting surprisingly good. Don't lean 100% on cloud AI without a backup plan.

The real lesson: learn HOW these tools think, not just WHAT they output. Study the patterns, understand the reasoning, apply the methods. That knowledge stays with you when the API goes down.

The Next Frontier: Cross-Organization Federation

While we're mastering team collaboration, the next wave involves cross-organizational AI agent cooperation.

Example: A bank's compliance agent detects suspicious activity patterns. It securely federates with:

  • Customer's accounting AI to verify legitimate business expenses
  • Merchant's inventory agent to confirm supply chain movements
  • Regulatory reporting agents to file real-time compliance updates

Each agent maintains its organization's data sovereignty while sharing just enough context to prevent fraud without exposing sensitive internals. Auth tokens expire after specific workflows. Audit trails span organizations. Legal agreements are encoded in the agent handshakes.

Parallelization solves individual productivity. Collaboration solves team productivity. Federation could solve ecosystem productivity--turning entire industries into coordinated, intelligent networks.

Reality Check: What Actually Matters

Beyond the hype, three uncomfortable truths emerged:

Human Bottlenecks > Tool Limitations

  • Context switching capacity remains the biggest constraint
  • Parallelization helps but requires new mental models
  • Best tools can't fix unclear specifications

Cost Reality (June 2025)

  • API usage for heavy development: $50-200/day easily
  • Subscription models create predictability
  • Most teams see positive returns within weeks (shipping 3-5x faster outweighs tool costs)

Security Blind Spots

  • AI agents with full file system access = nuclear option
  • Container isolation isn't optional anymore
  • Auth tokens in AI context => eventual breach


Final Insight: Specification as North Star

The future belongs to teams that rally around living specifications. When documentation becomes the primary artifact and AI amplifies collective intelligence rather than individual productivity, sustainable velocity emerges.

New engineering excellence:

  • Writing specifications so clear that both humans and AI execute flawlessly
  • Creating context so rich that any team member can own any task
  • Building review systems that elevate everyone's capabilities
  • Maintaining educatability as a core team value

Your competitive advantage isn't keeping knowledge siloed--it's how fast your entire team learns, adapts, and ships quality together. Specification is your shared language. Clear communication is your superpower. Collective educatability is your moat.


TL;DR: engineering excellence = articulation excellence

Write specs so clear that implementation becomes mechanical. Use living documents that evolve. Compare spec to reality constantly. Parallelize everything. Share context obsessively. Clear thinking beats fast typing.






Darryl Nelson

Co-Founder & CTO at JAD²

3mo

Having attended the 2025 AI Engineer conference, this is the best and most insightful take I've read.

To view or add a comment, sign in

Others also viewed

Explore content categories