vlad.build

The Zen of Claude Code

The Zen of Claude Code: How Simplicity Beat Complexity in AI Agents

It's amazing how quickly the world of AI agents has changed, especially in the last couple of years. My talk, "The Zen of Claude Code," dives into the evolution of AI agents from 2023 to 2025.

The main takeaway: Simplicity beats complexity. One strong model in a local-first think → tool → read → repeat loop outperforms vector stacks and multi-model orchestration. This is yet another example of Sutton's Bitter Lesson.

The Problem with Over-Engineering (2023)

Back in 2023, the big excitement came with models like GPT-4. Immediately, people started trying to build truly autonomous agents. The most famous early attempt was AutoGPT.

2023_agent_architecture-9140

The architecture of AutoGPT was complex. It had a task creation agent, a prioritization agent, an execution agent, memory modules, and more.

All that complexity, though, didn't actually translate to success. It was extremely expensive to run (I remember one humorous story of it costing $1,000 and an hour just to order a pizza — see the original tweet) and it ultimately couldn't reliably get complex tasks done. It was a great experiment, but it didn't really work in the real world.

autogpt_tweets

The High-Cost Hybrid Approach: Cursor

The next evolution came with tools like Cursor.

Cursor was and is a success, proven by its multi-billion-dollar valuation (Series C announcement). But why did it work where AutoGPT failed?

As Andrej Karpathy discussed in his June 2025 talk on the ā€œautonomy sliderā€ (video), Cursor focused on human-AI collaboration instead of full autonomy.

claude_code_autonomy_slider_demo

A deep dive into Cursor’s build (see Pragmatic Engineer’s piece ā€œReal-world engineering challenges: building Cursorā€) reveals that it relies on a very complex technical stack:

The complexity provided performance, but at a very high cost in terms of engineering and computational overhead.

The Zen of Claude Code

Then came Claude Code in early 2025 (announcement with Claude 3.7 Sonnet).

claude_code_startup_screen

It quickly gained huge success and adoption, becoming a major revenue driver for Anthropic. It’s an agent that lives right in your terminal and immediately impressed developers with its ability to actually get work done.

reddit_claude_code_testimonial

I was curious to find out what "wonders of engineering" made it possible. To my surprise, I discovered that Claude Code is an extremely simple application.

This is the "Zen of Claude Code"—an approach based on picking the simplest possible solution that works. A great article by Gergely Orosz confirmed this approach, emphasizing that the Claude Code team always chooses the simplest options, writing as little business logic as possible (How Claude Code is built).

claude_code_architecture_simplicity

In more detail:

The Power of the Tool Use Loop

At the core of Claude Code and other successful modern agents is a simple loop with tool use. This is not new, but the simplicity is key. Thorsten Ball showed you can build a fully functional agent in under 400 lines (ā€œHow to Build an Agent (or: The Emperor Has No Clothes)ā€).

gemini_cli_architecture_diagram

The loop works like this:

  1. User Request: The user sends a request (e.g., ā€œImplement this featureā€).
  2. Model Reasoning: The model reads the request and context and decides the next step.
  3. Tool Call: It calls a tool like read_file or run_terminal_command.
  4. Application Response: The app runs the tool and returns the result to the AI.
  5. Iterate: The model reads the tool output and loops.

One clever example is the TodoWrite tool. It provides a lightweight task list the model can create and manage to scaffold multi-step plans.

todowrite_tool_documentation

I used this tool to link API documentation and integrate a new banking API, which it did by creating and tackling a to-do list autonomously.

claude_code_api_integration_workflow

The TodoWrite tool demonstrates how a simple tool can enable new capability (long-term planning) and enhance the model's practical intelligence.

The Bitter Lesson: Intelligence Beats Engineering

The huge shift from the complex, brittle AutoGPT to the simple, effective Claude Code is a great example of Richard Sutton’s ā€œThe Bitter Lesson.ā€

In the past, when models were dumber, engineers tried to compensate by building incredibly complex architectures to guide them. They tried to manually program the how.

Today, the models are so much smarter that the complex scaffolding holds them back. The winning approach is to give the model the simplest architecture—an LLM in a loop with tools—and let it figure out the how. The simpler system is more effective because it gets out of the model's way.

Practical Takeaways for Engineers

  1. Intelligence Beats Engineering: Don’t spend months on brittle architectures when a more capable model is available or imminent. Prefer a stronger model over more scaffolding.
  2. Get Out of the Model's Way: Reduce complexity. Provide only minimal scaffolding.
  3. Tool Use is Effective: Simple tools like TodoWrite expand practical capability.
  4. Agents are General: Claude Code was marketed as a coding assistant, but it’s a general agent (McKay Wrigley; Latent Space). With connectors, you can use it well beyond coding (Claude connectors directory).

mckay_wrigley_claude_agent_insight

This post was based on my talk, "The Zen of Claude Code: The Evolution of AI Agents from 2023 to 2025," originally presented at AI Camp Berlin in August 2025.