Or as Elon said “Anthropic is now officially more open than OpenAI“. On this fine April Fools’ Day, the joke isn’t that AI is replacing developers. The joke is that the playbook for doing it just… slipped onto the internet.
Anthropic didn’t intend to publish a step-by-step manual for building AI coding agents.
But through a mix of repos, prompts, and system design breadcrumbs, they effectively did exactly that.
The TL;DR or Key Takeaways from Claude Code’s Source:
- Prompts in source code: Surprisingly, much of Claude’s system prompting lives directly in the codebase — not assembled server-side as expected for valuable IP.
- Supply chain risk: It uses axios (recently hacked), a reminder that closed-source tools are still vulnerable to dependency attacks.
- LLM-friendly comments: The code has excellent, detailed comments clearly written for LLMs to understand context — a smart practice beyond just AGENTS.md files.
- Fewer tools = better performance: Claude Code keeps it lean with under 20 tools for normal coding tasks.
- Bash Tool is king: The Bash tool stands out, with heavy deterministic parsing to understand and handle different command types.
- Tech stack: Entirely TypeScript/React with explicit Bun bindings.
- Not open source: The source is “available” but still proprietary. Do not copy, redistribute, or reuse their prompts — that violates the license.
Overall impression:
- It’s a very well-organized codebase designed for agents to work on effectively.
- Human engineering is visible, though some parts (like messy prompt assembly) feel surprisingly low-level for Anthropic.
- The fact that core prompts ship in the CLI tool itself is the biggest surprise.
Let’s take a step back… It is all started with this:
And if you’re paying attention, this is one of those rare moments where the industry quietly shifts under your feet.
The Real Insight: It’s Not About the Model
Everyone is still arguing about models:
- GPT vs Claude
- Context window sizes
- Benchmarks nobody understands
Meanwhile, Anthropic basically said:
“Yeah, the model matters… but orchestration matters more.”
What they exposed (intentionally or not) is that AI coding agents are just well-structured loops + tools + guardrails.
We’ve already seen a version of this idea in Andrej Karpathy’s autoresearch project, where an agent runs in loops.
Training experiments, keeps the winners and discards the losers.
The Blueprint (Decoded)
Let’s strip it down to what actually matters.
1. The Agent Loop
At the core:
while not done: think() act() observe()
This is everywhere in their examples.
- The model plans
- The system executes tools
- The model reflects and iterates
It’s less “magic AI” and more “LLM wrapped in a control system.”
2. Tool Use is the Whole Game
Anthropic leans heavily on tools:
- File system access
- Code execution
- Search / retrieval
- Git operations
This turns the model from “smart autocomplete” into:
“A junior engineer with terminal access and zero fear.”
Example patterns from their cookbook:
- Tool calling via structured JSON
- Explicit tool descriptions
- Controlled execution layer
Repo: https://github.com/anthropics/anthropic-cookbook
3. Prompt Engineering… But Actually Engineering
This is where it gets spicy.
Their prompts aren’t cute. They’re operational.
They define:
- Role: “You are an expert software engineer…”
- Constraints: “Do not hallucinate file paths…”
- Workflow: “First analyze, then propose, then implement…”
In other words:
Prompts are no longer prompts. They’re runtime policies.
4. Memory is Cheap, Structure is Not
Instead of infinite context dumping, they:
- Keep tight working memory
- Use external storage (files, logs, state)
- Re-inject only what’s needed
Translation:
Stop shoving your entire repo into the context window and hoping for the best.
5. Guardrails > Intelligence
The system is full of:
- Validation steps
- Output checks
- Retry loops
- Human-in-the-loop options
Because—shocking—LLMs still do dumb things.
The takeaway:
Reliability doesn’t come from a smarter model. It comes from a stricter system.
Why This Actually Matters
This “leak” kills a myth:
You don’t need secret sauce to build AI agents.
You need:
- A loop
- Tools
- Good prompts
- Basic discipline
That’s it.
Which means:
The barrier to entry just collapsed.
Every (decent) engineering team can now build:
- Internal copilots
- Code migration agents
- Debugging assistants
- PR reviewers that don’t complain about your variable names
But…
Let’s be real.
Even with the blueprint, most teams will:
- Over-index on the model
Congrats, you upgraded from GPT-4 to Claude and nothing changed. - Under-invest in tooling
Your agent can “think” but can’t actually do anything. - Skip guardrails
Enjoy your AI deleting production files with confidence. - Ignore UX
If engineers don’t trust it, they won’t use it.
The Non-Obvious Opportunity
Here’s the interesting angle:
The real moat is not the agent.
It’s the environment around it.
Think:
- Company-specific workflows
- Internal APIs
- Codebase conventions
- Historical context
The agent is just the interface/tool.
The value is everything it plugs into.
A Practical Stack
Here’s a working setup:
- LLM: Claude / GPT
- Agent loop: simple Python / Go / TypeScript orchestrator
- Tool layer:
- file system
- shell
- git
- State:
- local files
- lightweight DB
- Guardrails:
- JSON schema validation
- execution sandbox
- Interface:
- CLI first (don’t overbuild UI)
Start ugly.
Iterate fast.
Be happy.
Discover more from Ido Green
Subscribe to get the latest posts sent to your email.
