DeepClaude (aattaran/deepclaude) is a tiny shim — three environment variables and a ~50-line Node HTTP proxy — that points the Claude Code CLI at DeepSeek's Anthropic-compatible endpoint. The harness keeps every Claude Code feature (sub-agents, MCP, /resume, hooks, IDE integrations, settings.json) and swaps the model underneath. DeepSeek V4 Pro lists at $0.27/M input, $1.10/M output, and $0.014/M on cache hits — roughly 17× cheaper than Sonnet 4.6 on like-for-like workloads, and 90%+ savings on agent-loop sessions where the cache-hit path dominates. Hit HN #1 on May 4, 2026 with 606 points and 257 comments. The most explicit demonstration to date that the agent-loop and the model are now decoupled — the harness is one product and the brain is another.
DeepClaude is the cleanest current demonstration of the agent-loop-vs-model decoupling. The Claude Code harness already respects ANTHROPIC_BASE_URL; DeepSeek already ships an Anthropic-compatible endpoint at api.deepseek.com/anthropic. DeepClaude is the four-env-var bridge plus a small Node proxy that handles edge cases like sub-agent routing and remote-task auth.
The pitch on the repo is “same UX, 17× cheaper” — which is empirically right on cache-heavy agent-loop workloads. On benchmark, DeepSeek V4 Pro is within 0.2 points of Claude Opus 4.7 on SWE-bench Verified, beats Sonnet 4.6 on Terminal-Bench and LiveCodeBench, and trails Opus 4.7 noticeably on multi-file architectural reasoning and SWE-bench Pro.
The reason DeepClaude is the #1 HN story today isn’t the cost savings (those have existed via OpenRouter and Ollama for a while). It’s the cleanness of the swap. Three exports, one binary that doesn’t need to exist, and the entire Claude Code experience runs on a competitor’s brain at ~6% of the price. The implication for the broader market — that the harness is the product and the model is a swappable component — is the structural shift the cluster of Cursor SDK, Browserbase Skills, and OpenAI Apps SDK launches has been building toward all month.
For agent buyers, the read is: default to DeepClaude, route the architecture-heavy 20% of work to Anthropic. The harness is the constant, the brain is the variable.
export ANTHROPIC_BASE_URL=https://api.deepseek.com/anthropic
export ANTHROPIC_AUTH_TOKEN=sk-deepseek-...
export ANTHROPIC_MODEL=deepseek-v4-pro
export ANTHROPIC_DEFAULT_OPUS_MODEL=deepseek-v4-pro
export ANTHROPIC_DEFAULT_SONNET_MODEL=deepseek-v4-pro
export ANTHROPIC_DEFAULT_HAIKU_MODEL=deepseek-v4-flash
claude
The fourth-and-fifth-and-sixth env vars handle sub-agent routing — without them, sub-agents will try to call Anthropic-named models on the DeepSeek endpoint and 404.
The cleanest agent-loop cost-cutting move available in May 2026, and a structural signal that the substrate-vs-model decoupling has gone mainstream. Test it for one week on a normal workload before committing — the cost of being wrong is one Saturday afternoon and $5 of DeepSeek credits.
Open-source AI pair programming tool that works in your terminal to edit code across your entire repository.
AWS's AI-powered coding assistant that helps developers build, deploy, and optimize applications on AWS with code generation and transformation.
Open-source AI coding harness builder that makes AI coding workflows deterministic and repeatable via YAML-defined DAG workflows.