Puerto Jiménez, Costa Rica
info@gexpsoftware.com
© 2026 Marcelo Retana
Every time a new AI coding tool drops, devs jump ship. The problem isn't the tool — it's never going deep enough with any of them. Here's an honest comparison and why mastery beats tool-hopping.
I replied to a tweet last week: "Codex vs Claude Code?"
My answer: "Some devs are switching from one to the other without even mastering one. Stick to the one that gives you results."
The replies were split. Half agreed. The other half wanted a real comparison. So here's both — the comparison you asked for, and the argument for why it barely matters if you keep switching every two weeks.
Here's the cycle I see constantly:
The problem isn't Codex. It's not Claude Code. It's the habit of never going deep enough with any tool to actually get good with it.
I've watched developers switch between Cursor, Copilot, Claude Code, Codex, Aider, and Windsurf — all in the span of three months. They never configured a single one. They never built custom commands, never set up project memory, never learned the keyboard shortcuts. And then they complain that "AI coding tools are overhyped."
No. You just never stayed long enough to get past the tutorial phase.
That said — the tools are genuinely different, and the differences matter. Here's what I've found after using Claude Code daily for over a year and testing Codex extensively.
Open source (Apache 2.0). Originally built in TypeScript, fully rewritten in Rust in mid-2025. Full-screen TUI with a polished terminal interface.
Where it shines:
codex cloud lets you fire off tasks asynchronously to the cloud. Delegate work, come back later. Claude Code has nothing equivalent/review command — Dedicated code review that reads a diff and gives prioritized findings. Built-in, not a prompt hackAGENTS.md — Repository-level configuration, similar in concept to CLAUDE.mdModel: GPT-5.3-Codex (purpose-built for coding) with a full lineup of GPT-5 variants.
Closed source. Built on Node.js. Conversational pair-programming style — less "execute this task" and more "let's work through this together."
Where it shines:
/clear, /compact, session management with TASKS.md. The tools for managing long sessions are more matureModel: Claude Opus 4.6 / Sonnet 4.6 (configurable per task).
| Aspect | Codex CLI | Claude Code |
|---|---|---|
| License | Apache 2.0 (open source) | Proprietary |
| Language | Rust | Node.js |
| Cloud tasks | Yes | No |
| Code review | Built-in /review | Prompt-based |
| Subagents | Experimental multi-agent | Mature, automatic |
| Project memory | AGENTS.md | CLAUDE.md + skills + hooks |
| Sandbox | Native (3 levels) | Docker / system-level |
| Plan mode | No | Yes |
| Session resume | codex resume / codex fork | /resume with session IDs |
| Complex multi-step | Good | Stronger |
| Speed (tokens/sec) | Faster output | Slower but more deliberate |
| MCP support | Yes | Yes |
| Price | ChatGPT Plus ($20/mo) | Claude Pro ($20/mo) / Max ($100-200/mo) |
Be honest: Codex has real strengths. Cloud tasks are genuinely useful for async work. The open-source model means the community drives features fast. The Rust rewrite made it snappy. /review is something I wish Claude Code had built-in.
But for my workflow — shipping full-stack projects from architecture through deployment — Claude Code wins on the tasks that actually take time. The subagent system, the planning mode, and the consistency on complex refactors are what save me hours, not milliseconds of faster token output.
Here's what nobody talks about in these comparisons: the real productivity gains come from configuration, not capability.
Both tools are powerful enough to 10x your output. But only if you invest in learning them. That means:
I've spent hundreds of hours building my Claude Code configuration. Custom commands, 17 reusable skills, project templates, session management workflows. That investment compounds every single day.
When you switch tools, you reset that compound interest to zero.
I wrote about this in detail in How I 10x My Coding Speed with Claude Code. The speed doesn't come from Claude being faster than Codex. It comes from Claude knowing my conventions, my git identity rules, my project structures, and my coding patterns — because I took the time to teach it.
I've shipped 50+ projects with Claude Code. My configuration includes:
That's not something I rebuilt from scratch. It's hundreds of hours of iteration, refinement, and pattern detection. Every time Claude suggests a new automation based on my habits, the system gets a little better.
Switching to Codex would mean rebuilding all of that in AGENTS.md format. Different syntax, different capabilities, different mental model. For what? Faster token output?
The tool that gives you results is the one you've invested in mastering. For me, that's Claude Code. For someone else, it might be Codex. The wrong answer is switching every two weeks and mastering neither.
I'm not saying you should never switch. There are legitimate reasons:
What's not a good reason:
Benchmarks don't ship your project. Configuration and muscle memory do.
Yes, and some developers do. Codex for quick async tasks via codex cloud, Claude Code for complex multi-step work. But maintaining configuration for both tools is double the investment. Most developers are better served by going deep with one.
Token output speed is faster. But speed-to-completion on complex tasks depends more on reasoning quality, context management, and how well you've configured the tool. A well-configured Claude Code session that nails a refactor in one pass is faster than a quick Codex response that needs three corrections.
Either one will serve you well. If open source matters to you, or you're already paying for ChatGPT Plus, Codex is a natural starting point. If you want deeper configuration options, a mature subagent system, and plan mode for complex projects, Claude Code is stronger there. Pick one, commit to learning it for at least 3 months before evaluating.
Not directly. The concepts are similar — both are markdown files that give the AI context about your project — but the syntax, features, and capabilities differ. You'd need to rewrite your configuration, not copy-paste it.
Different category. Cursor and Copilot are IDE-integrated assistants — they work inside your editor. Codex and Claude Code are terminal-native agents — they work alongside your editor. Many developers use both: an IDE assistant for inline completions and a terminal agent for larger tasks.
Want to see the configuration that makes Claude Code work at this level? Check out awesome-claude-code — open source, battle-tested across 50+ projects.
Building something and want to talk through your AI workflow? Let's chat.