For those who have already tested the Codex, what do you think?
Opus 4.1 is a beast of a coding model, but I'd suggest to any Claude Max user to at least try Codex CLI for a day. It can also use your ChatGPT subscription now and I've been getting a ton of usage out of my Plus tier. Even with Sonnet, Claude Pro would have limited me LONG ago.
A few thoughts:
-
While I still prefer CC + Opus 4.1 overall, I actually prefer the code that Codex CLI + GPT-5 writes. It's closer to the code I'd also write.
-
I've used CC over Bedrock and Vertex for work and the rate limits were getting really ridiculous. Not sure this also happens with the Anthropic API, but it's really refreshing how quick and stable GPT-5 performs over Codex CLI.
-
As of today Claude Code is a much more feature rich and complete tool compared to Codex. I miss quite a few things coming from CC, but core functionality is there and works well.
-
GPT-5 seems to have a very clear edge on debugging.
-
GPT-5 finds errors/bugs while working on something else, which I haven't noticed this strongly with Claude.
-
Codex CLI now also supports MCP, although support for image inputs doesn't seem to work.
-
Codex doesn't ship with fetch or search, so be sure to add those via MCP. I'm using my own
-
If your budget ends at $20 per month, I think ChatGPT might be the best value for your money
What's your experience?
Videos
I've been testing OpenAI's Codex CLI vs Claude Code in a 500k codebase which has a React Vite frontend and a ASP .NET 9 API, MySQL DB hosted on Azure. My takeaways from my use cases (or watch them from the YT video link in the comments):
- Boy oh boy, Codex CLI has caught up BIG time with GPT5 High Reasoning, I even preferred it to Claude Code in some implementations
- Codex uses GPT 5 MUCH better than in other AI Coding tools like Cursor
- Vid: https://youtu.be/MBhG5__15b0
- Codex was lacking a simple YOLO mode when I tested. You had to acknowledge not running in a sandbox AND allow it to never ask for approvals, which is a bit annoying, but you can just create an alias like codex-yolo for it
- Claude Code actually had more shots (error feedback/turns) than Codex to get things done
- Claude Code still has more useful features, like subagents and hooks. Notifications from Codex are still in a bit of beta
- GPT5 in Codex stops less to ask questions than in other AI tools, it's probably because of the released official GPT5 Prompting Guide by OpenAI
What is your experience with both tools?
Which one do you like more?
I have now used Claude Code for gamedev. Claude Code is great but sometimes it gives too much features I don’t need or put code in really strange places. Sometimes it tried to make god objects.
Do you think Codex cli would be better?
Hey everyone!
I've been an avid Claude user for over 6 months now and I absolutely love the value it brings to my workflow. I've been seeing a lot of hype about Codex, specifically with the GPT-5.4 model. I've tried GPT-5.4 in Cursor and I've seen promising results but I'm unsure about committing to one model, since the Codex app brings a few advantages over CC.
I've heard codex has more efficient token usage and the app, for me, would be a much more intuitive workflow compared to the CLI. I'm curious to know you guys' takes if you've regularly used both and the key differences that are actually monumental and not just 5-10% performance increments. Would love to know your experiences.
*Just FYI: I run a dev shop with around 10 clients and I actively contribute to all of those projects if that helps you get an idea of scale and usage. Mostly varies, but I'd say I'm averaging 2-3M tokens/month.
Hi everyone,
I’m a bit torn and would love some honest opinions from people who’ve actually used these tools in real projects.
I currently have a ChatGPT Business subscription and a Claude Pro subscription.
I keep seeing conflicting advice. Some people say you should use Codex mainly for reviewing and let Claude (Sonnet) do the actual coding. Others say the opposite: let Codex implement and use Claude for reasoning and planning.
My use case is mostly Laravel projects (Vue, Inertia, Tailwind), plus some general PHP and JavaScript. These aren’t massive systems, but usually internal workflow apps: dashboards, request or intake forms, approvals, basic admin panels, that kind of thing.
My current flow looks like this:
I first write a very strict taskfile md file with clear steps and constraints with the help of ChatGPT. Based on that, I decide whether to use Codex CLI (gpt-5.2-codex) or Claude Code (Sonnet 4.5) for the actual implementation. I work entirely in VS Code with two terminals open.
Given this setup, I’m curious:
Who would you let do the planning?
Who would you let do the actual implementation?
Who would you trust more for reviewing or tightening code?
Any tips, patterns, or personal experiences are very welcome.
Thanks in advance, I really appreciate the input.
I want to start by making one thing clear: I’m not a fan of any AI.
I don’t care about the company name or the product name. I just want a tool that helps me work better.
I recently paid for Claude Pro to complement my Codex Plus plan. I’ve been using Codex for several months now, and honestly, I’ve been very satisfied with it. The mistakes it makes are usually minimal, and most of the time Codex fixes them itself or I solve them in just a few minutes.
So far, my experience with Codex has been very good, even better than I expected. I don’t use it for extremely intensive tasks, but last week I hit the weekly limit and decided to subscribe to Claude as a supplement. I was also very curious because people on social media say amazing things about Claude, and I wanted to see for myself whether it really lived up to the hype.
But the truth is that my experience has been deeply disappointing. And just to be clear, I’m not trying to convince anyone of anything, I’m only sharing my personal experience.
With Claude, I feel like it just does whatever it wants. A lot of the time it doesn’t follow instructions, it does things I didn’t ask for, it doesn’t stick to the plan, it breaks parts of the code, and overall I find it frustrating to work with. On top of that, I get the feeling that it struggles to see beyond the immediate task.
With Codex, I feel the exact opposite. Sometimes it surprises me in a very positive way, because it not only does what I ask, but it also understands the context better, anticipates problems, and suggests fairly complete and functional implementations. Sometimes when I read its feedback, I think, “wow, I had forgotten about that,” or “I hadn’t thought of that.”
Honestly, it’s a shame because I really wanted to like Claude, especially since Claude’s $100 plan seems reasonable to me.
Has anyone else had a similar experience?
Am I doing something wrong with Claude, or does it just not fit the way I work?
I have used mostly Windsurf and Kilo Code to build around 8 projects, the most complicated one is a flutter iOS & Android app with appr. 750 test users using firebase as backend and Gemini Flash 2.5 for AI functionalities.
Now I would like to start learning CLI AI coding tools. 2 months ago the choice would have been an obvious Claude Code (I have the pro subscription), but I've seen the hype around OpenAI's Codex CLI these days.
Would be great to hear from your experience:
-
What is the difference between these 2 right now besides the LLM models?
-
What are the usage limits for a mix of planning / coding / debugging usage? (for Claude Pro and OpenAI Plus sub)
-
Any tipps for switching from editor based coding to terminal based? I am slightly hesitant because I am a visual person and am afraid that I will lose the overview using the terminal. Or do you guys use terminal and editor at the same time?
-
Are there any other options you recommend?
Hey! Genuine question, why did you choose to purchase Claude Code and not OpenAI Codex? I’m just trying to decide between the two.
i was working on a Mediation System implemented in a +2M download published game, and oooh boy, i really needed some deep analysis of some flaws that were causing a lot of bad performance in the monetization, not your (overage mediation system)
the code is complex and very delicate and the smallest change could break everything
i spend 4 days just trying different llms to identify issues, edge cases of what could have happened that lead to the bad monetization performance of this system, i noticed claude is extremely fast , he can do +326 diff change in a blink of an eye, output a new 560 lines of code class, in few seconds, BUT, it may seem good and well done at a glance, but onse you dig deep into the code, there is a lot of bad imlementation, critical logical flaws,
today i desiced to test CODEX i got the pro sub, and i gave the agent a task to analyze the issues and logical flaws in the system, it went out for 30 minutes digging and reading every single file !! grepping every single method and fetching , wheres it called from and where its going, and it identified a lot of issues that were very spot ON, claude code would just read 2 or 3 files, maybe grep a few methods here and there, in a very lighting fast way, and just come up with garbage analysis that is lacking and useless, (this is an advanced C# mediation system that is used by +5M users ) !
now codex is doing its magic, i dont mind it being slow, taking its time, i'd rather wait an hour and be done with the task and i see clear improvement, than spend 4 days hitting my head to the wall with claude
This is very unfortunate that claude is at this low now, it used to be the SOTA in every single aspect of coding, and i wish they give us back OUR Beloved Claude !
but for now i'm joining the Codex Clan!
it may sound like i'm like telling you codex is better go ahead and famboying openai
i trully dont like OPENAI and i always prefered claude models, but the reality is that we are ANGRY About the current state of claude, and we want OUR KING BACK ! that's why we are shouting loud , hopefully anthropic will hear, and we will be glad to jump ship back to our beloved claude! but for now, it feels like a low level IQ model , too verbose and too much emojis in chat, and unnnessassary code comments,
codex feels like speaking to a mature senior that understands you, understands your need and saves you time imlementing whats in your mind, and even give you some insights that you may have missed, even tho experienced we are humans afterall ...
I’ve tried both, and for now I slightly prefer Codex. I can’t fully explain why, it mostly comes down to some personal benchmarks based on my day-to-day work.
One big plus for Codex is usage: on the $20 plan I’ve never hit usage limits or interruptions, while using the same plan on both.
With Codex I’m using AGENTS.md, some reusable prompts in a prompts folder, and I’m planning to experiment with skills. I also tried plugging in a simple MCP server I built, but I couldn’t get it to work with Codex, so it feels a bit less flexible in that area.
What do you think is better overall: Claude Code or Codex? In terms of output quality and features.
Let the fight begin
I know many of you are dealing with the same frustrations, but I just have to say how incredible Codex CLI feels right now compared to Claude Code. It reminds me of the excitement I had when I first started using CC. Over the past month, though, CC hasn’t been able to deliver much of any real value, often breaking more things than it fixes, and good luck trying to get new features implemented.
Then I tried Codex CLI, and wow. Everything I ask for just gets done, right the first time, no extra fluff. I’ve always loved CC and I really hope Anthropic gets it back on track, but if you’re looking to get real work done today, Codex is where it’s at. (And yes, I’ve used Cursor and Augment too, Codex still comes out on top.)
I was initially impressed with Claude Code: it felt sharper, faster, and more context aware.
But lately, it started downgrading - shorter answers, less consistency, and a weird obsession with creating random .md files.
So I decided to cancel my Max plan, try Codex instead (since it had a free month on Pro).
Big mistake. The difference is night and day - Codex feels unfinished, often cutting off mid-sentence.
I used Claude daily for product work: roadmaps, architecture, UI mockups, pitch decks; it became a genuine co-pilot for building.
Not sure if I’ll go back to Max yet, but I’m definitely renewing Claude Pro back.
Sometimes, you only realize how good something was after you switch.
I finally got convinced and figured I'd try Codex CLI with one week left on my CC Max plan. So I'm using them side by side at the moment, here are some of my thoughts:
-
Claude Code interface is much more mature, feels like you are part of the development, Codex CLI feels more like an agent that does things in the background and delivers the final code to you
-
Not hearing "you are absolutely right" 100 times a day has a therapeutic effect
-
GPT-5 High Vs Opus : So far they are very close, with different styles. CC with Opus 4.1 always over designs and complicates things, GPT 5 does less of that. GPT 5 has been better at debugging my technology stack so far. Opus writes more readable outputs, for example in architectural discussions I can follow Opus a little bit better.
Interesting to see how these services evolve over time, both really good, but getting pricey so I need to decide which one I keep a month from now. Moving the workflow (Hooks, etc) seems to be a pain.
From what I’ve seen so far, Claude Code seems to have the best overall reviews in terms of quality and performance. The main downside for me is that it’s locked behind a company and not open source (I know about the leak, but I’m more interested in something officially open and actively maintained).
Codex, on the other hand, looks really appealing because it’s open source and allows for forks, which gives it a lot more flexibility and long-term potential.
Then there’s OpenCode, probably the most interesting of the three. It has a huge community and a lot of momentum, but I’m not sure if it’s actually on par with the others in real-world use.
Curious to hear your thoughts, how do these compare in practice? Is OpenCode actually competitive, or is it more hype than substance?
Oh and by Claude i'm referring to the open sourced forks that are comming, which we don't know if will be updated or etc, not using the proprietary one ever
I was using Claude Code for a while, but after seeing some posts about Codex CLI, I decided to try it out, and I’m really glad I did.
Even with just the OpenAI Plus plan, I’m not constantly running into usage limits like I was with Claude. That alone makes a huge difference. GPT-5 feels a lot smarter to me. It handles complex stuff better imo.
Only thing that bugs me is how many permissions Codex CLI asks for (I think there's an option to stop asking for permissions?). But overall, it’s been a much smoother experience.
Anyone else switched?
So I have been mainly using ChatGPT cus of Codex in Vscode. I now wanna switch to Claude because of this whole fiasco but dont know if Claude code is better or not.
I want to make it clear upfront that I mean no disrespect with this post. I’ve been using Claude Code for a long time, and I honestly don’t even remember when I first subscribed to the Claude Code Max plan at $200.
However, with the recent issues and ongoing discussions around Claude models and their performance in CC, I’ve noticed that many users seem to be quietly switching over to Codex.
I’ve also seen a number of YouTubers talking about this and sharing their thoughts, which has left me a bit confused. At the end of the day, I’m spending my own money, and I just want to make sure it’s worth it. I’m an AI engineer with several projects underway, and for the past few months I’ve been using CC both in my work and on personal projects.
What I’m really trying to figure out is not which tool is “the best” overall, but which one delivers the most value in terms of quality results. My main concern is making sure I’m not wasting money and that I’m getting the most out of whichever tool I use.
Appreciate any opinions/insights.
I'm not ready to call Codex a "Claude killer" just yet, but I'm definitely impressed with what I've seen over the past six hours of use.
I'm currently on Anthropic's $200/month plan (Claude's highest tier) and ChatGPT's $20 plus plan. Since this was my first time trying ChatGPT, I started with the Plus tier to get a feel for it. There is also a $200 pro tier available for Chatgpt This past week, Claude has been underperforming significantly, and I'm not alone in noticing this. After seeing many users discuss ChatGPT's coding capabilities, I decided to give Codex a shot, and I was impressed. I had two persistent coding issues that Claude couldn't resolve and ChatGPT fixed both of them easily, in one prompt. There are also a few other things I like about Codex so far. It has Better listening skills. It pays closer attention to my specific requests, it admits mistakes, it collaborates better on troubleshooting by asking clarifying questions about my code, and its response is noticeably quicker than Claude Opus. However, ChatGPT isn't perfect either. I'm currently dealing with a state persistence issue that neither AI has been able to solve. Additionally, since I've only used ChatGPT for six hours, compared to months with Claude, I may have given it tasks it excels at. Bottom line: I'm genuinely impressed with ChatGPT's performance, but I'm not abandoning Claude just yet. However, if you haven't tried ChatGPT for coding, I'd definitely recommend giving it a shot – it performed exceptionally well for my specific use cases. It may be that going forward I use both to finish my projects.
Edit: to install make sure you have node.js installed and your computer then run
npm install -g @openai/codex
You can also install using homebrew by running.
brew install codex