The 200k context window is deflating especially when gpt and gemini are eating them for lunch. Even if they went to 500k would be better.
Benchmarks at this point in the A.I game are negligible at best and you sure don't "Feel" a 1% difference between the 3. It feels like we are getting to the point of diminishing returns.
Us as programmers should be able to see the forest from the trees here. We think differently than the normal person. We think outside of the box. We don't get caught in hype as we exist in the realm of research, facts and practicality.
This Claude release is more hype than practical.
Source: Code with Claude Opening Keynote
Videos
Been using these tools for the last few years. Can already tell opus and sonnet 4 have set a completely new benchmark, especially using Claude Code.
They just work, less hallucination, less infinite loops of confusion. You can set it off and come back with a 80-90% confidence it’s done what you asked. Maximum 3-4 iterations to get website/app component styling perfect (vs 5-10 before).
I’ve already seen too many of the classic ‘omg this doesn’t work for me they suck, overhyped’ posts. Fair enough if that’s your experience, but I completely disagree and can’t help but think your prompting is the problem.
Without using too much stereotypical AI hyperbole, I think these are the biggest step change since GPT 3.
Today, Anthropic is introducing the next generation of Claude models: Claude Opus 4 and Claude Sonnet 4, setting new standards for coding, advanced reasoning, and AI agents. Claude Opus 4 is the world’s best coding model, with sustained performance on complex, long-running tasks and agent workflows. Claude Sonnet 4 is a drop-in replacement for Claude Sonnet 3.7, delivering superior coding and reasoning while responding more precisely to your instructions.
Claude Opus 4 and Sonnet 4 are hybrid models offering two modes: near-instant responses and extended thinking for deeper reasoning. Both models can also alternate between reasoning and tool use—like web search—to improve responses.
Both Claude 4 models are available today for all paid plans. Additionally, Claude Sonnet 4 is available on the free plan.
Read more here: https://www.anthropic.com/news/claude-4
Anthropic recently unveiled Claude 4 (Opus and Sonnet), achieving record-breaking 72.7% performance on SWE-bench Verified and surpassing OpenAI’s latest models. Benchmarks aside, I wanted to see how Claude 4 holds up under real-world software engineering tasks. I spent the last 24 hours putting it through intensive testing with challenging refactoring scenarios.
I tested Claude 4 using a Rust codebase featuring complex, interconnected issues following a significant architectural refactor. These problems included asynchronous workflows, edge-case handling in parsers, and multi-module dependencies. Previous versions, such as Claude Sonnet 3.7, struggled here—often resorting to modifying test code rather than addressing the root architectural issues.
Claude 4 impressed me by resolving these problems correctly in just one attempt, never modifying tests or taking shortcuts. Both Opus and Sonnet variants demonstrated genuine comprehension of architectural logic, providing solutions that improved long-term code maintainability.
Key observations from practical testing:
Claude 4 consistently focused on the deeper architectural causes, not superficial fixes.
Both variants successfully fixed the problems on their first attempt, editing around 15 lines across multiple files, all relevant and correct.
Solutions were clear, maintainable, and reflected real software engineering discipline.
I was initially skeptical about Anthropic’s claims regarding their models' improved discipline and reduced tendency toward superficial fixes. However, based on this hands-on experience, Claude 4 genuinely delivers noticeable improvement over earlier models.
For developers seriously evaluating AI coding assistants—particularly for integration in more sophisticated workflows—Claude 4 seems to genuinely warrant attention.
A detailed write-up and deeper analysis are available here: Claude 4 First Impressions: Anthropic’s AI Coding Breakthrough
Interested to hear others' experiences with Claude 4, especially in similarly challenging development scenarios.
I mean how the heck literally Is Qwen-3 better than claude-4(the Claude who used to dog walk everyone). this is just disappointing 🫠
I have been extensively using Gemini 2.5 Pro for coding-related stuff and O3 for everything else, and it's crazy that within a month or so, they look kind of obsolete. Claude Opus 4 is the best overall model available right now.
I ran a quick coding test, Opus against Gemini 2.5 Pro and OpenAI o3. The intention was to create visually appealing and bug-free code.
Here are my observations
-
Claude Opus 4 leads in raw performance and prompt adherence.
-
It understands user intentions better, reminiscent of 3.6 Sonnet.
-
High taste. The generated outputs are tasteful. Retains the Opus 3 personality to an extent.
-
Though unrelated to code, the model feels nice, and I never enjoyed talking to Gemini and o3.
-
Gemini 2.5 is more affordable in pricing and takes much fewer API credits than Opus.
-
One million context length in Gemini is undefeatable for large codebase understanding.
-
Opus is the slowest in time to first token. You have to be patient with the thinking mode.
Check out the blog post for complete comparison analysis with codes: Claude 4 Opus vs. Gemini 2.5 vs. OpenAI o3
The vibes with Opus are the best; it has a personality and is stupidly capable. But too pricey; it's best used with the Claude app, the API cost will put a hole in your pocket. Gemini will always be your friend with free access and the cheapest SOTA model.
Would love to know your experience with Claude 4 Opus and how you would compare it with o3 and Gemini 2.5 pro in coding and non-coding tasks.
Finished benchmarking Claude 4 (Sonnet) across a range of document understanding tasks, and the results are… not that good. It's currently ranked 7th overall on the leaderboard.
Key takeaways:
Weak performance in OCR – Claude 4 lags behind even smaller models like GPT-4.1-nano and InternVL3-38B-Instruct.
Rotation sensitivity – We tested OCR robustness with slightly rotated images ([-5°, +5°]). Most large models had a 2–3% drop in accuracy. Claude 4 dropped 9%.
Poor on handwritten documents – Scored only 51.64%, while Gemini 2.0 Flash got 71.24%. It also struggled with handwritten datasets in other tasks like key information extraction.
Chart VQA and visual tasks – Performed decently but still behind Gemini, Claude 3.7, and GPT-4.5/o4-mini.
Long document understanding – Claude 3.7 Sonnet (reasoning:low) ranked 1st. Claude 4 Sonnet ranked 13th.
One bright spot: table extraction – Claude 4 Sonnet is currently ranked 1st, narrowly ahead of Claude 3.7 Sonnet.
Leaderboard: https://idp-leaderboard.org/
Codebase: https://github.com/NanoNets/docext
How has everyone’s experience with the models been so far?
Introducing the next generation: Claude Opus 4 and Claude Sonnet 4.
Claude Opus 4 is our most powerful model yet, and the world’s best coding model.
Claude Sonnet 4 is a significant upgrade from its predecessor, delivering superior coding and reasoning.
I don't know what magic you guys did, but holy crap, Claude 4 opus is freaking amazing, beyond amazing! Anthropic team is legendary in my books for this. I was able to solve a very specific graph database chatbot issue that was plaguing me in production.
Rock on Claude team!
Been using ChatGPT Plus with o3 and Gemini 2.5 Pro for coding the past months. Both are decent but always felt like something was missing, you know? Like they'd get me 80% there but then I'd waste time fixing their weird quirks or explaining context over and over or running in a endless error loop.
Just tried Claude 4 Opus and... damn. This is what I expected AI coding to be like.
The difference is night and day:
-
Actually understands my existing codebase instead of giving generic solutions that don't fit
-
Debugging is scary good - it literally found a memory leak in my React app that I'd been hunting for days
-
Code quality is just... clean. Like actually readable, properly structured code
-
Explains trade-offs instead of just spitting out the first solution
Real example: Had this mess of nested async calls in my Express API. ChatGPT kept suggesting Promise.all which wasn't what I needed. Gemini gave me some overcomplicated rxjs nonsense. Claude 4 looked at it for 2 seconds and suggested a clean async/await pattern with proper error boundaries. Worked perfectly.
The context window is massive too - I can literally paste my entire project and it gets it. No more "remember we discussed X in our previous conversation" BS.
I'm not trying to shill here but if you're doing serious development work, this thing is worth every penny. Been more productive this week than the entire last month.
Got an invite link if anyone wants to try it: https://claude.ai/referral/6UGWfPA1pQ
Anyone else tried it yet? Curious how it compares for different languages/frameworks.
EDIT: Just to be clear - I've tested basically every major AI coding tool out there. This is the first one that actually feels like it gets programming, not just text completion that happens to be code. This also takes Cursor to a whole new level!
I know, I know, whenever a model comes out you get people saying this, but it's on very concrete things for me, I'm not just biased against it. For reference, I'm comparing 4 Sonnet (concise) with 3.7 Sonnet (concise), no reasoning for either.
I asked it to calculate the total markup I paid at a gas station relative to the supermarket. I gave it quantities in a way I thought was clear ("I got three protein bars and three milks, one of the others each. What was the total markup I paid?", but that's later in the conversation after it searched for prices). And indeed, 3.7 understands this without any issue (and I regenerated the message to make sure it wasn't a fluke). But with 4, even with much back and forth and several regenerations, it kept interpreting this as 3 milk, 1 protein bar, 1 [other item], 1 [other item], until I very explicitly laid it out as I just did.
And then, another conversation, I ask it, "Does this seem correct, or too much?" with a photo of food, and macro estimates for the meal in a screenshot. Again, 3.7 understands this fine, as asking whether the figures seem to be an accurate estimate. Whereas 4, again with a couple regenerations to test, seems to think I'm asking whether it's an appropriate meal (as in, not too much food for dinner or whatever). And in one instance, misreads the screenshot (thinking that the number of calories I will have cumulatively eaten after that meal is the number of calories of that meal).
Is anyone else seeing any issues like this?
Are there any rumors about a release on Claude 4? I love 3.5 so I can only hope the next version comes out soon
I thought the rumor was that this was coming at some point in February.
3.7 seems to indicate no new frontier models for a few months at least, going off of how the gap from 3.5 new to 3.7 was a few months.
What do we think? Was this actually the big drop that was being rumored?