Tuesday, January 13, 2026
No Result
View All Result
The Crypto HODL
  • Home
  • Bitcoin
  • Crypto Updates
    • Altcoin
    • Ethereum
    • Crypto Updates
    • Crypto Mining
    • Crypto Exchanges
  • Blockchain
  • NFT
  • DeFi
  • Web3
  • Metaverse
  • Regulations
  • Scam Alert
  • Analysis
  • Videos
Marketcap
  • Home
  • Bitcoin
  • Crypto Updates
    • Altcoin
    • Ethereum
    • Crypto Updates
    • Crypto Mining
    • Crypto Exchanges
  • Blockchain
  • NFT
  • DeFi
  • Web3
  • Metaverse
  • Regulations
  • Scam Alert
  • Analysis
  • Videos
No Result
View All Result
The Crypto HODL
No Result
View All Result

Anthropic Claude 4 Review: Creative Genius Trapped by Old Limitations

May 25, 2025
in Web3
Reading Time: 23 mins read
0 0
A A
0
Home Web3
Share on FacebookShare on Twitter


San Francisco-based Anthropic simply dropped the fourth technology of its Claude AI fashions, and the outcomes are… sophisticated. Whereas Google pushes context home windows previous one million tokens and OpenAI builds multimodal programs that see, hear, and converse, Anthropic caught with the identical 200,000-token restrict and text-only strategy. It is now the odd one out amongst main AI firms.

The timing feels deliberate—Google introduced Gemini this week too, and OpenAI unveiled a brand new coding agent primarily based on its proprietary Codex mannequin. Claude’s reply? Hybrid fashions that shift between reasoning and non-reasoning modes relying on what you throw at them—delivering what OpenAI expects to deliver every time they launch GPT-5.

However here is one thing for API customers to noticeably think about: Anthropic is charging premium costs for that improve.

Picture: t3.gg

The chatbot app, nevertheless, stays the identical at $20 with Claude Max priced at $200 a month, with 20x greater utilization limits.

We put the brand new fashions via their paces throughout artistic writing, coding, math, and reasoning duties. The outcomes inform an attention-grabbing story with marginal enhancements in some areas, shocking enchancment in others, and a transparent shift in Anthropic’s priorities away from normal use towards developer-focused options.

Right here is how each Claude Sonnet 4 and Claude Opus 4 carried out in our totally different assessments. (You’ll be able to test them out, together with our prompts and outcomes, in our Github repository.)

Artistic writing

Artistic writing capabilities decide whether or not AI fashions can produce participating narratives, preserve constant tone, and combine factual components naturally. These abilities matter for content material creators, entrepreneurs, and anybody needing AI help with storytelling or persuasive writing.

As of now, there is no such thing as a mannequin that may beat Claude on this subjective take a look at (not contemplating Longwriter, after all). So it is senseless to match Claude towards third-party choices. For this process we determined to place Sonnet and Opus face-to-face.

We requested the fashions to write down a brief story about an individual who travels again in time to forestall a disaster however finally ends up realizing that their actions from the previous truly have been a part of the occasions that made existence lean in the direction of that particular future. The immediate added some particulars to think about and gave fashions sufficient liberty and creativity to arrange a narrative as they see match.

Claude Sonnet 4 produced vivid prose with the perfect atmospheric particulars and psychological nuance. The mannequin crafted immersive descriptions and supplied a compelling story, although the ending was not precisely as requested—but it surely match the narrative and the anticipated end result.

Total, Sonnet’s narrative building balanced motion, introspection, and philosophical insights about historic inevitability.

Rating: 9/10—undoubtedly higher than Claude 3.7 Sonnet

Claude Opus 4 grounded its speculative fiction in credible historic contexts, referencing indigenous worldviews and pre-colonial Tupi society with cautious consideration to cultural limitations. The mannequin built-in supply materials naturally and supplied an extended story than Sonnet, with out with the ability to match its poetic aptitude, sadly.

It additionally confirmed an attention-grabbing factor: The narrative began much more vividly and was extra immersive than what Sonnet supplied, however someplace across the center, it shifted to hurry the plot twist, making the entire end result boring and predictable.

Rating: 8/10

Sonnet 4 is the winner for artistic writing, although the margin remained slim. Writers, beware: Not like with earlier fashions, it seems that Anthropic hasn’t prioritized artistic writing enhancements, focusing improvement efforts elsewhere.

All of the tales can be found right here.

Coding

Coding analysis measures whether or not AI can generate practical, maintainable software program that follows greatest practices. This functionality impacts builders utilizing AI for code technology, debugging, and architectural selections.

Gemini 2.5 Professional is taken into account the king of AI-powered coding, so we examined it towards Claude Opus 4 with prolonged considering.

We zero-shot our directions for a recreation—a robotic that should keep away from journalists in its solution to merge with a pc and obtain AGI—and used one extra iteration to repair bugs and make clear totally different facets of the sport.

Claude Opus created a top-down stealth recreation with subtle mechanics, together with dynamic sound waves, investigative AI states, and imaginative and prescient cone occlusion. The implementation featured wealthy gameplay components: journalists responded to sounds via heardSound flags, obstacles blocked line-of-sight calculations, and procedural technology created distinctive ranges every playthrough.

Rating: 8/10

Google’s Gemini produced a side-scrolling platformer with cleaner structure utilizing ES6 lessons and named constants.

The sport was not practical after two iterations, however the implementation separated considerations successfully: degree.init() dealt with terrain technology, the Journalist class encapsulated patrol logic, and constants like PLAYER_JUMP_POWER enabled simple tuning. Whereas gameplay remained easier than Claude’s model, the maintainable construction and constant coding requirements earned notably excessive marks for readability and maintainability.

Verdict: Claude gained: It delivered superior gameplay performance that customers would favor.

Nevertheless, builders would possibly favor Gemini regardless of all this, because it created cleaner code that may be improved extra simply.

Our immediate and codes can be found right here. And you’ll click on right here to play the sport generated with Claude.

Mathematical reasoning

Mathematical problem-solving assessments AI fashions’ capacity to deal with complicated calculations, present reasoning steps, and arrive at appropriate solutions. This issues for academic functions, scientific analysis, and any area requiring exact computational considering.

We in contrast Claude and OpenAI’s newest reasoning mannequin, o3, asking the fashions to resolve an issue that appeared on the FrontierMath benchmark—designed particularly to be laborious for fashions to resolve:

“Assemble a level 19 polynomial p(x) ∈ C[x] such that X := {p(x) = p(y)} ⊂ P1 × P1 has a minimum of 3 (however not all linear) irreducible parts over C. Select p(x) to be odd, monic, have actual coefficients and linear coefficient -19 and calculate p(19).”

Claude Opus 4 displayed its full reasoning course of when tackling troublesome mathematical challenges. The transparency allowed evaluators to hint logic paths and establish the place calculations went mistaken. Regardless of displaying all of the work, the mannequin failed to attain good accuracy.

OpenAI’s o3 mannequin achieved 100% accuracy on an identical mathematical duties, marking the primary time any mannequin solved the take a look at issues fully. Nevertheless, o3 truncated its reasoning show, displaying solely ultimate solutions with out intermediate steps. This strategy prevented error evaluation and made it inconceivable for customers to confirm the logic or be taught from the answer course of.

Verdict: OpenAI o3 gained the mathematical reasoning class via good accuracy, although Claude’s clear strategy provided academic benefits. For instance, researchers can have a better time catching failures whereas analyzing the complete Chain of Thought, as an alternative of getting to both absolutely belief the mannequin or clear up the issue manually to corroborate outcomes.

You’ll be able to test Claude 4’s Chain of Thought right here.

Non-mathematical reasoning and communication

For this analysis, we wished to check the fashions’ capacity to know complexities, craft nuanced messages, and stability pursuits. These abilities show important for enterprise technique, public relations, and any state of affairs requiring subtle human communication.

We supplied Claude, Grok, and ChatGPT directions to craft a single communication technique that concurrently addresses 5 totally different stakeholder teams a couple of essential scenario at a big medical heart. Every group has vastly totally different views, emotional states, info wants, and communication preferences.

Claude demonstrated distinctive strategic considering via a three-pillar messaging framework for a hospital ransomware disaster: Affected person Security First, Lively Response, and Stronger Future. The response included particular useful resource allocations of $2.3 million emergency funding, detailed timelines for every stakeholder group, and culturally delicate variations for multilingual populations. Particular person board member considerations obtained tailor-made consideration whereas sustaining message consistency. The mannequin supplied an excellent set of opening statements to seize an thought of the right way to strategy every viewers.

ChatGPT was additionally good on the process, however not on the identical degree of element and practicality. Whereas offering strong frameworks with clear core ideas, GPT4.1 relied extra on tone variation than substantive content material adaptation. The responses have been in depth and detailed, anticipating questions and moods, and the way our actions could impression these being addressed. Nevertheless, it lacked particular useful resource allocations, detailed deliverables, and different particulars that Claude supplied.

Verdict: Claude wins

You’ll be able to test the outcomes and Chain of Thought for every mannequin, right here.

Needle within the haystack

Context retrieval capabilities decide how successfully AI fashions can find particular info inside prolonged paperwork or conversations. This ability proves essential for authorized analysis, doc evaluation, educational literature evaluations, and any state of affairs requiring exact info extraction from massive textual content volumes.

We examined Claude’s capacity to establish particular info buried inside progressively bigger context home windows utilizing the usual “needle in a haystack” methodology. This analysis concerned putting a focused piece of knowledge at numerous positions inside paperwork of various lengths and measuring retrieval accuracy.

Claude Sonnet 4 and Opus 4 efficiently recognized the needle when embedded inside an 85,000 token haystack. The fashions demonstrated dependable retrieval capabilities throughout totally different placement positions inside this context vary, sustaining accuracy whether or not the goal info appeared firstly, center, or finish of the doc. Response high quality remained constant, with the mannequin offering exact citations and related context across the retrieved info.

Nevertheless, the fashions’ efficiency hit a tough limitation when trying to course of the 200,000 token haystack take a look at. They might not full this analysis as a result of the doc dimension exceeded their most context window capability of 200,000 tokens. It is a important constraint in comparison with opponents like Google’s Gemini, which handles context home windows exceeding a million tokens, and OpenAI’s fashions with considerably bigger processing capabilities.

This limitation has sensible implications for customers working with in depth documentation. Authorized professionals analyzing prolonged contracts, researchers processing complete educational papers, or analysts reviewing detailed monetary experiences could discover Claude’s context restrictions problematic. The lack to course of the complete 200,000 token take a look at means that real-world paperwork approaching this dimension may set off truncation or require guide segmentation.

Verdict: Gemini is the higher mannequin for lengthy context duties

You’ll be able to test on each the necessity and the haystack, right here.



Conclusion

Claude 4 is nice, and higher than ever—but it surely’s not for everybody.

Energy customers who want its creativity and coding capabilities will likely be very happy. Its understanding of human dynamics additionally makes it best for enterprise strategists, communications professionals, and anybody needing subtle evaluation of multi-stakeholder eventualities. The mannequin’s clear reasoning course of additionally advantages educators and researchers who want to know AI decision-making paths.

Nevertheless, novice customers wanting the complete AI expertise could discover the chatbot a little bit lackluster. It does not generate video, you can not discuss to it, and the interface is much less polished than what yow will discover in Gemini or ChatGPT.

The 200,000 token context window limitation impacts Claude customers processing prolonged paperwork or sustaining prolonged conversations, and it additionally implements a really strict quota which will have an effect on customers anticipating lengthy classes.

In our opinion, it’s a strong “sure” for artistic writers and vibe coders. Different varieties of customers may have some consideration, evaluating professionals and cons towards options.

Edited by Andrew Hayward

Typically Clever E-newsletter

A weekly AI journey narrated by Gen, a generative AI mannequin.



Source link

Tags: AnthropicClaudecreativeGENIUSLimitationsReviewTrapped
Previous Post

XRP Price Flashes Signal That Triggered 600% Rally In November 2024

Next Post

Top Trending Cryptos on Solana Chain Today – QuStream, eZ1, YAKU

Related Posts

Why Bitcoin May Be Underpricing January Rate Cut Odds
Web3

Why Bitcoin May Be Underpricing January Rate Cut Odds

January 13, 2026
YouTuber Cracks Coca-Cola’s 139-Year-Old Secret Formula—Here ‘s the Recipe
Web3

YouTuber Cracks Coca-Cola’s 139-Year-Old Secret Formula—Here ‘s the Recipe

January 12, 2026
Two major crypto events canceled after city hit by 18 violent physical attacks on crypto holders amid market downturn
Web3

Two major crypto events canceled after city hit by 18 violent physical attacks on crypto holders amid market downturn

January 12, 2026
Bitcoin Shrugs Off Powell Probe as DOJ Targets Fed Chair
Web3

Bitcoin Shrugs Off Powell Probe as DOJ Targets Fed Chair

January 12, 2026
Should Politicians Be Able to Use Prediction Markets? House Bill Proposes Ban
Web3

Should Politicians Be Able to Use Prediction Markets? House Bill Proposes Ban

January 9, 2026
Insiders Say DeepSeek V4 Will Beat Claude and ChatGPT at Coding, Launch Within Weeks
Web3

Insiders Say DeepSeek V4 Will Beat Claude and ChatGPT at Coding, Launch Within Weeks

January 10, 2026
Next Post
Top Trending Cryptos on Solana Chain Today – QuStream, eZ1, YAKU

Top Trending Cryptos on Solana Chain Today - QuStream, eZ1, YAKU

Simplify Investing With Stock Recommendations App

Simplify Investing With Stock Recommendations App

Wall Street Gets Ethereum’s ‘Digital Oil’, Says Etherealize Co-Founder Vivek Raman

Wall Street Gets Ethereum's 'Digital Oil', Says Etherealize Co-Founder Vivek Raman

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Twitter Instagram LinkedIn Telegram RSS
The Crypto HODL

Find the latest Bitcoin, Ethereum, blockchain, crypto, Business, Fintech News, interviews, and price analysis at The Crypto HODL

CATEGORIES

  • Altcoin
  • Analysis
  • Bitcoin
  • Blockchain
  • Crypto Exchanges
  • Crypto Mining
  • Crypto Updates
  • DeFi
  • Ethereum
  • Metaverse
  • NFT
  • Regulations
  • Scam Alert
  • Uncategorized
  • Videos
  • Web3

SITE MAP

  • Disclaimer
  • Privacy Policy
  • DMCA
  • Cookie Privacy Policy
  • Terms and Conditions
  • Contact us

Copyright © 2023 The Crypto HODL.
The Crypto HODL is not responsible for the content of external sites.

No Result
View All Result
  • Home
  • Bitcoin
  • Crypto Updates
    • Altcoin
    • Ethereum
    • Crypto Updates
    • Crypto Mining
    • Crypto Exchanges
  • Blockchain
  • NFT
  • DeFi
  • Web3
  • Metaverse
  • Regulations
  • Scam Alert
  • Analysis
  • Videos
Crypto Marketcap

Copyright © 2023 The Crypto HODL.
The Crypto HODL is not responsible for the content of external sites.

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In