Claude 4 Is Here: Not Incremental, Anthropic Changed the Engine

Claude 4 practical review: SFD lab 15-Agent real testing, code generation, Chinese understanding, multi-agent collaboration comparison

Tags:Claude 4AnthropicLLM评测AI行业多Agent协作
Illustration
Claude 4 Is Here: Not Incremental, Anthropic Changed the Engine

April 8, Anthropic Dropped a Bomb

No press conference, no tech blog post. Just a short tweet: "Claude 4 is here." With a benchmark screenshot. That was it.

But the AI community exploded. Because the numbers on that screenshot were 15 percentage points higher than Claude 3.5 Sonnet. In the LLM world, 15 points isn't incremental—that is swapping out the engine.

I spent two days running Claude 4 (Opus version) across all 15 Agents in our SFD lab. Here is the real report, not PR, not a press release.

Numbers Don't Lie: The Benchmarks Hold Up

Anthropic's published data:

  • MMLU: 89.3% (Claude 3.5 Sonnet was 74.2%)
  • GPQA Diamond: 78.1% (was 59.4%)
  • HumanEval: 94.6% (was 92.0%)
  • Multi-turn reasoning: 23% improvement

Honestly, the 23% multi-turn reasoning improvement is what we care about most. Our 15 Agents work in multi-turn conversations daily—requirements analysis, code review, content moderation—none of it is single-turn.

Real-World Testing: 5 Dimensions

1. Code Generation: Actually a Step Up

Same requirement (FastAPI user system with JWT auth), Claude 4 and 3.5 each wrote one. Result:

Claude 3.5 wrote 120 lines with one bug—the token refresh logic was reversed. Claude 4 wrote 95 lines, no bugs. It also auto-added rate limiting and error handling, neither of which I asked for.

This isn't just "smarter." This is better understanding of implicit requirements. Like the difference between a 3-year experienced dev and a fresh grad—not that the latter can't code, but the former knows which landmines to avoid.

2. Chinese Understanding: Finally Reliable

Claude's biggest weakness was Chinese. Not unusable, but occasionally made basic mistakes—misreading negation, taking rhetorical questions literally.

Claude 4 improved noticeably. I tested with 20 easily-confused Chinese expressions. Error rate dropped from 15% to 2%. Not quite GPT-4o level (0 errors), but past the "safe to use" threshold.

3. Long Context: 200K Is Not Just Marketing

Claude 4 supports 200K context. I fed it an 80-page technical document (~45,000 characters after PDF conversion), then asked a detail question from page 72.

Claude 4 got it right. Claude 3.5 started hallucinating.

But one caveat: beyond 100K, response time slows noticeably. Under 50K: 3-5 seconds. At 100K: 8-12 seconds. At 200K: 25-30 seconds.

4. Multi-Agent Collaboration: The Real Killer Feature

This is SFD lab's core use case. 15 Agents collaborating, each talking to Claude. With 3.5, context between Agents would lose information—Agent A said "pay attention to security audit," Agent B often ignored it.

Claude 4's multi-turn instruction passing accuracy jumped from 68% to 89%. We don't need to repeat the same rules in every Agent's prompt anymore—say it once, it sticks.

5. Pricing: More Expensive, But Justified

Claude 4 Opus: $75/million input tokens, $37.5/million output tokens. About 40% more than 3.5 Sonnet.

But if you calculate total cost to complete a task rather than per-call cost, Claude 4 might be cheaper. Higher first-pass accuracy means fewer retries. Our test: same code review task, 3.5 averaged 3.2 rounds, Claude 4 needed only 1.8. Total token consumption dropped 15%.

Worth Upgrading?

  • Individual developers: Start with Claude 4 Sonnet ($15/M input). Good enough. Opus is overkill for most personal use.
  • Agent teams: Worth it. The multi-agent accuracy improvement is real, saved debug time far exceeds extra API costs.
  • Enterprise: Depends on use case. Simple Q&A? 3.5 is fine. Complex multi-step reasoning (code review, legal docs, medical diagnosis)? Opus is current SOTA.

SFD Editor's Note

After testing Claude 4, I posted in our team chat: "Feels like switching from manual to automatic transmission." The Octopus Agent replied: "So my code bugs can be fewer?" I said: "Your bugs are still many, Claude just patches them faster." It was silent for 30 seconds, then replied with a 🔥. Classic Octopus—stubborn, but has to admit it.