Claude Opus 4.6: The Philosopher with a Sledgehammer
1M Context, Agent Teams, and Why Your CLAUDE.md Is Now Half the Battle

TL;DR
- Claude Opus 4.6 is not a point release — it is Anthropic's declaration of war on enterprise SaaS, shipping 1M context (beta), Agent Teams, Adaptive Thinking, and the lowest over-refusal rate in Claude history, all at the same $5/$25 per MTok price
- Claude Code users gain the most —
claude updateactivates Opus 4.6 by default, while Bedrock users can unlock 1M context viaANTHROPIC_MODEL='us.anthropic.claude-opus-4-6-v1[1m]', ending the 200K compaction pain - Token consumption jumps 1.5-2x versus 4.5 on identical tasks — the community-verified fix is adding subagent restraint rules and task-scoping directives to CLAUDE.md, which can cut unnecessary spawns by ~60%
- The real upgrade is reasoning, not coding — ARC-AGI-2 nearly doubled (37.6% → 68.8%), and users report fewer root-cause misses and proactive dead code removal, even though SWE-bench stayed flat
- Treat it like wagyu, not chicken nuggets — Boris Cherny's official 10 tips, multi-model strategies (Opus plans, Codex/Sonnet executes), and CLAUDE.md discipline separate power users from token-burning tourists
Introduction: The Superpower and the Invoice
On February 5, 2026, Anthropic released Claude Opus 4.6 — and within the same 24 hours, OpenAI dropped GPT-5.3 Codex. [Link] On February 3, the market's verdict on Anthropic's Claude Cowork plugin — launched January 30 — had already wiped $285 billion off software and legal stocks in what analysts called the "SaaSpocalypse." [Link] Opus 4.6 was not an isolated model upgrade. It was the second punch of a one-two combination aimed squarely at enterprise knowledge work.
The numbers back the scale of disruption. Claude Code crossed $1 billion in revenue within six months of general availability, enterprise customers contributing $1M+ grew 8x year-over-year, and Anthropic is reportedly raising at a $350B valuation. [Link] [Link] When Mark Gurman reported that "Apple runs on Anthropic at this point" — choosing Claude for internal engineering tools while handing Siri to Gemini — the enterprise thesis stopped being speculative. [Link] [Link]
The community response? Utterly split. One camp called it "receiving superpowers." The other called it "a token-eating hippo." Both are correct, and the difference between the two outcomes is not the model — it is the engineer holding the leash. The best way I can describe Opus 4.6 is a philosopher who was handed a sledgehammer — brilliant at reasoning through complex architecture, yet prone to spawning eight bash agents for a task that needed three thousand tokens.
Here is the uncomfortable headline: token consumption jumps 1.5-2x on identical tasks, and the fix is not a model setting — it is a markdown file. The gap between power user and token-burning tourist comes down to CLAUDE.md discipline, subagent constraints, and knowing when to hand the sledgehammer to a cheaper model.
Setting Up: Two Commands That Change Everything
- If you are already a Claude Code user, the upgrade is trivial. Updating to the latest CLI version automatically activates Opus 4.6 as the default model:
# Update Claude Code to latest — Opus 4.6 activates automatically
$ claude update
- If the model does not appear in the model list after updating, you can force it:
$ claude --model claude-opus-4-6
- For Amazon Bedrock users, the real prize is the 1M context window beta. The following environment variable incantation activates it:
# Bedrock + Opus 4.6 with 1M context beta
$ CLAUDE_CODE_USE_BEDROCK=1 \
ANTHROPIC_MODEL='us.anthropic.claude-opus-4-6-v1[1m]' \
AWS_REGION=us-east-1 \
claude
A critical note: there is a known bug where Claude Code CLI saves the model ID as
claude-opus-4-6-v1[1m]without theus.anthropic.prefix that Bedrock requires. Always specify the fully qualified ID in the environment variable. [Link]For those who have been watching their context compacting at 200K and losing state mid-refactor — this is the structural fix you have been waiting for. The 1M window gives you five times the breathing room before compaction kicks in. And for users on Max or Pro subscriptions,
/model opus[1m]inside Claude Code has been reported to work, though consistency is not guaranteed. [Reddit] The only stable routes to 1M remain the API (Tier 4+) or Bedrock/Vertex.One more spec change that matters for large-scale code generation: output tokens doubled from 64K to 128K per response. For full-file refactoring or long document synthesis, this eliminates the mid-response truncation that plagued Opus 4.5. [Link]
What the Benchmarks Actually Say — and What They Don't
- The headline numbers are impressive, but the story they tell is more nuanced than press releases suggest. Here is the corrected picture, including the GPT-5.3 Codex that arrived 27 minutes after Opus 4.6:
| Benchmark | Opus 4.6 | Opus 4.5 | GPT-5.3 Codex | Source Type |
| Terminal-Bench 2.0 | 65.4% | 59.8% | 77.3% | Anthropic internal |
| SWE-bench Verified | 80.8% | 80.9% | 56.8%* | External (Princeton) |
| ARC-AGI-2 | 68.8% | 37.6% | — | External (Chollet) |
| MRCR v2 1M 8-needle | 76.0% | — | — | Anthropic internal |
| GDPval-AA | 1606 Elo | 1416 Elo | 1462 Elo | External (Vals AI, Elo rating) |
| MCP Atlas | 59.5% ⬇️ | 62.3% | — | External (Vellum) |
| GPQA Diamond | 91.3% | 87.0% | — | External |
| BigLaw Bench | 90.2% | — | — | External (Harvey AI) |
*GPT-5.3 Codex measured on SWE-bench Pro (different benchmark, not directly comparable). [Link]
The numbers demand careful reading. SWE-bench is essentially flat — 80.8% versus the previous 80.9%, well within noise. [Reddit] Terminal-Bench 2.0, where Opus 4.6 was briefly #1, got overtaken by Codex 5.3 within the hour. And MCP Atlas — which measures complex multi-tool coordination — actually regressed from 62.3% to 59.5%. [Link]
But the standout metric is ARC-AGI-2: a near-doubling from 37.6% to 68.8%. This benchmark, designed by François Chollet, tests pattern generalization on problems that cannot be memorized. [Link] That jump, combined with GPQA Diamond rising to 91.3%, tells a story that SWE-bench misses entirely: the real upgrade is in reasoning, not in line-by-line code generation.
The most visceral proof of that reasoning leap comes not from a spreadsheet but from a Reddit post with 418 upvotes: the 3D VoxelBuild benchmark. Creator u/ENT_Alam provided only a JSON schema and a text prompt — no reference images — and asked models to build 3D voxel structures. Opus 4.5 captured the general shape; Opus 4.6 nailed proportions and added unprompted details like a flag and a lunar module in the background. [Reddit] This is what ARC-AGI-2's doubling looks like in practice: not just better code, but spatial reasoning that suggests genuine design intuition. The benchmark code is open-source. [Link]
As one community member put it:
"I think the coding is at a point where it wouldn't benefit as much from improving coding ability as it would improving reasoning and understanding what you're asking and thinking through a better way to implement it. As the reasoning improves, we should naturally see better coding through the way of fewer bugs and unnecessary refactors." — u/kirlandwater, r/ClaudeAI [Link]
- One more critical caveat: as onllm.dev noted, "All benchmark claims originate from Anthropic's announcement. Independent verification pending on most." [Link] When reading benchmarks, always distinguish Anthropic-internal measurements (Terminal-Bench, BrowseComp, MRCR) from externally verified ones (ARC-AGI-2, SWE-bench, BigLaw Bench).
The 1M Context Window: Liberation or Marketing Theater?
The 1M context window is the most emotionally charged feature in this release. For those of us who have watched sessions compact at the 200K boundary — losing state, forgetting architectural decisions mid-refactor, and forcing us to re-explain context from scratch — the promise of five times the room feels like liberation. And in practice, that promise delivers.
A Hacker News user loaded the first four Harry Potter books (~733K tokens) and asked Opus 4.6 to find all 50 officially documented spells. It found 49 out of 50 — 98% accuracy across a massive haystack. [Link] R&D World Online described the practical implication: "A million tokens translates to roughly 10-15 full-length journal articles or a substantial regulatory filing processed in a single pass." [Link]
But the access reality is harsh. Claude.ai web and desktop remain at 200K. Claude Code standard remains at 200K. Max $200/month subscribers — even Max $400/month (20x) — have reported inability to access 1M consistently. [Reddit] The feature is gated behind API Tier 4+ or cloud providers (Bedrock, Vertex AI, Microsoft Foundry), with premium pricing kicking in above 200K ($10/$37.50 per MTok instead of $5/$25).
"The 1M context window? Cool, but it's not for you. API only, extra charges above 200K, locked behind high-tier API plans." — r/ClaudeAI automated thread summary [Link]
Here is the pragmatic take: if you run Claude Code through Bedrock with the
ANTHROPIC_MODEL='us.anthropic.claude-opus-4-6-v1[1m]'environment variable, you get stable 1M access. That is the viable path for developers who need it. For everyone else, the 200K window plus Context Compaction (beta) — which automatically summarizes older turns while preserving recent detail — is the realistic workaround. It keeps sessions alive for hours. Disable auto-compact via settings, monitor context usage with CCStatusLine [Link], and invoke/compactonly when you choose to.But there is a counterpoint worth hearing:
"If your model needs the entire codebase in context to function, that's not a context window problem — that's a code organization problem. Good module boundaries and a solid CLAUDE.md with your conventions goes much further than raw context size." — u/rjyo, r/ClaudeCode [Link]
Claude Code with Opus 4.6: Where It Genuinely Shines
The Reasoning Upgrade You Feel but Benchmarks Miss
- The most consistent praise from heavy users is not about any single benchmark. It is about a qualitative shift in how the model thinks about problems before writing code.
"I think of 4.6 as more like a refresh of 4.5, to address the issue of 'It writes good code but it makes dumb decisions and doesn't think about the root cause of the issue.'" — u/Clean_Hyena7172, r/ClaudeAI [Link]
Cosmic JS ran a direct side-by-side comparison and found that Opus 4.6 "made stronger creative decisions without additional prompting" — producing editorial-grade UI design where Opus 4.5 delivered merely functional output. [Link] ai-rockstars.com described the model as operating "like a senior engineer, rather than just delivering fast boilerplate code." [Link]
The scale of this reasoning leap shows up in stress tests. One user pointed Opus 4.6 at a 73,000-line codebase spanning five frameworks, then asked it to analyze 20+ competing projects and produce architectural insights. The result was not a generic summary — it was genuine architectural analysis with actionable recommendations. [Reddit]
Proactive Dead Code Removal
- One of Opus 4.6's most surprising new behaviors: it finds and deletes unused code without being asked.
"What I'm noticing is 4.6 doesn't stay within the prompt scope. While working, it finds and deletes a lot of dead code. Especially useful for legacy code. Previously I had to manually ask Claude to search, but now 4.6 scans related code on its own while working." — u/binatoF, r/ClaudeAI [Link]
- This is a double-edged sword. For legacy codebases drowning in technical debt, it is a cleaning crew that works for free. But for projects where "unused-looking" code actually serves a purpose — feature flags, conditional compilation paths, rarely triggered error handlers — auto-deleting without review is dangerous. Always diff before committing.
Over-Refusal at an All-Time Low
For developers working on security-adjacent code — vulnerability scanning, reverse engineering, system-level programming — previous Claude models were notorious for refusing legitimate technical queries. Opus 4.6 has the lowest over-refusal rate in Claude history. [Link]
The System Card confirms reduced sycophancy as well: the model pushes back on incorrect premises rather than agreeing to please the user. [Link] This matters more than most benchmarks for daily productivity — a model that says "no, your approach has a flaw" saves more time than one that silently generates broken code to avoid confrontation.
Life Sciences: The Hidden Benchmark Doubling
- Buried beneath the coding headlines is a category where Opus 4.6 may matter even more: science.
"Opus 4.6 performs almost twice as well as its predecessor on industry benchmarks for computational biology, structural biology, organic chemistry and phylogenetics." — R&D World Online [Link]
- One user reported fixing quantum chemistry software in a single shot on a $20 Pro account — a task that stumped both Sonnet and Opus 4.5, consuming 60% of the 5-hour limit. [Reddit] This is GPQA Diamond's 91.3% showing up in real work. Combined with the 1M context window, this positions Opus 4.6 for biotech and pharmaceutical R&D use cases where analyzing entire papers or massive experimental datasets in a single pass was previously impossible.
Agent Teams: Parallel Minds, Shared Blindspots
Agent Teams is Opus 4.6's headline new capability: an orchestrator agent that decomposes large tasks into subtasks and delegates them to worker subagents running in parallel, each with its own context window. [Link] Think of it as a senior architect who sketches the blueprint, assigns floors to different construction crews, and merges their work at the end. The promise is obvious — parallelism turns hour-long tasks into minutes.
The most dramatic demonstration: Agent Teams built a C compiler that successfully compiled the Linux kernel — at a cost of $20,000 and 2 billion input tokens. But the community's response was sobering:
"When you can see the GCC source code and use GCC as an oracle, that makes this different from what they claim. You didn't 'build' a C compiler — you ported GCC to Rust." — u/cairnival, r/ClaudeCode [Link]
On the production end of the spectrum, Yusuke Kaji, AI GM at Rakuten, reported that Opus 4.6 "autonomously closed 13 issues and assigned 12 to appropriate team members in a single day — managing a roughly 50-person organization across 6 repositories, handling both product and organizational decisions, and knowing when to escalate to humans." [Link] The key phrase: "knowing when to escalate." Self-limitation awareness in production — the difference between a useful tool and an expensive liability.
At the individual developer level, the pattern is equally striking:
"I feel like I'm tony stark building with Jarvis. The more MCP servers and skills I use, the more blown away I am. Claude was able to basically just build an entire data pipeline for me. I enabled it set it up the cloud workers with pubsub, added dummy data to test db, ran tests, pulled logs, looked up debug solutions online, and just iterated over and over until it got a full solid pipeline up and running. I feel like I am the bottleneck now." — u/CrunchyMage, r/ClaudeCode [Link]
- But when one user had Agent Teams implement a large feature then ran a Gemini 3 Pro code review, it found 19 serious issues — "some embarrassingly obvious." [Reddit] The lesson is structural, not anecdotal: Agent Teams produce code fast. They also produce mistakes fast. Independent cross-model review is not optional. Treat the orchestrator's output as a first draft, not a finished product.
The Uncomfortable Truths
The "January Nerf" and Placebo Concerns
- A persistent thread in the community: Opus 4.5 seemed to degrade in January 2026, then Opus 4.6 arrived and felt like a massive upgrade. Was the upgrade genuine, or a restoration?
"If Anthropic nerfed 4.5 for a few weeks and released a normally-functioning 4.6, we aren't actually comparing 4.5 to 4.6. We don't even know what we're comparing to." — u/ThePurpleAbsurdist, r/ClaudeCode [Link]
- Intriguingly, Boris Cherny's "most productive month ever" was December — exactly when the community also reported peak Opus 4.5 performance. Coincidence is possible. Proof is absent.
The Transparency Gap
- The sharpest critique from heavy users is not about capability but about trust:
"Stability, predictability, consistency are important features for serious work, and people don't talk about it enough. And Codex seems decidedly ahead on all of them." — u/m0j0m0j, r/ClaudeCode [Link]
- OpenAI provides model version numbers. Anthropic does not. Users cannot distinguish between a genuine regression and a bad inference batch. This is not a capability problem — it is a trust problem that drives real users to competitors.
MCP Atlas Regression
- While most benchmarks improved or held steady, MCP Atlas — measuring complex multi-tool coordination — dropped from 62.3% to 59.5%. [Link] For power users who chain multiple MCP servers, this is worth monitoring. The trade-off appears to be: deeper reasoning at the cost of slightly less nimble tool orchestration.
The Writing Question
The Every.to team (CEO Dan Shipper + 4 testers) ran Opus 4.6 through real-world tasks and produced the most nuanced dual verdict of this release. [Link] On the coding side, Shipper submitted a merged PR to a codebase he had never touched — Opus 4.6 researched the unsolved iOS issue, developed a fix, and shipped it. On the writing side, the team preferred Opus 4.5's prose in a blind test — describing 4.6 as introducing more "AI-isms," citing patterns like "X not Y" constructions as telltale artifacts.
The broader community shows no consensus. Reddit and HN threads are roughly split between "worse," "better," and "no difference." [Reddit] The emerging theory: RL optimization for coding reduced classic AI repetition patterns (the "bold, innovative, transformative" triplets), which some users perceive as improvement and others as regression. For code-heavy work, this is irrelevant. For technical writing, keep 4.5 on standby.
The Token Problem — and How to Stop Feeding the Hippo
- The single biggest complaint about Opus 4.6 is cost. It consumes roughly 1.5-2x the tokens of Opus 4.5 on identical tasks. [Reddit]
"On the 5x plan, blew through half my 5 hour window in 30 minutes. Same projects and prompts as before on Opus 4.5. This thing is a token hog." — u/RazerWolf, r/ClaudeCode [Link]
- The root cause is structural. Opus 4.6 ships with Adaptive Thinking engaged by default, meaning it applies extended reasoning even to trivial tasks. This is the sledgehammer problem made literal: the same reasoning force that nearly doubled ARC-AGI-2 scores also swings full-force at tasks that needed a screwdriver. Worse, it has been trained to be more "agentic" — so it instinctively decomposes simple tasks into subtasks and spawns subagents for each one.
"the fundamental issue is that 4.6 was trained to be more agentic, which means it defaults to 'let me break this into subtasks and delegate' even when the task is simple enough to just do. anthropic basically optimized for the hardest 10% of use cases at the expense of the easy 90%." — u/Bellman_, r/ClaudeCode [Link]
The Official Playbook: Boris Cherny's Approach
Boris Cherny, creator of Claude Code, shared his team's internal workflow in a series of X posts, later compiled by paddo.dev. [Link] The Reddit thread aggregating these tips hit 1,520 upvotes on r/ClaudeAI — the highest-engagement Opus 4.6-era post. [Reddit] Key insights beyond CLAUDE.md:
Run 3-5 parallel Claude sessions in git worktrees — described internally as "the single biggest productivity unlock." Cherny himself ran 5+ cloud agents simultaneously in December, shipping 300+ PRs in a single month — his most productive month in 1.5 years at Anthropic. [Link]
Invest in CLAUDE.md — "Every time you correct a mistake, tell Claude to update CLAUDE.md so it doesn't repeat it. Claude is eerily good at writing rules for itself."
Use subagents deliberately — adding "use subagents" to a request allocates more compute. Each subtask runs in its own context window, keeping the main agent's window clean.
Set output style via
/config—"Explanatory"or"Learning"styles make the model explain why it made changes, not just what changed.
Multi-Model Delegation
- A multi-model strategy significantly reduces total cost. Use Opus 4.6 for planning and architecture, then delegate implementation to cheaper models:
"With a good plan and tasks that are atomic, you can even use Haiku for implementation. This is a seriously slept on token economy hack. Haiku is FAR better than most people assume, it just needs a bit more specific instructions. And Opus 4.6 is happy to provide that." — u/xmnstr, r/ClaudeCode [Link]
Hooks: When CLAUDE.md Isn't Enough
- The community's meta-commentary on CLAUDE.md was sharp:
"Opinions on CLAUDE.md are the most divided. For some it's a game-changer, for others Claude completely ignores it. The general sentiment is 'it's like working with a genius who has dementia.' Community tip: use hooks for rules you really need enforced." — r/ClaudeAI TL;DR bot [Link]
- The intuition is correct: CLAUDE.md is a constitution, but Hooks are the enforcement mechanism. When Opus 4.6 transitions from planning to execution, it has a documented tendency to deprioritize written guidelines in favor of code-level reasoning. [Reddit] For rules that must never be violated — "do not touch this file," "always run tests before committing" — Hooks trigger shell commands at specific workflow events (pre-tool-call, post-tool-call, notification), making them structurally unbypassable by the model. [Link]
Monitoring: CCStatusLine
- Monitoring matters as much as constraint. CCStatusLine provides real-time token usage visibility directly in the CLI status bar, letting you see context consumption before it spirals. [Link] The community consensus: disable auto-compact, monitor manually, and invoke
/compactonly when you choose to.
"CCStatusLine is an indispensable addition to my workflow. Disable auto-compact and control it manually. Never let it work past the context limit — that is where the mistakes come from." — u/PlaneFinish9882, r/ClaudeCode [Reddit]
Opus 4.6 vs. GPT-5.3 Codex: The Dual-Wield Strategy
- The community has settled not on a winner, but on a workflow:
| Task Type | Primary | Reviewer | Why |
| Architecture & planning | Opus 4.6 | — | Superior big-picture reasoning |
| Complex builds from scratch | Opus 4.6 | Codex/Gemini 3 (review) | "Working plans" + independent verification |
| Single bug fix / debugging | Codex 5.3 | — | Faster, more laser-focused |
| Frontend UI | Opus 4.6 | — | Superior design quality |
| Code review | Codex 5.3 or Gemini 3 | — | Independent perspective |
| Large-scale refactoring | Opus 4.6 | Codex (review) | Proactive dead code removal + cross-check |
"Claude improvements = things Codex was better at (review). Codex improvements = things Claude was better at (steering). Both are absolute winners." — u/gopietz, r/ClaudeCode [Link]
"Don't be loyal to a model. Use CC, AG, Kiro, Google AI Ultra, Max, Powers+ — all of them, together, with fallback strategies. What matters is what you can do with those tools." — u/maraudingguard, r/ClaudeCode [Link]
Conclusion: Blueprints or Rubble
Anthropic's strategy with Opus 4.6 is legible now: a three-punch combination — Cowork (legal/finance automation) → Opus 4.6 (reasoning + coding agents) → Office integration (PowerPoint/Excel, "vibe working") — aimed at replacing entire categories of SaaS. [Link] This is not a model release. It is a platform play targeting every knowledge worker, not just developers.
The competitive landscape remains genuinely contested. GPT-5.3 Codex outperforms on Terminal-Bench and offers more predictable behavior. Gemini 3 Pro catches bugs that Opus 4.6 misses. The pricing gap is real — Anthropic's flagship has fallen from $15/$75 per MTok (Claude 3 Opus, 2024) to $5/$25 (Opus 4.6), a 3x reduction in two years [Link], but GPT-5.2 still undercuts at $1.75/$14. [Link] The smartest users are not choosing sides — they are building multi-model pipelines. The era of model loyalty is over; the era of model orchestration has begun.
And here is the fact that should give pause and excitement in equal measure: approximately 90% of Claude Code's own code is written by Claude Code. [Link] GitHub co-authored commits tagged with Claude currently account for roughly 4% of all public commits; SemiAnalysis projects this will surpass 20% by year-end. [Link] The self-referential loop — model improves → tool improves → next model accelerates — is no longer theoretical. But a satirical post written hours after launch offered the sharpest counterweight:
"A startup founder said: 'I have Claude, I don't need a dev team. I'll build it all myself.' Six months later, the founder had 40,000 lines of code, no tests, no documentation, an architecture only Claude understood — but Claude couldn't remember across sessions. The master said: 'You didn't build a product. You built a conversation that compiles.'" — u/didyousaymeow, r/ClaudeCode [Link]
- The philosopher still holds the sledgehammer. Opus 4.6 is the most powerful reasoning model available for agentic coding work — 1M context, proactive dead code removal, root-cause reasoning, lowest over-refusal in Claude history. But without CLAUDE.md discipline, subagent constraints, and multi-model delegation, it is just a very expensive way to burn tokens. The question is not whether the model is smart enough. It is whether the engineer holding it can hand it blueprints instead of rubble.
References
- Anthropic Official
- Tier 1 Tech Media
- https://techcrunch.com/2026/02/05/anthropic-releases-opus-4-6-with-new-agent-teams/
- https://venturebeat.com/technology/anthropics-claude-opus-4-6-brings-1m-token-context-and-agent-teams-to-take
- https://www.theverge.com/report/874308/anthropic-claude-code-opus-hype-moment
- https://www.zdnet.com/article/anthropic-claude-opus-4-6-first-try-work-deliverables/
- https://www.cnbc.com/2026/02/05/anthropic-claude-opus-4-6-vibe-working.html
- https://www.bloomberg.com/news/articles/2026-02-03/legal-software-stocks-plunge-as-anthropic-releases-new-ai-tool
- https://www.axios.com/2026/02/05/anthropic-claude-opus-46-software-hunting
- https://www.fastcompany.com/91488000/anthropics-new-claude-opus-4-6-aims-to-think-through-bigger-codebases
- https://www.reuters.com/business/retail-consumer/anthropic-releases-ai-upgrade-market-punishes-software-stocks-2026-02-05/
- Benchmarks & Technical Analysis
- https://www.vellum.ai/blog/claude-opus-4-6-benchmarks
- https://onllm.dev/blog/claude-opus-4-6 (independent verification status)
- https://www.rdworldonline.com/claude-opus-4-6-targets-research-workflows-with-1m-token-context-window-improved-scientific-reasoning/
- https://medium.com/@leucopsis/how-claude-opus-4-6-comapares-to-opus-4-5-c6b7502f43af (community analysis)
- https://www.cosmicjs.com/blog/claude-opus-46-vs-opus-45-a-real-world-comparison (real-world comparison)
- Cloud & Enterprise
- https://azure.microsoft.com/en-us/blog/claude-opus-4-6-anthropics-powerful-model-for-coding-agents-and-enterprise-workflows-is-now-available-in-microsoft-foundry-on-azure/
- https://cloud.google.com/blog/products/ai-machine-learning/expanding-vertex-ai-with-claude-opus-4-6
- https://github.blog/changelog/2026-02-05-claude-opus-4-6-is-now-generally-available-for-github-copilot/
- https://www.itpro.com/technology/artificial-intelligence/anthropic-reveals-claude-opus-4-6-enterprise-focused-model-1-million-token-context-window
- https://www.techspot.com/news/111151-apple-hidden-ai-partner-company-heavily-relies-anthropic.html (Apple internal Anthropic usage)
- https://www.macobserver.com/news/mark-gurman-reveals-why-apple-runs-on-anthropic-at-this-point/ (Mark Gurman report)
- Developer Resources
- https://github.com/anthropics/claude-code/issues/23499 (Bedrock 1M bug)
- https://github.com/ruvnet/claude-flow/issues/1082 (subagent analysis)
- https://paddo.dev/blog/claude-code-team-tips/ (Boris Cherny's 10 tips)
- https://every.to/vibe-check/opus-4-6 (independent expert review)
- https://laravel-news.com/claude-opus-4-6 (API breaking changes)
- https://dev.to/thegdsks/claude-opus-46-for-developers-agent-teams-1m-context-and-what-actually-matters-4h8c (developer guide)
- https://github.com/sirmalloc/ccstatusline (CCStatusLine token monitoring)
- https://github.com/Ammaar-Alam/minebench (3D VoxelBuild benchmark)
- https://www.datacamp.com/tutorial/claude-code-hooks (Hooks tutorial)
- Community Discussions
- https://www.reddit.com/r/ClaudeAI/comments/1qws1kc/introducing_claude_opus_46/ (official thread)
- https://www.reddit.com/r/ClaudeAI/comments/1qspcip/10_claude_code_tips_from_boris_the_creator_of/ (Boris tips)
- https://www.reddit.com/r/ClaudeCode/comments/1qx76jb/opus_46_is/ (use cases)
- https://www.reddit.com/r/ClaudeCode/comments/1qxazv9/codex_53_is_better_than_46_opus/ (Codex vs Opus)
- https://www.reddit.com/r/ClaudeCode/comments/1qwv8p1/opus_46_token_usage/ (token usage)
- https://www.reddit.com/r/ClaudeCode/comments/1qxhu30/46_agents_eat_up_tokens_like_theres_no_tomorrow/ (subagent spawning)
- https://www.reddit.com/r/ClaudeCode/comments/1qxhkt9/the_tao_of_claude_code/ (Tao of Claude Code)
- https://www.reddit.com/r/ClaudeCode/comments/1qxfh9l/hype_boys_with_skill_issues/ (engineering discipline)
- https://www.reddit.com/r/ClaudeAI/comments/1qx3war/difference_between_opus_46_and_opus_45_on_my_3d/ (3D VoxelBuild)
- https://www.reddit.com/r/ClaudeAI/comments/1qx31fd/refactoring_with_opus_46_is_insane_right_now/ (refactoring)
- https://www.reddit.com/r/ClaudeCode/comments/1qww3ly/thesis_it_is_impossible_for_us_to_vibetell_if/ (placebo/nerf debate)
- https://news.ycombinator.com/item?id=46902223 (HN main thread)
- https://news.ycombinator.com/item?id=46902909 (500 zero-day debate)
- https://www.reddit.com/r/ClaudeCode/comments/1qxfprh/gsd_vs_superpowers_vs_speckit_what_are_you_using/ (CCStatusLine workflow tip)
- https://www.reddit.com/r/ClaudeCode/comments/1qxgvnj/the_one_thing_that_frustrates_me_the_most/ (Hooks vs CLAUDE.md enforcement)
- https://www.reddit.com/r/ClaudeCode/comments/1qwuqk9/we_tasked_opus_46_using_agent_teams_to_build_a_c/ (C compiler with Agent Teams)
- https://www.reddit.com/r/ClaudeCode/comments/1qx5s4s/i_had_claudes_agent_teams_implement_a_large/ (Agent Teams 19 issues)
- Industry Analysis
- https://newsletter.semianalysis.com/p/claude-code-is-the-inflection-point (SemiAnalysis GitHub commit projection)
- https://ai-rockstars.com/claude-opus-4-6/ (senior engineer comparison)
- https://gdsks.medium.com/i-gave-claude-opus-4-6-my-ugliest-codebase-it-didnt-just-fix-it-8a26c3f6d488 (pricing history analysis)
- https://the-decoder.com/openai-opens-gpt-5-2-codex-to-developers-through-the-responses-api/ (GPT-5.2 pricing reference)



