Morning Overview

The 1 thing Claude does better than any AI and how to try it

Anthropic launched Claude 3.7 Sonnet with a standout feature it emphasizes as a differentiator: extended thinking, a user-selectable mode that shows additional reasoning detail before the model delivers a final answer. The feature sits at the intersection of two growing demands in AI, the push for better accuracy on hard problems and the call for transparency in how models reach their conclusions. For anyone who has ever wondered what happens inside the black box between a question and an answer, Claude now pulls back the curtain.

What Extended Thinking Actually Does

Most AI chatbots produce responses in a single pass. A user types a question, the model generates text token by token, and the answer appears as a finished product. Extended thinking breaks that pattern. When enabled, Claude produces extra “thinking” output that can make parts of its intermediate reasoning visible before assembling its response. According to Anthropic’s product announcement, the feature is designed as user-selectable deeper reasoning, meaning it is not always on by default. Users choose when they want the model to think harder, which makes it practical for complex coding, analysis, or multi-step logic problems without slowing down simple queries.

The distinction matters because transparency in AI reasoning has been a persistent gap across the industry. Other models may use chain-of-thought prompting internally, but they rarely expose that process to the end user in a structured, readable way. According to Anthropic’s developer documentation, thinking output can vary by model and configuration, and in some cases may be summarized rather than shown in full. That can give users and developers more visibility into how the model arrived at its conclusions, not just what those conclusions are.

The Safety Framework Behind the Feature

Extended thinking did not emerge in a vacuum. It builds on a research foundation Anthropic has been developing since at least 2022, when the company published a paper on Constitutional AI methods. That work describes a training method in which AI systems critique and revise their own outputs against a set of principles, reducing harmful or dishonest responses without relying entirely on human reviewers for every interaction. The paper describes a training approach Anthropic has used to explain how it aligns Claude’s behavior in public communications.

The connection between Constitutional AI and extended thinking is worth spelling out. When Claude reasons visibly, users can check whether the model’s intermediate steps reflect the kind of self-correction the alignment framework is designed to produce. If Claude catches a flawed assumption mid-thought and adjusts, that is the training working in real time, and the user can see it happen. This is a different value proposition than simply promising a model is safe. It gives users evidence they can evaluate on their own terms, which is especially relevant for professionals in fields like law, medicine, or finance where the reasoning behind a conclusion can matter as much as the conclusion itself.

How to Turn It On

Trying extended thinking requires no special access or technical setup. Inside the Claude chat interface, users can find a toggle that activates the feature. One practical detail: switching to extended thinking mode starts a new chat, so it cannot be activated mid-conversation. Once enabled, the interface displays a “Thinking” indicator while Claude works through its reasoning. After the thinking phase completes, an expandable section lets users review the thinking output the model provides.

There are a few limits to be aware of. Some thinking content may be truncated due to safety and usage-policy gating, according to Anthropic’s help documentation. That means users will not always see every single reasoning step, particularly if the model’s internal process touches on sensitive topics. For developers building on the API, the feature uses dedicated thinking content blocks with token budget controls, allowing fine-grained management of how much computational effort goes into the reasoning phase. The best use cases are problems where a wrong answer carries real consequences: debugging code, working through a legal argument, or verifying a chain of mathematical steps where each link needs to hold.

Why Visible Reasoning Changes the Equation

The standard criticism of AI assistants is that they are confident but opaque. A model can produce a polished, grammatically perfect answer that is completely wrong, and the user has no way to diagnose where the logic broke down. Extended thinking addresses this directly by making the reasoning auditable. If Claude makes a factual error in step three of a ten-step analysis, a user reviewing the thinking section can catch it there rather than discovering the mistake only in the final output. This shifts the relationship between user and AI from blind trust to informed verification, especially when users are willing to read and challenge the intermediate steps instead of treating them as infallible.

That shift has practical consequences for adoption in professional settings. Organizations evaluating AI tools for high-stakes work often cite explainability as a requirement, not a nice-to-have. In some regulated or high-stakes environments, teams may prefer decision-support tools that can show more of their work. A model that surfaces its reasoning process, even imperfectly, is easier to audit and defend than one that operates as a sealed unit. The tradeoff is speed: extended thinking takes longer than a standard response, and the additional token usage means higher costs for API users who pay per token. For quick factual lookups or casual conversation, the feature adds overhead without much benefit. The payoff comes on hard problems where getting the answer right the first time saves more time than the thinking phase costs.

What the Industry Still Lacks

For all its promise, extended thinking has not yet been subjected to independent, third-party benchmarks measuring whether visible reasoning actually reduces error rates compared to opaque models. Anthropic positions the feature as enabling deeper reasoning, but no publicly available study from an outside lab has quantified the accuracy improvement on standardized tasks. That gap matters. Until organizations such as academic AI research groups or independent testing bodies publish comparative results, the case for extended thinking rests on Anthropic’s self-reported data and on individual users’ qualitative experiences across different problem domains.

The absence of independent validation does not diminish the feature’s design logic, but it does mean users should treat it as a tool for their own verification rather than an automatic guarantee of correctness. A visible reasoning chain that contains a flawed premise is still wrong, just more inspectable. Extended thinking is best understood as shifting some responsibility back to the user: the model offers a trace of how it got from question to answer, and the human decides which parts to trust, which to question, and where to bring in outside sources. In that sense, Claude 3.7 Sonnet’s experiment with transparent reasoning is less about replacing human judgment and more about giving people enough visibility into the machine’s thought process to exercise that judgment effectively.

More from Morning Overview

*This article was researched with the help of AI, with human editors creating the final content.