Titikey
HomeTips & TricksClaudeClaude API New Feature Breakdown: Sonnet 3.5 Extended Output & Console Workbench Upgrade Guide

Claude API New Feature Breakdown: Sonnet 3.5 Extended Output & Console Workbench Upgrade Guide

2/23/2026
Claude

If you’re using the Claude API for customer support, content generation, or code analysis, this recent wave of updates is very practical: Claude Sonnet 3.5’s maximum output has been expanded, and the Claude Console Workbench has filled in prompt generation and evaluation capabilities. Below, from a “ready to use right away” perspective, I’ll clearly walk through the key new features of the Claude API in one go.

Sonnet 3.5 Extended Output: from 4096 to 8192

In the Claude API, Claude Sonnet 3.5’s maximum output token limit has doubled from 4096 to 8192, which is especially friendly for long-form summarization, long code explanations, and drafting batches of emails. To enable extended output, you need to add a beta request header to your request, rather than only changing the max_tokens parameter.

The official approach is: when calling the Claude API, include the request header "anthropic-beta": "max-tokens-3-5-sonnet-2024-07-15". It’s recommended that you first validate response length and cost changes on non-critical traffic, then gradually ramp up, to avoid cost fluctuations caused by maxing out long outputs all at once.

Claude Console Workbench Upgrade: the Prompt Generator saves more time

The Claude Console Workbench has added a “Prompt Generator.” You simply describe the task objective (for example, “classify and handle inbound customer support requests”), and it will help you produce a more complete prompt structure. For teams without dedicated prompt engineering, this is like compressing drafting time from “hours” down to “minutes.”

More importantly, it makes Claude API adoption more standardized: for the same task you can quickly get a “reusable template,” and later you only need to replace variable fields. You can also migrate the generated prompt directly into code, reducing inconsistencies between offline and online setups.

Evaluation Mode Launch: side-by-side comparison of prompt outputs, reducing guesswork

The Workbench has added “Evaluation Mode,” which supports comparing the outputs of two or more prompts side by side and rating Claude’s output on a 5-point scale. This feature is critical for iterating on Claude API prompts because it turns “it feels better” into a comparable result that can be recorded and traced back.

In practice, you can keep the same set of input samples fixed, test different instructions, different formatting requirements, or different constraints, and then standardize the high-scoring version as a team baseline. This approach often improves output quality more reliably than blindly making prompts longer.

Usage & Cost Dashboard: more intuitive tracking by dollars, tokens, and API key

The developer console has added “Usage” and “Cost” tabs, allowing you to track Claude API usage and billing by dollar amount, token count, and API key. For teams running multiple projects and multiple environments (test/production) in parallel, this makes it faster to pinpoint “who exactly blew through the usage.”

It’s recommended to split API keys by business line and regularly review alongside this dashboard: which endpoints have overly long output-side token usage, and which scenarios are better suited to shorter response formats. Shifting cost management earlier is usually more effective than reconciling accounts at month-end.

Release Notes & Learning Resources: Claude API docs are easier to find

The Claude API documentation now includes more comprehensive release notes, covering updates across the API, the Claude Console, and the Claude app, making it easier for you to keep version change records. At the same time, Anthropic has also launched new educational courses (Claude API Fundamentals, Using Claude Tools) and expanded the Claude Cookbook, adding core skill guides such as citations, RAG, and classification.

If you’re integrating the Claude API into an existing system, it’s recommended that you first use the courses to get the request structure, tool connections, and JSON structured output working end to end, and then use the Cookbook patterns to implement it in an engineering-ready way—overall it will be more stable.

HomeShopOrders