This Claude API update mainly addresses three things: longer answers, faster debugging, and more transparent costs. Below, following the most common developer workflow, we break down Claude API’s new features clearly and provide key invocation tips you can copy directly.
Claude API Extended Output: Sonnet 3.5 Maximum Output Doubled
In the Claude API, Claude Sonnet 3.5’s maximum output token limit has been raised from 4096 to 8192, making it suitable for longer summaries, code generation, and multi-step explanations. To enable extended output, you need to add the specified beta request header to your request.
The approach is straightforward: when calling the Claude API, add the request header anthropic-beta: max-tokens-3-5-sonnet-2024-07-15, and set max_tokens to the value you need. It’s also recommended to define a clear output structure (e.g., bullet points, paragraphs, JSON fields); otherwise, longer output won’t necessarily be easier to read.
Console Workbench Upgrade: The Prompt Generator Is Better for “Quick Drafting”
The Claude Console workbench now includes a prompt generator. You only need to describe the task in one sentence (e.g., “classify and handle inbound customer support requests”), and it will produce a more complete prompt template. For those who frequently write system instructions and need standardized output formats, this can save a lot of back-and-forth trial and error.
A more practical workflow is: first, let the prompt generator produce a “runnable” version, then add your real constraints—such as field validation, failure fallbacks, output length, and language style. Finally, paste the finalized prompt back into the Claude API’s system message or the first user instruction.
Evaluation Mode: Compare Prompts Before Using the Claude API
The workbench’s evaluation mode supports side-by-side output comparisons for two or more prompts, and lets you score results on a 5-point scale. It’s especially useful for A/B testing “same task, different wording”: for example, for the same field extraction task, one prompt may be more robust while another is more concise.


