Titikey
HomeTips & TricksClaudeClaude New Feature Roundup: How to Use Longer Outputs, the File API, and the Usage Dashboard

Claude New Feature Roundup: How to Use Longer Outputs, the File API, and the Usage Dashboard

2/28/2026
Claude

If you’ve been using Claude recently, you’ve probably noticed it no longer just “chats better,” but is moving toward becoming a practical, deployable work assistant. This article breaks down several of Claude’s most useful new features: the increased long-output limit, the File API, prompt caching, and the usage and cost dashboards in the console. After reading, you’ll know which tasks Claude is best suited for—and how to use it more smoothly.

Longer Claude outputs: long-form text and code are no longer frequently cut off

On the API side, Claude Sonnet 3.5 increased the maximum output from 4,096 to 8,192 tokens, making it suitable for writing more complete documentation, generating longer test cases, or producing a more usable code framework in one go. To enable extended output, you need to add Anthropic’s official beta request header to your request (the docs give "anthropic-beta": "max-tokens-3-5-sonnet-2024-07-15"). For content teams, the most direct benefit of longer output is fewer follow-up prompts in multiple chunks and a more coherent structure.

Claude File API: keeps long tasks “connected,” more like a real assistant

As Claude emphasizes continuity for “long-running tasks” in its new models and agent capabilities, one key component is the File API. The File API allows Claude to read and write files, saving key progress, constraints, and intermediate conclusions to prevent details from being lost as a conversation drags on. For tasks like requirement breakdowns, code migrations, and information organization, using the File API helps Claude turn “intermediate artifacts” into reusable assets.

Prompt caching upgrade: running large repeated contexts on Claude is cheaper and faster

For many teams, the most expensive part of using Claude isn’t a single generation, but repeatedly calling it with the same long background context. Prompt caching stores this kind of “fixed context.” With the TTL increased from 5 minutes to 1 hour, it becomes much friendlier for long conversations or long workflows. Anthropic says costs can be reduced by up to 90% and latency by 85%, which is especially noticeable for Claude workflows that require frequent iteration.

New usage and cost dashboards in the console: Claude bills finally make sense

With the new “Usage” and “Costs” tabs in the developer console, Claude’s API usage can be tracked and broken down by dollar amount, token count, and API key. This feature may look simple, but it’s extremely practical: you can quickly pinpoint which pipeline or which key is “burning tokens,” and then decide whether to enable caching, compress context, or adjust output length. Combined with the newly added release notes, Claude’s updates are also easier for teams to keep in sync.

HomeShopOrders