Claude has recently taken a big step beyond “just chatting” by adding a public beta “computer use” capability. Put simply, you can have Claude view the screen, move the cursor, click buttons, and enter text—completing tasks the way a person operates a computer. This article explains Claude’s new feature from an editor’s perspective: what it can do, how to use it, and who it’s for.
What exactly has been updated in Claude “computer use”
The highlight of this update is that Claude now offers a “computer use” capability on the API side, allowing developers to direct Claude to navigate UIs. Claude makes judgments based on what’s on the screen and then performs actions such as clicking, typing, and jumping between pages, chaining together steps that previously required manual work. It’s worth noting that Claude officially states this is still in an experimental stage, and it may occasionally lag, click the wrong thing, or behave inconsistently across steps.
Where can you access this capability
At the moment, “computer use” is available in beta via the Anthropic API, making it easier to integrate Claude into automation workflows or internal tools. Claude also supports building similar capabilities on Amazon Bedrock and Google Cloud Vertex AI, which should make enterprise deployment smoother. Meanwhile, the upgraded Claude 3.5 Sonnet is already available to all users, with a particular emphasis on improved coding performance.


