What Is Computer Use?
Computer Use is a feature from Anthropic that lets Claude perceive a screen through screenshots and take actions on it — clicking the mouse, typing on the keyboard, opening applications, and navigating multi-step graphical workflows just like a human operator. It first shipped as a public beta in October 2024 and matured into a production-ready capability during 2025 and 2026. With the release of Claude Sonnet 4.6, Computer Use scored 72.5% on the OSWorld benchmark, making it the state-of-the-art general-purpose GUI automation model.
In one sentence, Computer Use is “the API that turns Claude into a virtual teammate who can drive a PC”. Where a traditional LLM only returns text, a Computer Use-enabled Claude can open a browser, fill in a form, download a file, inspect the result, and decide what to do next. You should think of it as a digital worker you can rent by the hour via an API call, not as a faster chatbot. Important: this capability is especially valuable for workflows that resist classic scraping and RPA, such as post-login SaaS screens, legacy desktop applications, and cross-application routines that a pure browser automation tool cannot handle.
Computer Use represents a larger industry shift toward “agentic AI”, where language models no longer just speak but actively take actions in the real digital world. Keep in mind that the skill set Claude needs for this is very different from chat: it must reason about spatial layout, understand icons without text, track what changed on the screen between two steps, and recover when an unexpected dialog appears. Note that these are precisely the things human users do intuitively, and they are notoriously hard for classic automation scripts to replicate.
How to Pronounce Computer Use
kum-PYOO-ter yoos (/kəmˈpjuːtər juːs/)
How Computer Use Works
Computer Use is implemented as a perception-to-action loop that combines Claude’s vision capabilities with its Tool Use mechanism. When a user sends a request, Claude first receives a screenshot of the current state of the virtual machine, visually reasons about what is on the screen, and then decides the next action. It calls one of the dedicated tools (the Computer Tool, the Text Editor Tool, or the Bash Tool) to perform the action, and the application code executes that action on the target VM. The resulting new screenshot is then fed back as a tool result, and Claude iterates — deciding, acting, and re-observing — until the task is complete.
The key technical design choice is that Claude sees the screen as an image and specifies mouse coordinates numerically, rather than relying on HTML or an accessibility tree. This is a fundamental difference from browser-only agents: because the model does not depend on DOM structure, Computer Use works on native desktop applications, Canvas-drawn UIs, old Flash-style interfaces, and remote-desktop sessions that offer no DOM at all. Important: this makes Computer Use the most general automation primitive available today, at the cost of higher token usage per task.
The Perception-to-Action Loop
Computer Use core loop
take screenshot
Claude picks next action
click / type / shell
new screenshot back
The loop is driven on the client side — meaning the developer must implement “take a screenshot” and “run an action” functions in the caller application, and invoke them according to Claude’s tool output. Anthropic ships an official Docker-based reference implementation with Ubuntu, Firefox, and Xvfb pre-configured, and you should start from there rather than writing the entire harness from scratch. Keep in mind that producing a safe, sandboxed environment for an LLM to control is just as important as the model’s accuracy.
Three Core Tools
Computer Use is not a single tool but three complementary ones that Claude picks between based on the task.
- Computer Tool: performs mouse actions (move, click, drag), keyboard actions (type, hotkey), and takes screenshots. This is the primitive that covers any GUI interaction.
- Text Editor Tool: lets Claude read, create, and edit files directly, without having to navigate a text editor UI. It is much faster than driving a GUI editor for multi-file code changes.
- Bash Tool: runs shell commands for file operations, process management, or any task that would be slower via mouse and keyboard. Important: Bash Tool is typically safer to use than GUI navigation for deterministic tasks like copying files.
You should register all three tools in the same request so that Claude can switch between them naturally. Note that Claude is trained to pick the right tool per step — for instance, it will often use Bash to create a directory rather than open a file manager, because it is faster and more reliable. Keep in mind that this meta-skill of “picking the right tool” is one of the major reasons why Sonnet 4.6 outperforms earlier Claude models on OSWorld by a substantial margin.
Key Specifications
| Item | Value |
|---|---|
| Vendor | Anthropic |
| Tool type string | computer_20251124 |
| Beta header | computer-use-2025-11-24 |
| Supported models | Claude Sonnet 4.6 / Claude 3.7 Sonnet / Claude 3.5 Sonnet v2 |
| OSWorld score | 72.5% (Sonnet 4.6) |
| Initial release | October 2024 (beta) |
| Production readiness | Progressive GA through 2025-2026 |
| Platforms | Anthropic API / Amazon Bedrock / Google Cloud Vertex AI |
| Automation targets | Any desktop, browser, terminal, legacy GUI |
| Recommended environment | Sandboxed VM or container |
Computer Use Usage and Examples
To use Computer Use, you call the Anthropic API with a tool definition that includes the computer_20251124 type, along with the beta header. The response from Claude will contain a sequence of tool-use calls that your application must execute against a virtual machine, feeding the resulting screenshots back into the next API call. You should treat the entire exchange as a conversation where Claude is the pilot and your code is the plane’s controls.
Minimal Python Example
import anthropic
client = anthropic.Anthropic()
response = client.beta.messages.create(
model="claude-sonnet-4-6",
max_tokens=1024,
tools=[
{
"type": "computer_20251124",
"name": "computer",
"display_width_px": 1024,
"display_height_px": 768,
},
],
messages=[{"role": "user", "content": "Save an image to Downloads"}],
betas=["computer-use-2025-11-24"],
)
print(response.content)
This returns a list of content blocks that include tool-use calls such as “click at (450, 320)” or “type ‘filename'”. Your code must execute those actions on the target VM, capture the new screenshot, and send it back as a tool result. Important: always pass the beta header in the betas list or the request will be rejected.
Full Three-Tool Setup
tools = [
{
"type": "computer_20251124",
"name": "computer",
"display_width_px": 1280,
"display_height_px": 800,
},
{
"type": "text_editor_20250429",
"name": "str_replace_based_edit_tool",
},
{
"type": "bash_20250124",
"name": "bash",
},
]
response = client.beta.messages.create(
model="claude-sonnet-4-6",
max_tokens=4096,
tools=tools,
messages=[{
"role": "user",
"content": "Fix typos in README.md, commit the change, and push to origin."
}],
betas=["computer-use-2025-11-24"],
)
With all three tools registered, Claude will happily use Bash to check git status, the Text Editor Tool to patch the file directly, and fall back to the Computer Tool only when the task actually needs a GUI. You should register all three tools unless you have a specific reason not to, because the resulting agent is significantly faster and cheaper.
Screenshot Loop Pseudocode
messages = [{"role": "user", "content": "Open Downloads and show the latest PDF"}]
while True:
resp = client.beta.messages.create(
model="claude-sonnet-4-6",
max_tokens=1024,
tools=tools,
messages=messages,
betas=["computer-use-2025-11-24"],
)
if resp.stop_reason == "end_turn":
break
tool_use = next(b for b in resp.content if b.type == "tool_use")
run_action_on_vm(tool_use.input)
screenshot = take_screenshot()
messages.append({"role": "assistant", "content": resp.content})
messages.append({
"role": "user",
"content": [{
"type": "tool_result",
"tool_use_id": tool_use.id,
"content": [{"type": "image", "source": screenshot}],
}],
})
Here run_action_on_vm is your function that executes mouse and keyboard events through PyAutoGUI, xdotool, or an equivalent, and take_screenshot returns a fresh screenshot of the current VM state. Keep in mind that production deployments need to add timeouts, error recovery, session isolation, and structured logging on top of this minimal loop.
Advantages and Disadvantages of Computer Use
Advantages
- Works on GUI-only systems: covers SaaS without APIs and legacy desktop applications.
- Independent of DOM: handles Canvas, Flash-like, or remote desktop interfaces that classic automation cannot.
- Natural-language instructions: describe goals instead of recording click paths.
- Strong at cross-app workflows: gathers data in a browser and enters it into a desktop app in the same session.
- Resilient to UI changes: because it operates on meaning, not coordinates, minor redesigns rarely break it.
- Triple toolset: combining Computer, Text Editor, and Bash makes tasks dramatically faster than GUI-only automation.
- Observable: every action is logged and auditable, making it ideal for regulated environments once guardrails are added.
Disadvantages
- Screenshot-heavy loops use many tokens, so per-task cost is higher than a classic API call.
- Infrastructure (sandboxed VM, screenshot capture, input dispatch) takes real engineering effort.
- Latency per step is higher than deterministic RPA because the model reasons between each action.
- Security risk is non-trivial — the agent will follow whatever it sees on screen, so prompt injection is a real threat.
- Some features are still beta and may change, so you should wrap the API in your own abstraction layer.
Computer Use vs RPA
At first glance, Computer Use looks like a reinvented RPA tool, but the philosophies diverge sharply. Traditional RPA platforms such as UiPath, Automation Anywhere, or Blue Prism are built to replay recorded steps faithfully — they excel at high-volume, deterministic tasks, and they fail loudly when the UI changes. Computer Use is an AI agent that adapts to the screen it sees, which is the opposite trade-off: fewer guarantees per action, but far better coverage of messy, changing environments. You should plan for them to coexist rather than to compete outright.
| Aspect | Computer Use | Traditional RPA |
|---|---|---|
| How you instruct | Natural language | Recorder / GUI builder |
| Screen understanding | Visual + semantic | Coordinates / selectors |
| Resilience to UI change | High | Low |
| Unexpected dialogs | Can recover | Typically fails |
| Speed per step | Slower (reasoning time) | Very fast |
| Cost model | Pay-per-token | License per bot / seat |
| Best fit | Judgmental, cross-app tasks | High-frequency, fixed flows |
You should keep in mind that many organizations are now running hybrid architectures: the deterministic RPA bots handle the 90% of volume that never changes, while Computer Use handles the 10% of edge cases that used to require a human. Important: this hybrid model tends to minimize both cost and failure rate, and is becoming the dominant pattern in enterprise automation.
Common Misconceptions
Misconception 1: “Computer Use replaces RPA completely.”
Computer Use is not yet 100% reliable on any task, and its per-action cost exceeds RPA. For high-frequency, repetitive flows, RPA remains more stable and cheaper. You should think of the two as complementary, with RPA handling steady-state workloads and Computer Use stepping in for exceptions and judgment-heavy steps. Important: this hybrid design is what most successful enterprise deployments look like.
Misconception 2: “It works perfectly on every screen.”
An OSWorld score of 72.5% is impressive, but it also means roughly one in four tasks needs a retry or a human override. In production, you should design for retries, timeouts, and human-in-the-loop checkpoints. Keep in mind that critical steps — financial transactions, data deletion, production deployments — should never be fully delegated to any agent without human approval.
Misconception 3: “It is inherently unsafe and should not be used.”
When run in a sandbox with restricted permissions, Computer Use can actually be safer than a human operator because every action is logged and replayable. You should invest in isolation (separate VMs, network restrictions, read-only mounts) and in allow-lists for what the agent can touch. Note that Anthropic publishes specific guidance on prompt-injection defenses, such as avoiding untrusted web pages and never letting the agent log into a shared file-sharing service.
Misconception 4: “Once you have prompts, Computer Use just works.”
Building a Computer Use agent involves substantial engineering: provisioning VMs, orchestrating the perception-action loop, handling network timeouts, capturing logs, and designing retries. The Anthropic Docker reference implementation is a starting point, not a production system. Important: budget engineering time for the harness even if you do not budget any time for the model itself.
Real-World Use Cases
1. Automating Legacy Enterprise Systems
Many industries still run client/server applications that were built decades ago and never received a modern API. Computer Use is especially effective at automating these systems because it does not need HTML structure or keyboard shortcuts specified in advance — it simply looks at the screen like a human analyst would. In practice, this turns day-long data-entry jobs into overnight batch processes. Note that manufacturing and financial institutions, which still operate plenty of legacy GUIs, are leading adopters.
2. Expense Report and Invoice Processing
Reading a receipt image, entering the data into an expense SaaS, attaching evidence, and submitting the form is exactly the kind of image-plus-GUI flow that Computer Use excels at. It can even close out a month-end period by clicking dashboard buttons, freeing human accountants to focus on validation rather than data entry.
3. End-to-End QA Test Automation
Classic E2E test suites break whenever a button moves or a selector changes. With Computer Use, tests can be written as natural-language specifications such as “log in and verify that the monthly revenue summary loads”. Because the agent reasons semantically about the screen, it continues to pass even after small UI changes. You should keep in mind that flaky tests are one of the largest hidden maintenance costs in modern engineering organizations.
4. Customer-Support Copilot
Support agents often spend more time navigating CRMs and internal tools than actually talking to customers. Computer Use can take over the “look up the customer’s order history” or “open the relevant internal wiki” parts silently in the background, reducing handoffs and shortening response times dramatically.
5. Recruiting and HR Automation
Translating job postings, cross-posting roles to multiple job boards, scheduling interviews, and notifying candidates are fragmented tasks with no unified API. Computer Use stitches them together automatically, which lets recruiters focus on candidate relationships rather than data entry.
6. Competitive Intelligence Behind Logins
Logged-in pricing pages and members-only inventory data are notoriously hard to scrape because they require maintaining authenticated sessions on GUIs that actively resist automation. Computer Use can log in as an authorized researcher and capture the same information a human would see. Important: always confirm that your use complies with the target service’s terms of service before deploying this pattern.
7. Software Deployment and Release Automation
Some internal release tools still require a human to click through a series of dashboards and upload artifacts manually. A Computer Use agent, paired with Bash and Text Editor Tools, can automate this entire release rundown while still pausing at the critical “confirm deploy” button for human approval, combining speed with safety.
8. Production Best Practices and Tuning
When taking Computer Use from prototype to production, you should invest heavily in the harness that surrounds the model. A robust harness usually includes a screenshot buffer that downsamples large displays to the model’s preferred resolution, a deterministic action dispatcher that validates coordinates before executing them, and a structured logger that captures every screenshot, every tool call, and every response token for later audit. Important: the quality of your harness often matters more than the raw OSWorld score, because a brittle harness can turn an 80%-capable model into a 40%-reliable system.
Sandboxing is the single most important design decision. You should run Computer Use inside an ephemeral virtual machine that is destroyed after every task, with no network access to internal systems by default and no mounted volumes that contain sensitive data. When you do need network access, keep it on an allow-list of trusted hostnames rather than an open internet egress that a prompt injection could exploit. Keep in mind that a compromised Computer Use session, without sandboxing, can do anything the VM’s user can do on that machine.
Human-in-the-loop checkpoints are the second critical design element. For any task that touches money, production systems, personally identifiable information, or data deletion, you should insert a pause where a human reviews the plan before the agent executes destructive actions. This can be implemented as a simple webhook that notifies a reviewer on Slack and waits for approval, or as a structured approval queue tied to your identity platform. Important: even a 95%-reliable agent becomes a liability without review gates when the failure mode is financial or legal.
Cost control is the third pillar. You should cap the number of iterations per task — typical production ceilings are 30 to 80 steps — set a token budget per session, and use Prompt Caching for any shared system prompt that describes your environment. Note that with Prompt Caching a 2,000-token system prompt containing your task templates and safety rules can cost up to 90% less per call, which adds up fast at scale. You should also monitor average screenshot counts per task as an early indicator of regressions in your prompts, since a sudden jump often means the agent is getting stuck in a loop.
Finally, evaluation is non-negotiable. You should maintain a small, private suite of representative tasks that you can run against each new model version to detect regressions. Anthropic regularly improves the underlying models, and while most changes are beneficial, you need your own test harness to make informed decisions about when to upgrade. Keep in mind that publicly available benchmarks such as OSWorld do not fully represent your enterprise workflows, so building a 30- to 50-task internal eval is one of the highest-leverage investments your team can make when adopting Computer Use at scale.
Frequently Asked Questions (FAQ)
Q1. How is Computer Use priced?
A. It is billed like any other Anthropic API call — per input and output token. Each screenshot counts as an image input of roughly 1,000 to 1,500 tokens at 1024×768 resolution. A typical task uses 20 to 50 screenshots, so you should budget around $0.10 to $0.50 per task as a starting point. Keep in mind that enabling Prompt Caching can cut recurring system-prompt costs by up to 90%.
Q2. Which Claude models support Computer Use?
A. The latest and highest-scoring model is Claude Sonnet 4.6 at 72.5% on OSWorld. Older options include Claude 3.7 Sonnet and Claude 3.5 Sonnet v2. The Sonnet tier is the standard choice because it balances capability and cost. Note that Opus models also support Computer Use but are usually reserved for research-grade tasks because the per-call cost is much higher.
Q3. What are the security considerations?
A. You should always run the agent inside an isolated sandbox (a dedicated VM or container), restrict network access, disallow access to production secrets, and require human approval for destructive actions such as deletes, payments, or production deployments. Anthropic also publishes prompt-injection mitigation guidance — such as avoiding untrusted web content and blocking credentials access — which you should incorporate into your harness design.
Q4. Can I try it locally?
A. Yes. Anthropic provides an official Docker reference implementation that runs Ubuntu, Firefox, and Xvfb inside a container, so you can experiment on your laptop within an hour. Keep in mind that production workloads belong in isolated cloud VMs rather than on your developer machine, both for security and for reproducibility.
Q5. What is the roadmap for Computer Use?
A. Anthropic continues to push GUI understanding forward, with ongoing improvements in long-horizon task completion, multi-monitor support, and mobile device compatibility. Important: the long-term trajectory is an agent that can autonomously finish hours-long business workflows, so you should expect the line between RPA and Computer Use to blur further over the next year.
Conclusion
- Computer Use lets Claude see a screen and act on it via mouse, keyboard, and shell commands.
- It launched as a beta in October 2024 and matured through 2025-2026, with Claude Sonnet 4.6 hitting 72.5% on OSWorld.
- The API uses tool type
computer_20251124and the beta headercomputer-use-2025-11-24. - It provides three tools — Computer, Text Editor, and Bash — that Claude combines automatically for efficiency.
- Unlike traditional RPA, it is resilient to UI changes and thrives on judgment-heavy, cross-application tasks.
- Production deployments require sandboxing, human-in-the-loop checkpoints, and audit logging.
- Real-world use cases include legacy system automation, expense processing, QA testing, support copiloting, and recruiting.
References
References
- ・Anthropic API Docs “Computer use tool” https://platform.claude.com/docs/en/agents-and-tools/tool-use/computer-use-tool
- ・Anthropic “Introducing computer use, a new Claude 3.5 Sonnet, and Claude 3.5 Haiku” https://www.anthropic.com/news/3-5-models-and-computer-use
- ・AWS “Configure computer use for Anthropic’s Claude models” https://docs.aws.amazon.com/bedrock/latest/userguide/computer-use.html



































Leave a Reply