Every model release is a re-pricing event for endpoint policy, and Claude 5’s Computer Use 2.0 has just marked the book.

Why Claude 5 Computer Use Matters Now

Anthropic’s Claude 5 ships Computer Use 2.0 with finer-grained permission scopes and a measurably tighter action latency. For builders, that’s a feature release. For CISOs, it’s a re-pricing event — the controllable surface area on every endpoint just expanded, and the speed at which an agent resolves a plan into a chain of OS-level actions just compressed.

The mistake most teams will make is treating this as a vendor announcement. It is not. It is a structural shift in where the perimeter actually sits. When the model’s plan is encoded as a sequence of mouse, keyboard, and filesystem operations issued through a system-level scope, prompt guardrails are mark-to-market only at the moment of inference. By the time those actions reach the kernel, the plan that produced them is gone — and so is your audit trail.

Metric Pre-Claude-5 baseline Claude 5 / Computer Use 2.0
Permission granularity at the model Coarse (one “computer use” toggle) Per-tool, per-session scope
Median end-to-end action latency ~3–5s ~1–2s
Agents per dev endpoint (Ospiri telemetry) 8–15 Trending higher each release
Window for human-in-the-loop intervention Seconds Sub-second

The first two rows paraphrase Anthropic’s release notes and developer documentation. The bottom two are Ospiri’s signature-pipeline observations — the same fleet research where the median enterprise dev fleet of 1,000 endpoints surfaces 8–15 distinct AI agents, most unsanctioned.

Computer Use 1.0 vs Computer Use 2.0: The Control-Plane Delta

Computer Use 1.0 was a binary capability — the agent had a desktop session or it didn’t. The governance granularity that mattered lived in your IAM and your endpoint controls, not in the model. Computer Use 2.0 changes the contract, and that changes where your control plane has to live.

Dimension Computer Use 1.0 Computer Use 2.0 Governance implication
Permission scope Session-level Per-tool, per-session Kernel policy must match model granularity
Plan visibility Visible in prompt trace Resolves into raw actions before the OS Prompt guardrails lose telemetry at the boundary
Action speed Operator-watchable Faster than human glance UEBA windows must shrink to sub-second
Reversibility Forward-only Same — but more frequent, more parallel Copy-on-write becomes table stakes
Audit surface Inference logs Inference logs + raw OS calls Correlation must happen at the kernel
Deployment shape Researcher / power user Embedded in standard dev workflows Inventory drift accelerates

The unlock is real for builders. The exposure is real for security teams. Both are true at the same time.

The Failure Mode: Plan Loss at the OS Boundary

Every prompt-layer guardrail — Lakera, Protect AI, GuardrailsAI, internal model-vendor classifiers — operates above the OS. They see prompts and tool-call schemas. They do not see the actual filesystem operation that lands a few hundred milliseconds later. With Computer Use 2.0, the distance between “the model decided to do X” and “X happened on the disk” is now small enough that human review is fictional unless engineered for.

The pattern we see in the signature pipeline:

  1. Plan resolves at the model. The agent translates the user instruction into a sequence of Computer Use tool calls.
  2. Plan dissolves at the OS boundary. Tool calls are issued as discrete syscalls and UI events. Whatever semantic intent existed at the prompt layer is no longer attached to the operation by the time it hits write() or unlink().
  3. Existing controls miss the connection. EDR sees a process making syscalls. SIEM sees the syscalls. Neither knows it was an agent acting on behalf of a user, and neither can attribute the action back to the prompt that produced it.
  4. Forensics becomes reconstruction. You piece the chain together from inference logs and OS telemetry post-incident. If your inference provider is upstream of you, you may not have the inference half at all.

This isn’t theoretical. The agent-driven repository wipes that surfaced over the last 18 months — codebases force-deleted by agents that “thought they were helping” — fall squarely in this pattern. The agent’s plan was reasonable in context. The action was irreversible. The prompt guardrail was satisfied. The kernel had no idea anything unusual was happening.

Pricing the Risk: Permission Scope × Reversibility

So, what’s the moral. Compress the calculation.

Risk Score = (Permission Scope × Reversibility) + (Action Frequency × Drift Coefficient)

Factor Low-score example High-score example
Permission Scope Read-only browser session Per-tool computer-use scope including filesystem write and shell
Reversibility Drafts saved to a sandboxed dir rm -rf, force-push, account deletion
Action Frequency A handful of actions per session Continuous, parallel, sub-second tool calls
Drift Coefficient Same syscalls, same dirs, same hours New syscalls, new directories, off-hours, new endpoints

Computer Use 2.0 raises the first factor across the fleet by default. The other three are properties of how the agent is deployed. If you want to mark the change to your portfolio in a single sentence: every fleet that adopts Claude 5 with Computer Use 2.0 marks its agent governance book higher this quarter, whether it observes the move or not.

What This Requires Architecturally

The control point has to match the granularity of the capability. Per-tool permission scope at the model demands per-tool policy at the OS.

Control point What it sees What it can enforce Fit for Computer Use 2.0
Prompt guardrails (Lakera, Protect AI) Prompts and tool schemas Block unsafe prompts pre-inference Necessary, not sufficient
EDR (CrowdStrike, SentinelOne, Defender) Process and syscall telemetry Kill processes; quarantine Sees the action, not the agent
DLP (Symantec, Forcepoint, Microsoft Purview) Data movement Block exfil paths Misses agent-internal filesystem operations
CASB SaaS API calls Block sanctioned-app misuse Blind to local filesystem
Agent firewall (kernel-scope) Per-tool scope, per-action context Copy-on-write, segmentation, scope-limited execution Designed for this control surface

The distinction we keep coming back to is block-on-deny vs copy-on-write. Block-on-deny tools — the historical pattern from DLP and EDR — get ripped out within two quarters because engineering teams revolt. Copy-on-write preserves the agent’s productivity while bounding the blast radius. That is the architectural lesson the 2026 fleet keeps re-learning.

What CISOs Should Do This Quarter

Step Action Output Effort
1 Inventory which endpoints in your fleet have Claude 5 / Computer Use 2.0 enabled Current count, broken out by team 1 week
2 Map the per-tool scopes your developers are granting at the model layer A scope matrix you can hand to your auditors 2 weeks
3 Pilot kernel-scope policy on one team — enforce per-tool scope at the OS and mirror it to the model scope Working policy for one cohort 4–6 weeks
4 Mark agent-risk-score quarter-over-quarter, including the Computer Use 2.0 delta Quarterly board metric Ongoing

If you don’t have an agent inventory yet, step one is where the budget conversation usually changes. The CISOs we work with typically reset their thinking in the first 30 minutes after seeing the inventory.

The Bottom Line

Claude 5’s Computer Use 2.0 is the moment endpoint policy stops being a model-vendor problem and becomes a kernel-scope problem. Every model release re-prices the book; this one re-prices the control plane. Prompt guardrails see the plan. EDR sees the action. Neither sees the agent. The architectural answer is a Claude firewall that enforces per-tool scope at the OS — not because prompts are wrong, but because the plan is gone by the time it lands at the kernel.

If your team is sizing this for the FY26 Q3 budget cycle, request a working session. We will walk through your environment, build the per-tool scope matrix for your Claude 5 deployments, and scope a kernel-scope pilot. 90 minutes.