Category: Anthropic

News, analysis, and profiles covering Anthropic the company and its team.

  • Claude on Vertex AI: Why Route Through GCP Instead of Direct API

    Claude on Vertex AI: Why Route Through GCP Instead of Direct API

    Claude AI · Tygart Media
    Bottom line: Routing Claude through Google Cloud’s Vertex AI makes sense if you’re already on GCP, need enterprise compliance controls, want billing consolidated under your cloud account, or want to run Claude inside a private VPC. For individual users and small teams, the direct Anthropic API is simpler.

    Anthropic offers two ways to access Claude programmatically: directly through the Anthropic API, or through Google Cloud’s Vertex AI. They run the same models with the same capabilities. The difference is infrastructure, billing, compliance, and control. Here’s when each makes sense — and why teams running production AI workloads on GCP increasingly choose Vertex.

    What You Actually Get Through Vertex AI

    When you access Claude through Vertex AI, the request routes through Google Cloud infrastructure rather than Anthropic’s own endpoints. You get access to every Claude model — Opus 4.6, Sonnet 4.6, Haiku 4.5 — with the same capabilities including the 1M token context window on Opus and Sonnet. Nothing is stripped down. The key differences are on the infrastructure and billing side, not the model side.

    Five Reasons to Route Through GCP Instead of Direct API

    1. Consolidated GCP billing

    If your organization already runs on Google Cloud, adding Claude through Vertex AI means all AI spending appears on a single GCP bill. No separate Anthropic invoice, no separate API key management system, no separate budget approval process. For enterprise finance teams, this is often the deciding factor — Claude becomes a line item on the existing cloud budget rather than a new vendor relationship.

    2. Use existing GCP credits

    Google Cloud offers $300 in free credits to new accounts, startup credits through various programs, and committed use discounts for larger organizations. All of these apply to Claude usage through Vertex AI. Teams with unused GCP credit can run substantial Claude workloads at no incremental cost. New GCP accounts can effectively run Claude Code for free until credits are exhausted.

    3. IAM and access control

    Vertex AI integrates with Google Cloud IAM, meaning you can control who in your organization can access Claude using the same permission system you use for every other GCP service. Roles, service accounts, audit logs — all standard GCP tooling applies. This eliminates the need for a separate API key distribution system and makes access revocation immediate and centralized.

    4. VPC Service Controls and private networking

    For organizations with strict data residency or network isolation requirements, Vertex AI supports VPC Service Controls that prevent Claude API calls from leaving your private network perimeter. Claude requests originate from inside your GCP VPC rather than from an internet-facing endpoint. This is the core of what some teams call a “Fortress Architecture” — running AI inference inside a secured cloud environment where data never traverses the public internet. For regulated industries (healthcare, finance, legal), this is often a compliance requirement, not a preference. See The Fortress Architecture: Why Regulated Industries Need Their Own Cloud for the full architecture breakdown.

    5. Regional data residency

    Vertex AI lets you pin Claude requests to specific GCP regions — US, EU, or specific regional endpoints. For organizations subject to GDPR or other data residency requirements, this ensures AI processing stays within the required geographic boundary. The Anthropic direct API does not offer equivalent regional controls.

    When the Direct Anthropic API Is Better

    Vertex AI adds setup overhead — you need a GCP project, Vertex AI enabled, model access requested in Model Garden, and IAM configured. For individual developers, startups, and teams that don’t already run on GCP, this overhead isn’t worth it. The direct Anthropic API is faster to set up (generate a key, start calling), has the best rate limits for getting started, and doesn’t require cloud infrastructure knowledge.

    Also: new Claude models appear in the direct API before they appear in Vertex AI’s Model Garden. If you need day-one access to new releases, direct is faster.

    Pricing Comparison

    Model Anthropic Direct Vertex AI (Global) Vertex AI (Regional)
    Claude Opus 4.6 input $5/M tokens $5/M tokens +10% premium
    Claude Sonnet 4.6 input $3/M tokens $3/M tokens +10% premium
    Claude Haiku 4.5 input $0.80/M tokens $0.80/M tokens +10% premium

    Global endpoint pricing matches Anthropic direct. Regional endpoints add a 10% premium for the data residency guarantee. If you don’t need regional pinning, use the global endpoint and pay identical rates.

    Is Claude on Vertex AI the same as the Anthropic API?

    Same models, same capabilities, different infrastructure. Vertex AI runs on Google Cloud with GCP billing, IAM, and VPC controls. The direct Anthropic API is simpler to set up but lacks GCP-native enterprise controls.

    Can I use GCP free credits for Claude on Vertex AI?

    Yes. New GCP accounts receive $300 in free credits. Startup programs and other Google Cloud credits all apply to Claude usage through Vertex AI. Teams with existing GCP credits can run Claude workloads at no incremental cost until credits are exhausted.

    Is Claude on Vertex AI more expensive than the direct API?

    At the global endpoint, pricing is identical to Anthropic direct. Regional endpoints (for data residency) add a 10% premium. If you don’t need regional pinning, the cost difference is zero.

  • Current Claude Model Version Tracker — April 2026

    Current Claude Model Version Tracker — April 2026

    Claude AI · Tygart Media · Updated April 2026
    Latest models (April 16, 2026): Claude Opus 4.6 (claude-opus-4-6) and Claude Sonnet 4.6 (claude-sonnet-4-6) are current. Original Claude 4.0 models deprecated — retiring June 15, 2026.

    Anthropic releases model updates frequently and the naming can be confusing. This page tracks the current Claude model lineup, the exact API strings to use, what’s deprecated, and what’s coming next. Bookmark it and check back — it’s updated when Anthropic ships changes.

    Current Models (April 2026)

    Model API String Context Best For
    Claude Opus 4.6 claude-opus-4-6 200K (1M beta) Complex reasoning, long-horizon tasks, maximum capability
    Claude Sonnet 4.6 claude-sonnet-4-6 200K (1M beta) Production default — near-Opus performance at lower cost
    Claude Haiku 4.5 claude-haiku-4-5-20251001 200K Speed, cost efficiency, high-volume tasks

    Deprecated Models (Action Required)

    Model API String Retirement Date Migrate To
    Claude Sonnet 4 (original) claude-sonnet-4-20250514 June 15, 2026 claude-sonnet-4-6
    Claude Opus 4 (original) claude-opus-4-20250514 June 15, 2026 claude-opus-4-6

    If you have 20250514 in any API calls or model strings in production code, you have until June 15 to update them. Search your codebase for that date string now.

    What Changed From 4.0 to 4.6

    The Claude 4.6 models (released February 2026) are meaningful upgrades over the original 4.0 release (May 2025). Key improvements in Sonnet 4.6: near-Opus-level performance on coding and document comprehension, dramatically improved computer use (navigating browsers, filling forms, operating software), better instruction-following with fewer errors, and the 1M token context window in beta. Opus 4.6 adds the same 1M context with additional improvements to long-horizon reasoning and multi-step agentic tasks.

    Model Naming: How It Works

    Anthropic uses a generation.version format. The “4” is the major generation (fourth architecture generation). The “.6” is a version increment within that generation — a meaningful capability update without a full architecture change. Haiku, Sonnet, and Opus are tiers within each generation: speed/cost, balanced, and maximum capability respectively. The date suffix in API strings (like 20250514) is the training cutoff snapshot used for that specific release.

    What’s Coming Next

    Claude 5 is expected Q2-Q3 2026 based on Anthropic’s release cadence. No official announcement as of April 2026. Early signals from Vertex AI logs suggested a “Fennec” codename for Claude 5 Sonnet. As always with Anthropic releases, assume the new Sonnet tier will outperform current Opus on most benchmarks at a lower price point. See Claude 5 Release Date: What We Know for the latest.

    Model Selection for API Developers

    For most production use cases in April 2026: use claude-sonnet-4-6 as your default. It handles the vast majority of tasks at better economics than Opus. Use claude-opus-4-6 for tasks that require maximum reasoning depth — complex multi-step analysis, difficult coding problems, long-horizon agentic runs. Use claude-haiku-4-5-20251001 for high-volume, latency-sensitive, or cost-constrained tasks where raw capability is less critical than speed.

    What is the latest Claude model right now?

    As of April 2026: Claude Opus 4.6 (claude-opus-4-6) and Claude Sonnet 4.6 (claude-sonnet-4-6), both released February 2026. Claude Haiku 4.5 is the current speed/cost tier.

    Is Claude Sonnet 4.6 better than Claude Opus 4?

    Yes, in most practical benchmarks. Claude Sonnet 4.6 outperforms the original Opus 4.0 on coding, document comprehension, and instruction-following — at a lower price point. This follows Anthropic’s consistent pattern of new Sonnet tiers exceeding prior Opus tiers.

    What Claude model string should I use in my API calls?

    Use claude-sonnet-4-6 for most tasks. Use claude-opus-4-6 for maximum capability. Use claude-haiku-4-5-20251001 for speed and volume. Avoid claude-sonnet-4-20250514 and claude-opus-4-20250514 — these retire June 15, 2026.


  • Claude Managed Agents Integrations: Complete List (Notion, Asana, Sentry, and More)

    Claude Managed Agents Integrations: Complete List (Notion, Asana, Sentry, and More)

    Claude AI · Tygart Media
    Current supported integrations (April 2026): Notion, Asana, Sentry, Rakuten, Intercom, Cloudflare, Confluence, Jira, Linear, PagerDuty, Stripe, and dozens more via the MCP ecosystem. Anthropic is actively expanding the list.

    Claude Managed Agents is Anthropic’s enterprise agentic service — Claude running as an autonomous agent connected to your tools, taking multi-step actions without a human in the loop for every decision. The integrations list is what most teams are researching before adopting it, and it’s not clearly documented in one place. Here’s the complete breakdown.

    What “Integration” Means in This Context

    When Anthropic says Claude Managed Agents supports an integration, they mean Claude can authenticate with that service, read data from it, take actions in it (create, update, complete tasks), and reason across multiple services in a single agentic run. This is different from a simple API connection — Claude is actively using the tool the way a human would, not just pulling data from it.

    Confirmed Integrations at Launch

    Integration What Claude Can Do
    Notion Read/write pages, update databases, synthesize across workspaces, create meeting notes, manage project trackers
    Asana Create and update tasks, move items between projects, mark completions, generate status reports
    Sentry Triage errors, assign issues, summarize error patterns, escalate to relevant team members
    Rakuten Process affiliate data, update campaign parameters, generate performance summaries
    Intercom Draft support responses, route tickets, escalate complex issues, update customer records
    Cloudflare Monitor security alerts, update firewall rules, generate traffic reports
    Confluence Create and update documentation, summarize meeting notes into wiki pages
    Jira Create tickets, update sprint boards, generate burndown summaries, escalate blockers
    Linear Manage engineering issues, update cycle progress, triage incoming bugs
    PagerDuty Respond to incidents, escalate alerts, create post-mortems
    Stripe Query transaction data, generate revenue summaries, flag anomalies
    GitHub Review PRs, create issues, summarize commit history, manage release notes

    The MCP Layer: Extending Beyond the Default List

    Beyond the out-of-the-box integrations, Claude Managed Agents supports any service that exposes a Model Context Protocol (MCP) server. MCP is the open standard Anthropic developed for connecting AI models to external tools. If your internal systems, proprietary databases, or less common SaaS tools have an MCP server, Claude can integrate with them through the same managed agent infrastructure — no custom code required on the Claude side.

    This is why the integration list is effectively unbounded: the default set covers the most common enterprise tools, and MCP handles everything else.

    How This Differs from Claude in Chat with MCP Connectors

    Using Claude Chat with MCP servers configured requires a human actively running the conversation. Claude Managed Agents runs autonomously — you define the objective and the integrations, and Claude executes multi-step workflows without a human prompting each step. The agent can read from Notion, check Sentry for errors, create a Jira ticket, update Asana, and send a summary to Intercom in a single autonomous run.

    Pricing Note

    Claude Managed Agents is an enterprise-tier offering priced per session and per hour of agent runtime. It’s not available on individual Claude plans. For current pricing, see Claude Managed Agents Pricing: Complete Cost Analysis.

    Does Claude Managed Agents work with Notion?

    Yes. Notion is one of the confirmed launch integrations. Claude can read pages, write and update databases, synthesize across workspaces, and manage project trackers autonomously.

    Can Claude Managed Agents connect to custom internal tools?

    Yes, through the MCP (Model Context Protocol) layer. Any internal tool or proprietary system that exposes an MCP server can be connected to Claude Managed Agents without requiring changes on the Claude side.

    Is Asana supported in Claude Managed Agents?

    Yes. Asana is a confirmed integration. Claude can create and update tasks, move items between projects, mark completions, and generate status reports autonomously within Asana.


  • How Claude Cowork Task Scheduling Works

    How Claude Cowork Task Scheduling Works

    Claude AI · Tygart Media
    How it works in plain terms: Cowork tasks are stored instruction sets that Claude executes on a schedule. You write the prompt once; Claude runs it at the scheduled time using whatever tools and MCP connections you have configured.

    Claude Cowork’s scheduling feature is one of the least-documented parts of the product, but it’s the most powerful. Understanding how it actually works — what triggers tasks, what Claude has access to when running them, and what the limitations are — changes how you design automation with it.

    How Cowork Tasks Are Stored

    Each Cowork task is a named, persistent instruction set saved locally in your Claude Desktop environment. The task contains: a name, a prompt (the full instruction Claude follows each run), a schedule, and optionally a working directory and a set of enabled tools. Tasks are stored in JSON format under your Documents folder at ~/Documents/Claude/Scheduled/ alongside a scheduled-tasks.json index file.

    What Triggers a Scheduled Task

    Tasks run on cron-style schedules configured when you create the task. Common schedules include daily at a specific time, weekly on specific days, or on-demand (manual trigger only). When the scheduled time arrives, Claude Desktop wakes the Cowork runner, loads the task prompt, and executes it with the configured tools and MCP connections active.

    Critical limitation: Claude Desktop must be running and your machine must be awake when the scheduled time fires. Cowork is not a cloud scheduler — it depends on the local process being live. If your machine is asleep or Claude Desktop is closed, the task is skipped for that run with no retry.

    What Claude Has Access to During a Task Run

    When a Cowork task runs, Claude has access to everything configured in your Claude Desktop environment at that moment: all active MCP servers (Notion, Gmail, Google Drive, etc.), the Cowork bash VM for executing scripts and filesystem operations, any skill files mounted in the VM, and the working directory specified in the task config. It does not have access to the interactive chat thread — the task runs in its own isolated context.

    Task Memory: What Carries Over Between Runs

    Nothing carries over automatically. Each task run is stateless — Claude starts fresh with only the task prompt as its context. If your task needs to know what happened last time (what was published, what changed, what errors occurred), you have to build that logging into the task itself. The standard pattern: at the end of each run, write a log entry to a Notion page or local file; at the start of the next run, read that log to pick up context.

    This is why well-designed Cowork tasks always end with a Notion write and start with a Notion read.

    How to Design a Reliable Cowork Task

    Tasks that work well have four components: a clear single objective per task (do one thing, do it well), explicit context loading at the start (read the log, check what already exists), a defined success condition Claude can verify, and a logging step at the end that captures what ran and any errors. Tasks that try to do too many things in one run, or that assume Claude will remember previous runs without explicit context, fail inconsistently.

    When to Move Tasks to GCP Instead

    Cowork scheduling works well for tasks that need to run during your working day when your machine is on. For anything that needs to run at 3 AM, run on a strict schedule with zero missed executions, or process large amounts of data that would exhaust the local VM disk — those belong on GCP Cloud Run or a Compute Engine cron job, not Cowork. The architectural principle: Cowork for interactive-adjacent automation, GCP for always-on production pipelines.

    How do I create a scheduled task in Cowork?

    Open Claude Desktop, navigate to the Cowork section, create a new task, write your prompt, and set the schedule. Tasks are saved locally and run when Claude Desktop is open at the scheduled time.

    Why did my Cowork task not run at the scheduled time?

    Most likely Claude Desktop was closed or your machine was asleep. Cowork tasks require Claude Desktop to be running. Tasks that miss their scheduled time are skipped — there is no retry or catch-up mechanism.

    Can Cowork tasks run while I am using Claude Chat?

    Yes. Cowork tasks run in a separate context from the chat interface. Active Cowork task runs do not interrupt or share context with your current chat sessions.


  • Claude Cowork Not Working: 5 Common Errors and Fixes

    Claude Cowork Not Working: 5 Common Errors and Fixes

    Claude AI · Tygart Media
    Most common cause: The Cowork VM disk is full (sessiondata.img). Second most common: a scheduled task depends on a local process that stops when your machine sleeps. Both are fixable in minutes.

    Claude Cowork stops working for a small set of predictable reasons. This page covers the five most common failures, how to diagnose which one you’re hitting, and the exact fix for each.

    Error 1: “useradd failed: exit status 12”

    What it means: The Cowork VM’s internal disk (sessiondata.img) is full. No new sessions can be provisioned.

    Fix: Quit Claude Desktop. Move sessiondata.img from %APPDATA%\Claude\vm_bundles\claudevm.bundle\ (Windows) or ~/Library/Application Support/Claude/vm_bundles/claudevm.bundle/ (macOS) to your Desktop. Relaunch Claude Desktop — it recreates a fresh image. Full walkthrough: Claude Cowork useradd Failed Error Fix.

    Error 2: Scheduled Tasks Stop Running

    What it means: Tasks that were running on schedule suddenly stop firing. Often appears as tasks last running a few days ago with no new entries.

    Causes: Machine went to sleep, Claude Desktop was quit, or the local runner process died. Cowork tasks require Claude Desktop to be open and running on an active machine. They are not fully cloud-hosted — they depend on the local Cowork environment being live.

    Fix: Reopen Claude Desktop and manually trigger one task to verify it runs. For tasks that need to run reliably without the machine being awake, move them to a GCP Cloud Run cron job or a cloud VM instead of Cowork’s local scheduler.

    Error 3: MCP Tools Not Available in Cowork

    What it means: Cowork tasks can’t access Notion, Gmail, or other connected services that work fine in Chat.

    Fix: MCP servers must be configured in claude_desktop_config.json — the same config file Claude Desktop uses. If an MCP server appears in Chat but not in Cowork, verify it’s listed in the desktop config, not just the web interface. Restart Claude Desktop after any config changes.

    Error 4: File Access Denied or Path Not Found

    What it means: A Cowork task fails trying to read or write a file that should be accessible.

    Fix: Cowork’s VM mounts specific directories from your machine. If the file is outside a mounted path, Cowork can’t reach it. Check that the file path is within your configured working directories. On Windows, path separator issues (\ vs /) inside the Linux VM can also cause this — use forward slashes or escape backslashes in task prompts.

    Error 5: Tasks Run but Produce Wrong Output

    What it means: Cowork is running but the results are stale, wrong, or missing context from previous runs.

    Fix: Cowork tasks don’t have memory of previous runs by default. If your task depends on knowing what happened last time — what was published, what changed — you need to build that context explicitly into the task prompt, typically by reading a log from Notion or a local file at the start of each run. The task prompt is the only persistent instruction; Claude doesn’t remember prior task outputs.

    Why did Cowork tasks stop running overnight?

    Cowork requires Claude Desktop to be running on an active machine. If your computer slept, hibernated, or Claude Desktop was closed, scheduled tasks won’t fire. For always-on reliability, route tasks through a cloud runner instead.

    Why can Cowork not find my files?

    Cowork’s Linux VM only has access to directories you’ve configured as mount points. Files outside those paths are invisible to the VM. Verify your working directory configuration in Claude Desktop settings.

    Does Cowork work on Windows?

    Yes, Cowork is available on both Windows and macOS via Claude Desktop. The VM behavior and file paths differ slightly between platforms but the core functionality is the same.


  • Claude Cowork vs Claude Chat: When to Use Which

    Claude Cowork vs Claude Chat: When to Use Which

    Claude AI · Tygart Media
    Quick answer: Use Cowork for scheduled, recurring, or multi-step tasks that need to persist and run on their own. Use Claude Chat for interactive work, analysis, writing, and one-off tasks you’re doing right now.

    Anthropic now offers two distinct modes of working with Claude — the familiar chat interface and Cowork, a persistent task and agent environment. They look similar but serve fundamentally different purposes. Using the wrong one creates friction; knowing which to reach for first saves significant time.

    What Cowork Actually Is

    Cowork is a persistent agent environment inside Claude Desktop. It gives Claude access to your local filesystem, a sandboxed Linux VM with bash execution, your installed MCP servers, and a scheduler. Tasks you set up in Cowork can run on a schedule — daily, weekly, whenever you trigger them — without you being in the conversation. Claude operates autonomously against your instructions until the task is done.

    What Claude Chat Actually Is

    Claude Chat (claude.ai or the Claude app) is a stateless, interactive conversation interface. Each session is fresh. Claude has no persistent memory across sessions beyond what you’ve configured in memory settings. It’s optimized for real-time back-and-forth: you ask, Claude responds, you refine. The bash environment in Chat (used for file operations and code execution) is sandboxed and resets between sessions.

    Side-by-Side Comparison

    Factor Claude Chat Claude Cowork
    Runs without you No Yes — scheduled tasks
    Access to your files Upload only Direct filesystem access
    Persistent across sessions No (memory only) Yes — tasks and state persist
    Best for Interactive work, writing, analysis Recurring automation, pipelines
    MCP tool access Yes (if configured) Yes + local filesystem tools
    Runs on Anthropic’s cloud Your local machine
    Resource competition None (cloud-side) Shares your CPU/disk
    Skill files Yes (/mnt/skills/) Yes (same mount)

    When to Use Claude Chat

    Chat is the right tool when you’re actively involved in the work — drafting, editing, analyzing, strategizing. If you need to go back and forth, refine an output, or make judgment calls mid-task, Chat’s interactive model is faster and more natural. It’s also better for any task that’s genuinely one-off: you do it once, you’re done, there’s nothing to schedule or automate.

    Chat also runs in the cloud, meaning it doesn’t compete with your machine’s other processes and doesn’t run into the local VM disk limitations that Cowork can hit with heavy workloads.

    When to Use Cowork

    Cowork shines for work that should happen without you: daily newsroom publishing, weekly SEO reports, nightly data syncs, any pipeline that follows the same steps every run. If you find yourself doing the same Claude Chat session more than twice a week, it’s a candidate for a Cowork task.

    Cowork also makes sense for tasks that need direct access to files on your machine — reading from a local folder, processing downloads, interacting with local applications — since Chat requires you to explicitly upload files each session.

    Known Cowork Limitation to Be Aware Of

    Cowork runs on a local VM (the sessiondata.img file) with a fixed 8.5GB disk. Heavy users with many skills installed will periodically hit a disk-full error that prevents new sessions from launching. This is a known bug (GitHub #30751) with a manual workaround. See Claude Cowork useradd Failed Error: How to Fix It for the fix.

    Is Claude Cowork better than Claude Chat?

    Neither is better — they serve different purposes. Chat is optimized for interactive, real-time work. Cowork is for persistent, scheduled, autonomous tasks. Most power users use both regularly for different types of work.

    Can Claude Cowork access the internet?

    Yes, through MCP server integrations and web search tools. Cowork tasks can call APIs, search the web, read from connected services like Notion or Gmail, and interact with any MCP-connected tool you’ve configured.

    Does Claude Cowork use the same AI model as Chat?

    Yes — Cowork uses the same underlying Claude models (currently Opus 4.6 and Sonnet 4.6). The difference is the execution environment, not the model.


  • Claude Cowork ‘useradd Failed’ Error: How to Fix the sessiondata.img Full Bug

    Claude Cowork ‘useradd Failed’ Error: How to Fix the sessiondata.img Full Bug

    Claude AI · Tygart Media
    ⚠ Known Bug: This is GitHub issue #30751 — still open as of April 2026. Anthropic has not shipped a permanent fix. The workaround below is the only reliable solution.

    If every Cowork task is failing with useradd: cannot create directory /sessions/friendly-youthful-thompson or a similar error, your Cowork VM’s internal disk is full. This is not something you broke — it’s a known Anthropic bug that affects power users consistently. Here’s exactly what’s happening and how to fix it in under two minutes.

    What’s Causing the Error

    Cowork runs on a local VM with a fixed 8.5GB disk image called sessiondata.img. Every Cowork conversation creates a new directory under /sessions/<name>/ inside that VM and caches all your installed plugins and skills there. Those directories are never cleaned up automatically. Once the disk fills — roughly 80 sessions for light users, 40–50 sessions for users with many skills installed — every new task fails immediately with a useradd error. The session simply can’t be provisioned.

    If you have 20+ skills installed (the Tygart Media stack runs 40+), you’ll hit the cap significantly faster than the average user.

    The Fix: Move the Image File

    The fix is the same on macOS and Windows: move sessiondata.img out of its location so Claude Desktop rebuilds it fresh on next launch.

    Windows

    Quit Claude Desktop completely. Open Run (Win + R), paste this path and press Enter:

    %APPDATA%\Claude m_bundles\claudevm.bundle\

    Find sessiondata.img and move it to your Desktop as a backup. Relaunch Claude Desktop — it will recreate a fresh image automatically. Your first Cowork session after the reset may take slightly longer while plugins reinstall.

    macOS

    Quit Claude Desktop. In Finder, press Cmd + Shift + G and go to:

    ~/Library/Application Support/Claude/vm_bundles/claudevm.bundle/

    Move sessiondata.img to your Desktop. Relaunch Claude Desktop.

    What Gets Wiped vs What’s Preserved

    Data Location Wiped?
    Sidebar task list Electron IndexedDB ✅ Preserved
    Scheduled task definitions Documents/Claude/Scheduled/ ✅ Preserved
    MCP server config claude_desktop_config.json ✅ Preserved
    Chat conversation history Electron LevelDB ✅ Preserved
    VM plugin/skill cache Inside sessiondata.img ⚠ Wiped (auto re-downloads)
    VM session working dirs /sessions/<n>/ inside VM ⚠ Wiped (this is the fix)

    How Often Will You Need to Do This?

    Until Anthropic ships automatic session cleanup, this is a recurring task. With a heavy skill load, plan on running the fix every 4–6 weeks or whenever you see the useradd error return. Setting a calendar reminder is the most reliable approach.

    The Longer-Term Fix: Move Heavy Operations Off Cowork

    The root cause is that Cowork was designed for lighter, conversational task automation — not running dozens of skills across many parallel sessions. If you’re running content pipelines, batch WordPress operations, or multi-step automation workflows, moving those operations to a GCP Cloud Run cron job or Compute Engine VM eliminates the local VM bottleneck entirely. Cowork’s local sandbox competes for your machine’s resources; GCP runs isolated, always-on, and never fills up your laptop’s disk.

    Why does Cowork say “useradd failed: exit status 12”?

    The Cowork VM’s internal disk (sessiondata.img) is full. It can no longer create new session user directories. Moving the image file out and letting Claude Desktop recreate it clears the disk and resolves the error.

    Will I lose my Cowork tasks if I move sessiondata.img?

    No. Your task definitions, scheduled tasks, MCP config, and conversation history are all stored outside the VM image. Only the internal plugin/skill cache is wiped — it re-downloads automatically on the next session.

    How do I prevent Cowork from filling up again?

    Until Anthropic ships a permanent fix (GitHub issue #30751), the options are: run the reset script periodically, reduce your installed skill count, or route heavy operations to GCP instead of Cowork.


  • Claude 5 Release Date 2026: Leak Signals, Expected Features & Anthropic’s Timeline

    Claude 5 Release Date 2026: Leak Signals, Expected Features & Anthropic’s Timeline

    Claude AI · Tygart Media · Updated April 2026
    Current status (April 16, 2026): Claude 5 has not been officially announced by Anthropic. The current latest models are Claude Opus 4.6 and Claude Sonnet 4.6, released in February 2026. Based on Anthropic’s release cadence and early signals, Claude 5 is expected Q2–Q3 2026.

    Every few months, a new wave of “Claude 5 release date” searches spikes — and it makes sense. Anthropic moves fast, the gaps between major generations have been shortening, and early signals like Vertex AI log leaks have given the community something to speculate on. Here’s an honest breakdown of what’s confirmed, what’s leaked, and what the pattern suggests.

    What’s Confirmed About Claude 5

    As of April 2026, Anthropic has not officially announced Claude 5 by name in any public release notes, API documentation, or blog post. The company’s official model table shows the Claude 4.x family as current. No countdown page exists. No API model string beginning with claude-5 has appeared in public documentation.

    What is confirmed: Anthropic is actively deprecating the original Claude 4.0 models (retiring June 15, 2026), recommending migration to Claude Sonnet 4.6 and Opus 4.6. This is a routine generational housekeeping move, not a Claude 5 announcement.

    The Evidence For a Q2–Q3 2026 Release

    The strongest early signal came in early February 2026, when a model identifier — claude-sonnet-5@20260203 — appeared briefly in Google Vertex AI error logs. Independent sources cross-verified the leak, and the codename “Fennec” circulated alongside claimed benchmark scores of around 80.9% on SWE-bench Verified, compared to Opus 4.6’s current scores.

    Beyond the leak, the pattern is consistent: Anthropic has released a new major model generation roughly every 12–14 months since Claude 3. Claude 4.5 (the highest-capability 4.x model) reached 77.2% on SWE-bench Verified. A Claude 5 release that clearly exceeds that — not just marginally — would justify a major version bump and align with Anthropic’s stated commitment to releasing models that represent genuine capability leaps, not incremental updates.

    Anthropic’s Release Pattern

    Generation Initial Release Gap to Next Major
    Claude 2 July 2023 ~8 months
    Claude 3 March 2024 ~14 months
    Claude 4 May 2025 ~12–14 months → Q2–Q3 2026

    A 12-month gap from the Claude 4 launch (May 2025) points to May–July 2026 as the earliest likely window. Anthropic has been explicit that they won’t rush a release — Claude 5 will need to clearly establish a new capability tier to justify the version number.

    What Claude 5 Is Expected to Improve

    Based on leaked benchmark data and Anthropic’s public research direction, the Claude 5 generation is expected to push forward on: extended thinking and multi-step reasoning (building on the chain-of-thought work in Claude 3.5+), larger context handling, improved agentic reliability for long-horizon tasks, and faster inference at the Sonnet tier. Pricing is expected to follow the established pattern — Claude 5 Sonnet likely priced at or below current Opus 4.6 rates while outperforming it on most tasks.

    The Current Models Are Excellent — Don’t Wait

    If you’re evaluating whether to build on Claude now or wait for Claude 5, the answer is build now. Claude Sonnet 4.6 and Opus 4.6 are capable, stable, and well-documented. The 4.x API will remain live well after Claude 5 launches — Anthropic maintains parallel model availability for enterprise predictability. Waiting costs you months of production time for a model that may arrive on an uncertain schedule.

    For current model specs and API strings, see Claude API Model Strings — Complete Reference. For pricing on current models, see Claude AI Pricing: Every Plan Explained.

    When is Claude 5 coming out?

    Claude 5 has not been officially announced. Based on Anthropic’s release cadence and early Vertex AI log leaks, Q2–Q3 2026 (roughly May–September) is the most cited window. No confirmed date exists as of April 2026.

    Is Claude 5 confirmed?

    No. Anthropic has not officially announced Claude 5 by name. The “Fennec” codename and claude-sonnet-5@20260203 model string surfaced in third-party Vertex AI logs, but Anthropic has not confirmed a Claude 5 release.

    What is the latest Claude model right now (April 2026)?

    The current latest Claude models are Claude Opus 4.6 (claude-opus-4-6) and Claude Sonnet 4.6 (claude-sonnet-4-6), both released in February 2026. Claude Haiku 4.5 is the current speed/cost tier.

    Will Claude 5 Sonnet beat Claude Opus 4.6?

    That’s the expected pattern. With every prior generation, the mid-tier Sonnet model of the new generation outperformed the previous generation’s Opus on most benchmarks, at lower cost. Leaked benchmark data suggests Claude 5 Sonnet (“Fennec”) scores around 80.9% on SWE-bench Verified versus Opus 4.6’s current scores.


  • Claude 4 Deprecation: Sonnet 4 and Opus 4 Retire June 15, 2026

    Claude 4 Deprecation: Sonnet 4 and Opus 4 Retire June 15, 2026

    Claude AI · Tygart Media
    ⚠ Deprecation Notice (April 2026): Anthropic has announced that claude-sonnet-4-20250514 and claude-opus-4-20250514 — the original Claude 4.0 models — are deprecated. API retirement is scheduled for June 15, 2026. Anthropic recommends migrating to Claude Sonnet 4.6 and Claude Opus 4.6 respectively.

    If you’re still running the original Claude Sonnet 4 or Opus 4 model strings in production, you have a hard deadline: June 15, 2026. After that date, calls to claude-sonnet-4-20250514 and claude-opus-4-20250514 will fail on the Anthropic API. Here’s exactly what’s being deprecated, what to migrate to, and what you’ll gain from upgrading.

    What’s Being Deprecated

    Anthropic is retiring the original Claude 4.0 model versions — the ones that shipped in May 2025. These are distinct from the 4.x versions released since. The specific API strings going offline:

    Model API String (retiring) Retirement Date
    Claude Sonnet 4 (original) claude-sonnet-4-20250514 June 15, 2026
    Claude Opus 4 (original) claude-opus-4-20250514 June 15, 2026

    These are not the latest Claude 4 models. If you’ve been on Anthropic’s recommended defaults, you’re likely already on 4.6. This deprecation primarily affects teams that pinned specific model version strings in their API calls rather than using the alias endpoints.

    What to Migrate To

    Anthropic’s recommendation is a direct version bump within the same model tier:

    Retiring Migrate To API String
    claude-sonnet-4-20250514 Claude Sonnet 4.6 claude-sonnet-4-6
    claude-opus-4-20250514 Claude Opus 4.6 claude-opus-4-6

    The 4.6 models are meaningful upgrades — not just version bumps. Claude Sonnet 4.6 ships with near-Opus-level performance on coding and document tasks, dramatically improved computer use capabilities, and a 1 million token context window in beta. Claude Opus 4.6 adds the same 1M context window alongside improvements to long-horizon reasoning and multi-step agentic work.

    Why Anthropic Deprecates Models

    Anthropic follows a predictable model lifecycle: new versions within a generation ship as upgrades, and older version strings are retired on a roughly 12-month timeline after a successor is available. This keeps the API surface clean and pushes users toward better-performing models. The deprecation of the original Claude 4.0 strings follows the same pattern as prior Claude 3 and 3.5 retirements.

    For most API users, the migration is a one-line change — swap the model string. Prompting styles, tool use conventions, and JSON response formats are stable across 4.x generations. Anthropic has not announced breaking changes that would require prompt rewrites when moving from 4.0 to 4.6.

    How This Fits the Claude 4 Generation Timeline

    Model Released Status
    Claude Sonnet 4 (original) May 2025 ⚠ Deprecated — retires June 15, 2026
    Claude Opus 4 (original) May 2025 ⚠ Deprecated — retires June 15, 2026
    Claude Opus 4.6 February 5, 2026 ✅ Current flagship
    Claude Sonnet 4.6 February 17, 2026 ✅ Current production default
    Claude Haiku 4.5 October 2025 ✅ Current speed/cost tier

    What If You Don’t Migrate Before June 15?

    API calls to claude-sonnet-4-20250514 or claude-opus-4-20250514 after June 15, 2026 will return errors. There is no automatic failover to a newer model — the call simply fails. If you have any production systems, scheduled jobs, or automated pipelines using these version strings, audit them now. A global search for 20250514 in your codebase is the fastest way to find exposure.

    What Comes After Claude 4.x

    Claude 5 is expected in Q2-Q3 2026, based on Anthropic’s release cadence and early signals from Vertex AI deployment logs. As has been the pattern with prior generations, Claude 5’s mid-tier Sonnet model is expected to outperform the current Opus 4.6 on most benchmarks at a lower price point. No official announcement has been made as of April 2026.

    When does Claude 4 deprecate?

    The original Claude Sonnet 4 (claude-sonnet-4-20250514) and Claude Opus 4 (claude-opus-4-20250514) are deprecated and retire on June 15, 2026. Current 4.6 models are not affected.

    What should I migrate to from Claude Sonnet 4?

    Migrate to claude-sonnet-4-6 (Claude Sonnet 4.6). It’s a direct upgrade in the same model tier with significantly improved capabilities and a 1M token context window in beta.

    Will my prompts still work after migrating from 4.0 to 4.6?

    In most cases, yes. Anthropic has maintained API compatibility across the 4.x generation. The 4.6 models are more capable, not differently structured. Most production prompts migrate without changes.

    What’s the difference between Claude 4 and Claude 4.6?

    Claude 4.6 (released Feb 2026) is a meaningful upgrade over the original Claude 4.0 (released May 2025). Key improvements: near-Opus performance at Sonnet pricing, 1M token context window in beta, dramatically better computer use, and improved instruction-following accuracy.

  • The Restoration Talent Window Is Closing Faster Than You Think

    The Restoration Talent Window Is Closing Faster Than You Think

    A LinkedIn post from a restoration recruiter in Houston tipped me off this morning. He’s right — but the timeline is shorter than most people in the industry realize.

    Mitchell Riley LinkedIn post about Claude Managed Agents announcement
    Mitchell Riley’s LinkedIn post that started this train of thought.

    This article is part of The Restoration Operator’s Playbook — Tygart Media’s body of work on how the industry’s best restoration companies are actually thinking in 2026. Start with the pillar piece if this is your first read.

    The post that got me thinking

    This morning I logged into LinkedIn and saw a post from Mitchell Riley — a restoration industry recruiter in Houston who places PMs, GMs, and business development leaders for restoration contractors across the country. Mitchell flagged Anthropic’s Claude Managed Agents launch with the kind of casual enthusiasm only people who actually use this stuff every day can manage. He called it “pretty cool” and noted that Claude will now build you an agent based on natural language.

    He’s right. He’s also pointing at something most of the restoration industry hasn’t fully processed yet.

    What Anthropic actually shipped

    On April 8, 2026, Anthropic launched Claude Managed Agents in public beta. The short version: the infrastructure work that used to take three to six months of engineering — sandboxed code execution, credential management, long-running session persistence, error recovery, observability — is now a managed service. You define what the agent should do. Anthropic runs it.

    The companies already shipping production agents on it: Notion, Asana, Rakuten, and Sentry. Notion lets teams delegate coding, slides, and spreadsheets to Claude without leaving the workspace. Rakuten deployed specialist agents across product, sales, marketing, finance, and HR — each live in under a week. Sentry built an agent that goes from flagged bug to open pull request, fully autonomous.

    Internal Anthropic testing showed up to a 10-point improvement in task success on structured generation work versus a standard prompting loop, with the largest gains on the hardest problems.

    That’s the announcement. Here’s why it matters for restoration.

    The bottleneck just moved

    For the last two years, the question every restoration owner asked about AI was some version of: “Can it actually do the work?” The honest answer was usually “not yet, not without a developer team you don’t have.”

    That’s no longer the question. The infrastructure gap closed on April 8. The new bottleneck is not “can you build the agent” — it’s “do you have the human operators who know what the agent should be doing in the first place.”

    Restoration is an industry where the real intelligence lives in people. A senior PM who has worked five hundred losses knows things that have never been written down anywhere. How a Cat 3 storm response actually sequences when the carrier is dragging on TPA approvals. The difference between a contents pack-out that closes clean and one that becomes a six-month dispute. Which mitigation decisions buy you a profitable job and which ones bury you on the reconstruction side. None of that lives in a textbook. It lives in the heads of people who have been doing the work for fifteen or twenty years.

    That tribal knowledge is now the constraint. The companies that win the next three years will be the ones who pair Managed Agents (or something like it) with senior operators who can tell the agent what good looks like. The companies that try to skip that step — that try to hire generalists and teach them restoration on the fly while their competitors are distilling twenty-year veterans into operational systems — are going to get lapped.

    Buy the talent now

    This is where the recruiting angle gets interesting. Senior restoration talent has always been hard to find. It’s about to get much harder, for a reason most owners haven’t priced in yet: the value of a senior PM is no longer just the work that PM does directly. It’s the work an entire AI system does in their image once their judgment has been encoded into the workflow.

    Right now, that arbitrage is open. The market hasn’t repriced senior operators for what they’re actually worth in an AI-augmented restoration company. In twelve to twenty-four months, it will. The owners who hire the best PMs, GMs, and BD leaders now — and who pair them with someone like Mitchell who actually understands the placement game — are going to look like geniuses in 2027.

    Mitchell is one of the people who gets this from the inside. He uses the AI tools himself. He builds workflows. He analyzes things in dimensions and context that most recruiters never touch — most recruiters in this industry are still working from a spreadsheet of resumes and a cell phone. Mitchell is the kind of recruiter who notices when Anthropic ships something that’s going to change the value of every senior hire he places, and posts about it on a Wednesday morning. That’s the level of operator the smart restoration owners are going to want in their corner.

    What to actually do this quarter

    If you run a restoration company and you read this far, three concrete things:

    One. Identify your two or three most senior operators — the people whose judgment is load-bearing for the business. Start documenting how they think, not just what they do. The documentation is the raw material every future AI workflow will run on.

    Two. Open one or two senior hires you’ve been putting off. The talent market is going to tighten. Get in front of it.

    Three. Stop treating AI as an IT project. It’s an operational capability. The companies that figure this out are not waiting for their tech vendor to sell them an “AI feature.” They’re hiring the operators, capturing the judgment, and pointing the tooling at the result.

    Mitchell’s post was three sentences. The full version of what he was pointing at takes about a thousand words. This is that version.

    If you’re a restoration owner thinking about senior placements in the next two quarters, you should be talking to Mitchell. And if you’re thinking about how to operationalize AI inside your company — distilling senior judgment into systems your whole team can run — that’s the conversation we have at Tygart Media.

    Read next: The New Restoration Operator: How the Industry’s Best Companies Are Thinking in 2026 — the pillar piece this article belongs to.