Tag: AI Models 2026

  • Multi-Model Concentration: How Seven AI Models Reading Your Notion at Once Becomes a Writing Methodology

    The short version: If you ask one AI model to summarize your knowledge base, you get one editorial sensibility. If you ask seven different models the same question and feed all seven answers back to a synthesizer, you get something else entirely: a triangulated map of your own thinking, with the canon and the edges marked. This is a writing methodology I stumbled into while drafting an article. It is repeatable, it is cheap, and it produces material no single model can produce alone.

    I was trying to write a short post for LinkedIn. The post was fine. The post was also missing the actual insight that made the topic worth writing about. I asked one of the larger AI models to query my Notion workspace and bring back any material I had already written that touched on the topic. It returned a clean, organized summary. Useful. But I had a quiet hunch that the summary was less complete than it looked.

    So I asked six other AI models the same question. Different companies, different training data, different objective functions. Same workspace. Same prompt. Then I pasted all the responses back into one synthesizer model and asked it to compare them.

    What I found was not subtle. Each model walked into the same room and saw a different room. The agreement zone — what three or more models independently surfaced — turned out to be my actual canon. The divergence zone — the unique pulls only one model found — turned out to contain the most interesting material in the whole set.

    This is the writeup of that process, what worked, what did not, and why I think it is genuinely a new way to do research on your own corpus.

    The setup

    I have a Notion workspace that holds about three years of structured thinking, framework drafts, content strategy notes, and operational documentation. It is the operating brain of a content agency. Roughly 500 pages, a few thousand chunks of indexed text. The kind of corpus that is too big to re-read but too valuable to ignore.

    The standard way to get value out of a corpus this size is to use a single AI assistant — Notion AI, ChatGPT with workspace access, Claude with MCP, whatever — and ask it to summarize, search, or extract. This works. It is also limited in a specific way: you only get one model’s reading of your material. One editorial sensibility. One set of training-data biases shaping what gets surfaced and what gets walked past.

    The experiment was simple. Run the same comprehensive prompt across seven models in parallel. Paste each response into a single conversation with a synthesizer model. Compare.

    The prompt

    The prompt asked each model to sweep the workspace for any content related to a specific cluster of themes — personal branding, skill development, niche authority, content strategy, and learning systems. It instructed each model to skip generic logs and surface only specific frameworks, named concepts, distinctive sentences, and concrete examples already in the user’s voice. It explicitly asked them to ignore noise and return concentrated signal.

    The same prompt went to every model. No customization. No second pass. Just one query each, then their raw responses pasted into a synthesis conversation.

    The seven models

    1. Claude Opus 4.7
    2. Claude Opus 4.6
    3. Claude Sonnet 4.6
    4. Google Gemini 3.1 Pro
    5. OpenAI GPT 5.4
    6. OpenAI GPT 5.2
    7. Moonshot Kimi 2.6

    One additional model — Gemini 2.5 Flash — was queried but declined. It honestly reported that it could not access the workspace from chat mode. That non-result turned out to be useful information of its own kind, which I will come back to.

    What happened

    The agreement zone is the canon

    A small set of concepts showed up in three or more model responses. Same source pages. Same quotes. Same framing. When seven independently trained AI models — different companies, different architectures, different objective functions — converge on the same handful of ideas pulled from your own writing, that convergence is not coincidence. It is signal that those ideas are structurally important in your corpus.

    For my own workspace, the agreement zone surfaced about a dozen high-conviction concepts that had been scattered across hundreds of pages. I had written all of them. I had not realized which ones were structurally load-bearing in my own thinking. The triangulation made it obvious.

    This is the first practical use case: multi-model concentration tells you what your canon actually is. Not what you think it is. Not what you wish it was. What the corpus, read by neutral readers, demonstrably contains.

    The divergence zone is the edge

    The more interesting half of the experiment was where the models disagreed. Each model surfaced unique material the others walked past. Not because the others missed it accidentally. Because each model has a different training signature that shapes what it values reading.

    • One Claude model went structural. It proposed a spine for the article and called out gaps in the corpus where I would need to do net-new research.
    • A different Claude version went concept-cartographer. It found named framework clusters the others scattered across multiple sections.
    • A Sonnet model surfaced operational mechanics — the actual step-by-step inside frameworks the others mentioned at headline level.
    • Gemini found pragmatic material no one else touched, including specific productivity numbers from the corpus.
    • One GPT version played hidden-gem hunter, surfacing single sentences with article-grade force that other models read past.
    • The other GPT version restructured everything into a finished reference document — designed as something publishable, not just retrievable.
    • Kimi went deep-system archaeologist, finding named frameworks in corners of the workspace others did not reach.

    Reading the seven outputs in sequence felt like getting feedback from seven editors. None of them were wrong. None of them were complete. The full picture only emerged when I treated all seven as inputs to a synthesis layer.

    The negative result mattered

    Gemini Flash’s honest “I cannot access this workspace from chat mode” was, in a quiet way, the most useful single response. It told me that workspace access is not equally distributed across the models I have available. Future runs of this methodology need to verify connectivity first — otherwise I am not comparing models, I am comparing connection states.

    It also reminded me that an AI that says “I cannot” is, on average, more trustworthy with deeper work than one that hallucinates a confident-sounding pull from a workspace it could not see. Worth weighting that into model selection going forward.

    The complication: recursive consensus

    Partway through the experiment I noticed something I had not predicted. Three of the models cited previous AI synthesis pages already living in my workspace. Pages titled things like “Cross-Model Second Brain Analysis Round 1” or “Round 3: Embedding-Fed Generative Pass.” These were artifacts of earlier concentration sessions I had run weeks ago and saved into Notion as canonical pages.

    Which means: when models queried my workspace, they were sometimes finding pages where previous models had already done this exact exercise and reached conclusions. Those pages were then read back as “discovered” insight by the current round of models.

    This matters. It means the agreement zone is partially inflated. When four models all surface the same concept as “an undervalued piece of intellectual property,” some of that consensus might be coming from a Notion page that already says exactly that — written by a prior AI synthesis based on a still-earlier round of consensus.

    That is a feedback loop. Earlier AI conclusions become canonical workspace content that later AI reads back as independently-discovered insight. It is not bad — in some sense it is exactly how a knowledge system should compound over time — but it should be named, because if you do not name it, you mistake echo for verification.

    The two types of signal

    Once you know about the recursive consensus problem, you can sort the agreement zone into two cleaner buckets:

    Primary-source canon. Concepts that surface across multiple models because the models independently found them on pages you originally wrote. These are the cleanest possible signal. Multiple neutral readers, reading your original material, all flagged the same idea as structurally important.

    Recursive AI consensus. Concepts that surface across multiple models because the models found them on pages that were themselves AI syntheses of earlier AI rounds. These are not worthless — the original AI rounds were also synthesizing real material — but they should be weighted lower than primary-source canon.

    Practically, this means tagging synthesis pages clearly in your knowledge base. Something like a metadata field on each Notion page declaring whether it is primary-source thinking or AI-derived synthesis. Future model runs can then be instructed to weight primary higher than synthesis, or to exclude synthesis entirely on a given pull.

    Why this is a real methodology, not just a curiosity

    I want to be careful not to overclaim. This is not magic. It is a specific application of well-understood ensemble principles — the same logic that says combining multiple weak classifiers usually beats a single strong one — applied to retrieval and synthesis over a personal corpus.

    What makes it useful in practice is that the cost is near zero, the inputs are already sitting in your workspace, and the output is a brief that is grounded in your own material rather than confabulated by a single model. For anyone who writes long-form, builds frameworks, or runs a knowledge-driven business, this is a genuine upgrade over single-model summarization.

    The four properties that make it work

    1. Different training signatures. The models must come from different labs with different training data. Two Claude models from the same family produce more correlated readings than a Claude and a Gemini. The diversity of the readers is the entire point.
    2. Same prompt, no customization. The comparison only works if every model sees the identical query. Optimizing the prompt for each model defeats the purpose.
    3. Same workspace access. All models must have read access to the same corpus. Otherwise the divergence is a function of who could see what, not a function of editorial sensibility.
    4. A synthesizer that compares, not summarizes. The final layer is not “give me a summary of all seven outputs.” It is “tell me where they agree, where they diverge, and what each model uniquely contributed.” That second framing is what makes the canon and the edge visible.

    What you actually do with the output

    The synthesizer’s comparison is the deliverable, not the source pulls. The pulls are raw material. The synthesis tells you:

    • What is undeniably canonical in your corpus (3+ model agreement)
    • What is structurally important but only one model spotted (the article-grade gems)
    • What is missing from your corpus entirely and would require external research (the gap analysis)
    • Which models are best at which types of retrieval (so you can pick better next time)

    That output is the brief. Whatever you build next — an article, a pitch, a framework, a new product — starts from there.

    The methodology in five steps

    1. Decide what you want to extract. Pick a thematic cluster. Not “summarize my workspace” — too broad. Something like “everything related to my personal branding, skill development, and authority-building thinking.” Specific enough to focus the readers, broad enough to invite real coverage.
    2. Write one prompt. The prompt should ask for specifics — frameworks, distinctive phrases, named concepts, examples in your voice — and explicitly tell each model to filter out generic notes, meeting logs, and task lists. Tell it you want concentrated signal, not summary.
    3. Run the same prompt across as many cross-lab models as you have access to. Three is the minimum useful sample. Five to seven gives a much clearer picture. Pull in Anthropic, OpenAI, Google, and at least one frontier model from outside the big three.
    4. Paste every response into a single synthesis conversation. Tell the synthesizer to compare, identify the agreement zone, identify the divergence zone, flag any negative results (models that could not access the corpus), and call out where the consensus might be inflated by recursive AI synthesis pages.
    5. Use the synthesis as your brief. Whatever you build next starts from this output, not from a blank page or a single model’s summary.

    The honest caveats

    Three things to keep in mind before you try this.

    It only works on a corpus worth triangulating. If your knowledge base is small, generic, or mostly meeting notes, the multi-model approach will not surface anything more useful than a single model would. The methodology assumes you have done the work of building a substantive corpus first.

    Connectivity is not uniform. Not every model has the same access to your workspace. Some will refuse the query honestly. Some may try to answer without true workspace access and confabulate. Verify what each model actually had access to before you compare outputs.

    The recursive consensus is real. If your workspace contains prior AI syntheses, future syntheses will be partially echoing past ones. This is not a fatal flaw — it is how a knowledge system compounds — but you should know it is happening so you do not over-weight findings that are bouncing around inside your own AI history.

    Why this matters beyond writing one article

    The bigger frame is this: most of the value in any modern knowledge worker’s life lives inside a corpus they have written themselves but cannot fully see. Notes, drafts, frameworks, half-finished documents, scattered insights. The brain that produced all of it cannot reread all of it.

    Single-model retrieval lets you query that corpus through one editorial lens. Useful. Limited.

    Multi-model concentration lets you query that corpus through several editorial lenses simultaneously, then triangulate. The agreement zone reveals what is structurally important in your own thinking. The divergence zone reveals the high-value material that only some kinds of readers will catch. The negative results reveal capability gaps you should know about. The whole thing produces a much higher-resolution map of your own intellectual material than any one model can produce alone.

    It cost almost nothing to run. It took maybe two hours from first prompt to final synthesis. The output was substantively better than anything I have produced from a single-model query. And the meta-insight — that AI consensus over your own corpus is partially recursive and needs to be tagged accordingly — is itself the kind of finding I would not have noticed without running multiple models in parallel.

    This is a methodology, not a one-off trick. I will keep using it. If you have a corpus worth concentrating, you should try it too.

    Frequently asked questions

    How many models do I need?

    Three is the minimum. Five to seven is the sweet spot. Past about ten you hit diminishing returns and start spending more time managing the inputs than reading the synthesis.

    Do the models need to come from different companies?

    Yes. Two Claude models will produce more correlated readings than a Claude and a Gemini. The diversity of training data is what makes the triangulation work. Mix Anthropic, OpenAI, Google, and at least one frontier model from outside the three big labs.

    What if my models cannot access my workspace?

    Then the methodology does not run. Connectivity is the prerequisite. Verify each model’s access before you start. A model that confabulates a confident-sounding pull from a workspace it cannot see is worse than a model that honestly declines.

    How do I handle the recursive consensus problem?

    Tag synthesis pages in your workspace with a metadata field declaring them as AI-derived. Then either instruct future model runs to weight primary-source pages higher, or run two passes: one with all sources, one with synthesis pages excluded. The delta between the two passes shows you what is genuine new signal versus what is echo.

    What is the synthesizer model supposed to do differently than the source models?

    The synthesizer is not summarizing your corpus. It is comparing the seven readings of your corpus. Its job is to identify agreement, divergence, and gaps across the inputs, and to flag the methodological caveats. That is a different task than retrieval. Pick a model with strong reasoning over long context for the synthesis layer.

    Can I use this for things other than writing articles?

    Yes. Anywhere you need to extract a brief from a substantial corpus — pitch decks, framework design, product positioning, board prep, strategic planning — multi-model concentration gives you a higher-resolution starting point than single-model retrieval. The article use case is just where I noticed it. The methodology generalizes.

    The bottom line

    One AI reading of your knowledge base is one editor’s opinion. Seven AI readings, compared properly, is a triangulation. The agreement zone is your actual canon. The divergence zone contains the highest-value unique material. The negative results tell you about capability gaps. The recursive consensus problem tells you which conclusions to trust and which to weight lower.

    The whole thing is cheap, fast, and produces material no single model can produce alone. If you have a corpus worth thinking about, you have a corpus worth concentrating across multiple models. Start with three. Compare what they bring back. The methodology gets sharper from there.


  • When to Use Claude in Chrome vs When to Use the API

    When to Use Claude in Chrome vs When to Use the API

    The Decision Rule
    API first. Claude in Chrome when the API doesn’t exist or is blocked. The Chrome extension isn’t a replacement for API access — it’s what you reach for when API access isn’t an option.

    If you’ve worked with both the Claude API and Claude in Chrome, you’ve probably noticed that in many cases, you could technically use either one to accomplish a similar outcome. Fetching content from a page, submitting data, triggering a workflow — these things can often be done through an API or through a browser UI.

    The question of which to use isn’t primarily about capability. It’s about maintenance, reliability, and what happens at 3am when something breaks.

    What the API Gives You That Chrome Can’t

    Repeatability. An API call is deterministic. The same endpoint, the same payload, the same result. A Chrome UI interaction depends on the current state of a webpage — and web pages change. A button gets renamed. A modal gets added. A UI redesign ships. None of this breaks an API. All of it can break a Chrome automation.

    Scale. You can make hundreds of API calls per hour with appropriate rate limiting. Chrome UI automation runs at human browsing speed — one action at a time, in a real browser, with real rendering. That’s fine for occasional tasks. It doesn’t scale.

    No browser dependency. API calls run in code. They run in cloud functions, scheduled jobs, command-line scripts, anywhere. Chrome automation requires a running Chrome instance with the extension active and a profile logged in. That’s more fragile infrastructure.

    Reliability across time. A well-written API integration runs for years without maintenance. Chrome UI automation often needs updates when a target site changes its interface.

    What Chrome Gives You That the API Can’t

    Access to tools with no API. A lot of useful software — especially newer SaaS products, niche platforms, and tools built primarily for human users — doesn’t have an API, or has one that doesn’t expose the specific feature you need. Chrome is often the only programmatic path in.

    Access to authenticated browser sessions. Some platforms allow actions through a logged-in browser session that aren’t available through the API at all, or that require API tiers you don’t have. Chrome operates inside a real session with real cookies.

    No API key management. Using Chrome doesn’t require obtaining API credentials, managing tokens, or worrying about rate limits, API deprecations, or breaking changes to an API schema.

    Speed to first working automation. Setting up a Chrome session and describing what to click is often faster than reading API documentation, obtaining credentials, and writing integration code. For a one-time task, Chrome wins on speed.

    The Practical Decision Framework

    Ask these questions in order:

    1. Does this tool have an API that exposes what I need? If yes — use the API. Always.
    2. Will I need to run this more than once or on a schedule? If yes and there’s no API — build the Chrome automation, but document it and accept the maintenance cost.
    3. Is this a one-off task? If yes — Chrome is fine. Don’t over-engineer it.
    4. Is the tool’s UI likely to change frequently? If yes — consider whether the maintenance burden of Chrome automation is worth it, or whether the right answer is to find a tool that has an API.

    The Hybrid Pattern

    In practice, the cleanest architectures use both. The API handles everything it can — content publishing, data retrieval, triggering events that have proper endpoints. Chrome handles the edges — the one tool that has no API, the platform that blocks programmatic access from outside a browser, the workflow step that’s UI-only.

    One pattern that recurs: the main pipeline runs via API. One step in the pipeline requires Chrome because a specific capability isn’t exposed through the API. Chrome handles that one step, hands off back to the API-driven pipeline. The rest of the automation doesn’t care that one step used a browser.

    A Note on Reliability Expectations

    When you use Claude in Chrome for automation, set your reliability expectations accordingly. API-based automation can be built for 99%+ reliability. Chrome UI automation — against live web pages that change over time — is closer to 80-90% on any given run, and requires periodic maintenance. Plan for failures. Build retry logic. Log what fails. Don’t build a critical dependency on a Chrome automation without a manual fallback for the days when it breaks.

    ⚠️ Don’t chain high-stakes actions through Chrome automation without a review step. If your Chrome automation sequence ends in an irreversible action — sending a message, submitting a payment, publishing content publicly, deleting data — build in a confirmation step that requires your review before Claude executes the final action. Chrome automation moves fast. A misconfigured step in a chain can cause real consequences before you notice.

    The Summary

    Use the API when it exists and covers what you need. Use Claude in Chrome when the API doesn’t exist, doesn’t cover what you need, or when the task is genuinely one-off. Combine them when the right architecture calls for it. Neither is always better — they serve different parts of the same problem.

    Frequently Asked Questions

    Is Claude in Chrome slower than using the API?

    Yes. Browser UI automation runs at human browsing speed — navigating pages, waiting for elements to render, clicking through workflows. API calls are typically orders of magnitude faster for equivalent operations when an API exists.

    Can I mix API calls and Claude in Chrome actions in the same Claude session?

    Yes. Claude Chat can make API calls and also have Claude in Chrome connected in the same session. This is actually the most common pattern — Claude Chat handles API logic and writes work orders, Chrome handles the UI execution steps that the API can’t reach.

    If a tool has both an API and a web UI, should I ever use Chrome?

    Rarely, but sometimes yes. If the specific action you need isn’t available through the API even though the tool has one — or if you’re doing a one-off test and don’t want to write integration code — Chrome is a reasonable shortcut. For anything recurring, build the API integration instead.

    What happens when a site changes its UI and breaks my Chrome automation?

    Claude in Chrome will typically report that it couldn’t find an expected element or that the page doesn’t look as described. It won’t guess and won’t take unintended actions. You’ll need to update the instructions to reflect the new UI state.

    Is there a way to make Chrome automations more resilient to UI changes?

    Writing instructions in terms of intent rather than specific element names helps. “Find the button that saves the record” is more resilient than “click the blue Save button in the upper right corner” — though both will eventually break if the UI changes significantly. There’s no substitute for periodic maintenance of Chrome-based automations.

  • The Article-to-Video Pipeline — How We Automate Video Creation With Claude in Chrome

    The Article-to-Video Pipeline — How We Automate Video Creation With Claude in Chrome

    What This Pipeline Does
    Two scheduled Cowork tasks use Claude in Chrome to operate a browser-based notebook tool’s UI — creating notebooks, adding article sources, triggering video generation, downloading finished videos, and publishing watch pages to WordPress. Fully automated. Nobody clicks anything.

    This pipeline exists because a popular browser-based AI notebook tool generates high-quality cinematic videos from written content — but it has no API. The only way to operate it programmatically is through the browser UI. Claude in Chrome is the bridge.

    What follows is documentation of a running production pipeline, including the failure modes that actually occur and how they’re handled.

    The Architecture: Two Scheduled Tasks

    The pipeline runs as two complementary Cowork scheduled tasks, staggered 30 minutes apart on the same 3-hour cycle.

    Task 1 — Kickoff (runs at :00 on each scheduled hour)

    1. Calls the WordPress REST API to fetch recently published articles
    2. Checks the pipeline log (a Notion page) for articles already processed
    3. Selects one unprocessed article per run
    4. Uses Claude in Chrome to open the notebook tool in the browser
    5. Creates a new notebook, adds the article URL as a source
    6. Navigates to the video generation interface and triggers Cinematic generation
    7. Logs the article as “processing” in Notion with the notebook URL and timestamp

    Task 2 — Harvest (runs at :30 on each scheduled hour)

    1. Reads the Notion pipeline log for articles in “processing” status
    2. Filters for any that were kicked off more than 25 minutes ago
    3. Uses Claude in Chrome to open each notebook and check if the video is ready
    4. If ready: downloads the video file via Chrome
    5. Uploads the video to the WordPress media library via REST API
    6. Creates a draft watch page post with the embedded video, article summary, and schema markup
    7. Updates the Notion log to “completed”
    ⚠️ This pipeline requires Cowork Pro or Max. Scheduled, unattended Cowork tasks are a Pro/Max feature. Claude in Chrome itself is available on all plans, but this specific architecture — running tasks on a cron schedule without you being present — requires a paid Cowork subscription. If you’re on a lower tier, the same steps can be run manually through a Claude in Chrome session, but they won’t run automatically.

    The Account Rotation Layer

    Browser-based AI notebook tools typically impose daily limits on cinematic video generation per account. One account isn’t enough to process a continuous stream of articles.

    The pipeline handles this by rotating between two accounts. When the primary account hits its daily generation limit, the kickoff task switches to the secondary account. Both accounts have the notebook tool open in different Chrome profiles, with the extension installed in each.

    There’s also a notebook count limit per account. Old notebooks that have already been harvested get deleted periodically to stay under the cap.

    The Failure Modes — Documented From Production

    This is the part that most automation write-ups skip. Here are the real failure modes this pipeline encounters, in roughly descending frequency:

    Timeout (Most Common)

    Video generation on the notebook tool can take anywhere from 25 minutes to several hours, depending on server load. The harvest task has a 3-hour timeout window — if a video hasn’t finished after 3 hours, it’s marked as failed and the article is available for retry. In practice, a meaningful portion of generation runs take longer than the timeout window, especially during peak hours.

    Mitigation: failed articles are automatically available for re-kickoff in the next cycle.

    Chrome Tab Closure

    If the Chrome tab that Claude in Chrome is operating gets closed — by the user, by a browser crash, or by an accidental window close — Claude loses access and the harvest fails. The video may be ready in the notebook tool, but there’s no way to download it without re-establishing the browser connection.

    Mitigation: the pipeline marks the article as failed. Manual recovery: reopen the notebook tool in the correct Chrome profile, reinstall the extension if needed, and re-run the harvest for that article.

    ⚠️ Don’t close Chrome windows while a scheduled pipeline is running. Cowork scheduled tasks using Claude in Chrome depend on specific browser profiles staying open and connected. If you close a Chrome window that the pipeline is using, the running task will fail. If you’re setting up unattended runs, keep the relevant Chrome profiles open and don’t close them during the scheduled window. A dedicated browser profile that stays open is the cleanest solution.

    Daily Generation Limits

    Both accounts can hit their daily cinematic generation limit on high-volume days. When this happens, the kickoff task will fail to start new videos until the limit resets — which happens on a daily cycle. The pipeline logs these failures with a clear reason so they’re easy to spot.

    Mitigation: add a third account if volume consistently exceeds two accounts’ daily limits.

    Notebook Count Limits

    Notebook tools cap how many notebooks a single account can hold. When an account is at its limit, new notebook creation fails. Regular deletion of completed notebooks (those that have been harvested) keeps the account under the cap.

    What the Watch Page Looks Like

    After a successful harvest, the pipeline creates a draft WordPress post with:

    • The embedded video (hosted in the WordPress media library, not on an external service)
    • A summary of the source article
    • Chapter/segment markers if the tool generates them
    • Article schema markup
    • A link back to the original article

    The post goes up as a draft, not published directly. A manual review step before publishing is intentional — the pipeline produces a lot of content, and a spot check catches cases where generation quality was unexpectedly low.

    Why This Is Genuinely Novel

    The combination of Cowork scheduling + Claude in Chrome + a browser-based tool with no API is a pattern that isn’t widely documented. Most automation examples assume APIs exist. This one doesn’t — it treats the browser UI as the API, and Claude in Chrome as the adapter layer.

    The practical result: a pipeline that runs on a schedule, processes a backlog of articles at a rate of one per run, handles account rotation automatically, logs its own state, and surfaces failures with enough detail to recover from them manually.

    The tools involved are off-the-shelf. What makes it work is the architecture.

    Frequently Asked Questions

    Does the notebook tool need to be open in Chrome for this to work?

    Yes. Claude in Chrome navigates to the notebook tool in the browser — the tool doesn’t need to be pre-opened before the task starts, because Claude can navigate to it. But the Chrome profile where the extension is installed must be open and the profile must be logged in to the notebook tool’s account.

    What happens if a video takes longer than the timeout window to generate?

    The pipeline marks it as failed. The article becomes available for retry in the next kickoff cycle. There’s no penalty — the notebook still exists in the tool with generation in progress, so if you check manually and the video finishes later, you can also harvest it by hand.

    Can this pattern be adapted for other browser-based tools with no API?

    Yes. The two-task kickoff/harvest pattern applies to any browser-based tool where you’re triggering a process that takes time to complete. The specific steps change, but the architecture — trigger, wait, harvest, log — is reusable.

    Are the watch page posts published automatically?

    No. The pipeline creates them as drafts. A manual review step is built in before anything goes live. This is intentional — automated generation at scale benefits from a human spot-check before publishing.

    What do I do if a harvest fails because a Chrome tab was closed?

    Reopen the relevant Chrome profile. Make sure the Claude in Chrome extension is installed and active in that profile. Log in to the notebook tool if the session has expired. Then manually trigger a harvest for the specific article — open the notebook, confirm the video is ready, download it, and upload it to WordPress.

  • Claude in Chrome Across Multiple Chrome Profiles — The Multi-Account Workflow

    Claude in Chrome Across Multiple Chrome Profiles — The Multi-Account Workflow

    What This Covers
    Chrome profiles are separate browser identities — different logins, different extensions, different sessions. Claude in Chrome connects to one profile at a time via a manual click. Here is how to set that up for multi-account work, and where the friction still lives.

    Chrome profiles are one of Chrome’s most useful and most underused features. Each profile is an isolated browser identity: its own login state, its own saved passwords, its own open tabs, its own extensions. If you manage multiple Google accounts, multiple work environments, or need to keep different service logins separate, profiles are how you do it.

    Claude in Chrome works at the profile level. Understanding that changes how you think about setting it up.

    Each Chrome Profile Is Its Own Island

    When Claude in Chrome connects to a session, it connects to a specific Chrome profile — the one you’re running the extension in, the one where you clicked Connect. It can navigate any tab open in that profile. It cannot see or interact with tabs in other profiles, even if those profiles are open in other windows on your screen.

    This isolation is actually useful. It means you can set up dedicated Chrome profiles for different purposes:

    • One profile logged in to your primary work tools
    • One profile for a client’s services or a specific platform
    • One profile for personal accounts you don’t want mixed into work sessions

    When you want Claude to work in a specific environment, you connect it to that profile. It only sees what that profile sees.

    ⚠️ The extension must be installed on each profile separately. Installing Claude in Chrome on one profile does not install it on others — Chrome isolates extensions per profile. If you set up five profiles and want Claude to be available on all of them, you need to install and connect the extension five times. Check that it’s installed and active before starting any session.

    How switch_browser Works Across Profiles

    When Claude calls the switch_browser tool, it broadcasts a connection request to all Chrome instances that currently have the Claude in Chrome extension installed and active. Every eligible browser window shows a Connect prompt.

    You click Connect on the profile you want Claude to use. That profile becomes the active connection. The other windows are unaffected.

    A few practical notes:

    • Only one profile is connected at a time. Claude doesn’t maintain simultaneous connections to multiple profiles. If you need Claude to work in a different profile mid-session, it calls switch_browser again, and you click Connect in the new target.
    • The connection requires a manual click every time. Claude cannot silently hop between profiles. Each switch requires your action. This is intentional — it keeps you in control of which environment Claude is accessing at any given moment.
    • Pre-login matters. Once connected, Claude can only interact with services you’re already logged in to in that profile. Log in before the session starts, not during.

    A Working Multi-Profile Workflow

    In documented use, the multi-profile workflow looks like this:

    1. Open the Chrome profiles you’ll need for the session — each in its own window
    2. Log in to all the services you’ll need in each profile
    3. Confirm the Claude in Chrome extension is installed and active in each profile you’ll use
    4. Tell Claude Chat what you need done and which profile/environment to start in
    5. Claude calls switch_browser — you click Connect in the right profile
    6. Claude executes the task in that profile
    7. If you need Claude to switch profiles, it calls switch_browser again — you click in the next profile

    The manual click at each switch is the main friction point. It means truly automatic profile-hopping isn’t possible — Claude can initiate the switch, but you have to authorize it each time.

    ⚠️ Be deliberate about which profile you click Connect in. If you have multiple profiles open and multiple Connect prompts appear simultaneously, it’s easy to click the wrong one. The simplest prevention: when switch_browser fires, close or minimize the windows for profiles you don’t want Claude to access before clicking Connect. You can also open only the profile you need at that moment, run the task, then open the next one.

    The Chrome Profile Mapping Idea

    One capability that doesn’t exist yet but is worth building: a Chrome Profile Mapping skill that tells Claude which profile has which services logged in. Right now, Claude has to be told at the start of each task — “the Google account is in Profile 2, the platform admin is in Profile 4.” With a profile map, Claude would know this from context and could request the right profile without you specifying it every time.

    The idea is filed. It’s a one-time setup that would pay off across every multi-profile session afterward.

    How Many Profiles Is Practical?

    There’s no technical limit, but practical friction increases with the number of profiles you’re managing. The manual click requirement means every profile switch is a human action. Sessions that require frequent switching between more than two or three profiles become difficult to sustain without losing track of where Claude is.

    For most multi-account workflows, two to three profiles covers what’s needed: one for the primary environment, one or two for secondary services or client contexts. Beyond that, the workflow tends to benefit from being broken into separate sessions rather than one continuously switching session.

    Frequently Asked Questions

    Can Claude switch between Chrome profiles without me clicking anything?

    No. Every profile switch requires you to click Connect in the target profile. Claude can request the switch by calling switch_browser, but it cannot complete the connection without your action. This is a deliberate design decision, not a technical limitation that will be worked around.

    Do I need to install the Claude in Chrome extension on every profile?

    Yes. Chrome extensions are isolated per profile. The extension must be installed separately on each profile where you want Claude in Chrome to be available.

    What happens if I have multiple Chrome profiles open and I click Connect in the wrong one?

    Claude will connect to whichever profile you clicked in. If you realize you connected to the wrong one, disconnect, call switch_browser again, and click Connect in the correct profile. There’s no automatic way to undo actions Claude took while connected to the wrong profile, so stay attentive when multiple profiles are open.

    Can Claude be connected to two Chrome profiles at the same time?

    No. Claude in Chrome maintains one active connection at a time. To work in a different profile, you switch — which disconnects the current one.

    Is it safe to have Claude connected to a profile that’s logged in to my personal Google account?

    Use judgment. Claude in Chrome can see and interact with any tab open in the connected profile. If your personal profile has Gmail, Google Drive, or other personal services open, Claude has access to those tabs during the session. If you don’t want Claude to interact with personal accounts, use a dedicated work profile for Claude sessions and keep personal tabs in a separate profile that isn’t connected.

  • How to Use Claude in Chrome to Write Directly to a Web App

    How to Use Claude in Chrome to Write Directly to a Web App

    The Pattern
    Claude Chat writes the work order. Claude in Chrome navigates the UI and executes it. This combination lets you automate web apps that have no API — or where the API doesn’t expose what you need.

    A lot of the most useful tools on the web don’t have APIs. Or they have APIs, but specific features — a particular button, a workflow trigger, a UI-only setting — aren’t exposed through them. For years, the workaround was Zapier, custom scripts, or doing it manually.

    Claude in Chrome opens a different path: Claude navigates the UI directly, the same way you would, but you don’t have to be the one clicking.

    How the Two-Claude Pattern Works

    The workflow that works well in practice uses two Claude instances working together:

    1. Claude Chat (the claude.ai interface) handles planning, writing, API calls, and generating the specific instructions for what needs to happen in the browser
    2. Claude in Chrome (the browser extension) receives those instructions and executes them directly in the web app UI

    The typical flow: you describe the task to Claude Chat. Claude Chat writes a precise, step-by-step work order — what page to navigate to, what to click, what to fill in, what to confirm. You paste that into Claude in Chrome. Claude in Chrome executes it in the browser.

    It’s not magic. It’s division of labor: reasoning on one side, execution on the other.

    Real Situations Where This Applies

    In documented use, the Claude Chat → Chrome pattern has been used for:

    • Cloud console navigation — walking through multi-step infrastructure setup in a browser-based cloud console where the relevant actions weren’t exposed through the provider’s CLI or API
    • Domain registrar settings — updating DNS records through a registrar’s web interface. The registrar had an API, but the specific record type needed wasn’t in it.
    • Social scheduling tools — posting or scheduling content through a platform’s web UI when the API tier available didn’t include the scheduling endpoint
    • Web-based terminal environments — operating Cloud Shell or browser-based terminals without switching windows or copy-pasting
    • Browser-based AI notebook tools — creating notebooks, adding source URLs, navigating to generation features, and triggering video or audio generation through a UI

    The common thread: a logged-in browser session was required, and the action wasn’t available through an API.

    ⚠️ Pre-login before you hand off. Claude in Chrome can only interact with services where you’re already logged in in that Chrome profile. If Claude navigates to a page that requires a login it doesn’t have, it will stall or hit an error. Log in to every service you intend to use before starting the session, and make sure the session hasn’t expired. Also: close any tabs with services you don’t want Claude to interact with during this task.

    What Makes a Good Work Order

    The quality of the Chrome execution depends heavily on the quality of the instructions Claude Chat produces. A good work order is:

    • Sequential. Each step follows the last. Claude in Chrome doesn’t skip around.
    • Specific about UI elements. “Click the blue Save button in the upper right” is better than “save it.”
    • Includes what to do if something unexpected appears. Login screen, confirmation dialog, error message — Claude in Chrome handles these better if the work order anticipates them.
    • Ends with a confirmation step. “After completing, read the page and report what you see” closes the loop so you know whether the task actually finished.

    Claude Chat is good at generating this kind of structured instruction when you describe the task well. Give it the context of what tool you’re working in, what you’re trying to accomplish, and what you expect the UI to look like.

    The API-First Rule

    Using Claude in Chrome to operate a web UI is slower and less reliable than using an API. UI layouts change. Buttons get renamed. A platform update can break a workflow that worked yesterday.

    The rule that holds up in practice: API first, Chrome when the API fails or doesn’t exist.

    If a tool you use regularly exposes the action you need through an API, build the API integration and use that. Chrome UI automation is the fallback — valuable and often the only option, but a fallback nonetheless. Don’t default to Chrome just because it’s faster to set up today.

    ⚠️ Don’t leave Claude in Chrome running on high-stakes UI actions without reviewing first. If your work order includes steps like submitting a payment form, publishing content publicly, deleting records, or sending a message — review the work order carefully before Claude executes it, and stay present during execution. UI actions in Claude in Chrome are real. There is no undo button built in.

    When the Work Order Approach Doesn’t Work Well

    A few situations where the Claude Chat → Chrome hand-off runs into friction:

    • Dynamic UIs with inconsistent layouts. If the UI renders differently based on account state, screen size, or A/B tests, Chrome may not find the element the work order described.
    • Multi-factor authentication prompts. If a service triggers MFA mid-session, Chrome will stall waiting for input. You need to be present to handle it.
    • Very long multi-step tasks. The longer the chain of actions, the more likely something unexpected will interrupt it. For long tasks, build in manual check points rather than treating the whole thing as one uninterrupted run.
    • Anything involving CAPTCHA. Chrome cannot solve CAPTCHAs. Tasks that require CAPTCHA completion need manual intervention at that step.

    Frequently Asked Questions

    Does Claude in Chrome work with any website?

    It works with any website loaded in Chrome where you have the appropriate access. The extension interacts with the live DOM of whatever page is open. Some sites use security measures that prevent external scripts from interacting with certain elements, which can limit what Claude can click or read on those pages.

    Can Claude in Chrome interact with pop-up windows or modal dialogs?

    Yes, in most cases. Pop-ups and modals that are part of the page’s DOM are accessible. Browser-level dialogs (like the native file picker or browser alert boxes) have more limited interaction.

    What if the UI changes and Claude can’t find an element?

    Claude in Chrome will report that it couldn’t find the element and stop. It won’t guess or click something random. You’ll need to update the work order to reflect the current UI, or manually navigate to the right state and then reconnect.

    Is there a risk of Claude submitting forms I don’t want submitted?

    Yes, if the work order includes a form submission step. Always review work orders that include submit, confirm, send, or delete actions before execution. If you’re uncertain, break the work order into stages and review what Claude has done before authorizing the next stage.

    Can I use Claude in Chrome for a tool I use for work with sensitive data?

    Use judgment. Claude in Chrome processes what it sees in the browser tab, and the content of that interaction is processed by Anthropic’s systems under your account’s privacy settings. Review Anthropic’s privacy policy for your plan before using Claude in Chrome with tools containing confidential, regulated, or personally identifiable information.

  • Claude in Chrome vs Cowork Computer Use — What’s the Difference

    Claude in Chrome vs Cowork Computer Use — What’s the Difference

    The Short Version
    Claude in Chrome = browser only, any plan, you stay present. Cowork computer use = full desktop, scheduled, unattended, Pro or Max required. They solve different problems. The confusion comes from using the word “automation” for both.

    If you’ve tried Claude in Chrome and also explored Cowork’s computer use feature, you’ve probably noticed they feel completely different — even though both involve Claude “doing things” on a computer. That’s because they are fundamentally different tools, with different scope, different risk levels, and different use cases.

    This comparison is built from documented use of both. Not marketing copy.

    The Core Difference: Browser vs. Desktop

    Claude in Chrome operates exclusively inside the Chrome browser. It can read pages, click elements, fill forms, scroll, download files, and navigate between open tabs. That’s it. It has no awareness of your desktop, no access to your filesystem, and no ability to open applications outside the browser.

    Cowork computer use operates at the full desktop level. It can see and interact with any application on your machine — your file manager, terminal, spreadsheet software, desktop apps, system utilities. It treats your entire computer as its workspace.

    The practical difference: if you close Chrome, Claude in Chrome stops. If you close Chrome while Cowork computer use is running, Cowork keeps going in other applications.

    Scheduling and Presence

    Feature Claude in Chrome Cowork Computer Use
    Scope Browser only Full desktop
    Can run scheduled / unattended No Yes
    Requires you to be present Yes No (once configured)
    Available on free plan Yes No
    Requires Pro or Max No Yes
    Access to filesystem No Yes
    Can open desktop applications No Yes
    Connection method Manual click to connect Configured per task

    When Chrome Is the Right Tool

    Claude in Chrome is the better choice when:

    • The tool you’re working with is entirely browser-based and has no API (or an API that doesn’t expose what you need)
    • You want to work alongside Claude in real time — you’re co-piloting, not delegating
    • The task is one-off or occasional, not something you need to run on a schedule
    • You want Claude to interact with a logged-in browser session that you control
    • You’re on any Claude plan and don’t have access to Cowork computer use
    ⚠️ Stay present with Chrome. Claude in Chrome is not designed for unattended use. If Claude clicks something unexpected or a form submits mid-session, you need to be there to intervene. This isn’t a limitation you can safely work around by walking away — it’s the intended operating model.

    When Cowork Computer Use Is the Right Tool

    Cowork computer use is the better choice when:

    • The task needs to repeat on a schedule — daily, every few hours, weekly
    • The task spans multiple applications (browser plus desktop app plus filesystem)
    • You want it to run without you being present
    • The task involves file operations — reading, writing, moving, processing local files
    • You need multi-step pipelines that chain browser actions with non-browser actions
    ⚠️ Unattended computer use has a wider blast radius. When Cowork computer use runs a scheduled task, it has access to your full desktop — including applications, files, and anything else open on your machine. A misconfigured task or an unexpected UI change on a target website can cause Claude to interact with things it wasn’t supposed to. Review what’s open on your machine before scheduling unattended runs, and test new tasks manually before letting them run on a schedule.

    They Can Work Together

    One pattern that works well in practice: Claude Chat writes the instructions, Claude in Chrome executes the browser-side steps. Cowork handles the scheduled, recurring, multi-app pieces.

    Think of it as a three-tier model. Claude Chat is strategy and orchestration. Claude in Chrome is the field operator for browser-native tasks that require a logged-in session or a UI that has no API. Cowork is the autonomous layer for scheduled, repeating, multi-system work.

    A task that’s “too small for Cowork but too tedious to do manually” is usually a Claude in Chrome task. A task that runs every night at 11pm is usually a Cowork task. Most workflows eventually use all three.

    The Decision Rule

    One question resolves most cases: do you need it to run while you’re asleep?

    If yes — Cowork computer use (Pro or Max required).
    If no — Claude in Chrome, from any plan, with you present.

    Frequently Asked Questions

    Can I use Claude in Chrome instead of Cowork computer use to save money?

    For one-off browser tasks, yes — Claude in Chrome is available on all plans and covers a meaningful range of browser automation. But it can’t replace Cowork computer use for scheduled tasks, unattended runs, or anything that requires filesystem or desktop application access.

    Does Claude in Chrome work inside a Cowork session?

    They’re separate features. Claude in Chrome is a browser extension that works in claude.ai chat sessions. Cowork computer use is a separate capability within the Cowork product. They don’t directly compose with each other, though you can use both in complementary workflows.

    Is Cowork computer use riskier than Claude in Chrome?

    The surface area is larger with Cowork computer use because it has access to your full desktop, not just the browser. Whether that translates to more risk depends entirely on how you configure and test your tasks. Well-tested Cowork tasks running on a focused setup can be lower risk than an untested Claude in Chrome session with sensitive tabs open. The tool isn’t the risk — how you set it up is.

    Can Claude in Chrome run overnight or on a schedule?

    No. Claude in Chrome requires an active chat session and a manual connection per session. It is not designed for scheduled or unattended use. For overnight or scheduled automation, you need Cowork computer use.

    Which one should I start with?

    If you’re new to both, start with Claude in Chrome. It’s available on all plans, the blast radius is limited to your browser, and you stay in the loop during every session. Once you’re comfortable with how Claude navigates browser-based tools, you’ll have a much better sense of whether Cowork’s scheduled automation is worth setting up for your specific workflows.

    Related: How Claude Cowork Can Actually Train Your Staff to Think Better — a 7-part series on using Cowork as a training tool across industries.

  • What Is Claude in Chrome and How Does It Actually Work

    What Is Claude in Chrome and How Does It Actually Work

    Claude in Chrome — Quick Definition
    Claude in Chrome is a browser extension that gives Claude direct control over your active Chrome tab. It can read page content, click buttons, fill forms, scroll, and download files — all inside the browser, without touching your desktop or filesystem.

    There are now three distinct ways to work with Claude at the task level: through the chat interface, through Claude Cowork, and through Claude in Chrome. Most people know the first two. The third one is genuinely different, and genuinely useful — and most people writing about Claude haven’t actually used it yet.

    This article is built from documented operational use. Not theory.

    What Claude in Chrome Actually Is

    Claude in Chrome is a browser extension — separate from claude.ai, separate from Cowork — that connects Claude to your active Chrome tab. Once the extension is installed and connected, Claude gains a set of browser-native tools it doesn’t have in a standard chat session.

    Those tools include:

    • Reading page content — Claude can see what’s on the current tab, including text, links, form fields, and interactive elements
    • Clicking — Claude can click buttons, links, checkboxes, and UI controls
    • Filling forms — Claude can type into text fields, dropdowns, and inputs
    • Scrolling — Claude can scroll a page to load more content or navigate to a section
    • Downloading files — Claude can trigger downloads from web interfaces
    • Navigating — Claude can move between tabs that are open in the connected profile
    ⚠️ Before you experiment: When Claude has browser control, it can interact with any tab in the connected Chrome profile — including tabs where you’re logged in to banking, email, or other sensitive services. Before running any Claude in Chrome session, close or move tabs you don’t want Claude to have access to. Pre-login only to the services you intend to use in that session.

    What Claude in Chrome Is Not

    It’s worth being precise here, because there’s real confusion between Claude in Chrome and Claude Cowork’s computer use feature.

    Claude in Chrome is browser-only. It operates inside Chrome. It cannot access your filesystem, run terminal commands, open desktop applications, or do anything outside a browser window. If you need Claude to interact with files on your computer or run code locally, that’s a different tool entirely.

    Claude Cowork computer use is full-desktop. Cowork’s computer use feature gives Claude access to your entire desktop environment — applications, filesystem, terminal, everything. It’s also scheduled and can run unattended. That’s a much larger surface area.

    The comparison matters because the risk profile is different. Browser-only means the blast radius of any mistake is limited to what’s accessible through Chrome. Full computer use is a fundamentally different level of access. More on this comparison in the full breakdown article.

    How the Connection Works

    Claude in Chrome uses a tool called switch_browser. When Claude calls this tool, it broadcasts a connection request to all Chrome instances that have the extension installed. A small prompt appears in the browser — you click Connect — and Claude is now operating in that Chrome profile.

    A few things to understand about how this works in practice:

    • One profile at a time. Claude connects to one Chrome profile per session. If you have multiple Chrome profiles open, the connection goes to whichever one you click Connect in.
    • The extension must be installed on each profile separately. Chrome profiles are isolated environments. Installing the extension in one profile doesn’t propagate it to others.
    • The connection requires a manual click. This is intentional friction — Claude can’t silently connect to a Chrome profile without your action. You will always know when Claude is taking browser control.
    • Once connected, Claude can navigate between open tabs freely within that profile.
    ⚠️ Don’t walk away during a session. Claude in Chrome is designed for working with a human present. If Claude navigates to a tab where you’re logged in to a web app and something goes wrong — a form submits, an action fires — you need to be there to catch it. This is different from Cowork scheduled tasks, which are designed to run unattended. Treat Claude in Chrome sessions like you’re co-piloting, not delegating.

    What It’s Useful For

    Claude in Chrome’s sweet spot is situations where there’s no API. A lot of useful web tools — dashboards, admin panels, third-party platforms — don’t offer an API, or their API is locked behind an enterprise plan, or the specific action you need isn’t exposed via API even if the tool has one.

    In documented use, Claude in Chrome has been used to:

    • Navigate cloud console interfaces that require clicking through menus
    • Interact with domain registrar admin panels to update DNS settings
    • Operate social media scheduling tools through their web UI when the API doesn’t expose the specific feature needed
    • Use web-based terminal environments where copy/paste would be the alternative
    • Run automated notebook workflows in browser-based AI tools — creating notebooks, adding sources, triggering generation, downloading output

    The pattern is consistent: API first, Chrome when the API doesn’t exist or is blocked. Chrome is the fallback, not the default. But it’s a very capable fallback.

    Available on All Claude Plans

    One thing that surprises people: Claude in Chrome is available to all Claude subscribers, not just Pro or Max. This is different from Cowork computer use, which requires Pro or Max.

    If you’re on a free plan, you can still install the extension and use browser control in your chat sessions. The session limits of your plan still apply, but the capability itself isn’t gated.

    The Right Mental Model

    The cleanest way to think about Claude in Chrome: it’s Claude with a mouse and keyboard, but only inside the browser, and only when you hand it control.

    That framing clarifies both the power and the limits. It’s not autonomous. It doesn’t run in the background. It doesn’t have memory of previous browser sessions. Every connection is a deliberate, per-session handoff. You stay in the loop.

    When you need Claude to do something in a browser-based tool and you’re willing to be present while it runs — Claude in Chrome is the right tool. When you need scheduled, unattended, multi-application automation — that’s Cowork territory.

    Frequently Asked Questions

    Do I need a paid Claude plan to use Claude in Chrome?

    No. Claude in Chrome is available on all Claude plans, including free. You’ll still be subject to your plan’s message limits, but the browser control capability itself is not restricted to paid tiers.

    Can Claude in Chrome access my files or run programs on my computer?

    No. Claude in Chrome operates only inside the Chrome browser. It cannot access your filesystem, open desktop applications, or run terminal commands. If you need Claude to interact with files or run code locally, you’re looking for a different tool.

    Is it safe to use Claude in Chrome while logged in to sensitive accounts?

    Use caution. When Claude in Chrome is connected to a Chrome profile, it can see and interact with all open tabs in that profile — including any tabs where you’re logged in to banking, email, or other sensitive services. Best practice is to pre-close tabs you don’t want Claude to have access to before starting a session, and to stay present during the session.

    Can Claude connect to Chrome automatically without me doing anything?

    No. Every connection requires a manual click. When Claude calls the switch_browser tool, a Connect prompt appears in the browser — you have to click it. Claude cannot silently establish a browser connection without your action.

    What’s the difference between Claude in Chrome and Claude Cowork computer use?

    Claude in Chrome is browser-only, works in any chat session, and is available on all plans. Cowork computer use gives Claude access to your entire desktop — applications, filesystem, terminal — and can run scheduled, unattended tasks. It requires a Pro or Max subscription. The choice depends on what you’re trying to automate and whether you need to be present.

    What happens if I close a Chrome tab while Claude in Chrome is using it?

    Claude will lose access to that tab. If the tab was part of an active task — for example, a browser-based notebook generating output — the task will fail or stall. You’ll need to reopen the tab, reconnect the extension, and restart the relevant step. It’s one of the reasons Claude in Chrome is designed for sessions where you stay present.

  • Claude Cowork Changelog: What Changed in Q1 2026

    Claude Cowork Changelog: What Changed in Q1 2026

    Claude AI · Tygart Media · Updated April 2026
    Q1 2026 summary: Cowork went from research preview to generally available. Computer use launched for Pro/Max users. Scheduled and recurring tasks shipped. The sessiondata.img disk-full bug (GitHub #30751) remained open all quarter — the workaround is manual. Plugin marketplace launched in April.

    Claude Cowork shipped more meaningful features in Q1 2026 than in any prior quarter. This is the complete log of what changed, what shipped, and what stayed broken — documented for teams managing Cowork deployments who need to know what actually changed and when.

    January 2026: Foundation Stability

    January was primarily infrastructure hardening. The Cowork runner environment received reliability improvements addressing the most common mid-task failures — streams aborting on slow API responses, sub-agent MCP tool inheritance failures, and session cleanup bugs that left stale working directories. No major feature launches, but the stability improvements reduced the frequency of mid-run failures that had characterized late 2025 Cowork usage.

    Claude Code received the iOS app in October 2025 and the web version — both of which fed into Cowork’s remote dispatch capabilities in Q1. By January, the ability to assign Cowork tasks from a phone was stable enough for regular use.

    February 2026: Model Upgrades Change Everything

    February 5: Claude Opus 4.6 launched. February 17: Claude Sonnet 4.6 launched. Both significantly improved Cowork task quality — particularly for long-horizon agentic sessions where the original 4.0 models would lose coherence mid-task. Sonnet 4.6’s dramatically improved computer use capability (scoring 72.7% on OSWorld) made computer-use Cowork tasks reliable for the first time. Tasks that previously required constant human intervention to stay on track became genuinely autonomous.

    The 1M token context window entered beta on both models in February, enabling Cowork tasks to hold significantly more context across long sessions — particularly valuable for content pipelines processing large document sets or cross-database synthesis tasks in Notion.

    March 2026: Computer Use Reaches Cowork

    March brought the integration of computer use into Cowork for Pro and Max plan users. Claude gained the ability to open files, navigate browsers, click through interfaces, and operate software within Cowork sessions — no additional setup required for Pro/Max subscribers. This was the most significant capability expansion of the quarter: Cowork tasks could now interact with software that doesn’t have an API, including legacy desktop applications and web interfaces without structured data access.

    Dispatch — Cowork’s task queue feature — was extended to support computer use actions, allowing scheduled tasks to include browser automation and desktop interaction steps alongside the existing MCP tool calls and bash operations.

    The Cowork VM disk-full bug (GitHub issue #30751) was acknowledged by Anthropic during March but not resolved. Power users with many skills installed continued to hit the useradd: cannot create directory error every 40-50 sessions. The documented workaround — moving sessiondata.img to reset the VM — remained the only fix. See the full fix guide.

    April 2026: General Availability

    Cowork reached general availability on macOS and Windows via Claude Desktop in April, removing the “research preview” label it had carried since launch. The GA release added enterprise features that had been absent from the preview: usage analytics, OpenTelemetry support for monitoring Cowork activity, and role-based access controls for Enterprise plans allowing admins to define which capabilities each team group can access.

    A plugin marketplace launched for Team and Enterprise plans with admin controls. Admins can now approve, restrict, or block specific plugins org-wide. The Customize section in Claude Desktop was reorganized to group skills, plugins, and connectors in one place.

    Scheduled and recurring task creation was formalized in the UI — previously requiring config file editing, now accessible from within the app. This was the feature most requested by Cowork power users throughout Q1.

    What Remained Broken Through Q1

    The sessiondata.img disk-full bug was the most significant ongoing issue. It affected every power user with a substantial skill library and required periodic manual intervention. No automatic session cleanup shipped in Q1. The manual workaround is documented at Claude Cowork useradd Failed Error Fix.

    Machine-sleep task skipping also remained unresolved — scheduled tasks that fire when a machine is asleep are silently skipped with no retry. Teams running reliable scheduled automation continued to need an always-on machine or a cloud-side solution.

    Q2 2026 Outlook

    The disk-full bug fix and automatic session cleanup are the most anticipated Q2 items. Agent teams (available on Max plans) are expected to expand with better orchestration tooling. Claude 5, expected Q2-Q3, will bring model quality improvements that should further improve long-horizon Cowork task reliability.

    When did Claude Cowork become generally available?

    Claude Cowork reached general availability on macOS and Windows in April 2026. It had been in research preview since its initial launch in late 2025.

    What was the biggest Cowork improvement in Q1 2026?

    The February launch of Claude Sonnet 4.6 and Opus 4.6 most improved Cowork task quality — especially computer use tasks, which became reliably autonomous with Sonnet 4.6’s improved OSWorld scores. March brought computer use to Cowork for Pro/Max users directly.

    Was the Cowork disk-full bug fixed in Q1 2026?

    No. GitHub issue #30751 (sessiondata.img filling up) remained open through Q1 2026. The manual workaround — moving sessiondata.img to reset the VM — is the only fix as of April 2026.

    Related: How Claude Cowork Can Actually Train Your Staff to Think Better — a 7-part series on using Cowork as a training tool across industries.


  • WordPress REST API for Publishers: How to Connect Claude to WordPress Without Plugins

    WordPress REST API for Publishers: How to Connect Claude to WordPress Without Plugins

    Claude AI · Tygart Media
    What this enables: Publishing articles to WordPress programmatically from Claude, Python scripts, GCP Cloud Run jobs, or any HTTP client — without plugins, without Elementor, without touching the WP admin. The same pipeline that powers 27+ managed sites publishing thousands of articles per month.

    WordPress has a fully functional REST API built in. Most people never use it because they don’t know it’s there. For publishers, content operations teams, and anyone running Claude-powered content workflows, the REST API is the infrastructure that eliminates manual publishing and enables automation at scale. Here’s how it works and how to wire Claude to it.

    What the WordPress REST API Can Do

    The REST API exposes every major WordPress function over HTTP: create posts, update posts, get posts, manage categories and tags, upload media, manage users. Every action you can take in the WordPress admin can be done via API call. No plugin required — it’s built into WordPress core since version 4.7.

    Authentication: Application Passwords

    The simplest authentication method for Claude-to-WordPress connections is WordPress Application Passwords — a built-in feature (WordPress 5.6+) that generates a dedicated password for API access without exposing your main login credentials.

    To generate one: WP Admin → Users → Your Profile → Application Passwords → enter a name → click Add New. Copy the generated password immediately — it’s only shown once. The format it gives you has spaces; remove them before using in API calls.

    Authenticate using HTTP Basic Auth:

    Authorization: Basic base64(username:app_password)

    Publishing a Post via API

    A complete post publish call:

    POST https://yoursite.com/wp-json/wp/v2/posts
    Authorization: Basic [base64 credentials]
    Content-Type: application/json

    {
      "title": "Your Post Title",
      "content": "<p>Full HTML content here</p>",
      "excerpt": "Your SEO meta description (140-160 chars)",
      "status": "publish",
      "categories": [5, 12],
      "tags": [34, 67, 89],
      "slug": "your-post-slug"
    }

    The response returns the new post ID and URL. Log these — you need the post ID for any subsequent updates.

    Wiring Claude Into the Pipeline

    The standard Claude-to-WordPress pipeline: Claude generates the article content (with SEO optimization, schema markup, and FAQ sections baked in), a Python or Node.js script assembles the API payload, the payload POSTs to the WordPress REST endpoint, and the response confirms publication. For Cowork tasks, this runs on a schedule without human intervention.

    The critical rule: Notion first, WordPress second. Every article goes to a Notion page before publishing to WordPress. Notion is the storage and version control layer; WordPress is the distribution layer. If you ever need to republish, update, or audit, you have a source of truth that isn’t locked inside the WordPress database.

    Handling WAF Blocks

    Many managed WordPress hosts (WP Engine, SiteGround) run Web Application Firewalls that block API calls from cloud IP addresses. Symptoms: 403 Forbidden errors on POST requests, even with correct credentials. Two solutions: route API calls through a Cloud Run proxy service that presents a different IP profile, or whitelist your specific GCP IP range in the hosting provider’s WAF settings. For SiteGround specifically, direct whitelisting is the most reliable path — the proxy approach has mixed results due to SiteGround’s aggressive WAF configuration.

    Schema and SEO Metadata

    The WordPress REST API supports all Yoast SEO and Rank Math meta fields as post meta. To set SEO title, meta description, and schema markup programmatically, include the relevant meta fields in your POST payload. For Yoast: _yoast_wpseo_title and _yoast_wpseo_metadesc. For Rank Math: rank_math_title and rank_math_description. Inject JSON-LD schema directly into the post content as a <script type="application/ld+json"> block — it renders correctly on the front end and passes Google’s rich results validator.

    How do I publish to WordPress without logging in?

    Use the WordPress REST API with Application Password authentication. Generate an application password in WP Admin → Users → Your Profile, then POST to /wp-json/wp/v2/posts with Basic Auth credentials. No plugin required — the REST API is built into WordPress core.

    Can Claude publish directly to WordPress?

    Yes — through the WordPress REST API. Claude generates content, a script assembles the API payload, and the POST call publishes it. This is how automated content pipelines work at scale. Always write to Notion first; WordPress is the distribution layer.

    Why is my WordPress REST API returning 403?

    Most likely a WAF (Web Application Firewall) blocking the request — common on WP Engine and SiteGround. Either route API calls through a proxy service with a whitelisted IP or whitelist your specific IP range in the hosting provider’s firewall settings.

  • Claude on GCP: Billing, IAM, and Quota Setup for Teams

    Claude on GCP: Billing, IAM, and Quota Setup for Teams

    Claude AI · Tygart Media
    The three things teams get wrong: Using a shared GCP project for Claude and other workloads (makes cost attribution impossible), not requesting quota increases before launch (causes 429 errors at the worst time), and using overly broad IAM roles (security risk and audit problem). All three are fixable in an afternoon.

    Running Claude through Vertex AI on GCP is straightforward to set up for a solo developer. For a team deploying Claude in production, three infrastructure decisions matter significantly: project structure for billing, IAM configuration for access control, and quota management to avoid rate-limit failures. Here’s the setup that scales cleanly.

    Project Structure: One Project for Claude

    Create a dedicated GCP project for Claude workloads — separate from your main application project, your data pipeline project, and your development sandbox. This separation is the single most important decision for operational clarity. With a dedicated project you get: Claude API costs isolated on their own billing line, IAM permissions that only affect Claude access (not your entire infrastructure), quota limits and alerts scoped to Claude usage, and audit logs that only contain Claude-related activity.

    Naming convention: company-claude-prod for production, company-claude-dev for development. Keep them separate — dev workloads shouldn’t share quotas with production.

    IAM Configuration: Minimum Necessary Permissions

    The role that grants Claude API access through Vertex AI is roles/aiplatform.user. That’s the only role needed for model invocation and token counting. Don’t assign broader roles like roles/aiplatform.admin or roles/editor to service accounts that only need to call Claude.

    For team deployments, create one service account per application or environment — not one shared service account for everything. Example structure:

    Service Account Role Used By
    claude-prod-api@project.iam.gserviceaccount.com aiplatform.user Production app
    claude-dev-api@project.iam.gserviceaccount.com aiplatform.user Development
    claude-cowork@project.iam.gserviceaccount.com aiplatform.user Claude Code / Cowork

    If a service account is compromised, you rotate one key without affecting other applications. If a developer leaves, you disable their specific account without touching production credentials.

    Quota Management: Request Increases Before You Need Them

    Vertex AI Claude quotas are set conservatively by default. The default quota for most regions is enough for development and testing, but production workloads — especially automated pipelines running multiple requests per minute — will hit limits. The 429 error (Resource exhausted) at peak load is one of the most common production failure modes.

    Request quota increases before launch, not during an incident. Go to Cloud Console → IAM & Admin → Quotas, filter by “anthropic,” and request increases for the Claude models you’re deploying. Approval is typically same-day for standard business accounts. For the global endpoint, a good starting quota for a production team is 60 requests per minute for Sonnet 4.6 and 20 requests per minute for Opus 4.6.

    Budget Alerts: Know Before It’s a Problem

    Set a budget alert on your Claude GCP project before anything runs in production. Go to Billing → Budgets & Alerts, create a budget for the project, and set email alerts at 50%, 80%, and 100% of your expected monthly spend. Add a Pub/Sub notification if you want to automatically throttle or pause workloads when budget thresholds are hit.

    A Claude content pipeline running at unexpected volume can burn through budget quickly — especially with Opus 4.6 at $25/million output tokens. Budget alerts are the safety net that turns a potential billing surprise into a manageable alert.

    Cloud Logging: Keep the Audit Trail

    Vertex AI API calls are logged to Cloud Logging by default. For regulated industries, explicitly configure log retention to match your compliance requirements — the default 30-day retention may not be sufficient. For SOC 2 or HIPAA environments, export logs to Cloud Storage for long-term archival. The log entries include model called, project, timestamp, and token counts — enough for a complete audit trail without exposing prompt content.

    How do I set up billing for Claude on GCP?

    Create a dedicated GCP project for Claude workloads, set a budget alert before anything runs in production, and monitor spend at Billing → Budgets. Keeping Claude in its own project makes cost attribution clean and prevents unexpected spend from affecting other project budgets.

    What IAM role does Claude need on Vertex AI?

    The roles/aiplatform.user role is sufficient for model invocation and token counting. Use one service account per application or environment. Never assign broader roles like editor or aiplatform.admin to service accounts that only need to call Claude.

    How do I fix Claude 429 quota errors on Vertex AI?

    Go to Cloud Console → IAM & Admin → Quotas, filter by “anthropic,” and request a quota increase for the specific Claude model hitting limits. Request increases before production launch, not during an incident. Approvals are typically same-day for standard business accounts.