Category: GCP & Cloud Infrastructure

Practitioner guides to running AI-native operations on Google Cloud Platform — Cloud Run, Vertex AI, BigQuery, and the full stack.

  • Knowledge Cluster VM Setup — 5-Site WordPress Network on GCP Compute Engine

    What Is a Knowledge Cluster VM?
    A Knowledge Cluster VM is a single GCP Compute Engine instance running five WordPress sites on a shared LAMP stack — each site with its own domain, SSL certificate, and WordPress installation, all managed from one server with Claude Code deployed for AI-assisted content operations. Five sites, one VM, unified content architecture, fraction of the cost of five separate hosting accounts.

    Running five WordPress sites on five separate managed hosting accounts costs $200–$500/month and gives you five completely isolated environments with no shared infrastructure, no shared AI tooling, and no economies of scale. A dedicated GCP VM changes the math: one e2-standard-2 instance runs all five sites for around $30–$50/month, with Claude Code deployed directly on the server for zero-latency AI content operations.

    We run our own 5-site knowledge cluster this way — restorationintel.com, riskcoveragehub.com, continuityhub.org, bcesg.org, and healthcarefacilityhub.org are all on one VM. The hub-and-spoke content architecture connects them intentionally: each site covers a different facet of a shared knowledge domain, and internal cross-linking amplifies authority across all five.

    Who This Is For

    Operators building a network of related WordPress sites — knowledge hubs, geo-local networks, topic clusters across related domains — who want shared infrastructure, lower hosting costs, and a unified AI content operation rather than five separate managed accounts.

    What We Build

    • GCP Compute Engine VM — e2-standard-2 (2 vCPU, 8GB RAM) or larger depending on traffic requirements, configured in us-west1 or your preferred region
    • Shared LAMP stack — Apache with virtual hosts, MySQL with separate databases per site, PHP 8.x configured for WordPress
    • Five WordPress installations — Each in its own directory, individual wp-config, separate database credentials
    • SSL certificates — Certbot/Let’s Encrypt for all five domains with auto-renewal configured
    • Claude Code deployment — Anthropic API key stored in GCP Secret Manager, Claude Code installed and configured for WP-CLI integration
    • Hub-and-spoke content map — Architecture document defining which site is the hub, which are spokes, and the interlinking strategy
    • WP-CLI batch scripts — Common operations (plugin updates, bulk post operations, taxonomy management) scripted for all five sites

    What We Deliver

    Item Included
    GCP VM provisioning and configuration
    5 WordPress installations with SSL
    Shared LAMP stack with Apache virtual hosts
    Claude Code deployment + GCP Secret Manager integration
    Hub-and-spoke content architecture document
    WP-CLI batch operation scripts
    Monitoring + auto-restart configuration
    Technical handoff documentation

    Ready to Consolidate 5 Sites onto One Smart Server?

    Share the 5 domains you want to host and your current monthly hosting cost. We’ll scope the VM build and show you the cost reduction.

    will@tygartmedia.com

    Email only. No commitment to reply.

    Frequently Asked Questions

    What happens if the VM goes down?

    GCP Compute Engine has 99.9% uptime SLA. We configure automatic restart policies and GCP’s built-in monitoring with alerting. For production sites with stricter uptime requirements, we can add a load balancer with health checks.

    How is this different from WordPress Multisite?

    WordPress Multisite shares a single WordPress installation across all sites — changes to plugins or core affect all sites simultaneously and customization is limited. The cluster uses five independent WordPress installations that share only the server hardware. Each site is fully independent.

    Can more than 5 sites run on one VM?

    Yes — an e2-standard-2 instance comfortably handles 8–10 low-to-medium traffic WordPress sites. We scale the VM size based on your traffic requirements. The architecture pattern works for 3–15 sites.


    Last updated: April 2026

  • BigQuery Knowledge Ledger — Persistent AI Memory for Content Operations

    What Is a BigQuery Knowledge Ledger?
    A BigQuery Knowledge Ledger is a persistent AI memory layer — your content, decisions, SOPs, and operational history stored as vector embeddings in Google BigQuery, queryable in real time. When a Claude session opens, you query the ledger instead of re-pasting context. Your AI starts informed, not blank.

    Every Claude session starts from zero. You re-brief it on your clients, your sites, your decisions, your rules. Then the session ends and it forgets. For casual use, that’s fine. For an operation running 27 WordPress sites, 500+ published articles, and dozens of active decisions — that reset is an expensive tax on every session.

    The BigQuery Knowledge Ledger is the solution we built for ourselves. It stores operational knowledge as vector embeddings — 925 content chunks across 8 tables in our production ledger — and makes it queryable from any Claude session. The AI doesn’t start blank. It starts with history.

    Who This Is For

    Agency operators, publishers, and AI-native teams running multi-site content operations where the cost of re-briefing AI across sessions is measurable. If you’ve ever said “as I mentioned before” to Claude, you need this.

    What We Build

    • BigQuery datasetoperations_ledger schema with 8 tables: knowledge pages, embedded chunks, session history, client records, decision log, content index, site registry, and change log
    • Embedding pipeline — Vertex AI text-embedding-005 model processes your existing content (Notion pages, SOPs, articles) into vector chunks stored in BigQuery
    • Query interface — Simple Python function (or Cloud Run endpoint) that accepts a natural language query and returns the most relevant chunks for context injection
    • Claude integration guide — How to query the ledger at session start and inject results into your Claude context window
    • Initial seed — We process your existing Notion pages, key SOPs, and site documentation into the ledger on setup

    What We Deliver

    Item Included
    BigQuery dataset + 8-table schema deployed to your GCP project
    Vertex AI embedding pipeline (text-embedding-005)
    Query function (Python + optional Cloud Run endpoint)
    Initial content seed (up to 100 Notion pages or documents)
    Claude session integration guide
    Ongoing ingestion script (add new content to ledger)
    Technical walkthrough + handoff documentation

    Stop Re-Briefing Your AI Every Session

    Tell us how many sites, documents, or SOPs you’re managing and what your current re-briefing tax looks like. We’ll scope the ledger build.

    will@tygartmedia.com

    Email only. No sales call required.

    Frequently Asked Questions

    Does this require Google Cloud?

    Yes. BigQuery and Vertex AI are Google Cloud services. You need a GCP project with billing enabled. We handle all setup and deployment.

    What’s the ongoing cost in GCP?

    BigQuery storage for a 1,000-chunk ledger costs less than $1/month. Embedding runs (adding new content) cost fractions of a cent per chunk via Vertex AI. Query costs are negligible at typical session volumes.

    Can this work with tools other than Claude?

    Yes. The ledger is model-agnostic — it returns text chunks that can be injected into any LLM context. ChatGPT, Gemini, and Perplexity integrations all work with the same query interface.

    What format does my existing content need to be in?

    Notion pages (via API), plain text, markdown, or Google Docs. We handle the conversion and chunking during initial seed. PDFs and Word docs require an additional preprocessing step.

    Last updated: April 2026

  • GCP Content Pipeline Setup for AI-Native WordPress Publishers

    What Is a GCP Content Pipeline?
    A GCP Content Pipeline is a Google Cloud-hosted infrastructure stack that connects Claude AI to your WordPress sites — bypassing rate limits, WAF blocks, and IP restrictions — and automates content publishing, image generation, and knowledge storage at scale. It’s the back-end that lets a one-person operation run like a 10-person content team.

    Most content agencies are running Claude in a browser tab and copy-pasting into WordPress. That works until you’re managing 5 sites, 20 posts a week, and a client who needs 200 articles in 30 days.

    We run 122+ Cloud Run services across a single GCP project. WordPress REST API calls route through a proxy that handles authentication, IP allowlisting, and retry logic automatically. Imagen 4 generates featured images with IPTC metadata injected before upload. A BigQuery knowledge ledger stores 925 embedded content chunks for persistent AI memory across sessions.

    We’ve now productized this infrastructure so you can skip the 18 months it took us to build it.

    Who This Is For

    Content agencies, SEO publishers, and AI-native operators running multiple WordPress sites who need content velocity that exceeds what a human-in-the-loop browser session can deliver. If you’re publishing fewer than 20 posts a week across fewer than 3 sites, you probably don’t need this yet. If you’re above that threshold and still doing it manually — you’re leaving serious capacity on the table.

    What We Build

    • WP Proxy (Cloud Run) — Single authenticated gateway to all your WordPress sites. Handles Basic auth, app passwords, WAF bypass, and retry logic. One endpoint to rule all sites.
    • Claude AI Publisher — Cloud Run service that accepts article briefs, calls Claude API, optimizes for SEO/AEO/GEO, and publishes directly to WordPress REST API. Fully automated brief-to-publish.
    • Imagen 4 Proxy — GCP Vertex AI image generation endpoint. Accepts prompts, returns WebP images with IPTC/XMP metadata injected, uploads to WordPress media library. Four-tier quality routing: Fast → Standard → Ultra → Flagship.
    • BigQuery Knowledge Ledger — Persistent AI memory layer. Content chunks embedded via Vertex AI text-embedding-005, stored in BigQuery, queryable across sessions. Ends the “start from scratch” problem every time a new Claude session opens.
    • Batch API Router — Routes non-time-sensitive jobs (taxonomy, schema, meta cleanup) to Anthropic Batch API at 50% cost. Routes real-time jobs to standard API. Automatic tier selection.

    What You Get vs. DIY vs. n8n/Zapier

    Tygart Media GCP Build DIY from scratch No-code automation (n8n/Zapier)
    WordPress WAF bypass built in You figure it out
    Imagen 4 image generation
    BigQuery persistent AI memory
    Anthropic Batch API cost routing
    Claude model tier routing
    Proven at 20+ posts/day Unknown

    What We Deliver

    Item Included
    WP Proxy Cloud Run service deployed to your GCP project
    Claude AI Publisher Cloud Run service
    Imagen 4 proxy with IPTC injection
    BigQuery knowledge ledger (schema + initial seed)
    Batch API routing logic
    Model tier routing configuration (Haiku/Sonnet/Opus)
    Site credential registry for all your WordPress sites
    Technical walkthrough + handoff documentation
    30-day async support

    Prerequisites

    You need: a Google Cloud account (we can help set one up), at least one WordPress site with REST API enabled, and an Anthropic API key. Vertex AI access (for Imagen 4) requires a brief GCP onboarding — we walk you through it.

    Ready to Stop Copy-Pasting Into WordPress?

    Tell us how many sites you’re managing, your current publishing volume, and where the friction is. We’ll tell you exactly which services to build first.

    will@tygartmedia.com

    Email only. No sales call required. No commitment to reply.

    Frequently Asked Questions

    Do I need to know how to use Google Cloud?

    No. We build and deploy everything. You’ll need a GCP account and billing enabled — we handle the rest and document every service so you can maintain it independently.

    How is this different from using Claude directly in a browser?

    Browser sessions have no memory, no automation, no direct WordPress integration, and no cost optimization. This infrastructure runs asynchronously, publishes directly to WordPress via REST API, stores content history in BigQuery, and routes jobs to the cheapest model tier that can handle the task.

    Which WordPress hosting providers does the proxy support?

    We’ve tested and configured routing for WP Engine, Flywheel, SiteGround, Cloudflare-protected sites, Apache/ModSecurity servers, and GCP Compute Engine. Most hosting environments work out of the box — a handful need custom WAF bypass headers, which we configure per-site.

    What does the BigQuery knowledge ledger actually do?

    It stores content chunks (articles, SOPs, client notes, research) as vector embeddings. When you start a new AI session, you query the ledger instead of re-pasting context. Your AI assistant starts with history, not a blank slate.

    What’s the ongoing GCP cost?

    Highly variable by volume. For a 10-site agency publishing 50 posts/week with image generation, expect $50–$200/month in GCP costs. Cloud Run scales to zero when idle, so you’re not paying for downtime.

    Can this be expanded after initial setup?

    Yes — the architecture is modular. Each Cloud Run service is independent. We can add newsroom services, variant engines, social publishing pipelines, or site-specific publishers on top of the core stack.

    Last updated: April 2026