Category: Local AI & Automation

Building autonomous AI systems that run locally. Zero cloud cost, full data control, infinite scale.

  • The AI Stack That Replaced Our $12K/Month Tool Budget

    What We Were Paying For (And Why We Stopped)

    At our peak tool sprawl, Tygart Media was spending over twelve thousand dollars per month on SaaS subscriptions. SEO platforms, content generation tools, social media schedulers, analytics dashboards, CRM integrations, and monitoring services. Every tool solved one problem and created two more – data silos, redundant features, and the constant overhead of managing logins, billing, and updates.

    The turning point came when we realized that 80% of what these tools did could be replicated by a combination of local AI models, open-source software, and well-written automation scripts. Not a theoretical possibility – we actually built it and measured the results over 90 days.

    The Local AI Models That Do the Heavy La flooring companyng

    We run Ollama on a standard laptop – no GPU cluster, no cloud compute bills. The models handle content drafting, keyword analysis, meta description generation, and internal link suggestions. For tasks requiring deeper reasoning, we route to Claude via the Anthropic API, which costs pennies per article compared to enterprise content platforms.

    The cost comparison is stark: a single enterprise SEO tool charges $300-500/month per site. We manage 23 sites. Our AI stack – running locally – handles the same keyword tracking, content gap analysis, and optimization recommendations for the cost of electricity.

    The models we rely on most: Llama 3.1 for fast content drafts, Mistral for technical analysis, and Claude for complex reasoning tasks like content strategy and schema generation. Each model has a specific role, and none of them send a monthly invoice.

    The Automation Layer: PowerShell, Python, and Cloud Run

    AI models alone don’t replace tools – you need the orchestration layer that connects them to your actual workflows. We built ours on three technologies:

    PowerShell scripts handle Windows-side automation: file management, API calls to WordPress sites, batch processing of images, and scheduling tasks. Python scripts handle the heavier data work: SEO signal extraction, content analysis, and reporting. Google Cloud Run hosts the few services that need to be always-on, like our WordPress API proxy and our content publishing pipeline.

    Total cloud cost: under $50/month on Google Cloud’s free tier and minimal compute. Compare that to the $12K we were spending on tools that did less.

    What We Still Pay For (And Why)

    We didn’t eliminate every subscription. Some tools earn their keep:

    Metricool ($50/month) handles social media scheduling across multiple brands – the API integration alone saves hours. DataForSEO (pay-per-use) provides raw SERP data that would be impractical to scrape ourselves. Call Tracking Metrics handles call attribution for restoration clients where phone leads are the primary conversion.

    The principle: pay for data you can’t generate and distribution you can’t replicate. Everything else – content creation, SEO analysis, reporting, optimization – runs on our own stack.

    The 90-Day Results

    After 90 days of running the replacement stack across all client sites and our own properties, the numbers told a clear story. Content output increased by 340%. SEO performance held steady or improved across 21 of 23 sites. Total monthly tool spend dropped from $12,200 to under $800.

    The hidden benefit: ownership. When your tools are your own scripts and models, no vendor can raise prices, change APIs, or sunset features. You own the entire stack.

    Frequently Asked Questions

    Do you need technical skills to build a local AI stack?

    You need basic comfort with command-line tools and scripting. If you can install software and edit a configuration file, you can run Ollama. The automation layer requires Python or PowerShell knowledge, but most scripts are straightforward once the architecture is in place.

    Can local AI models really match enterprise SEO tools?

    For content generation, optimization recommendations, and gap analysis – yes. For real-time SERP tracking and backlink monitoring, you still need external data sources like DataForSEO. The key is understanding which tasks need live data and which can run on local intelligence.

    What about reliability compared to SaaS tools?

    SaaS tools go down too. Local tools run when your machine runs. For cloud-hosted components, Google Cloud Run has a 99.95% uptime SLA. Our stack has been more reliable than the vendor tools it replaced.

    How long did the migration take?

    About six weeks of active development to replace the core tools, plus another month of refinement. The investment pays for itself in the first billing cycle.

    Build or Buy? Build.

    The era of needing expensive SaaS tools for every marketing function is ending. Local AI, open-source automation, and minimal cloud infrastructure can replace the majority of your tool budget while giving you more control, better customization, and zero vendor lock-in.

  • What Happens When Claude Runs Your WordPress for 90 Days

    The Experiment: Full AI Site Management

    In January 2026, we gave Claude – Anthropic’s AI assistant – the keys to our WordPress operation. Not just content generation, but the full stack: SEO audits, content gap analysis, taxonomy management, schema injection, internal linking, meta optimization, and publishing. Across 23 sites. For 90 days.

    This wasn’t a theoretical exercise. We built Claude into our operational pipeline through custom skills, WordPress REST API connections, and a GCP proxy layer that routes all site management through Google Cloud. Every optimization, every published article, every schema update was executed by Claude with human oversight on strategy and final approval.

    What Claude Actually Did

    During the 90-day period, Claude executed over 2,400 individual WordPress operations across all sites. The breakdown: 847 SEO meta refreshes, 312 new articles published, 156 schema markup injections, 94 taxonomy reorganizations, and 1,000+ internal link additions.

    Each operation followed a skill-based protocol. Our wp-seo-refresh skill handles on-page SEO. The wp-schema-inject skill adds structured data. The wp-interlink skill builds the internal link graph. Claude doesn’t freestyle – it follows proven playbooks that encode our SEO, AEO, and GEO best practices.

    The Results That Surprised Us

    Organic traffic across all 23 sites increased 47% over the 90-day period. The more interesting metric was consistency. Before Claude, our sites had wildly uneven optimization – some posts had full schema markup and internal links, others had nothing. After 90 days, every post on every site met the same baseline quality standard.

    The sites that improved most were the ones neglected longest. a luxury lending firm saw a 120% increase in organic sessions after Claude refreshed every post’s meta data, added FAQ schema, and built the internal link structure. a restoration company went from 12 ranking keywords to over 340.

    Well-optimized sites saw smaller but meaningful gains – typically 15-25% improvements in click-through rates from better meta descriptions and featured snippet capture.

    What Claude Can’t Do (Yet)

    AI site management has clear limitations. Claude can’t make strategic decisions about which markets to enter. It can’t conduct original customer research. It can’t judge whether content truly resonates with a human audience – it can only optimize for signals that correlate with resonance.

    We also found that AI-generated internal links sometimes prioritize SEO logic over user experience. A link that makes sense for PageRank distribution might confuse a reader. Human review improved link quality significantly.

    The right model is AI as operator, human as strategist. Claude handles the repetitive, systematic work that scales linearly with site count. Humans handle the judgment calls.

    Frequently Asked Questions

    Is it safe to give an AI access to your WordPress sites?

    We use WordPress Application Passwords with editor-level permissions – Claude can create and edit content but can’t modify site settings or access user data. All operations route through our GCP proxy with full audit logs.

    How do you prevent AI from making SEO mistakes?

    Every operation follows a validated protocol. Claude doesn’t improvise – it executes predefined skills with guardrails. Critical operations go through a review queue. We run weekly audits comparing pre- and post-optimization metrics.

    Can any business replicate this setup?

    The individual skills work on any WordPress site with REST API access. The scale advantage comes from the orchestration layer. A single-site business can start with basic Claude plus WordPress automation and expand from there.

    What’s the cost of running Claude as a site manager?

    API costs run approximately $50-100/month for our 23-site operation. The GCP proxy adds under $10/month. Compare that to a junior SEO specialist at $4,000-5,000/month handling maybe 3-5 sites.

    The Verdict After 90 Days

    We’re not going back. AI-managed WordPress isn’t a gimmick – it’s a fundamental shift in how digital operations scale. The 90-day experiment became our permanent operating model.

  • The Entrepreneur’s Case for Vertical AI Over Generic Tools

    Why ChatGPT Isn’t Enough for Your Business

    Every small business owner has tried ChatGPT by now. Most found it useful for drafting emails and brainstorming – and then stopped. The gap between a generic AI chatbot and a business-changing AI tool is enormous, and it comes down to one thing: vertical specificity.

    A generic AI tool knows a little about everything. A vertical AI tool knows everything about your specific business operation. The difference in output quality is the difference between ‘here are some marketing tips’ and ‘here are the 15 articles your WordPress site needs next month, optimized for your specific keyword gaps, written in your brand voice, and ready to publish.’

    What Vertical AI Looks Like in Practice

    At Tygart Media, we don’t use AI generally – we use AI vertically. Every AI tool in our stack is configured for a specific business function with specific data, specific rules, and specific output formats.

    WordPress Site Management AI: Configured with site credentials, content inventories, SEO protocols, and publishing workflows. It doesn’t suggest things – it executes them. ‘Run a full SEO refresh on post 247 on a luxury lending firm’ produces immediate, measurable results.

    Content Intelligence AI: Trained on our gap analysis framework, persona detection model, and article generation protocol. Input: a WordPress site URL. Output: a prioritized content opportunity report with 15 ready-to-generate article briefs.

    Client Operations AI: Connected to our Notion Command Center with access to task databases, client portals, and content calendars. It can triage incoming requests, generate status reports, and draft client communications – all within the context of our specific operational data.

    None of these use cases work with a generic AI tool. They require configuration, integration, and domain-specific protocols that transform general intelligence into business-specific capability.

    Why Generic Tools Fail Small Businesses

    No business context: Generic AI doesn’t know your customers, your competitors, or your market position. Every interaction starts from zero. Vertical AI retains context about your business and builds on previous interactions.

    No workflow integration: Generic AI lives in a chat window. Vertical AI connects to your WordPress sites, your Notion workspace, your social media scheduler, and your analytics platform. It doesn’t just advise – it acts.

    No quality enforcement: Generic AI produces whatever you ask for, with no guardrails. Vertical AI follows protocols – every article meets your SEO standards, every meta description fits the character limit, every schema markup validates correctly. Quality is systematic, not dependent on prompt quality.

    No compound learning: Generic AI interactions are ephemeral. Vertical AI builds on a knowledge base that grows with every operation – your site inventories, performance data, content history, and strategic decisions all become part of the system’s context.

    Building Your Own Vertical AI Stack

    You don’t need to build everything from scratch. The path to vertical AI follows a predictable sequence:

    Step 1: Identify your highest-volume repetitive task. For most businesses, it’s content creation, reporting, or customer communication. Pick one.

    Step 2: Document the protocol. Write down exactly how a human performs this task – every step, every decision point, every quality check. This documentation becomes your AI’s operating manual.

    Step 3: Connect the AI to your data. API integrations, database connections, file access – give the AI the same information a human employee would need to do the job.

    Step 4: Build the execution layer. Scripts, automations, and API calls that let the AI take action – not just generate text, but actually publish content, update databases, send communications.

    Step 5: Add human checkpoints. Identify the 2-3 moments in the workflow where human judgment adds value. Everything else runs automatically.

    Frequently Asked Questions

    How much does it cost to build a vertical AI stack?

    Development time is the primary investment – typically 4-8 weeks for a first vertical AI tool, depending on complexity. Ongoing API costs range from $50-200/month depending on usage. Compare that to hiring a specialist for the same function at $4,000-8,000/month.

    Do I need a technical background to implement vertical AI?

    Basic technical comfort helps – ability to work with APIs, configure tools, and write simple scripts. Many businesses partner with an AI-savvy agency (like Tygart Media) for initial setup and then operate the system independently.

    What’s the ROI timeline for vertical AI?

    Most businesses see positive ROI within 60-90 days. The cost savings from automated execution and the revenue gains from improved output quality compound quickly. Our clients typically report 3-5x ROI within six months.

    Is vertical AI only for marketing operations?

    No. The same principles apply to sales operations, customer service, financial reporting, inventory management, and any business function with repetitive, protocol-driven tasks. Marketing is where we apply it, but the framework is universal.

    Stop Using AI Like a Search Engine

    The biggest mistake small businesses make with AI is treating it like a better Google – a place to ask questions and get answers. The real power of AI is in vertical application: connecting it to your specific data, your specific workflows, and your specific quality standards. That’s where AI stops being a novelty and starts being a competitive advantage.

  • 387 Cowork Sessions and Counting: What Happens When AI Becomes Your Daily Operating Partner

    This Is Not a Chatbot Story

    When people hear I use AI every day, they picture someone typing questions into ChatGPT and getting answers. That’s not what this is. I’ve run 387 working sessions with Claude in Cowork mode since December 2025. Each session is a full operating environment – a Linux VM with file access, tool execution, API connections, and persistent memory across sessions.

    These aren’t conversations. They’re deployments. Content publishes. Infrastructure builds. SEO audits across 18 WordPress sites. Notion database updates. Email monitors. Scheduled tasks. Real operational work that used to require a team of specialists.

    The number 387 isn’t bragging. It’s data. And what that data reveals about how AI actually integrates into daily business operations is more interesting than any demo or product launch.

    What a Typical Session Actually Looks Like

    A session starts when I open Cowork mode and describe what I need done. Not a vague prompt – a specific operational task. “Run the content intelligence audit on a storm protection company.com and generate 15 draft articles.” “Check all 18 WordPress sites for posts missing featured images and generate them using Vertex AI.” “Read my Gmail for VIP messages from the last 6 hours and summarize what needs attention.”

    Claude loads into a sandboxed Linux environment with access to my workspace folder, my installed skills (I have 60+), my MCP server connections (Notion, Gmail, Google Calendar, Metricool, Figma, and more), and a full bash/Python execution layer. It reads my CLAUDE.md file – a persistent memory document that carries context across sessions – and gets to work.

    A single session might involve 50-200 tool calls. Reading files, executing scripts, making API calls, writing content, publishing to WordPress, logging results to Notion. The average session runs 15-45 minutes of active work. Some complex ones – like a full site optimization pass – run over two hours.

    The Skill Layer Changed Everything

    Early sessions were inefficient. I’d explain the same process every time – how to connect to WordPress via the proxy, what format to use for articles, which Notion database to log results in. Repetitive context-setting that ate 30% of every session.

    Then I started building skills. A skill is a structured instruction file (SKILL.md) that Claude reads at the start of a session when the task matches its trigger conditions. I now have skills for WordPress publishing, SEO optimization, content generation, Notion logging, YouTube watch page creation, social media scheduling, site auditing, and dozens more.

    The impact was immediate. A task that took 20 minutes of back-and-forth setup now triggers in one sentence. “Run the wp-intelligence-audit on a luxury asset lender.com” – Claude reads the skill, loads the credentials from the site registry, connects via the proxy, pulls all posts, analyzes gaps, and generates a full report. No explanation needed. The skill contains everything.

    Building skills is the highest-leverage activity I’ve found in AI-assisted work. Every hour spent writing a skill saves 10+ hours across future sessions. At 387 sessions, the compound return is staggering.

    What 387 Sessions Taught Me About AI Workflow

    Specificity beats intelligence. The most productive sessions aren’t the ones where Claude is “smartest.” They’re the ones where I give the most specific instructions. “Optimize this post for SEO” produces mediocre results. “Run wp-seo-refresh on post 247 at a luxury asset lender.com, ensure the focus keyword is ‘luxury asset lending,’ update the meta description to 140-160 characters, and add internal links to posts 312 and 418” produces excellent results. AI amplifies clarity.

    Persistent memory is the unlock. CLAUDE.md – a markdown file that persists across sessions – is the most important file in my entire system. It contains my preferences, operational rules, business context, and standing instructions. Without it, every session starts from zero. With it, session 387 has the accumulated context of all 386 before it. This is the difference between using AI as a tool and using AI as a partner.

    Batch operations reveal true ROI. Publishing one article? AI saves maybe 30 minutes. Publishing 15 articles across 3 sites with full SEO/AEO/GEO optimization, taxonomy assignment, internal linking, and Notion logging? AI saves 15+ hours. The value curve is exponential with batch size. I now default to batch operations for everything – content, audits, meta updates, image generation.

    Failures are cheap and informative. At least 40 of my 387 sessions hit significant errors – API timeouts, disk space issues, credential failures, rate limiting. Each failure taught me something that made the system more resilient. The SSH workaround. The WP proxy to avoid IP blocking. The WinError 206 fix for long PowerShell commands. Failure at high volume is the fastest path to robust systems.

    The Numbers Behind 387 Sessions

    I tracked the data because the data tells the real story:

    Content produced: Approximately 400+ articles published across 18 WordPress sites. Each article is 1,200-1,800 words, SEO-optimized, AEO-formatted with FAQ sections, and GEO-ready with entity optimization. At market rates for this quality of content, that’s roughly ,000-,000 worth of content production.

    Sites managed: 18 WordPress properties across multiple industries – restoration, luxury lending, cold storage, interior design, comedy, training, technology. Each site gets regular content, SEO audits, taxonomy fixes, schema injection, and internal linking.

    Automations built: 7 autonomous AI agents (the droid fleet), 60+ skills, 3 scheduled tasks, a GCP Compute Engine cluster running 5 WordPress sites, a Cloud Run proxy for WordPress API routing, and a Vertex AI chatbot deployment.

    Time investment: Approximately 200 hours of active session time over three months. For context, a single full-time employee working those same 200 hours could not have produced a fraction of this output, because the bottleneck isn’t thinking time – it’s execution speed. Claude executes API calls, writes code, publishes content, and processes data at machine speed. I provide direction at human speed. The combination is multiplicative.

    Why Most People Won’t Do This

    The honest answer: it requires upfront investment that most people aren’t willing to make. Building the skill library took weeks. Configuring the MCP connections, setting up the proxy, provisioning the GCP infrastructure, writing the CLAUDE.md context file – that’s real work before you see any return.

    Most people want AI to be plug-and-play. Type a question, get an answer. And for simple tasks, it is. But for operational AI – AI that runs your business processes daily – the setup cost is significant and the learning curve is real.

    The payoff, though, is not incremental. It’s categorical. I’m not 10% more productive than I was before Cowork mode. I’m operating at a fundamentally different scale. Tasks that would require hiring 3-4 specialists – content writer, SEO analyst, site admin, automation engineer – are handled in daily sessions by one person with a well-configured AI partner.

    That’s not a productivity hack. That’s a structural advantage.

    Frequently Asked Questions

    What is Cowork mode and how is it different from regular Claude?

    Cowork mode is a feature of Claude’s desktop app that gives Claude access to a sandboxed Linux VM, file system, bash execution, and MCP server connections. Regular Claude is a chat interface. Cowork mode is an operating environment where Claude can read files, run code, make API calls, and produce deliverables – not just text responses.

    How much does running 387 sessions cost?

    Cowork mode is included in the Claude Pro subscription at /month. The MCP connections (Notion, Gmail, etc.) use free API tiers. The GCP infrastructure runs about /month. Total cost for three months of operations: approximately . The value produced is orders of magnitude higher.

    Can someone replicate this without technical skills?

    Partially. The basic Cowork mode works out of the box for content creation, research, and file management. The advanced setup – custom skills, GCP infrastructure, API integrations – requires comfort with command-line tools, APIs, and basic scripting. The barrier is falling fast as skills become shareable and MCP servers become plug-and-play.

    What’s the most impactful single skill you’ve built?

    The wp-site-registry skill – a single file containing credentials and connection methods for all 18 WordPress sites. Before this skill existed, every session required manually providing credentials. After it, any wp- skill can connect to any site automatically. It turned 18 separate workflows into one unified system.

    What Comes Next

    Session 387 is not a milestone. It’s a Tuesday. The system compounds. Every skill I build makes future sessions faster. Every failure I fix makes the system more resilient. Every batch I run produces data that informs the next batch.

    The question I get most often is “where do you start?” The answer is boring: start with one task you do repeatedly. Build one skill for it. Run it 10 times. Then build another. By session 50, you’ll have a system. By session 200, you’ll have an operating partner. By session 387, you’ll wonder how you ever worked without one.

  • The SEO Drift Detector: How I Built an Agent That Watches 18 Sites for Ranking Decay

    Rankings Don’t Crash – They Drift

    Nobody wakes up to a sudden SEO catastrophe. What actually happens is slower and more insidious. A page that ranked #4 for its target keyword three months ago is now #9. Another page that owned a featured snippet quietly lost it. A cluster of posts that drove 40% of a site’s organic traffic has collectively slipped 3-5 positions across 12 keywords.

    By the time you notice, the damage is done. Traffic is down 25%. Leads have thinned. And the fix – refreshing content, rebuilding authority, reclaiming positions – takes weeks. The problem with SEO drift isn’t that it’s hard to fix. It’s that it’s hard to see.

    I manage 18 WordPress sites across industries ranging from luxury lending to restoration services to cold storage logistics. Manually checking keyword rankings across all of them? Impossible. Waiting for Google Search Console to show a decline? Too late. So I built SD-06 – the SEO Drift Detector – an autonomous agent that monitors keyword positions daily, calculates drift velocity, and flags pages that need attention before the traffic impact hits.

    How SD-06 Works Under the Hood

    The architecture connects three systems: DataForSEO for ranking data, a local SQLite database for historical tracking, and Slack for alerts.

    Every morning at 6 AM, SD-06 runs a scheduled Python script that pulls current ranking positions for tracked keywords across all 18 sites. DataForSEO’s SERP API returns the current Google position for each keyword-URL pair. The script stores these daily snapshots in a SQLite database – one row per keyword per day, with fields for position, URL, SERP features present (featured snippet, People Also Ask, local pack), and the date.

    With 30+ days of historical data, the agent calculates three metrics for each tracked keyword:

    Position delta (7-day): The difference between today’s position and the position 7 days ago. A keyword that moved from #5 to #8 has a delta of -3. Simple, fast, catches sudden drops.

    Drift velocity (30-day): The average daily position change over the last 30 days. This is the metric that catches slow decay. A keyword losing 0.1 positions per day doesn’t trigger any single-day alarm, but over 30 days that’s a 3-position drop. SD-06 calculates this as a rolling regression slope and flags anything with negative drift velocity exceeding -0.05 positions per day.

    Feature loss: Did this URL have a featured snippet, PAA box, or other SERP feature last week that it no longer holds? Feature loss often precedes position loss – it’s an early warning signal that content freshness or authority is slipping.

    The Alert System That Changed My Workflow

    SD-06 sends three types of Slack alerts:

    Red alert (immediate attention): Any keyword that dropped 5+ positions in 7 days, or any URL that lost a featured snippet it held for 14+ consecutive days. These are rare but critical – usually indicating a technical issue, a Google algorithm update, or a competitor publishing a significantly better page.

    Yellow alert (weekly review): Keywords with negative drift velocity exceeding the threshold but no single dramatic drop. These are bundled into a weekly digest every Monday morning. The digest includes the keyword, current position, 30-day trend direction, the affected URL, and a recommended action (refresh content, add internal links, update statistics, or expand the article).

    Green report (monthly summary): A full portfolio health report showing total tracked keywords, percentage dra flooring companyng negative vs. positive, top gainers, top losers, and overall portfolio trajectory. This is the report I share with clients to show proactive SEO management.

    The critical insight was making the recommended action part of every alert. An alert that says “keyword X dropped 3 positions” is information. An alert that says “keyword X dropped 3 positions – recommend refreshing the statistics section and adding 2 internal links from recent posts” is a task I can execute immediately. SD-06 generates these recommendations using simple rules based on what type of drift it detects.

    What 90 Days of Drift Data Revealed

    After running SD-06 for three months across all 18 sites, the data patterns were illuminating.

    Content age is the #1 drift predictor. Posts older than 18 months drift negative at 3x the rate of posts under 12 months old. This isn’t surprising – Google rewards freshness – but the magnitude was larger than expected. It means my content refresh cadence needs to target any post approaching the 18-month mark, not waiting for visible ranking loss.

    Internal linking density correlates with drift resistance. Pages with 5+ inbound internal links from other site content drifted negative 60% less frequently than pages with 0-2 internal links. Orphan pages – content with zero inbound internal links – were the fastest to lose rankings. This validated my investment in the wp-interlink skill that systematically adds internal links across every site.

    Featured snippet loss is a 2-week leading indicator. When a page loses a featured snippet, it loses 2-5 organic positions within the following 14 days approximately 70% of the time. This made featured snippet monitoring the most valuable early warning signal in the entire system. When SD-06 detects snippet loss, I now have a 2-week window to refresh the content before the position drop fully materializes.

    Competitor content publishing causes measurable drift. Several drift events correlated with competitors publishing fresh content targeting the same keywords. Without SD-06, I would have discovered this weeks later through traffic decline. With it, I can see the drift starting within 3-5 days of the competitor publish and respond immediately.

    The Technical Stack

    DataForSEO API for SERP position tracking. The SERP API costs approximately .002 per keyword check. Tracking 200 keywords daily across 18 sites runs about /month – trivial compared to the SEO tools that charge +/month for similar monitoring.

    SQLite for historical data storage. Lightweight, zero-configuration, file-based database that lives on the local machine. After 90 days of daily tracking across 200 keywords, the database file is under 50MB. No server, no cloud database, no monthly cost.

    Python 3.11 with pandas for data analysis, scipy for regression calculations, and the requests library for API calls. The entire script is under 400 lines.

    Slack Incoming Webhook for alerts, same pattern as the VIP Email Monitor. One webhook URL, formatted JSON payloads, zero infrastructure.

    Windows Task Scheduler triggers the script at 6 AM daily. Could also run as a cron job on Linux or a Cloud Run scheduled task on GCP.

    Why I Didn’t Just Use Ahrefs or SEMrush

    I’ve used both. They’re excellent tools. But they have three limitations for my use case.

    First, cost at scale. Monitoring 18 sites with 200+ keywords each on Ahrefs would cost +/month. SD-06 costs /month in API calls.

    Second, custom alert logic. Ahrefs and SEMrush send generic position change alerts. They don’t calculate drift velocity, predict future position loss based on trajectory, or generate content-specific refresh recommendations. SD-06’s alert intelligence is tailored to how I actually work.

    Third, integration with my existing workflow. SD-06 pushes alerts to the same Slack channel where all my other agents report. It writes recommendations that align with my wp-seo-refresh and wp-content-expand skills. The data flows directly into my operational system rather than living in a separate dashboard I have to remember to check.

    Frequently Asked Questions

    How many keywords should you track per site?

    Start with 10-15 per site – your highest-traffic pages and their primary keywords. Expand to 20-30 after the first month once you understand which keywords actually drive business results. Tracking 100+ keywords per site creates noise without proportional signal. Focus on the keywords that drive revenue, not vanity metrics.

    Can drift detection work without DataForSEO?

    Yes, but with less precision. Google Search Console provides position data with a 2-3 day delay and averages positions over date ranges rather than giving exact daily snapshots. You can build a simpler version using the Search Console API, but the drift velocity calculations will be less granular. DataForSEO provides same-day position data at the individual keyword level.

    How quickly can you reverse SEO drift once detected?

    For content-based drift (stale statistics, outdated information, thin sections), a content refresh typically recovers positions within 2-4 weeks after Google recrawls. For authority-based drift (competitors building more backlinks), recovery takes longer – 4-8 weeks – and requires both content improvement and internal linking reinforcement.

    Does this work for local SEO keywords?

    Absolutely. DataForSEO supports location-specific SERP checks, so you can track “water damage restoration Houston” at the Houston geo-target level. Several of my sites are local service businesses, and the drift patterns for local keywords follow the same trajectory math – they just tend to be more volatile due to local pack algorithm updates.

    The Principle Behind the Agent

    SD-06 exists because of a simple belief: the best time to fix SEO is before it breaks. Reactive SEO – waiting for traffic to drop, then scrambling to diagnose and fix – is expensive, stressful, and often too late. Proactive SEO – monitoring drift in real time and refreshing content before positions collapse – costs almost nothing and preserves the compounding value of content that’s already ranking.

    Every piece of content on a website is a depreciating asset. It starts strong, holds for a while, then slowly loses value as competitors publish newer content and search algorithms reward freshness. SD-06 doesn’t stop depreciation. It tells me exactly which assets need maintenance, exactly when they need it, and exactly what the maintenance should look like. That’s not magic. That’s operations.

  • I Indexed 468 Files Into a Local Vector Database. Now My Laptop Answers Questions About My Business.

    The Problem With Having Too Many Files

    I have 468 files that define how my businesses operate. Skill files that tell AI how to connect to WordPress sites. Session transcripts from hundreds of Cowork conversations. Notion exports. API documentation. Configuration files. Project briefs. Meeting notes. Operational playbooks.

    These files contain everything – credentials, workflows, decisions, architecture diagrams, troubleshooting histories. The knowledge is comprehensive. The problem is retrieval. When I need to remember how I configured the WP proxy, or what the resolution was for that SiteGround blocking issue three months ago, or which Notion database stores client portal data – I’m grep-searching through hundreds of files, hoping I remember the right keyword.

    Grep works when you know exactly what you’re looking for. It fails completely when you need to ask a question like “what was the workaround we used when SSH broke on the knowledge cluster VM?” That’s a semantic query. It requires understanding, not string matching.

    So I built a local vector search system. Every file gets chunked, embedded into vectors using a local model, stored in a local database, and queried with natural language. My laptop now answers questions about my own business operations – instantly, accurately, and without sending any data to the cloud.

    The Architecture: Ollama + ChromaDB + Python

    The stack is deliberately minimal. Three components, all running locally, zero cloud dependencies.

    Ollama with nomic-embed-text handles the embedding. This is a 137M parameter model specifically designed for text embeddings – turning chunks of text into 768-dimensional vectors that capture semantic meaning. It runs locally on my laptop, processes about 50 chunks per second, and produces embeddings that rival OpenAI’s ada-002 for retrieval tasks. The entire model is 274MB on disk.

    ChromaDB is the vector database. It’s an open-source, embedded vector store that runs as a Python library – no server process, no Docker container, no infrastructure. Data is persisted to a local directory. The entire 468-file index, with all embeddings and metadata, takes up 180MB on disk. Queries return results in under 100 milliseconds.

    A Python script ties it together. The indexer walks through designated directories, reads each file, splits it into chunks of ~500 tokens with 50-token overlap, generates embeddings via Ollama, and stores them in ChromaDB with metadata (file path, chunk number, file type, last modified date). The query interface takes a natural language question, embeds it, searches for the 5 most similar chunks, and returns the relevant passages with source attribution.

    What Gets Indexed

    I index four categories of files:

    Skills (60+ files): Every SKILL.md file in my skills directory. These contain operational instructions for WordPress publishing, SEO optimization, content generation, site auditing, Notion logging, and more. When I ask “how do I connect to the a luxury asset lender WordPress site?” the system retrieves the exact credentials and connection method from the wp-site-registry skill.

    Session transcripts (200+ files): Exported transcripts from Cowork sessions. These contain the full history of decisions, troubleshooting, and solutions. When I ask “what was the fix for the WinError 206 issue?” it retrieves the exact conversation where we diagnosed and solved that problem – publish one article per PowerShell call, never combine multiple article bodies in a single command.

    Project documentation (100+ files): Architecture documents, API documentation, configuration files, and project briefs. Technical reference material that I wrote once and need to recall later.

    Notion exports (50+ files): Periodic exports of key Notion databases – the task board, client records, content calendars, and operational notes. This bridges the gap between Notion (where I plan) and local files (where I execute).

    How the Chunking Strategy Matters

    The most underrated part of building a RAG system is chunking – how you split documents into pieces before embedding them. Get this wrong and your retrieval is useless regardless of how good your embedding model is.

    I tested three approaches:

    Fixed-size chunks (500 tokens): Simple but crude. Splits mid-sentence, mid-paragraph, sometimes mid-code-block. Retrieval accuracy was around 65% on my test queries – too many chunks lacked enough context to be useful.

    Paragraph-based chunks: Split on double newlines. Better for prose documents but terrible for skill files and code, where a single paragraph might be 2,000 tokens (too large) or 10 tokens (too small). Retrieval accuracy improved to about 72%.

    Semantic chunking with overlap: Split at ~500 tokens but respect sentence boundaries, and include 50 tokens of overlap between consecutive chunks. This means the end of chunk N appears at the beginning of chunk N+1, providing continuity. Additionally, each chunk gets prepended with the document title and the nearest H2 heading for context. Retrieval accuracy jumped to 89%.

    The overlap and heading prepend were the critical improvements. Without overlap, answers that span two chunks get lost. Without heading context, a chunk about “connection method” could be about any of 18 sites – the heading tells the model which site it’s about.

    Real Queries I Run Daily

    This isn’t a science project. I use this system every day. Here are actual queries from the past week:

    “What are the credentials for the an events platform WordPress site?” – Returns the exact username (will@engagesimply.com), app password, and the note that an events platform uses an email as username, not “Will.” Found in the wp-site-registry skill file.

    “How does the 247RS GCP publisher work?” – Returns the service URL, auth header format, and the explanation that SiteGround blocks all direct and proxy calls, requiring the dedicated Cloud Run publisher. Pulled from both the 247rs-site-operations skill and a session transcript where we built it.

    “What was the disk space issue on the knowledge cluster VM?” – Returns the session transcript passage about SSH dying because the 20GB boot disk filled to 98%, the startup script workaround, and the IAP tunneling backup method we configured afterward.

    “Which sites use Flywheel hosting?” – Returns a list: a flooring company (a flooring company.com), a live comedy platform (a comedy streaming site), an events platform (an events platform.com). Cross-referenced across multiple skill files and assembled by the retrieval system.

    Each query takes under 2 seconds – embedding the question (~50ms), vector search (~80ms), and displaying results with source file paths. No API call. No internet required. No data leaves my machine.

    Why Local Beats Cloud for This Use Case

    Security is absolute. These files contain API credentials, client information, business strategies, and operational playbooks. Uploading them to a cloud embedding service – even a reputable one – introduces a data handling surface I don’t need. Local means the data never leaves the machine. Period.

    Speed is consistent. Cloud API calls for embeddings add 200-500ms of latency per query, plus they’re subject to rate limits and service availability. Local embedding via Ollama is 50ms every time. When I’m mid-session and need an answer fast, consistent sub-second response matters.

    Cost is zero. OpenAI charges .0001 per 1K tokens for ada-002 embeddings. That sounds cheap until you’re re-indexing 468 files (roughly 2M tokens) every week – .20 per re-index, /year. Trivial in isolation, but when every tool in my stack has a small recurring cost, they compound. Local eliminates the line item entirely.

    Availability is guaranteed. The system works on an airplane, in a coffee shop with no WiFi, during a cloud provider outage. My operational knowledge base is always accessible because it runs on the same machine I’m working on.

    Frequently Asked Questions

    Can this replace a full knowledge management system like Confluence or Notion?

    No – it complements them. Notion is where I create and organize information. The local vector system is where I retrieve it instantly. They serve different functions. Notion is the authoring environment; the vector database is the search layer. I export from Notion periodically and re-index to keep the retrieval system current.

    How often do you re-index the files?

    Weekly for a full re-index, which takes about 4 minutes for all 468 files. I also run incremental indexing – only re-embedding files modified since the last index – as part of my daily morning script. Incremental indexing typically processes 5-15 files and takes under 30 seconds.

    What hardware do you need to run this?

    Surprisingly modest. My Windows laptop has 16GB RAM and an Intel i7. The nomic-embed-text model uses about 600MB of RAM while running. ChromaDB adds another 200MB for the index. Total memory overhead: under 1GB. Any modern laptop from the last 3-4 years can handle this comfortably. No GPU required for embeddings – CPU performance is more than adequate.

    How does this compare to just using Ctrl+F or grep?

    Grep finds exact text matches. Vector search finds semantic matches. If I search for “SiteGround blocking” with grep, I find files that contain those exact words. If I search for “why can’t I connect to the a restoration company site” with vector search, I find the explanation about SiteGround’s WAF blocking API calls – even though the passage might not contain the words “connect” or “a restoration company site” explicitly. The difference is understanding context vs. matching strings.

    The Compound Effect

    Every file I create makes the system smarter. Every session transcript adds to the searchable history. Every skill I write becomes instantly retrievable. The vector database is a living index of accumulated operational knowledge – and it grows automatically as I work.

    Three months ago, the answer to “how did we solve X?” was “let me search through my files for 10 minutes.” Today, the answer takes 2 seconds. Multiply that time savings across 20-30 lookups per week, and the ROI is measured in hours reclaimed – hours that go back into building, not searching.

  • MP-04: The Agent That Turns Every Meeting Into Action Items Before I Close the Tab

    Meetings Produce Information. Most of It Evaporates.

    I sat in a client call last month where we agreed on three specific deliverables, a revised timeline, and a budget adjustment. Everyone nodded. Everyone agreed. Three days later, nobody could remember the exact numbers or who owned what. I had to dig through a transcript to reconstruct the meeting.

    This happens constantly. Meetings generate decisions, action items, and commitments at a rate that exceeds human note-taking capacity. Even when someone takes notes, the notes are incomplete, biased toward what the note-taker found interesting, and almost never get distributed in an actionable format. The transcript exists – most meetings are recorded now – but a 45-minute transcript is a 6,000-word wall of text that nobody will read.

    MP-04 solves this. It’s the fourth agent in my autonomous fleet, and its job is simple: take any meeting transcript, extract everything actionable, and route it to the right systems before the meeting fades from memory.

    What MP-04 Extracts

    The agent processes meeting transcripts through Ollama’s Llama 3.2 model with a structured extraction prompt. It pulls five categories of information:

    Action items: Anything that someone committed to doing. “I’ll send the proposal by Friday” becomes an action item assigned to the speaker with a Friday deadline. “We need to update the website copy” becomes an action item with no assignee – flagged for me to assign. The model distinguishes between firm commitments (someone said “I will”) and vague suggestions (“we should probably”) and tags them accordingly.

    Decisions: Any point where the group reached agreement. “Let’s go with Option B” is a decision. “The budget is ,000” is a decision. These get logged as immutable records – what was decided, when, and by whom. Decisions are critical for accountability. When someone later says “we never agreed to that,” the decision log settles it.

    Client mentions: Names of clients, companies, or projects discussed. Each mention gets cross-referenced against my client database to attach the meeting context to the right client record. If a client was discussed in three meetings this month, their record shows all three with relevant excerpts.

    Deadlines and dates: Any temporal commitment. “The launch is March 15th.” “We need this by end of quarter.” “Let’s review next Tuesday.” These get extracted with enough context to create calendar-ready events or task due dates.

    Open questions: Things raised but not resolved. “What’s the pricing for the enterprise tier?” with no answer in the transcript becomes an open question flagged for follow-up. These are the items that silently disappear after meetings if nobody tracks them.

    The Routing Layer

    Extraction is useful. Routing is what makes it operational.

    After extracting the five categories, MP-04 routes each item to the appropriate system:

    Action items become Notion tasks in my Tasks Database. Each task is pre-populated with the company (inferred from client mentions), priority (inferred from deadline proximity and language urgency), source (the meeting date and title), and a link back to the full transcript. I don’t create these tasks manually. They appear in my task board, ready to be triaged in my next planning session.

    Decisions get logged to the Knowledge Database in Notion. This creates a searchable decision history. Three months from now, when I need to recall what was agreed about the Q2 content strategy, I search the decisions log instead of scrubbing through transcripts.

    Client mentions update the Client Database with a meeting note. The note includes a 2-3 sentence summary of what was discussed about that client, automatically generated from the relevant transcript sections.

    Deadlines get posted to Slack with a reminder. If the deadline is within 7 days, it goes to my priority channel. If it’s further out, it goes to the weekly planning channel.

    Open questions become follow-up tasks in Notion, tagged with a “needs-answer” status that keeps them visible until resolved.

    The Technical Reality

    MP-04 runs locally on my Windows machine. The input is a text transcript – either pasted directly or loaded from a file. Most meeting platforms (Zoom, Google Meet, Teams) now generate transcripts automatically, so the input is free.

    The Ollama call uses a detailed system prompt that defines the extraction schema with examples. The prompt is about 800 tokens of instructions that tell the model exactly how to format each extracted item – as JSON objects with specific fields for each category. This structured output means the routing script can parse the results programmatically without any ambiguity.

    Processing time for a 45-minute meeting transcript (approximately 6,000 words): about 15 seconds on Llama 3.2 3B running locally. The Notion API calls to create tasks, update client records, and log decisions add another 5-10 seconds. Total time from transcript to fully routed outputs: under 30 seconds.

    Compare that to the manual process: read the transcript (15 minutes), identify action items (10 minutes), create tasks in Notion (5 minutes), update client records (5 minutes), set reminders for deadlines (5 minutes). That’s 40 minutes of administrative work per meeting, reduced to 30 seconds.

    The Client Name Guardrail Problem

    One unexpected challenge: client names in transcripts are messy. People use first names, company names, project codenames, and abbreviations interchangeably. “The Beverly project” and “a luxury lending firm” and “Sarah’s account” might all refer to the same client.

    I built a name resolution layer that maps common references to canonical client records. It’s a JSON lookup table: “Beverly” maps to “a luxury lending firm Company,” “Sarah” maps to “Sarah [Client Last Name] at a luxury lending firm,” “BL” maps to “a luxury lending firm.” The table has about 150 entries covering all active clients and common reference patterns.

    When the extraction model identifies a client mention, the name resolver checks it against this table before routing. If there’s no match, it flags the mention as “unresolved client reference” for manual review rather than creating a misattributed record. The guardrail prevents the worst outcome – action items attached to the wrong client – at the cost of occasionally requiring a 10-second manual resolution.

    What Changed After 60 Days of Running MP-04

    The obvious win: I stopped losing action items. In the 60 days before MP-04, I estimate that about 20% of meeting commitments fell through the cracks – not from negligence, but from the gap between hearing a commitment and recording it in a system. In the 60 days after, that dropped to under 3% (the remaining 3% are items the model misclassifies or that I manually deprioritize).

    The less obvious win: meeting quality improved. When you know every commitment will be automatically extracted and tracked, you’re more careful about what you commit to. Meetings became more precise. Fewer vague “we should probably” statements, more specific “I will deliver X by Y.” The agent didn’t just capture accountability – it created it.

    The unexpected win: the decision log became a strategic asset. Having a searchable history of every decision across every client turned out to be invaluable for quarterly reviews, contract renewals, and scope discussions. “Based on the decisions log, we’ve expanded scope three times without adjusting the retainer” is a powerful conversation to have with data behind it.

    Frequently Asked Questions

    What meeting platforms does MP-04 work with?

    Any platform that produces a text transcript. Zoom, Google Meet, Microsoft Teams, Otter.ai, and Fireflies all export transcripts. MP-04 doesn’t integrate with these platforms directly – it processes the transcript file. This keeps it platform-agnostic and avoids the complexity of OAuth integrations with every meeting tool.

    How accurate is the action item extraction?

    On my test set of 40 meeting transcripts, the model correctly identified 91% of action items I had manually tagged. The 9% it missed were typically very implicit commitments – things like “I’ll take care of that” without specifying what “that” refers to. It also occasionally generates false positives from hypothetical statements – “if we were to do X, we would need Y” getting tagged as a commitment. The false positive rate is about 5%, easily caught in the triage step.

    Can this work for meetings I didn’t attend?

    Yes – and that’s one of the most useful applications. Team members can drop a transcript into the processing queue and I get a structured summary with action items without having attended the meeting. This is especially valuable for the meetings I delegate but still need to track outcomes from.

    What about sensitive meeting content?

    Everything runs locally. The transcript is processed by Ollama on my machine, routed to my private Notion workspace, and posted to my private Slack channels. No third-party service sees the meeting content. This is critical for client meetings that discuss financials, legal issues, or strategic plans.

    The Agent Philosophy

    MP-04 embodies the principle that runs through my entire agent fleet: don’t automate decisions – automate the administrative overhead around decisions. The agent doesn’t decide what to prioritize or how to respond to a client request. It extracts the raw information, structures it, and routes it to where I can make those decisions quickly and with full context. The human judgment stays human. The administrative busywork disappears.

  • Plugins, Skills, and MCPs: The Three Layers That Make AI Actually Useful

    Prompts Are Not a Strategy

    The entire AI productivity discourse is stuck on prompts. Write better prompts. Use this template. Here is my secret prompt. It is the equivalent of teaching someone to type faster when what they need is a computer.

    Prompts are inputs. A command is worthless without an operating system to execute it, tools to interact with, and persistent memory to build on. The gap between AI as a chatbot and AI as a business tool is not better prompts. It is infrastructure.

    After 387+ Cowork sessions of AI-powered operations, I have identified three infrastructure layers that transform AI from fancy autocomplete into a genuine operational partner.

    Layer 1: MCP Servers – The Connections

    MCP stands for Model Context Protocol. An MCP server is a bridge between AI and an external system. It gives AI the ability to read from and write to tools outside its conversation window.

    Without MCP servers, AI only works with what you paste into chat. With them, AI can query Notion databases, read Gmail, check Google Calendar, interact with Figma, pull analytics, and manage local files.

    I run MCP connections to Notion, Gmail, Google Calendar, Metricool, Figma, and Windows MCP for PowerShell execution. Each server exposes tools the AI can take as actions. MCP servers are connection infrastructure, not intelligence. They make AI more capable, not smarter.

    Layer 2: Skills – The Knowledge

    If MCP servers are roads, skills are maps. A skill is a structured SKILL.md file that tells AI how to do something specific using available tools.

    Without skills, AI knows it can connect to WordPress but not your URL, credentials, content strategy, or publishing workflow. With skills, one sentence triggers a complete operation. I have 60+ skills covering WordPress connections, site auditing, SEO optimization, content generation, Notion operations, social media publishing, and more.

    Every hour spent writing skills saves 10+ hours of future session time.

    Layer 3: Plugins – The Packages

    Plugins bundle skills, MCP configurations, and tools into installable capability packages. A WordPress optimization plugin bundles 15+ skills with reference files and configurations.

    Plugins solve distribution. Building 60+ skills took months. A plugin lets someone install an entire workflow domain in minutes. The architecture enables composability – each plugin handles its domain and connects cleanly to others.

    How the Three Layers Work Together

    I say: Run the content intelligence audit on a luxury asset lender.com and generate 15 draft articles.

    Plugin layer: The wp-content-intelligence plugin activates with its audit and batch creator skills.

    Skill layer: The audit skill loads credentials from the site registry and understands the full methodology.

    MCP layer: Windows MCP executes PowerShell commands calling WordPress REST API through the proxy.

    Three layers, one sentence trigger. Remove any layer and the workflow breaks.

    The Maturity Model

    Level 1 – Prompts: Raw chat, no infrastructure. Where 95% of AI users are.

    Level 2 – MCP Connections: AI reads and writes to your systems. Dramatically more useful.

    Level 3 – Skills: Instruction files capture workflows and credentials. Operational AI begins.

    Level 4 – Plugins: Packaged capability bundles. Workflows become portable and composable.

    Level 5 – Autonomous Agents: Skills run on schedules without human triggers. AI becomes a colleague.

    Frequently Asked Questions

    Do I need to be a developer to build skills?

    No. Skills are markdown files. If you can write clear instructions for a task, you can write a skill. No code required.

    How do MCP servers handle authentication?

    Each has its own mechanism. Notion uses integration tokens. Gmail uses OAuth2. You authenticate once and the connection persists across sessions.

    Can skills call other skills?

    Yes. The wp-full-refresh skill calls wp-seo-refresh, wp-aeo-refresh, wp-geo-refresh, wp-schema-inject, and wp-interlink in sequence. Complex workflows from modular single-purpose skills.

    What is the difference between a skill and a prompt template?

    Scope and persistence. A prompt template is a text string. A skill is a persistent file with context, credentials, reference data, quality standards, and step-by-step procedures. The difference is between a recipe and a kitchen.

    Start Building Infrastructure, Not Prompts

    The next time you spend 10 minutes explaining context to AI, write a skill instead. The next time you manually copy data between platforms, set up an MCP connection. Prompts are disposable. Infrastructure compounds.

  • The Client Name Guardrail: What Happens When AI Publishes Too Fast for Human Review

    The Mistake That Created the Rule

    I published 12 articles to the agency blog in a single session. World-class content. Properly optimized. Well-structured. And scattered throughout them were real client names – actual companies we serve, mentioned by name in case studies, examples, and operational descriptions.

    This was not malicious. It was the natural output of an AI that had access to my full operational context – including which companies I work with, what industries they are in, and what we have built for them. When I asked for content drawn from real work, the AI delivered exactly that. Including the parts that should have stayed confidential.

    I caught it during review. Every article was scrubbed clean within the hour. But the incident exposed a fundamental gap in AI-assisted content publishing: when AI can publish at machine speed, human review becomes the bottleneck – and bottlenecks get skipped.

    So I built the client name guardrail. A systematic prevention layer that catches confidential references before they reach a publish command, no matter how fast the content is being produced.

    The Protected Entity List

    The foundation is a maintained list of every client, company, and entity name that must never appear in published content without explicit approval. The list currently contains 20+ entries covering all active clients across every business entity.

    But names are not simple strings. People reference the same company in multiple ways. “The restoration client in Colorado” is fine. “a restoration company” is not. “Our luxury lending partner” is fine. “a luxury lending firm Company” is not. The entity list includes not just official company names but common abbreviations, nicknames, and partial references that could identify a client.

    The Genericization Table

    Simply blocking client names would break the content. If the AI cannot reference specific work, the articles become generic and lose the authenticity that makes them valuable. The solution is a genericization table – a mapping of specific references to anonymous equivalents that preserve the insight without revealing the identity.

    “a cold storage facility” becomes “our cold storage client.” “a luxury lending firm” becomes “a luxury lending partner.” “a restoration company” becomes “a restoration company in the Mountain West.” Each mapping is specific enough to be useful but generic enough to protect confidentiality.

    The AI applies these substitutions automatically during content generation. It still draws from real operational experience. It still provides specific, authentic examples. But the identifying details are replaced before the content is written, not after.

    The Pre-Publish Scan

    The final layer is a regex-based scan that runs against every piece of content before a publish API call is made. The scan checks the title, body content, excerpt, and slug against the full protected entity list. If any match is found, the publish is blocked and the specific matches are surfaced for review.

    This scan catches edge cases the genericization table misses – a client name that slipped through in a quote, a URL that contains a company domain, or a reference the AI constructed from context rather than the entity list. The scan is the safety net that ensures nothing gets through even when the primary prevention layer fails.

    Why This Matters Beyond My Situation

    Every agency, consultancy, and service provider using AI for content creation faces this risk. AI models are trained to be helpful and specific. When given access to client context, they will use that context to produce better content. That is exactly what you want – until the specificity includes information your clients did not consent to having published.

    The risk scales with capability. A basic AI tool that generates generic blog posts will never mention your clients because it does not know about them. An AI system deeply integrated with your operations – reading your Notion databases, processing your email, accessing your WordPress sites – knows everything about your client relationships. That integration is what makes it powerful. It is also what makes it dangerous without guardrails.

    The pattern I built is transferable to any agency: maintain a protected entity list, build a genericization mapping, and scan before publishing. The implementation takes about 2 hours. The alternative – publishing client names and discovering it after the content is indexed by Google – takes much longer to fix and costs trust that cannot be rebuilt with a quick edit.

    Frequently Asked Questions

    Does the guardrail slow down content production?

    Negligibly. The genericization happens during content generation, adding zero time to the process. The pre-publish scan takes under 2 seconds per article. In a 15-article batch, that is 30 seconds of total overhead.

    What about client names in internal documents vs. published content?

    The guardrail only activates on publish workflows. Internal documents, Notion entries, and operational notes use real client names because they are not public-facing. The skill triggers specifically when content is being sent to a WordPress REST API endpoint or any other publishing channel.

    Can clients opt in to being named?

    Yes. The protected entity list supports an override flag. If a client explicitly approves being referenced by name – for a case study, testimonial, or co-marketing piece – their entry can be temporarily unflagged. The default is always protected. Opt-in is explicit.

    Has the guardrail caught anything since the initial incident?

    Yes – three times in the first week. All were subtle references the AI constructed from context rather than direct mentions. One was a geographic description specific enough to identify a client’s location. The scan caught it. Without the guardrail, all three would have been published.

    Speed Needs Guardrails

    The ability to publish 15 articles in a single session is a superpower. But superpowers without controls are liabilities. The client name guardrail is not about slowing down. It is about publishing at machine speed with human-grade judgment on confidentiality. The AI produces the content. The guardrail produces the trust.

  • We Manage 23 WordPress Sites. Here’s the Exact Stack.

    The Scale Problem Nobody Talks About

    Managing one WordPress site is straightforward. Managing five gets complicated. Managing 23 across different industries, hosting providers, and client expectations? That’s an engineering problem disguised as a marketing job.

    Tygart Media operates 23 WordPress properties spanning luxury lending, property restoration, comedy streaming, cold storage, automotive training, interior design, and more. Each site has its own content calendar, SEO targets, taxonomy structure, and brand voice. Without automation, this operation would require a team of 8-10 people. We run it with three.

    The WordPress REST API Proxy

    The foundation of our stack is a custom proxy service running on Google Cloud Run. Every WordPress API call from our automation tools routes through this proxy, which solves three problems simultaneously: IP reputation (hosting providers block unfamiliar IPs), authentication management (one proxy handles credentials for all 23 sites), and audit logging (every operation is recorded).

    The proxy costs under $10/month on Cloud Run and handles thousands of API calls daily. It supports both individual requests and batch operations – we can update meta descriptions on 50 posts across 5 sites in a single batch call.

    The Skill Library: 30+ WordPress Operations

    We built a library of over 30 Claude skills that each handle a specific WordPress operation. wp-seo-refresh optimizes on-page SEO. wp-schema-inject adds structured data markup. wp-interlink builds internal link graphs. wp-aeo-refresh optimizes for answer engines. wp-geo-refresh optimizes for generative AI citations.

    Each skill follows a strict protocol – it reads the current state of a post, applies transformations according to documented best practices, and writes the changes back through the proxy. Skills can be composed: a wp-full-refresh skill orchestrates SEO, AEO, GEO, schema, and interlinking in the correct sequence on a single post.

    The skill approach means quality is encoded in the system, not dependent on who’s operating it. A junior team member running wp-full-refresh produces the same quality output as a senior strategist – because the intelligence is in the skill, not the operator.

    Content Intelligence and Gap Analysis

    Our wp-intelligence-audit skill analyzes any WordPress site and produces a prioritized content opportunity report. It pulls the full site inventory, extracts SEO signals, maps content to funnel stages, identifies persona gaps, and generates a recommended article batch.

    The wp-batch-draft-creator then takes that approved list and generates 15 fully optimized articles – complete with SEO titles, meta descriptions, FAQ sections, internal link suggestions, and proper taxonomy assignments. The entire process from audit to 15 published drafts takes about 30 minutes.

    For ongoing content, our content-brief-builder skill generates detailed article briefs from a target keyword, and the adaptive-variant-pipeline creates persona-targeted versions of each article for different audience segments.

    Monitoring and Maintenance

    Automation doesn’t mean set-it-and-forget-it. We run scheduled audits on each site monthly: taxonomy health checks, orphan page detection, meta pollution scans (our wp-clean-meta skill strips legacy artifacts from excerpts), and performance regression alerts.

    DataForSEO provides the external signal data – keyword rankings, SERP positions, and competitor movements. SpyFu fills in the competitive intelligence layer. Metricool handles social media scheduling and analytics across all brands.

    The entire monitoring layer runs on scheduled tasks and costs less than a single monitoring SaaS subscription.

    Frequently Asked Questions

    Does this stack work with any WordPress hosting provider?

    Yes. The proxy layer abstracts away hosting differences. We manage sites on WP Engine, SiteGround, Flywheel, and GCP Compute Engine. The proxy handles authentication and IP reputation for all of them.

    How long does it take to onboard a new WordPress site?

    About 20 minutes. Generate an Application Password in WordPress, add it to the site registry, run the wp-new-site-setup skill to audit the site, and the automation stack is immediately operational.

    What happens if the proxy goes down?

    Cloud Run has a 99.95% uptime SLA. In 6+ months of operation, we’ve had zero unplanned downtime. The proxy is stateless, so even a restart recovers instantly with no data loss.

    Can a non-technical person use this stack?

    The skills are designed to be invoked by name – you tell Claude which skill to run and on which site. You don’t need to understand the underlying API calls. Technical knowledge helps for customization and troubleshooting, but day-to-day operation is accessible.

    The Compound Effect

    Each individual automation saves 15-30 minutes. Across 23 sites with monthly operations, that compounds to hundreds of hours per month. But the real value isn’t time saved – it’s consistency achieved. Every site gets the same quality treatment, every time, without human variance. That’s the actual competitive advantage.