The Night Shift That Never Calls In Sick
Every night at 2 AM, while I’m asleep, seven AI agents wake up on my laptop and go to work. One generates content briefs. One indexes every file I created that day. One scans 23 websites for SEO changes. One processes meeting transcripts. One digests emails. One monitors site uptime. One writes news articles for seven industry verticals.
By the time I open my laptop at 7 AM, the work is done. Briefs are written. Indexes are updated. Drift is detected. Transcripts are summarized. Total cloud cost: zero. Total API cost: zero. Everything runs on Ollama with local models.
The Fleet
I call them droids because that’s what they are – autonomous units with specific missions that execute without supervision. Each one is a PowerShell script scheduled as a Windows Task. No Docker. No Kubernetes. No cloud functions. Just scripts, a schedule, and a 16GB laptop running Ollama.
SM-01: Site Monitor. Runs hourly. Pings all 18 managed WordPress sites, measures response time, logs to CSV. If a site goes down, a Windows balloon notification fires. Takes 30 seconds. I know about downtime before any client does.
NB-02: Nightly Brief Generator. Runs at 2 AM. Reads a topic queue – 15 default topics across all client sites – and generates structured JSON content briefs using Llama 3.2 at 3 billion parameters. Processes 5 briefs per night. By Friday, the week’s content is planned.
AI-03: Auto-Indexer. Runs at 3 AM. Scans every text file across my working directories. Generates 768-dimension vector embeddings using nomic-embed-text. Updates a local vector index. Currently tracking 468 files. Incremental runs take 2 minutes. Full reindex takes 15.
MP-04: Meeting Processor. Runs at 6 AM. Scans for Gemini transcript files from the previous day. Extracts summary, key decisions, action items, follow-ups, and notable quotes via Ollama. I never re-read a transcript – the processor pulls out what matters.
ED-05: Email Digest. Runs at 6:30 AM. Categorizes emails by priority and generates a morning digest. Flags anything that needs immediate attention. Pairs with Gmail MCP in Cowork for full coverage across 4 email accounts.
SD-06: SEO Drift Detector. Runs at 7 AM. Checks all 23 WordPress sites for changes in title tags, meta descriptions, H1 tags, canonical URLs, and HTTP status codes. Compares against a saved baseline. If someone – a client, a plugin, a hacker – changes SEO-critical elements, I know within 24 hours.
NR-07: News Reporter. Runs at 5 AM. Scans Google News RSS for 7 industry verticals – restoration, luxury lending, cold storage, comedy, automotive training, healthcare, ESG. Generates news beat articles via Ollama. 42 seconds per article, about 1,700 characters each. Raw material for client newsletters and social content.
Why Local Beats Cloud for This
The obvious question: why not run these in the cloud? Three reasons.
Cost. Seven agents running daily on cloud infrastructure – even serverless – would cost -400/month in compute, storage, and API calls. On my laptop, the cost is the electricity to keep it plugged in overnight.
Privacy. These agents process client data, email content, meeting transcripts, and SEO baselines. Running locally means none of that data leaves my machine. No third-party processing agreements. No data residency concerns. No breach surface.
Speed of iteration. When I want to change how the brief generator works, I edit a PowerShell script and save it. No deployment pipeline. No CI/CD. No container builds. The change takes effect on the next scheduled run. I’ve iterated on these agents dozens of times in the past week – each iteration took under 60 seconds.
The Compounding Effect
The real power isn’t any single agent – it’s how they feed each other. The auto-indexer picks up briefs generated by the brief generator. The meeting processor extracts topics that feed into the brief queue. The SEO drift detector catches changes that trigger content refresh priorities. The news reporter surfaces industry developments that inform content strategy.
After 30 days, the compound knowledge base is substantial. After 90 days, it’s a competitive advantage that no competitor can buy off the shelf.
Frequently Asked Questions
What specs does your laptop need?
16GB RAM minimum for running Llama 3.2 at 3B parameters. I run on a standard Windows 11 machine – no GPU, no special hardware. The 8B parameter models work too but are slower. For the vector indexer, you need about 1GB of free disk per 1,000 indexed files.
Why PowerShell instead of Python?
Windows Task Scheduler runs PowerShell natively. No virtual environments, no dependency management, no conda headaches. PowerShell talks to COM objects (Outlook), REST APIs (WordPress), and the file system equally well. For a Windows-native automation stack, it’s the pragmatic choice.
How reliable is Ollama for production tasks?
For structured, protocol-driven tasks – very reliable. The models follow formatting instructions consistently when the prompt is specific. For creative or nuanced work, quality varies. I use local models for extraction and analysis, cloud models for creative generation. Match the model to the task.
Can I replicate this setup?
Every script is under 200 lines of PowerShell. The Ollama setup is one install command and one model pull. The Windows Task Scheduler configuration takes 5 minutes per task. Total setup time for all seven agents: under 2 hours if you know what you’re building.
The Future Runs on Your Machine
The narrative that AI requires cloud infrastructure and enterprise budgets is wrong. Seven autonomous agents. One laptop. Zero cloud cost. The work gets done while I sleep. If you’re paying monthly fees for automations that could run on hardware you already own, you’re subsidizing someone else’s margins.
Leave a Reply