Tag: Claude AI

  • How We Built a Complete AI Music Album in Two Sessions: The Red Dirt Sakura Story

    How We Built a Complete AI Music Album in Two Sessions: The Red Dirt Sakura Story



    What if you could build a complete music album — concept, lyrics, artwork, production notes, and a full listening experience — without a recording studio, without a label, and without months of planning? That’s exactly what we did with Red Dirt Sakura, an 8-track country-soul album written and produced by a fictional Japanese-American artist named Yuki Hayashi. Here’s how we built it, what broke, what we fixed, and why this system is repeatable.

    What Is Red Dirt Sakura?

    Red Dirt Sakura is a concept album exploring what happens when Japanese-American identity collides with American country music. Each of the 8 tracks blends traditional Japanese melodic structure with outlaw country instrumentation — steel guitar, banjo, fiddle — sung in both English and Japanese. The album lives entirely on tygartmedia.com, built and published using a three-model AI pipeline.

    The Three-Model Pipeline: How It Works

    Every track on the album was processed through a sequential three-model workflow. No single model did everything — each one handled what it does best.

    Model 1 — Gemini 2.0 Flash (Audio Analysis): Each MP3 was uploaded directly to Gemini for deep audio analysis. Gemini doesn’t just transcribe — it reads the emotional arc of the music, identifies instrumentation, characterizes the tempo shifts, and analyzes how the sonic elements interact. For a track like “The Road Home / 家路,” Gemini identified the specific interplay between the steel guitar’s melancholy sweep and the banjo’s hopeful pulse — details a human reviewer might take hours to articulate.

    Model 2 — Imagen 4 (Artwork Generation): Gemini’s analysis fed directly into Imagen 4 prompts. The artwork for each track was generated from scratch — no stock photos, no licensed images. The key was specificity: “worn cowboy boots beside a shamisen resting on a Japanese farmhouse porch at golden hour, warm amber light, dust motes in the air” produces something entirely different from “country music with Japanese influence.” We learned this the hard way — more on that below.

    Model 3 — Claude (Assembly, Optimization, and Publish): Claude took the Gemini analysis, the Imagen artwork, the lyrics, and the production notes, then assembled and published each listening page via the WordPress REST API. This included the HTML layout, CSS template system, SEO optimization, schema markup, and internal link structure.

    What We Built: The Full Album Architecture

    The album isn’t just 8 MP3 files sitting in a folder. Every track has its own listening page with a full visual identity — hero artwork, a narrative about the song’s meaning, the lyrics in both English and Japanese, production notes, and navigation linking every page to the full station hub. The architecture looks like this:

    • Station Hub/music/red-dirt-sakura/ — the album home with all 8 track cards
    • 8 Listening Pages — one per track, each with unique artwork and full song narrative
    • Consistent CSS Template — the lr- class system applied uniformly across all pages
    • Parent-Child Hierarchy — all pages properly nested in WordPress for clean URL structure

    The QA Lessons: What Broke and What We Fixed

    Building a content system at this scale surfaces edge cases that only exist at scale. Here are the failures we hit and how we solved them.

    Imagen Model String Deprecation

    The Imagen 4 model string documented in various API references — imagen-4.0-generate-preview-06-06 — returns a 404. The working model string is imagen-4.0-generate-001. This is not documented prominently anywhere. We hit this on the first artwork generation attempt and traced it through the API error response. Future sessions: use imagen-4.0-generate-001 for Imagen 4 via Vertex AI.

    Prompt Specificity and Baked-In Text Artifacts

    Generic Imagen prompts that describe mood or theme rather than concrete visual scenes sometimes produce images with Stable Diffusion-style watermarks or text artifacts baked directly into the pixel data. The fix is scene-level specificity: describe exactly what objects are in frame, where the light is coming from, what surfaces look like, and what the emotional weight of the composition should be — without using any words that could be interpreted as text to render. The addWatermark: false parameter in the API payload is also required.

    WordPress Theme CSS Specificity

    Tygart Media’s WordPress theme applies color: rgb(232, 232, 226) — a light off-white — to the .entry-content wrapper. This overrides any custom color applied to child elements unless the child uses !important. Custom colors like #C8B99A (a warm tan) read as darker than the theme default on a dark background, making text effectively invisible. Every custom inline color declaration in the album pages required !important to render correctly. This is now documented and the lr- template system includes it.

    URL Architecture and Broken Nav Links

    When a URL structure changes mid-build, every internal nav link needs to be audited. The old station URL (/music/japanese-country-station/) was referenced by Song 7’s navigation links after we renamed the station to Red Dirt Sakura. We created a JavaScript + meta-refresh redirect from the old URL to the new one, and audited all 8 listening pages for broken references. If you’re building a multi-page content system, establish your final URL structure before page 1 goes live.

    Template Consistency at Scale

    The CSS template system (lr-wrap, lr-hero, lr-story, lr-section-label, etc.) was essential for maintaining visual consistency across 8 pages built across two separate sessions. Without this system, each page would have required individual visual QA. With it, fixing one global issue (like color specificity) required updating the template definition, not 8 individual pages.

    The Content Engine: Why This Post Exists

    The album itself is the first layer. But a music album with no audience is a tree falling in an empty forest. The content engine built around it is what makes it a business asset.

    Every listening page is an SEO-optimized content node targeting specific long-tail queries: Japanese country music, country music with Japanese influence, bilingual Americana, AI-generated music albums. The station hub is the pillar page. This case study is the authority anchor — it explains the system, demonstrates expertise, and creates a link target that the individual listening pages can reference.

    From this architecture, the next layer is social: one piece of social content per track, each linking to its listening page, with the case study as the ultimate destination for anyone who wants to understand the “how.” Eight tracks means eight distinct social narratives — the loneliness of “Whiskey and Wabi-Sabi,” the homecoming of “The Road Home / 家路,” the defiant energy of “Outlaw Sakura.” Each one is a separate door into the same content house.

    What This Proves About AI Content Systems

    The Red Dirt Sakura project demonstrates something important: AI models aren’t just content generators — they’re a production pipeline when orchestrated correctly. The value isn’t in any single output. It’s in the system that connects audio analysis, visual generation, content assembly, SEO optimization, and publication into a single repeatable workflow.

    The system is already proven. Album 2 could start tomorrow with the same pipeline, the same template system, and the documented fixes already applied. That’s what a content engine actually means: not just content, but a machine that produces it reliably.

    Frequently Asked Questions

    What AI models were used to build Red Dirt Sakura?

    The album was built using three models in sequence: Gemini 2.0 Flash for audio analysis, Google Imagen 4 (via Vertex AI) for artwork generation, and Claude Sonnet for content assembly, SEO optimization, and WordPress publishing via REST API.

    How long did it take to build an 8-track AI music album?

    The entire album — concept, lyrics, production, artwork, listening pages, and publication — was completed across two working sessions. The pipeline handles each track in sequence, so speed scales with the number of tracks rather than the complexity of any single one.

    What is the Imagen 4 model string for Vertex AI?

    The working model string for Imagen 4 via Google Vertex AI is imagen-4.0-generate-001. Preview strings listed in older documentation are deprecated and return 404 errors.

    Can this AI music pipeline be used for other albums or artists?

    Yes. The pipeline is artist-agnostic and genre-agnostic. The CSS template system, WordPress page hierarchy, and three-model workflow can be applied to any music project with minor customization of the visual style and narrative voice.

    What is Red Dirt Sakura?

    Red Dirt Sakura is a concept album by the fictional Japanese-American artist Yuki Hayashi, blending American outlaw country with traditional Japanese musical elements and sung in both English and Japanese. The album lives on tygartmedia.com and was produced entirely using AI tools.

    Where can I listen to the Red Dirt Sakura album?

    All 8 tracks are available on the Red Dirt Sakura station hub on tygartmedia.com. Each track has its own dedicated listening page with artwork, lyrics, and production notes.

    Ready to Hear It?

    The full album is live. Eight tracks, eight stories, two languages. Start with the station hub and follow the trail.

    Listen to Red Dirt Sakura →



  • DataForSEO + Claude: The Keyword Research Stack That Replaced 3 Tools

    DataForSEO + Claude: The Keyword Research Stack That Replaced 3 Tools

    We used to pay for SEMrush, Ahrefs, and Moz. Then we discovered we could use the DataForSEO API with Claude to do better keyword research, at 1/10th the cost, with more control over the analysis.

    The Old Stack (and Why It Broke)
    We were paying $600+ monthly across three platforms. Each had different strengths—Ahrefs for backlink data, SEMrush for SERP features, Moz for authority metrics—but also massive overlap. And none of them understood our specific context: managing 19 WordPress sites with different verticals and different SEO strategies.

    The tools gave us data. Claude gives us intelligence.

    DataForSEO + Claude: The New Stack
    DataForSEO is an API that pulls real search data. We hit their endpoints for:
    – Keyword search volume and trend data
    – SERP features (snippets, People Also Ask, related searches)
    – Ranking difficulty and opportunity scores
    – Competitor keyword analysis
    – Local search data (essential for restoration verticals)

    We pay $300/month for enough API calls to cover all 19 sites’ keyword research. That’s it.

    Where Claude Comes In
    DataForSEO gives us raw data. Claude synthesizes it into strategy.

    I’ll ask: “Given the keyword data for ‘water damage restoration in Houston,’ show me the 5 best opportunities to rank where we can compete immediately.”

    Claude looks at:
    – Search volume
    – Current top 10 (from DataForSEO)
    – Our existing content
    – Difficulty-to-opportunity ratio
    – PAA questions and featured snippet targets
    – Local intent signals

    It returns prioritized keyword clusters with actionable insights: “These 3 keywords have 100-500 monthly searches, lower competition in local SERPs, and People Also Ask questions you can answer in depth.”

    Competitive Analysis Without the Black Box
    Instead of trusting a platform’s opaque “difficulty score,” we use Claude to analyze actual SERP data:

    – What’s the common word count in top results?
    – How many have video content? Backlinks?
    – What schema markup are they using?
    – Are they targeting the same user intent or different angles?
    – What questions do they answer that we don’t?

    This gives us real competitive insight, not a number from 1-100.

    The Workflow
    1. Give Claude a target keyword and our target site
    2. Claude queries DataForSEO API for volume, difficulty, SERP data
    3. Claude pulls our existing content on related topics
    4. Claude analyzes the competitive landscape
    5. Claude recommends specific keywords with strategy recommendations
    6. I approve the targets, Claude drafts the content brief
    7. The brief goes to our content pipeline

    This entire workflow happens in 10 minutes. With the old tools, it took 2 hours of hopping between platforms.

    Cost and Scale
    DataForSEO is billed per API call, not per “seat” or “account.” We do ~500 keyword researches per month across all 19 sites. Cost: ~$30-40. Traditional tools would cost the same regardless of usage.

    As we scale content, our tool cost stays flat. With SEMrush, we’d hit overages or need higher plans.

    The Limitations (and Why We Accept Them)
    DataForSEO doesn’t have the 5-year historical trend data that Ahrefs does. We don’t get detailed backlink analysis. We don’t have a competitor tracking dashboard.

    But here’s the truth: we never used those features. We needed keyword opportunity identification and competitive insight. DataForSEO + Claude does that better than expensive platforms because Claude can reason about the data instead of just displaying it.

    What This Enables
    – Continuous keyword research (no tool budget constraints)
    – Smarter targeting (Claude reasons about intent)
    – Faster decisions (10 minutes instead of 2 hours)
    – Transparent methodology (we see exactly how decisions are made)
    – Scalable to all 19 sites simultaneously

    If you’re paying for three SEO platforms, you’re probably paying for one platform and wasting the other two. Try DataForSEO + Claude for your next keyword research cycle. You’ll get more actionable intelligence and spend less than a single month of your current setup.

    {
    “@context”: “https://schema.org”,
    “@type”: “Article”,
    “headline”: “DataForSEO + Claude: The Keyword Research Stack That Replaced 3 Tools”,
    “description”: “DataForSEO API + Claude replaces $600/month in SEO tools with $30/month API costs and better analysis. Here’s the keyword research workflow we built.”,
    “datePublished”: “2026-03-30”,
    “dateModified”: “2026-04-03”,
    “author”: {
    “@type”: “Person”,
    “name”: “Will Tygart”,
    “url”: “https://tygartmedia.com/about”
    },
    “publisher”: {
    “@type”: “Organization”,
    “name”: “Tygart Media”,
    “url”: “https://tygartmedia.com”,
    “logo”: {
    “@type”: “ImageObject”,
    “url”: “https://tygartmedia.com/wp-content/uploads/tygart-media-logo.png”
    }
    },
    “mainEntityOfPage”: {
    “@type”: “WebPage”,
    “@id”: “https://tygartmedia.com/dataforseo-claude-the-keyword-research-stack-that-replaced-3-tools/”
    }
    }