Category: Uncategorized

  • Notion AI for Product Managers: Specs, Roadmaps, and Stakeholder Updates

    Notion AI for Product Managers: Specs, Roadmaps, and Stakeholder Updates

    The 60-second version

    PMs spend 60% of their time writing — specs, updates, briefs, summaries. Custom Agents take that down to 20%. The PM defines the problem and the strategic call; the agent produces the documentation. Specs draft from a problem statement. Stakeholder updates generate in three audience-specific versions from one source. User research synthesizes into themes automatically. The PM gets back to the work that PMs are actually hired for: deciding what to build.

    Four PM-specific agent patterns

    1. The spec drafting agent. Triggered when a new initiative is added with a problem statement. Pulls related research, prior similar specs, technical constraints from engineering pages. Drafts a structured spec with goals, non-goals, user stories, success metrics, open questions. PM reviews and decides; doesn’t start blank.
    2. The audience-tailored update agent. Single input: this week’s progress and risks. Three outputs: exec brief (3 paragraphs, headline-led), engineering update (technical detail, dependencies), customer-facing update (benefits framing). Audience-specific framing automated.
    3. The research synthesis agent. Triggered when interview notes land in the research database. Extracts themes, codes responses, identifies patterns across interviews, ranks insights by frequency and impact. PM gets a synthesis instead of a pile of raw notes.
    4. The roadmap maintenance agent. Reads the roadmap database. When initiatives change status or priority, updates the Now/Next/Later view, drafts the rationale for moves, flags timeline conflicts. The roadmap stays current without weekly reformatting.

    What stays PM

    • Strategic prioritization (what to build, what to kill)
    • Customer conversations
    • Cross-functional negotiation
    • Final spec approval
    • The judgment behind every roadmap move
      The agent makes the writing fast. It doesn’t make the deciding fast.

    The compounding effect

    PMs running this pattern report a category change in their work: less time on producing artifacts, more time on customer conversations and strategic calls. The artifacts still exist (specs, updates, roadmaps) but they’re produced faster and revised more often because revising is cheap.
    A weekly artifact that used to take 4 hours now takes 90 minutes. Across 50 weeks, that’s 125 hours reclaimed per PM per year. Most PMs spend that on the work they were always supposed to be doing.

    Where PMs go wrong

    1. Letting the agent draft success metrics. Metrics are strategic. The agent can suggest; the PM decides. Don’t outsource the metric definition.
    2. Trusting cross-team updates without verification. The agent might miss context from another team. Sample-check updates that go to engineering or sales for accuracy before sending.
    3. Producing more artifacts because production is cheap. Cheap production is a temptation to over-produce. The discipline of “what should we actually communicate” matters more, not less.

    What to read next

    Notion AI for Engineering, Synthesize Research piece, AI-Native Company Patterns.

  • Notion AI for Engineering: Standups, Postmortems, and Architecture Records

    Notion AI for Engineering: Standups, Postmortems, and Architecture Records

    The 60-second version

    Engineers hate documentation. Documentation rots. Custom Agents fix the documentation rot without making engineers do the documentation. Standups generate from commits and tickets. Postmortems draft from incident channels. ADRs and runbooks stay current because the agent updates them when related pages change. The engineering org gets the documentation discipline of a regulated industry without the cultural cost.

    Four engineering-specific agent patterns

    1. The standup synthesis agent. Runs daily at 9 AM. Reads each engineer’s commits since last standup, ticket movements, Slack #standup channel posts. Produces a structured “yesterday/today/blockers” entry for each engineer. The standup meeting becomes a 5-minute review of pre-generated content instead of a 30-minute round-robin.
    2. The incident postmortem agent. Triggered when an incident is marked resolved. Reads the incident channel, status page updates, related PRs, and prior incidents. Drafts a blameless postmortem in the team’s template. Engineering reviews and refines instead of starting blank.
    3. The ADR maintenance agent. Watches the ADR database. When an architecture page or related design doc changes, flags the related ADR for update. Suggests the diff. Drafts the supersession or amendment record.
    4. The on-call runbook agent. Reads operational runbooks, cross-references with recent incidents. When an incident pattern emerges that the runbook doesn’t cover, drafts the runbook update. On-call rotates with current docs, not stale ones.

    What stays human

    • Architecture decisions
    • Code review (for now — agent-assisted code review is a different topic)
    • Incident response in the moment
    • Hiring decisions on engineering candidates
    • The judgment about whether a draft postmortem captures the right lessons

    The standup transformation

    Pre-agent standups: 30 minutes, mostly people remembering what they did yesterday and reciting it.
    Post-agent standups: 5-10 minutes, reviewing pre-generated content and surfacing only the friction the agent missed.
    This isn’t theoretical. Teams running this pattern reclaim 25 minutes per engineer per day. At a 10-engineer team, that’s roughly 4 engineering hours daily. Real money.

    Where engineering teams go wrong

    1. Trusting the agent to identify root cause. Agents synthesize what happened. They don’t reliably identify why. Root cause analysis is human work; the agent prepares the timeline.
    2. Letting ADRs autofill without engineer review. ADRs document decisions. Decisions are human. Agents draft; engineers approve and sign.
    3. Skipping the standup discussion. The standup isn’t just status; it’s friction surfacing. If the agent-generated standup leads to skipping the meeting entirely, friction accumulates silently. Keep the meeting; just make it shorter.

    What to read next

    Workers for Agents in TypeScript, Notion AI for Product Managers, AI-Native Company Patterns, Editorial Surface Area.

  • Notion AI for Legal Ops: Contract Review Triage Without Replacing Counsel

    Notion AI for Legal Ops: Contract Review Triage Without Replacing Counsel

    The 60-second version

    Legal ops is constrained by counsel time. Custom Agents change which work counsel actually has to do. Routine NDAs that match the playbook? Triaged and approved. Contracts with non-standard clauses? Flagged with the specific deviations and counsel reviews only those. Vendor compliance trackers? Auto-updated. Meeting briefings? Drafted. Counsel reviews exceptions; agents handle volume. The split protects legal quality while massively expanding throughput.

    Four legal-ops-specific agent patterns

    1. The NDA triage agent. New NDA arrives. Agent compares it against the playbook (standard mutual NDA terms, acceptable carveouts, dealbreakers). Classifies as GREEN (auto-approve), YELLOW (counsel review), or RED (substantive negotiation). For GREEN, drafts the response. For YELLOW/RED, prepares a deviation report.
    2. The contract review preparation agent. Triggered for any contract not handled by NDA triage. Reads the contract, compares against playbook, marks every deviation, and produces a redline-ready summary for counsel. Counsel opens the document and starts reviewing the deviations directly, not the entire contract.
    3. The vendor compliance tracker. Maintains a database of vendor agreements, renewal dates, surviving obligations, and required documents (DPA, BAA, COI). Flags upcoming renewals 60 days out and missing documentation continuously.
    4. The meeting brief agent. Before any contract negotiation or compliance meeting, pulls relevant context: prior agreements with the counterparty, related correspondence, current playbook positions on the topics expected. Counsel walks in prepped without the prep work.

    What absolutely stays counsel

    The non-negotiable boundaries:
    – Legal advice (period — agents never deliver this)
    – Substantive contract negotiation strategy
    – Risk assessment on novel issues
    – Anything that gets sent to opposing counsel as the firm’s position
    – Privileged communications
    Agents prepare the inputs to counsel’s judgment. They never replace the judgment.

    The triage discipline

    The triage agent only works if the playbook is explicit. “Standard NDA” is not a playbook; “12-month confidentiality, mutual obligation, no non-solicit, US jurisdiction acceptable, EU DPA required if data crosses border” is. The discipline of writing the playbook is what makes the agent reliable.
    Most legal ops teams underestimate how much playbook documentation they need. The first 90 days of a legal-ops agent rollout is mostly playbook work, not agent building.

    Where this goes wrong

    1. Treating the agent’s classification as final. GREEN means “agent thinks this matches playbook.” It doesn’t mean “approved without review.” A spot-check on 10% of GREEN classifications keeps the system honest.
    2. Letting the agent draft anything that goes to opposing counsel. Even a “thank you, attached is our standard NDA” response should have counsel eyes before send for high-stakes counterparties.
    3. Building too aggressive a YELLOW threshold. If too much routes to counsel, the agent isn’t saving time. Tighten YELLOW criteria. If too little routes, the agent is missing things — loosen YELLOW.

    What to read next

    Notion AI for Operations Managers, Notion AI for Finance, Vendor Check, Editorial Surface Area.

  • Notion AI for HR: Onboarding Plans, Policy Lookups, and Performance Cycles

    Notion AI for HR: Onboarding Plans, Policy Lookups, and Performance Cycles

    The 60-second version

    HR is split between policy and people. The policy half is largely automatable. The people half isn’t. Custom Agents take over the lookup, documentation, and template-generation work that consumes HR teams, freeing them for the relationship and judgment work that requires being human. The result is HR teams that feel less like document processors and more like organizational coaches.

    Four HR-specific agent patterns

    1. The onboarding plan agent. Triggered when a new hire is added to the people database. Pulls role-specific onboarding template, customizes for team and start date, schedules Day 1 / Week 1 / 30/60/90-day milestones, drafts welcome communications. Manager arrives on Day 1 with a customized plan, not a generic one.
    2. The policy lookup agent. Anyone in the company asks: “Can I work remotely from another country?” or “What’s our PTO policy?” Agent answers in plain language, citing the specific policy page. Frees HR from being the policy answering desk.
    3. The performance review prep agent. Quarterly. Pulls each manager’s direct reports, drafts review templates with prior cycle ratings, recent project work, and feedback patterns. Manager opens a populated draft, not a blank one.
    4. The recruiting pipeline agent. Daily across the recruiting database. Updates candidate stage based on activity, flags candidates stalled in stages, drafts follow-up communications. Recruiting status meeting starts at “what about these flagged ones” instead of “where are we.”

    What stays human (and should)

    • Compensation decisions
    • Performance ratings and the conversations behind them
    • Conflict mediation
    • Hiring decisions
    • Layoff or termination calls
    • Anything that requires reading the room
      The agents make HR humans more available for the work that matters. They don’t replace them at it.

    The privacy layer matters more here

    HR data is sensitive. Three guardrails:
    – Scope agents tightly — an HR agent should not have access to engineering project pages, finance data, or anything outside HR’s lane.
    – Audit agent access logs monthly. Know what the agent has read.
    – Apply the company’s data handling policy to agent inputs and outputs the same way you would to any HR system.

    Where HR teams go wrong

    1. Letting agents draft sensitive communications. Termination letters, performance improvement plans, complaint responses — these need human authorship. Agents can pull templates; humans write them.
    2. Trusting policy answers without verification. Policy interpretation has nuance. The agent’s plain-language answer should always cite the underlying policy doc so users can verify. Sample-check 10% monthly.
    3. Replacing the recruiter’s judgment with the agent’s pipeline view. Agents update status; recruiters decide who to advance. Don’t let the agent close candidate records autonomously.

    What to read next

    Notion AI for Operations Managers, Notion AI for Legal Ops, AI-Native Company Patterns, When Not to Use a Notion Agent.

  • Notion AI for Operations Managers: SOPs, Runbooks, and the Audit Trail

    Notion AI for Operations Managers: SOPs, Runbooks, and the Audit Trail

    The 60-second version

    Ops managers spend their days holding the operational fabric together — keeping SOPs current, ensuring procedures get followed, catching exceptions, communicating status. Custom Agents excel at exactly this category of work because the patterns are well-defined and the value of consistency is high. The ops manager’s job shifts from “running procedures” to “designing the agents that run procedures and handling what they can’t.”

    Four agents every ops function needs

    1. The SOP currency agent. Runs weekly. Reads each SOP page. Cross-references it against recent activity in related databases. Flags SOPs that haven’t been updated in 90 days OR where the actual practice has drifted from the documented process. Output: a one-page report on SOP health.
    2. The procedure execution agent. Triggered by named events (onboarding new hire, incident response, monthly close). Walks through the procedure step by step, executing or assigning each step, logging completion to an audit trail database. Pauses when human input is required.
    3. The exception triage agent. Watches a designated “exceptions” database. Categorizes incoming exceptions by type, urgency, and owner. Drafts initial response. Flags pattern exceptions (multiple of the same type) for systemic review.
    4. The status synthesis agent. Reads across team databases. Produces the weekly ops report — what’s running, what’s at risk, what shipped, what’s behind. Goes to leadership. Saves the ops manager 4-6 hours weekly.

    The audit trail dividend

    Custom Agents write audit logs by default. Every step they take, every page they read, every change they make is logged. For ops functions in regulated environments — finance, healthcare, legal-adjacent — this is meaningful. The agent’s audit trail is more thorough than what humans typically log because humans cut corners on logging when they’re under time pressure. Agents don’t.
    This shifts the conversation with auditors. “Show me your procedure” becomes “here’s the procedure and here’s every execution log for the last 12 months.” That’s a posture change.

    Where ops managers go wrong with agents

    1. Building agents for procedures that aren’t documented well. If the SOP is vague, the agent’s execution will be vague. Tighten the SOP first. Then build the agent.
    2. Trusting agent execution without sampling. Sample 10% of agent runs monthly. Look at the audit trail. Verify it matches reality. Drift happens silently.
    3. Replacing exception handling with an agent. Exception handling is judgment work. Agents categorize and surface; humans decide. Don’t let the agent close exception tickets autonomously without review.

    What this enables

    Ops managers running this pattern report: more time on systemic improvement, less time on procedure execution. More confidence in audit posture, less anxiety about gaps. More leverage per ops headcount, fewer manual handoffs.

    What to read next

    SOX Testing pieces in finance cluster, Compliance, Editorial Surface Area, AI-Native Company Patterns.

  • Notion AI for Sales Teams: Pipeline Hygiene, Call Notes, and Account Research

    Notion AI for Sales Teams: Pipeline Hygiene, Call Notes, and Account Research

    The 60-second version

    Sales reps universally underdo documentation. The CRM never matches reality. Account research is shallow. Call notes are sparse. Custom Agents fix the documentation layer without taking selling time. Autofill keeps deal properties current as pages change. A research agent prepares accounts before discovery calls. A post-call agent processes notes into structured updates and action items. The rep stays selling; the documentation layer maintains itself.

    Four sales-specific agent patterns

    1. The pipeline hygiene agent. Daily Autofill across the pipeline database. Keeps deal stage, last-activity date, and risk flags current based on page content. Surfaces deals that haven’t moved in 14+ days. The pipeline review meeting starts 70% prepped.
    2. The pre-call research agent. Triggered 60 minutes before each external meeting. Pulls company info, attendee LinkedIn data (via integration), prior interactions, related industry notes. Drops a one-page brief in the rep’s inbox.
    3. The post-call summary agent. Triggered when call notes land in a designated database. Extracts action items, identifies decision-makers mentioned, updates the account record, drafts the follow-up email. Rep reviews and sends in 90 seconds.
    4. The deal-at-risk agent. Weekly. Reads the pipeline. Flags deals where engagement has dropped, decision timelines have slipped, or red flags emerged in recent communications. Surfaces 3-5 deals for the rep’s attention each Monday.

    What this changes for the rep’s day

    Pre-agent: 60-70% of the day is selling, 30-40% is documentation and prep.
    Post-agent: 80-85% selling, 15-20% review and judgment.
    The math is dramatic at the team level. A 10-person sales team gets back roughly two FTE-equivalents of selling capacity. That capacity converts directly to pipeline.

    What stays human

    • The actual conversations with prospects
    • Negotiation and closing
    • Relationship building
    • Strategic account decisions (which accounts to pursue, which to disqualify)
    • The judgment about whether agent-summarized notes captured the call accurately
      Agents make documentation cheaper. They don’t make selling cheaper.

    The CRM integration question

    Many sales teams already have Salesforce, HubSpot, or another CRM. The pattern that works: keep the CRM as system-of-record, use Notion + agents as the working layer. Sync key fields back to the CRM via integration. Don’t try to replace the CRM with Notion.
    The exception: small teams (<10 reps) sometimes find that Notion + agents is enough and they don’t need a separate CRM. Test before committing.

    Where this goes wrong

    1. Autofill on judgment fields. “Deal probability” or “deal health” are judgment calls. Don’t autofill them. Surface the inputs that inform the judgment; let the rep make the call.
    2. Trusting post-call summaries without review. The agent processes the notes, but the notes themselves can be sparse or wrong. Sample-check post-call summaries weekly.
    3. Replacing the rep’s judgment about which deals to focus on. The deal-at-risk agent surfaces; it doesn’t decide. Reps with full context outperform any agent’s prioritization.

    What to read next

    Notion AI for Customer Success, Account Research piece in Sales cluster, AI-Native Company Patterns.

  • Notion AI for Agency Owners: The Client Delivery Workflow That Scales

    Notion AI for Agency Owners: The Client Delivery Workflow That Scales

    The 60-second version

    Agency margins are bounded by what humans can produce per hour. Custom Agents change the unit economics. An agency that builds a per-client agent loadout — status reports, content production, intake triage, deliverable drafting — can serve more clients with the same headcount, or serve the same clients with better quality. The constraint shifts from “production capacity” to “exception handling capacity.” Agencies that figure this out first compound their advantage.

    The per-client agent pattern

    For each client, build:
    A status report agent that produces the weekly client update from project data
    A deliverable draft agent customized to the client’s voice and brand
    An intake/inbox agent that handles their incoming work (if you manage their queues)
    A QA agent that runs deliverables through a client-specific checklist before they ship
    Each agent is scoped to that client’s databases, voice samples, and brand guide. The setup is non-trivial — first client takes a week — but each subsequent client takes hours, not days.

    What changes in agency economics

    Pre-agent agencies: revenue = headcount x billable rate. Margins compressed by labor cost.
    Post-agent agencies: revenue = (headcount x judgment work) + (agents x operational work). Margins expand because the operational work scales without headcount.
    This isn’t speculative. The agencies running this pattern in 2026 are the ones quietly outperforming their peers on margin while charging similar rates.

    Three pitfalls to avoid

    1. Selling agent-produced work as bespoke. Clients smell it. Don’t pretend a templated digest is hand-written. Be transparent about which work is agent-assisted and which is human; charge accordingly.
    2. Skipping the QA layer. Agent output ships through a human gate. Always. The agency’s reputation rides on the QA gate, not the agent’s output.
    3. Building one mega-agent instead of per-client agents. A single agent serving all clients hits voice and context boundaries hard. Per-client agents perform meaningfully better.

    The pricing implication

    After May 4, 2026, agency credit budgets become real. A client whose agent loadout consumes \$50/month in credits should see that in the cost of service. Agencies that absorb credit costs silently are eating into their own margin. Agencies that pass them through transparently (or bundle them into a “Custom Agent layer” line item) protect margin and educate clients.

    Onboarding clients into this model

    Three things to communicate during onboarding:
    – Which deliverables are agent-assisted and which are human-led
    – How the QA layer works (what gets reviewed, by whom)
    – Why this produces better consistency than a junior staffer would (controlled vocabulary, standardized format)
    Done well, “agent-assisted delivery” becomes a selling point, not a hidden cost.

    What to read next

    Notion AI for Content Teams, ROI Math, From Drafts to WordPress Publish.

  • Notion AI for Content Teams: From Brief to Publish Without Leaving Notion

    Notion AI for Content Teams: From Brief to Publish Without Leaving Notion

    The 60-second version

    The pre-AI content workflow was tools sprawl: brief in one app, research in another, draft in Google Docs, edit in Word, publish in WordPress. The Notion-native AI workflow collapses all of that. Brief lives in a Notion database. An agent enriches it with research. A second agent drafts from the brief. A fact-check agent flags claims. An editor reviews in-line. Publish goes to WordPress via integration. The whole pipeline lives in one workspace, fully visible, fully auditable.

    The four-agent content pipeline

    1. The brief enrichment agent. Triggers when a new brief lands in the briefs database. Pulls related sources, prior coverage, current SEO data (via integration), and competitor context. Fills properties: target keyword cluster, related internal links, missing-coverage angle, recommended word count.
    2. The draft production agent. Skill-driven. Reads the enriched brief, produces a first draft to the team’s house format. Includes pull quotes, internal links, AEO snippet block, sources cited inline.
    3. The fact-check agent. Reads the draft, checks every numerical claim and named entity against sources. Flags unverifiable claims for human review. Outputs a fact-check report alongside the draft.
    4. The editor prep agent. Formats the draft for editorial review — adds the rubric, the review surface, a side-by-side change-tracker against the brief, and pulls the relevant style guide sections. The human editor opens this and starts work, doesn’t have to assemble it.

    What stays human

    • Editorial judgment (does this argument work)
    • Voice match (does it sound like us)
    • Structural decisions (is this the right shape for this idea)
    • Final approval before publish
      The agents handle volume; the editor handles judgment. That split is what makes the pipeline scale without losing voice.

    Volume math

    A four-person content team running this pipeline can ship 2-3x the volume of a same-size team without it. The bottleneck shifts from drafting to editing. That’s the right bottleneck — humans editing well-drafted material is a different speed than humans drafting from scratch.
    Concretely: a team that previously shipped 8 articles/week can ship 16-24 with the same headcount. Quality holds if the gates hold.

    Where this fails

    Three failure modes:
    Voice flatness over time. The pipeline produces consistent output. Consistent shades into bland. Ship in voice samples and varied prompt patterns to keep the corpus textured.
    Citation laziness. Fact-check agents are good but not perfect. Editorial spot-checks remain mandatory.
    Brief sloppiness compounding. A bad brief becomes a bad draft becomes wasted edit time. The brief is the most important gate in the pipeline.

    What to read next

    Editorial Surface Area, Gates Before Volume, From Drafts to WordPress Publish.

  • The Solo Operator’s Notion AI Stack: Running Multiple Businesses With One Agent Team

    The Solo Operator’s Notion AI Stack: Running Multiple Businesses With One Agent Team

    The 60-second version

    Running multiple businesses solo used to mean either hiring an assistant or accepting that things slipped through. Custom Agents change the math. A small agent team — three to seven specialized agents — handles the operational layer across all businesses simultaneously, leaving the operator to focus on relationships, strategy, and exception work. The cost is real (post-May 4, somewhere between a coffee budget and a low-end consultant invoice per month) but the leverage is dramatic. The skill isn’t building agents. It’s deciding what to delegate to them.

    The starter loadout

    Seven agents that earn their keep for a multi-business solo operator:
    1. The morning briefing agent. Runs at 6 AM. Reads overnight emails, calendar for the day, project status changes across all businesses. Drops a one-page digest in your daily notes. You read it with coffee.
    2. The intake triage agent. Triggers on new inbound (form submissions, sales leads, partnership inquiries). Categorizes by business, urgency, and type. Drafts a first response. Routes for review.
    3. The calendar prep agent. Runs 30 minutes before each meeting. Pulls relevant project context, prior meeting notes, action items, and any open threads. Briefing arrives in your inbox before the meeting.
    4. The weekly status agent. Runs Friday 4 PM. For each business, summarizes what happened, what shipped, what’s at risk. Output: one digest per business plus a meta-digest across all of them.
    5. The follow-up watcher. Runs daily. Scans all open conversations, projects, and commitments. Flags anything that’s been waiting on you for more than 48 hours.
    6. The content production agent. Runs on schedule per business. Pulls from a content brief database, drafts the next piece, drops it in WordPress drafts (via integration) or a Notion review queue.
    7. The end-of-day capture agent. Runs at 6 PM. Prompts you for a quick voice note on what happened. Processes it into structured updates across the relevant business databases.

    What this stack costs

    Rough credit math at \$10/1000 (post-May 4):
    – Morning briefing: 30 days x ~15 credits = ~\$4.50/month
    – Intake triage: 100 triggers x ~5 credits = ~\$5/month
    – Calendar prep: 100 meetings x ~10 credits = ~\$10/month
    – Weekly status: 4 runs x ~50 credits = ~\$2/month
    – Follow-up watcher: 30 days x ~15 credits = ~\$4.50/month
    – Content production: 12 runs x ~80 credits = ~\$9.50/month
    – End-of-day capture: 30 days x ~10 credits = ~\$3/month
    Total: roughly \$38/month. Add Business plan seat fee. Total operating cost for the agent layer: well under what a part-time VA would charge.

    What this stack doesn’t do

    Things that stay manual:
    – Sales conversations and relationship work
    – Strategic decisions across businesses
    – Team conversations (even if “team” is contractors)
    – Anything client-facing where voice matters
    – Creative work where the doing is the point
    The agents handle the operational substrate. You handle the layer above it.

    How to start

    Don’t build all seven on day one. Build the morning briefing first. Live with it for two weeks. Tighten the prompt. Then build the next one. Sequential beats parallel.

    What to read next

    What Notion AI Agents Are, How Skills Work, Custom Agents vs Basic, ROI Math.

  • Auto Model Selection in Notion 3.2: Letting Notion Pick Claude, GPT, or Gemini For You

    Auto Model Selection in Notion 3.2: Letting Notion Pick Claude, GPT, or Gemini For You

    The 60-second version

    You don’t have to pick the model anymore. Notion 3.2 added auto-selection, which routes each request to the best-fit model from the available pool — currently including Claude Opus 4.7, GPT-5.2, and Gemini 3. Simple tasks (rewrites, summaries, quick drafts) go to faster models. Complex tasks (multi-step reasoning, long-context analysis, tool-heavy agent runs) go to more capable ones. You can override the selection per request, but the default behavior is “let Notion pick” — and for most workflows, that’s the right call.

    Why auto-selection matters

    Three reasons it’s a meaningful shift:
    1. You stop being a model-picker. Before auto-selection, getting good output required knowing which model handled which task best. That’s expert knowledge most users don’t have. Auto-selection internalizes that knowledge.
    2. Cost-performance balance happens automatically. Faster models are cheaper to run; capable models are more expensive. Notion’s auto-selection routes simple work to cheap models and reserves expensive models for tasks that need them. After May 4, when credits start metering Custom Agent work, this matters financially.
    3. Model diversity becomes a feature, not friction. Different models have different strengths. Claude is consistently strong on long-form writing and tool use. GPT is strong on broad reasoning. Gemini is strong on multimodal and certain analytical tasks. Auto-selection uses the right tool without forcing you to know which is which.

    When to override the auto-selection

    Three cases where manual model choice still wins:
    1. You’ve measured a specific preference. If you’ve tested the same task across all three models and found one consistently better for your use case, lock to that one. Auto-selection optimizes for the average user; you may not be the average user.
    2. You’re working in a domain with a clear model strength. Long-form editorial work where Claude’s prose quality is meaningfully better. Code work where GPT’s tool use feels more natural. Visual analysis where Gemini’s multimodal handles your case better.
    3. Reproducibility matters. Auto-selection means today’s request might use Claude and tomorrow’s might use GPT. If you need consistent voice or behavior across runs, lock the model.
    For everything else, auto-selection is fine. Stop optimizing the optimizer.

    What auto-selection isn’t

    It isn’t infinite model access. The pool is curated by Notion. You don’t get every model on the market. You get the ones Notion has integrated and validated for the platform.
    It also isn’t a replacement for model expertise if you’re a developer building on the API. When you build with Workers or skills via the API, you may want explicit model selection because reproducibility matters more there than in interactive use.

    How to verify auto-selection is working

    A 5-minute test:
    1. Open a page with substantive content (a project doc, an article, a meeting transcript)
    2. Run three different prompts: a quick rewrite, a complex synthesis, and a multi-step extraction
    3. Look at the output quality for each
    4. If all three feel right for the task, auto-selection is doing its job
    5. If any feel off — outputs that are too brief or too verbose, missing the task’s complexity — that’s where to consider manual override

    Why Claude Opus 4.7 in particular matters

    The Claude Opus 4.7 addition is worth noting separately. Anthropic’s latest uses fewer tokens (cheaper to run), makes 3x fewer tool errors (more reliable for agents that call Workers), and handles complex workflows better. For Notion specifically, that means agents that previously hit edge cases when chaining multiple skills or Workers now have a more reliable backbone.
    If you’re heavy into Custom Agents and Workers, Opus 4.7 in the rotation is the quiet upgrade that makes everything more dependable.

    What to read next

    Corpus follow-ups: Mobile AI in Notion (where auto-selection also runs), Custom Agents foundation piece (where model selection has cost implications), and the comparison articles (Notion AI vs ChatGPT, Claude Projects, Gemini for Workspaces).