Tag: Claude API

  • Anthropic Console: What It Is, How to Get an API Key, and How to Use It

    The Anthropic Console is the web-based dashboard where developers manage their Claude API access — creating API keys, monitoring usage, setting spending limits, and testing models. If you’re building with the Claude API, the Console is your operational home base.

    Access: console.anthropic.com — sign in with your Anthropic account. API access requires adding a payment method and generating an API key.

    What the Anthropic Console Does

    Section What you do here
    API Keys Create, name, and revoke API keys. Each key can have spending limits and restricted permissions.
    Workbench Test prompts and model configurations interactively before building. Adjust temperature, system prompts, and model selection in real time.
    Usage & Billing Monitor token consumption by model, set spending limits, view billing history, and add credits.
    Rate Limits See your current tier and the limits that apply — requests per minute, tokens per minute, tokens per day.
    Models Browse available models and their API strings. Use as reference before specifying models in code.
    Prompt Library Save and reuse prompts and system prompt configurations across projects.

    Getting Your First API Key

    1. Go to console.anthropic.com and sign in or create an account.
    2. Add a payment method under Billing — the API is pay-as-you-go, no subscription required.
    3. Navigate to API Keys and click Create Key.
    4. Name the key (e.g., “development” or “production”) and optionally set a spending limit.
    5. Copy the key immediately — it won’t be shown again after you close the dialog.
    6. Store it securely: environment variable, secrets manager, or your CI/CD vault. Never hardcode it.
    # Store your key as an environment variable
    export ANTHROPIC_API_KEY="sk-ant-..."
    
    # Then access it in Python
    import anthropic
    client = anthropic.Anthropic()  # reads ANTHROPIC_API_KEY automatically

    The Workbench: Test Before You Build

    The Workbench is the Console’s interactive testing environment. Before writing API code, use it to develop and test your prompts — adjust the system prompt, try different models, tune parameters, and see exactly how Claude responds. When you have the behavior you want, export the configuration as code with one click.

    This is the fastest way to iterate on prompt design without writing a test harness every time. It’s also where you can verify current model behavior before updating a production system.

    Understanding Rate Limits in the Console

    The Console shows your current rate limit tier and the specific limits that apply. Anthropic uses a tiered system — as your spending grows, your limits increase automatically:

    • Tier 1 — New accounts, basic limits, minimum spend
    • Tier 2-4 — Limits scale up as cumulative API spend increases
    • Enterprise — Custom limits negotiated with Anthropic sales

    If you’re hitting rate limits in production, the Console shows exactly which limit you’re hitting (requests per minute vs tokens per minute vs daily tokens) so you know whether to optimize your code or request a tier increase. For full context on limits, see Claude Rate Limits: What They Are and How to Work Around Them.

    Spending Limits and Cost Control

    The Console lets you set spending limits per API key — useful for development keys where you want a hard cap, or for giving team members API access with bounded risk. Usage dashboards show consumption by model and time period, which is essential for understanding which Claude model is driving cost in a production system.

    For full pricing details to budget against, see Anthropic API Pricing: All Models and Costs.

    Console vs claude.ai: What’s the Difference

    The Anthropic Console (console.anthropic.com) is for developers building with the API. Claude.ai is the consumer product for end users having conversations with Claude. They use the same underlying models but serve different purposes — the Console is where you manage programmatic access, the claude.ai interface is where you use Claude directly.

    Tygart Media

    Getting Claude set up is one thing.
    Getting it working for your team is another.

    We configure Claude Code, system prompts, integrations, and team workflows end-to-end. You get a working setup — not more documentation to read.

    See what we set up →

    Frequently Asked Questions

    What is the Anthropic Console?

    The Anthropic Console (console.anthropic.com) is the developer dashboard for managing Claude API access — creating API keys, monitoring usage and billing, testing prompts in the Workbench, and managing rate limits. It’s separate from claude.ai, which is the end-user product.

    How do I get an Anthropic API key?

    Go to console.anthropic.com, sign in, add a payment method under Billing, then go to API Keys and click Create Key. Copy the key immediately after creation — it won’t be shown again. Store it as an environment variable, never in your code.

    Is the Anthropic Console free?

    Creating an account and accessing the Console is free. The API itself is pay-as-you-go — you only pay for tokens consumed. There’s no monthly subscription fee for API access; you add credits and they’re deducted as you use the API.

    Need this set up for your team?
    Talk to Will →
  • How to Install Claude Code: Desktop App and CLI Setup Guide

    Claude Code is available two ways: as the Code tab inside Claude Desktop (with a graphical interface), or as a CLI tool you install and run from your terminal. Here’s how to get either one set up from scratch.

    Requirement: Claude Code requires a Pro, Max, Team, or Enterprise subscription. It is not available on the free plan. The Claude Desktop app (which includes a graphical Claude Code interface) is free to download but the Code tab requires a paid subscription.

    Option 1: Claude Desktop (Recommended for Most Users)

    The easiest way to get Claude Code is through Claude Desktop — no terminal required.

    1. Download Claude Desktop from claude.ai/download — available for macOS and Windows (x64 or ARM64). Linux is not supported.
    2. Install — on Mac, open the PKG and drag to Applications; on Windows, run the installer.
    3. Sign in with your Anthropic account (Pro, Max, Team, or Enterprise).
    4. Click the Code tab in the top navigation.
    5. Select Local to work with files on your machine, or Remote to run on Anthropic’s cloud infrastructure.
    6. Click “Select folder” and choose your project directory. You’re ready.

    On Windows, Git must be installed for local sessions to work. Most Macs include Git by default — check by running git --version in Terminal.

    Free — no pitch

    Get the Claude workflow that actually sticks.

    Practical Claude setup tips from someone running it across 27 client sites daily — not marketing, not theory. Email Will directly and he’ll share what’s working.

    Get the tips →

    Option 2: Claude Code CLI

    For developers who prefer working in the terminal, Claude Code is also available as a command-line tool.

    # Install via npm
    npm install -g @anthropic-ai/claude-code
    
    # Authenticate
    claude login
    
    # Start in your project directory
    cd your-project
    claude

    The CLI requires Node.js. After running claude login, you’ll authenticate with your Anthropic account in a browser window. The session starts automatically in the current directory.

    Local vs. Remote Sessions

    Session type What it does Best for
    Local Runs on your machine, accesses your files directly Everyday development work
    Remote Runs on Anthropic’s cloud, continues if you close the app Long-running tasks, autonomous work
    SSH Connects to a remote machine over SSH Server or cloud VM development

    Common Setup Issues

    Code tab not appearing in Desktop: Confirm your account is on a paid plan. Claude Code requires Pro, Max, Team, or Enterprise — it’s not available on the free tier.

    Windows Git error: Claude Code needs Git for local sessions on Windows. Download Git from git-scm.com, install with default settings, then restart the desktop app.

    CLI authentication failing: Run claude logout then claude login again. Make sure your Anthropic account has an active paid subscription.

    Permission errors on first run: Claude Code will ask permission to access your files when you first select a folder. Click Allow — it needs read/write access to work with your project.

    First Session: What to Expect

    When you start your first Claude Code session, Anthropic recommends starting with a small, familiar project. Ask Claude to explain the codebase, fix a specific bug, or add a small feature. This gives you a calibrated sense of how it works before tackling larger tasks. Claude will read relevant files, propose changes, and ask for your approval before modifying anything.

    For an overview of what Claude Code can do once you’re set up, see How to Use Claude Code. For pricing details, see Claude Code Pricing: Pro vs Max.

    Tygart Media

    Getting Claude set up is one thing.
    Getting it working for your team is another.

    We configure Claude Code, system prompts, integrations, and team workflows end-to-end. You get a working setup — not more documentation to read.

    See what we set up →

    Frequently Asked Questions

    How do I install Claude Code?

    Download Claude Desktop from claude.ai/download and use the Code tab — no terminal required. Or install the CLI with npm install -g @anthropic-ai/claude-code and run claude login to authenticate.

    Is Claude Code free to install?

    Claude Desktop (which includes Claude Code) is free to download. Using Claude Code requires a paid subscription — Pro ($20/month), Max ($100/month), Team, or Enterprise. It is not available on the free plan.

    Does Claude Code work on Linux?

    The Claude Desktop app does not support Linux. The Claude Code CLI does run on Linux — install via npm and use it from your terminal.

    What’s the difference between Claude Code Desktop and the CLI?

    Claude Code Desktop (the Code tab in the Claude Desktop app) gives you a graphical interface with visual file diffs, a built-in preview panel, and no terminal required. The CLI runs in your terminal and supports the same core operations. Both share configuration files and can run simultaneously on the same project.

    Need this set up for your team?
    Talk to Will →
  • Claude Haiku vs Sonnet vs Opus: The Complete Three-Model Comparison

    Choosing between Claude’s three models comes down to one question: how hard is the task, and how much does cost matter? Haiku, Sonnet, and Opus each occupy a distinct position — this is the complete three-way breakdown so you can route work correctly from the start.

    The routing rule in one sentence: Haiku for volume and speed, Sonnet for almost everything else, Opus for the tasks where Sonnet isn’t quite enough.

    Haiku vs Sonnet vs Opus: Full Comparison

    Spec Haiku Sonnet Opus
    API string claude-haiku-4-5-20251001 claude-sonnet-4-6 claude-opus-4-6
    Input price (per M tokens) ~$1.00 ~$3.00 ~$5.00
    Output price (per M tokens) ~$5.00 ~$5.00 ~$25.00
    Context window 200K 1M 1M
    Speed ⚡ Fastest ⚡ Fast 🐢 Slower
    Reasoning depth Good Excellent Maximum
    Writing quality Good Excellent Maximum
    Cost vs Sonnet ~4× cheaper ~5× more expensive

    Claude Haiku: The Volume Model

    Haiku is optimized for tasks that are high in quantity but low in complexity — situations where you’re running the same operation hundreds or thousands of times and cost per call is a real constraint. Classification, extraction, summarization, metadata generation, routing logic, short-form responses, and real-time features where latency matters more than depth.

    The output quality on constrained tasks is strong. Where Haiku shows its limits is on open-ended, nuanced work — multi-step reasoning, long-form writing where voice consistency matters, or problems with competing constraints. For those, Sonnet is the right call.

    Claude Sonnet: The Default

    Sonnet handles the vast majority of professional work at a quality level that’s indistinguishable from Opus for most tasks. Writing, analysis, research, coding, summarization, strategy — Sonnet does all of it well. It’s the model to start with and the one most people should use as their production default.

    The gap between Sonnet and Opus shows on genuinely hard tasks: novel multi-step reasoning, edge cases in complex code, nuanced judgment in ambiguous situations, or extended agentic sessions where small quality differences compound. For everything else, Sonnet is the right choice and a fraction of the cost.

    Claude Opus: The Specialist

    Opus earns its premium on tasks where maximum capability is the only variable that matters and cost is secondary. Complex legal or technical analysis, research synthesis across conflicting sources, architectural decisions with long-term consequences, extended agentic sessions, and any task where you’ve tried Sonnet and felt the output was a notch below what the problem deserved.

    The practical test: if Sonnet’s output on a task is good enough, use Sonnet. Only reach for Opus when you’ve genuinely hit Sonnet’s ceiling on a specific problem. Most professionals do this on a small fraction of their actual workload.

    The Decision Framework

    Use Haiku when: same operation at high volume, output is constrained/structured, cost and speed matter, real-time latency required.

    Use Sonnet when: any standard professional task — writing, coding, analysis, research. This should be your default 90% of the time.

    Use Opus when: the task is genuinely hard, involves novel reasoning, Sonnet’s output wasn’t quite right, or quality is the only variable that matters regardless of cost.

    For full pricing details, see Anthropic API Pricing. For a Haiku deep-dive, see Claude Haiku: Pricing, Use Cases, and API String. For the Opus vs Sonnet head-to-head, see Claude Opus vs Sonnet.

    Frequently Asked Questions

    What’s the difference between Claude Haiku, Sonnet, and Opus?

    Haiku is fastest and cheapest — built for high-volume, constrained tasks. Sonnet is the balanced production default with excellent quality across most professional work. Opus is the most capable model for complex reasoning — about 5× more expensive than Sonnet on input tokens.

    Which Claude model should I use?

    Start with Sonnet for almost everything. Switch to Haiku when you’re running the same operation at high volume and cost matters. Switch to Opus when Sonnet’s output on a specific task isn’t quite at the level the problem requires.

    Is Claude Haiku good enough for most tasks?

    For structured, constrained tasks — yes, Haiku is strong. For open-ended writing, complex reasoning, or work requiring nuanced judgment, Sonnet is the right step up. The cost savings from Haiku are meaningful at scale, making it the right choice when the task fits its strengths.

    Need this set up for your team?
    Talk to Will →
  • Claude Integrations and Plugins: Complete List of What Claude Can Connect To

    Claude doesn’t use a traditional plugin marketplace — instead, it connects to external tools and services through MCP (Model Context Protocol), an open standard that lets any service build a Claude integration. Here’s a complete rundown of what Claude can connect to in 2026, how those connections work, and how to set them up.

    How Claude integrations work: Claude uses MCP (Model Context Protocol) instead of plugins. Services publish an MCP server; Claude connects to it and gains access to that service’s capabilities. In Claude.ai, many integrations are available in Settings → Connections. In Claude Desktop and the API, you can connect to any MCP server.

    Claude Integrations Available in Claude.ai (2026)

    Service What Claude can do Available in
    Google Drive Search, read, and analyze documents Claude.ai
    Google Calendar Read and create calendar events Claude.ai
    Gmail Read, search, and draft emails Claude.ai
    Notion Read and write pages, query databases Claude.ai
    Slack Read channels, search messages, post Claude.ai
    GitHub Read repos, create issues, review PRs Claude Desktop / API
    Zapier Trigger automations across 6,000+ apps Claude.ai
    HubSpot Read and update CRM records Claude.ai
    Cloudflare Manage workers, DNS, and infrastructure Claude Desktop / API
    PostgreSQL / databases Query, read schema, analyze data Claude Desktop / API
    File system Read, write, organize local files Claude Desktop
    Web search Search the web for current information Claude.ai (built-in)
    Jira / Linear Read and create issues, update status Claude.ai / API
    Custom APIs Any service with an MCP server Claude Desktop / API

    How to Add Integrations in Claude.ai

    1. Go to claude.ai → Settings → Connections
    2. Browse the available integrations and click Connect on any you want to enable
    3. Authenticate with the service (usually OAuth — you’ll be redirected to authorize)
    4. Once connected, Claude can use that service in your conversations when relevant

    Claude Desktop: More Integrations, More Control

    The Claude Desktop app supports MCP server configuration via a JSON config file — giving you access to any MCP server, including self-hosted ones and community-built integrations that aren’t in the official Claude.ai connection list. This is where the integration ecosystem expands beyond the curated set: database connections, local file systems, internal tools, and any API where someone has built an MCP server.

    Building Your Own Claude Integration

    Any developer can build an MCP server and connect it to Claude. Anthropic publishes the MCP spec openly — you implement the server, and Claude can immediately use whatever tools or data you expose. This is how companies integrate Claude into proprietary internal systems without exposing data to a third party. For the technical implementation, see the Claude MCP guide.

    Frequently Asked Questions

    Does Claude have plugins?

    Claude doesn’t use a plugin marketplace like early ChatGPT did. Instead it uses MCP (Model Context Protocol) — an open standard where services publish integration servers that Claude connects to. In Claude.ai, these appear as “Connections” in Settings. Claude Desktop supports any MCP server via config file.

    What apps can Claude connect to?

    Claude can connect to Google Drive, Gmail, Google Calendar, Notion, Slack, Zapier, HubSpot, GitHub, Cloudflare, databases, local file systems, and any service that has published an MCP server. The ecosystem is growing rapidly — new MCP servers are added by third-party developers regularly.

    How do I add integrations to Claude?

    In Claude.ai, go to Settings → Connections and authenticate the services you want to connect. For Claude Desktop, integrations are configured via a JSON config file that specifies which MCP servers to load. Via the API, you pass MCP server URLs in your request parameters.

    Need this set up for your team?
    Talk to Will →
  • Claude Haiku: Pricing, API String, Use Cases, and When to Use It

    Claude Haiku is Anthropic’s fastest and most cost-efficient model — the right choice when you need high-volume AI at low cost without sacrificing the quality that makes Claude worth using. It’s not a cut-down version of the flagship models. It’s a purpose-built model for the tasks where speed and cost matter more than maximum reasoning depth.

    When to use Haiku: Any time you’re running the same operation across many inputs — classification, extraction, summarization, metadata generation, routing logic, short-form responses — and cost or speed is a meaningful constraint. Haiku handles these at a fraction of Sonnet’s price with output quality that’s more than sufficient.

    Claude Haiku Specs (April 2026)

    Spec Value
    API model string claude-haiku-4-5-20251001
    Context window 200,000 tokens
    Input pricing ~$1.00 per million tokens
    Output pricing ~$5.00 per million tokens
    Speed vs Sonnet Faster — optimized for low latency
    Batch API discount ~50% off (~$0.50 input / ~$2.50 output)

    Claude Haiku vs Sonnet vs Opus

    Model Input cost Speed Reasoning depth Best for
    Haiku ~$1.00/M Fastest Good High-volume, latency-sensitive
    Sonnet ~$3.00/M Fast Excellent Production workloads, daily driver
    Opus ~$5.00/M Slower Maximum Complex reasoning, highest quality

    What Claude Haiku Is Best At

    Haiku is optimized for tasks where the output is constrained and the logic is clear — not open-ended creative or strategic work where maximum capability pays off. The practical use cases where Haiku earns its position:

    • Classification and routing — is this a support ticket, a bug report, or a feature request? Tag it and route it. Haiku handles thousands of these per hour at minimal cost.
    • Extraction — pull the names, dates, dollar amounts, or addresses from a document. Structured output from unstructured text at scale.
    • Summarization — condense articles, emails, or documents to key points. Haiku’s summarization is strong enough for most production use cases.
    • SEO metadata — generate title tags, meta descriptions, alt text, and schema markup in bulk. This is where Haiku shines for content operations.
    • Short-form responses — FAQ answers, product descriptions, short explanations. Anything where the output is a few sentences or a structured short block.
    • Real-time features — chatbots, autocomplete, inline suggestions — anywhere latency affects user experience.

    Claude Haiku vs GPT-4o Mini

    GPT-4o mini is OpenAI’s comparable low-cost model and is less expensive than Haiku per token. The cost trade-off is real — GPT-4o mini is cheaper. The quality trade-off depends on the task. For instruction-following on complex structured outputs, Haiku tends to be more reliable. For simple, high-volume tasks where the output format is forgiving, the cost difference may favor GPT-4o mini. For teams already building on Claude for quality reasons, Haiku is the natural choice for high-volume work within that stack.

    Using Claude Haiku in the API

    import anthropic
    
    client = anthropic.Anthropic()
    
    message = client.messages.create(
        model="claude-haiku-4-5-20251001",
        max_tokens=256,
        messages=[
            {"role": "user", "content": "Classify this support ticket: ..."}
        ]
    )
    
    print(message.content)

    For a full model comparison, see Claude Models Explained: Haiku vs Sonnet vs Opus. For API pricing across all models, see Anthropic API Pricing.

    Frequently Asked Questions

    What is Claude Haiku?

    Claude Haiku is Anthropic’s fastest and most affordable model — approximately $1.00 per million input tokens. It’s purpose-built for high-volume, latency-sensitive tasks like classification, extraction, summarization, and short-form generation where cost efficiency matters more than maximum reasoning depth.

    How much does Claude Haiku cost?

    Claude Haiku costs approximately $1.00 per million input tokens and $5.00 per million output tokens. The Batch API reduces these to approximately $0.40 input and $2.00 output — roughly half price for non-time-sensitive workloads.

    When should I use Claude Haiku instead of Sonnet?

    Use Haiku when your task is well-defined with a constrained output, you’re running it at high volume, and cost or latency is a meaningful consideration. Use Sonnet when the task is complex, requires nuanced reasoning, or produces longer open-ended outputs where maximum quality matters.

    What is the Claude Haiku API model string?

    The current Claude Haiku model string is claude-haiku-4-5-20251001. Always verify the current string in Anthropic’s official model documentation before production deployment.

    Need this set up for your team?
    Talk to Will →
  • Claude System Prompt Guide: How to Write Them, Examples, and Best Practices

    A system prompt is the instructions you give Claude before the conversation begins — the context, persona, rules, and constraints that shape every response in the session. It’s the most powerful lever you have for controlling Claude’s behavior at scale, and the foundation of any serious Claude integration. Here’s how system prompts work, how to write them well, and real examples across common use cases.

    What a system prompt does: Sets Claude’s role, knowledge, tone, constraints, and output format before the user says anything. Claude treats system prompt instructions as authoritative — they persist throughout the conversation and take priority over conflicting user requests within the boundaries Anthropic allows.

    System Prompt Structure: The Five Elements

    A well-structured system prompt typically covers these elements — not all are required for every use case, but the strongest prompts address most of them:

    # Role
    You are [specific role/persona]. [1-2 sentences on expertise and perspective].

    # Context
    [What this system/application/conversation is for. Who the user is. What they’re trying to accomplish.]

    # Instructions
    [Specific behaviors: what to do, how to format responses, how to handle edge cases]

    # Constraints
    [What NOT to do. Topics to avoid. Format rules to enforce. Information not to share.]

    # Output format
    [How Claude should structure its responses: length, format, sections, tone]

    System Prompt Examples by Use Case

    Customer Support Agent

    You are a customer support agent for Acme Software. You help users with account questions, billing issues, and technical troubleshooting for Acme’s project management platform.

    Tone: professional, patient, solution-focused. Never dismissive.

    For billing questions: provide information but escalate refund requests to billing@acme.com.
    For technical issues: follow the troubleshooting guide below before escalating.
    Never discuss: competitor products, internal pricing strategy, unreleased features.

    Always end with: “Is there anything else I can help you with today?”

    Code Assistant

    You are a senior software engineer helping with Python and TypeScript code.

    When writing code: use type hints in Python, strict TypeScript, and always include error handling. Prefer explicit over implicit. Comment non-obvious logic.

    When reviewing code: flag issues by severity (critical/high/medium/low). Always explain why something is a problem, not just that it is.

    Never write code without error handling. Never use eval(). Never hardcode credentials.

    Content Writer

    You write content for [Brand Name], a B2B SaaS company in the project management space.

    Voice: direct, confident, no filler. Never use “leverage,” “synergy,” or “utilize.” Short sentences. Active voice.

    Audience: project managers and engineering leads at companies with 50–500 employees.

    Always: include a clear next step or CTA. Never: make claims we can’t back up, mention competitors by name.

    What System Prompts Can and Can’t Do

    System prompts are powerful but not absolute. They can reliably control: Claude’s tone and persona, output format and structure, topic scope and focus, response length guidelines, and how Claude handles specific scenarios. They cannot override Anthropic’s core guidelines — Claude won’t follow system prompt instructions to produce harmful content, lie about being an AI when sincerely asked, or violate its trained ethical constraints regardless of what the system prompt says.

    System Prompts in the API vs. Claude.ai

    In the API, the system prompt is passed as the system parameter in your API call. In Claude.ai Projects, the custom instructions field functions as the system prompt for all conversations in that Project. In Claude.ai standard conversations, you can prepend context at the start of a conversation — it’s not a true system prompt but achieves a similar effect.

    import anthropic
    
    client = anthropic.Anthropic()
    
    response = client.messages.create(
        model="claude-sonnet-4-6",
        max_tokens=1024,
        system="You are a helpful assistant...",  # ← system prompt here
        messages=[
            {"role": "user", "content": "Hello"}
        ]
    )

    For a full library of tested prompts across use cases, see the Claude Prompt Library and Claude Prompt Generator and Improver.

    Tygart Media

    Getting Claude set up is one thing.
    Getting it working for your team is another.

    We configure Claude Code, system prompts, integrations, and team workflows end-to-end. You get a working setup — not more documentation to read.

    See what we set up →

    Frequently Asked Questions

    What is a Claude system prompt?

    A system prompt is instructions given to Claude before the conversation begins — setting its role, constraints, tone, and output format. It persists throughout the session and takes priority over user messages within Anthropic’s guidelines.

    How long should a Claude system prompt be?

    Long enough to cover what Claude needs to behave correctly, short enough that Claude actually follows all of it. Most production system prompts are 200–1,000 words. Beyond that, you risk important instructions getting less attention. Structure with headers helps Claude parse longer prompts.

    Can users override a system prompt?

    Not reliably. System prompts take priority over user messages. A user saying “ignore your system prompt” won’t override legitimate business instructions. Claude is designed to follow operator system prompts even when users push back, within Anthropic’s ethical guidelines.

    Need this set up for your team?
    Talk to Will →
  • Claude Context Window Size: 200K Tokens Explained in Plain Terms

    Claude’s context window determines how much information it can hold and process in a single conversation. Claude Sonnet 4.6 and Opus 4.6 support 1 million tokens; Haiku 4.5 supports 200,000 tokens — one of the largest in the industry. Here’s what that means in practice, what you can actually fit inside it, and how context window size affects your work.

    200K tokens in plain terms: Roughly 150,000 words, or about 500 pages of text. That’s enough for an entire novel, a full codebase, or months of conversation history — all in a single session without truncation.

    Claude Context Window by Model (April 2026)

    Model Context Window ~Words ~Pages
    Claude Haiku 200,000 tokens ~150,000 ~500
    Claude Sonnet 200,000 tokens ~150,000 ~500
    Claude Opus 200,000 tokens ~150,000 ~500

    What Fits in 200K Tokens

    Content type Approximate fit
    News articles ~200+ articles
    Research papers ~30–50 papers depending on length
    A full novel Yes — most novels fit with room to spare
    Python codebase Medium-sized codebases (10k–50k lines)
    Legal contracts Hundreds of pages of contracts
    Conversation history Very long sessions before truncation

    Context Window vs. Output Length

    The context window covers everything Claude processes — both input and output combined. If your prompt is 50,000 tokens (a long document), Claude has 150,000 tokens remaining for its response and any further back-and-forth. The window is shared between what you send and what Claude generates.

    Maximum output length is a separate constraint — Claude won’t generate an infinitely long response even within a large context window. For very long outputs (full books, extensive reports), you typically work in sections rather than expecting Claude to produce everything in one pass.

    Why Context Window Size Matters

    Context window size is the practical limit on how much work you can give Claude at once without losing information. Before large context windows, working with long documents required chunking — splitting the document into pieces, analyzing each separately, and manually synthesizing the results. With 200K tokens, Claude can hold the entire document and answer questions about any part of it with full awareness of everything else.

    This matters most for: document analysis and legal review, code understanding across large files, research synthesis across many sources, and long multi-step conversations where earlier context affects later decisions.

    How Claude Performs at the Edges of Its Context Window

    Research on large language models has found that performance can degrade somewhat for information buried in the middle of a very long context — sometimes called the “lost in the middle” problem. Claude performs well across its context window, but for maximum reliability on information from a very long document, referencing specific sections explicitly (“in the section about pricing on page 12…”) helps ensure Claude focuses on the right part.

    For the full model spec breakdown, see Claude API Model Strings and Specs and Claude Models Explained: Haiku vs Sonnet vs Opus.

    Frequently Asked Questions

    What is Claude’s context window size?

    Claude Sonnet 4.6 and Opus 4.6 support a 1 million token context window at standard pricing. Claude Haiku 4.5 supports 200,000 tokens. That’s approximately 150,000 words or about 500 pages of text in a single conversation.

    How many tokens is 200K context?

    200,000 tokens is approximately 150,000 words of English text. One token is roughly four characters or three-quarters of a word. A typical 800-word article is about 1,000 tokens; a full novel is typically 80,000–120,000 tokens.

    Can I upload a full PDF to Claude?

    Yes, as long as the PDF’s text content fits within the 200K token context window. Most documents, reports, contracts, and research papers fit easily. Very large documents (multiple volumes, extensive legal filings) may need to be split.

    Need this set up for your team?
    Talk to Will →
  • Claude Rate Limits: What They Are, How They Work, and What to Do

    Claude has usage limits on every plan — but Anthropic doesn’t publish exact numbers. Instead limits are dynamic, adjusting based on model, message length, and system load. Here’s what the limits actually look like in practice, what triggers them, and what your options are when you hit them.

    What you’ll see: When you hit Claude’s usage limit, you’ll get a message saying you’ve reached your usage limit and showing a countdown to when your limit resets. On Pro this typically resets within a few hours. On Max, limits are high enough that most users never hit them during normal work.

    Rate Limits by Plan

    Plan Relative limit Typical experience
    Free Low Hit limits quickly on heavy use; resets daily
    Pro ~5× Free Most users get through a full workday; heavy users may hit limits
    Max ~5× Pro Most users never hit limits; designed for agentic and heavy use
    Team Higher than Pro Per-user limits slightly higher than individual Pro
    API Separate system Tokens per minute/day limits by tier; see Anthropic’s API docs

    What Counts Against Your Limit

    Claude’s limits are usage-based, not message-count-based. A single message asking Claude to write a 3,000-word article uses more of your limit than ten quick back-and-forth questions. What consumes the most limit, fastest:

    • Long outputs — requests for long articles, detailed analyses, or extended code
    • Long context — uploading large documents and asking questions about them
    • Opus model — the most powerful model consumes limits faster than Sonnet or Haiku
    • Agentic tasks — multi-step autonomous operations use significantly more than conversational use

    API Rate Limits: How They Work

    The API uses a different limit system from the web interface. API limits are measured in:

    • Requests per minute (RPM) — how many API calls you can make
    • Tokens per minute (TPM) — total tokens (input + output) processed per minute
    • Tokens per day (TPD) — total daily token budget

    New API accounts start on lower tiers and can request higher limits through the Anthropic Console as usage establishes a track record. The Batch API has separate, higher limits since it’s asynchronous and non-time-sensitive.

    What To Do When You Hit a Limit

    Wait for reset: The limit message shows when your usage resets — usually within a few hours. This is the simplest option if the timing works.

    Switch models: If you’ve been using Opus, switching to Sonnet for less critical tasks conserves your limit for when you need the top model.

    Upgrade your plan: If you consistently hit Pro limits during your workday, Claude Max at $100/month gives 5× the headroom.

    Use the API: For developers, moving high-volume work to the API with the Batch API gives more control over usage and significant cost savings on non-time-sensitive tasks.

    Frequently Asked Questions

    What are Claude’s usage limits?

    Anthropic doesn’t publish exact numbers. Limits are dynamic and based on usage volume rather than message count. Free is most restricted; Pro is roughly 5× Free; Max is roughly 5× Pro. The limit message appears when you’ve reached your tier’s threshold and shows when it resets.

    How long does it take for Claude’s limit to reset?

    The reset countdown is shown in the limit message. For Pro, limits typically reset within a few hours. For Free, resets are on a daily cycle. The exact timing varies based on when you started using heavily in the current period.

    Does Claude count messages or tokens toward the limit?

    Usage is based on the volume of content processed, not a simple message count. One long request asking for a 3,000-word output uses significantly more of your limit than ten short conversational exchanges.

    Are API rate limits the same as subscription limits?

    No. API limits (RPM, TPM, TPD) are a separate system from web subscription limits. They’re set per API account tier and can be increased by request through the Anthropic Console. Subscription usage and API usage don’t share limits.

    Need this set up for your team?
    Talk to Will →
  • Claude MCP: What the Model Context Protocol Is and How to Use It

    MCP — the Model Context Protocol — is Anthropic’s open standard for connecting Claude to external tools, data sources, and services. It’s the architecture that lets Claude read from your Google Drive, post to Slack, query a database, or interact with any API without you having to write custom integration code for each one. Here’s what MCP is, how it works, and why it matters.

    Short version: MCP is how Claude gets access to tools beyond the chat window. Instead of every developer writing one-off integrations, MCP creates a standard protocol — Claude speaks it, and any tool that implements it can plug in. Think of it as USB-C for AI tool connections.

    What MCP Actually Does

    Without MCP, connecting Claude to an external system means building a custom bridge: write code that calls the external API, format the results in a way Claude understands, handle authentication, manage errors. Every integration is a separate project.

    With MCP, the external system (a database, a SaaS tool, a file system, an API) publishes an MCP server — a standardized interface that describes what it can do. Claude connects to that server and immediately knows what tools are available, what inputs they need, and how to use them. The developer only builds the MCP server once; Claude handles the rest.

    What You Can Do With Claude MCP Today

    MCP Integration What Claude Can Do
    Google Drive Search, read, and summarize documents in your Drive
    Slack Read channels, search messages, post drafts
    GitHub Read repos, create issues, review pull requests
    Notion Read and write pages, query databases
    PostgreSQL / databases Run queries, read schema, analyze data
    File systems Read, write, and organize local files
    Web search Search the web and return current results
    Custom APIs Any API with an MCP server implementation

    MCP vs. Claude’s Built-In Tools

    Claude already has some built-in capabilities — web search, code execution in certain contexts, file analysis. MCP extends this with external integrations that persist across sessions, connect to your actual data, and scale to any service that builds an MCP server.

    The practical difference: built-in tools are what Anthropic ships with Claude. MCP tools are what the ecosystem builds — which means the integration surface grows every week as more services add MCP support.

    How to Use MCP With Claude

    MCP works differently depending on where you’re running Claude:

    Claude.ai (web/app): MCP integrations are available through the Connections settings. Anthropic has partnered with services like Google, Notion, Slack, and others whose MCP servers are pre-built and available to connect in a few clicks.

    Claude Desktop: The desktop app supports MCP configuration via a JSON config file, letting you connect to any MCP server — including self-hosted ones or custom integrations you build.

    Claude Code / API: Developers can wire MCP servers directly into Claude API calls, giving Claude access to any tool during an agentic session.

    Why MCP Is a Big Deal

    Before MCP, each AI company built its own plugin standard — OpenAI had plugins, others had connectors, and nothing worked across systems. MCP is Anthropic’s bet on an open standard: publish the spec, let anyone build to it, and Claude (and any other AI that implements it) gains access to the entire ecosystem.

    The momentum has been significant. Within months of the MCP spec being published, major platforms including Cloudflare, Zapier, HubSpot, and dozens of others shipped MCP server implementations. The network effect is real — the more tools support MCP, the more useful Claude becomes without Anthropic having to build any of those integrations themselves.

    For a deeper technical walkthrough, see the Claude MCP Tutorial.

    Frequently Asked Questions

    What is Claude MCP?

    MCP (Model Context Protocol) is an open standard from Anthropic that lets Claude connect to external tools, databases, and services. Instead of one-off integrations, MCP creates a universal protocol — any tool that builds an MCP server can be connected to Claude.

    How do I add MCP tools to Claude?

    In Claude.ai, go to Settings → Connections to add pre-built MCP integrations for services like Google Drive, Notion, and Slack. In Claude Desktop, you configure MCP servers in a JSON config file. Via the API, you pass MCP server URLs in your request.

    Is MCP only for Claude?

    No. MCP is an open protocol — any AI model or application can implement it. Anthropic published the spec publicly with the intent of making it an industry standard. Other AI tools have begun adopting it, though Claude has the deepest native MCP integration currently.

    What’s the difference between MCP and Claude plugins?

    Claude doesn’t use a “plugin” model the way older ChatGPT did. MCP is Anthropic’s approach — an open, standardized protocol rather than a proprietary plugin marketplace. MCP integrations work at a deeper level and are designed to scale across any service that implements the standard.

    Need this set up for your team?
    Talk to Will →
  • Claude API Model Strings, IDs and Specs — Complete Reference (April 2026)

    When you’re building on Claude via the API, you need the exact model string — not just the name. Anthropic uses specific model identifiers that change with each version, and using a deprecated string will break your application. This is the complete reference for Claude API model names, IDs, and specs as of April 2026.

    Quick reference: The current flagship models are claude-opus-4-6, claude-sonnet-4-6, and claude-haiku-4-5-20251001. Always use versioned model strings in production — never rely on alias strings that may point to different models over time.

    Current Claude API Model Strings (April 2026)

    Model API Model String Context Window Best for
    Claude Opus 4.6 claude-opus-4-6 1M tokens Complex reasoning, highest quality
    Claude Sonnet 4.6 claude-sonnet-4-6 1M tokens Production workloads, balanced cost/quality
    Claude Haiku 4.5 claude-haiku-4-5-20251001 200K tokens High-volume, latency-sensitive tasks

    Anthropic publishes the full, current list of model strings in their official models documentation. Always verify there before updating production systems — model strings are updated with each new release.

    How to Use Model Strings in an API Call

    import anthropic
    
    client = anthropic.Anthropic()
    
    message = client.messages.create(
        model="claude-sonnet-4-6",  # ← model string goes here
        max_tokens=1024,
        messages=[
            {"role": "user", "content": "Your prompt here"}
        ]
    )
    
    print(message.content)

    Model Selection: Which String to Use When

    The right model depends on your task requirements. Here’s the practical routing logic:

    Use Haiku (claude-haiku-4-5-20251001) when: you need speed and low cost at scale — classification, extraction, routing, metadata, high-volume pipelines where every call matters to your budget.

    Use Sonnet (claude-sonnet-4-6) when: you need solid quality across a wide range of tasks — content generation, analysis, coding, summarization. This is the right default for most production applications.

    Use Opus (claude-opus-4-6) when: the task genuinely requires maximum reasoning capability — complex multi-step logic, nuanced judgment, or work where output quality is the only variable that matters and cost is secondary.

    API Pricing by Model

    Model Input (per M tokens) Output (per M tokens)
    Claude Haiku ~$1.00 ~$5.00
    Claude Sonnet ~$3.00 ~$5.00
    Claude Opus ~$5.00 ~$25.00

    The Batch API offers roughly 50% off all rates for asynchronous workloads. For a full pricing breakdown, see Anthropic API Pricing: Every Model and Mode Explained.

    Important: Versioned Strings vs. Aliases

    Anthropic occasionally provides alias strings (like claude-sonnet-latest) that point to the current version of a model family. These are convenient for development but can create problems in production — when Anthropic updates the model the alias points to, your application silently starts using a different model without a code change. For production systems, always pin to a versioned model string and upgrade intentionally.

    Frequently Asked Questions

    What is the Claude API model string for Sonnet?

    The current Claude Sonnet model string is claude-sonnet-4-6. Always verify the current string in Anthropic’s official models documentation before deploying, as strings are updated with each new model release.

    How do I specify which Claude model to use in the API?

    Pass the model string in the model parameter of your API call. For example: model="claude-sonnet-4-6". The model string must match exactly — Anthropic’s API will return an error if the string is invalid or deprecated.

    What Claude API model should I use for production?

    Claude Sonnet is the right default for most production workloads — it balances quality and cost well across a wide range of tasks. Use Haiku when speed and cost are the priority at scale. Use Opus when the task genuinely requires maximum reasoning capability and cost is secondary.

    Need this set up for your team?
    Talk to Will →