Claude’s context window determines how much information it can hold and process in a single conversation. Claude Sonnet 4.6 and Opus 4.6 support 1 million tokens; Haiku 4.5 supports 200,000 tokens — one of the largest in the industry. Here’s what that means in practice, what you can actually fit inside it, and how context window size affects your work.
Claude Context Window by Model (April 2026)
| Model | Context Window | ~Words | ~Pages |
|---|---|---|---|
| Claude Haiku | 200,000 tokens | ~150,000 | ~500 |
| Claude Sonnet | 200,000 tokens | ~150,000 | ~500 |
| Claude Opus | 200,000 tokens | ~150,000 | ~500 |
What Fits in 200K Tokens
| Content type | Approximate fit |
|---|---|
| News articles | ~200+ articles |
| Research papers | ~30–50 papers depending on length |
| A full novel | Yes — most novels fit with room to spare |
| Python codebase | Medium-sized codebases (10k–50k lines) |
| Legal contracts | Hundreds of pages of contracts |
| Conversation history | Very long sessions before truncation |
Context Window vs. Output Length
The context window covers everything Claude processes — both input and output combined. If your prompt is 50,000 tokens (a long document), Claude has 150,000 tokens remaining for its response and any further back-and-forth. The window is shared between what you send and what Claude generates.
Maximum output length is a separate constraint — Claude won’t generate an infinitely long response even within a large context window. For very long outputs (full books, extensive reports), you typically work in sections rather than expecting Claude to produce everything in one pass.
Why Context Window Size Matters
Context window size is the practical limit on how much work you can give Claude at once without losing information. Before large context windows, working with long documents required chunking — splitting the document into pieces, analyzing each separately, and manually synthesizing the results. With 200K tokens, Claude can hold the entire document and answer questions about any part of it with full awareness of everything else.
This matters most for: document analysis and legal review, code understanding across large files, research synthesis across many sources, and long multi-step conversations where earlier context affects later decisions.
How Claude Performs at the Edges of Its Context Window
Research on large language models has found that performance can degrade somewhat for information buried in the middle of a very long context — sometimes called the “lost in the middle” problem. Claude performs well across its context window, but for maximum reliability on information from a very long document, referencing specific sections explicitly (“in the section about pricing on page 12…”) helps ensure Claude focuses on the right part.
For the full model spec breakdown, see Claude API Model Strings and Specs and Claude Models Explained: Haiku vs Sonnet vs Opus.
Frequently Asked Questions
What is Claude’s context window size?
Claude Sonnet 4.6 and Opus 4.6 support a 1 million token context window at standard pricing. Claude Haiku 4.5 supports 200,000 tokens. That’s approximately 150,000 words or about 500 pages of text in a single conversation.
How many tokens is 200K context?
200,000 tokens is approximately 150,000 words of English text. One token is roughly four characters or three-quarters of a word. A typical 800-word article is about 1,000 tokens; a full novel is typically 80,000–120,000 tokens.
Can I upload a full PDF to Claude?
Yes, as long as the PDF’s text content fits within the 200K token context window. Most documents, reports, contracts, and research papers fit easily. Very large documents (multiple volumes, extensive legal filings) may need to be split.