Context Window #186494
Replies: 4 comments 2 replies
-
|
Beta Was this translation helpful? Give feedback.
-
|
Claude Opus 4.6 “supporting 1M context” is basically the model’s maximum architectural capability, but in the actual product Anthropic sets practical limits for speed, cost, and reliability. At 1M tokens the compute and latency blow up massively, and even if the model can technically read that much, its effective recall and consistency can become less stable. That’s why they cap it at 128K input and 64K output to keep performance predictable, inference fast enough, and pricing manageable. |
Beta Was this translation helpful? Give feedback.
-
|
Claude Opus 4.6 can support up to ~1M tokens at the model level. Copilot Chat does not expose the model’s maximum context. Copilot applies its own caps for:
So you see: 128K input / 64K output Important distinction Model capability ≠ Product limit. Vendors frequently gate large contexts behind:
Copilot currently uses a constrained configuration of Claude models. Why Copilot does this
GitHub optimizes for interactive developer workflows, not massive document ingestion. If you truly need >128K Use Claude directly through Anthropic’s API or a provider that exposes 1M context. Copilot is not the right surface today. |
Beta Was this translation helpful? Give feedback.
-
|
Claude Opus and Sonnet 4.6 both just changed 1M context window from beta to GA. Copilot should make use of it as well |
Beta Was this translation helpful? Give feedback.

Uh oh!
There was an error while loading. Please reload this page.
-
Select Topic Area
Question
Copilot Feature Area
VS Code
Body
Why does the Claude Opus4.6 token context window only have 128K input and 64K output, when the model can handle up to 1M?

Beta Was this translation helpful? Give feedback.
All reactions