Glenn's Digital Garden

Home

❯

sequence length

sequence length

Aug 27, 2025

  • artificial-intelligence
  • seedling

Sequence length or context length or context window is the maximum number of tokens that can be input into a transformer.

Here are a few examples:

ModelSequence LengthReference
GPT-4o128K
Llama-3.1 405b128K
DeepSeek-R1128K
GPT-4.11M1
Gemini 2.5 Pro1M2

Footnotes

  1. Introducing GPT-4.1 in the API ↩

  2. Gemini models. Gemini 2.5 Pro actually has a sequence length of 220=1,048,576 tokens. ↩


Graph View

Backlinks

  • Mamba
  • attention

Created with Quartz v4.5.2 © 2026

  • glennklockwood.com
  • @glennklockwood.com