Skip to content

Encodings

Token counting uses tiktoken. Encoding is auto-selected from the model name.

Model → Encoding

ModelContext WindowMax InputEncoding
gpt-5400K272Ko200k_base
gpt-4o128K112Ko200k_base
gpt-4o-mini128K112Ko200k_base
gpt-4-turbo128K124Kcl100k_base
gpt-48K4Kcl100k_base
gpt-3.5-turbo16K12Kcl100k_base

Override

To use a non-default encoding:

json
{
  "gpt-4o": {
    "encoding": "cl100k_base",
    "warning": 8000,
    "error": 12000
  }
}

Valid values: cl100k_base, o200k_base, p50k_base, r50k_base.

Released under the MIT License.