AI prompts glossary
Token Limit
A token limit is the maximum number of tokens that can be included in a single model call, counting prompts, context, and generated output. Exceeding this limit truncates input or responses. For prompt engineers and system designers, managing token limits is crucial for long conversations, detailed instructions, and RAG contexts, ensuring that critical information fits while controlling latency and cost. A token limit is the maximum number of tokens that a model call can include across prompts, context, and generated output. If the limit is exceeded, input may be truncated or responses shortened. For prompt engineers and architects, managing token limits is essential to keep conversations coherent while controlling latency and cost.

