Skip to content

AI prompts glossary

Latency (LLM API)

Latency in an LLM API is the time between sending a request and receiving a response from the model. It depends on model size, prompt length, decoding settings, and infrastructure. For conversational products and live campaigns, latency directly affects user experience and perceived responsiveness. Designers and engineers must balance output quality with speed to keep Ai Messages feeling timely and interactive. Latency in an LLM API is the time between sending a request and receiving a response from the model. It is influenced by model size, prompt length, and decoding settings. For conversational products and live workflows, latency directly affects user experience, making response-time optimization an important design factor.