Cache (LLM Cache)

Storing previous AI responses for reuse. Saves costs and speeds up repeated queries.

Why it matters

Many AI queries are repetitive. Caching means you pay for the first answer, then serve identical questions for free — cutting costs by 40-80% in typical business deployments.

In practice

Our LLM routing uses content-hash caching: identical queries return cached results instantly. FAQ matching handles ~40% of chat queries for free before any LLM is called.

Related terms

Back to glossary