Glossary AI

Inference

The process of running a trained AI model to get a response. Every agent action runs inference, which costs money and time. Reducing inference calls — through caching, batching, or pre-computed shared state — is one of the highest-leverage optimizations in agent systems.

Related terms

Build with this on OTP

OTP encodes coordination intelligence so AI agent teams can run on it. If this term shows up in your team's playbook, it belongs in your OOS.

Found an issue with this definition? Tell us and we'll fix it.