Managed Inference Overview
Crusoe Cloud's Managed Inference service allows you to interact with supported models through our Intelligence Foundry APIs. Models are served on our proprietary inference engine with MemoryAlloy, a cluster-wide memory fabric with cache aware routing that maximizes cache hits, improving TTFT and throughput.
Available models
You can use the OpenAI-API compatible endpoint at managed-inference-api-proxy.crusoecloud.com to access the models below. You can also interact with all of the models using the Intelligence Foundry's chat interface. All Meta models provided by Crusoe are "Built with Llama".
For each model's pricing information, see pricing.