Skip to main content

Models for Crusoe Managed Inference Service

With Crusoe’s managed inference service, you can test and deploy AI models for generative language, image, and processing tasks. All models provided by Meta are “Built with Llama”.

Available Models

For text generation

We provide an OpenAI-API compatible endpoint at api.crusoe.ai for the following models:

NameProviderTypeContext LengthLicenseAcceptable Use Policy
meta-llama/Meta-Llama-3.1-8B-Instruct (Model card¹)Metainstruct128kLlama 3.1 Community License AgreementLlama 3.1 Acceptable Use Policy
meta-llama/Llama-3.2-1B-Instruct (Model card¹)Metainstruct128kLlama 3.2 Community License AgreementLlama 3.2 Acceptable Use Policy
meta-llama/Llama-3.3-70B-Instruct (Model card¹)Metainstruct128kLlama 3.3 Community License AgreementLlama 3.3 Acceptable Use Policy

Getting Started

After retrieving an API key from the Crusoe Console, you can use the OpenAI SDK to make requests:

import os
from openai import OpenAI

CRUSOE_API_KEY = os.getenv("CRUSOE_API_KEY")
client = OpenAI(
api_key=CRUSOE_API_KEY,
base_url="https://api.crusoe.ai/v1",
)

completion = client.chat.completions.create(
model="meta-llama/Llama-3.3-70B-Instruct",
messages=[
{"role": "system", "content": "You are a helpful, concise assistant."},
{"role": "user", "content": "Who is Robinson Crusoe?"},
],
)

print(completion.choices[0].message.content)