Models for Crusoe Managed Inference Service
With Crusoe’s managed inference service, you can test and deploy AI models for generative language, image, and processing tasks. All models provided by Meta are “Built with Llama”.
Available Models
For text generation
We provide an OpenAI-API compatible endpoint at api.crusoe.ai
for the following models:
Name | Provider | Type | Context Length | License | Acceptable Use Policy |
---|---|---|---|---|---|
meta-llama/Meta-Llama-3.1-8B-Instruct (Model card¹) | Meta | instruct | 128k | Llama 3.1 Community License Agreement | Llama 3.1 Acceptable Use Policy |
meta-llama/Llama-3.2-1B-Instruct (Model card¹) | Meta | instruct | 128k | Llama 3.2 Community License Agreement | Llama 3.2 Acceptable Use Policy |
meta-llama/Llama-3.3-70B-Instruct (Model card¹) | Meta | instruct | 128k | Llama 3.3 Community License Agreement | Llama 3.3 Acceptable Use Policy |
Getting Started
After retrieving an API key from the Crusoe Console, you can use the OpenAI SDK to make requests:
import os
from openai import OpenAI
CRUSOE_API_KEY = os.getenv("CRUSOE_API_KEY")
client = OpenAI(
api_key=CRUSOE_API_KEY,
base_url="https://api.crusoe.ai/v1",
)
completion = client.chat.completions.create(
model="meta-llama/Llama-3.3-70B-Instruct",
messages=[
{"role": "system", "content": "You are a helpful, concise assistant."},
{"role": "user", "content": "Who is Robinson Crusoe?"},
],
)
print(completion.choices[0].message.content)