Cerebras Provider

Cerebras is exposed as a named preset over SwarmVault's OpenAI-compatible adapter. Cerebras runs inference on purpose-built wafer-scale hardware for fast throughput.

Configuration

{
  "providers": {
    "primary": {
      "type": "cerebras",
      "model": "llama3.1-70b"
    }
  }
}

Options

FieldDefaultDescription
model--Model ID (e.g., llama3.1-70b, llama3.1-8b)
apiKeyEnv"CEREBRAS_API_KEY"Environment variable for API key
baseUrlhttps://api.cerebras.ai/v1API base URL
apiStyle"chat"API style (chat for Chat Completions)
capabilities--Override auto-detected capabilities

Environment Variable

export CEREBRAS_API_KEY=csk-...

Notes

  • Speed: Cerebras is optimized for high-throughput inference on their custom hardware. Like Groq, it works well as a fast provider for query or lint tasks.
  • Models: Cerebras currently hosts Llama variants. Check the Cerebras API docs for the current model list.