Ollama Cloud
Access 29 Ollama Cloud models through Mastra's model router. Authentication is handled automatically using the OLLAMA_API_KEY environment variable.
Learn more in the Ollama Cloud documentation.
.env
OLLAMA_API_KEY=your-api-key
src/mastra/agents/my-agent.ts
import { Agent } from "@mastra/core/agent";
const agent = new Agent({
id: "my-agent",
name: "My Agent",
instructions: "You are a helpful assistant",
model: "ollama-cloud/cogito-2.1:671b"
});
// Generate a response
const response = await agent.generate("Hello!");
// Stream a response
const stream = await agent.stream("Tell me a story");
for await (const chunk of stream) {
console.log(chunk);
}
info
Mastra uses the OpenAI-compatible /chat/completions endpoint. Some provider-specific features may not be available. Check the Ollama Cloud documentation for details.
ModelsDirect link to Models
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|---|---|---|---|---|---|---|---|---|
ollama-cloud/cogito-2.1:671b | 164K | — | — | |||||
ollama-cloud/deepseek-v3.1:671b | 164K | — | — | |||||
ollama-cloud/deepseek-v3.2 | 164K | — | — | |||||
ollama-cloud/devstral-2:123b | 262K | — | — | |||||
ollama-cloud/devstral-small-2:24b | 262K | — | — | |||||
ollama-cloud/gemini-3-flash-preview | 1.0M | — | — | |||||
ollama-cloud/gemini-3-pro-preview | 1.0M | — | — | |||||
ollama-cloud/gemma3:12b | 131K | — | — | |||||
ollama-cloud/gemma3:27b | 131K | — | — | |||||
ollama-cloud/gemma3:4b | 131K | — | — | |||||
ollama-cloud/glm-4.6 | 203K | — | — | |||||
ollama-cloud/glm-4.7 | 203K | — | — | |||||
ollama-cloud/gpt-oss:120b | 131K | — | — | |||||
ollama-cloud/gpt-oss:20b | 131K | — | — | |||||
ollama-cloud/kimi-k2-thinking | 262K | — | — | |||||
ollama-cloud/kimi-k2:1t | 262K | — | — | |||||
ollama-cloud/kimi-k2.5 | 262K | — | — | |||||
ollama-cloud/minimax-m2 | 205K | — | — | |||||
ollama-cloud/minimax-m2.1 | 205K | — | — | |||||
ollama-cloud/ministral-3:14b | 262K | — | — | |||||
ollama-cloud/ministral-3:3b | 262K | — | — | |||||
ollama-cloud/ministral-3:8b | 262K | — | — | |||||
ollama-cloud/mistral-large-3:675b | 262K | — | — | |||||
ollama-cloud/nemotron-3-nano:30b | 1.0M | — | — | |||||
ollama-cloud/qwen3-coder:480b | 262K | — | — | |||||
ollama-cloud/qwen3-next:80b | 262K | — | — | |||||
ollama-cloud/qwen3-vl:235b | 262K | — | — | |||||
ollama-cloud/qwen3-vl:235b-instruct | 262K | — | — | |||||
ollama-cloud/rnj-1:8b | 33K | — | — |
Advanced ConfigurationDirect link to Advanced Configuration
Custom HeadersDirect link to Custom Headers
src/mastra/agents/my-agent.ts
const agent = new Agent({
id: "custom-agent",
name: "custom-agent",
model: {
url: "https://ollama.com/v1",
id: "ollama-cloud/cogito-2.1:671b",
apiKey: process.env.OLLAMA_API_KEY,
headers: {
"X-Custom-Header": "value"
}
}
});
Dynamic Model SelectionDirect link to Dynamic Model Selection
src/mastra/agents/my-agent.ts
const agent = new Agent({
id: "dynamic-agent",
name: "Dynamic Agent",
model: ({ requestContext }) => {
const useAdvanced = requestContext.task === "complex";
return useAdvanced
? "ollama-cloud/rnj-1:8b"
: "ollama-cloud/cogito-2.1:671b";
}
});