Run Blender MCP Locally with Ollama (Free, No API Key)
If you want free, private, and offline AI-powered 3D modeling with Blender MCP, Ollama is the answer. This guide explains the setup — and an important caveat: Ollama does not natively speak the MCP protocol. The solution is to use an MCP-compatible client (like Cursor or VSCode with Cline) configured with Ollama as the model provider, alongside the Blender MCP server.
You get: local AI (Ollama) + Blender tools (MCP) = free AI 3D modeling with no API keys or cloud dependency.
How It Works
| Component | Role |
|---|---|
| Ollama | Runs LLMs locally on your machine — no API key, no internet after model download |
| Cursor / VSCode + Cline | MCP-compatible client; uses Ollama for the AI brain and Blender MCP for tools |
| Blender MCP | Provides tools to create and modify 3D content in Blender |
Ollama is an inference engine, not an MCP client. So you need a client that:
- Supports MCP servers (to connect Blender MCP)
- Can use Ollama as the model backend
Cursor and VSCode with the Cline extension both satisfy this. Cursor has built-in support for Ollama. Cline lets you point the AI to Ollama’s API. In both cases, you add Blender MCP as a separate MCP server.
Benefits of Local AI
- Free — No API costs, no usage limits
- Private — Your prompts and data never leave your machine
- Offline — Works without internet after models are downloaded
- Unlimited — No rate limits or quota caps
Limitations
- Slower — Local models are typically slower than cloud APIs
- Less capable — Models like Llama 3.1 are good but often less capable than Claude or GPT-4 for complex 3D workflows
- Hardware — You need a decent GPU (8GB+ VRAM recommended) for smooth performance
- Setup complexity — Two pieces to configure: Ollama + MCP client
Prerequisites
- Blender 3.0+ — Download from blender.org
- Ollama — ollama.com
- Cursor or VSCode (with Cline extension)
- uv package manager — For running Blender MCP
- GPU — 8GB+ VRAM recommended for local models
Step 1: Install Ollama and Pull a Model
macOS / Linux
Download from ollama.com or:
curl -fsSL https://ollama.com/install.sh | shPull a model that supports tool use:
ollama pull llama3.1:8bOr: ollama pull mistral or ollama pull codellama for coding-heavy workflows.
Model choice: llama3.1:8b is a good balance of speed and capability. For more complex 3D tasks, try llama3.1:70b if you have the VRAM (40GB+ recommended). Smaller models (e.g., phi3) may struggle with Blender MCP’s tool-calling patterns.
Step 2: Install Your MCP-Compatible Client
Option A: Cursor (Recommended)
Download Cursor . Cursor has native MCP support and can use Ollama as a model provider.
Option B: VSCode + Cline
Install VSCode and the Cline extension . Cline supports MCP and can be configured to use Ollama.
Step 3: Configure the Client to Use Ollama
Cursor
- Open Cursor Settings → Models
- Add Ollama as a model provider (Cursor detects it when Ollama is running)
- Select your pulled model (e.g.,
llama3.1:8b) for the Agent/Composer
VSCode + Cline
- Open Cline settings
- Set the API endpoint to your local Ollama instance (typically
http://localhost:11434) - Choose your model from the Cline model selector
Ollama API: By default, Ollama serves at http://localhost:11434. Ensure Ollama is running (ollama serve or the Ollama app) before using it from Cursor or Cline.
Step 4: Add the Blender MCP Server
Install uv if you haven’t:
curl -LsSf https://astral.sh/uv/install.sh | shAdd the Blender MCP server to your client’s MCP config. The config is the same as in our other setup guides.
For Cursor — Go to Settings → MCP and add:
macOS / Linux
{
"mcpServers": {
"blender": {
"command": "uvx",
"args": ["blender-mcp"]
}
}
}For VSCode + Cline — Add the same config to Cline’s MCP settings (location varies by Cline version; typically in workspace or user settings).
Step 5: Install the Blender Addon and Connect
Download the addon
Download addon.py from the Blender MCP GitHub repository .
Install in Blender
Open Blender → Edit → Preferences → Add-ons → Install… → Select addon.py.
Enable the addon
Check “Interface: Blender MCP”.
Connect
Press N in the 3D Viewport → BlenderMCP tab → Click “Connect to Claude”.
Step 6: Test Your Setup
With Blender connected and your client configured:
- Open the Agent/Composer (Cursor) or Cline chat in VSCode
- Ensure the model is set to your Ollama model
- Verify Blender MCP appears in the tool list
- Try: “Create a red cube in Blender” or “Add a UV sphere at the origin”
Local models may need clearer, shorter prompts. If a complex request fails, break it into smaller steps.
Troubleshooting
Ollama model doesn’t use Blender tools
- Confirm the model supports function/tool calling (Llama 3.1, Mistral, and CodeLlama do)
- Run a simpler prompt first to verify the MCP connection
- Some smaller models struggle with multi-tool workflows — try a larger model
Blender MCP shows disconnected
- Same fixes as other guides: verify uv, Blender addon enabled, Connect clicked
- Restart the client after adding MCP config
- Only one MCP client should connect to Blender at a time
Slow or unresponsive
- Local inference is slower than cloud APIs
- Use a smaller model (
llama3.1:8b) for faster iteration - Ensure your GPU is being used (
nvidia-smion Linux to check)
Out of memory
- Use a smaller model or reduce context length
- Close other GPU-heavy applications
For a technical deep dive, see our MCP Server architecture guide.
Using a different AI client?
FAQ
Does Ollama natively support MCP?
No. Ollama is an LLM inference engine — it runs models locally but doesn't speak the MCP protocol. To use Blender MCP with Ollama, you need an MCP-compatible client (Cursor or VSCode with Cline) that uses Ollama as the model backend and connects to Blender MCP as a separate server.
Is Blender MCP free with Ollama?
Yes. Blender MCP is free and open source. Ollama is free. Blender is free. The full setup has zero API costs. You only need compatible hardware (ideally a GPU with 8GB+ VRAM) to run models locally.
Which Ollama models work best with Blender MCP?
Models that support tool/function calling work best: llama3.1:8b, llama3.1:70b, mistral, and codellama. Smaller models (phi3, tinyllama) may struggle with Blender MCP's tool-calling patterns. For complex 3D scenes, larger models perform better.
How does Blender MCP + Ollama compare to Claude?
Ollama gives you free, private, offline AI but with slower inference and generally lower capability than Claude for complex 3D tasks. Claude excels at multi-step scene creation and nuanced prompts. Ollama is ideal if you prioritize privacy, cost, or working offline.
Can I use Open WebUI or other Ollama frontends with Blender MCP?
Open WebUI and similar interfaces focus on chat with Ollama models. To use Blender MCP, you need a client that supports both MCP servers and Ollama as a model. Cursor and VSCode with Cline are the most straightforward options. Some MCP-Ollama bridge projects exist but may require extra setup.
Get Started with Blender MCP
Blender MCP is free, open-source, and community-driven. Star the repo, report issues, or contribute — all on GitHub.
View on GitHub →