Files
AutoGPT/docs/content/platform/ollama.md
Bently 79b38343c2 docs(ollama): Update to add info on how to properly setup ollama environment variables (#10089)
Update ollama docs to add info on how to setup ollama environment vars
for proper access

This includes properly setting the "OLLAMA_HOST" env var with the ip and
port "0.0.0.0:11434" which makes it accessible to AutoGPT thats running
inside of docker



#### For code changes:
- [x] I have clearly listed my changes in the PR description
- [x] I have made a test plan
- [x] I have tested my changes according to the test plan:
  <!-- Put your test plan here: -->
  - [x] Follow the latest setup to test Ollama to make sure it works
2025-06-06 10:45:31 +00:00

153 lines
4.3 KiB
Markdown

# Running Ollama with AutoGPT
> **Important**: Ollama integration is only available when self-hosting the AutoGPT platform. It cannot be used with the cloud-hosted version.
Follow these steps to set up and run Ollama with the AutoGPT platform.
## Prerequisites
1. Make sure you have gone through and completed the [AutoGPT Setup](/platform/getting-started) steps, if not please do so before continuing with this guide.
2. Before starting, ensure you have [Ollama installed](https://ollama.com/download) on your machine.
## Setup Steps
### 1. Launch Ollama
To properly set up Ollama for network access, follow these steps:
1. **Set the host environment variable:**
**Windows (Command Prompt):**
```
set OLLAMA_HOST=0.0.0.0:11434
```
**Linux/macOS (Terminal):**
```bash
export OLLAMA_HOST=0.0.0.0:11434
```
2. Start the Ollama server:
```
ollama serve
```
3. **Open a new terminal/command window** and download your desired model:
```
ollama pull llama3.2
```
> **Note**: This will download the [llama3.2](https://ollama.com/library/llama3.2) model. Keep the terminal with `ollama serve` running in the background throughout your session.
### 2. Start the Backend
Open a new terminal and navigate to the autogpt_platform directory:
```bash
cd autogpt_platform
docker compose up -d --build
```
### 3. Start the Frontend
Open a new terminal and navigate to the frontend directory:
```bash
cd autogpt_platform/frontend
corepack enable
pnpm i
pnpm dev
```
Then visit [http://localhost:3000](http://localhost:3000) to see the frontend running, after registering an account/logging in, navigate to the build page at [http://localhost:3000/build](http://localhost:3000/build)
### 4. Using Ollama with AutoGPT
Now that both Ollama and the AutoGPT platform are running we can move onto using Ollama with AutoGPT:
1. Add an AI Text Generator block to your workspace (it can work with any AI LLM block but for this example will be using the AI Text Generator block):
![Add AI Text Generator Block](../imgs/ollama/Select-AI-block.png)
2. In the "LLM Model" dropdown, select "llama3.2" (This is the model we downloaded earlier)
![Select Ollama Model](../imgs/ollama/Ollama-Select-Llama32.png)
> **Compatible Models**: Not all models work with Ollama in AutoGPT. Here are the models that are confirmed to work:
> - `llama3.2`
> - `llama3`
> - `llama3.1:405b`
> - `dolphin-mistral:latest`
3. **Set your local IP address** in the "Ollama Host" field:
**To find your local IP address:**
**Windows (Command Prompt):**
```
ipconfig
```
**Linux/macOS (Terminal):**
```bash
ip addr show
```
or
```bash
ipconfig
```
Look for your IPv4 address (e.g., `192.168.0.39`), then enter it with port `11434` in the "Ollama Host" field:
```
192.168.0.39:11434
```
![Ollama Remote Host](../imgs/ollama/Ollama-Remote-Host.png)
4. Now we need to add some prompts then save and then run the graph:
![Add Prompt](../imgs/ollama/Ollama-Add-Prompts.png)
That's it! You've successfully setup the AutoGPT platform and made a LLM call to Ollama.
![Ollama Output](../imgs/ollama/Ollama-Output.png)
### Using Ollama on a Remote Server with AutoGPT
For running Ollama on a remote server, simply make sure the Ollama server is running and is accessible from other devices on your network/remotely through the port 11434.
**To find your local IP address of the system running Ollama:**
**Windows (Command Prompt):**
```
ipconfig
```
**Linux/macOS (Terminal):**
```bash
ip addr show
```
or
```bash
ipconfig
```
Look for your IPv4 address (e.g., `192.168.0.39`).
Then you can use the same steps above but you need to add the Ollama server's IP address to the "Ollama Host" field in the block settings like so:
```
192.168.0.39:11434
```
![Ollama Remote Host](../imgs/ollama/Ollama-Remote-Host.png)
## Troubleshooting
If you encounter any issues, verify that:
- Ollama is properly installed and running
- All terminals remain open during operation
- Docker is running before starting the backend
For common errors:
1. **Connection Refused**: Make sure Ollama is running and the host address is correct (also make sure the port is correct, its default is 11434)
2. **Model Not Found**: Try running `ollama pull llama3.2` manually first
3. **Docker Issues**: Ensure Docker daemon is running with `docker ps`