feat(docs): Update Ollama setup docs with new methods (#10693)

### Changes 🏗️

Expanded the Ollama setup instructions to include desktop app, Docker,
and legacy CLI methods. Added new screenshots for network exposure and
model selection. Clarified steps for starting the AutoGPT platform,
configuring models, and troubleshooting. Included instructions for
adding custom models and improved overall documentation structure.


#### For code changes:
- [x] I have clearly listed my changes in the PR description
- [x] I have made a test plan
- [x] I have tested my changes according to the test plan:
  <!-- Put your test plan here: -->
- [x] Look at the new ollama doc to make sure it makes sense and is easy
to follow
This commit is contained in:
Bently
2025-08-21 22:32:53 +01:00
committed by GitHub
parent 7c908c10b8
commit fa5ff9ca3c
3 changed files with 167 additions and 53 deletions

Binary file not shown.

After

Width:  |  Height:  |  Size: 55 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 40 KiB

View File

@@ -13,76 +13,113 @@ Follow these steps to set up and run Ollama with the AutoGPT platform.
### 1. Launch Ollama
To properly set up Ollama for network access, follow these steps:
To properly set up Ollama for network access, choose one of these methods:
1. **Set the host environment variable:**
**Method A: Using Ollama Desktop App (Recommended)**
**Windows (Command Prompt):**
```
set OLLAMA_HOST=0.0.0.0:11434
```
**Linux/macOS (Terminal):**
```bash
export OLLAMA_HOST=0.0.0.0:11434
```
1. Open the Ollama desktop application
2. Go to **Settings** and toggle **"Expose Ollama to the network"**
![Expose Ollama to Network](../imgs/ollama/Ollama-Expose-Network.png)
3. Click on the model name field in the "New Chat" window
4. Search for "llama3.2" (or your preferred model)
![Select llama3.2 model](../imgs/ollama/Ollama-Select-llama3.2.png)
5. Click on it to start the download and load the model to be used
2. Start the Ollama server:
```
ollama serve
```
??? note "Method B: Using Docker (Alternative)"
3. **Open a new terminal/command window** and download your desired model:
```
ollama pull llama3.2
```
If you prefer to run Ollama via Docker instead of the desktop app, you can use the official [Ollama Docker image](https://hub.docker.com/r/ollama/ollama):
> **Note**: This will download the [llama3.2](https://ollama.com/library/llama3.2) model. Keep the terminal with `ollama serve` running in the background throughout your session.
1. **Start Ollama container** (choose based on your hardware):
### 2. Start the Backend
**CPU only:**
```bash
docker run -d -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama
```
Open a new terminal and navigate to the autogpt_platform directory:
**With NVIDIA GPU** (requires [NVIDIA Container Toolkit](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html)):
```bash
docker run -d --gpus=all -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama
```
**With AMD GPU:**
```bash
docker run -d --device /dev/kfd --device /dev/dri -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama:rocm
```
**Download your desired model:**
```bash
docker exec -it ollama ollama run llama3.2
```
!!! note
The Docker method automatically exposes Ollama on `0.0.0.0:11434`, making it accessible to AutoGPT. More models can be found on the [Ollama library](https://ollama.com/library).
??? warning "Method C: Using Ollama Via Command Line (Legacy)"
For users still using the traditional CLI approach or older Ollama installations:
1. **Set the host environment variable:**
**Windows (Command Prompt):**
```cmd
set OLLAMA_HOST=0.0.0.0:11434
```
**Linux/macOS (Terminal):**
```bash
export OLLAMA_HOST=0.0.0.0:11434
```
2. **Start the Ollama server:**
```bash
ollama serve
```
3. **Open a new terminal/command window** and download your desired model:
```bash
ollama pull llama3.2
```
!!! note
This will download the [llama3.2](https://ollama.com/library/llama3.2) model. Keep the terminal with `ollama serve` running in the background throughout your session.
### 2. Start the AutoGPT Platform
Navigate to the autogpt_platform directory and start all services:
```bash
cd autogpt_platform
docker compose up -d --build
```
### 3. Start the Frontend
This command starts both the backend and frontend services. Once running, visit [http://localhost:3000](http://localhost:3000) to access the platform. After registering/logging in, navigate to the build page at [http://localhost:3000/build](http://localhost:3000/build).
Open a new terminal and navigate to the frontend directory:
### 3. Using Ollama with AutoGPT
```bash
cd autogpt_platform/frontend
corepack enable
pnpm i
pnpm dev
```
Then visit [http://localhost:3000](http://localhost:3000) to see the frontend running, after registering an account/logging in, navigate to the build page at [http://localhost:3000/build](http://localhost:3000/build)
### 4. Using Ollama with AutoGPT
Now that both Ollama and the AutoGPT platform are running we can move onto using Ollama with AutoGPT:
Now that both Ollama and the AutoGPT platform are running, we can use Ollama with AutoGPT:
1. Add an AI Text Generator block to your workspace (it can work with any AI LLM block but for this example will be using the AI Text Generator block):
![Add AI Text Generator Block](../imgs/ollama/Select-AI-block.png)
2. In the "LLM Model" dropdown, select "llama3.2" (This is the model we downloaded earlier)
2. **Configure the API Key field**: Enter any value (e.g., "dummy" or "not-needed") since Ollama doesn't require authentication.
3. In the "LLM Model" dropdown, select "llama3.2" (This is the model we downloaded earlier)
![Select Ollama Model](../imgs/ollama/Ollama-Select-Llama32.png)
> **Compatible Models**: Not all models work with Ollama in AutoGPT. Here are the models that are confirmed to work:
> - `llama3.2`
> - `llama3`
> **Compatible Models**: The following Ollama models are available in AutoGPT by default:
> - `llama3.2` (Recommended for most use cases)
> - `llama3`
> - `llama3.1:405b`
> - `dolphin-mistral:latest`
>
> **Note**: To use other models, follow the "Add Custom Models" step above.
3. **Set your local IP address** in the "Ollama Host" field:
4. **Set your local IP address** in the "Ollama Host" field:
**To find your local IP address:**
**Windows (Command Prompt):**
```
```cmd
ipconfig
```
@@ -92,7 +129,7 @@ Now that both Ollama and the AutoGPT platform are running we can move onto using
```
or
```bash
ipconfig
ifconfig
```
Look for your IPv4 address (e.g., `192.168.0.39`), then enter it with port `11434` in the "Ollama Host" field:
@@ -102,7 +139,9 @@ Now that both Ollama and the AutoGPT platform are running we can move onto using
![Ollama Remote Host](../imgs/ollama/Ollama-Remote-Host.png)
4. Now we need to add some prompts then save and then run the graph:
> **Important**: Since AutoGPT runs in Docker containers, you must use your host machine's IP address instead of `localhost` or `127.0.0.1`. Docker containers cannot reach `localhost` on the host machine.
5. Add prompts to your AI block, save the graph, and run it:
![Add Prompt](../imgs/ollama/Ollama-Add-Prompts.png)
That's it! You've successfully setup the AutoGPT platform and made a LLM call to Ollama.
@@ -115,7 +154,7 @@ For running Ollama on a remote server, simply make sure the Ollama server is run
**To find your local IP address of the system running Ollama:**
**Windows (Command Prompt):**
```
```cmd
ipconfig
```
@@ -125,7 +164,7 @@ ip addr show
```
or
```bash
ipconfig
ifconfig
```
Look for your IPv4 address (e.g., `192.168.0.39`).
@@ -137,16 +176,91 @@ Then you can use the same steps above but you need to add the Ollama server's IP
![Ollama Remote Host](../imgs/ollama/Ollama-Remote-Host.png)
## Add Custom Models (Advanced)
If you want to use models other than the default ones, you'll need to add them to the model list. Follow these steps:
1. **Add the model to the LlmModel enum** in `autogpt_platform/backend/backend/blocks/llm.py`:
Find the Ollama models section (around line 119) and add your model like the other Ollama models:
```python
# Ollama models
OLLAMA_LLAMA3_3 = "llama3.3"
OLLAMA_LLAMA3_2 = "llama3.2"
OLLAMA_YOUR_MODEL = "The-model-name-from-ollama" # Add your model here
```
2. **Add model metadata** in the same file:
Find the `MODEL_METADATA` dictionary (around line 181) and add your model with its metadata:
```python
# In MODEL_METADATA dictionary, add:
LlmModel.OLLAMA_YOUR_MODEL: ModelMetadata("ollama", 8192, None),
```
Where:
- `"ollama"` = provider name
- `8192` = max context window (adjust based on your model)
- `None` = max output tokens (None means no specific limit)
3. **Add model cost configuration** in `autogpt_platform/backend/backend/data/block_cost_config.py`:
Find the `MODEL_COST` dictionary (around line 54) and add your model:
```python
# In MODEL_COST dictionary, add:
LlmModel.OLLAMA_YOUR_MODEL: 1,
```
> **Note**: Setting cost to `1` is fine for local usage as cost tracking is disabled for self-hosted instances.
4. **Rebuild the backend**:
```bash
docker compose up -d --build
```
5. **Pull the model in Ollama**:
```bash
ollama pull your-model-name
```
## Troubleshooting
If you encounter any issues, verify that:
- Ollama is properly installed and running
- All terminals remain open during operation
- Docker is running before starting the backend
- Ollama is properly installed and running with `ollama serve`
- Docker is running before starting the platform
- If running Ollama outside Docker, ensure it's set to `0.0.0.0:11434` for network access
For common errors:
### Common Issues
1. **Connection Refused**: Make sure Ollama is running and the host address is correct (also make sure the port is correct, its default is 11434)
2. **Model Not Found**: Try running `ollama pull llama3.2` manually first
3. **Docker Issues**: Ensure Docker daemon is running with `docker ps`
#### Connection Refused / Cannot Connect to Ollama
- **Most common cause**: Using `localhost` or `127.0.0.1` in the Ollama Host field
- **Solution**: Use your host machine's IP address (e.g., `192.168.0.39:11434`)
- **Why**: AutoGPT runs in Docker containers and cannot reach `localhost` on the host
- **Find your IP**: Use `ipconfig` (Windows) or `ifconfig` (Linux/macOS)
- **Test Ollama is running**: `curl http://localhost:11434/api/tags` should work from your host machine
#### Model Not Found
- Pull the model manually:
```bash
ollama pull llama3.2
```
- If using a custom model, ensure it's added to the model list in `backend/server/model.py`
#### Docker Issues
- Ensure Docker daemon is running:
```bash
docker ps
```
- Try rebuilding:
```bash
docker compose up -d --build
```
#### API Key Errors
- Remember that Ollama doesn't require authentication - any value works for the API key field
#### Model Selection Issues
- Look for models with "ollama" in their description in the dropdown
- Only the models listed in the "Compatible Models" section are guaranteed to work