# Running Ollama with AutoGPT > **Important**: Ollama integration is only available when self-hosting the AutoGPT platform. It cannot be used with the cloud-hosted version. Follow these steps to set up and run Ollama with the AutoGPT platform. ## Prerequisites 1. Make sure you have gone through and completed the [AutoGPT Setup](/platform/getting-started) steps, if not please do so before continuing with this guide. 2. Before starting, ensure you have [Ollama installed](https://ollama.com/download) on your machine. ## Setup Steps ### 1. Launch Ollama To properly set up Ollama for network access, follow these steps: 1. **Set the host environment variable:** **Windows (Command Prompt):** ``` set OLLAMA_HOST=0.0.0.0:11434 ``` **Linux/macOS (Terminal):** ```bash export OLLAMA_HOST=0.0.0.0:11434 ``` 2. Start the Ollama server: ``` ollama serve ``` 3. **Open a new terminal/command window** and download your desired model: ``` ollama pull llama3.2 ``` > **Note**: This will download the [llama3.2](https://ollama.com/library/llama3.2) model. Keep the terminal with `ollama serve` running in the background throughout your session. ### 2. Start the Backend Open a new terminal and navigate to the autogpt_platform directory: ```bash cd autogpt_platform docker compose up -d --build ``` ### 3. Start the Frontend Open a new terminal and navigate to the frontend directory: ```bash cd autogpt_platform/frontend corepack enable pnpm i pnpm dev ``` Then visit [http://localhost:3000](http://localhost:3000) to see the frontend running, after registering an account/logging in, navigate to the build page at [http://localhost:3000/build](http://localhost:3000/build) ### 4. Using Ollama with AutoGPT Now that both Ollama and the AutoGPT platform are running we can move onto using Ollama with AutoGPT: 1. Add an AI Text Generator block to your workspace (it can work with any AI LLM block but for this example will be using the AI Text Generator block): ![Add AI Text Generator Block](../imgs/ollama/Select-AI-block.png) 2. In the "LLM Model" dropdown, select "llama3.2" (This is the model we downloaded earlier) ![Select Ollama Model](../imgs/ollama/Ollama-Select-Llama32.png) > **Compatible Models**: Not all models work with Ollama in AutoGPT. Here are the models that are confirmed to work: > - `llama3.2` > - `llama3` > - `llama3.1:405b` > - `dolphin-mistral:latest` 3. **Set your local IP address** in the "Ollama Host" field: **To find your local IP address:** **Windows (Command Prompt):** ``` ipconfig ``` **Linux/macOS (Terminal):** ```bash ip addr show ``` or ```bash ipconfig ``` Look for your IPv4 address (e.g., `192.168.0.39`), then enter it with port `11434` in the "Ollama Host" field: ``` 192.168.0.39:11434 ``` ![Ollama Remote Host](../imgs/ollama/Ollama-Remote-Host.png) 4. Now we need to add some prompts then save and then run the graph: ![Add Prompt](../imgs/ollama/Ollama-Add-Prompts.png) That's it! You've successfully setup the AutoGPT platform and made a LLM call to Ollama. ![Ollama Output](../imgs/ollama/Ollama-Output.png) ### Using Ollama on a Remote Server with AutoGPT For running Ollama on a remote server, simply make sure the Ollama server is running and is accessible from other devices on your network/remotely through the port 11434. **To find your local IP address of the system running Ollama:** **Windows (Command Prompt):** ``` ipconfig ``` **Linux/macOS (Terminal):** ```bash ip addr show ``` or ```bash ipconfig ``` Look for your IPv4 address (e.g., `192.168.0.39`). Then you can use the same steps above but you need to add the Ollama server's IP address to the "Ollama Host" field in the block settings like so: ``` 192.168.0.39:11434 ``` ![Ollama Remote Host](../imgs/ollama/Ollama-Remote-Host.png) ## Troubleshooting If you encounter any issues, verify that: - Ollama is properly installed and running - All terminals remain open during operation - Docker is running before starting the backend For common errors: 1. **Connection Refused**: Make sure Ollama is running and the host address is correct (also make sure the port is correct, its default is 11434) 2. **Model Not Found**: Try running `ollama pull llama3.2` manually first 3. **Docker Issues**: Ensure Docker daemon is running with `docker ps`