Your Own Local ChatGPT β With Ollama + Open WebUI π€β¨

Letβs face it: AI is awesome.
But running AI locally, on your own server, offline, without sending anything to the cloud?
Thatβs next-level awesome.
In this post, we’re setting up our very own private LLM chat server using Ollama and Open WebUI.
Fully containerized with Docker. One YAML file. Local AI fun. π»π§
π€ Wait, Whatβs Ollama?
Ollama makes it super simple to run large language models (LLMs) locally. Think ChatGPT, but it lives on your own hardware.
Open WebUI gives you a clean browser interface to interact with your local models.
No API keys. No cloud. Just you and your AI.
π§± docker-compose.yml
Hereβs the full docker-compose.yml
:
version: "3.8"
services:
ollama:
image: ollama/ollama
container_name: ollama
ports:
- "11434:11434"
volumes:
- ollama_data:/root/.ollama
restart: unless-stopped
openwebui:
image: ghcr.io/open-webui/open-webui:main
container_name: openwebui
ports:
- "3000:8080"
environment:
- OLLAMA_BASE_URL=http://ollama:11434
depends_on:
- ollama
volumes:
- openwebui_data:/app/backend/data
restart: unless-stopped
volumes:
ollama_data:
openwebui_data:
Spin it up with:
sudo docker compose -p local-ai up -d
π€ Pulling Your AI Models
Once running, pull some models with:
docker exec -it ollama ollama pull deepseek-r1:latest
docker exec -it ollama ollama pull gemma3:latest
docker exec -it ollama ollama pull phi4:latest
(These can take some time and storage β models can be multiple GBs.)
π¬ Accessing the Chat Interface
Once it’s all up, open your browser and go to:
http://<your-server-ip>:3000
Choose a model and start chatting β locally and privately.
β οΈ Note on Performance (especially on Raspberry Pi)
Yes β you can run this on a Raspberry Pi (especially Pi 5 with active cooling).
Butβ¦ it’s slow. π’
RAM is limited, there’s no GPU, and response times can be frustrating.
π§ Recommended Setup for Smooth LLM Use
If you want faster and smoother experience, consider running this on:
- β At least 16 GB RAM
- β A modern x86 CPU (e.g., Ryzen 5 5600G or Intel i5+)
- β Optional: GPU support (NVIDIA RTX 3060 or better for even faster inference)
- β Fast SSD storage (models load faster)
- β Linux or WSL2 on Windows
Even an old laptop with decent specs will do better than a Pi.
π§ Why This Rocks
- π‘οΈ Privacy-first AI
- π No cloud required
- π§° Self-hosted and hackable
- π₯οΈ Clean web UI for everyday use
- π« No token limits or subscriptions