Docker Deployment
Deploy Obot using Docker for local development, testing, and proof-of-concept scenarios.
Overview
Docker deployment is the fastest way to get Obot running. It's ideal for:
- Local development and testing
- Single-machine deployments
- Proof-of-concept and evaluation
- Small team usage
For production deployments, see Kubernetes Deployment.
Prerequisites
- Docker installed
- 2+ CPU cores and 4GB RAM available
- 10GB disk space
Quick Start
Basic Deployment (Built-in PostgreSQL)
Run Obot with the built-in PostgreSQL instance (suitable for development and testing):
docker run -d \
  --name obot \
  -v obot-data:/data \
  -v /var/run/docker.sock:/var/run/docker.sock \
  -p 8080:8080 \
  -e OPENAI_API_KEY=your-openai-key \
  ghcr.io/obot-platform/obot:latest
With Authentication
docker run -d \
  --name obot \
  -v obot-data:/data \
  -v /var/run/docker.sock:/var/run/docker.sock \
  -p 8080:8080 \
  -e OPENAI_API_KEY=your-openai-key \
  -e OBOT_SERVER_ENABLE_AUTHENTICATION=true \
  ghcr.io/obot-platform/obot:latest
Accessing Obot
Once started, access Obot at:
- Web UI: http://localhost:8080
Bootstrap token
The first time you access Obot, you may need the bootstrap token found in the logs:
docker logs obot
You will need to look for an entry like:
--------------------------------------
| Bootstrap token: <BOOTSTRAP_TOKEN> |
--------------------------------------
Next Steps
- Configure Authentication: Set up auth providers for secure access
- Configure Model Providers: Configure model providers (OpenAI, Anthropic, etc.)
- Set Up MCP Tools: Deploy MCP servers for extended functionality