Skip to content

Instantly share code, notes, and snippets.

@EvilFreelancer
Created March 18, 2025 10:32
Show Gist options
  • Save EvilFreelancer/d0ea8bd0298b9107ff3f5bf402d3a737 to your computer and use it in GitHub Desktop.
Save EvilFreelancer/d0ea8bd0298b9107ff3f5bf402d3a737 to your computer and use it in GitHub Desktop.
Конфиг docker-compose.yml с ollama и open-webui
x-shared-logs: &shared-logs
logging:
driver: "json-file"
options:
max-size: "10k"
services:
ollama:
image: ollama/ollama:0.6.0
restart: unless-stopped
volumes:
- ./ollama_data:/root
environment:
OLLAMA_KEEP_ALIVE: 60m
OLLAMA_FLASH_ATTENTION: 1
OLLAMA_MAX_LOADED_MODELS: 2
OLLAMA_MAX_QUEUE: 8
OLLAMA_NUM_PARALLEL: 10
OLLAMA_GPU_OVERHEAD: 0
OLLAMA_SCHED_SPREAD: 1
OLLAMA_CONTEXT_LENGTH: 16000
ports:
- "11434:11434"
deploy:
resources:
reservations:
devices:
- driver: nvidia
count: all
capabilities: [ gpu ]
<<: *shared-logs
open-webui:
image: ghcr.io/open-webui/open-webui:main
restart: unless-stopped
ports:
- "8080:8080"
volumes:
- ./open-webui_data:/app/backend/data:rw
environment:
# WEBUI_AUTH: false
OLLAMA_BASE_URL: http://ollama:11434
<<: *shared-logs
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment