x-defaults: &defaults restart: unless-stopped logging: driver: json-file options: max-size: 100m max-file: "3" services: ollama: <<: *defaults image: ${GLOBAL_REGISTRY:-}ollama/ollama:${OLLAMA_VERSION:-0.14.3} ports: - "${OLLAMA_PORT_OVERRIDE:-11434}:11434" volumes: - ollama_models:/root/.ollama environment: - TZ=${TZ:-UTC} ipc: host healthcheck: test: [ "CMD", "wget", "--no-verbose", "--tries=1", "--spider", "http://localhost:11434/", ] interval: 30s timeout: 10s retries: 3 start_period: 30s deploy: resources: limits: cpus: ${OLLAMA_CPU_LIMIT:-8.0} memory: ${OLLAMA_MEMORY_LIMIT:-16G} reservations: cpus: ${OLLAMA_CPU_RESERVATION:-2.0} memory: ${OLLAMA_MEMORY_RESERVATION:-4G} devices: - driver: nvidia device_ids: ["0"] capabilities: [gpu] volumes: ollama_models: