diff --git a/docker-compose.yaml b/docker-compose.yaml index fce93be..3a04309 100644 --- a/docker-compose.yaml +++ b/docker-compose.yaml @@ -1,5 +1,5 @@ services: - ollama: + ollamasmall: image: ${OLLAMA:-ollama/ollama:latest} deploy: resources: @@ -7,15 +7,35 @@ services: devices: - driver: nvidia capabilities: [gpu] - count: all + device_ids: ['0'] ports: - "11434:11434" volumes: - - ollama:/root/.ollama + - ollama-small:/root/.ollama restart: always container_name: ollama networks: - shared_net + command: serve llava:7b granite3-dense:2b --no-timeout + + ollamalarge: + image: ${OLLAMA:-ollama/ollama:latest} + deploy: + resources: + reservations: + devices: + - driver: nvidia + capabilities: [gpu] + device_ids: ['1'] # Assign to GPU1 + ports: + - "11435:11434" # Different port to avoid conflicts + volumes: + - ollama:/root/.ollama + restart: always + container_name: ollamalarge + networks: + - shared_net + open_webui: container_name: openwebui @@ -75,6 +95,7 @@ services: OPENID_PROVIDER_URL: https://git.adamoutler.com/.well-known/openid-configuration OPENID_REDIRECT_URI: https://ai.hackedyour.info/oauth/oidc/callback OLLAMA_BASE_URL: http://ollama:11434 + OLLAMA_BASE_URLS: http://ollamasmall:11434,http://ollamalarge:11435 OPENAI_API_KEY: ${OPEN_AI_TOKEN} OAUTH_CLIENT_ID: ${CLIENT_ID} OAUTH_CLIENT_SECRET: ${CLIENT_SECRET} @@ -102,6 +123,7 @@ services: volumes: ollama: + ollama-small: data: chromadb: