forked from open-webui/open-webui
Merge pull request #246 from gnscc/main
Refactor docker-compose configuration for modularity
This commit is contained in:
commit
295ebb4f25
4 changed files with 33 additions and 13 deletions
14
README.md
14
README.md
|
@ -79,7 +79,19 @@ If you don't have Ollama installed yet, you can use the provided Docker Compose
|
|||
docker compose up -d --build
|
||||
```
|
||||
|
||||
This command will install both Ollama and Ollama Web UI on your system. Ensure to modify the `compose.yaml` file for GPU support and Exposing Ollama API outside the container stack if needed.
|
||||
This command will install both Ollama and Ollama Web UI on your system.
|
||||
|
||||
#### Enable GPU
|
||||
Use the additional Docker Compose file designed to enable GPU support by running the following command:
|
||||
```bash
|
||||
docker compose -f docker-compose.yml -f docker-compose.gpu.yml up -d --build
|
||||
```
|
||||
|
||||
#### Expose Ollama API outside the container stack
|
||||
Deploy the service with an additional Docker Compose file designed for API exposure:
|
||||
```bash
|
||||
docker compose -f docker-compose.yml -f docker-compose.api.yml up -d --build
|
||||
```
|
||||
|
||||
### Installing Ollama Web UI Only
|
||||
|
||||
|
|
7
docker-compose.api.yml
Normal file
7
docker-compose.api.yml
Normal file
|
@ -0,0 +1,7 @@
|
|||
version: '3.6'
|
||||
|
||||
services:
|
||||
ollama:
|
||||
# Expose Ollama API outside the container stack
|
||||
ports:
|
||||
- 11434:11434
|
13
docker-compose.gpu.yml
Normal file
13
docker-compose.gpu.yml
Normal file
|
@ -0,0 +1,13 @@
|
|||
version: '3.6'
|
||||
|
||||
services:
|
||||
ollama:
|
||||
# GPU support
|
||||
deploy:
|
||||
resources:
|
||||
reservations:
|
||||
devices:
|
||||
- driver: nvidia
|
||||
count: 1
|
||||
capabilities:
|
||||
- gpu
|
|
@ -2,20 +2,8 @@ version: '3.6'
|
|||
|
||||
services:
|
||||
ollama:
|
||||
# Uncomment below for GPU support
|
||||
# deploy:
|
||||
# resources:
|
||||
# reservations:
|
||||
# devices:
|
||||
# - driver: nvidia
|
||||
# count: 1
|
||||
# capabilities:
|
||||
# - gpu
|
||||
volumes:
|
||||
- ollama:/root/.ollama
|
||||
# Uncomment below to expose Ollama API outside the container stack
|
||||
# ports:
|
||||
# - 11434:11434
|
||||
container_name: ollama
|
||||
pull_policy: always
|
||||
tty: true
|
||||
|
|
Loading…
Reference in a new issue