Skip to content

Commit 4be772d

Browse files
authored
Merge pull request #23 from dappnode/dappnodebot/bump-upstream/ollama/ollama@v0.13.5,open-webui/open-webui@v0.6.43
bump ollama/ollama to v0.13.5, open-webui/open-webui to v0.6.43
2 parents ec6a9a8 + cafea31 commit 4be772d

File tree

5 files changed

+15
-36
lines changed

5 files changed

+15
-36
lines changed

dappnode_package.json

Lines changed: 5 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -4,12 +4,12 @@
44
"upstream": [
55
{
66
"repo": "ollama/ollama",
7-
"version": "v0.12.9",
7+
"version": "v0.13.5",
88
"arg": "OLLAMA_VERSION"
99
},
1010
{
1111
"repo": "open-webui/open-webui",
12-
"version": "v0.6.34",
12+
"version": "v0.6.43",
1313
"arg": "WEBUI_VERSION"
1414
}
1515
],
@@ -19,13 +19,9 @@
1919
"mainService": "webui",
2020
"author": "DAppNode Association <admin@dappnode.io> (https://github.com/dappnode)",
2121
"license": "GPL-3.0",
22-
"categories": [
23-
"AI"
24-
],
22+
"categories": ["AI"],
2523
"links": {
2624
"ui": "http://ollama-openwebui.dappnode:8080"
2725
},
28-
"architectures": [
29-
"linux/amd64"
30-
]
31-
}
26+
"architectures": ["linux/amd64"]
27+
}

docker-compose.yml

Lines changed: 3 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -1,37 +1,28 @@
11
version: "3.5"
2-
32
services:
43
webui:
54
build:
65
context: webui
76
container_name: openwebui.ollama-openwebui.dnp.dappnode.eth
87
environment:
9-
OLLAMA_BASE_URL: "http://ollama:11434"
8+
OLLAMA_BASE_URL: http://ollama:11434
109
WEBUI_AUTH: "true"
1110
volumes:
12-
- "webui:/app/backend/data"
11+
- webui:/app/backend/data
1312
restart: unless-stopped
1413
depends_on:
1514
- ollama
16-
1715
ollama:
1816
build:
1917
context: ollama
2018
container_name: ollama.ollama-openwebui.dnp.dappnode.eth
2119
volumes:
22-
- "ollama:/root/.ollama"
20+
- ollama:/root/.ollama
2321
restart: unless-stopped
2422
environment:
25-
# Show token‑throughput and other debug info in the container logs
2623
- OLLAMA_LOG_LEVEL=debug
27-
28-
# Enable the /metrics endpoint (Prometheus format)
2924
- OLLAMA_METRICS=1
30-
31-
# OPTIONAL – JSON‑formatted logs (easier to ship to Loki/Elastic)
3225
- OLLAMA_LOG_FORMAT=json
33-
34-
# OPTIONAL – Turn off outbound telemetry if you only want local metrics
3526
- OLLAMA_TELEMETRY=0
3627
volumes:
3728
ollama: {}

package_variants/amd/dappnode_package.json

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -5,5 +5,4 @@
55
"ui": "http://ollama-openwebui.dappnode:8080"
66
},
77
"description": "Run large language models locally on your DAppNode with GPU acceleration. This package combines Ollama (with AMD ROCm support for GPU inference) and Open WebUI (a ChatGPT-like interface) to provide a complete local AI solution.\n\n**Features:**\n- AMD GPU acceleration via ROCm\n- ChatGPT-like web interface\n- Complete privacy - all processing stays local\n- Support for multiple LLM models (Llama, Mistral, CodeLlama, etc.)\n\n**Requirements:**\n- AMD GPU with ROCm support\n- At least 8GB RAM (16GB+ recommended)\n- Sufficient storage for models (10GB+ recommended)\n"
8-
9-
}
8+
}
Lines changed: 3 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,9 @@
11
{
22
"name": "ollama-cpu-openwebui.dnp.dappnode.eth",
3-
"version": "0.1.0",
3+
"version": "0.1.1",
44
"links": {
55
"ui": "http://ollama-cpu-openwebui.dappnode:8080"
66
},
7-
"architectures": [
8-
"linux/amd64",
9-
"linux/arm64"
10-
],
7+
"architectures": ["linux/amd64", "linux/arm64"],
118
"description": "Run large language models locally on your DAppNode. This package combines Ollama and Open WebUI (a ChatGPT-like interface) to provide a complete local AI solution.\n\n**Features:**\n- CPU acceleration for inference\n- ChatGPT-like web interface\n- Complete privacy - all processing stays local\n- Support for multiple LLM models (Llama, Mistral, CodeLlama, etc.)\n\n**Requirements:**\n- At least 8GB RAM (16GB+ recommended)\n- Sufficient storage for models (10GB+ recommended)\n"
12-
}
9+
}
Lines changed: 3 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -1,13 +1,9 @@
11
{
22
"name": "ollama-nvidia-openwebui.dnp.dappnode.eth",
3-
"version": "0.1.0",
3+
"version": "0.1.1",
44
"links": {
55
"ui": "http://ollama-nvidia-openwebui.dappnode:8080"
66
},
7-
"architectures": [
8-
"linux/amd64",
9-
"linux/arm64"
10-
],
7+
"architectures": ["linux/amd64", "linux/arm64"],
118
"description": "Run large language models locally on your DAppNode with GPU acceleration. This package combines Ollama (with NVIDIA GPU support for GPU inference) and Open WebUI (a ChatGPT-like interface) to provide a complete local AI solution.\n\n**Features:**\n- NVIDIA GPU acceleration\n- ChatGPT-like web interface\n- Complete privacy - all processing stays local\n- Support for multiple LLM models (Llama, Mistral, CodeLlama, etc.)\n\n**Requirements:**\n- NVIDIA GPU with CUDA support\n- At least 8GB RAM (16GB+ recommended)\n- Sufficient storage for models (10GB+ recommended)\n"
12-
13-
}
9+
}

0 commit comments

Comments
 (0)