Skip to content

Commit 747b464

Browse files
committed
Added CPU only version
1 parent 467e1c9 commit 747b464

5 files changed

Lines changed: 46 additions & 6 deletions

File tree

dappnode_package.json

Lines changed: 13 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,17 +1,28 @@
11
{
22
"name": "ollama-openwebui.dnp.dappnode.eth",
33
"version": "0.1.0",
4+
"upstream": [
5+
{
6+
"repo": "ollama/ollama",
7+
"version": "v0.12.9",
8+
"arg": "OLLAMA_VERSION"
9+
},
10+
{
11+
"repo": "open-webui/open-webui",
12+
"version": "v0.6.34",
13+
"arg": "WEBUI_VERSION"
14+
}
15+
],
416
"shortDescription": "Local AI chat interface with Ollama and Open WebUI",
517
"description": "Run large language models locally on your DAppNode with GPU acceleration. This package combines Ollama (with AMD ROCm support for GPU inference) and Open WebUI (a ChatGPT-like interface) to provide a complete local AI solution.\n\n**Features:**\n- AMD GPU acceleration via ROCm\n- ChatGPT-like web interface\n- Complete privacy - all processing stays local\n- Support for multiple LLM models (Llama, Mistral, CodeLlama, etc.)\n\n**Requirements:**\n- AMD GPU with ROCm support\n- At least 8GB RAM (16GB+ recommended)\n- Sufficient storage for models (10GB+ recommended)\n\nAccess Open WebUI at http://ollama-openwebui.dappnode:8080",
618
"type": "service",
719
"mainService": "webui",
8-
"author": "DAppNode Community",
20+
"author": "DAppNode Association <admin@dappnode.io> (https://github.com/dappnode)",
921
"license": "GPL-3.0",
1022
"categories": [
1123
"AI"
1224
],
1325
"links": {
14-
"homepage": "https://github.com/open-webui/open-webui",
1526
"ui": "http://ollama-openwebui.dappnode:8080"
1627
},
1728
"architectures": [

package_variants/amd/dappnode_package.json

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,8 @@
22
"name": "ollama-amd-openwebui.dnp.dappnode.eth",
33
"version": "0.1.0",
44
"links": {
5-
"homepage": "https://github.com/open-webui/open-webui",
65
"ui": "http://ollama-openwebui.dappnode:8080"
7-
}
6+
},
7+
"description": "Run large language models locally on your DAppNode with GPU acceleration. This package combines Ollama (with AMD ROCm support for GPU inference) and Open WebUI (a ChatGPT-like interface) to provide a complete local AI solution.\n\n**Features:**\n- AMD GPU acceleration via ROCm\n- ChatGPT-like web interface\n- Complete privacy - all processing stays local\n- Support for multiple LLM models (Llama, Mistral, CodeLlama, etc.)\n\n**Requirements:**\n- AMD GPU with ROCm support\n- At least 8GB RAM (16GB+ recommended)\n- Sufficient storage for models (10GB+ recommended)\n"
8+
89
}
Lines changed: 12 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,12 @@
1+
{
2+
"name": "ollama-cpu-openwebui.dnp.dappnode.eth",
3+
"version": "0.1.0",
4+
"links": {
5+
"ui": "http://ollama-cpu-openwebui.dappnode:8080"
6+
},
7+
"architectures": [
8+
"linux/amd64",
9+
"linux/arm64"
10+
],
11+
"description": "Run large language models locally on your DAppNode. This package combines Ollama and Open WebUI (a ChatGPT-like interface) to provide a complete local AI solution.\n\n**Features:**\n- CPU acceleration for inference\n- ChatGPT-like web interface\n- Complete privacy - all processing stays local\n- Support for multiple LLM models (Llama, Mistral, CodeLlama, etc.)\n\n**Requirements:**\n- At least 8GB RAM (16GB+ recommended)\n- Sufficient storage for models (10GB+ recommended)\n"
12+
}
Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,15 @@
1+
version: "3.5"
2+
3+
services:
4+
webui:
5+
build:
6+
context: webui
7+
args:
8+
WEBUI_VERSION: v0.6.34
9+
container_name: openwebui-cpu.ollama-openwebui.dnp.dappnode.eth
10+
ollama:
11+
build:
12+
context: ollama
13+
args:
14+
OLLAMA_VERSION: 0.12.6
15+
container_name: ollama-cpu.ollama-openwebui.dnp.dappnode.eth

package_variants/nvidia/dappnode_package.json

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,11 +2,12 @@
22
"name": "ollama-nvidia-openwebui.dnp.dappnode.eth",
33
"version": "0.1.0",
44
"links": {
5-
"homepage": "https://github.com/open-webui/open-webui",
65
"ui": "http://ollama-nvidia-openwebui.dappnode:8080"
76
},
87
"architectures": [
98
"linux/amd64",
109
"linux/arm64"
11-
]
10+
],
11+
"description": "Run large language models locally on your DAppNode with GPU acceleration. This package combines Ollama (with NVIDIA GPU support for GPU inference) and Open WebUI (a ChatGPT-like interface) to provide a complete local AI solution.\n\n**Features:**\n- NVIDIA GPU acceleration\n- ChatGPT-like web interface\n- Complete privacy - all processing stays local\n- Support for multiple LLM models (Llama, Mistral, CodeLlama, etc.)\n\n**Requirements:**\n- NVIDIA GPU with CUDA support\n- At least 8GB RAM (16GB+ recommended)\n- Sufficient storage for models (10GB+ recommended)\n"
12+
1213
}

0 commit comments

Comments
 (0)