From e444bf368e93a4edaac37b00f2a87f2bccd92843 Mon Sep 17 00:00:00 2001 From: hcastc00 Date: Sat, 14 Mar 2026 04:33:18 +0000 Subject: [PATCH] bump ollama/ollama to v0.18.0 --- dappnode_package.json | 12 ++++-------- docker-compose.yml | 2 +- package_variants/cpu/dappnode_package.json | 7 ++----- package_variants/nvidia/dappnode_package.json | 7 ++----- 4 files changed, 9 insertions(+), 19 deletions(-) diff --git a/dappnode_package.json b/dappnode_package.json index 0c1c46a..ee6da71 100644 --- a/dappnode_package.json +++ b/dappnode_package.json @@ -4,7 +4,7 @@ "upstream": [ { "repo": "ollama/ollama", - "version": "v0.17.7", + "version": "v0.18.0", "arg": "OLLAMA_VERSION" } ], @@ -14,13 +14,9 @@ "type": "service", "author": "DAppNode Association (https://github.com/dappnode)", "license": "GPL-3.0", - "categories": [ - "AI" - ], + "categories": ["AI"], "links": { "Models library": "https://ollama.com/library" }, - "architectures": [ - "linux/amd64" - ] -} \ No newline at end of file + "architectures": ["linux/amd64"] +} diff --git a/docker-compose.yml b/docker-compose.yml index dc9bc61..9179e60 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -4,7 +4,7 @@ services: build: context: ollama args: - OLLAMA_VERSION: v0.17.7 + OLLAMA_VERSION: v0.18.0 container_name: ollama.ollama.dnp.dappnode.eth volumes: - ollama:/root/.ollama diff --git a/package_variants/cpu/dappnode_package.json b/package_variants/cpu/dappnode_package.json index 39898e2..30fd75c 100644 --- a/package_variants/cpu/dappnode_package.json +++ b/package_variants/cpu/dappnode_package.json @@ -5,9 +5,6 @@ "Model Manager": "http://ui.ollama-cpu.dappnode", "api": "http://ollama-cpu.dappnode:11434" }, - "architectures": [ - "linux/amd64", - "linux/arm64" - ], + "architectures": ["linux/amd64", "linux/arm64"], "description": "Run large language models locally on your DAppNode. Ollama inference engine with CPU-only support.\n\n**Features:**\n- CPU acceleration for inference\n- Complete privacy - all processing stays local\n- Support for multiple LLM models (Llama, Mistral, CodeLlama, etc.)\n\n**Requirements:**\n- At least 8GB RAM (16GB+ recommended)\n- Sufficient storage for models (10GB+ recommended)\n" -} \ No newline at end of file +} diff --git a/package_variants/nvidia/dappnode_package.json b/package_variants/nvidia/dappnode_package.json index adb55fd..87042ea 100644 --- a/package_variants/nvidia/dappnode_package.json +++ b/package_variants/nvidia/dappnode_package.json @@ -5,9 +5,6 @@ "Model Manager": "http://ui.ollama-nvidia.dappnode", "api": "http://ollama-nvidia.dappnode:11434" }, - "architectures": [ - "linux/amd64", - "linux/arm64" - ], + "architectures": ["linux/amd64", "linux/arm64"], "description": "Run large language models locally on your DAppNode with GPU acceleration. Ollama with NVIDIA GPU support for GPU inference.\n\n**Features:**\n- NVIDIA GPU acceleration\n- Complete privacy - all processing stays local\n- Support for multiple LLM models (Llama, Mistral, CodeLlama, etc.)\n\n**Requirements:**\n- NVIDIA GPU with CUDA support\n- At least 8GB RAM (16GB+ recommended)\n- Sufficient storage for models (10GB+ recommended)\n" -} \ No newline at end of file +}