From 1569a4372357cea0993e2d0e88f258ef8a6abe2a Mon Sep 17 00:00:00 2001 From: "renovate[bot]" <29139614+renovate[bot]@users.noreply.github.com> Date: Thu, 13 Jun 2024 22:24:22 +0000 Subject: [PATCH] chore(deps): update ollama/ollama docker tag to v0.1.44 (#3779) --- apps/ollama-amd/config.json | 4 ++-- apps/ollama-amd/docker-compose.yml | 2 +- apps/ollama-cpu/config.json | 4 ++-- apps/ollama-cpu/docker-compose.yml | 2 +- apps/ollama-nvidia/config.json | 4 ++-- apps/ollama-nvidia/docker-compose.yml | 2 +- 6 files changed, 9 insertions(+), 9 deletions(-) diff --git a/apps/ollama-amd/config.json b/apps/ollama-amd/config.json index 072c6a09..c2eab909 100755 --- a/apps/ollama-amd/config.json +++ b/apps/ollama-amd/config.json @@ -5,8 +5,8 @@ "exposable": true, "port": 11434, "id": "ollama-amd", - "tipi_version": 7, - "version": "0.1.43-rocm", + "tipi_version": 8, + "version": "0.1.44-rocm", "categories": ["ai"], "description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.", "short_desc": "LLMs inference server with OpenAI compatible API", diff --git a/apps/ollama-amd/docker-compose.yml b/apps/ollama-amd/docker-compose.yml index 8274cea6..2ab49d79 100755 --- a/apps/ollama-amd/docker-compose.yml +++ b/apps/ollama-amd/docker-compose.yml @@ -2,7 +2,7 @@ version: '3.7' services: ollama-amd: - image: ollama/ollama:0.1.43-rocm + image: ollama/ollama:0.1.44-rocm restart: unless-stopped container_name: ollama-amd environment: diff --git a/apps/ollama-cpu/config.json b/apps/ollama-cpu/config.json index 43c47d63..60e0e151 100755 --- a/apps/ollama-cpu/config.json +++ b/apps/ollama-cpu/config.json @@ -5,8 +5,8 @@ "exposable": true, "port": 11436, "id": "ollama-cpu", - "tipi_version": 7, - "version": "0.1.43", + "tipi_version": 8, + "version": "0.1.44", "categories": ["ai"], "description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.", "short_desc": "LLMs inference server with OpenAI compatible API", diff --git a/apps/ollama-cpu/docker-compose.yml b/apps/ollama-cpu/docker-compose.yml index b333effb..2ffe3794 100755 --- a/apps/ollama-cpu/docker-compose.yml +++ b/apps/ollama-cpu/docker-compose.yml @@ -2,7 +2,7 @@ version: '3.7' services: ollama-cpu: - image: ollama/ollama:0.1.43 + image: ollama/ollama:0.1.44 restart: unless-stopped container_name: ollama-cpu ports: diff --git a/apps/ollama-nvidia/config.json b/apps/ollama-nvidia/config.json index cec01d7d..9ca93ac5 100755 --- a/apps/ollama-nvidia/config.json +++ b/apps/ollama-nvidia/config.json @@ -5,8 +5,8 @@ "exposable": true, "port": 11435, "id": "ollama-nvidia", - "tipi_version": 7, - "version": "0.1.43", + "tipi_version": 8, + "version": "0.1.44", "categories": ["ai"], "description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.", "short_desc": "LLMs inference server with OpenAI compatible API", diff --git a/apps/ollama-nvidia/docker-compose.yml b/apps/ollama-nvidia/docker-compose.yml index cb01db94..6faee98d 100755 --- a/apps/ollama-nvidia/docker-compose.yml +++ b/apps/ollama-nvidia/docker-compose.yml @@ -2,7 +2,7 @@ version: '3.7' services: ollama-nvidia: - image: ollama/ollama:0.1.43 + image: ollama/ollama:0.1.44 restart: unless-stopped container_name: ollama-nvidia ports: