From 883a2a259c3b0fd33e0163ea1b1f15b1f5aee5e4 Mon Sep 17 00:00:00 2001 From: "renovate[bot]" <29139614+renovate[bot]@users.noreply.github.com> Date: Sun, 12 May 2024 00:30:03 +0000 Subject: [PATCH] chore(deps): update ollama/ollama docker tag to v0.1.37 (#3448) --- apps/ollama-amd/config.json | 13 +++++++++---- apps/ollama-amd/docker-compose.yml | 2 +- apps/ollama-cpu/config.json | 13 +++++++++---- apps/ollama-cpu/docker-compose.yml | 2 +- apps/ollama-nvidia/config.json | 13 +++++++++---- apps/ollama-nvidia/docker-compose.yml | 2 +- 6 files changed, 30 insertions(+), 15 deletions(-) diff --git a/apps/ollama-amd/config.json b/apps/ollama-amd/config.json index d6ba80aa..b0493d93 100755 --- a/apps/ollama-amd/config.json +++ b/apps/ollama-amd/config.json @@ -5,14 +5,19 @@ "exposable": true, "port": 11434, "id": "ollama-amd", - "tipi_version": 1, - "version": "0.1.33", - "categories": ["ai"], + "tipi_version": 2, + "version": "0.1.37-rocm", + "categories": [ + "ai" + ], "description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.", "short_desc": "LLMs inference server with OpenAI compatible API", "author": "ollama", "source": "https://github.com/ollama/ollama", "website": "https://ollama.com", "form_fields": [], - "supported_architectures": ["arm64", "amd64"] + "supported_architectures": [ + "arm64", + "amd64" + ] } diff --git a/apps/ollama-amd/docker-compose.yml b/apps/ollama-amd/docker-compose.yml index de618415..3a794977 100755 --- a/apps/ollama-amd/docker-compose.yml +++ b/apps/ollama-amd/docker-compose.yml @@ -2,7 +2,7 @@ version: '3.7' services: ollama-amd: - image: ollama/ollama:0.1.33-rocm + image: ollama/ollama:0.1.37-rocm restart: unless-stopped container_name: ollama-amd environment: diff --git a/apps/ollama-cpu/config.json b/apps/ollama-cpu/config.json index 73981ea8..0a27653c 100755 --- a/apps/ollama-cpu/config.json +++ b/apps/ollama-cpu/config.json @@ -5,14 +5,19 @@ "exposable": true, "port": 11436, "id": "ollama-cpu", - "tipi_version": 1, - "version": "0.1.33", - "categories": ["ai"], + "tipi_version": 2, + "version": "0.1.37", + "categories": [ + "ai" + ], "description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.", "short_desc": "LLMs inference server with OpenAI compatible API", "author": "ollama", "source": "https://github.com/ollama/ollama", "website": "https://ollama.com", "form_fields": [], - "supported_architectures": ["arm64", "amd64"] + "supported_architectures": [ + "arm64", + "amd64" + ] } diff --git a/apps/ollama-cpu/docker-compose.yml b/apps/ollama-cpu/docker-compose.yml index 8f31e896..77db3dde 100755 --- a/apps/ollama-cpu/docker-compose.yml +++ b/apps/ollama-cpu/docker-compose.yml @@ -2,7 +2,7 @@ version: '3.7' services: ollama-cpu: - image: ollama/ollama:0.1.33 + image: ollama/ollama:0.1.37 restart: unless-stopped container_name: ollama-cpu ports: diff --git a/apps/ollama-nvidia/config.json b/apps/ollama-nvidia/config.json index e6474a5e..43cd9fcc 100755 --- a/apps/ollama-nvidia/config.json +++ b/apps/ollama-nvidia/config.json @@ -5,14 +5,19 @@ "exposable": true, "port": 11435, "id": "ollama-nvidia", - "tipi_version": 1, - "version": "0.1.33", - "categories": ["ai"], + "tipi_version": 2, + "version": "0.1.37", + "categories": [ + "ai" + ], "description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.", "short_desc": "LLMs inference server with OpenAI compatible API", "author": "ollama", "source": "https://github.com/ollama/ollama", "website": "https://ollama.com", "form_fields": [], - "supported_architectures": ["arm64", "amd64"] + "supported_architectures": [ + "arm64", + "amd64" + ] } diff --git a/apps/ollama-nvidia/docker-compose.yml b/apps/ollama-nvidia/docker-compose.yml index 842c11d1..22912ccb 100755 --- a/apps/ollama-nvidia/docker-compose.yml +++ b/apps/ollama-nvidia/docker-compose.yml @@ -2,7 +2,7 @@ version: '3.7' services: ollama-nvidia: - image: ollama/ollama:0.1.33 + image: ollama/ollama:0.1.37 restart: unless-stopped container_name: ollama-nvidia ports: