chore(deps): update ollama/ollama docker tag to v0.1.46 (#3864)
This commit is contained in:
parent
8ef188a6f9
commit
0ded42c6b6
|
@ -5,8 +5,8 @@
|
|||
"exposable": true,
|
||||
"port": 11434,
|
||||
"id": "ollama-amd",
|
||||
"tipi_version": 9,
|
||||
"version": "0.1.45-rocm",
|
||||
"tipi_version": 10,
|
||||
"version": "0.1.46-rocm",
|
||||
"categories": ["ai"],
|
||||
"description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.",
|
||||
"short_desc": "LLMs inference server with OpenAI compatible API",
|
||||
|
|
|
@ -2,7 +2,7 @@ version: '3.7'
|
|||
|
||||
services:
|
||||
ollama-amd:
|
||||
image: ollama/ollama:0.1.45-rocm
|
||||
image: ollama/ollama:0.1.46-rocm
|
||||
restart: unless-stopped
|
||||
container_name: ollama-amd
|
||||
environment:
|
||||
|
|
|
@ -5,8 +5,8 @@
|
|||
"exposable": true,
|
||||
"port": 11436,
|
||||
"id": "ollama-cpu",
|
||||
"tipi_version": 9,
|
||||
"version": "0.1.45",
|
||||
"tipi_version": 10,
|
||||
"version": "0.1.46",
|
||||
"categories": ["ai"],
|
||||
"description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.",
|
||||
"short_desc": "LLMs inference server with OpenAI compatible API",
|
||||
|
|
|
@ -2,7 +2,7 @@ version: '3.7'
|
|||
|
||||
services:
|
||||
ollama-cpu:
|
||||
image: ollama/ollama:0.1.45
|
||||
image: ollama/ollama:0.1.46
|
||||
restart: unless-stopped
|
||||
container_name: ollama-cpu
|
||||
ports:
|
||||
|
|
|
@ -5,8 +5,8 @@
|
|||
"exposable": true,
|
||||
"port": 11435,
|
||||
"id": "ollama-nvidia",
|
||||
"tipi_version": 9,
|
||||
"version": "0.1.45",
|
||||
"tipi_version": 10,
|
||||
"version": "0.1.46",
|
||||
"categories": ["ai"],
|
||||
"description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.",
|
||||
"short_desc": "LLMs inference server with OpenAI compatible API",
|
||||
|
|
|
@ -2,7 +2,7 @@ version: '3.7'
|
|||
|
||||
services:
|
||||
ollama-nvidia:
|
||||
image: ollama/ollama:0.1.45
|
||||
image: ollama/ollama:0.1.46
|
||||
restart: unless-stopped
|
||||
container_name: ollama-nvidia
|
||||
ports:
|
||||
|
|
Loading…
Reference in New Issue
Block a user