chore(deps): update ollama/ollama docker tag to v0.1.38 (#3522)
This commit is contained in:
parent
7c97f7fded
commit
e27f30e1a5
|
@ -5,19 +5,14 @@
|
||||||
"exposable": true,
|
"exposable": true,
|
||||||
"port": 11434,
|
"port": 11434,
|
||||||
"id": "ollama-amd",
|
"id": "ollama-amd",
|
||||||
"tipi_version": 2,
|
"tipi_version": 3,
|
||||||
"version": "0.1.37-rocm",
|
"version": "0.1.38-rocm",
|
||||||
"categories": [
|
"categories": ["ai"],
|
||||||
"ai"
|
|
||||||
],
|
|
||||||
"description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.",
|
"description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.",
|
||||||
"short_desc": "LLMs inference server with OpenAI compatible API",
|
"short_desc": "LLMs inference server with OpenAI compatible API",
|
||||||
"author": "ollama",
|
"author": "ollama",
|
||||||
"source": "https://github.com/ollama/ollama",
|
"source": "https://github.com/ollama/ollama",
|
||||||
"website": "https://ollama.com",
|
"website": "https://ollama.com",
|
||||||
"form_fields": [],
|
"form_fields": [],
|
||||||
"supported_architectures": [
|
"supported_architectures": ["arm64", "amd64"]
|
||||||
"arm64",
|
|
||||||
"amd64"
|
|
||||||
]
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -2,7 +2,7 @@ version: '3.7'
|
||||||
|
|
||||||
services:
|
services:
|
||||||
ollama-amd:
|
ollama-amd:
|
||||||
image: ollama/ollama:0.1.37-rocm
|
image: ollama/ollama:0.1.38-rocm
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
container_name: ollama-amd
|
container_name: ollama-amd
|
||||||
environment:
|
environment:
|
||||||
|
|
|
@ -5,19 +5,14 @@
|
||||||
"exposable": true,
|
"exposable": true,
|
||||||
"port": 11436,
|
"port": 11436,
|
||||||
"id": "ollama-cpu",
|
"id": "ollama-cpu",
|
||||||
"tipi_version": 2,
|
"tipi_version": 3,
|
||||||
"version": "0.1.37",
|
"version": "0.1.38",
|
||||||
"categories": [
|
"categories": ["ai"],
|
||||||
"ai"
|
|
||||||
],
|
|
||||||
"description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.",
|
"description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.",
|
||||||
"short_desc": "LLMs inference server with OpenAI compatible API",
|
"short_desc": "LLMs inference server with OpenAI compatible API",
|
||||||
"author": "ollama",
|
"author": "ollama",
|
||||||
"source": "https://github.com/ollama/ollama",
|
"source": "https://github.com/ollama/ollama",
|
||||||
"website": "https://ollama.com",
|
"website": "https://ollama.com",
|
||||||
"form_fields": [],
|
"form_fields": [],
|
||||||
"supported_architectures": [
|
"supported_architectures": ["arm64", "amd64"]
|
||||||
"arm64",
|
|
||||||
"amd64"
|
|
||||||
]
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -2,7 +2,7 @@ version: '3.7'
|
||||||
|
|
||||||
services:
|
services:
|
||||||
ollama-cpu:
|
ollama-cpu:
|
||||||
image: ollama/ollama:0.1.37
|
image: ollama/ollama:0.1.38
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
container_name: ollama-cpu
|
container_name: ollama-cpu
|
||||||
ports:
|
ports:
|
||||||
|
|
|
@ -5,19 +5,14 @@
|
||||||
"exposable": true,
|
"exposable": true,
|
||||||
"port": 11435,
|
"port": 11435,
|
||||||
"id": "ollama-nvidia",
|
"id": "ollama-nvidia",
|
||||||
"tipi_version": 2,
|
"tipi_version": 3,
|
||||||
"version": "0.1.37",
|
"version": "0.1.38",
|
||||||
"categories": [
|
"categories": ["ai"],
|
||||||
"ai"
|
|
||||||
],
|
|
||||||
"description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.",
|
"description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.",
|
||||||
"short_desc": "LLMs inference server with OpenAI compatible API",
|
"short_desc": "LLMs inference server with OpenAI compatible API",
|
||||||
"author": "ollama",
|
"author": "ollama",
|
||||||
"source": "https://github.com/ollama/ollama",
|
"source": "https://github.com/ollama/ollama",
|
||||||
"website": "https://ollama.com",
|
"website": "https://ollama.com",
|
||||||
"form_fields": [],
|
"form_fields": [],
|
||||||
"supported_architectures": [
|
"supported_architectures": ["arm64", "amd64"]
|
||||||
"arm64",
|
|
||||||
"amd64"
|
|
||||||
]
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -2,7 +2,7 @@ version: '3.7'
|
||||||
|
|
||||||
services:
|
services:
|
||||||
ollama-nvidia:
|
ollama-nvidia:
|
||||||
image: ollama/ollama:0.1.37
|
image: ollama/ollama:0.1.38
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
container_name: ollama-nvidia
|
container_name: ollama-nvidia
|
||||||
ports:
|
ports:
|
||||||
|
|
Loading…
Reference in New Issue
Block a user