modified start commands for open-webui and localai to optimize GPU usage
This commit is contained in:
parent
2c6cb28b56
commit
a07cbe73a3
@ -195,31 +195,6 @@ clean_container() {
|
|||||||
# Handle INSTALL (no --update passed) ------------------------------------------------------------------------INSTALL SECTION BELOW
|
# Handle INSTALL (no --update passed) ------------------------------------------------------------------------INSTALL SECTION BELOW
|
||||||
|
|
||||||
|
|
||||||
# open-webui INSTALL
|
|
||||||
if [[ "$1" != "--update" ]]; then
|
|
||||||
if check_container_running "open-webui"; then
|
|
||||||
echo "✅ Skipping Open WebUI + Ollama install (already running)."
|
|
||||||
else
|
|
||||||
read -p "➕ Install Open WebUI + Ollama? [y/N]: " openui
|
|
||||||
if [[ "$openui" =~ ^[Yy]$ ]]; then
|
|
||||||
echo "🧠 Setting up Open WebUI + Ollama..."
|
|
||||||
docker stop open-webui 2>/dev/null || true
|
|
||||||
docker rm open-webui 2>/dev/null || true
|
|
||||||
docker run -d -p 3000:8080 --gpus all \
|
|
||||||
-e OPENAI_API_BASE_URL=http://pipelines:9099 \
|
|
||||||
-e OPENAI_API_KEY=0p3n-w3bu! \
|
|
||||||
-e OLLAMA_MAX_LOADED_MODELS=2 \
|
|
||||||
-v ollama:/root/.ollama \
|
|
||||||
-v open-webui:/app/backend/data \
|
|
||||||
--name open-webui \
|
|
||||||
--network ai-stack-net \
|
|
||||||
--restart always \
|
|
||||||
ghcr.io/open-webui/open-webui:ollama
|
|
||||||
echo "🌐 Open WebUI running at http://localhost:3000"
|
|
||||||
fi
|
|
||||||
fi
|
|
||||||
fi
|
|
||||||
|
|
||||||
|
|
||||||
# jupyterlab INSTALL
|
# jupyterlab INSTALL
|
||||||
if [[ "$1" != "--update" ]]; then
|
if [[ "$1" != "--update" ]]; then
|
||||||
@ -489,6 +464,32 @@ if [[ "$1" != "--update" ]]; then
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
|
|
||||||
|
# open-webui INSTALL
|
||||||
|
if [[ "$1" != "--update" ]]; then
|
||||||
|
if check_container_running "open-webui"; then
|
||||||
|
echo "✅ Skipping Open WebUI + Ollama install (already running)."
|
||||||
|
else
|
||||||
|
read -p "➕ Install Open WebUI + Ollama? [y/N]: " openui
|
||||||
|
if [[ "$openui" =~ ^[Yy]$ ]]; then
|
||||||
|
echo "🧠 Setting up Open WebUI + Ollama..."
|
||||||
|
docker stop open-webui 2>/dev/null || true
|
||||||
|
docker rm open-webui 2>/dev/null || true
|
||||||
|
docker run -d -p 3000:8080 --gpus all \
|
||||||
|
-e OPENAI_API_BASE_URL=http://pipelines:9099 \
|
||||||
|
-e OPENAI_API_KEY=0p3n-w3bu! \
|
||||||
|
-e OLLAMA_MAX_LOADED_MODELS=2 \
|
||||||
|
-v ollama:/root/.ollama \
|
||||||
|
-v open-webui:/app/backend/data \
|
||||||
|
--name open-webui \
|
||||||
|
--network ai-stack-net \
|
||||||
|
--restart always \
|
||||||
|
ghcr.io/open-webui/open-webui:ollama
|
||||||
|
echo "🌐 Open WebUI running at http://localhost:3000"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
|
||||||
# pipelines INSTALL
|
# pipelines INSTALL
|
||||||
if [[ "$1" != "--update" ]]; then
|
if [[ "$1" != "--update" ]]; then
|
||||||
if check_container_running "pipelines"; then
|
if check_container_running "pipelines"; then
|
||||||
|
Loading…
x
Reference in New Issue
Block a user