From 5801e5f800c3c6b2acb442b65ec6c3416d12e06a Mon Sep 17 00:00:00 2001 From: crowetic Date: Mon, 19 May 2025 09:43:39 -0700 Subject: [PATCH] Fixed incorrect localai setup --- setup-ai-stack.sh | 136 +++++++++++++++++++++------------------------- 1 file changed, 62 insertions(+), 74 deletions(-) diff --git a/setup-ai-stack.sh b/setup-ai-stack.sh index 2d08ed3..f2a30e2 100644 --- a/setup-ai-stack.sh +++ b/setup-ai-stack.sh @@ -2,6 +2,8 @@ set -e mkdir -p ~/ai-stack +mkdir -p ~/ai-stack/localai/models +mkdir -p ~/ai-stack/localai/config LOG_FILE=~/ai-stack/setup.log exec > >(tee -a "$LOG_FILE") 2>&1 @@ -41,71 +43,45 @@ if [[ "$1" == "--update" ]]; then else echo "โš ๏ธ JupyterLab is not running, skipping update." fi - - # # Whisper (whisper.cpp) - # if check_container_running "whisper"; then - # echo "๐Ÿงผ Updating Whisper..." - # docker stop whisper - # docker rm whisper - # docker pull ggerganov/whisper.cpp:latest - # docker run -d \ - # --name whisper \ - # --network ai-stack-net \ - # --gpus all \ - # -v whisper_data:/app/data \ - # ggerganov/whisper.cpp:latest - # echo "โœ… Whisper updated." - # else - # echo "โš ๏ธ Whisper is not running, skipping update." - # fi - - # # Stable Diffusion (A1111) - # if check_container_running "stable-diffusion"; then - # echo "๐Ÿงผ Updating Stable Diffusion..." - # docker stop stable-diffusion - # docker rm stable-diffusion - # docker pull abraham-ai/automatic1111-webui:latest - # docker run -d \ - # --name stable-diffusion \ - # --gpus all \ - # -p 7860:7860 \ - # -v sd_models:/data \ - # abraham-ai/automatic1111-webui:latest - # echo "โœ… Stable Diffusion updated." - # else - # echo "โš ๏ธ Stable Diffusion is not running, skipping update." - # fi - - # echo "๐Ÿ” Updating ComfyUI..." - # docker stop comfyui 2>/dev/null || true - # docker rm comfyui 2>/dev/null || true - # docker pull ghcr.io/cmdr2/comfyui:latest - # docker run -d --gpus all \ - # --name comfyui \ - # --network ai-stack-net \ - # -p 8188:8188 \ - # -v comfyui_data:/workspace \ - # ghcr.io/cmdr2/comfyui:latest - # echo "โœ… ComfyUI updated and restarted." - # echo "๐Ÿ” Updating LocalAI..." - # docker stop localai 2>/dev/null || true - # docker rm localai 2>/dev/null || true - mkdir -p ~/ai-stack/localai + # ๐Ÿ” Updating LocalAI (GPU build) + if check_container_running "localai"; then + echo "๐Ÿง  Updating LocalAI..." + docker stop localai + docker rm localai + else + echo "โž• Installing LocalAI..." + fi - docker pull quay.io/go-skynet/local-ai:latest - docker run -d \ - --name localai \ - --network ai-stack-net \ - -p 8080:8080 \ - -v ~/ai-stack/localai:/models \ - -e MODELS_PATH=/models \ - -e ENABLE_OOLLAMA_BACKEND=true \ - -v /var/run/docker.sock:/var/run/docker.sock \ - quay.io/go-skynet/local-ai:latest + docker pull localai/localai:latest-aio-gpu-nvidia-cuda-12 + + # Ensure default models.yaml exists + MODELS_YAML=~/ai-stack/localai/config/models.yaml + if [ ! -f "$MODELS_YAML" ]; then + echo "๐Ÿ“„ Creating default models.yaml..." +cat > "$MODELS_YAML" </dev/null || true - docker rm localai 2>/dev/null || true + echo "๐Ÿ“„ Creating default models.yaml..." + cat > ~/ai-stack/localai/config/models.yaml <