Add option to install Ollama in Open WebUI LXC (#3991)
This commit is contained in:
parent
77b06ee3f9
commit
8a21f6e7f0
1 changed files with 28 additions and 0 deletions
|
@ -49,12 +49,40 @@ cp .env.example .env
|
||||||
cat <<EOF >/opt/open-webui/.env
|
cat <<EOF >/opt/open-webui/.env
|
||||||
ENV=prod
|
ENV=prod
|
||||||
ENABLE_OLLAMA_API=false
|
ENABLE_OLLAMA_API=false
|
||||||
|
OLLAMA_BASE_URL=http://0.0.0.0:11434
|
||||||
EOF
|
EOF
|
||||||
$STD npm install
|
$STD npm install
|
||||||
export NODE_OPTIONS="--max-old-space-size=4096"
|
export NODE_OPTIONS="--max-old-space-size=4096"
|
||||||
$STD npm run build
|
$STD npm run build
|
||||||
msg_ok "Installed Open WebUI"
|
msg_ok "Installed Open WebUI"
|
||||||
|
|
||||||
|
read -r -p "Would you like to add Ollama? <y/N> " prompt
|
||||||
|
if [[ ${prompt,,} =~ ^(y|yes)$ ]]; then
|
||||||
|
msg_info "Installing Ollama"
|
||||||
|
curl -fsSLO https://ollama.com/download/ollama-linux-amd64.tgz
|
||||||
|
tar -C /usr -xzf ollama-linux-amd64.tgz
|
||||||
|
rm -rf ollama-linux-amd64.tgz
|
||||||
|
cat <<EOF >/etc/systemd/system/ollama.service
|
||||||
|
[Unit]
|
||||||
|
Description=Ollama Service
|
||||||
|
After=network-online.target
|
||||||
|
|
||||||
|
[Service]
|
||||||
|
Type=exec
|
||||||
|
ExecStart=/usr/bin/ollama serve
|
||||||
|
Environment=HOME=$HOME
|
||||||
|
Environment=OLLAMA_HOST=0.0.0.0
|
||||||
|
Restart=always
|
||||||
|
RestartSec=3
|
||||||
|
|
||||||
|
[Install]
|
||||||
|
WantedBy=multi-user.target
|
||||||
|
EOF
|
||||||
|
systemctl enable -q --now ollama.service
|
||||||
|
sed -i 's/ENABLE_OLLAMA_API=false/ENABLE_OLLAMA_API=true/g' /opt/open-webui/.env
|
||||||
|
msg_ok "Installed Ollama"
|
||||||
|
fi
|
||||||
|
|
||||||
msg_info "Creating Service"
|
msg_info "Creating Service"
|
||||||
cat <<EOF >/etc/systemd/system/open-webui.service
|
cat <<EOF >/etc/systemd/system/open-webui.service
|
||||||
[Unit]
|
[Unit]
|
||||||
|
|
Loading…
Reference in a new issue