scripts/service.ollama.md
2025-05-05 11:58:47 -06:00

83 lines
1.4 KiB
Markdown

---
tags:
- script
date: 2025-04-29
author: The Bunker Admin
---
Setting up the Ollama service, including a few starter models, as follows:
---
## Configure Ollama
> [!warning] Ollama on Network
> The following configuration will allow Ollama to be accessed on your local network. Exposing this endpoint can increase system exposure.
Create/edit the configuration file:
```
sudo mkdir -p /etc/ollama
sudo nano /etc/ollama/config
```
Add the following content:
```
{
"host": "0.0.0.0"
}
```
## Ollama System Service
```
sudo nano /etc/systemd/system/ollama.service
```
> [!note]
>
> For the following text, you can also just add the `Environment="OLLAMA_HOST=0.0.0.0"` itself and then your system should work fine.
```
[Unit]
Description=Ollama Service
After=network-online.target
[Service]
ExecStart=/usr/local/bin/ollama serve
User=ollama
Group=ollama
Restart=always
RestartSec=3
Environment="PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin:/snap/bin"
Environment="OLLAMA_HOST=0.0.0.0"
[Install]
WantedBy=default.target
```
```
sudo systemctl daemon-reload
sudo systemctl enable ollama
sudo systemctl start ollama
```
## Models
A series of models that all will run on the [[build.homelab]].
```
ollama pull gemma3:12b
```
```
ollama pull qwen3
```
```
ollama pull deepseek-r1
```
```
ollama pull mistral-small3.1
```
```
ollama pull llama3.2
```