Ollama¶
Ollama makes it easy to get up and running with large language models locally.
Setup¶
SSH into the machine running HomelabOS and install a model like so:
docker exec -it ollama_ollama_1 bash
ollama run llama2
You can then make direct requests within the container using standard ollama interface, or hit the external CURL endpoint, like:
curl https://brain.t3family.net/api/chat -d '{
"model": "llama2",
"messages": [
{ "role": "user", "content": "why is the sky blue?" }
]
}'
Warning!¶
This endpoint is publicly accessible, with no rate limiting. It could lead to pain.
Access¶
It is available at https://brain.t3family.net/ or http://brain.t3family.net/
Security enable/disable https_only and auth¶
To enable https_only or auth set the service config to True
settings/config.yml
ollama:
https_only: True
auth: True
Last update:
October 8, 2025