tangled
alpha
login
or
join now
devins.page
/
flake
0
fork
atom
configurations for my servers and desktops
nix
nixos
flake
dots
dotfiles
0
fork
atom
overview
issues
pulls
pipelines
fix: enable beets
devins.page
1 month ago
beb93ede
c83b39f3
+1
-26
2 changed files
expand all
collapse all
unified
split
hosts
bluepill
default.nix
modules
ollama.nix
+1
hosts/bluepill/default.nix
···
1
1
{pkgs, ...}: {
2
2
imports = [
3
3
./hardware.nix
4
4
+
./modules/beets.nix
4
5
./modules/blocky.nix
5
6
./modules/homepage.nix
6
7
./modules/immich.nix
-26
hosts/bluepill/modules/ollama.nix
···
1
1
-
{pkgs, ...}: {
2
2
-
services.ollama = {
3
3
-
enable = true;
4
4
-
loadModels = ["deepseek-r1:latest" "deepseek-v2:latest" "qwen3-coder:latest" "qwen3-vl:latest" "qwen3:latest" "gemma3:latest" "nemotron-3-nano:latest" "functiongemma:latest" "olmo-3:latest" "gtp-oss:latest" "llama3:latest" "devstral-small-2:latest" "mistral:latest" "mistral-nemo:latest"];
5
5
-
syncModels = true;
6
6
-
openFirewall = true;
7
7
-
};
8
8
-
9
9
-
virtualisation = {
10
10
-
docker.enable = true;
11
11
-
oci-containers = {
12
12
-
backend = "docker";
13
13
-
containers.open-webui = {
14
14
-
image = "ghcr.io/open-webui/open-webui:main";
15
15
-
extraOptions = ["--network=host"];
16
16
-
environment = {
17
17
-
OLLAMA_BASE_URL = "http://127.0.0.1:11434";
18
18
-
};
19
19
-
volumes = [
20
20
-
"/var/lib/open-webui:/app/backend/data"
21
21
-
];
22
22
-
autoStart = true;
23
23
-
};
24
24
-
};
25
25
-
};
26
26
-
}