From d0b66496a112d01cf727009bb071a87625af498d Mon Sep 17 00:00:00 2001 From: Richie Cahill Date: Sat, 3 Jan 2026 22:19:41 -0500 Subject: [PATCH] Major llm rework --- systems/bob/llms.nix | 54 +++++++++++--------- systems/jeeves/services/llms.nix | 63 ++++++++++++------------ systems/jeeves/services/open_webui.nix | 5 +- systems/rhapsody-in-green/default.nix | 3 +- systems/rhapsody-in-green/llms.nix | 2 + systems/rhapsody-in-green/open_webui.nix | 12 +++++ 6 files changed, 78 insertions(+), 61 deletions(-) create mode 100644 systems/rhapsody-in-green/open_webui.nix diff --git a/systems/bob/llms.nix b/systems/bob/llms.nix index ff18998..306278b 100644 --- a/systems/bob/llms.nix +++ b/systems/bob/llms.nix @@ -1,27 +1,33 @@ { - services = { - ollama = { - user = "ollama"; - enable = true; - loadModels = [ - "codellama:7b" - "deepseek-r1:8b" - "deepseek-r1:14b" - "deepseek-r1:32b" - "llama3.2:3b" - "llama2-uncensored:7b" - "mistral-nemo:12b" - "dolphin-mixtral:8x7b" - "qwq:32b" - "Qihoo360-Light-R1-32B" - ]; - models = "/zfs/models"; - openFirewall = true; - }; - # open-webui = { - # enable = true; - # openFirewall = true; - # host = "0.0.0.0"; - # }; + services.ollama = { + user = "ollama"; + host = "0.0.0.0"; + enable = true; + + syncModels = true; + loadModels = [ + "codellama:7b" + "deepscaler:1.5b" + "deepseek-r1:8b" + "deepseek-r1:14b" + "deepseek-r1:32b" + "devstral-small-2:24b" + "functiongemma:270m" + "gemma3:12b" + "gemma3:27b" + "gpt-oss:20b" + "llama3.1:8b" + "llama3.2:1b" + "llama3.2:3b" + "magistral:24b" + "ministral-3:14b" + "nemotron-3-nano:30b" + "qwen3-coder:30b" + "qwen3-vl:32b" + "qwen3:14b" + "qwen3:30b" + ]; + models = "/zfs/models"; + openFirewall = true; }; } diff --git a/systems/jeeves/services/llms.nix b/systems/jeeves/services/llms.nix index 1d441d4..3a078db 100644 --- a/systems/jeeves/services/llms.nix +++ b/systems/jeeves/services/llms.nix @@ -2,38 +2,37 @@ let vars = import ../vars.nix; in { - services = { - ollama = { - user = "ollama"; - enable = true; - host = "0.0.0.0"; - loadModels = [ - "codellama:7b" - "deepscaler:1.5b" - "deepseek-r1:14b" - "deepseek-r1:32b" - "deepseek-r1:8b" - "devstral-small-2:24b" - "functiongemma:270m" - "gemma3:12b" - "gemma3:27b" - "gpt-oss:120b" - "gpt-oss:20b" - "llama3.1:70b" - "llama3.1:8b" - "llama3.2:1b" - "llama3.2:3b" - "magistral:24b" - "ministral-3:14b" - "nemotron-3-nano:30b" - "qwen3-coder:30b" - "qwen3-vl:32b" - "qwen3:14b" - "qwen3:30b" - ]; - models = vars.ollama; - openFirewall = true; - }; + services.ollama = { + user = "ollama"; + enable = true; + host = "0.0.0.0"; + syncModels = true; + loadModels = [ + "codellama:7b" + "deepscaler:1.5b" + "deepseek-r1:14b" + "deepseek-r1:32b" + "deepseek-r1:8b" + "devstral-small-2:24b" + "functiongemma:270m" + "gemma3:12b" + "gemma3:27b" + "gpt-oss:120b" + "gpt-oss:20b" + "llama3.1:70b" + "llama3.1:8b" + "llama3.2:1b" + "llama3.2:3b" + "magistral:24b" + "ministral-3:14b" + "nemotron-3-nano:30b" + "qwen3-coder:30b" + "qwen3-vl:32b" + "qwen3:14b" + "qwen3:30b" + ]; + models = vars.ollama; + openFirewall = true; }; systemd.services = { ollama.serviceConfig = { diff --git a/systems/jeeves/services/open_webui.nix b/systems/jeeves/services/open_webui.nix index 663e633..152eeb6 100644 --- a/systems/jeeves/services/open_webui.nix +++ b/systems/jeeves/services/open_webui.nix @@ -1,10 +1,7 @@ -let - vars = import ../vars.nix; -in { services.open-webui = { - stateDir = "${vars.services}/open_webui/"; enable = true; + host = "0.0.0.0"; openFirewall = true; environment = { ANONYMIZED_TELEMETRY = "False"; diff --git a/systems/rhapsody-in-green/default.nix b/systems/rhapsody-in-green/default.nix index 73c2a94..4ff7ffc 100644 --- a/systems/rhapsody-in-green/default.nix +++ b/systems/rhapsody-in-green/default.nix @@ -12,8 +12,9 @@ "${inputs.self}/common/optional/zerotier.nix" ./hardware.nix ./llms.nix - ./syncthing.nix + ./open_webui.nix ./qmk.nix + ./syncthing.nix inputs.nixos-hardware.nixosModules.framework-13-7040-amd ]; diff --git a/systems/rhapsody-in-green/llms.nix b/systems/rhapsody-in-green/llms.nix index dd3115b..d90b067 100644 --- a/systems/rhapsody-in-green/llms.nix +++ b/systems/rhapsody-in-green/llms.nix @@ -3,8 +3,10 @@ user = "ollama"; enable = true; host = "127.0.0.1"; + syncModels = true; loadModels = [ "codellama:7b" + "deepscaler:1.5b" "deepseek-r1:14b" "deepseek-r1:32b" "deepseek-r1:8b" diff --git a/systems/rhapsody-in-green/open_webui.nix b/systems/rhapsody-in-green/open_webui.nix new file mode 100644 index 0000000..827543b --- /dev/null +++ b/systems/rhapsody-in-green/open_webui.nix @@ -0,0 +1,12 @@ +{ + services.open-webui = { + enable = true; + environment = { + ANONYMIZED_TELEMETRY = "False"; + DO_NOT_TRACK = "True"; + SCARF_NO_ANALYTICS = "True"; + OLLAMA_API_BASE_URL = "http://127.0.0.1:11434"; + WEBUI_AUTH = "False"; + }; + }; +}