{ pkgs, config, ... }: { services.ollama = { enable = true; loadModels = [ "llama3.2:3b" "phi4-reasoning:14b" "dolphin3:8b" "smallthinker:3b" "gemma3:4b" "gemma3:12b" "gemma3:27b" "deepcoder:14b" "qwen3:14b" "nomic-embed-text" ]; acceleration = "cuda"; }; services.open-webui = { enable = true; port = 8888; host = "127.0.0.1"; }; environment.systemPackages = with pkgs; [ oterm alpaca aichat fabric-ai aider-chat # tgpt # smartcat # nextjs-ollama-llm-ui # open-webui ]; }