diff --git a/home/modules/emacs/doom/config.el b/home/modules/emacs/doom/config.el index f008c5a..c71e8ad 100644 --- a/home/modules/emacs/doom/config.el +++ b/home/modules/emacs/doom/config.el @@ -83,12 +83,16 @@ (setq! gptel-api-key (my/get-rbw-password "openai-api-key-chatgpt-el") gptel-default-mode 'org-mode gptel-use-tools t - gptel-confirm-tool-calls 'always) + gptel-confirm-tool-calls 'always + gptel-include-reasoning 'ignore + gptel-model "qwen3:30b") - (gptel-make-ollama "Ollama-Local" - :host "localhost:11434" - :stream t - :models '(deepseek-r1 deepseek-r1-fullctx qwen3 qwen3-coder llama3.1 qwen2.5-coder mistral-nemo gpt-oss)) + ;; Set default backend to be Ollama-Local + (setq! gptel-backend + (gptel-make-ollama "Ollama-Local" + :host "localhost:11434" + :stream t + :models '(deepseek-r1 deepseek-r1-fullctx qwen3:30b qwen3:4b llama3.1 qwen2.5-coder mistral-nemo gpt-oss))) ;; Define custom tools (gptel-make-tool diff --git a/machines/zix790prors/configuration.nix b/machines/zix790prors/configuration.nix index f0132d1..9140ac9 100644 --- a/machines/zix790prors/configuration.nix +++ b/machines/zix790prors/configuration.nix @@ -71,9 +71,11 @@ with lib; powerManagement.finegrained = false; }; - environment.systemPackages = [ - (pkgs.ollama.override { acceleration = "cuda"; }) - ]; + services.ollama = { + enable = true; + acceleration = "cuda"; + loadModels = [ "gpt-oss" "deepseek-r1" "qwen3:30b" ]; + }; # This option defines the first version of NixOS you have installed on this particular machine, # and is used to maintain compatibility with application data (e.g. databases) created on older NixOS versions.