diff --git a/home/home.nix b/home/home.nix index 0d17a3b..58f672b 100644 --- a/home/home.nix +++ b/home/home.nix @@ -1,5 +1,11 @@ { pkgs, spicetify-nix, ... }: +let + localFastModel = "qwen2.5-coder-num_ctx"; + localReasoningModel = "deepseek-r1-num_ctx"; + remoteFastModel = "deepseek/deepseek-chat-v3-0324:free"; # "qwen/qwen-2.5-coder-32b-instruct:free"; + remoteReasoningModel = "deepseek/deepseek-r1-0528:free"; +in { home = { stateVersion = "23.11"; @@ -44,6 +50,14 @@ external.enable = true; }; }; + shellAliases = { + aichat_reasoning_remote = "aichat --model openrouter:${remoteReasoningModel}"; + aichat_reasoning_local = "aichat --model ollama:${localReasoningModel}"; + aichat_fast_remote = "aichat --model openrouter:${remoteFastModel}"; + aichat_fast_local = "aichat --model ollama:${localFastModel}"; + codex_remote = "codex --provider openrouter --model ${remoteFastModel}"; + codex_local = "codex --provider ollama --model ${localFastModel}"; + }; }; starship = { enable = true; @@ -69,7 +83,7 @@ codex = { enable = true; settings = { - model = "qwen2.5-coder"; + model = "${localFastModel}"; provider = "ollama"; providers = { ollama = { @@ -88,7 +102,7 @@ aichat = { enable = true; settings = { - model = "ollama:qwen2.5-coder:latest"; + model = "ollama:${localFastModel}"; clients = [ { type = "openai-compatible"; @@ -96,12 +110,12 @@ api_base = "http://localhost:11434/v1"; models = [ { - name = "qwen2.5-coder:latest"; + name = "${localFastModel}"; supports_function_calling = true; supports_vision = true; } { - name = "deepseek-r1:latest"; + name = "${localReasoningModel}"; supports_function_calling = true; supports_vision = true; } @@ -113,12 +127,12 @@ api_base = "https://openrouter.ai/api/v1"; models = [ { - name = "qwen/qwen-2.5-coder-32b-instruct:free"; + name = "${remoteFastModel}"; supports_function_calling = true; supports_vision = true; } { - name = "deepseek/deepseek-r1-0528:free"; + name = "${remoteReasoningModel}"; supports_function_calling = true; supports_vision = true; }