diff --git a/home/apps/default.nix b/home/apps/default.nix index 95761a65..51a5e1ed 100644 --- a/home/apps/default.nix +++ b/home/apps/default.nix @@ -16,6 +16,7 @@ lib.optionalAttrs device.hasGui { ./gimp.nix # ./guitarix.nix ./hyprpicker.nix + ./ida.nix # ./jellyflix.nix # ./kicad.nix ./kitty.nix @@ -27,19 +28,19 @@ lib.optionalAttrs device.hasGui { # ./openscad.nix ./orcaslicer.nix # ./pcsx2.nix + ./prismlauncher.nix # ./rpcs3.nix # ./shadps4.nix ./slack.nix # ./thunderbird.nix # ./tsukimi.nix # ./vial.nix + ./vicinae.nix ./vlc.nix ./vscode.nix ./wezterm.nix ./zathura.nix ./zed.nix ./zen.nix - ./vicinae.nix - ./ida.nix ]; } diff --git a/home/apps/prismlauncher.nix b/home/apps/prismlauncher.nix new file mode 100644 index 00000000..111345d0 --- /dev/null +++ b/home/apps/prismlauncher.nix @@ -0,0 +1,13 @@ +{pkgs, ...}: { + home.packages = with pkgs; [ + (prismlauncher.override { + additionalPrograms = [ffmpeg zenity]; + jdks = [ + # graalvm-ce + zulu8 + zulu17 + zulu + ]; + }) + ]; +} diff --git a/home/services/gui.nix b/home/services/gui.nix index 22af95e8..ea8a1c5a 100644 --- a/home/services/gui.nix +++ b/home/services/gui.nix @@ -32,14 +32,5 @@ spotify steam-run wl-clipboard - # (prismlauncher.override { - # additionalPrograms = [ffmpeg zenity]; - # jdks = [ - # # graalvm-ce - # zulu8 - # zulu17 - # zulu - # ]; - # }) ]; } diff --git a/nixos/ryu/services/default.nix b/nixos/ryu/services/default.nix index 0721ac8b..e6b527a9 100644 --- a/nixos/ryu/services/default.nix +++ b/nixos/ryu/services/default.nix @@ -4,8 +4,8 @@ # ./sunshine.nix # ./zerotier.nix # ./dnscrypt.nix - # ./ollama.nix ./llama.nix + ./ollama.nix ./tailscale.nix ./samba.nix ./mullvad.nix diff --git a/nixos/ryu/services/llama.nix b/nixos/ryu/services/llama.nix index a4ec558a..b441de39 100644 --- a/nixos/ryu/services/llama.nix +++ b/nixos/ryu/services/llama.nix @@ -4,35 +4,32 @@ inputs, ... }: { + # llama-cpp = { + # enable = false; + # port = 11345; + # # model = "/nix/store/ch6z9di3l0k54ad29pzv8k3zv47q30d1-Qwen3-Coder-30B-A3B-Instruct-Q4_K_M.gguf"; + # model = pkgs.fetchurl { + # # url = "https://huggingface.co/lmstudio-community/gpt-oss-20b-GGUF/resolve/main/gpt-oss-20b-MXFP4.gguf"; + # # sha256 = "65d06d31a3977d553cb3af137b5c26b5f1e9297a6aaa29ae7caa98788cde53ab"; + # url = "https://huggingface.co/lmstudio-community/Qwen3-Coder-30B-A3B-Instruct-GGUF/resolve/main/Qwen3-Coder-30B-A3B-Instruct-Q4_K_M.gguf"; + # sha256 = "79ad15a5ee3caddc3f4ff0db33a14454a5a3eb503d7fa1c1e35feafc579de486"; + # }; + # extraFlags = [ + # "-c" + # "98304" + # "--jinja" + # "--chat-template-file" + # "${../../../assets/chat.hbs}" + # # "/nix/store/4zk1p50hrzghp3jzzysz96pa64i2kmjl-promp.hbs" + # ]; + # # package = inputs.llama-cpp.packages.${pkgs.system}.cuda; + # }; services = { - llama-cpp = { - enable = false; - port = 11345; - # model = "/nix/store/ch6z9di3l0k54ad29pzv8k3zv47q30d1-Qwen3-Coder-30B-A3B-Instruct-Q4_K_M.gguf"; - model = pkgs.fetchurl { - # url = "https://huggingface.co/lmstudio-community/gpt-oss-20b-GGUF/resolve/main/gpt-oss-20b-MXFP4.gguf"; - # sha256 = "65d06d31a3977d553cb3af137b5c26b5f1e9297a6aaa29ae7caa98788cde53ab"; - url = "https://huggingface.co/lmstudio-community/Qwen3-Coder-30B-A3B-Instruct-GGUF/resolve/main/Qwen3-Coder-30B-A3B-Instruct-Q4_K_M.gguf"; - sha256 = "79ad15a5ee3caddc3f4ff0db33a14454a5a3eb503d7fa1c1e35feafc579de486"; - }; - extraFlags = [ - "-c" - "98304" - "--jinja" - "--chat-template-file" - "${../../../assets/chat.hbs}" - # "/nix/store/4zk1p50hrzghp3jzzysz96pa64i2kmjl-promp.hbs" - ]; - # package = inputs.llama-cpp.packages.${pkgs.system}.cuda; - }; caddy = { virtualHosts."llama.ryu.darksailor.dev".extraConfig = '' import cloudflare - reverse_proxy localhost:${builtins.toString config.services.llama-cpp.port} + reverse_proxy localhost:11345 ''; }; }; - environment.systemPackages = with pkgs; [ - llama-cpp - ]; } diff --git a/nixos/ryu/services/ollama.nix b/nixos/ryu/services/ollama.nix index e0db7e27..2c7d13b9 100644 --- a/nixos/ryu/services/ollama.nix +++ b/nixos/ryu/services/ollama.nix @@ -6,30 +6,19 @@ }: { services = { ollama = { - enable = false; + enable = true; host = "0.0.0.0"; # loadModels = ["deepseek-r1:7b" "deepseek-r1:14b" "RobinBially/nomic-embed-text-8k" "qwen3:8b" "qwen3:14b"]; # loadModels = ["deepseek-r1:7b" "deepseek-r1:14b" "RobinBially/nomic-embed-text-8k" "qwen3:8b" "qwen3:14b"]; port = 11434; - acceleration = "cuda"; + # acceleration = "cuda"; environmentVariables = { OLLAMA_ORIGINS = "*"; OLLAMA_LLM_LIBRARY = "cuda"; LD_LIBRARY_PATH = "run/opengl-driver/lib"; HTTP_PROXY = "https://ollama.ryu.darksailor.dev"; }; - # package = pkgs.ollama.overrideAttrs { - # version = "0.11.0"; - # src = pkgs.fetchFromGitHub { - # owner = "ollama"; - # repo = "ollama"; - # tag = "v0.11.0"; - # hash = "sha256-po7BxJAj9eOpOaXsLDmw6/1RyjXPtXza0YUv0pVojZ0="; - # fetchSubmodules = true; - # }; - # doCheck = false; - # vendorHash = "sha256-SlaDsu001TUW+t9WRp7LqxUSQSGDF1Lqu9M1bgILoX4="; - # }; + package = pkgs.ollama-cuda; }; # open-webui = { # enable = false;