Some checks failed
Flake checker / Build Nix targets (push) Has been cancelled
feat(nixos): enable ollama service with cuda support fix(nixos): update llama service to use fixed port and remove commented code chore(home): reorganize app imports and remove unused packages
36 lines
1.2 KiB
Nix
36 lines
1.2 KiB
Nix
{
|
|
pkgs,
|
|
config,
|
|
inputs,
|
|
...
|
|
}: {
|
|
# llama-cpp = {
|
|
# enable = false;
|
|
# port = 11345;
|
|
# # model = "/nix/store/ch6z9di3l0k54ad29pzv8k3zv47q30d1-Qwen3-Coder-30B-A3B-Instruct-Q4_K_M.gguf";
|
|
# model = pkgs.fetchurl {
|
|
# # url = "https://huggingface.co/lmstudio-community/gpt-oss-20b-GGUF/resolve/main/gpt-oss-20b-MXFP4.gguf";
|
|
# # sha256 = "65d06d31a3977d553cb3af137b5c26b5f1e9297a6aaa29ae7caa98788cde53ab";
|
|
# url = "https://huggingface.co/lmstudio-community/Qwen3-Coder-30B-A3B-Instruct-GGUF/resolve/main/Qwen3-Coder-30B-A3B-Instruct-Q4_K_M.gguf";
|
|
# sha256 = "79ad15a5ee3caddc3f4ff0db33a14454a5a3eb503d7fa1c1e35feafc579de486";
|
|
# };
|
|
# extraFlags = [
|
|
# "-c"
|
|
# "98304"
|
|
# "--jinja"
|
|
# "--chat-template-file"
|
|
# "${../../../assets/chat.hbs}"
|
|
# # "/nix/store/4zk1p50hrzghp3jzzysz96pa64i2kmjl-promp.hbs"
|
|
# ];
|
|
# # package = inputs.llama-cpp.packages.${pkgs.system}.cuda;
|
|
# };
|
|
services = {
|
|
caddy = {
|
|
virtualHosts."llama.ryu.darksailor.dev".extraConfig = ''
|
|
import cloudflare
|
|
reverse_proxy localhost:11345
|
|
'';
|
|
};
|
|
};
|
|
}
|