up llama
This commit is contained in:
@@ -2,6 +2,8 @@
|
||||
config,
|
||||
lib,
|
||||
pkgs,
|
||||
inputs,
|
||||
system,
|
||||
namespace,
|
||||
...
|
||||
}:
|
||||
@@ -28,7 +30,7 @@ let
|
||||
description = "AI Services";
|
||||
options = {
|
||||
llama-cpp = {
|
||||
model = mkOpt types.str "Qwen3-Coder-Next-UD-Q3_K_XL" "";
|
||||
model = mkOpt types.str "models--unsloth--gemma-4-26B-A4B-it-GGUF/snapshots/b8654b48d979f2853b7a81d6541ca64eea7dc3c5/gemma-4-26B-A4B-it-UD-Q8_K_XL" "";
|
||||
};
|
||||
};
|
||||
moduleConfig = {
|
||||
@@ -51,8 +53,8 @@ let
|
||||
port = 8127;
|
||||
host = "0.0.0.0";
|
||||
openFirewall = cfg.openFirewall;
|
||||
model = null;
|
||||
package = pkgs.llama-cpp-rocm;
|
||||
model = "${cfg.configDir}/llama-cpp/models/${cfg.llama-cpp.model}.gguf";
|
||||
package = inputs.llama-cpp.packages.${system}.rocm;
|
||||
extraFlags = [
|
||||
"--fit"
|
||||
"on"
|
||||
|
||||
@@ -73,6 +73,14 @@ let
|
||||
package = pkgs.sonarr;
|
||||
};
|
||||
|
||||
services.lidarr = {
|
||||
enable = true;
|
||||
openFirewall = cfg.openFirewall;
|
||||
user = "nix-apps";
|
||||
group = "jallen-nas";
|
||||
dataDir = "${cfg.configDir}/lidarr";
|
||||
};
|
||||
|
||||
# Enable Sabnzbd service
|
||||
services.sabnzbd = {
|
||||
enable = true;
|
||||
|
||||
Reference in New Issue
Block a user