caddy
This commit is contained in:
@@ -7,13 +7,19 @@
|
||||
}:
|
||||
with lib;
|
||||
let
|
||||
inherit (lib.${namespace}) mkOpt;
|
||||
|
||||
cfg = config.${namespace}.services.ai;
|
||||
|
||||
aiConfig = lib.${namespace}.mkModule {
|
||||
inherit config;
|
||||
name = "ai";
|
||||
description = "AI Services";
|
||||
options = { };
|
||||
options = {
|
||||
llama-cpp = {
|
||||
model = mkOpt types.str "Qwen3-Coder-Next-UD-Q3_K_XL" "";
|
||||
};
|
||||
};
|
||||
moduleConfig = {
|
||||
services = {
|
||||
ollama = {
|
||||
@@ -34,7 +40,7 @@ let
|
||||
port = 8127;
|
||||
host = "0.0.0.0";
|
||||
openFirewall = cfg.openFirewall;
|
||||
model = "${cfg.configDir}/llama-cpp/models/Qwen3-Coder-Next-Q4_0.gguf";
|
||||
model = "${cfg.configDir}/llama-cpp/models/${cfg.llama-cpp.model}.gguf";
|
||||
package = pkgs.llama-cpp-rocm;
|
||||
extraFlags = [
|
||||
"--fit"
|
||||
@@ -105,7 +111,7 @@ let
|
||||
set -euo pipefail
|
||||
|
||||
MODEL_DIR="${cfg.configDir}/llama-cpp/models"
|
||||
MODEL_NAME="Qwen3-Coder-Next-Q4_0.gguf"
|
||||
MODEL_NAME="${cfg.llama-cpp.model}.gguf"
|
||||
REPO_ID="unsloth/Qwen3-Coder-Next-GGUF"
|
||||
|
||||
# Create model directory if it doesn't exist
|
||||
|
||||
Reference in New Issue
Block a user