From 3249180cce721df8b56cd1335b4d58e6d5951794 Mon Sep 17 00:00:00 2001 From: mjallen18 Date: Wed, 19 Mar 2025 16:13:05 -0500 Subject: [PATCH] ollama --- hosts/nas/apps/ollama/default.nix | 21 +++++++++++++++++++-- 1 file changed, 19 insertions(+), 2 deletions(-) diff --git a/hosts/nas/apps/ollama/default.nix b/hosts/nas/apps/ollama/default.nix index 903b4d4..efdf5d7 100755 --- a/hosts/nas/apps/ollama/default.nix +++ b/hosts/nas/apps/ollama/default.nix @@ -1,4 +1,17 @@ -{ config, ... }: +{ config, pkgs, ... }: +let + llamaPackage = pkgs.llama-cpp.overrideAttrs (old: { + src = pkgs.fetchFromGitHub { + owner = "ggml-org"; + repo = "llama.cpp"; + rev = "b4920"; + sha256 = "sha256-SnQIeY74JpAPRMxWcpklDH5D4CQvAgi0GYx5+ECk2J4="; + }; + # Optionally override other attributes if you need to + # version = "my-fork-version"; + # pname = "llama-cpp-custom"; + }); +in { services.ollama = { enable = true; @@ -11,12 +24,15 @@ home = "/media/nas/ssd/nix-app-data/ollama"; }; + environment.systemPackages = [ llamaPackage ]; + services.llama-cpp = { enable = true; port = 8127; host = "0.0.0.0"; openFirewall = true; model = "/media/nas/ssd/nix-app-data/llama-cpp/models/functionary-small-v3.2-GGUF/functionary-small-v3.2.Q4_0.gguf"; + package = llamaPackage; # pkgs.unstable.llama-cpp; extraFlags = [ "--n_gpu-layers" "500" @@ -24,7 +40,8 @@ "0" "--numa" "numactl" - ] + "--jinja" + ]; }; services.open-webui = {