From a30d0f2e68ea6133b750ba22e6d914e826cbbb77 Mon Sep 17 00:00:00 2001 From: Giulio De Pasquale Date: Tue, 9 Jan 2024 14:59:25 +0100 Subject: [PATCH] llm: fixed upstream image for big-agi. use ollama with GPU support --- hosts/architect/llm.nix | 7 +++++-- 1 file changed, 5 insertions(+), 2 deletions(-) diff --git a/hosts/architect/llm.nix b/hosts/architect/llm.nix index ab2dd4a..73db576 100644 --- a/hosts/architect/llm.nix +++ b/hosts/architect/llm.nix @@ -1,11 +1,14 @@ -{ config, ... }: +{ config, pkgs, ... }: let domain = "pino.giugl.io"; backendPort = 3000; frontendPort = 3002; + llama-cpp = pkgs.unstablePkgs.llama-cpp.override { cudaSupport = true; }; + ollama = pkgs.unstablePkgs.ollama.override { inherit llama-cpp; }; in { + environment.systemPackages = [ ollama ]; architect.vhost.${domain} = { dnsInterfaces = [ "tailscale" ]; @@ -20,7 +23,7 @@ in virtualisation.oci-containers = { containers = { big-agi = { - image = "ghcr.io/enricoros/big-agi:main"; + image = "ghcr.io/enricoros/big-agi:latest"; autoStart = true; ports = [