Compare commits

..

No commits in common. "ba39859e01e241c9e0486671e69659b5fc5c2359" and "8436e03e88291c2787c5608720510ad6b55778cc" have entirely different histories.

3 changed files with 11 additions and 15 deletions

18
flake.lock generated
View File

@ -7,11 +7,11 @@
] ]
}, },
"locked": { "locked": {
"lastModified": 1704099619, "lastModified": 1703113038,
"narHash": "sha256-QRVMkdxLmv+aKGjcgeEg31xtJEIsYq4i1Kbyw5EPS6g=", "narHash": "sha256-oxkyzjpD+mNT7arzU/zHrkNHLuY9tKwmnD2MNaZiSDw=",
"owner": "nix-community", "owner": "nix-community",
"repo": "home-manager", "repo": "home-manager",
"rev": "7e398b3d76bc1503171b1364c9d4a07ac06f3851", "rev": "0c2353d5d930c3d93724df6858aef064a31b3c00",
"type": "github" "type": "github"
}, },
"original": { "original": {
@ -23,11 +23,11 @@
}, },
"nixos-unstable": { "nixos-unstable": {
"locked": { "locked": {
"lastModified": 1704538339, "lastModified": 1703013332,
"narHash": "sha256-1734d3mQuux9ySvwf6axRWZRBhtcZA9Q8eftD6EZg6U=", "narHash": "sha256-+tFNwMvlXLbJZXiMHqYq77z/RfmpfpiI3yjL6o/Zo9M=",
"owner": "NixOS", "owner": "NixOS",
"repo": "nixpkgs", "repo": "nixpkgs",
"rev": "46ae0210ce163b3cba6c7da08840c1d63de9c701", "rev": "54aac082a4d9bb5bbc5c4e899603abfb76a3f6d6",
"type": "github" "type": "github"
}, },
"original": { "original": {
@ -39,11 +39,11 @@
}, },
"nixpkgs": { "nixpkgs": {
"locked": { "locked": {
"lastModified": 1704795870, "lastModified": 1703163755,
"narHash": "sha256-M86cqEn65whEr2JC4yWVJXzGikqpt+2SzcSe3DOiRHQ=", "narHash": "sha256-5urf1hkKPpD+HfphQWUA4ogu2v5pfcWjxWArn/vYW4c=",
"owner": "NixOS", "owner": "NixOS",
"repo": "nixpkgs", "repo": "nixpkgs",
"rev": "72f7568a6d748d87e878098957077118a6af91f8", "rev": "34309fb17d5c1e96f97274b21654712c8396813a",
"type": "github" "type": "github"
}, },
"original": { "original": {

View File

@ -1,14 +1,11 @@
{ config, pkgs, ... }: { config, ... }:
let let
domain = "pino.giugl.io"; domain = "pino.giugl.io";
backendPort = 3000; backendPort = 3000;
frontendPort = 3002; frontendPort = 3002;
llama-cpp = pkgs.unstablePkgs.llama-cpp.override { cudaSupport = true; };
ollama = pkgs.unstablePkgs.ollama.override { inherit llama-cpp; };
in in
{ {
environment.systemPackages = [ ollama ];
architect.vhost.${domain} = { architect.vhost.${domain} = {
dnsInterfaces = [ "tailscale" ]; dnsInterfaces = [ "tailscale" ];
@ -23,7 +20,7 @@ in
virtualisation.oci-containers = { virtualisation.oci-containers = {
containers = { containers = {
big-agi = { big-agi = {
image = "ghcr.io/enricoros/big-agi:latest"; image = "ghcr.io/enricoros/big-agi:main";
autoStart = true; autoStart = true;
ports = [ ports = [

View File

@ -71,7 +71,6 @@ in
ruff ruff
shellcheck shellcheck
shfmt shfmt
gopls
] ++ pythonPkgs ++ nodePkgs; ] ++ pythonPkgs ++ nodePkgs;
}; };
} }