this repo has no description
4
fork

Configure Feed

Select the types of activity you want to include in your feed.

im updating it so good

dawn 269aacf6 4486c07f

+22 -50
+4 -4
_sources/generated.json
··· 42 42 }, 43 43 "blog": { 44 44 "cargoLock": null, 45 - "date": "2026-04-17", 45 + "date": "2026-04-24", 46 46 "extract": null, 47 47 "name": "blog", 48 48 "passthru": null, ··· 52 52 "fetchSubmodules": false, 53 53 "leaveDotGit": false, 54 54 "name": null, 55 - "rev": "07d745fb33cf580b79ff6d29631188eecd471357", 56 - "sha256": "sha256-2ThCN55JV1b4YbMaOtc7oFVNNSU/IC8wx6F7Ybj1+z4=", 55 + "rev": "9a31eccc91322536b4278f5b037f3c78e184abb3", 56 + "sha256": "sha256-9G10HuIkLvWL2dzwA2FVdq/q17XWGcL+wlMZXyMKkJk=", 57 57 "sparseCheckout": [], 58 58 "type": "git", 59 59 "url": "https://tangled.org/@ptr.pet/endpoint" 60 60 }, 61 - "version": "07d745fb33cf580b79ff6d29631188eecd471357" 61 + "version": "9a31eccc91322536b4278f5b037f3c78e184abb3" 62 62 }, 63 63 "clickee": { 64 64 "cargoLock": null,
+4 -4
_sources/generated.nix
··· 34 34 }; 35 35 blog = { 36 36 pname = "blog"; 37 - version = "07d745fb33cf580b79ff6d29631188eecd471357"; 37 + version = "9a31eccc91322536b4278f5b037f3c78e184abb3"; 38 38 src = fetchgit { 39 39 url = "https://tangled.org/@ptr.pet/endpoint"; 40 - rev = "07d745fb33cf580b79ff6d29631188eecd471357"; 40 + rev = "9a31eccc91322536b4278f5b037f3c78e184abb3"; 41 41 fetchSubmodules = false; 42 42 deepClone = false; 43 43 leaveDotGit = false; 44 44 sparseCheckout = [ ]; 45 - sha256 = "sha256-2ThCN55JV1b4YbMaOtc7oFVNNSU/IC8wx6F7Ybj1+z4="; 45 + sha256 = "sha256-9G10HuIkLvWL2dzwA2FVdq/q17XWGcL+wlMZXyMKkJk="; 46 46 }; 47 - date = "2026-04-17"; 47 + date = "2026-04-24"; 48 48 }; 49 49 clickee = { 50 50 pname = "clickee";
+2 -30
hosts/chernobog/modules/llm.nix
··· 1 - {pkgs, ...}: 1 + {pkgs, terra, ...}: 2 2 let 3 - llama = (pkgs.llama-cpp.override { 3 + llama = (terra.llama-cpp.override { 4 4 blasSupport = true; 5 5 vulkanSupport = true; 6 6 cudaSupport = false; ··· 9 9 doCheck = false; 10 10 cmakeFlags = (old.cmakeFlags or []) ++ ["-DGGML_AVX2=ON" "-DGGML_FMA=ON" "-DGGML_F16C=ON"]; 11 11 }); 12 - # llama = pkgs.stdenv.mkDerivation { 13 - # name = "llamacpp-rocm-b1229"; 14 - 15 - # src = pkgs.fetchurl { 16 - # url = "https://github.com/lemonade-sdk/llamacpp-rocm/releases/download/b1229/llama-b1229-ubuntu-rocm-gfx110X-x64.zip"; 17 - # hash = "sha256-RGKGfZ3HBssBhFaskb8IAQYMng+K4Fol7J0H5rCmgwA="; 18 - # }; 19 - # sourceRoot = "."; 20 - 21 - # nativeBuildInputs = [ pkgs.autoPatchelfHook pkgs.makeBinaryWrapper pkgs.unzip ]; 22 - # buildInputs = [ pkgs.stdenv.cc.cc.lib ]; # just glibc/libstdc++ 23 - 24 - # installPhase = '' 25 - # mkdir -p $out/bin $out/lib/rocblas 26 - # cp *.so* $out/lib || true 27 - # cp llama-* $out/bin 28 - # chmod +x $out/bin/* 29 - # ln -sf ${pkgs.rocmPackages.rocblas}/lib/rocblas/library $out/lib/rocblas/library 30 - # ''; 31 - 32 - # autoPatchelfIgnoreMissingDeps = true; 33 - # }; 34 12 in { 35 - # services.ollama = { 36 - # enable = true; 37 - # package = pkgs.ollama-rocm; 38 - # rocmOverrideGfx = "11.0.0"; 39 - # }; 40 - 41 13 hardware.amdgpu.opencl.enable = true; 42 14 43 15 # services.llama-cpp = {
+3 -3
hosts/trimounts/modules/website.nix
··· 5 5 ... 6 6 }: 7 7 let 8 - PUBLIC_BASE_URL = "https://gaze.systems"; 8 + PUBLIC_BASE_URL = "https://ptr.pet"; 9 9 modules = (pkgs.callPackage "${inputs.blog}/nix/modules.nix" { }).overrideAttrs (_: { 10 10 # outputHash = "sha256-rzfSfiK8FSNFR+1QTwM/ltLZBprG9BoQsPmOt6IdXFc="; 11 11 }); 12 12 pkg = pkgs.callPackage "${inputs.blog}/nix" { 13 13 inherit PUBLIC_BASE_URL; 14 - eunomia-modules = modules; 14 + endpoint-modules = modules; 15 15 }; 16 16 port = 3003; 17 17 vhostConfig = { ··· 59 59 }; 60 60 serviceConfig = { 61 61 User = "website"; 62 - ExecStart = "${pkg}/bin/eunomia"; 62 + ExecStart = "${pkg}/bin/endpoint"; 63 63 Restart = "on-failure"; 64 64 RestartSec = 5; 65 65 WorkingDirectory = "/var/lib/website";
+3 -3
pkgs-set/pkgs/helium.nix
··· 5 5 }: 6 6 pkgs.appimageTools.wrapType2 rec { 7 7 pname = "helium"; 8 - version = "0.10.8.1"; 8 + version = "0.11.4.1"; 9 9 10 10 src = let 11 11 platformMap = { ··· 16 16 platform = platformMap.${pkgs.system}; 17 17 18 18 hashes = { 19 - "x86_64-linux" = "sha256-pN/Iw1ANggDOxxFb2CN436qbcrs8/bDcEqjZC80grQs="; 20 - "aarch64-linux" = "sha256-B63tvOtSRlMRJozvzC7lqG2LM0ZgLIq2G/AHABl+Qqg="; 19 + "x86_64-linux" = "sha256-EavHZuleVkgD4J+gcUQ2ZgGHjvWXzdamUh7r42bUkas="; 20 + "aarch64-linux" = ""; 21 21 }; 22 22 23 23 hash = hashes.${pkgs.system};
+5 -5
pkgs-set/pkgs/llama-cpp.nix
··· 83 83 in 84 84 effectiveStdenv.mkDerivation (finalAttrs: { 85 85 pname = "llama-cpp"; 86 - version = "8664"; 86 + version = "8925"; 87 87 88 88 outputs = [ 89 89 "out" ··· 93 93 src = fetchFromGitHub { 94 94 owner = "ggml-org"; 95 95 repo = "llama.cpp"; 96 - # tag = "b${finalAttrs.version}"; 97 - rev = "b8635075ffe27b135c49afb9a8b5c434bd42c502"; 98 - hash = "sha256-5rioYM989O6wXEtC3SWg7v3ZgMCSzE6/RyF6ILzY9rQ="; 96 + tag = "b${finalAttrs.version}"; 97 + # rev = "b8635075ffe27b135c49afb9a8b5c434bd42c502"; 98 + hash = "sha256-BVRp+T4eKZYS0aT4SGx/M/k9HJ7V74M1z5OTOasFl8E="; 99 99 leaveDotGit = true; 100 100 postFetch = '' 101 101 git -C "$out" rev-parse --short HEAD > $out/COMMIT ··· 131 131 ++ [ openssl ]; 132 132 133 133 npmRoot = "tools/server/webui"; 134 - npmDepsHash = "sha256-DxgUDVr+kwtW55C4b89Pl+j3u2ILmACcQOvOBjKWAKQ="; 134 + npmDepsHash = "sha256-RAFtsbBGBjteCt5yXhrmHL39rIDJMCFBETgzId2eRRk="; 135 135 npmDeps = fetchNpmDeps { 136 136 name = "${finalAttrs.pname}-${finalAttrs.version}-npm-deps"; 137 137 inherit (finalAttrs) src patches;
+1 -1
users/modules/discord/default.nix
··· 5 5 }: 6 6 let 7 7 pkg = pkgs.discord.override { 8 - withMoonlight = true; 8 + withMoonlight = false; 9 9 inherit (terra) moonlight; 10 10 withOpenASAR = false; 11 11 withTTS = false;