Personal Nix setup
0
fork

Configure Feed

Select the types of activity you want to include in your feed.

Add prepackaged ollama 0.11.2 download

+29 -2
+6
home/apps/ollama.nix
··· 16 16 OLLAMA_FLASH_ATTENTION = if cfg.ollama.flashAttention then "1" else "0"; 17 17 OLLAMA_SCHED_SPREAD = if cfg.ollama.schedSpread then "1" else "0"; 18 18 OLLAMA_INTEL_GPU = if cfg.ollama.intelGpu then "1" else "0"; 19 + OLLAMA_NEW_ENGINE = if cfg.ollama.newEngine then "1" else "0"; 19 20 OLLAMA_KV_CACHE_TYPE = cfg.ollama.kvCacheType; 20 21 OLLAMA_CONTEXT_LENGTH = toString cfg.ollama.defaultContextLength; 21 22 OLLAMA_MAX_LOADED_MODELS = toString cfg.ollama.maxLoadedModels; ··· 83 84 Effect: Enables multi-GPU usage for model inference. 84 85 Scenario: Beneficial in high-performance computing environments with multiple GPUs to maximize hardware utilization. 85 86 ''; 87 + type = types.bool; 88 + }; 89 + 90 + newEngine = mkOption { 91 + default = true; 86 92 type = types.bool; 87 93 }; 88 94
+23 -2
machines/irnbru/home.nix
··· 1 - { ... }: 1 + { pkgs, ... }: 2 2 3 3 { 4 4 modules = { ··· 7 7 }; 8 8 apps = { 9 9 enable = true; 10 - ollama.enable = true; 11 10 ghostty.enable = true; 11 + ollama = { 12 + enable = true; 13 + package = with pkgs; stdenv.mkDerivation rec { 14 + pname = "ollama"; 15 + version = "0.11.2"; 16 + src = fetchurl { 17 + url = "https://github.com/ollama/ollama/releases/download/v${version}/ollama-darwin.tgz"; 18 + hash = "sha256-gUKwOmZ6oViIEzozQx5FB5090Zs0q8nQTdRqjqa0t8I="; 19 + }; 20 + sourceRoot = "."; 21 + dontBuild = true; 22 + dontConfigure = true; 23 + installPhase = '' 24 + runHook preInstall 25 + mkdir -p $out/bin $out/lib/ollama 26 + ls -lah 27 + install -Dm755 ollama $out/bin/ollama 28 + cp -r *.so *.dylib $out/lib/ollama/ 29 + runHook postInstall 30 + ''; 31 + }; 32 + }; 12 33 }; 13 34 }; 14 35 }