From 880780080ba75a94ef08228c871246c6bf248af7 Mon Sep 17 00:00:00 2001 From: Green Sky Date: Fri, 27 Oct 2023 22:35:05 +0200 Subject: [PATCH] flake : use even smaller version of torch --- flake.nix | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/flake.nix b/flake.nix index a969ac9ca..fa34394b2 100644 --- a/flake.nix +++ b/flake.nix @@ -53,7 +53,7 @@ pkgs.python3.withPackages (ps: with ps; [ numpy sentencepiece ]); # TODO(Green-Sky): find a better way to opt-into the heavy ml python runtime llama-python-extra = - pkgs.python3.withPackages (ps: with ps; [ numpy sentencepiece torch transformers ]); + pkgs.python3.withPackages (ps: with ps; [ numpy sentencepiece torchWithoutCuda transformers ]); postPatch = '' substituteInPlace ./ggml-metal.m \ --replace '[bundle pathForResource:@"ggml-metal" ofType:@"metal"];' "@\"$out/bin/ggml-metal.metal\";"