From 55170d93b783e80b9b6d5ac111408606d5327d3a Mon Sep 17 00:00:00 2001 From: Honkware <119620994+Honkware@users.noreply.github.com> Date: Sun, 27 Oct 2024 18:12:50 -0500 Subject: [PATCH] corrected node-llama-cpp command --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 9f0af7dcda..007de180d6 100644 --- a/README.md +++ b/README.md @@ -75,7 +75,7 @@ TWITTER_COOKIES= # Account cookies If you have an NVIDIA GPU, you can install CUDA to speed up local inference dramatically. ``` npm install -npx --no node-llama-cpp download --gpu cuda +npx --no node-llama-cpp source download --gpu cuda ``` Make sure that you've installed the CUDA Toolkit, including cuDNN and cuBLAS.