From cdb854db9ebeaca56e9d456700bd38166ebf2564 Mon Sep 17 00:00:00 2001 From: "q5sys (JT)" Date: Tue, 29 Aug 2023 16:56:50 -0400 Subject: [PATCH] Update llama.cpp.md instructions (#3702) --- docs/llama.cpp.md | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/docs/llama.cpp.md b/docs/llama.cpp.md index 68aa1cfa..c27ad65a 100644 --- a/docs/llama.cpp.md +++ b/docs/llama.cpp.md @@ -9,7 +9,10 @@ llama.cpp is the best backend in two important scenarios: #### Pre-converted -Download the ggml model directly into your `text-generation-webui/models` folder, making sure that its name contains `ggml` somewhere and ends in `.bin`. It's a single file. +Download the GGUF or GGML models directly into your `text-generation-webui/models` folder. It will be a single file. + +* For GGUF models, make sure its name contains `.gguf`. +* For GGML models, make sure its name contains `ggml` and ends in `.bin`. `q4_K_M` quantization is recommended.