mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2024-10-01 01:26:03 -04:00
Update llama.cpp.md instructions (#3702)
This commit is contained in:
parent
a9a1784420
commit
cdb854db9e
@ -9,7 +9,10 @@ llama.cpp is the best backend in two important scenarios:
|
||||
|
||||
#### Pre-converted
|
||||
|
||||
Download the ggml model directly into your `text-generation-webui/models` folder, making sure that its name contains `ggml` somewhere and ends in `.bin`. It's a single file.
|
||||
Download the GGUF or GGML models directly into your `text-generation-webui/models` folder. It will be a single file.
|
||||
|
||||
* For GGUF models, make sure its name contains `.gguf`.
|
||||
* For GGML models, make sure its name contains `ggml` and ends in `.bin`.
|
||||
|
||||
`q4_K_M` quantization is recommended.
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user