mirror of
https://github.com/nomic-ai/gpt4all.git
synced 2024-10-01 01:06:10 -04:00
mpt: allow q4_2 quantized models to load
This commit is contained in:
parent
832720dd27
commit
5002614b20
@ -236,6 +236,7 @@ bool mpt_model_load(const std::string &fname, std::istream &fin, mpt_model & mod
|
|||||||
case 1: wtype = GGML_TYPE_F16; break;
|
case 1: wtype = GGML_TYPE_F16; break;
|
||||||
case 2: wtype = GGML_TYPE_Q4_0; break;
|
case 2: wtype = GGML_TYPE_Q4_0; break;
|
||||||
case 3: wtype = GGML_TYPE_Q4_1; break;
|
case 3: wtype = GGML_TYPE_Q4_1; break;
|
||||||
|
case 5: wtype = GGML_TYPE_Q4_2; break;
|
||||||
default:
|
default:
|
||||||
{
|
{
|
||||||
fprintf(stderr, "%s: invalid model file '%s' (bad f16 value %d)\n",
|
fprintf(stderr, "%s: invalid model file '%s' (bad f16 value %d)\n",
|
||||||
|
Loading…
Reference in New Issue
Block a user