mirror of
https://github.com/nomic-ai/gpt4all.git
synced 2024-10-01 01:06:10 -04:00
backend: do not use Vulkan with non-LLaMA models
This commit is contained in:
parent
672cb850f9
commit
1534df3e9f
@ -309,8 +309,7 @@ bool ChatLLM::loadModel(const ModelInfo &modelInfo)
|
||||
// We might have had to fallback to CPU after load if the model is not possible to accelerate
|
||||
// for instance if the quantization method is not supported on Vulkan yet
|
||||
emit reportDevice("CPU");
|
||||
// TODO(cebtenzzre): report somewhere if llamamodel decided the model was not supported
|
||||
emit reportFallbackReason("<br>Using CPU: unsupported quantization type");
|
||||
emit reportFallbackReason("<br>Using CPU: unsupported model or quant");
|
||||
}
|
||||
|
||||
MySettings::globalInstance()->setAttemptModelLoad(QString());
|
||||
|
Loading…
Reference in New Issue
Block a user