backend: do not use Vulkan with non-LLaMA models

This commit is contained in:
Cebtenzzre 2023-10-04 16:21:30 -04:00 committed by Adam Treat
parent 672cb850f9
commit 1534df3e9f

View File

@ -309,8 +309,7 @@ bool ChatLLM::loadModel(const ModelInfo &modelInfo)
// We might have had to fallback to CPU after load if the model is not possible to accelerate
// for instance if the quantization method is not supported on Vulkan yet
emit reportDevice("CPU");
// TODO(cebtenzzre): report somewhere if llamamodel decided the model was not supported
emit reportFallbackReason("<br>Using CPU: unsupported quantization type");
emit reportFallbackReason("<br>Using CPU: unsupported model or quant");
}
MySettings::globalInstance()->setAttemptModelLoad(QString());