mirror of
https://github.com/nomic-ai/gpt4all.git
synced 2024-10-01 01:06:10 -04:00
Continue to shrink the API space for qml and the backend.
This commit is contained in:
parent
ed59190e48
commit
c0d4a9d426
12
chat.cpp
12
chat.cpp
@ -55,11 +55,6 @@ void Chat::regenerateResponse()
|
||||
emit regenerateResponseRequested(); // blocking queued connection
|
||||
}
|
||||
|
||||
void Chat::resetResponse()
|
||||
{
|
||||
emit resetResponseRequested(); // blocking queued connection
|
||||
}
|
||||
|
||||
void Chat::stopGenerating()
|
||||
{
|
||||
m_llmodel->stopGenerating();
|
||||
@ -109,6 +104,13 @@ int32_t Chat::threadCount() {
|
||||
return m_llmodel->threadCount();
|
||||
}
|
||||
|
||||
void Chat::newPromptResponsePair(const QString &prompt)
|
||||
{
|
||||
m_chatModel->appendPrompt(tr("Prompt: "), prompt);
|
||||
m_chatModel->appendResponse(tr("Response: "), prompt);
|
||||
emit resetResponseRequested(); // blocking queued connection
|
||||
}
|
||||
|
||||
bool Chat::isRecalc() const
|
||||
{
|
||||
return m_llmodel->isRecalc();
|
||||
|
2
chat.h
2
chat.h
@ -34,11 +34,11 @@ public:
|
||||
Q_INVOKABLE void prompt(const QString &prompt, const QString &prompt_template, int32_t n_predict, int32_t top_k, float top_p,
|
||||
float temp, int32_t n_batch, float repeat_penalty, int32_t repeat_penalty_tokens);
|
||||
Q_INVOKABLE void regenerateResponse();
|
||||
Q_INVOKABLE void resetResponse();
|
||||
Q_INVOKABLE void stopGenerating();
|
||||
Q_INVOKABLE void syncThreadCount();
|
||||
Q_INVOKABLE void setThreadCount(int32_t n_threads);
|
||||
Q_INVOKABLE int32_t threadCount();
|
||||
Q_INVOKABLE void newPromptResponsePair(const QString &prompt);
|
||||
|
||||
QString response() const;
|
||||
bool responseInProgress() const { return m_responseInProgress; }
|
||||
|
@ -101,7 +101,7 @@ public:
|
||||
return roles;
|
||||
}
|
||||
|
||||
Q_INVOKABLE void appendPrompt(const QString &name, const QString &value)
|
||||
void appendPrompt(const QString &name, const QString &value)
|
||||
{
|
||||
ChatItem item;
|
||||
item.name = name;
|
||||
@ -112,7 +112,7 @@ public:
|
||||
emit countChanged();
|
||||
}
|
||||
|
||||
Q_INVOKABLE void appendResponse(const QString &name, const QString &prompt)
|
||||
void appendResponse(const QString &name, const QString &prompt)
|
||||
{
|
||||
ChatItem item;
|
||||
item.id = m_chatItems.count(); // This is only relevant for responses
|
||||
|
4
main.qml
4
main.qml
@ -949,9 +949,7 @@ Window {
|
||||
chatModel.updateCurrentResponse(index, false);
|
||||
chatModel.updateValue(index, LLM.currentChat.response);
|
||||
}
|
||||
chatModel.appendPrompt(qsTr("Prompt: "), textInput.text);
|
||||
chatModel.appendResponse(qsTr("Response: "), textInput.text);
|
||||
LLM.currentChat.resetResponse()
|
||||
LLM.currentChat.newPromptResponsePair(textInput.text);
|
||||
LLM.currentChat.prompt(textInput.text, settingsDialog.promptTemplate,
|
||||
settingsDialog.maxLength,
|
||||
settingsDialog.topK,
|
||||
|
Loading…
Reference in New Issue
Block a user