mirror of
https://github.com/nomic-ai/gpt4all.git
synced 2024-10-01 01:06:10 -04:00
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
This commit is contained in:
parent
62bc84366b
commit
3025f9deff
@ -17,6 +17,7 @@ The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.1.0/).
|
|||||||
### Fixed
|
### Fixed
|
||||||
- Fix a crash when attempting to continue a chat loaded from disk ([#2995](https://github.com/nomic-ai/gpt4all/pull/2995))
|
- Fix a crash when attempting to continue a chat loaded from disk ([#2995](https://github.com/nomic-ai/gpt4all/pull/2995))
|
||||||
- Fix the local server rejecting min\_p/top\_p less than 1 ([#2996](https://github.com/nomic-ai/gpt4all/pull/2996))
|
- Fix the local server rejecting min\_p/top\_p less than 1 ([#2996](https://github.com/nomic-ai/gpt4all/pull/2996))
|
||||||
|
- Fix "regenerate" always forgetting the most recent message ([#3011](https://github.com/nomic-ai/gpt4all/pull/3011))
|
||||||
|
|
||||||
## [3.3.1] - 2024-09-27 ([v3.3.y](https://github.com/nomic-ai/gpt4all/tree/v3.3.y))
|
## [3.3.1] - 2024-09-27 ([v3.3.y](https://github.com/nomic-ai/gpt4all/tree/v3.3.y))
|
||||||
|
|
||||||
|
@ -238,8 +238,9 @@ void Chat::newPromptResponsePair(const QString &prompt)
|
|||||||
{
|
{
|
||||||
resetResponseState();
|
resetResponseState();
|
||||||
m_chatModel->updateCurrentResponse(m_chatModel->count() - 1, false);
|
m_chatModel->updateCurrentResponse(m_chatModel->count() - 1, false);
|
||||||
|
// the prompt is passed as the prompt item's value and the response item's prompt
|
||||||
m_chatModel->appendPrompt("Prompt: ", prompt);
|
m_chatModel->appendPrompt("Prompt: ", prompt);
|
||||||
m_chatModel->appendResponse("Response: ", QString());
|
m_chatModel->appendResponse("Response: ", prompt);
|
||||||
emit resetResponseRequested();
|
emit resetResponseRequested();
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -248,8 +249,9 @@ void Chat::serverNewPromptResponsePair(const QString &prompt)
|
|||||||
{
|
{
|
||||||
resetResponseState();
|
resetResponseState();
|
||||||
m_chatModel->updateCurrentResponse(m_chatModel->count() - 1, false);
|
m_chatModel->updateCurrentResponse(m_chatModel->count() - 1, false);
|
||||||
|
// the prompt is passed as the prompt item's value and the response item's prompt
|
||||||
m_chatModel->appendPrompt("Prompt: ", prompt);
|
m_chatModel->appendPrompt("Prompt: ", prompt);
|
||||||
m_chatModel->appendResponse("Response: ", QString());
|
m_chatModel->appendResponse("Response: ", prompt);
|
||||||
}
|
}
|
||||||
|
|
||||||
bool Chat::restoringFromText() const
|
bool Chat::restoringFromText() const
|
||||||
|
Loading…
Reference in New Issue
Block a user