gpt4all/gpt4all-chat/metadata/release.json

178 lines
5.9 KiB
JSON
Raw Normal View History

[
{
"version": "2.2.2",
"notes":
"
* repeat penalty for both gptj and llama models
* scroll the context window when conversation reaches context limit
* persistent thread count setting
* new default template
* new settings for model path, repeat penalty
* bugfix for settings dialog onEditingFinished
* new tab based settings dialog format
* bugfix for datalake when conversation contains forbidden json chars
* new C library API and split the backend into own separate lib for bindings
* apple signed/notarized dmg installer
* update llama.cpp submodule to latest
* bugfix for too large of a prompt
* support for opt-in only anonymous usage and statistics
* bugfixes for the model downloader and improve performance
* various UI bugfixes and enhancements including the send message textarea automatically wrapping by word
* new startup dialog on first start of a new release displaying release notes and opt-in buttons
* new logo and icons
* fixed apple installer so there is now a symlink in the applications folder
",
"contributors":
"
* Adam Treat (Nomic AI)
* Aaron Miller
* Matthieu Talbot
* Tim Jobbins
* chad (eachadea)
* Community (beta testers, bug reporters)
"
},
{
"version": "2.3.0",
"notes":
"
* repeat penalty for both gptj and llama models
* scroll the context window when conversation reaches context limit
* persistent thread count setting
* new default template
* new settings for model path, repeat penalty
* bugfix for settings dialog onEditingFinished
* new tab based settings dialog format
* bugfix for datalake when conversation contains forbidden json chars
* new C library API and split the backend into own separate lib for bindings
* apple signed/notarized dmg installer
* update llama.cpp submodule to latest
* bugfix for too large of a prompt
* support for opt-in only anonymous usage and statistics
* bugfixes for the model downloader and improve performance
* various UI bugfixes and enhancements including the send message textarea automatically wrapping by word
* new startup dialog on first start of a new release displaying release notes and opt-in buttons
* new logo and icons
* fixed apple installer so there is now a symlink in the applications folder
* fixed bug with versions
* fixed optout marking
",
"contributors":
"
* Adam Treat (Nomic AI)
* Aaron Miller
* Matthieu Talbot
* Tim Jobbins
* chad (eachadea)
* Community (beta testers, bug reporters)
"
},
{
"version": "2.4.0",
"notes":
"
* reverse prompt for both llama and gptj models which should help stop them from repeating the prompt template
* resumable downloads for models
* chat list in the drawer drop down
* add/remove/rename chats
* perist chats to disk and restore them with full context (WARNING: the average size of each chat on disk is ~1.5GB)
* NOTE: to turn on the persistent chats feature you need to do so via the settings dialog as it is off by default
* automatically rename chats using the AI after the first prompt/response pair
* new usage statistics including more detailed hardware info to help debug problems on older hardware
* fix dialog sizes for those with smaller displays
* add support for persistent contexts and internal model state to the C api
* add a confirm button for deletion of chats
* bugfix for blocking the gui when changing models
* datalake now captures all conversations when network opt-in is turned on
* new much shorter prompt template by default
",
"contributors":
"
* Adam Treat (Nomic AI)
* Aaron Miller
* Community (beta testers, bug reporters)
"
},
{
"version": "2.4.1",
"notes":
"
* compress persistent chats and save order of magnitude disk space on some small chats
* persistent chat files are now stored in same folder as models
* use a thread for deserializing chats on startup so the gui shows window faster
* fail gracefully and early when we detect incompatible hardware
* repeat penalty restore default bugfix
* new mpt backend for mosaic ml's new base model and chat model
* add mpt chat and base model to downloads
* lower memory required for gptj models by using f16 for kv cache
* better error handling for when a model is deleted by user and persistent chat remains
* add a user default model setting so the users preferred model comes up on startup
",
"contributors":
"
* Adam Treat (Nomic AI)
* Zach Nussbaum (Nomic AI)
* Aaron Miller
* Community (beta testers, bug reporters)
"
},
{
"version": "2.4.2",
"notes":
"
* add webserver feature that offers mirror api to chatgpt on localhost:4891
* add chatgpt models installed using openai key to chat client gui
* fixup the memory handling when switching between chats/models to decrease RAM load across the board
* fix bug in thread safety for mpt model and de-duplicated code
* uses compact json format for network
* add remove model option in download dialog
",
"contributors":
"
* Adam Treat (Nomic AI)
* Aaron Miller
* Community (beta testers, bug reporters)
"
},
{
"version": "2.4.3",
"notes":
"
* add webserver feature that offers mirror api to chatgpt on localhost:4891
* add chatgpt models installed using openai key to chat client gui
* fixup the memory handling when switching between chats/models to decrease RAM load across the board
* fix bug in thread safety for mpt model and de-duplicated code
* uses compact json format for network
* add remove model option in download dialog
* remove text-davinci-003 as it is not a chat model
* fix installers on mac and linux to include libllmodel versions
",
"contributors":
"
* Adam Treat (Nomic AI)
* Aaron Miller
* Community (beta testers, bug reporters)
"
},
{
"version": "2.4.4",
"notes":
"
* fix buffer overrun in backend
* bugfix for browse for model directory
* dedup of qml code
* revamp settings dialog UI
* add localdocs plugin (beta) feature allowing scanning of local docs
* various other bugfixes and performance improvements
",
"contributors":
"
* Adam Treat (Nomic AI)
* Aaron Miller
* Juuso Alasuutari
* Justin Wang
* Community (beta testers, bug reporters)
"
}
]