diff --git a/gpt4all-chat/CMakeLists.txt b/gpt4all-chat/CMakeLists.txt
index 470ca286..bdd64092 100644
--- a/gpt4all-chat/CMakeLists.txt
+++ b/gpt4all-chat/CMakeLists.txt
@@ -18,7 +18,7 @@ endif()
set(APP_VERSION_MAJOR 2)
set(APP_VERSION_MINOR 8)
-set(APP_VERSION_PATCH 0)
+set(APP_VERSION_PATCH 1)
set(APP_VERSION "${APP_VERSION_MAJOR}.${APP_VERSION_MINOR}.${APP_VERSION_PATCH}")
# Include the binary directory for the generated header file
diff --git a/gpt4all-chat/metadata/release.json b/gpt4all-chat/metadata/release.json
index d99cd806..bdc3b96d 100644
--- a/gpt4all-chat/metadata/release.json
+++ b/gpt4all-chat/metadata/release.json
@@ -810,6 +810,43 @@
* Jared Van Bortel (Nomic AI)
* Adam Treat (Nomic AI)
* Community (beta testers, bug reporters, bindings authors)
+"
+ },
+ {
+ "version": "2.8.0",
+ "notes":
+"
+— What's New —
+* Context Menu: Replace \"Select All\" on message with \"Copy Message\" (PR #2324)
+* Context Menu: Hide Copy/Cut when nothing is selected (PR #2324)
+* Improve speed of context switch after quickly switching between several chats (PR #2343)
+* New Chat: Always switch to the new chat when the button is clicked (PR #2330)
+* New Chat: Always scroll to the top of the list when the button is clicked (PR #2330)
+* Update to latest llama.cpp as of May 9, 2024 (PR #2310)
+* **Add support for the llama.cpp CUDA backend** (PR #2310, PR #2357)
+ * Nomic Vulkan is still used by default, but CUDA devices can now be selected in Settings
+ * When in use: Greatly improved prompt processing and generation speed on some devices
+ * When in use: GPU support for Q5\_0, Q5\_1, Q8\_0, K-quants, I-quants, and Mixtral
+* Add support for InternLM models (PR #2310)
+
+— Fixes —
+* Do not allow sending a message while the LLM is responding (PR #2323)
+* Fix poor quality of generated chat titles with many models (PR #2322)
+* Set the window icon correctly on Windows (PR #2321)
+* Fix a few memory leaks (PR #2328, PR #2348, PR #2310)
+* Do not crash if a model file has no architecture key (PR #2346)
+* Fix several instances of model loading progress displaying incorrectly (PR #2337, PR #2343)
+* New Chat: Fix the new chat being scrolled above the top of the list on startup (PR #2330)
+* macOS: Show a \"Metal\" device option, and actually use the CPU when \"CPU\" is selected (PR #2310)
+* Remove unsupported Mamba, Persimmon, and PLaMo models from the whitelist (PR #2310)
+* Fix GPT4All.desktop being created by offline installers on macOS (PR #2361)
+",
+ "contributors":
+"
+* Jared Van Bortel (Nomic AI)
+* Adam Treat (Nomic AI)
+* Tim453 (`@Tim453`)
+* Community (beta testers, bug reporters, bindings authors)
"
}
]