CharlesCNorton
ea4c5546d8
fix: typos in reset context comments ( #2453 )
...
Corrected two typos and a grammatical issue in the comments within the reset context code. Specifically:
- Changed "indictate" to "indicate"
- Corrected "me might" to "we might"
Signed-off-by: CharlesCNorton <135471798+CharlesCNorton@users.noreply.github.com>
2024-07-09 11:18:37 -04:00
Jared Van Bortel
01870b4a46
chat: fix blank device in UI and improve Mixpanel reporting ( #2409 )
...
Also remove LLModel::hasGPUDevice.
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
2024-06-26 15:26:27 -04:00
AT
9273b49b62
chat: major UI redesign for v3.0.0 ( #2396 )
...
Signed-off-by: Adam Treat <treat.adam@gmail.com>
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
Co-authored-by: Jared Van Bortel <jared@nomic.ai>
2024-06-24 18:49:23 -04:00
Jared Van Bortel
41c9013fa4
chat: don't use incomplete types with signals/slots/Q_INVOKABLE ( #2408 )
...
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
2024-06-06 11:59:28 -04:00
Jared Van Bortel
d3d777bc51
chat: fix #includes with include-what-you-use ( #2401 )
...
Also use qGuiApp instead of qApp.
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
2024-06-04 14:47:11 -04:00
Jared Van Bortel
7e1e00f331
chat: fix issues with quickly switching between multiple chats ( #2343 )
...
* prevent load progress from getting out of sync with the current chat
* fix memory leak on exit if the LLModelStore contains a model
* do not report cancellation as a failure in console/Mixpanel
* show "waiting for model" separately from "switching context" in UI
* do not show lower "reload" button on error
* skip context switch if unload is pending
* skip unnecessary calls to LLModel::saveState
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
2024-05-15 14:07:03 -04:00
Jared Van Bortel
c622921894
improve mixpanel usage statistics ( #2238 )
...
Other changes:
- Always display first start dialog if privacy options are unset (e.g. if the user closed GPT4All without selecting them)
- LocalDocs scanQueue is now always deferred
- Fix a potential crash in magic_match
- LocalDocs indexing is now started after the first start dialog is dismissed so usage stats are included
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
2024-04-25 13:16:52 -04:00
Adam Treat
94a9943782
Change the behavior of show references setting for localdocs.
...
Signed-off-by: Adam Treat <treat.adam@gmail.com>
2024-04-15 14:30:26 -05:00
Adam Treat
17dee02287
Fix for issue #2080 where the GUI appears to hang when a chat with a large
...
model is deleted. There is no reason to save the context for a chat that
is being deleted.
Signed-off-by: Adam Treat <treat.adam@gmail.com>
2024-03-06 16:52:17 -06:00
Jared Van Bortel
44717682a7
chat: implement display of model loading warnings ( #2034 )
...
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
2024-03-06 17:14:54 -05:00
Adam Treat
67099f80ba
Add comment to make this clear.
...
Signed-off-by: Adam Treat <treat.adam@gmail.com>
2024-02-21 10:15:20 -06:00
Adam Treat
d948a4f2ee
Complete revamp of model loading to allow for more discreet control by
...
the user of the models loading behavior.
Signed-off-by: Adam Treat <treat.adam@gmail.com>
2024-02-21 10:15:20 -06:00
Jared Van Bortel
29d2c936d1
chat: don't show "retrieving localdocs" for zero collections ( #1874 )
...
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
2024-01-29 13:57:42 -05:00
Gerhard Stein
3e99b90c0b
Some cleanps
2024-01-03 08:41:40 -06:00
Jared Van Bortel
0600f551b3
chatllm: do not attempt to serialize incompatible state ( #1742 )
2023-12-12 11:45:03 -05:00
Adam Treat
371e2a5cbc
LocalDocs version 2 with text embeddings.
2023-11-17 11:59:31 -05:00
Adam Treat
f529d55380
Move this logic to QML.
2023-10-30 09:57:21 -04:00
Adam Treat
131cfcdeae
Don't regenerate the name for deserialized chats.
2023-10-28 16:41:23 -04:00
Adam Treat
908aec27fe
Always save chats to disk, but save them as text by default. This also changes
...
the UI behavior to always open a 'New Chat' and setting it as current instead
of setting a restored chat as current. This improves usability by not requiring
the user to wait if they want to immediately start chatting.
2023-10-12 07:52:11 -04:00
Adam Treat
f0742c22f4
Restore state from text if necessary.
2023-10-11 09:16:02 -04:00
Cebtenzzre
2eb83b9f2a
chat: report reason for fallback to CPU
2023-10-05 18:16:19 -04:00
Adam Treat
1fa67a585c
Report the actual device we're using.
2023-09-14 08:25:37 -04:00
Adam Treat
9dccc96e70
Immediately signal when the model is in a new loading state.
2023-07-11 15:10:59 -04:00
Adam Treat
eab92a9d73
Fix typo and add new show references setting to localdocs.
2023-07-05 19:41:23 -04:00
Adam Treat
6d9cdf228c
Huge change that completely revamps the settings dialog and implements
...
per model settings as well as the ability to clone a model into a "character."
This also implements system prompts as well as quite a few bugfixes for
instance this fixes chatgpt.
2023-07-05 15:51:42 -04:00
Adam Treat
285aa50b60
Consolidate generation and application settings on the new settings object.
2023-06-28 20:36:43 -03:00
Adam Treat
64e98b8ea9
Fix bug with model loading on initial load.
2023-06-26 14:14:49 -04:00
Adam Treat
7f01b153b3
Modellist temp
2023-06-26 14:14:46 -04:00
Adam Treat
c8a590bc6f
Get rid of last blocking operations and make the chat/llm thread safe.
2023-06-20 18:18:10 -03:00
Adam Treat
7d2ce06029
Start working on more thread safety and model load error handling.
2023-06-20 14:39:22 -03:00
Adam Treat
a3a6a20146
Don't store db results in ChatLLM.
2023-06-19 15:38:01 -07:00
AT
2b6cc99a31
Show token generation speed in gui. ( #1020 )
2023-06-19 14:34:53 -04:00
Adam Treat
782e1e77a4
Fix up model names that don't begin with 'ggml-'
2023-06-16 14:43:14 -04:00
AT
a576220b18
Support loading files if 'ggml' is found anywhere in the name not just at ( #1001 )
...
the beginning and add deprecated flag to models.json so older versions will
show a model, but later versions don't. This will allow us to transition
away from models < ggmlv2 and still allow older installs of gpt4all to work.
2023-06-16 11:09:33 -04:00
Adam Treat
9f590db98d
Better error handling when the model fails to load.
2023-06-04 14:55:05 -04:00
niansa/tuxifan
f3564ac6b9
Fixed tons of warnings and clazy findings ( #811 )
2023-06-02 15:46:41 -04:00
Adam Treat
28944ac01b
Fix for stale references after we regenerate.
2023-06-01 17:23:57 -04:00
Adam Treat
aea94f756d
Better name for database results.
2023-06-01 17:14:17 -04:00
Adam Treat
f62e439a2d
Make localdocs work with server mode.
2023-06-01 17:14:17 -04:00
Adam Treat
912cb2a842
Get rid of blocking behavior for regenerate response.
2023-05-30 18:17:59 -04:00
Adam Treat
9b0629db8b
Add context link to references.
2023-05-25 11:28:06 -04:00
Adam Treat
db9eecdce4
Store the references separately so they are not sent to datalake.
2023-05-25 11:28:06 -04:00
Adam Treat
b5380c9b7f
Adds the collections to serialize and implement references for localdocs.
2023-05-25 11:28:06 -04:00
Adam Treat
01b8c7617f
Add more of the UI for selecting collections for chats.
2023-05-25 11:28:06 -04:00
Adam Treat
c800291e7f
Add prompt processing and localdocs to the busy indicator in UI.
2023-05-25 11:28:06 -04:00
Adam Treat
7e42af5f33
localdocs
2023-05-25 11:28:06 -04:00
Adam Treat
f931de21c5
Add save/restore to chatgpt chats and allow serialize/deseralize from disk.
2023-05-16 10:31:55 -04:00
Adam Treat
dd27c10f54
Preliminary support for chatgpt models.
2023-05-16 10:31:55 -04:00
Adam Treat
ddc24acf33
Much better memory mgmt for multi-threaded model loading/unloading.
2023-05-13 19:10:56 -04:00
Adam Treat
2989b74d43
httpserver
2023-05-13 19:07:06 -04:00