diff --git a/gpt4all-bindings/python/docs/gpt4all_chat.md b/gpt4all-bindings/python/docs/gpt4all_chat.md index 2638631d..7e3f2856 100644 --- a/gpt4all-bindings/python/docs/gpt4all_chat.md +++ b/gpt4all-bindings/python/docs/gpt4all_chat.md @@ -1,13 +1,40 @@ -# GPT4All Chat Client +# GPT4All Chat UI The [GPT4All Chat Client](https://gpt4all.io) lets you easily interact with any local large language model. It is optimized to run 7-13B parameter LLMs on the CPU's of any computer running OSX/Windows/Linux. +## Plugins +GPT4All Chat Plugins allow you to expand the capabilities of Local LLMs. All plugins are compatible with the +chat clients server mode. + +### LocalDocs Plugin (Chat With Your Data) +LocalDocs is a GPT4All plugin that allows you to chat with your local files and data. +It allows you to utilize powerful local LLMs to chat with private data without any data leaving your computer or server. +When using LocalDocs, your LLM will cite the sources that most likely contributed to a given output. Note, even an LLM equipped with LocalDocs can hallucinate. + +#### Enabling LocalDocs +1. Install the latest version of GPT4All Chat from https://gpt4all.io. +2. Go to `Settings > the LocalDocs tab`. +3. Configure a collection (folder) on your computer that contains the files your LLM should have access to. You can alter the contents of the folder/directory at anytime. As you +add more files to your collection, your LLM will dynamically be able to access them. +4. Spin up a chat session with any LLM (including external ones like ChatGPT but warning data will leave your machine!) +5. At the top right, click the database icon and select which collection you want your LLM to know about. +6. Start chatting! + +### How it works +LocalDocs works by maintaining an index of all data in the directory your collection is linked to. This index +consists of small chunks of each document that the LLM can receive as additional input when you ask it a question. +This helps it respond to your queries with knowledge about the contents of your data. +The number of chunks and the size of each chunk can be configured in the LocalDocs plugin settings tab. +For indexing speed purposes, LocalDocs uses pre-deep-learning n-gram and tfidf based retrieval when deciding +what documents your LLM should have as context in response to a question. You'll find its of comparable quality +with embedding based retrieval approaches but magnitudes faster to ingest data. Don't worry, embedding based semantic +search for retrieval is on the roadmap for those with more powerful computers - pick up the feature on Github! -## GPT4All Chat Server Mode +## Server Mode GPT4All Chat comes with a built-in server mode allowing you to programmatically interact with any supported local LLM through a *very familiar* HTTP API. You can find the API documentation [here](https://platform.openai.com/docs/api-reference/completions).