1. You need to have the conda environment manager installed into your system. If you don't have it already, [get miniconda here](https://docs.conda.io/en/latest/miniconda.html).
For AMD GPUs, you need the ROCm version of pytorch. If you don't have any GPU and want to run the webui in CPU mode, you just need the stock pytorch and should use this command instead:
Models should be placed under `models/model-name`. For instance, `models/gpt-j-6B` for [gpt-j-6B](https://huggingface.co/EleutherAI/gpt-j-6B/tree/main).
[Hugging Face](https://huggingface.co/models?pipeline_tag=text-generation&sort=downloads) is the main place to download models. These are some noteworthy examples:
[GPT-4chan](https://huggingface.co/ykilcher/gpt-4chan) has been shut down from Hugging Face, so you need to download it elsewhere. You have two options:
You don't really need all of GPT-J's files, just the tokenizer files, but you might as well download the whole thing. Those files will be automatically detected when you attempt to load GPT-4chan.
The output model will be saved to `torch-dumps/model-name.pt`. When you load a new model, the webui first looks for this .pt file; if it is not found, it loads the model as usual from `models/model-name`.
| `-h`, `--help` | show this help message and exit |
| `--model MODEL` | Name of the model to load by default. |
| `--notebook` | Launch the webui in notebook mode, where the output is written to the same text box as the input. |
| `--chat` | Launch the webui in chat mode.|
| `--cai-chat` | Launch the webui in chat mode with a style similar to Character.AI's. If the file profile.png exists in the same folder as server.py, this image will be used as the bot's profile picture.|
| `--cpu` | Use the CPU to generate text.|
| `--auto-devices` | Automatically split the model across the available GPU(s) and CPU.|
| `--load-in-8bit` | Load the model with 8-bit precision.|
| `--no-listen` | Make the webui unreachable from your local network.|
| `--settings-file SETTINGS\_FILE` | Load default interface settings from this json file. See settings-template.json for an example.|
Check the [wiki](https://github.com/oobabooga/text-generation-webui/wiki/System-requirements) for some examples of VRAM and RAM usage in both GPU and CPU mode.