mirror of
https://github.com/nomic-ai/gpt4all.git
synced 2024-10-01 01:06:10 -04:00
Update README.md
This commit is contained in:
parent
252676ff05
commit
2db43570ae
@ -28,6 +28,14 @@ Clone this repository down and place the quantized model in the `chat` directory
|
|||||||
|
|
||||||
To compile for custom hardware, see our fork of the [Alpaca C++](https://github.com/zanussbaum/gpt4all.cpp) repo.
|
To compile for custom hardware, see our fork of the [Alpaca C++](https://github.com/zanussbaum/gpt4all.cpp) repo.
|
||||||
|
|
||||||
|
-----------
|
||||||
|
|
||||||
|
[Secret Unfiltered Checkpoint](https://the-eye.eu/public/AI/models/nomic-ai/gpt4all/gpt4all-lora-unfiltered-quantized.bin)
|
||||||
|
|
||||||
|
This model had all refusal to answer responses removed from training. Try it with:
|
||||||
|
- `cd chat;./gpt4all-lora-quantized-OSX-m1 -m gpt4all-lora-unfiltered-quantized.bin`
|
||||||
|
|
||||||
|
-----------
|
||||||
Note: the full model on GPU (16GB of RAM required) performs much better in our qualitative evaluations.
|
Note: the full model on GPU (16GB of RAM required) performs much better in our qualitative evaluations.
|
||||||
|
|
||||||
# Reproducibility
|
# Reproducibility
|
||||||
|
Loading…
Reference in New Issue
Block a user