Commit Graph

1146 Commits

Author SHA1 Message Date
oobabooga
09b0a3aafb Add repetition_penalty 2023-03-31 14:45:17 -03:00
oobabooga
4c27562157 Minor changes 2023-03-31 14:33:46 -03:00
oobabooga
9d1dcf880a General improvements 2023-03-31 14:27:01 -03:00
Thomas Antony
7fa5d96c22 Update to use new llamacpp API 2023-03-30 11:23:05 +01:00
Thomas Antony
79fa2b6d7e Add support for alpaca 2023-03-30 11:23:04 +01:00
Thomas Antony
8953a262cb Add llamacpp to requirements.txt 2023-03-30 11:22:38 +01:00
Thomas Antony
a5f5736e74 Add to text_generation.py 2023-03-30 11:22:38 +01:00
Thomas Antony
7745faa7bb Add llamacpp to models.py 2023-03-30 11:22:37 +01:00
Thomas Antony
7a562481fa Initial version of llamacpp_model.py 2023-03-30 11:22:07 +01:00
Thomas Antony
53ab1e285d Update .gitignore 2023-03-30 11:22:07 +01:00
oobabooga
f0fdab08d3 Increase --chat height 2023-03-30 01:02:11 -03:00
oobabooga
bd65940a48 Increase --chat box height 2023-03-30 00:43:49 -03:00
oobabooga
131753fcf5 Save the sha256sum of downloaded models 2023-03-29 23:28:16 -03:00
oobabooga
a21e580782 Move an import 2023-03-29 22:50:58 -03:00
oobabooga
55755e27b9 Don't hardcode prompts in the settings dict/json 2023-03-29 22:47:01 -03:00
oobabooga
1cb9246160 Adapt to the new model names 2023-03-29 21:47:36 -03:00
oobabooga
0345e04249 Fix "Unknown argument(s): {'verbose': False}" 2023-03-29 21:17:48 -03:00
oobabooga
9104164297
Merge pull request #618 from nikita-skakun/optimize-download-model
Improve download-model.py progress bar with multiple threads
2023-03-29 20:54:19 -03:00
oobabooga
37754164eb Move argparse 2023-03-29 20:47:36 -03:00
oobabooga
6403e72062 Merge branch 'main' into nikita-skakun-optimize-download-model 2023-03-29 20:45:33 -03:00
oobabooga
1445ea86f7 Add --output and better metadata for downloading models 2023-03-29 20:26:44 -03:00
oobabooga
58349f44a0
Handle training exception for unsupported models 2023-03-29 11:55:34 -03:00
oobabooga
a6d0373063
Fix training dataset loading #636 2023-03-29 11:48:17 -03:00
oobabooga
41b58bc47e
Update README.md 2023-03-29 11:02:29 -03:00
oobabooga
3b4447a4fe
Update README.md 2023-03-29 02:24:11 -03:00
oobabooga
5d0b83c341
Update README.md 2023-03-29 02:22:19 -03:00
oobabooga
c2a863f87d
Mention the updated one-click installer 2023-03-29 02:11:51 -03:00
oobabooga
1edfb96778
Fix loading extensions from within the interface 2023-03-28 23:27:02 -03:00
Nikita Skakun
aaa218a102 Remove unused import. 2023-03-28 18:32:49 -07:00
Nikita Skakun
ff515ec2fe Improve progress bar visual style
This commit reverts the performance improvements of the previous commit for for improved visual style of multithreaded progress bars. The style of the progress bar has been modified to take up the same amount of size to align them.
2023-03-28 18:29:20 -07:00
oobabooga
304f812c63 Gracefully handle CUDA out of memory errors with streaming 2023-03-28 19:20:50 -03:00
Nikita Skakun
4d8e101006 Refactor download process to use multiprocessing
The previous implementation used threads to download files in parallel, which could lead to performance issues due to the Global Interpreter Lock (GIL).
This commit refactors the download process to use multiprocessing instead,
which allows for true parallelism across multiple CPUs.
This results in significantly faster downloads, particularly for large models.
2023-03-28 14:24:23 -07:00
oobabooga
b2f356a9ae
Generalize GPTQ_loader, support any model (#615 from mayaeary/feature/gpt-j-4bit-v2)
This includes Pygmalion 4bit
2023-03-28 18:00:09 -03:00
oobabooga
010b259dde Update documentation 2023-03-28 17:46:00 -03:00
oobabooga
0bec15ebcd Reorder imports 2023-03-28 17:34:15 -03:00
Maya Eary
41ec682834 Disable kernel threshold for gpt-j 2023-03-28 22:45:38 +03:00
Maya
1ac003d41c
Merge branch 'oobabooga:main' into feature/gpt-j-4bit-v2 2023-03-28 22:30:39 +03:00
oobabooga
aebd3cf110
Merge pull request #616 from mayaeary/fix/api-convert-params
Fixes for api server - chat mode and integer temperature
2023-03-28 15:21:58 -03:00
Maya Eary
d1377c37af Fixes for api server - chat mode and integer temperature 2023-03-28 20:57:16 +03:00
Maya Eary
1c075d8d21 Fix typo 2023-03-28 20:43:50 +03:00
Maya Eary
c8207d474f Generalized load_quantized 2023-03-28 20:38:55 +03:00
oobabooga
cac577d99f Fix interface reloading 2023-03-28 13:25:58 -03:00
oobabooga
88ad86249d Remove unnecessary file 2023-03-28 13:19:52 -03:00
oobabooga
91aa5b460e If both .pt and .safetensors are present, download only safetensors 2023-03-28 13:08:38 -03:00
oobabooga
8579fe51dd Fix new lines in the HTML tab 2023-03-28 12:59:34 -03:00
oobabooga
46f6536fae
Merge pull request #570 from mcmonkey4eva/add-train-lora-tab
LoRA Training Tab
2023-03-28 02:53:51 -03:00
Alex "mcmonkey" Goodwin
b0f05046b3 remove duplicate import 2023-03-27 22:50:37 -07:00
Alex "mcmonkey" Goodwin
e817fac542 better defaults 2023-03-27 22:29:23 -07:00
Alex "mcmonkey" Goodwin
9cc811a0e6 fix LoRA path typo in #549 2023-03-27 22:16:40 -07:00
Alex "mcmonkey" Goodwin
2e08af4edf implement initial Raw Text File Input
also bump default Rank & Alpha for values that will make sense in testing if you don't know what you're doing and leave the defaults.
2023-03-27 22:15:32 -07:00