Commit Graph

1071 Commits

Author SHA1 Message Date
oobabooga
7ab45fb54a
Merge pull request #296 from luiscosio/patch-1
Fix for issue #282
2023-03-13 14:45:58 -03:00
Luis Cosio
435a69e357
Fix for issue #282
RuntimeError: Tensors must have same number of dimensions: got 3 and 4
2023-03-13 11:41:35 -06:00
Ayanami Rei
e1c952c41c make argument non case-sensitive 2023-03-13 20:22:38 +03:00
Ayanami Rei
b746250b2f Update README 2023-03-13 20:20:45 +03:00
Ayanami Rei
3c9afd5ca3 rename method 2023-03-13 20:14:40 +03:00
Ayanami Rei
1b99ed61bc add argument --gptq-model-type and remove duplicate arguments 2023-03-13 20:01:34 +03:00
Ayanami Rei
edbc61139f use new quant loader 2023-03-13 20:00:38 +03:00
Ayanami Rei
345b6dee8c refactor quant models loader and add support of OPT 2023-03-13 19:59:57 +03:00
oobabooga
66b6971b61 Update README 2023-03-13 12:44:18 -03:00
oobabooga
ddea518e0f Document --auto-launch 2023-03-13 12:43:33 -03:00
oobabooga
d97bfb8713
Update README.md 2023-03-13 12:39:33 -03:00
oobabooga
372363bc3d Fix GPTQ load_quant call on Windows 2023-03-13 12:07:02 -03:00
oobabooga
bdff37f0bb
Update README.md 2023-03-13 11:05:51 -03:00
oobabooga
b6098e9ccb
Merge pull request #275 from stefanhamburger/patch-1
Fix: tuple object does not support item assignment
2023-03-13 11:01:31 -03:00
oobabooga
72757088fa
Create FUNDING.yml 2023-03-13 10:55:00 -03:00
oobabooga
0c224cf4f4 Fix GALACTICA (#285) 2023-03-13 10:32:28 -03:00
stefanhamburger
91c2a8e88d
Fix: tuple object does not support item assignment 2023-03-13 07:42:09 +01:00
oobabooga
2c4699a7e9 Change a comment 2023-03-13 00:20:02 -03:00
oobabooga
0a7acb3bd9 Remove redundant comments 2023-03-13 00:12:21 -03:00
oobabooga
77294b27dd Use str(Path) instead of os.path.abspath(Path) 2023-03-13 00:08:01 -03:00
oobabooga
b9e0712b92 Fix Open Assistant 2023-03-12 23:58:25 -03:00
oobabooga
1ddcd4d0ba Clean up silero_tts
This should only be used with --no-stream.

The shared.still_streaming implementation was faulty by design:
output_modifier should never be called when streaming is already over.
2023-03-12 23:42:49 -03:00
oobabooga
a95592fc56 Add back a progress indicator to --no-stream 2023-03-12 20:38:40 -03:00
oobabooga
d168b6e1f7
Update README.md 2023-03-12 17:54:07 -03:00
EliasVincent
48aa52849b use Gradio microphone input instead 2023-03-12 21:03:07 +01:00
oobabooga
54e8f0c31f
Update README.md 2023-03-12 16:58:00 -03:00
HideLord
683556f411 Adding markdown support and slight refactoring. 2023-03-12 21:34:09 +02:00
oobabooga
cebe8b390d Remove useless "substring_found" variable 2023-03-12 15:50:38 -03:00
oobabooga
4bcd675ccd Add *Is typing...* to regenerate as well 2023-03-12 15:23:33 -03:00
Elias Vincent Simon
3b4145966d
Merge branch 'oobabooga:main' into stt-extension 2023-03-12 19:19:43 +01:00
oobabooga
3375eaece0 Update README 2023-03-12 15:01:32 -03:00
oobabooga
c7aa51faa6 Use a list of eos_tokens instead of just a number
This might be the cause of LLaMA ramblings that some people have experienced.
2023-03-12 14:54:58 -03:00
oobabooga
17210ff88f
Update README.md 2023-03-12 14:31:24 -03:00
oobabooga
441e993c51 Bump accelerate, RWKV and safetensors 2023-03-12 14:25:14 -03:00
oobabooga
d8bea766d7
Merge pull request #192 from xanthousm/main
Add text generation stream status to shared module, use for better TTS with auto-play
2023-03-12 13:40:16 -03:00
oobabooga
4066ab4c0c Reorder the imports 2023-03-12 13:36:18 -03:00
oobabooga
4dc1d8c091
Update README.md 2023-03-12 12:46:53 -03:00
oobabooga
901dcba9b4
Merge pull request #263 from HideLord/main
Fixing compatibility with GPTQ repository
2023-03-12 12:42:08 -03:00
oobabooga
fda376d9c3 Use os.path.abspath() instead of str() 2023-03-12 12:41:04 -03:00
HideLord
8403152257 Fixing compatibility with GPTQ repo commit 2f667f7da051967566a5fb0546f8614bcd3a1ccd. Expects string and breaks on 2023-03-12 17:28:15 +02:00
HideLord
a27f98dbbc Merge branch 'main' of https://github.com/HideLord/text-generation-webui 2023-03-12 16:51:04 +02:00
oobabooga
f3b00dd165
Merge pull request #224 from ItsLogic/llama-bits
Allow users to load 2, 3 and 4 bit llama models
2023-03-12 11:23:50 -03:00
oobabooga
89e9493509 Update README 2023-03-12 11:23:20 -03:00
oobabooga
65dda28c9d Rename --llama-bits to --gptq-bits 2023-03-12 11:19:07 -03:00
oobabooga
fed3617f07 Move LLaMA 4-bit into a separate file 2023-03-12 11:12:34 -03:00
oobabooga
0ac562bdba Add a default prompt for OpenAssistant oasst-sft-1-pythia-12b #253 2023-03-12 10:46:16 -03:00
oobabooga
78901d522b Remove unused imports 2023-03-12 08:59:05 -03:00
oobabooga
35c14f31b2
Merge pull request #259 from hieultp/patch-1
Fix typo error in LLaMa prompts
2023-03-12 08:52:02 -03:00
oobabooga
3c25557ef0 Add tqdm to requirements.txt 2023-03-12 08:48:16 -03:00
Phuoc-Hieu Le
781c09235c
Fix typo error in script.py 2023-03-12 15:21:50 +07:00