Carl Kenner
c86231377b
Wizard Mega, Ziya, KoAlpaca, OpenBuddy, Chinese-Vicuna, Vigogne, Bactrian, H2O support, fix Baize ( #2159 )
2023-05-19 11:42:41 -03:00
Alex "mcmonkey" Goodwin
1f50dbe352
Experimental jank multiGPU inference that's 2x faster than native somehow ( #2100 )
2023-05-17 10:41:09 -03:00
Andrei
e657dd342d
Add in-memory cache support for llama.cpp ( #1936 )
2023-05-15 20:19:55 -03:00
AlphaAtlas
071f0776ad
Add llama.cpp GPU offload option ( #2060 )
2023-05-14 22:58:11 -03:00
oobabooga
23d3f6909a
Update README.md
2023-05-11 10:21:20 -03:00
oobabooga
2930e5a895
Update README.md
2023-05-11 10:04:38 -03:00
oobabooga
0ff38c994e
Update README.md
2023-05-11 09:58:58 -03:00
oobabooga
e6959a5d9a
Update README.md
2023-05-11 09:54:22 -03:00
oobabooga
dcfd09b61e
Update README.md
2023-05-11 09:49:57 -03:00
oobabooga
7a49ceab29
Update README.md
2023-05-11 09:42:39 -03:00
oobabooga
57dc44a995
Update README.md
2023-05-10 12:48:25 -03:00
oobabooga
181b102521
Update README.md
2023-05-10 12:09:47 -03:00
Carl Kenner
814f754451
Support for MPT, INCITE, WizardLM, StableLM, Galactica, Vicuna, Guanaco, and Baize instruction following ( #1596 )
2023-05-09 20:37:31 -03:00
Wojtab
e9e75a9ec7
Generalize multimodality (llava/minigpt4 7b and 13b now supported) ( #1741 )
2023-05-09 20:18:02 -03:00
oobabooga
00e333d790
Add MOSS support
2023-05-04 23:20:34 -03:00
oobabooga
b6ff138084
Add --checkpoint argument for GPTQ
2023-05-04 15:17:20 -03:00
Ahmed Said
fbcd32988e
added no_mmap & mlock parameters to llama.cpp and removed llamacpp_model_alternative ( #1649 )
...
---------
Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-05-02 18:25:28 -03:00
oobabooga
f39c99fa14
Load more than one LoRA with --lora, fix a bug
2023-04-25 22:58:48 -03:00
oobabooga
b6af2e56a2
Add --character flag, add character to settings.json
2023-04-24 13:19:42 -03:00
eiery
78d1977ebf
add n_batch support for llama.cpp ( #1115 )
2023-04-24 03:46:18 -03:00
Andy Salerno
654933c634
New universal API with streaming/blocking endpoints ( #990 )
...
Previous title: Add api_streaming extension and update api-example-stream to use it
* Merge with latest main
* Add parameter capturing encoder_repetition_penalty
* Change some defaults, minor fixes
* Add --api, --public-api flags
* remove unneeded/broken comment from blocking API startup. The comment is already correctly emitted in try_start_cloudflared by calling the lambda we pass in.
* Update on_start message for blocking_api, it should say 'non-streaming' and not 'streaming'
* Update the API examples
* Change a comment
* Update README
* Remove the gradio API
* Remove unused import
* Minor change
* Remove unused import
---------
Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-04-23 15:52:43 -03:00
oobabooga
7438f4f6ba
Change GPTQ triton default settings
2023-04-22 12:27:30 -03:00
oobabooga
fe02281477
Update README.md
2023-04-22 03:05:00 -03:00
oobabooga
038fa3eb39
Update README.md
2023-04-22 02:46:07 -03:00
oobabooga
505c2c73e8
Update README.md
2023-04-22 00:11:27 -03:00
oobabooga
f8da9a0424
Update README.md
2023-04-18 20:25:08 -03:00
oobabooga
c3f6e65554
Update README.md
2023-04-18 20:23:31 -03:00
oobabooga
eb15193327
Update README.md
2023-04-18 13:07:08 -03:00
oobabooga
7fbfc489e2
Update README.md
2023-04-18 12:56:37 -03:00
oobabooga
f559f9595b
Update README.md
2023-04-18 12:54:09 -03:00
loeken
89e22d4d6a
added windows/docker docs ( #1027 )
2023-04-18 12:47:43 -03:00
oobabooga
8275989f03
Add new 1-click installers for Linux and MacOS
2023-04-18 02:40:36 -03:00
oobabooga
301c687c64
Update README.md
2023-04-17 11:25:26 -03:00
oobabooga
89bc540557
Update README
2023-04-17 10:55:35 -03:00
practicaldreamer
3961f49524
Add note about --no-fused_mlp ignoring --gpu-memory ( #1301 )
2023-04-17 10:46:37 -03:00
sgsdxzy
b57ffc2ec9
Update to support GPTQ triton commit c90adef ( #1229 )
2023-04-17 01:11:18 -03:00
oobabooga
3e5cdd005f
Update README.md
2023-04-16 23:28:59 -03:00
oobabooga
39099663a0
Add 4-bit LoRA support ( #1200 )
2023-04-16 23:26:52 -03:00
oobabooga
705121161b
Update README.md
2023-04-16 20:03:03 -03:00
oobabooga
50c55a51fc
Update README.md
2023-04-16 19:22:31 -03:00
Forkoz
c6fe1ced01
Add ChatGLM support ( #1256 )
...
---------
Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-04-16 19:15:03 -03:00
oobabooga
c96529a1b3
Update README.md
2023-04-16 17:00:03 -03:00
oobabooga
004f275efe
Update README.md
2023-04-14 23:36:56 -03:00
oobabooga
83964ed354
Update README.md
2023-04-14 23:33:54 -03:00
oobabooga
c41037db68
Update README.md
2023-04-14 23:32:39 -03:00
v0xie
9d66957207
Add --listen-host launch option ( #1122 )
2023-04-13 21:35:08 -03:00
oobabooga
403be8a27f
Update README.md
2023-04-13 21:23:35 -03:00
Light
97e67d136b
Update README.md
2023-04-13 21:00:58 +08:00
Light
15d5a043f2
Merge remote-tracking branch 'origin/main' into triton
2023-04-13 19:38:51 +08:00
oobabooga
7dfbe54f42
Add --model-menu option
2023-04-12 21:24:26 -03:00