Commit Graph

27 Commits

Author SHA1 Message Date
Zach Nussbaum
2f6ecbe798 feat: build works + tokenizer 2023-05-08 12:21:30 -04:00
Zach Nussbaum
525b703984 feat: add ln 2, rename vars 2023-05-08 12:21:30 -04:00
Zach Nussbaum
aef524b460 feat: mpt wip 2023-05-08 12:21:30 -04:00
Adam Treat
159053be5a Scaffolding for the mpt <-> ggml project. 2023-05-08 12:21:30 -04:00
Adam Treat
6d943917f1 Fail early/gracefully if incompatible hardware detected. And default to universal builds on mac. 2023-05-08 08:23:00 -04:00
Adam Treat
7b66cb7119 Add debug for chatllm model loading and fix order of getting rid of the
dummy chat when no models are restored.
2023-05-07 14:40:02 -04:00
Adam Treat
d0d5d84e06 Add reverse prompt support for gptj too. 2023-05-05 11:16:24 -04:00
Adam Treat
659442394f Persistent state for gpt-j models too. 2023-05-05 10:00:17 -04:00
Aaron Miller
019f6d0103 include <cstdint> in llmodel.h 2023-05-04 20:36:19 -04:00
Adam Treat
f291853e51 First attempt at providing a persistent chat list experience.
Limitations:

1) Context is not restored for gpt-j models
2) When you switch between different model types in an existing chat
   the context and all the conversation is lost
3) The settings are not chat or conversation specific
4) The sizes of the chat persisted files are very large due to how much
   data the llama.cpp backend tries to persist. Need to investigate how
   we can shrink this.
2023-05-04 15:31:41 -04:00
Adam Treat
82c1d08b33 Add reverse prompts for llama models. 2023-05-03 11:58:26 -04:00
Adam Treat
8fe60c29fb Don't set the app version in the llmodel. 2023-04-29 10:31:12 -04:00
Adam Treat
69f92d8ea8 Load models from filepath only. 2023-04-28 20:15:10 -04:00
Adam Treat
d982dc0529 Update to latest llama.cpp 2023-04-28 11:03:16 -04:00
Adam Treat
5a7d40f604 Move the saving of the tokens to the impl and not the callbacks responsibility. 2023-04-27 11:16:51 -04:00
Adam Treat
ba4b28fcd5 Move the promptCallback to own function. 2023-04-27 11:08:15 -04:00
Adam Treat
0e9f85bcda Provide an initial impl. of the C interface. NOTE: has not been tested. 2023-04-27 09:43:24 -04:00
Adam Treat
b19d2f2c21 Add this and unbreak the build. 2023-04-26 22:45:10 -04:00
Adam Treat
ee5c58c26c Initial support for opt-in telemetry. 2023-04-26 22:05:56 -04:00
Adam Treat
a3d97fa009 Don't crash when prompt is too large. 2023-04-26 19:08:37 -04:00
Adam Treat
7da3bc07cc Update llama.cpp submodule to latest. 2023-04-26 11:50:05 -04:00
Adam Treat
fd0f92a94e Clean up the docs a bit more still. 2023-04-26 08:22:38 -04:00
Adam Treat
c89096ccb4 Clean up the docs a bit more. 2023-04-26 08:22:38 -04:00
Adam Treat
ac7ecd2cef Clean up the docs a bit. 2023-04-26 08:22:38 -04:00
Adam Treat
832b5d1a96 Only need one opaque pointer. 2023-04-26 08:22:38 -04:00
Adam Treat
102f68b18c Fixup the api a bit. 2023-04-26 08:22:38 -04:00
Adam Treat
3c9139b5d2 Move the backend code into own subdirectory and make it a shared library. Begin fleshing out the C api wrapper that bindings can use. 2023-04-26 08:22:38 -04:00