AT
48275d0dcc
Dlopen backend 5 ( #779 )
...
Major change to the backend that allows for pluggable versions of llama.cpp/ggml. This was squashed merged from dlopen_backend_5 where the history is preserved.
2023-05-31 17:04:01 -04:00
Adam Treat
912cb2a842
Get rid of blocking behavior for regenerate response.
2023-05-30 18:17:59 -04:00
FoivosC
337c7fecac
construct and return the correct reponse when the request is a chat completion
2023-05-30 16:34:23 -04:00
Adam Treat
748e7977ca
Generate the new prompt/response pair before model loading in server mode.
2023-05-16 10:31:55 -04:00
Adam Treat
b71c0ac3bd
The server has different lifetime mgmt than the other chats.
2023-05-13 19:34:54 -04:00
Adam Treat
2989b74d43
httpserver
2023-05-13 19:07:06 -04:00