llama.cpp/llama_cpp
gmcgoldr 09a8406c83
Fix streaming doesn't return finish reason (#798)
When streaming the yield that contains the finish can be skipped. This change ensures that yield isn't skipped.
2023-10-19 02:55:56 -04:00
..
server update value check for n_gpu_layers field (#826) 2023-10-18 18:25:25 -04:00
__init__.py Bump version 2023-09-30 16:04:46 -04:00
llama.py Fix streaming doesn't return finish reason (#798) 2023-10-19 02:55:56 -04:00
llama_chat_format.py Fix repeat greeting (#808) 2023-10-15 13:52:21 -04:00
llama_cpp.py Update llama.cpp 2023-10-19 02:55:08 -04:00
llama_grammar.py Fix typos in llama_grammar 2023-08-17 21:00:44 +09:00
llama_types.py Update llama_types and names to match openai api 2023-09-20 15:38:26 -04:00
py.typed Add py.typed 2023-08-11 09:58:48 +02:00
utils.py Suppress llama.cpp output when loading model. 2023-07-28 14:45:18 -04:00