llama.cpp/llama_cpp
2023-11-10 02:49:27 -05:00
..
server Fix: default max_tokens matches openai api (16 for completion, max length for chat completion) 2023-11-10 02:49:27 -05:00
__init__.py Bump version 2023-11-08 00:54:54 -05:00
_utils.py Clean up stdout / stderr suppression 2023-11-03 13:02:15 -04:00
llama.py Fix: default max_tokens matches openai api (16 for completion, max length for chat completion) 2023-11-10 02:49:27 -05:00
llama_chat_format.py Bugfix: missing response_format for functionary and llava chat handlers 2023-11-09 00:55:23 -05:00
llama_cpp.py Update llama.cpp 2023-11-05 16:57:10 -05:00
llama_grammar.py Fix built in GBNF grammar rules 2023-11-08 00:06:22 -05:00
llama_types.py Add JSON mode support. Closes #881 2023-11-08 00:07:16 -05:00
llava_cpp.py Multimodal Support (Llava 1.5) (#821) 2023-11-07 22:48:51 -05:00
py.typed Add py.typed 2023-08-11 09:58:48 +02:00