llama.cpp/llama_cpp
Brandon Roberts 62944df142
Bugfix: Remove f16_kv, add offload_kqv field (#1019)
F16_KV appears to have been removed here: af99c6fbfc

This addresses two issues:

 - #995 which just requests to add the KV cache offloading param
 - #1006 a NULL ptr exception when using the embeddings (introduced by
   leaving f16_kv in the fields struct)
2023-12-18 14:27:11 -05:00
..
server Bugfix: Remove f16_kv, add offload_kqv field (#1019) 2023-12-18 14:27:11 -05:00
__init__.py Bump version 2023-12-13 21:52:29 -05:00
_utils.py Fix UnsupportedOperation: fileno in suppress_stdout_stderr (#961) 2023-12-11 20:44:51 -05:00
llama.py Bugfix: Remove f16_kv, add offload_kqv field (#1019) 2023-12-18 14:27:11 -05:00
llama_chat_format.py Add qwen chat format (#1005) 2023-12-13 21:43:43 -05:00
llama_cpp.py Bugfix: Remove f16_kv, add offload_kqv field (#1019) 2023-12-18 14:27:11 -05:00
llama_grammar.py Add from_json_schema to LlamaGrammar 2023-11-23 00:27:00 -05:00
llama_types.py Add missing tool_calls finish_reason 2023-11-10 02:51:06 -05:00
llava_cpp.py Make building llava optional 2023-11-28 04:55:21 -05:00
py.typed Add py.typed 2023-08-11 09:58:48 +02:00