From d0aedfcff66ca243a5f817a940bfe33095422ebe Mon Sep 17 00:00:00 2001 From: Andrei Betlen Date: Mon, 18 Dec 2023 18:12:49 -0500 Subject: [PATCH] Fix type annotation --- llama_cpp/llama_cpp.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama_cpp/llama_cpp.py b/llama_cpp/llama_cpp.py index 172ddf5..b250c00 100644 --- a/llama_cpp/llama_cpp.py +++ b/llama_cpp/llama_cpp.py @@ -1059,7 +1059,7 @@ _lib.llama_kv_cache_view_init.restype = llama_kv_cache_view # // Free a KV cache view. (use only for debugging purposes) # LLAMA_API void llama_kv_cache_view_free(struct llama_kv_cache_view * view); -def llama_kv_cache_view_free(view: llama_kv_cache_view_p): +def llama_kv_cache_view_free(view: "ctypes._Pointer[llama_kv_cache_view]"): # type: ignore """Free a KV cache view. (use only for debugging purposes)""" return _lib.llama_kv_cache_view_free(view)