From b703aad79e7404bd30a5041f95879fe083fd29a8 Mon Sep 17 00:00:00 2001 From: Andrei Betlen Date: Mon, 18 Dec 2023 18:13:37 -0500 Subject: [PATCH] Fix type annotation --- llama_cpp/llama_cpp.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama_cpp/llama_cpp.py b/llama_cpp/llama_cpp.py index b250c00..5f9fd18 100644 --- a/llama_cpp/llama_cpp.py +++ b/llama_cpp/llama_cpp.py @@ -1070,7 +1070,7 @@ _lib.llama_kv_cache_view_free.restype = None # // Update the KV cache view structure with the current state of the KV cache. (use only for debugging purposes) # LLAMA_API void llama_kv_cache_view_update(const struct llama_context * ctx, struct llama_kv_cache_view * view); -def llama_kv_cache_view_update(ctx: llama_context_p, view: llama_kv_cache_view_p): +def llama_kv_cache_view_update(ctx: llama_context_p, view: "ctypes._Pointer[llama_kv_cache_view]"): # type: ignore """Update the KV cache view structure with the current state of the KV cache. (use only for debugging purposes)""" return _lib.llama_kv_cache_view_update(ctx, view)