From b206f7b646c736ff70a04c0ac68f7d1a7260158e Mon Sep 17 00:00:00 2001 From: Samuel Walker Date: Fri, 15 Dec 2023 10:43:12 -0500 Subject: [PATCH] Fix 1902-cuda.patch Fix patch against newer versions of llama.cpp Signed-off-by: Samuel Walker --- patches/1902-cuda.patch | 11 +++++++---- 1 file changed, 7 insertions(+), 4 deletions(-) diff --git a/patches/1902-cuda.patch b/patches/1902-cuda.patch index aed2fd4..0c2f4a0 100644 --- a/patches/1902-cuda.patch +++ b/patches/1902-cuda.patch @@ -223,14 +223,13 @@ index 2597ba0..e42ae73 100644 +} \ No newline at end of file diff --git a/common/common.h b/common/common.h -index 18aea38..ca7a168 100644 +index e87ce11..323d6f5 100644 --- a/common/common.h +++ b/common/common.h -@@ -209,3 +209,19 @@ std::string get_sortable_timestamp(); - void dump_non_result_info_yaml( +@@ -231,6 +231,22 @@ void dump_non_result_info_yaml( FILE * stream, const gpt_params & params, const llama_context * lctx, const std::string & timestamp, const std::vector & prompt_tokens, const char * model_desc); -+ + +struct llama_binding_state { + llama_context * ctx; + llama_model * model; @@ -246,3 +245,7 @@ index 18aea38..ca7a168 100644 + const std::vector & last_tokens, + std::vector & candidates, + int idx = 0); ++ + // + // KV cache utils + //