From c2d63a71481b0963cf1ba0c5be669b9d8e19f91e Mon Sep 17 00:00:00 2001 From: Gardner Bickford Date: Mon, 27 Nov 2023 09:39:18 +1300 Subject: [PATCH 1/2] fix: Typo in the Open Orca chat format #874 (#947) --- llama_cpp/llama_chat_format.py | 14 +++++++------- 1 file changed, 7 insertions(+), 7 deletions(-) diff --git a/llama_cpp/llama_chat_format.py b/llama_cpp/llama_chat_format.py index e927148..24ef9e2 100644 --- a/llama_cpp/llama_chat_format.py +++ b/llama_cpp/llama_chat_format.py @@ -587,14 +587,14 @@ def format_open_orca( system_template = "{system_message}" system_message = ( "You are a helpful assistant. Please answer truthfully and write out your " + "thinking step by step to be sure you get the right answer. If you make a mistake or encounter " + "an error in your thinking, say so out loud and attempt to correct it. If you don't know or " + "aren't sure about something, say so clearly. You will act as a professional logician, mathematician, " + "and physicist. You will also act as the most appropriate type of expert to answer any particular " + "question or solve the relevant problem; state which expert type your are, if so. Also think of " + "any particular named expert that would be ideal to answer the relevant question or solve the " + "relevant problem; name and act as them, if appropriate." ) - "thinking step by step to be sure you get the right answer. If you make a mistake or encounter " - "an error in your thinking, say so out loud and attempt to correct it. If you don't know or " - "aren't sure about something, say so clearly. You will act as a professional logician, mathematician, " - "and physicist. You will also act as the most appropriate type of expert to answer any particular " - "question or solve the relevant problem; state which expert type your are, if so. Also think of " - "any particular named expert that would be ideal to answer the relevant question or solve the " - "relevant problem; name and act as them, if appropriate." roles = ("User", "Assistant") sep = "<|end_of_turn|>\n" # stop_token_ids=[32000, 32001], # "<|end_of_turn|>" From aa5a7a188069ef227f78d125751829b3f82ad8c3 Mon Sep 17 00:00:00 2001 From: Anton Vice <118047001+antonvice@users.noreply.github.com> Date: Mon, 27 Nov 2023 04:39:38 +0800 Subject: [PATCH 2/2] Update README.md (#940) .ccp >> .cpp --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index c5c57e4..60be94b 100644 --- a/README.md +++ b/README.md @@ -45,7 +45,7 @@ Note: If you are using Apple Silicon (M1) Mac, make sure you have installed a ve wget https://github.com/conda-forge/miniforge/releases/latest/download/Miniforge3-MacOSX-arm64.sh bash Miniforge3-MacOSX-arm64.sh ``` -Otherwise, while installing it will build the llama.ccp x86 version which will be 10x slower on Apple Silicon (M1) Mac. +Otherwise, while installing it will build the llama.cpp x86 version which will be 10x slower on Apple Silicon (M1) Mac. ### Installation with Hardware Acceleration