diff --git a/CHANGELOG.md b/CHANGELOG.md index 728fbb2..49b3283 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -7,6 +7,23 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0 ## [Unreleased] +## [0.2.12] + +- Update llama.cpp to ggerganov/llama.cpp@50337961a678fce4081554b24e56e86b67660163 +- Fix missing `n_seq_id` in `llama_batch` by @NickAlgra in #842 +- Fix exception raised in `__del__` when freeing models by @cebtenzzre in #848 +- Performance improvement for logit bias by @zolastro in #851 +- Fix suffix check arbitrary code execution bug by @mtasic85 in #854 +- Fix typo in `function_call` parameter in `llama_types.py` by @akatora28 in #849 +- Fix streaming not returning `finish_reason` by @gmcgoldr in #798 +- Fix `n_gpu_layers` check to allow values less than 1 for server by @hxy9243 in #826 +- Supppress stdout and stderr when freeing model by @paschembri in #803 +- Fix `llama2` chat format by @delock in #808 +- Add validation for tensor_split size by @eric1932 #820 +- Print stack trace on server error by @abetlen in d6a130a052db3a50975a719088a9226abfebb266 +- Update docs for gguf by @johnccshen in #783 +- Add `chatml` chat format by @abetlen in 305482bd4156c70802fc054044119054806f4126 + ## [0.2.11] - Fix bug in `llama_model_params` object has no attribute `logits_all` by @abetlen in d696251fbe40015e8616ea7a7d7ad5257fd1b896 diff --git a/llama_cpp/__init__.py b/llama_cpp/__init__.py index 1c7a27d..4c41396 100644 --- a/llama_cpp/__init__.py +++ b/llama_cpp/__init__.py @@ -1,4 +1,4 @@ from .llama_cpp import * from .llama import * -__version__ = "0.2.11" \ No newline at end of file +__version__ = "0.2.12" \ No newline at end of file