Commit graph

1291 commits

Author SHA1 Message Date
MillionthOdin16 1e90597983 Add pydantic dep. Errors if pedantic isn't present. Also throws errors relating to TypeDict or subclass() if the version is too old or new... 2023-04-05 17:37:06 -04:00
Andrei Betlen 267d3648fc Bump version 2023-04-05 16:26:22 -04:00
Andrei Betlen 74bf043ddd Update llama.cpp 2023-04-05 16:25:54 -04:00
Andrei Betlen 44448fb3a8 Add server as a subpackage 2023-04-05 16:23:25 -04:00
Andrei Betlen e1b5b9bb04 Update fastapi server example 2023-04-05 14:44:26 -04:00
Mug 283e59c5e9 Fix bug in init_break not being set when exited via antiprompt and others. 2023-04-05 14:47:24 +02:00
Mug 99ceecfccd Move to new examples directory 2023-04-05 14:28:02 +02:00
Mug e3ea354547 Allow local llama library usage 2023-04-05 14:23:01 +02:00
Mug e4c6f34d95 Merge branch 'main' of https://github.com/abetlen/llama-cpp-python 2023-04-05 14:18:27 +02:00
Andrei Betlen 6de2f24aca Bump version 2023-04-05 06:53:43 -04:00
Andrei Betlen e96a5c5722 Make Llama instance pickleable. Closes #27 2023-04-05 06:52:17 -04:00
Andrei Betlen 152e4695c3 Bump Version 2023-04-05 04:43:51 -04:00
Andrei Betlen c177c807e5 Add supported python versions 2023-04-05 04:43:19 -04:00
Andrei Betlen 17fdd1547c Update workflow name and add badge to README 2023-04-05 04:41:24 -04:00
Andrei Betlen 7643f6677d Bugfix for Python3.7 2023-04-05 04:37:33 -04:00
Andrei Betlen 4d015c33bd Fix syntax error 2023-04-05 04:35:15 -04:00
Andrei Betlen 47570df17b Checkout submodules 2023-04-05 04:34:19 -04:00
Andrei Betlen e3f999e732 Add missing scikit-build install 2023-04-05 04:31:38 -04:00
Andrei Betlen 43c20d3282 Add initial github action to run automated tests 2023-04-05 04:30:32 -04:00
Andrei Betlen b1babcf56c Add quantize example 2023-04-05 04:17:26 -04:00
Andrei Betlen c8e13a78d0 Re-organize examples folder 2023-04-05 04:10:13 -04:00
Andrei Betlen c16bda5fb9 Add performance tuning notebook 2023-04-05 04:09:19 -04:00
Andrei Betlen cefc69ea43 Add runtime check to ensure embedding is enabled if trying to generate embeddings 2023-04-05 03:25:37 -04:00
Andrei Betlen 5c50af7462 Remove workaround 2023-04-05 03:25:09 -04:00
Andrei Betlen c3972b61ae Add basic tests. Closes #24 2023-04-05 03:23:15 -04:00
Andrei Betlen 51dbcf2693 Bugfix: wrong signature for quantize function 2023-04-04 22:36:59 -04:00
Andrei Betlen 8279fb7d92 Bump version 2023-04-04 17:17:11 -04:00
Andrei Betlen c137789143 Add verbose flag. Closes #19 2023-04-04 13:09:24 -04:00
Andrei Betlen 5075c16fcc Bugfix: n_batch should always be <= n_ctx 2023-04-04 13:08:21 -04:00
Mug c862e8bac5 Fix repeating instructions and an antiprompt bug 2023-04-04 17:54:47 +02:00
Andrei Betlen 248b0566fa Update README 2023-04-04 10:57:22 -04:00
Mug 9cde7973cc Fix stripping instruction prompt 2023-04-04 16:20:27 +02:00
Mug da5a6a7089 Added instruction mode, fixed infinite generation, and various other fixes 2023-04-04 16:18:26 +02:00
Mug 0b32bb3d43 Add instruction mode 2023-04-04 11:48:48 +02:00
Andrei Betlen ffe34cf64d Allow user to set llama config from env vars 2023-04-04 00:52:44 -04:00
Andrei Betlen 05eb2087d8 Small fixes for examples 2023-04-03 20:33:07 -04:00
Andrei Betlen caf3c0362b Add return type for default __call__ method 2023-04-03 20:26:08 -04:00
Andrei Betlen 4aa349d777 Add docstring for create_chat_completion 2023-04-03 20:24:20 -04:00
Andrei Betlen 4615f1e520 Add chat completion method to docs 2023-04-03 20:14:03 -04:00
Andrei Betlen 5cf29d0231 Bump version 2023-04-03 20:13:46 -04:00
Andrei Betlen 7fedf16531 Add support for chat completion 2023-04-03 20:12:44 -04:00
Andrei Betlen 3dec778c90 Update to more sensible return signature 2023-04-03 20:12:14 -04:00
Andrei Betlen f7ab8d55b2 Update context size defaults Close #11 2023-04-03 20:11:13 -04:00
Andrei Betlen c0a5c0171f Add embed back into documentation 2023-04-03 18:53:00 -04:00
Andrei Betlen adf656d542 Bump version 2023-04-03 18:46:49 -04:00
Andrei Betlen ae004eb69e Fix #16 2023-04-03 18:46:19 -04:00
Mug f1615f05e6 Chat llama.cpp example implementation 2023-04-03 22:54:46 +02:00
Andrei Betlen 7d1977e8f0 Bump version 2023-04-03 14:49:36 -04:00
Andrei Betlen 4530197629 Update llama.cpp 2023-04-03 14:49:07 -04:00
Andrei 1d9a988644
Merge pull request #10 from MillionthOdin16/patch-1
Improve Shared Library Loading Mechanism
2023-04-03 14:47:11 -04:00