Go to file
2023-03-23 15:51:05 -04:00
examples Update basic example 2023-03-23 14:57:31 -04:00
llama_cpp Add support for logprobs 2023-03-23 15:51:05 -04:00
vendor Add llama.cpp to vendor folder 2023-03-23 05:37:26 -04:00
.gitignore Updated package to build with skbuild 2023-03-23 13:54:14 -04:00
.gitmodules Add llama.cpp to vendor folder 2023-03-23 05:37:26 -04:00
CMakeLists.txt Updated package to build with skbuild 2023-03-23 13:54:14 -04:00
LICENSE.md Initial commit 2023-03-23 05:33:06 -04:00
poetry.lock Initial pypi release 2023-03-23 14:24:08 -04:00
pyproject.toml Update repo url 2023-03-23 14:57:18 -04:00
README.md Update pip instructions in readme 2023-03-23 14:24:34 -04:00
setup.py Initial pypi release 2023-03-23 14:24:08 -04:00

llama.cpp Python Bindings

Simple Python bindings for @ggerganov's llama.cpp library.

These bindings expose the low-level llama.cpp C API through a complete ctypes interface. This module also exposes a high-level Python API that is more convenient to use and follows a familiar format.

Install

pip install llama-cpp-python

Usage

>>> from llama_cpp import Llama
>>> llm = Llama(model_path="models/7B/...")
>>> output = llm("Q: Name the planets in the solar system? A: ", max_tokens=32, stop=["Q:", "\n"], echo=True)
>>> print(output)
{
  "id": "cmpl-xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx",
  "object": "text_completion",
  "created": 1679561337,
  "model": "models/7B/...",
  "choices": [
    {
      "text": "Q: Name the planets in the solar system? A: Mercury, Venus, Earth, Mars, Jupiter, Saturn, Uranus, Neptune and Pluto.",
      "index": 0,
      "logprobs": None,
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 14,
    "completion_tokens": 28,
    "total_tokens": 42
  }
}