llama.cpp/llama_cpp/server
2023-09-13 21:23:13 -04:00
..
__init__.py llama_cpp server: app is now importable, still runnable as a module 2023-04-29 11:41:25 -07:00
__main__.py Fix tensor_split cli option 2023-09-13 20:00:42 -04:00
app.py Update server params. Added lora_base, lora_path, low_vram, and main_gpu. Removed rms_norm_eps and n_gqa (deprecated in llama.cpp) 2023-09-13 21:23:13 -04:00