update readme
This commit is contained in:
parent
6bf6c1ef28
commit
6ed34de804
@ -81,7 +81,7 @@ This will automatically configure the API for you as well as the bot in two sepe
|
|||||||
# Docker Compose with GPU
|
# Docker Compose with GPU
|
||||||
This will automatically configure the API that supports cuBLAS and GPU inference for you as well as the bot in two seperate containers within a stack.
|
This will automatically configure the API that supports cuBLAS and GPU inference for you as well as the bot in two seperate containers within a stack.
|
||||||
|
|
||||||
NOTE: Caching is currently broken for cuBLAS: https://github.com/abetlen/llama-cpp-python/issues/253
|
NOTE: Caching for GPU has been fixed.
|
||||||
|
|
||||||
1. `git clone https://git.ssh.surf/snxraven/llama-cpp-python-djs-bot.git` - Clone the repo
|
1. `git clone https://git.ssh.surf/snxraven/llama-cpp-python-djs-bot.git` - Clone the repo
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user