diff --git a/README.md b/README.md index a84c937..2e463c1 100644 --- a/README.md +++ b/README.md @@ -81,7 +81,7 @@ This will automatically configure the API for you as well as the bot in two sepe # Docker Compose with GPU This will automatically configure the API that supports cuBLAS and GPU inference for you as well as the bot in two seperate containers within a stack. -NOTE: Caching is currently broken for cuBLAS: https://github.com/abetlen/llama-cpp-python/issues/253 +NOTE: Caching for GPU has been fixed. 1. `git clone https://git.ssh.surf/snxraven/llama-cpp-python-djs-bot.git` - Clone the repo