From 6ed34de804e810903ca57a531f9fe2a580e7c7bc Mon Sep 17 00:00:00 2001 From: Raven Scott Date: Sat, 27 May 2023 02:02:38 +0200 Subject: [PATCH] update readme --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index a84c937..2e463c1 100644 --- a/README.md +++ b/README.md @@ -81,7 +81,7 @@ This will automatically configure the API for you as well as the bot in two sepe # Docker Compose with GPU This will automatically configure the API that supports cuBLAS and GPU inference for you as well as the bot in two seperate containers within a stack. -NOTE: Caching is currently broken for cuBLAS: https://github.com/abetlen/llama-cpp-python/issues/253 +NOTE: Caching for GPU has been fixed. 1. `git clone https://git.ssh.surf/snxraven/llama-cpp-python-djs-bot.git` - Clone the repo