35 lines
876 B
YAML
35 lines
876 B
YAML
|
version: '3.9'
|
||
|
|
||
|
services:
|
||
|
llama-python-server:
|
||
|
container_name: llama-python-gpu-server
|
||
|
restart: unless-stopped
|
||
|
build:
|
||
|
context: ./gpu-server
|
||
|
env_file: .env
|
||
|
volumes:
|
||
|
- ${DATA_DIR}:/usr/src/app/models
|
||
|
environment:
|
||
|
- HOST=llama-python-gpu-server
|
||
|
- MODEL=./models/ggml-vic7b-q5_1.bin.1
|
||
|
- NVIDIA_VISIBLE_DEVICES=all
|
||
|
runtime: nvidia
|
||
|
|
||
|
llama-python-djs-bot:
|
||
|
container_name: llama-python-djs-bot
|
||
|
restart: unless-stopped
|
||
|
build:
|
||
|
context: .
|
||
|
depends_on:
|
||
|
- llama-python-server
|
||
|
environment:
|
||
|
- THE_TOKEN
|
||
|
- REFRESH_INTERVAL
|
||
|
- CHANNEL_IDS
|
||
|
- GPU
|
||
|
- ROOT_IP=llama-python-server
|
||
|
- ROOT_PORT=8000
|
||
|
- INIT_PROMPT='Assistant Name':' ChatAI. You code, write and provide any information without any mistakes.'
|
||
|
- NVIDIA_VISIBLE_DEVICES=all
|
||
|
runtime: nvidia
|