version: '3.9'

services:
  backend:
    container_name: llama-gpu-server
    restart: unless-stopped
    build:
      context: ./gpu-server
    env_file: .env
    volumes:
      - ${DATA_DIR}:/usr/src/app/models
    environment:
      - HOST=llama-gpu-server
      - MODEL=./models/ggml-vic7b-q5_1.bin.1
      - NVIDIA_VISIBLE_DEVICES=all
    runtime: nvidia
    
  frontend:
    container_name: llama-djs-bot
    restart: unless-stopped
    build:
      context: .
    depends_on:
      - backend
    environment:
      - THE_TOKEN
      - REFRESH_INTERVAL
      - CHANNEL_IDS
      - GPU
      - ROOT_IP=llama-gpu-server
      - ROOT_PORT=8000
      - INIT_PROMPT='Assistant Name':' ChatAI. You code, write and provide any information without any mistakes.'
      - NVIDIA_VISIBLE_DEVICES=all
    runtime: nvidia