version: '3.8' services: model-server: build: context: . dockerfile: Dockerfile image: nexasci-agent:latest container_name: nexasci-model-server command: uvicorn agent.model_server:app --host 0.0.0.0 --port 8001 ports: - "8001:8001" volumes: - ./models:/app/models - ./agent:/app/agent - ~/.cache/huggingface:/root/.cache/huggingface environment: - CUDA_VISIBLE_DEVICES=0 - TOKENIZERS_PARALLELISM=false deploy: resources: reservations: devices: - driver: nvidia count: 1 capabilities: [gpu] restart: unless-stopped tool-server: build: context: . dockerfile: Dockerfile image: nexasci-agent:latest container_name: nexasci-tool-server command: uvicorn tools.server:app --host 0.0.0.0 --port 8000 ports: - "8000:8000" volumes: - ./tools:/app/tools - ./pipeline:/app/pipeline - ./index:/app/index - ./tmp:/app/tmp environment: - TOKENIZERS_PARALLELISM=false restart: unless-stopped depends_on: - model-server agent: build: context: . dockerfile: Dockerfile image: nexasci-agent:latest container_name: nexasci-agent command: python examples/demo_agent.py --prompt "Hello, test the agent system." volumes: - ./examples:/app/examples - ./agent:/app/agent environment: - TOKENIZERS_PARALLELISM=false depends_on: - model-server - tool-server stdin_open: true tty: true