Files
StrikePackageGPT/docker-compose.yml
mblanke 8b51ba9108 v2.2: Network map improvements and OS filtering
- Fixed jumpy network map: nodes settle in 2 seconds and stay fixed
- Added click vs drag detection for better node interaction
- Made legend clickable as OS type filters (Windows, Linux, macOS, etc.)
- Multiple filters can be active simultaneously (OR logic)
- Added 'Clear filters' button when filters are active
- Added DELETE endpoints to clear network hosts from dashboard
- Fixed nmap parser to only include hosts with open ports
- Nodes stay in place after dragging
2025-12-08 10:17:06 -05:00

130 lines
3.7 KiB
YAML

services:
# Web Dashboard - Main user interface
dashboard:
build:
context: ./services/dashboard
dockerfile: Dockerfile
container_name: strikepackage-dashboard
ports:
- "8080:8080"
environment:
- HACKGPT_API_URL=http://strikepackage-hackgpt-api:8001
- LLM_ROUTER_URL=http://strikepackage-llm-router:8000
- KALI_EXECUTOR_URL=http://strikepackage-kali-executor:8002
depends_on:
- hackgpt-api
- llm-router
networks:
- strikepackage-net
restart: unless-stopped
# HackGPT API - Security-focused API service
hackgpt-api:
build:
context: ./services/hackgpt-api
dockerfile: Dockerfile
container_name: strikepackage-hackgpt-api
ports:
- "8001:8001"
environment:
- LLM_ROUTER_URL=http://strikepackage-llm-router:8000
- KALI_EXECUTOR_URL=http://strikepackage-kali-executor:8002
- DEFAULT_LLM_PROVIDER=${DEFAULT_LLM_PROVIDER:-ollama}
- DEFAULT_LLM_MODEL=${DEFAULT_LLM_MODEL:-llama3.2}
depends_on:
- llm-router
- kali-executor
networks:
- strikepackage-net
restart: unless-stopped
# Kali Executor - Command execution service
kali-executor:
build:
context: ./services/kali-executor
dockerfile: Dockerfile
container_name: strikepackage-kali-executor
ports:
- "8002:8002"
environment:
- KALI_CONTAINER_NAME=strikepackage-kali
volumes:
- /var/run/docker.sock:/var/run/docker.sock
depends_on:
- kali
networks:
- strikepackage-net
restart: unless-stopped
# LLM Router - Routes to different LLM providers with load balancing
llm-router:
build:
context: ./services/llm-router
dockerfile: Dockerfile
container_name: strikepackage-llm-router
ports:
- "8000:8000"
environment:
- OPENAI_API_KEY=${OPENAI_API_KEY:-}
- ANTHROPIC_API_KEY=${ANTHROPIC_API_KEY:-}
# Local Ollama on host machine (use host.docker.internal on Windows/Mac)
- OLLAMA_LOCAL_URL=${OLLAMA_LOCAL_URL:-http://host.docker.internal:11434}
# Network Ollama instances (Dell LLM box with larger models)
- OLLAMA_NETWORK_URLS=${OLLAMA_NETWORK_URLS:-http://192.168.1.50:11434}
# Legacy single endpoint (fallback)
- OLLAMA_ENDPOINTS=${OLLAMA_ENDPOINTS:-http://host.docker.internal:11434}
- OLLAMA_BASE_URL=${OLLAMA_BASE_URL:-http://host.docker.internal:11434}
# Load balancing: round-robin, random, failover
- LOAD_BALANCE_STRATEGY=${LOAD_BALANCE_STRATEGY:-failover}
extra_hosts:
- "host.docker.internal:host-gateway"
networks:
- strikepackage-net
restart: unless-stopped
# Kali Linux - Security tools container
kali:
build:
context: ./services/kali
dockerfile: Dockerfile
container_name: strikepackage-kali
stdin_open: true
tty: true
volumes:
- kali-workspace:/workspace
- ./data:/data
networks:
- strikepackage-net
cap_add:
- NET_ADMIN
- NET_RAW
restart: unless-stopped
# Ollama - Local LLM (disabled - using Dell LLM box at 192.168.1.50)
# Uncomment to use local Ollama instead
# ollama:
# image: ollama/ollama:latest
# container_name: strikepackage-ollama
# ports:
# - "11434:11434"
# volumes:
# - ollama-models:/root/.ollama
# networks:
# - strikepackage-net
# restart: unless-stopped
# # Uncomment for GPU support:
# # deploy:
# # resources:
# # reservations:
# # devices:
# # - driver: nvidia
# # count: all
# # capabilities: [gpu]
networks:
strikepackage-net:
driver: bridge
volumes:
kali-workspace:
ollama-models: