Files
StrikePackageGPT/docker-compose.yml
mblanke 5d5a4d4e20 V2.1: Network Map integration, Scan History, OS detection improvements
- Added nmap wrapper to auto-send scan results to Dashboard
- Network Map now displays hosts from terminal scans
- Scan History tab shows all scans (GUI and terminal)
- Load previous scans to Network Map feature
- Improved OS detection from nmap output (parses OS details, smb-os-discovery)
- Added determine_os_type() with OUI/MAC vendor lookup
- Static network map layout (no more jumpy D3 force simulation)
- Fixed docker-compose for Ollama connectivity (host.docker.internal)
- Added test_services.sh for comprehensive testing
2025-12-08 09:07:41 -05:00

129 lines
3.7 KiB
YAML

services:
# Web Dashboard - Main user interface
dashboard:
build:
context: ./services/dashboard
dockerfile: Dockerfile
container_name: strikepackage-dashboard
ports:
- "8080:8080"
environment:
- HACKGPT_API_URL=http://strikepackage-hackgpt-api:8001
- LLM_ROUTER_URL=http://strikepackage-llm-router:8000
- KALI_EXECUTOR_URL=http://strikepackage-kali-executor:8002
depends_on:
- hackgpt-api
- llm-router
networks:
- strikepackage-net
restart: unless-stopped
# HackGPT API - Security-focused API service
hackgpt-api:
build:
context: ./services/hackgpt-api
dockerfile: Dockerfile
container_name: strikepackage-hackgpt-api
ports:
- "8001:8001"
environment:
- LLM_ROUTER_URL=http://strikepackage-llm-router:8000
- KALI_EXECUTOR_URL=http://strikepackage-kali-executor:8002
- DEFAULT_LLM_PROVIDER=${DEFAULT_LLM_PROVIDER:-ollama}
- DEFAULT_LLM_MODEL=${DEFAULT_LLM_MODEL:-llama3.1:latest}
depends_on:
- llm-router
- kali-executor
networks:
- strikepackage-net
restart: unless-stopped
# Kali Executor - Command execution service
kali-executor:
build:
context: ./services/kali-executor
dockerfile: Dockerfile
container_name: strikepackage-kali-executor
ports:
- "8002:8002"
environment:
- KALI_CONTAINER_NAME=strikepackage-kali
volumes:
- /var/run/docker.sock:/var/run/docker.sock
# depends_on:
# - kali # Temporarily disabled due to Kali mirror SSL issues
networks:
- strikepackage-net
restart: unless-stopped
# LLM Router - Routes to different LLM providers with load balancing
llm-router:
build:
context: ./services/llm-router
dockerfile: Dockerfile
container_name: strikepackage-llm-router
ports:
- "8000:8000"
environment:
- OPENAI_API_KEY=${OPENAI_API_KEY:-}
- ANTHROPIC_API_KEY=${ANTHROPIC_API_KEY:-}
# Local Ollama endpoint (use host.docker.internal to reach host machine from container)
- OLLAMA_LOCAL_URL=${OLLAMA_LOCAL_URL:-http://host.docker.internal:11434}
# Network Ollama endpoints (comma-separated for multiple)
- OLLAMA_NETWORK_URLS=${OLLAMA_NETWORK_URLS:-http://192.168.1.50:11434}
# Legacy single endpoint (fallback)
- OLLAMA_BASE_URL=${OLLAMA_BASE_URL:-http://host.docker.internal:11434}
# Load balancing: round-robin, random, failover
- LOAD_BALANCE_STRATEGY=${LOAD_BALANCE_STRATEGY:-round-robin}
extra_hosts:
- "host.docker.internal:host-gateway"
networks:
- strikepackage-net
restart: unless-stopped
# Kali Linux - Security tools container
kali:
build:
context: ./services/kali
dockerfile: Dockerfile
container_name: strikepackage-kali
stdin_open: true
tty: true
volumes:
- kali-workspace:/workspace
- ./data:/data
networks:
- strikepackage-net
cap_add:
- NET_ADMIN
- NET_RAW
restart: unless-stopped
# Ollama - Local LLM (disabled - using Dell LLM box at 192.168.1.50)
# Uncomment to use local Ollama instead
# ollama:
# image: ollama/ollama:latest
# container_name: strikepackage-ollama
# ports:
# - "11434:11434"
# volumes:
# - ollama-models:/root/.ollama
# networks:
# - strikepackage-net
# restart: unless-stopped
# # Uncomment for GPU support:
# # deploy:
# # resources:
# # reservations:
# # devices:
# # - driver: nvidia
# # count: all
# # capabilities: [gpu]
networks:
strikepackage-net:
driver: bridge
volumes:
kali-workspace:
ollama-models: