* fix: add build-arg support for proxies and mirrors in Docker builds (#1260) Pin Debian images to bookworm, make UV source image configurable, and pass APT_MIRROR/NPM_REGISTRY/UV_IMAGE through docker-compose. * fix: ensure build args use consistent defaults across compose and Dockerfiles UV_IMAGE: ${UV_IMAGE:-} resolved to empty when unset, overriding the Dockerfile ARG default and breaking `FROM ${UV_IMAGE}`. Also configure COREPACK_NPM_REGISTRY before pnpm download and propagate NPM_REGISTRY into the prod stage. * fix: dearmor NodeSource GPG key to resolve signing error Pipe the downloaded key through gpg --dearmor so apt can verify the repository signature (fixes NO_PUBKEY 2F59B5F99B1BE0B4). --------- Co-authored-by: JeffJiang <for-eleven@hotmail.com>
DeerFlow Sandbox Provisioner
The Sandbox Provisioner is a FastAPI service that dynamically manages sandbox Pods in Kubernetes. It provides a REST API for the DeerFlow backend to create, monitor, and destroy isolated sandbox environments for code execution.
Architecture
┌────────────┐ HTTP ┌─────────────┐ K8s API ┌──────────────┐
│ Backend │ ─────▸ │ Provisioner │ ────────▸ │ Host K8s │
│ (gateway/ │ │ :8002 │ │ API Server │
│ langgraph) │ └─────────────┘ └──────┬───────┘
└────────────┘ │ creates
│
┌─────────────┐ ┌────▼─────┐
│ Backend │ ──────▸ │ Sandbox │
│ (via Docker │ NodePort│ Pod(s) │
│ network) │ └──────────┘
└─────────────┘
How It Works
-
Backend Request: When the backend needs to execute code, it sends a
POST /api/sandboxesrequest with asandbox_idandthread_id. -
Pod Creation: The provisioner creates a dedicated Pod in the
deer-flownamespace with:- The sandbox container image (all-in-one-sandbox)
- HostPath volumes mounted for:
/mnt/skills→ Read-only access to public skills/mnt/user-data→ Read-write access to thread-specific data
- Resource limits (CPU, memory, ephemeral storage)
- Readiness/liveness probes
-
Service Creation: A NodePort Service is created to expose the Pod, with Kubernetes auto-allocating a port from the NodePort range (typically 30000-32767).
-
Access URL: The provisioner returns
http://host.docker.internal:{NodePort}to the backend, which the backend containers can reach directly. -
Cleanup: When the session ends,
DELETE /api/sandboxes/{sandbox_id}removes both the Pod and Service.
Requirements
Host machine with a running Kubernetes cluster (Docker Desktop K8s, OrbStack, minikube, kind, etc.)
Enable Kubernetes in Docker Desktop
- Open Docker Desktop settings
- Go to "Kubernetes" tab
- Check "Enable Kubernetes"
- Click "Apply & Restart"
Enable Kubernetes in OrbStack
- Open OrbStack settings
- Go to "Kubernetes" tab
- Check "Enable Kubernetes"
API Endpoints
GET /health
Health check endpoint.
Response:
{
"status": "ok"
}
POST /api/sandboxes
Create a new sandbox Pod + Service.
Request:
{
"sandbox_id": "abc-123",
"thread_id": "thread-456"
}
Response:
{
"sandbox_id": "abc-123",
"sandbox_url": "http://host.docker.internal:32123",
"status": "Pending"
}
Idempotent: Calling with the same sandbox_id returns the existing sandbox info.
GET /api/sandboxes/{sandbox_id}
Get status and URL of a specific sandbox.
Response:
{
"sandbox_id": "abc-123",
"sandbox_url": "http://host.docker.internal:32123",
"status": "Running"
}
Status Values: Pending, Running, Succeeded, Failed, Unknown, NotFound
DELETE /api/sandboxes/{sandbox_id}
Destroy a sandbox Pod + Service.
Response:
{
"ok": true,
"sandbox_id": "abc-123"
}
GET /api/sandboxes
List all sandboxes currently managed.
Response:
{
"sandboxes": [
{
"sandbox_id": "abc-123",
"sandbox_url": "http://host.docker.internal:32123",
"status": "Running"
}
],
"count": 1
}
Configuration
The provisioner is configured via environment variables (set in docker-compose-dev.yaml):
| Variable | Default | Description |
|---|---|---|
K8S_NAMESPACE |
deer-flow |
Kubernetes namespace for sandbox resources |
SANDBOX_IMAGE |
enterprise-public-cn-beijing.cr.volces.com/vefaas-public/all-in-one-sandbox:latest |
Container image for sandbox Pods |
SKILLS_HOST_PATH |
- | Host machine path to skills directory (must be absolute) |
THREADS_HOST_PATH |
- | Host machine path to threads data directory (must be absolute) |
KUBECONFIG_PATH |
/root/.kube/config |
Path to kubeconfig inside the provisioner container |
NODE_HOST |
host.docker.internal |
Hostname that backend containers use to reach host NodePorts |
K8S_API_SERVER |
(from kubeconfig) | Override K8s API server URL (e.g., https://host.docker.internal:26443) |
Important: K8S_API_SERVER Override
If your kubeconfig uses localhost, 127.0.0.1, or 0.0.0.0 as the API server address (common with OrbStack, minikube, kind), the provisioner cannot reach it from inside the Docker container.
Solution: Set K8S_API_SERVER to use host.docker.internal:
# docker-compose-dev.yaml
provisioner:
environment:
- K8S_API_SERVER=https://host.docker.internal:26443 # Replace 26443 with your API port
Check your kubeconfig API server:
kubectl config view --minify -o jsonpath='{.clusters[0].cluster.server}'
Prerequisites
Host Machine Requirements
-
Kubernetes Cluster:
- Docker Desktop with Kubernetes enabled, or
- OrbStack (built-in K8s), or
- minikube, kind, k3s, etc.
-
kubectl Configured:
~/.kube/configmust exist and be valid- Current context should point to your local cluster
-
Kubernetes Access:
- The provisioner needs permissions to:
- Create/read/delete Pods in the
deer-flownamespace - Create/read/delete Services in the
deer-flownamespace - Read Namespaces (to create
deer-flowif missing)
- Create/read/delete Pods in the
- The provisioner needs permissions to:
-
Host Paths:
- The
SKILLS_HOST_PATHandTHREADS_HOST_PATHmust be absolute paths on the host machine - These paths are mounted into sandbox Pods via K8s HostPath volumes
- The paths must exist and be readable by the K8s node
- The
Docker Compose Setup
The provisioner runs as part of the docker-compose-dev stack:
# Start Docker services (provisioner starts only when config.yaml enables provisioner mode)
make docker-start
# Or start just the provisioner
docker compose -p deer-flow-dev -f docker/docker-compose-dev.yaml up -d provisioner
The compose file:
- Mounts your host's
~/.kube/configinto the container - Adds
extra_hostsentry forhost.docker.internal(required on Linux) - Configures environment variables for K8s access
Testing
Manual API Testing
# Health check
curl http://localhost:8002/health
# Create a sandbox (via provisioner container for internal DNS)
docker exec deer-flow-provisioner curl -X POST http://localhost:8002/api/sandboxes \
-H "Content-Type: application/json" \
-d '{"sandbox_id":"test-001","thread_id":"thread-001"}'
# Check sandbox status
docker exec deer-flow-provisioner curl http://localhost:8002/api/sandboxes/test-001
# List all sandboxes
docker exec deer-flow-provisioner curl http://localhost:8002/api/sandboxes
# Verify Pod and Service in K8s
kubectl get pod,svc -n deer-flow -l sandbox-id=test-001
# Delete sandbox
docker exec deer-flow-provisioner curl -X DELETE http://localhost:8002/api/sandboxes/test-001
Verify from Backend Containers
Once a sandbox is created, the backend containers (gateway, langgraph) can access it:
# Get sandbox URL from provisioner
SANDBOX_URL=$(docker exec deer-flow-provisioner curl -s http://localhost:8002/api/sandboxes/test-001 | jq -r .sandbox_url)
# Test from gateway container
docker exec deer-flow-gateway curl -s $SANDBOX_URL/v1/sandbox
Troubleshooting
Issue: "Kubeconfig not found"
Cause: The kubeconfig file doesn't exist at the mounted path.
Solution:
- Ensure
~/.kube/configexists on your host machine - Run
kubectl config viewto verify - Check the volume mount in docker-compose-dev.yaml
Issue: "Kubeconfig path is a directory"
Cause: The mounted KUBECONFIG_PATH points to a directory instead of a file.
Solution:
- Ensure the compose mount source is a file (e.g.,
~/.kube/config) not a directory - Verify inside container:
docker exec deer-flow-provisioner ls -ld /root/.kube/config - Expected output should indicate a regular file (
-), not a directory (d)
Issue: "Connection refused" to K8s API
Cause: The provisioner can't reach the K8s API server.
Solution:
- Check your kubeconfig server address:
kubectl config view --minify -o jsonpath='{.clusters[0].cluster.server}' - If it's
localhostor127.0.0.1, setK8S_API_SERVER:environment: - K8S_API_SERVER=https://host.docker.internal:PORT
Issue: "Unprocessable Entity" when creating Pod
Cause: HostPath volumes contain invalid paths (e.g., relative paths with ..).
Solution:
- Use absolute paths for
SKILLS_HOST_PATHandTHREADS_HOST_PATH - Verify the paths exist on your host machine:
ls -la /path/to/skills ls -la /path/to/backend/.deer-flow/threads
Issue: Pod stuck in "ContainerCreating"
Cause: Usually pulling the sandbox image from the registry.
Solution:
- Pre-pull the image:
make docker-init - Check Pod events:
kubectl describe pod sandbox-XXX -n deer-flow - Check node:
kubectl get nodes
Issue: Cannot access sandbox URL from backend
Cause: NodePort not reachable or NODE_HOST misconfigured.
Solution:
- Verify the Service exists:
kubectl get svc -n deer-flow - Test from host:
curl http://localhost:NODE_PORT/v1/sandbox - Ensure
extra_hostsis set in docker-compose (Linux) - Check
NODE_HOSTenv var matches how backend reaches host
Security Considerations
-
HostPath Volumes: The provisioner mounts host directories into sandbox Pods. Ensure these paths contain only trusted data.
-
Resource Limits: Each sandbox Pod has CPU, memory, and storage limits to prevent resource exhaustion.
-
Network Isolation: Sandbox Pods run in the
deer-flownamespace but share the host's network namespace via NodePort. Consider NetworkPolicies for stricter isolation. -
kubeconfig Access: The provisioner has full access to your Kubernetes cluster via the mounted kubeconfig. Run it only in trusted environments.
-
Image Trust: The sandbox image should come from a trusted registry. Review and audit the image contents.
Future Enhancements
- Support for custom resource requests/limits per sandbox
- PersistentVolume support for larger data requirements
- Automatic cleanup of stale sandboxes (timeout-based)
- Metrics and monitoring (Prometheus integration)
- Multi-cluster support (route to different K8s clusters)
- Pod affinity/anti-affinity rules for better placement
- NetworkPolicy templates for sandbox isolation