Files
dockhand/lib/server/docker.ts
Jarek Krochmalski 62e3c6439e Initial commit
2025-12-28 21:16:03 +01:00

3237 lines
93 KiB
TypeScript

/**
* Docker Operations Module
*
* Uses direct Docker API calls over Unix socket or HTTP/HTTPS.
* No external dependencies like dockerode - uses native Bun fetch.
*/
import { homedir } from 'node:os';
import { existsSync, mkdirSync, rmSync, readdirSync } from 'node:fs';
import { join, resolve } from 'node:path';
import type { Environment } from './db';
import { getStackEnvVarsAsRecord } from './db';
/**
* Custom error for when an environment is not found.
* API endpoints should catch this and return 404.
*/
export class EnvironmentNotFoundError extends Error {
public readonly envId: number;
constructor(envId: number) {
super(`Environment ${envId} not found`);
this.name = 'EnvironmentNotFoundError';
this.envId = envId;
}
}
/**
* Custom error for Docker connection failures with user-friendly messages.
* Wraps raw Bun fetch errors to hide technical details from users.
*/
export class DockerConnectionError extends Error {
public readonly originalError: unknown;
constructor(message: string, originalError: unknown) {
super(message);
this.name = 'DockerConnectionError';
this.originalError = originalError;
}
/**
* Create a DockerConnectionError from any error, sanitizing technical messages
*/
static fromError(error: unknown, context?: string): DockerConnectionError {
const errorStr = String(error);
let friendlyMessage: string;
if (errorStr.includes('FailedToOpenSocket') || errorStr.includes('ECONNREFUSED')) {
friendlyMessage = 'Docker socket not accessible';
} else if (errorStr.includes('ECONNRESET') || errorStr.includes('connection was closed')) {
friendlyMessage = 'Connection lost';
} else if (errorStr.includes('verbose') || errorStr.includes('typo')) {
friendlyMessage = 'Connection failed';
} else if (errorStr.includes('timeout') || errorStr.includes('Timeout') || errorStr.includes('ETIMEDOUT')) {
friendlyMessage = 'Connection timeout';
} else if (errorStr.includes('ENOTFOUND') || errorStr.includes('getaddrinfo')) {
friendlyMessage = 'Host not found';
} else if (errorStr.includes('EHOSTUNREACH')) {
friendlyMessage = 'Host unreachable';
} else {
friendlyMessage = 'Connection error';
}
if (context) {
friendlyMessage = `${context}: ${friendlyMessage}`;
}
return new DockerConnectionError(friendlyMessage, error);
}
}
/**
* Container inspect result from Docker API
*/
export interface ContainerInspectResult {
Id: string;
Name: string;
RestartCount: number;
State: {
Status: string;
Running: boolean;
Paused: boolean;
Restarting: boolean;
OOMKilled: boolean;
Dead: boolean;
Pid: number;
ExitCode: number;
Error: string;
StartedAt: string;
FinishedAt: string;
Health?: {
Status: string;
FailingStreak: number;
Log: Array<{
Start: string;
End: string;
ExitCode: number;
Output: string;
}>;
};
};
Config: {
Hostname: string;
User: string;
Tty: boolean;
Env: string[];
Cmd: string[];
Image: string;
Labels: Record<string, string>;
WorkingDir: string;
Entrypoint: string[] | null;
};
NetworkSettings: {
Networks: Record<string, {
IPAddress: string;
Gateway: string;
MacAddress: string;
}>;
Ports: Record<string, Array<{ HostIp: string; HostPort: string }> | null>;
};
Mounts: Array<{
Type: string;
Source: string;
Destination: string;
Mode: string;
RW: boolean;
}>;
HostConfig: {
Binds: string[] | null;
NetworkMode: string;
PortBindings: Record<string, Array<{ HostIp: string; HostPort: string }>> | null;
RestartPolicy: {
Name: string;
MaximumRetryCount: number;
};
Privileged: boolean;
Memory: number;
MemorySwap: number;
NanoCpus: number;
CpuShares: number;
};
}
// Detect Docker socket path for local connections
function detectDockerSocket(): string {
// Check environment variable first
if (process.env.DOCKER_SOCKET && existsSync(process.env.DOCKER_SOCKET)) {
console.log(`Using Docker socket from DOCKER_SOCKET env: ${process.env.DOCKER_SOCKET}`);
return process.env.DOCKER_SOCKET;
}
// Check DOCKER_HOST environment variable
if (process.env.DOCKER_HOST) {
const dockerHost = process.env.DOCKER_HOST;
if (dockerHost.startsWith('unix://')) {
const socketPath = dockerHost.replace('unix://', '');
if (existsSync(socketPath)) {
console.log(`Using Docker socket from DOCKER_HOST: ${socketPath}`);
return socketPath;
}
}
}
// List of possible socket locations in order of preference
const possibleSockets = [
'/var/run/docker.sock', // Standard Linux/Docker Desktop
`${homedir()}/.docker/run/docker.sock`, // Docker Desktop for Mac (new location)
`${homedir()}/.orbstack/run/docker.sock`, // OrbStack
'/run/docker.sock', // Alternative Linux location
];
for (const socket of possibleSockets) {
if (existsSync(socket)) {
console.log(`Detected Docker socket at: ${socket}`);
return socket;
}
}
// Fallback to default
console.warn('No Docker socket found, using default /var/run/docker.sock');
return '/var/run/docker.sock';
}
const socketPath = detectDockerSocket();
/**
* Demultiplex Docker stream output (strip 8-byte headers)
* Docker streams have: 1 byte type, 3 bytes padding, 4 bytes size BE, then payload
*/
function demuxDockerStream(buffer: Buffer, options?: { separateStreams?: boolean }): string | { stdout: string; stderr: string } {
const stdout: string[] = [];
const stderr: string[] = [];
let offset = 0;
while (offset < buffer.length) {
if (offset + 8 > buffer.length) break;
const streamType = buffer.readUInt8(offset);
const frameSize = buffer.readUInt32BE(offset + 4);
if (frameSize === 0 || frameSize > buffer.length - offset - 8) {
// Invalid frame, return raw content with control chars stripped
const raw = buffer.toString('utf-8').replace(/[\x00-\x08\x0B\x0C\x0E-\x1F]/g, '');
return options?.separateStreams ? { stdout: raw, stderr: '' } : raw;
}
const payload = buffer.slice(offset + 8, offset + 8 + frameSize).toString('utf-8');
if (streamType === 1) {
stdout.push(payload);
} else if (streamType === 2) {
stderr.push(payload);
} else {
stdout.push(payload); // Default to stdout for unknown types
}
offset += 8 + frameSize;
}
if (options?.separateStreams) {
return { stdout: stdout.join(''), stderr: stderr.join('') };
}
return [...stdout, ...stderr].join('');
}
/**
* Process Docker stream frames incrementally from a buffer
* Returns processed frames and remaining buffer
*/
function processStreamFrames(
buffer: Buffer,
onStdout?: (data: string) => void,
onStderr?: (data: string) => void
): { stdout: string; remaining: Buffer<ArrayBufferLike> } {
let stdout = '';
let offset = 0;
while (buffer.length >= offset + 8) {
const streamType = buffer.readUInt8(offset);
const frameSize = buffer.readUInt32BE(offset + 4);
if (buffer.length < offset + 8 + frameSize) break;
const payload = buffer.slice(offset + 8, offset + 8 + frameSize).toString('utf-8');
if (streamType === 1) {
stdout += payload;
onStdout?.(payload);
} else if (streamType === 2) {
onStderr?.(payload);
}
offset += 8 + frameSize;
}
return { stdout, remaining: buffer.slice(offset) };
}
// Cache for environment configurations with timestamps
interface CachedEnv {
env: Environment;
lastUsed: number;
}
const envCache = new Map<number, CachedEnv>();
// Cache TTL: 30 minutes (in milliseconds)
const CACHE_TTL = 30 * 60 * 1000;
// Cleanup stale cache entries periodically
function cleanupEnvCache() {
const now = Date.now();
const entries = Array.from(envCache.entries());
for (const [envId, cached] of entries) {
if (now - cached.lastUsed > CACHE_TTL) {
envCache.delete(envId);
}
}
}
// Guard against multiple intervals during HMR
declare global {
var __dockerEnvCacheCleanupInterval: ReturnType<typeof setInterval> | undefined;
}
// Run cleanup every 10 minutes (guarded to prevent HMR leaks)
if (!globalThis.__dockerEnvCacheCleanupInterval) {
globalThis.__dockerEnvCacheCleanupInterval = setInterval(cleanupEnvCache, 10 * 60 * 1000);
}
// Import db functions for environment lookup
import { getEnvironment } from './db';
// Import hawser edge connection manager for edge mode routing
import { sendEdgeRequest, sendEdgeStreamRequest, isEdgeConnected, type EdgeResponse } from './hawser';
/**
* Docker API client configuration
*/
interface DockerClientConfig {
type: 'socket' | 'http' | 'https';
socketPath?: string;
host?: string;
port?: number;
ca?: string;
cert?: string;
key?: string;
skipVerify?: boolean;
// Hawser connection settings
connectionType?: 'socket' | 'direct' | 'hawser-standard' | 'hawser-edge';
hawserToken?: string;
// Environment ID for edge mode routing
environmentId?: number;
}
/**
* Build Docker client config from an environment
*/
function buildConfigFromEnv(env: Environment): DockerClientConfig {
// Socket connection type - use Unix socket
if (env.connectionType === 'socket' || !env.connectionType) {
return {
type: 'socket',
socketPath: env.socketPath || '/var/run/docker.sock',
connectionType: 'socket',
environmentId: env.id
};
}
// Direct or Hawser connection types - use HTTP/HTTPS
const protocol = (env.protocol as 'http' | 'https') || 'http';
return {
type: protocol,
host: env.host || 'localhost',
port: env.port || 2375,
ca: env.tlsCa || undefined,
cert: env.tlsCert || undefined,
key: env.tlsKey || undefined,
skipVerify: env.tlsSkipVerify || undefined,
connectionType: env.connectionType as 'direct' | 'hawser-standard' | 'hawser-edge',
hawserToken: env.hawserToken || undefined,
environmentId: env.id
};
}
/**
* Get Docker client configuration for an environment
*/
async function getDockerConfig(envId?: number | null): Promise<DockerClientConfig> {
if (!envId) {
throw new Error('No environment specified');
}
// Check cache first
const cached = envCache.get(envId);
if (cached) {
cached.lastUsed = Date.now();
return buildConfigFromEnv(cached.env);
}
// Fetch and cache
const env = await getEnvironment(envId);
if (env) {
envCache.set(envId, { env, lastUsed: Date.now() });
return buildConfigFromEnv(env);
}
throw new EnvironmentNotFoundError(envId);
}
interface DockerFetchOptions extends RequestInit {
/** Set to true for long-lived streaming connections (disables Bun's idle timeout) */
streaming?: boolean;
}
/**
* Check if a string is valid base64
*/
function isBase64(str: string): boolean {
if (!str || str.length === 0) return false;
// Base64 strings have length divisible by 4 and contain only valid chars
if (str.length % 4 !== 0) return false;
return /^[A-Za-z0-9+/]*={0,2}$/.test(str);
}
/**
* Convert EdgeResponse from hawser WebSocket to a standard Response object
* Handles base64-encoded binary data from Go agent
*/
function edgeResponseToResponse(edgeResponse: EdgeResponse): Response {
let body: string | Uint8Array = edgeResponse.body;
// The Go agent sends isBinary flag to indicate if body is base64-encoded
if (edgeResponse.isBinary && typeof body === 'string' && body.length > 0) {
// Decode base64 to binary
body = Uint8Array.from(atob(body), c => c.charCodeAt(0));
}
return new Response(body as BodyInit, {
status: edgeResponse.statusCode,
headers: edgeResponse.headers
});
}
/**
* Make a request to the Docker API
* Exported for use by stacks.ts module
*/
export async function dockerFetch(
path: string,
options: DockerFetchOptions = {},
envId?: number | null
): Promise<Response> {
const startTime = Date.now();
const config = await getDockerConfig(envId);
const { streaming, ...fetchOptions } = options;
const method = (options.method || 'GET').toUpperCase();
// For streaming connections, disable Bun's idle timeout
// This prevents long-lived streams (like Docker events) from being terminated
const bunOptions = streaming ? { timeout: false } : {};
// Hawser Edge mode - route through WebSocket connection
if (config.connectionType === 'hawser-edge' && config.environmentId) {
// Check if agent is connected
if (!isEdgeConnected(config.environmentId)) {
const error = new Error('Hawser Edge agent is not connected');
// Log without stack trace for cleaner output
console.warn(`[Docker] Edge env ${config.environmentId}: agent not connected for ${method} ${path}`);
throw error;
}
// Extract request details
const headers: Record<string, string> = {};
// Convert Headers object to plain object
if (fetchOptions.headers) {
if (fetchOptions.headers instanceof Headers) {
fetchOptions.headers.forEach((value, key) => {
headers[key] = value;
});
} else if (typeof fetchOptions.headers === 'object') {
Object.assign(headers, fetchOptions.headers);
}
}
// Parse body if present
let body: unknown;
if (fetchOptions.body) {
if (typeof fetchOptions.body === 'string') {
try {
body = JSON.parse(fetchOptions.body);
} catch {
body = fetchOptions.body;
}
} else {
body = fetchOptions.body;
}
}
// Send request through edge connection
try {
const edgeResponse = await sendEdgeRequest(
config.environmentId,
method,
path,
body,
headers,
streaming || false,
streaming ? 300000 : 30000 // 5 min for streaming, 30s for normal requests
);
const elapsed = Date.now() - startTime;
if (elapsed > 5000) {
console.warn(`[Docker] Edge env ${config.environmentId}: ${method} ${path} took ${elapsed}ms`);
}
return edgeResponseToResponse(edgeResponse);
} catch (error) {
const elapsed = Date.now() - startTime;
console.error(`[Docker] Edge env ${config.environmentId}: ${method} ${path} failed after ${elapsed}ms:`, error);
throw DockerConnectionError.fromError(error);
}
}
if (config.type === 'socket') {
// Use Bun's native Unix socket support
const url = `http://localhost${path}`;
try {
const response = await fetch(url, {
...fetchOptions,
// @ts-ignore - Bun supports unix socket and timeout options
unix: config.socketPath,
...bunOptions
});
const elapsed = Date.now() - startTime;
if (elapsed > 5000) {
console.warn(`[Docker] Socket: ${method} ${path} took ${elapsed}ms`);
}
return response;
} catch (error) {
const elapsed = Date.now() - startTime;
console.error(`[Docker] Socket: ${method} ${path} failed after ${elapsed}ms:`, error);
throw DockerConnectionError.fromError(error);
}
} else {
// HTTP/HTTPS remote connection
const protocol = config.type;
const url = `${protocol}://${config.host}:${config.port}${path}`;
const finalOptions: RequestInit = { ...fetchOptions };
// For Hawser Standard mode with token authentication
if (config.connectionType === 'hawser-standard' && config.hawserToken) {
finalOptions.headers = {
...finalOptions.headers,
'X-Hawser-Token': config.hawserToken
};
}
// For HTTPS with TLS certificates, we need to configure TLS
// IMPORTANT: Bun requires certificates as Buffer objects, not strings
if (config.type === 'https') {
const tlsOptions: Record<string, unknown> = {};
// CA certificate - must be array of Buffers for Bun
if (config.ca) {
tlsOptions.ca = [Buffer.from(config.ca)];
}
// Client certificate and key for mTLS - must be Buffers
if (config.cert) {
tlsOptions.cert = Buffer.from(config.cert);
}
if (config.key) {
tlsOptions.key = Buffer.from(config.key);
}
// Skip verification (self-signed without CA)
if (config.skipVerify) {
tlsOptions.rejectUnauthorized = false;
} else {
tlsOptions.rejectUnauthorized = true;
}
if (Object.keys(tlsOptions).length > 0) {
// @ts-ignore - Bun supports tls options with Buffer certs
finalOptions.tls = tlsOptions;
}
}
// @ts-ignore - Bun supports timeout option
try {
const response = await fetch(url, { ...finalOptions, ...bunOptions });
const elapsed = Date.now() - startTime;
if (elapsed > 5000) {
console.warn(`[Docker] ${config.connectionType || 'direct'} ${config.host}: ${method} ${path} took ${elapsed}ms`);
}
return response;
} catch (error) {
const elapsed = Date.now() - startTime;
console.error(`[Docker] ${config.connectionType || 'direct'} ${config.host}: ${method} ${path} failed after ${elapsed}ms:`, error);
throw DockerConnectionError.fromError(error);
}
}
}
/**
* Make a JSON request to Docker API
*/
async function dockerJsonRequest<T>(
path: string,
options: RequestInit = {},
envId?: number | null
): Promise<T> {
const response = await dockerFetch(path, {
...options,
headers: {
'Content-Type': 'application/json',
...options.headers
}
}, envId);
if (!response.ok) {
const errorText = await response.text();
let errorJson: any = {};
try {
errorJson = JSON.parse(errorText);
} catch {
// Not JSON, use text as message
errorJson = { message: errorText };
}
const error: any = new Error(errorJson.message || `Docker API error: ${response.status}`);
error.statusCode = response.status;
error.json = errorJson;
throw error;
}
return response.json();
}
// Clear cached client for an environment (e.g., when settings change)
export function clearDockerClientCache(envId?: number) {
if (envId !== undefined) {
envCache.delete(envId);
} else {
envCache.clear();
}
}
export interface ContainerInfo {
id: string;
name: string;
image: string;
state: string;
status: string;
created: number;
ports: Array<{
IP?: string;
PrivatePort: number;
PublicPort?: number;
Type: string;
}>;
networks: { [networkName: string]: { ipAddress: string } };
health?: string;
restartCount: number;
mounts: Array<{ type: string; source: string; destination: string; mode: string; rw: boolean }>;
labels: { [key: string]: string };
command: string;
}
export interface ImageInfo {
id: string;
tags: string[];
size: number;
created: number;
}
// Container operations
export async function listContainers(all = true, envId?: number | null): Promise<ContainerInfo[]> {
const containers = await dockerJsonRequest<any[]>(
`/containers/json?all=${all}`,
{},
envId
);
// Fetch restart counts only for restarting containers
const restartCounts = new Map<string, number>();
const restartingContainers = containers.filter(c => c.State === 'restarting');
await Promise.all(
restartingContainers.map(async (container) => {
try {
const inspect = await inspectContainer(container.Id, envId);
restartCounts.set(container.Id, inspect.RestartCount || 0);
} catch {
// Ignore errors
}
})
);
return containers.map((container) => {
// Extract network info with IP addresses
const networks: { [networkName: string]: { ipAddress: string } } = {};
if (container.NetworkSettings?.Networks) {
for (const [networkName, networkData] of Object.entries(container.NetworkSettings.Networks)) {
networks[networkName] = {
ipAddress: (networkData as any).IPAddress || ''
};
}
}
// Extract mount info
const mounts = (container.Mounts || []).map((m: any) => ({
type: m.Type || 'unknown',
source: m.Source || m.Name || '',
destination: m.Destination || '',
mode: m.Mode || '',
rw: m.RW ?? true
}));
// Extract health status from Status string
let health: string | undefined;
const healthMatch = container.Status?.match(/\((healthy|unhealthy|starting)\)/i);
if (healthMatch) {
health = healthMatch[1].toLowerCase();
}
return {
id: container.Id,
name: container.Names[0]?.replace(/^\//, '') || 'unnamed',
image: container.Image,
state: container.State,
status: container.Status,
created: container.Created,
ports: container.Ports || [],
networks,
health,
restartCount: restartCounts.get(container.Id) || 0,
mounts,
labels: container.Labels || {},
command: container.Command || ''
};
});
}
export async function getContainerStats(id: string, envId?: number | null) {
return dockerJsonRequest(`/containers/${id}/stats?stream=false`, {}, envId);
}
export async function startContainer(id: string, envId?: number | null) {
await dockerFetch(`/containers/${id}/start`, { method: 'POST' }, envId);
}
export async function stopContainer(id: string, envId?: number | null) {
await dockerFetch(`/containers/${id}/stop`, { method: 'POST' }, envId);
}
export async function restartContainer(id: string, envId?: number | null) {
await dockerFetch(`/containers/${id}/restart`, { method: 'POST' }, envId);
}
export async function pauseContainer(id: string, envId?: number | null) {
await dockerFetch(`/containers/${id}/pause`, { method: 'POST' }, envId);
}
export async function unpauseContainer(id: string, envId?: number | null) {
await dockerFetch(`/containers/${id}/unpause`, { method: 'POST' }, envId);
}
export async function removeContainer(id: string, force = false, envId?: number | null) {
const response = await dockerFetch(`/containers/${id}?force=${force}`, { method: 'DELETE' }, envId);
if (!response.ok) {
const errorBody = await response.text();
let errorMessage = `Failed to remove container ${id}`;
try {
const parsed = JSON.parse(errorBody);
if (parsed.message) {
errorMessage = parsed.message;
}
} catch {
if (errorBody) {
errorMessage = errorBody;
}
}
throw new Error(errorMessage);
}
}
export async function renameContainer(id: string, newName: string, envId?: number | null) {
await dockerFetch(`/containers/${id}/rename?name=${encodeURIComponent(newName)}`, { method: 'POST' }, envId);
}
export async function getContainerLogs(id: string, tail = 100, envId?: number | null): Promise<string> {
// Check if container has TTY enabled
const info = await inspectContainer(id, envId);
const hasTty = info.Config?.Tty ?? false;
const response = await dockerFetch(
`/containers/${id}/logs?stdout=true&stderr=true&tail=${tail}&timestamps=true`,
{},
envId
);
const buffer = Buffer.from(await response.arrayBuffer());
// If TTY is enabled, logs are raw text (no demux needed)
if (hasTty) {
return buffer.toString('utf-8');
}
return demuxDockerStream(buffer) as string;
}
export async function inspectContainer(id: string, envId?: number | null): Promise<ContainerInspectResult> {
return dockerJsonRequest<ContainerInspectResult>(`/containers/${id}/json`, {}, envId);
}
export interface HealthcheckConfig {
test?: string[];
interval?: number;
timeout?: number;
retries?: number;
startPeriod?: number;
}
export interface UlimitConfig {
name: string;
soft: number;
hard: number;
}
export interface DeviceMapping {
hostPath: string;
containerPath: string;
permissions?: string;
}
export interface CreateContainerOptions {
name: string;
image: string;
ports?: { [key: string]: { HostPort: string } };
volumes?: { [key: string]: {} };
volumeBinds?: string[];
env?: string[];
labels?: { [key: string]: string };
cmd?: string[];
restartPolicy?: string;
networkMode?: string;
networks?: string[];
user?: string;
privileged?: boolean;
healthcheck?: HealthcheckConfig;
memory?: number;
memoryReservation?: number;
cpuShares?: number;
cpuQuota?: number;
cpuPeriod?: number;
nanoCpus?: number;
capAdd?: string[];
capDrop?: string[];
devices?: DeviceMapping[];
dns?: string[];
dnsSearch?: string[];
dnsOptions?: string[];
securityOpt?: string[];
ulimits?: UlimitConfig[];
}
export async function createContainer(options: CreateContainerOptions, envId?: number | null) {
const containerConfig: any = {
Image: options.image,
Env: options.env || [],
Labels: options.labels || {},
HostConfig: {
RestartPolicy: {
Name: options.restartPolicy || 'no'
}
}
};
if (options.cmd && options.cmd.length > 0) {
containerConfig.Cmd = options.cmd;
}
if (options.user) {
containerConfig.User = options.user;
}
if (options.healthcheck) {
containerConfig.Healthcheck = {};
if (options.healthcheck.test && options.healthcheck.test.length > 0) {
containerConfig.Healthcheck.Test = options.healthcheck.test;
}
if (options.healthcheck.interval !== undefined) {
containerConfig.Healthcheck.Interval = options.healthcheck.interval;
}
if (options.healthcheck.timeout !== undefined) {
containerConfig.Healthcheck.Timeout = options.healthcheck.timeout;
}
if (options.healthcheck.retries !== undefined) {
containerConfig.Healthcheck.Retries = options.healthcheck.retries;
}
if (options.healthcheck.startPeriod !== undefined) {
containerConfig.Healthcheck.StartPeriod = options.healthcheck.startPeriod;
}
}
if (options.ports) {
containerConfig.ExposedPorts = {};
containerConfig.HostConfig.PortBindings = {};
for (const [containerPort, hostConfig] of Object.entries(options.ports)) {
containerConfig.ExposedPorts[containerPort] = {};
containerConfig.HostConfig.PortBindings[containerPort] = [hostConfig];
}
}
if (options.volumeBinds && options.volumeBinds.length > 0) {
containerConfig.HostConfig.Binds = options.volumeBinds;
}
if (options.volumes) {
containerConfig.Volumes = options.volumes;
}
if (options.networkMode) {
containerConfig.HostConfig.NetworkMode = options.networkMode;
}
if (options.networks && options.networks.length > 0) {
containerConfig.HostConfig.NetworkMode = options.networks[0];
containerConfig.NetworkingConfig = {
EndpointsConfig: {
[options.networks[0]]: {}
}
};
}
if (options.privileged) {
containerConfig.HostConfig.Privileged = options.privileged;
}
if (options.memory) {
containerConfig.HostConfig.Memory = options.memory;
}
if (options.memoryReservation) {
containerConfig.HostConfig.MemoryReservation = options.memoryReservation;
}
if (options.cpuShares) {
containerConfig.HostConfig.CpuShares = options.cpuShares;
}
if (options.cpuQuota) {
containerConfig.HostConfig.CpuQuota = options.cpuQuota;
}
if (options.cpuPeriod) {
containerConfig.HostConfig.CpuPeriod = options.cpuPeriod;
}
if (options.nanoCpus) {
containerConfig.HostConfig.NanoCpus = options.nanoCpus;
}
if (options.capAdd && options.capAdd.length > 0) {
containerConfig.HostConfig.CapAdd = options.capAdd;
}
if (options.capDrop && options.capDrop.length > 0) {
containerConfig.HostConfig.CapDrop = options.capDrop;
}
if (options.devices && options.devices.length > 0) {
containerConfig.HostConfig.Devices = options.devices.map(d => ({
PathOnHost: d.hostPath,
PathInContainer: d.containerPath,
CgroupPermissions: d.permissions || 'rwm'
}));
}
if (options.dns && options.dns.length > 0) {
containerConfig.HostConfig.Dns = options.dns;
}
if (options.dnsSearch && options.dnsSearch.length > 0) {
containerConfig.HostConfig.DnsSearch = options.dnsSearch;
}
if (options.dnsOptions && options.dnsOptions.length > 0) {
containerConfig.HostConfig.DnsOptions = options.dnsOptions;
}
if (options.securityOpt && options.securityOpt.length > 0) {
containerConfig.HostConfig.SecurityOpt = options.securityOpt;
}
if (options.ulimits && options.ulimits.length > 0) {
containerConfig.HostConfig.Ulimits = options.ulimits.map(u => ({
Name: u.name,
Soft: u.soft,
Hard: u.hard
}));
}
const result = await dockerJsonRequest<{ Id: string }>(
`/containers/create?name=${encodeURIComponent(options.name)}`,
{
method: 'POST',
body: JSON.stringify(containerConfig)
},
envId
);
// Connect to additional networks after container creation
if (options.networks && options.networks.length > 1) {
for (let i = 1; i < options.networks.length; i++) {
await dockerFetch(
`/networks/${options.networks[i]}/connect`,
{
method: 'POST',
headers: { 'Content-Type': 'application/json' },
body: JSON.stringify({ Container: result.Id })
},
envId
);
}
}
return { id: result.Id, start: () => startContainer(result.Id, envId) };
}
export async function updateContainer(id: string, options: CreateContainerOptions, startAfterUpdate = false, envId?: number | null) {
const oldContainerInfo = await inspectContainer(id, envId);
const wasRunning = oldContainerInfo.State.Running;
if (wasRunning) {
await stopContainer(id, envId);
}
await removeContainer(id, true, envId);
const newContainer = await createContainer(options, envId);
if (startAfterUpdate || wasRunning) {
await newContainer.start();
}
return newContainer;
}
// Image operations
export async function listImages(envId?: number | null): Promise<ImageInfo[]> {
const images = await dockerJsonRequest<any[]>('/images/json', {}, envId);
return images.map((image) => ({
id: image.Id,
tags: image.RepoTags || [],
size: image.Size,
created: image.Created
}));
}
export async function pullImage(imageName: string, onProgress?: (data: any) => void, envId?: number | null) {
// Parse image name and tag to avoid pulling all tags
// Docker API: if tag is empty, it pulls ALL tags for the image
// Format can be: repo:tag, repo@digest, or just repo (defaults to :latest)
let fromImage = imageName;
let tag = 'latest';
if (imageName.includes('@')) {
// Image with digest: repo@sha256:abc123
// Don't split, pass as-is (digest is part of fromImage)
fromImage = imageName;
tag = ''; // Empty tag when using digest
} else if (imageName.includes(':')) {
// Image with tag: repo:tag or registry.example.com/repo:tag
const lastColonIndex = imageName.lastIndexOf(':');
const potentialTag = imageName.substring(lastColonIndex + 1);
// Make sure we're not splitting on a port number (e.g., registry.example.com:5000/repo)
// Tags don't contain slashes, but registry ports are followed by a path
if (!potentialTag.includes('/')) {
fromImage = imageName.substring(0, lastColonIndex);
tag = potentialTag;
}
}
// Build URL with explicit tag parameter to prevent pulling all tags
const url = tag
? `/images/create?fromImage=${encodeURIComponent(fromImage)}&tag=${encodeURIComponent(tag)}`
: `/images/create?fromImage=${encodeURIComponent(fromImage)}`;
// Look up registry credentials for authenticated pulls
const headers: Record<string, string> = {};
try {
const { registry } = parseImageReference(imageName);
const creds = await findRegistryCredentials(registry);
if (creds) {
console.log(`[Pull] Using credentials for ${registry} (user: ${creds.username})`);
// Docker API expects X-Registry-Auth header with base64-encoded JSON
const authConfig = {
username: creds.username,
password: creds.password,
serveraddress: registry
};
headers['X-Registry-Auth'] = Buffer.from(JSON.stringify(authConfig)).toString('base64');
} else {
console.log(`[Pull] No credentials found for ${registry}`);
}
} catch (e) {
console.error(`[Pull] Failed to lookup credentials:`, e);
}
// Use streaming: true for longer timeout on edge environments
const response = await dockerFetch(url, { method: 'POST', streaming: true, headers }, envId);
if (!response.ok) {
throw new Error(`Failed to pull image: ${await response.text()}`);
}
// Stream the response for progress updates
const reader = response.body?.getReader();
if (!reader) return;
const decoder = new TextDecoder();
let buffer = '';
while (true) {
const { done, value } = await reader.read();
if (done) break;
buffer += decoder.decode(value, { stream: true });
const lines = buffer.split('\n');
buffer = lines.pop() || '';
for (const line of lines) {
if (line.trim()) {
try {
const data = JSON.parse(line);
if (onProgress) onProgress(data);
} catch {
// Ignore parse errors
}
}
}
}
}
export async function removeImage(id: string, force = false, envId?: number | null) {
const response = await dockerFetch(`/images/${encodeURIComponent(id)}?force=${force}`, { method: 'DELETE' }, envId);
if (!response.ok) {
const data = await response.json().catch(() => ({}));
const error: any = new Error(data.message || 'Failed to remove image');
error.statusCode = response.status;
error.json = data;
throw error;
}
}
export async function getImageHistory(id: string, envId?: number | null) {
return dockerJsonRequest(`/images/${encodeURIComponent(id)}/history`, {}, envId);
}
export async function inspectImage(id: string, envId?: number | null) {
return dockerJsonRequest(`/images/${encodeURIComponent(id)}/json`, {}, envId);
}
/**
* Parse an image reference into registry, repository, and tag components.
* Follows Docker's reference parsing rules.
* Examples:
* nginx:latest -> { registry: 'index.docker.io', repo: 'library/nginx', tag: 'latest' }
* ghcr.io/user/image:v1 -> { registry: 'ghcr.io', repo: 'user/image', tag: 'v1' }
* registry.example.com:5000/repo:tag -> { registry: 'registry.example.com:5000', repo: 'repo', tag: 'tag' }
*/
function parseImageReference(imageName: string): { registry: string; repo: string; tag: string } {
let registry = 'index.docker.io'; // Docker Hub's actual host
let repo = imageName;
let tag = 'latest';
// Handle digest references (remove digest part for manifest lookup)
if (repo.includes('@')) {
const [repoWithoutDigest] = repo.split('@');
repo = repoWithoutDigest;
}
// Extract tag
const lastColon = repo.lastIndexOf(':');
if (lastColon > -1) {
const potentialTag = repo.substring(lastColon + 1);
// Make sure it's not a port number (no slashes in tags)
if (!potentialTag.includes('/')) {
tag = potentialTag;
repo = repo.substring(0, lastColon);
}
}
// Extract registry if present
const firstSlash = repo.indexOf('/');
if (firstSlash > -1) {
const firstPart = repo.substring(0, firstSlash);
// If the first part contains a dot, colon, or is "localhost", it's a registry
if (firstPart.includes('.') || firstPart.includes(':') || firstPart === 'localhost') {
registry = firstPart;
repo = repo.substring(firstSlash + 1);
}
}
// Docker Hub requires library/ prefix for official images
if (registry === 'index.docker.io' && !repo.includes('/')) {
repo = `library/${repo}`;
}
return { registry, repo, tag };
}
/**
* Find registry credentials from Dockhand's stored registries.
* Matches by registry host (url field).
*/
async function findRegistryCredentials(registryHost: string): Promise<{ username: string; password: string } | null> {
try {
// Import here to avoid circular dependency
const { getRegistries } = await import('./db.js');
const registries = await getRegistries();
for (const reg of registries) {
// Match by URL - extract host from stored URL
const storedHost = reg.url.replace(/^https?:\/\//, '').replace(/\/.*$/, '');
if (storedHost === registryHost || reg.url.includes(registryHost)) {
if (reg.username && reg.password) {
return { username: reg.username, password: reg.password };
}
}
}
// Also check for Docker Hub variations
if (registryHost === 'index.docker.io' || registryHost === 'registry-1.docker.io') {
for (const reg of registries) {
const storedHost = reg.url.replace(/^https?:\/\//, '').replace(/\/.*$/, '');
// Match all Docker Hub URL variations
if (storedHost === 'docker.io' || storedHost === 'hub.docker.com' ||
storedHost === 'registry.hub.docker.com' || storedHost === 'index.docker.io' ||
storedHost === 'registry-1.docker.io') {
if (reg.username && reg.password) {
return { username: reg.username, password: reg.password };
}
}
}
}
return null;
} catch (e) {
console.error('Failed to lookup registry credentials:', e);
return null;
}
}
/**
* Get bearer token from registry using challenge-response flow.
* This follows the Docker Registry v2 authentication spec:
* 1. Make request to /v2/ to get WWW-Authenticate challenge
* 2. Parse realm, service, scope from challenge
* 3. Request token from realm URL (with credentials if available)
*/
async function getRegistryBearerToken(registry: string, repo: string): Promise<string | null> {
try {
const registryUrl = `https://${registry}`;
// Look up stored credentials for this registry
const credentials = await findRegistryCredentials(registry);
// Step 1: Challenge request to /v2/
const challengeResponse = await fetch(`${registryUrl}/v2/`, {
method: 'GET',
headers: { 'User-Agent': 'Dockhand/1.0' }
});
// If 200, no auth needed
if (challengeResponse.ok) {
return null;
}
// If not 401, something else is wrong
if (challengeResponse.status !== 401) {
console.error(`Registry challenge failed: ${challengeResponse.status}`);
return null;
}
// Step 2: Parse WWW-Authenticate header
const wwwAuth = challengeResponse.headers.get('WWW-Authenticate') || '';
const challenge = wwwAuth.toLowerCase();
if (challenge.startsWith('basic')) {
// Basic auth - use credentials if we have them
if (credentials) {
const basicAuth = Buffer.from(`${credentials.username}:${credentials.password}`).toString('base64');
return `Basic ${basicAuth}`;
}
return null;
}
if (!challenge.startsWith('bearer')) {
console.error(`Unsupported auth type: ${wwwAuth}`);
return null;
}
// Parse bearer challenge: Bearer realm="...",service="...",scope="..."
const realmMatch = wwwAuth.match(/realm="([^"]+)"/i);
const serviceMatch = wwwAuth.match(/service="([^"]+)"/i);
if (!realmMatch) {
console.error('No realm in WWW-Authenticate header');
return null;
}
const realm = realmMatch[1];
const service = serviceMatch ? serviceMatch[1] : '';
const scope = `repository:${repo}:pull`;
// Step 3: Request token from realm (with credentials if available)
const tokenUrl = new URL(realm);
if (service) tokenUrl.searchParams.set('service', service);
tokenUrl.searchParams.set('scope', scope);
const tokenHeaders: Record<string, string> = { 'User-Agent': 'Dockhand/1.0' };
// Add Basic auth header if we have credentials
if (credentials) {
const basicAuth = Buffer.from(`${credentials.username}:${credentials.password}`).toString('base64');
tokenHeaders['Authorization'] = `Basic ${basicAuth}`;
}
const tokenResponse = await fetch(tokenUrl.toString(), {
headers: tokenHeaders
});
if (!tokenResponse.ok) {
console.error(`Token request failed: ${tokenResponse.status}`);
return null;
}
const tokenData = await tokenResponse.json() as { token?: string; access_token?: string };
const token = tokenData.token || tokenData.access_token || null;
return token ? `Bearer ${token}` : null;
} catch (e) {
console.error('Failed to get registry bearer token:', e);
return null;
}
}
/**
* Check the registry for the current manifest digest of an image.
* Simple HEAD request to get Docker-Content-Digest header.
* Docker stores the manifest list digest in RepoDigests, so we compare that directly.
*/
export async function getRegistryManifestDigest(imageName: string): Promise<string | null> {
try {
const { registry, repo, tag } = parseImageReference(imageName);
const token = await getRegistryBearerToken(registry, repo);
const manifestUrl = `https://${registry}/v2/${repo}/manifests/${tag}`;
const headers: Record<string, string> = {
'User-Agent': 'Dockhand/1.0',
'Accept': [
'application/vnd.docker.distribution.manifest.list.v2+json',
'application/vnd.oci.image.index.v1+json',
'application/vnd.docker.distribution.manifest.v2+json',
'application/vnd.oci.image.manifest.v1+json'
].join(', ')
};
if (token) headers['Authorization'] = token;
const response = await fetch(manifestUrl, { method: 'HEAD', headers });
if (!response.ok) {
if (response.status !== 429) {
console.error(`[Registry] ${imageName}: ${response.status}`);
}
return null;
}
return response.headers.get('Docker-Content-Digest');
} catch (e) {
console.error(`[Registry] ${imageName}: ${e}`);
return null;
}
}
export interface ImageUpdateCheckResult {
hasUpdate: boolean;
currentDigest?: string;
registryDigest?: string;
/** True if this is a local-only image (no registry) */
isLocalImage?: boolean;
/** Error message if check failed */
error?: string;
}
/**
* Check if an image has an update available by comparing local digests against registry.
* This is a lightweight check that doesn't pull the image.
*
* @param imageName - The image name with optional tag (e.g., "nginx:latest")
* @param currentImageId - The sha256 ID of the current local image
* @param envId - Optional environment ID for multi-environment support
* @returns Update check result with hasUpdate flag and digest info
*/
export async function checkImageUpdateAvailable(
imageName: string,
currentImageId: string,
envId?: number
): Promise<ImageUpdateCheckResult> {
try {
// Get current image info to get RepoDigests
let currentImageInfo: any;
try {
currentImageInfo = await inspectImage(currentImageId, envId);
} catch {
return { hasUpdate: false, error: 'Could not inspect current image' };
}
const currentRepoDigests: string[] = currentImageInfo?.RepoDigests || [];
// Extract digest part from RepoDigest (format: repo@sha256:...)
const extractDigest = (rd: string): string | null => {
const atIndex = rd.lastIndexOf('@');
return atIndex > -1 ? rd.substring(atIndex + 1) : null;
};
// Get ALL local digests - an image can have multiple RepoDigests
// (e.g., when a tag is updated but the content for your architecture is the same)
const localDigests = currentRepoDigests
.map(extractDigest)
.filter((d): d is string => d !== null);
// If no local digests, this is likely a local-only image
if (localDigests.length === 0) {
return {
hasUpdate: false,
isLocalImage: true,
currentDigest: currentImageId
};
}
// Query registry for current manifest digest
const registryDigest = await getRegistryManifestDigest(imageName);
if (!registryDigest) {
// Registry unreachable or image not found - can't determine update status
return {
hasUpdate: false,
currentDigest: currentRepoDigests[0],
error: 'Could not query registry'
};
}
// Check if registry digest matches ANY of the local digests
const matchesLocal = localDigests.includes(registryDigest);
const hasUpdate = !matchesLocal;
return {
hasUpdate,
currentDigest: currentRepoDigests[0],
registryDigest: hasUpdate ? registryDigest : undefined
};
} catch (e: any) {
return { hasUpdate: false, error: e.message };
}
}
export async function tagImage(id: string, repo: string, tag: string, envId?: number | null) {
await dockerFetch(
`/images/${encodeURIComponent(id)}/tag?repo=${encodeURIComponent(repo)}&tag=${encodeURIComponent(tag)}`,
{ method: 'POST' },
envId
);
}
/**
* Generate a temporary tag name for safe pulling during auto-updates.
* This allows scanning the new image before committing to the update.
* @param imageName - The original image name (e.g., "nginx:latest" or "nginx")
* @returns Temporary tag name (e.g., "nginx:latest-dockhand-pending")
*/
export function getTempImageTag(imageName: string): string {
// Handle images with digest (e.g., nginx@sha256:abc123)
if (imageName.includes('@')) {
// For digest-based images, we can't use temp tags - return as-is
return imageName;
}
// Find the last colon
const lastColon = imageName.lastIndexOf(':');
// No colon at all - simple image like "nginx"
if (lastColon === -1) {
return `${imageName}:latest-dockhand-pending`;
}
const afterColon = imageName.substring(lastColon + 1);
// If the part after the last colon contains a slash, it's a port number
// e.g., "registry:5000/nginx" -> afterColon = "5000/nginx"
// In this case, there's no tag, so we append :latest-dockhand-pending
if (afterColon.includes('/')) {
return `${imageName}:latest-dockhand-pending`;
}
// Otherwise, the last colon separates repo from tag
// e.g., "registry.bor6.pl/test:latest" -> repo="registry.bor6.pl/test", tag="latest"
const repo = imageName.substring(0, lastColon);
const tag = afterColon;
return `${repo}:${tag}-dockhand-pending`;
}
/**
* Check if an image name is using a digest (sha256) instead of a tag.
* Digest-based images don't need temp tag handling.
*/
export function isDigestBasedImage(imageName: string): boolean {
return imageName.includes('@sha256:');
}
/**
* Normalize an image tag for comparison.
* Docker Hub images can be represented as:
* - n8nio/n8n:latest
* - docker.io/n8nio/n8n:latest
* - docker.io/library/nginx:latest (for official images)
* - library/nginx:latest
* - nginx:latest
* Custom registries:
* - docker.n8n.io/n8nio/n8n (n8n's custom registry)
*/
function normalizeImageTag(tag: string): string {
let normalized = tag;
// Remove docker.io/ prefix
normalized = normalized.replace(/^docker\.io\//, '');
// Remove library/ prefix for official images
normalized = normalized.replace(/^library\//, '');
// Add :latest if no tag specified (and not a digest)
if (!normalized.includes(':') && !normalized.includes('@')) {
normalized = `${normalized}:latest`;
}
return normalized.toLowerCase();
}
/**
* Get image ID by tag name.
* Uses Docker's image inspect API which correctly resolves any image reference
* (docker.io, ghcr.io, custom registries, etc.)
* @returns Image ID (sha256:...) or null if not found
*/
export async function getImageIdByTag(tagName: string, envId?: number | null): Promise<string | null> {
try {
// First try: Use Docker's image inspect API - this is the most reliable
// as Docker knows exactly how to resolve the image name
const imageInfo = await inspectImage(tagName, envId) as { Id?: string } | null;
if (imageInfo?.Id) {
return imageInfo.Id;
}
} catch {
// Image inspect failed - fall back to listing images
}
try {
// Fallback: Search through listed images with normalization
const images = await listImages(envId);
const normalizedSearch = normalizeImageTag(tagName);
for (const image of images) {
if (image.tags) {
for (const tag of image.tags) {
if (normalizeImageTag(tag) === normalizedSearch) {
return image.id;
}
}
}
}
return null;
} catch {
return null;
}
}
/**
* Remove a temporary image by its tag.
* Used to clean up after a blocked auto-update.
* @param imageIdOrTag - Image ID or tag to remove
* @param force - Force removal even if image is in use
*/
export async function removeTempImage(imageIdOrTag: string, envId?: number | null, force = true): Promise<void> {
try {
await removeImage(imageIdOrTag, force, envId);
} catch (error: any) {
// Log but don't throw - cleanup failure shouldn't break the flow
console.warn(`[Docker] Failed to remove temp image ${imageIdOrTag}: ${error.message}`);
}
}
/**
* Export (save) an image as a tar archive stream.
* Uses Docker's GET /images/{name}/get endpoint.
* @returns Response object with tar stream body
*/
export async function exportImage(id: string, envId?: number | null): Promise<Response> {
const response = await dockerFetch(
`/images/${encodeURIComponent(id)}/get`,
{ method: 'GET', streaming: true },
envId
);
if (!response.ok) {
const error = await response.text().catch(() => 'Unknown error');
throw new Error(`Failed to export image: ${response.status} - ${error}`);
}
return response;
}
// System information
export async function getDockerInfo(envId?: number | null) {
return dockerJsonRequest('/info', {}, envId);
}
export async function getDockerVersion(envId?: number | null) {
return dockerJsonRequest('/version', {}, envId);
}
/**
* Get Hawser agent info (for hawser-standard mode)
* Returns agent info including uptime
*/
export async function getHawserInfo(envId: number): Promise<{
agentId: string;
agentName: string;
dockerVersion: string;
hawserVersion: string;
mode: string;
uptime: number;
} | null> {
try {
const response = await dockerFetch('/_hawser/info', {}, envId);
if (response.ok) {
return await response.json();
}
} catch {
// Hawser info not available
}
return null;
}
// Volume operations
export interface VolumeInfo {
name: string;
driver: string;
mountpoint: string;
scope: string;
created: string;
labels: { [key: string]: string };
}
export async function listVolumes(envId?: number | null): Promise<VolumeInfo[]> {
// Fetch volumes and containers in parallel
const [volumeResult, containers] = await Promise.all([
dockerJsonRequest<{ Volumes: any[] }>('/volumes', {}, envId),
dockerJsonRequest<any[]>('/containers/json?all=true', {}, envId)
]);
// Build a map of volume name -> containers using it
const volumeUsageMap = new Map<string, { containerId: string; containerName: string }[]>();
for (const container of containers) {
const containerName = container.Names?.[0]?.replace(/^\//, '') || 'unnamed';
const containerId = container.Id;
for (const mount of container.Mounts || []) {
// Check for volume-type mounts (not bind mounts)
if (mount.Type === 'volume' && mount.Name) {
const volumeName = mount.Name;
if (!volumeUsageMap.has(volumeName)) {
volumeUsageMap.set(volumeName, []);
}
volumeUsageMap.get(volumeName)!.push({ containerId, containerName });
}
}
}
return (volumeResult.Volumes || []).map((volume: any) => ({
name: volume.Name,
driver: volume.Driver,
mountpoint: volume.Mountpoint,
scope: volume.Scope,
created: volume.CreatedAt,
labels: volume.Labels || {},
usedBy: volumeUsageMap.get(volume.Name) || []
}));
}
/**
* Check if a volume is in use by any containers
* Returns list of containers using the volume
*/
export async function getVolumeUsage(
volumeName: string,
envId?: number | null
): Promise<{ containerId: string; containerName: string; state: string }[]> {
const containers = await dockerJsonRequest<any[]>('/containers/json?all=true', {}, envId);
const usage: { containerId: string; containerName: string; state: string }[] = [];
for (const container of containers) {
// Skip our own helper containers
if (container.Labels?.['dockhand.volume.helper'] === 'true') {
continue;
}
const containerName = container.Names?.[0]?.replace(/^\//, '') || 'unnamed';
const containerId = container.Id;
const state = container.State || 'unknown';
for (const mount of container.Mounts || []) {
if (mount.Type === 'volume' && mount.Name === volumeName) {
usage.push({ containerId, containerName, state });
break;
}
}
}
return usage;
}
export async function removeVolume(name: string, force = false, envId?: number | null) {
const response = await dockerFetch(`/volumes/${encodeURIComponent(name)}?force=${force}`, { method: 'DELETE' }, envId);
if (!response.ok) {
const data = await response.json().catch(() => ({}));
const error: any = new Error(data.message || 'Failed to remove volume');
error.statusCode = response.status;
error.json = data;
throw error;
}
}
export async function inspectVolume(name: string, envId?: number | null) {
return dockerJsonRequest(`/volumes/${encodeURIComponent(name)}`, {}, envId);
}
export interface CreateVolumeOptions {
name: string;
driver?: string;
driverOpts?: { [key: string]: string };
labels?: { [key: string]: string };
}
export async function createVolume(options: CreateVolumeOptions, envId?: number | null) {
const volumeConfig = {
Name: options.name,
Driver: options.driver || 'local',
DriverOpts: options.driverOpts || {},
Labels: options.labels || {}
};
return dockerJsonRequest('/volumes/create', {
method: 'POST',
body: JSON.stringify(volumeConfig)
}, envId);
}
// Network operations
export interface NetworkInfo {
id: string;
name: string;
driver: string;
scope: string;
internal: boolean;
ipam: {
driver: string;
config: Array<{ subnet?: string; gateway?: string }>;
};
containers: { [key: string]: { name: string; ipv4Address: string } };
}
export async function listNetworks(envId?: number | null): Promise<NetworkInfo[]> {
const networks = await dockerJsonRequest<any[]>('/networks', {}, envId);
// Docker's /networks endpoint returns empty Containers - we need to inspect each network
// to get the actual connected containers. Run inspections in parallel for performance.
const networkDetails = await Promise.all(
networks.map(async (network: any) => {
try {
const details = await dockerJsonRequest<any>(`/networks/${network.Id}`, {}, envId);
return {
...network,
Containers: details.Containers || {}
};
} catch {
// If inspection fails, return network with empty containers
return network;
}
})
);
return networkDetails.map((network: any) => ({
id: network.Id,
name: network.Name,
driver: network.Driver,
scope: network.Scope,
internal: network.Internal || false,
ipam: {
driver: network.IPAM?.Driver || 'default',
// Normalize IPAM config field names to lowercase for consistency
config: (network.IPAM?.Config || []).map((cfg: any) => ({
subnet: cfg.Subnet || cfg.subnet,
gateway: cfg.Gateway || cfg.gateway,
ipRange: cfg.IPRange || cfg.ipRange,
auxAddress: cfg.AuxAddress || cfg.auxAddress
}))
},
containers: Object.entries(network.Containers || {}).reduce((acc: any, [id, data]: [string, any]) => {
acc[id] = {
name: data.Name,
ipv4Address: data.IPv4Address
};
return acc;
}, {})
}));
}
export async function removeNetwork(id: string, envId?: number | null) {
const response = await dockerFetch(`/networks/${id}`, { method: 'DELETE' }, envId);
if (!response.ok) {
const data = await response.json().catch(() => ({}));
const error: any = new Error(data.message || 'Failed to remove network');
error.statusCode = response.status;
error.json = data;
throw error;
}
}
export async function inspectNetwork(id: string, envId?: number | null) {
return dockerJsonRequest(`/networks/${id}`, {}, envId);
}
export interface CreateNetworkOptions {
name: string;
driver?: string;
internal?: boolean;
attachable?: boolean;
ingress?: boolean;
enableIPv6?: boolean;
ipam?: {
driver?: string;
config?: Array<{
subnet?: string;
ipRange?: string;
gateway?: string;
auxAddress?: { [key: string]: string };
}>;
options?: { [key: string]: string };
};
options?: { [key: string]: string };
labels?: { [key: string]: string };
}
export async function createNetwork(options: CreateNetworkOptions, envId?: number | null) {
const networkConfig: any = {
Name: options.name,
Driver: options.driver || 'bridge',
Internal: options.internal || false,
Attachable: options.attachable || false,
Ingress: options.ingress || false,
EnableIPv6: options.enableIPv6 || false,
Options: options.options || {},
Labels: options.labels || {}
};
if (options.ipam) {
networkConfig.IPAM = {
Driver: options.ipam.driver || 'default',
Config: options.ipam.config?.map(cfg => ({
Subnet: cfg.subnet,
IPRange: cfg.ipRange,
Gateway: cfg.gateway,
AuxiliaryAddresses: cfg.auxAddress
})).filter(cfg => cfg.Subnet || cfg.Gateway) || [],
Options: options.ipam.options || {}
};
}
return dockerJsonRequest('/networks/create', {
method: 'POST',
body: JSON.stringify(networkConfig)
}, envId);
}
// Network connect/disconnect operations
export async function connectContainerToNetwork(
networkId: string,
containerId: string,
envId?: number | null
): Promise<void> {
const response = await dockerFetch(
`/networks/${networkId}/connect`,
{
method: 'POST',
headers: { 'Content-Type': 'application/json' },
body: JSON.stringify({ Container: containerId })
},
envId
);
if (!response.ok) {
const data = await response.json().catch(() => ({}));
throw new Error(data.message || 'Failed to connect container to network');
}
}
export async function disconnectContainerFromNetwork(
networkId: string,
containerId: string,
force = false,
envId?: number | null
): Promise<void> {
const response = await dockerFetch(
`/networks/${networkId}/disconnect`,
{
method: 'POST',
headers: { 'Content-Type': 'application/json' },
body: JSON.stringify({ Container: containerId, Force: force })
},
envId
);
if (!response.ok) {
const data = await response.json().catch(() => ({}));
throw new Error(data.message || 'Failed to disconnect container from network');
}
}
// Container exec operations
export interface ExecOptions {
containerId: string;
cmd: string[];
user?: string;
workingDir?: string;
envId?: number | null;
}
export async function createExec(options: ExecOptions): Promise<{ Id: string }> {
const execConfig = {
Cmd: options.cmd,
AttachStdin: true,
AttachStdout: true,
AttachStderr: true,
Tty: true,
User: options.user || 'root',
WorkingDir: options.workingDir
};
return dockerJsonRequest(`/containers/${options.containerId}/exec`, {
method: 'POST',
body: JSON.stringify(execConfig)
}, options.envId);
}
export async function resizeExec(execId: string, cols: number, rows: number, envId?: number | null) {
try {
await dockerFetch(`/exec/${execId}/resize?h=${rows}&w=${cols}`, { method: 'POST' }, envId);
} catch {
// Resize may fail if exec is not running, ignore
}
}
/**
* Get Docker connection info for direct WebSocket connections from the client
* This is used by the terminal to connect directly to the Docker API
*/
export async function getDockerConnectionInfo(envId?: number | null): Promise<{
type: 'socket' | 'http' | 'https';
socketPath?: string;
host?: string;
port?: number;
}> {
const config = await getDockerConfig(envId);
return {
type: config.type,
socketPath: config.socketPath,
host: config.host,
port: config.port
};
}
// System disk usage
export async function getDiskUsage(envId?: number | null) {
return dockerJsonRequest('/system/df', {}, envId);
}
// Prune operations
export async function pruneContainers(envId?: number | null) {
return dockerJsonRequest('/containers/prune', { method: 'POST' }, envId);
}
export async function pruneImages(dangling = true, envId?: number | null) {
const filters = dangling ? '{"dangling":["true"]}' : '{}';
return dockerJsonRequest(`/images/prune?filters=${encodeURIComponent(filters)}`, { method: 'POST' }, envId);
}
export async function pruneVolumes(envId?: number | null) {
return dockerJsonRequest('/volumes/prune', { method: 'POST' }, envId);
}
export async function pruneNetworks(envId?: number | null) {
return dockerJsonRequest('/networks/prune', { method: 'POST' }, envId);
}
export async function pruneAll(envId?: number | null) {
const containers = await pruneContainers(envId);
const images = await pruneImages(false, envId);
const volumes = await pruneVolumes(envId);
const networks = await pruneNetworks(envId);
return { containers, images, volumes, networks };
}
// Registry operations
export async function searchImages(term: string, limit = 25, envId?: number | null) {
return dockerJsonRequest(`/images/search?term=${encodeURIComponent(term)}&limit=${limit}`, {}, envId);
}
// List containers with size info (slower operation)
export async function listContainersWithSize(all = true, envId?: number | null): Promise<Record<string, { sizeRw: number; sizeRootFs: number }>> {
const containers = await dockerJsonRequest<any[]>(
`/containers/json?all=${all}&size=true`,
{},
envId
);
const sizes: Record<string, { sizeRw: number; sizeRootFs: number }> = {};
for (const container of containers) {
sizes[container.Id] = {
sizeRw: container.SizeRw || 0,
sizeRootFs: container.SizeRootFs || 0
};
}
return sizes;
}
// Get container top (process list)
export async function getContainerTop(id: string, envId?: number | null): Promise<{ Titles: string[]; Processes: string[][] }> {
return dockerJsonRequest(`/containers/${id}/top`, {}, envId);
}
// Execute a command in a container and return the output
export async function execInContainer(
containerId: string,
cmd: string[],
envId?: number | null
): Promise<string> {
// Create exec instance
const execCreate = await dockerJsonRequest<{ Id: string }>(
`/containers/${containerId}/exec`,
{
method: 'POST',
body: JSON.stringify({
Cmd: cmd,
AttachStdout: true,
AttachStderr: true,
Tty: false
})
},
envId
);
// Start exec and get output
const response = await dockerFetch(
`/exec/${execCreate.Id}/start`,
{
method: 'POST',
headers: { 'Content-Type': 'application/json' },
body: JSON.stringify({ Detach: false, Tty: false })
},
envId
);
const buffer = Buffer.from(await response.arrayBuffer());
const output = demuxDockerStream(buffer) as string;
// Check exit code by inspecting the exec instance
const execInfo = await dockerJsonRequest<{ ExitCode: number }>(
`/exec/${execCreate.Id}/json`,
{},
envId
);
if (execInfo.ExitCode !== 0) {
const errorMsg = output.trim() || `Command failed with exit code ${execInfo.ExitCode}`;
throw new Error(errorMsg);
}
return output;
}
// Get Docker events as a stream (for SSE)
export async function getDockerEvents(
filters: Record<string, string[]>,
envId?: number | null
): Promise<ReadableStream<Uint8Array> | null> {
const filterJson = JSON.stringify(filters);
try {
// Note: We use streaming: true to disable Bun's idle timeout for this long-lived connection.
// The Docker events API keeps the connection open indefinitely, sending events as they occur.
// Without streaming: true, Bun would terminate the connection after ~5 seconds of inactivity.
const response = await dockerFetch(
`/events?filters=${encodeURIComponent(filterJson)}`,
{ streaming: true },
envId
);
if (!response.ok) {
throw new Error(`Docker events API returned ${response.status}`);
}
return response.body;
} catch (error: any) {
throw error;
}
}
// Check if volume exists
export async function volumeExists(volumeName: string, envId?: number | null): Promise<boolean> {
try {
const volumes = await listVolumes(envId);
return volumes.some(v => v.name === volumeName);
} catch {
return false;
}
}
// Generate a random suffix for container names (avoids conflicts)
function randomSuffix(): string {
return Math.random().toString(36).substring(2, 8);
}
// Run a short-lived container and return stdout
export async function runContainer(options: {
image: string;
cmd: string[];
binds?: string[];
env?: string[];
name?: string;
autoRemove?: boolean;
envId?: number | null;
}): Promise<{ stdout: string; stderr: string }> {
// Add random suffix to avoid naming conflicts
const baseName = options.name || `dockhand-temp-${Date.now()}`;
const containerName = `${baseName}-${randomSuffix()}`;
// Create container
const containerConfig: any = {
Image: options.image,
Cmd: options.cmd,
Env: options.env || [],
Tty: false,
HostConfig: {
Binds: options.binds || [],
AutoRemove: options.autoRemove !== false
}
};
const createResult = await dockerJsonRequest<{ Id: string }>(
`/containers/create?name=${encodeURIComponent(containerName)}`,
{
method: 'POST',
body: JSON.stringify(containerConfig)
},
options.envId
);
const containerId = createResult.Id;
try {
// Start container
await dockerFetch(`/containers/${containerId}/start`, { method: 'POST' }, options.envId);
// Wait for container to finish
await dockerFetch(`/containers/${containerId}/wait`, { method: 'POST' }, options.envId);
// Get logs
const logsResponse = await dockerFetch(
`/containers/${containerId}/logs?stdout=true&stderr=true`,
{},
options.envId
);
const buffer = Buffer.from(await logsResponse.arrayBuffer());
return demuxDockerStream(buffer, { separateStreams: true }) as { stdout: string; stderr: string };
} finally {
// Cleanup container if not auto-removed
if (options.autoRemove === false) {
try {
await dockerFetch(`/containers/${containerId}?force=true`, { method: 'DELETE' }, options.envId);
} catch {
// Ignore cleanup errors
}
}
}
}
// Run a container with attached streams (for scanners that need real-time output)
export async function runContainerWithStreaming(options: {
image: string;
cmd: string[];
binds?: string[];
env?: string[];
name?: string;
envId?: number | null;
onStdout?: (data: string) => void;
onStderr?: (data: string) => void;
}): Promise<string> {
// Add random suffix to avoid naming conflicts
const baseName = options.name || `dockhand-stream-${Date.now()}`;
const containerName = `${baseName}-${randomSuffix()}`;
// Create container
const containerConfig: any = {
Image: options.image,
Cmd: options.cmd,
Env: options.env || [],
Tty: false,
HostConfig: {
Binds: options.binds || [],
AutoRemove: true
}
};
// Try to create container, handle 409 conflict by removing stale container
let createResult: { Id: string };
try {
createResult = await dockerJsonRequest<{ Id: string }>(
`/containers/create?name=${encodeURIComponent(containerName)}`,
{
method: 'POST',
body: JSON.stringify(containerConfig)
},
options.envId
);
} catch (error: any) {
// Check for 409 conflict (container name already in use)
if (error?.message?.includes('409') || error?.status === 409) {
console.log(`[Docker] Container name conflict for ${containerName}, attempting cleanup...`);
// Try to force remove the conflicting container
try {
await dockerFetch(`/containers/${containerName}?force=true`, { method: 'DELETE' }, options.envId);
console.log(`[Docker] Removed stale container ${containerName}`);
} catch (removeError) {
console.error(`[Docker] Failed to remove stale container:`, removeError);
}
// Retry with a new random suffix
const retryName = `${baseName}-${randomSuffix()}`;
createResult = await dockerJsonRequest<{ Id: string }>(
`/containers/create?name=${encodeURIComponent(retryName)}`,
{
method: 'POST',
body: JSON.stringify(containerConfig)
},
options.envId
);
} else {
throw error;
}
}
const containerId = createResult.Id;
// Start container
await dockerFetch(`/containers/${containerId}/start`, { method: 'POST' }, options.envId);
// Check if this is an edge environment for streaming approach
const config = await getDockerConfig(options.envId ?? undefined);
// Stream logs while container is running
if (config.connectionType === 'hawser-edge' && config.environmentId) {
// Edge mode: use sendEdgeStreamRequest for real-time streaming
return new Promise<string>((resolve, reject) => {
let stdout = '';
let buffer: Buffer<ArrayBufferLike> = Buffer.alloc(0);
const { cancel } = sendEdgeStreamRequest(
config.environmentId!,
'GET',
`/containers/${containerId}/logs?stdout=true&stderr=true&follow=true`,
{
onData: (data: string) => {
try {
// Data is base64 encoded from edge agent
const decoded = Buffer.from(data, 'base64');
buffer = Buffer.concat([buffer, decoded]);
// Process Docker stream frames
const result = processStreamFrames(buffer, options.onStdout, options.onStderr);
stdout += result.stdout;
buffer = result.remaining;
} catch {
// If not base64, try as raw data
const result = processStreamFrames(Buffer.from(data), options.onStdout, options.onStderr);
stdout += result.stdout;
}
},
onEnd: () => {
resolve(stdout);
},
onError: (error: string) => {
// If container finished, treat as success
if (error.includes('container') && (error.includes('exited') || error.includes('not running'))) {
resolve(stdout);
} else {
reject(new Error(error));
}
}
}
);
});
}
// Non-edge mode: use regular streaming
const logsResponse = await dockerFetch(
`/containers/${containerId}/logs?stdout=true&stderr=true&follow=true`,
{ streaming: true },
options.envId
);
let stdout = '';
const reader = logsResponse.body?.getReader();
if (reader) {
let buffer: Buffer<ArrayBufferLike> = Buffer.alloc(0);
while (true) {
const { done, value } = await reader.read();
if (done) break;
buffer = Buffer.concat([buffer, Buffer.from(value)]);
const result = processStreamFrames(buffer, options.onStdout, options.onStderr);
stdout += result.stdout;
buffer = result.remaining;
}
}
return stdout;
}
// Push image to registry
export async function pushImage(
imageTag: string,
authConfig: { username?: string; password?: string; serveraddress: string },
onProgress?: (data: any) => void,
envId?: number | null
): Promise<void> {
// Parse tag to get registry info
const [repo, tag = 'latest'] = imageTag.split(':');
// Create X-Registry-Auth header
const authHeader = Buffer.from(JSON.stringify(authConfig)).toString('base64');
const response = await dockerFetch(
`/images/${encodeURIComponent(imageTag)}/push`,
{
method: 'POST',
headers: {
'X-Registry-Auth': authHeader
}
},
envId
);
if (!response.ok) {
const error = await response.text();
throw new Error(`Failed to push image: ${error}`);
}
// Stream the response for progress updates
const reader = response.body?.getReader();
if (!reader) return;
const decoder = new TextDecoder();
let buffer = '';
while (true) {
const { done, value } = await reader.read();
if (done) break;
buffer += decoder.decode(value, { stream: true });
const lines = buffer.split('\n');
buffer = lines.pop() || '';
for (const line of lines) {
if (line.trim()) {
try {
const data = JSON.parse(line);
if (data.error) {
throw new Error(data.error);
}
if (onProgress) onProgress(data);
} catch (e: any) {
if (e.message && !e.message.includes('JSON')) {
throw e;
}
}
}
}
}
}
// Container filesystem operations
export interface FileEntry {
name: string;
type: 'file' | 'directory' | 'symlink' | 'other';
size: number;
permissions: string;
owner: string;
group: string;
modified: string;
linkTarget?: string;
readonly?: boolean;
}
/**
* Parse ls -la output into FileEntry array
* Handles multiple formats:
* - GNU ls with --time-style=iso: drwxr-xr-x 2 root root 4096 2024-12-08 10:30 dirname
* - Standard GNU ls: drwxr-xr-x 2 root root 4096 Dec 8 10:30 dirname
* - Busybox ls: drwxr-xr-x 2 root root 4096 Dec 8 10:30 dirname
*/
function parseLsOutput(output: string): FileEntry[] {
const lines = output.trim().split('\n');
const entries: FileEntry[] = [];
const currentYear = new Date().getFullYear();
// Month name to number mapping
const monthMap: Record<string, string> = {
Jan: '01',
Feb: '02',
Mar: '03',
Apr: '04',
May: '05',
Jun: '06',
Jul: '07',
Aug: '08',
Sep: '09',
Oct: '10',
Nov: '11',
Dec: '12'
};
for (const line of lines) {
// Skip total line, empty lines, and error messages
if (!line || line.startsWith('total ') || line.includes('cannot access') || line.includes('Permission denied')) continue;
let typeChar: string;
let perms: string;
let owner: string;
let group: string;
let sizeStr: string;
let date: string;
let time: string;
let nameAndLink: string;
// Try ISO format first (GNU ls with --time-style=iso)
// Format: drwxr-xr-x 2 root root 4096 2024-12-08 10:30 dirname
// With ACL: drwxr-xr-x+ 2 root root 4096 2024-12-08 10:30 dirname
// With extended attrs: drwxr-xr-x@ 2 root root 4096 2024-12-08 10:30 dirname
const isoMatch = line.match(
/^([dlcbps-])([rwxsStT-]{9})[+@.]?\s+\d+\s+(\S+)\s+(\S+)\s+(\d+)\s+(\d{2,4}-\d{2}(?:-\d{2})?)\s+(\d{2}:\d{2})\s+(.+)$/
);
if (isoMatch) {
[, typeChar, perms, owner, group, sizeStr, date, time, nameAndLink] = isoMatch;
// Normalize date to YYYY-MM-DD format
if (date.length <= 5) {
// Format: MM-DD (no year)
date = `${currentYear}-${date}`;
} else if (!date.includes('-', 4)) {
// Format: YYYY-MM (no day)
date = `${date}-01`;
}
} else {
// Try standard format (GNU/busybox without --time-style)
// Format: drwxr-xr-x 2 root root 4096 Dec 8 10:30 dirname
// Or: drwxr-xr-x 2 root root 4096 Dec 8 10:30 dirname
// Or with year: drwxr-xr-x 2 root root 4096 Dec 8 2023 dirname
// With ACL/attrs: drwxr-xr-x+ or drwxr-xr-x@ or drwxr-xr-x.
const stdMatch = line.match(
/^([dlcbps-])([rwxsStT-]{9})[+@.]?\s+\d+\s+(\S+)\s+(\S+)\s+(\d+)\s+(\w{3})\s+(\d{1,2})\s+(\d{1,2}:\d{2}|\d{4})\s+(.+)$/
);
if (!stdMatch) {
// Try device file format (block/char devices have major,minor instead of size)
// Format: crw-rw-rw- 1 root root 1, 3 Dec 8 10:30 null
const deviceMatch = line.match(
/^([cb])([rwxsStT-]{9})[+@.]?\s+\d+\s+(\S+)\s+(\S+)\s+(\d+),\s*(\d+)\s+(\w{3})\s+(\d{1,2})\s+(\d{1,2}:\d{2}|\d{4})\s+(.+)$/
);
if (deviceMatch) {
let monthStr: string;
let dayStr: string;
let timeOrYear: string;
[, typeChar, perms, owner, group, , , monthStr, dayStr, timeOrYear, nameAndLink] = deviceMatch;
sizeStr = '0'; // Device files don't have a traditional size
const month = monthMap[monthStr] || '01';
const day = dayStr.padStart(2, '0');
if (timeOrYear.includes(':')) {
time = timeOrYear;
date = `${currentYear}-${month}-${day}`;
} else {
time = '00:00';
date = `${timeOrYear}-${month}-${day}`;
}
} else {
continue;
}
} else {
let monthStr: string;
let dayStr: string;
let timeOrYear: string;
[, typeChar, perms, owner, group, sizeStr, monthStr, dayStr, timeOrYear, nameAndLink] =
stdMatch;
const month = monthMap[monthStr] || '01';
const day = dayStr.padStart(2, '0');
// timeOrYear is either "HH:MM" or "YYYY"
if (timeOrYear.includes(':')) {
time = timeOrYear;
date = `${currentYear}-${month}-${day}`;
} else {
time = '00:00';
date = `${timeOrYear}-${month}-${day}`;
}
}
}
let type: FileEntry['type'];
switch (typeChar) {
case 'd':
type = 'directory';
break;
case 'l':
type = 'symlink';
break;
case '-':
type = 'file';
break;
default:
type = 'other';
}
let name = nameAndLink;
let linkTarget: string | undefined;
// Handle symlinks: "name -> target"
if (type === 'symlink' && nameAndLink.includes(' -> ')) {
const parts = nameAndLink.split(' -> ');
name = parts[0];
linkTarget = parts.slice(1).join(' -> ');
}
// Skip . and .. entries
if (name === '.' || name === '..') continue;
// Check if file is read-only (owner doesn't have write permission)
// perms format: rwxrwxrwx - index 1 is owner write
const isReadonly = perms.charAt(1) !== 'w';
entries.push({
name,
type,
size: parseInt(sizeStr, 10),
permissions: perms,
owner,
group,
modified: `${date}T${time}:00`,
linkTarget,
readonly: isReadonly
});
}
return entries;
}
/**
* List files in a container directory
* Tries multiple ls command variants for compatibility with different containers.
*/
export async function listContainerDirectory(
containerId: string,
path: string,
envId?: number | null,
useSimpleLs?: boolean
): Promise<{ path: string; entries: FileEntry[] }> {
// Sanitize path to prevent command injection
const safePath = path.replace(/[;&|`$(){}[\]<>'"\\]/g, '');
// Commands to try in order of preference
const commands = useSimpleLs
? [
['ls', '-la', safePath],
['/bin/ls', '-la', safePath],
['/usr/bin/ls', '-la', safePath],
]
: [
['ls', '-la', '--time-style=iso', safePath],
['ls', '-la', safePath],
['/bin/ls', '-la', safePath],
['/usr/bin/ls', '-la', safePath],
];
let lastError: Error | null = null;
for (const cmd of commands) {
try {
const output = await execInContainer(containerId, cmd, envId);
const entries = parseLsOutput(output);
return { path: safePath, entries };
} catch (err: any) {
lastError = err;
continue;
}
}
throw lastError || new Error('Failed to list directory: no working ls command found');
}
/**
* Get file/directory archive from container (for download)
* Returns the raw Docker API response for streaming
*/
export async function getContainerArchive(
containerId: string,
path: string,
envId?: number | null
): Promise<Response> {
// Sanitize path
const safePath = path.replace(/[;&|`$(){}[\]<>'"\\]/g, '');
const response = await dockerFetch(
`/containers/${containerId}/archive?path=${encodeURIComponent(safePath)}`,
{},
envId
);
if (!response.ok) {
const error = await response.text();
throw new Error(`Failed to get archive: ${error}`);
}
return response;
}
/**
* Upload files to container (tar archive)
*/
export async function putContainerArchive(
containerId: string,
path: string,
tarData: ArrayBuffer | Uint8Array,
envId?: number | null
): Promise<void> {
// Sanitize path
const safePath = path.replace(/[;&|`$(){}[\]<>'"\\]/g, '');
const response = await dockerFetch(
`/containers/${containerId}/archive?path=${encodeURIComponent(safePath)}`,
{
method: 'PUT',
headers: {
'Content-Type': 'application/x-tar'
},
body: tarData as BodyInit
},
envId
);
if (!response.ok) {
const error = await response.text();
throw new Error(`Failed to upload archive: ${error}`);
}
}
/**
* Get stat info for a file/directory in container
*/
export async function statContainerPath(
containerId: string,
path: string,
envId?: number | null
): Promise<{ name: string; size: number; mode: number; mtime: string; linkTarget?: string }> {
// Sanitize path
const safePath = path.replace(/[;&|`$(){}[\]<>'"\\]/g, '');
const response = await dockerFetch(
`/containers/${containerId}/archive?path=${encodeURIComponent(safePath)}`,
{ method: 'HEAD' },
envId
);
if (!response.ok) {
throw new Error(`Path not found: ${safePath}`);
}
// Docker returns stat info in X-Docker-Container-Path-Stat header as base64 JSON
const statHeader = response.headers.get('X-Docker-Container-Path-Stat');
if (!statHeader) {
throw new Error('No stat info returned');
}
const statJson = Buffer.from(statHeader, 'base64').toString('utf-8');
return JSON.parse(statJson);
}
/**
* Read file content from container
* Uses cat command via exec to read file contents
*/
export async function readContainerFile(
containerId: string,
path: string,
envId?: number | null
): Promise<string> {
// Sanitize path to prevent command injection
const safePath = path.replace(/[;&|`$(){}[\]<>'"\\]/g, '');
// Use cat to read file content
const output = await execInContainer(containerId, ['cat', safePath], envId);
return output;
}
/**
* Write file content to container
* Uses Docker archive API to write file
*/
export async function writeContainerFile(
containerId: string,
path: string,
content: string,
envId?: number | null
): Promise<void> {
// Sanitize path
const safePath = path.replace(/[;&|`$(){}[\]<>'"\\]/g, '');
// Get directory and filename
const parts = safePath.split('/');
const filename = parts.pop() || 'file';
const directory = parts.join('/') || '/';
// Create a minimal tar archive with the file
// Tar format: 512-byte header + file content + padding to 512-byte boundary
const contentBytes = new TextEncoder().encode(content);
const fileSize = contentBytes.length;
// Calculate total tar size (header + content + padding + two 512-byte end blocks)
const paddedContentSize = Math.ceil(fileSize / 512) * 512;
const tarSize = 512 + paddedContentSize + 1024; // header + padded content + end blocks
const tarData = new Uint8Array(tarSize);
// Write tar header (512 bytes)
// File name (100 bytes)
const filenameBytes = new TextEncoder().encode(filename);
tarData.set(filenameBytes.slice(0, 100), 0);
// File mode (8 bytes octal) - 0644
tarData.set(new TextEncoder().encode('0000644\0'), 100);
// UID (8 bytes octal) - 0
tarData.set(new TextEncoder().encode('0000000\0'), 108);
// GID (8 bytes octal) - 0
tarData.set(new TextEncoder().encode('0000000\0'), 116);
// File size (12 bytes octal)
const sizeOctal = fileSize.toString(8).padStart(11, '0') + '\0';
tarData.set(new TextEncoder().encode(sizeOctal), 124);
// Mtime (12 bytes octal) - current time
const mtime = Math.floor(Date.now() / 1000).toString(8).padStart(11, '0') + '\0';
tarData.set(new TextEncoder().encode(mtime), 136);
// Checksum placeholder (8 bytes) - filled with spaces initially
tarData.set(new TextEncoder().encode(' '), 148);
// Type flag (1 byte) - '0' for regular file
tarData[156] = 48; // ASCII '0'
// Link name (100 bytes) - empty for regular files
// Already zeros
// USTAR magic (6 bytes) + version (2 bytes)
tarData.set(new TextEncoder().encode('ustar\0'), 257);
tarData.set(new TextEncoder().encode('00'), 263);
// Owner name (32 bytes) - root
tarData.set(new TextEncoder().encode('root'), 265);
// Group name (32 bytes) - root
tarData.set(new TextEncoder().encode('root'), 297);
// Calculate and write checksum
let checksum = 0;
for (let i = 0; i < 512; i++) {
checksum += tarData[i];
}
const checksumOctal = checksum.toString(8).padStart(6, '0') + '\0 ';
tarData.set(new TextEncoder().encode(checksumOctal), 148);
// Write file content after header
tarData.set(contentBytes, 512);
// Upload to container
await putContainerArchive(containerId, directory, tarData, envId);
}
/**
* Create an empty file in container
*/
export async function createContainerFile(
containerId: string,
path: string,
envId?: number | null
): Promise<void> {
// Sanitize path to prevent command injection
const safePath = path.replace(/[;&|`$(){}[\]<>'"\\]/g, '');
// Use touch to create empty file
await execInContainer(containerId, ['touch', safePath], envId);
}
/**
* Create a directory in container
*/
export async function createContainerDirectory(
containerId: string,
path: string,
envId?: number | null
): Promise<void> {
// Sanitize path to prevent command injection
const safePath = path.replace(/[;&|`$(){}[\]<>'"\\]/g, '');
// Use mkdir -p to create directory (and parents if needed)
await execInContainer(containerId, ['mkdir', '-p', safePath], envId);
}
/**
* Delete a file or directory in container
*/
export async function deleteContainerPath(
containerId: string,
path: string,
envId?: number | null
): Promise<void> {
// Sanitize path to prevent command injection
const safePath = path.replace(/[;&|`$(){}[\]<>'"\\]/g, '');
// Safety check: don't allow deleting root or critical paths
const dangerousPaths = ['/', '/bin', '/sbin', '/usr', '/lib', '/lib64', '/etc', '/var', '/root', '/home'];
if (dangerousPaths.includes(safePath) || safePath === '') {
throw new Error('Cannot delete critical system path');
}
// Use rm -rf to delete file or directory
await execInContainer(containerId, ['rm', '-rf', safePath], envId);
}
/**
* Rename/move a file or directory in container
*/
export async function renameContainerPath(
containerId: string,
oldPath: string,
newPath: string,
envId?: number | null
): Promise<void> {
// Sanitize paths to prevent command injection
const safeOldPath = oldPath.replace(/[;&|`$(){}[\]<>'"\\]/g, '');
const safeNewPath = newPath.replace(/[;&|`$(){}[\]<>'"\\]/g, '');
// Use mv to rename
await execInContainer(containerId, ['mv', safeOldPath, safeNewPath], envId);
}
/**
* Change permissions of a file or directory in container
*/
export async function chmodContainerPath(
containerId: string,
path: string,
mode: string,
recursive: boolean = false,
envId?: number | null
): Promise<void> {
// Sanitize path to prevent command injection
const safePath = path.replace(/[;&|`$(){}[\]<>'"\\]/g, '');
// Validate mode (should be octal like 755 or symbolic like u+x)
if (!/^[0-7]{3,4}$/.test(mode) && !/^[ugoa]*[+-=][rwxXst]+$/.test(mode)) {
throw new Error('Invalid chmod mode');
}
// Build command
const cmd = recursive ? ['chmod', '-R', mode, safePath] : ['chmod', mode, safePath];
await execInContainer(containerId, cmd, envId);
}
// Volume browsing and export helpers
const VOLUME_HELPER_IMAGE = 'busybox:latest';
const VOLUME_MOUNT_PATH = '/volume';
const VOLUME_HELPER_TTL_SECONDS = 300; // 5 minutes TTL for helper containers
// Cache for volume helper containers: key = `${volumeName}:${envId ?? 'local'}` -> containerId
const volumeHelperCache = new Map<string, { containerId: string; expiresAt: number }>();
/**
* Get cache key for a volume helper container
*/
function getVolumeCacheKey(volumeName: string, envId?: number | null): string {
return `${volumeName}:${envId ?? 'local'}`;
}
/**
* Ensure the volume helper image (busybox) is available, pulling if necessary
*/
async function ensureVolumeHelperImage(envId?: number | null): Promise<void> {
// Check if image exists
const response = await dockerFetch(`/images/${encodeURIComponent(VOLUME_HELPER_IMAGE)}/json`, {}, envId);
if (response.ok) {
return; // Image exists
}
// Image not found, pull it
console.log(`Pulling ${VOLUME_HELPER_IMAGE} for volume browsing...`);
const pullResponse = await dockerFetch(
`/images/create?fromImage=${encodeURIComponent(VOLUME_HELPER_IMAGE)}`,
{ method: 'POST' },
envId
);
if (!pullResponse.ok) {
const error = await pullResponse.text();
throw new Error(`Failed to pull ${VOLUME_HELPER_IMAGE}: ${error}`);
}
// Wait for pull to complete by consuming the stream
const reader = pullResponse.body?.getReader();
if (reader) {
while (true) {
const { done } = await reader.read();
if (done) break;
}
}
console.log(`Successfully pulled ${VOLUME_HELPER_IMAGE}`);
}
/**
* Check if a container exists and is running
*/
async function isContainerRunning(containerId: string, envId?: number | null): Promise<boolean> {
try {
const response = await dockerFetch(`/containers/${containerId}/json`, {}, envId);
if (!response.ok) return false;
const info = await response.json();
return info.State?.Running === true;
} catch {
return false;
}
}
/**
* Get or create a helper container for volume browsing.
* Reuses existing containers from cache for better performance.
* Returns the container ID.
* @param readOnly - If true, mount volume read-only (default). If false, mount writable.
*/
export async function getOrCreateVolumeHelperContainer(
volumeName: string,
envId?: number | null,
readOnly: boolean = true
): Promise<string> {
// Include readOnly in cache key since we need different containers for ro/rw
const cacheKey = `${getVolumeCacheKey(volumeName, envId)}:${readOnly ? 'ro' : 'rw'}`;
const now = Date.now();
// Check cache for existing container
const cached = volumeHelperCache.get(cacheKey);
if (cached && cached.expiresAt > now) {
// Verify container is still running
if (await isContainerRunning(cached.containerId, envId)) {
// Refresh expiry time on access
cached.expiresAt = now + VOLUME_HELPER_TTL_SECONDS * 1000;
return cached.containerId;
}
// Container no longer running, remove from cache
volumeHelperCache.delete(cacheKey);
}
// Ensure helper image is available (auto-pull if missing)
await ensureVolumeHelperImage(envId);
// Generate a unique container name based on volume name
const safeVolumeName = volumeName.replace(/[^a-zA-Z0-9_.-]/g, '_').slice(0, 50);
const rwSuffix = readOnly ? 'ro' : 'rw';
const containerName = `dockhand-browse-${safeVolumeName}-${rwSuffix}-${Date.now().toString(36)}`;
// Create a temporary container with the volume mounted
const bindMount = readOnly
? `${volumeName}:${VOLUME_MOUNT_PATH}:ro`
: `${volumeName}:${VOLUME_MOUNT_PATH}`;
const containerConfig = {
Image: VOLUME_HELPER_IMAGE,
Cmd: ['sleep', 'infinity'], // Keep alive indefinitely (managed by cache TTL)
HostConfig: {
Binds: [bindMount],
AutoRemove: false
},
Labels: {
'dockhand.volume.helper': 'true',
'dockhand.volume.name': volumeName,
'dockhand.volume.readonly': String(readOnly)
}
};
const response = await dockerJsonRequest<{ Id: string }>(
`/containers/create?name=${encodeURIComponent(containerName)}`,
{
method: 'POST',
body: JSON.stringify(containerConfig)
},
envId
);
const containerId = response.Id;
// Start the container
await dockerFetch(`/containers/${containerId}/start`, { method: 'POST' }, envId);
// Cache the container
volumeHelperCache.set(cacheKey, {
containerId,
expiresAt: now + VOLUME_HELPER_TTL_SECONDS * 1000
});
return containerId;
}
/**
* @deprecated Use getOrCreateVolumeHelperContainer instead
* Create a temporary container with a volume mounted for browsing/export
* Returns the container ID. Caller is responsible for removing the container.
*/
export async function createVolumeHelperContainer(
volumeName: string,
envId?: number | null
): Promise<string> {
return getOrCreateVolumeHelperContainer(volumeName, envId);
}
/**
* Release a cached volume helper container when done browsing.
* This removes the container from cache and stops/removes it from Docker.
* Cleans up both ro and rw variants if they exist.
*/
export async function releaseVolumeHelperContainer(
volumeName: string,
envId?: number | null
): Promise<void> {
const baseCacheKey = getVolumeCacheKey(volumeName, envId);
// Clean up both read-only and read-write variants
for (const suffix of [':ro', ':rw']) {
const cacheKey = baseCacheKey + suffix;
const cached = volumeHelperCache.get(cacheKey);
if (cached) {
volumeHelperCache.delete(cacheKey);
await removeVolumeHelperContainer(cached.containerId, envId).catch(err => {
console.warn('Failed to cleanup volume helper container:', err);
});
}
}
}
/**
* Cleanup expired volume helper containers.
* Called periodically to remove containers that have exceeded their TTL.
*/
export async function cleanupExpiredVolumeHelpers(): Promise<void> {
const now = Date.now();
const expiredEntries: Array<{ key: string; containerId: string; envId?: number | null }> = [];
for (const [key, cached] of volumeHelperCache.entries()) {
if (cached.expiresAt <= now) {
// Parse envId from key: "volumeName:envId" or "volumeName:local"
const [, envIdStr] = key.split(':');
const envId = envIdStr === 'local' ? null : parseInt(envIdStr);
expiredEntries.push({ key, containerId: cached.containerId, envId });
}
}
// Remove from cache and cleanup containers
for (const { key, containerId, envId } of expiredEntries) {
volumeHelperCache.delete(key);
removeVolumeHelperContainer(containerId, envId ?? undefined).catch(err => {
console.warn('Failed to cleanup expired volume helper container:', err);
});
}
if (expiredEntries.length > 0) {
console.log(`Cleaned up ${expiredEntries.length} expired volume helper container(s)`);
}
}
/**
* Remove a volume helper container
*/
export async function removeVolumeHelperContainer(
containerId: string,
envId?: number | null
): Promise<void> {
try {
// Stop the container first (force)
await dockerFetch(`/containers/${containerId}/stop?t=1`, { method: 'POST' }, envId);
} catch {
// Ignore stop errors
}
// Remove the container
await dockerFetch(`/containers/${containerId}?force=true`, { method: 'DELETE' }, envId);
}
/**
* Cleanup all stale volume helper containers on a specific environment.
* Finds containers with label dockhand.volume.helper=true and removes them.
* Called on startup to clean up containers from previous process runs.
*/
async function cleanupStaleVolumeHelpersForEnv(envId?: number | null): Promise<number> {
try {
// Query containers with our helper label
const filters = JSON.stringify({ label: ['dockhand.volume.helper=true'] });
const response = await dockerFetch(
`/containers/json?all=true&filters=${encodeURIComponent(filters)}`,
{},
envId
);
if (!response.ok) {
return 0;
}
const containers: Array<{ Id: string; Names: string[] }> = await response.json();
let removed = 0;
for (const container of containers) {
try {
await removeVolumeHelperContainer(container.Id, envId);
removed++;
} catch (err) {
console.warn(`Failed to remove stale helper container ${container.Names?.[0] || container.Id}:`, err);
}
}
return removed;
} catch (err) {
console.warn('Failed to query stale volume helpers:', err);
return 0;
}
}
/**
* Cleanup stale volume helper containers across all environments.
* Should be called on startup to clean up orphaned containers.
* @param environments - Optional pre-fetched environments (avoids dynamic import in production)
*/
export async function cleanupStaleVolumeHelpers(environments: Array<{ id: number }>): Promise<void> {
console.log('Cleaning up stale volume helper containers...');
if (!environments || environments.length === 0) {
console.log('No environments to clean up');
return;
}
let totalRemoved = 0;
// Clean up all configured environments
for (const env of environments) {
totalRemoved += await cleanupStaleVolumeHelpersForEnv(env.id);
}
if (totalRemoved > 0) {
console.log(`Removed ${totalRemoved} stale volume helper container(s)`);
}
}
/**
* List directory contents in a volume
* Uses cached helper containers for better performance.
*/
export async function listVolumeDirectory(
volumeName: string,
path: string,
envId?: number | null,
readOnly: boolean = true
): Promise<{ path: string; entries: FileEntry[]; containerId: string }> {
const containerId = await getOrCreateVolumeHelperContainer(volumeName, envId, readOnly);
// Sanitize path
const safePath = path.replace(/[;&|`$(){}[\]<>'"\\]/g, '');
const fullPath = `${VOLUME_MOUNT_PATH}${safePath.startsWith('/') ? safePath : '/' + safePath}`;
// Use simple ls since busybox doesn't support --time-style
const output = await execInContainer(containerId, ['ls', '-la', fullPath], envId);
const entries = parseLsOutput(output);
return {
path: safePath || '/',
entries,
containerId
};
// Note: Container is kept alive for reuse. It will be cleaned up
// when the cache TTL expires or when the volume browser modal closes.
}
/**
* Get archive of volume contents for download
* Uses cached helper containers for better performance.
*/
export async function getVolumeArchive(
volumeName: string,
path: string,
envId?: number | null,
readOnly: boolean = true
): Promise<{ response: Response; containerId: string }> {
const containerId = await getOrCreateVolumeHelperContainer(volumeName, envId, readOnly);
// Sanitize path
const safePath = path.replace(/[;&|`$(){}[\]<>'"\\]/g, '');
const fullPath = `${VOLUME_MOUNT_PATH}${safePath.startsWith('/') ? safePath : '/' + safePath}`;
const response = await dockerFetch(
`/containers/${containerId}/archive?path=${encodeURIComponent(fullPath)}`,
{},
envId
);
if (!response.ok) {
const error = await response.text();
throw new Error(`Failed to get archive: ${error}`);
}
return { response, containerId };
// Note: Container is kept alive for reuse. Cache TTL will handle cleanup.
}
/**
* Read file content from volume
* Uses cached helper containers for better performance.
*/
export async function readVolumeFile(
volumeName: string,
path: string,
envId?: number | null,
readOnly: boolean = true
): Promise<string> {
const containerId = await getOrCreateVolumeHelperContainer(volumeName, envId, readOnly);
// Sanitize path
const safePath = path.replace(/[;&|`$(){}[\]<>'"\\]/g, '');
const fullPath = `${VOLUME_MOUNT_PATH}${safePath.startsWith('/') ? safePath : '/' + safePath}`;
// Use cat to read file content
const output = await execInContainer(containerId, ['cat', fullPath], envId);
return output;
// Note: Container is kept alive for reuse. Cache TTL will handle cleanup.
}