Initial commit

This commit is contained in:
2026-04-05 15:28:04 +02:00
commit 0435b3d07d
43 changed files with 4394 additions and 0 deletions

0
.codex Normal file
View File

7
.dockerignore Normal file
View File

@@ -0,0 +1,7 @@
.git
.codex
node_modules
runtime
data
npm-debug.log

3
.gitignore vendored Normal file
View File

@@ -0,0 +1,3 @@
node_modules
runtime
artifacts

347
README.md Normal file
View File

@@ -0,0 +1,347 @@
# Skipper MVP
Skipper is a file-backed control plane and host agent pair for lightweight hosting orchestration.
## Structure
- `skipper-api/`: Express `/v1` control plane API
- `skippy-agent/`: polling agent that executes declarative work orders
- `shared/`: storage, tracing, auth, logging, events, snapshots
- `data/`: JSON-backed resources, work orders, logs, events, and auth records
## Local Run
```bash
docker compose up --build
```
## Skipper CLI
A minimal CLI is now available for common operator tasks.
Run it from the repo with:
```bash
npm run cli -- --help
```
Available commands:
```bash
npm run cli -- health --url http://127.0.0.1:3000
npm run cli -- deploy apply --url http://127.0.0.1:3000 --admin-token YOUR_ADMIN_TOKEN --tenant-id example-tenant
DATA_DIR=/opt/skipper/data npm run cli -- node register --node-id host-1 --role worker --region local
npm run cli -- node show --url http://127.0.0.1:3000 --admin-token YOUR_ADMIN_TOKEN --node-id host-1
```
If you install the repo as an executable package, the binary name is `skipper`.
## Deploy Skipper
The controller can be deployed on its own host from the internal registry.
First-start behavior:
- Skipper now bootstraps its own on-disk data layout on startup.
- An empty bind-mounted data directory is valid.
- You still need to supply an `ADMIN_TOKEN`.
Files:
- [`deploy/skipper/docker-compose.yml`](/home/sundown/Projekter/nodeJS/Skipper/deploy/skipper/docker-compose.yml#L1)
- [`deploy/skipper/.env.example`](/home/sundown/Projekter/nodeJS/Skipper/deploy/skipper/.env.example#L1)
Basic flow on the target server:
```bash
mkdir -p /opt/skipper
cd /opt/skipper
cp /path/to/repo/deploy/skipper/docker-compose.yml .
cp /path/to/repo/deploy/skipper/.env.example .env
```
Edit `.env`:
- set `SKIPPER_IMAGE` to the tag you released
- set a strong `ADMIN_TOKEN`
- optionally change `SKIPPER_PORT`
Start Skipper:
```bash
docker login registry.internal.budgethost.io
docker compose up -d
```
One-line first boot with `docker run`:
```bash
docker run -d \
--name skipper-api \
--restart unless-stopped \
-p 3000:3000 \
-e HOST=0.0.0.0 \
-e PORT=3000 \
-e DATA_DIR=/app/data \
-e ADMIN_TOKEN='replace-with-a-long-random-admin-token' \
-v /opt/skipper/data:/app/data \
registry.internal.budgethost.io/skipper/skipper-api:latest
```
That command is enough for first boot. The container will create the required `/app/data` subdirectories automatically.
Verify:
```bash
docker compose ps
curl http://127.0.0.1:3000/v1/health
```
Persisted controller state will live in:
```bash
/opt/skipper/data
```
The next step after this is deploying one or more `Skippy` agents pointed at the controller URL.
## Attach Skippy
Attaching a `Skippy` agent has two parts:
1. register the node on the Skipper host
2. start the agent container on the target host
### 1. Register The Node On Skipper
Run this on the Skipper host, in the same repo or deployment workspace that owns `/opt/skipper/data`:
```bash
DATA_DIR=/opt/skipper/data npm run cli -- node register --node-id host-1 --role worker --region se-sto-1
```
That writes:
- `/opt/skipper/data/resources/nodes/host-1.json`
- `/opt/skipper/data/auth/nodes/host-1.json`
The command prints the generated token. Save it and use it as `AGENT_TOKEN` on the target host.
You can also provide your own token:
```bash
DATA_DIR=/opt/skipper/data npm run cli -- node register --node-id host-1 --token 'your-long-random-token'
```
### 2. Start The Agent On The Target Host
Files:
- [`deploy/skippy/docker-compose.yml`](/home/sundown/Projekter/nodeJS/Skipper/deploy/skippy/docker-compose.yml#L1)
- [`deploy/skippy/.env.example`](/home/sundown/Projekter/nodeJS/Skipper/deploy/skippy/.env.example#L1)
Basic flow on the target host:
```bash
mkdir -p /opt/skippy
cd /opt/skippy
cp /path/to/repo/deploy/skippy/docker-compose.yml .
cp /path/to/repo/deploy/skippy/.env.example .env
```
Edit `.env`:
- set `SKIPPY_IMAGE`
- set `SKIPPER_URL`
- set `AGENT_ID`
- set `AGENT_TOKEN`
Start the agent:
```bash
docker login registry.internal.budgethost.io
docker compose up -d
```
One-line first boot with `docker run`:
```bash
docker run -d \
--name skippy-agent \
--restart unless-stopped \
-e DATA_DIR=/app/data \
-e SKIPPER_URL='http://your-skipper-host:3000' \
-e AGENT_ID='host-1' \
-e AGENT_TOKEN='replace-with-generated-token' \
-e POLL_INTERVAL_MS=5000 \
-e HEARTBEAT_INTERVAL_MS=15000 \
-e SKIPPY_COMPOSE_BASE_DIR=/opt/skipper/tenants \
-v /var/run/docker.sock:/var/run/docker.sock \
-v /opt/skippy/data:/app/data \
-v /opt/skippy/tenants:/opt/skipper/tenants \
registry.internal.budgethost.io/skipper/skippy-agent:latest
```
Verify:
```bash
docker logs skippy-agent --tail 50
curl -H "x-admin-token: <your-admin-token>" http://your-skipper-host:3000/v1/resources/node/host-1
```
## Image Build And Push
Default image targets:
- `registry.internal.budgethost.io/skipper/skipper-api`
- `registry.internal.budgethost.io/skipper/skippy-agent`
Prerequisites:
- Docker daemon access must work from your shell
- You must be logged into `registry.internal.budgethost.io`
Sanity checks:
```bash
docker info
docker login registry.internal.budgethost.io
```
Preview tags:
```bash
npm run images:print
```
Build locally:
```bash
IMAGE_TAG=0.1.0 npm run images:build
```
Build and push to the internal registry:
```bash
IMAGE_TAG=0.1.0 npm run images:release
```
Push prebuilt tags only:
```bash
IMAGE_TAG=0.1.0 npm run images:push
```
Optional overrides:
- `IMAGE_REGISTRY`
- `IMAGE_NAMESPACE`
- `IMAGE_TAG`
- `IMAGE_PLATFORM`
Run compose from registry-pushed images instead of local builds:
```bash
IMAGE_TAG=0.1.0 docker compose -f docker-compose.yml -f docker-compose.registry.yml up -d
```
First release sequence:
```bash
npm run smoke:test
docker login registry.internal.budgethost.io
IMAGE_TAG=0.1.0 npm run images:release
IMAGE_TAG=0.1.0 docker compose -f docker-compose.yml -f docker-compose.registry.yml up -d
```
Verification:
```bash
docker compose -f docker-compose.yml -f docker-compose.registry.yml ps
curl http://localhost:3000/v1/health
```
Current note:
- The application smoke test passed in this repository.
- The image release flow was prepared and invoked, but this environment could not access `/var/run/docker.sock`, so the actual build and push must be run from a host session with Docker daemon permissions.
## Release Manifest
A release-oriented wrapper is available to generate git-aware image tags and write a deployment manifest.
Print the next release plan:
```bash
npm run release:print
```
Write manifest files without building:
```bash
npm run release:plan
```
Build release-tagged images and write manifest:
```bash
RUN_SMOKE_TEST=1 npm run release:build
```
Build, push, and write manifest:
```bash
RUN_SMOKE_TEST=1 npm run release:publish
```
Generated manifest files:
- `artifacts/releases/latest.json`
- `artifacts/releases/<version_tag>.json`
Release environment overrides:
- `RELEASE_CHANNEL`
- `RELEASE_VERSION_TAG`
- `RELEASE_GIT_SHA`
- `IMAGE_REGISTRY`
- `IMAGE_NAMESPACE`
Tag behavior:
- If git metadata is available, the release tag includes the short commit SHA.
- If git metadata is not available, the release tag falls back to `nogit`.
- Unless overridden, the generated tag format is `<channel>-<git_sha>-<timestamp>`.
## Smoke Test
```bash
npm run smoke:test
```
This starts the API and agent as local subprocesses, uses a mocked `docker` binary, triggers one deploy, and verifies the completed job plus written compose artifacts.
Create a deployment work order:
```bash
curl -X POST \
-H 'x-admin-token: dev-admin-token' \
-H 'x-idempotency-key: local-run-1' \
-H 'x-request-id: 11111111-1111-1111-1111-111111111111' \
-H 'x-correlation-id: 22222222-2222-2222-2222-222222222222' \
http://localhost:3000/v1/deployments/example-tenant/apply
```
Inspect work orders:
```bash
find data/work-orders -type f | sort
```
## Example Flow
1. `POST /v1/deployments/example-tenant/apply` loads [`data/resources/tenants/example-tenant.json`](/home/sundown/Projekter/nodeJS/Skipper/data/resources/tenants/example-tenant.json#L1) and creates a declarative `deploy_service` work order in `data/work-orders/pending/`.
2. `skippy-agent` polls `GET /v1/nodes/host-1/work-orders/next` every 5 seconds and receives the next work order for `host-1`.
3. The agent writes the compose file to `/opt/skipper/tenants/example-tenant/docker-compose.yml`, writes `.env`, and runs `docker compose up -d`.
4. The agent sends `POST /v1/work-orders/:id/result` with a structured result object plus state updates.
5. `skipper-api` moves the finished work order into `data/work-orders/finished/`, updates resource state, writes structured logs, and emits events.

View File

@@ -0,0 +1,6 @@
{
"node_id": "host-1",
"token": "dev-node-token",
"schema_version": "v1",
"updated_at": "2026-04-05T00:00:00.000Z"
}

View File

@@ -0,0 +1,16 @@
{
"id": "host-1",
"resource_type": "node",
"schema_version": "v1",
"desired_state": {
"enabled": true,
"labels": {
"role": "local"
}
},
"current_state": {},
"last_applied_state": {},
"metadata": {},
"created_at": "2026-04-05T00:00:00.000Z",
"updated_at": "2026-04-05T00:00:00.000Z"
}

View File

@@ -0,0 +1,18 @@
{
"id": "service-web",
"resource_type": "service",
"schema_version": "v1",
"desired_state": {
"tenant_id": "example-tenant",
"service_kind": "nginx",
"image": "nginx:alpine",
"networks": [],
"volumes": [],
"resource_limits": null
},
"current_state": {},
"last_applied_state": {},
"metadata": {},
"created_at": "2026-04-05T00:00:00.000Z",
"updated_at": "2026-04-05T00:00:00.000Z"
}

View File

@@ -0,0 +1,26 @@
{
"id": "example-tenant",
"resource_type": "tenant",
"schema_version": "v1",
"desired_state": {
"display_name": "Example Tenant",
"deployment_policy": {
"target_node_id": "host-1"
},
"service_ids": [
"service-web"
],
"compose": {
"tenant_id": "example-tenant",
"compose_file": "services:\n web:\n image: nginx:alpine\n restart: unless-stopped\n ports:\n - \"${NGINX_PORT}:80\"\n",
"env": {
"NGINX_PORT": "8081"
}
}
},
"current_state": {},
"last_applied_state": {},
"metadata": {},
"created_at": "2026-04-05T00:00:00.000Z",
"updated_at": "2026-04-05T00:00:00.000Z"
}

View File

@@ -0,0 +1,4 @@
SKIPPER_IMAGE=registry.internal.budgethost.io/skipper/skipper-api:latest
ADMIN_TOKEN=replace-with-a-long-random-admin-token
SKIPPER_PORT=3000

View File

@@ -0,0 +1,15 @@
services:
skipper-api:
image: ${SKIPPER_IMAGE}
container_name: skipper-api
restart: unless-stopped
environment:
DATA_DIR: /app/data
PORT: 3000
HOST: 0.0.0.0
ADMIN_TOKEN: ${ADMIN_TOKEN}
ports:
- "${SKIPPER_PORT:-3000}:3000"
volumes:
- ./data:/app/data

View File

@@ -0,0 +1,7 @@
SKIPPY_IMAGE=registry.internal.budgethost.io/skipper/skippy-agent:latest
SKIPPER_URL=http://your-skipper-host:3000
AGENT_ID=host-1
AGENT_TOKEN=replace-with-the-token-generated-on-skipper
POLL_INTERVAL_MS=5000
HEARTBEAT_INTERVAL_MS=15000

View File

@@ -0,0 +1,18 @@
services:
skippy-agent:
image: ${SKIPPY_IMAGE}
container_name: skippy-agent
restart: unless-stopped
environment:
DATA_DIR: /app/data
SKIPPER_URL: ${SKIPPER_URL}
AGENT_ID: ${AGENT_ID}
AGENT_TOKEN: ${AGENT_TOKEN}
POLL_INTERVAL_MS: ${POLL_INTERVAL_MS:-5000}
HEARTBEAT_INTERVAL_MS: ${HEARTBEAT_INTERVAL_MS:-15000}
SKIPPY_COMPOSE_BASE_DIR: /opt/skipper/tenants
volumes:
- /var/run/docker.sock:/var/run/docker.sock
- ./data:/app/data
- ./tenants:/opt/skipper/tenants

View File

@@ -0,0 +1,8 @@
services:
skipper-api:
image: ${IMAGE_REGISTRY:-registry.internal.budgethost.io}/${IMAGE_NAMESPACE:-skipper}/skipper-api:${IMAGE_TAG:-latest}
build: null
skippy-agent:
image: ${IMAGE_REGISTRY:-registry.internal.budgethost.io}/${IMAGE_NAMESPACE:-skipper}/skippy-agent:${IMAGE_TAG:-latest}
build: null

34
docker-compose.yml Normal file
View File

@@ -0,0 +1,34 @@
services:
skipper-api:
build:
context: .
dockerfile: skipper-api/Dockerfile
container_name: skipper-api
environment:
DATA_DIR: /app/data
PORT: 3000
ADMIN_TOKEN: dev-admin-token
ports:
- "3000:3000"
volumes:
- ./data:/app/data
skippy-agent:
build:
context: .
dockerfile: skippy-agent/Dockerfile
container_name: skippy-agent
depends_on:
- skipper-api
environment:
DATA_DIR: /app/data
SKIPPER_URL: http://skipper-api:3000
AGENT_ID: host-1
AGENT_TOKEN: dev-node-token
POLL_INTERVAL_MS: 5000
HEARTBEAT_INTERVAL_MS: 15000
SKIPPY_COMPOSE_BASE_DIR: /opt/skipper/tenants
volumes:
- ./data:/app/data
- /var/run/docker.sock:/var/run/docker.sock
- ./runtime/agent-tenants:/opt/skipper/tenants

605
docs/architecture.md Normal file
View File

@@ -0,0 +1,605 @@
# Skipper Architecture
## Purpose
Skipper is a lightweight hosting orchestration system built for clarity, inspectability, and future AI-assisted operations.
- `Skipper` = control plane
- `Skippy` = host agent
- Communication model = HTTPS pull from agent to controller
- Implementation language = Node.js only
- Persistence = file-based JSON storage only
- Deployment target = Docker for all components
This document now reflects the current implementation in this repository, while also marking the remaining gaps to the broader target architecture.
## Design Principles
- No hidden state
- No implicit resource relationships
- No shell-script driven control flow as the primary orchestration model
- All operations must be idempotent and safe to retry
- All state transitions must be inspectable via API or persisted state
- Logs, events, and state must be sufficient for debugging
- Schemas must be explicit and versioned
- Extensibility is preferred over short-term convenience
- Observability is a first-class requirement
## System Topology
### Skipper
Skipper is the controller and source of truth for desired state.
Current responsibilities:
- Store resource definitions on disk
- Store `desired_state`, `current_state`, and `last_applied_state`
- Create declarative work orders targeted at nodes
- Accept structured work-order results and state reports
- Persist structured logs, events, idempotency records, and snapshots
- Expose a versioned REST API under `/v1`
### Skippy
Skippy is a node-local reconciliation agent.
Current responsibilities:
- Authenticate to Skipper with a node token
- Poll for work orders over HTTPS
- Apply desired state locally through Node.js modules
- Report structured results
- Report updated resource state
- Emit structured JSON logs locally while also driving persisted state changes through the API
## Communication Model
Communication is LAN-first and HTTPS-oriented.
- Agents initiate control-plane communication
- Skipper does not require inbound connectivity to managed nodes
- All implemented API endpoints are versioned under `/v1`
- All implemented requests and responses use JSON
- Authentication is token-based
- Request tracing is propagated with `request_id` and `correlation_id`
The current local development stack uses plain HTTP inside Docker and during smoke tests. The architecture remains HTTPS-first for production deployment.
## API Contract
### Versioning
All implemented control-plane endpoints live under `/v1`.
Implemented endpoints:
- `GET /v1/health`
- `GET /v1/resources`
- `GET /v1/resources/:resourceType/:resourceId`
- `GET /v1/work-orders`
- `GET /v1/work-orders/:workOrderId`
- `GET /v1/nodes/:nodeId/work-orders/next`
- `POST /v1/nodes/:nodeId/heartbeat`
- `POST /v1/work-orders/:workOrderId/result`
- `POST /v1/deployments/:tenantId/apply`
- `GET /v1/snapshots/system/latest`
- `GET /v1/snapshots/tenants/:tenantId/latest`
A compatibility `GET /health` endpoint also exists for simple health checks.
### Request Metadata
Every request is processed with:
- `request_id`
- `correlation_id`
These are accepted from:
- `x-request-id`
- `x-correlation-id`
If absent, Skipper generates them and returns them in both response headers and the response body envelope.
### Response Envelope
All API responses use a stable envelope:
```json
{
"schema_version": "v1",
"request_id": "6c4a5b1f-7f91-42cc-aef5-5ea4248fb2e8",
"correlation_id": "a0f84ecf-f8d6-4c4e-97a3-0ed68eb9c95d",
"data": {},
"error": null,
"metadata": {
"timestamp": "2026-04-05T12:00:00.000Z"
}
}
```
Error responses use the same envelope with `data: null`.
### Authentication
Two auth modes are currently implemented:
- admin API requests use `x-admin-token`
- node API requests use `Authorization: Bearer <node-token>`
Node tokens are stored in:
- `/data/auth/nodes/<node_id>.json`
### Idempotency
Idempotency is currently implemented for deployment apply requests through `x-idempotency-key`.
Persisted idempotency records live under:
- `/data/idempotency`
The main implemented idempotent flow is:
- `POST /v1/deployments/:tenantId/apply`
## Resource Model
All managed resources are explicit JSON documents.
Each resource document contains:
- `id`
- `resource_type`
- `schema_version`
- `desired_state`
- `current_state`
- `last_applied_state`
- `metadata`
- `created_at`
- `updated_at`
The three-state model is implemented and central:
- `desired_state`: what Skipper wants
- `current_state`: what Skippy or Skipper currently knows to be true
- `last_applied_state`: what Skippy most recently attempted or enforced
### Implemented Resource Types
The storage layer currently supports these resource types:
- `tenant`
- `node`
- `service`
- `deployment`
- `resource_limits`
- `network`
- `volume`
At the moment, the repository ships example data for:
- `tenant`
- `node`
- `service`
`deployment` resources are created dynamically when a deployment apply request is issued.
### Tenant
Current tenant usage:
- deployment target policy
- service references
- compose project specification
Example:
```json
{
"id": "example-tenant",
"resource_type": "tenant",
"schema_version": "v1",
"desired_state": {
"display_name": "Example Tenant",
"deployment_policy": {
"target_node_id": "host-1"
},
"service_ids": ["service-web"],
"compose": {
"tenant_id": "example-tenant",
"compose_file": "services:\n web:\n image: nginx:alpine\n",
"env": {
"NGINX_PORT": "8081"
}
}
},
"current_state": {},
"last_applied_state": {},
"metadata": {}
}
```
### Node
Current node usage:
- desired enablement and labels
- heartbeat status
- agent capabilities
- agent version
### Service
Current service usage:
- tenant ownership
- service kind
- image
- network and volume references
- resource limit reference
### Deployment
Current deployment usage:
- created during `POST /v1/deployments/:tenantId/apply`
- tracks deployment status
- tracks associated work order
- stores deployment-oriented desired state
## File-Based Persistence Layout
The current on-disk layout is:
```text
/data
/resources
/tenants
/nodes
/services
/deployments
/resource-limits
/networks
/volumes
/work-orders
/pending
/running
/finished
/events
/YYYY-MM-DD
/logs
/YYYY-MM-DD
/snapshots
/system
/tenants
/idempotency
/auth
/nodes
```
Rules implemented today:
- one JSON document per state file
- atomic JSON writes
- append-only event and log history
- stable file names derived from resource or work-order IDs
## Work Order Model
Skipper does not send direct commands. It issues declarative work orders.
### Work Order Schema
The implemented work-order model is:
```json
{
"id": "4b9f5e2a-cf65-4342-97f5-66f3fe5a54f7",
"resource_type": "work_order",
"schema_version": "v1",
"type": "deploy_service",
"target": {
"tenant_id": "example-tenant",
"node_id": "host-1"
},
"desired_state": {
"deployment_id": "deployment-123",
"tenant_id": "example-tenant",
"service_ids": ["service-web"],
"compose_project": {}
},
"status": "pending",
"result": null,
"request_id": "6c4a5b1f-7f91-42cc-aef5-5ea4248fb2e8",
"correlation_id": "a0f84ecf-f8d6-4c4e-97a3-0ed68eb9c95d",
"created_at": "2026-04-05T12:00:00.000Z",
"started_at": null,
"finished_at": null,
"metadata": {}
}
```
### Status Values
Implemented status values:
- `pending`
- `running`
- `success`
- `failed`
### Implemented Work Order Type
The current code implements:
- `deploy_service`
This currently reconciles a tenant compose project by:
1. writing `docker-compose.yml`
2. writing `.env`
3. running `docker compose up -d`
4. reporting structured output and state changes
### Work Order Result Schema
Work-order results are structured JSON, not free-form text:
```json
{
"success": true,
"code": "APPLY_OK",
"message": "Desired state applied",
"details": {
"duration_ms": 8,
"compose_path": "/opt/skipper/tenants/example-tenant/docker-compose.yml",
"changed_resources": ["service-web"],
"unchanged_resources": [],
"command": {
"program": "docker",
"args": ["compose", "-f", "...", "up", "-d"],
"exit_code": 0,
"stdout": "...",
"stderr": ""
}
}
}
```
This keeps results machine-readable while still preserving execution detail.
## Reconciliation Flow
The currently implemented flow is:
1. Admin calls `POST /v1/deployments/:tenantId/apply`
2. Skipper loads the tenant resource
3. Skipper creates a deployment resource
4. Skipper creates a `deploy_service` work order targeted at the node in tenant desired state
5. Skippy sends heartbeat and polls `GET /v1/nodes/:nodeId/work-orders/next`
6. Skippy claims and executes the work order
7. Skippy reports `POST /v1/work-orders/:workOrderId/result`
8. Skipper finishes the work order, updates resource state, writes events, and keeps logs/snapshots available
### Retry Safety
Current retry safety measures:
- deployment apply is idempotent through `x-idempotency-key`
- work-order execution is state-based and convergent at the compose level
- work-order completion is safe against duplicate result submission
- filesystem writes are atomic
## Structured Logging
All implemented operational logs are structured JSON.
Each log entry includes:
- `timestamp`
- `level`
- `service`
- `node_id`
- `tenant_id`
- `request_id`
- `correlation_id`
- `action`
- `result`
- `metadata`
Logs are written to:
- `/data/logs/YYYY-MM-DD/<service>.ndjson`
The logger also redacts common secret-shaped keys such as `token`, `secret`, `password`, and `authorization`.
## Event System
Every important state transition in the current flow emits an event.
Implemented event storage:
- `/data/events/YYYY-MM-DD/<timestamp>-<event_id>.json`
Implemented event types in the current code path:
- `resource_created`
- `work_order_created`
- `work_order_started`
- `work_order_succeeded`
- `work_order_failed`
- `deployment_started`
- `deployment_succeeded`
- `deployment_failed`
- `node_heartbeat_received`
- `snapshot_created`
The broader architecture still expects additional event coverage for all future resource mutations.
## State Snapshots
Snapshots are implemented and persisted as JSON documents.
Supported snapshot scopes:
- system
- per-tenant
Implemented endpoints:
- `GET /v1/snapshots/system/latest`
- `GET /v1/snapshots/tenants/:tenantId/latest`
Each snapshot includes:
- `snapshot_id`
- `scope`
- `created_at`
- `request_id`
- `correlation_id`
- `resources`
- `diffs`
Snapshot files are currently stored as:
- `/data/snapshots/system/latest.json`
- `/data/snapshots/tenants/<tenant_id>.json`
## Observability and AI Readiness
The current implementation is AI-ready at the core workflow level because it now preserves:
- request-level tracing across API and agent boundaries
- structured work-order lifecycle data
- historical logs
- historical events
- explicit desired/current/last-applied state
- exportable JSON/NDJSON persistence
For the implemented deployment path, the system can answer:
- what changed
- which work order applied it
- which node applied it
- what desired state was targeted
- what current and last applied state were recorded
## Error Handling
All API errors are structured and envelope-wrapped.
Implemented error shape:
```json
{
"code": "RESOURCE_NOT_FOUND",
"message": "Tenant not found",
"details": {
"resource_type": "tenant",
"resource_id": "example-tenant"
}
}
```
Implemented machine-readable error codes include:
- `INVALID_REQUEST`
- `UNAUTHORIZED`
- `RESOURCE_NOT_FOUND`
- `WORK_ORDER_NOT_CLAIMABLE`
- `INTERNAL_ERROR`
Raw stack traces are not returned in API responses.
## Security Model
### Implemented
- node token authentication
- admin token authentication
- correlation-aware structured logging
- redaction of common secret-shaped log fields
### Not Yet Implemented
- role-based authorization
- secret rotation workflows
- mTLS
- per-resource authorization policies
## Extensibility Model
The code is currently structured so new resource types and work-order types can be added without replacing the whole control flow.
Current extensibility anchors:
- resource storage by `resource_type`
- work-order execution by `type`
- stable response envelope
- versioned schemas
- shared storage and telemetry modules in `/shared`
## Implemented Internal Modules
### Shared
- [`shared/context.js`](/home/sundown/Projekter/nodeJS/Skipper/shared/context.js)
- [`shared/errors.js`](/home/sundown/Projekter/nodeJS/Skipper/shared/errors.js)
- [`shared/auth.js`](/home/sundown/Projekter/nodeJS/Skipper/shared/auth.js)
- [`shared/resources.js`](/home/sundown/Projekter/nodeJS/Skipper/shared/resources.js)
- [`shared/work-orders.js`](/home/sundown/Projekter/nodeJS/Skipper/shared/work-orders.js)
- [`shared/logs.js`](/home/sundown/Projekter/nodeJS/Skipper/shared/logs.js)
- [`shared/events.js`](/home/sundown/Projekter/nodeJS/Skipper/shared/events.js)
- [`shared/idempotency.js`](/home/sundown/Projekter/nodeJS/Skipper/shared/idempotency.js)
- [`shared/snapshots.js`](/home/sundown/Projekter/nodeJS/Skipper/shared/snapshots.js)
### Skipper API
- [`skipper-api/src/index.js`](/home/sundown/Projekter/nodeJS/Skipper/skipper-api/src/index.js)
### Skippy Agent
- [`skippy-agent/src/index.js`](/home/sundown/Projekter/nodeJS/Skipper/skippy-agent/src/index.js)
- [`skippy-agent/src/lib/http.js`](/home/sundown/Projekter/nodeJS/Skipper/skippy-agent/src/lib/http.js)
- [`skippy-agent/src/modules/docker.js`](/home/sundown/Projekter/nodeJS/Skipper/skippy-agent/src/modules/docker.js)
## Current Gaps
The code is now aligned with the architecture for the core deployment path, but it is not feature-complete across the full long-term vision.
Not yet implemented:
- full CRUD APIs for all resource types
- generic reconciliation across all future services
- `resource_updated` and `desired_state_changed` event coverage for every mutation path
- persisted state reports for all future resource kinds
- richer diffing beyond snapshot-level desired/current comparisons
- RBAC and richer authorization
- production HTTPS termination inside the app itself
- additional work-order types such as restart, migrate, nginx management, mysql provisioning, and systemd integration
## Current Compliance Summary
Implemented and aligned:
- `/v1` API contract
- request and correlation ID propagation
- envelope-based responses
- structured errors
- declarative work orders
- three-state resource model
- structured JSON logging
- event persistence
- snapshot persistence
- idempotent deployment apply
- token-based controller/agent auth
Still incomplete relative to the full target:
- broader resource coverage
- broader reconciliation coverage
- broader auth model
- full event coverage for every possible state mutation

22
package.json Normal file
View File

@@ -0,0 +1,22 @@
{
"name": "skipper",
"version": "1.0.0",
"private": true,
"bin": {
"skipper": "./scripts/skipper.js"
},
"scripts": {
"images:build": "node scripts/images.js build",
"images:push": "node scripts/images.js push",
"images:release": "node scripts/images.js release",
"images:print": "node scripts/images.js print",
"cli": "node scripts/skipper.js",
"node:register": "node scripts/register-node.js",
"release:print": "node scripts/release.js print",
"release:plan": "node scripts/release.js plan",
"release:build": "node scripts/release.js build",
"release:publish": "node scripts/release.js publish",
"smoke:test": "node scripts/smoke-test.js",
"test": "node scripts/smoke-test.js"
}
}

129
scripts/images.js Normal file
View File

@@ -0,0 +1,129 @@
const { spawn } = require('child_process');
const path = require('path');
const rootDir = path.resolve(__dirname, '..');
const config = {
registry: process.env.IMAGE_REGISTRY || 'registry.internal.budgethost.io',
namespace: process.env.IMAGE_NAMESPACE || 'skipper',
tag: process.env.IMAGE_TAG || 'latest',
platform: process.env.IMAGE_PLATFORM || '',
images: [
{
id: 'skipper-api',
context: '.',
dockerfile: 'skipper-api/Dockerfile',
repository: 'skipper-api',
},
{
id: 'skippy-agent',
context: '.',
dockerfile: 'skippy-agent/Dockerfile',
repository: 'skippy-agent',
},
],
};
function getAction() {
const action = process.argv[2] || 'print';
if (!['print', 'build', 'push', 'release'].includes(action)) {
throw new Error(`Unsupported action: ${action}`);
}
return action;
}
function imageRef(image, tag) {
return `${config.registry}/${config.namespace}/${image.repository}:${tag}`;
}
function spawnCommand(command, args) {
return new Promise((resolve, reject) => {
const child = spawn(command, args, {
cwd: rootDir,
stdio: 'inherit',
});
child.on('error', reject);
child.on('close', (code) => {
if (code === 0) {
resolve();
return;
}
reject(new Error(`${command} ${args.join(' ')} failed with exit code ${code}`));
});
});
}
async function buildImage(image) {
const tags = [imageRef(image, config.tag)];
if (config.tag !== 'latest') {
tags.push(imageRef(image, 'latest'));
}
const args = ['build', '-f', image.dockerfile];
if (config.platform) {
args.push('--platform', config.platform);
}
for (const tag of tags) {
args.push('-t', tag);
}
args.push(image.context);
console.log(`Building ${image.id}`);
await spawnCommand('docker', args);
}
async function pushImage(image) {
const tags = [config.tag];
if (config.tag !== 'latest') {
tags.push('latest');
}
for (const tag of tags) {
const ref = imageRef(image, tag);
console.log(`Pushing ${ref}`);
await spawnCommand('docker', ['push', ref]);
}
}
function printPlan() {
for (const image of config.images) {
console.log(`${image.id}: ${imageRef(image, config.tag)}`);
if (config.tag !== 'latest') {
console.log(`${image.id}: ${imageRef(image, 'latest')}`);
}
}
}
async function main() {
const action = getAction();
if (action === 'print') {
printPlan();
return;
}
for (const image of config.images) {
if (action === 'build' || action === 'release') {
await buildImage(image);
}
if (action === 'push' || action === 'release') {
await pushImage(image);
}
}
}
main().catch((error) => {
console.error(error.message);
process.exit(1);
});

86
scripts/register-node.js Normal file
View File

@@ -0,0 +1,86 @@
const path = require('path');
const { randomBytes } = require('crypto');
const { writeJson } = require('../shared/fs');
const { dataDir } = require('../shared/paths');
function getArg(flag, fallback) {
const index = process.argv.indexOf(flag);
if (index === -1 || index + 1 >= process.argv.length) {
return fallback;
}
return process.argv[index + 1];
}
function nowIso() {
return new Date().toISOString();
}
function randomTokenHex(bytes) {
return randomBytes(bytes).toString('hex');
}
async function registerNode({ nodeId, role, region, token }) {
const createdAt = nowIso();
const nodeFile = path.join(dataDir, 'resources', 'nodes', `${nodeId}.json`);
const tokenFile = path.join(dataDir, 'auth', 'nodes', `${nodeId}.json`);
await writeJson(nodeFile, {
id: nodeId,
resource_type: 'node',
schema_version: 'v1',
desired_state: {
enabled: true,
labels: {
role,
region,
},
},
current_state: {},
last_applied_state: {},
metadata: {},
created_at: createdAt,
updated_at: createdAt,
});
await writeJson(tokenFile, {
node_id: nodeId,
token,
schema_version: 'v1',
updated_at: createdAt,
});
return {
node_id: nodeId,
token,
node_file: nodeFile,
token_file: tokenFile,
};
}
async function main() {
const nodeId = getArg('--node-id', process.env.NODE_ID);
const role = getArg('--role', process.env.NODE_ROLE || 'worker');
const region = getArg('--region', process.env.NODE_REGION || 'default');
const token = getArg('--token', process.env.NODE_TOKEN || randomTokenHex(32));
if (!nodeId) {
throw new Error('Missing node id. Use --node-id <id> or NODE_ID=<id>.');
}
const result = await registerNode({ nodeId, role, region, token });
process.stdout.write(`${JSON.stringify(result, null, 2)}\n`);
}
if (require.main === module) {
main().catch((error) => {
process.stderr.write(`${error.message}\n`);
process.exit(1);
});
}
module.exports = {
registerNode,
randomTokenHex,
};

204
scripts/release.js Normal file
View File

@@ -0,0 +1,204 @@
const fs = require('fs/promises');
const path = require('path');
const { spawn } = require('child_process');
const rootDir = path.resolve(__dirname, '..');
const artifactsDir = path.join(rootDir, 'artifacts', 'releases');
const registry = process.env.IMAGE_REGISTRY || 'registry.internal.budgethost.io';
const namespace = process.env.IMAGE_NAMESPACE || 'skipper';
const channel = process.env.RELEASE_CHANNEL || 'stable';
function getAction() {
const action = process.argv[2] || 'print';
if (!['print', 'plan', 'build', 'publish'].includes(action)) {
throw new Error(`Unsupported action: ${action}`);
}
return action;
}
function timestampUtc() {
return new Date().toISOString();
}
function compactTimestamp(value) {
return value.replace(/[-:]/g, '').replace(/\.\d+Z$/, 'Z');
}
function sanitizeTagPart(value) {
return String(value || '')
.toLowerCase()
.replace(/[^a-z0-9._-]+/g, '-')
.replace(/^-+|-+$/g, '') || 'unknown';
}
function spawnCapture(command, args) {
return new Promise((resolve, reject) => {
const child = spawn(command, args, {
cwd: rootDir,
stdio: ['ignore', 'pipe', 'pipe'],
});
let stdout = '';
let stderr = '';
child.stdout.on('data', (chunk) => {
stdout += chunk.toString();
});
child.stderr.on('data', (chunk) => {
stderr += chunk.toString();
});
child.on('error', reject);
child.on('close', (code) => {
if (code === 0) {
resolve({ stdout: stdout.trim(), stderr: stderr.trim() });
return;
}
reject(new Error(stderr.trim() || `${command} ${args.join(' ')} failed with exit code ${code}`));
});
});
}
function spawnInherited(command, args, env) {
return new Promise((resolve, reject) => {
const child = spawn(command, args, {
cwd: rootDir,
stdio: 'inherit',
env: {
...process.env,
...env,
},
});
child.on('error', reject);
child.on('close', (code) => {
if (code === 0) {
resolve();
return;
}
reject(new Error(`${command} ${args.join(' ')} failed with exit code ${code}`));
});
});
}
async function detectGitSha() {
try {
const result = await spawnCapture('git', ['rev-parse', '--short', 'HEAD']);
return result.stdout || null;
} catch (error) {
return null;
}
}
async function buildReleaseMetadata() {
const createdAt = timestampUtc();
const gitSha = sanitizeTagPart(process.env.RELEASE_GIT_SHA || (await detectGitSha()) || 'nogit');
const versionTag = sanitizeTagPart(
process.env.RELEASE_VERSION_TAG || `${channel}-${gitSha}-${compactTimestamp(createdAt)}`
);
const images = [
{
id: 'skipper-api',
repository: 'skipper-api',
image: `${registry}/${namespace}/skipper-api:${versionTag}`,
latest_image: `${registry}/${namespace}/skipper-api:latest`,
},
{
id: 'skippy-agent',
repository: 'skippy-agent',
image: `${registry}/${namespace}/skippy-agent:${versionTag}`,
latest_image: `${registry}/${namespace}/skippy-agent:latest`,
},
];
return {
schema_version: 'v1',
created_at: createdAt,
registry,
namespace,
channel,
git_sha: gitSha,
version_tag: versionTag,
images,
compose: {
files: ['docker-compose.yml', 'docker-compose.registry.yml'],
image_tag: versionTag,
},
};
}
async function writeManifest(metadata) {
await fs.mkdir(artifactsDir, { recursive: true });
const versionPath = path.join(artifactsDir, `${metadata.version_tag}.json`);
const latestPath = path.join(artifactsDir, 'latest.json');
const payload = JSON.stringify(metadata, null, 2);
await fs.writeFile(versionPath, payload);
await fs.writeFile(latestPath, payload);
return {
versionPath,
latestPath,
};
}
async function runImagesRelease(mode, metadata) {
const scriptName = mode === 'build' ? 'images:build' : 'images:release';
await spawnInherited('npm', ['run', scriptName], {
IMAGE_REGISTRY: metadata.registry,
IMAGE_NAMESPACE: metadata.namespace,
IMAGE_TAG: metadata.version_tag,
});
}
function printMetadata(metadata, manifestPaths) {
console.log(`channel: ${metadata.channel}`);
console.log(`git_sha: ${metadata.git_sha}`);
console.log(`version_tag: ${metadata.version_tag}`);
for (const image of metadata.images) {
console.log(`${image.id}: ${image.image}`);
console.log(`${image.id}: ${image.latest_image}`);
}
if (manifestPaths) {
console.log(`manifest: ${manifestPaths.versionPath}`);
console.log(`manifest_latest: ${manifestPaths.latestPath}`);
}
}
async function main() {
const action = getAction();
const metadata = await buildReleaseMetadata();
if (action === 'print') {
printMetadata(metadata);
return;
}
const manifestPaths = await writeManifest(metadata);
if (action === 'plan') {
printMetadata(metadata, manifestPaths);
return;
}
if (process.env.RUN_SMOKE_TEST === '1') {
await spawnInherited('npm', ['run', 'smoke:test']);
}
await runImagesRelease(action, metadata);
printMetadata(metadata, manifestPaths);
}
main().catch((error) => {
console.error(error.message);
process.exit(1);
});

169
scripts/skipper.js Normal file
View File

@@ -0,0 +1,169 @@
#!/usr/bin/env node
const { randomUUID } = require('crypto');
const { registerNode, randomTokenHex } = require('./register-node');
function getArgs() {
return process.argv.slice(2);
}
function getFlagValue(args, flag, fallback) {
const index = args.indexOf(flag);
if (index === -1 || index + 1 >= args.length) {
return fallback;
}
return args[index + 1];
}
function hasFlag(args, flag) {
return args.includes(flag);
}
function jsonOut(value) {
process.stdout.write(`${JSON.stringify(value, null, 2)}\n`);
}
async function fetchEnvelope(url, options) {
const response = await fetch(url, options);
const payload = await response.json();
if (!response.ok || payload.error) {
const error = new Error(payload && payload.error ? payload.error.message : `Request failed: ${response.status}`);
error.payload = payload;
throw error;
}
return payload;
}
function buildContextHeaders() {
return {
'x-request-id': randomUUID(),
'x-correlation-id': randomUUID(),
};
}
function usage() {
process.stdout.write([
'Skipper CLI',
'',
'Commands:',
' skipper health --url <base_url>',
' skipper deploy apply --url <base_url> --admin-token <token> --tenant-id <tenant_id> [--idempotency-key <key>]',
' skipper node register --node-id <id> [--role <role>] [--region <region>] [--token <token>]',
' skipper node show --url <base_url> --admin-token <token> --node-id <id>',
'',
].join('\n'));
}
async function cmdHealth(args) {
const baseUrl = getFlagValue(args, '--url', process.env.SKIPPER_URL || 'http://127.0.0.1:3000');
const payload = await fetchEnvelope(`${baseUrl.replace(/\/$/, '')}/v1/health`, {
headers: buildContextHeaders(),
});
jsonOut(payload);
}
async function cmdDeployApply(args) {
const baseUrl = getFlagValue(args, '--url', process.env.SKIPPER_URL || 'http://127.0.0.1:3000');
const adminToken = getFlagValue(args, '--admin-token', process.env.ADMIN_TOKEN);
const tenantId = getFlagValue(args, '--tenant-id', process.env.TENANT_ID);
const idempotencyKey = getFlagValue(args, '--idempotency-key', process.env.IDEMPOTENCY_KEY || randomUUID());
if (!adminToken) {
throw new Error('Missing admin token. Use --admin-token or ADMIN_TOKEN.');
}
if (!tenantId) {
throw new Error('Missing tenant id. Use --tenant-id or TENANT_ID.');
}
const payload = await fetchEnvelope(`${baseUrl.replace(/\/$/, '')}/v1/deployments/${tenantId}/apply`, {
method: 'POST',
headers: {
...buildContextHeaders(),
'x-admin-token': adminToken,
'x-idempotency-key': idempotencyKey,
},
});
jsonOut(payload);
}
async function cmdNodeRegister(args) {
const nodeId = getFlagValue(args, '--node-id', process.env.NODE_ID);
const role = getFlagValue(args, '--role', process.env.NODE_ROLE || 'worker');
const region = getFlagValue(args, '--region', process.env.NODE_REGION || 'default');
const token = getFlagValue(args, '--token', process.env.NODE_TOKEN || randomTokenHex(32));
if (!nodeId) {
throw new Error('Missing node id. Use --node-id or NODE_ID.');
}
jsonOut(await registerNode({ nodeId, role, region, token }));
}
async function cmdNodeShow(args) {
const baseUrl = getFlagValue(args, '--url', process.env.SKIPPER_URL || 'http://127.0.0.1:3000');
const adminToken = getFlagValue(args, '--admin-token', process.env.ADMIN_TOKEN);
const nodeId = getFlagValue(args, '--node-id', process.env.NODE_ID);
if (!adminToken) {
throw new Error('Missing admin token. Use --admin-token or ADMIN_TOKEN.');
}
if (!nodeId) {
throw new Error('Missing node id. Use --node-id or NODE_ID.');
}
const payload = await fetchEnvelope(`${baseUrl.replace(/\/$/, '')}/v1/resources/node/${nodeId}`, {
headers: {
...buildContextHeaders(),
'x-admin-token': adminToken,
},
});
jsonOut(payload);
}
async function main() {
const args = getArgs();
if (args.length === 0 || hasFlag(args, '--help') || hasFlag(args, '-h')) {
usage();
return;
}
const [group, command, subcommand] = args;
if (group === 'health') {
await cmdHealth(args.slice(1));
return;
}
if (group === 'deploy' && command === 'apply') {
await cmdDeployApply(args.slice(2));
return;
}
if (group === 'node' && command === 'register') {
await cmdNodeRegister(args.slice(2));
return;
}
if (group === 'node' && command === 'show') {
await cmdNodeShow(args.slice(2));
return;
}
throw new Error(`Unknown command: ${args.join(' ')}`);
}
main().catch((error) => {
process.stderr.write(`${error.message}\n`);
if (error.payload) {
process.stderr.write(`${JSON.stringify(error.payload, null, 2)}\n`);
}
process.exit(1);
});

307
scripts/smoke-test.js Normal file
View File

@@ -0,0 +1,307 @@
const fs = require('fs/promises');
const path = require('path');
const os = require('os');
const { randomUUID } = require('crypto');
const { spawn } = require('child_process');
const rootDir = path.resolve(__dirname, '..');
const apiDir = path.join(rootDir, 'skipper-api');
const agentDir = path.join(rootDir, 'skippy-agent');
const adminToken = 'smoke-admin-token';
const agentToken = 'smoke-node-token';
function sleep(ms) {
return new Promise((resolve) => setTimeout(resolve, ms));
}
async function ensureDir(dirPath) {
await fs.mkdir(dirPath, { recursive: true });
}
async function writeJson(filePath, value) {
await ensureDir(path.dirname(filePath));
await fs.writeFile(filePath, JSON.stringify(value, null, 2));
}
function startProcess(name, command, args, options) {
const child = spawn(command, args, {
...options,
stdio: ['ignore', 'pipe', 'pipe'],
});
child.stdout.on('data', (chunk) => {
process.stdout.write(`[${name}] ${chunk.toString()}`);
});
child.stderr.on('data', (chunk) => {
process.stderr.write(`[${name}] ${chunk.toString()}`);
});
return { child };
}
async function stopProcess(proc) {
if (!proc || proc.child.exitCode !== null) {
return;
}
proc.child.kill('SIGINT');
await Promise.race([
new Promise((resolve) => proc.child.once('exit', resolve)),
sleep(3000).then(() => {
proc.child.kill('SIGKILL');
}),
]);
}
async function waitForHealth(url, timeoutMs) {
const startedAt = Date.now();
while (Date.now() - startedAt < timeoutMs) {
try {
const response = await fetch(url);
if (response.ok) {
const payload = await response.json();
if (payload.data && payload.data.ok) {
return;
}
}
} catch (error) {
// Retry until timeout.
}
await sleep(200);
}
throw new Error(`Timed out waiting for ${url}`);
}
async function waitForFinishedWorkOrder(finishedDir, timeoutMs) {
const startedAt = Date.now();
while (Date.now() - startedAt < timeoutMs) {
const entries = await fs.readdir(finishedDir);
if (entries.length > 0) {
const filePath = path.join(finishedDir, entries.sort()[0]);
return JSON.parse(await fs.readFile(filePath, 'utf8'));
}
await sleep(250);
}
throw new Error('Timed out waiting for finished work order');
}
async function main() {
const runId = randomUUID();
const testRoot = path.join(os.tmpdir(), `skipper-smoke-${runId}`);
const dataDir = path.join(testRoot, 'data');
const composeDir = path.join(testRoot, 'compose');
const mockBinDir = path.join(testRoot, 'mockbin');
const apiPort = 3100;
const apiUrl = `http://127.0.0.1:${apiPort}`;
const finishedDir = path.join(dataDir, 'work-orders', 'finished');
const deployLogsPath = path.join(testRoot, 'docker-invocations.log');
await ensureDir(path.join(dataDir, 'work-orders', 'pending'));
await ensureDir(finishedDir);
await ensureDir(path.join(dataDir, 'resources', 'tenants'));
await ensureDir(path.join(dataDir, 'resources', 'nodes'));
await ensureDir(path.join(dataDir, 'resources', 'services'));
await ensureDir(path.join(dataDir, 'resources', 'deployments'));
await ensureDir(path.join(dataDir, 'resources', 'resource-limits'));
await ensureDir(path.join(dataDir, 'resources', 'networks'));
await ensureDir(path.join(dataDir, 'resources', 'volumes'));
await ensureDir(path.join(dataDir, 'auth', 'nodes'));
await ensureDir(composeDir);
await ensureDir(mockBinDir);
await writeJson(path.join(dataDir, 'resources', 'nodes', 'host-1.json'), {
id: 'host-1',
resource_type: 'node',
schema_version: 'v1',
desired_state: {
enabled: true,
labels: {
role: 'smoke-test',
},
},
current_state: {},
last_applied_state: {},
metadata: {},
created_at: new Date().toISOString(),
updated_at: new Date().toISOString(),
});
await writeJson(path.join(dataDir, 'resources', 'tenants', 'example-tenant.json'), {
id: 'example-tenant',
resource_type: 'tenant',
schema_version: 'v1',
desired_state: {
display_name: 'Example Tenant',
deployment_policy: {
target_node_id: 'host-1',
},
service_ids: ['service-web'],
compose: {
tenant_id: 'example-tenant',
compose_file: [
'services:',
' web:',
' image: nginx:alpine',
' restart: unless-stopped',
' ports:',
' - "${NGINX_PORT}:80"',
'',
].join('\n'),
env: {
NGINX_PORT: '8081',
},
},
},
current_state: {},
last_applied_state: {},
metadata: {},
created_at: new Date().toISOString(),
updated_at: new Date().toISOString(),
});
await writeJson(path.join(dataDir, 'resources', 'services', 'service-web.json'), {
id: 'service-web',
resource_type: 'service',
schema_version: 'v1',
desired_state: {
tenant_id: 'example-tenant',
service_kind: 'nginx',
image: 'nginx:alpine',
networks: [],
volumes: [],
resource_limits: null,
},
current_state: {},
last_applied_state: {},
metadata: {},
created_at: new Date().toISOString(),
updated_at: new Date().toISOString(),
});
await writeJson(path.join(dataDir, 'auth', 'nodes', 'host-1.json'), {
node_id: 'host-1',
token: agentToken,
schema_version: 'v1',
updated_at: new Date().toISOString(),
});
await fs.writeFile(
path.join(mockBinDir, 'docker'),
[
'#!/bin/sh',
`echo "$@" >> "${deployLogsPath}"`,
'echo "mock docker $@"',
'exit 0',
'',
].join('\n'),
{ mode: 0o755 }
);
const api = startProcess('api', process.execPath, ['src/index.js'], {
cwd: apiDir,
env: {
...process.env,
DATA_DIR: dataDir,
PORT: String(apiPort),
HOST: '127.0.0.1',
ADMIN_TOKEN: adminToken,
},
});
let agent;
try {
await waitForHealth(`${apiUrl}/v1/health`, 10000);
agent = startProcess('agent', process.execPath, ['src/index.js'], {
cwd: agentDir,
env: {
...process.env,
DATA_DIR: dataDir,
SKIPPER_URL: apiUrl,
AGENT_ID: 'host-1',
AGENT_TOKEN: agentToken,
POLL_INTERVAL_MS: '500',
HEARTBEAT_INTERVAL_MS: '1000',
SKIPPY_COMPOSE_BASE_DIR: composeDir,
PATH: `${mockBinDir}:${process.env.PATH}`,
},
});
const deployResponse = await fetch(`${apiUrl}/v1/deployments/example-tenant/apply`, {
method: 'POST',
headers: {
'x-admin-token': adminToken,
'x-idempotency-key': `smoke-${runId}`,
'x-request-id': randomUUID(),
'x-correlation-id': randomUUID(),
},
});
if (!deployResponse.ok) {
throw new Error(`Deploy request failed: ${deployResponse.status} ${await deployResponse.text()}`);
}
const createdResponse = await deployResponse.json();
const createdWorkOrder = createdResponse.data.work_order;
const completedWorkOrder = await waitForFinishedWorkOrder(finishedDir, 10000);
const composeFile = await fs.readFile(path.join(composeDir, 'example-tenant', 'docker-compose.yml'), 'utf8');
const envFile = await fs.readFile(path.join(composeDir, 'example-tenant', '.env'), 'utf8');
const dockerLog = await fs.readFile(deployLogsPath, 'utf8');
const nodeState = JSON.parse(await fs.readFile(path.join(dataDir, 'resources', 'nodes', 'host-1.json'), 'utf8'));
const tenantState = JSON.parse(await fs.readFile(path.join(dataDir, 'resources', 'tenants', 'example-tenant.json'), 'utf8'));
if (completedWorkOrder.id !== createdWorkOrder.id) {
throw new Error(`Completed unexpected work order ${completedWorkOrder.id}`);
}
if (!completedWorkOrder.result || !completedWorkOrder.result.success) {
throw new Error('Work order did not complete successfully');
}
if (!dockerLog.includes(`compose -f ${path.join(composeDir, 'example-tenant', 'docker-compose.yml')} up -d`)) {
throw new Error('Mock docker command was not invoked as expected');
}
if (!composeFile.includes('image: nginx:alpine')) {
throw new Error('Compose file was not written correctly');
}
if (!envFile.includes('NGINX_PORT=8081')) {
throw new Error('.env file was not written correctly');
}
if (nodeState.id !== 'host-1' || !nodeState.current_state.heartbeat_at) {
throw new Error('Node heartbeat was not persisted');
}
if (tenantState.current_state.last_deployment_status !== 'success') {
throw new Error('Tenant current state was not updated');
}
console.log('');
console.log('Smoke test passed');
console.log(`Work order: ${completedWorkOrder.id}`);
console.log(`Duration: ${completedWorkOrder.result.details.duration_ms}ms`);
} finally {
await stopProcess(agent);
await stopProcess(api);
await fs.rm(testRoot, { recursive: true, force: true });
}
}
main().catch((error) => {
console.error(error.stack || error.message);
process.exit(1);
});

32
shared/auth.js Normal file
View File

@@ -0,0 +1,32 @@
const path = require('path');
const { authNodesDir } = require('./paths');
const { readJsonIfExists, writeJson } = require('./fs');
function nodeTokenPath(nodeId) {
return path.join(authNodesDir, `${nodeId}.json`);
}
async function getNodeToken(nodeId) {
return readJsonIfExists(nodeTokenPath(nodeId));
}
async function verifyNodeToken(nodeId, token) {
const record = await getNodeToken(nodeId);
return Boolean(record && record.node_id === nodeId && record.token === token);
}
async function saveNodeToken(nodeId, token) {
await writeJson(nodeTokenPath(nodeId), {
node_id: nodeId,
token,
schema_version: 'v1',
updated_at: new Date().toISOString(),
});
}
module.exports = {
getNodeToken,
verifyNodeToken,
saveNodeToken,
};

38
shared/bootstrap.js vendored Normal file
View File

@@ -0,0 +1,38 @@
const {
dataDir,
resourcesDir,
resourceTypeToDirName,
workOrdersPendingDir,
workOrdersRunningDir,
workOrdersFinishedDir,
eventsDir,
logsDir,
snapshotsSystemDir,
snapshotsTenantsDir,
idempotencyDir,
authNodesDir,
} = require('./paths');
const { ensureDir } = require('./fs');
async function bootstrapDataLayout() {
await ensureDir(dataDir);
await ensureDir(resourcesDir);
for (const dirName of Object.values(resourceTypeToDirName)) {
await ensureDir(`${resourcesDir}/${dirName}`);
}
await ensureDir(workOrdersPendingDir);
await ensureDir(workOrdersRunningDir);
await ensureDir(workOrdersFinishedDir);
await ensureDir(eventsDir);
await ensureDir(logsDir);
await ensureDir(snapshotsSystemDir);
await ensureDir(snapshotsTenantsDir);
await ensureDir(idempotencyDir);
await ensureDir(authNodesDir);
}
module.exports = {
bootstrapDataLayout,
};

17
shared/context.js Normal file
View File

@@ -0,0 +1,17 @@
const { randomUUID } = require('crypto');
function normalizeUuid(value) {
return typeof value === 'string' && value.trim() ? value.trim() : null;
}
function createContext(input) {
return {
request_id: normalizeUuid(input && input.request_id) || randomUUID(),
correlation_id: normalizeUuid(input && input.correlation_id) || randomUUID(),
};
}
module.exports = {
createContext,
};

14
shared/errors.js Normal file
View File

@@ -0,0 +1,14 @@
class AppError extends Error {
constructor(statusCode, code, message, details) {
super(message);
this.name = 'AppError';
this.statusCode = statusCode;
this.code = code;
this.details = details || {};
}
}
module.exports = {
AppError,
};

42
shared/events.js Normal file
View File

@@ -0,0 +1,42 @@
const path = require('path');
const { randomUUID } = require('crypto');
const { eventsDir } = require('./paths');
const { writeJson } = require('./fs');
function dayString(timestamp) {
return timestamp.slice(0, 10);
}
async function emitEvent({
type,
resource_type,
resource_id,
request_id,
correlation_id,
payload,
}) {
const timestamp = new Date().toISOString();
const event = {
event_id: randomUUID(),
schema_version: 'v1',
type,
resource_type,
resource_id,
timestamp,
request_id: request_id || null,
correlation_id: correlation_id || null,
payload: payload || {},
};
await writeJson(
path.join(eventsDir, dayString(timestamp), `${timestamp.replace(/[:.]/g, '-')}-${event.event_id}.json`),
event
);
return event;
}
module.exports = {
emitEvent,
};

72
shared/fs.js Normal file
View File

@@ -0,0 +1,72 @@
const fs = require('fs/promises');
const path = require('path');
const { randomUUID } = require('crypto');
async function ensureDir(dirPath) {
await fs.mkdir(dirPath, { recursive: true });
}
async function readJson(filePath) {
const raw = await fs.readFile(filePath, 'utf8');
return JSON.parse(raw);
}
async function readJsonIfExists(filePath) {
if (!(await fileExists(filePath))) {
return null;
}
return readJson(filePath);
}
async function writeJson(filePath, value) {
await ensureDir(path.dirname(filePath));
const tempPath = `${filePath}.${randomUUID()}.tmp`;
await fs.writeFile(tempPath, JSON.stringify(value, null, 2));
await fs.rename(tempPath, filePath);
}
async function fileExists(filePath) {
try {
await fs.access(filePath);
return true;
} catch (error) {
return false;
}
}
async function listJsonFiles(dirPath) {
await ensureDir(dirPath);
const entries = await fs.readdir(dirPath, { withFileTypes: true });
return entries
.filter((entry) => entry.isFile() && entry.name.endsWith('.json'))
.map((entry) => path.join(dirPath, entry.name))
.sort();
}
async function appendJsonLine(filePath, value) {
await ensureDir(path.dirname(filePath));
await fs.appendFile(filePath, `${JSON.stringify(value)}\n`);
}
async function moveFile(fromPath, toPath) {
await ensureDir(path.dirname(toPath));
await fs.rename(fromPath, toPath);
}
async function removeFile(filePath) {
await fs.rm(filePath, { force: true });
}
module.exports = {
ensureDir,
readJson,
readJsonIfExists,
writeJson,
fileExists,
listJsonFiles,
appendJsonLine,
moveFile,
removeFile,
};

29
shared/idempotency.js Normal file
View File

@@ -0,0 +1,29 @@
const path = require('path');
const { createHash } = require('crypto');
const { idempotencyDir } = require('./paths');
const { readJsonIfExists, writeJson } = require('./fs');
function recordPath(scope, key) {
const hash = createHash('sha256').update(`${scope}:${key}`).digest('hex');
return path.join(idempotencyDir, scope, `${hash}.json`);
}
async function getIdempotencyRecord(scope, key) {
return readJsonIfExists(recordPath(scope, key));
}
async function saveIdempotencyRecord(scope, key, value) {
await writeJson(recordPath(scope, key), {
schema_version: 'v1',
scope,
key,
value,
created_at: new Date().toISOString(),
});
}
module.exports = {
getIdempotencyRecord,
saveIdempotencyRecord,
};

71
shared/logs.js Normal file
View File

@@ -0,0 +1,71 @@
const path = require('path');
const { logsDir } = require('./paths');
const { appendJsonLine } = require('./fs');
const SECRET_PATTERNS = ['token', 'secret', 'password', 'authorization'];
function sanitizeValue(value) {
if (Array.isArray(value)) {
return value.map(sanitizeValue);
}
if (!value || typeof value !== 'object') {
return value;
}
return Object.fromEntries(
Object.entries(value).map(([key, entryValue]) => {
const lower = key.toLowerCase();
if (SECRET_PATTERNS.some((pattern) => lower.includes(pattern))) {
return [key, '[redacted]'];
}
return [key, sanitizeValue(entryValue)];
})
);
}
function dayString(timestamp) {
return timestamp.slice(0, 10);
}
async function log(entry) {
const timestamp = new Date().toISOString();
const logEntry = {
timestamp,
level: entry.level || 'info',
service: entry.service,
node_id: entry.node_id || null,
tenant_id: entry.tenant_id || null,
request_id: entry.request_id || null,
correlation_id: entry.correlation_id || null,
action: entry.action || 'unknown',
result: entry.result || 'unknown',
metadata: sanitizeValue(entry.metadata || {}),
};
await appendJsonLine(path.join(logsDir, dayString(timestamp), `${logEntry.service}.ndjson`), logEntry);
process.stdout.write(`${JSON.stringify(logEntry)}\n`);
return logEntry;
}
function createLogger(baseFields) {
return {
info(action, result, metadata) {
return log({ ...baseFields, level: 'info', action, result, metadata });
},
warn(action, result, metadata) {
return log({ ...baseFields, level: 'warn', action, result, metadata });
},
error(action, result, metadata) {
return log({ ...baseFields, level: 'error', action, result, metadata });
},
};
}
module.exports = {
log,
createLogger,
};

43
shared/paths.js Normal file
View File

@@ -0,0 +1,43 @@
const path = require('path');
const rootDir = path.resolve(__dirname, '..');
const dataDir = process.env.DATA_DIR
? path.resolve(process.env.DATA_DIR)
: path.join(rootDir, 'data');
const resourceTypeToDirName = {
tenant: 'tenants',
node: 'nodes',
service: 'services',
deployment: 'deployments',
resource_limits: 'resource-limits',
network: 'networks',
volume: 'volumes',
};
function getResourceDir(resourceType) {
const dirName = resourceTypeToDirName[resourceType];
if (!dirName) {
throw new Error(`Unknown resource type: ${resourceType}`);
}
return path.join(dataDir, 'resources', dirName);
}
module.exports = {
rootDir,
dataDir,
resourceTypeToDirName,
getResourceDir,
resourcesDir: path.join(dataDir, 'resources'),
workOrdersPendingDir: path.join(dataDir, 'work-orders', 'pending'),
workOrdersRunningDir: path.join(dataDir, 'work-orders', 'running'),
workOrdersFinishedDir: path.join(dataDir, 'work-orders', 'finished'),
eventsDir: path.join(dataDir, 'events'),
logsDir: path.join(dataDir, 'logs'),
snapshotsSystemDir: path.join(dataDir, 'snapshots', 'system'),
snapshotsTenantsDir: path.join(dataDir, 'snapshots', 'tenants'),
idempotencyDir: path.join(dataDir, 'idempotency'),
authNodesDir: path.join(dataDir, 'auth', 'nodes'),
};

94
shared/resources.js Normal file
View File

@@ -0,0 +1,94 @@
const path = require('path');
const { getResourceDir, resourceTypeToDirName } = require('./paths');
const { ensureDir, listJsonFiles, readJsonIfExists, writeJson } = require('./fs');
function resourcePath(resourceType, resourceId) {
return path.join(getResourceDir(resourceType), `${resourceId}.json`);
}
function nowIso() {
return new Date().toISOString();
}
function createResourceDocument({
id,
resource_type,
desired_state,
current_state,
last_applied_state,
metadata,
created_at,
updated_at,
}) {
const timestamp = nowIso();
return {
id,
resource_type,
schema_version: 'v1',
desired_state: desired_state || {},
current_state: current_state || {},
last_applied_state: last_applied_state || {},
metadata: metadata || {},
created_at: created_at || timestamp,
updated_at: updated_at || timestamp,
};
}
async function getResource(resourceType, resourceId) {
return readJsonIfExists(resourcePath(resourceType, resourceId));
}
async function saveResource(document) {
const existing = await getResource(document.resource_type, document.id);
const next = createResourceDocument({
...document,
created_at: existing ? existing.created_at : document.created_at,
updated_at: nowIso(),
});
await writeJson(resourcePath(next.resource_type, next.id), next);
return next;
}
async function patchResourceState(resourceType, resourceId, patch) {
const existing = await getResource(resourceType, resourceId);
if (!existing) {
return null;
}
return saveResource({
...existing,
desired_state: patch.desired_state || existing.desired_state,
current_state: patch.current_state || existing.current_state,
last_applied_state: patch.last_applied_state || existing.last_applied_state,
metadata: patch.metadata || existing.metadata,
});
}
async function listResources(resourceType) {
const files = await listJsonFiles(getResourceDir(resourceType));
return Promise.all(files.map((filePath) => readJsonIfExists(filePath)));
}
async function listAllResources() {
const result = {};
for (const resourceType of Object.keys(resourceTypeToDirName)) {
await ensureDir(getResourceDir(resourceType));
result[resourceType] = await listResources(resourceType);
}
return result;
}
module.exports = {
createResourceDocument,
getResource,
saveResource,
patchResourceState,
listResources,
listAllResources,
};

80
shared/snapshots.js Normal file
View File

@@ -0,0 +1,80 @@
const path = require('path');
const { randomUUID } = require('crypto');
const { snapshotsSystemDir, snapshotsTenantsDir } = require('./paths');
const { readJsonIfExists, writeJson } = require('./fs');
const { listAllResources } = require('./resources');
function buildResourceDiff(resource) {
return {
in_sync: JSON.stringify(resource.desired_state || {}) === JSON.stringify(resource.current_state || {}),
desired_state: resource.desired_state || {},
current_state: resource.current_state || {},
};
}
function filterTenantResources(resources, tenantId) {
const filtered = {};
for (const [resourceType, items] of Object.entries(resources)) {
filtered[resourceType] = items.filter((item) => {
if (!item) {
return false;
}
if (resourceType === 'tenant') {
return item.id === tenantId;
}
return item.desired_state && item.desired_state.tenant_id === tenantId;
});
}
return filtered;
}
async function createSnapshot(scope, context) {
const resources = await listAllResources();
const selected = scope === 'system'
? resources
: filterTenantResources(resources, scope.replace('tenant:', ''));
const snapshot = {
snapshot_id: randomUUID(),
schema_version: 'v1',
scope,
created_at: new Date().toISOString(),
request_id: context.request_id,
correlation_id: context.correlation_id,
resources: selected,
diffs: Object.fromEntries(
Object.entries(selected).map(([resourceType, items]) => [
resourceType,
items.map((item) => ({
resource_id: item.id,
diff: buildResourceDiff(item),
})),
])
),
};
const filePath = scope === 'system'
? path.join(snapshotsSystemDir, 'latest.json')
: path.join(snapshotsTenantsDir, `${scope.replace('tenant:', '')}.json`);
await writeJson(filePath, snapshot);
return snapshot;
}
async function getLatestSystemSnapshot() {
return readJsonIfExists(path.join(snapshotsSystemDir, 'latest.json'));
}
async function getLatestTenantSnapshot(tenantId) {
return readJsonIfExists(path.join(snapshotsTenantsDir, `${tenantId}.json`));
}
module.exports = {
createSnapshot,
getLatestSystemSnapshot,
getLatestTenantSnapshot,
};

167
shared/work-orders.js Normal file
View File

@@ -0,0 +1,167 @@
const fs = require('fs/promises');
const path = require('path');
const { randomUUID } = require('crypto');
const { workOrdersPendingDir, workOrdersRunningDir, workOrdersFinishedDir } = require('./paths');
const { ensureDir, fileExists, listJsonFiles, readJson, readJsonIfExists, writeJson } = require('./fs');
function pendingPath(workOrderId) {
return path.join(workOrdersPendingDir, `${workOrderId}.json`);
}
function runningPath(workOrderId) {
return path.join(workOrdersRunningDir, `${workOrderId}.json`);
}
function finishedPath(workOrderId) {
return path.join(workOrdersFinishedDir, `${workOrderId}.json`);
}
function lockPath(workOrderId) {
return path.join(workOrdersPendingDir, `${workOrderId}.lock`);
}
async function withLock(workOrderId, callback) {
const targetPath = lockPath(workOrderId);
await ensureDir(workOrdersPendingDir);
try {
await fs.mkdir(targetPath);
} catch (error) {
if (error.code === 'EEXIST') {
return null;
}
throw error;
}
try {
return await callback();
} finally {
await fs.rm(targetPath, { recursive: true, force: true });
}
}
async function createWorkOrder({ type, target, desired_state, request_id, correlation_id, metadata }) {
const timestamp = new Date().toISOString();
const workOrder = {
id: randomUUID(),
resource_type: 'work_order',
schema_version: 'v1',
type,
target,
desired_state: desired_state || {},
status: 'pending',
result: null,
request_id: request_id || null,
correlation_id: correlation_id || null,
created_at: timestamp,
started_at: null,
finished_at: null,
metadata: metadata || {},
};
await writeJson(pendingPath(workOrder.id), workOrder);
return workOrder;
}
async function getWorkOrder(workOrderId) {
return (
(await readJsonIfExists(pendingPath(workOrderId))) ||
(await readJsonIfExists(runningPath(workOrderId))) ||
(await readJsonIfExists(finishedPath(workOrderId)))
);
}
async function listWorkOrders() {
const files = [
...(await listJsonFiles(workOrdersPendingDir)),
...(await listJsonFiles(workOrdersRunningDir)),
...(await listJsonFiles(workOrdersFinishedDir)),
];
return Promise.all(files.map((filePath) => readJson(filePath)));
}
async function claimNextWorkOrder(nodeId, context) {
const files = await listJsonFiles(workOrdersPendingDir);
for (const filePath of files) {
const current = await readJson(filePath);
if (!current.target || current.target.node_id !== nodeId) {
continue;
}
const claimed = await withLock(current.id, async () => {
if (!(await fileExists(filePath))) {
return null;
}
const latest = await readJson(filePath);
if (!latest.target || latest.target.node_id !== nodeId || latest.status !== 'pending') {
return null;
}
const running = {
...latest,
status: 'running',
started_at: new Date().toISOString(),
request_id: context.request_id,
correlation_id: context.correlation_id,
};
await writeJson(runningPath(latest.id), running);
await fs.rm(filePath, { force: true });
return running;
});
if (claimed) {
return claimed;
}
}
return null;
}
async function finishWorkOrder(workOrderId, status, result) {
const existingFinished = await readJsonIfExists(finishedPath(workOrderId));
if (existingFinished) {
return existingFinished;
}
const activePath = runningPath(workOrderId);
if (!(await fileExists(activePath))) {
return null;
}
return withLock(workOrderId, async () => {
const running = await readJsonIfExists(activePath);
if (!running) {
return readJsonIfExists(finishedPath(workOrderId));
}
const finished = {
...running,
status,
result,
finished_at: new Date().toISOString(),
};
await writeJson(finishedPath(workOrderId), finished);
await fs.rm(activePath, { force: true });
return finished;
});
}
module.exports = {
createWorkOrder,
getWorkOrder,
listWorkOrders,
claimNextWorkOrder,
finishWorkOrder,
};

15
skipper-api/Dockerfile Normal file
View File

@@ -0,0 +1,15 @@
FROM node:20-alpine
WORKDIR /app/skipper-api
COPY skipper-api/package.json ./package.json
RUN npm install --omit=dev
COPY skipper-api/src ./src
COPY shared /app/shared
ENV PORT=3000
ENV DATA_DIR=/app/data
CMD ["npm", "start"]

759
skipper-api/package-lock.json generated Normal file
View File

@@ -0,0 +1,759 @@
{
"name": "skipper-api",
"version": "1.0.0",
"lockfileVersion": 3,
"requires": true,
"packages": {
"": {
"name": "skipper-api",
"version": "1.0.0",
"dependencies": {
"express": "^4.21.2"
}
},
"node_modules/accepts": {
"version": "1.3.8",
"resolved": "https://registry.npmjs.org/accepts/-/accepts-1.3.8.tgz",
"integrity": "sha512-PYAthTa2m2VKxuvSD3DPC/Gy+U+sOA1LAuT8mkmRuvw+NACSaeXEQ+NHcVF7rONl6qcaxV3Uuemwawk+7+SJLw==",
"dependencies": {
"mime-types": "~2.1.34",
"negotiator": "0.6.3"
},
"engines": {
"node": ">= 0.6"
}
},
"node_modules/array-flatten": {
"version": "1.1.1",
"resolved": "https://registry.npmjs.org/array-flatten/-/array-flatten-1.1.1.tgz",
"integrity": "sha512-PCVAQswWemu6UdxsDFFX/+gVeYqKAod3D3UVm91jHwynguOwAvYPhx8nNlM++NqRcK6CxxpUafjmhIdKiHibqg=="
},
"node_modules/body-parser": {
"version": "1.20.4",
"resolved": "https://registry.npmjs.org/body-parser/-/body-parser-1.20.4.tgz",
"integrity": "sha512-ZTgYYLMOXY9qKU/57FAo8F+HA2dGX7bqGc71txDRC1rS4frdFI5R7NhluHxH6M0YItAP0sHB4uqAOcYKxO6uGA==",
"dependencies": {
"bytes": "~3.1.2",
"content-type": "~1.0.5",
"debug": "2.6.9",
"depd": "2.0.0",
"destroy": "~1.2.0",
"http-errors": "~2.0.1",
"iconv-lite": "~0.4.24",
"on-finished": "~2.4.1",
"qs": "~6.14.0",
"raw-body": "~2.5.3",
"type-is": "~1.6.18",
"unpipe": "~1.0.0"
},
"engines": {
"node": ">= 0.8",
"npm": "1.2.8000 || >= 1.4.16"
}
},
"node_modules/bytes": {
"version": "3.1.2",
"resolved": "https://registry.npmjs.org/bytes/-/bytes-3.1.2.tgz",
"integrity": "sha512-/Nf7TyzTx6S3yRJObOAV7956r8cr2+Oj8AC5dt8wSP3BQAoeX58NoHyCU8P8zGkNXStjTSi6fzO6F0pBdcYbEg==",
"engines": {
"node": ">= 0.8"
}
},
"node_modules/call-bind-apply-helpers": {
"version": "1.0.2",
"resolved": "https://registry.npmjs.org/call-bind-apply-helpers/-/call-bind-apply-helpers-1.0.2.tgz",
"integrity": "sha512-Sp1ablJ0ivDkSzjcaJdxEunN5/XvksFJ2sMBFfq6x0ryhQV/2b/KwFe21cMpmHtPOSij8K99/wSfoEuTObmuMQ==",
"dependencies": {
"es-errors": "^1.3.0",
"function-bind": "^1.1.2"
},
"engines": {
"node": ">= 0.4"
}
},
"node_modules/call-bound": {
"version": "1.0.4",
"resolved": "https://registry.npmjs.org/call-bound/-/call-bound-1.0.4.tgz",
"integrity": "sha512-+ys997U96po4Kx/ABpBCqhA9EuxJaQWDQg7295H4hBphv3IZg0boBKuwYpt4YXp6MZ5AmZQnU/tyMTlRpaSejg==",
"dependencies": {
"call-bind-apply-helpers": "^1.0.2",
"get-intrinsic": "^1.3.0"
},
"engines": {
"node": ">= 0.4"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/content-disposition": {
"version": "0.5.4",
"resolved": "https://registry.npmjs.org/content-disposition/-/content-disposition-0.5.4.tgz",
"integrity": "sha512-FveZTNuGw04cxlAiWbzi6zTAL/lhehaWbTtgluJh4/E95DqMwTmha3KZN1aAWA8cFIhHzMZUvLevkw5Rqk+tSQ==",
"dependencies": {
"safe-buffer": "5.2.1"
},
"engines": {
"node": ">= 0.6"
}
},
"node_modules/content-type": {
"version": "1.0.5",
"resolved": "https://registry.npmjs.org/content-type/-/content-type-1.0.5.tgz",
"integrity": "sha512-nTjqfcBFEipKdXCv4YDQWCfmcLZKm81ldF0pAopTvyrFGVbcR6P/VAAd5G7N+0tTr8QqiU0tFadD6FK4NtJwOA==",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/cookie": {
"version": "0.7.2",
"resolved": "https://registry.npmjs.org/cookie/-/cookie-0.7.2.tgz",
"integrity": "sha512-yki5XnKuf750l50uGTllt6kKILY4nQ1eNIQatoXEByZ5dWgnKqbnqmTrBE5B4N7lrMJKQ2ytWMiTO2o0v6Ew/w==",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/cookie-signature": {
"version": "1.0.7",
"resolved": "https://registry.npmjs.org/cookie-signature/-/cookie-signature-1.0.7.tgz",
"integrity": "sha512-NXdYc3dLr47pBkpUCHtKSwIOQXLVn8dZEuywboCOJY/osA0wFSLlSawr3KN8qXJEyX66FcONTH8EIlVuK0yyFA=="
},
"node_modules/debug": {
"version": "2.6.9",
"resolved": "https://registry.npmjs.org/debug/-/debug-2.6.9.tgz",
"integrity": "sha512-bC7ElrdJaJnPbAP+1EotYvqZsb3ecl5wi6Bfi6BJTUcNowp6cvspg0jXznRTKDjm/E7AdgFBVeAPVMNcKGsHMA==",
"dependencies": {
"ms": "2.0.0"
}
},
"node_modules/depd": {
"version": "2.0.0",
"resolved": "https://registry.npmjs.org/depd/-/depd-2.0.0.tgz",
"integrity": "sha512-g7nH6P6dyDioJogAAGprGpCtVImJhpPk/roCzdb3fIh61/s/nPsfR6onyMwkCAR/OlC3yBC0lESvUoQEAssIrw==",
"engines": {
"node": ">= 0.8"
}
},
"node_modules/destroy": {
"version": "1.2.0",
"resolved": "https://registry.npmjs.org/destroy/-/destroy-1.2.0.tgz",
"integrity": "sha512-2sJGJTaXIIaR1w4iJSNoN0hnMY7Gpc/n8D4qSCJw8QqFWXf7cuAgnEHxBpweaVcPevC2l3KpjYCx3NypQQgaJg==",
"engines": {
"node": ">= 0.8",
"npm": "1.2.8000 || >= 1.4.16"
}
},
"node_modules/dunder-proto": {
"version": "1.0.1",
"resolved": "https://registry.npmjs.org/dunder-proto/-/dunder-proto-1.0.1.tgz",
"integrity": "sha512-KIN/nDJBQRcXw0MLVhZE9iQHmG68qAVIBg9CqmUYjmQIhgij9U5MFvrqkUL5FbtyyzZuOeOt0zdeRe4UY7ct+A==",
"dependencies": {
"call-bind-apply-helpers": "^1.0.1",
"es-errors": "^1.3.0",
"gopd": "^1.2.0"
},
"engines": {
"node": ">= 0.4"
}
},
"node_modules/ee-first": {
"version": "1.1.1",
"resolved": "https://registry.npmjs.org/ee-first/-/ee-first-1.1.1.tgz",
"integrity": "sha512-WMwm9LhRUo+WUaRN+vRuETqG89IgZphVSNkdFgeb6sS/E4OrDIN7t48CAewSHXc6C8lefD8KKfr5vY61brQlow=="
},
"node_modules/encodeurl": {
"version": "2.0.0",
"resolved": "https://registry.npmjs.org/encodeurl/-/encodeurl-2.0.0.tgz",
"integrity": "sha512-Q0n9HRi4m6JuGIV1eFlmvJB7ZEVxu93IrMyiMsGC0lrMJMWzRgx6WGquyfQgZVb31vhGgXnfmPNNXmxnOkRBrg==",
"engines": {
"node": ">= 0.8"
}
},
"node_modules/es-define-property": {
"version": "1.0.1",
"resolved": "https://registry.npmjs.org/es-define-property/-/es-define-property-1.0.1.tgz",
"integrity": "sha512-e3nRfgfUZ4rNGL232gUgX06QNyyez04KdjFrF+LTRoOXmrOgFKDg4BCdsjW8EnT69eqdYGmRpJwiPVYNrCaW3g==",
"engines": {
"node": ">= 0.4"
}
},
"node_modules/es-errors": {
"version": "1.3.0",
"resolved": "https://registry.npmjs.org/es-errors/-/es-errors-1.3.0.tgz",
"integrity": "sha512-Zf5H2Kxt2xjTvbJvP2ZWLEICxA6j+hAmMzIlypy4xcBg1vKVnx89Wy0GbS+kf5cwCVFFzdCFh2XSCFNULS6csw==",
"engines": {
"node": ">= 0.4"
}
},
"node_modules/es-object-atoms": {
"version": "1.1.1",
"resolved": "https://registry.npmjs.org/es-object-atoms/-/es-object-atoms-1.1.1.tgz",
"integrity": "sha512-FGgH2h8zKNim9ljj7dankFPcICIK9Cp5bm+c2gQSYePhpaG5+esrLODihIorn+Pe6FGJzWhXQotPv73jTaldXA==",
"dependencies": {
"es-errors": "^1.3.0"
},
"engines": {
"node": ">= 0.4"
}
},
"node_modules/escape-html": {
"version": "1.0.3",
"resolved": "https://registry.npmjs.org/escape-html/-/escape-html-1.0.3.tgz",
"integrity": "sha512-NiSupZ4OeuGwr68lGIeym/ksIZMJodUGOSCZ/FSnTxcrekbvqrgdUxlJOMpijaKZVjAJrWrGs/6Jy8OMuyj9ow=="
},
"node_modules/etag": {
"version": "1.8.1",
"resolved": "https://registry.npmjs.org/etag/-/etag-1.8.1.tgz",
"integrity": "sha512-aIL5Fx7mawVa300al2BnEE4iNvo1qETxLrPI/o05L7z6go7fCw1J6EQmbK4FmJ2AS7kgVF/KEZWufBfdClMcPg==",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/express": {
"version": "4.22.1",
"resolved": "https://registry.npmjs.org/express/-/express-4.22.1.tgz",
"integrity": "sha512-F2X8g9P1X7uCPZMA3MVf9wcTqlyNp7IhH5qPCI0izhaOIYXaW9L535tGA3qmjRzpH+bZczqq7hVKxTR4NWnu+g==",
"dependencies": {
"accepts": "~1.3.8",
"array-flatten": "1.1.1",
"body-parser": "~1.20.3",
"content-disposition": "~0.5.4",
"content-type": "~1.0.4",
"cookie": "~0.7.1",
"cookie-signature": "~1.0.6",
"debug": "2.6.9",
"depd": "2.0.0",
"encodeurl": "~2.0.0",
"escape-html": "~1.0.3",
"etag": "~1.8.1",
"finalhandler": "~1.3.1",
"fresh": "~0.5.2",
"http-errors": "~2.0.0",
"merge-descriptors": "1.0.3",
"methods": "~1.1.2",
"on-finished": "~2.4.1",
"parseurl": "~1.3.3",
"path-to-regexp": "~0.1.12",
"proxy-addr": "~2.0.7",
"qs": "~6.14.0",
"range-parser": "~1.2.1",
"safe-buffer": "5.2.1",
"send": "~0.19.0",
"serve-static": "~1.16.2",
"setprototypeof": "1.2.0",
"statuses": "~2.0.1",
"type-is": "~1.6.18",
"utils-merge": "1.0.1",
"vary": "~1.1.2"
},
"engines": {
"node": ">= 0.10.0"
},
"funding": {
"type": "opencollective",
"url": "https://opencollective.com/express"
}
},
"node_modules/finalhandler": {
"version": "1.3.2",
"resolved": "https://registry.npmjs.org/finalhandler/-/finalhandler-1.3.2.tgz",
"integrity": "sha512-aA4RyPcd3badbdABGDuTXCMTtOneUCAYH/gxoYRTZlIJdF0YPWuGqiAsIrhNnnqdXGswYk6dGujem4w80UJFhg==",
"dependencies": {
"debug": "2.6.9",
"encodeurl": "~2.0.0",
"escape-html": "~1.0.3",
"on-finished": "~2.4.1",
"parseurl": "~1.3.3",
"statuses": "~2.0.2",
"unpipe": "~1.0.0"
},
"engines": {
"node": ">= 0.8"
}
},
"node_modules/forwarded": {
"version": "0.2.0",
"resolved": "https://registry.npmjs.org/forwarded/-/forwarded-0.2.0.tgz",
"integrity": "sha512-buRG0fpBtRHSTCOASe6hD258tEubFoRLb4ZNA6NxMVHNw2gOcwHo9wyablzMzOA5z9xA9L1KNjk/Nt6MT9aYow==",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/fresh": {
"version": "0.5.2",
"resolved": "https://registry.npmjs.org/fresh/-/fresh-0.5.2.tgz",
"integrity": "sha512-zJ2mQYM18rEFOudeV4GShTGIQ7RbzA7ozbU9I/XBpm7kqgMywgmylMwXHxZJmkVoYkna9d2pVXVXPdYTP9ej8Q==",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/function-bind": {
"version": "1.1.2",
"resolved": "https://registry.npmjs.org/function-bind/-/function-bind-1.1.2.tgz",
"integrity": "sha512-7XHNxH7qX9xG5mIwxkhumTox/MIRNcOgDrxWsMt2pAr23WHp6MrRlN7FBSFpCpr+oVO0F744iUgR82nJMfG2SA==",
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/get-intrinsic": {
"version": "1.3.0",
"resolved": "https://registry.npmjs.org/get-intrinsic/-/get-intrinsic-1.3.0.tgz",
"integrity": "sha512-9fSjSaos/fRIVIp+xSJlE6lfwhES7LNtKaCBIamHsjr2na1BiABJPo0mOjjz8GJDURarmCPGqaiVg5mfjb98CQ==",
"dependencies": {
"call-bind-apply-helpers": "^1.0.2",
"es-define-property": "^1.0.1",
"es-errors": "^1.3.0",
"es-object-atoms": "^1.1.1",
"function-bind": "^1.1.2",
"get-proto": "^1.0.1",
"gopd": "^1.2.0",
"has-symbols": "^1.1.0",
"hasown": "^2.0.2",
"math-intrinsics": "^1.1.0"
},
"engines": {
"node": ">= 0.4"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/get-proto": {
"version": "1.0.1",
"resolved": "https://registry.npmjs.org/get-proto/-/get-proto-1.0.1.tgz",
"integrity": "sha512-sTSfBjoXBp89JvIKIefqw7U2CCebsc74kiY6awiGogKtoSGbgjYE/G/+l9sF3MWFPNc9IcoOC4ODfKHfxFmp0g==",
"dependencies": {
"dunder-proto": "^1.0.1",
"es-object-atoms": "^1.0.0"
},
"engines": {
"node": ">= 0.4"
}
},
"node_modules/gopd": {
"version": "1.2.0",
"resolved": "https://registry.npmjs.org/gopd/-/gopd-1.2.0.tgz",
"integrity": "sha512-ZUKRh6/kUFoAiTAtTYPZJ3hw9wNxx+BIBOijnlG9PnrJsCcSjs1wyyD6vJpaYtgnzDrKYRSqf3OO6Rfa93xsRg==",
"engines": {
"node": ">= 0.4"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/has-symbols": {
"version": "1.1.0",
"resolved": "https://registry.npmjs.org/has-symbols/-/has-symbols-1.1.0.tgz",
"integrity": "sha512-1cDNdwJ2Jaohmb3sg4OmKaMBwuC48sYni5HUw2DvsC8LjGTLK9h+eb1X6RyuOHe4hT0ULCW68iomhjUoKUqlPQ==",
"engines": {
"node": ">= 0.4"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/hasown": {
"version": "2.0.2",
"resolved": "https://registry.npmjs.org/hasown/-/hasown-2.0.2.tgz",
"integrity": "sha512-0hJU9SCPvmMzIBdZFqNPXWa6dqh7WdH0cII9y+CyS8rG3nL48Bclra9HmKhVVUHyPWNH5Y7xDwAB7bfgSjkUMQ==",
"dependencies": {
"function-bind": "^1.1.2"
},
"engines": {
"node": ">= 0.4"
}
},
"node_modules/http-errors": {
"version": "2.0.1",
"resolved": "https://registry.npmjs.org/http-errors/-/http-errors-2.0.1.tgz",
"integrity": "sha512-4FbRdAX+bSdmo4AUFuS0WNiPz8NgFt+r8ThgNWmlrjQjt1Q7ZR9+zTlce2859x4KSXrwIsaeTqDoKQmtP8pLmQ==",
"dependencies": {
"depd": "~2.0.0",
"inherits": "~2.0.4",
"setprototypeof": "~1.2.0",
"statuses": "~2.0.2",
"toidentifier": "~1.0.1"
},
"engines": {
"node": ">= 0.8"
},
"funding": {
"type": "opencollective",
"url": "https://opencollective.com/express"
}
},
"node_modules/iconv-lite": {
"version": "0.4.24",
"resolved": "https://registry.npmjs.org/iconv-lite/-/iconv-lite-0.4.24.tgz",
"integrity": "sha512-v3MXnZAcvnywkTUEZomIActle7RXXeedOR31wwl7VlyoXO4Qi9arvSenNQWne1TcRwhCL1HwLI21bEqdpj8/rA==",
"dependencies": {
"safer-buffer": ">= 2.1.2 < 3"
},
"engines": {
"node": ">=0.10.0"
}
},
"node_modules/inherits": {
"version": "2.0.4",
"resolved": "https://registry.npmjs.org/inherits/-/inherits-2.0.4.tgz",
"integrity": "sha512-k/vGaX4/Yla3WzyMCvTQOXYeIHvqOKtnqBduzTHpzpQZzAskKMhZ2K+EnBiSM9zGSoIFeMpXKxa4dYeZIQqewQ=="
},
"node_modules/ipaddr.js": {
"version": "1.9.1",
"resolved": "https://registry.npmjs.org/ipaddr.js/-/ipaddr.js-1.9.1.tgz",
"integrity": "sha512-0KI/607xoxSToH7GjN1FfSbLoU0+btTicjsQSWQlh/hZykN8KpmMf7uYwPW3R+akZ6R/w18ZlXSHBYXiYUPO3g==",
"engines": {
"node": ">= 0.10"
}
},
"node_modules/math-intrinsics": {
"version": "1.1.0",
"resolved": "https://registry.npmjs.org/math-intrinsics/-/math-intrinsics-1.1.0.tgz",
"integrity": "sha512-/IXtbwEk5HTPyEwyKX6hGkYXxM9nbj64B+ilVJnC/R6B0pH5G4V3b0pVbL7DBj4tkhBAppbQUlf6F6Xl9LHu1g==",
"engines": {
"node": ">= 0.4"
}
},
"node_modules/media-typer": {
"version": "0.3.0",
"resolved": "https://registry.npmjs.org/media-typer/-/media-typer-0.3.0.tgz",
"integrity": "sha512-dq+qelQ9akHpcOl/gUVRTxVIOkAJ1wR3QAvb4RsVjS8oVoFjDGTc679wJYmUmknUF5HwMLOgb5O+a3KxfWapPQ==",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/merge-descriptors": {
"version": "1.0.3",
"resolved": "https://registry.npmjs.org/merge-descriptors/-/merge-descriptors-1.0.3.tgz",
"integrity": "sha512-gaNvAS7TZ897/rVaZ0nMtAyxNyi/pdbjbAwUpFQpN70GqnVfOiXpeUUMKRBmzXaSQ8DdTX4/0ms62r2K+hE6mQ==",
"funding": {
"url": "https://github.com/sponsors/sindresorhus"
}
},
"node_modules/methods": {
"version": "1.1.2",
"resolved": "https://registry.npmjs.org/methods/-/methods-1.1.2.tgz",
"integrity": "sha512-iclAHeNqNm68zFtnZ0e+1L2yUIdvzNoauKU4WBA3VvH/vPFieF7qfRlwUZU+DA9P9bPXIS90ulxoUoCH23sV2w==",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/mime": {
"version": "1.6.0",
"resolved": "https://registry.npmjs.org/mime/-/mime-1.6.0.tgz",
"integrity": "sha512-x0Vn8spI+wuJ1O6S7gnbaQg8Pxh4NNHb7KSINmEWKiPE4RKOplvijn+NkmYmmRgP68mc70j2EbeTFRsrswaQeg==",
"bin": {
"mime": "cli.js"
},
"engines": {
"node": ">=4"
}
},
"node_modules/mime-db": {
"version": "1.52.0",
"resolved": "https://registry.npmjs.org/mime-db/-/mime-db-1.52.0.tgz",
"integrity": "sha512-sPU4uV7dYlvtWJxwwxHD0PuihVNiE7TyAbQ5SWxDCB9mUYvOgroQOwYQQOKPJ8CIbE+1ETVlOoK1UC2nU3gYvg==",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/mime-types": {
"version": "2.1.35",
"resolved": "https://registry.npmjs.org/mime-types/-/mime-types-2.1.35.tgz",
"integrity": "sha512-ZDY+bPm5zTTF+YpCrAU9nK0UgICYPT0QtT1NZWFv4s++TNkcgVaT0g6+4R2uI4MjQjzysHB1zxuWL50hzaeXiw==",
"dependencies": {
"mime-db": "1.52.0"
},
"engines": {
"node": ">= 0.6"
}
},
"node_modules/ms": {
"version": "2.0.0",
"resolved": "https://registry.npmjs.org/ms/-/ms-2.0.0.tgz",
"integrity": "sha512-Tpp60P6IUJDTuOq/5Z8cdskzJujfwqfOTkrwIwj7IRISpnkJnT6SyJ4PCPnGMoFjC9ddhal5KVIYtAt97ix05A=="
},
"node_modules/negotiator": {
"version": "0.6.3",
"resolved": "https://registry.npmjs.org/negotiator/-/negotiator-0.6.3.tgz",
"integrity": "sha512-+EUsqGPLsM+j/zdChZjsnX51g4XrHFOIXwfnCVPGlQk/k5giakcKsuxCObBRu6DSm9opw/O6slWbJdghQM4bBg==",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/object-inspect": {
"version": "1.13.4",
"resolved": "https://registry.npmjs.org/object-inspect/-/object-inspect-1.13.4.tgz",
"integrity": "sha512-W67iLl4J2EXEGTbfeHCffrjDfitvLANg0UlX3wFUUSTx92KXRFegMHUVgSqE+wvhAbi4WqjGg9czysTV2Epbew==",
"engines": {
"node": ">= 0.4"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/on-finished": {
"version": "2.4.1",
"resolved": "https://registry.npmjs.org/on-finished/-/on-finished-2.4.1.tgz",
"integrity": "sha512-oVlzkg3ENAhCk2zdv7IJwd/QUD4z2RxRwpkcGY8psCVcCYZNq4wYnVWALHM+brtuJjePWiYF/ClmuDr8Ch5+kg==",
"dependencies": {
"ee-first": "1.1.1"
},
"engines": {
"node": ">= 0.8"
}
},
"node_modules/parseurl": {
"version": "1.3.3",
"resolved": "https://registry.npmjs.org/parseurl/-/parseurl-1.3.3.tgz",
"integrity": "sha512-CiyeOxFT/JZyN5m0z9PfXw4SCBJ6Sygz1Dpl0wqjlhDEGGBP1GnsUVEL0p63hoG1fcj3fHynXi9NYO4nWOL+qQ==",
"engines": {
"node": ">= 0.8"
}
},
"node_modules/path-to-regexp": {
"version": "0.1.13",
"resolved": "https://registry.npmjs.org/path-to-regexp/-/path-to-regexp-0.1.13.tgz",
"integrity": "sha512-A/AGNMFN3c8bOlvV9RreMdrv7jsmF9XIfDeCd87+I8RNg6s78BhJxMu69NEMHBSJFxKidViTEdruRwEk/WIKqA=="
},
"node_modules/proxy-addr": {
"version": "2.0.7",
"resolved": "https://registry.npmjs.org/proxy-addr/-/proxy-addr-2.0.7.tgz",
"integrity": "sha512-llQsMLSUDUPT44jdrU/O37qlnifitDP+ZwrmmZcoSKyLKvtZxpyV0n2/bD/N4tBAAZ/gJEdZU7KMraoK1+XYAg==",
"dependencies": {
"forwarded": "0.2.0",
"ipaddr.js": "1.9.1"
},
"engines": {
"node": ">= 0.10"
}
},
"node_modules/qs": {
"version": "6.14.2",
"resolved": "https://registry.npmjs.org/qs/-/qs-6.14.2.tgz",
"integrity": "sha512-V/yCWTTF7VJ9hIh18Ugr2zhJMP01MY7c5kh4J870L7imm6/DIzBsNLTXzMwUA3yZ5b/KBqLx8Kp3uRvd7xSe3Q==",
"dependencies": {
"side-channel": "^1.1.0"
},
"engines": {
"node": ">=0.6"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/range-parser": {
"version": "1.2.1",
"resolved": "https://registry.npmjs.org/range-parser/-/range-parser-1.2.1.tgz",
"integrity": "sha512-Hrgsx+orqoygnmhFbKaHE6c296J+HTAQXoxEF6gNupROmmGJRoyzfG3ccAveqCBrwr/2yxQ5BVd/GTl5agOwSg==",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/raw-body": {
"version": "2.5.3",
"resolved": "https://registry.npmjs.org/raw-body/-/raw-body-2.5.3.tgz",
"integrity": "sha512-s4VSOf6yN0rvbRZGxs8Om5CWj6seneMwK3oDb4lWDH0UPhWcxwOWw5+qk24bxq87szX1ydrwylIOp2uG1ojUpA==",
"dependencies": {
"bytes": "~3.1.2",
"http-errors": "~2.0.1",
"iconv-lite": "~0.4.24",
"unpipe": "~1.0.0"
},
"engines": {
"node": ">= 0.8"
}
},
"node_modules/safe-buffer": {
"version": "5.2.1",
"resolved": "https://registry.npmjs.org/safe-buffer/-/safe-buffer-5.2.1.tgz",
"integrity": "sha512-rp3So07KcdmmKbGvgaNxQSJr7bGVSVk5S9Eq1F+ppbRo70+YeaDxkw5Dd8NPN+GD6bjnYm2VuPuCXmpuYvmCXQ==",
"funding": [
{
"type": "github",
"url": "https://github.com/sponsors/feross"
},
{
"type": "patreon",
"url": "https://www.patreon.com/feross"
},
{
"type": "consulting",
"url": "https://feross.org/support"
}
]
},
"node_modules/safer-buffer": {
"version": "2.1.2",
"resolved": "https://registry.npmjs.org/safer-buffer/-/safer-buffer-2.1.2.tgz",
"integrity": "sha512-YZo3K82SD7Riyi0E1EQPojLz7kpepnSQI9IyPbHHg1XXXevb5dJI7tpyN2ADxGcQbHG7vcyRHk0cbwqcQriUtg=="
},
"node_modules/send": {
"version": "0.19.2",
"resolved": "https://registry.npmjs.org/send/-/send-0.19.2.tgz",
"integrity": "sha512-VMbMxbDeehAxpOtWJXlcUS5E8iXh6QmN+BkRX1GARS3wRaXEEgzCcB10gTQazO42tpNIya8xIyNx8fll1OFPrg==",
"dependencies": {
"debug": "2.6.9",
"depd": "2.0.0",
"destroy": "1.2.0",
"encodeurl": "~2.0.0",
"escape-html": "~1.0.3",
"etag": "~1.8.1",
"fresh": "~0.5.2",
"http-errors": "~2.0.1",
"mime": "1.6.0",
"ms": "2.1.3",
"on-finished": "~2.4.1",
"range-parser": "~1.2.1",
"statuses": "~2.0.2"
},
"engines": {
"node": ">= 0.8.0"
}
},
"node_modules/send/node_modules/ms": {
"version": "2.1.3",
"resolved": "https://registry.npmjs.org/ms/-/ms-2.1.3.tgz",
"integrity": "sha512-6FlzubTLZG3J2a/NVCAleEhjzq5oxgHyaCU9yYXvcLsvoVaHJq/s5xXI6/XXP6tz7R9xAOtHnSO/tXtF3WRTlA=="
},
"node_modules/serve-static": {
"version": "1.16.3",
"resolved": "https://registry.npmjs.org/serve-static/-/serve-static-1.16.3.tgz",
"integrity": "sha512-x0RTqQel6g5SY7Lg6ZreMmsOzncHFU7nhnRWkKgWuMTu5NN0DR5oruckMqRvacAN9d5w6ARnRBXl9xhDCgfMeA==",
"dependencies": {
"encodeurl": "~2.0.0",
"escape-html": "~1.0.3",
"parseurl": "~1.3.3",
"send": "~0.19.1"
},
"engines": {
"node": ">= 0.8.0"
}
},
"node_modules/setprototypeof": {
"version": "1.2.0",
"resolved": "https://registry.npmjs.org/setprototypeof/-/setprototypeof-1.2.0.tgz",
"integrity": "sha512-E5LDX7Wrp85Kil5bhZv46j8jOeboKq5JMmYM3gVGdGH8xFpPWXUMsNrlODCrkoxMEeNi/XZIwuRvY4XNwYMJpw=="
},
"node_modules/side-channel": {
"version": "1.1.0",
"resolved": "https://registry.npmjs.org/side-channel/-/side-channel-1.1.0.tgz",
"integrity": "sha512-ZX99e6tRweoUXqR+VBrslhda51Nh5MTQwou5tnUDgbtyM0dBgmhEDtWGP/xbKn6hqfPRHujUNwz5fy/wbbhnpw==",
"dependencies": {
"es-errors": "^1.3.0",
"object-inspect": "^1.13.3",
"side-channel-list": "^1.0.0",
"side-channel-map": "^1.0.1",
"side-channel-weakmap": "^1.0.2"
},
"engines": {
"node": ">= 0.4"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/side-channel-list": {
"version": "1.0.0",
"resolved": "https://registry.npmjs.org/side-channel-list/-/side-channel-list-1.0.0.tgz",
"integrity": "sha512-FCLHtRD/gnpCiCHEiJLOwdmFP+wzCmDEkc9y7NsYxeF4u7Btsn1ZuwgwJGxImImHicJArLP4R0yX4c2KCrMrTA==",
"dependencies": {
"es-errors": "^1.3.0",
"object-inspect": "^1.13.3"
},
"engines": {
"node": ">= 0.4"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/side-channel-map": {
"version": "1.0.1",
"resolved": "https://registry.npmjs.org/side-channel-map/-/side-channel-map-1.0.1.tgz",
"integrity": "sha512-VCjCNfgMsby3tTdo02nbjtM/ewra6jPHmpThenkTYh8pG9ucZ/1P8So4u4FGBek/BjpOVsDCMoLA/iuBKIFXRA==",
"dependencies": {
"call-bound": "^1.0.2",
"es-errors": "^1.3.0",
"get-intrinsic": "^1.2.5",
"object-inspect": "^1.13.3"
},
"engines": {
"node": ">= 0.4"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/side-channel-weakmap": {
"version": "1.0.2",
"resolved": "https://registry.npmjs.org/side-channel-weakmap/-/side-channel-weakmap-1.0.2.tgz",
"integrity": "sha512-WPS/HvHQTYnHisLo9McqBHOJk2FkHO/tlpvldyrnem4aeQp4hai3gythswg6p01oSoTl58rcpiFAjF2br2Ak2A==",
"dependencies": {
"call-bound": "^1.0.2",
"es-errors": "^1.3.0",
"get-intrinsic": "^1.2.5",
"object-inspect": "^1.13.3",
"side-channel-map": "^1.0.1"
},
"engines": {
"node": ">= 0.4"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/statuses": {
"version": "2.0.2",
"resolved": "https://registry.npmjs.org/statuses/-/statuses-2.0.2.tgz",
"integrity": "sha512-DvEy55V3DB7uknRo+4iOGT5fP1slR8wQohVdknigZPMpMstaKJQWhwiYBACJE3Ul2pTnATihhBYnRhZQHGBiRw==",
"engines": {
"node": ">= 0.8"
}
},
"node_modules/toidentifier": {
"version": "1.0.1",
"resolved": "https://registry.npmjs.org/toidentifier/-/toidentifier-1.0.1.tgz",
"integrity": "sha512-o5sSPKEkg/DIQNmH43V0/uerLrpzVedkUh8tGNvaeXpfpuwjKenlSox/2O/BTlZUtEe+JG7s5YhEz608PlAHRA==",
"engines": {
"node": ">=0.6"
}
},
"node_modules/type-is": {
"version": "1.6.18",
"resolved": "https://registry.npmjs.org/type-is/-/type-is-1.6.18.tgz",
"integrity": "sha512-TkRKr9sUTxEH8MdfuCSP7VizJyzRNMjj2J2do2Jr3Kym598JVdEksuzPQCnlFPW4ky9Q+iA+ma9BGm06XQBy8g==",
"dependencies": {
"media-typer": "0.3.0",
"mime-types": "~2.1.24"
},
"engines": {
"node": ">= 0.6"
}
},
"node_modules/unpipe": {
"version": "1.0.0",
"resolved": "https://registry.npmjs.org/unpipe/-/unpipe-1.0.0.tgz",
"integrity": "sha512-pjy2bYhSsufwWlKwPc+l3cN7+wuJlK6uz0YdJEOlQDbl6jo/YlPi4mb8agUkVC8BF7V8NuzeyPNqRksA3hztKQ==",
"engines": {
"node": ">= 0.8"
}
},
"node_modules/utils-merge": {
"version": "1.0.1",
"resolved": "https://registry.npmjs.org/utils-merge/-/utils-merge-1.0.1.tgz",
"integrity": "sha512-pMZTvIkT1d+TFGvDOqodOclx0QWkkgi6Tdoa8gC8ffGAAqz9pzPTZWAybbsHHoED/ztMtkv/VoYTYyShUn81hA==",
"engines": {
"node": ">= 0.4.0"
}
},
"node_modules/vary": {
"version": "1.1.2",
"resolved": "https://registry.npmjs.org/vary/-/vary-1.1.2.tgz",
"integrity": "sha512-BNGbWLfd0eUPabhkXUVm0j8uuvREyTh5ovRa/dyow/BqAbZJyC+5fU+IzQOzmAKzYqYRAISoRhdQr3eIZ/PXqg==",
"engines": {
"node": ">= 0.8"
}
}
}
}

13
skipper-api/package.json Normal file
View File

@@ -0,0 +1,13 @@
{
"name": "skipper-api",
"version": "1.0.0",
"private": true,
"main": "src/index.js",
"scripts": {
"start": "node src/index.js"
},
"dependencies": {
"express": "^4.21.2"
}
}

472
skipper-api/src/index.js Normal file
View File

@@ -0,0 +1,472 @@
const express = require('express');
const { AppError } = require('../../shared/errors');
const { createContext } = require('../../shared/context');
const { verifyNodeToken } = require('../../shared/auth');
const { emitEvent } = require('../../shared/events');
const { createLogger } = require('../../shared/logs');
const { bootstrapDataLayout } = require('../../shared/bootstrap');
const { createSnapshot, getLatestSystemSnapshot, getLatestTenantSnapshot } = require('../../shared/snapshots');
const { createResourceDocument, getResource, listAllResources, saveResource, patchResourceState } = require('../../shared/resources');
const { getIdempotencyRecord, saveIdempotencyRecord } = require('../../shared/idempotency');
const { claimNextWorkOrder, createWorkOrder, finishWorkOrder, getWorkOrder, listWorkOrders } = require('../../shared/work-orders');
const { dataDir } = require('../../shared/paths');
const PORT = Number(process.env.PORT || 3000);
const HOST = process.env.HOST || '0.0.0.0';
const ADMIN_TOKEN = process.env.ADMIN_TOKEN || 'dev-admin-token';
const app = express();
function envelope(context, data, error) {
return {
schema_version: 'v1',
request_id: context.request_id,
correlation_id: context.correlation_id,
data: data === undefined ? null : data,
error: error || null,
metadata: {
timestamp: new Date().toISOString(),
},
};
}
function parseBearerToken(req) {
const value = req.headers.authorization || '';
if (!value.startsWith('Bearer ')) {
return null;
}
return value.slice('Bearer '.length).trim();
}
function asyncHandler(handler) {
return (req, res, next) => {
Promise.resolve(handler(req, res, next)).catch(next);
};
}
function requireAdmin(req, res, next) {
if (req.headers['x-admin-token'] !== ADMIN_TOKEN) {
return next(new AppError(401, 'UNAUTHORIZED', 'Admin token required'));
}
return next();
}
async function requireNodeAuth(req, res, next) {
let nodeId = req.params.nodeId;
if (!nodeId && req.params.workOrderId) {
const workOrder = await getWorkOrder(req.params.workOrderId);
nodeId = workOrder && workOrder.target ? workOrder.target.node_id : null;
}
const token = parseBearerToken(req);
if (!(await verifyNodeToken(nodeId, token))) {
return next(new AppError(401, 'UNAUTHORIZED', 'Node token invalid', { node_id: nodeId }));
}
return next();
}
app.use(express.json({ limit: '2mb' }));
app.use((req, res, next) => {
req.context = createContext({
request_id: req.headers['x-request-id'] || (req.body && req.body.request_id),
correlation_id: req.headers['x-correlation-id'] || (req.body && req.body.correlation_id),
});
res.setHeader('x-request-id', req.context.request_id);
res.setHeader('x-correlation-id', req.context.correlation_id);
next();
});
app.use((req, res, next) => {
const startedAt = Date.now();
res.on('finish', () => {
createLogger({ service: 'skipper' }).info('http.request', res.statusCode < 400 ? 'success' : 'error', {
method: req.method,
path: req.path,
status_code: res.statusCode,
duration_ms: Date.now() - startedAt,
request_id: req.context.request_id,
correlation_id: req.context.correlation_id,
}).catch(() => {});
});
next();
});
app.get('/health', asyncHandler(async (req, res) => {
res.json(envelope(req.context, { ok: true, data_dir: dataDir }));
}));
app.get('/v1/health', asyncHandler(async (req, res) => {
res.json(envelope(req.context, { ok: true, data_dir: dataDir }));
}));
app.get('/v1/resources', requireAdmin, asyncHandler(async (req, res) => {
res.json(envelope(req.context, await listAllResources()));
}));
app.get('/v1/resources/:resourceType/:resourceId', requireAdmin, asyncHandler(async (req, res) => {
const resource = await getResource(req.params.resourceType, req.params.resourceId);
if (!resource) {
throw new AppError(404, 'RESOURCE_NOT_FOUND', 'Resource not found', {
resource_type: req.params.resourceType,
resource_id: req.params.resourceId,
});
}
res.json(envelope(req.context, resource));
}));
app.get('/v1/work-orders', requireAdmin, asyncHandler(async (req, res) => {
res.json(envelope(req.context, await listWorkOrders()));
}));
app.get('/v1/work-orders/:workOrderId', requireAdmin, asyncHandler(async (req, res) => {
const workOrder = await getWorkOrder(req.params.workOrderId);
if (!workOrder) {
throw new AppError(404, 'RESOURCE_NOT_FOUND', 'Work order not found', {
resource_type: 'work_order',
resource_id: req.params.workOrderId,
});
}
res.json(envelope(req.context, workOrder));
}));
app.post('/v1/nodes/:nodeId/heartbeat', requireNodeAuth, asyncHandler(async (req, res) => {
const nodeId = req.params.nodeId;
const payload = req.body.data || {};
const node = await getResource('node', nodeId);
if (!node) {
throw new AppError(404, 'RESOURCE_NOT_FOUND', 'Node not found', {
resource_type: 'node',
resource_id: nodeId,
});
}
const updated = await saveResource({
...node,
current_state: {
...node.current_state,
heartbeat_at: new Date().toISOString(),
hostname: payload.hostname || node.current_state.hostname || nodeId,
capabilities: Array.isArray(payload.capabilities) ? payload.capabilities : node.current_state.capabilities || [],
agent_version: payload.agent_version || node.current_state.agent_version || null,
},
});
await emitEvent({
type: 'node_heartbeat_received',
resource_type: 'node',
resource_id: nodeId,
request_id: req.context.request_id,
correlation_id: req.context.correlation_id,
payload: {
hostname: updated.current_state.hostname,
capabilities: updated.current_state.capabilities || [],
},
});
res.json(envelope(req.context, updated));
}));
app.get('/v1/nodes/:nodeId/work-orders/next', requireNodeAuth, asyncHandler(async (req, res) => {
const workOrder = await claimNextWorkOrder(req.params.nodeId, req.context);
if (workOrder) {
await emitEvent({
type: 'work_order_started',
resource_type: 'work_order',
resource_id: workOrder.id,
request_id: req.context.request_id,
correlation_id: req.context.correlation_id,
payload: {
node_id: req.params.nodeId,
tenant_id: workOrder.target.tenant_id || null,
status: 'running',
},
});
if (workOrder.desired_state.deployment_id) {
await emitEvent({
type: 'deployment_started',
resource_type: 'deployment',
resource_id: workOrder.desired_state.deployment_id,
request_id: req.context.request_id,
correlation_id: req.context.correlation_id,
payload: {
node_id: req.params.nodeId,
work_order_id: workOrder.id,
},
});
}
}
res.json(envelope(req.context, workOrder));
}));
app.post('/v1/work-orders/:workOrderId/result', requireNodeAuth, asyncHandler(async (req, res) => {
const workOrderId = req.params.workOrderId;
const payload = req.body.data || {};
const existing = await getWorkOrder(workOrderId);
if (!existing) {
throw new AppError(404, 'RESOURCE_NOT_FOUND', 'Work order not found', {
resource_type: 'work_order',
resource_id: workOrderId,
});
}
const status = payload.status === 'success' ? 'success' : 'failed';
const completed = await finishWorkOrder(workOrderId, status, payload.result || null);
if (!completed) {
throw new AppError(409, 'WORK_ORDER_NOT_CLAIMABLE', 'Work order is not claimable', {
resource_id: workOrderId,
});
}
for (const stateUpdate of payload.state_report && Array.isArray(payload.state_report.resources)
? payload.state_report.resources
: []) {
await patchResourceState(stateUpdate.resource_type, stateUpdate.resource_id, {
current_state: stateUpdate.current_state,
last_applied_state: stateUpdate.last_applied_state,
});
}
if (completed.desired_state.deployment_id) {
const deployment = await getResource('deployment', completed.desired_state.deployment_id);
if (deployment) {
await saveResource({
...deployment,
current_state: {
...deployment.current_state,
status,
work_order_id: completed.id,
finished_at: completed.finished_at,
},
last_applied_state: completed.desired_state,
});
}
}
await emitEvent({
type: status === 'success' ? 'work_order_succeeded' : 'work_order_failed',
resource_type: 'work_order',
resource_id: completed.id,
request_id: req.context.request_id,
correlation_id: req.context.correlation_id,
payload: {
node_id: completed.target.node_id,
tenant_id: completed.target.tenant_id || null,
result: completed.result,
},
});
if (completed.desired_state.deployment_id) {
await emitEvent({
type: status === 'success' ? 'deployment_succeeded' : 'deployment_failed',
resource_type: 'deployment',
resource_id: completed.desired_state.deployment_id,
request_id: req.context.request_id,
correlation_id: req.context.correlation_id,
payload: {
node_id: completed.target.node_id,
work_order_id: completed.id,
result: completed.result,
},
});
}
res.json(envelope(req.context, completed));
}));
app.post('/v1/deployments/:tenantId/apply', requireAdmin, asyncHandler(async (req, res) => {
const tenantId = req.params.tenantId;
const idempotencyKey = req.headers['x-idempotency-key'] || (req.body && req.body.idempotency_key);
const tenant = await getResource('tenant', tenantId);
if (!tenant) {
throw new AppError(404, 'RESOURCE_NOT_FOUND', 'Tenant not found', {
resource_type: 'tenant',
resource_id: tenantId,
});
}
if (!idempotencyKey) {
throw new AppError(400, 'INVALID_REQUEST', 'Idempotency key required');
}
const replay = await getIdempotencyRecord('deployment_apply', idempotencyKey);
if (replay) {
return res.json(envelope(req.context, replay.value));
}
const targetNodeId = tenant.desired_state.deployment_policy && tenant.desired_state.deployment_policy.target_node_id;
const composeSpec = tenant.desired_state.compose;
if (!targetNodeId || !composeSpec || !composeSpec.compose_file) {
throw new AppError(400, 'INVALID_REQUEST', 'Tenant desired state is incomplete', {
tenant_id: tenant.id,
});
}
const deployment = await saveResource(createResourceDocument({
id: `deployment-${Date.now()}`,
resource_type: 'deployment',
desired_state: {
tenant_id: tenant.id,
service_ids: tenant.desired_state.service_ids || [],
target_node_id: targetNodeId,
strategy: 'apply',
compose: composeSpec,
},
current_state: {
status: 'pending',
},
last_applied_state: {},
metadata: {},
}));
const workOrder = await createWorkOrder({
type: 'deploy_service',
target: {
tenant_id: tenant.id,
node_id: targetNodeId,
},
desired_state: {
deployment_id: deployment.id,
tenant_id: tenant.id,
service_ids: tenant.desired_state.service_ids || [],
compose_project: composeSpec,
},
request_id: req.context.request_id,
correlation_id: req.context.correlation_id,
metadata: {},
});
await emitEvent({
type: 'resource_created',
resource_type: 'deployment',
resource_id: deployment.id,
request_id: req.context.request_id,
correlation_id: req.context.correlation_id,
payload: {
tenant_id: tenant.id,
target_node_id: targetNodeId,
},
});
await emitEvent({
type: 'work_order_created',
resource_type: 'work_order',
resource_id: workOrder.id,
request_id: req.context.request_id,
correlation_id: req.context.correlation_id,
payload: {
tenant_id: tenant.id,
node_id: targetNodeId,
deployment_id: deployment.id,
},
});
const responseData = {
deployment,
work_order: workOrder,
};
await saveIdempotencyRecord('deployment_apply', idempotencyKey, responseData);
res.status(201).json(envelope(req.context, responseData));
}));
app.get('/v1/snapshots/system/latest', requireAdmin, asyncHandler(async (req, res) => {
const snapshot = (await getLatestSystemSnapshot()) || (await createSnapshot('system', req.context));
await emitEvent({
type: 'snapshot_created',
resource_type: 'snapshot',
resource_id: snapshot.snapshot_id,
request_id: req.context.request_id,
correlation_id: req.context.correlation_id,
payload: {
scope: 'system',
},
});
res.json(envelope(req.context, snapshot));
}));
app.get('/v1/snapshots/tenants/:tenantId/latest', requireAdmin, asyncHandler(async (req, res) => {
const scope = `tenant:${req.params.tenantId}`;
const snapshot = (await getLatestTenantSnapshot(req.params.tenantId)) || (await createSnapshot(scope, req.context));
await emitEvent({
type: 'snapshot_created',
resource_type: 'snapshot',
resource_id: snapshot.snapshot_id,
request_id: req.context.request_id,
correlation_id: req.context.correlation_id,
payload: {
scope,
},
});
res.json(envelope(req.context, snapshot));
}));
app.use((error, req, res, next) => {
createLogger({
service: 'skipper',
request_id: req.context ? req.context.request_id : null,
correlation_id: req.context ? req.context.correlation_id : null,
}).error('http.error', 'failed', {
code: error.code || 'INTERNAL_ERROR',
message: error.message,
details: error.details || {},
}).catch(() => {});
if (error instanceof AppError) {
return res.status(error.statusCode).json(envelope(req.context || createContext(), null, {
code: error.code,
message: error.message,
details: error.details,
}));
}
return res.status(500).json(envelope(req.context || createContext(), null, {
code: 'INTERNAL_ERROR',
message: 'Internal server error',
details: {},
}));
});
async function start() {
await bootstrapDataLayout();
app.listen(PORT, HOST, () => {
createLogger({ service: 'skipper' }).info('server.start', 'success', {
host: HOST,
port: PORT,
data_dir: dataDir,
}).catch(() => {});
});
}
start().catch((error) => {
process.stderr.write(`${error.stack || error.message}\n`);
process.exit(1);
});

20
skippy-agent/Dockerfile Normal file
View File

@@ -0,0 +1,20 @@
FROM node:20-alpine
RUN apk add --no-cache docker-cli docker-cli-compose
WORKDIR /app/skippy-agent
COPY skippy-agent/package.json ./package.json
RUN npm install --omit=dev
COPY skippy-agent/src ./src
COPY shared /app/shared
ENV DATA_DIR=/app/data
ENV SKIPPER_URL=http://skipper-api:3000
ENV AGENT_ID=host-1
ENV POLL_INTERVAL_MS=5000
ENV HEARTBEAT_INTERVAL_MS=15000
CMD ["npm", "start"]

12
skippy-agent/package-lock.json generated Normal file
View File

@@ -0,0 +1,12 @@
{
"name": "skippy-agent",
"version": "1.0.0",
"lockfileVersion": 3,
"requires": true,
"packages": {
"": {
"name": "skippy-agent",
"version": "1.0.0"
}
}
}

10
skippy-agent/package.json Normal file
View File

@@ -0,0 +1,10 @@
{
"name": "skippy-agent",
"version": "1.0.0",
"private": true,
"main": "src/index.js",
"scripts": {
"start": "node src/index.js"
}
}

238
skippy-agent/src/index.js Normal file
View File

@@ -0,0 +1,238 @@
const fs = require('fs/promises');
const path = require('path');
const os = require('os');
const { createContext } = require('../../shared/context');
const { createLogger } = require('../../shared/logs');
const { getJson, postJson } = require('./lib/http');
const docker = require('./modules/docker');
const agentId = process.env.AGENT_ID || 'host-1';
const agentToken = process.env.AGENT_TOKEN || 'dev-node-token';
const skipperUrl = (process.env.SKIPPER_URL || 'http://localhost:3000').replace(/\/$/, '');
const pollIntervalMs = Number(process.env.POLL_INTERVAL_MS || 5000);
const heartbeatIntervalMs = Number(process.env.HEARTBEAT_INTERVAL_MS || 15000);
const logger = createLogger({ service: 'skippy', node_id: agentId });
let isPolling = false;
async function sendHeartbeat() {
const context = createContext();
await postJson(`${skipperUrl}/v1/nodes/${agentId}/heartbeat`, {
hostname: os.hostname(),
capabilities: ['deploy_service'],
agent_version: '1.0.0',
}, agentToken, context);
await logger.info('node.heartbeat', 'success', context);
}
function buildComposePath(workOrder) {
return path.join(docker.composeBaseDir, workOrder.desired_state.compose_project.tenant_id, 'docker-compose.yml');
}
async function executeDeployService(workOrder) {
const composeProject = workOrder.desired_state.compose_project;
const composePath = buildComposePath(workOrder);
await fs.mkdir(path.dirname(composePath), { recursive: true });
await fs.writeFile(composePath, composeProject.compose_file, 'utf8');
const commandResult = await docker.applyCompose({
path: composePath,
env: composeProject.env || {},
});
const status = commandResult.code === 0 ? 'success' : 'failed';
return {
status,
result: {
success: status === 'success',
code: status === 'success' ? 'APPLY_OK' : 'STATE_APPLY_FAILED',
message: status === 'success' ? 'Desired state applied' : 'Desired state apply failed',
details: {
duration_ms: 0,
compose_path: composePath,
changed_resources: workOrder.desired_state.service_ids || [],
unchanged_resources: [],
command: {
program: 'docker',
args: ['compose', '-f', composePath, 'up', '-d'],
exit_code: commandResult.code,
stdout: commandResult.stdout.trim(),
stderr: commandResult.stderr.trim(),
},
},
},
state_report: {
resources: [
{
resource_type: 'deployment',
resource_id: workOrder.desired_state.deployment_id,
current_state: {
status,
observed_at: new Date().toISOString(),
node_id: agentId,
compose_path: composePath,
},
last_applied_state: workOrder.desired_state,
},
{
resource_type: 'tenant',
resource_id: workOrder.target.tenant_id,
current_state: {
deployed_on_node_id: agentId,
last_deployment_status: status,
observed_at: new Date().toISOString(),
},
last_applied_state: {
deployment_id: workOrder.desired_state.deployment_id,
compose_project: {
tenant_id: composeProject.tenant_id,
env: composeProject.env || {},
},
},
},
],
},
};
}
async function executeWorkOrder(workOrder) {
const startedAt = Date.now();
const context = createContext({
request_id: workOrder.request_id,
correlation_id: workOrder.correlation_id,
});
await logger.info('work_order.execute', 'started', {
...context,
work_order_id: workOrder.id,
type: workOrder.type,
tenant_id: workOrder.target.tenant_id || null,
});
try {
let execution;
switch (workOrder.type) {
case 'deploy_service':
execution = await executeDeployService(workOrder);
break;
default:
execution = {
status: 'failed',
result: {
success: false,
code: 'STATE_APPLY_FAILED',
message: `Unsupported work order type: ${workOrder.type}`,
details: {
duration_ms: 0,
changed_resources: [],
unchanged_resources: [],
},
},
state_report: { resources: [] },
};
break;
}
execution.result.details.duration_ms = Date.now() - startedAt;
await logger.info('work_order.execute', execution.status, {
...context,
work_order_id: workOrder.id,
type: workOrder.type,
tenant_id: workOrder.target.tenant_id || null,
code: execution.result.code,
});
return {
...execution,
context,
};
} catch (error) {
await logger.error('work_order.execute', 'failed', {
...context,
work_order_id: workOrder.id,
type: workOrder.type,
tenant_id: workOrder.target.tenant_id || null,
message: error.message,
code: 'STATE_APPLY_FAILED',
});
return {
status: 'failed',
result: {
success: false,
code: 'STATE_APPLY_FAILED',
message: error.message,
details: {
duration_ms: Date.now() - startedAt,
changed_resources: [],
unchanged_resources: [],
},
},
state_report: {
resources: [],
},
context,
};
}
}
async function pollOnce() {
if (isPolling) {
return;
}
isPolling = true;
try {
const workOrder = await getJson(`${skipperUrl}/v1/nodes/${agentId}/work-orders/next`, agentToken);
if (!workOrder) {
return;
}
const execution = await executeWorkOrder(workOrder);
await postJson(`${skipperUrl}/v1/work-orders/${workOrder.id}/result`, {
status: execution.status,
result: execution.result,
state_report: execution.state_report,
}, agentToken, execution.context);
} finally {
isPolling = false;
}
}
async function start() {
await sendHeartbeat();
setInterval(() => {
sendHeartbeat().catch(async (error) => {
await logger.error('node.heartbeat', 'failed', {
node_id: agentId,
message: error.message,
});
});
}, heartbeatIntervalMs);
setInterval(() => {
pollOnce().catch(async (error) => {
await logger.error('work_order.poll', 'failed', {
node_id: agentId,
message: error.message,
});
});
}, pollIntervalMs);
await pollOnce();
}
start().catch(async (error) => {
await logger.error('agent.start', 'failed', {
node_id: agentId,
message: error.message,
});
process.exit(1);
});

View File

@@ -0,0 +1,74 @@
const { createContext } = require('../../../shared/context');
async function parseJson(response) {
const text = await response.text();
if (!text) {
return null;
}
return JSON.parse(text);
}
function buildHeaders(context, token) {
return {
'content-type': 'application/json',
'x-request-id': context.request_id,
'x-correlation-id': context.correlation_id,
authorization: `Bearer ${token}`,
};
}
function unwrapEnvelope(payload) {
if (!payload) {
return null;
}
if (payload.error) {
const error = new Error(payload.error.message);
error.code = payload.error.code;
error.details = payload.error.details;
throw error;
}
return payload.data;
}
async function getJson(url, token) {
const context = createContext();
const response = await fetch(url, {
headers: buildHeaders(context, token),
});
if (!response.ok) {
const body = await response.text();
throw new Error(`GET ${url} failed: ${response.status} ${body}`);
}
return unwrapEnvelope(await parseJson(response));
}
async function postJson(url, body, token, requestContext) {
const context = createContext(requestContext);
const response = await fetch(url, {
method: 'POST',
headers: buildHeaders(context, token),
body: JSON.stringify({
request_id: context.request_id,
correlation_id: context.correlation_id,
data: body,
}),
});
if (!response.ok) {
const text = await response.text();
throw new Error(`POST ${url} failed: ${response.status} ${text}`);
}
return unwrapEnvelope(await parseJson(response));
}
module.exports = {
getJson,
postJson,
};

View File

@@ -0,0 +1,51 @@
const fs = require('fs/promises');
const path = require('path');
const { spawn } = require('child_process');
const composeBaseDir = process.env.SKIPPY_COMPOSE_BASE_DIR || '/opt/skipper/tenants';
async function writeEnvFile(filePath, env) {
const entries = Object.entries(env || {}).map(([key, value]) => `${key}=${String(value)}`);
await fs.writeFile(filePath, `${entries.join('\n')}\n`);
}
function runCommand(command, args, options) {
return new Promise((resolve, reject) => {
const child = spawn(command, args, options);
let stdout = '';
let stderr = '';
child.stdout.on('data', (chunk) => {
stdout += chunk.toString();
});
child.stderr.on('data', (chunk) => {
stderr += chunk.toString();
});
child.on('error', reject);
child.on('close', (code) => {
resolve({ code, stdout, stderr });
});
});
}
async function applyCompose({ path: composePath, env }) {
const cwd = path.dirname(composePath);
await fs.mkdir(cwd, { recursive: true });
await writeEnvFile(path.join(cwd, '.env'), env || {});
return runCommand('docker', ['compose', '-f', composePath, 'up', '-d'], {
cwd,
env: {
...process.env,
...Object.fromEntries(Object.entries(env || {}).map(([key, value]) => [key, String(value)])),
},
});
}
module.exports = {
composeBaseDir,
applyCompose,
};