Compare commits
20 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| 3e4558abc5 | |||
| 3e0eb5e198 | |||
| 732e9e5cac | |||
| 5bf1779243 | |||
| 4908c21b84 | |||
| b1e2f0d8ea | |||
| d815915135 | |||
| 0f445b4c86 | |||
| 8f0514d10e | |||
| e1cf1768da | |||
| 4d32d5e71e | |||
| a4552498ac | |||
| 4585801f32 | |||
| 3dc75f5cda | |||
| 7591e0ed90 | |||
| d2c2a4c4dd | |||
| 89cd93cdff | |||
| 10aee5d4c5 | |||
| 53b7bd7048 | |||
| 101c4286c1 |
@@ -1,6 +0,0 @@
|
|||||||
FROM hosttoday/ht-docker-node:npmci
|
|
||||||
RUN yarn global add @git.zone/tsdocker
|
|
||||||
COPY ./ /workspace
|
|
||||||
WORKDIR /workspace
|
|
||||||
ENV CI=true
|
|
||||||
CMD ["tsdocker","runinside"];
|
|
||||||
71
changelog.md
71
changelog.md
@@ -1,5 +1,76 @@
|
|||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
|
## 2026-03-15 - 2.1.0 - feat(cli)
|
||||||
|
add global remote builder configuration and native SSH buildx nodes for multi-platform builds
|
||||||
|
|
||||||
|
- adds a new `tsdocker config` command with subcommands to add, remove, list, and show remote builder definitions
|
||||||
|
- introduces global config support for remote builders stored under `~/.git.zone/tsdocker/config.json`
|
||||||
|
- builds can now create multi-node buildx setups with remote SSH builders and open reverse SSH tunnels so remote nodes can push to the local staging registry
|
||||||
|
- updates the README and CLI help to document remote builder configuration and native cross-platform build workflows
|
||||||
|
|
||||||
|
## 2026-03-12 - 2.0.2 - fix(repo)
|
||||||
|
no changes to commit
|
||||||
|
|
||||||
|
|
||||||
|
## 2026-03-12 - 2.0.1 - fix(repository)
|
||||||
|
no changes to commit
|
||||||
|
|
||||||
|
|
||||||
|
## 2026-03-12 - 2.0.0 - BREAKING CHANGE(cli)
|
||||||
|
remove legacy container test runner and make the default command show the man page
|
||||||
|
|
||||||
|
- Removes legacy testing and VS Code commands, including `runinside`, `vscode`, generated Dockerfile assets, and related configuration fields (`baseImage`, `command`, `dockerSock`, `keyValueObject`)
|
||||||
|
- Simplifies configuration and dependencies by dropping qenv-based env loading and unused legacy packages
|
||||||
|
- Updates CLI and documentation to reflect default help output and the current build/push-focused workflow
|
||||||
|
|
||||||
|
## 2026-02-07 - 1.17.4 - fix()
|
||||||
|
no changes
|
||||||
|
|
||||||
|
|
||||||
|
## 2026-02-07 - 1.17.3 - fix(registry)
|
||||||
|
increase default maxRetries in fetchWithRetry from 3 to 6 to improve resilience when fetching registry resources
|
||||||
|
|
||||||
|
- Changed default maxRetries from 3 to 6 in ts/classes.registrycopy.ts
|
||||||
|
- Reduces failures from transient network or registry errors by allowing more retry attempts
|
||||||
|
- No API or behavior changes besides the increased default retry count
|
||||||
|
|
||||||
|
## 2026-02-07 - 1.17.2 - fix(registry)
|
||||||
|
improve HTTP fetch retry logging, backoff calculation, and token-cache warning
|
||||||
|
|
||||||
|
- Include HTTP method in logs and normalize method to uppercase for consistency
|
||||||
|
- Log retry attempts with method, URL and calculated exponential backoff delay
|
||||||
|
- Compute and reuse exponential backoff delay variable instead of inline calculation
|
||||||
|
- Log error when a 5xx response persists after all retry attempts and when fetch ultimately fails
|
||||||
|
- Add a warning log when clearing cached token after a 401 response
|
||||||
|
|
||||||
|
## 2026-02-07 - 1.17.1 - fix(registrycopy)
|
||||||
|
add fetchWithRetry wrapper to apply timeouts, retries with exponential backoff, and token cache handling; use it for registry HTTP requests
|
||||||
|
|
||||||
|
- Introduces fetchWithRetry(url, options, timeoutMs, maxRetries) to wrap fetch with AbortSignal timeout, exponential backoff retries, and retry behavior only for network errors and 5xx responses
|
||||||
|
- Replaces direct fetch calls for registry /v2 checks, token requests, and blob uploads with fetchWithRetry (30s for auth/token checks, 300s for blob operations)
|
||||||
|
- Clears token cache entry when a 401 response is received so the next attempt re-authenticates
|
||||||
|
- Adds logging on retry attempts and backoff delays to improve robustness and observability
|
||||||
|
|
||||||
|
## 2026-02-07 - 1.17.0 - feat(tsdocker)
|
||||||
|
add Dockerfile filtering, optional skip-build flow, and fallback Docker config credential loading
|
||||||
|
|
||||||
|
- Add TsDockerManager.filterDockerfiles(patterns) to filter discovered Dockerfiles by glob-style patterns and warn when no matches are found
|
||||||
|
- Allow skipping image build with --no-build (argvArg.build === false): discover Dockerfiles and apply filters without performing build
|
||||||
|
- Fallback to load Docker registry credentials from ~/.docker/config.json via RegistryCopy.getDockerConfigCredentials when env vars do not provide credentials
|
||||||
|
- Import RegistryCopy and add info/warn logs when credentials are loaded or missing
|
||||||
|
|
||||||
|
## 2026-02-07 - 1.16.0 - feat(core)
|
||||||
|
Introduce per-invocation TsDockerSession and session-aware local registry and build orchestration; stream and parse buildx output for improved logging and visibility; detect Docker topology and add CI-safe cleanup; update README with multi-arch, parallel-build, caching, and local registry usage and new CLI flags.
|
||||||
|
|
||||||
|
- Add TsDockerSession to allocate unique ports, container names and builder suffixes for concurrent runs (especially in CI).
|
||||||
|
- Make local registry session-aware: start/stop/use registry container and persistent storage per session; retry on port conflicts.
|
||||||
|
- Inject session into Dockerfile instances and TsDockerManager; use session.config.registryHost for tagging/pushing and test container naming.
|
||||||
|
- Stream and parse buildx/docker build output via createBuildOutputHandler for clearer step/platform/CACHED/DONE logging and --progress=plain usage.
|
||||||
|
- Detect Docker topology (socket-mount, dind, local) in DockerContext and expose it in context info.
|
||||||
|
- Add manager.cleanup to remove CI-scoped buildx builders and ensure CLI calls cleanup after build/push/test.
|
||||||
|
- Update interfaces to include topology and adjust many Dockerfile/manager methods to be session-aware.
|
||||||
|
- Large README improvements: multi-arch flow, persistent local registry, parallel builds, caching, new CLI and clean flags, and examples for CI integration.
|
||||||
|
|
||||||
## 2026-02-07 - 1.15.1 - fix(registry)
|
## 2026-02-07 - 1.15.1 - fix(registry)
|
||||||
use persistent local registry and OCI Distribution API image copy for pushes
|
use persistent local registry and OCI Distribution API image copy for pushes
|
||||||
|
|
||||||
|
|||||||
28
package.json
28
package.json
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@git.zone/tsdocker",
|
"name": "@git.zone/tsdocker",
|
||||||
"version": "1.15.1",
|
"version": "2.1.0",
|
||||||
"private": false,
|
"private": false,
|
||||||
"description": "develop npm modules cross platform with docker",
|
"description": "develop npm modules cross platform with docker",
|
||||||
"main": "dist_ts/index.js",
|
"main": "dist_ts/index.js",
|
||||||
@@ -11,13 +11,6 @@
|
|||||||
"scripts": {
|
"scripts": {
|
||||||
"test": "(npm run build)",
|
"test": "(npm run build)",
|
||||||
"build": "(tsbuild)",
|
"build": "(tsbuild)",
|
||||||
"testIntegration": "(npm run clean && npm run setupCheck && npm run testStandard)",
|
|
||||||
"testStandard": "(cd test/ && tsx ../ts/index.ts)",
|
|
||||||
"testClean": "(cd test/ && tsx ../ts/index.ts clean --all)",
|
|
||||||
"testVscode": "(cd test/ && tsx ../ts/index.ts vscode)",
|
|
||||||
"clean": "(rm -rf test/)",
|
|
||||||
"compile": "(npmts --notest)",
|
|
||||||
"setupCheck": "(git clone https://gitlab.com/sandboxzone/sandbox-npmts.git test/)",
|
|
||||||
"buildDocs": "tsdoc"
|
"buildDocs": "tsdoc"
|
||||||
},
|
},
|
||||||
"repository": {
|
"repository": {
|
||||||
@@ -34,27 +27,22 @@
|
|||||||
},
|
},
|
||||||
"homepage": "https://gitlab.com/gitzone/tsdocker#readme",
|
"homepage": "https://gitlab.com/gitzone/tsdocker#readme",
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
"@git.zone/tsbuild": "^4.1.2",
|
"@git.zone/tsbuild": "^4.3.0",
|
||||||
"@git.zone/tsrun": "^2.0.1",
|
"@git.zone/tsrun": "^2.0.1",
|
||||||
"@git.zone/tstest": "^3.1.6",
|
"@git.zone/tstest": "^3.3.2",
|
||||||
"@types/node": "^25.0.9"
|
"@types/node": "^25.5.0"
|
||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@push.rocks/lik": "^6.2.2",
|
"@push.rocks/lik": "^6.3.1",
|
||||||
"@push.rocks/npmextra": "^5.3.3",
|
"@push.rocks/npmextra": "^5.3.3",
|
||||||
"@push.rocks/projectinfo": "^5.0.2",
|
"@push.rocks/projectinfo": "^5.0.2",
|
||||||
"@push.rocks/qenv": "^6.1.3",
|
|
||||||
"@push.rocks/smartanalytics": "^2.0.15",
|
|
||||||
"@push.rocks/smartcli": "^4.0.20",
|
"@push.rocks/smartcli": "^4.0.20",
|
||||||
"@push.rocks/smartfs": "^1.3.1",
|
"@push.rocks/smartfs": "^1.5.0",
|
||||||
"@push.rocks/smartinteract": "^2.0.16",
|
"@push.rocks/smartinteract": "^2.0.16",
|
||||||
"@push.rocks/smartlog": "^3.1.10",
|
"@push.rocks/smartlog": "^3.2.1",
|
||||||
"@push.rocks/smartlog-destination-local": "^9.0.2",
|
"@push.rocks/smartlog-destination-local": "^9.0.2",
|
||||||
"@push.rocks/smartlog-source-ora": "^1.0.9",
|
"@push.rocks/smartlog-source-ora": "^1.0.9",
|
||||||
"@push.rocks/smartopen": "^2.0.0",
|
"@push.rocks/smartshell": "^3.3.7"
|
||||||
"@push.rocks/smartpromise": "^4.2.3",
|
|
||||||
"@push.rocks/smartshell": "^3.3.0",
|
|
||||||
"@push.rocks/smartstring": "^4.1.0"
|
|
||||||
},
|
},
|
||||||
"packageManager": "pnpm@10.18.1+sha512.77a884a165cbba2d8d1c19e3b4880eee6d2fcabd0d879121e282196b80042351d5eb3ca0935fa599da1dc51265cc68816ad2bddd2a2de5ea9fdf92adbec7cd34",
|
"packageManager": "pnpm@10.18.1+sha512.77a884a165cbba2d8d1c19e3b4880eee6d2fcabd0d879121e282196b80042351d5eb3ca0935fa599da1dc51265cc68816ad2bddd2a2de5ea9fdf92adbec7cd34",
|
||||||
"type": "module",
|
"type": "module",
|
||||||
|
|||||||
5595
pnpm-lock.yaml
generated
5595
pnpm-lock.yaml
generated
File diff suppressed because it is too large
Load Diff
@@ -3,7 +3,6 @@
|
|||||||
## Module Purpose
|
## Module Purpose
|
||||||
|
|
||||||
tsdocker is a comprehensive Docker development and building tool. It provides:
|
tsdocker is a comprehensive Docker development and building tool. It provides:
|
||||||
- Testing npm modules in clean Docker environments (legacy feature)
|
|
||||||
- Building Dockerfiles with dependency ordering
|
- Building Dockerfiles with dependency ordering
|
||||||
- Multi-registry push/pull support
|
- Multi-registry push/pull support
|
||||||
- Multi-architecture builds (amd64/arm64)
|
- Multi-architecture builds (amd64/arm64)
|
||||||
@@ -12,7 +11,7 @@ tsdocker is a comprehensive Docker development and building tool. It provides:
|
|||||||
|
|
||||||
| Command | Description |
|
| Command | Description |
|
||||||
|---------|-------------|
|
|---------|-------------|
|
||||||
| `tsdocker` | Run tests in container (legacy default behavior) |
|
| `tsdocker` | Show usage / man page |
|
||||||
| `tsdocker build` | Build all Dockerfiles with dependency ordering |
|
| `tsdocker build` | Build all Dockerfiles with dependency ordering |
|
||||||
| `tsdocker push [registry]` | Push images to configured registries |
|
| `tsdocker push [registry]` | Push images to configured registries |
|
||||||
| `tsdocker pull <registry>` | Pull images from registry |
|
| `tsdocker pull <registry>` | Pull images from registry |
|
||||||
@@ -20,7 +19,6 @@ tsdocker is a comprehensive Docker development and building tool. It provides:
|
|||||||
| `tsdocker login` | Login to configured registries |
|
| `tsdocker login` | Login to configured registries |
|
||||||
| `tsdocker list` | List discovered Dockerfiles and dependencies |
|
| `tsdocker list` | List discovered Dockerfiles and dependencies |
|
||||||
| `tsdocker clean --all` | Clean up Docker environment |
|
| `tsdocker clean --all` | Clean up Docker environment |
|
||||||
| `tsdocker vscode` | Start VS Code in Docker |
|
|
||||||
|
|
||||||
## Configuration
|
## Configuration
|
||||||
|
|
||||||
@@ -45,9 +43,6 @@ Configure in `package.json` under `@git.zone/tsdocker`:
|
|||||||
|
|
||||||
### Configuration Options
|
### Configuration Options
|
||||||
|
|
||||||
- `baseImage`: Base Docker image for testing (legacy)
|
|
||||||
- `command`: Command to run in container (legacy)
|
|
||||||
- `dockerSock`: Mount Docker socket (legacy)
|
|
||||||
- `registries`: Array of registry URLs to push to
|
- `registries`: Array of registry URLs to push to
|
||||||
- `registryRepoMap`: Map registry URLs to different repo paths
|
- `registryRepoMap`: Map registry URLs to different repo paths
|
||||||
- `buildArgEnvMap`: Map Docker build ARGs to environment variables
|
- `buildArgEnvMap`: Map Docker build ARGs to environment variables
|
||||||
@@ -78,8 +73,6 @@ ts/
|
|||||||
├── tsdocker.cli.ts (CLI commands)
|
├── tsdocker.cli.ts (CLI commands)
|
||||||
├── tsdocker.config.ts (configuration)
|
├── tsdocker.config.ts (configuration)
|
||||||
├── tsdocker.plugins.ts (plugin imports)
|
├── tsdocker.plugins.ts (plugin imports)
|
||||||
├── tsdocker.docker.ts (legacy test runner)
|
|
||||||
├── tsdocker.snippets.ts (Dockerfile generation)
|
|
||||||
├── classes.dockerfile.ts (Dockerfile management)
|
├── classes.dockerfile.ts (Dockerfile management)
|
||||||
├── classes.dockerregistry.ts (registry authentication)
|
├── classes.dockerregistry.ts (registry authentication)
|
||||||
├── classes.registrystorage.ts (registry storage)
|
├── classes.registrystorage.ts (registry storage)
|
||||||
@@ -123,8 +116,6 @@ The `config.push` field is now a no-op (kept for backward compat).
|
|||||||
## Build Status
|
## Build Status
|
||||||
|
|
||||||
- Build: ✅ Passes
|
- Build: ✅ Passes
|
||||||
- Legacy test functionality preserved
|
|
||||||
- New Docker build functionality added
|
|
||||||
|
|
||||||
## Previous Upgrades (2025-11-22)
|
## Previous Upgrades (2025-11-22)
|
||||||
|
|
||||||
|
|||||||
526
readme.md
526
readme.md
@@ -1,6 +1,6 @@
|
|||||||
# @git.zone/tsdocker
|
# @git.zone/tsdocker
|
||||||
|
|
||||||
> 🐳 The ultimate Docker development toolkit for TypeScript projects — build, test, and ship containerized applications with ease.
|
> 🐳 The ultimate Docker development toolkit for TypeScript projects — build, test, and ship multi-arch containerized applications with zero friction.
|
||||||
|
|
||||||
## Issue Reporting and Security
|
## Issue Reporting and Security
|
||||||
|
|
||||||
@@ -8,15 +8,20 @@ For reporting bugs, issues, or security vulnerabilities, please visit [community
|
|||||||
|
|
||||||
## What is tsdocker?
|
## What is tsdocker?
|
||||||
|
|
||||||
**tsdocker** is a comprehensive Docker development and building tool that handles everything from testing npm packages in clean environments to building and pushing multi-architecture Docker images across multiple registries.
|
**tsdocker** is a comprehensive Docker development and build tool that handles everything from testing npm packages in clean environments to building and pushing multi-architecture Docker images across multiple registries — all from a single CLI.
|
||||||
|
|
||||||
### 🎯 Key Capabilities
|
### 🎯 Key Capabilities
|
||||||
|
|
||||||
- 🧪 **Containerized Testing** — Run your tests in pristine Docker environments
|
|
||||||
- 🏗️ **Smart Docker Builds** — Automatically discover, sort, and build Dockerfiles by dependency
|
- 🏗️ **Smart Docker Builds** — Automatically discover, sort, and build Dockerfiles by dependency
|
||||||
- 🚀 **Multi-Registry Push** — Ship to Docker Hub, GitLab, GitHub Container Registry, and more
|
- 🌍 **True Multi-Architecture** — Build for `amd64` and `arm64` simultaneously with Docker Buildx
|
||||||
- 🔧 **Multi-Architecture** — Build for `amd64` and `arm64` with Docker Buildx
|
- 🚀 **Multi-Registry Push** — Ship to Docker Hub, GitLab, GitHub Container Registry, and more via OCI Distribution API
|
||||||
- ⚡ **Zero Config Start** — Works out of the box, scales with your needs
|
- ⚡ **Parallel Builds** — Level-based parallel builds with configurable concurrency
|
||||||
|
- 🗄️ **Persistent Local Registry** — All images flow through a local OCI registry with persistent storage
|
||||||
|
- 📦 **Build Caching** — Skip unchanged Dockerfiles with content-hash caching
|
||||||
|
- 🎯 **Dockerfile Filtering** — Build or push only specific Dockerfiles using glob patterns
|
||||||
|
- 🔁 **Resilient Push** — Automatic retry with exponential backoff, timeouts, and token refresh for rock-solid pushes
|
||||||
|
- 🏭 **CI-Safe Isolation** — Unique sessions per invocation prevent collisions in parallel CI pipelines
|
||||||
|
- 🔧 **Zero Config Start** — Works out of the box, scales with your needs
|
||||||
|
|
||||||
## Installation
|
## Installation
|
||||||
|
|
||||||
@@ -30,16 +35,6 @@ pnpm install --save-dev @git.zone/tsdocker
|
|||||||
|
|
||||||
## Quick Start
|
## Quick Start
|
||||||
|
|
||||||
### 🧪 Run Tests in Docker
|
|
||||||
|
|
||||||
The simplest use case — run your tests in a clean container:
|
|
||||||
|
|
||||||
```bash
|
|
||||||
tsdocker
|
|
||||||
```
|
|
||||||
|
|
||||||
This pulls your configured base image, mounts your project, and executes your test command in isolation.
|
|
||||||
|
|
||||||
### 🏗️ Build Docker Images
|
### 🏗️ Build Docker Images
|
||||||
|
|
||||||
Got `Dockerfile` files? Build them all with automatic dependency ordering:
|
Got `Dockerfile` files? Build them all with automatic dependency ordering:
|
||||||
@@ -53,6 +48,7 @@ tsdocker will:
|
|||||||
2. 📊 Analyze `FROM` dependencies between them
|
2. 📊 Analyze `FROM` dependencies between them
|
||||||
3. 🔄 Sort them topologically
|
3. 🔄 Sort them topologically
|
||||||
4. 🏗️ Build each image in the correct order
|
4. 🏗️ Build each image in the correct order
|
||||||
|
5. 📦 Push every image to a persistent local registry (`.nogit/docker-registry/`)
|
||||||
|
|
||||||
### 📤 Push to Registries
|
### 📤 Push to Registries
|
||||||
|
|
||||||
@@ -63,33 +59,97 @@ Ship your images to one or all configured registries:
|
|||||||
tsdocker push
|
tsdocker push
|
||||||
|
|
||||||
# Push to a specific registry
|
# Push to a specific registry
|
||||||
tsdocker push registry.gitlab.com
|
tsdocker push --registry=registry.gitlab.com
|
||||||
|
|
||||||
|
# Push without rebuilding (use existing images in local registry)
|
||||||
|
tsdocker push --no-build
|
||||||
|
```
|
||||||
|
|
||||||
|
Under the hood, `tsdocker push` uses the **OCI Distribution API** to copy images directly from the local registry to remote registries. This means multi-arch manifest lists are preserved end-to-end — no more single-platform-only pushes. Every request is protected with **automatic retry** (up to 6 attempts with exponential backoff) and **5-minute timeouts**, so transient network issues don't kill your push mid-transfer.
|
||||||
|
|
||||||
|
### 🎯 Build Only Specific Dockerfiles
|
||||||
|
|
||||||
|
Target specific Dockerfiles by name pattern — dependencies are resolved automatically:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Build only the base image
|
||||||
|
tsdocker build Dockerfile_base
|
||||||
|
|
||||||
|
# Build anything matching a glob pattern
|
||||||
|
tsdocker build Dockerfile_app*
|
||||||
|
|
||||||
|
# Push specific images only (skip build phase)
|
||||||
|
tsdocker push --no-build Dockerfile_api Dockerfile_web
|
||||||
```
|
```
|
||||||
|
|
||||||
## CLI Commands
|
## CLI Commands
|
||||||
|
|
||||||
| Command | Description |
|
| Command | Description |
|
||||||
|---------|-------------|
|
|---------|-------------|
|
||||||
| `tsdocker` | Run tests in a fresh Docker container |
|
| `tsdocker` | Show usage / man page |
|
||||||
| `tsdocker build` | Build all Dockerfiles with dependency ordering |
|
| `tsdocker build` | Build all Dockerfiles with dependency ordering |
|
||||||
| `tsdocker push [registry]` | Push images to configured registries |
|
| `tsdocker push` | Build + push images to configured registries |
|
||||||
| `tsdocker pull <registry>` | Pull images from a specific registry |
|
| `tsdocker pull <registry>` | Pull images from a specific registry |
|
||||||
| `tsdocker test` | Run container test scripts (test_*.sh) |
|
| `tsdocker test` | Build + run container test scripts (`test_*.sh`) |
|
||||||
| `tsdocker login` | Authenticate with configured registries |
|
| `tsdocker login` | Authenticate with configured registries |
|
||||||
| `tsdocker list` | Display discovered Dockerfiles and their dependencies |
|
| `tsdocker list` | Display discovered Dockerfiles and their dependencies |
|
||||||
| `tsdocker clean --all` | ⚠️ Aggressively clean Docker environment |
|
| `tsdocker config` | Manage global tsdocker configuration (remote builders, etc.) |
|
||||||
| `tsdocker vscode` | Launch containerized VS Code in browser |
|
| `tsdocker clean` | Interactively clean Docker environment |
|
||||||
|
|
||||||
|
### Build Flags
|
||||||
|
|
||||||
|
| Flag | Description |
|
||||||
|
|------|-------------|
|
||||||
|
| `<patterns>` | Positional Dockerfile name patterns (e.g. `Dockerfile_base`, `Dockerfile_app*`) |
|
||||||
|
| `--platform=linux/arm64` | Override build platform for a single architecture |
|
||||||
|
| `--timeout=600` | Build timeout in seconds |
|
||||||
|
| `--no-cache` | Force rebuild without Docker layer cache |
|
||||||
|
| `--cached` | Skip unchanged Dockerfiles (content-hash based) |
|
||||||
|
| `--verbose` | Stream raw `docker build` output |
|
||||||
|
| `--parallel` | Enable level-based parallel builds (default concurrency: 4) |
|
||||||
|
| `--parallel=8` | Parallel builds with custom concurrency |
|
||||||
|
| `--context=mycontext` | Use a specific Docker context |
|
||||||
|
|
||||||
|
### Push Flags
|
||||||
|
|
||||||
|
| Flag | Description |
|
||||||
|
|------|-------------|
|
||||||
|
| `<patterns>` | Positional Dockerfile name patterns to select which images to push |
|
||||||
|
| `--registry=<url>` | Push to a single specific registry instead of all configured |
|
||||||
|
| `--no-build` | Skip the build phase; only push existing images from local registry |
|
||||||
|
|
||||||
|
### Config Subcommands
|
||||||
|
|
||||||
|
| Subcommand | Description |
|
||||||
|
|------------|-------------|
|
||||||
|
| `add-builder` | Add or update a remote builder node |
|
||||||
|
| `remove-builder` | Remove a remote builder by name |
|
||||||
|
| `list-builders` | List all configured remote builders |
|
||||||
|
| `show` | Show the full global configuration |
|
||||||
|
|
||||||
|
**`add-builder` flags:**
|
||||||
|
|
||||||
|
| Flag | Description |
|
||||||
|
|------|-------------|
|
||||||
|
| `--name=<name>` | Builder name (e.g. `arm64-builder`) |
|
||||||
|
| `--host=<user@ip>` | SSH host (e.g. `armbuilder@192.168.1.100`) |
|
||||||
|
| `--platform=<p>` | Target platform (e.g. `linux/arm64`) |
|
||||||
|
| `--ssh-key=<path>` | SSH key path (optional, uses SSH agent/config by default) |
|
||||||
|
|
||||||
|
### Clean Flags
|
||||||
|
|
||||||
|
| Flag | Description |
|
||||||
|
|------|-------------|
|
||||||
|
| `--all` | Include all images and volumes (not just dangling) |
|
||||||
|
| `-y` | Auto-confirm all prompts |
|
||||||
|
|
||||||
## Configuration
|
## Configuration
|
||||||
|
|
||||||
Configure tsdocker in your `package.json` or `npmextra.json`:
|
Configure tsdocker in your `package.json` or `npmextra.json` under the `@git.zone/tsdocker` key:
|
||||||
|
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
"@git.zone/tsdocker": {
|
"@git.zone/tsdocker": {
|
||||||
"baseImage": "node:20",
|
|
||||||
"command": "npm test",
|
|
||||||
"dockerSock": false,
|
|
||||||
"registries": ["registry.gitlab.com", "docker.io"],
|
"registries": ["registry.gitlab.com", "docker.io"],
|
||||||
"registryRepoMap": {
|
"registryRepoMap": {
|
||||||
"registry.gitlab.com": "myorg/myproject"
|
"registry.gitlab.com": "myorg/myproject"
|
||||||
@@ -98,7 +158,6 @@ Configure tsdocker in your `package.json` or `npmextra.json`:
|
|||||||
"NODE_VERSION": "NODE_VERSION"
|
"NODE_VERSION": "NODE_VERSION"
|
||||||
},
|
},
|
||||||
"platforms": ["linux/amd64", "linux/arm64"],
|
"platforms": ["linux/amd64", "linux/arm64"],
|
||||||
"push": false,
|
|
||||||
"testDir": "./test"
|
"testDir": "./test"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -106,24 +165,107 @@ Configure tsdocker in your `package.json` or `npmextra.json`:
|
|||||||
|
|
||||||
### Configuration Options
|
### Configuration Options
|
||||||
|
|
||||||
#### Testing Options (Legacy)
|
|
||||||
|
|
||||||
| Option | Type | Description |
|
|
||||||
|--------|------|-------------|
|
|
||||||
| `baseImage` | `string` | Docker image for test environment (default: `hosttoday/ht-docker-node:npmdocker`) |
|
|
||||||
| `command` | `string` | Command to run inside container (default: `npmci npm test`) |
|
|
||||||
| `dockerSock` | `boolean` | Mount Docker socket for DinD scenarios (default: `false`) |
|
|
||||||
|
|
||||||
#### Build & Push Options
|
#### Build & Push Options
|
||||||
|
|
||||||
| Option | Type | Description |
|
| Option | Type | Default | Description |
|
||||||
|--------|------|-------------|
|
|--------|------|---------|-------------|
|
||||||
| `registries` | `string[]` | Registry URLs to push to |
|
| `registries` | `string[]` | `[]` | Registry URLs to push to |
|
||||||
| `registryRepoMap` | `object` | Map registries to different repository paths |
|
| `registryRepoMap` | `object` | `{}` | Map registries to different repository paths |
|
||||||
| `buildArgEnvMap` | `object` | Map Docker build ARGs to environment variables |
|
| `buildArgEnvMap` | `object` | `{}` | Map Docker build ARGs to environment variables |
|
||||||
| `platforms` | `string[]` | Target architectures (default: `["linux/amd64"]`) |
|
| `platforms` | `string[]` | `["linux/amd64"]` | Target architectures for multi-arch builds |
|
||||||
| `push` | `boolean` | Auto-push after build (default: `false`) |
|
| `testDir` | `string` | `./test` | Directory containing test scripts |
|
||||||
| `testDir` | `string` | Directory containing test scripts |
|
|
||||||
|
## Architecture: How tsdocker Works
|
||||||
|
|
||||||
|
tsdocker uses a **local OCI registry** as the canonical store for all built images. This design solves fundamental problems with Docker's local daemon, which cannot hold multi-architecture manifest lists.
|
||||||
|
|
||||||
|
### 📐 Build Flow
|
||||||
|
|
||||||
|
```
|
||||||
|
┌─────────────────────────────────────────────────────┐
|
||||||
|
│ tsdocker build │
|
||||||
|
│ │
|
||||||
|
│ 1. Start local registry (localhost:<dynamic-port>) │
|
||||||
|
│ └── Persistent volume: .nogit/docker-registry/ │
|
||||||
|
│ │
|
||||||
|
│ 2. For each Dockerfile (topological order): │
|
||||||
|
│ ├── Multi-platform: buildx --push → registry │
|
||||||
|
│ └── Single-platform: docker build → registry │
|
||||||
|
│ │
|
||||||
|
│ 3. Stop local registry (data persists on disk) │
|
||||||
|
└─────────────────────────────────────────────────────┘
|
||||||
|
```
|
||||||
|
|
||||||
|
### 📤 Push Flow
|
||||||
|
|
||||||
|
```
|
||||||
|
┌────────────────────────────────────────────────────────┐
|
||||||
|
│ tsdocker push │
|
||||||
|
│ │
|
||||||
|
│ 1. Start local registry (loads persisted data) │
|
||||||
|
│ │
|
||||||
|
│ 2. For each image × each remote registry: │
|
||||||
|
│ └── OCI Distribution API copy (with retry): │
|
||||||
|
│ ├── Fetch manifest (single or multi-arch) │
|
||||||
|
│ ├── Copy blobs (skip if already exist) │
|
||||||
|
│ ├── Retry up to 6× with exponential backoff │
|
||||||
|
│ └── Push manifest with destination tag │
|
||||||
|
│ │
|
||||||
|
│ 3. Stop local registry │
|
||||||
|
└────────────────────────────────────────────────────────┘
|
||||||
|
```
|
||||||
|
|
||||||
|
### 🔑 Why a Local Registry?
|
||||||
|
|
||||||
|
| Problem | Solution |
|
||||||
|
|---------|----------|
|
||||||
|
| `docker buildx --load` fails for multi-arch images | `buildx --push` to local registry works for any number of platforms |
|
||||||
|
| `docker push` only pushes single-platform manifests | OCI API copy preserves full manifest lists (multi-arch) |
|
||||||
|
| Images lost between build and push phases | Persistent storage at `.nogit/docker-registry/` survives restarts |
|
||||||
|
| Redundant blob uploads on incremental pushes | HEAD checks skip blobs that already exist on the remote |
|
||||||
|
|
||||||
|
### 🔁 Resilient Push
|
||||||
|
|
||||||
|
The OCI Distribution API client wraps every HTTP request with:
|
||||||
|
|
||||||
|
- **Timeouts** — 5-minute timeout for blob operations, 30-second timeout for auth/metadata calls via `AbortSignal.timeout()`
|
||||||
|
- **Automatic Retry** — Up to 6 attempts with exponential backoff (1s → 2s → 4s → 8s → 16s → 32s)
|
||||||
|
- **Smart Retry Logic** — Retries on network errors (`ECONNRESET`, `fetch failed`) and 5xx server errors; does NOT retry 4xx client errors
|
||||||
|
- **Token Refresh** — On 401 responses, the cached auth token is cleared so the next retry re-authenticates automatically
|
||||||
|
|
||||||
|
This means transient issues like stale connection pools, brief network blips, or token expiry during long multi-arch pushes (56+ blob operations) are handled gracefully instead of killing the entire transfer.
|
||||||
|
|
||||||
|
### 🏭 CI-Safe Session Isolation
|
||||||
|
|
||||||
|
Every tsdocker invocation gets its own **session** with unique:
|
||||||
|
|
||||||
|
- **Session ID** — Random 8-char hex (override with `TSDOCKER_SESSION_ID`)
|
||||||
|
- **Registry port** — Dynamically allocated (override with `TSDOCKER_REGISTRY_PORT`)
|
||||||
|
- **Registry container** — Named `tsdocker-registry-<sessionId>`
|
||||||
|
- **Builder suffix** — In CI, the buildx builder gets a `-<sessionId>` suffix to prevent collisions
|
||||||
|
|
||||||
|
This prevents resource conflicts when multiple CI jobs run tsdocker in parallel. Auto-detected CI systems:
|
||||||
|
|
||||||
|
| Environment Variable | CI System |
|
||||||
|
|---------------------|-----------|
|
||||||
|
| `GITEA_ACTIONS` | Gitea Actions |
|
||||||
|
| `GITHUB_ACTIONS` | GitHub Actions |
|
||||||
|
| `GITLAB_CI` | GitLab CI |
|
||||||
|
| `CI` | Generic CI |
|
||||||
|
|
||||||
|
In local dev, no suffix is added — keeping a persistent builder for faster rebuilds.
|
||||||
|
|
||||||
|
### 🔍 Docker Context & Topology Detection
|
||||||
|
|
||||||
|
tsdocker automatically detects your Docker environment topology:
|
||||||
|
|
||||||
|
| Topology | Detection | Meaning |
|
||||||
|
|----------|-----------|---------|
|
||||||
|
| `local` | Default | Standard Docker installation on the host |
|
||||||
|
| `socket-mount` | `/.dockerenv` exists | Running inside a container with Docker socket mounted |
|
||||||
|
| `dind` | `DOCKER_HOST` starts with `tcp://` | Docker-in-Docker setup |
|
||||||
|
|
||||||
|
Context-aware builder names (`tsdocker-builder-<context>`) prevent conflicts across Docker contexts. Rootless Docker configurations trigger appropriate warnings.
|
||||||
|
|
||||||
## Registry Authentication
|
## Registry Authentication
|
||||||
|
|
||||||
@@ -140,13 +282,17 @@ export DOCKER_REGISTRY_USER="username"
|
|||||||
export DOCKER_REGISTRY_PASSWORD="password"
|
export DOCKER_REGISTRY_PASSWORD="password"
|
||||||
```
|
```
|
||||||
|
|
||||||
|
### Docker Config Fallback
|
||||||
|
|
||||||
|
When pushing, tsdocker will also read credentials from `~/.docker/config.json` if no explicit credentials are provided via environment variables. This means `docker login` credentials work automatically. Docker Hub special cases (`docker.io`, `index.docker.io`, `registry-1.docker.io`) are all recognized.
|
||||||
|
|
||||||
### Login Command
|
### Login Command
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
tsdocker login
|
tsdocker login
|
||||||
```
|
```
|
||||||
|
|
||||||
Authenticates with all configured registries.
|
Authenticates with all configured registries using the provided environment variables.
|
||||||
|
|
||||||
## Advanced Usage
|
## Advanced Usage
|
||||||
|
|
||||||
@@ -162,7 +308,72 @@ Build for multiple platforms using Docker Buildx:
|
|||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
tsdocker automatically sets up a Buildx builder when multiple platforms are specified.
|
tsdocker automatically:
|
||||||
|
- Sets up a Buildx builder with `--driver-opt network=host` (so buildx can reach the local registry)
|
||||||
|
- Pushes multi-platform images to the local registry via `buildx --push`
|
||||||
|
- Copies the full manifest list (including all platform variants) to remote registries on `tsdocker push`
|
||||||
|
|
||||||
|
### 🖥️ Native Remote Builders
|
||||||
|
|
||||||
|
Instead of relying on slow QEMU emulation for cross-platform builds, tsdocker can use **native remote machines** via SSH as build nodes. For example, use a real arm64 machine for `linux/arm64` builds:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Add a remote arm64 builder
|
||||||
|
tsdocker config add-builder \
|
||||||
|
--name=arm64-builder \
|
||||||
|
--host=armbuilder@192.168.1.100 \
|
||||||
|
--platform=linux/arm64 \
|
||||||
|
--ssh-key=~/.ssh/id_ed25519
|
||||||
|
|
||||||
|
# List configured builders
|
||||||
|
tsdocker config list-builders
|
||||||
|
|
||||||
|
# Remove a builder
|
||||||
|
tsdocker config remove-builder --name=arm64-builder
|
||||||
|
|
||||||
|
# Show full global config
|
||||||
|
tsdocker config show
|
||||||
|
```
|
||||||
|
|
||||||
|
Global configuration is stored at `~/.git.zone/tsdocker/config.json`.
|
||||||
|
|
||||||
|
**How it works:**
|
||||||
|
|
||||||
|
When remote builders are configured and the project's `platforms` includes a matching platform, tsdocker automatically:
|
||||||
|
|
||||||
|
1. Creates a **multi-node buildx builder** — local node for `linux/amd64`, remote SSH node for `linux/arm64`
|
||||||
|
2. Opens **SSH reverse tunnels** so the remote builder can push to the local staging registry
|
||||||
|
3. Builds natively on each platform's hardware — no QEMU overhead
|
||||||
|
4. Tears down tunnels after the build completes
|
||||||
|
|
||||||
|
```
|
||||||
|
[Local machine] [Remote arm64 machine]
|
||||||
|
registry:2 on localhost:PORT <──── SSH reverse tunnel ──── localhost:PORT
|
||||||
|
BuildKit (amd64) ──push──> BuildKit (arm64) ──push──>
|
||||||
|
localhost:PORT localhost:PORT (tunneled)
|
||||||
|
```
|
||||||
|
|
||||||
|
**Prerequisites for the remote machine:**
|
||||||
|
- Docker installed and running
|
||||||
|
- A user with Docker group access (no sudo needed)
|
||||||
|
- SSH key access configured
|
||||||
|
|
||||||
|
### ⚡ Parallel Builds
|
||||||
|
|
||||||
|
Speed up builds by building independent images concurrently:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Default concurrency (4 workers)
|
||||||
|
tsdocker build --parallel
|
||||||
|
|
||||||
|
# Custom concurrency
|
||||||
|
tsdocker build --parallel=8
|
||||||
|
|
||||||
|
# Works with caching too
|
||||||
|
tsdocker build --parallel --cached
|
||||||
|
```
|
||||||
|
|
||||||
|
tsdocker groups Dockerfiles into **dependency levels** using topological analysis. Images within the same level have no dependencies on each other and build in parallel. Each level completes before the next begins.
|
||||||
|
|
||||||
### 📦 Dockerfile Naming Conventions
|
### 📦 Dockerfile Naming Conventions
|
||||||
|
|
||||||
@@ -175,6 +386,26 @@ tsdocker discovers files matching `Dockerfile*`:
|
|||||||
| `Dockerfile_alpine` | `alpine` |
|
| `Dockerfile_alpine` | `alpine` |
|
||||||
| `Dockerfile_##version##` | Uses `package.json` version |
|
| `Dockerfile_##version##` | Uses `package.json` version |
|
||||||
|
|
||||||
|
### 🎯 Dockerfile Filtering
|
||||||
|
|
||||||
|
Build or push only the Dockerfiles you need. Positional arguments are matched against Dockerfile basenames as glob patterns:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Build a single Dockerfile
|
||||||
|
tsdocker build Dockerfile_base
|
||||||
|
|
||||||
|
# Glob patterns with * and ? wildcards
|
||||||
|
tsdocker build Dockerfile_app*
|
||||||
|
|
||||||
|
# Multiple patterns
|
||||||
|
tsdocker build Dockerfile_base Dockerfile_web
|
||||||
|
|
||||||
|
# Push specific images without rebuilding
|
||||||
|
tsdocker push --no-build Dockerfile_api
|
||||||
|
```
|
||||||
|
|
||||||
|
When filtering for `build`, **dependencies are auto-resolved**: if `Dockerfile_app` depends on `Dockerfile_base`, specifying only `Dockerfile_app` will automatically include `Dockerfile_base` in the build order.
|
||||||
|
|
||||||
### 🔗 Dependency-Aware Builds
|
### 🔗 Dependency-Aware Builds
|
||||||
|
|
||||||
If you have multiple Dockerfiles that depend on each other:
|
If you have multiple Dockerfiles that depend on each other:
|
||||||
@@ -190,7 +421,7 @@ COPY . .
|
|||||||
RUN npm run build
|
RUN npm run build
|
||||||
```
|
```
|
||||||
|
|
||||||
tsdocker automatically detects that `Dockerfile_app` depends on `Dockerfile_base` and builds them in the correct order.
|
tsdocker automatically detects that `Dockerfile_app` depends on `Dockerfile_base`, builds them in the correct order, and makes the base image available to dependent builds via the local registry (using `--build-context` for buildx).
|
||||||
|
|
||||||
### 🧪 Container Test Scripts
|
### 🧪 Container Test Scripts
|
||||||
|
|
||||||
@@ -210,6 +441,8 @@ Run with:
|
|||||||
tsdocker test
|
tsdocker test
|
||||||
```
|
```
|
||||||
|
|
||||||
|
This builds all images, starts the local registry (so multi-arch images can be pulled), and runs each matching test script inside a container.
|
||||||
|
|
||||||
### 🔧 Build Args from Environment
|
### 🔧 Build Args from Environment
|
||||||
|
|
||||||
Pass environment variables as Docker build arguments:
|
Pass environment variables as Docker build arguments:
|
||||||
@@ -232,45 +465,6 @@ FROM node:${NODE_VERSION}
|
|||||||
RUN echo "//registry.npmjs.org/:_authToken=${NPM_TOKEN}" > ~/.npmrc
|
RUN echo "//registry.npmjs.org/:_authToken=${NPM_TOKEN}" > ~/.npmrc
|
||||||
```
|
```
|
||||||
|
|
||||||
### 🐳 Docker-in-Docker Testing
|
|
||||||
|
|
||||||
Test Docker-related tools by mounting the Docker socket:
|
|
||||||
|
|
||||||
```json
|
|
||||||
{
|
|
||||||
"@git.zone/tsdocker": {
|
|
||||||
"baseImage": "docker:latest",
|
|
||||||
"command": "docker version && docker ps",
|
|
||||||
"dockerSock": true
|
|
||||||
}
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
### 📋 Listing Dockerfiles
|
|
||||||
|
|
||||||
Inspect your project's Dockerfiles and their relationships:
|
|
||||||
|
|
||||||
```bash
|
|
||||||
tsdocker list
|
|
||||||
```
|
|
||||||
|
|
||||||
Output:
|
|
||||||
```
|
|
||||||
Discovered Dockerfiles:
|
|
||||||
========================
|
|
||||||
|
|
||||||
1. Dockerfile_base
|
|
||||||
Tag: myproject:base
|
|
||||||
Base Image: node:20-alpine
|
|
||||||
Version: base
|
|
||||||
|
|
||||||
2. Dockerfile_app
|
|
||||||
Tag: myproject:app
|
|
||||||
Base Image: myproject:base
|
|
||||||
Version: app
|
|
||||||
Depends on: myproject:base
|
|
||||||
```
|
|
||||||
|
|
||||||
### 🗺️ Registry Repo Mapping
|
### 🗺️ Registry Repo Mapping
|
||||||
|
|
||||||
Use different repository names for different registries:
|
Use different repository names for different registries:
|
||||||
@@ -287,40 +481,55 @@ Use different repository names for different registries:
|
|||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
## Environment Variables
|
When pushing, tsdocker maps the local repo name to the registry-specific path. For example, a locally built `myproject:latest` becomes `registry.gitlab.com/mygroup/myproject:latest` and `docker.io/myuser/myproject:latest`.
|
||||||
|
|
||||||
### qenv Integration
|
### 📋 Listing Dockerfiles
|
||||||
|
|
||||||
tsdocker automatically loads environment variables from `qenv.yml`:
|
Inspect your project's Dockerfiles and their relationships:
|
||||||
|
|
||||||
```yaml
|
```bash
|
||||||
# qenv.yml
|
tsdocker list
|
||||||
API_KEY: your-api-key
|
|
||||||
DATABASE_URL: postgres://localhost/test
|
|
||||||
```
|
```
|
||||||
|
|
||||||
These are injected into your test container automatically.
|
Output:
|
||||||
|
```
|
||||||
|
Discovered Dockerfiles:
|
||||||
|
========================
|
||||||
|
|
||||||
|
1. /path/to/Dockerfile_base
|
||||||
|
Tag: myproject:base
|
||||||
|
Base Image: node:20-alpine
|
||||||
|
Version: base
|
||||||
|
|
||||||
|
2. /path/to/Dockerfile_app
|
||||||
|
Tag: myproject:app
|
||||||
|
Base Image: myproject:base
|
||||||
|
Version: app
|
||||||
|
Depends on: myproject:base
|
||||||
|
```
|
||||||
|
|
||||||
## Examples
|
## Examples
|
||||||
|
|
||||||
### Basic Test Configuration
|
### Minimal Build & Push
|
||||||
|
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
"@git.zone/tsdocker": {
|
"@git.zone/tsdocker": {
|
||||||
"baseImage": "node:20",
|
"registries": ["docker.io"],
|
||||||
"command": "npm test"
|
"platforms": ["linux/amd64"]
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
|
```bash
|
||||||
|
tsdocker push
|
||||||
|
```
|
||||||
|
|
||||||
### Full Production Setup
|
### Full Production Setup
|
||||||
|
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
"@git.zone/tsdocker": {
|
"@git.zone/tsdocker": {
|
||||||
"baseImage": "node:20-alpine",
|
|
||||||
"command": "pnpm test",
|
|
||||||
"registries": ["registry.gitlab.com", "ghcr.io", "docker.io"],
|
"registries": ["registry.gitlab.com", "ghcr.io", "docker.io"],
|
||||||
"registryRepoMap": {
|
"registryRepoMap": {
|
||||||
"registry.gitlab.com": "myorg/myapp",
|
"registry.gitlab.com": "myorg/myapp",
|
||||||
@@ -338,73 +547,87 @@ These are injected into your test container automatically.
|
|||||||
|
|
||||||
### CI/CD Integration
|
### CI/CD Integration
|
||||||
|
|
||||||
|
**GitLab CI:**
|
||||||
|
|
||||||
```yaml
|
```yaml
|
||||||
# .gitlab-ci.yml
|
build-and-push:
|
||||||
build:
|
|
||||||
stage: build
|
stage: build
|
||||||
script:
|
script:
|
||||||
- npm install -g @git.zone/tsdocker
|
- npm install -g @git.zone/tsdocker
|
||||||
- tsdocker build
|
|
||||||
- tsdocker push
|
- tsdocker push
|
||||||
|
variables:
|
||||||
|
DOCKER_REGISTRY_1: "registry.gitlab.com|$CI_REGISTRY_USER|$CI_REGISTRY_PASSWORD"
|
||||||
|
```
|
||||||
|
|
||||||
# GitHub Actions
|
**GitHub Actions:**
|
||||||
|
|
||||||
|
```yaml
|
||||||
- name: Build and Push
|
- name: Build and Push
|
||||||
run: |
|
run: |
|
||||||
npm install -g @git.zone/tsdocker
|
npm install -g @git.zone/tsdocker
|
||||||
tsdocker login
|
tsdocker login
|
||||||
tsdocker build
|
|
||||||
tsdocker push
|
tsdocker push
|
||||||
env:
|
env:
|
||||||
DOCKER_REGISTRY_1: "ghcr.io|${{ github.actor }}|${{ secrets.GITHUB_TOKEN }}"
|
DOCKER_REGISTRY_1: "ghcr.io|${{ github.actor }}|${{ secrets.GITHUB_TOKEN }}"
|
||||||
```
|
```
|
||||||
|
|
||||||
## Requirements
|
**Gitea Actions:**
|
||||||
|
|
||||||
- **Docker** — Docker Engine or Docker Desktop must be installed
|
```yaml
|
||||||
- **Node.js** — Version 18 or higher (ESM support required)
|
- name: Build and Push
|
||||||
- **Docker Buildx** — Required for multi-architecture builds (included in Docker Desktop)
|
run: |
|
||||||
|
npm install -g @git.zone/tsdocker
|
||||||
|
tsdocker push
|
||||||
|
env:
|
||||||
|
DOCKER_REGISTRY_1: "gitea.example.com|${{ secrets.REGISTRY_USER }}|${{ secrets.REGISTRY_PASSWORD }}"
|
||||||
|
```
|
||||||
|
|
||||||
## Why tsdocker?
|
tsdocker auto-detects all three CI systems and enables session isolation automatically — no extra configuration needed.
|
||||||
|
|
||||||
### 🎯 The Problem
|
|
||||||
|
|
||||||
Managing Docker workflows manually is tedious:
|
|
||||||
- Remembering build order for dependent images
|
|
||||||
- Pushing to multiple registries with different credentials
|
|
||||||
- Setting up Buildx for multi-arch builds
|
|
||||||
- Ensuring consistent test environments
|
|
||||||
|
|
||||||
### ✨ The Solution
|
|
||||||
|
|
||||||
tsdocker automates the entire workflow:
|
|
||||||
- **One command** to build all images in dependency order
|
|
||||||
- **One command** to push to all registries
|
|
||||||
- **Automatic** Buildx setup for multi-platform builds
|
|
||||||
- **Consistent** containerized test environments
|
|
||||||
|
|
||||||
## TypeScript API
|
## TypeScript API
|
||||||
|
|
||||||
tsdocker exposes its types for programmatic use:
|
tsdocker can also be used programmatically:
|
||||||
|
|
||||||
```typescript
|
```typescript
|
||||||
import type { ITsDockerConfig } from '@git.zone/tsdocker/dist_ts/interfaces/index.js';
|
|
||||||
import { TsDockerManager } from '@git.zone/tsdocker/dist_ts/classes.tsdockermanager.js';
|
import { TsDockerManager } from '@git.zone/tsdocker/dist_ts/classes.tsdockermanager.js';
|
||||||
|
import type { ITsDockerConfig } from '@git.zone/tsdocker/dist_ts/interfaces/index.js';
|
||||||
|
|
||||||
const config: ITsDockerConfig = {
|
const config: ITsDockerConfig = {
|
||||||
baseImage: 'node:20',
|
|
||||||
command: 'npm test',
|
|
||||||
dockerSock: false,
|
|
||||||
keyValueObject: {},
|
|
||||||
registries: ['docker.io'],
|
registries: ['docker.io'],
|
||||||
platforms: ['linux/amd64'],
|
platforms: ['linux/amd64', 'linux/arm64'],
|
||||||
};
|
};
|
||||||
|
|
||||||
const manager = new TsDockerManager(config);
|
const manager = new TsDockerManager(config);
|
||||||
await manager.prepare();
|
await manager.prepare();
|
||||||
await manager.build();
|
await manager.build({ parallel: true });
|
||||||
await manager.push();
|
await manager.push();
|
||||||
```
|
```
|
||||||
|
|
||||||
|
## Environment Variables
|
||||||
|
|
||||||
|
### CI & Session Control
|
||||||
|
|
||||||
|
| Variable | Description |
|
||||||
|
|----------|-------------|
|
||||||
|
| `TSDOCKER_SESSION_ID` | Override the auto-generated session ID (default: random 8-char hex) |
|
||||||
|
| `TSDOCKER_REGISTRY_PORT` | Override the dynamically allocated local registry port |
|
||||||
|
| `CI` | Generic CI detection (also `GITHUB_ACTIONS`, `GITLAB_CI`, `GITEA_ACTIONS`) |
|
||||||
|
|
||||||
|
### Registry Credentials
|
||||||
|
|
||||||
|
| Variable | Description |
|
||||||
|
|----------|-------------|
|
||||||
|
| `DOCKER_REGISTRY_1` through `DOCKER_REGISTRY_10` | Pipe-delimited: `registry\|username\|password` |
|
||||||
|
| `DOCKER_REGISTRY_URL` | Registry URL for single-registry setup |
|
||||||
|
| `DOCKER_REGISTRY_USER` | Username for single-registry setup |
|
||||||
|
| `DOCKER_REGISTRY_PASSWORD` | Password for single-registry setup |
|
||||||
|
|
||||||
|
## Requirements
|
||||||
|
|
||||||
|
- **Docker** — Docker Engine 20+ or Docker Desktop
|
||||||
|
- **Node.js** — Version 18 or higher (for native `fetch` and ESM support)
|
||||||
|
- **Docker Buildx** — Required for multi-architecture builds (included in Docker Desktop)
|
||||||
|
|
||||||
## Troubleshooting
|
## Troubleshooting
|
||||||
|
|
||||||
### "docker not found"
|
### "docker not found"
|
||||||
@@ -417,11 +640,10 @@ docker --version
|
|||||||
|
|
||||||
### Multi-arch build fails
|
### Multi-arch build fails
|
||||||
|
|
||||||
Make sure Docker Buildx is available:
|
Make sure Docker Buildx is available. tsdocker will set up the builder automatically, but you can verify:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
docker buildx version
|
docker buildx version
|
||||||
docker buildx create --use
|
|
||||||
```
|
```
|
||||||
|
|
||||||
### Registry authentication fails
|
### Registry authentication fails
|
||||||
@@ -433,29 +655,31 @@ echo $DOCKER_REGISTRY_1
|
|||||||
tsdocker login
|
tsdocker login
|
||||||
```
|
```
|
||||||
|
|
||||||
|
tsdocker also falls back to `~/.docker/config.json` — ensure you've run `docker login` for your target registries.
|
||||||
|
|
||||||
|
### Push fails with "fetch failed"
|
||||||
|
|
||||||
|
tsdocker automatically retries failed requests up to 6 times with exponential backoff. If pushes still fail:
|
||||||
|
|
||||||
|
- Check network connectivity to the target registry
|
||||||
|
- Verify your credentials haven't expired
|
||||||
|
- Look for retry log messages (`fetch failed (attempt X/6)`) to diagnose the pattern
|
||||||
|
- Large layers may need longer timeouts — the default 5-minute timeout per request should cover most cases
|
||||||
|
|
||||||
### Circular dependency detected
|
### Circular dependency detected
|
||||||
|
|
||||||
Review your Dockerfiles' `FROM` statements — you have images depending on each other in a loop.
|
Review your Dockerfiles' `FROM` statements — you have images depending on each other in a loop.
|
||||||
|
|
||||||
## Performance Tips
|
### Build context too large
|
||||||
|
|
||||||
🚀 **Use specific tags**: `node:20-alpine` is smaller and faster than `node:latest`
|
Use a `.dockerignore` file to exclude `node_modules`, `.git`, `.nogit`, and other large directories:
|
||||||
|
|
||||||
🚀 **Leverage caching**: Docker layers are cached — your builds get faster over time
|
```
|
||||||
|
node_modules
|
||||||
🚀 **Prune regularly**: `docker system prune` reclaims disk space
|
.git
|
||||||
|
.nogit
|
||||||
🚀 **Use .dockerignore**: Exclude `node_modules`, `.git`, etc. from build context
|
dist_ts
|
||||||
|
```
|
||||||
## Migration from Legacy
|
|
||||||
|
|
||||||
Previously published as `npmdocker`, now `@git.zone/tsdocker`:
|
|
||||||
|
|
||||||
| Old | New |
|
|
||||||
|-----|-----|
|
|
||||||
| `npmdocker` command | `tsdocker` command |
|
|
||||||
| `"npmdocker"` config key | `"@git.zone/tsdocker"` config key |
|
|
||||||
| CommonJS | ESM with `.js` imports |
|
|
||||||
|
|
||||||
## License and Legal Information
|
## License and Legal Information
|
||||||
|
|
||||||
|
|||||||
@@ -3,6 +3,6 @@
|
|||||||
*/
|
*/
|
||||||
export const commitinfo = {
|
export const commitinfo = {
|
||||||
name: '@git.zone/tsdocker',
|
name: '@git.zone/tsdocker',
|
||||||
version: '1.15.1',
|
version: '2.1.0',
|
||||||
description: 'develop npm modules cross platform with docker'
|
description: 'develop npm modules cross platform with docker'
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,4 +1,5 @@
|
|||||||
import * as plugins from './tsdocker.plugins.js';
|
import * as plugins from './tsdocker.plugins.js';
|
||||||
|
import * as fs from 'fs';
|
||||||
import { logger } from './tsdocker.logging.js';
|
import { logger } from './tsdocker.logging.js';
|
||||||
import type { IDockerContextInfo } from './interfaces/index.js';
|
import type { IDockerContextInfo } from './interfaces/index.js';
|
||||||
|
|
||||||
@@ -38,19 +39,28 @@ export class DockerContext {
|
|||||||
isRootless = infoResult.stdout.includes('name=rootless');
|
isRootless = infoResult.stdout.includes('name=rootless');
|
||||||
}
|
}
|
||||||
|
|
||||||
this.contextInfo = { name, endpoint, isRootless, dockerHost: process.env.DOCKER_HOST };
|
// Detect topology
|
||||||
|
let topology: 'socket-mount' | 'dind' | 'local' = 'local';
|
||||||
|
if (process.env.DOCKER_HOST && process.env.DOCKER_HOST.startsWith('tcp://')) {
|
||||||
|
topology = 'dind';
|
||||||
|
} else if (fs.existsSync('/.dockerenv')) {
|
||||||
|
topology = 'socket-mount';
|
||||||
|
}
|
||||||
|
|
||||||
|
this.contextInfo = { name, endpoint, isRootless, dockerHost: process.env.DOCKER_HOST, topology };
|
||||||
return this.contextInfo;
|
return this.contextInfo;
|
||||||
}
|
}
|
||||||
|
|
||||||
/** Logs context info prominently. */
|
/** Logs context info prominently. */
|
||||||
public logContextInfo(): void {
|
public logContextInfo(): void {
|
||||||
if (!this.contextInfo) return;
|
if (!this.contextInfo) return;
|
||||||
const { name, endpoint, isRootless, dockerHost } = this.contextInfo;
|
const { name, endpoint, isRootless, dockerHost, topology } = this.contextInfo;
|
||||||
logger.log('info', '=== DOCKER CONTEXT ===');
|
logger.log('info', '=== DOCKER CONTEXT ===');
|
||||||
logger.log('info', `Context: ${name}`);
|
logger.log('info', `Context: ${name}`);
|
||||||
logger.log('info', `Endpoint: ${endpoint}`);
|
logger.log('info', `Endpoint: ${endpoint}`);
|
||||||
if (dockerHost) logger.log('info', `DOCKER_HOST: ${dockerHost}`);
|
if (dockerHost) logger.log('info', `DOCKER_HOST: ${dockerHost}`);
|
||||||
logger.log('info', `Rootless: ${isRootless ? 'yes' : 'no'}`);
|
logger.log('info', `Rootless: ${isRootless ? 'yes' : 'no'}`);
|
||||||
|
logger.log('info', `Topology: ${topology || 'local'}`);
|
||||||
}
|
}
|
||||||
|
|
||||||
/** Emits rootless-specific warnings. */
|
/** Emits rootless-specific warnings. */
|
||||||
|
|||||||
@@ -3,6 +3,7 @@ import * as paths from './tsdocker.paths.js';
|
|||||||
import { logger, formatDuration } from './tsdocker.logging.js';
|
import { logger, formatDuration } from './tsdocker.logging.js';
|
||||||
import { DockerRegistry } from './classes.dockerregistry.js';
|
import { DockerRegistry } from './classes.dockerregistry.js';
|
||||||
import { RegistryCopy } from './classes.registrycopy.js';
|
import { RegistryCopy } from './classes.registrycopy.js';
|
||||||
|
import { TsDockerSession } from './classes.tsdockersession.js';
|
||||||
import type { IDockerfileOptions, ITsDockerConfig, IBuildCommandOptions } from './interfaces/index.js';
|
import type { IDockerfileOptions, ITsDockerConfig, IBuildCommandOptions } from './interfaces/index.js';
|
||||||
import type { TsDockerManager } from './classes.tsdockermanager.js';
|
import type { TsDockerManager } from './classes.tsdockermanager.js';
|
||||||
import * as fs from 'fs';
|
import * as fs from 'fs';
|
||||||
@@ -11,9 +12,14 @@ const smartshellInstance = new plugins.smartshell.Smartshell({
|
|||||||
executor: 'bash',
|
executor: 'bash',
|
||||||
});
|
});
|
||||||
|
|
||||||
const LOCAL_REGISTRY_PORT = 5234;
|
/**
|
||||||
const LOCAL_REGISTRY_HOST = `localhost:${LOCAL_REGISTRY_PORT}`;
|
* Extracts a platform string (e.g. "linux/amd64") from a buildx bracket prefix.
|
||||||
const LOCAL_REGISTRY_CONTAINER = 'tsdocker-local-registry';
|
* The prefix may be like "linux/amd64 ", "linux/amd64 stage-1 ", "stage-1 ", or "".
|
||||||
|
*/
|
||||||
|
function extractPlatform(prefix: string): string | null {
|
||||||
|
const match = prefix.match(/linux\/\w+/);
|
||||||
|
return match ? match[0] : null;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Class Dockerfile represents a Dockerfile on disk
|
* Class Dockerfile represents a Dockerfile on disk
|
||||||
@@ -148,40 +154,55 @@ export class Dockerfile {
|
|||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
/** Starts a persistent registry:2 container on port 5234 with volume storage. */
|
/** Starts a persistent registry:2 container with session-unique port and name. */
|
||||||
public static async startLocalRegistry(isRootless?: boolean): Promise<void> {
|
public static async startLocalRegistry(session: TsDockerSession, isRootless?: boolean): Promise<void> {
|
||||||
// Ensure persistent storage directory exists
|
const { registryPort, registryHost, registryContainerName, isCI, sessionId } = session.config;
|
||||||
const registryDataDir = plugins.path.join(paths.cwd, '.nogit', 'docker-registry');
|
|
||||||
|
// Ensure persistent storage directory exists — isolate per session in CI
|
||||||
|
const registryDataDir = isCI
|
||||||
|
? plugins.path.join(paths.cwd, '.nogit', 'docker-registry', sessionId)
|
||||||
|
: plugins.path.join(paths.cwd, '.nogit', 'docker-registry');
|
||||||
fs.mkdirSync(registryDataDir, { recursive: true });
|
fs.mkdirSync(registryDataDir, { recursive: true });
|
||||||
|
|
||||||
await smartshellInstance.execSilent(
|
await smartshellInstance.execSilent(
|
||||||
`docker rm -f ${LOCAL_REGISTRY_CONTAINER} 2>/dev/null || true`
|
`docker rm -f ${registryContainerName} 2>/dev/null || true`
|
||||||
);
|
|
||||||
const result = await smartshellInstance.execSilent(
|
|
||||||
`docker run -d --name ${LOCAL_REGISTRY_CONTAINER} -p ${LOCAL_REGISTRY_PORT}:5000 -v "${registryDataDir}:/var/lib/registry" registry:2`
|
|
||||||
);
|
);
|
||||||
|
|
||||||
|
const runCmd = `docker run -d --name ${registryContainerName} -p ${registryPort}:5000 -v "${registryDataDir}:/var/lib/registry" registry:2`;
|
||||||
|
let result = await smartshellInstance.execSilent(runCmd);
|
||||||
|
|
||||||
|
// Port retry: if port was stolen between allocation and docker run, reallocate once
|
||||||
|
if (result.exitCode !== 0 && (result.stderr || result.stdout || '').includes('port is already allocated')) {
|
||||||
|
const newPort = await TsDockerSession.allocatePort();
|
||||||
|
logger.log('warn', `Port ${registryPort} taken, retrying with ${newPort}`);
|
||||||
|
session.config.registryPort = newPort;
|
||||||
|
session.config.registryHost = `localhost:${newPort}`;
|
||||||
|
const retryCmd = `docker run -d --name ${registryContainerName} -p ${newPort}:5000 -v "${registryDataDir}:/var/lib/registry" registry:2`;
|
||||||
|
result = await smartshellInstance.execSilent(retryCmd);
|
||||||
|
}
|
||||||
|
|
||||||
if (result.exitCode !== 0) {
|
if (result.exitCode !== 0) {
|
||||||
throw new Error(`Failed to start local registry: ${result.stderr || result.stdout}`);
|
throw new Error(`Failed to start local registry: ${result.stderr || result.stdout}`);
|
||||||
}
|
}
|
||||||
// registry:2 starts near-instantly; brief wait for readiness
|
// registry:2 starts near-instantly; brief wait for readiness
|
||||||
await new Promise(resolve => setTimeout(resolve, 1000));
|
await new Promise(resolve => setTimeout(resolve, 1000));
|
||||||
logger.log('info', `Started local registry at ${LOCAL_REGISTRY_HOST} (persistent storage at .nogit/docker-registry/)`);
|
logger.log('info', `Started local registry at ${session.config.registryHost} (container: ${registryContainerName})`);
|
||||||
if (isRootless) {
|
if (isRootless) {
|
||||||
logger.log('warn', `[rootless] Registry on port ${LOCAL_REGISTRY_PORT} — if buildx cannot reach localhost:${LOCAL_REGISTRY_PORT}, try 127.0.0.1:${LOCAL_REGISTRY_PORT}`);
|
logger.log('warn', `[rootless] Registry on port ${session.config.registryPort} — if buildx cannot reach localhost, try 127.0.0.1`);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/** Stops and removes the temporary local registry container. */
|
/** Stops and removes the session-specific local registry container. */
|
||||||
public static async stopLocalRegistry(): Promise<void> {
|
public static async stopLocalRegistry(session: TsDockerSession): Promise<void> {
|
||||||
await smartshellInstance.execSilent(
|
await smartshellInstance.execSilent(
|
||||||
`docker rm -f ${LOCAL_REGISTRY_CONTAINER} 2>/dev/null || true`
|
`docker rm -f ${session.config.registryContainerName} 2>/dev/null || true`
|
||||||
);
|
);
|
||||||
logger.log('info', 'Stopped local registry');
|
logger.log('info', `Stopped local registry (${session.config.registryContainerName})`);
|
||||||
}
|
}
|
||||||
|
|
||||||
/** Pushes a built image to the local registry for buildx consumption. */
|
/** Pushes a built image to the local registry for buildx consumption. */
|
||||||
public static async pushToLocalRegistry(dockerfile: Dockerfile): Promise<void> {
|
public static async pushToLocalRegistry(session: TsDockerSession, dockerfile: Dockerfile): Promise<void> {
|
||||||
const registryTag = `${LOCAL_REGISTRY_HOST}/${dockerfile.buildTag}`;
|
const registryTag = `${session.config.registryHost}/${dockerfile.buildTag}`;
|
||||||
await smartshellInstance.execSilent(`docker tag ${dockerfile.buildTag} ${registryTag}`);
|
await smartshellInstance.execSilent(`docker tag ${dockerfile.buildTag} ${registryTag}`);
|
||||||
const result = await smartshellInstance.execSilent(`docker push ${registryTag}`);
|
const result = await smartshellInstance.execSilent(`docker push ${registryTag}`);
|
||||||
if (result.exitCode !== 0) {
|
if (result.exitCode !== 0) {
|
||||||
@@ -244,12 +265,16 @@ export class Dockerfile {
|
|||||||
*/
|
*/
|
||||||
public static async buildDockerfiles(
|
public static async buildDockerfiles(
|
||||||
sortedArrayArg: Dockerfile[],
|
sortedArrayArg: Dockerfile[],
|
||||||
options?: { platform?: string; timeout?: number; noCache?: boolean; verbose?: boolean; isRootless?: boolean; parallel?: boolean; parallelConcurrency?: number },
|
session: TsDockerSession,
|
||||||
|
options?: { platform?: string; timeout?: number; noCache?: boolean; verbose?: boolean; isRootless?: boolean; parallel?: boolean; parallelConcurrency?: number; onRegistryStarted?: () => Promise<void>; onBeforeRegistryStop?: () => Promise<void> },
|
||||||
): Promise<Dockerfile[]> {
|
): Promise<Dockerfile[]> {
|
||||||
const total = sortedArrayArg.length;
|
const total = sortedArrayArg.length;
|
||||||
const overallStart = Date.now();
|
const overallStart = Date.now();
|
||||||
|
|
||||||
await Dockerfile.startLocalRegistry(options?.isRootless);
|
await Dockerfile.startLocalRegistry(session, options?.isRootless);
|
||||||
|
if (options?.onRegistryStarted) {
|
||||||
|
await options.onRegistryStarted();
|
||||||
|
}
|
||||||
|
|
||||||
try {
|
try {
|
||||||
if (options?.parallel) {
|
if (options?.parallel) {
|
||||||
@@ -296,7 +321,7 @@ export class Dockerfile {
|
|||||||
}
|
}
|
||||||
// Push ALL images to local registry (skip if already pushed via buildx)
|
// Push ALL images to local registry (skip if already pushed via buildx)
|
||||||
if (!df.localRegistryTag) {
|
if (!df.localRegistryTag) {
|
||||||
await Dockerfile.pushToLocalRegistry(df);
|
await Dockerfile.pushToLocalRegistry(session, df);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -324,12 +349,15 @@ export class Dockerfile {
|
|||||||
|
|
||||||
// Push ALL images to local registry (skip if already pushed via buildx)
|
// Push ALL images to local registry (skip if already pushed via buildx)
|
||||||
if (!dockerfileArg.localRegistryTag) {
|
if (!dockerfileArg.localRegistryTag) {
|
||||||
await Dockerfile.pushToLocalRegistry(dockerfileArg);
|
await Dockerfile.pushToLocalRegistry(session, dockerfileArg);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
} finally {
|
} finally {
|
||||||
await Dockerfile.stopLocalRegistry();
|
if (options?.onBeforeRegistryStop) {
|
||||||
|
await options.onBeforeRegistryStop();
|
||||||
|
}
|
||||||
|
await Dockerfile.stopLocalRegistry(session);
|
||||||
}
|
}
|
||||||
|
|
||||||
logger.log('info', `Total build time: ${formatDuration(Date.now() - overallStart)}`);
|
logger.log('info', `Total build time: ${formatDuration(Date.now() - overallStart)}`);
|
||||||
@@ -520,6 +548,7 @@ export class Dockerfile {
|
|||||||
|
|
||||||
// INSTANCE PROPERTIES
|
// INSTANCE PROPERTIES
|
||||||
public managerRef: TsDockerManager;
|
public managerRef: TsDockerManager;
|
||||||
|
public session?: TsDockerSession;
|
||||||
public filePath!: string;
|
public filePath!: string;
|
||||||
public repo: string;
|
public repo: string;
|
||||||
public version: string;
|
public version: string;
|
||||||
@@ -563,6 +592,79 @@ export class Dockerfile {
|
|||||||
this.localBaseImageDependent = false;
|
this.localBaseImageDependent = false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Creates a line-by-line handler for Docker build output that logs
|
||||||
|
* recognized layer/step lines in an emphasized format.
|
||||||
|
*/
|
||||||
|
private createBuildOutputHandler(verbose: boolean): {
|
||||||
|
handleChunk: (chunk: Buffer | string) => void;
|
||||||
|
} {
|
||||||
|
let buffer = '';
|
||||||
|
const tag = this.cleanTag;
|
||||||
|
|
||||||
|
const handleLine = (line: string) => {
|
||||||
|
// In verbose mode, write raw output prefixed with tag for identification
|
||||||
|
if (verbose) {
|
||||||
|
process.stdout.write(`[${tag}] ${line}\n`);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Buildx step: #N [platform step/total] INSTRUCTION
|
||||||
|
const bxStep = line.match(/^#\d+ \[([^\]]+?)(\d+\/\d+)\] (.+)/);
|
||||||
|
if (bxStep) {
|
||||||
|
const prefix = bxStep[1].trim();
|
||||||
|
const step = bxStep[2];
|
||||||
|
const instruction = bxStep[3];
|
||||||
|
const platform = extractPlatform(prefix);
|
||||||
|
const platStr = platform ? `${platform} ▸ ` : '';
|
||||||
|
logger.log('note', `[${tag}] ${platStr}[${step}] ${instruction}`);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Buildx CACHED: #N CACHED
|
||||||
|
const bxCached = line.match(/^#(\d+) CACHED/);
|
||||||
|
if (bxCached) {
|
||||||
|
logger.log('note', `[${tag}] CACHED`);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Buildx DONE: #N DONE 12.3s
|
||||||
|
const bxDone = line.match(/^#\d+ DONE (.+)/);
|
||||||
|
if (bxDone) {
|
||||||
|
const timing = bxDone[1];
|
||||||
|
if (!timing.startsWith('0.0')) {
|
||||||
|
logger.log('note', `[${tag}] DONE ${timing}`);
|
||||||
|
}
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Buildx export phase: #N exporting ...
|
||||||
|
const bxExport = line.match(/^#\d+ exporting (.+)/);
|
||||||
|
if (bxExport) {
|
||||||
|
logger.log('note', `[${tag}] exporting ${bxExport[1]}`);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Standard docker build: Step N/M : INSTRUCTION
|
||||||
|
const stdStep = line.match(/^Step (\d+\/\d+) : (.+)/);
|
||||||
|
if (stdStep) {
|
||||||
|
logger.log('note', `[${tag}] Step ${stdStep[1]}: ${stdStep[2]}`);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
return {
|
||||||
|
handleChunk: (chunk: Buffer | string) => {
|
||||||
|
buffer += chunk.toString();
|
||||||
|
const lines = buffer.split('\n');
|
||||||
|
buffer = lines.pop() || '';
|
||||||
|
for (const line of lines) {
|
||||||
|
const trimmed = line.replace(/\r$/, '').trim();
|
||||||
|
if (trimmed) handleLine(trimmed);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Builds the Dockerfile
|
* Builds the Dockerfile
|
||||||
*/
|
*/
|
||||||
@@ -590,27 +692,32 @@ export class Dockerfile {
|
|||||||
|
|
||||||
if (platformOverride) {
|
if (platformOverride) {
|
||||||
// Single platform override via buildx
|
// Single platform override via buildx
|
||||||
buildCommand = `docker buildx build --platform ${platformOverride}${noCacheFlag}${buildContextFlag} --load -t ${this.buildTag} -f ${this.filePath} ${buildArgsString} .`;
|
buildCommand = `docker buildx build --progress=plain --platform ${platformOverride}${noCacheFlag}${buildContextFlag} --load -t ${this.buildTag} -f ${this.filePath} ${buildArgsString} .`;
|
||||||
logger.log('info', `Build: buildx --platform ${platformOverride} --load`);
|
logger.log('info', `Build: buildx --platform ${platformOverride} --load`);
|
||||||
} else if (config.platforms && config.platforms.length > 1) {
|
} else if (config.platforms && config.platforms.length > 1) {
|
||||||
// Multi-platform build using buildx — always push to local registry
|
// Multi-platform build using buildx — always push to local registry
|
||||||
const platformString = config.platforms.join(',');
|
const platformString = config.platforms.join(',');
|
||||||
const localTag = `${LOCAL_REGISTRY_HOST}/${this.buildTag}`;
|
const registryHost = this.session?.config.registryHost || 'localhost:5234';
|
||||||
buildCommand = `docker buildx build --platform ${platformString}${noCacheFlag}${buildContextFlag} -t ${localTag} -f ${this.filePath} ${buildArgsString} --push .`;
|
const localTag = `${registryHost}/${this.buildTag}`;
|
||||||
|
buildCommand = `docker buildx build --progress=plain --platform ${platformString}${noCacheFlag}${buildContextFlag} -t ${localTag} -f ${this.filePath} ${buildArgsString} --push .`;
|
||||||
this.localRegistryTag = localTag;
|
this.localRegistryTag = localTag;
|
||||||
logger.log('info', `Build: buildx --platform ${platformString} --push to local registry`);
|
logger.log('info', `Build: buildx --platform ${platformString} --push to local registry`);
|
||||||
} else {
|
} else {
|
||||||
// Standard build
|
// Standard build
|
||||||
const versionLabel = this.managerRef.projectInfo?.npm?.version || 'unknown';
|
const versionLabel = this.managerRef.projectInfo?.npm?.version || 'unknown';
|
||||||
buildCommand = `docker build --label="version=${versionLabel}"${noCacheFlag} -t ${this.buildTag} -f ${this.filePath} ${buildArgsString} .`;
|
buildCommand = `docker build --progress=plain --label="version=${versionLabel}"${noCacheFlag} -t ${this.buildTag} -f ${this.filePath} ${buildArgsString} .`;
|
||||||
logger.log('info', 'Build: docker build (standard)');
|
logger.log('info', 'Build: docker build (standard)');
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Execute build with real-time layer logging
|
||||||
|
const handler = this.createBuildOutputHandler(verbose);
|
||||||
|
const streaming = await smartshellInstance.execStreamingSilent(buildCommand);
|
||||||
|
|
||||||
|
// Intercept output for layer logging
|
||||||
|
streaming.childProcess.stdout?.on('data', handler.handleChunk);
|
||||||
|
streaming.childProcess.stderr?.on('data', handler.handleChunk);
|
||||||
|
|
||||||
if (timeout) {
|
if (timeout) {
|
||||||
// Use streaming execution with timeout
|
|
||||||
const streaming = verbose
|
|
||||||
? await smartshellInstance.execStreaming(buildCommand)
|
|
||||||
: await smartshellInstance.execStreamingSilent(buildCommand);
|
|
||||||
const timeoutPromise = new Promise<never>((_, reject) => {
|
const timeoutPromise = new Promise<never>((_, reject) => {
|
||||||
setTimeout(() => {
|
setTimeout(() => {
|
||||||
streaming.childProcess.kill();
|
streaming.childProcess.kill();
|
||||||
@@ -623,9 +730,7 @@ export class Dockerfile {
|
|||||||
throw new Error(`Build failed for ${this.cleanTag}`);
|
throw new Error(`Build failed for ${this.cleanTag}`);
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
const result = verbose
|
const result = await streaming.finalPromise;
|
||||||
? await smartshellInstance.exec(buildCommand)
|
|
||||||
: await smartshellInstance.execSilent(buildCommand);
|
|
||||||
if (result.exitCode !== 0) {
|
if (result.exitCode !== 0) {
|
||||||
logger.log('error', `Build failed for ${this.cleanTag}`);
|
logger.log('error', `Build failed for ${this.cleanTag}`);
|
||||||
if (!verbose && result.stdout) {
|
if (!verbose && result.stdout) {
|
||||||
@@ -646,12 +751,13 @@ export class Dockerfile {
|
|||||||
const destRepo = this.getDestRepo(dockerRegistryArg.registryUrl);
|
const destRepo = this.getDestRepo(dockerRegistryArg.registryUrl);
|
||||||
const destTag = versionSuffix ? `${this.version}_${versionSuffix}` : this.version;
|
const destTag = versionSuffix ? `${this.version}_${versionSuffix}` : this.version;
|
||||||
const registryCopy = new RegistryCopy();
|
const registryCopy = new RegistryCopy();
|
||||||
|
const registryHost = this.session?.config.registryHost || 'localhost:5234';
|
||||||
|
|
||||||
this.pushTag = `${dockerRegistryArg.registryUrl}/${destRepo}:${destTag}`;
|
this.pushTag = `${dockerRegistryArg.registryUrl}/${destRepo}:${destTag}`;
|
||||||
logger.log('info', `Pushing ${this.pushTag} via OCI copy from local registry...`);
|
logger.log('info', `Pushing ${this.pushTag} via OCI copy from local registry...`);
|
||||||
|
|
||||||
await registryCopy.copyImage(
|
await registryCopy.copyImage(
|
||||||
LOCAL_REGISTRY_HOST,
|
registryHost,
|
||||||
this.repo,
|
this.repo,
|
||||||
this.version,
|
this.version,
|
||||||
dockerRegistryArg.registryUrl,
|
dockerRegistryArg.registryUrl,
|
||||||
@@ -701,23 +807,27 @@ export class Dockerfile {
|
|||||||
// Use local registry tag for multi-platform images (not in daemon), otherwise buildTag
|
// Use local registry tag for multi-platform images (not in daemon), otherwise buildTag
|
||||||
const imageRef = this.localRegistryTag || this.buildTag;
|
const imageRef = this.localRegistryTag || this.buildTag;
|
||||||
|
|
||||||
|
const sessionId = this.session?.config.sessionId || 'default';
|
||||||
|
const testContainerName = `tsdocker_test_${sessionId}`;
|
||||||
|
const testImageName = `tsdocker_test_image_${sessionId}`;
|
||||||
|
|
||||||
const testFileExists = fs.existsSync(testFile);
|
const testFileExists = fs.existsSync(testFile);
|
||||||
|
|
||||||
if (testFileExists) {
|
if (testFileExists) {
|
||||||
// Run tests in container
|
// Run tests in container
|
||||||
await smartshellInstance.exec(
|
await smartshellInstance.exec(
|
||||||
`docker run --name tsdocker_test_container --entrypoint="bash" ${imageRef} -c "mkdir /tsdocker_test"`
|
`docker run --name ${testContainerName} --entrypoint="bash" ${imageRef} -c "mkdir /tsdocker_test"`
|
||||||
);
|
);
|
||||||
await smartshellInstance.exec(`docker cp ${testFile} tsdocker_test_container:/tsdocker_test/test.sh`);
|
await smartshellInstance.exec(`docker cp ${testFile} ${testContainerName}:/tsdocker_test/test.sh`);
|
||||||
await smartshellInstance.exec(`docker commit tsdocker_test_container tsdocker_test_image`);
|
await smartshellInstance.exec(`docker commit ${testContainerName} ${testImageName}`);
|
||||||
|
|
||||||
const testResult = await smartshellInstance.exec(
|
const testResult = await smartshellInstance.exec(
|
||||||
`docker run --entrypoint="bash" tsdocker_test_image -x /tsdocker_test/test.sh`
|
`docker run --entrypoint="bash" ${testImageName} -x /tsdocker_test/test.sh`
|
||||||
);
|
);
|
||||||
|
|
||||||
// Cleanup
|
// Cleanup
|
||||||
await smartshellInstance.exec(`docker rm tsdocker_test_container`);
|
await smartshellInstance.exec(`docker rm ${testContainerName}`);
|
||||||
await smartshellInstance.exec(`docker rmi --force tsdocker_test_image`);
|
await smartshellInstance.exec(`docker rmi --force ${testImageName}`);
|
||||||
|
|
||||||
if (testResult.exitCode !== 0) {
|
if (testResult.exitCode !== 0) {
|
||||||
throw new Error(`Tests failed for ${this.cleanTag}`);
|
throw new Error(`Tests failed for ${this.cleanTag}`);
|
||||||
|
|||||||
76
ts/classes.globalconfig.ts
Normal file
76
ts/classes.globalconfig.ts
Normal file
@@ -0,0 +1,76 @@
|
|||||||
|
import * as fs from 'fs';
|
||||||
|
import * as plugins from './tsdocker.plugins.js';
|
||||||
|
import { logger } from './tsdocker.logging.js';
|
||||||
|
import type { IGlobalConfig, IRemoteBuilder } from './interfaces/index.js';
|
||||||
|
|
||||||
|
const CONFIG_DIR = plugins.path.join(
|
||||||
|
process.env.HOME || process.env.USERPROFILE || '~',
|
||||||
|
'.git.zone',
|
||||||
|
'tsdocker',
|
||||||
|
);
|
||||||
|
const CONFIG_PATH = plugins.path.join(CONFIG_DIR, 'config.json');
|
||||||
|
|
||||||
|
const DEFAULT_CONFIG: IGlobalConfig = {
|
||||||
|
remoteBuilders: [],
|
||||||
|
};
|
||||||
|
|
||||||
|
export class GlobalConfig {
|
||||||
|
static getConfigPath(): string {
|
||||||
|
return CONFIG_PATH;
|
||||||
|
}
|
||||||
|
|
||||||
|
static load(): IGlobalConfig {
|
||||||
|
try {
|
||||||
|
const raw = fs.readFileSync(CONFIG_PATH, 'utf-8');
|
||||||
|
const parsed = JSON.parse(raw);
|
||||||
|
return {
|
||||||
|
...DEFAULT_CONFIG,
|
||||||
|
...parsed,
|
||||||
|
};
|
||||||
|
} catch {
|
||||||
|
return { ...DEFAULT_CONFIG };
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
static save(config: IGlobalConfig): void {
|
||||||
|
fs.mkdirSync(CONFIG_DIR, { recursive: true });
|
||||||
|
fs.writeFileSync(CONFIG_PATH, JSON.stringify(config, null, 2) + '\n', 'utf-8');
|
||||||
|
}
|
||||||
|
|
||||||
|
static addBuilder(builder: IRemoteBuilder): void {
|
||||||
|
const config = GlobalConfig.load();
|
||||||
|
const existing = config.remoteBuilders.findIndex((b) => b.name === builder.name);
|
||||||
|
if (existing >= 0) {
|
||||||
|
config.remoteBuilders[existing] = builder;
|
||||||
|
logger.log('info', `Updated remote builder: ${builder.name}`);
|
||||||
|
} else {
|
||||||
|
config.remoteBuilders.push(builder);
|
||||||
|
logger.log('info', `Added remote builder: ${builder.name}`);
|
||||||
|
}
|
||||||
|
GlobalConfig.save(config);
|
||||||
|
}
|
||||||
|
|
||||||
|
static removeBuilder(name: string): void {
|
||||||
|
const config = GlobalConfig.load();
|
||||||
|
const before = config.remoteBuilders.length;
|
||||||
|
config.remoteBuilders = config.remoteBuilders.filter((b) => b.name !== name);
|
||||||
|
if (config.remoteBuilders.length < before) {
|
||||||
|
logger.log('info', `Removed remote builder: ${name}`);
|
||||||
|
} else {
|
||||||
|
logger.log('warn', `Remote builder not found: ${name}`);
|
||||||
|
}
|
||||||
|
GlobalConfig.save(config);
|
||||||
|
}
|
||||||
|
|
||||||
|
static getBuilders(): IRemoteBuilder[] {
|
||||||
|
return GlobalConfig.load().remoteBuilders;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns remote builders that match any of the requested platforms
|
||||||
|
*/
|
||||||
|
static getBuildersForPlatforms(platforms: string[]): IRemoteBuilder[] {
|
||||||
|
const builders = GlobalConfig.getBuilders();
|
||||||
|
return builders.filter((b) => platforms.includes(b.platform));
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -20,6 +20,53 @@ interface ITokenCache {
|
|||||||
export class RegistryCopy {
|
export class RegistryCopy {
|
||||||
private tokenCache: ITokenCache = {};
|
private tokenCache: ITokenCache = {};
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Wraps fetch() with timeout (via AbortSignal) and retry with exponential backoff.
|
||||||
|
* Retries on network errors and 5xx; does NOT retry on 4xx client errors.
|
||||||
|
* On 401, clears the token cache entry so the next attempt re-authenticates.
|
||||||
|
*/
|
||||||
|
private async fetchWithRetry(
|
||||||
|
url: string,
|
||||||
|
options: RequestInit & { duplex?: string },
|
||||||
|
timeoutMs: number = 300_000,
|
||||||
|
maxRetries: number = 6,
|
||||||
|
): Promise<Response> {
|
||||||
|
const method = (options.method || 'GET').toUpperCase();
|
||||||
|
let lastError: Error | null = null;
|
||||||
|
for (let attempt = 1; attempt <= maxRetries; attempt++) {
|
||||||
|
try {
|
||||||
|
if (attempt > 1) {
|
||||||
|
logger.log('info', `Retry ${attempt}/${maxRetries} for ${method} ${url}`);
|
||||||
|
}
|
||||||
|
const resp = await fetch(url, {
|
||||||
|
...options,
|
||||||
|
signal: AbortSignal.timeout(timeoutMs),
|
||||||
|
});
|
||||||
|
// Retry on 5xx server errors (but not 4xx)
|
||||||
|
if (resp.status >= 500 && attempt < maxRetries) {
|
||||||
|
const delay = 1000 * Math.pow(2, attempt - 1);
|
||||||
|
logger.log('warn', `${method} ${url} returned ${resp.status}, retrying in ${delay}ms (attempt ${attempt}/${maxRetries})...`);
|
||||||
|
await new Promise(r => setTimeout(r, delay));
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
if (resp.status >= 500) {
|
||||||
|
logger.log('error', `${method} ${url} returned ${resp.status} after ${maxRetries} attempts, giving up`);
|
||||||
|
}
|
||||||
|
return resp;
|
||||||
|
} catch (err) {
|
||||||
|
lastError = err as Error;
|
||||||
|
if (attempt < maxRetries) {
|
||||||
|
const delay = 1000 * Math.pow(2, attempt - 1);
|
||||||
|
logger.log('warn', `${method} ${url} failed (attempt ${attempt}/${maxRetries}): ${lastError.message}, retrying in ${delay}ms...`);
|
||||||
|
await new Promise(r => setTimeout(r, delay));
|
||||||
|
} else {
|
||||||
|
logger.log('error', `${method} ${url} failed after ${maxRetries} attempts: ${lastError.message}`);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
throw lastError!;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Reads Docker credentials from ~/.docker/config.json for a given registry.
|
* Reads Docker credentials from ~/.docker/config.json for a given registry.
|
||||||
* Supports base64-encoded "auth" field in the config.
|
* Supports base64-encoded "auth" field in the config.
|
||||||
@@ -109,7 +156,7 @@ export class RegistryCopy {
|
|||||||
}
|
}
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const checkResp = await fetch(`${apiBase}/v2/`, { method: 'GET' });
|
const checkResp = await this.fetchWithRetry(`${apiBase}/v2/`, { method: 'GET' }, 30_000);
|
||||||
if (checkResp.ok) return null; // No auth needed
|
if (checkResp.ok) return null; // No auth needed
|
||||||
|
|
||||||
const wwwAuth = checkResp.headers.get('www-authenticate') || '';
|
const wwwAuth = checkResp.headers.get('www-authenticate') || '';
|
||||||
@@ -131,7 +178,7 @@ export class RegistryCopy {
|
|||||||
headers['Authorization'] = 'Basic ' + Buffer.from(`${creds.username}:${creds.password}`).toString('base64');
|
headers['Authorization'] = 'Basic ' + Buffer.from(`${creds.username}:${creds.password}`).toString('base64');
|
||||||
}
|
}
|
||||||
|
|
||||||
const tokenResp = await fetch(tokenUrl.toString(), { headers });
|
const tokenResp = await this.fetchWithRetry(tokenUrl.toString(), { headers }, 30_000);
|
||||||
if (!tokenResp.ok) {
|
if (!tokenResp.ok) {
|
||||||
const body = await tokenResp.text();
|
const body = await tokenResp.text();
|
||||||
throw new Error(`Token request failed (${tokenResp.status}): ${body}`);
|
throw new Error(`Token request failed (${tokenResp.status}): ${body}`);
|
||||||
@@ -189,7 +236,16 @@ export class RegistryCopy {
|
|||||||
fetchOptions.duplex = 'half'; // Required for streaming body in Node
|
fetchOptions.duplex = 'half'; // Required for streaming body in Node
|
||||||
}
|
}
|
||||||
|
|
||||||
return fetch(url, fetchOptions);
|
const resp = await this.fetchWithRetry(url, fetchOptions, 300_000);
|
||||||
|
|
||||||
|
// Token expired — clear cache so next call re-authenticates
|
||||||
|
if (resp.status === 401 && token) {
|
||||||
|
const cacheKey = `${registry}/${`repository:${repo}:${actions}`}`;
|
||||||
|
logger.log('warn', `Got 401 for ${registry}${path} — clearing cached token for ${cacheKey}`);
|
||||||
|
delete this.tokenCache[cacheKey];
|
||||||
|
}
|
||||||
|
|
||||||
|
return resp;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -320,11 +376,11 @@ export class RegistryCopy {
|
|||||||
putHeaders['Authorization'] = `Bearer ${token}`;
|
putHeaders['Authorization'] = `Bearer ${token}`;
|
||||||
}
|
}
|
||||||
|
|
||||||
const putResp = await fetch(putUrl, {
|
const putResp = await this.fetchWithRetry(putUrl, {
|
||||||
method: 'PUT',
|
method: 'PUT',
|
||||||
headers: putHeaders,
|
headers: putHeaders,
|
||||||
body: blobData,
|
body: blobData,
|
||||||
});
|
}, 300_000);
|
||||||
|
|
||||||
if (!putResp.ok) {
|
if (!putResp.ok) {
|
||||||
const body = await putResp.text();
|
const body = await putResp.text();
|
||||||
|
|||||||
77
ts/classes.sshtunnel.ts
Normal file
77
ts/classes.sshtunnel.ts
Normal file
@@ -0,0 +1,77 @@
|
|||||||
|
import * as plugins from './tsdocker.plugins.js';
|
||||||
|
import { logger } from './tsdocker.logging.js';
|
||||||
|
import type { IRemoteBuilder } from './interfaces/index.js';
|
||||||
|
|
||||||
|
const smartshellInstance = new plugins.smartshell.Smartshell({
|
||||||
|
executor: 'bash',
|
||||||
|
});
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Manages SSH reverse tunnels for remote builder nodes.
|
||||||
|
* Opens tunnels so that the local staging registry (localhost:<port>)
|
||||||
|
* is accessible as localhost:<port> on each remote machine.
|
||||||
|
*/
|
||||||
|
export class SshTunnelManager {
|
||||||
|
private tunnelPids: number[] = [];
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Opens a reverse SSH tunnel to make localPort accessible on the remote machine.
|
||||||
|
* ssh -f -N -o StrictHostKeyChecking=no -o ExitOnForwardFailure=yes
|
||||||
|
* -R <localPort>:localhost:<localPort> [-i keyPath] user@host
|
||||||
|
*/
|
||||||
|
async openTunnel(builder: IRemoteBuilder, localPort: number): Promise<void> {
|
||||||
|
const keyOpt = builder.sshKeyPath ? `-i ${builder.sshKeyPath} ` : '';
|
||||||
|
const cmd = [
|
||||||
|
'ssh -f -N',
|
||||||
|
'-o StrictHostKeyChecking=no',
|
||||||
|
'-o ExitOnForwardFailure=yes',
|
||||||
|
`-R ${localPort}:localhost:${localPort}`,
|
||||||
|
`${keyOpt}${builder.host}`,
|
||||||
|
].join(' ');
|
||||||
|
|
||||||
|
logger.log('info', `Opening SSH tunnel to ${builder.host} for port ${localPort}...`);
|
||||||
|
const result = await smartshellInstance.exec(cmd);
|
||||||
|
|
||||||
|
if (result.exitCode !== 0) {
|
||||||
|
throw new Error(
|
||||||
|
`Failed to open SSH tunnel to ${builder.host}: ${result.stderr || 'unknown error'}`
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Find the PID of the tunnel process we just started
|
||||||
|
const pidResult = await smartshellInstance.exec(
|
||||||
|
`pgrep -f "ssh.*-R ${localPort}:localhost:${localPort}.*${builder.host}" | tail -1`
|
||||||
|
);
|
||||||
|
if (pidResult.exitCode === 0 && pidResult.stdout.trim()) {
|
||||||
|
const pid = parseInt(pidResult.stdout.trim(), 10);
|
||||||
|
if (!isNaN(pid)) {
|
||||||
|
this.tunnelPids.push(pid);
|
||||||
|
logger.log('ok', `SSH tunnel to ${builder.host} established (PID ${pid})`);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Opens tunnels for all provided remote builders
|
||||||
|
*/
|
||||||
|
async openTunnels(builders: IRemoteBuilder[], localPort: number): Promise<void> {
|
||||||
|
for (const builder of builders) {
|
||||||
|
await this.openTunnel(builder, localPort);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Closes all tunnel processes
|
||||||
|
*/
|
||||||
|
async closeAll(): Promise<void> {
|
||||||
|
for (const pid of this.tunnelPids) {
|
||||||
|
try {
|
||||||
|
process.kill(pid, 'SIGTERM');
|
||||||
|
logger.log('info', `Closed SSH tunnel (PID ${pid})`);
|
||||||
|
} catch {
|
||||||
|
// Process may have already exited
|
||||||
|
}
|
||||||
|
}
|
||||||
|
this.tunnelPids = [];
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -6,7 +6,11 @@ import { DockerRegistry } from './classes.dockerregistry.js';
|
|||||||
import { RegistryStorage } from './classes.registrystorage.js';
|
import { RegistryStorage } from './classes.registrystorage.js';
|
||||||
import { TsDockerCache } from './classes.tsdockercache.js';
|
import { TsDockerCache } from './classes.tsdockercache.js';
|
||||||
import { DockerContext } from './classes.dockercontext.js';
|
import { DockerContext } from './classes.dockercontext.js';
|
||||||
import type { ITsDockerConfig, IBuildCommandOptions } from './interfaces/index.js';
|
import { TsDockerSession } from './classes.tsdockersession.js';
|
||||||
|
import { RegistryCopy } from './classes.registrycopy.js';
|
||||||
|
import { GlobalConfig } from './classes.globalconfig.js';
|
||||||
|
import { SshTunnelManager } from './classes.sshtunnel.js';
|
||||||
|
import type { ITsDockerConfig, IBuildCommandOptions, IRemoteBuilder } from './interfaces/index.js';
|
||||||
|
|
||||||
const smartshellInstance = new plugins.smartshell.Smartshell({
|
const smartshellInstance = new plugins.smartshell.Smartshell({
|
||||||
executor: 'bash',
|
executor: 'bash',
|
||||||
@@ -20,7 +24,10 @@ export class TsDockerManager {
|
|||||||
public config: ITsDockerConfig;
|
public config: ITsDockerConfig;
|
||||||
public projectInfo: any;
|
public projectInfo: any;
|
||||||
public dockerContext: DockerContext;
|
public dockerContext: DockerContext;
|
||||||
|
public session!: TsDockerSession;
|
||||||
private dockerfiles: Dockerfile[] = [];
|
private dockerfiles: Dockerfile[] = [];
|
||||||
|
private activeRemoteBuilders: IRemoteBuilder[] = [];
|
||||||
|
private sshTunnelManager?: SshTunnelManager;
|
||||||
|
|
||||||
constructor(config: ITsDockerConfig) {
|
constructor(config: ITsDockerConfig) {
|
||||||
this.config = config;
|
this.config = config;
|
||||||
@@ -74,9 +81,28 @@ export class TsDockerManager {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Fallback: check ~/.docker/config.json if env vars didn't provide credentials
|
||||||
|
if (!this.registryStorage.getRegistryByUrl(registryUrl)) {
|
||||||
|
const dockerConfigCreds = RegistryCopy.getDockerConfigCredentials(registryUrl);
|
||||||
|
if (dockerConfigCreds) {
|
||||||
|
const registry = new DockerRegistry({
|
||||||
|
registryUrl,
|
||||||
|
username: dockerConfigCreds.username,
|
||||||
|
password: dockerConfigCreds.password,
|
||||||
|
});
|
||||||
|
this.registryStorage.addRegistry(registry);
|
||||||
|
logger.log('info', `Loaded credentials for ${registryUrl} from ~/.docker/config.json`);
|
||||||
|
} else {
|
||||||
|
logger.log('warn', `No credentials found for ${registryUrl} (checked env vars and ~/.docker/config.json)`);
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Create session identity (unique ports, names for CI concurrency)
|
||||||
|
this.session = await TsDockerSession.create();
|
||||||
|
|
||||||
logger.log('info', `Prepared TsDockerManager with ${this.registryStorage.getAllRegistries().length} registries`);
|
logger.log('info', `Prepared TsDockerManager with ${this.registryStorage.getAllRegistries().length} registries`);
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -98,9 +124,34 @@ export class TsDockerManager {
|
|||||||
this.dockerfiles = await Dockerfile.readDockerfiles(this);
|
this.dockerfiles = await Dockerfile.readDockerfiles(this);
|
||||||
this.dockerfiles = await Dockerfile.sortDockerfiles(this.dockerfiles);
|
this.dockerfiles = await Dockerfile.sortDockerfiles(this.dockerfiles);
|
||||||
this.dockerfiles = await Dockerfile.mapDockerfiles(this.dockerfiles);
|
this.dockerfiles = await Dockerfile.mapDockerfiles(this.dockerfiles);
|
||||||
|
// Inject session into each Dockerfile
|
||||||
|
for (const df of this.dockerfiles) {
|
||||||
|
df.session = this.session;
|
||||||
|
}
|
||||||
return this.dockerfiles;
|
return this.dockerfiles;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Filters discovered Dockerfiles by name patterns (glob-style).
|
||||||
|
* Mutates this.dockerfiles in place.
|
||||||
|
*/
|
||||||
|
public filterDockerfiles(patterns: string[]): void {
|
||||||
|
const matched = this.dockerfiles.filter((df) => {
|
||||||
|
const basename = plugins.path.basename(df.filePath);
|
||||||
|
return patterns.some((pattern) => {
|
||||||
|
if (pattern.includes('*') || pattern.includes('?')) {
|
||||||
|
const regexStr = '^' + pattern.replace(/\*/g, '.*').replace(/\?/g, '.') + '$';
|
||||||
|
return new RegExp(regexStr).test(basename);
|
||||||
|
}
|
||||||
|
return basename === pattern;
|
||||||
|
});
|
||||||
|
});
|
||||||
|
if (matched.length === 0) {
|
||||||
|
logger.log('warn', `No Dockerfiles matched patterns: ${patterns.join(', ')}`);
|
||||||
|
}
|
||||||
|
this.dockerfiles = matched;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Builds discovered Dockerfiles in dependency order.
|
* Builds discovered Dockerfiles in dependency order.
|
||||||
* When options.patterns is provided, only matching Dockerfiles (and their dependencies) are built.
|
* When options.patterns is provided, only matching Dockerfiles (and their dependencies) are built.
|
||||||
@@ -187,7 +238,8 @@ export class TsDockerManager {
|
|||||||
|
|
||||||
const total = toBuild.length;
|
const total = toBuild.length;
|
||||||
const overallStart = Date.now();
|
const overallStart = Date.now();
|
||||||
await Dockerfile.startLocalRegistry(this.dockerContext.contextInfo?.isRootless);
|
await Dockerfile.startLocalRegistry(this.session, this.dockerContext.contextInfo?.isRootless);
|
||||||
|
await this.openRemoteTunnels();
|
||||||
|
|
||||||
try {
|
try {
|
||||||
if (options?.parallel) {
|
if (options?.parallel) {
|
||||||
@@ -240,7 +292,7 @@ export class TsDockerManager {
|
|||||||
}
|
}
|
||||||
// Push ALL images to local registry (skip if already pushed via buildx)
|
// Push ALL images to local registry (skip if already pushed via buildx)
|
||||||
if (!df.localRegistryTag) {
|
if (!df.localRegistryTag) {
|
||||||
await Dockerfile.pushToLocalRegistry(df);
|
await Dockerfile.pushToLocalRegistry(this.session, df);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -280,19 +332,20 @@ export class TsDockerManager {
|
|||||||
|
|
||||||
// Push ALL images to local registry (skip if already pushed via buildx)
|
// Push ALL images to local registry (skip if already pushed via buildx)
|
||||||
if (!dockerfileArg.localRegistryTag) {
|
if (!dockerfileArg.localRegistryTag) {
|
||||||
await Dockerfile.pushToLocalRegistry(dockerfileArg);
|
await Dockerfile.pushToLocalRegistry(this.session, dockerfileArg);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
} finally {
|
} finally {
|
||||||
await Dockerfile.stopLocalRegistry();
|
await this.closeRemoteTunnels();
|
||||||
|
await Dockerfile.stopLocalRegistry(this.session);
|
||||||
}
|
}
|
||||||
|
|
||||||
logger.log('info', `Total build time: ${formatDuration(Date.now() - overallStart)}`);
|
logger.log('info', `Total build time: ${formatDuration(Date.now() - overallStart)}`);
|
||||||
cache.save();
|
cache.save();
|
||||||
} else {
|
} else {
|
||||||
// === STANDARD MODE: build all via static helper ===
|
// === STANDARD MODE: build all via static helper ===
|
||||||
await Dockerfile.buildDockerfiles(toBuild, {
|
await Dockerfile.buildDockerfiles(toBuild, this.session, {
|
||||||
platform: options?.platform,
|
platform: options?.platform,
|
||||||
timeout: options?.timeout,
|
timeout: options?.timeout,
|
||||||
noCache: options?.noCache,
|
noCache: options?.noCache,
|
||||||
@@ -300,6 +353,8 @@ export class TsDockerManager {
|
|||||||
isRootless: this.dockerContext.contextInfo?.isRootless,
|
isRootless: this.dockerContext.contextInfo?.isRootless,
|
||||||
parallel: options?.parallel,
|
parallel: options?.parallel,
|
||||||
parallelConcurrency: options?.parallelConcurrency,
|
parallelConcurrency: options?.parallelConcurrency,
|
||||||
|
onRegistryStarted: () => this.openRemoteTunnels(),
|
||||||
|
onBeforeRegistryStop: () => this.closeRemoteTunnels(),
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -326,13 +381,76 @@ export class TsDockerManager {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Ensures Docker buildx is set up for multi-architecture builds
|
* Ensures Docker buildx is set up for multi-architecture builds.
|
||||||
|
* When remote builders are configured in the global config, creates a multi-node
|
||||||
|
* builder with native nodes instead of relying on QEMU emulation.
|
||||||
*/
|
*/
|
||||||
private async ensureBuildx(): Promise<void> {
|
private async ensureBuildx(): Promise<void> {
|
||||||
const builderName = this.dockerContext.getBuilderName();
|
const builderName = this.dockerContext.getBuilderName() + (this.session?.config.builderSuffix || '');
|
||||||
const platforms = this.config.platforms?.join(', ') || 'default';
|
const platforms = this.config.platforms?.join(', ') || 'default';
|
||||||
logger.log('info', `Setting up Docker buildx [${platforms}]...`);
|
logger.log('info', `Setting up Docker buildx [${platforms}]...`);
|
||||||
logger.log('info', `Builder: ${builderName}`);
|
logger.log('info', `Builder: ${builderName}`);
|
||||||
|
|
||||||
|
// Check for remote builders matching our target platforms
|
||||||
|
const requestedPlatforms = this.config.platforms || ['linux/amd64'];
|
||||||
|
const remoteBuilders = GlobalConfig.getBuildersForPlatforms(requestedPlatforms);
|
||||||
|
|
||||||
|
if (remoteBuilders.length > 0) {
|
||||||
|
await this.ensureBuildxWithRemoteNodes(builderName, requestedPlatforms, remoteBuilders);
|
||||||
|
} else {
|
||||||
|
await this.ensureBuildxLocal(builderName);
|
||||||
|
}
|
||||||
|
|
||||||
|
logger.log('ok', `Docker buildx ready (builder: ${builderName}, platforms: ${platforms})`);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Creates a multi-node buildx builder with local + remote SSH nodes.
|
||||||
|
*/
|
||||||
|
private async ensureBuildxWithRemoteNodes(
|
||||||
|
builderName: string,
|
||||||
|
requestedPlatforms: string[],
|
||||||
|
remoteBuilders: IRemoteBuilder[],
|
||||||
|
): Promise<void> {
|
||||||
|
const remotePlatforms = new Set(remoteBuilders.map((b) => b.platform));
|
||||||
|
const localPlatforms = requestedPlatforms.filter((p) => !remotePlatforms.has(p));
|
||||||
|
|
||||||
|
logger.log('info', `Remote builders: ${remoteBuilders.map((b) => `${b.name} (${b.platform} @ ${b.host})`).join(', ')}`);
|
||||||
|
if (localPlatforms.length > 0) {
|
||||||
|
logger.log('info', `Local platforms: ${localPlatforms.join(', ')}`);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Always recreate the builder to ensure correct node topology
|
||||||
|
await smartshellInstance.execSilent(`docker buildx rm ${builderName} 2>/dev/null || true`);
|
||||||
|
|
||||||
|
// Create the local node
|
||||||
|
const localPlatformFlag = localPlatforms.length > 0 ? ` --platform ${localPlatforms.join(',')}` : '';
|
||||||
|
await smartshellInstance.exec(
|
||||||
|
`docker buildx create --name ${builderName} --driver docker-container --driver-opt network=host${localPlatformFlag} --use`
|
||||||
|
);
|
||||||
|
|
||||||
|
// Append remote nodes
|
||||||
|
for (const builder of remoteBuilders) {
|
||||||
|
logger.log('info', `Appending remote node: ${builder.name} (${builder.platform}) via ssh://${builder.host}`);
|
||||||
|
const appendResult = await smartshellInstance.exec(
|
||||||
|
`docker buildx create --append --name ${builderName} --driver docker-container --driver-opt network=host --platform ${builder.platform} --node ${builder.name} ssh://${builder.host}`
|
||||||
|
);
|
||||||
|
if (appendResult.exitCode !== 0) {
|
||||||
|
throw new Error(`Failed to append remote builder ${builder.name}: ${appendResult.stderr}`);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Bootstrap all nodes
|
||||||
|
await smartshellInstance.exec('docker buildx inspect --bootstrap');
|
||||||
|
|
||||||
|
// Store active remote builders for SSH tunnel setup during build
|
||||||
|
this.activeRemoteBuilders = remoteBuilders;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Creates a single-node local buildx builder (original behavior, uses QEMU for cross-platform).
|
||||||
|
*/
|
||||||
|
private async ensureBuildxLocal(builderName: string): Promise<void> {
|
||||||
const inspectResult = await smartshellInstance.exec(`docker buildx inspect ${builderName} 2>/dev/null`);
|
const inspectResult = await smartshellInstance.exec(`docker buildx inspect ${builderName} 2>/dev/null`);
|
||||||
|
|
||||||
if (inspectResult.exitCode !== 0) {
|
if (inspectResult.exitCode !== 0) {
|
||||||
@@ -354,7 +472,30 @@ export class TsDockerManager {
|
|||||||
await smartshellInstance.exec(`docker buildx use ${builderName}`);
|
await smartshellInstance.exec(`docker buildx use ${builderName}`);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
logger.log('ok', `Docker buildx ready (builder: ${builderName}, platforms: ${platforms})`);
|
this.activeRemoteBuilders = [];
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Opens SSH reverse tunnels for remote builders so they can reach the local registry.
|
||||||
|
*/
|
||||||
|
private async openRemoteTunnels(): Promise<void> {
|
||||||
|
if (this.activeRemoteBuilders.length === 0) return;
|
||||||
|
|
||||||
|
this.sshTunnelManager = new SshTunnelManager();
|
||||||
|
await this.sshTunnelManager.openTunnels(
|
||||||
|
this.activeRemoteBuilders,
|
||||||
|
this.session.config.registryPort,
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Closes any active SSH tunnels.
|
||||||
|
*/
|
||||||
|
private async closeRemoteTunnels(): Promise<void> {
|
||||||
|
if (this.sshTunnelManager) {
|
||||||
|
await this.sshTunnelManager.closeAll();
|
||||||
|
this.sshTunnelManager = undefined;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -394,7 +535,7 @@ export class TsDockerManager {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Start local registry (reads from persistent .nogit/docker-registry/)
|
// Start local registry (reads from persistent .nogit/docker-registry/)
|
||||||
await Dockerfile.startLocalRegistry(this.dockerContext.contextInfo?.isRootless);
|
await Dockerfile.startLocalRegistry(this.session, this.dockerContext.contextInfo?.isRootless);
|
||||||
try {
|
try {
|
||||||
// Push each Dockerfile to each registry via OCI copy
|
// Push each Dockerfile to each registry via OCI copy
|
||||||
for (const dockerfile of this.dockerfiles) {
|
for (const dockerfile of this.dockerfiles) {
|
||||||
@@ -403,7 +544,7 @@ export class TsDockerManager {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
} finally {
|
} finally {
|
||||||
await Dockerfile.stopLocalRegistry();
|
await Dockerfile.stopLocalRegistry(this.session);
|
||||||
}
|
}
|
||||||
|
|
||||||
logger.log('success', 'All images pushed successfully');
|
logger.log('success', 'All images pushed successfully');
|
||||||
@@ -446,11 +587,11 @@ export class TsDockerManager {
|
|||||||
logger.log('info', '');
|
logger.log('info', '');
|
||||||
logger.log('info', '=== TEST PHASE ===');
|
logger.log('info', '=== TEST PHASE ===');
|
||||||
|
|
||||||
await Dockerfile.startLocalRegistry(this.dockerContext.contextInfo?.isRootless);
|
await Dockerfile.startLocalRegistry(this.session, this.dockerContext.contextInfo?.isRootless);
|
||||||
try {
|
try {
|
||||||
await Dockerfile.testDockerfiles(this.dockerfiles);
|
await Dockerfile.testDockerfiles(this.dockerfiles);
|
||||||
} finally {
|
} finally {
|
||||||
await Dockerfile.stopLocalRegistry();
|
await Dockerfile.stopLocalRegistry(this.session);
|
||||||
}
|
}
|
||||||
|
|
||||||
logger.log('success', 'All tests completed');
|
logger.log('success', 'All tests completed');
|
||||||
@@ -490,4 +631,16 @@ export class TsDockerManager {
|
|||||||
public getDockerfiles(): Dockerfile[] {
|
public getDockerfiles(): Dockerfile[] {
|
||||||
return this.dockerfiles;
|
return this.dockerfiles;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Cleans up session-specific resources.
|
||||||
|
* In CI, removes the session-specific buildx builder to avoid accumulation.
|
||||||
|
*/
|
||||||
|
public async cleanup(): Promise<void> {
|
||||||
|
if (this.session?.config.isCI && this.session.config.builderSuffix) {
|
||||||
|
const builderName = this.dockerContext.getBuilderName() + this.session.config.builderSuffix;
|
||||||
|
logger.log('info', `CI cleanup: removing buildx builder ${builderName}`);
|
||||||
|
await smartshellInstance.execSilent(`docker buildx rm ${builderName} 2>/dev/null || true`);
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
107
ts/classes.tsdockersession.ts
Normal file
107
ts/classes.tsdockersession.ts
Normal file
@@ -0,0 +1,107 @@
|
|||||||
|
import * as crypto from 'crypto';
|
||||||
|
import * as net from 'net';
|
||||||
|
import { logger } from './tsdocker.logging.js';
|
||||||
|
|
||||||
|
export interface ISessionConfig {
|
||||||
|
sessionId: string;
|
||||||
|
registryPort: number;
|
||||||
|
registryHost: string;
|
||||||
|
registryContainerName: string;
|
||||||
|
isCI: boolean;
|
||||||
|
ciSystem: string | null;
|
||||||
|
builderSuffix: string;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Per-invocation session identity for tsdocker.
|
||||||
|
* Generates unique ports, container names, and builder names so that
|
||||||
|
* concurrent CI jobs on the same Docker host don't collide.
|
||||||
|
*
|
||||||
|
* In local (non-CI) dev the builder suffix is empty, preserving the
|
||||||
|
* persistent builder behavior.
|
||||||
|
*/
|
||||||
|
export class TsDockerSession {
|
||||||
|
public config: ISessionConfig;
|
||||||
|
|
||||||
|
private constructor(config: ISessionConfig) {
|
||||||
|
this.config = config;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Creates a new session. Allocates a dynamic port unless overridden
|
||||||
|
* via `TSDOCKER_REGISTRY_PORT`.
|
||||||
|
*/
|
||||||
|
public static async create(): Promise<TsDockerSession> {
|
||||||
|
const sessionId =
|
||||||
|
process.env.TSDOCKER_SESSION_ID || crypto.randomBytes(4).toString('hex');
|
||||||
|
|
||||||
|
const registryPort = await TsDockerSession.allocatePort();
|
||||||
|
const registryHost = `localhost:${registryPort}`;
|
||||||
|
const registryContainerName = `tsdocker-registry-${sessionId}`;
|
||||||
|
|
||||||
|
const { isCI, ciSystem } = TsDockerSession.detectCI();
|
||||||
|
const builderSuffix = isCI ? `-${sessionId}` : '';
|
||||||
|
|
||||||
|
const config: ISessionConfig = {
|
||||||
|
sessionId,
|
||||||
|
registryPort,
|
||||||
|
registryHost,
|
||||||
|
registryContainerName,
|
||||||
|
isCI,
|
||||||
|
ciSystem,
|
||||||
|
builderSuffix,
|
||||||
|
};
|
||||||
|
|
||||||
|
const session = new TsDockerSession(config);
|
||||||
|
session.logInfo();
|
||||||
|
return session;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Allocates a free TCP port. Respects `TSDOCKER_REGISTRY_PORT` override.
|
||||||
|
*/
|
||||||
|
public static async allocatePort(): Promise<number> {
|
||||||
|
const envPort = process.env.TSDOCKER_REGISTRY_PORT;
|
||||||
|
if (envPort) {
|
||||||
|
const parsed = parseInt(envPort, 10);
|
||||||
|
if (!isNaN(parsed) && parsed > 0) {
|
||||||
|
return parsed;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return new Promise<number>((resolve, reject) => {
|
||||||
|
const srv = net.createServer();
|
||||||
|
srv.listen(0, '127.0.0.1', () => {
|
||||||
|
const addr = srv.address() as net.AddressInfo;
|
||||||
|
const port = addr.port;
|
||||||
|
srv.close((err) => {
|
||||||
|
if (err) reject(err);
|
||||||
|
else resolve(port);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
srv.on('error', reject);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Detects whether we're running inside a CI system.
|
||||||
|
*/
|
||||||
|
private static detectCI(): { isCI: boolean; ciSystem: string | null } {
|
||||||
|
if (process.env.GITEA_ACTIONS) return { isCI: true, ciSystem: 'gitea-actions' };
|
||||||
|
if (process.env.GITHUB_ACTIONS) return { isCI: true, ciSystem: 'github-actions' };
|
||||||
|
if (process.env.GITLAB_CI) return { isCI: true, ciSystem: 'gitlab-ci' };
|
||||||
|
if (process.env.CI) return { isCI: true, ciSystem: 'generic' };
|
||||||
|
return { isCI: false, ciSystem: null };
|
||||||
|
}
|
||||||
|
|
||||||
|
private logInfo(): void {
|
||||||
|
const c = this.config;
|
||||||
|
logger.log('info', '=== TSDOCKER SESSION ===');
|
||||||
|
logger.log('info', `Session ID: ${c.sessionId}`);
|
||||||
|
logger.log('info', `Registry: ${c.registryHost} (container: ${c.registryContainerName})`);
|
||||||
|
if (c.isCI) {
|
||||||
|
logger.log('info', `CI detected: ${c.ciSystem}`);
|
||||||
|
logger.log('info', `Builder suffix: ${c.builderSuffix}`);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -1,15 +1,7 @@
|
|||||||
/**
|
/**
|
||||||
* Configuration interface for tsdocker
|
* Configuration interface for tsdocker
|
||||||
* Extends legacy config with new Docker build capabilities
|
|
||||||
*/
|
*/
|
||||||
export interface ITsDockerConfig {
|
export interface ITsDockerConfig {
|
||||||
// Legacy (backward compatible)
|
|
||||||
baseImage: string;
|
|
||||||
command: string;
|
|
||||||
dockerSock: boolean;
|
|
||||||
keyValueObject: { [key: string]: any };
|
|
||||||
|
|
||||||
// New Docker build config
|
|
||||||
registries?: string[];
|
registries?: string[];
|
||||||
registryRepoMap?: { [registry: string]: string };
|
registryRepoMap?: { [registry: string]: string };
|
||||||
buildArgEnvMap?: { [dockerArg: string]: string };
|
buildArgEnvMap?: { [dockerArg: string]: string };
|
||||||
@@ -101,4 +93,22 @@ export interface IDockerContextInfo {
|
|||||||
endpoint: string; // 'unix:///var/run/docker.sock'
|
endpoint: string; // 'unix:///var/run/docker.sock'
|
||||||
isRootless: boolean;
|
isRootless: boolean;
|
||||||
dockerHost?: string; // value of DOCKER_HOST env var, if set
|
dockerHost?: string; // value of DOCKER_HOST env var, if set
|
||||||
|
topology?: 'socket-mount' | 'dind' | 'local';
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* A remote builder node for native cross-platform builds
|
||||||
|
*/
|
||||||
|
export interface IRemoteBuilder {
|
||||||
|
name: string; // e.g., "arm64-builder"
|
||||||
|
host: string; // e.g., "armbuilder@192.168.190.216"
|
||||||
|
platform: string; // e.g., "linux/arm64"
|
||||||
|
sshKeyPath?: string; // e.g., "~/.ssh/id_ed25519"
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Global tsdocker configuration stored at ~/.git.zone/tsdocker/config.json
|
||||||
|
*/
|
||||||
|
export interface IGlobalConfig {
|
||||||
|
remoteBuilders: IRemoteBuilder[];
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -3,27 +3,97 @@ import * as paths from './tsdocker.paths.js';
|
|||||||
|
|
||||||
// modules
|
// modules
|
||||||
import * as ConfigModule from './tsdocker.config.js';
|
import * as ConfigModule from './tsdocker.config.js';
|
||||||
import * as DockerModule from './tsdocker.docker.js';
|
|
||||||
|
|
||||||
import { logger, ora } from './tsdocker.logging.js';
|
import { logger, ora } from './tsdocker.logging.js';
|
||||||
import { TsDockerManager } from './classes.tsdockermanager.js';
|
import { TsDockerManager } from './classes.tsdockermanager.js';
|
||||||
import { DockerContext } from './classes.dockercontext.js';
|
import { DockerContext } from './classes.dockercontext.js';
|
||||||
|
import { GlobalConfig } from './classes.globalconfig.js';
|
||||||
import type { IBuildCommandOptions } from './interfaces/index.js';
|
import type { IBuildCommandOptions } from './interfaces/index.js';
|
||||||
import { commitinfo } from './00_commitinfo_data.js';
|
import { commitinfo } from './00_commitinfo_data.js';
|
||||||
|
|
||||||
const tsdockerCli = new plugins.smartcli.Smartcli();
|
const tsdockerCli = new plugins.smartcli.Smartcli();
|
||||||
tsdockerCli.addVersion(commitinfo.version);
|
tsdockerCli.addVersion(commitinfo.version);
|
||||||
|
|
||||||
|
const printManPage = () => {
|
||||||
|
const manPage = `
|
||||||
|
TSDOCKER(1) User Commands TSDOCKER(1)
|
||||||
|
|
||||||
|
NAME
|
||||||
|
tsdocker - build, test, and push Docker images
|
||||||
|
|
||||||
|
VERSION
|
||||||
|
${commitinfo.version}
|
||||||
|
|
||||||
|
SYNOPSIS
|
||||||
|
tsdocker <command> [options]
|
||||||
|
|
||||||
|
COMMANDS
|
||||||
|
build [patterns...] [flags] Build Dockerfiles in dependency order
|
||||||
|
push [patterns...] [flags] Build and push images to registries
|
||||||
|
pull <registry-url> Pull images from a registry
|
||||||
|
test [flags] Build and run container test scripts
|
||||||
|
login Authenticate with configured registries
|
||||||
|
list List discovered Dockerfiles
|
||||||
|
config <subcommand> [flags] Manage global tsdocker configuration
|
||||||
|
clean [-y] [--all] Interactive Docker resource cleanup
|
||||||
|
|
||||||
|
BUILD / PUSH OPTIONS
|
||||||
|
--platform=<p> Target platform (e.g. linux/arm64)
|
||||||
|
--timeout=<s> Build timeout in seconds
|
||||||
|
--no-cache Rebuild without Docker layer cache
|
||||||
|
--cached Skip builds when Dockerfile is unchanged
|
||||||
|
--verbose Stream raw docker build output
|
||||||
|
--parallel[=<n>] Parallel builds (optional concurrency limit)
|
||||||
|
--context=<name> Docker context to use
|
||||||
|
|
||||||
|
PUSH-ONLY OPTIONS
|
||||||
|
--registry=<url> Push to a specific registry
|
||||||
|
--no-build Push already-built images (skip build step)
|
||||||
|
|
||||||
|
CLEAN OPTIONS
|
||||||
|
-y Auto-confirm all prompts
|
||||||
|
--all Include all images and volumes (not just dangling)
|
||||||
|
|
||||||
|
CONFIG SUBCOMMANDS
|
||||||
|
add-builder Add a remote builder node
|
||||||
|
--name=<n> Builder name (e.g. arm64-builder)
|
||||||
|
--host=<h> SSH host (e.g. user@192.168.1.100)
|
||||||
|
--platform=<p> Platform (e.g. linux/arm64)
|
||||||
|
--ssh-key=<path> SSH key path (optional)
|
||||||
|
remove-builder Remove a remote builder by name
|
||||||
|
--name=<n> Builder name to remove
|
||||||
|
list-builders List all configured remote builders
|
||||||
|
show Show full global config
|
||||||
|
|
||||||
|
CONFIGURATION
|
||||||
|
Configure via npmextra.json under the "@git.zone/tsdocker" key:
|
||||||
|
|
||||||
|
registries Array of registry URLs to push to
|
||||||
|
registryRepoMap Map of registry URL to repo path overrides
|
||||||
|
buildArgEnvMap Map of Docker build-arg names to env var names
|
||||||
|
platforms Array of target platforms (default: ["linux/amd64"])
|
||||||
|
push Boolean, auto-push after build
|
||||||
|
testDir Directory containing test_*.sh scripts
|
||||||
|
|
||||||
|
Global config is stored at ~/.git.zone/tsdocker/config.json
|
||||||
|
and managed via the "config" command.
|
||||||
|
|
||||||
|
EXAMPLES
|
||||||
|
tsdocker build
|
||||||
|
tsdocker build Dockerfile_app --platform=linux/arm64
|
||||||
|
tsdocker push --registry=ghcr.io
|
||||||
|
tsdocker test --verbose
|
||||||
|
tsdocker clean -y --all
|
||||||
|
tsdocker config add-builder --name=arm64 --host=user@host --platform=linux/arm64
|
||||||
|
tsdocker config list-builders
|
||||||
|
`;
|
||||||
|
console.log(manPage);
|
||||||
|
};
|
||||||
|
|
||||||
export let run = () => {
|
export let run = () => {
|
||||||
// Default command: run tests in container (legacy behavior)
|
// Default command: print man page
|
||||||
tsdockerCli.standardCommand().subscribe(async argvArg => {
|
tsdockerCli.standardCommand().subscribe(async () => {
|
||||||
const configArg = await ConfigModule.run().then(DockerModule.run);
|
printManPage();
|
||||||
if (configArg.exitCode === 0) {
|
|
||||||
logger.log('success', 'container ended all right!');
|
|
||||||
} else {
|
|
||||||
logger.log('error', `container ended with error! Exit Code is ${configArg.exitCode}`);
|
|
||||||
process.exit(1);
|
|
||||||
}
|
|
||||||
});
|
});
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -64,6 +134,7 @@ export let run = () => {
|
|||||||
}
|
}
|
||||||
|
|
||||||
await manager.build(buildOptions);
|
await manager.build(buildOptions);
|
||||||
|
await manager.cleanup();
|
||||||
logger.log('success', 'Build completed successfully');
|
logger.log('success', 'Build completed successfully');
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
logger.log('error', `Build failed: ${(err as Error).message}`);
|
logger.log('error', `Build failed: ${(err as Error).message}`);
|
||||||
@@ -109,14 +180,22 @@ export let run = () => {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Build images first (if not already built)
|
// Build images first, unless --no-build is set
|
||||||
await manager.build(buildOptions);
|
if (argvArg.build === false) {
|
||||||
|
await manager.discoverDockerfiles();
|
||||||
|
if (buildOptions.patterns?.length) {
|
||||||
|
manager.filterDockerfiles(buildOptions.patterns);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
await manager.build(buildOptions);
|
||||||
|
}
|
||||||
|
|
||||||
// Get registry from --registry flag
|
// Get registry from --registry flag
|
||||||
const registryArg = argvArg.registry as string | undefined;
|
const registryArg = argvArg.registry as string | undefined;
|
||||||
const registries = registryArg ? [registryArg] : undefined;
|
const registries = registryArg ? [registryArg] : undefined;
|
||||||
|
|
||||||
await manager.push(registries);
|
await manager.push(registries);
|
||||||
|
await manager.cleanup();
|
||||||
logger.log('success', 'Push completed successfully');
|
logger.log('success', 'Push completed successfully');
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
logger.log('error', `Push failed: ${(err as Error).message}`);
|
logger.log('error', `Push failed: ${(err as Error).message}`);
|
||||||
@@ -180,6 +259,7 @@ export let run = () => {
|
|||||||
|
|
||||||
// Run tests
|
// Run tests
|
||||||
await manager.test();
|
await manager.test();
|
||||||
|
await manager.cleanup();
|
||||||
logger.log('success', 'Tests completed successfully');
|
logger.log('success', 'Tests completed successfully');
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
logger.log('error', `Tests failed: ${(err as Error).message}`);
|
logger.log('error', `Tests failed: ${(err as Error).message}`);
|
||||||
@@ -219,21 +299,73 @@ export let run = () => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* this command is executed inside docker and meant for use from outside docker
|
* Manage global tsdocker configuration (remote builders, etc.)
|
||||||
|
* Usage: tsdocker config <subcommand> [--name=...] [--host=...] [--platform=...] [--ssh-key=...]
|
||||||
*/
|
*/
|
||||||
tsdockerCli.addCommand('runinside').subscribe(async argvArg => {
|
tsdockerCli.addCommand('config').subscribe(async argvArg => {
|
||||||
logger.log('ok', 'Allright. We are now in Docker!');
|
try {
|
||||||
ora.text('now trying to run your specified command');
|
const subcommand = argvArg._[1] as string;
|
||||||
const configArg = await ConfigModule.run();
|
|
||||||
const smartshellInstance = new plugins.smartshell.Smartshell({
|
switch (subcommand) {
|
||||||
executor: 'bash'
|
case 'add-builder': {
|
||||||
});
|
const name = argvArg.name as string;
|
||||||
ora.stop();
|
const host = argvArg.host as string;
|
||||||
await smartshellInstance.exec(configArg.command).then(response => {
|
const platform = argvArg.platform as string;
|
||||||
if (response.exitCode !== 0) {
|
const sshKeyPath = argvArg['ssh-key'] as string | undefined;
|
||||||
process.exit(1);
|
|
||||||
|
if (!name || !host || !platform) {
|
||||||
|
logger.log('error', 'Required: --name, --host, --platform');
|
||||||
|
logger.log('info', 'Usage: tsdocker config add-builder --name=arm64-builder --host=user@host --platform=linux/arm64 [--ssh-key=~/.ssh/id_ed25519]');
|
||||||
|
process.exit(1);
|
||||||
|
}
|
||||||
|
|
||||||
|
GlobalConfig.addBuilder({ name, host, platform, sshKeyPath });
|
||||||
|
logger.log('success', `Remote builder "${name}" configured: ${platform} via ssh://${host}`);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
case 'remove-builder': {
|
||||||
|
const name = argvArg.name as string;
|
||||||
|
if (!name) {
|
||||||
|
logger.log('error', 'Required: --name');
|
||||||
|
logger.log('info', 'Usage: tsdocker config remove-builder --name=arm64-builder');
|
||||||
|
process.exit(1);
|
||||||
|
}
|
||||||
|
GlobalConfig.removeBuilder(name);
|
||||||
|
logger.log('success', `Remote builder "${name}" removed`);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
case 'list-builders': {
|
||||||
|
const builders = GlobalConfig.getBuilders();
|
||||||
|
if (builders.length === 0) {
|
||||||
|
logger.log('info', 'No remote builders configured');
|
||||||
|
} else {
|
||||||
|
logger.log('info', `${builders.length} remote builder(s):`);
|
||||||
|
for (const b of builders) {
|
||||||
|
const keyInfo = b.sshKeyPath ? ` (key: ${b.sshKeyPath})` : '';
|
||||||
|
logger.log('info', ` ${b.name}: ${b.platform} via ssh://${b.host}${keyInfo}`);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
case 'show': {
|
||||||
|
const config = GlobalConfig.load();
|
||||||
|
logger.log('info', `Config file: ${GlobalConfig.getConfigPath()}`);
|
||||||
|
console.log(JSON.stringify(config, null, 2));
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
default:
|
||||||
|
logger.log('error', `Unknown config subcommand: ${subcommand || '(none)'}`);
|
||||||
|
logger.log('info', 'Available: add-builder, remove-builder, list-builders, show');
|
||||||
|
process.exit(1);
|
||||||
}
|
}
|
||||||
});
|
} catch (err) {
|
||||||
|
logger.log('error', `Config failed: ${(err as Error).message}`);
|
||||||
|
process.exit(1);
|
||||||
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
tsdockerCli.addCommand('clean').subscribe(async argvArg => {
|
tsdockerCli.addCommand('clean').subscribe(async argvArg => {
|
||||||
@@ -433,19 +565,5 @@ export let run = () => {
|
|||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
tsdockerCli.addCommand('vscode').subscribe(async argvArg => {
|
|
||||||
const smartshellInstance = new plugins.smartshell.Smartshell({
|
|
||||||
executor: 'bash'
|
|
||||||
});
|
|
||||||
logger.log('ok', `Starting vscode in cwd ${paths.cwd}`);
|
|
||||||
await smartshellInstance.execAndWaitForLine(
|
|
||||||
`docker run -p 127.0.0.1:8443:8443 -v "${
|
|
||||||
paths.cwd
|
|
||||||
}:/home/coder/project" registry.gitlab.com/hosttoday/ht-docker-vscode --allow-http --no-auth`,
|
|
||||||
/Connected to shared process/
|
|
||||||
);
|
|
||||||
await plugins.smartopen.openUrl('testing-vscode.git.zone:8443');
|
|
||||||
});
|
|
||||||
|
|
||||||
tsdockerCli.startParse();
|
tsdockerCli.startParse();
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -1,34 +1,10 @@
|
|||||||
import * as plugins from './tsdocker.plugins.js';
|
import * as plugins from './tsdocker.plugins.js';
|
||||||
import * as paths from './tsdocker.paths.js';
|
import * as paths from './tsdocker.paths.js';
|
||||||
import * as fs from 'fs';
|
|
||||||
import type { ITsDockerConfig } from './interfaces/index.js';
|
import type { ITsDockerConfig } from './interfaces/index.js';
|
||||||
|
|
||||||
// Re-export ITsDockerConfig as IConfig for backward compatibility
|
const buildConfig = async (): Promise<ITsDockerConfig> => {
|
||||||
export type IConfig = ITsDockerConfig & {
|
|
||||||
exitCode?: number;
|
|
||||||
};
|
|
||||||
|
|
||||||
const getQenvKeyValueObject = async () => {
|
|
||||||
let qenvKeyValueObjectArray: { [key: string]: string | number };
|
|
||||||
if (fs.existsSync(plugins.path.join(paths.cwd, 'qenv.yml'))) {
|
|
||||||
qenvKeyValueObjectArray = new plugins.qenv.Qenv(paths.cwd, '.nogit/').keyValueObject;
|
|
||||||
} else {
|
|
||||||
qenvKeyValueObjectArray = {};
|
|
||||||
}
|
|
||||||
return qenvKeyValueObjectArray;
|
|
||||||
};
|
|
||||||
|
|
||||||
const buildConfig = async (qenvKeyValueObjectArg: { [key: string]: string | number }) => {
|
|
||||||
const npmextra = new plugins.npmextra.Npmextra(paths.cwd);
|
const npmextra = new plugins.npmextra.Npmextra(paths.cwd);
|
||||||
const config = npmextra.dataFor<IConfig>('@git.zone/tsdocker', {
|
const config = npmextra.dataFor<ITsDockerConfig>('@git.zone/tsdocker', {
|
||||||
// Legacy options (backward compatible)
|
|
||||||
baseImage: 'hosttoday/ht-docker-node:npmdocker',
|
|
||||||
init: 'rm -rf node_nodules/ && yarn install',
|
|
||||||
command: 'npmci npm test',
|
|
||||||
dockerSock: false,
|
|
||||||
keyValueObject: qenvKeyValueObjectArg,
|
|
||||||
|
|
||||||
// New Docker build options
|
|
||||||
registries: [],
|
registries: [],
|
||||||
registryRepoMap: {},
|
registryRepoMap: {},
|
||||||
buildArgEnvMap: {},
|
buildArgEnvMap: {},
|
||||||
@@ -39,7 +15,6 @@ const buildConfig = async (qenvKeyValueObjectArg: { [key: string]: string | numb
|
|||||||
return config;
|
return config;
|
||||||
};
|
};
|
||||||
|
|
||||||
export let run = async (): Promise<IConfig> => {
|
export let run = async (): Promise<ITsDockerConfig> => {
|
||||||
const config = await getQenvKeyValueObject().then(buildConfig);
|
return buildConfig();
|
||||||
return config;
|
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -1,169 +0,0 @@
|
|||||||
import * as plugins from './tsdocker.plugins.js';
|
|
||||||
import * as paths from './tsdocker.paths.js';
|
|
||||||
import * as snippets from './tsdocker.snippets.js';
|
|
||||||
|
|
||||||
import { logger, ora } from './tsdocker.logging.js';
|
|
||||||
|
|
||||||
const smartshellInstance = new plugins.smartshell.Smartshell({
|
|
||||||
executor: 'bash'
|
|
||||||
});
|
|
||||||
|
|
||||||
// interfaces
|
|
||||||
import type { IConfig } from './tsdocker.config.js';
|
|
||||||
|
|
||||||
let config: IConfig;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* the docker data used to build the internal testing container
|
|
||||||
*/
|
|
||||||
const dockerData = {
|
|
||||||
imageTag: 'npmdocker-temp-image:latest',
|
|
||||||
containerName: 'npmdocker-temp-container',
|
|
||||||
dockerProjectMountString: '',
|
|
||||||
dockerSockString: '',
|
|
||||||
dockerEnvString: ''
|
|
||||||
};
|
|
||||||
|
|
||||||
/**
|
|
||||||
* check if docker is available
|
|
||||||
*/
|
|
||||||
const checkDocker = () => {
|
|
||||||
const done = plugins.smartpromise.defer();
|
|
||||||
ora.text('checking docker...');
|
|
||||||
|
|
||||||
if (smartshellInstance.exec('which docker')) {
|
|
||||||
logger.log('ok', 'Docker found!');
|
|
||||||
done.resolve();
|
|
||||||
} else {
|
|
||||||
done.reject(new Error('docker not found on this machine'));
|
|
||||||
}
|
|
||||||
return done.promise;
|
|
||||||
};
|
|
||||||
|
|
||||||
/**
|
|
||||||
* builds the Dockerfile according to the config in the project
|
|
||||||
*/
|
|
||||||
const buildDockerFile = async () => {
|
|
||||||
const done = plugins.smartpromise.defer();
|
|
||||||
ora.text('building Dockerfile...');
|
|
||||||
const dockerfile: string = snippets.dockerfileSnippet({
|
|
||||||
baseImage: config.baseImage,
|
|
||||||
command: config.command
|
|
||||||
});
|
|
||||||
logger.log('info', `Base image is: ${config.baseImage}`);
|
|
||||||
logger.log('info', `Command is: ${config.command}`);
|
|
||||||
await plugins.smartfs.file(plugins.path.join(paths.cwd, 'npmdocker')).write(dockerfile);
|
|
||||||
logger.log('ok', 'Dockerfile created!');
|
|
||||||
ora.stop();
|
|
||||||
done.resolve();
|
|
||||||
return done.promise;
|
|
||||||
};
|
|
||||||
|
|
||||||
/**
|
|
||||||
* builds the Dockerimage from the built Dockerfile
|
|
||||||
*/
|
|
||||||
const buildDockerImage = async () => {
|
|
||||||
logger.log('info', 'pulling latest base image from registry...');
|
|
||||||
await smartshellInstance.exec(`docker pull ${config.baseImage}`);
|
|
||||||
ora.text('building Dockerimage...');
|
|
||||||
const execResult = await smartshellInstance.execSilent(
|
|
||||||
`docker build --load -f npmdocker -t ${dockerData.imageTag} ${paths.cwd}`
|
|
||||||
);
|
|
||||||
if (execResult.exitCode !== 0) {
|
|
||||||
console.log(execResult.stdout);
|
|
||||||
process.exit(1);
|
|
||||||
}
|
|
||||||
logger.log('ok', 'Dockerimage built!');
|
|
||||||
};
|
|
||||||
|
|
||||||
const buildDockerProjectMountString = async () => {
|
|
||||||
if (process.env.CI !== 'true') {
|
|
||||||
dockerData.dockerProjectMountString = `-v ${paths.cwd}:/workspace`;
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
/**
|
|
||||||
* builds an environment string that docker cli understands
|
|
||||||
*/
|
|
||||||
const buildDockerEnvString = async () => {
|
|
||||||
for (const key of Object.keys(config.keyValueObject)) {
|
|
||||||
const envString = (dockerData.dockerEnvString =
|
|
||||||
dockerData.dockerEnvString + `-e ${key}=${config.keyValueObject[key]} `);
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
/**
|
|
||||||
* creates string to mount the docker.sock inside the testcontainer
|
|
||||||
*/
|
|
||||||
const buildDockerSockString = async () => {
|
|
||||||
if (config.dockerSock) {
|
|
||||||
dockerData.dockerSockString = `-v /var/run/docker.sock:/var/run/docker.sock`;
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
/**
|
|
||||||
* creates a container by running the built Dockerimage
|
|
||||||
*/
|
|
||||||
const runDockerImage = async () => {
|
|
||||||
const done = plugins.smartpromise.defer();
|
|
||||||
ora.text('starting Container...');
|
|
||||||
ora.stop();
|
|
||||||
logger.log('info', 'now running Dockerimage');
|
|
||||||
config.exitCode = (await smartshellInstance.exec(
|
|
||||||
`docker run ${dockerData.dockerProjectMountString} ${dockerData.dockerSockString} ${
|
|
||||||
dockerData.dockerEnvString
|
|
||||||
} --name ${dockerData.containerName} ${dockerData.imageTag}`
|
|
||||||
)).exitCode;
|
|
||||||
};
|
|
||||||
|
|
||||||
/**
|
|
||||||
* cleans up: deletes the test container
|
|
||||||
*/
|
|
||||||
const deleteDockerContainer = async () => {
|
|
||||||
await smartshellInstance.execSilent(`docker rm -f ${dockerData.containerName}`);
|
|
||||||
};
|
|
||||||
|
|
||||||
/**
|
|
||||||
* cleans up deletes the test image
|
|
||||||
*/
|
|
||||||
const deleteDockerImage = async () => {
|
|
||||||
await smartshellInstance.execSilent(`docker rmi ${dockerData.imageTag}`).then(async response => {
|
|
||||||
if (response.exitCode !== 0) {
|
|
||||||
console.log(response.stdout);
|
|
||||||
}
|
|
||||||
});
|
|
||||||
};
|
|
||||||
|
|
||||||
const preClean = async () => {
|
|
||||||
await deleteDockerImage()
|
|
||||||
.then(deleteDockerContainer)
|
|
||||||
.then(async () => {
|
|
||||||
logger.log('ok', 'ensured clean Docker environment!');
|
|
||||||
});
|
|
||||||
};
|
|
||||||
|
|
||||||
const postClean = async () => {
|
|
||||||
await deleteDockerContainer()
|
|
||||||
.then(deleteDockerImage)
|
|
||||||
.then(async () => {
|
|
||||||
logger.log('ok', 'cleaned up!');
|
|
||||||
});
|
|
||||||
await plugins.smartfs.file(paths.npmdockerFile).delete();
|
|
||||||
};
|
|
||||||
|
|
||||||
export let run = async (configArg: IConfig): Promise<IConfig> => {
|
|
||||||
config = configArg;
|
|
||||||
const resultConfig = await checkDocker()
|
|
||||||
.then(preClean)
|
|
||||||
.then(buildDockerFile)
|
|
||||||
.then(buildDockerImage)
|
|
||||||
.then(buildDockerProjectMountString)
|
|
||||||
.then(buildDockerEnvString)
|
|
||||||
.then(buildDockerSockString)
|
|
||||||
.then(runDockerImage)
|
|
||||||
.then(postClean)
|
|
||||||
.catch(err => {
|
|
||||||
console.log(err);
|
|
||||||
});
|
|
||||||
return config;
|
|
||||||
};
|
|
||||||
@@ -11,4 +11,3 @@ export let cwd = process.cwd();
|
|||||||
export let packageBase = plugins.path.join(__dirname, '../');
|
export let packageBase = plugins.path.join(__dirname, '../');
|
||||||
export let assets = plugins.path.join(packageBase, 'assets/');
|
export let assets = plugins.path.join(packageBase, 'assets/');
|
||||||
fs.mkdirSync(assets, { recursive: true });
|
fs.mkdirSync(assets, { recursive: true });
|
||||||
export let npmdockerFile = plugins.path.join(cwd, 'npmdocker');
|
|
||||||
|
|||||||
@@ -3,17 +3,13 @@ import * as lik from '@push.rocks/lik';
|
|||||||
import * as npmextra from '@push.rocks/npmextra';
|
import * as npmextra from '@push.rocks/npmextra';
|
||||||
import * as path from 'path';
|
import * as path from 'path';
|
||||||
import * as projectinfo from '@push.rocks/projectinfo';
|
import * as projectinfo from '@push.rocks/projectinfo';
|
||||||
import * as smartpromise from '@push.rocks/smartpromise';
|
|
||||||
import * as qenv from '@push.rocks/qenv';
|
|
||||||
import * as smartcli from '@push.rocks/smartcli';
|
import * as smartcli from '@push.rocks/smartcli';
|
||||||
import { SmartFs, SmartFsProviderNode } from '@push.rocks/smartfs';
|
import { SmartFs, SmartFsProviderNode } from '@push.rocks/smartfs';
|
||||||
import * as smartlog from '@push.rocks/smartlog';
|
import * as smartlog from '@push.rocks/smartlog';
|
||||||
import * as smartlogDestinationLocal from '@push.rocks/smartlog-destination-local';
|
import * as smartlogDestinationLocal from '@push.rocks/smartlog-destination-local';
|
||||||
import * as smartlogSouceOra from '@push.rocks/smartlog-source-ora';
|
import * as smartlogSouceOra from '@push.rocks/smartlog-source-ora';
|
||||||
import * as smartopen from '@push.rocks/smartopen';
|
|
||||||
import * as smartinteract from '@push.rocks/smartinteract';
|
import * as smartinteract from '@push.rocks/smartinteract';
|
||||||
import * as smartshell from '@push.rocks/smartshell';
|
import * as smartshell from '@push.rocks/smartshell';
|
||||||
import * as smartstring from '@push.rocks/smartstring';
|
|
||||||
|
|
||||||
// Create smartfs instance
|
// Create smartfs instance
|
||||||
export const smartfs = new SmartFs(new SmartFsProviderNode());
|
export const smartfs = new SmartFs(new SmartFsProviderNode());
|
||||||
@@ -23,14 +19,10 @@ export {
|
|||||||
npmextra,
|
npmextra,
|
||||||
path,
|
path,
|
||||||
projectinfo,
|
projectinfo,
|
||||||
smartpromise,
|
|
||||||
qenv,
|
|
||||||
smartcli,
|
smartcli,
|
||||||
smartinteract,
|
smartinteract,
|
||||||
smartlog,
|
smartlog,
|
||||||
smartlogDestinationLocal,
|
smartlogDestinationLocal,
|
||||||
smartlogSouceOra,
|
smartlogSouceOra,
|
||||||
smartopen,
|
|
||||||
smartshell,
|
smartshell,
|
||||||
smartstring
|
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -1,34 +0,0 @@
|
|||||||
import * as plugins from './tsdocker.plugins.js';
|
|
||||||
|
|
||||||
export interface IDockerfileSnippet {
|
|
||||||
baseImage: string;
|
|
||||||
command: string;
|
|
||||||
}
|
|
||||||
|
|
||||||
let getMountSolutionString = (optionsArg: IDockerfileSnippet) => {
|
|
||||||
if (process.env.CI) {
|
|
||||||
return 'COPY ./ /workspace';
|
|
||||||
} else {
|
|
||||||
return '# not copying workspcae since not in CI';
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
let getGlobalPreparationString = (optionsArg: IDockerfileSnippet) => {
|
|
||||||
// Always install tsdocker to ensure the latest version is available
|
|
||||||
return 'RUN npm install -g @git.zone/tsdocker';
|
|
||||||
};
|
|
||||||
|
|
||||||
export let dockerfileSnippet = (optionsArg: IDockerfileSnippet): string => {
|
|
||||||
return plugins.smartstring.indent.normalize(
|
|
||||||
`
|
|
||||||
FROM ${optionsArg.baseImage}
|
|
||||||
# For info about what tsdocker does read the docs at https://gitzone.github.io/tsdocker
|
|
||||||
${getGlobalPreparationString(optionsArg)}
|
|
||||||
${getMountSolutionString(optionsArg)}
|
|
||||||
WORKDIR /workspace
|
|
||||||
ENV CI=true
|
|
||||||
ENTRYPOINT ["tsdocker"]
|
|
||||||
CMD ["runinside"]
|
|
||||||
`
|
|
||||||
);
|
|
||||||
};
|
|
||||||
Reference in New Issue
Block a user