Compare commits
14 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| d1785aab86 | |||
| 31fb4aea3c | |||
| 907048fa87 | |||
| 02b267ee10 | |||
| 16cd0bbd87 | |||
| cc83743f9a | |||
| 7131c16f80 | |||
| 02688861f4 | |||
| 3a8b301b3e | |||
| c09bef33c3 | |||
| 32eb0d1d77 | |||
| 7cac628975 | |||
| c279dbd55e | |||
| 7b7064864e |
57
changelog.md
57
changelog.md
@@ -1,5 +1,62 @@
|
||||
# Changelog
|
||||
|
||||
## 2026-02-06 - 1.10.0 - feat(classes.dockerfile)
|
||||
support using a local base image as a build context in buildx commands
|
||||
|
||||
- Adds --build-context flag mapping base image to docker-image://<localTag> when localBaseImageDependent && localBaseDockerfile are set
|
||||
- Appends the build context flag to both single-platform and multi-platform docker buildx commands
|
||||
- Logs an info message indicating the local build context mapping
|
||||
|
||||
## 2026-02-06 - 1.9.0 - feat(build)
|
||||
add verbose build output, progress logging, and timing for builds/tests
|
||||
|
||||
- Add 'verbose' option to build/test flows (interfaces, CLI, and method signatures) to allow streaming raw docker build output or run silently
|
||||
- Log per-item progress for build and test phases (e.g. (1/N) Building/Testing <tag>) and report individual durations
|
||||
- Return elapsed time from Dockerfile.build() and Dockerfile.test() and aggregate total build/test times in manager
|
||||
- Introduce formatDuration(ms) helper in logging module to format timings
|
||||
- Switch from console.log to structured logger calls across cache, manager, dockerfile and push paths
|
||||
- Use silent exec variants when verbose is false and stream exec when verbose is true
|
||||
|
||||
## 2026-02-06 - 1.8.0 - feat(build)
|
||||
add optional content-hash based build cache to skip rebuilding unchanged Dockerfiles
|
||||
|
||||
- Introduce TsDockerCache to compute SHA-256 of Dockerfile content and persist cache to .nogit/tsdocker_support.json
|
||||
- Add ICacheEntry and ICacheData interfaces and a cached flag to IBuildCommandOptions
|
||||
- Integrate cached mode in TsDockerManager: skip builds on cache hits, verify image presence, record builds on misses, and still perform dependency tagging
|
||||
- Expose --cached option in CLI to enable the cached build flow
|
||||
- Cache records store contentHash, imageId, buildTag and timestamp
|
||||
|
||||
## 2026-02-06 - 1.7.0 - feat(cli)
|
||||
add CLI version display using commitinfo
|
||||
|
||||
- Imported commitinfo from './00_commitinfo_data.js' and called tsdockerCli.addVersion(commitinfo.version) to surface package/commit version in the Smartcli instance
|
||||
- Change made in ts/tsdocker.cli.ts — small user-facing CLI enhancement; no breaking changes
|
||||
|
||||
## 2026-02-06 - 1.6.0 - feat(docker)
|
||||
add support for no-cache builds and tag built images for local dependency resolution
|
||||
|
||||
- Introduce IBuildCommandOptions.noCache to control --no-cache behavior
|
||||
- Propagate noCache from CLI (via cache flag) through TsDockerManager to Dockerfile.build
|
||||
- Append --no-cache to docker build/buildx commands when noCache is true
|
||||
- After building an image, tag it with full base image references used by dependent Dockerfiles so their FROM lines resolve to the locally-built image
|
||||
- Log tagging actions and execute docker tag via smartshellInstance
|
||||
|
||||
## 2026-02-06 - 1.5.0 - feat(build)
|
||||
add support for selective builds, platform override and build timeout
|
||||
|
||||
- Introduce IBuildCommandOptions with patterns, platform and timeout to control build behavior
|
||||
- Allow manager.build() to accept options and build only matching Dockerfiles (including dependencies) preserving topological order
|
||||
- Add CLI parsing for build/push to accept positional Dockerfile patterns and --platform/--timeout flags
|
||||
- Support single-platform override via docker buildx and multi-platform buildx detection
|
||||
- Implement streaming exec with timeout to kill long-running builds and surface timeout errors
|
||||
|
||||
## 2026-02-04 - 1.4.3 - fix(dockerfile)
|
||||
fix matching of base images to local Dockerfiles by stripping registry prefixes when comparing image references
|
||||
|
||||
- Added Dockerfile.extractRepoVersion(imageRef) to normalize image references by removing registry prefixes (detects registries containing '.' or ':' or 'localhost').
|
||||
- Use extractRepoVersion when checking tagToDockerfile and when mapping local base dockerfiles to ensure comparisons use repo:tag keys rather than full registry-prefixed references.
|
||||
- Prevents mismatches when baseImage includes a registry (e.g. "host.today/repo:version") so it correctly matches a local cleanTag like "repo:version".
|
||||
|
||||
## 2026-01-21 - 1.4.2 - fix(classes.dockerfile)
|
||||
use a single top-level fs import instead of requiring fs inside methods
|
||||
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "@git.zone/tsdocker",
|
||||
"version": "1.4.2",
|
||||
"version": "1.10.0",
|
||||
"private": false,
|
||||
"description": "develop npm modules cross platform with docker",
|
||||
"main": "dist_ts/index.js",
|
||||
|
||||
@@ -3,6 +3,6 @@
|
||||
*/
|
||||
export const commitinfo = {
|
||||
name: '@git.zone/tsdocker',
|
||||
version: '1.4.2',
|
||||
version: '1.10.0',
|
||||
description: 'develop npm modules cross platform with docker'
|
||||
}
|
||||
|
||||
@@ -1,8 +1,8 @@
|
||||
import * as plugins from './tsdocker.plugins.js';
|
||||
import * as paths from './tsdocker.paths.js';
|
||||
import { logger } from './tsdocker.logging.js';
|
||||
import { logger, formatDuration } from './tsdocker.logging.js';
|
||||
import { DockerRegistry } from './classes.dockerregistry.js';
|
||||
import type { IDockerfileOptions, ITsDockerConfig } from './interfaces/index.js';
|
||||
import type { IDockerfileOptions, ITsDockerConfig, IBuildCommandOptions } from './interfaces/index.js';
|
||||
import type { TsDockerManager } from './classes.tsdockermanager.js';
|
||||
import * as fs from 'fs';
|
||||
|
||||
@@ -26,8 +26,10 @@ export class Dockerfile {
|
||||
.map(entry => plugins.path.join(paths.cwd, entry.name));
|
||||
|
||||
const readDockerfilesArray: Dockerfile[] = [];
|
||||
logger.log('info', `found ${fileTree.length} Dockerfiles:`);
|
||||
console.log(fileTree);
|
||||
logger.log('info', `found ${fileTree.length} Dockerfile(s):`);
|
||||
for (const filePath of fileTree) {
|
||||
logger.log('info', ` ${plugins.path.basename(filePath)}`);
|
||||
}
|
||||
|
||||
for (const dockerfilePath of fileTree) {
|
||||
const myDockerfile = new Dockerfile(managerRef, {
|
||||
@@ -58,9 +60,14 @@ export class Dockerfile {
|
||||
const dependencies: Dockerfile[] = [];
|
||||
const baseImage = dockerfile.baseImage;
|
||||
|
||||
// Extract repo:version from baseImage for comparison with cleanTag
|
||||
// baseImage may include a registry prefix (e.g., "host.today/repo:version")
|
||||
// but cleanTag is just "repo:version", so we strip the registry prefix
|
||||
const baseImageKey = Dockerfile.extractRepoVersion(baseImage);
|
||||
|
||||
// Check if the baseImage is among the local Dockerfiles
|
||||
if (tagToDockerfile.has(baseImage)) {
|
||||
const baseDockerfile = tagToDockerfile.get(baseImage)!;
|
||||
if (tagToDockerfile.has(baseImageKey)) {
|
||||
const baseDockerfile = tagToDockerfile.get(baseImageKey)!;
|
||||
dependencies.push(baseDockerfile);
|
||||
dockerfile.localBaseImageDependent = true;
|
||||
dockerfile.localBaseDockerfile = baseDockerfile;
|
||||
@@ -116,8 +123,10 @@ export class Dockerfile {
|
||||
public static async mapDockerfiles(sortedDockerfileArray: Dockerfile[]): Promise<Dockerfile[]> {
|
||||
sortedDockerfileArray.forEach((dockerfileArg) => {
|
||||
if (dockerfileArg.localBaseImageDependent) {
|
||||
// Extract repo:version from baseImage for comparison with cleanTag
|
||||
const baseImageKey = Dockerfile.extractRepoVersion(dockerfileArg.baseImage);
|
||||
sortedDockerfileArray.forEach((dockfile2: Dockerfile) => {
|
||||
if (dockfile2.cleanTag === dockerfileArg.baseImage) {
|
||||
if (dockfile2.cleanTag === baseImageKey) {
|
||||
dockerfileArg.localBaseDockerfile = dockfile2;
|
||||
}
|
||||
});
|
||||
@@ -129,10 +138,36 @@ export class Dockerfile {
|
||||
/**
|
||||
* Builds the corresponding real docker image for each Dockerfile class instance
|
||||
*/
|
||||
public static async buildDockerfiles(sortedArrayArg: Dockerfile[]): Promise<Dockerfile[]> {
|
||||
for (const dockerfileArg of sortedArrayArg) {
|
||||
await dockerfileArg.build();
|
||||
public static async buildDockerfiles(
|
||||
sortedArrayArg: Dockerfile[],
|
||||
options?: { platform?: string; timeout?: number; noCache?: boolean; verbose?: boolean },
|
||||
): Promise<Dockerfile[]> {
|
||||
const total = sortedArrayArg.length;
|
||||
const overallStart = Date.now();
|
||||
|
||||
for (let i = 0; i < total; i++) {
|
||||
const dockerfileArg = sortedArrayArg[i];
|
||||
const progress = `(${i + 1}/${total})`;
|
||||
logger.log('info', `${progress} Building ${dockerfileArg.cleanTag}...`);
|
||||
|
||||
const elapsed = await dockerfileArg.build(options);
|
||||
logger.log('ok', `${progress} Built ${dockerfileArg.cleanTag} in ${formatDuration(elapsed)}`);
|
||||
|
||||
// Tag the built image with the full base image references used by dependent Dockerfiles,
|
||||
// so their FROM lines resolve to the locally-built image instead of pulling from a registry.
|
||||
const dependentBaseImages = new Set<string>();
|
||||
for (const other of sortedArrayArg) {
|
||||
if (other.localBaseDockerfile === dockerfileArg && other.baseImage !== dockerfileArg.buildTag) {
|
||||
dependentBaseImages.add(other.baseImage);
|
||||
}
|
||||
}
|
||||
for (const fullTag of dependentBaseImages) {
|
||||
logger.log('info', `Tagging ${dockerfileArg.buildTag} as ${fullTag} for local dependency resolution`);
|
||||
await smartshellInstance.exec(`docker tag ${dockerfileArg.buildTag} ${fullTag}`);
|
||||
}
|
||||
}
|
||||
|
||||
logger.log('info', `Total build time: ${formatDuration(Date.now() - overallStart)}`);
|
||||
return sortedArrayArg;
|
||||
}
|
||||
|
||||
@@ -140,9 +175,19 @@ export class Dockerfile {
|
||||
* Tests all Dockerfiles by calling Dockerfile.test()
|
||||
*/
|
||||
public static async testDockerfiles(sortedArrayArg: Dockerfile[]): Promise<Dockerfile[]> {
|
||||
for (const dockerfileArg of sortedArrayArg) {
|
||||
await dockerfileArg.test();
|
||||
const total = sortedArrayArg.length;
|
||||
const overallStart = Date.now();
|
||||
|
||||
for (let i = 0; i < total; i++) {
|
||||
const dockerfileArg = sortedArrayArg[i];
|
||||
const progress = `(${i + 1}/${total})`;
|
||||
logger.log('info', `${progress} Testing ${dockerfileArg.cleanTag}...`);
|
||||
|
||||
const elapsed = await dockerfileArg.test();
|
||||
logger.log('ok', `${progress} Tested ${dockerfileArg.cleanTag} in ${formatDuration(elapsed)}`);
|
||||
}
|
||||
|
||||
logger.log('info', `Total test time: ${formatDuration(Date.now() - overallStart)}`);
|
||||
return sortedArrayArg;
|
||||
}
|
||||
|
||||
@@ -231,6 +276,34 @@ export class Dockerfile {
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Extracts the repo:version part from a full image reference, stripping any registry prefix.
|
||||
* Examples:
|
||||
* "registry.example.com/repo:version" -> "repo:version"
|
||||
* "repo:version" -> "repo:version"
|
||||
* "host.today/ht-docker-node:npmci" -> "ht-docker-node:npmci"
|
||||
*/
|
||||
private static extractRepoVersion(imageRef: string): string {
|
||||
const parts = imageRef.split('/');
|
||||
if (parts.length === 1) {
|
||||
// No registry prefix: "repo:version"
|
||||
return imageRef;
|
||||
}
|
||||
|
||||
// Check if first part looks like a registry (contains '.' or ':' or is 'localhost')
|
||||
const firstPart = parts[0];
|
||||
const looksLikeRegistry =
|
||||
firstPart.includes('.') || firstPart.includes(':') || firstPart === 'localhost';
|
||||
|
||||
if (looksLikeRegistry) {
|
||||
// Strip registry: "registry.example.com/repo:version" -> "repo:version"
|
||||
return parts.slice(1).join('/');
|
||||
}
|
||||
|
||||
// No registry prefix, could be "org/repo:version"
|
||||
return imageRef;
|
||||
}
|
||||
|
||||
/**
|
||||
* Returns the docker tag string for a given registry and repo
|
||||
*/
|
||||
@@ -327,18 +400,32 @@ export class Dockerfile {
|
||||
/**
|
||||
* Builds the Dockerfile
|
||||
*/
|
||||
public async build(): Promise<void> {
|
||||
logger.log('info', 'now building Dockerfile for ' + this.cleanTag);
|
||||
public async build(options?: { platform?: string; timeout?: number; noCache?: boolean; verbose?: boolean }): Promise<number> {
|
||||
const startTime = Date.now();
|
||||
const buildArgsString = await Dockerfile.getDockerBuildArgs(this.managerRef);
|
||||
const config = this.managerRef.config;
|
||||
const platformOverride = options?.platform;
|
||||
const timeout = options?.timeout;
|
||||
const noCacheFlag = options?.noCache ? ' --no-cache' : '';
|
||||
const verbose = options?.verbose ?? false;
|
||||
|
||||
let buildContextFlag = '';
|
||||
if (this.localBaseImageDependent && this.localBaseDockerfile) {
|
||||
const fromImage = this.baseImage;
|
||||
const localTag = this.localBaseDockerfile.buildTag;
|
||||
buildContextFlag = ` --build-context "${fromImage}=docker-image://${localTag}"`;
|
||||
logger.log('info', `Using local build context: ${fromImage} -> docker-image://${localTag}`);
|
||||
}
|
||||
|
||||
let buildCommand: string;
|
||||
|
||||
// Check if multi-platform build is needed
|
||||
if (config.platforms && config.platforms.length > 1) {
|
||||
if (platformOverride) {
|
||||
// Single platform override via buildx
|
||||
buildCommand = `docker buildx build --platform ${platformOverride}${noCacheFlag}${buildContextFlag} --load -t ${this.buildTag} -f ${this.filePath} ${buildArgsString} .`;
|
||||
} else if (config.platforms && config.platforms.length > 1) {
|
||||
// Multi-platform build using buildx
|
||||
const platformString = config.platforms.join(',');
|
||||
buildCommand = `docker buildx build --platform ${platformString} -t ${this.buildTag} -f ${this.filePath} ${buildArgsString} .`;
|
||||
buildCommand = `docker buildx build --platform ${platformString}${noCacheFlag}${buildContextFlag} -t ${this.buildTag} -f ${this.filePath} ${buildArgsString} .`;
|
||||
|
||||
if (config.push) {
|
||||
buildCommand += ' --push';
|
||||
@@ -348,17 +435,39 @@ export class Dockerfile {
|
||||
} else {
|
||||
// Standard build
|
||||
const versionLabel = this.managerRef.projectInfo?.npm?.version || 'unknown';
|
||||
buildCommand = `docker build --label="version=${versionLabel}" -t ${this.buildTag} -f ${this.filePath} ${buildArgsString} .`;
|
||||
buildCommand = `docker build --label="version=${versionLabel}"${noCacheFlag} -t ${this.buildTag} -f ${this.filePath} ${buildArgsString} .`;
|
||||
}
|
||||
|
||||
const result = await smartshellInstance.exec(buildCommand);
|
||||
if (result.exitCode !== 0) {
|
||||
logger.log('error', `Build failed for ${this.cleanTag}`);
|
||||
console.log(result.stdout);
|
||||
throw new Error(`Build failed for ${this.cleanTag}`);
|
||||
if (timeout) {
|
||||
// Use streaming execution with timeout
|
||||
const streaming = verbose
|
||||
? await smartshellInstance.execStreaming(buildCommand)
|
||||
: await smartshellInstance.execStreamingSilent(buildCommand);
|
||||
const timeoutPromise = new Promise<never>((_, reject) => {
|
||||
setTimeout(() => {
|
||||
streaming.childProcess.kill();
|
||||
reject(new Error(`Build timed out after ${timeout}s for ${this.cleanTag}`));
|
||||
}, timeout * 1000);
|
||||
});
|
||||
const result = await Promise.race([streaming.finalPromise, timeoutPromise]);
|
||||
if (result.exitCode !== 0) {
|
||||
logger.log('error', `Build failed for ${this.cleanTag}`);
|
||||
throw new Error(`Build failed for ${this.cleanTag}`);
|
||||
}
|
||||
} else {
|
||||
const result = verbose
|
||||
? await smartshellInstance.exec(buildCommand)
|
||||
: await smartshellInstance.execSilent(buildCommand);
|
||||
if (result.exitCode !== 0) {
|
||||
logger.log('error', `Build failed for ${this.cleanTag}`);
|
||||
if (!verbose && result.stdout) {
|
||||
logger.log('error', `Build output:\n${result.stdout}`);
|
||||
}
|
||||
throw new Error(`Build failed for ${this.cleanTag}`);
|
||||
}
|
||||
}
|
||||
|
||||
logger.log('ok', `Built ${this.cleanTag}`);
|
||||
return Date.now() - startTime;
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -388,7 +497,7 @@ export class Dockerfile {
|
||||
|
||||
if (inspectResult.exitCode === 0 && inspectResult.stdout.includes('@')) {
|
||||
const imageDigest = inspectResult.stdout.split('@')[1]?.trim();
|
||||
console.log(`The image ${this.pushTag} has digest ${imageDigest}`);
|
||||
logger.log('info', `The image ${this.pushTag} has digest ${imageDigest}`);
|
||||
}
|
||||
|
||||
logger.log('ok', `Pushed ${this.pushTag}`);
|
||||
@@ -415,15 +524,14 @@ export class Dockerfile {
|
||||
/**
|
||||
* Tests the Dockerfile by running a test script if it exists
|
||||
*/
|
||||
public async test(): Promise<void> {
|
||||
public async test(): Promise<number> {
|
||||
const startTime = Date.now();
|
||||
const testDir = this.managerRef.config.testDir || plugins.path.join(paths.cwd, 'test');
|
||||
const testFile = plugins.path.join(testDir, 'test_' + this.version + '.sh');
|
||||
|
||||
const testFileExists = fs.existsSync(testFile);
|
||||
|
||||
if (testFileExists) {
|
||||
logger.log('info', `Running tests for ${this.cleanTag}`);
|
||||
|
||||
// Run tests in container
|
||||
await smartshellInstance.exec(
|
||||
`docker run --name tsdocker_test_container --entrypoint="bash" ${this.buildTag} -c "mkdir /tsdocker_test"`
|
||||
@@ -442,11 +550,11 @@ export class Dockerfile {
|
||||
if (testResult.exitCode !== 0) {
|
||||
throw new Error(`Tests failed for ${this.cleanTag}`);
|
||||
}
|
||||
|
||||
logger.log('ok', `Tests passed for ${this.cleanTag}`);
|
||||
} else {
|
||||
logger.log('warn', `Skipping tests for ${this.cleanTag} because no test file was found at ${testFile}`);
|
||||
logger.log('warn', `Skipping tests for ${this.cleanTag} — no test file at ${testFile}`);
|
||||
}
|
||||
|
||||
return Date.now() - startTime;
|
||||
}
|
||||
|
||||
/**
|
||||
|
||||
108
ts/classes.tsdockercache.ts
Normal file
108
ts/classes.tsdockercache.ts
Normal file
@@ -0,0 +1,108 @@
|
||||
import * as crypto from 'crypto';
|
||||
import * as fs from 'fs';
|
||||
import * as path from 'path';
|
||||
import * as plugins from './tsdocker.plugins.js';
|
||||
import * as paths from './tsdocker.paths.js';
|
||||
import { logger } from './tsdocker.logging.js';
|
||||
import type { ICacheData, ICacheEntry } from './interfaces/index.js';
|
||||
|
||||
const smartshellInstance = new plugins.smartshell.Smartshell({
|
||||
executor: 'bash',
|
||||
});
|
||||
|
||||
/**
|
||||
* Manages content-hash-based build caching for Dockerfiles.
|
||||
* Cache is stored in .nogit/tsdocker_support.json.
|
||||
*/
|
||||
export class TsDockerCache {
|
||||
private cacheFilePath: string;
|
||||
private data: ICacheData;
|
||||
|
||||
constructor() {
|
||||
this.cacheFilePath = path.join(paths.cwd, '.nogit', 'tsdocker_support.json');
|
||||
this.data = { version: 1, entries: {} };
|
||||
}
|
||||
|
||||
/**
|
||||
* Loads cache data from disk. Falls back to empty cache on missing/corrupt file.
|
||||
*/
|
||||
public load(): void {
|
||||
try {
|
||||
const raw = fs.readFileSync(this.cacheFilePath, 'utf-8');
|
||||
const parsed = JSON.parse(raw);
|
||||
if (parsed && parsed.version === 1 && parsed.entries) {
|
||||
this.data = parsed;
|
||||
} else {
|
||||
logger.log('warn', '[cache] Cache file has unexpected format, starting fresh');
|
||||
this.data = { version: 1, entries: {} };
|
||||
}
|
||||
} catch {
|
||||
// Missing or corrupt file — start fresh
|
||||
this.data = { version: 1, entries: {} };
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Saves cache data to disk. Creates .nogit directory if needed.
|
||||
*/
|
||||
public save(): void {
|
||||
const dir = path.dirname(this.cacheFilePath);
|
||||
fs.mkdirSync(dir, { recursive: true });
|
||||
fs.writeFileSync(this.cacheFilePath, JSON.stringify(this.data, null, 2), 'utf-8');
|
||||
}
|
||||
|
||||
/**
|
||||
* Computes SHA-256 hash of Dockerfile content.
|
||||
*/
|
||||
public computeContentHash(content: string): string {
|
||||
return crypto.createHash('sha256').update(content).digest('hex');
|
||||
}
|
||||
|
||||
/**
|
||||
* Checks whether a build can be skipped for the given Dockerfile.
|
||||
* Logs detailed diagnostics and returns true if the build should be skipped.
|
||||
*/
|
||||
public async shouldSkipBuild(cleanTag: string, content: string): Promise<boolean> {
|
||||
const contentHash = this.computeContentHash(content);
|
||||
const entry = this.data.entries[cleanTag];
|
||||
|
||||
if (!entry) {
|
||||
logger.log('info', `[cache] ${cleanTag}: no cached entry, will build`);
|
||||
return false;
|
||||
}
|
||||
|
||||
const hashMatch = entry.contentHash === contentHash;
|
||||
logger.log('info', `[cache] ${cleanTag}: hash ${hashMatch ? 'matches' : 'changed'}`);
|
||||
|
||||
if (!hashMatch) {
|
||||
logger.log('info', `[cache] ${cleanTag}: content changed, will build`);
|
||||
return false;
|
||||
}
|
||||
|
||||
// Hash matches — verify the image still exists locally
|
||||
const inspectResult = await smartshellInstance.exec(
|
||||
`docker image inspect ${entry.imageId} > /dev/null 2>&1`
|
||||
);
|
||||
const available = inspectResult.exitCode === 0;
|
||||
|
||||
if (available) {
|
||||
logger.log('info', `[cache] ${cleanTag}: cache hit, skipping build`);
|
||||
return true;
|
||||
}
|
||||
|
||||
logger.log('info', `[cache] ${cleanTag}: image no longer available, will build`);
|
||||
return false;
|
||||
}
|
||||
|
||||
/**
|
||||
* Records a successful build in the cache.
|
||||
*/
|
||||
public recordBuild(cleanTag: string, content: string, imageId: string, buildTag: string): void {
|
||||
this.data.entries[cleanTag] = {
|
||||
contentHash: this.computeContentHash(content),
|
||||
imageId,
|
||||
buildTag,
|
||||
timestamp: Date.now(),
|
||||
};
|
||||
}
|
||||
}
|
||||
@@ -1,10 +1,11 @@
|
||||
import * as plugins from './tsdocker.plugins.js';
|
||||
import * as paths from './tsdocker.paths.js';
|
||||
import { logger } from './tsdocker.logging.js';
|
||||
import { logger, formatDuration } from './tsdocker.logging.js';
|
||||
import { Dockerfile } from './classes.dockerfile.js';
|
||||
import { DockerRegistry } from './classes.dockerregistry.js';
|
||||
import { RegistryStorage } from './classes.registrystorage.js';
|
||||
import type { ITsDockerConfig } from './interfaces/index.js';
|
||||
import { TsDockerCache } from './classes.tsdockercache.js';
|
||||
import type { ITsDockerConfig, IBuildCommandOptions } from './interfaces/index.js';
|
||||
|
||||
const smartshellInstance = new plugins.smartshell.Smartshell({
|
||||
executor: 'bash',
|
||||
@@ -90,9 +91,10 @@ export class TsDockerManager {
|
||||
}
|
||||
|
||||
/**
|
||||
* Builds all discovered Dockerfiles in dependency order
|
||||
* Builds discovered Dockerfiles in dependency order.
|
||||
* When options.patterns is provided, only matching Dockerfiles (and their dependencies) are built.
|
||||
*/
|
||||
public async build(): Promise<Dockerfile[]> {
|
||||
public async build(options?: IBuildCommandOptions): Promise<Dockerfile[]> {
|
||||
if (this.dockerfiles.length === 0) {
|
||||
await this.discoverDockerfiles();
|
||||
}
|
||||
@@ -102,16 +104,121 @@ export class TsDockerManager {
|
||||
return [];
|
||||
}
|
||||
|
||||
// Determine which Dockerfiles to build
|
||||
let toBuild = this.dockerfiles;
|
||||
|
||||
if (options?.patterns && options.patterns.length > 0) {
|
||||
// Filter to matching Dockerfiles
|
||||
const matched = this.dockerfiles.filter((df) => {
|
||||
const basename = plugins.path.basename(df.filePath);
|
||||
return options.patterns!.some((pattern) => {
|
||||
if (pattern.includes('*') || pattern.includes('?')) {
|
||||
// Convert glob pattern to regex
|
||||
const regexStr = '^' + pattern.replace(/\*/g, '.*').replace(/\?/g, '.') + '$';
|
||||
return new RegExp(regexStr).test(basename);
|
||||
}
|
||||
return basename === pattern;
|
||||
});
|
||||
});
|
||||
|
||||
if (matched.length === 0) {
|
||||
logger.log('warn', `No Dockerfiles matched patterns: ${options.patterns.join(', ')}`);
|
||||
return [];
|
||||
}
|
||||
|
||||
// Resolve dependency chain and preserve topological order
|
||||
toBuild = this.resolveWithDependencies(matched, this.dockerfiles);
|
||||
logger.log('info', `Matched ${matched.length} Dockerfile(s), building ${toBuild.length} (including dependencies)`);
|
||||
}
|
||||
|
||||
// Check if buildx is needed
|
||||
if (this.config.platforms && this.config.platforms.length > 1) {
|
||||
if (options?.platform || (this.config.platforms && this.config.platforms.length > 1)) {
|
||||
await this.ensureBuildx();
|
||||
}
|
||||
|
||||
logger.log('info', `Building ${this.dockerfiles.length} Dockerfiles...`);
|
||||
await Dockerfile.buildDockerfiles(this.dockerfiles);
|
||||
logger.log('info', '');
|
||||
logger.log('info', '=== BUILD PHASE ===');
|
||||
logger.log('info', `Building ${toBuild.length} Dockerfile(s)...`);
|
||||
|
||||
if (options?.cached) {
|
||||
// === CACHED MODE: skip builds for unchanged Dockerfiles ===
|
||||
logger.log('info', '(cached mode active)');
|
||||
const cache = new TsDockerCache();
|
||||
cache.load();
|
||||
|
||||
const total = toBuild.length;
|
||||
const overallStart = Date.now();
|
||||
|
||||
for (let i = 0; i < total; i++) {
|
||||
const dockerfileArg = toBuild[i];
|
||||
const progress = `(${i + 1}/${total})`;
|
||||
const skip = await cache.shouldSkipBuild(dockerfileArg.cleanTag, dockerfileArg.content);
|
||||
if (skip) {
|
||||
logger.log('ok', `${progress} Skipped ${dockerfileArg.cleanTag} (cached)`);
|
||||
continue;
|
||||
}
|
||||
|
||||
// Cache miss — build this Dockerfile
|
||||
logger.log('info', `${progress} Building ${dockerfileArg.cleanTag}...`);
|
||||
const elapsed = await dockerfileArg.build({
|
||||
platform: options?.platform,
|
||||
timeout: options?.timeout,
|
||||
noCache: options?.noCache,
|
||||
verbose: options?.verbose,
|
||||
});
|
||||
logger.log('ok', `${progress} Built ${dockerfileArg.cleanTag} in ${formatDuration(elapsed)}`);
|
||||
|
||||
const imageId = await dockerfileArg.getId();
|
||||
cache.recordBuild(dockerfileArg.cleanTag, dockerfileArg.content, imageId, dockerfileArg.buildTag);
|
||||
}
|
||||
|
||||
logger.log('info', `Total build time: ${formatDuration(Date.now() - overallStart)}`);
|
||||
|
||||
// Perform dependency tagging for all Dockerfiles (even cache hits, since tags may be stale)
|
||||
for (const dockerfileArg of toBuild) {
|
||||
const dependentBaseImages = new Set<string>();
|
||||
for (const other of toBuild) {
|
||||
if (other.localBaseDockerfile === dockerfileArg && other.baseImage !== dockerfileArg.buildTag) {
|
||||
dependentBaseImages.add(other.baseImage);
|
||||
}
|
||||
}
|
||||
for (const fullTag of dependentBaseImages) {
|
||||
logger.log('info', `Tagging ${dockerfileArg.buildTag} as ${fullTag} for local dependency resolution`);
|
||||
await smartshellInstance.exec(`docker tag ${dockerfileArg.buildTag} ${fullTag}`);
|
||||
}
|
||||
}
|
||||
|
||||
cache.save();
|
||||
} else {
|
||||
// === STANDARD MODE: build all via static helper ===
|
||||
await Dockerfile.buildDockerfiles(toBuild, {
|
||||
platform: options?.platform,
|
||||
timeout: options?.timeout,
|
||||
noCache: options?.noCache,
|
||||
verbose: options?.verbose,
|
||||
});
|
||||
}
|
||||
|
||||
logger.log('success', 'All Dockerfiles built successfully');
|
||||
|
||||
return this.dockerfiles;
|
||||
return toBuild;
|
||||
}
|
||||
|
||||
/**
|
||||
* Resolves a set of target Dockerfiles to include all their local base image dependencies,
|
||||
* preserving the original topological build order.
|
||||
*/
|
||||
private resolveWithDependencies(targets: Dockerfile[], allSorted: Dockerfile[]): Dockerfile[] {
|
||||
const needed = new Set<Dockerfile>();
|
||||
const addWithDeps = (df: Dockerfile) => {
|
||||
if (needed.has(df)) return;
|
||||
needed.add(df);
|
||||
if (df.localBaseImageDependent && df.localBaseDockerfile) {
|
||||
addWithDeps(df.localBaseDockerfile);
|
||||
}
|
||||
};
|
||||
for (const df of targets) addWithDeps(df);
|
||||
return allSorted.filter((df) => needed.has(df));
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -215,6 +322,8 @@ export class TsDockerManager {
|
||||
return;
|
||||
}
|
||||
|
||||
logger.log('info', '');
|
||||
logger.log('info', '=== TEST PHASE ===');
|
||||
await Dockerfile.testDockerfiles(this.dockerfiles);
|
||||
logger.log('success', 'All tests completed');
|
||||
}
|
||||
@@ -227,19 +336,21 @@ export class TsDockerManager {
|
||||
await this.discoverDockerfiles();
|
||||
}
|
||||
|
||||
console.log('\nDiscovered Dockerfiles:');
|
||||
console.log('========================\n');
|
||||
logger.log('info', '');
|
||||
logger.log('info', 'Discovered Dockerfiles:');
|
||||
logger.log('info', '========================');
|
||||
logger.log('info', '');
|
||||
|
||||
for (let i = 0; i < this.dockerfiles.length; i++) {
|
||||
const df = this.dockerfiles[i];
|
||||
console.log(`${i + 1}. ${df.filePath}`);
|
||||
console.log(` Tag: ${df.cleanTag}`);
|
||||
console.log(` Base Image: ${df.baseImage}`);
|
||||
console.log(` Version: ${df.version}`);
|
||||
logger.log('info', `${i + 1}. ${df.filePath}`);
|
||||
logger.log('info', ` Tag: ${df.cleanTag}`);
|
||||
logger.log('info', ` Base Image: ${df.baseImage}`);
|
||||
logger.log('info', ` Version: ${df.version}`);
|
||||
if (df.localBaseImageDependent) {
|
||||
console.log(` Depends on: ${df.localBaseDockerfile?.cleanTag}`);
|
||||
logger.log('info', ` Depends on: ${df.localBaseDockerfile?.cleanTag}`);
|
||||
}
|
||||
console.log('');
|
||||
logger.log('info', '');
|
||||
}
|
||||
|
||||
return this.dockerfiles;
|
||||
|
||||
@@ -68,3 +68,27 @@ export interface IPushResult {
|
||||
digest?: string;
|
||||
error?: string;
|
||||
}
|
||||
|
||||
/**
|
||||
* Options for the build command
|
||||
*/
|
||||
export interface IBuildCommandOptions {
|
||||
patterns?: string[]; // Dockerfile name patterns (e.g., ['Dockerfile_base', 'Dockerfile_*'])
|
||||
platform?: string; // Single platform override (e.g., 'linux/arm64')
|
||||
timeout?: number; // Build timeout in seconds
|
||||
noCache?: boolean; // Force rebuild without Docker layer cache (--no-cache)
|
||||
cached?: boolean; // Skip builds when Dockerfile content hasn't changed
|
||||
verbose?: boolean; // Stream raw docker build output (default: silent)
|
||||
}
|
||||
|
||||
export interface ICacheEntry {
|
||||
contentHash: string; // SHA-256 hex of Dockerfile content
|
||||
imageId: string; // Docker image ID (sha256:...)
|
||||
buildTag: string;
|
||||
timestamp: number; // Unix ms
|
||||
}
|
||||
|
||||
export interface ICacheData {
|
||||
version: 1;
|
||||
entries: { [cleanTag: string]: ICacheEntry };
|
||||
}
|
||||
|
||||
@@ -7,8 +7,11 @@ import * as DockerModule from './tsdocker.docker.js';
|
||||
|
||||
import { logger, ora } from './tsdocker.logging.js';
|
||||
import { TsDockerManager } from './classes.tsdockermanager.js';
|
||||
import type { IBuildCommandOptions } from './interfaces/index.js';
|
||||
import { commitinfo } from './00_commitinfo_data.js';
|
||||
|
||||
const tsdockerCli = new plugins.smartcli.Smartcli();
|
||||
tsdockerCli.addVersion(commitinfo.version);
|
||||
|
||||
export let run = () => {
|
||||
// Default command: run tests in container (legacy behavior)
|
||||
@@ -23,14 +26,37 @@ export let run = () => {
|
||||
});
|
||||
|
||||
/**
|
||||
* Build all Dockerfiles in dependency order
|
||||
* Build Dockerfiles in dependency order
|
||||
* Usage: tsdocker build [Dockerfile_patterns...] [--platform=linux/arm64] [--timeout=600]
|
||||
*/
|
||||
tsdockerCli.addCommand('build').subscribe(async argvArg => {
|
||||
try {
|
||||
const config = await ConfigModule.run();
|
||||
const manager = new TsDockerManager(config);
|
||||
await manager.prepare();
|
||||
await manager.build();
|
||||
|
||||
const buildOptions: IBuildCommandOptions = {};
|
||||
const patterns = argvArg._.slice(1) as string[];
|
||||
if (patterns.length > 0) {
|
||||
buildOptions.patterns = patterns;
|
||||
}
|
||||
if (argvArg.platform) {
|
||||
buildOptions.platform = argvArg.platform as string;
|
||||
}
|
||||
if (argvArg.timeout) {
|
||||
buildOptions.timeout = Number(argvArg.timeout);
|
||||
}
|
||||
if (argvArg.cache === false) {
|
||||
buildOptions.noCache = true;
|
||||
}
|
||||
if (argvArg.cached) {
|
||||
buildOptions.cached = true;
|
||||
}
|
||||
if (argvArg.verbose) {
|
||||
buildOptions.verbose = true;
|
||||
}
|
||||
|
||||
await manager.build(buildOptions);
|
||||
logger.log('success', 'Build completed successfully');
|
||||
} catch (err) {
|
||||
logger.log('error', `Build failed: ${(err as Error).message}`);
|
||||
@@ -40,6 +66,7 @@ export let run = () => {
|
||||
|
||||
/**
|
||||
* Push built images to configured registries
|
||||
* Usage: tsdocker push [Dockerfile_patterns...] [--platform=linux/arm64] [--timeout=600] [--registry=url]
|
||||
*/
|
||||
tsdockerCli.addCommand('push').subscribe(async argvArg => {
|
||||
try {
|
||||
@@ -50,11 +77,30 @@ export let run = () => {
|
||||
// Login first
|
||||
await manager.login();
|
||||
|
||||
// Build images first (if not already built)
|
||||
await manager.build();
|
||||
// Parse build options from positional args and flags
|
||||
const buildOptions: IBuildCommandOptions = {};
|
||||
const patterns = argvArg._.slice(1) as string[];
|
||||
if (patterns.length > 0) {
|
||||
buildOptions.patterns = patterns;
|
||||
}
|
||||
if (argvArg.platform) {
|
||||
buildOptions.platform = argvArg.platform as string;
|
||||
}
|
||||
if (argvArg.timeout) {
|
||||
buildOptions.timeout = Number(argvArg.timeout);
|
||||
}
|
||||
if (argvArg.cache === false) {
|
||||
buildOptions.noCache = true;
|
||||
}
|
||||
if (argvArg.verbose) {
|
||||
buildOptions.verbose = true;
|
||||
}
|
||||
|
||||
// Get registry from arguments if specified
|
||||
const registryArg = argvArg._[1]; // e.g., tsdocker push registry.gitlab.com
|
||||
// Build images first (if not already built)
|
||||
await manager.build(buildOptions);
|
||||
|
||||
// Get registry from --registry flag
|
||||
const registryArg = argvArg.registry as string | undefined;
|
||||
const registries = registryArg ? [registryArg] : undefined;
|
||||
|
||||
await manager.push(registries);
|
||||
@@ -101,7 +147,17 @@ export let run = () => {
|
||||
await manager.prepare();
|
||||
|
||||
// Build images first
|
||||
await manager.build();
|
||||
const buildOptions: IBuildCommandOptions = {};
|
||||
if (argvArg.cache === false) {
|
||||
buildOptions.noCache = true;
|
||||
}
|
||||
if (argvArg.cached) {
|
||||
buildOptions.cached = true;
|
||||
}
|
||||
if (argvArg.verbose) {
|
||||
buildOptions.verbose = true;
|
||||
}
|
||||
await manager.build(buildOptions);
|
||||
|
||||
// Run tests
|
||||
await manager.test();
|
||||
|
||||
@@ -15,3 +15,12 @@ export const logger = new plugins.smartlog.Smartlog({
|
||||
logger.addLogDestination(new plugins.smartlogDestinationLocal.DestinationLocal());
|
||||
|
||||
export const ora = new plugins.smartlogSouceOra.SmartlogSourceOra();
|
||||
|
||||
export function formatDuration(ms: number): string {
|
||||
if (ms < 1000) return `${ms}ms`;
|
||||
const totalSeconds = ms / 1000;
|
||||
if (totalSeconds < 60) return `${totalSeconds.toFixed(1)}s`;
|
||||
const minutes = Math.floor(totalSeconds / 60);
|
||||
const seconds = Math.round(totalSeconds % 60);
|
||||
return `${minutes}m ${seconds}s`;
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user