Compare commits
12 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| 16cd0bbd87 | |||
| cc83743f9a | |||
| 7131c16f80 | |||
| 02688861f4 | |||
| 3a8b301b3e | |||
| c09bef33c3 | |||
| 32eb0d1d77 | |||
| 7cac628975 | |||
| c279dbd55e | |||
| 7b7064864e | |||
| 36f06cef09 | |||
| b0f87deb4b |
47
changelog.md
47
changelog.md
@@ -1,5 +1,52 @@
|
|||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
|
## 2026-02-06 - 1.8.0 - feat(build)
|
||||||
|
add optional content-hash based build cache to skip rebuilding unchanged Dockerfiles
|
||||||
|
|
||||||
|
- Introduce TsDockerCache to compute SHA-256 of Dockerfile content and persist cache to .nogit/tsdocker_support.json
|
||||||
|
- Add ICacheEntry and ICacheData interfaces and a cached flag to IBuildCommandOptions
|
||||||
|
- Integrate cached mode in TsDockerManager: skip builds on cache hits, verify image presence, record builds on misses, and still perform dependency tagging
|
||||||
|
- Expose --cached option in CLI to enable the cached build flow
|
||||||
|
- Cache records store contentHash, imageId, buildTag and timestamp
|
||||||
|
|
||||||
|
## 2026-02-06 - 1.7.0 - feat(cli)
|
||||||
|
add CLI version display using commitinfo
|
||||||
|
|
||||||
|
- Imported commitinfo from './00_commitinfo_data.js' and called tsdockerCli.addVersion(commitinfo.version) to surface package/commit version in the Smartcli instance
|
||||||
|
- Change made in ts/tsdocker.cli.ts — small user-facing CLI enhancement; no breaking changes
|
||||||
|
|
||||||
|
## 2026-02-06 - 1.6.0 - feat(docker)
|
||||||
|
add support for no-cache builds and tag built images for local dependency resolution
|
||||||
|
|
||||||
|
- Introduce IBuildCommandOptions.noCache to control --no-cache behavior
|
||||||
|
- Propagate noCache from CLI (via cache flag) through TsDockerManager to Dockerfile.build
|
||||||
|
- Append --no-cache to docker build/buildx commands when noCache is true
|
||||||
|
- After building an image, tag it with full base image references used by dependent Dockerfiles so their FROM lines resolve to the locally-built image
|
||||||
|
- Log tagging actions and execute docker tag via smartshellInstance
|
||||||
|
|
||||||
|
## 2026-02-06 - 1.5.0 - feat(build)
|
||||||
|
add support for selective builds, platform override and build timeout
|
||||||
|
|
||||||
|
- Introduce IBuildCommandOptions with patterns, platform and timeout to control build behavior
|
||||||
|
- Allow manager.build() to accept options and build only matching Dockerfiles (including dependencies) preserving topological order
|
||||||
|
- Add CLI parsing for build/push to accept positional Dockerfile patterns and --platform/--timeout flags
|
||||||
|
- Support single-platform override via docker buildx and multi-platform buildx detection
|
||||||
|
- Implement streaming exec with timeout to kill long-running builds and surface timeout errors
|
||||||
|
|
||||||
|
## 2026-02-04 - 1.4.3 - fix(dockerfile)
|
||||||
|
fix matching of base images to local Dockerfiles by stripping registry prefixes when comparing image references
|
||||||
|
|
||||||
|
- Added Dockerfile.extractRepoVersion(imageRef) to normalize image references by removing registry prefixes (detects registries containing '.' or ':' or 'localhost').
|
||||||
|
- Use extractRepoVersion when checking tagToDockerfile and when mapping local base dockerfiles to ensure comparisons use repo:tag keys rather than full registry-prefixed references.
|
||||||
|
- Prevents mismatches when baseImage includes a registry (e.g. "host.today/repo:version") so it correctly matches a local cleanTag like "repo:version".
|
||||||
|
|
||||||
|
## 2026-01-21 - 1.4.2 - fix(classes.dockerfile)
|
||||||
|
use a single top-level fs import instead of requiring fs inside methods
|
||||||
|
|
||||||
|
- Added top-level import: import * as fs from 'fs' in ts/classes.dockerfile.ts
|
||||||
|
- Removed inline require('fs') calls and replaced with the imported fs in constructor and test() to keep imports consistent
|
||||||
|
- No behavioral change expected; this is a cleanup/refactor to standardize module usage
|
||||||
|
|
||||||
## 2026-01-20 - 1.4.1 - fix(docs)
|
## 2026-01-20 - 1.4.1 - fix(docs)
|
||||||
update README: expand usage, installation, quick start, features, troubleshooting and migration notes
|
update README: expand usage, installation, quick start, features, troubleshooting and migration notes
|
||||||
|
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@git.zone/tsdocker",
|
"name": "@git.zone/tsdocker",
|
||||||
"version": "1.4.1",
|
"version": "1.8.0",
|
||||||
"private": false,
|
"private": false,
|
||||||
"description": "develop npm modules cross platform with docker",
|
"description": "develop npm modules cross platform with docker",
|
||||||
"main": "dist_ts/index.js",
|
"main": "dist_ts/index.js",
|
||||||
|
|||||||
@@ -3,6 +3,6 @@
|
|||||||
*/
|
*/
|
||||||
export const commitinfo = {
|
export const commitinfo = {
|
||||||
name: '@git.zone/tsdocker',
|
name: '@git.zone/tsdocker',
|
||||||
version: '1.4.1',
|
version: '1.8.0',
|
||||||
description: 'develop npm modules cross platform with docker'
|
description: 'develop npm modules cross platform with docker'
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -2,8 +2,9 @@ import * as plugins from './tsdocker.plugins.js';
|
|||||||
import * as paths from './tsdocker.paths.js';
|
import * as paths from './tsdocker.paths.js';
|
||||||
import { logger } from './tsdocker.logging.js';
|
import { logger } from './tsdocker.logging.js';
|
||||||
import { DockerRegistry } from './classes.dockerregistry.js';
|
import { DockerRegistry } from './classes.dockerregistry.js';
|
||||||
import type { IDockerfileOptions, ITsDockerConfig } from './interfaces/index.js';
|
import type { IDockerfileOptions, ITsDockerConfig, IBuildCommandOptions } from './interfaces/index.js';
|
||||||
import type { TsDockerManager } from './classes.tsdockermanager.js';
|
import type { TsDockerManager } from './classes.tsdockermanager.js';
|
||||||
|
import * as fs from 'fs';
|
||||||
|
|
||||||
const smartshellInstance = new plugins.smartshell.Smartshell({
|
const smartshellInstance = new plugins.smartshell.Smartshell({
|
||||||
executor: 'bash',
|
executor: 'bash',
|
||||||
@@ -57,9 +58,14 @@ export class Dockerfile {
|
|||||||
const dependencies: Dockerfile[] = [];
|
const dependencies: Dockerfile[] = [];
|
||||||
const baseImage = dockerfile.baseImage;
|
const baseImage = dockerfile.baseImage;
|
||||||
|
|
||||||
|
// Extract repo:version from baseImage for comparison with cleanTag
|
||||||
|
// baseImage may include a registry prefix (e.g., "host.today/repo:version")
|
||||||
|
// but cleanTag is just "repo:version", so we strip the registry prefix
|
||||||
|
const baseImageKey = Dockerfile.extractRepoVersion(baseImage);
|
||||||
|
|
||||||
// Check if the baseImage is among the local Dockerfiles
|
// Check if the baseImage is among the local Dockerfiles
|
||||||
if (tagToDockerfile.has(baseImage)) {
|
if (tagToDockerfile.has(baseImageKey)) {
|
||||||
const baseDockerfile = tagToDockerfile.get(baseImage)!;
|
const baseDockerfile = tagToDockerfile.get(baseImageKey)!;
|
||||||
dependencies.push(baseDockerfile);
|
dependencies.push(baseDockerfile);
|
||||||
dockerfile.localBaseImageDependent = true;
|
dockerfile.localBaseImageDependent = true;
|
||||||
dockerfile.localBaseDockerfile = baseDockerfile;
|
dockerfile.localBaseDockerfile = baseDockerfile;
|
||||||
@@ -115,8 +121,10 @@ export class Dockerfile {
|
|||||||
public static async mapDockerfiles(sortedDockerfileArray: Dockerfile[]): Promise<Dockerfile[]> {
|
public static async mapDockerfiles(sortedDockerfileArray: Dockerfile[]): Promise<Dockerfile[]> {
|
||||||
sortedDockerfileArray.forEach((dockerfileArg) => {
|
sortedDockerfileArray.forEach((dockerfileArg) => {
|
||||||
if (dockerfileArg.localBaseImageDependent) {
|
if (dockerfileArg.localBaseImageDependent) {
|
||||||
|
// Extract repo:version from baseImage for comparison with cleanTag
|
||||||
|
const baseImageKey = Dockerfile.extractRepoVersion(dockerfileArg.baseImage);
|
||||||
sortedDockerfileArray.forEach((dockfile2: Dockerfile) => {
|
sortedDockerfileArray.forEach((dockfile2: Dockerfile) => {
|
||||||
if (dockfile2.cleanTag === dockerfileArg.baseImage) {
|
if (dockfile2.cleanTag === baseImageKey) {
|
||||||
dockerfileArg.localBaseDockerfile = dockfile2;
|
dockerfileArg.localBaseDockerfile = dockfile2;
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
@@ -128,9 +136,25 @@ export class Dockerfile {
|
|||||||
/**
|
/**
|
||||||
* Builds the corresponding real docker image for each Dockerfile class instance
|
* Builds the corresponding real docker image for each Dockerfile class instance
|
||||||
*/
|
*/
|
||||||
public static async buildDockerfiles(sortedArrayArg: Dockerfile[]): Promise<Dockerfile[]> {
|
public static async buildDockerfiles(
|
||||||
|
sortedArrayArg: Dockerfile[],
|
||||||
|
options?: { platform?: string; timeout?: number; noCache?: boolean },
|
||||||
|
): Promise<Dockerfile[]> {
|
||||||
for (const dockerfileArg of sortedArrayArg) {
|
for (const dockerfileArg of sortedArrayArg) {
|
||||||
await dockerfileArg.build();
|
await dockerfileArg.build(options);
|
||||||
|
|
||||||
|
// Tag the built image with the full base image references used by dependent Dockerfiles,
|
||||||
|
// so their FROM lines resolve to the locally-built image instead of pulling from a registry.
|
||||||
|
const dependentBaseImages = new Set<string>();
|
||||||
|
for (const other of sortedArrayArg) {
|
||||||
|
if (other.localBaseDockerfile === dockerfileArg && other.baseImage !== dockerfileArg.buildTag) {
|
||||||
|
dependentBaseImages.add(other.baseImage);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for (const fullTag of dependentBaseImages) {
|
||||||
|
logger.log('info', `Tagging ${dockerfileArg.buildTag} as ${fullTag} for local dependency resolution`);
|
||||||
|
await smartshellInstance.exec(`docker tag ${dockerfileArg.buildTag} ${fullTag}`);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
return sortedArrayArg;
|
return sortedArrayArg;
|
||||||
}
|
}
|
||||||
@@ -230,6 +254,34 @@ export class Dockerfile {
|
|||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Extracts the repo:version part from a full image reference, stripping any registry prefix.
|
||||||
|
* Examples:
|
||||||
|
* "registry.example.com/repo:version" -> "repo:version"
|
||||||
|
* "repo:version" -> "repo:version"
|
||||||
|
* "host.today/ht-docker-node:npmci" -> "ht-docker-node:npmci"
|
||||||
|
*/
|
||||||
|
private static extractRepoVersion(imageRef: string): string {
|
||||||
|
const parts = imageRef.split('/');
|
||||||
|
if (parts.length === 1) {
|
||||||
|
// No registry prefix: "repo:version"
|
||||||
|
return imageRef;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check if first part looks like a registry (contains '.' or ':' or is 'localhost')
|
||||||
|
const firstPart = parts[0];
|
||||||
|
const looksLikeRegistry =
|
||||||
|
firstPart.includes('.') || firstPart.includes(':') || firstPart === 'localhost';
|
||||||
|
|
||||||
|
if (looksLikeRegistry) {
|
||||||
|
// Strip registry: "registry.example.com/repo:version" -> "repo:version"
|
||||||
|
return parts.slice(1).join('/');
|
||||||
|
}
|
||||||
|
|
||||||
|
// No registry prefix, could be "org/repo:version"
|
||||||
|
return imageRef;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Returns the docker tag string for a given registry and repo
|
* Returns the docker tag string for a given registry and repo
|
||||||
*/
|
*/
|
||||||
@@ -314,7 +366,6 @@ export class Dockerfile {
|
|||||||
this.containerName = 'dockerfile-' + this.version;
|
this.containerName = 'dockerfile-' + this.version;
|
||||||
|
|
||||||
if (options.filePath && options.read) {
|
if (options.filePath && options.read) {
|
||||||
const fs = require('fs');
|
|
||||||
this.content = fs.readFileSync(plugins.path.resolve(options.filePath), 'utf-8');
|
this.content = fs.readFileSync(plugins.path.resolve(options.filePath), 'utf-8');
|
||||||
} else if (options.fileContents) {
|
} else if (options.fileContents) {
|
||||||
this.content = options.fileContents;
|
this.content = options.fileContents;
|
||||||
@@ -327,18 +378,23 @@ export class Dockerfile {
|
|||||||
/**
|
/**
|
||||||
* Builds the Dockerfile
|
* Builds the Dockerfile
|
||||||
*/
|
*/
|
||||||
public async build(): Promise<void> {
|
public async build(options?: { platform?: string; timeout?: number; noCache?: boolean }): Promise<void> {
|
||||||
logger.log('info', 'now building Dockerfile for ' + this.cleanTag);
|
logger.log('info', 'now building Dockerfile for ' + this.cleanTag);
|
||||||
const buildArgsString = await Dockerfile.getDockerBuildArgs(this.managerRef);
|
const buildArgsString = await Dockerfile.getDockerBuildArgs(this.managerRef);
|
||||||
const config = this.managerRef.config;
|
const config = this.managerRef.config;
|
||||||
|
const platformOverride = options?.platform;
|
||||||
|
const timeout = options?.timeout;
|
||||||
|
const noCacheFlag = options?.noCache ? ' --no-cache' : '';
|
||||||
|
|
||||||
let buildCommand: string;
|
let buildCommand: string;
|
||||||
|
|
||||||
// Check if multi-platform build is needed
|
if (platformOverride) {
|
||||||
if (config.platforms && config.platforms.length > 1) {
|
// Single platform override via buildx
|
||||||
|
buildCommand = `docker buildx build --platform ${platformOverride}${noCacheFlag} --load -t ${this.buildTag} -f ${this.filePath} ${buildArgsString} .`;
|
||||||
|
} else if (config.platforms && config.platforms.length > 1) {
|
||||||
// Multi-platform build using buildx
|
// Multi-platform build using buildx
|
||||||
const platformString = config.platforms.join(',');
|
const platformString = config.platforms.join(',');
|
||||||
buildCommand = `docker buildx build --platform ${platformString} -t ${this.buildTag} -f ${this.filePath} ${buildArgsString} .`;
|
buildCommand = `docker buildx build --platform ${platformString}${noCacheFlag} -t ${this.buildTag} -f ${this.filePath} ${buildArgsString} .`;
|
||||||
|
|
||||||
if (config.push) {
|
if (config.push) {
|
||||||
buildCommand += ' --push';
|
buildCommand += ' --push';
|
||||||
@@ -348,14 +404,30 @@ export class Dockerfile {
|
|||||||
} else {
|
} else {
|
||||||
// Standard build
|
// Standard build
|
||||||
const versionLabel = this.managerRef.projectInfo?.npm?.version || 'unknown';
|
const versionLabel = this.managerRef.projectInfo?.npm?.version || 'unknown';
|
||||||
buildCommand = `docker build --label="version=${versionLabel}" -t ${this.buildTag} -f ${this.filePath} ${buildArgsString} .`;
|
buildCommand = `docker build --label="version=${versionLabel}"${noCacheFlag} -t ${this.buildTag} -f ${this.filePath} ${buildArgsString} .`;
|
||||||
}
|
}
|
||||||
|
|
||||||
const result = await smartshellInstance.exec(buildCommand);
|
if (timeout) {
|
||||||
if (result.exitCode !== 0) {
|
// Use streaming execution with timeout
|
||||||
logger.log('error', `Build failed for ${this.cleanTag}`);
|
const streaming = await smartshellInstance.execStreaming(buildCommand);
|
||||||
console.log(result.stdout);
|
const timeoutPromise = new Promise<never>((_, reject) => {
|
||||||
throw new Error(`Build failed for ${this.cleanTag}`);
|
setTimeout(() => {
|
||||||
|
streaming.childProcess.kill();
|
||||||
|
reject(new Error(`Build timed out after ${timeout}s for ${this.cleanTag}`));
|
||||||
|
}, timeout * 1000);
|
||||||
|
});
|
||||||
|
const result = await Promise.race([streaming.finalPromise, timeoutPromise]);
|
||||||
|
if (result.exitCode !== 0) {
|
||||||
|
logger.log('error', `Build failed for ${this.cleanTag}`);
|
||||||
|
throw new Error(`Build failed for ${this.cleanTag}`);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
const result = await smartshellInstance.exec(buildCommand);
|
||||||
|
if (result.exitCode !== 0) {
|
||||||
|
logger.log('error', `Build failed for ${this.cleanTag}`);
|
||||||
|
console.log(result.stdout);
|
||||||
|
throw new Error(`Build failed for ${this.cleanTag}`);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
logger.log('ok', `Built ${this.cleanTag}`);
|
logger.log('ok', `Built ${this.cleanTag}`);
|
||||||
@@ -419,7 +491,6 @@ export class Dockerfile {
|
|||||||
const testDir = this.managerRef.config.testDir || plugins.path.join(paths.cwd, 'test');
|
const testDir = this.managerRef.config.testDir || plugins.path.join(paths.cwd, 'test');
|
||||||
const testFile = plugins.path.join(testDir, 'test_' + this.version + '.sh');
|
const testFile = plugins.path.join(testDir, 'test_' + this.version + '.sh');
|
||||||
|
|
||||||
const fs = require('fs');
|
|
||||||
const testFileExists = fs.existsSync(testFile);
|
const testFileExists = fs.existsSync(testFile);
|
||||||
|
|
||||||
if (testFileExists) {
|
if (testFileExists) {
|
||||||
|
|||||||
117
ts/classes.tsdockercache.ts
Normal file
117
ts/classes.tsdockercache.ts
Normal file
@@ -0,0 +1,117 @@
|
|||||||
|
import * as crypto from 'crypto';
|
||||||
|
import * as fs from 'fs';
|
||||||
|
import * as path from 'path';
|
||||||
|
import * as plugins from './tsdocker.plugins.js';
|
||||||
|
import * as paths from './tsdocker.paths.js';
|
||||||
|
import { logger } from './tsdocker.logging.js';
|
||||||
|
import type { ICacheData, ICacheEntry } from './interfaces/index.js';
|
||||||
|
|
||||||
|
const smartshellInstance = new plugins.smartshell.Smartshell({
|
||||||
|
executor: 'bash',
|
||||||
|
});
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Manages content-hash-based build caching for Dockerfiles.
|
||||||
|
* Cache is stored in .nogit/tsdocker_support.json.
|
||||||
|
*/
|
||||||
|
export class TsDockerCache {
|
||||||
|
private cacheFilePath: string;
|
||||||
|
private data: ICacheData;
|
||||||
|
|
||||||
|
constructor() {
|
||||||
|
this.cacheFilePath = path.join(paths.cwd, '.nogit', 'tsdocker_support.json');
|
||||||
|
this.data = { version: 1, entries: {} };
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Loads cache data from disk. Falls back to empty cache on missing/corrupt file.
|
||||||
|
*/
|
||||||
|
public load(): void {
|
||||||
|
try {
|
||||||
|
const raw = fs.readFileSync(this.cacheFilePath, 'utf-8');
|
||||||
|
const parsed = JSON.parse(raw);
|
||||||
|
if (parsed && parsed.version === 1 && parsed.entries) {
|
||||||
|
this.data = parsed;
|
||||||
|
} else {
|
||||||
|
logger.log('warn', '[cache] Cache file has unexpected format, starting fresh');
|
||||||
|
this.data = { version: 1, entries: {} };
|
||||||
|
}
|
||||||
|
} catch {
|
||||||
|
// Missing or corrupt file — start fresh
|
||||||
|
this.data = { version: 1, entries: {} };
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Saves cache data to disk. Creates .nogit directory if needed.
|
||||||
|
*/
|
||||||
|
public save(): void {
|
||||||
|
const dir = path.dirname(this.cacheFilePath);
|
||||||
|
fs.mkdirSync(dir, { recursive: true });
|
||||||
|
fs.writeFileSync(this.cacheFilePath, JSON.stringify(this.data, null, 2), 'utf-8');
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Computes SHA-256 hash of Dockerfile content.
|
||||||
|
*/
|
||||||
|
public computeContentHash(content: string): string {
|
||||||
|
return crypto.createHash('sha256').update(content).digest('hex');
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Checks whether a build can be skipped for the given Dockerfile.
|
||||||
|
* Logs detailed diagnostics and returns true if the build should be skipped.
|
||||||
|
*/
|
||||||
|
public async shouldSkipBuild(cleanTag: string, content: string): Promise<boolean> {
|
||||||
|
const contentHash = this.computeContentHash(content);
|
||||||
|
const entry = this.data.entries[cleanTag];
|
||||||
|
|
||||||
|
if (!entry) {
|
||||||
|
console.log(`[cache] ${cleanTag}:`);
|
||||||
|
console.log(`[cache] Content hash: ${contentHash}`);
|
||||||
|
console.log(`[cache] Stored hash: (none)`);
|
||||||
|
console.log(`[cache] Hash match: no`);
|
||||||
|
console.log(`→ Building ${cleanTag}`);
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
const hashMatch = entry.contentHash === contentHash;
|
||||||
|
console.log(`[cache] ${cleanTag}:`);
|
||||||
|
console.log(`[cache] Content hash: ${contentHash}`);
|
||||||
|
console.log(`[cache] Stored hash: ${entry.contentHash}`);
|
||||||
|
console.log(`[cache] Image ID: ${entry.imageId}`);
|
||||||
|
console.log(`[cache] Hash match: ${hashMatch ? 'yes' : 'no'}`);
|
||||||
|
|
||||||
|
if (!hashMatch) {
|
||||||
|
console.log(`→ Building ${cleanTag}`);
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Hash matches — verify the image still exists locally
|
||||||
|
const inspectResult = await smartshellInstance.exec(
|
||||||
|
`docker image inspect ${entry.imageId} > /dev/null 2>&1`
|
||||||
|
);
|
||||||
|
const available = inspectResult.exitCode === 0;
|
||||||
|
console.log(`[cache] Available: ${available ? 'yes' : 'no'}`);
|
||||||
|
|
||||||
|
if (available) {
|
||||||
|
console.log(`→ Skipping build for ${cleanTag} (cache hit)`);
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
console.log(`→ Building ${cleanTag} (image no longer available)`);
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Records a successful build in the cache.
|
||||||
|
*/
|
||||||
|
public recordBuild(cleanTag: string, content: string, imageId: string, buildTag: string): void {
|
||||||
|
this.data.entries[cleanTag] = {
|
||||||
|
contentHash: this.computeContentHash(content),
|
||||||
|
imageId,
|
||||||
|
buildTag,
|
||||||
|
timestamp: Date.now(),
|
||||||
|
};
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -4,7 +4,8 @@ import { logger } from './tsdocker.logging.js';
|
|||||||
import { Dockerfile } from './classes.dockerfile.js';
|
import { Dockerfile } from './classes.dockerfile.js';
|
||||||
import { DockerRegistry } from './classes.dockerregistry.js';
|
import { DockerRegistry } from './classes.dockerregistry.js';
|
||||||
import { RegistryStorage } from './classes.registrystorage.js';
|
import { RegistryStorage } from './classes.registrystorage.js';
|
||||||
import type { ITsDockerConfig } from './interfaces/index.js';
|
import { TsDockerCache } from './classes.tsdockercache.js';
|
||||||
|
import type { ITsDockerConfig, IBuildCommandOptions } from './interfaces/index.js';
|
||||||
|
|
||||||
const smartshellInstance = new plugins.smartshell.Smartshell({
|
const smartshellInstance = new plugins.smartshell.Smartshell({
|
||||||
executor: 'bash',
|
executor: 'bash',
|
||||||
@@ -90,9 +91,10 @@ export class TsDockerManager {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Builds all discovered Dockerfiles in dependency order
|
* Builds discovered Dockerfiles in dependency order.
|
||||||
|
* When options.patterns is provided, only matching Dockerfiles (and their dependencies) are built.
|
||||||
*/
|
*/
|
||||||
public async build(): Promise<Dockerfile[]> {
|
public async build(options?: IBuildCommandOptions): Promise<Dockerfile[]> {
|
||||||
if (this.dockerfiles.length === 0) {
|
if (this.dockerfiles.length === 0) {
|
||||||
await this.discoverDockerfiles();
|
await this.discoverDockerfiles();
|
||||||
}
|
}
|
||||||
@@ -102,16 +104,107 @@ export class TsDockerManager {
|
|||||||
return [];
|
return [];
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Determine which Dockerfiles to build
|
||||||
|
let toBuild = this.dockerfiles;
|
||||||
|
|
||||||
|
if (options?.patterns && options.patterns.length > 0) {
|
||||||
|
// Filter to matching Dockerfiles
|
||||||
|
const matched = this.dockerfiles.filter((df) => {
|
||||||
|
const basename = plugins.path.basename(df.filePath);
|
||||||
|
return options.patterns!.some((pattern) => {
|
||||||
|
if (pattern.includes('*') || pattern.includes('?')) {
|
||||||
|
// Convert glob pattern to regex
|
||||||
|
const regexStr = '^' + pattern.replace(/\*/g, '.*').replace(/\?/g, '.') + '$';
|
||||||
|
return new RegExp(regexStr).test(basename);
|
||||||
|
}
|
||||||
|
return basename === pattern;
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
if (matched.length === 0) {
|
||||||
|
logger.log('warn', `No Dockerfiles matched patterns: ${options.patterns.join(', ')}`);
|
||||||
|
return [];
|
||||||
|
}
|
||||||
|
|
||||||
|
// Resolve dependency chain and preserve topological order
|
||||||
|
toBuild = this.resolveWithDependencies(matched, this.dockerfiles);
|
||||||
|
logger.log('info', `Matched ${matched.length} Dockerfile(s), building ${toBuild.length} (including dependencies)`);
|
||||||
|
}
|
||||||
|
|
||||||
// Check if buildx is needed
|
// Check if buildx is needed
|
||||||
if (this.config.platforms && this.config.platforms.length > 1) {
|
if (options?.platform || (this.config.platforms && this.config.platforms.length > 1)) {
|
||||||
await this.ensureBuildx();
|
await this.ensureBuildx();
|
||||||
}
|
}
|
||||||
|
|
||||||
logger.log('info', `Building ${this.dockerfiles.length} Dockerfiles...`);
|
logger.log('info', `Building ${toBuild.length} Dockerfiles...`);
|
||||||
await Dockerfile.buildDockerfiles(this.dockerfiles);
|
|
||||||
|
if (options?.cached) {
|
||||||
|
// === CACHED MODE: skip builds for unchanged Dockerfiles ===
|
||||||
|
logger.log('info', '=== CACHED MODE ACTIVE ===');
|
||||||
|
const cache = new TsDockerCache();
|
||||||
|
cache.load();
|
||||||
|
|
||||||
|
for (const dockerfileArg of toBuild) {
|
||||||
|
const skip = await cache.shouldSkipBuild(dockerfileArg.cleanTag, dockerfileArg.content);
|
||||||
|
if (skip) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Cache miss — build this Dockerfile
|
||||||
|
await dockerfileArg.build({
|
||||||
|
platform: options?.platform,
|
||||||
|
timeout: options?.timeout,
|
||||||
|
noCache: options?.noCache,
|
||||||
|
});
|
||||||
|
|
||||||
|
const imageId = await dockerfileArg.getId();
|
||||||
|
cache.recordBuild(dockerfileArg.cleanTag, dockerfileArg.content, imageId, dockerfileArg.buildTag);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Perform dependency tagging for all Dockerfiles (even cache hits, since tags may be stale)
|
||||||
|
for (const dockerfileArg of toBuild) {
|
||||||
|
const dependentBaseImages = new Set<string>();
|
||||||
|
for (const other of toBuild) {
|
||||||
|
if (other.localBaseDockerfile === dockerfileArg && other.baseImage !== dockerfileArg.buildTag) {
|
||||||
|
dependentBaseImages.add(other.baseImage);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for (const fullTag of dependentBaseImages) {
|
||||||
|
logger.log('info', `Tagging ${dockerfileArg.buildTag} as ${fullTag} for local dependency resolution`);
|
||||||
|
await smartshellInstance.exec(`docker tag ${dockerfileArg.buildTag} ${fullTag}`);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
cache.save();
|
||||||
|
} else {
|
||||||
|
// === STANDARD MODE: build all via static helper ===
|
||||||
|
await Dockerfile.buildDockerfiles(toBuild, {
|
||||||
|
platform: options?.platform,
|
||||||
|
timeout: options?.timeout,
|
||||||
|
noCache: options?.noCache,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
logger.log('success', 'All Dockerfiles built successfully');
|
logger.log('success', 'All Dockerfiles built successfully');
|
||||||
|
|
||||||
return this.dockerfiles;
|
return toBuild;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Resolves a set of target Dockerfiles to include all their local base image dependencies,
|
||||||
|
* preserving the original topological build order.
|
||||||
|
*/
|
||||||
|
private resolveWithDependencies(targets: Dockerfile[], allSorted: Dockerfile[]): Dockerfile[] {
|
||||||
|
const needed = new Set<Dockerfile>();
|
||||||
|
const addWithDeps = (df: Dockerfile) => {
|
||||||
|
if (needed.has(df)) return;
|
||||||
|
needed.add(df);
|
||||||
|
if (df.localBaseImageDependent && df.localBaseDockerfile) {
|
||||||
|
addWithDeps(df.localBaseDockerfile);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
for (const df of targets) addWithDeps(df);
|
||||||
|
return allSorted.filter((df) => needed.has(df));
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|||||||
@@ -68,3 +68,26 @@ export interface IPushResult {
|
|||||||
digest?: string;
|
digest?: string;
|
||||||
error?: string;
|
error?: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Options for the build command
|
||||||
|
*/
|
||||||
|
export interface IBuildCommandOptions {
|
||||||
|
patterns?: string[]; // Dockerfile name patterns (e.g., ['Dockerfile_base', 'Dockerfile_*'])
|
||||||
|
platform?: string; // Single platform override (e.g., 'linux/arm64')
|
||||||
|
timeout?: number; // Build timeout in seconds
|
||||||
|
noCache?: boolean; // Force rebuild without Docker layer cache (--no-cache)
|
||||||
|
cached?: boolean; // Skip builds when Dockerfile content hasn't changed
|
||||||
|
}
|
||||||
|
|
||||||
|
export interface ICacheEntry {
|
||||||
|
contentHash: string; // SHA-256 hex of Dockerfile content
|
||||||
|
imageId: string; // Docker image ID (sha256:...)
|
||||||
|
buildTag: string;
|
||||||
|
timestamp: number; // Unix ms
|
||||||
|
}
|
||||||
|
|
||||||
|
export interface ICacheData {
|
||||||
|
version: 1;
|
||||||
|
entries: { [cleanTag: string]: ICacheEntry };
|
||||||
|
}
|
||||||
|
|||||||
@@ -7,8 +7,11 @@ import * as DockerModule from './tsdocker.docker.js';
|
|||||||
|
|
||||||
import { logger, ora } from './tsdocker.logging.js';
|
import { logger, ora } from './tsdocker.logging.js';
|
||||||
import { TsDockerManager } from './classes.tsdockermanager.js';
|
import { TsDockerManager } from './classes.tsdockermanager.js';
|
||||||
|
import type { IBuildCommandOptions } from './interfaces/index.js';
|
||||||
|
import { commitinfo } from './00_commitinfo_data.js';
|
||||||
|
|
||||||
const tsdockerCli = new plugins.smartcli.Smartcli();
|
const tsdockerCli = new plugins.smartcli.Smartcli();
|
||||||
|
tsdockerCli.addVersion(commitinfo.version);
|
||||||
|
|
||||||
export let run = () => {
|
export let run = () => {
|
||||||
// Default command: run tests in container (legacy behavior)
|
// Default command: run tests in container (legacy behavior)
|
||||||
@@ -23,14 +26,34 @@ export let run = () => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Build all Dockerfiles in dependency order
|
* Build Dockerfiles in dependency order
|
||||||
|
* Usage: tsdocker build [Dockerfile_patterns...] [--platform=linux/arm64] [--timeout=600]
|
||||||
*/
|
*/
|
||||||
tsdockerCli.addCommand('build').subscribe(async argvArg => {
|
tsdockerCli.addCommand('build').subscribe(async argvArg => {
|
||||||
try {
|
try {
|
||||||
const config = await ConfigModule.run();
|
const config = await ConfigModule.run();
|
||||||
const manager = new TsDockerManager(config);
|
const manager = new TsDockerManager(config);
|
||||||
await manager.prepare();
|
await manager.prepare();
|
||||||
await manager.build();
|
|
||||||
|
const buildOptions: IBuildCommandOptions = {};
|
||||||
|
const patterns = argvArg._.slice(1) as string[];
|
||||||
|
if (patterns.length > 0) {
|
||||||
|
buildOptions.patterns = patterns;
|
||||||
|
}
|
||||||
|
if (argvArg.platform) {
|
||||||
|
buildOptions.platform = argvArg.platform as string;
|
||||||
|
}
|
||||||
|
if (argvArg.timeout) {
|
||||||
|
buildOptions.timeout = Number(argvArg.timeout);
|
||||||
|
}
|
||||||
|
if (argvArg.cache === false) {
|
||||||
|
buildOptions.noCache = true;
|
||||||
|
}
|
||||||
|
if (argvArg.cached) {
|
||||||
|
buildOptions.cached = true;
|
||||||
|
}
|
||||||
|
|
||||||
|
await manager.build(buildOptions);
|
||||||
logger.log('success', 'Build completed successfully');
|
logger.log('success', 'Build completed successfully');
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
logger.log('error', `Build failed: ${(err as Error).message}`);
|
logger.log('error', `Build failed: ${(err as Error).message}`);
|
||||||
@@ -40,6 +63,7 @@ export let run = () => {
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Push built images to configured registries
|
* Push built images to configured registries
|
||||||
|
* Usage: tsdocker push [Dockerfile_patterns...] [--platform=linux/arm64] [--timeout=600] [--registry=url]
|
||||||
*/
|
*/
|
||||||
tsdockerCli.addCommand('push').subscribe(async argvArg => {
|
tsdockerCli.addCommand('push').subscribe(async argvArg => {
|
||||||
try {
|
try {
|
||||||
@@ -50,11 +74,27 @@ export let run = () => {
|
|||||||
// Login first
|
// Login first
|
||||||
await manager.login();
|
await manager.login();
|
||||||
|
|
||||||
// Build images first (if not already built)
|
// Parse build options from positional args and flags
|
||||||
await manager.build();
|
const buildOptions: IBuildCommandOptions = {};
|
||||||
|
const patterns = argvArg._.slice(1) as string[];
|
||||||
|
if (patterns.length > 0) {
|
||||||
|
buildOptions.patterns = patterns;
|
||||||
|
}
|
||||||
|
if (argvArg.platform) {
|
||||||
|
buildOptions.platform = argvArg.platform as string;
|
||||||
|
}
|
||||||
|
if (argvArg.timeout) {
|
||||||
|
buildOptions.timeout = Number(argvArg.timeout);
|
||||||
|
}
|
||||||
|
if (argvArg.cache === false) {
|
||||||
|
buildOptions.noCache = true;
|
||||||
|
}
|
||||||
|
|
||||||
// Get registry from arguments if specified
|
// Build images first (if not already built)
|
||||||
const registryArg = argvArg._[1]; // e.g., tsdocker push registry.gitlab.com
|
await manager.build(buildOptions);
|
||||||
|
|
||||||
|
// Get registry from --registry flag
|
||||||
|
const registryArg = argvArg.registry as string | undefined;
|
||||||
const registries = registryArg ? [registryArg] : undefined;
|
const registries = registryArg ? [registryArg] : undefined;
|
||||||
|
|
||||||
await manager.push(registries);
|
await manager.push(registries);
|
||||||
@@ -101,7 +141,14 @@ export let run = () => {
|
|||||||
await manager.prepare();
|
await manager.prepare();
|
||||||
|
|
||||||
// Build images first
|
// Build images first
|
||||||
await manager.build();
|
const buildOptions: IBuildCommandOptions = {};
|
||||||
|
if (argvArg.cache === false) {
|
||||||
|
buildOptions.noCache = true;
|
||||||
|
}
|
||||||
|
if (argvArg.cached) {
|
||||||
|
buildOptions.cached = true;
|
||||||
|
}
|
||||||
|
await manager.build(buildOptions);
|
||||||
|
|
||||||
// Run tests
|
// Run tests
|
||||||
await manager.test();
|
await manager.test();
|
||||||
|
|||||||
Reference in New Issue
Block a user