fix(plugins): Migrate filesystem usage to Node fs/fsPromises and upgrade smartfile to v13; add listFileTree helper and update tests
This commit is contained in:
@@ -1,7 +1,33 @@
|
||||
import * as path from 'path';
|
||||
import * as path from 'node:path';
|
||||
import * as fs from 'node:fs';
|
||||
import * as fsPromises from 'node:fs/promises';
|
||||
import * as smartpath from '@push.rocks/smartpath';
|
||||
import * as smartfile from '@push.rocks/smartfile';
|
||||
import * as smartrequest from '@push.rocks/smartrequest';
|
||||
import * as smartstream from '@push.rocks/smartstream';
|
||||
|
||||
export { path, smartpath, smartfile, smartrequest, smartstream };
|
||||
export { path, fs, fsPromises, smartpath, smartfile, smartrequest, smartstream };
|
||||
|
||||
/**
|
||||
* List files in a directory recursively, returning relative paths
|
||||
*/
|
||||
export async function listFileTree(dirPath: string, _pattern: string = '**/*'): Promise<string[]> {
|
||||
const results: string[] = [];
|
||||
|
||||
async function walkDir(currentPath: string, relativePath: string = '') {
|
||||
const entries = await fsPromises.readdir(currentPath, { withFileTypes: true });
|
||||
for (const entry of entries) {
|
||||
const entryRelPath = relativePath ? path.join(relativePath, entry.name) : entry.name;
|
||||
const entryFullPath = path.join(currentPath, entry.name);
|
||||
|
||||
if (entry.isDirectory()) {
|
||||
await walkDir(entryFullPath, entryRelPath);
|
||||
} else if (entry.isFile()) {
|
||||
results.push(entryRelPath);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
await walkDir(dirPath);
|
||||
return results;
|
||||
}
|
||||
|
||||
@@ -14,7 +14,7 @@ const testPaths = {
|
||||
};
|
||||
|
||||
tap.preTask('should prepare test directories', async () => {
|
||||
await plugins.smartfile.fs.ensureDir(testPaths.gzipTestDir);
|
||||
await plugins.fsPromises.mkdir(testPaths.gzipTestDir, { recursive: true });
|
||||
});
|
||||
|
||||
tap.test('should create and extract a gzip file', async () => {
|
||||
@@ -24,23 +24,17 @@ tap.test('should create and extract a gzip file', async () => {
|
||||
const gzipFileName = 'test-file.txt.gz';
|
||||
|
||||
// Write the original file
|
||||
await plugins.smartfile.memory.toFs(
|
||||
testContent,
|
||||
plugins.path.join(testPaths.gzipTestDir, testFileName)
|
||||
await plugins.fsPromises.writeFile(
|
||||
plugins.path.join(testPaths.gzipTestDir, testFileName),
|
||||
testContent
|
||||
);
|
||||
|
||||
// Compress the file using gzip
|
||||
const originalFile = await plugins.smartfile.fs.fileTreeToObject(
|
||||
testPaths.gzipTestDir,
|
||||
testFileName
|
||||
);
|
||||
|
||||
|
||||
// Create gzip compressed version using fflate directly
|
||||
const fflate = await import('fflate');
|
||||
const compressed = fflate.gzipSync(Buffer.from(testContent));
|
||||
await plugins.smartfile.memory.toFs(
|
||||
Buffer.from(compressed),
|
||||
plugins.path.join(testPaths.gzipTestDir, gzipFileName)
|
||||
await plugins.fsPromises.writeFile(
|
||||
plugins.path.join(testPaths.gzipTestDir, gzipFileName),
|
||||
Buffer.from(compressed)
|
||||
);
|
||||
|
||||
// Now test extraction using SmartArchive
|
||||
@@ -50,13 +44,14 @@ tap.test('should create and extract a gzip file', async () => {
|
||||
|
||||
// Export to a new location
|
||||
const extractPath = plugins.path.join(testPaths.gzipTestDir, 'extracted');
|
||||
await plugins.smartfile.fs.ensureDir(extractPath);
|
||||
await plugins.fsPromises.mkdir(extractPath, { recursive: true });
|
||||
// Provide a filename since gzip doesn't contain filename metadata
|
||||
await gzipArchive.exportToFs(extractPath, 'test-file.txt');
|
||||
|
||||
|
||||
// Read the extracted file
|
||||
const extractedContent = await plugins.smartfile.fs.toStringSync(
|
||||
plugins.path.join(extractPath, 'test-file.txt')
|
||||
const extractedContent = await plugins.fsPromises.readFile(
|
||||
plugins.path.join(extractPath, 'test-file.txt'),
|
||||
'utf8'
|
||||
);
|
||||
|
||||
// Verify the content matches
|
||||
@@ -71,13 +66,13 @@ tap.test('should handle gzip stream extraction', async () => {
|
||||
// Create gzip compressed version
|
||||
const fflate = await import('fflate');
|
||||
const compressed = fflate.gzipSync(Buffer.from(testContent));
|
||||
await plugins.smartfile.memory.toFs(
|
||||
Buffer.from(compressed),
|
||||
plugins.path.join(testPaths.gzipTestDir, gzipFileName)
|
||||
await plugins.fsPromises.writeFile(
|
||||
plugins.path.join(testPaths.gzipTestDir, gzipFileName),
|
||||
Buffer.from(compressed)
|
||||
);
|
||||
|
||||
|
||||
// Create a read stream for the gzip file
|
||||
const gzipStream = plugins.smartfile.fsStream.createReadStream(
|
||||
const gzipStream = plugins.fs.createReadStream(
|
||||
plugins.path.join(testPaths.gzipTestDir, gzipFileName)
|
||||
);
|
||||
|
||||
@@ -121,7 +116,7 @@ tap.test('should handle gzip files with original filename in header', async () =
|
||||
const gzipFileName = 'compressed.gz';
|
||||
|
||||
// Create a proper gzip with filename header using Node's zlib
|
||||
const zlib = await import('zlib');
|
||||
const zlib = await import('node:zlib');
|
||||
const gzipBuffer = await new Promise<Buffer>((resolve, reject) => {
|
||||
zlib.gzip(Buffer.from(testContent), {
|
||||
level: 9,
|
||||
@@ -133,29 +128,30 @@ tap.test('should handle gzip files with original filename in header', async () =
|
||||
});
|
||||
});
|
||||
|
||||
await plugins.smartfile.memory.toFs(
|
||||
gzipBuffer,
|
||||
plugins.path.join(testPaths.gzipTestDir, gzipFileName)
|
||||
await plugins.fsPromises.writeFile(
|
||||
plugins.path.join(testPaths.gzipTestDir, gzipFileName),
|
||||
gzipBuffer
|
||||
);
|
||||
|
||||
|
||||
// Test extraction
|
||||
const gzipArchive = await smartarchive.SmartArchive.fromArchiveFile(
|
||||
plugins.path.join(testPaths.gzipTestDir, gzipFileName)
|
||||
);
|
||||
|
||||
|
||||
const extractPath = plugins.path.join(testPaths.gzipTestDir, 'header-test');
|
||||
await plugins.smartfile.fs.ensureDir(extractPath);
|
||||
await plugins.fsPromises.mkdir(extractPath, { recursive: true });
|
||||
// Provide a filename since gzip doesn't reliably contain filename metadata
|
||||
await gzipArchive.exportToFs(extractPath, 'compressed.txt');
|
||||
|
||||
|
||||
// Check if file was extracted (name might be derived from archive name)
|
||||
const files = await plugins.smartfile.fs.listFileTree(extractPath, '**/*');
|
||||
const files = await plugins.listFileTree(extractPath, '**/*');
|
||||
expect(files.length).toBeGreaterThan(0);
|
||||
|
||||
|
||||
// Read and verify content
|
||||
const extractedFile = files[0];
|
||||
const extractedContent = await plugins.smartfile.fs.toStringSync(
|
||||
plugins.path.join(extractPath, extractedFile || 'compressed.txt')
|
||||
const extractedContent = await plugins.fsPromises.readFile(
|
||||
plugins.path.join(extractPath, extractedFile || 'compressed.txt'),
|
||||
'utf8'
|
||||
);
|
||||
expect(extractedContent).toEqual(testContent);
|
||||
});
|
||||
@@ -168,27 +164,28 @@ tap.test('should handle large gzip files', async () => {
|
||||
// Compress the large file
|
||||
const fflate = await import('fflate');
|
||||
const compressed = fflate.gzipSync(Buffer.from(largeContent));
|
||||
await plugins.smartfile.memory.toFs(
|
||||
Buffer.from(compressed),
|
||||
plugins.path.join(testPaths.gzipTestDir, gzipFileName)
|
||||
await plugins.fsPromises.writeFile(
|
||||
plugins.path.join(testPaths.gzipTestDir, gzipFileName),
|
||||
Buffer.from(compressed)
|
||||
);
|
||||
|
||||
|
||||
// Test extraction
|
||||
const gzipArchive = await smartarchive.SmartArchive.fromArchiveFile(
|
||||
plugins.path.join(testPaths.gzipTestDir, gzipFileName)
|
||||
);
|
||||
|
||||
|
||||
const extractPath = plugins.path.join(testPaths.gzipTestDir, 'large-extracted');
|
||||
await plugins.smartfile.fs.ensureDir(extractPath);
|
||||
await plugins.fsPromises.mkdir(extractPath, { recursive: true });
|
||||
// Provide a filename since gzip doesn't contain filename metadata
|
||||
await gzipArchive.exportToFs(extractPath, 'large-file.txt');
|
||||
|
||||
|
||||
// Verify the extracted content
|
||||
const files = await plugins.smartfile.fs.listFileTree(extractPath, '**/*');
|
||||
const files = await plugins.listFileTree(extractPath, '**/*');
|
||||
expect(files.length).toBeGreaterThan(0);
|
||||
|
||||
const extractedContent = await plugins.smartfile.fs.toStringSync(
|
||||
plugins.path.join(extractPath, files[0] || 'large-file.txt')
|
||||
|
||||
const extractedContent = await plugins.fsPromises.readFile(
|
||||
plugins.path.join(extractPath, files[0] || 'large-file.txt'),
|
||||
'utf8'
|
||||
);
|
||||
expect(extractedContent.length).toEqual(largeContent.length);
|
||||
expect(extractedContent).toEqual(largeContent);
|
||||
@@ -200,60 +197,64 @@ tap.test('should handle real-world multi-chunk gzip from URL', async () => {
|
||||
|
||||
// Download and extract the archive
|
||||
const testArchive = await smartarchive.SmartArchive.fromArchiveUrl(testUrl);
|
||||
|
||||
|
||||
const extractPath = plugins.path.join(testPaths.gzipTestDir, 'real-world-test');
|
||||
await plugins.smartfile.fs.ensureDir(extractPath);
|
||||
|
||||
await plugins.fsPromises.mkdir(extractPath, { recursive: true });
|
||||
|
||||
// This will test multi-chunk decompression as the file is larger
|
||||
await testArchive.exportToFs(extractPath);
|
||||
|
||||
|
||||
// Verify extraction worked
|
||||
const files = await plugins.smartfile.fs.listFileTree(extractPath, '**/*');
|
||||
const files = await plugins.listFileTree(extractPath, '**/*');
|
||||
expect(files.length).toBeGreaterThan(0);
|
||||
|
||||
|
||||
// Check for expected package structure
|
||||
const hasPackageJson = files.some(f => f.includes('package.json'));
|
||||
expect(hasPackageJson).toBeTrue();
|
||||
|
||||
|
||||
// Read and verify package.json content
|
||||
const packageJsonPath = files.find(f => f.includes('package.json'));
|
||||
if (packageJsonPath) {
|
||||
const packageJsonContent = await plugins.smartfile.fs.toStringSync(
|
||||
plugins.path.join(extractPath, packageJsonPath)
|
||||
const packageJsonContent = await plugins.fsPromises.readFile(
|
||||
plugins.path.join(extractPath, packageJsonPath),
|
||||
'utf8'
|
||||
);
|
||||
const packageJson = JSON.parse(packageJsonContent);
|
||||
expect(packageJson.name).toEqual('@push.rocks/smartfile');
|
||||
expect(packageJson.version).toEqual('11.2.7');
|
||||
}
|
||||
|
||||
|
||||
// Read and verify a TypeScript file
|
||||
const tsFilePath = files.find(f => f.endsWith('.ts'));
|
||||
if (tsFilePath) {
|
||||
const tsFileContent = await plugins.smartfile.fs.toStringSync(
|
||||
plugins.path.join(extractPath, tsFilePath)
|
||||
const tsFileContent = await plugins.fsPromises.readFile(
|
||||
plugins.path.join(extractPath, tsFilePath),
|
||||
'utf8'
|
||||
);
|
||||
// TypeScript files should have content
|
||||
expect(tsFileContent.length).toBeGreaterThan(10);
|
||||
console.log(` ✓ TypeScript file ${tsFilePath} has ${tsFileContent.length} bytes`);
|
||||
}
|
||||
|
||||
|
||||
// Read and verify license file
|
||||
const licensePath = files.find(f => f.includes('license'));
|
||||
if (licensePath) {
|
||||
const licenseContent = await plugins.smartfile.fs.toStringSync(
|
||||
plugins.path.join(extractPath, licensePath)
|
||||
const licenseContent = await plugins.fsPromises.readFile(
|
||||
plugins.path.join(extractPath, licensePath),
|
||||
'utf8'
|
||||
);
|
||||
expect(licenseContent).toContain('MIT');
|
||||
}
|
||||
|
||||
|
||||
// Verify we can read multiple files without corruption
|
||||
const readableFiles = files.filter(f =>
|
||||
const readableFiles = files.filter(f =>
|
||||
f.endsWith('.json') || f.endsWith('.md') || f.endsWith('.ts') || f.endsWith('.js')
|
||||
).slice(0, 5); // Test first 5 readable files
|
||||
|
||||
|
||||
for (const file of readableFiles) {
|
||||
const content = await plugins.smartfile.fs.toStringSync(
|
||||
plugins.path.join(extractPath, file)
|
||||
const content = await plugins.fsPromises.readFile(
|
||||
plugins.path.join(extractPath, file),
|
||||
'utf8'
|
||||
);
|
||||
expect(content).toBeDefined();
|
||||
expect(content.length).toBeGreaterThan(0);
|
||||
@@ -270,7 +271,7 @@ tap.test('should handle gzip extraction fully in memory', async () => {
|
||||
const compressed = fflate.gzipSync(Buffer.from(testContent));
|
||||
|
||||
// Create a stream from the compressed data
|
||||
const { Readable } = await import('stream');
|
||||
const { Readable } = await import('node:stream');
|
||||
const compressedStream = Readable.from(Buffer.from(compressed));
|
||||
|
||||
// Process through SmartArchive without touching filesystem
|
||||
@@ -318,7 +319,7 @@ tap.test('should handle real tgz file fully in memory', async (tools) => {
|
||||
console.log(` Downloaded ${tgzBuffer.length} bytes into memory`);
|
||||
|
||||
// Create stream from buffer
|
||||
const { Readable: Readable2 } = await import('stream');
|
||||
const { Readable: Readable2 } = await import('node:stream');
|
||||
const tgzStream = Readable2.from(tgzBuffer);
|
||||
|
||||
// Process through SmartArchive in memory
|
||||
@@ -16,7 +16,7 @@ const testPaths = {
|
||||
import * as smartarchive from '../ts/index.js';
|
||||
|
||||
tap.preTask('should prepare .nogit dir', async () => {
|
||||
await plugins.smartfile.fs.ensureDir(testPaths.remoteDir);
|
||||
await plugins.fsPromises.mkdir(testPaths.remoteDir, { recursive: true });
|
||||
});
|
||||
|
||||
tap.preTask('should prepare downloads', async (tools) => {
|
||||
@@ -26,9 +26,9 @@ tap.preTask('should prepare downloads', async (tools) => {
|
||||
)
|
||||
.get();
|
||||
const downloadedFile: Buffer = Buffer.from(await response.arrayBuffer());
|
||||
await plugins.smartfile.memory.toFs(
|
||||
downloadedFile,
|
||||
await plugins.fsPromises.writeFile(
|
||||
plugins.path.join(testPaths.nogitDir, 'test.tgz'),
|
||||
downloadedFile,
|
||||
);
|
||||
});
|
||||
|
||||
Reference in New Issue
Block a user