fix(core): update
This commit is contained in:
parent
f13db1e422
commit
da2191bb96
@ -9,27 +9,27 @@
|
|||||||
"githost": "code.foss.global",
|
"githost": "code.foss.global",
|
||||||
"gitscope": "push.rocks",
|
"gitscope": "push.rocks",
|
||||||
"gitrepo": "smartstream",
|
"gitrepo": "smartstream",
|
||||||
"description": "simplifies access to node streams",
|
"description": "A library to simplify the creation and manipulation of Node.js streams, providing utilities for handling transform, duplex, and readable/writable streams effectively in TypeScript.",
|
||||||
"npmPackagename": "@push.rocks/smartstream",
|
"npmPackagename": "@push.rocks/smartstream",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"keywords": [
|
"keywords": [
|
||||||
|
"stream",
|
||||||
"node.js",
|
"node.js",
|
||||||
"streams",
|
"typescript",
|
||||||
"stream manipulation",
|
"stream manipulation",
|
||||||
"pipeline",
|
|
||||||
"data processing",
|
"data processing",
|
||||||
|
"pipeline",
|
||||||
"async transformation",
|
"async transformation",
|
||||||
"event handling",
|
"event handling",
|
||||||
"backpressure management",
|
"backpressure",
|
||||||
"readable streams",
|
"readable stream",
|
||||||
"writable streams",
|
"writable stream",
|
||||||
"duplex streams",
|
"duplex stream",
|
||||||
"transform streams",
|
"transform stream",
|
||||||
"file streaming",
|
"file streaming",
|
||||||
"buffer streams",
|
"buffer",
|
||||||
"stream utilities",
|
"stream utilities",
|
||||||
"stream intake",
|
"esm"
|
||||||
"stream output"
|
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
22
package.json
22
package.json
@ -2,7 +2,7 @@
|
|||||||
"name": "@push.rocks/smartstream",
|
"name": "@push.rocks/smartstream",
|
||||||
"version": "3.0.35",
|
"version": "3.0.35",
|
||||||
"private": false,
|
"private": false,
|
||||||
"description": "simplifies access to node streams",
|
"description": "A library to simplify the creation and manipulation of Node.js streams, providing utilities for handling transform, duplex, and readable/writable streams effectively in TypeScript.",
|
||||||
"main": "dist_ts/index.js",
|
"main": "dist_ts/index.js",
|
||||||
"typings": "dist_ts/index.d.ts",
|
"typings": "dist_ts/index.d.ts",
|
||||||
"type": "module",
|
"type": "module",
|
||||||
@ -51,22 +51,22 @@
|
|||||||
"readme.md"
|
"readme.md"
|
||||||
],
|
],
|
||||||
"keywords": [
|
"keywords": [
|
||||||
|
"stream",
|
||||||
"node.js",
|
"node.js",
|
||||||
"streams",
|
"typescript",
|
||||||
"stream manipulation",
|
"stream manipulation",
|
||||||
"pipeline",
|
|
||||||
"data processing",
|
"data processing",
|
||||||
|
"pipeline",
|
||||||
"async transformation",
|
"async transformation",
|
||||||
"event handling",
|
"event handling",
|
||||||
"backpressure management",
|
"backpressure",
|
||||||
"readable streams",
|
"readable stream",
|
||||||
"writable streams",
|
"writable stream",
|
||||||
"duplex streams",
|
"duplex stream",
|
||||||
"transform streams",
|
"transform stream",
|
||||||
"file streaming",
|
"file streaming",
|
||||||
"buffer streams",
|
"buffer",
|
||||||
"stream utilities",
|
"stream utilities",
|
||||||
"stream intake",
|
"esm"
|
||||||
"stream output"
|
|
||||||
]
|
]
|
||||||
}
|
}
|
@ -1 +1 @@
|
|||||||
|
- make sure to respect backpressure handling.
|
242
readme.md
242
readme.md
@ -1,5 +1,6 @@
|
|||||||
|
```markdown
|
||||||
# @push.rocks/smartstream
|
# @push.rocks/smartstream
|
||||||
simplifies access to node streams
|
A TypeScript library to simplify the creation and manipulation of Node.js streams, providing utilities for transform, duplex, and readable/writable stream handling while managing backpressure effectively.
|
||||||
|
|
||||||
## Install
|
## Install
|
||||||
To install `@push.rocks/smartstream`, you can use npm or yarn as follows:
|
To install `@push.rocks/smartstream`, you can use npm or yarn as follows:
|
||||||
@ -14,7 +15,7 @@ This will add `@push.rocks/smartstream` to your project's dependencies.
|
|||||||
|
|
||||||
## Usage
|
## Usage
|
||||||
|
|
||||||
The `@push.rocks/smartstream` module is designed to simplify working with Node.js streams by providing a set of utilities for creating and manipulating streams. This module makes heavy use of TypeScript for improved code quality, readability, and maintenance. ESM syntax is utilized throughout the examples.
|
The `@push.rocks/smartstream` module is designed to simplify working with Node.js streams by providing a set of utilities for creating and manipulating streams. This module makes extensive use of TypeScript for improved code quality, readability, and maintenance. ESM syntax is utilized throughout the examples.
|
||||||
|
|
||||||
### Importing the Module
|
### Importing the Module
|
||||||
|
|
||||||
@ -24,6 +25,12 @@ Start by importing the module into your TypeScript file:
|
|||||||
import * as smartstream from '@push.rocks/smartstream';
|
import * as smartstream from '@push.rocks/smartstream';
|
||||||
```
|
```
|
||||||
|
|
||||||
|
For a more specific import, you may do the following:
|
||||||
|
|
||||||
|
```typescript
|
||||||
|
import { SmartDuplex, StreamWrapper, StreamIntake, createTransformFunction, createPassThrough } from '@push.rocks/smartstream';
|
||||||
|
```
|
||||||
|
|
||||||
### Creating Basic Transform Streams
|
### Creating Basic Transform Streams
|
||||||
|
|
||||||
The module provides utilities for creating transform streams. For example, to create a transform stream that modifies chunks of data, you can use the `createTransformFunction` utility:
|
The module provides utilities for creating transform streams. For example, to create a transform stream that modifies chunks of data, you can use the `createTransformFunction` utility:
|
||||||
@ -58,7 +65,7 @@ const processDataDuplex = new SmartDuplex({
|
|||||||
sourceStream.pipe(processDataDuplex).pipe(destinationStream);
|
sourceStream.pipe(processDataDuplex).pipe(destinationStream);
|
||||||
```
|
```
|
||||||
|
|
||||||
### Stream Combiners
|
### Combining Multiple Streams
|
||||||
|
|
||||||
`Smartstream` facilitates easy combining of multiple streams into a single pipeline, handling errors and cleanup automatically. Here's how you can combine multiple streams:
|
`Smartstream` facilitates easy combining of multiple streams into a single pipeline, handling errors and cleanup automatically. Here's how you can combine multiple streams:
|
||||||
|
|
||||||
@ -101,7 +108,7 @@ Consider a scenario where you need to process a large CSV file, transform the da
|
|||||||
```typescript
|
```typescript
|
||||||
import { SmartDuplex, createTransformFunction } from '@push.rocks/smartstream';
|
import { SmartDuplex, createTransformFunction } from '@push.rocks/smartstream';
|
||||||
import fs from 'fs';
|
import fs from 'fs';
|
||||||
import csvParser from 'csv-parser'; // Assume this is a CSV parsing library
|
import csvParser from 'csv-parser';
|
||||||
|
|
||||||
const csvReadTransform = createTransformFunction<any, any>(async (row) => {
|
const csvReadTransform = createTransformFunction<any, any>(async (row) => {
|
||||||
// Process row
|
// Process row
|
||||||
@ -121,11 +128,232 @@ fs.createReadStream('path/to/largeFile.csv')
|
|||||||
|
|
||||||
This example demonstrates reading a large CSV file, transforming each row with `createTransformFunction`, and using a `SmartDuplex` to manage the processed data flow efficiently, ensuring no data is lost due to backpressure issues.
|
This example demonstrates reading a large CSV file, transforming each row with `createTransformFunction`, and using a `SmartDuplex` to manage the processed data flow efficiently, ensuring no data is lost due to backpressure issues.
|
||||||
|
|
||||||
### Conclusion
|
### Advanced Use Case: Backpressure Handling
|
||||||
|
|
||||||
`@push.rocks/smartstream` offers a robust set of tools for working with Node.js streams, providing a more intuitive and reliable way to create, manipulate, and combine streams. By leveraging TypeScript and ESM syntax, `smartstream` enables developers to build more maintainable and type-safe stream-based solutions.
|
Effective backpressure handling is crucial when working with streams to avoid overwhelming the downstream consumers. Here’s a comprehensive example that demonstrates handling backpressure in a pipeline with multiple `SmartDuplex` instances:
|
||||||
|
|
||||||
|
```typescript
|
||||||
|
import { SmartDuplex } from '@push.rocks/smartstream';
|
||||||
|
|
||||||
|
// Define the first SmartDuplex, which writes data slowly to simulate backpressure
|
||||||
|
const slowProcessingStream = new SmartDuplex({
|
||||||
|
name: 'SlowProcessor',
|
||||||
|
objectMode: true,
|
||||||
|
writeFunction: async (chunk, { push }) => {
|
||||||
|
await new Promise(resolve => setTimeout(resolve, 100)); // Simulated delay
|
||||||
|
console.log('Processed chunk:', chunk);
|
||||||
|
push(chunk);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
// Define the second SmartDuplex as a fast processor
|
||||||
|
const fastProcessingStream = new SmartDuplex({
|
||||||
|
name: 'FastProcessor',
|
||||||
|
objectMode: true,
|
||||||
|
writeFunction: async (chunk, { push }) => {
|
||||||
|
console.log('Fast processing chunk:', chunk);
|
||||||
|
push(chunk);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
// Create a StreamIntake to dynamically handle incoming data
|
||||||
|
const streamIntake = new StreamIntake<string>();
|
||||||
|
|
||||||
|
// Chain the streams together and handle the backpressure scenario
|
||||||
|
streamIntake
|
||||||
|
.pipe(fastProcessingStream)
|
||||||
|
.pipe(slowProcessingStream)
|
||||||
|
.pipe(createPassThrough()) // Use Pass-Through to provide intermediary handling
|
||||||
|
.on('data', data => console.log('Final output:', data))
|
||||||
|
.on('error', error => console.error('Stream encountered an error:', error));
|
||||||
|
|
||||||
|
// Simulate data pushing with intervals to observe backpressure handling
|
||||||
|
let counter = 0;
|
||||||
|
const interval = setInterval(() => {
|
||||||
|
if (counter >= 10) {
|
||||||
|
streamIntake.signalEnd();
|
||||||
|
clearInterval(interval);
|
||||||
|
} else {
|
||||||
|
streamIntake.pushData(`Chunk ${counter}`);
|
||||||
|
counter++;
|
||||||
|
}
|
||||||
|
}, 50);
|
||||||
|
```
|
||||||
|
|
||||||
|
In this advanced use case, a `SlowProcessor` and `FastProcessor` are created using `SmartDuplex`, simulating a situation where one stream is slower than another. The `StreamIntake` dynamically handles incoming chunks of data and the intermediary Pass-Through handles any potential interruptions.
|
||||||
|
|
||||||
|
### Transform Streams in Parallel
|
||||||
|
|
||||||
|
For scenarios where you need to process data in parallel:
|
||||||
|
|
||||||
|
```typescript
|
||||||
|
import { SmartDuplex, createTransformFunction } from '@push.rocks/smartstream';
|
||||||
|
|
||||||
|
const parallelTransform = createTransformFunction<any, any>(async (chunk) => {
|
||||||
|
// Parallel Processing
|
||||||
|
const results = await Promise.all(chunk.map(async item => await processItem(item)));
|
||||||
|
return results;
|
||||||
|
});
|
||||||
|
|
||||||
|
const streamIntake = new StreamIntake<any[]>();
|
||||||
|
|
||||||
|
streamIntake
|
||||||
|
.pipe(parallelTransform)
|
||||||
|
.pipe(new SmartDuplex({
|
||||||
|
async writeFunction(chunk, { push }) {
|
||||||
|
console.log('Processed parallel chunk:', chunk);
|
||||||
|
push(chunk);
|
||||||
|
}
|
||||||
|
}))
|
||||||
|
.on('finish', () => console.log('Parallel processing completed.'));
|
||||||
|
|
||||||
|
// Simulate data pushing
|
||||||
|
streamIntake.pushData([1, 2, 3, 4]);
|
||||||
|
streamIntake.pushData([5, 6, 7, 8]);
|
||||||
|
streamIntake.signalEnd();
|
||||||
|
```
|
||||||
|
|
||||||
|
### Error Handling in Stream Pipelines
|
||||||
|
|
||||||
|
Error handling is an essential part of working with streams. The `StreamWrapper` assists in combining multiple streams while managing errors seamlessly:
|
||||||
|
|
||||||
|
```typescript
|
||||||
|
import { StreamWrapper } from '@push.rocks/smartstream';
|
||||||
|
|
||||||
|
const faultyStream = new SmartDuplex({
|
||||||
|
async writeFunction(chunk, { push }) {
|
||||||
|
if (chunk === 'bad data') {
|
||||||
|
throw new Error('Faulty data encountered');
|
||||||
|
}
|
||||||
|
push(chunk);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
const readStream = new StreamIntake<string>();
|
||||||
|
const writeStream = new SmartDuplex({
|
||||||
|
async writeFunction(chunk) {
|
||||||
|
console.log('Written chunk:', chunk);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
const combinedStream = new StreamWrapper([readStream, faultyStream, writeStream]);
|
||||||
|
|
||||||
|
combinedStream.run()
|
||||||
|
.then(() => console.log('Stream processing completed.'))
|
||||||
|
.catch(err => console.error('Stream error:', err.message));
|
||||||
|
|
||||||
|
// Push Data
|
||||||
|
readStream.pushData('good data');
|
||||||
|
readStream.pushData('bad data'); // This will throw an error
|
||||||
|
readStream.pushData('more good data');
|
||||||
|
readStream.signalEnd();
|
||||||
|
```
|
||||||
|
|
||||||
|
### Testing Streams
|
||||||
|
|
||||||
|
Here's an example test case using the `tap` testing framework to verify the integrity of the `SmartDuplex` from a buffer:
|
||||||
|
|
||||||
|
```typescript
|
||||||
|
import { expect, tap } from '@push.rocks/tapbundle';
|
||||||
|
import { SmartDuplex } from '@push.rocks/smartstream';
|
||||||
|
|
||||||
|
tap.test('should create a SmartStream from a Buffer', async () => {
|
||||||
|
const bufferData = Buffer.from('This is a test buffer');
|
||||||
|
const smartStream = SmartDuplex.fromBuffer(bufferData, {});
|
||||||
|
|
||||||
|
let receivedData = Buffer.alloc(0);
|
||||||
|
|
||||||
|
return new Promise<void>((resolve) => {
|
||||||
|
smartStream.on('data', (chunk: Buffer) => {
|
||||||
|
receivedData = Buffer.concat([receivedData, chunk]);
|
||||||
|
});
|
||||||
|
|
||||||
|
smartStream.on('end', () => {
|
||||||
|
expect(receivedData.toString()).toEqual(bufferData.toString());
|
||||||
|
resolve();
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
tap.start();
|
||||||
|
```
|
||||||
|
|
||||||
|
### Working with Files and Buffers
|
||||||
|
|
||||||
|
You can easily stream files and buffers with `smartstream`. Here’s a test illustrating reading and writing with file streams using `smartfile` combined with `smartstream` utilities:
|
||||||
|
|
||||||
|
```typescript
|
||||||
|
import { tap } from '@push.rocks/tapbundle';
|
||||||
|
import * as smartfile from '@push.rocks/smartfile';
|
||||||
|
import { SmartDuplex, StreamWrapper } from '@push.rocks/smartstream';
|
||||||
|
|
||||||
|
tap.test('should handle file read and write streams', async () => {
|
||||||
|
const readStream = smartfile.fsStream.createReadStream('./test/assets/readabletext.txt');
|
||||||
|
const writeStream = smartfile.fsStream.createWriteStream('./test/assets/writabletext.txt');
|
||||||
|
|
||||||
|
const transformStream = new SmartDuplex({
|
||||||
|
async writeFunction(chunk, { push }) {
|
||||||
|
const transformedChunk = chunk.toString().toUpperCase();
|
||||||
|
push(transformedChunk);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
const streamWrapper = new StreamWrapper([readStream, transformStream, writeStream]);
|
||||||
|
|
||||||
|
await streamWrapper.run();
|
||||||
|
|
||||||
|
const outputContent = await smartfile.fs.promises.readFile('./test/assets/writabletext.txt', 'utf-8');
|
||||||
|
console.log('Output Content:', outputContent);
|
||||||
|
});
|
||||||
|
|
||||||
|
tap.start();
|
||||||
|
```
|
||||||
|
|
||||||
|
### Modular and Scoped Transformations
|
||||||
|
|
||||||
|
Creating modular and scoped transformations is straightforward with `SmartDuplex`:
|
||||||
|
|
||||||
|
```typescript
|
||||||
|
import { SmartDuplex } from '@push.rocks/smartstream';
|
||||||
|
|
||||||
|
type DataChunk = {
|
||||||
|
id: number;
|
||||||
|
data: string;
|
||||||
|
};
|
||||||
|
|
||||||
|
const transformationStream1 = new SmartDuplex<DataChunk, DataChunk>({
|
||||||
|
async writeFunction(chunk, { push }) {
|
||||||
|
chunk.data = chunk.data.toUpperCase();
|
||||||
|
push(chunk);
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
const transformationStream2 = new SmartDuplex<DataChunk, DataChunk>({
|
||||||
|
async writeFunction(chunk, { push }) {
|
||||||
|
chunk.data = `${chunk.data} processed with transformation 2`;
|
||||||
|
push(chunk);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
const initialData: DataChunk[] = [
|
||||||
|
{ id: 1, data: 'first' },
|
||||||
|
{ id: 2, data: 'second' }
|
||||||
|
];
|
||||||
|
|
||||||
|
const intakeStream = new StreamIntake<DataChunk>();
|
||||||
|
|
||||||
|
intakeStream
|
||||||
|
.pipe(transformationStream1)
|
||||||
|
.pipe(transformationStream2)
|
||||||
|
.on('data', data => console.log('Transformed Data:', data));
|
||||||
|
|
||||||
|
initialData.forEach(item => intakeStream.pushData(item));
|
||||||
|
intakeStream.signalEnd();
|
||||||
|
```
|
||||||
|
|
||||||
|
By leveraging `SmartDuplex`, `StreamWrapper`, and `StreamIntake`, you can streamline and enhance your data transformation pipelines in Node.js with a clear, efficient, and backpressure-friendly approach.
|
||||||
|
```
|
||||||
|
|
||||||
For more detailed examples and documentation, visit the [GitLab Repository](https://gitlab.com/push.rocks/smartstream) or the [GitHub Mirror](https://github.com/pushrocks/smartstream).
|
|
||||||
|
|
||||||
## License and Legal Information
|
## License and Legal Information
|
||||||
|
|
||||||
|
@ -3,6 +3,6 @@
|
|||||||
*/
|
*/
|
||||||
export const commitinfo = {
|
export const commitinfo = {
|
||||||
name: '@push.rocks/smartstream',
|
name: '@push.rocks/smartstream',
|
||||||
version: '3.0.35',
|
version: '3.0.36',
|
||||||
description: 'simplifies access to node streams'
|
description: 'A library to simplify the creation and manipulation of Node.js streams, providing utilities for handling transform, duplex, and readable/writable streams effectively in TypeScript.'
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user