Compare commits
63 Commits
Author | SHA1 | Date | |
---|---|---|---|
3d13cb76f6 | |||
9e3fd28c4a | |||
673f5c86fb | |||
a225188e24 | |||
4fc82d0dc6 | |||
3d58a01b29 | |||
f7e9636bf6 | |||
f211cc8ddd | |||
60c8824f33 | |||
40e8e06ff1 | |||
30f2facd59 | |||
ddb7d4af03 | |||
22d93b4c07 | |||
e138bca39d | |||
6a2ef1b152 | |||
7b1d2199e9 | |||
04c22f73df | |||
c8dc791c83 | |||
9c30e5bab1 | |||
5f2c5f9380 | |||
f9b8bf33b0 | |||
a55b2548d7 | |||
c8465b82be | |||
b593e3a32c | |||
a490f521ab | |||
59027782dc | |||
8c7dd7970c | |||
22d18dc21f | |||
1cb6f727af | |||
824c44d165 | |||
3e062103f8 | |||
6451e93c12 | |||
70cf93595c | |||
17e03e9790 | |||
e52ce7af61 | |||
f548f4b6cb | |||
23a7a77a73 | |||
13d2fc78b8 | |||
898cc0407d | |||
8a3f43a11a | |||
da2191bb96 | |||
f13db1e422 | |||
42a90e804a | |||
413e2af717 | |||
267a76af13 | |||
7834b7e6d2 | |||
ae643708e7 | |||
d9d96b8bb7 | |||
a961eea431 | |||
edb58ade28 | |||
753a481765 | |||
bbbd1b73b9 | |||
271d0be106 | |||
0ceeacd5a0 | |||
287695e445 | |||
60f9e541a5 | |||
96ea67e135 | |||
ba0a2023ad | |||
a09c359847 | |||
e2b4d772b3 | |||
0f46b62b2d | |||
9bf37469c6 | |||
12bb125bdc |
140
.gitlab-ci.yml
140
.gitlab-ci.yml
@ -1,140 +0,0 @@
|
||||
# gitzone ci_default
|
||||
image: registry.gitlab.com/hosttoday/ht-docker-node:npmci
|
||||
|
||||
cache:
|
||||
paths:
|
||||
- .npmci_cache/
|
||||
key: '$CI_BUILD_STAGE'
|
||||
|
||||
stages:
|
||||
- security
|
||||
- test
|
||||
- release
|
||||
- metadata
|
||||
|
||||
before_script:
|
||||
- npm install -g @shipzone/npmci
|
||||
|
||||
# ====================
|
||||
# security stage
|
||||
# ====================
|
||||
mirror:
|
||||
stage: security
|
||||
script:
|
||||
- npmci git mirror
|
||||
only:
|
||||
- tags
|
||||
tags:
|
||||
- lossless
|
||||
- docker
|
||||
- notpriv
|
||||
|
||||
auditProductionDependencies:
|
||||
image: registry.gitlab.com/hosttoday/ht-docker-node:npmci
|
||||
stage: security
|
||||
script:
|
||||
- npmci npm prepare
|
||||
- npmci command npm install --production --ignore-scripts
|
||||
- npmci command npm config set registry https://registry.npmjs.org
|
||||
- npmci command npm audit --audit-level=high --only=prod --production
|
||||
tags:
|
||||
- docker
|
||||
allow_failure: true
|
||||
|
||||
auditDevDependencies:
|
||||
image: registry.gitlab.com/hosttoday/ht-docker-node:npmci
|
||||
stage: security
|
||||
script:
|
||||
- npmci npm prepare
|
||||
- npmci command npm install --ignore-scripts
|
||||
- npmci command npm config set registry https://registry.npmjs.org
|
||||
- npmci command npm audit --audit-level=high --only=dev
|
||||
tags:
|
||||
- docker
|
||||
allow_failure: true
|
||||
|
||||
# ====================
|
||||
# test stage
|
||||
# ====================
|
||||
|
||||
testStable:
|
||||
stage: test
|
||||
script:
|
||||
- npmci npm prepare
|
||||
- npmci node install stable
|
||||
- npmci npm install
|
||||
- npmci npm test
|
||||
coverage: /\d+.?\d+?\%\s*coverage/
|
||||
tags:
|
||||
- docker
|
||||
|
||||
testBuild:
|
||||
stage: test
|
||||
script:
|
||||
- npmci npm prepare
|
||||
- npmci node install stable
|
||||
- npmci npm install
|
||||
- npmci command npm run build
|
||||
coverage: /\d+.?\d+?\%\s*coverage/
|
||||
tags:
|
||||
- docker
|
||||
|
||||
release:
|
||||
stage: release
|
||||
script:
|
||||
- npmci node install stable
|
||||
- npmci npm publish
|
||||
only:
|
||||
- tags
|
||||
tags:
|
||||
- lossless
|
||||
- docker
|
||||
- notpriv
|
||||
|
||||
# ====================
|
||||
# metadata stage
|
||||
# ====================
|
||||
codequality:
|
||||
stage: metadata
|
||||
allow_failure: true
|
||||
only:
|
||||
- tags
|
||||
script:
|
||||
- npmci command npm install -g typescript
|
||||
- npmci npm prepare
|
||||
- npmci npm install
|
||||
tags:
|
||||
- lossless
|
||||
- docker
|
||||
- priv
|
||||
|
||||
trigger:
|
||||
stage: metadata
|
||||
script:
|
||||
- npmci trigger
|
||||
only:
|
||||
- tags
|
||||
tags:
|
||||
- lossless
|
||||
- docker
|
||||
- notpriv
|
||||
|
||||
pages:
|
||||
stage: metadata
|
||||
script:
|
||||
- npmci node install lts
|
||||
- npmci command npm install -g @git.zone/tsdoc
|
||||
- npmci npm prepare
|
||||
- npmci npm install
|
||||
- npmci command tsdoc
|
||||
tags:
|
||||
- lossless
|
||||
- docker
|
||||
- notpriv
|
||||
only:
|
||||
- tags
|
||||
artifacts:
|
||||
expire_in: 1 week
|
||||
paths:
|
||||
- public
|
||||
allow_failure: true
|
95
changelog.md
Normal file
95
changelog.md
Normal file
@ -0,0 +1,95 @@
|
||||
# Changelog
|
||||
|
||||
## 2024-11-19 - 3.2.5 - fix(nodewebhelpers)
|
||||
Fix import and use correct module structure for Node.js streams in smartstream.nodewebhelpers.ts
|
||||
|
||||
- Corrected the import statement for the fs module.
|
||||
- Updated the use of the fs.createReadStream method.
|
||||
|
||||
## 2024-10-16 - 3.2.4 - fix(SmartDuplex)
|
||||
Fix stream termination when reading from a web readable stream
|
||||
|
||||
- Resolved an issue in SmartDuplex where the stream did not properly terminate after reaching the end of a web readable stream.
|
||||
|
||||
## 2024-10-16 - 3.2.3 - fix(smartduplex)
|
||||
Enhance documentation for read function in SmartDuplex
|
||||
|
||||
- Added inline comments to clarify the behavior and importance of unlocking the reader in the readFunction of SmartDuplex.fromWebReadableStream.
|
||||
|
||||
## 2024-10-16 - 3.2.2 - fix(SmartDuplex)
|
||||
Fix issue with SmartDuplex fromWebReadableStream method
|
||||
|
||||
- Resolved a potential unhandled promise rejection in fromWebReadableStream method
|
||||
- Ensured proper release of stream reader lock in case of read completion
|
||||
|
||||
## 2024-10-16 - 3.2.1 - fix(core)
|
||||
Fix the order of operations in SmartDuplex _read method to ensure proper waiting for items.
|
||||
|
||||
- Adjusted the order of reading function execution and waiting for items in the SmartDuplex _read method.
|
||||
- Fixed potential issues with stream data processing timing.
|
||||
|
||||
## 2024-10-16 - 3.2.0 - feat(SmartDuplex)
|
||||
Added method to create SmartDuplex from a WebReadableStream.
|
||||
|
||||
- Implemented a static method in SmartDuplex to allow creating an instance from a WebReadableStream.
|
||||
- This addition enhances the capability of SmartDuplex to integrate with web streams, facilitating seamless stream manipulation across environments.
|
||||
|
||||
## 2024-10-14 - 3.1.2 - fix(WebDuplexStream)
|
||||
Fix variable naming inconsistency in WebDuplexStream test
|
||||
|
||||
- Changed variable names from 'transformStream' to 'webDuplexStream' for consistency.
|
||||
- Renamed 'writableStream' and 'readableStream' to 'writer' and 'reader' respectively.
|
||||
|
||||
## 2024-10-13 - 3.1.1 - fix(WebDuplexStream)
|
||||
Improved read/write interface and error handling in WebDuplexStream
|
||||
|
||||
- Enhanced the IStreamToolsRead and IStreamToolsWrite interfaces for better Promise handling
|
||||
- Refined readFunction and writeFunction handling to accommodate asynchronous data processing and error propagation
|
||||
- Added internal _startReading method to facilitate initial data handling if readFunction is present
|
||||
- Maintained backward compatibility while ensuring data continuity when no writeFunction is specified
|
||||
|
||||
## 2024-10-13 - 3.1.0 - feat(core)
|
||||
Add support for creating Web ReadableStream from a file
|
||||
|
||||
- Introduced a new helper function `createWebReadableStreamFromFile` that allows for creating a Web ReadableStream from a file path.
|
||||
- Updated exports in `ts/index.ts` to include `nodewebhelpers` which provides the new web stream feature.
|
||||
|
||||
## 2024-10-13 - 3.0.46 - fix(WebDuplexStream)
|
||||
Fix errors in WebDuplexStream transformation and test logic
|
||||
|
||||
- Corrected async handling in WebDuplexStream write function
|
||||
- Fixed `WebDuplexStream` tests to properly handle asynchronous reading and writing
|
||||
|
||||
## 2024-10-13 - 3.0.45 - fix(ts)
|
||||
Fixed formatting issues in SmartDuplex class
|
||||
|
||||
- Resolved inconsistent spacing in SmartDuplex class methods and constructor.
|
||||
- Ensured consistent formatting in the getWebStreams method.
|
||||
|
||||
## 2024-06-02 - 3.0.39 - smartduplex
|
||||
Add .getWebStreams method
|
||||
|
||||
- Introduced a new `.getWebStreams` method in the smartduplex module, providing compatibility with the web streams API.
|
||||
|
||||
## 2024-03-16 - 3.0.34 - configuration
|
||||
Update project configuration files
|
||||
|
||||
- Updated `tsconfig` for optimization.
|
||||
- Modified `npmextra.json` to set the `githost` attribute.
|
||||
|
||||
## 2023-11-03 - 3.0.0 to 3.0.8 - core
|
||||
Transition to major version 3.x
|
||||
|
||||
- Implemented breaking changes in the core system for better performance and feature set.
|
||||
- Continuous core updates to improve stability and performance across minor version increments.
|
||||
|
||||
## 2023-11-02 - 2.0.4 to 2.0.8 - core
|
||||
Core updates and a major fix
|
||||
|
||||
- Implemented core updates addressing minor bugs and enhancements.
|
||||
- A significant breaking change update transitioning from 2.0.x to 3.0.0.
|
||||
|
||||
## 2022-03-31 - 2.0.0 - core
|
||||
Major esm transition
|
||||
|
||||
- Implemented a breaking change by switching the core to ESM (ECMAScript Module) format for modernized module handling.
|
@ -6,12 +6,34 @@
|
||||
"gitzone": {
|
||||
"projectType": "npm",
|
||||
"module": {
|
||||
"githost": "gitlab.com",
|
||||
"githost": "code.foss.global",
|
||||
"gitscope": "push.rocks",
|
||||
"gitrepo": "smartstream",
|
||||
"description": "simplifies access to node streams",
|
||||
"description": "A library to simplify the creation and manipulation of Node.js streams, providing utilities for handling transform, duplex, and readable/writable streams effectively in TypeScript.",
|
||||
"npmPackagename": "@push.rocks/smartstream",
|
||||
"license": "MIT"
|
||||
"license": "MIT",
|
||||
"keywords": [
|
||||
"stream",
|
||||
"node.js",
|
||||
"typescript",
|
||||
"stream manipulation",
|
||||
"data processing",
|
||||
"pipeline",
|
||||
"async transformation",
|
||||
"event handling",
|
||||
"backpressure",
|
||||
"readable stream",
|
||||
"writable stream",
|
||||
"duplex stream",
|
||||
"transform stream",
|
||||
"file streaming",
|
||||
"buffer",
|
||||
"stream utilities",
|
||||
"esm"
|
||||
]
|
||||
}
|
||||
},
|
||||
"tsdoc": {
|
||||
"legal": "\n## License and Legal Information\n\nThis repository contains open-source code that is licensed under the MIT License. A copy of the MIT License can be found in the [license](license) file within this repository. \n\n**Please note:** The MIT License does not grant permission to use the trade names, trademarks, service marks, or product names of the project, except as required for reasonable and customary use in describing the origin of the work and reproducing the content of the NOTICE file.\n\n### Trademarks\n\nThis project is owned and maintained by Task Venture Capital GmbH. The names and logos associated with Task Venture Capital GmbH and any related products or services are trademarks of Task Venture Capital GmbH and are not included within the scope of the MIT license granted herein. Use of these trademarks must comply with Task Venture Capital GmbH's Trademark Guidelines, and any usage must be approved in writing by Task Venture Capital GmbH.\n\n### Company Information\n\nTask Venture Capital GmbH \nRegistered at District court Bremen HRB 35230 HB, Germany\n\nFor any legal inquiries or if you require further information, please contact us via email at hello@task.vc.\n\nBy using this repository, you acknowledge that you have read this section, agree to comply with its terms, and understand that the licensing of the code does not imply endorsement by Task Venture Capital GmbH of any derivative works.\n"
|
||||
}
|
||||
}
|
49
package.json
49
package.json
@ -1,36 +1,38 @@
|
||||
{
|
||||
"name": "@push.rocks/smartstream",
|
||||
"version": "3.0.26",
|
||||
"version": "3.2.5",
|
||||
"private": false,
|
||||
"description": "simplifies access to node streams",
|
||||
"main": "dist_ts/index.js",
|
||||
"typings": "dist_ts/index.d.ts",
|
||||
"description": "A library to simplify the creation and manipulation of Node.js streams, providing utilities for handling transform, duplex, and readable/writable streams effectively in TypeScript.",
|
||||
"type": "module",
|
||||
"exports": {
|
||||
".": "./dist_ts/index.js",
|
||||
"./web": "./dist_ts_web/index.js"
|
||||
},
|
||||
"scripts": {
|
||||
"test": "(tstest test/)",
|
||||
"build": "(tsbuild)",
|
||||
"buildDocs": "tsdoc"
|
||||
"build": "(tsbuild tsfolders --web --allowimplicitany)"
|
||||
},
|
||||
"repository": {
|
||||
"type": "git",
|
||||
"url": "git+https://gitlab.com/push.rocks/smartstream.git"
|
||||
"url": "https://code.foss.global/push.rocks/smartstream.git"
|
||||
},
|
||||
"author": "Lossless GmbH",
|
||||
"license": "MIT",
|
||||
"bugs": {
|
||||
"url": "https://gitlab.com/push.rocks/smartstream/issues"
|
||||
},
|
||||
"homepage": "https://gitlab.com/push.rocks/smartstream#readme",
|
||||
"homepage": "https://code.foss.global/push.rocks/smartstream",
|
||||
"devDependencies": {
|
||||
"@git.zone/tsbuild": "^2.1.66",
|
||||
"@git.zone/tsbuild": "^2.1.80",
|
||||
"@git.zone/tsrun": "^1.2.44",
|
||||
"@git.zone/tstest": "^1.0.84",
|
||||
"@push.rocks/smartfile": "^11.0.0",
|
||||
"@push.rocks/tapbundle": "^5.0.15",
|
||||
"@types/node": "^20.9.0"
|
||||
"@git.zone/tstest": "^1.0.90",
|
||||
"@push.rocks/smartfile": "^11.0.15",
|
||||
"@push.rocks/tapbundle": "^5.0.23",
|
||||
"@types/node": "^20.12.12"
|
||||
},
|
||||
"dependencies": {
|
||||
"@push.rocks/lik": "^6.0.12",
|
||||
"@push.rocks/lik": "^6.0.15",
|
||||
"@push.rocks/smartenv": "^5.0.12",
|
||||
"@push.rocks/smartpromise": "^4.0.3",
|
||||
"@push.rocks/smartrx": "^3.0.7"
|
||||
},
|
||||
@ -48,5 +50,24 @@
|
||||
"cli.js",
|
||||
"npmextra.json",
|
||||
"readme.md"
|
||||
],
|
||||
"keywords": [
|
||||
"stream",
|
||||
"node.js",
|
||||
"typescript",
|
||||
"stream manipulation",
|
||||
"data processing",
|
||||
"pipeline",
|
||||
"async transformation",
|
||||
"event handling",
|
||||
"backpressure",
|
||||
"readable stream",
|
||||
"writable stream",
|
||||
"duplex stream",
|
||||
"transform stream",
|
||||
"file streaming",
|
||||
"buffer",
|
||||
"stream utilities",
|
||||
"esm"
|
||||
]
|
||||
}
|
||||
|
8706
pnpm-lock.yaml
generated
8706
pnpm-lock.yaml
generated
File diff suppressed because it is too large
Load Diff
1
readme.hints.md
Normal file
1
readme.hints.md
Normal file
@ -0,0 +1 @@
|
||||
- make sure to respect backpressure handling.
|
401
readme.md
401
readme.md
@ -1,54 +1,375 @@
|
||||
# @pushrocks/smartstream
|
||||
simplifies access to node streams
|
||||
```markdown
|
||||
# @push.rocks/smartstream
|
||||
A TypeScript library to simplify the creation and manipulation of Node.js streams, providing utilities for transform, duplex, and readable/writable stream handling while managing backpressure effectively.
|
||||
|
||||
## Availabililty and Links
|
||||
* [npmjs.org (npm package)](https://www.npmjs.com/package/@pushrocks/smartstream)
|
||||
* [gitlab.com (source)](https://gitlab.com/pushrocks/smartstream)
|
||||
* [github.com (source mirror)](https://github.com/pushrocks/smartstream)
|
||||
* [docs (typedoc)](https://pushrocks.gitlab.io/smartstream/)
|
||||
## Install
|
||||
To install `@push.rocks/smartstream`, you can use npm or yarn as follows:
|
||||
|
||||
## Status for master
|
||||
```bash
|
||||
npm install @push.rocks/smartstream --save
|
||||
# OR
|
||||
yarn add @push.rocks/smartstream
|
||||
```
|
||||
|
||||
Status Category | Status Badge
|
||||
-- | --
|
||||
GitLab Pipelines | [](https://lossless.cloud)
|
||||
GitLab Pipline Test Coverage | [](https://lossless.cloud)
|
||||
npm | [](https://lossless.cloud)
|
||||
Snyk | [](https://lossless.cloud)
|
||||
TypeScript Support | [](https://lossless.cloud)
|
||||
node Support | [](https://nodejs.org/dist/latest-v10.x/docs/api/)
|
||||
Code Style | [](https://lossless.cloud)
|
||||
PackagePhobia (total standalone install weight) | [](https://lossless.cloud)
|
||||
PackagePhobia (package size on registry) | [](https://lossless.cloud)
|
||||
BundlePhobia (total size when bundled) | [](https://lossless.cloud)
|
||||
Platform support | [](https://lossless.cloud) [](https://lossless.cloud)
|
||||
This will add `@push.rocks/smartstream` to your project's dependencies.
|
||||
|
||||
## Usage
|
||||
|
||||
Use TypeScript for best in class instellisense.
|
||||
The `@push.rocks/smartstream` module is designed to simplify working with Node.js streams by providing a set of utilities for creating and manipulating streams. This module makes extensive use of TypeScript for improved code quality, readability, and maintenance. ESM syntax is utilized throughout the examples.
|
||||
|
||||
### Importing the Module
|
||||
|
||||
Start by importing the module into your TypeScript file:
|
||||
|
||||
```typescript
|
||||
import { Smartstream } from 'smartstream'
|
||||
import * as gUglify from 'gulp-uglify'
|
||||
|
||||
let mySmartstream = new Smartstream([
|
||||
gulp.src(['./file1.js','./file2.js']),
|
||||
gUglify(),
|
||||
gulp.dest('./some/output/path')
|
||||
])
|
||||
|
||||
mySmartstream.onError((err) => { /* handle error */ }) // handles all errors in stream
|
||||
myStream.onCustomEvent('myeventname', (args...) => { /* Do something */ }) // emit an custom event anywhere in your stream
|
||||
mySmartstream.run().then(() => {/* do something when stream is finished */})
|
||||
import * as smartstream from '@push.rocks/smartstream';
|
||||
```
|
||||
|
||||
## Contribution
|
||||
For a more specific import, you may do the following:
|
||||
|
||||
We are always happy for code contributions. If you are not the code contributing type that is ok. Still, maintaining Open Source repositories takes considerable time and thought. If you like the quality of what we do and our modules are useful to you we would appreciate a little monthly contribution: You can [contribute one time](https://lossless.link/contribute-onetime) or [contribute monthly](https://lossless.link/contribute). :)
|
||||
```typescript
|
||||
import { SmartDuplex, StreamWrapper, StreamIntake, createTransformFunction, createPassThrough } from '@push.rocks/smartstream';
|
||||
```
|
||||
|
||||
For further information read the linked docs at the top of this readme.
|
||||
### Creating Basic Transform Streams
|
||||
|
||||
> MIT licensed | **©** [Lossless GmbH](https://lossless.gmbh)
|
||||
| By using this npm module you agree to our [privacy policy](https://lossless.gmbH/privacy)
|
||||
The module provides utilities for creating transform streams. For example, to create a transform stream that modifies chunks of data, you can use the `createTransformFunction` utility:
|
||||
|
||||
[](https://maintainedby.lossless.com)
|
||||
```typescript
|
||||
import { createTransformFunction } from '@push.rocks/smartstream';
|
||||
|
||||
const upperCaseTransform = createTransformFunction<string, string>(async (chunk) => {
|
||||
return chunk.toUpperCase();
|
||||
});
|
||||
|
||||
// Usage with pipe
|
||||
readableStream
|
||||
.pipe(upperCaseTransform)
|
||||
.pipe(writableStream);
|
||||
```
|
||||
|
||||
### Handling Backpressure with SmartDuplex
|
||||
|
||||
`SmartDuplex` is a powerful part of the `smartstream` module designed to handle backpressure effectively. Here's an example of how to create a `SmartDuplex` stream that processes data and respects the consumer's pace:
|
||||
|
||||
```typescript
|
||||
import { SmartDuplex } from '@push.rocks/smartstream';
|
||||
|
||||
const processDataDuplex = new SmartDuplex({
|
||||
async writeFunction(chunk, { push }) {
|
||||
const processedChunk = await processChunk(chunk); // Assume this is a defined asynchronous function
|
||||
push(processedChunk);
|
||||
}
|
||||
});
|
||||
|
||||
sourceStream.pipe(processDataDuplex).pipe(destinationStream);
|
||||
```
|
||||
|
||||
### Combining Multiple Streams
|
||||
|
||||
`Smartstream` facilitates easy combining of multiple streams into a single pipeline, handling errors and cleanup automatically. Here's how you can combine multiple streams:
|
||||
|
||||
```typescript
|
||||
import { StreamWrapper } from '@push.rocks/smartstream';
|
||||
|
||||
const combinedStream = new StreamWrapper([
|
||||
readStream, // Source stream
|
||||
transformStream1, // Transformation
|
||||
transformStream2, // Another transformation
|
||||
writeStream // Destination stream
|
||||
]);
|
||||
|
||||
combinedStream.run()
|
||||
.then(() => console.log('Processing completed.'))
|
||||
.catch(err => console.error('An error occurred:', err));
|
||||
```
|
||||
|
||||
### Working with StreamIntake
|
||||
|
||||
`StreamIntake` allows for more dynamic control of the reading process, facilitating scenarios where data is not continuously available:
|
||||
|
||||
```typescript
|
||||
import { StreamIntake } from '@push.rocks/smartstream';
|
||||
|
||||
const streamIntake = new StreamIntake<string>();
|
||||
|
||||
// Dynamically push data into the intake
|
||||
streamIntake.pushData('Hello, World!');
|
||||
streamIntake.pushData('Another message');
|
||||
|
||||
// Signal end when no more data is to be pushed
|
||||
streamIntake.signalEnd();
|
||||
```
|
||||
|
||||
### Real-world Scenario: Processing Large Files
|
||||
|
||||
Consider a scenario where you need to process a large CSV file, transform the data row-by-row, and then write the results to a database or another file. With `smartstream`, you could create a pipe that reads the CSV, processes each row, and handles backpressure, ensuring efficient use of resources.
|
||||
|
||||
```typescript
|
||||
import { SmartDuplex, createTransformFunction } from '@push.rocks/smartstream';
|
||||
import fs from 'fs';
|
||||
import csvParser from 'csv-parser';
|
||||
|
||||
const csvReadTransform = createTransformFunction<any, any>(async (row) => {
|
||||
// Process row
|
||||
return processedRow;
|
||||
});
|
||||
|
||||
fs.createReadStream('path/to/largeFile.csv')
|
||||
.pipe(csvParser())
|
||||
.pipe(csvReadTransform)
|
||||
.pipe(new SmartDuplex({
|
||||
async writeFunction(chunk, { push }) {
|
||||
await writeToDatabase(chunk); // Assume this writes to a database
|
||||
}
|
||||
}))
|
||||
.on('finish', () => console.log('File processed successfully.'));
|
||||
```
|
||||
|
||||
This example demonstrates reading a large CSV file, transforming each row with `createTransformFunction`, and using a `SmartDuplex` to manage the processed data flow efficiently, ensuring no data is lost due to backpressure issues.
|
||||
|
||||
### Advanced Use Case: Backpressure Handling
|
||||
|
||||
Effective backpressure handling is crucial when working with streams to avoid overwhelming the downstream consumers. Here’s a comprehensive example that demonstrates handling backpressure in a pipeline with multiple `SmartDuplex` instances:
|
||||
|
||||
```typescript
|
||||
import { SmartDuplex } from '@push.rocks/smartstream';
|
||||
|
||||
// Define the first SmartDuplex, which writes data slowly to simulate backpressure
|
||||
const slowProcessingStream = new SmartDuplex({
|
||||
name: 'SlowProcessor',
|
||||
objectMode: true,
|
||||
writeFunction: async (chunk, { push }) => {
|
||||
await new Promise(resolve => setTimeout(resolve, 100)); // Simulated delay
|
||||
console.log('Processed chunk:', chunk);
|
||||
push(chunk);
|
||||
}
|
||||
});
|
||||
|
||||
// Define the second SmartDuplex as a fast processor
|
||||
const fastProcessingStream = new SmartDuplex({
|
||||
name: 'FastProcessor',
|
||||
objectMode: true,
|
||||
writeFunction: async (chunk, { push }) => {
|
||||
console.log('Fast processing chunk:', chunk);
|
||||
push(chunk);
|
||||
}
|
||||
});
|
||||
|
||||
// Create a StreamIntake to dynamically handle incoming data
|
||||
const streamIntake = new StreamIntake<string>();
|
||||
|
||||
// Chain the streams together and handle the backpressure scenario
|
||||
streamIntake
|
||||
.pipe(fastProcessingStream)
|
||||
.pipe(slowProcessingStream)
|
||||
.pipe(createPassThrough()) // Use Pass-Through to provide intermediary handling
|
||||
.on('data', data => console.log('Final output:', data))
|
||||
.on('error', error => console.error('Stream encountered an error:', error));
|
||||
|
||||
// Simulate data pushing with intervals to observe backpressure handling
|
||||
let counter = 0;
|
||||
const interval = setInterval(() => {
|
||||
if (counter >= 10) {
|
||||
streamIntake.signalEnd();
|
||||
clearInterval(interval);
|
||||
} else {
|
||||
streamIntake.pushData(`Chunk ${counter}`);
|
||||
counter++;
|
||||
}
|
||||
}, 50);
|
||||
```
|
||||
|
||||
In this advanced use case, a `SlowProcessor` and `FastProcessor` are created using `SmartDuplex`, simulating a situation where one stream is slower than another. The `StreamIntake` dynamically handles incoming chunks of data and the intermediary Pass-Through handles any potential interruptions.
|
||||
|
||||
### Transform Streams in Parallel
|
||||
|
||||
For scenarios where you need to process data in parallel:
|
||||
|
||||
```typescript
|
||||
import { SmartDuplex, createTransformFunction } from '@push.rocks/smartstream';
|
||||
|
||||
const parallelTransform = createTransformFunction<any, any>(async (chunk) => {
|
||||
// Parallel Processing
|
||||
const results = await Promise.all(chunk.map(async item => await processItem(item)));
|
||||
return results;
|
||||
});
|
||||
|
||||
const streamIntake = new StreamIntake<any[]>();
|
||||
|
||||
streamIntake
|
||||
.pipe(parallelTransform)
|
||||
.pipe(new SmartDuplex({
|
||||
async writeFunction(chunk, { push }) {
|
||||
console.log('Processed parallel chunk:', chunk);
|
||||
push(chunk);
|
||||
}
|
||||
}))
|
||||
.on('finish', () => console.log('Parallel processing completed.'));
|
||||
|
||||
// Simulate data pushing
|
||||
streamIntake.pushData([1, 2, 3, 4]);
|
||||
streamIntake.pushData([5, 6, 7, 8]);
|
||||
streamIntake.signalEnd();
|
||||
```
|
||||
|
||||
### Error Handling in Stream Pipelines
|
||||
|
||||
Error handling is an essential part of working with streams. The `StreamWrapper` assists in combining multiple streams while managing errors seamlessly:
|
||||
|
||||
```typescript
|
||||
import { StreamWrapper } from '@push.rocks/smartstream';
|
||||
|
||||
const faultyStream = new SmartDuplex({
|
||||
async writeFunction(chunk, { push }) {
|
||||
if (chunk === 'bad data') {
|
||||
throw new Error('Faulty data encountered');
|
||||
}
|
||||
push(chunk);
|
||||
}
|
||||
});
|
||||
|
||||
const readStream = new StreamIntake<string>();
|
||||
const writeStream = new SmartDuplex({
|
||||
async writeFunction(chunk) {
|
||||
console.log('Written chunk:', chunk);
|
||||
}
|
||||
});
|
||||
|
||||
const combinedStream = new StreamWrapper([readStream, faultyStream, writeStream]);
|
||||
|
||||
combinedStream.run()
|
||||
.then(() => console.log('Stream processing completed.'))
|
||||
.catch(err => console.error('Stream error:', err.message));
|
||||
|
||||
// Push Data
|
||||
readStream.pushData('good data');
|
||||
readStream.pushData('bad data'); // This will throw an error
|
||||
readStream.pushData('more good data');
|
||||
readStream.signalEnd();
|
||||
```
|
||||
|
||||
### Testing Streams
|
||||
|
||||
Here's an example test case using the `tap` testing framework to verify the integrity of the `SmartDuplex` from a buffer:
|
||||
|
||||
```typescript
|
||||
import { expect, tap } from '@push.rocks/tapbundle';
|
||||
import { SmartDuplex } from '@push.rocks/smartstream';
|
||||
|
||||
tap.test('should create a SmartStream from a Buffer', async () => {
|
||||
const bufferData = Buffer.from('This is a test buffer');
|
||||
const smartStream = SmartDuplex.fromBuffer(bufferData, {});
|
||||
|
||||
let receivedData = Buffer.alloc(0);
|
||||
|
||||
return new Promise<void>((resolve) => {
|
||||
smartStream.on('data', (chunk: Buffer) => {
|
||||
receivedData = Buffer.concat([receivedData, chunk]);
|
||||
});
|
||||
|
||||
smartStream.on('end', () => {
|
||||
expect(receivedData.toString()).toEqual(bufferData.toString());
|
||||
resolve();
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
tap.start();
|
||||
```
|
||||
|
||||
### Working with Files and Buffers
|
||||
|
||||
You can easily stream files and buffers with `smartstream`. Here’s a test illustrating reading and writing with file streams using `smartfile` combined with `smartstream` utilities:
|
||||
|
||||
```typescript
|
||||
import { tap } from '@push.rocks/tapbundle';
|
||||
import * as smartfile from '@push.rocks/smartfile';
|
||||
import { SmartDuplex, StreamWrapper } from '@push.rocks/smartstream';
|
||||
|
||||
tap.test('should handle file read and write streams', async () => {
|
||||
const readStream = smartfile.fsStream.createReadStream('./test/assets/readabletext.txt');
|
||||
const writeStream = smartfile.fsStream.createWriteStream('./test/assets/writabletext.txt');
|
||||
|
||||
const transformStream = new SmartDuplex({
|
||||
async writeFunction(chunk, { push }) {
|
||||
const transformedChunk = chunk.toString().toUpperCase();
|
||||
push(transformedChunk);
|
||||
}
|
||||
});
|
||||
|
||||
const streamWrapper = new StreamWrapper([readStream, transformStream, writeStream]);
|
||||
|
||||
await streamWrapper.run();
|
||||
|
||||
const outputContent = await smartfile.fs.promises.readFile('./test/assets/writabletext.txt', 'utf-8');
|
||||
console.log('Output Content:', outputContent);
|
||||
});
|
||||
|
||||
tap.start();
|
||||
```
|
||||
|
||||
### Modular and Scoped Transformations
|
||||
|
||||
Creating modular and scoped transformations is straightforward with `SmartDuplex`:
|
||||
|
||||
```typescript
|
||||
import { SmartDuplex } from '@push.rocks/smartstream';
|
||||
|
||||
type DataChunk = {
|
||||
id: number;
|
||||
data: string;
|
||||
};
|
||||
|
||||
const transformationStream1 = new SmartDuplex<DataChunk, DataChunk>({
|
||||
async writeFunction(chunk, { push }) {
|
||||
chunk.data = chunk.data.toUpperCase();
|
||||
push(chunk);
|
||||
}
|
||||
})
|
||||
|
||||
const transformationStream2 = new SmartDuplex<DataChunk, DataChunk>({
|
||||
async writeFunction(chunk, { push }) {
|
||||
chunk.data = `${chunk.data} processed with transformation 2`;
|
||||
push(chunk);
|
||||
}
|
||||
});
|
||||
|
||||
const initialData: DataChunk[] = [
|
||||
{ id: 1, data: 'first' },
|
||||
{ id: 2, data: 'second' }
|
||||
];
|
||||
|
||||
const intakeStream = new StreamIntake<DataChunk>();
|
||||
|
||||
intakeStream
|
||||
.pipe(transformationStream1)
|
||||
.pipe(transformationStream2)
|
||||
.on('data', data => console.log('Transformed Data:', data));
|
||||
|
||||
initialData.forEach(item => intakeStream.pushData(item));
|
||||
intakeStream.signalEnd();
|
||||
```
|
||||
|
||||
By leveraging `SmartDuplex`, `StreamWrapper`, and `StreamIntake`, you can streamline and enhance your data transformation pipelines in Node.js with a clear, efficient, and backpressure-friendly approach.
|
||||
```
|
||||
|
||||
|
||||
## License and Legal Information
|
||||
|
||||
This repository contains open-source code that is licensed under the MIT License. A copy of the MIT License can be found in the [license](license) file within this repository.
|
||||
|
||||
**Please note:** The MIT License does not grant permission to use the trade names, trademarks, service marks, or product names of the project, except as required for reasonable and customary use in describing the origin of the work and reproducing the content of the NOTICE file.
|
||||
|
||||
### Trademarks
|
||||
|
||||
This project is owned and maintained by Task Venture Capital GmbH. The names and logos associated with Task Venture Capital GmbH and any related products or services are trademarks of Task Venture Capital GmbH and are not included within the scope of the MIT license granted herein. Use of these trademarks must comply with Task Venture Capital GmbH's Trademark Guidelines, and any usage must be approved in writing by Task Venture Capital GmbH.
|
||||
|
||||
### Company Information
|
||||
|
||||
Task Venture Capital GmbH
|
||||
Registered at District court Bremen HRB 35230 HB, Germany
|
||||
|
||||
For any legal inquiries or if you require further information, please contact us via email at hello@task.vc.
|
||||
|
||||
By using this repository, you acknowledge that you have read this section, agree to comply with its terms, and understand that the licensing of the code does not imply endorsement by Task Venture Capital GmbH of any derivative works.
|
||||
|
@ -19,15 +19,15 @@ tap.test('should run backpressure test', async (toolsArg) => {
|
||||
writeFunction: async (chunk, tools) => {
|
||||
await new Promise((resolve) => setTimeout(resolve, 20)); // Slow processing
|
||||
console.log(`processed chunk ${chunk} in stream 2`);
|
||||
// await tools.push(chunk);
|
||||
return chunk;
|
||||
await tools.push(chunk);
|
||||
// return chunk, optionally return ;
|
||||
},
|
||||
}); // This stream processes data more slowly
|
||||
const stream3 = new SmartDuplex({
|
||||
objectMode: true,
|
||||
name: 'stream3',
|
||||
writeFunction: async (chunk, tools) => {
|
||||
await new Promise((resolve) => setTimeout(resolve, 100)); // Slow processing
|
||||
await new Promise((resolve) => setTimeout(resolve, 200)); // Slow processing
|
||||
console.log(`processed chunk ${chunk} in stream 3`);
|
||||
},
|
||||
});
|
67
test/test.ts_web.both.ts
Normal file
67
test/test.ts_web.both.ts
Normal file
@ -0,0 +1,67 @@
|
||||
import { expect, tap } from '@push.rocks/tapbundle';
|
||||
import * as webstream from '../ts_web/index.js';
|
||||
|
||||
tap.test('WebDuplexStream fromUInt8Array should read back the same Uint8Array', async () => {
|
||||
const inputUint8Array = new Uint8Array([1, 2, 3, 4, 5]);
|
||||
const stream = webstream.WebDuplexStream.fromUInt8Array(inputUint8Array);
|
||||
|
||||
const reader = stream.readable.getReader();
|
||||
let readUint8Array = new Uint8Array();
|
||||
|
||||
// Read from the stream
|
||||
while (true) {
|
||||
const { value, done } = await reader.read();
|
||||
if (done) break;
|
||||
if (value) {
|
||||
// Concatenate value to readUint8Array
|
||||
const tempArray = new Uint8Array(readUint8Array.length + value.length);
|
||||
tempArray.set(readUint8Array, 0);
|
||||
tempArray.set(value, readUint8Array.length);
|
||||
readUint8Array = tempArray;
|
||||
}
|
||||
}
|
||||
|
||||
expect(readUint8Array).toEqual(inputUint8Array);
|
||||
});
|
||||
|
||||
tap.test('WebDuplexStream should handle transform with a write function', async () => {
|
||||
const input = [1, 2, 3, 4, 5];
|
||||
const expectedOutput = [2, 4, 6, 8, 10];
|
||||
|
||||
const webDuplexStream = new webstream.WebDuplexStream<number, number>({
|
||||
writeFunction: async (chunk, { push }) => {
|
||||
// Push the doubled number into the stream
|
||||
push(chunk * 2);
|
||||
},
|
||||
});
|
||||
|
||||
const writer = webDuplexStream.writable.getWriter();
|
||||
const reader = webDuplexStream.readable.getReader();
|
||||
|
||||
const output: number[] = [];
|
||||
|
||||
// Read from the stream asynchronously
|
||||
const readPromise = (async () => {
|
||||
while (true) {
|
||||
const { value, done } = await reader.read();
|
||||
if (done) break;
|
||||
if (value !== undefined) {
|
||||
output.push(value);
|
||||
}
|
||||
}
|
||||
})();
|
||||
|
||||
// Write to the stream
|
||||
for (const num of input) {
|
||||
await writer.write(num);
|
||||
}
|
||||
await writer.close();
|
||||
|
||||
// Wait for the reading to complete
|
||||
await readPromise;
|
||||
|
||||
// Assert that the output matches the expected transformed data
|
||||
expect(output).toEqual(expectedOutput);
|
||||
});
|
||||
|
||||
tap.start();
|
@ -1,8 +1,8 @@
|
||||
/**
|
||||
* autocreated commitinfo by @pushrocks/commitinfo
|
||||
* autocreated commitinfo by @push.rocks/commitinfo
|
||||
*/
|
||||
export const commitinfo = {
|
||||
name: '@push.rocks/smartstream',
|
||||
version: '3.0.26',
|
||||
description: 'simplifies access to node streams'
|
||||
version: '3.2.5',
|
||||
description: 'A library to simplify the creation and manipulation of Node.js streams, providing utilities for handling transform, duplex, and readable/writable streams effectively in TypeScript.'
|
||||
}
|
||||
|
15
ts/index.ts
15
ts/index.ts
@ -1,6 +1,17 @@
|
||||
export * from './smartstream.classes.passthrough.js';
|
||||
import { stream } from './smartstream.plugins.js';
|
||||
export {
|
||||
stream,
|
||||
}
|
||||
|
||||
export * from './smartstream.classes.smartduplex.js';
|
||||
export * from './smartstream.classes.streamwrapper.js';
|
||||
export * from './smartstream.classes.streamintake.js';
|
||||
|
||||
export * from './smartstream.functions.js'
|
||||
export * from './smartstream.functions.js';
|
||||
|
||||
import * as plugins from './smartstream.plugins.js';
|
||||
export const webstream = plugins.webstream;
|
||||
import * as nodewebhelpers from './smartstream.nodewebhelpers.js';
|
||||
export {
|
||||
nodewebhelpers,
|
||||
}
|
||||
|
@ -1,21 +0,0 @@
|
||||
import * as plugins from './smartstream.plugins.js';
|
||||
|
||||
export class PassThrough extends plugins.stream.Duplex {
|
||||
constructor(options?: plugins.stream.DuplexOptions) {
|
||||
super(options);
|
||||
}
|
||||
|
||||
_read(size: number): void {
|
||||
// No-op: Data written will be automatically available for reading.
|
||||
}
|
||||
|
||||
_write(chunk: any, encoding: BufferEncoding, callback: (error?: Error | null) => void): void {
|
||||
if (this.push(chunk, encoding)) {
|
||||
callback();
|
||||
} else {
|
||||
this.once('drain', () => {
|
||||
callback();
|
||||
});
|
||||
}
|
||||
}
|
||||
}
|
@ -3,7 +3,7 @@ import { Duplex, type DuplexOptions } from 'stream';
|
||||
|
||||
export interface IStreamTools {
|
||||
truncate: () => void;
|
||||
push: (pipeObject: any) => Promise<void>;
|
||||
push: (pipeObject: any) => Promise<boolean>;
|
||||
}
|
||||
|
||||
export interface IStreamWriteFunction<T, rT> {
|
||||
@ -15,12 +15,30 @@ export interface IStreamFinalFunction<rT> {
|
||||
}
|
||||
|
||||
export interface ISmartDuplexOptions<TInput, TOutput> extends DuplexOptions {
|
||||
/**
|
||||
* wether to print debug logs
|
||||
*/
|
||||
debug?: boolean;
|
||||
/**
|
||||
* the name of the stream
|
||||
*/
|
||||
name?: string;
|
||||
/**
|
||||
* a function that is being called to read more stuff from whereever to be processed by the stream
|
||||
* @returns
|
||||
*/
|
||||
readFunction?: () => Promise<void>;
|
||||
|
||||
/**
|
||||
* the write function is called for every chunk that is being written to the stream
|
||||
* it can push or return chunks (but does not have to) to be written to the readable side of the stream
|
||||
*/
|
||||
writeFunction?: IStreamWriteFunction<TInput, TOutput>;
|
||||
|
||||
/**
|
||||
* a final function that is run at the end of the stream
|
||||
*/
|
||||
finalFunction?: IStreamFinalFunction<TOutput>;
|
||||
// Add other custom options if necessary
|
||||
}
|
||||
|
||||
export class SmartDuplex<TInput = any, TOutput = any> extends Duplex {
|
||||
@ -34,38 +52,79 @@ export class SmartDuplex<TInput = any, TOutput = any> extends Duplex {
|
||||
return smartDuplex;
|
||||
}
|
||||
|
||||
public static fromWebReadableStream<T = any>(
|
||||
readableStream: ReadableStream<T>
|
||||
): SmartDuplex<T, T> {
|
||||
const smartDuplex = new SmartDuplex<T, T>({
|
||||
/**
|
||||
* this function is called whenever the stream is being read from and at the same time if nothing is enqueued
|
||||
* therefor it is important to always unlock the reader after reading
|
||||
*/
|
||||
readFunction: async () => {
|
||||
const reader = readableStream.getReader();
|
||||
const { value, done } = await reader.read();
|
||||
if (value !== undefined) {
|
||||
smartDuplex.push(value);
|
||||
}
|
||||
reader.releaseLock();
|
||||
if (done) {
|
||||
smartDuplex.push(null);
|
||||
}
|
||||
},
|
||||
});
|
||||
return smartDuplex;
|
||||
}
|
||||
|
||||
// INSTANCE
|
||||
private backpressuredArray: plugins.lik.BackpressuredArray<TOutput>;
|
||||
private backpressuredArray: plugins.lik.BackpressuredArray<TOutput>; // an array that only takes a defined amount of items
|
||||
public options: ISmartDuplexOptions<TInput, TOutput>;
|
||||
private observableSubscription?: plugins.smartrx.rxjs.Subscription;
|
||||
private debugLog(messageArg: string) {
|
||||
// optional debug log
|
||||
if (this.options.debug) {
|
||||
console.log(messageArg);
|
||||
}
|
||||
}
|
||||
|
||||
constructor(optionsArg?: ISmartDuplexOptions<TInput, TOutput>) {
|
||||
super(Object.assign({
|
||||
highWaterMark: 1,
|
||||
}, optionsArg));
|
||||
super(
|
||||
Object.assign(
|
||||
{
|
||||
highWaterMark: 1,
|
||||
},
|
||||
optionsArg
|
||||
)
|
||||
);
|
||||
this.options = optionsArg;
|
||||
this.backpressuredArray = new plugins.lik.BackpressuredArray<TOutput>(this.options.highWaterMark || 1)
|
||||
this.backpressuredArray = new plugins.lik.BackpressuredArray<TOutput>(
|
||||
this.options.highWaterMark || 1
|
||||
);
|
||||
}
|
||||
|
||||
public async _read(size: number): Promise<void> {
|
||||
this.debugLog(`${this.options.name}: read was called`);
|
||||
await this.backpressuredArray.waitForItems();
|
||||
this.debugLog(`${this.options.name}: successfully waited for items.`);
|
||||
if (this.options.readFunction) {
|
||||
await this.options.readFunction();
|
||||
}
|
||||
let canPushMore = true;
|
||||
while(this.backpressuredArray.data.length > 0 && canPushMore) {
|
||||
await this.backpressuredArray.waitForItems();
|
||||
this.debugLog(`${this.options.name}: successfully waited for items.`);
|
||||
let canPushMore = true;
|
||||
while (this.backpressuredArray.data.length > 0 && canPushMore) {
|
||||
const nextChunk = this.backpressuredArray.shift();
|
||||
canPushMore = this.push(nextChunk);
|
||||
}
|
||||
}
|
||||
|
||||
public async backpressuredPush(pushArg: TOutput) {
|
||||
const canPushMore = this.backpressuredArray.push(pushArg);
|
||||
if (!canPushMore) {
|
||||
this.debugLog(`${this.options.name}: cannot push more`);
|
||||
await this.backpressuredArray.waitForSpace();
|
||||
this.debugLog(`${this.options.name}: can push more again`);
|
||||
}
|
||||
return canPushMore;
|
||||
}
|
||||
|
||||
private asyncWritePromiseObjectmap = new plugins.lik.ObjectMap<Promise<any>>();
|
||||
// Ensure the _write method types the chunk as TInput and encodes TOutput
|
||||
public async _write(chunk: TInput, encoding: string, callback: (error?: Error | null) => void) {
|
||||
@ -81,12 +140,7 @@ export class SmartDuplex<TInput = any, TOutput = any> extends Duplex {
|
||||
callback();
|
||||
},
|
||||
push: async (pushArg: TOutput) => {
|
||||
const canPushMore = this.backpressuredArray.push(pushArg);
|
||||
if (!canPushMore) {
|
||||
this.debugLog(`${this.options.name}: cannot push more`);
|
||||
await this.backpressuredArray.waitForSpace();
|
||||
this.debugLog(`${this.options.name}: can push more again`);
|
||||
}
|
||||
return await this.backpressuredPush(pushArg);
|
||||
},
|
||||
};
|
||||
|
||||
@ -116,7 +170,7 @@ export class SmartDuplex<TInput = any, TOutput = any> extends Duplex {
|
||||
const tools: IStreamTools = {
|
||||
truncate: () => callback(),
|
||||
push: async (pipeObject) => {
|
||||
this.backpressuredArray.push(pipeObject);
|
||||
return this.backpressuredArray.push(pipeObject);
|
||||
},
|
||||
};
|
||||
|
||||
@ -134,4 +188,53 @@ export class SmartDuplex<TInput = any, TOutput = any> extends Duplex {
|
||||
this.backpressuredArray.push(null);
|
||||
callback();
|
||||
}
|
||||
|
||||
public async getWebStreams(): Promise<{ readable: ReadableStream; writable: WritableStream }> {
|
||||
const duplex = this;
|
||||
const readable = new ReadableStream({
|
||||
start(controller) {
|
||||
duplex.on('readable', () => {
|
||||
let chunk;
|
||||
while (null !== (chunk = duplex.read())) {
|
||||
controller.enqueue(chunk);
|
||||
}
|
||||
});
|
||||
|
||||
duplex.on('end', () => {
|
||||
controller.close();
|
||||
});
|
||||
},
|
||||
cancel(reason) {
|
||||
duplex.destroy(new Error(reason));
|
||||
},
|
||||
});
|
||||
|
||||
const writable = new WritableStream({
|
||||
write(chunk) {
|
||||
return new Promise<void>((resolve, reject) => {
|
||||
const isBackpressured = !duplex.write(chunk, (error) => {
|
||||
if (error) {
|
||||
reject(error);
|
||||
} else {
|
||||
resolve();
|
||||
}
|
||||
});
|
||||
|
||||
if (isBackpressured) {
|
||||
duplex.once('drain', resolve);
|
||||
}
|
||||
});
|
||||
},
|
||||
close() {
|
||||
return new Promise<void>((resolve, reject) => {
|
||||
duplex.end(resolve);
|
||||
});
|
||||
},
|
||||
abort(reason) {
|
||||
duplex.destroy(new Error(reason));
|
||||
},
|
||||
});
|
||||
|
||||
return { readable, writable };
|
||||
}
|
||||
}
|
||||
|
@ -1,6 +1,45 @@
|
||||
import * as plugins from './smartstream.plugins.js';
|
||||
|
||||
export class StreamIntake<T> extends plugins.stream.Readable {
|
||||
// STATIC
|
||||
public static async fromStream<U>(inputStream: plugins.stream.Readable | ReadableStream, options?: plugins.stream.ReadableOptions): Promise<StreamIntake<U>> {
|
||||
const intakeStream = new StreamIntake<U>(options);
|
||||
|
||||
if (inputStream instanceof plugins.stream.Readable) {
|
||||
inputStream.on('data', (chunk: U) => {
|
||||
intakeStream.pushData(chunk);
|
||||
});
|
||||
|
||||
inputStream.on('end', () => {
|
||||
intakeStream.signalEnd();
|
||||
});
|
||||
|
||||
inputStream.on('error', (err: Error) => {
|
||||
intakeStream.destroy(err);
|
||||
});
|
||||
} else {
|
||||
const reader = (inputStream as ReadableStream).getReader();
|
||||
|
||||
const readChunk = () => {
|
||||
reader.read().then(({ done, value }) => {
|
||||
if (done) {
|
||||
intakeStream.signalEnd();
|
||||
} else {
|
||||
intakeStream.pushData(value);
|
||||
readChunk();
|
||||
}
|
||||
}).catch((err) => {
|
||||
intakeStream.destroy(err);
|
||||
});
|
||||
};
|
||||
|
||||
readChunk();
|
||||
}
|
||||
|
||||
return intakeStream;
|
||||
}
|
||||
|
||||
// INSTANCE
|
||||
private signalEndBoolean = false;
|
||||
private chunkStore: T[] = [];
|
||||
public pushNextObservable = new plugins.smartrx.ObservableIntake<any>();
|
||||
@ -14,7 +53,7 @@ export class StreamIntake<T> extends plugins.stream.Readable {
|
||||
_read(size: number): void {
|
||||
// console.log('get next');
|
||||
const pushChunk = (): void => {
|
||||
if (this.chunkStore.length > 0) {
|
||||
while (this.chunkStore.length > 0) {
|
||||
// If push returns false, then we should stop reading
|
||||
if (!this.push(this.chunkStore.shift())) {
|
||||
return;
|
||||
|
@ -1,4 +1,4 @@
|
||||
import { Transform, type TransformCallback, type TransformOptions } from 'stream';
|
||||
import { type TransformOptions } from 'stream';
|
||||
import { SmartDuplex } from './smartstream.classes.smartduplex.js';
|
||||
|
||||
export interface AsyncTransformFunction<TInput, TOutput> {
|
||||
@ -18,4 +18,13 @@ export function createTransformFunction<TInput, TOutput>(
|
||||
});
|
||||
|
||||
return smartDuplexStream;
|
||||
}
|
||||
}
|
||||
|
||||
export const createPassThrough = () => {
|
||||
return new SmartDuplex({
|
||||
objectMode: true,
|
||||
writeFunction: async (chunkArg, toolsArg) => {
|
||||
return chunkArg;
|
||||
}
|
||||
})
|
||||
}
|
||||
|
151
ts/smartstream.nodewebhelpers.ts
Normal file
151
ts/smartstream.nodewebhelpers.ts
Normal file
@ -0,0 +1,151 @@
|
||||
import * as plugins from './smartstream.plugins.js';
|
||||
|
||||
/**
|
||||
* Creates a Web ReadableStream from a file.
|
||||
*
|
||||
* @param filePath - The path to the file to be read
|
||||
* @returns A Web ReadableStream that reads the file in chunks
|
||||
*/
|
||||
export function createWebReadableStreamFromFile(filePath: string): ReadableStream<Uint8Array> {
|
||||
const fileStream = plugins.fs.createReadStream(filePath);
|
||||
|
||||
return new ReadableStream({
|
||||
start(controller) {
|
||||
// When data is available, enqueue it into the Web ReadableStream
|
||||
fileStream.on('data', (chunk) => {
|
||||
controller.enqueue(chunk as Uint8Array);
|
||||
});
|
||||
|
||||
// When the file stream ends, close the Web ReadableStream
|
||||
fileStream.on('end', () => {
|
||||
controller.close();
|
||||
});
|
||||
|
||||
// If there's an error, error the Web ReadableStream
|
||||
fileStream.on('error', (err) => {
|
||||
controller.error(err);
|
||||
});
|
||||
},
|
||||
cancel() {
|
||||
// If the Web ReadableStream is canceled, destroy the file stream
|
||||
fileStream.destroy();
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Converts a Web ReadableStream to a Node.js Readable stream.
|
||||
*
|
||||
* @param webStream - The Web ReadableStream to convert
|
||||
* @returns A Node.js Readable stream that reads data from the Web ReadableStream
|
||||
*/
|
||||
export function convertWebReadableToNodeReadable(webStream: ReadableStream<Uint8Array>): plugins.stream.Readable {
|
||||
const reader = webStream.getReader();
|
||||
|
||||
return new plugins.stream.Readable({
|
||||
async read() {
|
||||
try {
|
||||
const { value, done } = await reader.read();
|
||||
if (done) {
|
||||
this.push(null); // Signal end of stream
|
||||
} else {
|
||||
this.push(Buffer.from(value)); // Convert Uint8Array to Buffer for Node.js Readable
|
||||
}
|
||||
} catch (err) {
|
||||
this.destroy(err); // Handle errors by destroying the stream
|
||||
}
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Converts a Node.js Readable stream to a Web ReadableStream.
|
||||
*
|
||||
* @param nodeStream - The Node.js Readable stream to convert
|
||||
* @returns A Web ReadableStream that reads data from the Node.js Readable stream
|
||||
*/
|
||||
export function convertNodeReadableToWebReadable(nodeStream: plugins.stream.Readable): ReadableStream<Uint8Array> {
|
||||
return new ReadableStream({
|
||||
start(controller) {
|
||||
nodeStream.on('data', (chunk) => {
|
||||
controller.enqueue(new Uint8Array(chunk));
|
||||
});
|
||||
|
||||
nodeStream.on('end', () => {
|
||||
controller.close();
|
||||
});
|
||||
|
||||
nodeStream.on('error', (err) => {
|
||||
controller.error(err);
|
||||
});
|
||||
},
|
||||
cancel() {
|
||||
nodeStream.destroy();
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Converts a Web WritableStream to a Node.js Writable stream.
|
||||
*
|
||||
* @param webWritable - The Web WritableStream to convert
|
||||
* @returns A Node.js Writable stream that writes data to the Web WritableStream
|
||||
*/
|
||||
export function convertWebWritableToNodeWritable(webWritable: WritableStream<Uint8Array>): plugins.stream.Writable {
|
||||
const writer = webWritable.getWriter();
|
||||
|
||||
return new plugins.stream.Writable({
|
||||
async write(chunk, encoding, callback) {
|
||||
try {
|
||||
await writer.write(new Uint8Array(chunk));
|
||||
callback();
|
||||
} catch (err) {
|
||||
callback(err);
|
||||
}
|
||||
},
|
||||
final(callback) {
|
||||
writer.close().then(() => callback()).catch(callback);
|
||||
},
|
||||
destroy(err, callback) {
|
||||
writer.abort(err).then(() => callback(err)).catch(callback);
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Converts a Node.js Writable stream to a Web WritableStream.
|
||||
*
|
||||
* @param nodeWritable - The Node.js Writable stream to convert
|
||||
* @returns A Web WritableStream that writes data to the Node.js Writable stream
|
||||
*/
|
||||
export function convertNodeWritableToWebWritable(nodeWritable: plugins.stream.Writable): WritableStream<Uint8Array> {
|
||||
return new WritableStream({
|
||||
write(chunk) {
|
||||
return new Promise((resolve, reject) => {
|
||||
nodeWritable.write(Buffer.from(chunk), (err) => {
|
||||
if (err) {
|
||||
reject(err);
|
||||
} else {
|
||||
resolve();
|
||||
}
|
||||
});
|
||||
});
|
||||
},
|
||||
close() {
|
||||
return new Promise((resolve, reject) => {
|
||||
nodeWritable.end((err) => {
|
||||
if (err) {
|
||||
reject(err);
|
||||
} else {
|
||||
resolve();
|
||||
}
|
||||
});
|
||||
});
|
||||
},
|
||||
abort(reason) {
|
||||
return new Promise((resolve, reject) => {
|
||||
nodeWritable.destroy(reason);
|
||||
});
|
||||
}
|
||||
});
|
||||
}
|
@ -1,12 +1,14 @@
|
||||
// node native
|
||||
import * as fs from 'fs';
|
||||
import * as stream from 'stream';
|
||||
|
||||
export { stream };
|
||||
export { fs, stream };
|
||||
|
||||
// pushrocks scope
|
||||
import * as lik from '@push.rocks/lik';
|
||||
import * as smartpromise from '@push.rocks/smartpromise';
|
||||
import * as smartrx from '@push.rocks/smartrx';
|
||||
import * as webstream from '../dist_ts_web/index.js';
|
||||
|
||||
export { lik, smartpromise, smartrx };
|
||||
export { lik, smartpromise, smartrx, webstream };
|
||||
|
||||
|
8
ts_web/00_commitinfo_data.ts
Normal file
8
ts_web/00_commitinfo_data.ts
Normal file
@ -0,0 +1,8 @@
|
||||
/**
|
||||
* autocreated commitinfo by @push.rocks/commitinfo
|
||||
*/
|
||||
export const commitinfo = {
|
||||
name: '@push.rocks/smartstream',
|
||||
version: '3.2.5',
|
||||
description: 'A library to simplify the creation and manipulation of Node.js streams, providing utilities for handling transform, duplex, and readable/writable streams effectively in TypeScript.'
|
||||
}
|
135
ts_web/classes.webduplexstream.ts
Normal file
135
ts_web/classes.webduplexstream.ts
Normal file
@ -0,0 +1,135 @@
|
||||
import * as plugins from './plugins.js';
|
||||
|
||||
// ========================================
|
||||
// Interfaces for Read functionality
|
||||
// ========================================
|
||||
export interface IStreamToolsRead<TInput, TOutput> {
|
||||
done: () => void;
|
||||
write: (writeArg: TInput) => Promise<void>;
|
||||
}
|
||||
|
||||
/**
|
||||
* The read function is called when data needs to be read into the stream.
|
||||
*/
|
||||
export interface IStreamReadFunction<TInput, TOutput> {
|
||||
(toolsArg: IStreamToolsRead<TInput, TOutput>): Promise<void>;
|
||||
}
|
||||
|
||||
// ========================================
|
||||
// Interfaces for Write functionality
|
||||
// ========================================
|
||||
export interface IStreamToolsWrite<TInput, TOutput> {
|
||||
truncate: () => void;
|
||||
push: (pushArg: TOutput) => void;
|
||||
}
|
||||
|
||||
/**
|
||||
* The write function is called whenever a chunk is written to the stream.
|
||||
*/
|
||||
export interface IStreamWriteFunction<TInput, TOutput> {
|
||||
(chunkArg: TInput, toolsArg: IStreamToolsWrite<TInput, TOutput>): Promise<any>;
|
||||
}
|
||||
|
||||
export interface IStreamFinalFunction<TInput, TOutput> {
|
||||
(toolsArg: IStreamToolsWrite<TInput, TOutput>): Promise<TOutput | void>;
|
||||
}
|
||||
|
||||
export interface WebDuplexStreamOptions<TInput, TOutput> {
|
||||
readFunction?: IStreamReadFunction<TInput, TOutput>;
|
||||
writeFunction?: IStreamWriteFunction<TInput, TOutput>;
|
||||
finalFunction?: IStreamFinalFunction<TInput, TOutput>;
|
||||
}
|
||||
|
||||
export class WebDuplexStream<TInput = any, TOutput = any> extends TransformStream<TInput, TOutput> {
|
||||
// INSTANCE
|
||||
options: WebDuplexStreamOptions<TInput, TOutput>;
|
||||
|
||||
constructor(optionsArg: WebDuplexStreamOptions<TInput, TOutput>) {
|
||||
super({
|
||||
async start(controller) {
|
||||
// Optionally initialize any state here
|
||||
},
|
||||
async transform(chunk, controller) {
|
||||
if (optionsArg?.writeFunction) {
|
||||
const tools: IStreamToolsWrite<TInput, TOutput> = {
|
||||
truncate: () => controller.terminate(),
|
||||
push: (pushArg: TOutput) => controller.enqueue(pushArg),
|
||||
};
|
||||
|
||||
try {
|
||||
const writeReturnChunk = await optionsArg.writeFunction(chunk, tools);
|
||||
if (writeReturnChunk !== undefined && writeReturnChunk !== null) {
|
||||
controller.enqueue(writeReturnChunk);
|
||||
}
|
||||
} catch (err) {
|
||||
controller.error(err);
|
||||
}
|
||||
} else {
|
||||
// If no writeFunction is provided, pass the chunk through
|
||||
controller.enqueue(chunk as unknown as TOutput);
|
||||
}
|
||||
},
|
||||
async flush(controller) {
|
||||
if (optionsArg?.finalFunction) {
|
||||
const tools: IStreamToolsWrite<TInput, TOutput> = {
|
||||
truncate: () => controller.terminate(),
|
||||
push: (pushArg) => controller.enqueue(pushArg),
|
||||
};
|
||||
|
||||
try {
|
||||
const finalChunk = await optionsArg.finalFunction(tools);
|
||||
if (finalChunk) {
|
||||
controller.enqueue(finalChunk);
|
||||
}
|
||||
} catch (err) {
|
||||
controller.error(err);
|
||||
} finally {
|
||||
controller.terminate();
|
||||
}
|
||||
} else {
|
||||
controller.terminate();
|
||||
}
|
||||
},
|
||||
});
|
||||
|
||||
this.options = optionsArg;
|
||||
|
||||
// Start producing data if readFunction is provided
|
||||
if (this.options.readFunction) {
|
||||
this._startReading();
|
||||
}
|
||||
}
|
||||
|
||||
private async _startReading() {
|
||||
const writable = this.writable;
|
||||
const writer = writable.getWriter();
|
||||
|
||||
const tools: IStreamToolsRead<TInput, TOutput> = {
|
||||
done: () => writer.close(),
|
||||
write: async (writeArg) => await writer.write(writeArg),
|
||||
};
|
||||
|
||||
try {
|
||||
await this.options.readFunction(tools);
|
||||
} catch (err) {
|
||||
writer.abort(err);
|
||||
} finally {
|
||||
writer.releaseLock();
|
||||
}
|
||||
}
|
||||
|
||||
// Static method example (adjust as needed)
|
||||
static fromUInt8Array(uint8Array: Uint8Array): WebDuplexStream<Uint8Array, Uint8Array> {
|
||||
const stream = new WebDuplexStream<Uint8Array, Uint8Array>({
|
||||
writeFunction: async (chunk, { push }) => {
|
||||
push(chunk); // Directly push the chunk as is
|
||||
return null;
|
||||
},
|
||||
});
|
||||
|
||||
const writer = stream.writable.getWriter();
|
||||
writer.write(uint8Array).then(() => writer.close());
|
||||
|
||||
return stream;
|
||||
}
|
||||
}
|
2
ts_web/index.ts
Normal file
2
ts_web/index.ts
Normal file
@ -0,0 +1,2 @@
|
||||
import './plugins.js';
|
||||
export * from './classes.webduplexstream.js';
|
15
ts_web/plugins.ts
Normal file
15
ts_web/plugins.ts
Normal file
@ -0,0 +1,15 @@
|
||||
// @push.rocks scope
|
||||
import * as smartenv from '@push.rocks/smartenv';
|
||||
|
||||
export {
|
||||
smartenv,
|
||||
}
|
||||
|
||||
// lets setup dependencies
|
||||
const smartenvInstance = new smartenv.Smartenv();
|
||||
|
||||
await smartenvInstance.getSafeNodeModule<typeof import('stream/web')>('stream/web', async (moduleArg) => {
|
||||
globalThis.ReadableStream = moduleArg.ReadableStream;
|
||||
globalThis.WritableStream = moduleArg.WritableStream;
|
||||
globalThis.TransformStream = moduleArg.TransformStream;
|
||||
})
|
Loading…
x
Reference in New Issue
Block a user