Compare commits
No commits in common. "master" and "v3.0.14" have entirely different histories.
150
changelog.md
150
changelog.md
@ -1,150 +0,0 @@
|
||||
# Changelog
|
||||
|
||||
## 2024-12-02 - 3.3.7 - fix(package)
|
||||
Update author field in package.json
|
||||
|
||||
- Corrected the author field from 'Lossless GmbH' to 'Task Venture Capital GmbH' in the package.json file.
|
||||
|
||||
## 2024-12-02 - 3.3.6 - fix(package)
|
||||
Fix license field in package.json to reflect MIT licensing
|
||||
|
||||
|
||||
## 2024-11-25 - 3.3.5 - fix(test)
|
||||
Refactor trash test to improve metadata validation
|
||||
|
||||
- Added new checks in trash tests to ensure metadata files are correctly moved to trash.
|
||||
- Validated the presence and integrity of metadata within trashed files.
|
||||
|
||||
## 2024-11-25 - 3.3.4 - fix(core)
|
||||
Minor refactoring and cleanup of TypeScript source files for improved readability and maintainability.
|
||||
|
||||
|
||||
## 2024-11-24 - 3.3.3 - fix(documentation)
|
||||
Improved documentation accuracy and consistency
|
||||
|
||||
- Updated the project description to reflect the cloud-agnostic nature and advanced capabilities
|
||||
- Enhanced the README with detailed explanations and code examples for advanced features like trash management
|
||||
- Clarified the handling and importance of metadata using the MetaData utility
|
||||
|
||||
## 2024-11-24 - 3.3.2 - fix(documentation)
|
||||
Updated keywords and description for clarity and consistency.
|
||||
|
||||
- Modified keywords and description in package.json and npmextra.json.
|
||||
- Enhanced readme.md file structure and examples
|
||||
|
||||
## 2024-11-24 - 3.3.1 - fix(File)
|
||||
Fixed issue with file restore metadata operations.
|
||||
|
||||
- Corrected the order of operations in the file restore function to ensure custom metadata is appropriately deleted after moving the file.
|
||||
|
||||
## 2024-11-24 - 3.3.0 - feat(core)
|
||||
Enhanced directory handling and file restoration from trash
|
||||
|
||||
- Refined getSubDirectoryByName to handle file paths treated as directories.
|
||||
- Introduced file restoration function from trash to original or specified paths.
|
||||
|
||||
## 2024-11-24 - 3.2.2 - fix(core)
|
||||
Refactor Bucket class for improved error handling
|
||||
|
||||
- Ensured safe access using non-null assertions when finding a bucket.
|
||||
- Enhanced fastPut method by adding fastPutStrict for safer operations.
|
||||
- Added explicit error handling and type checking in fastExists method.
|
||||
|
||||
## 2024-11-24 - 3.2.1 - fix(metadata)
|
||||
Fix metadata handling for deleted files
|
||||
|
||||
- Ensured metadata is correctly stored and managed when files are deleted into the trash.
|
||||
|
||||
## 2024-11-24 - 3.2.0 - feat(bucket)
|
||||
Enhanced SmartBucket with trash management and metadata handling
|
||||
|
||||
- Added functionality to move files to a trash directory.
|
||||
- Introduced methods to handle file metadata more robustly.
|
||||
- Implemented a method to clean all contents from a bucket.
|
||||
- Enhanced directory retrieval to handle non-existent directories with options.
|
||||
- Improved handling of file paths and metadata within the storage system.
|
||||
|
||||
## 2024-11-18 - 3.1.0 - feat(file)
|
||||
Added functionality to retrieve magic bytes from files and detect file types using magic bytes.
|
||||
|
||||
- Introduced method `getMagicBytes` in `File` and `Bucket` classes to retrieve a specific number of bytes from a file.
|
||||
- Enhanced file type detection by utilizing magic bytes in `MetaData` class.
|
||||
- Updated dependencies for better performance and compatibility.
|
||||
|
||||
## 2024-11-18 - 3.0.24 - fix(metadata)
|
||||
Fix metadata handling to address type assertion and data retrieval.
|
||||
|
||||
- Fixed type assertion issues in `MetaData` class properties with type non-null assertions.
|
||||
- Corrected the handling of JSON data retrieval in `MetaData.storeCustomMetaData` function.
|
||||
|
||||
## 2024-10-16 - 3.0.23 - fix(dependencies)
|
||||
Update package dependencies for improved functionality and security.
|
||||
|
||||
- Updated @aws-sdk/client-s3 to version ^3.670.0 for enhanced S3 client capabilities.
|
||||
- Updated @push.rocks/smartstream to version ^3.2.4.
|
||||
- Updated the dev dependency @push.rocks/tapbundle to version ^5.3.0.
|
||||
|
||||
## 2024-07-28 - 3.0.22 - fix(dependencies)
|
||||
Update dependencies and improve bucket retrieval logging
|
||||
|
||||
- Updated @aws-sdk/client-s3 to ^3.620.0
|
||||
- Updated @git.zone/tsbuild to ^2.1.84
|
||||
- Updated @git.zone/tsrun to ^1.2.49
|
||||
- Updated @push.rocks/smartpromise to ^4.0.4
|
||||
- Updated @tsclass/tsclass to ^4.1.2
|
||||
- Added a log for when a bucket is not found by name in getBucketByName method
|
||||
|
||||
## 2024-07-04 - 3.0.21 - fix(test)
|
||||
Update endpoint configuration in tests to use environment variable
|
||||
|
||||
- Modified `qenv.yml` to include `S3_ENDPOINT` as a required environment variable.
|
||||
- Updated test files to fetch `S3_ENDPOINT` from environment instead of hardcoding.
|
||||
|
||||
## 2024-06-19 - 3.0.20 - Fix and Stability Updates
|
||||
Improved overall stability and consistency.
|
||||
|
||||
## 2024-06-18 - 3.0.18 - Delete Functions Consistency
|
||||
Ensured more consistency between delete methods and trash behavior.
|
||||
|
||||
## 2024-06-17 - 3.0.17 to 3.0.16 - Fix and Update
|
||||
Routine updates and fixes performed.
|
||||
|
||||
## 2024-06-11 - 3.0.15 to 3.0.14 - Fix and Update
|
||||
Routine updates and fixes performed.
|
||||
|
||||
## 2024-06-10 - 3.0.13 - Trash Feature Completion
|
||||
Finished work on trash feature.
|
||||
|
||||
## 2024-06-09 - 3.0.12 - Fix and Update
|
||||
Routine updates and fixes performed.
|
||||
|
||||
## 2024-06-08 - 3.0.11 to 3.0.10 - Fix and Update
|
||||
Routine updates and fixes performed.
|
||||
|
||||
## 2024-06-03 - 3.0.10 - Fix and Update
|
||||
Routine updates and fixes performed.
|
||||
|
||||
## 2024-05-29 - 3.0.9 - Update Description
|
||||
Updated project description.
|
||||
|
||||
## 2024-05-27 - 3.0.8 to 3.0.6 - Pathing and Core Updates
|
||||
Routine updates and fixes performed.
|
||||
- S3 paths' pathing differences now correctly handled with a reducePath method.
|
||||
|
||||
## 2024-05-21 - 3.0.5 to 3.0.4 - Fix and Update
|
||||
Routine updates and fixes performed.
|
||||
|
||||
## 2024-05-17 - 3.0.3 to 3.0.2 - Fix and Update
|
||||
Routine updates and fixes performed.
|
||||
|
||||
## 2024-05-17 - 3.0.0 - Major Release
|
||||
Introduced breaking changes in core and significant improvements.
|
||||
|
||||
## 2024-05-05 - 2.0.5 - Breaking Changes
|
||||
Introduced breaking changes in core functionality.
|
||||
|
||||
## 2024-04-14 - 2.0.4 - TSConfig Update
|
||||
Updated TypeScript configuration.
|
||||
|
||||
## 2024-01-01 - 2.0.2 - Organization Scheme Update
|
||||
Switched to the new organizational scheme.
|
19
license
19
license
@ -1,19 +0,0 @@
|
||||
Copyright (c) 2014 Task Venture Capital GmbH (hello@task.vc)
|
||||
|
||||
Permission is hereby granted, free of charge, to any person obtaining a copy
|
||||
of this software and associated documentation files (the "Software"), to deal
|
||||
in the Software without restriction, including without limitation the rights
|
||||
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
||||
copies of the Software, and to permit persons to whom the Software is
|
||||
furnished to do so, subject to the following conditions:
|
||||
|
||||
The above copyright notice and this permission notice shall be included in all
|
||||
copies or substantial portions of the Software.
|
||||
|
||||
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
||||
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
||||
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
||||
SOFTWARE.
|
@ -8,30 +8,28 @@
|
||||
"githost": "code.foss.global",
|
||||
"gitscope": "push.rocks",
|
||||
"gitrepo": "smartbucket",
|
||||
"description": "A TypeScript library providing a cloud-agnostic interface for managing object storage with functionalities like bucket management, file and directory operations, and advanced features such as metadata handling and file locking.",
|
||||
"description": "A TypeScript library offering simple and cloud-agnostic object storage with advanced features like bucket creation, file and directory management, and data streaming.",
|
||||
"npmPackagename": "@push.rocks/smartbucket",
|
||||
"license": "MIT",
|
||||
"keywords": [
|
||||
"TypeScript",
|
||||
"cloud agnostic",
|
||||
"cloud storage",
|
||||
"object storage",
|
||||
"bucket management",
|
||||
"file operations",
|
||||
"bucket creation",
|
||||
"file management",
|
||||
"directory management",
|
||||
"data streaming",
|
||||
"S3",
|
||||
"multi-cloud",
|
||||
"file locking",
|
||||
"metadata management",
|
||||
"buffer handling",
|
||||
"access control",
|
||||
"environment configuration",
|
||||
"API",
|
||||
"unified storage",
|
||||
"bucket policies",
|
||||
"trash management",
|
||||
"file transfer",
|
||||
"data management",
|
||||
"streaming"
|
||||
"S3",
|
||||
"minio",
|
||||
"file locking",
|
||||
"metadata",
|
||||
"buffer handling",
|
||||
"access key",
|
||||
"secret key",
|
||||
"cloud agnostic"
|
||||
]
|
||||
}
|
||||
},
|
||||
|
4
package-lock.json
generated
4
package-lock.json
generated
@ -1,12 +1,12 @@
|
||||
{
|
||||
"name": "@push.rocks/smartbucket",
|
||||
"version": "3.3.7",
|
||||
"version": "3.0.14",
|
||||
"lockfileVersion": 3,
|
||||
"requires": true,
|
||||
"packages": {
|
||||
"": {
|
||||
"name": "@push.rocks/smartbucket",
|
||||
"version": "3.3.7",
|
||||
"version": "3.0.14",
|
||||
"license": "UNLICENSED",
|
||||
"dependencies": {
|
||||
"@push.rocks/smartpath": "^5.0.18",
|
||||
|
52
package.json
52
package.json
@ -1,33 +1,33 @@
|
||||
{
|
||||
"name": "@push.rocks/smartbucket",
|
||||
"version": "3.3.7",
|
||||
"description": "A TypeScript library providing a cloud-agnostic interface for managing object storage with functionalities like bucket management, file and directory operations, and advanced features such as metadata handling and file locking.",
|
||||
"version": "3.0.14",
|
||||
"description": "A TypeScript library offering simple and cloud-agnostic object storage with advanced features like bucket creation, file and directory management, and data streaming.",
|
||||
"main": "dist_ts/index.js",
|
||||
"typings": "dist_ts/index.d.ts",
|
||||
"type": "module",
|
||||
"author": "Task Venture Capital GmbH",
|
||||
"license": "MIT",
|
||||
"author": "Lossless GmbH",
|
||||
"license": "UNLICENSED",
|
||||
"scripts": {
|
||||
"test": "(tstest test/)",
|
||||
"build": "(tsbuild --web --allowimplicitany)"
|
||||
},
|
||||
"devDependencies": {
|
||||
"@git.zone/tsbuild": "^2.1.84",
|
||||
"@git.zone/tsrun": "^1.2.49",
|
||||
"@git.zone/tsbuild": "^2.1.80",
|
||||
"@git.zone/tsrun": "^1.2.46",
|
||||
"@git.zone/tstest": "^1.0.90",
|
||||
"@push.rocks/qenv": "^6.1.0",
|
||||
"@push.rocks/tapbundle": "^5.5.3"
|
||||
"@push.rocks/qenv": "^6.0.5",
|
||||
"@push.rocks/tapbundle": "^5.0.23"
|
||||
},
|
||||
"dependencies": {
|
||||
"@aws-sdk/client-s3": "^3.699.0",
|
||||
"@push.rocks/smartmime": "^2.0.4",
|
||||
"@push.rocks/smartmime": "^2.0.2",
|
||||
"@push.rocks/smartpath": "^5.0.18",
|
||||
"@push.rocks/smartpromise": "^4.0.4",
|
||||
"@push.rocks/smartpromise": "^4.0.3",
|
||||
"@push.rocks/smartrx": "^3.0.7",
|
||||
"@push.rocks/smartstream": "^3.2.5",
|
||||
"@push.rocks/smartstream": "^3.0.44",
|
||||
"@push.rocks/smartstring": "^4.0.15",
|
||||
"@push.rocks/smartunique": "^3.0.9",
|
||||
"@tsclass/tsclass": "^4.1.2"
|
||||
"@tsclass/tsclass": "^4.0.55",
|
||||
"minio": "^8.0.0"
|
||||
},
|
||||
"private": false,
|
||||
"files": [
|
||||
@ -47,25 +47,23 @@
|
||||
],
|
||||
"keywords": [
|
||||
"TypeScript",
|
||||
"cloud agnostic",
|
||||
"cloud storage",
|
||||
"object storage",
|
||||
"bucket management",
|
||||
"file operations",
|
||||
"bucket creation",
|
||||
"file management",
|
||||
"directory management",
|
||||
"data streaming",
|
||||
"S3",
|
||||
"multi-cloud",
|
||||
"file locking",
|
||||
"metadata management",
|
||||
"buffer handling",
|
||||
"access control",
|
||||
"environment configuration",
|
||||
"API",
|
||||
"unified storage",
|
||||
"bucket policies",
|
||||
"trash management",
|
||||
"file transfer",
|
||||
"data management",
|
||||
"streaming"
|
||||
"S3",
|
||||
"minio",
|
||||
"file locking",
|
||||
"metadata",
|
||||
"buffer handling",
|
||||
"access key",
|
||||
"secret key",
|
||||
"cloud agnostic"
|
||||
],
|
||||
"homepage": "https://code.foss.global/push.rocks/smartbucket",
|
||||
"repository": {
|
||||
|
5826
pnpm-lock.yaml
generated
5826
pnpm-lock.yaml
generated
File diff suppressed because it is too large
Load Diff
@ -1,3 +1 @@
|
||||
* The project uses the official s3 client, not the minio client.
|
||||
* notice the difference between *Strict methods and the normal methods.
|
||||
* metadata is handled though the MetaData class. Important!
|
||||
|
341
readme.md
341
readme.md
@ -1,49 +1,41 @@
|
||||
```markdown
|
||||
# @push.rocks/smartbucket
|
||||
|
||||
A comprehensive TypeScript library for cloud-agnostic object storage offering bucket management, file operations, and advanced data streaming.
|
||||
A TypeScript library for cloud-independent object storage, providing features like bucket creation, file and directory management, and data streaming.
|
||||
|
||||
## Install
|
||||
|
||||
To install `@push.rocks/smartbucket`, ensure you have Node.js and npm installed. Then, run the following command in your project directory:
|
||||
To install `@push.rocks/smartbucket`, you need to have Node.js and npm (Node Package Manager) installed. If they are installed, you can add `@push.rocks/smartbucket` to your project by running the following command in your project's root directory:
|
||||
|
||||
```bash
|
||||
npm install @push.rocks/smartbucket --save
|
||||
```
|
||||
|
||||
This command will add `@push.rocks/smartbucket` to your project's dependencies and install it along with its requirements in the `node_modules` directory.
|
||||
This command will download and install `@push.rocks/smartbucket` along with its required dependencies into your project's `node_modules` directory and save it as a dependency in your project's `package.json` file.
|
||||
|
||||
## Usage
|
||||
|
||||
### Introduction
|
||||
`@push.rocks/smartbucket` is a TypeScript module designed to provide simple cloud-independent object storage functionality. It wraps various cloud storage providers such as AWS S3, Google Cloud Storage, and others, offering a unified API to manage storage buckets and objects within those buckets.
|
||||
|
||||
`@push.rocks/smartbucket` provides a robust set of features to manage cloud storage operations in a cloud-agnostic manner. By leveraging this library, you can seamlessly interact with object storage services like AWS S3, without being tied to any vendor-specific implementations. This library not only abstracts basic file operations but also integrates advanced capabilities such as metadata management, data streaming, file locking, and bucket policies, all through a simplified API.
|
||||
In this guide, we will delve into the usage of SmartBucket, covering its full range of features from setting up the library to advanced usage scenarios.
|
||||
|
||||
### Table of Contents
|
||||
|
||||
1. [Setting Up](#setting-up)
|
||||
2. [Working with Buckets](#working-with-buckets)
|
||||
- [Creating a New Bucket](#creating-a-new-bucket)
|
||||
- [Listing Buckets](#listing-buckets)
|
||||
- [Deleting Buckets](#deleting-buckets)
|
||||
3. [File Operations in Buckets](#file-operations-in-buckets)
|
||||
2. [Creating a New Bucket](#creating-a-new-bucket)
|
||||
3. [Listing Buckets](#listing-buckets)
|
||||
4. [Working with Files](#working-with-files)
|
||||
- [Uploading Files](#uploading-files)
|
||||
- [Downloading Files](#downloading-files)
|
||||
- [Streaming Files](#streaming-files)
|
||||
- [Deleting Files](#deleting-files)
|
||||
4. [Directory Operations](#directory-operations)
|
||||
- [Listing Directories and Files](#listing-directories-and-files)
|
||||
- [Managing Files in Directories](#managing-files-in-directories)
|
||||
5. [Advanced Features](#advanced-features)
|
||||
- [Streaming Files](#streaming-files)
|
||||
5. [Working with Directories](#working-with-directories)
|
||||
6. [Advanced Features](#advanced-features)
|
||||
- [Bucket Policies](#bucket-policies)
|
||||
- [Metadata Management](#metadata-management)
|
||||
- [File Locking](#file-locking)
|
||||
- [Trash Management](#trash-management)
|
||||
6. [Cloud Agnosticism](#cloud-agnosticism)
|
||||
- [Object Metadata](#object-metadata)
|
||||
- [Cloud Agnostic](#cloud-agnostic)
|
||||
|
||||
### Setting Up
|
||||
|
||||
Begin by importing the necessary classes from the `@push.rocks/smartbucket` package into your TypeScript file. Create an instance of `SmartBucket` with your storage configuration:
|
||||
First, ensure you are using ECMAScript modules (ESM) and TypeScript in your project for best compatibility. Here's how to import and initialize SmartBucket in a TypeScript file:
|
||||
|
||||
```typescript
|
||||
import {
|
||||
@ -57,223 +49,264 @@ const mySmartBucket = new SmartBucket({
|
||||
accessKey: "yourAccessKey",
|
||||
accessSecret: "yourSecretKey",
|
||||
endpoint: "yourEndpointURL",
|
||||
port: 443,
|
||||
useSsl: true
|
||||
port: 443, // Default is 443, can be customized for specific endpoint
|
||||
useSsl: true // Defaults to true
|
||||
});
|
||||
```
|
||||
|
||||
Replace `"yourAccessKey"`, `"yourSecretKey"`, and `"yourEndpointURL"` with actual data specific to your cloud provider.
|
||||
Make sure to replace `"yourAccessKey"`, `"yourSecretKey"`, and `"yourEndpointURL"` with your actual credentials and endpoint URL. The `port` and `useSsl` options are optional and can be omitted if the defaults are acceptable.
|
||||
|
||||
### Working with Buckets
|
||||
### Creating a New Bucket
|
||||
|
||||
#### Creating a New Bucket
|
||||
|
||||
Creating a bucket involves invoking the `createBucket` method. Note that bucket names are unique and follow the rules of the cloud provider:
|
||||
To create a new bucket:
|
||||
|
||||
```typescript
|
||||
async function createBucket(bucketName: string) {
|
||||
try {
|
||||
const newBucket: Bucket = await mySmartBucket.createBucket(bucketName);
|
||||
const myBucket: Bucket = await mySmartBucket.createBucket(bucketName);
|
||||
console.log(`Bucket ${bucketName} created successfully.`);
|
||||
} catch (error) {
|
||||
console.error("Error creating bucket:", error);
|
||||
}
|
||||
}
|
||||
|
||||
createBucket("myNewBucket");
|
||||
// Use the function
|
||||
createBucket("exampleBucket");
|
||||
```
|
||||
|
||||
#### Listing Buckets
|
||||
Bucket names must be unique across the storage service.
|
||||
|
||||
While the library uses cloud-provider capabilities like AWS SDK to list existing buckets, `smartbucket` is aimed at simplifying content management within them.
|
||||
### Listing Buckets
|
||||
|
||||
#### Deleting Buckets
|
||||
Currently, SmartBucket does not include a direct method to list all buckets, but you can access the underlying client provided by the cloud storage SDK to perform such operations, depending on the SDK's capabilities.
|
||||
|
||||
To delete a bucket, simply call the `removeBucket` function:
|
||||
|
||||
```typescript
|
||||
async function deleteBucket(bucketName: string) {
|
||||
try {
|
||||
await mySmartBucket.removeBucket(bucketName);
|
||||
console.log(`Bucket ${bucketName} deleted successfully.`);
|
||||
} catch (error) {
|
||||
console.error("Error deleting bucket:", error);
|
||||
}
|
||||
}
|
||||
|
||||
deleteBucket("anotherBucketName");
|
||||
```
|
||||
|
||||
### File Operations in Buckets
|
||||
|
||||
SmartBucket offers a unified API to execute file-based operations efficiently.
|
||||
### Working with Files
|
||||
|
||||
#### Uploading Files
|
||||
|
||||
Upload a file using the `fastPut` method, specifying the bucket name, file path, and content:
|
||||
To upload an object to a bucket:
|
||||
|
||||
```typescript
|
||||
async function uploadFile(bucketName: string, filePath: string, fileContent: Buffer | string) {
|
||||
const bucket: Bucket = await mySmartBucket.getBucketByName(bucketName);
|
||||
await bucket.fastPut({ path: filePath, contents: fileContent });
|
||||
console.log(`File uploaded to ${filePath}`);
|
||||
const myBucket: Bucket = await mySmartBucket.getBucketByName(bucketName);
|
||||
if (myBucket) {
|
||||
await myBucket.fastPut({ path: filePath, contents: fileContent });
|
||||
console.log(`File uploaded to ${bucketName} at ${filePath}`);
|
||||
} else {
|
||||
console.error(`Bucket ${bucketName} does not exist.`);
|
||||
}
|
||||
}
|
||||
|
||||
uploadFile("myBucket", "example.txt", "This is a sample file content.");
|
||||
// Use the function
|
||||
uploadFile("exampleBucket", "path/to/object.txt", "Hello, world!");
|
||||
```
|
||||
|
||||
#### Downloading Files
|
||||
|
||||
Download files using `fastGet`. It retrieves the file content as a buffer:
|
||||
To download an object:
|
||||
|
||||
```typescript
|
||||
async function downloadFile(bucketName: string, filePath: string) {
|
||||
const bucket: Bucket = await mySmartBucket.getBucketByName(bucketName);
|
||||
const content: Buffer = await bucket.fastGet({ path: filePath });
|
||||
console.log("Downloaded content:", content.toString());
|
||||
const myBucket: Bucket = await mySmartBucket.getBucketByName(bucketName);
|
||||
if (myBucket) {
|
||||
const fileContent: Buffer = await myBucket.fastGet({ path: filePath });
|
||||
console.log("Downloaded file content:", fileContent.toString());
|
||||
} else {
|
||||
console.error(`Bucket ${bucketName} does not exist.`);
|
||||
}
|
||||
}
|
||||
|
||||
downloadFile("myBucket", "example.txt");
|
||||
```
|
||||
|
||||
#### Streaming Files
|
||||
|
||||
For large-scale applications, stream files without loading them fully into memory:
|
||||
|
||||
```typescript
|
||||
async function streamFile(bucketName: string, filePath: string) {
|
||||
const bucket: Bucket = await mySmartBucket.getBucketByName(bucketName);
|
||||
const stream = await bucket.fastGetStream({ path: filePath }, "nodestream");
|
||||
stream.on('data', chunk => console.log("Chunk:", chunk.toString()));
|
||||
stream.on('end', () => console.log("Download completed."));
|
||||
}
|
||||
|
||||
streamFile("myBucket", "largefile.txt");
|
||||
// Use the function
|
||||
downloadFile("exampleBucket", "path/to/object.txt");
|
||||
```
|
||||
|
||||
#### Deleting Files
|
||||
|
||||
Delete files with precision using `fastRemove`:
|
||||
To delete an object from a bucket:
|
||||
|
||||
```typescript
|
||||
async function deleteFile(bucketName: string, filePath: string) {
|
||||
const bucket: Bucket = await mySmartBucket.getBucketByName(bucketName);
|
||||
await bucket.fastRemove({ path: filePath });
|
||||
console.log(`File ${filePath} deleted.`);
|
||||
const myBucket: Bucket = await mySmartBucket.getBucketByName(bucketName);
|
||||
if (myBucket) {
|
||||
await myBucket.fastRemove({ path: filePath });
|
||||
console.log(`File at ${filePath} deleted from ${bucketName}.`);
|
||||
} else {
|
||||
console.error(`Bucket ${bucketName} does not exist.`);
|
||||
}
|
||||
}
|
||||
|
||||
deleteFile("myBucket", "example.txt");
|
||||
// Use the function
|
||||
deleteFile("exampleBucket", "path/to/object.txt");
|
||||
```
|
||||
|
||||
### Directory Operations
|
||||
#### Streaming Files
|
||||
|
||||
Leverage directory functionalities to better organize and manage files within buckets.
|
||||
SmartBucket allows you to work with file streams, which can be useful for handling large files.
|
||||
|
||||
#### Listing Directories and Files
|
||||
|
||||
Listing contents showcases a directory’s structure and file contents:
|
||||
To read a file as a stream:
|
||||
|
||||
```typescript
|
||||
async function listDirectory(bucketName: string, directoryPath: string) {
|
||||
const bucket: Bucket = await mySmartBucket.getBucketByName(bucketName);
|
||||
const baseDirectory: Directory = await bucket.getBaseDirectory();
|
||||
const targetDirectory = await baseDirectory.getSubDirectoryByName(directoryPath);
|
||||
import { ReplaySubject } from '@push.rocks/smartrx';
|
||||
|
||||
console.log('Directories:');
|
||||
(await targetDirectory.listDirectories()).forEach(dir => console.log(dir.name));
|
||||
|
||||
console.log('Files:');
|
||||
(await targetDirectory.listFiles()).forEach(file => console.log(file.name));
|
||||
async function readFileStream(bucketName: string, filePath: string) {
|
||||
const myBucket: Bucket = await mySmartBucket.getBucketByName(bucketName);
|
||||
if (myBucket) {
|
||||
const fileStream: ReplaySubject<Buffer> = await myBucket.fastGetStream({ path: filePath });
|
||||
fileStream.subscribe({
|
||||
next(chunk: Buffer) {
|
||||
console.log("Chunk received:", chunk.toString());
|
||||
},
|
||||
complete() {
|
||||
console.log("File read completed.");
|
||||
},
|
||||
error(err) {
|
||||
console.error("Error reading file stream:", err);
|
||||
}
|
||||
});
|
||||
} else {
|
||||
console.error(`Bucket ${bucketName} does not exist.`);
|
||||
}
|
||||
}
|
||||
|
||||
listDirectory("myBucket", "path/to/directory");
|
||||
// Use the function
|
||||
readFileStream("exampleBucket", "path/to/object.txt");
|
||||
```
|
||||
|
||||
#### Managing Files in Directories
|
||||
|
||||
Additional functionalities allow file management, inclusive of handling sub-directories:
|
||||
To write a file as a stream:
|
||||
|
||||
```typescript
|
||||
async function manageFilesInDirectory(bucketName: string, directoryPath: string, fileName: string, content: string) {
|
||||
const bucket: Bucket = await mySmartBucket.getBucketByName(bucketName);
|
||||
const baseDirectory: Directory = await bucket.getBaseDirectory();
|
||||
const directory = await baseDirectory.getSubDirectoryByName(directoryPath) ?? baseDirectory;
|
||||
import { Readable } from 'stream';
|
||||
|
||||
await directory.fastPut({ path: fileName, contents: content });
|
||||
console.log(`File ${fileName} created in ${directoryPath}`);
|
||||
|
||||
const fileContent = await directory.fastGet({ path: fileName });
|
||||
console.log(`Content of ${fileName}: ${fileContent.toString()}`);
|
||||
async function writeFileStream(bucketName: string, filePath: string, readableStream: Readable) {
|
||||
const myBucket: Bucket = await mySmartBucket.getBucketByName(bucketName);
|
||||
if (myBucket) {
|
||||
await myBucket.fastPutStream({ path: filePath, dataStream: readableStream });
|
||||
console.log(`File streamed to ${bucketName} at ${filePath}`);
|
||||
} else {
|
||||
console.error(`Bucket ${bucketName} does not exist.`);
|
||||
}
|
||||
}
|
||||
|
||||
manageFilesInDirectory("myBucket", "myDir", "example.txt", "File content here");
|
||||
// Create a readable stream from a string
|
||||
const readable = new Readable();
|
||||
readable.push('Hello world streamed as a file!');
|
||||
readable.push(null); // End of stream
|
||||
|
||||
// Use the function
|
||||
writeFileStream("exampleBucket", "path/to/streamedObject.txt", readable);
|
||||
```
|
||||
|
||||
### Working with Directories
|
||||
|
||||
`@push.rocks/smartbucket` offers abstractions for directories within buckets for easier object management. You can create, list, and delete directories using the `Directory` class.
|
||||
|
||||
To list the contents of a directory:
|
||||
|
||||
```typescript
|
||||
async function listDirectoryContents(bucketName: string, directoryPath: string) {
|
||||
const myBucket: Bucket = await mySmartBucket.getBucketByName(bucketName);
|
||||
if (myBucket) {
|
||||
const baseDirectory: Directory = await myBucket.getBaseDirectory();
|
||||
const targetDirectory: Directory = await baseDirectory.getSubDirectoryByName(directoryPath);
|
||||
|
||||
console.log('Listing directories:');
|
||||
const directories = await targetDirectory.listDirectories();
|
||||
directories.forEach(dir => {
|
||||
console.log(`- ${dir.name}`);
|
||||
});
|
||||
|
||||
console.log('Listing files:');
|
||||
const files = await targetDirectory.listFiles();
|
||||
files.forEach(file => {
|
||||
console.log(`- ${file.name}`);
|
||||
});
|
||||
} else {
|
||||
console.error(`Bucket ${bucketName} does not exist.`);
|
||||
}
|
||||
}
|
||||
|
||||
// Use the function
|
||||
listDirectoryContents("exampleBucket", "some/directory/path");
|
||||
```
|
||||
|
||||
To create a file within a directory:
|
||||
|
||||
```typescript
|
||||
async function createFileInDirectory(bucketName: string, directoryPath: string, fileName: string, fileContent: string) {
|
||||
const myBucket: Bucket = await mySmartBucket.getBucketByName(bucketName);
|
||||
if (myBucket) {
|
||||
const baseDirectory: Directory = await myBucket.getBaseDirectory();
|
||||
const targetDirectory: Directory = await baseDirectory.getSubDirectoryByName(directoryPath);
|
||||
await targetDirectory.createEmptyFile(fileName); // Create an empty file
|
||||
const file = new File({ directoryRefArg: targetDirectory, fileName });
|
||||
await file.updateWithContents({ contents: fileContent });
|
||||
console.log(`File created: ${fileName}`);
|
||||
} else {
|
||||
console.error(`Bucket ${bucketName} does not exist.`);
|
||||
}
|
||||
}
|
||||
|
||||
// Use the function
|
||||
createFileInDirectory("exampleBucket", "some/directory", "newfile.txt", "Hello, world!");
|
||||
```
|
||||
|
||||
### Advanced Features
|
||||
|
||||
The library’s advanced features streamline intricate cloud storage workflows.
|
||||
|
||||
#### Bucket Policies
|
||||
|
||||
The library offers tools for maintaining consistent bucket policies across storage providers, assisting in defining access roles and permissions.
|
||||
Manage bucket policies to control access permissions. This feature depends on the policies provided by the storage service (e.g., AWS S3, MinIO).
|
||||
|
||||
#### Metadata Management
|
||||
#### Object Metadata
|
||||
|
||||
Easily manage and store metadata by using the `MetaData` utility:
|
||||
Retrieve and modify object metadata. Metadata can be useful for storing additional information about an object.
|
||||
|
||||
To retrieve metadata:
|
||||
|
||||
```typescript
|
||||
async function handleMetadata(bucketName: string, filePath: string) {
|
||||
const bucket: Bucket = await mySmartBucket.getBucketByName(bucketName);
|
||||
const meta = await bucket.fastStat({ path: filePath });
|
||||
console.log("Metadata:", meta.Metadata);
|
||||
async function getObjectMetadata(bucketName: string, filePath: string) {
|
||||
const myBucket: Bucket = await mySmartBucket.getBucketByName(bucketName);
|
||||
if (myBucket) {
|
||||
const metadata = await mySmartBucket.minioClient.statObject(bucketName, filePath);
|
||||
console.log("Object metadata:", metadata);
|
||||
} else {
|
||||
console.error(`Bucket ${bucketName} does not exist.`);
|
||||
}
|
||||
}
|
||||
|
||||
handleMetadata("myBucket", "example.txt");
|
||||
// Use the function
|
||||
getObjectMetadata("exampleBucket", "path/to/object.txt");
|
||||
```
|
||||
|
||||
#### File Locking
|
||||
|
||||
Prevent accidental writes by locking files:
|
||||
To update metadata:
|
||||
|
||||
```typescript
|
||||
async function lockFile(bucketName: string, filePath: string) {
|
||||
const bucket: Bucket = await mySmartBucket.getBucketByName(bucketName);
|
||||
const file: File = await bucket.getBaseDirectory().getFileStrict({ path: filePath });
|
||||
await file.lock({ timeoutMillis: 600000 }); // Lock for 10 minutes
|
||||
console.log(`File ${filePath} locked.`);
|
||||
async function updateObjectMetadata(bucketName: string, filePath: string, newMetadata: { [key: string]: string }) {
|
||||
const myBucket: Bucket = await mySmartBucket.getBucketByName(bucketName);
|
||||
if (myBucket) {
|
||||
await myBucket.copyObject({
|
||||
objectKey: filePath,
|
||||
nativeMetadata: newMetadata,
|
||||
deleteExistingNativeMetadata: false,
|
||||
});
|
||||
console.log(`Metadata updated for ${filePath}`);
|
||||
} else {
|
||||
console.error(`Bucket ${bucketName} does not exist.`);
|
||||
}
|
||||
}
|
||||
|
||||
lockFile("myBucket", "example.txt");
|
||||
// Use the function
|
||||
updateObjectMetadata("exampleBucket", "path/to/object.txt", {
|
||||
customKey: "customValue"
|
||||
});
|
||||
```
|
||||
|
||||
#### Trash Management
|
||||
#### Cloud Agnostic
|
||||
|
||||
SmartBucket enables a safe deletion mode where files can be moved to a recycling bin, allowing for restoration:
|
||||
`@push.rocks/smartbucket` is designed to work with multiple cloud providers, allowing for easier migration or multi-cloud strategies. This means you can switch from one provider to another with minimal changes to your codebase.
|
||||
|
||||
```typescript
|
||||
async function trashAndRestoreFile(bucketName: string, filePath: string) {
|
||||
const bucket: Bucket = await mySmartBucket.getBucketByName(bucketName);
|
||||
const file: File = await bucket.getBaseDirectory().getFileStrict({ path: filePath });
|
||||
Remember, each cloud provider has specific features and limitations. `@push.rocks/smartbucket` aims to abstract common functionalities, but always refer to the specific cloud provider's documentation for advanced features or limitations.
|
||||
|
||||
// Move the file to trash
|
||||
await file.delete({ mode: 'trash' });
|
||||
console.log(`File ${filePath} moved to trash.`);
|
||||
|
||||
// Retrieve the file from the trash
|
||||
const trashFile = await bucket.getTrash().getTrashedFileByOriginalName({ path: filePath });
|
||||
await trashFile.restore();
|
||||
console.log(`File ${filePath} restored from trash.`);
|
||||
}
|
||||
|
||||
trashAndRestoreFile("myBucket", "example.txt");
|
||||
```
|
||||
|
||||
### Cloud Agnosticism
|
||||
|
||||
`@push.rocks/smartbucket` supports a multitude of cloud providers, enhancing flexibility in adopting different cloud strategies without the need for extensive code rewrite. It offers a uniform interface allowing to perform operations seamlessly between different storage solutions such as AWS S3, Google Cloud Storage, and more. This aspect empowers organizations to align their storage decisions with business needs rather than technical constraints.
|
||||
|
||||
By following this guide, you should be well-equipped to handle cloud storage operations using the `@push.rocks/smartbucket` library. Diligently constructed code examples elucidate the extensive functionalities offered by the library, aligned with best practices in cloud storage. For a deeper dive into any specific feature, refer to the comprehensive documentation provided with the library and the official documentation of the cloud providers you are integrating with.
|
||||
```
|
||||
This guide covers the basic to advanced scenarios of using `@push.rocks/smartbucket`. For further details, refer to the API documentation and examples.
|
||||
|
||||
## License and Legal Information
|
||||
|
||||
|
@ -1,7 +0,0 @@
|
||||
import { tap, expect } from '@push.rocks/tapbundle';
|
||||
|
||||
tap.test('test metadata functionality', async () => {
|
||||
|
||||
})
|
||||
|
||||
tap.start();
|
@ -1,92 +0,0 @@
|
||||
import { expect, expectAsync, tap } from '@push.rocks/tapbundle';
|
||||
import { jestExpect } from '@push.rocks/tapbundle/node';
|
||||
import { Qenv } from '@push.rocks/qenv';
|
||||
|
||||
import * as smartbucket from '../ts/index.js';
|
||||
|
||||
const testQenv = new Qenv('./', './.nogit/');
|
||||
|
||||
let testSmartbucket: smartbucket.SmartBucket;
|
||||
let myBucket: smartbucket.Bucket;
|
||||
let baseDirectory: smartbucket.Directory;
|
||||
|
||||
tap.test('should create a valid smartbucket', async () => {
|
||||
testSmartbucket = new smartbucket.SmartBucket({
|
||||
accessKey: await testQenv.getEnvVarOnDemandStrict('S3_ACCESSKEY'),
|
||||
accessSecret: await testQenv.getEnvVarOnDemandStrict('S3_ACCESSSECRET'),
|
||||
endpoint: await testQenv.getEnvVarOnDemandStrict('S3_ENDPOINT'),
|
||||
});
|
||||
expect(testSmartbucket).toBeInstanceOf(smartbucket.SmartBucket);
|
||||
myBucket = await testSmartbucket.getBucketByNameStrict(await testQenv.getEnvVarOnDemandStrict('S3_BUCKET'),);
|
||||
expect(myBucket).toBeInstanceOf(smartbucket.Bucket);
|
||||
expect(myBucket.name).toEqual('test-pushrocks-smartbucket');
|
||||
});
|
||||
|
||||
tap.test('should clean all contents', async () => {
|
||||
await myBucket.cleanAllContents();
|
||||
expect(await myBucket.fastExists({ path: 'hithere/socool.txt' })).toBeFalse();
|
||||
expect(await myBucket.fastExists({ path: 'trashtest/trashme.txt' })).toBeFalse();
|
||||
});
|
||||
|
||||
tap.test('should delete a file into the normally', async () => {
|
||||
const path = 'trashtest/trashme.txt';
|
||||
const file = await myBucket.fastPutStrict({
|
||||
path,
|
||||
contents: 'I\'m in the trash test content!',
|
||||
});
|
||||
const fileMetadata = await (await file.getMetaData()).metadataFile.getContents();
|
||||
console.log(fileMetadata.toString());
|
||||
expect(await file.getMetaData().then((meta) => meta.metadataFile.getJsonData())).toEqual({});
|
||||
await file.delete({ mode: 'permanent' });
|
||||
expect((await (await myBucket.getBaseDirectory()).listFiles()).length).toEqual(0);
|
||||
expect((await (await myBucket.getBaseDirectory()).listDirectories()).length).toEqual(0);
|
||||
});
|
||||
|
||||
tap.test('should put a file into the trash', async () => {
|
||||
const path = 'trashtest/trashme.txt';
|
||||
const file = await myBucket.fastPutStrict({
|
||||
path,
|
||||
contents: 'I\'m in the trash test content!',
|
||||
});
|
||||
const fileMetadata = await (await file.getMetaData()).metadataFile.getContents();
|
||||
console.log(fileMetadata.toString());
|
||||
expect(await file.getMetaData().then((meta) => meta.metadataFile.getJsonData())).toEqual({});
|
||||
await file.delete({ mode: 'trash' });
|
||||
|
||||
const getTrashContents = async () => {
|
||||
const trash = await myBucket.getTrash();
|
||||
const trashDir = await trash.getTrashDir();
|
||||
return await trashDir.listFiles();
|
||||
}
|
||||
|
||||
const trashedFiles = await getTrashContents();
|
||||
expect(trashedFiles.length).toEqual(2);
|
||||
|
||||
const trashedMetaFile = trashedFiles.find(file => file.name.endsWith('.metadata'));
|
||||
expect(trashedMetaFile).toBeDefined();
|
||||
expect(trashedMetaFile).toBeInstanceOf(smartbucket.File);
|
||||
|
||||
jestExpect(await trashedMetaFile!.getJsonData()).toEqual({
|
||||
custom_recycle: {
|
||||
deletedAt: jestExpect.any(Number),
|
||||
originalPath: "trashtest/trashme.txt",
|
||||
},
|
||||
});
|
||||
});
|
||||
|
||||
tap.test('should restore a file from trash', async () => {
|
||||
const baseDirectory = await myBucket.getBaseDirectory();
|
||||
const file = await baseDirectory.getFileStrict({
|
||||
path: 'trashtest/trashme.txt',
|
||||
getFromTrash: true
|
||||
});
|
||||
const trashFileMeta = await file.getMetaData();
|
||||
const data = await trashFileMeta.getCustomMetaData({
|
||||
key: 'recycle'
|
||||
});
|
||||
expect(file).toBeInstanceOf(smartbucket.File);
|
||||
await file.restore();
|
||||
});
|
||||
|
||||
|
||||
export default tap.start();
|
36
test/test.ts
36
test/test.ts
@ -11,28 +11,24 @@ let baseDirectory: smartbucket.Directory;
|
||||
|
||||
tap.test('should create a valid smartbucket', async () => {
|
||||
testSmartbucket = new smartbucket.SmartBucket({
|
||||
accessKey: await testQenv.getEnvVarOnDemandStrict('S3_ACCESSKEY'),
|
||||
accessSecret: await testQenv.getEnvVarOnDemandStrict('S3_ACCESSSECRET'),
|
||||
endpoint: await testQenv.getEnvVarOnDemandStrict('S3_ENDPOINT'),
|
||||
accessKey: await testQenv.getEnvVarOnDemand('S3_KEY'),
|
||||
accessSecret: await testQenv.getEnvVarOnDemand('S3_SECRET'),
|
||||
endpoint: 's3.eu-central-1.wasabisys.com',
|
||||
});
|
||||
expect(testSmartbucket).toBeInstanceOf(smartbucket.SmartBucket);
|
||||
myBucket = await testSmartbucket.getBucketByNameStrict(await testQenv.getEnvVarOnDemandStrict('S3_BUCKET'),);
|
||||
expect(myBucket).toBeInstanceOf(smartbucket.Bucket);
|
||||
expect(myBucket.name).toEqual('test-pushrocks-smartbucket');
|
||||
});
|
||||
|
||||
tap.test('should clean all contents', async () => {
|
||||
await myBucket.cleanAllContents();
|
||||
expect(await myBucket.fastExists({ path: 'hithere/socool.txt' })).toBeFalse();
|
||||
expect(await myBucket.fastExists({ path: 'trashtest/trashme.txt' })).toBeFalse();
|
||||
});
|
||||
|
||||
tap.skip.test('should create testbucket', async () => {
|
||||
// await testSmartbucket.createBucket('testzone2');
|
||||
// await testSmartbucket.createBucket('testzone');
|
||||
});
|
||||
|
||||
tap.skip.test('should remove testbucket', async () => {
|
||||
// await testSmartbucket.removeBucket('testzone2');
|
||||
// await testSmartbucket.removeBucket('testzone');
|
||||
});
|
||||
|
||||
tap.test('should get a bucket', async () => {
|
||||
myBucket = await testSmartbucket.getBucketByName('testzone');
|
||||
expect(myBucket).toBeInstanceOf(smartbucket.Bucket);
|
||||
expect(myBucket.name).toEqual('testzone');
|
||||
});
|
||||
|
||||
// Fast operations
|
||||
@ -47,12 +43,9 @@ tap.test('should get data in bucket', async () => {
|
||||
const fileString = await myBucket.fastGet({
|
||||
path: 'hithere/socool.txt',
|
||||
});
|
||||
const fileStringStream = await myBucket.fastGetStream(
|
||||
{
|
||||
const fileStringStream = await myBucket.fastGetStream({
|
||||
path: 'hithere/socool.txt',
|
||||
},
|
||||
'nodestream'
|
||||
);
|
||||
}, 'nodestream');
|
||||
console.log(fileString);
|
||||
});
|
||||
|
||||
@ -106,9 +99,8 @@ tap.test('should get base directory', async () => {
|
||||
tap.test('should correctly build paths for sub directories', async () => {
|
||||
const dir4 = await baseDirectory.getSubDirectoryByName('dir3/dir4');
|
||||
expect(dir4).toBeInstanceOf(smartbucket.Directory);
|
||||
const dir4BasePath = dir4?.getBasePath();
|
||||
const dir4BasePath = dir4.getBasePath();
|
||||
console.log(dir4BasePath);
|
||||
expect(dir4BasePath).toEqual('dir3/dir4/');
|
||||
});
|
||||
|
||||
tap.test('clean up directory style tests', async () => {
|
||||
|
@ -1,8 +1,8 @@
|
||||
/**
|
||||
* autocreated commitinfo by @push.rocks/commitinfo
|
||||
* autocreated commitinfo by @pushrocks/commitinfo
|
||||
*/
|
||||
export const commitinfo = {
|
||||
name: '@push.rocks/smartbucket',
|
||||
version: '3.3.7',
|
||||
description: 'A TypeScript library providing a cloud-agnostic interface for managing object storage with functionalities like bucket management, file and directory operations, and advanced features such as metadata handling and file locking.'
|
||||
version: '3.0.14',
|
||||
description: 'A TypeScript library offering simple and cloud-agnostic object storage with advanced features like bucket creation, file and directory management, and data streaming.'
|
||||
}
|
||||
|
@ -1,5 +1,3 @@
|
||||
// classes.bucket.ts
|
||||
|
||||
import * as plugins from './plugins.js';
|
||||
import * as helpers from './helpers.js';
|
||||
import * as interfaces from './interfaces.js';
|
||||
@ -9,35 +7,33 @@ import { File } from './classes.file.js';
|
||||
import { Trash } from './classes.trash.js';
|
||||
|
||||
/**
|
||||
* The bucket class exposes the basic functionality of a bucket.
|
||||
* The bucket class exposes the basc functionality of a bucket.
|
||||
* The functions of the bucket alone are enough to
|
||||
* operate in S3 basic fashion on blobs of data.
|
||||
* operate in s3 basic fashion on blobs of data.
|
||||
*/
|
||||
export class Bucket {
|
||||
public static async getBucketByName(smartbucketRef: SmartBucket, bucketNameArg: string) {
|
||||
const command = new plugins.s3.ListBucketsCommand({});
|
||||
const buckets = await smartbucketRef.s3Client.send(command);
|
||||
const foundBucket = buckets.Buckets!.find((bucket) => bucket.Name === bucketNameArg);
|
||||
const buckets = await smartbucketRef.minioClient.listBuckets();
|
||||
const foundBucket = buckets.find((bucket) => {
|
||||
return bucket.name === bucketNameArg;
|
||||
});
|
||||
|
||||
if (foundBucket) {
|
||||
console.log(`bucket with name ${bucketNameArg} exists.`);
|
||||
console.log(`Taking this as base for new Bucket instance`);
|
||||
return new this(smartbucketRef, bucketNameArg);
|
||||
} else {
|
||||
console.log(`did not find bucket by name: ${bucketNameArg}`);
|
||||
return null;
|
||||
}
|
||||
}
|
||||
|
||||
public static async createBucketByName(smartbucketRef: SmartBucket, bucketName: string) {
|
||||
const command = new plugins.s3.CreateBucketCommand({ Bucket: bucketName });
|
||||
await smartbucketRef.s3Client.send(command).catch((e) => console.log(e));
|
||||
await smartbucketRef.minioClient.makeBucket(bucketName, 'ams3').catch((e) => console.log(e));
|
||||
return new Bucket(smartbucketRef, bucketName);
|
||||
}
|
||||
|
||||
public static async removeBucketByName(smartbucketRef: SmartBucket, bucketName: string) {
|
||||
const command = new plugins.s3.DeleteBucketCommand({ Bucket: bucketName });
|
||||
await smartbucketRef.s3Client.send(command).catch((e) => console.log(e));
|
||||
await smartbucketRef.minioClient.removeBucket(bucketName).catch((e) => console.log(e));
|
||||
}
|
||||
|
||||
public smartbucketRef: SmartBucket;
|
||||
@ -52,7 +48,7 @@ export class Bucket {
|
||||
* gets the base directory of the bucket
|
||||
*/
|
||||
public async getBaseDirectory(): Promise<Directory> {
|
||||
return new Directory(this, null!, '');
|
||||
return new Directory(this, null, '');
|
||||
}
|
||||
|
||||
/**
|
||||
@ -69,11 +65,9 @@ export class Bucket {
|
||||
if (!pathDescriptorArg.path && !pathDescriptorArg.directory) {
|
||||
return this.getBaseDirectory();
|
||||
}
|
||||
const checkPath = await helpers.reducePathDescriptorToPath(pathDescriptorArg);
|
||||
let checkPath = await helpers.reducePathDescriptorToPath(pathDescriptorArg);
|
||||
const baseDirectory = await this.getBaseDirectory();
|
||||
return await baseDirectory.getSubDirectoryByNameStrict(checkPath, {
|
||||
getEmptyDirectory: true,
|
||||
});
|
||||
return await baseDirectory.getSubDirectoryByName(checkPath);
|
||||
}
|
||||
|
||||
// ===============
|
||||
@ -83,20 +77,18 @@ export class Bucket {
|
||||
/**
|
||||
* store file
|
||||
*/
|
||||
public async fastPut(
|
||||
optionsArg: interfaces.IPathDecriptor & {
|
||||
public async fastPut(optionsArg: interfaces.IPathDecriptor & {
|
||||
contents: string | Buffer;
|
||||
overwrite?: boolean;
|
||||
}
|
||||
): Promise<File | null> {
|
||||
}): Promise<File> {
|
||||
try {
|
||||
const reducedPath = await helpers.reducePathDescriptorToPath(optionsArg);
|
||||
// Check if the object already exists
|
||||
const exists = await this.fastExists({ path: reducedPath });
|
||||
|
||||
if (exists && !optionsArg.overwrite) {
|
||||
const errorText = `Object already exists at path '${reducedPath}' in bucket '${this.name}'.`;
|
||||
console.error(errorText);
|
||||
return null;
|
||||
console.error(`Object already exists at path '${reducedPath}' in bucket '${this.name}'.`);
|
||||
return;
|
||||
} else if (exists && optionsArg.overwrite) {
|
||||
console.log(
|
||||
`Overwriting existing object at path '${reducedPath}' in bucket '${this.name}'.`
|
||||
@ -105,12 +97,16 @@ export class Bucket {
|
||||
console.log(`Creating new object at path '${reducedPath}' in bucket '${this.name}'.`);
|
||||
}
|
||||
|
||||
const command = new plugins.s3.PutObjectCommand({
|
||||
Bucket: this.name,
|
||||
Key: reducedPath,
|
||||
Body: optionsArg.contents,
|
||||
});
|
||||
await this.smartbucketRef.s3Client.send(command);
|
||||
// Proceed with putting the object
|
||||
const streamIntake = new plugins.smartstream.StreamIntake();
|
||||
const putPromise = this.smartbucketRef.minioClient.putObject(
|
||||
this.name,
|
||||
reducedPath,
|
||||
streamIntake
|
||||
);
|
||||
streamIntake.pushData(optionsArg.contents);
|
||||
streamIntake.signalEnd();
|
||||
await putPromise;
|
||||
|
||||
console.log(`Object '${reducedPath}' has been successfully stored in bucket '${this.name}'.`);
|
||||
const parsedPath = plugins.path.parse(reducedPath);
|
||||
@ -129,14 +125,6 @@ export class Bucket {
|
||||
}
|
||||
}
|
||||
|
||||
public async fastPutStrict(...args: Parameters<Bucket['fastPut']>) {
|
||||
const file = await this.fastPut(...args);
|
||||
if (!file) {
|
||||
throw new Error(`File not stored at path '${args[0].path}'`);
|
||||
}
|
||||
return file;
|
||||
}
|
||||
|
||||
/**
|
||||
* get file
|
||||
*/
|
||||
@ -161,7 +149,7 @@ export class Bucket {
|
||||
},
|
||||
});
|
||||
await done.promise;
|
||||
return completeFile!;
|
||||
return completeFile;
|
||||
}
|
||||
|
||||
/**
|
||||
@ -173,16 +161,10 @@ export class Bucket {
|
||||
public async fastGetReplaySubject(optionsArg: {
|
||||
path: string;
|
||||
}): Promise<plugins.smartrx.rxjs.ReplaySubject<Buffer>> {
|
||||
const command = new plugins.s3.GetObjectCommand({
|
||||
Bucket: this.name,
|
||||
Key: optionsArg.path,
|
||||
});
|
||||
const response = await this.smartbucketRef.s3Client.send(command);
|
||||
const fileStream = await this.smartbucketRef.minioClient
|
||||
.getObject(this.name, optionsArg.path)
|
||||
.catch((e) => console.log(e));
|
||||
const replaySubject = new plugins.smartrx.rxjs.ReplaySubject<Buffer>();
|
||||
|
||||
// Convert the stream to a format that supports piping
|
||||
const stream = response.Body as any; // SdkStreamMixin includes readable stream
|
||||
if (typeof stream.pipe === 'function') {
|
||||
const duplexStream = new plugins.smartstream.SmartDuplex<Buffer, void>({
|
||||
writeFunction: async (chunk) => {
|
||||
replaySubject.next(chunk);
|
||||
@ -194,9 +176,12 @@ export class Bucket {
|
||||
},
|
||||
});
|
||||
|
||||
stream.pipe(duplexStream);
|
||||
if (!fileStream) {
|
||||
return null;
|
||||
}
|
||||
|
||||
const smartstream = new plugins.smartstream.StreamWrapper([fileStream, duplexStream]);
|
||||
smartstream.run();
|
||||
return replaySubject;
|
||||
}
|
||||
|
||||
@ -213,37 +198,39 @@ export class Bucket {
|
||||
typeArg: 'nodestream'
|
||||
): Promise<plugins.stream.Readable>;
|
||||
|
||||
/**
|
||||
* fastGetStream
|
||||
* @param optionsArg
|
||||
* @returns
|
||||
*/
|
||||
public async fastGetStream(
|
||||
optionsArg: { path: string },
|
||||
typeArg: 'webstream' | 'nodestream' = 'nodestream'
|
||||
): Promise<ReadableStream | plugins.stream.Readable> {
|
||||
const command = new plugins.s3.GetObjectCommand({
|
||||
Bucket: this.name,
|
||||
Key: optionsArg.path,
|
||||
});
|
||||
const response = await this.smartbucketRef.s3Client.send(command);
|
||||
const stream = response.Body as any; // SdkStreamMixin includes readable stream
|
||||
|
||||
const fileStream = await this.smartbucketRef.minioClient
|
||||
.getObject(this.name, optionsArg.path)
|
||||
.catch((e) => console.log(e));
|
||||
const duplexStream = new plugins.smartstream.SmartDuplex<Buffer, Buffer>({
|
||||
writeFunction: async (chunk) => {
|
||||
return chunk;
|
||||
},
|
||||
finalFunction: async (cb) => {
|
||||
return null!;
|
||||
return null;
|
||||
},
|
||||
});
|
||||
|
||||
if (typeof stream.pipe === 'function') {
|
||||
stream.pipe(duplexStream);
|
||||
if (!fileStream) {
|
||||
return null;
|
||||
}
|
||||
|
||||
const smartstream = new plugins.smartstream.StreamWrapper([fileStream, duplexStream]);
|
||||
smartstream.run();
|
||||
if (typeArg === 'nodestream') {
|
||||
return duplexStream;
|
||||
}
|
||||
if (typeArg === 'webstream') {
|
||||
return (await duplexStream.getWebStreams()).readable;
|
||||
}
|
||||
throw new Error('unknown typeArg');
|
||||
}
|
||||
|
||||
/**
|
||||
@ -256,6 +243,7 @@ export class Bucket {
|
||||
overwrite?: boolean;
|
||||
}): Promise<void> {
|
||||
try {
|
||||
// Check if the object already exists
|
||||
const exists = await this.fastExists({ path: optionsArg.path });
|
||||
|
||||
if (exists && !optionsArg.overwrite) {
|
||||
@ -271,13 +259,18 @@ export class Bucket {
|
||||
console.log(`Creating new object at path '${optionsArg.path}' in bucket '${this.name}'.`);
|
||||
}
|
||||
|
||||
const command = new plugins.s3.PutObjectCommand({
|
||||
Bucket: this.name,
|
||||
Key: optionsArg.path,
|
||||
Body: optionsArg.readableStream,
|
||||
Metadata: optionsArg.nativeMetadata,
|
||||
});
|
||||
await this.smartbucketRef.s3Client.send(command);
|
||||
const streamIntake = await plugins.smartstream.StreamIntake.fromStream<Uint8Array>(
|
||||
optionsArg.readableStream
|
||||
);
|
||||
|
||||
// Proceed with putting the object
|
||||
await this.smartbucketRef.minioClient.putObject(
|
||||
this.name,
|
||||
optionsArg.path,
|
||||
streamIntake,
|
||||
null,
|
||||
null // TODO: Add support for custom metadata once proper support is in minio.
|
||||
);
|
||||
|
||||
console.log(
|
||||
`Object '${optionsArg.path}' has been successfully stored in bucket '${this.name}'.`
|
||||
@ -302,29 +295,28 @@ export class Bucket {
|
||||
const targetBucketName = optionsArg.targetBucket ? optionsArg.targetBucket.name : this.name;
|
||||
|
||||
// Retrieve current object information to use in copy conditions
|
||||
const currentObjInfo = await this.smartbucketRef.s3Client.send(
|
||||
new plugins.s3.HeadObjectCommand({
|
||||
Bucket: this.name,
|
||||
Key: optionsArg.sourcePath,
|
||||
})
|
||||
const currentObjInfo = await this.smartbucketRef.minioClient.statObject(
|
||||
targetBucketName,
|
||||
optionsArg.sourcePath
|
||||
);
|
||||
|
||||
// Setting up copy conditions
|
||||
const copyConditions = new plugins.minio.CopyConditions();
|
||||
|
||||
// Prepare new metadata
|
||||
const newNativeMetadata = {
|
||||
...(optionsArg.deleteExistingNativeMetadata ? {} : currentObjInfo.Metadata),
|
||||
...(optionsArg.deleteExistingNativeMetadata ? {} : currentObjInfo.metaData),
|
||||
...optionsArg.nativeMetadata,
|
||||
};
|
||||
|
||||
// Define the copy operation
|
||||
const copySource = `${this.name}/${optionsArg.sourcePath}`;
|
||||
const command = new plugins.s3.CopyObjectCommand({
|
||||
Bucket: targetBucketName,
|
||||
CopySource: copySource,
|
||||
Key: optionsArg.destinationPath || optionsArg.sourcePath,
|
||||
Metadata: newNativeMetadata,
|
||||
MetadataDirective: optionsArg.deleteExistingNativeMetadata ? 'REPLACE' : 'COPY',
|
||||
});
|
||||
await this.smartbucketRef.s3Client.send(command);
|
||||
// Define the copy operation as a Promise
|
||||
// TODO: check on issue here: https://github.com/minio/minio-js/issues/1286
|
||||
await this.smartbucketRef.minioClient.copyObject(
|
||||
this.name,
|
||||
optionsArg.sourcePath,
|
||||
`/${targetBucketName}/${optionsArg.destinationPath || optionsArg.sourcePath}`,
|
||||
copyConditions
|
||||
);
|
||||
} catch (err) {
|
||||
console.error('Error updating metadata:', err);
|
||||
throw err; // rethrow to allow caller to handle
|
||||
@ -341,10 +333,9 @@ export class Bucket {
|
||||
overwrite?: boolean;
|
||||
}): Promise<void> {
|
||||
try {
|
||||
// Check if the destination object already exists
|
||||
const destinationBucket = optionsArg.targetBucket || this;
|
||||
const exists = await destinationBucket.fastExists({
|
||||
path: optionsArg.destinationPath,
|
||||
});
|
||||
const exists = await destinationBucket.fastExists({ path: optionsArg.destinationPath });
|
||||
|
||||
if (exists && !optionsArg.overwrite) {
|
||||
console.error(
|
||||
@ -361,7 +352,10 @@ export class Bucket {
|
||||
);
|
||||
}
|
||||
|
||||
// Proceed with copying the object to the new path
|
||||
await this.fastCopy(optionsArg);
|
||||
|
||||
// Remove the original object after successful copy
|
||||
await this.fastRemove({ path: optionsArg.sourcePath });
|
||||
|
||||
console.log(
|
||||
@ -380,29 +374,21 @@ export class Bucket {
|
||||
* removeObject
|
||||
*/
|
||||
public async fastRemove(optionsArg: { path: string }) {
|
||||
const command = new plugins.s3.DeleteObjectCommand({
|
||||
Bucket: this.name,
|
||||
Key: optionsArg.path,
|
||||
});
|
||||
await this.smartbucketRef.s3Client.send(command);
|
||||
await this.smartbucketRef.minioClient.removeObject(this.name, optionsArg.path);
|
||||
}
|
||||
|
||||
/**
|
||||
* check whether file exists
|
||||
* check wether file exists
|
||||
* @param optionsArg
|
||||
* @returns
|
||||
*/
|
||||
public async fastExists(optionsArg: { path: string }): Promise<boolean> {
|
||||
try {
|
||||
const command = new plugins.s3.HeadObjectCommand({
|
||||
Bucket: this.name,
|
||||
Key: optionsArg.path,
|
||||
});
|
||||
await this.smartbucketRef.s3Client.send(command);
|
||||
await this.smartbucketRef.minioClient.statObject(this.name, optionsArg.path);
|
||||
console.log(`Object '${optionsArg.path}' exists in bucket '${this.name}'.`);
|
||||
return true;
|
||||
} catch (error: any) {
|
||||
if (error?.name === 'NotFound') {
|
||||
} catch (error) {
|
||||
if (error.code === 'NotFound') {
|
||||
console.log(`Object '${optionsArg.path}' does not exist in bucket '${this.name}'.`);
|
||||
return false;
|
||||
} else {
|
||||
@ -416,111 +402,59 @@ export class Bucket {
|
||||
* deletes this bucket
|
||||
*/
|
||||
public async delete() {
|
||||
await this.smartbucketRef.s3Client.send(
|
||||
new plugins.s3.DeleteBucketCommand({ Bucket: this.name })
|
||||
);
|
||||
await this.smartbucketRef.minioClient.removeBucket(this.name);
|
||||
}
|
||||
|
||||
public async fastStat(pathDescriptor: interfaces.IPathDecriptor) {
|
||||
const checkPath = await helpers.reducePathDescriptorToPath(pathDescriptor);
|
||||
const command = new plugins.s3.HeadObjectCommand({
|
||||
Bucket: this.name,
|
||||
Key: checkPath,
|
||||
});
|
||||
return this.smartbucketRef.s3Client.send(command);
|
||||
let checkPath = await helpers.reducePathDescriptorToPath(pathDescriptor);
|
||||
return this.smartbucketRef.minioClient.statObject(this.name, checkPath);
|
||||
}
|
||||
|
||||
public async isDirectory(pathDescriptor: interfaces.IPathDecriptor): Promise<boolean> {
|
||||
const checkPath = await helpers.reducePathDescriptorToPath(pathDescriptor);
|
||||
const command = new plugins.s3.ListObjectsV2Command({
|
||||
Bucket: this.name,
|
||||
Prefix: checkPath,
|
||||
Delimiter: '/',
|
||||
let checkPath = await helpers.reducePathDescriptorToPath(pathDescriptor);
|
||||
|
||||
// lets check if the checkPath is a directory
|
||||
const stream = this.smartbucketRef.minioClient.listObjectsV2(this.name, checkPath, true);
|
||||
const done = plugins.smartpromise.defer<boolean>();
|
||||
stream.on('data', (dataArg) => {
|
||||
stream.destroy(); // Stop the stream early if we find at least one object
|
||||
if (dataArg.prefix.startsWith(checkPath + '/')) {
|
||||
done.resolve(true);
|
||||
}
|
||||
});
|
||||
const { CommonPrefixes } = await this.smartbucketRef.s3Client.send(command);
|
||||
return !!CommonPrefixes && CommonPrefixes.length > 0;
|
||||
|
||||
stream.on('end', () => {
|
||||
done.resolve(false);
|
||||
});
|
||||
|
||||
stream.on('error', (err) => {
|
||||
done.reject(err);
|
||||
});
|
||||
|
||||
return done.promise;
|
||||
}
|
||||
|
||||
public async isFile(pathDescriptor: interfaces.IPathDecriptor): Promise<boolean> {
|
||||
const checkPath = await helpers.reducePathDescriptorToPath(pathDescriptor);
|
||||
const command = new plugins.s3.ListObjectsV2Command({
|
||||
Bucket: this.name,
|
||||
Prefix: checkPath,
|
||||
Delimiter: '/',
|
||||
});
|
||||
const { Contents } = await this.smartbucketRef.s3Client.send(command);
|
||||
return !!Contents && Contents.length > 0;
|
||||
let checkPath = await helpers.reducePathDescriptorToPath(pathDescriptor);
|
||||
|
||||
// lets check if the checkPath is a directory
|
||||
const stream = this.smartbucketRef.minioClient.listObjectsV2(this.name, checkPath, true);
|
||||
const done = plugins.smartpromise.defer<boolean>();
|
||||
stream.on('data', (dataArg) => {
|
||||
stream.destroy(); // Stop the stream early if we find at least one object
|
||||
if (dataArg.prefix === checkPath) {
|
||||
done.resolve(true);
|
||||
}
|
||||
|
||||
public async getMagicBytes(optionsArg: { path: string; length: number }): Promise<Buffer> {
|
||||
try {
|
||||
const command = new plugins.s3.GetObjectCommand({
|
||||
Bucket: this.name,
|
||||
Key: optionsArg.path,
|
||||
Range: `bytes=0-${optionsArg.length - 1}`,
|
||||
});
|
||||
const response = await this.smartbucketRef.s3Client.send(command);
|
||||
const chunks = [];
|
||||
const stream = response.Body as any; // SdkStreamMixin includes readable stream
|
||||
|
||||
for await (const chunk of stream) {
|
||||
chunks.push(chunk);
|
||||
}
|
||||
return Buffer.concat(chunks);
|
||||
} catch (error) {
|
||||
console.error(
|
||||
`Error retrieving magic bytes from object at path '${optionsArg.path}' in bucket '${this.name}':`,
|
||||
error
|
||||
);
|
||||
throw error;
|
||||
}
|
||||
}
|
||||
|
||||
public async cleanAllContents(): Promise<void> {
|
||||
try {
|
||||
// Define the command type explicitly
|
||||
const listCommandInput: plugins.s3.ListObjectsV2CommandInput = {
|
||||
Bucket: this.name,
|
||||
};
|
||||
|
||||
let isTruncated = true;
|
||||
let continuationToken: string | undefined = undefined;
|
||||
|
||||
while (isTruncated) {
|
||||
// Add the continuation token to the input if present
|
||||
const listCommand = new plugins.s3.ListObjectsV2Command({
|
||||
...listCommandInput,
|
||||
ContinuationToken: continuationToken,
|
||||
});
|
||||
|
||||
// Explicitly type the response
|
||||
const response: plugins.s3.ListObjectsV2Output =
|
||||
await this.smartbucketRef.s3Client.send(listCommand);
|
||||
|
||||
console.log(`Cleaning contents of bucket '${this.name}': Now deleting ${response.Contents?.length} items...`);
|
||||
|
||||
if (response.Contents && response.Contents.length > 0) {
|
||||
// Delete objects in batches, mapping each item to { Key: string }
|
||||
const deleteCommand = new plugins.s3.DeleteObjectsCommand({
|
||||
Bucket: this.name,
|
||||
Delete: {
|
||||
Objects: response.Contents.map((item) => ({ Key: item.Key! })),
|
||||
Quiet: true,
|
||||
},
|
||||
stream.on('end', () => {
|
||||
done.resolve(false);
|
||||
});
|
||||
|
||||
await this.smartbucketRef.s3Client.send(deleteCommand);
|
||||
}
|
||||
stream.on('error', (err) => {
|
||||
done.reject(err);
|
||||
});
|
||||
|
||||
// Update continuation token and truncation status
|
||||
isTruncated = response.IsTruncated || false;
|
||||
continuationToken = response.NextContinuationToken;
|
||||
}
|
||||
|
||||
console.log(`All contents in bucket '${this.name}' have been deleted.`);
|
||||
} catch (error) {
|
||||
console.error(`Error cleaning contents of bucket '${this.name}':`, error);
|
||||
throw error;
|
||||
}
|
||||
return done.promise;
|
||||
}
|
||||
}
|
||||
|
@ -1,8 +1,7 @@
|
||||
// classes.directory.ts
|
||||
|
||||
import * as plugins from './plugins.js';
|
||||
import { Bucket } from './classes.bucket.js';
|
||||
import { File } from './classes.file.js';
|
||||
|
||||
import * as helpers from './helpers.js';
|
||||
|
||||
export class Directory {
|
||||
@ -10,13 +9,13 @@ export class Directory {
|
||||
public parentDirectoryRef: Directory;
|
||||
public name: string;
|
||||
|
||||
public tree!: string[];
|
||||
public files!: string[];
|
||||
public folders!: string[];
|
||||
public tree: string[];
|
||||
public files: string[];
|
||||
public folders: string[];
|
||||
|
||||
constructor(bucketRefArg: Bucket, parentDirectory: Directory, name: string) {
|
||||
constructor(bucketRefArg: Bucket, parentDiretory: Directory, name: string) {
|
||||
this.bucketRef = bucketRefArg;
|
||||
this.parentDirectoryRef = parentDirectory;
|
||||
this.parentDirectoryRef = parentDiretory;
|
||||
this.name = name;
|
||||
}
|
||||
|
||||
@ -66,20 +65,23 @@ export class Directory {
|
||||
* gets a file by name
|
||||
*/
|
||||
public async getFile(optionsArg: {
|
||||
path: string;
|
||||
name: string;
|
||||
createWithContents?: string | Buffer;
|
||||
getFromTrash?: boolean;
|
||||
}): Promise<File | null> {
|
||||
}): Promise<File> {
|
||||
const pathDescriptor = {
|
||||
directory: this,
|
||||
path: optionsArg.path,
|
||||
path: optionsArg.name,
|
||||
};
|
||||
// check wether the file exists
|
||||
const exists = await this.bucketRef.fastExists({
|
||||
path: await helpers.reducePathDescriptorToPath(pathDescriptor),
|
||||
});
|
||||
if (!exists && optionsArg.getFromTrash) {
|
||||
const trash = await this.bucketRef.getTrash();
|
||||
const trashedFile = await trash.getTrashedFileByOriginalName(pathDescriptor);
|
||||
const trashedFile = await trash.getTrashedFileByOriginalName(
|
||||
pathDescriptor
|
||||
)
|
||||
return trashedFile;
|
||||
}
|
||||
if (!exists && !optionsArg.createWithContents) {
|
||||
@ -88,44 +90,40 @@ export class Directory {
|
||||
if (!exists && optionsArg.createWithContents) {
|
||||
await File.create({
|
||||
directory: this,
|
||||
name: optionsArg.path,
|
||||
name: optionsArg.name,
|
||||
contents: optionsArg.createWithContents,
|
||||
});
|
||||
}
|
||||
return new File({
|
||||
directoryRefArg: this,
|
||||
fileName: optionsArg.path,
|
||||
fileName: optionsArg.name,
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* gets a file strictly
|
||||
* @param args
|
||||
* @returns
|
||||
*/
|
||||
public async getFileStrict(...args: Parameters<Directory['getFile']>) {
|
||||
const file = await this.getFile(...args);
|
||||
if (!file) {
|
||||
throw new Error(`File not found at path '${args[0].path}'`);
|
||||
}
|
||||
return file;
|
||||
}
|
||||
|
||||
/**
|
||||
* lists all files
|
||||
*/
|
||||
public async listFiles(): Promise<File[]> {
|
||||
const command = new plugins.s3.ListObjectsV2Command({
|
||||
Bucket: this.bucketRef.name,
|
||||
Prefix: this.getBasePath(),
|
||||
Delimiter: '/',
|
||||
});
|
||||
const response = await this.bucketRef.smartbucketRef.s3Client.send(command);
|
||||
const done = plugins.smartpromise.defer();
|
||||
const fileNameStream = await this.bucketRef.smartbucketRef.minioClient.listObjectsV2(
|
||||
this.bucketRef.name,
|
||||
this.getBasePath(),
|
||||
false
|
||||
);
|
||||
const fileArray: File[] = [];
|
||||
|
||||
response.Contents?.forEach((item) => {
|
||||
if (item.Key && !item.Key.endsWith('/')) {
|
||||
const subtractedPath = item.Key.replace(this.getBasePath(), '');
|
||||
const duplexStream = new plugins.smartstream.SmartDuplex<plugins.minio.BucketItem, void>({
|
||||
objectMode: true,
|
||||
writeFunction: async (bucketItem) => {
|
||||
if (bucketItem.prefix) {
|
||||
return;
|
||||
}
|
||||
if (!bucketItem.name) {
|
||||
return;
|
||||
}
|
||||
let subtractedPath = bucketItem.name.replace(this.getBasePath(), '');
|
||||
if (subtractedPath.startsWith('/')) {
|
||||
subtractedPath = subtractedPath.substr(1);
|
||||
}
|
||||
if (!subtractedPath.includes('/')) {
|
||||
fileArray.push(
|
||||
new File({
|
||||
@ -134,9 +132,13 @@ export class Directory {
|
||||
})
|
||||
);
|
||||
}
|
||||
}
|
||||
},
|
||||
finalFunction: async (tools) => {
|
||||
done.resolve();
|
||||
},
|
||||
});
|
||||
|
||||
fileNameStream.pipe(duplexStream);
|
||||
await done.promise;
|
||||
return fileArray;
|
||||
}
|
||||
|
||||
@ -144,124 +146,70 @@ export class Directory {
|
||||
* lists all folders
|
||||
*/
|
||||
public async listDirectories(): Promise<Directory[]> {
|
||||
try {
|
||||
const command = new plugins.s3.ListObjectsV2Command({
|
||||
Bucket: this.bucketRef.name,
|
||||
Prefix: this.getBasePath(),
|
||||
Delimiter: '/',
|
||||
});
|
||||
const response = await this.bucketRef.smartbucketRef.s3Client.send(command);
|
||||
const done = plugins.smartpromise.defer();
|
||||
const basePath = this.getBasePath();
|
||||
const completeDirStream = await this.bucketRef.smartbucketRef.minioClient.listObjectsV2(
|
||||
this.bucketRef.name,
|
||||
this.getBasePath(),
|
||||
false
|
||||
);
|
||||
const directoryArray: Directory[] = [];
|
||||
|
||||
if (response.CommonPrefixes) {
|
||||
response.CommonPrefixes.forEach((item) => {
|
||||
if (item.Prefix) {
|
||||
const subtractedPath = item.Prefix.replace(this.getBasePath(), '');
|
||||
if (subtractedPath.endsWith('/')) {
|
||||
const dirName = subtractedPath.slice(0, -1);
|
||||
// Ensure the directory name is not empty (which would indicate the base directory itself)
|
||||
if (dirName) {
|
||||
const duplexStream = new plugins.smartstream.SmartDuplex<plugins.minio.BucketItem, void>({
|
||||
objectMode: true,
|
||||
writeFunction: async (bucketItem) => {
|
||||
if (bucketItem.name) {
|
||||
return;
|
||||
}
|
||||
let subtractedPath = bucketItem.prefix.replace(this.getBasePath(), '');
|
||||
if (subtractedPath.startsWith('/')) {
|
||||
subtractedPath = subtractedPath.substr(1);
|
||||
}
|
||||
if (subtractedPath.includes('/')) {
|
||||
const dirName = subtractedPath.split('/')[0];
|
||||
if (directoryArray.find((directory) => directory.name === dirName)) {
|
||||
return;
|
||||
}
|
||||
directoryArray.push(new Directory(this.bucketRef, this, dirName));
|
||||
}
|
||||
}
|
||||
}
|
||||
},
|
||||
finalFunction: async (tools) => {
|
||||
done.resolve();
|
||||
},
|
||||
});
|
||||
}
|
||||
|
||||
completeDirStream.pipe(duplexStream);
|
||||
await done.promise;
|
||||
return directoryArray;
|
||||
} catch (error) {
|
||||
console.error('Error listing directories:', error);
|
||||
throw error;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* gets an array that has all objects with a certain prefix
|
||||
* gets an array that has all objects with a certain prefix;
|
||||
*/
|
||||
public async getTreeArray() {
|
||||
const command = new plugins.s3.ListObjectsV2Command({
|
||||
Bucket: this.bucketRef.name,
|
||||
Prefix: this.getBasePath(),
|
||||
Delimiter: '/',
|
||||
});
|
||||
const response = await this.bucketRef.smartbucketRef.s3Client.send(command);
|
||||
return response.Contents;
|
||||
const treeArray = await this.bucketRef.smartbucketRef.minioClient.listObjectsV2(
|
||||
this.bucketRef.name,
|
||||
this.getBasePath(),
|
||||
true
|
||||
);
|
||||
}
|
||||
|
||||
/**
|
||||
* gets a sub directory by name
|
||||
* gets a sub directory
|
||||
*/
|
||||
public async getSubDirectoryByName(dirNameArg: string, optionsArg: {
|
||||
/**
|
||||
* in s3 a directory does not exist if it is empty
|
||||
* this option returns a directory even if it is empty
|
||||
*/
|
||||
getEmptyDirectory?: boolean;
|
||||
/**
|
||||
* in s3 a directory does not exist if it is empty
|
||||
* this option creates a directory even if it is empty using a initializer file
|
||||
*/
|
||||
createWithInitializerFile?: boolean;
|
||||
/**
|
||||
* if the path is a file path, it will be treated as a file and the parent directory will be returned
|
||||
*/
|
||||
couldBeFilePath?: boolean;
|
||||
} = {}): Promise<Directory | null> {
|
||||
public async getSubDirectoryByName(dirNameArg: string): Promise<Directory> {
|
||||
const dirNameArray = dirNameArg.split('/');
|
||||
|
||||
const dirNameArray = dirNameArg.split('/').filter(str => str.trim() !== "");
|
||||
|
||||
optionsArg = {
|
||||
getEmptyDirectory: false,
|
||||
createWithInitializerFile: false,
|
||||
...optionsArg,
|
||||
}
|
||||
|
||||
|
||||
const getDirectory = async (directoryArg: Directory, dirNameToSearch: string, isFinalDirectory: boolean) => {
|
||||
const getDirectory = async (directoryArg: Directory, dirNameToSearch: string) => {
|
||||
const directories = await directoryArg.listDirectories();
|
||||
let returnDirectory = directories.find((directory) => {
|
||||
return directories.find((directory) => {
|
||||
return directory.name === dirNameToSearch;
|
||||
});
|
||||
if (returnDirectory) {
|
||||
return returnDirectory;
|
||||
}
|
||||
if (optionsArg.getEmptyDirectory || optionsArg.createWithInitializerFile) {
|
||||
returnDirectory = new Directory(this.bucketRef, this, dirNameToSearch);
|
||||
}
|
||||
if (isFinalDirectory && optionsArg.createWithInitializerFile) {
|
||||
returnDirectory?.createEmptyFile('00init.txt');
|
||||
}
|
||||
return returnDirectory || null;
|
||||
};
|
||||
|
||||
if (optionsArg.couldBeFilePath) {
|
||||
const baseDirectory = await this.bucketRef.getBaseDirectory();
|
||||
const existingFile = await baseDirectory.getFile({
|
||||
path: dirNameArg,
|
||||
});
|
||||
if (existingFile) {
|
||||
const adjustedPath = dirNameArg.substring(0, dirNameArg.lastIndexOf('/'));
|
||||
return this.getSubDirectoryByName(adjustedPath);
|
||||
}
|
||||
}
|
||||
|
||||
let wantedDirectory: Directory | null = null;
|
||||
let counter = 0;
|
||||
let wantedDirectory: Directory;
|
||||
for (const dirNameToSearch of dirNameArray) {
|
||||
counter++;
|
||||
const directoryToSearchIn = wantedDirectory ? wantedDirectory : this;
|
||||
wantedDirectory = await getDirectory(directoryToSearchIn, dirNameToSearch, counter === dirNameArray.length);
|
||||
wantedDirectory = await getDirectory(directoryToSearchIn, dirNameToSearch);
|
||||
}
|
||||
|
||||
return wantedDirectory || null;
|
||||
}
|
||||
|
||||
public async getSubDirectoryByNameStrict(...args: Parameters<Directory['getSubDirectoryByName']>) {
|
||||
const directory = await this.getSubDirectoryByName(...args);
|
||||
if (!directory) {
|
||||
throw new Error(`Directory not found at path '${args[0]}'`);
|
||||
}
|
||||
return directory;
|
||||
return wantedDirectory;
|
||||
}
|
||||
|
||||
/**
|
||||
@ -269,20 +217,19 @@ export class Directory {
|
||||
*/
|
||||
public async move() {
|
||||
// TODO
|
||||
throw new Error('Moving a directory is not yet implemented');
|
||||
throw new Error('moving a directory is not yet implemented');
|
||||
}
|
||||
|
||||
/**
|
||||
* creates an empty file within this directory
|
||||
* creates a file within this directory
|
||||
* @param relativePathArg
|
||||
*/
|
||||
public async createEmptyFile(relativePathArg: string) {
|
||||
const emptyFile = await File.create({
|
||||
const emtpyFile = await File.create({
|
||||
directory: this,
|
||||
name: relativePathArg,
|
||||
contents: '',
|
||||
});
|
||||
return emptyFile;
|
||||
}
|
||||
|
||||
// file operations
|
||||
@ -350,34 +297,23 @@ export class Directory {
|
||||
|
||||
/**
|
||||
* removes a file within the directory
|
||||
* uses file class to make sure effects for metadata etc. are handled correctly
|
||||
* @param optionsArg
|
||||
*/
|
||||
public async fastRemove(optionsArg: {
|
||||
path: string
|
||||
/**
|
||||
* wether the file should be placed into trash. Default is false.
|
||||
*/
|
||||
mode?: 'permanent' | 'trash';
|
||||
}) {
|
||||
const file = await this.getFileStrict({
|
||||
path: optionsArg.path,
|
||||
});
|
||||
await file.delete({
|
||||
mode: optionsArg.mode ? optionsArg.mode : 'permanent',
|
||||
public async fastRemove(optionsArg: { path: string }) {
|
||||
const path = plugins.path.join(this.getBasePath(), optionsArg.path);
|
||||
await this.bucketRef.fastRemove({
|
||||
path,
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* deletes the directory with all its contents
|
||||
*/
|
||||
public async delete(optionsArg: {
|
||||
mode?: 'permanent' | 'trash';
|
||||
}) {
|
||||
public async delete() {
|
||||
const deleteDirectory = async (directoryArg: Directory) => {
|
||||
const childDirectories = await directoryArg.listDirectories();
|
||||
if (childDirectories.length === 0) {
|
||||
console.log('Directory empty! Path complete!');
|
||||
console.log('directory empty! Path complete!');
|
||||
} else {
|
||||
for (const childDir of childDirectories) {
|
||||
await deleteDirectory(childDir);
|
||||
@ -385,9 +321,9 @@ export class Directory {
|
||||
}
|
||||
const files = await directoryArg.listFiles();
|
||||
for (const file of files) {
|
||||
await file.delete({
|
||||
mode: optionsArg.mode ? optionsArg.mode : 'permanent',
|
||||
})
|
||||
await directoryArg.fastRemove({
|
||||
path: file.name,
|
||||
});
|
||||
}
|
||||
};
|
||||
await deleteDirectory(this);
|
||||
|
@ -50,10 +50,6 @@ export class File {
|
||||
public parentDirectoryRef: Directory;
|
||||
public name: string;
|
||||
|
||||
/**
|
||||
* get the full path to the file
|
||||
* @returns the full path to the file
|
||||
*/
|
||||
public getBasePath(): string {
|
||||
return plugins.path.join(this.parentDirectoryRef.getBasePath(), this.name);
|
||||
}
|
||||
@ -92,24 +88,25 @@ export class File {
|
||||
/**
|
||||
* deletes this file
|
||||
*/
|
||||
public async delete(optionsArg?: { mode: 'trash' | 'permanent' }) {
|
||||
public async delete(optionsArg?: {
|
||||
mode: 'trash' | 'permanent';
|
||||
}) {
|
||||
|
||||
optionsArg = {
|
||||
...{
|
||||
... {
|
||||
mode: 'permanent',
|
||||
},
|
||||
...optionsArg,
|
||||
};
|
||||
}
|
||||
|
||||
if (optionsArg.mode === 'permanent') {
|
||||
await this.parentDirectoryRef.bucketRef.fastRemove({
|
||||
path: this.getBasePath(),
|
||||
});
|
||||
if (!this.name.endsWith('.metadata')) {
|
||||
if (await this.hasMetaData()) {
|
||||
const metadata = await this.getMetaData();
|
||||
await metadata.metadataFile.delete(optionsArg);
|
||||
}
|
||||
}
|
||||
} else if (optionsArg.mode === 'trash') {
|
||||
const metadata = await this.getMetaData();
|
||||
await metadata.storeCustomMetaData({
|
||||
@ -120,9 +117,8 @@ export class File {
|
||||
},
|
||||
});
|
||||
const trash = await this.parentDirectoryRef.bucketRef.getTrash();
|
||||
const trashDir = await trash.getTrashDir();
|
||||
await this.move({
|
||||
directory: trashDir,
|
||||
directory: await trash.getTrashDir(),
|
||||
path: await trash.getTrashKeyByOriginalBasePath(this.getBasePath()),
|
||||
});
|
||||
}
|
||||
@ -130,33 +126,6 @@ export class File {
|
||||
await this.parentDirectoryRef.listFiles();
|
||||
}
|
||||
|
||||
/**
|
||||
* restores
|
||||
*/
|
||||
public async restore(optionsArg: {
|
||||
useOriginalPath?: boolean;
|
||||
toPath?: string;
|
||||
overwrite?: boolean;
|
||||
} = {}) {
|
||||
optionsArg = {
|
||||
useOriginalPath: (() => {
|
||||
return optionsArg.toPath ? false : true;
|
||||
})(),
|
||||
overwrite: false,
|
||||
...optionsArg,
|
||||
};
|
||||
const metadata = await this.getMetaData();
|
||||
const moveToPath = optionsArg.toPath || (await metadata.getCustomMetaData({
|
||||
key: 'recycle'
|
||||
})).originalPath;
|
||||
await metadata.deleteCustomMetaData({
|
||||
key: 'recycle'
|
||||
})
|
||||
await this.move({
|
||||
path: moveToPath,
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* allows locking the file
|
||||
* @param optionsArg
|
||||
@ -181,7 +150,7 @@ export class File {
|
||||
}) {
|
||||
const metadata = await this.getMetaData();
|
||||
await metadata.removeLock({
|
||||
force: optionsArg?.force || false,
|
||||
force: optionsArg?.force,
|
||||
});
|
||||
}
|
||||
|
||||
@ -196,19 +165,16 @@ export class File {
|
||||
await this.parentDirectoryRef.bucketRef.fastPutStream({
|
||||
path: this.getBasePath(),
|
||||
readableStream: optionsArg.contents,
|
||||
overwrite: true,
|
||||
});
|
||||
} else if (Buffer.isBuffer(optionsArg.contents)) {
|
||||
await this.parentDirectoryRef.bucketRef.fastPut({
|
||||
path: this.getBasePath(),
|
||||
contents: optionsArg.contents,
|
||||
overwrite: true,
|
||||
});
|
||||
} else if (typeof optionsArg.contents === 'string') {
|
||||
await this.parentDirectoryRef.bucketRef.fastPut({
|
||||
path: this.getBasePath(),
|
||||
contents: Buffer.from(optionsArg.contents, optionsArg.encoding),
|
||||
overwrite: true,
|
||||
});
|
||||
}
|
||||
}
|
||||
@ -217,52 +183,23 @@ export class File {
|
||||
* moves the file to another directory
|
||||
*/
|
||||
public async move(pathDescriptorArg: interfaces.IPathDecriptor) {
|
||||
let moveToPath: string = '';
|
||||
let moveToPath = '';
|
||||
const isDirectory = await this.parentDirectoryRef.bucketRef.isDirectory(pathDescriptorArg);
|
||||
if (isDirectory) {
|
||||
moveToPath = await helpers.reducePathDescriptorToPath({
|
||||
...pathDescriptorArg,
|
||||
path: plugins.path.join(pathDescriptorArg.path!, this.name),
|
||||
path: plugins.path.join(pathDescriptorArg.path, this.name),
|
||||
});
|
||||
} else {
|
||||
moveToPath = await helpers.reducePathDescriptorToPath(pathDescriptorArg);
|
||||
}
|
||||
// lets move the file
|
||||
await this.parentDirectoryRef.bucketRef.fastMove({
|
||||
sourcePath: this.getBasePath(),
|
||||
destinationPath: moveToPath,
|
||||
overwrite: true,
|
||||
});
|
||||
|
||||
// lets move the metadatafile
|
||||
if (!this.name.endsWith('.metadata')) {
|
||||
const metadata = await this.getMetaData();
|
||||
await this.parentDirectoryRef.bucketRef.fastMove({
|
||||
sourcePath: metadata.metadataFile.getBasePath(),
|
||||
destinationPath: moveToPath + '.metadata',
|
||||
overwrite: true,
|
||||
});
|
||||
}
|
||||
|
||||
// lets update references of this
|
||||
const baseDirectory = await this.parentDirectoryRef.bucketRef.getBaseDirectory();
|
||||
this.parentDirectoryRef = await baseDirectory.getSubDirectoryByNameStrict(
|
||||
await helpers.reducePathDescriptorToPath(pathDescriptorArg),
|
||||
{
|
||||
couldBeFilePath: true,
|
||||
}
|
||||
);
|
||||
this.name = pathDescriptorArg.path!;
|
||||
}
|
||||
|
||||
public async hasMetaData(): Promise<boolean> {
|
||||
if (!this.name.endsWith('.metadata')) {
|
||||
const hasMetadataBool = MetaData.hasMetaData({
|
||||
file: this,
|
||||
});
|
||||
return hasMetadataBool;
|
||||
} else {
|
||||
return false;
|
||||
}
|
||||
await metadata.metadataFile.move(pathDescriptorArg);
|
||||
}
|
||||
|
||||
/**
|
||||
@ -293,11 +230,4 @@ export class File {
|
||||
contents: JSON.stringify(dataArg),
|
||||
});
|
||||
}
|
||||
|
||||
public async getMagicBytes(optionsArg: { length: number }): Promise<Buffer> {
|
||||
return this.parentDirectoryRef.bucketRef.getMagicBytes({
|
||||
path: this.getBasePath(),
|
||||
length: optionsArg.length,
|
||||
});
|
||||
}
|
||||
}
|
||||
|
@ -3,22 +3,14 @@ import * as plugins from './plugins.js';
|
||||
import { File } from './classes.file.js';
|
||||
|
||||
export class MetaData {
|
||||
public static async hasMetaData(optionsArg: { file: File }) {
|
||||
// lets find the existing metadata file
|
||||
const existingFile = await optionsArg.file.parentDirectoryRef.getFile({
|
||||
path: optionsArg.file.name + '.metadata',
|
||||
});
|
||||
return !!existingFile;
|
||||
}
|
||||
|
||||
// static
|
||||
public static async createForFile(optionsArg: { file: File }) {
|
||||
const metaData = new MetaData();
|
||||
metaData.fileRef = optionsArg.file;
|
||||
|
||||
// lets find the existing metadata file
|
||||
metaData.metadataFile = await metaData.fileRef.parentDirectoryRef.getFileStrict({
|
||||
path: metaData.fileRef.name + '.metadata',
|
||||
metaData.metadataFile = await metaData.fileRef.parentDirectoryRef.getFile({
|
||||
name: metaData.fileRef.name + '.metadata',
|
||||
createWithContents: '{}',
|
||||
});
|
||||
|
||||
@ -29,34 +21,20 @@ export class MetaData {
|
||||
/**
|
||||
* the file that contains the metadata
|
||||
*/
|
||||
metadataFile!: File;
|
||||
metadataFile: File;
|
||||
|
||||
/**
|
||||
* the file that the metadata is for
|
||||
*/
|
||||
fileRef!: File;
|
||||
fileRef: File;
|
||||
|
||||
public async getFileType(optionsArg?: {
|
||||
useFileExtension?: boolean;
|
||||
useMagicBytes?: boolean;
|
||||
}): Promise<plugins.smartmime.IFileTypeResult | undefined> {
|
||||
if ((optionsArg && optionsArg.useFileExtension) || !optionsArg) {
|
||||
const fileType = await plugins.smartmime.detectMimeType({
|
||||
path: this.fileRef.name,
|
||||
});
|
||||
|
||||
return fileType;
|
||||
}): Promise<string> {
|
||||
if ((optionsArg && optionsArg.useFileExtension) || optionsArg.useFileExtension === undefined) {
|
||||
return plugins.path.extname(this.fileRef.name);
|
||||
}
|
||||
if (optionsArg && optionsArg.useMagicBytes) {
|
||||
const fileType = await plugins.smartmime.detectMimeType({
|
||||
buffer: await this.fileRef.getMagicBytes({
|
||||
length: 100,
|
||||
})
|
||||
});
|
||||
|
||||
return fileType;
|
||||
}
|
||||
throw new Error('optionsArg.useFileExtension and optionsArg.useMagicBytes cannot both be false');
|
||||
}
|
||||
|
||||
/**
|
||||
@ -66,13 +44,13 @@ export class MetaData {
|
||||
const stat = await this.fileRef.parentDirectoryRef.bucketRef.fastStat({
|
||||
path: this.fileRef.getBasePath(),
|
||||
});
|
||||
return stat.ContentLength!;
|
||||
return stat.size;
|
||||
}
|
||||
|
||||
private prefixCustomMetaData = 'custom_';
|
||||
|
||||
public async storeCustomMetaData<T = any>(optionsArg: { key: string; value: T }) {
|
||||
const data = await this.metadataFile.getJsonData();
|
||||
const data = await this.metadataFile.getContentsAsString();
|
||||
data[this.prefixCustomMetaData + optionsArg.key] = optionsArg.value;
|
||||
await this.metadataFile.writeJsonData(data);
|
||||
}
|
||||
|
@ -1,34 +1,22 @@
|
||||
// classes.smartbucket.ts
|
||||
|
||||
import * as plugins from './plugins.js';
|
||||
import { Bucket } from './classes.bucket.js';
|
||||
|
||||
export class SmartBucket {
|
||||
public config: plugins.tsclass.storage.IS3Descriptor;
|
||||
|
||||
public s3Client: plugins.s3.S3Client;
|
||||
public minioClient: plugins.minio.Client;
|
||||
|
||||
/**
|
||||
* the constructor of SmartBucket
|
||||
*/
|
||||
/**
|
||||
* the constructor of SmartBucket
|
||||
*/
|
||||
constructor(configArg: plugins.tsclass.storage.IS3Descriptor) {
|
||||
this.config = configArg;
|
||||
|
||||
const protocol = configArg.useSsl === false ? 'http' : 'https';
|
||||
const port = configArg.port ? `:${configArg.port}` : '';
|
||||
const endpoint = `${protocol}://${configArg.endpoint}${port}`;
|
||||
|
||||
this.s3Client = new plugins.s3.S3Client({
|
||||
endpoint,
|
||||
region: configArg.region || 'us-east-1',
|
||||
credentials: {
|
||||
accessKeyId: configArg.accessKey,
|
||||
secretAccessKey: configArg.accessSecret,
|
||||
},
|
||||
forcePathStyle: true, // Necessary for S3-compatible storage like MinIO or Wasabi
|
||||
this.minioClient = new plugins.minio.Client({
|
||||
endPoint: this.config.endpoint,
|
||||
port: configArg.port || 443,
|
||||
useSSL: configArg.useSsl !== undefined ? configArg.useSsl : true,
|
||||
accessKey: this.config.accessKey,
|
||||
secretKey: this.config.accessSecret,
|
||||
});
|
||||
}
|
||||
|
||||
@ -41,15 +29,7 @@ export class SmartBucket {
|
||||
await Bucket.removeBucketByName(this, bucketName);
|
||||
}
|
||||
|
||||
public async getBucketByName(bucketNameArg: string) {
|
||||
return Bucket.getBucketByName(this, bucketNameArg);
|
||||
}
|
||||
|
||||
public async getBucketByNameStrict(...args: Parameters<SmartBucket['getBucketByName']>) {
|
||||
const bucket = await this.getBucketByName(...args);
|
||||
if (!bucket) {
|
||||
throw new Error(`Bucket ${args[0]} does not exist.`);
|
||||
}
|
||||
return bucket;
|
||||
public async getBucketByName(bucketName: string) {
|
||||
return Bucket.getBucketByName(this, bucketName);
|
||||
}
|
||||
}
|
||||
|
@ -21,7 +21,7 @@ export class Trash {
|
||||
const trashDir = await this.getTrashDir();
|
||||
const originalPath = await helpers.reducePathDescriptorToPath(pathDescriptor);
|
||||
const trashKey = await this.getTrashKeyByOriginalBasePath(originalPath);
|
||||
return trashDir.getFileStrict({ path: trashKey });
|
||||
return trashDir.getFile({ name: trashKey });
|
||||
}
|
||||
|
||||
public async getTrashKeyByOriginalBasePath (originalPath: string): Promise<string> {
|
||||
|
@ -1,5 +1,3 @@
|
||||
// plugins.ts
|
||||
|
||||
// node native
|
||||
import * as path from 'path';
|
||||
import * as stream from 'stream';
|
||||
@ -25,8 +23,6 @@ export {
|
||||
}
|
||||
|
||||
// third party scope
|
||||
import * as s3 from '@aws-sdk/client-s3';
|
||||
import * as minio from 'minio';
|
||||
|
||||
export {
|
||||
s3,
|
||||
}
|
||||
export { minio };
|
||||
|
@ -6,8 +6,7 @@
|
||||
"module": "NodeNext",
|
||||
"moduleResolution": "NodeNext",
|
||||
"esModuleInterop": true,
|
||||
"verbatimModuleSyntax": true,
|
||||
"strict": true
|
||||
"verbatimModuleSyntax": true
|
||||
},
|
||||
"exclude": [
|
||||
"dist_*/**/*.d.ts"
|
||||
|
Loading…
x
Reference in New Issue
Block a user