# @lossless.zone/objectstorage > 🚀 S3-compatible object storage server with clustering, erasure coding, and a slick management UI — powered by [`smartstorage`](https://code.foss.global/push.rocks/smartstorage). **objectstorage** gives you a fully featured, self-hosted S3-compatible storage server with a beautiful web-based management interface — all in a single Docker image. No Java, no bloat, no fuss. Built on Deno for the backend and [`@design.estate/dees-catalog`](https://code.foss.global/design.estate/dees-catalog) for a polished UI, it speaks the S3 protocol out of the box while adding powerful management features on top. Scale from a single node to a distributed cluster with erasure coding and multi-drive support. ## Issue Reporting and Security For reporting bugs, issues, or security vulnerabilities, please visit [community.foss.global/](https://community.foss.global/). This is the central community hub for all issue reporting. Developers who sign and comply with our contribution agreement and go through identification can also get a [code.foss.global/](https://code.foss.global/) account to submit Pull Requests directly. ## ✨ Features - **Full S3 API compatibility** — Works with any S3 client, SDK, or tool (AWS CLI, boto3, etc.) - **🔗 Cluster mode** — Distribute storage across multiple nodes with QUIC transport, automatic discovery, and quorum writes/reads - **🛡️ Erasure coding** — Reed-Solomon erasure coding (default 4+2) for data durability with minimal overhead - **💾 Multi-drive support** — Stripe data across multiple disks per node with per-drive health monitoring - **🔄 Self-healing** — Background scanner detects and reconstructs missing or corrupt shards automatically - **Management UI** — Web dashboard for buckets, objects, policies, credentials, cluster config, and storage drives - **Finder-style object browser** — Column view with file preview, drag-and-drop upload, move/rename, context menus - **Inline code editing** — Built-in Monaco editor with syntax highlighting and save-back-to-storage - **PDF viewer** — Render PDFs inline with page navigation, zoom, and thumbnails - **Named policy management** — Create reusable IAM-style policies, attach them to multiple buckets - **Credential management** — Add/remove access keys through the UI with live-reload - **Single Docker image** — Multi-arch (`amd64` + `arm64`), tiny Alpine-based image - **Fast** — Rust-powered storage engine via `smartstorage`, streaming I/O with zero-copy and backpressure - **Secure by default** — JWT-based admin auth, S3 SigV4 authentication, bucket policies - **🌙 Dark theme** — Automatic dark mode following your system preference ## ⚡ Quick Start ### Docker (recommended) ```bash docker run -d \ --name objectstorage \ -p 9000:9000 \ -p 3000:3000 \ -v objstdata:/data \ -e OBJST_ACCESS_KEY=myadminkey \ -e OBJST_SECRET_KEY=mysupersecret \ -e OBJST_ADMIN_PASSWORD=myuipassword \ code.foss.global/lossless.zone/objectstorage:latest ``` Then open **http://localhost:3000** for the management UI and use **http://localhost:9000** as your S3 endpoint. ### Deno (development) ```bash # Clone and install frontend dependencies git clone ssh://git@code.foss.global:29419/lossless.zone/objectstorage.git cd objectstorage pnpm install pnpm run build # Run in ephemeral mode (data stored in .nogit/objstdata) deno run --allow-all mod.ts server --ephemeral ``` ## ⚙️ Configuration objectstorage is configured through environment variables, CLI flags, or programmatic config. **Environment variables take precedence** over CLI flags. ### Server Environment Variables | Variable | Description | Default | |---|---|---| | `OBJST_PORT` | Storage API port | `9000` | | `UI_PORT` | Management UI port | `3000` | | `OBJST_STORAGE_DIR` | Data storage directory | `/data` | | `OBJST_ACCESS_KEY` | Access key ID | `admin` | | `OBJST_SECRET_KEY` | Secret access key | `admin` | | `OBJST_ADMIN_PASSWORD` | Admin UI password | `admin` | | `OBJST_REGION` | Storage region identifier | `us-east-1` | ### Cluster Environment Variables | Variable | Description | Default | |---|---|---| | `OBJST_CLUSTER_ENABLED` | Enable cluster mode (`true`/`false`) | `false` | | `OBJST_CLUSTER_NODE_ID` | Unique node identifier | auto-generated | | `OBJST_CLUSTER_QUIC_PORT` | QUIC transport port | `4433` | | `OBJST_CLUSTER_SEED_NODES` | Comma-separated seed node addresses | _(empty)_ | | `OBJST_DRIVE_PATHS` | Comma-separated drive mount paths | storage dir | | `OBJST_ERASURE_DATA_SHARDS` | Erasure coding data shards | `4` | | `OBJST_ERASURE_PARITY_SHARDS` | Erasure coding parity shards | `2` | | `OBJST_ERASURE_CHUNK_SIZE` | Erasure chunk size in bytes | `4194304` (4 MB) | | `OBJST_HEARTBEAT_INTERVAL_MS` | Cluster heartbeat interval | `5000` | | `OBJST_HEARTBEAT_TIMEOUT_MS` | Cluster heartbeat timeout | `30000` | ### CLI Flags ```bash deno run --allow-all mod.ts server [options] Server Options: --storage-port Storage API port (default: 9000) --ui-port Management UI port (default: 3000) --storage-dir Storage directory (default: /data) --ephemeral Use ./.nogit/objstdata for storage (dev mode) Clustering Options: --cluster-enabled Enable cluster mode --cluster-node-id Unique node identifier --cluster-quic-port QUIC transport port (default: 4433) --cluster-seed-nodes Comma-separated seed node addresses --drive-paths Comma-separated drive mount paths --erasure-data-shards Erasure coding data shards (default: 4) --erasure-parity-shards Erasure coding parity shards (default: 2) ``` ## 🔗 Cluster Mode objectstorage supports distributed storage across multiple nodes with automatic failover and data redundancy. ### How it works 1. **Enable clustering** on each node with `OBJST_CLUSTER_ENABLED=true` 2. **Point nodes at each other** using `OBJST_CLUSTER_SEED_NODES` — nodes discover the full cluster from any seed 3. **Configure drives** per node with `OBJST_DRIVE_PATHS` — each drive is independently managed 4. **Erasure coding** splits objects into data + parity shards across drives and nodes ### Example: 3-node cluster ```bash # Node 1 docker run -d --name objst-node1 \ -p 9000:9000 -p 3000:3000 -p 4433:4433/udp \ -v /mnt/disk1:/drive1 -v /mnt/disk2:/drive2 \ -e OBJST_CLUSTER_ENABLED=true \ -e OBJST_CLUSTER_NODE_ID=node-1 \ -e OBJST_CLUSTER_SEED_NODES=node2:4433,node3:4433 \ -e OBJST_DRIVE_PATHS=/drive1,/drive2 \ -e OBJST_ACCESS_KEY=myadminkey \ -e OBJST_SECRET_KEY=mysupersecret \ code.foss.global/lossless.zone/objectstorage:latest # Node 2 and Node 3 — same pattern, different node IDs and seed nodes ``` ### Erasure coding presets | Config | Data Shards | Parity Shards | Overhead | Fault Tolerance | |---|---|---|---|---| | Default | 4 | 2 | 50% | 2 failures | | High durability | 6 | 3 | 50% | 3 failures | | Minimal | 2 | 1 | 50% | 1 failure | ### Inter-node transport Cluster communication uses **QUIC** (UDP port 4433 by default) with: - Auto-generated TLS certificates - Multiplexed streams with flow-control backpressure - Heartbeat-based failure detection (default: 5s interval, 30s timeout) ## 🖥️ Management UI The web-based management UI is served on the UI port (default: `3000`). Log in with username `admin` and the configured admin password. ### Overview Dashboard showing server status, uptime, storage usage, bucket count, and connection info. ![Overview](./docs/01-overview.png) ### Buckets Create/delete buckets. View object counts and sizes. Attach/detach named policies per bucket. ![Buckets](./docs/02-buckets.png) ### Browser Finder-style column browser for objects. Upload, download, preview, move, rename, and delete files and folders — with syntax-highlighted code preview. ![Browser](./docs/03-browser.png) Right-click any item for quick actions: ![Context Menu](./docs/11-context-menu.png) #### Inline Code Editing Click Edit on any text file to open the built-in Monaco editor with syntax highlighting, language detection, and save-back-to-storage. ![Code Editing](./docs/06-code-editing.png) #### PDF Viewer PDF files render inline with a full-featured viewer — page navigation, zoom, fit-to-page, thumbnails, download, and print. ![PDF Viewer](./docs/07-pdf-viewer.png) ### Policies Create reusable named policies with IAM-style S3 statements. Attach policies to multiple buckets at once. ![Policies](./docs/04-policies.png) #### Attaching Policies to Buckets From the Buckets view, click the policy icon on any bucket to see attached and available policies. Attach or detach with one click. ![Attach Policy](./docs/08-attach-policy.png) ### Access Keys Add and remove access credentials. Secret keys are masked. Changes take effect immediately — no server restart needed. ![Access Keys](./docs/05-access-keys.png) #### Adding Access Keys Click "Add Key" to create new access credentials. They're immediately available for API authentication. ![Add Access Key](./docs/09-add-access-key.png) ### Configuration View your server's current configuration at a glance — ports, region, storage directory, auth/CORS status, cluster configuration, erasure coding settings, and storage drive paths. The config view also includes an environment variable reference guide for cluster setup. ![Configuration](./docs/06-config.png) ### Dark Theme Full dark theme support — automatically follows your system preference via `prefers-color-scheme`. ![Dark Theme](./docs/10-dark-theme.png) ## 📋 Named Policy System objectstorage adds a **named policy** abstraction on top of standard S3 bucket policies. Instead of editing raw JSON per bucket, you define reusable policy templates and attach them to any number of buckets. ### How it works 1. **Create a named policy** in the Policies view — give it a name, description, and S3 policy statements 2. **Attach it to buckets** — from the Policies view or the Buckets view 3. **objectstorage merges** all attached policy statements into a single S3 policy document and applies it to the bucket automatically ### `${bucket}` placeholder Use `${bucket}` in your policy's `Resource` ARN and it will be replaced with the actual bucket name when applied: ```json [ { "Sid": "PublicRead", "Effect": "Allow", "Principal": "*", "Action": "s3:GetObject", "Resource": "arn:aws:s3:::${bucket}/*" } ] ``` This lets one policy like "Public Read" work across many buckets without hardcoding names. ### Lifecycle - **Updating a policy** automatically recomputes and re-applies the merged S3 policy on all attached buckets - **Deleting a policy** detaches it from all buckets and recomputes each - **Deleting a bucket** cleans up its policy attachments automatically ## 🔌 S3 API Usage Use any S3-compatible client to interact with the storage. Here are some examples: ### AWS CLI ```bash # Configure AWS CLI aws configure set aws_access_key_id admin aws configure set aws_secret_access_key admin aws configure set default.region us-east-1 # Create a bucket aws --endpoint-url http://localhost:9000 s3 mb s3://my-bucket # Upload a file aws --endpoint-url http://localhost:9000 s3 cp myfile.txt s3://my-bucket/ # List objects aws --endpoint-url http://localhost:9000 s3 ls s3://my-bucket/ # Download a file aws --endpoint-url http://localhost:9000 s3 cp s3://my-bucket/myfile.txt ./downloaded.txt ``` ### Node.js / TypeScript (AWS SDK v3) ```typescript import { S3Client, PutObjectCommand, ListObjectsV2Command } from '@aws-sdk/client-s3'; const s3 = new S3Client({ endpoint: 'http://localhost:9000', region: 'us-east-1', credentials: { accessKeyId: 'admin', secretAccessKey: 'admin', }, forcePathStyle: true, }); // Upload an object await s3.send(new PutObjectCommand({ Bucket: 'my-bucket', Key: 'hello.txt', Body: 'Hello, S3!', })); // List objects in a bucket const result = await s3.send(new ListObjectsV2Command({ Bucket: 'my-bucket', })); console.log(result.Contents); ``` ### Python (boto3) ```python import boto3 s3 = boto3.client( 's3', endpoint_url='http://localhost:9000', aws_access_key_id='admin', aws_secret_access_key='admin', region_name='us-east-1', ) # Upload a file s3.put_object(Bucket='my-bucket', Key='hello.txt', Body=b'Hello from Python!') # List objects response = s3.list_objects_v2(Bucket='my-bucket') for obj in response.get('Contents', []): print(obj['Key'], obj['Size']) ``` ## 🐳 Docker ### Build ```bash # Build for the native platform pnpm run build:docker # Or build and run directly pnpm run start:docker ``` ### Docker Compose (standalone) ```yaml services: objectstorage: image: code.foss.global/lossless.zone/objectstorage:latest ports: - "9000:9000" # S3 API - "3000:3000" # Management UI volumes: - objstdata:/data environment: OBJST_ACCESS_KEY: myadminkey OBJST_SECRET_KEY: mysupersecret OBJST_ADMIN_PASSWORD: securepw123 volumes: objstdata: ``` ### Docker Compose (3-node cluster) ```yaml services: node1: image: code.foss.global/lossless.zone/objectstorage:latest ports: - "9001:9000" - "3001:3000" - "4433:4433/udp" volumes: - node1-drive1:/drive1 - node1-drive2:/drive2 environment: OBJST_CLUSTER_ENABLED: "true" OBJST_CLUSTER_NODE_ID: node-1 OBJST_CLUSTER_QUIC_PORT: "4433" OBJST_CLUSTER_SEED_NODES: node2:4433,node3:4433 OBJST_DRIVE_PATHS: /drive1,/drive2 OBJST_ACCESS_KEY: myadminkey OBJST_SECRET_KEY: mysupersecret node2: image: code.foss.global/lossless.zone/objectstorage:latest ports: - "9002:9000" - "3002:3000" - "4434:4433/udp" volumes: - node2-drive1:/drive1 - node2-drive2:/drive2 environment: OBJST_CLUSTER_ENABLED: "true" OBJST_CLUSTER_NODE_ID: node-2 OBJST_CLUSTER_QUIC_PORT: "4433" OBJST_CLUSTER_SEED_NODES: node1:4433,node3:4433 OBJST_DRIVE_PATHS: /drive1,/drive2 OBJST_ACCESS_KEY: myadminkey OBJST_SECRET_KEY: mysupersecret node3: image: code.foss.global/lossless.zone/objectstorage:latest ports: - "9003:9000" - "3003:3000" - "4435:4433/udp" volumes: - node3-drive1:/drive1 - node3-drive2:/drive2 environment: OBJST_CLUSTER_ENABLED: "true" OBJST_CLUSTER_NODE_ID: node-3 OBJST_CLUSTER_QUIC_PORT: "4433" OBJST_CLUSTER_SEED_NODES: node1:4433,node2:4433 OBJST_DRIVE_PATHS: /drive1,/drive2 OBJST_ACCESS_KEY: myadminkey OBJST_SECRET_KEY: mysupersecret volumes: node1-drive1: node1-drive2: node2-drive1: node2-drive2: node3-drive1: node3-drive2: ``` ### Image Details - **Base**: `alpine:edge` with Deno runtime - **Architectures**: `linux/amd64`, `linux/arm64` - **Size**: ~150 MB compressed - **Init system**: `tini` for proper signal handling - **Exposed ports**: `9000` (S3), `3000` (UI), `4433` (QUIC cluster transport) - **Volume**: `/data` — all bucket data and config persisted here ## 🏗️ Architecture ``` ┌──────────────────────────────────────────────────────────────┐ │ objectstorage │ │ │ │ ┌──────────────┐ ┌────────────────────────────────────┐ │ │ │ Management │ │ Storage Engine (smartstorage) │ │ │ │ UI (port │ │ Rust binary via ruststorage │ │ │ │ 3000) │ │ (port 9000) │ │ │ │ │ │ │ │ │ │ dees-catalog │ │ • S3 API (path-style routing) │ │ │ │ SPA bundle │ │ • SigV4 authentication │ │ │ └──────┬───────┘ │ • Bucket policies │ │ │ │ │ • Streaming I/O (zero-copy) │ │ │ ┌──────▼───────┐ │ • Multipart upload support │ │ │ │ OpsServer │ └─────────────┬──────────────────────┘ │ │ │ (TypedReq │ │ │ │ │ handlers) │── S3 Client ──────┘ │ │ │ │ │ │ │ • Admin auth │ ┌────────────────────────────────────┐ │ │ │ • CRUD APIs │ │ Cluster Layer (optional) │ │ │ │ • Policy mgr │ │ │ │ │ └──────────────┘ │ • QUIC transport (port 4433) │ │ │ │ • Reed-Solomon erasure coding │ │ │ │ • Quorum writes / reads │ │ │ │ • Heartbeat failure detection │ │ │ │ • Self-healing shard repair │ │ │ └────────────────────────────────────┘ │ │ │ │ ┌──────────────────────────────────────────────────────────┐ │ │ │ Storage Drives │ │ │ │ /drive1 /drive2 /drive3 ... (or single /data) │ │ │ └──────────────────────────────────────────────────────────┘ │ └──────────────────────────────────────────────────────────────┘ ``` ### Tech Stack | Layer | Technology | |---|---| | **Storage Engine** | [`@push.rocks/smartstorage`](https://code.foss.global/push.rocks/smartstorage) (Rust binary via `ruststorage`) | | **Cluster Transport** | QUIC via `quinn` (auto-TLS, multiplexed streams, backpressure) | | **Erasure Coding** | Reed-Solomon (configurable data + parity shards) | | **Runtime** | Deno | | **Management API** | [`@api.global/typedrequest`](https://code.foss.global/api.global/typedrequest) + [`@api.global/typedserver`](https://code.foss.global/api.global/typedserver) | | **Auth** | JWT via [`@push.rocks/smartjwt`](https://code.foss.global/push.rocks/smartjwt), S3 SigV4 | | **Frontend** | [`@design.estate/dees-element`](https://code.foss.global/design.estate/dees-element) (LitElement) + [`@design.estate/dees-catalog`](https://code.foss.global/design.estate/dees-catalog) | | **Frontend Build** | esbuild via [`@git.zone/tsbundle`](https://code.foss.global/git.zone/tsbundle) | | **Docker** | Multi-stage (Node.js build → Alpine + Deno runtime) | ## 🛠️ Development ```bash # Install dependencies pnpm install # Watch mode — auto-rebuilds frontend + restarts backend pnpm run watch # Build frontend bundle only pnpm run build # Type check backend deno check mod.ts # Run tests pnpm test # Run in development mode deno run --allow-all mod.ts server --ephemeral ``` ## License and Legal Information This repository contains open-source code licensed under the MIT License. A copy of the license can be found in the [license](./license) file. **Please note:** The MIT License does not grant permission to use the trade names, trademarks, service marks, or product names of the project, except as required for reasonable and customary use in describing the origin of the work and reproducing the content of the NOTICE file. ### Trademarks This project is owned and maintained by Task Venture Capital GmbH. The names and logos associated with Task Venture Capital GmbH and any related products or services are trademarks of Task Venture Capital GmbH or third parties, and are not included within the scope of the MIT license granted herein. Use of these trademarks must comply with Task Venture Capital GmbH's Trademark Guidelines or the guidelines of the respective third-party owners, and any usage must be approved in writing. Third-party trademarks used herein are the property of their respective owners and used only in a descriptive manner, e.g. for an implementation of an API or similar. ### Company Information Task Venture Capital GmbH Registered at District Court Bremen HRB 35230 HB, Germany For any legal inquiries or further information, please contact us via email at hello@task.vc. By using this repository, you acknowledge that you have read this section, agree to comply with its terms, and understand that the licensing of the code does not imply endorsement by Task Venture Capital GmbH of any derivative works.