6 Commits

Author SHA1 Message Date
4cf570afbd 1.0.86 2024-06-13 15:12:08 +02:00
8913faebde fix(core): update 2024-06-13 15:12:07 +02:00
9e523de620 1.0.85 2024-05-08 21:31:32 +02:00
435b6e24a1 fix(core): update 2024-05-08 21:31:31 +02:00
02e68ebb68 1.0.84 2024-05-08 20:53:35 +02:00
4e73bf7a0e fix(core): update 2024-05-08 20:53:35 +02:00
12 changed files with 5595 additions and 3612 deletions

View File

@ -1,132 +0,0 @@
# gitzone ci_default_private
image: registry.gitlab.com/hosttoday/ht-docker-node:npmci
cache:
paths:
- .npmci_cache/
key: '$CI_BUILD_STAGE'
stages:
- security
- test
- release
- metadata
before_script:
- pnpm install -g pnpm
- pnpm install -g @shipzone/npmci
- npmci npm prepare
# ====================
# security stage
# ====================
# ====================
# security stage
# ====================
auditProductionDependencies:
image: registry.gitlab.com/hosttoday/ht-docker-node:npmci
stage: security
script:
- npmci command npm config set registry https://registry.npmjs.org
- npmci command pnpm audit --audit-level=high --prod
tags:
- lossless
- docker
auditDevDependencies:
image: registry.gitlab.com/hosttoday/ht-docker-node:npmci
stage: security
script:
- npmci command npm config set registry https://registry.npmjs.org
- npmci command pnpm audit --audit-level=high --dev
tags:
- lossless
- docker
allow_failure: true
# ====================
# test stage
# ====================
testStable:
stage: test
script:
- npmci node install stable
- npmci npm install
- npmci npm test
coverage: /\d+.?\d+?\%\s*coverage/
tags:
- lossless
- docker
- notpriv
testBuild:
stage: test
script:
- npmci node install stable
- npmci npm install
- npmci command npm run build
coverage: /\d+.?\d+?\%\s*coverage/
tags:
- lossless
- docker
- notpriv
release:
stage: release
script:
- npmci node install stable
- npmci npm publish
only:
- tags
tags:
- lossless
- docker
- notpriv
# ====================
# metadata stage
# ====================
codequality:
stage: metadata
allow_failure: true
only:
- tags
script:
- npmci command npm install -g typescript
- npmci npm prepare
- npmci npm install
tags:
- lossless
- docker
- priv
trigger:
stage: metadata
script:
- npmci trigger
only:
- tags
tags:
- lossless
- docker
- notpriv
pages:
stage: metadata
script:
- npmci node install lts
- npmci command npm install -g @git.zone/tsdoc
- npmci npm install
- npmci command tsdoc
tags:
- lossless
- docker
- notpriv
only:
- tags
artifacts:
expire_in: 1 week
paths:
- public
allow_failure: true

View File

@ -1,4 +1,6 @@
Copyright (c) 2019 Lossless GmbH (hello@lossless.com) The MIT License (MIT)
Copyright (c) 2019 Task Venture Capital GmbH (hello@task.vc)
Permission is hereby granted, free of charge, to any person obtaining a copy Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal of this software and associated documentation files (the "Software"), to deal

View File

@ -5,10 +5,23 @@
"githost": "gitlab.com", "githost": "gitlab.com",
"gitscope": "losslessone/services/initzone", "gitscope": "losslessone/services/initzone",
"gitrepo": "spark", "gitrepo": "spark",
"description": "sparks the servezone services", "description": "A tool to maintain and configure servers on the base OS level for the Servezone infrastructure.",
"npmPackagename": "@losslessone_private/spark", "npmPackagename": "@losslessone_private/spark",
"license": "MIT", "license": "MIT",
"projectDomain": "https://lossless.one" "projectDomain": "https://lossless.one",
"keywords": [
"server management",
"devops",
"automation",
"docker",
"configuration management",
"daemon service",
"continuous integration",
"continuous deployment",
"deployment automation",
"service orchestration",
"node.js"
]
} }
}, },
"npmci": { "npmci": {

View File

@ -1,8 +1,8 @@
{ {
"name": "@serve.zone/spark", "name": "@serve.zone/spark",
"version": "1.0.83", "version": "1.0.86",
"private": false, "private": false,
"description": "sparks the servezone services", "description": "A tool to maintain and configure servers on the base OS level for the Servezone infrastructure.",
"main": "dist_ts/index.js", "main": "dist_ts/index.js",
"typings": "dist_ts/index.d.ts", "typings": "dist_ts/index.d.ts",
"author": "Task Venture Capital GmbH", "author": "Task Venture Capital GmbH",
@ -16,28 +16,29 @@
"spark": "./cli.js" "spark": "./cli.js"
}, },
"devDependencies": { "devDependencies": {
"@git.zone/tsbuild": "^2.1.29", "@git.zone/tsbuild": "^2.1.80",
"@git.zone/tsrun": "^1.2.39", "@git.zone/tsrun": "^1.2.39",
"@git.zone/tstest": "^1.0.60", "@git.zone/tstest": "^1.0.60",
"@push.rocks/tapbundle": "^5.0.4", "@push.rocks/tapbundle": "^5.0.4",
"@types/node": "20.10.0" "@types/node": "20.14.2"
}, },
"dependencies": { "dependencies": {
"@apiclient.xyz/docker": "^1.0.103", "@apiclient.xyz/docker": "^1.2.2",
"@push.rocks/npmextra": "^5.0.13", "@push.rocks/npmextra": "^5.0.17",
"@push.rocks/projectinfo": "^5.0.1", "@push.rocks/projectinfo": "^5.0.1",
"@push.rocks/qenv": "^6.0.5", "@push.rocks/qenv": "^6.0.5",
"@push.rocks/smartcli": "^4.0.6", "@push.rocks/smartcli": "^4.0.11",
"@push.rocks/smartdaemon": "^2.0.3", "@push.rocks/smartdaemon": "^2.0.3",
"@push.rocks/smartdelay": "^3.0.5", "@push.rocks/smartdelay": "^3.0.5",
"@push.rocks/smartfile": "^11.0.14", "@push.rocks/smartfile": "^11.0.20",
"@push.rocks/smartjson": "^5.0.5", "@push.rocks/smartjson": "^5.0.20",
"@push.rocks/smartlog": "^3.0.3", "@push.rocks/smartlog": "^3.0.7",
"@push.rocks/smartlog-destination-local": "^9.0.0", "@push.rocks/smartlog-destination-local": "^9.0.0",
"@push.rocks/smartpath": "^5.0.5", "@push.rocks/smartpath": "^5.0.5",
"@push.rocks/smartshell": "^3.0.5", "@push.rocks/smartshell": "^3.0.5",
"@push.rocks/smartupdate": "^2.0.4", "@push.rocks/smartupdate": "^2.0.4",
"@push.rocks/taskbuffer": "^3.0.10" "@push.rocks/taskbuffer": "^3.0.10",
"@serve.zone/interfaces": "^1.0.74"
}, },
"files": [ "files": [
"ts/**/*", "ts/**/*",
@ -54,5 +55,18 @@
"browserslist": [ "browserslist": [
"last 1 chrome versions" "last 1 chrome versions"
], ],
"type": "module" "type": "module",
"keywords": [
"server management",
"devops",
"automation",
"docker",
"configuration management",
"daemon service",
"continuous integration",
"continuous deployment",
"deployment automation",
"service orchestration",
"node.js"
]
} }

8729
pnpm-lock.yaml generated

File diff suppressed because it is too large Load Diff

2
readme.hints.md Normal file
View File

@ -0,0 +1,2 @@
- this package is part of serve.zone
- it is used to maintain and configure servers on the base OS level

111
readme.md
View File

@ -1,40 +1,93 @@
# @losslessone/services/initzone/spark # @serve.zone/spark
sparks the servezone services sparks the servezone services
## Availabililty and Links ## Install
* [npmjs.org (npm package)](https://www.npmjs.com/package/@losslessone_private/spark) To install `@serve.zone/spark`, run the following command in your terminal:
* [gitlab.com (source)](https://gitlab.com/losslessone/services/initzone/spark) ```sh
* [github.com (source mirror)](https://github.com/losslessone/services/initzone/spark) npm install @serve.zone/spark --save
* [docs (typedoc)](https://losslessone/services/initzone.gitlab.io/spark/) ```
## Status for master
Status Category | Status Badge
-- | --
GitLab Pipelines | [![pipeline status](https://gitlab.com/losslessone/services/initzone/spark/badges/master/pipeline.svg)](https://lossless.cloud)
GitLab Pipline Test Coverage | [![coverage report](https://gitlab.com/losslessone/services/initzone/spark/badges/master/coverage.svg)](https://lossless.cloud)
npm | [![npm downloads per month](https://badgen.net/npm/dy/@losslessone_private/spark)](https://lossless.cloud)
Snyk | [![Known Vulnerabilities](https://badgen.net/snyk/losslessone/services/initzone/spark)](https://lossless.cloud)
TypeScript Support | [![TypeScript](https://badgen.net/badge/TypeScript/>=%203.x/blue?icon=typescript)](https://lossless.cloud)
node Support | [![node](https://img.shields.io/badge/node->=%2010.x.x-blue.svg)](https://nodejs.org/dist/latest-v10.x/docs/api/)
Code Style | [![Code Style](https://badgen.net/badge/style/prettier/purple)](https://lossless.cloud)
PackagePhobia (total standalone install weight) | [![PackagePhobia](https://badgen.net/packagephobia/install/@losslessone_private/spark)](https://lossless.cloud)
PackagePhobia (package size on registry) | [![PackagePhobia](https://badgen.net/packagephobia/publish/@losslessone_private/spark)](https://lossless.cloud)
BundlePhobia (total size when bundled) | [![BundlePhobia](https://badgen.net/bundlephobia/minzip/@losslessone_private/spark)](https://lossless.cloud)
## Usage ## Usage
## Contribution ### Getting Started
To use `@serve.zone/spark` in your project, you need to include and initiate it in your TypeScript project. Ensure you have TypeScript and the necessary build tools set up in your project.
We are always happy for code contributions. If you are not the code contributing type that is ok. Still, maintaining Open Source repositories takes considerable time and thought. If you like the quality of what we do and our modules are useful to you we would appreciate a little monthly contribution: You can [contribute one time](https://lossless.link/contribute-onetime) or [contribute monthly](https://lossless.link/contribute). :) First, import `@serve.zone/spark`:
```typescript
import { Spark } from '@serve.zone/spark';
```
### Initializing Spark
Create an instance of the `Spark` class to start using Spark. This instance will serve as the main entry point for interacting with the Spark functionalities.
## Contribution ```typescript
const sparkInstance = new Spark();
```
We are always happy for code contributions. If you are not the code contributing type that is ok. Still, maintaining Open Source repositories takes considerable time and thought. If you like the quality of what we do and our modules are useful to you we would appreciate a little monthly contribution: You can [contribute one time](https://lossless.link/contribute-onetime) or [contribute monthly](https://lossless.link/contribute). :) ### Running Spark as a Daemon
To run Spark as a daemon, which is useful for maintaining and configuring servers on the base OS level, use the CLI feature bundled with Spark. This should ideally be handled outside of your code through a command-line terminal but can also be automated within your Node.js scripts if required.
For further information read the linked docs at the top of this readme. ```shell
spark installdaemon
```
## Legal The command above sets up Spark as a system service, enabling it to run and maintain server configurations automatically.
> MIT licensed | **©** [Task Venture Capital GmbH](https://task.vc)
| By using this npm module you agree to our [privacy policy](https://lossless.gmbH/privacy) ### Updating Spark or Maintained Services
Spark can self-update and manage updates for its maintained services. Trigger an update check and process by calling the `updateServices` method on the Spark instance.
```typescript
await sparkInstance.sparkUpdateManager.updateServices();
```
### Managing Configuration and Logging
Spark allows for extensive configuration and logging customization. Use the `SparkLocalConfig` and logging features to tailor Spark's operation to your needs.
```typescript
// Accessing the local configuration
const localConfig = sparkInstance.sparkLocalConfig;
// Utilizing the logger for custom log messages
import { logger } from '@serve.zone/spark';
logger.log('info', 'Custom log message');
```
### Advanced Usage
`@serve.zone/spark` offers a suite of tools for detailed server and service management, including but not limited to task scheduling, daemon management, and service updates. Explore the `SparkTaskManager` for scheduling specific tasks, `SparkUpdateManager` for handling service updates, and `SparkLocalConfig` for configuration.
### Example: Scheduling Custom Tasks
```typescript
import { SparkTaskManager } from '@serve.zone/spark';
const sparkInstance = new Spark();
const myTask = {
name: 'customTask',
taskFunction: async () => {
console.log('Running custom task');
},
};
sparkInstance.sparkTaskManager.taskmanager.addAndScheduleTask(myTask, '* * * * * *');
```
The example above creates a simple task that logs a message every second, demonstrating how to use Spark's task manager for custom scheduled tasks.
### Advanced Configuration
For advanced configurations, including Docker and service management:
- Use `SparkUpdateManager` to handle Docker image updates, service creation, and management.
- Access and modify Docker and service configurations through Spark's integration with configuration files and environment variables.
```typescript
// Managing Docker services with Spark
await sparkInstance.sparkUpdateManager.dockerHost.someDockerMethod();
// Example: Creating a Docker service
const newServiceDefinition = {...};
await sparkInstance.sparkUpdateManager.createService(newServiceDefinition);
```
### Conclusion
`@serve.zone/spark` provides a comprehensive toolkit for orchestrating and managing server environments and Docker-based services. By leveraging its CLI and programmatic interfaces, you can automate and streamline server operations, configurations, updates, and task scheduling, ensuring your infrastructure is responsive, updated, and maintained efficiently.
undefined

View File

@ -3,6 +3,6 @@
*/ */
export const commitinfo = { export const commitinfo = {
name: '@serve.zone/spark', name: '@serve.zone/spark',
version: '1.0.83', version: '1.0.86',
description: 'sparks the servezone services' description: 'A tool to maintain and configure servers on the base OS level for the Servezone infrastructure.'
} }

View File

@ -7,6 +7,9 @@ export class SparkLocalConfig {
constructor(sparkRefArg: Spark) { constructor(sparkRefArg: Spark) {
this.sparkRef = sparkRefArg; this.sparkRef = sparkRefArg;
this.kvStore = new plugins.npmextra.KeyValueStore('custom', 'spark'); this.kvStore = new plugins.npmextra.KeyValueStore({
typeArg: 'userHomeDir',
identityArg: 'spark',
});
} }
} }

View File

@ -9,7 +9,7 @@ export class SparkUpdateManager {
public smartupdate: plugins.smartupdate.SmartUpdate; public smartupdate: plugins.smartupdate.SmartUpdate;
constructor(sparkrefArg: Spark) { constructor(sparkrefArg: Spark) {
this.sparkRef = sparkrefArg; this.sparkRef = sparkrefArg;
this.dockerHost = new plugins.docker.DockerHost(); this.dockerHost = new plugins.docker.DockerHost({});
this.smartupdate = new plugins.smartupdate.SmartUpdate(); this.smartupdate = new plugins.smartupdate.SmartUpdate();
} }
@ -27,9 +27,6 @@ export class SparkUpdateManager {
plugins.path.join(paths.homeDir, 'serve.zone/spark/spark.json') plugins.path.join(paths.homeDir, 'serve.zone/spark/spark.json')
)) ))
) { ) {
const sparkJson = plugins.smartfile.fs.toObjectSync(
plugins.path.join(paths.homeDir, 'serve.zone/spark/spark.json')
);
const services: Array<{ const services: Array<{
name: string; name: string;
image: string; image: string;
@ -38,16 +35,45 @@ export class SparkUpdateManager {
environment: string; environment: string;
secretJson: any; secretJson: any;
}> = []; }> = [];
for (const serviceKey of Object.keys(sparkJson.services)) { // lets add coreflow
services.push({ services.push({
name: serviceKey, name: `coreflow`,
image: sparkJson.services[serviceKey].image, image: `code.foss.global/serve.zone/coreflow`,
url: sparkJson.services[serviceKey].url, url: `coreflow`,
environment: sparkJson.services[serviceKey].environment, environment: `production`,
port: sparkJson.services[serviceKey].port, port: `3000`,
secretJson: sparkJson.services[serviceKey].secretJson, secretJson: {
SERVEZONE_PORT: `3000`,
SERVEZONE_ENVIRONMENT: `production`,
},
}); });
}
services.push({
name: `coretraffic`,
image: `code.foss.global/serve.zone/coretraffic`,
url: `coreflow`,
environment: `production`,
port: `3000`,
secretJson: {
SERVEZONE_PORT: `3000`,
SERVEZONE_ENVIRONMENT: `production`,
},
});
services.push({
name: `corelog`,
image: `code.foss.global/serve.zone/corelog`,
url: `coreflow`,
environment: `production`,
port: `3000`,
secretJson: {
SERVEZONE_PORT: `3000`,
SERVEZONE_ENVIRONMENT: `production`,
},
});
// lets add coretraffic
for (const service of services) { for (const service of services) {
const existingService = await plugins.docker.DockerService.getServiceByName( const existingService = await plugins.docker.DockerService.getServiceByName(
this.dockerHost, this.dockerHost,
@ -61,6 +87,7 @@ export class SparkUpdateManager {
const needsUpdate: boolean = await existingService.needsUpdate(); const needsUpdate: boolean = await existingService.needsUpdate();
if (!needsUpdate) { if (!needsUpdate) {
logger.log('info', `not needing update.`); logger.log('info', `not needing update.`);
// we simply return here to end the functions
return; return;
} }
logger.log('ok', `${service.name} needs to be updated!`); logger.log('ok', `${service.name} needs to be updated!`);
@ -74,7 +101,9 @@ export class SparkUpdateManager {
const newServiceImage = await plugins.docker.DockerImage.createFromRegistry( const newServiceImage = await plugins.docker.DockerImage.createFromRegistry(
this.dockerHost, this.dockerHost,
{ {
creationObject: {
imageUrl: service.image, imageUrl: service.image,
},
} }
); );
const newServiceSecret = await plugins.docker.DockerSecret.createSecret(this.dockerHost, { const newServiceSecret = await plugins.docker.DockerSecret.createSecret(this.dockerHost, {
@ -92,8 +121,9 @@ export class SparkUpdateManager {
secrets: [newServiceSecret], secrets: [newServiceSecret],
ports: [`${service.port}:${service.secretJson.SERVEZONE_PORT}`], ports: [`${service.port}:${service.secretJson.SERVEZONE_PORT}`],
}); });
logger.log('success', 'updated all services!'); logger.log('ok', `updated service >>${newService.Spec.Name}<<!`);
} }
logger.log('success', `updated ${services.length} services!`);
} }
} }
} }

View File

@ -1,15 +1,7 @@
import * as plugins from './spark.plugins.js'; import * as plugins from './spark.plugins.js';
import * as paths from './spark.paths.js'; import * as paths from './spark.paths.js';
import { commitinfo } from './00_commitinfo_data.js';
const projectInfoNpm = new plugins.projectinfo.ProjectinfoNpm(paths.packageDir); const projectInfoNpm = new plugins.projectinfo.ProjectinfoNpm(paths.packageDir);
export const logger = new plugins.smartlog.Smartlog({ export const logger = plugins.smartlog.Smartlog.createForCommitinfo(commitinfo);
logContext: {
environment: 'production',
runtime: 'node',
zone: 'baremetal',
company: null,
companyunit: null,
containerName: 'spark',
}
});

View File

@ -3,6 +3,11 @@ import * as path from 'path';
export { path }; export { path };
// @serve.zone scope
import * as servezoneInterfaces from '@serve.zone/interfaces';
export { servezoneInterfaces };
// @apiclient.xyz scope // @apiclient.xyz scope
import * as docker from '@apiclient.xyz/docker'; import * as docker from '@apiclient.xyz/docker';