Compare commits
No commits in common. "master" and "v1.1.0" have entirely different histories.
20
changelog.md
20
changelog.md
@ -1,25 +1,5 @@
|
|||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
## 2024-12-20 - 1.2.2 - fix(core)
|
|
||||||
Refactored configuration management classes and improved service update handling
|
|
||||||
|
|
||||||
- Replaced SparkLocalConfig with SparkConfig for configuration management.
|
|
||||||
- Improved service handling and update check logic.
|
|
||||||
- Consolidated service definition and update logic for better maintainability.
|
|
||||||
|
|
||||||
## 2024-12-19 - 1.2.1 - fix(taskmanager)
|
|
||||||
Remove checkinSlackTask from SparkTaskManager for streamlined task management
|
|
||||||
|
|
||||||
- checkinSlackTask has been removed from the task manager class.
|
|
||||||
- Removal of the slack check-in task allows the system to focus on essential update tasks.
|
|
||||||
|
|
||||||
## 2024-12-18 - 1.2.0 - feat(core)
|
|
||||||
Initial commit of the Spark project with core functionalities for server management and integration with Docker.
|
|
||||||
|
|
||||||
- Add core functionalities for server maintenance and configuration.
|
|
||||||
- Integrate Docker for advanced task scheduling and service management.
|
|
||||||
- Provide CLI commands for daemon management and task execution.
|
|
||||||
|
|
||||||
## 2024-12-18 - 1.1.0 - feat(core)
|
## 2024-12-18 - 1.1.0 - feat(core)
|
||||||
Update package dependencies and improve API integration.
|
Update package dependencies and improve API integration.
|
||||||
|
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@serve.zone/spark",
|
"name": "@serve.zone/spark",
|
||||||
"version": "1.2.2",
|
"version": "1.1.0",
|
||||||
"private": false,
|
"private": false,
|
||||||
"description": "A comprehensive tool for maintaining and configuring servers, integrating with Docker and supporting advanced task scheduling, targeted at the Servezone infrastructure and used by @serve.zone/cloudly as a cluster node server system manager.",
|
"description": "A comprehensive tool for maintaining and configuring servers, integrating with Docker and supporting advanced task scheduling, targeted at the Servezone infrastructure and used by @serve.zone/cloudly as a cluster node server system manager.",
|
||||||
"main": "dist_ts/index.js",
|
"main": "dist_ts/index.js",
|
||||||
|
2647
pnpm-lock.yaml
generated
2647
pnpm-lock.yaml
generated
File diff suppressed because it is too large
Load Diff
@ -3,6 +3,6 @@
|
|||||||
*/
|
*/
|
||||||
export const commitinfo = {
|
export const commitinfo = {
|
||||||
name: '@serve.zone/spark',
|
name: '@serve.zone/spark',
|
||||||
version: '1.2.2',
|
version: '1.1.0',
|
||||||
description: 'A comprehensive tool for maintaining and configuring servers, integrating with Docker and supporting advanced task scheduling, targeted at the Servezone infrastructure and used by @serve.zone/cloudly as a cluster node server system manager.'
|
description: 'A comprehensive tool for maintaining and configuring servers, integrating with Docker and supporting advanced task scheduling, targeted at the Servezone infrastructure and used by @serve.zone/cloudly as a cluster node server system manager.'
|
||||||
}
|
}
|
||||||
|
@ -3,12 +3,7 @@ import { Spark } from './index.js';
|
|||||||
|
|
||||||
export class SparkConfig {
|
export class SparkConfig {
|
||||||
public sparkRef: Spark;
|
public sparkRef: Spark;
|
||||||
public kvStore: plugins.npmextra.KeyValueStore;
|
|
||||||
constructor(sparkRefArg: Spark) {
|
constructor(sparkRefArg: Spark) {
|
||||||
this.sparkRef = sparkRefArg;
|
this.sparkRef = sparkRefArg;
|
||||||
this.kvStore = new plugins.npmextra.KeyValueStore({
|
|
||||||
typeArg: 'userHomeDir',
|
|
||||||
identityArg: 'servezone_spark',
|
|
||||||
});
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
15
ts/spark.classes.localconfig.ts
Normal file
15
ts/spark.classes.localconfig.ts
Normal file
@ -0,0 +1,15 @@
|
|||||||
|
import * as plugins from './spark.plugins.js';
|
||||||
|
import { Spark } from './index.js';
|
||||||
|
|
||||||
|
export class SparkLocalConfig {
|
||||||
|
public sparkRef: Spark;
|
||||||
|
private kvStore: plugins.npmextra.KeyValueStore;
|
||||||
|
|
||||||
|
constructor(sparkRefArg: Spark) {
|
||||||
|
this.sparkRef = sparkRefArg;
|
||||||
|
this.kvStore = new plugins.npmextra.KeyValueStore({
|
||||||
|
typeArg: 'userHomeDir',
|
||||||
|
identityArg: 'servezone_spark',
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
@ -1,23 +1,23 @@
|
|||||||
import * as plugins from './spark.plugins.js';
|
import * as plugins from './spark.plugins.js';
|
||||||
import { SparkTaskManager } from './spark.classes.taskmanager.js';
|
import { SparkTaskManager } from './spark.classes.taskmanager.js';
|
||||||
import { SparkInfo } from './spark.classes.info.js';
|
import { SparkInfo } from './spark.classes.info.js';
|
||||||
import { SparkServicesManager } from './spark.classes.updatemanager.js';
|
import { SparkUpdateManager } from './spark.classes.updatemanager.js';
|
||||||
import { logger } from './spark.logging.js';
|
import { logger } from './spark.logging.js';
|
||||||
import { SparkConfig } from './spark.classes.config.js';
|
import { SparkLocalConfig } from './spark.classes.localconfig.js';
|
||||||
|
|
||||||
export class Spark {
|
export class Spark {
|
||||||
public smartdaemon: plugins.smartdaemon.SmartDaemon;
|
public smartdaemon: plugins.smartdaemon.SmartDaemon;
|
||||||
public sparkConfig: SparkConfig;
|
public sparkLocalConfig: SparkLocalConfig;
|
||||||
public sparkTaskManager: SparkTaskManager;
|
public sparkTaskManager: SparkTaskManager;
|
||||||
public sparkInfo: SparkInfo;
|
public sparkInfo: SparkInfo;
|
||||||
public sparkUpdateManager: SparkServicesManager;
|
public sparkUpdateManager: SparkUpdateManager;
|
||||||
|
|
||||||
constructor() {
|
constructor() {
|
||||||
this.smartdaemon = new plugins.smartdaemon.SmartDaemon();
|
this.smartdaemon = new plugins.smartdaemon.SmartDaemon();
|
||||||
this.sparkConfig = new SparkConfig(this);
|
this.sparkLocalConfig = new SparkLocalConfig(this);
|
||||||
this.sparkInfo = new SparkInfo(this);
|
this.sparkInfo = new SparkInfo(this);
|
||||||
this.sparkTaskManager = new SparkTaskManager(this);
|
this.sparkTaskManager = new SparkTaskManager(this);
|
||||||
this.sparkUpdateManager = new SparkServicesManager(this);
|
this.sparkUpdateManager = new SparkUpdateManager(this);
|
||||||
}
|
}
|
||||||
|
|
||||||
public async daemonStart() {
|
public async daemonStart() {
|
||||||
|
@ -8,14 +8,25 @@ export class SparkTaskManager {
|
|||||||
public taskmanager: plugins.taskbuffer.TaskManager;
|
public taskmanager: plugins.taskbuffer.TaskManager;
|
||||||
|
|
||||||
// tasks
|
// tasks
|
||||||
|
public checkinSlackTask: plugins.taskbuffer.Task;
|
||||||
public updateSpark: plugins.taskbuffer.Task;
|
public updateSpark: plugins.taskbuffer.Task;
|
||||||
public updateHost: plugins.taskbuffer.Task;
|
public updateHost: plugins.taskbuffer.Task;
|
||||||
public updateServices: plugins.taskbuffer.Task;
|
public updateCloudly: plugins.taskbuffer.Task;
|
||||||
|
|
||||||
constructor(sparkRefArg: Spark) {
|
constructor(sparkRefArg: Spark) {
|
||||||
this.sparkRef = sparkRefArg;
|
this.sparkRef = sparkRefArg;
|
||||||
this.taskmanager = new plugins.taskbuffer.TaskManager();
|
this.taskmanager = new plugins.taskbuffer.TaskManager();
|
||||||
|
|
||||||
|
// checkinOnSlack
|
||||||
|
this.checkinSlackTask = new plugins.taskbuffer.Task({
|
||||||
|
name: 'checkinSlack',
|
||||||
|
taskFunction: async () => {
|
||||||
|
logger.log('ok', 'running hourly checkin now');
|
||||||
|
|
||||||
|
logger.log('info', 'completed hourly checkin');
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
// updateSpark
|
// updateSpark
|
||||||
this.updateSpark = new plugins.taskbuffer.Task({
|
this.updateSpark = new plugins.taskbuffer.Task({
|
||||||
name: 'updateSpark',
|
name: 'updateSpark',
|
||||||
@ -56,10 +67,7 @@ export class SparkTaskManager {
|
|||||||
},
|
},
|
||||||
});
|
});
|
||||||
|
|
||||||
/**
|
this.updateCloudly = new plugins.taskbuffer.Task({
|
||||||
* only being run when mode is cloudly
|
|
||||||
*/
|
|
||||||
this.updateServices = new plugins.taskbuffer.Task({
|
|
||||||
name: 'updateCloudly',
|
name: 'updateCloudly',
|
||||||
taskFunction: async () => {
|
taskFunction: async () => {
|
||||||
logger.log('info', 'now running updateCloudly task');
|
logger.log('info', 'now running updateCloudly task');
|
||||||
@ -72,9 +80,10 @@ export class SparkTaskManager {
|
|||||||
* start the taskmanager
|
* start the taskmanager
|
||||||
*/
|
*/
|
||||||
public async start() {
|
public async start() {
|
||||||
this.taskmanager.addAndScheduleTask(this.updateServices, '30 */2 * * * *');
|
this.taskmanager.addAndScheduleTask(this.checkinSlackTask, '0 0 * * * *');
|
||||||
this.taskmanager.addAndScheduleTask(this.updateSpark, '0 * * * * *');
|
this.taskmanager.addAndScheduleTask(this.updateSpark, '0 * * * * *');
|
||||||
this.taskmanager.addAndScheduleTask(this.updateHost, '0 0 0 * * *');
|
this.taskmanager.addAndScheduleTask(this.updateHost, '0 0 0 * * *');
|
||||||
|
this.taskmanager.addAndScheduleTask(this.updateCloudly, '30 */2 * * * *');
|
||||||
this.taskmanager.start();
|
this.taskmanager.start();
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -82,9 +91,10 @@ export class SparkTaskManager {
|
|||||||
* stops the taskmanager
|
* stops the taskmanager
|
||||||
*/
|
*/
|
||||||
public async stop() {
|
public async stop() {
|
||||||
|
this.taskmanager.descheduleTask(this.checkinSlackTask);
|
||||||
this.taskmanager.descheduleTask(this.updateSpark);
|
this.taskmanager.descheduleTask(this.updateSpark);
|
||||||
this.taskmanager.descheduleTask(this.updateHost);
|
this.taskmanager.descheduleTask(this.updateHost);
|
||||||
this.taskmanager.descheduleTask(this.updateServices);
|
this.taskmanager.descheduleTask(this.updateCloudly);
|
||||||
this.taskmanager.stop();
|
this.taskmanager.stop();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -3,26 +3,10 @@ import * as paths from './spark.paths.js';
|
|||||||
import { Spark } from './spark.classes.spark.js';
|
import { Spark } from './spark.classes.spark.js';
|
||||||
import { logger } from './spark.logging.js';
|
import { logger } from './spark.logging.js';
|
||||||
|
|
||||||
/**
|
export class SparkUpdateManager {
|
||||||
* this class takes care of updating the services that are managed by spark
|
|
||||||
*/
|
|
||||||
export class SparkServicesManager {
|
|
||||||
public sparkRef: Spark;
|
public sparkRef: Spark;
|
||||||
public dockerHost: plugins.docker.DockerHost;
|
public dockerHost: plugins.docker.DockerHost;
|
||||||
public smartupdate: plugins.smartupdate.SmartUpdate;
|
public smartupdate: plugins.smartupdate.SmartUpdate;
|
||||||
|
|
||||||
/**
|
|
||||||
* the services that are managed by spark
|
|
||||||
*/
|
|
||||||
services: Array<{
|
|
||||||
name: string;
|
|
||||||
image: string;
|
|
||||||
url: string;
|
|
||||||
port: string;
|
|
||||||
environment: string;
|
|
||||||
secretJson: any;
|
|
||||||
}> = [];
|
|
||||||
|
|
||||||
constructor(sparkrefArg: Spark) {
|
constructor(sparkrefArg: Spark) {
|
||||||
this.sparkRef = sparkrefArg;
|
this.sparkRef = sparkrefArg;
|
||||||
this.dockerHost = new plugins.docker.DockerHost({});
|
this.dockerHost = new plugins.docker.DockerHost({});
|
||||||
@ -37,7 +21,60 @@ export class SparkServicesManager {
|
|||||||
}
|
}
|
||||||
|
|
||||||
public async updateServices() {
|
public async updateServices() {
|
||||||
for (const service of this.services) {
|
if (
|
||||||
|
plugins.smartfile.fs.isDirectory(plugins.path.join(paths.homeDir, 'serve.zone/spark')) &&
|
||||||
|
(await plugins.smartfile.fs.fileExists(
|
||||||
|
plugins.path.join(paths.homeDir, 'serve.zone/spark/spark.json')
|
||||||
|
))
|
||||||
|
) {
|
||||||
|
const services: Array<{
|
||||||
|
name: string;
|
||||||
|
image: string;
|
||||||
|
url: string;
|
||||||
|
port: string;
|
||||||
|
environment: string;
|
||||||
|
secretJson: any;
|
||||||
|
}> = [];
|
||||||
|
// lets add coreflow
|
||||||
|
services.push({
|
||||||
|
name: `coreflow`,
|
||||||
|
image: `code.foss.global/serve.zone/coreflow`,
|
||||||
|
url: `coreflow`,
|
||||||
|
environment: `production`,
|
||||||
|
port: `3000`,
|
||||||
|
secretJson: {
|
||||||
|
SERVEZONE_PORT: `3000`,
|
||||||
|
SERVEZONE_ENVIRONMENT: `production`,
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
|
services.push({
|
||||||
|
name: `coretraffic`,
|
||||||
|
image: `code.foss.global/serve.zone/coretraffic`,
|
||||||
|
url: `coreflow`,
|
||||||
|
environment: `production`,
|
||||||
|
port: `3000`,
|
||||||
|
secretJson: {
|
||||||
|
SERVEZONE_PORT: `3000`,
|
||||||
|
SERVEZONE_ENVIRONMENT: `production`,
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
|
services.push({
|
||||||
|
name: `corelog`,
|
||||||
|
image: `code.foss.global/serve.zone/corelog`,
|
||||||
|
url: `coreflow`,
|
||||||
|
environment: `production`,
|
||||||
|
port: `3000`,
|
||||||
|
secretJson: {
|
||||||
|
SERVEZONE_PORT: `3000`,
|
||||||
|
SERVEZONE_ENVIRONMENT: `production`,
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
|
// lets add coretraffic
|
||||||
|
|
||||||
|
for (const service of services) {
|
||||||
const existingService = await plugins.docker.DockerService.getServiceByName(
|
const existingService = await plugins.docker.DockerService.getServiceByName(
|
||||||
this.dockerHost,
|
this.dockerHost,
|
||||||
service.name
|
service.name
|
||||||
@ -49,11 +86,10 @@ export class SparkServicesManager {
|
|||||||
if (existingService) {
|
if (existingService) {
|
||||||
const needsUpdate: boolean = await existingService.needsUpdate();
|
const needsUpdate: boolean = await existingService.needsUpdate();
|
||||||
if (!needsUpdate) {
|
if (!needsUpdate) {
|
||||||
logger.log('info', `service >>${service.name}<< not needing update.`);
|
logger.log('info', `not needing update.`);
|
||||||
// we simply return here to end the functions
|
// we simply return here to end the functions
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
// continuing here means we need to update the service
|
|
||||||
logger.log('ok', `${service.name} needs to be updated!`);
|
logger.log('ok', `${service.name} needs to be updated!`);
|
||||||
await existingService.remove();
|
await existingService.remove();
|
||||||
await existingServiceSecret.remove();
|
await existingServiceSecret.remove();
|
||||||
@ -76,7 +112,6 @@ export class SparkServicesManager {
|
|||||||
version: await newServiceImage.getVersion(),
|
version: await newServiceImage.getVersion(),
|
||||||
labels: {},
|
labels: {},
|
||||||
});
|
});
|
||||||
|
|
||||||
const newService = await plugins.docker.DockerService.createService(this.dockerHost, {
|
const newService = await plugins.docker.DockerService.createService(this.dockerHost, {
|
||||||
image: newServiceImage,
|
image: newServiceImage,
|
||||||
labels: {},
|
labels: {},
|
||||||
@ -88,7 +123,7 @@ export class SparkServicesManager {
|
|||||||
});
|
});
|
||||||
logger.log('ok', `updated service >>${newService.Spec.Name}<<!`);
|
logger.log('ok', `updated service >>${newService.Spec.Name}<<!`);
|
||||||
}
|
}
|
||||||
|
logger.log('success', `updated ${services.length} services!`);
|
||||||
logger.log('success', `updated ${this.services.length} services!`);
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -45,50 +45,6 @@ export const runCli = async () => {
|
|||||||
smartcliInstance.addCommand('asdaemon').subscribe(async (argvArg) => {
|
smartcliInstance.addCommand('asdaemon').subscribe(async (argvArg) => {
|
||||||
logger.log('success', 'looks like we are running as daemon now');
|
logger.log('success', 'looks like we are running as daemon now');
|
||||||
logger.log('info', 'starting spark in daemon mode');
|
logger.log('info', 'starting spark in daemon mode');
|
||||||
|
|
||||||
// lets determine the mode if specified
|
|
||||||
let mode = argvArg.mode;
|
|
||||||
if (mode === 'cloudly') {
|
|
||||||
await sparkInstance.sparkConfig.kvStore.writeKey('mode', 'cloudly');
|
|
||||||
} else if (mode === 'coreflow-node') {
|
|
||||||
await sparkInstance.sparkConfig.kvStore.writeKey('mode', 'coreflow-node');
|
|
||||||
} else if (mode) {
|
|
||||||
logger.log('error', 'unknown mode specified');
|
|
||||||
process.exit(1);
|
|
||||||
} else {
|
|
||||||
// mode is not specified by cli, lets get it from the config
|
|
||||||
mode = await sparkInstance.sparkConfig.kvStore.readKey('mode');
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!mode) {
|
|
||||||
logger.log('error', 'no mode specified by either cli or config');
|
|
||||||
process.exit(1);
|
|
||||||
} else if (mode === 'cloudly') {
|
|
||||||
sparkInstance.sparkUpdateManager.services.push({
|
|
||||||
name: `coreflow`,
|
|
||||||
image: `code.foss.global/serve.zone/cloudly`,
|
|
||||||
url: `cloudly`,
|
|
||||||
environment: `production`,
|
|
||||||
port: `3000`,
|
|
||||||
secretJson: {
|
|
||||||
SERVEZONE_PORT: `3000`,
|
|
||||||
SERVEZONE_ENVIRONMENT: `production`,
|
|
||||||
},
|
|
||||||
});
|
|
||||||
} else if (mode === 'coreflow-node') {
|
|
||||||
sparkInstance.sparkUpdateManager.services.push({
|
|
||||||
name: `coreflow`,
|
|
||||||
image: `code.foss.global/serve.zone/coreflow`,
|
|
||||||
url: `coreflow`,
|
|
||||||
environment: `production`,
|
|
||||||
port: `3000`,
|
|
||||||
secretJson: {
|
|
||||||
SERVEZONE_PORT: `3000`,
|
|
||||||
SERVEZONE_ENVIRONMENT: `production`,
|
|
||||||
},
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
await sparkInstance.daemonStart();
|
await sparkInstance.daemonStart();
|
||||||
});
|
});
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user