Compare commits

..

3 Commits

Author SHA1 Message Date
Ivan Li
0d71700f11 chore: debug log 仅输出app的log 2021-03-28 10:17:29 +08:00
Ivan Li
bd045c6abe fix(commons): fix sanitize not return value. 2021-03-27 22:31:21 +08:00
Ivan
8e3dea7099 feat(webhooks): add gitea webhooks api. 2021-03-27 21:47:32 +08:00
76 changed files with 9028 additions and 13424 deletions

1
.gitignore vendored
View File

@ -34,4 +34,3 @@ lerna-debug.log*
!.vscode/extensions.json !.vscode/extensions.json
/config.yml /config.yml
tsconfig.build.tsbuildinfo

View File

@ -1,3 +0,0 @@
{
"recommendations": []
}

0
.vscode/launch.json vendored
View File

View File

@ -1,19 +1,11 @@
{ {
"cSpell.words": [ "cSpell.words": [
"Mutex",
"Repos", "Repos",
"amqp",
"boardcat", "boardcat",
"errout",
"fanout",
"gitea", "gitea",
"golevelup",
"lpush", "lpush",
"lrange", "lrange",
"metatype", "metatype",
"pmessage",
"psubscribe",
"rabbitmq",
"rpop", "rpop",
"rpush" "rpush"
] ]

View File

@ -14,11 +14,5 @@ db:
port: 6379 port: 6379
password: password:
prefix: fennec prefix: fennec
rabbitmq:
uri: 'amqp://fennec:fennec@192.168.31.194:5672'
etcd:
hosts:
- 'http://192.168.31.194:2379'
workspaces: workspaces:
root: '/Users/ivanli/Projects/fennec/workspaces' root: '/Users/ivanli/Projects/fennec/workspaces'

View File

@ -1,14 +0,0 @@
module.exports = {
apps: [
{
name: 'fennec-be',
script: 'npm',
args: 'run start:prod',
watch: false,
ignore_watch: ['node_modules'],
log_date_format: 'MM-DD HH:mm:ss.SSS Z',
env: {},
max_restarts: 5,
},
],
};

19266
package-lock.json generated

File diff suppressed because it is too large Load Diff

View File

@ -1,8 +1,8 @@
{ {
"name": "fennec-be", "name": "fennec-be",
"version": "0.1.1", "version": "0.0.1",
"description": "a ci/cd project.", "description": "",
"author": "Ivan Li\b<ivanli2048@gmail.com>", "author": "",
"private": true, "private": true,
"license": "UNLICENSED", "license": "UNLICENSED",
"scripts": { "scripts": {
@ -21,34 +21,29 @@
"test:e2e": "jest --config ./test/jest-e2e.json" "test:e2e": "jest --config ./test/jest-e2e.json"
}, },
"dependencies": { "dependencies": {
"@golevelup/nestjs-rabbitmq": "^1.16.1", "@nestjs/bull": "^0.3.1",
"@nestjs-lib/auth": "^0.2.1",
"@nestjs/common": "^7.5.1", "@nestjs/common": "^7.5.1",
"@nestjs/config": "^0.6.2", "@nestjs/config": "^0.6.2",
"@nestjs/core": "^7.5.1", "@nestjs/core": "^7.5.1",
"@nestjs/graphql": "^7.9.8", "@nestjs/graphql": "^7.9.8",
"@nestjs/platform-express": "^7.5.1", "@nestjs/platform-express": "^7.5.1",
"@nestjs/typeorm": "^7.1.5", "@nestjs/typeorm": "^7.1.5",
"@types/amqplib": "^0.8.0", "@neuralegion/class-sanitizer": "^0.3.2",
"@types/bull": "^3.15.0",
"@types/ramda": "^0.27.38", "@types/ramda": "^0.27.38",
"apollo-server-express": "^2.19.2", "apollo-server-express": "^2.19.2",
"bcrypt": "^5.0.0", "bcrypt": "^5.0.0",
"body-parser": "^1.19.0", "body-parser": "^1.19.0",
"bull": "^3.20.1",
"class-transformer": "^0.3.2", "class-transformer": "^0.3.2",
"class-validator": "^0.13.1", "class-validator": "^0.13.1",
"debug": "^4.3.1", "debug": "^4.3.1",
"graphql": "^15.5.0", "graphql": "^15.5.0",
"graphql-tools": "^8.1.0", "graphql-tools": "^7.0.2",
"ioredis": "^4.25.0",
"jose": "^3.14.0",
"js-yaml": "^4.0.0", "js-yaml": "^4.0.0",
"nestjs-etcd": "^0.2.0",
"nestjs-pino": "^1.4.0",
"nestjs-redis": "^1.2.8", "nestjs-redis": "^1.2.8",
"observable-to-async-generator": "^1.0.1-rc", "observable-to-async-generator": "^1.0.1-rc",
"pg": "^8.5.1", "pg": "^8.5.1",
"pino-pretty": "^4.7.1",
"pm2": "^5.1.0",
"ramda": "^0.27.1", "ramda": "^0.27.1",
"reflect-metadata": "^0.1.13", "reflect-metadata": "^0.1.13",
"rimraf": "^3.0.2", "rimraf": "^3.0.2",
@ -57,17 +52,15 @@
"typeorm": "^0.2.30" "typeorm": "^0.2.30"
}, },
"devDependencies": { "devDependencies": {
"@nestjs/cli": "^7.6.0", "@nestjs/cli": "^7.5.7",
"@nestjs/schematics": "^7.1.3", "@nestjs/schematics": "^7.1.3",
"@nestjs/testing": "^7.5.1", "@nestjs/testing": "^7.5.1",
"@types/body-parser": "^1.19.0", "@types/body-parser": "^1.19.0",
"@types/debug": "^4.1.5", "@types/debug": "^4.1.5",
"@types/express": "^4.17.8", "@types/express": "^4.17.8",
"@types/ioredis": "^4.22.2",
"@types/jest": "^26.0.15", "@types/jest": "^26.0.15",
"@types/js-yaml": "^4.0.0", "@types/js-yaml": "^4.0.0",
"@types/node": "^14.14.6", "@types/node": "^14.14.6",
"@types/pino-pretty": "^4.7.0",
"@types/supertest": "^2.0.10", "@types/supertest": "^2.0.10",
"@typescript-eslint/eslint-plugin": "^4.6.1", "@typescript-eslint/eslint-plugin": "^4.6.1",
"@typescript-eslint/parser": "^4.6.1", "@typescript-eslint/parser": "^4.6.1",
@ -97,9 +90,6 @@
"collectCoverageFrom": [ "collectCoverageFrom": [
"**/*.(t|j)s" "**/*.(t|j)s"
], ],
"moduleNameMapper": {
"^jose/(.*)$": "<rootDir>/../node_modules/jose/dist/node/cjs/$1"
},
"coverageDirectory": "../coverage", "coverageDirectory": "../coverage",
"testEnvironment": "node" "testEnvironment": "node"
} }

View File

@ -1,4 +1,3 @@
import { CommonsModule } from './commons/commons.module';
import { MiddlewareConsumer, Module, NestModule } from '@nestjs/common'; import { MiddlewareConsumer, Module, NestModule } from '@nestjs/common';
import { ConfigModule, ConfigService } from '@nestjs/config'; import { ConfigModule, ConfigService } from '@nestjs/config';
import { GraphQLModule } from '@nestjs/graphql'; import { GraphQLModule } from '@nestjs/graphql';
@ -13,36 +12,15 @@ import { PipelineTasksModule } from './pipeline-tasks/pipeline-tasks.module';
import configuration from './commons/config/configuration'; import configuration from './commons/config/configuration';
import { RedisModule } from 'nestjs-redis'; import { RedisModule } from 'nestjs-redis';
import { WebhooksModule } from './webhooks/webhooks.module'; import { WebhooksModule } from './webhooks/webhooks.module';
import { RawBodyMiddleware } from './commons/middleware/raw-body.middleware'; import { RawBodyMiddleware } from './commons/middlewares/raw-body.middleware';
import { GiteaWebhooksController } from './webhooks/gitea-webhooks.controller'; import { GiteaWebhooksController } from './webhooks/gitea-webhooks.controller';
import { ParseBodyMiddleware } from './commons/middleware/parse-body.middleware'; import { ParseBodyMiddleware } from './commons/middlewares/parse-body.middleware';
import { LoggerModule } from 'nestjs-pino';
import { EtcdModule } from 'nestjs-etcd';
import pinoPretty from 'pino-pretty';
import { fromPairs, map, pipe, toPairs } from 'ramda';
@Module({ @Module({
imports: [ imports: [
ConfigModule.forRoot({ ConfigModule.forRoot({
load: [configuration], load: [configuration],
}), }),
LoggerModule.forRootAsync({
imports: [ConfigModule],
useFactory: (configService: ConfigService) => {
const isDev = configService.get<'dev' | 'prod'>('env') === 'dev';
return {
pinoHttp: {
prettyPrint: isDev
? {
levelFirst: true,
}
: false,
prettifier: pinoPretty,
},
};
},
inject: [ConfigService],
}),
TypeOrmModule.forRootAsync({ TypeOrmModule.forRootAsync({
imports: [ConfigModule], imports: [ConfigModule],
useFactory: (configService: ConfigService) => ({ useFactory: (configService: ConfigService) => ({
@ -64,22 +42,6 @@ import { fromPairs, map, pipe, toPairs } from 'ramda';
playground: true, playground: true,
autoSchemaFile: true, autoSchemaFile: true,
installSubscriptionHandlers: true, installSubscriptionHandlers: true,
context: ({ req, connection, ...args }) => {
return connection ? { req: connection.context } : { req };
},
subscriptions: {
onConnect: (connectionParams: Record<string, string>) => {
const connectionParamsWithLowerKeys = pipe(
toPairs,
map(([key, value]) => [key.toLowerCase(), value]),
fromPairs,
)(connectionParams);
return {
headers: connectionParamsWithLowerKeys,
};
},
},
}), }),
inject: [ConfigService], inject: [ConfigService],
}), }),
@ -97,15 +59,7 @@ import { fromPairs, map, pipe, toPairs } from 'ramda';
}), }),
inject: [ConfigService], inject: [ConfigService],
}), }),
EtcdModule.forRootAsync({
imports: [ConfigModule],
useFactory: (configService: ConfigService) => ({
hosts: configService.get<string>('db.etcd.hosts', 'localhost:2379'),
}),
inject: [ConfigService],
}),
WebhooksModule, WebhooksModule,
CommonsModule,
], ],
controllers: [AppController], controllers: [AppController],
providers: [AppService, AppResolver], providers: [AppService, AppResolver],

View File

@ -1,11 +1,8 @@
import { Module } from '@nestjs/common'; import { Module } from '@nestjs/common';
import { PasswordConverter } from './services/password-converter'; import { PasswordConverter } from './services/password-converter';
import { RedisMutexModule } from './redis-mutex/redis-mutex.module';
import { AuthModule } from '@nestjs-lib/auth';
@Module({ @Module({
imports: [RedisMutexModule, AuthModule],
providers: [PasswordConverter], providers: [PasswordConverter],
exports: [PasswordConverter, RedisMutexModule, AuthModule], exports: [PasswordConverter],
}) })
export class CommonsModule {} export class CommonsModule {}

View File

@ -1,11 +1,11 @@
import { pick } from 'ramda';
export class ApplicationException extends Error { export class ApplicationException extends Error {
code: number; code: number;
error: Error; error: Error;
constructor( constructor(
message: string | { error?: Error; message?: string | any; code?: number }, message:
| string
| { error?: Error; message?: string | object; code?: number },
) { ) {
if (message instanceof Object) { if (message instanceof Object) {
super(); super();
@ -18,8 +18,4 @@ export class ApplicationException extends Error {
super((message as unknown) as any); super((message as unknown) as any);
} }
} }
toJSON() {
return pick(['code', 'message'], this);
}
} }

View File

@ -6,18 +6,13 @@ import {
HttpStatus, HttpStatus,
} from '@nestjs/common'; } from '@nestjs/common';
import { ApolloError } from 'apollo-server-errors'; import { ApolloError } from 'apollo-server-errors';
import { PinoLogger, InjectPinoLogger } from 'nestjs-pino';
@Catch(HttpException) @Catch(HttpException)
export class HttpExceptionFilter implements ExceptionFilter { export class HttpExceptionFilter implements ExceptionFilter {
constructor(
@InjectPinoLogger(HttpExceptionFilter.name)
private readonly logger: PinoLogger,
) {}
catch(exception: HttpException, host: ArgumentsHost) { catch(exception: HttpException, host: ArgumentsHost) {
switch (host.getType<'http' | 'graphql' | string>()) { switch (host.getType<'http' | 'graphql' | string>()) {
case 'graphql': { case 'graphql': {
const errorName = exception.message; const message = exception.message;
const extensions: Record<string, any> = {}; const extensions: Record<string, any> = {};
const err = exception.getResponse(); const err = exception.getResponse();
if (typeof err === 'string') { if (typeof err === 'string') {
@ -26,10 +21,8 @@ export class HttpExceptionFilter implements ExceptionFilter {
Object.assign(extensions, (err as any).extension); Object.assign(extensions, (err as any).extension);
extensions.message = (err as any).message; extensions.message = (err as any).message;
} }
extensions.error = errorName;
this.logger.error(extensions);
return new ApolloError( return new ApolloError(
extensions.message, message,
exception.getStatus().toString(), exception.getStatus().toString(),
extensions, extensions,
); );

View File

@ -1,5 +1,5 @@
import { ArgumentMetadata, Injectable, PipeTransform } from '@nestjs/common'; import { ArgumentMetadata, Injectable, PipeTransform } from '@nestjs/common';
import { plainToClass } from 'class-transformer'; import { sanitize } from '@neuralegion/class-sanitizer/dist';
@Injectable() @Injectable()
export class SanitizePipe implements PipeTransform { export class SanitizePipe implements PipeTransform {
@ -12,11 +12,13 @@ export class SanitizePipe implements PipeTransform {
return value; return value;
} }
const constructorFunction = metadata.metatype; const constructorFunction = metadata.metatype;
if (!constructorFunction || value instanceof constructorFunction) { if (!constructorFunction) {
return value; return value;
} }
value = Object.assign(new constructorFunction(), value);
try { try {
return plainToClass(constructorFunction, value); sanitize(value);
return value;
} catch (err) { } catch (err) {
console.error(err); console.error(err);
throw err; throw err;

View File

@ -1,10 +0,0 @@
import { Module } from '@nestjs/common';
import { RedisMutexService } from './redis-mutex.service';
import { RedisModule } from 'nestjs-redis';
@Module({
imports: [RedisModule],
providers: [RedisMutexService],
exports: [RedisMutexService],
})
export class RedisMutexModule {}

View File

@ -1,71 +0,0 @@
import { Injectable } from '@nestjs/common';
import { RedisService } from 'nestjs-redis';
import * as uuid from 'uuid';
import { ApplicationException } from '../exceptions/application.exception';
export interface RedisMutexOption {
/**
* seconds
*/
expires?: number;
/**
* seconds
*/
timeout?: number | null;
/**
* milliseconds
*/
retryDelay?: number;
}
@Injectable()
export class RedisMutexService {
constructor(private readonly redisClient: RedisService) {}
public async lock(
key: string,
{ expires = 100, timeout = 10, retryDelay = 100 }: RedisMutexOption = {
expires: 100,
timeout: 10,
retryDelay: 100,
},
) {
const redisKey = `${'mutex-lock'}:${key}`;
const redis = this.redisClient.getClient();
const value = uuid.v4();
const timeoutAt = timeout ? Date.now() + timeout * 1000 : null;
while (
!(await redis
.set(redisKey, value, 'EX', expires, 'NX')
.then(() => true)
.catch(() => false))
) {
if (timeoutAt && timeoutAt > Date.now()) {
throw new ApplicationException('lock timeout');
}
await new Promise((resolve) => setTimeout(resolve, retryDelay));
}
const renewTimer = setInterval(() => {
redis.expire(redisKey, expires);
}, (expires * 1000) / 2);
return async () => {
clearInterval(renewTimer);
await redis.eval(
`
if redis.call("get", KEYS[1]) == ARGV[1]
then
return redis.call("del", KEYS[1])
else
return 0
end
`,
1,
redisKey,
value,
);
};
}
}

View File

@ -53,22 +53,8 @@ export class BaseDbService<Entity extends AppBaseEntity> extends TypeormHelper {
async isDuplicateEntityForUpdate<Dto extends Entity>( async isDuplicateEntityForUpdate<Dto extends Entity>(
id: string, id: string,
dto: Partial<Dto>, dto: Partial<Dto>,
extendsFields?: Array<keyof Dto & string>,
): Promise<false | never>;
async isDuplicateEntityForUpdate<Dto extends Entity>(
old: Entity,
dto: Partial<Dto>,
extendsFields?: Array<keyof Dto & string>,
): Promise<false | never>;
async isDuplicateEntityForUpdate<Dto extends Entity>(
id: string | Entity,
dto: Partial<Dto>,
extendsFields: Array<keyof Dto & string> = [], extendsFields: Array<keyof Dto & string> = [],
): Promise<false | never> { ): Promise<false | never> {
if (typeof id !== 'string') {
dto = Object.assign({}, id, dto);
id = id.id;
}
const qb = this.repository.createQueryBuilder('entity'); const qb = this.repository.createQueryBuilder('entity');
const compareFields = this.getCompareFields(dto, [ const compareFields = this.getCompareFields(dto, [
...this.uniqueFields, ...this.uniqueFields,

View File

@ -1,21 +0,0 @@
import { hostname } from 'os';
export function getInstanceName() {
return hostname();
}
export function getSelfInstanceRouteKey(key: string) {
return getAppInstanceRouteKey(key, getInstanceName());
}
export function getAppInstanceRouteKey(key: string, appInstance?: string) {
return appInstance ? `${key}.${appInstance}` : key;
}
export function getSelfInstanceQueueKey(key: string) {
return getAppInstanceQueueKey(key, getInstanceName());
}
export function getAppInstanceQueueKey(key: string, appInstance?: string) {
return appInstance ? `${key}.${appInstance}` : key;
}

View File

@ -1,4 +1,3 @@
import { PinoLogger } from 'nestjs-pino';
import { ValidationPipe } from '@nestjs/common'; import { ValidationPipe } from '@nestjs/common';
import { ConfigService } from '@nestjs/config'; import { ConfigService } from '@nestjs/config';
import { NestFactory } from '@nestjs/core'; import { NestFactory } from '@nestjs/core';
@ -15,8 +14,7 @@ async function bootstrap() {
transform: true, transform: true,
}), }),
); );
const httpExceptionFilterLogger = await app.resolve(PinoLogger); app.useGlobalFilters(new HttpExceptionFilter());
app.useGlobalFilters(new HttpExceptionFilter(httpExceptionFilterLogger));
await app.listen(configService.get<number>('http.port')); await app.listen(configService.get<number>('http.port'));
} }
bootstrap(); bootstrap();

View File

@ -11,5 +11,3 @@ registerEnumType(TaskStatuses, {
name: 'TaskStatuses', name: 'TaskStatuses',
description: '任务状态', description: '任务状态',
}); });
export const terminalTaskStatuses = [TaskStatuses.success, TaskStatuses.failed];

View File

@ -1,25 +0,0 @@
import { Field, ObjectType } from '@nestjs/graphql';
import { PipelineUnits } from '../enums/pipeline-units.enum';
import { TaskStatuses } from '../enums/task-statuses.enum';
import { Type } from 'class-transformer';
@ObjectType()
export class PipelineTaskEvent {
@Field()
taskId: string;
@Field()
pipelineId: string;
@Field()
projectId: string;
@Field(() => PipelineUnits, { nullable: true })
unit: PipelineUnits | null;
@Field()
@Type(() => Date)
emittedAt: Date;
@Field()
message: string;
@Field()
messageType: 'stdout' | 'stderr' | 'stdin';
@Field(() => TaskStatuses)
status: TaskStatuses;
}

View File

@ -1,7 +1,6 @@
import { PipelineTask } from './../pipeline-task.entity'; import { PipelineTask } from './../pipeline-task.entity';
import { PipelineUnits } from '../enums/pipeline-units.enum'; import { PipelineUnits } from '../enums/pipeline-units.enum';
import { Field, HideField, ObjectType } from '@nestjs/graphql'; import { Field, HideField, ObjectType } from '@nestjs/graphql';
import { Type } from 'class-transformer';
@ObjectType() @ObjectType()
export class PipelineTaskLogMessage { export class PipelineTaskLogMessage {
@ -10,7 +9,6 @@ export class PipelineTaskLogMessage {
@Field(() => PipelineUnits, { nullable: true }) @Field(() => PipelineUnits, { nullable: true })
unit?: PipelineUnits; unit?: PipelineUnits;
@Field() @Field()
@Type(() => Date)
time: Date; time: Date;
@Field() @Field()
message: string; message: string;

View File

@ -1,7 +1,6 @@
import { TaskStatuses } from '../enums/task-statuses.enum'; import { TaskStatuses } from '../enums/task-statuses.enum';
import { PipelineUnits } from '../enums/pipeline-units.enum'; import { PipelineUnits } from '../enums/pipeline-units.enum';
import { Field, ObjectType } from '@nestjs/graphql'; import { Field, ObjectType } from '@nestjs/graphql';
import { Type } from 'class-transformer';
@ObjectType() @ObjectType()
export class PipelineTaskLogs { export class PipelineTaskLogs {
@ -9,9 +8,7 @@ export class PipelineTaskLogs {
unit: PipelineUnits; unit: PipelineUnits;
@Field(() => TaskStatuses) @Field(() => TaskStatuses)
status: TaskStatuses; status: TaskStatuses;
@Type(() => Date)
startedAt?: Date; startedAt?: Date;
@Type(() => Date)
endedAt?: Date; endedAt?: Date;
logs = ''; logs = '';
} }

View File

@ -1,16 +1,9 @@
import { Field, InputType, Int, ObjectType } from '@nestjs/graphql'; import { InputType, ObjectType } from '@nestjs/graphql';
import { Type } from 'class-transformer';
import { IsInstance, isInstance, ValidateNested } from 'class-validator';
import { WorkUnit } from './work-unit.model'; import { WorkUnit } from './work-unit.model';
@InputType('WorkUnitMetadataInput') @InputType('WorkUnitMetadataInput')
@ObjectType() @ObjectType()
export class WorkUnitMetadata { export class WorkUnitMetadata {
@Field(() => Int)
version = 1; version = 1;
@Type(() => WorkUnit)
@IsInstance(WorkUnit, { each: true })
@ValidateNested({ each: true })
units: WorkUnit[]; units: WorkUnit[];
} }

View File

@ -1,5 +1,4 @@
import { Field, InputType, ObjectType } from '@nestjs/graphql'; import { Field, InputType, ObjectType } from '@nestjs/graphql';
import { IsNotEmpty } from 'class-validator';
import { import {
PipelineUnits, PipelineUnits,
PipelineUnits as PipelineUnitTypes, PipelineUnits as PipelineUnitTypes,
@ -10,7 +9,5 @@ import {
export class WorkUnit { export class WorkUnit {
@Field(() => PipelineUnits) @Field(() => PipelineUnits)
type: PipelineUnitTypes; type: PipelineUnitTypes;
@IsNotEmpty({ each: true })
scripts: string[]; scripts: string[];
} }

View File

@ -1,88 +0,0 @@
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
import { Test, TestingModule } from '@nestjs/testing';
import { RedisService } from 'nestjs-redis';
import { PipelineTaskFlushService } from './pipeline-task-flush.service';
import { PipelineTaskEvent } from './models/pipeline-task-event';
import { TaskStatuses } from './enums/task-statuses.enum';
import {
EXCHANGE_PIPELINE_TASK_TOPIC,
ROUTE_PIPELINE_TASK_DONE,
} from './pipeline-tasks.constants';
describe('PipelineTaskFlushService', () => {
let service: PipelineTaskFlushService;
let redisService: RedisService;
let amqpConnection: AmqpConnection;
beforeEach(async () => {
const redisClient = {
rpush: jest.fn(() => Promise.resolve()),
lrange: jest.fn(() => Promise.resolve()),
expire: jest.fn(() => Promise.resolve()),
};
const module: TestingModule = await Test.createTestingModule({
providers: [
PipelineTaskFlushService,
{
provide: RedisService,
useValue: {
getClient() {
return redisClient;
},
},
},
{
provide: AmqpConnection,
useValue: {
request: jest.fn(() => Promise.resolve()),
},
},
],
}).compile();
service = module.get<PipelineTaskFlushService>(PipelineTaskFlushService);
redisService = module.get<RedisService>(RedisService);
amqpConnection = module.get<AmqpConnection>(AmqpConnection);
});
it('should be defined', () => {
expect(service).toBeDefined();
});
describe('write', () => {
const amqpMsg = {
properties: { headers: { sender: 'test' } },
} as any;
it('normal', async () => {
const testEvent = new PipelineTaskEvent();
testEvent.taskId = 'test';
testEvent.status = TaskStatuses.working;
const rpush = jest.spyOn(redisService.getClient(), 'rpush');
const request = jest.spyOn(amqpConnection, 'request');
await service.write(testEvent, amqpMsg);
expect(rpush).toBeCalledTimes(1);
expect(rpush.mock.calls[0][0]).toEqual('p-task:log:test');
expect(rpush.mock.calls[0][1]).toEqual(JSON.stringify(testEvent));
expect(request).toBeCalledTimes(1);
});
it('event for which task done', async () => {
const testEvent = new PipelineTaskEvent();
testEvent.taskId = 'test';
testEvent.status = TaskStatuses.success;
const rpush = jest.spyOn(redisService.getClient(), 'rpush');
const request = jest.spyOn(amqpConnection, 'request');
await service.write(testEvent, amqpMsg);
expect(rpush).toBeCalledTimes(1);
expect(request).toBeCalledTimes(1);
expect(request.mock.calls[0][0]).toMatchObject({
exchange: EXCHANGE_PIPELINE_TASK_TOPIC,
routingKey: ROUTE_PIPELINE_TASK_DONE,
payload: {
taskId: 'test',
status: TaskStatuses.success,
runOn: 'test',
},
});
});
});
});

View File

@ -1,66 +0,0 @@
import { AmqpConnection, RabbitSubscribe } from '@golevelup/nestjs-rabbitmq';
import { Injectable } from '@nestjs/common';
import { ConsumeMessage } from 'amqplib';
import { deserialize } from 'class-transformer';
import { RedisService } from 'nestjs-redis';
import { isNil } from 'ramda';
import { getSelfInstanceQueueKey } from '../commons/utils/rabbit-mq';
import { PipelineTaskEvent } from './models/pipeline-task-event';
import {
EXCHANGE_PIPELINE_TASK_TOPIC,
ROUTE_PIPELINE_TASK_DONE,
} from './pipeline-tasks.constants';
import {
EXCHANGE_PIPELINE_TASK_FANOUT,
ROUTE_PIPELINE_TASK_LOG,
QUEUE_WRITE_PIPELINE_TASK_LOG,
} from './pipeline-tasks.constants';
@Injectable()
export class PipelineTaskFlushService {
constructor(
private readonly redisService: RedisService,
private readonly amqpConnection: AmqpConnection,
) {}
@RabbitSubscribe({
exchange: EXCHANGE_PIPELINE_TASK_FANOUT,
routingKey: ROUTE_PIPELINE_TASK_LOG,
queue: getSelfInstanceQueueKey(QUEUE_WRITE_PIPELINE_TASK_LOG),
queueOptions: {
autoDelete: true,
durable: true,
},
})
async write(message: PipelineTaskEvent, amqpMsg: ConsumeMessage) {
const client = this.redisService.getClient();
await client.rpush(this.getKey(message.taskId), JSON.stringify(message));
await client.expire(this.getKey(message.taskId), 600); // ten minutes
if (isNil(message.unit)) {
try {
await this.amqpConnection.request({
exchange: EXCHANGE_PIPELINE_TASK_TOPIC,
routingKey: ROUTE_PIPELINE_TASK_DONE,
payload: {
taskId: message.taskId,
status: message.status,
runOn: amqpMsg.properties.headers.sender,
},
});
} catch (error) {
console.log(error);
}
}
}
async read(taskId: string) {
const raw = await this.redisService
.getClient()
.lrange(this.getKey(taskId), 0, -1);
return raw.map((it) => deserialize(PipelineTaskEvent, it));
}
private getKey(taskId: string) {
return `p-task:log:${taskId}`;
}
}

View File

@ -1,14 +1,14 @@
import { Test, TestingModule } from '@nestjs/testing'; import { Test, TestingModule } from '@nestjs/testing';
import { PipelineTaskLogsService } from './pipeline-task-logs.service';
import { RedisService } from 'nestjs-redis'; import { RedisService } from 'nestjs-redis';
import { RedisMutexService } from './redis-mutex.service';
describe('RedisMutexService', () => { describe('PipelineTaskLogsService', () => {
let service: RedisMutexService; let service: PipelineTaskLogsService;
beforeEach(async () => { beforeEach(async () => {
const module: TestingModule = await Test.createTestingModule({ const module: TestingModule = await Test.createTestingModule({
providers: [ providers: [
RedisMutexService, PipelineTaskLogsService,
{ {
provide: RedisService, provide: RedisService,
useValue: {}, useValue: {},
@ -16,7 +16,7 @@ describe('RedisMutexService', () => {
], ],
}).compile(); }).compile();
service = module.get<RedisMutexService>(RedisMutexService); service = module.get<PipelineTaskLogsService>(PipelineTaskLogsService);
}); });
it('should be defined', () => { it('should be defined', () => {

View File

@ -0,0 +1,78 @@
import { Injectable } from '@nestjs/common';
import { log } from 'console';
import { PubSub } from 'graphql-subscriptions';
import { RedisService } from 'nestjs-redis';
import { find, omit, propEq } from 'ramda';
import { PipelineUnits } from './enums/pipeline-units.enum';
import { TaskStatuses } from './enums/task-statuses.enum';
import { PipelineTaskLogMessage } from './models/pipeline-task-log-message.module';
import { PipelineTaskLogs } from './models/pipeline-task-logs.model';
import { PipelineTask } from './pipeline-task.entity';
const LOG_TIMEOUT_SECONDS = 10_000;
@Injectable()
export class PipelineTaskLogsService {
constructor(private readonly redisService: RedisService) {}
pubSub = new PubSub();
get redis() {
return this.redisService.getClient();
}
getKeys(task: PipelineTask) {
return `ptl:${task.id}`;
}
async recordLog(log: PipelineTaskLogMessage) {
const logDto = omit(['task'], log);
await Promise.all([
this.pubSub.publish(this.getKeys(log.task), logDto),
this.redis
.expire(this.getKeys(log.task), LOG_TIMEOUT_SECONDS)
.then(() =>
this.redis.rpush(this.getKeys(log.task), JSON.stringify(logDto)),
),
]);
}
async readLog(task: PipelineTask): Promise<PipelineTaskLogMessage[]> {
return await this.redis.lrange(this.getKeys(task), 0, -1).then((items) =>
items.map((item) => {
const log = JSON.parse(item) as PipelineTaskLogMessage;
log.task = task;
log.time = new Date(log.time);
return log;
}),
);
}
async readLogsAsPipelineTaskLogs(
task: PipelineTask,
): Promise<PipelineTaskLogs[]> {
const logs = await this.readLog(task);
const taskLogs: PipelineTaskLogs[] = [];
for (const log of logs) {
const taskLog = find<PipelineTaskLogs>(
propEq('unit', log.unit),
taskLogs,
);
if (!taskLog) {
taskLogs.push({
unit: (log.unit as unknown) as PipelineUnits,
status: TaskStatuses.working,
startedAt: log.time,
logs: log.message,
});
} else {
taskLog.logs += log.message;
}
}
return taskLogs;
}
watchLogs(task: PipelineTask) {
return this.pubSub.asyncIterator(this.getKeys(task));
}
}

View File

@ -0,0 +1,242 @@
import { Test, TestingModule } from '@nestjs/testing';
import { Job } from 'bull';
import { join } from 'path';
import { ReposService } from '../repos/repos.service';
import { PipelineUnits } from './enums/pipeline-units.enum';
import { PipelineTaskConsumer } from './pipeline-task.consumer';
import { PipelineTask } from './pipeline-task.entity';
import { PipelineTasksService } from './pipeline-tasks.service';
import { PipelineTaskLogMessage } from './models/pipeline-task-log-message.module';
import { Pipeline } from '../pipelines/pipeline.entity';
import { Project } from '../projects/project.entity';
import { TaskStatuses } from './enums/task-statuses.enum';
import { PipelineTaskLogsService } from './pipeline-task-logs.service';
import { ApplicationException } from '../commons/exceptions/application.exception';
describe('PipelineTaskConsumer', () => {
let consumer: PipelineTaskConsumer;
let tasksService: PipelineTasksService;
let logsService: PipelineTaskLogsService;
const getJob = () =>
({
data: {
pipelineId: 'test',
units: [PipelineUnits.checkout, PipelineUnits.test],
},
} as Job<PipelineTask>);
beforeEach(async () => {
const module: TestingModule = await Test.createTestingModule({
providers: [
{
provide: PipelineTasksService,
useValue: {
doNextTask: () => undefined,
updateTask: async (value) => value,
},
},
{
provide: ReposService,
useValue: {
getWorkspaceRootByTask: () => 'workspace-root',
checkout: async () => undefined,
},
},
{
provide: PipelineTaskLogsService,
useValue: {
recordLog: async () => undefined,
readLogsAsPipelineTaskLogs: async () => [],
},
},
PipelineTaskConsumer,
],
}).compile();
tasksService = module.get(PipelineTasksService);
logsService = module.get(PipelineTaskLogsService);
consumer = module.get(PipelineTaskConsumer);
});
it('should be defined', () => {
expect(consumer).toBeDefined();
});
describe('onCompleted', () => {
it('should call doNextTask()', () => {
const job = getJob();
const doNextTask = jest.spyOn(tasksService, 'doNextTask');
consumer.onCompleted(job);
expect(doNextTask).toHaveBeenCalledTimes(1);
});
});
describe('runScript', () => {
let logText: string;
let errorText: string;
let recordLog: jest.SpyInstance;
beforeEach(() => {
logText = '';
errorText = '';
recordLog = jest
.spyOn(logsService, 'recordLog')
.mockImplementation(async (log: PipelineTaskLogMessage) => {
logText += log.message;
if (log.isError) {
errorText += log.message;
}
});
});
it('should success and log right message', async () => {
await consumer.runScript(
'node one-second-work.js',
join(__dirname, '../../test/data'),
);
expect(logText).toMatch(/10.+20.+30.+40.+50.+60.+70.+80.+90/s);
expect(recordLog).toHaveBeenCalledTimes(10);
expect(
((recordLog.mock.calls[8][0] as unknown) as PipelineTaskLogMessage)
.message,
).toMatch(/^90/);
});
it('should failed and log right message', async () => {
await expect(
consumer.runScript(
'node bad-work.js',
join(__dirname, '../../test/data'),
),
).rejects.toThrowError(/exec script failed/);
expect(errorText).toMatch(/Error Message/);
const logs = recordLog.mock.calls
.map((call) => ((call[0] as unknown) as PipelineTaskLogMessage).message)
.join('');
expect(logs).toMatch(/10.+20.+30.+40.+50/s);
});
it('should log with task', async () => {
const task = new PipelineTask();
task.id = 'test';
const recordLog = jest.spyOn(logsService, 'recordLog');
await expect(
consumer.runScript(
'node bad-work.js',
join(__dirname, '../../test/data'),
task,
),
).rejects.toThrowError(/exec script failed/);
expect(errorText).toMatch(/Error Message 2/);
expect(
((recordLog.mock.calls[2][0] as unknown) as PipelineTaskLogMessage)
.task,
).toMatchObject(task);
});
});
describe('doTask', () => {
let task: PipelineTask;
beforeEach(() => {
task = new PipelineTask();
task.id = 'test-id';
task.logs = [];
task.pipeline = new Pipeline();
task.pipeline.workUnitMetadata = {
version: 1,
units: [
{
type: PipelineUnits.checkout,
scripts: [],
},
{
type: PipelineUnits.installDependencies,
scripts: ["echo ' Hello, Fennec!'"],
},
],
};
task.units = task.pipeline.workUnitMetadata.units.map(
(unit) => unit.type,
);
task.pipeline.project = new Project();
task.pipeline.project.name = 'test-project';
});
it('success and update task on db', async () => {
const job: Job = ({
data: task,
update: jest.fn().mockImplementation(() => undefined),
} as unknown) as Job;
jest
.spyOn(consumer, 'runScript')
.mockImplementation(async () => undefined);
const updateTask = jest.spyOn(tasksService, 'updateTask');
await consumer.doTask(job);
expect(updateTask).toHaveBeenCalledTimes(2);
expect(updateTask.mock.calls[0][0].startedAt).toBeDefined();
expect(updateTask.mock.calls[1][0].endedAt).toBeDefined();
expect(updateTask.mock.calls[1][0].status).toEqual(TaskStatuses.success);
});
it('failed and update task on db', async () => {
const job: Job = ({
data: task,
update: jest.fn().mockImplementation(() => undefined),
} as unknown) as Job;
jest.spyOn(consumer, 'runScript').mockImplementation(async () => {
throw new ApplicationException('exec script failed');
});
const updateTask = jest.spyOn(tasksService, 'updateTask');
await consumer.doTask(job);
expect(updateTask).toHaveBeenCalledTimes(2);
expect(updateTask.mock.calls[0][0].startedAt).toBeDefined();
expect(updateTask.mock.calls[1][0].endedAt).toBeDefined();
expect(updateTask.mock.calls[1][0].status).toEqual(TaskStatuses.failed);
});
it('should do all task', async () => {
const job: Job = ({
data: task,
update: jest.fn().mockImplementation(() => undefined),
} as unknown) as Job;
const runScript = jest
.spyOn(consumer, 'runScript')
.mockImplementation(async () => undefined);
const updateTask = jest.spyOn(tasksService, 'updateTask');
await consumer.doTask(job);
expect(runScript).toHaveBeenCalledTimes(1);
expect(updateTask).toHaveBeenCalledTimes(2);
const taskDto: PipelineTask = updateTask.mock.calls[0][0];
expect(taskDto.logs).toHaveLength(2);
expect(taskDto.logs[0].status).toEqual(TaskStatuses.success);
expect(taskDto.logs[0].unit).toEqual(PipelineUnits.checkout);
});
it('should log error message', async () => {
const job: Job = ({
data: task,
update: jest.fn().mockImplementation(() => undefined),
} as unknown) as Job;
const runScript = jest
.spyOn(consumer, 'runScript')
.mockImplementation(async () => {
throw new Error('bad message');
});
const updateTask = jest.spyOn(tasksService, 'updateTask');
await consumer.doTask(job);
expect(updateTask).toHaveBeenCalledTimes(2);
const taskDto: PipelineTask = updateTask.mock.calls[0][0];
expect(taskDto.logs).toHaveLength(2);
expect(taskDto.logs[0].status).toEqual(TaskStatuses.success);
expect(taskDto.logs[1].status).toEqual(TaskStatuses.failed);
});
});
});

View File

@ -0,0 +1,147 @@
import { PipelineTaskLogs } from './models/pipeline-task-logs.model';
import { ReposService } from './../repos/repos.service';
import {
OnQueueCompleted,
OnQueueFailed,
Process,
Processor,
} from '@nestjs/bull';
import { Job } from 'bull';
import { spawn } from 'child_process';
import { PipelineTask } from './pipeline-task.entity';
import { PIPELINE_TASK_QUEUE } from './pipeline-tasks.constants';
import { PipelineTasksService } from './pipeline-tasks.service';
import { ApplicationException } from '../commons/exceptions/application.exception';
import { PipelineUnits } from './enums/pipeline-units.enum';
import { PipelineTaskLogMessage } from './models/pipeline-task-log-message.module';
import { TaskStatuses } from './enums/task-statuses.enum';
import { PipelineTaskLogsService } from './pipeline-task-logs.service';
import debug from 'debug';
const log = debug('fennec:pipeline-tasks:consumer');
@Processor(PIPELINE_TASK_QUEUE)
export class PipelineTaskConsumer {
constructor(
private readonly service: PipelineTasksService,
private readonly reposService: ReposService,
private readonly logsService: PipelineTaskLogsService,
) {}
@Process()
async doTask(job: Job<PipelineTask>) {
let task = job.data;
if (task.pipeline.workUnitMetadata.version !== 1) {
throw new ApplicationException(
'work unit metadata version is not match.',
);
}
task.startedAt = new Date();
task.status = TaskStatuses.working;
task = await this.service.updateTask(task);
log('start job');
await job.update(task);
const workspaceRoot = this.reposService.getWorkspaceRootByTask(task);
const units = task.units.map(
(type) =>
task.pipeline.workUnitMetadata.units.find(
(unit) => unit.type === type,
) ?? { type: type, scripts: [] },
);
log('task have [%o] units', units);
try {
for (const unit of units) {
const unitLog = new PipelineTaskLogs();
unitLog.unit = unit.type;
unitLog.startedAt = new Date();
log('curr unit is %s', unit.type);
try {
// 检出代码前执行 git checkout
if (unit.type === PipelineUnits.checkout) {
log('begin checkout');
await this.reposService.checkout(task, workspaceRoot);
unitLog.status = TaskStatuses.success;
log('end checkout');
}
for (const script of unit.scripts) {
unitLog.logs += `[RUN SCRIPT] ${script}`;
log('begin runScript %s', script);
await this.runScript(script, workspaceRoot, task, unit.type);
log('end runScript %s', script);
}
unitLog.status = TaskStatuses.success;
} catch (err) {
unitLog.status = TaskStatuses.failed;
unitLog.logs += err.message;
throw err;
} finally {
unitLog.endedAt = new Date();
unitLog.logs = await this.logsService
.readLogsAsPipelineTaskLogs(task)
.then(
(taskLogs) =>
taskLogs.find((tl) => tl.unit === unit.type)?.logs ?? '',
);
task.logs.push(unitLog);
await job.update(task);
}
}
task.status = TaskStatuses.success;
} catch (err) {
task.status = TaskStatuses.failed;
log('task is failed', err);
} finally {
task.endedAt = new Date();
task = await this.service.updateTask(task);
await job.update(task);
}
}
async runScript(
script: string,
workspaceRoot: string,
task?: PipelineTask,
unit?: PipelineUnits,
): Promise<void> {
return new Promise((resolve, reject) => {
const sub = spawn(script, {
shell: true,
cwd: workspaceRoot,
});
sub.stderr.on('data', (data: Buffer) => {
const str = data.toString();
this.logsService.recordLog(
PipelineTaskLogMessage.create(task, unit, str, true),
);
});
sub.stdout.on('data', (data: Buffer) => {
const str = data.toString();
this.logsService.recordLog(
PipelineTaskLogMessage.create(task, unit, str, false),
);
});
sub.addListener('close', (code) => {
if (code === 0) {
return resolve();
}
return reject(new ApplicationException('exec script failed'));
});
});
}
@OnQueueCompleted()
onCompleted(job: Job<PipelineTask>) {
log('queue onCompleted');
this.service.doNextTask(job.data.pipeline);
}
@OnQueueFailed()
onFailed(job: Job<PipelineTask>) {
log('queue onFailed');
this.service.doNextTask(job.data.pipeline);
}
}

View File

@ -1,16 +1,11 @@
import { AppBaseEntity } from './../commons/entities/app-base-entity'; import { AppBaseEntity } from './../commons/entities/app-base-entity';
import { ObjectType } from '@nestjs/graphql'; import { Field, ObjectType } from '@nestjs/graphql';
import { Column, Entity, ManyToOne, ValueTransformer } from 'typeorm'; import { Column, Entity, ManyToOne } from 'typeorm';
import { Pipeline } from '../pipelines/pipeline.entity'; import { Pipeline } from '../pipelines/pipeline.entity';
import { PipelineTaskLogs } from './models/pipeline-task-logs.model'; import { PipelineTaskLogs } from './models/pipeline-task-logs.model';
import { TaskStatuses } from './enums/task-statuses.enum'; import { TaskStatuses } from './enums/task-statuses.enum';
import { PipelineUnits } from './enums/pipeline-units.enum'; import { PipelineUnits } from './enums/pipeline-units.enum';
import { plainToClass } from 'class-transformer';
const logsTransformer: ValueTransformer = {
from: (value) => plainToClass(PipelineTaskLogs, value),
to: (value) => value,
};
@ObjectType() @ObjectType()
@Entity() @Entity()
export class PipelineTask extends AppBaseEntity { export class PipelineTask extends AppBaseEntity {
@ -25,7 +20,7 @@ export class PipelineTask extends AppBaseEntity {
@Column({ type: 'enum', enum: PipelineUnits, array: true }) @Column({ type: 'enum', enum: PipelineUnits, array: true })
units: PipelineUnits[]; units: PipelineUnits[];
@Column({ type: 'jsonb', default: '[]', transformer: logsTransformer }) @Column({ type: 'jsonb', default: '[]' })
logs: PipelineTaskLogs[]; logs: PipelineTaskLogs[];
@Column({ type: 'enum', enum: TaskStatuses, default: TaskStatuses.pending }) @Column({ type: 'enum', enum: TaskStatuses, default: TaskStatuses.pending })
@ -36,7 +31,4 @@ export class PipelineTask extends AppBaseEntity {
@Column({ nullable: true }) @Column({ nullable: true })
endedAt?: Date; endedAt?: Date;
@Column({ nullable: true })
runOn: string;
} }

View File

@ -1,77 +0,0 @@
import { Test, TestingModule } from '@nestjs/testing';
import { PipelineTaskLogger } from './pipeline-task.logger';
import { PipelineTaskEvent } from './models/pipeline-task-event';
import { take, timeout } from 'rxjs/operators';
describe('PipelineTaskRunner', () => {
let logger: PipelineTaskLogger;
let module: TestingModule;
beforeEach(async () => {
module = await Test.createTestingModule({
providers: [PipelineTaskLogger],
}).compile();
logger = module.get(PipelineTaskLogger);
});
it('should be defined', () => {
expect(logger).toBeDefined();
});
describe('getMessage$', () => {
it('normal', async () => {
const event = new PipelineTaskEvent();
event.taskId = 'test';
const emittedAt = new Date();
event.emittedAt = emittedAt.toISOString() as any;
const message$ = logger.getMessage$('test');
let receiveEvent;
message$.pipe(take(1)).subscribe((value) => (receiveEvent = value));
await logger.handleEvent(event);
expect(receiveEvent).toMatchObject({
...event,
emittedAt,
});
});
it('no match', async () => {
const event = new PipelineTaskEvent();
event.taskId = 'test';
const message$ = logger.getMessage$('other');
setTimeout(() => {
logger.handleEvent(event);
});
await expect(
message$.pipe(take(1), timeout(100)).toPromise(),
).rejects.toThrow(/timeout/i);
});
it('multiple subscribers', async () => {
const event = new PipelineTaskEvent();
event.taskId = 'test';
const message$ = logger.getMessage$('test');
const message2$ = logger.getMessage$('test');
setTimeout(() => {
logger.handleEvent(event);
logger.handleEvent(event);
});
await Promise.all([
expect(
message$.pipe(take(1), timeout(100)).toPromise(),
).resolves.toEqual(event),
expect(
message2$.pipe(take(1), timeout(100)).toPromise(),
).resolves.toEqual(event),
]);
});
});
describe('onModuleDestroy', () => {
it('complete observable when destroying module', async () => {
logger.onModuleDestroy();
await expect(
(logger as any).message$.toPromise(),
).resolves.toBeUndefined();
});
});
});

View File

@ -1,38 +0,0 @@
import { RabbitSubscribe } from '@golevelup/nestjs-rabbitmq';
import { Injectable, OnModuleDestroy } from '@nestjs/common';
import { plainToClass } from 'class-transformer';
import { Observable, Subject } from 'rxjs';
import { filter } from 'rxjs/operators';
import { PipelineTaskEvent } from './models/pipeline-task-event';
import {
EXCHANGE_PIPELINE_TASK_FANOUT,
QUEUE_HANDLE_PIPELINE_TASK_LOG_EVENT,
ROUTE_PIPELINE_TASK_LOG,
} from './pipeline-tasks.constants';
@Injectable()
export class PipelineTaskLogger implements OnModuleDestroy {
private readonly messageSubject = new Subject<PipelineTaskEvent>();
private readonly message$: Observable<PipelineTaskEvent> =
this.messageSubject.pipe();
@RabbitSubscribe({
exchange: EXCHANGE_PIPELINE_TASK_FANOUT,
routingKey: ROUTE_PIPELINE_TASK_LOG,
queue: QUEUE_HANDLE_PIPELINE_TASK_LOG_EVENT,
queueOptions: {
autoDelete: true,
},
})
async handleEvent(message: PipelineTaskEvent) {
this.messageSubject.next(plainToClass(PipelineTaskEvent, message));
}
getMessage$(taskId: string) {
return this.message$.pipe(filter((event) => event.taskId === taskId));
}
onModuleDestroy() {
this.messageSubject.complete();
}
}

View File

@ -1,351 +0,0 @@
import { DeployByPm2Service } from './runners/deploy-by-pm2/deploy-by-pm2.service';
import { Test, TestingModule } from '@nestjs/testing';
import { ReposService } from '../repos/repos.service';
import { PipelineUnits } from './enums/pipeline-units.enum';
import { PipelineTask } from './pipeline-task.entity';
import { Pipeline } from '../pipelines/pipeline.entity';
import { Project } from '../projects/project.entity';
import { TaskStatuses } from './enums/task-statuses.enum';
import { getLoggerToken, PinoLogger } from 'nestjs-pino';
import { PipelineTaskRunner } from './pipeline-task.runner';
import { WorkUnitMetadata } from './models/work-unit-metadata.model';
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
describe('PipelineTaskRunner', () => {
let runner: PipelineTaskRunner;
let deployByPM2Service: DeployByPm2Service;
beforeEach(async () => {
const module: TestingModule = await Test.createTestingModule({
providers: [
{
provide: ReposService,
useValue: {
getWorkspaceRootByTask: () => 'workspace-root',
checkout: async () => undefined,
},
},
{
provide: getLoggerToken(PipelineTaskRunner.name),
useValue: new PinoLogger({}),
},
{
provide: 'spawn',
useValue: () => undefined,
},
PipelineTaskRunner,
{
provide: AmqpConnection,
useValue: {},
},
{
provide: DeployByPm2Service,
useValue: {
deploy: () => Promise.resolve(),
},
},
],
}).compile();
module.get(ReposService);
runner = module.get(PipelineTaskRunner);
deployByPM2Service = module.get(DeployByPm2Service);
});
it('should be defined', () => {
expect(runner).toBeDefined();
});
it('onNewTask', async () => {
const task = new PipelineTask();
let tmpTask;
const doTask = jest
.spyOn(runner, 'doTask')
.mockImplementation(async (task) => {
tmpTask = task;
});
await runner.onNewTask(task);
expect(tmpTask).toEqual(task);
expect(doTask).toBeCalledTimes(1);
});
describe('test biz', () => {
let emitEvent: jest.SpyInstance;
beforeEach(() => {
emitEvent = jest
.spyOn(runner, 'emitEvent')
.mockImplementation(() => Promise.resolve());
});
describe('doTask', () => {
let checkout: jest.SpyInstance;
let doTaskUnit: jest.SpyInstance;
beforeEach(() => {
checkout = jest
.spyOn(runner, 'checkout')
.mockImplementation(() => Promise.resolve('/null'));
doTaskUnit = jest
.spyOn(runner, 'doTaskUnit')
.mockImplementation(() => Promise.resolve());
});
it('only checkout', async () => {
const task = new PipelineTask();
(task.id = 'taskId'), (task.pipeline = new Pipeline());
task.units = [PipelineUnits.checkout];
task.pipeline.id = 'pipelineId';
task.pipeline.project = new Project();
task.pipeline.project.id = 'projectId';
task.pipeline.workUnitMetadata = new WorkUnitMetadata();
task.pipeline.workUnitMetadata.version = 1;
task.pipeline.workUnitMetadata.units = [
{
type: PipelineUnits.checkout,
scripts: [],
},
];
await runner.doTask(task);
expect(checkout).toBeCalledTimes(1);
expect(doTaskUnit).toBeCalledTimes(0);
expect(emitEvent).toBeCalledTimes(2);
expect(emitEvent.mock.calls[0][0]).toMatchObject(task);
expect(emitEvent.mock.calls[0][1]).toBeNull();
expect(emitEvent.mock.calls[0][2]).toEqual(TaskStatuses.working);
expect(emitEvent.mock.calls[1][0]).toMatchObject(task);
expect(emitEvent.mock.calls[1][1]).toBeNull();
expect(emitEvent.mock.calls[1][2]).toEqual(TaskStatuses.success);
});
it('many units', async () => {
const task = new PipelineTask();
(task.id = 'taskId'), (task.pipeline = new Pipeline());
task.units = [
PipelineUnits.checkout,
PipelineUnits.test,
PipelineUnits.deploy,
];
task.pipeline.id = 'pipelineId';
task.pipeline.project = new Project();
task.pipeline.project.id = 'projectId';
task.pipeline.workUnitMetadata = new WorkUnitMetadata();
task.pipeline.workUnitMetadata.version = 1;
task.pipeline.workUnitMetadata.units = [
{
type: PipelineUnits.checkout,
scripts: [],
},
{
type: PipelineUnits.installDependencies,
scripts: ['pwd'],
},
{
type: PipelineUnits.test,
scripts: ['pwd'],
},
{
type: PipelineUnits.deploy,
scripts: ['pwd', 'uname'],
},
];
await runner.doTask(task);
expect(checkout).toBeCalledTimes(1);
expect(doTaskUnit).toBeCalledTimes(1);
expect(emitEvent).toBeCalledTimes(2);
});
it('unit work failed', async () => {
const task = new PipelineTask();
(task.id = 'taskId'), (task.pipeline = new Pipeline());
task.units = [PipelineUnits.checkout, PipelineUnits.test];
task.pipeline.id = 'pipelineId';
task.pipeline.project = new Project();
task.pipeline.project.id = 'projectId';
task.pipeline.workUnitMetadata = new WorkUnitMetadata();
task.pipeline.workUnitMetadata.version = 1;
task.pipeline.workUnitMetadata.units = [
{
type: PipelineUnits.checkout,
scripts: [],
},
{
type: PipelineUnits.test,
scripts: ['pwd'],
},
];
doTaskUnit = jest
.spyOn(runner, 'doTaskUnit')
.mockImplementation(() => Promise.reject(new Error('test error')));
await runner.doTask(task);
expect(checkout).toBeCalledTimes(1);
expect(doTaskUnit).toBeCalledTimes(1);
expect(emitEvent).toBeCalledTimes(2);
expect(emitEvent.mock.calls[1][0]).toMatchObject(task);
expect(emitEvent.mock.calls[1][1]).toBeNull();
expect(emitEvent.mock.calls[1][2]).toEqual(TaskStatuses.failed);
});
});
describe('doTaskUnit', () => {
it('success', async () => {
const runScript = jest
.spyOn(runner, 'runScript')
.mockImplementation(() => Promise.resolve());
const task = new PipelineTask();
const unit = PipelineUnits.test;
const workspacePath = '/null';
await runner.doTaskUnit(unit, ['pwd'], task, workspacePath);
expect(emitEvent.mock.calls[0][0]).toEqual(task);
expect(emitEvent.mock.calls[0][1]).toEqual(unit);
expect(emitEvent.mock.calls[0][2]).toEqual(TaskStatuses.working);
expect(emitEvent.mock.calls[1][0]).toEqual(task);
expect(emitEvent.mock.calls[1][1]).toEqual(unit);
expect(emitEvent.mock.calls[1][2]).toEqual(TaskStatuses.success);
expect(runScript.mock.calls[0][0]).toEqual('pwd');
expect(runScript.mock.calls[0][1]).toEqual(workspacePath);
expect(runScript.mock.calls[0][2]).toEqual(task);
expect(runScript.mock.calls[0][3]).toEqual(unit);
});
it('failed', async () => {
const runScript = jest
.spyOn(runner, 'runScript')
.mockImplementation(() => Promise.reject(new Error('test error')));
const task = new PipelineTask();
const unit = PipelineUnits.test;
const workspacePath = '/null';
await expect(
runner.doTaskUnit(unit, ['pwd'], task, workspacePath),
).rejects.toThrow('test error');
expect(emitEvent.mock.calls[1]?.[0]).toEqual(task);
expect(emitEvent.mock.calls[1]?.[1]).toEqual(unit);
expect(emitEvent.mock.calls[1]?.[2]).toEqual(TaskStatuses.failed);
expect(runScript).toBeCalledTimes(1);
});
});
describe('runScript', () => {
it('normal', async () => {
const spawn = jest.fn<any, any>(() => ({
stdout: {
on: () => undefined,
},
stderr: {
on: () => undefined,
},
addListener: (_: any, fn: (code: number) => void) => {
fn(0);
},
}));
(runner as any).spawn = spawn;
const task = new PipelineTask();
task.id = 'taskId';
const unit = PipelineUnits.deploy;
await runner.runScript('script name', 'workspaceRoot', task, unit);
expect(spawn).toHaveBeenCalledTimes(1);
expect(spawn.mock.calls[0][0]).toEqual('script name');
expect(spawn.mock.calls[0][1]).toMatchObject({
shell: true,
cwd: 'workspaceRoot',
});
});
it('failed', async () => {
const spawn = jest.fn(() => ({
stdout: {
on: () => undefined,
},
stderr: {
on: () => undefined,
},
addListener: (_: any, fn: (code: number) => void) => {
fn(1);
},
}));
(runner as any).spawn = spawn;
const task = new PipelineTask();
task.id = 'taskId';
const unit = PipelineUnits.deploy;
expect(
runner.runScript('script name', 'workspaceRoot', task, unit),
).rejects.toThrowError();
});
it('wait emit message done', async () => {
let finishedFn: () => void;
const on = jest.fn((_: any, fn: (buff: Buffer) => void) => {
setTimeout(() => {
fn(Buffer.from('message 1'));
setTimeout(() => {
fn(Buffer.from('message 2'));
setTimeout(() => {
fn(Buffer.from('message 3'));
finishedFn();
}, 1000);
}, 10);
}, 10);
});
const spawn = jest.fn(() => ({
stdout: {
on,
},
stderr: {
on,
},
addListener: (_: any, fn: (code: number) => void) => {
finishedFn = () => fn(0);
},
}));
let emitSuccessCount = 0;
jest.spyOn(runner, 'emitEvent').mockImplementation(() => {
return new Promise((resolve) => {
setTimeout(() => {
emitSuccessCount++;
resolve();
}, 1000);
});
});
(runner as any).spawn = spawn;
const task = new PipelineTask();
task.id = 'taskId';
const unit = PipelineUnits.deploy;
await runner.runScript('script name', 'workspaceRoot', task, unit);
expect(emitSuccessCount).toEqual(1 + 6);
});
});
});
describe('tryRunDeployScript', () => {
it('should be call deploy with right args', async () => {
const deploy = jest.spyOn(deployByPM2Service, 'deploy');
await expect(
runner['tryRunDeployScript'](
'/test/dir',
'@@DEPLOY ecosystem.config.js',
),
).resolves.toBe(true);
expect(deploy.mock.calls[0][0]).toEqual('/test/dir/ecosystem.config.js');
});
it('should return false', async () => {
await expect(
runner['tryRunDeployScript'](
'/test/dir',
'pm2 start ecosystem.config.js',
),
).resolves.toBe(false);
});
});
});

View File

@ -1,339 +0,0 @@
import { DeployByPm2Service } from './runners/deploy-by-pm2/deploy-by-pm2.service';
import { ReposService } from '../repos/repos.service';
import { spawn, ChildProcessWithoutNullStreams } from 'child_process';
import { PipelineTask } from './pipeline-task.entity';
import { ApplicationException } from '../commons/exceptions/application.exception';
import { PipelineUnits } from './enums/pipeline-units.enum';
import { TaskStatuses } from './enums/task-statuses.enum';
import { InjectPinoLogger, PinoLogger } from 'nestjs-pino';
import {
AmqpConnection,
RabbitRPC,
RabbitSubscribe,
} from '@golevelup/nestjs-rabbitmq';
import { PipelineTaskEvent } from './models/pipeline-task-event';
import { last } from 'ramda';
import { Inject } from '@nestjs/common';
import {
EXCHANGE_PIPELINE_TASK_TOPIC,
QUEUE_PIPELINE_TASK_KILL,
ROUTE_PIPELINE_TASK_KILL,
} from './pipeline-tasks.constants';
import {
EXCHANGE_PIPELINE_TASK_FANOUT,
ROUTE_PIPELINE_TASK_LOG,
} from './pipeline-tasks.constants';
import {
getInstanceName,
getSelfInstanceQueueKey,
getSelfInstanceRouteKey,
} from '../commons/utils/rabbit-mq';
import { rm } from 'fs/promises';
import { rename } from 'fs/promises';
import { join } from 'path';
type Spawn = typeof spawn;
export class PipelineTaskRunner {
readonly processes = new Map<string, ChildProcessWithoutNullStreams>();
readonly stopTaskIds = new Set<string>();
constructor(
private readonly reposService: ReposService,
@InjectPinoLogger(PipelineTaskRunner.name)
private readonly logger: PinoLogger,
@Inject('spawn')
private readonly spawn: Spawn,
private readonly amqpConnection: AmqpConnection,
private readonly deployByPM2Service: DeployByPm2Service,
) {}
@RabbitSubscribe({
exchange: 'new-pipeline-task',
routingKey: 'mac',
queue: 'mac.new-pipeline-task',
})
async onNewTask(task: PipelineTask) {
this.logger.info({ task }, 'on new task [%s].', task.id);
try {
await this.doTask(task);
} catch (err) {
this.logger.error({ task, err }, err.message);
}
}
@RabbitRPC({
exchange: EXCHANGE_PIPELINE_TASK_TOPIC,
routingKey: getSelfInstanceRouteKey(ROUTE_PIPELINE_TASK_KILL),
queue: getSelfInstanceQueueKey(QUEUE_PIPELINE_TASK_KILL),
queueOptions: {
autoDelete: true,
durable: true,
},
})
async onStopTask(task: PipelineTask) {
this.logger.info({ task }, 'on stop task [%s].', task.id);
this.stopTaskIds.add(task.id);
const process = this.processes.get(task.id);
if (process) {
this.logger.info({ task }, 'send signal SIGINT to child process.');
process.kill('SIGINT');
setTimeout(() => {
setTimeout(() => {
this.stopTaskIds.delete(task.id);
}, 10_000);
if (process === this.processes.get(task.id)) {
this.logger.info({ task }, 'send signal SIGKILL to child process.');
process.kill('SIGKILL');
return;
}
if (this.processes.has(task.id)) {
this.logger.error(
{ task },
'this pipeline task not stop yet. there is a new process running, maybe is a bug about error capture',
);
}
}, 10_000);
} else {
this.logger.info({ task }, 'child process is not running.');
}
return true;
}
async doTask(task: PipelineTask) {
if (task.pipeline.workUnitMetadata.version !== 1) {
throw new ApplicationException(
'work unit metadata version is not match.',
);
}
await this.emitEvent(
task,
null,
TaskStatuses.working,
`[start task]`,
'stdout',
);
this.logger.info('running task [%s].', task.id);
try {
let workspaceRoot = await this.checkout(task);
const units = task.units
.filter((unit) => unit !== PipelineUnits.checkout)
.map(
(type) =>
task.pipeline.workUnitMetadata.units.find(
(unit) => unit.type === type,
) ?? { type: type, scripts: [] },
);
this.logger.info({ units }, 'begin run units.');
for (const unit of units) {
if (unit.type === PipelineUnits.deploy) {
const oldRoot = workspaceRoot;
workspaceRoot = this.reposService.getDeployRoot(task);
if (oldRoot !== workspaceRoot) {
await rm(workspaceRoot, { force: true, recursive: true });
await rename(oldRoot, workspaceRoot);
}
await this.emitEvent(
task,
unit.type,
TaskStatuses.success,
`[deploy] change deploy folder content success`,
'stdout',
);
}
await this.doTaskUnit(unit.type, unit.scripts, task, workspaceRoot);
}
await this.emitEvent(
task,
null,
TaskStatuses.success,
`[finished task] success`,
'stdout',
);
this.logger.info({ task }, 'task [%s] completed.', task.id);
} catch (err) {
await this.emitEvent(
task,
null,
TaskStatuses.failed,
`[finished unit] ${err.message}`,
'stderr',
);
this.logger.error({ task, error: err }, 'task [%s] failed.', task.id);
} finally {
}
}
async doTaskUnit(
unit: PipelineUnits,
scripts: string[],
task: PipelineTask,
workspaceRoot: string,
) {
await this.emitEvent(
task,
unit,
TaskStatuses.working,
`[begin unit] ${unit}`,
'stdin',
);
this.logger.info({ task }, 'curr unit is %s', unit);
try {
for (const script of scripts) {
this.logger.debug('begin runScript %s', script);
if (this.stopTaskIds.has(task.id)) {
throw new ApplicationException('Task is be KILLED');
}
await this.runScript(script, workspaceRoot, task, unit);
this.logger.debug('end runScript %s', script);
}
await this.emitEvent(
task,
unit,
TaskStatuses.success,
`[finished unit] ${unit}`,
'stdout',
);
} catch (err) {
await this.emitEvent(
task,
unit,
TaskStatuses.failed,
`[finished unit] ${err.message}`,
'stderr',
);
throw err;
}
}
async checkout(task: PipelineTask) {
await this.emitEvent(
task,
PipelineUnits.checkout,
TaskStatuses.working,
'[begin unit] checkout',
'stdin',
);
try {
const path = await this.reposService.checkout4Task(task);
await this.emitEvent(
task,
PipelineUnits.checkout,
TaskStatuses.success,
'checkout success.',
'stdout',
);
return path;
} catch (err) {
await this.emitEvent(
task,
PipelineUnits.checkout,
TaskStatuses.failed,
'checkout failed.',
'stderr',
);
throw err;
}
}
async emitEvent(
task: PipelineTask,
unit: PipelineUnits | null,
status: TaskStatuses,
message: string,
messageType: 'stderr' | 'stdout' | 'stdin',
) {
const event: PipelineTaskEvent = {
taskId: task.id,
pipelineId: task.pipeline.id,
projectId: task.pipeline.project.id,
unit,
emittedAt: new Date(),
message: last(message) === '\n' ? message : message + '\n',
messageType,
status,
};
this.amqpConnection
.publish(EXCHANGE_PIPELINE_TASK_FANOUT, ROUTE_PIPELINE_TASK_LOG, event, {
headers: {
sender: getInstanceName(),
},
})
.catch((error) => {
this.logger.error(
{ error, event },
'send event message to queue failed. %s',
error.message,
);
});
}
async runScript(
script: string,
workspaceRoot: string,
task: PipelineTask,
unit: PipelineUnits,
): Promise<void> {
await this.emitEvent(task, unit, TaskStatuses.working, script, 'stdin');
if (await this.tryRunDeployScript(workspaceRoot, script)) {
return;
}
return new Promise((resolve, reject) => {
const sub = this.spawn(script, {
shell: true,
cwd: workspaceRoot,
});
this.processes.set(task.id, sub);
let loggingCount = 0; // semaphore
sub.stderr.on('data', (data: Buffer) => {
const str = data.toString();
loggingCount++;
this.emitEvent(task, unit, TaskStatuses.working, str, 'stdout').finally(
() => loggingCount--,
);
});
sub.stdout.on('data', (data: Buffer) => {
const str = data.toString();
loggingCount++;
this.emitEvent(task, unit, TaskStatuses.working, str, 'stderr').finally(
() => loggingCount--,
);
});
sub.addListener('close', async (code) => {
this.processes.delete(task.id);
await new Promise<void>(async (resolve) => {
for (let i = 0; i < 10 && loggingCount > 0; i++) {
await new Promise((resolve) => setTimeout(resolve, 500));
this.logger.debug('waiting logging... (%dx500ms)', i);
}
resolve();
});
if (code === 0) {
return resolve();
}
if (this.stopTaskIds.has(task.id)) {
throw reject(new ApplicationException('Task is be KILLED'));
}
return reject(new ApplicationException('exec script failed'));
});
});
}
private async tryRunDeployScript(workspaceRoot: string, script: string) {
const match = /^@@DEPLOY\s+(\S*)/.exec(script);
if (match) {
await this.deployByPM2Service.deploy(
join(workspaceRoot, match[1]),
workspaceRoot,
);
return true;
} else {
return false;
}
}
}

View File

@ -1,11 +1,3 @@
export const EXCHANGE_PIPELINE_TASK_TOPIC = 'pipeline-task.topic'; export const PIPELINE_TASK_QUEUE = 'PIPELINE_TASK_QUEUE';
export const EXCHANGE_PIPELINE_TASK_FANOUT = 'pipeline-task.fanout'; export const PIPELINE_TASK_LOG_QUEUE = 'PIPELINE_TASK_LOG_QUEUE';
export const ROUTE_PIPELINE_TASK_LOG = 'pipeline-task-log'; export const PIPELINE_TASK_LOG_PUBSUB = 'PIPELINE_TASK_LOG_PUBSUB';
export const QUEUE_HANDLE_PIPELINE_TASK_LOG_EVENT = 'pipeline-task-log';
export const QUEUE_WRITE_PIPELINE_TASK_LOG = 'write-pipeline-task-log';
export const ROUTE_PIPELINE_TASK_DONE = 'pipeline-task-done';
export const QUEUE_PIPELINE_TASK_DONE = 'pipeline-task-done';
export const ROUTE_PIPELINE_TASK_KILL = 'pipeline-task-kill';
export const QUEUE_PIPELINE_TASK_KILL = 'pipeline-task-kill';
export const PM2_TOKEN = Symbol('pm2-token');

View File

@ -6,70 +6,33 @@ import { PipelineTask } from './pipeline-task.entity';
import { Pipeline } from '../pipelines/pipeline.entity'; import { Pipeline } from '../pipelines/pipeline.entity';
import { ReposModule } from '../repos/repos.module'; import { ReposModule } from '../repos/repos.module';
import { RedisModule } from 'nestjs-redis'; import { RedisModule } from 'nestjs-redis';
import { RabbitMQModule } from '@golevelup/nestjs-rabbitmq'; import { BullModule } from '@nestjs/bull';
import { ConfigModule, ConfigService } from '@nestjs/config'; import { PipelineTaskConsumer } from './pipeline-task.consumer';
import { PipelineTaskRunner } from './pipeline-task.runner';
import { spawn } from 'child_process';
import { import {
EXCHANGE_PIPELINE_TASK_FANOUT, PIPELINE_TASK_QUEUE,
EXCHANGE_PIPELINE_TASK_TOPIC, PIPELINE_TASK_LOG_PUBSUB,
} from './pipeline-tasks.constants'; } from './pipeline-tasks.constants';
import { PipelineTaskLogger } from './pipeline-task.logger'; import { PipelineTaskLogsService } from './pipeline-task-logs.service';
import { PipelineTaskFlushService } from './pipeline-task-flush.service'; import { PubSub } from 'apollo-server-express';
import { CommonsModule } from '../commons/commons.module';
import { DeployByPm2Service } from './runners/deploy-by-pm2/deploy-by-pm2.service';
@Module({ @Module({
imports: [ imports: [
CommonsModule,
TypeOrmModule.forFeature([PipelineTask, Pipeline]), TypeOrmModule.forFeature([PipelineTask, Pipeline]),
BullModule.registerQueue({
name: PIPELINE_TASK_QUEUE,
}),
RedisModule, RedisModule,
ReposModule, ReposModule,
RabbitMQModule.forRootAsync(RabbitMQModule, {
imports: [ConfigModule],
useFactory: (configService: ConfigService) => ({
uri: configService.get<string>('db.rabbitmq.uri'),
exchanges: [
{
name: 'new-pipeline-task',
type: 'fanout',
options: {
durable: true,
autoDelete: true,
},
},
{
name: EXCHANGE_PIPELINE_TASK_FANOUT,
type: 'fanout',
options: {
durable: false,
autoDelete: true,
},
},
{
name: EXCHANGE_PIPELINE_TASK_TOPIC,
type: 'topic',
options: {
durable: false,
autoDelete: true,
},
},
],
}),
inject: [ConfigService],
}),
], ],
providers: [ providers: [
PipelineTasksService, PipelineTasksService,
PipelineTasksResolver, PipelineTasksResolver,
PipelineTaskRunner, PipelineTaskConsumer,
PipelineTaskLogger, PipelineTaskLogsService,
{ {
provide: 'spawn', provide: Symbol(PIPELINE_TASK_LOG_PUBSUB),
useValue: spawn, useValue: new PubSub(),
}, },
PipelineTaskFlushService,
DeployByPm2Service,
], ],
exports: [PipelineTasksService], exports: [PipelineTasksService],
}) })

View File

@ -1,7 +1,6 @@
import { JwtService } from '@nestjs-lib/auth';
import { Test, TestingModule } from '@nestjs/testing'; import { Test, TestingModule } from '@nestjs/testing';
import { PipelineTaskLogger } from './pipeline-task.logger';
import { PipelineTasksResolver } from './pipeline-tasks.resolver'; import { PipelineTasksResolver } from './pipeline-tasks.resolver';
import { PipelineTaskLogsService } from './pipeline-task-logs.service';
import { PipelineTasksService } from './pipeline-tasks.service'; import { PipelineTasksService } from './pipeline-tasks.service';
describe('PipelineTasksResolver', () => { describe('PipelineTasksResolver', () => {
@ -16,11 +15,7 @@ describe('PipelineTasksResolver', () => {
useValue: {}, useValue: {},
}, },
{ {
provide: PipelineTaskLogger, provide: PipelineTaskLogsService,
useValue: {},
},
{
provide: JwtService,
useValue: {}, useValue: {},
}, },
], ],

View File

@ -2,19 +2,15 @@ import { Resolver, Args, Mutation, Subscription, Query } from '@nestjs/graphql';
import { PipelineTask } from './pipeline-task.entity'; import { PipelineTask } from './pipeline-task.entity';
import { PipelineTasksService } from './pipeline-tasks.service'; import { PipelineTasksService } from './pipeline-tasks.service';
import { CreatePipelineTaskInput } from './dtos/create-pipeline-task.input'; import { CreatePipelineTaskInput } from './dtos/create-pipeline-task.input';
import { PipelineTaskLogMessage } from './models/pipeline-task-log-message.module';
import { PipelineTaskLogArgs } from './dtos/pipeline-task-log.args'; import { PipelineTaskLogArgs } from './dtos/pipeline-task-log.args';
import { plainToClass } from 'class-transformer'; import { PipelineTaskLogsService } from './pipeline-task-logs.service';
import { PipelineTaskLogger } from './pipeline-task.logger';
import { observableToAsyncIterable } from '@graphql-tools/utils';
import { PipelineTaskEvent } from './models/pipeline-task-event';
import { Roles, AccountRole } from '@nestjs-lib/auth';
@Roles(AccountRole.admin, AccountRole.super)
@Resolver() @Resolver()
export class PipelineTasksResolver { export class PipelineTasksResolver {
constructor( constructor(
private readonly service: PipelineTasksService, private readonly service: PipelineTasksService,
private readonly taskLogger: PipelineTaskLogger, private readonly logsService: PipelineTaskLogsService,
) {} ) {}
@Mutation(() => PipelineTask) @Mutation(() => PipelineTask)
@ -22,17 +18,15 @@ export class PipelineTasksResolver {
return await this.service.addTask(taskDto); return await this.service.addTask(taskDto);
} }
@Subscription(() => PipelineTaskEvent, { @Subscription(() => PipelineTaskLogMessage, {
resolve: (value) => { resolve: (value) => {
const data = plainToClass(PipelineTaskEvent, value); return value;
return data;
}, },
}) })
async pipelineTaskEvent(@Args() args: PipelineTaskLogArgs) { async pipelineTaskLog(@Args() args: PipelineTaskLogArgs) {
const task = await this.service.findTaskById(args.taskId); const task = await this.service.findTaskById(args.taskId);
return observableToAsyncIterable<PipelineTaskEvent>( const asyncIterator = this.logsService.watchLogs(task);
this.taskLogger.getMessage$(task.id), return asyncIterator;
);
} }
@Subscription(() => PipelineTask, { @Subscription(() => PipelineTask, {
@ -41,7 +35,7 @@ export class PipelineTasksResolver {
}, },
}) })
async pipelineTaskChanged(@Args('id') id: string) { async pipelineTaskChanged(@Args('id') id: string) {
// return await this.service.watchTaskUpdated(id); return await this.service.watchTaskUpdated(id);
} }
@Query(() => [PipelineTask]) @Query(() => [PipelineTask])
@ -50,14 +44,7 @@ export class PipelineTasksResolver {
} }
@Query(() => PipelineTask) @Query(() => PipelineTask)
async pipelineTask(@Args('id') id: string) { async findPipelineTask(@Args('id') id: string) {
return await this.service.findTaskById(id); return await this.service.findTaskById(id);
} }
@Mutation(() => Boolean)
async stopPipelineTask(@Args('id') id: string) {
const task = await this.service.findTaskById(id);
await this.service.stopTask(task);
return true;
}
} }

View File

@ -2,19 +2,51 @@ import { Test, TestingModule } from '@nestjs/testing';
import { PipelineTasksService } from './pipeline-tasks.service'; import { PipelineTasksService } from './pipeline-tasks.service';
import { getRepositoryToken } from '@nestjs/typeorm'; import { getRepositoryToken } from '@nestjs/typeorm';
import { PipelineTask } from './pipeline-task.entity'; import { PipelineTask } from './pipeline-task.entity';
import { PIPELINE_TASK_QUEUE } from './pipeline-tasks.constants';
import { getQueueToken } from '@nestjs/bull';
import { RedisService } from 'nestjs-redis';
import { Pipeline } from '../pipelines/pipeline.entity'; import { Pipeline } from '../pipelines/pipeline.entity';
import { EntityNotFoundError } from 'typeorm/error/EntityNotFoundError';
import { Repository } from 'typeorm'; import { Repository } from 'typeorm';
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq'; import { Queue } from 'bull';
import { PipelineTaskFlushService } from './pipeline-task-flush.service'; import { LockFailedException } from '../commons/exceptions/lock-failed.exception';
import { getLoggerToken, PinoLogger } from 'nestjs-pino';
describe('PipelineTasksService', () => { describe('PipelineTasksService', () => {
let service: PipelineTasksService; let service: PipelineTasksService;
let module: TestingModule; let module: TestingModule;
let taskRepository: Repository<PipelineTask>; let taskRepository: Repository<PipelineTask>;
let pipelineRepository: Repository<Pipeline>; let pipelineRepository: Repository<Pipeline>;
const getBasePipeline = () =>
({
id: 'test',
name: '测试流水线',
branch: 'master',
workUnitMetadata: {},
project: {
id: 'test-project',
},
} as Pipeline);
let redisClient;
let taskQueue: Queue;
const getTask = () =>
({
pipelineId: 'test',
commit: 'test',
pipeline: { branch: 'master' },
units: [],
} as PipelineTask);
beforeEach(async () => { beforeEach(async () => {
redisClient = (() => ({
set: jest.fn().mockImplementation(async () => 'OK'),
del: jest.fn().mockImplementation(async () => 'test'),
get: jest.fn().mockImplementation(async () => 'test'),
lpush: jest.fn().mockImplementation(async () => 1),
rpop: jest.fn().mockImplementation(async () => JSON.stringify(getTask())),
}))() as any;
taskQueue = (() => ({
add: jest.fn().mockImplementation(async () => null),
}))() as any;
module = await Test.createTestingModule({ module = await Test.createTestingModule({
providers: [ providers: [
PipelineTasksService, PipelineTasksService,
@ -27,16 +59,14 @@ describe('PipelineTasksService', () => {
useValue: new Repository(), useValue: new Repository(),
}, },
{ {
provide: AmqpConnection, provide: getQueueToken(PIPELINE_TASK_QUEUE),
useValue: {}, useValue: taskQueue,
}, },
{ {
provide: PipelineTaskFlushService, provide: RedisService,
useValue: {}, useValue: {
getClient: jest.fn(() => redisClient),
}, },
{
provide: getLoggerToken(PipelineTasksService.name),
useValue: new PinoLogger({}),
}, },
], ],
}).compile(); }).compile();
@ -57,43 +87,119 @@ describe('PipelineTasksService', () => {
expect(service).toBeDefined(); expect(service).toBeDefined();
}); });
// describe('addTask', () => { describe('addTask', () => {
// beforeEach(() => { beforeEach(() => {
// jest jest
// .spyOn(pipelineRepository, 'findOneOrFail') .spyOn(pipelineRepository, 'findOneOrFail')
// .mockImplementation(async () => getBasePipeline()); .mockImplementation(async () => getBasePipeline());
// }); });
// it('pipeline not found', async () => { it('pipeline not found', async () => {
// jest.spyOn(taskRepository, 'findOneOrFail').mockImplementation(() => { jest.spyOn(taskRepository, 'findOneOrFail').mockImplementation(() => {
// throw new EntityNotFoundError(Pipeline, {}); throw new EntityNotFoundError(Pipeline, {});
// }); });
// await expect( await expect(
// service.addTask({ pipelineId: 'test', commit: 'test', units: [] }), service.addTask({ pipelineId: 'test', commit: 'test', units: [] }),
// ).rejects; ).rejects;
// }); });
// it('create task on db', async () => { it('create task on db', async () => {
// const save = jest const save = jest
// .spyOn(taskRepository, 'save') .spyOn(taskRepository, 'save')
// .mockImplementation(async (data: any) => data); .mockImplementation(async (data: any) => data);
// const findOne = jest.spyOn(taskRepository, 'findOne'); const findOne = jest.spyOn(taskRepository, 'findOne');
// await service.addTask({ pipelineId: 'test', commit: 'test', units: [] }), jest
// expect(save.mock.calls[0][0]).toMatchObject({ .spyOn(service, 'doNextTask')
// pipelineId: 'test', .mockImplementation(async () => undefined);
// commit: 'test', await service.addTask({ pipelineId: 'test', commit: 'test', units: [] }),
// units: [], expect(save.mock.calls[0][0]).toMatchObject({
// }); pipelineId: 'test',
// expect(findOne).toBeCalled(); commit: 'test',
// }); units: [],
// it('add task', async () => { });
// const lpush = jest.spyOn(redisClient, 'lpush'); expect(findOne).toBeCalled();
// await service.addTask({ pipelineId: 'test', commit: 'test', units: [] }); });
// expect(typeof lpush.mock.calls[0][1] === 'string').toBeTruthy(); it('add task', async () => {
// expect(JSON.parse(lpush.mock.calls[0][1] as string)).toMatchObject({ const lpush = jest.spyOn(redisClient, 'lpush');
// pipelineId: 'test', const doNextTask = jest.spyOn(service, 'doNextTask');
// commit: 'test', jest
// units: [], .spyOn(service, 'doNextTask')
// pipeline: getBasePipeline(), .mockImplementation(async () => undefined);
// }); await service.addTask({ pipelineId: 'test', commit: 'test', units: [] });
// }); expect(typeof lpush.mock.calls[0][1] === 'string').toBeTruthy();
// }); expect(JSON.parse(lpush.mock.calls[0][1] as string)).toMatchObject({
pipelineId: 'test',
commit: 'test',
units: [],
pipeline: getBasePipeline(),
});
expect(doNextTask).toHaveBeenCalledWith(getBasePipeline());
});
});
describe('doNextTask', () => {
it('add task to queue', async () => {
let lckValue: string;
const set = jest
.spyOn(redisClient, 'set')
.mockImplementation(async (...args) => (lckValue = args[3] as string));
const get = jest
.spyOn(redisClient, 'get')
.mockImplementation(async () => lckValue);
const del = jest.spyOn(redisClient, 'del');
const rpop = jest.spyOn(redisClient, 'rpop');
const add = jest.spyOn(taskQueue, 'add');
await service.doNextTask(getBasePipeline());
expect(add).toHaveBeenCalledWith(getTask());
expect(set).toHaveBeenCalledTimes(1);
expect(rpop).toHaveBeenCalledTimes(1);
expect(get).toHaveBeenCalledTimes(1);
expect(del).toHaveBeenCalledTimes(1);
});
it('pipeline is busy', async () => {
let remainTimes = 3;
let lckValue: string;
const set = jest
.spyOn(redisClient, 'set')
.mockImplementation(async (...args) => {
if (remainTimes-- > 0) {
throw new Error();
} else {
lckValue = args[3] as string;
}
});
const get = jest
.spyOn(redisClient, 'get')
.mockImplementation(async () => lckValue);
const del = jest.spyOn(redisClient, 'del');
const rpop = jest.spyOn(redisClient, 'rpop');
const add = jest.spyOn(taskQueue, 'add');
await service.doNextTask(getBasePipeline());
expect(rpop).toHaveBeenCalledTimes(1);
expect(set).toHaveBeenCalledTimes(4);
expect(get).toHaveBeenCalledTimes(1);
expect(del).toHaveBeenCalledTimes(1);
expect(add).toHaveBeenCalledWith(getTask());
}, 10_000);
it('pipeline always busy and timeout', async () => {
const set = jest
.spyOn(redisClient, 'set')
.mockImplementation(async () => {
throw new Error();
});
const get = jest.spyOn(redisClient, 'get');
const del = jest.spyOn(redisClient, 'del');
await expect(
service.doNextTask(getBasePipeline()),
).rejects.toBeInstanceOf(LockFailedException);
expect(set).toHaveBeenCalledTimes(5);
expect(get).toHaveBeenCalledTimes(0);
expect(del).toHaveBeenCalledTimes(0);
}, 15_000);
});
}); });

View File

@ -1,59 +1,63 @@
import { BadRequestException, Injectable } from '@nestjs/common'; import { ConflictException, Injectable } from '@nestjs/common';
import { InjectRepository } from '@nestjs/typeorm'; import { InjectRepository } from '@nestjs/typeorm';
import { PipelineTask } from './pipeline-task.entity'; import { PipelineTask } from './pipeline-task.entity';
import { Repository } from 'typeorm'; import { In, Repository } from 'typeorm';
import { CreatePipelineTaskInput } from './dtos/create-pipeline-task.input'; import { CreatePipelineTaskInput } from './dtos/create-pipeline-task.input';
import { RedisService } from 'nestjs-redis';
import { Pipeline } from '../pipelines/pipeline.entity'; import { Pipeline } from '../pipelines/pipeline.entity';
import { InjectQueue } from '@nestjs/bull';
import { PIPELINE_TASK_QUEUE } from './pipeline-tasks.constants';
import { Queue } from 'bull';
import { LockFailedException } from '../commons/exceptions/lock-failed.exception';
import { PubSub } from 'apollo-server-express';
import { TaskStatuses } from './enums/task-statuses.enum';
import { isNil } from 'ramda';
import debug from 'debug'; import debug from 'debug';
import { AmqpConnection, RabbitRPC } from '@golevelup/nestjs-rabbitmq';
import {
EXCHANGE_PIPELINE_TASK_TOPIC,
QUEUE_PIPELINE_TASK_DONE,
ROUTE_PIPELINE_TASK_DONE,
} from './pipeline-tasks.constants';
import { PipelineTaskFlushService } from './pipeline-task-flush.service';
import { find, isNil, propEq } from 'ramda';
import { PipelineTaskLogs } from './models/pipeline-task-logs.model';
import { TaskStatuses, terminalTaskStatuses } from './enums/task-statuses.enum';
import { InjectPinoLogger, PinoLogger } from 'nestjs-pino';
import { getAppInstanceRouteKey } from '../commons/utils/rabbit-mq';
import { ROUTE_PIPELINE_TASK_KILL } from './pipeline-tasks.constants';
const log = debug('fennec:pipeline-tasks:service'); const log = debug('fennec:pipeline-tasks:service');
@Injectable() @Injectable()
export class PipelineTasksService { export class PipelineTasksService {
pubSub = new PubSub();
constructor( constructor(
@InjectRepository(PipelineTask) @InjectRepository(PipelineTask)
private readonly repository: Repository<PipelineTask>, private readonly repository: Repository<PipelineTask>,
@InjectRepository(Pipeline) @InjectRepository(Pipeline)
private readonly pipelineRepository: Repository<Pipeline>, private readonly pipelineRepository: Repository<Pipeline>,
private readonly amqpConnection: AmqpConnection, @InjectQueue(PIPELINE_TASK_QUEUE)
private readonly eventFlushService: PipelineTaskFlushService, private readonly queue: Queue<PipelineTask>,
@InjectPinoLogger(PipelineTasksService.name) private readonly redis: RedisService,
private readonly logger: PinoLogger,
) {} ) {}
async addTask(dto: CreatePipelineTaskInput) { async addTask(dto: CreatePipelineTaskInput) {
const pipeline = await this.pipelineRepository.findOneOrFail({ const pipeline = await this.pipelineRepository.findOneOrFail({
where: { id: dto.pipelineId }, where: { id: dto.pipelineId },
relations: ['project'], relations: ['project'],
}); });
// const hasUnfinishedTask = await this.repository const hasUnfinishedTask = await this.repository
// .findOne({ .findOne({
// pipelineId: dto.pipelineId, pipelineId: dto.pipelineId,
// commit: dto.commit, commit: dto.commit,
// status: In([TaskStatuses.pending, TaskStatuses.working]), status: In([TaskStatuses.pending, TaskStatuses.working]),
// }) })
// .then((val) => !isNil(val)); .then((val) => !isNil(val));
// if (hasUnfinishedTask) { if (hasUnfinishedTask) {
// throw new ConflictException( throw new ConflictException(
// 'There are the same tasks among the unfinished tasks!', 'There are the same tasks among the unfinished tasks!',
// ); );
// } }
const task = await this.repository.save(this.repository.create(dto)); const task = await this.repository.save(this.repository.create(dto));
task.pipeline = pipeline; task.pipeline = pipeline;
this.amqpConnection.publish('new-pipeline-task', 'mac', task); const tasksKey = this.getRedisTokens(pipeline)[1];
const redis = this.redis.getClient();
await redis.lpush(tasksKey, JSON.stringify(task));
log(
'add task %s:%s-%s',
task.id,
task.pipeline.branch,
task.commit.slice(0, 6),
);
await this.doNextTask(pipeline);
return task; return task;
} }
@ -65,94 +69,60 @@ export class PipelineTasksService {
return await this.repository.find({ pipelineId }); return await this.repository.find({ pipelineId });
} }
async listTasksByCommitHash(hash: string) { async doNextTask(pipeline: Pipeline) {
return await this.repository.find({ const [lckKey, tasksKey] = this.getRedisTokens(pipeline);
where: { commit: hash }, const redis = this.redis.getClient();
order: { createdAt: 'DESC' },
log('doNextTask()');
const unLck = await new Promise<() => Promise<void>>(
async (resolve, reject) => {
const lckValue = Date.now().toString();
for (let i = 0; i < 5; i++) {
if (
await redis
.set(lckKey, 0, 'EX', lckValue, 'NX')
.then(() => true)
.catch(() => false)
) {
resolve(async () => {
if ((await redis.get(lckKey)) === lckValue) {
await redis.del(lckKey);
}
}); });
return;
}
await new Promise((resolve) => setTimeout(resolve, 2000));
}
reject(new LockFailedException(lckKey));
},
);
const task = JSON.parse(
(await redis.rpop(tasksKey).finally(() => unLck())) ?? 'null',
);
if (task) {
log(
'add task (%s:%s-%s) to queue',
task.id,
task.pipeline.branch,
task.commit.slice(0, 6),
);
await this.queue.add(task);
} else {
log('task is empty');
}
}
async updateTask(task: PipelineTask) {
this.pubSub.publish(task.id, task);
return await this.repository.save(task);
}
async watchTaskUpdated(id: string) {
return this.pubSub.asyncIterator(id);
} }
getRedisTokens(pipeline: Pipeline): [string, string] { getRedisTokens(pipeline: Pipeline): [string, string] {
return [`pipeline-${pipeline.id}:lck`, `pipeline-${pipeline.id}:tasks`]; return [`pipeline-${pipeline.id}:lck`, `pipeline-${pipeline.id}:tasks`];
} }
@RabbitRPC({
exchange: EXCHANGE_PIPELINE_TASK_TOPIC,
routingKey: ROUTE_PIPELINE_TASK_DONE,
queue: QUEUE_PIPELINE_TASK_DONE,
queueOptions: {
autoDelete: true,
durable: true,
},
})
async updateByEvent({ taskId, runOn }: { taskId: string; runOn: string }) {
try {
const [events, task] = await Promise.all([
this.eventFlushService.read(taskId),
this.findTaskById(taskId),
]);
this.logger.info('[updateByEvent] start. taskId: %s', taskId);
for (const event of events) {
if (isNil(event.unit)) {
if (
event.status !== TaskStatuses.pending &&
task.status === TaskStatuses.pending
) {
task.startedAt = event.emittedAt;
} else if (terminalTaskStatuses.includes(event.status)) {
task.endedAt = event.emittedAt;
}
task.status = event.status;
} else {
let l: PipelineTaskLogs = find<PipelineTaskLogs>(
propEq('unit', event.unit),
task.logs,
);
if (isNil(l)) {
l = {
unit: event.unit,
startedAt: event.emittedAt,
endedAt: null,
logs: event.message,
status: event.status,
};
task.logs.push(l);
} else {
l.logs += event.message;
}
if (terminalTaskStatuses.includes(event.status)) {
l.endedAt = event.emittedAt;
}
l.status = event.status;
}
}
task.runOn = runOn;
await this.repository.update({ id: taskId }, task);
this.logger.info('[updateByEvent] success. taskId: %s', taskId);
return task;
} catch (error) {
this.logger.error(
{ error },
'[updateByEvent] failed. taskId: %s',
taskId,
);
}
}
async stopTask(task: PipelineTask) {
if (isNil(task.runOn)) {
throw new BadRequestException(
"the task have not running instance on database. field 'runOn' is nil",
);
}
await this.amqpConnection.request({
exchange: EXCHANGE_PIPELINE_TASK_TOPIC,
routingKey: getAppInstanceRouteKey(ROUTE_PIPELINE_TASK_KILL, task.runOn),
payload: task,
});
}
} }

View File

@ -1,117 +0,0 @@
import { Test, TestingModule } from '@nestjs/testing';
import { getLoggerToken, PinoLogger } from 'nestjs-pino';
import { join } from 'path';
import { DeployByPm2Service } from './deploy-by-pm2.service';
describe('DeployByPm2Service', () => {
let service: DeployByPm2Service;
beforeEach(async () => {
const module: TestingModule = await Test.createTestingModule({
providers: [
DeployByPm2Service,
{
provide: getLoggerToken(DeployByPm2Service.name),
useValue: new PinoLogger({
pinoHttp: {
level: 'silent',
},
}),
},
],
}).compile();
service = module.get<DeployByPm2Service>(DeployByPm2Service);
});
it('should be defined', () => {
expect(service).toBeDefined();
});
describe('getAppsSn', () => {
it('should return right value', () => {
expect(
service['getAppsSn']([
{ name: 'app' },
{ name: 'app#4' },
{ name: 'app#1' },
]),
).toEqual(4 + 1);
});
it('should return 1 when no match', () => {
expect(
service['getAppsSn']([
{ name: 'bar' },
{ name: 'foo#4' },
{ name: 'foo#1' },
]),
).toEqual(4 + 1);
});
});
describe('filterOldApps', () => {
it('should return right value', () => {
expect(
service['filterOldApps'](
[{ name: 'app' }],
[
{ name: 'app' },
{ name: 'app#4' },
{ name: 'foo#2' },
{ name: 'bar' },
],
),
).toEqual([{ name: 'app' }, { name: 'app#4' }]);
});
it('should return [] when no match', () => {
expect(
service['filterOldApps'](
[{ name: 'app' }],
[{ name: 'foo#2' }, { name: 'bar' }],
),
).toEqual([]);
});
});
describe('replaceAppName', () => {
it('should be replaced with right value', () => {
const getAppsSn = jest
.spyOn(service, 'getAppsSn' as any)
.mockImplementation(() => 1);
const options = [{ name: 'app' }, { name: 'foo' }];
service['replaceAppName'](options, []);
expect(options).toEqual([{ name: 'app#1' }, { name: 'foo#1' }]);
expect(getAppsSn).toBeCalledTimes(1);
expect(getAppsSn.mock.calls[0][0]).toEqual([]);
});
});
describe('deploy', () => {
it('should be success', async () => {
const filterOldApps = jest
.spyOn(service, 'filterOldApps' as any)
.mockImplementation(() => [{ name: 'app#1' }, { name: 'app#2' }]);
const replaceAppName = jest
.spyOn(service, 'replaceAppName' as any)
.mockImplementation((options) => (options[0].name = 'app#2'));
const stopApps = jest
.spyOn(service, 'stopApps' as any)
.mockImplementation(() => Promise.resolve());
await expect(
service['deploy'](
join(
__dirname,
'../../../../test/__mocks__/deploy-service/ecosystem.config.js',
),
join(__dirname, '../../../../test/__mocks__/deploy-service'),
),
).resolves.toBeFalsy();
expect(filterOldApps).toBeCalledTimes(1);
expect(replaceAppName).toBeCalledTimes(1);
expect(stopApps).toBeCalledTimes(1);
stopApps.mockReset();
await service['stopApps']([{ name: 'app#2' }]);
}, 10_000);
});
});

View File

@ -1,98 +0,0 @@
import { InjectPinoLogger, PinoLogger } from 'nestjs-pino';
import { Injectable } from '@nestjs/common';
import { promisify } from 'util';
import * as pm2 from 'pm2';
import { Proc, ProcessDescription, StartOptions } from 'pm2';
import { clone, last } from 'ramda';
@Injectable()
export class DeployByPm2Service {
constructor(
@InjectPinoLogger(DeployByPm2Service.name)
private readonly logger: PinoLogger,
) {}
async deploy(filePath: string, workspace: string) {
const baseConfig: { apps: StartOptions[] } = await import(filePath);
const appOptionsList: StartOptions[] = clone(baseConfig.apps);
await promisify<void>(pm2.connect.bind(pm2))();
const allApps = await promisify(pm2.list.bind(pm2))();
try {
if (!Array.isArray(baseConfig.apps)) {
this.logger.error(
'the "apps" in the PM2 ecosystem configuration is not array',
);
throw new Error('apps is not array');
}
const oldApps = this.filterOldApps(appOptionsList, allApps);
this.replaceAppName(appOptionsList, oldApps);
for (const appOptions of appOptionsList) {
const proc = await promisify<StartOptions, Proc>(pm2.start.bind(pm2))({
...appOptions,
cwd: workspace,
});
this.logger.info({ proc }, `start ${appOptions.name}`);
}
await this.stopApps(oldApps);
} catch (err) {
await this.stopApps(appOptionsList);
throw err;
} finally {
pm2.disconnect();
}
}
private async stopApps(apps: ProcessDescription[] | StartOptions[]) {
await Promise.all(
apps.map(async (app: ProcessDescription | StartOptions) => {
let procAtStop: ProcessDescription;
let procAtDelete: ProcessDescription;
try {
const idOrName = 'pm_id' in app ? app.pm_id : app.name;
procAtStop = await promisify(pm2.stop.bind(pm2))(idOrName);
procAtDelete = await promisify(pm2.delete.bind(pm2))(idOrName);
this.logger.info('stop & delete %s success', app.name);
} catch (error) {
this.logger.error(
{ error, procAtStop, procAtDelete },
'stop & delete %s error',
app.name,
);
}
}),
);
}
private replaceAppName(
optionsList: StartOptions[],
oldApps: ProcessDescription[],
) {
const appSn = this.getAppsSn(oldApps);
optionsList.forEach((options) => {
if (!options.name) {
this.logger.error('please give a name for application');
throw new Error('app name is not given');
}
options.name = `${options.name}#${appSn}`;
});
}
private filterOldApps(
optionsList: StartOptions[],
apps: ProcessDescription[],
) {
return apps.filter((app) =>
optionsList.some((options) => app.name.split('#')[0] === options.name),
);
}
private getAppsSn(oldApps: ProcessDescription[]) {
const appsSn: number[] = oldApps.map(
(app) => +(app.name.split('#')?.[1] ?? 0),
);
return (last(appsSn.sort()) ?? 0) + 1;
}
}

View File

@ -1,35 +0,0 @@
import { JwtService } from '@nestjs-lib/auth';
import { Test, TestingModule } from '@nestjs/testing';
import { PipelineTasksService } from '../pipeline-tasks/pipeline-tasks.service';
import { CommitLogsResolver } from './commit-logs.resolver';
import { PipelinesService } from './pipelines.service';
describe('CommitLogsResolver', () => {
let resolver: CommitLogsResolver;
beforeEach(async () => {
const module: TestingModule = await Test.createTestingModule({
providers: [
CommitLogsResolver,
{
provide: PipelinesService,
useValue: {},
},
{
provide: PipelineTasksService,
useValue: {},
},
{
provide: JwtService,
useValue: {},
},
],
}).compile();
resolver = module.get<CommitLogsResolver>(CommitLogsResolver);
});
it('should be defined', () => {
expect(resolver).toBeDefined();
});
});

View File

@ -1,44 +0,0 @@
import { Roles, AccountRole } from '@nestjs-lib/auth';
import { Query } from '@nestjs/graphql';
import {
Args,
Parent,
ResolveField,
Resolver,
Subscription,
} from '@nestjs/graphql';
import { PipelineTasksService } from '../pipeline-tasks/pipeline-tasks.service';
import { Commit, LogFields } from '../repos/dtos/log-list.model';
import { PipelinesService } from './pipelines.service';
@Roles(AccountRole.admin, AccountRole.super)
@Resolver(() => Commit)
export class CommitLogsResolver {
constructor(
private readonly service: PipelinesService,
private readonly taskServices: PipelineTasksService,
) {}
@Subscription(() => String, { resolve: (val) => val, nullable: true })
async syncCommits(
@Args('pipelineId', { type: () => String })
pipelineId: string,
@Args('appInstance', { type: () => String, nullable: true })
appInstance?: string,
) {
const pipeline = await this.service.findOneWithProject(pipelineId);
const syncCommitsPromise = this.service.syncCommits(pipeline, appInstance);
return (async function* () {
yield await syncCommitsPromise;
})();
}
@ResolveField()
async tasks(@Parent() commit: LogFields) {
return await this.taskServices.listTasksByCommitHash(commit.hash);
}
@Query(() => [Commit], { nullable: true })
async commits(@Args('pipelineId', { type: () => String }) id: string) {
const pipeline = await this.service.findOneWithProject(id);
return await this.service.listCommits(pipeline);
}
}

View File

@ -1,13 +1,11 @@
import { Type } from 'class-transformer';
import { InputType } from '@nestjs/graphql'; import { InputType } from '@nestjs/graphql';
import { WorkUnitMetadata } from '../../pipeline-tasks/models/work-unit-metadata.model'; import { WorkUnitMetadata } from '../../pipeline-tasks/models/work-unit-metadata.model';
import { import {
IsInstance, IsObject,
IsOptional, IsOptional,
IsString, IsString,
IsUUID, IsUUID,
MaxLength, MaxLength,
ValidateNested,
} from 'class-validator'; } from 'class-validator';
@InputType({ isAbstract: true }) @InputType({ isAbstract: true })
@ -23,9 +21,7 @@ export class CreatePipelineInput {
@MaxLength(32) @MaxLength(32)
name: string; name: string;
@Type(() => WorkUnitMetadata)
@IsOptional() @IsOptional()
@ValidateNested() @IsObject()
@IsInstance(WorkUnitMetadata)
workUnitMetadata: WorkUnitMetadata; workUnitMetadata: WorkUnitMetadata;
} }

View File

@ -1,9 +1,5 @@
import { InputType, OmitType } from '@nestjs/graphql'; import { InputType } from '@nestjs/graphql';
import { CreatePipelineInput } from './create-pipeline.input'; import { CreatePipelineInput } from './create-pipeline.input';
@InputType() @InputType()
export class UpdatePipelineInput extends OmitType(CreatePipelineInput, [ export class UpdatePipelineInput extends CreatePipelineInput {}
'projectId',
]) {
id: string;
}

View File

@ -3,27 +3,16 @@ import { PipelinesResolver } from './pipelines.resolver';
import { PipelinesService } from './pipelines.service'; import { PipelinesService } from './pipelines.service';
import { TypeOrmModule } from '@nestjs/typeorm'; import { TypeOrmModule } from '@nestjs/typeorm';
import { Pipeline } from './pipeline.entity'; import { Pipeline } from './pipeline.entity';
import { CommitLogsResolver } from './commit-logs.resolver'; import { BullModule } from '@nestjs/bull';
import { PipelineTasksModule } from '../pipeline-tasks/pipeline-tasks.module'; import { LIST_LOGS_TASK } from '../repos/repos.constants';
import { ReposModule } from '../repos/repos.module';
import { RabbitMQModule } from '@golevelup/nestjs-rabbitmq';
import { ConfigModule, ConfigService } from '@nestjs/config';
import { CommonsModule } from '../commons/commons.module';
@Module({ @Module({
imports: [ imports: [
CommonsModule,
TypeOrmModule.forFeature([Pipeline]), TypeOrmModule.forFeature([Pipeline]),
PipelineTasksModule, BullModule.registerQueue({
RabbitMQModule.forRootAsync(RabbitMQModule, { name: LIST_LOGS_TASK,
imports: [ConfigModule],
useFactory: (configService: ConfigService) => ({
uri: configService.get<string>('db.rabbitmq.uri'),
exchanges: [],
}),
inject: [ConfigService],
}), }),
], ],
providers: [PipelinesResolver, PipelinesService, CommitLogsResolver], providers: [PipelinesResolver, PipelinesService],
}) })
export class PipelinesModule {} export class PipelinesModule {}

View File

@ -1,4 +1,3 @@
import { JwtService } from '@nestjs-lib/auth';
import { Test, TestingModule } from '@nestjs/testing'; import { Test, TestingModule } from '@nestjs/testing';
import { PipelinesResolver } from './pipelines.resolver'; import { PipelinesResolver } from './pipelines.resolver';
import { PipelinesService } from './pipelines.service'; import { PipelinesService } from './pipelines.service';
@ -14,10 +13,6 @@ describe('PipelinesResolver', () => {
provide: PipelinesService, provide: PipelinesService,
useValue: {}, useValue: {},
}, },
{
provide: JwtService,
useValue: {},
},
], ],
}).compile(); }).compile();

View File

@ -1,39 +1,39 @@
import { Args, Mutation, Query, Resolver } from '@nestjs/graphql'; import { Args, Mutation, Query, Resolver, Subscription } from '@nestjs/graphql';
import { CreatePipelineInput } from './dtos/create-pipeline.input'; import { CreatePipelineInput } from './dtos/create-pipeline.input';
import { UpdatePipelineInput } from './dtos/update-pipeline.input'; import { UpdatePipelineInput } from './dtos/update-pipeline.input';
import { Pipeline } from './pipeline.entity'; import { Pipeline } from './pipeline.entity';
import { PipelinesService } from './pipelines.service'; import { PipelinesService } from './pipelines.service';
import { ListPipelineArgs } from './dtos/list-pipelines.args'; import { ListPipelineArgs } from './dtos/list-pipelines.args';
import { Roles, AccountRole } from '@nestjs-lib/auth'; import { LogList } from '../repos/dtos/log-list.model';
@Roles(AccountRole.admin, AccountRole.super)
@Resolver() @Resolver()
export class PipelinesResolver { export class PipelinesResolver {
constructor(private readonly service: PipelinesService) {} constructor(private readonly service: PipelinesService) {}
@Query(() => [Pipeline]) @Query(() => [Pipeline])
async pipelines(@Args() dto: ListPipelineArgs) { async listPipelines(@Args() dto: ListPipelineArgs) {
return await this.service.list(dto); return await this.service.list(dto);
} }
@Query(() => Pipeline) @Query(() => Pipeline)
async pipeline(@Args('id', { type: () => String }) id: string) { async findPipeline(@Args('id', { type: () => String }) id: string) {
return await this.service.findOne(id); return await this.service.findOne(id);
} }
@Mutation(() => Pipeline) @Mutation(() => Pipeline)
async createPipeline( async createPipeline(
@Args('pipeline', { type: () => CreatePipelineInput }) @Args('pipeline', { type: () => CreatePipelineInput })
dto: CreatePipelineInput, dto: UpdatePipelineInput,
) { ) {
return await this.service.create(dto); return await this.service.create(dto);
} }
@Mutation(() => Pipeline) @Mutation(() => Pipeline)
async updatePipeline( async modifyPipeline(
@Args('pipeline', { type: () => UpdatePipelineInput }) @Args('id', { type: () => String }) id: string,
@Args('Pipeline', { type: () => UpdatePipelineInput })
dto: UpdatePipelineInput, dto: UpdatePipelineInput,
) { ) {
const tmp = await this.service.update(dto); const tmp = await this.service.update(id, dto);
console.log(tmp); console.log(tmp);
return tmp; return tmp;
} }
@ -42,4 +42,16 @@ export class PipelinesResolver {
async deletePipeline(@Args('id', { type: () => String }) id: string) { async deletePipeline(@Args('id', { type: () => String }) id: string) {
return await this.service.remove(id); return await this.service.remove(id);
} }
@Subscription(() => LogList, {
resolve: (value) => {
return value;
},
})
async listLogsForPipeline(@Args('id', { type: () => String }) id: string) {
const job = await this.service.listLogsForPipeline(id);
return (async function* () {
yield await job.finished();
})();
}
} }

View File

@ -2,14 +2,18 @@ import { Test, TestingModule } from '@nestjs/testing';
import { PipelinesService } from './pipelines.service'; import { PipelinesService } from './pipelines.service';
import { Pipeline } from './pipeline.entity'; import { Pipeline } from './pipeline.entity';
import { getRepositoryToken } from '@nestjs/typeorm'; import { getRepositoryToken } from '@nestjs/typeorm';
import { getQueueToken } from '@nestjs/bull';
import { LIST_LOGS_TASK } from '../repos/repos.constants';
import { Repository } from 'typeorm'; import { Repository } from 'typeorm';
import { Project } from '../projects/project.entity'; import { Project } from '../projects/project.entity';
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq'; import { Job, Queue } from 'bull';
import { ListLogsOption } from '../repos/models/list-logs.options';
describe('PipelinesService', () => { describe('PipelinesService', () => {
let service: PipelinesService; let service: PipelinesService;
let repository: Repository<Pipeline>; let repository: Repository<Pipeline>;
let pipeline: Pipeline; let pipeline: Pipeline;
let queue: Queue<ListLogsOption>;
beforeEach(async () => { beforeEach(async () => {
pipeline = Object.assign(new Pipeline(), { pipeline = Object.assign(new Pipeline(), {
@ -33,17 +37,33 @@ describe('PipelinesService', () => {
}, },
}, },
{ {
provide: AmqpConnection, provide: getQueueToken(LIST_LOGS_TASK),
useValue: {}, useValue: {
add: jest.fn().mockImplementation(() => ({ id: 1 } as Job)),
},
}, },
], ],
}).compile(); }).compile();
service = module.get<PipelinesService>(PipelinesService); service = module.get<PipelinesService>(PipelinesService);
repository = module.get(getRepositoryToken(Pipeline)); repository = module.get(getRepositoryToken(Pipeline));
queue = module.get(getQueueToken(LIST_LOGS_TASK));
}); });
it('should be defined', () => { it('should be defined', () => {
expect(service).toBeDefined(); expect(service).toBeDefined();
}); });
describe('listLogsForPipeline', () => {
it('should send task to queue.', async () => {
const add = jest.spyOn(queue, 'add');
await expect(
service.listLogsForPipeline('test-pipeline'),
).resolves.toEqual({ id: 1 });
expect(add).toBeCalledWith({
project: pipeline.project,
branch: pipeline.branch,
});
});
});
}); });

View File

@ -6,16 +6,10 @@ import { BaseDbService } from '../commons/services/base-db.service';
import { CreatePipelineInput } from './dtos/create-pipeline.input'; import { CreatePipelineInput } from './dtos/create-pipeline.input';
import { UpdatePipelineInput } from './dtos/update-pipeline.input'; import { UpdatePipelineInput } from './dtos/update-pipeline.input';
import { ListPipelineArgs } from './dtos/list-pipelines.args'; import { ListPipelineArgs } from './dtos/list-pipelines.args';
import { import { InjectQueue } from '@nestjs/bull';
EXCHANGE_REPO, import { LIST_LOGS_TASK } from '../repos/repos.constants';
ROUTE_FETCH, import { Queue } from 'bull';
ROUTE_LIST_COMMITS, import { ListLogsOption } from '../repos/models/list-logs.options';
} from '../repos/repos.constants';
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
import { Commit } from '../repos/dtos/log-list.model';
import { getAppInstanceRouteKey } from '../commons/utils/rabbit-mq';
import { ApplicationException } from '../commons/exceptions/application.exception';
import { plainToClass } from 'class-transformer';
@Injectable() @Injectable()
export class PipelinesService extends BaseDbService<Pipeline> { export class PipelinesService extends BaseDbService<Pipeline> {
@ -23,7 +17,8 @@ export class PipelinesService extends BaseDbService<Pipeline> {
constructor( constructor(
@InjectRepository(Pipeline) @InjectRepository(Pipeline)
readonly repository: Repository<Pipeline>, readonly repository: Repository<Pipeline>,
private readonly amqpConnection: AmqpConnection, @InjectQueue(LIST_LOGS_TASK)
private readonly listLogsQueue: Queue<ListLogsOption>,
) { ) {
super(repository); super(repository);
} }
@ -31,48 +26,30 @@ export class PipelinesService extends BaseDbService<Pipeline> {
return this.repository.find(dto); return this.repository.find(dto);
} }
async findOneWithProject(id: string) {
return await this.repository.findOne({
where: { id },
relations: ['project'],
});
}
async create(dto: CreatePipelineInput) { async create(dto: CreatePipelineInput) {
await this.isDuplicateEntity(dto); await this.isDuplicateEntity(dto);
return await this.repository.save(this.repository.create(dto)); return await this.repository.save(this.repository.create(dto));
} }
async update(dto: UpdatePipelineInput) { async update(id: string, dto: UpdatePipelineInput) {
const old = await this.findOne(dto.id); await this.isDuplicateEntityForUpdate(id, dto);
await this.isDuplicateEntityForUpdate(old, dto); const old = await this.findOne(id);
return await this.repository.save(this.repository.merge(old, dto)); return await this.repository.save(this.repository.merge(old, dto));
} }
async remove(id: string) { async remove(id: string) {
return (await this.repository.softDelete({ id })).affected; return (await this.repository.softDelete({ id })).affected;
} }
async syncCommits(pipeline: Pipeline, appInstance?: string) {
return await this.amqpConnection.request<string | null>({ async listLogsForPipeline(id: string) {
exchange: EXCHANGE_REPO, const pipeline = await this.repository.findOneOrFail({
routingKey: getAppInstanceRouteKey(ROUTE_FETCH, appInstance), where: { id },
payload: pipeline, relations: ['project'],
timeout: 120_000,
}); });
} const job = await this.listLogsQueue.add({
async listCommits(pipeline: Pipeline) { project: pipeline.project,
return await this.amqpConnection branch: pipeline.branch,
.request<[Error, Commit[]]>({
exchange: EXCHANGE_REPO,
routingKey: ROUTE_LIST_COMMITS,
payload: pipeline,
timeout: 30_000,
})
.then(([error, list]) => {
if (error) {
throw new ApplicationException(error);
}
return plainToClass(Commit, list);
}); });
return job;
} }
} }

View File

@ -21,10 +21,7 @@ export class CreateProjectInput {
comment: string; comment: string;
@Matches( @Matches(
/^(?:ssh:\/\/)?(?:[\w\d-_]+@)?(?:[\w\d-_]+\.)*\w{2,10}(?::\d{1,5})?(?:\/[\w\d-_.]+)*/, /^(?:ssh:\/\/)?(?:[\w\d-_]+@)(?:[\w\d-_]+\.)*\w{2,10}(?::\d{1,5})?(?:\/[\w\d-_.]+)*/,
{
message: 'wrong ssh url',
},
) )
@MaxLength(256) @MaxLength(256)
sshUrl: string; sshUrl: string;

View File

@ -1,9 +1,5 @@
import { InputType } from '@nestjs/graphql'; import { InputType } from '@nestjs/graphql';
import { IsUUID } from 'class-validator';
import { CreateProjectInput } from './create-project.input'; import { CreateProjectInput } from './create-project.input';
@InputType() @InputType()
export class UpdateProjectInput extends CreateProjectInput { export class UpdateProjectInput extends CreateProjectInput {}
@IsUUID()
id: string;
}

View File

@ -1,3 +0,0 @@
export const EXCHANGE_PROJECT_TOPIC = 'project.topic';
export const EXCHANGE_PROJECT_FANOUT = 'project.fanout';
export const ROUTE_PROJECT_CHANGE = 'project-change';

View File

@ -3,33 +3,9 @@ import { ProjectsService } from './projects.service';
import { ProjectsResolver } from './projects.resolver'; import { ProjectsResolver } from './projects.resolver';
import { TypeOrmModule } from '@nestjs/typeorm'; import { TypeOrmModule } from '@nestjs/typeorm';
import { Project } from './project.entity'; import { Project } from './project.entity';
import { RabbitMQModule } from '@golevelup/nestjs-rabbitmq';
import { ConfigModule, ConfigService } from '@nestjs/config';
import { EXCHANGE_PROJECT_FANOUT } from './projects.constants';
import { CommonsModule } from '../commons/commons.module';
@Module({ @Module({
imports: [ imports: [TypeOrmModule.forFeature([Project])],
CommonsModule,
TypeOrmModule.forFeature([Project]),
RabbitMQModule.forRootAsync(RabbitMQModule, {
imports: [ConfigModule],
useFactory: (configService: ConfigService) => ({
uri: configService.get<string>('db.rabbitmq.uri'),
exchanges: [
{
name: EXCHANGE_PROJECT_FANOUT,
type: 'fanout',
options: {
durable: false,
autoDelete: true,
},
},
],
}),
inject: [ConfigService],
}),
],
providers: [ProjectsService, ProjectsResolver], providers: [ProjectsService, ProjectsResolver],
exports: [ProjectsService], exports: [ProjectsService],
}) })

View File

@ -1,4 +1,3 @@
import { JwtService } from '@nestjs-lib/auth';
import { Test, TestingModule } from '@nestjs/testing'; import { Test, TestingModule } from '@nestjs/testing';
import { ProjectsResolver } from './projects.resolver'; import { ProjectsResolver } from './projects.resolver';
import { ProjectsService } from './projects.service'; import { ProjectsService } from './projects.service';
@ -14,10 +13,6 @@ describe('ProjectsResolver', () => {
provide: ProjectsService, provide: ProjectsService,
useValue: {}, useValue: {},
}, },
{
provide: JwtService,
useValue: {},
},
], ],
}).compile(); }).compile();

View File

@ -1,44 +1,43 @@
import { AccountRole, Roles } from '@nestjs-lib/auth';
import { Args, Mutation, Query, Resolver } from '@nestjs/graphql'; import { Args, Mutation, Query, Resolver } from '@nestjs/graphql';
import { CreateProjectInput } from './dtos/create-project.input'; import { CreateProjectInput } from './dtos/create-project.input';
import { UpdateProjectInput } from './dtos/update-project.input'; import { UpdateProjectInput } from './dtos/update-project.input';
import { Project } from './project.entity'; import { Project } from './project.entity';
import { ProjectsService } from './projects.service'; import { ProjectsService } from './projects.service';
@Roles(AccountRole.admin, AccountRole.super)
@Resolver(() => Project) @Resolver(() => Project)
export class ProjectsResolver { export class ProjectsResolver {
constructor(private readonly service: ProjectsService) {} constructor(private readonly service: ProjectsService) {}
@Query(() => [Project]) @Query(() => [Project])
async projects() { async findProjects() {
return await this.service.list(); return await this.service.list();
} }
@Query(() => Project) @Query(() => Project)
async project(@Args('id', { type: () => String }) id: string) { async findProject(@Args('id', { type: () => String }) id: string) {
return await this.service.findOne(id); return await this.service.findOne(id);
} }
@Mutation(() => Project) @Mutation(() => Project)
async createProject( async createProject(
@Args('project', { type: () => CreateProjectInput }) @Args('project', { type: () => CreateProjectInput })
dto: CreateProjectInput, dto: UpdateProjectInput,
) { ) {
return await this.service.create(dto); return await this.service.create(dto);
} }
@Mutation(() => Project) @Mutation(() => Project)
async updateProject( async modifyProject(
@Args('id', { type: () => String }) id: string,
@Args('project', { type: () => UpdateProjectInput }) @Args('project', { type: () => UpdateProjectInput })
dto: UpdateProjectInput, dto: UpdateProjectInput,
) { ) {
const tmp = await this.service.update(dto); const tmp = await this.service.update(id, dto);
console.log(tmp); console.log(tmp);
return tmp; return tmp;
} }
@Mutation(() => Number) @Mutation(() => Number)
async removeProject(@Args('id', { type: () => String }) id: string) { async deleteProject(@Args('id', { type: () => String }) id: string) {
return await this.service.remove(id); return await this.service.remove(id);
} }
} }

View File

@ -2,7 +2,6 @@ import { Test, TestingModule } from '@nestjs/testing';
import { ProjectsService } from './projects.service'; import { ProjectsService } from './projects.service';
import { getRepositoryToken } from '@nestjs/typeorm'; import { getRepositoryToken } from '@nestjs/typeorm';
import { Project } from './project.entity'; import { Project } from './project.entity';
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
describe('ProjectsService', () => { describe('ProjectsService', () => {
let service: ProjectsService; let service: ProjectsService;
@ -15,10 +14,6 @@ describe('ProjectsService', () => {
provide: getRepositoryToken(Project), provide: getRepositoryToken(Project),
useValue: {}, useValue: {},
}, },
{
provide: AmqpConnection,
useValue: {},
},
], ],
}).compile(); }).compile();

View File

@ -5,11 +5,6 @@ import { Repository } from 'typeorm';
import { CreateProjectInput } from './dtos/create-project.input'; import { CreateProjectInput } from './dtos/create-project.input';
import { Project } from './project.entity'; import { Project } from './project.entity';
import { UpdateProjectInput } from './dtos/update-project.input'; import { UpdateProjectInput } from './dtos/update-project.input';
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
import {
EXCHANGE_PROJECT_FANOUT,
ROUTE_PROJECT_CHANGE,
} from './projects.constants';
@Injectable() @Injectable()
export class ProjectsService extends BaseDbService<Project> { export class ProjectsService extends BaseDbService<Project> {
@ -17,7 +12,6 @@ export class ProjectsService extends BaseDbService<Project> {
constructor( constructor(
@InjectRepository(Project) @InjectRepository(Project)
readonly repository: Repository<Project>, readonly repository: Repository<Project>,
private readonly amqpConnection: AmqpConnection,
) { ) {
super(repository); super(repository);
} }
@ -31,15 +25,10 @@ export class ProjectsService extends BaseDbService<Project> {
return await this.repository.save(this.repository.create(dto)); return await this.repository.save(this.repository.create(dto));
} }
async update(dto: UpdateProjectInput) { async update(id: string, dto: UpdateProjectInput) {
await this.isDuplicateEntityForUpdate(dto.id, dto); await this.isDuplicateEntityForUpdate(id, dto);
const old = await this.findOne(dto.id); const old = await this.findOne(id);
const project = await this.repository.save(this.repository.merge(old, dto)); return await this.repository.save(this.repository.merge(old, dto));
this.amqpConnection.publish(EXCHANGE_PROJECT_FANOUT, ROUTE_PROJECT_CHANGE, [
project,
old,
]);
return project;
} }
async remove(id: string) { async remove(id: string) {

View File

@ -1,20 +1,6 @@
import { ObjectType, Field } from '@nestjs/graphql'; import { ObjectType, Field } from '@nestjs/graphql';
import { Type } from 'class-transformer';
import { LogResult, DefaultLogFields } from 'simple-git'; import { LogResult, DefaultLogFields } from 'simple-git';
import { PipelineTask } from '../../pipeline-tasks/pipeline-task.entity';
@ObjectType()
export class Commit {
hash: string;
@Type(() => Date)
date: Date;
message: string;
refs: string;
body: string;
author_name: string;
author_email: string;
tasks: PipelineTask[];
}
@ObjectType() @ObjectType()
export class LogFields { export class LogFields {
hash: string; hash: string;
@ -24,7 +10,6 @@ export class LogFields {
body: string; body: string;
author_name: string; author_name: string;
author_email: string; author_email: string;
tasks: PipelineTask[];
} }
@ObjectType() @ObjectType()

View File

@ -0,0 +1,14 @@
import { ReposService } from './repos.service';
import { Processor, Process } from '@nestjs/bull';
import { Job } from 'bull';
import { ListLogsOption } from './models/list-logs.options';
import { LIST_LOGS_TASK } from './repos.constants';
@Processor(LIST_LOGS_TASK)
export class ListLogsConsumer {
constructor(private readonly service: ReposService) {}
@Process()
async listLogs(job: Job<ListLogsOption>) {
const logs = await this.service.listLogs(job.data);
return logs;
}
}

View File

@ -1,6 +1,3 @@
export const EXCHANGE_REPO = 'fennec.repo'; export const LIST_LOGS_TASK = 'LIST_LOGS_TASK';
export const ROUTE_FETCH = 'fetch'; export const LIST_LOGS_PUB_SUB = 'LIST_LOGS_PUB_SUB';
export const ROUTE_LIST_COMMITS = 'list-commits'; export const LIST_LOGS_DONE = 'LIST_LOGS_DONE';
export const QUEUE_LIST_COMMITS = 'list-commits';
export const QUEUE_FETCH = 'repo-fetch';
export const QUEUE_REFRESH_REPO = 'refresh-repo';

View File

@ -3,37 +3,23 @@ import { TypeOrmModule } from '@nestjs/typeorm';
import { Project } from '../projects/project.entity'; import { Project } from '../projects/project.entity';
import { ReposResolver } from './repos.resolver'; import { ReposResolver } from './repos.resolver';
import { ReposService } from './repos.service'; import { ReposService } from './repos.service';
import { ConfigModule, ConfigService } from '@nestjs/config'; import { ConfigModule } from '@nestjs/config';
import { ProjectsModule } from '../projects/projects.module'; import { ProjectsModule } from '../projects/projects.module';
import { EXCHANGE_REPO } from './repos.constants'; import { BullModule } from '@nestjs/bull';
import { RabbitMQModule } from '@golevelup/nestjs-rabbitmq'; import { LIST_LOGS_TASK, LIST_LOGS_PUB_SUB } from './repos.constants';
import { CommonsModule } from '../commons/commons.module'; import { PubSub } from 'graphql-subscriptions';
import { ListLogsConsumer } from './list-logs.consumer';
@Module({ @Module({
imports: [ imports: [
TypeOrmModule.forFeature([Project]), TypeOrmModule.forFeature([Project]),
ConfigModule, ConfigModule,
ProjectsModule, ProjectsModule,
CommonsModule, BullModule.registerQueue({
RabbitMQModule.forRootAsync(RabbitMQModule, { name: LIST_LOGS_TASK,
imports: [ConfigModule],
useFactory: (configService: ConfigService) => ({
uri: configService.get<string>('db.rabbitmq.uri'),
exchanges: [
{
name: EXCHANGE_REPO,
type: 'topic',
options: {
durable: true,
autoDelete: true,
},
},
],
}),
inject: [ConfigService],
}), }),
], ],
providers: [ReposResolver, ReposService], providers: [ReposResolver, ReposService, ListLogsConsumer],
exports: [ReposService], exports: [ReposService],
}) })
export class ReposModule {} export class ReposModule {}

View File

@ -9,10 +9,6 @@ import configuration from '../commons/config/configuration';
import { PipelineTask } from '../pipeline-tasks/pipeline-task.entity'; import { PipelineTask } from '../pipeline-tasks/pipeline-task.entity';
import { join } from 'path'; import { join } from 'path';
import { readFile } from 'fs/promises'; import { readFile } from 'fs/promises';
import { getLoggerToken, PinoLogger } from 'nestjs-pino';
import { Nack } from '@golevelup/nestjs-rabbitmq';
import { getInstanceName } from '../commons/utils/rabbit-mq';
import { RedisMutexService } from '../commons/redis-mutex/redis-mutex.service';
const getTest1Project = () => const getTest1Project = () =>
({ ({
@ -49,18 +45,6 @@ describe('ReposService', () => {
provide: getRepositoryToken(Project), provide: getRepositoryToken(Project),
useFactory: repositoryMockFactory, useFactory: repositoryMockFactory,
}, },
{
provide: getLoggerToken(ReposService.name),
useValue: new PinoLogger({}),
},
{
provide: RedisMutexService,
useValue: {
lock: jest.fn(() =>
Promise.resolve(() => Promise.resolve(undefined)),
),
},
},
], ],
}).compile(); }).compile();
@ -155,57 +139,4 @@ describe('ReposService', () => {
); );
}); });
}); });
describe('fetch', () => {
it('success', async () => {
const project = new Project();
const pipeline = new Pipeline();
pipeline.branch = 'test';
const fetch = jest.fn((_: any) => Promise.resolve());
pipeline.project = project;
const getGit = jest.spyOn(service, 'getGit').mockImplementation(() =>
Promise.resolve({
fetch,
} as any),
);
await expect(service.fetch(pipeline)).resolves.toEqual(getInstanceName());
expect(getGit).toBeCalledTimes(1);
expect(getGit.mock.calls[0]?.[0]).toEqual(project);
expect(fetch).toBeCalledTimes(1);
expect(fetch.mock.calls[0]?.[0]).toMatchObject([
'origin',
'test',
'--depth=100',
]);
});
it('failed a', async () => {
const project = new Project();
const pipeline = new Pipeline();
pipeline.branch = 'test';
const fetch = jest.fn((_: any) => Promise.resolve());
pipeline.project = project;
const getGit = jest
.spyOn(service, 'getGit')
.mockImplementation(() => Promise.reject('error'));
await expect(service.fetch(pipeline)).resolves.toMatchObject(new Nack());
expect(getGit).toBeCalledTimes(1);
expect(getGit.mock.calls[0]?.[0]).toEqual(project);
expect(fetch).toBeCalledTimes(0);
});
it('failed b', async () => {
const project = new Project();
const pipeline = new Pipeline();
pipeline.branch = 'test';
const fetch = jest.fn((_: any) => Promise.reject('error'));
pipeline.project = project;
const getGit = jest.spyOn(service, 'getGit').mockImplementation(() =>
Promise.resolve({
fetch,
} as any),
);
await expect(service.fetch(pipeline)).resolves.toMatchObject(new Nack());
expect(getGit).toBeCalledTimes(1);
expect(fetch).toBeCalledTimes(1);
});
});
}); });

View File

@ -1,4 +1,5 @@
import { ListLogsOption } from './models/list-logs.options'; import { ListLogsOption } from './models/list-logs.options';
import { Pipeline } from './../pipelines/pipeline.entity';
import { PipelineTask } from './../pipeline-tasks/pipeline-task.entity'; import { PipelineTask } from './../pipeline-tasks/pipeline-task.entity';
import { Injectable, NotFoundException } from '@nestjs/common'; import { Injectable, NotFoundException } from '@nestjs/common';
import { InjectRepository } from '@nestjs/typeorm'; import { InjectRepository } from '@nestjs/typeorm';
@ -9,31 +10,8 @@ import { gitP } from 'simple-git';
import { Repository } from 'typeorm'; import { Repository } from 'typeorm';
import { Project } from '../projects/project.entity'; import { Project } from '../projects/project.entity';
import { ListBranchesArgs } from './dtos/list-branches.args'; import { ListBranchesArgs } from './dtos/list-branches.args';
import { ListLogsArgs } from './dtos/list-logs.args';
import { ConfigService } from '@nestjs/config'; import { ConfigService } from '@nestjs/config';
import { Commit } from './dtos/log-list.model';
import { Nack, RabbitRPC, RabbitSubscribe } from '@golevelup/nestjs-rabbitmq';
import { Pipeline } from '../pipelines/pipeline.entity';
import { InjectPinoLogger, PinoLogger } from 'nestjs-pino';
import {
EXCHANGE_REPO,
QUEUE_FETCH,
QUEUE_LIST_COMMITS,
QUEUE_REFRESH_REPO,
ROUTE_FETCH,
ROUTE_LIST_COMMITS,
} from './repos.constants';
import { getSelfInstanceQueueKey } from '../commons/utils/rabbit-mq';
import {
getInstanceName,
getSelfInstanceRouteKey,
} from '../commons/utils/rabbit-mq';
import { ApplicationException } from '../commons/exceptions/application.exception';
import {
EXCHANGE_PROJECT_FANOUT,
ROUTE_PROJECT_CHANGE,
} from '../projects/projects.constants';
import { RedisMutexService } from '../commons/redis-mutex/redis-mutex.service';
import { rm } from 'fs/promises';
const DEFAULT_REMOTE_NAME = 'origin'; const DEFAULT_REMOTE_NAME = 'origin';
const INFO_PATH = '@info'; const INFO_PATH = '@info';
@ -43,9 +21,6 @@ export class ReposService {
@InjectRepository(Project) @InjectRepository(Project)
private readonly projectRepository: Repository<Project>, private readonly projectRepository: Repository<Project>,
private readonly configService: ConfigService, private readonly configService: ConfigService,
@InjectPinoLogger(ReposService.name)
private readonly logger: PinoLogger,
private readonly redisMutexService: RedisMutexService,
) {} ) {}
getWorkspaceRoot(project: Project): string { getWorkspaceRoot(project: Project): string {
@ -56,19 +31,7 @@ export class ReposService {
); );
} }
getDeployRoot(task: PipelineTask) { async getGit(project: Project, workspaceRoot?: string) {
return join(
this.configService.get<string>('workspaces.root'),
encodeURIComponent(task.pipeline.project.name),
encodeURIComponent(`deploy-${task.pipeline.name}`),
);
}
async getGit(
project: Project,
workspaceRoot?: string,
{ fetch = true } = {},
) {
if (!workspaceRoot) { if (!workspaceRoot) {
workspaceRoot = this.getWorkspaceRoot(project); workspaceRoot = this.getWorkspaceRoot(project);
} }
@ -81,9 +44,7 @@ export class ReposService {
await git.init(); await git.init();
await git.addRemote(DEFAULT_REMOTE_NAME, project.sshUrl); await git.addRemote(DEFAULT_REMOTE_NAME, project.sshUrl);
} }
if (fetch) {
await git.fetch(); await git.fetch();
}
return git; return git;
} }
@ -128,113 +89,4 @@ export class ReposService {
encodeURIComponent(`${task.pipeline.name}-${task.commit}`), encodeURIComponent(`${task.pipeline.name}-${task.commit}`),
); );
} }
async checkout4Task(task: PipelineTask): Promise<string> {
const path = this.getWorkspaceRootByTask(task);
await this.checkout(task, path);
return path;
}
@RabbitRPC({
exchange: EXCHANGE_REPO,
routingKey: [
ROUTE_LIST_COMMITS,
getSelfInstanceRouteKey(ROUTE_LIST_COMMITS),
],
queue: getSelfInstanceQueueKey(QUEUE_LIST_COMMITS),
queueOptions: {
autoDelete: true,
},
})
async listCommits(pipeline: Pipeline): Promise<[Error, Commit[]?]> {
const git = await this.getGit(pipeline.project, undefined, {
fetch: false,
});
try {
const data = await git.log([
'-100',
'--branches',
`remotes/origin/${pipeline.branch}`,
'--',
]);
return [
null,
data.all.map(
(it) =>
({
...it,
date: new Date(it.date),
} as Commit),
),
];
} catch (error) {
this.logger.error(
{ error, pipeline },
'[listCommits] %s',
error?.message,
);
return [new ApplicationException(error)];
}
}
@RabbitRPC({
exchange: EXCHANGE_REPO,
routingKey: [ROUTE_FETCH, getSelfInstanceRouteKey(ROUTE_FETCH)],
queue: getSelfInstanceQueueKey(QUEUE_FETCH),
queueOptions: {
autoDelete: true,
},
})
async fetch(pipeline: Pipeline): Promise<string | null | Nack> {
const unlock = await this.redisMutexService.lock(
`repo-project-${pipeline.projectId}`,
);
try {
const git = await this.getGit(pipeline.project, undefined, {
fetch: false,
});
await git.fetch(['origin', pipeline.branch, '--depth=100']);
return getInstanceName();
} catch (error) {
this.logger.error({ error, pipeline }, '[fetch] %s', error?.message);
return new Nack();
} finally {
await unlock();
}
}
@RabbitSubscribe({
exchange: EXCHANGE_PROJECT_FANOUT,
routingKey: ROUTE_PROJECT_CHANGE,
queue: QUEUE_REFRESH_REPO,
queueOptions: {
autoDelete: true,
durable: true,
},
})
async refreshRepo([project]: [Project]) {
this.logger.info({ project }, '[refreshRepo] start');
const unlock = await this.redisMutexService.lock(
`repo-project-${project.id}`,
{
timeout: null,
},
);
try {
const path = join(
this.configService.get<string>('workspaces.root'),
encodeURIComponent(project.name),
);
await rm(path, { recursive: true });
this.logger.info({ project }, '[refreshRepo] success');
} catch (error) {
this.logger.error(
{ project, error },
'[refreshRepo] failed. $s',
error.message,
);
} finally {
await unlock();
}
}
} }

View File

@ -4,6 +4,7 @@ import { PipelineTasksModule } from '../pipeline-tasks/pipeline-tasks.module';
import { GiteaWebhooksController } from './gitea-webhooks.controller'; import { GiteaWebhooksController } from './gitea-webhooks.controller';
import { WebhookLog } from './webhook-log.entity'; import { WebhookLog } from './webhook-log.entity';
import { WebhooksService } from './webhooks.service'; import { WebhooksService } from './webhooks.service';
import { raw } from 'body-parser';
@Module({ @Module({
imports: [TypeOrmModule.forFeature([WebhookLog]), PipelineTasksModule], imports: [TypeOrmModule.forFeature([WebhookLog]), PipelineTasksModule],
@ -11,4 +12,9 @@ import { WebhooksService } from './webhooks.service';
providers: [WebhooksService], providers: [WebhooksService],
}) })
export class WebhooksModule { export class WebhooksModule {
// configure(consumer: MiddlewareConsumer) {
// consumer
// .apply(raw({ type: 'application/json' }))
// .forRoutes(GiteaWebhooksController);
// }
} }

View File

@ -1,15 +0,0 @@
module.exports = {
apps: [
{
name: 'app',
script: __dirname + '/index.js',
watch: false,
ignore_watch: ['node_modules'],
log_date_format: 'MM-DD HH:mm:ss.SSS Z',
env: {},
max_restarts: 5,
kill_timeout: 10_000,
wait_ready: true,
},
],
};

View File

@ -1,23 +0,0 @@
import { createServer } from 'http';
var app = createServer(function (req, res) {
res.writeHead(200);
setTimeout(() => {
res.end('hey');
}, 2000);
});
var listener = app.listen(0, function () {
console.log('Listening on port ' + listener.address().port);
setTimeout(() => {
// Here we send the ready signal to PM2
process.send('ready');
}, 5000);
});
process.on('SIGINT', function () {
listener.close();
setTimeout(() => {
process.exit(0);
}, 2000);
});

View File

@ -1,10 +0,0 @@
{
"name": "deploy-service",
"version": "1.0.0",
"description": "For Test",
"main": "index.js",
"type": "module",
"scripts": {},
"author": "Ivan Li",
"license": "ISC"
}

View File

@ -1,16 +0,0 @@
import { createServer } from 'http';
var app = createServer(function (req, res) {
res.writeHead(200);
setTimeout(() => {
res.end('pm2');
}, 2000);
});
var listener = app.listen(33333, function () {
console.log('Listening on port ' + listener.address().port);
setTimeout(() => {
// Here we send the ready signal to PM2
process.send('ready');
}, 5000);
});

View File

@ -7,10 +7,9 @@
"experimentalDecorators": true, "experimentalDecorators": true,
"allowSyntheticDefaultImports": true, "allowSyntheticDefaultImports": true,
"target": "es2017", "target": "es2017",
"lib": ["ES2021"],
"sourceMap": true, "sourceMap": true,
"outDir": "./dist", "outDir": "./dist",
"rootDir": "./src",
"baseUrl": "./", "baseUrl": "./",
}, "incremental": true
}
} }