Compare commits
21 Commits
246c0bd8f8
...
07f19101a5
Author | SHA1 | Date | |
---|---|---|---|
|
07f19101a5 | ||
|
7d84017f9e | ||
|
a231a02c28 | ||
|
7e17de0f15 | ||
|
9d735c582c | ||
|
b626eed859 | ||
|
5b5a657651 | ||
|
a510f411a7 | ||
|
246623b5db | ||
|
37f8ae19be | ||
|
133439bb49 | ||
|
646f68d298 | ||
|
0c3310d3a5 | ||
|
ead32a1204 | ||
|
20612d4301 | ||
|
7091f9df6a | ||
|
3ee41ece67 | ||
|
b3a2b11db9 | ||
|
4041a6fd2a | ||
|
5a8b699e2f | ||
|
86c8bce9ea |
0
.vscode/launch.json
vendored
Normal file
0
.vscode/launch.json
vendored
Normal file
6
.vscode/settings.json
vendored
6
.vscode/settings.json
vendored
@ -1,13 +1,19 @@
|
|||||||
{
|
{
|
||||||
"cSpell.words": [
|
"cSpell.words": [
|
||||||
|
"Mutex",
|
||||||
"Repos",
|
"Repos",
|
||||||
|
"amqp",
|
||||||
"boardcat",
|
"boardcat",
|
||||||
|
"errout",
|
||||||
|
"fanout",
|
||||||
"gitea",
|
"gitea",
|
||||||
|
"golevelup",
|
||||||
"lpush",
|
"lpush",
|
||||||
"lrange",
|
"lrange",
|
||||||
"metatype",
|
"metatype",
|
||||||
"pmessage",
|
"pmessage",
|
||||||
"psubscribe",
|
"psubscribe",
|
||||||
|
"rabbitmq",
|
||||||
"rpop",
|
"rpop",
|
||||||
"rpush"
|
"rpush"
|
||||||
]
|
]
|
||||||
|
@ -14,5 +14,7 @@ db:
|
|||||||
port: 6379
|
port: 6379
|
||||||
password:
|
password:
|
||||||
prefix: fennec
|
prefix: fennec
|
||||||
|
rabbitmq:
|
||||||
|
uri: 'amqp://fennec:fennec@192.168.31.194:5672'
|
||||||
workspaces:
|
workspaces:
|
||||||
root: '/Users/ivanli/Projects/fennec/workspaces'
|
root: '/Users/ivanli/Projects/fennec/workspaces'
|
15609
package-lock.json
generated
15609
package-lock.json
generated
File diff suppressed because it is too large
Load Diff
@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "fennec-be",
|
"name": "fennec-be",
|
||||||
"version": "0.0.1",
|
"version": "0.1.0",
|
||||||
"description": "",
|
"description": "",
|
||||||
"author": "",
|
"author": "",
|
||||||
"private": true,
|
"private": true,
|
||||||
@ -21,6 +21,7 @@
|
|||||||
"test:e2e": "jest --config ./test/jest-e2e.json"
|
"test:e2e": "jest --config ./test/jest-e2e.json"
|
||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
|
"@golevelup/nestjs-rabbitmq": "^1.16.1",
|
||||||
"@nestjs/bull": "^0.3.1",
|
"@nestjs/bull": "^0.3.1",
|
||||||
"@nestjs/common": "^7.5.1",
|
"@nestjs/common": "^7.5.1",
|
||||||
"@nestjs/config": "^0.6.2",
|
"@nestjs/config": "^0.6.2",
|
||||||
@ -28,6 +29,7 @@
|
|||||||
"@nestjs/graphql": "^7.9.8",
|
"@nestjs/graphql": "^7.9.8",
|
||||||
"@nestjs/platform-express": "^7.5.1",
|
"@nestjs/platform-express": "^7.5.1",
|
||||||
"@nestjs/typeorm": "^7.1.5",
|
"@nestjs/typeorm": "^7.1.5",
|
||||||
|
"@types/amqplib": "^0.8.0",
|
||||||
"@types/bull": "^3.15.0",
|
"@types/bull": "^3.15.0",
|
||||||
"@types/ramda": "^0.27.38",
|
"@types/ramda": "^0.27.38",
|
||||||
"apollo-server-express": "^2.19.2",
|
"apollo-server-express": "^2.19.2",
|
||||||
@ -41,9 +43,11 @@
|
|||||||
"graphql-tools": "^7.0.2",
|
"graphql-tools": "^7.0.2",
|
||||||
"ioredis": "^4.25.0",
|
"ioredis": "^4.25.0",
|
||||||
"js-yaml": "^4.0.0",
|
"js-yaml": "^4.0.0",
|
||||||
|
"nestjs-pino": "^1.4.0",
|
||||||
"nestjs-redis": "^1.2.8",
|
"nestjs-redis": "^1.2.8",
|
||||||
"observable-to-async-generator": "^1.0.1-rc",
|
"observable-to-async-generator": "^1.0.1-rc",
|
||||||
"pg": "^8.5.1",
|
"pg": "^8.5.1",
|
||||||
|
"pino-pretty": "^4.7.1",
|
||||||
"ramda": "^0.27.1",
|
"ramda": "^0.27.1",
|
||||||
"reflect-metadata": "^0.1.13",
|
"reflect-metadata": "^0.1.13",
|
||||||
"rimraf": "^3.0.2",
|
"rimraf": "^3.0.2",
|
||||||
@ -62,6 +66,7 @@
|
|||||||
"@types/jest": "^26.0.15",
|
"@types/jest": "^26.0.15",
|
||||||
"@types/js-yaml": "^4.0.0",
|
"@types/js-yaml": "^4.0.0",
|
||||||
"@types/node": "^14.14.6",
|
"@types/node": "^14.14.6",
|
||||||
|
"@types/pino-pretty": "^4.7.0",
|
||||||
"@types/supertest": "^2.0.10",
|
"@types/supertest": "^2.0.10",
|
||||||
"@typescript-eslint/eslint-plugin": "^4.6.1",
|
"@typescript-eslint/eslint-plugin": "^4.6.1",
|
||||||
"@typescript-eslint/parser": "^4.6.1",
|
"@typescript-eslint/parser": "^4.6.1",
|
||||||
|
@ -16,13 +16,32 @@ import { RawBodyMiddleware } from './commons/middlewares/raw-body.middleware';
|
|||||||
import { GiteaWebhooksController } from './webhooks/gitea-webhooks.controller';
|
import { GiteaWebhooksController } from './webhooks/gitea-webhooks.controller';
|
||||||
import { ParseBodyMiddleware } from './commons/middlewares/parse-body.middleware';
|
import { ParseBodyMiddleware } from './commons/middlewares/parse-body.middleware';
|
||||||
import { BullModule } from '@nestjs/bull';
|
import { BullModule } from '@nestjs/bull';
|
||||||
import { PubSubModule } from './commons/pub-sub/pub-sub.module';
|
import { LoggerModule } from 'nestjs-pino';
|
||||||
|
|
||||||
|
import pinoPretty from 'pino-pretty';
|
||||||
|
|
||||||
@Module({
|
@Module({
|
||||||
imports: [
|
imports: [
|
||||||
ConfigModule.forRoot({
|
ConfigModule.forRoot({
|
||||||
load: [configuration],
|
load: [configuration],
|
||||||
}),
|
}),
|
||||||
|
LoggerModule.forRootAsync({
|
||||||
|
imports: [ConfigModule],
|
||||||
|
useFactory: (configService: ConfigService) => {
|
||||||
|
const isDev = configService.get<'dev' | 'prod'>('env') === 'dev';
|
||||||
|
return {
|
||||||
|
pinoHttp: {
|
||||||
|
prettyPrint: isDev
|
||||||
|
? {
|
||||||
|
levelFirst: true,
|
||||||
|
}
|
||||||
|
: false,
|
||||||
|
prettifier: pinoPretty,
|
||||||
|
},
|
||||||
|
};
|
||||||
|
},
|
||||||
|
inject: [ConfigService],
|
||||||
|
}),
|
||||||
TypeOrmModule.forRootAsync({
|
TypeOrmModule.forRootAsync({
|
||||||
imports: [ConfigModule],
|
imports: [ConfigModule],
|
||||||
useFactory: (configService: ConfigService) => ({
|
useFactory: (configService: ConfigService) => ({
|
||||||
@ -58,17 +77,6 @@ import { PubSubModule } from './commons/pub-sub/pub-sub.module';
|
|||||||
}),
|
}),
|
||||||
inject: [ConfigService],
|
inject: [ConfigService],
|
||||||
}),
|
}),
|
||||||
PubSubModule.forRootAsync({
|
|
||||||
imports: [ConfigModule],
|
|
||||||
useFactory: (configService: ConfigService) => ({
|
|
||||||
redis: {
|
|
||||||
host: configService.get<string>('db.redis.host', 'localhost'),
|
|
||||||
port: configService.get<number>('db.redis.port', undefined),
|
|
||||||
password: configService.get<string>('db.redis.password', undefined),
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
inject: [ConfigService],
|
|
||||||
}),
|
|
||||||
ProjectsModule,
|
ProjectsModule,
|
||||||
ReposModule,
|
ReposModule,
|
||||||
PipelinesModule,
|
PipelinesModule,
|
||||||
|
@ -1,10 +1,10 @@
|
|||||||
import { Module } from '@nestjs/common';
|
import { Module } from '@nestjs/common';
|
||||||
import { PasswordConverter } from './services/password-converter';
|
import { PasswordConverter } from './services/password-converter';
|
||||||
import { PubSubModule } from './pub-sub/pub-sub.module';
|
import { RedisMutexModule } from './redis-mutex/redis-mutex.module';
|
||||||
|
|
||||||
@Module({
|
@Module({
|
||||||
providers: [PasswordConverter],
|
providers: [PasswordConverter],
|
||||||
exports: [PasswordConverter],
|
exports: [PasswordConverter, RedisMutexModule],
|
||||||
imports: [PubSubModule],
|
imports: [RedisMutexModule],
|
||||||
})
|
})
|
||||||
export class CommonsModule {}
|
export class CommonsModule {}
|
||||||
|
@ -1,11 +1,11 @@
|
|||||||
|
import { pick } from 'ramda';
|
||||||
|
|
||||||
export class ApplicationException extends Error {
|
export class ApplicationException extends Error {
|
||||||
code: number;
|
code: number;
|
||||||
error: Error;
|
error: Error;
|
||||||
|
|
||||||
constructor(
|
constructor(
|
||||||
message:
|
message: string | { error?: Error; message?: string | any; code?: number },
|
||||||
| string
|
|
||||||
| { error?: Error; message?: string | object; code?: number },
|
|
||||||
) {
|
) {
|
||||||
if (message instanceof Object) {
|
if (message instanceof Object) {
|
||||||
super();
|
super();
|
||||||
@ -18,4 +18,8 @@ export class ApplicationException extends Error {
|
|||||||
super((message as unknown) as any);
|
super((message as unknown) as any);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
toJSON() {
|
||||||
|
return pick(['code', 'message'], this);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
@ -1,5 +0,0 @@
|
|||||||
import { Inject } from '@nestjs/common';
|
|
||||||
import { getPubSubToken } from '../utils/token';
|
|
||||||
|
|
||||||
export const InjectPubSub = (name?: string): ParameterDecorator =>
|
|
||||||
Inject(getPubSubToken(name));
|
|
@ -1,8 +0,0 @@
|
|||||||
import { ModuleMetadata } from '@nestjs/common';
|
|
||||||
import { PubSubOptions } from './pub-sub-options.interface';
|
|
||||||
|
|
||||||
export interface PubSubAsyncConfig extends Pick<ModuleMetadata, 'imports'> {
|
|
||||||
name?: string;
|
|
||||||
useFactory: (...args: any[]) => Promise<PubSubOptions> | PubSubOptions;
|
|
||||||
inject?: any[];
|
|
||||||
}
|
|
@ -1,6 +0,0 @@
|
|||||||
import { RedisOptions } from 'ioredis';
|
|
||||||
|
|
||||||
export interface PubSubOptions {
|
|
||||||
name?: string;
|
|
||||||
redis: RedisOptions;
|
|
||||||
}
|
|
@ -1,15 +0,0 @@
|
|||||||
export type PubSubRawMessage<T> =
|
|
||||||
| PubSubRawNextMessage<T>
|
|
||||||
| PubSubRawErrorMessage<T>
|
|
||||||
| PubSubRawCompleteMessage<T>;
|
|
||||||
export interface PubSubRawNextMessage<T> {
|
|
||||||
type: 'next';
|
|
||||||
message: T;
|
|
||||||
}
|
|
||||||
export interface PubSubRawErrorMessage<T> {
|
|
||||||
type: 'error';
|
|
||||||
error: string;
|
|
||||||
}
|
|
||||||
export interface PubSubRawCompleteMessage<T> {
|
|
||||||
type: 'complete';
|
|
||||||
}
|
|
@ -1 +0,0 @@
|
|||||||
export const DEFAULT_PUB_SUB_NAME = 'default';
|
|
@ -1,48 +0,0 @@
|
|||||||
import { DynamicModule, Module } from '@nestjs/common';
|
|
||||||
import { PubSubService } from './pub-sub.service';
|
|
||||||
import {
|
|
||||||
createAsyncPubSubProviders,
|
|
||||||
createPubSubProvider,
|
|
||||||
} from './pub-sub.providers';
|
|
||||||
import { PubSubOptions } from './interfaces/pub-sub-options.interface';
|
|
||||||
import { PubSubAsyncConfig } from './interfaces/pub-sub-async-config.interface';
|
|
||||||
import { getPubSubConfigToken } from './utils/token';
|
|
||||||
|
|
||||||
@Module({
|
|
||||||
providers: [PubSubService],
|
|
||||||
})
|
|
||||||
export class PubSubModule {
|
|
||||||
public static forRoot(options: PubSubOptions): DynamicModule {
|
|
||||||
const providers = [
|
|
||||||
{
|
|
||||||
provide: getPubSubConfigToken(options.name),
|
|
||||||
useValue: options,
|
|
||||||
},
|
|
||||||
createPubSubProvider(options.name),
|
|
||||||
];
|
|
||||||
return {
|
|
||||||
global: true,
|
|
||||||
module: PubSubModule,
|
|
||||||
providers,
|
|
||||||
exports: providers,
|
|
||||||
};
|
|
||||||
}
|
|
||||||
public static forRootAsync(...configs: PubSubAsyncConfig[]): DynamicModule {
|
|
||||||
const providers = createAsyncPubSubProviders(configs);
|
|
||||||
return {
|
|
||||||
global: true,
|
|
||||||
module: PubSubModule,
|
|
||||||
imports: configs
|
|
||||||
.map((config) => config.imports)
|
|
||||||
.flat()
|
|
||||||
.filter((o, i, a) => a.indexOf(o) === i),
|
|
||||||
providers,
|
|
||||||
exports: providers,
|
|
||||||
};
|
|
||||||
}
|
|
||||||
public static forFeature(): DynamicModule {
|
|
||||||
return {
|
|
||||||
module: PubSubModule,
|
|
||||||
};
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,31 +0,0 @@
|
|||||||
import { Provider } from '@nestjs/common';
|
|
||||||
import { PubSubAsyncConfig } from './interfaces/pub-sub-async-config.interface';
|
|
||||||
import { PubSubOptions } from './interfaces/pub-sub-options.interface';
|
|
||||||
import { PubSub } from './pub-sub';
|
|
||||||
import { getPubSubConfigToken, getPubSubToken } from './utils/token';
|
|
||||||
|
|
||||||
export function createPubSubProvider(name: string): Provider {
|
|
||||||
return {
|
|
||||||
provide: getPubSubToken(name),
|
|
||||||
useFactory: (option: PubSubOptions) => new PubSub(option),
|
|
||||||
inject: [getPubSubConfigToken(name)],
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
export function createOptionsProvider(config: PubSubAsyncConfig): Provider {
|
|
||||||
return {
|
|
||||||
provide: getPubSubConfigToken(config.name),
|
|
||||||
useFactory: config.useFactory,
|
|
||||||
inject: config.inject || [],
|
|
||||||
};
|
|
||||||
}
|
|
||||||
export function createAsyncPubSubProviders(
|
|
||||||
configs: PubSubAsyncConfig[],
|
|
||||||
): Provider[] {
|
|
||||||
return configs
|
|
||||||
.map((config) => [
|
|
||||||
createOptionsProvider(config),
|
|
||||||
createPubSubProvider(config.name),
|
|
||||||
])
|
|
||||||
.flat();
|
|
||||||
}
|
|
@ -1,9 +0,0 @@
|
|||||||
import { Injectable } from '@nestjs/common';
|
|
||||||
import { PubSubOptions } from './interfaces/pub-sub-options.interface';
|
|
||||||
|
|
||||||
@Injectable()
|
|
||||||
export class PubSubService {
|
|
||||||
private options = new Map<string, PubSubOptions>();
|
|
||||||
private pubClient;
|
|
||||||
private;
|
|
||||||
}
|
|
@ -1,87 +0,0 @@
|
|||||||
import debug from 'debug';
|
|
||||||
import { tap } from 'rxjs/operators';
|
|
||||||
import { PubSub } from './pub-sub';
|
|
||||||
|
|
||||||
debug.enable('app:pubsub:*');
|
|
||||||
|
|
||||||
describe('PubSub', () => {
|
|
||||||
let instance: PubSub;
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
|
||||||
instance = new PubSub({
|
|
||||||
name: 'default',
|
|
||||||
redis: {
|
|
||||||
host: 'localhost',
|
|
||||||
},
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should be defined', () => {
|
|
||||||
expect(instance).toBeDefined();
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should can send and receive message', async () => {
|
|
||||||
const arr = new Array(10)
|
|
||||||
.fill(undefined)
|
|
||||||
.map(() => Math.random().toString(36).slice(2, 8));
|
|
||||||
const results: string[] = [];
|
|
||||||
instance
|
|
||||||
.message$<string>('test')
|
|
||||||
.pipe(
|
|
||||||
tap((val) => {
|
|
||||||
console.log(val);
|
|
||||||
}),
|
|
||||||
)
|
|
||||||
.subscribe((val) => results.push(val));
|
|
||||||
await new Promise((r) => setTimeout(r, 1000));
|
|
||||||
await Promise.all([...arr.map((str) => instance.publish('test', str))]);
|
|
||||||
await new Promise((r) => setTimeout(r, 1000));
|
|
||||||
expect(results).toMatchObject(arr);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should complete', async () => {
|
|
||||||
const arr = new Array(10)
|
|
||||||
.fill(undefined)
|
|
||||||
.map(() => Math.random().toString(36).slice(2, 8));
|
|
||||||
const results: string[] = [];
|
|
||||||
instance
|
|
||||||
.message$<string>('test')
|
|
||||||
.pipe(
|
|
||||||
tap((val) => {
|
|
||||||
console.log(val);
|
|
||||||
}),
|
|
||||||
)
|
|
||||||
.subscribe((val) => results.push(val));
|
|
||||||
await new Promise((r) => setTimeout(r, 1000));
|
|
||||||
await Promise.all([...arr.map((str) => instance.publish('test', str))]);
|
|
||||||
await instance.finish('test');
|
|
||||||
await Promise.all([...arr.map((str) => instance.publish('test', str))]);
|
|
||||||
await new Promise((r) => setTimeout(r, 1000));
|
|
||||||
expect(results).toMatchObject(arr);
|
|
||||||
});
|
|
||||||
it('should error', async () => {
|
|
||||||
const arr = new Array(10)
|
|
||||||
.fill(undefined)
|
|
||||||
.map(() => Math.random().toString(36).slice(2, 8));
|
|
||||||
const results: string[] = [];
|
|
||||||
let error: string;
|
|
||||||
instance
|
|
||||||
.message$<string>('test')
|
|
||||||
.pipe(
|
|
||||||
tap((val) => {
|
|
||||||
console.log(val);
|
|
||||||
}),
|
|
||||||
)
|
|
||||||
.subscribe({
|
|
||||||
next: (val) => results.push(val),
|
|
||||||
error: (err) => (error = err.message),
|
|
||||||
});
|
|
||||||
await new Promise((r) => setTimeout(r, 1000));
|
|
||||||
await Promise.all([...arr.map((str) => instance.publish('test', str))]);
|
|
||||||
await instance.throwError('test', 'TEST ERROR MESSAGE');
|
|
||||||
await Promise.all([...arr.map((str) => instance.publish('test', str))]);
|
|
||||||
await new Promise((r) => setTimeout(r, 1000));
|
|
||||||
expect(results).toMatchObject(arr);
|
|
||||||
expect(error).toEqual('TEST ERROR MESSAGE');
|
|
||||||
});
|
|
||||||
});
|
|
@ -1,115 +0,0 @@
|
|||||||
import debug from 'debug';
|
|
||||||
import { EventEmitter } from 'events';
|
|
||||||
import IORedis, { Redis } from 'ioredis';
|
|
||||||
import { from, fromEvent, Observable } from 'rxjs';
|
|
||||||
import { filter, map, share, switchMap, takeWhile, tap } from 'rxjs/operators';
|
|
||||||
import { ApplicationException } from '../exceptions/application.exception';
|
|
||||||
import { PubSubOptions } from './interfaces/pub-sub-options.interface';
|
|
||||||
import {
|
|
||||||
PubSubRawMessage,
|
|
||||||
PubSubRawNextMessage,
|
|
||||||
} from './interfaces/pub-sub-raw-message.interface';
|
|
||||||
|
|
||||||
const log = debug('fennec:pubsub:instance');
|
|
||||||
export class PubSub extends EventEmitter {
|
|
||||||
pubRedis: Redis;
|
|
||||||
pSubRedis: Redis;
|
|
||||||
channels: string[] = [];
|
|
||||||
event$: Observable<[string, string, any]>;
|
|
||||||
|
|
||||||
constructor(private readonly options: PubSubOptions) {
|
|
||||||
super();
|
|
||||||
this.pubRedis = new IORedis(this.options.redis);
|
|
||||||
this.pSubRedis = new IORedis(this.options.redis);
|
|
||||||
|
|
||||||
this.pSubRedis.on('pmessage', (...args) =>
|
|
||||||
log.extend('raw')('%s %s %o', ...args),
|
|
||||||
);
|
|
||||||
|
|
||||||
this.event$ = fromEvent<[string, string, string]>(
|
|
||||||
this.pSubRedis,
|
|
||||||
'pmessage',
|
|
||||||
).pipe(
|
|
||||||
map((ev) => {
|
|
||||||
try {
|
|
||||||
ev[2] = JSON.parse(ev[2]);
|
|
||||||
} catch (err) {
|
|
||||||
log('WARN: is not json');
|
|
||||||
return null;
|
|
||||||
}
|
|
||||||
log('on message: %s %s %o', ...ev);
|
|
||||||
return ev;
|
|
||||||
}),
|
|
||||||
filter((v) => !!v),
|
|
||||||
share(),
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
async disconnect() {
|
|
||||||
log('disconnecting to redis...');
|
|
||||||
this.pubRedis.disconnect();
|
|
||||||
this.pSubRedis.disconnect();
|
|
||||||
log('disconnected');
|
|
||||||
}
|
|
||||||
|
|
||||||
async registerChannel(channel: string) {
|
|
||||||
if (this.channels.includes(channel)) {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
this.channels.push(channel);
|
|
||||||
|
|
||||||
return await this.pSubRedis.psubscribe(channel);
|
|
||||||
}
|
|
||||||
|
|
||||||
private async redisPublish<T>(channel: string, message: PubSubRawMessage<T>) {
|
|
||||||
log.extend('publish')('channel: %s, message: %O', channel, message);
|
|
||||||
return await this.pubRedis.publish(channel, JSON.stringify(message));
|
|
||||||
}
|
|
||||||
|
|
||||||
async publish(channel: string, message: any): Promise<number> {
|
|
||||||
return await this.redisPublish(channel, {
|
|
||||||
type: 'next',
|
|
||||||
message,
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
async finish(channel: string): Promise<number> {
|
|
||||||
return await this.redisPublish(channel, {
|
|
||||||
type: 'complete',
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
async throwError(channel: string, error: string): Promise<number> {
|
|
||||||
return await this.redisPublish(channel, {
|
|
||||||
type: 'error',
|
|
||||||
error,
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
message$ = <T>(channel: string): Observable<T> => {
|
|
||||||
return from(this.registerChannel(channel))
|
|
||||||
.pipe(switchMap(() => this.event$))
|
|
||||||
.pipe(
|
|
||||||
filter(([pattern]) => pattern === channel),
|
|
||||||
tap(([pattern, channel, message]) => {
|
|
||||||
log.extend('subscribe')(
|
|
||||||
'channel: %s, match: %s, message: %O',
|
|
||||||
channel,
|
|
||||||
pattern,
|
|
||||||
message,
|
|
||||||
);
|
|
||||||
}),
|
|
||||||
takeWhile(([pattern, channel, message]) => {
|
|
||||||
if (pattern === channel) {
|
|
||||||
if (message.type === 'error') {
|
|
||||||
throw new ApplicationException(message.error);
|
|
||||||
}
|
|
||||||
return message.type !== 'complete';
|
|
||||||
}
|
|
||||||
return true;
|
|
||||||
}),
|
|
||||||
map((ev) => ev[2] as PubSubRawMessage<T>),
|
|
||||||
map((message: PubSubRawNextMessage<T>) => message.message),
|
|
||||||
);
|
|
||||||
};
|
|
||||||
}
|
|
@ -1,8 +0,0 @@
|
|||||||
import { DEFAULT_PUB_SUB_NAME } from '../pub-sub.constants';
|
|
||||||
|
|
||||||
export function getPubSubToken(name = DEFAULT_PUB_SUB_NAME) {
|
|
||||||
return `app:pub-usb:${name || DEFAULT_PUB_SUB_NAME}`;
|
|
||||||
}
|
|
||||||
export function getPubSubConfigToken(name = DEFAULT_PUB_SUB_NAME) {
|
|
||||||
return `app:pub-usb:config:${name || DEFAULT_PUB_SUB_NAME}`;
|
|
||||||
}
|
|
10
src/commons/redis-mutex/redis-mutex.module.ts
Normal file
10
src/commons/redis-mutex/redis-mutex.module.ts
Normal file
@ -0,0 +1,10 @@
|
|||||||
|
import { Module } from '@nestjs/common';
|
||||||
|
import { RedisMutexService } from './redis-mutex.service';
|
||||||
|
import { RedisModule } from 'nestjs-redis';
|
||||||
|
|
||||||
|
@Module({
|
||||||
|
imports: [RedisModule],
|
||||||
|
providers: [RedisMutexService],
|
||||||
|
exports: [RedisMutexService],
|
||||||
|
})
|
||||||
|
export class RedisMutexModule {}
|
@ -1,15 +1,15 @@
|
|||||||
import { Test, TestingModule } from '@nestjs/testing';
|
import { Test, TestingModule } from '@nestjs/testing';
|
||||||
import { PubSubService } from './pub-sub.service';
|
import { RedisMutexService } from './redis-mutex.service';
|
||||||
|
|
||||||
describe('PubsubService', () => {
|
describe('RedisMutexService', () => {
|
||||||
let service: PubSubService;
|
let service: RedisMutexService;
|
||||||
|
|
||||||
beforeEach(async () => {
|
beforeEach(async () => {
|
||||||
const module: TestingModule = await Test.createTestingModule({
|
const module: TestingModule = await Test.createTestingModule({
|
||||||
providers: [PubSubService],
|
providers: [RedisMutexService],
|
||||||
}).compile();
|
}).compile();
|
||||||
|
|
||||||
service = module.get<PubSubService>(PubSubService);
|
service = module.get<RedisMutexService>(RedisMutexService);
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should be defined', () => {
|
it('should be defined', () => {
|
71
src/commons/redis-mutex/redis-mutex.service.ts
Normal file
71
src/commons/redis-mutex/redis-mutex.service.ts
Normal file
@ -0,0 +1,71 @@
|
|||||||
|
import { Injectable } from '@nestjs/common';
|
||||||
|
import { RedisService } from 'nestjs-redis';
|
||||||
|
import * as uuid from 'uuid';
|
||||||
|
import { ApplicationException } from '../exceptions/application.exception';
|
||||||
|
|
||||||
|
export interface RedisMutexOption {
|
||||||
|
/**
|
||||||
|
* seconds
|
||||||
|
*/
|
||||||
|
expires?: number;
|
||||||
|
/**
|
||||||
|
* seconds
|
||||||
|
*/
|
||||||
|
timeout?: number | null;
|
||||||
|
/**
|
||||||
|
* milliseconds
|
||||||
|
*/
|
||||||
|
retryDelay?: number;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Injectable()
|
||||||
|
export class RedisMutexService {
|
||||||
|
constructor(private readonly redisClient: RedisService) {}
|
||||||
|
|
||||||
|
public async lock(
|
||||||
|
key: string,
|
||||||
|
{ expires = 100, timeout = 10, retryDelay = 100 }: RedisMutexOption = {
|
||||||
|
expires: 100,
|
||||||
|
timeout: 10,
|
||||||
|
retryDelay: 100,
|
||||||
|
},
|
||||||
|
) {
|
||||||
|
const redisKey = `${'mutex-lock'}:${key}`;
|
||||||
|
const redis = this.redisClient.getClient();
|
||||||
|
const value = uuid.v4();
|
||||||
|
const timeoutAt = timeout ? Date.now() + timeout * 1000 : null;
|
||||||
|
|
||||||
|
while (
|
||||||
|
!(await redis
|
||||||
|
.set(redisKey, value, 'EX', expires, 'NX')
|
||||||
|
.then(() => true)
|
||||||
|
.catch(() => false))
|
||||||
|
) {
|
||||||
|
if (timeoutAt && timeoutAt > Date.now()) {
|
||||||
|
throw new ApplicationException('lock timeout');
|
||||||
|
}
|
||||||
|
await new Promise((resolve) => setTimeout(resolve, retryDelay));
|
||||||
|
}
|
||||||
|
|
||||||
|
const renewTimer = setInterval(() => {
|
||||||
|
redis.expire(redisKey, expires);
|
||||||
|
}, (expires * 1000) / 2);
|
||||||
|
|
||||||
|
return async () => {
|
||||||
|
clearInterval(renewTimer);
|
||||||
|
await redis.eval(
|
||||||
|
`
|
||||||
|
if redis.call("get", KEYS[1]) == ARGV[1]
|
||||||
|
then
|
||||||
|
return redis.call("del", KEYS[1])
|
||||||
|
else
|
||||||
|
return 0
|
||||||
|
end
|
||||||
|
`,
|
||||||
|
1,
|
||||||
|
redisKey,
|
||||||
|
value,
|
||||||
|
);
|
||||||
|
};
|
||||||
|
}
|
||||||
|
}
|
@ -53,8 +53,22 @@ export class BaseDbService<Entity extends AppBaseEntity> extends TypeormHelper {
|
|||||||
async isDuplicateEntityForUpdate<Dto extends Entity>(
|
async isDuplicateEntityForUpdate<Dto extends Entity>(
|
||||||
id: string,
|
id: string,
|
||||||
dto: Partial<Dto>,
|
dto: Partial<Dto>,
|
||||||
|
extendsFields?: Array<keyof Dto & string>,
|
||||||
|
): Promise<false | never>;
|
||||||
|
async isDuplicateEntityForUpdate<Dto extends Entity>(
|
||||||
|
old: Entity,
|
||||||
|
dto: Partial<Dto>,
|
||||||
|
extendsFields?: Array<keyof Dto & string>,
|
||||||
|
): Promise<false | never>;
|
||||||
|
async isDuplicateEntityForUpdate<Dto extends Entity>(
|
||||||
|
id: string | Entity,
|
||||||
|
dto: Partial<Dto>,
|
||||||
extendsFields: Array<keyof Dto & string> = [],
|
extendsFields: Array<keyof Dto & string> = [],
|
||||||
): Promise<false | never> {
|
): Promise<false | never> {
|
||||||
|
if (typeof id !== 'string') {
|
||||||
|
dto = Object.assign({}, id, dto);
|
||||||
|
id = id.id;
|
||||||
|
}
|
||||||
const qb = this.repository.createQueryBuilder('entity');
|
const qb = this.repository.createQueryBuilder('entity');
|
||||||
const compareFields = this.getCompareFields(dto, [
|
const compareFields = this.getCompareFields(dto, [
|
||||||
...this.uniqueFields,
|
...this.uniqueFields,
|
||||||
|
21
src/commons/utils/rabbit-mq.ts
Normal file
21
src/commons/utils/rabbit-mq.ts
Normal file
@ -0,0 +1,21 @@
|
|||||||
|
import { hostname } from 'os';
|
||||||
|
|
||||||
|
export function getInstanceName() {
|
||||||
|
return hostname();
|
||||||
|
}
|
||||||
|
|
||||||
|
export function getSelfInstanceRouteKey(key: string) {
|
||||||
|
return getAppInstanceRouteKey(key, getInstanceName());
|
||||||
|
}
|
||||||
|
|
||||||
|
export function getAppInstanceRouteKey(key: string, appInstance?: string) {
|
||||||
|
return appInstance ? `${key}.${appInstance}` : key;
|
||||||
|
}
|
||||||
|
|
||||||
|
export function getSelfInstanceQueueKey(key: string) {
|
||||||
|
return getAppInstanceQueueKey(key, getInstanceName());
|
||||||
|
}
|
||||||
|
|
||||||
|
export function getAppInstanceQueueKey(key: string, appInstance?: string) {
|
||||||
|
return appInstance ? `${key}.${appInstance}` : key;
|
||||||
|
}
|
@ -11,3 +11,5 @@ registerEnumType(TaskStatuses, {
|
|||||||
name: 'TaskStatuses',
|
name: 'TaskStatuses',
|
||||||
description: '任务状态',
|
description: '任务状态',
|
||||||
});
|
});
|
||||||
|
|
||||||
|
export const terminalTaskStatuses = [TaskStatuses.success, TaskStatuses.failed];
|
||||||
|
25
src/pipeline-tasks/models/pipeline-task-event.ts
Normal file
25
src/pipeline-tasks/models/pipeline-task-event.ts
Normal file
@ -0,0 +1,25 @@
|
|||||||
|
import { Field, ObjectType } from '@nestjs/graphql';
|
||||||
|
import { PipelineUnits } from '../enums/pipeline-units.enum';
|
||||||
|
import { TaskStatuses } from '../enums/task-statuses.enum';
|
||||||
|
import { Type } from 'class-transformer';
|
||||||
|
|
||||||
|
@ObjectType()
|
||||||
|
export class PipelineTaskEvent {
|
||||||
|
@Field()
|
||||||
|
taskId: string;
|
||||||
|
@Field()
|
||||||
|
pipelineId: string;
|
||||||
|
@Field()
|
||||||
|
projectId: string;
|
||||||
|
@Field(() => PipelineUnits, { nullable: true })
|
||||||
|
unit: PipelineUnits | null;
|
||||||
|
@Field()
|
||||||
|
@Type(() => Date)
|
||||||
|
emittedAt: Date;
|
||||||
|
@Field()
|
||||||
|
message: string;
|
||||||
|
@Field()
|
||||||
|
messageType: 'stdout' | 'stderr' | 'stdin';
|
||||||
|
@Field(() => TaskStatuses)
|
||||||
|
status: TaskStatuses;
|
||||||
|
}
|
@ -1,9 +1,10 @@
|
|||||||
import { InputType, ObjectType } from '@nestjs/graphql';
|
import { Field, InputType, Int, ObjectType } from '@nestjs/graphql';
|
||||||
import { WorkUnit } from './work-unit.model';
|
import { WorkUnit } from './work-unit.model';
|
||||||
|
|
||||||
@InputType('WorkUnitMetadataInput')
|
@InputType('WorkUnitMetadataInput')
|
||||||
@ObjectType()
|
@ObjectType()
|
||||||
export class WorkUnitMetadata {
|
export class WorkUnitMetadata {
|
||||||
|
@Field(() => Int)
|
||||||
version = 1;
|
version = 1;
|
||||||
units: WorkUnit[];
|
units: WorkUnit[];
|
||||||
}
|
}
|
||||||
|
88
src/pipeline-tasks/pipeline-task-flush.service.spec.ts
Normal file
88
src/pipeline-tasks/pipeline-task-flush.service.spec.ts
Normal file
@ -0,0 +1,88 @@
|
|||||||
|
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
|
||||||
|
import { Test, TestingModule } from '@nestjs/testing';
|
||||||
|
import { RedisService } from 'nestjs-redis';
|
||||||
|
import { PipelineTaskFlushService } from './pipeline-task-flush.service';
|
||||||
|
import { PipelineTaskEvent } from './models/pipeline-task-event';
|
||||||
|
import { TaskStatuses } from './enums/task-statuses.enum';
|
||||||
|
import {
|
||||||
|
EXCHANGE_PIPELINE_TASK_TOPIC,
|
||||||
|
ROUTE_PIPELINE_TASK_DONE,
|
||||||
|
} from './pipeline-tasks.constants';
|
||||||
|
|
||||||
|
describe('PipelineTaskFlushService', () => {
|
||||||
|
let service: PipelineTaskFlushService;
|
||||||
|
let redisService: RedisService;
|
||||||
|
let amqpConnection: AmqpConnection;
|
||||||
|
|
||||||
|
beforeEach(async () => {
|
||||||
|
const redisClient = {
|
||||||
|
rpush: jest.fn(() => Promise.resolve()),
|
||||||
|
lrange: jest.fn(() => Promise.resolve()),
|
||||||
|
expire: jest.fn(() => Promise.resolve()),
|
||||||
|
};
|
||||||
|
const module: TestingModule = await Test.createTestingModule({
|
||||||
|
providers: [
|
||||||
|
PipelineTaskFlushService,
|
||||||
|
{
|
||||||
|
provide: RedisService,
|
||||||
|
useValue: {
|
||||||
|
getClient() {
|
||||||
|
return redisClient;
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
provide: AmqpConnection,
|
||||||
|
useValue: {
|
||||||
|
request: jest.fn(() => Promise.resolve()),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
],
|
||||||
|
}).compile();
|
||||||
|
|
||||||
|
service = module.get<PipelineTaskFlushService>(PipelineTaskFlushService);
|
||||||
|
redisService = module.get<RedisService>(RedisService);
|
||||||
|
amqpConnection = module.get<AmqpConnection>(AmqpConnection);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should be defined', () => {
|
||||||
|
expect(service).toBeDefined();
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('write', () => {
|
||||||
|
const amqpMsg = {
|
||||||
|
properties: { headers: { sender: 'test' } },
|
||||||
|
} as any;
|
||||||
|
it('normal', async () => {
|
||||||
|
const testEvent = new PipelineTaskEvent();
|
||||||
|
testEvent.taskId = 'test';
|
||||||
|
testEvent.status = TaskStatuses.working;
|
||||||
|
const rpush = jest.spyOn(redisService.getClient(), 'rpush');
|
||||||
|
const request = jest.spyOn(amqpConnection, 'request');
|
||||||
|
await service.write(testEvent, amqpMsg);
|
||||||
|
expect(rpush).toBeCalledTimes(1);
|
||||||
|
expect(rpush.mock.calls[0][0]).toEqual('p-task:log:test');
|
||||||
|
expect(rpush.mock.calls[0][1]).toEqual(JSON.stringify(testEvent));
|
||||||
|
expect(request).toBeCalledTimes(1);
|
||||||
|
});
|
||||||
|
it('event for which task done', async () => {
|
||||||
|
const testEvent = new PipelineTaskEvent();
|
||||||
|
testEvent.taskId = 'test';
|
||||||
|
testEvent.status = TaskStatuses.success;
|
||||||
|
const rpush = jest.spyOn(redisService.getClient(), 'rpush');
|
||||||
|
const request = jest.spyOn(amqpConnection, 'request');
|
||||||
|
await service.write(testEvent, amqpMsg);
|
||||||
|
expect(rpush).toBeCalledTimes(1);
|
||||||
|
expect(request).toBeCalledTimes(1);
|
||||||
|
expect(request.mock.calls[0][0]).toMatchObject({
|
||||||
|
exchange: EXCHANGE_PIPELINE_TASK_TOPIC,
|
||||||
|
routingKey: ROUTE_PIPELINE_TASK_DONE,
|
||||||
|
payload: {
|
||||||
|
taskId: 'test',
|
||||||
|
status: TaskStatuses.success,
|
||||||
|
runOn: 'test',
|
||||||
|
},
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
66
src/pipeline-tasks/pipeline-task-flush.service.ts
Normal file
66
src/pipeline-tasks/pipeline-task-flush.service.ts
Normal file
@ -0,0 +1,66 @@
|
|||||||
|
import { AmqpConnection, RabbitSubscribe } from '@golevelup/nestjs-rabbitmq';
|
||||||
|
import { Injectable } from '@nestjs/common';
|
||||||
|
import { ConsumeMessage } from 'amqplib';
|
||||||
|
import { deserialize } from 'class-transformer';
|
||||||
|
import { RedisService } from 'nestjs-redis';
|
||||||
|
import { isNil } from 'ramda';
|
||||||
|
import { getSelfInstanceQueueKey } from '../commons/utils/rabbit-mq';
|
||||||
|
import { PipelineTaskEvent } from './models/pipeline-task-event';
|
||||||
|
import {
|
||||||
|
EXCHANGE_PIPELINE_TASK_TOPIC,
|
||||||
|
ROUTE_PIPELINE_TASK_DONE,
|
||||||
|
} from './pipeline-tasks.constants';
|
||||||
|
import {
|
||||||
|
EXCHANGE_PIPELINE_TASK_FANOUT,
|
||||||
|
ROUTE_PIPELINE_TASK_LOG,
|
||||||
|
QUEUE_WRITE_PIPELINE_TASK_LOG,
|
||||||
|
} from './pipeline-tasks.constants';
|
||||||
|
|
||||||
|
@Injectable()
|
||||||
|
export class PipelineTaskFlushService {
|
||||||
|
constructor(
|
||||||
|
private readonly redisService: RedisService,
|
||||||
|
private readonly amqpConnection: AmqpConnection,
|
||||||
|
) {}
|
||||||
|
|
||||||
|
@RabbitSubscribe({
|
||||||
|
exchange: EXCHANGE_PIPELINE_TASK_FANOUT,
|
||||||
|
routingKey: ROUTE_PIPELINE_TASK_LOG,
|
||||||
|
queue: getSelfInstanceQueueKey(QUEUE_WRITE_PIPELINE_TASK_LOG),
|
||||||
|
queueOptions: {
|
||||||
|
autoDelete: true,
|
||||||
|
durable: true,
|
||||||
|
},
|
||||||
|
})
|
||||||
|
async write(message: PipelineTaskEvent, amqpMsg: ConsumeMessage) {
|
||||||
|
const client = this.redisService.getClient();
|
||||||
|
await client.rpush(this.getKey(message.taskId), JSON.stringify(message));
|
||||||
|
await client.expire(this.getKey(message.taskId), 600); // ten minutes
|
||||||
|
if (isNil(message.unit)) {
|
||||||
|
try {
|
||||||
|
await this.amqpConnection.request({
|
||||||
|
exchange: EXCHANGE_PIPELINE_TASK_TOPIC,
|
||||||
|
routingKey: ROUTE_PIPELINE_TASK_DONE,
|
||||||
|
payload: {
|
||||||
|
taskId: message.taskId,
|
||||||
|
status: message.status,
|
||||||
|
runOn: amqpMsg.properties.headers.sender,
|
||||||
|
},
|
||||||
|
});
|
||||||
|
} catch (error) {
|
||||||
|
console.log(error);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async read(taskId: string) {
|
||||||
|
const raw = await this.redisService
|
||||||
|
.getClient()
|
||||||
|
.lrange(this.getKey(taskId), 0, -1);
|
||||||
|
return raw.map((it) => deserialize(PipelineTaskEvent, it));
|
||||||
|
}
|
||||||
|
|
||||||
|
private getKey(taskId: string) {
|
||||||
|
return `p-task:log:${taskId}`;
|
||||||
|
}
|
||||||
|
}
|
@ -1,30 +0,0 @@
|
|||||||
import { Test, TestingModule } from '@nestjs/testing';
|
|
||||||
import { PipelineTaskLogsService } from './pipeline-task-logs.service';
|
|
||||||
import { RedisService } from 'nestjs-redis';
|
|
||||||
import { getPubSubToken } from '../commons/pub-sub/utils/token';
|
|
||||||
|
|
||||||
describe('PipelineTaskLogsService', () => {
|
|
||||||
let service: PipelineTaskLogsService;
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
|
||||||
const module: TestingModule = await Test.createTestingModule({
|
|
||||||
providers: [
|
|
||||||
PipelineTaskLogsService,
|
|
||||||
{
|
|
||||||
provide: RedisService,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: getPubSubToken(),
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
],
|
|
||||||
}).compile();
|
|
||||||
|
|
||||||
service = module.get<PipelineTaskLogsService>(PipelineTaskLogsService);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should be defined', () => {
|
|
||||||
expect(service).toBeDefined();
|
|
||||||
});
|
|
||||||
});
|
|
@ -1,81 +0,0 @@
|
|||||||
import { Injectable } from '@nestjs/common';
|
|
||||||
import { observableToAsyncIterable } from 'graphql-tools';
|
|
||||||
import { RedisService } from 'nestjs-redis';
|
|
||||||
import { find, omit, propEq } from 'ramda';
|
|
||||||
import { InjectPubSub } from '../commons/pub-sub/decorators/inject-pub-sub.decorator';
|
|
||||||
import { PubSub } from '../commons/pub-sub/pub-sub';
|
|
||||||
import { PipelineUnits } from './enums/pipeline-units.enum';
|
|
||||||
import { TaskStatuses } from './enums/task-statuses.enum';
|
|
||||||
import { PipelineTaskLogMessage } from './models/pipeline-task-log-message.module';
|
|
||||||
import { PipelineTaskLogs } from './models/pipeline-task-logs.model';
|
|
||||||
import { PipelineTask } from './pipeline-task.entity';
|
|
||||||
|
|
||||||
const LOG_TIMEOUT_SECONDS = 10_000;
|
|
||||||
|
|
||||||
@Injectable()
|
|
||||||
export class PipelineTaskLogsService {
|
|
||||||
constructor(
|
|
||||||
private readonly redisService: RedisService,
|
|
||||||
@InjectPubSub()
|
|
||||||
private readonly pubSub: PubSub,
|
|
||||||
) {}
|
|
||||||
|
|
||||||
get redis() {
|
|
||||||
return this.redisService.getClient();
|
|
||||||
}
|
|
||||||
|
|
||||||
getKeys(task: PipelineTask) {
|
|
||||||
return `ptl:${task.id}`;
|
|
||||||
}
|
|
||||||
|
|
||||||
async recordLog(log: PipelineTaskLogMessage) {
|
|
||||||
const logDto = omit(['task'], log);
|
|
||||||
await Promise.all([
|
|
||||||
this.pubSub.publish(this.getKeys(log.task), logDto),
|
|
||||||
this.redis
|
|
||||||
.expire(this.getKeys(log.task), LOG_TIMEOUT_SECONDS)
|
|
||||||
.then(() =>
|
|
||||||
this.redis.rpush(this.getKeys(log.task), JSON.stringify(logDto)),
|
|
||||||
),
|
|
||||||
]);
|
|
||||||
}
|
|
||||||
|
|
||||||
async readLog(task: PipelineTask): Promise<PipelineTaskLogMessage[]> {
|
|
||||||
return await this.redis.lrange(this.getKeys(task), 0, -1).then((items) =>
|
|
||||||
items.map((item) => {
|
|
||||||
const log = JSON.parse(item) as PipelineTaskLogMessage;
|
|
||||||
log.task = task;
|
|
||||||
log.time = new Date(log.time);
|
|
||||||
return log;
|
|
||||||
}),
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
async readLogsAsPipelineTaskLogs(
|
|
||||||
task: PipelineTask,
|
|
||||||
): Promise<PipelineTaskLogs[]> {
|
|
||||||
const logs = await this.readLog(task);
|
|
||||||
const taskLogs: PipelineTaskLogs[] = [];
|
|
||||||
for (const log of logs) {
|
|
||||||
const taskLog = find<PipelineTaskLogs>(
|
|
||||||
propEq('unit', log.unit),
|
|
||||||
taskLogs,
|
|
||||||
);
|
|
||||||
if (!taskLog) {
|
|
||||||
taskLogs.push({
|
|
||||||
unit: (log.unit as unknown) as PipelineUnits,
|
|
||||||
status: TaskStatuses.working,
|
|
||||||
startedAt: log.time,
|
|
||||||
logs: log.message,
|
|
||||||
});
|
|
||||||
} else {
|
|
||||||
taskLog.logs += log.message;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return taskLogs;
|
|
||||||
}
|
|
||||||
|
|
||||||
watchLogs(task: PipelineTask) {
|
|
||||||
return observableToAsyncIterable(this.pubSub.message$(this.getKeys(task)));
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,240 +0,0 @@
|
|||||||
import { Test, TestingModule } from '@nestjs/testing';
|
|
||||||
import { Job } from 'bull';
|
|
||||||
import { join } from 'path';
|
|
||||||
import { ReposService } from '../repos/repos.service';
|
|
||||||
import { PipelineUnits } from './enums/pipeline-units.enum';
|
|
||||||
import { PipelineTaskConsumer } from './pipeline-task.consumer';
|
|
||||||
import { PipelineTask } from './pipeline-task.entity';
|
|
||||||
import { PipelineTasksService } from './pipeline-tasks.service';
|
|
||||||
import { PipelineTaskLogMessage } from './models/pipeline-task-log-message.module';
|
|
||||||
import { Pipeline } from '../pipelines/pipeline.entity';
|
|
||||||
import { Project } from '../projects/project.entity';
|
|
||||||
import { TaskStatuses } from './enums/task-statuses.enum';
|
|
||||||
import { PipelineTaskLogsService } from './pipeline-task-logs.service';
|
|
||||||
import { ApplicationException } from '../commons/exceptions/application.exception';
|
|
||||||
|
|
||||||
describe('PipelineTaskConsumer', () => {
|
|
||||||
let consumer: PipelineTaskConsumer;
|
|
||||||
let tasksService: PipelineTasksService;
|
|
||||||
let logsService: PipelineTaskLogsService;
|
|
||||||
const getJob = () =>
|
|
||||||
({
|
|
||||||
data: {
|
|
||||||
pipelineId: 'test',
|
|
||||||
units: [PipelineUnits.checkout, PipelineUnits.test],
|
|
||||||
},
|
|
||||||
} as Job<PipelineTask>);
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
|
||||||
const module: TestingModule = await Test.createTestingModule({
|
|
||||||
providers: [
|
|
||||||
{
|
|
||||||
provide: PipelineTasksService,
|
|
||||||
useValue: {
|
|
||||||
doNextTask: () => undefined,
|
|
||||||
updateTask: async (value) => value,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: ReposService,
|
|
||||||
useValue: {
|
|
||||||
getWorkspaceRootByTask: () => 'workspace-root',
|
|
||||||
checkout: async () => undefined,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: PipelineTaskLogsService,
|
|
||||||
useValue: {
|
|
||||||
recordLog: async () => undefined,
|
|
||||||
readLogsAsPipelineTaskLogs: async () => [],
|
|
||||||
},
|
|
||||||
},
|
|
||||||
PipelineTaskConsumer,
|
|
||||||
],
|
|
||||||
}).compile();
|
|
||||||
|
|
||||||
tasksService = module.get(PipelineTasksService);
|
|
||||||
logsService = module.get(PipelineTaskLogsService);
|
|
||||||
consumer = module.get(PipelineTaskConsumer);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should be defined', () => {
|
|
||||||
expect(consumer).toBeDefined();
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('onCompleted', () => {
|
|
||||||
it('should call doNextTask()', () => {
|
|
||||||
const job = getJob();
|
|
||||||
const doNextTask = jest.spyOn(tasksService, 'doNextTask');
|
|
||||||
consumer.onCompleted(job);
|
|
||||||
expect(doNextTask).toHaveBeenCalledTimes(1);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('runScript', () => {
|
|
||||||
let logText: string;
|
|
||||||
let errorText: string;
|
|
||||||
let recordLog: jest.SpyInstance;
|
|
||||||
beforeEach(() => {
|
|
||||||
logText = '';
|
|
||||||
errorText = '';
|
|
||||||
recordLog = jest
|
|
||||||
.spyOn(logsService, 'recordLog')
|
|
||||||
.mockImplementation(async (log: PipelineTaskLogMessage) => {
|
|
||||||
logText += log.message;
|
|
||||||
if (log.isError) {
|
|
||||||
errorText += log.message;
|
|
||||||
}
|
|
||||||
});
|
|
||||||
});
|
|
||||||
it('should success and log right message', async () => {
|
|
||||||
await consumer.runScript(
|
|
||||||
'node one-second-work.js',
|
|
||||||
join(__dirname, '../../test/data'),
|
|
||||||
);
|
|
||||||
expect(logText).toMatch(/10.+20.+30.+40.+50.+60.+70.+80.+90/s);
|
|
||||||
expect(recordLog).toHaveBeenCalledTimes(10);
|
|
||||||
expect(
|
|
||||||
((recordLog.mock.calls[8][0] as unknown) as PipelineTaskLogMessage)
|
|
||||||
.message,
|
|
||||||
).toMatch(/^90/);
|
|
||||||
});
|
|
||||||
it('should failed and log right message', async () => {
|
|
||||||
await expect(
|
|
||||||
consumer.runScript(
|
|
||||||
'node bad-work.js',
|
|
||||||
join(__dirname, '../../test/data'),
|
|
||||||
),
|
|
||||||
).rejects.toThrowError(/exec script failed/);
|
|
||||||
expect(errorText).toMatch(/Error Message/);
|
|
||||||
const logs = recordLog.mock.calls
|
|
||||||
.map((call) => ((call[0] as unknown) as PipelineTaskLogMessage).message)
|
|
||||||
.join('');
|
|
||||||
expect(logs).toMatch(/10.+20.+30.+40.+50/s);
|
|
||||||
});
|
|
||||||
it('should log with task', async () => {
|
|
||||||
const task = new PipelineTask();
|
|
||||||
task.id = 'test';
|
|
||||||
|
|
||||||
const recordLog = jest.spyOn(logsService, 'recordLog');
|
|
||||||
await expect(
|
|
||||||
consumer.runScript(
|
|
||||||
'node bad-work.js',
|
|
||||||
join(__dirname, '../../test/data'),
|
|
||||||
task,
|
|
||||||
),
|
|
||||||
).rejects.toThrowError(/exec script failed/);
|
|
||||||
|
|
||||||
expect(errorText).toMatch(/Error Message 2/);
|
|
||||||
expect(
|
|
||||||
((recordLog.mock.calls[2][0] as unknown) as PipelineTaskLogMessage)
|
|
||||||
.task,
|
|
||||||
).toMatchObject(task);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('doTask', () => {
|
|
||||||
let task: PipelineTask;
|
|
||||||
|
|
||||||
beforeEach(() => {
|
|
||||||
task = new PipelineTask();
|
|
||||||
task.id = 'test-id';
|
|
||||||
task.logs = [];
|
|
||||||
task.pipeline = new Pipeline();
|
|
||||||
task.pipeline.workUnitMetadata = {
|
|
||||||
version: 1,
|
|
||||||
units: [
|
|
||||||
{
|
|
||||||
type: PipelineUnits.checkout,
|
|
||||||
scripts: [],
|
|
||||||
},
|
|
||||||
{
|
|
||||||
type: PipelineUnits.installDependencies,
|
|
||||||
scripts: ["echo ' Hello, Fennec!'"],
|
|
||||||
},
|
|
||||||
],
|
|
||||||
};
|
|
||||||
task.units = task.pipeline.workUnitMetadata.units.map(
|
|
||||||
(unit) => unit.type,
|
|
||||||
);
|
|
||||||
task.pipeline.project = new Project();
|
|
||||||
task.pipeline.project.name = 'test-project';
|
|
||||||
});
|
|
||||||
|
|
||||||
it('success and update task on db', async () => {
|
|
||||||
const job: Job = ({
|
|
||||||
data: task,
|
|
||||||
update: jest.fn().mockImplementation(() => undefined),
|
|
||||||
} as unknown) as Job;
|
|
||||||
|
|
||||||
jest
|
|
||||||
.spyOn(consumer, 'runScript')
|
|
||||||
.mockImplementation(async () => undefined);
|
|
||||||
const updateTask = jest.spyOn(tasksService, 'updateTask');
|
|
||||||
|
|
||||||
await consumer.doTask(job);
|
|
||||||
|
|
||||||
expect(updateTask).toHaveBeenCalledTimes(4);
|
|
||||||
expect(updateTask.mock.calls[0][0].startedAt).toBeDefined();
|
|
||||||
expect(updateTask.mock.calls[1][0].endedAt).toBeDefined();
|
|
||||||
expect(updateTask.mock.calls[1][0].status).toEqual(TaskStatuses.success);
|
|
||||||
});
|
|
||||||
it('failed and update task on db', async () => {
|
|
||||||
const job: Job = ({
|
|
||||||
data: task,
|
|
||||||
update: jest.fn().mockImplementation(() => undefined),
|
|
||||||
} as unknown) as Job;
|
|
||||||
|
|
||||||
jest.spyOn(consumer, 'runScript').mockImplementation(async () => {
|
|
||||||
throw new ApplicationException('exec script failed');
|
|
||||||
});
|
|
||||||
const updateTask = jest.spyOn(tasksService, 'updateTask');
|
|
||||||
|
|
||||||
await consumer.doTask(job);
|
|
||||||
|
|
||||||
expect(updateTask).toHaveBeenCalledTimes(4);
|
|
||||||
expect(updateTask.mock.calls[0][0].startedAt).toBeDefined();
|
|
||||||
expect(updateTask.mock.calls[1][0].endedAt).toBeDefined();
|
|
||||||
expect(updateTask.mock.calls[1][0].status).toEqual(TaskStatuses.failed);
|
|
||||||
});
|
|
||||||
it('should do all task', async () => {
|
|
||||||
const job: Job = ({
|
|
||||||
data: task,
|
|
||||||
update: jest.fn().mockImplementation(() => undefined),
|
|
||||||
} as unknown) as Job;
|
|
||||||
|
|
||||||
const runScript = jest
|
|
||||||
.spyOn(consumer, 'runScript')
|
|
||||||
.mockImplementation(async () => undefined);
|
|
||||||
const updateTask = jest.spyOn(tasksService, 'updateTask');
|
|
||||||
|
|
||||||
await consumer.doTask(job);
|
|
||||||
|
|
||||||
expect(runScript).toHaveBeenCalledTimes(1);
|
|
||||||
expect(updateTask).toHaveBeenCalledTimes(4);
|
|
||||||
const taskDto: PipelineTask = updateTask.mock.calls[0][0];
|
|
||||||
expect(taskDto.logs).toHaveLength(2);
|
|
||||||
expect(taskDto.logs[0].status).toEqual(TaskStatuses.success);
|
|
||||||
expect(taskDto.logs[0].unit).toEqual(PipelineUnits.checkout);
|
|
||||||
});
|
|
||||||
it('should log error message', async () => {
|
|
||||||
const job: Job = ({
|
|
||||||
data: task,
|
|
||||||
update: jest.fn().mockImplementation(() => undefined),
|
|
||||||
} as unknown) as Job;
|
|
||||||
|
|
||||||
jest.spyOn(consumer, 'runScript').mockImplementation(async () => {
|
|
||||||
throw new Error('bad message');
|
|
||||||
});
|
|
||||||
const updateTask = jest.spyOn(tasksService, 'updateTask');
|
|
||||||
|
|
||||||
await consumer.doTask(job);
|
|
||||||
|
|
||||||
expect(updateTask).toHaveBeenCalledTimes(4);
|
|
||||||
const taskDto: PipelineTask = updateTask.mock.calls[0][0];
|
|
||||||
expect(taskDto.logs).toHaveLength(2);
|
|
||||||
expect(taskDto.logs[0].status).toEqual(TaskStatuses.success);
|
|
||||||
expect(taskDto.logs[1].status).toEqual(TaskStatuses.failed);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
@ -1,159 +0,0 @@
|
|||||||
import { PipelineTaskLogs } from './models/pipeline-task-logs.model';
|
|
||||||
import { ReposService } from './../repos/repos.service';
|
|
||||||
import {
|
|
||||||
OnQueueCompleted,
|
|
||||||
OnQueueFailed,
|
|
||||||
Process,
|
|
||||||
Processor,
|
|
||||||
} from '@nestjs/bull';
|
|
||||||
import { Job } from 'bull';
|
|
||||||
import { spawn } from 'child_process';
|
|
||||||
import { PipelineTask } from './pipeline-task.entity';
|
|
||||||
import { PIPELINE_TASK_QUEUE } from './pipeline-tasks.constants';
|
|
||||||
import { PipelineTasksService } from './pipeline-tasks.service';
|
|
||||||
import { ApplicationException } from '../commons/exceptions/application.exception';
|
|
||||||
import { PipelineUnits } from './enums/pipeline-units.enum';
|
|
||||||
import { PipelineTaskLogMessage } from './models/pipeline-task-log-message.module';
|
|
||||||
import { TaskStatuses } from './enums/task-statuses.enum';
|
|
||||||
import { PipelineTaskLogsService } from './pipeline-task-logs.service';
|
|
||||||
import debug from 'debug';
|
|
||||||
|
|
||||||
const log = debug('fennec:pipeline-tasks:consumer');
|
|
||||||
|
|
||||||
@Processor(PIPELINE_TASK_QUEUE)
|
|
||||||
export class PipelineTaskConsumer {
|
|
||||||
constructor(
|
|
||||||
private readonly service: PipelineTasksService,
|
|
||||||
private readonly reposService: ReposService,
|
|
||||||
private readonly logsService: PipelineTaskLogsService,
|
|
||||||
) {}
|
|
||||||
@Process()
|
|
||||||
async doTask(job: Job<PipelineTask>) {
|
|
||||||
let task = job.data;
|
|
||||||
if (task.pipeline.workUnitMetadata.version !== 1) {
|
|
||||||
throw new ApplicationException(
|
|
||||||
'work unit metadata version is not match.',
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
task.startedAt = new Date();
|
|
||||||
task.status = TaskStatuses.working;
|
|
||||||
task = await this.service.updateTask(task);
|
|
||||||
log('start job');
|
|
||||||
await job.update(task);
|
|
||||||
|
|
||||||
const workspaceRoot = this.reposService.getWorkspaceRootByTask(task);
|
|
||||||
|
|
||||||
const units = task.units.map(
|
|
||||||
(type) =>
|
|
||||||
task.pipeline.workUnitMetadata.units.find(
|
|
||||||
(unit) => unit.type === type,
|
|
||||||
) ?? { type: type, scripts: [] },
|
|
||||||
);
|
|
||||||
|
|
||||||
log('task have [%o] units', units);
|
|
||||||
try {
|
|
||||||
for (const unit of units) {
|
|
||||||
const unitLog = new PipelineTaskLogs();
|
|
||||||
unitLog.unit = unit.type;
|
|
||||||
unitLog.startedAt = new Date();
|
|
||||||
log('curr unit is %s', unit.type);
|
|
||||||
try {
|
|
||||||
// 检出代码前执行 git checkout
|
|
||||||
if (unit.type === PipelineUnits.checkout) {
|
|
||||||
log('begin checkout');
|
|
||||||
await this.reposService.checkout(task, workspaceRoot);
|
|
||||||
unitLog.status = TaskStatuses.success;
|
|
||||||
log('end checkout');
|
|
||||||
}
|
|
||||||
for (const script of unit.scripts) {
|
|
||||||
unitLog.logs += `[RUN SCRIPT] ${script}`;
|
|
||||||
log('begin runScript %s', script);
|
|
||||||
await this.runScript(script, workspaceRoot, task, unit.type);
|
|
||||||
log('end runScript %s', script);
|
|
||||||
}
|
|
||||||
unitLog.status = TaskStatuses.success;
|
|
||||||
} catch (err) {
|
|
||||||
unitLog.status = TaskStatuses.failed;
|
|
||||||
unitLog.logs += err.message;
|
|
||||||
throw err;
|
|
||||||
} finally {
|
|
||||||
unitLog.endedAt = new Date();
|
|
||||||
unitLog.logs = await this.logsService
|
|
||||||
.readLogsAsPipelineTaskLogs(task)
|
|
||||||
.then(
|
|
||||||
(taskLogs) =>
|
|
||||||
taskLogs.find((tl) => tl.unit === unit.type)?.logs ?? '',
|
|
||||||
);
|
|
||||||
task.logs.push(unitLog);
|
|
||||||
task = await this.service.updateTask(task);
|
|
||||||
await job.update(task);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
task.status = TaskStatuses.success;
|
|
||||||
} catch (err) {
|
|
||||||
task.status = TaskStatuses.failed;
|
|
||||||
log('task is failed', err);
|
|
||||||
} finally {
|
|
||||||
task.endedAt = new Date();
|
|
||||||
task = await this.service.updateTask(task);
|
|
||||||
await job.update(task);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
async runScript(
|
|
||||||
script: string,
|
|
||||||
workspaceRoot: string,
|
|
||||||
task?: PipelineTask,
|
|
||||||
unit?: PipelineUnits,
|
|
||||||
): Promise<void> {
|
|
||||||
return new Promise((resolve, reject) => {
|
|
||||||
const sub = spawn(script, {
|
|
||||||
shell: true,
|
|
||||||
cwd: workspaceRoot,
|
|
||||||
});
|
|
||||||
let loggingCount = 0; // semaphore
|
|
||||||
|
|
||||||
sub.stderr.on('data', (data: Buffer) => {
|
|
||||||
const str = data.toString();
|
|
||||||
loggingCount++;
|
|
||||||
this.logsService
|
|
||||||
.recordLog(PipelineTaskLogMessage.create(task, unit, str, true))
|
|
||||||
.finally(() => loggingCount--);
|
|
||||||
});
|
|
||||||
sub.stdout.on('data', (data: Buffer) => {
|
|
||||||
const str = data.toString();
|
|
||||||
loggingCount++;
|
|
||||||
this.logsService
|
|
||||||
.recordLog(PipelineTaskLogMessage.create(task, unit, str, false))
|
|
||||||
.finally(() => loggingCount--);
|
|
||||||
});
|
|
||||||
sub.addListener('close', async (code) => {
|
|
||||||
await new Promise<void>(async (resolve) => {
|
|
||||||
for (let i = 0; i < 10 && loggingCount > 0; i++) {
|
|
||||||
await new Promise((resolve) => setTimeout(resolve, 500));
|
|
||||||
log('waiting logging... (%dx500ms)', i);
|
|
||||||
}
|
|
||||||
resolve();
|
|
||||||
});
|
|
||||||
if (code === 0) {
|
|
||||||
return resolve();
|
|
||||||
}
|
|
||||||
return reject(new ApplicationException('exec script failed'));
|
|
||||||
});
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
@OnQueueCompleted()
|
|
||||||
onCompleted(job: Job<PipelineTask>) {
|
|
||||||
log('queue onCompleted');
|
|
||||||
this.service.doNextTask(job.data.pipeline);
|
|
||||||
}
|
|
||||||
|
|
||||||
@OnQueueFailed()
|
|
||||||
onFailed(job: Job<PipelineTask>) {
|
|
||||||
log('queue onFailed');
|
|
||||||
this.service.doNextTask(job.data.pipeline);
|
|
||||||
}
|
|
||||||
}
|
|
@ -36,4 +36,7 @@ export class PipelineTask extends AppBaseEntity {
|
|||||||
|
|
||||||
@Column({ nullable: true })
|
@Column({ nullable: true })
|
||||||
endedAt?: Date;
|
endedAt?: Date;
|
||||||
|
|
||||||
|
@Column({ nullable: true })
|
||||||
|
runOn: string;
|
||||||
}
|
}
|
||||||
|
74
src/pipeline-tasks/pipeline-task.logger.spec.ts
Normal file
74
src/pipeline-tasks/pipeline-task.logger.spec.ts
Normal file
@ -0,0 +1,74 @@
|
|||||||
|
import { Test, TestingModule } from '@nestjs/testing';
|
||||||
|
import { PipelineTaskLogger } from './pipeline-task.logger';
|
||||||
|
import { PipelineTaskEvent } from './models/pipeline-task-event';
|
||||||
|
import { take, timeout } from 'rxjs/operators';
|
||||||
|
|
||||||
|
describe('PipelineTaskRunner', () => {
|
||||||
|
let logger: PipelineTaskLogger;
|
||||||
|
let module: TestingModule;
|
||||||
|
|
||||||
|
beforeEach(async () => {
|
||||||
|
module = await Test.createTestingModule({
|
||||||
|
providers: [PipelineTaskLogger],
|
||||||
|
}).compile();
|
||||||
|
|
||||||
|
logger = module.get(PipelineTaskLogger);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should be defined', () => {
|
||||||
|
expect(logger).toBeDefined();
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('getMessage$', () => {
|
||||||
|
it('normal', async () => {
|
||||||
|
const event = new PipelineTaskEvent();
|
||||||
|
event.taskId = 'test';
|
||||||
|
const emittedAt = new Date();
|
||||||
|
event.emittedAt = emittedAt.toISOString() as any;
|
||||||
|
const message$ = logger.getMessage$('test');
|
||||||
|
|
||||||
|
let receiveEvent;
|
||||||
|
message$.pipe(take(1)).subscribe((value) => (receiveEvent = value));
|
||||||
|
await logger.handleEvent(event);
|
||||||
|
expect(receiveEvent).toMatchObject({
|
||||||
|
...event,
|
||||||
|
emittedAt,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
it('no match', async () => {
|
||||||
|
const event = new PipelineTaskEvent();
|
||||||
|
event.taskId = 'test';
|
||||||
|
const message$ = logger.getMessage$('other');
|
||||||
|
setTimeout(() => {
|
||||||
|
logger.handleEvent(event);
|
||||||
|
});
|
||||||
|
expect(message$.pipe(take(1), timeout(100)).toPromise()).rejects.toMatch(
|
||||||
|
'timeout',
|
||||||
|
);
|
||||||
|
});
|
||||||
|
it('multiple subscribers', async () => {
|
||||||
|
const event = new PipelineTaskEvent();
|
||||||
|
event.taskId = 'test';
|
||||||
|
const message$ = logger.getMessage$('test');
|
||||||
|
const message2$ = logger.getMessage$('test');
|
||||||
|
setTimeout(() => {
|
||||||
|
logger.handleEvent(event);
|
||||||
|
});
|
||||||
|
expect(message$.pipe(take(1), timeout(100)).toPromise()).resolves.toEqual(
|
||||||
|
event,
|
||||||
|
);
|
||||||
|
expect(
|
||||||
|
message2$.pipe(take(1), timeout(100)).toPromise(),
|
||||||
|
).resolves.toEqual(event);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('onModuleDestroy', () => {
|
||||||
|
it('complete observable when destroying module', async () => {
|
||||||
|
logger.onModuleDestroy();
|
||||||
|
await expect(
|
||||||
|
(logger as any).message$.toPromise(),
|
||||||
|
).resolves.toBeUndefined();
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
37
src/pipeline-tasks/pipeline-task.logger.ts
Normal file
37
src/pipeline-tasks/pipeline-task.logger.ts
Normal file
@ -0,0 +1,37 @@
|
|||||||
|
import { RabbitSubscribe } from '@golevelup/nestjs-rabbitmq';
|
||||||
|
import { Injectable, OnModuleDestroy } from '@nestjs/common';
|
||||||
|
import { plainToClass } from 'class-transformer';
|
||||||
|
import { Observable, Subject } from 'rxjs';
|
||||||
|
import { filter } from 'rxjs/operators';
|
||||||
|
import { PipelineTaskEvent } from './models/pipeline-task-event';
|
||||||
|
import {
|
||||||
|
EXCHANGE_PIPELINE_TASK_FANOUT,
|
||||||
|
QUEUE_HANDLE_PIPELINE_TASK_LOG_EVENT,
|
||||||
|
ROUTE_PIPELINE_TASK_LOG,
|
||||||
|
} from './pipeline-tasks.constants';
|
||||||
|
|
||||||
|
@Injectable()
|
||||||
|
export class PipelineTaskLogger implements OnModuleDestroy {
|
||||||
|
private readonly messageSubject = new Subject<PipelineTaskEvent>();
|
||||||
|
private readonly message$: Observable<PipelineTaskEvent> = this.messageSubject.pipe();
|
||||||
|
|
||||||
|
@RabbitSubscribe({
|
||||||
|
exchange: EXCHANGE_PIPELINE_TASK_FANOUT,
|
||||||
|
routingKey: ROUTE_PIPELINE_TASK_LOG,
|
||||||
|
queue: QUEUE_HANDLE_PIPELINE_TASK_LOG_EVENT,
|
||||||
|
queueOptions: {
|
||||||
|
autoDelete: true,
|
||||||
|
},
|
||||||
|
})
|
||||||
|
async handleEvent(message: PipelineTaskEvent) {
|
||||||
|
this.messageSubject.next(plainToClass(PipelineTaskEvent, message));
|
||||||
|
}
|
||||||
|
|
||||||
|
getMessage$(taskId: string) {
|
||||||
|
return this.message$.pipe(filter((event) => event.taskId === taskId));
|
||||||
|
}
|
||||||
|
|
||||||
|
onModuleDestroy() {
|
||||||
|
this.messageSubject.complete();
|
||||||
|
}
|
||||||
|
}
|
326
src/pipeline-tasks/pipeline-task.runner.spec.ts
Normal file
326
src/pipeline-tasks/pipeline-task.runner.spec.ts
Normal file
@ -0,0 +1,326 @@
|
|||||||
|
import { Test, TestingModule } from '@nestjs/testing';
|
||||||
|
import { ReposService } from '../repos/repos.service';
|
||||||
|
import { PipelineUnits } from './enums/pipeline-units.enum';
|
||||||
|
import { PipelineTask } from './pipeline-task.entity';
|
||||||
|
import { Pipeline } from '../pipelines/pipeline.entity';
|
||||||
|
import { Project } from '../projects/project.entity';
|
||||||
|
import { TaskStatuses } from './enums/task-statuses.enum';
|
||||||
|
import { getLoggerToken, PinoLogger } from 'nestjs-pino';
|
||||||
|
import { PipelineTaskRunner } from './pipeline-task.runner';
|
||||||
|
import { WorkUnitMetadata } from './models/work-unit-metadata.model';
|
||||||
|
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
|
||||||
|
describe('PipelineTaskRunner', () => {
|
||||||
|
let runner: PipelineTaskRunner;
|
||||||
|
let reposService: ReposService;
|
||||||
|
|
||||||
|
beforeEach(async () => {
|
||||||
|
const module: TestingModule = await Test.createTestingModule({
|
||||||
|
providers: [
|
||||||
|
{
|
||||||
|
provide: ReposService,
|
||||||
|
useValue: {
|
||||||
|
getWorkspaceRootByTask: () => 'workspace-root',
|
||||||
|
checkout: async () => undefined,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
provide: getLoggerToken(PipelineTaskRunner.name),
|
||||||
|
useValue: new PinoLogger({}),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
provide: 'spawn',
|
||||||
|
useValue: () => undefined,
|
||||||
|
},
|
||||||
|
PipelineTaskRunner,
|
||||||
|
{
|
||||||
|
provide: AmqpConnection,
|
||||||
|
useValue: {},
|
||||||
|
},
|
||||||
|
],
|
||||||
|
}).compile();
|
||||||
|
|
||||||
|
reposService = module.get(ReposService);
|
||||||
|
runner = module.get(PipelineTaskRunner);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('should be defined', () => {
|
||||||
|
expect(runner).toBeDefined();
|
||||||
|
});
|
||||||
|
|
||||||
|
it('onNewTask', async () => {
|
||||||
|
const task = new PipelineTask();
|
||||||
|
let tmpTask;
|
||||||
|
const doTask = jest
|
||||||
|
.spyOn(runner, 'doTask')
|
||||||
|
.mockImplementation(async (task) => {
|
||||||
|
tmpTask = task;
|
||||||
|
});
|
||||||
|
await runner.onNewTask(task);
|
||||||
|
expect(tmpTask).toEqual(task);
|
||||||
|
expect(doTask).toBeCalledTimes(1);
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('test biz', () => {
|
||||||
|
let emitEvent: jest.SpyInstance;
|
||||||
|
beforeEach(() => {
|
||||||
|
emitEvent = jest
|
||||||
|
.spyOn(runner, 'emitEvent')
|
||||||
|
.mockImplementation((..._) => Promise.resolve());
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('doTask', () => {
|
||||||
|
let checkout: jest.SpyInstance;
|
||||||
|
let doTaskUnit: jest.SpyInstance;
|
||||||
|
|
||||||
|
beforeEach(() => {
|
||||||
|
checkout = jest
|
||||||
|
.spyOn(runner, 'checkout')
|
||||||
|
.mockImplementation((..._) => Promise.resolve('/null'));
|
||||||
|
doTaskUnit = jest
|
||||||
|
.spyOn(runner, 'doTaskUnit')
|
||||||
|
.mockImplementation((..._) => Promise.resolve());
|
||||||
|
});
|
||||||
|
|
||||||
|
it('only checkout', async () => {
|
||||||
|
const task = new PipelineTask();
|
||||||
|
(task.id = 'taskId'), (task.pipeline = new Pipeline());
|
||||||
|
task.units = [PipelineUnits.checkout];
|
||||||
|
task.pipeline.id = 'pipelineId';
|
||||||
|
task.pipeline.project = new Project();
|
||||||
|
task.pipeline.project.id = 'projectId';
|
||||||
|
task.pipeline.workUnitMetadata = new WorkUnitMetadata();
|
||||||
|
task.pipeline.workUnitMetadata.version = 1;
|
||||||
|
task.pipeline.workUnitMetadata.units = [
|
||||||
|
{
|
||||||
|
type: PipelineUnits.checkout,
|
||||||
|
scripts: [],
|
||||||
|
},
|
||||||
|
];
|
||||||
|
|
||||||
|
await runner.doTask(task);
|
||||||
|
|
||||||
|
expect(checkout).toBeCalledTimes(1);
|
||||||
|
expect(doTaskUnit).toBeCalledTimes(0);
|
||||||
|
expect(emitEvent).toBeCalledTimes(2);
|
||||||
|
expect(emitEvent.mock.calls[0][0]).toMatchObject(task);
|
||||||
|
expect(emitEvent.mock.calls[0][1]).toBeNull();
|
||||||
|
expect(emitEvent.mock.calls[0][2]).toEqual(TaskStatuses.working);
|
||||||
|
expect(emitEvent.mock.calls[1][0]).toMatchObject(task);
|
||||||
|
expect(emitEvent.mock.calls[1][1]).toBeNull();
|
||||||
|
expect(emitEvent.mock.calls[1][2]).toEqual(TaskStatuses.success);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('many units', async () => {
|
||||||
|
const task = new PipelineTask();
|
||||||
|
(task.id = 'taskId'), (task.pipeline = new Pipeline());
|
||||||
|
task.units = [
|
||||||
|
PipelineUnits.checkout,
|
||||||
|
PipelineUnits.test,
|
||||||
|
PipelineUnits.deploy,
|
||||||
|
];
|
||||||
|
task.pipeline.id = 'pipelineId';
|
||||||
|
task.pipeline.project = new Project();
|
||||||
|
task.pipeline.project.id = 'projectId';
|
||||||
|
task.pipeline.workUnitMetadata = new WorkUnitMetadata();
|
||||||
|
task.pipeline.workUnitMetadata.version = 1;
|
||||||
|
task.pipeline.workUnitMetadata.units = [
|
||||||
|
{
|
||||||
|
type: PipelineUnits.checkout,
|
||||||
|
scripts: [],
|
||||||
|
},
|
||||||
|
{
|
||||||
|
type: PipelineUnits.installDependencies,
|
||||||
|
scripts: ['pwd'],
|
||||||
|
},
|
||||||
|
{
|
||||||
|
type: PipelineUnits.test,
|
||||||
|
scripts: ['pwd'],
|
||||||
|
},
|
||||||
|
{
|
||||||
|
type: PipelineUnits.deploy,
|
||||||
|
scripts: ['pwd', 'uname'],
|
||||||
|
},
|
||||||
|
];
|
||||||
|
|
||||||
|
await runner.doTask(task);
|
||||||
|
|
||||||
|
expect(checkout).toBeCalledTimes(1);
|
||||||
|
expect(doTaskUnit).toBeCalledTimes(2);
|
||||||
|
expect(emitEvent).toBeCalledTimes(2);
|
||||||
|
});
|
||||||
|
|
||||||
|
it('unit work failed', async () => {
|
||||||
|
const task = new PipelineTask();
|
||||||
|
(task.id = 'taskId'), (task.pipeline = new Pipeline());
|
||||||
|
task.units = [PipelineUnits.checkout, PipelineUnits.test];
|
||||||
|
task.pipeline.id = 'pipelineId';
|
||||||
|
task.pipeline.project = new Project();
|
||||||
|
task.pipeline.project.id = 'projectId';
|
||||||
|
task.pipeline.workUnitMetadata = new WorkUnitMetadata();
|
||||||
|
task.pipeline.workUnitMetadata.version = 1;
|
||||||
|
task.pipeline.workUnitMetadata.units = [
|
||||||
|
{
|
||||||
|
type: PipelineUnits.checkout,
|
||||||
|
scripts: [],
|
||||||
|
},
|
||||||
|
{
|
||||||
|
type: PipelineUnits.test,
|
||||||
|
scripts: ['pwd'],
|
||||||
|
},
|
||||||
|
];
|
||||||
|
|
||||||
|
doTaskUnit = jest
|
||||||
|
.spyOn(runner, 'doTaskUnit')
|
||||||
|
.mockImplementation((..._) =>
|
||||||
|
Promise.reject(new Error('test error')),
|
||||||
|
);
|
||||||
|
await runner.doTask(task);
|
||||||
|
|
||||||
|
expect(checkout).toBeCalledTimes(1);
|
||||||
|
expect(doTaskUnit).toBeCalledTimes(1);
|
||||||
|
expect(emitEvent).toBeCalledTimes(2);
|
||||||
|
expect(emitEvent.mock.calls[1][0]).toMatchObject(task);
|
||||||
|
expect(emitEvent.mock.calls[1][1]).toBeNull();
|
||||||
|
expect(emitEvent.mock.calls[1][2]).toEqual(TaskStatuses.failed);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('doTaskUnit', () => {
|
||||||
|
it('success', async () => {
|
||||||
|
const runScript = jest
|
||||||
|
.spyOn(runner, 'runScript')
|
||||||
|
.mockImplementation((..._) => Promise.resolve());
|
||||||
|
const task = new PipelineTask();
|
||||||
|
|
||||||
|
const unit = PipelineUnits.test;
|
||||||
|
const workspacePath = '/null';
|
||||||
|
await runner.doTaskUnit(unit, ['pwd'], task, workspacePath);
|
||||||
|
|
||||||
|
expect(emitEvent.mock.calls[0][0]).toEqual(task);
|
||||||
|
expect(emitEvent.mock.calls[0][1]).toEqual(unit);
|
||||||
|
expect(emitEvent.mock.calls[0][2]).toEqual(TaskStatuses.working);
|
||||||
|
expect(emitEvent.mock.calls[1][0]).toEqual(task);
|
||||||
|
expect(emitEvent.mock.calls[1][1]).toEqual(unit);
|
||||||
|
expect(emitEvent.mock.calls[1][2]).toEqual(TaskStatuses.success);
|
||||||
|
expect(runScript.mock.calls[0][0]).toEqual('pwd');
|
||||||
|
expect(runScript.mock.calls[0][1]).toEqual(workspacePath);
|
||||||
|
expect(runScript.mock.calls[0][2]).toEqual(task);
|
||||||
|
expect(runScript.mock.calls[0][3]).toEqual(unit);
|
||||||
|
});
|
||||||
|
it('failed', async () => {
|
||||||
|
const runScript = jest
|
||||||
|
.spyOn(runner, 'runScript')
|
||||||
|
.mockImplementation((..._) =>
|
||||||
|
Promise.reject(new Error('test error')),
|
||||||
|
);
|
||||||
|
const task = new PipelineTask();
|
||||||
|
|
||||||
|
const unit = PipelineUnits.test;
|
||||||
|
const workspacePath = '/null';
|
||||||
|
await expect(
|
||||||
|
runner.doTaskUnit(unit, ['pwd'], task, workspacePath),
|
||||||
|
).rejects.toThrow('test error');
|
||||||
|
|
||||||
|
expect(emitEvent.mock.calls[1]?.[0]).toEqual(task);
|
||||||
|
expect(emitEvent.mock.calls[1]?.[1]).toEqual(unit);
|
||||||
|
expect(emitEvent.mock.calls[1]?.[2]).toEqual(TaskStatuses.failed);
|
||||||
|
expect(runScript).toBeCalledTimes(1);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('runScript', () => {
|
||||||
|
it('normal', async () => {
|
||||||
|
const spawn = jest.fn((..._: any[]) => ({
|
||||||
|
stdout: {
|
||||||
|
on: () => undefined,
|
||||||
|
},
|
||||||
|
stderr: {
|
||||||
|
on: () => undefined,
|
||||||
|
},
|
||||||
|
addListener: (_: any, fn: (code: number) => void) => {
|
||||||
|
fn(0);
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
(runner as any).spawn = spawn;
|
||||||
|
|
||||||
|
const task = new PipelineTask();
|
||||||
|
task.id = 'taskId';
|
||||||
|
const unit = PipelineUnits.deploy;
|
||||||
|
|
||||||
|
await runner.runScript('script name', 'workspaceRoot', task, unit);
|
||||||
|
expect(spawn).toHaveBeenCalledTimes(1);
|
||||||
|
expect(spawn.mock.calls[0][0]).toEqual('script name');
|
||||||
|
expect(spawn.mock.calls[0][1]).toMatchObject({
|
||||||
|
shell: true,
|
||||||
|
cwd: 'workspaceRoot',
|
||||||
|
});
|
||||||
|
});
|
||||||
|
it('failed', async () => {
|
||||||
|
const spawn = jest.fn((..._: any[]) => ({
|
||||||
|
stdout: {
|
||||||
|
on: () => undefined,
|
||||||
|
},
|
||||||
|
stderr: {
|
||||||
|
on: () => undefined,
|
||||||
|
},
|
||||||
|
addListener: (_: any, fn: (code: number) => void) => {
|
||||||
|
fn(1);
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
(runner as any).spawn = spawn;
|
||||||
|
|
||||||
|
const task = new PipelineTask();
|
||||||
|
task.id = 'taskId';
|
||||||
|
const unit = PipelineUnits.deploy;
|
||||||
|
|
||||||
|
expect(
|
||||||
|
runner.runScript('script name', 'workspaceRoot', task, unit),
|
||||||
|
).rejects.toThrowError();
|
||||||
|
});
|
||||||
|
it('wait emit message done', async () => {
|
||||||
|
let finishedFn: () => void;
|
||||||
|
const on = jest.fn((_: any, fn: (buff: Buffer) => void) => {
|
||||||
|
setTimeout(() => {
|
||||||
|
fn(Buffer.from('message 1'));
|
||||||
|
setTimeout(() => {
|
||||||
|
fn(Buffer.from('message 2'));
|
||||||
|
setTimeout(() => {
|
||||||
|
fn(Buffer.from('message 3'));
|
||||||
|
finishedFn();
|
||||||
|
}, 1000);
|
||||||
|
}, 10);
|
||||||
|
}, 10);
|
||||||
|
});
|
||||||
|
const spawn = jest.fn((..._: any[]) => ({
|
||||||
|
stdout: {
|
||||||
|
on,
|
||||||
|
},
|
||||||
|
stderr: {
|
||||||
|
on,
|
||||||
|
},
|
||||||
|
addListener: (_: any, fn: (code: number) => void) => {
|
||||||
|
finishedFn = () => fn(0);
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
let emitSuccessCount = 0;
|
||||||
|
jest.spyOn(runner, 'emitEvent').mockImplementation((..._: any[]) => {
|
||||||
|
return new Promise((resolve) => {
|
||||||
|
setTimeout(() => {
|
||||||
|
emitSuccessCount++;
|
||||||
|
resolve();
|
||||||
|
}, 1000);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
(runner as any).spawn = spawn;
|
||||||
|
|
||||||
|
const task = new PipelineTask();
|
||||||
|
task.id = 'taskId';
|
||||||
|
const unit = PipelineUnits.deploy;
|
||||||
|
|
||||||
|
await runner.runScript('script name', 'workspaceRoot', task, unit);
|
||||||
|
expect(emitSuccessCount).toEqual(1 + 6);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
301
src/pipeline-tasks/pipeline-task.runner.ts
Normal file
301
src/pipeline-tasks/pipeline-task.runner.ts
Normal file
@ -0,0 +1,301 @@
|
|||||||
|
import { ReposService } from '../repos/repos.service';
|
||||||
|
import { spawn, ChildProcessWithoutNullStreams } from 'child_process';
|
||||||
|
import { PipelineTask } from './pipeline-task.entity';
|
||||||
|
import { ApplicationException } from '../commons/exceptions/application.exception';
|
||||||
|
import { PipelineUnits } from './enums/pipeline-units.enum';
|
||||||
|
import { TaskStatuses } from './enums/task-statuses.enum';
|
||||||
|
import { InjectPinoLogger, PinoLogger } from 'nestjs-pino';
|
||||||
|
import {
|
||||||
|
AmqpConnection,
|
||||||
|
RabbitRPC,
|
||||||
|
RabbitSubscribe,
|
||||||
|
} from '@golevelup/nestjs-rabbitmq';
|
||||||
|
import { PipelineTaskEvent } from './models/pipeline-task-event';
|
||||||
|
import { last } from 'ramda';
|
||||||
|
import { Inject } from '@nestjs/common';
|
||||||
|
import {
|
||||||
|
EXCHANGE_PIPELINE_TASK_TOPIC,
|
||||||
|
QUEUE_PIPELINE_TASK_KILL,
|
||||||
|
ROUTE_PIPELINE_TASK_KILL,
|
||||||
|
} from './pipeline-tasks.constants';
|
||||||
|
import {
|
||||||
|
EXCHANGE_PIPELINE_TASK_FANOUT,
|
||||||
|
ROUTE_PIPELINE_TASK_LOG,
|
||||||
|
} from './pipeline-tasks.constants';
|
||||||
|
import {
|
||||||
|
getInstanceName,
|
||||||
|
getSelfInstanceQueueKey,
|
||||||
|
getSelfInstanceRouteKey,
|
||||||
|
} from '../commons/utils/rabbit-mq';
|
||||||
|
|
||||||
|
type Spawn = typeof spawn;
|
||||||
|
|
||||||
|
export class PipelineTaskRunner {
|
||||||
|
readonly processes = new Map<string, ChildProcessWithoutNullStreams>();
|
||||||
|
readonly stopTaskIds = new Set<string>();
|
||||||
|
|
||||||
|
constructor(
|
||||||
|
private readonly reposService: ReposService,
|
||||||
|
@InjectPinoLogger(PipelineTaskRunner.name)
|
||||||
|
private readonly logger: PinoLogger,
|
||||||
|
@Inject('spawn')
|
||||||
|
private readonly spawn: Spawn,
|
||||||
|
private readonly amqpConnection: AmqpConnection,
|
||||||
|
) {}
|
||||||
|
@RabbitSubscribe({
|
||||||
|
exchange: 'new-pipeline-task',
|
||||||
|
routingKey: 'mac',
|
||||||
|
queue: 'mac.new-pipeline-task',
|
||||||
|
})
|
||||||
|
async onNewTask(task: PipelineTask) {
|
||||||
|
this.logger.info({ task }, 'on new task [%s].', task.id);
|
||||||
|
try {
|
||||||
|
await this.doTask(task);
|
||||||
|
} catch (err) {
|
||||||
|
this.logger.error({ task, err }, err.message);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
@RabbitRPC({
|
||||||
|
exchange: EXCHANGE_PIPELINE_TASK_TOPIC,
|
||||||
|
routingKey: getSelfInstanceRouteKey(ROUTE_PIPELINE_TASK_KILL),
|
||||||
|
queue: getSelfInstanceQueueKey(QUEUE_PIPELINE_TASK_KILL),
|
||||||
|
queueOptions: {
|
||||||
|
autoDelete: true,
|
||||||
|
durable: true,
|
||||||
|
},
|
||||||
|
})
|
||||||
|
async onStopTask(task: PipelineTask) {
|
||||||
|
this.logger.info({ task }, 'on stop task [%s].', task.id);
|
||||||
|
this.stopTaskIds.add(task.id);
|
||||||
|
const process = this.processes.get(task.id);
|
||||||
|
if (process) {
|
||||||
|
this.logger.info({ task }, 'send signal SIGINT to child process.');
|
||||||
|
process.kill('SIGINT');
|
||||||
|
|
||||||
|
setTimeout(() => {
|
||||||
|
setTimeout(() => {
|
||||||
|
this.stopTaskIds.delete(task.id);
|
||||||
|
}, 10_000);
|
||||||
|
if (process === this.processes.get(task.id)) {
|
||||||
|
this.logger.info({ task }, 'send signal SIGKILL to child process.');
|
||||||
|
process.kill('SIGKILL');
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
if (this.processes.has(task.id)) {
|
||||||
|
this.logger.error(
|
||||||
|
{ task },
|
||||||
|
'this pipeline task not stop yet. there is a new process running, maybe is a bug about error capture',
|
||||||
|
);
|
||||||
|
}
|
||||||
|
}, 10_000);
|
||||||
|
} else {
|
||||||
|
this.logger.info({ task }, 'child process is not running.');
|
||||||
|
}
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
async doTask(task: PipelineTask) {
|
||||||
|
if (task.pipeline.workUnitMetadata.version !== 1) {
|
||||||
|
throw new ApplicationException(
|
||||||
|
'work unit metadata version is not match.',
|
||||||
|
);
|
||||||
|
}
|
||||||
|
await this.emitEvent(
|
||||||
|
task,
|
||||||
|
null,
|
||||||
|
TaskStatuses.working,
|
||||||
|
`[start task]`,
|
||||||
|
'stdout',
|
||||||
|
);
|
||||||
|
|
||||||
|
this.logger.info('running task [%s].', task.id);
|
||||||
|
try {
|
||||||
|
const workspaceRoot = await this.checkout(task);
|
||||||
|
const units = task.units
|
||||||
|
.filter((unit) => unit !== PipelineUnits.checkout)
|
||||||
|
.map(
|
||||||
|
(type) =>
|
||||||
|
task.pipeline.workUnitMetadata.units.find(
|
||||||
|
(unit) => unit.type === type,
|
||||||
|
) ?? { type: type, scripts: [] },
|
||||||
|
);
|
||||||
|
this.logger.info({ units }, 'begin run units.');
|
||||||
|
for (const unit of units) {
|
||||||
|
await this.doTaskUnit(unit.type, unit.scripts, task, workspaceRoot);
|
||||||
|
}
|
||||||
|
await this.emitEvent(
|
||||||
|
task,
|
||||||
|
null,
|
||||||
|
TaskStatuses.success,
|
||||||
|
`[finished task] success`,
|
||||||
|
'stdout',
|
||||||
|
);
|
||||||
|
this.logger.info({ task }, 'task [%s] completed.', task.id);
|
||||||
|
} catch (err) {
|
||||||
|
await this.emitEvent(
|
||||||
|
task,
|
||||||
|
null,
|
||||||
|
TaskStatuses.failed,
|
||||||
|
`[finished unit] ${err.message}`,
|
||||||
|
'stderr',
|
||||||
|
);
|
||||||
|
this.logger.error({ task, error: err }, 'task [%s] failed.', task.id);
|
||||||
|
} finally {
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async doTaskUnit(
|
||||||
|
unit: PipelineUnits,
|
||||||
|
scripts: string[],
|
||||||
|
task: PipelineTask,
|
||||||
|
workspaceRoot: string,
|
||||||
|
) {
|
||||||
|
await this.emitEvent(
|
||||||
|
task,
|
||||||
|
unit,
|
||||||
|
TaskStatuses.working,
|
||||||
|
`[begin unit] ${unit}`,
|
||||||
|
'stdin',
|
||||||
|
);
|
||||||
|
this.logger.info({ task }, 'curr unit is %s', unit);
|
||||||
|
try {
|
||||||
|
for (const script of scripts) {
|
||||||
|
this.logger.debug('begin runScript %s', script);
|
||||||
|
if (this.stopTaskIds.has(task.id)) {
|
||||||
|
throw new ApplicationException('Task is be KILLED');
|
||||||
|
}
|
||||||
|
await this.runScript(script, workspaceRoot, task, unit);
|
||||||
|
this.logger.debug('end runScript %s', script);
|
||||||
|
}
|
||||||
|
|
||||||
|
await this.emitEvent(
|
||||||
|
task,
|
||||||
|
unit,
|
||||||
|
TaskStatuses.success,
|
||||||
|
`[finished unit] ${unit}`,
|
||||||
|
'stdout',
|
||||||
|
);
|
||||||
|
} catch (err) {
|
||||||
|
await this.emitEvent(
|
||||||
|
task,
|
||||||
|
unit,
|
||||||
|
TaskStatuses.failed,
|
||||||
|
`[finished unit] ${err.message}`,
|
||||||
|
'stderr',
|
||||||
|
);
|
||||||
|
throw err;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async checkout(task: PipelineTask) {
|
||||||
|
await this.emitEvent(
|
||||||
|
task,
|
||||||
|
PipelineUnits.checkout,
|
||||||
|
TaskStatuses.working,
|
||||||
|
'[begin unit] checkout',
|
||||||
|
'stdin',
|
||||||
|
);
|
||||||
|
try {
|
||||||
|
const path = await this.reposService.checkout4Task(task);
|
||||||
|
await this.emitEvent(
|
||||||
|
task,
|
||||||
|
PipelineUnits.checkout,
|
||||||
|
TaskStatuses.success,
|
||||||
|
'checkout success.',
|
||||||
|
'stdout',
|
||||||
|
);
|
||||||
|
return path;
|
||||||
|
} catch (err) {
|
||||||
|
await this.emitEvent(
|
||||||
|
task,
|
||||||
|
PipelineUnits.checkout,
|
||||||
|
TaskStatuses.failed,
|
||||||
|
'checkout failed.',
|
||||||
|
'stderr',
|
||||||
|
);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async emitEvent(
|
||||||
|
task: PipelineTask,
|
||||||
|
unit: PipelineUnits | null,
|
||||||
|
status: TaskStatuses,
|
||||||
|
message: string,
|
||||||
|
messageType: 'stderr' | 'stdout' | 'stdin',
|
||||||
|
) {
|
||||||
|
const event: PipelineTaskEvent = {
|
||||||
|
taskId: task.id,
|
||||||
|
pipelineId: task.pipeline.id,
|
||||||
|
projectId: task.pipeline.project.id,
|
||||||
|
unit,
|
||||||
|
emittedAt: new Date(),
|
||||||
|
message: last(message) === '\n' ? message : message + '\n',
|
||||||
|
messageType,
|
||||||
|
status,
|
||||||
|
};
|
||||||
|
this.amqpConnection
|
||||||
|
.publish(EXCHANGE_PIPELINE_TASK_FANOUT, ROUTE_PIPELINE_TASK_LOG, event, {
|
||||||
|
headers: {
|
||||||
|
sender: getInstanceName(),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
.catch((error) => {
|
||||||
|
this.logger.error(
|
||||||
|
{ error, event },
|
||||||
|
'send event message to queue failed. %s',
|
||||||
|
error.message,
|
||||||
|
);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
async runScript(
|
||||||
|
script: string,
|
||||||
|
workspaceRoot: string,
|
||||||
|
task: PipelineTask,
|
||||||
|
unit: PipelineUnits,
|
||||||
|
): Promise<void> {
|
||||||
|
await this.emitEvent(task, unit, TaskStatuses.working, script, 'stdin');
|
||||||
|
return new Promise((resolve, reject) => {
|
||||||
|
const sub = this.spawn(script, {
|
||||||
|
shell: true,
|
||||||
|
cwd: workspaceRoot,
|
||||||
|
});
|
||||||
|
this.processes.set(task.id, sub);
|
||||||
|
let loggingCount = 0; // semaphore
|
||||||
|
|
||||||
|
sub.stderr.on('data', (data: Buffer) => {
|
||||||
|
const str = data.toString();
|
||||||
|
loggingCount++;
|
||||||
|
|
||||||
|
this.emitEvent(task, unit, TaskStatuses.working, str, 'stdout').finally(
|
||||||
|
() => loggingCount--,
|
||||||
|
);
|
||||||
|
});
|
||||||
|
sub.stdout.on('data', (data: Buffer) => {
|
||||||
|
const str = data.toString();
|
||||||
|
loggingCount++;
|
||||||
|
|
||||||
|
this.emitEvent(task, unit, TaskStatuses.working, str, 'stderr').finally(
|
||||||
|
() => loggingCount--,
|
||||||
|
);
|
||||||
|
});
|
||||||
|
sub.addListener('close', async (code) => {
|
||||||
|
this.processes.delete(task.id);
|
||||||
|
await new Promise<void>(async (resolve) => {
|
||||||
|
for (let i = 0; i < 10 && loggingCount > 0; i++) {
|
||||||
|
await new Promise((resolve) => setTimeout(resolve, 500));
|
||||||
|
this.logger.debug('waiting logging... (%dx500ms)', i);
|
||||||
|
}
|
||||||
|
resolve();
|
||||||
|
});
|
||||||
|
if (code === 0) {
|
||||||
|
return resolve();
|
||||||
|
}
|
||||||
|
if (this.stopTaskIds.has(task.id)) {
|
||||||
|
throw reject(new ApplicationException('Task is be KILLED'));
|
||||||
|
}
|
||||||
|
return reject(new ApplicationException('exec script failed'));
|
||||||
|
});
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
@ -1,2 +1,9 @@
|
|||||||
export const PIPELINE_TASK_QUEUE = 'PIPELINE_TASK_QUEUE';
|
export const EXCHANGE_PIPELINE_TASK_TOPIC = 'pipeline-task.topic';
|
||||||
export const PIPELINE_TASK_LOG_QUEUE = 'PIPELINE_TASK_LOG_QUEUE';
|
export const EXCHANGE_PIPELINE_TASK_FANOUT = 'pipeline-task.fanout';
|
||||||
|
export const ROUTE_PIPELINE_TASK_LOG = 'pipeline-task-log';
|
||||||
|
export const QUEUE_HANDLE_PIPELINE_TASK_LOG_EVENT = 'pipeline-task-log';
|
||||||
|
export const QUEUE_WRITE_PIPELINE_TASK_LOG = 'write-pipeline-task-log';
|
||||||
|
export const ROUTE_PIPELINE_TASK_DONE = 'pipeline-task-done';
|
||||||
|
export const QUEUE_PIPELINE_TASK_DONE = 'pipeline-task-done';
|
||||||
|
export const ROUTE_PIPELINE_TASK_KILL = 'pipeline-task-kill';
|
||||||
|
export const QUEUE_PIPELINE_TASK_KILL = 'pipeline-task-kill';
|
||||||
|
@ -6,27 +6,66 @@ import { PipelineTask } from './pipeline-task.entity';
|
|||||||
import { Pipeline } from '../pipelines/pipeline.entity';
|
import { Pipeline } from '../pipelines/pipeline.entity';
|
||||||
import { ReposModule } from '../repos/repos.module';
|
import { ReposModule } from '../repos/repos.module';
|
||||||
import { RedisModule } from 'nestjs-redis';
|
import { RedisModule } from 'nestjs-redis';
|
||||||
import { BullModule } from '@nestjs/bull';
|
import { RabbitMQModule } from '@golevelup/nestjs-rabbitmq';
|
||||||
import { PipelineTaskConsumer } from './pipeline-task.consumer';
|
import { ConfigModule, ConfigService } from '@nestjs/config';
|
||||||
import { PIPELINE_TASK_QUEUE } from './pipeline-tasks.constants';
|
import { PipelineTaskRunner } from './pipeline-task.runner';
|
||||||
import { PipelineTaskLogsService } from './pipeline-task-logs.service';
|
import { spawn } from 'child_process';
|
||||||
import { PubSubModule } from '../commons/pub-sub/pub-sub.module';
|
import {
|
||||||
|
EXCHANGE_PIPELINE_TASK_FANOUT,
|
||||||
|
EXCHANGE_PIPELINE_TASK_TOPIC,
|
||||||
|
} from './pipeline-tasks.constants';
|
||||||
|
import { PipelineTaskLogger } from './pipeline-task.logger';
|
||||||
|
import { PipelineTaskFlushService } from './pipeline-task-flush.service';
|
||||||
|
|
||||||
@Module({
|
@Module({
|
||||||
imports: [
|
imports: [
|
||||||
TypeOrmModule.forFeature([PipelineTask, Pipeline]),
|
TypeOrmModule.forFeature([PipelineTask, Pipeline]),
|
||||||
BullModule.registerQueue({
|
|
||||||
name: PIPELINE_TASK_QUEUE,
|
|
||||||
}),
|
|
||||||
PubSubModule.forFeature(),
|
|
||||||
RedisModule,
|
RedisModule,
|
||||||
ReposModule,
|
ReposModule,
|
||||||
|
RabbitMQModule.forRootAsync(RabbitMQModule, {
|
||||||
|
imports: [ConfigModule],
|
||||||
|
useFactory: (configService: ConfigService) => ({
|
||||||
|
uri: configService.get<string>('db.rabbitmq.uri'),
|
||||||
|
exchanges: [
|
||||||
|
{
|
||||||
|
name: 'new-pipeline-task',
|
||||||
|
type: 'fanout',
|
||||||
|
options: {
|
||||||
|
durable: true,
|
||||||
|
autoDelete: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: EXCHANGE_PIPELINE_TASK_FANOUT,
|
||||||
|
type: 'fanout',
|
||||||
|
options: {
|
||||||
|
durable: false,
|
||||||
|
autoDelete: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: EXCHANGE_PIPELINE_TASK_TOPIC,
|
||||||
|
type: 'topic',
|
||||||
|
options: {
|
||||||
|
durable: false,
|
||||||
|
autoDelete: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
],
|
||||||
|
}),
|
||||||
|
inject: [ConfigService],
|
||||||
|
}),
|
||||||
],
|
],
|
||||||
providers: [
|
providers: [
|
||||||
PipelineTasksService,
|
PipelineTasksService,
|
||||||
PipelineTasksResolver,
|
PipelineTasksResolver,
|
||||||
PipelineTaskConsumer,
|
PipelineTaskRunner,
|
||||||
PipelineTaskLogsService,
|
PipelineTaskLogger,
|
||||||
|
{
|
||||||
|
provide: 'spawn',
|
||||||
|
useValue: spawn,
|
||||||
|
},
|
||||||
|
PipelineTaskFlushService,
|
||||||
],
|
],
|
||||||
exports: [PipelineTasksService],
|
exports: [PipelineTasksService],
|
||||||
})
|
})
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
import { Test, TestingModule } from '@nestjs/testing';
|
import { Test, TestingModule } from '@nestjs/testing';
|
||||||
|
import { PipelineTaskLogger } from './pipeline-task.logger';
|
||||||
import { PipelineTasksResolver } from './pipeline-tasks.resolver';
|
import { PipelineTasksResolver } from './pipeline-tasks.resolver';
|
||||||
import { PipelineTaskLogsService } from './pipeline-task-logs.service';
|
|
||||||
import { PipelineTasksService } from './pipeline-tasks.service';
|
import { PipelineTasksService } from './pipeline-tasks.service';
|
||||||
|
|
||||||
describe('PipelineTasksResolver', () => {
|
describe('PipelineTasksResolver', () => {
|
||||||
@ -15,7 +15,7 @@ describe('PipelineTasksResolver', () => {
|
|||||||
useValue: {},
|
useValue: {},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
provide: PipelineTaskLogsService,
|
provide: PipelineTaskLogger,
|
||||||
useValue: {},
|
useValue: {},
|
||||||
},
|
},
|
||||||
],
|
],
|
||||||
|
@ -2,16 +2,17 @@ import { Resolver, Args, Mutation, Subscription, Query } from '@nestjs/graphql';
|
|||||||
import { PipelineTask } from './pipeline-task.entity';
|
import { PipelineTask } from './pipeline-task.entity';
|
||||||
import { PipelineTasksService } from './pipeline-tasks.service';
|
import { PipelineTasksService } from './pipeline-tasks.service';
|
||||||
import { CreatePipelineTaskInput } from './dtos/create-pipeline-task.input';
|
import { CreatePipelineTaskInput } from './dtos/create-pipeline-task.input';
|
||||||
import { PipelineTaskLogMessage } from './models/pipeline-task-log-message.module';
|
|
||||||
import { PipelineTaskLogArgs } from './dtos/pipeline-task-log.args';
|
import { PipelineTaskLogArgs } from './dtos/pipeline-task-log.args';
|
||||||
import { PipelineTaskLogsService } from './pipeline-task-logs.service';
|
|
||||||
import { plainToClass } from 'class-transformer';
|
import { plainToClass } from 'class-transformer';
|
||||||
|
import { PipelineTaskLogger } from './pipeline-task.logger';
|
||||||
|
import { observableToAsyncIterable } from '@graphql-tools/utils';
|
||||||
|
import { PipelineTaskEvent } from './models/pipeline-task-event';
|
||||||
|
|
||||||
@Resolver()
|
@Resolver()
|
||||||
export class PipelineTasksResolver {
|
export class PipelineTasksResolver {
|
||||||
constructor(
|
constructor(
|
||||||
private readonly service: PipelineTasksService,
|
private readonly service: PipelineTasksService,
|
||||||
private readonly logsService: PipelineTaskLogsService,
|
private readonly taskLogger: PipelineTaskLogger,
|
||||||
) {}
|
) {}
|
||||||
|
|
||||||
@Mutation(() => PipelineTask)
|
@Mutation(() => PipelineTask)
|
||||||
@ -19,16 +20,17 @@ export class PipelineTasksResolver {
|
|||||||
return await this.service.addTask(taskDto);
|
return await this.service.addTask(taskDto);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Subscription(() => PipelineTaskLogMessage, {
|
@Subscription(() => PipelineTaskEvent, {
|
||||||
resolve: (value) => {
|
resolve: (value) => {
|
||||||
const data = plainToClass(PipelineTaskLogMessage, value);
|
const data = plainToClass(PipelineTaskEvent, value);
|
||||||
return data;
|
return data;
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
async pipelineTaskLog(@Args() args: PipelineTaskLogArgs) {
|
async pipelineTaskEvent(@Args() args: PipelineTaskLogArgs) {
|
||||||
const task = await this.service.findTaskById(args.taskId);
|
const task = await this.service.findTaskById(args.taskId);
|
||||||
const asyncIterator = this.logsService.watchLogs(task);
|
return observableToAsyncIterable<PipelineTaskEvent>(
|
||||||
return asyncIterator;
|
this.taskLogger.getMessage$(task.id),
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Subscription(() => PipelineTask, {
|
@Subscription(() => PipelineTask, {
|
||||||
@ -37,7 +39,7 @@ export class PipelineTasksResolver {
|
|||||||
},
|
},
|
||||||
})
|
})
|
||||||
async pipelineTaskChanged(@Args('id') id: string) {
|
async pipelineTaskChanged(@Args('id') id: string) {
|
||||||
return await this.service.watchTaskUpdated(id);
|
// return await this.service.watchTaskUpdated(id);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Query(() => [PipelineTask])
|
@Query(() => [PipelineTask])
|
||||||
@ -46,7 +48,14 @@ export class PipelineTasksResolver {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Query(() => PipelineTask)
|
@Query(() => PipelineTask)
|
||||||
async findPipelineTask(@Args('id') id: string) {
|
async pipelineTask(@Args('id') id: string) {
|
||||||
return await this.service.findTaskById(id);
|
return await this.service.findTaskById(id);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Mutation(() => Boolean)
|
||||||
|
async stopPipelineTask(@Args('id') id: string) {
|
||||||
|
const task = await this.service.findTaskById(id);
|
||||||
|
await this.service.stopTask(task);
|
||||||
|
return true;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
@ -2,52 +2,19 @@ import { Test, TestingModule } from '@nestjs/testing';
|
|||||||
import { PipelineTasksService } from './pipeline-tasks.service';
|
import { PipelineTasksService } from './pipeline-tasks.service';
|
||||||
import { getRepositoryToken } from '@nestjs/typeorm';
|
import { getRepositoryToken } from '@nestjs/typeorm';
|
||||||
import { PipelineTask } from './pipeline-task.entity';
|
import { PipelineTask } from './pipeline-task.entity';
|
||||||
import { PIPELINE_TASK_QUEUE } from './pipeline-tasks.constants';
|
|
||||||
import { getQueueToken } from '@nestjs/bull';
|
|
||||||
import { RedisService } from 'nestjs-redis';
|
|
||||||
import { Pipeline } from '../pipelines/pipeline.entity';
|
import { Pipeline } from '../pipelines/pipeline.entity';
|
||||||
import { EntityNotFoundError } from 'typeorm/error/EntityNotFoundError';
|
|
||||||
import { Repository } from 'typeorm';
|
import { Repository } from 'typeorm';
|
||||||
import { Queue } from 'bull';
|
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
|
||||||
import { LockFailedException } from '../commons/exceptions/lock-failed.exception';
|
import { PipelineTaskFlushService } from './pipeline-task-flush.service';
|
||||||
import { getPubSubToken } from '../commons/pub-sub/utils/token';
|
import { getLoggerToken, PinoLogger } from 'nestjs-pino';
|
||||||
|
|
||||||
describe('PipelineTasksService', () => {
|
describe('PipelineTasksService', () => {
|
||||||
let service: PipelineTasksService;
|
let service: PipelineTasksService;
|
||||||
let module: TestingModule;
|
let module: TestingModule;
|
||||||
let taskRepository: Repository<PipelineTask>;
|
let taskRepository: Repository<PipelineTask>;
|
||||||
let pipelineRepository: Repository<Pipeline>;
|
let pipelineRepository: Repository<Pipeline>;
|
||||||
const getBasePipeline = () =>
|
|
||||||
({
|
|
||||||
id: 'test',
|
|
||||||
name: '测试流水线',
|
|
||||||
branch: 'master',
|
|
||||||
workUnitMetadata: {},
|
|
||||||
project: {
|
|
||||||
id: 'test-project',
|
|
||||||
},
|
|
||||||
} as Pipeline);
|
|
||||||
let redisClient;
|
|
||||||
let taskQueue: Queue;
|
|
||||||
const getTask = () =>
|
|
||||||
({
|
|
||||||
pipelineId: 'test',
|
|
||||||
commit: 'test',
|
|
||||||
pipeline: { branch: 'master' },
|
|
||||||
units: [],
|
|
||||||
} as PipelineTask);
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
beforeEach(async () => {
|
||||||
redisClient = (() => ({
|
|
||||||
set: jest.fn().mockImplementation(async () => 'OK'),
|
|
||||||
del: jest.fn().mockImplementation(async () => 'test'),
|
|
||||||
get: jest.fn().mockImplementation(async () => 'test'),
|
|
||||||
lpush: jest.fn().mockImplementation(async () => 1),
|
|
||||||
rpop: jest.fn().mockImplementation(async () => JSON.stringify(getTask())),
|
|
||||||
}))() as any;
|
|
||||||
taskQueue = (() => ({
|
|
||||||
add: jest.fn().mockImplementation(async () => null),
|
|
||||||
}))() as any;
|
|
||||||
module = await Test.createTestingModule({
|
module = await Test.createTestingModule({
|
||||||
providers: [
|
providers: [
|
||||||
PipelineTasksService,
|
PipelineTasksService,
|
||||||
@ -60,19 +27,17 @@ describe('PipelineTasksService', () => {
|
|||||||
useValue: new Repository(),
|
useValue: new Repository(),
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
provide: getQueueToken(PIPELINE_TASK_QUEUE),
|
provide: AmqpConnection,
|
||||||
useValue: taskQueue,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: RedisService,
|
|
||||||
useValue: {
|
|
||||||
getClient: jest.fn(() => redisClient),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: getPubSubToken(),
|
|
||||||
useValue: {},
|
useValue: {},
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
provide: PipelineTaskFlushService,
|
||||||
|
useValue: {},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
provide: getLoggerToken(PipelineTasksService.name),
|
||||||
|
useValue: new PinoLogger({}),
|
||||||
|
},
|
||||||
],
|
],
|
||||||
}).compile();
|
}).compile();
|
||||||
|
|
||||||
@ -92,119 +57,43 @@ describe('PipelineTasksService', () => {
|
|||||||
expect(service).toBeDefined();
|
expect(service).toBeDefined();
|
||||||
});
|
});
|
||||||
|
|
||||||
describe('addTask', () => {
|
// describe('addTask', () => {
|
||||||
beforeEach(() => {
|
// beforeEach(() => {
|
||||||
jest
|
// jest
|
||||||
.spyOn(pipelineRepository, 'findOneOrFail')
|
// .spyOn(pipelineRepository, 'findOneOrFail')
|
||||||
.mockImplementation(async () => getBasePipeline());
|
// .mockImplementation(async () => getBasePipeline());
|
||||||
});
|
// });
|
||||||
it('pipeline not found', async () => {
|
// it('pipeline not found', async () => {
|
||||||
jest.spyOn(taskRepository, 'findOneOrFail').mockImplementation(() => {
|
// jest.spyOn(taskRepository, 'findOneOrFail').mockImplementation(() => {
|
||||||
throw new EntityNotFoundError(Pipeline, {});
|
// throw new EntityNotFoundError(Pipeline, {});
|
||||||
});
|
// });
|
||||||
await expect(
|
// await expect(
|
||||||
service.addTask({ pipelineId: 'test', commit: 'test', units: [] }),
|
// service.addTask({ pipelineId: 'test', commit: 'test', units: [] }),
|
||||||
).rejects;
|
// ).rejects;
|
||||||
});
|
// });
|
||||||
it('create task on db', async () => {
|
// it('create task on db', async () => {
|
||||||
const save = jest
|
// const save = jest
|
||||||
.spyOn(taskRepository, 'save')
|
// .spyOn(taskRepository, 'save')
|
||||||
.mockImplementation(async (data: any) => data);
|
// .mockImplementation(async (data: any) => data);
|
||||||
const findOne = jest.spyOn(taskRepository, 'findOne');
|
// const findOne = jest.spyOn(taskRepository, 'findOne');
|
||||||
jest
|
// await service.addTask({ pipelineId: 'test', commit: 'test', units: [] }),
|
||||||
.spyOn(service, 'doNextTask')
|
// expect(save.mock.calls[0][0]).toMatchObject({
|
||||||
.mockImplementation(async () => undefined);
|
// pipelineId: 'test',
|
||||||
await service.addTask({ pipelineId: 'test', commit: 'test', units: [] }),
|
// commit: 'test',
|
||||||
expect(save.mock.calls[0][0]).toMatchObject({
|
// units: [],
|
||||||
pipelineId: 'test',
|
// });
|
||||||
commit: 'test',
|
// expect(findOne).toBeCalled();
|
||||||
units: [],
|
// });
|
||||||
});
|
// it('add task', async () => {
|
||||||
expect(findOne).toBeCalled();
|
// const lpush = jest.spyOn(redisClient, 'lpush');
|
||||||
});
|
// await service.addTask({ pipelineId: 'test', commit: 'test', units: [] });
|
||||||
it('add task', async () => {
|
// expect(typeof lpush.mock.calls[0][1] === 'string').toBeTruthy();
|
||||||
const lpush = jest.spyOn(redisClient, 'lpush');
|
// expect(JSON.parse(lpush.mock.calls[0][1] as string)).toMatchObject({
|
||||||
const doNextTask = jest.spyOn(service, 'doNextTask');
|
// pipelineId: 'test',
|
||||||
jest
|
// commit: 'test',
|
||||||
.spyOn(service, 'doNextTask')
|
// units: [],
|
||||||
.mockImplementation(async () => undefined);
|
// pipeline: getBasePipeline(),
|
||||||
await service.addTask({ pipelineId: 'test', commit: 'test', units: [] });
|
// });
|
||||||
expect(typeof lpush.mock.calls[0][1] === 'string').toBeTruthy();
|
// });
|
||||||
expect(JSON.parse(lpush.mock.calls[0][1] as string)).toMatchObject({
|
// });
|
||||||
pipelineId: 'test',
|
|
||||||
commit: 'test',
|
|
||||||
units: [],
|
|
||||||
pipeline: getBasePipeline(),
|
|
||||||
});
|
|
||||||
expect(doNextTask).toHaveBeenCalledWith(getBasePipeline());
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('doNextTask', () => {
|
|
||||||
it('add task to queue', async () => {
|
|
||||||
let lckValue: string;
|
|
||||||
const set = jest
|
|
||||||
.spyOn(redisClient, 'set')
|
|
||||||
.mockImplementation(async (...args) => (lckValue = args[3] as string));
|
|
||||||
const get = jest
|
|
||||||
.spyOn(redisClient, 'get')
|
|
||||||
.mockImplementation(async () => lckValue);
|
|
||||||
const del = jest.spyOn(redisClient, 'del');
|
|
||||||
const rpop = jest.spyOn(redisClient, 'rpop');
|
|
||||||
const add = jest.spyOn(taskQueue, 'add');
|
|
||||||
|
|
||||||
await service.doNextTask(getBasePipeline());
|
|
||||||
|
|
||||||
expect(add).toHaveBeenCalledWith(getTask());
|
|
||||||
expect(set).toHaveBeenCalledTimes(1);
|
|
||||||
expect(rpop).toHaveBeenCalledTimes(1);
|
|
||||||
expect(get).toHaveBeenCalledTimes(1);
|
|
||||||
expect(del).toHaveBeenCalledTimes(1);
|
|
||||||
});
|
|
||||||
it('pipeline is busy', async () => {
|
|
||||||
let remainTimes = 3;
|
|
||||||
|
|
||||||
let lckValue: string;
|
|
||||||
const set = jest
|
|
||||||
.spyOn(redisClient, 'set')
|
|
||||||
.mockImplementation(async (...args) => {
|
|
||||||
if (remainTimes-- > 0) {
|
|
||||||
throw new Error();
|
|
||||||
} else {
|
|
||||||
lckValue = args[3] as string;
|
|
||||||
}
|
|
||||||
});
|
|
||||||
const get = jest
|
|
||||||
.spyOn(redisClient, 'get')
|
|
||||||
.mockImplementation(async () => lckValue);
|
|
||||||
const del = jest.spyOn(redisClient, 'del');
|
|
||||||
const rpop = jest.spyOn(redisClient, 'rpop');
|
|
||||||
const add = jest.spyOn(taskQueue, 'add');
|
|
||||||
|
|
||||||
await service.doNextTask(getBasePipeline());
|
|
||||||
|
|
||||||
expect(rpop).toHaveBeenCalledTimes(1);
|
|
||||||
expect(set).toHaveBeenCalledTimes(4);
|
|
||||||
expect(get).toHaveBeenCalledTimes(1);
|
|
||||||
expect(del).toHaveBeenCalledTimes(1);
|
|
||||||
expect(add).toHaveBeenCalledWith(getTask());
|
|
||||||
}, 10_000);
|
|
||||||
it('pipeline always busy and timeout', async () => {
|
|
||||||
const set = jest
|
|
||||||
.spyOn(redisClient, 'set')
|
|
||||||
.mockImplementation(async () => {
|
|
||||||
throw new Error();
|
|
||||||
});
|
|
||||||
const get = jest.spyOn(redisClient, 'get');
|
|
||||||
const del = jest.spyOn(redisClient, 'del');
|
|
||||||
|
|
||||||
await expect(
|
|
||||||
service.doNextTask(getBasePipeline()),
|
|
||||||
).rejects.toBeInstanceOf(LockFailedException);
|
|
||||||
|
|
||||||
expect(set).toHaveBeenCalledTimes(5);
|
|
||||||
expect(get).toHaveBeenCalledTimes(0);
|
|
||||||
expect(del).toHaveBeenCalledTimes(0);
|
|
||||||
}, 15_000);
|
|
||||||
});
|
|
||||||
});
|
});
|
||||||
|
@ -1,20 +1,23 @@
|
|||||||
import { ConflictException, Injectable } from '@nestjs/common';
|
import { BadRequestException, Injectable } from '@nestjs/common';
|
||||||
import { InjectRepository } from '@nestjs/typeorm';
|
import { InjectRepository } from '@nestjs/typeorm';
|
||||||
import { PipelineTask } from './pipeline-task.entity';
|
import { PipelineTask } from './pipeline-task.entity';
|
||||||
import { In, Repository } from 'typeorm';
|
import { Repository } from 'typeorm';
|
||||||
import { CreatePipelineTaskInput } from './dtos/create-pipeline-task.input';
|
import { CreatePipelineTaskInput } from './dtos/create-pipeline-task.input';
|
||||||
import { RedisService } from 'nestjs-redis';
|
|
||||||
import { Pipeline } from '../pipelines/pipeline.entity';
|
import { Pipeline } from '../pipelines/pipeline.entity';
|
||||||
import { InjectQueue } from '@nestjs/bull';
|
|
||||||
import { PIPELINE_TASK_QUEUE } from './pipeline-tasks.constants';
|
|
||||||
import { Queue } from 'bull';
|
|
||||||
import { LockFailedException } from '../commons/exceptions/lock-failed.exception';
|
|
||||||
import { TaskStatuses } from './enums/task-statuses.enum';
|
|
||||||
import { isNil } from 'ramda';
|
|
||||||
import debug from 'debug';
|
import debug from 'debug';
|
||||||
import { InjectPubSub } from '../commons/pub-sub/decorators/inject-pub-sub.decorator';
|
import { AmqpConnection, RabbitRPC } from '@golevelup/nestjs-rabbitmq';
|
||||||
import { PubSub } from '../commons/pub-sub/pub-sub';
|
import {
|
||||||
import { observableToAsyncIterable } from '@graphql-tools/utils';
|
EXCHANGE_PIPELINE_TASK_TOPIC,
|
||||||
|
QUEUE_PIPELINE_TASK_DONE,
|
||||||
|
ROUTE_PIPELINE_TASK_DONE,
|
||||||
|
} from './pipeline-tasks.constants';
|
||||||
|
import { PipelineTaskFlushService } from './pipeline-task-flush.service';
|
||||||
|
import { find, isNil, propEq } from 'ramda';
|
||||||
|
import { PipelineTaskLogs } from './models/pipeline-task-logs.model';
|
||||||
|
import { TaskStatuses, terminalTaskStatuses } from './enums/task-statuses.enum';
|
||||||
|
import { InjectPinoLogger, PinoLogger } from 'nestjs-pino';
|
||||||
|
import { getAppInstanceRouteKey } from '../commons/utils/rabbit-mq';
|
||||||
|
import { ROUTE_PIPELINE_TASK_KILL } from './pipeline-tasks.constants';
|
||||||
|
|
||||||
const log = debug('fennec:pipeline-tasks:service');
|
const log = debug('fennec:pipeline-tasks:service');
|
||||||
|
|
||||||
@ -25,42 +28,32 @@ export class PipelineTasksService {
|
|||||||
private readonly repository: Repository<PipelineTask>,
|
private readonly repository: Repository<PipelineTask>,
|
||||||
@InjectRepository(Pipeline)
|
@InjectRepository(Pipeline)
|
||||||
private readonly pipelineRepository: Repository<Pipeline>,
|
private readonly pipelineRepository: Repository<Pipeline>,
|
||||||
@InjectQueue(PIPELINE_TASK_QUEUE)
|
private readonly amqpConnection: AmqpConnection,
|
||||||
private readonly queue: Queue<PipelineTask>,
|
private readonly eventFlushService: PipelineTaskFlushService,
|
||||||
private readonly redis: RedisService,
|
@InjectPinoLogger(PipelineTasksService.name)
|
||||||
@InjectPubSub()
|
private readonly logger: PinoLogger,
|
||||||
private readonly pubSub: PubSub,
|
|
||||||
) {}
|
) {}
|
||||||
async addTask(dto: CreatePipelineTaskInput) {
|
async addTask(dto: CreatePipelineTaskInput) {
|
||||||
const pipeline = await this.pipelineRepository.findOneOrFail({
|
const pipeline = await this.pipelineRepository.findOneOrFail({
|
||||||
where: { id: dto.pipelineId },
|
where: { id: dto.pipelineId },
|
||||||
relations: ['project'],
|
relations: ['project'],
|
||||||
});
|
});
|
||||||
const hasUnfinishedTask = await this.repository
|
// const hasUnfinishedTask = await this.repository
|
||||||
.findOne({
|
// .findOne({
|
||||||
pipelineId: dto.pipelineId,
|
// pipelineId: dto.pipelineId,
|
||||||
commit: dto.commit,
|
// commit: dto.commit,
|
||||||
status: In([TaskStatuses.pending, TaskStatuses.working]),
|
// status: In([TaskStatuses.pending, TaskStatuses.working]),
|
||||||
})
|
// })
|
||||||
.then((val) => !isNil(val));
|
// .then((val) => !isNil(val));
|
||||||
if (hasUnfinishedTask) {
|
// if (hasUnfinishedTask) {
|
||||||
throw new ConflictException(
|
// throw new ConflictException(
|
||||||
'There are the same tasks among the unfinished tasks!',
|
// 'There are the same tasks among the unfinished tasks!',
|
||||||
);
|
// );
|
||||||
}
|
// }
|
||||||
const task = await this.repository.save(this.repository.create(dto));
|
const task = await this.repository.save(this.repository.create(dto));
|
||||||
task.pipeline = pipeline;
|
task.pipeline = pipeline;
|
||||||
|
|
||||||
const tasksKey = this.getRedisTokens(pipeline)[1];
|
this.amqpConnection.publish('new-pipeline-task', 'mac', task);
|
||||||
const redis = this.redis.getClient();
|
|
||||||
await redis.lpush(tasksKey, JSON.stringify(task));
|
|
||||||
log(
|
|
||||||
'add task %s:%s-%s',
|
|
||||||
task.id,
|
|
||||||
task.pipeline.branch,
|
|
||||||
task.commit.slice(0, 6),
|
|
||||||
);
|
|
||||||
await this.doNextTask(pipeline);
|
|
||||||
return task;
|
return task;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -73,65 +66,93 @@ export class PipelineTasksService {
|
|||||||
}
|
}
|
||||||
|
|
||||||
async listTasksByCommitHash(hash: string) {
|
async listTasksByCommitHash(hash: string) {
|
||||||
return await this.repository.find({ commit: hash });
|
return await this.repository.find({
|
||||||
}
|
where: { commit: hash },
|
||||||
|
order: { createdAt: 'DESC' },
|
||||||
async doNextTask(pipeline: Pipeline) {
|
|
||||||
const [lckKey, tasksKey] = this.getRedisTokens(pipeline);
|
|
||||||
const redis = this.redis.getClient();
|
|
||||||
|
|
||||||
log('doNextTask()');
|
|
||||||
const unLck = await new Promise<() => Promise<void>>(
|
|
||||||
async (resolve, reject) => {
|
|
||||||
const lckValue = Date.now().toString();
|
|
||||||
for (let i = 0; i < 5; i++) {
|
|
||||||
if (
|
|
||||||
await redis
|
|
||||||
.set(lckKey, 0, 'EX', lckValue, 'NX')
|
|
||||||
.then(() => true)
|
|
||||||
.catch(() => false)
|
|
||||||
) {
|
|
||||||
resolve(async () => {
|
|
||||||
if ((await redis.get(lckKey)) === lckValue) {
|
|
||||||
await redis.del(lckKey);
|
|
||||||
}
|
|
||||||
});
|
});
|
||||||
return;
|
|
||||||
}
|
|
||||||
await new Promise((resolve) => setTimeout(resolve, 2000));
|
|
||||||
}
|
|
||||||
reject(new LockFailedException(lckKey));
|
|
||||||
},
|
|
||||||
);
|
|
||||||
|
|
||||||
const task = JSON.parse(
|
|
||||||
(await redis.rpop(tasksKey).finally(() => unLck())) ?? 'null',
|
|
||||||
);
|
|
||||||
if (task) {
|
|
||||||
log(
|
|
||||||
'add task (%s:%s-%s) to queue',
|
|
||||||
task.id,
|
|
||||||
task.pipeline.branch,
|
|
||||||
task.commit.slice(0, 6),
|
|
||||||
);
|
|
||||||
await this.queue.add(task);
|
|
||||||
} else {
|
|
||||||
log('task is empty');
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
async updateTask(task: PipelineTask) {
|
|
||||||
this.pubSub.publish(`pipeline-task:${task.id}`, task);
|
|
||||||
return await this.repository.save(task);
|
|
||||||
}
|
|
||||||
|
|
||||||
async watchTaskUpdated(id: string) {
|
|
||||||
return observableToAsyncIterable(
|
|
||||||
this.pubSub.message$(`pipeline-task:${id}`),
|
|
||||||
);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
getRedisTokens(pipeline: Pipeline): [string, string] {
|
getRedisTokens(pipeline: Pipeline): [string, string] {
|
||||||
return [`pipeline-${pipeline.id}:lck`, `pipeline-${pipeline.id}:tasks`];
|
return [`pipeline-${pipeline.id}:lck`, `pipeline-${pipeline.id}:tasks`];
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@RabbitRPC({
|
||||||
|
exchange: EXCHANGE_PIPELINE_TASK_TOPIC,
|
||||||
|
routingKey: ROUTE_PIPELINE_TASK_DONE,
|
||||||
|
queue: QUEUE_PIPELINE_TASK_DONE,
|
||||||
|
queueOptions: {
|
||||||
|
autoDelete: true,
|
||||||
|
durable: true,
|
||||||
|
},
|
||||||
|
})
|
||||||
|
async updateByEvent({ taskId, runOn }: { taskId: string; runOn: string }) {
|
||||||
|
try {
|
||||||
|
const [events, task] = await Promise.all([
|
||||||
|
this.eventFlushService.read(taskId),
|
||||||
|
this.findTaskById(taskId),
|
||||||
|
]);
|
||||||
|
this.logger.info('[updateByEvent] start. taskId: %s', taskId);
|
||||||
|
|
||||||
|
for (const event of events) {
|
||||||
|
if (isNil(event.unit)) {
|
||||||
|
if (
|
||||||
|
event.status !== TaskStatuses.pending &&
|
||||||
|
task.status === TaskStatuses.pending
|
||||||
|
) {
|
||||||
|
task.startedAt = event.emittedAt;
|
||||||
|
} else if (terminalTaskStatuses.includes(event.status)) {
|
||||||
|
task.endedAt = event.emittedAt;
|
||||||
|
}
|
||||||
|
task.status = event.status;
|
||||||
|
} else {
|
||||||
|
let l: PipelineTaskLogs = find<PipelineTaskLogs>(
|
||||||
|
propEq('unit', event.unit),
|
||||||
|
task.logs,
|
||||||
|
);
|
||||||
|
|
||||||
|
if (isNil(l)) {
|
||||||
|
l = {
|
||||||
|
unit: event.unit,
|
||||||
|
startedAt: event.emittedAt,
|
||||||
|
endedAt: null,
|
||||||
|
logs: event.message,
|
||||||
|
status: event.status,
|
||||||
|
};
|
||||||
|
|
||||||
|
task.logs.push(l);
|
||||||
|
} else {
|
||||||
|
l.logs += event.message;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (terminalTaskStatuses.includes(event.status)) {
|
||||||
|
l.endedAt = event.emittedAt;
|
||||||
|
}
|
||||||
|
l.status = event.status;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
task.runOn = runOn;
|
||||||
|
await this.repository.update({ id: taskId }, task);
|
||||||
|
this.logger.info('[updateByEvent] success. taskId: %s', taskId);
|
||||||
|
return task;
|
||||||
|
} catch (error) {
|
||||||
|
this.logger.error(
|
||||||
|
{ error },
|
||||||
|
'[updateByEvent] failed. taskId: %s',
|
||||||
|
taskId,
|
||||||
|
);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async stopTask(task: PipelineTask) {
|
||||||
|
if (isNil(task.runOn)) {
|
||||||
|
throw new BadRequestException(
|
||||||
|
"the task have not running instance on database. field 'runOn' is nil",
|
||||||
|
);
|
||||||
|
}
|
||||||
|
await this.amqpConnection.request({
|
||||||
|
exchange: EXCHANGE_PIPELINE_TASK_TOPIC,
|
||||||
|
routingKey: getAppInstanceRouteKey(ROUTE_PIPELINE_TASK_KILL, task.runOn),
|
||||||
|
payload: task,
|
||||||
|
});
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
@ -1,3 +1,4 @@
|
|||||||
|
import { Query } from '@nestjs/graphql';
|
||||||
import {
|
import {
|
||||||
Args,
|
Args,
|
||||||
Parent,
|
Parent,
|
||||||
@ -6,28 +7,36 @@ import {
|
|||||||
Subscription,
|
Subscription,
|
||||||
} from '@nestjs/graphql';
|
} from '@nestjs/graphql';
|
||||||
import { PipelineTasksService } from '../pipeline-tasks/pipeline-tasks.service';
|
import { PipelineTasksService } from '../pipeline-tasks/pipeline-tasks.service';
|
||||||
import { LogFields, LogList } from '../repos/dtos/log-list.model';
|
import { Commit, LogFields } from '../repos/dtos/log-list.model';
|
||||||
import { PipelinesService } from './pipelines.service';
|
import { PipelinesService } from './pipelines.service';
|
||||||
|
|
||||||
@Resolver(() => LogFields)
|
@Resolver(() => Commit)
|
||||||
export class CommitLogsResolver {
|
export class CommitLogsResolver {
|
||||||
constructor(
|
constructor(
|
||||||
private readonly service: PipelinesService,
|
private readonly service: PipelinesService,
|
||||||
private readonly taskServices: PipelineTasksService,
|
private readonly taskServices: PipelineTasksService,
|
||||||
) {}
|
) {}
|
||||||
@Subscription(() => LogList, {
|
@Subscription(() => String, { resolve: (val) => val, nullable: true })
|
||||||
resolve: (value) => {
|
async syncCommits(
|
||||||
return value;
|
@Args('pipelineId', { type: () => String })
|
||||||
},
|
pipelineId: string,
|
||||||
})
|
@Args('appInstance', { type: () => String, nullable: true })
|
||||||
async listLogsForPipeline(@Args('id', { type: () => String }) id: string) {
|
appInstance?: string,
|
||||||
const job = await this.service.listLogsForPipeline(id);
|
) {
|
||||||
|
const pipeline = await this.service.findOneWithProject(pipelineId);
|
||||||
|
const syncCommitsPromise = this.service.syncCommits(pipeline, appInstance);
|
||||||
return (async function* () {
|
return (async function* () {
|
||||||
yield await job.finished();
|
yield await syncCommitsPromise;
|
||||||
})();
|
})();
|
||||||
}
|
}
|
||||||
@ResolveField()
|
@ResolveField()
|
||||||
async tasks(@Parent() commit: LogFields) {
|
async tasks(@Parent() commit: LogFields) {
|
||||||
return await this.taskServices.listTasksByCommitHash(commit.hash);
|
return await this.taskServices.listTasksByCommitHash(commit.hash);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Query(() => [Commit], { nullable: true })
|
||||||
|
async commits(@Args('pipelineId', { type: () => String }) id: string) {
|
||||||
|
const pipeline = await this.service.findOneWithProject(id);
|
||||||
|
return await this.service.listCommits(pipeline);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
@ -1,5 +1,9 @@
|
|||||||
import { InputType } from '@nestjs/graphql';
|
import { InputType, OmitType } from '@nestjs/graphql';
|
||||||
import { CreatePipelineInput } from './create-pipeline.input';
|
import { CreatePipelineInput } from './create-pipeline.input';
|
||||||
|
|
||||||
@InputType()
|
@InputType()
|
||||||
export class UpdatePipelineInput extends CreatePipelineInput {}
|
export class UpdatePipelineInput extends OmitType(CreatePipelineInput, [
|
||||||
|
'projectId',
|
||||||
|
]) {
|
||||||
|
id: string;
|
||||||
|
}
|
||||||
|
@ -3,18 +3,24 @@ import { PipelinesResolver } from './pipelines.resolver';
|
|||||||
import { PipelinesService } from './pipelines.service';
|
import { PipelinesService } from './pipelines.service';
|
||||||
import { TypeOrmModule } from '@nestjs/typeorm';
|
import { TypeOrmModule } from '@nestjs/typeorm';
|
||||||
import { Pipeline } from './pipeline.entity';
|
import { Pipeline } from './pipeline.entity';
|
||||||
import { BullModule } from '@nestjs/bull';
|
|
||||||
import { LIST_LOGS_TASK } from '../repos/repos.constants';
|
|
||||||
import { CommitLogsResolver } from './commit-logs.resolver';
|
import { CommitLogsResolver } from './commit-logs.resolver';
|
||||||
import { PipelineTasksModule } from '../pipeline-tasks/pipeline-tasks.module';
|
import { PipelineTasksModule } from '../pipeline-tasks/pipeline-tasks.module';
|
||||||
|
import { ReposModule } from '../repos/repos.module';
|
||||||
|
import { RabbitMQModule } from '@golevelup/nestjs-rabbitmq';
|
||||||
|
import { ConfigModule, ConfigService } from '@nestjs/config';
|
||||||
|
|
||||||
@Module({
|
@Module({
|
||||||
imports: [
|
imports: [
|
||||||
TypeOrmModule.forFeature([Pipeline]),
|
TypeOrmModule.forFeature([Pipeline]),
|
||||||
BullModule.registerQueue({
|
|
||||||
name: LIST_LOGS_TASK,
|
|
||||||
}),
|
|
||||||
PipelineTasksModule,
|
PipelineTasksModule,
|
||||||
|
RabbitMQModule.forRootAsync(RabbitMQModule, {
|
||||||
|
imports: [ConfigModule],
|
||||||
|
useFactory: (configService: ConfigService) => ({
|
||||||
|
uri: configService.get<string>('db.rabbitmq.uri'),
|
||||||
|
exchanges: [],
|
||||||
|
}),
|
||||||
|
inject: [ConfigService],
|
||||||
|
}),
|
||||||
],
|
],
|
||||||
providers: [PipelinesResolver, PipelinesService, CommitLogsResolver],
|
providers: [PipelinesResolver, PipelinesService, CommitLogsResolver],
|
||||||
})
|
})
|
||||||
|
@ -9,30 +9,29 @@ import { ListPipelineArgs } from './dtos/list-pipelines.args';
|
|||||||
export class PipelinesResolver {
|
export class PipelinesResolver {
|
||||||
constructor(private readonly service: PipelinesService) {}
|
constructor(private readonly service: PipelinesService) {}
|
||||||
@Query(() => [Pipeline])
|
@Query(() => [Pipeline])
|
||||||
async listPipelines(@Args() dto: ListPipelineArgs) {
|
async pipelines(@Args() dto: ListPipelineArgs) {
|
||||||
return await this.service.list(dto);
|
return await this.service.list(dto);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Query(() => Pipeline)
|
@Query(() => Pipeline)
|
||||||
async findPipeline(@Args('id', { type: () => String }) id: string) {
|
async pipeline(@Args('id', { type: () => String }) id: string) {
|
||||||
return await this.service.findOne(id);
|
return await this.service.findOne(id);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Mutation(() => Pipeline)
|
@Mutation(() => Pipeline)
|
||||||
async createPipeline(
|
async createPipeline(
|
||||||
@Args('pipeline', { type: () => CreatePipelineInput })
|
@Args('pipeline', { type: () => CreatePipelineInput })
|
||||||
dto: UpdatePipelineInput,
|
dto: CreatePipelineInput,
|
||||||
) {
|
) {
|
||||||
return await this.service.create(dto);
|
return await this.service.create(dto);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Mutation(() => Pipeline)
|
@Mutation(() => Pipeline)
|
||||||
async modifyPipeline(
|
async updatePipeline(
|
||||||
@Args('id', { type: () => String }) id: string,
|
@Args('pipeline', { type: () => UpdatePipelineInput })
|
||||||
@Args('Pipeline', { type: () => UpdatePipelineInput })
|
|
||||||
dto: UpdatePipelineInput,
|
dto: UpdatePipelineInput,
|
||||||
) {
|
) {
|
||||||
const tmp = await this.service.update(id, dto);
|
const tmp = await this.service.update(dto);
|
||||||
console.log(tmp);
|
console.log(tmp);
|
||||||
return tmp;
|
return tmp;
|
||||||
}
|
}
|
||||||
|
@ -2,18 +2,14 @@ import { Test, TestingModule } from '@nestjs/testing';
|
|||||||
import { PipelinesService } from './pipelines.service';
|
import { PipelinesService } from './pipelines.service';
|
||||||
import { Pipeline } from './pipeline.entity';
|
import { Pipeline } from './pipeline.entity';
|
||||||
import { getRepositoryToken } from '@nestjs/typeorm';
|
import { getRepositoryToken } from '@nestjs/typeorm';
|
||||||
import { getQueueToken } from '@nestjs/bull';
|
|
||||||
import { LIST_LOGS_TASK } from '../repos/repos.constants';
|
|
||||||
import { Repository } from 'typeorm';
|
import { Repository } from 'typeorm';
|
||||||
import { Project } from '../projects/project.entity';
|
import { Project } from '../projects/project.entity';
|
||||||
import { Job, Queue } from 'bull';
|
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
|
||||||
import { ListLogsOption } from '../repos/models/list-logs.options';
|
|
||||||
|
|
||||||
describe('PipelinesService', () => {
|
describe('PipelinesService', () => {
|
||||||
let service: PipelinesService;
|
let service: PipelinesService;
|
||||||
let repository: Repository<Pipeline>;
|
let repository: Repository<Pipeline>;
|
||||||
let pipeline: Pipeline;
|
let pipeline: Pipeline;
|
||||||
let queue: Queue<ListLogsOption>;
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
beforeEach(async () => {
|
||||||
pipeline = Object.assign(new Pipeline(), {
|
pipeline = Object.assign(new Pipeline(), {
|
||||||
@ -37,33 +33,17 @@ describe('PipelinesService', () => {
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
provide: getQueueToken(LIST_LOGS_TASK),
|
provide: AmqpConnection,
|
||||||
useValue: {
|
useValue: {},
|
||||||
add: jest.fn().mockImplementation(() => ({ id: 1 } as Job)),
|
|
||||||
},
|
|
||||||
},
|
},
|
||||||
],
|
],
|
||||||
}).compile();
|
}).compile();
|
||||||
|
|
||||||
service = module.get<PipelinesService>(PipelinesService);
|
service = module.get<PipelinesService>(PipelinesService);
|
||||||
repository = module.get(getRepositoryToken(Pipeline));
|
repository = module.get(getRepositoryToken(Pipeline));
|
||||||
queue = module.get(getQueueToken(LIST_LOGS_TASK));
|
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should be defined', () => {
|
it('should be defined', () => {
|
||||||
expect(service).toBeDefined();
|
expect(service).toBeDefined();
|
||||||
});
|
});
|
||||||
|
|
||||||
describe('listLogsForPipeline', () => {
|
|
||||||
it('should send task to queue.', async () => {
|
|
||||||
const add = jest.spyOn(queue, 'add');
|
|
||||||
await expect(
|
|
||||||
service.listLogsForPipeline('test-pipeline'),
|
|
||||||
).resolves.toEqual({ id: 1 });
|
|
||||||
expect(add).toBeCalledWith({
|
|
||||||
project: pipeline.project,
|
|
||||||
branch: pipeline.branch,
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
});
|
||||||
|
@ -6,10 +6,16 @@ import { BaseDbService } from '../commons/services/base-db.service';
|
|||||||
import { CreatePipelineInput } from './dtos/create-pipeline.input';
|
import { CreatePipelineInput } from './dtos/create-pipeline.input';
|
||||||
import { UpdatePipelineInput } from './dtos/update-pipeline.input';
|
import { UpdatePipelineInput } from './dtos/update-pipeline.input';
|
||||||
import { ListPipelineArgs } from './dtos/list-pipelines.args';
|
import { ListPipelineArgs } from './dtos/list-pipelines.args';
|
||||||
import { InjectQueue } from '@nestjs/bull';
|
import {
|
||||||
import { LIST_LOGS_TASK } from '../repos/repos.constants';
|
EXCHANGE_REPO,
|
||||||
import { Queue } from 'bull';
|
ROUTE_FETCH,
|
||||||
import { ListLogsOption } from '../repos/models/list-logs.options';
|
ROUTE_LIST_COMMITS,
|
||||||
|
} from '../repos/repos.constants';
|
||||||
|
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
|
||||||
|
import { Commit } from '../repos/dtos/log-list.model';
|
||||||
|
import { getAppInstanceRouteKey } from '../commons/utils/rabbit-mq';
|
||||||
|
import { ApplicationException } from '../commons/exceptions/application.exception';
|
||||||
|
import { plainToClass } from 'class-transformer';
|
||||||
|
|
||||||
@Injectable()
|
@Injectable()
|
||||||
export class PipelinesService extends BaseDbService<Pipeline> {
|
export class PipelinesService extends BaseDbService<Pipeline> {
|
||||||
@ -17,8 +23,7 @@ export class PipelinesService extends BaseDbService<Pipeline> {
|
|||||||
constructor(
|
constructor(
|
||||||
@InjectRepository(Pipeline)
|
@InjectRepository(Pipeline)
|
||||||
readonly repository: Repository<Pipeline>,
|
readonly repository: Repository<Pipeline>,
|
||||||
@InjectQueue(LIST_LOGS_TASK)
|
private readonly amqpConnection: AmqpConnection,
|
||||||
private readonly listLogsQueue: Queue<ListLogsOption>,
|
|
||||||
) {
|
) {
|
||||||
super(repository);
|
super(repository);
|
||||||
}
|
}
|
||||||
@ -26,30 +31,48 @@ export class PipelinesService extends BaseDbService<Pipeline> {
|
|||||||
return this.repository.find(dto);
|
return this.repository.find(dto);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
async findOneWithProject(id: string) {
|
||||||
|
return await this.repository.findOne({
|
||||||
|
where: { id },
|
||||||
|
relations: ['project'],
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
async create(dto: CreatePipelineInput) {
|
async create(dto: CreatePipelineInput) {
|
||||||
await this.isDuplicateEntity(dto);
|
await this.isDuplicateEntity(dto);
|
||||||
return await this.repository.save(this.repository.create(dto));
|
return await this.repository.save(this.repository.create(dto));
|
||||||
}
|
}
|
||||||
|
|
||||||
async update(id: string, dto: UpdatePipelineInput) {
|
async update(dto: UpdatePipelineInput) {
|
||||||
await this.isDuplicateEntityForUpdate(id, dto);
|
const old = await this.findOne(dto.id);
|
||||||
const old = await this.findOne(id);
|
await this.isDuplicateEntityForUpdate(old, dto);
|
||||||
return await this.repository.save(this.repository.merge(old, dto));
|
return await this.repository.save(this.repository.merge(old, dto));
|
||||||
}
|
}
|
||||||
|
|
||||||
async remove(id: string) {
|
async remove(id: string) {
|
||||||
return (await this.repository.softDelete({ id })).affected;
|
return (await this.repository.softDelete({ id })).affected;
|
||||||
}
|
}
|
||||||
|
async syncCommits(pipeline: Pipeline, appInstance?: string) {
|
||||||
async listLogsForPipeline(id: string) {
|
return await this.amqpConnection.request<string | null>({
|
||||||
const pipeline = await this.repository.findOneOrFail({
|
exchange: EXCHANGE_REPO,
|
||||||
where: { id },
|
routingKey: getAppInstanceRouteKey(ROUTE_FETCH, appInstance),
|
||||||
relations: ['project'],
|
payload: pipeline,
|
||||||
|
timeout: 120_000,
|
||||||
});
|
});
|
||||||
const job = await this.listLogsQueue.add({
|
}
|
||||||
project: pipeline.project,
|
async listCommits(pipeline: Pipeline) {
|
||||||
branch: pipeline.branch,
|
return await this.amqpConnection
|
||||||
|
.request<[Error, Commit[]]>({
|
||||||
|
exchange: EXCHANGE_REPO,
|
||||||
|
routingKey: ROUTE_LIST_COMMITS,
|
||||||
|
payload: pipeline,
|
||||||
|
timeout: 30_000,
|
||||||
|
})
|
||||||
|
.then(([error, list]) => {
|
||||||
|
if (error) {
|
||||||
|
throw new ApplicationException(error);
|
||||||
|
}
|
||||||
|
return plainToClass(Commit, list);
|
||||||
});
|
});
|
||||||
return job;
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -21,7 +21,10 @@ export class CreateProjectInput {
|
|||||||
comment: string;
|
comment: string;
|
||||||
|
|
||||||
@Matches(
|
@Matches(
|
||||||
/^(?:ssh:\/\/)?(?:[\w\d-_]+@)(?:[\w\d-_]+\.)*\w{2,10}(?::\d{1,5})?(?:\/[\w\d-_.]+)*/,
|
/^(?:ssh:\/\/)?(?:[\w\d-_]+@)?(?:[\w\d-_]+\.)*\w{2,10}(?::\d{1,5})?(?:\/[\w\d-_.]+)*/,
|
||||||
|
{
|
||||||
|
message: 'wrong ssh url',
|
||||||
|
},
|
||||||
)
|
)
|
||||||
@MaxLength(256)
|
@MaxLength(256)
|
||||||
sshUrl: string;
|
sshUrl: string;
|
||||||
|
@ -1,5 +1,9 @@
|
|||||||
import { InputType } from '@nestjs/graphql';
|
import { InputType } from '@nestjs/graphql';
|
||||||
|
import { IsUUID } from 'class-validator';
|
||||||
import { CreateProjectInput } from './create-project.input';
|
import { CreateProjectInput } from './create-project.input';
|
||||||
|
|
||||||
@InputType()
|
@InputType()
|
||||||
export class UpdateProjectInput extends CreateProjectInput {}
|
export class UpdateProjectInput extends CreateProjectInput {
|
||||||
|
@IsUUID()
|
||||||
|
id: string;
|
||||||
|
}
|
||||||
|
3
src/projects/projects.constants.ts
Normal file
3
src/projects/projects.constants.ts
Normal file
@ -0,0 +1,3 @@
|
|||||||
|
export const EXCHANGE_PROJECT_TOPIC = 'project.topic';
|
||||||
|
export const EXCHANGE_PROJECT_FANOUT = 'project.fanout';
|
||||||
|
export const ROUTE_PROJECT_CHANGE = 'project-change';
|
@ -3,9 +3,31 @@ import { ProjectsService } from './projects.service';
|
|||||||
import { ProjectsResolver } from './projects.resolver';
|
import { ProjectsResolver } from './projects.resolver';
|
||||||
import { TypeOrmModule } from '@nestjs/typeorm';
|
import { TypeOrmModule } from '@nestjs/typeorm';
|
||||||
import { Project } from './project.entity';
|
import { Project } from './project.entity';
|
||||||
|
import { RabbitMQModule } from '@golevelup/nestjs-rabbitmq';
|
||||||
|
import { ConfigModule, ConfigService } from '@nestjs/config';
|
||||||
|
import { EXCHANGE_PROJECT_FANOUT } from './projects.constants';
|
||||||
|
|
||||||
@Module({
|
@Module({
|
||||||
imports: [TypeOrmModule.forFeature([Project])],
|
imports: [
|
||||||
|
TypeOrmModule.forFeature([Project]),
|
||||||
|
RabbitMQModule.forRootAsync(RabbitMQModule, {
|
||||||
|
imports: [ConfigModule],
|
||||||
|
useFactory: (configService: ConfigService) => ({
|
||||||
|
uri: configService.get<string>('db.rabbitmq.uri'),
|
||||||
|
exchanges: [
|
||||||
|
{
|
||||||
|
name: EXCHANGE_PROJECT_FANOUT,
|
||||||
|
type: 'fanout',
|
||||||
|
options: {
|
||||||
|
durable: false,
|
||||||
|
autoDelete: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
],
|
||||||
|
}),
|
||||||
|
inject: [ConfigService],
|
||||||
|
}),
|
||||||
|
],
|
||||||
providers: [ProjectsService, ProjectsResolver],
|
providers: [ProjectsService, ProjectsResolver],
|
||||||
exports: [ProjectsService],
|
exports: [ProjectsService],
|
||||||
})
|
})
|
||||||
|
@ -8,12 +8,12 @@ import { ProjectsService } from './projects.service';
|
|||||||
export class ProjectsResolver {
|
export class ProjectsResolver {
|
||||||
constructor(private readonly service: ProjectsService) {}
|
constructor(private readonly service: ProjectsService) {}
|
||||||
@Query(() => [Project])
|
@Query(() => [Project])
|
||||||
async findProjects() {
|
async projects() {
|
||||||
return await this.service.list();
|
return await this.service.list();
|
||||||
}
|
}
|
||||||
|
|
||||||
@Query(() => Project)
|
@Query(() => Project)
|
||||||
async findProject(@Args('id', { type: () => String }) id: string) {
|
async project(@Args('id', { type: () => String }) id: string) {
|
||||||
return await this.service.findOne(id);
|
return await this.service.findOne(id);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -26,18 +26,17 @@ export class ProjectsResolver {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Mutation(() => Project)
|
@Mutation(() => Project)
|
||||||
async modifyProject(
|
async updateProject(
|
||||||
@Args('id', { type: () => String }) id: string,
|
|
||||||
@Args('project', { type: () => UpdateProjectInput })
|
@Args('project', { type: () => UpdateProjectInput })
|
||||||
dto: UpdateProjectInput,
|
dto: UpdateProjectInput,
|
||||||
) {
|
) {
|
||||||
const tmp = await this.service.update(id, dto);
|
const tmp = await this.service.update(dto);
|
||||||
console.log(tmp);
|
console.log(tmp);
|
||||||
return tmp;
|
return tmp;
|
||||||
}
|
}
|
||||||
|
|
||||||
@Mutation(() => Number)
|
@Mutation(() => Number)
|
||||||
async deleteProject(@Args('id', { type: () => String }) id: string) {
|
async removeProject(@Args('id', { type: () => String }) id: string) {
|
||||||
return await this.service.remove(id);
|
return await this.service.remove(id);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -2,6 +2,7 @@ import { Test, TestingModule } from '@nestjs/testing';
|
|||||||
import { ProjectsService } from './projects.service';
|
import { ProjectsService } from './projects.service';
|
||||||
import { getRepositoryToken } from '@nestjs/typeorm';
|
import { getRepositoryToken } from '@nestjs/typeorm';
|
||||||
import { Project } from './project.entity';
|
import { Project } from './project.entity';
|
||||||
|
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
|
||||||
|
|
||||||
describe('ProjectsService', () => {
|
describe('ProjectsService', () => {
|
||||||
let service: ProjectsService;
|
let service: ProjectsService;
|
||||||
@ -14,6 +15,10 @@ describe('ProjectsService', () => {
|
|||||||
provide: getRepositoryToken(Project),
|
provide: getRepositoryToken(Project),
|
||||||
useValue: {},
|
useValue: {},
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
provide: AmqpConnection,
|
||||||
|
useValue: {},
|
||||||
|
},
|
||||||
],
|
],
|
||||||
}).compile();
|
}).compile();
|
||||||
|
|
||||||
|
@ -5,6 +5,11 @@ import { Repository } from 'typeorm';
|
|||||||
import { CreateProjectInput } from './dtos/create-project.input';
|
import { CreateProjectInput } from './dtos/create-project.input';
|
||||||
import { Project } from './project.entity';
|
import { Project } from './project.entity';
|
||||||
import { UpdateProjectInput } from './dtos/update-project.input';
|
import { UpdateProjectInput } from './dtos/update-project.input';
|
||||||
|
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
|
||||||
|
import {
|
||||||
|
EXCHANGE_PROJECT_FANOUT,
|
||||||
|
ROUTE_PROJECT_CHANGE,
|
||||||
|
} from './projects.constants';
|
||||||
|
|
||||||
@Injectable()
|
@Injectable()
|
||||||
export class ProjectsService extends BaseDbService<Project> {
|
export class ProjectsService extends BaseDbService<Project> {
|
||||||
@ -12,6 +17,7 @@ export class ProjectsService extends BaseDbService<Project> {
|
|||||||
constructor(
|
constructor(
|
||||||
@InjectRepository(Project)
|
@InjectRepository(Project)
|
||||||
readonly repository: Repository<Project>,
|
readonly repository: Repository<Project>,
|
||||||
|
private readonly amqpConnection: AmqpConnection,
|
||||||
) {
|
) {
|
||||||
super(repository);
|
super(repository);
|
||||||
}
|
}
|
||||||
@ -25,10 +31,15 @@ export class ProjectsService extends BaseDbService<Project> {
|
|||||||
return await this.repository.save(this.repository.create(dto));
|
return await this.repository.save(this.repository.create(dto));
|
||||||
}
|
}
|
||||||
|
|
||||||
async update(id: string, dto: UpdateProjectInput) {
|
async update(dto: UpdateProjectInput) {
|
||||||
await this.isDuplicateEntityForUpdate(id, dto);
|
await this.isDuplicateEntityForUpdate(dto.id, dto);
|
||||||
const old = await this.findOne(id);
|
const old = await this.findOne(dto.id);
|
||||||
return await this.repository.save(this.repository.merge(old, dto));
|
const project = await this.repository.save(this.repository.merge(old, dto));
|
||||||
|
this.amqpConnection.publish(EXCHANGE_PROJECT_FANOUT, ROUTE_PROJECT_CHANGE, [
|
||||||
|
project,
|
||||||
|
old,
|
||||||
|
]);
|
||||||
|
return project;
|
||||||
}
|
}
|
||||||
|
|
||||||
async remove(id: string) {
|
async remove(id: string) {
|
||||||
|
@ -1,7 +1,20 @@
|
|||||||
import { ObjectType, Field } from '@nestjs/graphql';
|
import { ObjectType, Field } from '@nestjs/graphql';
|
||||||
|
import { Type } from 'class-transformer';
|
||||||
import { LogResult, DefaultLogFields } from 'simple-git';
|
import { LogResult, DefaultLogFields } from 'simple-git';
|
||||||
import { PipelineTask } from '../../pipeline-tasks/pipeline-task.entity';
|
import { PipelineTask } from '../../pipeline-tasks/pipeline-task.entity';
|
||||||
|
|
||||||
|
@ObjectType()
|
||||||
|
export class Commit {
|
||||||
|
hash: string;
|
||||||
|
@Type(() => Date)
|
||||||
|
date: Date;
|
||||||
|
message: string;
|
||||||
|
refs: string;
|
||||||
|
body: string;
|
||||||
|
author_name: string;
|
||||||
|
author_email: string;
|
||||||
|
tasks: PipelineTask[];
|
||||||
|
}
|
||||||
@ObjectType()
|
@ObjectType()
|
||||||
export class LogFields {
|
export class LogFields {
|
||||||
hash: string;
|
hash: string;
|
||||||
|
@ -1,14 +0,0 @@
|
|||||||
import { ReposService } from './repos.service';
|
|
||||||
import { Processor, Process } from '@nestjs/bull';
|
|
||||||
import { Job } from 'bull';
|
|
||||||
import { ListLogsOption } from './models/list-logs.options';
|
|
||||||
import { LIST_LOGS_TASK } from './repos.constants';
|
|
||||||
@Processor(LIST_LOGS_TASK)
|
|
||||||
export class ListLogsConsumer {
|
|
||||||
constructor(private readonly service: ReposService) {}
|
|
||||||
@Process()
|
|
||||||
async listLogs(job: Job<ListLogsOption>) {
|
|
||||||
const logs = await this.service.listLogs(job.data);
|
|
||||||
return logs;
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,3 +1,6 @@
|
|||||||
export const LIST_LOGS_TASK = 'LIST_LOGS_TASK';
|
export const EXCHANGE_REPO = 'fennec.repo';
|
||||||
export const LIST_LOGS_PUB_SUB = 'LIST_LOGS_PUB_SUB';
|
export const ROUTE_FETCH = 'fetch';
|
||||||
export const LIST_LOGS_DONE = 'LIST_LOGS_DONE';
|
export const ROUTE_LIST_COMMITS = 'list-commits';
|
||||||
|
export const QUEUE_LIST_COMMITS = 'list-commits';
|
||||||
|
export const QUEUE_FETCH = 'repo-fetch';
|
||||||
|
export const QUEUE_REFRESH_REPO = 'refresh-repo';
|
||||||
|
@ -3,23 +3,37 @@ import { TypeOrmModule } from '@nestjs/typeorm';
|
|||||||
import { Project } from '../projects/project.entity';
|
import { Project } from '../projects/project.entity';
|
||||||
import { ReposResolver } from './repos.resolver';
|
import { ReposResolver } from './repos.resolver';
|
||||||
import { ReposService } from './repos.service';
|
import { ReposService } from './repos.service';
|
||||||
import { ConfigModule } from '@nestjs/config';
|
import { ConfigModule, ConfigService } from '@nestjs/config';
|
||||||
import { ProjectsModule } from '../projects/projects.module';
|
import { ProjectsModule } from '../projects/projects.module';
|
||||||
import { BullModule } from '@nestjs/bull';
|
import { EXCHANGE_REPO } from './repos.constants';
|
||||||
import { LIST_LOGS_TASK, LIST_LOGS_PUB_SUB } from './repos.constants';
|
import { RabbitMQModule } from '@golevelup/nestjs-rabbitmq';
|
||||||
import { PubSub } from 'graphql-subscriptions';
|
import { CommonsModule } from '../commons/commons.module';
|
||||||
import { ListLogsConsumer } from './list-logs.consumer';
|
|
||||||
|
|
||||||
@Module({
|
@Module({
|
||||||
imports: [
|
imports: [
|
||||||
TypeOrmModule.forFeature([Project]),
|
TypeOrmModule.forFeature([Project]),
|
||||||
ConfigModule,
|
ConfigModule,
|
||||||
ProjectsModule,
|
ProjectsModule,
|
||||||
BullModule.registerQueue({
|
CommonsModule,
|
||||||
name: LIST_LOGS_TASK,
|
RabbitMQModule.forRootAsync(RabbitMQModule, {
|
||||||
|
imports: [ConfigModule],
|
||||||
|
useFactory: (configService: ConfigService) => ({
|
||||||
|
uri: configService.get<string>('db.rabbitmq.uri'),
|
||||||
|
exchanges: [
|
||||||
|
{
|
||||||
|
name: EXCHANGE_REPO,
|
||||||
|
type: 'topic',
|
||||||
|
options: {
|
||||||
|
durable: true,
|
||||||
|
autoDelete: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
],
|
||||||
|
}),
|
||||||
|
inject: [ConfigService],
|
||||||
}),
|
}),
|
||||||
],
|
],
|
||||||
providers: [ReposResolver, ReposService, ListLogsConsumer],
|
providers: [ReposResolver, ReposService],
|
||||||
exports: [ReposService],
|
exports: [ReposService],
|
||||||
})
|
})
|
||||||
export class ReposModule {}
|
export class ReposModule {}
|
||||||
|
@ -9,6 +9,9 @@ import configuration from '../commons/config/configuration';
|
|||||||
import { PipelineTask } from '../pipeline-tasks/pipeline-task.entity';
|
import { PipelineTask } from '../pipeline-tasks/pipeline-task.entity';
|
||||||
import { join } from 'path';
|
import { join } from 'path';
|
||||||
import { readFile } from 'fs/promises';
|
import { readFile } from 'fs/promises';
|
||||||
|
import { getLoggerToken, PinoLogger } from 'nestjs-pino';
|
||||||
|
import { Nack } from '@golevelup/nestjs-rabbitmq';
|
||||||
|
import { getInstanceName } from '../commons/utils/rabbit-mq';
|
||||||
|
|
||||||
const getTest1Project = () =>
|
const getTest1Project = () =>
|
||||||
({
|
({
|
||||||
@ -45,6 +48,10 @@ describe('ReposService', () => {
|
|||||||
provide: getRepositoryToken(Project),
|
provide: getRepositoryToken(Project),
|
||||||
useFactory: repositoryMockFactory,
|
useFactory: repositoryMockFactory,
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
provide: getLoggerToken(ReposService.name),
|
||||||
|
useValue: new PinoLogger({}),
|
||||||
|
},
|
||||||
],
|
],
|
||||||
}).compile();
|
}).compile();
|
||||||
|
|
||||||
@ -139,4 +146,57 @@ describe('ReposService', () => {
|
|||||||
);
|
);
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
|
describe('fetch', () => {
|
||||||
|
it('success', async () => {
|
||||||
|
const project = new Project();
|
||||||
|
const pipeline = new Pipeline();
|
||||||
|
pipeline.branch = 'test';
|
||||||
|
const fetch = jest.fn((_: any) => Promise.resolve());
|
||||||
|
pipeline.project = project;
|
||||||
|
const getGit = jest.spyOn(service, 'getGit').mockImplementation(() =>
|
||||||
|
Promise.resolve({
|
||||||
|
fetch,
|
||||||
|
} as any),
|
||||||
|
);
|
||||||
|
await expect(service.fetch(pipeline)).resolves.toEqual(getInstanceName());
|
||||||
|
expect(getGit).toBeCalledTimes(1);
|
||||||
|
expect(getGit.mock.calls[0]?.[0]).toEqual(project);
|
||||||
|
expect(fetch).toBeCalledTimes(1);
|
||||||
|
expect(fetch.mock.calls[0]?.[0]).toMatchObject([
|
||||||
|
'origin',
|
||||||
|
'test',
|
||||||
|
'--depth=100',
|
||||||
|
]);
|
||||||
|
});
|
||||||
|
it('failed a', async () => {
|
||||||
|
const project = new Project();
|
||||||
|
const pipeline = new Pipeline();
|
||||||
|
pipeline.branch = 'test';
|
||||||
|
const fetch = jest.fn((_: any) => Promise.resolve());
|
||||||
|
pipeline.project = project;
|
||||||
|
const getGit = jest
|
||||||
|
.spyOn(service, 'getGit')
|
||||||
|
.mockImplementation(() => Promise.reject('error'));
|
||||||
|
await expect(service.fetch(pipeline)).resolves.toMatchObject(new Nack());
|
||||||
|
expect(getGit).toBeCalledTimes(1);
|
||||||
|
expect(getGit.mock.calls[0]?.[0]).toEqual(project);
|
||||||
|
expect(fetch).toBeCalledTimes(0);
|
||||||
|
});
|
||||||
|
it('failed b', async () => {
|
||||||
|
const project = new Project();
|
||||||
|
const pipeline = new Pipeline();
|
||||||
|
pipeline.branch = 'test';
|
||||||
|
const fetch = jest.fn((_: any) => Promise.reject('error'));
|
||||||
|
pipeline.project = project;
|
||||||
|
const getGit = jest.spyOn(service, 'getGit').mockImplementation(() =>
|
||||||
|
Promise.resolve({
|
||||||
|
fetch,
|
||||||
|
} as any),
|
||||||
|
);
|
||||||
|
await expect(service.fetch(pipeline)).resolves.toMatchObject(new Nack());
|
||||||
|
expect(getGit).toBeCalledTimes(1);
|
||||||
|
expect(fetch).toBeCalledTimes(1);
|
||||||
|
});
|
||||||
|
});
|
||||||
});
|
});
|
||||||
|
@ -1,5 +1,4 @@
|
|||||||
import { ListLogsOption } from './models/list-logs.options';
|
import { ListLogsOption } from './models/list-logs.options';
|
||||||
import { Pipeline } from './../pipelines/pipeline.entity';
|
|
||||||
import { PipelineTask } from './../pipeline-tasks/pipeline-task.entity';
|
import { PipelineTask } from './../pipeline-tasks/pipeline-task.entity';
|
||||||
import { Injectable, NotFoundException } from '@nestjs/common';
|
import { Injectable, NotFoundException } from '@nestjs/common';
|
||||||
import { InjectRepository } from '@nestjs/typeorm';
|
import { InjectRepository } from '@nestjs/typeorm';
|
||||||
@ -10,8 +9,31 @@ import { gitP } from 'simple-git';
|
|||||||
import { Repository } from 'typeorm';
|
import { Repository } from 'typeorm';
|
||||||
import { Project } from '../projects/project.entity';
|
import { Project } from '../projects/project.entity';
|
||||||
import { ListBranchesArgs } from './dtos/list-branches.args';
|
import { ListBranchesArgs } from './dtos/list-branches.args';
|
||||||
import { ListLogsArgs } from './dtos/list-logs.args';
|
|
||||||
import { ConfigService } from '@nestjs/config';
|
import { ConfigService } from '@nestjs/config';
|
||||||
|
import { Commit } from './dtos/log-list.model';
|
||||||
|
import { Nack, RabbitRPC, RabbitSubscribe } from '@golevelup/nestjs-rabbitmq';
|
||||||
|
import { Pipeline } from '../pipelines/pipeline.entity';
|
||||||
|
import { InjectPinoLogger, PinoLogger } from 'nestjs-pino';
|
||||||
|
import {
|
||||||
|
EXCHANGE_REPO,
|
||||||
|
QUEUE_FETCH,
|
||||||
|
QUEUE_LIST_COMMITS,
|
||||||
|
QUEUE_REFRESH_REPO,
|
||||||
|
ROUTE_FETCH,
|
||||||
|
ROUTE_LIST_COMMITS,
|
||||||
|
} from './repos.constants';
|
||||||
|
import { getSelfInstanceQueueKey } from '../commons/utils/rabbit-mq';
|
||||||
|
import {
|
||||||
|
getInstanceName,
|
||||||
|
getSelfInstanceRouteKey,
|
||||||
|
} from '../commons/utils/rabbit-mq';
|
||||||
|
import { ApplicationException } from '../commons/exceptions/application.exception';
|
||||||
|
import {
|
||||||
|
EXCHANGE_PROJECT_FANOUT,
|
||||||
|
ROUTE_PROJECT_CHANGE,
|
||||||
|
} from '../projects/projects.constants';
|
||||||
|
import { RedisMutexService } from '../commons/redis-mutex/redis-mutex.service';
|
||||||
|
import { rm } from 'fs/promises';
|
||||||
|
|
||||||
const DEFAULT_REMOTE_NAME = 'origin';
|
const DEFAULT_REMOTE_NAME = 'origin';
|
||||||
const INFO_PATH = '@info';
|
const INFO_PATH = '@info';
|
||||||
@ -21,6 +43,9 @@ export class ReposService {
|
|||||||
@InjectRepository(Project)
|
@InjectRepository(Project)
|
||||||
private readonly projectRepository: Repository<Project>,
|
private readonly projectRepository: Repository<Project>,
|
||||||
private readonly configService: ConfigService,
|
private readonly configService: ConfigService,
|
||||||
|
@InjectPinoLogger(ReposService.name)
|
||||||
|
private readonly logger: PinoLogger,
|
||||||
|
private readonly redisMutexService: RedisMutexService,
|
||||||
) {}
|
) {}
|
||||||
|
|
||||||
getWorkspaceRoot(project: Project): string {
|
getWorkspaceRoot(project: Project): string {
|
||||||
@ -31,7 +56,11 @@ export class ReposService {
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
async getGit(project: Project, workspaceRoot?: string) {
|
async getGit(
|
||||||
|
project: Project,
|
||||||
|
workspaceRoot?: string,
|
||||||
|
{ fetch = true } = {},
|
||||||
|
) {
|
||||||
if (!workspaceRoot) {
|
if (!workspaceRoot) {
|
||||||
workspaceRoot = this.getWorkspaceRoot(project);
|
workspaceRoot = this.getWorkspaceRoot(project);
|
||||||
}
|
}
|
||||||
@ -44,7 +73,9 @@ export class ReposService {
|
|||||||
await git.init();
|
await git.init();
|
||||||
await git.addRemote(DEFAULT_REMOTE_NAME, project.sshUrl);
|
await git.addRemote(DEFAULT_REMOTE_NAME, project.sshUrl);
|
||||||
}
|
}
|
||||||
|
if (fetch) {
|
||||||
await git.fetch();
|
await git.fetch();
|
||||||
|
}
|
||||||
return git;
|
return git;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -89,4 +120,113 @@ export class ReposService {
|
|||||||
encodeURIComponent(`${task.pipeline.name}-${task.commit}`),
|
encodeURIComponent(`${task.pipeline.name}-${task.commit}`),
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
async checkout4Task(task: PipelineTask): Promise<string> {
|
||||||
|
const path = this.getWorkspaceRootByTask(task);
|
||||||
|
await this.checkout(task, path);
|
||||||
|
return path;
|
||||||
|
}
|
||||||
|
|
||||||
|
@RabbitRPC({
|
||||||
|
exchange: EXCHANGE_REPO,
|
||||||
|
routingKey: [
|
||||||
|
ROUTE_LIST_COMMITS,
|
||||||
|
getSelfInstanceRouteKey(ROUTE_LIST_COMMITS),
|
||||||
|
],
|
||||||
|
queue: getSelfInstanceQueueKey(QUEUE_LIST_COMMITS),
|
||||||
|
queueOptions: {
|
||||||
|
autoDelete: true,
|
||||||
|
},
|
||||||
|
})
|
||||||
|
async listCommits(pipeline: Pipeline): Promise<[Error, Commit[]?]> {
|
||||||
|
const git = await this.getGit(pipeline.project, undefined, {
|
||||||
|
fetch: false,
|
||||||
|
});
|
||||||
|
try {
|
||||||
|
const data = await git.log([
|
||||||
|
'-100',
|
||||||
|
'--branches',
|
||||||
|
`remotes/origin/${pipeline.branch}`,
|
||||||
|
'--',
|
||||||
|
]);
|
||||||
|
return [
|
||||||
|
null,
|
||||||
|
data.all.map(
|
||||||
|
(it) =>
|
||||||
|
({
|
||||||
|
...it,
|
||||||
|
date: new Date(it.date),
|
||||||
|
} as Commit),
|
||||||
|
),
|
||||||
|
];
|
||||||
|
} catch (error) {
|
||||||
|
this.logger.error(
|
||||||
|
{ error, pipeline },
|
||||||
|
'[listCommits] %s',
|
||||||
|
error?.message,
|
||||||
|
);
|
||||||
|
return [new ApplicationException(error)];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@RabbitRPC({
|
||||||
|
exchange: EXCHANGE_REPO,
|
||||||
|
routingKey: [ROUTE_FETCH, getSelfInstanceRouteKey(ROUTE_FETCH)],
|
||||||
|
queue: getSelfInstanceQueueKey(QUEUE_FETCH),
|
||||||
|
queueOptions: {
|
||||||
|
autoDelete: true,
|
||||||
|
},
|
||||||
|
})
|
||||||
|
async fetch(pipeline: Pipeline): Promise<string | null | Nack> {
|
||||||
|
const unlock = await this.redisMutexService.lock(
|
||||||
|
`repo-project-${pipeline.projectId}`,
|
||||||
|
);
|
||||||
|
try {
|
||||||
|
const git = await this.getGit(pipeline.project, undefined, {
|
||||||
|
fetch: false,
|
||||||
|
});
|
||||||
|
await git.fetch(['origin', pipeline.branch, '--depth=100']);
|
||||||
|
return getInstanceName();
|
||||||
|
} catch (error) {
|
||||||
|
this.logger.error({ error, pipeline }, '[fetch] %s', error?.message);
|
||||||
|
return new Nack();
|
||||||
|
} finally {
|
||||||
|
await unlock();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@RabbitSubscribe({
|
||||||
|
exchange: EXCHANGE_PROJECT_FANOUT,
|
||||||
|
routingKey: ROUTE_PROJECT_CHANGE,
|
||||||
|
queue: QUEUE_REFRESH_REPO,
|
||||||
|
queueOptions: {
|
||||||
|
autoDelete: true,
|
||||||
|
durable: true,
|
||||||
|
},
|
||||||
|
})
|
||||||
|
async refreshRepo([project]: [Project]) {
|
||||||
|
this.logger.info({ project }, '[refreshRepo] start');
|
||||||
|
const unlock = await this.redisMutexService.lock(
|
||||||
|
`repo-project-${project.id}`,
|
||||||
|
{
|
||||||
|
timeout: null,
|
||||||
|
},
|
||||||
|
);
|
||||||
|
try {
|
||||||
|
const path = join(
|
||||||
|
this.configService.get<string>('workspaces.root'),
|
||||||
|
encodeURIComponent(project.name),
|
||||||
|
);
|
||||||
|
await rm(path, { recursive: true });
|
||||||
|
this.logger.info({ project }, '[refreshRepo] success');
|
||||||
|
} catch (error) {
|
||||||
|
this.logger.error(
|
||||||
|
{ project, error },
|
||||||
|
'[refreshRepo] failed. $s',
|
||||||
|
error.message,
|
||||||
|
);
|
||||||
|
} finally {
|
||||||
|
await unlock();
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user